{"adam_epsilon": 1e-08, "alphabet": "gene_prot", "append_eos": true, "asl_gamma_neg": 4.0, "asl_gamma_pos": 1.0, "best_metric_type": "sp_statistic", "beta1": 0.9, "beta2": 0.98, "buffer_size": 4096, "cache_dir": null, "classifier_activate_func": "gelu", "classifier_size": 128, "codes_file": null, "config_path": "../config/luca_base/luca_base_config.json", "cross_atten": false, "dataset_name": "DMS_Bind_Reps_Strain", "dataset_type": "protein", "delete_old": true, "dev_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/dev/", "device": "cuda", "do_eval": true, "do_lower_case": false, "do_metrics": true, "do_predict": true, "do_train": true, "dropout_prob": 0.1, "early_stop_epoch": -1, "emb_activate_func": "gelu", "embedding_complete": true, "embedding_complete_seg_overlap": true, "embedding_fixed_len_a_time": 3072, "embedding_input_size": 2560, "embedding_input_size_a": null, "embedding_input_size_b": null, "eval_all_checkpoints": false, "evaluate_during_training": true, "evaluate_steps": 2000, "evaluate_strategy": "epoch", "fc_activate_func": "gelu", "focal_loss_alpha": 0.7, "focal_loss_gamma": 2.0, "focal_loss_reduce": false, "fp16": false, "fp16_embedding": false, "fp16_opt_level": "O1", "fusion_type": "concat", "gradient_accumulation_steps": 1, "hidden_size": 1024, "ignore_index": -100, "input_mode": "single", "input_type": "seq", "intermediate_size": 4096, "label_filepath": "../dataset/DMS_Bind_Reps_Strain/protein/regression/label.txt", "label_size": 1, "label_type": "DMS_Bind_Reps_Strain", "learning_rate": 0.0001, "llm_dir": "..", "llm_dirpath": "../llm/models/lucaone_virus/v1.0/token_level,span_level,seq_level/lucaone_virus/20240815023346/checkpoint-step3800000", "llm_step": "3800000", "llm_task_level": "token_level,span_level,seq_level", "llm_time_str": "20240815023346", "llm_type": "lucaone_virus", "llm_version": "v1.0", "lmdb_path": null, "local_rank": -1, "log_dir": "../logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/seq/20250410011916", "logging_steps": 10, "loss_reduction": "mean", "loss_type": "l2", "lr_decay_rate": 0.9, "lr_update_strategy": "step", "matrix_add_special_token": true, "matrix_dirpath": "../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000", "matrix_embedding_exists": false, "matrix_encoder": false, "matrix_encoder_act": false, "matrix_fc_size": "128", "matrix_max_length": 100000, "matrix_max_length_a": null, "matrix_max_length_b": null, "matrix_pooling_type": "value_attention", "max_grad_norm": 1.0, "max_sentence_length": null, "max_sentences": null, "max_steps": -1, "model_dirpath": null, "model_type": "luca_base", "n_gpu": 1, "no_cuda": false, "no_position_embeddings": true, "no_token_embeddings": false, "no_token_type_embeddings": true, "non_ignore": false, "not_append_eos": false, "not_matrix_encoder_shared": false, "not_prepend_bos": false, "not_save_emb_to_disk": false, "not_seq_encoder_shared": false, "num_attention_heads": 4, "num_hidden_layers": 2, "num_train_epochs": 50, "output_dir": "../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/seq/20250410011916", "output_mode": "regression", "overwrite_cache": false, "overwrite_output_dir": true, "per_gpu_eval_batch_size": 16, "per_gpu_train_batch_size": 16, "pos_weight": 1.0, "position_embedding_type": "absolute", "prepend_bos": true, "save_all": true, "save_steps": -1, "seed": 1221, "self_atten": false, "seq_fc_size": "null", "seq_max_length": 100000, "seq_max_length_a": null, "seq_max_length_b": null, "seq_pooling_type": "value_attention", "seq_subword": false, "seq_vocab_path": "gene_prot", "sigmoid": false, "task_level_type": "seq_level", "task_type": "regression", "tb_log_dir": "../tb-logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/seq/20250410011916", "test_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/test/", "time_str": "20250410011927", "train_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/train/", "trunc_type": "right", "vector_dirpath": "../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000", "vector_fc_size": "null", "vocab_size": 39, "warmup_steps": 1000, "weight": null, "weight_decay": 0.01, "worker_num": 0} ################################################## n_gpu: 1 ################################################## Inputs: Input Name List: protein,seq ################################################## Encoder Config: {'llm_type': 'lucaone_virus', 'llm_version': 'v1.0', 'llm_step': '3800000', 'llm_dirpath': '../llm/models/lucaone_virus/v1.0/token_level,span_level,seq_level/lucaone_virus/20240815023346/checkpoint-step3800000', 'input_type': 'seq', 'trunc_type': 'right', 'seq_max_length': 100000, 'atom_seq_max_length': None, 'vector_dirpath': '../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000', 'matrix_dirpath': '../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000', 'local_rank': -1, 'max_sentence_length': None, 'max_sentences': None, 'matrix_add_special_token': True, 'embedding_complete': True, 'embedding_complete_seg_overlap': True, 'embedding_fixed_len_a_time': 3072, 'matrix_embedding_exists': False, 'save_emb_to_disk': True, 'fp16_embedding': False} ################################################## Model Config: LucaConfig { "alphabet": "gene_prot", "attention_probs_dropout_prob": 0.1, "classifier_activate_func": "gelu", "classifier_dropout_prob": 0.1, "classifier_size": 128, "cls_token_id": 2, "cross_atten": false, "directionality": "bidi", "emb_activate_func": "gelu", "embedding_input_size": 2560, "fc_activate_func": "gelu", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "LABEL_0" }, "ignore_index": -100, "initializer_range": 0.02, "intermediate_size": 4096, "kernel_size": 7, "label2id": { "LABEL_0": 0 }, "layer_norm_eps": 1e-12, "loss_reduction": "mean", "matrix_fc_size": [ 128 ], "matrix_max_length": 100000, "matrix_pooling_type": "value_attention", "max_position_embeddings": 100000, "no_position_embeddings": true, "no_token_embeddings": false, "no_token_type_embeddings": true, "num_attention_heads": 4, "num_hidden_layers": 2, "pad_token_id": 0, "pos_weight": 1.0, "position_embedding_type": "absolute", "self_atten": false, "sep_token_id": 3, "seq_fc_size": null, "seq_max_length": 100000, "seq_pooling_type": "value_attention", "token_dropout": null, "transformers_version": "4.29.0", "type_vocab_size": 2, "use_luca_layer_norm_v2": true, "vector_fc_size": null, "vocab_size": 39 } ################################################## Mode Architecture: LucaBase( (seq_encoder): LucaTransformer( (embeddings): LucaEmbeddings( (word_embeddings): Embedding(39, 1024, padding_idx=0) (dropout): Dropout(p=0.1, inplace=False) ) (encoder): ModuleList( (0): LucaTransformerLayer( (pre_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True) (self_attn): LucaMultiHeadAttention( (k_proj): Linear(in_features=1024, out_features=1024, bias=True) (v_proj): Linear(in_features=1024, out_features=1024, bias=True) (q_proj): Linear(in_features=1024, out_features=1024, bias=True) (out_proj): Linear(in_features=1024, out_features=1024, bias=True) (rot_emb): RotaryEmbedding() ) (post_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True) (fc1): Linear(in_features=1024, out_features=4096, bias=True) (fc2): Linear(in_features=4096, out_features=1024, bias=True) ) (1): LucaTransformerLayer( (pre_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True) (self_attn): LucaMultiHeadAttention( (k_proj): Linear(in_features=1024, out_features=1024, bias=True) (v_proj): Linear(in_features=1024, out_features=1024, bias=True) (q_proj): Linear(in_features=1024, out_features=1024, bias=True) (out_proj): Linear(in_features=1024, out_features=1024, bias=True) (rot_emb): RotaryEmbedding() ) (post_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True) (fc1): Linear(in_features=1024, out_features=4096, bias=True) (fc2): Linear(in_features=4096, out_features=1024, bias=True) ) ) (last_layer_norm): LayerNorm((1024,), eps=1e-05, elementwise_affine=True) ) (seq_pooler): GlobalMaskValueAttentionPooling1D (1024 -> 1024) (dropout): Dropout(p=0.1, inplace=False) (hidden_layer): Linear(in_features=1024, out_features=128, bias=True) (hidden_act): GELU(approximate='none') (classifier): Linear(in_features=128, out_features=1, bias=True) (loss_fct): MaskedMSELoss( (criterion): MSELoss() ) ) ################################################## Model parameters: 28511489 ################################################## {"total_num": "27.190000M", "total_size": "108.760000MB", "param_sum": "27.190000M", "param_size": "108.760000MB", "buffer_sum": "0.000000M", "buffer_size": "0.000000MB", "trainable_num": "27.190675M", "trainable_size": "108.762699MB"} ################################################## Train dataset len: 34055, batch size: 16, batch num: 2129 Train dataset t_total: 106450, max_steps: -1 ***** Running training ***** Train Dataset Num examples = 34055 Train Dataset Num Epochs = 50 Logging Steps = 10 Saving Steps = -1 Evaluating Strategy = epoch Train Dataset Instantaneous batch size per GPU = 16 Train Dataset Total train batch size (w. parallel, distributed & accumulation) = 16 Train Dataset Gradient Accumulation steps = 1 Train Dataset Total optimization steps = 106450 ################################################## Training, Epoch: 0001, Batch: 000010, Sample Num: 160, Cur Loss: 71.14894867, Cur Avg Loss: 73.03455353, Log Avg loss: 73.03455353, Global Avg Loss: 73.03455353, Time: 0.0522 Steps: 10, Updated lr: 0.000001 Training, Epoch: 0001, Batch: 000020, Sample Num: 320, Cur Loss: 69.24745941, Cur Avg Loss: 71.03140621, Log Avg loss: 69.02825890, Global Avg Loss: 71.03140621, Time: 0.0517 Steps: 20, Updated lr: 0.000002 Training, Epoch: 0001, Batch: 000030, Sample Num: 480, Cur Loss: 62.20997620, Cur Avg Loss: 69.86681010, Log Avg loss: 67.53761787, Global Avg Loss: 69.86681010, Time: 0.0550 Steps: 30, Updated lr: 0.000003 Training, Epoch: 0001, Batch: 000040, Sample Num: 640, Cur Loss: 49.46105957, Cur Avg Loss: 66.31812353, Log Avg loss: 55.67206383, Global Avg Loss: 66.31812353, Time: 0.0514 Steps: 40, Updated lr: 0.000004 Training, Epoch: 0001, Batch: 000050, Sample Num: 800, Cur Loss: 39.33003235, Cur Avg Loss: 62.78356766, Log Avg loss: 48.64534416, Global Avg Loss: 62.78356766, Time: 0.0520 Steps: 50, Updated lr: 0.000005 Training, Epoch: 0001, Batch: 000060, Sample Num: 960, Cur Loss: 43.09672165, Cur Avg Loss: 59.52227370, Log Avg loss: 43.21580391, Global Avg Loss: 59.52227370, Time: 0.0606 Steps: 60, Updated lr: 0.000006 Training, Epoch: 0001, Batch: 000070, Sample Num: 1120, Cur Loss: 32.23752213, Cur Avg Loss: 56.41950602, Log Avg loss: 37.80289993, Global Avg Loss: 56.41950602, Time: 0.0720 Steps: 70, Updated lr: 0.000007 Training, Epoch: 0001, Batch: 000080, Sample Num: 1280, Cur Loss: 37.93776321, Cur Avg Loss: 53.67785540, Log Avg loss: 34.48630104, Global Avg Loss: 53.67785540, Time: 0.0621 Steps: 80, Updated lr: 0.000008 Training, Epoch: 0001, Batch: 000090, Sample Num: 1440, Cur Loss: 34.89184570, Cur Avg Loss: 51.29469975, Log Avg loss: 32.22945461, Global Avg Loss: 51.29469975, Time: 0.0712 Steps: 90, Updated lr: 0.000009 Training, Epoch: 0001, Batch: 000100, Sample Num: 1600, Cur Loss: 36.13318253, Cur Avg Loss: 49.57297775, Log Avg loss: 34.07747974, Global Avg Loss: 49.57297775, Time: 0.0825 Steps: 100, Updated lr: 0.000010 Training, Epoch: 0001, Batch: 000110, Sample Num: 1760, Cur Loss: 29.13734245, Cur Avg Loss: 47.94535911, Log Avg loss: 31.66917267, Global Avg Loss: 47.94535911, Time: 0.0487 Steps: 110, Updated lr: 0.000011 Training, Epoch: 0001, Batch: 000120, Sample Num: 1920, Cur Loss: 30.19417191, Cur Avg Loss: 46.58153137, Log Avg loss: 31.57942619, Global Avg Loss: 46.58153137, Time: 0.0585 Steps: 120, Updated lr: 0.000012 Training, Epoch: 0001, Batch: 000130, Sample Num: 2080, Cur Loss: 32.55385971, Cur Avg Loss: 45.33840244, Log Avg loss: 30.42085533, Global Avg Loss: 45.33840244, Time: 0.0413 Steps: 130, Updated lr: 0.000013 Training, Epoch: 0001, Batch: 000140, Sample Num: 2240, Cur Loss: 30.55818558, Cur Avg Loss: 44.17868805, Log Avg loss: 29.10240097, Global Avg Loss: 44.17868805, Time: 0.0646 Steps: 140, Updated lr: 0.000014 Training, Epoch: 0001, Batch: 000150, Sample Num: 2400, Cur Loss: 27.61389923, Cur Avg Loss: 43.17020561, Log Avg loss: 29.05145149, Global Avg Loss: 43.17020561, Time: 0.0402 Steps: 150, Updated lr: 0.000015 Training, Epoch: 0001, Batch: 000160, Sample Num: 2560, Cur Loss: 21.76577568, Cur Avg Loss: 42.03182172, Log Avg loss: 24.95606327, Global Avg Loss: 42.03182172, Time: 0.0402 Steps: 160, Updated lr: 0.000016 Training, Epoch: 0001, Batch: 000170, Sample Num: 2720, Cur Loss: 23.05492401, Cur Avg Loss: 40.93174637, Log Avg loss: 23.33054085, Global Avg Loss: 40.93174637, Time: 0.0402 Steps: 170, Updated lr: 0.000017 Training, Epoch: 0001, Batch: 000180, Sample Num: 2880, Cur Loss: 18.56377029, Cur Avg Loss: 39.91387710, Log Avg loss: 22.61009941, Global Avg Loss: 39.91387710, Time: 0.0402 Steps: 180, Updated lr: 0.000018 Training, Epoch: 0001, Batch: 000190, Sample Num: 3040, Cur Loss: 23.47644806, Cur Avg Loss: 38.96343419, Log Avg loss: 21.85546188, Global Avg Loss: 38.96343419, Time: 0.0402 Steps: 190, Updated lr: 0.000019 Training, Epoch: 0001, Batch: 000200, Sample Num: 3200, Cur Loss: 15.72817421, Cur Avg Loss: 38.00501175, Log Avg loss: 19.79498539, Global Avg Loss: 38.00501175, Time: 0.0402 Steps: 200, Updated lr: 0.000020 Training, Epoch: 0001, Batch: 000210, Sample Num: 3360, Cur Loss: 16.31772995, Cur Avg Loss: 37.03458045, Log Avg loss: 17.62595444, Global Avg Loss: 37.03458045, Time: 0.0402 Steps: 210, Updated lr: 0.000021 Training, Epoch: 0001, Batch: 000220, Sample Num: 3520, Cur Loss: 12.57945442, Cur Avg Loss: 36.06495706, Log Avg loss: 15.70286598, Global Avg Loss: 36.06495706, Time: 0.0402 Steps: 220, Updated lr: 0.000022 Training, Epoch: 0001, Batch: 000230, Sample Num: 3680, Cur Loss: 16.25167084, Cur Avg Loss: 35.08649716, Log Avg loss: 13.56037922, Global Avg Loss: 35.08649716, Time: 0.0402 Steps: 230, Updated lr: 0.000023 Training, Epoch: 0001, Batch: 000240, Sample Num: 3840, Cur Loss: 12.25489521, Cur Avg Loss: 34.19729117, Log Avg loss: 13.74555349, Global Avg Loss: 34.19729117, Time: 0.0402 Steps: 240, Updated lr: 0.000024 Training, Epoch: 0001, Batch: 000250, Sample Num: 4000, Cur Loss: 11.02290630, Cur Avg Loss: 33.28087657, Log Avg loss: 11.28692617, Global Avg Loss: 33.28087657, Time: 0.0402 Steps: 250, Updated lr: 0.000025 Training, Epoch: 0001, Batch: 000260, Sample Num: 4160, Cur Loss: 8.89455223, Cur Avg Loss: 32.35894426, Log Avg loss: 9.31063652, Global Avg Loss: 32.35894426, Time: 0.0402 Steps: 260, Updated lr: 0.000026 Training, Epoch: 0001, Batch: 000270, Sample Num: 4320, Cur Loss: 6.00930977, Cur Avg Loss: 31.43731322, Log Avg loss: 7.47490611, Global Avg Loss: 31.43731322, Time: 0.0403 Steps: 270, Updated lr: 0.000027 Training, Epoch: 0001, Batch: 000280, Sample Num: 4480, Cur Loss: 4.68542814, Cur Avg Loss: 30.54026564, Log Avg loss: 6.31998110, Global Avg Loss: 30.54026564, Time: 0.0402 Steps: 280, Updated lr: 0.000028 Training, Epoch: 0001, Batch: 000290, Sample Num: 4640, Cur Loss: 3.96164680, Cur Avg Loss: 29.66276215, Log Avg loss: 5.09266431, Global Avg Loss: 29.66276215, Time: 0.0402 Steps: 290, Updated lr: 0.000029 Training, Epoch: 0001, Batch: 000300, Sample Num: 4800, Cur Loss: 3.03303790, Cur Avg Loss: 28.79063998, Log Avg loss: 3.49909701, Global Avg Loss: 28.79063998, Time: 0.0402 Steps: 300, Updated lr: 0.000030 Training, Epoch: 0001, Batch: 000310, Sample Num: 4960, Cur Loss: 1.67109609, Cur Avg Loss: 27.94220390, Log Avg loss: 2.48912157, Global Avg Loss: 27.94220390, Time: 0.0402 Steps: 310, Updated lr: 0.000031 Training, Epoch: 0001, Batch: 000320, Sample Num: 5120, Cur Loss: 1.56751728, Cur Avg Loss: 27.12985519, Log Avg loss: 1.94704523, Global Avg Loss: 27.12985519, Time: 0.0403 Steps: 320, Updated lr: 0.000032 Training, Epoch: 0001, Batch: 000330, Sample Num: 5280, Cur Loss: 2.88235378, Cur Avg Loss: 26.38118303, Log Avg loss: 2.42367371, Global Avg Loss: 26.38118303, Time: 0.0402 Steps: 330, Updated lr: 0.000033 Training, Epoch: 0001, Batch: 000340, Sample Num: 5440, Cur Loss: 1.79011524, Cur Avg Loss: 25.65449490, Log Avg loss: 1.67378691, Global Avg Loss: 25.65449490, Time: 0.0402 Steps: 340, Updated lr: 0.000034 Training, Epoch: 0001, Batch: 000350, Sample Num: 5600, Cur Loss: 0.78489280, Cur Avg Loss: 24.97072655, Log Avg loss: 1.72260238, Global Avg Loss: 24.97072655, Time: 0.0402 Steps: 350, Updated lr: 0.000035 Training, Epoch: 0001, Batch: 000360, Sample Num: 5760, Cur Loss: 2.20396495, Cur Avg Loss: 24.32457303, Log Avg loss: 1.70919985, Global Avg Loss: 24.32457303, Time: 0.0402 Steps: 360, Updated lr: 0.000036 Training, Epoch: 0001, Batch: 000370, Sample Num: 5920, Cur Loss: 2.78235960, Cur Avg Loss: 23.72895582, Log Avg loss: 2.28673648, Global Avg Loss: 23.72895582, Time: 0.0402 Steps: 370, Updated lr: 0.000037 Training, Epoch: 0001, Batch: 000380, Sample Num: 6080, Cur Loss: 1.74071896, Cur Avg Loss: 23.15058806, Log Avg loss: 1.75098085, Global Avg Loss: 23.15058806, Time: 0.0402 Steps: 380, Updated lr: 0.000038 Training, Epoch: 0001, Batch: 000390, Sample Num: 6240, Cur Loss: 3.39026260, Cur Avg Loss: 22.61695615, Log Avg loss: 2.33894363, Global Avg Loss: 22.61695615, Time: 0.0403 Steps: 390, Updated lr: 0.000039 Training, Epoch: 0001, Batch: 000400, Sample Num: 6400, Cur Loss: 0.88832474, Cur Avg Loss: 22.09818510, Log Avg loss: 1.86611419, Global Avg Loss: 22.09818510, Time: 0.0403 Steps: 400, Updated lr: 0.000040 Training, Epoch: 0001, Batch: 000410, Sample Num: 6560, Cur Loss: 1.95176327, Cur Avg Loss: 21.60732935, Log Avg loss: 1.97309922, Global Avg Loss: 21.60732935, Time: 0.0402 Steps: 410, Updated lr: 0.000041 Training, Epoch: 0001, Batch: 000420, Sample Num: 6720, Cur Loss: 2.08386016, Cur Avg Loss: 21.13637407, Log Avg loss: 1.82720777, Global Avg Loss: 21.13637407, Time: 0.0550 Steps: 420, Updated lr: 0.000042 Training, Epoch: 0001, Batch: 000430, Sample Num: 6880, Cur Loss: 1.88403654, Cur Avg Loss: 20.68677178, Log Avg loss: 1.80347539, Global Avg Loss: 20.68677178, Time: 0.0507 Steps: 430, Updated lr: 0.000043 Training, Epoch: 0001, Batch: 000440, Sample Num: 7040, Cur Loss: 1.76535964, Cur Avg Loss: 20.25368109, Log Avg loss: 1.63078151, Global Avg Loss: 20.25368109, Time: 0.1002 Steps: 440, Updated lr: 0.000044 Training, Epoch: 0001, Batch: 000450, Sample Num: 7200, Cur Loss: 1.69031906, Cur Avg Loss: 19.83986342, Log Avg loss: 1.63188602, Global Avg Loss: 19.83986342, Time: 0.0612 Steps: 450, Updated lr: 0.000045 Training, Epoch: 0001, Batch: 000460, Sample Num: 7360, Cur Loss: 2.15460372, Cur Avg Loss: 19.45276467, Log Avg loss: 2.03332069, Global Avg Loss: 19.45276467, Time: 0.0594 Steps: 460, Updated lr: 0.000046 Training, Epoch: 0001, Batch: 000470, Sample Num: 7520, Cur Loss: 0.84985846, Cur Avg Loss: 19.08145694, Log Avg loss: 2.00130154, Global Avg Loss: 19.08145694, Time: 0.0713 Steps: 470, Updated lr: 0.000047 Training, Epoch: 0001, Batch: 000480, Sample Num: 7680, Cur Loss: 2.48041892, Cur Avg Loss: 18.72857421, Log Avg loss: 2.14308571, Global Avg Loss: 18.72857421, Time: 0.0423 Steps: 480, Updated lr: 0.000048 Training, Epoch: 0001, Batch: 000490, Sample Num: 7840, Cur Loss: 1.22021008, Cur Avg Loss: 18.38170207, Log Avg loss: 1.73183954, Global Avg Loss: 18.38170207, Time: 0.0629 Steps: 490, Updated lr: 0.000049 Training, Epoch: 0001, Batch: 000500, Sample Num: 8000, Cur Loss: 0.89156950, Cur Avg Loss: 18.04927642, Log Avg loss: 1.76041961, Global Avg Loss: 18.04927642, Time: 0.0403 Steps: 500, Updated lr: 0.000050 Training, Epoch: 0001, Batch: 000510, Sample Num: 8160, Cur Loss: 1.95452452, Cur Avg Loss: 17.73862498, Log Avg loss: 2.20605298, Global Avg Loss: 17.73862498, Time: 0.0403 Steps: 510, Updated lr: 0.000051 Training, Epoch: 0001, Batch: 000520, Sample Num: 8320, Cur Loss: 2.12379694, Cur Avg Loss: 17.43523365, Log Avg loss: 1.96227547, Global Avg Loss: 17.43523365, Time: 0.0403 Steps: 520, Updated lr: 0.000052 Training, Epoch: 0001, Batch: 000530, Sample Num: 8480, Cur Loss: 1.21361494, Cur Avg Loss: 17.14314113, Log Avg loss: 1.95433049, Global Avg Loss: 17.14314113, Time: 0.0402 Steps: 530, Updated lr: 0.000053 Training, Epoch: 0001, Batch: 000540, Sample Num: 8640, Cur Loss: 1.06973648, Cur Avg Loss: 16.85409276, Log Avg loss: 1.53452870, Global Avg Loss: 16.85409276, Time: 0.0403 Steps: 540, Updated lr: 0.000054 Training, Epoch: 0001, Batch: 000550, Sample Num: 8800, Cur Loss: 1.51438034, Cur Avg Loss: 16.58180894, Log Avg loss: 1.87848289, Global Avg Loss: 16.58180894, Time: 0.0404 Steps: 550, Updated lr: 0.000055 Training, Epoch: 0001, Batch: 000560, Sample Num: 8960, Cur Loss: 1.61576080, Cur Avg Loss: 16.31472554, Log Avg loss: 1.62513857, Global Avg Loss: 16.31472554, Time: 0.0402 Steps: 560, Updated lr: 0.000056 Training, Epoch: 0001, Batch: 000570, Sample Num: 9120, Cur Loss: 1.00663221, Cur Avg Loss: 16.05972773, Log Avg loss: 1.77985026, Global Avg Loss: 16.05972773, Time: 0.0402 Steps: 570, Updated lr: 0.000057 Training, Epoch: 0001, Batch: 000580, Sample Num: 9280, Cur Loss: 1.02778745, Cur Avg Loss: 15.81299854, Log Avg loss: 1.74943504, Global Avg Loss: 15.81299854, Time: 0.0402 Steps: 580, Updated lr: 0.000058 Training, Epoch: 0001, Batch: 000590, Sample Num: 9440, Cur Loss: 0.77695692, Cur Avg Loss: 15.57434407, Log Avg loss: 1.73238457, Global Avg Loss: 15.57434407, Time: 0.0402 Steps: 590, Updated lr: 0.000059 Training, Epoch: 0001, Batch: 000600, Sample Num: 9600, Cur Loss: 2.45042467, Cur Avg Loss: 15.34516776, Log Avg loss: 1.82376572, Global Avg Loss: 15.34516776, Time: 0.0402 Steps: 600, Updated lr: 0.000060 Training, Epoch: 0001, Batch: 000610, Sample Num: 9760, Cur Loss: 1.67299557, Cur Avg Loss: 15.11669691, Log Avg loss: 1.40844538, Global Avg Loss: 15.11669691, Time: 0.0403 Steps: 610, Updated lr: 0.000061 Training, Epoch: 0001, Batch: 000620, Sample Num: 9920, Cur Loss: 1.36762595, Cur Avg Loss: 14.90203567, Log Avg loss: 1.80770034, Global Avg Loss: 14.90203567, Time: 0.0402 Steps: 620, Updated lr: 0.000062 Training, Epoch: 0001, Batch: 000630, Sample Num: 10080, Cur Loss: 0.80855286, Cur Avg Loss: 14.69447570, Log Avg loss: 1.82575734, Global Avg Loss: 14.69447570, Time: 0.0403 Steps: 630, Updated lr: 0.000063 Training, Epoch: 0001, Batch: 000640, Sample Num: 10240, Cur Loss: 1.30649197, Cur Avg Loss: 14.49443939, Log Avg loss: 1.89215229, Global Avg Loss: 14.49443939, Time: 0.0402 Steps: 640, Updated lr: 0.000064 Training, Epoch: 0001, Batch: 000650, Sample Num: 10400, Cur Loss: 1.55614543, Cur Avg Loss: 14.30085002, Log Avg loss: 1.91112990, Global Avg Loss: 14.30085002, Time: 0.0403 Steps: 650, Updated lr: 0.000065 Training, Epoch: 0001, Batch: 000660, Sample Num: 10560, Cur Loss: 1.68380141, Cur Avg Loss: 14.11604145, Log Avg loss: 2.10348477, Global Avg Loss: 14.11604145, Time: 0.0402 Steps: 660, Updated lr: 0.000066 Training, Epoch: 0001, Batch: 000670, Sample Num: 10720, Cur Loss: 0.94162977, Cur Avg Loss: 13.92680081, Log Avg loss: 1.43691837, Global Avg Loss: 13.92680081, Time: 0.0402 Steps: 670, Updated lr: 0.000067 Training, Epoch: 0001, Batch: 000680, Sample Num: 10880, Cur Loss: 1.88713741, Cur Avg Loss: 13.74821336, Log Avg loss: 1.78285394, Global Avg Loss: 13.74821336, Time: 0.0402 Steps: 680, Updated lr: 0.000068 Training, Epoch: 0001, Batch: 000690, Sample Num: 11040, Cur Loss: 0.96594489, Cur Avg Loss: 13.57344370, Log Avg loss: 1.68910695, Global Avg Loss: 13.57344370, Time: 0.0402 Steps: 690, Updated lr: 0.000069 Training, Epoch: 0001, Batch: 000700, Sample Num: 11200, Cur Loss: 2.03539777, Cur Avg Loss: 13.40506565, Log Avg loss: 1.78698004, Global Avg Loss: 13.40506565, Time: 0.0403 Steps: 700, Updated lr: 0.000070 Training, Epoch: 0001, Batch: 000710, Sample Num: 11360, Cur Loss: 1.90716243, Cur Avg Loss: 13.24346658, Log Avg loss: 1.93153237, Global Avg Loss: 13.24346658, Time: 0.0402 Steps: 710, Updated lr: 0.000071 Training, Epoch: 0001, Batch: 000720, Sample Num: 11520, Cur Loss: 1.65020502, Cur Avg Loss: 13.08109484, Log Avg loss: 1.55270074, Global Avg Loss: 13.08109484, Time: 0.0402 Steps: 720, Updated lr: 0.000072 Training, Epoch: 0001, Batch: 000730, Sample Num: 11680, Cur Loss: 3.20545197, Cur Avg Loss: 12.93002103, Log Avg loss: 2.05270712, Global Avg Loss: 12.93002103, Time: 0.0401 Steps: 730, Updated lr: 0.000073 Training, Epoch: 0001, Batch: 000740, Sample Num: 11840, Cur Loss: 1.58589888, Cur Avg Loss: 12.78613107, Log Avg loss: 2.28216369, Global Avg Loss: 12.78613107, Time: 0.0402 Steps: 740, Updated lr: 0.000074 Training, Epoch: 0001, Batch: 000750, Sample Num: 12000, Cur Loss: 2.65515637, Cur Avg Loss: 12.64141437, Log Avg loss: 1.93237884, Global Avg Loss: 12.64141437, Time: 0.0402 Steps: 750, Updated lr: 0.000075 Training, Epoch: 0001, Batch: 000760, Sample Num: 12160, Cur Loss: 1.10249949, Cur Avg Loss: 12.49905439, Log Avg loss: 1.82205562, Global Avg Loss: 12.49905439, Time: 0.0401 Steps: 760, Updated lr: 0.000076 Training, Epoch: 0001, Batch: 000770, Sample Num: 12320, Cur Loss: 2.06337690, Cur Avg Loss: 12.36138461, Log Avg loss: 1.89848107, Global Avg Loss: 12.36138461, Time: 0.0403 Steps: 770, Updated lr: 0.000077 Training, Epoch: 0001, Batch: 000780, Sample Num: 12480, Cur Loss: 1.89739084, Cur Avg Loss: 12.22701644, Log Avg loss: 1.88066793, Global Avg Loss: 12.22701644, Time: 0.0402 Steps: 780, Updated lr: 0.000078 Training, Epoch: 0001, Batch: 000790, Sample Num: 12640, Cur Loss: 0.84317636, Cur Avg Loss: 12.09190976, Log Avg loss: 1.55358855, Global Avg Loss: 12.09190976, Time: 0.0846 Steps: 790, Updated lr: 0.000079 Training, Epoch: 0001, Batch: 000800, Sample Num: 12800, Cur Loss: 1.60177755, Cur Avg Loss: 11.96276871, Log Avg loss: 1.76062579, Global Avg Loss: 11.96276871, Time: 0.0947 Steps: 800, Updated lr: 0.000080 Training, Epoch: 0001, Batch: 000810, Sample Num: 12960, Cur Loss: 1.98383141, Cur Avg Loss: 11.83919302, Log Avg loss: 1.95313808, Global Avg Loss: 11.83919302, Time: 0.0619 Steps: 810, Updated lr: 0.000081 Training, Epoch: 0001, Batch: 000820, Sample Num: 13120, Cur Loss: 1.29766870, Cur Avg Loss: 11.71748687, Log Avg loss: 1.85928869, Global Avg Loss: 11.71748687, Time: 0.0407 Steps: 820, Updated lr: 0.000082 Training, Epoch: 0001, Batch: 000830, Sample Num: 13280, Cur Loss: 1.19590390, Cur Avg Loss: 11.59659262, Log Avg loss: 1.68326353, Global Avg Loss: 11.59659262, Time: 0.0586 Steps: 830, Updated lr: 0.000083 Training, Epoch: 0001, Batch: 000840, Sample Num: 13440, Cur Loss: 0.94475609, Cur Avg Loss: 11.47810758, Log Avg loss: 1.64384933, Global Avg Loss: 11.47810758, Time: 0.0606 Steps: 840, Updated lr: 0.000084 Training, Epoch: 0001, Batch: 000850, Sample Num: 13600, Cur Loss: 2.69317675, Cur Avg Loss: 11.36120073, Log Avg loss: 1.54102560, Global Avg Loss: 11.36120073, Time: 0.0682 Steps: 850, Updated lr: 0.000085 Training, Epoch: 0001, Batch: 000860, Sample Num: 13760, Cur Loss: 1.90405905, Cur Avg Loss: 11.24981726, Log Avg loss: 1.78222257, Global Avg Loss: 11.24981726, Time: 0.0458 Steps: 860, Updated lr: 0.000086 Training, Epoch: 0001, Batch: 000870, Sample Num: 13920, Cur Loss: 1.17476046, Cur Avg Loss: 11.14172683, Log Avg loss: 1.84594966, Global Avg Loss: 11.14172683, Time: 0.0637 Steps: 870, Updated lr: 0.000087 Training, Epoch: 0001, Batch: 000880, Sample Num: 14080, Cur Loss: 0.79157495, Cur Avg Loss: 11.03542416, Log Avg loss: 1.78709158, Global Avg Loss: 11.03542416, Time: 0.0402 Steps: 880, Updated lr: 0.000088 Training, Epoch: 0001, Batch: 000890, Sample Num: 14240, Cur Loss: 0.81335413, Cur Avg Loss: 10.93147399, Log Avg loss: 1.78385957, Global Avg Loss: 10.93147399, Time: 0.0402 Steps: 890, Updated lr: 0.000089 Training, Epoch: 0001, Batch: 000900, Sample Num: 14400, Cur Loss: 1.19467616, Cur Avg Loss: 10.83120210, Log Avg loss: 1.90700307, Global Avg Loss: 10.83120210, Time: 0.0402 Steps: 900, Updated lr: 0.000090 Training, Epoch: 0001, Batch: 000910, Sample Num: 14560, Cur Loss: 1.94166958, Cur Avg Loss: 10.73340059, Log Avg loss: 1.93126522, Global Avg Loss: 10.73340059, Time: 0.0403 Steps: 910, Updated lr: 0.000091 Training, Epoch: 0001, Batch: 000920, Sample Num: 14720, Cur Loss: 2.75092983, Cur Avg Loss: 10.63703183, Log Avg loss: 1.86747472, Global Avg Loss: 10.63703183, Time: 0.0402 Steps: 920, Updated lr: 0.000092 Training, Epoch: 0001, Batch: 000930, Sample Num: 14880, Cur Loss: 2.44894195, Cur Avg Loss: 10.54173836, Log Avg loss: 1.77473910, Global Avg Loss: 10.54173836, Time: 0.0402 Steps: 930, Updated lr: 0.000093 Training, Epoch: 0001, Batch: 000940, Sample Num: 15040, Cur Loss: 2.96418023, Cur Avg Loss: 10.45143678, Log Avg loss: 2.05338951, Global Avg Loss: 10.45143678, Time: 0.0402 Steps: 940, Updated lr: 0.000094 Training, Epoch: 0001, Batch: 000950, Sample Num: 15200, Cur Loss: 2.63856173, Cur Avg Loss: 10.36261575, Log Avg loss: 2.01343926, Global Avg Loss: 10.36261575, Time: 0.0402 Steps: 950, Updated lr: 0.000095 Training, Epoch: 0001, Batch: 000960, Sample Num: 15360, Cur Loss: 2.49548769, Cur Avg Loss: 10.27283254, Log Avg loss: 1.74342766, Global Avg Loss: 10.27283254, Time: 0.0402 Steps: 960, Updated lr: 0.000096 Training, Epoch: 0001, Batch: 000970, Sample Num: 15520, Cur Loss: 1.46422756, Cur Avg Loss: 10.19027319, Log Avg loss: 2.26457492, Global Avg Loss: 10.19027319, Time: 0.0402 Steps: 970, Updated lr: 0.000097 Training, Epoch: 0001, Batch: 000980, Sample Num: 15680, Cur Loss: 1.81395459, Cur Avg Loss: 10.10712431, Log Avg loss: 2.04168367, Global Avg Loss: 10.10712431, Time: 0.0402 Steps: 980, Updated lr: 0.000098 Training, Epoch: 0001, Batch: 000990, Sample Num: 15840, Cur Loss: 2.06306767, Cur Avg Loss: 10.02431448, Log Avg loss: 1.90895096, Global Avg Loss: 10.02431448, Time: 0.0402 Steps: 990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001000, Sample Num: 16000, Cur Loss: 0.96932518, Cur Avg Loss: 9.93963145, Log Avg loss: 1.55601095, Global Avg Loss: 9.93963145, Time: 0.0402 Steps: 1000, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001010, Sample Num: 16160, Cur Loss: 1.07395720, Cur Avg Loss: 9.85769626, Log Avg loss: 1.66417755, Global Avg Loss: 9.85769626, Time: 0.0402 Steps: 1010, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001020, Sample Num: 16320, Cur Loss: 1.43062806, Cur Avg Loss: 9.77614342, Log Avg loss: 1.53930724, Global Avg Loss: 9.77614342, Time: 0.0402 Steps: 1020, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001030, Sample Num: 16480, Cur Loss: 1.11335862, Cur Avg Loss: 9.69533574, Log Avg loss: 1.45295193, Global Avg Loss: 9.69533574, Time: 0.0402 Steps: 1030, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001040, Sample Num: 16640, Cur Loss: 1.27645564, Cur Avg Loss: 9.61521900, Log Avg loss: 1.36319470, Global Avg Loss: 9.61521900, Time: 0.0403 Steps: 1040, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001050, Sample Num: 16800, Cur Loss: 1.06321120, Cur Avg Loss: 9.54308940, Log Avg loss: 2.04161059, Global Avg Loss: 9.54308940, Time: 0.0402 Steps: 1050, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001060, Sample Num: 16960, Cur Loss: 1.72523832, Cur Avg Loss: 9.46750367, Log Avg loss: 1.53100244, Global Avg Loss: 9.46750367, Time: 0.0403 Steps: 1060, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001070, Sample Num: 17120, Cur Loss: 2.82821226, Cur Avg Loss: 9.39939487, Log Avg loss: 2.17986190, Global Avg Loss: 9.39939487, Time: 0.0402 Steps: 1070, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001080, Sample Num: 17280, Cur Loss: 3.16603613, Cur Avg Loss: 9.33183978, Log Avg loss: 2.10344529, Global Avg Loss: 9.33183978, Time: 0.0402 Steps: 1080, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001090, Sample Num: 17440, Cur Loss: 2.45856452, Cur Avg Loss: 9.26231168, Log Avg loss: 1.75327646, Global Avg Loss: 9.26231168, Time: 0.0402 Steps: 1090, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001100, Sample Num: 17600, Cur Loss: 2.46825695, Cur Avg Loss: 9.19641251, Log Avg loss: 2.01340312, Global Avg Loss: 9.19641251, Time: 0.0402 Steps: 1100, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001110, Sample Num: 17760, Cur Loss: 1.14804506, Cur Avg Loss: 9.12822267, Log Avg loss: 1.62734044, Global Avg Loss: 9.12822267, Time: 0.0402 Steps: 1110, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001120, Sample Num: 17920, Cur Loss: 1.58950496, Cur Avg Loss: 9.06267924, Log Avg loss: 1.78735852, Global Avg Loss: 9.06267924, Time: 0.0402 Steps: 1120, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001130, Sample Num: 18080, Cur Loss: 1.54136384, Cur Avg Loss: 8.99693093, Log Avg loss: 1.63312046, Global Avg Loss: 8.99693093, Time: 0.0405 Steps: 1130, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001140, Sample Num: 18240, Cur Loss: 3.51944518, Cur Avg Loss: 8.93295122, Log Avg loss: 1.70324435, Global Avg Loss: 8.93295122, Time: 0.0606 Steps: 1140, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001150, Sample Num: 18400, Cur Loss: 1.97582173, Cur Avg Loss: 8.87146082, Log Avg loss: 1.86155469, Global Avg Loss: 8.87146082, Time: 0.0403 Steps: 1150, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001160, Sample Num: 18560, Cur Loss: 2.27796459, Cur Avg Loss: 8.81054868, Log Avg loss: 1.80565282, Global Avg Loss: 8.81054868, Time: 0.0694 Steps: 1160, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001170, Sample Num: 18720, Cur Loss: 1.77076817, Cur Avg Loss: 8.74716591, Log Avg loss: 1.39476421, Global Avg Loss: 8.74716591, Time: 0.0714 Steps: 1170, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001180, Sample Num: 18880, Cur Loss: 1.24365795, Cur Avg Loss: 8.69103197, Log Avg loss: 2.12336154, Global Avg Loss: 8.69103197, Time: 0.0494 Steps: 1180, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001190, Sample Num: 19040, Cur Loss: 1.31879902, Cur Avg Loss: 8.62997181, Log Avg loss: 1.42487259, Global Avg Loss: 8.62997181, Time: 0.0745 Steps: 1190, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001200, Sample Num: 19200, Cur Loss: 1.81050754, Cur Avg Loss: 8.57190572, Log Avg loss: 1.66204093, Global Avg Loss: 8.57190572, Time: 0.0600 Steps: 1200, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001210, Sample Num: 19360, Cur Loss: 0.55185771, Cur Avg Loss: 8.51376273, Log Avg loss: 1.53660387, Global Avg Loss: 8.51376273, Time: 0.0580 Steps: 1210, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001220, Sample Num: 19520, Cur Loss: 0.95611745, Cur Avg Loss: 8.45678874, Log Avg loss: 1.56293656, Global Avg Loss: 8.45678874, Time: 0.0402 Steps: 1220, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001230, Sample Num: 19680, Cur Loss: 2.28764296, Cur Avg Loss: 8.40318389, Log Avg loss: 1.86339155, Global Avg Loss: 8.40318389, Time: 0.0402 Steps: 1230, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001240, Sample Num: 19840, Cur Loss: 1.33264840, Cur Avg Loss: 8.34983637, Log Avg loss: 1.78809144, Global Avg Loss: 8.34983637, Time: 0.0402 Steps: 1240, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001250, Sample Num: 20000, Cur Loss: 2.39080811, Cur Avg Loss: 8.29575065, Log Avg loss: 1.58912110, Global Avg Loss: 8.29575065, Time: 0.0402 Steps: 1250, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001260, Sample Num: 20160, Cur Loss: 1.23603773, Cur Avg Loss: 8.23851531, Log Avg loss: 1.08409833, Global Avg Loss: 8.23851531, Time: 0.0402 Steps: 1260, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001270, Sample Num: 20320, Cur Loss: 1.99267817, Cur Avg Loss: 8.18639207, Log Avg loss: 1.61886392, Global Avg Loss: 8.18639207, Time: 0.0402 Steps: 1270, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001280, Sample Num: 20480, Cur Loss: 0.96967006, Cur Avg Loss: 8.13843777, Log Avg loss: 2.04824188, Global Avg Loss: 8.13843777, Time: 0.0403 Steps: 1280, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001290, Sample Num: 20640, Cur Loss: 1.19643009, Cur Avg Loss: 8.08756163, Log Avg loss: 1.57541534, Global Avg Loss: 8.08756163, Time: 0.0402 Steps: 1290, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001300, Sample Num: 20800, Cur Loss: 2.56330347, Cur Avg Loss: 8.03859226, Log Avg loss: 1.72154402, Global Avg Loss: 8.03859226, Time: 0.0402 Steps: 1300, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001310, Sample Num: 20960, Cur Loss: 1.15920007, Cur Avg Loss: 7.98704087, Log Avg loss: 1.28535949, Global Avg Loss: 7.98704087, Time: 0.0402 Steps: 1310, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001320, Sample Num: 21120, Cur Loss: 1.70641804, Cur Avg Loss: 7.93993578, Log Avg loss: 1.76916875, Global Avg Loss: 7.93993578, Time: 0.0401 Steps: 1320, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001330, Sample Num: 21280, Cur Loss: 0.63976377, Cur Avg Loss: 7.89336416, Log Avg loss: 1.74591065, Global Avg Loss: 7.89336416, Time: 0.0402 Steps: 1330, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001340, Sample Num: 21440, Cur Loss: 1.91141164, Cur Avg Loss: 7.84543563, Log Avg loss: 1.47094096, Global Avg Loss: 7.84543563, Time: 0.0401 Steps: 1340, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001350, Sample Num: 21600, Cur Loss: 1.66499209, Cur Avg Loss: 7.79933532, Log Avg loss: 1.62189399, Global Avg Loss: 7.79933532, Time: 0.0402 Steps: 1350, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001360, Sample Num: 21760, Cur Loss: 1.52390265, Cur Avg Loss: 7.75532715, Log Avg loss: 1.81422471, Global Avg Loss: 7.75532715, Time: 0.0402 Steps: 1360, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001370, Sample Num: 21920, Cur Loss: 1.58981490, Cur Avg Loss: 7.71258146, Log Avg loss: 1.89916657, Global Avg Loss: 7.71258146, Time: 0.0403 Steps: 1370, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001380, Sample Num: 22080, Cur Loss: 1.42000580, Cur Avg Loss: 7.66989083, Log Avg loss: 1.82127519, Global Avg Loss: 7.66989083, Time: 0.0402 Steps: 1380, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001390, Sample Num: 22240, Cur Loss: 1.41031432, Cur Avg Loss: 7.62779104, Log Avg loss: 1.81802011, Global Avg Loss: 7.62779104, Time: 0.0402 Steps: 1390, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001400, Sample Num: 22400, Cur Loss: 2.96620989, Cur Avg Loss: 7.58846829, Log Avg loss: 2.12260531, Global Avg Loss: 7.58846829, Time: 0.0402 Steps: 1400, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001410, Sample Num: 22560, Cur Loss: 1.04800534, Cur Avg Loss: 7.54655055, Log Avg loss: 1.67806810, Global Avg Loss: 7.54655055, Time: 0.0402 Steps: 1410, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001420, Sample Num: 22720, Cur Loss: 1.38704431, Cur Avg Loss: 7.50439942, Log Avg loss: 1.56108956, Global Avg Loss: 7.50439942, Time: 0.0402 Steps: 1420, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001430, Sample Num: 22880, Cur Loss: 1.37766206, Cur Avg Loss: 7.46560175, Log Avg loss: 1.95633264, Global Avg Loss: 7.46560175, Time: 0.0402 Steps: 1430, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001440, Sample Num: 23040, Cur Loss: 0.92229837, Cur Avg Loss: 7.42392644, Log Avg loss: 1.46435655, Global Avg Loss: 7.42392644, Time: 0.0402 Steps: 1440, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001450, Sample Num: 23200, Cur Loss: 0.64682591, Cur Avg Loss: 7.38372937, Log Avg loss: 1.59535128, Global Avg Loss: 7.38372937, Time: 0.0401 Steps: 1450, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001460, Sample Num: 23360, Cur Loss: 2.07187462, Cur Avg Loss: 7.34459513, Log Avg loss: 1.67013100, Global Avg Loss: 7.34459513, Time: 0.0402 Steps: 1460, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001470, Sample Num: 23520, Cur Loss: 2.83746338, Cur Avg Loss: 7.30421743, Log Avg loss: 1.40907258, Global Avg Loss: 7.30421743, Time: 0.0402 Steps: 1470, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001480, Sample Num: 23680, Cur Loss: 1.34018564, Cur Avg Loss: 7.26593008, Log Avg loss: 1.63768956, Global Avg Loss: 7.26593008, Time: 0.0402 Steps: 1480, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001490, Sample Num: 23840, Cur Loss: 2.05876207, Cur Avg Loss: 7.22811936, Log Avg loss: 1.63213398, Global Avg Loss: 7.22811936, Time: 0.0535 Steps: 1490, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001500, Sample Num: 24000, Cur Loss: 1.89364159, Cur Avg Loss: 7.19168017, Log Avg loss: 1.76223981, Global Avg Loss: 7.19168017, Time: 0.0405 Steps: 1500, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001510, Sample Num: 24160, Cur Loss: 1.75466299, Cur Avg Loss: 7.15827010, Log Avg loss: 2.14676062, Global Avg Loss: 7.15827010, Time: 0.0585 Steps: 1510, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001520, Sample Num: 24320, Cur Loss: 0.99211848, Cur Avg Loss: 7.12246886, Log Avg loss: 1.71648026, Global Avg Loss: 7.12246886, Time: 0.0404 Steps: 1520, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001530, Sample Num: 24480, Cur Loss: 1.21025610, Cur Avg Loss: 7.08486674, Log Avg loss: 1.36934469, Global Avg Loss: 7.08486674, Time: 0.0559 Steps: 1530, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001540, Sample Num: 24640, Cur Loss: 2.75506020, Cur Avg Loss: 7.05231085, Log Avg loss: 2.07125978, Global Avg Loss: 7.05231085, Time: 0.0428 Steps: 1540, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001550, Sample Num: 24800, Cur Loss: 2.37949085, Cur Avg Loss: 7.01754807, Log Avg loss: 1.66408017, Global Avg Loss: 7.01754807, Time: 0.0670 Steps: 1550, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001560, Sample Num: 24960, Cur Loss: 0.93894315, Cur Avg Loss: 6.98243679, Log Avg loss: 1.54018847, Global Avg Loss: 6.98243679, Time: 0.0788 Steps: 1560, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001570, Sample Num: 25120, Cur Loss: 1.73139095, Cur Avg Loss: 6.94809713, Log Avg loss: 1.59111063, Global Avg Loss: 6.94809713, Time: 0.0764 Steps: 1570, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001580, Sample Num: 25280, Cur Loss: 1.51084149, Cur Avg Loss: 6.91434705, Log Avg loss: 1.61558444, Global Avg Loss: 6.91434705, Time: 0.0402 Steps: 1580, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001590, Sample Num: 25440, Cur Loss: 1.26158750, Cur Avg Loss: 6.88073423, Log Avg loss: 1.56990790, Global Avg Loss: 6.88073423, Time: 0.0402 Steps: 1590, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001600, Sample Num: 25600, Cur Loss: 1.30209374, Cur Avg Loss: 6.84770044, Log Avg loss: 1.59532772, Global Avg Loss: 6.84770044, Time: 0.0402 Steps: 1600, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001610, Sample Num: 25760, Cur Loss: 2.06854057, Cur Avg Loss: 6.81535590, Log Avg loss: 1.64023072, Global Avg Loss: 6.81535590, Time: 0.0402 Steps: 1610, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001620, Sample Num: 25920, Cur Loss: 1.69886446, Cur Avg Loss: 6.78518820, Log Avg loss: 1.92818766, Global Avg Loss: 6.78518820, Time: 0.0402 Steps: 1620, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001630, Sample Num: 26080, Cur Loss: 1.59974146, Cur Avg Loss: 6.75126228, Log Avg loss: 1.25526295, Global Avg Loss: 6.75126228, Time: 0.0402 Steps: 1630, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001640, Sample Num: 26240, Cur Loss: 2.88771415, Cur Avg Loss: 6.72046586, Log Avg loss: 1.70065026, Global Avg Loss: 6.72046586, Time: 0.0402 Steps: 1640, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001650, Sample Num: 26400, Cur Loss: 1.27925098, Cur Avg Loss: 6.68855401, Log Avg loss: 1.45501107, Global Avg Loss: 6.68855401, Time: 0.0402 Steps: 1650, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001660, Sample Num: 26560, Cur Loss: 1.70886123, Cur Avg Loss: 6.65938027, Log Avg loss: 1.84571182, Global Avg Loss: 6.65938027, Time: 0.0402 Steps: 1660, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001670, Sample Num: 26720, Cur Loss: 1.14101720, Cur Avg Loss: 6.62926117, Log Avg loss: 1.62949067, Global Avg Loss: 6.62926117, Time: 0.0402 Steps: 1670, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001680, Sample Num: 26880, Cur Loss: 1.52369392, Cur Avg Loss: 6.59984963, Log Avg loss: 1.68812284, Global Avg Loss: 6.59984963, Time: 0.0401 Steps: 1680, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001690, Sample Num: 27040, Cur Loss: 1.79576564, Cur Avg Loss: 6.57227486, Log Avg loss: 1.93971385, Global Avg Loss: 6.57227486, Time: 0.0402 Steps: 1690, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001700, Sample Num: 27200, Cur Loss: 1.96101391, Cur Avg Loss: 6.54537326, Log Avg loss: 1.99900179, Global Avg Loss: 6.54537326, Time: 0.0402 Steps: 1700, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001710, Sample Num: 27360, Cur Loss: 2.02814603, Cur Avg Loss: 6.51844215, Log Avg loss: 1.94015453, Global Avg Loss: 6.51844215, Time: 0.0402 Steps: 1710, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001720, Sample Num: 27520, Cur Loss: 0.98896426, Cur Avg Loss: 6.48910497, Log Avg loss: 1.47244617, Global Avg Loss: 6.48910497, Time: 0.0401 Steps: 1720, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001730, Sample Num: 27680, Cur Loss: 2.91323090, Cur Avg Loss: 6.46407320, Log Avg loss: 2.15861011, Global Avg Loss: 6.46407320, Time: 0.0401 Steps: 1730, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001740, Sample Num: 27840, Cur Loss: 1.38085568, Cur Avg Loss: 6.43601931, Log Avg loss: 1.58269571, Global Avg Loss: 6.43601931, Time: 0.0402 Steps: 1740, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001750, Sample Num: 28000, Cur Loss: 1.53979921, Cur Avg Loss: 6.40911791, Log Avg loss: 1.72827512, Global Avg Loss: 6.40911791, Time: 0.0402 Steps: 1750, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001760, Sample Num: 28160, Cur Loss: 2.24001026, Cur Avg Loss: 6.38262092, Log Avg loss: 1.74564604, Global Avg Loss: 6.38262092, Time: 0.0402 Steps: 1760, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001770, Sample Num: 28320, Cur Loss: 2.80166817, Cur Avg Loss: 6.35674041, Log Avg loss: 1.80177109, Global Avg Loss: 6.35674041, Time: 0.0402 Steps: 1770, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001780, Sample Num: 28480, Cur Loss: 1.31315660, Cur Avg Loss: 6.32998468, Log Avg loss: 1.59422067, Global Avg Loss: 6.32998468, Time: 0.0402 Steps: 1780, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001790, Sample Num: 28640, Cur Loss: 1.30281723, Cur Avg Loss: 6.30332765, Log Avg loss: 1.55837700, Global Avg Loss: 6.30332765, Time: 0.0402 Steps: 1790, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001800, Sample Num: 28800, Cur Loss: 2.19638395, Cur Avg Loss: 6.27568435, Log Avg loss: 1.32753271, Global Avg Loss: 6.27568435, Time: 0.0403 Steps: 1800, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001810, Sample Num: 28960, Cur Loss: 2.69499946, Cur Avg Loss: 6.24990386, Log Avg loss: 1.60941693, Global Avg Loss: 6.24990386, Time: 0.0402 Steps: 1810, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001820, Sample Num: 29120, Cur Loss: 1.86630487, Cur Avg Loss: 6.22407085, Log Avg loss: 1.54829464, Global Avg Loss: 6.22407085, Time: 0.0402 Steps: 1820, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001830, Sample Num: 29280, Cur Loss: 1.78177714, Cur Avg Loss: 6.20016472, Log Avg loss: 1.84924971, Global Avg Loss: 6.20016472, Time: 0.0402 Steps: 1830, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001840, Sample Num: 29440, Cur Loss: 1.52330780, Cur Avg Loss: 6.17619640, Log Avg loss: 1.78999360, Global Avg Loss: 6.17619640, Time: 0.0402 Steps: 1840, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001850, Sample Num: 29600, Cur Loss: 2.17282104, Cur Avg Loss: 6.15103361, Log Avg loss: 1.52108102, Global Avg Loss: 6.15103361, Time: 0.0681 Steps: 1850, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001860, Sample Num: 29760, Cur Loss: 1.73996961, Cur Avg Loss: 6.12757290, Log Avg loss: 1.78734047, Global Avg Loss: 6.12757290, Time: 0.0403 Steps: 1860, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001870, Sample Num: 29920, Cur Loss: 2.28169584, Cur Avg Loss: 6.10425328, Log Avg loss: 1.76680475, Global Avg Loss: 6.10425328, Time: 0.0406 Steps: 1870, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001880, Sample Num: 30080, Cur Loss: 2.06261539, Cur Avg Loss: 6.08046423, Log Avg loss: 1.63191059, Global Avg Loss: 6.08046423, Time: 0.0727 Steps: 1880, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001890, Sample Num: 30240, Cur Loss: 0.74958467, Cur Avg Loss: 6.05685358, Log Avg loss: 1.61805139, Global Avg Loss: 6.05685358, Time: 0.0580 Steps: 1890, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001900, Sample Num: 30400, Cur Loss: 1.71274567, Cur Avg Loss: 6.03375283, Log Avg loss: 1.66771162, Global Avg Loss: 6.03375283, Time: 0.0404 Steps: 1900, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001910, Sample Num: 30560, Cur Loss: 1.55153561, Cur Avg Loss: 6.01294467, Log Avg loss: 2.05939473, Global Avg Loss: 6.01294467, Time: 0.0658 Steps: 1910, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001920, Sample Num: 30720, Cur Loss: 0.72776794, Cur Avg Loss: 5.99129304, Log Avg loss: 1.85583215, Global Avg Loss: 5.99129304, Time: 0.1001 Steps: 1920, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001930, Sample Num: 30880, Cur Loss: 1.17325497, Cur Avg Loss: 5.96969867, Log Avg loss: 1.82357892, Global Avg Loss: 5.96969867, Time: 0.0724 Steps: 1930, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001940, Sample Num: 31040, Cur Loss: 0.76996189, Cur Avg Loss: 5.94769211, Log Avg loss: 1.70042691, Global Avg Loss: 5.94769211, Time: 0.0405 Steps: 1940, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001950, Sample Num: 31200, Cur Loss: 1.56518650, Cur Avg Loss: 5.92599018, Log Avg loss: 1.71581538, Global Avg Loss: 5.92599018, Time: 0.0402 Steps: 1950, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001960, Sample Num: 31360, Cur Loss: 1.44175076, Cur Avg Loss: 5.90574259, Log Avg loss: 1.95746130, Global Avg Loss: 5.90574259, Time: 0.0402 Steps: 1960, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001970, Sample Num: 31520, Cur Loss: 2.00443363, Cur Avg Loss: 5.88500091, Log Avg loss: 1.81963328, Global Avg Loss: 5.88500091, Time: 0.0402 Steps: 1970, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001980, Sample Num: 31680, Cur Loss: 2.72659063, Cur Avg Loss: 5.86634282, Log Avg loss: 2.19069809, Global Avg Loss: 5.86634282, Time: 0.0402 Steps: 1980, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001990, Sample Num: 31840, Cur Loss: 1.28339100, Cur Avg Loss: 5.84635788, Log Avg loss: 1.88934031, Global Avg Loss: 5.84635788, Time: 0.0402 Steps: 1990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002000, Sample Num: 32000, Cur Loss: 1.27682924, Cur Avg Loss: 5.82524681, Log Avg loss: 1.62414345, Global Avg Loss: 5.82524681, Time: 0.0401 Steps: 2000, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002010, Sample Num: 32160, Cur Loss: 1.66924846, Cur Avg Loss: 5.80414643, Log Avg loss: 1.58407106, Global Avg Loss: 5.80414643, Time: 0.0402 Steps: 2010, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002020, Sample Num: 32320, Cur Loss: 1.37096000, Cur Avg Loss: 5.78395851, Log Avg loss: 1.72618575, Global Avg Loss: 5.78395851, Time: 0.0402 Steps: 2020, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002030, Sample Num: 32480, Cur Loss: 0.93907082, Cur Avg Loss: 5.76347969, Log Avg loss: 1.62675779, Global Avg Loss: 5.76347969, Time: 0.0402 Steps: 2030, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002040, Sample Num: 32640, Cur Loss: 2.04687548, Cur Avg Loss: 5.74633085, Log Avg loss: 2.26511775, Global Avg Loss: 5.74633085, Time: 0.0402 Steps: 2040, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002050, Sample Num: 32800, Cur Loss: 2.45229387, Cur Avg Loss: 5.72721710, Log Avg loss: 1.82801067, Global Avg Loss: 5.72721710, Time: 0.0401 Steps: 2050, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002060, Sample Num: 32960, Cur Loss: 2.21565843, Cur Avg Loss: 5.70847002, Log Avg loss: 1.86531857, Global Avg Loss: 5.70847002, Time: 0.0402 Steps: 2060, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002070, Sample Num: 33120, Cur Loss: 1.54463756, Cur Avg Loss: 5.68931327, Log Avg loss: 1.74302260, Global Avg Loss: 5.68931327, Time: 0.0402 Steps: 2070, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002080, Sample Num: 33280, Cur Loss: 1.52777565, Cur Avg Loss: 5.66961729, Log Avg loss: 1.59255078, Global Avg Loss: 5.66961729, Time: 0.0402 Steps: 2080, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002090, Sample Num: 33440, Cur Loss: 2.08339167, Cur Avg Loss: 5.65056130, Log Avg loss: 1.68691515, Global Avg Loss: 5.65056130, Time: 0.0403 Steps: 2090, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002100, Sample Num: 33600, Cur Loss: 1.23581827, Cur Avg Loss: 5.63153839, Log Avg loss: 1.65575097, Global Avg Loss: 5.63153839, Time: 0.0402 Steps: 2100, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002110, Sample Num: 33760, Cur Loss: 0.79258800, Cur Avg Loss: 5.61169383, Log Avg loss: 1.44433531, Global Avg Loss: 5.61169383, Time: 0.0402 Steps: 2110, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002120, Sample Num: 33920, Cur Loss: 1.86375690, Cur Avg Loss: 5.59426244, Log Avg loss: 1.91623861, Global Avg Loss: 5.59426244, Time: 0.0402 Steps: 2120, Updated lr: 0.000099 ***** Running evaluation checkpoint-2129 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-2129 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 98.013967, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.854827, "eval_total_loss": 1303.943257, "eval_mae": 0.940875, "eval_mse": 1.855472, "eval_r2": -0.17946, "eval_sp_statistic": 0.124956, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.129463, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.585111, "test_total_loss": 795.725496, "test_mae": 0.870811, "test_mse": 1.585547, "test_r2": -0.023326, "test_sp_statistic": 0.135823, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.125082, "test_ps_pvalue": 0.0, "lr": 9.892935040303462e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 5.578205149335557, "train_cur_epoch_loss": 11875.9987629354, "train_cur_epoch_avg_loss": 5.578205149335557, "train_cur_epoch_time": 98.0139672756195, "train_cur_epoch_avg_time": 0.0460375609561388, "epoch": 1, "step": 2129} ################################################## Training, Epoch: 0002, Batch: 000001, Sample Num: 16, Cur Loss: 1.47586226, Cur Avg Loss: 1.47586226, Log Avg loss: 1.76382568, Global Avg Loss: 5.57627917, Time: 0.0406 Steps: 2130, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000011, Sample Num: 176, Cur Loss: 2.10221624, Cur Avg Loss: 1.74907265, Log Avg loss: 1.77639369, Global Avg Loss: 5.55852269, Time: 0.0438 Steps: 2140, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000021, Sample Num: 336, Cur Loss: 3.96819520, Cur Avg Loss: 1.71364293, Log Avg loss: 1.67467023, Global Avg Loss: 5.54045826, Time: 0.0402 Steps: 2150, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000031, Sample Num: 496, Cur Loss: 2.49972343, Cur Avg Loss: 1.75420245, Log Avg loss: 1.83937746, Global Avg Loss: 5.52332363, Time: 0.0402 Steps: 2160, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000041, Sample Num: 656, Cur Loss: 1.34506845, Cur Avg Loss: 1.74950088, Log Avg loss: 1.73492600, Global Avg Loss: 5.50586558, Time: 0.0402 Steps: 2170, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000051, Sample Num: 816, Cur Loss: 0.71235335, Cur Avg Loss: 1.71505485, Log Avg loss: 1.57382610, Global Avg Loss: 5.48782870, Time: 0.0401 Steps: 2180, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000061, Sample Num: 976, Cur Loss: 1.20514250, Cur Avg Loss: 1.66396315, Log Avg loss: 1.40339552, Global Avg Loss: 5.46917832, Time: 0.0402 Steps: 2190, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000071, Sample Num: 1136, Cur Loss: 1.85340977, Cur Avg Loss: 1.69496326, Log Avg loss: 1.88406389, Global Avg Loss: 5.45288234, Time: 0.0402 Steps: 2200, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000081, Sample Num: 1296, Cur Loss: 1.44271159, Cur Avg Loss: 1.70385050, Log Avg loss: 1.76694996, Global Avg Loss: 5.43620392, Time: 0.0401 Steps: 2210, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000091, Sample Num: 1456, Cur Loss: 2.07062316, Cur Avg Loss: 1.71436042, Log Avg loss: 1.79949073, Global Avg Loss: 5.41982232, Time: 0.0402 Steps: 2220, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000101, Sample Num: 1616, Cur Loss: 2.73856878, Cur Avg Loss: 1.74351727, Log Avg loss: 2.00884461, Global Avg Loss: 5.40452646, Time: 0.0402 Steps: 2230, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000111, Sample Num: 1776, Cur Loss: 1.90746856, Cur Avg Loss: 1.76341567, Log Avg loss: 1.96438956, Global Avg Loss: 5.38916871, Time: 0.0402 Steps: 2240, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000121, Sample Num: 1936, Cur Loss: 1.47235990, Cur Avg Loss: 1.74344386, Log Avg loss: 1.52175672, Global Avg Loss: 5.37198021, Time: 0.0402 Steps: 2250, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000131, Sample Num: 2096, Cur Loss: 1.94062436, Cur Avg Loss: 1.75594330, Log Avg loss: 1.90718652, Global Avg Loss: 5.35664926, Time: 0.0402 Steps: 2260, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000141, Sample Num: 2256, Cur Loss: 1.72058344, Cur Avg Loss: 1.75628099, Log Avg loss: 1.76070466, Global Avg Loss: 5.34080810, Time: 0.0402 Steps: 2270, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000151, Sample Num: 2416, Cur Loss: 1.51551509, Cur Avg Loss: 1.75843848, Log Avg loss: 1.78885920, Global Avg Loss: 5.32522937, Time: 0.0403 Steps: 2280, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000161, Sample Num: 2576, Cur Loss: 2.23319006, Cur Avg Loss: 1.76590572, Log Avg loss: 1.87866107, Global Avg Loss: 5.31017886, Time: 0.0404 Steps: 2290, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000171, Sample Num: 2736, Cur Loss: 1.79684126, Cur Avg Loss: 1.75940441, Log Avg loss: 1.65473325, Global Avg Loss: 5.29428562, Time: 0.0404 Steps: 2300, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000181, Sample Num: 2896, Cur Loss: 0.86200833, Cur Avg Loss: 1.74478726, Log Avg loss: 1.49483397, Global Avg Loss: 5.27783777, Time: 0.0507 Steps: 2310, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000191, Sample Num: 3056, Cur Loss: 1.61944032, Cur Avg Loss: 1.74658024, Log Avg loss: 1.77903328, Global Avg Loss: 5.26275672, Time: 0.0431 Steps: 2320, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000201, Sample Num: 3216, Cur Loss: 2.22166443, Cur Avg Loss: 1.75079802, Log Avg loss: 1.83135744, Global Avg Loss: 5.24802968, Time: 0.0538 Steps: 2330, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000211, Sample Num: 3376, Cur Loss: 1.77532911, Cur Avg Loss: 1.73421596, Log Avg loss: 1.40091663, Global Avg Loss: 5.23158903, Time: 0.0404 Steps: 2340, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000221, Sample Num: 3536, Cur Loss: 1.06981373, Cur Avg Loss: 1.73044730, Log Avg loss: 1.65092868, Global Avg Loss: 5.21635218, Time: 0.0667 Steps: 2350, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000231, Sample Num: 3696, Cur Loss: 0.45492125, Cur Avg Loss: 1.73282626, Log Avg loss: 1.78540120, Global Avg Loss: 5.20181425, Time: 0.0403 Steps: 2360, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000241, Sample Num: 3856, Cur Loss: 1.31771803, Cur Avg Loss: 1.73182537, Log Avg loss: 1.70870471, Global Avg Loss: 5.18707539, Time: 0.0404 Steps: 2370, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000251, Sample Num: 4016, Cur Loss: 1.17796373, Cur Avg Loss: 1.72245300, Log Avg loss: 1.49657895, Global Avg Loss: 5.17156910, Time: 0.0687 Steps: 2380, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000261, Sample Num: 4176, Cur Loss: 1.97602856, Cur Avg Loss: 1.72384941, Log Avg loss: 1.75889946, Global Avg Loss: 5.15729015, Time: 0.0402 Steps: 2390, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000271, Sample Num: 4336, Cur Loss: 1.96450949, Cur Avg Loss: 1.72650222, Log Avg loss: 1.79574047, Global Avg Loss: 5.14328369, Time: 0.0402 Steps: 2400, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000281, Sample Num: 4496, Cur Loss: 1.18232441, Cur Avg Loss: 1.72333715, Log Avg loss: 1.63756369, Global Avg Loss: 5.12873714, Time: 0.0402 Steps: 2410, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000291, Sample Num: 4656, Cur Loss: 2.65389872, Cur Avg Loss: 1.73889782, Log Avg loss: 2.17615259, Global Avg Loss: 5.11653638, Time: 0.0402 Steps: 2420, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000301, Sample Num: 4816, Cur Loss: 0.70189750, Cur Avg Loss: 1.73199636, Log Avg loss: 1.53116393, Global Avg Loss: 5.10178176, Time: 0.0402 Steps: 2430, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000311, Sample Num: 4976, Cur Loss: 1.54033399, Cur Avg Loss: 1.73546707, Log Avg loss: 1.83993559, Global Avg Loss: 5.08841353, Time: 0.0402 Steps: 2440, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000321, Sample Num: 5136, Cur Loss: 1.59510970, Cur Avg Loss: 1.73594484, Log Avg loss: 1.75080338, Global Avg Loss: 5.07479064, Time: 0.0402 Steps: 2450, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000331, Sample Num: 5296, Cur Loss: 2.62615204, Cur Avg Loss: 1.74871957, Log Avg loss: 2.15878841, Global Avg Loss: 5.06293697, Time: 0.0403 Steps: 2460, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000341, Sample Num: 5456, Cur Loss: 1.51392698, Cur Avg Loss: 1.74286488, Log Avg loss: 1.54907471, Global Avg Loss: 5.04871080, Time: 0.0403 Steps: 2470, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000351, Sample Num: 5616, Cur Loss: 1.39225411, Cur Avg Loss: 1.74975889, Log Avg loss: 1.98484458, Global Avg Loss: 5.03635651, Time: 0.0403 Steps: 2480, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000361, Sample Num: 5776, Cur Loss: 1.16271198, Cur Avg Loss: 1.74002068, Log Avg loss: 1.39820947, Global Avg Loss: 5.02174547, Time: 0.0402 Steps: 2490, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000371, Sample Num: 5936, Cur Loss: 1.70329642, Cur Avg Loss: 1.74281647, Log Avg loss: 1.84374450, Global Avg Loss: 5.00903347, Time: 0.0402 Steps: 2500, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000381, Sample Num: 6096, Cur Loss: 1.30673051, Cur Avg Loss: 1.74071022, Log Avg loss: 1.66256834, Global Avg Loss: 4.99570094, Time: 0.0402 Steps: 2510, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000391, Sample Num: 6256, Cur Loss: 2.01568317, Cur Avg Loss: 1.74031661, Log Avg loss: 1.72532014, Global Avg Loss: 4.98272324, Time: 0.0402 Steps: 2520, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000401, Sample Num: 6416, Cur Loss: 1.27202141, Cur Avg Loss: 1.73287298, Log Avg loss: 1.44182717, Global Avg Loss: 4.96872760, Time: 0.0402 Steps: 2530, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000411, Sample Num: 6576, Cur Loss: 1.54126215, Cur Avg Loss: 1.73135734, Log Avg loss: 1.67057992, Global Avg Loss: 4.95574277, Time: 0.0402 Steps: 2540, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000421, Sample Num: 6736, Cur Loss: 0.91983843, Cur Avg Loss: 1.73016069, Log Avg loss: 1.68097851, Global Avg Loss: 4.94290055, Time: 0.0402 Steps: 2550, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000431, Sample Num: 6896, Cur Loss: 1.85698509, Cur Avg Loss: 1.72842893, Log Avg loss: 1.65552197, Global Avg Loss: 4.93005923, Time: 0.0402 Steps: 2560, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000441, Sample Num: 7056, Cur Loss: 2.51518130, Cur Avg Loss: 1.73417389, Log Avg loss: 1.98178127, Global Avg Loss: 4.91858733, Time: 0.0402 Steps: 2570, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000451, Sample Num: 7216, Cur Loss: 2.47551394, Cur Avg Loss: 1.72540308, Log Avg loss: 1.33861062, Global Avg Loss: 4.90471145, Time: 0.0402 Steps: 2580, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000461, Sample Num: 7376, Cur Loss: 0.80408287, Cur Avg Loss: 1.72070449, Log Avg loss: 1.50879797, Global Avg Loss: 4.89159982, Time: 0.0403 Steps: 2590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000471, Sample Num: 7536, Cur Loss: 2.22583151, Cur Avg Loss: 1.71641158, Log Avg loss: 1.51850870, Global Avg Loss: 4.87862639, Time: 0.0403 Steps: 2600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000481, Sample Num: 7696, Cur Loss: 2.12761211, Cur Avg Loss: 1.71612882, Log Avg loss: 1.70281055, Global Avg Loss: 4.86645852, Time: 0.0402 Steps: 2610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000491, Sample Num: 7856, Cur Loss: 1.07745087, Cur Avg Loss: 1.71576301, Log Avg loss: 1.69816744, Global Avg Loss: 4.85436580, Time: 0.0402 Steps: 2620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000501, Sample Num: 8016, Cur Loss: 1.39159584, Cur Avg Loss: 1.71724141, Log Avg loss: 1.78983084, Global Avg Loss: 4.84271358, Time: 0.0402 Steps: 2630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000511, Sample Num: 8176, Cur Loss: 0.94099855, Cur Avg Loss: 1.71228576, Log Avg loss: 1.46400777, Global Avg Loss: 4.82991545, Time: 0.0402 Steps: 2640, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000521, Sample Num: 8336, Cur Loss: 1.66062737, Cur Avg Loss: 1.71168055, Log Avg loss: 1.68075442, Global Avg Loss: 4.81803182, Time: 0.0498 Steps: 2650, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000531, Sample Num: 8496, Cur Loss: 2.64220119, Cur Avg Loss: 1.70829772, Log Avg loss: 1.53205211, Global Avg Loss: 4.80567852, Time: 0.0404 Steps: 2660, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000541, Sample Num: 8656, Cur Loss: 1.47710299, Cur Avg Loss: 1.71185801, Log Avg loss: 1.90090972, Global Avg Loss: 4.79479923, Time: 0.0825 Steps: 2670, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000551, Sample Num: 8816, Cur Loss: 1.41299796, Cur Avg Loss: 1.71340014, Log Avg loss: 1.79682912, Global Avg Loss: 4.78361278, Time: 0.0629 Steps: 2680, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000561, Sample Num: 8976, Cur Loss: 3.67004585, Cur Avg Loss: 1.71324061, Log Avg loss: 1.70445064, Global Avg Loss: 4.77216608, Time: 0.0423 Steps: 2690, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000571, Sample Num: 9136, Cur Loss: 1.77122867, Cur Avg Loss: 1.71376380, Log Avg loss: 1.74311474, Global Avg Loss: 4.76094737, Time: 0.0619 Steps: 2700, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000581, Sample Num: 9296, Cur Loss: 0.58138180, Cur Avg Loss: 1.70593791, Log Avg loss: 1.25907978, Global Avg Loss: 4.74802535, Time: 0.0404 Steps: 2710, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000591, Sample Num: 9456, Cur Loss: 1.52933335, Cur Avg Loss: 1.70211952, Log Avg loss: 1.48027083, Global Avg Loss: 4.73601154, Time: 0.0815 Steps: 2720, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000601, Sample Num: 9616, Cur Loss: 2.10015869, Cur Avg Loss: 1.69663856, Log Avg loss: 1.37271385, Global Avg Loss: 4.72369177, Time: 0.0408 Steps: 2730, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000611, Sample Num: 9776, Cur Loss: 1.58867335, Cur Avg Loss: 1.69368741, Log Avg loss: 1.51632316, Global Avg Loss: 4.71198605, Time: 0.0402 Steps: 2740, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000621, Sample Num: 9936, Cur Loss: 0.76957160, Cur Avg Loss: 1.69044916, Log Avg loss: 1.49259253, Global Avg Loss: 4.70027916, Time: 0.0402 Steps: 2750, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000631, Sample Num: 10096, Cur Loss: 2.22148705, Cur Avg Loss: 1.69779470, Log Avg loss: 2.15395213, Global Avg Loss: 4.69105334, Time: 0.0402 Steps: 2760, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000641, Sample Num: 10256, Cur Loss: 1.33808684, Cur Avg Loss: 1.69851932, Log Avg loss: 1.74424340, Global Avg Loss: 4.68041504, Time: 0.0402 Steps: 2770, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000651, Sample Num: 10416, Cur Loss: 2.97716904, Cur Avg Loss: 1.69882432, Log Avg loss: 1.71837472, Global Avg Loss: 4.66976021, Time: 0.0402 Steps: 2780, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000661, Sample Num: 10576, Cur Loss: 1.78383613, Cur Avg Loss: 1.69766869, Log Avg loss: 1.62243723, Global Avg Loss: 4.65883791, Time: 0.0402 Steps: 2790, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000671, Sample Num: 10736, Cur Loss: 2.10033655, Cur Avg Loss: 1.70112401, Log Avg loss: 1.92952014, Global Avg Loss: 4.64909035, Time: 0.0403 Steps: 2800, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000681, Sample Num: 10896, Cur Loss: 1.56312060, Cur Avg Loss: 1.69767131, Log Avg loss: 1.46599514, Global Avg Loss: 4.63776261, Time: 0.0402 Steps: 2810, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000691, Sample Num: 11056, Cur Loss: 1.24687362, Cur Avg Loss: 1.69563815, Log Avg loss: 1.55718029, Global Avg Loss: 4.62683855, Time: 0.0402 Steps: 2820, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000701, Sample Num: 11216, Cur Loss: 1.94169950, Cur Avg Loss: 1.69599703, Log Avg loss: 1.72079593, Global Avg Loss: 4.61656985, Time: 0.0403 Steps: 2830, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000711, Sample Num: 11376, Cur Loss: 1.24137592, Cur Avg Loss: 1.69496965, Log Avg loss: 1.62295012, Global Avg Loss: 4.60602894, Time: 0.0402 Steps: 2840, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000721, Sample Num: 11536, Cur Loss: 1.57191038, Cur Avg Loss: 1.69442218, Log Avg loss: 1.65549672, Global Avg Loss: 4.59567619, Time: 0.0402 Steps: 2850, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000731, Sample Num: 11696, Cur Loss: 1.94060564, Cur Avg Loss: 1.69448886, Log Avg loss: 1.69929651, Global Avg Loss: 4.58554899, Time: 0.0402 Steps: 2860, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000741, Sample Num: 11856, Cur Loss: 1.38816023, Cur Avg Loss: 1.69010153, Log Avg loss: 1.36938760, Global Avg Loss: 4.57434285, Time: 0.0402 Steps: 2870, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000751, Sample Num: 12016, Cur Loss: 1.54493904, Cur Avg Loss: 1.69023307, Log Avg loss: 1.69998045, Global Avg Loss: 4.56436243, Time: 0.0402 Steps: 2880, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000761, Sample Num: 12176, Cur Loss: 1.07166505, Cur Avg Loss: 1.68493916, Log Avg loss: 1.28736658, Global Avg Loss: 4.55302334, Time: 0.0402 Steps: 2890, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000771, Sample Num: 12336, Cur Loss: 1.33259439, Cur Avg Loss: 1.68341106, Log Avg loss: 1.56712253, Global Avg Loss: 4.54272713, Time: 0.0402 Steps: 2900, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000781, Sample Num: 12496, Cur Loss: 1.50327027, Cur Avg Loss: 1.68706400, Log Avg loss: 1.96870608, Global Avg Loss: 4.53388170, Time: 0.0401 Steps: 2910, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000791, Sample Num: 12656, Cur Loss: 2.06172514, Cur Avg Loss: 1.68734604, Log Avg loss: 1.70937314, Global Avg Loss: 4.52420873, Time: 0.0402 Steps: 2920, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000801, Sample Num: 12816, Cur Loss: 1.25564647, Cur Avg Loss: 1.69246598, Log Avg loss: 2.09745332, Global Avg Loss: 4.51592628, Time: 0.0402 Steps: 2930, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000811, Sample Num: 12976, Cur Loss: 2.16684008, Cur Avg Loss: 1.69286843, Log Avg loss: 1.72510427, Global Avg Loss: 4.50643369, Time: 0.0402 Steps: 2940, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000821, Sample Num: 13136, Cur Loss: 2.31237149, Cur Avg Loss: 1.69924536, Log Avg loss: 2.21641489, Global Avg Loss: 4.49867092, Time: 0.0402 Steps: 2950, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000831, Sample Num: 13296, Cur Loss: 2.99537182, Cur Avg Loss: 1.69930294, Log Avg loss: 1.70402992, Global Avg Loss: 4.48922956, Time: 0.0401 Steps: 2960, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000841, Sample Num: 13456, Cur Loss: 1.72264302, Cur Avg Loss: 1.69891720, Log Avg loss: 1.66686230, Global Avg Loss: 4.47972664, Time: 0.0402 Steps: 2970, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000851, Sample Num: 13616, Cur Loss: 2.08211803, Cur Avg Loss: 1.69724962, Log Avg loss: 1.55700619, Global Avg Loss: 4.46991886, Time: 0.0401 Steps: 2980, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000861, Sample Num: 13776, Cur Loss: 1.25816607, Cur Avg Loss: 1.69738698, Log Avg loss: 1.70907655, Global Avg Loss: 4.46068527, Time: 0.0606 Steps: 2990, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000871, Sample Num: 13936, Cur Loss: 0.97254586, Cur Avg Loss: 1.69282027, Log Avg loss: 1.29962627, Global Avg Loss: 4.45014841, Time: 0.0995 Steps: 3000, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000881, Sample Num: 14096, Cur Loss: 1.53207040, Cur Avg Loss: 1.69720624, Log Avg loss: 2.07922455, Global Avg Loss: 4.44227158, Time: 0.0742 Steps: 3010, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000891, Sample Num: 14256, Cur Loss: 2.13432527, Cur Avg Loss: 1.69826232, Log Avg loss: 1.79130292, Global Avg Loss: 4.43349354, Time: 0.0507 Steps: 3020, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000901, Sample Num: 14416, Cur Loss: 1.57823372, Cur Avg Loss: 1.69696930, Log Avg loss: 1.58176057, Global Avg Loss: 4.42408188, Time: 0.0587 Steps: 3030, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000911, Sample Num: 14576, Cur Loss: 1.43918157, Cur Avg Loss: 1.69976917, Log Avg loss: 1.95203745, Global Avg Loss: 4.41595016, Time: 0.0851 Steps: 3040, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000921, Sample Num: 14736, Cur Loss: 1.85929787, Cur Avg Loss: 1.70298184, Log Avg loss: 1.99565661, Global Avg Loss: 4.40801477, Time: 0.0406 Steps: 3050, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000931, Sample Num: 14896, Cur Loss: 2.95291185, Cur Avg Loss: 1.70219920, Log Avg loss: 1.63011746, Global Avg Loss: 4.39893667, Time: 0.0777 Steps: 3060, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000941, Sample Num: 15056, Cur Loss: 1.43691826, Cur Avg Loss: 1.70355368, Log Avg loss: 1.82965642, Global Avg Loss: 4.39056768, Time: 0.0515 Steps: 3070, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000951, Sample Num: 15216, Cur Loss: 1.74565089, Cur Avg Loss: 1.70792656, Log Avg loss: 2.11941386, Global Avg Loss: 4.38319380, Time: 0.0402 Steps: 3080, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000961, Sample Num: 15376, Cur Loss: 1.52557790, Cur Avg Loss: 1.70937447, Log Avg loss: 1.84707092, Global Avg Loss: 4.37498629, Time: 0.0402 Steps: 3090, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000971, Sample Num: 15536, Cur Loss: 2.06395698, Cur Avg Loss: 1.70946998, Log Avg loss: 1.71864896, Global Avg Loss: 4.36641746, Time: 0.0402 Steps: 3100, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000981, Sample Num: 15696, Cur Loss: 1.21656060, Cur Avg Loss: 1.70962064, Log Avg loss: 1.72424930, Global Avg Loss: 4.35792174, Time: 0.0403 Steps: 3110, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000991, Sample Num: 15856, Cur Loss: 1.91268110, Cur Avg Loss: 1.70848199, Log Avg loss: 1.59678030, Global Avg Loss: 4.34907193, Time: 0.0402 Steps: 3120, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001001, Sample Num: 16016, Cur Loss: 1.30775571, Cur Avg Loss: 1.70390235, Log Avg loss: 1.25006061, Global Avg Loss: 4.33917093, Time: 0.0403 Steps: 3130, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001011, Sample Num: 16176, Cur Loss: 2.27030706, Cur Avg Loss: 1.70453111, Log Avg loss: 1.76747013, Global Avg Loss: 4.33098080, Time: 0.0402 Steps: 3140, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001021, Sample Num: 16336, Cur Loss: 1.61534262, Cur Avg Loss: 1.70583803, Log Avg loss: 1.83796734, Global Avg Loss: 4.32306647, Time: 0.0403 Steps: 3150, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001031, Sample Num: 16496, Cur Loss: 0.45242232, Cur Avg Loss: 1.70528687, Log Avg loss: 1.64901379, Global Avg Loss: 4.31460428, Time: 0.0402 Steps: 3160, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001041, Sample Num: 16656, Cur Loss: 1.96729910, Cur Avg Loss: 1.70548320, Log Avg loss: 1.72572475, Global Avg Loss: 4.30643747, Time: 0.0402 Steps: 3170, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001051, Sample Num: 16816, Cur Loss: 1.31885266, Cur Avg Loss: 1.70794855, Log Avg loss: 1.96459099, Global Avg Loss: 4.29907317, Time: 0.0402 Steps: 3180, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001061, Sample Num: 16976, Cur Loss: 1.13885391, Cur Avg Loss: 1.70589817, Log Avg loss: 1.49040384, Global Avg Loss: 4.29026857, Time: 0.0402 Steps: 3190, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001071, Sample Num: 17136, Cur Loss: 3.10543418, Cur Avg Loss: 1.70537070, Log Avg loss: 1.64940533, Global Avg Loss: 4.28201587, Time: 0.0402 Steps: 3200, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001081, Sample Num: 17296, Cur Loss: 1.94178677, Cur Avg Loss: 1.70599645, Log Avg loss: 1.77301418, Global Avg Loss: 4.27419966, Time: 0.0402 Steps: 3210, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001091, Sample Num: 17456, Cur Loss: 2.36103797, Cur Avg Loss: 1.70640660, Log Avg loss: 1.75074408, Global Avg Loss: 4.26636285, Time: 0.0403 Steps: 3220, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001101, Sample Num: 17616, Cur Loss: 1.12035394, Cur Avg Loss: 1.70439110, Log Avg loss: 1.48450001, Global Avg Loss: 4.25775027, Time: 0.0402 Steps: 3230, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001111, Sample Num: 17776, Cur Loss: 2.07742453, Cur Avg Loss: 1.70519888, Log Avg loss: 1.79413582, Global Avg Loss: 4.25014652, Time: 0.0403 Steps: 3240, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001121, Sample Num: 17936, Cur Loss: 1.81109071, Cur Avg Loss: 1.70570250, Log Avg loss: 1.76165425, Global Avg Loss: 4.24248962, Time: 0.0403 Steps: 3250, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001131, Sample Num: 18096, Cur Loss: 1.13237059, Cur Avg Loss: 1.70629750, Log Avg loss: 1.77299763, Global Avg Loss: 4.23491449, Time: 0.0403 Steps: 3260, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001141, Sample Num: 18256, Cur Loss: 1.79552293, Cur Avg Loss: 1.70611280, Log Avg loss: 1.68522239, Global Avg Loss: 4.22711727, Time: 0.0402 Steps: 3270, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001151, Sample Num: 18416, Cur Loss: 1.75689745, Cur Avg Loss: 1.70552877, Log Avg loss: 1.63889149, Global Avg Loss: 4.21922633, Time: 0.0402 Steps: 3280, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001161, Sample Num: 18576, Cur Loss: 1.56999159, Cur Avg Loss: 1.70589101, Log Avg loss: 1.74758444, Global Avg Loss: 4.21171375, Time: 0.0402 Steps: 3290, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001171, Sample Num: 18736, Cur Loss: 1.68261623, Cur Avg Loss: 1.70785106, Log Avg loss: 1.93541375, Global Avg Loss: 4.20481587, Time: 0.0402 Steps: 3300, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001181, Sample Num: 18896, Cur Loss: 2.28056192, Cur Avg Loss: 1.70870564, Log Avg loss: 1.80877593, Global Avg Loss: 4.19757708, Time: 0.0402 Steps: 3310, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001191, Sample Num: 19056, Cur Loss: 2.20090127, Cur Avg Loss: 1.71193582, Log Avg loss: 2.09342084, Global Avg Loss: 4.19123926, Time: 0.0402 Steps: 3320, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001201, Sample Num: 19216, Cur Loss: 2.24313831, Cur Avg Loss: 1.70977765, Log Avg loss: 1.45273896, Global Avg Loss: 4.18301553, Time: 0.0403 Steps: 3330, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001211, Sample Num: 19376, Cur Loss: 3.30134106, Cur Avg Loss: 1.70976341, Log Avg loss: 1.70805381, Global Avg Loss: 4.17560547, Time: 0.0403 Steps: 3340, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001221, Sample Num: 19536, Cur Loss: 1.00482583, Cur Avg Loss: 1.71125333, Log Avg loss: 1.89168236, Global Avg Loss: 4.16878778, Time: 0.0481 Steps: 3350, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001231, Sample Num: 19696, Cur Loss: 0.49957713, Cur Avg Loss: 1.70988136, Log Avg loss: 1.54236339, Global Avg Loss: 4.16097105, Time: 0.0832 Steps: 3360, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001241, Sample Num: 19856, Cur Loss: 2.52352834, Cur Avg Loss: 1.70922658, Log Avg loss: 1.62862307, Global Avg Loss: 4.15345666, Time: 0.0428 Steps: 3370, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001251, Sample Num: 20016, Cur Loss: 1.23750222, Cur Avg Loss: 1.71019453, Log Avg loss: 1.83031822, Global Avg Loss: 4.14658347, Time: 0.0733 Steps: 3380, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001261, Sample Num: 20176, Cur Loss: 1.81345248, Cur Avg Loss: 1.71214582, Log Avg loss: 1.95625182, Global Avg Loss: 4.14012231, Time: 0.0574 Steps: 3390, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001271, Sample Num: 20336, Cur Loss: 2.25962901, Cur Avg Loss: 1.71105735, Log Avg loss: 1.57380100, Global Avg Loss: 4.13257431, Time: 0.0880 Steps: 3400, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001281, Sample Num: 20496, Cur Loss: 3.37772512, Cur Avg Loss: 1.71056685, Log Avg loss: 1.64822466, Global Avg Loss: 4.12528883, Time: 0.0607 Steps: 3410, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001291, Sample Num: 20656, Cur Loss: 1.95514047, Cur Avg Loss: 1.70996946, Log Avg loss: 1.63344404, Global Avg Loss: 4.11800273, Time: 0.0403 Steps: 3420, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001301, Sample Num: 20816, Cur Loss: 1.36766577, Cur Avg Loss: 1.71251087, Log Avg loss: 2.04060577, Global Avg Loss: 4.11194618, Time: 0.0636 Steps: 3430, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001311, Sample Num: 20976, Cur Loss: 2.29934406, Cur Avg Loss: 1.71367746, Log Avg loss: 1.86545116, Global Avg Loss: 4.10541567, Time: 0.1003 Steps: 3440, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001321, Sample Num: 21136, Cur Loss: 1.20109510, Cur Avg Loss: 1.71334484, Log Avg loss: 1.66973864, Global Avg Loss: 4.09835574, Time: 0.0403 Steps: 3450, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001331, Sample Num: 21296, Cur Loss: 0.76740491, Cur Avg Loss: 1.71298604, Log Avg loss: 1.66558830, Global Avg Loss: 4.09132462, Time: 0.0403 Steps: 3460, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001341, Sample Num: 21456, Cur Loss: 1.98350036, Cur Avg Loss: 1.71352716, Log Avg loss: 1.78555082, Global Avg Loss: 4.08467974, Time: 0.0402 Steps: 3470, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001351, Sample Num: 21616, Cur Loss: 1.12211275, Cur Avg Loss: 1.71304725, Log Avg loss: 1.64869081, Global Avg Loss: 4.07767977, Time: 0.0402 Steps: 3480, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001361, Sample Num: 21776, Cur Loss: 1.67115891, Cur Avg Loss: 1.71208974, Log Avg loss: 1.58273043, Global Avg Loss: 4.07053092, Time: 0.0403 Steps: 3490, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001371, Sample Num: 21936, Cur Loss: 2.99862623, Cur Avg Loss: 1.71412523, Log Avg loss: 1.99115490, Global Avg Loss: 4.06458984, Time: 0.0402 Steps: 3500, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001381, Sample Num: 22096, Cur Loss: 3.06656504, Cur Avg Loss: 1.71465537, Log Avg loss: 1.78733775, Global Avg Loss: 4.05810194, Time: 0.0402 Steps: 3510, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001391, Sample Num: 22256, Cur Loss: 1.50544167, Cur Avg Loss: 1.71385057, Log Avg loss: 1.60270736, Global Avg Loss: 4.05112639, Time: 0.0402 Steps: 3520, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001401, Sample Num: 22416, Cur Loss: 2.36487746, Cur Avg Loss: 1.71594911, Log Avg loss: 2.00785728, Global Avg Loss: 4.04533809, Time: 0.0402 Steps: 3530, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001411, Sample Num: 22576, Cur Loss: 3.29696727, Cur Avg Loss: 1.71779636, Log Avg loss: 1.97659578, Global Avg Loss: 4.03949419, Time: 0.0402 Steps: 3540, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001421, Sample Num: 22736, Cur Loss: 1.02940214, Cur Avg Loss: 1.71959033, Log Avg loss: 1.97271912, Global Avg Loss: 4.03367229, Time: 0.0402 Steps: 3550, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001431, Sample Num: 22896, Cur Loss: 1.00080752, Cur Avg Loss: 1.72037130, Log Avg loss: 1.83134669, Global Avg Loss: 4.02748598, Time: 0.0402 Steps: 3560, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001441, Sample Num: 23056, Cur Loss: 1.03858495, Cur Avg Loss: 1.72081622, Log Avg loss: 1.78448539, Global Avg Loss: 4.02120307, Time: 0.0403 Steps: 3570, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001451, Sample Num: 23216, Cur Loss: 1.86417556, Cur Avg Loss: 1.72002156, Log Avg loss: 1.60551107, Global Avg Loss: 4.01445532, Time: 0.0403 Steps: 3580, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001461, Sample Num: 23376, Cur Loss: 1.82243299, Cur Avg Loss: 1.72078563, Log Avg loss: 1.83165216, Global Avg Loss: 4.00837509, Time: 0.0402 Steps: 3590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001471, Sample Num: 23536, Cur Loss: 0.59025013, Cur Avg Loss: 1.71702015, Log Avg loss: 1.16688339, Global Avg Loss: 4.00048206, Time: 0.0403 Steps: 3600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001481, Sample Num: 23696, Cur Loss: 2.94520020, Cur Avg Loss: 1.71709610, Log Avg loss: 1.72826751, Global Avg Loss: 3.99418783, Time: 0.0402 Steps: 3610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001491, Sample Num: 23856, Cur Loss: 1.20078313, Cur Avg Loss: 1.71843834, Log Avg loss: 1.91722388, Global Avg Loss: 3.98845037, Time: 0.0402 Steps: 3620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001501, Sample Num: 24016, Cur Loss: 2.04793978, Cur Avg Loss: 1.71634568, Log Avg loss: 1.40433026, Global Avg Loss: 3.98133158, Time: 0.0404 Steps: 3630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001511, Sample Num: 24176, Cur Loss: 2.05972528, Cur Avg Loss: 1.71554665, Log Avg loss: 1.59561257, Global Avg Loss: 3.97477740, Time: 0.0402 Steps: 3640, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001521, Sample Num: 24336, Cur Loss: 0.58716249, Cur Avg Loss: 1.71526868, Log Avg loss: 1.67326745, Global Avg Loss: 3.96847190, Time: 0.0402 Steps: 3650, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001531, Sample Num: 24496, Cur Loss: 1.41881418, Cur Avg Loss: 1.71533802, Log Avg loss: 1.72588425, Global Avg Loss: 3.96234461, Time: 0.0402 Steps: 3660, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001541, Sample Num: 24656, Cur Loss: 0.88637233, Cur Avg Loss: 1.71357465, Log Avg loss: 1.44360293, Global Avg Loss: 3.95548155, Time: 0.0402 Steps: 3670, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001551, Sample Num: 24816, Cur Loss: 1.53258097, Cur Avg Loss: 1.71355383, Log Avg loss: 1.71034481, Global Avg Loss: 3.94938064, Time: 0.0402 Steps: 3680, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001561, Sample Num: 24976, Cur Loss: 1.43106675, Cur Avg Loss: 1.71214665, Log Avg loss: 1.49389367, Global Avg Loss: 3.94272620, Time: 0.0403 Steps: 3690, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001571, Sample Num: 25136, Cur Loss: 1.85120249, Cur Avg Loss: 1.71203690, Log Avg loss: 1.69490569, Global Avg Loss: 3.93665101, Time: 0.0402 Steps: 3700, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001581, Sample Num: 25296, Cur Loss: 0.92679983, Cur Avg Loss: 1.71061935, Log Avg loss: 1.48792177, Global Avg Loss: 3.93005066, Time: 0.0405 Steps: 3710, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001591, Sample Num: 25456, Cur Loss: 2.14349031, Cur Avg Loss: 1.71073974, Log Avg loss: 1.72977354, Global Avg Loss: 3.92413594, Time: 0.0406 Steps: 3720, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001601, Sample Num: 25616, Cur Loss: 0.63727587, Cur Avg Loss: 1.70887304, Log Avg loss: 1.41188114, Global Avg Loss: 3.91740067, Time: 0.0404 Steps: 3730, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001611, Sample Num: 25776, Cur Loss: 1.55865908, Cur Avg Loss: 1.71131222, Log Avg loss: 2.10182468, Global Avg Loss: 3.91254619, Time: 0.0613 Steps: 3740, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001621, Sample Num: 25936, Cur Loss: 2.72866535, Cur Avg Loss: 1.71208140, Log Avg loss: 1.83599644, Global Avg Loss: 3.90700872, Time: 0.0769 Steps: 3750, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001631, Sample Num: 26096, Cur Loss: 2.30455923, Cur Avg Loss: 1.71435279, Log Avg loss: 2.08254514, Global Avg Loss: 3.90215643, Time: 0.0554 Steps: 3760, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001641, Sample Num: 26256, Cur Loss: 0.70866799, Cur Avg Loss: 1.71187219, Log Avg loss: 1.30728580, Global Avg Loss: 3.89527348, Time: 0.0635 Steps: 3770, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001651, Sample Num: 26416, Cur Loss: 2.06681132, Cur Avg Loss: 1.71151709, Log Avg loss: 1.65324519, Global Avg Loss: 3.88934219, Time: 0.0756 Steps: 3780, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001661, Sample Num: 26576, Cur Loss: 0.67039686, Cur Avg Loss: 1.70993709, Log Avg loss: 1.44907869, Global Avg Loss: 3.88290350, Time: 0.0405 Steps: 3790, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001671, Sample Num: 26736, Cur Loss: 1.68427885, Cur Avg Loss: 1.70984641, Log Avg loss: 1.69478514, Global Avg Loss: 3.87714529, Time: 0.0402 Steps: 3800, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001681, Sample Num: 26896, Cur Loss: 1.39941049, Cur Avg Loss: 1.70982410, Log Avg loss: 1.70609666, Global Avg Loss: 3.87144700, Time: 0.0403 Steps: 3810, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001691, Sample Num: 27056, Cur Loss: 1.61029983, Cur Avg Loss: 1.71144524, Log Avg loss: 1.98395783, Global Avg Loss: 3.86650593, Time: 0.0402 Steps: 3820, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001701, Sample Num: 27216, Cur Loss: 1.65693545, Cur Avg Loss: 1.71270401, Log Avg loss: 1.92556252, Global Avg Loss: 3.86143819, Time: 0.0402 Steps: 3830, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001711, Sample Num: 27376, Cur Loss: 0.63259804, Cur Avg Loss: 1.71149958, Log Avg loss: 1.50662555, Global Avg Loss: 3.85530587, Time: 0.0402 Steps: 3840, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001721, Sample Num: 27536, Cur Loss: 1.40384352, Cur Avg Loss: 1.71433122, Log Avg loss: 2.19882559, Global Avg Loss: 3.85100332, Time: 0.0403 Steps: 3850, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001731, Sample Num: 27696, Cur Loss: 1.73375416, Cur Avg Loss: 1.71404278, Log Avg loss: 1.66440116, Global Avg Loss: 3.84533855, Time: 0.0403 Steps: 3860, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001741, Sample Num: 27856, Cur Loss: 0.63327664, Cur Avg Loss: 1.71386117, Log Avg loss: 1.68242550, Global Avg Loss: 3.83974963, Time: 0.0402 Steps: 3870, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001751, Sample Num: 28016, Cur Loss: 1.89448762, Cur Avg Loss: 1.71319467, Log Avg loss: 1.59715699, Global Avg Loss: 3.83396975, Time: 0.0402 Steps: 3880, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001761, Sample Num: 28176, Cur Loss: 1.13189411, Cur Avg Loss: 1.71417942, Log Avg loss: 1.88660935, Global Avg Loss: 3.82896368, Time: 0.0403 Steps: 3890, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001771, Sample Num: 28336, Cur Loss: 2.86109829, Cur Avg Loss: 1.71364682, Log Avg loss: 1.61985553, Global Avg Loss: 3.82329930, Time: 0.0402 Steps: 3900, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001781, Sample Num: 28496, Cur Loss: 2.77665234, Cur Avg Loss: 1.71464894, Log Avg loss: 1.89212413, Global Avg Loss: 3.81836024, Time: 0.0402 Steps: 3910, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001791, Sample Num: 28656, Cur Loss: 2.60930943, Cur Avg Loss: 1.71595564, Log Avg loss: 1.94867879, Global Avg Loss: 3.81359064, Time: 0.0402 Steps: 3920, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001801, Sample Num: 28816, Cur Loss: 1.17785501, Cur Avg Loss: 1.71517806, Log Avg loss: 1.57591419, Global Avg Loss: 3.80789681, Time: 0.0403 Steps: 3930, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001811, Sample Num: 28976, Cur Loss: 2.33990026, Cur Avg Loss: 1.71675315, Log Avg loss: 2.00042641, Global Avg Loss: 3.80330932, Time: 0.0402 Steps: 3940, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001821, Sample Num: 29136, Cur Loss: 2.29729390, Cur Avg Loss: 1.71806809, Log Avg loss: 1.95620315, Global Avg Loss: 3.79863310, Time: 0.0403 Steps: 3950, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001831, Sample Num: 29296, Cur Loss: 2.63490987, Cur Avg Loss: 1.71943313, Log Avg loss: 1.96800786, Global Avg Loss: 3.79401031, Time: 0.0402 Steps: 3960, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001841, Sample Num: 29456, Cur Loss: 1.37945890, Cur Avg Loss: 1.71773433, Log Avg loss: 1.40668336, Global Avg Loss: 3.78799689, Time: 0.0403 Steps: 3970, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001851, Sample Num: 29616, Cur Loss: 0.83653015, Cur Avg Loss: 1.71701253, Log Avg loss: 1.58413016, Global Avg Loss: 3.78245954, Time: 0.0403 Steps: 3980, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001861, Sample Num: 29776, Cur Loss: 1.64035988, Cur Avg Loss: 1.71773206, Log Avg loss: 1.85091577, Global Avg Loss: 3.77761858, Time: 0.0402 Steps: 3990, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001871, Sample Num: 29936, Cur Loss: 1.54062796, Cur Avg Loss: 1.71774747, Log Avg loss: 1.72061604, Global Avg Loss: 3.77247607, Time: 0.0402 Steps: 4000, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001881, Sample Num: 30096, Cur Loss: 2.34649801, Cur Avg Loss: 1.71906995, Log Avg loss: 1.96650549, Global Avg Loss: 3.76797240, Time: 0.0402 Steps: 4010, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001891, Sample Num: 30256, Cur Loss: 1.43852699, Cur Avg Loss: 1.72035141, Log Avg loss: 1.96139402, Global Avg Loss: 3.76347843, Time: 0.0403 Steps: 4020, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001901, Sample Num: 30416, Cur Loss: 2.38860083, Cur Avg Loss: 1.71956834, Log Avg loss: 1.57149047, Global Avg Loss: 3.75803925, Time: 0.0402 Steps: 4030, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001911, Sample Num: 30576, Cur Loss: 1.44460857, Cur Avg Loss: 1.71938743, Log Avg loss: 1.68499607, Global Avg Loss: 3.75290796, Time: 0.0402 Steps: 4040, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001921, Sample Num: 30736, Cur Loss: 0.79812491, Cur Avg Loss: 1.71923878, Log Avg loss: 1.69083185, Global Avg Loss: 3.74781641, Time: 0.0402 Steps: 4050, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001931, Sample Num: 30896, Cur Loss: 1.37431014, Cur Avg Loss: 1.72189304, Log Avg loss: 2.23177706, Global Avg Loss: 3.74408232, Time: 0.0665 Steps: 4060, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001941, Sample Num: 31056, Cur Loss: 1.90242743, Cur Avg Loss: 1.72138360, Log Avg loss: 1.62301099, Global Avg Loss: 3.73887085, Time: 0.0437 Steps: 4070, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001951, Sample Num: 31216, Cur Loss: 2.93741989, Cur Avg Loss: 1.72246355, Log Avg loss: 1.93208013, Global Avg Loss: 3.73444244, Time: 0.0589 Steps: 4080, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001961, Sample Num: 31376, Cur Loss: 2.11931229, Cur Avg Loss: 1.72273077, Log Avg loss: 1.77486635, Global Avg Loss: 3.72965130, Time: 0.0752 Steps: 4090, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001971, Sample Num: 31536, Cur Loss: 2.09537292, Cur Avg Loss: 1.72284674, Log Avg loss: 1.74558905, Global Avg Loss: 3.72481212, Time: 0.0913 Steps: 4100, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001981, Sample Num: 31696, Cur Loss: 2.11973047, Cur Avg Loss: 1.72281207, Log Avg loss: 1.71597861, Global Avg Loss: 3.71992445, Time: 0.0548 Steps: 4110, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001991, Sample Num: 31856, Cur Loss: 3.10230565, Cur Avg Loss: 1.72106432, Log Avg loss: 1.37483392, Global Avg Loss: 3.71423248, Time: 0.0501 Steps: 4120, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002001, Sample Num: 32016, Cur Loss: 2.62264085, Cur Avg Loss: 1.72193471, Log Avg loss: 1.89522961, Global Avg Loss: 3.70982812, Time: 0.0976 Steps: 4130, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002011, Sample Num: 32176, Cur Loss: 0.89033329, Cur Avg Loss: 1.72151426, Log Avg loss: 1.63738140, Global Avg Loss: 3.70482221, Time: 0.0403 Steps: 4140, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002021, Sample Num: 32336, Cur Loss: 1.83457494, Cur Avg Loss: 1.72102324, Log Avg loss: 1.62228042, Global Avg Loss: 3.69980403, Time: 0.0403 Steps: 4150, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002031, Sample Num: 32496, Cur Loss: 2.09108663, Cur Avg Loss: 1.72053749, Log Avg loss: 1.62236638, Global Avg Loss: 3.69481019, Time: 0.0402 Steps: 4160, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002041, Sample Num: 32656, Cur Loss: 1.79951000, Cur Avg Loss: 1.72158619, Log Avg loss: 1.93457754, Global Avg Loss: 3.69058901, Time: 0.0402 Steps: 4170, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002051, Sample Num: 32816, Cur Loss: 1.63371551, Cur Avg Loss: 1.72146347, Log Avg loss: 1.69641657, Global Avg Loss: 3.68581826, Time: 0.0402 Steps: 4180, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002061, Sample Num: 32976, Cur Loss: 1.42078114, Cur Avg Loss: 1.72148600, Log Avg loss: 1.72610694, Global Avg Loss: 3.68114115, Time: 0.0402 Steps: 4190, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002071, Sample Num: 33136, Cur Loss: 1.35826385, Cur Avg Loss: 1.72004480, Log Avg loss: 1.42301371, Global Avg Loss: 3.67576465, Time: 0.0402 Steps: 4200, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002081, Sample Num: 33296, Cur Loss: 1.45189440, Cur Avg Loss: 1.72165508, Log Avg loss: 2.05514299, Global Avg Loss: 3.67191520, Time: 0.0402 Steps: 4210, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002091, Sample Num: 33456, Cur Loss: 2.02935195, Cur Avg Loss: 1.72256858, Log Avg loss: 1.91266981, Global Avg Loss: 3.66774637, Time: 0.0403 Steps: 4220, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002101, Sample Num: 33616, Cur Loss: 1.93272913, Cur Avg Loss: 1.72243960, Log Avg loss: 1.69546989, Global Avg Loss: 3.66308378, Time: 0.0402 Steps: 4230, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002111, Sample Num: 33776, Cur Loss: 1.53834009, Cur Avg Loss: 1.72197915, Log Avg loss: 1.62523805, Global Avg Loss: 3.65827754, Time: 0.0402 Steps: 4240, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002121, Sample Num: 33936, Cur Loss: 1.88108182, Cur Avg Loss: 1.72200324, Log Avg loss: 1.72708923, Global Avg Loss: 3.65373356, Time: 0.0402 Steps: 4250, Updated lr: 0.000097 ***** Running evaluation checkpoint-4258 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-4258 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.597164, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.575116, "eval_total_loss": 1107.306784, "eval_mae": 1.083501, "eval_mse": 1.575083, "eval_r2": -0.001226, "eval_sp_statistic": 0.099379, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.095274, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.729425, "test_total_loss": 868.17119, "test_mae": 1.191374, "test_mse": 1.729446, "test_r2": -0.1162, "test_sp_statistic": 0.123249, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.084481, "test_ps_pvalue": 0.0, "lr": 9.691038406827881e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.6502676062800337, "train_cur_epoch_loss": 3666.8407046049833, "train_cur_epoch_avg_loss": 1.7223300632245107, "train_cur_epoch_time": 95.59716415405273, "train_cur_epoch_avg_time": 0.04490237865385286, "epoch": 2, "step": 4258} ################################################## Training, Epoch: 0003, Batch: 000002, Sample Num: 32, Cur Loss: 1.53153396, Cur Avg Loss: 1.91135156, Log Avg loss: 1.82945255, Global Avg Loss: 3.64945121, Time: 0.0404 Steps: 4260, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000012, Sample Num: 192, Cur Loss: 1.56524074, Cur Avg Loss: 1.69925099, Log Avg loss: 1.65683087, Global Avg Loss: 3.64478466, Time: 0.0402 Steps: 4270, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000022, Sample Num: 352, Cur Loss: 2.59623241, Cur Avg Loss: 1.59120336, Log Avg loss: 1.46154620, Global Avg Loss: 3.63968363, Time: 0.0402 Steps: 4280, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000032, Sample Num: 512, Cur Loss: 1.32710385, Cur Avg Loss: 1.55757084, Log Avg loss: 1.48357931, Global Avg Loss: 3.63465775, Time: 0.0402 Steps: 4290, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000042, Sample Num: 672, Cur Loss: 0.90632701, Cur Avg Loss: 1.59845572, Log Avg loss: 1.72928731, Global Avg Loss: 3.63022665, Time: 0.0402 Steps: 4300, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000052, Sample Num: 832, Cur Loss: 3.26148510, Cur Avg Loss: 1.65817264, Log Avg loss: 1.90898374, Global Avg Loss: 3.62623305, Time: 0.0402 Steps: 4310, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000062, Sample Num: 992, Cur Loss: 2.11803532, Cur Avg Loss: 1.64853201, Log Avg loss: 1.59840071, Global Avg Loss: 3.62153899, Time: 0.0402 Steps: 4320, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000072, Sample Num: 1152, Cur Loss: 1.87227309, Cur Avg Loss: 1.62449200, Log Avg loss: 1.47544395, Global Avg Loss: 3.61658265, Time: 0.0402 Steps: 4330, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000082, Sample Num: 1312, Cur Loss: 3.36503887, Cur Avg Loss: 1.64288043, Log Avg loss: 1.77527716, Global Avg Loss: 3.61234001, Time: 0.0402 Steps: 4340, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000092, Sample Num: 1472, Cur Loss: 1.09323978, Cur Avg Loss: 1.65283341, Log Avg loss: 1.73444781, Global Avg Loss: 3.60802302, Time: 0.0402 Steps: 4350, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000102, Sample Num: 1632, Cur Loss: 1.72966862, Cur Avg Loss: 1.68577688, Log Avg loss: 1.98885678, Global Avg Loss: 3.60430934, Time: 0.0404 Steps: 4360, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000112, Sample Num: 1792, Cur Loss: 1.25702941, Cur Avg Loss: 1.68628011, Log Avg loss: 1.69141311, Global Avg Loss: 3.59993200, Time: 0.0405 Steps: 4370, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000122, Sample Num: 1952, Cur Loss: 2.63202095, Cur Avg Loss: 1.69624115, Log Avg loss: 1.80780472, Global Avg Loss: 3.59584039, Time: 0.0402 Steps: 4380, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000132, Sample Num: 2112, Cur Loss: 2.08640218, Cur Avg Loss: 1.71173226, Log Avg loss: 1.90072379, Global Avg Loss: 3.59197907, Time: 0.0402 Steps: 4390, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000142, Sample Num: 2272, Cur Loss: 1.09085870, Cur Avg Loss: 1.70929873, Log Avg loss: 1.67717614, Global Avg Loss: 3.58762725, Time: 0.0402 Steps: 4400, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000152, Sample Num: 2432, Cur Loss: 1.43914449, Cur Avg Loss: 1.73749388, Log Avg loss: 2.13786512, Global Avg Loss: 3.58433980, Time: 0.0402 Steps: 4410, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000162, Sample Num: 2592, Cur Loss: 2.22437930, Cur Avg Loss: 1.75039627, Log Avg loss: 1.94651258, Global Avg Loss: 3.58063431, Time: 0.0402 Steps: 4420, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000172, Sample Num: 2752, Cur Loss: 1.45979059, Cur Avg Loss: 1.74482510, Log Avg loss: 1.65457203, Global Avg Loss: 3.57628654, Time: 0.0402 Steps: 4430, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000182, Sample Num: 2912, Cur Loss: 1.11200356, Cur Avg Loss: 1.73438276, Log Avg loss: 1.55477457, Global Avg Loss: 3.57173359, Time: 0.0401 Steps: 4440, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000192, Sample Num: 3072, Cur Loss: 1.21214271, Cur Avg Loss: 1.73340327, Log Avg loss: 1.71557653, Global Avg Loss: 3.56756245, Time: 0.0402 Steps: 4450, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000202, Sample Num: 3232, Cur Loss: 1.57130635, Cur Avg Loss: 1.74227634, Log Avg loss: 1.91263940, Global Avg Loss: 3.56385186, Time: 0.0402 Steps: 4460, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000212, Sample Num: 3392, Cur Loss: 1.15751457, Cur Avg Loss: 1.71654132, Log Avg loss: 1.19669384, Global Avg Loss: 3.55855620, Time: 0.0402 Steps: 4470, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000222, Sample Num: 3552, Cur Loss: 2.59583807, Cur Avg Loss: 1.73725991, Log Avg loss: 2.17649392, Global Avg Loss: 3.55547124, Time: 0.0617 Steps: 4480, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000232, Sample Num: 3712, Cur Loss: 4.49852800, Cur Avg Loss: 1.75180736, Log Avg loss: 2.07476091, Global Avg Loss: 3.55217345, Time: 0.0648 Steps: 4490, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000242, Sample Num: 3872, Cur Loss: 1.21145248, Cur Avg Loss: 1.74037032, Log Avg loss: 1.47503085, Global Avg Loss: 3.54755757, Time: 0.0647 Steps: 4500, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000252, Sample Num: 4032, Cur Loss: 0.61425018, Cur Avg Loss: 1.71878276, Log Avg loss: 1.19636379, Global Avg Loss: 3.54234428, Time: 0.0492 Steps: 4510, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000262, Sample Num: 4192, Cur Loss: 1.51276207, Cur Avg Loss: 1.70999873, Log Avg loss: 1.48864116, Global Avg Loss: 3.53780069, Time: 0.0416 Steps: 4520, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000272, Sample Num: 4352, Cur Loss: 0.58982158, Cur Avg Loss: 1.71387572, Log Avg loss: 1.81545296, Global Avg Loss: 3.53399860, Time: 0.0411 Steps: 4530, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000282, Sample Num: 4512, Cur Loss: 1.86641371, Cur Avg Loss: 1.71650124, Log Avg loss: 1.78791529, Global Avg Loss: 3.53015260, Time: 0.0457 Steps: 4540, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000292, Sample Num: 4672, Cur Loss: 1.83982658, Cur Avg Loss: 1.71350438, Log Avg loss: 1.62899299, Global Avg Loss: 3.52597423, Time: 0.0630 Steps: 4550, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000302, Sample Num: 4832, Cur Loss: 1.74119115, Cur Avg Loss: 1.70681588, Log Avg loss: 1.51151162, Global Avg Loss: 3.52155655, Time: 0.0474 Steps: 4560, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000312, Sample Num: 4992, Cur Loss: 2.57190800, Cur Avg Loss: 1.71638409, Log Avg loss: 2.00534415, Global Avg Loss: 3.51823880, Time: 0.0403 Steps: 4570, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000322, Sample Num: 5152, Cur Loss: 0.92464149, Cur Avg Loss: 1.70830732, Log Avg loss: 1.45631200, Global Avg Loss: 3.51373677, Time: 0.0402 Steps: 4580, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000332, Sample Num: 5312, Cur Loss: 2.67072845, Cur Avg Loss: 1.70613444, Log Avg loss: 1.63616763, Global Avg Loss: 3.50964621, Time: 0.0402 Steps: 4590, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000342, Sample Num: 5472, Cur Loss: 1.20158505, Cur Avg Loss: 1.71022045, Log Avg loss: 1.84587612, Global Avg Loss: 3.50602932, Time: 0.0402 Steps: 4600, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000352, Sample Num: 5632, Cur Loss: 1.23356259, Cur Avg Loss: 1.70850463, Log Avg loss: 1.64982345, Global Avg Loss: 3.50200284, Time: 0.0403 Steps: 4610, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000362, Sample Num: 5792, Cur Loss: 1.95665193, Cur Avg Loss: 1.71113393, Log Avg loss: 1.80368544, Global Avg Loss: 3.49832683, Time: 0.0402 Steps: 4620, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000372, Sample Num: 5952, Cur Loss: 1.14369738, Cur Avg Loss: 1.70595960, Log Avg loss: 1.51864903, Global Avg Loss: 3.49405107, Time: 0.0403 Steps: 4630, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000382, Sample Num: 6112, Cur Loss: 3.87831020, Cur Avg Loss: 1.71185824, Log Avg loss: 1.93128766, Global Avg Loss: 3.49068304, Time: 0.0402 Steps: 4640, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000392, Sample Num: 6272, Cur Loss: 2.03580475, Cur Avg Loss: 1.71351634, Log Avg loss: 1.77685558, Global Avg Loss: 3.48699739, Time: 0.0402 Steps: 4650, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000402, Sample Num: 6432, Cur Loss: 2.83310056, Cur Avg Loss: 1.71497371, Log Avg loss: 1.77210253, Global Avg Loss: 3.48331736, Time: 0.0403 Steps: 4660, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000412, Sample Num: 6592, Cur Loss: 1.92308474, Cur Avg Loss: 1.72039483, Log Avg loss: 1.93832389, Global Avg Loss: 3.48000902, Time: 0.0402 Steps: 4670, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000422, Sample Num: 6752, Cur Loss: 1.15249527, Cur Avg Loss: 1.71542193, Log Avg loss: 1.51053830, Global Avg Loss: 3.47580075, Time: 0.0402 Steps: 4680, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000432, Sample Num: 6912, Cur Loss: 1.30165458, Cur Avg Loss: 1.71747084, Log Avg loss: 1.80393510, Global Avg Loss: 3.47223601, Time: 0.0402 Steps: 4690, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000442, Sample Num: 7072, Cur Loss: 0.55573535, Cur Avg Loss: 1.72170811, Log Avg loss: 1.90475788, Global Avg Loss: 3.46890095, Time: 0.0402 Steps: 4700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000452, Sample Num: 7232, Cur Loss: 0.87624878, Cur Avg Loss: 1.72778443, Log Avg loss: 1.99635797, Global Avg Loss: 3.46577453, Time: 0.0402 Steps: 4710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000462, Sample Num: 7392, Cur Loss: 1.49805546, Cur Avg Loss: 1.72983427, Log Avg loss: 1.82248700, Global Avg Loss: 3.46229299, Time: 0.0402 Steps: 4720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000472, Sample Num: 7552, Cur Loss: 1.39662671, Cur Avg Loss: 1.72729071, Log Avg loss: 1.60977815, Global Avg Loss: 3.45837647, Time: 0.0402 Steps: 4730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000482, Sample Num: 7712, Cur Loss: 2.02519703, Cur Avg Loss: 1.72869046, Log Avg loss: 1.79475857, Global Avg Loss: 3.45486672, Time: 0.0402 Steps: 4740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000492, Sample Num: 7872, Cur Loss: 0.86747527, Cur Avg Loss: 1.72870936, Log Avg loss: 1.72962066, Global Avg Loss: 3.45123463, Time: 0.0403 Steps: 4750, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000502, Sample Num: 8032, Cur Loss: 1.66664207, Cur Avg Loss: 1.71703784, Log Avg loss: 1.14279892, Global Avg Loss: 3.44638497, Time: 0.0402 Steps: 4760, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000512, Sample Num: 8192, Cur Loss: 1.00712740, Cur Avg Loss: 1.71675639, Log Avg loss: 1.70262768, Global Avg Loss: 3.44272930, Time: 0.0402 Steps: 4770, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000522, Sample Num: 8352, Cur Loss: 2.10102725, Cur Avg Loss: 1.71515118, Log Avg loss: 1.63296431, Global Avg Loss: 3.43894318, Time: 0.0402 Steps: 4780, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000532, Sample Num: 8512, Cur Loss: 1.62874365, Cur Avg Loss: 1.70805034, Log Avg loss: 1.33738677, Global Avg Loss: 3.43455579, Time: 0.0403 Steps: 4790, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000542, Sample Num: 8672, Cur Loss: 2.73427463, Cur Avg Loss: 1.71585973, Log Avg loss: 2.13131918, Global Avg Loss: 3.43184072, Time: 0.0402 Steps: 4800, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000552, Sample Num: 8832, Cur Loss: 2.12444353, Cur Avg Loss: 1.71504542, Log Avg loss: 1.67090972, Global Avg Loss: 3.42817974, Time: 0.0402 Steps: 4810, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000562, Sample Num: 8992, Cur Loss: 1.73983800, Cur Avg Loss: 1.71430818, Log Avg loss: 1.67361245, Global Avg Loss: 3.42453956, Time: 0.0577 Steps: 4820, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000572, Sample Num: 9152, Cur Loss: 1.36986387, Cur Avg Loss: 1.71162774, Log Avg loss: 1.56098706, Global Avg Loss: 3.42068127, Time: 0.0588 Steps: 4830, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000582, Sample Num: 9312, Cur Loss: 2.88210893, Cur Avg Loss: 1.71195782, Log Avg loss: 1.73083835, Global Avg Loss: 3.41718986, Time: 0.0420 Steps: 4840, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000592, Sample Num: 9472, Cur Loss: 2.66546750, Cur Avg Loss: 1.70948185, Log Avg loss: 1.56538070, Global Avg Loss: 3.41337170, Time: 0.0404 Steps: 4850, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000602, Sample Num: 9632, Cur Loss: 1.21379137, Cur Avg Loss: 1.70573754, Log Avg loss: 1.48407395, Global Avg Loss: 3.40940195, Time: 0.0565 Steps: 4860, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000612, Sample Num: 9792, Cur Loss: 1.79771626, Cur Avg Loss: 1.70232309, Log Avg loss: 1.49677361, Global Avg Loss: 3.40547458, Time: 0.0521 Steps: 4870, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000622, Sample Num: 9952, Cur Loss: 0.78065968, Cur Avg Loss: 1.70238829, Log Avg loss: 1.70637854, Global Avg Loss: 3.40199282, Time: 0.0509 Steps: 4880, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000632, Sample Num: 10112, Cur Loss: 2.35663319, Cur Avg Loss: 1.70030091, Log Avg loss: 1.57046592, Global Avg Loss: 3.39824737, Time: 0.0460 Steps: 4890, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000642, Sample Num: 10272, Cur Loss: 1.40950274, Cur Avg Loss: 1.69908617, Log Avg loss: 1.62231438, Global Avg Loss: 3.39462302, Time: 0.0404 Steps: 4900, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000652, Sample Num: 10432, Cur Loss: 2.35442781, Cur Avg Loss: 1.70090000, Log Avg loss: 1.81734802, Global Avg Loss: 3.39141065, Time: 0.0402 Steps: 4910, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000662, Sample Num: 10592, Cur Loss: 2.63956165, Cur Avg Loss: 1.69968266, Log Avg loss: 1.62031195, Global Avg Loss: 3.38781085, Time: 0.0402 Steps: 4920, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000672, Sample Num: 10752, Cur Loss: 2.88476491, Cur Avg Loss: 1.70493864, Log Avg loss: 2.05288447, Global Avg Loss: 3.38510309, Time: 0.0403 Steps: 4930, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000682, Sample Num: 10912, Cur Loss: 1.05965137, Cur Avg Loss: 1.70428135, Log Avg loss: 1.66011130, Global Avg Loss: 3.38161120, Time: 0.0402 Steps: 4940, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000692, Sample Num: 11072, Cur Loss: 0.91134763, Cur Avg Loss: 1.69673961, Log Avg loss: 1.18239284, Global Avg Loss: 3.37716834, Time: 0.0402 Steps: 4950, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000702, Sample Num: 11232, Cur Loss: 1.44804466, Cur Avg Loss: 1.69110500, Log Avg loss: 1.30119064, Global Avg Loss: 3.37298290, Time: 0.0402 Steps: 4960, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000712, Sample Num: 11392, Cur Loss: 1.30642605, Cur Avg Loss: 1.69295389, Log Avg loss: 1.82274571, Global Avg Loss: 3.36986371, Time: 0.0402 Steps: 4970, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000722, Sample Num: 11552, Cur Loss: 1.76748657, Cur Avg Loss: 1.69237266, Log Avg loss: 1.65098877, Global Avg Loss: 3.36641215, Time: 0.0402 Steps: 4980, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000732, Sample Num: 11712, Cur Loss: 1.86398017, Cur Avg Loss: 1.68916961, Log Avg loss: 1.45790969, Global Avg Loss: 3.36258750, Time: 0.0403 Steps: 4990, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000742, Sample Num: 11872, Cur Loss: 1.84245515, Cur Avg Loss: 1.68521656, Log Avg loss: 1.39585361, Global Avg Loss: 3.35865403, Time: 0.0403 Steps: 5000, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000752, Sample Num: 12032, Cur Loss: 1.73033714, Cur Avg Loss: 1.68692598, Log Avg loss: 1.81376481, Global Avg Loss: 3.35557042, Time: 0.0402 Steps: 5010, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000762, Sample Num: 12192, Cur Loss: 2.47363448, Cur Avg Loss: 1.68481207, Log Avg loss: 1.52584559, Global Avg Loss: 3.35192555, Time: 0.0402 Steps: 5020, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000772, Sample Num: 12352, Cur Loss: 2.56110716, Cur Avg Loss: 1.68774941, Log Avg loss: 1.91157472, Global Avg Loss: 3.34906203, Time: 0.0402 Steps: 5030, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000782, Sample Num: 12512, Cur Loss: 1.49692488, Cur Avg Loss: 1.68782570, Log Avg loss: 1.69371527, Global Avg Loss: 3.34577761, Time: 0.0403 Steps: 5040, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000792, Sample Num: 12672, Cur Loss: 1.96184123, Cur Avg Loss: 1.68741989, Log Avg loss: 1.65568546, Global Avg Loss: 3.34243089, Time: 0.0402 Steps: 5050, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000802, Sample Num: 12832, Cur Loss: 1.15755415, Cur Avg Loss: 1.68400004, Log Avg loss: 1.41314853, Global Avg Loss: 3.33861808, Time: 0.0402 Steps: 5060, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000812, Sample Num: 12992, Cur Loss: 1.96340263, Cur Avg Loss: 1.68640183, Log Avg loss: 1.87902504, Global Avg Loss: 3.33573920, Time: 0.0403 Steps: 5070, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000822, Sample Num: 13152, Cur Loss: 1.61865032, Cur Avg Loss: 1.68733131, Log Avg loss: 1.76280516, Global Avg Loss: 3.33264287, Time: 0.0402 Steps: 5080, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000832, Sample Num: 13312, Cur Loss: 0.47250015, Cur Avg Loss: 1.68638625, Log Avg loss: 1.60870262, Global Avg Loss: 3.32925596, Time: 0.0403 Steps: 5090, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000842, Sample Num: 13472, Cur Loss: 1.24321330, Cur Avg Loss: 1.69039466, Log Avg loss: 2.02389421, Global Avg Loss: 3.32669643, Time: 0.0402 Steps: 5100, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000852, Sample Num: 13632, Cur Loss: 0.61452961, Cur Avg Loss: 1.69183197, Log Avg loss: 1.81285322, Global Avg Loss: 3.32373391, Time: 0.0402 Steps: 5110, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000862, Sample Num: 13792, Cur Loss: 2.04338217, Cur Avg Loss: 1.69224552, Log Avg loss: 1.72748011, Global Avg Loss: 3.32061623, Time: 0.0402 Steps: 5120, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000872, Sample Num: 13952, Cur Loss: 1.36903000, Cur Avg Loss: 1.69665224, Log Avg loss: 2.07651147, Global Avg Loss: 3.31819108, Time: 0.0402 Steps: 5130, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000882, Sample Num: 14112, Cur Loss: 0.78536779, Cur Avg Loss: 1.69299573, Log Avg loss: 1.37414830, Global Avg Loss: 3.31440889, Time: 0.0402 Steps: 5140, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000892, Sample Num: 14272, Cur Loss: 1.20649445, Cur Avg Loss: 1.69073713, Log Avg loss: 1.49152852, Global Avg Loss: 3.31086932, Time: 0.0403 Steps: 5150, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000902, Sample Num: 14432, Cur Loss: 1.53551245, Cur Avg Loss: 1.69203874, Log Avg loss: 1.80814192, Global Avg Loss: 3.30795706, Time: 0.0403 Steps: 5160, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000912, Sample Num: 14592, Cur Loss: 1.08096790, Cur Avg Loss: 1.69376074, Log Avg loss: 1.84908568, Global Avg Loss: 3.30513525, Time: 0.0535 Steps: 5170, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000922, Sample Num: 14752, Cur Loss: 1.43910646, Cur Avg Loss: 1.69733025, Log Avg loss: 2.02286905, Global Avg Loss: 3.30265984, Time: 0.0649 Steps: 5180, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000932, Sample Num: 14912, Cur Loss: 1.05655479, Cur Avg Loss: 1.69395364, Log Avg loss: 1.38263034, Global Avg Loss: 3.29896036, Time: 0.0548 Steps: 5190, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000942, Sample Num: 15072, Cur Loss: 1.92576802, Cur Avg Loss: 1.69654646, Log Avg loss: 1.93819716, Global Avg Loss: 3.29634351, Time: 0.0590 Steps: 5200, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000952, Sample Num: 15232, Cur Loss: 1.46020484, Cur Avg Loss: 1.69478071, Log Avg loss: 1.52844731, Global Avg Loss: 3.29295023, Time: 0.0405 Steps: 5210, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000962, Sample Num: 15392, Cur Loss: 1.09510851, Cur Avg Loss: 1.69487384, Log Avg loss: 1.70373973, Global Avg Loss: 3.28990577, Time: 0.0825 Steps: 5220, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000972, Sample Num: 15552, Cur Loss: 1.36699152, Cur Avg Loss: 1.69378351, Log Avg loss: 1.58889392, Global Avg Loss: 3.28665335, Time: 0.0563 Steps: 5230, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000982, Sample Num: 15712, Cur Loss: 0.81054616, Cur Avg Loss: 1.69462323, Log Avg loss: 1.77624420, Global Avg Loss: 3.28377089, Time: 0.0603 Steps: 5240, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000992, Sample Num: 15872, Cur Loss: 1.85505021, Cur Avg Loss: 1.69575851, Log Avg loss: 1.80724258, Global Avg Loss: 3.28095846, Time: 0.0827 Steps: 5250, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001002, Sample Num: 16032, Cur Loss: 0.69539058, Cur Avg Loss: 1.69616798, Log Avg loss: 1.73678714, Global Avg Loss: 3.27802277, Time: 0.0403 Steps: 5260, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001012, Sample Num: 16192, Cur Loss: 1.86208677, Cur Avg Loss: 1.69592774, Log Avg loss: 1.67185622, Global Avg Loss: 3.27497502, Time: 0.0403 Steps: 5270, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001022, Sample Num: 16352, Cur Loss: 1.05269253, Cur Avg Loss: 1.69606692, Log Avg loss: 1.71015133, Global Avg Loss: 3.27201134, Time: 0.0403 Steps: 5280, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001032, Sample Num: 16512, Cur Loss: 2.63685632, Cur Avg Loss: 1.69991069, Log Avg loss: 2.09274476, Global Avg Loss: 3.26978210, Time: 0.0402 Steps: 5290, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001042, Sample Num: 16672, Cur Loss: 1.54019678, Cur Avg Loss: 1.69788608, Log Avg loss: 1.48894625, Global Avg Loss: 3.26642203, Time: 0.0402 Steps: 5300, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001052, Sample Num: 16832, Cur Loss: 2.38385010, Cur Avg Loss: 1.69786778, Log Avg loss: 1.69596063, Global Avg Loss: 3.26346448, Time: 0.0402 Steps: 5310, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001062, Sample Num: 16992, Cur Loss: 3.04500794, Cur Avg Loss: 1.69797793, Log Avg loss: 1.70956613, Global Avg Loss: 3.26054362, Time: 0.0402 Steps: 5320, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001072, Sample Num: 17152, Cur Loss: 0.83199823, Cur Avg Loss: 1.69675181, Log Avg loss: 1.56653782, Global Avg Loss: 3.25736537, Time: 0.0402 Steps: 5330, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001082, Sample Num: 17312, Cur Loss: 2.31448364, Cur Avg Loss: 1.69437824, Log Avg loss: 1.43993133, Global Avg Loss: 3.25396193, Time: 0.0403 Steps: 5340, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001092, Sample Num: 17472, Cur Loss: 1.66665721, Cur Avg Loss: 1.69334737, Log Avg loss: 1.58180728, Global Avg Loss: 3.25083641, Time: 0.0402 Steps: 5350, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001102, Sample Num: 17632, Cur Loss: 1.70226920, Cur Avg Loss: 1.69197559, Log Avg loss: 1.54217694, Global Avg Loss: 3.24764861, Time: 0.0403 Steps: 5360, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001112, Sample Num: 17792, Cur Loss: 2.47427130, Cur Avg Loss: 1.68887357, Log Avg loss: 1.34703085, Global Avg Loss: 3.24410929, Time: 0.0403 Steps: 5370, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001122, Sample Num: 17952, Cur Loss: 1.42260146, Cur Avg Loss: 1.68983917, Log Avg loss: 1.79721421, Global Avg Loss: 3.24141989, Time: 0.0403 Steps: 5380, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001132, Sample Num: 18112, Cur Loss: 0.79490745, Cur Avg Loss: 1.69395582, Log Avg loss: 2.15584383, Global Avg Loss: 3.23940584, Time: 0.0402 Steps: 5390, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001142, Sample Num: 18272, Cur Loss: 0.71349752, Cur Avg Loss: 1.69238403, Log Avg loss: 1.51445752, Global Avg Loss: 3.23621149, Time: 0.0403 Steps: 5400, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001152, Sample Num: 18432, Cur Loss: 2.03799868, Cur Avg Loss: 1.69056738, Log Avg loss: 1.48310533, Global Avg Loss: 3.23297100, Time: 0.0402 Steps: 5410, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001162, Sample Num: 18592, Cur Loss: 1.11213362, Cur Avg Loss: 1.69513226, Log Avg loss: 2.22100701, Global Avg Loss: 3.23110390, Time: 0.0403 Steps: 5420, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001172, Sample Num: 18752, Cur Loss: 0.68020838, Cur Avg Loss: 1.68996816, Log Avg loss: 1.08989982, Global Avg Loss: 3.22716062, Time: 0.0402 Steps: 5430, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001182, Sample Num: 18912, Cur Loss: 1.32341039, Cur Avg Loss: 1.68863846, Log Avg loss: 1.53279746, Global Avg Loss: 3.22404598, Time: 0.0402 Steps: 5440, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001192, Sample Num: 19072, Cur Loss: 1.51071239, Cur Avg Loss: 1.69020450, Log Avg loss: 1.87531052, Global Avg Loss: 3.22157124, Time: 0.0403 Steps: 5450, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001202, Sample Num: 19232, Cur Loss: 1.36249042, Cur Avg Loss: 1.68929351, Log Avg loss: 1.58070291, Global Avg Loss: 3.21856598, Time: 0.0402 Steps: 5460, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001212, Sample Num: 19392, Cur Loss: 2.14245033, Cur Avg Loss: 1.68790870, Log Avg loss: 1.52145544, Global Avg Loss: 3.21546340, Time: 0.0402 Steps: 5470, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001222, Sample Num: 19552, Cur Loss: 0.54840046, Cur Avg Loss: 1.68689491, Log Avg loss: 1.56402314, Global Avg Loss: 3.21244983, Time: 0.0402 Steps: 5480, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001232, Sample Num: 19712, Cur Loss: 1.36882055, Cur Avg Loss: 1.68937177, Log Avg loss: 1.99204370, Global Avg Loss: 3.21022686, Time: 0.0402 Steps: 5490, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001242, Sample Num: 19872, Cur Loss: 3.04173350, Cur Avg Loss: 1.68829004, Log Avg loss: 1.55502094, Global Avg Loss: 3.20721740, Time: 0.0402 Steps: 5500, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001252, Sample Num: 20032, Cur Loss: 1.64968133, Cur Avg Loss: 1.68724973, Log Avg loss: 1.55804315, Global Avg Loss: 3.20422434, Time: 0.0402 Steps: 5510, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001262, Sample Num: 20192, Cur Loss: 2.25699759, Cur Avg Loss: 1.68749914, Log Avg loss: 1.71872579, Global Avg Loss: 3.20153322, Time: 0.0405 Steps: 5520, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001272, Sample Num: 20352, Cur Loss: 3.24227691, Cur Avg Loss: 1.68691880, Log Avg loss: 1.61367942, Global Avg Loss: 3.19866188, Time: 0.0510 Steps: 5530, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001282, Sample Num: 20512, Cur Loss: 1.45707417, Cur Avg Loss: 1.68659033, Log Avg loss: 1.64480952, Global Avg Loss: 3.19585709, Time: 0.0704 Steps: 5540, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001292, Sample Num: 20672, Cur Loss: 0.37865648, Cur Avg Loss: 1.68282130, Log Avg loss: 1.19963194, Global Avg Loss: 3.19226029, Time: 0.0404 Steps: 5550, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001302, Sample Num: 20832, Cur Loss: 1.31456935, Cur Avg Loss: 1.68197224, Log Avg loss: 1.57227387, Global Avg Loss: 3.18934664, Time: 0.0596 Steps: 5560, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001312, Sample Num: 20992, Cur Loss: 2.01556110, Cur Avg Loss: 1.68609800, Log Avg loss: 2.22327089, Global Avg Loss: 3.18761222, Time: 0.0519 Steps: 5570, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001322, Sample Num: 21152, Cur Loss: 1.57410240, Cur Avg Loss: 1.68456118, Log Avg loss: 1.48293065, Global Avg Loss: 3.18455723, Time: 0.0536 Steps: 5580, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001332, Sample Num: 21312, Cur Loss: 1.90311837, Cur Avg Loss: 1.68378972, Log Avg loss: 1.58180310, Global Avg Loss: 3.18169005, Time: 0.0852 Steps: 5590, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001342, Sample Num: 21472, Cur Loss: 2.18367124, Cur Avg Loss: 1.68427975, Log Avg loss: 1.74955218, Global Avg Loss: 3.17913266, Time: 0.0719 Steps: 5600, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001352, Sample Num: 21632, Cur Loss: 1.96865547, Cur Avg Loss: 1.68707313, Log Avg loss: 2.06194376, Global Avg Loss: 3.17714124, Time: 0.0403 Steps: 5610, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001362, Sample Num: 21792, Cur Loss: 2.00197268, Cur Avg Loss: 1.68681638, Log Avg loss: 1.65210352, Global Avg Loss: 3.17442765, Time: 0.0402 Steps: 5620, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001372, Sample Num: 21952, Cur Loss: 1.38441157, Cur Avg Loss: 1.68384214, Log Avg loss: 1.27875115, Global Avg Loss: 3.17106055, Time: 0.0402 Steps: 5630, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001382, Sample Num: 22112, Cur Loss: 2.35798407, Cur Avg Loss: 1.68213708, Log Avg loss: 1.44820247, Global Avg Loss: 3.16800583, Time: 0.0402 Steps: 5640, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001392, Sample Num: 22272, Cur Loss: 2.47740531, Cur Avg Loss: 1.68140554, Log Avg loss: 1.58030722, Global Avg Loss: 3.16519575, Time: 0.0403 Steps: 5650, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001402, Sample Num: 22432, Cur Loss: 1.63706017, Cur Avg Loss: 1.68202645, Log Avg loss: 1.76845657, Global Avg Loss: 3.16272801, Time: 0.0402 Steps: 5660, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001412, Sample Num: 22592, Cur Loss: 1.47955358, Cur Avg Loss: 1.68485906, Log Avg loss: 2.08199111, Global Avg Loss: 3.16082195, Time: 0.0402 Steps: 5670, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001422, Sample Num: 22752, Cur Loss: 1.52552688, Cur Avg Loss: 1.68607194, Log Avg loss: 1.85733131, Global Avg Loss: 3.15852707, Time: 0.0402 Steps: 5680, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001432, Sample Num: 22912, Cur Loss: 0.79425144, Cur Avg Loss: 1.68496778, Log Avg loss: 1.52795655, Global Avg Loss: 3.15566139, Time: 0.0403 Steps: 5690, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001442, Sample Num: 23072, Cur Loss: 1.32916880, Cur Avg Loss: 1.68843716, Log Avg loss: 2.18525120, Global Avg Loss: 3.15395892, Time: 0.0403 Steps: 5700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001452, Sample Num: 23232, Cur Loss: 0.47099018, Cur Avg Loss: 1.68765580, Log Avg loss: 1.57498437, Global Avg Loss: 3.15119364, Time: 0.0403 Steps: 5710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001462, Sample Num: 23392, Cur Loss: 2.94720984, Cur Avg Loss: 1.68638047, Log Avg loss: 1.50120280, Global Avg Loss: 3.14830904, Time: 0.0402 Steps: 5720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001472, Sample Num: 23552, Cur Loss: 1.86420417, Cur Avg Loss: 1.68746478, Log Avg loss: 1.84599112, Global Avg Loss: 3.14603624, Time: 0.0403 Steps: 5730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001482, Sample Num: 23712, Cur Loss: 1.98136628, Cur Avg Loss: 1.68594763, Log Avg loss: 1.46262258, Global Avg Loss: 3.14310346, Time: 0.0402 Steps: 5740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001492, Sample Num: 23872, Cur Loss: 2.77646875, Cur Avg Loss: 1.68856899, Log Avg loss: 2.07705461, Global Avg Loss: 3.14124946, Time: 0.0404 Steps: 5750, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001502, Sample Num: 24032, Cur Loss: 1.41423905, Cur Avg Loss: 1.68997810, Log Avg loss: 1.90021700, Global Avg Loss: 3.13909489, Time: 0.0402 Steps: 5760, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001512, Sample Num: 24192, Cur Loss: 1.35549200, Cur Avg Loss: 1.69054496, Log Avg loss: 1.77568736, Global Avg Loss: 3.13673197, Time: 0.0402 Steps: 5770, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001522, Sample Num: 24352, Cur Loss: 1.72472429, Cur Avg Loss: 1.69212323, Log Avg loss: 1.93075762, Global Avg Loss: 3.13464551, Time: 0.0403 Steps: 5780, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001532, Sample Num: 24512, Cur Loss: 2.01159787, Cur Avg Loss: 1.69099657, Log Avg loss: 1.51951840, Global Avg Loss: 3.13185599, Time: 0.0402 Steps: 5790, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001542, Sample Num: 24672, Cur Loss: 1.16719556, Cur Avg Loss: 1.69102076, Log Avg loss: 1.69472799, Global Avg Loss: 3.12937819, Time: 0.0403 Steps: 5800, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001552, Sample Num: 24832, Cur Loss: 0.91911525, Cur Avg Loss: 1.68917034, Log Avg loss: 1.40383468, Global Avg Loss: 3.12640823, Time: 0.0403 Steps: 5810, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001562, Sample Num: 24992, Cur Loss: 1.40275240, Cur Avg Loss: 1.68964457, Log Avg loss: 1.76324537, Global Avg Loss: 3.12406603, Time: 0.0402 Steps: 5820, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001572, Sample Num: 25152, Cur Loss: 0.70315719, Cur Avg Loss: 1.68788292, Log Avg loss: 1.41271287, Global Avg Loss: 3.12113060, Time: 0.0402 Steps: 5830, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001582, Sample Num: 25312, Cur Loss: 0.34794480, Cur Avg Loss: 1.68872073, Log Avg loss: 1.82042434, Global Avg Loss: 3.11890337, Time: 0.0402 Steps: 5840, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001592, Sample Num: 25472, Cur Loss: 1.86130881, Cur Avg Loss: 1.68955480, Log Avg loss: 1.82150444, Global Avg Loss: 3.11668559, Time: 0.0402 Steps: 5850, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001602, Sample Num: 25632, Cur Loss: 1.50807035, Cur Avg Loss: 1.69064576, Log Avg loss: 1.86432784, Global Avg Loss: 3.11454846, Time: 0.0402 Steps: 5860, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001612, Sample Num: 25792, Cur Loss: 2.82005405, Cur Avg Loss: 1.68990207, Log Avg loss: 1.57076312, Global Avg Loss: 3.11191850, Time: 0.0402 Steps: 5870, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001622, Sample Num: 25952, Cur Loss: 1.01926088, Cur Avg Loss: 1.69085248, Log Avg loss: 1.84405773, Global Avg Loss: 3.10976228, Time: 0.0404 Steps: 5880, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001632, Sample Num: 26112, Cur Loss: 1.09509802, Cur Avg Loss: 1.68874129, Log Avg loss: 1.34630617, Global Avg Loss: 3.10676829, Time: 0.0405 Steps: 5890, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001642, Sample Num: 26272, Cur Loss: 1.46798933, Cur Avg Loss: 1.68826932, Log Avg loss: 1.61124401, Global Avg Loss: 3.10423351, Time: 0.0428 Steps: 5900, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001652, Sample Num: 26432, Cur Loss: 1.29721391, Cur Avg Loss: 1.69031179, Log Avg loss: 2.02568606, Global Avg Loss: 3.10240855, Time: 0.0403 Steps: 5910, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001662, Sample Num: 26592, Cur Loss: 1.51393938, Cur Avg Loss: 1.69182363, Log Avg loss: 1.94157910, Global Avg Loss: 3.10044769, Time: 0.0467 Steps: 5920, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001672, Sample Num: 26752, Cur Loss: 1.50418568, Cur Avg Loss: 1.69216644, Log Avg loss: 1.74914101, Global Avg Loss: 3.09816893, Time: 0.0863 Steps: 5930, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001682, Sample Num: 26912, Cur Loss: 1.13115251, Cur Avg Loss: 1.69234221, Log Avg loss: 1.72173204, Global Avg Loss: 3.09585170, Time: 0.0752 Steps: 5940, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001692, Sample Num: 27072, Cur Loss: 1.40906858, Cur Avg Loss: 1.69269577, Log Avg loss: 1.75216455, Global Avg Loss: 3.09359340, Time: 0.0970 Steps: 5950, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001702, Sample Num: 27232, Cur Loss: 1.59445024, Cur Avg Loss: 1.69335200, Log Avg loss: 1.80438570, Global Avg Loss: 3.09143030, Time: 0.0403 Steps: 5960, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001712, Sample Num: 27392, Cur Loss: 1.98243237, Cur Avg Loss: 1.69071690, Log Avg loss: 1.24222250, Global Avg Loss: 3.08833280, Time: 0.0402 Steps: 5970, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001722, Sample Num: 27552, Cur Loss: 1.13996089, Cur Avg Loss: 1.69234086, Log Avg loss: 1.97036307, Global Avg Loss: 3.08646328, Time: 0.0403 Steps: 5980, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001732, Sample Num: 27712, Cur Loss: 2.89667511, Cur Avg Loss: 1.69232455, Log Avg loss: 1.68951505, Global Avg Loss: 3.08413115, Time: 0.0402 Steps: 5990, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001742, Sample Num: 27872, Cur Loss: 1.56140327, Cur Avg Loss: 1.69477720, Log Avg loss: 2.11957723, Global Avg Loss: 3.08252356, Time: 0.0402 Steps: 6000, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001752, Sample Num: 28032, Cur Loss: 2.28005028, Cur Avg Loss: 1.69352968, Log Avg loss: 1.47621191, Global Avg Loss: 3.07985083, Time: 0.0402 Steps: 6010, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001762, Sample Num: 28192, Cur Loss: 1.83346736, Cur Avg Loss: 1.69457768, Log Avg loss: 1.87818655, Global Avg Loss: 3.07785471, Time: 0.0402 Steps: 6020, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001772, Sample Num: 28352, Cur Loss: 2.31342602, Cur Avg Loss: 1.69488098, Log Avg loss: 1.74832227, Global Avg Loss: 3.07564984, Time: 0.0402 Steps: 6030, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001782, Sample Num: 28512, Cur Loss: 1.06789291, Cur Avg Loss: 1.69474252, Log Avg loss: 1.67020704, Global Avg Loss: 3.07332295, Time: 0.0402 Steps: 6040, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001792, Sample Num: 28672, Cur Loss: 1.21383059, Cur Avg Loss: 1.69329906, Log Avg loss: 1.43607596, Global Avg Loss: 3.07061676, Time: 0.0403 Steps: 6050, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001802, Sample Num: 28832, Cur Loss: 1.58142030, Cur Avg Loss: 1.69380967, Log Avg loss: 1.78531103, Global Avg Loss: 3.06849579, Time: 0.0403 Steps: 6060, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001812, Sample Num: 28992, Cur Loss: 1.21839070, Cur Avg Loss: 1.69441387, Log Avg loss: 1.80329044, Global Avg Loss: 3.06641143, Time: 0.0403 Steps: 6070, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001822, Sample Num: 29152, Cur Loss: 1.48033762, Cur Avg Loss: 1.69507208, Log Avg loss: 1.81433929, Global Avg Loss: 3.06435211, Time: 0.0402 Steps: 6080, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001832, Sample Num: 29312, Cur Loss: 0.99968261, Cur Avg Loss: 1.69424186, Log Avg loss: 1.54297625, Global Avg Loss: 3.06185395, Time: 0.0402 Steps: 6090, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001842, Sample Num: 29472, Cur Loss: 2.27442408, Cur Avg Loss: 1.69311466, Log Avg loss: 1.48661104, Global Avg Loss: 3.05927159, Time: 0.0402 Steps: 6100, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001852, Sample Num: 29632, Cur Loss: 1.56979322, Cur Avg Loss: 1.69593153, Log Avg loss: 2.21479800, Global Avg Loss: 3.05788947, Time: 0.0402 Steps: 6110, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001862, Sample Num: 29792, Cur Loss: 2.36990118, Cur Avg Loss: 1.69458645, Log Avg loss: 1.44547802, Global Avg Loss: 3.05525481, Time: 0.0401 Steps: 6120, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001872, Sample Num: 29952, Cur Loss: 0.62174159, Cur Avg Loss: 1.69305666, Log Avg loss: 1.40821019, Global Avg Loss: 3.05256795, Time: 0.0402 Steps: 6130, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001882, Sample Num: 30112, Cur Loss: 3.47088909, Cur Avg Loss: 1.69507749, Log Avg loss: 2.07337773, Global Avg Loss: 3.05097318, Time: 0.0402 Steps: 6140, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001892, Sample Num: 30272, Cur Loss: 2.71136093, Cur Avg Loss: 1.69618827, Log Avg loss: 1.90523552, Global Avg Loss: 3.04911019, Time: 0.0402 Steps: 6150, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001902, Sample Num: 30432, Cur Loss: 0.99988663, Cur Avg Loss: 1.69667683, Log Avg loss: 1.78911281, Global Avg Loss: 3.04706474, Time: 0.0402 Steps: 6160, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001912, Sample Num: 30592, Cur Loss: 2.49092770, Cur Avg Loss: 1.69784585, Log Avg loss: 1.92019407, Global Avg Loss: 3.04523837, Time: 0.0402 Steps: 6170, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001922, Sample Num: 30752, Cur Loss: 2.03019571, Cur Avg Loss: 1.69716219, Log Avg loss: 1.56644598, Global Avg Loss: 3.04284550, Time: 0.0402 Steps: 6180, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001932, Sample Num: 30912, Cur Loss: 1.89416635, Cur Avg Loss: 1.69671050, Log Avg loss: 1.60989538, Global Avg Loss: 3.04053056, Time: 0.0402 Steps: 6190, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001942, Sample Num: 31072, Cur Loss: 1.48274112, Cur Avg Loss: 1.69771615, Log Avg loss: 1.89200768, Global Avg Loss: 3.03867810, Time: 0.0402 Steps: 6200, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001952, Sample Num: 31232, Cur Loss: 1.66405129, Cur Avg Loss: 1.69685548, Log Avg loss: 1.52971334, Global Avg Loss: 3.03624821, Time: 0.0438 Steps: 6210, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001962, Sample Num: 31392, Cur Loss: 2.85446048, Cur Avg Loss: 1.69707242, Log Avg loss: 1.73941942, Global Avg Loss: 3.03416327, Time: 0.0552 Steps: 6220, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001972, Sample Num: 31552, Cur Loss: 0.80417824, Cur Avg Loss: 1.69625721, Log Avg loss: 1.53631309, Global Avg Loss: 3.03175902, Time: 0.0552 Steps: 6230, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001982, Sample Num: 31712, Cur Loss: 2.11514711, Cur Avg Loss: 1.69576199, Log Avg loss: 1.59810502, Global Avg Loss: 3.02946150, Time: 0.0515 Steps: 6240, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001992, Sample Num: 31872, Cur Loss: 1.81546628, Cur Avg Loss: 1.69585956, Log Avg loss: 1.71519769, Global Avg Loss: 3.02735867, Time: 0.0655 Steps: 6250, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002002, Sample Num: 32032, Cur Loss: 2.19649935, Cur Avg Loss: 1.69617971, Log Avg loss: 1.75995417, Global Avg Loss: 3.02533407, Time: 0.0927 Steps: 6260, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002012, Sample Num: 32192, Cur Loss: 0.68596202, Cur Avg Loss: 1.69618262, Log Avg loss: 1.69676381, Global Avg Loss: 3.02321513, Time: 0.0522 Steps: 6270, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002022, Sample Num: 32352, Cur Loss: 1.57947135, Cur Avg Loss: 1.69468718, Log Avg loss: 1.39380544, Global Avg Loss: 3.02062053, Time: 0.0504 Steps: 6280, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002032, Sample Num: 32512, Cur Loss: 2.18519402, Cur Avg Loss: 1.69656151, Log Avg loss: 2.07555088, Global Avg Loss: 3.01911804, Time: 0.0402 Steps: 6290, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002042, Sample Num: 32672, Cur Loss: 1.06531572, Cur Avg Loss: 1.69750868, Log Avg loss: 1.88997403, Global Avg Loss: 3.01732575, Time: 0.0402 Steps: 6300, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002052, Sample Num: 32832, Cur Loss: 1.08738184, Cur Avg Loss: 1.69789412, Log Avg loss: 1.77660158, Global Avg Loss: 3.01535946, Time: 0.0402 Steps: 6310, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002062, Sample Num: 32992, Cur Loss: 1.89465249, Cur Avg Loss: 1.69797754, Log Avg loss: 1.71509352, Global Avg Loss: 3.01330208, Time: 0.0403 Steps: 6320, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002072, Sample Num: 33152, Cur Loss: 1.17290533, Cur Avg Loss: 1.69825583, Log Avg loss: 1.75563998, Global Avg Loss: 3.01131525, Time: 0.0402 Steps: 6330, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002082, Sample Num: 33312, Cur Loss: 1.06504226, Cur Avg Loss: 1.69853268, Log Avg loss: 1.75589527, Global Avg Loss: 3.00933509, Time: 0.0402 Steps: 6340, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002092, Sample Num: 33472, Cur Loss: 2.42835093, Cur Avg Loss: 1.69707746, Log Avg loss: 1.39410056, Global Avg Loss: 3.00679142, Time: 0.0402 Steps: 6350, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002102, Sample Num: 33632, Cur Loss: 1.33906865, Cur Avg Loss: 1.69598805, Log Avg loss: 1.46808548, Global Avg Loss: 3.00437207, Time: 0.0402 Steps: 6360, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002112, Sample Num: 33792, Cur Loss: 1.92622352, Cur Avg Loss: 1.69762017, Log Avg loss: 2.04069058, Global Avg Loss: 3.00285923, Time: 0.0402 Steps: 6370, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002122, Sample Num: 33952, Cur Loss: 1.79444277, Cur Avg Loss: 1.69771405, Log Avg loss: 1.71754217, Global Avg Loss: 3.00084462, Time: 0.0402 Steps: 6380, Updated lr: 0.000095 ***** Running evaluation checkpoint-6387 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-6387 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 94.806626, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.556019, "eval_total_loss": 1093.881481, "eval_mae": 0.990496, "eval_mse": 1.556267, "eval_r2": 0.010734, "eval_sp_statistic": 0.134851, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.176643, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.535567, "test_total_loss": 770.854652, "test_mae": 1.021768, "test_mse": 1.535767, "test_r2": 0.008803, "test_sp_statistic": 0.13294, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.151448, "test_ps_pvalue": 0.0, "lr": 9.489141773352301e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.9990199859365902, "train_cur_epoch_loss": 3611.9011826366186, "train_cur_epoch_avg_loss": 1.6965247452497034, "train_cur_epoch_time": 94.8066258430481, "train_cur_epoch_avg_time": 0.044531059578698025, "epoch": 3, "step": 6387} ################################################## Training, Epoch: 0004, Batch: 000003, Sample Num: 48, Cur Loss: 1.38678479, Cur Avg Loss: 2.59344467, Log Avg loss: 1.71322982, Global Avg Loss: 2.99882957, Time: 0.0402 Steps: 6390, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000013, Sample Num: 208, Cur Loss: 1.12808561, Cur Avg Loss: 2.00695335, Log Avg loss: 1.83100595, Global Avg Loss: 2.99700485, Time: 0.0403 Steps: 6400, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000023, Sample Num: 368, Cur Loss: 0.57128406, Cur Avg Loss: 1.72195074, Log Avg loss: 1.35144736, Global Avg Loss: 2.99443768, Time: 0.0403 Steps: 6410, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000033, Sample Num: 528, Cur Loss: 2.44075441, Cur Avg Loss: 1.77719204, Log Avg loss: 1.90424701, Global Avg Loss: 2.99273956, Time: 0.0402 Steps: 6420, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000043, Sample Num: 688, Cur Loss: 1.13467491, Cur Avg Loss: 1.73644477, Log Avg loss: 1.60197881, Global Avg Loss: 2.99057664, Time: 0.0402 Steps: 6430, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000053, Sample Num: 848, Cur Loss: 2.57135129, Cur Avg Loss: 1.73861050, Log Avg loss: 1.74792312, Global Avg Loss: 2.98864705, Time: 0.0402 Steps: 6440, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000063, Sample Num: 1008, Cur Loss: 2.02712750, Cur Avg Loss: 1.73994546, Log Avg loss: 1.74702076, Global Avg Loss: 2.98672205, Time: 0.0402 Steps: 6450, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000073, Sample Num: 1168, Cur Loss: 2.73102617, Cur Avg Loss: 1.71220679, Log Avg loss: 1.53745313, Global Avg Loss: 2.98447860, Time: 0.0402 Steps: 6460, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000083, Sample Num: 1328, Cur Loss: 2.41003847, Cur Avg Loss: 1.68501193, Log Avg loss: 1.48648945, Global Avg Loss: 2.98216331, Time: 0.0403 Steps: 6470, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000093, Sample Num: 1488, Cur Loss: 1.84948587, Cur Avg Loss: 1.67123493, Log Avg loss: 1.55688587, Global Avg Loss: 2.97996381, Time: 0.0402 Steps: 6480, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000103, Sample Num: 1648, Cur Loss: 1.70486081, Cur Avg Loss: 1.65299492, Log Avg loss: 1.48336284, Global Avg Loss: 2.97765780, Time: 0.0402 Steps: 6490, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000113, Sample Num: 1808, Cur Loss: 2.19560599, Cur Avg Loss: 1.67672815, Log Avg loss: 1.92118040, Global Avg Loss: 2.97603245, Time: 0.0402 Steps: 6500, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000123, Sample Num: 1968, Cur Loss: 1.69251847, Cur Avg Loss: 1.67621818, Log Avg loss: 1.67045556, Global Avg Loss: 2.97402696, Time: 0.0402 Steps: 6510, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000133, Sample Num: 2128, Cur Loss: 1.50124907, Cur Avg Loss: 1.67329992, Log Avg loss: 1.63740523, Global Avg Loss: 2.97197692, Time: 0.0402 Steps: 6520, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000143, Sample Num: 2288, Cur Loss: 0.57563859, Cur Avg Loss: 1.69285598, Log Avg loss: 1.95295166, Global Avg Loss: 2.97041639, Time: 0.0402 Steps: 6530, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000153, Sample Num: 2448, Cur Loss: 0.84232366, Cur Avg Loss: 1.67156731, Log Avg loss: 1.36713926, Global Avg Loss: 2.96796490, Time: 0.0402 Steps: 6540, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000163, Sample Num: 2608, Cur Loss: 1.66502798, Cur Avg Loss: 1.65257596, Log Avg loss: 1.36200833, Global Avg Loss: 2.96551306, Time: 0.0402 Steps: 6550, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000173, Sample Num: 2768, Cur Loss: 1.06634641, Cur Avg Loss: 1.64639430, Log Avg loss: 1.54563321, Global Avg Loss: 2.96334861, Time: 0.0402 Steps: 6560, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000183, Sample Num: 2928, Cur Loss: 1.67102098, Cur Avg Loss: 1.65262886, Log Avg loss: 1.76048675, Global Avg Loss: 2.96151777, Time: 0.0403 Steps: 6570, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000193, Sample Num: 3088, Cur Loss: 0.94102383, Cur Avg Loss: 1.64675773, Log Avg loss: 1.53931613, Global Avg Loss: 2.95935637, Time: 0.0402 Steps: 6580, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000203, Sample Num: 3248, Cur Loss: 1.79425716, Cur Avg Loss: 1.64451460, Log Avg loss: 1.60122212, Global Avg Loss: 2.95729546, Time: 0.0402 Steps: 6590, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000213, Sample Num: 3408, Cur Loss: 0.97748923, Cur Avg Loss: 1.62390157, Log Avg loss: 1.20545711, Global Avg Loss: 2.95464116, Time: 0.0402 Steps: 6600, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000223, Sample Num: 3568, Cur Loss: 0.88400936, Cur Avg Loss: 1.62916011, Log Avg loss: 1.74116700, Global Avg Loss: 2.95280535, Time: 0.0403 Steps: 6610, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000233, Sample Num: 3728, Cur Loss: 1.22484505, Cur Avg Loss: 1.62669866, Log Avg loss: 1.57180843, Global Avg Loss: 2.95071925, Time: 0.0662 Steps: 6620, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000243, Sample Num: 3888, Cur Loss: 0.91228187, Cur Avg Loss: 1.63283825, Log Avg loss: 1.77589060, Global Avg Loss: 2.94894726, Time: 0.0685 Steps: 6630, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000253, Sample Num: 4048, Cur Loss: 1.55957794, Cur Avg Loss: 1.65020260, Log Avg loss: 2.07215624, Global Avg Loss: 2.94762679, Time: 0.0404 Steps: 6640, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000263, Sample Num: 4208, Cur Loss: 2.74650145, Cur Avg Loss: 1.65716015, Log Avg loss: 1.83318632, Global Avg Loss: 2.94595094, Time: 0.0441 Steps: 6650, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000273, Sample Num: 4368, Cur Loss: 2.08340168, Cur Avg Loss: 1.66394976, Log Avg loss: 1.84251649, Global Avg Loss: 2.94429413, Time: 0.0610 Steps: 6660, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000283, Sample Num: 4528, Cur Loss: 1.87083089, Cur Avg Loss: 1.66839388, Log Avg loss: 1.78971825, Global Avg Loss: 2.94256314, Time: 0.0406 Steps: 6670, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000293, Sample Num: 4688, Cur Loss: 1.45307088, Cur Avg Loss: 1.66748696, Log Avg loss: 1.64182104, Global Avg Loss: 2.94061592, Time: 0.0611 Steps: 6680, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000303, Sample Num: 4848, Cur Loss: 1.08971000, Cur Avg Loss: 1.66758865, Log Avg loss: 1.67056825, Global Avg Loss: 2.93871749, Time: 0.0533 Steps: 6690, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000313, Sample Num: 5008, Cur Loss: 1.37486076, Cur Avg Loss: 1.67719176, Log Avg loss: 1.96816608, Global Avg Loss: 2.93726891, Time: 0.0403 Steps: 6700, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000323, Sample Num: 5168, Cur Loss: 1.71490645, Cur Avg Loss: 1.68082641, Log Avg loss: 1.79459102, Global Avg Loss: 2.93556596, Time: 0.0403 Steps: 6710, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000333, Sample Num: 5328, Cur Loss: 1.20963383, Cur Avg Loss: 1.68263463, Log Avg loss: 1.74103999, Global Avg Loss: 2.93378839, Time: 0.0403 Steps: 6720, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000343, Sample Num: 5488, Cur Loss: 2.27716613, Cur Avg Loss: 1.68238789, Log Avg loss: 1.67417140, Global Avg Loss: 2.93191675, Time: 0.0402 Steps: 6730, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000353, Sample Num: 5648, Cur Loss: 0.39106506, Cur Avg Loss: 1.67096428, Log Avg loss: 1.27913460, Global Avg Loss: 2.92946455, Time: 0.0402 Steps: 6740, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000363, Sample Num: 5808, Cur Loss: 1.50361943, Cur Avg Loss: 1.66613859, Log Avg loss: 1.49579162, Global Avg Loss: 2.92734059, Time: 0.0402 Steps: 6750, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000373, Sample Num: 5968, Cur Loss: 1.63382852, Cur Avg Loss: 1.67079044, Log Avg loss: 1.83965259, Global Avg Loss: 2.92573158, Time: 0.0404 Steps: 6760, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000383, Sample Num: 6128, Cur Loss: 1.26615143, Cur Avg Loss: 1.66729343, Log Avg loss: 1.53685504, Global Avg Loss: 2.92368006, Time: 0.0403 Steps: 6770, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000393, Sample Num: 6288, Cur Loss: 1.55868900, Cur Avg Loss: 1.66346824, Log Avg loss: 1.51696342, Global Avg Loss: 2.92160526, Time: 0.0403 Steps: 6780, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000403, Sample Num: 6448, Cur Loss: 1.65354311, Cur Avg Loss: 1.66382215, Log Avg loss: 1.67773079, Global Avg Loss: 2.91977334, Time: 0.0402 Steps: 6790, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000413, Sample Num: 6608, Cur Loss: 1.86575258, Cur Avg Loss: 1.66266045, Log Avg loss: 1.61584406, Global Avg Loss: 2.91785580, Time: 0.0403 Steps: 6800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000423, Sample Num: 6768, Cur Loss: 1.89432478, Cur Avg Loss: 1.66910386, Log Avg loss: 1.93521661, Global Avg Loss: 2.91641286, Time: 0.0403 Steps: 6810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000433, Sample Num: 6928, Cur Loss: 1.79291534, Cur Avg Loss: 1.67201251, Log Avg loss: 1.79504830, Global Avg Loss: 2.91476863, Time: 0.0402 Steps: 6820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000443, Sample Num: 7088, Cur Loss: 2.10812163, Cur Avg Loss: 1.67343223, Log Avg loss: 1.73490599, Global Avg Loss: 2.91304116, Time: 0.0402 Steps: 6830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000453, Sample Num: 7248, Cur Loss: 1.72045779, Cur Avg Loss: 1.67389160, Log Avg loss: 1.69424211, Global Avg Loss: 2.91125929, Time: 0.0403 Steps: 6840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000463, Sample Num: 7408, Cur Loss: 1.60839033, Cur Avg Loss: 1.66737077, Log Avg loss: 1.37197685, Global Avg Loss: 2.90901216, Time: 0.0403 Steps: 6850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000473, Sample Num: 7568, Cur Loss: 0.86728245, Cur Avg Loss: 1.66328262, Log Avg loss: 1.47400118, Global Avg Loss: 2.90692031, Time: 0.0402 Steps: 6860, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000483, Sample Num: 7728, Cur Loss: 1.64258361, Cur Avg Loss: 1.66455282, Log Avg loss: 1.72463357, Global Avg Loss: 2.90519937, Time: 0.0403 Steps: 6870, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000493, Sample Num: 7888, Cur Loss: 0.96879661, Cur Avg Loss: 1.65797711, Log Avg loss: 1.34037014, Global Avg Loss: 2.90292491, Time: 0.0402 Steps: 6880, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000503, Sample Num: 8048, Cur Loss: 1.31985188, Cur Avg Loss: 1.65276365, Log Avg loss: 1.39573992, Global Avg Loss: 2.90073741, Time: 0.0402 Steps: 6890, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000513, Sample Num: 8208, Cur Loss: 2.12727785, Cur Avg Loss: 1.65502854, Log Avg loss: 1.76895287, Global Avg Loss: 2.89909714, Time: 0.0402 Steps: 6900, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000523, Sample Num: 8368, Cur Loss: 1.54451275, Cur Avg Loss: 1.65645410, Log Avg loss: 1.72958528, Global Avg Loss: 2.89740465, Time: 0.0402 Steps: 6910, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000533, Sample Num: 8528, Cur Loss: 1.92783046, Cur Avg Loss: 1.65834256, Log Avg loss: 1.75710906, Global Avg Loss: 2.89575683, Time: 0.0402 Steps: 6920, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000543, Sample Num: 8688, Cur Loss: 1.23124456, Cur Avg Loss: 1.65210654, Log Avg loss: 1.31972665, Global Avg Loss: 2.89348261, Time: 0.0402 Steps: 6930, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000553, Sample Num: 8848, Cur Loss: 2.84980178, Cur Avg Loss: 1.65458399, Log Avg loss: 1.78910959, Global Avg Loss: 2.89189130, Time: 0.0402 Steps: 6940, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000563, Sample Num: 9008, Cur Loss: 1.21587956, Cur Avg Loss: 1.64938340, Log Avg loss: 1.36179036, Global Avg Loss: 2.88968971, Time: 0.0402 Steps: 6950, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000573, Sample Num: 9168, Cur Loss: 1.09827745, Cur Avg Loss: 1.64814038, Log Avg loss: 1.57815883, Global Avg Loss: 2.88780533, Time: 0.0402 Steps: 6960, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000583, Sample Num: 9328, Cur Loss: 1.39103103, Cur Avg Loss: 1.64644725, Log Avg loss: 1.54943048, Global Avg Loss: 2.88588514, Time: 0.0678 Steps: 6970, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000593, Sample Num: 9488, Cur Loss: 0.96264046, Cur Avg Loss: 1.64409995, Log Avg loss: 1.50725265, Global Avg Loss: 2.88391002, Time: 0.0428 Steps: 6980, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000603, Sample Num: 9648, Cur Loss: 0.89386237, Cur Avg Loss: 1.63725897, Log Avg loss: 1.23158860, Global Avg Loss: 2.88154618, Time: 0.0429 Steps: 6990, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000613, Sample Num: 9808, Cur Loss: 1.97494054, Cur Avg Loss: 1.64294806, Log Avg loss: 1.98600044, Global Avg Loss: 2.88026683, Time: 0.0634 Steps: 7000, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000623, Sample Num: 9968, Cur Loss: 1.63423431, Cur Avg Loss: 1.64051704, Log Avg loss: 1.49149565, Global Avg Loss: 2.87828570, Time: 0.0561 Steps: 7010, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000633, Sample Num: 10128, Cur Loss: 1.29240477, Cur Avg Loss: 1.64408163, Log Avg loss: 1.86615553, Global Avg Loss: 2.87684392, Time: 0.0493 Steps: 7020, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000643, Sample Num: 10288, Cur Loss: 1.57554829, Cur Avg Loss: 1.64127636, Log Avg loss: 1.46370265, Global Avg Loss: 2.87483376, Time: 0.0553 Steps: 7030, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000653, Sample Num: 10448, Cur Loss: 2.60202169, Cur Avg Loss: 1.63797533, Log Avg loss: 1.42571936, Global Avg Loss: 2.87277536, Time: 0.0685 Steps: 7040, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000663, Sample Num: 10608, Cur Loss: 2.64387965, Cur Avg Loss: 1.63509869, Log Avg loss: 1.44725404, Global Avg Loss: 2.87075335, Time: 0.0404 Steps: 7050, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000673, Sample Num: 10768, Cur Loss: 1.35434008, Cur Avg Loss: 1.62950721, Log Avg loss: 1.25879201, Global Avg Loss: 2.86847011, Time: 0.0403 Steps: 7060, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000683, Sample Num: 10928, Cur Loss: 1.65496409, Cur Avg Loss: 1.63102187, Log Avg loss: 1.73295866, Global Avg Loss: 2.86686402, Time: 0.0402 Steps: 7070, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000693, Sample Num: 11088, Cur Loss: 1.22041595, Cur Avg Loss: 1.62444555, Log Avg loss: 1.17528251, Global Avg Loss: 2.86447478, Time: 0.0404 Steps: 7080, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000703, Sample Num: 11248, Cur Loss: 1.02113068, Cur Avg Loss: 1.62144641, Log Avg loss: 1.41360581, Global Avg Loss: 2.86242842, Time: 0.0402 Steps: 7090, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000713, Sample Num: 11408, Cur Loss: 1.29492283, Cur Avg Loss: 1.62460600, Log Avg loss: 1.84672546, Global Avg Loss: 2.86099785, Time: 0.0403 Steps: 7100, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000723, Sample Num: 11568, Cur Loss: 1.20051289, Cur Avg Loss: 1.62261961, Log Avg loss: 1.48098997, Global Avg Loss: 2.85905691, Time: 0.0402 Steps: 7110, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000733, Sample Num: 11728, Cur Loss: 0.99059278, Cur Avg Loss: 1.61708053, Log Avg loss: 1.21660488, Global Avg Loss: 2.85675010, Time: 0.0402 Steps: 7120, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000743, Sample Num: 11888, Cur Loss: 2.49759722, Cur Avg Loss: 1.61843327, Log Avg loss: 1.71758913, Global Avg Loss: 2.85515239, Time: 0.0402 Steps: 7130, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000753, Sample Num: 12048, Cur Loss: 2.31966639, Cur Avg Loss: 1.61974117, Log Avg loss: 1.71691830, Global Avg Loss: 2.85355823, Time: 0.0402 Steps: 7140, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000763, Sample Num: 12208, Cur Loss: 1.51431990, Cur Avg Loss: 1.61574008, Log Avg loss: 1.31445829, Global Avg Loss: 2.85140564, Time: 0.0402 Steps: 7150, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000773, Sample Num: 12368, Cur Loss: 1.99128747, Cur Avg Loss: 1.61913794, Log Avg loss: 1.87839438, Global Avg Loss: 2.85004669, Time: 0.0402 Steps: 7160, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000783, Sample Num: 12528, Cur Loss: 1.71120393, Cur Avg Loss: 1.61759629, Log Avg loss: 1.49842702, Global Avg Loss: 2.84816158, Time: 0.0402 Steps: 7170, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000793, Sample Num: 12688, Cur Loss: 1.56161141, Cur Avg Loss: 1.62096339, Log Avg loss: 1.88460703, Global Avg Loss: 2.84681958, Time: 0.0402 Steps: 7180, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000803, Sample Num: 12848, Cur Loss: 0.69180512, Cur Avg Loss: 1.62015933, Log Avg loss: 1.55639748, Global Avg Loss: 2.84502484, Time: 0.0402 Steps: 7190, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000813, Sample Num: 13008, Cur Loss: 1.39970779, Cur Avg Loss: 1.62302941, Log Avg loss: 1.85349704, Global Avg Loss: 2.84364772, Time: 0.0402 Steps: 7200, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000823, Sample Num: 13168, Cur Loss: 1.17065001, Cur Avg Loss: 1.62024445, Log Avg loss: 1.39382729, Global Avg Loss: 2.84163687, Time: 0.0402 Steps: 7210, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000833, Sample Num: 13328, Cur Loss: 2.53562522, Cur Avg Loss: 1.61845899, Log Avg loss: 1.47151558, Global Avg Loss: 2.83973920, Time: 0.0402 Steps: 7220, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000843, Sample Num: 13488, Cur Loss: 1.38164973, Cur Avg Loss: 1.62186841, Log Avg loss: 1.90587237, Global Avg Loss: 2.83844754, Time: 0.0402 Steps: 7230, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000853, Sample Num: 13648, Cur Loss: 2.59780359, Cur Avg Loss: 1.62375277, Log Avg loss: 1.78260436, Global Avg Loss: 2.83698919, Time: 0.0402 Steps: 7240, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000863, Sample Num: 13808, Cur Loss: 1.84603000, Cur Avg Loss: 1.62460828, Log Avg loss: 1.69758331, Global Avg Loss: 2.83541760, Time: 0.0402 Steps: 7250, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000873, Sample Num: 13968, Cur Loss: 1.48406410, Cur Avg Loss: 1.62689182, Log Avg loss: 1.82396200, Global Avg Loss: 2.83402441, Time: 0.0402 Steps: 7260, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000883, Sample Num: 14128, Cur Loss: 0.90748560, Cur Avg Loss: 1.62725761, Log Avg loss: 1.65919046, Global Avg Loss: 2.83240841, Time: 0.0402 Steps: 7270, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000893, Sample Num: 14288, Cur Loss: 2.26288486, Cur Avg Loss: 1.62938670, Log Avg loss: 1.81738561, Global Avg Loss: 2.83101414, Time: 0.0402 Steps: 7280, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000903, Sample Num: 14448, Cur Loss: 1.79309058, Cur Avg Loss: 1.63093365, Log Avg loss: 1.76907626, Global Avg Loss: 2.82955744, Time: 0.0402 Steps: 7290, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000913, Sample Num: 14608, Cur Loss: 1.60199451, Cur Avg Loss: 1.63268469, Log Avg loss: 1.79080361, Global Avg Loss: 2.82813449, Time: 0.0401 Steps: 7300, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000923, Sample Num: 14768, Cur Loss: 2.08352327, Cur Avg Loss: 1.63723837, Log Avg loss: 2.05298904, Global Avg Loss: 2.82707410, Time: 0.0402 Steps: 7310, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000933, Sample Num: 14928, Cur Loss: 2.16213608, Cur Avg Loss: 1.63550717, Log Avg loss: 1.47571786, Global Avg Loss: 2.82522798, Time: 0.0808 Steps: 7320, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000943, Sample Num: 15088, Cur Loss: 1.65720439, Cur Avg Loss: 1.63878711, Log Avg loss: 1.94480489, Global Avg Loss: 2.82402686, Time: 0.0639 Steps: 7330, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000953, Sample Num: 15248, Cur Loss: 1.75852132, Cur Avg Loss: 1.63851951, Log Avg loss: 1.61328487, Global Avg Loss: 2.82237735, Time: 0.0429 Steps: 7340, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000963, Sample Num: 15408, Cur Loss: 1.35454988, Cur Avg Loss: 1.64022163, Log Avg loss: 1.80243363, Global Avg Loss: 2.82098967, Time: 0.0779 Steps: 7350, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000973, Sample Num: 15568, Cur Loss: 1.27154660, Cur Avg Loss: 1.63922570, Log Avg loss: 1.54331839, Global Avg Loss: 2.81925370, Time: 0.0406 Steps: 7360, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000983, Sample Num: 15728, Cur Loss: 1.90261745, Cur Avg Loss: 1.63994216, Log Avg loss: 1.70965295, Global Avg Loss: 2.81774814, Time: 0.0773 Steps: 7370, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000993, Sample Num: 15888, Cur Loss: 1.12896252, Cur Avg Loss: 1.63977695, Log Avg loss: 1.62353702, Global Avg Loss: 2.81612997, Time: 0.0854 Steps: 7380, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001003, Sample Num: 16048, Cur Loss: 2.62876797, Cur Avg Loss: 1.63983799, Log Avg loss: 1.64589960, Global Avg Loss: 2.81454644, Time: 0.0869 Steps: 7390, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001013, Sample Num: 16208, Cur Loss: 1.17044628, Cur Avg Loss: 1.63812735, Log Avg loss: 1.46654968, Global Avg Loss: 2.81272482, Time: 0.0809 Steps: 7400, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001023, Sample Num: 16368, Cur Loss: 1.33040392, Cur Avg Loss: 1.63904937, Log Avg loss: 1.73245062, Global Avg Loss: 2.81126696, Time: 0.0402 Steps: 7410, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001033, Sample Num: 16528, Cur Loss: 1.00597668, Cur Avg Loss: 1.63691321, Log Avg loss: 1.41838369, Global Avg Loss: 2.80938976, Time: 0.0403 Steps: 7420, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001043, Sample Num: 16688, Cur Loss: 1.83237004, Cur Avg Loss: 1.63634793, Log Avg loss: 1.57795470, Global Avg Loss: 2.80773237, Time: 0.0402 Steps: 7430, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001053, Sample Num: 16848, Cur Loss: 1.34763288, Cur Avg Loss: 1.63639803, Log Avg loss: 1.64162322, Global Avg Loss: 2.80616502, Time: 0.0402 Steps: 7440, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001063, Sample Num: 17008, Cur Loss: 1.05652475, Cur Avg Loss: 1.63413730, Log Avg loss: 1.39608266, Global Avg Loss: 2.80427230, Time: 0.0402 Steps: 7450, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001073, Sample Num: 17168, Cur Loss: 1.76455665, Cur Avg Loss: 1.63292411, Log Avg loss: 1.50396215, Global Avg Loss: 2.80252925, Time: 0.0403 Steps: 7460, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001083, Sample Num: 17328, Cur Loss: 1.37584257, Cur Avg Loss: 1.63283982, Log Avg loss: 1.62379519, Global Avg Loss: 2.80095130, Time: 0.0402 Steps: 7470, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001093, Sample Num: 17488, Cur Loss: 1.89561689, Cur Avg Loss: 1.63206532, Log Avg loss: 1.54818686, Global Avg Loss: 2.79927648, Time: 0.0402 Steps: 7480, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001103, Sample Num: 17648, Cur Loss: 2.68063807, Cur Avg Loss: 1.63273445, Log Avg loss: 1.70587088, Global Avg Loss: 2.79781666, Time: 0.0402 Steps: 7490, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001113, Sample Num: 17808, Cur Loss: 2.13822985, Cur Avg Loss: 1.63164670, Log Avg loss: 1.51166714, Global Avg Loss: 2.79610179, Time: 0.0403 Steps: 7500, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001123, Sample Num: 17968, Cur Loss: 3.05108404, Cur Avg Loss: 1.63697809, Log Avg loss: 2.23036186, Global Avg Loss: 2.79534847, Time: 0.0401 Steps: 7510, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001133, Sample Num: 18128, Cur Loss: 3.01071644, Cur Avg Loss: 1.63777125, Log Avg loss: 1.72684398, Global Avg Loss: 2.79392759, Time: 0.0402 Steps: 7520, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001143, Sample Num: 18288, Cur Loss: 1.86173010, Cur Avg Loss: 1.63582093, Log Avg loss: 1.41484947, Global Avg Loss: 2.79209615, Time: 0.0403 Steps: 7530, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001153, Sample Num: 18448, Cur Loss: 2.34453058, Cur Avg Loss: 1.63638545, Log Avg loss: 1.70090928, Global Avg Loss: 2.79064895, Time: 0.0403 Steps: 7540, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001163, Sample Num: 18608, Cur Loss: 1.74919438, Cur Avg Loss: 1.63599932, Log Avg loss: 1.59147866, Global Avg Loss: 2.78906064, Time: 0.0402 Steps: 7550, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001173, Sample Num: 18768, Cur Loss: 1.44149137, Cur Avg Loss: 1.63639309, Log Avg loss: 1.68218923, Global Avg Loss: 2.78759653, Time: 0.0403 Steps: 7560, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001183, Sample Num: 18928, Cur Loss: 1.08478582, Cur Avg Loss: 1.63714304, Log Avg loss: 1.72511249, Global Avg Loss: 2.78619298, Time: 0.0402 Steps: 7570, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001193, Sample Num: 19088, Cur Loss: 1.20263076, Cur Avg Loss: 1.64131468, Log Avg loss: 2.13481922, Global Avg Loss: 2.78533365, Time: 0.0402 Steps: 7580, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001203, Sample Num: 19248, Cur Loss: 1.48844099, Cur Avg Loss: 1.63958451, Log Avg loss: 1.43317518, Global Avg Loss: 2.78355215, Time: 0.0402 Steps: 7590, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001213, Sample Num: 19408, Cur Loss: 1.23413122, Cur Avg Loss: 1.63835883, Log Avg loss: 1.49090949, Global Avg Loss: 2.78185130, Time: 0.0402 Steps: 7600, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001223, Sample Num: 19568, Cur Loss: 2.14262438, Cur Avg Loss: 1.63555127, Log Avg loss: 1.29499411, Global Avg Loss: 2.77989748, Time: 0.0402 Steps: 7610, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001233, Sample Num: 19728, Cur Loss: 2.83182144, Cur Avg Loss: 1.63435761, Log Avg loss: 1.48837283, Global Avg Loss: 2.77820257, Time: 0.0402 Steps: 7620, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001243, Sample Num: 19888, Cur Loss: 1.72782564, Cur Avg Loss: 1.63742396, Log Avg loss: 2.01550578, Global Avg Loss: 2.77720297, Time: 0.0402 Steps: 7630, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001253, Sample Num: 20048, Cur Loss: 1.60904002, Cur Avg Loss: 1.63796598, Log Avg loss: 1.70533863, Global Avg Loss: 2.77580000, Time: 0.0402 Steps: 7640, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001263, Sample Num: 20208, Cur Loss: 0.89537811, Cur Avg Loss: 1.63926036, Log Avg loss: 1.80144643, Global Avg Loss: 2.77452634, Time: 0.0402 Steps: 7650, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001273, Sample Num: 20368, Cur Loss: 1.15810728, Cur Avg Loss: 1.64022238, Log Avg loss: 1.76172515, Global Avg Loss: 2.77320414, Time: 0.0769 Steps: 7660, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001283, Sample Num: 20528, Cur Loss: 1.23335814, Cur Avg Loss: 1.63922740, Log Avg loss: 1.51256677, Global Avg Loss: 2.77156055, Time: 0.0915 Steps: 7670, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001293, Sample Num: 20688, Cur Loss: 1.67482114, Cur Avg Loss: 1.63987887, Log Avg loss: 1.72346250, Global Avg Loss: 2.77019584, Time: 0.0481 Steps: 7680, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001303, Sample Num: 20848, Cur Loss: 0.94204187, Cur Avg Loss: 1.63700777, Log Avg loss: 1.26577378, Global Avg Loss: 2.76823950, Time: 0.0457 Steps: 7690, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001313, Sample Num: 21008, Cur Loss: 0.80205095, Cur Avg Loss: 1.63791607, Log Avg loss: 1.75626782, Global Avg Loss: 2.76692525, Time: 0.0562 Steps: 7700, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001323, Sample Num: 21168, Cur Loss: 1.83079731, Cur Avg Loss: 1.63751568, Log Avg loss: 1.58494499, Global Avg Loss: 2.76539220, Time: 0.0424 Steps: 7710, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001333, Sample Num: 21328, Cur Loss: 2.86043096, Cur Avg Loss: 1.63965550, Log Avg loss: 1.92275267, Global Avg Loss: 2.76430070, Time: 0.0564 Steps: 7720, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001343, Sample Num: 21488, Cur Loss: 1.03545809, Cur Avg Loss: 1.64076424, Log Avg loss: 1.78855993, Global Avg Loss: 2.76303842, Time: 0.0409 Steps: 7730, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001353, Sample Num: 21648, Cur Loss: 1.35457063, Cur Avg Loss: 1.64108373, Log Avg loss: 1.68399059, Global Avg Loss: 2.76164431, Time: 0.0799 Steps: 7740, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001363, Sample Num: 21808, Cur Loss: 3.14393520, Cur Avg Loss: 1.64189261, Log Avg loss: 1.75133508, Global Avg Loss: 2.76034068, Time: 0.0402 Steps: 7750, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001373, Sample Num: 21968, Cur Loss: 0.99881780, Cur Avg Loss: 1.63908058, Log Avg loss: 1.25580044, Global Avg Loss: 2.75840184, Time: 0.0403 Steps: 7760, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001383, Sample Num: 22128, Cur Loss: 0.95847154, Cur Avg Loss: 1.63900051, Log Avg loss: 1.62800753, Global Avg Loss: 2.75694702, Time: 0.0402 Steps: 7770, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001393, Sample Num: 22288, Cur Loss: 1.63610125, Cur Avg Loss: 1.63755175, Log Avg loss: 1.43718702, Global Avg Loss: 2.75525067, Time: 0.0403 Steps: 7780, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001403, Sample Num: 22448, Cur Loss: 2.60554218, Cur Avg Loss: 1.63813577, Log Avg loss: 1.71949034, Global Avg Loss: 2.75392107, Time: 0.0402 Steps: 7790, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001413, Sample Num: 22608, Cur Loss: 1.67322803, Cur Avg Loss: 1.63904442, Log Avg loss: 1.76652813, Global Avg Loss: 2.75265518, Time: 0.0402 Steps: 7800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001423, Sample Num: 22768, Cur Loss: 1.95680833, Cur Avg Loss: 1.63969227, Log Avg loss: 1.73123323, Global Avg Loss: 2.75134734, Time: 0.0403 Steps: 7810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001433, Sample Num: 22928, Cur Loss: 1.88671100, Cur Avg Loss: 1.63898029, Log Avg loss: 1.53766632, Global Avg Loss: 2.74979532, Time: 0.0402 Steps: 7820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001443, Sample Num: 23088, Cur Loss: 2.13053942, Cur Avg Loss: 1.64213222, Log Avg loss: 2.09380301, Global Avg Loss: 2.74895753, Time: 0.0402 Steps: 7830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001453, Sample Num: 23248, Cur Loss: 2.04544687, Cur Avg Loss: 1.64230216, Log Avg loss: 1.66682438, Global Avg Loss: 2.74757726, Time: 0.0402 Steps: 7840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001463, Sample Num: 23408, Cur Loss: 1.83328152, Cur Avg Loss: 1.64223908, Log Avg loss: 1.63307314, Global Avg Loss: 2.74615751, Time: 0.0403 Steps: 7850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001473, Sample Num: 23568, Cur Loss: 2.08768702, Cur Avg Loss: 1.63962026, Log Avg loss: 1.25648745, Global Avg Loss: 2.74426225, Time: 0.0402 Steps: 7860, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001483, Sample Num: 23728, Cur Loss: 1.96389866, Cur Avg Loss: 1.63962761, Log Avg loss: 1.64070994, Global Avg Loss: 2.74286002, Time: 0.0402 Steps: 7870, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001493, Sample Num: 23888, Cur Loss: 1.04676986, Cur Avg Loss: 1.63967585, Log Avg loss: 1.64683062, Global Avg Loss: 2.74146912, Time: 0.0402 Steps: 7880, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001503, Sample Num: 24048, Cur Loss: 1.47948492, Cur Avg Loss: 1.63993166, Log Avg loss: 1.67812407, Global Avg Loss: 2.74012141, Time: 0.0403 Steps: 7890, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001513, Sample Num: 24208, Cur Loss: 1.64846134, Cur Avg Loss: 1.63902184, Log Avg loss: 1.50227618, Global Avg Loss: 2.73855452, Time: 0.0402 Steps: 7900, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001523, Sample Num: 24368, Cur Loss: 1.50721574, Cur Avg Loss: 1.64006905, Log Avg loss: 1.79851129, Global Avg Loss: 2.73736610, Time: 0.0403 Steps: 7910, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001533, Sample Num: 24528, Cur Loss: 2.77139163, Cur Avg Loss: 1.64161000, Log Avg loss: 1.87629670, Global Avg Loss: 2.73627889, Time: 0.0402 Steps: 7920, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001543, Sample Num: 24688, Cur Loss: 0.76083922, Cur Avg Loss: 1.64176478, Log Avg loss: 1.66549333, Global Avg Loss: 2.73492859, Time: 0.0403 Steps: 7930, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001553, Sample Num: 24848, Cur Loss: 1.31925702, Cur Avg Loss: 1.64227798, Log Avg loss: 1.72146378, Global Avg Loss: 2.73365219, Time: 0.0402 Steps: 7940, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001563, Sample Num: 25008, Cur Loss: 1.35048342, Cur Avg Loss: 1.64134636, Log Avg loss: 1.49666564, Global Avg Loss: 2.73209623, Time: 0.0402 Steps: 7950, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001573, Sample Num: 25168, Cur Loss: 1.50853038, Cur Avg Loss: 1.64155928, Log Avg loss: 1.67483895, Global Avg Loss: 2.73076801, Time: 0.0403 Steps: 7960, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001583, Sample Num: 25328, Cur Loss: 2.20006990, Cur Avg Loss: 1.64295678, Log Avg loss: 1.86278402, Global Avg Loss: 2.72967895, Time: 0.0402 Steps: 7970, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001593, Sample Num: 25488, Cur Loss: 2.31070590, Cur Avg Loss: 1.64297763, Log Avg loss: 1.64627764, Global Avg Loss: 2.72832130, Time: 0.0402 Steps: 7980, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001603, Sample Num: 25648, Cur Loss: 0.75548756, Cur Avg Loss: 1.64173166, Log Avg loss: 1.44324913, Global Avg Loss: 2.72671295, Time: 0.0402 Steps: 7990, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001613, Sample Num: 25808, Cur Loss: 2.25802374, Cur Avg Loss: 1.64363249, Log Avg loss: 1.94833496, Global Avg Loss: 2.72573998, Time: 0.0402 Steps: 8000, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001623, Sample Num: 25968, Cur Loss: 1.22538674, Cur Avg Loss: 1.64178776, Log Avg loss: 1.34423310, Global Avg Loss: 2.72401525, Time: 0.0402 Steps: 8010, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001633, Sample Num: 26128, Cur Loss: 2.55021667, Cur Avg Loss: 1.64192697, Log Avg loss: 1.66452020, Global Avg Loss: 2.72269419, Time: 0.0656 Steps: 8020, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001643, Sample Num: 26288, Cur Loss: 1.53419602, Cur Avg Loss: 1.64089476, Log Avg loss: 1.47233524, Global Avg Loss: 2.72113708, Time: 0.0432 Steps: 8030, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001653, Sample Num: 26448, Cur Loss: 2.10005307, Cur Avg Loss: 1.64189591, Log Avg loss: 1.80638572, Global Avg Loss: 2.71999933, Time: 0.0765 Steps: 8040, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001663, Sample Num: 26608, Cur Loss: 1.57498348, Cur Avg Loss: 1.64109739, Log Avg loss: 1.50910208, Global Avg Loss: 2.71849511, Time: 0.0479 Steps: 8050, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001673, Sample Num: 26768, Cur Loss: 1.21854138, Cur Avg Loss: 1.64239125, Log Avg loss: 1.85755867, Global Avg Loss: 2.71742695, Time: 0.0405 Steps: 8060, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001683, Sample Num: 26928, Cur Loss: 1.52052808, Cur Avg Loss: 1.64207541, Log Avg loss: 1.58923601, Global Avg Loss: 2.71602894, Time: 0.0406 Steps: 8070, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001693, Sample Num: 27088, Cur Loss: 1.91532302, Cur Avg Loss: 1.64188345, Log Avg loss: 1.60957720, Global Avg Loss: 2.71465957, Time: 0.0505 Steps: 8080, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001703, Sample Num: 27248, Cur Loss: 1.46341324, Cur Avg Loss: 1.64094856, Log Avg loss: 1.48267132, Global Avg Loss: 2.71313672, Time: 0.0406 Steps: 8090, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001713, Sample Num: 27408, Cur Loss: 1.45256305, Cur Avg Loss: 1.64056106, Log Avg loss: 1.57456886, Global Avg Loss: 2.71173108, Time: 0.0477 Steps: 8100, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001723, Sample Num: 27568, Cur Loss: 1.87046480, Cur Avg Loss: 1.64186030, Log Avg loss: 1.86442022, Global Avg Loss: 2.71068631, Time: 0.0403 Steps: 8110, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001733, Sample Num: 27728, Cur Loss: 1.08056498, Cur Avg Loss: 1.64034836, Log Avg loss: 1.37984245, Global Avg Loss: 2.70904734, Time: 0.0403 Steps: 8120, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001743, Sample Num: 27888, Cur Loss: 2.01579666, Cur Avg Loss: 1.64046593, Log Avg loss: 1.66084063, Global Avg Loss: 2.70775803, Time: 0.0402 Steps: 8130, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001753, Sample Num: 28048, Cur Loss: 1.93587792, Cur Avg Loss: 1.64216674, Log Avg loss: 1.93861687, Global Avg Loss: 2.70681314, Time: 0.0402 Steps: 8140, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001763, Sample Num: 28208, Cur Loss: 0.78684080, Cur Avg Loss: 1.63911945, Log Avg loss: 1.10493008, Global Avg Loss: 2.70484764, Time: 0.0402 Steps: 8150, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001773, Sample Num: 28368, Cur Loss: 2.42028880, Cur Avg Loss: 1.64341184, Log Avg loss: 2.40015946, Global Avg Loss: 2.70447424, Time: 0.0403 Steps: 8160, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001783, Sample Num: 28528, Cur Loss: 0.86640888, Cur Avg Loss: 1.64275251, Log Avg loss: 1.52585421, Global Avg Loss: 2.70303163, Time: 0.0402 Steps: 8170, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001793, Sample Num: 28688, Cur Loss: 2.49431014, Cur Avg Loss: 1.64352405, Log Avg loss: 1.78108995, Global Avg Loss: 2.70190456, Time: 0.0403 Steps: 8180, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001803, Sample Num: 28848, Cur Loss: 1.63811851, Cur Avg Loss: 1.64526274, Log Avg loss: 1.95700852, Global Avg Loss: 2.70099504, Time: 0.0402 Steps: 8190, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001813, Sample Num: 29008, Cur Loss: 0.73957789, Cur Avg Loss: 1.64426050, Log Avg loss: 1.46355668, Global Avg Loss: 2.69948597, Time: 0.0402 Steps: 8200, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001823, Sample Num: 29168, Cur Loss: 3.03078318, Cur Avg Loss: 1.64499888, Log Avg loss: 1.77886724, Global Avg Loss: 2.69836463, Time: 0.0402 Steps: 8210, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001833, Sample Num: 29328, Cur Loss: 1.52811432, Cur Avg Loss: 1.64595042, Log Avg loss: 1.81941692, Global Avg Loss: 2.69729535, Time: 0.0403 Steps: 8220, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001843, Sample Num: 29488, Cur Loss: 1.33266497, Cur Avg Loss: 1.64515123, Log Avg loss: 1.49866006, Global Avg Loss: 2.69583893, Time: 0.0402 Steps: 8230, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001853, Sample Num: 29648, Cur Loss: 0.85106748, Cur Avg Loss: 1.64495532, Log Avg loss: 1.60884885, Global Avg Loss: 2.69451976, Time: 0.0403 Steps: 8240, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001863, Sample Num: 29808, Cur Loss: 2.76814675, Cur Avg Loss: 1.64367575, Log Avg loss: 1.40657132, Global Avg Loss: 2.69295861, Time: 0.0403 Steps: 8250, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001873, Sample Num: 29968, Cur Loss: 1.67357039, Cur Avg Loss: 1.64442762, Log Avg loss: 1.78450143, Global Avg Loss: 2.69185879, Time: 0.0403 Steps: 8260, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001883, Sample Num: 30128, Cur Loss: 1.21252823, Cur Avg Loss: 1.64355990, Log Avg loss: 1.48103535, Global Avg Loss: 2.69039467, Time: 0.0403 Steps: 8270, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001893, Sample Num: 30288, Cur Loss: 1.53450608, Cur Avg Loss: 1.64411456, Log Avg loss: 1.74855810, Global Avg Loss: 2.68925719, Time: 0.0403 Steps: 8280, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001903, Sample Num: 30448, Cur Loss: 1.74870217, Cur Avg Loss: 1.64469324, Log Avg loss: 1.75423619, Global Avg Loss: 2.68812930, Time: 0.0403 Steps: 8290, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001913, Sample Num: 30608, Cur Loss: 1.61176586, Cur Avg Loss: 1.64645491, Log Avg loss: 1.98170053, Global Avg Loss: 2.68727818, Time: 0.0402 Steps: 8300, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001923, Sample Num: 30768, Cur Loss: 1.05612493, Cur Avg Loss: 1.64598688, Log Avg loss: 1.55645392, Global Avg Loss: 2.68591738, Time: 0.0402 Steps: 8310, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001933, Sample Num: 30928, Cur Loss: 2.15113854, Cur Avg Loss: 1.64598560, Log Avg loss: 1.64573803, Global Avg Loss: 2.68466716, Time: 0.0403 Steps: 8320, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001943, Sample Num: 31088, Cur Loss: 1.12083828, Cur Avg Loss: 1.64462687, Log Avg loss: 1.38198564, Global Avg Loss: 2.68310332, Time: 0.0402 Steps: 8330, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001953, Sample Num: 31248, Cur Loss: 0.47543329, Cur Avg Loss: 1.64516153, Log Avg loss: 1.74904518, Global Avg Loss: 2.68198335, Time: 0.0402 Steps: 8340, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001963, Sample Num: 31408, Cur Loss: 1.55348706, Cur Avg Loss: 1.64647131, Log Avg loss: 1.90227206, Global Avg Loss: 2.68104956, Time: 0.0402 Steps: 8350, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001973, Sample Num: 31568, Cur Loss: 1.11825383, Cur Avg Loss: 1.64484608, Log Avg loss: 1.32581218, Global Avg Loss: 2.67942846, Time: 0.0588 Steps: 8360, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001983, Sample Num: 31728, Cur Loss: 2.28168416, Cur Avg Loss: 1.64872714, Log Avg loss: 2.41446185, Global Avg Loss: 2.67911190, Time: 0.0548 Steps: 8370, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001993, Sample Num: 31888, Cur Loss: 1.96371162, Cur Avg Loss: 1.64969803, Log Avg loss: 1.84222382, Global Avg Loss: 2.67811322, Time: 0.0403 Steps: 8380, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002003, Sample Num: 32048, Cur Loss: 2.23206663, Cur Avg Loss: 1.65063971, Log Avg loss: 1.83831829, Global Avg Loss: 2.67711228, Time: 0.0890 Steps: 8390, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002013, Sample Num: 32208, Cur Loss: 1.28771305, Cur Avg Loss: 1.64942600, Log Avg loss: 1.40631831, Global Avg Loss: 2.67559943, Time: 0.0467 Steps: 8400, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002023, Sample Num: 32368, Cur Loss: 0.81077451, Cur Avg Loss: 1.64969893, Log Avg loss: 1.70464098, Global Avg Loss: 2.67444490, Time: 0.0660 Steps: 8410, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002033, Sample Num: 32528, Cur Loss: 1.15645576, Cur Avg Loss: 1.64959414, Log Avg loss: 1.62839428, Global Avg Loss: 2.67320256, Time: 0.0757 Steps: 8420, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002043, Sample Num: 32688, Cur Loss: 2.00143528, Cur Avg Loss: 1.65115580, Log Avg loss: 1.96864177, Global Avg Loss: 2.67236678, Time: 0.0849 Steps: 8430, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002053, Sample Num: 32848, Cur Loss: 0.43918231, Cur Avg Loss: 1.64955606, Log Avg loss: 1.32272969, Global Avg Loss: 2.67076768, Time: 0.0402 Steps: 8440, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002063, Sample Num: 33008, Cur Loss: 0.99970043, Cur Avg Loss: 1.65103837, Log Avg loss: 1.95535548, Global Avg Loss: 2.66992104, Time: 0.0403 Steps: 8450, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002073, Sample Num: 33168, Cur Loss: 1.71288359, Cur Avg Loss: 1.65071987, Log Avg loss: 1.58501290, Global Avg Loss: 2.66863864, Time: 0.0402 Steps: 8460, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002083, Sample Num: 33328, Cur Loss: 2.05849457, Cur Avg Loss: 1.64964724, Log Avg loss: 1.42729126, Global Avg Loss: 2.66717306, Time: 0.0402 Steps: 8470, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002093, Sample Num: 33488, Cur Loss: 1.16081500, Cur Avg Loss: 1.65078333, Log Avg loss: 1.88743168, Global Avg Loss: 2.66625356, Time: 0.0402 Steps: 8480, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002103, Sample Num: 33648, Cur Loss: 0.95069796, Cur Avg Loss: 1.64977743, Log Avg loss: 1.43924136, Global Avg Loss: 2.66480831, Time: 0.0403 Steps: 8490, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002113, Sample Num: 33808, Cur Loss: 1.35727811, Cur Avg Loss: 1.64882664, Log Avg loss: 1.44887704, Global Avg Loss: 2.66337781, Time: 0.0403 Steps: 8500, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002123, Sample Num: 33968, Cur Loss: 2.63385344, Cur Avg Loss: 1.64785449, Log Avg loss: 1.44243845, Global Avg Loss: 2.66194309, Time: 0.0402 Steps: 8510, Updated lr: 0.000093 ***** Running evaluation checkpoint-8516 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-8516 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.294317, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.478172, "eval_total_loss": 1039.154813, "eval_mae": 1.024173, "eval_mse": 1.478231, "eval_r2": 0.060339, "eval_sp_statistic": 0.206291, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.287303, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.515598, "test_total_loss": 760.829992, "test_mae": 1.059205, "test_mse": 1.515761, "test_r2": 0.021714, "test_sp_statistic": 0.207885, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.292692, "test_ps_pvalue": 0.0, "lr": 9.287245139876719e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.6613093581787295, "train_cur_epoch_loss": 3508.969844073057, "train_cur_epoch_avg_loss": 1.6481774749051465, "train_cur_epoch_time": 95.29431653022766, "train_cur_epoch_avg_time": 0.04476012988737795, "epoch": 4, "step": 8516} ################################################## Training, Epoch: 0005, Batch: 000004, Sample Num: 64, Cur Loss: 2.13376832, Cur Avg Loss: 2.34916639, Log Avg loss: 1.99714297, Global Avg Loss: 2.66116281, Time: 0.0402 Steps: 8520, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000014, Sample Num: 224, Cur Loss: 1.10528195, Cur Avg Loss: 1.91374327, Log Avg loss: 1.73957403, Global Avg Loss: 2.66008240, Time: 0.0403 Steps: 8530, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000024, Sample Num: 384, Cur Loss: 2.87396264, Cur Avg Loss: 1.84442721, Log Avg loss: 1.74738472, Global Avg Loss: 2.65901367, Time: 0.0401 Steps: 8540, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000034, Sample Num: 544, Cur Loss: 1.19540286, Cur Avg Loss: 1.81130112, Log Avg loss: 1.73179851, Global Avg Loss: 2.65792921, Time: 0.0403 Steps: 8550, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000044, Sample Num: 704, Cur Loss: 1.55976248, Cur Avg Loss: 1.74235684, Log Avg loss: 1.50794628, Global Avg Loss: 2.65658577, Time: 0.0402 Steps: 8560, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000054, Sample Num: 864, Cur Loss: 0.96684831, Cur Avg Loss: 1.69537003, Log Avg loss: 1.48862810, Global Avg Loss: 2.65522293, Time: 0.0403 Steps: 8570, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000064, Sample Num: 1024, Cur Loss: 1.48556054, Cur Avg Loss: 1.72059322, Log Avg loss: 1.85679846, Global Avg Loss: 2.65429236, Time: 0.0402 Steps: 8580, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000074, Sample Num: 1184, Cur Loss: 1.13214731, Cur Avg Loss: 1.68906759, Log Avg loss: 1.48730357, Global Avg Loss: 2.65293382, Time: 0.0402 Steps: 8590, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000084, Sample Num: 1344, Cur Loss: 1.08542645, Cur Avg Loss: 1.66958643, Log Avg loss: 1.52542585, Global Avg Loss: 2.65162276, Time: 0.0404 Steps: 8600, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000094, Sample Num: 1504, Cur Loss: 1.00110769, Cur Avg Loss: 1.64418119, Log Avg loss: 1.43077711, Global Avg Loss: 2.65020482, Time: 0.0402 Steps: 8610, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000104, Sample Num: 1664, Cur Loss: 0.66135669, Cur Avg Loss: 1.60193826, Log Avg loss: 1.20485474, Global Avg Loss: 2.64852808, Time: 0.0402 Steps: 8620, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000114, Sample Num: 1824, Cur Loss: 0.78364801, Cur Avg Loss: 1.59596331, Log Avg loss: 1.53382381, Global Avg Loss: 2.64723642, Time: 0.0402 Steps: 8630, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000124, Sample Num: 1984, Cur Loss: 0.85868913, Cur Avg Loss: 1.60657316, Log Avg loss: 1.72752544, Global Avg Loss: 2.64617194, Time: 0.0402 Steps: 8640, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000134, Sample Num: 2144, Cur Loss: 1.69512558, Cur Avg Loss: 1.61058155, Log Avg loss: 1.66028556, Global Avg Loss: 2.64503219, Time: 0.0403 Steps: 8650, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000144, Sample Num: 2304, Cur Loss: 1.16129708, Cur Avg Loss: 1.62161052, Log Avg loss: 1.76939875, Global Avg Loss: 2.64402106, Time: 0.0402 Steps: 8660, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000154, Sample Num: 2464, Cur Loss: 1.74183989, Cur Avg Loss: 1.62995668, Log Avg loss: 1.75014146, Global Avg Loss: 2.64299006, Time: 0.0402 Steps: 8670, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000164, Sample Num: 2624, Cur Loss: 1.37104011, Cur Avg Loss: 1.61876376, Log Avg loss: 1.44639270, Global Avg Loss: 2.64161149, Time: 0.0402 Steps: 8680, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000174, Sample Num: 2784, Cur Loss: 1.66741061, Cur Avg Loss: 1.61102281, Log Avg loss: 1.48407130, Global Avg Loss: 2.64027945, Time: 0.0402 Steps: 8690, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000184, Sample Num: 2944, Cur Loss: 1.58048248, Cur Avg Loss: 1.61498821, Log Avg loss: 1.68398616, Global Avg Loss: 2.63918027, Time: 0.0402 Steps: 8700, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000194, Sample Num: 3104, Cur Loss: 0.97888017, Cur Avg Loss: 1.61330264, Log Avg loss: 1.58228816, Global Avg Loss: 2.63796684, Time: 0.0402 Steps: 8710, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000204, Sample Num: 3264, Cur Loss: 0.48859906, Cur Avg Loss: 1.60274064, Log Avg loss: 1.39783785, Global Avg Loss: 2.63654468, Time: 0.0402 Steps: 8720, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000214, Sample Num: 3424, Cur Loss: 1.52702773, Cur Avg Loss: 1.61487246, Log Avg loss: 1.86236157, Global Avg Loss: 2.63565787, Time: 0.0479 Steps: 8730, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000224, Sample Num: 3584, Cur Loss: 0.88473344, Cur Avg Loss: 1.60727991, Log Avg loss: 1.44479943, Global Avg Loss: 2.63429533, Time: 0.0403 Steps: 8740, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000234, Sample Num: 3744, Cur Loss: 2.45654726, Cur Avg Loss: 1.61133461, Log Avg loss: 1.70215986, Global Avg Loss: 2.63323003, Time: 0.0923 Steps: 8750, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000244, Sample Num: 3904, Cur Loss: 1.24755776, Cur Avg Loss: 1.61202991, Log Avg loss: 1.62829990, Global Avg Loss: 2.63208285, Time: 0.0405 Steps: 8760, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000254, Sample Num: 4064, Cur Loss: 1.38751841, Cur Avg Loss: 1.59451201, Log Avg loss: 1.16707520, Global Avg Loss: 2.63041238, Time: 0.0404 Steps: 8770, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000264, Sample Num: 4224, Cur Loss: 1.42187130, Cur Avg Loss: 1.61222018, Log Avg loss: 2.06200767, Global Avg Loss: 2.62976499, Time: 0.0434 Steps: 8780, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000274, Sample Num: 4384, Cur Loss: 0.39839858, Cur Avg Loss: 1.61560437, Log Avg loss: 1.70494709, Global Avg Loss: 2.62871287, Time: 0.0448 Steps: 8790, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000284, Sample Num: 4544, Cur Loss: 2.49471426, Cur Avg Loss: 1.62536643, Log Avg loss: 1.89284684, Global Avg Loss: 2.62787665, Time: 0.0799 Steps: 8800, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000294, Sample Num: 4704, Cur Loss: 0.89993989, Cur Avg Loss: 1.63276522, Log Avg loss: 1.84289085, Global Avg Loss: 2.62698564, Time: 0.0479 Steps: 8810, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000304, Sample Num: 4864, Cur Loss: 1.90390301, Cur Avg Loss: 1.62935830, Log Avg loss: 1.52919477, Global Avg Loss: 2.62574098, Time: 0.0405 Steps: 8820, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000314, Sample Num: 5024, Cur Loss: 1.60684228, Cur Avg Loss: 1.63098827, Log Avg loss: 1.68053936, Global Avg Loss: 2.62467053, Time: 0.0402 Steps: 8830, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000324, Sample Num: 5184, Cur Loss: 1.02823615, Cur Avg Loss: 1.61278926, Log Avg loss: 1.04134053, Global Avg Loss: 2.62287944, Time: 0.0402 Steps: 8840, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000334, Sample Num: 5344, Cur Loss: 2.79671025, Cur Avg Loss: 1.61370163, Log Avg loss: 1.64326215, Global Avg Loss: 2.62177252, Time: 0.0401 Steps: 8850, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000344, Sample Num: 5504, Cur Loss: 0.93682885, Cur Avg Loss: 1.61693959, Log Avg loss: 1.72508751, Global Avg Loss: 2.62076046, Time: 0.0402 Steps: 8860, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000354, Sample Num: 5664, Cur Loss: 1.73782229, Cur Avg Loss: 1.61895043, Log Avg loss: 1.68812337, Global Avg Loss: 2.61970901, Time: 0.0402 Steps: 8870, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000364, Sample Num: 5824, Cur Loss: 1.19219339, Cur Avg Loss: 1.61416113, Log Avg loss: 1.44462000, Global Avg Loss: 2.61838571, Time: 0.0402 Steps: 8880, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000374, Sample Num: 5984, Cur Loss: 2.04297209, Cur Avg Loss: 1.60634444, Log Avg loss: 1.32181696, Global Avg Loss: 2.61692726, Time: 0.0402 Steps: 8890, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000384, Sample Num: 6144, Cur Loss: 1.85594499, Cur Avg Loss: 1.60729880, Log Avg loss: 1.64299197, Global Avg Loss: 2.61583295, Time: 0.0402 Steps: 8900, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000394, Sample Num: 6304, Cur Loss: 1.24360561, Cur Avg Loss: 1.60812101, Log Avg loss: 1.63969375, Global Avg Loss: 2.61473739, Time: 0.0402 Steps: 8910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000404, Sample Num: 6464, Cur Loss: 0.56500840, Cur Avg Loss: 1.60174803, Log Avg loss: 1.35065242, Global Avg Loss: 2.61332026, Time: 0.0402 Steps: 8920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000414, Sample Num: 6624, Cur Loss: 0.82590622, Cur Avg Loss: 1.59373471, Log Avg loss: 1.26999669, Global Avg Loss: 2.61181598, Time: 0.0402 Steps: 8930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000424, Sample Num: 6784, Cur Loss: 1.63786387, Cur Avg Loss: 1.60019000, Log Avg loss: 1.86743897, Global Avg Loss: 2.61098334, Time: 0.0402 Steps: 8940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000434, Sample Num: 6944, Cur Loss: 1.52869022, Cur Avg Loss: 1.60242686, Log Avg loss: 1.69726959, Global Avg Loss: 2.60996243, Time: 0.0403 Steps: 8950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000444, Sample Num: 7104, Cur Loss: 1.50418544, Cur Avg Loss: 1.60187058, Log Avg loss: 1.57772810, Global Avg Loss: 2.60881038, Time: 0.0402 Steps: 8960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000454, Sample Num: 7264, Cur Loss: 0.40425727, Cur Avg Loss: 1.59649070, Log Avg loss: 1.35762425, Global Avg Loss: 2.60741553, Time: 0.0403 Steps: 8970, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000464, Sample Num: 7424, Cur Loss: 1.62084877, Cur Avg Loss: 1.59080365, Log Avg loss: 1.33261147, Global Avg Loss: 2.60599592, Time: 0.0402 Steps: 8980, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000474, Sample Num: 7584, Cur Loss: 2.58981180, Cur Avg Loss: 1.59199651, Log Avg loss: 1.64734529, Global Avg Loss: 2.60492957, Time: 0.0402 Steps: 8990, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000484, Sample Num: 7744, Cur Loss: 0.66949129, Cur Avg Loss: 1.59436159, Log Avg loss: 1.70646638, Global Avg Loss: 2.60393128, Time: 0.0402 Steps: 9000, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000494, Sample Num: 7904, Cur Loss: 1.68215787, Cur Avg Loss: 1.60201437, Log Avg loss: 1.97240911, Global Avg Loss: 2.60323037, Time: 0.0403 Steps: 9010, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000504, Sample Num: 8064, Cur Loss: 2.81867242, Cur Avg Loss: 1.60177772, Log Avg loss: 1.59008700, Global Avg Loss: 2.60210715, Time: 0.0402 Steps: 9020, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000514, Sample Num: 8224, Cur Loss: 0.94388938, Cur Avg Loss: 1.59577350, Log Avg loss: 1.29316095, Global Avg Loss: 2.60065759, Time: 0.0402 Steps: 9030, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000524, Sample Num: 8384, Cur Loss: 0.99787086, Cur Avg Loss: 1.59882890, Log Avg loss: 1.75587617, Global Avg Loss: 2.59972310, Time: 0.0402 Steps: 9040, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000534, Sample Num: 8544, Cur Loss: 2.05639386, Cur Avg Loss: 1.59716860, Log Avg loss: 1.51016886, Global Avg Loss: 2.59851917, Time: 0.0402 Steps: 9050, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000544, Sample Num: 8704, Cur Loss: 1.28225517, Cur Avg Loss: 1.59610332, Log Avg loss: 1.53921766, Global Avg Loss: 2.59734997, Time: 0.0402 Steps: 9060, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000554, Sample Num: 8864, Cur Loss: 0.85325766, Cur Avg Loss: 1.59499202, Log Avg loss: 1.53453718, Global Avg Loss: 2.59617818, Time: 0.0402 Steps: 9070, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000564, Sample Num: 9024, Cur Loss: 0.97450930, Cur Avg Loss: 1.58802867, Log Avg loss: 1.20225918, Global Avg Loss: 2.59464302, Time: 0.0402 Steps: 9080, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000574, Sample Num: 9184, Cur Loss: 2.01581931, Cur Avg Loss: 1.58883368, Log Avg loss: 1.63423629, Global Avg Loss: 2.59358647, Time: 0.0464 Steps: 9090, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000584, Sample Num: 9344, Cur Loss: 0.78021151, Cur Avg Loss: 1.58369577, Log Avg loss: 1.28877977, Global Avg Loss: 2.59215262, Time: 0.0456 Steps: 9100, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000594, Sample Num: 9504, Cur Loss: 1.00627172, Cur Avg Loss: 1.58314957, Log Avg loss: 1.55125152, Global Avg Loss: 2.59101003, Time: 0.0624 Steps: 9110, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000604, Sample Num: 9664, Cur Loss: 3.07179761, Cur Avg Loss: 1.58408067, Log Avg loss: 1.63938750, Global Avg Loss: 2.58996658, Time: 0.0780 Steps: 9120, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000614, Sample Num: 9824, Cur Loss: 1.46980691, Cur Avg Loss: 1.58312418, Log Avg loss: 1.52535256, Global Avg Loss: 2.58880052, Time: 0.0714 Steps: 9130, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000624, Sample Num: 9984, Cur Loss: 2.14140511, Cur Avg Loss: 1.58468589, Log Avg loss: 1.68057449, Global Avg Loss: 2.58780684, Time: 0.0405 Steps: 9140, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000634, Sample Num: 10144, Cur Loss: 1.63026154, Cur Avg Loss: 1.58414581, Log Avg loss: 1.55044531, Global Avg Loss: 2.58667311, Time: 0.0659 Steps: 9150, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000644, Sample Num: 10304, Cur Loss: 1.31240487, Cur Avg Loss: 1.58633169, Log Avg loss: 1.72491612, Global Avg Loss: 2.58573233, Time: 0.0479 Steps: 9160, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000654, Sample Num: 10464, Cur Loss: 1.19631338, Cur Avg Loss: 1.58047636, Log Avg loss: 1.20339328, Global Avg Loss: 2.58422487, Time: 0.0404 Steps: 9170, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000664, Sample Num: 10624, Cur Loss: 1.89880931, Cur Avg Loss: 1.57717190, Log Avg loss: 1.36105992, Global Avg Loss: 2.58289244, Time: 0.0402 Steps: 9180, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000674, Sample Num: 10784, Cur Loss: 0.56393361, Cur Avg Loss: 1.57320868, Log Avg loss: 1.31005110, Global Avg Loss: 2.58150741, Time: 0.0401 Steps: 9190, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000684, Sample Num: 10944, Cur Loss: 0.35322624, Cur Avg Loss: 1.56620067, Log Avg loss: 1.09386071, Global Avg Loss: 2.57989041, Time: 0.0402 Steps: 9200, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000694, Sample Num: 11104, Cur Loss: 1.43838000, Cur Avg Loss: 1.56285561, Log Avg loss: 1.33405344, Global Avg Loss: 2.57853771, Time: 0.0402 Steps: 9210, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000704, Sample Num: 11264, Cur Loss: 1.27535486, Cur Avg Loss: 1.56630428, Log Avg loss: 1.80564249, Global Avg Loss: 2.57769943, Time: 0.0402 Steps: 9220, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000714, Sample Num: 11424, Cur Loss: 0.92541325, Cur Avg Loss: 1.56570004, Log Avg loss: 1.52316110, Global Avg Loss: 2.57655691, Time: 0.0402 Steps: 9230, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000724, Sample Num: 11584, Cur Loss: 1.48199666, Cur Avg Loss: 1.56678424, Log Avg loss: 1.64419614, Global Avg Loss: 2.57554787, Time: 0.0402 Steps: 9240, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000734, Sample Num: 11744, Cur Loss: 1.58184326, Cur Avg Loss: 1.56392078, Log Avg loss: 1.35660611, Global Avg Loss: 2.57423009, Time: 0.0401 Steps: 9250, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000744, Sample Num: 11904, Cur Loss: 1.56233025, Cur Avg Loss: 1.57308819, Log Avg loss: 2.24597609, Global Avg Loss: 2.57387561, Time: 0.0403 Steps: 9260, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000754, Sample Num: 12064, Cur Loss: 0.79197264, Cur Avg Loss: 1.57159732, Log Avg loss: 1.46067693, Global Avg Loss: 2.57267474, Time: 0.0402 Steps: 9270, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000764, Sample Num: 12224, Cur Loss: 2.78891134, Cur Avg Loss: 1.57237034, Log Avg loss: 1.63065598, Global Avg Loss: 2.57165964, Time: 0.0403 Steps: 9280, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000774, Sample Num: 12384, Cur Loss: 1.73007786, Cur Avg Loss: 1.56838234, Log Avg loss: 1.26369887, Global Avg Loss: 2.57025171, Time: 0.0402 Steps: 9290, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000784, Sample Num: 12544, Cur Loss: 1.48778796, Cur Avg Loss: 1.56608149, Log Avg loss: 1.38799570, Global Avg Loss: 2.56898047, Time: 0.0402 Steps: 9300, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000794, Sample Num: 12704, Cur Loss: 2.28188920, Cur Avg Loss: 1.56474273, Log Avg loss: 1.45978428, Global Avg Loss: 2.56778907, Time: 0.0402 Steps: 9310, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000804, Sample Num: 12864, Cur Loss: 1.57028329, Cur Avg Loss: 1.56423587, Log Avg loss: 1.52399099, Global Avg Loss: 2.56666911, Time: 0.0402 Steps: 9320, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000814, Sample Num: 13024, Cur Loss: 1.96901178, Cur Avg Loss: 1.56538514, Log Avg loss: 1.65778636, Global Avg Loss: 2.56569496, Time: 0.0402 Steps: 9330, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000824, Sample Num: 13184, Cur Loss: 1.49813366, Cur Avg Loss: 1.56643145, Log Avg loss: 1.65160151, Global Avg Loss: 2.56471628, Time: 0.0402 Steps: 9340, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000834, Sample Num: 13344, Cur Loss: 2.82622600, Cur Avg Loss: 1.56725289, Log Avg loss: 1.63493915, Global Avg Loss: 2.56372186, Time: 0.0402 Steps: 9350, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000844, Sample Num: 13504, Cur Loss: 2.05692816, Cur Avg Loss: 1.56865772, Log Avg loss: 1.68582066, Global Avg Loss: 2.56278393, Time: 0.0403 Steps: 9360, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000854, Sample Num: 13664, Cur Loss: 1.29553831, Cur Avg Loss: 1.56909998, Log Avg loss: 1.60642680, Global Avg Loss: 2.56176327, Time: 0.0402 Steps: 9370, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000864, Sample Num: 13824, Cur Loss: 2.05107689, Cur Avg Loss: 1.56946163, Log Avg loss: 1.60034633, Global Avg Loss: 2.56073831, Time: 0.0402 Steps: 9380, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000874, Sample Num: 13984, Cur Loss: 1.55465221, Cur Avg Loss: 1.56877553, Log Avg loss: 1.50949652, Global Avg Loss: 2.55961878, Time: 0.0402 Steps: 9390, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000884, Sample Num: 14144, Cur Loss: 1.41008759, Cur Avg Loss: 1.57350068, Log Avg loss: 1.98647890, Global Avg Loss: 2.55900905, Time: 0.0402 Steps: 9400, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000894, Sample Num: 14304, Cur Loss: 3.39675093, Cur Avg Loss: 1.57488321, Log Avg loss: 1.69709901, Global Avg Loss: 2.55809310, Time: 0.0403 Steps: 9410, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000904, Sample Num: 14464, Cur Loss: 1.37598419, Cur Avg Loss: 1.57381381, Log Avg loss: 1.47820907, Global Avg Loss: 2.55694673, Time: 0.0401 Steps: 9420, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000914, Sample Num: 14624, Cur Loss: 1.67082572, Cur Avg Loss: 1.57466972, Log Avg loss: 1.65204446, Global Avg Loss: 2.55598713, Time: 0.0402 Steps: 9430, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000924, Sample Num: 14784, Cur Loss: 1.67530465, Cur Avg Loss: 1.57620043, Log Avg loss: 1.71610730, Global Avg Loss: 2.55509743, Time: 0.0404 Steps: 9440, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000934, Sample Num: 14944, Cur Loss: 2.82494164, Cur Avg Loss: 1.57641634, Log Avg loss: 1.59636629, Global Avg Loss: 2.55408289, Time: 0.0418 Steps: 9450, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000944, Sample Num: 15104, Cur Loss: 0.91556692, Cur Avg Loss: 1.57696532, Log Avg loss: 1.62823974, Global Avg Loss: 2.55310420, Time: 0.0432 Steps: 9460, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000954, Sample Num: 15264, Cur Loss: 2.13372755, Cur Avg Loss: 1.57797256, Log Avg loss: 1.67305666, Global Avg Loss: 2.55217490, Time: 0.0404 Steps: 9470, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000964, Sample Num: 15424, Cur Loss: 0.72739041, Cur Avg Loss: 1.57692804, Log Avg loss: 1.47728089, Global Avg Loss: 2.55104105, Time: 0.0529 Steps: 9480, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000974, Sample Num: 15584, Cur Loss: 1.87824678, Cur Avg Loss: 1.57707888, Log Avg loss: 1.59161940, Global Avg Loss: 2.55003007, Time: 0.0742 Steps: 9490, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000984, Sample Num: 15744, Cur Loss: 1.18694329, Cur Avg Loss: 1.57782734, Log Avg loss: 1.65072719, Global Avg Loss: 2.54908343, Time: 0.0418 Steps: 9500, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000994, Sample Num: 15904, Cur Loss: 3.08405113, Cur Avg Loss: 1.57998458, Log Avg loss: 1.79225749, Global Avg Loss: 2.54828761, Time: 0.0501 Steps: 9510, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001004, Sample Num: 16064, Cur Loss: 1.38135421, Cur Avg Loss: 1.58067697, Log Avg loss: 1.64950006, Global Avg Loss: 2.54734351, Time: 0.0445 Steps: 9520, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001014, Sample Num: 16224, Cur Loss: 1.54805565, Cur Avg Loss: 1.57979132, Log Avg loss: 1.49087235, Global Avg Loss: 2.54623493, Time: 0.0405 Steps: 9530, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001024, Sample Num: 16384, Cur Loss: 1.67836905, Cur Avg Loss: 1.58421007, Log Avg loss: 2.03227103, Global Avg Loss: 2.54569618, Time: 0.0404 Steps: 9540, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001034, Sample Num: 16544, Cur Loss: 1.76023471, Cur Avg Loss: 1.58306537, Log Avg loss: 1.46584829, Global Avg Loss: 2.54456545, Time: 0.0403 Steps: 9550, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001044, Sample Num: 16704, Cur Loss: 1.64443898, Cur Avg Loss: 1.58796459, Log Avg loss: 2.09454409, Global Avg Loss: 2.54409472, Time: 0.0402 Steps: 9560, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001054, Sample Num: 16864, Cur Loss: 1.53112030, Cur Avg Loss: 1.58838950, Log Avg loss: 1.63275027, Global Avg Loss: 2.54314243, Time: 0.0403 Steps: 9570, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001064, Sample Num: 17024, Cur Loss: 1.62576926, Cur Avg Loss: 1.58903564, Log Avg loss: 1.65713893, Global Avg Loss: 2.54221758, Time: 0.0404 Steps: 9580, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001074, Sample Num: 17184, Cur Loss: 2.02484393, Cur Avg Loss: 1.59074294, Log Avg loss: 1.77239948, Global Avg Loss: 2.54141485, Time: 0.0402 Steps: 9590, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001084, Sample Num: 17344, Cur Loss: 0.73918718, Cur Avg Loss: 1.59026306, Log Avg loss: 1.53872389, Global Avg Loss: 2.54037038, Time: 0.0403 Steps: 9600, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001094, Sample Num: 17504, Cur Loss: 1.37941027, Cur Avg Loss: 1.59007277, Log Avg loss: 1.56944478, Global Avg Loss: 2.53936005, Time: 0.0403 Steps: 9610, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001104, Sample Num: 17664, Cur Loss: 1.53734708, Cur Avg Loss: 1.59101360, Log Avg loss: 1.69394060, Global Avg Loss: 2.53848124, Time: 0.0402 Steps: 9620, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001114, Sample Num: 17824, Cur Loss: 2.43124962, Cur Avg Loss: 1.59240499, Log Avg loss: 1.74601417, Global Avg Loss: 2.53765832, Time: 0.0403 Steps: 9630, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001124, Sample Num: 17984, Cur Loss: 0.88174021, Cur Avg Loss: 1.59228402, Log Avg loss: 1.57880889, Global Avg Loss: 2.53666367, Time: 0.0403 Steps: 9640, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001134, Sample Num: 18144, Cur Loss: 1.21716487, Cur Avg Loss: 1.59093395, Log Avg loss: 1.43918553, Global Avg Loss: 2.53552638, Time: 0.0402 Steps: 9650, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001144, Sample Num: 18304, Cur Loss: 1.42906606, Cur Avg Loss: 1.59000984, Log Avg loss: 1.48521639, Global Avg Loss: 2.53443911, Time: 0.0402 Steps: 9660, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001154, Sample Num: 18464, Cur Loss: 1.48269868, Cur Avg Loss: 1.59286922, Log Avg loss: 1.91998211, Global Avg Loss: 2.53380368, Time: 0.0402 Steps: 9670, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001164, Sample Num: 18624, Cur Loss: 1.39618254, Cur Avg Loss: 1.59058407, Log Avg loss: 1.32687709, Global Avg Loss: 2.53255685, Time: 0.0403 Steps: 9680, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001174, Sample Num: 18784, Cur Loss: 2.99491358, Cur Avg Loss: 1.58921185, Log Avg loss: 1.42948558, Global Avg Loss: 2.53141849, Time: 0.0404 Steps: 9690, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001184, Sample Num: 18944, Cur Loss: 2.45262575, Cur Avg Loss: 1.59080425, Log Avg loss: 1.77775202, Global Avg Loss: 2.53064152, Time: 0.0402 Steps: 9700, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001194, Sample Num: 19104, Cur Loss: 1.26394010, Cur Avg Loss: 1.59076167, Log Avg loss: 1.58571981, Global Avg Loss: 2.52966838, Time: 0.0402 Steps: 9710, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001204, Sample Num: 19264, Cur Loss: 1.72104239, Cur Avg Loss: 1.59041712, Log Avg loss: 1.54927852, Global Avg Loss: 2.52865974, Time: 0.0402 Steps: 9720, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001214, Sample Num: 19424, Cur Loss: 1.71237576, Cur Avg Loss: 1.59089240, Log Avg loss: 1.64811606, Global Avg Loss: 2.52775477, Time: 0.0402 Steps: 9730, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001224, Sample Num: 19584, Cur Loss: 1.44790220, Cur Avg Loss: 1.59228611, Log Avg loss: 1.76148279, Global Avg Loss: 2.52696804, Time: 0.0403 Steps: 9740, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001234, Sample Num: 19744, Cur Loss: 1.31819046, Cur Avg Loss: 1.58917291, Log Avg loss: 1.20811633, Global Avg Loss: 2.52561537, Time: 0.0401 Steps: 9750, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001244, Sample Num: 19904, Cur Loss: 1.58564651, Cur Avg Loss: 1.59031237, Log Avg loss: 1.73092177, Global Avg Loss: 2.52480113, Time: 0.0403 Steps: 9760, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001254, Sample Num: 20064, Cur Loss: 1.13100410, Cur Avg Loss: 1.58966004, Log Avg loss: 1.50851066, Global Avg Loss: 2.52376092, Time: 0.0402 Steps: 9770, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001264, Sample Num: 20224, Cur Loss: 0.94087040, Cur Avg Loss: 1.58866302, Log Avg loss: 1.46363612, Global Avg Loss: 2.52267695, Time: 0.0401 Steps: 9780, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001274, Sample Num: 20384, Cur Loss: 2.00493097, Cur Avg Loss: 1.58792757, Log Avg loss: 1.49496775, Global Avg Loss: 2.52162719, Time: 0.0402 Steps: 9790, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001284, Sample Num: 20544, Cur Loss: 0.71165800, Cur Avg Loss: 1.58784676, Log Avg loss: 1.57755123, Global Avg Loss: 2.52066385, Time: 0.0406 Steps: 9800, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001294, Sample Num: 20704, Cur Loss: 1.69559789, Cur Avg Loss: 1.58606015, Log Avg loss: 1.35665913, Global Avg Loss: 2.51947730, Time: 0.0503 Steps: 9810, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001304, Sample Num: 20864, Cur Loss: 1.21794653, Cur Avg Loss: 1.58766100, Log Avg loss: 1.79481098, Global Avg Loss: 2.51873935, Time: 0.0489 Steps: 9820, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001314, Sample Num: 21024, Cur Loss: 2.22528958, Cur Avg Loss: 1.58855529, Log Avg loss: 1.70517062, Global Avg Loss: 2.51791171, Time: 0.0434 Steps: 9830, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001324, Sample Num: 21184, Cur Loss: 0.64958543, Cur Avg Loss: 1.58660859, Log Avg loss: 1.33081276, Global Avg Loss: 2.51670531, Time: 0.0519 Steps: 9840, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001334, Sample Num: 21344, Cur Loss: 1.08541167, Cur Avg Loss: 1.58641142, Log Avg loss: 1.56030620, Global Avg Loss: 2.51573435, Time: 0.0459 Steps: 9850, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001344, Sample Num: 21504, Cur Loss: 2.13993168, Cur Avg Loss: 1.58974654, Log Avg loss: 2.03465144, Global Avg Loss: 2.51524643, Time: 0.0730 Steps: 9860, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001354, Sample Num: 21664, Cur Loss: 1.54405427, Cur Avg Loss: 1.58897313, Log Avg loss: 1.48502656, Global Avg Loss: 2.51420265, Time: 0.0432 Steps: 9870, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001364, Sample Num: 21824, Cur Loss: 1.41395783, Cur Avg Loss: 1.58786408, Log Avg loss: 1.43769907, Global Avg Loss: 2.51311307, Time: 0.0404 Steps: 9880, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001374, Sample Num: 21984, Cur Loss: 1.84037113, Cur Avg Loss: 1.58736372, Log Avg loss: 1.51911467, Global Avg Loss: 2.51210801, Time: 0.0402 Steps: 9890, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001384, Sample Num: 22144, Cur Loss: 2.45487833, Cur Avg Loss: 1.58734292, Log Avg loss: 1.58448429, Global Avg Loss: 2.51117102, Time: 0.0402 Steps: 9900, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001394, Sample Num: 22304, Cur Loss: 1.43894255, Cur Avg Loss: 1.58818125, Log Avg loss: 1.70420653, Global Avg Loss: 2.51035673, Time: 0.0402 Steps: 9910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001404, Sample Num: 22464, Cur Loss: 2.06664729, Cur Avg Loss: 1.58774382, Log Avg loss: 1.52676668, Global Avg Loss: 2.50936520, Time: 0.0402 Steps: 9920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001414, Sample Num: 22624, Cur Loss: 0.85235929, Cur Avg Loss: 1.58697257, Log Avg loss: 1.47868819, Global Avg Loss: 2.50832726, Time: 0.0402 Steps: 9930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001424, Sample Num: 22784, Cur Loss: 1.88629103, Cur Avg Loss: 1.58742449, Log Avg loss: 1.65132579, Global Avg Loss: 2.50746509, Time: 0.0403 Steps: 9940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001434, Sample Num: 22944, Cur Loss: 0.89043599, Cur Avg Loss: 1.58651571, Log Avg loss: 1.45710617, Global Avg Loss: 2.50640945, Time: 0.0403 Steps: 9950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001444, Sample Num: 23104, Cur Loss: 2.36642814, Cur Avg Loss: 1.58687282, Log Avg loss: 1.63808253, Global Avg Loss: 2.50553764, Time: 0.0402 Steps: 9960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001454, Sample Num: 23264, Cur Loss: 1.60681725, Cur Avg Loss: 1.59105576, Log Avg loss: 2.19507223, Global Avg Loss: 2.50522624, Time: 0.0402 Steps: 9970, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001464, Sample Num: 23424, Cur Loss: 1.73400223, Cur Avg Loss: 1.59215691, Log Avg loss: 1.75226427, Global Avg Loss: 2.50447177, Time: 0.0402 Steps: 9980, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001474, Sample Num: 23584, Cur Loss: 1.72577691, Cur Avg Loss: 1.59235768, Log Avg loss: 1.62174988, Global Avg Loss: 2.50358816, Time: 0.0402 Steps: 9990, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001484, Sample Num: 23744, Cur Loss: 3.19634151, Cur Avg Loss: 1.59471061, Log Avg loss: 1.94153267, Global Avg Loss: 2.50302610, Time: 0.0402 Steps: 10000, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001494, Sample Num: 23904, Cur Loss: 1.47997046, Cur Avg Loss: 1.59355566, Log Avg loss: 1.42216069, Global Avg Loss: 2.50194632, Time: 0.0403 Steps: 10010, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001504, Sample Num: 24064, Cur Loss: 1.73826265, Cur Avg Loss: 1.59597385, Log Avg loss: 1.95725141, Global Avg Loss: 2.50140271, Time: 0.0403 Steps: 10020, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001514, Sample Num: 24224, Cur Loss: 0.86684954, Cur Avg Loss: 1.59641842, Log Avg loss: 1.66328219, Global Avg Loss: 2.50056710, Time: 0.0402 Steps: 10030, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001524, Sample Num: 24384, Cur Loss: 1.80984008, Cur Avg Loss: 1.59739615, Log Avg loss: 1.74542462, Global Avg Loss: 2.49981496, Time: 0.0403 Steps: 10040, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001534, Sample Num: 24544, Cur Loss: 1.76976275, Cur Avg Loss: 1.59659133, Log Avg loss: 1.47393702, Global Avg Loss: 2.49879419, Time: 0.0402 Steps: 10050, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001544, Sample Num: 24704, Cur Loss: 0.75364947, Cur Avg Loss: 1.59561918, Log Avg loss: 1.44649097, Global Avg Loss: 2.49774816, Time: 0.0402 Steps: 10060, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001554, Sample Num: 24864, Cur Loss: 0.78841400, Cur Avg Loss: 1.59747402, Log Avg loss: 1.88386053, Global Avg Loss: 2.49713854, Time: 0.0403 Steps: 10070, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001564, Sample Num: 25024, Cur Loss: 1.19952047, Cur Avg Loss: 1.59766728, Log Avg loss: 1.62769995, Global Avg Loss: 2.49627600, Time: 0.0402 Steps: 10080, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001574, Sample Num: 25184, Cur Loss: 1.84108806, Cur Avg Loss: 1.59810810, Log Avg loss: 1.66705289, Global Avg Loss: 2.49545418, Time: 0.0401 Steps: 10090, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001584, Sample Num: 25344, Cur Loss: 1.26565325, Cur Avg Loss: 1.59704402, Log Avg loss: 1.42955787, Global Avg Loss: 2.49439883, Time: 0.0403 Steps: 10100, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001594, Sample Num: 25504, Cur Loss: 1.82058072, Cur Avg Loss: 1.59739896, Log Avg loss: 1.65362153, Global Avg Loss: 2.49356720, Time: 0.0402 Steps: 10110, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001604, Sample Num: 25664, Cur Loss: 0.98599076, Cur Avg Loss: 1.59852461, Log Avg loss: 1.77795281, Global Avg Loss: 2.49286008, Time: 0.0402 Steps: 10120, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001614, Sample Num: 25824, Cur Loss: 1.69657421, Cur Avg Loss: 1.59794648, Log Avg loss: 1.50521463, Global Avg Loss: 2.49188510, Time: 0.0402 Steps: 10130, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001624, Sample Num: 25984, Cur Loss: 1.89993215, Cur Avg Loss: 1.59738160, Log Avg loss: 1.50621068, Global Avg Loss: 2.49091304, Time: 0.0402 Steps: 10140, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001634, Sample Num: 26144, Cur Loss: 1.33329356, Cur Avg Loss: 1.59982401, Log Avg loss: 1.99647110, Global Avg Loss: 2.49042590, Time: 0.0406 Steps: 10150, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001644, Sample Num: 26304, Cur Loss: 1.49206424, Cur Avg Loss: 1.60013643, Log Avg loss: 1.65118526, Global Avg Loss: 2.48959988, Time: 0.0748 Steps: 10160, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001654, Sample Num: 26464, Cur Loss: 2.38860822, Cur Avg Loss: 1.60050549, Log Avg loss: 1.66117954, Global Avg Loss: 2.48878531, Time: 0.0679 Steps: 10170, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001664, Sample Num: 26624, Cur Loss: 1.41815400, Cur Avg Loss: 1.60011413, Log Avg loss: 1.53538348, Global Avg Loss: 2.48784876, Time: 0.0431 Steps: 10180, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001674, Sample Num: 26784, Cur Loss: 2.47767448, Cur Avg Loss: 1.59969472, Log Avg loss: 1.52990473, Global Avg Loss: 2.48690868, Time: 0.0405 Steps: 10190, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001684, Sample Num: 26944, Cur Loss: 1.86206400, Cur Avg Loss: 1.59959313, Log Avg loss: 1.58258580, Global Avg Loss: 2.48602209, Time: 0.0835 Steps: 10200, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001694, Sample Num: 27104, Cur Loss: 1.03723681, Cur Avg Loss: 1.60086575, Log Avg loss: 1.81517528, Global Avg Loss: 2.48536504, Time: 0.0917 Steps: 10210, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001704, Sample Num: 27264, Cur Loss: 1.61285532, Cur Avg Loss: 1.59991814, Log Avg loss: 1.43939360, Global Avg Loss: 2.48434159, Time: 0.0496 Steps: 10220, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001714, Sample Num: 27424, Cur Loss: 2.70242119, Cur Avg Loss: 1.59989600, Log Avg loss: 1.59612324, Global Avg Loss: 2.48347334, Time: 0.0552 Steps: 10230, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001724, Sample Num: 27584, Cur Loss: 2.09458899, Cur Avg Loss: 1.60047335, Log Avg loss: 1.69943124, Global Avg Loss: 2.48270767, Time: 0.0403 Steps: 10240, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001734, Sample Num: 27744, Cur Loss: 2.18565512, Cur Avg Loss: 1.59967993, Log Avg loss: 1.46289487, Global Avg Loss: 2.48171273, Time: 0.0402 Steps: 10250, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001744, Sample Num: 27904, Cur Loss: 2.35666132, Cur Avg Loss: 1.60211136, Log Avg loss: 2.02372085, Global Avg Loss: 2.48126635, Time: 0.0402 Steps: 10260, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001754, Sample Num: 28064, Cur Loss: 1.74378610, Cur Avg Loss: 1.60259446, Log Avg loss: 1.68684599, Global Avg Loss: 2.48049281, Time: 0.0402 Steps: 10270, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001764, Sample Num: 28224, Cur Loss: 2.26757050, Cur Avg Loss: 1.60225013, Log Avg loss: 1.54185619, Global Avg Loss: 2.47957974, Time: 0.0403 Steps: 10280, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001774, Sample Num: 28384, Cur Loss: 1.18503571, Cur Avg Loss: 1.60426321, Log Avg loss: 1.95936999, Global Avg Loss: 2.47907419, Time: 0.0402 Steps: 10290, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001784, Sample Num: 28544, Cur Loss: 2.01349258, Cur Avg Loss: 1.60547539, Log Avg loss: 1.82051506, Global Avg Loss: 2.47843481, Time: 0.0402 Steps: 10300, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001794, Sample Num: 28704, Cur Loss: 1.94991577, Cur Avg Loss: 1.60492124, Log Avg loss: 1.50606219, Global Avg Loss: 2.47749168, Time: 0.0403 Steps: 10310, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001804, Sample Num: 28864, Cur Loss: 2.25289369, Cur Avg Loss: 1.60566233, Log Avg loss: 1.73861369, Global Avg Loss: 2.47677571, Time: 0.0405 Steps: 10320, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001814, Sample Num: 29024, Cur Loss: 1.37073946, Cur Avg Loss: 1.60595869, Log Avg loss: 1.65942168, Global Avg Loss: 2.47598447, Time: 0.0403 Steps: 10330, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001824, Sample Num: 29184, Cur Loss: 3.49283314, Cur Avg Loss: 1.60602448, Log Avg loss: 1.61795803, Global Avg Loss: 2.47515466, Time: 0.0403 Steps: 10340, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001834, Sample Num: 29344, Cur Loss: 2.06464648, Cur Avg Loss: 1.60794020, Log Avg loss: 1.95736867, Global Avg Loss: 2.47465438, Time: 0.0403 Steps: 10350, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001844, Sample Num: 29504, Cur Loss: 1.94292188, Cur Avg Loss: 1.60858763, Log Avg loss: 1.72732654, Global Avg Loss: 2.47393302, Time: 0.0403 Steps: 10360, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001854, Sample Num: 29664, Cur Loss: 2.62954187, Cur Avg Loss: 1.60881570, Log Avg loss: 1.65087057, Global Avg Loss: 2.47313932, Time: 0.0403 Steps: 10370, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001864, Sample Num: 29824, Cur Loss: 1.35068440, Cur Avg Loss: 1.61141175, Log Avg loss: 2.09271964, Global Avg Loss: 2.47277283, Time: 0.0403 Steps: 10380, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001874, Sample Num: 29984, Cur Loss: 1.30595767, Cur Avg Loss: 1.61132276, Log Avg loss: 1.59473619, Global Avg Loss: 2.47192775, Time: 0.0403 Steps: 10390, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001884, Sample Num: 30144, Cur Loss: 1.67538357, Cur Avg Loss: 1.61222321, Log Avg loss: 1.78096602, Global Avg Loss: 2.47126337, Time: 0.0402 Steps: 10400, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001894, Sample Num: 30304, Cur Loss: 2.64271498, Cur Avg Loss: 1.61202050, Log Avg loss: 1.57383094, Global Avg Loss: 2.47040128, Time: 0.0403 Steps: 10410, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001904, Sample Num: 30464, Cur Loss: 2.69335842, Cur Avg Loss: 1.61226923, Log Avg loss: 1.65937846, Global Avg Loss: 2.46962295, Time: 0.0402 Steps: 10420, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001914, Sample Num: 30624, Cur Loss: 0.86412174, Cur Avg Loss: 1.61264175, Log Avg loss: 1.68356952, Global Avg Loss: 2.46886930, Time: 0.0402 Steps: 10430, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001924, Sample Num: 30784, Cur Loss: 0.43528146, Cur Avg Loss: 1.61242413, Log Avg loss: 1.57077160, Global Avg Loss: 2.46800905, Time: 0.0402 Steps: 10440, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001934, Sample Num: 30944, Cur Loss: 1.45791507, Cur Avg Loss: 1.61105345, Log Avg loss: 1.34733564, Global Avg Loss: 2.46693664, Time: 0.0402 Steps: 10450, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001944, Sample Num: 31104, Cur Loss: 1.88282037, Cur Avg Loss: 1.61163978, Log Avg loss: 1.72503430, Global Avg Loss: 2.46622736, Time: 0.0402 Steps: 10460, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001954, Sample Num: 31264, Cur Loss: 1.07514095, Cur Avg Loss: 1.61140514, Log Avg loss: 1.56579120, Global Avg Loss: 2.46536735, Time: 0.0402 Steps: 10470, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001964, Sample Num: 31424, Cur Loss: 1.44080281, Cur Avg Loss: 1.61180515, Log Avg loss: 1.68996814, Global Avg Loss: 2.46462746, Time: 0.0404 Steps: 10480, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001974, Sample Num: 31584, Cur Loss: 1.04029834, Cur Avg Loss: 1.61089188, Log Avg loss: 1.43152609, Global Avg Loss: 2.46364262, Time: 0.0564 Steps: 10490, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001984, Sample Num: 31744, Cur Loss: 1.48996556, Cur Avg Loss: 1.61124227, Log Avg loss: 1.68040854, Global Avg Loss: 2.46289668, Time: 0.0660 Steps: 10500, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001994, Sample Num: 31904, Cur Loss: 0.51010048, Cur Avg Loss: 1.61116954, Log Avg loss: 1.59674021, Global Avg Loss: 2.46207256, Time: 0.0641 Steps: 10510, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002004, Sample Num: 32064, Cur Loss: 1.51167631, Cur Avg Loss: 1.61133429, Log Avg loss: 1.64418493, Global Avg Loss: 2.46129510, Time: 0.0668 Steps: 10520, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002014, Sample Num: 32224, Cur Loss: 1.66063476, Cur Avg Loss: 1.61255872, Log Avg loss: 1.85793438, Global Avg Loss: 2.46072210, Time: 0.0409 Steps: 10530, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002024, Sample Num: 32384, Cur Loss: 1.18174052, Cur Avg Loss: 1.61188163, Log Avg loss: 1.47551520, Global Avg Loss: 2.45978737, Time: 0.0692 Steps: 10540, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002034, Sample Num: 32544, Cur Loss: 3.27365589, Cur Avg Loss: 1.61249255, Log Avg loss: 1.73614390, Global Avg Loss: 2.45910145, Time: 0.0414 Steps: 10550, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002044, Sample Num: 32704, Cur Loss: 1.47433281, Cur Avg Loss: 1.61198850, Log Avg loss: 1.50946504, Global Avg Loss: 2.45820218, Time: 0.0493 Steps: 10560, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002054, Sample Num: 32864, Cur Loss: 1.50531459, Cur Avg Loss: 1.61133360, Log Avg loss: 1.47747196, Global Avg Loss: 2.45727433, Time: 0.0403 Steps: 10570, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002064, Sample Num: 33024, Cur Loss: 0.58420157, Cur Avg Loss: 1.61186627, Log Avg loss: 1.72127510, Global Avg Loss: 2.45657868, Time: 0.0402 Steps: 10580, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002074, Sample Num: 33184, Cur Loss: 0.79169691, Cur Avg Loss: 1.61124945, Log Avg loss: 1.48393964, Global Avg Loss: 2.45566023, Time: 0.0402 Steps: 10590, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002084, Sample Num: 33344, Cur Loss: 1.51834917, Cur Avg Loss: 1.61102509, Log Avg loss: 1.56449255, Global Avg Loss: 2.45481951, Time: 0.0403 Steps: 10600, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002094, Sample Num: 33504, Cur Loss: 0.93639100, Cur Avg Loss: 1.61009837, Log Avg loss: 1.41696984, Global Avg Loss: 2.45384133, Time: 0.0402 Steps: 10610, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002104, Sample Num: 33664, Cur Loss: 1.97422695, Cur Avg Loss: 1.61078142, Log Avg loss: 1.75381112, Global Avg Loss: 2.45318217, Time: 0.0403 Steps: 10620, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002114, Sample Num: 33824, Cur Loss: 1.37058187, Cur Avg Loss: 1.61197418, Log Avg loss: 1.86293119, Global Avg Loss: 2.45262690, Time: 0.0403 Steps: 10630, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002124, Sample Num: 33984, Cur Loss: 1.04289258, Cur Avg Loss: 1.61181782, Log Avg loss: 1.57876390, Global Avg Loss: 2.45180560, Time: 0.0402 Steps: 10640, Updated lr: 0.000091 ***** Running evaluation checkpoint-10645 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-10645 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.455377, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.445505, "eval_total_loss": 1016.190272, "eval_mae": 1.017726, "eval_mse": 1.445523, "eval_r2": 0.08113, "eval_sp_statistic": 0.269382, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.329787, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.624612, "test_total_loss": 815.555076, "test_mae": 1.156123, "test_mse": 1.624653, "test_r2": -0.048566, "test_sp_statistic": 0.267971, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.328068, "test_ps_pvalue": 0.0, "lr": 9.085348506401138e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.4513770493194498, "train_cur_epoch_loss": 3431.1981957554817, "train_cur_epoch_avg_loss": 1.6116478138823305, "train_cur_epoch_time": 95.45537686347961, "train_cur_epoch_avg_time": 0.04483578058406745, "epoch": 5, "step": 10645} ################################################## Training, Epoch: 0006, Batch: 000005, Sample Num: 80, Cur Loss: 0.86383832, Cur Avg Loss: 1.34871143, Log Avg loss: 1.44406990, Global Avg Loss: 2.45085937, Time: 0.0500 Steps: 10650, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000015, Sample Num: 240, Cur Loss: 0.78990299, Cur Avg Loss: 1.40899106, Log Avg loss: 1.43913088, Global Avg Loss: 2.44991028, Time: 0.0403 Steps: 10660, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000025, Sample Num: 400, Cur Loss: 1.36705041, Cur Avg Loss: 1.37785634, Log Avg loss: 1.33115426, Global Avg Loss: 2.44886177, Time: 0.0402 Steps: 10670, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000035, Sample Num: 560, Cur Loss: 1.94650567, Cur Avg Loss: 1.42154777, Log Avg loss: 1.53077636, Global Avg Loss: 2.44800214, Time: 0.0402 Steps: 10680, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000045, Sample Num: 720, Cur Loss: 2.93315625, Cur Avg Loss: 1.50318080, Log Avg loss: 1.78889639, Global Avg Loss: 2.44738558, Time: 0.0402 Steps: 10690, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000055, Sample Num: 880, Cur Loss: 1.86371422, Cur Avg Loss: 1.49198040, Log Avg loss: 1.44157860, Global Avg Loss: 2.44644557, Time: 0.0402 Steps: 10700, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000065, Sample Num: 1040, Cur Loss: 2.09808969, Cur Avg Loss: 1.51139303, Log Avg loss: 1.61816247, Global Avg Loss: 2.44567220, Time: 0.0402 Steps: 10710, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000075, Sample Num: 1200, Cur Loss: 1.63183224, Cur Avg Loss: 1.51512941, Log Avg loss: 1.53941590, Global Avg Loss: 2.44482681, Time: 0.0402 Steps: 10720, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000085, Sample Num: 1360, Cur Loss: 0.84745336, Cur Avg Loss: 1.54089189, Log Avg loss: 1.73411052, Global Avg Loss: 2.44416445, Time: 0.0403 Steps: 10730, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000095, Sample Num: 1520, Cur Loss: 0.86257333, Cur Avg Loss: 1.54657479, Log Avg loss: 1.59487942, Global Avg Loss: 2.44337368, Time: 0.0402 Steps: 10740, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000105, Sample Num: 1680, Cur Loss: 2.00743079, Cur Avg Loss: 1.55119812, Log Avg loss: 1.59511971, Global Avg Loss: 2.44258460, Time: 0.0402 Steps: 10750, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000115, Sample Num: 1840, Cur Loss: 1.30128813, Cur Avg Loss: 1.51937389, Log Avg loss: 1.18521957, Global Avg Loss: 2.44141605, Time: 0.0403 Steps: 10760, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000125, Sample Num: 2000, Cur Loss: 1.64123893, Cur Avg Loss: 1.53789426, Log Avg loss: 1.75087848, Global Avg Loss: 2.44077488, Time: 0.0403 Steps: 10770, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000135, Sample Num: 2160, Cur Loss: 0.99340403, Cur Avg Loss: 1.55094450, Log Avg loss: 1.71407249, Global Avg Loss: 2.44010076, Time: 0.0402 Steps: 10780, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000145, Sample Num: 2320, Cur Loss: 1.39031792, Cur Avg Loss: 1.56360504, Log Avg loss: 1.73452240, Global Avg Loss: 2.43944684, Time: 0.0402 Steps: 10790, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000155, Sample Num: 2480, Cur Loss: 1.82639098, Cur Avg Loss: 1.58687214, Log Avg loss: 1.92424507, Global Avg Loss: 2.43896980, Time: 0.0402 Steps: 10800, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000165, Sample Num: 2640, Cur Loss: 1.24950337, Cur Avg Loss: 1.57600022, Log Avg loss: 1.40748543, Global Avg Loss: 2.43801561, Time: 0.0403 Steps: 10810, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000175, Sample Num: 2800, Cur Loss: 1.28608191, Cur Avg Loss: 1.58055507, Log Avg loss: 1.65571001, Global Avg Loss: 2.43729259, Time: 0.0402 Steps: 10820, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000185, Sample Num: 2960, Cur Loss: 1.84295750, Cur Avg Loss: 1.58114750, Log Avg loss: 1.59151504, Global Avg Loss: 2.43651163, Time: 0.0402 Steps: 10830, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000195, Sample Num: 3120, Cur Loss: 1.12607801, Cur Avg Loss: 1.56698551, Log Avg loss: 1.30498866, Global Avg Loss: 2.43546779, Time: 0.0402 Steps: 10840, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000205, Sample Num: 3280, Cur Loss: 1.29336512, Cur Avg Loss: 1.57057968, Log Avg loss: 1.64066606, Global Avg Loss: 2.43473526, Time: 0.0402 Steps: 10850, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000215, Sample Num: 3440, Cur Loss: 1.39540100, Cur Avg Loss: 1.57037367, Log Avg loss: 1.56615053, Global Avg Loss: 2.43393545, Time: 0.0402 Steps: 10860, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000225, Sample Num: 3600, Cur Loss: 1.97793329, Cur Avg Loss: 1.57706183, Log Avg loss: 1.72085728, Global Avg Loss: 2.43327945, Time: 0.0402 Steps: 10870, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000235, Sample Num: 3760, Cur Loss: 1.21513462, Cur Avg Loss: 1.57257033, Log Avg loss: 1.47151153, Global Avg Loss: 2.43239547, Time: 0.0402 Steps: 10880, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000245, Sample Num: 3920, Cur Loss: 1.20212579, Cur Avg Loss: 1.57182901, Log Avg loss: 1.55440789, Global Avg Loss: 2.43158924, Time: 0.0402 Steps: 10890, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000255, Sample Num: 4080, Cur Loss: 1.24309087, Cur Avg Loss: 1.58383301, Log Avg loss: 1.87793101, Global Avg Loss: 2.43108129, Time: 0.0402 Steps: 10900, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000265, Sample Num: 4240, Cur Loss: 0.73176086, Cur Avg Loss: 1.59527383, Log Avg loss: 1.88701496, Global Avg Loss: 2.43058261, Time: 0.0579 Steps: 10910, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000275, Sample Num: 4400, Cur Loss: 1.41192758, Cur Avg Loss: 1.59395281, Log Avg loss: 1.55894555, Global Avg Loss: 2.42978441, Time: 0.0545 Steps: 10920, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000285, Sample Num: 4560, Cur Loss: 0.84953922, Cur Avg Loss: 1.61534063, Log Avg loss: 2.20350565, Global Avg Loss: 2.42957738, Time: 0.0413 Steps: 10930, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000295, Sample Num: 4720, Cur Loss: 2.68436813, Cur Avg Loss: 1.61523015, Log Avg loss: 1.61208150, Global Avg Loss: 2.42883013, Time: 0.0406 Steps: 10940, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000305, Sample Num: 4880, Cur Loss: 1.07430482, Cur Avg Loss: 1.61456243, Log Avg loss: 1.59486468, Global Avg Loss: 2.42806851, Time: 0.0662 Steps: 10950, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000315, Sample Num: 5040, Cur Loss: 1.57900715, Cur Avg Loss: 1.61517275, Log Avg loss: 1.63378772, Global Avg Loss: 2.42734381, Time: 0.0441 Steps: 10960, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000325, Sample Num: 5200, Cur Loss: 1.75430799, Cur Avg Loss: 1.61059968, Log Avg loss: 1.46654788, Global Avg Loss: 2.42646797, Time: 0.0674 Steps: 10970, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000335, Sample Num: 5360, Cur Loss: 1.11724973, Cur Avg Loss: 1.60378989, Log Avg loss: 1.38247162, Global Avg Loss: 2.42551715, Time: 0.0625 Steps: 10980, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000345, Sample Num: 5520, Cur Loss: 1.41755176, Cur Avg Loss: 1.60781385, Log Avg loss: 1.74261659, Global Avg Loss: 2.42489577, Time: 0.0405 Steps: 10990, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000355, Sample Num: 5680, Cur Loss: 0.83011013, Cur Avg Loss: 1.59486738, Log Avg loss: 1.14821420, Global Avg Loss: 2.42373515, Time: 0.0402 Steps: 11000, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000365, Sample Num: 5840, Cur Loss: 1.95086873, Cur Avg Loss: 1.59709059, Log Avg loss: 1.67601468, Global Avg Loss: 2.42305602, Time: 0.0403 Steps: 11010, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000375, Sample Num: 6000, Cur Loss: 1.25586843, Cur Avg Loss: 1.60255124, Log Avg loss: 1.80186462, Global Avg Loss: 2.42249232, Time: 0.0402 Steps: 11020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000385, Sample Num: 6160, Cur Loss: 1.32507539, Cur Avg Loss: 1.60627139, Log Avg loss: 1.74577724, Global Avg Loss: 2.42187880, Time: 0.0403 Steps: 11030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000395, Sample Num: 6320, Cur Loss: 1.53159440, Cur Avg Loss: 1.60579578, Log Avg loss: 1.58748457, Global Avg Loss: 2.42112301, Time: 0.0402 Steps: 11040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000405, Sample Num: 6480, Cur Loss: 1.23341227, Cur Avg Loss: 1.59820418, Log Avg loss: 1.29833624, Global Avg Loss: 2.42010691, Time: 0.0402 Steps: 11050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000415, Sample Num: 6640, Cur Loss: 1.27657151, Cur Avg Loss: 1.59559691, Log Avg loss: 1.49000224, Global Avg Loss: 2.41926595, Time: 0.0402 Steps: 11060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000425, Sample Num: 6800, Cur Loss: 0.87064457, Cur Avg Loss: 1.59850101, Log Avg loss: 1.71902119, Global Avg Loss: 2.41863339, Time: 0.0403 Steps: 11070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000435, Sample Num: 6960, Cur Loss: 1.72930908, Cur Avg Loss: 1.60204270, Log Avg loss: 1.75256442, Global Avg Loss: 2.41803224, Time: 0.0402 Steps: 11080, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000445, Sample Num: 7120, Cur Loss: 0.94036829, Cur Avg Loss: 1.59538933, Log Avg loss: 1.30596790, Global Avg Loss: 2.41702948, Time: 0.0402 Steps: 11090, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000455, Sample Num: 7280, Cur Loss: 1.37217677, Cur Avg Loss: 1.59201859, Log Avg loss: 1.44202088, Global Avg Loss: 2.41615109, Time: 0.0402 Steps: 11100, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000465, Sample Num: 7440, Cur Loss: 1.02901149, Cur Avg Loss: 1.58832029, Log Avg loss: 1.42004747, Global Avg Loss: 2.41525451, Time: 0.0402 Steps: 11110, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000475, Sample Num: 7600, Cur Loss: 1.84816229, Cur Avg Loss: 1.58959849, Log Avg loss: 1.64903495, Global Avg Loss: 2.41456547, Time: 0.0402 Steps: 11120, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000485, Sample Num: 7760, Cur Loss: 1.86636162, Cur Avg Loss: 1.58955911, Log Avg loss: 1.58768855, Global Avg Loss: 2.41382254, Time: 0.0403 Steps: 11130, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000495, Sample Num: 7920, Cur Loss: 1.33267546, Cur Avg Loss: 1.59761864, Log Avg loss: 1.98850591, Global Avg Loss: 2.41344075, Time: 0.0402 Steps: 11140, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000505, Sample Num: 8080, Cur Loss: 2.37817240, Cur Avg Loss: 1.60142882, Log Avg loss: 1.79003227, Global Avg Loss: 2.41288164, Time: 0.0403 Steps: 11150, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000515, Sample Num: 8240, Cur Loss: 1.96034884, Cur Avg Loss: 1.60203650, Log Avg loss: 1.63272481, Global Avg Loss: 2.41218257, Time: 0.0403 Steps: 11160, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000525, Sample Num: 8400, Cur Loss: 3.05343914, Cur Avg Loss: 1.61146227, Log Avg loss: 2.09688910, Global Avg Loss: 2.41190030, Time: 0.0402 Steps: 11170, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000535, Sample Num: 8560, Cur Loss: 1.56119514, Cur Avg Loss: 1.60639286, Log Avg loss: 1.34024909, Global Avg Loss: 2.41094176, Time: 0.0403 Steps: 11180, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000545, Sample Num: 8720, Cur Loss: 1.86570954, Cur Avg Loss: 1.60033371, Log Avg loss: 1.27616879, Global Avg Loss: 2.40992766, Time: 0.0403 Steps: 11190, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000555, Sample Num: 8880, Cur Loss: 1.09621310, Cur Avg Loss: 1.59702994, Log Avg loss: 1.41697456, Global Avg Loss: 2.40904110, Time: 0.0402 Steps: 11200, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000565, Sample Num: 9040, Cur Loss: 1.01451516, Cur Avg Loss: 1.59226648, Log Avg loss: 1.32789463, Global Avg Loss: 2.40807665, Time: 0.0402 Steps: 11210, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000575, Sample Num: 9200, Cur Loss: 0.40361875, Cur Avg Loss: 1.58002946, Log Avg loss: 0.88863775, Global Avg Loss: 2.40672243, Time: 0.0403 Steps: 11220, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000585, Sample Num: 9360, Cur Loss: 1.86061299, Cur Avg Loss: 1.58051715, Log Avg loss: 1.60855955, Global Avg Loss: 2.40601169, Time: 0.0402 Steps: 11230, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000595, Sample Num: 9520, Cur Loss: 1.91014409, Cur Avg Loss: 1.57784025, Log Avg loss: 1.42124145, Global Avg Loss: 2.40513556, Time: 0.0402 Steps: 11240, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000605, Sample Num: 9680, Cur Loss: 1.21162546, Cur Avg Loss: 1.58034305, Log Avg loss: 1.72925987, Global Avg Loss: 2.40453478, Time: 0.0402 Steps: 11250, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000615, Sample Num: 9840, Cur Loss: 0.74362993, Cur Avg Loss: 1.58081373, Log Avg loss: 1.60928949, Global Avg Loss: 2.40382852, Time: 0.0403 Steps: 11260, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000625, Sample Num: 10000, Cur Loss: 1.52727938, Cur Avg Loss: 1.58304819, Log Avg loss: 1.72046735, Global Avg Loss: 2.40322217, Time: 0.0726 Steps: 11270, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000635, Sample Num: 10160, Cur Loss: 1.24956584, Cur Avg Loss: 1.57997500, Log Avg loss: 1.38790118, Global Avg Loss: 2.40232206, Time: 0.0763 Steps: 11280, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000645, Sample Num: 10320, Cur Loss: 4.45071507, Cur Avg Loss: 1.58618747, Log Avg loss: 1.98067915, Global Avg Loss: 2.40194859, Time: 0.0831 Steps: 11290, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000655, Sample Num: 10480, Cur Loss: 1.20140409, Cur Avg Loss: 1.58636258, Log Avg loss: 1.59765673, Global Avg Loss: 2.40123683, Time: 0.0404 Steps: 11300, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000665, Sample Num: 10640, Cur Loss: 0.83338529, Cur Avg Loss: 1.58467620, Log Avg loss: 1.47421853, Global Avg Loss: 2.40041718, Time: 0.0523 Steps: 11310, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000675, Sample Num: 10800, Cur Loss: 0.96059418, Cur Avg Loss: 1.58195629, Log Avg loss: 1.40108217, Global Avg Loss: 2.39953438, Time: 0.0429 Steps: 11320, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000685, Sample Num: 10960, Cur Loss: 1.30193567, Cur Avg Loss: 1.57639504, Log Avg loss: 1.20101108, Global Avg Loss: 2.39847655, Time: 0.0551 Steps: 11330, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000695, Sample Num: 11120, Cur Loss: 1.59297931, Cur Avg Loss: 1.57109220, Log Avg loss: 1.20784736, Global Avg Loss: 2.39742661, Time: 0.0407 Steps: 11340, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000705, Sample Num: 11280, Cur Loss: 1.98315549, Cur Avg Loss: 1.57377807, Log Avg loss: 1.76044598, Global Avg Loss: 2.39686539, Time: 0.0403 Steps: 11350, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000715, Sample Num: 11440, Cur Loss: 1.83876967, Cur Avg Loss: 1.57709252, Log Avg loss: 1.81076134, Global Avg Loss: 2.39634946, Time: 0.0403 Steps: 11360, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000725, Sample Num: 11600, Cur Loss: 1.78840399, Cur Avg Loss: 1.57381889, Log Avg loss: 1.33975425, Global Avg Loss: 2.39542017, Time: 0.0402 Steps: 11370, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000735, Sample Num: 11760, Cur Loss: 2.85513115, Cur Avg Loss: 1.57201158, Log Avg loss: 1.44098165, Global Avg Loss: 2.39458148, Time: 0.0402 Steps: 11380, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000745, Sample Num: 11920, Cur Loss: 1.97309399, Cur Avg Loss: 1.57122622, Log Avg loss: 1.51350231, Global Avg Loss: 2.39380792, Time: 0.0402 Steps: 11390, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000755, Sample Num: 12080, Cur Loss: 0.97652113, Cur Avg Loss: 1.56935645, Log Avg loss: 1.43005828, Global Avg Loss: 2.39296253, Time: 0.0402 Steps: 11400, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000765, Sample Num: 12240, Cur Loss: 0.51995677, Cur Avg Loss: 1.56875848, Log Avg loss: 1.52361192, Global Avg Loss: 2.39220061, Time: 0.0402 Steps: 11410, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000775, Sample Num: 12400, Cur Loss: 0.74814773, Cur Avg Loss: 1.56442255, Log Avg loss: 1.23272424, Global Avg Loss: 2.39118530, Time: 0.0403 Steps: 11420, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000785, Sample Num: 12560, Cur Loss: 1.99024367, Cur Avg Loss: 1.57089733, Log Avg loss: 2.07269243, Global Avg Loss: 2.39090666, Time: 0.0403 Steps: 11430, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000795, Sample Num: 12720, Cur Loss: 1.24547780, Cur Avg Loss: 1.56779250, Log Avg loss: 1.32406361, Global Avg Loss: 2.38997410, Time: 0.0402 Steps: 11440, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000805, Sample Num: 12880, Cur Loss: 2.19617128, Cur Avg Loss: 1.57585606, Log Avg loss: 2.21690867, Global Avg Loss: 2.38982295, Time: 0.0403 Steps: 11450, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000815, Sample Num: 13040, Cur Loss: 2.10909176, Cur Avg Loss: 1.57598660, Log Avg loss: 1.58649551, Global Avg Loss: 2.38912197, Time: 0.0403 Steps: 11460, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000825, Sample Num: 13200, Cur Loss: 1.88007641, Cur Avg Loss: 1.57421765, Log Avg loss: 1.43004801, Global Avg Loss: 2.38828581, Time: 0.0403 Steps: 11470, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000835, Sample Num: 13360, Cur Loss: 3.36091185, Cur Avg Loss: 1.57494475, Log Avg loss: 1.63493090, Global Avg Loss: 2.38762958, Time: 0.0402 Steps: 11480, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000845, Sample Num: 13520, Cur Loss: 1.34466326, Cur Avg Loss: 1.57510526, Log Avg loss: 1.58850793, Global Avg Loss: 2.38693409, Time: 0.0402 Steps: 11490, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000855, Sample Num: 13680, Cur Loss: 1.61391640, Cur Avg Loss: 1.57419561, Log Avg loss: 1.49732952, Global Avg Loss: 2.38616052, Time: 0.0402 Steps: 11500, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000865, Sample Num: 13840, Cur Loss: 0.90423596, Cur Avg Loss: 1.57557340, Log Avg loss: 1.69337485, Global Avg Loss: 2.38555862, Time: 0.0403 Steps: 11510, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000875, Sample Num: 14000, Cur Loss: 1.59625375, Cur Avg Loss: 1.57884628, Log Avg loss: 1.86195042, Global Avg Loss: 2.38510410, Time: 0.0402 Steps: 11520, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000885, Sample Num: 14160, Cur Loss: 1.06445527, Cur Avg Loss: 1.57693118, Log Avg loss: 1.40935967, Global Avg Loss: 2.38425783, Time: 0.0403 Steps: 11530, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000895, Sample Num: 14320, Cur Loss: 1.44363880, Cur Avg Loss: 1.57589553, Log Avg loss: 1.48424079, Global Avg Loss: 2.38347792, Time: 0.0403 Steps: 11540, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000905, Sample Num: 14480, Cur Loss: 1.82769251, Cur Avg Loss: 1.57819418, Log Avg loss: 1.78392282, Global Avg Loss: 2.38295882, Time: 0.0402 Steps: 11550, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000915, Sample Num: 14640, Cur Loss: 1.90829504, Cur Avg Loss: 1.58022262, Log Avg loss: 1.76379672, Global Avg Loss: 2.38242322, Time: 0.0402 Steps: 11560, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000925, Sample Num: 14800, Cur Loss: 2.61356878, Cur Avg Loss: 1.58315479, Log Avg loss: 1.85144802, Global Avg Loss: 2.38196429, Time: 0.0402 Steps: 11570, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000935, Sample Num: 14960, Cur Loss: 3.17488289, Cur Avg Loss: 1.58153441, Log Avg loss: 1.43164970, Global Avg Loss: 2.38114364, Time: 0.0402 Steps: 11580, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000945, Sample Num: 15120, Cur Loss: 0.53261721, Cur Avg Loss: 1.58134283, Log Avg loss: 1.56343036, Global Avg Loss: 2.38043811, Time: 0.0402 Steps: 11590, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000955, Sample Num: 15280, Cur Loss: 2.84590125, Cur Avg Loss: 1.58058716, Log Avg loss: 1.50917601, Global Avg Loss: 2.37968702, Time: 0.0403 Steps: 11600, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000965, Sample Num: 15440, Cur Loss: 1.26908422, Cur Avg Loss: 1.57858507, Log Avg loss: 1.38738528, Global Avg Loss: 2.37883232, Time: 0.0417 Steps: 11610, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000975, Sample Num: 15600, Cur Loss: 0.82066655, Cur Avg Loss: 1.58219889, Log Avg loss: 1.93093312, Global Avg Loss: 2.37844687, Time: 0.0566 Steps: 11620, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000985, Sample Num: 15760, Cur Loss: 2.12388659, Cur Avg Loss: 1.58706845, Log Avg loss: 2.06184998, Global Avg Loss: 2.37817464, Time: 0.0504 Steps: 11630, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000995, Sample Num: 15920, Cur Loss: 0.94604456, Cur Avg Loss: 1.58411364, Log Avg loss: 1.29306485, Global Avg Loss: 2.37724242, Time: 0.0450 Steps: 11640, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001005, Sample Num: 16080, Cur Loss: 2.14975214, Cur Avg Loss: 1.58803754, Log Avg loss: 1.97846559, Global Avg Loss: 2.37690012, Time: 0.0734 Steps: 11650, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001015, Sample Num: 16240, Cur Loss: 1.37513876, Cur Avg Loss: 1.59020388, Log Avg loss: 1.80792102, Global Avg Loss: 2.37641215, Time: 0.0602 Steps: 11660, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001025, Sample Num: 16400, Cur Loss: 1.78793764, Cur Avg Loss: 1.58901407, Log Avg loss: 1.46824864, Global Avg Loss: 2.37563394, Time: 0.0557 Steps: 11670, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001035, Sample Num: 16560, Cur Loss: 1.25538993, Cur Avg Loss: 1.58609562, Log Avg loss: 1.28695405, Global Avg Loss: 2.37470185, Time: 0.0806 Steps: 11680, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001045, Sample Num: 16720, Cur Loss: 0.84909499, Cur Avg Loss: 1.58943918, Log Avg loss: 1.93549790, Global Avg Loss: 2.37432614, Time: 0.0404 Steps: 11690, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001055, Sample Num: 16880, Cur Loss: 1.41417909, Cur Avg Loss: 1.58766830, Log Avg loss: 1.40261196, Global Avg Loss: 2.37349562, Time: 0.0405 Steps: 11700, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001065, Sample Num: 17040, Cur Loss: 1.31344724, Cur Avg Loss: 1.58827080, Log Avg loss: 1.65183451, Global Avg Loss: 2.37287934, Time: 0.0402 Steps: 11710, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001075, Sample Num: 17200, Cur Loss: 1.55569839, Cur Avg Loss: 1.58732776, Log Avg loss: 1.48689361, Global Avg Loss: 2.37212338, Time: 0.0402 Steps: 11720, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001085, Sample Num: 17360, Cur Loss: 2.19304609, Cur Avg Loss: 1.58756815, Log Avg loss: 1.61341046, Global Avg Loss: 2.37147657, Time: 0.0403 Steps: 11730, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001095, Sample Num: 17520, Cur Loss: 2.53306198, Cur Avg Loss: 1.58894417, Log Avg loss: 1.73824220, Global Avg Loss: 2.37093719, Time: 0.0402 Steps: 11740, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001105, Sample Num: 17680, Cur Loss: 1.23209238, Cur Avg Loss: 1.58809394, Log Avg loss: 1.49499308, Global Avg Loss: 2.37019170, Time: 0.0403 Steps: 11750, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001115, Sample Num: 17840, Cur Loss: 2.50471830, Cur Avg Loss: 1.58752099, Log Avg loss: 1.52420999, Global Avg Loss: 2.36947233, Time: 0.0403 Steps: 11760, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001125, Sample Num: 18000, Cur Loss: 2.21784067, Cur Avg Loss: 1.58793401, Log Avg loss: 1.63398562, Global Avg Loss: 2.36884745, Time: 0.0403 Steps: 11770, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001135, Sample Num: 18160, Cur Loss: 0.89587086, Cur Avg Loss: 1.58434948, Log Avg loss: 1.18109041, Global Avg Loss: 2.36783916, Time: 0.0403 Steps: 11780, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001145, Sample Num: 18320, Cur Loss: 1.56913960, Cur Avg Loss: 1.58716355, Log Avg loss: 1.90656070, Global Avg Loss: 2.36744792, Time: 0.0404 Steps: 11790, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001155, Sample Num: 18480, Cur Loss: 2.58972788, Cur Avg Loss: 1.58884953, Log Avg loss: 1.78189419, Global Avg Loss: 2.36695169, Time: 0.0403 Steps: 11800, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001165, Sample Num: 18640, Cur Loss: 2.87161970, Cur Avg Loss: 1.58970869, Log Avg loss: 1.68894097, Global Avg Loss: 2.36637759, Time: 0.0402 Steps: 11810, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001175, Sample Num: 18800, Cur Loss: 1.51932776, Cur Avg Loss: 1.58861071, Log Avg loss: 1.46069652, Global Avg Loss: 2.36561136, Time: 0.0402 Steps: 11820, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001185, Sample Num: 18960, Cur Loss: 0.99272674, Cur Avg Loss: 1.58896247, Log Avg loss: 1.63029485, Global Avg Loss: 2.36498979, Time: 0.0403 Steps: 11830, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001195, Sample Num: 19120, Cur Loss: 0.79057670, Cur Avg Loss: 1.58784464, Log Avg loss: 1.45538110, Global Avg Loss: 2.36422154, Time: 0.0403 Steps: 11840, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001205, Sample Num: 19280, Cur Loss: 1.22294486, Cur Avg Loss: 1.58686673, Log Avg loss: 1.47000715, Global Avg Loss: 2.36346693, Time: 0.0403 Steps: 11850, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001215, Sample Num: 19440, Cur Loss: 0.55529267, Cur Avg Loss: 1.58490216, Log Avg loss: 1.34817094, Global Avg Loss: 2.36261086, Time: 0.0402 Steps: 11860, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001225, Sample Num: 19600, Cur Loss: 0.88455993, Cur Avg Loss: 1.58506800, Log Avg loss: 1.60521757, Global Avg Loss: 2.36197279, Time: 0.0402 Steps: 11870, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001235, Sample Num: 19760, Cur Loss: 1.06835747, Cur Avg Loss: 1.58425160, Log Avg loss: 1.48424208, Global Avg Loss: 2.36123396, Time: 0.0403 Steps: 11880, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001245, Sample Num: 19920, Cur Loss: 1.49984264, Cur Avg Loss: 1.58359583, Log Avg loss: 1.50260832, Global Avg Loss: 2.36051182, Time: 0.0402 Steps: 11890, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001255, Sample Num: 20080, Cur Loss: 1.35141647, Cur Avg Loss: 1.58443992, Log Avg loss: 1.68952950, Global Avg Loss: 2.35994797, Time: 0.0402 Steps: 11900, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001265, Sample Num: 20240, Cur Loss: 2.06077409, Cur Avg Loss: 1.58013640, Log Avg loss: 1.04004532, Global Avg Loss: 2.35883973, Time: 0.0402 Steps: 11910, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001275, Sample Num: 20400, Cur Loss: 2.83754849, Cur Avg Loss: 1.58048334, Log Avg loss: 1.62437093, Global Avg Loss: 2.35822357, Time: 0.0402 Steps: 11920, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001285, Sample Num: 20560, Cur Loss: 1.02522314, Cur Avg Loss: 1.57897432, Log Avg loss: 1.38657349, Global Avg Loss: 2.35740911, Time: 0.0402 Steps: 11930, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001295, Sample Num: 20720, Cur Loss: 2.08462429, Cur Avg Loss: 1.57860843, Log Avg loss: 1.53159242, Global Avg Loss: 2.35671747, Time: 0.0402 Steps: 11940, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001305, Sample Num: 20880, Cur Loss: 1.84698749, Cur Avg Loss: 1.57895477, Log Avg loss: 1.62380538, Global Avg Loss: 2.35610416, Time: 0.0402 Steps: 11950, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001315, Sample Num: 21040, Cur Loss: 1.31128788, Cur Avg Loss: 1.57973016, Log Avg loss: 1.68091909, Global Avg Loss: 2.35553962, Time: 0.0465 Steps: 11960, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001325, Sample Num: 21200, Cur Loss: 1.54280210, Cur Avg Loss: 1.57856345, Log Avg loss: 1.42514079, Global Avg Loss: 2.35476234, Time: 0.0612 Steps: 11970, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001335, Sample Num: 21360, Cur Loss: 1.47754121, Cur Avg Loss: 1.57929335, Log Avg loss: 1.67600515, Global Avg Loss: 2.35419577, Time: 0.0498 Steps: 11980, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001345, Sample Num: 21520, Cur Loss: 3.16758513, Cur Avg Loss: 1.57913113, Log Avg loss: 1.55747489, Global Avg Loss: 2.35353128, Time: 0.0829 Steps: 11990, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001355, Sample Num: 21680, Cur Loss: 1.80041075, Cur Avg Loss: 1.57699445, Log Avg loss: 1.28961058, Global Avg Loss: 2.35264468, Time: 0.0575 Steps: 12000, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001365, Sample Num: 21840, Cur Loss: 1.58131945, Cur Avg Loss: 1.57661759, Log Avg loss: 1.52555263, Global Avg Loss: 2.35195601, Time: 0.0605 Steps: 12010, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001375, Sample Num: 22000, Cur Loss: 1.34873426, Cur Avg Loss: 1.57647052, Log Avg loss: 1.55639569, Global Avg Loss: 2.35129415, Time: 0.0683 Steps: 12020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001385, Sample Num: 22160, Cur Loss: 1.62355113, Cur Avg Loss: 1.57689189, Log Avg loss: 1.63483096, Global Avg Loss: 2.35069858, Time: 0.0687 Steps: 12030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001395, Sample Num: 22320, Cur Loss: 1.59128475, Cur Avg Loss: 1.57664287, Log Avg loss: 1.54215341, Global Avg Loss: 2.35002703, Time: 0.0403 Steps: 12040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001405, Sample Num: 22480, Cur Loss: 1.17233801, Cur Avg Loss: 1.57519914, Log Avg loss: 1.37379913, Global Avg Loss: 2.34921689, Time: 0.0403 Steps: 12050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001415, Sample Num: 22640, Cur Loss: 1.13543010, Cur Avg Loss: 1.57449178, Log Avg loss: 1.47510680, Global Avg Loss: 2.34849209, Time: 0.0404 Steps: 12060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001425, Sample Num: 22800, Cur Loss: 1.24417746, Cur Avg Loss: 1.57343223, Log Avg loss: 1.42350645, Global Avg Loss: 2.34772573, Time: 0.0402 Steps: 12070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001435, Sample Num: 22960, Cur Loss: 1.14426279, Cur Avg Loss: 1.57311129, Log Avg loss: 1.52737678, Global Avg Loss: 2.34704664, Time: 0.0402 Steps: 12080, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001445, Sample Num: 23120, Cur Loss: 1.32903838, Cur Avg Loss: 1.57525724, Log Avg loss: 1.88320156, Global Avg Loss: 2.34666298, Time: 0.0403 Steps: 12090, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001455, Sample Num: 23280, Cur Loss: 2.92306757, Cur Avg Loss: 1.57903181, Log Avg loss: 2.12445756, Global Avg Loss: 2.34647934, Time: 0.0402 Steps: 12100, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001465, Sample Num: 23440, Cur Loss: 2.55305099, Cur Avg Loss: 1.58296520, Log Avg loss: 2.15527346, Global Avg Loss: 2.34632145, Time: 0.0402 Steps: 12110, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001475, Sample Num: 23600, Cur Loss: 2.04583621, Cur Avg Loss: 1.58376571, Log Avg loss: 1.70103940, Global Avg Loss: 2.34578904, Time: 0.0402 Steps: 12120, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001485, Sample Num: 23760, Cur Loss: 3.18387485, Cur Avg Loss: 1.58404358, Log Avg loss: 1.62503057, Global Avg Loss: 2.34519484, Time: 0.0402 Steps: 12130, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001495, Sample Num: 23920, Cur Loss: 2.06977963, Cur Avg Loss: 1.58516321, Log Avg loss: 1.75142826, Global Avg Loss: 2.34470574, Time: 0.0402 Steps: 12140, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001505, Sample Num: 24080, Cur Loss: 1.43104792, Cur Avg Loss: 1.58733876, Log Avg loss: 1.91258305, Global Avg Loss: 2.34435008, Time: 0.0402 Steps: 12150, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001515, Sample Num: 24240, Cur Loss: 0.79200155, Cur Avg Loss: 1.58722759, Log Avg loss: 1.57049624, Global Avg Loss: 2.34371369, Time: 0.0403 Steps: 12160, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001525, Sample Num: 24400, Cur Loss: 2.11614037, Cur Avg Loss: 1.58758260, Log Avg loss: 1.64136677, Global Avg Loss: 2.34313658, Time: 0.0403 Steps: 12170, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001535, Sample Num: 24560, Cur Loss: 1.30209315, Cur Avg Loss: 1.58390030, Log Avg loss: 1.02235016, Global Avg Loss: 2.34205219, Time: 0.0402 Steps: 12180, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001545, Sample Num: 24720, Cur Loss: 1.64669275, Cur Avg Loss: 1.58340533, Log Avg loss: 1.50742716, Global Avg Loss: 2.34136751, Time: 0.0402 Steps: 12190, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001555, Sample Num: 24880, Cur Loss: 1.49979830, Cur Avg Loss: 1.58286992, Log Avg loss: 1.50014793, Global Avg Loss: 2.34067798, Time: 0.0402 Steps: 12200, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001565, Sample Num: 25040, Cur Loss: 0.32379383, Cur Avg Loss: 1.58108515, Log Avg loss: 1.30355430, Global Avg Loss: 2.33982858, Time: 0.0402 Steps: 12210, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001575, Sample Num: 25200, Cur Loss: 1.38720191, Cur Avg Loss: 1.57991864, Log Avg loss: 1.39735952, Global Avg Loss: 2.33905733, Time: 0.0402 Steps: 12220, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001585, Sample Num: 25360, Cur Loss: 1.04942727, Cur Avg Loss: 1.57825436, Log Avg loss: 1.31613098, Global Avg Loss: 2.33822092, Time: 0.0402 Steps: 12230, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001595, Sample Num: 25520, Cur Loss: 0.92690098, Cur Avg Loss: 1.57723045, Log Avg loss: 1.41494061, Global Avg Loss: 2.33746661, Time: 0.0402 Steps: 12240, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001605, Sample Num: 25680, Cur Loss: 1.76150012, Cur Avg Loss: 1.57928060, Log Avg loss: 1.90627910, Global Avg Loss: 2.33711462, Time: 0.0402 Steps: 12250, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001615, Sample Num: 25840, Cur Loss: 2.47804689, Cur Avg Loss: 1.57884065, Log Avg loss: 1.50822836, Global Avg Loss: 2.33643853, Time: 0.0402 Steps: 12260, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001625, Sample Num: 26000, Cur Loss: 1.81506026, Cur Avg Loss: 1.57906956, Log Avg loss: 1.61603805, Global Avg Loss: 2.33585140, Time: 0.0404 Steps: 12270, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001635, Sample Num: 26160, Cur Loss: 1.77407122, Cur Avg Loss: 1.57921305, Log Avg loss: 1.60253133, Global Avg Loss: 2.33525424, Time: 0.0402 Steps: 12280, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001645, Sample Num: 26320, Cur Loss: 2.02878618, Cur Avg Loss: 1.57642473, Log Avg loss: 1.12053461, Global Avg Loss: 2.33426586, Time: 0.0402 Steps: 12290, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001655, Sample Num: 26480, Cur Loss: 2.08661389, Cur Avg Loss: 1.57841039, Log Avg loss: 1.90505054, Global Avg Loss: 2.33391690, Time: 0.0458 Steps: 12300, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001665, Sample Num: 26640, Cur Loss: 2.91968536, Cur Avg Loss: 1.57935687, Log Avg loss: 1.73599917, Global Avg Loss: 2.33343118, Time: 0.0408 Steps: 12310, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001675, Sample Num: 26800, Cur Loss: 1.94989264, Cur Avg Loss: 1.58129279, Log Avg loss: 1.90362371, Global Avg Loss: 2.33308231, Time: 0.0637 Steps: 12320, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001685, Sample Num: 26960, Cur Loss: 1.50442839, Cur Avg Loss: 1.58182692, Log Avg loss: 1.67129465, Global Avg Loss: 2.33254558, Time: 0.0471 Steps: 12330, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001695, Sample Num: 27120, Cur Loss: 1.36469436, Cur Avg Loss: 1.58370682, Log Avg loss: 1.90046834, Global Avg Loss: 2.33219544, Time: 0.0631 Steps: 12340, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001705, Sample Num: 27280, Cur Loss: 1.18755698, Cur Avg Loss: 1.58384244, Log Avg loss: 1.60683031, Global Avg Loss: 2.33160810, Time: 0.0457 Steps: 12350, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001715, Sample Num: 27440, Cur Loss: 1.68261433, Cur Avg Loss: 1.58379529, Log Avg loss: 1.57575598, Global Avg Loss: 2.33099657, Time: 0.0623 Steps: 12360, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001725, Sample Num: 27600, Cur Loss: 1.76834512, Cur Avg Loss: 1.58567634, Log Avg loss: 1.90827688, Global Avg Loss: 2.33065484, Time: 0.0406 Steps: 12370, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001735, Sample Num: 27760, Cur Loss: 2.17211652, Cur Avg Loss: 1.58634481, Log Avg loss: 1.70165680, Global Avg Loss: 2.33014676, Time: 0.0657 Steps: 12380, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001745, Sample Num: 27920, Cur Loss: 1.23724437, Cur Avg Loss: 1.58725605, Log Avg loss: 1.74535506, Global Avg Loss: 2.32967478, Time: 0.0402 Steps: 12390, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001755, Sample Num: 28080, Cur Loss: 0.59818268, Cur Avg Loss: 1.58567200, Log Avg loss: 1.30925589, Global Avg Loss: 2.32885186, Time: 0.0404 Steps: 12400, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001765, Sample Num: 28240, Cur Loss: 1.42607987, Cur Avg Loss: 1.58600481, Log Avg loss: 1.64441274, Global Avg Loss: 2.32830034, Time: 0.0404 Steps: 12410, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001775, Sample Num: 28400, Cur Loss: 1.10151839, Cur Avg Loss: 1.58460484, Log Avg loss: 1.33751047, Global Avg Loss: 2.32750260, Time: 0.0403 Steps: 12420, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001785, Sample Num: 28560, Cur Loss: 1.06125140, Cur Avg Loss: 1.58285272, Log Avg loss: 1.27185179, Global Avg Loss: 2.32665332, Time: 0.0402 Steps: 12430, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001795, Sample Num: 28720, Cur Loss: 1.51004195, Cur Avg Loss: 1.58412560, Log Avg loss: 1.81133398, Global Avg Loss: 2.32623908, Time: 0.0402 Steps: 12440, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001805, Sample Num: 28880, Cur Loss: 1.24916172, Cur Avg Loss: 1.58533438, Log Avg loss: 1.80230987, Global Avg Loss: 2.32581825, Time: 0.0403 Steps: 12450, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001815, Sample Num: 29040, Cur Loss: 0.98809630, Cur Avg Loss: 1.58538975, Log Avg loss: 1.59538501, Global Avg Loss: 2.32523203, Time: 0.0403 Steps: 12460, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001825, Sample Num: 29200, Cur Loss: 1.21955526, Cur Avg Loss: 1.58438728, Log Avg loss: 1.40243818, Global Avg Loss: 2.32449202, Time: 0.0403 Steps: 12470, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001835, Sample Num: 29360, Cur Loss: 0.64362311, Cur Avg Loss: 1.58542113, Log Avg loss: 1.77409845, Global Avg Loss: 2.32405100, Time: 0.0403 Steps: 12480, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001845, Sample Num: 29520, Cur Loss: 1.50650096, Cur Avg Loss: 1.58592969, Log Avg loss: 1.67925131, Global Avg Loss: 2.32353475, Time: 0.0403 Steps: 12490, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001855, Sample Num: 29680, Cur Loss: 1.98760581, Cur Avg Loss: 1.58584516, Log Avg loss: 1.57024855, Global Avg Loss: 2.32293212, Time: 0.0403 Steps: 12500, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001865, Sample Num: 29840, Cur Loss: 1.19237077, Cur Avg Loss: 1.58659552, Log Avg loss: 1.72578723, Global Avg Loss: 2.32245478, Time: 0.0403 Steps: 12510, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001875, Sample Num: 30000, Cur Loss: 0.91690451, Cur Avg Loss: 1.58456718, Log Avg loss: 1.20628252, Global Avg Loss: 2.32156327, Time: 0.0403 Steps: 12520, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001885, Sample Num: 30160, Cur Loss: 1.95049286, Cur Avg Loss: 1.58586854, Log Avg loss: 1.82987363, Global Avg Loss: 2.32117086, Time: 0.0403 Steps: 12530, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001895, Sample Num: 30320, Cur Loss: 2.53005958, Cur Avg Loss: 1.58564166, Log Avg loss: 1.54287532, Global Avg Loss: 2.32055021, Time: 0.0403 Steps: 12540, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001905, Sample Num: 30480, Cur Loss: 0.65171897, Cur Avg Loss: 1.58487606, Log Avg loss: 1.43979464, Global Avg Loss: 2.31984841, Time: 0.0402 Steps: 12550, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001915, Sample Num: 30640, Cur Loss: 1.04052746, Cur Avg Loss: 1.58537378, Log Avg loss: 1.68018924, Global Avg Loss: 2.31933913, Time: 0.0403 Steps: 12560, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001925, Sample Num: 30800, Cur Loss: 0.99238837, Cur Avg Loss: 1.58503742, Log Avg loss: 1.52062350, Global Avg Loss: 2.31870372, Time: 0.0402 Steps: 12570, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001935, Sample Num: 30960, Cur Loss: 1.31218767, Cur Avg Loss: 1.58608346, Log Avg loss: 1.78744688, Global Avg Loss: 2.31828141, Time: 0.0402 Steps: 12580, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001945, Sample Num: 31120, Cur Loss: 2.07744312, Cur Avg Loss: 1.58643799, Log Avg loss: 1.65503911, Global Avg Loss: 2.31775461, Time: 0.0402 Steps: 12590, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001955, Sample Num: 31280, Cur Loss: 1.20355141, Cur Avg Loss: 1.58417304, Log Avg loss: 1.14364044, Global Avg Loss: 2.31682278, Time: 0.0402 Steps: 12600, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001965, Sample Num: 31440, Cur Loss: 1.96494651, Cur Avg Loss: 1.58480191, Log Avg loss: 1.70774701, Global Avg Loss: 2.31633977, Time: 0.0403 Steps: 12610, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001975, Sample Num: 31600, Cur Loss: 0.77140445, Cur Avg Loss: 1.58522197, Log Avg loss: 1.66776310, Global Avg Loss: 2.31582584, Time: 0.0402 Steps: 12620, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001985, Sample Num: 31760, Cur Loss: 2.14717317, Cur Avg Loss: 1.58730696, Log Avg loss: 1.99909254, Global Avg Loss: 2.31557506, Time: 0.0402 Steps: 12630, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001995, Sample Num: 31920, Cur Loss: 1.58444226, Cur Avg Loss: 1.58733762, Log Avg loss: 1.59342434, Global Avg Loss: 2.31500374, Time: 0.0402 Steps: 12640, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002005, Sample Num: 32080, Cur Loss: 2.54117203, Cur Avg Loss: 1.58670683, Log Avg loss: 1.46086400, Global Avg Loss: 2.31432853, Time: 0.0402 Steps: 12650, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002015, Sample Num: 32240, Cur Loss: 1.33217156, Cur Avg Loss: 1.58672744, Log Avg loss: 1.59085872, Global Avg Loss: 2.31375707, Time: 0.0406 Steps: 12660, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002025, Sample Num: 32400, Cur Loss: 3.04390407, Cur Avg Loss: 1.58751396, Log Avg loss: 1.74599899, Global Avg Loss: 2.31330896, Time: 0.0408 Steps: 12670, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002035, Sample Num: 32560, Cur Loss: 1.70097744, Cur Avg Loss: 1.58814561, Log Avg loss: 1.71605468, Global Avg Loss: 2.31283793, Time: 0.0595 Steps: 12680, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002045, Sample Num: 32720, Cur Loss: 0.55967343, Cur Avg Loss: 1.58935740, Log Avg loss: 1.83595672, Global Avg Loss: 2.31246214, Time: 0.0408 Steps: 12690, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002055, Sample Num: 32880, Cur Loss: 3.10367489, Cur Avg Loss: 1.58866197, Log Avg loss: 1.44644631, Global Avg Loss: 2.31178024, Time: 0.0405 Steps: 12700, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002065, Sample Num: 33040, Cur Loss: 1.38690960, Cur Avg Loss: 1.58820893, Log Avg loss: 1.49510769, Global Avg Loss: 2.31113770, Time: 0.0775 Steps: 12710, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002075, Sample Num: 33200, Cur Loss: 1.59446120, Cur Avg Loss: 1.58672764, Log Avg loss: 1.28084181, Global Avg Loss: 2.31032772, Time: 0.0430 Steps: 12720, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002085, Sample Num: 33360, Cur Loss: 3.06761265, Cur Avg Loss: 1.58752126, Log Avg loss: 1.75219709, Global Avg Loss: 2.30988928, Time: 0.0613 Steps: 12730, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002095, Sample Num: 33520, Cur Loss: 0.90951216, Cur Avg Loss: 1.58764088, Log Avg loss: 1.61258219, Global Avg Loss: 2.30934194, Time: 0.0993 Steps: 12740, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002105, Sample Num: 33680, Cur Loss: 1.75690138, Cur Avg Loss: 1.58815401, Log Avg loss: 1.69565458, Global Avg Loss: 2.30886062, Time: 0.0402 Steps: 12750, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002115, Sample Num: 33840, Cur Loss: 1.48629761, Cur Avg Loss: 1.58796027, Log Avg loss: 1.54717867, Global Avg Loss: 2.30826369, Time: 0.0404 Steps: 12760, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002125, Sample Num: 34000, Cur Loss: 1.37133253, Cur Avg Loss: 1.58963639, Log Avg loss: 1.94413526, Global Avg Loss: 2.30797854, Time: 0.0402 Steps: 12770, Updated lr: 0.000089 ***** Running evaluation checkpoint-12774 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-12774 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.189345, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.444096, "eval_total_loss": 1015.199814, "eval_mae": 1.010303, "eval_mse": 1.444195, "eval_r2": 0.081975, "eval_sp_statistic": 0.25392, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.329373, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.596888, "test_total_loss": 801.637765, "test_mae": 1.139106, "test_mse": 1.597005, "test_r2": -0.030721, "test_sp_statistic": 0.252574, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.333613, "test_ps_pvalue": 0.0, "lr": 8.883451872925558e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.307747711562972, "train_cur_epoch_loss": 3384.2605774998665, "train_cur_epoch_avg_loss": 1.5896010227805855, "train_cur_epoch_time": 95.18934464454651, "train_cur_epoch_avg_time": 0.04471082416371372, "epoch": 6, "step": 12774} ################################################## Training, Epoch: 0007, Batch: 000006, Sample Num: 96, Cur Loss: 1.20727670, Cur Avg Loss: 1.27744778, Log Avg loss: 1.39479378, Global Avg Loss: 2.30726400, Time: 0.0682 Steps: 12780, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000016, Sample Num: 256, Cur Loss: 1.67098951, Cur Avg Loss: 1.52397522, Log Avg loss: 1.67189168, Global Avg Loss: 2.30676723, Time: 0.0853 Steps: 12790, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000026, Sample Num: 416, Cur Loss: 1.95598865, Cur Avg Loss: 1.53344693, Log Avg loss: 1.54860167, Global Avg Loss: 2.30617491, Time: 0.0407 Steps: 12800, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000036, Sample Num: 576, Cur Loss: 0.73831522, Cur Avg Loss: 1.51709776, Log Avg loss: 1.47458991, Global Avg Loss: 2.30552574, Time: 0.0405 Steps: 12810, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000046, Sample Num: 736, Cur Loss: 1.68975139, Cur Avg Loss: 1.55815174, Log Avg loss: 1.70594606, Global Avg Loss: 2.30505805, Time: 0.0402 Steps: 12820, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000056, Sample Num: 896, Cur Loss: 2.29640293, Cur Avg Loss: 1.53946447, Log Avg loss: 1.45350305, Global Avg Loss: 2.30439433, Time: 0.0403 Steps: 12830, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000066, Sample Num: 1056, Cur Loss: 1.99059117, Cur Avg Loss: 1.53303173, Log Avg loss: 1.49700838, Global Avg Loss: 2.30376553, Time: 0.0402 Steps: 12840, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000076, Sample Num: 1216, Cur Loss: 2.22357178, Cur Avg Loss: 1.59576621, Log Avg loss: 2.00981381, Global Avg Loss: 2.30353677, Time: 0.0403 Steps: 12850, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000086, Sample Num: 1376, Cur Loss: 1.11098111, Cur Avg Loss: 1.55821114, Log Avg loss: 1.27279258, Global Avg Loss: 2.30273526, Time: 0.0403 Steps: 12860, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000096, Sample Num: 1536, Cur Loss: 0.79790604, Cur Avg Loss: 1.56145613, Log Avg loss: 1.58936304, Global Avg Loss: 2.30218097, Time: 0.0403 Steps: 12870, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000106, Sample Num: 1696, Cur Loss: 2.80254841, Cur Avg Loss: 1.58865002, Log Avg loss: 1.84971133, Global Avg Loss: 2.30182967, Time: 0.0403 Steps: 12880, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000116, Sample Num: 1856, Cur Loss: 2.30735922, Cur Avg Loss: 1.60445487, Log Avg loss: 1.77198628, Global Avg Loss: 2.30141862, Time: 0.0402 Steps: 12890, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000126, Sample Num: 2016, Cur Loss: 3.56079865, Cur Avg Loss: 1.61733937, Log Avg loss: 1.76679963, Global Avg Loss: 2.30100419, Time: 0.0403 Steps: 12900, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000136, Sample Num: 2176, Cur Loss: 1.59101725, Cur Avg Loss: 1.61331961, Log Avg loss: 1.56267059, Global Avg Loss: 2.30043228, Time: 0.0403 Steps: 12910, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000146, Sample Num: 2336, Cur Loss: 2.42682028, Cur Avg Loss: 1.62973563, Log Avg loss: 1.85299350, Global Avg Loss: 2.30008597, Time: 0.0402 Steps: 12920, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000156, Sample Num: 2496, Cur Loss: 0.92376995, Cur Avg Loss: 1.61696698, Log Avg loss: 1.43054465, Global Avg Loss: 2.29941347, Time: 0.0402 Steps: 12930, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000166, Sample Num: 2656, Cur Loss: 1.77923846, Cur Avg Loss: 1.62118161, Log Avg loss: 1.68692988, Global Avg Loss: 2.29894014, Time: 0.0403 Steps: 12940, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000176, Sample Num: 2816, Cur Loss: 1.56586897, Cur Avg Loss: 1.61921652, Log Avg loss: 1.58659595, Global Avg Loss: 2.29839007, Time: 0.0403 Steps: 12950, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000186, Sample Num: 2976, Cur Loss: 1.86314964, Cur Avg Loss: 1.61399594, Log Avg loss: 1.52211387, Global Avg Loss: 2.29779109, Time: 0.0402 Steps: 12960, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000196, Sample Num: 3136, Cur Loss: 1.53977334, Cur Avg Loss: 1.62847259, Log Avg loss: 1.89773828, Global Avg Loss: 2.29748264, Time: 0.0402 Steps: 12970, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000206, Sample Num: 3296, Cur Loss: 1.58397710, Cur Avg Loss: 1.63188907, Log Avg loss: 1.69885198, Global Avg Loss: 2.29702145, Time: 0.0403 Steps: 12980, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000216, Sample Num: 3456, Cur Loss: 1.42238533, Cur Avg Loss: 1.63034952, Log Avg loss: 1.59863490, Global Avg Loss: 2.29648382, Time: 0.0403 Steps: 12990, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000226, Sample Num: 3616, Cur Loss: 1.21191347, Cur Avg Loss: 1.63205710, Log Avg loss: 1.66894078, Global Avg Loss: 2.29600109, Time: 0.0403 Steps: 13000, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000236, Sample Num: 3776, Cur Loss: 0.78951180, Cur Avg Loss: 1.61548193, Log Avg loss: 1.24088303, Global Avg Loss: 2.29519008, Time: 0.0402 Steps: 13010, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000246, Sample Num: 3936, Cur Loss: 2.88706255, Cur Avg Loss: 1.62351686, Log Avg loss: 1.81314134, Global Avg Loss: 2.29481985, Time: 0.0402 Steps: 13020, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000256, Sample Num: 4096, Cur Loss: 2.76558900, Cur Avg Loss: 1.63051072, Log Avg loss: 1.80255972, Global Avg Loss: 2.29444206, Time: 0.0404 Steps: 13030, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000266, Sample Num: 4256, Cur Loss: 0.74137175, Cur Avg Loss: 1.62282092, Log Avg loss: 1.42596196, Global Avg Loss: 2.29377605, Time: 0.0403 Steps: 13040, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000276, Sample Num: 4416, Cur Loss: 0.38371131, Cur Avg Loss: 1.61545599, Log Avg loss: 1.41954884, Global Avg Loss: 2.29310614, Time: 0.0402 Steps: 13050, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000286, Sample Num: 4576, Cur Loss: 1.49231052, Cur Avg Loss: 1.62387009, Log Avg loss: 1.85609909, Global Avg Loss: 2.29277152, Time: 0.0402 Steps: 13060, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000296, Sample Num: 4736, Cur Loss: 1.74810410, Cur Avg Loss: 1.62605131, Log Avg loss: 1.68843441, Global Avg Loss: 2.29230914, Time: 0.0402 Steps: 13070, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000306, Sample Num: 4896, Cur Loss: 1.07078588, Cur Avg Loss: 1.61672541, Log Avg loss: 1.34067864, Global Avg Loss: 2.29158159, Time: 0.0643 Steps: 13080, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000316, Sample Num: 5056, Cur Loss: 0.78930694, Cur Avg Loss: 1.62196677, Log Avg loss: 1.78235251, Global Avg Loss: 2.29119257, Time: 0.0535 Steps: 13090, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000326, Sample Num: 5216, Cur Loss: 0.75701368, Cur Avg Loss: 1.62041462, Log Avg loss: 1.57136661, Global Avg Loss: 2.29064309, Time: 0.0432 Steps: 13100, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000336, Sample Num: 5376, Cur Loss: 1.54487252, Cur Avg Loss: 1.61263096, Log Avg loss: 1.35888349, Global Avg Loss: 2.28993236, Time: 0.0784 Steps: 13110, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000346, Sample Num: 5536, Cur Loss: 2.39306760, Cur Avg Loss: 1.62369400, Log Avg loss: 1.99541218, Global Avg Loss: 2.28970788, Time: 0.0404 Steps: 13120, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000356, Sample Num: 5696, Cur Loss: 0.85793817, Cur Avg Loss: 1.61421786, Log Avg loss: 1.28634357, Global Avg Loss: 2.28894370, Time: 0.0406 Steps: 13130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000366, Sample Num: 5856, Cur Loss: 2.45863676, Cur Avg Loss: 1.62987210, Log Avg loss: 2.18716304, Global Avg Loss: 2.28886624, Time: 0.0403 Steps: 13140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000376, Sample Num: 6016, Cur Loss: 2.06391716, Cur Avg Loss: 1.63188388, Log Avg loss: 1.70551488, Global Avg Loss: 2.28842263, Time: 0.0406 Steps: 13150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000386, Sample Num: 6176, Cur Loss: 0.52707160, Cur Avg Loss: 1.62087989, Log Avg loss: 1.20712987, Global Avg Loss: 2.28760098, Time: 0.0833 Steps: 13160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000396, Sample Num: 6336, Cur Loss: 1.45946872, Cur Avg Loss: 1.62846490, Log Avg loss: 1.92124648, Global Avg Loss: 2.28732281, Time: 0.0402 Steps: 13170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000406, Sample Num: 6496, Cur Loss: 1.18930066, Cur Avg Loss: 1.62619382, Log Avg loss: 1.53625886, Global Avg Loss: 2.28675296, Time: 0.0402 Steps: 13180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000416, Sample Num: 6656, Cur Loss: 0.70980394, Cur Avg Loss: 1.61629236, Log Avg loss: 1.21429334, Global Avg Loss: 2.28593987, Time: 0.0402 Steps: 13190, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000426, Sample Num: 6816, Cur Loss: 1.91555250, Cur Avg Loss: 1.62070757, Log Avg loss: 1.80438014, Global Avg Loss: 2.28557505, Time: 0.0402 Steps: 13200, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000436, Sample Num: 6976, Cur Loss: 1.81828082, Cur Avg Loss: 1.61808446, Log Avg loss: 1.50633998, Global Avg Loss: 2.28498517, Time: 0.0402 Steps: 13210, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000446, Sample Num: 7136, Cur Loss: 2.42597914, Cur Avg Loss: 1.62195278, Log Avg loss: 1.79061149, Global Avg Loss: 2.28461121, Time: 0.0402 Steps: 13220, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000456, Sample Num: 7296, Cur Loss: 1.51533103, Cur Avg Loss: 1.62358652, Log Avg loss: 1.69645133, Global Avg Loss: 2.28416665, Time: 0.0402 Steps: 13230, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000466, Sample Num: 7456, Cur Loss: 1.89859819, Cur Avg Loss: 1.62541392, Log Avg loss: 1.70874341, Global Avg Loss: 2.28373204, Time: 0.0402 Steps: 13240, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000476, Sample Num: 7616, Cur Loss: 1.19551444, Cur Avg Loss: 1.62660396, Log Avg loss: 1.68205961, Global Avg Loss: 2.28327794, Time: 0.0402 Steps: 13250, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000486, Sample Num: 7776, Cur Loss: 1.88012278, Cur Avg Loss: 1.63230914, Log Avg loss: 1.90387570, Global Avg Loss: 2.28299182, Time: 0.0402 Steps: 13260, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000496, Sample Num: 7936, Cur Loss: 0.94979107, Cur Avg Loss: 1.63590483, Log Avg loss: 1.81065571, Global Avg Loss: 2.28263588, Time: 0.0402 Steps: 13270, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000506, Sample Num: 8096, Cur Loss: 1.02487803, Cur Avg Loss: 1.63426818, Log Avg loss: 1.55309015, Global Avg Loss: 2.28208652, Time: 0.0402 Steps: 13280, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000516, Sample Num: 8256, Cur Loss: 1.29736090, Cur Avg Loss: 1.63200771, Log Avg loss: 1.51762771, Global Avg Loss: 2.28151130, Time: 0.0403 Steps: 13290, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000526, Sample Num: 8416, Cur Loss: 1.05514312, Cur Avg Loss: 1.62756783, Log Avg loss: 1.39847018, Global Avg Loss: 2.28084736, Time: 0.0402 Steps: 13300, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000536, Sample Num: 8576, Cur Loss: 1.80066431, Cur Avg Loss: 1.62522238, Log Avg loss: 1.50185196, Global Avg Loss: 2.28026209, Time: 0.0402 Steps: 13310, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000546, Sample Num: 8736, Cur Loss: 1.43300676, Cur Avg Loss: 1.62839420, Log Avg loss: 1.79840376, Global Avg Loss: 2.27990034, Time: 0.0402 Steps: 13320, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000556, Sample Num: 8896, Cur Loss: 1.16013908, Cur Avg Loss: 1.62178927, Log Avg loss: 1.26116014, Global Avg Loss: 2.27913609, Time: 0.0402 Steps: 13330, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000566, Sample Num: 9056, Cur Loss: 2.18773174, Cur Avg Loss: 1.62436376, Log Avg loss: 1.76750515, Global Avg Loss: 2.27875256, Time: 0.0402 Steps: 13340, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000576, Sample Num: 9216, Cur Loss: 0.63457525, Cur Avg Loss: 1.61771492, Log Avg loss: 1.24139085, Global Avg Loss: 2.27797551, Time: 0.0403 Steps: 13350, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000586, Sample Num: 9376, Cur Loss: 2.42238903, Cur Avg Loss: 1.61778529, Log Avg loss: 1.62183844, Global Avg Loss: 2.27748439, Time: 0.0403 Steps: 13360, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000596, Sample Num: 9536, Cur Loss: 2.49152136, Cur Avg Loss: 1.61458638, Log Avg loss: 1.42713001, Global Avg Loss: 2.27684837, Time: 0.0402 Steps: 13370, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000606, Sample Num: 9696, Cur Loss: 0.77334762, Cur Avg Loss: 1.61034094, Log Avg loss: 1.35731279, Global Avg Loss: 2.27616113, Time: 0.0403 Steps: 13380, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000616, Sample Num: 9856, Cur Loss: 0.80899668, Cur Avg Loss: 1.60689141, Log Avg loss: 1.39784985, Global Avg Loss: 2.27550518, Time: 0.0402 Steps: 13390, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000626, Sample Num: 10016, Cur Loss: 1.22417736, Cur Avg Loss: 1.60187077, Log Avg loss: 1.29259930, Global Avg Loss: 2.27477167, Time: 0.0402 Steps: 13400, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000636, Sample Num: 10176, Cur Loss: 2.30577469, Cur Avg Loss: 1.60415931, Log Avg loss: 1.74742202, Global Avg Loss: 2.27437842, Time: 0.0402 Steps: 13410, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000646, Sample Num: 10336, Cur Loss: 1.19188762, Cur Avg Loss: 1.60042879, Log Avg loss: 1.36316766, Global Avg Loss: 2.27369942, Time: 0.0403 Steps: 13420, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000656, Sample Num: 10496, Cur Loss: 2.69209647, Cur Avg Loss: 1.60181148, Log Avg loss: 1.69113349, Global Avg Loss: 2.27326564, Time: 0.0402 Steps: 13430, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000666, Sample Num: 10656, Cur Loss: 1.33721626, Cur Avg Loss: 1.59618099, Log Avg loss: 1.22682068, Global Avg Loss: 2.27248704, Time: 0.0544 Steps: 13440, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000676, Sample Num: 10816, Cur Loss: 1.70752263, Cur Avg Loss: 1.60023611, Log Avg loss: 1.87030723, Global Avg Loss: 2.27218802, Time: 0.0510 Steps: 13450, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000686, Sample Num: 10976, Cur Loss: 1.80427992, Cur Avg Loss: 1.60656006, Log Avg loss: 2.03405874, Global Avg Loss: 2.27201110, Time: 0.0462 Steps: 13460, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000696, Sample Num: 11136, Cur Loss: 0.53614062, Cur Avg Loss: 1.60249052, Log Avg loss: 1.32332032, Global Avg Loss: 2.27130681, Time: 0.0404 Steps: 13470, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000706, Sample Num: 11296, Cur Loss: 1.71181476, Cur Avg Loss: 1.60015781, Log Avg loss: 1.43780127, Global Avg Loss: 2.27068848, Time: 0.0549 Steps: 13480, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000716, Sample Num: 11456, Cur Loss: 0.64144522, Cur Avg Loss: 1.59490792, Log Avg loss: 1.22426587, Global Avg Loss: 2.26991278, Time: 0.0991 Steps: 13490, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000726, Sample Num: 11616, Cur Loss: 2.44063759, Cur Avg Loss: 1.59461644, Log Avg loss: 1.57374592, Global Avg Loss: 2.26939710, Time: 0.0408 Steps: 13500, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000736, Sample Num: 11776, Cur Loss: 1.49647141, Cur Avg Loss: 1.59093505, Log Avg loss: 1.32366635, Global Avg Loss: 2.26869707, Time: 0.0403 Steps: 13510, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000746, Sample Num: 11936, Cur Loss: 2.99612927, Cur Avg Loss: 1.59198052, Log Avg loss: 1.66892730, Global Avg Loss: 2.26825346, Time: 0.0404 Steps: 13520, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000756, Sample Num: 12096, Cur Loss: 2.29314566, Cur Avg Loss: 1.59459807, Log Avg loss: 1.78986746, Global Avg Loss: 2.26789988, Time: 0.0784 Steps: 13530, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000766, Sample Num: 12256, Cur Loss: 1.16959763, Cur Avg Loss: 1.58843808, Log Avg loss: 1.12274276, Global Avg Loss: 2.26705412, Time: 0.0403 Steps: 13540, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000776, Sample Num: 12416, Cur Loss: 2.34627080, Cur Avg Loss: 1.58733245, Log Avg loss: 1.50264066, Global Avg Loss: 2.26648998, Time: 0.0403 Steps: 13550, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000786, Sample Num: 12576, Cur Loss: 0.82025886, Cur Avg Loss: 1.58574397, Log Avg loss: 1.46247823, Global Avg Loss: 2.26589705, Time: 0.0402 Steps: 13560, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000796, Sample Num: 12736, Cur Loss: 2.86337137, Cur Avg Loss: 1.58720496, Log Avg loss: 1.70203867, Global Avg Loss: 2.26548153, Time: 0.0402 Steps: 13570, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000806, Sample Num: 12896, Cur Loss: 0.77004439, Cur Avg Loss: 1.58249825, Log Avg loss: 1.20784443, Global Avg Loss: 2.26470271, Time: 0.0403 Steps: 13580, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000816, Sample Num: 13056, Cur Loss: 1.94986510, Cur Avg Loss: 1.58598879, Log Avg loss: 1.86732641, Global Avg Loss: 2.26441031, Time: 0.0403 Steps: 13590, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000826, Sample Num: 13216, Cur Loss: 2.04643011, Cur Avg Loss: 1.59339199, Log Avg loss: 2.19749311, Global Avg Loss: 2.26436111, Time: 0.0402 Steps: 13600, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000836, Sample Num: 13376, Cur Loss: 1.57703352, Cur Avg Loss: 1.59175041, Log Avg loss: 1.45615602, Global Avg Loss: 2.26376728, Time: 0.0402 Steps: 13610, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000846, Sample Num: 13536, Cur Loss: 1.62308431, Cur Avg Loss: 1.59674748, Log Avg loss: 2.01450194, Global Avg Loss: 2.26358426, Time: 0.0403 Steps: 13620, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000856, Sample Num: 13696, Cur Loss: 0.90878797, Cur Avg Loss: 1.59480297, Log Avg loss: 1.43029757, Global Avg Loss: 2.26297290, Time: 0.0402 Steps: 13630, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000866, Sample Num: 13856, Cur Loss: 0.78495008, Cur Avg Loss: 1.59258790, Log Avg loss: 1.40297778, Global Avg Loss: 2.26234240, Time: 0.0402 Steps: 13640, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000876, Sample Num: 14016, Cur Loss: 1.80279231, Cur Avg Loss: 1.59396503, Log Avg loss: 1.71322499, Global Avg Loss: 2.26194012, Time: 0.0402 Steps: 13650, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000886, Sample Num: 14176, Cur Loss: 1.80675960, Cur Avg Loss: 1.59454082, Log Avg loss: 1.64497947, Global Avg Loss: 2.26148846, Time: 0.0402 Steps: 13660, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000896, Sample Num: 14336, Cur Loss: 1.35469425, Cur Avg Loss: 1.60058835, Log Avg loss: 2.13640001, Global Avg Loss: 2.26139696, Time: 0.0402 Steps: 13670, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000906, Sample Num: 14496, Cur Loss: 1.26922870, Cur Avg Loss: 1.60287373, Log Avg loss: 1.80764366, Global Avg Loss: 2.26106527, Time: 0.0402 Steps: 13680, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000916, Sample Num: 14656, Cur Loss: 1.31437719, Cur Avg Loss: 1.60082060, Log Avg loss: 1.41480737, Global Avg Loss: 2.26044711, Time: 0.0403 Steps: 13690, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000926, Sample Num: 14816, Cur Loss: 1.28382111, Cur Avg Loss: 1.59897081, Log Avg loss: 1.42952980, Global Avg Loss: 2.25984060, Time: 0.0402 Steps: 13700, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000936, Sample Num: 14976, Cur Loss: 1.92754555, Cur Avg Loss: 1.59633106, Log Avg loss: 1.35189031, Global Avg Loss: 2.25917835, Time: 0.0402 Steps: 13710, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000946, Sample Num: 15136, Cur Loss: 1.55044055, Cur Avg Loss: 1.59757769, Log Avg loss: 1.71426192, Global Avg Loss: 2.25878118, Time: 0.0402 Steps: 13720, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000956, Sample Num: 15296, Cur Loss: 2.00174665, Cur Avg Loss: 1.59990042, Log Avg loss: 1.81963120, Global Avg Loss: 2.25846133, Time: 0.0402 Steps: 13730, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000966, Sample Num: 15456, Cur Loss: 1.82075298, Cur Avg Loss: 1.60187781, Log Avg loss: 1.79091574, Global Avg Loss: 2.25812105, Time: 0.0402 Steps: 13740, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000976, Sample Num: 15616, Cur Loss: 1.26048613, Cur Avg Loss: 1.59927381, Log Avg loss: 1.34772736, Global Avg Loss: 2.25745895, Time: 0.0402 Steps: 13750, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000986, Sample Num: 15776, Cur Loss: 1.65276611, Cur Avg Loss: 1.60253554, Log Avg loss: 1.92088057, Global Avg Loss: 2.25721434, Time: 0.0402 Steps: 13760, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000996, Sample Num: 15936, Cur Loss: 0.82392019, Cur Avg Loss: 1.60342107, Log Avg loss: 1.69073411, Global Avg Loss: 2.25680295, Time: 0.0402 Steps: 13770, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001006, Sample Num: 16096, Cur Loss: 1.59759939, Cur Avg Loss: 1.60161260, Log Avg loss: 1.42148877, Global Avg Loss: 2.25619677, Time: 0.0402 Steps: 13780, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001016, Sample Num: 16256, Cur Loss: 2.96492910, Cur Avg Loss: 1.60046109, Log Avg loss: 1.48461932, Global Avg Loss: 2.25563725, Time: 0.0578 Steps: 13790, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001026, Sample Num: 16416, Cur Loss: 1.76669943, Cur Avg Loss: 1.60077989, Log Avg loss: 1.63316994, Global Avg Loss: 2.25518619, Time: 0.0622 Steps: 13800, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001036, Sample Num: 16576, Cur Loss: 1.18058944, Cur Avg Loss: 1.60124953, Log Avg loss: 1.64943523, Global Avg Loss: 2.25474756, Time: 0.0406 Steps: 13810, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001046, Sample Num: 16736, Cur Loss: 1.50620151, Cur Avg Loss: 1.60122836, Log Avg loss: 1.59903481, Global Avg Loss: 2.25427309, Time: 0.0644 Steps: 13820, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001056, Sample Num: 16896, Cur Loss: 1.41226161, Cur Avg Loss: 1.60169743, Log Avg loss: 1.65076202, Global Avg Loss: 2.25383671, Time: 0.0780 Steps: 13830, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001066, Sample Num: 17056, Cur Loss: 2.53613710, Cur Avg Loss: 1.60689772, Log Avg loss: 2.15604818, Global Avg Loss: 2.25376606, Time: 0.0797 Steps: 13840, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001076, Sample Num: 17216, Cur Loss: 2.91435623, Cur Avg Loss: 1.60478139, Log Avg loss: 1.37918036, Global Avg Loss: 2.25313459, Time: 0.0491 Steps: 13850, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001086, Sample Num: 17376, Cur Loss: 1.36482334, Cur Avg Loss: 1.60698630, Log Avg loss: 1.84423522, Global Avg Loss: 2.25283957, Time: 0.0733 Steps: 13860, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001096, Sample Num: 17536, Cur Loss: 0.85705572, Cur Avg Loss: 1.60409143, Log Avg loss: 1.28970855, Global Avg Loss: 2.25214517, Time: 0.0404 Steps: 13870, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001106, Sample Num: 17696, Cur Loss: 1.36388969, Cur Avg Loss: 1.60310251, Log Avg loss: 1.49471678, Global Avg Loss: 2.25159947, Time: 0.0403 Steps: 13880, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001116, Sample Num: 17856, Cur Loss: 1.81686687, Cur Avg Loss: 1.60402538, Log Avg loss: 1.70609506, Global Avg Loss: 2.25120674, Time: 0.0403 Steps: 13890, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001126, Sample Num: 18016, Cur Loss: 1.92470407, Cur Avg Loss: 1.60564420, Log Avg loss: 1.78630448, Global Avg Loss: 2.25087228, Time: 0.0402 Steps: 13900, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001136, Sample Num: 18176, Cur Loss: 2.73163891, Cur Avg Loss: 1.60684218, Log Avg loss: 1.74173500, Global Avg Loss: 2.25050625, Time: 0.0402 Steps: 13910, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001146, Sample Num: 18336, Cur Loss: 0.57077551, Cur Avg Loss: 1.60840369, Log Avg loss: 1.78579104, Global Avg Loss: 2.25017241, Time: 0.0403 Steps: 13920, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001156, Sample Num: 18496, Cur Loss: 1.00619805, Cur Avg Loss: 1.60825675, Log Avg loss: 1.59141657, Global Avg Loss: 2.24969950, Time: 0.0402 Steps: 13930, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001166, Sample Num: 18656, Cur Loss: 1.01352191, Cur Avg Loss: 1.60846872, Log Avg loss: 1.63297296, Global Avg Loss: 2.24925709, Time: 0.0402 Steps: 13940, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001176, Sample Num: 18816, Cur Loss: 1.91013873, Cur Avg Loss: 1.60662334, Log Avg loss: 1.39145235, Global Avg Loss: 2.24864217, Time: 0.0402 Steps: 13950, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001186, Sample Num: 18976, Cur Loss: 1.99607122, Cur Avg Loss: 1.60383743, Log Avg loss: 1.27621430, Global Avg Loss: 2.24794559, Time: 0.0403 Steps: 13960, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001196, Sample Num: 19136, Cur Loss: 1.65968549, Cur Avg Loss: 1.60361475, Log Avg loss: 1.57720443, Global Avg Loss: 2.24746546, Time: 0.0402 Steps: 13970, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001206, Sample Num: 19296, Cur Loss: 1.48034668, Cur Avg Loss: 1.60292497, Log Avg loss: 1.52042753, Global Avg Loss: 2.24694541, Time: 0.0402 Steps: 13980, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001216, Sample Num: 19456, Cur Loss: 0.81694722, Cur Avg Loss: 1.60185840, Log Avg loss: 1.47323026, Global Avg Loss: 2.24639236, Time: 0.0404 Steps: 13990, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001226, Sample Num: 19616, Cur Loss: 1.50150752, Cur Avg Loss: 1.60422976, Log Avg loss: 1.89258707, Global Avg Loss: 2.24613964, Time: 0.0403 Steps: 14000, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001236, Sample Num: 19776, Cur Loss: 0.40292543, Cur Avg Loss: 1.60421085, Log Avg loss: 1.60189273, Global Avg Loss: 2.24567979, Time: 0.0402 Steps: 14010, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001246, Sample Num: 19936, Cur Loss: 1.69565260, Cur Avg Loss: 1.60613883, Log Avg loss: 1.84443706, Global Avg Loss: 2.24539360, Time: 0.0402 Steps: 14020, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001256, Sample Num: 20096, Cur Loss: 1.71253157, Cur Avg Loss: 1.60601227, Log Avg loss: 1.59024228, Global Avg Loss: 2.24492663, Time: 0.0403 Steps: 14030, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001266, Sample Num: 20256, Cur Loss: 2.96497679, Cur Avg Loss: 1.60701152, Log Avg loss: 1.73251774, Global Avg Loss: 2.24456167, Time: 0.0402 Steps: 14040, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001276, Sample Num: 20416, Cur Loss: 1.37616372, Cur Avg Loss: 1.60639821, Log Avg loss: 1.52875259, Global Avg Loss: 2.24405220, Time: 0.0403 Steps: 14050, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001286, Sample Num: 20576, Cur Loss: 1.06159294, Cur Avg Loss: 1.60429161, Log Avg loss: 1.33549032, Global Avg Loss: 2.24340599, Time: 0.0403 Steps: 14060, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001296, Sample Num: 20736, Cur Loss: 2.58046222, Cur Avg Loss: 1.60589714, Log Avg loss: 1.81236848, Global Avg Loss: 2.24309964, Time: 0.0403 Steps: 14070, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001306, Sample Num: 20896, Cur Loss: 1.17175400, Cur Avg Loss: 1.60736010, Log Avg loss: 1.79695954, Global Avg Loss: 2.24278278, Time: 0.0402 Steps: 14080, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001316, Sample Num: 21056, Cur Loss: 0.64524239, Cur Avg Loss: 1.60800661, Log Avg loss: 1.69244099, Global Avg Loss: 2.24239219, Time: 0.0403 Steps: 14090, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001326, Sample Num: 21216, Cur Loss: 2.27581716, Cur Avg Loss: 1.61040614, Log Avg loss: 1.92618412, Global Avg Loss: 2.24216793, Time: 0.0403 Steps: 14100, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001336, Sample Num: 21376, Cur Loss: 1.50660598, Cur Avg Loss: 1.61113777, Log Avg loss: 1.70815213, Global Avg Loss: 2.24178946, Time: 0.0402 Steps: 14110, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001346, Sample Num: 21536, Cur Loss: 0.65011889, Cur Avg Loss: 1.61196323, Log Avg loss: 1.72224481, Global Avg Loss: 2.24142151, Time: 0.0402 Steps: 14120, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001356, Sample Num: 21696, Cur Loss: 2.42728448, Cur Avg Loss: 1.61135695, Log Avg loss: 1.52975136, Global Avg Loss: 2.24091786, Time: 0.0402 Steps: 14130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001366, Sample Num: 21856, Cur Loss: 1.98348820, Cur Avg Loss: 1.61198470, Log Avg loss: 1.69710666, Global Avg Loss: 2.24053326, Time: 0.0404 Steps: 14140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001376, Sample Num: 22016, Cur Loss: 0.77225924, Cur Avg Loss: 1.61319515, Log Avg loss: 1.77854381, Global Avg Loss: 2.24020677, Time: 0.0897 Steps: 14150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001386, Sample Num: 22176, Cur Loss: 2.44197559, Cur Avg Loss: 1.61176971, Log Avg loss: 1.41562898, Global Avg Loss: 2.23962444, Time: 0.0582 Steps: 14160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001396, Sample Num: 22336, Cur Loss: 1.63309252, Cur Avg Loss: 1.61130810, Log Avg loss: 1.54732860, Global Avg Loss: 2.23913588, Time: 0.0481 Steps: 14170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001406, Sample Num: 22496, Cur Loss: 1.30721164, Cur Avg Loss: 1.61174996, Log Avg loss: 1.67343406, Global Avg Loss: 2.23873693, Time: 0.0405 Steps: 14180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001416, Sample Num: 22656, Cur Loss: 0.69866073, Cur Avg Loss: 1.61163083, Log Avg loss: 1.59488102, Global Avg Loss: 2.23828319, Time: 0.0869 Steps: 14190, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001426, Sample Num: 22816, Cur Loss: 1.37137890, Cur Avg Loss: 1.61479745, Log Avg loss: 2.06318978, Global Avg Loss: 2.23815989, Time: 0.0511 Steps: 14200, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001436, Sample Num: 22976, Cur Loss: 0.56485498, Cur Avg Loss: 1.61407793, Log Avg loss: 1.51147451, Global Avg Loss: 2.23764850, Time: 0.0828 Steps: 14210, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001446, Sample Num: 23136, Cur Loss: 2.38554454, Cur Avg Loss: 1.61279332, Log Avg loss: 1.42832371, Global Avg Loss: 2.23707935, Time: 0.0555 Steps: 14220, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001456, Sample Num: 23296, Cur Loss: 1.04788303, Cur Avg Loss: 1.61120318, Log Avg loss: 1.38126916, Global Avg Loss: 2.23647794, Time: 0.0404 Steps: 14230, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001466, Sample Num: 23456, Cur Loss: 1.27815282, Cur Avg Loss: 1.61109054, Log Avg loss: 1.59468998, Global Avg Loss: 2.23602725, Time: 0.0402 Steps: 14240, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001476, Sample Num: 23616, Cur Loss: 2.43350434, Cur Avg Loss: 1.61118328, Log Avg loss: 1.62477956, Global Avg Loss: 2.23559830, Time: 0.0402 Steps: 14250, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001486, Sample Num: 23776, Cur Loss: 1.26272595, Cur Avg Loss: 1.60971813, Log Avg loss: 1.39346169, Global Avg Loss: 2.23500774, Time: 0.0403 Steps: 14260, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001496, Sample Num: 23936, Cur Loss: 3.39180684, Cur Avg Loss: 1.61000199, Log Avg loss: 1.65218402, Global Avg Loss: 2.23459932, Time: 0.0402 Steps: 14270, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001506, Sample Num: 24096, Cur Loss: 1.40122795, Cur Avg Loss: 1.60779208, Log Avg loss: 1.27718922, Global Avg Loss: 2.23392886, Time: 0.0403 Steps: 14280, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001516, Sample Num: 24256, Cur Loss: 1.50773108, Cur Avg Loss: 1.60699987, Log Avg loss: 1.48769248, Global Avg Loss: 2.23340665, Time: 0.0402 Steps: 14290, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001526, Sample Num: 24416, Cur Loss: 2.50314879, Cur Avg Loss: 1.60488325, Log Avg loss: 1.28400444, Global Avg Loss: 2.23274274, Time: 0.0403 Steps: 14300, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001536, Sample Num: 24576, Cur Loss: 1.77686954, Cur Avg Loss: 1.60475937, Log Avg loss: 1.58585421, Global Avg Loss: 2.23229068, Time: 0.0404 Steps: 14310, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001546, Sample Num: 24736, Cur Loss: 2.15174055, Cur Avg Loss: 1.60402306, Log Avg loss: 1.49092684, Global Avg Loss: 2.23177297, Time: 0.0404 Steps: 14320, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001556, Sample Num: 24896, Cur Loss: 2.07131505, Cur Avg Loss: 1.60566556, Log Avg loss: 1.85959533, Global Avg Loss: 2.23151325, Time: 0.0401 Steps: 14330, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001566, Sample Num: 25056, Cur Loss: 1.69117641, Cur Avg Loss: 1.60531149, Log Avg loss: 1.55021907, Global Avg Loss: 2.23103815, Time: 0.0402 Steps: 14340, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001576, Sample Num: 25216, Cur Loss: 1.74306774, Cur Avg Loss: 1.60413701, Log Avg loss: 1.42021307, Global Avg Loss: 2.23047311, Time: 0.0402 Steps: 14350, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001586, Sample Num: 25376, Cur Loss: 1.43070900, Cur Avg Loss: 1.60368649, Log Avg loss: 1.53268403, Global Avg Loss: 2.22998719, Time: 0.0403 Steps: 14360, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001596, Sample Num: 25536, Cur Loss: 1.76948905, Cur Avg Loss: 1.60488152, Log Avg loss: 1.79441286, Global Avg Loss: 2.22968408, Time: 0.0403 Steps: 14370, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001606, Sample Num: 25696, Cur Loss: 1.44827914, Cur Avg Loss: 1.60507206, Log Avg loss: 1.63548238, Global Avg Loss: 2.22927086, Time: 0.0402 Steps: 14380, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001616, Sample Num: 25856, Cur Loss: 0.97931117, Cur Avg Loss: 1.60611290, Log Avg loss: 1.77327276, Global Avg Loss: 2.22895398, Time: 0.0402 Steps: 14390, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001626, Sample Num: 26016, Cur Loss: 2.44671845, Cur Avg Loss: 1.60779081, Log Avg loss: 1.87894009, Global Avg Loss: 2.22871091, Time: 0.0402 Steps: 14400, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001636, Sample Num: 26176, Cur Loss: 1.85284460, Cur Avg Loss: 1.60806435, Log Avg loss: 1.65254327, Global Avg Loss: 2.22831107, Time: 0.0403 Steps: 14410, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001646, Sample Num: 26336, Cur Loss: 0.74737525, Cur Avg Loss: 1.60680141, Log Avg loss: 1.40018346, Global Avg Loss: 2.22773678, Time: 0.0402 Steps: 14420, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001656, Sample Num: 26496, Cur Loss: 1.58878183, Cur Avg Loss: 1.60738922, Log Avg loss: 1.70414231, Global Avg Loss: 2.22737393, Time: 0.0402 Steps: 14430, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001666, Sample Num: 26656, Cur Loss: 1.03077853, Cur Avg Loss: 1.60672034, Log Avg loss: 1.49595522, Global Avg Loss: 2.22686741, Time: 0.0402 Steps: 14440, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001676, Sample Num: 26816, Cur Loss: 1.85580552, Cur Avg Loss: 1.60608699, Log Avg loss: 1.50056971, Global Avg Loss: 2.22636478, Time: 0.0402 Steps: 14450, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001686, Sample Num: 26976, Cur Loss: 2.77528024, Cur Avg Loss: 1.60577585, Log Avg loss: 1.55362992, Global Avg Loss: 2.22589954, Time: 0.0402 Steps: 14460, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001696, Sample Num: 27136, Cur Loss: 1.49159336, Cur Avg Loss: 1.60618109, Log Avg loss: 1.67450456, Global Avg Loss: 2.22551848, Time: 0.0403 Steps: 14470, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001706, Sample Num: 27296, Cur Loss: 1.15593600, Cur Avg Loss: 1.60509233, Log Avg loss: 1.42043713, Global Avg Loss: 2.22496248, Time: 0.0402 Steps: 14480, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001716, Sample Num: 27456, Cur Loss: 1.98075318, Cur Avg Loss: 1.60447248, Log Avg loss: 1.49872647, Global Avg Loss: 2.22446129, Time: 0.0407 Steps: 14490, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001726, Sample Num: 27616, Cur Loss: 2.49189258, Cur Avg Loss: 1.60350278, Log Avg loss: 1.43710257, Global Avg Loss: 2.22391828, Time: 0.0632 Steps: 14500, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001736, Sample Num: 27776, Cur Loss: 1.31107557, Cur Avg Loss: 1.60311215, Log Avg loss: 1.53568987, Global Avg Loss: 2.22344397, Time: 0.0670 Steps: 14510, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001746, Sample Num: 27936, Cur Loss: 0.99744511, Cur Avg Loss: 1.60210662, Log Avg loss: 1.42754688, Global Avg Loss: 2.22289583, Time: 0.0842 Steps: 14520, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001756, Sample Num: 28096, Cur Loss: 0.92711091, Cur Avg Loss: 1.60187631, Log Avg loss: 1.56166399, Global Avg Loss: 2.22244075, Time: 0.0587 Steps: 14530, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001766, Sample Num: 28256, Cur Loss: 2.43588543, Cur Avg Loss: 1.60344740, Log Avg loss: 1.87933059, Global Avg Loss: 2.22220477, Time: 0.1003 Steps: 14540, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001776, Sample Num: 28416, Cur Loss: 2.42282963, Cur Avg Loss: 1.60410315, Log Avg loss: 1.71990908, Global Avg Loss: 2.22185955, Time: 0.0404 Steps: 14550, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001786, Sample Num: 28576, Cur Loss: 0.46588692, Cur Avg Loss: 1.60262218, Log Avg loss: 1.33960173, Global Avg Loss: 2.22125360, Time: 0.0420 Steps: 14560, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001796, Sample Num: 28736, Cur Loss: 0.50108910, Cur Avg Loss: 1.60091028, Log Avg loss: 1.29516363, Global Avg Loss: 2.22061799, Time: 0.0463 Steps: 14570, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001806, Sample Num: 28896, Cur Loss: 2.19188881, Cur Avg Loss: 1.60074644, Log Avg loss: 1.57132112, Global Avg Loss: 2.22017266, Time: 0.0402 Steps: 14580, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001816, Sample Num: 29056, Cur Loss: 1.70221913, Cur Avg Loss: 1.60163407, Log Avg loss: 1.76193972, Global Avg Loss: 2.21985858, Time: 0.0403 Steps: 14590, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001826, Sample Num: 29216, Cur Loss: 1.31857467, Cur Avg Loss: 1.60317856, Log Avg loss: 1.88365901, Global Avg Loss: 2.21962831, Time: 0.0402 Steps: 14600, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001836, Sample Num: 29376, Cur Loss: 2.33775139, Cur Avg Loss: 1.60462867, Log Avg loss: 1.86941881, Global Avg Loss: 2.21938860, Time: 0.0402 Steps: 14610, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001846, Sample Num: 29536, Cur Loss: 0.57431674, Cur Avg Loss: 1.60374946, Log Avg loss: 1.44232672, Global Avg Loss: 2.21885710, Time: 0.0402 Steps: 14620, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001856, Sample Num: 29696, Cur Loss: 1.78473938, Cur Avg Loss: 1.60391588, Log Avg loss: 1.63463582, Global Avg Loss: 2.21845777, Time: 0.0403 Steps: 14630, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001866, Sample Num: 29856, Cur Loss: 1.35776532, Cur Avg Loss: 1.60177292, Log Avg loss: 1.20403960, Global Avg Loss: 2.21776486, Time: 0.0402 Steps: 14640, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001876, Sample Num: 30016, Cur Loss: 1.50045657, Cur Avg Loss: 1.60206199, Log Avg loss: 1.65600350, Global Avg Loss: 2.21738140, Time: 0.0402 Steps: 14650, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001886, Sample Num: 30176, Cur Loss: 1.42446625, Cur Avg Loss: 1.60083320, Log Avg loss: 1.37031200, Global Avg Loss: 2.21680359, Time: 0.0403 Steps: 14660, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001896, Sample Num: 30336, Cur Loss: 1.56377006, Cur Avg Loss: 1.60032769, Log Avg loss: 1.50498865, Global Avg Loss: 2.21631838, Time: 0.0403 Steps: 14670, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001906, Sample Num: 30496, Cur Loss: 1.85982203, Cur Avg Loss: 1.60091916, Log Avg loss: 1.71306230, Global Avg Loss: 2.21597556, Time: 0.0403 Steps: 14680, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001916, Sample Num: 30656, Cur Loss: 1.87263691, Cur Avg Loss: 1.60116669, Log Avg loss: 1.64834580, Global Avg Loss: 2.21558915, Time: 0.0402 Steps: 14690, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001926, Sample Num: 30816, Cur Loss: 1.71925330, Cur Avg Loss: 1.59993735, Log Avg loss: 1.36439586, Global Avg Loss: 2.21501011, Time: 0.0402 Steps: 14700, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001936, Sample Num: 30976, Cur Loss: 1.15074134, Cur Avg Loss: 1.60090114, Log Avg loss: 1.78652558, Global Avg Loss: 2.21471882, Time: 0.0402 Steps: 14710, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001946, Sample Num: 31136, Cur Loss: 1.38898528, Cur Avg Loss: 1.60105777, Log Avg loss: 1.63138161, Global Avg Loss: 2.21432253, Time: 0.0402 Steps: 14720, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001956, Sample Num: 31296, Cur Loss: 1.16004515, Cur Avg Loss: 1.60107237, Log Avg loss: 1.60391312, Global Avg Loss: 2.21390813, Time: 0.0403 Steps: 14730, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001966, Sample Num: 31456, Cur Loss: 1.66601753, Cur Avg Loss: 1.60114576, Log Avg loss: 1.61550205, Global Avg Loss: 2.21350216, Time: 0.0403 Steps: 14740, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001976, Sample Num: 31616, Cur Loss: 1.38536036, Cur Avg Loss: 1.60092236, Log Avg loss: 1.55700163, Global Avg Loss: 2.21305707, Time: 0.0402 Steps: 14750, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001986, Sample Num: 31776, Cur Loss: 1.12340271, Cur Avg Loss: 1.60199549, Log Avg loss: 1.81404636, Global Avg Loss: 2.21278674, Time: 0.0403 Steps: 14760, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001996, Sample Num: 31936, Cur Loss: 1.24204922, Cur Avg Loss: 1.60163197, Log Avg loss: 1.52943716, Global Avg Loss: 2.21232408, Time: 0.0402 Steps: 14770, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002006, Sample Num: 32096, Cur Loss: 1.02213287, Cur Avg Loss: 1.60219874, Log Avg loss: 1.71532443, Global Avg Loss: 2.21198782, Time: 0.0402 Steps: 14780, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002016, Sample Num: 32256, Cur Loss: 2.13426328, Cur Avg Loss: 1.60231890, Log Avg loss: 1.62642401, Global Avg Loss: 2.21159190, Time: 0.0402 Steps: 14790, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002026, Sample Num: 32416, Cur Loss: 1.78413081, Cur Avg Loss: 1.60260124, Log Avg loss: 1.65952117, Global Avg Loss: 2.21121888, Time: 0.0402 Steps: 14800, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002036, Sample Num: 32576, Cur Loss: 1.16944146, Cur Avg Loss: 1.60208380, Log Avg loss: 1.49724914, Global Avg Loss: 2.21073679, Time: 0.0403 Steps: 14810, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002046, Sample Num: 32736, Cur Loss: 0.73176992, Cur Avg Loss: 1.60163286, Log Avg loss: 1.50982296, Global Avg Loss: 2.21026384, Time: 0.0402 Steps: 14820, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002056, Sample Num: 32896, Cur Loss: 1.37884402, Cur Avg Loss: 1.60076696, Log Avg loss: 1.42360293, Global Avg Loss: 2.20973339, Time: 0.0402 Steps: 14830, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002066, Sample Num: 33056, Cur Loss: 2.03703904, Cur Avg Loss: 1.60027899, Log Avg loss: 1.49995242, Global Avg Loss: 2.20925510, Time: 0.0488 Steps: 14840, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002076, Sample Num: 33216, Cur Loss: 1.38865352, Cur Avg Loss: 1.59980400, Log Avg loss: 1.50167157, Global Avg Loss: 2.20877861, Time: 0.0640 Steps: 14850, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002086, Sample Num: 33376, Cur Loss: 1.22939444, Cur Avg Loss: 1.59681855, Log Avg loss: 0.97703900, Global Avg Loss: 2.20794971, Time: 0.0481 Steps: 14860, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002096, Sample Num: 33536, Cur Loss: 2.06744719, Cur Avg Loss: 1.59836614, Log Avg loss: 1.92119337, Global Avg Loss: 2.20775687, Time: 0.0405 Steps: 14870, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002106, Sample Num: 33696, Cur Loss: 1.49560809, Cur Avg Loss: 1.59959895, Log Avg loss: 1.85799618, Global Avg Loss: 2.20752182, Time: 0.0404 Steps: 14880, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002116, Sample Num: 33856, Cur Loss: 1.88375843, Cur Avg Loss: 1.60155959, Log Avg loss: 2.01447092, Global Avg Loss: 2.20739217, Time: 0.0666 Steps: 14890, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002126, Sample Num: 34016, Cur Loss: 1.76375818, Cur Avg Loss: 1.60116486, Log Avg loss: 1.51763817, Global Avg Loss: 2.20692924, Time: 0.0507 Steps: 14900, Updated lr: 0.000087 ***** Running evaluation checkpoint-14903 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-14903 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.928875, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.433221, "eval_total_loss": 1007.554287, "eval_mae": 1.028285, "eval_mse": 1.433286, "eval_r2": 0.088909, "eval_sp_statistic": 0.293626, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.343759, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.509631, "test_total_loss": 757.834952, "test_mae": 1.082312, "test_mse": 1.509823, "test_r2": 0.025547, "test_sp_statistic": 0.300368, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.351031, "test_ps_pvalue": 0.0, "lr": 8.681555239449977e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.2068154055562457, "train_cur_epoch_loss": 3409.000721499324, "train_cur_epoch_avg_loss": 1.6012215695158871, "train_cur_epoch_time": 95.92887473106384, "train_cur_epoch_avg_time": 0.045058184467385555, "epoch": 7, "step": 14903} ################################################## Training, Epoch: 0008, Batch: 000007, Sample Num: 112, Cur Loss: 1.95513272, Cur Avg Loss: 1.81542884, Log Avg loss: 1.76322412, Global Avg Loss: 2.20663166, Time: 0.0455 Steps: 14910, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000017, Sample Num: 272, Cur Loss: 0.96749711, Cur Avg Loss: 1.67101958, Log Avg loss: 1.56993309, Global Avg Loss: 2.20620491, Time: 0.0404 Steps: 14920, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000027, Sample Num: 432, Cur Loss: 1.84299612, Cur Avg Loss: 1.67111274, Log Avg loss: 1.67127111, Global Avg Loss: 2.20584662, Time: 0.0421 Steps: 14930, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000037, Sample Num: 592, Cur Loss: 0.43592873, Cur Avg Loss: 1.63487016, Log Avg loss: 1.53701522, Global Avg Loss: 2.20539894, Time: 0.0481 Steps: 14940, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000047, Sample Num: 752, Cur Loss: 1.54082716, Cur Avg Loss: 1.68402222, Log Avg loss: 1.86588484, Global Avg Loss: 2.20517184, Time: 0.0631 Steps: 14950, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000057, Sample Num: 912, Cur Loss: 2.11069083, Cur Avg Loss: 1.66649947, Log Avg loss: 1.58414251, Global Avg Loss: 2.20475672, Time: 0.0405 Steps: 14960, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000067, Sample Num: 1072, Cur Loss: 1.55183125, Cur Avg Loss: 1.65912069, Log Avg loss: 1.61706165, Global Avg Loss: 2.20436413, Time: 0.0480 Steps: 14970, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000077, Sample Num: 1232, Cur Loss: 0.59715271, Cur Avg Loss: 1.64260173, Log Avg loss: 1.53192468, Global Avg Loss: 2.20391524, Time: 0.0598 Steps: 14980, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000087, Sample Num: 1392, Cur Loss: 1.19699955, Cur Avg Loss: 1.66875959, Log Avg loss: 1.87017517, Global Avg Loss: 2.20369260, Time: 0.0619 Steps: 14990, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000097, Sample Num: 1552, Cur Loss: 1.52022254, Cur Avg Loss: 1.66799679, Log Avg loss: 1.66136041, Global Avg Loss: 2.20333105, Time: 0.0712 Steps: 15000, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000107, Sample Num: 1712, Cur Loss: 0.45443058, Cur Avg Loss: 1.64666897, Log Avg loss: 1.43978909, Global Avg Loss: 2.20282236, Time: 0.0403 Steps: 15010, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000117, Sample Num: 1872, Cur Loss: 1.43893814, Cur Avg Loss: 1.62419746, Log Avg loss: 1.38375228, Global Avg Loss: 2.20227704, Time: 0.0402 Steps: 15020, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000127, Sample Num: 2032, Cur Loss: 2.35768294, Cur Avg Loss: 1.62754758, Log Avg loss: 1.66674405, Global Avg Loss: 2.20192073, Time: 0.0402 Steps: 15030, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000137, Sample Num: 2192, Cur Loss: 1.73088121, Cur Avg Loss: 1.60889043, Log Avg loss: 1.37194453, Global Avg Loss: 2.20136888, Time: 0.0403 Steps: 15040, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000147, Sample Num: 2352, Cur Loss: 1.10541868, Cur Avg Loss: 1.58679992, Log Avg loss: 1.28416000, Global Avg Loss: 2.20075944, Time: 0.0403 Steps: 15050, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000157, Sample Num: 2512, Cur Loss: 1.95603240, Cur Avg Loss: 1.59376611, Log Avg loss: 1.69616914, Global Avg Loss: 2.20042439, Time: 0.0402 Steps: 15060, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000167, Sample Num: 2672, Cur Loss: 0.78904480, Cur Avg Loss: 1.56460867, Log Avg loss: 1.10683674, Global Avg Loss: 2.19969872, Time: 0.0402 Steps: 15070, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000177, Sample Num: 2832, Cur Loss: 1.01897454, Cur Avg Loss: 1.56824335, Log Avg loss: 1.62894267, Global Avg Loss: 2.19932023, Time: 0.0403 Steps: 15080, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000187, Sample Num: 2992, Cur Loss: 1.54676688, Cur Avg Loss: 1.57188860, Log Avg loss: 1.63640949, Global Avg Loss: 2.19894719, Time: 0.0402 Steps: 15090, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000197, Sample Num: 3152, Cur Loss: 0.77199674, Cur Avg Loss: 1.55022709, Log Avg loss: 1.14515674, Global Avg Loss: 2.19824932, Time: 0.0403 Steps: 15100, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000207, Sample Num: 3312, Cur Loss: 0.84556371, Cur Avg Loss: 1.55417728, Log Avg loss: 1.63199614, Global Avg Loss: 2.19787457, Time: 0.0402 Steps: 15110, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000217, Sample Num: 3472, Cur Loss: 2.53335953, Cur Avg Loss: 1.55010642, Log Avg loss: 1.46583947, Global Avg Loss: 2.19739042, Time: 0.0402 Steps: 15120, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000227, Sample Num: 3632, Cur Loss: 0.53233176, Cur Avg Loss: 1.54810888, Log Avg loss: 1.50476239, Global Avg Loss: 2.19693263, Time: 0.0402 Steps: 15130, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000237, Sample Num: 3792, Cur Loss: 0.81116652, Cur Avg Loss: 1.55459468, Log Avg loss: 1.70182239, Global Avg Loss: 2.19660561, Time: 0.0403 Steps: 15140, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000247, Sample Num: 3952, Cur Loss: 1.50612271, Cur Avg Loss: 1.54287249, Log Avg loss: 1.26505660, Global Avg Loss: 2.19599073, Time: 0.0403 Steps: 15150, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000257, Sample Num: 4112, Cur Loss: 1.46244895, Cur Avg Loss: 1.55051776, Log Avg loss: 1.73935583, Global Avg Loss: 2.19568952, Time: 0.0403 Steps: 15160, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000267, Sample Num: 4272, Cur Loss: 1.11685741, Cur Avg Loss: 1.55143971, Log Avg loss: 1.57513380, Global Avg Loss: 2.19528045, Time: 0.0402 Steps: 15170, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000277, Sample Num: 4432, Cur Loss: 1.20374799, Cur Avg Loss: 1.54780384, Log Avg loss: 1.45072610, Global Avg Loss: 2.19478996, Time: 0.0402 Steps: 15180, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000287, Sample Num: 4592, Cur Loss: 1.30188334, Cur Avg Loss: 1.54824536, Log Avg loss: 1.56047565, Global Avg Loss: 2.19437238, Time: 0.0403 Steps: 15190, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000297, Sample Num: 4752, Cur Loss: 1.70023155, Cur Avg Loss: 1.54821213, Log Avg loss: 1.54725838, Global Avg Loss: 2.19394664, Time: 0.0402 Steps: 15200, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000307, Sample Num: 4912, Cur Loss: 1.58061528, Cur Avg Loss: 1.55490930, Log Avg loss: 1.75381531, Global Avg Loss: 2.19365727, Time: 0.0402 Steps: 15210, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000317, Sample Num: 5072, Cur Loss: 0.60928261, Cur Avg Loss: 1.55839722, Log Avg loss: 1.66547629, Global Avg Loss: 2.19331024, Time: 0.0402 Steps: 15220, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000327, Sample Num: 5232, Cur Loss: 1.17275023, Cur Avg Loss: 1.56274009, Log Avg loss: 1.70040895, Global Avg Loss: 2.19298661, Time: 0.0402 Steps: 15230, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000337, Sample Num: 5392, Cur Loss: 1.51612186, Cur Avg Loss: 1.56133495, Log Avg loss: 1.51538687, Global Avg Loss: 2.19254199, Time: 0.0401 Steps: 15240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000347, Sample Num: 5552, Cur Loss: 2.20394421, Cur Avg Loss: 1.56684445, Log Avg loss: 1.75251451, Global Avg Loss: 2.19225344, Time: 0.0402 Steps: 15250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000357, Sample Num: 5712, Cur Loss: 1.61433399, Cur Avg Loss: 1.56809938, Log Avg loss: 1.61164547, Global Avg Loss: 2.19187297, Time: 0.0402 Steps: 15260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000367, Sample Num: 5872, Cur Loss: 2.09897852, Cur Avg Loss: 1.57174435, Log Avg loss: 1.70187002, Global Avg Loss: 2.19155207, Time: 0.0664 Steps: 15270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000377, Sample Num: 6032, Cur Loss: 0.85286117, Cur Avg Loss: 1.56871346, Log Avg loss: 1.45747981, Global Avg Loss: 2.19107166, Time: 0.0431 Steps: 15280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000387, Sample Num: 6192, Cur Loss: 1.33888030, Cur Avg Loss: 1.56008001, Log Avg loss: 1.23459881, Global Avg Loss: 2.19044611, Time: 0.0449 Steps: 15290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000397, Sample Num: 6352, Cur Loss: 0.89771283, Cur Avg Loss: 1.55136369, Log Avg loss: 1.21404209, Global Avg Loss: 2.18980793, Time: 0.0723 Steps: 15300, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000407, Sample Num: 6512, Cur Loss: 1.35583317, Cur Avg Loss: 1.55210143, Log Avg loss: 1.58138964, Global Avg Loss: 2.18941053, Time: 0.0662 Steps: 15310, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000417, Sample Num: 6672, Cur Loss: 2.75563765, Cur Avg Loss: 1.55390358, Log Avg loss: 1.62725123, Global Avg Loss: 2.18904359, Time: 0.0730 Steps: 15320, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000427, Sample Num: 6832, Cur Loss: 0.60947907, Cur Avg Loss: 1.54448402, Log Avg loss: 1.15168852, Global Avg Loss: 2.18836691, Time: 0.0555 Steps: 15330, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000437, Sample Num: 6992, Cur Loss: 1.73967957, Cur Avg Loss: 1.54332694, Log Avg loss: 1.49391953, Global Avg Loss: 2.18791420, Time: 0.0404 Steps: 15340, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000447, Sample Num: 7152, Cur Loss: 1.67029440, Cur Avg Loss: 1.54229567, Log Avg loss: 1.49722906, Global Avg Loss: 2.18746424, Time: 0.0402 Steps: 15350, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000457, Sample Num: 7312, Cur Loss: 1.12477982, Cur Avg Loss: 1.54622219, Log Avg loss: 1.72173786, Global Avg Loss: 2.18716104, Time: 0.0402 Steps: 15360, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000467, Sample Num: 7472, Cur Loss: 2.31304693, Cur Avg Loss: 1.55378228, Log Avg loss: 1.89927837, Global Avg Loss: 2.18697374, Time: 0.0402 Steps: 15370, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000477, Sample Num: 7632, Cur Loss: 2.31124449, Cur Avg Loss: 1.55333163, Log Avg loss: 1.53228595, Global Avg Loss: 2.18654806, Time: 0.0403 Steps: 15380, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000487, Sample Num: 7792, Cur Loss: 0.75654435, Cur Avg Loss: 1.55534979, Log Avg loss: 1.65161600, Global Avg Loss: 2.18620048, Time: 0.0402 Steps: 15390, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000497, Sample Num: 7952, Cur Loss: 1.31465042, Cur Avg Loss: 1.55488320, Log Avg loss: 1.53216052, Global Avg Loss: 2.18577578, Time: 0.0402 Steps: 15400, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000507, Sample Num: 8112, Cur Loss: 2.92843080, Cur Avg Loss: 1.55804120, Log Avg loss: 1.71499355, Global Avg Loss: 2.18547027, Time: 0.0402 Steps: 15410, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000517, Sample Num: 8272, Cur Loss: 0.44769055, Cur Avg Loss: 1.55286512, Log Avg loss: 1.29043789, Global Avg Loss: 2.18488983, Time: 0.0402 Steps: 15420, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000527, Sample Num: 8432, Cur Loss: 1.35435343, Cur Avg Loss: 1.55785577, Log Avg loss: 1.81587255, Global Avg Loss: 2.18465068, Time: 0.0402 Steps: 15430, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000537, Sample Num: 8592, Cur Loss: 0.89539289, Cur Avg Loss: 1.55579464, Log Avg loss: 1.44717326, Global Avg Loss: 2.18417304, Time: 0.0402 Steps: 15440, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000547, Sample Num: 8752, Cur Loss: 1.34930038, Cur Avg Loss: 1.55533791, Log Avg loss: 1.53081139, Global Avg Loss: 2.18375015, Time: 0.0403 Steps: 15450, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000557, Sample Num: 8912, Cur Loss: 1.18571138, Cur Avg Loss: 1.55382907, Log Avg loss: 1.47129536, Global Avg Loss: 2.18328931, Time: 0.0402 Steps: 15460, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000567, Sample Num: 9072, Cur Loss: 1.92555428, Cur Avg Loss: 1.55722675, Log Avg loss: 1.74647745, Global Avg Loss: 2.18300695, Time: 0.0402 Steps: 15470, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000577, Sample Num: 9232, Cur Loss: 2.42777824, Cur Avg Loss: 1.55778599, Log Avg loss: 1.58949475, Global Avg Loss: 2.18262355, Time: 0.0402 Steps: 15480, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000587, Sample Num: 9392, Cur Loss: 1.79296935, Cur Avg Loss: 1.55427067, Log Avg loss: 1.35143677, Global Avg Loss: 2.18208695, Time: 0.0402 Steps: 15490, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000597, Sample Num: 9552, Cur Loss: 1.13358796, Cur Avg Loss: 1.55360216, Log Avg loss: 1.51436069, Global Avg Loss: 2.18165616, Time: 0.0402 Steps: 15500, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000607, Sample Num: 9712, Cur Loss: 2.06069231, Cur Avg Loss: 1.55273765, Log Avg loss: 1.50112650, Global Avg Loss: 2.18121739, Time: 0.0401 Steps: 15510, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000617, Sample Num: 9872, Cur Loss: 1.30659890, Cur Avg Loss: 1.55162263, Log Avg loss: 1.48394096, Global Avg Loss: 2.18076812, Time: 0.0402 Steps: 15520, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000627, Sample Num: 10032, Cur Loss: 1.71191537, Cur Avg Loss: 1.55092688, Log Avg loss: 1.50799885, Global Avg Loss: 2.18033491, Time: 0.0402 Steps: 15530, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000637, Sample Num: 10192, Cur Loss: 0.80265498, Cur Avg Loss: 1.54583283, Log Avg loss: 1.22643639, Global Avg Loss: 2.17972107, Time: 0.0402 Steps: 15540, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000647, Sample Num: 10352, Cur Loss: 2.41387296, Cur Avg Loss: 1.54363065, Log Avg loss: 1.40335135, Global Avg Loss: 2.17922180, Time: 0.0402 Steps: 15550, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000657, Sample Num: 10512, Cur Loss: 1.97680855, Cur Avg Loss: 1.54297087, Log Avg loss: 1.50028323, Global Avg Loss: 2.17878547, Time: 0.0402 Steps: 15560, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000667, Sample Num: 10672, Cur Loss: 1.10153139, Cur Avg Loss: 1.54368005, Log Avg loss: 1.59027338, Global Avg Loss: 2.17840749, Time: 0.0402 Steps: 15570, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000677, Sample Num: 10832, Cur Loss: 2.75328636, Cur Avg Loss: 1.54451756, Log Avg loss: 1.60037923, Global Avg Loss: 2.17803648, Time: 0.0402 Steps: 15580, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000687, Sample Num: 10992, Cur Loss: 1.28730512, Cur Avg Loss: 1.54376895, Log Avg loss: 1.49308796, Global Avg Loss: 2.17759713, Time: 0.0402 Steps: 15590, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000697, Sample Num: 11152, Cur Loss: 0.49961567, Cur Avg Loss: 1.53740350, Log Avg loss: 1.10009716, Global Avg Loss: 2.17690642, Time: 0.0402 Steps: 15600, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000707, Sample Num: 11312, Cur Loss: 1.82430255, Cur Avg Loss: 1.53548982, Log Avg loss: 1.40210628, Global Avg Loss: 2.17641008, Time: 0.0402 Steps: 15610, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000717, Sample Num: 11472, Cur Loss: 2.62641478, Cur Avg Loss: 1.53518695, Log Avg loss: 1.51377448, Global Avg Loss: 2.17598585, Time: 0.0677 Steps: 15620, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000727, Sample Num: 11632, Cur Loss: 1.80238032, Cur Avg Loss: 1.53689256, Log Avg loss: 1.65918483, Global Avg Loss: 2.17565521, Time: 0.0796 Steps: 15630, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000737, Sample Num: 11792, Cur Loss: 3.14246178, Cur Avg Loss: 1.53580794, Log Avg loss: 1.45695574, Global Avg Loss: 2.17519568, Time: 0.0696 Steps: 15640, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000747, Sample Num: 11952, Cur Loss: 1.42979968, Cur Avg Loss: 1.53894390, Log Avg loss: 1.77006456, Global Avg Loss: 2.17493681, Time: 0.0994 Steps: 15650, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000757, Sample Num: 12112, Cur Loss: 1.55462849, Cur Avg Loss: 1.54209451, Log Avg loss: 1.77744490, Global Avg Loss: 2.17468298, Time: 0.0799 Steps: 15660, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000767, Sample Num: 12272, Cur Loss: 1.85787642, Cur Avg Loss: 1.53721568, Log Avg loss: 1.16788812, Global Avg Loss: 2.17404049, Time: 0.0560 Steps: 15670, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000777, Sample Num: 12432, Cur Loss: 1.19155169, Cur Avg Loss: 1.53726892, Log Avg loss: 1.54135201, Global Avg Loss: 2.17363699, Time: 0.0668 Steps: 15680, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000787, Sample Num: 12592, Cur Loss: 1.86967814, Cur Avg Loss: 1.53620627, Log Avg loss: 1.45363858, Global Avg Loss: 2.17317810, Time: 0.0662 Steps: 15690, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000797, Sample Num: 12752, Cur Loss: 0.52767855, Cur Avg Loss: 1.53663581, Log Avg loss: 1.57044099, Global Avg Loss: 2.17279419, Time: 0.0456 Steps: 15700, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000807, Sample Num: 12912, Cur Loss: 1.89990449, Cur Avg Loss: 1.54257980, Log Avg loss: 2.01631585, Global Avg Loss: 2.17269458, Time: 0.0706 Steps: 15710, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000817, Sample Num: 13072, Cur Loss: 3.36284328, Cur Avg Loss: 1.54494125, Log Avg loss: 1.73551000, Global Avg Loss: 2.17241648, Time: 0.0402 Steps: 15720, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000827, Sample Num: 13232, Cur Loss: 0.76603711, Cur Avg Loss: 1.54227303, Log Avg loss: 1.32427968, Global Avg Loss: 2.17187729, Time: 0.0402 Steps: 15730, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000837, Sample Num: 13392, Cur Loss: 1.36233306, Cur Avg Loss: 1.54029040, Log Avg loss: 1.37632640, Global Avg Loss: 2.17137186, Time: 0.0403 Steps: 15740, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000847, Sample Num: 13552, Cur Loss: 2.58762479, Cur Avg Loss: 1.54041760, Log Avg loss: 1.55106407, Global Avg Loss: 2.17097801, Time: 0.0403 Steps: 15750, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000857, Sample Num: 13712, Cur Loss: 0.85829896, Cur Avg Loss: 1.54214783, Log Avg loss: 1.68869904, Global Avg Loss: 2.17067200, Time: 0.0402 Steps: 15760, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000867, Sample Num: 13872, Cur Loss: 1.69414949, Cur Avg Loss: 1.53929249, Log Avg loss: 1.29458969, Global Avg Loss: 2.17011646, Time: 0.0402 Steps: 15770, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000877, Sample Num: 14032, Cur Loss: 1.75333166, Cur Avg Loss: 1.54256105, Log Avg loss: 1.82594537, Global Avg Loss: 2.16989835, Time: 0.0402 Steps: 15780, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000887, Sample Num: 14192, Cur Loss: 1.29904771, Cur Avg Loss: 1.54286887, Log Avg loss: 1.56986411, Global Avg Loss: 2.16951835, Time: 0.0402 Steps: 15790, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000897, Sample Num: 14352, Cur Loss: 2.17472506, Cur Avg Loss: 1.54530991, Log Avg loss: 1.76183012, Global Avg Loss: 2.16926031, Time: 0.0403 Steps: 15800, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000907, Sample Num: 14512, Cur Loss: 1.39296305, Cur Avg Loss: 1.55444092, Log Avg loss: 2.37349294, Global Avg Loss: 2.16938949, Time: 0.0402 Steps: 15810, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000917, Sample Num: 14672, Cur Loss: 1.33967447, Cur Avg Loss: 1.55550016, Log Avg loss: 1.65157304, Global Avg Loss: 2.16906218, Time: 0.0402 Steps: 15820, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000927, Sample Num: 14832, Cur Loss: 1.59991121, Cur Avg Loss: 1.55279428, Log Avg loss: 1.30466483, Global Avg Loss: 2.16851613, Time: 0.0402 Steps: 15830, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000937, Sample Num: 14992, Cur Loss: 1.36157203, Cur Avg Loss: 1.55317074, Log Avg loss: 1.58806916, Global Avg Loss: 2.16814968, Time: 0.0402 Steps: 15840, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000947, Sample Num: 15152, Cur Loss: 1.74647832, Cur Avg Loss: 1.55339265, Log Avg loss: 1.57418528, Global Avg Loss: 2.16777494, Time: 0.0402 Steps: 15850, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000957, Sample Num: 15312, Cur Loss: 2.47058773, Cur Avg Loss: 1.55264399, Log Avg loss: 1.48174614, Global Avg Loss: 2.16734239, Time: 0.0402 Steps: 15860, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000967, Sample Num: 15472, Cur Loss: 0.85201538, Cur Avg Loss: 1.55171915, Log Avg loss: 1.46321135, Global Avg Loss: 2.16689870, Time: 0.0402 Steps: 15870, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000977, Sample Num: 15632, Cur Loss: 1.44307911, Cur Avg Loss: 1.55005002, Log Avg loss: 1.38864594, Global Avg Loss: 2.16640862, Time: 0.0402 Steps: 15880, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000987, Sample Num: 15792, Cur Loss: 1.98842418, Cur Avg Loss: 1.54846675, Log Avg loss: 1.39378136, Global Avg Loss: 2.16592238, Time: 0.0402 Steps: 15890, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000997, Sample Num: 15952, Cur Loss: 2.19802642, Cur Avg Loss: 1.55015418, Log Avg loss: 1.71670266, Global Avg Loss: 2.16563986, Time: 0.0402 Steps: 15900, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001007, Sample Num: 16112, Cur Loss: 0.93433422, Cur Avg Loss: 1.54893452, Log Avg loss: 1.42733496, Global Avg Loss: 2.16517580, Time: 0.0402 Steps: 15910, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001017, Sample Num: 16272, Cur Loss: 0.86870044, Cur Avg Loss: 1.55012410, Log Avg loss: 1.66991518, Global Avg Loss: 2.16486471, Time: 0.0402 Steps: 15920, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001027, Sample Num: 16432, Cur Loss: 2.23394203, Cur Avg Loss: 1.55249290, Log Avg loss: 1.79339907, Global Avg Loss: 2.16463153, Time: 0.0402 Steps: 15930, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001037, Sample Num: 16592, Cur Loss: 1.74714053, Cur Avg Loss: 1.55166349, Log Avg loss: 1.46648296, Global Avg Loss: 2.16419354, Time: 0.0402 Steps: 15940, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001047, Sample Num: 16752, Cur Loss: 1.39119565, Cur Avg Loss: 1.54925821, Log Avg loss: 1.29983074, Global Avg Loss: 2.16365162, Time: 0.0402 Steps: 15950, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001057, Sample Num: 16912, Cur Loss: 1.21161938, Cur Avg Loss: 1.55024609, Log Avg loss: 1.65367750, Global Avg Loss: 2.16333209, Time: 0.0402 Steps: 15960, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001067, Sample Num: 17072, Cur Loss: 1.28309262, Cur Avg Loss: 1.55274774, Log Avg loss: 1.81717194, Global Avg Loss: 2.16311533, Time: 0.0537 Steps: 15970, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001077, Sample Num: 17232, Cur Loss: 1.73877001, Cur Avg Loss: 1.55292129, Log Avg loss: 1.57143888, Global Avg Loss: 2.16274507, Time: 0.0687 Steps: 15980, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001087, Sample Num: 17392, Cur Loss: 1.35355103, Cur Avg Loss: 1.55313325, Log Avg loss: 1.57596139, Global Avg Loss: 2.16237810, Time: 0.0405 Steps: 15990, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001097, Sample Num: 17552, Cur Loss: 1.24117970, Cur Avg Loss: 1.55248979, Log Avg loss: 1.48254605, Global Avg Loss: 2.16195321, Time: 0.0474 Steps: 16000, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001107, Sample Num: 17712, Cur Loss: 2.45551968, Cur Avg Loss: 1.55283350, Log Avg loss: 1.59053879, Global Avg Loss: 2.16159629, Time: 0.0405 Steps: 16010, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001117, Sample Num: 17872, Cur Loss: 0.76911497, Cur Avg Loss: 1.55203336, Log Avg loss: 1.46345701, Global Avg Loss: 2.16116050, Time: 0.0514 Steps: 16020, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001127, Sample Num: 18032, Cur Loss: 2.57772112, Cur Avg Loss: 1.54986565, Log Avg loss: 1.30773329, Global Avg Loss: 2.16062811, Time: 0.0898 Steps: 16030, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001137, Sample Num: 18192, Cur Loss: 1.76379561, Cur Avg Loss: 1.54856942, Log Avg loss: 1.40248423, Global Avg Loss: 2.16015545, Time: 0.0679 Steps: 16040, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001147, Sample Num: 18352, Cur Loss: 2.08855057, Cur Avg Loss: 1.54877487, Log Avg loss: 1.57213418, Global Avg Loss: 2.15978908, Time: 0.0656 Steps: 16050, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001157, Sample Num: 18512, Cur Loss: 0.79255164, Cur Avg Loss: 1.54949217, Log Avg loss: 1.63176677, Global Avg Loss: 2.15946030, Time: 0.0403 Steps: 16060, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001167, Sample Num: 18672, Cur Loss: 1.36176205, Cur Avg Loss: 1.55132484, Log Avg loss: 1.76336488, Global Avg Loss: 2.15921382, Time: 0.0402 Steps: 16070, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001177, Sample Num: 18832, Cur Loss: 0.68359733, Cur Avg Loss: 1.54919595, Log Avg loss: 1.30075360, Global Avg Loss: 2.15867995, Time: 0.0403 Steps: 16080, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001187, Sample Num: 18992, Cur Loss: 1.88817334, Cur Avg Loss: 1.55035865, Log Avg loss: 1.68720927, Global Avg Loss: 2.15838693, Time: 0.0402 Steps: 16090, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001197, Sample Num: 19152, Cur Loss: 1.00858593, Cur Avg Loss: 1.55153037, Log Avg loss: 1.69061313, Global Avg Loss: 2.15809639, Time: 0.0403 Steps: 16100, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001207, Sample Num: 19312, Cur Loss: 2.24018335, Cur Avg Loss: 1.55179800, Log Avg loss: 1.58383399, Global Avg Loss: 2.15773992, Time: 0.0402 Steps: 16110, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001217, Sample Num: 19472, Cur Loss: 1.03257549, Cur Avg Loss: 1.55075451, Log Avg loss: 1.42480475, Global Avg Loss: 2.15728525, Time: 0.0403 Steps: 16120, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001227, Sample Num: 19632, Cur Loss: 1.86360526, Cur Avg Loss: 1.55200259, Log Avg loss: 1.70389434, Global Avg Loss: 2.15700416, Time: 0.0402 Steps: 16130, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001237, Sample Num: 19792, Cur Loss: 1.06528878, Cur Avg Loss: 1.55362295, Log Avg loss: 1.75244057, Global Avg Loss: 2.15675351, Time: 0.0402 Steps: 16140, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001247, Sample Num: 19952, Cur Loss: 1.94212437, Cur Avg Loss: 1.55214016, Log Avg loss: 1.36871948, Global Avg Loss: 2.15626556, Time: 0.0402 Steps: 16150, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001257, Sample Num: 20112, Cur Loss: 2.24868679, Cur Avg Loss: 1.55320303, Log Avg loss: 1.68574249, Global Avg Loss: 2.15597439, Time: 0.0402 Steps: 16160, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001267, Sample Num: 20272, Cur Loss: 1.42121613, Cur Avg Loss: 1.55507093, Log Avg loss: 1.78986554, Global Avg Loss: 2.15574798, Time: 0.0402 Steps: 16170, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001277, Sample Num: 20432, Cur Loss: 1.67162561, Cur Avg Loss: 1.55412681, Log Avg loss: 1.43450764, Global Avg Loss: 2.15530222, Time: 0.0402 Steps: 16180, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001287, Sample Num: 20592, Cur Loss: 1.45640433, Cur Avg Loss: 1.55578059, Log Avg loss: 1.76696817, Global Avg Loss: 2.15506236, Time: 0.0402 Steps: 16190, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001297, Sample Num: 20752, Cur Loss: 1.42522502, Cur Avg Loss: 1.55450781, Log Avg loss: 1.39070070, Global Avg Loss: 2.15459053, Time: 0.0402 Steps: 16200, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001307, Sample Num: 20912, Cur Loss: 1.57232797, Cur Avg Loss: 1.55497898, Log Avg loss: 1.61609019, Global Avg Loss: 2.15425833, Time: 0.0402 Steps: 16210, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001317, Sample Num: 21072, Cur Loss: 0.85181409, Cur Avg Loss: 1.55323658, Log Avg loss: 1.32550417, Global Avg Loss: 2.15374738, Time: 0.0402 Steps: 16220, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001327, Sample Num: 21232, Cur Loss: 1.37546587, Cur Avg Loss: 1.55183109, Log Avg loss: 1.36672834, Global Avg Loss: 2.15326247, Time: 0.0402 Steps: 16230, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001337, Sample Num: 21392, Cur Loss: 1.36141741, Cur Avg Loss: 1.54929960, Log Avg loss: 1.21337053, Global Avg Loss: 2.15268372, Time: 0.0402 Steps: 16240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001347, Sample Num: 21552, Cur Loss: 1.69573569, Cur Avg Loss: 1.55345777, Log Avg loss: 2.10940619, Global Avg Loss: 2.15265708, Time: 0.0402 Steps: 16250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001357, Sample Num: 21712, Cur Loss: 0.59949195, Cur Avg Loss: 1.55170069, Log Avg loss: 1.31502141, Global Avg Loss: 2.15214193, Time: 0.0402 Steps: 16260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001367, Sample Num: 21872, Cur Loss: 1.81623173, Cur Avg Loss: 1.55356953, Log Avg loss: 1.80717095, Global Avg Loss: 2.15192990, Time: 0.0402 Steps: 16270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001377, Sample Num: 22032, Cur Loss: 2.14291334, Cur Avg Loss: 1.55485502, Log Avg loss: 1.73058149, Global Avg Loss: 2.15167109, Time: 0.0402 Steps: 16280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001387, Sample Num: 22192, Cur Loss: 0.72174895, Cur Avg Loss: 1.55383187, Log Avg loss: 1.41294456, Global Avg Loss: 2.15121761, Time: 0.0402 Steps: 16290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001397, Sample Num: 22352, Cur Loss: 2.83870316, Cur Avg Loss: 1.55363989, Log Avg loss: 1.52701247, Global Avg Loss: 2.15083466, Time: 0.0402 Steps: 16300, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001407, Sample Num: 22512, Cur Loss: 1.25147474, Cur Avg Loss: 1.55433406, Log Avg loss: 1.65130879, Global Avg Loss: 2.15052839, Time: 0.0560 Steps: 16310, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001417, Sample Num: 22672, Cur Loss: 1.46033680, Cur Avg Loss: 1.55570084, Log Avg loss: 1.74800680, Global Avg Loss: 2.15028174, Time: 0.0403 Steps: 16320, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001427, Sample Num: 22832, Cur Loss: 1.29206002, Cur Avg Loss: 1.55428765, Log Avg loss: 1.35403837, Global Avg Loss: 2.14979415, Time: 0.0434 Steps: 16330, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001437, Sample Num: 22992, Cur Loss: 1.07186437, Cur Avg Loss: 1.55571671, Log Avg loss: 1.75964478, Global Avg Loss: 2.14955538, Time: 0.0982 Steps: 16340, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001447, Sample Num: 23152, Cur Loss: 1.49346375, Cur Avg Loss: 1.55442542, Log Avg loss: 1.36886710, Global Avg Loss: 2.14907789, Time: 0.0527 Steps: 16350, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001457, Sample Num: 23312, Cur Loss: 1.49043500, Cur Avg Loss: 1.55330944, Log Avg loss: 1.39182678, Global Avg Loss: 2.14861503, Time: 0.0735 Steps: 16360, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001467, Sample Num: 23472, Cur Loss: 1.90768456, Cur Avg Loss: 1.55638377, Log Avg loss: 2.00431367, Global Avg Loss: 2.14852688, Time: 0.0604 Steps: 16370, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001477, Sample Num: 23632, Cur Loss: 2.81748223, Cur Avg Loss: 1.55764829, Log Avg loss: 1.74315314, Global Avg Loss: 2.14827940, Time: 0.0443 Steps: 16380, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001487, Sample Num: 23792, Cur Loss: 0.89663702, Cur Avg Loss: 1.55492634, Log Avg loss: 1.15289376, Global Avg Loss: 2.14767208, Time: 0.0402 Steps: 16390, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001497, Sample Num: 23952, Cur Loss: 1.00898647, Cur Avg Loss: 1.55677063, Log Avg loss: 1.83101678, Global Avg Loss: 2.14747900, Time: 0.0402 Steps: 16400, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001507, Sample Num: 24112, Cur Loss: 0.52357936, Cur Avg Loss: 1.55345544, Log Avg loss: 1.05717210, Global Avg Loss: 2.14681459, Time: 0.0402 Steps: 16410, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001517, Sample Num: 24272, Cur Loss: 1.34278917, Cur Avg Loss: 1.55264549, Log Avg loss: 1.43058607, Global Avg Loss: 2.14637839, Time: 0.0402 Steps: 16420, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001527, Sample Num: 24432, Cur Loss: 1.04067898, Cur Avg Loss: 1.55253493, Log Avg loss: 1.53576192, Global Avg Loss: 2.14600675, Time: 0.0402 Steps: 16430, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001537, Sample Num: 24592, Cur Loss: 1.21462369, Cur Avg Loss: 1.55302676, Log Avg loss: 1.62813061, Global Avg Loss: 2.14569174, Time: 0.0403 Steps: 16440, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001547, Sample Num: 24752, Cur Loss: 2.11548018, Cur Avg Loss: 1.55347856, Log Avg loss: 1.62291924, Global Avg Loss: 2.14537394, Time: 0.0402 Steps: 16450, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001557, Sample Num: 24912, Cur Loss: 0.81734020, Cur Avg Loss: 1.55303044, Log Avg loss: 1.48370710, Global Avg Loss: 2.14497196, Time: 0.0402 Steps: 16460, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001567, Sample Num: 25072, Cur Loss: 2.21673894, Cur Avg Loss: 1.55187732, Log Avg loss: 1.37233648, Global Avg Loss: 2.14450284, Time: 0.0402 Steps: 16470, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001577, Sample Num: 25232, Cur Loss: 1.12486863, Cur Avg Loss: 1.55310702, Log Avg loss: 1.74580092, Global Avg Loss: 2.14426091, Time: 0.0403 Steps: 16480, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001587, Sample Num: 25392, Cur Loss: 2.17972183, Cur Avg Loss: 1.55295915, Log Avg loss: 1.52964033, Global Avg Loss: 2.14388818, Time: 0.0402 Steps: 16490, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001597, Sample Num: 25552, Cur Loss: 1.93371224, Cur Avg Loss: 1.55165033, Log Avg loss: 1.34393913, Global Avg Loss: 2.14340337, Time: 0.0402 Steps: 16500, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001607, Sample Num: 25712, Cur Loss: 0.55250734, Cur Avg Loss: 1.55019453, Log Avg loss: 1.31770461, Global Avg Loss: 2.14290325, Time: 0.0402 Steps: 16510, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001617, Sample Num: 25872, Cur Loss: 2.52236676, Cur Avg Loss: 1.54863699, Log Avg loss: 1.29833984, Global Avg Loss: 2.14239201, Time: 0.0402 Steps: 16520, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001627, Sample Num: 26032, Cur Loss: 1.32295728, Cur Avg Loss: 1.54841559, Log Avg loss: 1.51261466, Global Avg Loss: 2.14201102, Time: 0.0403 Steps: 16530, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001637, Sample Num: 26192, Cur Loss: 1.08064091, Cur Avg Loss: 1.54847455, Log Avg loss: 1.55806809, Global Avg Loss: 2.14165797, Time: 0.0402 Steps: 16540, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001647, Sample Num: 26352, Cur Loss: 1.50714707, Cur Avg Loss: 1.55049909, Log Avg loss: 1.88191640, Global Avg Loss: 2.14150103, Time: 0.0402 Steps: 16550, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001657, Sample Num: 26512, Cur Loss: 1.29298162, Cur Avg Loss: 1.55109432, Log Avg loss: 1.64912781, Global Avg Loss: 2.14120370, Time: 0.0402 Steps: 16560, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001667, Sample Num: 26672, Cur Loss: 1.14185047, Cur Avg Loss: 1.54953747, Log Avg loss: 1.29156741, Global Avg Loss: 2.14069094, Time: 0.0402 Steps: 16570, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001677, Sample Num: 26832, Cur Loss: 1.57828772, Cur Avg Loss: 1.54958396, Log Avg loss: 1.55733348, Global Avg Loss: 2.14033910, Time: 0.0402 Steps: 16580, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001687, Sample Num: 26992, Cur Loss: 2.60239363, Cur Avg Loss: 1.55136954, Log Avg loss: 1.85081243, Global Avg Loss: 2.14016458, Time: 0.0402 Steps: 16590, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001697, Sample Num: 27152, Cur Loss: 0.80692577, Cur Avg Loss: 1.55145030, Log Avg loss: 1.56507505, Global Avg Loss: 2.13981814, Time: 0.0402 Steps: 16600, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001707, Sample Num: 27312, Cur Loss: 0.78150195, Cur Avg Loss: 1.55353038, Log Avg loss: 1.90651921, Global Avg Loss: 2.13967769, Time: 0.0402 Steps: 16610, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001717, Sample Num: 27472, Cur Loss: 1.42019916, Cur Avg Loss: 1.55281371, Log Avg loss: 1.43047863, Global Avg Loss: 2.13925097, Time: 0.0402 Steps: 16620, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001727, Sample Num: 27632, Cur Loss: 0.91236949, Cur Avg Loss: 1.55516752, Log Avg loss: 1.95931560, Global Avg Loss: 2.13914277, Time: 0.0402 Steps: 16630, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001737, Sample Num: 27792, Cur Loss: 1.41837978, Cur Avg Loss: 1.55348768, Log Avg loss: 1.26337999, Global Avg Loss: 2.13861647, Time: 0.0583 Steps: 16640, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001747, Sample Num: 27952, Cur Loss: 2.36229992, Cur Avg Loss: 1.55424225, Log Avg loss: 1.68531152, Global Avg Loss: 2.13834422, Time: 0.0602 Steps: 16650, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001757, Sample Num: 28112, Cur Loss: 0.93113887, Cur Avg Loss: 1.55284958, Log Avg loss: 1.30955033, Global Avg Loss: 2.13784674, Time: 0.0744 Steps: 16660, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001767, Sample Num: 28272, Cur Loss: 2.50487924, Cur Avg Loss: 1.55304293, Log Avg loss: 1.58701397, Global Avg Loss: 2.13751631, Time: 0.0505 Steps: 16670, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001777, Sample Num: 28432, Cur Loss: 1.72205472, Cur Avg Loss: 1.55361288, Log Avg loss: 1.65432292, Global Avg Loss: 2.13722662, Time: 0.0661 Steps: 16680, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001787, Sample Num: 28592, Cur Loss: 0.96427882, Cur Avg Loss: 1.55442783, Log Avg loss: 1.69924479, Global Avg Loss: 2.13696420, Time: 0.0406 Steps: 16690, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001797, Sample Num: 28752, Cur Loss: 1.70899129, Cur Avg Loss: 1.55437989, Log Avg loss: 1.54581325, Global Avg Loss: 2.13661022, Time: 0.0620 Steps: 16700, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001807, Sample Num: 28912, Cur Loss: 0.92938125, Cur Avg Loss: 1.55551913, Log Avg loss: 1.76023941, Global Avg Loss: 2.13638498, Time: 0.0498 Steps: 16710, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001817, Sample Num: 29072, Cur Loss: 1.59799564, Cur Avg Loss: 1.55572684, Log Avg loss: 1.59325967, Global Avg Loss: 2.13606015, Time: 0.0403 Steps: 16720, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001827, Sample Num: 29232, Cur Loss: 1.31206155, Cur Avg Loss: 1.55527083, Log Avg loss: 1.47241530, Global Avg Loss: 2.13566347, Time: 0.0402 Steps: 16730, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001837, Sample Num: 29392, Cur Loss: 2.10770273, Cur Avg Loss: 1.55643196, Log Avg loss: 1.76856915, Global Avg Loss: 2.13544418, Time: 0.0402 Steps: 16740, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001847, Sample Num: 29552, Cur Loss: 0.93278819, Cur Avg Loss: 1.55594688, Log Avg loss: 1.46683865, Global Avg Loss: 2.13504501, Time: 0.0403 Steps: 16750, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001857, Sample Num: 29712, Cur Loss: 1.07582200, Cur Avg Loss: 1.55642655, Log Avg loss: 1.64502126, Global Avg Loss: 2.13475263, Time: 0.0402 Steps: 16760, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001867, Sample Num: 29872, Cur Loss: 1.05333304, Cur Avg Loss: 1.55713582, Log Avg loss: 1.68884665, Global Avg Loss: 2.13448674, Time: 0.0403 Steps: 16770, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001877, Sample Num: 30032, Cur Loss: 1.12615490, Cur Avg Loss: 1.55686585, Log Avg loss: 1.50646327, Global Avg Loss: 2.13411247, Time: 0.0402 Steps: 16780, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001887, Sample Num: 30192, Cur Loss: 1.43182039, Cur Avg Loss: 1.55668549, Log Avg loss: 1.52283165, Global Avg Loss: 2.13374839, Time: 0.0402 Steps: 16790, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001897, Sample Num: 30352, Cur Loss: 0.93696445, Cur Avg Loss: 1.55545880, Log Avg loss: 1.32398299, Global Avg Loss: 2.13326639, Time: 0.0402 Steps: 16800, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001907, Sample Num: 30512, Cur Loss: 1.88834512, Cur Avg Loss: 1.55518054, Log Avg loss: 1.50239415, Global Avg Loss: 2.13289109, Time: 0.0402 Steps: 16810, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001917, Sample Num: 30672, Cur Loss: 0.73239374, Cur Avg Loss: 1.55454157, Log Avg loss: 1.43268920, Global Avg Loss: 2.13247480, Time: 0.0402 Steps: 16820, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001927, Sample Num: 30832, Cur Loss: 1.08424175, Cur Avg Loss: 1.55584805, Log Avg loss: 1.80630192, Global Avg Loss: 2.13228100, Time: 0.0405 Steps: 16830, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001937, Sample Num: 30992, Cur Loss: 2.26585817, Cur Avg Loss: 1.55710917, Log Avg loss: 1.80012540, Global Avg Loss: 2.13208376, Time: 0.0403 Steps: 16840, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001947, Sample Num: 31152, Cur Loss: 2.73104048, Cur Avg Loss: 1.55954189, Log Avg loss: 2.03076091, Global Avg Loss: 2.13202362, Time: 0.0403 Steps: 16850, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001957, Sample Num: 31312, Cur Loss: 1.25980496, Cur Avg Loss: 1.55860574, Log Avg loss: 1.37633666, Global Avg Loss: 2.13157541, Time: 0.0403 Steps: 16860, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001967, Sample Num: 31472, Cur Loss: 1.53861976, Cur Avg Loss: 1.55941684, Log Avg loss: 1.71814841, Global Avg Loss: 2.13133034, Time: 0.0402 Steps: 16870, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001977, Sample Num: 31632, Cur Loss: 0.79901809, Cur Avg Loss: 1.55863545, Log Avg loss: 1.40493617, Global Avg Loss: 2.13090002, Time: 0.0402 Steps: 16880, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001987, Sample Num: 31792, Cur Loss: 2.36986303, Cur Avg Loss: 1.56068325, Log Avg loss: 1.96553413, Global Avg Loss: 2.13080211, Time: 0.0401 Steps: 16890, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001997, Sample Num: 31952, Cur Loss: 1.89163470, Cur Avg Loss: 1.56156394, Log Avg loss: 1.73655616, Global Avg Loss: 2.13056883, Time: 0.0403 Steps: 16900, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002007, Sample Num: 32112, Cur Loss: 2.17519307, Cur Avg Loss: 1.56290693, Log Avg loss: 1.83110377, Global Avg Loss: 2.13039173, Time: 0.0403 Steps: 16910, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002017, Sample Num: 32272, Cur Loss: 0.90056789, Cur Avg Loss: 1.56156754, Log Avg loss: 1.29275073, Global Avg Loss: 2.12989667, Time: 0.0403 Steps: 16920, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002027, Sample Num: 32432, Cur Loss: 1.61808515, Cur Avg Loss: 1.56149684, Log Avg loss: 1.54723651, Global Avg Loss: 2.12955252, Time: 0.0402 Steps: 16930, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002037, Sample Num: 32592, Cur Loss: 0.86615223, Cur Avg Loss: 1.55921149, Log Avg loss: 1.09597194, Global Avg Loss: 2.12894237, Time: 0.0402 Steps: 16940, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002047, Sample Num: 32752, Cur Loss: 2.22021198, Cur Avg Loss: 1.56059438, Log Avg loss: 1.84228903, Global Avg Loss: 2.12877326, Time: 0.0402 Steps: 16950, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002057, Sample Num: 32912, Cur Loss: 1.03996313, Cur Avg Loss: 1.56026653, Log Avg loss: 1.49315569, Global Avg Loss: 2.12839848, Time: 0.0402 Steps: 16960, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002067, Sample Num: 33072, Cur Loss: 1.52986693, Cur Avg Loss: 1.56099555, Log Avg loss: 1.71095409, Global Avg Loss: 2.12815249, Time: 0.0402 Steps: 16970, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002077, Sample Num: 33232, Cur Loss: 2.33417511, Cur Avg Loss: 1.56170369, Log Avg loss: 1.70807687, Global Avg Loss: 2.12790510, Time: 0.0402 Steps: 16980, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002087, Sample Num: 33392, Cur Loss: 1.13772237, Cur Avg Loss: 1.56065425, Log Avg loss: 1.34268540, Global Avg Loss: 2.12744293, Time: 0.0402 Steps: 16990, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002097, Sample Num: 33552, Cur Loss: 1.63448095, Cur Avg Loss: 1.56107458, Log Avg loss: 1.64879813, Global Avg Loss: 2.12716138, Time: 0.1002 Steps: 17000, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002107, Sample Num: 33712, Cur Loss: 2.46951437, Cur Avg Loss: 1.56192452, Log Avg loss: 1.74015635, Global Avg Loss: 2.12693386, Time: 0.0514 Steps: 17010, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002117, Sample Num: 33872, Cur Loss: 0.85825872, Cur Avg Loss: 1.56001966, Log Avg loss: 1.15866494, Global Avg Loss: 2.12636496, Time: 0.0521 Steps: 17020, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002127, Sample Num: 34032, Cur Loss: 1.44506228, Cur Avg Loss: 1.55920289, Log Avg loss: 1.38629425, Global Avg Loss: 2.12593039, Time: 0.0602 Steps: 17030, Updated lr: 0.000085 ***** Running evaluation checkpoint-17032 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-17032 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 96.300931, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.69238, "eval_total_loss": 1189.743039, "eval_mae": 1.187361, "eval_mse": 1.692093, "eval_r2": -0.075605, "eval_sp_statistic": 0.290532, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.349313, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.480177, "test_total_loss": 743.048862, "test_mae": 1.062025, "test_mse": 1.480376, "test_r2": 0.044552, "test_sp_statistic": 0.301587, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.358573, "test_ps_pvalue": 0.0, "lr": 8.479658605974395e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.1258738539822506, "train_cur_epoch_loss": 3319.7134920209646, "train_cur_epoch_avg_loss": 1.559282992964286, "train_cur_epoch_time": 96.30093097686768, "train_cur_epoch_avg_time": 0.045232940806419764, "epoch": 8, "step": 17032} ################################################## Training, Epoch: 0009, Batch: 000008, Sample Num: 128, Cur Loss: 1.23628664, Cur Avg Loss: 1.67042640, Log Avg loss: 1.66523459, Global Avg Loss: 2.12566003, Time: 0.0402 Steps: 17040, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000018, Sample Num: 288, Cur Loss: 1.10538530, Cur Avg Loss: 1.65965266, Log Avg loss: 1.65103368, Global Avg Loss: 2.12538166, Time: 0.0402 Steps: 17050, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000028, Sample Num: 448, Cur Loss: 0.79527038, Cur Avg Loss: 1.63479984, Log Avg loss: 1.59006476, Global Avg Loss: 2.12506787, Time: 0.0404 Steps: 17060, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000038, Sample Num: 608, Cur Loss: 1.26831901, Cur Avg Loss: 1.57869358, Log Avg loss: 1.42159605, Global Avg Loss: 2.12465576, Time: 0.0647 Steps: 17070, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000048, Sample Num: 768, Cur Loss: 3.12952185, Cur Avg Loss: 1.58410570, Log Avg loss: 1.60467175, Global Avg Loss: 2.12435132, Time: 0.0452 Steps: 17080, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000058, Sample Num: 928, Cur Loss: 1.20602965, Cur Avg Loss: 1.55411519, Log Avg loss: 1.41016074, Global Avg Loss: 2.12393342, Time: 0.0648 Steps: 17090, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000068, Sample Num: 1088, Cur Loss: 1.55109954, Cur Avg Loss: 1.54663002, Log Avg loss: 1.50321603, Global Avg Loss: 2.12357043, Time: 0.0480 Steps: 17100, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000078, Sample Num: 1248, Cur Loss: 0.81925851, Cur Avg Loss: 1.51132493, Log Avg loss: 1.27125037, Global Avg Loss: 2.12307229, Time: 0.0891 Steps: 17110, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000088, Sample Num: 1408, Cur Loss: 1.09246135, Cur Avg Loss: 1.53985716, Log Avg loss: 1.76240852, Global Avg Loss: 2.12286162, Time: 0.0653 Steps: 17120, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000098, Sample Num: 1568, Cur Loss: 1.01112032, Cur Avg Loss: 1.50204593, Log Avg loss: 1.16930714, Global Avg Loss: 2.12230496, Time: 0.0570 Steps: 17130, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000108, Sample Num: 1728, Cur Loss: 1.02033830, Cur Avg Loss: 1.51041098, Log Avg loss: 1.59238839, Global Avg Loss: 2.12199579, Time: 0.0622 Steps: 17140, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000118, Sample Num: 1888, Cur Loss: 1.53466392, Cur Avg Loss: 1.51949796, Log Avg loss: 1.61763742, Global Avg Loss: 2.12170170, Time: 0.0403 Steps: 17150, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000128, Sample Num: 2048, Cur Loss: 1.70649588, Cur Avg Loss: 1.51614472, Log Avg loss: 1.47657639, Global Avg Loss: 2.12132576, Time: 0.0402 Steps: 17160, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000138, Sample Num: 2208, Cur Loss: 2.43398237, Cur Avg Loss: 1.50369129, Log Avg loss: 1.34428739, Global Avg Loss: 2.12087320, Time: 0.0402 Steps: 17170, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000148, Sample Num: 2368, Cur Loss: 1.33833396, Cur Avg Loss: 1.49492917, Log Avg loss: 1.37401192, Global Avg Loss: 2.12043847, Time: 0.0402 Steps: 17180, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000158, Sample Num: 2528, Cur Loss: 0.73140311, Cur Avg Loss: 1.49993693, Log Avg loss: 1.57405176, Global Avg Loss: 2.12012062, Time: 0.0402 Steps: 17190, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000168, Sample Num: 2688, Cur Loss: 1.39742279, Cur Avg Loss: 1.53151237, Log Avg loss: 2.03040440, Global Avg Loss: 2.12006846, Time: 0.0402 Steps: 17200, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000178, Sample Num: 2848, Cur Loss: 1.59774637, Cur Avg Loss: 1.54717536, Log Avg loss: 1.81031365, Global Avg Loss: 2.11988848, Time: 0.0402 Steps: 17210, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000188, Sample Num: 3008, Cur Loss: 2.33350325, Cur Avg Loss: 1.53888025, Log Avg loss: 1.39122731, Global Avg Loss: 2.11946533, Time: 0.0402 Steps: 17220, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000198, Sample Num: 3168, Cur Loss: 2.53898764, Cur Avg Loss: 1.53174625, Log Avg loss: 1.39762696, Global Avg Loss: 2.11904639, Time: 0.0402 Steps: 17230, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000208, Sample Num: 3328, Cur Loss: 2.33751535, Cur Avg Loss: 1.54859525, Log Avg loss: 1.88220553, Global Avg Loss: 2.11890901, Time: 0.0402 Steps: 17240, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000218, Sample Num: 3488, Cur Loss: 2.69646549, Cur Avg Loss: 1.56225171, Log Avg loss: 1.84630609, Global Avg Loss: 2.11875098, Time: 0.0402 Steps: 17250, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000228, Sample Num: 3648, Cur Loss: 2.82523346, Cur Avg Loss: 1.56270890, Log Avg loss: 1.57267544, Global Avg Loss: 2.11843459, Time: 0.0402 Steps: 17260, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000238, Sample Num: 3808, Cur Loss: 1.71585286, Cur Avg Loss: 1.55882709, Log Avg loss: 1.47032183, Global Avg Loss: 2.11805931, Time: 0.0402 Steps: 17270, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000248, Sample Num: 3968, Cur Loss: 1.64165187, Cur Avg Loss: 1.56014450, Log Avg loss: 1.59149900, Global Avg Loss: 2.11775459, Time: 0.0402 Steps: 17280, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000258, Sample Num: 4128, Cur Loss: 1.52870250, Cur Avg Loss: 1.55906086, Log Avg loss: 1.53218649, Global Avg Loss: 2.11741592, Time: 0.0402 Steps: 17290, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000268, Sample Num: 4288, Cur Loss: 0.79778469, Cur Avg Loss: 1.56415878, Log Avg loss: 1.69568510, Global Avg Loss: 2.11717214, Time: 0.0402 Steps: 17300, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000278, Sample Num: 4448, Cur Loss: 1.69496953, Cur Avg Loss: 1.56042239, Log Avg loss: 1.46028733, Global Avg Loss: 2.11679266, Time: 0.0402 Steps: 17310, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000288, Sample Num: 4608, Cur Loss: 0.79734468, Cur Avg Loss: 1.56067493, Log Avg loss: 1.56769544, Global Avg Loss: 2.11647563, Time: 0.0402 Steps: 17320, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000298, Sample Num: 4768, Cur Loss: 1.65325677, Cur Avg Loss: 1.55486577, Log Avg loss: 1.38756183, Global Avg Loss: 2.11605502, Time: 0.0402 Steps: 17330, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000308, Sample Num: 4928, Cur Loss: 1.47672558, Cur Avg Loss: 1.55742556, Log Avg loss: 1.63370749, Global Avg Loss: 2.11577685, Time: 0.0402 Steps: 17340, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000318, Sample Num: 5088, Cur Loss: 2.05052257, Cur Avg Loss: 1.55928019, Log Avg loss: 1.61640260, Global Avg Loss: 2.11548902, Time: 0.0402 Steps: 17350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000328, Sample Num: 5248, Cur Loss: 0.57526064, Cur Avg Loss: 1.55185156, Log Avg loss: 1.31562122, Global Avg Loss: 2.11502827, Time: 0.0402 Steps: 17360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000338, Sample Num: 5408, Cur Loss: 2.03109980, Cur Avg Loss: 1.55374121, Log Avg loss: 1.61572180, Global Avg Loss: 2.11474082, Time: 0.0402 Steps: 17370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000348, Sample Num: 5568, Cur Loss: 1.02926350, Cur Avg Loss: 1.54727601, Log Avg loss: 1.32875228, Global Avg Loss: 2.11428858, Time: 0.0401 Steps: 17380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000358, Sample Num: 5728, Cur Loss: 0.66988057, Cur Avg Loss: 1.55330861, Log Avg loss: 1.76324294, Global Avg Loss: 2.11408671, Time: 0.0402 Steps: 17390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000368, Sample Num: 5888, Cur Loss: 0.79162169, Cur Avg Loss: 1.54782263, Log Avg loss: 1.35142460, Global Avg Loss: 2.11364840, Time: 0.0402 Steps: 17400, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000378, Sample Num: 6048, Cur Loss: 1.25694418, Cur Avg Loss: 1.54885822, Log Avg loss: 1.58696814, Global Avg Loss: 2.11334589, Time: 0.0406 Steps: 17410, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000388, Sample Num: 6208, Cur Loss: 1.29500604, Cur Avg Loss: 1.55484740, Log Avg loss: 1.78123812, Global Avg Loss: 2.11315524, Time: 0.0406 Steps: 17420, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000398, Sample Num: 6368, Cur Loss: 1.52580345, Cur Avg Loss: 1.56448287, Log Avg loss: 1.93833930, Global Avg Loss: 2.11305494, Time: 0.0613 Steps: 17430, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000408, Sample Num: 6528, Cur Loss: 1.49230361, Cur Avg Loss: 1.56242167, Log Avg loss: 1.48038570, Global Avg Loss: 2.11269217, Time: 0.0625 Steps: 17440, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000418, Sample Num: 6688, Cur Loss: 2.07156372, Cur Avg Loss: 1.56518464, Log Avg loss: 1.67791402, Global Avg Loss: 2.11244302, Time: 0.0406 Steps: 17450, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000428, Sample Num: 6848, Cur Loss: 0.97850388, Cur Avg Loss: 1.56189286, Log Avg loss: 1.42429654, Global Avg Loss: 2.11204889, Time: 0.0411 Steps: 17460, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000438, Sample Num: 7008, Cur Loss: 0.62219989, Cur Avg Loss: 1.55838875, Log Avg loss: 1.40841271, Global Avg Loss: 2.11164612, Time: 0.0536 Steps: 17470, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000448, Sample Num: 7168, Cur Loss: 0.82888639, Cur Avg Loss: 1.55779063, Log Avg loss: 1.53159286, Global Avg Loss: 2.11131428, Time: 0.0639 Steps: 17480, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000458, Sample Num: 7328, Cur Loss: 1.42858768, Cur Avg Loss: 1.55846421, Log Avg loss: 1.58864046, Global Avg Loss: 2.11101544, Time: 0.0403 Steps: 17490, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000468, Sample Num: 7488, Cur Loss: 0.58912206, Cur Avg Loss: 1.55490709, Log Avg loss: 1.39199135, Global Avg Loss: 2.11060457, Time: 0.0403 Steps: 17500, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000478, Sample Num: 7648, Cur Loss: 1.64923298, Cur Avg Loss: 1.55296269, Log Avg loss: 1.46196444, Global Avg Loss: 2.11023413, Time: 0.0402 Steps: 17510, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000488, Sample Num: 7808, Cur Loss: 0.77542317, Cur Avg Loss: 1.55392915, Log Avg loss: 1.60012594, Global Avg Loss: 2.10994297, Time: 0.0402 Steps: 17520, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000498, Sample Num: 7968, Cur Loss: 0.77268851, Cur Avg Loss: 1.55444829, Log Avg loss: 1.57978253, Global Avg Loss: 2.10964054, Time: 0.0402 Steps: 17530, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000508, Sample Num: 8128, Cur Loss: 1.75117743, Cur Avg Loss: 1.55584055, Log Avg loss: 1.62517518, Global Avg Loss: 2.10936434, Time: 0.0403 Steps: 17540, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000518, Sample Num: 8288, Cur Loss: 1.25817680, Cur Avg Loss: 1.55085748, Log Avg loss: 1.29771719, Global Avg Loss: 2.10890186, Time: 0.0402 Steps: 17550, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000528, Sample Num: 8448, Cur Loss: 0.63093483, Cur Avg Loss: 1.54936243, Log Avg loss: 1.47191917, Global Avg Loss: 2.10853911, Time: 0.0402 Steps: 17560, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000538, Sample Num: 8608, Cur Loss: 0.73253965, Cur Avg Loss: 1.53925168, Log Avg loss: 1.00540369, Global Avg Loss: 2.10791126, Time: 0.0402 Steps: 17570, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000548, Sample Num: 8768, Cur Loss: 1.36085284, Cur Avg Loss: 1.53484898, Log Avg loss: 1.29798398, Global Avg Loss: 2.10745055, Time: 0.0402 Steps: 17580, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000558, Sample Num: 8928, Cur Loss: 2.21822476, Cur Avg Loss: 1.53404485, Log Avg loss: 1.48997863, Global Avg Loss: 2.10709952, Time: 0.0402 Steps: 17590, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000568, Sample Num: 9088, Cur Loss: 1.26471114, Cur Avg Loss: 1.52862126, Log Avg loss: 1.22598473, Global Avg Loss: 2.10659888, Time: 0.0402 Steps: 17600, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000578, Sample Num: 9248, Cur Loss: 2.75331235, Cur Avg Loss: 1.52923206, Log Avg loss: 1.56392556, Global Avg Loss: 2.10629072, Time: 0.0402 Steps: 17610, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000588, Sample Num: 9408, Cur Loss: 0.88336873, Cur Avg Loss: 1.53002044, Log Avg loss: 1.57558860, Global Avg Loss: 2.10598953, Time: 0.0403 Steps: 17620, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000598, Sample Num: 9568, Cur Loss: 1.43200541, Cur Avg Loss: 1.53173324, Log Avg loss: 1.63244623, Global Avg Loss: 2.10572093, Time: 0.0403 Steps: 17630, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000608, Sample Num: 9728, Cur Loss: 1.23330057, Cur Avg Loss: 1.53406665, Log Avg loss: 1.67360423, Global Avg Loss: 2.10547596, Time: 0.0402 Steps: 17640, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000618, Sample Num: 9888, Cur Loss: 1.21473002, Cur Avg Loss: 1.53269889, Log Avg loss: 1.44953932, Global Avg Loss: 2.10510433, Time: 0.0402 Steps: 17650, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000628, Sample Num: 10048, Cur Loss: 0.92871594, Cur Avg Loss: 1.53113206, Log Avg loss: 1.43430169, Global Avg Loss: 2.10472449, Time: 0.0402 Steps: 17660, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000638, Sample Num: 10208, Cur Loss: 1.81048465, Cur Avg Loss: 1.53111794, Log Avg loss: 1.53023149, Global Avg Loss: 2.10439936, Time: 0.0402 Steps: 17670, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000648, Sample Num: 10368, Cur Loss: 0.38703400, Cur Avg Loss: 1.52926402, Log Avg loss: 1.41098416, Global Avg Loss: 2.10400716, Time: 0.0402 Steps: 17680, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000658, Sample Num: 10528, Cur Loss: 0.92057157, Cur Avg Loss: 1.52837705, Log Avg loss: 1.47090107, Global Avg Loss: 2.10364927, Time: 0.0402 Steps: 17690, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000668, Sample Num: 10688, Cur Loss: 1.22257316, Cur Avg Loss: 1.52877360, Log Avg loss: 1.55486648, Global Avg Loss: 2.10333922, Time: 0.0402 Steps: 17700, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000678, Sample Num: 10848, Cur Loss: 0.48684075, Cur Avg Loss: 1.52879077, Log Avg loss: 1.52993805, Global Avg Loss: 2.10301545, Time: 0.0402 Steps: 17710, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000688, Sample Num: 11008, Cur Loss: 2.72570086, Cur Avg Loss: 1.52719552, Log Avg loss: 1.41903761, Global Avg Loss: 2.10262946, Time: 0.0402 Steps: 17720, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000698, Sample Num: 11168, Cur Loss: 2.84081101, Cur Avg Loss: 1.52824558, Log Avg loss: 1.60048975, Global Avg Loss: 2.10234624, Time: 0.0401 Steps: 17730, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000708, Sample Num: 11328, Cur Loss: 0.63541353, Cur Avg Loss: 1.52360961, Log Avg loss: 1.20001835, Global Avg Loss: 2.10183760, Time: 0.0402 Steps: 17740, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000718, Sample Num: 11488, Cur Loss: 2.05458164, Cur Avg Loss: 1.52568234, Log Avg loss: 1.67243185, Global Avg Loss: 2.10159568, Time: 0.0401 Steps: 17750, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000728, Sample Num: 11648, Cur Loss: 2.18742752, Cur Avg Loss: 1.52360569, Log Avg loss: 1.37450264, Global Avg Loss: 2.10118629, Time: 0.0401 Steps: 17760, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000738, Sample Num: 11808, Cur Loss: 0.58701265, Cur Avg Loss: 1.51880320, Log Avg loss: 1.16918175, Global Avg Loss: 2.10066180, Time: 0.0801 Steps: 17770, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000748, Sample Num: 11968, Cur Loss: 0.96564555, Cur Avg Loss: 1.51701592, Log Avg loss: 1.38511481, Global Avg Loss: 2.10025936, Time: 0.0596 Steps: 17780, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000758, Sample Num: 12128, Cur Loss: 2.27895641, Cur Avg Loss: 1.52228372, Log Avg loss: 1.91631505, Global Avg Loss: 2.10015596, Time: 0.0486 Steps: 17790, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000768, Sample Num: 12288, Cur Loss: 0.67816240, Cur Avg Loss: 1.51866367, Log Avg loss: 1.24426362, Global Avg Loss: 2.09967512, Time: 0.0770 Steps: 17800, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000778, Sample Num: 12448, Cur Loss: 0.99162239, Cur Avg Loss: 1.52224556, Log Avg loss: 1.79733459, Global Avg Loss: 2.09950536, Time: 0.0855 Steps: 17810, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000788, Sample Num: 12608, Cur Loss: 1.48980260, Cur Avg Loss: 1.52442266, Log Avg loss: 1.69380115, Global Avg Loss: 2.09927770, Time: 0.0610 Steps: 17820, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000798, Sample Num: 12768, Cur Loss: 1.57868147, Cur Avg Loss: 1.52334617, Log Avg loss: 1.43851861, Global Avg Loss: 2.09890711, Time: 0.0739 Steps: 17830, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000808, Sample Num: 12928, Cur Loss: 0.97948498, Cur Avg Loss: 1.52252172, Log Avg loss: 1.45673083, Global Avg Loss: 2.09854714, Time: 0.0404 Steps: 17840, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000818, Sample Num: 13088, Cur Loss: 0.58740306, Cur Avg Loss: 1.51870324, Log Avg loss: 1.21016998, Global Avg Loss: 2.09804945, Time: 0.0833 Steps: 17850, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000828, Sample Num: 13248, Cur Loss: 2.47479892, Cur Avg Loss: 1.51558925, Log Avg loss: 1.26086473, Global Avg Loss: 2.09758070, Time: 0.0846 Steps: 17860, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000838, Sample Num: 13408, Cur Loss: 1.38678122, Cur Avg Loss: 1.51926680, Log Avg loss: 1.82376860, Global Avg Loss: 2.09742748, Time: 0.0402 Steps: 17870, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000848, Sample Num: 13568, Cur Loss: 0.76986176, Cur Avg Loss: 1.51381471, Log Avg loss: 1.05692946, Global Avg Loss: 2.09684555, Time: 0.0402 Steps: 17880, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000858, Sample Num: 13728, Cur Loss: 0.50417560, Cur Avg Loss: 1.51440665, Log Avg loss: 1.56460306, Global Avg Loss: 2.09654804, Time: 0.0402 Steps: 17890, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000868, Sample Num: 13888, Cur Loss: 1.30606675, Cur Avg Loss: 1.51499364, Log Avg loss: 1.56535733, Global Avg Loss: 2.09625128, Time: 0.0402 Steps: 17900, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000878, Sample Num: 14048, Cur Loss: 2.36449385, Cur Avg Loss: 1.51577577, Log Avg loss: 1.58366466, Global Avg Loss: 2.09596508, Time: 0.0403 Steps: 17910, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000888, Sample Num: 14208, Cur Loss: 2.85364532, Cur Avg Loss: 1.51577707, Log Avg loss: 1.51589142, Global Avg Loss: 2.09564138, Time: 0.0402 Steps: 17920, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000898, Sample Num: 14368, Cur Loss: 0.68493617, Cur Avg Loss: 1.51603125, Log Avg loss: 1.53860168, Global Avg Loss: 2.09533070, Time: 0.0402 Steps: 17930, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000908, Sample Num: 14528, Cur Loss: 1.51255512, Cur Avg Loss: 1.51680454, Log Avg loss: 1.58624669, Global Avg Loss: 2.09504693, Time: 0.0402 Steps: 17940, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000918, Sample Num: 14688, Cur Loss: 2.66275740, Cur Avg Loss: 1.51935554, Log Avg loss: 1.75098641, Global Avg Loss: 2.09485526, Time: 0.0402 Steps: 17950, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000928, Sample Num: 14848, Cur Loss: 1.80153823, Cur Avg Loss: 1.52006414, Log Avg loss: 1.58511317, Global Avg Loss: 2.09457144, Time: 0.0402 Steps: 17960, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000938, Sample Num: 15008, Cur Loss: 2.51872301, Cur Avg Loss: 1.52126411, Log Avg loss: 1.63262100, Global Avg Loss: 2.09431437, Time: 0.0402 Steps: 17970, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000948, Sample Num: 15168, Cur Loss: 1.71597278, Cur Avg Loss: 1.52365383, Log Avg loss: 1.74780983, Global Avg Loss: 2.09412165, Time: 0.0402 Steps: 17980, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000958, Sample Num: 15328, Cur Loss: 0.92237115, Cur Avg Loss: 1.52704103, Log Avg loss: 1.84814803, Global Avg Loss: 2.09398492, Time: 0.0402 Steps: 17990, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000968, Sample Num: 15488, Cur Loss: 1.30886102, Cur Avg Loss: 1.53172246, Log Avg loss: 1.98020278, Global Avg Loss: 2.09392171, Time: 0.0402 Steps: 18000, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000978, Sample Num: 15648, Cur Loss: 0.86314797, Cur Avg Loss: 1.52897800, Log Avg loss: 1.26331513, Global Avg Loss: 2.09346052, Time: 0.0402 Steps: 18010, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000988, Sample Num: 15808, Cur Loss: 1.58982396, Cur Avg Loss: 1.53187628, Log Avg loss: 1.81532804, Global Avg Loss: 2.09330617, Time: 0.0401 Steps: 18020, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000998, Sample Num: 15968, Cur Loss: 1.41022635, Cur Avg Loss: 1.53815115, Log Avg loss: 2.15810804, Global Avg Loss: 2.09334211, Time: 0.0402 Steps: 18030, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001008, Sample Num: 16128, Cur Loss: 1.76850772, Cur Avg Loss: 1.54118618, Log Avg loss: 1.84408150, Global Avg Loss: 2.09320394, Time: 0.0402 Steps: 18040, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001018, Sample Num: 16288, Cur Loss: 0.40723804, Cur Avg Loss: 1.53996222, Log Avg loss: 1.41658739, Global Avg Loss: 2.09282909, Time: 0.0402 Steps: 18050, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001028, Sample Num: 16448, Cur Loss: 1.40861857, Cur Avg Loss: 1.54084392, Log Avg loss: 1.63060100, Global Avg Loss: 2.09257315, Time: 0.0401 Steps: 18060, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001038, Sample Num: 16608, Cur Loss: 1.57108045, Cur Avg Loss: 1.54257841, Log Avg loss: 1.72088362, Global Avg Loss: 2.09236745, Time: 0.0402 Steps: 18070, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001048, Sample Num: 16768, Cur Loss: 1.27603030, Cur Avg Loss: 1.54317046, Log Avg loss: 1.60462575, Global Avg Loss: 2.09209768, Time: 0.0402 Steps: 18080, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001058, Sample Num: 16928, Cur Loss: 2.75057411, Cur Avg Loss: 1.54383697, Log Avg loss: 1.61368709, Global Avg Loss: 2.09183322, Time: 0.0402 Steps: 18090, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001068, Sample Num: 17088, Cur Loss: 0.94655871, Cur Avg Loss: 1.54625531, Log Avg loss: 1.80211541, Global Avg Loss: 2.09167316, Time: 0.0402 Steps: 18100, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001078, Sample Num: 17248, Cur Loss: 1.28086174, Cur Avg Loss: 1.54335981, Log Avg loss: 1.23412110, Global Avg Loss: 2.09119963, Time: 0.0403 Steps: 18110, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001088, Sample Num: 17408, Cur Loss: 1.85249436, Cur Avg Loss: 1.54342108, Log Avg loss: 1.55002547, Global Avg Loss: 2.09090097, Time: 0.0402 Steps: 18120, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001098, Sample Num: 17568, Cur Loss: 1.56739378, Cur Avg Loss: 1.54470020, Log Avg loss: 1.68386886, Global Avg Loss: 2.09067646, Time: 0.0461 Steps: 18130, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001108, Sample Num: 17728, Cur Loss: 2.35636854, Cur Avg Loss: 1.54583078, Log Avg loss: 1.66996795, Global Avg Loss: 2.09044454, Time: 0.0762 Steps: 18140, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001118, Sample Num: 17888, Cur Loss: 1.18301368, Cur Avg Loss: 1.54559102, Log Avg loss: 1.51902619, Global Avg Loss: 2.09012971, Time: 0.0457 Steps: 18150, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001128, Sample Num: 18048, Cur Loss: 1.33197784, Cur Avg Loss: 1.54565356, Log Avg loss: 1.55264541, Global Avg Loss: 2.08983374, Time: 0.0667 Steps: 18160, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001138, Sample Num: 18208, Cur Loss: 1.18105769, Cur Avg Loss: 1.54316326, Log Avg loss: 1.26225750, Global Avg Loss: 2.08937828, Time: 0.0401 Steps: 18170, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001148, Sample Num: 18368, Cur Loss: 1.81184220, Cur Avg Loss: 1.54251132, Log Avg loss: 1.46831981, Global Avg Loss: 2.08903666, Time: 0.0537 Steps: 18180, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001158, Sample Num: 18528, Cur Loss: 1.35166991, Cur Avg Loss: 1.54070572, Log Avg loss: 1.33342367, Global Avg Loss: 2.08862126, Time: 0.0406 Steps: 18190, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001168, Sample Num: 18688, Cur Loss: 0.66436493, Cur Avg Loss: 1.54086444, Log Avg loss: 1.55924333, Global Avg Loss: 2.08833039, Time: 0.0407 Steps: 18200, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001178, Sample Num: 18848, Cur Loss: 1.97461951, Cur Avg Loss: 1.54093260, Log Avg loss: 1.54889455, Global Avg Loss: 2.08803416, Time: 0.0407 Steps: 18210, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001188, Sample Num: 19008, Cur Loss: 1.34156251, Cur Avg Loss: 1.54045629, Log Avg loss: 1.48434653, Global Avg Loss: 2.08770283, Time: 0.0403 Steps: 18220, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001198, Sample Num: 19168, Cur Loss: 0.54896617, Cur Avg Loss: 1.53886229, Log Avg loss: 1.34949498, Global Avg Loss: 2.08729789, Time: 0.0402 Steps: 18230, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001208, Sample Num: 19328, Cur Loss: 1.37579846, Cur Avg Loss: 1.53546278, Log Avg loss: 1.12820184, Global Avg Loss: 2.08677207, Time: 0.0402 Steps: 18240, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001218, Sample Num: 19488, Cur Loss: 1.07419682, Cur Avg Loss: 1.53469808, Log Avg loss: 1.44232168, Global Avg Loss: 2.08641894, Time: 0.0402 Steps: 18250, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001228, Sample Num: 19648, Cur Loss: 1.07959044, Cur Avg Loss: 1.53504110, Log Avg loss: 1.57682182, Global Avg Loss: 2.08613987, Time: 0.0403 Steps: 18260, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001238, Sample Num: 19808, Cur Loss: 1.65883529, Cur Avg Loss: 1.53300234, Log Avg loss: 1.28264206, Global Avg Loss: 2.08570008, Time: 0.0402 Steps: 18270, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001248, Sample Num: 19968, Cur Loss: 2.04415989, Cur Avg Loss: 1.53477558, Log Avg loss: 1.75430262, Global Avg Loss: 2.08551879, Time: 0.0402 Steps: 18280, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001258, Sample Num: 20128, Cur Loss: 1.45931745, Cur Avg Loss: 1.53644264, Log Avg loss: 1.74449264, Global Avg Loss: 2.08533233, Time: 0.0402 Steps: 18290, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001268, Sample Num: 20288, Cur Loss: 1.04301071, Cur Avg Loss: 1.53571804, Log Avg loss: 1.44456230, Global Avg Loss: 2.08498218, Time: 0.0402 Steps: 18300, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001278, Sample Num: 20448, Cur Loss: 0.91192591, Cur Avg Loss: 1.53363200, Log Avg loss: 1.26912241, Global Avg Loss: 2.08453660, Time: 0.0402 Steps: 18310, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001288, Sample Num: 20608, Cur Loss: 1.18345261, Cur Avg Loss: 1.53296529, Log Avg loss: 1.44776006, Global Avg Loss: 2.08418902, Time: 0.0402 Steps: 18320, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001298, Sample Num: 20768, Cur Loss: 1.60681820, Cur Avg Loss: 1.53185907, Log Avg loss: 1.38937790, Global Avg Loss: 2.08380996, Time: 0.0403 Steps: 18330, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001308, Sample Num: 20928, Cur Loss: 2.27358294, Cur Avg Loss: 1.53191432, Log Avg loss: 1.53908587, Global Avg Loss: 2.08351295, Time: 0.0402 Steps: 18340, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001318, Sample Num: 21088, Cur Loss: 1.61086893, Cur Avg Loss: 1.53216605, Log Avg loss: 1.56509152, Global Avg Loss: 2.08323043, Time: 0.0402 Steps: 18350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001328, Sample Num: 21248, Cur Loss: 1.54382074, Cur Avg Loss: 1.53411195, Log Avg loss: 1.79058273, Global Avg Loss: 2.08307103, Time: 0.0402 Steps: 18360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001338, Sample Num: 21408, Cur Loss: 2.21677256, Cur Avg Loss: 1.53565441, Log Avg loss: 1.74049191, Global Avg Loss: 2.08288454, Time: 0.0402 Steps: 18370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001348, Sample Num: 21568, Cur Loss: 0.57940674, Cur Avg Loss: 1.53682353, Log Avg loss: 1.69325217, Global Avg Loss: 2.08267256, Time: 0.0402 Steps: 18380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001358, Sample Num: 21728, Cur Loss: 1.71797407, Cur Avg Loss: 1.53600961, Log Avg loss: 1.42629278, Global Avg Loss: 2.08231563, Time: 0.0402 Steps: 18390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001368, Sample Num: 21888, Cur Loss: 0.79785764, Cur Avg Loss: 1.53180466, Log Avg loss: 0.96077257, Global Avg Loss: 2.08170610, Time: 0.0402 Steps: 18400, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001378, Sample Num: 22048, Cur Loss: 1.64611316, Cur Avg Loss: 1.53141880, Log Avg loss: 1.47863422, Global Avg Loss: 2.08137852, Time: 0.0403 Steps: 18410, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001388, Sample Num: 22208, Cur Loss: 2.35252619, Cur Avg Loss: 1.53358947, Log Avg loss: 1.83270791, Global Avg Loss: 2.08124352, Time: 0.0402 Steps: 18420, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001398, Sample Num: 22368, Cur Loss: 0.70731556, Cur Avg Loss: 1.53337819, Log Avg loss: 1.50405124, Global Avg Loss: 2.08093034, Time: 0.0402 Steps: 18430, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001408, Sample Num: 22528, Cur Loss: 1.15245795, Cur Avg Loss: 1.53514336, Log Avg loss: 1.78191539, Global Avg Loss: 2.08076819, Time: 0.0402 Steps: 18440, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001418, Sample Num: 22688, Cur Loss: 0.99668109, Cur Avg Loss: 1.53466308, Log Avg loss: 1.46703846, Global Avg Loss: 2.08043554, Time: 0.0402 Steps: 18450, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001428, Sample Num: 22848, Cur Loss: 1.31221616, Cur Avg Loss: 1.53489954, Log Avg loss: 1.56843001, Global Avg Loss: 2.08015818, Time: 0.0403 Steps: 18460, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001438, Sample Num: 23008, Cur Loss: 2.90315151, Cur Avg Loss: 1.53496728, Log Avg loss: 1.54463999, Global Avg Loss: 2.07986824, Time: 0.0402 Steps: 18470, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001448, Sample Num: 23168, Cur Loss: 1.32482958, Cur Avg Loss: 1.53526629, Log Avg loss: 1.57826424, Global Avg Loss: 2.07959681, Time: 0.0405 Steps: 18480, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001458, Sample Num: 23328, Cur Loss: 1.25215673, Cur Avg Loss: 1.53466175, Log Avg loss: 1.44712453, Global Avg Loss: 2.07925475, Time: 0.0406 Steps: 18490, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001468, Sample Num: 23488, Cur Loss: 1.82138896, Cur Avg Loss: 1.53942650, Log Avg loss: 2.23412737, Global Avg Loss: 2.07933846, Time: 0.0547 Steps: 18500, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001478, Sample Num: 23648, Cur Loss: 1.02297688, Cur Avg Loss: 1.53754575, Log Avg loss: 1.26145120, Global Avg Loss: 2.07889660, Time: 0.0718 Steps: 18510, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001488, Sample Num: 23808, Cur Loss: 1.42133844, Cur Avg Loss: 1.53753833, Log Avg loss: 1.53644143, Global Avg Loss: 2.07860370, Time: 0.0429 Steps: 18520, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001498, Sample Num: 23968, Cur Loss: 1.36757803, Cur Avg Loss: 1.54049520, Log Avg loss: 1.98047752, Global Avg Loss: 2.07855074, Time: 0.0486 Steps: 18530, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001508, Sample Num: 24128, Cur Loss: 1.30716443, Cur Avg Loss: 1.53848856, Log Avg loss: 1.23789403, Global Avg Loss: 2.07809732, Time: 0.0785 Steps: 18540, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001518, Sample Num: 24288, Cur Loss: 2.02624464, Cur Avg Loss: 1.54018089, Log Avg loss: 1.79538502, Global Avg Loss: 2.07794491, Time: 0.0667 Steps: 18550, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001528, Sample Num: 24448, Cur Loss: 1.60122323, Cur Avg Loss: 1.53991773, Log Avg loss: 1.49997013, Global Avg Loss: 2.07763350, Time: 0.0694 Steps: 18560, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001538, Sample Num: 24608, Cur Loss: 2.35328865, Cur Avg Loss: 1.53988853, Log Avg loss: 1.53542673, Global Avg Loss: 2.07734152, Time: 0.0402 Steps: 18570, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001548, Sample Num: 24768, Cur Loss: 1.15706491, Cur Avg Loss: 1.53907279, Log Avg loss: 1.41361144, Global Avg Loss: 2.07698429, Time: 0.0402 Steps: 18580, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001558, Sample Num: 24928, Cur Loss: 1.38123965, Cur Avg Loss: 1.53914624, Log Avg loss: 1.55051595, Global Avg Loss: 2.07670109, Time: 0.0402 Steps: 18590, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001568, Sample Num: 25088, Cur Loss: 1.34002852, Cur Avg Loss: 1.53610316, Log Avg loss: 1.06199097, Global Avg Loss: 2.07615555, Time: 0.0402 Steps: 18600, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001578, Sample Num: 25248, Cur Loss: 2.12194133, Cur Avg Loss: 1.53575988, Log Avg loss: 1.48193470, Global Avg Loss: 2.07583625, Time: 0.0403 Steps: 18610, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001588, Sample Num: 25408, Cur Loss: 1.13884568, Cur Avg Loss: 1.53431099, Log Avg loss: 1.30567592, Global Avg Loss: 2.07542263, Time: 0.0402 Steps: 18620, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001598, Sample Num: 25568, Cur Loss: 1.50142503, Cur Avg Loss: 1.53502839, Log Avg loss: 1.64895205, Global Avg Loss: 2.07519371, Time: 0.0403 Steps: 18630, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001608, Sample Num: 25728, Cur Loss: 1.36381698, Cur Avg Loss: 1.53404909, Log Avg loss: 1.37755686, Global Avg Loss: 2.07481944, Time: 0.0403 Steps: 18640, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001618, Sample Num: 25888, Cur Loss: 3.18446684, Cur Avg Loss: 1.53485582, Log Avg loss: 1.66457713, Global Avg Loss: 2.07459947, Time: 0.0402 Steps: 18650, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001628, Sample Num: 26048, Cur Loss: 0.85067570, Cur Avg Loss: 1.53465231, Log Avg loss: 1.50172542, Global Avg Loss: 2.07429247, Time: 0.0403 Steps: 18660, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001638, Sample Num: 26208, Cur Loss: 1.01969194, Cur Avg Loss: 1.53441360, Log Avg loss: 1.49555156, Global Avg Loss: 2.07398248, Time: 0.0401 Steps: 18670, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001648, Sample Num: 26368, Cur Loss: 2.50350809, Cur Avg Loss: 1.53687137, Log Avg loss: 1.93945263, Global Avg Loss: 2.07391047, Time: 0.0403 Steps: 18680, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001658, Sample Num: 26528, Cur Loss: 1.52476406, Cur Avg Loss: 1.53851050, Log Avg loss: 1.80863978, Global Avg Loss: 2.07376853, Time: 0.0402 Steps: 18690, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001668, Sample Num: 26688, Cur Loss: 0.91164470, Cur Avg Loss: 1.53928695, Log Avg loss: 1.66802284, Global Avg Loss: 2.07355156, Time: 0.0402 Steps: 18700, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001678, Sample Num: 26848, Cur Loss: 1.00427997, Cur Avg Loss: 1.53847061, Log Avg loss: 1.40230522, Global Avg Loss: 2.07319279, Time: 0.0402 Steps: 18710, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001688, Sample Num: 27008, Cur Loss: 2.33233261, Cur Avg Loss: 1.54007911, Log Avg loss: 1.80998410, Global Avg Loss: 2.07305219, Time: 0.0402 Steps: 18720, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001698, Sample Num: 27168, Cur Loss: 1.59797847, Cur Avg Loss: 1.54099216, Log Avg loss: 1.69511571, Global Avg Loss: 2.07285041, Time: 0.0402 Steps: 18730, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001708, Sample Num: 27328, Cur Loss: 0.37452853, Cur Avg Loss: 1.53906036, Log Avg loss: 1.21104081, Global Avg Loss: 2.07239053, Time: 0.0402 Steps: 18740, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001718, Sample Num: 27488, Cur Loss: 1.45192242, Cur Avg Loss: 1.53942950, Log Avg loss: 1.60247917, Global Avg Loss: 2.07213991, Time: 0.0403 Steps: 18750, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001728, Sample Num: 27648, Cur Loss: 1.84095585, Cur Avg Loss: 1.53990101, Log Avg loss: 1.62090546, Global Avg Loss: 2.07189938, Time: 0.0402 Steps: 18760, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001738, Sample Num: 27808, Cur Loss: 2.31735563, Cur Avg Loss: 1.54002304, Log Avg loss: 1.56110942, Global Avg Loss: 2.07162725, Time: 0.0402 Steps: 18770, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001748, Sample Num: 27968, Cur Loss: 1.52373719, Cur Avg Loss: 1.54108173, Log Avg loss: 1.72508231, Global Avg Loss: 2.07144272, Time: 0.0402 Steps: 18780, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001758, Sample Num: 28128, Cur Loss: 1.27198410, Cur Avg Loss: 1.53837717, Log Avg loss: 1.06562087, Global Avg Loss: 2.07090743, Time: 0.0403 Steps: 18790, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001768, Sample Num: 28288, Cur Loss: 1.02402127, Cur Avg Loss: 1.53657322, Log Avg loss: 1.21943898, Global Avg Loss: 2.07045452, Time: 0.0402 Steps: 18800, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001778, Sample Num: 28448, Cur Loss: 1.11442018, Cur Avg Loss: 1.53655351, Log Avg loss: 1.53306880, Global Avg Loss: 2.07016883, Time: 0.0401 Steps: 18810, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001788, Sample Num: 28608, Cur Loss: 1.56228340, Cur Avg Loss: 1.53804686, Log Avg loss: 1.80356491, Global Avg Loss: 2.07002717, Time: 0.0402 Steps: 18820, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001798, Sample Num: 28768, Cur Loss: 0.97119272, Cur Avg Loss: 1.53729506, Log Avg loss: 1.40287186, Global Avg Loss: 2.06967286, Time: 0.0402 Steps: 18830, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001808, Sample Num: 28928, Cur Loss: 1.53355646, Cur Avg Loss: 1.53802246, Log Avg loss: 1.66880986, Global Avg Loss: 2.06946009, Time: 0.0585 Steps: 18840, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001818, Sample Num: 29088, Cur Loss: 1.35172915, Cur Avg Loss: 1.53935480, Log Avg loss: 1.78024230, Global Avg Loss: 2.06930666, Time: 0.0405 Steps: 18850, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001828, Sample Num: 29248, Cur Loss: 2.21131945, Cur Avg Loss: 1.53853264, Log Avg loss: 1.38906339, Global Avg Loss: 2.06894598, Time: 0.0776 Steps: 18860, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001838, Sample Num: 29408, Cur Loss: 2.99671793, Cur Avg Loss: 1.53740769, Log Avg loss: 1.33176707, Global Avg Loss: 2.06855532, Time: 0.0850 Steps: 18870, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001848, Sample Num: 29568, Cur Loss: 1.21554220, Cur Avg Loss: 1.53801707, Log Avg loss: 1.65001988, Global Avg Loss: 2.06833363, Time: 0.0406 Steps: 18880, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001858, Sample Num: 29728, Cur Loss: 2.13730383, Cur Avg Loss: 1.53872405, Log Avg loss: 1.66937459, Global Avg Loss: 2.06812243, Time: 0.0836 Steps: 18890, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001868, Sample Num: 29888, Cur Loss: 0.79334611, Cur Avg Loss: 1.53823196, Log Avg loss: 1.44680097, Global Avg Loss: 2.06779369, Time: 0.0582 Steps: 18900, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001878, Sample Num: 30048, Cur Loss: 0.28634325, Cur Avg Loss: 1.53662827, Log Avg loss: 1.23705907, Global Avg Loss: 2.06735438, Time: 0.0404 Steps: 18910, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001888, Sample Num: 30208, Cur Loss: 0.97636849, Cur Avg Loss: 1.53737566, Log Avg loss: 1.67773640, Global Avg Loss: 2.06714845, Time: 0.0555 Steps: 18920, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001898, Sample Num: 30368, Cur Loss: 1.62478137, Cur Avg Loss: 1.53751026, Log Avg loss: 1.56292338, Global Avg Loss: 2.06688209, Time: 0.0403 Steps: 18930, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001908, Sample Num: 30528, Cur Loss: 2.63472843, Cur Avg Loss: 1.53625307, Log Avg loss: 1.29763734, Global Avg Loss: 2.06647594, Time: 0.0404 Steps: 18940, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001918, Sample Num: 30688, Cur Loss: 1.13269162, Cur Avg Loss: 1.53592769, Log Avg loss: 1.47384477, Global Avg Loss: 2.06616321, Time: 0.0403 Steps: 18950, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001928, Sample Num: 30848, Cur Loss: 2.61735439, Cur Avg Loss: 1.53664756, Log Avg loss: 1.67471859, Global Avg Loss: 2.06595675, Time: 0.0402 Steps: 18960, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001938, Sample Num: 31008, Cur Loss: 1.74712992, Cur Avg Loss: 1.53781850, Log Avg loss: 1.76357689, Global Avg Loss: 2.06579735, Time: 0.0403 Steps: 18970, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001948, Sample Num: 31168, Cur Loss: 1.59933591, Cur Avg Loss: 1.53704040, Log Avg loss: 1.38624406, Global Avg Loss: 2.06543931, Time: 0.0403 Steps: 18980, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001958, Sample Num: 31328, Cur Loss: 1.53992939, Cur Avg Loss: 1.53808590, Log Avg loss: 1.74174991, Global Avg Loss: 2.06526886, Time: 0.0403 Steps: 18990, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001968, Sample Num: 31488, Cur Loss: 1.46750414, Cur Avg Loss: 1.53765486, Log Avg loss: 1.45325747, Global Avg Loss: 2.06494675, Time: 0.0403 Steps: 19000, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001978, Sample Num: 31648, Cur Loss: 0.41968715, Cur Avg Loss: 1.53841131, Log Avg loss: 1.68728023, Global Avg Loss: 2.06474808, Time: 0.0404 Steps: 19010, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001988, Sample Num: 31808, Cur Loss: 1.84097171, Cur Avg Loss: 1.53975733, Log Avg loss: 1.80599917, Global Avg Loss: 2.06461204, Time: 0.0402 Steps: 19020, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001998, Sample Num: 31968, Cur Loss: 0.89188403, Cur Avg Loss: 1.53897065, Log Avg loss: 1.38257906, Global Avg Loss: 2.06425364, Time: 0.0403 Steps: 19030, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002008, Sample Num: 32128, Cur Loss: 1.74265993, Cur Avg Loss: 1.54098749, Log Avg loss: 1.94395162, Global Avg Loss: 2.06419046, Time: 0.0403 Steps: 19040, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002018, Sample Num: 32288, Cur Loss: 1.36599755, Cur Avg Loss: 1.54131042, Log Avg loss: 1.60615579, Global Avg Loss: 2.06395002, Time: 0.0403 Steps: 19050, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002028, Sample Num: 32448, Cur Loss: 1.40922999, Cur Avg Loss: 1.54182776, Log Avg loss: 1.64622580, Global Avg Loss: 2.06373086, Time: 0.0402 Steps: 19060, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002038, Sample Num: 32608, Cur Loss: 2.05745077, Cur Avg Loss: 1.54225538, Log Avg loss: 1.62897736, Global Avg Loss: 2.06350288, Time: 0.0403 Steps: 19070, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002048, Sample Num: 32768, Cur Loss: 1.26788270, Cur Avg Loss: 1.54176682, Log Avg loss: 1.44219780, Global Avg Loss: 2.06317725, Time: 0.0404 Steps: 19080, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002058, Sample Num: 32928, Cur Loss: 2.62184477, Cur Avg Loss: 1.54128732, Log Avg loss: 1.44308664, Global Avg Loss: 2.06285242, Time: 0.0402 Steps: 19090, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002068, Sample Num: 33088, Cur Loss: 1.03058863, Cur Avg Loss: 1.54033599, Log Avg loss: 1.34455308, Global Avg Loss: 2.06247635, Time: 0.0402 Steps: 19100, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002078, Sample Num: 33248, Cur Loss: 1.23286903, Cur Avg Loss: 1.54093970, Log Avg loss: 1.66578693, Global Avg Loss: 2.06226877, Time: 0.0402 Steps: 19110, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002088, Sample Num: 33408, Cur Loss: 1.22740829, Cur Avg Loss: 1.54065272, Log Avg loss: 1.48101822, Global Avg Loss: 2.06196477, Time: 0.0403 Steps: 19120, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002098, Sample Num: 33568, Cur Loss: 0.89289874, Cur Avg Loss: 1.54092232, Log Avg loss: 1.59721484, Global Avg Loss: 2.06172183, Time: 0.0403 Steps: 19130, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002108, Sample Num: 33728, Cur Loss: 1.54149497, Cur Avg Loss: 1.54026534, Log Avg loss: 1.40242922, Global Avg Loss: 2.06137737, Time: 0.0402 Steps: 19140, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002118, Sample Num: 33888, Cur Loss: 2.94257617, Cur Avg Loss: 1.54150001, Log Avg loss: 1.80176839, Global Avg Loss: 2.06124180, Time: 0.0403 Steps: 19150, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002128, Sample Num: 34048, Cur Loss: 0.74406302, Cur Avg Loss: 1.54033385, Log Avg loss: 1.29334287, Global Avg Loss: 2.06084102, Time: 0.0403 Steps: 19160, Updated lr: 0.000083 ***** Running evaluation checkpoint-19161 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-19161 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.910193, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.523727, "eval_total_loss": 1071.180265, "eval_mae": 1.105774, "eval_mse": 1.52355, "eval_r2": 0.031531, "eval_sp_statistic": 0.328114, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.374051, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.414053, "test_total_loss": 709.854497, "test_mae": 1.013984, "test_mse": 1.41422, "test_r2": 0.08725, "test_sp_statistic": 0.338807, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.391943, "test_ps_pvalue": 0.0, "lr": 8.277761972498815e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.060874927263843, "train_cur_epoch_loss": 3280.541000276804, "train_cur_epoch_avg_loss": 1.5408835135165824, "train_cur_epoch_time": 95.9101927280426, "train_cur_epoch_avg_time": 0.045049409454223865, "epoch": 9, "step": 19161} ################################################## Training, Epoch: 0010, Batch: 000009, Sample Num: 144, Cur Loss: 0.90631330, Cur Avg Loss: 1.26767595, Log Avg loss: 1.41196417, Global Avg Loss: 2.06050253, Time: 0.0403 Steps: 19170, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000019, Sample Num: 304, Cur Loss: 1.38494742, Cur Avg Loss: 1.59332444, Log Avg loss: 1.88640808, Global Avg Loss: 2.06041176, Time: 0.0403 Steps: 19180, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000029, Sample Num: 464, Cur Loss: 1.86525893, Cur Avg Loss: 1.62610491, Log Avg loss: 1.68838782, Global Avg Loss: 2.06021790, Time: 0.0403 Steps: 19190, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000039, Sample Num: 624, Cur Loss: 0.88973963, Cur Avg Loss: 1.50614885, Log Avg loss: 1.15827626, Global Avg Loss: 2.05974814, Time: 0.0403 Steps: 19200, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000049, Sample Num: 784, Cur Loss: 0.72196734, Cur Avg Loss: 1.45369124, Log Avg loss: 1.24910656, Global Avg Loss: 2.05932615, Time: 0.0404 Steps: 19210, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000059, Sample Num: 944, Cur Loss: 1.30685043, Cur Avg Loss: 1.41980653, Log Avg loss: 1.25377144, Global Avg Loss: 2.05890703, Time: 0.0403 Steps: 19220, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000069, Sample Num: 1104, Cur Loss: 1.91133559, Cur Avg Loss: 1.51825305, Log Avg loss: 2.09908753, Global Avg Loss: 2.05892792, Time: 0.0403 Steps: 19230, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000079, Sample Num: 1264, Cur Loss: 0.88529003, Cur Avg Loss: 1.52250999, Log Avg loss: 1.55188289, Global Avg Loss: 2.05866439, Time: 0.0403 Steps: 19240, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000089, Sample Num: 1424, Cur Loss: 1.45030642, Cur Avg Loss: 1.51648080, Log Avg loss: 1.46885016, Global Avg Loss: 2.05835799, Time: 0.0529 Steps: 19250, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000099, Sample Num: 1584, Cur Loss: 2.33984590, Cur Avg Loss: 1.53207417, Log Avg loss: 1.67085519, Global Avg Loss: 2.05815679, Time: 0.0726 Steps: 19260, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000109, Sample Num: 1744, Cur Loss: 2.02558732, Cur Avg Loss: 1.54869148, Log Avg loss: 1.71320290, Global Avg Loss: 2.05797778, Time: 0.0656 Steps: 19270, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000119, Sample Num: 1904, Cur Loss: 1.78609526, Cur Avg Loss: 1.55127380, Log Avg loss: 1.57942103, Global Avg Loss: 2.05772957, Time: 0.0561 Steps: 19280, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000129, Sample Num: 2064, Cur Loss: 2.92466974, Cur Avg Loss: 1.54862207, Log Avg loss: 1.51706644, Global Avg Loss: 2.05744929, Time: 0.0403 Steps: 19290, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000139, Sample Num: 2224, Cur Loss: 1.11850858, Cur Avg Loss: 1.53921279, Log Avg loss: 1.41783308, Global Avg Loss: 2.05711788, Time: 0.0409 Steps: 19300, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000149, Sample Num: 2384, Cur Loss: 1.38551784, Cur Avg Loss: 1.54645917, Log Avg loss: 1.64718388, Global Avg Loss: 2.05690559, Time: 0.0683 Steps: 19310, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000159, Sample Num: 2544, Cur Loss: 2.53963399, Cur Avg Loss: 1.54696580, Log Avg loss: 1.55451468, Global Avg Loss: 2.05664555, Time: 0.0517 Steps: 19320, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000169, Sample Num: 2704, Cur Loss: 0.92647296, Cur Avg Loss: 1.52728851, Log Avg loss: 1.21441946, Global Avg Loss: 2.05620984, Time: 0.0501 Steps: 19330, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000179, Sample Num: 2864, Cur Loss: 1.29583907, Cur Avg Loss: 1.52656850, Log Avg loss: 1.51440031, Global Avg Loss: 2.05592969, Time: 0.0403 Steps: 19340, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000189, Sample Num: 3024, Cur Loss: 1.37006330, Cur Avg Loss: 1.52727728, Log Avg loss: 1.53996460, Global Avg Loss: 2.05566304, Time: 0.0403 Steps: 19350, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000199, Sample Num: 3184, Cur Loss: 2.07587242, Cur Avg Loss: 1.52746987, Log Avg loss: 1.53110976, Global Avg Loss: 2.05539210, Time: 0.0403 Steps: 19360, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000209, Sample Num: 3344, Cur Loss: 0.65740418, Cur Avg Loss: 1.53023094, Log Avg loss: 1.58517632, Global Avg Loss: 2.05514934, Time: 0.0404 Steps: 19370, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000219, Sample Num: 3504, Cur Loss: 1.08122635, Cur Avg Loss: 1.52075021, Log Avg loss: 1.32260292, Global Avg Loss: 2.05477135, Time: 0.0403 Steps: 19380, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000229, Sample Num: 3664, Cur Loss: 0.76582402, Cur Avg Loss: 1.50646558, Log Avg loss: 1.19363220, Global Avg Loss: 2.05432724, Time: 0.0403 Steps: 19390, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000239, Sample Num: 3824, Cur Loss: 0.71469784, Cur Avg Loss: 1.49988296, Log Avg loss: 1.34914083, Global Avg Loss: 2.05396374, Time: 0.0402 Steps: 19400, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000249, Sample Num: 3984, Cur Loss: 2.37811613, Cur Avg Loss: 1.51093676, Log Avg loss: 1.77512265, Global Avg Loss: 2.05382008, Time: 0.0402 Steps: 19410, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000259, Sample Num: 4144, Cur Loss: 0.25004885, Cur Avg Loss: 1.50635442, Log Avg loss: 1.39225410, Global Avg Loss: 2.05347942, Time: 0.0403 Steps: 19420, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000269, Sample Num: 4304, Cur Loss: 0.84311783, Cur Avg Loss: 1.51190922, Log Avg loss: 1.65577861, Global Avg Loss: 2.05327473, Time: 0.0403 Steps: 19430, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000279, Sample Num: 4464, Cur Loss: 1.91011333, Cur Avg Loss: 1.50149753, Log Avg loss: 1.22142295, Global Avg Loss: 2.05284683, Time: 0.0403 Steps: 19440, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000289, Sample Num: 4624, Cur Loss: 2.25445485, Cur Avg Loss: 1.50601406, Log Avg loss: 1.63202542, Global Avg Loss: 2.05263047, Time: 0.0402 Steps: 19450, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000299, Sample Num: 4784, Cur Loss: 1.14078832, Cur Avg Loss: 1.49697623, Log Avg loss: 1.23578296, Global Avg Loss: 2.05221071, Time: 0.0403 Steps: 19460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000309, Sample Num: 4944, Cur Loss: 0.91258830, Cur Avg Loss: 1.50340869, Log Avg loss: 1.69573913, Global Avg Loss: 2.05202762, Time: 0.0403 Steps: 19470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000319, Sample Num: 5104, Cur Loss: 0.69365555, Cur Avg Loss: 1.49864690, Log Avg loss: 1.35150743, Global Avg Loss: 2.05166801, Time: 0.0402 Steps: 19480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000329, Sample Num: 5264, Cur Loss: 1.26493073, Cur Avg Loss: 1.50278809, Log Avg loss: 1.63489206, Global Avg Loss: 2.05145417, Time: 0.0402 Steps: 19490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000339, Sample Num: 5424, Cur Loss: 1.78372145, Cur Avg Loss: 1.52014101, Log Avg loss: 2.09105215, Global Avg Loss: 2.05147448, Time: 0.0403 Steps: 19500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000349, Sample Num: 5584, Cur Loss: 1.73071015, Cur Avg Loss: 1.52413927, Log Avg loss: 1.65968042, Global Avg Loss: 2.05127366, Time: 0.0402 Steps: 19510, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000359, Sample Num: 5744, Cur Loss: 2.78624249, Cur Avg Loss: 1.52028830, Log Avg loss: 1.38588954, Global Avg Loss: 2.05093279, Time: 0.0403 Steps: 19520, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000369, Sample Num: 5904, Cur Loss: 1.15649402, Cur Avg Loss: 1.53214080, Log Avg loss: 1.95764537, Global Avg Loss: 2.05088502, Time: 0.0402 Steps: 19530, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000379, Sample Num: 6064, Cur Loss: 0.74347031, Cur Avg Loss: 1.52693619, Log Avg loss: 1.33488601, Global Avg Loss: 2.05051859, Time: 0.0403 Steps: 19540, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000389, Sample Num: 6224, Cur Loss: 1.28181267, Cur Avg Loss: 1.52906157, Log Avg loss: 1.60961355, Global Avg Loss: 2.05029307, Time: 0.0402 Steps: 19550, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000399, Sample Num: 6384, Cur Loss: 2.29204178, Cur Avg Loss: 1.52501831, Log Avg loss: 1.36773545, Global Avg Loss: 2.04994411, Time: 0.0402 Steps: 19560, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000409, Sample Num: 6544, Cur Loss: 2.08932734, Cur Avg Loss: 1.52785068, Log Avg loss: 1.64086223, Global Avg Loss: 2.04973507, Time: 0.0402 Steps: 19570, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000419, Sample Num: 6704, Cur Loss: 1.36227977, Cur Avg Loss: 1.52307848, Log Avg loss: 1.32789569, Global Avg Loss: 2.04936641, Time: 0.0402 Steps: 19580, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000429, Sample Num: 6864, Cur Loss: 1.09195483, Cur Avg Loss: 1.52738536, Log Avg loss: 1.70784354, Global Avg Loss: 2.04919208, Time: 0.0402 Steps: 19590, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000439, Sample Num: 7024, Cur Loss: 1.52307904, Cur Avg Loss: 1.53094739, Log Avg loss: 1.68375843, Global Avg Loss: 2.04900563, Time: 0.0405 Steps: 19600, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000449, Sample Num: 7184, Cur Loss: 1.34577310, Cur Avg Loss: 1.52849391, Log Avg loss: 1.42078605, Global Avg Loss: 2.04868528, Time: 0.0669 Steps: 19610, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000459, Sample Num: 7344, Cur Loss: 1.07020891, Cur Avg Loss: 1.53295287, Log Avg loss: 1.73316008, Global Avg Loss: 2.04852446, Time: 0.0563 Steps: 19620, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000469, Sample Num: 7504, Cur Loss: 2.31810856, Cur Avg Loss: 1.53868635, Log Avg loss: 1.80185339, Global Avg Loss: 2.04839880, Time: 0.0999 Steps: 19630, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000479, Sample Num: 7664, Cur Loss: 1.38292491, Cur Avg Loss: 1.53881541, Log Avg loss: 1.54486842, Global Avg Loss: 2.04814242, Time: 0.0432 Steps: 19640, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000489, Sample Num: 7824, Cur Loss: 3.35365868, Cur Avg Loss: 1.54161089, Log Avg loss: 1.67551398, Global Avg Loss: 2.04795278, Time: 0.0411 Steps: 19650, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000499, Sample Num: 7984, Cur Loss: 1.70420599, Cur Avg Loss: 1.54328850, Log Avg loss: 1.62532391, Global Avg Loss: 2.04773782, Time: 0.0888 Steps: 19660, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000509, Sample Num: 8144, Cur Loss: 1.27463508, Cur Avg Loss: 1.54072809, Log Avg loss: 1.41296359, Global Avg Loss: 2.04741510, Time: 0.0430 Steps: 19670, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000519, Sample Num: 8304, Cur Loss: 1.12282372, Cur Avg Loss: 1.53008257, Log Avg loss: 0.98822575, Global Avg Loss: 2.04687690, Time: 0.0641 Steps: 19680, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000529, Sample Num: 8464, Cur Loss: 3.98412895, Cur Avg Loss: 1.53145674, Log Avg loss: 1.60277606, Global Avg Loss: 2.04665135, Time: 0.0402 Steps: 19690, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000539, Sample Num: 8624, Cur Loss: 1.46119618, Cur Avg Loss: 1.53408343, Log Avg loss: 1.67303506, Global Avg Loss: 2.04646170, Time: 0.0402 Steps: 19700, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000549, Sample Num: 8784, Cur Loss: 2.12864947, Cur Avg Loss: 1.53306564, Log Avg loss: 1.47820688, Global Avg Loss: 2.04617339, Time: 0.0403 Steps: 19710, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000559, Sample Num: 8944, Cur Loss: 1.43686199, Cur Avg Loss: 1.52771832, Log Avg loss: 1.23415077, Global Avg Loss: 2.04576161, Time: 0.0402 Steps: 19720, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000569, Sample Num: 9104, Cur Loss: 2.42058420, Cur Avg Loss: 1.52356632, Log Avg loss: 1.29146952, Global Avg Loss: 2.04537931, Time: 0.0403 Steps: 19730, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000579, Sample Num: 9264, Cur Loss: 2.11023045, Cur Avg Loss: 1.52217731, Log Avg loss: 1.44314218, Global Avg Loss: 2.04507422, Time: 0.0402 Steps: 19740, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000589, Sample Num: 9424, Cur Loss: 1.73543668, Cur Avg Loss: 1.52203974, Log Avg loss: 1.51407476, Global Avg Loss: 2.04480536, Time: 0.0403 Steps: 19750, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000599, Sample Num: 9584, Cur Loss: 1.15727544, Cur Avg Loss: 1.52202772, Log Avg loss: 1.52131979, Global Avg Loss: 2.04454044, Time: 0.0402 Steps: 19760, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000609, Sample Num: 9744, Cur Loss: 1.55864525, Cur Avg Loss: 1.52552557, Log Avg loss: 1.73504658, Global Avg Loss: 2.04438389, Time: 0.0403 Steps: 19770, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000619, Sample Num: 9904, Cur Loss: 1.09297323, Cur Avg Loss: 1.52549657, Log Avg loss: 1.52373059, Global Avg Loss: 2.04412067, Time: 0.0403 Steps: 19780, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000629, Sample Num: 10064, Cur Loss: 2.33793974, Cur Avg Loss: 1.52860047, Log Avg loss: 1.72073171, Global Avg Loss: 2.04395726, Time: 0.0402 Steps: 19790, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000639, Sample Num: 10224, Cur Loss: 2.14287758, Cur Avg Loss: 1.53105233, Log Avg loss: 1.68527417, Global Avg Loss: 2.04377611, Time: 0.0402 Steps: 19800, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000649, Sample Num: 10384, Cur Loss: 1.79253626, Cur Avg Loss: 1.53077603, Log Avg loss: 1.51312054, Global Avg Loss: 2.04350823, Time: 0.0403 Steps: 19810, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000659, Sample Num: 10544, Cur Loss: 1.95301342, Cur Avg Loss: 1.53361226, Log Avg loss: 1.71768389, Global Avg Loss: 2.04334384, Time: 0.0402 Steps: 19820, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000669, Sample Num: 10704, Cur Loss: 2.60572004, Cur Avg Loss: 1.53272581, Log Avg loss: 1.47430849, Global Avg Loss: 2.04305689, Time: 0.0402 Steps: 19830, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000679, Sample Num: 10864, Cur Loss: 3.17653656, Cur Avg Loss: 1.53884190, Log Avg loss: 1.94800817, Global Avg Loss: 2.04300898, Time: 0.0402 Steps: 19840, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000689, Sample Num: 11024, Cur Loss: 1.86293077, Cur Avg Loss: 1.54497499, Log Avg loss: 1.96141235, Global Avg Loss: 2.04296787, Time: 0.0402 Steps: 19850, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000699, Sample Num: 11184, Cur Loss: 1.85480213, Cur Avg Loss: 1.54454863, Log Avg loss: 1.51517235, Global Avg Loss: 2.04270211, Time: 0.0402 Steps: 19860, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000709, Sample Num: 11344, Cur Loss: 1.12852097, Cur Avg Loss: 1.54225693, Log Avg loss: 1.38206697, Global Avg Loss: 2.04236963, Time: 0.0402 Steps: 19870, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000719, Sample Num: 11504, Cur Loss: 0.95406729, Cur Avg Loss: 1.53906534, Log Avg loss: 1.31278123, Global Avg Loss: 2.04200264, Time: 0.0402 Steps: 19880, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000729, Sample Num: 11664, Cur Loss: 0.77303588, Cur Avg Loss: 1.53663621, Log Avg loss: 1.36198187, Global Avg Loss: 2.04166075, Time: 0.0403 Steps: 19890, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000739, Sample Num: 11824, Cur Loss: 1.31110120, Cur Avg Loss: 1.53598147, Log Avg loss: 1.48825148, Global Avg Loss: 2.04138265, Time: 0.0403 Steps: 19900, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000749, Sample Num: 11984, Cur Loss: 2.03228498, Cur Avg Loss: 1.53499009, Log Avg loss: 1.46172652, Global Avg Loss: 2.04109151, Time: 0.0403 Steps: 19910, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000759, Sample Num: 12144, Cur Loss: 1.04695332, Cur Avg Loss: 1.53365083, Log Avg loss: 1.43334076, Global Avg Loss: 2.04078642, Time: 0.0402 Steps: 19920, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000769, Sample Num: 12304, Cur Loss: 2.32155418, Cur Avg Loss: 1.53348576, Log Avg loss: 1.52095679, Global Avg Loss: 2.04052559, Time: 0.0402 Steps: 19930, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000779, Sample Num: 12464, Cur Loss: 1.24561954, Cur Avg Loss: 1.53343752, Log Avg loss: 1.52972798, Global Avg Loss: 2.04026942, Time: 0.0402 Steps: 19940, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000789, Sample Num: 12624, Cur Loss: 1.56327128, Cur Avg Loss: 1.53469946, Log Avg loss: 1.63300468, Global Avg Loss: 2.04006528, Time: 0.0568 Steps: 19950, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000799, Sample Num: 12784, Cur Loss: 1.04644954, Cur Avg Loss: 1.53601408, Log Avg loss: 1.63973773, Global Avg Loss: 2.03986472, Time: 0.0406 Steps: 19960, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000809, Sample Num: 12944, Cur Loss: 0.97967923, Cur Avg Loss: 1.53196701, Log Avg loss: 1.20860606, Global Avg Loss: 2.03944846, Time: 0.0590 Steps: 19970, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000819, Sample Num: 13104, Cur Loss: 1.59677243, Cur Avg Loss: 1.53208789, Log Avg loss: 1.54186682, Global Avg Loss: 2.03919942, Time: 0.0985 Steps: 19980, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000829, Sample Num: 13264, Cur Loss: 1.41804790, Cur Avg Loss: 1.53120627, Log Avg loss: 1.45900165, Global Avg Loss: 2.03890918, Time: 0.0537 Steps: 19990, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000839, Sample Num: 13424, Cur Loss: 1.96427393, Cur Avg Loss: 1.53477957, Log Avg loss: 1.83100621, Global Avg Loss: 2.03880523, Time: 0.0506 Steps: 20000, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000849, Sample Num: 13584, Cur Loss: 1.39279640, Cur Avg Loss: 1.53421917, Log Avg loss: 1.48720174, Global Avg Loss: 2.03852956, Time: 0.0694 Steps: 20010, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000859, Sample Num: 13744, Cur Loss: 2.39829588, Cur Avg Loss: 1.53487337, Log Avg loss: 1.59041467, Global Avg Loss: 2.03830573, Time: 0.0586 Steps: 20020, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000869, Sample Num: 13904, Cur Loss: 1.60058999, Cur Avg Loss: 1.53571872, Log Avg loss: 1.60833400, Global Avg Loss: 2.03809107, Time: 0.0544 Steps: 20030, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000879, Sample Num: 14064, Cur Loss: 0.63310993, Cur Avg Loss: 1.53387941, Log Avg loss: 1.37404368, Global Avg Loss: 2.03775970, Time: 0.0402 Steps: 20040, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000889, Sample Num: 14224, Cur Loss: 1.45246518, Cur Avg Loss: 1.53434604, Log Avg loss: 1.57536289, Global Avg Loss: 2.03752908, Time: 0.0402 Steps: 20050, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000899, Sample Num: 14384, Cur Loss: 1.73457849, Cur Avg Loss: 1.53735118, Log Avg loss: 1.80450783, Global Avg Loss: 2.03741292, Time: 0.0402 Steps: 20060, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000909, Sample Num: 14544, Cur Loss: 1.46204591, Cur Avg Loss: 1.54054832, Log Avg loss: 1.82797091, Global Avg Loss: 2.03730857, Time: 0.0402 Steps: 20070, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000919, Sample Num: 14704, Cur Loss: 0.85838783, Cur Avg Loss: 1.53832903, Log Avg loss: 1.33659639, Global Avg Loss: 2.03695960, Time: 0.0402 Steps: 20080, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000929, Sample Num: 14864, Cur Loss: 1.34586442, Cur Avg Loss: 1.53753781, Log Avg loss: 1.46482393, Global Avg Loss: 2.03667482, Time: 0.0402 Steps: 20090, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000939, Sample Num: 15024, Cur Loss: 1.90387011, Cur Avg Loss: 1.53884194, Log Avg loss: 1.65999557, Global Avg Loss: 2.03648742, Time: 0.0402 Steps: 20100, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000949, Sample Num: 15184, Cur Loss: 1.10411525, Cur Avg Loss: 1.54154615, Log Avg loss: 1.79547232, Global Avg Loss: 2.03636757, Time: 0.0402 Steps: 20110, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000959, Sample Num: 15344, Cur Loss: 1.73914552, Cur Avg Loss: 1.54299918, Log Avg loss: 1.68089129, Global Avg Loss: 2.03619089, Time: 0.0402 Steps: 20120, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000969, Sample Num: 15504, Cur Loss: 1.11175966, Cur Avg Loss: 1.54183092, Log Avg loss: 1.42979477, Global Avg Loss: 2.03588965, Time: 0.0401 Steps: 20130, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000979, Sample Num: 15664, Cur Loss: 2.81484580, Cur Avg Loss: 1.54674638, Log Avg loss: 2.02305478, Global Avg Loss: 2.03588328, Time: 0.0403 Steps: 20140, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000989, Sample Num: 15824, Cur Loss: 3.07126832, Cur Avg Loss: 1.54953662, Log Avg loss: 1.82270057, Global Avg Loss: 2.03577748, Time: 0.0403 Steps: 20150, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000999, Sample Num: 15984, Cur Loss: 1.77239132, Cur Avg Loss: 1.55078896, Log Avg loss: 1.67464536, Global Avg Loss: 2.03559835, Time: 0.0403 Steps: 20160, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001009, Sample Num: 16144, Cur Loss: 2.14946127, Cur Avg Loss: 1.55010559, Log Avg loss: 1.48183692, Global Avg Loss: 2.03532380, Time: 0.0402 Steps: 20170, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001019, Sample Num: 16304, Cur Loss: 1.60197163, Cur Avg Loss: 1.55299437, Log Avg loss: 1.84447232, Global Avg Loss: 2.03522922, Time: 0.0402 Steps: 20180, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001029, Sample Num: 16464, Cur Loss: 2.83102322, Cur Avg Loss: 1.55393217, Log Avg loss: 1.64949409, Global Avg Loss: 2.03503817, Time: 0.0402 Steps: 20190, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001039, Sample Num: 16624, Cur Loss: 2.58433723, Cur Avg Loss: 1.55683729, Log Avg loss: 1.85577458, Global Avg Loss: 2.03494943, Time: 0.0403 Steps: 20200, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001049, Sample Num: 16784, Cur Loss: 1.72887969, Cur Avg Loss: 1.55476642, Log Avg loss: 1.33960292, Global Avg Loss: 2.03460537, Time: 0.0402 Steps: 20210, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001059, Sample Num: 16944, Cur Loss: 1.88063741, Cur Avg Loss: 1.55437200, Log Avg loss: 1.51299705, Global Avg Loss: 2.03434740, Time: 0.0402 Steps: 20220, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001069, Sample Num: 17104, Cur Loss: 1.00708055, Cur Avg Loss: 1.55379126, Log Avg loss: 1.49229137, Global Avg Loss: 2.03407945, Time: 0.0402 Steps: 20230, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001079, Sample Num: 17264, Cur Loss: 1.52778912, Cur Avg Loss: 1.55046989, Log Avg loss: 1.19541492, Global Avg Loss: 2.03366509, Time: 0.0402 Steps: 20240, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001089, Sample Num: 17424, Cur Loss: 1.81472254, Cur Avg Loss: 1.55170424, Log Avg loss: 1.68489081, Global Avg Loss: 2.03349286, Time: 0.0402 Steps: 20250, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001099, Sample Num: 17584, Cur Loss: 0.50497764, Cur Avg Loss: 1.54928481, Log Avg loss: 1.28580894, Global Avg Loss: 2.03312381, Time: 0.0402 Steps: 20260, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001109, Sample Num: 17744, Cur Loss: 0.99631000, Cur Avg Loss: 1.54954378, Log Avg loss: 1.57800487, Global Avg Loss: 2.03289929, Time: 0.0403 Steps: 20270, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001119, Sample Num: 17904, Cur Loss: 1.27458262, Cur Avg Loss: 1.54864545, Log Avg loss: 1.44902013, Global Avg Loss: 2.03261138, Time: 0.0403 Steps: 20280, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001129, Sample Num: 18064, Cur Loss: 2.84618974, Cur Avg Loss: 1.54780547, Log Avg loss: 1.45381179, Global Avg Loss: 2.03232611, Time: 0.0402 Steps: 20290, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001139, Sample Num: 18224, Cur Loss: 1.57604373, Cur Avg Loss: 1.54849976, Log Avg loss: 1.62688510, Global Avg Loss: 2.03212639, Time: 0.0406 Steps: 20300, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001149, Sample Num: 18384, Cur Loss: 1.62785959, Cur Avg Loss: 1.54950662, Log Avg loss: 1.66418787, Global Avg Loss: 2.03194523, Time: 0.0588 Steps: 20310, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001159, Sample Num: 18544, Cur Loss: 1.93791091, Cur Avg Loss: 1.54868479, Log Avg loss: 1.45425637, Global Avg Loss: 2.03166093, Time: 0.0614 Steps: 20320, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001169, Sample Num: 18704, Cur Loss: 0.53986132, Cur Avg Loss: 1.54620995, Log Avg loss: 1.25937612, Global Avg Loss: 2.03128106, Time: 0.0406 Steps: 20330, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001179, Sample Num: 18864, Cur Loss: 1.19997096, Cur Avg Loss: 1.54570644, Log Avg loss: 1.48684659, Global Avg Loss: 2.03101339, Time: 0.0512 Steps: 20340, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001189, Sample Num: 19024, Cur Loss: 2.64918065, Cur Avg Loss: 1.54916622, Log Avg loss: 1.95707336, Global Avg Loss: 2.03097706, Time: 0.0407 Steps: 20350, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001199, Sample Num: 19184, Cur Loss: 1.25074255, Cur Avg Loss: 1.54945290, Log Avg loss: 1.58354003, Global Avg Loss: 2.03075729, Time: 0.0739 Steps: 20360, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001209, Sample Num: 19344, Cur Loss: 1.98590469, Cur Avg Loss: 1.54905160, Log Avg loss: 1.50093577, Global Avg Loss: 2.03049720, Time: 0.0522 Steps: 20370, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001219, Sample Num: 19504, Cur Loss: 1.41621566, Cur Avg Loss: 1.55101046, Log Avg loss: 1.78783619, Global Avg Loss: 2.03037813, Time: 0.0406 Steps: 20380, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001229, Sample Num: 19664, Cur Loss: 0.67817974, Cur Avg Loss: 1.54970849, Log Avg loss: 1.39099791, Global Avg Loss: 2.03006455, Time: 0.0402 Steps: 20390, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001239, Sample Num: 19824, Cur Loss: 1.41098428, Cur Avg Loss: 1.54889492, Log Avg loss: 1.44890717, Global Avg Loss: 2.02977967, Time: 0.0402 Steps: 20400, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001249, Sample Num: 19984, Cur Loss: 1.84077930, Cur Avg Loss: 1.55258605, Log Avg loss: 2.00991738, Global Avg Loss: 2.02976994, Time: 0.0403 Steps: 20410, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001259, Sample Num: 20144, Cur Loss: 0.94422811, Cur Avg Loss: 1.55288448, Log Avg loss: 1.59015799, Global Avg Loss: 2.02955465, Time: 0.0402 Steps: 20420, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001269, Sample Num: 20304, Cur Loss: 1.02378178, Cur Avg Loss: 1.55209316, Log Avg loss: 1.45246663, Global Avg Loss: 2.02927218, Time: 0.0402 Steps: 20430, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001279, Sample Num: 20464, Cur Loss: 1.92185438, Cur Avg Loss: 1.55485801, Log Avg loss: 1.90571700, Global Avg Loss: 2.02921174, Time: 0.0402 Steps: 20440, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001289, Sample Num: 20624, Cur Loss: 1.84753656, Cur Avg Loss: 1.55400657, Log Avg loss: 1.44510797, Global Avg Loss: 2.02892611, Time: 0.0402 Steps: 20450, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001299, Sample Num: 20784, Cur Loss: 0.90376675, Cur Avg Loss: 1.55550222, Log Avg loss: 1.74829147, Global Avg Loss: 2.02878895, Time: 0.0403 Steps: 20460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001309, Sample Num: 20944, Cur Loss: 1.55078483, Cur Avg Loss: 1.55460454, Log Avg loss: 1.43799636, Global Avg Loss: 2.02850033, Time: 0.0403 Steps: 20470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001319, Sample Num: 21104, Cur Loss: 1.06642127, Cur Avg Loss: 1.55420285, Log Avg loss: 1.50162109, Global Avg Loss: 2.02824307, Time: 0.0402 Steps: 20480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001329, Sample Num: 21264, Cur Loss: 0.63023895, Cur Avg Loss: 1.55505151, Log Avg loss: 1.66698920, Global Avg Loss: 2.02806676, Time: 0.0402 Steps: 20490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001339, Sample Num: 21424, Cur Loss: 1.74581623, Cur Avg Loss: 1.55358282, Log Avg loss: 1.35839447, Global Avg Loss: 2.02774009, Time: 0.0402 Steps: 20500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001349, Sample Num: 21584, Cur Loss: 0.78338933, Cur Avg Loss: 1.55297571, Log Avg loss: 1.47168303, Global Avg Loss: 2.02746898, Time: 0.0402 Steps: 20510, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001359, Sample Num: 21744, Cur Loss: 2.98203230, Cur Avg Loss: 1.55204760, Log Avg loss: 1.42684619, Global Avg Loss: 2.02717628, Time: 0.0402 Steps: 20520, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001369, Sample Num: 21904, Cur Loss: 1.44868493, Cur Avg Loss: 1.55218014, Log Avg loss: 1.57019233, Global Avg Loss: 2.02695368, Time: 0.0403 Steps: 20530, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001379, Sample Num: 22064, Cur Loss: 2.63321400, Cur Avg Loss: 1.55357799, Log Avg loss: 1.74494331, Global Avg Loss: 2.02681638, Time: 0.0402 Steps: 20540, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001389, Sample Num: 22224, Cur Loss: 0.92165613, Cur Avg Loss: 1.55262404, Log Avg loss: 1.42107514, Global Avg Loss: 2.02652162, Time: 0.0402 Steps: 20550, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001399, Sample Num: 22384, Cur Loss: 4.51944876, Cur Avg Loss: 1.55395385, Log Avg loss: 1.73866340, Global Avg Loss: 2.02638161, Time: 0.0403 Steps: 20560, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001409, Sample Num: 22544, Cur Loss: 2.68092942, Cur Avg Loss: 1.55608824, Log Avg loss: 1.85468953, Global Avg Loss: 2.02629814, Time: 0.0402 Steps: 20570, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001419, Sample Num: 22704, Cur Loss: 1.24372923, Cur Avg Loss: 1.55446585, Log Avg loss: 1.32587175, Global Avg Loss: 2.02595780, Time: 0.0402 Steps: 20580, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001429, Sample Num: 22864, Cur Loss: 3.10445309, Cur Avg Loss: 1.55432427, Log Avg loss: 1.53423395, Global Avg Loss: 2.02571898, Time: 0.0402 Steps: 20590, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001439, Sample Num: 23024, Cur Loss: 1.26808918, Cur Avg Loss: 1.55128039, Log Avg loss: 1.11631055, Global Avg Loss: 2.02527752, Time: 0.0403 Steps: 20600, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001449, Sample Num: 23184, Cur Loss: 2.04944015, Cur Avg Loss: 1.55183977, Log Avg loss: 1.63233408, Global Avg Loss: 2.02508687, Time: 0.0402 Steps: 20610, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001459, Sample Num: 23344, Cur Loss: 0.47235426, Cur Avg Loss: 1.55091577, Log Avg loss: 1.41702807, Global Avg Loss: 2.02479198, Time: 0.0402 Steps: 20620, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001469, Sample Num: 23504, Cur Loss: 2.55108738, Cur Avg Loss: 1.54994383, Log Avg loss: 1.40813764, Global Avg Loss: 2.02449307, Time: 0.0402 Steps: 20630, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001479, Sample Num: 23664, Cur Loss: 1.44252920, Cur Avg Loss: 1.55039059, Log Avg loss: 1.61602029, Global Avg Loss: 2.02429516, Time: 0.0707 Steps: 20640, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001489, Sample Num: 23824, Cur Loss: 1.01494789, Cur Avg Loss: 1.55018590, Log Avg loss: 1.51991138, Global Avg Loss: 2.02405091, Time: 0.0859 Steps: 20650, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001499, Sample Num: 23984, Cur Loss: 2.21416521, Cur Avg Loss: 1.55121142, Log Avg loss: 1.70391127, Global Avg Loss: 2.02389595, Time: 0.0407 Steps: 20660, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001509, Sample Num: 24144, Cur Loss: 2.19962263, Cur Avg Loss: 1.55347022, Log Avg loss: 1.89206430, Global Avg Loss: 2.02383217, Time: 0.0407 Steps: 20670, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001519, Sample Num: 24304, Cur Loss: 1.11843681, Cur Avg Loss: 1.55433735, Log Avg loss: 1.68518714, Global Avg Loss: 2.02366842, Time: 0.0401 Steps: 20680, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001529, Sample Num: 24464, Cur Loss: 1.00512695, Cur Avg Loss: 1.55597826, Log Avg loss: 1.80523299, Global Avg Loss: 2.02356284, Time: 0.0407 Steps: 20690, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001539, Sample Num: 24624, Cur Loss: 0.87090158, Cur Avg Loss: 1.55670670, Log Avg loss: 1.66808507, Global Avg Loss: 2.02339112, Time: 0.0501 Steps: 20700, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001549, Sample Num: 24784, Cur Loss: 1.86510324, Cur Avg Loss: 1.55819154, Log Avg loss: 1.78670911, Global Avg Loss: 2.02327683, Time: 0.0567 Steps: 20710, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001559, Sample Num: 24944, Cur Loss: 1.07942474, Cur Avg Loss: 1.55785549, Log Avg loss: 1.50580006, Global Avg Loss: 2.02302708, Time: 0.0406 Steps: 20720, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001569, Sample Num: 25104, Cur Loss: 0.74889553, Cur Avg Loss: 1.55764852, Log Avg loss: 1.52538258, Global Avg Loss: 2.02278702, Time: 0.0402 Steps: 20730, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001579, Sample Num: 25264, Cur Loss: 1.81330478, Cur Avg Loss: 1.55746740, Log Avg loss: 1.52904954, Global Avg Loss: 2.02254896, Time: 0.0402 Steps: 20740, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001589, Sample Num: 25424, Cur Loss: 2.26193714, Cur Avg Loss: 1.55693946, Log Avg loss: 1.47357860, Global Avg Loss: 2.02228440, Time: 0.0402 Steps: 20750, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001599, Sample Num: 25584, Cur Loss: 2.10276031, Cur Avg Loss: 1.55595623, Log Avg loss: 1.39972105, Global Avg Loss: 2.02198451, Time: 0.0403 Steps: 20760, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001609, Sample Num: 25744, Cur Loss: 1.10588586, Cur Avg Loss: 1.55420730, Log Avg loss: 1.27455274, Global Avg Loss: 2.02162465, Time: 0.0402 Steps: 20770, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001619, Sample Num: 25904, Cur Loss: 1.35303605, Cur Avg Loss: 1.55397665, Log Avg loss: 1.51686507, Global Avg Loss: 2.02138175, Time: 0.0402 Steps: 20780, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001629, Sample Num: 26064, Cur Loss: 2.19471836, Cur Avg Loss: 1.55566266, Log Avg loss: 1.82862730, Global Avg Loss: 2.02128903, Time: 0.0402 Steps: 20790, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001639, Sample Num: 26224, Cur Loss: 1.56890535, Cur Avg Loss: 1.55851716, Log Avg loss: 2.02351619, Global Avg Loss: 2.02129010, Time: 0.0402 Steps: 20800, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001649, Sample Num: 26384, Cur Loss: 2.90292311, Cur Avg Loss: 1.55848882, Log Avg loss: 1.55384386, Global Avg Loss: 2.02106548, Time: 0.0402 Steps: 20810, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001659, Sample Num: 26544, Cur Loss: 1.57405388, Cur Avg Loss: 1.55834533, Log Avg loss: 1.53468297, Global Avg Loss: 2.02083186, Time: 0.0403 Steps: 20820, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001669, Sample Num: 26704, Cur Loss: 0.92242789, Cur Avg Loss: 1.55778536, Log Avg loss: 1.46488711, Global Avg Loss: 2.02056497, Time: 0.0403 Steps: 20830, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001679, Sample Num: 26864, Cur Loss: 1.46014524, Cur Avg Loss: 1.55724737, Log Avg loss: 1.46745593, Global Avg Loss: 2.02029956, Time: 0.0403 Steps: 20840, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001689, Sample Num: 27024, Cur Loss: 0.34775406, Cur Avg Loss: 1.55657093, Log Avg loss: 1.44299718, Global Avg Loss: 2.02002268, Time: 0.0403 Steps: 20850, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001699, Sample Num: 27184, Cur Loss: 0.61327690, Cur Avg Loss: 1.55539430, Log Avg loss: 1.35666208, Global Avg Loss: 2.01970467, Time: 0.0402 Steps: 20860, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001709, Sample Num: 27344, Cur Loss: 1.35061574, Cur Avg Loss: 1.55688762, Log Avg loss: 1.81060249, Global Avg Loss: 2.01960448, Time: 0.0402 Steps: 20870, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001719, Sample Num: 27504, Cur Loss: 1.48897755, Cur Avg Loss: 1.55719875, Log Avg loss: 1.61037052, Global Avg Loss: 2.01940848, Time: 0.0402 Steps: 20880, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001729, Sample Num: 27664, Cur Loss: 1.43297458, Cur Avg Loss: 1.55783382, Log Avg loss: 1.66700152, Global Avg Loss: 2.01923979, Time: 0.0402 Steps: 20890, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001739, Sample Num: 27824, Cur Loss: 1.56215882, Cur Avg Loss: 1.55746135, Log Avg loss: 1.49306169, Global Avg Loss: 2.01898803, Time: 0.0402 Steps: 20900, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001749, Sample Num: 27984, Cur Loss: 1.08294249, Cur Avg Loss: 1.55713999, Log Avg loss: 1.50125542, Global Avg Loss: 2.01874043, Time: 0.0402 Steps: 20910, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001759, Sample Num: 28144, Cur Loss: 0.59940159, Cur Avg Loss: 1.55631979, Log Avg loss: 1.41286706, Global Avg Loss: 2.01845081, Time: 0.0401 Steps: 20920, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001769, Sample Num: 28304, Cur Loss: 1.17384195, Cur Avg Loss: 1.55687010, Log Avg loss: 1.65366876, Global Avg Loss: 2.01827653, Time: 0.0402 Steps: 20930, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001779, Sample Num: 28464, Cur Loss: 1.78633118, Cur Avg Loss: 1.55644036, Log Avg loss: 1.48042068, Global Avg Loss: 2.01801967, Time: 0.0402 Steps: 20940, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001789, Sample Num: 28624, Cur Loss: 1.61820698, Cur Avg Loss: 1.55529367, Log Avg loss: 1.35129662, Global Avg Loss: 2.01770142, Time: 0.0402 Steps: 20950, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001799, Sample Num: 28784, Cur Loss: 1.86281967, Cur Avg Loss: 1.55562036, Log Avg loss: 1.61406517, Global Avg Loss: 2.01750885, Time: 0.0402 Steps: 20960, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001809, Sample Num: 28944, Cur Loss: 0.91793883, Cur Avg Loss: 1.55432013, Log Avg loss: 1.32040932, Global Avg Loss: 2.01717642, Time: 0.0402 Steps: 20970, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001819, Sample Num: 29104, Cur Loss: 2.54019237, Cur Avg Loss: 1.55818615, Log Avg loss: 2.25754900, Global Avg Loss: 2.01729100, Time: 0.0401 Steps: 20980, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001829, Sample Num: 29264, Cur Loss: 1.09112740, Cur Avg Loss: 1.55826194, Log Avg loss: 1.57204759, Global Avg Loss: 2.01707887, Time: 0.0407 Steps: 20990, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001839, Sample Num: 29424, Cur Loss: 1.38162827, Cur Avg Loss: 1.55700180, Log Avg loss: 1.32652292, Global Avg Loss: 2.01675004, Time: 0.0402 Steps: 21000, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001849, Sample Num: 29584, Cur Loss: 1.64070237, Cur Avg Loss: 1.55696027, Log Avg loss: 1.54932337, Global Avg Loss: 2.01652756, Time: 0.0405 Steps: 21010, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001859, Sample Num: 29744, Cur Loss: 1.32963586, Cur Avg Loss: 1.55530975, Log Avg loss: 1.25012866, Global Avg Loss: 2.01616295, Time: 0.0403 Steps: 21020, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001869, Sample Num: 29904, Cur Loss: 0.59716523, Cur Avg Loss: 1.55568319, Log Avg loss: 1.62510481, Global Avg Loss: 2.01597700, Time: 0.0916 Steps: 21030, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001879, Sample Num: 30064, Cur Loss: 0.64740354, Cur Avg Loss: 1.55555608, Log Avg loss: 1.53180034, Global Avg Loss: 2.01574688, Time: 0.0431 Steps: 21040, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001889, Sample Num: 30224, Cur Loss: 2.32288790, Cur Avg Loss: 1.55718804, Log Avg loss: 1.86383167, Global Avg Loss: 2.01567471, Time: 0.0721 Steps: 21050, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001899, Sample Num: 30384, Cur Loss: 0.98837316, Cur Avg Loss: 1.55837156, Log Avg loss: 1.78193975, Global Avg Loss: 2.01556373, Time: 0.0560 Steps: 21060, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001909, Sample Num: 30544, Cur Loss: 0.84029830, Cur Avg Loss: 1.55782509, Log Avg loss: 1.45405017, Global Avg Loss: 2.01529723, Time: 0.0406 Steps: 21070, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001919, Sample Num: 30704, Cur Loss: 1.15794468, Cur Avg Loss: 1.55835463, Log Avg loss: 1.65944426, Global Avg Loss: 2.01512842, Time: 0.0402 Steps: 21080, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001929, Sample Num: 30864, Cur Loss: 1.06600845, Cur Avg Loss: 1.55840558, Log Avg loss: 1.56818127, Global Avg Loss: 2.01491649, Time: 0.0402 Steps: 21090, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001939, Sample Num: 31024, Cur Loss: 1.47874796, Cur Avg Loss: 1.55795676, Log Avg loss: 1.47138111, Global Avg Loss: 2.01465889, Time: 0.0402 Steps: 21100, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001949, Sample Num: 31184, Cur Loss: 1.08840799, Cur Avg Loss: 1.55758196, Log Avg loss: 1.48490664, Global Avg Loss: 2.01440794, Time: 0.0402 Steps: 21110, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001959, Sample Num: 31344, Cur Loss: 1.37190974, Cur Avg Loss: 1.55738570, Log Avg loss: 1.51913577, Global Avg Loss: 2.01417344, Time: 0.0402 Steps: 21120, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001969, Sample Num: 31504, Cur Loss: 1.25138128, Cur Avg Loss: 1.55607978, Log Avg loss: 1.30024984, Global Avg Loss: 2.01383557, Time: 0.0402 Steps: 21130, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001979, Sample Num: 31664, Cur Loss: 1.29171109, Cur Avg Loss: 1.55454039, Log Avg loss: 1.25143395, Global Avg Loss: 2.01347492, Time: 0.0402 Steps: 21140, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001989, Sample Num: 31824, Cur Loss: 1.60303426, Cur Avg Loss: 1.55376632, Log Avg loss: 1.40057923, Global Avg Loss: 2.01318514, Time: 0.0402 Steps: 21150, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001999, Sample Num: 31984, Cur Loss: 2.26930618, Cur Avg Loss: 1.55519785, Log Avg loss: 1.83992726, Global Avg Loss: 2.01310326, Time: 0.0402 Steps: 21160, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002009, Sample Num: 32144, Cur Loss: 2.98445368, Cur Avg Loss: 1.55720376, Log Avg loss: 1.95818604, Global Avg Loss: 2.01307732, Time: 0.0402 Steps: 21170, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002019, Sample Num: 32304, Cur Loss: 2.20921469, Cur Avg Loss: 1.55785327, Log Avg loss: 1.68834027, Global Avg Loss: 2.01292400, Time: 0.0402 Steps: 21180, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002029, Sample Num: 32464, Cur Loss: 1.42052078, Cur Avg Loss: 1.55825926, Log Avg loss: 1.64022844, Global Avg Loss: 2.01274811, Time: 0.0402 Steps: 21190, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002039, Sample Num: 32624, Cur Loss: 2.43736649, Cur Avg Loss: 1.55704081, Log Avg loss: 1.30981756, Global Avg Loss: 2.01241654, Time: 0.0402 Steps: 21200, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002049, Sample Num: 32784, Cur Loss: 1.61323071, Cur Avg Loss: 1.55756591, Log Avg loss: 1.66463366, Global Avg Loss: 2.01225257, Time: 0.0403 Steps: 21210, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002059, Sample Num: 32944, Cur Loss: 0.47414714, Cur Avg Loss: 1.55713690, Log Avg loss: 1.46923326, Global Avg Loss: 2.01199667, Time: 0.0402 Steps: 21220, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002069, Sample Num: 33104, Cur Loss: 1.32159162, Cur Avg Loss: 1.55696612, Log Avg loss: 1.52180107, Global Avg Loss: 2.01176577, Time: 0.0402 Steps: 21230, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002079, Sample Num: 33264, Cur Loss: 1.62740946, Cur Avg Loss: 1.55848939, Log Avg loss: 1.87365457, Global Avg Loss: 2.01170075, Time: 0.0402 Steps: 21240, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002089, Sample Num: 33424, Cur Loss: 0.91310489, Cur Avg Loss: 1.55800611, Log Avg loss: 1.45753133, Global Avg Loss: 2.01143996, Time: 0.0402 Steps: 21250, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002099, Sample Num: 33584, Cur Loss: 1.89933848, Cur Avg Loss: 1.55686089, Log Avg loss: 1.31762632, Global Avg Loss: 2.01111362, Time: 0.0402 Steps: 21260, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002109, Sample Num: 33744, Cur Loss: 1.30673504, Cur Avg Loss: 1.55820190, Log Avg loss: 1.83967816, Global Avg Loss: 2.01103302, Time: 0.0402 Steps: 21270, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002119, Sample Num: 33904, Cur Loss: 2.39323759, Cur Avg Loss: 1.55650811, Log Avg loss: 1.19928850, Global Avg Loss: 2.01065156, Time: 0.0402 Steps: 21280, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002129, Sample Num: 34055, Cur Loss: 1.41136932, Cur Avg Loss: 1.55710308, Log Avg loss: 1.68317659, Global Avg Loss: 2.01049774, Time: 0.0198 Steps: 21290, Updated lr: 0.000081 ***** Running evaluation checkpoint-21290 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-21290 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.341603, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.416407, "eval_total_loss": 995.733915, "eval_mae": 1.021268, "eval_mse": 1.416638, "eval_r2": 0.099492, "eval_sp_statistic": 0.298681, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.355902, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.371199, "test_total_loss": 688.341749, "test_mae": 0.865972, "test_mse": 1.371653, "test_r2": 0.114723, "test_sp_statistic": 0.304683, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.361987, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.0104977421934977, "train_cur_epoch_loss": 3315.0724499970675, "train_cur_epoch_avg_loss": 1.5571030765603886, "train_cur_epoch_time": 95.34160327911377, "train_cur_epoch_avg_time": 0.04478234066656354, "epoch": 10, "step": 21290} ################################################## Training, Epoch: 0011, Batch: 000010, Sample Num: 160, Cur Loss: 1.51038456, Cur Avg Loss: 1.32286580, Log Avg loss: 1.32286580, Global Avg Loss: 2.01017491, Time: 0.0404 Steps: 21300, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000020, Sample Num: 320, Cur Loss: 1.02935696, Cur Avg Loss: 1.34205539, Log Avg loss: 1.36124498, Global Avg Loss: 2.00987039, Time: 0.0403 Steps: 21310, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000030, Sample Num: 480, Cur Loss: 1.66248429, Cur Avg Loss: 1.59083534, Log Avg loss: 2.08839525, Global Avg Loss: 2.00990722, Time: 0.0404 Steps: 21320, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000040, Sample Num: 640, Cur Loss: 1.76886201, Cur Avg Loss: 1.50775192, Log Avg loss: 1.25850163, Global Avg Loss: 2.00955495, Time: 0.0402 Steps: 21330, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000050, Sample Num: 800, Cur Loss: 1.10655844, Cur Avg Loss: 1.50891599, Log Avg loss: 1.51357228, Global Avg Loss: 2.00932253, Time: 0.0402 Steps: 21340, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000060, Sample Num: 960, Cur Loss: 1.78994060, Cur Avg Loss: 1.53566389, Log Avg loss: 1.66940342, Global Avg Loss: 2.00916331, Time: 0.0402 Steps: 21350, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000070, Sample Num: 1120, Cur Loss: 1.85326111, Cur Avg Loss: 1.57035332, Log Avg loss: 1.77848985, Global Avg Loss: 2.00905532, Time: 0.0402 Steps: 21360, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000080, Sample Num: 1280, Cur Loss: 2.16204643, Cur Avg Loss: 1.56386993, Log Avg loss: 1.51848621, Global Avg Loss: 2.00882576, Time: 0.0402 Steps: 21370, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000090, Sample Num: 1440, Cur Loss: 0.83074784, Cur Avg Loss: 1.56073898, Log Avg loss: 1.53569138, Global Avg Loss: 2.00860446, Time: 0.0402 Steps: 21380, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000100, Sample Num: 1600, Cur Loss: 1.33925295, Cur Avg Loss: 1.56840049, Log Avg loss: 1.63735406, Global Avg Loss: 2.00843090, Time: 0.0402 Steps: 21390, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000110, Sample Num: 1760, Cur Loss: 2.67882323, Cur Avg Loss: 1.54838691, Log Avg loss: 1.34825114, Global Avg Loss: 2.00812241, Time: 0.0402 Steps: 21400, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000120, Sample Num: 1920, Cur Loss: 1.01365495, Cur Avg Loss: 1.55761368, Log Avg loss: 1.65910820, Global Avg Loss: 2.00795939, Time: 0.0907 Steps: 21410, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000130, Sample Num: 2080, Cur Loss: 1.84720302, Cur Avg Loss: 1.56605592, Log Avg loss: 1.66736269, Global Avg Loss: 2.00780038, Time: 0.0850 Steps: 21420, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000140, Sample Num: 2240, Cur Loss: 1.04400516, Cur Avg Loss: 1.54569921, Log Avg loss: 1.28106209, Global Avg Loss: 2.00746126, Time: 0.0407 Steps: 21430, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000150, Sample Num: 2400, Cur Loss: 1.06415153, Cur Avg Loss: 1.53280898, Log Avg loss: 1.35234574, Global Avg Loss: 2.00715570, Time: 0.0499 Steps: 21440, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000160, Sample Num: 2560, Cur Loss: 2.25698090, Cur Avg Loss: 1.52045571, Log Avg loss: 1.33515657, Global Avg Loss: 2.00684242, Time: 0.0506 Steps: 21450, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000170, Sample Num: 2720, Cur Loss: 0.77036870, Cur Avg Loss: 1.50428346, Log Avg loss: 1.24552749, Global Avg Loss: 2.00648766, Time: 0.0915 Steps: 21460, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000180, Sample Num: 2880, Cur Loss: 1.41273963, Cur Avg Loss: 1.52726789, Log Avg loss: 1.91800326, Global Avg Loss: 2.00644644, Time: 0.0404 Steps: 21470, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000190, Sample Num: 3040, Cur Loss: 2.62667298, Cur Avg Loss: 1.54597966, Log Avg loss: 1.88279157, Global Avg Loss: 2.00638888, Time: 0.0893 Steps: 21480, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000200, Sample Num: 3200, Cur Loss: 1.00408101, Cur Avg Loss: 1.57122130, Log Avg loss: 2.05081241, Global Avg Loss: 2.00640955, Time: 0.0646 Steps: 21490, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000210, Sample Num: 3360, Cur Loss: 1.34271741, Cur Avg Loss: 1.56863754, Log Avg loss: 1.51696222, Global Avg Loss: 2.00618190, Time: 0.0402 Steps: 21500, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000220, Sample Num: 3520, Cur Loss: 2.09326482, Cur Avg Loss: 1.56017287, Log Avg loss: 1.38241490, Global Avg Loss: 2.00589191, Time: 0.0402 Steps: 21510, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000230, Sample Num: 3680, Cur Loss: 1.07103908, Cur Avg Loss: 1.55836285, Log Avg loss: 1.51854247, Global Avg Loss: 2.00566545, Time: 0.0402 Steps: 21520, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000240, Sample Num: 3840, Cur Loss: 1.33766055, Cur Avg Loss: 1.55290159, Log Avg loss: 1.42729263, Global Avg Loss: 2.00539681, Time: 0.0402 Steps: 21530, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000250, Sample Num: 4000, Cur Loss: 1.56961429, Cur Avg Loss: 1.54413217, Log Avg loss: 1.33366599, Global Avg Loss: 2.00508496, Time: 0.0402 Steps: 21540, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000260, Sample Num: 4160, Cur Loss: 1.49150109, Cur Avg Loss: 1.54410349, Log Avg loss: 1.54338641, Global Avg Loss: 2.00487071, Time: 0.0403 Steps: 21550, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000270, Sample Num: 4320, Cur Loss: 0.76396698, Cur Avg Loss: 1.53218084, Log Avg loss: 1.22219198, Global Avg Loss: 2.00450769, Time: 0.0402 Steps: 21560, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000280, Sample Num: 4480, Cur Loss: 1.82452178, Cur Avg Loss: 1.53337444, Log Avg loss: 1.56560165, Global Avg Loss: 2.00430421, Time: 0.0404 Steps: 21570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000290, Sample Num: 4640, Cur Loss: 1.90313387, Cur Avg Loss: 1.53712639, Log Avg loss: 1.64218104, Global Avg Loss: 2.00413640, Time: 0.0403 Steps: 21580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000300, Sample Num: 4800, Cur Loss: 0.91132283, Cur Avg Loss: 1.54014325, Log Avg loss: 1.62763227, Global Avg Loss: 2.00396202, Time: 0.0402 Steps: 21590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000310, Sample Num: 4960, Cur Loss: 0.80654645, Cur Avg Loss: 1.55236752, Log Avg loss: 1.91909556, Global Avg Loss: 2.00392273, Time: 0.0402 Steps: 21600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000320, Sample Num: 5120, Cur Loss: 1.27346587, Cur Avg Loss: 1.56135259, Log Avg loss: 1.83988985, Global Avg Loss: 2.00384682, Time: 0.0403 Steps: 21610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000330, Sample Num: 5280, Cur Loss: 1.91178715, Cur Avg Loss: 1.56225283, Log Avg loss: 1.59106035, Global Avg Loss: 2.00365589, Time: 0.0403 Steps: 21620, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000340, Sample Num: 5440, Cur Loss: 1.55901825, Cur Avg Loss: 1.56683210, Log Avg loss: 1.71794818, Global Avg Loss: 2.00352380, Time: 0.0402 Steps: 21630, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000350, Sample Num: 5600, Cur Loss: 1.12729287, Cur Avg Loss: 1.56190554, Log Avg loss: 1.39440237, Global Avg Loss: 2.00324232, Time: 0.0403 Steps: 21640, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000360, Sample Num: 5760, Cur Loss: 0.84794950, Cur Avg Loss: 1.55242212, Log Avg loss: 1.22050240, Global Avg Loss: 2.00288078, Time: 0.0403 Steps: 21650, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000370, Sample Num: 5920, Cur Loss: 1.54653931, Cur Avg Loss: 1.55791622, Log Avg loss: 1.75570384, Global Avg Loss: 2.00276666, Time: 0.0402 Steps: 21660, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000380, Sample Num: 6080, Cur Loss: 2.16116714, Cur Avg Loss: 1.55967928, Log Avg loss: 1.62491249, Global Avg Loss: 2.00259230, Time: 0.0402 Steps: 21670, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000390, Sample Num: 6240, Cur Loss: 3.04380536, Cur Avg Loss: 1.55991813, Log Avg loss: 1.56899441, Global Avg Loss: 2.00239230, Time: 0.0402 Steps: 21680, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000400, Sample Num: 6400, Cur Loss: 1.93379378, Cur Avg Loss: 1.55614226, Log Avg loss: 1.40888350, Global Avg Loss: 2.00211866, Time: 0.0402 Steps: 21690, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000410, Sample Num: 6560, Cur Loss: 1.09554601, Cur Avg Loss: 1.55605124, Log Avg loss: 1.55241016, Global Avg Loss: 2.00191143, Time: 0.0403 Steps: 21700, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000420, Sample Num: 6720, Cur Loss: 1.23134971, Cur Avg Loss: 1.54917628, Log Avg loss: 1.26730298, Global Avg Loss: 2.00157305, Time: 0.0403 Steps: 21710, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000430, Sample Num: 6880, Cur Loss: 0.79018629, Cur Avg Loss: 1.54845915, Log Avg loss: 1.51833987, Global Avg Loss: 2.00135057, Time: 0.0404 Steps: 21720, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000440, Sample Num: 7040, Cur Loss: 0.89327592, Cur Avg Loss: 1.54706829, Log Avg loss: 1.48726098, Global Avg Loss: 2.00111399, Time: 0.0404 Steps: 21730, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000450, Sample Num: 7200, Cur Loss: 1.86462486, Cur Avg Loss: 1.54653290, Log Avg loss: 1.52297614, Global Avg Loss: 2.00089405, Time: 0.0406 Steps: 21740, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000460, Sample Num: 7360, Cur Loss: 1.54650271, Cur Avg Loss: 1.54483653, Log Avg loss: 1.46849971, Global Avg Loss: 2.00064928, Time: 0.0403 Steps: 21750, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000470, Sample Num: 7520, Cur Loss: 0.79474413, Cur Avg Loss: 1.54355305, Log Avg loss: 1.48451287, Global Avg Loss: 2.00041208, Time: 0.0657 Steps: 21760, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000480, Sample Num: 7680, Cur Loss: 2.46393871, Cur Avg Loss: 1.54683962, Log Avg loss: 1.70130863, Global Avg Loss: 2.00027469, Time: 0.0412 Steps: 21770, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000490, Sample Num: 7840, Cur Loss: 1.82716846, Cur Avg Loss: 1.54611260, Log Avg loss: 1.51121529, Global Avg Loss: 2.00005014, Time: 0.0475 Steps: 21780, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000500, Sample Num: 8000, Cur Loss: 1.27772307, Cur Avg Loss: 1.54380505, Log Avg loss: 1.43073538, Global Avg Loss: 1.99978887, Time: 0.0404 Steps: 21790, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000510, Sample Num: 8160, Cur Loss: 1.19268489, Cur Avg Loss: 1.54237422, Log Avg loss: 1.47083271, Global Avg Loss: 1.99954623, Time: 0.0407 Steps: 21800, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000520, Sample Num: 8320, Cur Loss: 1.30521429, Cur Avg Loss: 1.54026982, Log Avg loss: 1.43294525, Global Avg Loss: 1.99928644, Time: 0.0481 Steps: 21810, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000530, Sample Num: 8480, Cur Loss: 0.95643222, Cur Avg Loss: 1.53816413, Log Avg loss: 1.42866832, Global Avg Loss: 1.99902493, Time: 0.0865 Steps: 21820, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000540, Sample Num: 8640, Cur Loss: 2.07581139, Cur Avg Loss: 1.53646641, Log Avg loss: 1.44648715, Global Avg Loss: 1.99877182, Time: 0.0655 Steps: 21830, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000550, Sample Num: 8800, Cur Loss: 2.59629583, Cur Avg Loss: 1.53227036, Log Avg loss: 1.30568395, Global Avg Loss: 1.99845447, Time: 0.0639 Steps: 21840, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000560, Sample Num: 8960, Cur Loss: 1.64171052, Cur Avg Loss: 1.52869669, Log Avg loss: 1.33214466, Global Avg Loss: 1.99814952, Time: 0.0402 Steps: 21850, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000570, Sample Num: 9120, Cur Loss: 1.38593662, Cur Avg Loss: 1.53085588, Log Avg loss: 1.65177063, Global Avg Loss: 1.99799107, Time: 0.0403 Steps: 21860, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000580, Sample Num: 9280, Cur Loss: 1.33115673, Cur Avg Loss: 1.52886860, Log Avg loss: 1.41559372, Global Avg Loss: 1.99772477, Time: 0.0402 Steps: 21870, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000590, Sample Num: 9440, Cur Loss: 1.19347000, Cur Avg Loss: 1.52808520, Log Avg loss: 1.48264768, Global Avg Loss: 1.99748936, Time: 0.0402 Steps: 21880, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000600, Sample Num: 9600, Cur Loss: 1.99348259, Cur Avg Loss: 1.52797042, Log Avg loss: 1.52119848, Global Avg Loss: 1.99727178, Time: 0.0403 Steps: 21890, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000610, Sample Num: 9760, Cur Loss: 1.45969200, Cur Avg Loss: 1.53147627, Log Avg loss: 1.74182726, Global Avg Loss: 1.99715513, Time: 0.0403 Steps: 21900, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000620, Sample Num: 9920, Cur Loss: 1.52982068, Cur Avg Loss: 1.53346008, Log Avg loss: 1.65447281, Global Avg Loss: 1.99699873, Time: 0.0402 Steps: 21910, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000630, Sample Num: 10080, Cur Loss: 4.29922247, Cur Avg Loss: 1.53674156, Log Avg loss: 1.74019318, Global Avg Loss: 1.99688157, Time: 0.0402 Steps: 21920, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000640, Sample Num: 10240, Cur Loss: 1.37793446, Cur Avg Loss: 1.52692190, Log Avg loss: 0.90828344, Global Avg Loss: 1.99638518, Time: 0.0402 Steps: 21930, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000650, Sample Num: 10400, Cur Loss: 1.56165183, Cur Avg Loss: 1.51964856, Log Avg loss: 1.05415451, Global Avg Loss: 1.99595572, Time: 0.0402 Steps: 21940, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000660, Sample Num: 10560, Cur Loss: 1.46120584, Cur Avg Loss: 1.51616526, Log Avg loss: 1.28975060, Global Avg Loss: 1.99563399, Time: 0.0402 Steps: 21950, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000670, Sample Num: 10720, Cur Loss: 1.75956321, Cur Avg Loss: 1.51602045, Log Avg loss: 1.50646345, Global Avg Loss: 1.99541123, Time: 0.0403 Steps: 21960, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000680, Sample Num: 10880, Cur Loss: 1.42931271, Cur Avg Loss: 1.51312794, Log Avg loss: 1.31932957, Global Avg Loss: 1.99510350, Time: 0.0402 Steps: 21970, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000690, Sample Num: 11040, Cur Loss: 0.69732779, Cur Avg Loss: 1.51109631, Log Avg loss: 1.37294564, Global Avg Loss: 1.99482045, Time: 0.0403 Steps: 21980, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000700, Sample Num: 11200, Cur Loss: 1.87994599, Cur Avg Loss: 1.51849575, Log Avg loss: 2.02905716, Global Avg Loss: 1.99483601, Time: 0.0403 Steps: 21990, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000710, Sample Num: 11360, Cur Loss: 0.67440367, Cur Avg Loss: 1.51471955, Log Avg loss: 1.25038517, Global Avg Loss: 1.99449763, Time: 0.0403 Steps: 22000, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000720, Sample Num: 11520, Cur Loss: 0.95200640, Cur Avg Loss: 1.51095767, Log Avg loss: 1.24386465, Global Avg Loss: 1.99415659, Time: 0.0402 Steps: 22010, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000730, Sample Num: 11680, Cur Loss: 2.40207863, Cur Avg Loss: 1.51271024, Log Avg loss: 1.63889467, Global Avg Loss: 1.99399525, Time: 0.0403 Steps: 22020, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000740, Sample Num: 11840, Cur Loss: 0.83801007, Cur Avg Loss: 1.51344242, Log Avg loss: 1.56689210, Global Avg Loss: 1.99380138, Time: 0.0401 Steps: 22030, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000750, Sample Num: 12000, Cur Loss: 3.71823287, Cur Avg Loss: 1.51323595, Log Avg loss: 1.49795692, Global Avg Loss: 1.99357640, Time: 0.0402 Steps: 22040, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000760, Sample Num: 12160, Cur Loss: 2.06480813, Cur Avg Loss: 1.51543818, Log Avg loss: 1.68060575, Global Avg Loss: 1.99343446, Time: 0.0402 Steps: 22050, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000770, Sample Num: 12320, Cur Loss: 0.99261761, Cur Avg Loss: 1.51631951, Log Avg loss: 1.58329991, Global Avg Loss: 1.99324855, Time: 0.0401 Steps: 22060, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000780, Sample Num: 12480, Cur Loss: 3.82179689, Cur Avg Loss: 1.51861003, Log Avg loss: 1.69498050, Global Avg Loss: 1.99311340, Time: 0.0402 Steps: 22070, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000790, Sample Num: 12640, Cur Loss: 0.73448205, Cur Avg Loss: 1.52191858, Log Avg loss: 1.77998565, Global Avg Loss: 1.99301688, Time: 0.0402 Steps: 22080, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000800, Sample Num: 12800, Cur Loss: 2.11498308, Cur Avg Loss: 1.52097428, Log Avg loss: 1.44637418, Global Avg Loss: 1.99276941, Time: 0.0402 Steps: 22090, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000810, Sample Num: 12960, Cur Loss: 1.47239780, Cur Avg Loss: 1.52082631, Log Avg loss: 1.50898906, Global Avg Loss: 1.99255051, Time: 0.0402 Steps: 22100, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000820, Sample Num: 13120, Cur Loss: 1.71653819, Cur Avg Loss: 1.52530220, Log Avg loss: 1.88784940, Global Avg Loss: 1.99250315, Time: 0.0402 Steps: 22110, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000830, Sample Num: 13280, Cur Loss: 1.12606752, Cur Avg Loss: 1.52599046, Log Avg loss: 1.58242769, Global Avg Loss: 1.99231777, Time: 0.0872 Steps: 22120, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000840, Sample Num: 13440, Cur Loss: 0.90001863, Cur Avg Loss: 1.52673564, Log Avg loss: 1.58858526, Global Avg Loss: 1.99213533, Time: 0.0791 Steps: 22130, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000850, Sample Num: 13600, Cur Loss: 2.08824635, Cur Avg Loss: 1.52969406, Log Avg loss: 1.77820134, Global Avg Loss: 1.99203870, Time: 0.0767 Steps: 22140, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000860, Sample Num: 13760, Cur Loss: 1.48709989, Cur Avg Loss: 1.52930060, Log Avg loss: 1.49585620, Global Avg Loss: 1.99181469, Time: 0.0481 Steps: 22150, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000870, Sample Num: 13920, Cur Loss: 1.76352143, Cur Avg Loss: 1.53404523, Log Avg loss: 1.94208394, Global Avg Loss: 1.99179225, Time: 0.0404 Steps: 22160, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000880, Sample Num: 14080, Cur Loss: 3.14531326, Cur Avg Loss: 1.53260954, Log Avg loss: 1.40770465, Global Avg Loss: 1.99152879, Time: 0.0788 Steps: 22170, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000890, Sample Num: 14240, Cur Loss: 2.50973701, Cur Avg Loss: 1.53715135, Log Avg loss: 1.93683041, Global Avg Loss: 1.99150413, Time: 0.0406 Steps: 22180, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000900, Sample Num: 14400, Cur Loss: 1.17542768, Cur Avg Loss: 1.54003190, Log Avg loss: 1.79640063, Global Avg Loss: 1.99141621, Time: 0.0412 Steps: 22190, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000910, Sample Num: 14560, Cur Loss: 2.21848345, Cur Avg Loss: 1.53892293, Log Avg loss: 1.43911614, Global Avg Loss: 1.99116742, Time: 0.0695 Steps: 22200, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000920, Sample Num: 14720, Cur Loss: 2.06287146, Cur Avg Loss: 1.53716899, Log Avg loss: 1.37756005, Global Avg Loss: 1.99089115, Time: 0.0402 Steps: 22210, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000930, Sample Num: 14880, Cur Loss: 2.12895536, Cur Avg Loss: 1.53924680, Log Avg loss: 1.73040515, Global Avg Loss: 1.99077392, Time: 0.0403 Steps: 22220, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000940, Sample Num: 15040, Cur Loss: 1.52578342, Cur Avg Loss: 1.53986090, Log Avg loss: 1.59697286, Global Avg Loss: 1.99059677, Time: 0.0404 Steps: 22230, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000950, Sample Num: 15200, Cur Loss: 1.40838802, Cur Avg Loss: 1.53901793, Log Avg loss: 1.45977857, Global Avg Loss: 1.99035809, Time: 0.0402 Steps: 22240, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000960, Sample Num: 15360, Cur Loss: 2.31434345, Cur Avg Loss: 1.53935379, Log Avg loss: 1.57126016, Global Avg Loss: 1.99016973, Time: 0.0402 Steps: 22250, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000970, Sample Num: 15520, Cur Loss: 1.47755647, Cur Avg Loss: 1.53528608, Log Avg loss: 1.14478595, Global Avg Loss: 1.98978996, Time: 0.0402 Steps: 22260, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000980, Sample Num: 15680, Cur Loss: 1.42965066, Cur Avg Loss: 1.53884300, Log Avg loss: 1.88386452, Global Avg Loss: 1.98974239, Time: 0.0402 Steps: 22270, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000990, Sample Num: 15840, Cur Loss: 1.94310582, Cur Avg Loss: 1.54062705, Log Avg loss: 1.71546339, Global Avg Loss: 1.98961929, Time: 0.0403 Steps: 22280, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001000, Sample Num: 16000, Cur Loss: 1.91512322, Cur Avg Loss: 1.53925987, Log Avg loss: 1.40390955, Global Avg Loss: 1.98935652, Time: 0.0402 Steps: 22290, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001010, Sample Num: 16160, Cur Loss: 1.85169363, Cur Avg Loss: 1.53881609, Log Avg loss: 1.49443806, Global Avg Loss: 1.98913458, Time: 0.0402 Steps: 22300, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001020, Sample Num: 16320, Cur Loss: 1.51209497, Cur Avg Loss: 1.53914871, Log Avg loss: 1.57274363, Global Avg Loss: 1.98894794, Time: 0.0402 Steps: 22310, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001030, Sample Num: 16480, Cur Loss: 1.93548667, Cur Avg Loss: 1.53808002, Log Avg loss: 1.42907371, Global Avg Loss: 1.98869710, Time: 0.0402 Steps: 22320, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001040, Sample Num: 16640, Cur Loss: 1.20940375, Cur Avg Loss: 1.53800243, Log Avg loss: 1.53001010, Global Avg Loss: 1.98849169, Time: 0.0403 Steps: 22330, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001050, Sample Num: 16800, Cur Loss: 1.45959830, Cur Avg Loss: 1.53717054, Log Avg loss: 1.45065389, Global Avg Loss: 1.98825094, Time: 0.0402 Steps: 22340, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001060, Sample Num: 16960, Cur Loss: 1.12816942, Cur Avg Loss: 1.53717718, Log Avg loss: 1.53787448, Global Avg Loss: 1.98804943, Time: 0.0402 Steps: 22350, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001070, Sample Num: 17120, Cur Loss: 1.52894354, Cur Avg Loss: 1.53981898, Log Avg loss: 1.81984980, Global Avg Loss: 1.98797421, Time: 0.0402 Steps: 22360, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001080, Sample Num: 17280, Cur Loss: 2.62920809, Cur Avg Loss: 1.53813930, Log Avg loss: 1.35841361, Global Avg Loss: 1.98769277, Time: 0.0402 Steps: 22370, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001090, Sample Num: 17440, Cur Loss: 1.63038135, Cur Avg Loss: 1.53786241, Log Avg loss: 1.50795863, Global Avg Loss: 1.98747842, Time: 0.0402 Steps: 22380, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001100, Sample Num: 17600, Cur Loss: 2.04758024, Cur Avg Loss: 1.53854601, Log Avg loss: 1.61305757, Global Avg Loss: 1.98731119, Time: 0.0402 Steps: 22390, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001110, Sample Num: 17760, Cur Loss: 0.85652381, Cur Avg Loss: 1.53430679, Log Avg loss: 1.06799312, Global Avg Loss: 1.98690078, Time: 0.0402 Steps: 22400, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001120, Sample Num: 17920, Cur Loss: 0.74057651, Cur Avg Loss: 1.53470690, Log Avg loss: 1.57911924, Global Avg Loss: 1.98671882, Time: 0.0402 Steps: 22410, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001130, Sample Num: 18080, Cur Loss: 0.99452955, Cur Avg Loss: 1.53493434, Log Avg loss: 1.56040730, Global Avg Loss: 1.98652867, Time: 0.0402 Steps: 22420, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001140, Sample Num: 18240, Cur Loss: 0.94296551, Cur Avg Loss: 1.53285551, Log Avg loss: 1.29794842, Global Avg Loss: 1.98622168, Time: 0.0402 Steps: 22430, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001150, Sample Num: 18400, Cur Loss: 3.02644634, Cur Avg Loss: 1.53331946, Log Avg loss: 1.58620961, Global Avg Loss: 1.98604342, Time: 0.0402 Steps: 22440, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001160, Sample Num: 18560, Cur Loss: 1.71377242, Cur Avg Loss: 1.53262834, Log Avg loss: 1.45314901, Global Avg Loss: 1.98580605, Time: 0.0402 Steps: 22450, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001170, Sample Num: 18720, Cur Loss: 0.67845351, Cur Avg Loss: 1.53102879, Log Avg loss: 1.34548067, Global Avg Loss: 1.98552095, Time: 0.0402 Steps: 22460, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001180, Sample Num: 18880, Cur Loss: 0.96459228, Cur Avg Loss: 1.53253882, Log Avg loss: 1.70921252, Global Avg Loss: 1.98539799, Time: 0.0402 Steps: 22470, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001190, Sample Num: 19040, Cur Loss: 1.23970413, Cur Avg Loss: 1.53672476, Log Avg loss: 2.03066546, Global Avg Loss: 1.98541812, Time: 0.0411 Steps: 22480, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001200, Sample Num: 19200, Cur Loss: 1.67714620, Cur Avg Loss: 1.53932456, Log Avg loss: 1.84870073, Global Avg Loss: 1.98535733, Time: 0.0622 Steps: 22490, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001210, Sample Num: 19360, Cur Loss: 1.45275402, Cur Avg Loss: 1.54001033, Log Avg loss: 1.62230265, Global Avg Loss: 1.98519597, Time: 0.0817 Steps: 22500, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001220, Sample Num: 19520, Cur Loss: 0.84590626, Cur Avg Loss: 1.53911691, Log Avg loss: 1.43101307, Global Avg Loss: 1.98494978, Time: 0.0411 Steps: 22510, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001230, Sample Num: 19680, Cur Loss: 1.88160419, Cur Avg Loss: 1.54034330, Log Avg loss: 1.68996309, Global Avg Loss: 1.98481879, Time: 0.0601 Steps: 22520, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001240, Sample Num: 19840, Cur Loss: 1.26438117, Cur Avg Loss: 1.53995187, Log Avg loss: 1.49180676, Global Avg Loss: 1.98459997, Time: 0.0708 Steps: 22530, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001250, Sample Num: 20000, Cur Loss: 0.92115891, Cur Avg Loss: 1.53901013, Log Avg loss: 1.42223350, Global Avg Loss: 1.98435047, Time: 0.0489 Steps: 22540, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001260, Sample Num: 20160, Cur Loss: 1.19454300, Cur Avg Loss: 1.53909625, Log Avg loss: 1.54986206, Global Avg Loss: 1.98415779, Time: 0.0882 Steps: 22550, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001270, Sample Num: 20320, Cur Loss: 2.16053987, Cur Avg Loss: 1.53989481, Log Avg loss: 1.64051247, Global Avg Loss: 1.98400547, Time: 0.0405 Steps: 22560, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001280, Sample Num: 20480, Cur Loss: 1.27766538, Cur Avg Loss: 1.53851435, Log Avg loss: 1.36319682, Global Avg Loss: 1.98373041, Time: 0.0410 Steps: 22570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001290, Sample Num: 20640, Cur Loss: 1.52650881, Cur Avg Loss: 1.53748149, Log Avg loss: 1.40527472, Global Avg Loss: 1.98347423, Time: 0.0403 Steps: 22580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001300, Sample Num: 20800, Cur Loss: 2.08810282, Cur Avg Loss: 1.53583173, Log Avg loss: 1.32301353, Global Avg Loss: 1.98318186, Time: 0.0403 Steps: 22590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001310, Sample Num: 20960, Cur Loss: 1.52727437, Cur Avg Loss: 1.53648351, Log Avg loss: 1.62121465, Global Avg Loss: 1.98302170, Time: 0.0402 Steps: 22600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001320, Sample Num: 21120, Cur Loss: 1.77401304, Cur Avg Loss: 1.53483652, Log Avg loss: 1.31908023, Global Avg Loss: 1.98272805, Time: 0.0403 Steps: 22610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001330, Sample Num: 21280, Cur Loss: 2.06007910, Cur Avg Loss: 1.53419334, Log Avg loss: 1.44929383, Global Avg Loss: 1.98249222, Time: 0.0403 Steps: 22620, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001340, Sample Num: 21440, Cur Loss: 0.99219722, Cur Avg Loss: 1.53514737, Log Avg loss: 1.66203365, Global Avg Loss: 1.98235061, Time: 0.0403 Steps: 22630, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001350, Sample Num: 21600, Cur Loss: 0.96594673, Cur Avg Loss: 1.53560252, Log Avg loss: 1.59659268, Global Avg Loss: 1.98218023, Time: 0.0402 Steps: 22640, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001360, Sample Num: 21760, Cur Loss: 1.21928811, Cur Avg Loss: 1.53402030, Log Avg loss: 1.32042030, Global Avg Loss: 1.98188806, Time: 0.0403 Steps: 22650, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001370, Sample Num: 21920, Cur Loss: 3.01344681, Cur Avg Loss: 1.53750120, Log Avg loss: 2.01090314, Global Avg Loss: 1.98190086, Time: 0.0403 Steps: 22660, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001380, Sample Num: 22080, Cur Loss: 1.40466559, Cur Avg Loss: 1.53776484, Log Avg loss: 1.57388361, Global Avg Loss: 1.98172088, Time: 0.0403 Steps: 22670, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001390, Sample Num: 22240, Cur Loss: 0.78318608, Cur Avg Loss: 1.53552914, Log Avg loss: 1.22700342, Global Avg Loss: 1.98138811, Time: 0.0402 Steps: 22680, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001400, Sample Num: 22400, Cur Loss: 2.54684281, Cur Avg Loss: 1.53456554, Log Avg loss: 1.40062487, Global Avg Loss: 1.98113216, Time: 0.0402 Steps: 22690, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001410, Sample Num: 22560, Cur Loss: 0.64403772, Cur Avg Loss: 1.53399915, Log Avg loss: 1.45470459, Global Avg Loss: 1.98090025, Time: 0.0402 Steps: 22700, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001420, Sample Num: 22720, Cur Loss: 2.16587067, Cur Avg Loss: 1.53297407, Log Avg loss: 1.38843699, Global Avg Loss: 1.98063937, Time: 0.0403 Steps: 22710, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001430, Sample Num: 22880, Cur Loss: 2.76769543, Cur Avg Loss: 1.53415794, Log Avg loss: 1.70226818, Global Avg Loss: 1.98051685, Time: 0.0402 Steps: 22720, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001440, Sample Num: 23040, Cur Loss: 1.50717807, Cur Avg Loss: 1.53429731, Log Avg loss: 1.55422760, Global Avg Loss: 1.98032930, Time: 0.0403 Steps: 22730, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001450, Sample Num: 23200, Cur Loss: 0.95692480, Cur Avg Loss: 1.53514508, Log Avg loss: 1.65722333, Global Avg Loss: 1.98018722, Time: 0.0402 Steps: 22740, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001460, Sample Num: 23360, Cur Loss: 1.65276229, Cur Avg Loss: 1.53512250, Log Avg loss: 1.53184785, Global Avg Loss: 1.97999014, Time: 0.0403 Steps: 22750, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001470, Sample Num: 23520, Cur Loss: 1.28459954, Cur Avg Loss: 1.53753333, Log Avg loss: 1.88951542, Global Avg Loss: 1.97995039, Time: 0.0402 Steps: 22760, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001480, Sample Num: 23680, Cur Loss: 1.13151538, Cur Avg Loss: 1.53777336, Log Avg loss: 1.57305708, Global Avg Loss: 1.97977170, Time: 0.0402 Steps: 22770, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001490, Sample Num: 23840, Cur Loss: 1.21172404, Cur Avg Loss: 1.53643764, Log Avg loss: 1.33875173, Global Avg Loss: 1.97949030, Time: 0.0402 Steps: 22780, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001500, Sample Num: 24000, Cur Loss: 0.59898931, Cur Avg Loss: 1.53654853, Log Avg loss: 1.55307090, Global Avg Loss: 1.97930319, Time: 0.0402 Steps: 22790, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001510, Sample Num: 24160, Cur Loss: 2.75404835, Cur Avg Loss: 1.53909782, Log Avg loss: 1.92149158, Global Avg Loss: 1.97927784, Time: 0.0402 Steps: 22800, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001520, Sample Num: 24320, Cur Loss: 2.39264369, Cur Avg Loss: 1.54039051, Log Avg loss: 1.73558611, Global Avg Loss: 1.97917100, Time: 0.0403 Steps: 22810, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001530, Sample Num: 24480, Cur Loss: 2.40184116, Cur Avg Loss: 1.54266508, Log Avg loss: 1.88840001, Global Avg Loss: 1.97913122, Time: 0.0402 Steps: 22820, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001540, Sample Num: 24640, Cur Loss: 0.70042145, Cur Avg Loss: 1.54320661, Log Avg loss: 1.62606009, Global Avg Loss: 1.97897657, Time: 0.0402 Steps: 22830, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001550, Sample Num: 24800, Cur Loss: 1.60638952, Cur Avg Loss: 1.54213267, Log Avg loss: 1.37674688, Global Avg Loss: 1.97871290, Time: 0.0637 Steps: 22840, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001560, Sample Num: 24960, Cur Loss: 0.82567817, Cur Avg Loss: 1.54119697, Log Avg loss: 1.39616275, Global Avg Loss: 1.97845795, Time: 0.0969 Steps: 22850, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001570, Sample Num: 25120, Cur Loss: 1.04626822, Cur Avg Loss: 1.53941183, Log Avg loss: 1.26093109, Global Avg Loss: 1.97814407, Time: 0.0442 Steps: 22860, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001580, Sample Num: 25280, Cur Loss: 2.92638040, Cur Avg Loss: 1.54265525, Log Avg loss: 2.05187192, Global Avg Loss: 1.97817631, Time: 0.0653 Steps: 22870, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001590, Sample Num: 25440, Cur Loss: 2.08114195, Cur Avg Loss: 1.54329017, Log Avg loss: 1.64360663, Global Avg Loss: 1.97803008, Time: 0.0403 Steps: 22880, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001600, Sample Num: 25600, Cur Loss: 0.99075913, Cur Avg Loss: 1.54242374, Log Avg loss: 1.40466209, Global Avg Loss: 1.97777959, Time: 0.0426 Steps: 22890, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001610, Sample Num: 25760, Cur Loss: 1.32755995, Cur Avg Loss: 1.54229863, Log Avg loss: 1.52228163, Global Avg Loss: 1.97758069, Time: 0.0778 Steps: 22900, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001620, Sample Num: 25920, Cur Loss: 1.15411592, Cur Avg Loss: 1.54207708, Log Avg loss: 1.50640705, Global Avg Loss: 1.97737502, Time: 0.0679 Steps: 22910, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001630, Sample Num: 26080, Cur Loss: 2.40997243, Cur Avg Loss: 1.54450397, Log Avg loss: 1.93766049, Global Avg Loss: 1.97735770, Time: 0.0585 Steps: 22920, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001640, Sample Num: 26240, Cur Loss: 1.47372425, Cur Avg Loss: 1.54472344, Log Avg loss: 1.58049650, Global Avg Loss: 1.97718462, Time: 0.0402 Steps: 22930, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001650, Sample Num: 26400, Cur Loss: 0.82345903, Cur Avg Loss: 1.54298335, Log Avg loss: 1.25760817, Global Avg Loss: 1.97687094, Time: 0.0403 Steps: 22940, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001660, Sample Num: 26560, Cur Loss: 1.17586327, Cur Avg Loss: 1.54345468, Log Avg loss: 1.62122474, Global Avg Loss: 1.97671598, Time: 0.0403 Steps: 22950, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001670, Sample Num: 26720, Cur Loss: 1.00548756, Cur Avg Loss: 1.54384833, Log Avg loss: 1.60919420, Global Avg Loss: 1.97655591, Time: 0.0402 Steps: 22960, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001680, Sample Num: 26880, Cur Loss: 1.84490395, Cur Avg Loss: 1.54249991, Log Avg loss: 1.31731299, Global Avg Loss: 1.97626891, Time: 0.0402 Steps: 22970, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001690, Sample Num: 27040, Cur Loss: 0.85780382, Cur Avg Loss: 1.54198521, Log Avg loss: 1.45551613, Global Avg Loss: 1.97604229, Time: 0.0402 Steps: 22980, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001700, Sample Num: 27200, Cur Loss: 0.99979228, Cur Avg Loss: 1.54205898, Log Avg loss: 1.55452618, Global Avg Loss: 1.97585895, Time: 0.0403 Steps: 22990, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001710, Sample Num: 27360, Cur Loss: 0.95467877, Cur Avg Loss: 1.54290886, Log Avg loss: 1.68738806, Global Avg Loss: 1.97573353, Time: 0.0403 Steps: 23000, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001720, Sample Num: 27520, Cur Loss: 1.68194008, Cur Avg Loss: 1.54126185, Log Avg loss: 1.25962412, Global Avg Loss: 1.97542231, Time: 0.0402 Steps: 23010, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001730, Sample Num: 27680, Cur Loss: 3.61700153, Cur Avg Loss: 1.54396809, Log Avg loss: 2.00944068, Global Avg Loss: 1.97543709, Time: 0.0402 Steps: 23020, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001740, Sample Num: 27840, Cur Loss: 1.66930091, Cur Avg Loss: 1.54414703, Log Avg loss: 1.57510304, Global Avg Loss: 1.97526325, Time: 0.0403 Steps: 23030, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001750, Sample Num: 28000, Cur Loss: 1.48147321, Cur Avg Loss: 1.54585382, Log Avg loss: 1.84283603, Global Avg Loss: 1.97520578, Time: 0.0402 Steps: 23040, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001760, Sample Num: 28160, Cur Loss: 2.63489151, Cur Avg Loss: 1.54742929, Log Avg loss: 1.82313637, Global Avg Loss: 1.97513980, Time: 0.0402 Steps: 23050, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001770, Sample Num: 28320, Cur Loss: 1.83920217, Cur Avg Loss: 1.54697118, Log Avg loss: 1.46634445, Global Avg Loss: 1.97491916, Time: 0.0403 Steps: 23060, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001780, Sample Num: 28480, Cur Loss: 1.49867845, Cur Avg Loss: 1.54752672, Log Avg loss: 1.64585763, Global Avg Loss: 1.97477653, Time: 0.0403 Steps: 23070, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001790, Sample Num: 28640, Cur Loss: 2.15961051, Cur Avg Loss: 1.54914582, Log Avg loss: 1.83734525, Global Avg Loss: 1.97471698, Time: 0.0402 Steps: 23080, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001800, Sample Num: 28800, Cur Loss: 3.10976171, Cur Avg Loss: 1.55052057, Log Avg loss: 1.79660121, Global Avg Loss: 1.97463984, Time: 0.0403 Steps: 23090, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001810, Sample Num: 28960, Cur Loss: 1.10646629, Cur Avg Loss: 1.54986607, Log Avg loss: 1.43205581, Global Avg Loss: 1.97440496, Time: 0.0402 Steps: 23100, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001820, Sample Num: 29120, Cur Loss: 2.52387238, Cur Avg Loss: 1.54984975, Log Avg loss: 1.54689499, Global Avg Loss: 1.97421997, Time: 0.0402 Steps: 23110, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001830, Sample Num: 29280, Cur Loss: 1.95866048, Cur Avg Loss: 1.54974924, Log Avg loss: 1.53145576, Global Avg Loss: 1.97402846, Time: 0.0402 Steps: 23120, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001840, Sample Num: 29440, Cur Loss: 2.47055006, Cur Avg Loss: 1.55095583, Log Avg loss: 1.77176319, Global Avg Loss: 1.97394101, Time: 0.0402 Steps: 23130, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001850, Sample Num: 29600, Cur Loss: 1.58037066, Cur Avg Loss: 1.55224886, Log Avg loss: 1.79016559, Global Avg Loss: 1.97386160, Time: 0.0402 Steps: 23140, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001860, Sample Num: 29760, Cur Loss: 1.56534243, Cur Avg Loss: 1.55255240, Log Avg loss: 1.60870848, Global Avg Loss: 1.97370386, Time: 0.0403 Steps: 23150, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001870, Sample Num: 29920, Cur Loss: 2.33332181, Cur Avg Loss: 1.55391719, Log Avg loss: 1.80776651, Global Avg Loss: 1.97363221, Time: 0.0402 Steps: 23160, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001880, Sample Num: 30080, Cur Loss: 1.14771867, Cur Avg Loss: 1.55341931, Log Avg loss: 1.46031641, Global Avg Loss: 1.97341067, Time: 0.0402 Steps: 23170, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001890, Sample Num: 30240, Cur Loss: 0.99156487, Cur Avg Loss: 1.55369209, Log Avg loss: 1.60497562, Global Avg Loss: 1.97325173, Time: 0.0636 Steps: 23180, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001900, Sample Num: 30400, Cur Loss: 0.64920610, Cur Avg Loss: 1.55302876, Log Avg loss: 1.42765942, Global Avg Loss: 1.97301645, Time: 0.0544 Steps: 23190, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001910, Sample Num: 30560, Cur Loss: 1.79066098, Cur Avg Loss: 1.55215223, Log Avg loss: 1.38561040, Global Avg Loss: 1.97276326, Time: 0.0406 Steps: 23200, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001920, Sample Num: 30720, Cur Loss: 1.72801900, Cur Avg Loss: 1.55275072, Log Avg loss: 1.66706300, Global Avg Loss: 1.97263155, Time: 0.0999 Steps: 23210, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001930, Sample Num: 30880, Cur Loss: 2.11657095, Cur Avg Loss: 1.55482087, Log Avg loss: 1.95228869, Global Avg Loss: 1.97262279, Time: 0.0658 Steps: 23220, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001940, Sample Num: 31040, Cur Loss: 0.62482524, Cur Avg Loss: 1.55561057, Log Avg loss: 1.70802356, Global Avg Loss: 1.97250889, Time: 0.0405 Steps: 23230, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001950, Sample Num: 31200, Cur Loss: 3.33565784, Cur Avg Loss: 1.55559711, Log Avg loss: 1.55298495, Global Avg Loss: 1.97232837, Time: 0.0423 Steps: 23240, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001960, Sample Num: 31360, Cur Loss: 1.56965542, Cur Avg Loss: 1.55369642, Log Avg loss: 1.18306295, Global Avg Loss: 1.97198890, Time: 0.0612 Steps: 23250, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001970, Sample Num: 31520, Cur Loss: 0.83396250, Cur Avg Loss: 1.55158508, Log Avg loss: 1.13776285, Global Avg Loss: 1.97163025, Time: 0.0405 Steps: 23260, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001980, Sample Num: 31680, Cur Loss: 2.22913599, Cur Avg Loss: 1.55177939, Log Avg loss: 1.59005662, Global Avg Loss: 1.97146627, Time: 0.0403 Steps: 23270, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001990, Sample Num: 31840, Cur Loss: 1.79260170, Cur Avg Loss: 1.55082194, Log Avg loss: 1.36124712, Global Avg Loss: 1.97120415, Time: 0.0402 Steps: 23280, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002000, Sample Num: 32000, Cur Loss: 1.46707153, Cur Avg Loss: 1.55161253, Log Avg loss: 1.70893995, Global Avg Loss: 1.97109154, Time: 0.0403 Steps: 23290, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002010, Sample Num: 32160, Cur Loss: 1.08523464, Cur Avg Loss: 1.55060022, Log Avg loss: 1.34813904, Global Avg Loss: 1.97082418, Time: 0.0402 Steps: 23300, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002020, Sample Num: 32320, Cur Loss: 1.73568153, Cur Avg Loss: 1.55192094, Log Avg loss: 1.81738486, Global Avg Loss: 1.97075835, Time: 0.0402 Steps: 23310, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002030, Sample Num: 32480, Cur Loss: 2.40128326, Cur Avg Loss: 1.55195551, Log Avg loss: 1.55893948, Global Avg Loss: 1.97058176, Time: 0.0404 Steps: 23320, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002040, Sample Num: 32640, Cur Loss: 2.29724455, Cur Avg Loss: 1.55232222, Log Avg loss: 1.62676452, Global Avg Loss: 1.97043439, Time: 0.0403 Steps: 23330, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002050, Sample Num: 32800, Cur Loss: 1.35708547, Cur Avg Loss: 1.55281265, Log Avg loss: 1.65286076, Global Avg Loss: 1.97029832, Time: 0.0403 Steps: 23340, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002060, Sample Num: 32960, Cur Loss: 0.46318573, Cur Avg Loss: 1.55264326, Log Avg loss: 1.51791749, Global Avg Loss: 1.97010458, Time: 0.0403 Steps: 23350, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002070, Sample Num: 33120, Cur Loss: 0.88995576, Cur Avg Loss: 1.55271184, Log Avg loss: 1.56684033, Global Avg Loss: 1.96993195, Time: 0.0403 Steps: 23360, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002080, Sample Num: 33280, Cur Loss: 1.25432265, Cur Avg Loss: 1.55269238, Log Avg loss: 1.54866336, Global Avg Loss: 1.96975169, Time: 0.0402 Steps: 23370, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002090, Sample Num: 33440, Cur Loss: 2.84148264, Cur Avg Loss: 1.55180601, Log Avg loss: 1.36744091, Global Avg Loss: 1.96949408, Time: 0.0402 Steps: 23380, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002100, Sample Num: 33600, Cur Loss: 0.92546010, Cur Avg Loss: 1.55231157, Log Avg loss: 1.65797309, Global Avg Loss: 1.96936089, Time: 0.0403 Steps: 23390, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002110, Sample Num: 33760, Cur Loss: 1.15140188, Cur Avg Loss: 1.55162635, Log Avg loss: 1.40773054, Global Avg Loss: 1.96912088, Time: 0.0402 Steps: 23400, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002120, Sample Num: 33920, Cur Loss: 1.42245674, Cur Avg Loss: 1.55143853, Log Avg loss: 1.51180819, Global Avg Loss: 1.96892553, Time: 0.0403 Steps: 23410, Updated lr: 0.000079 ***** Running evaluation checkpoint-23419 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-23419 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.730594, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.405937, "eval_total_loss": 988.373377, "eval_mae": 0.928992, "eval_mse": 1.406324, "eval_r2": 0.106048, "eval_sp_statistic": 0.292347, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.350654, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.451619, "test_total_loss": 728.712655, "test_mae": 0.779312, "test_mse": 1.452135, "test_r2": 0.062779, "test_sp_statistic": 0.298765, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.359585, "test_ps_pvalue": 0.0, "lr": 7.873968705547652e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.9687662087175684, "train_cur_epoch_loss": 3303.0389106571674, "train_cur_epoch_avg_loss": 1.551450873958275, "train_cur_epoch_time": 95.73059439659119, "train_cur_epoch_avg_time": 0.04496505138402592, "epoch": 11, "step": 23419} ################################################## Training, Epoch: 0012, Batch: 000001, Sample Num: 16, Cur Loss: 1.13403440, Cur Avg Loss: 1.13403440, Log Avg loss: 1.51232659, Global Avg Loss: 1.96873057, Time: 0.0407 Steps: 23420, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000011, Sample Num: 176, Cur Loss: 1.46324301, Cur Avg Loss: 1.87808600, Log Avg loss: 1.95249116, Global Avg Loss: 1.96872364, Time: 0.0404 Steps: 23430, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000021, Sample Num: 336, Cur Loss: 1.52785492, Cur Avg Loss: 1.68969412, Log Avg loss: 1.48246306, Global Avg Loss: 1.96851619, Time: 0.0403 Steps: 23440, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000031, Sample Num: 496, Cur Loss: 2.18175316, Cur Avg Loss: 1.55577202, Log Avg loss: 1.27453562, Global Avg Loss: 1.96822025, Time: 0.0403 Steps: 23450, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000041, Sample Num: 656, Cur Loss: 1.15623939, Cur Avg Loss: 1.51694376, Log Avg loss: 1.39657615, Global Avg Loss: 1.96797658, Time: 0.0404 Steps: 23460, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000051, Sample Num: 816, Cur Loss: 1.90104342, Cur Avg Loss: 1.52030791, Log Avg loss: 1.53410093, Global Avg Loss: 1.96779171, Time: 0.0404 Steps: 23470, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000061, Sample Num: 976, Cur Loss: 1.59443510, Cur Avg Loss: 1.56851134, Log Avg loss: 1.81434884, Global Avg Loss: 1.96772636, Time: 0.0403 Steps: 23480, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000071, Sample Num: 1136, Cur Loss: 0.48128569, Cur Avg Loss: 1.52725813, Log Avg loss: 1.27561353, Global Avg Loss: 1.96743172, Time: 0.0403 Steps: 23490, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000081, Sample Num: 1296, Cur Loss: 2.13587618, Cur Avg Loss: 1.57213305, Log Avg loss: 1.89074495, Global Avg Loss: 1.96739909, Time: 0.0402 Steps: 23500, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000091, Sample Num: 1456, Cur Loss: 1.04306257, Cur Avg Loss: 1.57150063, Log Avg loss: 1.56637803, Global Avg Loss: 1.96722852, Time: 0.0402 Steps: 23510, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000101, Sample Num: 1616, Cur Loss: 0.31514773, Cur Avg Loss: 1.54645074, Log Avg loss: 1.31849678, Global Avg Loss: 1.96695269, Time: 0.0403 Steps: 23520, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000111, Sample Num: 1776, Cur Loss: 2.25747204, Cur Avg Loss: 1.52384747, Log Avg loss: 1.29555438, Global Avg Loss: 1.96666736, Time: 0.0403 Steps: 23530, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000121, Sample Num: 1936, Cur Loss: 1.47525406, Cur Avg Loss: 1.51580363, Log Avg loss: 1.42651699, Global Avg Loss: 1.96643790, Time: 0.0402 Steps: 23540, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000131, Sample Num: 2096, Cur Loss: 1.06214130, Cur Avg Loss: 1.50561138, Log Avg loss: 1.38228519, Global Avg Loss: 1.96618985, Time: 0.0403 Steps: 23550, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000141, Sample Num: 2256, Cur Loss: 1.14493454, Cur Avg Loss: 1.50002188, Log Avg loss: 1.42679946, Global Avg Loss: 1.96596091, Time: 0.0402 Steps: 23560, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000151, Sample Num: 2416, Cur Loss: 1.57551968, Cur Avg Loss: 1.51195919, Log Avg loss: 1.68027531, Global Avg Loss: 1.96583970, Time: 0.0557 Steps: 23570, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000161, Sample Num: 2576, Cur Loss: 2.40740323, Cur Avg Loss: 1.51563518, Log Avg loss: 1.57114262, Global Avg Loss: 1.96567231, Time: 0.0459 Steps: 23580, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000171, Sample Num: 2736, Cur Loss: 1.52385545, Cur Avg Loss: 1.52007353, Log Avg loss: 1.59153091, Global Avg Loss: 1.96551371, Time: 0.0582 Steps: 23590, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000181, Sample Num: 2896, Cur Loss: 2.06616402, Cur Avg Loss: 1.52170422, Log Avg loss: 1.54958899, Global Avg Loss: 1.96533747, Time: 0.0712 Steps: 23600, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000191, Sample Num: 3056, Cur Loss: 2.05929065, Cur Avg Loss: 1.50151206, Log Avg loss: 1.13603402, Global Avg Loss: 1.96498622, Time: 0.0406 Steps: 23610, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000201, Sample Num: 3216, Cur Loss: 1.96073520, Cur Avg Loss: 1.50286276, Log Avg loss: 1.52866109, Global Avg Loss: 1.96480149, Time: 0.0694 Steps: 23620, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000211, Sample Num: 3376, Cur Loss: 0.68348432, Cur Avg Loss: 1.48962683, Log Avg loss: 1.22358462, Global Avg Loss: 1.96448782, Time: 0.0519 Steps: 23630, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000221, Sample Num: 3536, Cur Loss: 0.83411705, Cur Avg Loss: 1.50793452, Log Avg loss: 1.89422681, Global Avg Loss: 1.96445810, Time: 0.0464 Steps: 23640, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000231, Sample Num: 3696, Cur Loss: 1.42146349, Cur Avg Loss: 1.51985496, Log Avg loss: 1.78329660, Global Avg Loss: 1.96438149, Time: 0.0404 Steps: 23650, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000241, Sample Num: 3856, Cur Loss: 1.67439437, Cur Avg Loss: 1.52312586, Log Avg loss: 1.59868383, Global Avg Loss: 1.96422693, Time: 0.0403 Steps: 23660, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000251, Sample Num: 4016, Cur Loss: 1.73529637, Cur Avg Loss: 1.51654988, Log Avg loss: 1.35806857, Global Avg Loss: 1.96397084, Time: 0.0402 Steps: 23670, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000261, Sample Num: 4176, Cur Loss: 1.51850224, Cur Avg Loss: 1.51921788, Log Avg loss: 1.58618487, Global Avg Loss: 1.96381131, Time: 0.0402 Steps: 23680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000271, Sample Num: 4336, Cur Loss: 1.14727771, Cur Avg Loss: 1.51163487, Log Avg loss: 1.31371809, Global Avg Loss: 1.96353689, Time: 0.0402 Steps: 23690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000281, Sample Num: 4496, Cur Loss: 1.55177915, Cur Avg Loss: 1.50298921, Log Avg loss: 1.26869200, Global Avg Loss: 1.96324371, Time: 0.0403 Steps: 23700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000291, Sample Num: 4656, Cur Loss: 1.80562854, Cur Avg Loss: 1.49439118, Log Avg loss: 1.25278661, Global Avg Loss: 1.96294406, Time: 0.0402 Steps: 23710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000301, Sample Num: 4816, Cur Loss: 1.35724068, Cur Avg Loss: 1.48788988, Log Avg loss: 1.29870198, Global Avg Loss: 1.96266403, Time: 0.0402 Steps: 23720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000311, Sample Num: 4976, Cur Loss: 0.49156606, Cur Avg Loss: 1.49157411, Log Avg loss: 1.60246927, Global Avg Loss: 1.96251224, Time: 0.0403 Steps: 23730, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000321, Sample Num: 5136, Cur Loss: 0.63198888, Cur Avg Loss: 1.48018262, Log Avg loss: 1.12590745, Global Avg Loss: 1.96215983, Time: 0.0402 Steps: 23740, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000331, Sample Num: 5296, Cur Loss: 1.85304022, Cur Avg Loss: 1.49525212, Log Avg loss: 1.97898291, Global Avg Loss: 1.96216692, Time: 0.0403 Steps: 23750, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000341, Sample Num: 5456, Cur Loss: 1.86147094, Cur Avg Loss: 1.49453780, Log Avg loss: 1.47089379, Global Avg Loss: 1.96196015, Time: 0.0402 Steps: 23760, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000351, Sample Num: 5616, Cur Loss: 0.80882752, Cur Avg Loss: 1.50478929, Log Avg loss: 1.85436507, Global Avg Loss: 1.96191489, Time: 0.0402 Steps: 23770, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000361, Sample Num: 5776, Cur Loss: 2.76735973, Cur Avg Loss: 1.51566428, Log Avg loss: 1.89737650, Global Avg Loss: 1.96188775, Time: 0.0402 Steps: 23780, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000371, Sample Num: 5936, Cur Loss: 0.86231315, Cur Avg Loss: 1.52110064, Log Avg loss: 1.71735318, Global Avg Loss: 1.96178496, Time: 0.0403 Steps: 23790, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000381, Sample Num: 6096, Cur Loss: 1.78631282, Cur Avg Loss: 1.51811499, Log Avg loss: 1.40734756, Global Avg Loss: 1.96155200, Time: 0.0403 Steps: 23800, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000391, Sample Num: 6256, Cur Loss: 1.29526949, Cur Avg Loss: 1.51658769, Log Avg loss: 1.45839755, Global Avg Loss: 1.96134068, Time: 0.0404 Steps: 23810, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000401, Sample Num: 6416, Cur Loss: 1.10608947, Cur Avg Loss: 1.50858725, Log Avg loss: 1.19576987, Global Avg Loss: 1.96101928, Time: 0.0403 Steps: 23820, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000411, Sample Num: 6576, Cur Loss: 1.23374927, Cur Avg Loss: 1.51501851, Log Avg loss: 1.77291213, Global Avg Loss: 1.96094035, Time: 0.0402 Steps: 23830, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000421, Sample Num: 6736, Cur Loss: 1.94086552, Cur Avg Loss: 1.51670805, Log Avg loss: 1.58614838, Global Avg Loss: 1.96078313, Time: 0.0403 Steps: 23840, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000431, Sample Num: 6896, Cur Loss: 1.48547864, Cur Avg Loss: 1.52145229, Log Avg loss: 1.72118445, Global Avg Loss: 1.96068267, Time: 0.0403 Steps: 23850, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000441, Sample Num: 7056, Cur Loss: 1.55550182, Cur Avg Loss: 1.52356142, Log Avg loss: 1.61446490, Global Avg Loss: 1.96053757, Time: 0.0403 Steps: 23860, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000451, Sample Num: 7216, Cur Loss: 1.15828872, Cur Avg Loss: 1.53135665, Log Avg loss: 1.87512642, Global Avg Loss: 1.96050179, Time: 0.0402 Steps: 23870, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000461, Sample Num: 7376, Cur Loss: 1.34180760, Cur Avg Loss: 1.52623029, Log Avg loss: 1.29503151, Global Avg Loss: 1.96022312, Time: 0.0403 Steps: 23880, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000471, Sample Num: 7536, Cur Loss: 1.45660591, Cur Avg Loss: 1.53117210, Log Avg loss: 1.75898947, Global Avg Loss: 1.96013888, Time: 0.0404 Steps: 23890, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000481, Sample Num: 7696, Cur Loss: 1.55703914, Cur Avg Loss: 1.53925209, Log Avg loss: 1.91981982, Global Avg Loss: 1.96012201, Time: 0.0402 Steps: 23900, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000491, Sample Num: 7856, Cur Loss: 1.17798138, Cur Avg Loss: 1.53768094, Log Avg loss: 1.46210867, Global Avg Loss: 1.95991373, Time: 0.0404 Steps: 23910, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000501, Sample Num: 8016, Cur Loss: 1.47110462, Cur Avg Loss: 1.53039353, Log Avg loss: 1.17258130, Global Avg Loss: 1.95958457, Time: 0.0821 Steps: 23920, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000511, Sample Num: 8176, Cur Loss: 0.21822268, Cur Avg Loss: 1.52982064, Log Avg loss: 1.50111917, Global Avg Loss: 1.95939299, Time: 0.0589 Steps: 23930, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000521, Sample Num: 8336, Cur Loss: 1.52388716, Cur Avg Loss: 1.52677985, Log Avg loss: 1.37139530, Global Avg Loss: 1.95914737, Time: 0.0432 Steps: 23940, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000531, Sample Num: 8496, Cur Loss: 0.73016083, Cur Avg Loss: 1.52735789, Log Avg loss: 1.55747383, Global Avg Loss: 1.95897966, Time: 0.0723 Steps: 23950, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000541, Sample Num: 8656, Cur Loss: 1.37103975, Cur Avg Loss: 1.52891986, Log Avg loss: 1.61186069, Global Avg Loss: 1.95883479, Time: 0.0404 Steps: 23960, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000551, Sample Num: 8816, Cur Loss: 0.85773766, Cur Avg Loss: 1.52982567, Log Avg loss: 1.57882993, Global Avg Loss: 1.95867625, Time: 0.0431 Steps: 23970, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000561, Sample Num: 8976, Cur Loss: 0.32683465, Cur Avg Loss: 1.52346514, Log Avg loss: 1.17299974, Global Avg Loss: 1.95834861, Time: 0.0544 Steps: 23980, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000571, Sample Num: 9136, Cur Loss: 0.63153148, Cur Avg Loss: 1.52548762, Log Avg loss: 1.63894892, Global Avg Loss: 1.95821548, Time: 0.0564 Steps: 23990, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000581, Sample Num: 9296, Cur Loss: 1.72184718, Cur Avg Loss: 1.52538839, Log Avg loss: 1.51972237, Global Avg Loss: 1.95803277, Time: 0.0403 Steps: 24000, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000591, Sample Num: 9456, Cur Loss: 0.68880343, Cur Avg Loss: 1.52512141, Log Avg loss: 1.50960963, Global Avg Loss: 1.95784601, Time: 0.0403 Steps: 24010, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000601, Sample Num: 9616, Cur Loss: 2.71202302, Cur Avg Loss: 1.52439642, Log Avg loss: 1.48154968, Global Avg Loss: 1.95764771, Time: 0.0402 Steps: 24020, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000611, Sample Num: 9776, Cur Loss: 0.67206049, Cur Avg Loss: 1.52304018, Log Avg loss: 1.44153013, Global Avg Loss: 1.95743293, Time: 0.0402 Steps: 24030, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000621, Sample Num: 9936, Cur Loss: 1.48936439, Cur Avg Loss: 1.52513433, Log Avg loss: 1.65308654, Global Avg Loss: 1.95730633, Time: 0.0403 Steps: 24040, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000631, Sample Num: 10096, Cur Loss: 1.01634061, Cur Avg Loss: 1.52061174, Log Avg loss: 1.23975922, Global Avg Loss: 1.95700798, Time: 0.0403 Steps: 24050, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000641, Sample Num: 10256, Cur Loss: 1.43557358, Cur Avg Loss: 1.51766694, Log Avg loss: 1.33184992, Global Avg Loss: 1.95674814, Time: 0.0402 Steps: 24060, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000651, Sample Num: 10416, Cur Loss: 1.58197641, Cur Avg Loss: 1.51656812, Log Avg loss: 1.44613397, Global Avg Loss: 1.95653601, Time: 0.0402 Steps: 24070, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000661, Sample Num: 10576, Cur Loss: 1.20102239, Cur Avg Loss: 1.51433232, Log Avg loss: 1.36878167, Global Avg Loss: 1.95629192, Time: 0.0403 Steps: 24080, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000671, Sample Num: 10736, Cur Loss: 0.37937573, Cur Avg Loss: 1.51005086, Log Avg loss: 1.22704628, Global Avg Loss: 1.95598921, Time: 0.0402 Steps: 24090, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000681, Sample Num: 10896, Cur Loss: 1.66165543, Cur Avg Loss: 1.50816448, Log Avg loss: 1.38158835, Global Avg Loss: 1.95575087, Time: 0.0402 Steps: 24100, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000691, Sample Num: 11056, Cur Loss: 1.44179320, Cur Avg Loss: 1.51056800, Log Avg loss: 1.67424765, Global Avg Loss: 1.95563411, Time: 0.0402 Steps: 24110, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000701, Sample Num: 11216, Cur Loss: 0.70464957, Cur Avg Loss: 1.51392872, Log Avg loss: 1.74615484, Global Avg Loss: 1.95554726, Time: 0.0403 Steps: 24120, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000711, Sample Num: 11376, Cur Loss: 1.57890821, Cur Avg Loss: 1.51091027, Log Avg loss: 1.29931669, Global Avg Loss: 1.95527530, Time: 0.0402 Steps: 24130, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000721, Sample Num: 11536, Cur Loss: 0.31090203, Cur Avg Loss: 1.50576883, Log Avg loss: 1.14021239, Global Avg Loss: 1.95493766, Time: 0.0402 Steps: 24140, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000731, Sample Num: 11696, Cur Loss: 1.76739657, Cur Avg Loss: 1.50694565, Log Avg loss: 1.59179472, Global Avg Loss: 1.95478729, Time: 0.0403 Steps: 24150, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000741, Sample Num: 11856, Cur Loss: 1.66933811, Cur Avg Loss: 1.50652541, Log Avg loss: 1.47580573, Global Avg Loss: 1.95458904, Time: 0.0402 Steps: 24160, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000751, Sample Num: 12016, Cur Loss: 1.36098289, Cur Avg Loss: 1.50647521, Log Avg loss: 1.50275511, Global Avg Loss: 1.95440210, Time: 0.0402 Steps: 24170, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000761, Sample Num: 12176, Cur Loss: 1.89979064, Cur Avg Loss: 1.50818417, Log Avg loss: 1.63652689, Global Avg Loss: 1.95427064, Time: 0.0403 Steps: 24180, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000771, Sample Num: 12336, Cur Loss: 1.78198004, Cur Avg Loss: 1.50890640, Log Avg loss: 1.56386815, Global Avg Loss: 1.95410925, Time: 0.0402 Steps: 24190, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000781, Sample Num: 12496, Cur Loss: 1.88651943, Cur Avg Loss: 1.50729978, Log Avg loss: 1.38342928, Global Avg Loss: 1.95387343, Time: 0.0402 Steps: 24200, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000791, Sample Num: 12656, Cur Loss: 1.36132896, Cur Avg Loss: 1.51071856, Log Avg loss: 1.77772529, Global Avg Loss: 1.95380067, Time: 0.0402 Steps: 24210, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000801, Sample Num: 12816, Cur Loss: 1.07913589, Cur Avg Loss: 1.51559033, Log Avg loss: 1.90094740, Global Avg Loss: 1.95377885, Time: 0.0402 Steps: 24220, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000811, Sample Num: 12976, Cur Loss: 1.81836402, Cur Avg Loss: 1.51665818, Log Avg loss: 1.60219288, Global Avg Loss: 1.95363374, Time: 0.0402 Steps: 24230, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000821, Sample Num: 13136, Cur Loss: 1.73608184, Cur Avg Loss: 1.51588667, Log Avg loss: 1.45331746, Global Avg Loss: 1.95342734, Time: 0.0402 Steps: 24240, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000831, Sample Num: 13296, Cur Loss: 1.53966188, Cur Avg Loss: 1.51750164, Log Avg loss: 1.65009121, Global Avg Loss: 1.95330226, Time: 0.0402 Steps: 24250, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000841, Sample Num: 13456, Cur Loss: 1.80481184, Cur Avg Loss: 1.51770391, Log Avg loss: 1.53451177, Global Avg Loss: 1.95312963, Time: 0.0402 Steps: 24260, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000851, Sample Num: 13616, Cur Loss: 2.46889186, Cur Avg Loss: 1.51485668, Log Avg loss: 1.27540506, Global Avg Loss: 1.95285039, Time: 0.0699 Steps: 24270, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000861, Sample Num: 13776, Cur Loss: 1.27375841, Cur Avg Loss: 1.51950069, Log Avg loss: 1.91470577, Global Avg Loss: 1.95283468, Time: 0.0560 Steps: 24280, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000871, Sample Num: 13936, Cur Loss: 1.90072298, Cur Avg Loss: 1.52020592, Log Avg loss: 1.58092664, Global Avg Loss: 1.95268156, Time: 0.0403 Steps: 24290, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000881, Sample Num: 14096, Cur Loss: 1.79119515, Cur Avg Loss: 1.52263512, Log Avg loss: 1.73421857, Global Avg Loss: 1.95259166, Time: 0.0658 Steps: 24300, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000891, Sample Num: 14256, Cur Loss: 1.80853665, Cur Avg Loss: 1.52791192, Log Avg loss: 1.99279780, Global Avg Loss: 1.95260820, Time: 0.0562 Steps: 24310, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000901, Sample Num: 14416, Cur Loss: 1.76347852, Cur Avg Loss: 1.52756088, Log Avg loss: 1.49628283, Global Avg Loss: 1.95242057, Time: 0.0519 Steps: 24320, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000911, Sample Num: 14576, Cur Loss: 1.97588897, Cur Avg Loss: 1.52480519, Log Avg loss: 1.27651769, Global Avg Loss: 1.95214276, Time: 0.0610 Steps: 24330, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000921, Sample Num: 14736, Cur Loss: 0.47676232, Cur Avg Loss: 1.52617218, Log Avg loss: 1.65070514, Global Avg Loss: 1.95201892, Time: 0.0481 Steps: 24340, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000931, Sample Num: 14896, Cur Loss: 1.59192491, Cur Avg Loss: 1.52517457, Log Avg loss: 1.43329411, Global Avg Loss: 1.95180589, Time: 0.0682 Steps: 24350, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000941, Sample Num: 15056, Cur Loss: 2.20922470, Cur Avg Loss: 1.52656631, Log Avg loss: 1.65613760, Global Avg Loss: 1.95168451, Time: 0.0403 Steps: 24360, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000951, Sample Num: 15216, Cur Loss: 1.41505647, Cur Avg Loss: 1.52927925, Log Avg loss: 1.78456721, Global Avg Loss: 1.95161594, Time: 0.0402 Steps: 24370, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000961, Sample Num: 15376, Cur Loss: 1.19439554, Cur Avg Loss: 1.52788272, Log Avg loss: 1.39507278, Global Avg Loss: 1.95138766, Time: 0.0403 Steps: 24380, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000971, Sample Num: 15536, Cur Loss: 0.64678228, Cur Avg Loss: 1.52559830, Log Avg loss: 1.30606564, Global Avg Loss: 1.95112307, Time: 0.0402 Steps: 24390, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000981, Sample Num: 15696, Cur Loss: 1.97805262, Cur Avg Loss: 1.52644807, Log Avg loss: 1.60896087, Global Avg Loss: 1.95098284, Time: 0.0403 Steps: 24400, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000991, Sample Num: 15856, Cur Loss: 0.77566254, Cur Avg Loss: 1.52784523, Log Avg loss: 1.66490635, Global Avg Loss: 1.95086565, Time: 0.0402 Steps: 24410, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001001, Sample Num: 16016, Cur Loss: 1.46990418, Cur Avg Loss: 1.52884515, Log Avg loss: 1.62793660, Global Avg Loss: 1.95073341, Time: 0.0403 Steps: 24420, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001011, Sample Num: 16176, Cur Loss: 0.95455867, Cur Avg Loss: 1.52625132, Log Avg loss: 1.26660983, Global Avg Loss: 1.95045337, Time: 0.0402 Steps: 24430, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001021, Sample Num: 16336, Cur Loss: 1.43283916, Cur Avg Loss: 1.53108051, Log Avg loss: 2.01931142, Global Avg Loss: 1.95048155, Time: 0.0403 Steps: 24440, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001031, Sample Num: 16496, Cur Loss: 1.80331349, Cur Avg Loss: 1.53046362, Log Avg loss: 1.46747862, Global Avg Loss: 1.95028400, Time: 0.0403 Steps: 24450, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001041, Sample Num: 16656, Cur Loss: 1.07735610, Cur Avg Loss: 1.53036225, Log Avg loss: 1.51991149, Global Avg Loss: 1.95010805, Time: 0.0403 Steps: 24460, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001051, Sample Num: 16816, Cur Loss: 1.90979242, Cur Avg Loss: 1.53265952, Log Avg loss: 1.77180491, Global Avg Loss: 1.95003519, Time: 0.0402 Steps: 24470, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001061, Sample Num: 16976, Cur Loss: 1.58867860, Cur Avg Loss: 1.53328826, Log Avg loss: 1.59936920, Global Avg Loss: 1.94989194, Time: 0.0402 Steps: 24480, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001071, Sample Num: 17136, Cur Loss: 0.78439188, Cur Avg Loss: 1.53079381, Log Avg loss: 1.26613277, Global Avg Loss: 1.94961274, Time: 0.0402 Steps: 24490, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001081, Sample Num: 17296, Cur Loss: 1.62101138, Cur Avg Loss: 1.53040161, Log Avg loss: 1.48839721, Global Avg Loss: 1.94942449, Time: 0.0403 Steps: 24500, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001091, Sample Num: 17456, Cur Loss: 1.18955708, Cur Avg Loss: 1.52826858, Log Avg loss: 1.29768769, Global Avg Loss: 1.94915858, Time: 0.0402 Steps: 24510, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001101, Sample Num: 17616, Cur Loss: 0.93749273, Cur Avg Loss: 1.53107623, Log Avg loss: 1.83739076, Global Avg Loss: 1.94911300, Time: 0.0403 Steps: 24520, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001111, Sample Num: 17776, Cur Loss: 3.03120518, Cur Avg Loss: 1.53368066, Log Avg loss: 1.82042799, Global Avg Loss: 1.94906054, Time: 0.0402 Steps: 24530, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001121, Sample Num: 17936, Cur Loss: 0.69015145, Cur Avg Loss: 1.53646708, Log Avg loss: 1.84603820, Global Avg Loss: 1.94901856, Time: 0.0402 Steps: 24540, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001131, Sample Num: 18096, Cur Loss: 1.24677014, Cur Avg Loss: 1.53719479, Log Avg loss: 1.61877191, Global Avg Loss: 1.94888404, Time: 0.0402 Steps: 24550, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001141, Sample Num: 18256, Cur Loss: 2.58163357, Cur Avg Loss: 1.53615437, Log Avg loss: 1.41848280, Global Avg Loss: 1.94866808, Time: 0.0402 Steps: 24560, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001151, Sample Num: 18416, Cur Loss: 1.52957559, Cur Avg Loss: 1.53764829, Log Avg loss: 1.70810402, Global Avg Loss: 1.94857017, Time: 0.0402 Steps: 24570, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001161, Sample Num: 18576, Cur Loss: 1.21671104, Cur Avg Loss: 1.53796972, Log Avg loss: 1.57496662, Global Avg Loss: 1.94841817, Time: 0.0402 Steps: 24580, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001171, Sample Num: 18736, Cur Loss: 2.59612489, Cur Avg Loss: 1.53557387, Log Avg loss: 1.25741594, Global Avg Loss: 1.94813716, Time: 0.0402 Steps: 24590, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001181, Sample Num: 18896, Cur Loss: 1.10690212, Cur Avg Loss: 1.53184923, Log Avg loss: 1.09569395, Global Avg Loss: 1.94779064, Time: 0.0402 Steps: 24600, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001191, Sample Num: 19056, Cur Loss: 2.09395599, Cur Avg Loss: 1.53173452, Log Avg loss: 1.51818680, Global Avg Loss: 1.94761608, Time: 0.0402 Steps: 24610, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001201, Sample Num: 19216, Cur Loss: 2.95833349, Cur Avg Loss: 1.53062355, Log Avg loss: 1.39830666, Global Avg Loss: 1.94739296, Time: 0.0604 Steps: 24620, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001211, Sample Num: 19376, Cur Loss: 0.64723682, Cur Avg Loss: 1.53020171, Log Avg loss: 1.47953967, Global Avg Loss: 1.94720301, Time: 0.0573 Steps: 24630, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001221, Sample Num: 19536, Cur Loss: 0.77306080, Cur Avg Loss: 1.53183726, Log Avg loss: 1.72990240, Global Avg Loss: 1.94711482, Time: 0.0653 Steps: 24640, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001231, Sample Num: 19696, Cur Loss: 1.03989315, Cur Avg Loss: 1.52913787, Log Avg loss: 1.19954168, Global Avg Loss: 1.94681154, Time: 0.0406 Steps: 24650, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001241, Sample Num: 19856, Cur Loss: 1.16292679, Cur Avg Loss: 1.52660094, Log Avg loss: 1.21430463, Global Avg Loss: 1.94651450, Time: 0.0483 Steps: 24660, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001251, Sample Num: 20016, Cur Loss: 1.21707034, Cur Avg Loss: 1.52736609, Log Avg loss: 1.62232098, Global Avg Loss: 1.94638309, Time: 0.0406 Steps: 24670, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001261, Sample Num: 20176, Cur Loss: 1.82422161, Cur Avg Loss: 1.52641276, Log Avg loss: 1.40715215, Global Avg Loss: 1.94616460, Time: 0.0454 Steps: 24680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001271, Sample Num: 20336, Cur Loss: 1.79427779, Cur Avg Loss: 1.52671874, Log Avg loss: 1.56530246, Global Avg Loss: 1.94601034, Time: 0.0581 Steps: 24690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001281, Sample Num: 20496, Cur Loss: 0.58992934, Cur Avg Loss: 1.52553117, Log Avg loss: 1.37459065, Global Avg Loss: 1.94577900, Time: 0.0406 Steps: 24700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001291, Sample Num: 20656, Cur Loss: 2.41878653, Cur Avg Loss: 1.52741111, Log Avg loss: 1.76823230, Global Avg Loss: 1.94570715, Time: 0.0402 Steps: 24710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001301, Sample Num: 20816, Cur Loss: 0.72213405, Cur Avg Loss: 1.52683112, Log Avg loss: 1.45195383, Global Avg Loss: 1.94550741, Time: 0.0403 Steps: 24720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001311, Sample Num: 20976, Cur Loss: 1.25935435, Cur Avg Loss: 1.52725357, Log Avg loss: 1.58221437, Global Avg Loss: 1.94536050, Time: 0.0403 Steps: 24730, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001321, Sample Num: 21136, Cur Loss: 1.22371984, Cur Avg Loss: 1.52765519, Log Avg loss: 1.58030750, Global Avg Loss: 1.94521295, Time: 0.0402 Steps: 24740, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001331, Sample Num: 21296, Cur Loss: 1.89824724, Cur Avg Loss: 1.53021409, Log Avg loss: 1.86824549, Global Avg Loss: 1.94518185, Time: 0.0402 Steps: 24750, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001341, Sample Num: 21456, Cur Loss: 1.80030918, Cur Avg Loss: 1.52971319, Log Avg loss: 1.46304246, Global Avg Loss: 1.94498713, Time: 0.0403 Steps: 24760, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001351, Sample Num: 21616, Cur Loss: 1.62188900, Cur Avg Loss: 1.52896269, Log Avg loss: 1.42832059, Global Avg Loss: 1.94477854, Time: 0.0403 Steps: 24770, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001361, Sample Num: 21776, Cur Loss: 1.58528674, Cur Avg Loss: 1.52960456, Log Avg loss: 1.61632166, Global Avg Loss: 1.94464599, Time: 0.0403 Steps: 24780, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001371, Sample Num: 21936, Cur Loss: 1.52982354, Cur Avg Loss: 1.52935986, Log Avg loss: 1.49605550, Global Avg Loss: 1.94446503, Time: 0.0403 Steps: 24790, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001381, Sample Num: 22096, Cur Loss: 1.24358213, Cur Avg Loss: 1.52922083, Log Avg loss: 1.51016067, Global Avg Loss: 1.94428991, Time: 0.0403 Steps: 24800, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001391, Sample Num: 22256, Cur Loss: 0.85770220, Cur Avg Loss: 1.52739387, Log Avg loss: 1.27509083, Global Avg Loss: 1.94402018, Time: 0.0402 Steps: 24810, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001401, Sample Num: 22416, Cur Loss: 1.37405729, Cur Avg Loss: 1.52765626, Log Avg loss: 1.56415432, Global Avg Loss: 1.94386713, Time: 0.0403 Steps: 24820, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001411, Sample Num: 22576, Cur Loss: 0.73703063, Cur Avg Loss: 1.52787328, Log Avg loss: 1.55827743, Global Avg Loss: 1.94371184, Time: 0.0402 Steps: 24830, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001421, Sample Num: 22736, Cur Loss: 1.70594192, Cur Avg Loss: 1.52866930, Log Avg loss: 1.64098821, Global Avg Loss: 1.94358997, Time: 0.0403 Steps: 24840, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001431, Sample Num: 22896, Cur Loss: 0.78343391, Cur Avg Loss: 1.52850037, Log Avg loss: 1.50449545, Global Avg Loss: 1.94341327, Time: 0.0403 Steps: 24850, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001441, Sample Num: 23056, Cur Loss: 1.33454704, Cur Avg Loss: 1.52711984, Log Avg loss: 1.32956551, Global Avg Loss: 1.94316635, Time: 0.0403 Steps: 24860, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001451, Sample Num: 23216, Cur Loss: 1.41904831, Cur Avg Loss: 1.52627515, Log Avg loss: 1.40455567, Global Avg Loss: 1.94294978, Time: 0.0402 Steps: 24870, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001461, Sample Num: 23376, Cur Loss: 0.92349088, Cur Avg Loss: 1.52669382, Log Avg loss: 1.58744317, Global Avg Loss: 1.94280689, Time: 0.0403 Steps: 24880, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001471, Sample Num: 23536, Cur Loss: 1.26005602, Cur Avg Loss: 1.52654405, Log Avg loss: 1.50466235, Global Avg Loss: 1.94263086, Time: 0.0402 Steps: 24890, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001481, Sample Num: 23696, Cur Loss: 0.84051400, Cur Avg Loss: 1.52391731, Log Avg loss: 1.13752435, Global Avg Loss: 1.94230753, Time: 0.0404 Steps: 24900, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001491, Sample Num: 23856, Cur Loss: 2.33136392, Cur Avg Loss: 1.52677416, Log Avg loss: 1.94987360, Global Avg Loss: 1.94231056, Time: 0.0403 Steps: 24910, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001501, Sample Num: 24016, Cur Loss: 1.87690198, Cur Avg Loss: 1.52594178, Log Avg loss: 1.40183287, Global Avg Loss: 1.94209368, Time: 0.0403 Steps: 24920, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001511, Sample Num: 24176, Cur Loss: 2.45313454, Cur Avg Loss: 1.52669672, Log Avg loss: 1.64001305, Global Avg Loss: 1.94197251, Time: 0.0403 Steps: 24930, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001521, Sample Num: 24336, Cur Loss: 1.89245057, Cur Avg Loss: 1.52660980, Log Avg loss: 1.51347731, Global Avg Loss: 1.94180070, Time: 0.0403 Steps: 24940, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001531, Sample Num: 24496, Cur Loss: 1.08224428, Cur Avg Loss: 1.52585985, Log Avg loss: 1.41179143, Global Avg Loss: 1.94158827, Time: 0.0402 Steps: 24950, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001541, Sample Num: 24656, Cur Loss: 2.43237090, Cur Avg Loss: 1.52565958, Log Avg loss: 1.49499939, Global Avg Loss: 1.94140935, Time: 0.0756 Steps: 24960, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001551, Sample Num: 24816, Cur Loss: 2.17118669, Cur Avg Loss: 1.52695664, Log Avg loss: 1.72683228, Global Avg Loss: 1.94132341, Time: 0.0747 Steps: 24970, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001561, Sample Num: 24976, Cur Loss: 2.74956679, Cur Avg Loss: 1.52735025, Log Avg loss: 1.58840001, Global Avg Loss: 1.94118213, Time: 0.0826 Steps: 24980, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001571, Sample Num: 25136, Cur Loss: 2.53934169, Cur Avg Loss: 1.52922797, Log Avg loss: 1.82234075, Global Avg Loss: 1.94113457, Time: 0.0404 Steps: 24990, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001581, Sample Num: 25296, Cur Loss: 1.28243899, Cur Avg Loss: 1.53045869, Log Avg loss: 1.72380326, Global Avg Loss: 1.94104764, Time: 0.0649 Steps: 25000, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001591, Sample Num: 25456, Cur Loss: 2.26599431, Cur Avg Loss: 1.53227531, Log Avg loss: 1.81948335, Global Avg Loss: 1.94099903, Time: 0.0463 Steps: 25010, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001601, Sample Num: 25616, Cur Loss: 1.31036735, Cur Avg Loss: 1.53183106, Log Avg loss: 1.46115162, Global Avg Loss: 1.94080725, Time: 0.0858 Steps: 25020, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001611, Sample Num: 25776, Cur Loss: 1.85120738, Cur Avg Loss: 1.53124478, Log Avg loss: 1.43738153, Global Avg Loss: 1.94060612, Time: 0.0438 Steps: 25030, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001621, Sample Num: 25936, Cur Loss: 1.63422143, Cur Avg Loss: 1.53039151, Log Avg loss: 1.39292939, Global Avg Loss: 1.94038740, Time: 0.0406 Steps: 25040, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001631, Sample Num: 26096, Cur Loss: 0.76173437, Cur Avg Loss: 1.53039250, Log Avg loss: 1.53055318, Global Avg Loss: 1.94022379, Time: 0.0403 Steps: 25050, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001641, Sample Num: 26256, Cur Loss: 0.95676368, Cur Avg Loss: 1.52825051, Log Avg loss: 1.17889079, Global Avg Loss: 1.93991999, Time: 0.0403 Steps: 25060, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001651, Sample Num: 26416, Cur Loss: 2.12851334, Cur Avg Loss: 1.52874225, Log Avg loss: 1.60943791, Global Avg Loss: 1.93978817, Time: 0.0403 Steps: 25070, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001661, Sample Num: 26576, Cur Loss: 1.04374421, Cur Avg Loss: 1.52647103, Log Avg loss: 1.15149156, Global Avg Loss: 1.93947385, Time: 0.0403 Steps: 25080, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001671, Sample Num: 26736, Cur Loss: 1.76388454, Cur Avg Loss: 1.52596733, Log Avg loss: 1.44230353, Global Avg Loss: 1.93927570, Time: 0.0404 Steps: 25090, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001681, Sample Num: 26896, Cur Loss: 1.45506918, Cur Avg Loss: 1.52679844, Log Avg loss: 1.66567597, Global Avg Loss: 1.93916669, Time: 0.0403 Steps: 25100, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001691, Sample Num: 27056, Cur Loss: 1.63944030, Cur Avg Loss: 1.52640404, Log Avg loss: 1.46010695, Global Avg Loss: 1.93897591, Time: 0.0404 Steps: 25110, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001701, Sample Num: 27216, Cur Loss: 0.80908030, Cur Avg Loss: 1.52692480, Log Avg loss: 1.61498388, Global Avg Loss: 1.93884693, Time: 0.0403 Steps: 25120, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001711, Sample Num: 27376, Cur Loss: 0.79678977, Cur Avg Loss: 1.52575365, Log Avg loss: 1.32654250, Global Avg Loss: 1.93860328, Time: 0.0403 Steps: 25130, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001721, Sample Num: 27536, Cur Loss: 1.99582386, Cur Avg Loss: 1.52562288, Log Avg loss: 1.50324749, Global Avg Loss: 1.93843010, Time: 0.0403 Steps: 25140, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001731, Sample Num: 27696, Cur Loss: 0.69314373, Cur Avg Loss: 1.52435648, Log Avg loss: 1.30640891, Global Avg Loss: 1.93817880, Time: 0.0404 Steps: 25150, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001741, Sample Num: 27856, Cur Loss: 0.47211069, Cur Avg Loss: 1.52402654, Log Avg loss: 1.46691477, Global Avg Loss: 1.93799150, Time: 0.0404 Steps: 25160, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001751, Sample Num: 28016, Cur Loss: 0.82811850, Cur Avg Loss: 1.52295856, Log Avg loss: 1.33702224, Global Avg Loss: 1.93775273, Time: 0.0403 Steps: 25170, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001761, Sample Num: 28176, Cur Loss: 1.51130986, Cur Avg Loss: 1.52312509, Log Avg loss: 1.55228488, Global Avg Loss: 1.93759965, Time: 0.0403 Steps: 25180, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001771, Sample Num: 28336, Cur Loss: 1.41577637, Cur Avg Loss: 1.52263254, Log Avg loss: 1.43589453, Global Avg Loss: 1.93740048, Time: 0.0402 Steps: 25190, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001781, Sample Num: 28496, Cur Loss: 1.40555739, Cur Avg Loss: 1.52281237, Log Avg loss: 1.55465931, Global Avg Loss: 1.93724860, Time: 0.0403 Steps: 25200, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001791, Sample Num: 28656, Cur Loss: 1.32164824, Cur Avg Loss: 1.52384359, Log Avg loss: 1.70750437, Global Avg Loss: 1.93715747, Time: 0.0403 Steps: 25210, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001801, Sample Num: 28816, Cur Loss: 1.73634577, Cur Avg Loss: 1.52586951, Log Avg loss: 1.88871187, Global Avg Loss: 1.93713826, Time: 0.0403 Steps: 25220, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001811, Sample Num: 28976, Cur Loss: 2.23899531, Cur Avg Loss: 1.52659915, Log Avg loss: 1.65800672, Global Avg Loss: 1.93702762, Time: 0.0404 Steps: 25230, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001821, Sample Num: 29136, Cur Loss: 2.53867626, Cur Avg Loss: 1.52712531, Log Avg loss: 1.62241324, Global Avg Loss: 1.93690297, Time: 0.0403 Steps: 25240, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001831, Sample Num: 29296, Cur Loss: 1.64122045, Cur Avg Loss: 1.52875785, Log Avg loss: 1.82604460, Global Avg Loss: 1.93685907, Time: 0.0403 Steps: 25250, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001841, Sample Num: 29456, Cur Loss: 1.54859507, Cur Avg Loss: 1.53093630, Log Avg loss: 1.92981035, Global Avg Loss: 1.93685628, Time: 0.0403 Steps: 25260, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001851, Sample Num: 29616, Cur Loss: 0.66525459, Cur Avg Loss: 1.53123095, Log Avg loss: 1.58547447, Global Avg Loss: 1.93671723, Time: 0.0402 Steps: 25270, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001861, Sample Num: 29776, Cur Loss: 2.33405542, Cur Avg Loss: 1.53144695, Log Avg loss: 1.57142954, Global Avg Loss: 1.93657273, Time: 0.0403 Steps: 25280, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001871, Sample Num: 29936, Cur Loss: 1.13857293, Cur Avg Loss: 1.52954996, Log Avg loss: 1.17652078, Global Avg Loss: 1.93627220, Time: 0.0514 Steps: 25290, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001881, Sample Num: 30096, Cur Loss: 1.36001229, Cur Avg Loss: 1.52916326, Log Avg loss: 1.45681032, Global Avg Loss: 1.93608268, Time: 0.0553 Steps: 25300, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001891, Sample Num: 30256, Cur Loss: 0.79782522, Cur Avg Loss: 1.52849716, Log Avg loss: 1.40320356, Global Avg Loss: 1.93587214, Time: 0.0661 Steps: 25310, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001901, Sample Num: 30416, Cur Loss: 1.99172771, Cur Avg Loss: 1.53013032, Log Avg loss: 1.83896141, Global Avg Loss: 1.93583387, Time: 0.0405 Steps: 25320, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001911, Sample Num: 30576, Cur Loss: 1.98975933, Cur Avg Loss: 1.53188975, Log Avg loss: 1.86635789, Global Avg Loss: 1.93580644, Time: 0.0402 Steps: 25330, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001921, Sample Num: 30736, Cur Loss: 0.62068969, Cur Avg Loss: 1.52947053, Log Avg loss: 1.06715755, Global Avg Loss: 1.93546364, Time: 0.0511 Steps: 25340, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001931, Sample Num: 30896, Cur Loss: 1.78846121, Cur Avg Loss: 1.52865941, Log Avg loss: 1.37284393, Global Avg Loss: 1.93524170, Time: 0.0420 Steps: 25350, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001941, Sample Num: 31056, Cur Loss: 2.42897749, Cur Avg Loss: 1.53009906, Log Avg loss: 1.80809541, Global Avg Loss: 1.93519157, Time: 0.0544 Steps: 25360, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001951, Sample Num: 31216, Cur Loss: 0.89982116, Cur Avg Loss: 1.52921197, Log Avg loss: 1.35702620, Global Avg Loss: 1.93496367, Time: 0.0405 Steps: 25370, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001961, Sample Num: 31376, Cur Loss: 1.14645147, Cur Avg Loss: 1.52793837, Log Avg loss: 1.27945946, Global Avg Loss: 1.93470540, Time: 0.0403 Steps: 25380, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001971, Sample Num: 31536, Cur Loss: 1.52562356, Cur Avg Loss: 1.52802332, Log Avg loss: 1.54468229, Global Avg Loss: 1.93455178, Time: 0.0403 Steps: 25390, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001981, Sample Num: 31696, Cur Loss: 1.26495075, Cur Avg Loss: 1.52903564, Log Avg loss: 1.72856303, Global Avg Loss: 1.93447069, Time: 0.0403 Steps: 25400, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001991, Sample Num: 31856, Cur Loss: 0.79257607, Cur Avg Loss: 1.52849044, Log Avg loss: 1.42048793, Global Avg Loss: 1.93426841, Time: 0.0403 Steps: 25410, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002001, Sample Num: 32016, Cur Loss: 1.38267040, Cur Avg Loss: 1.52780787, Log Avg loss: 1.39190825, Global Avg Loss: 1.93405505, Time: 0.0403 Steps: 25420, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002011, Sample Num: 32176, Cur Loss: 1.85526919, Cur Avg Loss: 1.52675588, Log Avg loss: 1.31625134, Global Avg Loss: 1.93381211, Time: 0.0403 Steps: 25430, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002021, Sample Num: 32336, Cur Loss: 1.51757908, Cur Avg Loss: 1.52627928, Log Avg loss: 1.43043512, Global Avg Loss: 1.93361424, Time: 0.0403 Steps: 25440, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002031, Sample Num: 32496, Cur Loss: 0.59533399, Cur Avg Loss: 1.52543661, Log Avg loss: 1.35513336, Global Avg Loss: 1.93338694, Time: 0.0403 Steps: 25450, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002041, Sample Num: 32656, Cur Loss: 3.28058314, Cur Avg Loss: 1.52672250, Log Avg loss: 1.78788767, Global Avg Loss: 1.93332979, Time: 0.0403 Steps: 25460, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002051, Sample Num: 32816, Cur Loss: 2.13615012, Cur Avg Loss: 1.52792193, Log Avg loss: 1.77272502, Global Avg Loss: 1.93326673, Time: 0.0403 Steps: 25470, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002061, Sample Num: 32976, Cur Loss: 0.51376641, Cur Avg Loss: 1.52738109, Log Avg loss: 1.41645509, Global Avg Loss: 1.93306390, Time: 0.0403 Steps: 25480, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002071, Sample Num: 33136, Cur Loss: 2.73474026, Cur Avg Loss: 1.52721308, Log Avg loss: 1.49258670, Global Avg Loss: 1.93289110, Time: 0.0403 Steps: 25490, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002081, Sample Num: 33296, Cur Loss: 2.16504169, Cur Avg Loss: 1.52819379, Log Avg loss: 1.73129704, Global Avg Loss: 1.93281204, Time: 0.0404 Steps: 25500, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002091, Sample Num: 33456, Cur Loss: 1.31210542, Cur Avg Loss: 1.52673799, Log Avg loss: 1.22378621, Global Avg Loss: 1.93253410, Time: 0.0403 Steps: 25510, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002101, Sample Num: 33616, Cur Loss: 1.95436263, Cur Avg Loss: 1.52579591, Log Avg loss: 1.32880681, Global Avg Loss: 1.93229753, Time: 0.0404 Steps: 25520, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002111, Sample Num: 33776, Cur Loss: 2.33512592, Cur Avg Loss: 1.52605213, Log Avg loss: 1.57988485, Global Avg Loss: 1.93215949, Time: 0.0403 Steps: 25530, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002121, Sample Num: 33936, Cur Loss: 1.21141243, Cur Avg Loss: 1.52695015, Log Avg loss: 1.71652212, Global Avg Loss: 1.93207506, Time: 0.0404 Steps: 25540, Updated lr: 0.000077 ***** Running evaluation checkpoint-25548 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-25548 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 94.977155, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.829798, "eval_total_loss": 1286.347717, "eval_mae": 1.240614, "eval_mse": 1.829338, "eval_r2": -0.162848, "eval_sp_statistic": 0.334398, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.382775, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.444899, "test_total_loss": 725.33907, "test_mae": 1.046814, "test_mse": 1.445087, "test_r2": 0.067328, "test_sp_statistic": 0.332223, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.38971, "test_ps_pvalue": 0.0, "lr": 7.672072072072073e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.931983087052617, "train_cur_epoch_loss": 3251.7680660635233, "train_cur_epoch_avg_loss": 1.5273687487381509, "train_cur_epoch_time": 94.97715520858765, "train_cur_epoch_avg_time": 0.04461115791854751, "epoch": 12, "step": 25548} ################################################## Training, Epoch: 0013, Batch: 000002, Sample Num: 32, Cur Loss: 0.91114306, Cur Avg Loss: 1.04893911, Log Avg loss: 1.52046750, Global Avg Loss: 1.93191396, Time: 0.0403 Steps: 25550, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000012, Sample Num: 192, Cur Loss: 0.38418770, Cur Avg Loss: 1.54331731, Log Avg loss: 1.64219295, Global Avg Loss: 1.93180061, Time: 0.0402 Steps: 25560, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000022, Sample Num: 352, Cur Loss: 1.35069966, Cur Avg Loss: 1.61597061, Log Avg loss: 1.70315458, Global Avg Loss: 1.93171120, Time: 0.0403 Steps: 25570, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000032, Sample Num: 512, Cur Loss: 1.10728049, Cur Avg Loss: 1.59012405, Log Avg loss: 1.53326161, Global Avg Loss: 1.93155543, Time: 0.0403 Steps: 25580, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000042, Sample Num: 672, Cur Loss: 1.59160304, Cur Avg Loss: 1.57575851, Log Avg loss: 1.52978878, Global Avg Loss: 1.93139843, Time: 0.0404 Steps: 25590, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000052, Sample Num: 832, Cur Loss: 1.19515491, Cur Avg Loss: 1.56265432, Log Avg loss: 1.50761672, Global Avg Loss: 1.93123289, Time: 0.0403 Steps: 25600, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000062, Sample Num: 992, Cur Loss: 1.27216506, Cur Avg Loss: 1.55738700, Log Avg loss: 1.52999693, Global Avg Loss: 1.93107622, Time: 0.0403 Steps: 25610, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000072, Sample Num: 1152, Cur Loss: 1.61171246, Cur Avg Loss: 1.54546734, Log Avg loss: 1.47156549, Global Avg Loss: 1.93089686, Time: 0.0402 Steps: 25620, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000082, Sample Num: 1312, Cur Loss: 1.44048619, Cur Avg Loss: 1.51652994, Log Avg loss: 1.30818062, Global Avg Loss: 1.93065390, Time: 0.0402 Steps: 25630, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000092, Sample Num: 1472, Cur Loss: 1.25003409, Cur Avg Loss: 1.51469828, Log Avg loss: 1.49967873, Global Avg Loss: 1.93048581, Time: 0.0403 Steps: 25640, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000102, Sample Num: 1632, Cur Loss: 0.81170344, Cur Avg Loss: 1.52969531, Log Avg loss: 1.66766791, Global Avg Loss: 1.93038335, Time: 0.0402 Steps: 25650, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000112, Sample Num: 1792, Cur Loss: 0.63826722, Cur Avg Loss: 1.53425982, Log Avg loss: 1.58081784, Global Avg Loss: 1.93024712, Time: 0.0403 Steps: 25660, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000122, Sample Num: 1952, Cur Loss: 1.27765012, Cur Avg Loss: 1.50920591, Log Avg loss: 1.22860209, Global Avg Loss: 1.92997378, Time: 0.0402 Steps: 25670, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000132, Sample Num: 2112, Cur Loss: 2.53173733, Cur Avg Loss: 1.54054193, Log Avg loss: 1.92284144, Global Avg Loss: 1.92997101, Time: 0.0920 Steps: 25680, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000142, Sample Num: 2272, Cur Loss: 1.45929217, Cur Avg Loss: 1.53241145, Log Avg loss: 1.42508910, Global Avg Loss: 1.92977448, Time: 0.0404 Steps: 25690, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000152, Sample Num: 2432, Cur Loss: 2.24278188, Cur Avg Loss: 1.55082769, Log Avg loss: 1.81233832, Global Avg Loss: 1.92972878, Time: 0.0646 Steps: 25700, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000162, Sample Num: 2592, Cur Loss: 1.30998945, Cur Avg Loss: 1.53911037, Log Avg loss: 1.36100714, Global Avg Loss: 1.92950758, Time: 0.0634 Steps: 25710, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000172, Sample Num: 2752, Cur Loss: 1.13015568, Cur Avg Loss: 1.53205740, Log Avg loss: 1.41779929, Global Avg Loss: 1.92930862, Time: 0.0650 Steps: 25720, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000182, Sample Num: 2912, Cur Loss: 1.66399121, Cur Avg Loss: 1.53200418, Log Avg loss: 1.53108872, Global Avg Loss: 1.92915385, Time: 0.0634 Steps: 25730, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000192, Sample Num: 3072, Cur Loss: 1.64658403, Cur Avg Loss: 1.52907362, Log Avg loss: 1.47573746, Global Avg Loss: 1.92897770, Time: 0.0405 Steps: 25740, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000202, Sample Num: 3232, Cur Loss: 2.04111505, Cur Avg Loss: 1.51095238, Log Avg loss: 1.16302463, Global Avg Loss: 1.92868024, Time: 0.0794 Steps: 25750, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000212, Sample Num: 3392, Cur Loss: 1.31031549, Cur Avg Loss: 1.51814655, Log Avg loss: 1.66346868, Global Avg Loss: 1.92857729, Time: 0.0687 Steps: 25760, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000222, Sample Num: 3552, Cur Loss: 2.40024710, Cur Avg Loss: 1.51679902, Log Avg loss: 1.48823135, Global Avg Loss: 1.92840641, Time: 0.0534 Steps: 25770, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000232, Sample Num: 3712, Cur Loss: 0.93011737, Cur Avg Loss: 1.51754465, Log Avg loss: 1.53409778, Global Avg Loss: 1.92825346, Time: 0.0403 Steps: 25780, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000242, Sample Num: 3872, Cur Loss: 1.47398806, Cur Avg Loss: 1.53014106, Log Avg loss: 1.82237761, Global Avg Loss: 1.92821241, Time: 0.0403 Steps: 25790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000252, Sample Num: 4032, Cur Loss: 3.40309381, Cur Avg Loss: 1.53391440, Log Avg loss: 1.62522928, Global Avg Loss: 1.92809497, Time: 0.0402 Steps: 25800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000262, Sample Num: 4192, Cur Loss: 1.26814365, Cur Avg Loss: 1.54270731, Log Avg loss: 1.76428878, Global Avg Loss: 1.92803151, Time: 0.0402 Steps: 25810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000272, Sample Num: 4352, Cur Loss: 0.43946201, Cur Avg Loss: 1.53806574, Log Avg loss: 1.41645650, Global Avg Loss: 1.92783338, Time: 0.0402 Steps: 25820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000282, Sample Num: 4512, Cur Loss: 1.45884252, Cur Avg Loss: 1.54627042, Log Avg loss: 1.76943756, Global Avg Loss: 1.92777205, Time: 0.0402 Steps: 25830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000292, Sample Num: 4672, Cur Loss: 1.92565250, Cur Avg Loss: 1.55847551, Log Avg loss: 1.90265906, Global Avg Loss: 1.92776234, Time: 0.0403 Steps: 25840, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000302, Sample Num: 4832, Cur Loss: 3.88448668, Cur Avg Loss: 1.56110422, Log Avg loss: 1.63786260, Global Avg Loss: 1.92765019, Time: 0.0403 Steps: 25850, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000312, Sample Num: 4992, Cur Loss: 0.95033044, Cur Avg Loss: 1.56663269, Log Avg loss: 1.73359244, Global Avg Loss: 1.92757515, Time: 0.0402 Steps: 25860, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000322, Sample Num: 5152, Cur Loss: 1.07715154, Cur Avg Loss: 1.56410486, Log Avg loss: 1.48523685, Global Avg Loss: 1.92740416, Time: 0.0402 Steps: 25870, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000332, Sample Num: 5312, Cur Loss: 0.96436095, Cur Avg Loss: 1.56150465, Log Avg loss: 1.47777787, Global Avg Loss: 1.92723043, Time: 0.0402 Steps: 25880, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000342, Sample Num: 5472, Cur Loss: 0.51690805, Cur Avg Loss: 1.57067755, Log Avg loss: 1.87521767, Global Avg Loss: 1.92721034, Time: 0.0403 Steps: 25890, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000352, Sample Num: 5632, Cur Loss: 1.15008759, Cur Avg Loss: 1.55811301, Log Avg loss: 1.12840568, Global Avg Loss: 1.92690192, Time: 0.0403 Steps: 25900, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000362, Sample Num: 5792, Cur Loss: 1.17867899, Cur Avg Loss: 1.55724813, Log Avg loss: 1.52680436, Global Avg Loss: 1.92674750, Time: 0.0402 Steps: 25910, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000372, Sample Num: 5952, Cur Loss: 1.55766129, Cur Avg Loss: 1.55770822, Log Avg loss: 1.57436372, Global Avg Loss: 1.92661155, Time: 0.0402 Steps: 25920, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000382, Sample Num: 6112, Cur Loss: 1.85699534, Cur Avg Loss: 1.55570533, Log Avg loss: 1.48119783, Global Avg Loss: 1.92643977, Time: 0.0402 Steps: 25930, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000392, Sample Num: 6272, Cur Loss: 3.90223265, Cur Avg Loss: 1.55585169, Log Avg loss: 1.56144249, Global Avg Loss: 1.92629907, Time: 0.0402 Steps: 25940, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000402, Sample Num: 6432, Cur Loss: 0.78366977, Cur Avg Loss: 1.54764978, Log Avg loss: 1.22613482, Global Avg Loss: 1.92602925, Time: 0.0402 Steps: 25950, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000412, Sample Num: 6592, Cur Loss: 1.98798680, Cur Avg Loss: 1.53651081, Log Avg loss: 1.08872434, Global Avg Loss: 1.92570672, Time: 0.0403 Steps: 25960, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000422, Sample Num: 6752, Cur Loss: 1.06997371, Cur Avg Loss: 1.52864272, Log Avg loss: 1.20447724, Global Avg Loss: 1.92542900, Time: 0.0403 Steps: 25970, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000432, Sample Num: 6912, Cur Loss: 1.21359479, Cur Avg Loss: 1.52874484, Log Avg loss: 1.53305452, Global Avg Loss: 1.92527797, Time: 0.0402 Steps: 25980, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000442, Sample Num: 7072, Cur Loss: 2.35940099, Cur Avg Loss: 1.53762990, Log Avg loss: 1.92146446, Global Avg Loss: 1.92527650, Time: 0.0401 Steps: 25990, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000452, Sample Num: 7232, Cur Loss: 2.25573134, Cur Avg Loss: 1.53238293, Log Avg loss: 1.30046673, Global Avg Loss: 1.92503619, Time: 0.0403 Steps: 26000, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000462, Sample Num: 7392, Cur Loss: 1.28221226, Cur Avg Loss: 1.53860440, Log Avg loss: 1.81981479, Global Avg Loss: 1.92499574, Time: 0.0403 Steps: 26010, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000472, Sample Num: 7552, Cur Loss: 1.48405194, Cur Avg Loss: 1.53773710, Log Avg loss: 1.49766789, Global Avg Loss: 1.92483151, Time: 0.0403 Steps: 26020, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000482, Sample Num: 7712, Cur Loss: 1.92718244, Cur Avg Loss: 1.54219821, Log Avg loss: 1.75276260, Global Avg Loss: 1.92476540, Time: 0.0616 Steps: 26030, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000492, Sample Num: 7872, Cur Loss: 1.57560492, Cur Avg Loss: 1.54506688, Log Avg loss: 1.68333675, Global Avg Loss: 1.92467269, Time: 0.0711 Steps: 26040, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000502, Sample Num: 8032, Cur Loss: 1.98766696, Cur Avg Loss: 1.55211884, Log Avg loss: 1.89907539, Global Avg Loss: 1.92466286, Time: 0.0567 Steps: 26050, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000512, Sample Num: 8192, Cur Loss: 1.08213663, Cur Avg Loss: 1.55117395, Log Avg loss: 1.50374060, Global Avg Loss: 1.92450134, Time: 0.0406 Steps: 26060, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000522, Sample Num: 8352, Cur Loss: 0.56999922, Cur Avg Loss: 1.54214850, Log Avg loss: 1.08004519, Global Avg Loss: 1.92417742, Time: 0.0405 Steps: 26070, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000532, Sample Num: 8512, Cur Loss: 1.03512311, Cur Avg Loss: 1.53161435, Log Avg loss: 0.98173179, Global Avg Loss: 1.92381606, Time: 0.0412 Steps: 26080, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000542, Sample Num: 8672, Cur Loss: 1.70788288, Cur Avg Loss: 1.53817070, Log Avg loss: 1.88696836, Global Avg Loss: 1.92380193, Time: 0.0476 Steps: 26090, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000552, Sample Num: 8832, Cur Loss: 0.68201548, Cur Avg Loss: 1.53079742, Log Avg loss: 1.13116584, Global Avg Loss: 1.92349824, Time: 0.0662 Steps: 26100, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000562, Sample Num: 8992, Cur Loss: 1.39171362, Cur Avg Loss: 1.52954930, Log Avg loss: 1.46065319, Global Avg Loss: 1.92332097, Time: 0.0406 Steps: 26110, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000572, Sample Num: 9152, Cur Loss: 1.50842202, Cur Avg Loss: 1.52655271, Log Avg loss: 1.35814438, Global Avg Loss: 1.92310460, Time: 0.0403 Steps: 26120, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000582, Sample Num: 9312, Cur Loss: 1.62089849, Cur Avg Loss: 1.52447457, Log Avg loss: 1.40560492, Global Avg Loss: 1.92290655, Time: 0.0403 Steps: 26130, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000592, Sample Num: 9472, Cur Loss: 1.58062792, Cur Avg Loss: 1.52684140, Log Avg loss: 1.66459102, Global Avg Loss: 1.92280773, Time: 0.0402 Steps: 26140, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000602, Sample Num: 9632, Cur Loss: 3.53460002, Cur Avg Loss: 1.52739972, Log Avg loss: 1.56045213, Global Avg Loss: 1.92266916, Time: 0.0402 Steps: 26150, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000612, Sample Num: 9792, Cur Loss: 3.09132338, Cur Avg Loss: 1.53201402, Log Avg loss: 1.80979452, Global Avg Loss: 1.92262601, Time: 0.0402 Steps: 26160, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000622, Sample Num: 9952, Cur Loss: 0.69800854, Cur Avg Loss: 1.52978120, Log Avg loss: 1.39313288, Global Avg Loss: 1.92242368, Time: 0.0403 Steps: 26170, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000632, Sample Num: 10112, Cur Loss: 0.97503674, Cur Avg Loss: 1.52999405, Log Avg loss: 1.54323351, Global Avg Loss: 1.92227884, Time: 0.0402 Steps: 26180, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000642, Sample Num: 10272, Cur Loss: 1.77533543, Cur Avg Loss: 1.53137501, Log Avg loss: 1.61865122, Global Avg Loss: 1.92216291, Time: 0.0403 Steps: 26190, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000652, Sample Num: 10432, Cur Loss: 1.44422448, Cur Avg Loss: 1.53256081, Log Avg loss: 1.60868933, Global Avg Loss: 1.92204327, Time: 0.0403 Steps: 26200, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000662, Sample Num: 10592, Cur Loss: 0.89586830, Cur Avg Loss: 1.52850619, Log Avg loss: 1.26414478, Global Avg Loss: 1.92179225, Time: 0.0402 Steps: 26210, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000672, Sample Num: 10752, Cur Loss: 1.00940728, Cur Avg Loss: 1.52817720, Log Avg loss: 1.50639804, Global Avg Loss: 1.92163383, Time: 0.0402 Steps: 26220, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000682, Sample Num: 10912, Cur Loss: 2.31456995, Cur Avg Loss: 1.52699867, Log Avg loss: 1.44780144, Global Avg Loss: 1.92145318, Time: 0.0404 Steps: 26230, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000692, Sample Num: 11072, Cur Loss: 2.44647598, Cur Avg Loss: 1.52633005, Log Avg loss: 1.48073078, Global Avg Loss: 1.92128523, Time: 0.0403 Steps: 26240, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000702, Sample Num: 11232, Cur Loss: 2.01136971, Cur Avg Loss: 1.52581529, Log Avg loss: 1.49019333, Global Avg Loss: 1.92112100, Time: 0.0402 Steps: 26250, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000712, Sample Num: 11392, Cur Loss: 1.24494028, Cur Avg Loss: 1.52048408, Log Avg loss: 1.14623358, Global Avg Loss: 1.92082592, Time: 0.0403 Steps: 26260, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000722, Sample Num: 11552, Cur Loss: 1.54826736, Cur Avg Loss: 1.52034364, Log Avg loss: 1.51034441, Global Avg Loss: 1.92066966, Time: 0.0403 Steps: 26270, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000732, Sample Num: 11712, Cur Loss: 1.52389598, Cur Avg Loss: 1.51982036, Log Avg loss: 1.48203926, Global Avg Loss: 1.92050276, Time: 0.0403 Steps: 26280, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000742, Sample Num: 11872, Cur Loss: 0.55652225, Cur Avg Loss: 1.52108752, Log Avg loss: 1.61384358, Global Avg Loss: 1.92038611, Time: 0.0403 Steps: 26290, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000752, Sample Num: 12032, Cur Loss: 0.75381035, Cur Avg Loss: 1.52092369, Log Avg loss: 1.50876725, Global Avg Loss: 1.92022960, Time: 0.0406 Steps: 26300, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000762, Sample Num: 12192, Cur Loss: 1.30908644, Cur Avg Loss: 1.51882785, Log Avg loss: 1.36122088, Global Avg Loss: 1.92001713, Time: 0.0402 Steps: 26310, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000772, Sample Num: 12352, Cur Loss: 1.66752994, Cur Avg Loss: 1.51995027, Log Avg loss: 1.60547875, Global Avg Loss: 1.91989763, Time: 0.0403 Steps: 26320, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000782, Sample Num: 12512, Cur Loss: 1.91081798, Cur Avg Loss: 1.52380817, Log Avg loss: 1.82163816, Global Avg Loss: 1.91986031, Time: 0.0403 Steps: 26330, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000792, Sample Num: 12672, Cur Loss: 1.66849184, Cur Avg Loss: 1.52800700, Log Avg loss: 1.85635535, Global Avg Loss: 1.91983620, Time: 0.0403 Steps: 26340, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000802, Sample Num: 12832, Cur Loss: 1.37587702, Cur Avg Loss: 1.52837409, Log Avg loss: 1.55744747, Global Avg Loss: 1.91969867, Time: 0.0403 Steps: 26350, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000812, Sample Num: 12992, Cur Loss: 1.59179902, Cur Avg Loss: 1.52965740, Log Avg loss: 1.63257946, Global Avg Loss: 1.91958975, Time: 0.0403 Steps: 26360, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000822, Sample Num: 13152, Cur Loss: 0.89847589, Cur Avg Loss: 1.52734972, Log Avg loss: 1.33996596, Global Avg Loss: 1.91936994, Time: 0.0403 Steps: 26370, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000832, Sample Num: 13312, Cur Loss: 1.66791761, Cur Avg Loss: 1.53051038, Log Avg loss: 1.79031603, Global Avg Loss: 1.91932102, Time: 0.0844 Steps: 26380, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000842, Sample Num: 13472, Cur Loss: 2.33235168, Cur Avg Loss: 1.53081694, Log Avg loss: 1.55632352, Global Avg Loss: 1.91918347, Time: 0.0537 Steps: 26390, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000852, Sample Num: 13632, Cur Loss: 2.43981290, Cur Avg Loss: 1.53311229, Log Avg loss: 1.72638050, Global Avg Loss: 1.91911044, Time: 0.0404 Steps: 26400, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000862, Sample Num: 13792, Cur Loss: 1.72534966, Cur Avg Loss: 1.53958692, Log Avg loss: 2.09122519, Global Avg Loss: 1.91917561, Time: 0.0552 Steps: 26410, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000872, Sample Num: 13952, Cur Loss: 2.86322403, Cur Avg Loss: 1.54044152, Log Avg loss: 1.61410770, Global Avg Loss: 1.91906014, Time: 0.0665 Steps: 26420, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000882, Sample Num: 14112, Cur Loss: 1.47284842, Cur Avg Loss: 1.54375353, Log Avg loss: 1.83256108, Global Avg Loss: 1.91902741, Time: 0.0771 Steps: 26430, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000892, Sample Num: 14272, Cur Loss: 1.75392032, Cur Avg Loss: 1.54304375, Log Avg loss: 1.48044140, Global Avg Loss: 1.91886153, Time: 0.0926 Steps: 26440, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000902, Sample Num: 14432, Cur Loss: 1.34671330, Cur Avg Loss: 1.54267030, Log Avg loss: 1.50935888, Global Avg Loss: 1.91870671, Time: 0.0406 Steps: 26450, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000912, Sample Num: 14592, Cur Loss: 1.24170876, Cur Avg Loss: 1.54222587, Log Avg loss: 1.50213795, Global Avg Loss: 1.91854928, Time: 0.0404 Steps: 26460, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000922, Sample Num: 14752, Cur Loss: 1.31708026, Cur Avg Loss: 1.54313751, Log Avg loss: 1.62627892, Global Avg Loss: 1.91843886, Time: 0.0404 Steps: 26470, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000932, Sample Num: 14912, Cur Loss: 1.07290864, Cur Avg Loss: 1.54309306, Log Avg loss: 1.53899503, Global Avg Loss: 1.91829557, Time: 0.0404 Steps: 26480, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000942, Sample Num: 15072, Cur Loss: 1.32332098, Cur Avg Loss: 1.54367857, Log Avg loss: 1.59824754, Global Avg Loss: 1.91817475, Time: 0.0403 Steps: 26490, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000952, Sample Num: 15232, Cur Loss: 3.09404516, Cur Avg Loss: 1.54654313, Log Avg loss: 1.81638500, Global Avg Loss: 1.91813634, Time: 0.0402 Steps: 26500, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000962, Sample Num: 15392, Cur Loss: 1.41491342, Cur Avg Loss: 1.54815672, Log Avg loss: 1.70177074, Global Avg Loss: 1.91805472, Time: 0.0403 Steps: 26510, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000972, Sample Num: 15552, Cur Loss: 1.77439976, Cur Avg Loss: 1.54666866, Log Avg loss: 1.40351717, Global Avg Loss: 1.91786070, Time: 0.0403 Steps: 26520, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000982, Sample Num: 15712, Cur Loss: 1.14223111, Cur Avg Loss: 1.54267116, Log Avg loss: 1.15411376, Global Avg Loss: 1.91757282, Time: 0.0403 Steps: 26530, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000992, Sample Num: 15872, Cur Loss: 1.27593064, Cur Avg Loss: 1.54380675, Log Avg loss: 1.65532200, Global Avg Loss: 1.91747401, Time: 0.0403 Steps: 26540, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001002, Sample Num: 16032, Cur Loss: 2.34251666, Cur Avg Loss: 1.54214697, Log Avg loss: 1.37749653, Global Avg Loss: 1.91727063, Time: 0.0403 Steps: 26550, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001012, Sample Num: 16192, Cur Loss: 1.93834996, Cur Avg Loss: 1.54490032, Log Avg loss: 1.82078584, Global Avg Loss: 1.91723430, Time: 0.0402 Steps: 26560, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001022, Sample Num: 16352, Cur Loss: 1.42649186, Cur Avg Loss: 1.54430275, Log Avg loss: 1.48382854, Global Avg Loss: 1.91707118, Time: 0.0403 Steps: 26570, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001032, Sample Num: 16512, Cur Loss: 3.43118715, Cur Avg Loss: 1.54540043, Log Avg loss: 1.65758424, Global Avg Loss: 1.91697356, Time: 0.0402 Steps: 26580, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001042, Sample Num: 16672, Cur Loss: 1.26098549, Cur Avg Loss: 1.54239940, Log Avg loss: 1.23269226, Global Avg Loss: 1.91671621, Time: 0.0403 Steps: 26590, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001052, Sample Num: 16832, Cur Loss: 1.29220223, Cur Avg Loss: 1.54225849, Log Avg loss: 1.52757650, Global Avg Loss: 1.91656992, Time: 0.0403 Steps: 26600, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001062, Sample Num: 16992, Cur Loss: 1.05668616, Cur Avg Loss: 1.53975577, Log Avg loss: 1.27646921, Global Avg Loss: 1.91632937, Time: 0.0402 Steps: 26610, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001072, Sample Num: 17152, Cur Loss: 3.01315308, Cur Avg Loss: 1.53858311, Log Avg loss: 1.41404677, Global Avg Loss: 1.91614068, Time: 0.0403 Steps: 26620, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001082, Sample Num: 17312, Cur Loss: 2.03421307, Cur Avg Loss: 1.54121892, Log Avg loss: 1.82377766, Global Avg Loss: 1.91610600, Time: 0.0402 Steps: 26630, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001092, Sample Num: 17472, Cur Loss: 0.77434707, Cur Avg Loss: 1.54096208, Log Avg loss: 1.51317164, Global Avg Loss: 1.91595475, Time: 0.0403 Steps: 26640, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001102, Sample Num: 17632, Cur Loss: 1.19446743, Cur Avg Loss: 1.53903310, Log Avg loss: 1.32838830, Global Avg Loss: 1.91573427, Time: 0.0402 Steps: 26650, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001112, Sample Num: 17792, Cur Loss: 2.23425722, Cur Avg Loss: 1.54141202, Log Avg loss: 1.80356941, Global Avg Loss: 1.91569220, Time: 0.0402 Steps: 26660, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001122, Sample Num: 17952, Cur Loss: 1.47317564, Cur Avg Loss: 1.54183252, Log Avg loss: 1.58859191, Global Avg Loss: 1.91556955, Time: 0.0402 Steps: 26670, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001132, Sample Num: 18112, Cur Loss: 0.87339890, Cur Avg Loss: 1.53844118, Log Avg loss: 1.15793264, Global Avg Loss: 1.91528558, Time: 0.0402 Steps: 26680, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001142, Sample Num: 18272, Cur Loss: 1.69854867, Cur Avg Loss: 1.53813040, Log Avg loss: 1.50295066, Global Avg Loss: 1.91513109, Time: 0.0402 Steps: 26690, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001152, Sample Num: 18432, Cur Loss: 1.01068592, Cur Avg Loss: 1.53763802, Log Avg loss: 1.48140868, Global Avg Loss: 1.91496865, Time: 0.0402 Steps: 26700, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001162, Sample Num: 18592, Cur Loss: 1.09910202, Cur Avg Loss: 1.53765459, Log Avg loss: 1.53956280, Global Avg Loss: 1.91482810, Time: 0.0402 Steps: 26710, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001172, Sample Num: 18752, Cur Loss: 1.57091236, Cur Avg Loss: 1.53854971, Log Avg loss: 1.64256300, Global Avg Loss: 1.91472620, Time: 0.0402 Steps: 26720, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001182, Sample Num: 18912, Cur Loss: 3.06154966, Cur Avg Loss: 1.53676960, Log Avg loss: 1.32813991, Global Avg Loss: 1.91450676, Time: 0.0480 Steps: 26730, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001192, Sample Num: 19072, Cur Loss: 0.76135969, Cur Avg Loss: 1.53815010, Log Avg loss: 1.70132607, Global Avg Loss: 1.91442703, Time: 0.0405 Steps: 26740, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001202, Sample Num: 19232, Cur Loss: 1.52569199, Cur Avg Loss: 1.53956886, Log Avg loss: 1.70868419, Global Avg Loss: 1.91435012, Time: 0.0486 Steps: 26750, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001212, Sample Num: 19392, Cur Loss: 1.12080777, Cur Avg Loss: 1.53877838, Log Avg loss: 1.44376374, Global Avg Loss: 1.91417426, Time: 0.0799 Steps: 26760, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001222, Sample Num: 19552, Cur Loss: 1.82315040, Cur Avg Loss: 1.53940253, Log Avg loss: 1.61504950, Global Avg Loss: 1.91406253, Time: 0.0463 Steps: 26770, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001232, Sample Num: 19712, Cur Loss: 3.09584093, Cur Avg Loss: 1.53841972, Log Avg loss: 1.41832025, Global Avg Loss: 1.91387741, Time: 0.0455 Steps: 26780, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001242, Sample Num: 19872, Cur Loss: 1.79394007, Cur Avg Loss: 1.53907799, Log Avg loss: 1.62017653, Global Avg Loss: 1.91376778, Time: 0.0406 Steps: 26790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001252, Sample Num: 20032, Cur Loss: 2.22056389, Cur Avg Loss: 1.53855694, Log Avg loss: 1.47384241, Global Avg Loss: 1.91360363, Time: 0.0521 Steps: 26800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001262, Sample Num: 20192, Cur Loss: 2.52396560, Cur Avg Loss: 1.54088382, Log Avg loss: 1.83220919, Global Avg Loss: 1.91357327, Time: 0.0405 Steps: 26810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001272, Sample Num: 20352, Cur Loss: 1.36258268, Cur Avg Loss: 1.54279167, Log Avg loss: 1.78356260, Global Avg Loss: 1.91352479, Time: 0.0403 Steps: 26820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001282, Sample Num: 20512, Cur Loss: 0.71208441, Cur Avg Loss: 1.54198184, Log Avg loss: 1.43897188, Global Avg Loss: 1.91334792, Time: 0.0403 Steps: 26830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001292, Sample Num: 20672, Cur Loss: 2.43982172, Cur Avg Loss: 1.54478583, Log Avg loss: 1.90425624, Global Avg Loss: 1.91334453, Time: 0.0403 Steps: 26840, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001302, Sample Num: 20832, Cur Loss: 1.60254216, Cur Avg Loss: 1.54264111, Log Avg loss: 1.26554435, Global Avg Loss: 1.91310326, Time: 0.0402 Steps: 26850, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001312, Sample Num: 20992, Cur Loss: 1.50177622, Cur Avg Loss: 1.54375178, Log Avg loss: 1.68836079, Global Avg Loss: 1.91301959, Time: 0.0403 Steps: 26860, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001322, Sample Num: 21152, Cur Loss: 2.27205229, Cur Avg Loss: 1.54512529, Log Avg loss: 1.72532985, Global Avg Loss: 1.91294974, Time: 0.0403 Steps: 26870, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001332, Sample Num: 21312, Cur Loss: 1.10916281, Cur Avg Loss: 1.54416945, Log Avg loss: 1.41780763, Global Avg Loss: 1.91276554, Time: 0.0403 Steps: 26880, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001342, Sample Num: 21472, Cur Loss: 1.95014286, Cur Avg Loss: 1.54369714, Log Avg loss: 1.48078420, Global Avg Loss: 1.91260489, Time: 0.0403 Steps: 26890, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001352, Sample Num: 21632, Cur Loss: 1.33335114, Cur Avg Loss: 1.54650850, Log Avg loss: 1.92379390, Global Avg Loss: 1.91260905, Time: 0.0403 Steps: 26900, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001362, Sample Num: 21792, Cur Loss: 1.14425433, Cur Avg Loss: 1.54413910, Log Avg loss: 1.22379550, Global Avg Loss: 1.91235308, Time: 0.0403 Steps: 26910, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001372, Sample Num: 21952, Cur Loss: 1.54789340, Cur Avg Loss: 1.54300078, Log Avg loss: 1.38796252, Global Avg Loss: 1.91215828, Time: 0.0403 Steps: 26920, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001382, Sample Num: 22112, Cur Loss: 1.78255022, Cur Avg Loss: 1.54307888, Log Avg loss: 1.55379322, Global Avg Loss: 1.91202521, Time: 0.0402 Steps: 26930, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001392, Sample Num: 22272, Cur Loss: 2.04510522, Cur Avg Loss: 1.54317395, Log Avg loss: 1.55631350, Global Avg Loss: 1.91189317, Time: 0.0402 Steps: 26940, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001402, Sample Num: 22432, Cur Loss: 1.04496932, Cur Avg Loss: 1.54213185, Log Avg loss: 1.39707129, Global Avg Loss: 1.91170214, Time: 0.0403 Steps: 26950, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001412, Sample Num: 22592, Cur Loss: 1.10488915, Cur Avg Loss: 1.54095711, Log Avg loss: 1.37625898, Global Avg Loss: 1.91150354, Time: 0.0404 Steps: 26960, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001422, Sample Num: 22752, Cur Loss: 0.38088578, Cur Avg Loss: 1.53869743, Log Avg loss: 1.21963069, Global Avg Loss: 1.91124700, Time: 0.0405 Steps: 26970, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001432, Sample Num: 22912, Cur Loss: 1.08257544, Cur Avg Loss: 1.53821949, Log Avg loss: 1.47025601, Global Avg Loss: 1.91108355, Time: 0.0403 Steps: 26980, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001442, Sample Num: 23072, Cur Loss: 2.29164314, Cur Avg Loss: 1.53909343, Log Avg loss: 1.66424081, Global Avg Loss: 1.91099209, Time: 0.0403 Steps: 26990, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001452, Sample Num: 23232, Cur Loss: 3.17322779, Cur Avg Loss: 1.53973397, Log Avg loss: 1.63210088, Global Avg Loss: 1.91088880, Time: 0.0403 Steps: 27000, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001462, Sample Num: 23392, Cur Loss: 0.73710001, Cur Avg Loss: 1.53882878, Log Avg loss: 1.40739421, Global Avg Loss: 1.91070239, Time: 0.0403 Steps: 27010, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001472, Sample Num: 23552, Cur Loss: 0.74873960, Cur Avg Loss: 1.53785754, Log Avg loss: 1.39586312, Global Avg Loss: 1.91051185, Time: 0.0403 Steps: 27020, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001482, Sample Num: 23712, Cur Loss: 2.28055286, Cur Avg Loss: 1.53644221, Log Avg loss: 1.32810596, Global Avg Loss: 1.91029638, Time: 0.0403 Steps: 27030, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001492, Sample Num: 23872, Cur Loss: 1.86393785, Cur Avg Loss: 1.53483597, Log Avg loss: 1.29679116, Global Avg Loss: 1.91006950, Time: 0.0403 Steps: 27040, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001502, Sample Num: 24032, Cur Loss: 0.92575532, Cur Avg Loss: 1.53393549, Log Avg loss: 1.39958347, Global Avg Loss: 1.90988078, Time: 0.0403 Steps: 27050, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001512, Sample Num: 24192, Cur Loss: 2.43549609, Cur Avg Loss: 1.53426336, Log Avg loss: 1.58350880, Global Avg Loss: 1.90976017, Time: 0.0403 Steps: 27060, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001522, Sample Num: 24352, Cur Loss: 1.91940022, Cur Avg Loss: 1.53513339, Log Avg loss: 1.66668280, Global Avg Loss: 1.90967037, Time: 0.0567 Steps: 27070, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001532, Sample Num: 24512, Cur Loss: 2.89757490, Cur Avg Loss: 1.53860394, Log Avg loss: 2.06682087, Global Avg Loss: 1.90972840, Time: 0.0529 Steps: 27080, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001542, Sample Num: 24672, Cur Loss: 1.12763464, Cur Avg Loss: 1.53847110, Log Avg loss: 1.51812030, Global Avg Loss: 1.90958384, Time: 0.0511 Steps: 27090, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001552, Sample Num: 24832, Cur Loss: 0.68352318, Cur Avg Loss: 1.53768989, Log Avg loss: 1.41722685, Global Avg Loss: 1.90940216, Time: 0.0588 Steps: 27100, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001562, Sample Num: 24992, Cur Loss: 1.27732217, Cur Avg Loss: 1.53829219, Log Avg loss: 1.63176980, Global Avg Loss: 1.90929975, Time: 0.0797 Steps: 27110, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001572, Sample Num: 25152, Cur Loss: 1.90258002, Cur Avg Loss: 1.53828521, Log Avg loss: 1.53719435, Global Avg Loss: 1.90916255, Time: 0.0696 Steps: 27120, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001582, Sample Num: 25312, Cur Loss: 1.94181705, Cur Avg Loss: 1.53869727, Log Avg loss: 1.60347325, Global Avg Loss: 1.90904987, Time: 0.0781 Steps: 27130, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001592, Sample Num: 25472, Cur Loss: 0.64226902, Cur Avg Loss: 1.53907088, Log Avg loss: 1.59817669, Global Avg Loss: 1.90893533, Time: 0.0594 Steps: 27140, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001602, Sample Num: 25632, Cur Loss: 1.56464338, Cur Avg Loss: 1.54094019, Log Avg loss: 1.83853443, Global Avg Loss: 1.90890940, Time: 0.0516 Steps: 27150, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001612, Sample Num: 25792, Cur Loss: 1.13073325, Cur Avg Loss: 1.53936053, Log Avg loss: 1.28629782, Global Avg Loss: 1.90868016, Time: 0.0404 Steps: 27160, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001622, Sample Num: 25952, Cur Loss: 1.87780440, Cur Avg Loss: 1.54112743, Log Avg loss: 1.82595302, Global Avg Loss: 1.90864971, Time: 0.0403 Steps: 27170, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001632, Sample Num: 26112, Cur Loss: 2.16777086, Cur Avg Loss: 1.54050129, Log Avg loss: 1.43894007, Global Avg Loss: 1.90847689, Time: 0.0403 Steps: 27180, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001642, Sample Num: 26272, Cur Loss: 1.88628936, Cur Avg Loss: 1.54149958, Log Avg loss: 1.70442185, Global Avg Loss: 1.90840185, Time: 0.0403 Steps: 27190, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001652, Sample Num: 26432, Cur Loss: 1.21556437, Cur Avg Loss: 1.54141075, Log Avg loss: 1.52682501, Global Avg Loss: 1.90826156, Time: 0.0403 Steps: 27200, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001662, Sample Num: 26592, Cur Loss: 1.31082726, Cur Avg Loss: 1.54133198, Log Avg loss: 1.52831787, Global Avg Loss: 1.90812193, Time: 0.0403 Steps: 27210, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001672, Sample Num: 26752, Cur Loss: 1.22881269, Cur Avg Loss: 1.54231391, Log Avg loss: 1.70551075, Global Avg Loss: 1.90804749, Time: 0.0404 Steps: 27220, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001682, Sample Num: 26912, Cur Loss: 0.40215391, Cur Avg Loss: 1.54014566, Log Avg loss: 1.17761486, Global Avg Loss: 1.90777925, Time: 0.0404 Steps: 27230, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001692, Sample Num: 27072, Cur Loss: 1.99535978, Cur Avg Loss: 1.54174803, Log Avg loss: 1.81126643, Global Avg Loss: 1.90774382, Time: 0.0404 Steps: 27240, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001702, Sample Num: 27232, Cur Loss: 2.25196314, Cur Avg Loss: 1.54187135, Log Avg loss: 1.56273662, Global Avg Loss: 1.90761721, Time: 0.0402 Steps: 27250, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001712, Sample Num: 27392, Cur Loss: 1.23693979, Cur Avg Loss: 1.54175517, Log Avg loss: 1.52198232, Global Avg Loss: 1.90747574, Time: 0.0404 Steps: 27260, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001722, Sample Num: 27552, Cur Loss: 0.20548910, Cur Avg Loss: 1.53910408, Log Avg loss: 1.08523626, Global Avg Loss: 1.90717423, Time: 0.0404 Steps: 27270, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001732, Sample Num: 27712, Cur Loss: 1.40159106, Cur Avg Loss: 1.53810253, Log Avg loss: 1.36563613, Global Avg Loss: 1.90697571, Time: 0.0403 Steps: 27280, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001742, Sample Num: 27872, Cur Loss: 1.49971652, Cur Avg Loss: 1.54019176, Log Avg loss: 1.90204663, Global Avg Loss: 1.90697391, Time: 0.0403 Steps: 27290, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001752, Sample Num: 28032, Cur Loss: 1.56116903, Cur Avg Loss: 1.54022354, Log Avg loss: 1.54575952, Global Avg Loss: 1.90684160, Time: 0.0404 Steps: 27300, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001762, Sample Num: 28192, Cur Loss: 1.59447479, Cur Avg Loss: 1.54188267, Log Avg loss: 1.83256195, Global Avg Loss: 1.90681440, Time: 0.0404 Steps: 27310, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001772, Sample Num: 28352, Cur Loss: 0.87555391, Cur Avg Loss: 1.54103703, Log Avg loss: 1.39203615, Global Avg Loss: 1.90662597, Time: 0.0404 Steps: 27320, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001782, Sample Num: 28512, Cur Loss: 1.36018288, Cur Avg Loss: 1.54142553, Log Avg loss: 1.61026775, Global Avg Loss: 1.90651753, Time: 0.0403 Steps: 27330, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001792, Sample Num: 28672, Cur Loss: 0.81827116, Cur Avg Loss: 1.53992435, Log Avg loss: 1.27241383, Global Avg Loss: 1.90628560, Time: 0.0405 Steps: 27340, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001802, Sample Num: 28832, Cur Loss: 1.21520376, Cur Avg Loss: 1.53974488, Log Avg loss: 1.50758424, Global Avg Loss: 1.90613982, Time: 0.0403 Steps: 27350, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001812, Sample Num: 28992, Cur Loss: 1.51143956, Cur Avg Loss: 1.54030156, Log Avg loss: 1.64061478, Global Avg Loss: 1.90604278, Time: 0.0403 Steps: 27360, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001822, Sample Num: 29152, Cur Loss: 0.65649629, Cur Avg Loss: 1.53956596, Log Avg loss: 1.40627455, Global Avg Loss: 1.90586018, Time: 0.0403 Steps: 27370, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001832, Sample Num: 29312, Cur Loss: 0.94941628, Cur Avg Loss: 1.53883391, Log Avg loss: 1.40545452, Global Avg Loss: 1.90567742, Time: 0.0403 Steps: 27380, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001842, Sample Num: 29472, Cur Loss: 2.04923463, Cur Avg Loss: 1.54009399, Log Avg loss: 1.77094159, Global Avg Loss: 1.90562822, Time: 0.0402 Steps: 27390, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001852, Sample Num: 29632, Cur Loss: 0.91975772, Cur Avg Loss: 1.53920389, Log Avg loss: 1.37524756, Global Avg Loss: 1.90543465, Time: 0.0403 Steps: 27400, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001862, Sample Num: 29792, Cur Loss: 1.86350191, Cur Avg Loss: 1.54014344, Log Avg loss: 1.71414802, Global Avg Loss: 1.90536487, Time: 0.0403 Steps: 27410, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001872, Sample Num: 29952, Cur Loss: 1.35542655, Cur Avg Loss: 1.54049143, Log Avg loss: 1.60528703, Global Avg Loss: 1.90525543, Time: 0.0682 Steps: 27420, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001882, Sample Num: 30112, Cur Loss: 1.90949464, Cur Avg Loss: 1.54249021, Log Avg loss: 1.91666130, Global Avg Loss: 1.90525959, Time: 0.0404 Steps: 27430, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001892, Sample Num: 30272, Cur Loss: 1.68952608, Cur Avg Loss: 1.54350763, Log Avg loss: 1.73498564, Global Avg Loss: 1.90519753, Time: 0.0404 Steps: 27440, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001902, Sample Num: 30432, Cur Loss: 2.35106802, Cur Avg Loss: 1.54437069, Log Avg loss: 1.70766314, Global Avg Loss: 1.90512557, Time: 0.0807 Steps: 27450, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001912, Sample Num: 30592, Cur Loss: 1.73946691, Cur Avg Loss: 1.54504350, Log Avg loss: 1.67301183, Global Avg Loss: 1.90504104, Time: 0.0556 Steps: 27460, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001922, Sample Num: 30752, Cur Loss: 2.06033301, Cur Avg Loss: 1.54650263, Log Avg loss: 1.82548841, Global Avg Loss: 1.90501208, Time: 0.0655 Steps: 27470, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001932, Sample Num: 30912, Cur Loss: 0.96166086, Cur Avg Loss: 1.54515555, Log Avg loss: 1.28624646, Global Avg Loss: 1.90478692, Time: 0.0576 Steps: 27480, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001942, Sample Num: 31072, Cur Loss: 1.53650331, Cur Avg Loss: 1.54582096, Log Avg loss: 1.67437732, Global Avg Loss: 1.90470310, Time: 0.0713 Steps: 27490, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001952, Sample Num: 31232, Cur Loss: 0.54160416, Cur Avg Loss: 1.54581778, Log Avg loss: 1.54520039, Global Avg Loss: 1.90457237, Time: 0.0525 Steps: 27500, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001962, Sample Num: 31392, Cur Loss: 2.11772394, Cur Avg Loss: 1.54830921, Log Avg loss: 2.03463599, Global Avg Loss: 1.90461965, Time: 0.0432 Steps: 27510, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001972, Sample Num: 31552, Cur Loss: 0.78585404, Cur Avg Loss: 1.54651633, Log Avg loss: 1.19475318, Global Avg Loss: 1.90436170, Time: 0.0404 Steps: 27520, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001982, Sample Num: 31712, Cur Loss: 1.45888209, Cur Avg Loss: 1.54735347, Log Avg loss: 1.71243792, Global Avg Loss: 1.90429199, Time: 0.0403 Steps: 27530, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001992, Sample Num: 31872, Cur Loss: 1.30881059, Cur Avg Loss: 1.54671156, Log Avg loss: 1.41948574, Global Avg Loss: 1.90411595, Time: 0.0403 Steps: 27540, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002002, Sample Num: 32032, Cur Loss: 2.06410718, Cur Avg Loss: 1.54525196, Log Avg loss: 1.25449998, Global Avg Loss: 1.90388016, Time: 0.0403 Steps: 27550, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002012, Sample Num: 32192, Cur Loss: 1.54698253, Cur Avg Loss: 1.54519318, Log Avg loss: 1.53342365, Global Avg Loss: 1.90374574, Time: 0.0403 Steps: 27560, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002022, Sample Num: 32352, Cur Loss: 2.17837381, Cur Avg Loss: 1.54523492, Log Avg loss: 1.55363355, Global Avg Loss: 1.90361875, Time: 0.0404 Steps: 27570, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002032, Sample Num: 32512, Cur Loss: 2.91103482, Cur Avg Loss: 1.54591257, Log Avg loss: 1.68293288, Global Avg Loss: 1.90353873, Time: 0.0403 Steps: 27580, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002042, Sample Num: 32672, Cur Loss: 0.91114211, Cur Avg Loss: 1.54428709, Log Avg loss: 1.21399112, Global Avg Loss: 1.90328881, Time: 0.0404 Steps: 27590, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002052, Sample Num: 32832, Cur Loss: 2.24991679, Cur Avg Loss: 1.54316333, Log Avg loss: 1.31369002, Global Avg Loss: 1.90307518, Time: 0.0403 Steps: 27600, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002062, Sample Num: 32992, Cur Loss: 1.24277055, Cur Avg Loss: 1.54275480, Log Avg loss: 1.45892451, Global Avg Loss: 1.90291432, Time: 0.0403 Steps: 27610, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002072, Sample Num: 33152, Cur Loss: 1.75211453, Cur Avg Loss: 1.54204297, Log Avg loss: 1.39526520, Global Avg Loss: 1.90273052, Time: 0.0403 Steps: 27620, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002082, Sample Num: 33312, Cur Loss: 2.55069351, Cur Avg Loss: 1.54228849, Log Avg loss: 1.59315853, Global Avg Loss: 1.90261848, Time: 0.0403 Steps: 27630, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002092, Sample Num: 33472, Cur Loss: 0.89437157, Cur Avg Loss: 1.54102606, Log Avg loss: 1.27818874, Global Avg Loss: 1.90239256, Time: 0.0403 Steps: 27640, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002102, Sample Num: 33632, Cur Loss: 0.99170840, Cur Avg Loss: 1.54156153, Log Avg loss: 1.65358225, Global Avg Loss: 1.90230258, Time: 0.0403 Steps: 27650, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002112, Sample Num: 33792, Cur Loss: 1.14324832, Cur Avg Loss: 1.54140991, Log Avg loss: 1.50953909, Global Avg Loss: 1.90216058, Time: 0.0403 Steps: 27660, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002122, Sample Num: 33952, Cur Loss: 1.03136253, Cur Avg Loss: 1.54144214, Log Avg loss: 1.54824851, Global Avg Loss: 1.90203268, Time: 0.0404 Steps: 27670, Updated lr: 0.000075 ***** Running evaluation checkpoint-27677 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-27677 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.727008, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.532771, "eval_total_loss": 1077.538056, "eval_mae": 1.111255, "eval_mse": 1.532618, "eval_r2": 0.025767, "eval_sp_statistic": 0.314327, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.375381, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.417345, "test_total_loss": 711.507297, "test_mae": 1.017216, "test_mse": 1.417582, "test_r2": 0.08508, "test_sp_statistic": 0.327445, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.383204, "test_ps_pvalue": 0.0, "lr": 7.470175438596491e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.9019593804155113, "train_cur_epoch_loss": 3282.225863739848, "train_cur_epoch_avg_loss": 1.5416749007702433, "train_cur_epoch_time": 95.72700834274292, "train_cur_epoch_avg_time": 0.044963366999879247, "epoch": 13, "step": 27677} ################################################## Training, Epoch: 0014, Batch: 000003, Sample Num: 48, Cur Loss: 0.64620471, Cur Avg Loss: 1.22405505, Log Avg loss: 1.49578151, Global Avg Loss: 1.90188591, Time: 0.0403 Steps: 27680, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000013, Sample Num: 208, Cur Loss: 2.40876341, Cur Avg Loss: 1.57232854, Log Avg loss: 1.67681058, Global Avg Loss: 1.90180462, Time: 0.0403 Steps: 27690, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000023, Sample Num: 368, Cur Loss: 1.39964581, Cur Avg Loss: 1.37407693, Log Avg loss: 1.11634984, Global Avg Loss: 1.90152107, Time: 0.0403 Steps: 27700, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000033, Sample Num: 528, Cur Loss: 0.47948045, Cur Avg Loss: 1.38318035, Log Avg loss: 1.40411822, Global Avg Loss: 1.90134156, Time: 0.0403 Steps: 27710, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000043, Sample Num: 688, Cur Loss: 0.93920648, Cur Avg Loss: 1.46837796, Log Avg loss: 1.74953005, Global Avg Loss: 1.90128680, Time: 0.0403 Steps: 27720, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000053, Sample Num: 848, Cur Loss: 1.55951476, Cur Avg Loss: 1.48332705, Log Avg loss: 1.54760816, Global Avg Loss: 1.90115925, Time: 0.0403 Steps: 27730, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000063, Sample Num: 1008, Cur Loss: 1.50387776, Cur Avg Loss: 1.55617042, Log Avg loss: 1.94224027, Global Avg Loss: 1.90117406, Time: 0.0403 Steps: 27740, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000073, Sample Num: 1168, Cur Loss: 2.17566657, Cur Avg Loss: 1.63611827, Log Avg loss: 2.13978974, Global Avg Loss: 1.90126005, Time: 0.0403 Steps: 27750, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000083, Sample Num: 1328, Cur Loss: 3.01694441, Cur Avg Loss: 1.63885757, Log Avg loss: 1.65885441, Global Avg Loss: 1.90117273, Time: 0.0403 Steps: 27760, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000093, Sample Num: 1488, Cur Loss: 3.36635685, Cur Avg Loss: 1.65569871, Log Avg loss: 1.79548023, Global Avg Loss: 1.90113467, Time: 0.0403 Steps: 27770, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000103, Sample Num: 1648, Cur Loss: 1.19049573, Cur Avg Loss: 1.63674466, Log Avg loss: 1.46047201, Global Avg Loss: 1.90097604, Time: 0.0403 Steps: 27780, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000113, Sample Num: 1808, Cur Loss: 0.75820887, Cur Avg Loss: 1.62310551, Log Avg loss: 1.48262226, Global Avg Loss: 1.90082550, Time: 0.0403 Steps: 27790, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000123, Sample Num: 1968, Cur Loss: 0.51314259, Cur Avg Loss: 1.57960962, Log Avg loss: 1.08810605, Global Avg Loss: 1.90053316, Time: 0.0638 Steps: 27800, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000133, Sample Num: 2128, Cur Loss: 0.88670725, Cur Avg Loss: 1.56079528, Log Avg loss: 1.32937893, Global Avg Loss: 1.90032778, Time: 0.0696 Steps: 27810, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000143, Sample Num: 2288, Cur Loss: 1.09524107, Cur Avg Loss: 1.54911983, Log Avg loss: 1.39383636, Global Avg Loss: 1.90014572, Time: 0.0405 Steps: 27820, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000153, Sample Num: 2448, Cur Loss: 0.75548375, Cur Avg Loss: 1.54400020, Log Avg loss: 1.47078949, Global Avg Loss: 1.89999144, Time: 0.0478 Steps: 27830, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000163, Sample Num: 2608, Cur Loss: 1.61860812, Cur Avg Loss: 1.55310456, Log Avg loss: 1.69240126, Global Avg Loss: 1.89991688, Time: 0.0656 Steps: 27840, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000173, Sample Num: 2768, Cur Loss: 1.62527418, Cur Avg Loss: 1.55151459, Log Avg loss: 1.52559797, Global Avg Loss: 1.89978247, Time: 0.0460 Steps: 27850, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000183, Sample Num: 2928, Cur Loss: 0.67330980, Cur Avg Loss: 1.53866448, Log Avg loss: 1.31635765, Global Avg Loss: 1.89957306, Time: 0.0435 Steps: 27860, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000193, Sample Num: 3088, Cur Loss: 0.95129853, Cur Avg Loss: 1.52534451, Log Avg loss: 1.28158901, Global Avg Loss: 1.89935132, Time: 0.0648 Steps: 27870, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000203, Sample Num: 3248, Cur Loss: 1.13877738, Cur Avg Loss: 1.50809336, Log Avg loss: 1.17514622, Global Avg Loss: 1.89909156, Time: 0.0741 Steps: 27880, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000213, Sample Num: 3408, Cur Loss: 0.74900794, Cur Avg Loss: 1.51093707, Log Avg loss: 1.56866444, Global Avg Loss: 1.89897309, Time: 0.0404 Steps: 27890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000223, Sample Num: 3568, Cur Loss: 0.90464181, Cur Avg Loss: 1.50746763, Log Avg loss: 1.43356842, Global Avg Loss: 1.89880627, Time: 0.0403 Steps: 27900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000233, Sample Num: 3728, Cur Loss: 2.62840295, Cur Avg Loss: 1.51408096, Log Avg loss: 1.66155840, Global Avg Loss: 1.89872127, Time: 0.0403 Steps: 27910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000243, Sample Num: 3888, Cur Loss: 1.16717649, Cur Avg Loss: 1.51684853, Log Avg loss: 1.58133277, Global Avg Loss: 1.89860759, Time: 0.0404 Steps: 27920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000253, Sample Num: 4048, Cur Loss: 1.44878137, Cur Avg Loss: 1.51156935, Log Avg loss: 1.38328530, Global Avg Loss: 1.89842309, Time: 0.0403 Steps: 27930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000263, Sample Num: 4208, Cur Loss: 2.17146587, Cur Avg Loss: 1.51140835, Log Avg loss: 1.50733498, Global Avg Loss: 1.89828311, Time: 0.0403 Steps: 27940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000273, Sample Num: 4368, Cur Loss: 1.89353812, Cur Avg Loss: 1.51690264, Log Avg loss: 1.66140264, Global Avg Loss: 1.89819836, Time: 0.0403 Steps: 27950, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000283, Sample Num: 4528, Cur Loss: 0.93600881, Cur Avg Loss: 1.51678554, Log Avg loss: 1.51358855, Global Avg Loss: 1.89806080, Time: 0.0402 Steps: 27960, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000293, Sample Num: 4688, Cur Loss: 2.22656727, Cur Avg Loss: 1.52295375, Log Avg loss: 1.69751405, Global Avg Loss: 1.89798910, Time: 0.0402 Steps: 27970, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000303, Sample Num: 4848, Cur Loss: 1.41227233, Cur Avg Loss: 1.53254764, Log Avg loss: 1.81364883, Global Avg Loss: 1.89795896, Time: 0.0402 Steps: 27980, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000313, Sample Num: 5008, Cur Loss: 1.78000045, Cur Avg Loss: 1.52881781, Log Avg loss: 1.41580378, Global Avg Loss: 1.89778670, Time: 0.0403 Steps: 27990, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000323, Sample Num: 5168, Cur Loss: 1.56551194, Cur Avg Loss: 1.53517083, Log Avg loss: 1.73402034, Global Avg Loss: 1.89772821, Time: 0.0403 Steps: 28000, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000333, Sample Num: 5328, Cur Loss: 2.67614412, Cur Avg Loss: 1.53427618, Log Avg loss: 1.50537890, Global Avg Loss: 1.89758814, Time: 0.0403 Steps: 28010, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000343, Sample Num: 5488, Cur Loss: 2.04788327, Cur Avg Loss: 1.53632601, Log Avg loss: 1.60458556, Global Avg Loss: 1.89748357, Time: 0.0403 Steps: 28020, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000353, Sample Num: 5648, Cur Loss: 1.00371587, Cur Avg Loss: 1.53939834, Log Avg loss: 1.64477913, Global Avg Loss: 1.89739341, Time: 0.0403 Steps: 28030, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000363, Sample Num: 5808, Cur Loss: 0.94114858, Cur Avg Loss: 1.53851839, Log Avg loss: 1.50745611, Global Avg Loss: 1.89725435, Time: 0.0402 Steps: 28040, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000373, Sample Num: 5968, Cur Loss: 2.72361469, Cur Avg Loss: 1.53635094, Log Avg loss: 1.45767274, Global Avg Loss: 1.89709764, Time: 0.0403 Steps: 28050, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000383, Sample Num: 6128, Cur Loss: 1.84614289, Cur Avg Loss: 1.53031518, Log Avg loss: 1.30518114, Global Avg Loss: 1.89688669, Time: 0.0403 Steps: 28060, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000393, Sample Num: 6288, Cur Loss: 1.27053273, Cur Avg Loss: 1.53334946, Log Avg loss: 1.64956248, Global Avg Loss: 1.89679858, Time: 0.0403 Steps: 28070, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000403, Sample Num: 6448, Cur Loss: 1.36424899, Cur Avg Loss: 1.53344586, Log Avg loss: 1.53723447, Global Avg Loss: 1.89667053, Time: 0.0402 Steps: 28080, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000413, Sample Num: 6608, Cur Loss: 1.84201801, Cur Avg Loss: 1.52655359, Log Avg loss: 1.24879481, Global Avg Loss: 1.89643989, Time: 0.0403 Steps: 28090, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000423, Sample Num: 6768, Cur Loss: 1.32948554, Cur Avg Loss: 1.52526228, Log Avg loss: 1.47193140, Global Avg Loss: 1.89628882, Time: 0.0403 Steps: 28100, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000433, Sample Num: 6928, Cur Loss: 1.00980234, Cur Avg Loss: 1.52358588, Log Avg loss: 1.45267419, Global Avg Loss: 1.89613100, Time: 0.0403 Steps: 28110, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000443, Sample Num: 7088, Cur Loss: 1.61663032, Cur Avg Loss: 1.52246473, Log Avg loss: 1.47391893, Global Avg Loss: 1.89598086, Time: 0.0403 Steps: 28120, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000453, Sample Num: 7248, Cur Loss: 1.21164644, Cur Avg Loss: 1.51777593, Log Avg loss: 1.31006221, Global Avg Loss: 1.89577257, Time: 0.0402 Steps: 28130, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000463, Sample Num: 7408, Cur Loss: 1.99782777, Cur Avg Loss: 1.52679723, Log Avg loss: 1.93546188, Global Avg Loss: 1.89578667, Time: 0.0402 Steps: 28140, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000473, Sample Num: 7568, Cur Loss: 1.89856648, Cur Avg Loss: 1.53136457, Log Avg loss: 1.74283230, Global Avg Loss: 1.89573233, Time: 0.0570 Steps: 28150, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000483, Sample Num: 7728, Cur Loss: 2.25448990, Cur Avg Loss: 1.53956003, Log Avg loss: 1.92720533, Global Avg Loss: 1.89574351, Time: 0.0561 Steps: 28160, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000493, Sample Num: 7888, Cur Loss: 1.44015718, Cur Avg Loss: 1.53527617, Log Avg loss: 1.32836577, Global Avg Loss: 1.89554210, Time: 0.0627 Steps: 28170, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000503, Sample Num: 8048, Cur Loss: 1.21796727, Cur Avg Loss: 1.53227546, Log Avg loss: 1.38434060, Global Avg Loss: 1.89536069, Time: 0.0407 Steps: 28180, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000513, Sample Num: 8208, Cur Loss: 0.77370203, Cur Avg Loss: 1.52738636, Log Avg loss: 1.28146466, Global Avg Loss: 1.89514292, Time: 0.0450 Steps: 28190, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000523, Sample Num: 8368, Cur Loss: 1.01993537, Cur Avg Loss: 1.52722990, Log Avg loss: 1.51920354, Global Avg Loss: 1.89500961, Time: 0.0481 Steps: 28200, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000533, Sample Num: 8528, Cur Loss: 0.57278150, Cur Avg Loss: 1.52230610, Log Avg loss: 1.26479138, Global Avg Loss: 1.89478621, Time: 0.0819 Steps: 28210, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000543, Sample Num: 8688, Cur Loss: 0.47197229, Cur Avg Loss: 1.51933848, Log Avg loss: 1.36116429, Global Avg Loss: 1.89459711, Time: 0.0489 Steps: 28220, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000553, Sample Num: 8848, Cur Loss: 0.93072933, Cur Avg Loss: 1.51776364, Log Avg loss: 1.43224990, Global Avg Loss: 1.89443334, Time: 0.0548 Steps: 28230, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000563, Sample Num: 9008, Cur Loss: 0.93587333, Cur Avg Loss: 1.51973393, Log Avg loss: 1.62869053, Global Avg Loss: 1.89433923, Time: 0.0403 Steps: 28240, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000573, Sample Num: 9168, Cur Loss: 1.03279352, Cur Avg Loss: 1.52058187, Log Avg loss: 1.56832103, Global Avg Loss: 1.89422383, Time: 0.0404 Steps: 28250, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000583, Sample Num: 9328, Cur Loss: 2.71545601, Cur Avg Loss: 1.51847953, Log Avg loss: 1.39801546, Global Avg Loss: 1.89404824, Time: 0.0403 Steps: 28260, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000593, Sample Num: 9488, Cur Loss: 1.81085801, Cur Avg Loss: 1.51572645, Log Avg loss: 1.35522227, Global Avg Loss: 1.89385764, Time: 0.0403 Steps: 28270, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000603, Sample Num: 9648, Cur Loss: 1.78007424, Cur Avg Loss: 1.51327711, Log Avg loss: 1.36803120, Global Avg Loss: 1.89367171, Time: 0.0403 Steps: 28280, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000613, Sample Num: 9808, Cur Loss: 1.60540915, Cur Avg Loss: 1.51305846, Log Avg loss: 1.49987345, Global Avg Loss: 1.89353251, Time: 0.0403 Steps: 28290, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000623, Sample Num: 9968, Cur Loss: 0.95366824, Cur Avg Loss: 1.50755877, Log Avg loss: 1.17042800, Global Avg Loss: 1.89327699, Time: 0.0403 Steps: 28300, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000633, Sample Num: 10128, Cur Loss: 1.54941022, Cur Avg Loss: 1.50458578, Log Avg loss: 1.31936819, Global Avg Loss: 1.89307427, Time: 0.0403 Steps: 28310, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000643, Sample Num: 10288, Cur Loss: 1.55111647, Cur Avg Loss: 1.50117328, Log Avg loss: 1.28516254, Global Avg Loss: 1.89285961, Time: 0.0403 Steps: 28320, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000653, Sample Num: 10448, Cur Loss: 1.71913922, Cur Avg Loss: 1.50240721, Log Avg loss: 1.58174861, Global Avg Loss: 1.89274979, Time: 0.0403 Steps: 28330, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000663, Sample Num: 10608, Cur Loss: 1.33048165, Cur Avg Loss: 1.50358981, Log Avg loss: 1.58081393, Global Avg Loss: 1.89263973, Time: 0.0403 Steps: 28340, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000673, Sample Num: 10768, Cur Loss: 0.68024468, Cur Avg Loss: 1.50571894, Log Avg loss: 1.64687990, Global Avg Loss: 1.89255304, Time: 0.0403 Steps: 28350, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000683, Sample Num: 10928, Cur Loss: 2.11039972, Cur Avg Loss: 1.50422040, Log Avg loss: 1.40336875, Global Avg Loss: 1.89238055, Time: 0.0403 Steps: 28360, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000693, Sample Num: 11088, Cur Loss: 1.15077424, Cur Avg Loss: 1.50022593, Log Avg loss: 1.22740375, Global Avg Loss: 1.89214615, Time: 0.0403 Steps: 28370, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000703, Sample Num: 11248, Cur Loss: 1.66736972, Cur Avg Loss: 1.49417565, Log Avg loss: 1.07489085, Global Avg Loss: 1.89185818, Time: 0.0403 Steps: 28380, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000713, Sample Num: 11408, Cur Loss: 2.23325872, Cur Avg Loss: 1.49486654, Log Avg loss: 1.54343618, Global Avg Loss: 1.89173546, Time: 0.0404 Steps: 28390, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000723, Sample Num: 11568, Cur Loss: 1.38507271, Cur Avg Loss: 1.49261005, Log Avg loss: 1.33172215, Global Avg Loss: 1.89153827, Time: 0.0402 Steps: 28400, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000733, Sample Num: 11728, Cur Loss: 1.31342793, Cur Avg Loss: 1.48754712, Log Avg loss: 1.12149763, Global Avg Loss: 1.89126722, Time: 0.0403 Steps: 28410, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000743, Sample Num: 11888, Cur Loss: 0.66109234, Cur Avg Loss: 1.48670466, Log Avg loss: 1.42495269, Global Avg Loss: 1.89110314, Time: 0.0403 Steps: 28420, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000753, Sample Num: 12048, Cur Loss: 1.14836669, Cur Avg Loss: 1.48667324, Log Avg loss: 1.48433867, Global Avg Loss: 1.89096007, Time: 0.0402 Steps: 28430, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000763, Sample Num: 12208, Cur Loss: 1.40935898, Cur Avg Loss: 1.48634018, Log Avg loss: 1.46126029, Global Avg Loss: 1.89080898, Time: 0.0402 Steps: 28440, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000773, Sample Num: 12368, Cur Loss: 1.10801160, Cur Avg Loss: 1.48357707, Log Avg loss: 1.27275218, Global Avg Loss: 1.89059173, Time: 0.0402 Steps: 28450, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000783, Sample Num: 12528, Cur Loss: 1.81239021, Cur Avg Loss: 1.48372023, Log Avg loss: 1.49478659, Global Avg Loss: 1.89045266, Time: 0.0402 Steps: 28460, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000793, Sample Num: 12688, Cur Loss: 1.06963587, Cur Avg Loss: 1.48244113, Log Avg loss: 1.38228765, Global Avg Loss: 1.89027417, Time: 0.0403 Steps: 28470, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000803, Sample Num: 12848, Cur Loss: 1.29018807, Cur Avg Loss: 1.48098025, Log Avg loss: 1.36513228, Global Avg Loss: 1.89008978, Time: 0.0403 Steps: 28480, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000813, Sample Num: 13008, Cur Loss: 1.15397942, Cur Avg Loss: 1.48010116, Log Avg loss: 1.40951025, Global Avg Loss: 1.88992110, Time: 0.0404 Steps: 28490, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000823, Sample Num: 13168, Cur Loss: 1.33834994, Cur Avg Loss: 1.48099611, Log Avg loss: 1.55375566, Global Avg Loss: 1.88980314, Time: 0.0407 Steps: 28500, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000833, Sample Num: 13328, Cur Loss: 1.67059314, Cur Avg Loss: 1.48008729, Log Avg loss: 1.40529147, Global Avg Loss: 1.88963320, Time: 0.0491 Steps: 28510, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000843, Sample Num: 13488, Cur Loss: 1.24155283, Cur Avg Loss: 1.48084374, Log Avg loss: 1.54385580, Global Avg Loss: 1.88951196, Time: 0.0433 Steps: 28520, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000853, Sample Num: 13648, Cur Loss: 1.96190810, Cur Avg Loss: 1.48258297, Log Avg loss: 1.62919977, Global Avg Loss: 1.88942072, Time: 0.0700 Steps: 28530, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000863, Sample Num: 13808, Cur Loss: 0.65956104, Cur Avg Loss: 1.48535168, Log Avg loss: 1.72152293, Global Avg Loss: 1.88936189, Time: 0.0407 Steps: 28540, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000873, Sample Num: 13968, Cur Loss: 1.73565173, Cur Avg Loss: 1.48683806, Log Avg loss: 1.61511281, Global Avg Loss: 1.88926583, Time: 0.0586 Steps: 28550, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000883, Sample Num: 14128, Cur Loss: 1.29610896, Cur Avg Loss: 1.48670351, Log Avg loss: 1.47495711, Global Avg Loss: 1.88912076, Time: 0.0432 Steps: 28560, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000893, Sample Num: 14288, Cur Loss: 1.77618015, Cur Avg Loss: 1.48678058, Log Avg loss: 1.49358553, Global Avg Loss: 1.88898232, Time: 0.0608 Steps: 28570, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000903, Sample Num: 14448, Cur Loss: 0.68195534, Cur Avg Loss: 1.48772784, Log Avg loss: 1.57231821, Global Avg Loss: 1.88887152, Time: 0.0404 Steps: 28580, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000913, Sample Num: 14608, Cur Loss: 1.87859249, Cur Avg Loss: 1.48950123, Log Avg loss: 1.64963877, Global Avg Loss: 1.88878784, Time: 0.0680 Steps: 28590, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000923, Sample Num: 14768, Cur Loss: 1.70983231, Cur Avg Loss: 1.48848735, Log Avg loss: 1.39591973, Global Avg Loss: 1.88861551, Time: 0.0403 Steps: 28600, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000933, Sample Num: 14928, Cur Loss: 1.46261573, Cur Avg Loss: 1.48839782, Log Avg loss: 1.48013433, Global Avg Loss: 1.88847273, Time: 0.0404 Steps: 28610, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000943, Sample Num: 15088, Cur Loss: 2.18409395, Cur Avg Loss: 1.49209590, Log Avg loss: 1.83712705, Global Avg Loss: 1.88845479, Time: 0.0403 Steps: 28620, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000953, Sample Num: 15248, Cur Loss: 1.31044805, Cur Avg Loss: 1.49188377, Log Avg loss: 1.47187998, Global Avg Loss: 1.88830929, Time: 0.0403 Steps: 28630, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000963, Sample Num: 15408, Cur Loss: 2.62544775, Cur Avg Loss: 1.49301245, Log Avg loss: 1.60057478, Global Avg Loss: 1.88820883, Time: 0.0404 Steps: 28640, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000973, Sample Num: 15568, Cur Loss: 1.74833775, Cur Avg Loss: 1.49547721, Log Avg loss: 1.73283424, Global Avg Loss: 1.88815459, Time: 0.0403 Steps: 28650, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000983, Sample Num: 15728, Cur Loss: 0.86056513, Cur Avg Loss: 1.49459930, Log Avg loss: 1.40917811, Global Avg Loss: 1.88798747, Time: 0.0403 Steps: 28660, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000993, Sample Num: 15888, Cur Loss: 0.96078110, Cur Avg Loss: 1.49289655, Log Avg loss: 1.32551660, Global Avg Loss: 1.88779128, Time: 0.0403 Steps: 28670, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001003, Sample Num: 16048, Cur Loss: 3.52182555, Cur Avg Loss: 1.49248659, Log Avg loss: 1.45177717, Global Avg Loss: 1.88763925, Time: 0.0403 Steps: 28680, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001013, Sample Num: 16208, Cur Loss: 1.61610258, Cur Avg Loss: 1.49226212, Log Avg loss: 1.46974862, Global Avg Loss: 1.88749360, Time: 0.0403 Steps: 28690, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001023, Sample Num: 16368, Cur Loss: 2.68083382, Cur Avg Loss: 1.49314962, Log Avg loss: 1.58305269, Global Avg Loss: 1.88738752, Time: 0.0403 Steps: 28700, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001033, Sample Num: 16528, Cur Loss: 2.48891592, Cur Avg Loss: 1.49155037, Log Avg loss: 1.32794687, Global Avg Loss: 1.88719266, Time: 0.0403 Steps: 28710, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001043, Sample Num: 16688, Cur Loss: 1.77689004, Cur Avg Loss: 1.48903602, Log Avg loss: 1.22930456, Global Avg Loss: 1.88696359, Time: 0.0403 Steps: 28720, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001053, Sample Num: 16848, Cur Loss: 1.10997641, Cur Avg Loss: 1.48858055, Log Avg loss: 1.44107455, Global Avg Loss: 1.88680839, Time: 0.0403 Steps: 28730, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001063, Sample Num: 17008, Cur Loss: 2.05972600, Cur Avg Loss: 1.48705065, Log Avg loss: 1.32595203, Global Avg Loss: 1.88661324, Time: 0.0404 Steps: 28740, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001073, Sample Num: 17168, Cur Loss: 2.57426453, Cur Avg Loss: 1.48899564, Log Avg loss: 1.69574789, Global Avg Loss: 1.88654686, Time: 0.0403 Steps: 28750, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001083, Sample Num: 17328, Cur Loss: 0.79035449, Cur Avg Loss: 1.48771777, Log Avg loss: 1.35060276, Global Avg Loss: 1.88636050, Time: 0.0403 Steps: 28760, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001093, Sample Num: 17488, Cur Loss: 1.30641520, Cur Avg Loss: 1.48905824, Log Avg loss: 1.63423076, Global Avg Loss: 1.88627287, Time: 0.0404 Steps: 28770, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001103, Sample Num: 17648, Cur Loss: 0.62188089, Cur Avg Loss: 1.48666595, Log Avg loss: 1.22518902, Global Avg Loss: 1.88604317, Time: 0.0404 Steps: 28780, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001113, Sample Num: 17808, Cur Loss: 1.12453985, Cur Avg Loss: 1.48699718, Log Avg loss: 1.52353135, Global Avg Loss: 1.88591725, Time: 0.0403 Steps: 28790, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001123, Sample Num: 17968, Cur Loss: 0.64744520, Cur Avg Loss: 1.48493468, Log Avg loss: 1.25537914, Global Avg Loss: 1.88569831, Time: 0.0403 Steps: 28800, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001133, Sample Num: 18128, Cur Loss: 1.19802272, Cur Avg Loss: 1.48512875, Log Avg loss: 1.50692228, Global Avg Loss: 1.88556684, Time: 0.0403 Steps: 28810, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001143, Sample Num: 18288, Cur Loss: 0.89212245, Cur Avg Loss: 1.48771265, Log Avg loss: 1.78046901, Global Avg Loss: 1.88553037, Time: 0.0403 Steps: 28820, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001153, Sample Num: 18448, Cur Loss: 1.26340270, Cur Avg Loss: 1.48807127, Log Avg loss: 1.52906094, Global Avg Loss: 1.88540673, Time: 0.0403 Steps: 28830, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001163, Sample Num: 18608, Cur Loss: 0.96742153, Cur Avg Loss: 1.48931470, Log Avg loss: 1.63268313, Global Avg Loss: 1.88531910, Time: 0.0405 Steps: 28840, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001173, Sample Num: 18768, Cur Loss: 1.49743938, Cur Avg Loss: 1.48497729, Log Avg loss: 0.98053595, Global Avg Loss: 1.88500548, Time: 0.0719 Steps: 28850, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001183, Sample Num: 18928, Cur Loss: 0.79434693, Cur Avg Loss: 1.48489244, Log Avg loss: 1.47494012, Global Avg Loss: 1.88486339, Time: 0.0622 Steps: 28860, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001193, Sample Num: 19088, Cur Loss: 1.00447655, Cur Avg Loss: 1.48529462, Log Avg loss: 1.53287163, Global Avg Loss: 1.88474147, Time: 0.0643 Steps: 28870, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001203, Sample Num: 19248, Cur Loss: 1.15799403, Cur Avg Loss: 1.48736309, Log Avg loss: 1.73413128, Global Avg Loss: 1.88468932, Time: 0.0407 Steps: 28880, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001213, Sample Num: 19408, Cur Loss: 0.88443804, Cur Avg Loss: 1.48638294, Log Avg loss: 1.36847194, Global Avg Loss: 1.88451064, Time: 0.0847 Steps: 28890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001223, Sample Num: 19568, Cur Loss: 1.00727677, Cur Avg Loss: 1.48640253, Log Avg loss: 1.48877892, Global Avg Loss: 1.88437370, Time: 0.0678 Steps: 28900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001233, Sample Num: 19728, Cur Loss: 0.97012532, Cur Avg Loss: 1.48539950, Log Avg loss: 1.36272861, Global Avg Loss: 1.88419327, Time: 0.0671 Steps: 28910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001243, Sample Num: 19888, Cur Loss: 1.05840492, Cur Avg Loss: 1.48684193, Log Avg loss: 1.66469331, Global Avg Loss: 1.88411737, Time: 0.0712 Steps: 28920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001253, Sample Num: 20048, Cur Loss: 1.18722081, Cur Avg Loss: 1.48697102, Log Avg loss: 1.50301689, Global Avg Loss: 1.88398564, Time: 0.0687 Steps: 28930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001263, Sample Num: 20208, Cur Loss: 3.00617576, Cur Avg Loss: 1.48969448, Log Avg loss: 1.83094401, Global Avg Loss: 1.88396731, Time: 0.0403 Steps: 28940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001273, Sample Num: 20368, Cur Loss: 3.16322470, Cur Avg Loss: 1.48753584, Log Avg loss: 1.21489977, Global Avg Loss: 1.88373620, Time: 0.0404 Steps: 28950, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001283, Sample Num: 20528, Cur Loss: 1.12681055, Cur Avg Loss: 1.49314080, Log Avg loss: 2.20665183, Global Avg Loss: 1.88384770, Time: 0.0403 Steps: 28960, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001293, Sample Num: 20688, Cur Loss: 1.53256261, Cur Avg Loss: 1.49436100, Log Avg loss: 1.65091325, Global Avg Loss: 1.88376730, Time: 0.0403 Steps: 28970, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001303, Sample Num: 20848, Cur Loss: 2.20671988, Cur Avg Loss: 1.49540616, Log Avg loss: 1.63054510, Global Avg Loss: 1.88367992, Time: 0.0404 Steps: 28980, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001313, Sample Num: 21008, Cur Loss: 1.34717989, Cur Avg Loss: 1.49682990, Log Avg loss: 1.68234293, Global Avg Loss: 1.88361047, Time: 0.0404 Steps: 28990, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001323, Sample Num: 21168, Cur Loss: 0.45026526, Cur Avg Loss: 1.49487740, Log Avg loss: 1.23851431, Global Avg Loss: 1.88338802, Time: 0.0403 Steps: 29000, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001333, Sample Num: 21328, Cur Loss: 1.69742084, Cur Avg Loss: 1.49533780, Log Avg loss: 1.55624872, Global Avg Loss: 1.88327525, Time: 0.0403 Steps: 29010, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001343, Sample Num: 21488, Cur Loss: 1.87660015, Cur Avg Loss: 1.49541158, Log Avg loss: 1.50524655, Global Avg Loss: 1.88314499, Time: 0.0404 Steps: 29020, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001353, Sample Num: 21648, Cur Loss: 1.69531763, Cur Avg Loss: 1.49961020, Log Avg loss: 2.06348430, Global Avg Loss: 1.88320711, Time: 0.0403 Steps: 29030, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001363, Sample Num: 21808, Cur Loss: 0.94249624, Cur Avg Loss: 1.50152827, Log Avg loss: 1.76104418, Global Avg Loss: 1.88316504, Time: 0.0404 Steps: 29040, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001373, Sample Num: 21968, Cur Loss: 1.62380552, Cur Avg Loss: 1.50034413, Log Avg loss: 1.33894538, Global Avg Loss: 1.88297770, Time: 0.0404 Steps: 29050, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001383, Sample Num: 22128, Cur Loss: 2.57054090, Cur Avg Loss: 1.50001977, Log Avg loss: 1.45548503, Global Avg Loss: 1.88283060, Time: 0.0403 Steps: 29060, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001393, Sample Num: 22288, Cur Loss: 1.82088482, Cur Avg Loss: 1.50101627, Log Avg loss: 1.63883204, Global Avg Loss: 1.88274666, Time: 0.0403 Steps: 29070, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001403, Sample Num: 22448, Cur Loss: 1.12671506, Cur Avg Loss: 1.50069010, Log Avg loss: 1.45525495, Global Avg Loss: 1.88259966, Time: 0.0403 Steps: 29080, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001413, Sample Num: 22608, Cur Loss: 0.42529303, Cur Avg Loss: 1.50124218, Log Avg loss: 1.57869844, Global Avg Loss: 1.88249519, Time: 0.0404 Steps: 29090, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001423, Sample Num: 22768, Cur Loss: 0.63238806, Cur Avg Loss: 1.49983930, Log Avg loss: 1.30161288, Global Avg Loss: 1.88229557, Time: 0.0403 Steps: 29100, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001433, Sample Num: 22928, Cur Loss: 0.63156021, Cur Avg Loss: 1.49742395, Log Avg loss: 1.15371896, Global Avg Loss: 1.88204529, Time: 0.0403 Steps: 29110, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001443, Sample Num: 23088, Cur Loss: 1.40360940, Cur Avg Loss: 1.49807740, Log Avg loss: 1.59171766, Global Avg Loss: 1.88194559, Time: 0.0403 Steps: 29120, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001453, Sample Num: 23248, Cur Loss: 1.25087476, Cur Avg Loss: 1.49919073, Log Avg loss: 1.65984434, Global Avg Loss: 1.88186934, Time: 0.0403 Steps: 29130, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001463, Sample Num: 23408, Cur Loss: 0.80673075, Cur Avg Loss: 1.49966591, Log Avg loss: 1.56870970, Global Avg Loss: 1.88176187, Time: 0.0402 Steps: 29140, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001473, Sample Num: 23568, Cur Loss: 2.91125298, Cur Avg Loss: 1.49934236, Log Avg loss: 1.45200686, Global Avg Loss: 1.88161445, Time: 0.0403 Steps: 29150, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001483, Sample Num: 23728, Cur Loss: 2.13459706, Cur Avg Loss: 1.49852044, Log Avg loss: 1.37745151, Global Avg Loss: 1.88144155, Time: 0.0403 Steps: 29160, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001493, Sample Num: 23888, Cur Loss: 2.16397691, Cur Avg Loss: 1.49867045, Log Avg loss: 1.52091650, Global Avg Loss: 1.88131796, Time: 0.0402 Steps: 29170, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001503, Sample Num: 24048, Cur Loss: 2.48756599, Cur Avg Loss: 1.49875706, Log Avg loss: 1.51168766, Global Avg Loss: 1.88119128, Time: 0.0403 Steps: 29180, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001513, Sample Num: 24208, Cur Loss: 1.47029305, Cur Avg Loss: 1.49976620, Log Avg loss: 1.65144083, Global Avg Loss: 1.88111257, Time: 0.0403 Steps: 29190, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001523, Sample Num: 24368, Cur Loss: 0.88908690, Cur Avg Loss: 1.49796596, Log Avg loss: 1.22558886, Global Avg Loss: 1.88088808, Time: 0.0620 Steps: 29200, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001533, Sample Num: 24528, Cur Loss: 0.70855159, Cur Avg Loss: 1.49650219, Log Avg loss: 1.27357106, Global Avg Loss: 1.88068017, Time: 0.0406 Steps: 29210, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001543, Sample Num: 24688, Cur Loss: 0.50399315, Cur Avg Loss: 1.49442480, Log Avg loss: 1.17596014, Global Avg Loss: 1.88043899, Time: 0.0404 Steps: 29220, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001553, Sample Num: 24848, Cur Loss: 0.81532961, Cur Avg Loss: 1.49242431, Log Avg loss: 1.18374934, Global Avg Loss: 1.88020064, Time: 0.0559 Steps: 29230, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001563, Sample Num: 25008, Cur Loss: 2.72771597, Cur Avg Loss: 1.49356019, Log Avg loss: 1.66996123, Global Avg Loss: 1.88012874, Time: 0.0404 Steps: 29240, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001573, Sample Num: 25168, Cur Loss: 1.47481632, Cur Avg Loss: 1.49525603, Log Avg loss: 1.76031685, Global Avg Loss: 1.88008778, Time: 0.0406 Steps: 29250, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001583, Sample Num: 25328, Cur Loss: 2.99837828, Cur Avg Loss: 1.49659894, Log Avg loss: 1.70783759, Global Avg Loss: 1.88002891, Time: 0.0439 Steps: 29260, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001593, Sample Num: 25488, Cur Loss: 1.73786223, Cur Avg Loss: 1.49622649, Log Avg loss: 1.43726878, Global Avg Loss: 1.87987764, Time: 0.0456 Steps: 29270, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001603, Sample Num: 25648, Cur Loss: 0.45427179, Cur Avg Loss: 1.49770969, Log Avg loss: 1.73398366, Global Avg Loss: 1.87982781, Time: 0.0410 Steps: 29280, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001613, Sample Num: 25808, Cur Loss: 1.38614988, Cur Avg Loss: 1.49669593, Log Avg loss: 1.33418878, Global Avg Loss: 1.87964153, Time: 0.0404 Steps: 29290, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001623, Sample Num: 25968, Cur Loss: 1.57818329, Cur Avg Loss: 1.49658139, Log Avg loss: 1.47810697, Global Avg Loss: 1.87950448, Time: 0.0404 Steps: 29300, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001633, Sample Num: 26128, Cur Loss: 0.95960176, Cur Avg Loss: 1.49476420, Log Avg loss: 1.19983363, Global Avg Loss: 1.87927259, Time: 0.0404 Steps: 29310, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001643, Sample Num: 26288, Cur Loss: 1.84941483, Cur Avg Loss: 1.49567354, Log Avg loss: 1.64416862, Global Avg Loss: 1.87919241, Time: 0.0403 Steps: 29320, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001653, Sample Num: 26448, Cur Loss: 2.50772452, Cur Avg Loss: 1.49564614, Log Avg loss: 1.49114411, Global Avg Loss: 1.87906010, Time: 0.0403 Steps: 29330, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001663, Sample Num: 26608, Cur Loss: 1.16210532, Cur Avg Loss: 1.49392721, Log Avg loss: 1.20978941, Global Avg Loss: 1.87883199, Time: 0.0403 Steps: 29340, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001673, Sample Num: 26768, Cur Loss: 1.13704956, Cur Avg Loss: 1.49439570, Log Avg loss: 1.57230440, Global Avg Loss: 1.87872756, Time: 0.0403 Steps: 29350, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001683, Sample Num: 26928, Cur Loss: 3.33808422, Cur Avg Loss: 1.49567415, Log Avg loss: 1.70955890, Global Avg Loss: 1.87866994, Time: 0.0403 Steps: 29360, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001693, Sample Num: 27088, Cur Loss: 1.61275482, Cur Avg Loss: 1.49612839, Log Avg loss: 1.57257812, Global Avg Loss: 1.87856572, Time: 0.0403 Steps: 29370, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001703, Sample Num: 27248, Cur Loss: 2.13126945, Cur Avg Loss: 1.49608633, Log Avg loss: 1.48896546, Global Avg Loss: 1.87843311, Time: 0.0402 Steps: 29380, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001713, Sample Num: 27408, Cur Loss: 1.39809585, Cur Avg Loss: 1.49558591, Log Avg loss: 1.41036432, Global Avg Loss: 1.87827385, Time: 0.0403 Steps: 29390, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001723, Sample Num: 27568, Cur Loss: 2.26347113, Cur Avg Loss: 1.49671147, Log Avg loss: 1.68951961, Global Avg Loss: 1.87820965, Time: 0.0404 Steps: 29400, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001733, Sample Num: 27728, Cur Loss: 1.77793813, Cur Avg Loss: 1.49930164, Log Avg loss: 1.94558793, Global Avg Loss: 1.87823256, Time: 0.0403 Steps: 29410, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001743, Sample Num: 27888, Cur Loss: 1.58166337, Cur Avg Loss: 1.49985356, Log Avg loss: 1.59550074, Global Avg Loss: 1.87813646, Time: 0.0403 Steps: 29420, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001753, Sample Num: 28048, Cur Loss: 0.64041245, Cur Avg Loss: 1.49958071, Log Avg loss: 1.45202354, Global Avg Loss: 1.87799167, Time: 0.0403 Steps: 29430, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001763, Sample Num: 28208, Cur Loss: 1.16917646, Cur Avg Loss: 1.49734672, Log Avg loss: 1.10572838, Global Avg Loss: 1.87772935, Time: 0.0403 Steps: 29440, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001773, Sample Num: 28368, Cur Loss: 1.90438509, Cur Avg Loss: 1.49699725, Log Avg loss: 1.43538575, Global Avg Loss: 1.87757915, Time: 0.0403 Steps: 29450, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001783, Sample Num: 28528, Cur Loss: 1.86358011, Cur Avg Loss: 1.49639822, Log Avg loss: 1.39018984, Global Avg Loss: 1.87741371, Time: 0.0403 Steps: 29460, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001793, Sample Num: 28688, Cur Loss: 2.72193813, Cur Avg Loss: 1.49677691, Log Avg loss: 1.56429691, Global Avg Loss: 1.87730746, Time: 0.0403 Steps: 29470, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001803, Sample Num: 28848, Cur Loss: 2.42142916, Cur Avg Loss: 1.49766434, Log Avg loss: 1.65678069, Global Avg Loss: 1.87723265, Time: 0.0403 Steps: 29480, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001813, Sample Num: 29008, Cur Loss: 0.87392217, Cur Avg Loss: 1.49600188, Log Avg loss: 1.19626061, Global Avg Loss: 1.87700174, Time: 0.0403 Steps: 29490, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001823, Sample Num: 29168, Cur Loss: 1.58055115, Cur Avg Loss: 1.49654176, Log Avg loss: 1.59442276, Global Avg Loss: 1.87690595, Time: 0.0403 Steps: 29500, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001833, Sample Num: 29328, Cur Loss: 0.49893358, Cur Avg Loss: 1.49671862, Log Avg loss: 1.52896000, Global Avg Loss: 1.87678804, Time: 0.0403 Steps: 29510, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001843, Sample Num: 29488, Cur Loss: 1.89405513, Cur Avg Loss: 1.49952858, Log Avg loss: 2.01459423, Global Avg Loss: 1.87683472, Time: 0.0403 Steps: 29520, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001853, Sample Num: 29648, Cur Loss: 1.66598392, Cur Avg Loss: 1.50189585, Log Avg loss: 1.93818256, Global Avg Loss: 1.87685550, Time: 0.0403 Steps: 29530, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001863, Sample Num: 29808, Cur Loss: 1.33264971, Cur Avg Loss: 1.50251311, Log Avg loss: 1.61689281, Global Avg Loss: 1.87676749, Time: 0.0403 Steps: 29540, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001873, Sample Num: 29968, Cur Loss: 1.40868878, Cur Avg Loss: 1.50234528, Log Avg loss: 1.47107836, Global Avg Loss: 1.87663020, Time: 0.0729 Steps: 29550, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001883, Sample Num: 30128, Cur Loss: 1.13145137, Cur Avg Loss: 1.50065841, Log Avg loss: 1.18470643, Global Avg Loss: 1.87639613, Time: 0.0784 Steps: 29560, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001893, Sample Num: 30288, Cur Loss: 2.35206318, Cur Avg Loss: 1.50140299, Log Avg loss: 1.64160748, Global Avg Loss: 1.87631673, Time: 0.0403 Steps: 29570, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001903, Sample Num: 30448, Cur Loss: 1.77616525, Cur Avg Loss: 1.50034908, Log Avg loss: 1.30084461, Global Avg Loss: 1.87612218, Time: 0.0508 Steps: 29580, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001913, Sample Num: 30608, Cur Loss: 1.30363619, Cur Avg Loss: 1.49996096, Log Avg loss: 1.42610221, Global Avg Loss: 1.87597009, Time: 0.0860 Steps: 29590, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001923, Sample Num: 30768, Cur Loss: 1.70516109, Cur Avg Loss: 1.50120095, Log Avg loss: 1.73841109, Global Avg Loss: 1.87592362, Time: 0.0841 Steps: 29600, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001933, Sample Num: 30928, Cur Loss: 1.42946315, Cur Avg Loss: 1.50136975, Log Avg loss: 1.53382951, Global Avg Loss: 1.87580809, Time: 0.0485 Steps: 29610, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001943, Sample Num: 31088, Cur Loss: 1.63354945, Cur Avg Loss: 1.50302087, Log Avg loss: 1.82218246, Global Avg Loss: 1.87578998, Time: 0.0403 Steps: 29620, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001953, Sample Num: 31248, Cur Loss: 1.56095123, Cur Avg Loss: 1.50200495, Log Avg loss: 1.30461115, Global Avg Loss: 1.87559721, Time: 0.0404 Steps: 29630, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001963, Sample Num: 31408, Cur Loss: 1.51295280, Cur Avg Loss: 1.50208461, Log Avg loss: 1.51764229, Global Avg Loss: 1.87547645, Time: 0.0403 Steps: 29640, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001973, Sample Num: 31568, Cur Loss: 1.23709345, Cur Avg Loss: 1.50264119, Log Avg loss: 1.61189824, Global Avg Loss: 1.87538755, Time: 0.0404 Steps: 29650, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001983, Sample Num: 31728, Cur Loss: 1.35314631, Cur Avg Loss: 1.50105029, Log Avg loss: 1.18716620, Global Avg Loss: 1.87515551, Time: 0.0404 Steps: 29660, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001993, Sample Num: 31888, Cur Loss: 2.53370333, Cur Avg Loss: 1.50130576, Log Avg loss: 1.55196531, Global Avg Loss: 1.87504658, Time: 0.0403 Steps: 29670, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002003, Sample Num: 32048, Cur Loss: 1.21125638, Cur Avg Loss: 1.50096572, Log Avg loss: 1.43319442, Global Avg Loss: 1.87489771, Time: 0.0404 Steps: 29680, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002013, Sample Num: 32208, Cur Loss: 2.52515578, Cur Avg Loss: 1.50077273, Log Avg loss: 1.46211779, Global Avg Loss: 1.87475868, Time: 0.0404 Steps: 29690, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002023, Sample Num: 32368, Cur Loss: 1.37184107, Cur Avg Loss: 1.49916865, Log Avg loss: 1.17626671, Global Avg Loss: 1.87452350, Time: 0.0403 Steps: 29700, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002033, Sample Num: 32528, Cur Loss: 0.86108613, Cur Avg Loss: 1.50008817, Log Avg loss: 1.68610733, Global Avg Loss: 1.87446008, Time: 0.0403 Steps: 29710, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002043, Sample Num: 32688, Cur Loss: 2.19409418, Cur Avg Loss: 1.50091420, Log Avg loss: 1.66884681, Global Avg Loss: 1.87439090, Time: 0.0404 Steps: 29720, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002053, Sample Num: 32848, Cur Loss: 1.55889988, Cur Avg Loss: 1.50058803, Log Avg loss: 1.43395120, Global Avg Loss: 1.87424275, Time: 0.0403 Steps: 29730, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002063, Sample Num: 33008, Cur Loss: 1.22543013, Cur Avg Loss: 1.50177042, Log Avg loss: 1.74451588, Global Avg Loss: 1.87419913, Time: 0.0404 Steps: 29740, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002073, Sample Num: 33168, Cur Loss: 1.17321372, Cur Avg Loss: 1.50100560, Log Avg loss: 1.34322242, Global Avg Loss: 1.87402065, Time: 0.0404 Steps: 29750, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002083, Sample Num: 33328, Cur Loss: 1.25349069, Cur Avg Loss: 1.50105862, Log Avg loss: 1.51204960, Global Avg Loss: 1.87389902, Time: 0.0403 Steps: 29760, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002093, Sample Num: 33488, Cur Loss: 1.03633547, Cur Avg Loss: 1.50026860, Log Avg loss: 1.33570791, Global Avg Loss: 1.87371824, Time: 0.0403 Steps: 29770, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002103, Sample Num: 33648, Cur Loss: 3.03985310, Cur Avg Loss: 1.50163187, Log Avg loss: 1.78696409, Global Avg Loss: 1.87368911, Time: 0.0402 Steps: 29780, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002113, Sample Num: 33808, Cur Loss: 0.73781472, Cur Avg Loss: 1.50012972, Log Avg loss: 1.18422728, Global Avg Loss: 1.87345767, Time: 0.0404 Steps: 29790, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002123, Sample Num: 33968, Cur Loss: 1.16882730, Cur Avg Loss: 1.50120510, Log Avg loss: 1.72843184, Global Avg Loss: 1.87340900, Time: 0.0403 Steps: 29800, Updated lr: 0.000073 ***** Running evaluation checkpoint-29806 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-29806 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.492477, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.712923, "eval_total_loss": 1204.184639, "eval_mae": 1.191852, "eval_mse": 1.712573, "eval_r2": -0.088624, "eval_sp_statistic": 0.29088, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.368235, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.329896, "test_total_loss": 667.607743, "test_mae": 0.831493, "test_mse": 1.330285, "test_r2": 0.141422, "test_sp_statistic": 0.302926, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.394293, "test_ps_pvalue": 0.0, "lr": 7.26827880512091e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.873293094123013, "train_cur_epoch_loss": 3194.844191670418, "train_cur_epoch_avg_loss": 1.5006313723205345, "train_cur_epoch_time": 95.49247670173645, "train_cur_epoch_avg_time": 0.04485320652970242, "epoch": 14, "step": 29806} ################################################## Training, Epoch: 0015, Batch: 000004, Sample Num: 64, Cur Loss: 1.18540907, Cur Avg Loss: 1.35667090, Log Avg loss: 1.32124584, Global Avg Loss: 1.87322377, Time: 0.0404 Steps: 29810, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000014, Sample Num: 224, Cur Loss: 1.70398211, Cur Avg Loss: 1.45249749, Log Avg loss: 1.49082812, Global Avg Loss: 1.87309554, Time: 0.0404 Steps: 29820, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000024, Sample Num: 384, Cur Loss: 0.95192063, Cur Avg Loss: 1.47765887, Log Avg loss: 1.51288480, Global Avg Loss: 1.87297478, Time: 0.0403 Steps: 29830, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000034, Sample Num: 544, Cur Loss: 0.85667050, Cur Avg Loss: 1.53144444, Log Avg loss: 1.66052981, Global Avg Loss: 1.87290359, Time: 0.0403 Steps: 29840, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000044, Sample Num: 704, Cur Loss: 0.61013502, Cur Avg Loss: 1.46556592, Log Avg loss: 1.24157895, Global Avg Loss: 1.87269209, Time: 0.0403 Steps: 29850, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000054, Sample Num: 864, Cur Loss: 0.28879860, Cur Avg Loss: 1.48102023, Log Avg loss: 1.54901921, Global Avg Loss: 1.87258369, Time: 0.0404 Steps: 29860, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000064, Sample Num: 1024, Cur Loss: 2.25428247, Cur Avg Loss: 1.53576031, Log Avg loss: 1.83135673, Global Avg Loss: 1.87256989, Time: 0.0403 Steps: 29870, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000074, Sample Num: 1184, Cur Loss: 0.76148522, Cur Avg Loss: 1.52750603, Log Avg loss: 1.47467865, Global Avg Loss: 1.87243673, Time: 0.0404 Steps: 29880, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000084, Sample Num: 1344, Cur Loss: 2.94752336, Cur Avg Loss: 1.53466718, Log Avg loss: 1.58765966, Global Avg Loss: 1.87234145, Time: 0.0403 Steps: 29890, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000094, Sample Num: 1504, Cur Loss: 1.34980452, Cur Avg Loss: 1.57550127, Log Avg loss: 1.91850764, Global Avg Loss: 1.87235689, Time: 0.0403 Steps: 29900, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000104, Sample Num: 1664, Cur Loss: 2.15768075, Cur Avg Loss: 1.61181471, Log Avg loss: 1.95316104, Global Avg Loss: 1.87238391, Time: 0.0403 Steps: 29910, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000114, Sample Num: 1824, Cur Loss: 3.09758425, Cur Avg Loss: 1.62856596, Log Avg loss: 1.80277901, Global Avg Loss: 1.87236064, Time: 0.0531 Steps: 29920, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000124, Sample Num: 1984, Cur Loss: 1.48668289, Cur Avg Loss: 1.63040836, Log Avg loss: 1.65141166, Global Avg Loss: 1.87228682, Time: 0.0457 Steps: 29930, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000134, Sample Num: 2144, Cur Loss: 0.93931860, Cur Avg Loss: 1.64204480, Log Avg loss: 1.78633671, Global Avg Loss: 1.87225812, Time: 0.0406 Steps: 29940, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000144, Sample Num: 2304, Cur Loss: 1.35038400, Cur Avg Loss: 1.63873713, Log Avg loss: 1.59441426, Global Avg Loss: 1.87216535, Time: 0.0550 Steps: 29950, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000154, Sample Num: 2464, Cur Loss: 1.11226332, Cur Avg Loss: 1.61895529, Log Avg loss: 1.33409680, Global Avg Loss: 1.87198575, Time: 0.0664 Steps: 29960, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000164, Sample Num: 2624, Cur Loss: 1.35580182, Cur Avg Loss: 1.61752174, Log Avg loss: 1.59544508, Global Avg Loss: 1.87189348, Time: 0.0440 Steps: 29970, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000174, Sample Num: 2784, Cur Loss: 1.52761579, Cur Avg Loss: 1.59644573, Log Avg loss: 1.25079918, Global Avg Loss: 1.87168631, Time: 0.0568 Steps: 29980, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000184, Sample Num: 2944, Cur Loss: 1.66791105, Cur Avg Loss: 1.59790814, Log Avg loss: 1.62335414, Global Avg Loss: 1.87160350, Time: 0.0636 Steps: 29990, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000194, Sample Num: 3104, Cur Loss: 1.44469118, Cur Avg Loss: 1.59806691, Log Avg loss: 1.60098826, Global Avg Loss: 1.87151330, Time: 0.0403 Steps: 30000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000204, Sample Num: 3264, Cur Loss: 0.84448826, Cur Avg Loss: 1.58197344, Log Avg loss: 1.26976000, Global Avg Loss: 1.87131278, Time: 0.0403 Steps: 30010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000214, Sample Num: 3424, Cur Loss: 0.90436023, Cur Avg Loss: 1.56842183, Log Avg loss: 1.29196902, Global Avg Loss: 1.87111979, Time: 0.0404 Steps: 30020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000224, Sample Num: 3584, Cur Loss: 2.03433776, Cur Avg Loss: 1.55135650, Log Avg loss: 1.18615850, Global Avg Loss: 1.87089170, Time: 0.0404 Steps: 30030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000234, Sample Num: 3744, Cur Loss: 1.94344425, Cur Avg Loss: 1.53582549, Log Avg loss: 1.18793077, Global Avg Loss: 1.87066435, Time: 0.0403 Steps: 30040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000244, Sample Num: 3904, Cur Loss: 1.31607032, Cur Avg Loss: 1.52943074, Log Avg loss: 1.37979361, Global Avg Loss: 1.87050100, Time: 0.0403 Steps: 30050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000254, Sample Num: 4064, Cur Loss: 1.51427603, Cur Avg Loss: 1.52413372, Log Avg loss: 1.39488648, Global Avg Loss: 1.87034278, Time: 0.0404 Steps: 30060, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000264, Sample Num: 4224, Cur Loss: 1.63932252, Cur Avg Loss: 1.52190793, Log Avg loss: 1.46537293, Global Avg Loss: 1.87020810, Time: 0.0403 Steps: 30070, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000274, Sample Num: 4384, Cur Loss: 1.03102374, Cur Avg Loss: 1.52385763, Log Avg loss: 1.57532973, Global Avg Loss: 1.87011007, Time: 0.0404 Steps: 30080, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000284, Sample Num: 4544, Cur Loss: 2.59149408, Cur Avg Loss: 1.51924667, Log Avg loss: 1.39290633, Global Avg Loss: 1.86995148, Time: 0.0404 Steps: 30090, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000294, Sample Num: 4704, Cur Loss: 1.36358130, Cur Avg Loss: 1.51675887, Log Avg loss: 1.44610543, Global Avg Loss: 1.86981067, Time: 0.0403 Steps: 30100, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000304, Sample Num: 4864, Cur Loss: 0.60108340, Cur Avg Loss: 1.51468287, Log Avg loss: 1.45364822, Global Avg Loss: 1.86967245, Time: 0.0403 Steps: 30110, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000314, Sample Num: 5024, Cur Loss: 0.86848450, Cur Avg Loss: 1.51591934, Log Avg loss: 1.55350813, Global Avg Loss: 1.86956748, Time: 0.0403 Steps: 30120, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000324, Sample Num: 5184, Cur Loss: 1.64320183, Cur Avg Loss: 1.51934329, Log Avg loss: 1.62685521, Global Avg Loss: 1.86948693, Time: 0.0404 Steps: 30130, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000334, Sample Num: 5344, Cur Loss: 1.21751595, Cur Avg Loss: 1.51146131, Log Avg loss: 1.25608518, Global Avg Loss: 1.86928341, Time: 0.0403 Steps: 30140, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000344, Sample Num: 5504, Cur Loss: 0.99395639, Cur Avg Loss: 1.51222676, Log Avg loss: 1.53779292, Global Avg Loss: 1.86917346, Time: 0.0403 Steps: 30150, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000354, Sample Num: 5664, Cur Loss: 1.34290195, Cur Avg Loss: 1.51121565, Log Avg loss: 1.47643350, Global Avg Loss: 1.86904325, Time: 0.0404 Steps: 30160, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000364, Sample Num: 5824, Cur Loss: 1.14528286, Cur Avg Loss: 1.50591063, Log Avg loss: 1.31811304, Global Avg Loss: 1.86886064, Time: 0.0403 Steps: 30170, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000374, Sample Num: 5984, Cur Loss: 0.96393293, Cur Avg Loss: 1.49870539, Log Avg loss: 1.23643453, Global Avg Loss: 1.86865109, Time: 0.0403 Steps: 30180, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000384, Sample Num: 6144, Cur Loss: 1.82403421, Cur Avg Loss: 1.50155699, Log Avg loss: 1.60820682, Global Avg Loss: 1.86856482, Time: 0.0403 Steps: 30190, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000394, Sample Num: 6304, Cur Loss: 1.46543622, Cur Avg Loss: 1.50643492, Log Avg loss: 1.69374744, Global Avg Loss: 1.86850693, Time: 0.0403 Steps: 30200, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000404, Sample Num: 6464, Cur Loss: 1.73621774, Cur Avg Loss: 1.51477441, Log Avg loss: 1.84335027, Global Avg Loss: 1.86849860, Time: 0.0402 Steps: 30210, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000414, Sample Num: 6624, Cur Loss: 1.37312186, Cur Avg Loss: 1.51807664, Log Avg loss: 1.65148666, Global Avg Loss: 1.86842679, Time: 0.0403 Steps: 30220, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000424, Sample Num: 6784, Cur Loss: 1.06698072, Cur Avg Loss: 1.51179077, Log Avg loss: 1.25155564, Global Avg Loss: 1.86822273, Time: 0.0403 Steps: 30230, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000434, Sample Num: 6944, Cur Loss: 0.58623564, Cur Avg Loss: 1.51248577, Log Avg loss: 1.54195390, Global Avg Loss: 1.86811484, Time: 0.0402 Steps: 30240, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000444, Sample Num: 7104, Cur Loss: 1.61343932, Cur Avg Loss: 1.51008425, Log Avg loss: 1.40585818, Global Avg Loss: 1.86796203, Time: 0.0404 Steps: 30250, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000454, Sample Num: 7264, Cur Loss: 1.47031736, Cur Avg Loss: 1.50489500, Log Avg loss: 1.27449235, Global Avg Loss: 1.86776591, Time: 0.0403 Steps: 30260, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000464, Sample Num: 7424, Cur Loss: 1.63031209, Cur Avg Loss: 1.50250451, Log Avg loss: 1.39397656, Global Avg Loss: 1.86760938, Time: 0.0946 Steps: 30270, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000474, Sample Num: 7584, Cur Loss: 1.01747370, Cur Avg Loss: 1.49760184, Log Avg loss: 1.27011782, Global Avg Loss: 1.86741206, Time: 0.0528 Steps: 30280, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000484, Sample Num: 7744, Cur Loss: 0.59678352, Cur Avg Loss: 1.48971805, Log Avg loss: 1.11602618, Global Avg Loss: 1.86716400, Time: 0.0406 Steps: 30290, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000494, Sample Num: 7904, Cur Loss: 2.86684799, Cur Avg Loss: 1.49261558, Log Avg loss: 1.63285638, Global Avg Loss: 1.86708667, Time: 0.0563 Steps: 30300, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000504, Sample Num: 8064, Cur Loss: 1.25857615, Cur Avg Loss: 1.48815647, Log Avg loss: 1.26787622, Global Avg Loss: 1.86688897, Time: 0.0777 Steps: 30310, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000514, Sample Num: 8224, Cur Loss: 0.62728894, Cur Avg Loss: 1.49132796, Log Avg loss: 1.65117088, Global Avg Loss: 1.86681783, Time: 0.0856 Steps: 30320, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000524, Sample Num: 8384, Cur Loss: 1.48997140, Cur Avg Loss: 1.49504341, Log Avg loss: 1.68601753, Global Avg Loss: 1.86675822, Time: 0.0562 Steps: 30330, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000534, Sample Num: 8544, Cur Loss: 2.09726715, Cur Avg Loss: 1.49092124, Log Avg loss: 1.27491998, Global Avg Loss: 1.86656315, Time: 0.0409 Steps: 30340, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000544, Sample Num: 8704, Cur Loss: 2.40953517, Cur Avg Loss: 1.48696053, Log Avg loss: 1.27545866, Global Avg Loss: 1.86636839, Time: 0.0484 Steps: 30350, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000554, Sample Num: 8864, Cur Loss: 1.03786540, Cur Avg Loss: 1.47920466, Log Avg loss: 1.05728487, Global Avg Loss: 1.86610189, Time: 0.0403 Steps: 30360, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000564, Sample Num: 9024, Cur Loss: 1.92624283, Cur Avg Loss: 1.47681774, Log Avg loss: 1.34458272, Global Avg Loss: 1.86593017, Time: 0.0403 Steps: 30370, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000574, Sample Num: 9184, Cur Loss: 1.49875116, Cur Avg Loss: 1.47475460, Log Avg loss: 1.35839341, Global Avg Loss: 1.86576310, Time: 0.0404 Steps: 30380, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000584, Sample Num: 9344, Cur Loss: 0.99474382, Cur Avg Loss: 1.47238219, Log Avg loss: 1.33620599, Global Avg Loss: 1.86558885, Time: 0.0403 Steps: 30390, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000594, Sample Num: 9504, Cur Loss: 0.94619370, Cur Avg Loss: 1.47064580, Log Avg loss: 1.36924033, Global Avg Loss: 1.86542558, Time: 0.0403 Steps: 30400, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000604, Sample Num: 9664, Cur Loss: 1.76144886, Cur Avg Loss: 1.46979459, Log Avg loss: 1.41923273, Global Avg Loss: 1.86527885, Time: 0.0404 Steps: 30410, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000614, Sample Num: 9824, Cur Loss: 0.56837291, Cur Avg Loss: 1.47104665, Log Avg loss: 1.54667144, Global Avg Loss: 1.86517412, Time: 0.0402 Steps: 30420, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000624, Sample Num: 9984, Cur Loss: 0.95121884, Cur Avg Loss: 1.46978325, Log Avg loss: 1.39221036, Global Avg Loss: 1.86501869, Time: 0.0403 Steps: 30430, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000634, Sample Num: 10144, Cur Loss: 1.40857220, Cur Avg Loss: 1.46782229, Log Avg loss: 1.34545804, Global Avg Loss: 1.86484801, Time: 0.0403 Steps: 30440, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000644, Sample Num: 10304, Cur Loss: 0.97661704, Cur Avg Loss: 1.46234226, Log Avg loss: 1.11490880, Global Avg Loss: 1.86460172, Time: 0.0403 Steps: 30450, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000654, Sample Num: 10464, Cur Loss: 0.41299120, Cur Avg Loss: 1.45616046, Log Avg loss: 1.05805233, Global Avg Loss: 1.86433693, Time: 0.0403 Steps: 30460, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000664, Sample Num: 10624, Cur Loss: 1.12565255, Cur Avg Loss: 1.46082642, Log Avg loss: 1.76598014, Global Avg Loss: 1.86430465, Time: 0.0404 Steps: 30470, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000674, Sample Num: 10784, Cur Loss: 1.14628720, Cur Avg Loss: 1.45696510, Log Avg loss: 1.20057322, Global Avg Loss: 1.86408689, Time: 0.0403 Steps: 30480, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000684, Sample Num: 10944, Cur Loss: 2.75216222, Cur Avg Loss: 1.45989543, Log Avg loss: 1.65739996, Global Avg Loss: 1.86401910, Time: 0.0403 Steps: 30490, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000694, Sample Num: 11104, Cur Loss: 1.42688286, Cur Avg Loss: 1.45977185, Log Avg loss: 1.45131918, Global Avg Loss: 1.86388379, Time: 0.0402 Steps: 30500, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000704, Sample Num: 11264, Cur Loss: 1.97405457, Cur Avg Loss: 1.46016616, Log Avg loss: 1.48753096, Global Avg Loss: 1.86376044, Time: 0.0403 Steps: 30510, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000714, Sample Num: 11424, Cur Loss: 0.97641325, Cur Avg Loss: 1.45741538, Log Avg loss: 1.26376076, Global Avg Loss: 1.86356384, Time: 0.0403 Steps: 30520, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000724, Sample Num: 11584, Cur Loss: 1.35560274, Cur Avg Loss: 1.45521600, Log Avg loss: 1.29818004, Global Avg Loss: 1.86337866, Time: 0.0402 Steps: 30530, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000734, Sample Num: 11744, Cur Loss: 1.43816781, Cur Avg Loss: 1.45380990, Log Avg loss: 1.35200819, Global Avg Loss: 1.86321121, Time: 0.0402 Steps: 30540, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000744, Sample Num: 11904, Cur Loss: 1.11214757, Cur Avg Loss: 1.45562697, Log Avg loss: 1.58899997, Global Avg Loss: 1.86312145, Time: 0.0402 Steps: 30550, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000754, Sample Num: 12064, Cur Loss: 0.46548122, Cur Avg Loss: 1.45452024, Log Avg loss: 1.37217949, Global Avg Loss: 1.86296081, Time: 0.0402 Steps: 30560, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000764, Sample Num: 12224, Cur Loss: 1.41507840, Cur Avg Loss: 1.45816643, Log Avg loss: 1.73308915, Global Avg Loss: 1.86291832, Time: 0.0403 Steps: 30570, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000774, Sample Num: 12384, Cur Loss: 0.47462627, Cur Avg Loss: 1.45724959, Log Avg loss: 1.38720306, Global Avg Loss: 1.86276276, Time: 0.0403 Steps: 30580, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000784, Sample Num: 12544, Cur Loss: 2.28444505, Cur Avg Loss: 1.45477725, Log Avg loss: 1.26341802, Global Avg Loss: 1.86256683, Time: 0.0403 Steps: 30590, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000794, Sample Num: 12704, Cur Loss: 1.60670042, Cur Avg Loss: 1.45509373, Log Avg loss: 1.47990595, Global Avg Loss: 1.86244178, Time: 0.0404 Steps: 30600, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000804, Sample Num: 12864, Cur Loss: 0.48040247, Cur Avg Loss: 1.45282495, Log Avg loss: 1.27268345, Global Avg Loss: 1.86224911, Time: 0.0403 Steps: 30610, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000814, Sample Num: 13024, Cur Loss: 2.02164555, Cur Avg Loss: 1.45584873, Log Avg loss: 1.69896089, Global Avg Loss: 1.86219578, Time: 0.0849 Steps: 30620, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000824, Sample Num: 13184, Cur Loss: 1.45110309, Cur Avg Loss: 1.45639027, Log Avg loss: 1.50047143, Global Avg Loss: 1.86207769, Time: 0.0512 Steps: 30630, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000834, Sample Num: 13344, Cur Loss: 1.18695211, Cur Avg Loss: 1.45873959, Log Avg loss: 1.65232366, Global Avg Loss: 1.86200923, Time: 0.0658 Steps: 30640, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000844, Sample Num: 13504, Cur Loss: 1.15662754, Cur Avg Loss: 1.46310732, Log Avg loss: 1.82737657, Global Avg Loss: 1.86199793, Time: 0.0404 Steps: 30650, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000854, Sample Num: 13664, Cur Loss: 1.53790379, Cur Avg Loss: 1.46436498, Log Avg loss: 1.57051096, Global Avg Loss: 1.86190286, Time: 0.0436 Steps: 30660, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000864, Sample Num: 13824, Cur Loss: 1.00097883, Cur Avg Loss: 1.46442351, Log Avg loss: 1.46942170, Global Avg Loss: 1.86177489, Time: 0.0492 Steps: 30670, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000874, Sample Num: 13984, Cur Loss: 1.40553904, Cur Avg Loss: 1.46830846, Log Avg loss: 1.80396835, Global Avg Loss: 1.86175605, Time: 0.0581 Steps: 30680, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000884, Sample Num: 14144, Cur Loss: 1.18326318, Cur Avg Loss: 1.46633505, Log Avg loss: 1.29385890, Global Avg Loss: 1.86157101, Time: 0.0738 Steps: 30690, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000894, Sample Num: 14304, Cur Loss: 1.50980830, Cur Avg Loss: 1.46432716, Log Avg loss: 1.28682956, Global Avg Loss: 1.86138379, Time: 0.1011 Steps: 30700, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000904, Sample Num: 14464, Cur Loss: 1.01785016, Cur Avg Loss: 1.46197066, Log Avg loss: 1.25130016, Global Avg Loss: 1.86118513, Time: 0.0403 Steps: 30710, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000914, Sample Num: 14624, Cur Loss: 3.04965544, Cur Avg Loss: 1.46341056, Log Avg loss: 1.59357749, Global Avg Loss: 1.86109802, Time: 0.0403 Steps: 30720, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000924, Sample Num: 14784, Cur Loss: 0.90799040, Cur Avg Loss: 1.46688287, Log Avg loss: 1.78425179, Global Avg Loss: 1.86107301, Time: 0.0403 Steps: 30730, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000934, Sample Num: 14944, Cur Loss: 1.76903033, Cur Avg Loss: 1.46730889, Log Avg loss: 1.50667365, Global Avg Loss: 1.86095773, Time: 0.0404 Steps: 30740, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000944, Sample Num: 15104, Cur Loss: 2.08001709, Cur Avg Loss: 1.46761647, Log Avg loss: 1.49634386, Global Avg Loss: 1.86083915, Time: 0.0403 Steps: 30750, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000954, Sample Num: 15264, Cur Loss: 1.13212872, Cur Avg Loss: 1.46841092, Log Avg loss: 1.54340680, Global Avg Loss: 1.86073596, Time: 0.0403 Steps: 30760, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000964, Sample Num: 15424, Cur Loss: 1.15054774, Cur Avg Loss: 1.46820310, Log Avg loss: 1.44837765, Global Avg Loss: 1.86060194, Time: 0.0403 Steps: 30770, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000974, Sample Num: 15584, Cur Loss: 0.95206958, Cur Avg Loss: 1.46903875, Log Avg loss: 1.54959524, Global Avg Loss: 1.86050090, Time: 0.0403 Steps: 30780, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000984, Sample Num: 15744, Cur Loss: 1.48118496, Cur Avg Loss: 1.47012062, Log Avg loss: 1.57549475, Global Avg Loss: 1.86040834, Time: 0.0403 Steps: 30790, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000994, Sample Num: 15904, Cur Loss: 2.18667245, Cur Avg Loss: 1.47339748, Log Avg loss: 1.79584081, Global Avg Loss: 1.86038737, Time: 0.0403 Steps: 30800, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001004, Sample Num: 16064, Cur Loss: 2.42227316, Cur Avg Loss: 1.47311345, Log Avg loss: 1.44488076, Global Avg Loss: 1.86025251, Time: 0.0403 Steps: 30810, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001014, Sample Num: 16224, Cur Loss: 3.13462400, Cur Avg Loss: 1.47524861, Log Avg loss: 1.68961862, Global Avg Loss: 1.86019715, Time: 0.0403 Steps: 30820, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001024, Sample Num: 16384, Cur Loss: 0.53676373, Cur Avg Loss: 1.47504945, Log Avg loss: 1.45485422, Global Avg Loss: 1.86006567, Time: 0.0404 Steps: 30830, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001034, Sample Num: 16544, Cur Loss: 1.45240617, Cur Avg Loss: 1.47350931, Log Avg loss: 1.31579944, Global Avg Loss: 1.85988919, Time: 0.0403 Steps: 30840, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001044, Sample Num: 16704, Cur Loss: 2.43335676, Cur Avg Loss: 1.47164630, Log Avg loss: 1.27901107, Global Avg Loss: 1.85970090, Time: 0.0402 Steps: 30850, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001054, Sample Num: 16864, Cur Loss: 2.36671591, Cur Avg Loss: 1.47428192, Log Avg loss: 1.74944041, Global Avg Loss: 1.85966517, Time: 0.0404 Steps: 30860, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001064, Sample Num: 17024, Cur Loss: 1.65315270, Cur Avg Loss: 1.47493753, Log Avg loss: 1.54403841, Global Avg Loss: 1.85956292, Time: 0.0404 Steps: 30870, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001074, Sample Num: 17184, Cur Loss: 2.12051249, Cur Avg Loss: 1.47435318, Log Avg loss: 1.41217908, Global Avg Loss: 1.85941805, Time: 0.0403 Steps: 30880, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001084, Sample Num: 17344, Cur Loss: 1.07485962, Cur Avg Loss: 1.47464813, Log Avg loss: 1.50632561, Global Avg Loss: 1.85930374, Time: 0.0403 Steps: 30890, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001094, Sample Num: 17504, Cur Loss: 1.41320753, Cur Avg Loss: 1.47451396, Log Avg loss: 1.45996975, Global Avg Loss: 1.85917451, Time: 0.0403 Steps: 30900, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001104, Sample Num: 17664, Cur Loss: 0.94654572, Cur Avg Loss: 1.47354237, Log Avg loss: 1.36725093, Global Avg Loss: 1.85901536, Time: 0.0402 Steps: 30910, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001114, Sample Num: 17824, Cur Loss: 0.89360058, Cur Avg Loss: 1.47166885, Log Avg loss: 1.26483154, Global Avg Loss: 1.85882319, Time: 0.0403 Steps: 30920, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001124, Sample Num: 17984, Cur Loss: 1.35200024, Cur Avg Loss: 1.47087967, Log Avg loss: 1.38296478, Global Avg Loss: 1.85866934, Time: 0.0403 Steps: 30930, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001134, Sample Num: 18144, Cur Loss: 1.73089576, Cur Avg Loss: 1.46944457, Log Avg loss: 1.30813929, Global Avg Loss: 1.85849141, Time: 0.0402 Steps: 30940, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001144, Sample Num: 18304, Cur Loss: 1.42775488, Cur Avg Loss: 1.47161510, Log Avg loss: 1.71775327, Global Avg Loss: 1.85844593, Time: 0.0402 Steps: 30950, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001154, Sample Num: 18464, Cur Loss: 1.90894282, Cur Avg Loss: 1.46865032, Log Avg loss: 1.12948015, Global Avg Loss: 1.85821048, Time: 0.0403 Steps: 30960, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001164, Sample Num: 18624, Cur Loss: 1.40543199, Cur Avg Loss: 1.47063273, Log Avg loss: 1.69940314, Global Avg Loss: 1.85815920, Time: 0.0403 Steps: 30970, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001174, Sample Num: 18784, Cur Loss: 3.41635060, Cur Avg Loss: 1.47117063, Log Avg loss: 1.53378195, Global Avg Loss: 1.85805450, Time: 0.0755 Steps: 30980, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001184, Sample Num: 18944, Cur Loss: 0.96902174, Cur Avg Loss: 1.46929998, Log Avg loss: 1.24968547, Global Avg Loss: 1.85785818, Time: 0.0549 Steps: 30990, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001194, Sample Num: 19104, Cur Loss: 0.93867505, Cur Avg Loss: 1.46843597, Log Avg loss: 1.36613744, Global Avg Loss: 1.85769957, Time: 0.0545 Steps: 31000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001204, Sample Num: 19264, Cur Loss: 1.94172418, Cur Avg Loss: 1.47096877, Log Avg loss: 1.77338524, Global Avg Loss: 1.85767238, Time: 0.0406 Steps: 31010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001214, Sample Num: 19424, Cur Loss: 1.01141417, Cur Avg Loss: 1.47288131, Log Avg loss: 1.70315078, Global Avg Loss: 1.85762256, Time: 0.0625 Steps: 31020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001224, Sample Num: 19584, Cur Loss: 1.79982328, Cur Avg Loss: 1.47423912, Log Avg loss: 1.63907713, Global Avg Loss: 1.85755213, Time: 0.0791 Steps: 31030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001234, Sample Num: 19744, Cur Loss: 0.67908263, Cur Avg Loss: 1.47382773, Log Avg loss: 1.42347300, Global Avg Loss: 1.85741229, Time: 0.0515 Steps: 31040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001244, Sample Num: 19904, Cur Loss: 0.84956205, Cur Avg Loss: 1.47209899, Log Avg loss: 1.25877305, Global Avg Loss: 1.85721949, Time: 0.0678 Steps: 31050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001254, Sample Num: 20064, Cur Loss: 1.42834675, Cur Avg Loss: 1.47470024, Log Avg loss: 1.79829632, Global Avg Loss: 1.85720052, Time: 0.0463 Steps: 31060, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001264, Sample Num: 20224, Cur Loss: 1.15499353, Cur Avg Loss: 1.47397863, Log Avg loss: 1.38348859, Global Avg Loss: 1.85704805, Time: 0.0402 Steps: 31070, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001274, Sample Num: 20384, Cur Loss: 2.15048885, Cur Avg Loss: 1.47268977, Log Avg loss: 1.30977719, Global Avg Loss: 1.85687197, Time: 0.0403 Steps: 31080, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001284, Sample Num: 20544, Cur Loss: 1.50281560, Cur Avg Loss: 1.47201753, Log Avg loss: 1.38637412, Global Avg Loss: 1.85672063, Time: 0.0404 Steps: 31090, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001294, Sample Num: 20704, Cur Loss: 1.13541102, Cur Avg Loss: 1.47105281, Log Avg loss: 1.34718300, Global Avg Loss: 1.85655679, Time: 0.0403 Steps: 31100, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001304, Sample Num: 20864, Cur Loss: 0.99378371, Cur Avg Loss: 1.47113319, Log Avg loss: 1.48153462, Global Avg Loss: 1.85643625, Time: 0.0404 Steps: 31110, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001314, Sample Num: 21024, Cur Loss: 1.98591685, Cur Avg Loss: 1.47107135, Log Avg loss: 1.46300750, Global Avg Loss: 1.85630982, Time: 0.0403 Steps: 31120, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001324, Sample Num: 21184, Cur Loss: 0.88266623, Cur Avg Loss: 1.46805306, Log Avg loss: 1.07144989, Global Avg Loss: 1.85605770, Time: 0.0403 Steps: 31130, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001334, Sample Num: 21344, Cur Loss: 0.98924196, Cur Avg Loss: 1.46916959, Log Avg loss: 1.61699722, Global Avg Loss: 1.85598093, Time: 0.0403 Steps: 31140, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001344, Sample Num: 21504, Cur Loss: 0.85650170, Cur Avg Loss: 1.47103245, Log Avg loss: 1.71953856, Global Avg Loss: 1.85593713, Time: 0.0402 Steps: 31150, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001354, Sample Num: 21664, Cur Loss: 1.21306348, Cur Avg Loss: 1.46969642, Log Avg loss: 1.29013438, Global Avg Loss: 1.85575555, Time: 0.0403 Steps: 31160, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001364, Sample Num: 21824, Cur Loss: 1.80145812, Cur Avg Loss: 1.47405322, Log Avg loss: 2.06396393, Global Avg Loss: 1.85582235, Time: 0.0403 Steps: 31170, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001374, Sample Num: 21984, Cur Loss: 2.00085425, Cur Avg Loss: 1.47193532, Log Avg loss: 1.18305281, Global Avg Loss: 1.85560658, Time: 0.0403 Steps: 31180, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001384, Sample Num: 22144, Cur Loss: 1.76218450, Cur Avg Loss: 1.47462624, Log Avg loss: 1.84435948, Global Avg Loss: 1.85560297, Time: 0.0403 Steps: 31190, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001394, Sample Num: 22304, Cur Loss: 3.07357955, Cur Avg Loss: 1.47749087, Log Avg loss: 1.87395542, Global Avg Loss: 1.85560885, Time: 0.0402 Steps: 31200, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001404, Sample Num: 22464, Cur Loss: 1.08001363, Cur Avg Loss: 1.47630266, Log Avg loss: 1.31066591, Global Avg Loss: 1.85543425, Time: 0.0403 Steps: 31210, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001414, Sample Num: 22624, Cur Loss: 1.04837990, Cur Avg Loss: 1.47706645, Log Avg loss: 1.58430277, Global Avg Loss: 1.85534740, Time: 0.0403 Steps: 31220, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001424, Sample Num: 22784, Cur Loss: 2.17044806, Cur Avg Loss: 1.47864590, Log Avg loss: 1.70198044, Global Avg Loss: 1.85529829, Time: 0.0404 Steps: 31230, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001434, Sample Num: 22944, Cur Loss: 1.56261659, Cur Avg Loss: 1.47949451, Log Avg loss: 1.60033677, Global Avg Loss: 1.85521668, Time: 0.0403 Steps: 31240, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001444, Sample Num: 23104, Cur Loss: 1.95464921, Cur Avg Loss: 1.48023133, Log Avg loss: 1.58589145, Global Avg Loss: 1.85513050, Time: 0.0403 Steps: 31250, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001454, Sample Num: 23264, Cur Loss: 2.27245474, Cur Avg Loss: 1.47929334, Log Avg loss: 1.34384664, Global Avg Loss: 1.85496694, Time: 0.0403 Steps: 31260, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001464, Sample Num: 23424, Cur Loss: 1.23969984, Cur Avg Loss: 1.48047683, Log Avg loss: 1.65255671, Global Avg Loss: 1.85490221, Time: 0.0403 Steps: 31270, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001474, Sample Num: 23584, Cur Loss: 1.89568186, Cur Avg Loss: 1.47985601, Log Avg loss: 1.38896759, Global Avg Loss: 1.85475325, Time: 0.0402 Steps: 31280, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001484, Sample Num: 23744, Cur Loss: 0.69480276, Cur Avg Loss: 1.47831698, Log Avg loss: 1.25146352, Global Avg Loss: 1.85456045, Time: 0.0402 Steps: 31290, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001494, Sample Num: 23904, Cur Loss: 1.09320021, Cur Avg Loss: 1.47763706, Log Avg loss: 1.37673692, Global Avg Loss: 1.85440779, Time: 0.0402 Steps: 31300, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001504, Sample Num: 24064, Cur Loss: 0.78298008, Cur Avg Loss: 1.47809971, Log Avg loss: 1.54722033, Global Avg Loss: 1.85430968, Time: 0.0403 Steps: 31310, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001514, Sample Num: 24224, Cur Loss: 1.64569449, Cur Avg Loss: 1.47851225, Log Avg loss: 1.54055833, Global Avg Loss: 1.85420950, Time: 0.0404 Steps: 31320, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001524, Sample Num: 24384, Cur Loss: 0.99816895, Cur Avg Loss: 1.47678514, Log Avg loss: 1.21529993, Global Avg Loss: 1.85400557, Time: 0.0620 Steps: 31330, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001534, Sample Num: 24544, Cur Loss: 1.43544245, Cur Avg Loss: 1.47507183, Log Avg loss: 1.21396476, Global Avg Loss: 1.85380135, Time: 0.0553 Steps: 31340, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001544, Sample Num: 24704, Cur Loss: 1.21318018, Cur Avg Loss: 1.47407789, Log Avg loss: 1.32160690, Global Avg Loss: 1.85363159, Time: 0.0590 Steps: 31350, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001554, Sample Num: 24864, Cur Loss: 1.21183753, Cur Avg Loss: 1.47384385, Log Avg loss: 1.43770868, Global Avg Loss: 1.85349896, Time: 0.0714 Steps: 31360, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001564, Sample Num: 25024, Cur Loss: 1.57893658, Cur Avg Loss: 1.47372106, Log Avg loss: 1.45463942, Global Avg Loss: 1.85337181, Time: 0.0488 Steps: 31370, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001574, Sample Num: 25184, Cur Loss: 1.30450964, Cur Avg Loss: 1.47590296, Log Avg loss: 1.81715065, Global Avg Loss: 1.85336027, Time: 0.0604 Steps: 31380, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001584, Sample Num: 25344, Cur Loss: 1.23956251, Cur Avg Loss: 1.47576312, Log Avg loss: 1.45375358, Global Avg Loss: 1.85323296, Time: 0.0579 Steps: 31390, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001594, Sample Num: 25504, Cur Loss: 2.52463937, Cur Avg Loss: 1.47647355, Log Avg loss: 1.58900445, Global Avg Loss: 1.85314882, Time: 0.0586 Steps: 31400, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001604, Sample Num: 25664, Cur Loss: 0.88354599, Cur Avg Loss: 1.47614131, Log Avg loss: 1.42318230, Global Avg Loss: 1.85301193, Time: 0.0403 Steps: 31410, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001614, Sample Num: 25824, Cur Loss: 0.82250798, Cur Avg Loss: 1.47361950, Log Avg loss: 1.06912111, Global Avg Loss: 1.85276244, Time: 0.0403 Steps: 31420, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001624, Sample Num: 25984, Cur Loss: 2.98918581, Cur Avg Loss: 1.47622832, Log Avg loss: 1.89729246, Global Avg Loss: 1.85277661, Time: 0.0403 Steps: 31430, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001634, Sample Num: 26144, Cur Loss: 2.05911326, Cur Avg Loss: 1.47653564, Log Avg loss: 1.52644449, Global Avg Loss: 1.85267281, Time: 0.0403 Steps: 31440, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001644, Sample Num: 26304, Cur Loss: 0.94855523, Cur Avg Loss: 1.47553051, Log Avg loss: 1.31129205, Global Avg Loss: 1.85250067, Time: 0.0403 Steps: 31450, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001654, Sample Num: 26464, Cur Loss: 1.71489322, Cur Avg Loss: 1.47750125, Log Avg loss: 1.80149200, Global Avg Loss: 1.85248446, Time: 0.0403 Steps: 31460, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001664, Sample Num: 26624, Cur Loss: 0.93085724, Cur Avg Loss: 1.47659290, Log Avg loss: 1.32635092, Global Avg Loss: 1.85231727, Time: 0.0403 Steps: 31470, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001674, Sample Num: 26784, Cur Loss: 1.99810386, Cur Avg Loss: 1.47691207, Log Avg loss: 1.53002151, Global Avg Loss: 1.85221489, Time: 0.0403 Steps: 31480, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001684, Sample Num: 26944, Cur Loss: 0.68004477, Cur Avg Loss: 1.47363317, Log Avg loss: 0.92474614, Global Avg Loss: 1.85192036, Time: 0.0404 Steps: 31490, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001694, Sample Num: 27104, Cur Loss: 1.10027862, Cur Avg Loss: 1.47334113, Log Avg loss: 1.42416109, Global Avg Loss: 1.85178457, Time: 0.0403 Steps: 31500, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001704, Sample Num: 27264, Cur Loss: 1.49248266, Cur Avg Loss: 1.47397067, Log Avg loss: 1.58061461, Global Avg Loss: 1.85169851, Time: 0.0403 Steps: 31510, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001714, Sample Num: 27424, Cur Loss: 1.90487456, Cur Avg Loss: 1.47435548, Log Avg loss: 1.53992733, Global Avg Loss: 1.85159960, Time: 0.0402 Steps: 31520, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001724, Sample Num: 27584, Cur Loss: 2.22437143, Cur Avg Loss: 1.47608975, Log Avg loss: 1.77334298, Global Avg Loss: 1.85157478, Time: 0.0403 Steps: 31530, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001734, Sample Num: 27744, Cur Loss: 1.07166886, Cur Avg Loss: 1.47599381, Log Avg loss: 1.45945473, Global Avg Loss: 1.85145045, Time: 0.0403 Steps: 31540, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001744, Sample Num: 27904, Cur Loss: 1.05250597, Cur Avg Loss: 1.47572968, Log Avg loss: 1.42993009, Global Avg Loss: 1.85131685, Time: 0.0404 Steps: 31550, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001754, Sample Num: 28064, Cur Loss: 0.95113164, Cur Avg Loss: 1.47536041, Log Avg loss: 1.41095871, Global Avg Loss: 1.85117732, Time: 0.0403 Steps: 31560, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001764, Sample Num: 28224, Cur Loss: 0.99962103, Cur Avg Loss: 1.47652707, Log Avg loss: 1.68115993, Global Avg Loss: 1.85112346, Time: 0.0403 Steps: 31570, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001774, Sample Num: 28384, Cur Loss: 1.34397054, Cur Avg Loss: 1.47511826, Log Avg loss: 1.22660407, Global Avg Loss: 1.85092570, Time: 0.0403 Steps: 31580, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001784, Sample Num: 28544, Cur Loss: 1.58907115, Cur Avg Loss: 1.47500290, Log Avg loss: 1.45453721, Global Avg Loss: 1.85080023, Time: 0.0403 Steps: 31590, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001794, Sample Num: 28704, Cur Loss: 1.82563353, Cur Avg Loss: 1.47507360, Log Avg loss: 1.48768767, Global Avg Loss: 1.85068532, Time: 0.0403 Steps: 31600, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001804, Sample Num: 28864, Cur Loss: 1.98180902, Cur Avg Loss: 1.47585576, Log Avg loss: 1.61617389, Global Avg Loss: 1.85061113, Time: 0.0402 Steps: 31610, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001814, Sample Num: 29024, Cur Loss: 1.29433465, Cur Avg Loss: 1.47540130, Log Avg loss: 1.39341703, Global Avg Loss: 1.85046654, Time: 0.0403 Steps: 31620, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001824, Sample Num: 29184, Cur Loss: 2.83062840, Cur Avg Loss: 1.47741601, Log Avg loss: 1.84288441, Global Avg Loss: 1.85046414, Time: 0.0403 Steps: 31630, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001834, Sample Num: 29344, Cur Loss: 0.94870347, Cur Avg Loss: 1.47565504, Log Avg loss: 1.15445489, Global Avg Loss: 1.85024416, Time: 0.0402 Steps: 31640, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001844, Sample Num: 29504, Cur Loss: 1.35440111, Cur Avg Loss: 1.47645187, Log Avg loss: 1.62259079, Global Avg Loss: 1.85017223, Time: 0.0402 Steps: 31650, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001854, Sample Num: 29664, Cur Loss: 2.07695174, Cur Avg Loss: 1.47720339, Log Avg loss: 1.61578350, Global Avg Loss: 1.85009820, Time: 0.0402 Steps: 31660, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001864, Sample Num: 29824, Cur Loss: 0.72634792, Cur Avg Loss: 1.47643691, Log Avg loss: 1.33433089, Global Avg Loss: 1.84993534, Time: 0.0558 Steps: 31670, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001874, Sample Num: 29984, Cur Loss: 1.39038181, Cur Avg Loss: 1.47632404, Log Avg loss: 1.45528462, Global Avg Loss: 1.84981077, Time: 0.0673 Steps: 31680, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001884, Sample Num: 30144, Cur Loss: 1.80173206, Cur Avg Loss: 1.47623046, Log Avg loss: 1.45869420, Global Avg Loss: 1.84968735, Time: 0.0702 Steps: 31690, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001894, Sample Num: 30304, Cur Loss: 2.36120892, Cur Avg Loss: 1.47561755, Log Avg loss: 1.36014560, Global Avg Loss: 1.84953292, Time: 0.0887 Steps: 31700, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001904, Sample Num: 30464, Cur Loss: 1.71633363, Cur Avg Loss: 1.47548648, Log Avg loss: 1.45066144, Global Avg Loss: 1.84940713, Time: 0.0403 Steps: 31710, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001914, Sample Num: 30624, Cur Loss: 0.93450487, Cur Avg Loss: 1.47432073, Log Avg loss: 1.25236230, Global Avg Loss: 1.84921891, Time: 0.0638 Steps: 31720, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001924, Sample Num: 30784, Cur Loss: 1.63073683, Cur Avg Loss: 1.47375391, Log Avg loss: 1.36526507, Global Avg Loss: 1.84906639, Time: 0.1016 Steps: 31730, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001934, Sample Num: 30944, Cur Loss: 1.08455658, Cur Avg Loss: 1.47494979, Log Avg loss: 1.70503637, Global Avg Loss: 1.84902101, Time: 0.0690 Steps: 31740, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001944, Sample Num: 31104, Cur Loss: 3.19587040, Cur Avg Loss: 1.47673717, Log Avg loss: 1.82241576, Global Avg Loss: 1.84901263, Time: 0.0415 Steps: 31750, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001954, Sample Num: 31264, Cur Loss: 0.79849422, Cur Avg Loss: 1.47539032, Log Avg loss: 1.21356350, Global Avg Loss: 1.84881255, Time: 0.0478 Steps: 31760, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001964, Sample Num: 31424, Cur Loss: 0.49768341, Cur Avg Loss: 1.47472700, Log Avg loss: 1.34511305, Global Avg Loss: 1.84865401, Time: 0.0402 Steps: 31770, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001974, Sample Num: 31584, Cur Loss: 0.89483643, Cur Avg Loss: 1.47423186, Log Avg loss: 1.37698821, Global Avg Loss: 1.84850559, Time: 0.0403 Steps: 31780, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001984, Sample Num: 31744, Cur Loss: 1.39107072, Cur Avg Loss: 1.47476230, Log Avg loss: 1.57947071, Global Avg Loss: 1.84842096, Time: 0.0402 Steps: 31790, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001994, Sample Num: 31904, Cur Loss: 1.63574612, Cur Avg Loss: 1.47593095, Log Avg loss: 1.70779094, Global Avg Loss: 1.84837674, Time: 0.0403 Steps: 31800, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002004, Sample Num: 32064, Cur Loss: 0.39540297, Cur Avg Loss: 1.47682249, Log Avg loss: 1.65459602, Global Avg Loss: 1.84831582, Time: 0.0402 Steps: 31810, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002014, Sample Num: 32224, Cur Loss: 2.29042864, Cur Avg Loss: 1.47619239, Log Avg loss: 1.34991857, Global Avg Loss: 1.84815919, Time: 0.0402 Steps: 31820, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002024, Sample Num: 32384, Cur Loss: 0.97236586, Cur Avg Loss: 1.47692158, Log Avg loss: 1.62378176, Global Avg Loss: 1.84808870, Time: 0.0402 Steps: 31830, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002034, Sample Num: 32544, Cur Loss: 0.90932167, Cur Avg Loss: 1.47708452, Log Avg loss: 1.51006292, Global Avg Loss: 1.84798253, Time: 0.0403 Steps: 31840, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002044, Sample Num: 32704, Cur Loss: 1.24832606, Cur Avg Loss: 1.47881125, Log Avg loss: 1.83002828, Global Avg Loss: 1.84797690, Time: 0.0403 Steps: 31850, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002054, Sample Num: 32864, Cur Loss: 2.13438725, Cur Avg Loss: 1.48013134, Log Avg loss: 1.74995700, Global Avg Loss: 1.84794613, Time: 0.0403 Steps: 31860, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002064, Sample Num: 33024, Cur Loss: 1.06698966, Cur Avg Loss: 1.47836894, Log Avg loss: 1.11637284, Global Avg Loss: 1.84771658, Time: 0.0403 Steps: 31870, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002074, Sample Num: 33184, Cur Loss: 1.89866567, Cur Avg Loss: 1.47740233, Log Avg loss: 1.27789331, Global Avg Loss: 1.84753784, Time: 0.0402 Steps: 31880, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002084, Sample Num: 33344, Cur Loss: 0.85481143, Cur Avg Loss: 1.47557728, Log Avg loss: 1.09706245, Global Avg Loss: 1.84730251, Time: 0.0402 Steps: 31890, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002094, Sample Num: 33504, Cur Loss: 2.04961872, Cur Avg Loss: 1.47546441, Log Avg loss: 1.45194235, Global Avg Loss: 1.84717857, Time: 0.0403 Steps: 31900, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002104, Sample Num: 33664, Cur Loss: 1.14984465, Cur Avg Loss: 1.47773066, Log Avg loss: 1.95228323, Global Avg Loss: 1.84721151, Time: 0.0404 Steps: 31910, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002114, Sample Num: 33824, Cur Loss: 1.52955151, Cur Avg Loss: 1.47895755, Log Avg loss: 1.73709632, Global Avg Loss: 1.84717701, Time: 0.0403 Steps: 31920, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002124, Sample Num: 33984, Cur Loss: 1.56961846, Cur Avg Loss: 1.47966122, Log Avg loss: 1.62841628, Global Avg Loss: 1.84710850, Time: 0.0403 Steps: 31930, Updated lr: 0.000071 ***** Running evaluation checkpoint-31935 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-31935 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.235957, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.532897, "eval_total_loss": 1077.626327, "eval_mae": 1.109046, "eval_mse": 1.532693, "eval_r2": 0.02572, "eval_sp_statistic": 0.29529, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.376657, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.370019, "test_total_loss": 687.749582, "test_mae": 0.98717, "test_mse": 1.37024, "test_r2": 0.115635, "test_sp_statistic": 0.303212, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.403272, "test_ps_pvalue": 0.0, "lr": 7.066382171645329e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.8470245722772856, "train_cur_epoch_loss": 3149.3557522445917, "train_cur_epoch_avg_loss": 1.4792652664371027, "train_cur_epoch_time": 95.23595714569092, "train_cur_epoch_avg_time": 0.04473271824597976, "epoch": 15, "step": 31935} ################################################## Training, Epoch: 0016, Batch: 000005, Sample Num: 80, Cur Loss: 1.82147479, Cur Avg Loss: 2.02921717, Log Avg loss: 1.67014073, Global Avg Loss: 1.84705309, Time: 0.0403 Steps: 31940, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000015, Sample Num: 240, Cur Loss: 2.25325847, Cur Avg Loss: 1.80409404, Log Avg loss: 1.69153247, Global Avg Loss: 1.84700442, Time: 0.0404 Steps: 31950, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000025, Sample Num: 400, Cur Loss: 2.51416874, Cur Avg Loss: 1.57526603, Log Avg loss: 1.23202403, Global Avg Loss: 1.84681200, Time: 0.0403 Steps: 31960, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000035, Sample Num: 560, Cur Loss: 0.45596832, Cur Avg Loss: 1.50594530, Log Avg loss: 1.33264349, Global Avg Loss: 1.84665117, Time: 0.0403 Steps: 31970, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000045, Sample Num: 720, Cur Loss: 0.91074240, Cur Avg Loss: 1.47782146, Log Avg loss: 1.37938802, Global Avg Loss: 1.84650506, Time: 0.0402 Steps: 31980, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000055, Sample Num: 880, Cur Loss: 1.22742414, Cur Avg Loss: 1.49905591, Log Avg loss: 1.59461090, Global Avg Loss: 1.84642631, Time: 0.0402 Steps: 31990, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000065, Sample Num: 1040, Cur Loss: 0.60942185, Cur Avg Loss: 1.48269664, Log Avg loss: 1.39272065, Global Avg Loss: 1.84628453, Time: 0.0402 Steps: 32000, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000075, Sample Num: 1200, Cur Loss: 0.81272876, Cur Avg Loss: 1.44507769, Log Avg loss: 1.20055456, Global Avg Loss: 1.84608280, Time: 0.0402 Steps: 32010, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000085, Sample Num: 1360, Cur Loss: 1.11347628, Cur Avg Loss: 1.47406609, Log Avg loss: 1.69147904, Global Avg Loss: 1.84603452, Time: 0.0403 Steps: 32020, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000095, Sample Num: 1520, Cur Loss: 2.17665625, Cur Avg Loss: 1.45386329, Log Avg loss: 1.28213949, Global Avg Loss: 1.84585847, Time: 0.0403 Steps: 32030, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000105, Sample Num: 1680, Cur Loss: 1.38359392, Cur Avg Loss: 1.45766779, Log Avg loss: 1.49381058, Global Avg Loss: 1.84574859, Time: 0.0403 Steps: 32040, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000115, Sample Num: 1840, Cur Loss: 1.00157082, Cur Avg Loss: 1.47616789, Log Avg loss: 1.67041898, Global Avg Loss: 1.84569389, Time: 0.0403 Steps: 32050, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000125, Sample Num: 2000, Cur Loss: 1.62457252, Cur Avg Loss: 1.47947094, Log Avg loss: 1.51745592, Global Avg Loss: 1.84559150, Time: 0.0402 Steps: 32060, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000135, Sample Num: 2160, Cur Loss: 1.35518980, Cur Avg Loss: 1.48007267, Log Avg loss: 1.48759441, Global Avg Loss: 1.84547987, Time: 0.0403 Steps: 32070, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000145, Sample Num: 2320, Cur Loss: 1.87314200, Cur Avg Loss: 1.47009302, Log Avg loss: 1.33536767, Global Avg Loss: 1.84532086, Time: 0.0402 Steps: 32080, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000155, Sample Num: 2480, Cur Loss: 2.15703511, Cur Avg Loss: 1.48487606, Log Avg loss: 1.69923012, Global Avg Loss: 1.84527534, Time: 0.0562 Steps: 32090, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000165, Sample Num: 2640, Cur Loss: 0.78450203, Cur Avg Loss: 1.48128729, Log Avg loss: 1.42566139, Global Avg Loss: 1.84514461, Time: 0.0475 Steps: 32100, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000175, Sample Num: 2800, Cur Loss: 0.56098396, Cur Avg Loss: 1.47131206, Log Avg loss: 1.30672072, Global Avg Loss: 1.84497693, Time: 0.0511 Steps: 32110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000185, Sample Num: 2960, Cur Loss: 1.08320236, Cur Avg Loss: 1.47548080, Log Avg loss: 1.54843373, Global Avg Loss: 1.84488461, Time: 0.0995 Steps: 32120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000195, Sample Num: 3120, Cur Loss: 0.64242083, Cur Avg Loss: 1.44940306, Log Avg loss: 0.96696493, Global Avg Loss: 1.84461137, Time: 0.0402 Steps: 32130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000205, Sample Num: 3280, Cur Loss: 0.59368944, Cur Avg Loss: 1.44102425, Log Avg loss: 1.27763745, Global Avg Loss: 1.84443496, Time: 0.0997 Steps: 32140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000215, Sample Num: 3440, Cur Loss: 1.29632461, Cur Avg Loss: 1.45023150, Log Avg loss: 1.63898016, Global Avg Loss: 1.84437106, Time: 0.0533 Steps: 32150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000225, Sample Num: 3600, Cur Loss: 1.75292063, Cur Avg Loss: 1.46120058, Log Avg loss: 1.69703585, Global Avg Loss: 1.84432524, Time: 0.0455 Steps: 32160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000235, Sample Num: 3760, Cur Loss: 2.66058922, Cur Avg Loss: 1.48681977, Log Avg loss: 2.06325142, Global Avg Loss: 1.84439330, Time: 0.0405 Steps: 32170, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000245, Sample Num: 3920, Cur Loss: 0.65948278, Cur Avg Loss: 1.48424094, Log Avg loss: 1.42363842, Global Avg Loss: 1.84426255, Time: 0.0404 Steps: 32180, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000255, Sample Num: 4080, Cur Loss: 1.50548410, Cur Avg Loss: 1.48532570, Log Avg loss: 1.51190243, Global Avg Loss: 1.84415930, Time: 0.0402 Steps: 32190, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000265, Sample Num: 4240, Cur Loss: 2.85438728, Cur Avg Loss: 1.48562432, Log Avg loss: 1.49323917, Global Avg Loss: 1.84405032, Time: 0.0402 Steps: 32200, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000275, Sample Num: 4400, Cur Loss: 1.22700548, Cur Avg Loss: 1.49175832, Log Avg loss: 1.65430915, Global Avg Loss: 1.84399141, Time: 0.0402 Steps: 32210, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000285, Sample Num: 4560, Cur Loss: 1.34768677, Cur Avg Loss: 1.48996460, Log Avg loss: 1.44063747, Global Avg Loss: 1.84386622, Time: 0.0403 Steps: 32220, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000295, Sample Num: 4720, Cur Loss: 1.01462293, Cur Avg Loss: 1.48974936, Log Avg loss: 1.48361489, Global Avg Loss: 1.84375445, Time: 0.0403 Steps: 32230, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000305, Sample Num: 4880, Cur Loss: 1.24467301, Cur Avg Loss: 1.48795457, Log Avg loss: 1.43500821, Global Avg Loss: 1.84362766, Time: 0.0403 Steps: 32240, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000315, Sample Num: 5040, Cur Loss: 1.60748947, Cur Avg Loss: 1.49833532, Log Avg loss: 1.81494828, Global Avg Loss: 1.84361877, Time: 0.0402 Steps: 32250, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000325, Sample Num: 5200, Cur Loss: 2.19084406, Cur Avg Loss: 1.49566083, Log Avg loss: 1.41141444, Global Avg Loss: 1.84348479, Time: 0.0402 Steps: 32260, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000335, Sample Num: 5360, Cur Loss: 0.77754664, Cur Avg Loss: 1.48311433, Log Avg loss: 1.07535312, Global Avg Loss: 1.84324676, Time: 0.0403 Steps: 32270, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000345, Sample Num: 5520, Cur Loss: 2.27378607, Cur Avg Loss: 1.48681688, Log Avg loss: 1.61085213, Global Avg Loss: 1.84317477, Time: 0.0403 Steps: 32280, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000355, Sample Num: 5680, Cur Loss: 1.32486606, Cur Avg Loss: 1.47328486, Log Avg loss: 1.00643021, Global Avg Loss: 1.84291563, Time: 0.0403 Steps: 32290, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000365, Sample Num: 5840, Cur Loss: 2.19579792, Cur Avg Loss: 1.47204572, Log Avg loss: 1.42805637, Global Avg Loss: 1.84278720, Time: 0.0402 Steps: 32300, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000375, Sample Num: 6000, Cur Loss: 1.83215463, Cur Avg Loss: 1.46379725, Log Avg loss: 1.16272812, Global Avg Loss: 1.84257672, Time: 0.0403 Steps: 32310, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000385, Sample Num: 6160, Cur Loss: 1.35919189, Cur Avg Loss: 1.46144158, Log Avg loss: 1.37310385, Global Avg Loss: 1.84243146, Time: 0.0404 Steps: 32320, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000395, Sample Num: 6320, Cur Loss: 1.36188161, Cur Avg Loss: 1.46913594, Log Avg loss: 1.76536883, Global Avg Loss: 1.84240762, Time: 0.0403 Steps: 32330, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000405, Sample Num: 6480, Cur Loss: 1.13327718, Cur Avg Loss: 1.47442570, Log Avg loss: 1.68337106, Global Avg Loss: 1.84235845, Time: 0.0403 Steps: 32340, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000415, Sample Num: 6640, Cur Loss: 1.01814115, Cur Avg Loss: 1.47327891, Log Avg loss: 1.42683424, Global Avg Loss: 1.84223000, Time: 0.0402 Steps: 32350, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000425, Sample Num: 6800, Cur Loss: 1.81302428, Cur Avg Loss: 1.47282611, Log Avg loss: 1.45403486, Global Avg Loss: 1.84211004, Time: 0.0402 Steps: 32360, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000435, Sample Num: 6960, Cur Loss: 1.61773992, Cur Avg Loss: 1.47214949, Log Avg loss: 1.44339288, Global Avg Loss: 1.84198686, Time: 0.0402 Steps: 32370, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000445, Sample Num: 7120, Cur Loss: 0.73971605, Cur Avg Loss: 1.47265134, Log Avg loss: 1.49448209, Global Avg Loss: 1.84187954, Time: 0.0402 Steps: 32380, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000455, Sample Num: 7280, Cur Loss: 1.87070107, Cur Avg Loss: 1.47482279, Log Avg loss: 1.57145206, Global Avg Loss: 1.84179605, Time: 0.0402 Steps: 32390, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000465, Sample Num: 7440, Cur Loss: 2.03169823, Cur Avg Loss: 1.47530856, Log Avg loss: 1.49741136, Global Avg Loss: 1.84168976, Time: 0.0402 Steps: 32400, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000475, Sample Num: 7600, Cur Loss: 1.68051159, Cur Avg Loss: 1.47956893, Log Avg loss: 1.67767620, Global Avg Loss: 1.84163915, Time: 0.0402 Steps: 32410, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000485, Sample Num: 7760, Cur Loss: 0.92309278, Cur Avg Loss: 1.48205955, Log Avg loss: 1.60036356, Global Avg Loss: 1.84156473, Time: 0.0402 Steps: 32420, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000495, Sample Num: 7920, Cur Loss: 2.69329691, Cur Avg Loss: 1.49054625, Log Avg loss: 1.90215145, Global Avg Loss: 1.84158341, Time: 0.0402 Steps: 32430, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000505, Sample Num: 8080, Cur Loss: 2.32405114, Cur Avg Loss: 1.49093557, Log Avg loss: 1.51020694, Global Avg Loss: 1.84148126, Time: 0.0403 Steps: 32440, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000515, Sample Num: 8240, Cur Loss: 2.16578388, Cur Avg Loss: 1.49373422, Log Avg loss: 1.63506615, Global Avg Loss: 1.84141765, Time: 0.0629 Steps: 32450, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000525, Sample Num: 8400, Cur Loss: 0.75655085, Cur Avg Loss: 1.48686848, Log Avg loss: 1.13328295, Global Avg Loss: 1.84119950, Time: 0.0508 Steps: 32460, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000535, Sample Num: 8560, Cur Loss: 1.56173825, Cur Avg Loss: 1.48723562, Log Avg loss: 1.50651001, Global Avg Loss: 1.84109642, Time: 0.0607 Steps: 32470, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000545, Sample Num: 8720, Cur Loss: 0.76370394, Cur Avg Loss: 1.48722118, Log Avg loss: 1.48644861, Global Avg Loss: 1.84098723, Time: 0.0687 Steps: 32480, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000555, Sample Num: 8880, Cur Loss: 1.59384882, Cur Avg Loss: 1.48260301, Log Avg loss: 1.23091319, Global Avg Loss: 1.84079946, Time: 0.0477 Steps: 32490, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000565, Sample Num: 9040, Cur Loss: 0.54540002, Cur Avg Loss: 1.47869348, Log Avg loss: 1.26171430, Global Avg Loss: 1.84062128, Time: 0.0656 Steps: 32500, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000575, Sample Num: 9200, Cur Loss: 1.89339316, Cur Avg Loss: 1.48513384, Log Avg loss: 1.84901444, Global Avg Loss: 1.84062386, Time: 0.0664 Steps: 32510, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000585, Sample Num: 9360, Cur Loss: 0.88822854, Cur Avg Loss: 1.48476977, Log Avg loss: 1.46383586, Global Avg Loss: 1.84050800, Time: 0.0404 Steps: 32520, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000595, Sample Num: 9520, Cur Loss: 1.47115803, Cur Avg Loss: 1.48270432, Log Avg loss: 1.36187518, Global Avg Loss: 1.84036086, Time: 0.0405 Steps: 32530, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000605, Sample Num: 9680, Cur Loss: 0.79216051, Cur Avg Loss: 1.47912621, Log Avg loss: 1.26622854, Global Avg Loss: 1.84018442, Time: 0.0403 Steps: 32540, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000615, Sample Num: 9840, Cur Loss: 1.89009023, Cur Avg Loss: 1.48399985, Log Avg loss: 1.77885503, Global Avg Loss: 1.84016558, Time: 0.0403 Steps: 32550, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000625, Sample Num: 10000, Cur Loss: 1.64406204, Cur Avg Loss: 1.48026504, Log Avg loss: 1.25057412, Global Avg Loss: 1.83998450, Time: 0.0402 Steps: 32560, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000635, Sample Num: 10160, Cur Loss: 1.70573342, Cur Avg Loss: 1.48277731, Log Avg loss: 1.63979461, Global Avg Loss: 1.83992304, Time: 0.0402 Steps: 32570, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000645, Sample Num: 10320, Cur Loss: 0.41114098, Cur Avg Loss: 1.48396713, Log Avg loss: 1.55952080, Global Avg Loss: 1.83983697, Time: 0.0402 Steps: 32580, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000655, Sample Num: 10480, Cur Loss: 0.46040523, Cur Avg Loss: 1.48171866, Log Avg loss: 1.33669207, Global Avg Loss: 1.83968258, Time: 0.0403 Steps: 32590, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000665, Sample Num: 10640, Cur Loss: 1.91576886, Cur Avg Loss: 1.47751364, Log Avg loss: 1.20208491, Global Avg Loss: 1.83948700, Time: 0.0403 Steps: 32600, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000675, Sample Num: 10800, Cur Loss: 0.78044677, Cur Avg Loss: 1.47525770, Log Avg loss: 1.32523777, Global Avg Loss: 1.83932931, Time: 0.0403 Steps: 32610, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000685, Sample Num: 10960, Cur Loss: 1.10088539, Cur Avg Loss: 1.47441899, Log Avg loss: 1.41780580, Global Avg Loss: 1.83920008, Time: 0.0403 Steps: 32620, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000695, Sample Num: 11120, Cur Loss: 1.34290385, Cur Avg Loss: 1.47448211, Log Avg loss: 1.47880582, Global Avg Loss: 1.83908963, Time: 0.0402 Steps: 32630, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000705, Sample Num: 11280, Cur Loss: 0.62006307, Cur Avg Loss: 1.46790391, Log Avg loss: 1.01071882, Global Avg Loss: 1.83883584, Time: 0.0403 Steps: 32640, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000715, Sample Num: 11440, Cur Loss: 0.91572571, Cur Avg Loss: 1.46542899, Log Avg loss: 1.29094774, Global Avg Loss: 1.83866804, Time: 0.0402 Steps: 32650, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000725, Sample Num: 11600, Cur Loss: 1.53794086, Cur Avg Loss: 1.46766830, Log Avg loss: 1.62777879, Global Avg Loss: 1.83860347, Time: 0.0403 Steps: 32660, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000735, Sample Num: 11760, Cur Loss: 3.06341124, Cur Avg Loss: 1.47071506, Log Avg loss: 1.69160533, Global Avg Loss: 1.83855847, Time: 0.0403 Steps: 32670, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000745, Sample Num: 11920, Cur Loss: 1.52648222, Cur Avg Loss: 1.47357931, Log Avg loss: 1.68410153, Global Avg Loss: 1.83851121, Time: 0.0403 Steps: 32680, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000755, Sample Num: 12080, Cur Loss: 1.60782146, Cur Avg Loss: 1.47288192, Log Avg loss: 1.42092595, Global Avg Loss: 1.83838347, Time: 0.0403 Steps: 32690, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000765, Sample Num: 12240, Cur Loss: 1.09133768, Cur Avg Loss: 1.46927818, Log Avg loss: 1.19719597, Global Avg Loss: 1.83818739, Time: 0.0403 Steps: 32700, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000775, Sample Num: 12400, Cur Loss: 1.00730872, Cur Avg Loss: 1.47162058, Log Avg loss: 1.65081469, Global Avg Loss: 1.83813010, Time: 0.0403 Steps: 32710, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000785, Sample Num: 12560, Cur Loss: 1.70902300, Cur Avg Loss: 1.47163395, Log Avg loss: 1.47266972, Global Avg Loss: 1.83801841, Time: 0.0403 Steps: 32720, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000795, Sample Num: 12720, Cur Loss: 1.08372128, Cur Avg Loss: 1.47336005, Log Avg loss: 1.60885919, Global Avg Loss: 1.83794839, Time: 0.0403 Steps: 32730, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000805, Sample Num: 12880, Cur Loss: 1.37552905, Cur Avg Loss: 1.47898876, Log Avg loss: 1.92647128, Global Avg Loss: 1.83797543, Time: 0.0404 Steps: 32740, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000815, Sample Num: 13040, Cur Loss: 1.37655544, Cur Avg Loss: 1.48102607, Log Avg loss: 1.64502895, Global Avg Loss: 1.83791652, Time: 0.0404 Steps: 32750, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000825, Sample Num: 13200, Cur Loss: 2.53008270, Cur Avg Loss: 1.48452627, Log Avg loss: 1.76979258, Global Avg Loss: 1.83789572, Time: 0.0403 Steps: 32760, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000835, Sample Num: 13360, Cur Loss: 1.07240379, Cur Avg Loss: 1.48599062, Log Avg loss: 1.60679988, Global Avg Loss: 1.83782520, Time: 0.0403 Steps: 32770, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000845, Sample Num: 13520, Cur Loss: 1.86852741, Cur Avg Loss: 1.48470596, Log Avg loss: 1.37743653, Global Avg Loss: 1.83768475, Time: 0.0403 Steps: 32780, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000855, Sample Num: 13680, Cur Loss: 0.94693160, Cur Avg Loss: 1.48615394, Log Avg loss: 1.60850857, Global Avg Loss: 1.83761486, Time: 0.0733 Steps: 32790, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000865, Sample Num: 13840, Cur Loss: 1.34756494, Cur Avg Loss: 1.48484978, Log Avg loss: 1.37334361, Global Avg Loss: 1.83747332, Time: 0.0407 Steps: 32800, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000875, Sample Num: 14000, Cur Loss: 1.54717982, Cur Avg Loss: 1.48721601, Log Avg loss: 1.69189517, Global Avg Loss: 1.83742895, Time: 0.0802 Steps: 32810, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000885, Sample Num: 14160, Cur Loss: 1.71837258, Cur Avg Loss: 1.48519286, Log Avg loss: 1.30816753, Global Avg Loss: 1.83726768, Time: 0.0610 Steps: 32820, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000895, Sample Num: 14320, Cur Loss: 0.91524315, Cur Avg Loss: 1.48616892, Log Avg loss: 1.57254982, Global Avg Loss: 1.83718705, Time: 0.0589 Steps: 32830, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000905, Sample Num: 14480, Cur Loss: 0.84658492, Cur Avg Loss: 1.48455833, Log Avg loss: 1.34041044, Global Avg Loss: 1.83703578, Time: 0.0535 Steps: 32840, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000915, Sample Num: 14640, Cur Loss: 1.95107532, Cur Avg Loss: 1.48500194, Log Avg loss: 1.52514925, Global Avg Loss: 1.83694084, Time: 0.0407 Steps: 32850, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000925, Sample Num: 14800, Cur Loss: 3.13339329, Cur Avg Loss: 1.49174172, Log Avg loss: 2.10843130, Global Avg Loss: 1.83702346, Time: 0.0404 Steps: 32860, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000935, Sample Num: 14960, Cur Loss: 2.81430769, Cur Avg Loss: 1.49321758, Log Avg loss: 1.62973447, Global Avg Loss: 1.83696039, Time: 0.0402 Steps: 32870, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000945, Sample Num: 15120, Cur Loss: 1.45639443, Cur Avg Loss: 1.49489307, Log Avg loss: 1.65155116, Global Avg Loss: 1.83690400, Time: 0.0403 Steps: 32880, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000955, Sample Num: 15280, Cur Loss: 1.77671862, Cur Avg Loss: 1.50069632, Log Avg loss: 2.04910386, Global Avg Loss: 1.83696852, Time: 0.0402 Steps: 32890, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000965, Sample Num: 15440, Cur Loss: 2.66511035, Cur Avg Loss: 1.50206747, Log Avg loss: 1.63301216, Global Avg Loss: 1.83690653, Time: 0.0404 Steps: 32900, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000975, Sample Num: 15600, Cur Loss: 1.15649629, Cur Avg Loss: 1.49992084, Log Avg loss: 1.29277127, Global Avg Loss: 1.83674119, Time: 0.0403 Steps: 32910, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000985, Sample Num: 15760, Cur Loss: 2.41539383, Cur Avg Loss: 1.50086045, Log Avg loss: 1.59247229, Global Avg Loss: 1.83666699, Time: 0.0404 Steps: 32920, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000995, Sample Num: 15920, Cur Loss: 0.99801242, Cur Avg Loss: 1.49764347, Log Avg loss: 1.18077080, Global Avg Loss: 1.83646781, Time: 0.0402 Steps: 32930, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001005, Sample Num: 16080, Cur Loss: 1.13720191, Cur Avg Loss: 1.49992074, Log Avg loss: 1.72650962, Global Avg Loss: 1.83643443, Time: 0.0403 Steps: 32940, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001015, Sample Num: 16240, Cur Loss: 1.11396313, Cur Avg Loss: 1.49944322, Log Avg loss: 1.45145165, Global Avg Loss: 1.83631759, Time: 0.0404 Steps: 32950, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001025, Sample Num: 16400, Cur Loss: 2.19184446, Cur Avg Loss: 1.50016593, Log Avg loss: 1.57352178, Global Avg Loss: 1.83623786, Time: 0.0404 Steps: 32960, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001035, Sample Num: 16560, Cur Loss: 1.30962396, Cur Avg Loss: 1.49674088, Log Avg loss: 1.14567297, Global Avg Loss: 1.83602841, Time: 0.0403 Steps: 32970, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001045, Sample Num: 16720, Cur Loss: 1.80519962, Cur Avg Loss: 1.49380495, Log Avg loss: 1.18993614, Global Avg Loss: 1.83583250, Time: 0.0402 Steps: 32980, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001055, Sample Num: 16880, Cur Loss: 1.40931296, Cur Avg Loss: 1.49491022, Log Avg loss: 1.61041123, Global Avg Loss: 1.83576417, Time: 0.0403 Steps: 32990, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001065, Sample Num: 17040, Cur Loss: 2.03663158, Cur Avg Loss: 1.49446356, Log Avg loss: 1.44734088, Global Avg Loss: 1.83564647, Time: 0.0403 Steps: 33000, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001075, Sample Num: 17200, Cur Loss: 1.34186900, Cur Avg Loss: 1.49055931, Log Avg loss: 1.07475638, Global Avg Loss: 1.83541596, Time: 0.0403 Steps: 33010, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001085, Sample Num: 17360, Cur Loss: 1.89136815, Cur Avg Loss: 1.49001284, Log Avg loss: 1.43126765, Global Avg Loss: 1.83529357, Time: 0.0403 Steps: 33020, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001095, Sample Num: 17520, Cur Loss: 0.82167828, Cur Avg Loss: 1.48822689, Log Avg loss: 1.29445103, Global Avg Loss: 1.83512983, Time: 0.0402 Steps: 33030, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001105, Sample Num: 17680, Cur Loss: 1.67965138, Cur Avg Loss: 1.48848206, Log Avg loss: 1.51642267, Global Avg Loss: 1.83503337, Time: 0.0402 Steps: 33040, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001115, Sample Num: 17840, Cur Loss: 1.59927559, Cur Avg Loss: 1.48868980, Log Avg loss: 1.51164607, Global Avg Loss: 1.83493552, Time: 0.0402 Steps: 33050, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001125, Sample Num: 18000, Cur Loss: 1.68143773, Cur Avg Loss: 1.49050262, Log Avg loss: 1.69263114, Global Avg Loss: 1.83489247, Time: 0.0404 Steps: 33060, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001135, Sample Num: 18160, Cur Loss: 1.82434320, Cur Avg Loss: 1.49155969, Log Avg loss: 1.61048082, Global Avg Loss: 1.83482461, Time: 0.0403 Steps: 33070, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001145, Sample Num: 18320, Cur Loss: 1.97449732, Cur Avg Loss: 1.49004340, Log Avg loss: 1.31794452, Global Avg Loss: 1.83466836, Time: 0.0402 Steps: 33080, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001155, Sample Num: 18480, Cur Loss: 0.94885474, Cur Avg Loss: 1.48803225, Log Avg loss: 1.25775505, Global Avg Loss: 1.83449402, Time: 0.0402 Steps: 33090, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001165, Sample Num: 18640, Cur Loss: 2.35742021, Cur Avg Loss: 1.48917933, Log Avg loss: 1.62166760, Global Avg Loss: 1.83442972, Time: 0.0403 Steps: 33100, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001175, Sample Num: 18800, Cur Loss: 0.55847365, Cur Avg Loss: 1.48576865, Log Avg loss: 1.08842386, Global Avg Loss: 1.83420441, Time: 0.0405 Steps: 33110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001185, Sample Num: 18960, Cur Loss: 1.26278186, Cur Avg Loss: 1.48980194, Log Avg loss: 1.96371336, Global Avg Loss: 1.83424351, Time: 0.0486 Steps: 33120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001195, Sample Num: 19120, Cur Loss: 1.62072563, Cur Avg Loss: 1.49023860, Log Avg loss: 1.54198331, Global Avg Loss: 1.83415529, Time: 0.0833 Steps: 33130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001205, Sample Num: 19280, Cur Loss: 1.26353920, Cur Avg Loss: 1.49268521, Log Avg loss: 1.78505554, Global Avg Loss: 1.83414048, Time: 0.0404 Steps: 33140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001215, Sample Num: 19440, Cur Loss: 0.86335528, Cur Avg Loss: 1.49263157, Log Avg loss: 1.48616677, Global Avg Loss: 1.83403551, Time: 0.0590 Steps: 33150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001225, Sample Num: 19600, Cur Loss: 0.60473704, Cur Avg Loss: 1.49218047, Log Avg loss: 1.43737284, Global Avg Loss: 1.83391589, Time: 0.0406 Steps: 33160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001235, Sample Num: 19760, Cur Loss: 0.83063853, Cur Avg Loss: 1.48777165, Log Avg loss: 0.94769078, Global Avg Loss: 1.83364871, Time: 0.0729 Steps: 33170, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001245, Sample Num: 19920, Cur Loss: 1.32195854, Cur Avg Loss: 1.48801760, Log Avg loss: 1.51839185, Global Avg Loss: 1.83355370, Time: 0.0404 Steps: 33180, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001255, Sample Num: 20080, Cur Loss: 1.77809048, Cur Avg Loss: 1.48796547, Log Avg loss: 1.48147620, Global Avg Loss: 1.83344762, Time: 0.0701 Steps: 33190, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001265, Sample Num: 20240, Cur Loss: 1.43035924, Cur Avg Loss: 1.49158652, Log Avg loss: 1.94602818, Global Avg Loss: 1.83348153, Time: 0.0403 Steps: 33200, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001275, Sample Num: 20400, Cur Loss: 2.27559805, Cur Avg Loss: 1.49310542, Log Avg loss: 1.68524644, Global Avg Loss: 1.83343689, Time: 0.0403 Steps: 33210, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001285, Sample Num: 20560, Cur Loss: 1.21830988, Cur Avg Loss: 1.49389018, Log Avg loss: 1.59394646, Global Avg Loss: 1.83336480, Time: 0.0403 Steps: 33220, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001295, Sample Num: 20720, Cur Loss: 0.93926692, Cur Avg Loss: 1.49474858, Log Avg loss: 1.60505304, Global Avg Loss: 1.83329609, Time: 0.0404 Steps: 33230, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001305, Sample Num: 20880, Cur Loss: 1.87852859, Cur Avg Loss: 1.49430064, Log Avg loss: 1.43629215, Global Avg Loss: 1.83317666, Time: 0.0402 Steps: 33240, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001315, Sample Num: 21040, Cur Loss: 0.85605532, Cur Avg Loss: 1.49387311, Log Avg loss: 1.43808032, Global Avg Loss: 1.83305783, Time: 0.0403 Steps: 33250, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001325, Sample Num: 21200, Cur Loss: 0.61943519, Cur Avg Loss: 1.49373154, Log Avg loss: 1.47511602, Global Avg Loss: 1.83295021, Time: 0.0402 Steps: 33260, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001335, Sample Num: 21360, Cur Loss: 1.84765577, Cur Avg Loss: 1.49443778, Log Avg loss: 1.58801420, Global Avg Loss: 1.83287659, Time: 0.0403 Steps: 33270, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001345, Sample Num: 21520, Cur Loss: 1.93360317, Cur Avg Loss: 1.49454206, Log Avg loss: 1.50846312, Global Avg Loss: 1.83277911, Time: 0.0404 Steps: 33280, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001355, Sample Num: 21680, Cur Loss: 0.92244440, Cur Avg Loss: 1.49498917, Log Avg loss: 1.55512596, Global Avg Loss: 1.83269571, Time: 0.0402 Steps: 33290, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001365, Sample Num: 21840, Cur Loss: 0.85445356, Cur Avg Loss: 1.49576577, Log Avg loss: 1.60099454, Global Avg Loss: 1.83262613, Time: 0.0403 Steps: 33300, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001375, Sample Num: 22000, Cur Loss: 1.49126339, Cur Avg Loss: 1.49549323, Log Avg loss: 1.45829126, Global Avg Loss: 1.83251375, Time: 0.0403 Steps: 33310, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001385, Sample Num: 22160, Cur Loss: 0.80505633, Cur Avg Loss: 1.49526876, Log Avg loss: 1.46440482, Global Avg Loss: 1.83240327, Time: 0.0403 Steps: 33320, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001395, Sample Num: 22320, Cur Loss: 2.09707689, Cur Avg Loss: 1.49846477, Log Avg loss: 1.94111154, Global Avg Loss: 1.83243589, Time: 0.0403 Steps: 33330, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001405, Sample Num: 22480, Cur Loss: 0.95595211, Cur Avg Loss: 1.49790015, Log Avg loss: 1.41913664, Global Avg Loss: 1.83231192, Time: 0.0403 Steps: 33340, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001415, Sample Num: 22640, Cur Loss: 1.41049135, Cur Avg Loss: 1.49710899, Log Avg loss: 1.38595070, Global Avg Loss: 1.83217808, Time: 0.0402 Steps: 33350, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001425, Sample Num: 22800, Cur Loss: 1.06405330, Cur Avg Loss: 1.49940119, Log Avg loss: 1.82374698, Global Avg Loss: 1.83217555, Time: 0.0403 Steps: 33360, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001435, Sample Num: 22960, Cur Loss: 2.38910055, Cur Avg Loss: 1.49946183, Log Avg loss: 1.50810335, Global Avg Loss: 1.83207844, Time: 0.0403 Steps: 33370, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001445, Sample Num: 23120, Cur Loss: 0.99826914, Cur Avg Loss: 1.49690279, Log Avg loss: 1.12968115, Global Avg Loss: 1.83186801, Time: 0.0403 Steps: 33380, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001455, Sample Num: 23280, Cur Loss: 0.60130626, Cur Avg Loss: 1.49592102, Log Avg loss: 1.35405443, Global Avg Loss: 1.83172491, Time: 0.0402 Steps: 33390, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001465, Sample Num: 23440, Cur Loss: 1.51117229, Cur Avg Loss: 1.49601818, Log Avg loss: 1.51015521, Global Avg Loss: 1.83162863, Time: 0.0402 Steps: 33400, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001475, Sample Num: 23600, Cur Loss: 1.24284315, Cur Avg Loss: 1.49742274, Log Avg loss: 1.70319103, Global Avg Loss: 1.83159019, Time: 0.0403 Steps: 33410, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001485, Sample Num: 23760, Cur Loss: 1.65244770, Cur Avg Loss: 1.49798010, Log Avg loss: 1.58019089, Global Avg Loss: 1.83151497, Time: 0.0402 Steps: 33420, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001495, Sample Num: 23920, Cur Loss: 0.46701312, Cur Avg Loss: 1.49785809, Log Avg loss: 1.47973927, Global Avg Loss: 1.83140974, Time: 0.0403 Steps: 33430, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001505, Sample Num: 24080, Cur Loss: 1.45462608, Cur Avg Loss: 1.49642289, Log Avg loss: 1.28186025, Global Avg Loss: 1.83124540, Time: 0.0403 Steps: 33440, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001515, Sample Num: 24240, Cur Loss: 2.11323643, Cur Avg Loss: 1.49850434, Log Avg loss: 1.81176327, Global Avg Loss: 1.83123958, Time: 0.0482 Steps: 33450, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001525, Sample Num: 24400, Cur Loss: 1.91729093, Cur Avg Loss: 1.49581807, Log Avg loss: 1.08884690, Global Avg Loss: 1.83101770, Time: 0.0991 Steps: 33460, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001535, Sample Num: 24560, Cur Loss: 1.60388386, Cur Avg Loss: 1.49542477, Log Avg loss: 1.43544772, Global Avg Loss: 1.83089951, Time: 0.0771 Steps: 33470, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001545, Sample Num: 24720, Cur Loss: 0.72670186, Cur Avg Loss: 1.49632155, Log Avg loss: 1.63397712, Global Avg Loss: 1.83084070, Time: 0.0994 Steps: 33480, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001555, Sample Num: 24880, Cur Loss: 0.62619054, Cur Avg Loss: 1.49757490, Log Avg loss: 1.69121789, Global Avg Loss: 1.83079901, Time: 0.0415 Steps: 33490, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001565, Sample Num: 25040, Cur Loss: 1.80616355, Cur Avg Loss: 1.49693058, Log Avg loss: 1.39673812, Global Avg Loss: 1.83066944, Time: 0.0560 Steps: 33500, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001575, Sample Num: 25200, Cur Loss: 1.80348289, Cur Avg Loss: 1.49670823, Log Avg loss: 1.46191038, Global Avg Loss: 1.83055939, Time: 0.0422 Steps: 33510, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001585, Sample Num: 25360, Cur Loss: 2.12355709, Cur Avg Loss: 1.49557406, Log Avg loss: 1.31694214, Global Avg Loss: 1.83040616, Time: 0.0532 Steps: 33520, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001595, Sample Num: 25520, Cur Loss: 1.08777165, Cur Avg Loss: 1.49605618, Log Avg loss: 1.57247275, Global Avg Loss: 1.83032924, Time: 0.0590 Steps: 33530, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001605, Sample Num: 25680, Cur Loss: 1.94769812, Cur Avg Loss: 1.49791068, Log Avg loss: 1.79370307, Global Avg Loss: 1.83031832, Time: 0.0402 Steps: 33540, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001615, Sample Num: 25840, Cur Loss: 2.40000796, Cur Avg Loss: 1.49784820, Log Avg loss: 1.48781970, Global Avg Loss: 1.83021623, Time: 0.0402 Steps: 33550, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001625, Sample Num: 26000, Cur Loss: 2.82143736, Cur Avg Loss: 1.49781313, Log Avg loss: 1.49215026, Global Avg Loss: 1.83011550, Time: 0.0403 Steps: 33560, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001635, Sample Num: 26160, Cur Loss: 0.64508539, Cur Avg Loss: 1.49627655, Log Avg loss: 1.24658190, Global Avg Loss: 1.82994167, Time: 0.0403 Steps: 33570, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001645, Sample Num: 26320, Cur Loss: 2.13932467, Cur Avg Loss: 1.49599917, Log Avg loss: 1.45064766, Global Avg Loss: 1.82982872, Time: 0.0402 Steps: 33580, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001655, Sample Num: 26480, Cur Loss: 1.13540912, Cur Avg Loss: 1.49557946, Log Avg loss: 1.42653692, Global Avg Loss: 1.82970865, Time: 0.0402 Steps: 33590, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001665, Sample Num: 26640, Cur Loss: 1.60627222, Cur Avg Loss: 1.49474476, Log Avg loss: 1.35660131, Global Avg Loss: 1.82956785, Time: 0.0402 Steps: 33600, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001675, Sample Num: 26800, Cur Loss: 1.94615579, Cur Avg Loss: 1.49598732, Log Avg loss: 1.70287474, Global Avg Loss: 1.82953015, Time: 0.0403 Steps: 33610, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001685, Sample Num: 26960, Cur Loss: 2.23137736, Cur Avg Loss: 1.49511778, Log Avg loss: 1.34946896, Global Avg Loss: 1.82938736, Time: 0.0402 Steps: 33620, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001695, Sample Num: 27120, Cur Loss: 2.45897269, Cur Avg Loss: 1.49396848, Log Avg loss: 1.30031136, Global Avg Loss: 1.82923004, Time: 0.0402 Steps: 33630, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001705, Sample Num: 27280, Cur Loss: 1.65045714, Cur Avg Loss: 1.49659050, Log Avg loss: 1.94102364, Global Avg Loss: 1.82926327, Time: 0.0403 Steps: 33640, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001715, Sample Num: 27440, Cur Loss: 2.03141999, Cur Avg Loss: 1.49807119, Log Avg loss: 1.75052773, Global Avg Loss: 1.82923988, Time: 0.0402 Steps: 33650, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001725, Sample Num: 27600, Cur Loss: 1.01724887, Cur Avg Loss: 1.49824946, Log Avg loss: 1.52882373, Global Avg Loss: 1.82915062, Time: 0.0403 Steps: 33660, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001735, Sample Num: 27760, Cur Loss: 0.81896126, Cur Avg Loss: 1.49602926, Log Avg loss: 1.11304446, Global Avg Loss: 1.82893794, Time: 0.0402 Steps: 33670, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001745, Sample Num: 27920, Cur Loss: 1.39193869, Cur Avg Loss: 1.49553108, Log Avg loss: 1.40909752, Global Avg Loss: 1.82881329, Time: 0.0403 Steps: 33680, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001755, Sample Num: 28080, Cur Loss: 1.96960437, Cur Avg Loss: 1.49610144, Log Avg loss: 1.59562890, Global Avg Loss: 1.82874407, Time: 0.0403 Steps: 33690, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001765, Sample Num: 28240, Cur Loss: 1.68138599, Cur Avg Loss: 1.49627835, Log Avg loss: 1.52732586, Global Avg Loss: 1.82865463, Time: 0.0403 Steps: 33700, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001775, Sample Num: 28400, Cur Loss: 1.06256652, Cur Avg Loss: 1.49681953, Log Avg loss: 1.59233755, Global Avg Loss: 1.82858453, Time: 0.0403 Steps: 33710, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001785, Sample Num: 28560, Cur Loss: 1.45888317, Cur Avg Loss: 1.49700197, Log Avg loss: 1.52938505, Global Avg Loss: 1.82849580, Time: 0.0403 Steps: 33720, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001795, Sample Num: 28720, Cur Loss: 1.29082620, Cur Avg Loss: 1.49600370, Log Avg loss: 1.31781291, Global Avg Loss: 1.82834439, Time: 0.0403 Steps: 33730, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001805, Sample Num: 28880, Cur Loss: 0.38464302, Cur Avg Loss: 1.49390533, Log Avg loss: 1.11724719, Global Avg Loss: 1.82813363, Time: 0.0403 Steps: 33740, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001815, Sample Num: 29040, Cur Loss: 1.09185839, Cur Avg Loss: 1.49563049, Log Avg loss: 1.80702194, Global Avg Loss: 1.82812738, Time: 0.0402 Steps: 33750, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001825, Sample Num: 29200, Cur Loss: 1.79628491, Cur Avg Loss: 1.49423055, Log Avg loss: 1.24014241, Global Avg Loss: 1.82795321, Time: 0.0402 Steps: 33760, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001835, Sample Num: 29360, Cur Loss: 0.69692320, Cur Avg Loss: 1.49499174, Log Avg loss: 1.63390820, Global Avg Loss: 1.82789575, Time: 0.0403 Steps: 33770, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001845, Sample Num: 29520, Cur Loss: 1.62847304, Cur Avg Loss: 1.49497413, Log Avg loss: 1.49174260, Global Avg Loss: 1.82779624, Time: 0.0403 Steps: 33780, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001855, Sample Num: 29680, Cur Loss: 1.70852530, Cur Avg Loss: 1.49504019, Log Avg loss: 1.50722831, Global Avg Loss: 1.82770137, Time: 0.0403 Steps: 33790, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001865, Sample Num: 29840, Cur Loss: 0.93427861, Cur Avg Loss: 1.49661936, Log Avg loss: 1.78955503, Global Avg Loss: 1.82769008, Time: 0.0593 Steps: 33800, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001875, Sample Num: 30000, Cur Loss: 1.86388767, Cur Avg Loss: 1.49653331, Log Avg loss: 1.48048481, Global Avg Loss: 1.82758739, Time: 0.0548 Steps: 33810, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001885, Sample Num: 30160, Cur Loss: 1.16500390, Cur Avg Loss: 1.49555174, Log Avg loss: 1.31150843, Global Avg Loss: 1.82743479, Time: 0.0403 Steps: 33820, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001895, Sample Num: 30320, Cur Loss: 0.99471718, Cur Avg Loss: 1.49424993, Log Avg loss: 1.24885897, Global Avg Loss: 1.82726377, Time: 0.0683 Steps: 33830, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001905, Sample Num: 30480, Cur Loss: 1.06108880, Cur Avg Loss: 1.49595239, Log Avg loss: 1.81856768, Global Avg Loss: 1.82726120, Time: 0.0991 Steps: 33840, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001915, Sample Num: 30640, Cur Loss: 2.17918110, Cur Avg Loss: 1.49678352, Log Avg loss: 1.65511452, Global Avg Loss: 1.82721034, Time: 0.0489 Steps: 33850, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001925, Sample Num: 30800, Cur Loss: 1.17650676, Cur Avg Loss: 1.49835374, Log Avg loss: 1.79905072, Global Avg Loss: 1.82720203, Time: 0.0404 Steps: 33860, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001935, Sample Num: 30960, Cur Loss: 2.12734628, Cur Avg Loss: 1.49836980, Log Avg loss: 1.50146109, Global Avg Loss: 1.82710585, Time: 0.0644 Steps: 33870, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001945, Sample Num: 31120, Cur Loss: 1.48192143, Cur Avg Loss: 1.49782757, Log Avg loss: 1.39290636, Global Avg Loss: 1.82697770, Time: 0.0487 Steps: 33880, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001955, Sample Num: 31280, Cur Loss: 1.52643466, Cur Avg Loss: 1.49668572, Log Avg loss: 1.27459599, Global Avg Loss: 1.82681470, Time: 0.0403 Steps: 33890, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001965, Sample Num: 31440, Cur Loss: 3.08188963, Cur Avg Loss: 1.49788616, Log Avg loss: 1.73257170, Global Avg Loss: 1.82678690, Time: 0.0403 Steps: 33900, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001975, Sample Num: 31600, Cur Loss: 1.29173887, Cur Avg Loss: 1.49848697, Log Avg loss: 1.61654595, Global Avg Loss: 1.82672490, Time: 0.0403 Steps: 33910, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001985, Sample Num: 31760, Cur Loss: 2.99337435, Cur Avg Loss: 1.49971814, Log Avg loss: 1.74287544, Global Avg Loss: 1.82670018, Time: 0.0404 Steps: 33920, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001995, Sample Num: 31920, Cur Loss: 1.71262896, Cur Avg Loss: 1.50044043, Log Avg loss: 1.64381385, Global Avg Loss: 1.82664628, Time: 0.0403 Steps: 33930, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002005, Sample Num: 32080, Cur Loss: 1.64534092, Cur Avg Loss: 1.50164200, Log Avg loss: 1.74135449, Global Avg Loss: 1.82662115, Time: 0.0403 Steps: 33940, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002015, Sample Num: 32240, Cur Loss: 0.84912288, Cur Avg Loss: 1.50143620, Log Avg loss: 1.46017433, Global Avg Loss: 1.82651322, Time: 0.0402 Steps: 33950, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002025, Sample Num: 32400, Cur Loss: 1.66745496, Cur Avg Loss: 1.50221093, Log Avg loss: 1.65831949, Global Avg Loss: 1.82646369, Time: 0.0403 Steps: 33960, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002035, Sample Num: 32560, Cur Loss: 1.13158393, Cur Avg Loss: 1.50126015, Log Avg loss: 1.30872673, Global Avg Loss: 1.82631128, Time: 0.0403 Steps: 33970, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002045, Sample Num: 32720, Cur Loss: 1.18413603, Cur Avg Loss: 1.49967411, Log Avg loss: 1.17691460, Global Avg Loss: 1.82612017, Time: 0.0403 Steps: 33980, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002055, Sample Num: 32880, Cur Loss: 1.14442563, Cur Avg Loss: 1.49868838, Log Avg loss: 1.29710592, Global Avg Loss: 1.82596453, Time: 0.0403 Steps: 33990, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002065, Sample Num: 33040, Cur Loss: 1.25486374, Cur Avg Loss: 1.49676243, Log Avg loss: 1.10097969, Global Avg Loss: 1.82575130, Time: 0.0403 Steps: 34000, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002075, Sample Num: 33200, Cur Loss: 0.97050512, Cur Avg Loss: 1.49733599, Log Avg loss: 1.61577761, Global Avg Loss: 1.82568956, Time: 0.0402 Steps: 34010, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002085, Sample Num: 33360, Cur Loss: 1.52831793, Cur Avg Loss: 1.49777855, Log Avg loss: 1.58960840, Global Avg Loss: 1.82562016, Time: 0.0403 Steps: 34020, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002095, Sample Num: 33520, Cur Loss: 0.95393938, Cur Avg Loss: 1.49784236, Log Avg loss: 1.51114710, Global Avg Loss: 1.82552775, Time: 0.0403 Steps: 34030, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002105, Sample Num: 33680, Cur Loss: 1.48737514, Cur Avg Loss: 1.49773226, Log Avg loss: 1.47466735, Global Avg Loss: 1.82542468, Time: 0.0404 Steps: 34040, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002115, Sample Num: 33840, Cur Loss: 1.03437781, Cur Avg Loss: 1.49781914, Log Avg loss: 1.51610693, Global Avg Loss: 1.82533384, Time: 0.0403 Steps: 34050, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002125, Sample Num: 34000, Cur Loss: 1.88306355, Cur Avg Loss: 1.49588530, Log Avg loss: 1.08687733, Global Avg Loss: 1.82511703, Time: 0.0403 Steps: 34060, Updated lr: 0.000069 ***** Running evaluation checkpoint-34064 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-34064 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.079904, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.482088, "eval_total_loss": 1041.907678, "eval_mae": 0.857985, "eval_mse": 1.482568, "eval_r2": 0.057583, "eval_sp_statistic": 0.279542, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.383775, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.435894, "test_total_loss": 720.819014, "test_mae": 0.741607, "test_mse": 1.43639, "test_r2": 0.072941, "test_sp_statistic": 0.291084, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.401998, "test_ps_pvalue": 0.0, "lr": 6.86448553816975e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.8251655492494159, "train_cur_epoch_loss": 3187.7095539569855, "train_cur_epoch_avg_loss": 1.4972802038313695, "train_cur_epoch_time": 95.07990384101868, "train_cur_epoch_avg_time": 0.044659419371075, "epoch": 16, "step": 34064} ################################################## Training, Epoch: 0017, Batch: 000006, Sample Num: 96, Cur Loss: 0.80956125, Cur Avg Loss: 1.50833177, Log Avg loss: 1.80032898, Global Avg Loss: 1.82510975, Time: 0.0403 Steps: 34070, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000016, Sample Num: 256, Cur Loss: 1.53467035, Cur Avg Loss: 1.39816583, Log Avg loss: 1.33206626, Global Avg Loss: 1.82496508, Time: 0.0403 Steps: 34080, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000026, Sample Num: 416, Cur Loss: 0.56272650, Cur Avg Loss: 1.45081551, Log Avg loss: 1.53505499, Global Avg Loss: 1.82488004, Time: 0.0403 Steps: 34090, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000036, Sample Num: 576, Cur Loss: 1.00480175, Cur Avg Loss: 1.39368549, Log Avg loss: 1.24514746, Global Avg Loss: 1.82471003, Time: 0.0404 Steps: 34100, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000046, Sample Num: 736, Cur Loss: 1.00379372, Cur Avg Loss: 1.39942829, Log Avg loss: 1.42010238, Global Avg Loss: 1.82459141, Time: 0.0403 Steps: 34110, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000056, Sample Num: 896, Cur Loss: 2.21331882, Cur Avg Loss: 1.49245953, Log Avg loss: 1.92040321, Global Avg Loss: 1.82461949, Time: 0.0402 Steps: 34120, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000066, Sample Num: 1056, Cur Loss: 1.43010962, Cur Avg Loss: 1.48919470, Log Avg loss: 1.47091166, Global Avg Loss: 1.82451585, Time: 0.0404 Steps: 34130, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000076, Sample Num: 1216, Cur Loss: 2.13636422, Cur Avg Loss: 1.47649170, Log Avg loss: 1.39265193, Global Avg Loss: 1.82438936, Time: 0.0403 Steps: 34140, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000086, Sample Num: 1376, Cur Loss: 0.80988318, Cur Avg Loss: 1.49507242, Log Avg loss: 1.63628590, Global Avg Loss: 1.82433428, Time: 0.0403 Steps: 34150, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000096, Sample Num: 1536, Cur Loss: 1.79945540, Cur Avg Loss: 1.49191962, Log Avg loss: 1.46480552, Global Avg Loss: 1.82422903, Time: 0.0403 Steps: 34160, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000106, Sample Num: 1696, Cur Loss: 0.74151236, Cur Avg Loss: 1.45053695, Log Avg loss: 1.05326330, Global Avg Loss: 1.82400340, Time: 0.0403 Steps: 34170, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000116, Sample Num: 1856, Cur Loss: 1.14519751, Cur Avg Loss: 1.45680645, Log Avg loss: 1.52326311, Global Avg Loss: 1.82391541, Time: 0.0403 Steps: 34180, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000126, Sample Num: 2016, Cur Loss: 1.25962448, Cur Avg Loss: 1.48024060, Log Avg loss: 1.75207675, Global Avg Loss: 1.82389440, Time: 0.0403 Steps: 34190, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000136, Sample Num: 2176, Cur Loss: 1.13167036, Cur Avg Loss: 1.46181708, Log Avg loss: 1.22968073, Global Avg Loss: 1.82372065, Time: 0.0574 Steps: 34200, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000146, Sample Num: 2336, Cur Loss: 1.15206921, Cur Avg Loss: 1.44474847, Log Avg loss: 1.21261548, Global Avg Loss: 1.82354202, Time: 0.0444 Steps: 34210, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000156, Sample Num: 2496, Cur Loss: 1.01494932, Cur Avg Loss: 1.42699490, Log Avg loss: 1.16779279, Global Avg Loss: 1.82335039, Time: 0.0474 Steps: 34220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000166, Sample Num: 2656, Cur Loss: 1.57606065, Cur Avg Loss: 1.42937682, Log Avg loss: 1.46653475, Global Avg Loss: 1.82324615, Time: 0.0932 Steps: 34230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000176, Sample Num: 2816, Cur Loss: 1.85477209, Cur Avg Loss: 1.42796567, Log Avg loss: 1.40454057, Global Avg Loss: 1.82312387, Time: 0.0447 Steps: 34240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000186, Sample Num: 2976, Cur Loss: 2.31841946, Cur Avg Loss: 1.44016581, Log Avg loss: 1.65488828, Global Avg Loss: 1.82307475, Time: 0.0483 Steps: 34250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000196, Sample Num: 3136, Cur Loss: 2.15262222, Cur Avg Loss: 1.44648807, Log Avg loss: 1.56408212, Global Avg Loss: 1.82299915, Time: 0.0873 Steps: 34260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000206, Sample Num: 3296, Cur Loss: 1.23760021, Cur Avg Loss: 1.45475535, Log Avg loss: 1.61679404, Global Avg Loss: 1.82293898, Time: 0.0404 Steps: 34270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000216, Sample Num: 3456, Cur Loss: 1.08390296, Cur Avg Loss: 1.45741763, Log Avg loss: 1.51226053, Global Avg Loss: 1.82284835, Time: 0.0404 Steps: 34280, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000226, Sample Num: 3616, Cur Loss: 3.11568880, Cur Avg Loss: 1.46106222, Log Avg loss: 1.53978527, Global Avg Loss: 1.82276580, Time: 0.0656 Steps: 34290, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000236, Sample Num: 3776, Cur Loss: 1.37966180, Cur Avg Loss: 1.45953517, Log Avg loss: 1.42502385, Global Avg Loss: 1.82264984, Time: 0.0403 Steps: 34300, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000246, Sample Num: 3936, Cur Loss: 2.02490377, Cur Avg Loss: 1.46314387, Log Avg loss: 1.54830937, Global Avg Loss: 1.82256988, Time: 0.0402 Steps: 34310, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000256, Sample Num: 4096, Cur Loss: 2.75736570, Cur Avg Loss: 1.45773842, Log Avg loss: 1.32476416, Global Avg Loss: 1.82242483, Time: 0.0405 Steps: 34320, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000266, Sample Num: 4256, Cur Loss: 0.67251599, Cur Avg Loss: 1.46170355, Log Avg loss: 1.56321102, Global Avg Loss: 1.82234933, Time: 0.0402 Steps: 34330, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000276, Sample Num: 4416, Cur Loss: 0.63062149, Cur Avg Loss: 1.45018763, Log Avg loss: 1.14386407, Global Avg Loss: 1.82215175, Time: 0.0403 Steps: 34340, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000286, Sample Num: 4576, Cur Loss: 1.65135932, Cur Avg Loss: 1.44824278, Log Avg loss: 1.39456496, Global Avg Loss: 1.82202727, Time: 0.0403 Steps: 34350, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000296, Sample Num: 4736, Cur Loss: 1.33331931, Cur Avg Loss: 1.44600280, Log Avg loss: 1.38193946, Global Avg Loss: 1.82189919, Time: 0.0402 Steps: 34360, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000306, Sample Num: 4896, Cur Loss: 2.16858625, Cur Avg Loss: 1.44761914, Log Avg loss: 1.49546269, Global Avg Loss: 1.82180421, Time: 0.0402 Steps: 34370, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000316, Sample Num: 5056, Cur Loss: 1.15534651, Cur Avg Loss: 1.45276303, Log Avg loss: 1.61016605, Global Avg Loss: 1.82174265, Time: 0.0402 Steps: 34380, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000326, Sample Num: 5216, Cur Loss: 0.61699510, Cur Avg Loss: 1.44891839, Log Avg loss: 1.32742780, Global Avg Loss: 1.82159891, Time: 0.0402 Steps: 34390, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000336, Sample Num: 5376, Cur Loss: 1.53801298, Cur Avg Loss: 1.45102476, Log Avg loss: 1.51969232, Global Avg Loss: 1.82151115, Time: 0.0402 Steps: 34400, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000346, Sample Num: 5536, Cur Loss: 2.00943446, Cur Avg Loss: 1.44785992, Log Avg loss: 1.34152154, Global Avg Loss: 1.82137166, Time: 0.0402 Steps: 34410, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000356, Sample Num: 5696, Cur Loss: 0.87401938, Cur Avg Loss: 1.44595918, Log Avg loss: 1.38019336, Global Avg Loss: 1.82124348, Time: 0.0403 Steps: 34420, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000366, Sample Num: 5856, Cur Loss: 1.68862808, Cur Avg Loss: 1.45003635, Log Avg loss: 1.59518368, Global Avg Loss: 1.82117783, Time: 0.0403 Steps: 34430, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000376, Sample Num: 6016, Cur Loss: 1.36543703, Cur Avg Loss: 1.44204581, Log Avg loss: 1.14959221, Global Avg Loss: 1.82098283, Time: 0.0403 Steps: 34440, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000386, Sample Num: 6176, Cur Loss: 1.16067088, Cur Avg Loss: 1.43671897, Log Avg loss: 1.23642974, Global Avg Loss: 1.82081314, Time: 0.0402 Steps: 34450, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000396, Sample Num: 6336, Cur Loss: 1.17311859, Cur Avg Loss: 1.43418172, Log Avg loss: 1.33624386, Global Avg Loss: 1.82067253, Time: 0.0402 Steps: 34460, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000406, Sample Num: 6496, Cur Loss: 1.90708470, Cur Avg Loss: 1.43931122, Log Avg loss: 1.64243948, Global Avg Loss: 1.82062082, Time: 0.0402 Steps: 34470, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000416, Sample Num: 6656, Cur Loss: 1.27994847, Cur Avg Loss: 1.44358066, Log Avg loss: 1.61691973, Global Avg Loss: 1.82056174, Time: 0.0404 Steps: 34480, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000426, Sample Num: 6816, Cur Loss: 2.12967634, Cur Avg Loss: 1.44974698, Log Avg loss: 1.70626595, Global Avg Loss: 1.82052860, Time: 0.0403 Steps: 34490, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000436, Sample Num: 6976, Cur Loss: 2.99335623, Cur Avg Loss: 1.44821135, Log Avg loss: 1.38279362, Global Avg Loss: 1.82040172, Time: 0.0403 Steps: 34500, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000446, Sample Num: 7136, Cur Loss: 1.52828801, Cur Avg Loss: 1.44471870, Log Avg loss: 1.29243917, Global Avg Loss: 1.82024873, Time: 0.0403 Steps: 34510, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000456, Sample Num: 7296, Cur Loss: 3.26021624, Cur Avg Loss: 1.45393328, Log Avg loss: 1.86490361, Global Avg Loss: 1.82026167, Time: 0.0402 Steps: 34520, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000466, Sample Num: 7456, Cur Loss: 0.96750224, Cur Avg Loss: 1.44924618, Log Avg loss: 1.23551436, Global Avg Loss: 1.82009233, Time: 0.0402 Steps: 34530, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000476, Sample Num: 7616, Cur Loss: 2.09125328, Cur Avg Loss: 1.45040662, Log Avg loss: 1.50448286, Global Avg Loss: 1.82000095, Time: 0.0403 Steps: 34540, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000486, Sample Num: 7776, Cur Loss: 1.76747119, Cur Avg Loss: 1.45456282, Log Avg loss: 1.65239806, Global Avg Loss: 1.81995244, Time: 0.0403 Steps: 34550, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000496, Sample Num: 7936, Cur Loss: 1.16499090, Cur Avg Loss: 1.44904467, Log Avg loss: 1.18086246, Global Avg Loss: 1.81976752, Time: 0.0563 Steps: 34560, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000506, Sample Num: 8096, Cur Loss: 1.17080009, Cur Avg Loss: 1.45156897, Log Avg loss: 1.57677428, Global Avg Loss: 1.81969723, Time: 0.0737 Steps: 34570, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000516, Sample Num: 8256, Cur Loss: 1.26672101, Cur Avg Loss: 1.45031597, Log Avg loss: 1.38691427, Global Avg Loss: 1.81957207, Time: 0.0407 Steps: 34580, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000526, Sample Num: 8416, Cur Loss: 2.62787867, Cur Avg Loss: 1.44967710, Log Avg loss: 1.41671150, Global Avg Loss: 1.81945561, Time: 0.0570 Steps: 34590, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000536, Sample Num: 8576, Cur Loss: 1.28876412, Cur Avg Loss: 1.45066112, Log Avg loss: 1.50242063, Global Avg Loss: 1.81936398, Time: 0.0407 Steps: 34600, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000546, Sample Num: 8736, Cur Loss: 1.67381728, Cur Avg Loss: 1.44946161, Log Avg loss: 1.38516758, Global Avg Loss: 1.81923852, Time: 0.0404 Steps: 34610, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000556, Sample Num: 8896, Cur Loss: 0.93539989, Cur Avg Loss: 1.45060607, Log Avg loss: 1.51309382, Global Avg Loss: 1.81915009, Time: 0.0509 Steps: 34620, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000566, Sample Num: 9056, Cur Loss: 2.00787830, Cur Avg Loss: 1.44872812, Log Avg loss: 1.34431396, Global Avg Loss: 1.81901298, Time: 0.0530 Steps: 34630, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000576, Sample Num: 9216, Cur Loss: 2.42407155, Cur Avg Loss: 1.44837859, Log Avg loss: 1.42859528, Global Avg Loss: 1.81890027, Time: 0.0736 Steps: 34640, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000586, Sample Num: 9376, Cur Loss: 1.73895454, Cur Avg Loss: 1.44297828, Log Avg loss: 1.13192042, Global Avg Loss: 1.81870201, Time: 0.0403 Steps: 34650, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000596, Sample Num: 9536, Cur Loss: 0.84628373, Cur Avg Loss: 1.43762788, Log Avg loss: 1.12409447, Global Avg Loss: 1.81850160, Time: 0.0403 Steps: 34660, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000606, Sample Num: 9696, Cur Loss: 0.89758492, Cur Avg Loss: 1.43607988, Log Avg loss: 1.34381901, Global Avg Loss: 1.81836469, Time: 0.0402 Steps: 34670, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000616, Sample Num: 9856, Cur Loss: 1.21727824, Cur Avg Loss: 1.43514392, Log Avg loss: 1.37842469, Global Avg Loss: 1.81823783, Time: 0.0403 Steps: 34680, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000626, Sample Num: 10016, Cur Loss: 0.71470213, Cur Avg Loss: 1.43490524, Log Avg loss: 1.42020257, Global Avg Loss: 1.81812309, Time: 0.0402 Steps: 34690, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000636, Sample Num: 10176, Cur Loss: 0.93501127, Cur Avg Loss: 1.42913887, Log Avg loss: 1.06816394, Global Avg Loss: 1.81790696, Time: 0.0403 Steps: 34700, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000646, Sample Num: 10336, Cur Loss: 1.09252691, Cur Avg Loss: 1.43115212, Log Avg loss: 1.55919485, Global Avg Loss: 1.81783243, Time: 0.0402 Steps: 34710, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000656, Sample Num: 10496, Cur Loss: 1.44211984, Cur Avg Loss: 1.43301067, Log Avg loss: 1.55307305, Global Avg Loss: 1.81775617, Time: 0.0403 Steps: 34720, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000666, Sample Num: 10656, Cur Loss: 1.98252892, Cur Avg Loss: 1.43248769, Log Avg loss: 1.39818001, Global Avg Loss: 1.81763536, Time: 0.0403 Steps: 34730, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000676, Sample Num: 10816, Cur Loss: 1.20465803, Cur Avg Loss: 1.42492485, Log Avg loss: 0.92124011, Global Avg Loss: 1.81737733, Time: 0.0403 Steps: 34740, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000686, Sample Num: 10976, Cur Loss: 1.94448328, Cur Avg Loss: 1.42638778, Log Avg loss: 1.52528190, Global Avg Loss: 1.81729327, Time: 0.0403 Steps: 34750, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000696, Sample Num: 11136, Cur Loss: 1.51645112, Cur Avg Loss: 1.42555691, Log Avg loss: 1.36855937, Global Avg Loss: 1.81716418, Time: 0.0403 Steps: 34760, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000706, Sample Num: 11296, Cur Loss: 0.82676286, Cur Avg Loss: 1.42335687, Log Avg loss: 1.27023402, Global Avg Loss: 1.81700688, Time: 0.0403 Steps: 34770, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000716, Sample Num: 11456, Cur Loss: 0.62333125, Cur Avg Loss: 1.41921075, Log Avg loss: 1.12649446, Global Avg Loss: 1.81680834, Time: 0.0404 Steps: 34780, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000726, Sample Num: 11616, Cur Loss: 1.37557602, Cur Avg Loss: 1.42205145, Log Avg loss: 1.62544526, Global Avg Loss: 1.81675334, Time: 0.0402 Steps: 34790, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000736, Sample Num: 11776, Cur Loss: 1.22665524, Cur Avg Loss: 1.42193368, Log Avg loss: 1.41338378, Global Avg Loss: 1.81663743, Time: 0.0402 Steps: 34800, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000746, Sample Num: 11936, Cur Loss: 3.34917259, Cur Avg Loss: 1.42500096, Log Avg loss: 1.65075321, Global Avg Loss: 1.81658977, Time: 0.0403 Steps: 34810, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000756, Sample Num: 12096, Cur Loss: 1.78979826, Cur Avg Loss: 1.42380996, Log Avg loss: 1.33496119, Global Avg Loss: 1.81645145, Time: 0.0403 Steps: 34820, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000766, Sample Num: 12256, Cur Loss: 1.78453267, Cur Avg Loss: 1.42519271, Log Avg loss: 1.52972837, Global Avg Loss: 1.81636913, Time: 0.0404 Steps: 34830, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000776, Sample Num: 12416, Cur Loss: 1.83657885, Cur Avg Loss: 1.42619035, Log Avg loss: 1.50260957, Global Avg Loss: 1.81627908, Time: 0.0402 Steps: 34840, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000786, Sample Num: 12576, Cur Loss: 1.08125091, Cur Avg Loss: 1.42429052, Log Avg loss: 1.27686355, Global Avg Loss: 1.81612429, Time: 0.0402 Steps: 34850, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000796, Sample Num: 12736, Cur Loss: 1.33264399, Cur Avg Loss: 1.42631499, Log Avg loss: 1.58543850, Global Avg Loss: 1.81605812, Time: 0.0402 Steps: 34860, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000806, Sample Num: 12896, Cur Loss: 2.14659691, Cur Avg Loss: 1.42924972, Log Avg loss: 1.66285449, Global Avg Loss: 1.81601418, Time: 0.0402 Steps: 34870, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000816, Sample Num: 13056, Cur Loss: 1.34871268, Cur Avg Loss: 1.42917005, Log Avg loss: 1.42274809, Global Avg Loss: 1.81590143, Time: 0.0402 Steps: 34880, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000826, Sample Num: 13216, Cur Loss: 1.06862903, Cur Avg Loss: 1.42868094, Log Avg loss: 1.38876975, Global Avg Loss: 1.81577901, Time: 0.0402 Steps: 34890, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000836, Sample Num: 13376, Cur Loss: 1.64452744, Cur Avg Loss: 1.42831227, Log Avg loss: 1.39785995, Global Avg Loss: 1.81565926, Time: 0.0402 Steps: 34900, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000846, Sample Num: 13536, Cur Loss: 1.86832297, Cur Avg Loss: 1.43063051, Log Avg loss: 1.62443553, Global Avg Loss: 1.81560449, Time: 0.0541 Steps: 34910, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000856, Sample Num: 13696, Cur Loss: 0.75370985, Cur Avg Loss: 1.42964296, Log Avg loss: 1.34609661, Global Avg Loss: 1.81547004, Time: 0.0840 Steps: 34920, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000866, Sample Num: 13856, Cur Loss: 2.02676916, Cur Avg Loss: 1.42821881, Log Avg loss: 1.30631164, Global Avg Loss: 1.81532427, Time: 0.0643 Steps: 34930, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000876, Sample Num: 14016, Cur Loss: 3.59643006, Cur Avg Loss: 1.42962627, Log Avg loss: 1.55151190, Global Avg Loss: 1.81524877, Time: 0.0523 Steps: 34940, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000886, Sample Num: 14176, Cur Loss: 0.62604272, Cur Avg Loss: 1.42903753, Log Avg loss: 1.37746397, Global Avg Loss: 1.81512351, Time: 0.0581 Steps: 34950, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000896, Sample Num: 14336, Cur Loss: 2.42763329, Cur Avg Loss: 1.43444543, Log Avg loss: 1.91358581, Global Avg Loss: 1.81515167, Time: 0.0676 Steps: 34960, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000906, Sample Num: 14496, Cur Loss: 1.31563640, Cur Avg Loss: 1.43663851, Log Avg loss: 1.63313818, Global Avg Loss: 1.81509962, Time: 0.0764 Steps: 34970, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000916, Sample Num: 14656, Cur Loss: 1.88296115, Cur Avg Loss: 1.43722982, Log Avg loss: 1.49080225, Global Avg Loss: 1.81500691, Time: 0.0598 Steps: 34980, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000926, Sample Num: 14816, Cur Loss: 1.48735237, Cur Avg Loss: 1.43626560, Log Avg loss: 1.34794289, Global Avg Loss: 1.81487343, Time: 0.0518 Steps: 34990, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000936, Sample Num: 14976, Cur Loss: 1.76247954, Cur Avg Loss: 1.43506651, Log Avg loss: 1.32403130, Global Avg Loss: 1.81473319, Time: 0.0402 Steps: 35000, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000946, Sample Num: 15136, Cur Loss: 1.74753642, Cur Avg Loss: 1.43522394, Log Avg loss: 1.44995931, Global Avg Loss: 1.81462900, Time: 0.0403 Steps: 35010, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000956, Sample Num: 15296, Cur Loss: 0.84323293, Cur Avg Loss: 1.43425612, Log Avg loss: 1.34270014, Global Avg Loss: 1.81449424, Time: 0.0403 Steps: 35020, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000966, Sample Num: 15456, Cur Loss: 1.00587463, Cur Avg Loss: 1.43462340, Log Avg loss: 1.46973518, Global Avg Loss: 1.81439582, Time: 0.0402 Steps: 35030, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000976, Sample Num: 15616, Cur Loss: 1.72800398, Cur Avg Loss: 1.43450967, Log Avg loss: 1.42352342, Global Avg Loss: 1.81428427, Time: 0.0403 Steps: 35040, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000986, Sample Num: 15776, Cur Loss: 3.18735456, Cur Avg Loss: 1.43825348, Log Avg loss: 1.80364995, Global Avg Loss: 1.81428123, Time: 0.0402 Steps: 35050, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000996, Sample Num: 15936, Cur Loss: 1.73814416, Cur Avg Loss: 1.44062790, Log Avg loss: 1.67474564, Global Avg Loss: 1.81424143, Time: 0.0403 Steps: 35060, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001006, Sample Num: 16096, Cur Loss: 0.78687179, Cur Avg Loss: 1.43771555, Log Avg loss: 1.14764474, Global Avg Loss: 1.81405136, Time: 0.0404 Steps: 35070, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001016, Sample Num: 16256, Cur Loss: 2.14060450, Cur Avg Loss: 1.43831559, Log Avg loss: 1.49868023, Global Avg Loss: 1.81396146, Time: 0.0403 Steps: 35080, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001026, Sample Num: 16416, Cur Loss: 1.17834973, Cur Avg Loss: 1.44255234, Log Avg loss: 1.87300643, Global Avg Loss: 1.81397828, Time: 0.0403 Steps: 35090, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001036, Sample Num: 16576, Cur Loss: 1.45555592, Cur Avg Loss: 1.44031245, Log Avg loss: 1.21049919, Global Avg Loss: 1.81380635, Time: 0.0403 Steps: 35100, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001046, Sample Num: 16736, Cur Loss: 2.38020420, Cur Avg Loss: 1.44122858, Log Avg loss: 1.53613946, Global Avg Loss: 1.81372727, Time: 0.0403 Steps: 35110, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001056, Sample Num: 16896, Cur Loss: 0.56493592, Cur Avg Loss: 1.43909016, Log Avg loss: 1.21541194, Global Avg Loss: 1.81355690, Time: 0.0403 Steps: 35120, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001066, Sample Num: 17056, Cur Loss: 2.08105326, Cur Avg Loss: 1.44003426, Log Avg loss: 1.53973145, Global Avg Loss: 1.81347896, Time: 0.0403 Steps: 35130, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001076, Sample Num: 17216, Cur Loss: 2.68228197, Cur Avg Loss: 1.44493351, Log Avg loss: 1.96719261, Global Avg Loss: 1.81352270, Time: 0.0403 Steps: 35140, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001086, Sample Num: 17376, Cur Loss: 0.92736435, Cur Avg Loss: 1.44382535, Log Avg loss: 1.32458774, Global Avg Loss: 1.81338360, Time: 0.0402 Steps: 35150, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001096, Sample Num: 17536, Cur Loss: 1.63258648, Cur Avg Loss: 1.44317499, Log Avg loss: 1.37254634, Global Avg Loss: 1.81325822, Time: 0.0403 Steps: 35160, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001106, Sample Num: 17696, Cur Loss: 1.11416769, Cur Avg Loss: 1.44355582, Log Avg loss: 1.48529441, Global Avg Loss: 1.81316497, Time: 0.0403 Steps: 35170, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001116, Sample Num: 17856, Cur Loss: 1.36385858, Cur Avg Loss: 1.44350997, Log Avg loss: 1.43843868, Global Avg Loss: 1.81305845, Time: 0.0402 Steps: 35180, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001126, Sample Num: 18016, Cur Loss: 1.39427900, Cur Avg Loss: 1.44420575, Log Avg loss: 1.52185479, Global Avg Loss: 1.81297570, Time: 0.0402 Steps: 35190, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001136, Sample Num: 18176, Cur Loss: 0.79345775, Cur Avg Loss: 1.44395250, Log Avg loss: 1.41543702, Global Avg Loss: 1.81286276, Time: 0.0402 Steps: 35200, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001146, Sample Num: 18336, Cur Loss: 1.49841118, Cur Avg Loss: 1.44512380, Log Avg loss: 1.57818310, Global Avg Loss: 1.81279611, Time: 0.0404 Steps: 35210, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001156, Sample Num: 18496, Cur Loss: 1.34363556, Cur Avg Loss: 1.44538102, Log Avg loss: 1.47485871, Global Avg Loss: 1.81270016, Time: 0.0402 Steps: 35220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001166, Sample Num: 18656, Cur Loss: 0.86323953, Cur Avg Loss: 1.44327848, Log Avg loss: 1.20022437, Global Avg Loss: 1.81252631, Time: 0.0402 Steps: 35230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001176, Sample Num: 18816, Cur Loss: 0.61368197, Cur Avg Loss: 1.44345845, Log Avg loss: 1.46444366, Global Avg Loss: 1.81242754, Time: 0.0402 Steps: 35240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001186, Sample Num: 18976, Cur Loss: 1.01122499, Cur Avg Loss: 1.44242539, Log Avg loss: 1.32093775, Global Avg Loss: 1.81228811, Time: 0.0402 Steps: 35250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001196, Sample Num: 19136, Cur Loss: 1.33818531, Cur Avg Loss: 1.44034667, Log Avg loss: 1.19380974, Global Avg Loss: 1.81211270, Time: 0.0402 Steps: 35260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001206, Sample Num: 19296, Cur Loss: 1.90284276, Cur Avg Loss: 1.44183665, Log Avg loss: 1.62003885, Global Avg Loss: 1.81205824, Time: 0.0949 Steps: 35270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001216, Sample Num: 19456, Cur Loss: 2.47364473, Cur Avg Loss: 1.44253117, Log Avg loss: 1.52628965, Global Avg Loss: 1.81197724, Time: 0.0532 Steps: 35280, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001226, Sample Num: 19616, Cur Loss: 1.17707753, Cur Avg Loss: 1.44383819, Log Avg loss: 1.60277214, Global Avg Loss: 1.81191796, Time: 0.1008 Steps: 35290, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001236, Sample Num: 19776, Cur Loss: 0.95263720, Cur Avg Loss: 1.44209296, Log Avg loss: 1.22812740, Global Avg Loss: 1.81175258, Time: 0.0875 Steps: 35300, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001246, Sample Num: 19936, Cur Loss: 1.48548830, Cur Avg Loss: 1.44245284, Log Avg loss: 1.48693420, Global Avg Loss: 1.81166059, Time: 0.0993 Steps: 35310, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001256, Sample Num: 20096, Cur Loss: 0.67822534, Cur Avg Loss: 1.44055794, Log Avg loss: 1.20445369, Global Avg Loss: 1.81148868, Time: 0.0858 Steps: 35320, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001266, Sample Num: 20256, Cur Loss: 0.65036827, Cur Avg Loss: 1.44219764, Log Avg loss: 1.64814364, Global Avg Loss: 1.81144244, Time: 0.0535 Steps: 35330, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001276, Sample Num: 20416, Cur Loss: 0.88869989, Cur Avg Loss: 1.44431118, Log Avg loss: 1.71188576, Global Avg Loss: 1.81141427, Time: 0.0578 Steps: 35340, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001286, Sample Num: 20576, Cur Loss: 1.20438266, Cur Avg Loss: 1.44466265, Log Avg loss: 1.48950948, Global Avg Loss: 1.81132321, Time: 0.0554 Steps: 35350, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001296, Sample Num: 20736, Cur Loss: 1.20358205, Cur Avg Loss: 1.44527941, Log Avg loss: 1.52459460, Global Avg Loss: 1.81124212, Time: 0.0530 Steps: 35360, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001306, Sample Num: 20896, Cur Loss: 1.30257738, Cur Avg Loss: 1.44628873, Log Avg loss: 1.57709674, Global Avg Loss: 1.81117592, Time: 0.0402 Steps: 35370, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001316, Sample Num: 21056, Cur Loss: 1.56756842, Cur Avg Loss: 1.44779151, Log Avg loss: 1.64405509, Global Avg Loss: 1.81112869, Time: 0.0403 Steps: 35380, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001326, Sample Num: 21216, Cur Loss: 2.25308943, Cur Avg Loss: 1.44806844, Log Avg loss: 1.48451198, Global Avg Loss: 1.81103639, Time: 0.0403 Steps: 35390, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001336, Sample Num: 21376, Cur Loss: 1.16171741, Cur Avg Loss: 1.44519968, Log Avg loss: 1.06480255, Global Avg Loss: 1.81082559, Time: 0.0402 Steps: 35400, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001346, Sample Num: 21536, Cur Loss: 2.06289482, Cur Avg Loss: 1.44663174, Log Avg loss: 1.63795543, Global Avg Loss: 1.81077677, Time: 0.0403 Steps: 35410, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001356, Sample Num: 21696, Cur Loss: 1.42539942, Cur Avg Loss: 1.44654030, Log Avg loss: 1.43423258, Global Avg Loss: 1.81067047, Time: 0.0403 Steps: 35420, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001366, Sample Num: 21856, Cur Loss: 1.16450644, Cur Avg Loss: 1.44696107, Log Avg loss: 1.50401673, Global Avg Loss: 1.81058391, Time: 0.0403 Steps: 35430, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001376, Sample Num: 22016, Cur Loss: 0.76122332, Cur Avg Loss: 1.44665883, Log Avg loss: 1.40537237, Global Avg Loss: 1.81046958, Time: 0.0402 Steps: 35440, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001386, Sample Num: 22176, Cur Loss: 1.59913468, Cur Avg Loss: 1.44629085, Log Avg loss: 1.39565702, Global Avg Loss: 1.81035256, Time: 0.0402 Steps: 35450, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001396, Sample Num: 22336, Cur Loss: 2.04592776, Cur Avg Loss: 1.44686739, Log Avg loss: 1.52677692, Global Avg Loss: 1.81027259, Time: 0.0403 Steps: 35460, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001406, Sample Num: 22496, Cur Loss: 0.76372963, Cur Avg Loss: 1.44567765, Log Avg loss: 1.27958878, Global Avg Loss: 1.81012298, Time: 0.0403 Steps: 35470, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001416, Sample Num: 22656, Cur Loss: 0.52441865, Cur Avg Loss: 1.44503875, Log Avg loss: 1.35521018, Global Avg Loss: 1.80999476, Time: 0.0403 Steps: 35480, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001426, Sample Num: 22816, Cur Loss: 0.25261226, Cur Avg Loss: 1.44407322, Log Avg loss: 1.30735451, Global Avg Loss: 1.80985313, Time: 0.0402 Steps: 35490, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001436, Sample Num: 22976, Cur Loss: 2.06197548, Cur Avg Loss: 1.44595645, Log Avg loss: 1.71450488, Global Avg Loss: 1.80982627, Time: 0.0403 Steps: 35500, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001446, Sample Num: 23136, Cur Loss: 2.72067165, Cur Avg Loss: 1.44676856, Log Avg loss: 1.56338767, Global Avg Loss: 1.80975687, Time: 0.0403 Steps: 35510, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001456, Sample Num: 23296, Cur Loss: 1.28578222, Cur Avg Loss: 1.44783309, Log Avg loss: 1.60176308, Global Avg Loss: 1.80969832, Time: 0.0402 Steps: 35520, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001466, Sample Num: 23456, Cur Loss: 0.52851629, Cur Avg Loss: 1.44629306, Log Avg loss: 1.22206575, Global Avg Loss: 1.80953293, Time: 0.0402 Steps: 35530, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001476, Sample Num: 23616, Cur Loss: 1.19673133, Cur Avg Loss: 1.44677564, Log Avg loss: 1.51752073, Global Avg Loss: 1.80945076, Time: 0.0402 Steps: 35540, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001486, Sample Num: 23776, Cur Loss: 0.95810342, Cur Avg Loss: 1.44668299, Log Avg loss: 1.43300788, Global Avg Loss: 1.80934487, Time: 0.0402 Steps: 35550, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001496, Sample Num: 23936, Cur Loss: 0.85411036, Cur Avg Loss: 1.44731397, Log Avg loss: 1.54107744, Global Avg Loss: 1.80926943, Time: 0.0402 Steps: 35560, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001506, Sample Num: 24096, Cur Loss: 1.65449560, Cur Avg Loss: 1.44609637, Log Avg loss: 1.26394390, Global Avg Loss: 1.80911612, Time: 0.0402 Steps: 35570, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001516, Sample Num: 24256, Cur Loss: 0.81747246, Cur Avg Loss: 1.44467989, Log Avg loss: 1.23135856, Global Avg Loss: 1.80895374, Time: 0.0402 Steps: 35580, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001526, Sample Num: 24416, Cur Loss: 1.43652546, Cur Avg Loss: 1.44488400, Log Avg loss: 1.47582707, Global Avg Loss: 1.80886014, Time: 0.0402 Steps: 35590, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001536, Sample Num: 24576, Cur Loss: 0.98714054, Cur Avg Loss: 1.44420740, Log Avg loss: 1.34095846, Global Avg Loss: 1.80872870, Time: 0.0404 Steps: 35600, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001546, Sample Num: 24736, Cur Loss: 2.06801224, Cur Avg Loss: 1.44423665, Log Avg loss: 1.44872887, Global Avg Loss: 1.80862761, Time: 0.0402 Steps: 35610, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001556, Sample Num: 24896, Cur Loss: 1.12607288, Cur Avg Loss: 1.44519316, Log Avg loss: 1.59306977, Global Avg Loss: 1.80856709, Time: 0.0402 Steps: 35620, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001566, Sample Num: 25056, Cur Loss: 0.70269418, Cur Avg Loss: 1.44641574, Log Avg loss: 1.63664825, Global Avg Loss: 1.80851884, Time: 0.0515 Steps: 35630, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001576, Sample Num: 25216, Cur Loss: 1.62830389, Cur Avg Loss: 1.44538997, Log Avg loss: 1.28475491, Global Avg Loss: 1.80837188, Time: 0.0552 Steps: 35640, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001586, Sample Num: 25376, Cur Loss: 1.20398760, Cur Avg Loss: 1.44598031, Log Avg loss: 1.53901739, Global Avg Loss: 1.80829633, Time: 0.0412 Steps: 35650, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001596, Sample Num: 25536, Cur Loss: 0.59211957, Cur Avg Loss: 1.44260308, Log Avg loss: 0.90697543, Global Avg Loss: 1.80804357, Time: 0.0655 Steps: 35660, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001606, Sample Num: 25696, Cur Loss: 1.44865513, Cur Avg Loss: 1.44196554, Log Avg loss: 1.34021347, Global Avg Loss: 1.80791242, Time: 0.0657 Steps: 35670, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001616, Sample Num: 25856, Cur Loss: 1.72077727, Cur Avg Loss: 1.44256110, Log Avg loss: 1.53820810, Global Avg Loss: 1.80783683, Time: 0.0472 Steps: 35680, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001626, Sample Num: 26016, Cur Loss: 1.60208762, Cur Avg Loss: 1.44291097, Log Avg loss: 1.49945041, Global Avg Loss: 1.80775042, Time: 0.0663 Steps: 35690, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001636, Sample Num: 26176, Cur Loss: 1.62494898, Cur Avg Loss: 1.44187227, Log Avg loss: 1.27297994, Global Avg Loss: 1.80760062, Time: 0.0404 Steps: 35700, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001646, Sample Num: 26336, Cur Loss: 2.15055847, Cur Avg Loss: 1.44426121, Log Avg loss: 1.83509185, Global Avg Loss: 1.80760832, Time: 0.0405 Steps: 35710, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001656, Sample Num: 26496, Cur Loss: 0.67681819, Cur Avg Loss: 1.44256659, Log Avg loss: 1.16363159, Global Avg Loss: 1.80742804, Time: 0.0402 Steps: 35720, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001666, Sample Num: 26656, Cur Loss: 0.92128736, Cur Avg Loss: 1.44205989, Log Avg loss: 1.35815089, Global Avg Loss: 1.80730230, Time: 0.0402 Steps: 35730, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001676, Sample Num: 26816, Cur Loss: 0.79071814, Cur Avg Loss: 1.44297673, Log Avg loss: 1.59572222, Global Avg Loss: 1.80724310, Time: 0.0403 Steps: 35740, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001686, Sample Num: 26976, Cur Loss: 0.92685997, Cur Avg Loss: 1.44041493, Log Avg loss: 1.01105750, Global Avg Loss: 1.80702039, Time: 0.0403 Steps: 35750, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001696, Sample Num: 27136, Cur Loss: 1.63905835, Cur Avg Loss: 1.44018022, Log Avg loss: 1.40060774, Global Avg Loss: 1.80690674, Time: 0.0404 Steps: 35760, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001706, Sample Num: 27296, Cur Loss: 2.55429721, Cur Avg Loss: 1.44170681, Log Avg loss: 1.70061545, Global Avg Loss: 1.80687702, Time: 0.0403 Steps: 35770, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001716, Sample Num: 27456, Cur Loss: 0.46755531, Cur Avg Loss: 1.44188407, Log Avg loss: 1.47212575, Global Avg Loss: 1.80678346, Time: 0.0403 Steps: 35780, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001726, Sample Num: 27616, Cur Loss: 0.98851514, Cur Avg Loss: 1.44325993, Log Avg loss: 1.67935634, Global Avg Loss: 1.80674786, Time: 0.0403 Steps: 35790, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001736, Sample Num: 27776, Cur Loss: 0.42734838, Cur Avg Loss: 1.44264510, Log Avg loss: 1.33652671, Global Avg Loss: 1.80661651, Time: 0.0403 Steps: 35800, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001746, Sample Num: 27936, Cur Loss: 1.19049478, Cur Avg Loss: 1.44213947, Log Avg loss: 1.35436217, Global Avg Loss: 1.80649022, Time: 0.0403 Steps: 35810, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001756, Sample Num: 28096, Cur Loss: 3.38080645, Cur Avg Loss: 1.44509952, Log Avg loss: 1.96192399, Global Avg Loss: 1.80653361, Time: 0.0402 Steps: 35820, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001766, Sample Num: 28256, Cur Loss: 1.87471545, Cur Avg Loss: 1.44536287, Log Avg loss: 1.49160638, Global Avg Loss: 1.80644572, Time: 0.0404 Steps: 35830, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001776, Sample Num: 28416, Cur Loss: 1.25371063, Cur Avg Loss: 1.44469368, Log Avg loss: 1.32651482, Global Avg Loss: 1.80631181, Time: 0.0403 Steps: 35840, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001786, Sample Num: 28576, Cur Loss: 2.18168330, Cur Avg Loss: 1.44532578, Log Avg loss: 1.55758787, Global Avg Loss: 1.80624243, Time: 0.0403 Steps: 35850, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001796, Sample Num: 28736, Cur Loss: 1.61145878, Cur Avg Loss: 1.44577391, Log Avg loss: 1.52580888, Global Avg Loss: 1.80616423, Time: 0.0403 Steps: 35860, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001806, Sample Num: 28896, Cur Loss: 0.33588681, Cur Avg Loss: 1.44593051, Log Avg loss: 1.47405650, Global Avg Loss: 1.80607164, Time: 0.0403 Steps: 35870, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001816, Sample Num: 29056, Cur Loss: 1.32957101, Cur Avg Loss: 1.44615808, Log Avg loss: 1.48725742, Global Avg Loss: 1.80598279, Time: 0.0402 Steps: 35880, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001826, Sample Num: 29216, Cur Loss: 1.01083755, Cur Avg Loss: 1.44701567, Log Avg loss: 1.60275289, Global Avg Loss: 1.80592616, Time: 0.0403 Steps: 35890, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001836, Sample Num: 29376, Cur Loss: 1.45172358, Cur Avg Loss: 1.44707544, Log Avg loss: 1.45799081, Global Avg Loss: 1.80582924, Time: 0.0403 Steps: 35900, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001846, Sample Num: 29536, Cur Loss: 1.60291541, Cur Avg Loss: 1.44979792, Log Avg loss: 1.94964472, Global Avg Loss: 1.80586929, Time: 0.0402 Steps: 35910, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001856, Sample Num: 29696, Cur Loss: 2.56047869, Cur Avg Loss: 1.45042732, Log Avg loss: 1.56661426, Global Avg Loss: 1.80580268, Time: 0.0403 Steps: 35920, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001866, Sample Num: 29856, Cur Loss: 1.48863804, Cur Avg Loss: 1.44953761, Log Avg loss: 1.28440810, Global Avg Loss: 1.80565757, Time: 0.0403 Steps: 35930, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001876, Sample Num: 30016, Cur Loss: 1.44116032, Cur Avg Loss: 1.45032133, Log Avg loss: 1.59656251, Global Avg Loss: 1.80559939, Time: 0.0402 Steps: 35940, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001886, Sample Num: 30176, Cur Loss: 1.68917120, Cur Avg Loss: 1.45275245, Log Avg loss: 1.90883183, Global Avg Loss: 1.80562811, Time: 0.0403 Steps: 35950, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001896, Sample Num: 30336, Cur Loss: 0.97349602, Cur Avg Loss: 1.45383567, Log Avg loss: 1.65812966, Global Avg Loss: 1.80558709, Time: 0.0403 Steps: 35960, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001906, Sample Num: 30496, Cur Loss: 0.88399851, Cur Avg Loss: 1.45290302, Log Avg loss: 1.27607343, Global Avg Loss: 1.80543988, Time: 0.0476 Steps: 35970, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001916, Sample Num: 30656, Cur Loss: 0.46801057, Cur Avg Loss: 1.45222776, Log Avg loss: 1.32352357, Global Avg Loss: 1.80530594, Time: 0.0757 Steps: 35980, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001926, Sample Num: 30816, Cur Loss: 0.83228642, Cur Avg Loss: 1.45202031, Log Avg loss: 1.41227248, Global Avg Loss: 1.80519673, Time: 0.0711 Steps: 35990, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001936, Sample Num: 30976, Cur Loss: 2.35325742, Cur Avg Loss: 1.45246041, Log Avg loss: 1.53722285, Global Avg Loss: 1.80512229, Time: 0.0413 Steps: 36000, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001946, Sample Num: 31136, Cur Loss: 1.73167503, Cur Avg Loss: 1.45391270, Log Avg loss: 1.73507652, Global Avg Loss: 1.80510284, Time: 0.0804 Steps: 36010, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001956, Sample Num: 31296, Cur Loss: 1.58556604, Cur Avg Loss: 1.45461607, Log Avg loss: 1.59149182, Global Avg Loss: 1.80504354, Time: 0.0828 Steps: 36020, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001966, Sample Num: 31456, Cur Loss: 2.46179438, Cur Avg Loss: 1.45519261, Log Avg loss: 1.56796392, Global Avg Loss: 1.80497774, Time: 0.0624 Steps: 36030, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001976, Sample Num: 31616, Cur Loss: 0.73032439, Cur Avg Loss: 1.45408253, Log Avg loss: 1.23583973, Global Avg Loss: 1.80481982, Time: 0.0492 Steps: 36040, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001986, Sample Num: 31776, Cur Loss: 2.33366132, Cur Avg Loss: 1.45513587, Log Avg loss: 1.66327696, Global Avg Loss: 1.80478056, Time: 0.0403 Steps: 36050, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001996, Sample Num: 31936, Cur Loss: 0.47540000, Cur Avg Loss: 1.45382984, Log Avg loss: 1.19445252, Global Avg Loss: 1.80461130, Time: 0.0402 Steps: 36060, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002006, Sample Num: 32096, Cur Loss: 1.02115738, Cur Avg Loss: 1.45367751, Log Avg loss: 1.42327113, Global Avg Loss: 1.80450558, Time: 0.0402 Steps: 36070, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002016, Sample Num: 32256, Cur Loss: 1.33445632, Cur Avg Loss: 1.45394474, Log Avg loss: 1.50755148, Global Avg Loss: 1.80442328, Time: 0.0403 Steps: 36080, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002026, Sample Num: 32416, Cur Loss: 1.98381615, Cur Avg Loss: 1.45483697, Log Avg loss: 1.63471190, Global Avg Loss: 1.80437625, Time: 0.0403 Steps: 36090, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002036, Sample Num: 32576, Cur Loss: 1.55240643, Cur Avg Loss: 1.45463947, Log Avg loss: 1.41462551, Global Avg Loss: 1.80426829, Time: 0.0402 Steps: 36100, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002046, Sample Num: 32736, Cur Loss: 1.56542003, Cur Avg Loss: 1.45593113, Log Avg loss: 1.71891254, Global Avg Loss: 1.80424465, Time: 0.0403 Steps: 36110, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002056, Sample Num: 32896, Cur Loss: 1.21360862, Cur Avg Loss: 1.45648458, Log Avg loss: 1.56972016, Global Avg Loss: 1.80417972, Time: 0.0402 Steps: 36120, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002066, Sample Num: 33056, Cur Loss: 1.96595740, Cur Avg Loss: 1.45592750, Log Avg loss: 1.34139178, Global Avg Loss: 1.80405163, Time: 0.0402 Steps: 36130, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002076, Sample Num: 33216, Cur Loss: 1.27840972, Cur Avg Loss: 1.45559399, Log Avg loss: 1.38669178, Global Avg Loss: 1.80393615, Time: 0.0403 Steps: 36140, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002086, Sample Num: 33376, Cur Loss: 0.52319729, Cur Avg Loss: 1.45659268, Log Avg loss: 1.66391929, Global Avg Loss: 1.80389742, Time: 0.0402 Steps: 36150, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002096, Sample Num: 33536, Cur Loss: 1.71281099, Cur Avg Loss: 1.45628068, Log Avg loss: 1.39119868, Global Avg Loss: 1.80378328, Time: 0.0402 Steps: 36160, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002106, Sample Num: 33696, Cur Loss: 1.29738176, Cur Avg Loss: 1.45633157, Log Avg loss: 1.46699758, Global Avg Loss: 1.80369017, Time: 0.0402 Steps: 36170, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002116, Sample Num: 33856, Cur Loss: 0.73695874, Cur Avg Loss: 1.45547710, Log Avg loss: 1.27552557, Global Avg Loss: 1.80354419, Time: 0.0402 Steps: 36180, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002126, Sample Num: 34016, Cur Loss: 2.18829179, Cur Avg Loss: 1.45603980, Log Avg loss: 1.57510856, Global Avg Loss: 1.80348107, Time: 0.0402 Steps: 36190, Updated lr: 0.000067 ***** Running evaluation checkpoint-36193 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-36193 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.799215, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.373519, "eval_total_loss": 965.584113, "eval_mae": 0.993062, "eval_mse": 1.373496, "eval_r2": 0.126916, "eval_sp_statistic": 0.320192, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.375769, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.32408, "test_total_loss": 664.688004, "test_mae": 0.935126, "test_mse": 1.324302, "test_r2": 0.145284, "test_sp_statistic": 0.327713, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.407876, "test_ps_pvalue": 0.0, "lr": 6.662588904694168e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.803470366343115, "train_cur_epoch_loss": 3100.563699424267, "train_cur_epoch_avg_loss": 1.4563474398423049, "train_cur_epoch_time": 95.79921460151672, "train_cur_epoch_avg_time": 0.0449972825746908, "epoch": 17, "step": 36193} ################################################## Training, Epoch: 0018, Batch: 000007, Sample Num: 112, Cur Loss: 1.68072891, Cur Avg Loss: 1.24685620, Log Avg loss: 1.37510687, Global Avg Loss: 1.80336273, Time: 0.0405 Steps: 36200, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000017, Sample Num: 272, Cur Loss: 2.12270355, Cur Avg Loss: 1.29254646, Log Avg loss: 1.32452964, Global Avg Loss: 1.80323050, Time: 0.0403 Steps: 36210, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000027, Sample Num: 432, Cur Loss: 1.97751725, Cur Avg Loss: 1.43802763, Log Avg loss: 1.68534562, Global Avg Loss: 1.80319795, Time: 0.0403 Steps: 36220, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000037, Sample Num: 592, Cur Loss: 1.34635496, Cur Avg Loss: 1.46578303, Log Avg loss: 1.54072260, Global Avg Loss: 1.80312550, Time: 0.0403 Steps: 36230, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000047, Sample Num: 752, Cur Loss: 3.59949160, Cur Avg Loss: 1.50146893, Log Avg loss: 1.63350679, Global Avg Loss: 1.80307870, Time: 0.0403 Steps: 36240, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000057, Sample Num: 912, Cur Loss: 1.91062641, Cur Avg Loss: 1.52774209, Log Avg loss: 1.65122594, Global Avg Loss: 1.80303681, Time: 0.0402 Steps: 36250, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000067, Sample Num: 1072, Cur Loss: 0.85667551, Cur Avg Loss: 1.50271823, Log Avg loss: 1.36008223, Global Avg Loss: 1.80291465, Time: 0.0402 Steps: 36260, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000077, Sample Num: 1232, Cur Loss: 0.42507029, Cur Avg Loss: 1.45466277, Log Avg loss: 1.13269119, Global Avg Loss: 1.80272986, Time: 0.0402 Steps: 36270, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000087, Sample Num: 1392, Cur Loss: 2.12279606, Cur Avg Loss: 1.46886020, Log Avg loss: 1.57818040, Global Avg Loss: 1.80266797, Time: 0.0402 Steps: 36280, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000097, Sample Num: 1552, Cur Loss: 1.06587446, Cur Avg Loss: 1.47756449, Log Avg loss: 1.55329179, Global Avg Loss: 1.80259925, Time: 0.0402 Steps: 36290, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000107, Sample Num: 1712, Cur Loss: 1.34613526, Cur Avg Loss: 1.47557394, Log Avg loss: 1.45626565, Global Avg Loss: 1.80250384, Time: 0.0402 Steps: 36300, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000117, Sample Num: 1872, Cur Loss: 1.36687541, Cur Avg Loss: 1.48296297, Log Avg loss: 1.56202552, Global Avg Loss: 1.80243761, Time: 0.0403 Steps: 36310, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000127, Sample Num: 2032, Cur Loss: 1.06918430, Cur Avg Loss: 1.46677898, Log Avg loss: 1.27742634, Global Avg Loss: 1.80229306, Time: 0.0402 Steps: 36320, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000137, Sample Num: 2192, Cur Loss: 0.83408880, Cur Avg Loss: 1.47548607, Log Avg loss: 1.58606616, Global Avg Loss: 1.80223354, Time: 0.0402 Steps: 36330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000147, Sample Num: 2352, Cur Loss: 0.90945327, Cur Avg Loss: 1.46859106, Log Avg loss: 1.37412941, Global Avg Loss: 1.80211574, Time: 0.0402 Steps: 36340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000157, Sample Num: 2512, Cur Loss: 1.82711661, Cur Avg Loss: 1.48507131, Log Avg loss: 1.72733102, Global Avg Loss: 1.80209516, Time: 0.0456 Steps: 36350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000167, Sample Num: 2672, Cur Loss: 1.28877211, Cur Avg Loss: 1.46932070, Log Avg loss: 1.22203611, Global Avg Loss: 1.80193563, Time: 0.0636 Steps: 36360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000177, Sample Num: 2832, Cur Loss: 0.46944657, Cur Avg Loss: 1.44153920, Log Avg loss: 0.97758804, Global Avg Loss: 1.80170897, Time: 0.0477 Steps: 36370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000187, Sample Num: 2992, Cur Loss: 1.52842677, Cur Avg Loss: 1.43826825, Log Avg loss: 1.38037251, Global Avg Loss: 1.80159316, Time: 0.0432 Steps: 36380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000197, Sample Num: 3152, Cur Loss: 1.62530851, Cur Avg Loss: 1.44134951, Log Avg loss: 1.49896898, Global Avg Loss: 1.80151000, Time: 0.0480 Steps: 36390, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000207, Sample Num: 3312, Cur Loss: 1.59585619, Cur Avg Loss: 1.44225094, Log Avg loss: 1.46000922, Global Avg Loss: 1.80141618, Time: 0.0637 Steps: 36400, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000217, Sample Num: 3472, Cur Loss: 1.90298593, Cur Avg Loss: 1.43708239, Log Avg loss: 1.33009336, Global Avg Loss: 1.80128673, Time: 0.0782 Steps: 36410, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000227, Sample Num: 3632, Cur Loss: 1.63500726, Cur Avg Loss: 1.42239569, Log Avg loss: 1.10369428, Global Avg Loss: 1.80109519, Time: 0.0406 Steps: 36420, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000237, Sample Num: 3792, Cur Loss: 1.95151448, Cur Avg Loss: 1.42931667, Log Avg loss: 1.58642304, Global Avg Loss: 1.80103626, Time: 0.1008 Steps: 36430, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000247, Sample Num: 3952, Cur Loss: 1.34914863, Cur Avg Loss: 1.42983352, Log Avg loss: 1.44208282, Global Avg Loss: 1.80093776, Time: 0.0403 Steps: 36440, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000257, Sample Num: 4112, Cur Loss: 2.55641818, Cur Avg Loss: 1.42529493, Log Avg loss: 1.31319179, Global Avg Loss: 1.80080394, Time: 0.0403 Steps: 36450, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000267, Sample Num: 4272, Cur Loss: 0.90318662, Cur Avg Loss: 1.43239586, Log Avg loss: 1.61488972, Global Avg Loss: 1.80075295, Time: 0.0402 Steps: 36460, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000277, Sample Num: 4432, Cur Loss: 0.51245183, Cur Avg Loss: 1.43468342, Log Avg loss: 1.49576117, Global Avg Loss: 1.80066932, Time: 0.0402 Steps: 36470, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000287, Sample Num: 4592, Cur Loss: 2.06533718, Cur Avg Loss: 1.43713792, Log Avg loss: 1.50512748, Global Avg Loss: 1.80058831, Time: 0.0402 Steps: 36480, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000297, Sample Num: 4752, Cur Loss: 0.93421835, Cur Avg Loss: 1.43750488, Log Avg loss: 1.44803662, Global Avg Loss: 1.80049169, Time: 0.0402 Steps: 36490, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000307, Sample Num: 4912, Cur Loss: 1.56383634, Cur Avg Loss: 1.44166994, Log Avg loss: 1.56537238, Global Avg Loss: 1.80042728, Time: 0.0403 Steps: 36500, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000317, Sample Num: 5072, Cur Loss: 0.69522369, Cur Avg Loss: 1.44020164, Log Avg loss: 1.39512478, Global Avg Loss: 1.80031627, Time: 0.0402 Steps: 36510, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000327, Sample Num: 5232, Cur Loss: 1.78353298, Cur Avg Loss: 1.44848146, Log Avg loss: 1.71095186, Global Avg Loss: 1.80029180, Time: 0.0403 Steps: 36520, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000337, Sample Num: 5392, Cur Loss: 2.86336017, Cur Avg Loss: 1.45513106, Log Avg loss: 1.67257302, Global Avg Loss: 1.80025683, Time: 0.0402 Steps: 36530, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000347, Sample Num: 5552, Cur Loss: 1.73094440, Cur Avg Loss: 1.45880431, Log Avg loss: 1.58259259, Global Avg Loss: 1.80019727, Time: 0.0402 Steps: 36540, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000357, Sample Num: 5712, Cur Loss: 2.05977488, Cur Avg Loss: 1.45768514, Log Avg loss: 1.41885022, Global Avg Loss: 1.80009293, Time: 0.0402 Steps: 36550, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000367, Sample Num: 5872, Cur Loss: 1.45461428, Cur Avg Loss: 1.45979708, Log Avg loss: 1.53519313, Global Avg Loss: 1.80002047, Time: 0.0402 Steps: 36560, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000377, Sample Num: 6032, Cur Loss: 1.89657855, Cur Avg Loss: 1.46849375, Log Avg loss: 1.78766160, Global Avg Loss: 1.80001709, Time: 0.0402 Steps: 36570, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000387, Sample Num: 6192, Cur Loss: 1.71160054, Cur Avg Loss: 1.46181866, Log Avg loss: 1.21016790, Global Avg Loss: 1.79985584, Time: 0.0402 Steps: 36580, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000397, Sample Num: 6352, Cur Loss: 1.01594996, Cur Avg Loss: 1.46323068, Log Avg loss: 1.51787565, Global Avg Loss: 1.79977878, Time: 0.0402 Steps: 36590, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000407, Sample Num: 6512, Cur Loss: 2.56115055, Cur Avg Loss: 1.46802762, Log Avg loss: 1.65846604, Global Avg Loss: 1.79974017, Time: 0.0402 Steps: 36600, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000417, Sample Num: 6672, Cur Loss: 1.25420547, Cur Avg Loss: 1.46172831, Log Avg loss: 1.20534644, Global Avg Loss: 1.79957781, Time: 0.0402 Steps: 36610, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000427, Sample Num: 6832, Cur Loss: 1.40355980, Cur Avg Loss: 1.46018880, Log Avg loss: 1.39599122, Global Avg Loss: 1.79946760, Time: 0.0402 Steps: 36620, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000437, Sample Num: 6992, Cur Loss: 1.17586589, Cur Avg Loss: 1.45527231, Log Avg loss: 1.24533841, Global Avg Loss: 1.79931632, Time: 0.0404 Steps: 36630, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000447, Sample Num: 7152, Cur Loss: 1.46067238, Cur Avg Loss: 1.45902666, Log Avg loss: 1.62309176, Global Avg Loss: 1.79926823, Time: 0.0403 Steps: 36640, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000457, Sample Num: 7312, Cur Loss: 0.69250697, Cur Avg Loss: 1.45073407, Log Avg loss: 1.08005511, Global Avg Loss: 1.79907199, Time: 0.0402 Steps: 36650, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000467, Sample Num: 7472, Cur Loss: 1.23599923, Cur Avg Loss: 1.44768713, Log Avg loss: 1.30844225, Global Avg Loss: 1.79893816, Time: 0.0402 Steps: 36660, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000477, Sample Num: 7632, Cur Loss: 1.15228069, Cur Avg Loss: 1.44928023, Log Avg loss: 1.52367793, Global Avg Loss: 1.79886309, Time: 0.0402 Steps: 36670, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000487, Sample Num: 7792, Cur Loss: 1.15616524, Cur Avg Loss: 1.44017417, Log Avg loss: 1.00581495, Global Avg Loss: 1.79864689, Time: 0.0402 Steps: 36680, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000497, Sample Num: 7952, Cur Loss: 0.82204288, Cur Avg Loss: 1.43559884, Log Avg loss: 1.21278051, Global Avg Loss: 1.79848721, Time: 0.0402 Steps: 36690, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000507, Sample Num: 8112, Cur Loss: 0.61062270, Cur Avg Loss: 1.43053099, Log Avg loss: 1.17865889, Global Avg Loss: 1.79831832, Time: 0.0589 Steps: 36700, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000517, Sample Num: 8272, Cur Loss: 1.16198206, Cur Avg Loss: 1.42709175, Log Avg loss: 1.25272231, Global Avg Loss: 1.79816969, Time: 0.0633 Steps: 36710, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000527, Sample Num: 8432, Cur Loss: 1.56361103, Cur Avg Loss: 1.41956303, Log Avg loss: 1.03032777, Global Avg Loss: 1.79796059, Time: 0.0660 Steps: 36720, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000537, Sample Num: 8592, Cur Loss: 2.07423329, Cur Avg Loss: 1.41658481, Log Avg loss: 1.25963299, Global Avg Loss: 1.79781402, Time: 0.0404 Steps: 36730, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000547, Sample Num: 8752, Cur Loss: 1.71080184, Cur Avg Loss: 1.41908301, Log Avg loss: 1.55323589, Global Avg Loss: 1.79774745, Time: 0.0404 Steps: 36740, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000557, Sample Num: 8912, Cur Loss: 2.21484804, Cur Avg Loss: 1.41526316, Log Avg loss: 1.20631782, Global Avg Loss: 1.79758652, Time: 0.0405 Steps: 36750, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000567, Sample Num: 9072, Cur Loss: 1.22228396, Cur Avg Loss: 1.41251165, Log Avg loss: 1.25925217, Global Avg Loss: 1.79744007, Time: 0.0930 Steps: 36760, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000577, Sample Num: 9232, Cur Loss: 0.87555391, Cur Avg Loss: 1.40537589, Log Avg loss: 1.00077849, Global Avg Loss: 1.79722341, Time: 0.0532 Steps: 36770, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000587, Sample Num: 9392, Cur Loss: 0.79196888, Cur Avg Loss: 1.40406343, Log Avg loss: 1.32833424, Global Avg Loss: 1.79709593, Time: 0.0633 Steps: 36780, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000597, Sample Num: 9552, Cur Loss: 1.03270042, Cur Avg Loss: 1.39872882, Log Avg loss: 1.08558722, Global Avg Loss: 1.79690253, Time: 0.0402 Steps: 36790, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000607, Sample Num: 9712, Cur Loss: 1.32246387, Cur Avg Loss: 1.39356154, Log Avg loss: 1.08507536, Global Avg Loss: 1.79670910, Time: 0.0402 Steps: 36800, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000617, Sample Num: 9872, Cur Loss: 0.45373720, Cur Avg Loss: 1.39813050, Log Avg loss: 1.67546631, Global Avg Loss: 1.79667616, Time: 0.0402 Steps: 36810, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000627, Sample Num: 10032, Cur Loss: 2.39430499, Cur Avg Loss: 1.40213080, Log Avg loss: 1.64894946, Global Avg Loss: 1.79663604, Time: 0.0402 Steps: 36820, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000637, Sample Num: 10192, Cur Loss: 1.59036922, Cur Avg Loss: 1.40181186, Log Avg loss: 1.38181415, Global Avg Loss: 1.79652341, Time: 0.0403 Steps: 36830, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000647, Sample Num: 10352, Cur Loss: 1.51190114, Cur Avg Loss: 1.40244063, Log Avg loss: 1.44249302, Global Avg Loss: 1.79642731, Time: 0.0402 Steps: 36840, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000657, Sample Num: 10512, Cur Loss: 1.42676711, Cur Avg Loss: 1.40697523, Log Avg loss: 1.70036377, Global Avg Loss: 1.79640124, Time: 0.0402 Steps: 36850, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000667, Sample Num: 10672, Cur Loss: 2.02063465, Cur Avg Loss: 1.40933512, Log Avg loss: 1.56437995, Global Avg Loss: 1.79633829, Time: 0.0402 Steps: 36860, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000677, Sample Num: 10832, Cur Loss: 1.12743497, Cur Avg Loss: 1.40611018, Log Avg loss: 1.19100702, Global Avg Loss: 1.79617411, Time: 0.0402 Steps: 36870, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000687, Sample Num: 10992, Cur Loss: 2.05491400, Cur Avg Loss: 1.40558435, Log Avg loss: 1.36998537, Global Avg Loss: 1.79605855, Time: 0.0402 Steps: 36880, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000697, Sample Num: 11152, Cur Loss: 1.24849272, Cur Avg Loss: 1.40251652, Log Avg loss: 1.19175683, Global Avg Loss: 1.79589474, Time: 0.0403 Steps: 36890, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000707, Sample Num: 11312, Cur Loss: 2.12430763, Cur Avg Loss: 1.40128263, Log Avg loss: 1.31528031, Global Avg Loss: 1.79576449, Time: 0.0402 Steps: 36900, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000717, Sample Num: 11472, Cur Loss: 0.74178338, Cur Avg Loss: 1.40429339, Log Avg loss: 1.61715391, Global Avg Loss: 1.79571610, Time: 0.0403 Steps: 36910, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000727, Sample Num: 11632, Cur Loss: 1.30401754, Cur Avg Loss: 1.40620219, Log Avg loss: 1.54306369, Global Avg Loss: 1.79564767, Time: 0.0403 Steps: 36920, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000737, Sample Num: 11792, Cur Loss: 1.16121411, Cur Avg Loss: 1.40648971, Log Avg loss: 1.42739199, Global Avg Loss: 1.79554795, Time: 0.0403 Steps: 36930, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000747, Sample Num: 11952, Cur Loss: 1.19094610, Cur Avg Loss: 1.40301172, Log Avg loss: 1.14668420, Global Avg Loss: 1.79537230, Time: 0.0403 Steps: 36940, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000757, Sample Num: 12112, Cur Loss: 1.06263101, Cur Avg Loss: 1.39680170, Log Avg loss: 0.93291305, Global Avg Loss: 1.79513889, Time: 0.0402 Steps: 36950, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000767, Sample Num: 12272, Cur Loss: 0.54990554, Cur Avg Loss: 1.39686165, Log Avg loss: 1.40139987, Global Avg Loss: 1.79503236, Time: 0.0403 Steps: 36960, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000777, Sample Num: 12432, Cur Loss: 1.20143569, Cur Avg Loss: 1.39845011, Log Avg loss: 1.52028505, Global Avg Loss: 1.79495804, Time: 0.0403 Steps: 36970, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000787, Sample Num: 12592, Cur Loss: 1.61059904, Cur Avg Loss: 1.40052594, Log Avg loss: 1.56181786, Global Avg Loss: 1.79489499, Time: 0.0403 Steps: 36980, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000797, Sample Num: 12752, Cur Loss: 0.78507531, Cur Avg Loss: 1.40309178, Log Avg loss: 1.60502372, Global Avg Loss: 1.79484366, Time: 0.0403 Steps: 36990, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000807, Sample Num: 12912, Cur Loss: 2.65526128, Cur Avg Loss: 1.40537635, Log Avg loss: 1.58745639, Global Avg Loss: 1.79478761, Time: 0.0402 Steps: 37000, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000817, Sample Num: 13072, Cur Loss: 1.09802163, Cur Avg Loss: 1.40377212, Log Avg loss: 1.27431075, Global Avg Loss: 1.79464698, Time: 0.0402 Steps: 37010, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000827, Sample Num: 13232, Cur Loss: 1.49066091, Cur Avg Loss: 1.40287585, Log Avg loss: 1.32965021, Global Avg Loss: 1.79452137, Time: 0.0402 Steps: 37020, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000837, Sample Num: 13392, Cur Loss: 2.87622905, Cur Avg Loss: 1.40376808, Log Avg loss: 1.47755569, Global Avg Loss: 1.79443578, Time: 0.0402 Steps: 37030, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000847, Sample Num: 13552, Cur Loss: 2.17340374, Cur Avg Loss: 1.40458972, Log Avg loss: 1.47336128, Global Avg Loss: 1.79434909, Time: 0.0402 Steps: 37040, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000857, Sample Num: 13712, Cur Loss: 2.31409836, Cur Avg Loss: 1.40294533, Log Avg loss: 1.26366506, Global Avg Loss: 1.79420586, Time: 0.0676 Steps: 37050, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000867, Sample Num: 13872, Cur Loss: 0.60756737, Cur Avg Loss: 1.40230700, Log Avg loss: 1.34760262, Global Avg Loss: 1.79408535, Time: 0.0740 Steps: 37060, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000877, Sample Num: 14032, Cur Loss: 1.65372300, Cur Avg Loss: 1.40443327, Log Avg loss: 1.58878061, Global Avg Loss: 1.79402997, Time: 0.0863 Steps: 37070, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000887, Sample Num: 14192, Cur Loss: 0.85107255, Cur Avg Loss: 1.40855699, Log Avg loss: 1.77020726, Global Avg Loss: 1.79402354, Time: 0.0988 Steps: 37080, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000897, Sample Num: 14352, Cur Loss: 2.77959108, Cur Avg Loss: 1.40930174, Log Avg loss: 1.47536096, Global Avg Loss: 1.79393763, Time: 0.0404 Steps: 37090, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000907, Sample Num: 14512, Cur Loss: 0.99226457, Cur Avg Loss: 1.40796185, Log Avg loss: 1.28777390, Global Avg Loss: 1.79380120, Time: 0.0567 Steps: 37100, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000917, Sample Num: 14672, Cur Loss: 1.63271260, Cur Avg Loss: 1.40785109, Log Avg loss: 1.39780523, Global Avg Loss: 1.79369449, Time: 0.0589 Steps: 37110, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000927, Sample Num: 14832, Cur Loss: 0.95433605, Cur Avg Loss: 1.40603730, Log Avg loss: 1.23971286, Global Avg Loss: 1.79354525, Time: 0.0763 Steps: 37120, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000937, Sample Num: 14992, Cur Loss: 1.35213852, Cur Avg Loss: 1.40856578, Log Avg loss: 1.64295581, Global Avg Loss: 1.79350469, Time: 0.0404 Steps: 37130, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000947, Sample Num: 15152, Cur Loss: 1.40618253, Cur Avg Loss: 1.40662662, Log Avg loss: 1.22492732, Global Avg Loss: 1.79335160, Time: 0.0402 Steps: 37140, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000957, Sample Num: 15312, Cur Loss: 1.04753578, Cur Avg Loss: 1.41034736, Log Avg loss: 1.76270111, Global Avg Loss: 1.79334335, Time: 0.0402 Steps: 37150, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000967, Sample Num: 15472, Cur Loss: 2.06810546, Cur Avg Loss: 1.41316892, Log Avg loss: 1.68319273, Global Avg Loss: 1.79331371, Time: 0.0403 Steps: 37160, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000977, Sample Num: 15632, Cur Loss: 1.01056683, Cur Avg Loss: 1.41341286, Log Avg loss: 1.43700129, Global Avg Loss: 1.79321785, Time: 0.0403 Steps: 37170, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000987, Sample Num: 15792, Cur Loss: 1.67363465, Cur Avg Loss: 1.41215368, Log Avg loss: 1.28913237, Global Avg Loss: 1.79308227, Time: 0.0403 Steps: 37180, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000997, Sample Num: 15952, Cur Loss: 0.78322935, Cur Avg Loss: 1.40956662, Log Avg loss: 1.15422312, Global Avg Loss: 1.79291048, Time: 0.0403 Steps: 37190, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001007, Sample Num: 16112, Cur Loss: 1.01309741, Cur Avg Loss: 1.41324049, Log Avg loss: 1.77952589, Global Avg Loss: 1.79290689, Time: 0.0402 Steps: 37200, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001017, Sample Num: 16272, Cur Loss: 0.72066247, Cur Avg Loss: 1.41666883, Log Avg loss: 1.76190187, Global Avg Loss: 1.79289855, Time: 0.0402 Steps: 37210, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001027, Sample Num: 16432, Cur Loss: 3.01612043, Cur Avg Loss: 1.41846271, Log Avg loss: 1.60090062, Global Avg Loss: 1.79284697, Time: 0.0402 Steps: 37220, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001037, Sample Num: 16592, Cur Loss: 2.34421802, Cur Avg Loss: 1.41954854, Log Avg loss: 1.53106340, Global Avg Loss: 1.79277665, Time: 0.0402 Steps: 37230, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001047, Sample Num: 16752, Cur Loss: 1.31256342, Cur Avg Loss: 1.42026823, Log Avg loss: 1.49489988, Global Avg Loss: 1.79269666, Time: 0.0402 Steps: 37240, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001057, Sample Num: 16912, Cur Loss: 1.71588552, Cur Avg Loss: 1.42084527, Log Avg loss: 1.48126136, Global Avg Loss: 1.79261306, Time: 0.0402 Steps: 37250, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001067, Sample Num: 17072, Cur Loss: 1.10914087, Cur Avg Loss: 1.41952317, Log Avg loss: 1.27977757, Global Avg Loss: 1.79247542, Time: 0.0402 Steps: 37260, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001077, Sample Num: 17232, Cur Loss: 1.92154551, Cur Avg Loss: 1.42049550, Log Avg loss: 1.52424265, Global Avg Loss: 1.79240345, Time: 0.0403 Steps: 37270, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001087, Sample Num: 17392, Cur Loss: 1.58131051, Cur Avg Loss: 1.42217502, Log Avg loss: 1.60305969, Global Avg Loss: 1.79235266, Time: 0.0402 Steps: 37280, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001097, Sample Num: 17552, Cur Loss: 1.70522726, Cur Avg Loss: 1.42128259, Log Avg loss: 1.32427532, Global Avg Loss: 1.79222714, Time: 0.0403 Steps: 37290, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001107, Sample Num: 17712, Cur Loss: 2.35575581, Cur Avg Loss: 1.42195926, Log Avg loss: 1.49619017, Global Avg Loss: 1.79214777, Time: 0.0403 Steps: 37300, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001117, Sample Num: 17872, Cur Loss: 1.47684860, Cur Avg Loss: 1.42217642, Log Avg loss: 1.44621547, Global Avg Loss: 1.79205505, Time: 0.0403 Steps: 37310, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001127, Sample Num: 18032, Cur Loss: 0.64360118, Cur Avg Loss: 1.42293477, Log Avg loss: 1.50764301, Global Avg Loss: 1.79197884, Time: 0.0402 Steps: 37320, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001137, Sample Num: 18192, Cur Loss: 0.94932783, Cur Avg Loss: 1.42220771, Log Avg loss: 1.34026747, Global Avg Loss: 1.79185784, Time: 0.0403 Steps: 37330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001147, Sample Num: 18352, Cur Loss: 0.79220808, Cur Avg Loss: 1.41938420, Log Avg loss: 1.09835178, Global Avg Loss: 1.79167211, Time: 0.0402 Steps: 37340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001157, Sample Num: 18512, Cur Loss: 3.13486814, Cur Avg Loss: 1.42201230, Log Avg loss: 1.72345482, Global Avg Loss: 1.79165385, Time: 0.0403 Steps: 37350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001167, Sample Num: 18672, Cur Loss: 1.25720048, Cur Avg Loss: 1.42124793, Log Avg loss: 1.33281096, Global Avg Loss: 1.79153103, Time: 0.0403 Steps: 37360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001177, Sample Num: 18832, Cur Loss: 1.16884589, Cur Avg Loss: 1.42086374, Log Avg loss: 1.37602812, Global Avg Loss: 1.79141984, Time: 0.0402 Steps: 37370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001187, Sample Num: 18992, Cur Loss: 1.70942235, Cur Avg Loss: 1.42005648, Log Avg loss: 1.32504278, Global Avg Loss: 1.79129508, Time: 0.0403 Steps: 37380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001197, Sample Num: 19152, Cur Loss: 1.94343519, Cur Avg Loss: 1.42448098, Log Avg loss: 1.94966898, Global Avg Loss: 1.79133744, Time: 0.0654 Steps: 37390, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001207, Sample Num: 19312, Cur Loss: 1.58745909, Cur Avg Loss: 1.42596596, Log Avg loss: 1.60371721, Global Avg Loss: 1.79128727, Time: 0.1009 Steps: 37400, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001217, Sample Num: 19472, Cur Loss: 0.40147781, Cur Avg Loss: 1.42579945, Log Avg loss: 1.40570253, Global Avg Loss: 1.79118420, Time: 0.0502 Steps: 37410, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001227, Sample Num: 19632, Cur Loss: 1.84549379, Cur Avg Loss: 1.42706921, Log Avg loss: 1.58159847, Global Avg Loss: 1.79112819, Time: 0.0467 Steps: 37420, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001237, Sample Num: 19792, Cur Loss: 1.32503819, Cur Avg Loss: 1.42619117, Log Avg loss: 1.31845624, Global Avg Loss: 1.79100191, Time: 0.0504 Steps: 37430, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001247, Sample Num: 19952, Cur Loss: 0.39818710, Cur Avg Loss: 1.42569396, Log Avg loss: 1.36418900, Global Avg Loss: 1.79088791, Time: 0.0694 Steps: 37440, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001257, Sample Num: 20112, Cur Loss: 1.19231832, Cur Avg Loss: 1.42470429, Log Avg loss: 1.30129231, Global Avg Loss: 1.79075718, Time: 0.0997 Steps: 37450, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001267, Sample Num: 20272, Cur Loss: 0.43598273, Cur Avg Loss: 1.42377607, Log Avg loss: 1.30709898, Global Avg Loss: 1.79062806, Time: 0.0449 Steps: 37460, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001277, Sample Num: 20432, Cur Loss: 1.43406200, Cur Avg Loss: 1.42302606, Log Avg loss: 1.32799895, Global Avg Loss: 1.79050460, Time: 0.0532 Steps: 37470, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001287, Sample Num: 20592, Cur Loss: 2.32850027, Cur Avg Loss: 1.42474970, Log Avg loss: 1.64485895, Global Avg Loss: 1.79046574, Time: 0.0403 Steps: 37480, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001297, Sample Num: 20752, Cur Loss: 1.48939073, Cur Avg Loss: 1.42323910, Log Avg loss: 1.22882482, Global Avg Loss: 1.79031593, Time: 0.0402 Steps: 37490, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001307, Sample Num: 20912, Cur Loss: 1.08640027, Cur Avg Loss: 1.42519832, Log Avg loss: 1.67930938, Global Avg Loss: 1.79028632, Time: 0.0402 Steps: 37500, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001317, Sample Num: 21072, Cur Loss: 1.31508696, Cur Avg Loss: 1.42703854, Log Avg loss: 1.66755552, Global Avg Loss: 1.79025361, Time: 0.0402 Steps: 37510, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001327, Sample Num: 21232, Cur Loss: 3.94748354, Cur Avg Loss: 1.42772456, Log Avg loss: 1.51807284, Global Avg Loss: 1.79018106, Time: 0.0402 Steps: 37520, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001337, Sample Num: 21392, Cur Loss: 0.91415083, Cur Avg Loss: 1.42859439, Log Avg loss: 1.54402137, Global Avg Loss: 1.79011547, Time: 0.0403 Steps: 37530, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001347, Sample Num: 21552, Cur Loss: 0.35024899, Cur Avg Loss: 1.42730755, Log Avg loss: 1.25525705, Global Avg Loss: 1.78997300, Time: 0.0402 Steps: 37540, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001357, Sample Num: 21712, Cur Loss: 0.58149529, Cur Avg Loss: 1.42788298, Log Avg loss: 1.50539377, Global Avg Loss: 1.78989721, Time: 0.0403 Steps: 37550, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001367, Sample Num: 21872, Cur Loss: 2.58200073, Cur Avg Loss: 1.42737168, Log Avg loss: 1.35798711, Global Avg Loss: 1.78978222, Time: 0.0402 Steps: 37560, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001377, Sample Num: 22032, Cur Loss: 0.64778566, Cur Avg Loss: 1.42681567, Log Avg loss: 1.35080912, Global Avg Loss: 1.78966538, Time: 0.0403 Steps: 37570, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001387, Sample Num: 22192, Cur Loss: 1.84626913, Cur Avg Loss: 1.42787354, Log Avg loss: 1.57354312, Global Avg Loss: 1.78960787, Time: 0.0402 Steps: 37580, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001397, Sample Num: 22352, Cur Loss: 1.66709054, Cur Avg Loss: 1.42841819, Log Avg loss: 1.50396068, Global Avg Loss: 1.78953187, Time: 0.0403 Steps: 37590, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001407, Sample Num: 22512, Cur Loss: 0.85580134, Cur Avg Loss: 1.42728659, Log Avg loss: 1.26920164, Global Avg Loss: 1.78939349, Time: 0.0403 Steps: 37600, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001417, Sample Num: 22672, Cur Loss: 2.47352767, Cur Avg Loss: 1.42721808, Log Avg loss: 1.41757880, Global Avg Loss: 1.78929463, Time: 0.0402 Steps: 37610, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001427, Sample Num: 22832, Cur Loss: 0.52923703, Cur Avg Loss: 1.42610607, Log Avg loss: 1.26853414, Global Avg Loss: 1.78915620, Time: 0.0402 Steps: 37620, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001437, Sample Num: 22992, Cur Loss: 1.51355195, Cur Avg Loss: 1.42857426, Log Avg loss: 1.78078631, Global Avg Loss: 1.78915398, Time: 0.0402 Steps: 37630, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001447, Sample Num: 23152, Cur Loss: 2.46084833, Cur Avg Loss: 1.42906240, Log Avg loss: 1.49920794, Global Avg Loss: 1.78907695, Time: 0.0402 Steps: 37640, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001457, Sample Num: 23312, Cur Loss: 2.70698404, Cur Avg Loss: 1.42884948, Log Avg loss: 1.39803966, Global Avg Loss: 1.78897309, Time: 0.0402 Steps: 37650, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001467, Sample Num: 23472, Cur Loss: 0.74493647, Cur Avg Loss: 1.42710935, Log Avg loss: 1.17357181, Global Avg Loss: 1.78880968, Time: 0.0402 Steps: 37660, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001477, Sample Num: 23632, Cur Loss: 2.59685469, Cur Avg Loss: 1.43031551, Log Avg loss: 1.90066013, Global Avg Loss: 1.78883937, Time: 0.0403 Steps: 37670, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001487, Sample Num: 23792, Cur Loss: 1.06572330, Cur Avg Loss: 1.43031597, Log Avg loss: 1.43038275, Global Avg Loss: 1.78874424, Time: 0.0402 Steps: 37680, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001497, Sample Num: 23952, Cur Loss: 2.12179804, Cur Avg Loss: 1.43098696, Log Avg loss: 1.53076303, Global Avg Loss: 1.78867579, Time: 0.0402 Steps: 37690, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001507, Sample Num: 24112, Cur Loss: 1.30685854, Cur Avg Loss: 1.43090594, Log Avg loss: 1.41877851, Global Avg Loss: 1.78857767, Time: 0.0402 Steps: 37700, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001517, Sample Num: 24272, Cur Loss: 1.61553335, Cur Avg Loss: 1.42974311, Log Avg loss: 1.25450356, Global Avg Loss: 1.78843605, Time: 0.0403 Steps: 37710, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001527, Sample Num: 24432, Cur Loss: 2.42267394, Cur Avg Loss: 1.43049293, Log Avg loss: 1.54424037, Global Avg Loss: 1.78837131, Time: 0.0403 Steps: 37720, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001537, Sample Num: 24592, Cur Loss: 0.99291265, Cur Avg Loss: 1.43002658, Log Avg loss: 1.35881597, Global Avg Loss: 1.78825746, Time: 0.0402 Steps: 37730, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001547, Sample Num: 24752, Cur Loss: 1.23860002, Cur Avg Loss: 1.43075230, Log Avg loss: 1.54229583, Global Avg Loss: 1.78819228, Time: 0.0403 Steps: 37740, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001557, Sample Num: 24912, Cur Loss: 1.07761276, Cur Avg Loss: 1.43235931, Log Avg loss: 1.68096310, Global Avg Loss: 1.78816388, Time: 0.0662 Steps: 37750, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001567, Sample Num: 25072, Cur Loss: 1.12551916, Cur Avg Loss: 1.43136949, Log Avg loss: 1.27725504, Global Avg Loss: 1.78802857, Time: 0.0508 Steps: 37760, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001577, Sample Num: 25232, Cur Loss: 2.47903371, Cur Avg Loss: 1.43263215, Log Avg loss: 1.63049060, Global Avg Loss: 1.78798686, Time: 0.0747 Steps: 37770, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001587, Sample Num: 25392, Cur Loss: 1.94579387, Cur Avg Loss: 1.43207245, Log Avg loss: 1.34380788, Global Avg Loss: 1.78786929, Time: 0.0766 Steps: 37780, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001597, Sample Num: 25552, Cur Loss: 1.42948902, Cur Avg Loss: 1.43186489, Log Avg loss: 1.39892433, Global Avg Loss: 1.78776637, Time: 0.0835 Steps: 37790, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001607, Sample Num: 25712, Cur Loss: 1.45691752, Cur Avg Loss: 1.43254519, Log Avg loss: 1.54118967, Global Avg Loss: 1.78770114, Time: 0.0816 Steps: 37800, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001617, Sample Num: 25872, Cur Loss: 0.71964145, Cur Avg Loss: 1.43151934, Log Avg loss: 1.26666477, Global Avg Loss: 1.78756334, Time: 0.0531 Steps: 37810, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001627, Sample Num: 26032, Cur Loss: 0.50635529, Cur Avg Loss: 1.43093452, Log Avg loss: 1.33636956, Global Avg Loss: 1.78744404, Time: 0.0405 Steps: 37820, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001637, Sample Num: 26192, Cur Loss: 1.44546175, Cur Avg Loss: 1.43074201, Log Avg loss: 1.39942015, Global Avg Loss: 1.78734147, Time: 0.0427 Steps: 37830, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001647, Sample Num: 26352, Cur Loss: 1.92451429, Cur Avg Loss: 1.43101301, Log Avg loss: 1.47537561, Global Avg Loss: 1.78725902, Time: 0.0403 Steps: 37840, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001657, Sample Num: 26512, Cur Loss: 1.18266606, Cur Avg Loss: 1.43192740, Log Avg loss: 1.58252876, Global Avg Loss: 1.78720493, Time: 0.0403 Steps: 37850, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001667, Sample Num: 26672, Cur Loss: 1.11874676, Cur Avg Loss: 1.43129604, Log Avg loss: 1.32667838, Global Avg Loss: 1.78708329, Time: 0.0403 Steps: 37860, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001677, Sample Num: 26832, Cur Loss: 0.54340822, Cur Avg Loss: 1.42845076, Log Avg loss: 0.95414268, Global Avg Loss: 1.78686335, Time: 0.0403 Steps: 37870, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001687, Sample Num: 26992, Cur Loss: 1.07414985, Cur Avg Loss: 1.42955257, Log Avg loss: 1.61432672, Global Avg Loss: 1.78681780, Time: 0.0403 Steps: 37880, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001697, Sample Num: 27152, Cur Loss: 1.15969217, Cur Avg Loss: 1.42808162, Log Avg loss: 1.17993302, Global Avg Loss: 1.78665763, Time: 0.0403 Steps: 37890, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001707, Sample Num: 27312, Cur Loss: 0.73234802, Cur Avg Loss: 1.42603559, Log Avg loss: 1.07882375, Global Avg Loss: 1.78647086, Time: 0.0402 Steps: 37900, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001717, Sample Num: 27472, Cur Loss: 1.67343056, Cur Avg Loss: 1.42601193, Log Avg loss: 1.42197298, Global Avg Loss: 1.78637472, Time: 0.0403 Steps: 37910, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001727, Sample Num: 27632, Cur Loss: 1.25694060, Cur Avg Loss: 1.42489760, Log Avg loss: 1.23356786, Global Avg Loss: 1.78622893, Time: 0.0402 Steps: 37920, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001737, Sample Num: 27792, Cur Loss: 0.84888375, Cur Avg Loss: 1.42337234, Log Avg loss: 1.15995927, Global Avg Loss: 1.78606382, Time: 0.0402 Steps: 37930, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001747, Sample Num: 27952, Cur Loss: 2.19240713, Cur Avg Loss: 1.42450284, Log Avg loss: 1.62087122, Global Avg Loss: 1.78602028, Time: 0.0403 Steps: 37940, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001757, Sample Num: 28112, Cur Loss: 1.94225693, Cur Avg Loss: 1.42488271, Log Avg loss: 1.49124559, Global Avg Loss: 1.78594261, Time: 0.0402 Steps: 37950, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001767, Sample Num: 28272, Cur Loss: 1.56593192, Cur Avg Loss: 1.42484286, Log Avg loss: 1.41784179, Global Avg Loss: 1.78584564, Time: 0.0402 Steps: 37960, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001777, Sample Num: 28432, Cur Loss: 0.54420263, Cur Avg Loss: 1.42474257, Log Avg loss: 1.40702050, Global Avg Loss: 1.78574587, Time: 0.0403 Steps: 37970, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001787, Sample Num: 28592, Cur Loss: 1.92142737, Cur Avg Loss: 1.42679986, Log Avg loss: 1.79238106, Global Avg Loss: 1.78574761, Time: 0.0402 Steps: 37980, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001797, Sample Num: 28752, Cur Loss: 1.78673840, Cur Avg Loss: 1.42792423, Log Avg loss: 1.62884772, Global Avg Loss: 1.78570631, Time: 0.0402 Steps: 37990, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001807, Sample Num: 28912, Cur Loss: 3.17138696, Cur Avg Loss: 1.42796036, Log Avg loss: 1.43445362, Global Avg Loss: 1.78561388, Time: 0.0403 Steps: 38000, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001817, Sample Num: 29072, Cur Loss: 1.72955585, Cur Avg Loss: 1.42722891, Log Avg loss: 1.29505513, Global Avg Loss: 1.78548482, Time: 0.0402 Steps: 38010, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001827, Sample Num: 29232, Cur Loss: 1.22357857, Cur Avg Loss: 1.42624968, Log Avg loss: 1.24832404, Global Avg Loss: 1.78534353, Time: 0.0402 Steps: 38020, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001837, Sample Num: 29392, Cur Loss: 1.82098269, Cur Avg Loss: 1.42601005, Log Avg loss: 1.38222956, Global Avg Loss: 1.78523753, Time: 0.0402 Steps: 38030, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001847, Sample Num: 29552, Cur Loss: 0.91906363, Cur Avg Loss: 1.42484092, Log Avg loss: 1.21007149, Global Avg Loss: 1.78508633, Time: 0.0402 Steps: 38040, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001857, Sample Num: 29712, Cur Loss: 1.26509500, Cur Avg Loss: 1.42555219, Log Avg loss: 1.55692405, Global Avg Loss: 1.78502637, Time: 0.0402 Steps: 38050, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001867, Sample Num: 29872, Cur Loss: 1.07738400, Cur Avg Loss: 1.42494997, Log Avg loss: 1.31311886, Global Avg Loss: 1.78490238, Time: 0.0402 Steps: 38060, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001877, Sample Num: 30032, Cur Loss: 1.76492941, Cur Avg Loss: 1.42800370, Log Avg loss: 1.99813360, Global Avg Loss: 1.78495839, Time: 0.0402 Steps: 38070, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001887, Sample Num: 30192, Cur Loss: 2.24964714, Cur Avg Loss: 1.42732617, Log Avg loss: 1.30015464, Global Avg Loss: 1.78483108, Time: 0.0402 Steps: 38080, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001897, Sample Num: 30352, Cur Loss: 2.35196733, Cur Avg Loss: 1.42885410, Log Avg loss: 1.71717442, Global Avg Loss: 1.78481332, Time: 0.0402 Steps: 38090, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001907, Sample Num: 30512, Cur Loss: 1.66753757, Cur Avg Loss: 1.43020738, Log Avg loss: 1.68692408, Global Avg Loss: 1.78478762, Time: 0.0667 Steps: 38100, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001917, Sample Num: 30672, Cur Loss: 1.98730254, Cur Avg Loss: 1.42933652, Log Avg loss: 1.26326312, Global Avg Loss: 1.78465078, Time: 0.0709 Steps: 38110, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001927, Sample Num: 30832, Cur Loss: 2.31991529, Cur Avg Loss: 1.43064599, Log Avg loss: 1.68167281, Global Avg Loss: 1.78462376, Time: 0.0403 Steps: 38120, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001937, Sample Num: 30992, Cur Loss: 0.86686701, Cur Avg Loss: 1.43039261, Log Avg loss: 1.38156517, Global Avg Loss: 1.78451806, Time: 0.0884 Steps: 38130, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001947, Sample Num: 31152, Cur Loss: 1.17116797, Cur Avg Loss: 1.42964262, Log Avg loss: 1.28437095, Global Avg Loss: 1.78438692, Time: 0.0769 Steps: 38140, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001957, Sample Num: 31312, Cur Loss: 1.68411350, Cur Avg Loss: 1.43098936, Log Avg loss: 1.69319971, Global Avg Loss: 1.78436302, Time: 0.0525 Steps: 38150, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001967, Sample Num: 31472, Cur Loss: 1.71640849, Cur Avg Loss: 1.43298082, Log Avg loss: 1.82270926, Global Avg Loss: 1.78437307, Time: 0.0403 Steps: 38160, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001977, Sample Num: 31632, Cur Loss: 1.01332235, Cur Avg Loss: 1.43264279, Log Avg loss: 1.36615155, Global Avg Loss: 1.78426350, Time: 0.0406 Steps: 38170, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001987, Sample Num: 31792, Cur Loss: 1.38235629, Cur Avg Loss: 1.43274465, Log Avg loss: 1.45288154, Global Avg Loss: 1.78417670, Time: 0.0761 Steps: 38180, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001997, Sample Num: 31952, Cur Loss: 0.92568660, Cur Avg Loss: 1.43258469, Log Avg loss: 1.40080165, Global Avg Loss: 1.78407632, Time: 0.0403 Steps: 38190, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002007, Sample Num: 32112, Cur Loss: 1.64897954, Cur Avg Loss: 1.43219463, Log Avg loss: 1.35429942, Global Avg Loss: 1.78396381, Time: 0.0403 Steps: 38200, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002017, Sample Num: 32272, Cur Loss: 0.70383120, Cur Avg Loss: 1.43022856, Log Avg loss: 1.03563764, Global Avg Loss: 1.78376797, Time: 0.0402 Steps: 38210, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002027, Sample Num: 32432, Cur Loss: 1.05649245, Cur Avg Loss: 1.42912000, Log Avg loss: 1.20552468, Global Avg Loss: 1.78361667, Time: 0.0402 Steps: 38220, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002037, Sample Num: 32592, Cur Loss: 2.01302338, Cur Avg Loss: 1.42782119, Log Avg loss: 1.16455269, Global Avg Loss: 1.78345474, Time: 0.0403 Steps: 38230, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002047, Sample Num: 32752, Cur Loss: 2.29830885, Cur Avg Loss: 1.42886175, Log Avg loss: 1.64082323, Global Avg Loss: 1.78341744, Time: 0.0402 Steps: 38240, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002057, Sample Num: 32912, Cur Loss: 0.63617456, Cur Avg Loss: 1.42986635, Log Avg loss: 1.63550701, Global Avg Loss: 1.78337877, Time: 0.0403 Steps: 38250, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002067, Sample Num: 33072, Cur Loss: 2.02325487, Cur Avg Loss: 1.42933970, Log Avg loss: 1.32100841, Global Avg Loss: 1.78325792, Time: 0.0403 Steps: 38260, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002077, Sample Num: 33232, Cur Loss: 2.17753267, Cur Avg Loss: 1.42816161, Log Avg loss: 1.18465083, Global Avg Loss: 1.78310151, Time: 0.0404 Steps: 38270, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002087, Sample Num: 33392, Cur Loss: 1.87972188, Cur Avg Loss: 1.42989284, Log Avg loss: 1.78946883, Global Avg Loss: 1.78310317, Time: 0.0403 Steps: 38280, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002097, Sample Num: 33552, Cur Loss: 0.70669395, Cur Avg Loss: 1.42813984, Log Avg loss: 1.06228815, Global Avg Loss: 1.78291492, Time: 0.0403 Steps: 38290, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002107, Sample Num: 33712, Cur Loss: 3.17580795, Cur Avg Loss: 1.42756177, Log Avg loss: 1.30634217, Global Avg Loss: 1.78279049, Time: 0.0402 Steps: 38300, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002117, Sample Num: 33872, Cur Loss: 1.48707771, Cur Avg Loss: 1.42877534, Log Avg loss: 1.68447376, Global Avg Loss: 1.78276482, Time: 0.0403 Steps: 38310, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002127, Sample Num: 34032, Cur Loss: 1.58520436, Cur Avg Loss: 1.42973938, Log Avg loss: 1.63382704, Global Avg Loss: 1.78272596, Time: 0.0403 Steps: 38320, Updated lr: 0.000065 ***** Running evaluation checkpoint-38322 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-38322 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.177710, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.284458, "eval_total_loss": 902.973988, "eval_mae": 0.920207, "eval_mse": 1.28458, "eval_r2": 0.183437, "eval_sp_statistic": 0.347121, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.431023, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.237304, "test_total_loss": 621.126632, "test_mae": 0.830189, "test_mse": 1.237619, "test_r2": 0.20123, "test_sp_statistic": 0.355455, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.451007, "test_ps_pvalue": 0.0, "lr": 6.460692271218587e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7826788389244468, "train_cur_epoch_loss": 3042.8154962062836, "train_cur_epoch_avg_loss": 1.4292228728070848, "train_cur_epoch_time": 95.17770957946777, "train_cur_epoch_avg_time": 0.04470535912610041, "epoch": 18, "step": 38322} ################################################## Training, Epoch: 0019, Batch: 000008, Sample Num: 128, Cur Loss: 2.25704646, Cur Avg Loss: 1.80269523, Log Avg loss: 1.61813923, Global Avg Loss: 1.78268302, Time: 0.0403 Steps: 38330, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000018, Sample Num: 288, Cur Loss: 2.11352158, Cur Avg Loss: 1.89626607, Log Avg loss: 1.97112274, Global Avg Loss: 1.78273217, Time: 0.0402 Steps: 38340, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000028, Sample Num: 448, Cur Loss: 1.05212951, Cur Avg Loss: 1.86413762, Log Avg loss: 1.80630640, Global Avg Loss: 1.78273831, Time: 0.0402 Steps: 38350, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000038, Sample Num: 608, Cur Loss: 2.92817760, Cur Avg Loss: 1.83478352, Log Avg loss: 1.75259206, Global Avg Loss: 1.78273045, Time: 0.0402 Steps: 38360, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000048, Sample Num: 768, Cur Loss: 1.94974840, Cur Avg Loss: 1.73684710, Log Avg loss: 1.36468869, Global Avg Loss: 1.78262150, Time: 0.0403 Steps: 38370, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000058, Sample Num: 928, Cur Loss: 1.29135251, Cur Avg Loss: 1.67389044, Log Avg loss: 1.37169849, Global Avg Loss: 1.78251444, Time: 0.0403 Steps: 38380, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000068, Sample Num: 1088, Cur Loss: 0.71877140, Cur Avg Loss: 1.58020106, Log Avg loss: 1.03680263, Global Avg Loss: 1.78232019, Time: 0.0402 Steps: 38390, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000078, Sample Num: 1248, Cur Loss: 0.84272373, Cur Avg Loss: 1.55826819, Log Avg loss: 1.40912468, Global Avg Loss: 1.78222300, Time: 0.0402 Steps: 38400, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000088, Sample Num: 1408, Cur Loss: 3.09724545, Cur Avg Loss: 1.57429453, Log Avg loss: 1.69930002, Global Avg Loss: 1.78220142, Time: 0.0402 Steps: 38410, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000098, Sample Num: 1568, Cur Loss: 0.92754346, Cur Avg Loss: 1.55166181, Log Avg loss: 1.35249388, Global Avg Loss: 1.78208957, Time: 0.0402 Steps: 38420, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000108, Sample Num: 1728, Cur Loss: 0.92478585, Cur Avg Loss: 1.56521135, Log Avg loss: 1.69799681, Global Avg Loss: 1.78206769, Time: 0.0403 Steps: 38430, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000118, Sample Num: 1888, Cur Loss: 2.25638437, Cur Avg Loss: 1.53749543, Log Avg loss: 1.23816352, Global Avg Loss: 1.78192619, Time: 0.0402 Steps: 38440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000128, Sample Num: 2048, Cur Loss: 1.75832331, Cur Avg Loss: 1.53743922, Log Avg loss: 1.53677596, Global Avg Loss: 1.78186244, Time: 0.0402 Steps: 38450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000138, Sample Num: 2208, Cur Loss: 0.49447885, Cur Avg Loss: 1.50621120, Log Avg loss: 1.10649248, Global Avg Loss: 1.78168683, Time: 0.0403 Steps: 38460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000148, Sample Num: 2368, Cur Loss: 2.03627324, Cur Avg Loss: 1.54478548, Log Avg loss: 2.07711053, Global Avg Loss: 1.78176363, Time: 0.0403 Steps: 38470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000158, Sample Num: 2528, Cur Loss: 0.87352836, Cur Avg Loss: 1.52529253, Log Avg loss: 1.23679686, Global Avg Loss: 1.78162200, Time: 0.0402 Steps: 38480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000168, Sample Num: 2688, Cur Loss: 1.05709124, Cur Avg Loss: 1.51647281, Log Avg loss: 1.37712120, Global Avg Loss: 1.78151691, Time: 0.0402 Steps: 38490, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000178, Sample Num: 2848, Cur Loss: 1.13718557, Cur Avg Loss: 1.50737527, Log Avg loss: 1.35453671, Global Avg Loss: 1.78140601, Time: 0.0402 Steps: 38500, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000188, Sample Num: 3008, Cur Loss: 0.80860186, Cur Avg Loss: 1.49937735, Log Avg loss: 1.35701441, Global Avg Loss: 1.78129580, Time: 0.0511 Steps: 38510, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000198, Sample Num: 3168, Cur Loss: 1.00293779, Cur Avg Loss: 1.48249359, Log Avg loss: 1.16507882, Global Avg Loss: 1.78113583, Time: 0.0504 Steps: 38520, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000208, Sample Num: 3328, Cur Loss: 0.87842226, Cur Avg Loss: 1.46200808, Log Avg loss: 1.05639508, Global Avg Loss: 1.78094773, Time: 0.0402 Steps: 38530, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000218, Sample Num: 3488, Cur Loss: 1.09874701, Cur Avg Loss: 1.45974501, Log Avg loss: 1.41267299, Global Avg Loss: 1.78085218, Time: 0.0638 Steps: 38540, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000228, Sample Num: 3648, Cur Loss: 0.38378447, Cur Avg Loss: 1.44772851, Log Avg loss: 1.18576895, Global Avg Loss: 1.78069781, Time: 0.0531 Steps: 38550, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000238, Sample Num: 3808, Cur Loss: 0.94321108, Cur Avg Loss: 1.44285738, Log Avg loss: 1.33179562, Global Avg Loss: 1.78058139, Time: 0.0401 Steps: 38560, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000248, Sample Num: 3968, Cur Loss: 0.78753942, Cur Avg Loss: 1.43024100, Log Avg loss: 1.12997115, Global Avg Loss: 1.78041271, Time: 0.0925 Steps: 38570, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000258, Sample Num: 4128, Cur Loss: 1.54751611, Cur Avg Loss: 1.42895628, Log Avg loss: 1.39709526, Global Avg Loss: 1.78031335, Time: 0.0692 Steps: 38580, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000268, Sample Num: 4288, Cur Loss: 2.58402729, Cur Avg Loss: 1.43457577, Log Avg loss: 1.57955864, Global Avg Loss: 1.78026133, Time: 0.0405 Steps: 38590, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000278, Sample Num: 4448, Cur Loss: 2.84978247, Cur Avg Loss: 1.43566797, Log Avg loss: 1.46493870, Global Avg Loss: 1.78017964, Time: 0.0403 Steps: 38600, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000288, Sample Num: 4608, Cur Loss: 0.93398273, Cur Avg Loss: 1.43709563, Log Avg loss: 1.47678459, Global Avg Loss: 1.78010106, Time: 0.0403 Steps: 38610, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000298, Sample Num: 4768, Cur Loss: 1.03866315, Cur Avg Loss: 1.44559789, Log Avg loss: 1.69046297, Global Avg Loss: 1.78007785, Time: 0.0402 Steps: 38620, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000308, Sample Num: 4928, Cur Loss: 1.11738253, Cur Avg Loss: 1.45087113, Log Avg loss: 1.60801373, Global Avg Loss: 1.78003331, Time: 0.0401 Steps: 38630, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000318, Sample Num: 5088, Cur Loss: 1.22021210, Cur Avg Loss: 1.45189897, Log Avg loss: 1.48355651, Global Avg Loss: 1.77995658, Time: 0.0402 Steps: 38640, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000328, Sample Num: 5248, Cur Loss: 1.61736202, Cur Avg Loss: 1.44053552, Log Avg loss: 1.07917776, Global Avg Loss: 1.77977527, Time: 0.0402 Steps: 38650, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000338, Sample Num: 5408, Cur Loss: 0.96759415, Cur Avg Loss: 1.43772972, Log Avg loss: 1.34569943, Global Avg Loss: 1.77966299, Time: 0.0402 Steps: 38660, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000348, Sample Num: 5568, Cur Loss: 1.11767209, Cur Avg Loss: 1.43433519, Log Avg loss: 1.31960014, Global Avg Loss: 1.77954402, Time: 0.0402 Steps: 38670, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000358, Sample Num: 5728, Cur Loss: 1.31910586, Cur Avg Loss: 1.43666101, Log Avg loss: 1.51759960, Global Avg Loss: 1.77947630, Time: 0.0402 Steps: 38680, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000368, Sample Num: 5888, Cur Loss: 1.10479760, Cur Avg Loss: 1.43680702, Log Avg loss: 1.44203398, Global Avg Loss: 1.77938908, Time: 0.0403 Steps: 38690, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000378, Sample Num: 6048, Cur Loss: 0.95775288, Cur Avg Loss: 1.44189497, Log Avg loss: 1.62913152, Global Avg Loss: 1.77935025, Time: 0.0404 Steps: 38700, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000388, Sample Num: 6208, Cur Loss: 2.65411234, Cur Avg Loss: 1.44375195, Log Avg loss: 1.51394598, Global Avg Loss: 1.77928169, Time: 0.0403 Steps: 38710, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000398, Sample Num: 6368, Cur Loss: 0.97817469, Cur Avg Loss: 1.44200737, Log Avg loss: 1.37431754, Global Avg Loss: 1.77917710, Time: 0.0403 Steps: 38720, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000408, Sample Num: 6528, Cur Loss: 1.09501529, Cur Avg Loss: 1.44361641, Log Avg loss: 1.50765644, Global Avg Loss: 1.77910700, Time: 0.0402 Steps: 38730, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000418, Sample Num: 6688, Cur Loss: 1.93974745, Cur Avg Loss: 1.43675219, Log Avg loss: 1.15669186, Global Avg Loss: 1.77894633, Time: 0.0402 Steps: 38740, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000428, Sample Num: 6848, Cur Loss: 0.75864851, Cur Avg Loss: 1.43304764, Log Avg loss: 1.27819732, Global Avg Loss: 1.77881711, Time: 0.0402 Steps: 38750, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000438, Sample Num: 7008, Cur Loss: 2.07617283, Cur Avg Loss: 1.43722848, Log Avg loss: 1.61616864, Global Avg Loss: 1.77877514, Time: 0.0403 Steps: 38760, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000448, Sample Num: 7168, Cur Loss: 1.70535576, Cur Avg Loss: 1.43591343, Log Avg loss: 1.37831436, Global Avg Loss: 1.77867185, Time: 0.0402 Steps: 38770, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000458, Sample Num: 7328, Cur Loss: 0.43930623, Cur Avg Loss: 1.43202911, Log Avg loss: 1.25801149, Global Avg Loss: 1.77853759, Time: 0.0402 Steps: 38780, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000468, Sample Num: 7488, Cur Loss: 1.48796391, Cur Avg Loss: 1.43470825, Log Avg loss: 1.55741289, Global Avg Loss: 1.77848059, Time: 0.0403 Steps: 38790, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000478, Sample Num: 7648, Cur Loss: 0.56003904, Cur Avg Loss: 1.43386006, Log Avg loss: 1.39416485, Global Avg Loss: 1.77838154, Time: 0.0402 Steps: 38800, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000488, Sample Num: 7808, Cur Loss: 0.94942242, Cur Avg Loss: 1.42826940, Log Avg loss: 1.16103584, Global Avg Loss: 1.77822247, Time: 0.0403 Steps: 38810, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000498, Sample Num: 7968, Cur Loss: 1.57700169, Cur Avg Loss: 1.43720100, Log Avg loss: 1.87306312, Global Avg Loss: 1.77824690, Time: 0.0402 Steps: 38820, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000508, Sample Num: 8128, Cur Loss: 1.13653874, Cur Avg Loss: 1.43885169, Log Avg loss: 1.52105605, Global Avg Loss: 1.77818066, Time: 0.0402 Steps: 38830, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000518, Sample Num: 8288, Cur Loss: 1.05432212, Cur Avg Loss: 1.43593337, Log Avg loss: 1.28768250, Global Avg Loss: 1.77805438, Time: 0.0403 Steps: 38840, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000528, Sample Num: 8448, Cur Loss: 1.09486413, Cur Avg Loss: 1.43428349, Log Avg loss: 1.34881967, Global Avg Loss: 1.77794389, Time: 0.0402 Steps: 38850, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000538, Sample Num: 8608, Cur Loss: 1.01253033, Cur Avg Loss: 1.44004078, Log Avg loss: 1.74402575, Global Avg Loss: 1.77793516, Time: 0.0402 Steps: 38860, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000548, Sample Num: 8768, Cur Loss: 1.73364794, Cur Avg Loss: 1.43871748, Log Avg loss: 1.36752388, Global Avg Loss: 1.77782958, Time: 0.0648 Steps: 38870, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000558, Sample Num: 8928, Cur Loss: 1.16682792, Cur Avg Loss: 1.43397493, Log Avg loss: 1.17408324, Global Avg Loss: 1.77767429, Time: 0.0519 Steps: 38880, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000568, Sample Num: 9088, Cur Loss: 0.79047489, Cur Avg Loss: 1.43003023, Log Avg loss: 1.20991566, Global Avg Loss: 1.77752830, Time: 0.0404 Steps: 38890, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000578, Sample Num: 9248, Cur Loss: 0.77305621, Cur Avg Loss: 1.42287236, Log Avg loss: 1.01630542, Global Avg Loss: 1.77733261, Time: 0.0684 Steps: 38900, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000588, Sample Num: 9408, Cur Loss: 0.83292037, Cur Avg Loss: 1.41355374, Log Avg loss: 0.87493749, Global Avg Loss: 1.77710070, Time: 0.0584 Steps: 38910, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000598, Sample Num: 9568, Cur Loss: 0.96551722, Cur Avg Loss: 1.41400313, Log Avg loss: 1.44042727, Global Avg Loss: 1.77701419, Time: 0.0955 Steps: 38920, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000608, Sample Num: 9728, Cur Loss: 1.52171183, Cur Avg Loss: 1.41308663, Log Avg loss: 1.35828012, Global Avg Loss: 1.77690663, Time: 0.0406 Steps: 38930, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000618, Sample Num: 9888, Cur Loss: 0.81405973, Cur Avg Loss: 1.40761730, Log Avg loss: 1.07508207, Global Avg Loss: 1.77672640, Time: 0.0456 Steps: 38940, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000628, Sample Num: 10048, Cur Loss: 2.33861756, Cur Avg Loss: 1.40801018, Log Avg loss: 1.43229014, Global Avg Loss: 1.77663797, Time: 0.0772 Steps: 38950, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000638, Sample Num: 10208, Cur Loss: 2.87474823, Cur Avg Loss: 1.41780424, Log Avg loss: 2.03287115, Global Avg Loss: 1.77670374, Time: 0.0403 Steps: 38960, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000648, Sample Num: 10368, Cur Loss: 0.90326202, Cur Avg Loss: 1.41618985, Log Avg loss: 1.31319165, Global Avg Loss: 1.77658480, Time: 0.0402 Steps: 38970, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000658, Sample Num: 10528, Cur Loss: 0.89093804, Cur Avg Loss: 1.41438345, Log Avg loss: 1.29732873, Global Avg Loss: 1.77646185, Time: 0.0402 Steps: 38980, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000668, Sample Num: 10688, Cur Loss: 1.64718843, Cur Avg Loss: 1.41538183, Log Avg loss: 1.48107535, Global Avg Loss: 1.77638609, Time: 0.0402 Steps: 38990, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000678, Sample Num: 10848, Cur Loss: 1.48585975, Cur Avg Loss: 1.41169355, Log Avg loss: 1.16531624, Global Avg Loss: 1.77622940, Time: 0.0403 Steps: 39000, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000688, Sample Num: 11008, Cur Loss: 1.63205004, Cur Avg Loss: 1.40988060, Log Avg loss: 1.28696288, Global Avg Loss: 1.77610398, Time: 0.0403 Steps: 39010, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000698, Sample Num: 11168, Cur Loss: 1.28574467, Cur Avg Loss: 1.41027542, Log Avg loss: 1.43743898, Global Avg Loss: 1.77601719, Time: 0.0404 Steps: 39020, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000708, Sample Num: 11328, Cur Loss: 1.30810642, Cur Avg Loss: 1.41292130, Log Avg loss: 1.59760345, Global Avg Loss: 1.77597148, Time: 0.0403 Steps: 39030, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000718, Sample Num: 11488, Cur Loss: 1.00658011, Cur Avg Loss: 1.41294709, Log Avg loss: 1.41477312, Global Avg Loss: 1.77587896, Time: 0.0402 Steps: 39040, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000728, Sample Num: 11648, Cur Loss: 1.23827147, Cur Avg Loss: 1.41487969, Log Avg loss: 1.55364060, Global Avg Loss: 1.77582205, Time: 0.0403 Steps: 39050, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000738, Sample Num: 11808, Cur Loss: 1.36629939, Cur Avg Loss: 1.41017114, Log Avg loss: 1.06738858, Global Avg Loss: 1.77564067, Time: 0.0403 Steps: 39060, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000748, Sample Num: 11968, Cur Loss: 0.28736681, Cur Avg Loss: 1.40581268, Log Avg loss: 1.08415847, Global Avg Loss: 1.77546369, Time: 0.0404 Steps: 39070, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000758, Sample Num: 12128, Cur Loss: 1.72502756, Cur Avg Loss: 1.40578954, Log Avg loss: 1.40405857, Global Avg Loss: 1.77536865, Time: 0.0402 Steps: 39080, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000768, Sample Num: 12288, Cur Loss: 0.66246951, Cur Avg Loss: 1.40222121, Log Avg loss: 1.13174225, Global Avg Loss: 1.77520400, Time: 0.0403 Steps: 39090, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000778, Sample Num: 12448, Cur Loss: 0.81220472, Cur Avg Loss: 1.40201843, Log Avg loss: 1.38644448, Global Avg Loss: 1.77510457, Time: 0.0402 Steps: 39100, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000788, Sample Num: 12608, Cur Loss: 2.18965435, Cur Avg Loss: 1.40366597, Log Avg loss: 1.53184498, Global Avg Loss: 1.77504237, Time: 0.0403 Steps: 39110, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000798, Sample Num: 12768, Cur Loss: 0.90495116, Cur Avg Loss: 1.40255540, Log Avg loss: 1.31504176, Global Avg Loss: 1.77492479, Time: 0.0402 Steps: 39120, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000808, Sample Num: 12928, Cur Loss: 1.85835075, Cur Avg Loss: 1.40504976, Log Avg loss: 1.60409983, Global Avg Loss: 1.77488113, Time: 0.0402 Steps: 39130, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000818, Sample Num: 13088, Cur Loss: 1.02838588, Cur Avg Loss: 1.40132623, Log Avg loss: 1.10046505, Global Avg Loss: 1.77470882, Time: 0.0402 Steps: 39140, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000828, Sample Num: 13248, Cur Loss: 1.26195240, Cur Avg Loss: 1.40146830, Log Avg loss: 1.41309016, Global Avg Loss: 1.77461646, Time: 0.0402 Steps: 39150, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000838, Sample Num: 13408, Cur Loss: 1.91880083, Cur Avg Loss: 1.40440761, Log Avg loss: 1.64778233, Global Avg Loss: 1.77458407, Time: 0.0402 Steps: 39160, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000848, Sample Num: 13568, Cur Loss: 0.94760352, Cur Avg Loss: 1.40333654, Log Avg loss: 1.31358082, Global Avg Loss: 1.77446637, Time: 0.0402 Steps: 39170, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000858, Sample Num: 13728, Cur Loss: 3.44075704, Cur Avg Loss: 1.40633508, Log Avg loss: 1.66061096, Global Avg Loss: 1.77443731, Time: 0.0402 Steps: 39180, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000868, Sample Num: 13888, Cur Loss: 2.23638034, Cur Avg Loss: 1.40718547, Log Avg loss: 1.48014886, Global Avg Loss: 1.77436222, Time: 0.0402 Steps: 39190, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000878, Sample Num: 14048, Cur Loss: 1.58146715, Cur Avg Loss: 1.40789584, Log Avg loss: 1.46955606, Global Avg Loss: 1.77428446, Time: 0.0402 Steps: 39200, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000888, Sample Num: 14208, Cur Loss: 2.91262102, Cur Avg Loss: 1.40805129, Log Avg loss: 1.42170010, Global Avg Loss: 1.77419454, Time: 0.0402 Steps: 39210, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000898, Sample Num: 14368, Cur Loss: 1.43957937, Cur Avg Loss: 1.41140871, Log Avg loss: 1.70954781, Global Avg Loss: 1.77417806, Time: 0.0504 Steps: 39220, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000908, Sample Num: 14528, Cur Loss: 2.09513617, Cur Avg Loss: 1.41437865, Log Avg loss: 1.68107914, Global Avg Loss: 1.77415433, Time: 0.0673 Steps: 39230, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000918, Sample Num: 14688, Cur Loss: 1.14968812, Cur Avg Loss: 1.41446603, Log Avg loss: 1.42239994, Global Avg Loss: 1.77406469, Time: 0.0543 Steps: 39240, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000928, Sample Num: 14848, Cur Loss: 1.54148555, Cur Avg Loss: 1.41377553, Log Avg loss: 1.35038740, Global Avg Loss: 1.77395674, Time: 0.0405 Steps: 39250, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000938, Sample Num: 15008, Cur Loss: 1.31043565, Cur Avg Loss: 1.41432359, Log Avg loss: 1.46518376, Global Avg Loss: 1.77387809, Time: 0.0602 Steps: 39260, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000948, Sample Num: 15168, Cur Loss: 1.32363057, Cur Avg Loss: 1.41660003, Log Avg loss: 1.63012977, Global Avg Loss: 1.77384149, Time: 0.0406 Steps: 39270, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000958, Sample Num: 15328, Cur Loss: 2.55880666, Cur Avg Loss: 1.41463767, Log Avg loss: 1.22860665, Global Avg Loss: 1.77370268, Time: 0.0466 Steps: 39280, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000968, Sample Num: 15488, Cur Loss: 1.15465081, Cur Avg Loss: 1.41696424, Log Avg loss: 1.63984954, Global Avg Loss: 1.77366861, Time: 0.0985 Steps: 39290, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000978, Sample Num: 15648, Cur Loss: 1.02677023, Cur Avg Loss: 1.41665911, Log Avg loss: 1.38712232, Global Avg Loss: 1.77357026, Time: 0.0995 Steps: 39300, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000988, Sample Num: 15808, Cur Loss: 1.79124105, Cur Avg Loss: 1.41587134, Log Avg loss: 1.33882761, Global Avg Loss: 1.77345966, Time: 0.0403 Steps: 39310, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000998, Sample Num: 15968, Cur Loss: 1.20508587, Cur Avg Loss: 1.41501264, Log Avg loss: 1.33017254, Global Avg Loss: 1.77334692, Time: 0.0402 Steps: 39320, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001008, Sample Num: 16128, Cur Loss: 3.24366570, Cur Avg Loss: 1.41584572, Log Avg loss: 1.49898729, Global Avg Loss: 1.77327717, Time: 0.0403 Steps: 39330, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001018, Sample Num: 16288, Cur Loss: 1.94172168, Cur Avg Loss: 1.41587902, Log Avg loss: 1.41923586, Global Avg Loss: 1.77318717, Time: 0.0403 Steps: 39340, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001028, Sample Num: 16448, Cur Loss: 1.89576888, Cur Avg Loss: 1.41655686, Log Avg loss: 1.48556064, Global Avg Loss: 1.77311408, Time: 0.0403 Steps: 39350, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001038, Sample Num: 16608, Cur Loss: 1.77769661, Cur Avg Loss: 1.41578172, Log Avg loss: 1.33609793, Global Avg Loss: 1.77300305, Time: 0.0402 Steps: 39360, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001048, Sample Num: 16768, Cur Loss: 0.74352068, Cur Avg Loss: 1.41594377, Log Avg loss: 1.43276430, Global Avg Loss: 1.77291663, Time: 0.0403 Steps: 39370, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001058, Sample Num: 16928, Cur Loss: 1.13455963, Cur Avg Loss: 1.41460866, Log Avg loss: 1.27468918, Global Avg Loss: 1.77279011, Time: 0.0402 Steps: 39380, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001068, Sample Num: 17088, Cur Loss: 2.10981846, Cur Avg Loss: 1.41794882, Log Avg loss: 1.77133794, Global Avg Loss: 1.77278974, Time: 0.0402 Steps: 39390, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001078, Sample Num: 17248, Cur Loss: 1.74307156, Cur Avg Loss: 1.41868705, Log Avg loss: 1.49752958, Global Avg Loss: 1.77271988, Time: 0.0403 Steps: 39400, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001088, Sample Num: 17408, Cur Loss: 2.38738656, Cur Avg Loss: 1.41937462, Log Avg loss: 1.49349462, Global Avg Loss: 1.77264902, Time: 0.0402 Steps: 39410, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001098, Sample Num: 17568, Cur Loss: 1.15234411, Cur Avg Loss: 1.41898191, Log Avg loss: 1.37625502, Global Avg Loss: 1.77254847, Time: 0.0404 Steps: 39420, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001108, Sample Num: 17728, Cur Loss: 1.31764054, Cur Avg Loss: 1.42068311, Log Avg loss: 1.60747465, Global Avg Loss: 1.77250660, Time: 0.0403 Steps: 39430, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001118, Sample Num: 17888, Cur Loss: 1.04963636, Cur Avg Loss: 1.41982113, Log Avg loss: 1.32431442, Global Avg Loss: 1.77239296, Time: 0.0402 Steps: 39440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001128, Sample Num: 18048, Cur Loss: 0.35144734, Cur Avg Loss: 1.41875461, Log Avg loss: 1.29951701, Global Avg Loss: 1.77227310, Time: 0.0402 Steps: 39450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001138, Sample Num: 18208, Cur Loss: 1.60506403, Cur Avg Loss: 1.41653736, Log Avg loss: 1.16643189, Global Avg Loss: 1.77211956, Time: 0.0402 Steps: 39460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001148, Sample Num: 18368, Cur Loss: 1.23806608, Cur Avg Loss: 1.41687881, Log Avg loss: 1.45573590, Global Avg Loss: 1.77203941, Time: 0.0403 Steps: 39470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001158, Sample Num: 18528, Cur Loss: 0.59445417, Cur Avg Loss: 1.41639227, Log Avg loss: 1.36053731, Global Avg Loss: 1.77193518, Time: 0.0402 Steps: 39480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001168, Sample Num: 18688, Cur Loss: 2.10208941, Cur Avg Loss: 1.41651722, Log Avg loss: 1.43098698, Global Avg Loss: 1.77184884, Time: 0.0402 Steps: 39490, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001178, Sample Num: 18848, Cur Loss: 1.55472660, Cur Avg Loss: 1.41806518, Log Avg loss: 1.59886695, Global Avg Loss: 1.77180504, Time: 0.0402 Steps: 39500, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001188, Sample Num: 19008, Cur Loss: 1.23229146, Cur Avg Loss: 1.41841683, Log Avg loss: 1.45984108, Global Avg Loss: 1.77172609, Time: 0.0403 Steps: 39510, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001198, Sample Num: 19168, Cur Loss: 0.78288376, Cur Avg Loss: 1.41674440, Log Avg loss: 1.21805883, Global Avg Loss: 1.77158599, Time: 0.0402 Steps: 39520, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001208, Sample Num: 19328, Cur Loss: 0.68294680, Cur Avg Loss: 1.41459296, Log Avg loss: 1.15685147, Global Avg Loss: 1.77143048, Time: 0.0402 Steps: 39530, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001218, Sample Num: 19488, Cur Loss: 2.76944542, Cur Avg Loss: 1.41415315, Log Avg loss: 1.36102405, Global Avg Loss: 1.77132668, Time: 0.0402 Steps: 39540, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001228, Sample Num: 19648, Cur Loss: 2.74924326, Cur Avg Loss: 1.41453576, Log Avg loss: 1.46113754, Global Avg Loss: 1.77124825, Time: 0.0402 Steps: 39550, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001238, Sample Num: 19808, Cur Loss: 0.67854190, Cur Avg Loss: 1.41129633, Log Avg loss: 1.01349442, Global Avg Loss: 1.77105671, Time: 0.0835 Steps: 39560, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001248, Sample Num: 19968, Cur Loss: 0.63149536, Cur Avg Loss: 1.41325176, Log Avg loss: 1.65533378, Global Avg Loss: 1.77102746, Time: 0.0517 Steps: 39570, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001258, Sample Num: 20128, Cur Loss: 1.75712943, Cur Avg Loss: 1.41542436, Log Avg loss: 1.68656436, Global Avg Loss: 1.77100612, Time: 0.0555 Steps: 39580, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001268, Sample Num: 20288, Cur Loss: 0.93670988, Cur Avg Loss: 1.41438096, Log Avg loss: 1.28312151, Global Avg Loss: 1.77088289, Time: 0.0650 Steps: 39590, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001278, Sample Num: 20448, Cur Loss: 1.91335583, Cur Avg Loss: 1.41556827, Log Avg loss: 1.56611955, Global Avg Loss: 1.77083118, Time: 0.0684 Steps: 39600, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001288, Sample Num: 20608, Cur Loss: 1.16877699, Cur Avg Loss: 1.41569704, Log Avg loss: 1.43215297, Global Avg Loss: 1.77074568, Time: 0.0464 Steps: 39610, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001298, Sample Num: 20768, Cur Loss: 0.32341141, Cur Avg Loss: 1.41323252, Log Avg loss: 1.09580287, Global Avg Loss: 1.77057532, Time: 0.0405 Steps: 39620, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001308, Sample Num: 20928, Cur Loss: 1.90196478, Cur Avg Loss: 1.41418651, Log Avg loss: 1.53801380, Global Avg Loss: 1.77051664, Time: 0.0797 Steps: 39630, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001318, Sample Num: 21088, Cur Loss: 0.70297319, Cur Avg Loss: 1.41344468, Log Avg loss: 1.31641413, Global Avg Loss: 1.77040208, Time: 0.0403 Steps: 39640, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001328, Sample Num: 21248, Cur Loss: 0.74665976, Cur Avg Loss: 1.41307548, Log Avg loss: 1.36441520, Global Avg Loss: 1.77029969, Time: 0.0403 Steps: 39650, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001338, Sample Num: 21408, Cur Loss: 2.04265404, Cur Avg Loss: 1.41299813, Log Avg loss: 1.40272548, Global Avg Loss: 1.77020701, Time: 0.0402 Steps: 39660, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001348, Sample Num: 21568, Cur Loss: 1.98274589, Cur Avg Loss: 1.41298496, Log Avg loss: 1.41122338, Global Avg Loss: 1.77011652, Time: 0.0402 Steps: 39670, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001358, Sample Num: 21728, Cur Loss: 0.88209927, Cur Avg Loss: 1.41197209, Log Avg loss: 1.27543696, Global Avg Loss: 1.76999185, Time: 0.0403 Steps: 39680, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001368, Sample Num: 21888, Cur Loss: 0.62213486, Cur Avg Loss: 1.41458208, Log Avg loss: 1.76901831, Global Avg Loss: 1.76999160, Time: 0.0402 Steps: 39690, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001378, Sample Num: 22048, Cur Loss: 1.41409314, Cur Avg Loss: 1.41478235, Log Avg loss: 1.44217917, Global Avg Loss: 1.76990903, Time: 0.0402 Steps: 39700, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001388, Sample Num: 22208, Cur Loss: 1.45626318, Cur Avg Loss: 1.41451935, Log Avg loss: 1.37827883, Global Avg Loss: 1.76981041, Time: 0.0402 Steps: 39710, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001398, Sample Num: 22368, Cur Loss: 2.38832569, Cur Avg Loss: 1.41536181, Log Avg loss: 1.53229532, Global Avg Loss: 1.76975061, Time: 0.0403 Steps: 39720, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001408, Sample Num: 22528, Cur Loss: 1.50658226, Cur Avg Loss: 1.41601823, Log Avg loss: 1.50778558, Global Avg Loss: 1.76968468, Time: 0.0404 Steps: 39730, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001418, Sample Num: 22688, Cur Loss: 2.10036564, Cur Avg Loss: 1.41726151, Log Avg loss: 1.59231492, Global Avg Loss: 1.76964004, Time: 0.0403 Steps: 39740, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001428, Sample Num: 22848, Cur Loss: 1.14512885, Cur Avg Loss: 1.41716915, Log Avg loss: 1.40407248, Global Avg Loss: 1.76954808, Time: 0.0403 Steps: 39750, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001438, Sample Num: 23008, Cur Loss: 1.19512641, Cur Avg Loss: 1.41580524, Log Avg loss: 1.22103915, Global Avg Loss: 1.76941012, Time: 0.0402 Steps: 39760, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001448, Sample Num: 23168, Cur Loss: 2.14509249, Cur Avg Loss: 1.41615615, Log Avg loss: 1.46661666, Global Avg Loss: 1.76933398, Time: 0.0402 Steps: 39770, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001458, Sample Num: 23328, Cur Loss: 2.48867226, Cur Avg Loss: 1.41619988, Log Avg loss: 1.42253193, Global Avg Loss: 1.76924680, Time: 0.0403 Steps: 39780, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001468, Sample Num: 23488, Cur Loss: 2.37354159, Cur Avg Loss: 1.41765689, Log Avg loss: 1.63008962, Global Avg Loss: 1.76921183, Time: 0.0403 Steps: 39790, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001478, Sample Num: 23648, Cur Loss: 1.49685717, Cur Avg Loss: 1.41821904, Log Avg loss: 1.50074140, Global Avg Loss: 1.76914438, Time: 0.0402 Steps: 39800, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001488, Sample Num: 23808, Cur Loss: 1.14413857, Cur Avg Loss: 1.41988794, Log Avg loss: 1.66655214, Global Avg Loss: 1.76911861, Time: 0.0402 Steps: 39810, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001498, Sample Num: 23968, Cur Loss: 0.57698667, Cur Avg Loss: 1.41753889, Log Avg loss: 1.06799974, Global Avg Loss: 1.76894253, Time: 0.0403 Steps: 39820, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001508, Sample Num: 24128, Cur Loss: 2.09406090, Cur Avg Loss: 1.41736295, Log Avg loss: 1.39100740, Global Avg Loss: 1.76884765, Time: 0.0402 Steps: 39830, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001518, Sample Num: 24288, Cur Loss: 0.77677488, Cur Avg Loss: 1.41754322, Log Avg loss: 1.44472796, Global Avg Loss: 1.76876629, Time: 0.0402 Steps: 39840, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001528, Sample Num: 24448, Cur Loss: 1.13550484, Cur Avg Loss: 1.41553736, Log Avg loss: 1.11104800, Global Avg Loss: 1.76860124, Time: 0.0403 Steps: 39850, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001538, Sample Num: 24608, Cur Loss: 1.21332157, Cur Avg Loss: 1.41578814, Log Avg loss: 1.45410745, Global Avg Loss: 1.76852234, Time: 0.0403 Steps: 39860, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001548, Sample Num: 24768, Cur Loss: 1.75315356, Cur Avg Loss: 1.41634455, Log Avg loss: 1.50191948, Global Avg Loss: 1.76845548, Time: 0.0402 Steps: 39870, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001558, Sample Num: 24928, Cur Loss: 0.74923909, Cur Avg Loss: 1.41457834, Log Avg loss: 1.14117014, Global Avg Loss: 1.76829818, Time: 0.0402 Steps: 39880, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001568, Sample Num: 25088, Cur Loss: 0.68631291, Cur Avg Loss: 1.41447714, Log Avg loss: 1.39870933, Global Avg Loss: 1.76820553, Time: 0.0403 Steps: 39890, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001578, Sample Num: 25248, Cur Loss: 2.09735894, Cur Avg Loss: 1.41795677, Log Avg loss: 1.96356288, Global Avg Loss: 1.76825449, Time: 0.0403 Steps: 39900, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001588, Sample Num: 25408, Cur Loss: 0.90925664, Cur Avg Loss: 1.41558930, Log Avg loss: 1.04200326, Global Avg Loss: 1.76807252, Time: 0.0406 Steps: 39910, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001598, Sample Num: 25568, Cur Loss: 0.73478818, Cur Avg Loss: 1.41475992, Log Avg loss: 1.28305475, Global Avg Loss: 1.76795102, Time: 0.0408 Steps: 39920, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001608, Sample Num: 25728, Cur Loss: 0.90833819, Cur Avg Loss: 1.41450061, Log Avg loss: 1.37306206, Global Avg Loss: 1.76785213, Time: 0.0994 Steps: 39930, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001618, Sample Num: 25888, Cur Loss: 1.26771426, Cur Avg Loss: 1.41228134, Log Avg loss: 1.05542281, Global Avg Loss: 1.76767375, Time: 0.0713 Steps: 39940, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001628, Sample Num: 26048, Cur Loss: 1.48221862, Cur Avg Loss: 1.41361660, Log Avg loss: 1.62966093, Global Avg Loss: 1.76763921, Time: 0.0404 Steps: 39950, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001638, Sample Num: 26208, Cur Loss: 1.25928760, Cur Avg Loss: 1.41276936, Log Avg loss: 1.27483866, Global Avg Loss: 1.76751588, Time: 0.0596 Steps: 39960, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001648, Sample Num: 26368, Cur Loss: 1.47906768, Cur Avg Loss: 1.41088908, Log Avg loss: 1.10289945, Global Avg Loss: 1.76734960, Time: 0.0595 Steps: 39970, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001658, Sample Num: 26528, Cur Loss: 2.39796782, Cur Avg Loss: 1.41333324, Log Avg loss: 1.81613197, Global Avg Loss: 1.76736181, Time: 0.0999 Steps: 39980, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001668, Sample Num: 26688, Cur Loss: 0.30384415, Cur Avg Loss: 1.41125296, Log Avg loss: 1.06634144, Global Avg Loss: 1.76718651, Time: 0.0855 Steps: 39990, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001678, Sample Num: 26848, Cur Loss: 0.45139399, Cur Avg Loss: 1.40982756, Log Avg loss: 1.17207164, Global Avg Loss: 1.76703773, Time: 0.0402 Steps: 40000, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001688, Sample Num: 27008, Cur Loss: 1.21577668, Cur Avg Loss: 1.41068031, Log Avg loss: 1.55377135, Global Avg Loss: 1.76698442, Time: 0.0403 Steps: 40010, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001698, Sample Num: 27168, Cur Loss: 1.01952672, Cur Avg Loss: 1.41073151, Log Avg loss: 1.41937419, Global Avg Loss: 1.76689757, Time: 0.0402 Steps: 40020, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001708, Sample Num: 27328, Cur Loss: 1.35670435, Cur Avg Loss: 1.41175132, Log Avg loss: 1.58491529, Global Avg Loss: 1.76685210, Time: 0.0403 Steps: 40030, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001718, Sample Num: 27488, Cur Loss: 1.58778286, Cur Avg Loss: 1.40992061, Log Avg loss: 1.09723511, Global Avg Loss: 1.76668487, Time: 0.0403 Steps: 40040, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001728, Sample Num: 27648, Cur Loss: 2.21397877, Cur Avg Loss: 1.40892652, Log Avg loss: 1.23814166, Global Avg Loss: 1.76655290, Time: 0.0403 Steps: 40050, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001738, Sample Num: 27808, Cur Loss: 0.67261231, Cur Avg Loss: 1.40745691, Log Avg loss: 1.15350879, Global Avg Loss: 1.76639986, Time: 0.0402 Steps: 40060, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001748, Sample Num: 27968, Cur Loss: 1.01031709, Cur Avg Loss: 1.40791931, Log Avg loss: 1.48828404, Global Avg Loss: 1.76633046, Time: 0.0403 Steps: 40070, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001758, Sample Num: 28128, Cur Loss: 1.76357341, Cur Avg Loss: 1.40809786, Log Avg loss: 1.43930891, Global Avg Loss: 1.76624886, Time: 0.0402 Steps: 40080, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001768, Sample Num: 28288, Cur Loss: 2.09318018, Cur Avg Loss: 1.40784930, Log Avg loss: 1.36415237, Global Avg Loss: 1.76614857, Time: 0.0403 Steps: 40090, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001778, Sample Num: 28448, Cur Loss: 1.04477143, Cur Avg Loss: 1.40752000, Log Avg loss: 1.34929885, Global Avg Loss: 1.76604461, Time: 0.0402 Steps: 40100, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001788, Sample Num: 28608, Cur Loss: 1.35100555, Cur Avg Loss: 1.40732678, Log Avg loss: 1.37297291, Global Avg Loss: 1.76594662, Time: 0.0402 Steps: 40110, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001798, Sample Num: 28768, Cur Loss: 1.80697107, Cur Avg Loss: 1.40946019, Log Avg loss: 1.79091446, Global Avg Loss: 1.76595284, Time: 0.0402 Steps: 40120, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001808, Sample Num: 28928, Cur Loss: 1.51319599, Cur Avg Loss: 1.41088881, Log Avg loss: 1.66775423, Global Avg Loss: 1.76592837, Time: 0.0402 Steps: 40130, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001818, Sample Num: 29088, Cur Loss: 0.66346705, Cur Avg Loss: 1.41132697, Log Avg loss: 1.49054663, Global Avg Loss: 1.76585976, Time: 0.0403 Steps: 40140, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001828, Sample Num: 29248, Cur Loss: 1.09824693, Cur Avg Loss: 1.41230481, Log Avg loss: 1.59007589, Global Avg Loss: 1.76581598, Time: 0.0403 Steps: 40150, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001838, Sample Num: 29408, Cur Loss: 1.39290142, Cur Avg Loss: 1.41027627, Log Avg loss: 1.03945833, Global Avg Loss: 1.76563512, Time: 0.0402 Steps: 40160, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001848, Sample Num: 29568, Cur Loss: 2.39528751, Cur Avg Loss: 1.41175222, Log Avg loss: 1.68303292, Global Avg Loss: 1.76561455, Time: 0.0402 Steps: 40170, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001858, Sample Num: 29728, Cur Loss: 1.33173192, Cur Avg Loss: 1.41306456, Log Avg loss: 1.65558462, Global Avg Loss: 1.76558717, Time: 0.0402 Steps: 40180, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001868, Sample Num: 29888, Cur Loss: 2.23412561, Cur Avg Loss: 1.41353161, Log Avg loss: 1.50030858, Global Avg Loss: 1.76552116, Time: 0.0402 Steps: 40190, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001878, Sample Num: 30048, Cur Loss: 0.78235114, Cur Avg Loss: 1.41198940, Log Avg loss: 1.12390621, Global Avg Loss: 1.76536156, Time: 0.0402 Steps: 40200, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001888, Sample Num: 30208, Cur Loss: 2.37521052, Cur Avg Loss: 1.41200720, Log Avg loss: 1.41534874, Global Avg Loss: 1.76527451, Time: 0.0403 Steps: 40210, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001898, Sample Num: 30368, Cur Loss: 0.28728166, Cur Avg Loss: 1.41047252, Log Avg loss: 1.12072580, Global Avg Loss: 1.76511425, Time: 0.0403 Steps: 40220, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001908, Sample Num: 30528, Cur Loss: 1.24728131, Cur Avg Loss: 1.41127116, Log Avg loss: 1.56285332, Global Avg Loss: 1.76506398, Time: 0.0402 Steps: 40230, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001918, Sample Num: 30688, Cur Loss: 1.26471603, Cur Avg Loss: 1.40985118, Log Avg loss: 1.13891850, Global Avg Loss: 1.76490838, Time: 0.0402 Steps: 40240, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001928, Sample Num: 30848, Cur Loss: 1.34671187, Cur Avg Loss: 1.41029640, Log Avg loss: 1.49568872, Global Avg Loss: 1.76484149, Time: 0.0402 Steps: 40250, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001938, Sample Num: 31008, Cur Loss: 2.18090701, Cur Avg Loss: 1.41057693, Log Avg loss: 1.46466291, Global Avg Loss: 1.76476693, Time: 0.0453 Steps: 40260, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001948, Sample Num: 31168, Cur Loss: 1.00250840, Cur Avg Loss: 1.41012159, Log Avg loss: 1.32187831, Global Avg Loss: 1.76465695, Time: 0.0452 Steps: 40270, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001958, Sample Num: 31328, Cur Loss: 2.35220528, Cur Avg Loss: 1.40954359, Log Avg loss: 1.29694784, Global Avg Loss: 1.76454083, Time: 0.0405 Steps: 40280, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001968, Sample Num: 31488, Cur Loss: 0.88907170, Cur Avg Loss: 1.40850319, Log Avg loss: 1.20479318, Global Avg Loss: 1.76440190, Time: 0.0673 Steps: 40290, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001978, Sample Num: 31648, Cur Loss: 1.66334176, Cur Avg Loss: 1.40996387, Log Avg loss: 1.69742600, Global Avg Loss: 1.76438529, Time: 0.0531 Steps: 40300, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001988, Sample Num: 31808, Cur Loss: 0.38912413, Cur Avg Loss: 1.40870852, Log Avg loss: 1.16040069, Global Avg Loss: 1.76423545, Time: 0.0635 Steps: 40310, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001998, Sample Num: 31968, Cur Loss: 1.08543825, Cur Avg Loss: 1.40813045, Log Avg loss: 1.29320974, Global Avg Loss: 1.76411863, Time: 0.0480 Steps: 40320, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002008, Sample Num: 32128, Cur Loss: 1.18917799, Cur Avg Loss: 1.40784180, Log Avg loss: 1.35016862, Global Avg Loss: 1.76401599, Time: 0.0437 Steps: 40330, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002018, Sample Num: 32288, Cur Loss: 1.04656911, Cur Avg Loss: 1.40827017, Log Avg loss: 1.49428754, Global Avg Loss: 1.76394912, Time: 0.0402 Steps: 40340, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002028, Sample Num: 32448, Cur Loss: 0.69520497, Cur Avg Loss: 1.40866371, Log Avg loss: 1.48807969, Global Avg Loss: 1.76388075, Time: 0.0403 Steps: 40350, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002038, Sample Num: 32608, Cur Loss: 1.02210617, Cur Avg Loss: 1.40723246, Log Avg loss: 1.11697462, Global Avg Loss: 1.76372047, Time: 0.0402 Steps: 40360, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002048, Sample Num: 32768, Cur Loss: 0.91955316, Cur Avg Loss: 1.40721649, Log Avg loss: 1.40396360, Global Avg Loss: 1.76363136, Time: 0.0403 Steps: 40370, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002058, Sample Num: 32928, Cur Loss: 3.53979254, Cur Avg Loss: 1.40716003, Log Avg loss: 1.39559577, Global Avg Loss: 1.76354021, Time: 0.0403 Steps: 40380, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002068, Sample Num: 33088, Cur Loss: 1.78911150, Cur Avg Loss: 1.40716294, Log Avg loss: 1.40776142, Global Avg Loss: 1.76345213, Time: 0.0402 Steps: 40390, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002078, Sample Num: 33248, Cur Loss: 0.78631091, Cur Avg Loss: 1.40655466, Log Avg loss: 1.28076366, Global Avg Loss: 1.76333265, Time: 0.0403 Steps: 40400, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002088, Sample Num: 33408, Cur Loss: 0.96770537, Cur Avg Loss: 1.40662551, Log Avg loss: 1.42134838, Global Avg Loss: 1.76324802, Time: 0.0402 Steps: 40410, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002098, Sample Num: 33568, Cur Loss: 1.32088387, Cur Avg Loss: 1.40664265, Log Avg loss: 1.41021997, Global Avg Loss: 1.76316068, Time: 0.0402 Steps: 40420, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002108, Sample Num: 33728, Cur Loss: 1.27233243, Cur Avg Loss: 1.40618783, Log Avg loss: 1.31076806, Global Avg Loss: 1.76304879, Time: 0.0402 Steps: 40430, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002118, Sample Num: 33888, Cur Loss: 2.35399103, Cur Avg Loss: 1.40672865, Log Avg loss: 1.52073238, Global Avg Loss: 1.76298887, Time: 0.0403 Steps: 40440, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002128, Sample Num: 34048, Cur Loss: 1.34026217, Cur Avg Loss: 1.40654133, Log Avg loss: 1.36686775, Global Avg Loss: 1.76289094, Time: 0.0402 Steps: 40450, Updated lr: 0.000063 ***** Running evaluation checkpoint-40451 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-40451 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.176504, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.319092, "eval_total_loss": 927.321361, "eval_mae": 0.982577, "eval_mse": 1.31904, "eval_r2": 0.161531, "eval_sp_statistic": 0.357441, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.447311, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.250481, "test_total_loss": 627.741317, "test_mae": 0.875388, "test_mse": 1.250754, "test_r2": 0.192752, "test_sp_statistic": 0.370343, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.454385, "test_ps_pvalue": 0.0, "lr": 6.258795637743007e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7628738692934258, "train_cur_epoch_loss": 2994.192421525717, "train_cur_epoch_avg_loss": 1.4063844159350478, "train_cur_epoch_time": 95.17650365829468, "train_cur_epoch_avg_time": 0.0447047926999975, "epoch": 19, "step": 40451} ################################################## Training, Epoch: 0020, Batch: 000009, Sample Num: 144, Cur Loss: 1.12668014, Cur Avg Loss: 1.26036908, Log Avg loss: 1.24157887, Global Avg Loss: 1.76276209, Time: 0.0403 Steps: 40460, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000019, Sample Num: 304, Cur Loss: 0.84433585, Cur Avg Loss: 1.50040109, Log Avg loss: 1.71642990, Global Avg Loss: 1.76275064, Time: 0.0402 Steps: 40470, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000029, Sample Num: 464, Cur Loss: 1.18906939, Cur Avg Loss: 1.47324566, Log Avg loss: 1.42165036, Global Avg Loss: 1.76266638, Time: 0.0402 Steps: 40480, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000039, Sample Num: 624, Cur Loss: 0.41480690, Cur Avg Loss: 1.44915708, Log Avg loss: 1.37930019, Global Avg Loss: 1.76257170, Time: 0.0403 Steps: 40490, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000049, Sample Num: 784, Cur Loss: 2.07237244, Cur Avg Loss: 1.42967841, Log Avg loss: 1.35371159, Global Avg Loss: 1.76247074, Time: 0.0402 Steps: 40500, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000059, Sample Num: 944, Cur Loss: 0.44557959, Cur Avg Loss: 1.44210617, Log Avg loss: 1.50300218, Global Avg Loss: 1.76240669, Time: 0.0402 Steps: 40510, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000069, Sample Num: 1104, Cur Loss: 1.37070572, Cur Avg Loss: 1.37635633, Log Avg loss: 0.98843228, Global Avg Loss: 1.76221568, Time: 0.0402 Steps: 40520, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000079, Sample Num: 1264, Cur Loss: 1.32089102, Cur Avg Loss: 1.33653511, Log Avg loss: 1.06176873, Global Avg Loss: 1.76204286, Time: 0.0403 Steps: 40530, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000089, Sample Num: 1424, Cur Loss: 1.51962376, Cur Avg Loss: 1.33462333, Log Avg loss: 1.31952027, Global Avg Loss: 1.76193370, Time: 0.0403 Steps: 40540, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000099, Sample Num: 1584, Cur Loss: 0.75274229, Cur Avg Loss: 1.33466280, Log Avg loss: 1.33501406, Global Avg Loss: 1.76182842, Time: 0.0403 Steps: 40550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000109, Sample Num: 1744, Cur Loss: 1.52838385, Cur Avg Loss: 1.35743783, Log Avg loss: 1.58291067, Global Avg Loss: 1.76178431, Time: 0.0402 Steps: 40560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000119, Sample Num: 1904, Cur Loss: 1.05539358, Cur Avg Loss: 1.34868691, Log Avg loss: 1.25330188, Global Avg Loss: 1.76165898, Time: 0.0402 Steps: 40570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000129, Sample Num: 2064, Cur Loss: 0.45936054, Cur Avg Loss: 1.34744997, Log Avg loss: 1.33273035, Global Avg Loss: 1.76155328, Time: 0.0402 Steps: 40580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000139, Sample Num: 2224, Cur Loss: 1.14781559, Cur Avg Loss: 1.33391168, Log Avg loss: 1.15926779, Global Avg Loss: 1.76140489, Time: 0.0402 Steps: 40590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000149, Sample Num: 2384, Cur Loss: 1.69752657, Cur Avg Loss: 1.34379633, Log Avg loss: 1.48119285, Global Avg Loss: 1.76133588, Time: 0.0402 Steps: 40600, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000159, Sample Num: 2544, Cur Loss: 2.07969332, Cur Avg Loss: 1.34316966, Log Avg loss: 1.33383236, Global Avg Loss: 1.76123060, Time: 0.0403 Steps: 40610, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000169, Sample Num: 2704, Cur Loss: 0.76300025, Cur Avg Loss: 1.34318042, Log Avg loss: 1.34335146, Global Avg Loss: 1.76112773, Time: 0.0402 Steps: 40620, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000179, Sample Num: 2864, Cur Loss: 1.43920159, Cur Avg Loss: 1.33538368, Log Avg loss: 1.20361874, Global Avg Loss: 1.76099051, Time: 0.0402 Steps: 40630, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000189, Sample Num: 3024, Cur Loss: 0.59908491, Cur Avg Loss: 1.35617705, Log Avg loss: 1.72837838, Global Avg Loss: 1.76098249, Time: 0.0402 Steps: 40640, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000199, Sample Num: 3184, Cur Loss: 0.95620036, Cur Avg Loss: 1.39137515, Log Avg loss: 2.05661929, Global Avg Loss: 1.76105522, Time: 0.0402 Steps: 40650, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000209, Sample Num: 3344, Cur Loss: 1.03832757, Cur Avg Loss: 1.37162929, Log Avg loss: 0.97868675, Global Avg Loss: 1.76086280, Time: 0.0402 Steps: 40660, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000219, Sample Num: 3504, Cur Loss: 1.59250164, Cur Avg Loss: 1.37727299, Log Avg loss: 1.49522622, Global Avg Loss: 1.76079748, Time: 0.0713 Steps: 40670, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000229, Sample Num: 3664, Cur Loss: 1.47336364, Cur Avg Loss: 1.36684902, Log Avg loss: 1.13856412, Global Avg Loss: 1.76064453, Time: 0.0635 Steps: 40680, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000239, Sample Num: 3824, Cur Loss: 1.82003522, Cur Avg Loss: 1.37164494, Log Avg loss: 1.48147159, Global Avg Loss: 1.76057592, Time: 0.0755 Steps: 40690, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000249, Sample Num: 3984, Cur Loss: 1.12432075, Cur Avg Loss: 1.37728557, Log Avg loss: 1.51209664, Global Avg Loss: 1.76051486, Time: 0.0772 Steps: 40700, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000259, Sample Num: 4144, Cur Loss: 1.99586225, Cur Avg Loss: 1.37884058, Log Avg loss: 1.41756029, Global Avg Loss: 1.76043062, Time: 0.0611 Steps: 40710, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000269, Sample Num: 4304, Cur Loss: 1.34063625, Cur Avg Loss: 1.37642119, Log Avg loss: 1.31375880, Global Avg Loss: 1.76032093, Time: 0.0675 Steps: 40720, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000279, Sample Num: 4464, Cur Loss: 1.44293690, Cur Avg Loss: 1.37061121, Log Avg loss: 1.21432295, Global Avg Loss: 1.76018687, Time: 0.0403 Steps: 40730, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000289, Sample Num: 4624, Cur Loss: 1.27820528, Cur Avg Loss: 1.36948809, Log Avg loss: 1.33815306, Global Avg Loss: 1.76008328, Time: 0.0514 Steps: 40740, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000299, Sample Num: 4784, Cur Loss: 2.02734971, Cur Avg Loss: 1.37141111, Log Avg loss: 1.42698621, Global Avg Loss: 1.76000154, Time: 0.0728 Steps: 40750, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000309, Sample Num: 4944, Cur Loss: 0.52757657, Cur Avg Loss: 1.37072774, Log Avg loss: 1.35029495, Global Avg Loss: 1.75990102, Time: 0.0402 Steps: 40760, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000319, Sample Num: 5104, Cur Loss: 0.34421766, Cur Avg Loss: 1.36655910, Log Avg loss: 1.23774812, Global Avg Loss: 1.75977295, Time: 0.0403 Steps: 40770, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000329, Sample Num: 5264, Cur Loss: 1.87870646, Cur Avg Loss: 1.37286637, Log Avg loss: 1.57406857, Global Avg Loss: 1.75972741, Time: 0.0403 Steps: 40780, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000339, Sample Num: 5424, Cur Loss: 1.41313362, Cur Avg Loss: 1.38028045, Log Avg loss: 1.62420347, Global Avg Loss: 1.75969419, Time: 0.0403 Steps: 40790, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000349, Sample Num: 5584, Cur Loss: 1.83730555, Cur Avg Loss: 1.37791045, Log Avg loss: 1.29756732, Global Avg Loss: 1.75958092, Time: 0.0402 Steps: 40800, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000359, Sample Num: 5744, Cur Loss: 0.50174618, Cur Avg Loss: 1.37421460, Log Avg loss: 1.24522951, Global Avg Loss: 1.75945489, Time: 0.0403 Steps: 40810, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000369, Sample Num: 5904, Cur Loss: 1.58248031, Cur Avg Loss: 1.36282339, Log Avg loss: 0.95387911, Global Avg Loss: 1.75925754, Time: 0.0403 Steps: 40820, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000379, Sample Num: 6064, Cur Loss: 0.79932606, Cur Avg Loss: 1.37347378, Log Avg loss: 1.76647317, Global Avg Loss: 1.75925931, Time: 0.0402 Steps: 40830, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000389, Sample Num: 6224, Cur Loss: 2.92432547, Cur Avg Loss: 1.37146712, Log Avg loss: 1.29541455, Global Avg Loss: 1.75914573, Time: 0.0402 Steps: 40840, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000399, Sample Num: 6384, Cur Loss: 1.06609237, Cur Avg Loss: 1.37372056, Log Avg loss: 1.46137928, Global Avg Loss: 1.75907284, Time: 0.0402 Steps: 40850, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000409, Sample Num: 6544, Cur Loss: 0.94471121, Cur Avg Loss: 1.38197152, Log Avg loss: 1.71118512, Global Avg Loss: 1.75906112, Time: 0.0402 Steps: 40860, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000419, Sample Num: 6704, Cur Loss: 1.14012575, Cur Avg Loss: 1.38225205, Log Avg loss: 1.39372557, Global Avg Loss: 1.75897173, Time: 0.0402 Steps: 40870, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000429, Sample Num: 6864, Cur Loss: 0.84876031, Cur Avg Loss: 1.37658488, Log Avg loss: 1.13913052, Global Avg Loss: 1.75882010, Time: 0.0403 Steps: 40880, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000439, Sample Num: 7024, Cur Loss: 1.38862681, Cur Avg Loss: 1.37828654, Log Avg loss: 1.45128760, Global Avg Loss: 1.75874489, Time: 0.0402 Steps: 40890, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000449, Sample Num: 7184, Cur Loss: 0.49683651, Cur Avg Loss: 1.36970025, Log Avg loss: 0.99276227, Global Avg Loss: 1.75855761, Time: 0.0403 Steps: 40900, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000459, Sample Num: 7344, Cur Loss: 1.70517135, Cur Avg Loss: 1.36606039, Log Avg loss: 1.20263056, Global Avg Loss: 1.75842172, Time: 0.0402 Steps: 40910, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000469, Sample Num: 7504, Cur Loss: 0.75129485, Cur Avg Loss: 1.36912299, Log Avg loss: 1.50969657, Global Avg Loss: 1.75836094, Time: 0.0403 Steps: 40920, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000479, Sample Num: 7664, Cur Loss: 2.69237304, Cur Avg Loss: 1.37750709, Log Avg loss: 1.77072145, Global Avg Loss: 1.75836396, Time: 0.0402 Steps: 40930, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000489, Sample Num: 7824, Cur Loss: 2.72417665, Cur Avg Loss: 1.37924885, Log Avg loss: 1.46267877, Global Avg Loss: 1.75829173, Time: 0.0402 Steps: 40940, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000499, Sample Num: 7984, Cur Loss: 1.38963628, Cur Avg Loss: 1.38495187, Log Avg loss: 1.66382995, Global Avg Loss: 1.75826867, Time: 0.0402 Steps: 40950, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000509, Sample Num: 8144, Cur Loss: 0.81987011, Cur Avg Loss: 1.38691653, Log Avg loss: 1.48495298, Global Avg Loss: 1.75820194, Time: 0.0403 Steps: 40960, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000519, Sample Num: 8304, Cur Loss: 1.52487504, Cur Avg Loss: 1.38971209, Log Avg loss: 1.53200580, Global Avg Loss: 1.75814673, Time: 0.0403 Steps: 40970, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000529, Sample Num: 8464, Cur Loss: 1.53002405, Cur Avg Loss: 1.38749014, Log Avg loss: 1.27217124, Global Avg Loss: 1.75802814, Time: 0.0402 Steps: 40980, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000539, Sample Num: 8624, Cur Loss: 0.65978688, Cur Avg Loss: 1.38595505, Log Avg loss: 1.30474857, Global Avg Loss: 1.75791756, Time: 0.0402 Steps: 40990, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000549, Sample Num: 8784, Cur Loss: 1.68645740, Cur Avg Loss: 1.38183836, Log Avg loss: 1.15994856, Global Avg Loss: 1.75777171, Time: 0.0402 Steps: 41000, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000559, Sample Num: 8944, Cur Loss: 1.66750455, Cur Avg Loss: 1.37956324, Log Avg loss: 1.25465959, Global Avg Loss: 1.75764903, Time: 0.0402 Steps: 41010, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000569, Sample Num: 9104, Cur Loss: 0.59318769, Cur Avg Loss: 1.38045494, Log Avg loss: 1.43030102, Global Avg Loss: 1.75756923, Time: 0.0565 Steps: 41020, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000579, Sample Num: 9264, Cur Loss: 1.59364843, Cur Avg Loss: 1.37654909, Log Avg loss: 1.15430600, Global Avg Loss: 1.75742220, Time: 0.0659 Steps: 41030, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000589, Sample Num: 9424, Cur Loss: 2.07547712, Cur Avg Loss: 1.37141921, Log Avg loss: 1.07439938, Global Avg Loss: 1.75725577, Time: 0.0405 Steps: 41040, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000599, Sample Num: 9584, Cur Loss: 2.04828644, Cur Avg Loss: 1.37049320, Log Avg loss: 1.31595076, Global Avg Loss: 1.75714827, Time: 0.0856 Steps: 41050, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000609, Sample Num: 9744, Cur Loss: 1.91407514, Cur Avg Loss: 1.36696609, Log Avg loss: 1.15569241, Global Avg Loss: 1.75700178, Time: 0.0403 Steps: 41060, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000619, Sample Num: 9904, Cur Loss: 1.61088717, Cur Avg Loss: 1.37052185, Log Avg loss: 1.58706785, Global Avg Loss: 1.75696041, Time: 0.0549 Steps: 41070, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000629, Sample Num: 10064, Cur Loss: 0.94448304, Cur Avg Loss: 1.36803174, Log Avg loss: 1.21389403, Global Avg Loss: 1.75682821, Time: 0.0456 Steps: 41080, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000639, Sample Num: 10224, Cur Loss: 0.87670225, Cur Avg Loss: 1.36405500, Log Avg loss: 1.11391805, Global Avg Loss: 1.75667175, Time: 0.0404 Steps: 41090, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000649, Sample Num: 10384, Cur Loss: 0.56900424, Cur Avg Loss: 1.35905568, Log Avg loss: 1.03959891, Global Avg Loss: 1.75649728, Time: 0.0500 Steps: 41100, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000659, Sample Num: 10544, Cur Loss: 1.53341746, Cur Avg Loss: 1.35897999, Log Avg loss: 1.35406789, Global Avg Loss: 1.75639938, Time: 0.0959 Steps: 41110, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000669, Sample Num: 10704, Cur Loss: 1.36395872, Cur Avg Loss: 1.35620620, Log Avg loss: 1.17341347, Global Avg Loss: 1.75625761, Time: 0.0403 Steps: 41120, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000679, Sample Num: 10864, Cur Loss: 3.35711408, Cur Avg Loss: 1.35743886, Log Avg loss: 1.43990385, Global Avg Loss: 1.75618069, Time: 0.0403 Steps: 41130, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000689, Sample Num: 11024, Cur Loss: 1.72504520, Cur Avg Loss: 1.35275319, Log Avg loss: 1.03459555, Global Avg Loss: 1.75600529, Time: 0.0403 Steps: 41140, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000699, Sample Num: 11184, Cur Loss: 1.05813336, Cur Avg Loss: 1.35337050, Log Avg loss: 1.39590324, Global Avg Loss: 1.75591779, Time: 0.0403 Steps: 41150, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000709, Sample Num: 11344, Cur Loss: 0.42459673, Cur Avg Loss: 1.35035801, Log Avg loss: 1.13978539, Global Avg Loss: 1.75576809, Time: 0.0403 Steps: 41160, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000719, Sample Num: 11504, Cur Loss: 0.75340188, Cur Avg Loss: 1.35252623, Log Avg loss: 1.50625259, Global Avg Loss: 1.75570749, Time: 0.0403 Steps: 41170, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000729, Sample Num: 11664, Cur Loss: 1.39371562, Cur Avg Loss: 1.35252584, Log Avg loss: 1.35249804, Global Avg Loss: 1.75560957, Time: 0.0402 Steps: 41180, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000739, Sample Num: 11824, Cur Loss: 0.33843184, Cur Avg Loss: 1.34868325, Log Avg loss: 1.06855864, Global Avg Loss: 1.75544277, Time: 0.0403 Steps: 41190, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000749, Sample Num: 11984, Cur Loss: 1.60292768, Cur Avg Loss: 1.35093929, Log Avg loss: 1.51766027, Global Avg Loss: 1.75538506, Time: 0.0403 Steps: 41200, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000759, Sample Num: 12144, Cur Loss: 0.64031816, Cur Avg Loss: 1.35062133, Log Avg loss: 1.32680618, Global Avg Loss: 1.75528106, Time: 0.0402 Steps: 41210, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000769, Sample Num: 12304, Cur Loss: 0.89195162, Cur Avg Loss: 1.34887134, Log Avg loss: 1.21604736, Global Avg Loss: 1.75515024, Time: 0.0403 Steps: 41220, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000779, Sample Num: 12464, Cur Loss: 1.08820319, Cur Avg Loss: 1.34653127, Log Avg loss: 1.16657955, Global Avg Loss: 1.75500749, Time: 0.0402 Steps: 41230, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000789, Sample Num: 12624, Cur Loss: 1.00164759, Cur Avg Loss: 1.34648341, Log Avg loss: 1.34275503, Global Avg Loss: 1.75490752, Time: 0.0403 Steps: 41240, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000799, Sample Num: 12784, Cur Loss: 1.93737614, Cur Avg Loss: 1.34839766, Log Avg loss: 1.49943261, Global Avg Loss: 1.75484559, Time: 0.0402 Steps: 41250, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000809, Sample Num: 12944, Cur Loss: 1.17492521, Cur Avg Loss: 1.35051966, Log Avg loss: 1.52006726, Global Avg Loss: 1.75478869, Time: 0.0403 Steps: 41260, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000819, Sample Num: 13104, Cur Loss: 1.82300007, Cur Avg Loss: 1.35273672, Log Avg loss: 1.53209693, Global Avg Loss: 1.75473473, Time: 0.0402 Steps: 41270, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000829, Sample Num: 13264, Cur Loss: 1.30657125, Cur Avg Loss: 1.35500086, Log Avg loss: 1.54043381, Global Avg Loss: 1.75468281, Time: 0.0402 Steps: 41280, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000839, Sample Num: 13424, Cur Loss: 1.71376240, Cur Avg Loss: 1.35669608, Log Avg loss: 1.49722995, Global Avg Loss: 1.75462046, Time: 0.0402 Steps: 41290, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000849, Sample Num: 13584, Cur Loss: 1.02410984, Cur Avg Loss: 1.35692830, Log Avg loss: 1.37641118, Global Avg Loss: 1.75452889, Time: 0.0402 Steps: 41300, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000859, Sample Num: 13744, Cur Loss: 1.69932973, Cur Avg Loss: 1.35694322, Log Avg loss: 1.35820982, Global Avg Loss: 1.75443295, Time: 0.0402 Steps: 41310, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000869, Sample Num: 13904, Cur Loss: 1.24206007, Cur Avg Loss: 1.35533017, Log Avg loss: 1.21676919, Global Avg Loss: 1.75430283, Time: 0.0402 Steps: 41320, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000879, Sample Num: 14064, Cur Loss: 1.72643805, Cur Avg Loss: 1.35349659, Log Avg loss: 1.19415919, Global Avg Loss: 1.75416730, Time: 0.0402 Steps: 41330, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000889, Sample Num: 14224, Cur Loss: 1.18017471, Cur Avg Loss: 1.35380423, Log Avg loss: 1.38084552, Global Avg Loss: 1.75407699, Time: 0.0402 Steps: 41340, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000899, Sample Num: 14384, Cur Loss: 1.04298115, Cur Avg Loss: 1.35275396, Log Avg loss: 1.25938479, Global Avg Loss: 1.75395736, Time: 0.0402 Steps: 41350, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000909, Sample Num: 14544, Cur Loss: 1.13289249, Cur Avg Loss: 1.35200060, Log Avg loss: 1.28427330, Global Avg Loss: 1.75384380, Time: 0.0402 Steps: 41360, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000919, Sample Num: 14704, Cur Loss: 0.94091284, Cur Avg Loss: 1.35257613, Log Avg loss: 1.40489246, Global Avg Loss: 1.75375945, Time: 0.0402 Steps: 41370, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000929, Sample Num: 14864, Cur Loss: 1.05936813, Cur Avg Loss: 1.35248297, Log Avg loss: 1.34392156, Global Avg Loss: 1.75366041, Time: 0.0564 Steps: 41380, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000939, Sample Num: 15024, Cur Loss: 1.09838068, Cur Avg Loss: 1.35082154, Log Avg loss: 1.19647451, Global Avg Loss: 1.75352579, Time: 0.0541 Steps: 41390, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000949, Sample Num: 15184, Cur Loss: 2.12446856, Cur Avg Loss: 1.35149210, Log Avg loss: 1.41445751, Global Avg Loss: 1.75344389, Time: 0.0405 Steps: 41400, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000959, Sample Num: 15344, Cur Loss: 0.88167119, Cur Avg Loss: 1.35135743, Log Avg loss: 1.33857766, Global Avg Loss: 1.75334370, Time: 0.0432 Steps: 41410, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000969, Sample Num: 15504, Cur Loss: 0.76359725, Cur Avg Loss: 1.35433015, Log Avg loss: 1.63941379, Global Avg Loss: 1.75331620, Time: 0.0657 Steps: 41420, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000979, Sample Num: 15664, Cur Loss: 0.65384030, Cur Avg Loss: 1.35564485, Log Avg loss: 1.48303894, Global Avg Loss: 1.75325096, Time: 0.0404 Steps: 41430, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000989, Sample Num: 15824, Cur Loss: 1.54295337, Cur Avg Loss: 1.35603746, Log Avg loss: 1.39447446, Global Avg Loss: 1.75316438, Time: 0.0405 Steps: 41440, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000999, Sample Num: 15984, Cur Loss: 1.66708970, Cur Avg Loss: 1.35794632, Log Avg loss: 1.54673212, Global Avg Loss: 1.75311458, Time: 0.0404 Steps: 41450, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001009, Sample Num: 16144, Cur Loss: 1.76313424, Cur Avg Loss: 1.36275585, Log Avg loss: 1.84322816, Global Avg Loss: 1.75313631, Time: 0.0466 Steps: 41460, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001019, Sample Num: 16304, Cur Loss: 1.99068594, Cur Avg Loss: 1.36384229, Log Avg loss: 1.47346394, Global Avg Loss: 1.75306887, Time: 0.0402 Steps: 41470, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001029, Sample Num: 16464, Cur Loss: 0.79418081, Cur Avg Loss: 1.36400391, Log Avg loss: 1.38047258, Global Avg Loss: 1.75297905, Time: 0.0403 Steps: 41480, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001039, Sample Num: 16624, Cur Loss: 1.86308575, Cur Avg Loss: 1.36408003, Log Avg loss: 1.37191303, Global Avg Loss: 1.75288720, Time: 0.0402 Steps: 41490, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001049, Sample Num: 16784, Cur Loss: 1.23299599, Cur Avg Loss: 1.35976788, Log Avg loss: 0.91173589, Global Avg Loss: 1.75268452, Time: 0.0402 Steps: 41500, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001059, Sample Num: 16944, Cur Loss: 0.97744477, Cur Avg Loss: 1.35937439, Log Avg loss: 1.31809660, Global Avg Loss: 1.75257982, Time: 0.0403 Steps: 41510, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001069, Sample Num: 17104, Cur Loss: 1.49367583, Cur Avg Loss: 1.36169627, Log Avg loss: 1.60758405, Global Avg Loss: 1.75254490, Time: 0.0403 Steps: 41520, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001079, Sample Num: 17264, Cur Loss: 0.79998636, Cur Avg Loss: 1.36098939, Log Avg loss: 1.28542326, Global Avg Loss: 1.75243242, Time: 0.0403 Steps: 41530, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001089, Sample Num: 17424, Cur Loss: 0.53239560, Cur Avg Loss: 1.36247740, Log Avg loss: 1.52303444, Global Avg Loss: 1.75237720, Time: 0.0403 Steps: 41540, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001099, Sample Num: 17584, Cur Loss: 1.25365722, Cur Avg Loss: 1.36154755, Log Avg loss: 1.26028645, Global Avg Loss: 1.75225876, Time: 0.0402 Steps: 41550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001109, Sample Num: 17744, Cur Loss: 2.15755415, Cur Avg Loss: 1.36184487, Log Avg loss: 1.39451985, Global Avg Loss: 1.75217269, Time: 0.0403 Steps: 41560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001119, Sample Num: 17904, Cur Loss: 0.52871889, Cur Avg Loss: 1.36093471, Log Avg loss: 1.25999884, Global Avg Loss: 1.75205429, Time: 0.0403 Steps: 41570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001129, Sample Num: 18064, Cur Loss: 1.98863328, Cur Avg Loss: 1.36087089, Log Avg loss: 1.35372859, Global Avg Loss: 1.75195849, Time: 0.0403 Steps: 41580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001139, Sample Num: 18224, Cur Loss: 0.96601892, Cur Avg Loss: 1.36221222, Log Avg loss: 1.51364859, Global Avg Loss: 1.75190119, Time: 0.0402 Steps: 41590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001149, Sample Num: 18384, Cur Loss: 1.03989804, Cur Avg Loss: 1.36376270, Log Avg loss: 1.54036279, Global Avg Loss: 1.75185034, Time: 0.0403 Steps: 41600, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001159, Sample Num: 18544, Cur Loss: 1.00867128, Cur Avg Loss: 1.36215040, Log Avg loss: 1.17689700, Global Avg Loss: 1.75171217, Time: 0.0403 Steps: 41610, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001169, Sample Num: 18704, Cur Loss: 2.15022802, Cur Avg Loss: 1.36297459, Log Avg loss: 1.45849784, Global Avg Loss: 1.75164172, Time: 0.0402 Steps: 41620, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001179, Sample Num: 18864, Cur Loss: 1.02123976, Cur Avg Loss: 1.36342493, Log Avg loss: 1.41606996, Global Avg Loss: 1.75156111, Time: 0.0402 Steps: 41630, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001189, Sample Num: 19024, Cur Loss: 2.50426531, Cur Avg Loss: 1.36597407, Log Avg loss: 1.66651787, Global Avg Loss: 1.75154068, Time: 0.0402 Steps: 41640, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001199, Sample Num: 19184, Cur Loss: 2.45435786, Cur Avg Loss: 1.36796730, Log Avg loss: 1.60496281, Global Avg Loss: 1.75150549, Time: 0.0403 Steps: 41650, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001209, Sample Num: 19344, Cur Loss: 0.93456531, Cur Avg Loss: 1.36810133, Log Avg loss: 1.38417065, Global Avg Loss: 1.75141732, Time: 0.0403 Steps: 41660, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001219, Sample Num: 19504, Cur Loss: 2.43006015, Cur Avg Loss: 1.36982120, Log Avg loss: 1.57775381, Global Avg Loss: 1.75137564, Time: 0.0402 Steps: 41670, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001229, Sample Num: 19664, Cur Loss: 0.31770751, Cur Avg Loss: 1.36811686, Log Avg loss: 1.16035724, Global Avg Loss: 1.75123384, Time: 0.0403 Steps: 41680, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001239, Sample Num: 19824, Cur Loss: 1.00024772, Cur Avg Loss: 1.36894396, Log Avg loss: 1.47059516, Global Avg Loss: 1.75116653, Time: 0.0402 Steps: 41690, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001249, Sample Num: 19984, Cur Loss: 1.21858883, Cur Avg Loss: 1.36861008, Log Avg loss: 1.32724189, Global Avg Loss: 1.75106487, Time: 0.0402 Steps: 41700, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001259, Sample Num: 20144, Cur Loss: 1.59395242, Cur Avg Loss: 1.36706722, Log Avg loss: 1.17436390, Global Avg Loss: 1.75092660, Time: 0.0402 Steps: 41710, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001269, Sample Num: 20304, Cur Loss: 1.34161413, Cur Avg Loss: 1.36800101, Log Avg loss: 1.48556522, Global Avg Loss: 1.75086300, Time: 0.0403 Steps: 41720, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001279, Sample Num: 20464, Cur Loss: 1.03903341, Cur Avg Loss: 1.36949663, Log Avg loss: 1.55929176, Global Avg Loss: 1.75081709, Time: 0.0405 Steps: 41730, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001289, Sample Num: 20624, Cur Loss: 0.31662032, Cur Avg Loss: 1.36723599, Log Avg loss: 1.07809994, Global Avg Loss: 1.75065592, Time: 0.0614 Steps: 41740, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001299, Sample Num: 20784, Cur Loss: 0.91394734, Cur Avg Loss: 1.36657655, Log Avg loss: 1.28157443, Global Avg Loss: 1.75054356, Time: 0.0794 Steps: 41750, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001309, Sample Num: 20944, Cur Loss: 1.98922968, Cur Avg Loss: 1.36733159, Log Avg loss: 1.46541085, Global Avg Loss: 1.75047529, Time: 0.0404 Steps: 41760, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001319, Sample Num: 21104, Cur Loss: 1.51356268, Cur Avg Loss: 1.36756770, Log Avg loss: 1.39847471, Global Avg Loss: 1.75039101, Time: 0.0484 Steps: 41770, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001329, Sample Num: 21264, Cur Loss: 2.51479530, Cur Avg Loss: 1.36758823, Log Avg loss: 1.37029686, Global Avg Loss: 1.75030004, Time: 0.0840 Steps: 41780, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001339, Sample Num: 21424, Cur Loss: 1.39982617, Cur Avg Loss: 1.36847262, Log Avg loss: 1.48600723, Global Avg Loss: 1.75023680, Time: 0.0607 Steps: 41790, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001349, Sample Num: 21584, Cur Loss: 2.46537662, Cur Avg Loss: 1.37149043, Log Avg loss: 1.77557579, Global Avg Loss: 1.75024286, Time: 0.0443 Steps: 41800, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001359, Sample Num: 21744, Cur Loss: 1.50909400, Cur Avg Loss: 1.37448824, Log Avg loss: 1.77889217, Global Avg Loss: 1.75024971, Time: 0.0405 Steps: 41810, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001369, Sample Num: 21904, Cur Loss: 2.56479669, Cur Avg Loss: 1.37510473, Log Avg loss: 1.45888543, Global Avg Loss: 1.75018004, Time: 0.0402 Steps: 41820, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001379, Sample Num: 22064, Cur Loss: 1.94248366, Cur Avg Loss: 1.37597048, Log Avg loss: 1.49449190, Global Avg Loss: 1.75011891, Time: 0.0403 Steps: 41830, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001389, Sample Num: 22224, Cur Loss: 1.11625171, Cur Avg Loss: 1.37634824, Log Avg loss: 1.42844223, Global Avg Loss: 1.75004203, Time: 0.0403 Steps: 41840, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001399, Sample Num: 22384, Cur Loss: 1.12236190, Cur Avg Loss: 1.37601473, Log Avg loss: 1.32968935, Global Avg Loss: 1.74994159, Time: 0.0402 Steps: 41850, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001409, Sample Num: 22544, Cur Loss: 2.49887156, Cur Avg Loss: 1.37775523, Log Avg loss: 1.62125171, Global Avg Loss: 1.74991085, Time: 0.0402 Steps: 41860, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001419, Sample Num: 22704, Cur Loss: 1.26166761, Cur Avg Loss: 1.37715411, Log Avg loss: 1.29245548, Global Avg Loss: 1.74980159, Time: 0.0403 Steps: 41870, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001429, Sample Num: 22864, Cur Loss: 1.70376873, Cur Avg Loss: 1.37872089, Log Avg loss: 1.60104823, Global Avg Loss: 1.74976607, Time: 0.0403 Steps: 41880, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001439, Sample Num: 23024, Cur Loss: 1.37056470, Cur Avg Loss: 1.37953618, Log Avg loss: 1.49604125, Global Avg Loss: 1.74970550, Time: 0.0403 Steps: 41890, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001449, Sample Num: 23184, Cur Loss: 1.24166262, Cur Avg Loss: 1.38088501, Log Avg loss: 1.57498044, Global Avg Loss: 1.74966380, Time: 0.0402 Steps: 41900, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001459, Sample Num: 23344, Cur Loss: 1.32426071, Cur Avg Loss: 1.38145956, Log Avg loss: 1.46471279, Global Avg Loss: 1.74959581, Time: 0.0403 Steps: 41910, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001469, Sample Num: 23504, Cur Loss: 0.84759247, Cur Avg Loss: 1.38054339, Log Avg loss: 1.24687324, Global Avg Loss: 1.74947589, Time: 0.0402 Steps: 41920, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001479, Sample Num: 23664, Cur Loss: 1.93668485, Cur Avg Loss: 1.38117850, Log Avg loss: 1.47447734, Global Avg Loss: 1.74941030, Time: 0.0402 Steps: 41930, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001489, Sample Num: 23824, Cur Loss: 1.41850305, Cur Avg Loss: 1.38075811, Log Avg loss: 1.31858125, Global Avg Loss: 1.74930758, Time: 0.0402 Steps: 41940, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001499, Sample Num: 23984, Cur Loss: 1.02205062, Cur Avg Loss: 1.37853592, Log Avg loss: 1.04765219, Global Avg Loss: 1.74914032, Time: 0.0403 Steps: 41950, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001509, Sample Num: 24144, Cur Loss: 1.33317828, Cur Avg Loss: 1.38075350, Log Avg loss: 1.71316863, Global Avg Loss: 1.74913174, Time: 0.0402 Steps: 41960, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001519, Sample Num: 24304, Cur Loss: 1.25250578, Cur Avg Loss: 1.37979187, Log Avg loss: 1.23468245, Global Avg Loss: 1.74900917, Time: 0.0403 Steps: 41970, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001529, Sample Num: 24464, Cur Loss: 0.81905723, Cur Avg Loss: 1.37746132, Log Avg loss: 1.02345084, Global Avg Loss: 1.74883633, Time: 0.0402 Steps: 41980, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001539, Sample Num: 24624, Cur Loss: 0.95295942, Cur Avg Loss: 1.37815396, Log Avg loss: 1.48405891, Global Avg Loss: 1.74877328, Time: 0.0402 Steps: 41990, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001549, Sample Num: 24784, Cur Loss: 0.79921961, Cur Avg Loss: 1.37722086, Log Avg loss: 1.23361635, Global Avg Loss: 1.74865062, Time: 0.0402 Steps: 42000, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001559, Sample Num: 24944, Cur Loss: 0.52523035, Cur Avg Loss: 1.37638158, Log Avg loss: 1.24637650, Global Avg Loss: 1.74853106, Time: 0.0402 Steps: 42010, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001569, Sample Num: 25104, Cur Loss: 0.92110282, Cur Avg Loss: 1.37413510, Log Avg loss: 1.02390996, Global Avg Loss: 1.74835861, Time: 0.0402 Steps: 42020, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001579, Sample Num: 25264, Cur Loss: 1.54451418, Cur Avg Loss: 1.37426199, Log Avg loss: 1.39417077, Global Avg Loss: 1.74827434, Time: 0.0402 Steps: 42030, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001589, Sample Num: 25424, Cur Loss: 0.92876852, Cur Avg Loss: 1.37417161, Log Avg loss: 1.35990089, Global Avg Loss: 1.74818196, Time: 0.0403 Steps: 42040, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001599, Sample Num: 25584, Cur Loss: 1.55827296, Cur Avg Loss: 1.37233838, Log Avg loss: 1.08103665, Global Avg Loss: 1.74802330, Time: 0.0402 Steps: 42050, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001609, Sample Num: 25744, Cur Loss: 1.03058577, Cur Avg Loss: 1.37007563, Log Avg loss: 1.00826296, Global Avg Loss: 1.74784742, Time: 0.0402 Steps: 42060, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001619, Sample Num: 25904, Cur Loss: 1.41290176, Cur Avg Loss: 1.37066624, Log Avg loss: 1.46569513, Global Avg Loss: 1.74778035, Time: 0.0402 Steps: 42070, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001629, Sample Num: 26064, Cur Loss: 1.47593331, Cur Avg Loss: 1.37072121, Log Avg loss: 1.37962092, Global Avg Loss: 1.74769286, Time: 0.0458 Steps: 42080, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001639, Sample Num: 26224, Cur Loss: 2.98413086, Cur Avg Loss: 1.37086066, Log Avg loss: 1.39357634, Global Avg Loss: 1.74760873, Time: 0.0403 Steps: 42090, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001649, Sample Num: 26384, Cur Loss: 1.66936398, Cur Avg Loss: 1.37074229, Log Avg loss: 1.35134201, Global Avg Loss: 1.74751461, Time: 0.0875 Steps: 42100, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001659, Sample Num: 26544, Cur Loss: 0.55572498, Cur Avg Loss: 1.36874892, Log Avg loss: 1.04004290, Global Avg Loss: 1.74734660, Time: 0.0648 Steps: 42110, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001669, Sample Num: 26704, Cur Loss: 0.86586356, Cur Avg Loss: 1.36951308, Log Avg loss: 1.49628611, Global Avg Loss: 1.74728699, Time: 0.0407 Steps: 42120, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001679, Sample Num: 26864, Cur Loss: 0.98112303, Cur Avg Loss: 1.36901101, Log Avg loss: 1.28521668, Global Avg Loss: 1.74717732, Time: 0.0406 Steps: 42130, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001689, Sample Num: 27024, Cur Loss: 1.22957098, Cur Avg Loss: 1.36732234, Log Avg loss: 1.08379311, Global Avg Loss: 1.74701989, Time: 0.0559 Steps: 42140, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001699, Sample Num: 27184, Cur Loss: 1.24149394, Cur Avg Loss: 1.36825363, Log Avg loss: 1.52554986, Global Avg Loss: 1.74696735, Time: 0.0917 Steps: 42150, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001709, Sample Num: 27344, Cur Loss: 1.94557714, Cur Avg Loss: 1.36980467, Log Avg loss: 1.63332523, Global Avg Loss: 1.74694040, Time: 0.0405 Steps: 42160, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001719, Sample Num: 27504, Cur Loss: 1.35972285, Cur Avg Loss: 1.36890482, Log Avg loss: 1.21512105, Global Avg Loss: 1.74681428, Time: 0.0624 Steps: 42170, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001729, Sample Num: 27664, Cur Loss: 0.98906362, Cur Avg Loss: 1.36954008, Log Avg loss: 1.47874067, Global Avg Loss: 1.74675073, Time: 0.0403 Steps: 42180, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001739, Sample Num: 27824, Cur Loss: 0.63185209, Cur Avg Loss: 1.36932830, Log Avg loss: 1.33271186, Global Avg Loss: 1.74665259, Time: 0.0402 Steps: 42190, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001749, Sample Num: 27984, Cur Loss: 2.43026423, Cur Avg Loss: 1.37058978, Log Avg loss: 1.58996198, Global Avg Loss: 1.74661546, Time: 0.0402 Steps: 42200, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001759, Sample Num: 28144, Cur Loss: 1.29703271, Cur Avg Loss: 1.36963684, Log Avg loss: 1.20296643, Global Avg Loss: 1.74648666, Time: 0.0403 Steps: 42210, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001769, Sample Num: 28304, Cur Loss: 0.25505641, Cur Avg Loss: 1.36843454, Log Avg loss: 1.15695057, Global Avg Loss: 1.74634703, Time: 0.0402 Steps: 42220, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001779, Sample Num: 28464, Cur Loss: 0.88410068, Cur Avg Loss: 1.36805280, Log Avg loss: 1.30052245, Global Avg Loss: 1.74624146, Time: 0.0402 Steps: 42230, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001789, Sample Num: 28624, Cur Loss: 1.17853200, Cur Avg Loss: 1.36900797, Log Avg loss: 1.53893256, Global Avg Loss: 1.74619238, Time: 0.0402 Steps: 42240, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001799, Sample Num: 28784, Cur Loss: 1.22799098, Cur Avg Loss: 1.37146040, Log Avg loss: 1.81020126, Global Avg Loss: 1.74620753, Time: 0.0402 Steps: 42250, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001809, Sample Num: 28944, Cur Loss: 0.88502973, Cur Avg Loss: 1.37227975, Log Avg loss: 1.51968000, Global Avg Loss: 1.74615393, Time: 0.0403 Steps: 42260, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001819, Sample Num: 29104, Cur Loss: 0.68030196, Cur Avg Loss: 1.37113752, Log Avg loss: 1.16450810, Global Avg Loss: 1.74601632, Time: 0.0403 Steps: 42270, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001829, Sample Num: 29264, Cur Loss: 1.09472966, Cur Avg Loss: 1.37070447, Log Avg loss: 1.29193334, Global Avg Loss: 1.74590893, Time: 0.0402 Steps: 42280, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001839, Sample Num: 29424, Cur Loss: 0.57336485, Cur Avg Loss: 1.37174119, Log Avg loss: 1.56135715, Global Avg Loss: 1.74586529, Time: 0.0402 Steps: 42290, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001849, Sample Num: 29584, Cur Loss: 1.03421950, Cur Avg Loss: 1.37153270, Log Avg loss: 1.33319089, Global Avg Loss: 1.74576773, Time: 0.0402 Steps: 42300, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001859, Sample Num: 29744, Cur Loss: 3.84524345, Cur Avg Loss: 1.37286479, Log Avg loss: 1.61916798, Global Avg Loss: 1.74573780, Time: 0.0403 Steps: 42310, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001869, Sample Num: 29904, Cur Loss: 1.01694965, Cur Avg Loss: 1.37464997, Log Avg loss: 1.70651609, Global Avg Loss: 1.74572854, Time: 0.0403 Steps: 42320, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001879, Sample Num: 30064, Cur Loss: 1.71306300, Cur Avg Loss: 1.37560370, Log Avg loss: 1.55385486, Global Avg Loss: 1.74568321, Time: 0.0403 Steps: 42330, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001889, Sample Num: 30224, Cur Loss: 0.51090962, Cur Avg Loss: 1.37480746, Log Avg loss: 1.22519477, Global Avg Loss: 1.74556028, Time: 0.0402 Steps: 42340, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001899, Sample Num: 30384, Cur Loss: 1.66796184, Cur Avg Loss: 1.37521948, Log Avg loss: 1.45305023, Global Avg Loss: 1.74549121, Time: 0.0402 Steps: 42350, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001909, Sample Num: 30544, Cur Loss: 1.97703087, Cur Avg Loss: 1.37575679, Log Avg loss: 1.47779125, Global Avg Loss: 1.74542801, Time: 0.0403 Steps: 42360, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001919, Sample Num: 30704, Cur Loss: 1.95986819, Cur Avg Loss: 1.37436034, Log Avg loss: 1.10777882, Global Avg Loss: 1.74527752, Time: 0.0403 Steps: 42370, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001929, Sample Num: 30864, Cur Loss: 0.54153895, Cur Avg Loss: 1.37415176, Log Avg loss: 1.33412504, Global Avg Loss: 1.74518050, Time: 0.0403 Steps: 42380, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001939, Sample Num: 31024, Cur Loss: 1.96373820, Cur Avg Loss: 1.37365282, Log Avg loss: 1.27740752, Global Avg Loss: 1.74507015, Time: 0.0403 Steps: 42390, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001949, Sample Num: 31184, Cur Loss: 1.19209838, Cur Avg Loss: 1.37368776, Log Avg loss: 1.38046224, Global Avg Loss: 1.74498416, Time: 0.0403 Steps: 42400, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001959, Sample Num: 31344, Cur Loss: 1.09931827, Cur Avg Loss: 1.37484651, Log Avg loss: 1.60068734, Global Avg Loss: 1.74495013, Time: 0.0403 Steps: 42410, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001969, Sample Num: 31504, Cur Loss: 1.14521098, Cur Avg Loss: 1.37585628, Log Avg loss: 1.57366925, Global Avg Loss: 1.74490976, Time: 0.0402 Steps: 42420, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001979, Sample Num: 31664, Cur Loss: 0.63284433, Cur Avg Loss: 1.37469533, Log Avg loss: 1.14610518, Global Avg Loss: 1.74476863, Time: 0.0403 Steps: 42430, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001989, Sample Num: 31824, Cur Loss: 1.06794310, Cur Avg Loss: 1.37268353, Log Avg loss: 0.97454814, Global Avg Loss: 1.74458714, Time: 0.0698 Steps: 42440, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001999, Sample Num: 31984, Cur Loss: 1.36938763, Cur Avg Loss: 1.37281391, Log Avg loss: 1.39874569, Global Avg Loss: 1.74450567, Time: 0.0455 Steps: 42450, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002009, Sample Num: 32144, Cur Loss: 1.27825034, Cur Avg Loss: 1.37264606, Log Avg loss: 1.33909264, Global Avg Loss: 1.74441019, Time: 0.0403 Steps: 42460, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002019, Sample Num: 32304, Cur Loss: 0.90503144, Cur Avg Loss: 1.37351609, Log Avg loss: 1.54830655, Global Avg Loss: 1.74436402, Time: 0.0873 Steps: 42470, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002029, Sample Num: 32464, Cur Loss: 0.19865879, Cur Avg Loss: 1.37318202, Log Avg loss: 1.30573328, Global Avg Loss: 1.74426076, Time: 0.0405 Steps: 42480, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002039, Sample Num: 32624, Cur Loss: 1.55258524, Cur Avg Loss: 1.37363880, Log Avg loss: 1.46631861, Global Avg Loss: 1.74419535, Time: 0.0404 Steps: 42490, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002049, Sample Num: 32784, Cur Loss: 1.41324258, Cur Avg Loss: 1.37344281, Log Avg loss: 1.33347945, Global Avg Loss: 1.74409871, Time: 0.0560 Steps: 42500, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002059, Sample Num: 32944, Cur Loss: 1.98988986, Cur Avg Loss: 1.37354519, Log Avg loss: 1.39452408, Global Avg Loss: 1.74401648, Time: 0.0407 Steps: 42510, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002069, Sample Num: 33104, Cur Loss: 2.93807673, Cur Avg Loss: 1.37484940, Log Avg loss: 1.64338585, Global Avg Loss: 1.74399281, Time: 0.0848 Steps: 42520, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002079, Sample Num: 33264, Cur Loss: 1.12436557, Cur Avg Loss: 1.37484251, Log Avg loss: 1.37341740, Global Avg Loss: 1.74390568, Time: 0.0599 Steps: 42530, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002089, Sample Num: 33424, Cur Loss: 1.84686553, Cur Avg Loss: 1.37414346, Log Avg loss: 1.22881078, Global Avg Loss: 1.74378459, Time: 0.0403 Steps: 42540, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002099, Sample Num: 33584, Cur Loss: 1.75091314, Cur Avg Loss: 1.37452168, Log Avg loss: 1.45353153, Global Avg Loss: 1.74371638, Time: 0.0402 Steps: 42550, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002109, Sample Num: 33744, Cur Loss: 1.01850009, Cur Avg Loss: 1.37455125, Log Avg loss: 1.38075901, Global Avg Loss: 1.74363110, Time: 0.0402 Steps: 42560, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002119, Sample Num: 33904, Cur Loss: 3.24517441, Cur Avg Loss: 1.37394555, Log Avg loss: 1.24620346, Global Avg Loss: 1.74351425, Time: 0.0402 Steps: 42570, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002129, Sample Num: 34055, Cur Loss: 0.54236901, Cur Avg Loss: 1.37523890, Log Avg loss: 1.64929948, Global Avg Loss: 1.74349212, Time: 0.0198 Steps: 42580, Updated lr: 0.000061 ***** Running evaluation checkpoint-42580 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-42580 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.927957, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.329362, "eval_total_loss": 934.541452, "eval_mae": 0.93822, "eval_mse": 1.329453, "eval_r2": 0.154913, "eval_sp_statistic": 0.33364, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.419999, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.290179, "test_total_loss": 647.670038, "test_mae": 0.796872, "test_mse": 1.29053, "test_r2": 0.167081, "test_sp_statistic": 0.341681, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.446863, "test_ps_pvalue": 0.0, "lr": 6.056899004267425e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7434921209544236, "train_cur_epoch_loss": 2927.8836234509945, "train_cur_epoch_avg_loss": 1.375238902513384, "train_cur_epoch_time": 95.9279568195343, "train_cur_epoch_avg_time": 0.04505775332058915, "epoch": 20, "step": 42580} ################################################## Training, Epoch: 0021, Batch: 000010, Sample Num: 160, Cur Loss: 1.16519880, Cur Avg Loss: 1.42175985, Log Avg loss: 1.42175985, Global Avg Loss: 1.74341658, Time: 0.0403 Steps: 42590, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000020, Sample Num: 320, Cur Loss: 0.89399451, Cur Avg Loss: 1.46192422, Log Avg loss: 1.50208859, Global Avg Loss: 1.74335993, Time: 0.0403 Steps: 42600, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000030, Sample Num: 480, Cur Loss: 0.64782405, Cur Avg Loss: 1.37848345, Log Avg loss: 1.21160190, Global Avg Loss: 1.74323513, Time: 0.0404 Steps: 42610, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000040, Sample Num: 640, Cur Loss: 0.65845942, Cur Avg Loss: 1.39141926, Log Avg loss: 1.43022671, Global Avg Loss: 1.74316169, Time: 0.0404 Steps: 42620, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000050, Sample Num: 800, Cur Loss: 0.75729704, Cur Avg Loss: 1.35662920, Log Avg loss: 1.21746895, Global Avg Loss: 1.74303838, Time: 0.0404 Steps: 42630, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000060, Sample Num: 960, Cur Loss: 1.86797833, Cur Avg Loss: 1.32770681, Log Avg loss: 1.18309488, Global Avg Loss: 1.74290706, Time: 0.0403 Steps: 42640, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000070, Sample Num: 1120, Cur Loss: 0.55483830, Cur Avg Loss: 1.30226446, Log Avg loss: 1.14961035, Global Avg Loss: 1.74276795, Time: 0.0403 Steps: 42650, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000080, Sample Num: 1280, Cur Loss: 1.82553804, Cur Avg Loss: 1.33015563, Log Avg loss: 1.52539381, Global Avg Loss: 1.74271699, Time: 0.0404 Steps: 42660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000090, Sample Num: 1440, Cur Loss: 1.38545418, Cur Avg Loss: 1.35532761, Log Avg loss: 1.55670347, Global Avg Loss: 1.74267340, Time: 0.0402 Steps: 42670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000100, Sample Num: 1600, Cur Loss: 1.23628986, Cur Avg Loss: 1.37023989, Log Avg loss: 1.50445043, Global Avg Loss: 1.74261758, Time: 0.0402 Steps: 42680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000110, Sample Num: 1760, Cur Loss: 1.27228224, Cur Avg Loss: 1.35015622, Log Avg loss: 1.14931942, Global Avg Loss: 1.74247861, Time: 0.0403 Steps: 42690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000120, Sample Num: 1920, Cur Loss: 1.17562413, Cur Avg Loss: 1.33766684, Log Avg loss: 1.20028369, Global Avg Loss: 1.74235163, Time: 0.0403 Steps: 42700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000130, Sample Num: 2080, Cur Loss: 1.12671614, Cur Avg Loss: 1.32355564, Log Avg loss: 1.15422124, Global Avg Loss: 1.74221393, Time: 0.0403 Steps: 42710, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000140, Sample Num: 2240, Cur Loss: 1.76320791, Cur Avg Loss: 1.34485371, Log Avg loss: 1.62172859, Global Avg Loss: 1.74218572, Time: 0.0402 Steps: 42720, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000150, Sample Num: 2400, Cur Loss: 1.19660139, Cur Avg Loss: 1.35157420, Log Avg loss: 1.44566118, Global Avg Loss: 1.74211633, Time: 0.0402 Steps: 42730, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000160, Sample Num: 2560, Cur Loss: 1.59420478, Cur Avg Loss: 1.35066177, Log Avg loss: 1.33697527, Global Avg Loss: 1.74202153, Time: 0.0402 Steps: 42740, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000170, Sample Num: 2720, Cur Loss: 1.42126572, Cur Avg Loss: 1.36361466, Log Avg loss: 1.57086094, Global Avg Loss: 1.74198150, Time: 0.0402 Steps: 42750, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000180, Sample Num: 2880, Cur Loss: 1.95730722, Cur Avg Loss: 1.36771499, Log Avg loss: 1.43742057, Global Avg Loss: 1.74191027, Time: 0.0402 Steps: 42760, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000190, Sample Num: 3040, Cur Loss: 2.18457985, Cur Avg Loss: 1.36343583, Log Avg loss: 1.28641086, Global Avg Loss: 1.74180377, Time: 0.0402 Steps: 42770, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000200, Sample Num: 3200, Cur Loss: 1.89485407, Cur Avg Loss: 1.37595267, Log Avg loss: 1.61377262, Global Avg Loss: 1.74177384, Time: 0.0403 Steps: 42780, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000210, Sample Num: 3360, Cur Loss: 4.32739544, Cur Avg Loss: 1.39739278, Log Avg loss: 1.82619499, Global Avg Loss: 1.74179357, Time: 0.0402 Steps: 42790, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000220, Sample Num: 3520, Cur Loss: 0.61793000, Cur Avg Loss: 1.39546424, Log Avg loss: 1.35496507, Global Avg Loss: 1.74170319, Time: 0.0403 Steps: 42800, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000230, Sample Num: 3680, Cur Loss: 2.39420938, Cur Avg Loss: 1.41460361, Log Avg loss: 1.83566970, Global Avg Loss: 1.74172514, Time: 0.0402 Steps: 42810, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000240, Sample Num: 3840, Cur Loss: 1.81181252, Cur Avg Loss: 1.41803324, Log Avg loss: 1.49691470, Global Avg Loss: 1.74166797, Time: 0.0402 Steps: 42820, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000250, Sample Num: 4000, Cur Loss: 1.33201385, Cur Avg Loss: 1.40803110, Log Avg loss: 1.16797979, Global Avg Loss: 1.74153402, Time: 0.0402 Steps: 42830, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000260, Sample Num: 4160, Cur Loss: 1.17214584, Cur Avg Loss: 1.39837489, Log Avg loss: 1.15696965, Global Avg Loss: 1.74139757, Time: 0.1008 Steps: 42840, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000270, Sample Num: 4320, Cur Loss: 1.61484849, Cur Avg Loss: 1.38674387, Log Avg loss: 1.08433718, Global Avg Loss: 1.74124423, Time: 0.0535 Steps: 42850, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000280, Sample Num: 4480, Cur Loss: 1.23336494, Cur Avg Loss: 1.39035644, Log Avg loss: 1.48789591, Global Avg Loss: 1.74118512, Time: 0.0683 Steps: 42860, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000290, Sample Num: 4640, Cur Loss: 1.44190502, Cur Avg Loss: 1.37950198, Log Avg loss: 1.07557698, Global Avg Loss: 1.74102986, Time: 0.0580 Steps: 42870, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000300, Sample Num: 4800, Cur Loss: 1.34678364, Cur Avg Loss: 1.38891360, Log Avg loss: 1.66185070, Global Avg Loss: 1.74101139, Time: 0.0652 Steps: 42880, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000310, Sample Num: 4960, Cur Loss: 2.21863747, Cur Avg Loss: 1.38696408, Log Avg loss: 1.32847846, Global Avg Loss: 1.74091521, Time: 0.0404 Steps: 42890, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000320, Sample Num: 5120, Cur Loss: 1.26832950, Cur Avg Loss: 1.38016669, Log Avg loss: 1.16944776, Global Avg Loss: 1.74078200, Time: 0.0468 Steps: 42900, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000330, Sample Num: 5280, Cur Loss: 1.01774502, Cur Avg Loss: 1.38130784, Log Avg loss: 1.41782460, Global Avg Loss: 1.74070674, Time: 0.0741 Steps: 42910, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000340, Sample Num: 5440, Cur Loss: 0.99238378, Cur Avg Loss: 1.37961538, Log Avg loss: 1.32376423, Global Avg Loss: 1.74060959, Time: 0.0504 Steps: 42920, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000350, Sample Num: 5600, Cur Loss: 3.23110056, Cur Avg Loss: 1.38670219, Log Avg loss: 1.62765357, Global Avg Loss: 1.74058328, Time: 0.0403 Steps: 42930, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000360, Sample Num: 5760, Cur Loss: 2.84116840, Cur Avg Loss: 1.39557352, Log Avg loss: 1.70607012, Global Avg Loss: 1.74057524, Time: 0.0402 Steps: 42940, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000370, Sample Num: 5920, Cur Loss: 1.99357605, Cur Avg Loss: 1.39449816, Log Avg loss: 1.35578531, Global Avg Loss: 1.74048565, Time: 0.0403 Steps: 42950, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000380, Sample Num: 6080, Cur Loss: 1.70309615, Cur Avg Loss: 1.39290655, Log Avg loss: 1.33401695, Global Avg Loss: 1.74039104, Time: 0.0401 Steps: 42960, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000390, Sample Num: 6240, Cur Loss: 1.54996240, Cur Avg Loss: 1.39643099, Log Avg loss: 1.53035942, Global Avg Loss: 1.74034216, Time: 0.0403 Steps: 42970, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000400, Sample Num: 6400, Cur Loss: 0.88577956, Cur Avg Loss: 1.39773328, Log Avg loss: 1.44852281, Global Avg Loss: 1.74027426, Time: 0.0402 Steps: 42980, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000410, Sample Num: 6560, Cur Loss: 1.70635068, Cur Avg Loss: 1.39990731, Log Avg loss: 1.48686830, Global Avg Loss: 1.74021532, Time: 0.0402 Steps: 42990, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000420, Sample Num: 6720, Cur Loss: 1.01562274, Cur Avg Loss: 1.40309968, Log Avg loss: 1.53398696, Global Avg Loss: 1.74016736, Time: 0.0402 Steps: 43000, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000430, Sample Num: 6880, Cur Loss: 1.46553206, Cur Avg Loss: 1.40788055, Log Avg loss: 1.60867729, Global Avg Loss: 1.74013679, Time: 0.0402 Steps: 43010, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000440, Sample Num: 7040, Cur Loss: 2.07219791, Cur Avg Loss: 1.40221341, Log Avg loss: 1.15852604, Global Avg Loss: 1.74000159, Time: 0.0403 Steps: 43020, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000450, Sample Num: 7200, Cur Loss: 0.77597189, Cur Avg Loss: 1.40361606, Log Avg loss: 1.46533273, Global Avg Loss: 1.73993776, Time: 0.0402 Steps: 43030, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000460, Sample Num: 7360, Cur Loss: 2.36343026, Cur Avg Loss: 1.40517120, Log Avg loss: 1.47515286, Global Avg Loss: 1.73987624, Time: 0.0403 Steps: 43040, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000470, Sample Num: 7520, Cur Loss: 0.53285670, Cur Avg Loss: 1.40000000, Log Avg loss: 1.16212466, Global Avg Loss: 1.73974203, Time: 0.0402 Steps: 43050, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000480, Sample Num: 7680, Cur Loss: 0.59463036, Cur Avg Loss: 1.39446685, Log Avg loss: 1.13440854, Global Avg Loss: 1.73960145, Time: 0.0402 Steps: 43060, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000490, Sample Num: 7840, Cur Loss: 1.32113075, Cur Avg Loss: 1.39853928, Log Avg loss: 1.59401606, Global Avg Loss: 1.73956765, Time: 0.0402 Steps: 43070, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000500, Sample Num: 8000, Cur Loss: 1.06904829, Cur Avg Loss: 1.40656649, Log Avg loss: 1.79989988, Global Avg Loss: 1.73958166, Time: 0.0402 Steps: 43080, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000510, Sample Num: 8160, Cur Loss: 0.89348489, Cur Avg Loss: 1.40898357, Log Avg loss: 1.52983731, Global Avg Loss: 1.73953298, Time: 0.0403 Steps: 43090, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000520, Sample Num: 8320, Cur Loss: 0.29996276, Cur Avg Loss: 1.40403058, Log Avg loss: 1.15142804, Global Avg Loss: 1.73939653, Time: 0.0402 Steps: 43100, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000530, Sample Num: 8480, Cur Loss: 0.95484090, Cur Avg Loss: 1.40385619, Log Avg loss: 1.39478838, Global Avg Loss: 1.73931659, Time: 0.0403 Steps: 43110, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000540, Sample Num: 8640, Cur Loss: 1.19032097, Cur Avg Loss: 1.40320873, Log Avg loss: 1.36889297, Global Avg Loss: 1.73923069, Time: 0.0402 Steps: 43120, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000550, Sample Num: 8800, Cur Loss: 1.46621585, Cur Avg Loss: 1.39795873, Log Avg loss: 1.11445889, Global Avg Loss: 1.73908583, Time: 0.0402 Steps: 43130, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000560, Sample Num: 8960, Cur Loss: 0.76517910, Cur Avg Loss: 1.39440310, Log Avg loss: 1.19884339, Global Avg Loss: 1.73896060, Time: 0.0402 Steps: 43140, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000570, Sample Num: 9120, Cur Loss: 0.77737874, Cur Avg Loss: 1.38737524, Log Avg loss: 0.99381523, Global Avg Loss: 1.73878791, Time: 0.0402 Steps: 43150, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000580, Sample Num: 9280, Cur Loss: 0.22994815, Cur Avg Loss: 1.38257925, Log Avg loss: 1.10920752, Global Avg Loss: 1.73864204, Time: 0.0402 Steps: 43160, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000590, Sample Num: 9440, Cur Loss: 2.34191227, Cur Avg Loss: 1.38707913, Log Avg loss: 1.64807242, Global Avg Loss: 1.73862106, Time: 0.0402 Steps: 43170, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000600, Sample Num: 9600, Cur Loss: 0.75410354, Cur Avg Loss: 1.38842878, Log Avg loss: 1.46805786, Global Avg Loss: 1.73855840, Time: 0.0402 Steps: 43180, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000610, Sample Num: 9760, Cur Loss: 1.15861523, Cur Avg Loss: 1.38340223, Log Avg loss: 1.08180965, Global Avg Loss: 1.73840634, Time: 0.0505 Steps: 43190, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000620, Sample Num: 9920, Cur Loss: 1.19697547, Cur Avg Loss: 1.38105415, Log Avg loss: 1.23782086, Global Avg Loss: 1.73829046, Time: 0.0866 Steps: 43200, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000630, Sample Num: 10080, Cur Loss: 1.51725197, Cur Avg Loss: 1.37812023, Log Avg loss: 1.19621769, Global Avg Loss: 1.73816501, Time: 0.0455 Steps: 43210, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000640, Sample Num: 10240, Cur Loss: 1.09205425, Cur Avg Loss: 1.37951514, Log Avg loss: 1.46739421, Global Avg Loss: 1.73810236, Time: 0.0416 Steps: 43220, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000650, Sample Num: 10400, Cur Loss: 0.70828855, Cur Avg Loss: 1.37627779, Log Avg loss: 1.16908706, Global Avg Loss: 1.73797074, Time: 0.0406 Steps: 43230, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000660, Sample Num: 10560, Cur Loss: 1.10141563, Cur Avg Loss: 1.37635431, Log Avg loss: 1.38132858, Global Avg Loss: 1.73788826, Time: 0.0536 Steps: 43240, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000670, Sample Num: 10720, Cur Loss: 1.24227393, Cur Avg Loss: 1.37295172, Log Avg loss: 1.14838034, Global Avg Loss: 1.73775196, Time: 0.0486 Steps: 43250, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000680, Sample Num: 10880, Cur Loss: 1.39447188, Cur Avg Loss: 1.37028625, Log Avg loss: 1.19170011, Global Avg Loss: 1.73762573, Time: 0.0669 Steps: 43260, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000690, Sample Num: 11040, Cur Loss: 2.03322172, Cur Avg Loss: 1.37135967, Log Avg loss: 1.44435194, Global Avg Loss: 1.73755795, Time: 0.0512 Steps: 43270, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000700, Sample Num: 11200, Cur Loss: 0.88997376, Cur Avg Loss: 1.36718961, Log Avg loss: 1.07945593, Global Avg Loss: 1.73740590, Time: 0.0481 Steps: 43280, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000710, Sample Num: 11360, Cur Loss: 0.76666081, Cur Avg Loss: 1.36788106, Log Avg loss: 1.41628219, Global Avg Loss: 1.73733172, Time: 0.0403 Steps: 43290, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000720, Sample Num: 11520, Cur Loss: 0.45556706, Cur Avg Loss: 1.36713179, Log Avg loss: 1.31393394, Global Avg Loss: 1.73723394, Time: 0.0403 Steps: 43300, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000730, Sample Num: 11680, Cur Loss: 0.67079282, Cur Avg Loss: 1.36641511, Log Avg loss: 1.31481422, Global Avg Loss: 1.73713640, Time: 0.0402 Steps: 43310, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000740, Sample Num: 11840, Cur Loss: 1.58015430, Cur Avg Loss: 1.36737415, Log Avg loss: 1.43738393, Global Avg Loss: 1.73706721, Time: 0.0403 Steps: 43320, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000750, Sample Num: 12000, Cur Loss: 1.52904439, Cur Avg Loss: 1.36836907, Log Avg loss: 1.44199278, Global Avg Loss: 1.73699911, Time: 0.0403 Steps: 43330, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000760, Sample Num: 12160, Cur Loss: 1.34939742, Cur Avg Loss: 1.36637679, Log Avg loss: 1.21695583, Global Avg Loss: 1.73687912, Time: 0.0403 Steps: 43340, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000770, Sample Num: 12320, Cur Loss: 0.90269268, Cur Avg Loss: 1.36686782, Log Avg loss: 1.40418620, Global Avg Loss: 1.73680237, Time: 0.0402 Steps: 43350, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000780, Sample Num: 12480, Cur Loss: 1.69775426, Cur Avg Loss: 1.36859005, Log Avg loss: 1.50120186, Global Avg Loss: 1.73674803, Time: 0.0403 Steps: 43360, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000790, Sample Num: 12640, Cur Loss: 1.85628986, Cur Avg Loss: 1.37408062, Log Avg loss: 1.80234497, Global Avg Loss: 1.73676316, Time: 0.0402 Steps: 43370, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000800, Sample Num: 12800, Cur Loss: 1.80367041, Cur Avg Loss: 1.37710187, Log Avg loss: 1.61578043, Global Avg Loss: 1.73673527, Time: 0.0402 Steps: 43380, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000810, Sample Num: 12960, Cur Loss: 1.22032762, Cur Avg Loss: 1.37667727, Log Avg loss: 1.34270952, Global Avg Loss: 1.73664446, Time: 0.0402 Steps: 43390, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000820, Sample Num: 13120, Cur Loss: 1.10788822, Cur Avg Loss: 1.37915550, Log Avg loss: 1.57989216, Global Avg Loss: 1.73660834, Time: 0.0402 Steps: 43400, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000830, Sample Num: 13280, Cur Loss: 2.22933960, Cur Avg Loss: 1.37978472, Log Avg loss: 1.43138067, Global Avg Loss: 1.73653803, Time: 0.0402 Steps: 43410, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000840, Sample Num: 13440, Cur Loss: 1.22240734, Cur Avg Loss: 1.38173105, Log Avg loss: 1.54327687, Global Avg Loss: 1.73649352, Time: 0.0403 Steps: 43420, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000850, Sample Num: 13600, Cur Loss: 1.29771352, Cur Avg Loss: 1.38389542, Log Avg loss: 1.56570251, Global Avg Loss: 1.73645419, Time: 0.0403 Steps: 43430, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000860, Sample Num: 13760, Cur Loss: 0.63051689, Cur Avg Loss: 1.38456370, Log Avg loss: 1.44136747, Global Avg Loss: 1.73638626, Time: 0.0403 Steps: 43440, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000870, Sample Num: 13920, Cur Loss: 0.91750997, Cur Avg Loss: 1.38741945, Log Avg loss: 1.63301371, Global Avg Loss: 1.73636247, Time: 0.0402 Steps: 43450, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000880, Sample Num: 14080, Cur Loss: 0.78793550, Cur Avg Loss: 1.38620076, Log Avg loss: 1.28017483, Global Avg Loss: 1.73625751, Time: 0.0402 Steps: 43460, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000890, Sample Num: 14240, Cur Loss: 1.69242930, Cur Avg Loss: 1.38649457, Log Avg loss: 1.41234987, Global Avg Loss: 1.73618299, Time: 0.0402 Steps: 43470, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000900, Sample Num: 14400, Cur Loss: 1.62401354, Cur Avg Loss: 1.38543040, Log Avg loss: 1.29071882, Global Avg Loss: 1.73608054, Time: 0.0402 Steps: 43480, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000910, Sample Num: 14560, Cur Loss: 0.75806105, Cur Avg Loss: 1.38368769, Log Avg loss: 1.22684442, Global Avg Loss: 1.73596345, Time: 0.0403 Steps: 43490, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000920, Sample Num: 14720, Cur Loss: 1.77840149, Cur Avg Loss: 1.38744375, Log Avg loss: 1.72924517, Global Avg Loss: 1.73596190, Time: 0.0403 Steps: 43500, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000930, Sample Num: 14880, Cur Loss: 0.85924840, Cur Avg Loss: 1.38522224, Log Avg loss: 1.18084301, Global Avg Loss: 1.73583432, Time: 0.0402 Steps: 43510, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000940, Sample Num: 15040, Cur Loss: 0.70529574, Cur Avg Loss: 1.38619935, Log Avg loss: 1.47707106, Global Avg Loss: 1.73577486, Time: 0.0402 Steps: 43520, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000950, Sample Num: 15200, Cur Loss: 1.97367775, Cur Avg Loss: 1.39065844, Log Avg loss: 1.80981215, Global Avg Loss: 1.73579187, Time: 0.0402 Steps: 43530, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000960, Sample Num: 15360, Cur Loss: 0.88832712, Cur Avg Loss: 1.38827580, Log Avg loss: 1.16192535, Global Avg Loss: 1.73566007, Time: 0.0402 Steps: 43540, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000970, Sample Num: 15520, Cur Loss: 3.09008503, Cur Avg Loss: 1.38822225, Log Avg loss: 1.38308094, Global Avg Loss: 1.73557911, Time: 0.0685 Steps: 43550, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000980, Sample Num: 15680, Cur Loss: 1.46174431, Cur Avg Loss: 1.39121497, Log Avg loss: 1.68150924, Global Avg Loss: 1.73556669, Time: 0.0959 Steps: 43560, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000990, Sample Num: 15840, Cur Loss: 0.47092909, Cur Avg Loss: 1.38778182, Log Avg loss: 1.05133347, Global Avg Loss: 1.73540965, Time: 0.0534 Steps: 43570, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001000, Sample Num: 16000, Cur Loss: 0.38413143, Cur Avg Loss: 1.38637102, Log Avg loss: 1.24670182, Global Avg Loss: 1.73529751, Time: 0.0645 Steps: 43580, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001010, Sample Num: 16160, Cur Loss: 0.69479477, Cur Avg Loss: 1.38694966, Log Avg loss: 1.44481317, Global Avg Loss: 1.73523087, Time: 0.0747 Steps: 43590, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001020, Sample Num: 16320, Cur Loss: 2.05094028, Cur Avg Loss: 1.38722258, Log Avg loss: 1.41478816, Global Avg Loss: 1.73515737, Time: 0.0403 Steps: 43600, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001030, Sample Num: 16480, Cur Loss: 0.94111347, Cur Avg Loss: 1.38574363, Log Avg loss: 1.23488992, Global Avg Loss: 1.73504266, Time: 0.0673 Steps: 43610, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001040, Sample Num: 16640, Cur Loss: 1.33087754, Cur Avg Loss: 1.38744322, Log Avg loss: 1.56250132, Global Avg Loss: 1.73500311, Time: 0.0468 Steps: 43620, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001050, Sample Num: 16800, Cur Loss: 1.22058809, Cur Avg Loss: 1.38297885, Log Avg loss: 0.91868470, Global Avg Loss: 1.73481601, Time: 0.0405 Steps: 43630, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001060, Sample Num: 16960, Cur Loss: 1.22781789, Cur Avg Loss: 1.38247191, Log Avg loss: 1.32924304, Global Avg Loss: 1.73472307, Time: 0.0802 Steps: 43640, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001070, Sample Num: 17120, Cur Loss: 1.55182576, Cur Avg Loss: 1.38216551, Log Avg loss: 1.34968672, Global Avg Loss: 1.73463486, Time: 0.0402 Steps: 43650, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001080, Sample Num: 17280, Cur Loss: 0.82826823, Cur Avg Loss: 1.37997625, Log Avg loss: 1.14572530, Global Avg Loss: 1.73449997, Time: 0.0402 Steps: 43660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001090, Sample Num: 17440, Cur Loss: 2.18316865, Cur Avg Loss: 1.38384425, Log Avg loss: 1.80158852, Global Avg Loss: 1.73451534, Time: 0.0402 Steps: 43670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001100, Sample Num: 17600, Cur Loss: 1.03643036, Cur Avg Loss: 1.38506395, Log Avg loss: 1.51801109, Global Avg Loss: 1.73446577, Time: 0.0403 Steps: 43680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001110, Sample Num: 17760, Cur Loss: 1.19937110, Cur Avg Loss: 1.38486787, Log Avg loss: 1.36329923, Global Avg Loss: 1.73438082, Time: 0.0403 Steps: 43690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001120, Sample Num: 17920, Cur Loss: 0.49033618, Cur Avg Loss: 1.38570948, Log Avg loss: 1.47912841, Global Avg Loss: 1.73432241, Time: 0.0402 Steps: 43700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001130, Sample Num: 18080, Cur Loss: 0.72905105, Cur Avg Loss: 1.38328924, Log Avg loss: 1.11222258, Global Avg Loss: 1.73418008, Time: 0.0403 Steps: 43710, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001140, Sample Num: 18240, Cur Loss: 1.72952318, Cur Avg Loss: 1.38539977, Log Avg loss: 1.62388890, Global Avg Loss: 1.73415485, Time: 0.0404 Steps: 43720, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001150, Sample Num: 18400, Cur Loss: 1.13049078, Cur Avg Loss: 1.38430550, Log Avg loss: 1.25955876, Global Avg Loss: 1.73404633, Time: 0.0404 Steps: 43730, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001160, Sample Num: 18560, Cur Loss: 1.69677758, Cur Avg Loss: 1.38360863, Log Avg loss: 1.30346885, Global Avg Loss: 1.73394789, Time: 0.0402 Steps: 43740, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001170, Sample Num: 18720, Cur Loss: 1.69302201, Cur Avg Loss: 1.38422480, Log Avg loss: 1.45570022, Global Avg Loss: 1.73388429, Time: 0.0403 Steps: 43750, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001180, Sample Num: 18880, Cur Loss: 1.43039644, Cur Avg Loss: 1.38672190, Log Avg loss: 1.67888260, Global Avg Loss: 1.73387172, Time: 0.0403 Steps: 43760, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001190, Sample Num: 19040, Cur Loss: 0.70499778, Cur Avg Loss: 1.38779946, Log Avg loss: 1.51495159, Global Avg Loss: 1.73382170, Time: 0.0403 Steps: 43770, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001200, Sample Num: 19200, Cur Loss: 1.27707613, Cur Avg Loss: 1.38863402, Log Avg loss: 1.48794714, Global Avg Loss: 1.73376554, Time: 0.0403 Steps: 43780, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001210, Sample Num: 19360, Cur Loss: 1.62608528, Cur Avg Loss: 1.38688763, Log Avg loss: 1.17732046, Global Avg Loss: 1.73363847, Time: 0.0406 Steps: 43790, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001220, Sample Num: 19520, Cur Loss: 0.94561183, Cur Avg Loss: 1.38819202, Log Avg loss: 1.54602334, Global Avg Loss: 1.73359563, Time: 0.0402 Steps: 43800, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001230, Sample Num: 19680, Cur Loss: 0.92695272, Cur Avg Loss: 1.39118246, Log Avg loss: 1.75601588, Global Avg Loss: 1.73360075, Time: 0.0402 Steps: 43810, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001240, Sample Num: 19840, Cur Loss: 1.36664009, Cur Avg Loss: 1.39267925, Log Avg loss: 1.57678539, Global Avg Loss: 1.73356497, Time: 0.0403 Steps: 43820, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001250, Sample Num: 20000, Cur Loss: 1.33168042, Cur Avg Loss: 1.39169106, Log Avg loss: 1.26915508, Global Avg Loss: 1.73345901, Time: 0.0402 Steps: 43830, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001260, Sample Num: 20160, Cur Loss: 2.30617714, Cur Avg Loss: 1.39032370, Log Avg loss: 1.21940395, Global Avg Loss: 1.73334175, Time: 0.0402 Steps: 43840, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001270, Sample Num: 20320, Cur Loss: 1.91324806, Cur Avg Loss: 1.39021779, Log Avg loss: 1.37687328, Global Avg Loss: 1.73326046, Time: 0.0402 Steps: 43850, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001280, Sample Num: 20480, Cur Loss: 1.79475045, Cur Avg Loss: 1.39115392, Log Avg loss: 1.51004180, Global Avg Loss: 1.73320957, Time: 0.0403 Steps: 43860, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001290, Sample Num: 20640, Cur Loss: 1.41282773, Cur Avg Loss: 1.38833881, Log Avg loss: 1.02800518, Global Avg Loss: 1.73304882, Time: 0.0402 Steps: 43870, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001300, Sample Num: 20800, Cur Loss: 0.64119583, Cur Avg Loss: 1.38778648, Log Avg loss: 1.31653528, Global Avg Loss: 1.73295390, Time: 0.0402 Steps: 43880, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001310, Sample Num: 20960, Cur Loss: 1.50544941, Cur Avg Loss: 1.38784426, Log Avg loss: 1.39535539, Global Avg Loss: 1.73287698, Time: 0.0401 Steps: 43890, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001320, Sample Num: 21120, Cur Loss: 1.76145709, Cur Avg Loss: 1.38842715, Log Avg loss: 1.46478642, Global Avg Loss: 1.73281591, Time: 0.0999 Steps: 43900, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001330, Sample Num: 21280, Cur Loss: 1.82024920, Cur Avg Loss: 1.38893562, Log Avg loss: 1.45605283, Global Avg Loss: 1.73275288, Time: 0.0525 Steps: 43910, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001340, Sample Num: 21440, Cur Loss: 0.81184769, Cur Avg Loss: 1.38740385, Log Avg loss: 1.18367872, Global Avg Loss: 1.73262786, Time: 0.0646 Steps: 43920, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001350, Sample Num: 21600, Cur Loss: 1.60175276, Cur Avg Loss: 1.38975289, Log Avg loss: 1.70452513, Global Avg Loss: 1.73262146, Time: 0.0933 Steps: 43930, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001360, Sample Num: 21760, Cur Loss: 2.43535924, Cur Avg Loss: 1.39163805, Log Avg loss: 1.64613375, Global Avg Loss: 1.73260178, Time: 0.0656 Steps: 43940, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001370, Sample Num: 21920, Cur Loss: 0.71422648, Cur Avg Loss: 1.39186574, Log Avg loss: 1.42283179, Global Avg Loss: 1.73253130, Time: 0.0549 Steps: 43950, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001380, Sample Num: 22080, Cur Loss: 0.90245903, Cur Avg Loss: 1.39305745, Log Avg loss: 1.55632174, Global Avg Loss: 1.73249121, Time: 0.0411 Steps: 43960, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001390, Sample Num: 22240, Cur Loss: 1.13914502, Cur Avg Loss: 1.39348668, Log Avg loss: 1.45271996, Global Avg Loss: 1.73242759, Time: 0.0717 Steps: 43970, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001400, Sample Num: 22400, Cur Loss: 0.81133866, Cur Avg Loss: 1.39357797, Log Avg loss: 1.40626843, Global Avg Loss: 1.73235343, Time: 0.0403 Steps: 43980, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001410, Sample Num: 22560, Cur Loss: 2.01381135, Cur Avg Loss: 1.39368563, Log Avg loss: 1.40875791, Global Avg Loss: 1.73227986, Time: 0.0402 Steps: 43990, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001420, Sample Num: 22720, Cur Loss: 2.89577961, Cur Avg Loss: 1.39462368, Log Avg loss: 1.52688842, Global Avg Loss: 1.73223318, Time: 0.0402 Steps: 44000, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001430, Sample Num: 22880, Cur Loss: 1.18575537, Cur Avg Loss: 1.39381839, Log Avg loss: 1.27946653, Global Avg Loss: 1.73213031, Time: 0.0402 Steps: 44010, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001440, Sample Num: 23040, Cur Loss: 1.43359256, Cur Avg Loss: 1.39394430, Log Avg loss: 1.41195056, Global Avg Loss: 1.73205757, Time: 0.0404 Steps: 44020, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001450, Sample Num: 23200, Cur Loss: 0.95081872, Cur Avg Loss: 1.39514276, Log Avg loss: 1.56772066, Global Avg Loss: 1.73202025, Time: 0.0402 Steps: 44030, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001460, Sample Num: 23360, Cur Loss: 1.27804041, Cur Avg Loss: 1.39800583, Log Avg loss: 1.81315135, Global Avg Loss: 1.73203867, Time: 0.0403 Steps: 44040, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001470, Sample Num: 23520, Cur Loss: 0.78978264, Cur Avg Loss: 1.39665934, Log Avg loss: 1.20007111, Global Avg Loss: 1.73191791, Time: 0.0403 Steps: 44050, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001480, Sample Num: 23680, Cur Loss: 1.80722857, Cur Avg Loss: 1.39763837, Log Avg loss: 1.54155624, Global Avg Loss: 1.73187470, Time: 0.0402 Steps: 44060, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001490, Sample Num: 23840, Cur Loss: 1.74548030, Cur Avg Loss: 1.39859596, Log Avg loss: 1.54031857, Global Avg Loss: 1.73183123, Time: 0.0403 Steps: 44070, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001500, Sample Num: 24000, Cur Loss: 0.69429147, Cur Avg Loss: 1.39680928, Log Avg loss: 1.13059428, Global Avg Loss: 1.73169484, Time: 0.0402 Steps: 44080, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001510, Sample Num: 24160, Cur Loss: 1.79603755, Cur Avg Loss: 1.39497731, Log Avg loss: 1.12018250, Global Avg Loss: 1.73155614, Time: 0.0403 Steps: 44090, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001520, Sample Num: 24320, Cur Loss: 1.83519578, Cur Avg Loss: 1.39454538, Log Avg loss: 1.32932363, Global Avg Loss: 1.73146493, Time: 0.0404 Steps: 44100, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001530, Sample Num: 24480, Cur Loss: 0.51879656, Cur Avg Loss: 1.39356625, Log Avg loss: 1.24473758, Global Avg Loss: 1.73135459, Time: 0.0402 Steps: 44110, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001540, Sample Num: 24640, Cur Loss: 1.30333018, Cur Avg Loss: 1.39454547, Log Avg loss: 1.54436683, Global Avg Loss: 1.73131221, Time: 0.0403 Steps: 44120, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001550, Sample Num: 24800, Cur Loss: 1.04969299, Cur Avg Loss: 1.39259736, Log Avg loss: 1.09258887, Global Avg Loss: 1.73116747, Time: 0.0402 Steps: 44130, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001560, Sample Num: 24960, Cur Loss: 1.72028542, Cur Avg Loss: 1.39213895, Log Avg loss: 1.32108515, Global Avg Loss: 1.73107456, Time: 0.0402 Steps: 44140, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001570, Sample Num: 25120, Cur Loss: 1.02552903, Cur Avg Loss: 1.39103685, Log Avg loss: 1.21910819, Global Avg Loss: 1.73095860, Time: 0.0402 Steps: 44150, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001580, Sample Num: 25280, Cur Loss: 1.23582435, Cur Avg Loss: 1.39281937, Log Avg loss: 1.67267617, Global Avg Loss: 1.73094541, Time: 0.0403 Steps: 44160, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001590, Sample Num: 25440, Cur Loss: 1.63844812, Cur Avg Loss: 1.39231450, Log Avg loss: 1.31254419, Global Avg Loss: 1.73085068, Time: 0.0403 Steps: 44170, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001600, Sample Num: 25600, Cur Loss: 2.43199182, Cur Avg Loss: 1.39288560, Log Avg loss: 1.48369049, Global Avg Loss: 1.73079474, Time: 0.0403 Steps: 44180, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001610, Sample Num: 25760, Cur Loss: 0.47341073, Cur Avg Loss: 1.39119236, Log Avg loss: 1.12027448, Global Avg Loss: 1.73065658, Time: 0.0402 Steps: 44190, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001620, Sample Num: 25920, Cur Loss: 1.77367425, Cur Avg Loss: 1.39067486, Log Avg loss: 1.30735722, Global Avg Loss: 1.73056081, Time: 0.0403 Steps: 44200, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001630, Sample Num: 26080, Cur Loss: 0.32194448, Cur Avg Loss: 1.38976926, Log Avg loss: 1.24306192, Global Avg Loss: 1.73045054, Time: 0.0403 Steps: 44210, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001640, Sample Num: 26240, Cur Loss: 1.08339918, Cur Avg Loss: 1.39131499, Log Avg loss: 1.64326874, Global Avg Loss: 1.73043083, Time: 0.0402 Steps: 44220, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001650, Sample Num: 26400, Cur Loss: 1.41742575, Cur Avg Loss: 1.39103222, Log Avg loss: 1.34465886, Global Avg Loss: 1.73034361, Time: 0.0402 Steps: 44230, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001660, Sample Num: 26560, Cur Loss: 1.12683415, Cur Avg Loss: 1.39145619, Log Avg loss: 1.46141096, Global Avg Loss: 1.73028282, Time: 0.0403 Steps: 44240, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001670, Sample Num: 26720, Cur Loss: 1.52174163, Cur Avg Loss: 1.39050174, Log Avg loss: 1.23206350, Global Avg Loss: 1.73017022, Time: 0.0788 Steps: 44250, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001680, Sample Num: 26880, Cur Loss: 1.51684785, Cur Avg Loss: 1.39079599, Log Avg loss: 1.43993549, Global Avg Loss: 1.73010465, Time: 0.0509 Steps: 44260, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001690, Sample Num: 27040, Cur Loss: 0.54151851, Cur Avg Loss: 1.38910274, Log Avg loss: 1.10463618, Global Avg Loss: 1.72996336, Time: 0.0481 Steps: 44270, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001700, Sample Num: 27200, Cur Loss: 3.58982563, Cur Avg Loss: 1.39008185, Log Avg loss: 1.55555094, Global Avg Loss: 1.72992398, Time: 0.0669 Steps: 44280, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001710, Sample Num: 27360, Cur Loss: 1.26966739, Cur Avg Loss: 1.38897905, Log Avg loss: 1.20150298, Global Avg Loss: 1.72980467, Time: 0.0637 Steps: 44290, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001720, Sample Num: 27520, Cur Loss: 1.00389373, Cur Avg Loss: 1.38881395, Log Avg loss: 1.36058301, Global Avg Loss: 1.72972132, Time: 0.0575 Steps: 44300, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001730, Sample Num: 27680, Cur Loss: 1.16218150, Cur Avg Loss: 1.39013403, Log Avg loss: 1.61718732, Global Avg Loss: 1.72969592, Time: 0.0706 Steps: 44310, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001740, Sample Num: 27840, Cur Loss: 1.54869747, Cur Avg Loss: 1.39227687, Log Avg loss: 1.76298804, Global Avg Loss: 1.72970344, Time: 0.0401 Steps: 44320, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001750, Sample Num: 28000, Cur Loss: 1.65481544, Cur Avg Loss: 1.39295361, Log Avg loss: 1.51070709, Global Avg Loss: 1.72965403, Time: 0.0921 Steps: 44330, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001760, Sample Num: 28160, Cur Loss: 1.64013660, Cur Avg Loss: 1.39202909, Log Avg loss: 1.23023748, Global Avg Loss: 1.72954140, Time: 0.0597 Steps: 44340, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001770, Sample Num: 28320, Cur Loss: 0.73199642, Cur Avg Loss: 1.39023928, Log Avg loss: 1.07523299, Global Avg Loss: 1.72939387, Time: 0.0403 Steps: 44350, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001780, Sample Num: 28480, Cur Loss: 1.65568972, Cur Avg Loss: 1.39001508, Log Avg loss: 1.35033217, Global Avg Loss: 1.72930842, Time: 0.0402 Steps: 44360, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001790, Sample Num: 28640, Cur Loss: 0.87812716, Cur Avg Loss: 1.38951396, Log Avg loss: 1.30031380, Global Avg Loss: 1.72921173, Time: 0.0402 Steps: 44370, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001800, Sample Num: 28800, Cur Loss: 1.24232543, Cur Avg Loss: 1.39033395, Log Avg loss: 1.53711204, Global Avg Loss: 1.72916845, Time: 0.0402 Steps: 44380, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001810, Sample Num: 28960, Cur Loss: 2.19100094, Cur Avg Loss: 1.39004839, Log Avg loss: 1.33864844, Global Avg Loss: 1.72908047, Time: 0.0403 Steps: 44390, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001820, Sample Num: 29120, Cur Loss: 2.20458364, Cur Avg Loss: 1.39110121, Log Avg loss: 1.58166173, Global Avg Loss: 1.72904727, Time: 0.0402 Steps: 44400, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001830, Sample Num: 29280, Cur Loss: 1.69460106, Cur Avg Loss: 1.39203208, Log Avg loss: 1.56144997, Global Avg Loss: 1.72900953, Time: 0.0403 Steps: 44410, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001840, Sample Num: 29440, Cur Loss: 0.63103199, Cur Avg Loss: 1.39175206, Log Avg loss: 1.34050885, Global Avg Loss: 1.72892207, Time: 0.0403 Steps: 44420, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001850, Sample Num: 29600, Cur Loss: 1.28558660, Cur Avg Loss: 1.39195402, Log Avg loss: 1.42911444, Global Avg Loss: 1.72885459, Time: 0.0403 Steps: 44430, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001860, Sample Num: 29760, Cur Loss: 0.24550183, Cur Avg Loss: 1.39119779, Log Avg loss: 1.25129498, Global Avg Loss: 1.72874713, Time: 0.0402 Steps: 44440, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001870, Sample Num: 29920, Cur Loss: 1.94776094, Cur Avg Loss: 1.39093581, Log Avg loss: 1.34220717, Global Avg Loss: 1.72866017, Time: 0.0403 Steps: 44450, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001880, Sample Num: 30080, Cur Loss: 2.07530332, Cur Avg Loss: 1.39081566, Log Avg loss: 1.36834852, Global Avg Loss: 1.72857913, Time: 0.0402 Steps: 44460, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001890, Sample Num: 30240, Cur Loss: 0.93413782, Cur Avg Loss: 1.39153282, Log Avg loss: 1.52635878, Global Avg Loss: 1.72853365, Time: 0.0403 Steps: 44470, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001900, Sample Num: 30400, Cur Loss: 0.61934400, Cur Avg Loss: 1.39149945, Log Avg loss: 1.38519211, Global Avg Loss: 1.72845646, Time: 0.0403 Steps: 44480, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001910, Sample Num: 30560, Cur Loss: 2.09501338, Cur Avg Loss: 1.39233558, Log Avg loss: 1.55120007, Global Avg Loss: 1.72841662, Time: 0.0403 Steps: 44490, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001920, Sample Num: 30720, Cur Loss: 0.95126033, Cur Avg Loss: 1.39212111, Log Avg loss: 1.35115682, Global Avg Loss: 1.72833184, Time: 0.0402 Steps: 44500, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001930, Sample Num: 30880, Cur Loss: 2.23158932, Cur Avg Loss: 1.39195782, Log Avg loss: 1.36060566, Global Avg Loss: 1.72824923, Time: 0.0402 Steps: 44510, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001940, Sample Num: 31040, Cur Loss: 1.43104863, Cur Avg Loss: 1.39275353, Log Avg loss: 1.54632602, Global Avg Loss: 1.72820836, Time: 0.0402 Steps: 44520, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001950, Sample Num: 31200, Cur Loss: 1.52989435, Cur Avg Loss: 1.39332871, Log Avg loss: 1.50491332, Global Avg Loss: 1.72815822, Time: 0.0402 Steps: 44530, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001960, Sample Num: 31360, Cur Loss: 0.98813832, Cur Avg Loss: 1.39273547, Log Avg loss: 1.27705492, Global Avg Loss: 1.72805694, Time: 0.0402 Steps: 44540, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001970, Sample Num: 31520, Cur Loss: 2.47731733, Cur Avg Loss: 1.39354192, Log Avg loss: 1.55160660, Global Avg Loss: 1.72801733, Time: 0.0402 Steps: 44550, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001980, Sample Num: 31680, Cur Loss: 1.02784789, Cur Avg Loss: 1.39415724, Log Avg loss: 1.51537409, Global Avg Loss: 1.72796961, Time: 0.0402 Steps: 44560, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001990, Sample Num: 31840, Cur Loss: 0.67864299, Cur Avg Loss: 1.39404382, Log Avg loss: 1.37158739, Global Avg Loss: 1.72788965, Time: 0.0402 Steps: 44570, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002000, Sample Num: 32000, Cur Loss: 2.76683164, Cur Avg Loss: 1.39377992, Log Avg loss: 1.34126248, Global Avg Loss: 1.72780292, Time: 0.0402 Steps: 44580, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002010, Sample Num: 32160, Cur Loss: 2.19477820, Cur Avg Loss: 1.39574576, Log Avg loss: 1.78891470, Global Avg Loss: 1.72781663, Time: 0.0402 Steps: 44590, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002020, Sample Num: 32320, Cur Loss: 0.67495352, Cur Avg Loss: 1.39545972, Log Avg loss: 1.33796619, Global Avg Loss: 1.72772922, Time: 0.0405 Steps: 44600, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002030, Sample Num: 32480, Cur Loss: 1.17595792, Cur Avg Loss: 1.39536595, Log Avg loss: 1.37642481, Global Avg Loss: 1.72765047, Time: 0.0504 Steps: 44610, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002040, Sample Num: 32640, Cur Loss: 1.95830691, Cur Avg Loss: 1.39562886, Log Avg loss: 1.44899881, Global Avg Loss: 1.72758802, Time: 0.0711 Steps: 44620, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002050, Sample Num: 32800, Cur Loss: 3.00162172, Cur Avg Loss: 1.39689861, Log Avg loss: 1.65592723, Global Avg Loss: 1.72757196, Time: 0.0550 Steps: 44630, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002060, Sample Num: 32960, Cur Loss: 1.89563918, Cur Avg Loss: 1.39629223, Log Avg loss: 1.27198525, Global Avg Loss: 1.72746990, Time: 0.0794 Steps: 44640, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002070, Sample Num: 33120, Cur Loss: 1.79732513, Cur Avg Loss: 1.39713954, Log Avg loss: 1.57168478, Global Avg Loss: 1.72743501, Time: 0.0583 Steps: 44650, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002080, Sample Num: 33280, Cur Loss: 2.54244995, Cur Avg Loss: 1.39715855, Log Avg loss: 1.40109305, Global Avg Loss: 1.72736194, Time: 0.0806 Steps: 44660, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002090, Sample Num: 33440, Cur Loss: 0.97762537, Cur Avg Loss: 1.39736159, Log Avg loss: 1.43959417, Global Avg Loss: 1.72729752, Time: 0.0519 Steps: 44670, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002100, Sample Num: 33600, Cur Loss: 1.64660573, Cur Avg Loss: 1.39636577, Log Avg loss: 1.18824055, Global Avg Loss: 1.72717687, Time: 0.0404 Steps: 44680, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002110, Sample Num: 33760, Cur Loss: 1.95501876, Cur Avg Loss: 1.39621257, Log Avg loss: 1.36404041, Global Avg Loss: 1.72709562, Time: 0.0403 Steps: 44690, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002120, Sample Num: 33920, Cur Loss: 0.93694556, Cur Avg Loss: 1.39643282, Log Avg loss: 1.44290373, Global Avg Loss: 1.72703204, Time: 0.0403 Steps: 44700, Updated lr: 0.000059 ***** Running evaluation checkpoint-44709 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-44709 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.490062, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.306457, "eval_total_loss": 918.439069, "eval_mae": 0.84966, "eval_mse": 1.306785, "eval_r2": 0.169322, "eval_sp_statistic": 0.351549, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.435886, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.22355, "test_total_loss": 614.22188, "test_mae": 0.819939, "test_mse": 1.223843, "test_r2": 0.210121, "test_sp_statistic": 0.371787, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.459507, "test_ps_pvalue": 0.0, "lr": 5.855002370791844e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7269673706525503, "train_cur_epoch_loss": 2973.0896642655134, "train_cur_epoch_avg_loss": 1.3964723646150838, "train_cur_epoch_time": 95.49006247520447, "train_cur_epoch_avg_time": 0.04485207255763479, "epoch": 21, "step": 44709} ################################################## Training, Epoch: 0022, Batch: 000001, Sample Num: 16, Cur Loss: 1.55969179, Cur Avg Loss: 1.55969179, Log Avg loss: 1.42117876, Global Avg Loss: 1.72696363, Time: 0.0821 Steps: 44710, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000011, Sample Num: 176, Cur Loss: 0.97976017, Cur Avg Loss: 1.23490624, Log Avg loss: 1.20242768, Global Avg Loss: 1.72684634, Time: 0.0543 Steps: 44720, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000021, Sample Num: 336, Cur Loss: 1.61998081, Cur Avg Loss: 1.38948556, Log Avg loss: 1.55952283, Global Avg Loss: 1.72680893, Time: 0.0575 Steps: 44730, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000031, Sample Num: 496, Cur Loss: 0.96263444, Cur Avg Loss: 1.38563746, Log Avg loss: 1.37755643, Global Avg Loss: 1.72673087, Time: 0.1013 Steps: 44740, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000041, Sample Num: 656, Cur Loss: 1.05060220, Cur Avg Loss: 1.32151687, Log Avg loss: 1.12274306, Global Avg Loss: 1.72659590, Time: 0.0691 Steps: 44750, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000051, Sample Num: 816, Cur Loss: 1.47461271, Cur Avg Loss: 1.35346863, Log Avg loss: 1.48447086, Global Avg Loss: 1.72654180, Time: 0.0403 Steps: 44760, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000061, Sample Num: 976, Cur Loss: 1.31551957, Cur Avg Loss: 1.37037637, Log Avg loss: 1.45660580, Global Avg Loss: 1.72648151, Time: 0.0403 Steps: 44770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000071, Sample Num: 1136, Cur Loss: 1.01540995, Cur Avg Loss: 1.37152202, Log Avg loss: 1.37851053, Global Avg Loss: 1.72640380, Time: 0.0403 Steps: 44780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000081, Sample Num: 1296, Cur Loss: 1.38695765, Cur Avg Loss: 1.39868276, Log Avg loss: 1.59152397, Global Avg Loss: 1.72637369, Time: 0.0402 Steps: 44790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000091, Sample Num: 1456, Cur Loss: 0.88297057, Cur Avg Loss: 1.36423028, Log Avg loss: 1.08516518, Global Avg Loss: 1.72623056, Time: 0.0402 Steps: 44800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000101, Sample Num: 1616, Cur Loss: 1.12018871, Cur Avg Loss: 1.37541331, Log Avg loss: 1.47717894, Global Avg Loss: 1.72617498, Time: 0.0402 Steps: 44810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000111, Sample Num: 1776, Cur Loss: 1.12146020, Cur Avg Loss: 1.37162520, Log Avg loss: 1.33336523, Global Avg Loss: 1.72608734, Time: 0.0403 Steps: 44820, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000121, Sample Num: 1936, Cur Loss: 0.76956123, Cur Avg Loss: 1.40272925, Log Avg loss: 1.74798430, Global Avg Loss: 1.72609222, Time: 0.0403 Steps: 44830, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000131, Sample Num: 2096, Cur Loss: 0.44772732, Cur Avg Loss: 1.39191259, Log Avg loss: 1.26103095, Global Avg Loss: 1.72598851, Time: 0.0402 Steps: 44840, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000141, Sample Num: 2256, Cur Loss: 2.12864733, Cur Avg Loss: 1.38634579, Log Avg loss: 1.31342064, Global Avg Loss: 1.72589652, Time: 0.0403 Steps: 44850, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000151, Sample Num: 2416, Cur Loss: 1.01789272, Cur Avg Loss: 1.38984421, Log Avg loss: 1.43917193, Global Avg Loss: 1.72583260, Time: 0.0402 Steps: 44860, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000161, Sample Num: 2576, Cur Loss: 1.00377035, Cur Avg Loss: 1.39805321, Log Avg loss: 1.52200910, Global Avg Loss: 1.72578718, Time: 0.0402 Steps: 44870, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000171, Sample Num: 2736, Cur Loss: 0.89013785, Cur Avg Loss: 1.39243310, Log Avg loss: 1.30194941, Global Avg Loss: 1.72569274, Time: 0.0403 Steps: 44880, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000181, Sample Num: 2896, Cur Loss: 1.79906249, Cur Avg Loss: 1.40235689, Log Avg loss: 1.57205371, Global Avg Loss: 1.72565852, Time: 0.0403 Steps: 44890, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000191, Sample Num: 3056, Cur Loss: 0.71033680, Cur Avg Loss: 1.40063313, Log Avg loss: 1.36943297, Global Avg Loss: 1.72557918, Time: 0.0403 Steps: 44900, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000201, Sample Num: 3216, Cur Loss: 1.13558364, Cur Avg Loss: 1.38669122, Log Avg loss: 1.12040086, Global Avg Loss: 1.72544442, Time: 0.0402 Steps: 44910, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000211, Sample Num: 3376, Cur Loss: 1.21430039, Cur Avg Loss: 1.38533406, Log Avg loss: 1.35805519, Global Avg Loss: 1.72536264, Time: 0.0402 Steps: 44920, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000221, Sample Num: 3536, Cur Loss: 0.98105913, Cur Avg Loss: 1.38986278, Log Avg loss: 1.48541869, Global Avg Loss: 1.72530923, Time: 0.0402 Steps: 44930, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000231, Sample Num: 3696, Cur Loss: 2.79614067, Cur Avg Loss: 1.38894985, Log Avg loss: 1.36877400, Global Avg Loss: 1.72522990, Time: 0.0403 Steps: 44940, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000241, Sample Num: 3856, Cur Loss: 1.27165604, Cur Avg Loss: 1.38471848, Log Avg loss: 1.28697386, Global Avg Loss: 1.72513240, Time: 0.0403 Steps: 44950, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000251, Sample Num: 4016, Cur Loss: 1.36162400, Cur Avg Loss: 1.38375187, Log Avg loss: 1.36045675, Global Avg Loss: 1.72505129, Time: 0.0402 Steps: 44960, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000261, Sample Num: 4176, Cur Loss: 1.02143645, Cur Avg Loss: 1.37961304, Log Avg loss: 1.27572837, Global Avg Loss: 1.72495137, Time: 0.0402 Steps: 44970, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000271, Sample Num: 4336, Cur Loss: 1.37857831, Cur Avg Loss: 1.38415963, Log Avg loss: 1.50282550, Global Avg Loss: 1.72490199, Time: 0.0401 Steps: 44980, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000281, Sample Num: 4496, Cur Loss: 0.29247549, Cur Avg Loss: 1.39070808, Log Avg loss: 1.56817125, Global Avg Loss: 1.72486715, Time: 0.0403 Steps: 44990, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000291, Sample Num: 4656, Cur Loss: 0.31142384, Cur Avg Loss: 1.38516928, Log Avg loss: 1.22952890, Global Avg Loss: 1.72475708, Time: 0.0403 Steps: 45000, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000301, Sample Num: 4816, Cur Loss: 2.50310183, Cur Avg Loss: 1.39790419, Log Avg loss: 1.76849015, Global Avg Loss: 1.72476679, Time: 0.0402 Steps: 45010, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000311, Sample Num: 4976, Cur Loss: 0.75123227, Cur Avg Loss: 1.39370955, Log Avg loss: 1.26745086, Global Avg Loss: 1.72466521, Time: 0.0410 Steps: 45020, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000321, Sample Num: 5136, Cur Loss: 1.83313394, Cur Avg Loss: 1.40168687, Log Avg loss: 1.64978137, Global Avg Loss: 1.72464858, Time: 0.0605 Steps: 45030, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000331, Sample Num: 5296, Cur Loss: 0.38036782, Cur Avg Loss: 1.39574576, Log Avg loss: 1.20503604, Global Avg Loss: 1.72453322, Time: 0.0546 Steps: 45040, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000341, Sample Num: 5456, Cur Loss: 1.06571376, Cur Avg Loss: 1.38906780, Log Avg loss: 1.16802757, Global Avg Loss: 1.72440968, Time: 0.0663 Steps: 45050, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000351, Sample Num: 5616, Cur Loss: 2.54820776, Cur Avg Loss: 1.39272866, Log Avg loss: 1.51756378, Global Avg Loss: 1.72436378, Time: 0.0548 Steps: 45060, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000361, Sample Num: 5776, Cur Loss: 0.76410854, Cur Avg Loss: 1.39098358, Log Avg loss: 1.32973152, Global Avg Loss: 1.72427622, Time: 0.0405 Steps: 45070, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000371, Sample Num: 5936, Cur Loss: 0.67292905, Cur Avg Loss: 1.39031337, Log Avg loss: 1.36611865, Global Avg Loss: 1.72419677, Time: 0.0485 Steps: 45080, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000381, Sample Num: 6096, Cur Loss: 1.47185457, Cur Avg Loss: 1.39341534, Log Avg loss: 1.50849850, Global Avg Loss: 1.72414893, Time: 0.0663 Steps: 45090, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000391, Sample Num: 6256, Cur Loss: 1.29220724, Cur Avg Loss: 1.39321618, Log Avg loss: 1.38562822, Global Avg Loss: 1.72407387, Time: 0.0619 Steps: 45100, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000401, Sample Num: 6416, Cur Loss: 1.56810176, Cur Avg Loss: 1.40413633, Log Avg loss: 1.83111410, Global Avg Loss: 1.72409760, Time: 0.0404 Steps: 45110, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000411, Sample Num: 6576, Cur Loss: 2.53198099, Cur Avg Loss: 1.41037727, Log Avg loss: 1.66063882, Global Avg Loss: 1.72408354, Time: 0.0403 Steps: 45120, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000421, Sample Num: 6736, Cur Loss: 1.10616779, Cur Avg Loss: 1.40087581, Log Avg loss: 1.01036604, Global Avg Loss: 1.72392539, Time: 0.0403 Steps: 45130, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000431, Sample Num: 6896, Cur Loss: 1.63486934, Cur Avg Loss: 1.39839189, Log Avg loss: 1.29381860, Global Avg Loss: 1.72383011, Time: 0.0402 Steps: 45140, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000441, Sample Num: 7056, Cur Loss: 1.31566596, Cur Avg Loss: 1.38872047, Log Avg loss: 0.97188223, Global Avg Loss: 1.72366356, Time: 0.0404 Steps: 45150, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000451, Sample Num: 7216, Cur Loss: 0.37850899, Cur Avg Loss: 1.38721328, Log Avg loss: 1.32074658, Global Avg Loss: 1.72357434, Time: 0.0402 Steps: 45160, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000461, Sample Num: 7376, Cur Loss: 1.96569800, Cur Avg Loss: 1.38429885, Log Avg loss: 1.25285799, Global Avg Loss: 1.72347013, Time: 0.0404 Steps: 45170, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000471, Sample Num: 7536, Cur Loss: 1.41190517, Cur Avg Loss: 1.38725953, Log Avg loss: 1.52374698, Global Avg Loss: 1.72342593, Time: 0.0403 Steps: 45180, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000481, Sample Num: 7696, Cur Loss: 1.19142413, Cur Avg Loss: 1.38482546, Log Avg loss: 1.27018044, Global Avg Loss: 1.72332563, Time: 0.0403 Steps: 45190, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000491, Sample Num: 7856, Cur Loss: 0.57002938, Cur Avg Loss: 1.38336573, Log Avg loss: 1.31315301, Global Avg Loss: 1.72323488, Time: 0.0403 Steps: 45200, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000501, Sample Num: 8016, Cur Loss: 0.94831347, Cur Avg Loss: 1.39401495, Log Avg loss: 1.91689138, Global Avg Loss: 1.72327772, Time: 0.0403 Steps: 45210, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000511, Sample Num: 8176, Cur Loss: 1.03178513, Cur Avg Loss: 1.38852144, Log Avg loss: 1.11329671, Global Avg Loss: 1.72314283, Time: 0.0404 Steps: 45220, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000521, Sample Num: 8336, Cur Loss: 0.57423997, Cur Avg Loss: 1.38844174, Log Avg loss: 1.38436884, Global Avg Loss: 1.72306793, Time: 0.0403 Steps: 45230, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000531, Sample Num: 8496, Cur Loss: 1.40830600, Cur Avg Loss: 1.38613035, Log Avg loss: 1.26570735, Global Avg Loss: 1.72296683, Time: 0.0403 Steps: 45240, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000541, Sample Num: 8656, Cur Loss: 0.26327744, Cur Avg Loss: 1.38132410, Log Avg loss: 1.12611210, Global Avg Loss: 1.72283493, Time: 0.0402 Steps: 45250, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000551, Sample Num: 8816, Cur Loss: 1.13918757, Cur Avg Loss: 1.38311772, Log Avg loss: 1.48015265, Global Avg Loss: 1.72278131, Time: 0.0402 Steps: 45260, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000561, Sample Num: 8976, Cur Loss: 1.63629758, Cur Avg Loss: 1.39115192, Log Avg loss: 1.83383590, Global Avg Loss: 1.72280584, Time: 0.0403 Steps: 45270, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000571, Sample Num: 9136, Cur Loss: 1.00677490, Cur Avg Loss: 1.39325116, Log Avg loss: 1.51101883, Global Avg Loss: 1.72275907, Time: 0.0403 Steps: 45280, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000581, Sample Num: 9296, Cur Loss: 1.42581379, Cur Avg Loss: 1.39204301, Log Avg loss: 1.32305771, Global Avg Loss: 1.72267081, Time: 0.0402 Steps: 45290, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000591, Sample Num: 9456, Cur Loss: 1.19171631, Cur Avg Loss: 1.38969773, Log Avg loss: 1.25343706, Global Avg Loss: 1.72256723, Time: 0.0402 Steps: 45300, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000601, Sample Num: 9616, Cur Loss: 0.80219960, Cur Avg Loss: 1.38879195, Log Avg loss: 1.33526020, Global Avg Loss: 1.72248175, Time: 0.0402 Steps: 45310, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000611, Sample Num: 9776, Cur Loss: 1.91064799, Cur Avg Loss: 1.38958842, Log Avg loss: 1.43745593, Global Avg Loss: 1.72241886, Time: 0.0403 Steps: 45320, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000621, Sample Num: 9936, Cur Loss: 0.91351926, Cur Avg Loss: 1.38469636, Log Avg loss: 1.08579154, Global Avg Loss: 1.72227842, Time: 0.0402 Steps: 45330, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000631, Sample Num: 10096, Cur Loss: 1.34495318, Cur Avg Loss: 1.38396030, Log Avg loss: 1.33825135, Global Avg Loss: 1.72219372, Time: 0.0402 Steps: 45340, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000641, Sample Num: 10256, Cur Loss: 1.76549196, Cur Avg Loss: 1.38477191, Log Avg loss: 1.43598434, Global Avg Loss: 1.72213061, Time: 0.0402 Steps: 45350, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000651, Sample Num: 10416, Cur Loss: 0.83740515, Cur Avg Loss: 1.37655137, Log Avg loss: 0.84961466, Global Avg Loss: 1.72193825, Time: 0.0477 Steps: 45360, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000661, Sample Num: 10576, Cur Loss: 0.83257699, Cur Avg Loss: 1.37624979, Log Avg loss: 1.35661718, Global Avg Loss: 1.72185773, Time: 0.0402 Steps: 45370, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000671, Sample Num: 10736, Cur Loss: 1.39151907, Cur Avg Loss: 1.37213025, Log Avg loss: 1.09982852, Global Avg Loss: 1.72172066, Time: 0.0455 Steps: 45380, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000681, Sample Num: 10896, Cur Loss: 0.66549903, Cur Avg Loss: 1.36264777, Log Avg loss: 0.72637323, Global Avg Loss: 1.72150137, Time: 0.0611 Steps: 45390, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000691, Sample Num: 11056, Cur Loss: 0.69765168, Cur Avg Loss: 1.35886930, Log Avg loss: 1.10155577, Global Avg Loss: 1.72136482, Time: 0.0676 Steps: 45400, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000701, Sample Num: 11216, Cur Loss: 1.45740771, Cur Avg Loss: 1.36112887, Log Avg loss: 1.51726519, Global Avg Loss: 1.72131987, Time: 0.0557 Steps: 45410, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000711, Sample Num: 11376, Cur Loss: 0.99938095, Cur Avg Loss: 1.36346443, Log Avg loss: 1.52718719, Global Avg Loss: 1.72127713, Time: 0.0405 Steps: 45420, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000721, Sample Num: 11536, Cur Loss: 2.24640059, Cur Avg Loss: 1.36424345, Log Avg loss: 1.41963149, Global Avg Loss: 1.72121074, Time: 0.0469 Steps: 45430, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000731, Sample Num: 11696, Cur Loss: 1.36335552, Cur Avg Loss: 1.36360694, Log Avg loss: 1.31771449, Global Avg Loss: 1.72112194, Time: 0.0674 Steps: 45440, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000741, Sample Num: 11856, Cur Loss: 0.60038674, Cur Avg Loss: 1.36180392, Log Avg loss: 1.23000299, Global Avg Loss: 1.72101388, Time: 0.0403 Steps: 45450, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000751, Sample Num: 12016, Cur Loss: 0.54052281, Cur Avg Loss: 1.36074828, Log Avg loss: 1.28252581, Global Avg Loss: 1.72091742, Time: 0.0403 Steps: 45460, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000761, Sample Num: 12176, Cur Loss: 1.46074772, Cur Avg Loss: 1.36192211, Log Avg loss: 1.45007653, Global Avg Loss: 1.72085786, Time: 0.0402 Steps: 45470, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000771, Sample Num: 12336, Cur Loss: 1.41235590, Cur Avg Loss: 1.36099823, Log Avg loss: 1.29069142, Global Avg Loss: 1.72076328, Time: 0.0402 Steps: 45480, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000781, Sample Num: 12496, Cur Loss: 0.64377326, Cur Avg Loss: 1.35890888, Log Avg loss: 1.19781928, Global Avg Loss: 1.72064832, Time: 0.0402 Steps: 45490, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000791, Sample Num: 12656, Cur Loss: 0.61732596, Cur Avg Loss: 1.36044932, Log Avg loss: 1.48075816, Global Avg Loss: 1.72059560, Time: 0.0404 Steps: 45500, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000801, Sample Num: 12816, Cur Loss: 1.59888411, Cur Avg Loss: 1.36275113, Log Avg loss: 1.54482379, Global Avg Loss: 1.72055697, Time: 0.0402 Steps: 45510, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000811, Sample Num: 12976, Cur Loss: 2.22799969, Cur Avg Loss: 1.36585626, Log Avg loss: 1.61457785, Global Avg Loss: 1.72053369, Time: 0.0403 Steps: 45520, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000821, Sample Num: 13136, Cur Loss: 1.64065862, Cur Avg Loss: 1.36633390, Log Avg loss: 1.40507029, Global Avg Loss: 1.72046440, Time: 0.0403 Steps: 45530, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000831, Sample Num: 13296, Cur Loss: 1.46630657, Cur Avg Loss: 1.36590881, Log Avg loss: 1.33100891, Global Avg Loss: 1.72037888, Time: 0.0403 Steps: 45540, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000841, Sample Num: 13456, Cur Loss: 0.48272437, Cur Avg Loss: 1.36452891, Log Avg loss: 1.24985933, Global Avg Loss: 1.72027559, Time: 0.0402 Steps: 45550, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000851, Sample Num: 13616, Cur Loss: 0.61687177, Cur Avg Loss: 1.36091244, Log Avg loss: 1.05676703, Global Avg Loss: 1.72012995, Time: 0.0403 Steps: 45560, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000861, Sample Num: 13776, Cur Loss: 0.66687030, Cur Avg Loss: 1.36255370, Log Avg loss: 1.50222489, Global Avg Loss: 1.72008214, Time: 0.0403 Steps: 45570, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000871, Sample Num: 13936, Cur Loss: 1.48924994, Cur Avg Loss: 1.36205354, Log Avg loss: 1.31899000, Global Avg Loss: 1.71999414, Time: 0.0403 Steps: 45580, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000881, Sample Num: 14096, Cur Loss: 1.13598633, Cur Avg Loss: 1.36249605, Log Avg loss: 1.40103844, Global Avg Loss: 1.71992418, Time: 0.0403 Steps: 45590, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000891, Sample Num: 14256, Cur Loss: 2.11816549, Cur Avg Loss: 1.36378499, Log Avg loss: 1.47734030, Global Avg Loss: 1.71987098, Time: 0.0402 Steps: 45600, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000901, Sample Num: 14416, Cur Loss: 1.43148887, Cur Avg Loss: 1.36029029, Log Avg loss: 1.04891297, Global Avg Loss: 1.71972387, Time: 0.0403 Steps: 45610, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000911, Sample Num: 14576, Cur Loss: 0.99120802, Cur Avg Loss: 1.35763868, Log Avg loss: 1.11872844, Global Avg Loss: 1.71959213, Time: 0.0402 Steps: 45620, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000921, Sample Num: 14736, Cur Loss: 0.72846556, Cur Avg Loss: 1.35730856, Log Avg loss: 1.32723470, Global Avg Loss: 1.71950614, Time: 0.0402 Steps: 45630, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000931, Sample Num: 14896, Cur Loss: 1.35335958, Cur Avg Loss: 1.36039661, Log Avg loss: 1.64480625, Global Avg Loss: 1.71948978, Time: 0.0402 Steps: 45640, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000941, Sample Num: 15056, Cur Loss: 1.91071844, Cur Avg Loss: 1.36225898, Log Avg loss: 1.53564502, Global Avg Loss: 1.71944950, Time: 0.0403 Steps: 45650, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000951, Sample Num: 15216, Cur Loss: 1.10267174, Cur Avg Loss: 1.36248522, Log Avg loss: 1.38377448, Global Avg Loss: 1.71937599, Time: 0.0402 Steps: 45660, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000961, Sample Num: 15376, Cur Loss: 2.40628004, Cur Avg Loss: 1.36248914, Log Avg loss: 1.36286253, Global Avg Loss: 1.71929793, Time: 0.0405 Steps: 45670, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000971, Sample Num: 15536, Cur Loss: 1.58382893, Cur Avg Loss: 1.36320080, Log Avg loss: 1.43159159, Global Avg Loss: 1.71923494, Time: 0.0402 Steps: 45680, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000981, Sample Num: 15696, Cur Loss: 2.89283061, Cur Avg Loss: 1.36446707, Log Avg loss: 1.48742094, Global Avg Loss: 1.71918421, Time: 0.0402 Steps: 45690, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000991, Sample Num: 15856, Cur Loss: 0.85009658, Cur Avg Loss: 1.36308145, Log Avg loss: 1.22715220, Global Avg Loss: 1.71907654, Time: 0.0403 Steps: 45700, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001001, Sample Num: 16016, Cur Loss: 3.42809343, Cur Avg Loss: 1.36693537, Log Avg loss: 1.74885946, Global Avg Loss: 1.71908306, Time: 0.0402 Steps: 45710, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001011, Sample Num: 16176, Cur Loss: 1.60054052, Cur Avg Loss: 1.36929696, Log Avg loss: 1.60569206, Global Avg Loss: 1.71905825, Time: 0.0402 Steps: 45720, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001021, Sample Num: 16336, Cur Loss: 1.17699683, Cur Avg Loss: 1.36875222, Log Avg loss: 1.31367885, Global Avg Loss: 1.71896961, Time: 0.1024 Steps: 45730, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001031, Sample Num: 16496, Cur Loss: 1.92288923, Cur Avg Loss: 1.37168433, Log Avg loss: 1.67105326, Global Avg Loss: 1.71895913, Time: 0.0658 Steps: 45740, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001041, Sample Num: 16656, Cur Loss: 2.16832924, Cur Avg Loss: 1.37102235, Log Avg loss: 1.30277209, Global Avg Loss: 1.71886816, Time: 0.0558 Steps: 45750, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001051, Sample Num: 16816, Cur Loss: 1.84811831, Cur Avg Loss: 1.37014419, Log Avg loss: 1.27872705, Global Avg Loss: 1.71877198, Time: 0.0606 Steps: 45760, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001061, Sample Num: 16976, Cur Loss: 1.41617477, Cur Avg Loss: 1.37347610, Log Avg loss: 1.72366059, Global Avg Loss: 1.71877305, Time: 0.0456 Steps: 45770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001071, Sample Num: 17136, Cur Loss: 1.09723163, Cur Avg Loss: 1.37444290, Log Avg loss: 1.47701980, Global Avg Loss: 1.71872024, Time: 0.0412 Steps: 45780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001081, Sample Num: 17296, Cur Loss: 1.51294947, Cur Avg Loss: 1.37446268, Log Avg loss: 1.37658141, Global Avg Loss: 1.71864552, Time: 0.0862 Steps: 45790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001091, Sample Num: 17456, Cur Loss: 1.12546015, Cur Avg Loss: 1.37276407, Log Avg loss: 1.18914365, Global Avg Loss: 1.71852991, Time: 0.0425 Steps: 45800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001101, Sample Num: 17616, Cur Loss: 0.46271157, Cur Avg Loss: 1.37373940, Log Avg loss: 1.48014805, Global Avg Loss: 1.71847787, Time: 0.0471 Steps: 45810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001111, Sample Num: 17776, Cur Loss: 2.70910454, Cur Avg Loss: 1.37596282, Log Avg loss: 1.62076134, Global Avg Loss: 1.71845654, Time: 0.0401 Steps: 45820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001121, Sample Num: 17936, Cur Loss: 1.90492451, Cur Avg Loss: 1.37633946, Log Avg loss: 1.41818480, Global Avg Loss: 1.71839103, Time: 0.0403 Steps: 45830, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001131, Sample Num: 18096, Cur Loss: 0.60850537, Cur Avg Loss: 1.37356706, Log Avg loss: 1.06278061, Global Avg Loss: 1.71824800, Time: 0.0402 Steps: 45840, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001141, Sample Num: 18256, Cur Loss: 1.88337135, Cur Avg Loss: 1.37520563, Log Avg loss: 1.56052806, Global Avg Loss: 1.71821361, Time: 0.0403 Steps: 45850, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001151, Sample Num: 18416, Cur Loss: 0.43423739, Cur Avg Loss: 1.37478268, Log Avg loss: 1.32652368, Global Avg Loss: 1.71812820, Time: 0.0403 Steps: 45860, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001161, Sample Num: 18576, Cur Loss: 1.39479709, Cur Avg Loss: 1.37615874, Log Avg loss: 1.53454334, Global Avg Loss: 1.71808817, Time: 0.0402 Steps: 45870, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001171, Sample Num: 18736, Cur Loss: 1.24478412, Cur Avg Loss: 1.37539789, Log Avg loss: 1.28706336, Global Avg Loss: 1.71799423, Time: 0.0402 Steps: 45880, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001181, Sample Num: 18896, Cur Loss: 1.54347539, Cur Avg Loss: 1.37865034, Log Avg loss: 1.75951232, Global Avg Loss: 1.71800327, Time: 0.0402 Steps: 45890, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001191, Sample Num: 19056, Cur Loss: 1.24127614, Cur Avg Loss: 1.37881593, Log Avg loss: 1.39837247, Global Avg Loss: 1.71793364, Time: 0.0402 Steps: 45900, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001201, Sample Num: 19216, Cur Loss: 0.59346616, Cur Avg Loss: 1.37794311, Log Avg loss: 1.27399023, Global Avg Loss: 1.71783694, Time: 0.0403 Steps: 45910, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001211, Sample Num: 19376, Cur Loss: 1.08972609, Cur Avg Loss: 1.37876250, Log Avg loss: 1.47717151, Global Avg Loss: 1.71778453, Time: 0.0403 Steps: 45920, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001221, Sample Num: 19536, Cur Loss: 0.58388609, Cur Avg Loss: 1.37653421, Log Avg loss: 1.10668789, Global Avg Loss: 1.71765148, Time: 0.0403 Steps: 45930, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001231, Sample Num: 19696, Cur Loss: 1.98013449, Cur Avg Loss: 1.37899098, Log Avg loss: 1.67896195, Global Avg Loss: 1.71764306, Time: 0.0402 Steps: 45940, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001241, Sample Num: 19856, Cur Loss: 0.57682627, Cur Avg Loss: 1.37986872, Log Avg loss: 1.48791918, Global Avg Loss: 1.71759306, Time: 0.0402 Steps: 45950, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001251, Sample Num: 20016, Cur Loss: 1.53016281, Cur Avg Loss: 1.38257701, Log Avg loss: 1.71867625, Global Avg Loss: 1.71759330, Time: 0.0402 Steps: 45960, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001261, Sample Num: 20176, Cur Loss: 1.11351442, Cur Avg Loss: 1.38067785, Log Avg loss: 1.14309293, Global Avg Loss: 1.71746833, Time: 0.0403 Steps: 45970, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001271, Sample Num: 20336, Cur Loss: 1.39562798, Cur Avg Loss: 1.38032532, Log Avg loss: 1.33587078, Global Avg Loss: 1.71738533, Time: 0.0403 Steps: 45980, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001281, Sample Num: 20496, Cur Loss: 0.74102497, Cur Avg Loss: 1.37898582, Log Avg loss: 1.20873517, Global Avg Loss: 1.71727473, Time: 0.0403 Steps: 45990, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001291, Sample Num: 20656, Cur Loss: 1.20464635, Cur Avg Loss: 1.37905426, Log Avg loss: 1.38782089, Global Avg Loss: 1.71720311, Time: 0.0403 Steps: 46000, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001301, Sample Num: 20816, Cur Loss: 1.23857582, Cur Avg Loss: 1.37874767, Log Avg loss: 1.33916754, Global Avg Loss: 1.71712095, Time: 0.0403 Steps: 46010, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001311, Sample Num: 20976, Cur Loss: 1.13162768, Cur Avg Loss: 1.37726447, Log Avg loss: 1.18430001, Global Avg Loss: 1.71700517, Time: 0.0403 Steps: 46020, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001321, Sample Num: 21136, Cur Loss: 2.09619713, Cur Avg Loss: 1.37695094, Log Avg loss: 1.33584715, Global Avg Loss: 1.71692236, Time: 0.0403 Steps: 46030, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001331, Sample Num: 21296, Cur Loss: 1.60139251, Cur Avg Loss: 1.37751648, Log Avg loss: 1.45222445, Global Avg Loss: 1.71686487, Time: 0.0405 Steps: 46040, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001341, Sample Num: 21456, Cur Loss: 1.83444285, Cur Avg Loss: 1.37730998, Log Avg loss: 1.34982505, Global Avg Loss: 1.71678517, Time: 0.0403 Steps: 46050, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001351, Sample Num: 21616, Cur Loss: 3.30526423, Cur Avg Loss: 1.37709081, Log Avg loss: 1.34769967, Global Avg Loss: 1.71670503, Time: 0.0402 Steps: 46060, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001361, Sample Num: 21776, Cur Loss: 0.93413031, Cur Avg Loss: 1.37712777, Log Avg loss: 1.38212166, Global Avg Loss: 1.71663241, Time: 0.0403 Steps: 46070, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001371, Sample Num: 21936, Cur Loss: 1.87387109, Cur Avg Loss: 1.38032902, Log Avg loss: 1.81601881, Global Avg Loss: 1.71665398, Time: 0.0402 Steps: 46080, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001381, Sample Num: 22096, Cur Loss: 1.44265771, Cur Avg Loss: 1.37912794, Log Avg loss: 1.21445986, Global Avg Loss: 1.71654502, Time: 0.0649 Steps: 46090, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001391, Sample Num: 22256, Cur Loss: 2.37617278, Cur Avg Loss: 1.37877691, Log Avg loss: 1.33029968, Global Avg Loss: 1.71646123, Time: 0.0404 Steps: 46100, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001401, Sample Num: 22416, Cur Loss: 2.31609678, Cur Avg Loss: 1.37953535, Log Avg loss: 1.48503481, Global Avg Loss: 1.71641104, Time: 0.0405 Steps: 46110, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001411, Sample Num: 22576, Cur Loss: 1.17375398, Cur Avg Loss: 1.37893100, Log Avg loss: 1.29426137, Global Avg Loss: 1.71631951, Time: 0.0406 Steps: 46120, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001421, Sample Num: 22736, Cur Loss: 1.33143079, Cur Avg Loss: 1.37645135, Log Avg loss: 1.02657307, Global Avg Loss: 1.71616999, Time: 0.0491 Steps: 46130, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001431, Sample Num: 22896, Cur Loss: 0.84536707, Cur Avg Loss: 1.37527708, Log Avg loss: 1.20841232, Global Avg Loss: 1.71605994, Time: 0.0662 Steps: 46140, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001441, Sample Num: 23056, Cur Loss: 1.09350467, Cur Avg Loss: 1.37570984, Log Avg loss: 1.43763776, Global Avg Loss: 1.71599961, Time: 0.0634 Steps: 46150, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001451, Sample Num: 23216, Cur Loss: 2.57132769, Cur Avg Loss: 1.37730503, Log Avg loss: 1.60717175, Global Avg Loss: 1.71597603, Time: 0.0667 Steps: 46160, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001461, Sample Num: 23376, Cur Loss: 1.15869057, Cur Avg Loss: 1.37735777, Log Avg loss: 1.38501044, Global Avg Loss: 1.71590435, Time: 0.0405 Steps: 46170, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001471, Sample Num: 23536, Cur Loss: 0.85596931, Cur Avg Loss: 1.37562351, Log Avg loss: 1.12224851, Global Avg Loss: 1.71577580, Time: 0.0402 Steps: 46180, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001481, Sample Num: 23696, Cur Loss: 2.47575378, Cur Avg Loss: 1.37633882, Log Avg loss: 1.48156083, Global Avg Loss: 1.71572509, Time: 0.0402 Steps: 46190, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001491, Sample Num: 23856, Cur Loss: 1.97015452, Cur Avg Loss: 1.37592158, Log Avg loss: 1.31412908, Global Avg Loss: 1.71563817, Time: 0.0402 Steps: 46200, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001501, Sample Num: 24016, Cur Loss: 1.53367829, Cur Avg Loss: 1.37743029, Log Avg loss: 1.60237899, Global Avg Loss: 1.71561366, Time: 0.0403 Steps: 46210, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001511, Sample Num: 24176, Cur Loss: 0.91096377, Cur Avg Loss: 1.37681768, Log Avg loss: 1.28486485, Global Avg Loss: 1.71552046, Time: 0.0403 Steps: 46220, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001521, Sample Num: 24336, Cur Loss: 0.69624221, Cur Avg Loss: 1.37675064, Log Avg loss: 1.36662043, Global Avg Loss: 1.71544499, Time: 0.0403 Steps: 46230, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001531, Sample Num: 24496, Cur Loss: 1.38232744, Cur Avg Loss: 1.37691578, Log Avg loss: 1.40203298, Global Avg Loss: 1.71537721, Time: 0.0402 Steps: 46240, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001541, Sample Num: 24656, Cur Loss: 0.83570790, Cur Avg Loss: 1.37662452, Log Avg loss: 1.33203261, Global Avg Loss: 1.71529433, Time: 0.0403 Steps: 46250, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001551, Sample Num: 24816, Cur Loss: 0.83711362, Cur Avg Loss: 1.37438357, Log Avg loss: 1.02905298, Global Avg Loss: 1.71514598, Time: 0.0403 Steps: 46260, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001561, Sample Num: 24976, Cur Loss: 2.41637039, Cur Avg Loss: 1.37424330, Log Avg loss: 1.35248864, Global Avg Loss: 1.71506760, Time: 0.0403 Steps: 46270, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001571, Sample Num: 25136, Cur Loss: 1.94287217, Cur Avg Loss: 1.37459104, Log Avg loss: 1.42887210, Global Avg Loss: 1.71500576, Time: 0.0403 Steps: 46280, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001581, Sample Num: 25296, Cur Loss: 2.07949090, Cur Avg Loss: 1.37563778, Log Avg loss: 1.54008140, Global Avg Loss: 1.71496797, Time: 0.0402 Steps: 46290, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001591, Sample Num: 25456, Cur Loss: 1.41830432, Cur Avg Loss: 1.37561162, Log Avg loss: 1.37147591, Global Avg Loss: 1.71489379, Time: 0.0403 Steps: 46300, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001601, Sample Num: 25616, Cur Loss: 1.23070335, Cur Avg Loss: 1.37754459, Log Avg loss: 1.68508049, Global Avg Loss: 1.71488735, Time: 0.0403 Steps: 46310, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001611, Sample Num: 25776, Cur Loss: 1.21769285, Cur Avg Loss: 1.37772341, Log Avg loss: 1.40635105, Global Avg Loss: 1.71482074, Time: 0.0403 Steps: 46320, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001621, Sample Num: 25936, Cur Loss: 1.29845726, Cur Avg Loss: 1.37771795, Log Avg loss: 1.37683985, Global Avg Loss: 1.71474779, Time: 0.0403 Steps: 46330, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001631, Sample Num: 26096, Cur Loss: 2.92770982, Cur Avg Loss: 1.37786025, Log Avg loss: 1.40092571, Global Avg Loss: 1.71468007, Time: 0.0402 Steps: 46340, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001641, Sample Num: 26256, Cur Loss: 1.20724285, Cur Avg Loss: 1.37816396, Log Avg loss: 1.42770014, Global Avg Loss: 1.71461815, Time: 0.0403 Steps: 46350, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001651, Sample Num: 26416, Cur Loss: 0.72991014, Cur Avg Loss: 1.37650852, Log Avg loss: 1.10484977, Global Avg Loss: 1.71448662, Time: 0.0403 Steps: 46360, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001661, Sample Num: 26576, Cur Loss: 1.50718963, Cur Avg Loss: 1.37505093, Log Avg loss: 1.13440320, Global Avg Loss: 1.71436152, Time: 0.0403 Steps: 46370, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001671, Sample Num: 26736, Cur Loss: 1.41542339, Cur Avg Loss: 1.37432964, Log Avg loss: 1.25452417, Global Avg Loss: 1.71426238, Time: 0.0402 Steps: 46380, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001681, Sample Num: 26896, Cur Loss: 0.69442284, Cur Avg Loss: 1.37445939, Log Avg loss: 1.39613971, Global Avg Loss: 1.71419380, Time: 0.0403 Steps: 46390, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001691, Sample Num: 27056, Cur Loss: 3.00104928, Cur Avg Loss: 1.37573664, Log Avg loss: 1.59044287, Global Avg Loss: 1.71416713, Time: 0.0402 Steps: 46400, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001701, Sample Num: 27216, Cur Loss: 1.40646791, Cur Avg Loss: 1.37796801, Log Avg loss: 1.75529292, Global Avg Loss: 1.71417599, Time: 0.0404 Steps: 46410, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001711, Sample Num: 27376, Cur Loss: 0.33161920, Cur Avg Loss: 1.37772834, Log Avg loss: 1.33695962, Global Avg Loss: 1.71409473, Time: 0.0402 Steps: 46420, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001721, Sample Num: 27536, Cur Loss: 1.29906547, Cur Avg Loss: 1.37895336, Log Avg loss: 1.58855426, Global Avg Loss: 1.71406769, Time: 0.0402 Steps: 46430, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001731, Sample Num: 27696, Cur Loss: 0.73733151, Cur Avg Loss: 1.37963408, Log Avg loss: 1.49678651, Global Avg Loss: 1.71402090, Time: 0.0407 Steps: 46440, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001741, Sample Num: 27856, Cur Loss: 0.84211695, Cur Avg Loss: 1.37902967, Log Avg loss: 1.27440566, Global Avg Loss: 1.71392626, Time: 0.0880 Steps: 46450, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001751, Sample Num: 28016, Cur Loss: 1.29649413, Cur Avg Loss: 1.37936587, Log Avg loss: 1.43789815, Global Avg Loss: 1.71386685, Time: 0.0616 Steps: 46460, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001761, Sample Num: 28176, Cur Loss: 1.81811202, Cur Avg Loss: 1.38042191, Log Avg loss: 1.56533503, Global Avg Loss: 1.71383489, Time: 0.0629 Steps: 46470, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001771, Sample Num: 28336, Cur Loss: 0.91910493, Cur Avg Loss: 1.37901487, Log Avg loss: 1.13123554, Global Avg Loss: 1.71370954, Time: 0.0506 Steps: 46480, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001781, Sample Num: 28496, Cur Loss: 1.25735259, Cur Avg Loss: 1.37826802, Log Avg loss: 1.24600081, Global Avg Loss: 1.71360894, Time: 0.0404 Steps: 46490, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001791, Sample Num: 28656, Cur Loss: 1.49919236, Cur Avg Loss: 1.37688247, Log Avg loss: 1.13011612, Global Avg Loss: 1.71348346, Time: 0.0524 Steps: 46500, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001801, Sample Num: 28816, Cur Loss: 1.06100440, Cur Avg Loss: 1.37661876, Log Avg loss: 1.32938776, Global Avg Loss: 1.71340087, Time: 0.0530 Steps: 46510, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001811, Sample Num: 28976, Cur Loss: 1.17112768, Cur Avg Loss: 1.37770634, Log Avg loss: 1.57357958, Global Avg Loss: 1.71337082, Time: 0.1014 Steps: 46520, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001821, Sample Num: 29136, Cur Loss: 1.26940644, Cur Avg Loss: 1.37739821, Log Avg loss: 1.32159618, Global Avg Loss: 1.71328662, Time: 0.0402 Steps: 46530, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001831, Sample Num: 29296, Cur Loss: 2.85070801, Cur Avg Loss: 1.37776196, Log Avg loss: 1.44400042, Global Avg Loss: 1.71322876, Time: 0.0402 Steps: 46540, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001841, Sample Num: 29456, Cur Loss: 2.15731168, Cur Avg Loss: 1.37986188, Log Avg loss: 1.76435689, Global Avg Loss: 1.71323974, Time: 0.0402 Steps: 46550, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001851, Sample Num: 29616, Cur Loss: 3.42960596, Cur Avg Loss: 1.38058769, Log Avg loss: 1.51421001, Global Avg Loss: 1.71319699, Time: 0.0402 Steps: 46560, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001861, Sample Num: 29776, Cur Loss: 2.45327139, Cur Avg Loss: 1.38285427, Log Avg loss: 1.80239840, Global Avg Loss: 1.71321615, Time: 0.0404 Steps: 46570, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001871, Sample Num: 29936, Cur Loss: 1.87060976, Cur Avg Loss: 1.38304789, Log Avg loss: 1.41908048, Global Avg Loss: 1.71315300, Time: 0.0402 Steps: 46580, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001881, Sample Num: 30096, Cur Loss: 1.23658872, Cur Avg Loss: 1.38154271, Log Avg loss: 1.09992350, Global Avg Loss: 1.71302138, Time: 0.0404 Steps: 46590, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001891, Sample Num: 30256, Cur Loss: 0.36584401, Cur Avg Loss: 1.38109825, Log Avg loss: 1.29749509, Global Avg Loss: 1.71293221, Time: 0.0402 Steps: 46600, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001901, Sample Num: 30416, Cur Loss: 1.07976902, Cur Avg Loss: 1.38144024, Log Avg loss: 1.44611087, Global Avg Loss: 1.71287496, Time: 0.0403 Steps: 46610, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001911, Sample Num: 30576, Cur Loss: 0.78117406, Cur Avg Loss: 1.38327867, Log Avg loss: 1.73276405, Global Avg Loss: 1.71287923, Time: 0.0402 Steps: 46620, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001921, Sample Num: 30736, Cur Loss: 1.99709082, Cur Avg Loss: 1.38406037, Log Avg loss: 1.53344389, Global Avg Loss: 1.71284075, Time: 0.0402 Steps: 46630, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001931, Sample Num: 30896, Cur Loss: 3.03486371, Cur Avg Loss: 1.38468508, Log Avg loss: 1.50469169, Global Avg Loss: 1.71279612, Time: 0.0402 Steps: 46640, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001941, Sample Num: 31056, Cur Loss: 2.56622934, Cur Avg Loss: 1.38733842, Log Avg loss: 1.89969831, Global Avg Loss: 1.71283619, Time: 0.0402 Steps: 46650, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001951, Sample Num: 31216, Cur Loss: 1.25405347, Cur Avg Loss: 1.38887727, Log Avg loss: 1.68756794, Global Avg Loss: 1.71283077, Time: 0.0403 Steps: 46660, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001961, Sample Num: 31376, Cur Loss: 0.72800988, Cur Avg Loss: 1.38925308, Log Avg loss: 1.46257278, Global Avg Loss: 1.71277715, Time: 0.0403 Steps: 46670, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001971, Sample Num: 31536, Cur Loss: 1.42314065, Cur Avg Loss: 1.39015179, Log Avg loss: 1.56638947, Global Avg Loss: 1.71274579, Time: 0.0403 Steps: 46680, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001981, Sample Num: 31696, Cur Loss: 1.25135887, Cur Avg Loss: 1.39101483, Log Avg loss: 1.56111927, Global Avg Loss: 1.71271331, Time: 0.0403 Steps: 46690, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001991, Sample Num: 31856, Cur Loss: 1.53616798, Cur Avg Loss: 1.39047248, Log Avg loss: 1.28303374, Global Avg Loss: 1.71262130, Time: 0.0402 Steps: 46700, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002001, Sample Num: 32016, Cur Loss: 0.63124341, Cur Avg Loss: 1.39040144, Log Avg loss: 1.37625796, Global Avg Loss: 1.71254929, Time: 0.0402 Steps: 46710, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002011, Sample Num: 32176, Cur Loss: 2.01713967, Cur Avg Loss: 1.39063086, Log Avg loss: 1.43653733, Global Avg Loss: 1.71249021, Time: 0.0402 Steps: 46720, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002021, Sample Num: 32336, Cur Loss: 0.46386734, Cur Avg Loss: 1.38936391, Log Avg loss: 1.13457986, Global Avg Loss: 1.71236654, Time: 0.0403 Steps: 46730, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002031, Sample Num: 32496, Cur Loss: 1.39486504, Cur Avg Loss: 1.38981584, Log Avg loss: 1.48115013, Global Avg Loss: 1.71231708, Time: 0.0403 Steps: 46740, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002041, Sample Num: 32656, Cur Loss: 3.77547979, Cur Avg Loss: 1.38997111, Log Avg loss: 1.42150780, Global Avg Loss: 1.71225487, Time: 0.0403 Steps: 46750, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002051, Sample Num: 32816, Cur Loss: 1.98890948, Cur Avg Loss: 1.39067381, Log Avg loss: 1.53409526, Global Avg Loss: 1.71221677, Time: 0.0402 Steps: 46760, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002061, Sample Num: 32976, Cur Loss: 0.80448097, Cur Avg Loss: 1.38858954, Log Avg loss: 0.96110571, Global Avg Loss: 1.71205617, Time: 0.0402 Steps: 46770, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002071, Sample Num: 33136, Cur Loss: 3.38507366, Cur Avg Loss: 1.38960153, Log Avg loss: 1.59817116, Global Avg Loss: 1.71203183, Time: 0.0403 Steps: 46780, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002081, Sample Num: 33296, Cur Loss: 0.99132216, Cur Avg Loss: 1.39040602, Log Avg loss: 1.55701740, Global Avg Loss: 1.71199870, Time: 0.0402 Steps: 46790, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002091, Sample Num: 33456, Cur Loss: 0.55961013, Cur Avg Loss: 1.39061785, Log Avg loss: 1.43469831, Global Avg Loss: 1.71193945, Time: 0.0640 Steps: 46800, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002101, Sample Num: 33616, Cur Loss: 0.89178634, Cur Avg Loss: 1.39082695, Log Avg loss: 1.43454972, Global Avg Loss: 1.71188019, Time: 0.0656 Steps: 46810, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002111, Sample Num: 33776, Cur Loss: 1.58605981, Cur Avg Loss: 1.39097948, Log Avg loss: 1.42302722, Global Avg Loss: 1.71181849, Time: 0.0562 Steps: 46820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002121, Sample Num: 33936, Cur Loss: 1.61198926, Cur Avg Loss: 1.39189930, Log Avg loss: 1.58607168, Global Avg Loss: 1.71179164, Time: 0.0455 Steps: 46830, Updated lr: 0.000057 ***** Running evaluation checkpoint-46838 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-46838 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 96.310151, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.25592, "eval_total_loss": 882.912078, "eval_mae": 0.915655, "eval_mse": 1.256013, "eval_r2": 0.201596, "eval_sp_statistic": 0.34768, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.454833, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.213085, "test_total_loss": 608.968794, "test_mae": 0.874231, "test_mse": 1.213331, "test_r2": 0.216905, "test_sp_statistic": 0.364861, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.487671, "test_ps_pvalue": 0.0, "lr": 5.653105737316264e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7116897925788603, "train_cur_epoch_loss": 2961.142330303788, "train_cur_epoch_avg_loss": 1.3908606530313707, "train_cur_epoch_time": 96.31015062332153, "train_cur_epoch_avg_time": 0.04523727131203454, "epoch": 22, "step": 46838} ################################################## Training, Epoch: 0023, Batch: 000002, Sample Num: 32, Cur Loss: 2.28335142, Cur Avg Loss: 1.39190042, Log Avg loss: 1.17077262, Global Avg Loss: 1.71167614, Time: 0.0519 Steps: 46840, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000012, Sample Num: 192, Cur Loss: 0.40532625, Cur Avg Loss: 1.28525314, Log Avg loss: 1.26392369, Global Avg Loss: 1.71158057, Time: 0.0637 Steps: 46850, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000022, Sample Num: 352, Cur Loss: 0.73472506, Cur Avg Loss: 1.19361210, Log Avg loss: 1.08364285, Global Avg Loss: 1.71144656, Time: 0.0490 Steps: 46860, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000032, Sample Num: 512, Cur Loss: 1.58073997, Cur Avg Loss: 1.26183579, Log Avg loss: 1.41192790, Global Avg Loss: 1.71138266, Time: 0.0606 Steps: 46870, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000042, Sample Num: 672, Cur Loss: 1.12739468, Cur Avg Loss: 1.28430260, Log Avg loss: 1.35619640, Global Avg Loss: 1.71130689, Time: 0.0432 Steps: 46880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000052, Sample Num: 832, Cur Loss: 1.74996066, Cur Avg Loss: 1.32549695, Log Avg loss: 1.49851323, Global Avg Loss: 1.71126151, Time: 0.0405 Steps: 46890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000062, Sample Num: 992, Cur Loss: 3.22198367, Cur Avg Loss: 1.36739785, Log Avg loss: 1.58528252, Global Avg Loss: 1.71123465, Time: 0.0407 Steps: 46900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000072, Sample Num: 1152, Cur Loss: 1.27771020, Cur Avg Loss: 1.42843779, Log Avg loss: 1.80688543, Global Avg Loss: 1.71125504, Time: 0.0539 Steps: 46910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000082, Sample Num: 1312, Cur Loss: 0.71671838, Cur Avg Loss: 1.41915861, Log Avg loss: 1.35234852, Global Avg Loss: 1.71117855, Time: 0.0404 Steps: 46920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000092, Sample Num: 1472, Cur Loss: 1.07099283, Cur Avg Loss: 1.41725883, Log Avg loss: 1.40168062, Global Avg Loss: 1.71111260, Time: 0.0403 Steps: 46930, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000102, Sample Num: 1632, Cur Loss: 2.28132749, Cur Avg Loss: 1.41819623, Log Avg loss: 1.42682027, Global Avg Loss: 1.71105203, Time: 0.0404 Steps: 46940, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000112, Sample Num: 1792, Cur Loss: 1.12654662, Cur Avg Loss: 1.41644200, Log Avg loss: 1.39854887, Global Avg Loss: 1.71098547, Time: 0.0404 Steps: 46950, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000122, Sample Num: 1952, Cur Loss: 0.69049680, Cur Avg Loss: 1.41663337, Log Avg loss: 1.41877669, Global Avg Loss: 1.71092325, Time: 0.0403 Steps: 46960, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000132, Sample Num: 2112, Cur Loss: 1.50155914, Cur Avg Loss: 1.40112750, Log Avg loss: 1.21195591, Global Avg Loss: 1.71081702, Time: 0.0403 Steps: 46970, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000142, Sample Num: 2272, Cur Loss: 1.65555739, Cur Avg Loss: 1.40409517, Log Avg loss: 1.44326837, Global Avg Loss: 1.71076007, Time: 0.0403 Steps: 46980, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000152, Sample Num: 2432, Cur Loss: 1.72873628, Cur Avg Loss: 1.39602084, Log Avg loss: 1.28136545, Global Avg Loss: 1.71066869, Time: 0.0403 Steps: 46990, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000162, Sample Num: 2592, Cur Loss: 1.24708235, Cur Avg Loss: 1.40049624, Log Avg loss: 1.46852221, Global Avg Loss: 1.71061717, Time: 0.0403 Steps: 47000, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000172, Sample Num: 2752, Cur Loss: 1.74776101, Cur Avg Loss: 1.40159681, Log Avg loss: 1.41942616, Global Avg Loss: 1.71055523, Time: 0.0402 Steps: 47010, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000182, Sample Num: 2912, Cur Loss: 1.04027271, Cur Avg Loss: 1.40772848, Log Avg loss: 1.51319323, Global Avg Loss: 1.71051325, Time: 0.0403 Steps: 47020, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000192, Sample Num: 3072, Cur Loss: 1.88366699, Cur Avg Loss: 1.40831233, Log Avg loss: 1.41893835, Global Avg Loss: 1.71045125, Time: 0.0403 Steps: 47030, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000202, Sample Num: 3232, Cur Loss: 0.55666250, Cur Avg Loss: 1.39670620, Log Avg loss: 1.17386853, Global Avg Loss: 1.71033718, Time: 0.0403 Steps: 47040, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000212, Sample Num: 3392, Cur Loss: 1.52039504, Cur Avg Loss: 1.40263692, Log Avg loss: 1.52243752, Global Avg Loss: 1.71029725, Time: 0.0402 Steps: 47050, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000222, Sample Num: 3552, Cur Loss: 1.81158853, Cur Avg Loss: 1.39791306, Log Avg loss: 1.29776708, Global Avg Loss: 1.71020959, Time: 0.0402 Steps: 47060, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000232, Sample Num: 3712, Cur Loss: 1.98221350, Cur Avg Loss: 1.39968351, Log Avg loss: 1.43898754, Global Avg Loss: 1.71015197, Time: 0.0402 Steps: 47070, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000242, Sample Num: 3872, Cur Loss: 1.82824051, Cur Avg Loss: 1.40588273, Log Avg loss: 1.54970473, Global Avg Loss: 1.71011789, Time: 0.0403 Steps: 47080, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000252, Sample Num: 4032, Cur Loss: 2.65545678, Cur Avg Loss: 1.40521266, Log Avg loss: 1.38899685, Global Avg Loss: 1.71004969, Time: 0.0403 Steps: 47090, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000262, Sample Num: 4192, Cur Loss: 0.82699525, Cur Avg Loss: 1.39927550, Log Avg loss: 1.24965913, Global Avg Loss: 1.70995195, Time: 0.0402 Steps: 47100, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000272, Sample Num: 4352, Cur Loss: 2.59457588, Cur Avg Loss: 1.40704439, Log Avg loss: 1.61058934, Global Avg Loss: 1.70993086, Time: 0.0402 Steps: 47110, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000282, Sample Num: 4512, Cur Loss: 1.30877733, Cur Avg Loss: 1.42108155, Log Avg loss: 1.80289212, Global Avg Loss: 1.70995058, Time: 0.0402 Steps: 47120, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000292, Sample Num: 4672, Cur Loss: 1.04495406, Cur Avg Loss: 1.41564850, Log Avg loss: 1.26243667, Global Avg Loss: 1.70985563, Time: 0.0402 Steps: 47130, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000302, Sample Num: 4832, Cur Loss: 2.02360415, Cur Avg Loss: 1.41657228, Log Avg loss: 1.44354650, Global Avg Loss: 1.70979914, Time: 0.0402 Steps: 47140, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000312, Sample Num: 4992, Cur Loss: 0.71526432, Cur Avg Loss: 1.41413156, Log Avg loss: 1.34042200, Global Avg Loss: 1.70972080, Time: 0.0402 Steps: 47150, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000322, Sample Num: 5152, Cur Loss: 0.84382069, Cur Avg Loss: 1.41107173, Log Avg loss: 1.31560481, Global Avg Loss: 1.70963723, Time: 0.0402 Steps: 47160, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000332, Sample Num: 5312, Cur Loss: 2.74033308, Cur Avg Loss: 1.41776211, Log Avg loss: 1.63319233, Global Avg Loss: 1.70962102, Time: 0.0402 Steps: 47170, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000342, Sample Num: 5472, Cur Loss: 0.60131466, Cur Avg Loss: 1.41517174, Log Avg loss: 1.32917168, Global Avg Loss: 1.70954038, Time: 0.0627 Steps: 47180, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000352, Sample Num: 5632, Cur Loss: 1.41927981, Cur Avg Loss: 1.41494445, Log Avg loss: 1.40717104, Global Avg Loss: 1.70947631, Time: 0.0561 Steps: 47190, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000362, Sample Num: 5792, Cur Loss: 1.99239218, Cur Avg Loss: 1.41130855, Log Avg loss: 1.28332483, Global Avg Loss: 1.70938602, Time: 0.0875 Steps: 47200, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000372, Sample Num: 5952, Cur Loss: 0.52683699, Cur Avg Loss: 1.40074658, Log Avg loss: 1.01840317, Global Avg Loss: 1.70923966, Time: 0.0566 Steps: 47210, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000382, Sample Num: 6112, Cur Loss: 0.78898275, Cur Avg Loss: 1.39962211, Log Avg loss: 1.35779182, Global Avg Loss: 1.70916523, Time: 0.0426 Steps: 47220, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000392, Sample Num: 6272, Cur Loss: 1.69936275, Cur Avg Loss: 1.40186496, Log Avg loss: 1.48754213, Global Avg Loss: 1.70911831, Time: 0.0406 Steps: 47230, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000402, Sample Num: 6432, Cur Loss: 1.12404919, Cur Avg Loss: 1.40221550, Log Avg loss: 1.41595670, Global Avg Loss: 1.70905625, Time: 0.0690 Steps: 47240, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000412, Sample Num: 6592, Cur Loss: 1.16577339, Cur Avg Loss: 1.39827031, Log Avg loss: 1.23967366, Global Avg Loss: 1.70895691, Time: 0.0665 Steps: 47250, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000422, Sample Num: 6752, Cur Loss: 0.56192422, Cur Avg Loss: 1.39471449, Log Avg loss: 1.24821461, Global Avg Loss: 1.70885942, Time: 0.0675 Steps: 47260, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000432, Sample Num: 6912, Cur Loss: 1.11083341, Cur Avg Loss: 1.38675475, Log Avg loss: 1.05085364, Global Avg Loss: 1.70872021, Time: 0.0402 Steps: 47270, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000442, Sample Num: 7072, Cur Loss: 1.68376136, Cur Avg Loss: 1.38673994, Log Avg loss: 1.38610030, Global Avg Loss: 1.70865198, Time: 0.0403 Steps: 47280, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000452, Sample Num: 7232, Cur Loss: 0.55214226, Cur Avg Loss: 1.37967229, Log Avg loss: 1.06728199, Global Avg Loss: 1.70851635, Time: 0.0403 Steps: 47290, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000462, Sample Num: 7392, Cur Loss: 2.38412094, Cur Avg Loss: 1.38349875, Log Avg loss: 1.55645474, Global Avg Loss: 1.70848421, Time: 0.0402 Steps: 47300, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000472, Sample Num: 7552, Cur Loss: 0.58842421, Cur Avg Loss: 1.37784427, Log Avg loss: 1.11660721, Global Avg Loss: 1.70835910, Time: 0.0403 Steps: 47310, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000482, Sample Num: 7712, Cur Loss: 0.83276707, Cur Avg Loss: 1.37604435, Log Avg loss: 1.29108845, Global Avg Loss: 1.70827092, Time: 0.0402 Steps: 47320, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000492, Sample Num: 7872, Cur Loss: 1.87101531, Cur Avg Loss: 1.37778242, Log Avg loss: 1.46155711, Global Avg Loss: 1.70821879, Time: 0.0403 Steps: 47330, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000502, Sample Num: 8032, Cur Loss: 1.00688243, Cur Avg Loss: 1.37199603, Log Avg loss: 1.08730553, Global Avg Loss: 1.70808763, Time: 0.0402 Steps: 47340, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000512, Sample Num: 8192, Cur Loss: 1.96985614, Cur Avg Loss: 1.37313853, Log Avg loss: 1.43049211, Global Avg Loss: 1.70802901, Time: 0.0404 Steps: 47350, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000522, Sample Num: 8352, Cur Loss: 2.49463320, Cur Avg Loss: 1.37304526, Log Avg loss: 1.36826969, Global Avg Loss: 1.70795727, Time: 0.0402 Steps: 47360, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000532, Sample Num: 8512, Cur Loss: 1.15470099, Cur Avg Loss: 1.36885951, Log Avg loss: 1.15036366, Global Avg Loss: 1.70783956, Time: 0.0403 Steps: 47370, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000542, Sample Num: 8672, Cur Loss: 0.86428893, Cur Avg Loss: 1.36877169, Log Avg loss: 1.36409966, Global Avg Loss: 1.70776701, Time: 0.0402 Steps: 47380, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000552, Sample Num: 8832, Cur Loss: 1.03964067, Cur Avg Loss: 1.36733079, Log Avg loss: 1.28923381, Global Avg Loss: 1.70767869, Time: 0.0402 Steps: 47390, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000562, Sample Num: 8992, Cur Loss: 0.81654787, Cur Avg Loss: 1.36852026, Log Avg loss: 1.43417896, Global Avg Loss: 1.70762099, Time: 0.0403 Steps: 47400, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000572, Sample Num: 9152, Cur Loss: 0.58867276, Cur Avg Loss: 1.36403362, Log Avg loss: 1.11188459, Global Avg Loss: 1.70749533, Time: 0.0402 Steps: 47410, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000582, Sample Num: 9312, Cur Loss: 1.12465000, Cur Avg Loss: 1.36558239, Log Avg loss: 1.45417188, Global Avg Loss: 1.70744191, Time: 0.0402 Steps: 47420, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000592, Sample Num: 9472, Cur Loss: 0.85413218, Cur Avg Loss: 1.36631244, Log Avg loss: 1.40880166, Global Avg Loss: 1.70737895, Time: 0.0403 Steps: 47430, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000602, Sample Num: 9632, Cur Loss: 0.64107203, Cur Avg Loss: 1.36187122, Log Avg loss: 1.09895063, Global Avg Loss: 1.70725070, Time: 0.0403 Steps: 47440, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000612, Sample Num: 9792, Cur Loss: 1.21779764, Cur Avg Loss: 1.36109071, Log Avg loss: 1.31410431, Global Avg Loss: 1.70716784, Time: 0.0402 Steps: 47450, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000622, Sample Num: 9952, Cur Loss: 1.69304717, Cur Avg Loss: 1.35697060, Log Avg loss: 1.10481980, Global Avg Loss: 1.70704092, Time: 0.0402 Steps: 47460, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000632, Sample Num: 10112, Cur Loss: 0.95484334, Cur Avg Loss: 1.35809228, Log Avg loss: 1.42786103, Global Avg Loss: 1.70698211, Time: 0.0402 Steps: 47470, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000642, Sample Num: 10272, Cur Loss: 2.38445044, Cur Avg Loss: 1.35893152, Log Avg loss: 1.41197154, Global Avg Loss: 1.70691998, Time: 0.0402 Steps: 47480, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000652, Sample Num: 10432, Cur Loss: 1.77381587, Cur Avg Loss: 1.35951105, Log Avg loss: 1.39671654, Global Avg Loss: 1.70685466, Time: 0.0402 Steps: 47490, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000662, Sample Num: 10592, Cur Loss: 0.86761034, Cur Avg Loss: 1.35601840, Log Avg loss: 1.12829773, Global Avg Loss: 1.70673286, Time: 0.0402 Steps: 47500, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000672, Sample Num: 10752, Cur Loss: 1.28064394, Cur Avg Loss: 1.35859656, Log Avg loss: 1.52927055, Global Avg Loss: 1.70669550, Time: 0.0402 Steps: 47510, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000682, Sample Num: 10912, Cur Loss: 1.09883535, Cur Avg Loss: 1.35644564, Log Avg loss: 1.21190428, Global Avg Loss: 1.70659138, Time: 0.0402 Steps: 47520, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000692, Sample Num: 11072, Cur Loss: 1.52870536, Cur Avg Loss: 1.35831916, Log Avg loss: 1.48609285, Global Avg Loss: 1.70654499, Time: 0.0617 Steps: 47530, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000702, Sample Num: 11232, Cur Loss: 0.95916289, Cur Avg Loss: 1.35585572, Log Avg loss: 1.18538558, Global Avg Loss: 1.70643536, Time: 0.0821 Steps: 47540, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000712, Sample Num: 11392, Cur Loss: 0.49275693, Cur Avg Loss: 1.35543915, Log Avg loss: 1.32619641, Global Avg Loss: 1.70635540, Time: 0.0406 Steps: 47550, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000722, Sample Num: 11552, Cur Loss: 0.87881374, Cur Avg Loss: 1.35178706, Log Avg loss: 1.09175805, Global Avg Loss: 1.70622617, Time: 0.0568 Steps: 47560, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000732, Sample Num: 11712, Cur Loss: 0.68639654, Cur Avg Loss: 1.35110011, Log Avg loss: 1.30150194, Global Avg Loss: 1.70614109, Time: 0.0467 Steps: 47570, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000742, Sample Num: 11872, Cur Loss: 0.97649348, Cur Avg Loss: 1.35066138, Log Avg loss: 1.31854653, Global Avg Loss: 1.70605963, Time: 0.0612 Steps: 47580, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000752, Sample Num: 12032, Cur Loss: 1.06497097, Cur Avg Loss: 1.34935641, Log Avg loss: 1.25252767, Global Avg Loss: 1.70596433, Time: 0.0449 Steps: 47590, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000762, Sample Num: 12192, Cur Loss: 0.73819268, Cur Avg Loss: 1.34536225, Log Avg loss: 1.04500160, Global Avg Loss: 1.70582547, Time: 0.0618 Steps: 47600, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000772, Sample Num: 12352, Cur Loss: 1.53456330, Cur Avg Loss: 1.34521322, Log Avg loss: 1.33385700, Global Avg Loss: 1.70574735, Time: 0.0404 Steps: 47610, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000782, Sample Num: 12512, Cur Loss: 1.77191615, Cur Avg Loss: 1.34901235, Log Avg loss: 1.64230499, Global Avg Loss: 1.70573402, Time: 0.0403 Steps: 47620, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000792, Sample Num: 12672, Cur Loss: 1.32972288, Cur Avg Loss: 1.35018627, Log Avg loss: 1.44198726, Global Avg Loss: 1.70567865, Time: 0.0403 Steps: 47630, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000802, Sample Num: 12832, Cur Loss: 0.84826154, Cur Avg Loss: 1.35216836, Log Avg loss: 1.50914977, Global Avg Loss: 1.70563740, Time: 0.0402 Steps: 47640, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000812, Sample Num: 12992, Cur Loss: 2.35491037, Cur Avg Loss: 1.35209244, Log Avg loss: 1.34600332, Global Avg Loss: 1.70556192, Time: 0.0402 Steps: 47650, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000822, Sample Num: 13152, Cur Loss: 1.57757437, Cur Avg Loss: 1.35454303, Log Avg loss: 1.55353143, Global Avg Loss: 1.70553002, Time: 0.0402 Steps: 47660, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000832, Sample Num: 13312, Cur Loss: 1.34826159, Cur Avg Loss: 1.35124526, Log Avg loss: 1.08016841, Global Avg Loss: 1.70539884, Time: 0.0402 Steps: 47670, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000842, Sample Num: 13472, Cur Loss: 2.00588703, Cur Avg Loss: 1.34982194, Log Avg loss: 1.23140166, Global Avg Loss: 1.70529942, Time: 0.0403 Steps: 47680, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000852, Sample Num: 13632, Cur Loss: 0.36629954, Cur Avg Loss: 1.34653887, Log Avg loss: 1.07010396, Global Avg Loss: 1.70516623, Time: 0.0403 Steps: 47690, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000862, Sample Num: 13792, Cur Loss: 1.09770072, Cur Avg Loss: 1.34753932, Log Avg loss: 1.43277780, Global Avg Loss: 1.70510913, Time: 0.0402 Steps: 47700, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000872, Sample Num: 13952, Cur Loss: 1.56711245, Cur Avg Loss: 1.34543016, Log Avg loss: 1.16362060, Global Avg Loss: 1.70499563, Time: 0.0403 Steps: 47710, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000882, Sample Num: 14112, Cur Loss: 1.34450054, Cur Avg Loss: 1.34784532, Log Avg loss: 1.55844705, Global Avg Loss: 1.70496492, Time: 0.0402 Steps: 47720, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000892, Sample Num: 14272, Cur Loss: 2.32289791, Cur Avg Loss: 1.34974096, Log Avg loss: 1.51693644, Global Avg Loss: 1.70492553, Time: 0.0403 Steps: 47730, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000902, Sample Num: 14432, Cur Loss: 0.87665009, Cur Avg Loss: 1.35052682, Log Avg loss: 1.42062601, Global Avg Loss: 1.70486598, Time: 0.0403 Steps: 47740, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000912, Sample Num: 14592, Cur Loss: 0.86575377, Cur Avg Loss: 1.34825058, Log Avg loss: 1.14293340, Global Avg Loss: 1.70474829, Time: 0.0402 Steps: 47750, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000922, Sample Num: 14752, Cur Loss: 1.30686212, Cur Avg Loss: 1.34929748, Log Avg loss: 1.44477523, Global Avg Loss: 1.70469386, Time: 0.0403 Steps: 47760, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000932, Sample Num: 14912, Cur Loss: 1.38567829, Cur Avg Loss: 1.34891409, Log Avg loss: 1.31356500, Global Avg Loss: 1.70461198, Time: 0.0402 Steps: 47770, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000942, Sample Num: 15072, Cur Loss: 1.44544220, Cur Avg Loss: 1.35066633, Log Avg loss: 1.51397513, Global Avg Loss: 1.70457208, Time: 0.0403 Steps: 47780, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000952, Sample Num: 15232, Cur Loss: 0.63249689, Cur Avg Loss: 1.34986248, Log Avg loss: 1.27413992, Global Avg Loss: 1.70448202, Time: 0.0402 Steps: 47790, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000962, Sample Num: 15392, Cur Loss: 1.61472058, Cur Avg Loss: 1.35128062, Log Avg loss: 1.48628744, Global Avg Loss: 1.70443637, Time: 0.0403 Steps: 47800, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000972, Sample Num: 15552, Cur Loss: 2.18245602, Cur Avg Loss: 1.35461341, Log Avg loss: 1.67522848, Global Avg Loss: 1.70443026, Time: 0.0402 Steps: 47810, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000982, Sample Num: 15712, Cur Loss: 1.12152719, Cur Avg Loss: 1.35636940, Log Avg loss: 1.52705116, Global Avg Loss: 1.70439317, Time: 0.0402 Steps: 47820, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000992, Sample Num: 15872, Cur Loss: 1.57479763, Cur Avg Loss: 1.35455643, Log Avg loss: 1.17652304, Global Avg Loss: 1.70428280, Time: 0.0402 Steps: 47830, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001002, Sample Num: 16032, Cur Loss: 0.98283124, Cur Avg Loss: 1.35522286, Log Avg loss: 1.42133236, Global Avg Loss: 1.70422366, Time: 0.0402 Steps: 47840, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001012, Sample Num: 16192, Cur Loss: 1.35201907, Cur Avg Loss: 1.35346352, Log Avg loss: 1.17717750, Global Avg Loss: 1.70411351, Time: 0.0403 Steps: 47850, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001022, Sample Num: 16352, Cur Loss: 1.75938272, Cur Avg Loss: 1.35194713, Log Avg loss: 1.19848905, Global Avg Loss: 1.70400787, Time: 0.0403 Steps: 47860, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001032, Sample Num: 16512, Cur Loss: 2.72535348, Cur Avg Loss: 1.35374645, Log Avg loss: 1.53763695, Global Avg Loss: 1.70397311, Time: 0.0402 Steps: 47870, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001042, Sample Num: 16672, Cur Loss: 1.15050423, Cur Avg Loss: 1.35249301, Log Avg loss: 1.22313731, Global Avg Loss: 1.70387269, Time: 0.0849 Steps: 47880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001052, Sample Num: 16832, Cur Loss: 1.00412416, Cur Avg Loss: 1.35140331, Log Avg loss: 1.23785732, Global Avg Loss: 1.70377538, Time: 0.0524 Steps: 47890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001062, Sample Num: 16992, Cur Loss: 1.26296473, Cur Avg Loss: 1.35339498, Log Avg loss: 1.56291784, Global Avg Loss: 1.70374597, Time: 0.0507 Steps: 47900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001072, Sample Num: 17152, Cur Loss: 2.38755393, Cur Avg Loss: 1.35626733, Log Avg loss: 1.66131181, Global Avg Loss: 1.70373711, Time: 0.0626 Steps: 47910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001082, Sample Num: 17312, Cur Loss: 1.92695677, Cur Avg Loss: 1.35521389, Log Avg loss: 1.24228460, Global Avg Loss: 1.70364082, Time: 0.0504 Steps: 47920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001092, Sample Num: 17472, Cur Loss: 0.60260737, Cur Avg Loss: 1.35472022, Log Avg loss: 1.30130535, Global Avg Loss: 1.70355687, Time: 0.0425 Steps: 47930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001102, Sample Num: 17632, Cur Loss: 1.10033286, Cur Avg Loss: 1.35429247, Log Avg loss: 1.30758160, Global Avg Loss: 1.70347428, Time: 0.0406 Steps: 47940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001112, Sample Num: 17792, Cur Loss: 2.20775700, Cur Avg Loss: 1.35607120, Log Avg loss: 1.55208747, Global Avg Loss: 1.70344270, Time: 0.0572 Steps: 47950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001122, Sample Num: 17952, Cur Loss: 0.33238134, Cur Avg Loss: 1.35437516, Log Avg loss: 1.16577580, Global Avg Loss: 1.70333060, Time: 0.0537 Steps: 47960, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001132, Sample Num: 18112, Cur Loss: 0.28687823, Cur Avg Loss: 1.35103385, Log Avg loss: 0.97613907, Global Avg Loss: 1.70317900, Time: 0.0403 Steps: 47970, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001142, Sample Num: 18272, Cur Loss: 1.69629836, Cur Avg Loss: 1.35124743, Log Avg loss: 1.37542393, Global Avg Loss: 1.70311069, Time: 0.0404 Steps: 47980, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001152, Sample Num: 18432, Cur Loss: 1.39818144, Cur Avg Loss: 1.35459720, Log Avg loss: 1.73714157, Global Avg Loss: 1.70311778, Time: 0.0402 Steps: 47990, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001162, Sample Num: 18592, Cur Loss: 1.17138684, Cur Avg Loss: 1.35413683, Log Avg loss: 1.30110239, Global Avg Loss: 1.70303403, Time: 0.0402 Steps: 48000, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001172, Sample Num: 18752, Cur Loss: 0.68627393, Cur Avg Loss: 1.35320240, Log Avg loss: 1.24462153, Global Avg Loss: 1.70293855, Time: 0.0402 Steps: 48010, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001182, Sample Num: 18912, Cur Loss: 0.35295460, Cur Avg Loss: 1.35260059, Log Avg loss: 1.28206805, Global Avg Loss: 1.70285090, Time: 0.0403 Steps: 48020, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001192, Sample Num: 19072, Cur Loss: 1.82945728, Cur Avg Loss: 1.35237803, Log Avg loss: 1.32607175, Global Avg Loss: 1.70277246, Time: 0.0402 Steps: 48030, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001202, Sample Num: 19232, Cur Loss: 2.18498945, Cur Avg Loss: 1.35164783, Log Avg loss: 1.26460799, Global Avg Loss: 1.70268125, Time: 0.0402 Steps: 48040, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001212, Sample Num: 19392, Cur Loss: 1.55831289, Cur Avg Loss: 1.35112458, Log Avg loss: 1.28823029, Global Avg Loss: 1.70259499, Time: 0.0403 Steps: 48050, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001222, Sample Num: 19552, Cur Loss: 2.80284786, Cur Avg Loss: 1.35305435, Log Avg loss: 1.58694146, Global Avg Loss: 1.70257093, Time: 0.0402 Steps: 48060, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001232, Sample Num: 19712, Cur Loss: 0.89072013, Cur Avg Loss: 1.35280855, Log Avg loss: 1.32277224, Global Avg Loss: 1.70249192, Time: 0.0402 Steps: 48070, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001242, Sample Num: 19872, Cur Loss: 1.48447013, Cur Avg Loss: 1.35277746, Log Avg loss: 1.34894679, Global Avg Loss: 1.70241839, Time: 0.0402 Steps: 48080, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001252, Sample Num: 20032, Cur Loss: 1.92915010, Cur Avg Loss: 1.35459743, Log Avg loss: 1.58063790, Global Avg Loss: 1.70239306, Time: 0.0403 Steps: 48090, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001262, Sample Num: 20192, Cur Loss: 0.96542221, Cur Avg Loss: 1.35395113, Log Avg loss: 1.27303518, Global Avg Loss: 1.70230380, Time: 0.0402 Steps: 48100, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001272, Sample Num: 20352, Cur Loss: 1.66731191, Cur Avg Loss: 1.35333853, Log Avg loss: 1.27602801, Global Avg Loss: 1.70221520, Time: 0.0402 Steps: 48110, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001282, Sample Num: 20512, Cur Loss: 1.37490082, Cur Avg Loss: 1.35353575, Log Avg loss: 1.37862211, Global Avg Loss: 1.70214795, Time: 0.0403 Steps: 48120, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001292, Sample Num: 20672, Cur Loss: 1.20562387, Cur Avg Loss: 1.35143002, Log Avg loss: 1.08147503, Global Avg Loss: 1.70201899, Time: 0.0402 Steps: 48130, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001302, Sample Num: 20832, Cur Loss: 1.74665809, Cur Avg Loss: 1.35248329, Log Avg loss: 1.48856592, Global Avg Loss: 1.70197465, Time: 0.0402 Steps: 48140, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001312, Sample Num: 20992, Cur Loss: 2.47733212, Cur Avg Loss: 1.35487698, Log Avg loss: 1.66653566, Global Avg Loss: 1.70196729, Time: 0.0403 Steps: 48150, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001322, Sample Num: 21152, Cur Loss: 1.25934684, Cur Avg Loss: 1.35812337, Log Avg loss: 1.78404924, Global Avg Loss: 1.70198434, Time: 0.0402 Steps: 48160, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001332, Sample Num: 21312, Cur Loss: 1.28501642, Cur Avg Loss: 1.35705074, Log Avg loss: 1.21524913, Global Avg Loss: 1.70188329, Time: 0.0402 Steps: 48170, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001342, Sample Num: 21472, Cur Loss: 1.45494103, Cur Avg Loss: 1.35708898, Log Avg loss: 1.36218300, Global Avg Loss: 1.70181278, Time: 0.0402 Steps: 48180, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001352, Sample Num: 21632, Cur Loss: 1.51103747, Cur Avg Loss: 1.35690144, Log Avg loss: 1.33173354, Global Avg Loss: 1.70173599, Time: 0.0403 Steps: 48190, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001362, Sample Num: 21792, Cur Loss: 0.84103298, Cur Avg Loss: 1.35656051, Log Avg loss: 1.31046609, Global Avg Loss: 1.70165481, Time: 0.0402 Steps: 48200, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001372, Sample Num: 21952, Cur Loss: 1.40389609, Cur Avg Loss: 1.35843400, Log Avg loss: 1.61360446, Global Avg Loss: 1.70163655, Time: 0.0403 Steps: 48210, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001382, Sample Num: 22112, Cur Loss: 1.16067314, Cur Avg Loss: 1.35922010, Log Avg loss: 1.46707312, Global Avg Loss: 1.70158790, Time: 0.0402 Steps: 48220, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001392, Sample Num: 22272, Cur Loss: 0.87136954, Cur Avg Loss: 1.35715166, Log Avg loss: 1.07129254, Global Avg Loss: 1.70145722, Time: 0.0422 Steps: 48230, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001402, Sample Num: 22432, Cur Loss: 1.51073635, Cur Avg Loss: 1.35803698, Log Avg loss: 1.48127328, Global Avg Loss: 1.70141157, Time: 0.0827 Steps: 48240, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001412, Sample Num: 22592, Cur Loss: 1.84150434, Cur Avg Loss: 1.35894944, Log Avg loss: 1.48687654, Global Avg Loss: 1.70136711, Time: 0.0506 Steps: 48250, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001422, Sample Num: 22752, Cur Loss: 0.24417922, Cur Avg Loss: 1.35694039, Log Avg loss: 1.07326316, Global Avg Loss: 1.70123696, Time: 0.0608 Steps: 48260, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001432, Sample Num: 22912, Cur Loss: 1.09181929, Cur Avg Loss: 1.35608747, Log Avg loss: 1.23480240, Global Avg Loss: 1.70114033, Time: 0.0463 Steps: 48270, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001442, Sample Num: 23072, Cur Loss: 0.78280759, Cur Avg Loss: 1.35571883, Log Avg loss: 1.30292849, Global Avg Loss: 1.70105785, Time: 0.0643 Steps: 48280, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001452, Sample Num: 23232, Cur Loss: 1.29380822, Cur Avg Loss: 1.35702925, Log Avg loss: 1.54599234, Global Avg Loss: 1.70102574, Time: 0.0771 Steps: 48290, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001462, Sample Num: 23392, Cur Loss: 0.68123949, Cur Avg Loss: 1.35790770, Log Avg loss: 1.48545867, Global Avg Loss: 1.70098111, Time: 0.0750 Steps: 48300, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001472, Sample Num: 23552, Cur Loss: 0.57600832, Cur Avg Loss: 1.35717762, Log Avg loss: 1.25044015, Global Avg Loss: 1.70088785, Time: 0.0645 Steps: 48310, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001482, Sample Num: 23712, Cur Loss: 4.20025682, Cur Avg Loss: 1.36057732, Log Avg loss: 1.86101280, Global Avg Loss: 1.70092099, Time: 0.0403 Steps: 48320, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001492, Sample Num: 23872, Cur Loss: 0.96540201, Cur Avg Loss: 1.35809426, Log Avg loss: 0.99010448, Global Avg Loss: 1.70077391, Time: 0.0402 Steps: 48330, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001502, Sample Num: 24032, Cur Loss: 1.40224540, Cur Avg Loss: 1.35713668, Log Avg loss: 1.21426611, Global Avg Loss: 1.70067327, Time: 0.0402 Steps: 48340, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001512, Sample Num: 24192, Cur Loss: 0.85904491, Cur Avg Loss: 1.35905036, Log Avg loss: 1.64648565, Global Avg Loss: 1.70066206, Time: 0.0403 Steps: 48350, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001522, Sample Num: 24352, Cur Loss: 1.03555298, Cur Avg Loss: 1.36034930, Log Avg loss: 1.55674839, Global Avg Loss: 1.70063230, Time: 0.0403 Steps: 48360, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001532, Sample Num: 24512, Cur Loss: 1.53380513, Cur Avg Loss: 1.35939902, Log Avg loss: 1.21476649, Global Avg Loss: 1.70053185, Time: 0.0403 Steps: 48370, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001542, Sample Num: 24672, Cur Loss: 2.03960204, Cur Avg Loss: 1.36016918, Log Avg loss: 1.47815776, Global Avg Loss: 1.70048589, Time: 0.0404 Steps: 48380, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001552, Sample Num: 24832, Cur Loss: 1.40051758, Cur Avg Loss: 1.36046821, Log Avg loss: 1.40657771, Global Avg Loss: 1.70042515, Time: 0.0404 Steps: 48390, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001562, Sample Num: 24992, Cur Loss: 0.92358816, Cur Avg Loss: 1.36118011, Log Avg loss: 1.47166830, Global Avg Loss: 1.70037789, Time: 0.0403 Steps: 48400, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001572, Sample Num: 25152, Cur Loss: 0.57933986, Cur Avg Loss: 1.36127146, Log Avg loss: 1.37553924, Global Avg Loss: 1.70031079, Time: 0.0403 Steps: 48410, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001582, Sample Num: 25312, Cur Loss: 0.51378655, Cur Avg Loss: 1.35869539, Log Avg loss: 0.95373759, Global Avg Loss: 1.70015660, Time: 0.0403 Steps: 48420, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001592, Sample Num: 25472, Cur Loss: 1.74046969, Cur Avg Loss: 1.35763595, Log Avg loss: 1.19003311, Global Avg Loss: 1.70005127, Time: 0.0402 Steps: 48430, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001602, Sample Num: 25632, Cur Loss: 1.89892185, Cur Avg Loss: 1.35967385, Log Avg loss: 1.68410697, Global Avg Loss: 1.70004798, Time: 0.0403 Steps: 48440, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001612, Sample Num: 25792, Cur Loss: 2.12693071, Cur Avg Loss: 1.36095025, Log Avg loss: 1.56542898, Global Avg Loss: 1.70002019, Time: 0.0403 Steps: 48450, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001622, Sample Num: 25952, Cur Loss: 0.77880013, Cur Avg Loss: 1.35883893, Log Avg loss: 1.01849527, Global Avg Loss: 1.69987956, Time: 0.0403 Steps: 48460, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001632, Sample Num: 26112, Cur Loss: 1.35507810, Cur Avg Loss: 1.36038692, Log Avg loss: 1.61146988, Global Avg Loss: 1.69986132, Time: 0.0403 Steps: 48470, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001642, Sample Num: 26272, Cur Loss: 1.85450184, Cur Avg Loss: 1.36148289, Log Avg loss: 1.54034501, Global Avg Loss: 1.69982841, Time: 0.0402 Steps: 48480, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001652, Sample Num: 26432, Cur Loss: 2.86673737, Cur Avg Loss: 1.36070353, Log Avg loss: 1.23273349, Global Avg Loss: 1.69973208, Time: 0.0402 Steps: 48490, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001662, Sample Num: 26592, Cur Loss: 2.17739511, Cur Avg Loss: 1.36151259, Log Avg loss: 1.49516920, Global Avg Loss: 1.69968991, Time: 0.0403 Steps: 48500, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001672, Sample Num: 26752, Cur Loss: 1.35310602, Cur Avg Loss: 1.36202105, Log Avg loss: 1.44652638, Global Avg Loss: 1.69963772, Time: 0.0403 Steps: 48510, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001682, Sample Num: 26912, Cur Loss: 0.70549929, Cur Avg Loss: 1.35995907, Log Avg loss: 1.01519589, Global Avg Loss: 1.69949665, Time: 0.0403 Steps: 48520, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001692, Sample Num: 27072, Cur Loss: 0.96813411, Cur Avg Loss: 1.36152021, Log Avg loss: 1.62410496, Global Avg Loss: 1.69948112, Time: 0.0403 Steps: 48530, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001702, Sample Num: 27232, Cur Loss: 1.43770671, Cur Avg Loss: 1.36078134, Log Avg loss: 1.23576481, Global Avg Loss: 1.69938559, Time: 0.0402 Steps: 48540, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001712, Sample Num: 27392, Cur Loss: 2.69968748, Cur Avg Loss: 1.36284361, Log Avg loss: 1.71384195, Global Avg Loss: 1.69938856, Time: 0.0403 Steps: 48550, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001722, Sample Num: 27552, Cur Loss: 0.57709497, Cur Avg Loss: 1.36091651, Log Avg loss: 1.03099636, Global Avg Loss: 1.69925092, Time: 0.0403 Steps: 48560, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001732, Sample Num: 27712, Cur Loss: 1.27720690, Cur Avg Loss: 1.35740832, Log Avg loss: 0.75329774, Global Avg Loss: 1.69905616, Time: 0.0827 Steps: 48570, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001742, Sample Num: 27872, Cur Loss: 0.71586758, Cur Avg Loss: 1.35723894, Log Avg loss: 1.32790335, Global Avg Loss: 1.69897976, Time: 0.0791 Steps: 48580, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001752, Sample Num: 28032, Cur Loss: 1.69937849, Cur Avg Loss: 1.35716881, Log Avg loss: 1.34495165, Global Avg Loss: 1.69890690, Time: 0.0439 Steps: 48590, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001762, Sample Num: 28192, Cur Loss: 0.76480055, Cur Avg Loss: 1.35587051, Log Avg loss: 1.12840916, Global Avg Loss: 1.69878951, Time: 0.0406 Steps: 48600, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001772, Sample Num: 28352, Cur Loss: 1.34325624, Cur Avg Loss: 1.35786129, Log Avg loss: 1.70863619, Global Avg Loss: 1.69879154, Time: 0.0406 Steps: 48610, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001782, Sample Num: 28512, Cur Loss: 1.61827326, Cur Avg Loss: 1.35832808, Log Avg loss: 1.44104266, Global Avg Loss: 1.69873853, Time: 0.0404 Steps: 48620, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001792, Sample Num: 28672, Cur Loss: 1.64361787, Cur Avg Loss: 1.35866705, Log Avg loss: 1.41907142, Global Avg Loss: 1.69868102, Time: 0.0898 Steps: 48630, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001802, Sample Num: 28832, Cur Loss: 0.92265141, Cur Avg Loss: 1.35850158, Log Avg loss: 1.32885013, Global Avg Loss: 1.69860498, Time: 0.0429 Steps: 48640, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001812, Sample Num: 28992, Cur Loss: 0.35595089, Cur Avg Loss: 1.35859165, Log Avg loss: 1.37482212, Global Avg Loss: 1.69853843, Time: 0.0405 Steps: 48650, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001822, Sample Num: 29152, Cur Loss: 0.57626331, Cur Avg Loss: 1.35722093, Log Avg loss: 1.10884555, Global Avg Loss: 1.69841724, Time: 0.0402 Steps: 48660, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001832, Sample Num: 29312, Cur Loss: 1.41440678, Cur Avg Loss: 1.35633777, Log Avg loss: 1.19542598, Global Avg Loss: 1.69831390, Time: 0.0403 Steps: 48670, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001842, Sample Num: 29472, Cur Loss: 1.14657605, Cur Avg Loss: 1.35638585, Log Avg loss: 1.36519421, Global Avg Loss: 1.69824546, Time: 0.0403 Steps: 48680, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001852, Sample Num: 29632, Cur Loss: 1.12472677, Cur Avg Loss: 1.35626887, Log Avg loss: 1.33472160, Global Avg Loss: 1.69817080, Time: 0.0402 Steps: 48690, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001862, Sample Num: 29792, Cur Loss: 2.08083892, Cur Avg Loss: 1.35488420, Log Avg loss: 1.09844342, Global Avg Loss: 1.69804766, Time: 0.0402 Steps: 48700, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001872, Sample Num: 29952, Cur Loss: 1.70231199, Cur Avg Loss: 1.35525586, Log Avg loss: 1.42445846, Global Avg Loss: 1.69799149, Time: 0.0403 Steps: 48710, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001882, Sample Num: 30112, Cur Loss: 1.07992589, Cur Avg Loss: 1.35562396, Log Avg loss: 1.42453344, Global Avg Loss: 1.69793536, Time: 0.0403 Steps: 48720, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001892, Sample Num: 30272, Cur Loss: 1.99345446, Cur Avg Loss: 1.35616296, Log Avg loss: 1.45760136, Global Avg Loss: 1.69788604, Time: 0.0402 Steps: 48730, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001902, Sample Num: 30432, Cur Loss: 2.88087893, Cur Avg Loss: 1.35901711, Log Avg loss: 1.89902283, Global Avg Loss: 1.69792731, Time: 0.0402 Steps: 48740, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001912, Sample Num: 30592, Cur Loss: 2.00147223, Cur Avg Loss: 1.35972749, Log Avg loss: 1.49484114, Global Avg Loss: 1.69788565, Time: 0.0403 Steps: 48750, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001922, Sample Num: 30752, Cur Loss: 0.58736664, Cur Avg Loss: 1.35802173, Log Avg loss: 1.03188047, Global Avg Loss: 1.69774906, Time: 0.0403 Steps: 48760, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001932, Sample Num: 30912, Cur Loss: 0.91756690, Cur Avg Loss: 1.35815518, Log Avg loss: 1.38380463, Global Avg Loss: 1.69768469, Time: 0.0402 Steps: 48770, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001942, Sample Num: 31072, Cur Loss: 2.29485798, Cur Avg Loss: 1.35769201, Log Avg loss: 1.26820846, Global Avg Loss: 1.69759665, Time: 0.0403 Steps: 48780, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001952, Sample Num: 31232, Cur Loss: 1.05502498, Cur Avg Loss: 1.35531834, Log Avg loss: 0.89435036, Global Avg Loss: 1.69743201, Time: 0.0403 Steps: 48790, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001962, Sample Num: 31392, Cur Loss: 1.20500660, Cur Avg Loss: 1.35732721, Log Avg loss: 1.74946001, Global Avg Loss: 1.69744267, Time: 0.0402 Steps: 48800, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001972, Sample Num: 31552, Cur Loss: 0.47377199, Cur Avg Loss: 1.35636704, Log Avg loss: 1.16798118, Global Avg Loss: 1.69733420, Time: 0.0402 Steps: 48810, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001982, Sample Num: 31712, Cur Loss: 2.22704482, Cur Avg Loss: 1.35821350, Log Avg loss: 1.72233482, Global Avg Loss: 1.69733932, Time: 0.0403 Steps: 48820, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001992, Sample Num: 31872, Cur Loss: 1.20555520, Cur Avg Loss: 1.36032236, Log Avg loss: 1.77829907, Global Avg Loss: 1.69735590, Time: 0.0403 Steps: 48830, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002002, Sample Num: 32032, Cur Loss: 0.42548972, Cur Avg Loss: 1.35973917, Log Avg loss: 1.24356790, Global Avg Loss: 1.69726299, Time: 0.0403 Steps: 48840, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002012, Sample Num: 32192, Cur Loss: 0.98629928, Cur Avg Loss: 1.35892835, Log Avg loss: 1.19660183, Global Avg Loss: 1.69716050, Time: 0.0402 Steps: 48850, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002022, Sample Num: 32352, Cur Loss: 1.85331047, Cur Avg Loss: 1.36054918, Log Avg loss: 1.68666061, Global Avg Loss: 1.69715835, Time: 0.0403 Steps: 48860, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002032, Sample Num: 32512, Cur Loss: 1.21780705, Cur Avg Loss: 1.36178328, Log Avg loss: 1.61131759, Global Avg Loss: 1.69714078, Time: 0.0403 Steps: 48870, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002042, Sample Num: 32672, Cur Loss: 2.06217027, Cur Avg Loss: 1.36173406, Log Avg loss: 1.35173284, Global Avg Loss: 1.69707012, Time: 0.0402 Steps: 48880, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002052, Sample Num: 32832, Cur Loss: 1.30999827, Cur Avg Loss: 1.36243172, Log Avg loss: 1.50489386, Global Avg Loss: 1.69703081, Time: 0.0402 Steps: 48890, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002062, Sample Num: 32992, Cur Loss: 1.18379951, Cur Avg Loss: 1.36310503, Log Avg loss: 1.50126898, Global Avg Loss: 1.69699078, Time: 0.0402 Steps: 48900, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002072, Sample Num: 33152, Cur Loss: 1.08498085, Cur Avg Loss: 1.36226878, Log Avg loss: 1.18983243, Global Avg Loss: 1.69688709, Time: 0.0402 Steps: 48910, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002082, Sample Num: 33312, Cur Loss: 1.09230566, Cur Avg Loss: 1.36158553, Log Avg loss: 1.22001647, Global Avg Loss: 1.69678961, Time: 0.0416 Steps: 48920, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002092, Sample Num: 33472, Cur Loss: 3.24230623, Cur Avg Loss: 1.36256681, Log Avg loss: 1.56686910, Global Avg Loss: 1.69676305, Time: 0.0433 Steps: 48930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002102, Sample Num: 33632, Cur Loss: 2.50598192, Cur Avg Loss: 1.36329311, Log Avg loss: 1.51523519, Global Avg Loss: 1.69672596, Time: 0.0624 Steps: 48940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002112, Sample Num: 33792, Cur Loss: 2.01136684, Cur Avg Loss: 1.36426626, Log Avg loss: 1.56882327, Global Avg Loss: 1.69669983, Time: 0.0405 Steps: 48950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002122, Sample Num: 33952, Cur Loss: 0.59047860, Cur Avg Loss: 1.36262484, Log Avg loss: 1.01595736, Global Avg Loss: 1.69656079, Time: 0.0577 Steps: 48960, Updated lr: 0.000055 ***** Running evaluation checkpoint-48967 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-48967 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.564506, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.294821, "eval_total_loss": 910.258875, "eval_mae": 0.814303, "eval_mse": 1.295205, "eval_r2": 0.176682, "eval_sp_statistic": 0.372507, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.460366, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.195142, "test_total_loss": 599.961153, "test_mae": 0.786095, "test_mse": 1.19547, "test_r2": 0.228433, "test_sp_statistic": 0.382319, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.478222, "test_ps_pvalue": 0.0, "lr": 5.451209103840683e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6964696987714494, "train_cur_epoch_loss": 2898.9052349328995, "train_cur_epoch_avg_loss": 1.3616276350084076, "train_cur_epoch_time": 95.56450605392456, "train_cur_epoch_avg_time": 0.04488703901076776, "epoch": 23, "step": 48967} ################################################## Training, Epoch: 0024, Batch: 000003, Sample Num: 48, Cur Loss: 1.71679151, Cur Avg Loss: 1.00937066, Log Avg loss: 1.04434303, Global Avg Loss: 1.69642761, Time: 0.0455 Steps: 48970, Updated lr: 0.000055 Training, Epoch: 0024, Batch: 000013, Sample Num: 208, Cur Loss: 0.97893536, Cur Avg Loss: 1.30597729, Log Avg loss: 1.39495928, Global Avg Loss: 1.69636606, Time: 0.0843 Steps: 48980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000023, Sample Num: 368, Cur Loss: 0.95770133, Cur Avg Loss: 1.20798641, Log Avg loss: 1.08059827, Global Avg Loss: 1.69624036, Time: 0.0889 Steps: 48990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000033, Sample Num: 528, Cur Loss: 0.84596974, Cur Avg Loss: 1.27051927, Log Avg loss: 1.41434485, Global Avg Loss: 1.69618283, Time: 0.0619 Steps: 49000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000043, Sample Num: 688, Cur Loss: 1.26932955, Cur Avg Loss: 1.28426452, Log Avg loss: 1.32962382, Global Avg Loss: 1.69610804, Time: 0.0638 Steps: 49010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000053, Sample Num: 848, Cur Loss: 2.48198390, Cur Avg Loss: 1.32715641, Log Avg loss: 1.51159153, Global Avg Loss: 1.69607040, Time: 0.0407 Steps: 49020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000063, Sample Num: 1008, Cur Loss: 2.19710302, Cur Avg Loss: 1.32857258, Log Avg loss: 1.33607829, Global Avg Loss: 1.69599698, Time: 0.0558 Steps: 49030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000073, Sample Num: 1168, Cur Loss: 1.28054881, Cur Avg Loss: 1.32974253, Log Avg loss: 1.33711324, Global Avg Loss: 1.69592380, Time: 0.0402 Steps: 49040, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000083, Sample Num: 1328, Cur Loss: 0.56755561, Cur Avg Loss: 1.33294438, Log Avg loss: 1.35631784, Global Avg Loss: 1.69585456, Time: 0.0404 Steps: 49050, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000093, Sample Num: 1488, Cur Loss: 1.29951477, Cur Avg Loss: 1.33755128, Log Avg loss: 1.37578861, Global Avg Loss: 1.69578932, Time: 0.0402 Steps: 49060, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000103, Sample Num: 1648, Cur Loss: 2.21371651, Cur Avg Loss: 1.35751797, Log Avg loss: 1.54320822, Global Avg Loss: 1.69575822, Time: 0.0403 Steps: 49070, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000113, Sample Num: 1808, Cur Loss: 0.77118480, Cur Avg Loss: 1.34114928, Log Avg loss: 1.17255170, Global Avg Loss: 1.69565162, Time: 0.0402 Steps: 49080, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000123, Sample Num: 1968, Cur Loss: 0.81644809, Cur Avg Loss: 1.32934756, Log Avg loss: 1.19598814, Global Avg Loss: 1.69554984, Time: 0.0402 Steps: 49090, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000133, Sample Num: 2128, Cur Loss: 1.87578547, Cur Avg Loss: 1.31336884, Log Avg loss: 1.11683059, Global Avg Loss: 1.69543197, Time: 0.0403 Steps: 49100, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000143, Sample Num: 2288, Cur Loss: 1.19017696, Cur Avg Loss: 1.31241122, Log Avg loss: 1.29967488, Global Avg Loss: 1.69535139, Time: 0.0402 Steps: 49110, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000153, Sample Num: 2448, Cur Loss: 0.45309007, Cur Avg Loss: 1.31173214, Log Avg loss: 1.30202126, Global Avg Loss: 1.69527131, Time: 0.0403 Steps: 49120, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000163, Sample Num: 2608, Cur Loss: 0.33086726, Cur Avg Loss: 1.29918397, Log Avg loss: 1.10719701, Global Avg Loss: 1.69515161, Time: 0.0402 Steps: 49130, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000173, Sample Num: 2768, Cur Loss: 2.28637862, Cur Avg Loss: 1.32831595, Log Avg loss: 1.80316713, Global Avg Loss: 1.69517359, Time: 0.0402 Steps: 49140, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000183, Sample Num: 2928, Cur Loss: 1.79089594, Cur Avg Loss: 1.34435649, Log Avg loss: 1.62185795, Global Avg Loss: 1.69515868, Time: 0.0402 Steps: 49150, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000193, Sample Num: 3088, Cur Loss: 0.90808225, Cur Avg Loss: 1.34122252, Log Avg loss: 1.28387088, Global Avg Loss: 1.69507501, Time: 0.0402 Steps: 49160, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000203, Sample Num: 3248, Cur Loss: 0.96720225, Cur Avg Loss: 1.33479149, Log Avg loss: 1.21067265, Global Avg Loss: 1.69497650, Time: 0.0402 Steps: 49170, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000213, Sample Num: 3408, Cur Loss: 2.07919216, Cur Avg Loss: 1.34077857, Log Avg loss: 1.46231615, Global Avg Loss: 1.69492919, Time: 0.0402 Steps: 49180, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000223, Sample Num: 3568, Cur Loss: 1.03076220, Cur Avg Loss: 1.33614534, Log Avg loss: 1.23745753, Global Avg Loss: 1.69483619, Time: 0.0402 Steps: 49190, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000233, Sample Num: 3728, Cur Loss: 1.18059242, Cur Avg Loss: 1.33558066, Log Avg loss: 1.32298826, Global Avg Loss: 1.69476061, Time: 0.0402 Steps: 49200, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000243, Sample Num: 3888, Cur Loss: 0.92091310, Cur Avg Loss: 1.35046188, Log Avg loss: 1.69719453, Global Avg Loss: 1.69476111, Time: 0.0402 Steps: 49210, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000253, Sample Num: 4048, Cur Loss: 2.42173791, Cur Avg Loss: 1.35627442, Log Avg loss: 1.49751914, Global Avg Loss: 1.69472103, Time: 0.0402 Steps: 49220, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000263, Sample Num: 4208, Cur Loss: 1.68748379, Cur Avg Loss: 1.34955727, Log Avg loss: 1.17961316, Global Avg Loss: 1.69461640, Time: 0.0402 Steps: 49230, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000273, Sample Num: 4368, Cur Loss: 1.17699146, Cur Avg Loss: 1.34357837, Log Avg loss: 1.18633354, Global Avg Loss: 1.69451317, Time: 0.0402 Steps: 49240, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000283, Sample Num: 4528, Cur Loss: 0.97847891, Cur Avg Loss: 1.34179517, Log Avg loss: 1.29311360, Global Avg Loss: 1.69443167, Time: 0.0403 Steps: 49250, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000293, Sample Num: 4688, Cur Loss: 0.87721628, Cur Avg Loss: 1.33110039, Log Avg loss: 1.02843808, Global Avg Loss: 1.69429647, Time: 0.0402 Steps: 49260, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000303, Sample Num: 4848, Cur Loss: 0.58616346, Cur Avg Loss: 1.33065633, Log Avg loss: 1.31764559, Global Avg Loss: 1.69422002, Time: 0.0402 Steps: 49270, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000313, Sample Num: 5008, Cur Loss: 0.99672616, Cur Avg Loss: 1.33844296, Log Avg loss: 1.57437783, Global Avg Loss: 1.69419571, Time: 0.0402 Steps: 49280, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000323, Sample Num: 5168, Cur Loss: 1.56241977, Cur Avg Loss: 1.34752170, Log Avg loss: 1.63168632, Global Avg Loss: 1.69418302, Time: 0.0402 Steps: 49290, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000333, Sample Num: 5328, Cur Loss: 0.79890418, Cur Avg Loss: 1.33787046, Log Avg loss: 1.02613527, Global Avg Loss: 1.69404752, Time: 0.0495 Steps: 49300, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000343, Sample Num: 5488, Cur Loss: 1.14368665, Cur Avg Loss: 1.32831591, Log Avg loss: 1.01014953, Global Avg Loss: 1.69390882, Time: 0.0652 Steps: 49310, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000353, Sample Num: 5648, Cur Loss: 1.05900741, Cur Avg Loss: 1.33748288, Log Avg loss: 1.65190988, Global Avg Loss: 1.69390031, Time: 0.0611 Steps: 49320, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000363, Sample Num: 5808, Cur Loss: 1.95007968, Cur Avg Loss: 1.34359647, Log Avg loss: 1.55940613, Global Avg Loss: 1.69387304, Time: 0.0407 Steps: 49330, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000373, Sample Num: 5968, Cur Loss: 1.28047752, Cur Avg Loss: 1.34266730, Log Avg loss: 1.30893832, Global Avg Loss: 1.69379503, Time: 0.0562 Steps: 49340, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000383, Sample Num: 6128, Cur Loss: 1.75857687, Cur Avg Loss: 1.34573290, Log Avg loss: 1.46007995, Global Avg Loss: 1.69374767, Time: 0.1005 Steps: 49350, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000393, Sample Num: 6288, Cur Loss: 1.36028886, Cur Avg Loss: 1.35262552, Log Avg loss: 1.61661273, Global Avg Loss: 1.69373204, Time: 0.0614 Steps: 49360, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000403, Sample Num: 6448, Cur Loss: 1.20323372, Cur Avg Loss: 1.35125650, Log Avg loss: 1.29745397, Global Avg Loss: 1.69365177, Time: 0.0617 Steps: 49370, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000413, Sample Num: 6608, Cur Loss: 1.05546618, Cur Avg Loss: 1.35071940, Log Avg loss: 1.32907461, Global Avg Loss: 1.69357794, Time: 0.0405 Steps: 49380, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000423, Sample Num: 6768, Cur Loss: 1.52848148, Cur Avg Loss: 1.35816907, Log Avg loss: 1.66584030, Global Avg Loss: 1.69357233, Time: 0.0402 Steps: 49390, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000433, Sample Num: 6928, Cur Loss: 0.48207152, Cur Avg Loss: 1.35236877, Log Avg loss: 1.10701594, Global Avg Loss: 1.69345359, Time: 0.0402 Steps: 49400, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000443, Sample Num: 7088, Cur Loss: 1.67829788, Cur Avg Loss: 1.34629169, Log Avg loss: 1.08315408, Global Avg Loss: 1.69333007, Time: 0.0402 Steps: 49410, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000453, Sample Num: 7248, Cur Loss: 1.09894991, Cur Avg Loss: 1.34135952, Log Avg loss: 1.12286438, Global Avg Loss: 1.69321464, Time: 0.0403 Steps: 49420, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000463, Sample Num: 7408, Cur Loss: 2.19906712, Cur Avg Loss: 1.34234569, Log Avg loss: 1.38701931, Global Avg Loss: 1.69315270, Time: 0.0402 Steps: 49430, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000473, Sample Num: 7568, Cur Loss: 0.56103742, Cur Avg Loss: 1.34333803, Log Avg loss: 1.38928343, Global Avg Loss: 1.69309123, Time: 0.0403 Steps: 49440, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000483, Sample Num: 7728, Cur Loss: 0.63982832, Cur Avg Loss: 1.33996455, Log Avg loss: 1.18039896, Global Avg Loss: 1.69298756, Time: 0.0402 Steps: 49450, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000493, Sample Num: 7888, Cur Loss: 1.29577935, Cur Avg Loss: 1.33442607, Log Avg loss: 1.06691726, Global Avg Loss: 1.69286097, Time: 0.0403 Steps: 49460, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000503, Sample Num: 8048, Cur Loss: 1.52997065, Cur Avg Loss: 1.34129240, Log Avg loss: 1.67980267, Global Avg Loss: 1.69285833, Time: 0.0403 Steps: 49470, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000513, Sample Num: 8208, Cur Loss: 1.33577800, Cur Avg Loss: 1.34058957, Log Avg loss: 1.30523694, Global Avg Loss: 1.69278000, Time: 0.0403 Steps: 49480, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000523, Sample Num: 8368, Cur Loss: 1.11926103, Cur Avg Loss: 1.33512591, Log Avg loss: 1.05484035, Global Avg Loss: 1.69265109, Time: 0.0402 Steps: 49490, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000533, Sample Num: 8528, Cur Loss: 1.41175890, Cur Avg Loss: 1.33472218, Log Avg loss: 1.31360734, Global Avg Loss: 1.69257452, Time: 0.0403 Steps: 49500, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000543, Sample Num: 8688, Cur Loss: 0.83824772, Cur Avg Loss: 1.33117276, Log Avg loss: 1.14198856, Global Avg Loss: 1.69246331, Time: 0.0404 Steps: 49510, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000553, Sample Num: 8848, Cur Loss: 0.46348861, Cur Avg Loss: 1.33473596, Log Avg loss: 1.52821783, Global Avg Loss: 1.69243014, Time: 0.0401 Steps: 49520, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000563, Sample Num: 9008, Cur Loss: 0.82163233, Cur Avg Loss: 1.32970798, Log Avg loss: 1.05166063, Global Avg Loss: 1.69230077, Time: 0.0402 Steps: 49530, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000573, Sample Num: 9168, Cur Loss: 0.78955346, Cur Avg Loss: 1.32320564, Log Avg loss: 0.95712363, Global Avg Loss: 1.69215237, Time: 0.0402 Steps: 49540, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000583, Sample Num: 9328, Cur Loss: 2.47103381, Cur Avg Loss: 1.33265915, Log Avg loss: 1.87434554, Global Avg Loss: 1.69218914, Time: 0.0402 Steps: 49550, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000593, Sample Num: 9488, Cur Loss: 1.38800991, Cur Avg Loss: 1.33316784, Log Avg loss: 1.36282412, Global Avg Loss: 1.69212268, Time: 0.0402 Steps: 49560, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000603, Sample Num: 9648, Cur Loss: 0.30850524, Cur Avg Loss: 1.33132783, Log Avg loss: 1.22221542, Global Avg Loss: 1.69202789, Time: 0.0402 Steps: 49570, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000613, Sample Num: 9808, Cur Loss: 0.58966321, Cur Avg Loss: 1.32735268, Log Avg loss: 1.08765100, Global Avg Loss: 1.69190599, Time: 0.0401 Steps: 49580, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000623, Sample Num: 9968, Cur Loss: 0.95181876, Cur Avg Loss: 1.32444642, Log Avg loss: 1.14629266, Global Avg Loss: 1.69179596, Time: 0.0402 Steps: 49590, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000633, Sample Num: 10128, Cur Loss: 1.32834017, Cur Avg Loss: 1.32459336, Log Avg loss: 1.33374774, Global Avg Loss: 1.69172378, Time: 0.0402 Steps: 49600, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000643, Sample Num: 10288, Cur Loss: 1.47449625, Cur Avg Loss: 1.31978308, Log Avg loss: 1.01529250, Global Avg Loss: 1.69158743, Time: 0.0402 Steps: 49610, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000653, Sample Num: 10448, Cur Loss: 1.20836842, Cur Avg Loss: 1.31626832, Log Avg loss: 1.09026957, Global Avg Loss: 1.69146624, Time: 0.0403 Steps: 49620, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000663, Sample Num: 10608, Cur Loss: 1.21812630, Cur Avg Loss: 1.31436558, Log Avg loss: 1.19011665, Global Avg Loss: 1.69136523, Time: 0.0402 Steps: 49630, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000673, Sample Num: 10768, Cur Loss: 1.37894726, Cur Avg Loss: 1.31572976, Log Avg loss: 1.40617487, Global Avg Loss: 1.69130777, Time: 0.0643 Steps: 49640, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000683, Sample Num: 10928, Cur Loss: 2.86495423, Cur Avg Loss: 1.32134268, Log Avg loss: 1.69909175, Global Avg Loss: 1.69130934, Time: 0.0765 Steps: 49650, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000693, Sample Num: 11088, Cur Loss: 1.52517855, Cur Avg Loss: 1.32308867, Log Avg loss: 1.44233996, Global Avg Loss: 1.69125921, Time: 0.0899 Steps: 49660, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000703, Sample Num: 11248, Cur Loss: 0.72647363, Cur Avg Loss: 1.31950791, Log Avg loss: 1.07136118, Global Avg Loss: 1.69113440, Time: 0.0405 Steps: 49670, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000713, Sample Num: 11408, Cur Loss: 1.19229102, Cur Avg Loss: 1.31634537, Log Avg loss: 1.09401899, Global Avg Loss: 1.69101421, Time: 0.0769 Steps: 49680, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000723, Sample Num: 11568, Cur Loss: 1.08210707, Cur Avg Loss: 1.31379517, Log Avg loss: 1.13196600, Global Avg Loss: 1.69090170, Time: 0.0554 Steps: 49690, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000733, Sample Num: 11728, Cur Loss: 1.05595148, Cur Avg Loss: 1.31111284, Log Avg loss: 1.11718049, Global Avg Loss: 1.69078627, Time: 0.0406 Steps: 49700, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000743, Sample Num: 11888, Cur Loss: 0.87878716, Cur Avg Loss: 1.31161415, Log Avg loss: 1.34835975, Global Avg Loss: 1.69071738, Time: 0.0432 Steps: 49710, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000753, Sample Num: 12048, Cur Loss: 2.10650516, Cur Avg Loss: 1.31344311, Log Avg loss: 1.44933507, Global Avg Loss: 1.69066883, Time: 0.0587 Steps: 49720, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000763, Sample Num: 12208, Cur Loss: 1.04338288, Cur Avg Loss: 1.31275924, Log Avg loss: 1.26126373, Global Avg Loss: 1.69058249, Time: 0.0402 Steps: 49730, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000773, Sample Num: 12368, Cur Loss: 1.66555512, Cur Avg Loss: 1.31127887, Log Avg loss: 1.19832665, Global Avg Loss: 1.69048352, Time: 0.0402 Steps: 49740, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000783, Sample Num: 12528, Cur Loss: 1.79240072, Cur Avg Loss: 1.31063008, Log Avg loss: 1.26047831, Global Avg Loss: 1.69039709, Time: 0.0403 Steps: 49750, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000793, Sample Num: 12688, Cur Loss: 1.76686788, Cur Avg Loss: 1.31143590, Log Avg loss: 1.37453173, Global Avg Loss: 1.69033361, Time: 0.0402 Steps: 49760, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000803, Sample Num: 12848, Cur Loss: 1.51938605, Cur Avg Loss: 1.31667275, Log Avg loss: 1.73195473, Global Avg Loss: 1.69034197, Time: 0.0402 Steps: 49770, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000813, Sample Num: 13008, Cur Loss: 1.35730863, Cur Avg Loss: 1.31560615, Log Avg loss: 1.22995863, Global Avg Loss: 1.69024949, Time: 0.0402 Steps: 49780, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000823, Sample Num: 13168, Cur Loss: 0.97770399, Cur Avg Loss: 1.31814081, Log Avg loss: 1.52420822, Global Avg Loss: 1.69021614, Time: 0.0402 Steps: 49790, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000833, Sample Num: 13328, Cur Loss: 1.15845132, Cur Avg Loss: 1.31672476, Log Avg loss: 1.20018405, Global Avg Loss: 1.69011774, Time: 0.0403 Steps: 49800, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000843, Sample Num: 13488, Cur Loss: 0.50687230, Cur Avg Loss: 1.31412848, Log Avg loss: 1.09785831, Global Avg Loss: 1.68999884, Time: 0.0402 Steps: 49810, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000853, Sample Num: 13648, Cur Loss: 1.38337052, Cur Avg Loss: 1.31640885, Log Avg loss: 1.50864444, Global Avg Loss: 1.68996243, Time: 0.0402 Steps: 49820, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000863, Sample Num: 13808, Cur Loss: 0.86948091, Cur Avg Loss: 1.31566392, Log Avg loss: 1.25212126, Global Avg Loss: 1.68987457, Time: 0.0402 Steps: 49830, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000873, Sample Num: 13968, Cur Loss: 0.72580564, Cur Avg Loss: 1.31477895, Log Avg loss: 1.23840598, Global Avg Loss: 1.68978398, Time: 0.0403 Steps: 49840, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000883, Sample Num: 14128, Cur Loss: 1.65067363, Cur Avg Loss: 1.31302519, Log Avg loss: 1.15992214, Global Avg Loss: 1.68967769, Time: 0.0402 Steps: 49850, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000893, Sample Num: 14288, Cur Loss: 0.88954371, Cur Avg Loss: 1.31439866, Log Avg loss: 1.43567616, Global Avg Loss: 1.68962675, Time: 0.0402 Steps: 49860, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000903, Sample Num: 14448, Cur Loss: 1.75009382, Cur Avg Loss: 1.31575557, Log Avg loss: 1.43692719, Global Avg Loss: 1.68957608, Time: 0.0402 Steps: 49870, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000913, Sample Num: 14608, Cur Loss: 1.17015529, Cur Avg Loss: 1.31773780, Log Avg loss: 1.49673334, Global Avg Loss: 1.68953742, Time: 0.0402 Steps: 49880, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000923, Sample Num: 14768, Cur Loss: 1.68014431, Cur Avg Loss: 1.31673914, Log Avg loss: 1.22556104, Global Avg Loss: 1.68944442, Time: 0.0403 Steps: 49890, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000933, Sample Num: 14928, Cur Loss: 0.96364045, Cur Avg Loss: 1.31211895, Log Avg loss: 0.88567589, Global Avg Loss: 1.68928334, Time: 0.0403 Steps: 49900, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000943, Sample Num: 15088, Cur Loss: 1.55569649, Cur Avg Loss: 1.31210671, Log Avg loss: 1.31096448, Global Avg Loss: 1.68920754, Time: 0.0402 Steps: 49910, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000953, Sample Num: 15248, Cur Loss: 1.60071158, Cur Avg Loss: 1.31249975, Log Avg loss: 1.34956334, Global Avg Loss: 1.68913950, Time: 0.0403 Steps: 49920, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000963, Sample Num: 15408, Cur Loss: 1.07832372, Cur Avg Loss: 1.30889703, Log Avg loss: 0.96555791, Global Avg Loss: 1.68899458, Time: 0.0402 Steps: 49930, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000973, Sample Num: 15568, Cur Loss: 2.93999100, Cur Avg Loss: 1.31318842, Log Avg loss: 1.72644953, Global Avg Loss: 1.68900208, Time: 0.0403 Steps: 49940, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000983, Sample Num: 15728, Cur Loss: 1.10258269, Cur Avg Loss: 1.31597968, Log Avg loss: 1.58756899, Global Avg Loss: 1.68898178, Time: 0.0403 Steps: 49950, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000993, Sample Num: 15888, Cur Loss: 1.62025201, Cur Avg Loss: 1.31874485, Log Avg loss: 1.59056127, Global Avg Loss: 1.68896208, Time: 0.0403 Steps: 49960, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001003, Sample Num: 16048, Cur Loss: 1.10275280, Cur Avg Loss: 1.32050886, Log Avg loss: 1.49567513, Global Avg Loss: 1.68892340, Time: 0.0402 Steps: 49970, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001013, Sample Num: 16208, Cur Loss: 1.10703969, Cur Avg Loss: 1.32219478, Log Avg loss: 1.49129198, Global Avg Loss: 1.68888385, Time: 0.0403 Steps: 49980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001023, Sample Num: 16368, Cur Loss: 1.37809324, Cur Avg Loss: 1.31959195, Log Avg loss: 1.05592535, Global Avg Loss: 1.68875724, Time: 0.0430 Steps: 49990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001033, Sample Num: 16528, Cur Loss: 1.32789218, Cur Avg Loss: 1.32315524, Log Avg loss: 1.68767996, Global Avg Loss: 1.68875702, Time: 0.0412 Steps: 50000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001043, Sample Num: 16688, Cur Loss: 1.81022048, Cur Avg Loss: 1.32301853, Log Avg loss: 1.30889661, Global Avg Loss: 1.68868107, Time: 0.0533 Steps: 50010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001053, Sample Num: 16848, Cur Loss: 2.02943611, Cur Avg Loss: 1.32399777, Log Avg loss: 1.42613291, Global Avg Loss: 1.68862858, Time: 0.0747 Steps: 50020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001063, Sample Num: 17008, Cur Loss: 1.48946488, Cur Avg Loss: 1.32510036, Log Avg loss: 1.44120256, Global Avg Loss: 1.68857912, Time: 0.0508 Steps: 50030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001073, Sample Num: 17168, Cur Loss: 1.96955609, Cur Avg Loss: 1.32788646, Log Avg loss: 1.62404886, Global Avg Loss: 1.68856623, Time: 0.0851 Steps: 50040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001083, Sample Num: 17328, Cur Loss: 1.45623446, Cur Avg Loss: 1.32864024, Log Avg loss: 1.40952077, Global Avg Loss: 1.68851047, Time: 0.0403 Steps: 50050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001093, Sample Num: 17488, Cur Loss: 2.47082233, Cur Avg Loss: 1.32872641, Log Avg loss: 1.33805910, Global Avg Loss: 1.68844047, Time: 0.0775 Steps: 50060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001103, Sample Num: 17648, Cur Loss: 0.94723117, Cur Avg Loss: 1.32740466, Log Avg loss: 1.18293727, Global Avg Loss: 1.68833951, Time: 0.0827 Steps: 50070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001113, Sample Num: 17808, Cur Loss: 1.02176964, Cur Avg Loss: 1.32709284, Log Avg loss: 1.29269890, Global Avg Loss: 1.68826050, Time: 0.0646 Steps: 50080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001123, Sample Num: 17968, Cur Loss: 3.15430975, Cur Avg Loss: 1.32848120, Log Avg loss: 1.48300531, Global Avg Loss: 1.68821953, Time: 0.0402 Steps: 50090, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001133, Sample Num: 18128, Cur Loss: 0.54517603, Cur Avg Loss: 1.32668346, Log Avg loss: 1.12479732, Global Avg Loss: 1.68810707, Time: 0.0403 Steps: 50100, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001143, Sample Num: 18288, Cur Loss: 2.29704714, Cur Avg Loss: 1.33039954, Log Avg loss: 1.75143175, Global Avg Loss: 1.68811970, Time: 0.0403 Steps: 50110, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001153, Sample Num: 18448, Cur Loss: 1.80294824, Cur Avg Loss: 1.32977424, Log Avg loss: 1.25830196, Global Avg Loss: 1.68803395, Time: 0.0402 Steps: 50120, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001163, Sample Num: 18608, Cur Loss: 0.98120791, Cur Avg Loss: 1.32928565, Log Avg loss: 1.27295144, Global Avg Loss: 1.68795115, Time: 0.0403 Steps: 50130, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001173, Sample Num: 18768, Cur Loss: 1.19114387, Cur Avg Loss: 1.33089923, Log Avg loss: 1.51855843, Global Avg Loss: 1.68791736, Time: 0.0402 Steps: 50140, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001183, Sample Num: 18928, Cur Loss: 1.08020258, Cur Avg Loss: 1.33020555, Log Avg loss: 1.24883707, Global Avg Loss: 1.68782981, Time: 0.0403 Steps: 50150, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001193, Sample Num: 19088, Cur Loss: 0.80479503, Cur Avg Loss: 1.33098209, Log Avg loss: 1.42284662, Global Avg Loss: 1.68777698, Time: 0.0403 Steps: 50160, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001203, Sample Num: 19248, Cur Loss: 0.66201532, Cur Avg Loss: 1.32815842, Log Avg loss: 0.99129561, Global Avg Loss: 1.68763816, Time: 0.0403 Steps: 50170, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001213, Sample Num: 19408, Cur Loss: 2.58863044, Cur Avg Loss: 1.32430046, Log Avg loss: 0.86018677, Global Avg Loss: 1.68747326, Time: 0.0419 Steps: 50180, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001223, Sample Num: 19568, Cur Loss: 1.87785399, Cur Avg Loss: 1.32647504, Log Avg loss: 1.59025232, Global Avg Loss: 1.68745389, Time: 0.0465 Steps: 50190, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001233, Sample Num: 19728, Cur Loss: 2.35294747, Cur Avg Loss: 1.32769361, Log Avg loss: 1.47672422, Global Avg Loss: 1.68741191, Time: 0.0402 Steps: 50200, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001243, Sample Num: 19888, Cur Loss: 0.91846162, Cur Avg Loss: 1.32528891, Log Avg loss: 1.02878982, Global Avg Loss: 1.68728074, Time: 0.0403 Steps: 50210, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001253, Sample Num: 20048, Cur Loss: 0.58333629, Cur Avg Loss: 1.32664422, Log Avg loss: 1.49510952, Global Avg Loss: 1.68724247, Time: 0.0402 Steps: 50220, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001263, Sample Num: 20208, Cur Loss: 1.02195728, Cur Avg Loss: 1.32651751, Log Avg loss: 1.31063984, Global Avg Loss: 1.68716750, Time: 0.0403 Steps: 50230, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001273, Sample Num: 20368, Cur Loss: 1.56962538, Cur Avg Loss: 1.32872215, Log Avg loss: 1.60716869, Global Avg Loss: 1.68715157, Time: 0.0403 Steps: 50240, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001283, Sample Num: 20528, Cur Loss: 0.79526043, Cur Avg Loss: 1.32738870, Log Avg loss: 1.15764092, Global Avg Loss: 1.68704620, Time: 0.0402 Steps: 50250, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001293, Sample Num: 20688, Cur Loss: 0.74672854, Cur Avg Loss: 1.32612758, Log Avg loss: 1.16432611, Global Avg Loss: 1.68694219, Time: 0.0402 Steps: 50260, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001303, Sample Num: 20848, Cur Loss: 0.53555846, Cur Avg Loss: 1.32367259, Log Avg loss: 1.00624194, Global Avg Loss: 1.68680679, Time: 0.0403 Steps: 50270, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001313, Sample Num: 21008, Cur Loss: 2.33931470, Cur Avg Loss: 1.32761773, Log Avg loss: 1.84166961, Global Avg Loss: 1.68683759, Time: 0.0403 Steps: 50280, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001323, Sample Num: 21168, Cur Loss: 1.13936257, Cur Avg Loss: 1.32862640, Log Avg loss: 1.46106467, Global Avg Loss: 1.68679269, Time: 0.0403 Steps: 50290, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001333, Sample Num: 21328, Cur Loss: 0.84374845, Cur Avg Loss: 1.32944085, Log Avg loss: 1.43719196, Global Avg Loss: 1.68674307, Time: 0.0403 Steps: 50300, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001343, Sample Num: 21488, Cur Loss: 1.69880676, Cur Avg Loss: 1.33254026, Log Avg loss: 1.74569177, Global Avg Loss: 1.68675479, Time: 0.0403 Steps: 50310, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001353, Sample Num: 21648, Cur Loss: 0.98411369, Cur Avg Loss: 1.33188268, Log Avg loss: 1.24356968, Global Avg Loss: 1.68666671, Time: 0.0403 Steps: 50320, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001363, Sample Num: 21808, Cur Loss: 0.35921642, Cur Avg Loss: 1.32955875, Log Avg loss: 1.01513127, Global Avg Loss: 1.68653329, Time: 0.0402 Steps: 50330, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001373, Sample Num: 21968, Cur Loss: 0.68292445, Cur Avg Loss: 1.32897362, Log Avg loss: 1.24922023, Global Avg Loss: 1.68644641, Time: 0.0403 Steps: 50340, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001383, Sample Num: 22128, Cur Loss: 2.03058147, Cur Avg Loss: 1.33159473, Log Avg loss: 1.69147347, Global Avg Loss: 1.68644741, Time: 0.0402 Steps: 50350, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001393, Sample Num: 22288, Cur Loss: 3.11138749, Cur Avg Loss: 1.33197653, Log Avg loss: 1.38477904, Global Avg Loss: 1.68638751, Time: 0.0403 Steps: 50360, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001403, Sample Num: 22448, Cur Loss: 0.90402228, Cur Avg Loss: 1.33148895, Log Avg loss: 1.26356964, Global Avg Loss: 1.68630357, Time: 0.0402 Steps: 50370, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001413, Sample Num: 22608, Cur Loss: 1.10507500, Cur Avg Loss: 1.33230581, Log Avg loss: 1.44691094, Global Avg Loss: 1.68625605, Time: 0.0402 Steps: 50380, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001423, Sample Num: 22768, Cur Loss: 0.52456409, Cur Avg Loss: 1.33354156, Log Avg loss: 1.50815361, Global Avg Loss: 1.68622071, Time: 0.0403 Steps: 50390, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001433, Sample Num: 22928, Cur Loss: 2.51024270, Cur Avg Loss: 1.33502538, Log Avg loss: 1.54617221, Global Avg Loss: 1.68619292, Time: 0.0403 Steps: 50400, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001443, Sample Num: 23088, Cur Loss: 1.92722368, Cur Avg Loss: 1.33698460, Log Avg loss: 1.61774137, Global Avg Loss: 1.68617934, Time: 0.0403 Steps: 50410, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001453, Sample Num: 23248, Cur Loss: 0.81608784, Cur Avg Loss: 1.33619549, Log Avg loss: 1.22232643, Global Avg Loss: 1.68608734, Time: 0.0403 Steps: 50420, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001463, Sample Num: 23408, Cur Loss: 1.96561122, Cur Avg Loss: 1.33676525, Log Avg loss: 1.41955169, Global Avg Loss: 1.68603449, Time: 0.0402 Steps: 50430, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001473, Sample Num: 23568, Cur Loss: 1.28396177, Cur Avg Loss: 1.33776428, Log Avg loss: 1.48392180, Global Avg Loss: 1.68599442, Time: 0.0402 Steps: 50440, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001483, Sample Num: 23728, Cur Loss: 0.99818218, Cur Avg Loss: 1.33781059, Log Avg loss: 1.34463203, Global Avg Loss: 1.68592676, Time: 0.0402 Steps: 50450, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001493, Sample Num: 23888, Cur Loss: 0.83459938, Cur Avg Loss: 1.33615067, Log Avg loss: 1.08998503, Global Avg Loss: 1.68580865, Time: 0.0403 Steps: 50460, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001503, Sample Num: 24048, Cur Loss: 1.13526917, Cur Avg Loss: 1.33753867, Log Avg loss: 1.54476694, Global Avg Loss: 1.68578071, Time: 0.0402 Steps: 50470, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001513, Sample Num: 24208, Cur Loss: 1.69089746, Cur Avg Loss: 1.33585898, Log Avg loss: 1.08340143, Global Avg Loss: 1.68566138, Time: 0.0403 Steps: 50480, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001523, Sample Num: 24368, Cur Loss: 2.57348037, Cur Avg Loss: 1.33503608, Log Avg loss: 1.21053123, Global Avg Loss: 1.68556727, Time: 0.0402 Steps: 50490, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001533, Sample Num: 24528, Cur Loss: 1.40311873, Cur Avg Loss: 1.33411047, Log Avg loss: 1.19313961, Global Avg Loss: 1.68546976, Time: 0.0402 Steps: 50500, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001543, Sample Num: 24688, Cur Loss: 0.97726107, Cur Avg Loss: 1.33346771, Log Avg loss: 1.23493327, Global Avg Loss: 1.68538057, Time: 0.0402 Steps: 50510, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001553, Sample Num: 24848, Cur Loss: 0.41000226, Cur Avg Loss: 1.33453591, Log Avg loss: 1.49935850, Global Avg Loss: 1.68534375, Time: 0.0402 Steps: 50520, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001563, Sample Num: 25008, Cur Loss: 1.85406184, Cur Avg Loss: 1.33369051, Log Avg loss: 1.20240066, Global Avg Loss: 1.68524817, Time: 0.0402 Steps: 50530, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001573, Sample Num: 25168, Cur Loss: 1.77426577, Cur Avg Loss: 1.33595269, Log Avg loss: 1.68953069, Global Avg Loss: 1.68524902, Time: 0.0402 Steps: 50540, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001583, Sample Num: 25328, Cur Loss: 1.48562503, Cur Avg Loss: 1.33496603, Log Avg loss: 1.17976468, Global Avg Loss: 1.68514902, Time: 0.0402 Steps: 50550, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001593, Sample Num: 25488, Cur Loss: 1.69672811, Cur Avg Loss: 1.33499257, Log Avg loss: 1.33919459, Global Avg Loss: 1.68508060, Time: 0.0402 Steps: 50560, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001603, Sample Num: 25648, Cur Loss: 1.51858640, Cur Avg Loss: 1.33582280, Log Avg loss: 1.46807888, Global Avg Loss: 1.68503768, Time: 0.0402 Steps: 50570, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001613, Sample Num: 25808, Cur Loss: 2.27533245, Cur Avg Loss: 1.33584507, Log Avg loss: 1.33941363, Global Avg Loss: 1.68496935, Time: 0.0402 Steps: 50580, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001623, Sample Num: 25968, Cur Loss: 1.27394879, Cur Avg Loss: 1.33554155, Log Avg loss: 1.28658417, Global Avg Loss: 1.68489060, Time: 0.0402 Steps: 50590, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001633, Sample Num: 26128, Cur Loss: 1.33025396, Cur Avg Loss: 1.33482886, Log Avg loss: 1.21916007, Global Avg Loss: 1.68479856, Time: 0.0402 Steps: 50600, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001643, Sample Num: 26288, Cur Loss: 1.09625745, Cur Avg Loss: 1.33451425, Log Avg loss: 1.28313772, Global Avg Loss: 1.68471920, Time: 0.0402 Steps: 50610, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001653, Sample Num: 26448, Cur Loss: 1.42313528, Cur Avg Loss: 1.33316376, Log Avg loss: 1.11127899, Global Avg Loss: 1.68460592, Time: 0.0402 Steps: 50620, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001663, Sample Num: 26608, Cur Loss: 1.41013753, Cur Avg Loss: 1.33420607, Log Avg loss: 1.50649971, Global Avg Loss: 1.68457074, Time: 0.0402 Steps: 50630, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001673, Sample Num: 26768, Cur Loss: 0.65871787, Cur Avg Loss: 1.33219105, Log Avg loss: 0.99709257, Global Avg Loss: 1.68443498, Time: 0.0402 Steps: 50640, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001683, Sample Num: 26928, Cur Loss: 1.28669322, Cur Avg Loss: 1.33337825, Log Avg loss: 1.53199637, Global Avg Loss: 1.68440488, Time: 0.0402 Steps: 50650, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001693, Sample Num: 27088, Cur Loss: 1.98516667, Cur Avg Loss: 1.33421533, Log Avg loss: 1.47509586, Global Avg Loss: 1.68436357, Time: 0.0402 Steps: 50660, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001703, Sample Num: 27248, Cur Loss: 3.10633278, Cur Avg Loss: 1.33381199, Log Avg loss: 1.26552755, Global Avg Loss: 1.68428091, Time: 0.0402 Steps: 50670, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001713, Sample Num: 27408, Cur Loss: 1.14307451, Cur Avg Loss: 1.33511099, Log Avg loss: 1.55632966, Global Avg Loss: 1.68425566, Time: 0.0402 Steps: 50680, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001723, Sample Num: 27568, Cur Loss: 0.91815931, Cur Avg Loss: 1.33507736, Log Avg loss: 1.32931809, Global Avg Loss: 1.68418564, Time: 0.0402 Steps: 50690, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001733, Sample Num: 27728, Cur Loss: 2.02591419, Cur Avg Loss: 1.33674706, Log Avg loss: 1.62443630, Global Avg Loss: 1.68417385, Time: 0.0402 Steps: 50700, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001743, Sample Num: 27888, Cur Loss: 2.13929987, Cur Avg Loss: 1.33742539, Log Avg loss: 1.45497875, Global Avg Loss: 1.68412866, Time: 0.0402 Steps: 50710, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001753, Sample Num: 28048, Cur Loss: 0.84226882, Cur Avg Loss: 1.33693886, Log Avg loss: 1.25213761, Global Avg Loss: 1.68404349, Time: 0.0402 Steps: 50720, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001763, Sample Num: 28208, Cur Loss: 0.86274624, Cur Avg Loss: 1.33579481, Log Avg loss: 1.13524207, Global Avg Loss: 1.68393530, Time: 0.0402 Steps: 50730, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001773, Sample Num: 28368, Cur Loss: 0.72658920, Cur Avg Loss: 1.33546782, Log Avg loss: 1.27782050, Global Avg Loss: 1.68385527, Time: 0.0402 Steps: 50740, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001783, Sample Num: 28528, Cur Loss: 0.84364188, Cur Avg Loss: 1.33587922, Log Avg loss: 1.40881907, Global Avg Loss: 1.68380107, Time: 0.0401 Steps: 50750, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001793, Sample Num: 28688, Cur Loss: 0.90365976, Cur Avg Loss: 1.33608464, Log Avg loss: 1.37271148, Global Avg Loss: 1.68373979, Time: 0.0403 Steps: 50760, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001803, Sample Num: 28848, Cur Loss: 1.08150995, Cur Avg Loss: 1.33553421, Log Avg loss: 1.23684158, Global Avg Loss: 1.68365176, Time: 0.0402 Steps: 50770, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001813, Sample Num: 29008, Cur Loss: 1.92848778, Cur Avg Loss: 1.33713316, Log Avg loss: 1.62542488, Global Avg Loss: 1.68364029, Time: 0.0402 Steps: 50780, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001823, Sample Num: 29168, Cur Loss: 1.38492322, Cur Avg Loss: 1.33642947, Log Avg loss: 1.20884979, Global Avg Loss: 1.68354681, Time: 0.0402 Steps: 50790, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001833, Sample Num: 29328, Cur Loss: 0.96443945, Cur Avg Loss: 1.33622325, Log Avg loss: 1.29863057, Global Avg Loss: 1.68347104, Time: 0.0402 Steps: 50800, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001843, Sample Num: 29488, Cur Loss: 3.83900309, Cur Avg Loss: 1.33705119, Log Avg loss: 1.48881164, Global Avg Loss: 1.68343273, Time: 0.0401 Steps: 50810, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001853, Sample Num: 29648, Cur Loss: 1.14349949, Cur Avg Loss: 1.33732015, Log Avg loss: 1.38689021, Global Avg Loss: 1.68337438, Time: 0.0402 Steps: 50820, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001863, Sample Num: 29808, Cur Loss: 1.54159093, Cur Avg Loss: 1.33688699, Log Avg loss: 1.25662217, Global Avg Loss: 1.68329042, Time: 0.0402 Steps: 50830, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001873, Sample Num: 29968, Cur Loss: 0.46649081, Cur Avg Loss: 1.33649910, Log Avg loss: 1.26423473, Global Avg Loss: 1.68320800, Time: 0.0402 Steps: 50840, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001883, Sample Num: 30128, Cur Loss: 1.85268056, Cur Avg Loss: 1.33689056, Log Avg loss: 1.41021062, Global Avg Loss: 1.68315431, Time: 0.0402 Steps: 50850, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001893, Sample Num: 30288, Cur Loss: 1.21593940, Cur Avg Loss: 1.33696805, Log Avg loss: 1.35155907, Global Avg Loss: 1.68308911, Time: 0.0403 Steps: 50860, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001903, Sample Num: 30448, Cur Loss: 0.98985541, Cur Avg Loss: 1.33758897, Log Avg loss: 1.45512930, Global Avg Loss: 1.68304430, Time: 0.0402 Steps: 50870, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001913, Sample Num: 30608, Cur Loss: 0.73674214, Cur Avg Loss: 1.33788086, Log Avg loss: 1.39342790, Global Avg Loss: 1.68298738, Time: 0.0403 Steps: 50880, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001923, Sample Num: 30768, Cur Loss: 1.84027719, Cur Avg Loss: 1.33748664, Log Avg loss: 1.26207255, Global Avg Loss: 1.68290467, Time: 0.0402 Steps: 50890, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001933, Sample Num: 30928, Cur Loss: 0.44260222, Cur Avg Loss: 1.33615282, Log Avg loss: 1.07965958, Global Avg Loss: 1.68278615, Time: 0.0402 Steps: 50900, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001943, Sample Num: 31088, Cur Loss: 0.90516973, Cur Avg Loss: 1.33713247, Log Avg loss: 1.52649928, Global Avg Loss: 1.68275545, Time: 0.0401 Steps: 50910, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001953, Sample Num: 31248, Cur Loss: 0.52961600, Cur Avg Loss: 1.33694456, Log Avg loss: 1.30043224, Global Avg Loss: 1.68268037, Time: 0.0402 Steps: 50920, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001963, Sample Num: 31408, Cur Loss: 3.64729881, Cur Avg Loss: 1.34041767, Log Avg loss: 2.01871712, Global Avg Loss: 1.68274635, Time: 0.0402 Steps: 50930, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001973, Sample Num: 31568, Cur Loss: 1.37492466, Cur Avg Loss: 1.34044410, Log Avg loss: 1.34563105, Global Avg Loss: 1.68268017, Time: 0.0402 Steps: 50940, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001983, Sample Num: 31728, Cur Loss: 1.57430720, Cur Avg Loss: 1.34221191, Log Avg loss: 1.69100220, Global Avg Loss: 1.68268180, Time: 0.0402 Steps: 50950, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001993, Sample Num: 31888, Cur Loss: 1.00397789, Cur Avg Loss: 1.34188908, Log Avg loss: 1.27787237, Global Avg Loss: 1.68260237, Time: 0.0403 Steps: 50960, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002003, Sample Num: 32048, Cur Loss: 0.58980328, Cur Avg Loss: 1.34130078, Log Avg loss: 1.22405102, Global Avg Loss: 1.68251240, Time: 0.0402 Steps: 50970, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002013, Sample Num: 32208, Cur Loss: 1.45779347, Cur Avg Loss: 1.34011294, Log Avg loss: 1.10218910, Global Avg Loss: 1.68239857, Time: 0.0402 Steps: 50980, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002023, Sample Num: 32368, Cur Loss: 1.48130143, Cur Avg Loss: 1.33876321, Log Avg loss: 1.06706277, Global Avg Loss: 1.68227789, Time: 0.0402 Steps: 50990, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002033, Sample Num: 32528, Cur Loss: 0.98748350, Cur Avg Loss: 1.33979826, Log Avg loss: 1.54918963, Global Avg Loss: 1.68225180, Time: 0.0402 Steps: 51000, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002043, Sample Num: 32688, Cur Loss: 1.10638678, Cur Avg Loss: 1.33942751, Log Avg loss: 1.26405302, Global Avg Loss: 1.68216981, Time: 0.0401 Steps: 51010, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002053, Sample Num: 32848, Cur Loss: 0.65220469, Cur Avg Loss: 1.33991902, Log Avg loss: 1.44033523, Global Avg Loss: 1.68212241, Time: 0.0405 Steps: 51020, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002063, Sample Num: 33008, Cur Loss: 0.89651537, Cur Avg Loss: 1.33964827, Log Avg loss: 1.28406361, Global Avg Loss: 1.68204441, Time: 0.0401 Steps: 51030, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002073, Sample Num: 33168, Cur Loss: 0.98030752, Cur Avg Loss: 1.34008178, Log Avg loss: 1.42951354, Global Avg Loss: 1.68199493, Time: 0.0409 Steps: 51040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002083, Sample Num: 33328, Cur Loss: 0.45270446, Cur Avg Loss: 1.33876937, Log Avg loss: 1.06670815, Global Avg Loss: 1.68187440, Time: 0.0403 Steps: 51050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002093, Sample Num: 33488, Cur Loss: 0.26124012, Cur Avg Loss: 1.33688489, Log Avg loss: 0.94434626, Global Avg Loss: 1.68172996, Time: 0.0414 Steps: 51060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002103, Sample Num: 33648, Cur Loss: 0.66335791, Cur Avg Loss: 1.33772137, Log Avg loss: 1.51279761, Global Avg Loss: 1.68169688, Time: 0.0402 Steps: 51070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002113, Sample Num: 33808, Cur Loss: 1.18137681, Cur Avg Loss: 1.33782991, Log Avg loss: 1.36065553, Global Avg Loss: 1.68163403, Time: 0.0401 Steps: 51080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002123, Sample Num: 33968, Cur Loss: 1.31255817, Cur Avg Loss: 1.33915565, Log Avg loss: 1.61928342, Global Avg Loss: 1.68162183, Time: 0.0600 Steps: 51090, Updated lr: 0.000052 ***** Running evaluation checkpoint-51096 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-51096 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 92.779663, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.507502, "eval_total_loss": 1059.774202, "eval_mae": 1.098753, "eval_mse": 1.507251, "eval_r2": 0.041892, "eval_sp_statistic": 0.351859, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.453098, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.538352, "test_total_loss": 772.252689, "test_mae": 1.117085, "test_mse": 1.538426, "test_r2": 0.007086, "test_sp_statistic": 0.371519, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.46892, "test_ps_pvalue": 0.0, "lr": 5.249312470365102e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6816205645144262, "train_cur_epoch_loss": 2853.0526246875525, "train_cur_epoch_avg_loss": 1.3400904766028898, "train_cur_epoch_time": 92.7796630859375, "train_cur_epoch_avg_time": 0.04357898688865078, "epoch": 24, "step": 51096} ################################################## Training, Epoch: 0025, Batch: 000004, Sample Num: 64, Cur Loss: 1.75988507, Cur Avg Loss: 1.65531540, Log Avg loss: 1.66464515, Global Avg Loss: 1.68161851, Time: 0.0407 Steps: 51100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000014, Sample Num: 224, Cur Loss: 1.05087328, Cur Avg Loss: 1.23455546, Log Avg loss: 1.06625148, Global Avg Loss: 1.68149810, Time: 0.0484 Steps: 51110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000024, Sample Num: 384, Cur Loss: 1.26233518, Cur Avg Loss: 1.35881640, Log Avg loss: 1.53278171, Global Avg Loss: 1.68146901, Time: 0.0522 Steps: 51120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000034, Sample Num: 544, Cur Loss: 1.55870962, Cur Avg Loss: 1.42719054, Log Avg loss: 1.59128850, Global Avg Loss: 1.68145138, Time: 0.0503 Steps: 51130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000044, Sample Num: 704, Cur Loss: 3.12423563, Cur Avg Loss: 1.43598222, Log Avg loss: 1.46587390, Global Avg Loss: 1.68140922, Time: 0.0402 Steps: 51140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000054, Sample Num: 864, Cur Loss: 1.06349611, Cur Avg Loss: 1.45918269, Log Avg loss: 1.56126477, Global Avg Loss: 1.68138573, Time: 0.0402 Steps: 51150, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000064, Sample Num: 1024, Cur Loss: 0.46147740, Cur Avg Loss: 1.44463541, Log Avg loss: 1.36608011, Global Avg Loss: 1.68132410, Time: 0.0403 Steps: 51160, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000074, Sample Num: 1184, Cur Loss: 2.90452194, Cur Avg Loss: 1.45574255, Log Avg loss: 1.52682825, Global Avg Loss: 1.68129391, Time: 0.0402 Steps: 51170, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000084, Sample Num: 1344, Cur Loss: 2.06469202, Cur Avg Loss: 1.42954297, Log Avg loss: 1.23566609, Global Avg Loss: 1.68120684, Time: 0.0402 Steps: 51180, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000094, Sample Num: 1504, Cur Loss: 2.34791803, Cur Avg Loss: 1.43814856, Log Avg loss: 1.51043548, Global Avg Loss: 1.68117348, Time: 0.0402 Steps: 51190, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000104, Sample Num: 1664, Cur Loss: 0.73545641, Cur Avg Loss: 1.41356446, Log Avg loss: 1.18247391, Global Avg Loss: 1.68107608, Time: 0.0402 Steps: 51200, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000114, Sample Num: 1824, Cur Loss: 2.38064432, Cur Avg Loss: 1.43536518, Log Avg loss: 1.66209264, Global Avg Loss: 1.68107237, Time: 0.0402 Steps: 51210, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000124, Sample Num: 1984, Cur Loss: 1.16568112, Cur Avg Loss: 1.41848056, Log Avg loss: 1.22599596, Global Avg Loss: 1.68098352, Time: 0.0402 Steps: 51220, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000134, Sample Num: 2144, Cur Loss: 2.46376753, Cur Avg Loss: 1.41116496, Log Avg loss: 1.32045146, Global Avg Loss: 1.68091315, Time: 0.0402 Steps: 51230, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000144, Sample Num: 2304, Cur Loss: 1.62177575, Cur Avg Loss: 1.42266236, Log Avg loss: 1.57672749, Global Avg Loss: 1.68089281, Time: 0.0402 Steps: 51240, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000154, Sample Num: 2464, Cur Loss: 1.29998612, Cur Avg Loss: 1.41496064, Log Avg loss: 1.30405597, Global Avg Loss: 1.68081928, Time: 0.0402 Steps: 51250, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000164, Sample Num: 2624, Cur Loss: 0.64021206, Cur Avg Loss: 1.38616306, Log Avg loss: 0.94268022, Global Avg Loss: 1.68067528, Time: 0.0402 Steps: 51260, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000174, Sample Num: 2784, Cur Loss: 1.62280238, Cur Avg Loss: 1.39260328, Log Avg loss: 1.49822291, Global Avg Loss: 1.68063970, Time: 0.0402 Steps: 51270, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000184, Sample Num: 2944, Cur Loss: 0.76559150, Cur Avg Loss: 1.39833536, Log Avg loss: 1.49807365, Global Avg Loss: 1.68060410, Time: 0.0402 Steps: 51280, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000194, Sample Num: 3104, Cur Loss: 1.06762898, Cur Avg Loss: 1.38099948, Log Avg loss: 1.06201929, Global Avg Loss: 1.68048349, Time: 0.0402 Steps: 51290, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000204, Sample Num: 3264, Cur Loss: 1.44236994, Cur Avg Loss: 1.38075587, Log Avg loss: 1.37602972, Global Avg Loss: 1.68042414, Time: 0.0402 Steps: 51300, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000214, Sample Num: 3424, Cur Loss: 1.62964141, Cur Avg Loss: 1.38212402, Log Avg loss: 1.41003428, Global Avg Loss: 1.68037145, Time: 0.0401 Steps: 51310, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000224, Sample Num: 3584, Cur Loss: 0.47677806, Cur Avg Loss: 1.38002437, Log Avg loss: 1.33509199, Global Avg Loss: 1.68030417, Time: 0.0402 Steps: 51320, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000234, Sample Num: 3744, Cur Loss: 1.36150289, Cur Avg Loss: 1.37547763, Log Avg loss: 1.27363052, Global Avg Loss: 1.68022494, Time: 0.0401 Steps: 51330, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000244, Sample Num: 3904, Cur Loss: 1.03942347, Cur Avg Loss: 1.36500306, Log Avg loss: 1.11989828, Global Avg Loss: 1.68011580, Time: 0.0402 Steps: 51340, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000254, Sample Num: 4064, Cur Loss: 0.84208649, Cur Avg Loss: 1.37510961, Log Avg loss: 1.62170932, Global Avg Loss: 1.68010442, Time: 0.0402 Steps: 51350, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000264, Sample Num: 4224, Cur Loss: 1.01678514, Cur Avg Loss: 1.38047740, Log Avg loss: 1.51681932, Global Avg Loss: 1.68007263, Time: 0.0456 Steps: 51360, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000274, Sample Num: 4384, Cur Loss: 0.68249667, Cur Avg Loss: 1.38314732, Log Avg loss: 1.45363327, Global Avg Loss: 1.68002855, Time: 0.0523 Steps: 51370, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000284, Sample Num: 4544, Cur Loss: 1.78833807, Cur Avg Loss: 1.38768008, Log Avg loss: 1.51187754, Global Avg Loss: 1.67999583, Time: 0.0407 Steps: 51380, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000294, Sample Num: 4704, Cur Loss: 1.43994689, Cur Avg Loss: 1.39085233, Log Avg loss: 1.48094435, Global Avg Loss: 1.67995709, Time: 0.0557 Steps: 51390, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000304, Sample Num: 4864, Cur Loss: 1.11296952, Cur Avg Loss: 1.38627006, Log Avg loss: 1.25155140, Global Avg Loss: 1.67987374, Time: 0.0506 Steps: 51400, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000314, Sample Num: 5024, Cur Loss: 1.37433553, Cur Avg Loss: 1.38845364, Log Avg loss: 1.45483439, Global Avg Loss: 1.67982997, Time: 0.0514 Steps: 51410, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000324, Sample Num: 5184, Cur Loss: 1.22883034, Cur Avg Loss: 1.38842727, Log Avg loss: 1.38759915, Global Avg Loss: 1.67977314, Time: 0.0402 Steps: 51420, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000334, Sample Num: 5344, Cur Loss: 1.11189663, Cur Avg Loss: 1.39182237, Log Avg loss: 1.50182379, Global Avg Loss: 1.67973854, Time: 0.0402 Steps: 51430, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000344, Sample Num: 5504, Cur Loss: 0.85220182, Cur Avg Loss: 1.38083322, Log Avg loss: 1.01379541, Global Avg Loss: 1.67960908, Time: 0.0403 Steps: 51440, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000354, Sample Num: 5664, Cur Loss: 0.91414821, Cur Avg Loss: 1.36899715, Log Avg loss: 0.96183640, Global Avg Loss: 1.67946957, Time: 0.0402 Steps: 51450, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000364, Sample Num: 5824, Cur Loss: 1.31832612, Cur Avg Loss: 1.37132292, Log Avg loss: 1.45365528, Global Avg Loss: 1.67942569, Time: 0.0402 Steps: 51460, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000374, Sample Num: 5984, Cur Loss: 1.78304768, Cur Avg Loss: 1.38126284, Log Avg loss: 1.74307581, Global Avg Loss: 1.67943805, Time: 0.0402 Steps: 51470, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000384, Sample Num: 6144, Cur Loss: 1.92221248, Cur Avg Loss: 1.38342989, Log Avg loss: 1.46447740, Global Avg Loss: 1.67939630, Time: 0.0402 Steps: 51480, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000394, Sample Num: 6304, Cur Loss: 1.50071478, Cur Avg Loss: 1.39032172, Log Avg loss: 1.65496821, Global Avg Loss: 1.67939155, Time: 0.0402 Steps: 51490, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000404, Sample Num: 6464, Cur Loss: 1.30796397, Cur Avg Loss: 1.39314300, Log Avg loss: 1.50430127, Global Avg Loss: 1.67935756, Time: 0.0402 Steps: 51500, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000414, Sample Num: 6624, Cur Loss: 3.38499069, Cur Avg Loss: 1.39337775, Log Avg loss: 1.40286176, Global Avg Loss: 1.67930388, Time: 0.0402 Steps: 51510, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000424, Sample Num: 6784, Cur Loss: 2.28174305, Cur Avg Loss: 1.39031713, Log Avg loss: 1.26360735, Global Avg Loss: 1.67922319, Time: 0.0402 Steps: 51520, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000434, Sample Num: 6944, Cur Loss: 1.15211987, Cur Avg Loss: 1.38816848, Log Avg loss: 1.29706581, Global Avg Loss: 1.67914903, Time: 0.0402 Steps: 51530, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000444, Sample Num: 7104, Cur Loss: 1.53626847, Cur Avg Loss: 1.38252030, Log Avg loss: 1.13738943, Global Avg Loss: 1.67904391, Time: 0.0402 Steps: 51540, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000454, Sample Num: 7264, Cur Loss: 1.77101731, Cur Avg Loss: 1.37535824, Log Avg loss: 1.05736285, Global Avg Loss: 1.67892332, Time: 0.0402 Steps: 51550, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000464, Sample Num: 7424, Cur Loss: 1.25817013, Cur Avg Loss: 1.37365353, Log Avg loss: 1.29625952, Global Avg Loss: 1.67884910, Time: 0.0402 Steps: 51560, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000474, Sample Num: 7584, Cur Loss: 0.95855415, Cur Avg Loss: 1.37120208, Log Avg loss: 1.25745499, Global Avg Loss: 1.67876739, Time: 0.0402 Steps: 51570, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000484, Sample Num: 7744, Cur Loss: 1.97404480, Cur Avg Loss: 1.35967650, Log Avg loss: 0.81336391, Global Avg Loss: 1.67859961, Time: 0.0402 Steps: 51580, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000494, Sample Num: 7904, Cur Loss: 0.82929653, Cur Avg Loss: 1.36299497, Log Avg loss: 1.52360880, Global Avg Loss: 1.67856957, Time: 0.0401 Steps: 51590, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000504, Sample Num: 8064, Cur Loss: 1.42206478, Cur Avg Loss: 1.35797702, Log Avg loss: 1.11009039, Global Avg Loss: 1.67845940, Time: 0.0402 Steps: 51600, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000514, Sample Num: 8224, Cur Loss: 0.85617161, Cur Avg Loss: 1.35184849, Log Avg loss: 1.04297068, Global Avg Loss: 1.67833626, Time: 0.0402 Steps: 51610, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000524, Sample Num: 8384, Cur Loss: 1.11969209, Cur Avg Loss: 1.34965238, Log Avg loss: 1.23677233, Global Avg Loss: 1.67825072, Time: 0.0402 Steps: 51620, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000534, Sample Num: 8544, Cur Loss: 1.59102428, Cur Avg Loss: 1.34740309, Log Avg loss: 1.22954025, Global Avg Loss: 1.67816381, Time: 0.0402 Steps: 51630, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000544, Sample Num: 8704, Cur Loss: 0.65291476, Cur Avg Loss: 1.34771451, Log Avg loss: 1.36434405, Global Avg Loss: 1.67810304, Time: 0.0402 Steps: 51640, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000554, Sample Num: 8864, Cur Loss: 1.16631794, Cur Avg Loss: 1.34629079, Log Avg loss: 1.26884065, Global Avg Loss: 1.67802380, Time: 0.0560 Steps: 51650, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000564, Sample Num: 9024, Cur Loss: 1.12798154, Cur Avg Loss: 1.34380408, Log Avg loss: 1.20604048, Global Avg Loss: 1.67793244, Time: 0.0489 Steps: 51660, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000574, Sample Num: 9184, Cur Loss: 1.30270910, Cur Avg Loss: 1.34442248, Log Avg loss: 1.37929978, Global Avg Loss: 1.67787464, Time: 0.0407 Steps: 51670, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000584, Sample Num: 9344, Cur Loss: 0.62420237, Cur Avg Loss: 1.34468186, Log Avg loss: 1.35957059, Global Avg Loss: 1.67781305, Time: 0.0588 Steps: 51680, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000594, Sample Num: 9504, Cur Loss: 0.94363141, Cur Avg Loss: 1.34627257, Log Avg loss: 1.43917002, Global Avg Loss: 1.67776688, Time: 0.0516 Steps: 51690, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000604, Sample Num: 9664, Cur Loss: 0.43676668, Cur Avg Loss: 1.33834920, Log Avg loss: 0.86770082, Global Avg Loss: 1.67761020, Time: 0.0407 Steps: 51700, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000614, Sample Num: 9824, Cur Loss: 1.06239963, Cur Avg Loss: 1.34043695, Log Avg loss: 1.46653750, Global Avg Loss: 1.67756938, Time: 0.0521 Steps: 51710, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000624, Sample Num: 9984, Cur Loss: 0.58165878, Cur Avg Loss: 1.33797569, Log Avg loss: 1.18685407, Global Avg Loss: 1.67747450, Time: 0.0402 Steps: 51720, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000634, Sample Num: 10144, Cur Loss: 1.81184709, Cur Avg Loss: 1.33975236, Log Avg loss: 1.45061684, Global Avg Loss: 1.67743065, Time: 0.0402 Steps: 51730, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000644, Sample Num: 10304, Cur Loss: 1.50592196, Cur Avg Loss: 1.34206120, Log Avg loss: 1.48844160, Global Avg Loss: 1.67739412, Time: 0.0402 Steps: 51740, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000654, Sample Num: 10464, Cur Loss: 1.35412800, Cur Avg Loss: 1.34189870, Log Avg loss: 1.33143362, Global Avg Loss: 1.67732727, Time: 0.0402 Steps: 51750, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000664, Sample Num: 10624, Cur Loss: 0.99638373, Cur Avg Loss: 1.34040509, Log Avg loss: 1.24272256, Global Avg Loss: 1.67724330, Time: 0.0402 Steps: 51760, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000674, Sample Num: 10784, Cur Loss: 2.08262730, Cur Avg Loss: 1.33695094, Log Avg loss: 1.10759531, Global Avg Loss: 1.67713327, Time: 0.0403 Steps: 51770, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000684, Sample Num: 10944, Cur Loss: 1.24575782, Cur Avg Loss: 1.33869011, Log Avg loss: 1.45591063, Global Avg Loss: 1.67709054, Time: 0.0402 Steps: 51780, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000694, Sample Num: 11104, Cur Loss: 0.58987886, Cur Avg Loss: 1.34102979, Log Avg loss: 1.50106388, Global Avg Loss: 1.67705656, Time: 0.0402 Steps: 51790, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000704, Sample Num: 11264, Cur Loss: 1.59310663, Cur Avg Loss: 1.33711025, Log Avg loss: 1.06509405, Global Avg Loss: 1.67693842, Time: 0.0402 Steps: 51800, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000714, Sample Num: 11424, Cur Loss: 0.73003983, Cur Avg Loss: 1.33903556, Log Avg loss: 1.47457755, Global Avg Loss: 1.67689936, Time: 0.0402 Steps: 51810, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000724, Sample Num: 11584, Cur Loss: 1.02102780, Cur Avg Loss: 1.33713263, Log Avg loss: 1.20126294, Global Avg Loss: 1.67680757, Time: 0.0401 Steps: 51820, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000734, Sample Num: 11744, Cur Loss: 1.73342693, Cur Avg Loss: 1.33794823, Log Avg loss: 1.39699834, Global Avg Loss: 1.67675359, Time: 0.0402 Steps: 51830, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000744, Sample Num: 11904, Cur Loss: 2.33272767, Cur Avg Loss: 1.33360279, Log Avg loss: 1.01464715, Global Avg Loss: 1.67662586, Time: 0.0402 Steps: 51840, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000754, Sample Num: 12064, Cur Loss: 0.63146603, Cur Avg Loss: 1.33031476, Log Avg loss: 1.08568548, Global Avg Loss: 1.67651189, Time: 0.0402 Steps: 51850, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000764, Sample Num: 12224, Cur Loss: 1.09648693, Cur Avg Loss: 1.33105706, Log Avg loss: 1.38702661, Global Avg Loss: 1.67645607, Time: 0.0402 Steps: 51860, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000774, Sample Num: 12384, Cur Loss: 1.57018840, Cur Avg Loss: 1.33244433, Log Avg loss: 1.43843153, Global Avg Loss: 1.67641018, Time: 0.0402 Steps: 51870, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000784, Sample Num: 12544, Cur Loss: 0.94702971, Cur Avg Loss: 1.33363807, Log Avg loss: 1.42603317, Global Avg Loss: 1.67636192, Time: 0.0402 Steps: 51880, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000794, Sample Num: 12704, Cur Loss: 1.63169622, Cur Avg Loss: 1.33426232, Log Avg loss: 1.38320406, Global Avg Loss: 1.67630543, Time: 0.0402 Steps: 51890, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000804, Sample Num: 12864, Cur Loss: 0.47658962, Cur Avg Loss: 1.33392672, Log Avg loss: 1.30727957, Global Avg Loss: 1.67623432, Time: 0.0402 Steps: 51900, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000814, Sample Num: 13024, Cur Loss: 1.56138217, Cur Avg Loss: 1.33417805, Log Avg loss: 1.35438523, Global Avg Loss: 1.67617232, Time: 0.0401 Steps: 51910, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000824, Sample Num: 13184, Cur Loss: 0.55578625, Cur Avg Loss: 1.33153420, Log Avg loss: 1.11632514, Global Avg Loss: 1.67606449, Time: 0.0402 Steps: 51920, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000834, Sample Num: 13344, Cur Loss: 0.98357362, Cur Avg Loss: 1.33020315, Log Avg loss: 1.22052408, Global Avg Loss: 1.67597677, Time: 0.0401 Steps: 51930, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000844, Sample Num: 13504, Cur Loss: 0.59505421, Cur Avg Loss: 1.33051992, Log Avg loss: 1.35693845, Global Avg Loss: 1.67591535, Time: 0.0402 Steps: 51940, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000854, Sample Num: 13664, Cur Loss: 0.49925151, Cur Avg Loss: 1.32946729, Log Avg loss: 1.24062569, Global Avg Loss: 1.67583156, Time: 0.0403 Steps: 51950, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000864, Sample Num: 13824, Cur Loss: 2.69244647, Cur Avg Loss: 1.33170106, Log Avg loss: 1.52246526, Global Avg Loss: 1.67580204, Time: 0.0812 Steps: 51960, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000874, Sample Num: 13984, Cur Loss: 1.78577232, Cur Avg Loss: 1.33230591, Log Avg loss: 1.38456511, Global Avg Loss: 1.67574600, Time: 0.0405 Steps: 51970, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000884, Sample Num: 14144, Cur Loss: 2.33717108, Cur Avg Loss: 1.33489808, Log Avg loss: 1.56145341, Global Avg Loss: 1.67572401, Time: 0.0499 Steps: 51980, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000894, Sample Num: 14304, Cur Loss: 2.42384481, Cur Avg Loss: 1.33585879, Log Avg loss: 1.42078584, Global Avg Loss: 1.67567498, Time: 0.0495 Steps: 51990, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000904, Sample Num: 14464, Cur Loss: 1.98566461, Cur Avg Loss: 1.33405505, Log Avg loss: 1.17279993, Global Avg Loss: 1.67557827, Time: 0.0457 Steps: 52000, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000914, Sample Num: 14624, Cur Loss: 1.63245654, Cur Avg Loss: 1.33422869, Log Avg loss: 1.34992579, Global Avg Loss: 1.67551566, Time: 0.0637 Steps: 52010, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000924, Sample Num: 14784, Cur Loss: 2.39119172, Cur Avg Loss: 1.33742247, Log Avg loss: 1.62933445, Global Avg Loss: 1.67550678, Time: 0.0489 Steps: 52020, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000934, Sample Num: 14944, Cur Loss: 0.82551384, Cur Avg Loss: 1.33617550, Log Avg loss: 1.22095500, Global Avg Loss: 1.67541942, Time: 0.0402 Steps: 52030, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000944, Sample Num: 15104, Cur Loss: 0.84565341, Cur Avg Loss: 1.33650180, Log Avg loss: 1.36697871, Global Avg Loss: 1.67536015, Time: 0.0403 Steps: 52040, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000954, Sample Num: 15264, Cur Loss: 0.32832992, Cur Avg Loss: 1.33522082, Log Avg loss: 1.21429614, Global Avg Loss: 1.67527157, Time: 0.0402 Steps: 52050, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000964, Sample Num: 15424, Cur Loss: 1.13854206, Cur Avg Loss: 1.33729941, Log Avg loss: 1.53559659, Global Avg Loss: 1.67524474, Time: 0.0401 Steps: 52060, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000974, Sample Num: 15584, Cur Loss: 1.68193603, Cur Avg Loss: 1.33994021, Log Avg loss: 1.59451407, Global Avg Loss: 1.67522923, Time: 0.0402 Steps: 52070, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000984, Sample Num: 15744, Cur Loss: 1.17493975, Cur Avg Loss: 1.34270261, Log Avg loss: 1.61176021, Global Avg Loss: 1.67521705, Time: 0.0402 Steps: 52080, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000994, Sample Num: 15904, Cur Loss: 1.46664691, Cur Avg Loss: 1.34486881, Log Avg loss: 1.55802240, Global Avg Loss: 1.67519455, Time: 0.0402 Steps: 52090, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001004, Sample Num: 16064, Cur Loss: 2.83796334, Cur Avg Loss: 1.34567826, Log Avg loss: 1.42613747, Global Avg Loss: 1.67514674, Time: 0.0402 Steps: 52100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001014, Sample Num: 16224, Cur Loss: 1.00512922, Cur Avg Loss: 1.34763447, Log Avg loss: 1.54403797, Global Avg Loss: 1.67512158, Time: 0.0402 Steps: 52110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001024, Sample Num: 16384, Cur Loss: 1.23809040, Cur Avg Loss: 1.34934338, Log Avg loss: 1.52262763, Global Avg Loss: 1.67509233, Time: 0.0403 Steps: 52120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001034, Sample Num: 16544, Cur Loss: 1.55406249, Cur Avg Loss: 1.34793322, Log Avg loss: 1.20353219, Global Avg Loss: 1.67500187, Time: 0.0401 Steps: 52130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001044, Sample Num: 16704, Cur Loss: 0.60197985, Cur Avg Loss: 1.34448479, Log Avg loss: 0.98791729, Global Avg Loss: 1.67487009, Time: 0.0402 Steps: 52140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001054, Sample Num: 16864, Cur Loss: 0.63142562, Cur Avg Loss: 1.34534375, Log Avg loss: 1.43501922, Global Avg Loss: 1.67482410, Time: 0.0402 Steps: 52150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001064, Sample Num: 17024, Cur Loss: 0.39129364, Cur Avg Loss: 1.34467645, Log Avg loss: 1.27434267, Global Avg Loss: 1.67474732, Time: 0.0402 Steps: 52160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001074, Sample Num: 17184, Cur Loss: 1.93367314, Cur Avg Loss: 1.34635797, Log Avg loss: 1.52527156, Global Avg Loss: 1.67471867, Time: 0.0402 Steps: 52170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001084, Sample Num: 17344, Cur Loss: 1.32352519, Cur Avg Loss: 1.34527822, Log Avg loss: 1.22931408, Global Avg Loss: 1.67463331, Time: 0.0402 Steps: 52180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001094, Sample Num: 17504, Cur Loss: 1.21234429, Cur Avg Loss: 1.34567455, Log Avg loss: 1.38863567, Global Avg Loss: 1.67457851, Time: 0.0402 Steps: 52190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001104, Sample Num: 17664, Cur Loss: 1.03885853, Cur Avg Loss: 1.34364223, Log Avg loss: 1.12130738, Global Avg Loss: 1.67447252, Time: 0.0402 Steps: 52200, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001114, Sample Num: 17824, Cur Loss: 1.61869287, Cur Avg Loss: 1.34405409, Log Avg loss: 1.38952270, Global Avg Loss: 1.67441794, Time: 0.0402 Steps: 52210, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001124, Sample Num: 17984, Cur Loss: 0.68103552, Cur Avg Loss: 1.34306979, Log Avg loss: 1.23341918, Global Avg Loss: 1.67433349, Time: 0.0402 Steps: 52220, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001134, Sample Num: 18144, Cur Loss: 0.89231527, Cur Avg Loss: 1.34283784, Log Avg loss: 1.31676658, Global Avg Loss: 1.67426503, Time: 0.0402 Steps: 52230, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001144, Sample Num: 18304, Cur Loss: 1.23777878, Cur Avg Loss: 1.34134722, Log Avg loss: 1.17231135, Global Avg Loss: 1.67416894, Time: 0.0401 Steps: 52240, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001154, Sample Num: 18464, Cur Loss: 2.88931370, Cur Avg Loss: 1.34307795, Log Avg loss: 1.54107350, Global Avg Loss: 1.67414347, Time: 0.0402 Steps: 52250, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001164, Sample Num: 18624, Cur Loss: 1.38537145, Cur Avg Loss: 1.34208501, Log Avg loss: 1.22749975, Global Avg Loss: 1.67405800, Time: 0.0407 Steps: 52260, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001174, Sample Num: 18784, Cur Loss: 1.79709327, Cur Avg Loss: 1.34148983, Log Avg loss: 1.27221026, Global Avg Loss: 1.67398113, Time: 0.0536 Steps: 52270, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001184, Sample Num: 18944, Cur Loss: 0.87248051, Cur Avg Loss: 1.34449563, Log Avg loss: 1.69737630, Global Avg Loss: 1.67398560, Time: 0.0551 Steps: 52280, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001194, Sample Num: 19104, Cur Loss: 1.12952554, Cur Avg Loss: 1.34598128, Log Avg loss: 1.52188312, Global Avg Loss: 1.67395651, Time: 0.0407 Steps: 52290, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001204, Sample Num: 19264, Cur Loss: 1.71163607, Cur Avg Loss: 1.34669813, Log Avg loss: 1.43228901, Global Avg Loss: 1.67391030, Time: 0.0405 Steps: 52300, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001214, Sample Num: 19424, Cur Loss: 0.95311856, Cur Avg Loss: 1.34670002, Log Avg loss: 1.34692807, Global Avg Loss: 1.67384780, Time: 0.0630 Steps: 52310, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001224, Sample Num: 19584, Cur Loss: 0.69944572, Cur Avg Loss: 1.34765028, Log Avg loss: 1.46301223, Global Avg Loss: 1.67380750, Time: 0.0587 Steps: 52320, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001234, Sample Num: 19744, Cur Loss: 0.43260723, Cur Avg Loss: 1.34563962, Log Avg loss: 1.09953429, Global Avg Loss: 1.67369776, Time: 0.0402 Steps: 52330, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001244, Sample Num: 19904, Cur Loss: 0.92265904, Cur Avg Loss: 1.34479798, Log Avg loss: 1.24094015, Global Avg Loss: 1.67361508, Time: 0.0402 Steps: 52340, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001254, Sample Num: 20064, Cur Loss: 1.46482301, Cur Avg Loss: 1.34551287, Log Avg loss: 1.43444481, Global Avg Loss: 1.67356939, Time: 0.0402 Steps: 52350, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001264, Sample Num: 20224, Cur Loss: 1.57803226, Cur Avg Loss: 1.34781696, Log Avg loss: 1.63675017, Global Avg Loss: 1.67356236, Time: 0.0402 Steps: 52360, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001274, Sample Num: 20384, Cur Loss: 2.80824304, Cur Avg Loss: 1.34984280, Log Avg loss: 1.60590814, Global Avg Loss: 1.67354944, Time: 0.0402 Steps: 52370, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001284, Sample Num: 20544, Cur Loss: 1.16937733, Cur Avg Loss: 1.35349625, Log Avg loss: 1.81894586, Global Avg Loss: 1.67357720, Time: 0.0402 Steps: 52380, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001294, Sample Num: 20704, Cur Loss: 1.11929882, Cur Avg Loss: 1.35206762, Log Avg loss: 1.16863232, Global Avg Loss: 1.67348081, Time: 0.0401 Steps: 52390, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001304, Sample Num: 20864, Cur Loss: 1.01808035, Cur Avg Loss: 1.35090486, Log Avg loss: 1.20044284, Global Avg Loss: 1.67339054, Time: 0.0402 Steps: 52400, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001314, Sample Num: 21024, Cur Loss: 2.74148750, Cur Avg Loss: 1.35205360, Log Avg loss: 1.50185021, Global Avg Loss: 1.67335781, Time: 0.0402 Steps: 52410, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001324, Sample Num: 21184, Cur Loss: 1.15718102, Cur Avg Loss: 1.35427796, Log Avg loss: 1.64655866, Global Avg Loss: 1.67335270, Time: 0.0403 Steps: 52420, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001334, Sample Num: 21344, Cur Loss: 0.98615986, Cur Avg Loss: 1.35450329, Log Avg loss: 1.38433734, Global Avg Loss: 1.67329757, Time: 0.0402 Steps: 52430, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001344, Sample Num: 21504, Cur Loss: 2.30931115, Cur Avg Loss: 1.35509154, Log Avg loss: 1.43356417, Global Avg Loss: 1.67325186, Time: 0.0402 Steps: 52440, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001354, Sample Num: 21664, Cur Loss: 1.16187477, Cur Avg Loss: 1.35723651, Log Avg loss: 1.64552015, Global Avg Loss: 1.67324657, Time: 0.0402 Steps: 52450, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001364, Sample Num: 21824, Cur Loss: 0.70468724, Cur Avg Loss: 1.35643213, Log Avg loss: 1.24751914, Global Avg Loss: 1.67316542, Time: 0.0403 Steps: 52460, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001374, Sample Num: 21984, Cur Loss: 1.12695730, Cur Avg Loss: 1.35735042, Log Avg loss: 1.48260437, Global Avg Loss: 1.67312910, Time: 0.0402 Steps: 52470, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001384, Sample Num: 22144, Cur Loss: 0.63082027, Cur Avg Loss: 1.35803497, Log Avg loss: 1.45209252, Global Avg Loss: 1.67308698, Time: 0.0402 Steps: 52480, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001394, Sample Num: 22304, Cur Loss: 1.35574973, Cur Avg Loss: 1.35784030, Log Avg loss: 1.33089736, Global Avg Loss: 1.67302179, Time: 0.0402 Steps: 52490, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001404, Sample Num: 22464, Cur Loss: 0.81764090, Cur Avg Loss: 1.35530277, Log Avg loss: 1.00157229, Global Avg Loss: 1.67289389, Time: 0.0403 Steps: 52500, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001414, Sample Num: 22624, Cur Loss: 2.57958722, Cur Avg Loss: 1.35761889, Log Avg loss: 1.68280103, Global Avg Loss: 1.67289578, Time: 0.0402 Steps: 52510, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001424, Sample Num: 22784, Cur Loss: 2.04445767, Cur Avg Loss: 1.35691802, Log Avg loss: 1.25781612, Global Avg Loss: 1.67281675, Time: 0.0402 Steps: 52520, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001434, Sample Num: 22944, Cur Loss: 1.30636966, Cur Avg Loss: 1.35618163, Log Avg loss: 1.25131848, Global Avg Loss: 1.67273651, Time: 0.0402 Steps: 52530, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001444, Sample Num: 23104, Cur Loss: 1.83113599, Cur Avg Loss: 1.35543123, Log Avg loss: 1.24782518, Global Avg Loss: 1.67265564, Time: 0.0403 Steps: 52540, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001454, Sample Num: 23264, Cur Loss: 2.19371104, Cur Avg Loss: 1.35769060, Log Avg loss: 1.68394237, Global Avg Loss: 1.67265778, Time: 0.0501 Steps: 52550, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001464, Sample Num: 23424, Cur Loss: 1.56960356, Cur Avg Loss: 1.35621464, Log Avg loss: 1.14161060, Global Avg Loss: 1.67255675, Time: 0.0551 Steps: 52560, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001474, Sample Num: 23584, Cur Loss: 1.70416117, Cur Avg Loss: 1.35823197, Log Avg loss: 1.65356866, Global Avg Loss: 1.67255313, Time: 0.0504 Steps: 52570, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001484, Sample Num: 23744, Cur Loss: 1.23845112, Cur Avg Loss: 1.35902909, Log Avg loss: 1.47652439, Global Avg Loss: 1.67251585, Time: 0.0457 Steps: 52580, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001494, Sample Num: 23904, Cur Loss: 1.37106693, Cur Avg Loss: 1.35938157, Log Avg loss: 1.41169098, Global Avg Loss: 1.67246626, Time: 0.0407 Steps: 52590, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001504, Sample Num: 24064, Cur Loss: 1.32254457, Cur Avg Loss: 1.36049830, Log Avg loss: 1.52733720, Global Avg Loss: 1.67243867, Time: 0.0457 Steps: 52600, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001514, Sample Num: 24224, Cur Loss: 1.06112432, Cur Avg Loss: 1.36050580, Log Avg loss: 1.36163298, Global Avg Loss: 1.67237959, Time: 0.0614 Steps: 52610, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001524, Sample Num: 24384, Cur Loss: 1.74599481, Cur Avg Loss: 1.36170515, Log Avg loss: 1.54328701, Global Avg Loss: 1.67235506, Time: 0.0402 Steps: 52620, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001534, Sample Num: 24544, Cur Loss: 1.70499146, Cur Avg Loss: 1.36120690, Log Avg loss: 1.28527410, Global Avg Loss: 1.67228151, Time: 0.0403 Steps: 52630, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001544, Sample Num: 24704, Cur Loss: 0.55017513, Cur Avg Loss: 1.36020237, Log Avg loss: 1.20610761, Global Avg Loss: 1.67219295, Time: 0.0402 Steps: 52640, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001554, Sample Num: 24864, Cur Loss: 0.29559663, Cur Avg Loss: 1.35851183, Log Avg loss: 1.09749245, Global Avg Loss: 1.67208379, Time: 0.0402 Steps: 52650, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001564, Sample Num: 25024, Cur Loss: 1.45511127, Cur Avg Loss: 1.35876942, Log Avg loss: 1.39879854, Global Avg Loss: 1.67203190, Time: 0.0402 Steps: 52660, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001574, Sample Num: 25184, Cur Loss: 0.59219170, Cur Avg Loss: 1.35780155, Log Avg loss: 1.20642605, Global Avg Loss: 1.67194350, Time: 0.0402 Steps: 52670, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001584, Sample Num: 25344, Cur Loss: 2.24352956, Cur Avg Loss: 1.35977635, Log Avg loss: 1.67061097, Global Avg Loss: 1.67194324, Time: 0.0402 Steps: 52680, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001594, Sample Num: 25504, Cur Loss: 1.11341155, Cur Avg Loss: 1.35946232, Log Avg loss: 1.30971885, Global Avg Loss: 1.67187450, Time: 0.0401 Steps: 52690, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001604, Sample Num: 25664, Cur Loss: 1.17728865, Cur Avg Loss: 1.36101433, Log Avg loss: 1.60840587, Global Avg Loss: 1.67186245, Time: 0.0402 Steps: 52700, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001614, Sample Num: 25824, Cur Loss: 2.30932975, Cur Avg Loss: 1.36009246, Log Avg loss: 1.21222360, Global Avg Loss: 1.67177525, Time: 0.0402 Steps: 52710, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001624, Sample Num: 25984, Cur Loss: 2.25950670, Cur Avg Loss: 1.36052160, Log Avg loss: 1.42978525, Global Avg Loss: 1.67172935, Time: 0.0402 Steps: 52720, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001634, Sample Num: 26144, Cur Loss: 0.75734866, Cur Avg Loss: 1.36162211, Log Avg loss: 1.54034445, Global Avg Loss: 1.67170444, Time: 0.0402 Steps: 52730, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001644, Sample Num: 26304, Cur Loss: 0.87207657, Cur Avg Loss: 1.36216324, Log Avg loss: 1.45058494, Global Avg Loss: 1.67166251, Time: 0.0402 Steps: 52740, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001654, Sample Num: 26464, Cur Loss: 1.21206808, Cur Avg Loss: 1.36184864, Log Avg loss: 1.31012873, Global Avg Loss: 1.67159397, Time: 0.0402 Steps: 52750, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001664, Sample Num: 26624, Cur Loss: 1.15291131, Cur Avg Loss: 1.36317609, Log Avg loss: 1.58273544, Global Avg Loss: 1.67157713, Time: 0.0402 Steps: 52760, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001674, Sample Num: 26784, Cur Loss: 3.41524076, Cur Avg Loss: 1.36473942, Log Avg loss: 1.62487692, Global Avg Loss: 1.67156828, Time: 0.0402 Steps: 52770, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001684, Sample Num: 26944, Cur Loss: 1.88904929, Cur Avg Loss: 1.36490875, Log Avg loss: 1.39325467, Global Avg Loss: 1.67151555, Time: 0.0404 Steps: 52780, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001694, Sample Num: 27104, Cur Loss: 0.79536605, Cur Avg Loss: 1.36113661, Log Avg loss: 0.72590832, Global Avg Loss: 1.67133642, Time: 0.0401 Steps: 52790, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001704, Sample Num: 27264, Cur Loss: 1.22379351, Cur Avg Loss: 1.36195546, Log Avg loss: 1.50066994, Global Avg Loss: 1.67130410, Time: 0.0403 Steps: 52800, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001714, Sample Num: 27424, Cur Loss: 1.47737503, Cur Avg Loss: 1.36154096, Log Avg loss: 1.29090913, Global Avg Loss: 1.67123207, Time: 0.0402 Steps: 52810, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001724, Sample Num: 27584, Cur Loss: 0.96320796, Cur Avg Loss: 1.36150252, Log Avg loss: 1.35491360, Global Avg Loss: 1.67117218, Time: 0.0403 Steps: 52820, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001734, Sample Num: 27744, Cur Loss: 2.12544394, Cur Avg Loss: 1.36130614, Log Avg loss: 1.32745156, Global Avg Loss: 1.67110712, Time: 0.0402 Steps: 52830, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001744, Sample Num: 27904, Cur Loss: 0.84953487, Cur Avg Loss: 1.36151174, Log Avg loss: 1.39716135, Global Avg Loss: 1.67105528, Time: 0.0432 Steps: 52840, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001754, Sample Num: 28064, Cur Loss: 1.45760310, Cur Avg Loss: 1.36111311, Log Avg loss: 1.29159269, Global Avg Loss: 1.67098348, Time: 0.0405 Steps: 52850, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001764, Sample Num: 28224, Cur Loss: 1.55116570, Cur Avg Loss: 1.36123936, Log Avg loss: 1.38338316, Global Avg Loss: 1.67092907, Time: 0.0462 Steps: 52860, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001774, Sample Num: 28384, Cur Loss: 0.85718417, Cur Avg Loss: 1.36002257, Log Avg loss: 1.14538054, Global Avg Loss: 1.67082967, Time: 0.0407 Steps: 52870, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001784, Sample Num: 28544, Cur Loss: 2.57650232, Cur Avg Loss: 1.35892473, Log Avg loss: 1.16416919, Global Avg Loss: 1.67073385, Time: 0.0491 Steps: 52880, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001794, Sample Num: 28704, Cur Loss: 3.20310926, Cur Avg Loss: 1.35966062, Log Avg loss: 1.49094369, Global Avg Loss: 1.67069986, Time: 0.0407 Steps: 52890, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001804, Sample Num: 28864, Cur Loss: 2.21694517, Cur Avg Loss: 1.36024037, Log Avg loss: 1.46424659, Global Avg Loss: 1.67066083, Time: 0.0552 Steps: 52900, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001814, Sample Num: 29024, Cur Loss: 0.54593796, Cur Avg Loss: 1.35909558, Log Avg loss: 1.15257623, Global Avg Loss: 1.67056291, Time: 0.0402 Steps: 52910, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001824, Sample Num: 29184, Cur Loss: 1.03665292, Cur Avg Loss: 1.35856928, Log Avg loss: 1.26309757, Global Avg Loss: 1.67048592, Time: 0.0403 Steps: 52920, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001834, Sample Num: 29344, Cur Loss: 1.16785884, Cur Avg Loss: 1.35725247, Log Avg loss: 1.11706615, Global Avg Loss: 1.67038136, Time: 0.0403 Steps: 52930, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001844, Sample Num: 29504, Cur Loss: 1.37896478, Cur Avg Loss: 1.35686259, Log Avg loss: 1.28535813, Global Avg Loss: 1.67030863, Time: 0.0402 Steps: 52940, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001854, Sample Num: 29664, Cur Loss: 1.54994619, Cur Avg Loss: 1.35641692, Log Avg loss: 1.27423590, Global Avg Loss: 1.67023383, Time: 0.0402 Steps: 52950, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001864, Sample Num: 29824, Cur Loss: 0.74939680, Cur Avg Loss: 1.35610276, Log Avg loss: 1.29785728, Global Avg Loss: 1.67016352, Time: 0.0404 Steps: 52960, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001874, Sample Num: 29984, Cur Loss: 1.49035215, Cur Avg Loss: 1.35694322, Log Avg loss: 1.51360481, Global Avg Loss: 1.67013396, Time: 0.0402 Steps: 52970, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001884, Sample Num: 30144, Cur Loss: 2.89734030, Cur Avg Loss: 1.35767957, Log Avg loss: 1.49567220, Global Avg Loss: 1.67010103, Time: 0.0402 Steps: 52980, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001894, Sample Num: 30304, Cur Loss: 0.65703797, Cur Avg Loss: 1.35774660, Log Avg loss: 1.37037526, Global Avg Loss: 1.67004447, Time: 0.0402 Steps: 52990, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001904, Sample Num: 30464, Cur Loss: 0.92428648, Cur Avg Loss: 1.35770186, Log Avg loss: 1.34922778, Global Avg Loss: 1.66998394, Time: 0.0402 Steps: 53000, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001914, Sample Num: 30624, Cur Loss: 0.32652837, Cur Avg Loss: 1.35730394, Log Avg loss: 1.28154046, Global Avg Loss: 1.66991066, Time: 0.0402 Steps: 53010, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001924, Sample Num: 30784, Cur Loss: 2.41599083, Cur Avg Loss: 1.35814335, Log Avg loss: 1.51880563, Global Avg Loss: 1.66988216, Time: 0.0402 Steps: 53020, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001934, Sample Num: 30944, Cur Loss: 1.36888289, Cur Avg Loss: 1.35844406, Log Avg loss: 1.41630079, Global Avg Loss: 1.66983434, Time: 0.0402 Steps: 53030, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001944, Sample Num: 31104, Cur Loss: 0.98260361, Cur Avg Loss: 1.35857008, Log Avg loss: 1.38294344, Global Avg Loss: 1.66978025, Time: 0.0402 Steps: 53040, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001954, Sample Num: 31264, Cur Loss: 2.46760535, Cur Avg Loss: 1.35999805, Log Avg loss: 1.63759475, Global Avg Loss: 1.66977419, Time: 0.0402 Steps: 53050, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001964, Sample Num: 31424, Cur Loss: 0.93471074, Cur Avg Loss: 1.36074734, Log Avg loss: 1.50715768, Global Avg Loss: 1.66974354, Time: 0.0403 Steps: 53060, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001974, Sample Num: 31584, Cur Loss: 0.62082797, Cur Avg Loss: 1.36008689, Log Avg loss: 1.23037569, Global Avg Loss: 1.66966075, Time: 0.0402 Steps: 53070, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001984, Sample Num: 31744, Cur Loss: 2.24263334, Cur Avg Loss: 1.36129001, Log Avg loss: 1.59878560, Global Avg Loss: 1.66964740, Time: 0.0402 Steps: 53080, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001994, Sample Num: 31904, Cur Loss: 1.72202909, Cur Avg Loss: 1.35974429, Log Avg loss: 1.05307359, Global Avg Loss: 1.66953126, Time: 0.0401 Steps: 53090, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002004, Sample Num: 32064, Cur Loss: 1.65785980, Cur Avg Loss: 1.35961864, Log Avg loss: 1.33456393, Global Avg Loss: 1.66946818, Time: 0.0402 Steps: 53100, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002014, Sample Num: 32224, Cur Loss: 2.17560363, Cur Avg Loss: 1.35996340, Log Avg loss: 1.42905353, Global Avg Loss: 1.66942291, Time: 0.0402 Steps: 53110, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002024, Sample Num: 32384, Cur Loss: 0.76753509, Cur Avg Loss: 1.35819789, Log Avg loss: 1.00262439, Global Avg Loss: 1.66929738, Time: 0.0403 Steps: 53120, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002034, Sample Num: 32544, Cur Loss: 1.66550016, Cur Avg Loss: 1.35844900, Log Avg loss: 1.40927215, Global Avg Loss: 1.66924844, Time: 0.0402 Steps: 53130, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002044, Sample Num: 32704, Cur Loss: 0.86590463, Cur Avg Loss: 1.35860834, Log Avg loss: 1.39101902, Global Avg Loss: 1.66919608, Time: 0.0402 Steps: 53140, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002054, Sample Num: 32864, Cur Loss: 0.73502076, Cur Avg Loss: 1.35753658, Log Avg loss: 1.13846945, Global Avg Loss: 1.66909623, Time: 0.0531 Steps: 53150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002064, Sample Num: 33024, Cur Loss: 0.63891840, Cur Avg Loss: 1.35888775, Log Avg loss: 1.63641801, Global Avg Loss: 1.66909008, Time: 0.0431 Steps: 53160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002074, Sample Num: 33184, Cur Loss: 1.17518091, Cur Avg Loss: 1.35935260, Log Avg loss: 1.45529718, Global Avg Loss: 1.66904987, Time: 0.0520 Steps: 53170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002084, Sample Num: 33344, Cur Loss: 0.25849104, Cur Avg Loss: 1.35735376, Log Avg loss: 0.94279517, Global Avg Loss: 1.66891331, Time: 0.0488 Steps: 53180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002094, Sample Num: 33504, Cur Loss: 0.35841024, Cur Avg Loss: 1.35689929, Log Avg loss: 1.26218601, Global Avg Loss: 1.66883684, Time: 0.0507 Steps: 53190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002104, Sample Num: 33664, Cur Loss: 1.87837791, Cur Avg Loss: 1.35700985, Log Avg loss: 1.38016243, Global Avg Loss: 1.66878258, Time: 0.0404 Steps: 53200, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002114, Sample Num: 33824, Cur Loss: 2.98682594, Cur Avg Loss: 1.35948170, Log Avg loss: 1.87955763, Global Avg Loss: 1.66882219, Time: 0.0509 Steps: 53210, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002124, Sample Num: 33984, Cur Loss: 0.43898600, Cur Avg Loss: 1.35914234, Log Avg loss: 1.28740265, Global Avg Loss: 1.66875052, Time: 0.0705 Steps: 53220, Updated lr: 0.000050 ***** Running evaluation checkpoint-53225 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-53225 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 90.075518, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.317864, "eval_total_loss": 926.45831, "eval_mae": 0.9945, "eval_mse": 1.317857, "eval_r2": 0.162284, "eval_sp_statistic": 0.38613, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.478603, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.13257, "test_total_loss": 568.549891, "test_mae": 0.753926, "test_mse": 1.132924, "test_r2": 0.268801, "test_sp_statistic": 0.412381, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.519151, "test_ps_pvalue": 0.0, "lr": 5.0474158368895205e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6686850298366507, "train_cur_epoch_loss": 2891.6763486266136, "train_cur_epoch_avg_loss": 1.3582321975700393, "train_cur_epoch_time": 90.07551789283752, "train_cur_epoch_avg_time": 0.0423088388411637, "epoch": 25, "step": 53225} ################################################## Training, Epoch: 0026, Batch: 000005, Sample Num: 80, Cur Loss: 2.29969192, Cur Avg Loss: 1.39428456, Log Avg loss: 1.18294398, Global Avg Loss: 1.66865925, Time: 0.0403 Steps: 53230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000015, Sample Num: 240, Cur Loss: 0.56431657, Cur Avg Loss: 1.50002373, Log Avg loss: 1.55289331, Global Avg Loss: 1.66863751, Time: 0.0402 Steps: 53240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000025, Sample Num: 400, Cur Loss: 0.51871216, Cur Avg Loss: 1.37234537, Log Avg loss: 1.18082784, Global Avg Loss: 1.66854590, Time: 0.0402 Steps: 53250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000035, Sample Num: 560, Cur Loss: 0.79183948, Cur Avg Loss: 1.38221464, Log Avg loss: 1.40688779, Global Avg Loss: 1.66849677, Time: 0.0402 Steps: 53260, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000045, Sample Num: 720, Cur Loss: 1.79838991, Cur Avg Loss: 1.39364090, Log Avg loss: 1.43363281, Global Avg Loss: 1.66845269, Time: 0.0403 Steps: 53270, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000055, Sample Num: 880, Cur Loss: 2.69963479, Cur Avg Loss: 1.45668000, Log Avg loss: 1.74035595, Global Avg Loss: 1.66846618, Time: 0.0402 Steps: 53280, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000065, Sample Num: 1040, Cur Loss: 0.64302647, Cur Avg Loss: 1.45155292, Log Avg loss: 1.42335398, Global Avg Loss: 1.66842018, Time: 0.0402 Steps: 53290, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000075, Sample Num: 1200, Cur Loss: 0.51158917, Cur Avg Loss: 1.43171148, Log Avg loss: 1.30274213, Global Avg Loss: 1.66835158, Time: 0.0402 Steps: 53300, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000085, Sample Num: 1360, Cur Loss: 1.88707566, Cur Avg Loss: 1.45078258, Log Avg loss: 1.59381585, Global Avg Loss: 1.66833760, Time: 0.0402 Steps: 53310, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000095, Sample Num: 1520, Cur Loss: 1.08410072, Cur Avg Loss: 1.45758933, Log Avg loss: 1.51544667, Global Avg Loss: 1.66830892, Time: 0.0402 Steps: 53320, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000105, Sample Num: 1680, Cur Loss: 0.58689535, Cur Avg Loss: 1.44769779, Log Avg loss: 1.35372817, Global Avg Loss: 1.66824993, Time: 0.0402 Steps: 53330, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000115, Sample Num: 1840, Cur Loss: 1.51298785, Cur Avg Loss: 1.46842692, Log Avg loss: 1.68608279, Global Avg Loss: 1.66825328, Time: 0.0402 Steps: 53340, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000125, Sample Num: 2000, Cur Loss: 0.65912354, Cur Avg Loss: 1.44878245, Log Avg loss: 1.22287107, Global Avg Loss: 1.66816979, Time: 0.0402 Steps: 53350, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000135, Sample Num: 2160, Cur Loss: 0.34612617, Cur Avg Loss: 1.41441322, Log Avg loss: 0.98479784, Global Avg Loss: 1.66804173, Time: 0.0402 Steps: 53360, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000145, Sample Num: 2320, Cur Loss: 2.31989360, Cur Avg Loss: 1.38734006, Log Avg loss: 1.02185237, Global Avg Loss: 1.66792065, Time: 0.0402 Steps: 53370, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000155, Sample Num: 2480, Cur Loss: 0.94517618, Cur Avg Loss: 1.39873293, Log Avg loss: 1.56392958, Global Avg Loss: 1.66790117, Time: 0.0402 Steps: 53380, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000165, Sample Num: 2640, Cur Loss: 1.41355777, Cur Avg Loss: 1.39098351, Log Avg loss: 1.27086743, Global Avg Loss: 1.66782680, Time: 0.0402 Steps: 53390, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000175, Sample Num: 2800, Cur Loss: 0.53801906, Cur Avg Loss: 1.37312777, Log Avg loss: 1.07850806, Global Avg Loss: 1.66771644, Time: 0.0402 Steps: 53400, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000185, Sample Num: 2960, Cur Loss: 1.34394979, Cur Avg Loss: 1.38698974, Log Avg loss: 1.62957420, Global Avg Loss: 1.66770930, Time: 0.0510 Steps: 53410, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000195, Sample Num: 3120, Cur Loss: 1.29546809, Cur Avg Loss: 1.38960081, Log Avg loss: 1.43790573, Global Avg Loss: 1.66766628, Time: 0.0430 Steps: 53420, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000205, Sample Num: 3280, Cur Loss: 1.16059089, Cur Avg Loss: 1.38742175, Log Avg loss: 1.34492993, Global Avg Loss: 1.66760588, Time: 0.0533 Steps: 53430, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000215, Sample Num: 3440, Cur Loss: 1.05112791, Cur Avg Loss: 1.37587442, Log Avg loss: 1.13915417, Global Avg Loss: 1.66750699, Time: 0.0489 Steps: 53440, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000225, Sample Num: 3600, Cur Loss: 1.86916471, Cur Avg Loss: 1.38390198, Log Avg loss: 1.55649460, Global Avg Loss: 1.66748622, Time: 0.0406 Steps: 53450, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000235, Sample Num: 3760, Cur Loss: 1.10003901, Cur Avg Loss: 1.37848606, Log Avg loss: 1.25662792, Global Avg Loss: 1.66740937, Time: 0.0632 Steps: 53460, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000245, Sample Num: 3920, Cur Loss: 0.71676290, Cur Avg Loss: 1.37778603, Log Avg loss: 1.36133525, Global Avg Loss: 1.66735213, Time: 0.0404 Steps: 53470, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000255, Sample Num: 4080, Cur Loss: 1.09086502, Cur Avg Loss: 1.35825857, Log Avg loss: 0.87983590, Global Avg Loss: 1.66720487, Time: 0.0402 Steps: 53480, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000265, Sample Num: 4240, Cur Loss: 0.65263492, Cur Avg Loss: 1.36074045, Log Avg loss: 1.42402839, Global Avg Loss: 1.66715941, Time: 0.0402 Steps: 53490, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000275, Sample Num: 4400, Cur Loss: 0.41679615, Cur Avg Loss: 1.34432252, Log Avg loss: 0.90924727, Global Avg Loss: 1.66701775, Time: 0.0402 Steps: 53500, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000285, Sample Num: 4560, Cur Loss: 1.24821806, Cur Avg Loss: 1.35386610, Log Avg loss: 1.61631461, Global Avg Loss: 1.66700827, Time: 0.0402 Steps: 53510, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000295, Sample Num: 4720, Cur Loss: 1.53978264, Cur Avg Loss: 1.36591605, Log Avg loss: 1.70933946, Global Avg Loss: 1.66701618, Time: 0.0402 Steps: 53520, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000305, Sample Num: 4880, Cur Loss: 0.85847723, Cur Avg Loss: 1.36098682, Log Avg loss: 1.21557468, Global Avg Loss: 1.66693185, Time: 0.0402 Steps: 53530, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000315, Sample Num: 5040, Cur Loss: 2.97191811, Cur Avg Loss: 1.36852081, Log Avg loss: 1.59830759, Global Avg Loss: 1.66691903, Time: 0.0402 Steps: 53540, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000325, Sample Num: 5200, Cur Loss: 0.76511002, Cur Avg Loss: 1.36396411, Log Avg loss: 1.22042782, Global Avg Loss: 1.66683565, Time: 0.0402 Steps: 53550, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000335, Sample Num: 5360, Cur Loss: 0.71882993, Cur Avg Loss: 1.35442187, Log Avg loss: 1.04429912, Global Avg Loss: 1.66671942, Time: 0.0402 Steps: 53560, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000345, Sample Num: 5520, Cur Loss: 2.52981758, Cur Avg Loss: 1.35580679, Log Avg loss: 1.40220185, Global Avg Loss: 1.66667004, Time: 0.0402 Steps: 53570, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000355, Sample Num: 5680, Cur Loss: 0.46576434, Cur Avg Loss: 1.34326116, Log Avg loss: 0.91043680, Global Avg Loss: 1.66652890, Time: 0.0402 Steps: 53580, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000365, Sample Num: 5840, Cur Loss: 1.35596299, Cur Avg Loss: 1.34222236, Log Avg loss: 1.30534494, Global Avg Loss: 1.66646150, Time: 0.0402 Steps: 53590, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000375, Sample Num: 6000, Cur Loss: 1.25551212, Cur Avg Loss: 1.34703218, Log Avg loss: 1.52259052, Global Avg Loss: 1.66643466, Time: 0.0402 Steps: 53600, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000385, Sample Num: 6160, Cur Loss: 0.90376759, Cur Avg Loss: 1.35189420, Log Avg loss: 1.53421990, Global Avg Loss: 1.66641000, Time: 0.0402 Steps: 53610, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000395, Sample Num: 6320, Cur Loss: 0.64098597, Cur Avg Loss: 1.35271624, Log Avg loss: 1.38436486, Global Avg Loss: 1.66635740, Time: 0.0402 Steps: 53620, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000405, Sample Num: 6480, Cur Loss: 1.12300229, Cur Avg Loss: 1.35030026, Log Avg loss: 1.25486916, Global Avg Loss: 1.66628067, Time: 0.0404 Steps: 53630, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000415, Sample Num: 6640, Cur Loss: 0.74543101, Cur Avg Loss: 1.34183216, Log Avg loss: 0.99887388, Global Avg Loss: 1.66615625, Time: 0.0402 Steps: 53640, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000425, Sample Num: 6800, Cur Loss: 0.50791824, Cur Avg Loss: 1.34691225, Log Avg loss: 1.55773618, Global Avg Loss: 1.66613604, Time: 0.0402 Steps: 53650, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000435, Sample Num: 6960, Cur Loss: 0.79652941, Cur Avg Loss: 1.34630453, Log Avg loss: 1.32047653, Global Avg Loss: 1.66607162, Time: 0.0402 Steps: 53660, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000445, Sample Num: 7120, Cur Loss: 0.92244226, Cur Avg Loss: 1.34975902, Log Avg loss: 1.50002907, Global Avg Loss: 1.66604068, Time: 0.0402 Steps: 53670, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000455, Sample Num: 7280, Cur Loss: 1.29404414, Cur Avg Loss: 1.35175814, Log Avg loss: 1.44071900, Global Avg Loss: 1.66599871, Time: 0.0402 Steps: 53680, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000465, Sample Num: 7440, Cur Loss: 0.73569000, Cur Avg Loss: 1.34952698, Log Avg loss: 1.24800931, Global Avg Loss: 1.66592086, Time: 0.0404 Steps: 53690, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000475, Sample Num: 7600, Cur Loss: 0.73100150, Cur Avg Loss: 1.35048012, Log Avg loss: 1.39480091, Global Avg Loss: 1.66587037, Time: 0.0402 Steps: 53700, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000485, Sample Num: 7760, Cur Loss: 1.48833442, Cur Avg Loss: 1.34368861, Log Avg loss: 1.02109188, Global Avg Loss: 1.66575032, Time: 0.0616 Steps: 53710, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000495, Sample Num: 7920, Cur Loss: 1.58862853, Cur Avg Loss: 1.34533156, Log Avg loss: 1.42501487, Global Avg Loss: 1.66570551, Time: 0.0518 Steps: 53720, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000505, Sample Num: 8080, Cur Loss: 1.32391047, Cur Avg Loss: 1.34147627, Log Avg loss: 1.15063917, Global Avg Loss: 1.66560965, Time: 0.0407 Steps: 53730, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000515, Sample Num: 8240, Cur Loss: 1.67096055, Cur Avg Loss: 1.34176935, Log Avg loss: 1.35657023, Global Avg Loss: 1.66555214, Time: 0.0503 Steps: 53740, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000525, Sample Num: 8400, Cur Loss: 1.66175997, Cur Avg Loss: 1.33436595, Log Avg loss: 0.95309065, Global Avg Loss: 1.66541959, Time: 0.0528 Steps: 53750, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000535, Sample Num: 8560, Cur Loss: 2.66438961, Cur Avg Loss: 1.33762622, Log Avg loss: 1.50879064, Global Avg Loss: 1.66539045, Time: 0.0405 Steps: 53760, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000545, Sample Num: 8720, Cur Loss: 0.61767304, Cur Avg Loss: 1.33640009, Log Avg loss: 1.27080216, Global Avg Loss: 1.66531707, Time: 0.0542 Steps: 53770, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000555, Sample Num: 8880, Cur Loss: 2.00614905, Cur Avg Loss: 1.33831458, Log Avg loss: 1.44265419, Global Avg Loss: 1.66527567, Time: 0.0403 Steps: 53780, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000565, Sample Num: 9040, Cur Loss: 0.89227206, Cur Avg Loss: 1.33679803, Log Avg loss: 1.25262925, Global Avg Loss: 1.66519895, Time: 0.0402 Steps: 53790, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000575, Sample Num: 9200, Cur Loss: 2.23241162, Cur Avg Loss: 1.33455537, Log Avg loss: 1.20784517, Global Avg Loss: 1.66511394, Time: 0.0402 Steps: 53800, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000585, Sample Num: 9360, Cur Loss: 0.64766550, Cur Avg Loss: 1.32482735, Log Avg loss: 0.76546622, Global Avg Loss: 1.66494675, Time: 0.0403 Steps: 53810, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000595, Sample Num: 9520, Cur Loss: 2.30484676, Cur Avg Loss: 1.32122306, Log Avg loss: 1.11037216, Global Avg Loss: 1.66484371, Time: 0.0402 Steps: 53820, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000605, Sample Num: 9680, Cur Loss: 0.54504740, Cur Avg Loss: 1.31974629, Log Avg loss: 1.23187829, Global Avg Loss: 1.66476328, Time: 0.0402 Steps: 53830, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000615, Sample Num: 9840, Cur Loss: 1.03504920, Cur Avg Loss: 1.31829859, Log Avg loss: 1.23071299, Global Avg Loss: 1.66468266, Time: 0.0401 Steps: 53840, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000625, Sample Num: 10000, Cur Loss: 0.48407987, Cur Avg Loss: 1.31757980, Log Avg loss: 1.27337394, Global Avg Loss: 1.66460999, Time: 0.0402 Steps: 53850, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000635, Sample Num: 10160, Cur Loss: 0.93244374, Cur Avg Loss: 1.31527639, Log Avg loss: 1.17131356, Global Avg Loss: 1.66451840, Time: 0.0402 Steps: 53860, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000645, Sample Num: 10320, Cur Loss: 1.45073891, Cur Avg Loss: 1.31508575, Log Avg loss: 1.30297980, Global Avg Loss: 1.66445129, Time: 0.0402 Steps: 53870, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000655, Sample Num: 10480, Cur Loss: 0.78492576, Cur Avg Loss: 1.31260446, Log Avg loss: 1.15256135, Global Avg Loss: 1.66435628, Time: 0.0402 Steps: 53880, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000665, Sample Num: 10640, Cur Loss: 1.83670473, Cur Avg Loss: 1.31674664, Log Avg loss: 1.58805952, Global Avg Loss: 1.66434213, Time: 0.0402 Steps: 53890, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000675, Sample Num: 10800, Cur Loss: 1.95339489, Cur Avg Loss: 1.31610277, Log Avg loss: 1.27328577, Global Avg Loss: 1.66426957, Time: 0.0402 Steps: 53900, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000685, Sample Num: 10960, Cur Loss: 1.89668655, Cur Avg Loss: 1.31268875, Log Avg loss: 1.08224238, Global Avg Loss: 1.66416161, Time: 0.0402 Steps: 53910, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000695, Sample Num: 11120, Cur Loss: 1.23320699, Cur Avg Loss: 1.31278217, Log Avg loss: 1.31918136, Global Avg Loss: 1.66409763, Time: 0.0402 Steps: 53920, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000705, Sample Num: 11280, Cur Loss: 1.90726650, Cur Avg Loss: 1.31257824, Log Avg loss: 1.29840502, Global Avg Loss: 1.66402982, Time: 0.0402 Steps: 53930, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000715, Sample Num: 11440, Cur Loss: 1.04848552, Cur Avg Loss: 1.31007842, Log Avg loss: 1.13384102, Global Avg Loss: 1.66393153, Time: 0.0402 Steps: 53940, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000725, Sample Num: 11600, Cur Loss: 0.38145253, Cur Avg Loss: 1.30322360, Log Avg loss: 0.81310376, Global Avg Loss: 1.66377382, Time: 0.0402 Steps: 53950, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000735, Sample Num: 11760, Cur Loss: 2.02232456, Cur Avg Loss: 1.30489381, Log Avg loss: 1.42598410, Global Avg Loss: 1.66372976, Time: 0.0402 Steps: 53960, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000745, Sample Num: 11920, Cur Loss: 1.17558169, Cur Avg Loss: 1.30552464, Log Avg loss: 1.35189106, Global Avg Loss: 1.66367198, Time: 0.0402 Steps: 53970, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000755, Sample Num: 12080, Cur Loss: 0.93765247, Cur Avg Loss: 1.30668679, Log Avg loss: 1.39326683, Global Avg Loss: 1.66362188, Time: 0.0402 Steps: 53980, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000765, Sample Num: 12240, Cur Loss: 1.18118262, Cur Avg Loss: 1.30639622, Log Avg loss: 1.28445764, Global Avg Loss: 1.66355165, Time: 0.0402 Steps: 53990, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000775, Sample Num: 12400, Cur Loss: 1.07453239, Cur Avg Loss: 1.30471599, Log Avg loss: 1.17617856, Global Avg Loss: 1.66346140, Time: 0.0520 Steps: 54000, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000785, Sample Num: 12560, Cur Loss: 0.96915323, Cur Avg Loss: 1.30456131, Log Avg loss: 1.29257362, Global Avg Loss: 1.66339273, Time: 0.0462 Steps: 54010, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000795, Sample Num: 12720, Cur Loss: 0.23903151, Cur Avg Loss: 1.30065280, Log Avg loss: 0.99383459, Global Avg Loss: 1.66326878, Time: 0.0732 Steps: 54020, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000805, Sample Num: 12880, Cur Loss: 1.32423925, Cur Avg Loss: 1.30213934, Log Avg loss: 1.42031955, Global Avg Loss: 1.66322382, Time: 0.0480 Steps: 54030, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000815, Sample Num: 13040, Cur Loss: 0.79506755, Cur Avg Loss: 1.30259178, Log Avg loss: 1.33901294, Global Avg Loss: 1.66316382, Time: 0.0457 Steps: 54040, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000825, Sample Num: 13200, Cur Loss: 0.82991445, Cur Avg Loss: 1.30562560, Log Avg loss: 1.55288225, Global Avg Loss: 1.66314342, Time: 0.0406 Steps: 54050, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000835, Sample Num: 13360, Cur Loss: 2.09628105, Cur Avg Loss: 1.30493198, Log Avg loss: 1.24770850, Global Avg Loss: 1.66306657, Time: 0.0520 Steps: 54060, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000845, Sample Num: 13520, Cur Loss: 1.04890418, Cur Avg Loss: 1.30384607, Log Avg loss: 1.21317282, Global Avg Loss: 1.66298337, Time: 0.0403 Steps: 54070, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000855, Sample Num: 13680, Cur Loss: 1.48852015, Cur Avg Loss: 1.30677016, Log Avg loss: 1.55385571, Global Avg Loss: 1.66296319, Time: 0.0402 Steps: 54080, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000865, Sample Num: 13840, Cur Loss: 1.75452900, Cur Avg Loss: 1.30896014, Log Avg loss: 1.49620346, Global Avg Loss: 1.66293236, Time: 0.0402 Steps: 54090, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000875, Sample Num: 14000, Cur Loss: 0.51192701, Cur Avg Loss: 1.30802278, Log Avg loss: 1.22694110, Global Avg Loss: 1.66285177, Time: 0.0402 Steps: 54100, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000885, Sample Num: 14160, Cur Loss: 1.69747150, Cur Avg Loss: 1.30879000, Log Avg loss: 1.37592162, Global Avg Loss: 1.66279874, Time: 0.0402 Steps: 54110, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000895, Sample Num: 14320, Cur Loss: 2.49604082, Cur Avg Loss: 1.30733423, Log Avg loss: 1.17849813, Global Avg Loss: 1.66270925, Time: 0.0401 Steps: 54120, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000905, Sample Num: 14480, Cur Loss: 1.01173842, Cur Avg Loss: 1.30841140, Log Avg loss: 1.40481791, Global Avg Loss: 1.66266161, Time: 0.0402 Steps: 54130, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000915, Sample Num: 14640, Cur Loss: 2.20763969, Cur Avg Loss: 1.30847378, Log Avg loss: 1.31411958, Global Avg Loss: 1.66259723, Time: 0.0402 Steps: 54140, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000925, Sample Num: 14800, Cur Loss: 1.57812309, Cur Avg Loss: 1.31371544, Log Avg loss: 1.79332756, Global Avg Loss: 1.66262138, Time: 0.0402 Steps: 54150, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000935, Sample Num: 14960, Cur Loss: 1.74837518, Cur Avg Loss: 1.31341773, Log Avg loss: 1.28587956, Global Avg Loss: 1.66255181, Time: 0.0402 Steps: 54160, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000945, Sample Num: 15120, Cur Loss: 1.65022755, Cur Avg Loss: 1.31631538, Log Avg loss: 1.58724561, Global Avg Loss: 1.66253791, Time: 0.0402 Steps: 54170, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000955, Sample Num: 15280, Cur Loss: 2.55478048, Cur Avg Loss: 1.31651348, Log Avg loss: 1.33523350, Global Avg Loss: 1.66247750, Time: 0.0403 Steps: 54180, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000965, Sample Num: 15440, Cur Loss: 1.42183971, Cur Avg Loss: 1.31452004, Log Avg loss: 1.12414647, Global Avg Loss: 1.66237816, Time: 0.0402 Steps: 54190, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000975, Sample Num: 15600, Cur Loss: 0.59483093, Cur Avg Loss: 1.31244139, Log Avg loss: 1.11185211, Global Avg Loss: 1.66227659, Time: 0.0402 Steps: 54200, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000985, Sample Num: 15760, Cur Loss: 0.80597395, Cur Avg Loss: 1.31372626, Log Avg loss: 1.43900133, Global Avg Loss: 1.66223540, Time: 0.0402 Steps: 54210, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000995, Sample Num: 15920, Cur Loss: 2.68663454, Cur Avg Loss: 1.31553369, Log Avg loss: 1.49356555, Global Avg Loss: 1.66220429, Time: 0.0403 Steps: 54220, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001005, Sample Num: 16080, Cur Loss: 0.44565850, Cur Avg Loss: 1.31286859, Log Avg loss: 1.04769082, Global Avg Loss: 1.66209098, Time: 0.0402 Steps: 54230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001015, Sample Num: 16240, Cur Loss: 2.37582302, Cur Avg Loss: 1.31535315, Log Avg loss: 1.56505158, Global Avg Loss: 1.66207309, Time: 0.0402 Steps: 54240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001025, Sample Num: 16400, Cur Loss: 2.80724311, Cur Avg Loss: 1.31633745, Log Avg loss: 1.41624336, Global Avg Loss: 1.66202777, Time: 0.0402 Steps: 54250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001035, Sample Num: 16560, Cur Loss: 1.34748209, Cur Avg Loss: 1.31835940, Log Avg loss: 1.52560995, Global Avg Loss: 1.66200263, Time: 0.0401 Steps: 54260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001045, Sample Num: 16720, Cur Loss: 3.29692841, Cur Avg Loss: 1.31779261, Log Avg loss: 1.25912910, Global Avg Loss: 1.66192839, Time: 0.0402 Steps: 54270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001055, Sample Num: 16880, Cur Loss: 1.78338397, Cur Avg Loss: 1.32112401, Log Avg loss: 1.66925548, Global Avg Loss: 1.66192974, Time: 0.0402 Steps: 54280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001065, Sample Num: 17040, Cur Loss: 1.87571454, Cur Avg Loss: 1.31888451, Log Avg loss: 1.08261741, Global Avg Loss: 1.66182304, Time: 0.0402 Steps: 54290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001075, Sample Num: 17200, Cur Loss: 0.70448995, Cur Avg Loss: 1.31597425, Log Avg loss: 1.00603200, Global Avg Loss: 1.66170227, Time: 0.0506 Steps: 54300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001085, Sample Num: 17360, Cur Loss: 0.90634185, Cur Avg Loss: 1.31673646, Log Avg loss: 1.39867371, Global Avg Loss: 1.66165383, Time: 0.0541 Steps: 54310, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001095, Sample Num: 17520, Cur Loss: 0.62609965, Cur Avg Loss: 1.31590414, Log Avg loss: 1.22559705, Global Avg Loss: 1.66157356, Time: 0.0406 Steps: 54320, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001105, Sample Num: 17680, Cur Loss: 2.22782803, Cur Avg Loss: 1.31635257, Log Avg loss: 1.36545557, Global Avg Loss: 1.66151906, Time: 0.0520 Steps: 54330, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001115, Sample Num: 17840, Cur Loss: 2.04861736, Cur Avg Loss: 1.31549910, Log Avg loss: 1.22119070, Global Avg Loss: 1.66143802, Time: 0.0524 Steps: 54340, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001125, Sample Num: 18000, Cur Loss: 0.96928906, Cur Avg Loss: 1.31772152, Log Avg loss: 1.56552146, Global Avg Loss: 1.66142038, Time: 0.0457 Steps: 54350, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001135, Sample Num: 18160, Cur Loss: 1.93979418, Cur Avg Loss: 1.31649797, Log Avg loss: 1.17884926, Global Avg Loss: 1.66133160, Time: 0.0404 Steps: 54360, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001145, Sample Num: 18320, Cur Loss: 1.14052141, Cur Avg Loss: 1.31771199, Log Avg loss: 1.45550236, Global Avg Loss: 1.66129375, Time: 0.0403 Steps: 54370, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001155, Sample Num: 18480, Cur Loss: 1.02384448, Cur Avg Loss: 1.31869022, Log Avg loss: 1.43069866, Global Avg Loss: 1.66125134, Time: 0.0402 Steps: 54380, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001165, Sample Num: 18640, Cur Loss: 1.13278675, Cur Avg Loss: 1.31874088, Log Avg loss: 1.32459149, Global Avg Loss: 1.66118944, Time: 0.0402 Steps: 54390, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001175, Sample Num: 18800, Cur Loss: 1.99959970, Cur Avg Loss: 1.32088913, Log Avg loss: 1.57116027, Global Avg Loss: 1.66117289, Time: 0.0402 Steps: 54400, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001185, Sample Num: 18960, Cur Loss: 1.44117594, Cur Avg Loss: 1.32119117, Log Avg loss: 1.35668115, Global Avg Loss: 1.66111693, Time: 0.0403 Steps: 54410, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001195, Sample Num: 19120, Cur Loss: 0.92034233, Cur Avg Loss: 1.31959132, Log Avg loss: 1.13000917, Global Avg Loss: 1.66101934, Time: 0.0403 Steps: 54420, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001205, Sample Num: 19280, Cur Loss: 1.18244243, Cur Avg Loss: 1.32136643, Log Avg loss: 1.53349229, Global Avg Loss: 1.66099591, Time: 0.0402 Steps: 54430, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001215, Sample Num: 19440, Cur Loss: 0.95804679, Cur Avg Loss: 1.32377621, Log Avg loss: 1.61415426, Global Avg Loss: 1.66098730, Time: 0.0402 Steps: 54440, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001225, Sample Num: 19600, Cur Loss: 1.75236201, Cur Avg Loss: 1.32545781, Log Avg loss: 1.52977186, Global Avg Loss: 1.66096321, Time: 0.0402 Steps: 54450, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001235, Sample Num: 19760, Cur Loss: 0.83065307, Cur Avg Loss: 1.32488580, Log Avg loss: 1.25481452, Global Avg Loss: 1.66088863, Time: 0.0402 Steps: 54460, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001245, Sample Num: 19920, Cur Loss: 1.87180126, Cur Avg Loss: 1.32290518, Log Avg loss: 1.07829954, Global Avg Loss: 1.66078167, Time: 0.0402 Steps: 54470, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001255, Sample Num: 20080, Cur Loss: 0.55705720, Cur Avg Loss: 1.32064929, Log Avg loss: 1.03979017, Global Avg Loss: 1.66066769, Time: 0.0401 Steps: 54480, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001265, Sample Num: 20240, Cur Loss: 1.31044173, Cur Avg Loss: 1.31932924, Log Avg loss: 1.15366302, Global Avg Loss: 1.66057464, Time: 0.0402 Steps: 54490, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001275, Sample Num: 20400, Cur Loss: 1.51818585, Cur Avg Loss: 1.32072089, Log Avg loss: 1.49676493, Global Avg Loss: 1.66054458, Time: 0.0402 Steps: 54500, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001285, Sample Num: 20560, Cur Loss: 0.65616548, Cur Avg Loss: 1.31918016, Log Avg loss: 1.12273682, Global Avg Loss: 1.66044592, Time: 0.0402 Steps: 54510, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001295, Sample Num: 20720, Cur Loss: 0.87155753, Cur Avg Loss: 1.31982371, Log Avg loss: 1.40251962, Global Avg Loss: 1.66039861, Time: 0.0402 Steps: 54520, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001305, Sample Num: 20880, Cur Loss: 1.42957079, Cur Avg Loss: 1.32161301, Log Avg loss: 1.55332790, Global Avg Loss: 1.66037898, Time: 0.0402 Steps: 54530, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001315, Sample Num: 21040, Cur Loss: 0.54801369, Cur Avg Loss: 1.31961576, Log Avg loss: 1.05897472, Global Avg Loss: 1.66026871, Time: 0.0402 Steps: 54540, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001325, Sample Num: 21200, Cur Loss: 1.07116365, Cur Avg Loss: 1.32002047, Log Avg loss: 1.37323905, Global Avg Loss: 1.66021609, Time: 0.0402 Steps: 54550, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001335, Sample Num: 21360, Cur Loss: 0.63196468, Cur Avg Loss: 1.32103127, Log Avg loss: 1.45496300, Global Avg Loss: 1.66017847, Time: 0.0402 Steps: 54560, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001345, Sample Num: 21520, Cur Loss: 0.79572487, Cur Avg Loss: 1.32135747, Log Avg loss: 1.36490486, Global Avg Loss: 1.66012436, Time: 0.0402 Steps: 54570, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001355, Sample Num: 21680, Cur Loss: 1.66884875, Cur Avg Loss: 1.32209073, Log Avg loss: 1.42071370, Global Avg Loss: 1.66008050, Time: 0.0402 Steps: 54580, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001365, Sample Num: 21840, Cur Loss: 1.10468590, Cur Avg Loss: 1.32257016, Log Avg loss: 1.38753340, Global Avg Loss: 1.66003057, Time: 0.0402 Steps: 54590, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001375, Sample Num: 22000, Cur Loss: 1.10078979, Cur Avg Loss: 1.32452635, Log Avg loss: 1.59154589, Global Avg Loss: 1.66001803, Time: 0.0402 Steps: 54600, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001385, Sample Num: 22160, Cur Loss: 2.58572960, Cur Avg Loss: 1.32662440, Log Avg loss: 1.61510718, Global Avg Loss: 1.66000981, Time: 0.0558 Steps: 54610, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001395, Sample Num: 22320, Cur Loss: 1.04804683, Cur Avg Loss: 1.32561436, Log Avg loss: 1.18572298, Global Avg Loss: 1.65992297, Time: 0.0642 Steps: 54620, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001405, Sample Num: 22480, Cur Loss: 1.27329481, Cur Avg Loss: 1.32532638, Log Avg loss: 1.28515381, Global Avg Loss: 1.65985437, Time: 0.0581 Steps: 54630, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001415, Sample Num: 22640, Cur Loss: 1.63666773, Cur Avg Loss: 1.32666071, Log Avg loss: 1.51413441, Global Avg Loss: 1.65982770, Time: 0.0612 Steps: 54640, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001425, Sample Num: 22800, Cur Loss: 2.71694136, Cur Avg Loss: 1.32880725, Log Avg loss: 1.63254241, Global Avg Loss: 1.65982271, Time: 0.0458 Steps: 54650, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001435, Sample Num: 22960, Cur Loss: 0.65478122, Cur Avg Loss: 1.32876344, Log Avg loss: 1.32252024, Global Avg Loss: 1.65976100, Time: 0.0482 Steps: 54660, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001445, Sample Num: 23120, Cur Loss: 1.00433457, Cur Avg Loss: 1.32803038, Log Avg loss: 1.22283681, Global Avg Loss: 1.65968108, Time: 0.0558 Steps: 54670, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001455, Sample Num: 23280, Cur Loss: 0.32225659, Cur Avg Loss: 1.32565661, Log Avg loss: 0.98264582, Global Avg Loss: 1.65955726, Time: 0.0402 Steps: 54680, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001465, Sample Num: 23440, Cur Loss: 1.24827242, Cur Avg Loss: 1.32453862, Log Avg loss: 1.16187130, Global Avg Loss: 1.65946626, Time: 0.0402 Steps: 54690, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001475, Sample Num: 23600, Cur Loss: 1.22018051, Cur Avg Loss: 1.32459328, Log Avg loss: 1.33260134, Global Avg Loss: 1.65940650, Time: 0.0402 Steps: 54700, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001485, Sample Num: 23760, Cur Loss: 1.53382003, Cur Avg Loss: 1.32572859, Log Avg loss: 1.49318616, Global Avg Loss: 1.65937612, Time: 0.0402 Steps: 54710, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001495, Sample Num: 23920, Cur Loss: 1.08605278, Cur Avg Loss: 1.32514204, Log Avg loss: 1.23803921, Global Avg Loss: 1.65929912, Time: 0.0402 Steps: 54720, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001505, Sample Num: 24080, Cur Loss: 1.66674137, Cur Avg Loss: 1.32501425, Log Avg loss: 1.30591057, Global Avg Loss: 1.65923455, Time: 0.0402 Steps: 54730, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001515, Sample Num: 24240, Cur Loss: 2.02668571, Cur Avg Loss: 1.32411584, Log Avg loss: 1.18890491, Global Avg Loss: 1.65914863, Time: 0.0402 Steps: 54740, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001525, Sample Num: 24400, Cur Loss: 2.11286521, Cur Avg Loss: 1.32564134, Log Avg loss: 1.55675387, Global Avg Loss: 1.65912993, Time: 0.0402 Steps: 54750, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001535, Sample Num: 24560, Cur Loss: 0.81690085, Cur Avg Loss: 1.32619866, Log Avg loss: 1.41119015, Global Avg Loss: 1.65908465, Time: 0.0402 Steps: 54760, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001545, Sample Num: 24720, Cur Loss: 2.70150661, Cur Avg Loss: 1.32923467, Log Avg loss: 1.79526349, Global Avg Loss: 1.65910952, Time: 0.0402 Steps: 54770, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001555, Sample Num: 24880, Cur Loss: 2.08334064, Cur Avg Loss: 1.32817490, Log Avg loss: 1.16444019, Global Avg Loss: 1.65901922, Time: 0.0401 Steps: 54780, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001565, Sample Num: 25040, Cur Loss: 1.84629428, Cur Avg Loss: 1.32735138, Log Avg loss: 1.19929348, Global Avg Loss: 1.65893531, Time: 0.0402 Steps: 54790, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001575, Sample Num: 25200, Cur Loss: 0.52957940, Cur Avg Loss: 1.32467571, Log Avg loss: 0.90593337, Global Avg Loss: 1.65879790, Time: 0.0401 Steps: 54800, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001585, Sample Num: 25360, Cur Loss: 1.76345813, Cur Avg Loss: 1.32402707, Log Avg loss: 1.22186604, Global Avg Loss: 1.65871818, Time: 0.0402 Steps: 54810, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001595, Sample Num: 25520, Cur Loss: 0.91611838, Cur Avg Loss: 1.32565963, Log Avg loss: 1.58441986, Global Avg Loss: 1.65870463, Time: 0.0402 Steps: 54820, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001605, Sample Num: 25680, Cur Loss: 1.00854325, Cur Avg Loss: 1.32427217, Log Avg loss: 1.10297294, Global Avg Loss: 1.65860327, Time: 0.0402 Steps: 54830, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001615, Sample Num: 25840, Cur Loss: 1.26604104, Cur Avg Loss: 1.32378908, Log Avg loss: 1.24625306, Global Avg Loss: 1.65852808, Time: 0.0401 Steps: 54840, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001625, Sample Num: 26000, Cur Loss: 0.57210773, Cur Avg Loss: 1.32377859, Log Avg loss: 1.32208461, Global Avg Loss: 1.65846674, Time: 0.0402 Steps: 54850, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001635, Sample Num: 26160, Cur Loss: 1.39087415, Cur Avg Loss: 1.32293030, Log Avg loss: 1.18508314, Global Avg Loss: 1.65838045, Time: 0.0401 Steps: 54860, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001645, Sample Num: 26320, Cur Loss: 0.47829100, Cur Avg Loss: 1.32168548, Log Avg loss: 1.11815774, Global Avg Loss: 1.65828200, Time: 0.0402 Steps: 54870, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001655, Sample Num: 26480, Cur Loss: 0.53149408, Cur Avg Loss: 1.32007226, Log Avg loss: 1.05469679, Global Avg Loss: 1.65817202, Time: 0.0402 Steps: 54880, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001665, Sample Num: 26640, Cur Loss: 1.24373281, Cur Avg Loss: 1.31936488, Log Avg loss: 1.20229413, Global Avg Loss: 1.65808896, Time: 0.0402 Steps: 54890, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001675, Sample Num: 26800, Cur Loss: 0.64170438, Cur Avg Loss: 1.31914613, Log Avg loss: 1.28272474, Global Avg Loss: 1.65802059, Time: 0.0533 Steps: 54900, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001685, Sample Num: 26960, Cur Loss: 1.46864331, Cur Avg Loss: 1.31937886, Log Avg loss: 1.35836073, Global Avg Loss: 1.65796602, Time: 0.0493 Steps: 54910, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001695, Sample Num: 27120, Cur Loss: 2.47082663, Cur Avg Loss: 1.32102815, Log Avg loss: 1.59893394, Global Avg Loss: 1.65795527, Time: 0.0406 Steps: 54920, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001705, Sample Num: 27280, Cur Loss: 2.09107542, Cur Avg Loss: 1.32143655, Log Avg loss: 1.39066038, Global Avg Loss: 1.65790661, Time: 0.0515 Steps: 54930, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001715, Sample Num: 27440, Cur Loss: 0.92276394, Cur Avg Loss: 1.32149576, Log Avg loss: 1.33159093, Global Avg Loss: 1.65784721, Time: 0.0519 Steps: 54940, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001725, Sample Num: 27600, Cur Loss: 1.95918560, Cur Avg Loss: 1.32040691, Log Avg loss: 1.13366850, Global Avg Loss: 1.65775182, Time: 0.0547 Steps: 54950, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001735, Sample Num: 27760, Cur Loss: 1.97154927, Cur Avg Loss: 1.32100545, Log Avg loss: 1.42425432, Global Avg Loss: 1.65770934, Time: 0.0402 Steps: 54960, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001745, Sample Num: 27920, Cur Loss: 1.53661585, Cur Avg Loss: 1.32220509, Log Avg loss: 1.53034283, Global Avg Loss: 1.65768617, Time: 0.0402 Steps: 54970, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001755, Sample Num: 28080, Cur Loss: 1.36179495, Cur Avg Loss: 1.32368768, Log Avg loss: 1.58239805, Global Avg Loss: 1.65767247, Time: 0.0402 Steps: 54980, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001765, Sample Num: 28240, Cur Loss: 1.38175404, Cur Avg Loss: 1.32365706, Log Avg loss: 1.31828467, Global Avg Loss: 1.65761076, Time: 0.0402 Steps: 54990, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001775, Sample Num: 28400, Cur Loss: 1.19619238, Cur Avg Loss: 1.32487531, Log Avg loss: 1.53989531, Global Avg Loss: 1.65758935, Time: 0.0402 Steps: 55000, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001785, Sample Num: 28560, Cur Loss: 2.40575051, Cur Avg Loss: 1.32642726, Log Avg loss: 1.60189875, Global Avg Loss: 1.65757923, Time: 0.0402 Steps: 55010, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001795, Sample Num: 28720, Cur Loss: 1.64557767, Cur Avg Loss: 1.32656588, Log Avg loss: 1.35130932, Global Avg Loss: 1.65752356, Time: 0.0402 Steps: 55020, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001805, Sample Num: 28880, Cur Loss: 0.36293161, Cur Avg Loss: 1.32576399, Log Avg loss: 1.18182468, Global Avg Loss: 1.65743712, Time: 0.0401 Steps: 55030, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001815, Sample Num: 29040, Cur Loss: 1.45908916, Cur Avg Loss: 1.32676224, Log Avg loss: 1.50694622, Global Avg Loss: 1.65740978, Time: 0.0402 Steps: 55040, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001825, Sample Num: 29200, Cur Loss: 0.68793076, Cur Avg Loss: 1.32662391, Log Avg loss: 1.30151677, Global Avg Loss: 1.65734513, Time: 0.0402 Steps: 55050, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001835, Sample Num: 29360, Cur Loss: 1.35390854, Cur Avg Loss: 1.32816020, Log Avg loss: 1.60853307, Global Avg Loss: 1.65733626, Time: 0.0402 Steps: 55060, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001845, Sample Num: 29520, Cur Loss: 1.49151731, Cur Avg Loss: 1.32747359, Log Avg loss: 1.20148091, Global Avg Loss: 1.65725349, Time: 0.0402 Steps: 55070, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001855, Sample Num: 29680, Cur Loss: 1.06701434, Cur Avg Loss: 1.32704246, Log Avg loss: 1.24749998, Global Avg Loss: 1.65717909, Time: 0.0402 Steps: 55080, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001865, Sample Num: 29840, Cur Loss: 1.12283373, Cur Avg Loss: 1.32618168, Log Avg loss: 1.16650738, Global Avg Loss: 1.65709003, Time: 0.0402 Steps: 55090, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001875, Sample Num: 30000, Cur Loss: 1.27618062, Cur Avg Loss: 1.32571052, Log Avg loss: 1.23783794, Global Avg Loss: 1.65701394, Time: 0.0402 Steps: 55100, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001885, Sample Num: 30160, Cur Loss: 0.74156320, Cur Avg Loss: 1.32741598, Log Avg loss: 1.64718964, Global Avg Loss: 1.65701215, Time: 0.0402 Steps: 55110, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001895, Sample Num: 30320, Cur Loss: 0.59627640, Cur Avg Loss: 1.32725977, Log Avg loss: 1.29781404, Global Avg Loss: 1.65694699, Time: 0.0402 Steps: 55120, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001905, Sample Num: 30480, Cur Loss: 1.58051240, Cur Avg Loss: 1.32837105, Log Avg loss: 1.53895925, Global Avg Loss: 1.65692559, Time: 0.0403 Steps: 55130, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001915, Sample Num: 30640, Cur Loss: 0.68822682, Cur Avg Loss: 1.32782298, Log Avg loss: 1.22341532, Global Avg Loss: 1.65684697, Time: 0.0402 Steps: 55140, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001925, Sample Num: 30800, Cur Loss: 2.65299439, Cur Avg Loss: 1.32706687, Log Avg loss: 1.18227222, Global Avg Loss: 1.65676091, Time: 0.0402 Steps: 55150, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001935, Sample Num: 30960, Cur Loss: 0.30078435, Cur Avg Loss: 1.32616847, Log Avg loss: 1.15322594, Global Avg Loss: 1.65666963, Time: 0.0402 Steps: 55160, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001945, Sample Num: 31120, Cur Loss: 1.24732685, Cur Avg Loss: 1.32731719, Log Avg loss: 1.54959561, Global Avg Loss: 1.65665022, Time: 0.0402 Steps: 55170, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001955, Sample Num: 31280, Cur Loss: 1.64449465, Cur Avg Loss: 1.32819498, Log Avg loss: 1.49892407, Global Avg Loss: 1.65662164, Time: 0.0401 Steps: 55180, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001965, Sample Num: 31440, Cur Loss: 0.94303566, Cur Avg Loss: 1.32927475, Log Avg loss: 1.54037019, Global Avg Loss: 1.65660057, Time: 0.0491 Steps: 55190, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001975, Sample Num: 31600, Cur Loss: 2.41629124, Cur Avg Loss: 1.33199884, Log Avg loss: 1.86728159, Global Avg Loss: 1.65663874, Time: 0.0405 Steps: 55200, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001985, Sample Num: 31760, Cur Loss: 0.82525063, Cur Avg Loss: 1.33347423, Log Avg loss: 1.62486570, Global Avg Loss: 1.65663298, Time: 0.0430 Steps: 55210, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001995, Sample Num: 31920, Cur Loss: 1.68526101, Cur Avg Loss: 1.33352062, Log Avg loss: 1.34272884, Global Avg Loss: 1.65657614, Time: 0.0533 Steps: 55220, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002005, Sample Num: 32080, Cur Loss: 0.46430254, Cur Avg Loss: 1.33321009, Log Avg loss: 1.27125763, Global Avg Loss: 1.65650637, Time: 0.0405 Steps: 55230, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002015, Sample Num: 32240, Cur Loss: 1.47573805, Cur Avg Loss: 1.33169183, Log Avg loss: 1.02728186, Global Avg Loss: 1.65639246, Time: 0.0422 Steps: 55240, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002025, Sample Num: 32400, Cur Loss: 1.65843761, Cur Avg Loss: 1.33289686, Log Avg loss: 1.57571081, Global Avg Loss: 1.65637786, Time: 0.0433 Steps: 55250, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002035, Sample Num: 32560, Cur Loss: 1.02157807, Cur Avg Loss: 1.33325216, Log Avg loss: 1.40520038, Global Avg Loss: 1.65633241, Time: 0.0403 Steps: 55260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002045, Sample Num: 32720, Cur Loss: 0.92808878, Cur Avg Loss: 1.33435762, Log Avg loss: 1.55931792, Global Avg Loss: 1.65631486, Time: 0.0402 Steps: 55270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002055, Sample Num: 32880, Cur Loss: 1.56412470, Cur Avg Loss: 1.33571916, Log Avg loss: 1.61415352, Global Avg Loss: 1.65630723, Time: 0.0402 Steps: 55280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002065, Sample Num: 33040, Cur Loss: 1.02919316, Cur Avg Loss: 1.33549660, Log Avg loss: 1.28976169, Global Avg Loss: 1.65624093, Time: 0.0402 Steps: 55290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002075, Sample Num: 33200, Cur Loss: 1.42044592, Cur Avg Loss: 1.33512691, Log Avg loss: 1.25878464, Global Avg Loss: 1.65616906, Time: 0.0402 Steps: 55300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002085, Sample Num: 33360, Cur Loss: 0.63983274, Cur Avg Loss: 1.33367828, Log Avg loss: 1.03308790, Global Avg Loss: 1.65605641, Time: 0.0403 Steps: 55310, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002095, Sample Num: 33520, Cur Loss: 1.38730526, Cur Avg Loss: 1.33424393, Log Avg loss: 1.45218339, Global Avg Loss: 1.65601955, Time: 0.0402 Steps: 55320, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002105, Sample Num: 33680, Cur Loss: 0.50126970, Cur Avg Loss: 1.33463359, Log Avg loss: 1.41626662, Global Avg Loss: 1.65597622, Time: 0.0403 Steps: 55330, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002115, Sample Num: 33840, Cur Loss: 2.47833395, Cur Avg Loss: 1.33465222, Log Avg loss: 1.33857378, Global Avg Loss: 1.65591887, Time: 0.0402 Steps: 55340, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002125, Sample Num: 34000, Cur Loss: 1.42548954, Cur Avg Loss: 1.33531346, Log Avg loss: 1.47516628, Global Avg Loss: 1.65588621, Time: 0.0402 Steps: 55350, Updated lr: 0.000048 ***** Running evaluation checkpoint-55354 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-55354 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 89.903707, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.223648, "eval_total_loss": 860.224427, "eval_mae": 0.913543, "eval_mse": 1.223695, "eval_r2": 0.222139, "eval_sp_statistic": 0.395253, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.483318, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.157599, "test_total_loss": 581.114901, "test_mae": 0.706076, "test_mse": 1.157965, "test_r2": 0.252639, "test_sp_statistic": 0.413526, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.516428, "test_ps_pvalue": 0.0, "lr": 4.845519203413941e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6558214167637921, "train_cur_epoch_loss": 2840.577990487218, "train_cur_epoch_avg_loss": 1.3342310899423286, "train_cur_epoch_time": 89.90370726585388, "train_cur_epoch_avg_time": 0.04222813868757815, "epoch": 26, "step": 55354} ################################################## Training, Epoch: 0027, Batch: 000006, Sample Num: 96, Cur Loss: 2.35033989, Cur Avg Loss: 1.28339863, Log Avg loss: 1.07372733, Global Avg Loss: 1.65578105, Time: 0.0402 Steps: 55360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000016, Sample Num: 256, Cur Loss: 1.40393496, Cur Avg Loss: 1.30561837, Log Avg loss: 1.31895021, Global Avg Loss: 1.65572022, Time: 0.0401 Steps: 55370, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000026, Sample Num: 416, Cur Loss: 2.17483568, Cur Avg Loss: 1.31740357, Log Avg loss: 1.33625990, Global Avg Loss: 1.65566254, Time: 0.0402 Steps: 55380, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000036, Sample Num: 576, Cur Loss: 1.41796660, Cur Avg Loss: 1.24990328, Log Avg loss: 1.07440251, Global Avg Loss: 1.65555760, Time: 0.0484 Steps: 55390, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000046, Sample Num: 736, Cur Loss: 0.80157363, Cur Avg Loss: 1.28497000, Log Avg loss: 1.41121020, Global Avg Loss: 1.65551349, Time: 0.0457 Steps: 55400, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000056, Sample Num: 896, Cur Loss: 1.15118480, Cur Avg Loss: 1.27892434, Log Avg loss: 1.25111430, Global Avg Loss: 1.65544051, Time: 0.0457 Steps: 55410, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000066, Sample Num: 1056, Cur Loss: 1.78081894, Cur Avg Loss: 1.27972700, Log Avg loss: 1.28422190, Global Avg Loss: 1.65537352, Time: 0.0499 Steps: 55420, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000076, Sample Num: 1216, Cur Loss: 2.53886533, Cur Avg Loss: 1.30723892, Log Avg loss: 1.48881763, Global Avg Loss: 1.65534348, Time: 0.0521 Steps: 55430, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000086, Sample Num: 1376, Cur Loss: 0.74943084, Cur Avg Loss: 1.30001362, Log Avg loss: 1.24510134, Global Avg Loss: 1.65526948, Time: 0.0402 Steps: 55440, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000096, Sample Num: 1536, Cur Loss: 1.31164813, Cur Avg Loss: 1.29579879, Log Avg loss: 1.25955118, Global Avg Loss: 1.65519811, Time: 0.0402 Steps: 55450, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000106, Sample Num: 1696, Cur Loss: 1.87228131, Cur Avg Loss: 1.29392127, Log Avg loss: 1.27589710, Global Avg Loss: 1.65512972, Time: 0.0402 Steps: 55460, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000116, Sample Num: 1856, Cur Loss: 0.66051024, Cur Avg Loss: 1.28306722, Log Avg loss: 1.16801429, Global Avg Loss: 1.65504191, Time: 0.0402 Steps: 55470, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000126, Sample Num: 2016, Cur Loss: 3.45388126, Cur Avg Loss: 1.30012097, Log Avg loss: 1.49794454, Global Avg Loss: 1.65501359, Time: 0.0402 Steps: 55480, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000136, Sample Num: 2176, Cur Loss: 1.40820670, Cur Avg Loss: 1.29446448, Log Avg loss: 1.22319263, Global Avg Loss: 1.65493577, Time: 0.0403 Steps: 55490, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000146, Sample Num: 2336, Cur Loss: 1.13124812, Cur Avg Loss: 1.30447259, Log Avg loss: 1.44058292, Global Avg Loss: 1.65489715, Time: 0.0402 Steps: 55500, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000156, Sample Num: 2496, Cur Loss: 1.40618992, Cur Avg Loss: 1.31631443, Log Avg loss: 1.48920535, Global Avg Loss: 1.65486730, Time: 0.0402 Steps: 55510, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000166, Sample Num: 2656, Cur Loss: 0.72774220, Cur Avg Loss: 1.33324121, Log Avg loss: 1.59729884, Global Avg Loss: 1.65485693, Time: 0.0402 Steps: 55520, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000176, Sample Num: 2816, Cur Loss: 1.39296675, Cur Avg Loss: 1.33214915, Log Avg loss: 1.31402099, Global Avg Loss: 1.65479555, Time: 0.0402 Steps: 55530, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000186, Sample Num: 2976, Cur Loss: 0.80844170, Cur Avg Loss: 1.32945678, Log Avg loss: 1.28207118, Global Avg Loss: 1.65472844, Time: 0.0402 Steps: 55540, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000196, Sample Num: 3136, Cur Loss: 1.18298149, Cur Avg Loss: 1.31806500, Log Avg loss: 1.10617784, Global Avg Loss: 1.65462969, Time: 0.0403 Steps: 55550, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000206, Sample Num: 3296, Cur Loss: 0.45079327, Cur Avg Loss: 1.30466083, Log Avg loss: 1.04193910, Global Avg Loss: 1.65451942, Time: 0.0402 Steps: 55560, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000216, Sample Num: 3456, Cur Loss: 3.18454933, Cur Avg Loss: 1.32898820, Log Avg loss: 1.83013196, Global Avg Loss: 1.65455102, Time: 0.0402 Steps: 55570, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000226, Sample Num: 3616, Cur Loss: 1.04462731, Cur Avg Loss: 1.32741692, Log Avg loss: 1.29347725, Global Avg Loss: 1.65448605, Time: 0.0402 Steps: 55580, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000236, Sample Num: 3776, Cur Loss: 1.08051085, Cur Avg Loss: 1.32744414, Log Avg loss: 1.32805944, Global Avg Loss: 1.65442733, Time: 0.0403 Steps: 55590, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000246, Sample Num: 3936, Cur Loss: 1.57742047, Cur Avg Loss: 1.33590669, Log Avg loss: 1.53562292, Global Avg Loss: 1.65440597, Time: 0.0402 Steps: 55600, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000256, Sample Num: 4096, Cur Loss: 0.94030178, Cur Avg Loss: 1.33368006, Log Avg loss: 1.27890495, Global Avg Loss: 1.65433844, Time: 0.0402 Steps: 55610, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000266, Sample Num: 4256, Cur Loss: 0.78421891, Cur Avg Loss: 1.33444409, Log Avg loss: 1.35400308, Global Avg Loss: 1.65428444, Time: 0.0402 Steps: 55620, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000276, Sample Num: 4416, Cur Loss: 0.70130116, Cur Avg Loss: 1.32351607, Log Avg loss: 1.03283085, Global Avg Loss: 1.65417273, Time: 0.0402 Steps: 55630, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000286, Sample Num: 4576, Cur Loss: 1.79560351, Cur Avg Loss: 1.32395707, Log Avg loss: 1.33612849, Global Avg Loss: 1.65411557, Time: 0.0402 Steps: 55640, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000296, Sample Num: 4736, Cur Loss: 1.74879992, Cur Avg Loss: 1.33742519, Log Avg loss: 1.72261367, Global Avg Loss: 1.65412788, Time: 0.0402 Steps: 55650, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000306, Sample Num: 4896, Cur Loss: 1.73297429, Cur Avg Loss: 1.34148813, Log Avg loss: 1.46175111, Global Avg Loss: 1.65409332, Time: 0.0402 Steps: 55660, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000316, Sample Num: 5056, Cur Loss: 1.99189734, Cur Avg Loss: 1.33558292, Log Avg loss: 1.15488356, Global Avg Loss: 1.65400364, Time: 0.0402 Steps: 55670, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000326, Sample Num: 5216, Cur Loss: 1.48489690, Cur Avg Loss: 1.32398419, Log Avg loss: 0.95746421, Global Avg Loss: 1.65387855, Time: 0.0402 Steps: 55680, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000336, Sample Num: 5376, Cur Loss: 1.60311365, Cur Avg Loss: 1.32769433, Log Avg loss: 1.44864493, Global Avg Loss: 1.65384170, Time: 0.0402 Steps: 55690, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000346, Sample Num: 5536, Cur Loss: 3.06197023, Cur Avg Loss: 1.34020428, Log Avg loss: 1.76053870, Global Avg Loss: 1.65386085, Time: 0.0402 Steps: 55700, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000356, Sample Num: 5696, Cur Loss: 0.57234144, Cur Avg Loss: 1.33400921, Log Avg loss: 1.11965963, Global Avg Loss: 1.65376496, Time: 0.0402 Steps: 55710, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000366, Sample Num: 5856, Cur Loss: 0.82430238, Cur Avg Loss: 1.33329326, Log Avg loss: 1.30780554, Global Avg Loss: 1.65370287, Time: 0.0402 Steps: 55720, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000376, Sample Num: 6016, Cur Loss: 3.21370482, Cur Avg Loss: 1.33738057, Log Avg loss: 1.48697616, Global Avg Loss: 1.65367296, Time: 0.0402 Steps: 55730, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000386, Sample Num: 6176, Cur Loss: 1.32025766, Cur Avg Loss: 1.33227690, Log Avg loss: 1.14037881, Global Avg Loss: 1.65358087, Time: 0.0402 Steps: 55740, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000396, Sample Num: 6336, Cur Loss: 0.74867135, Cur Avg Loss: 1.33855005, Log Avg loss: 1.58069372, Global Avg Loss: 1.65356779, Time: 0.0402 Steps: 55750, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000406, Sample Num: 6496, Cur Loss: 0.37609363, Cur Avg Loss: 1.33249237, Log Avg loss: 1.09260796, Global Avg Loss: 1.65346719, Time: 0.0402 Steps: 55760, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000416, Sample Num: 6656, Cur Loss: 2.87716436, Cur Avg Loss: 1.33102826, Log Avg loss: 1.27158561, Global Avg Loss: 1.65339872, Time: 0.0402 Steps: 55770, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000426, Sample Num: 6816, Cur Loss: 1.44769454, Cur Avg Loss: 1.33308359, Log Avg loss: 1.41858513, Global Avg Loss: 1.65335662, Time: 0.0401 Steps: 55780, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000436, Sample Num: 6976, Cur Loss: 0.81761652, Cur Avg Loss: 1.34018424, Log Avg loss: 1.64267222, Global Avg Loss: 1.65335471, Time: 0.0402 Steps: 55790, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000446, Sample Num: 7136, Cur Loss: 1.71788383, Cur Avg Loss: 1.34390914, Log Avg loss: 1.50631473, Global Avg Loss: 1.65332835, Time: 0.0402 Steps: 55800, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000456, Sample Num: 7296, Cur Loss: 0.56169063, Cur Avg Loss: 1.33999328, Log Avg loss: 1.16534564, Global Avg Loss: 1.65324092, Time: 0.0402 Steps: 55810, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000466, Sample Num: 7456, Cur Loss: 0.99054599, Cur Avg Loss: 1.33982326, Log Avg loss: 1.33207048, Global Avg Loss: 1.65318338, Time: 0.0402 Steps: 55820, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000476, Sample Num: 7616, Cur Loss: 0.47245246, Cur Avg Loss: 1.33625008, Log Avg loss: 1.16974014, Global Avg Loss: 1.65309679, Time: 0.0402 Steps: 55830, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000486, Sample Num: 7776, Cur Loss: 1.36391091, Cur Avg Loss: 1.33956396, Log Avg loss: 1.49730465, Global Avg Loss: 1.65306889, Time: 0.0401 Steps: 55840, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000496, Sample Num: 7936, Cur Loss: 0.79244852, Cur Avg Loss: 1.33888901, Log Avg loss: 1.30608617, Global Avg Loss: 1.65300676, Time: 0.0402 Steps: 55850, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000506, Sample Num: 8096, Cur Loss: 1.36792052, Cur Avg Loss: 1.34409638, Log Avg loss: 1.60238175, Global Avg Loss: 1.65299770, Time: 0.0402 Steps: 55860, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000516, Sample Num: 8256, Cur Loss: 1.67452753, Cur Avg Loss: 1.34545701, Log Avg loss: 1.41430500, Global Avg Loss: 1.65295498, Time: 0.0403 Steps: 55870, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000526, Sample Num: 8416, Cur Loss: 1.51046348, Cur Avg Loss: 1.34479473, Log Avg loss: 1.31062142, Global Avg Loss: 1.65289371, Time: 0.0402 Steps: 55880, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000536, Sample Num: 8576, Cur Loss: 1.65403795, Cur Avg Loss: 1.34946333, Log Avg loss: 1.59503120, Global Avg Loss: 1.65288336, Time: 0.0403 Steps: 55890, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000546, Sample Num: 8736, Cur Loss: 0.57986283, Cur Avg Loss: 1.35328899, Log Avg loss: 1.55834438, Global Avg Loss: 1.65286645, Time: 0.0402 Steps: 55900, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000556, Sample Num: 8896, Cur Loss: 0.17262633, Cur Avg Loss: 1.34604324, Log Avg loss: 0.95042581, Global Avg Loss: 1.65274081, Time: 0.0401 Steps: 55910, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000566, Sample Num: 9056, Cur Loss: 1.77327764, Cur Avg Loss: 1.34415974, Log Avg loss: 1.23943694, Global Avg Loss: 1.65266690, Time: 0.0402 Steps: 55920, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000576, Sample Num: 9216, Cur Loss: 1.76890123, Cur Avg Loss: 1.34136583, Log Avg loss: 1.18323032, Global Avg Loss: 1.65258297, Time: 0.0402 Steps: 55930, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000586, Sample Num: 9376, Cur Loss: 0.90769720, Cur Avg Loss: 1.33573866, Log Avg loss: 1.01161387, Global Avg Loss: 1.65246839, Time: 0.0402 Steps: 55940, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000596, Sample Num: 9536, Cur Loss: 2.10483098, Cur Avg Loss: 1.33549110, Log Avg loss: 1.32098375, Global Avg Loss: 1.65240914, Time: 0.0402 Steps: 55950, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000606, Sample Num: 9696, Cur Loss: 0.98365062, Cur Avg Loss: 1.33714099, Log Avg loss: 1.43547465, Global Avg Loss: 1.65237037, Time: 0.0401 Steps: 55960, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000616, Sample Num: 9856, Cur Loss: 0.51930058, Cur Avg Loss: 1.33907204, Log Avg loss: 1.45609398, Global Avg Loss: 1.65233531, Time: 0.0401 Steps: 55970, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000626, Sample Num: 10016, Cur Loss: 1.21540821, Cur Avg Loss: 1.33575220, Log Avg loss: 1.13124993, Global Avg Loss: 1.65224222, Time: 0.0401 Steps: 55980, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000636, Sample Num: 10176, Cur Loss: 2.40999341, Cur Avg Loss: 1.34151885, Log Avg loss: 1.70251095, Global Avg Loss: 1.65225120, Time: 0.0401 Steps: 55990, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000646, Sample Num: 10336, Cur Loss: 1.57356596, Cur Avg Loss: 1.33678488, Log Avg loss: 1.03570447, Global Avg Loss: 1.65214110, Time: 0.0401 Steps: 56000, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000656, Sample Num: 10496, Cur Loss: 2.54632306, Cur Avg Loss: 1.33692089, Log Avg loss: 1.34570701, Global Avg Loss: 1.65208639, Time: 0.0559 Steps: 56010, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000666, Sample Num: 10656, Cur Loss: 1.55666411, Cur Avg Loss: 1.33271589, Log Avg loss: 1.05686795, Global Avg Loss: 1.65198014, Time: 0.0409 Steps: 56020, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000676, Sample Num: 10816, Cur Loss: 2.09853220, Cur Avg Loss: 1.33284982, Log Avg loss: 1.34176953, Global Avg Loss: 1.65192478, Time: 0.0402 Steps: 56030, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000686, Sample Num: 10976, Cur Loss: 0.64217049, Cur Avg Loss: 1.33167762, Log Avg loss: 1.25243729, Global Avg Loss: 1.65185349, Time: 0.0400 Steps: 56040, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000696, Sample Num: 11136, Cur Loss: 2.38919306, Cur Avg Loss: 1.33091913, Log Avg loss: 1.27888618, Global Avg Loss: 1.65178695, Time: 0.0400 Steps: 56050, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000706, Sample Num: 11296, Cur Loss: 2.78817034, Cur Avg Loss: 1.32887496, Log Avg loss: 1.18660088, Global Avg Loss: 1.65170397, Time: 0.0404 Steps: 56060, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000716, Sample Num: 11456, Cur Loss: 1.52387643, Cur Avg Loss: 1.32697185, Log Avg loss: 1.19261261, Global Avg Loss: 1.65162209, Time: 0.0406 Steps: 56070, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000726, Sample Num: 11616, Cur Loss: 0.23220372, Cur Avg Loss: 1.32297297, Log Avg loss: 1.03665255, Global Avg Loss: 1.65151243, Time: 0.0401 Steps: 56080, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000736, Sample Num: 11776, Cur Loss: 0.94229114, Cur Avg Loss: 1.32432706, Log Avg loss: 1.42263456, Global Avg Loss: 1.65147162, Time: 0.0510 Steps: 56090, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000746, Sample Num: 11936, Cur Loss: 2.36329794, Cur Avg Loss: 1.32365785, Log Avg loss: 1.27440391, Global Avg Loss: 1.65140441, Time: 0.0493 Steps: 56100, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000756, Sample Num: 12096, Cur Loss: 1.81040859, Cur Avg Loss: 1.31919603, Log Avg loss: 0.98634398, Global Avg Loss: 1.65128588, Time: 0.0497 Steps: 56110, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000766, Sample Num: 12256, Cur Loss: 1.29543447, Cur Avg Loss: 1.31904892, Log Avg loss: 1.30792719, Global Avg Loss: 1.65122470, Time: 0.0482 Steps: 56120, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000776, Sample Num: 12416, Cur Loss: 0.99519867, Cur Avg Loss: 1.32102896, Log Avg loss: 1.47270012, Global Avg Loss: 1.65119289, Time: 0.0402 Steps: 56130, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000786, Sample Num: 12576, Cur Loss: 0.90284461, Cur Avg Loss: 1.32334461, Log Avg loss: 1.50303932, Global Avg Loss: 1.65116650, Time: 0.0402 Steps: 56140, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000796, Sample Num: 12736, Cur Loss: 1.46487439, Cur Avg Loss: 1.32668503, Log Avg loss: 1.58924209, Global Avg Loss: 1.65115548, Time: 0.0402 Steps: 56150, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000806, Sample Num: 12896, Cur Loss: 1.16693616, Cur Avg Loss: 1.32428202, Log Avg loss: 1.13300241, Global Avg Loss: 1.65106321, Time: 0.0402 Steps: 56160, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000816, Sample Num: 13056, Cur Loss: 1.24825478, Cur Avg Loss: 1.32527706, Log Avg loss: 1.40547751, Global Avg Loss: 1.65101949, Time: 0.0402 Steps: 56170, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000826, Sample Num: 13216, Cur Loss: 0.99519205, Cur Avg Loss: 1.32298090, Log Avg loss: 1.13561435, Global Avg Loss: 1.65092775, Time: 0.0402 Steps: 56180, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000836, Sample Num: 13376, Cur Loss: 1.03110969, Cur Avg Loss: 1.32461989, Log Avg loss: 1.46000004, Global Avg Loss: 1.65089377, Time: 0.0402 Steps: 56190, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000846, Sample Num: 13536, Cur Loss: 0.94618678, Cur Avg Loss: 1.32476892, Log Avg loss: 1.33722821, Global Avg Loss: 1.65083796, Time: 0.0402 Steps: 56200, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000856, Sample Num: 13696, Cur Loss: 0.61682570, Cur Avg Loss: 1.32216286, Log Avg loss: 1.10169013, Global Avg Loss: 1.65074026, Time: 0.0402 Steps: 56210, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000866, Sample Num: 13856, Cur Loss: 0.71014249, Cur Avg Loss: 1.32636901, Log Avg loss: 1.68641507, Global Avg Loss: 1.65074661, Time: 0.0402 Steps: 56220, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000876, Sample Num: 14016, Cur Loss: 3.13966036, Cur Avg Loss: 1.32765219, Log Avg loss: 1.43877528, Global Avg Loss: 1.65070891, Time: 0.0402 Steps: 56230, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000886, Sample Num: 14176, Cur Loss: 1.01597822, Cur Avg Loss: 1.32494751, Log Avg loss: 1.08801763, Global Avg Loss: 1.65060886, Time: 0.0402 Steps: 56240, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000896, Sample Num: 14336, Cur Loss: 1.37300694, Cur Avg Loss: 1.32525329, Log Avg loss: 1.35234574, Global Avg Loss: 1.65055583, Time: 0.0402 Steps: 56250, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000906, Sample Num: 14496, Cur Loss: 0.62166208, Cur Avg Loss: 1.32621675, Log Avg loss: 1.41254281, Global Avg Loss: 1.65051353, Time: 0.0402 Steps: 56260, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000916, Sample Num: 14656, Cur Loss: 1.48837376, Cur Avg Loss: 1.32896438, Log Avg loss: 1.57789953, Global Avg Loss: 1.65050062, Time: 0.0402 Steps: 56270, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000926, Sample Num: 14816, Cur Loss: 0.70792282, Cur Avg Loss: 1.33098261, Log Avg loss: 1.51585218, Global Avg Loss: 1.65047670, Time: 0.0402 Steps: 56280, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000936, Sample Num: 14976, Cur Loss: 0.64995921, Cur Avg Loss: 1.33162742, Log Avg loss: 1.39133745, Global Avg Loss: 1.65043066, Time: 0.0402 Steps: 56290, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000946, Sample Num: 15136, Cur Loss: 2.13638473, Cur Avg Loss: 1.33266076, Log Avg loss: 1.42938152, Global Avg Loss: 1.65039140, Time: 0.0402 Steps: 56300, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000956, Sample Num: 15296, Cur Loss: 1.63902855, Cur Avg Loss: 1.33271815, Log Avg loss: 1.33814709, Global Avg Loss: 1.65033595, Time: 0.0402 Steps: 56310, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000966, Sample Num: 15456, Cur Loss: 0.69550985, Cur Avg Loss: 1.33008794, Log Avg loss: 1.07863935, Global Avg Loss: 1.65023444, Time: 0.0402 Steps: 56320, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000976, Sample Num: 15616, Cur Loss: 2.01378775, Cur Avg Loss: 1.33010036, Log Avg loss: 1.33130004, Global Avg Loss: 1.65017782, Time: 0.0402 Steps: 56330, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000986, Sample Num: 15776, Cur Loss: 0.57155585, Cur Avg Loss: 1.33293665, Log Avg loss: 1.60975890, Global Avg Loss: 1.65017065, Time: 0.0402 Steps: 56340, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000996, Sample Num: 15936, Cur Loss: 1.24121630, Cur Avg Loss: 1.33429519, Log Avg loss: 1.46824766, Global Avg Loss: 1.65013836, Time: 0.0402 Steps: 56350, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001006, Sample Num: 16096, Cur Loss: 1.06895232, Cur Avg Loss: 1.33676255, Log Avg loss: 1.58251122, Global Avg Loss: 1.65012636, Time: 0.0509 Steps: 56360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001016, Sample Num: 16256, Cur Loss: 0.60982156, Cur Avg Loss: 1.33475109, Log Avg loss: 1.13239827, Global Avg Loss: 1.65003452, Time: 0.0501 Steps: 56370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001026, Sample Num: 16416, Cur Loss: 0.58005852, Cur Avg Loss: 1.33027523, Log Avg loss: 0.87552814, Global Avg Loss: 1.64989715, Time: 0.0457 Steps: 56380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001036, Sample Num: 16576, Cur Loss: 2.15238404, Cur Avg Loss: 1.33242468, Log Avg loss: 1.55295764, Global Avg Loss: 1.64987996, Time: 0.0501 Steps: 56390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001046, Sample Num: 16736, Cur Loss: 1.05502677, Cur Avg Loss: 1.33110844, Log Avg loss: 1.19474635, Global Avg Loss: 1.64979926, Time: 0.0455 Steps: 56400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001056, Sample Num: 16896, Cur Loss: 0.46525347, Cur Avg Loss: 1.33164050, Log Avg loss: 1.38729380, Global Avg Loss: 1.64975272, Time: 0.0734 Steps: 56410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001066, Sample Num: 17056, Cur Loss: 0.79522723, Cur Avg Loss: 1.32863023, Log Avg loss: 1.01074555, Global Avg Loss: 1.64963946, Time: 0.0507 Steps: 56420, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001076, Sample Num: 17216, Cur Loss: 2.01793385, Cur Avg Loss: 1.33282722, Log Avg loss: 1.78022704, Global Avg Loss: 1.64966260, Time: 0.0402 Steps: 56430, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001086, Sample Num: 17376, Cur Loss: 1.98014235, Cur Avg Loss: 1.33066811, Log Avg loss: 1.09834687, Global Avg Loss: 1.64956492, Time: 0.0402 Steps: 56440, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001096, Sample Num: 17536, Cur Loss: 1.28146362, Cur Avg Loss: 1.32995335, Log Avg loss: 1.25233050, Global Avg Loss: 1.64949455, Time: 0.0402 Steps: 56450, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001106, Sample Num: 17696, Cur Loss: 2.64268470, Cur Avg Loss: 1.33361320, Log Avg loss: 1.73473306, Global Avg Loss: 1.64950965, Time: 0.0402 Steps: 56460, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001116, Sample Num: 17856, Cur Loss: 1.60258365, Cur Avg Loss: 1.33520410, Log Avg loss: 1.51115767, Global Avg Loss: 1.64948515, Time: 0.0401 Steps: 56470, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001126, Sample Num: 18016, Cur Loss: 0.80096626, Cur Avg Loss: 1.33584486, Log Avg loss: 1.40735326, Global Avg Loss: 1.64944228, Time: 0.0403 Steps: 56480, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001136, Sample Num: 18176, Cur Loss: 1.11203599, Cur Avg Loss: 1.33600231, Log Avg loss: 1.35373146, Global Avg Loss: 1.64938993, Time: 0.0401 Steps: 56490, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001146, Sample Num: 18336, Cur Loss: 1.15660727, Cur Avg Loss: 1.33751323, Log Avg loss: 1.50915440, Global Avg Loss: 1.64936511, Time: 0.0401 Steps: 56500, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001156, Sample Num: 18496, Cur Loss: 1.57983124, Cur Avg Loss: 1.33586089, Log Avg loss: 1.14650211, Global Avg Loss: 1.64927613, Time: 0.0402 Steps: 56510, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001166, Sample Num: 18656, Cur Loss: 1.06974626, Cur Avg Loss: 1.33935255, Log Avg loss: 1.74298844, Global Avg Loss: 1.64929271, Time: 0.0401 Steps: 56520, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001176, Sample Num: 18816, Cur Loss: 0.92682898, Cur Avg Loss: 1.33859840, Log Avg loss: 1.25066448, Global Avg Loss: 1.64922219, Time: 0.0402 Steps: 56530, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001186, Sample Num: 18976, Cur Loss: 1.26432991, Cur Avg Loss: 1.33756201, Log Avg loss: 1.21568256, Global Avg Loss: 1.64914551, Time: 0.0403 Steps: 56540, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001196, Sample Num: 19136, Cur Loss: 0.82901293, Cur Avg Loss: 1.34039771, Log Avg loss: 1.67671233, Global Avg Loss: 1.64915039, Time: 0.0402 Steps: 56550, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001206, Sample Num: 19296, Cur Loss: 1.16973233, Cur Avg Loss: 1.33848787, Log Avg loss: 1.11007010, Global Avg Loss: 1.64905508, Time: 0.0402 Steps: 56560, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001216, Sample Num: 19456, Cur Loss: 1.27694416, Cur Avg Loss: 1.33637003, Log Avg loss: 1.08095947, Global Avg Loss: 1.64895465, Time: 0.0402 Steps: 56570, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001226, Sample Num: 19616, Cur Loss: 1.25965178, Cur Avg Loss: 1.33891397, Log Avg loss: 1.64825664, Global Avg Loss: 1.64895453, Time: 0.0402 Steps: 56580, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001236, Sample Num: 19776, Cur Loss: 1.20177066, Cur Avg Loss: 1.33958068, Log Avg loss: 1.42131982, Global Avg Loss: 1.64891430, Time: 0.0402 Steps: 56590, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001246, Sample Num: 19936, Cur Loss: 0.49654061, Cur Avg Loss: 1.34027946, Log Avg loss: 1.42664875, Global Avg Loss: 1.64887503, Time: 0.0402 Steps: 56600, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001256, Sample Num: 20096, Cur Loss: 0.89632034, Cur Avg Loss: 1.34053678, Log Avg loss: 1.37259797, Global Avg Loss: 1.64882623, Time: 0.0403 Steps: 56610, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001266, Sample Num: 20256, Cur Loss: 1.26418710, Cur Avg Loss: 1.34232959, Log Avg loss: 1.56750658, Global Avg Loss: 1.64881187, Time: 0.0402 Steps: 56620, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001276, Sample Num: 20416, Cur Loss: 2.36767244, Cur Avg Loss: 1.34127890, Log Avg loss: 1.20826197, Global Avg Loss: 1.64873407, Time: 0.0402 Steps: 56630, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001286, Sample Num: 20576, Cur Loss: 0.86594468, Cur Avg Loss: 1.33972151, Log Avg loss: 1.14099878, Global Avg Loss: 1.64864443, Time: 0.0403 Steps: 56640, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001296, Sample Num: 20736, Cur Loss: 1.06893039, Cur Avg Loss: 1.34110976, Log Avg loss: 1.51963789, Global Avg Loss: 1.64862166, Time: 0.0402 Steps: 56650, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001306, Sample Num: 20896, Cur Loss: 0.92301100, Cur Avg Loss: 1.33905693, Log Avg loss: 1.07301053, Global Avg Loss: 1.64852007, Time: 0.0407 Steps: 56660, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001316, Sample Num: 21056, Cur Loss: 2.44007540, Cur Avg Loss: 1.34068251, Log Avg loss: 1.55298350, Global Avg Loss: 1.64850321, Time: 0.0458 Steps: 56670, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001326, Sample Num: 21216, Cur Loss: 1.18315268, Cur Avg Loss: 1.34188758, Log Avg loss: 1.50047398, Global Avg Loss: 1.64847709, Time: 0.0455 Steps: 56680, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001336, Sample Num: 21376, Cur Loss: 1.97922587, Cur Avg Loss: 1.34189132, Log Avg loss: 1.34238749, Global Avg Loss: 1.64842310, Time: 0.0545 Steps: 56690, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001346, Sample Num: 21536, Cur Loss: 0.30741143, Cur Avg Loss: 1.34231506, Log Avg loss: 1.39892648, Global Avg Loss: 1.64837910, Time: 0.0407 Steps: 56700, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001356, Sample Num: 21696, Cur Loss: 1.20303786, Cur Avg Loss: 1.34176002, Log Avg loss: 1.26705207, Global Avg Loss: 1.64831185, Time: 0.0408 Steps: 56710, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001366, Sample Num: 21856, Cur Loss: 1.15245068, Cur Avg Loss: 1.34271222, Log Avg loss: 1.47183034, Global Avg Loss: 1.64828074, Time: 0.0529 Steps: 56720, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001376, Sample Num: 22016, Cur Loss: 1.43679190, Cur Avg Loss: 1.34383013, Log Avg loss: 1.49653661, Global Avg Loss: 1.64825399, Time: 0.0403 Steps: 56730, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001386, Sample Num: 22176, Cur Loss: 0.98305982, Cur Avg Loss: 1.34371247, Log Avg loss: 1.32752240, Global Avg Loss: 1.64819747, Time: 0.0402 Steps: 56740, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001396, Sample Num: 22336, Cur Loss: 0.69927531, Cur Avg Loss: 1.34390897, Log Avg loss: 1.37114417, Global Avg Loss: 1.64814865, Time: 0.0403 Steps: 56750, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001406, Sample Num: 22496, Cur Loss: 0.80986893, Cur Avg Loss: 1.34290919, Log Avg loss: 1.20333996, Global Avg Loss: 1.64807028, Time: 0.0403 Steps: 56760, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001416, Sample Num: 22656, Cur Loss: 1.19123197, Cur Avg Loss: 1.34400099, Log Avg loss: 1.49750878, Global Avg Loss: 1.64804376, Time: 0.0402 Steps: 56770, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001426, Sample Num: 22816, Cur Loss: 0.61540788, Cur Avg Loss: 1.34345863, Log Avg loss: 1.26666011, Global Avg Loss: 1.64797659, Time: 0.0402 Steps: 56780, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001436, Sample Num: 22976, Cur Loss: 1.11736012, Cur Avg Loss: 1.34211164, Log Avg loss: 1.15003068, Global Avg Loss: 1.64788891, Time: 0.0402 Steps: 56790, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001446, Sample Num: 23136, Cur Loss: 0.85911036, Cur Avg Loss: 1.34414009, Log Avg loss: 1.63542580, Global Avg Loss: 1.64788671, Time: 0.0402 Steps: 56800, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001456, Sample Num: 23296, Cur Loss: 0.58205140, Cur Avg Loss: 1.34465394, Log Avg loss: 1.41895562, Global Avg Loss: 1.64784641, Time: 0.0402 Steps: 56810, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001466, Sample Num: 23456, Cur Loss: 0.65509593, Cur Avg Loss: 1.34528814, Log Avg loss: 1.43762762, Global Avg Loss: 1.64780942, Time: 0.0402 Steps: 56820, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001476, Sample Num: 23616, Cur Loss: 1.40705085, Cur Avg Loss: 1.34555015, Log Avg loss: 1.38396145, Global Avg Loss: 1.64776299, Time: 0.0402 Steps: 56830, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001486, Sample Num: 23776, Cur Loss: 0.95541787, Cur Avg Loss: 1.34592536, Log Avg loss: 1.40130712, Global Avg Loss: 1.64771963, Time: 0.0402 Steps: 56840, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001496, Sample Num: 23936, Cur Loss: 2.92978692, Cur Avg Loss: 1.34820193, Log Avg loss: 1.68649942, Global Avg Loss: 1.64772645, Time: 0.0402 Steps: 56850, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001506, Sample Num: 24096, Cur Loss: 2.13445759, Cur Avg Loss: 1.34942058, Log Avg loss: 1.53173059, Global Avg Loss: 1.64770605, Time: 0.0402 Steps: 56860, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001516, Sample Num: 24256, Cur Loss: 1.72956300, Cur Avg Loss: 1.34969107, Log Avg loss: 1.39042776, Global Avg Loss: 1.64766081, Time: 0.0402 Steps: 56870, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001526, Sample Num: 24416, Cur Loss: 0.99581623, Cur Avg Loss: 1.34865164, Log Avg loss: 1.19107318, Global Avg Loss: 1.64758054, Time: 0.0401 Steps: 56880, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001536, Sample Num: 24576, Cur Loss: 1.03375483, Cur Avg Loss: 1.34856398, Log Avg loss: 1.33518704, Global Avg Loss: 1.64752563, Time: 0.0403 Steps: 56890, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001546, Sample Num: 24736, Cur Loss: 0.62759018, Cur Avg Loss: 1.34679115, Log Avg loss: 1.07448445, Global Avg Loss: 1.64742492, Time: 0.0402 Steps: 56900, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001556, Sample Num: 24896, Cur Loss: 0.78504813, Cur Avg Loss: 1.34429045, Log Avg loss: 0.95768177, Global Avg Loss: 1.64730372, Time: 0.0402 Steps: 56910, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001566, Sample Num: 25056, Cur Loss: 0.65460068, Cur Avg Loss: 1.34256692, Log Avg loss: 1.07438689, Global Avg Loss: 1.64720307, Time: 0.0401 Steps: 56920, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001576, Sample Num: 25216, Cur Loss: 0.97004557, Cur Avg Loss: 1.34100933, Log Avg loss: 1.09708946, Global Avg Loss: 1.64710644, Time: 0.0402 Steps: 56930, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001586, Sample Num: 25376, Cur Loss: 1.70722580, Cur Avg Loss: 1.33917905, Log Avg loss: 1.05072708, Global Avg Loss: 1.64700170, Time: 0.0402 Steps: 56940, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001596, Sample Num: 25536, Cur Loss: 2.01985073, Cur Avg Loss: 1.33935529, Log Avg loss: 1.36730681, Global Avg Loss: 1.64695259, Time: 0.0402 Steps: 56950, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001606, Sample Num: 25696, Cur Loss: 2.61865044, Cur Avg Loss: 1.34133883, Log Avg loss: 1.65791224, Global Avg Loss: 1.64695451, Time: 0.0432 Steps: 56960, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001616, Sample Num: 25856, Cur Loss: 1.84365869, Cur Avg Loss: 1.34233110, Log Avg loss: 1.50168938, Global Avg Loss: 1.64692901, Time: 0.0490 Steps: 56970, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001626, Sample Num: 26016, Cur Loss: 1.69467187, Cur Avg Loss: 1.34157856, Log Avg loss: 1.21996882, Global Avg Loss: 1.64685408, Time: 0.0539 Steps: 56980, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001636, Sample Num: 26176, Cur Loss: 1.98001301, Cur Avg Loss: 1.34151079, Log Avg loss: 1.33049112, Global Avg Loss: 1.64679857, Time: 0.0407 Steps: 56990, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001646, Sample Num: 26336, Cur Loss: 1.38700747, Cur Avg Loss: 1.34102364, Log Avg loss: 1.26132535, Global Avg Loss: 1.64673094, Time: 0.0602 Steps: 57000, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001656, Sample Num: 26496, Cur Loss: 1.26948833, Cur Avg Loss: 1.34213218, Log Avg loss: 1.52459913, Global Avg Loss: 1.64670952, Time: 0.0407 Steps: 57010, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001666, Sample Num: 26656, Cur Loss: 0.69457096, Cur Avg Loss: 1.34208637, Log Avg loss: 1.33450025, Global Avg Loss: 1.64665476, Time: 0.0402 Steps: 57020, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001676, Sample Num: 26816, Cur Loss: 1.58484936, Cur Avg Loss: 1.34241666, Log Avg loss: 1.39744306, Global Avg Loss: 1.64661106, Time: 0.0402 Steps: 57030, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001686, Sample Num: 26976, Cur Loss: 0.87339622, Cur Avg Loss: 1.34289418, Log Avg loss: 1.42292606, Global Avg Loss: 1.64657185, Time: 0.0402 Steps: 57040, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001696, Sample Num: 27136, Cur Loss: 1.60407722, Cur Avg Loss: 1.34269694, Log Avg loss: 1.30944211, Global Avg Loss: 1.64651276, Time: 0.0402 Steps: 57050, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001706, Sample Num: 27296, Cur Loss: 1.06642962, Cur Avg Loss: 1.34220662, Log Avg loss: 1.25904880, Global Avg Loss: 1.64644485, Time: 0.0402 Steps: 57060, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001716, Sample Num: 27456, Cur Loss: 1.45599318, Cur Avg Loss: 1.34199694, Log Avg loss: 1.30622410, Global Avg Loss: 1.64638524, Time: 0.0403 Steps: 57070, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001726, Sample Num: 27616, Cur Loss: 1.39003992, Cur Avg Loss: 1.34198615, Log Avg loss: 1.34013469, Global Avg Loss: 1.64633158, Time: 0.0402 Steps: 57080, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001736, Sample Num: 27776, Cur Loss: 1.14148057, Cur Avg Loss: 1.34305153, Log Avg loss: 1.52693752, Global Avg Loss: 1.64631067, Time: 0.0402 Steps: 57090, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001746, Sample Num: 27936, Cur Loss: 1.01436925, Cur Avg Loss: 1.34492592, Log Avg loss: 1.67031969, Global Avg Loss: 1.64631488, Time: 0.0402 Steps: 57100, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001756, Sample Num: 28096, Cur Loss: 1.99901485, Cur Avg Loss: 1.34579980, Log Avg loss: 1.49837942, Global Avg Loss: 1.64628897, Time: 0.0402 Steps: 57110, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001766, Sample Num: 28256, Cur Loss: 1.40549183, Cur Avg Loss: 1.34560004, Log Avg loss: 1.31052185, Global Avg Loss: 1.64623019, Time: 0.0403 Steps: 57120, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001776, Sample Num: 28416, Cur Loss: 1.34037447, Cur Avg Loss: 1.34423134, Log Avg loss: 1.10251956, Global Avg Loss: 1.64613502, Time: 0.0402 Steps: 57130, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001786, Sample Num: 28576, Cur Loss: 1.78248048, Cur Avg Loss: 1.34351168, Log Avg loss: 1.21569875, Global Avg Loss: 1.64605969, Time: 0.0402 Steps: 57140, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001796, Sample Num: 28736, Cur Loss: 0.57862514, Cur Avg Loss: 1.34420436, Log Avg loss: 1.46791704, Global Avg Loss: 1.64602852, Time: 0.0402 Steps: 57150, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001806, Sample Num: 28896, Cur Loss: 3.36756396, Cur Avg Loss: 1.34446039, Log Avg loss: 1.39044415, Global Avg Loss: 1.64598380, Time: 0.0402 Steps: 57160, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001816, Sample Num: 29056, Cur Loss: 1.29473627, Cur Avg Loss: 1.34463717, Log Avg loss: 1.37656415, Global Avg Loss: 1.64593668, Time: 0.0402 Steps: 57170, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001826, Sample Num: 29216, Cur Loss: 1.31291723, Cur Avg Loss: 1.34529146, Log Avg loss: 1.46410968, Global Avg Loss: 1.64590488, Time: 0.0401 Steps: 57180, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001836, Sample Num: 29376, Cur Loss: 0.56605923, Cur Avg Loss: 1.34634886, Log Avg loss: 1.53943069, Global Avg Loss: 1.64588626, Time: 0.0402 Steps: 57190, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001846, Sample Num: 29536, Cur Loss: 0.85604548, Cur Avg Loss: 1.34555814, Log Avg loss: 1.20038202, Global Avg Loss: 1.64580837, Time: 0.0402 Steps: 57200, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001856, Sample Num: 29696, Cur Loss: 1.92238414, Cur Avg Loss: 1.34570736, Log Avg loss: 1.37325315, Global Avg Loss: 1.64576073, Time: 0.0402 Steps: 57210, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001866, Sample Num: 29856, Cur Loss: 0.78811461, Cur Avg Loss: 1.34493200, Log Avg loss: 1.20102505, Global Avg Loss: 1.64568301, Time: 0.0402 Steps: 57220, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001876, Sample Num: 30016, Cur Loss: 0.76327085, Cur Avg Loss: 1.34388103, Log Avg loss: 1.14776967, Global Avg Loss: 1.64559601, Time: 0.0401 Steps: 57230, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001886, Sample Num: 30176, Cur Loss: 0.75117010, Cur Avg Loss: 1.34313139, Log Avg loss: 1.20249823, Global Avg Loss: 1.64551860, Time: 0.0401 Steps: 57240, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001896, Sample Num: 30336, Cur Loss: 0.71584761, Cur Avg Loss: 1.34536148, Log Avg loss: 1.76595680, Global Avg Loss: 1.64553963, Time: 0.0402 Steps: 57250, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001906, Sample Num: 30496, Cur Loss: 1.49011791, Cur Avg Loss: 1.34575238, Log Avg loss: 1.41986669, Global Avg Loss: 1.64550022, Time: 0.0565 Steps: 57260, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001916, Sample Num: 30656, Cur Loss: 0.85101390, Cur Avg Loss: 1.34437936, Log Avg loss: 1.08268311, Global Avg Loss: 1.64540195, Time: 0.0484 Steps: 57270, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001926, Sample Num: 30816, Cur Loss: 1.59988105, Cur Avg Loss: 1.34442316, Log Avg loss: 1.35281499, Global Avg Loss: 1.64535087, Time: 0.0532 Steps: 57280, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001936, Sample Num: 30976, Cur Loss: 1.66091657, Cur Avg Loss: 1.34416829, Log Avg loss: 1.29507976, Global Avg Loss: 1.64528973, Time: 0.0482 Steps: 57290, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001946, Sample Num: 31136, Cur Loss: 0.41197729, Cur Avg Loss: 1.34490234, Log Avg loss: 1.48701379, Global Avg Loss: 1.64526211, Time: 0.0539 Steps: 57300, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001956, Sample Num: 31296, Cur Loss: 1.47672844, Cur Avg Loss: 1.34583285, Log Avg loss: 1.52691091, Global Avg Loss: 1.64524145, Time: 0.0502 Steps: 57310, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001966, Sample Num: 31456, Cur Loss: 1.31615019, Cur Avg Loss: 1.34575152, Log Avg loss: 1.32984346, Global Avg Loss: 1.64518643, Time: 0.0402 Steps: 57320, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001976, Sample Num: 31616, Cur Loss: 2.39054918, Cur Avg Loss: 1.34500533, Log Avg loss: 1.19830336, Global Avg Loss: 1.64510848, Time: 0.0402 Steps: 57330, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001986, Sample Num: 31776, Cur Loss: 0.77098429, Cur Avg Loss: 1.34429336, Log Avg loss: 1.20360928, Global Avg Loss: 1.64503148, Time: 0.0402 Steps: 57340, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001996, Sample Num: 31936, Cur Loss: 1.10653818, Cur Avg Loss: 1.34316935, Log Avg loss: 1.11994062, Global Avg Loss: 1.64493993, Time: 0.0402 Steps: 57350, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002006, Sample Num: 32096, Cur Loss: 1.95531166, Cur Avg Loss: 1.34268643, Log Avg loss: 1.24629484, Global Avg Loss: 1.64487043, Time: 0.0401 Steps: 57360, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002016, Sample Num: 32256, Cur Loss: 0.55228043, Cur Avg Loss: 1.34210883, Log Avg loss: 1.22624244, Global Avg Loss: 1.64479746, Time: 0.0402 Steps: 57370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002026, Sample Num: 32416, Cur Loss: 1.85896647, Cur Avg Loss: 1.34208442, Log Avg loss: 1.33716407, Global Avg Loss: 1.64474384, Time: 0.0402 Steps: 57380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002036, Sample Num: 32576, Cur Loss: 1.94521308, Cur Avg Loss: 1.34366577, Log Avg loss: 1.66404775, Global Avg Loss: 1.64474721, Time: 0.0402 Steps: 57390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002046, Sample Num: 32736, Cur Loss: 1.02236974, Cur Avg Loss: 1.34360745, Log Avg loss: 1.33173322, Global Avg Loss: 1.64469268, Time: 0.0402 Steps: 57400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002056, Sample Num: 32896, Cur Loss: 2.26075625, Cur Avg Loss: 1.34319045, Log Avg loss: 1.25787234, Global Avg Loss: 1.64462530, Time: 0.0402 Steps: 57410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002066, Sample Num: 33056, Cur Loss: 0.62430316, Cur Avg Loss: 1.34337919, Log Avg loss: 1.38218442, Global Avg Loss: 1.64457959, Time: 0.0402 Steps: 57420, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002076, Sample Num: 33216, Cur Loss: 1.80010092, Cur Avg Loss: 1.34417034, Log Avg loss: 1.50762057, Global Avg Loss: 1.64455574, Time: 0.0402 Steps: 57430, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002086, Sample Num: 33376, Cur Loss: 1.68190050, Cur Avg Loss: 1.34365788, Log Avg loss: 1.23727218, Global Avg Loss: 1.64448484, Time: 0.0402 Steps: 57440, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002096, Sample Num: 33536, Cur Loss: 0.48162773, Cur Avg Loss: 1.34332179, Log Avg loss: 1.27321216, Global Avg Loss: 1.64442021, Time: 0.0402 Steps: 57450, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002106, Sample Num: 33696, Cur Loss: 0.99788767, Cur Avg Loss: 1.34410997, Log Avg loss: 1.50931347, Global Avg Loss: 1.64439670, Time: 0.0401 Steps: 57460, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002116, Sample Num: 33856, Cur Loss: 0.45222694, Cur Avg Loss: 1.34343200, Log Avg loss: 1.20065076, Global Avg Loss: 1.64431949, Time: 0.0402 Steps: 57470, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002126, Sample Num: 34016, Cur Loss: 0.66513824, Cur Avg Loss: 1.34284832, Log Avg loss: 1.21934255, Global Avg Loss: 1.64424555, Time: 0.0402 Steps: 57480, Updated lr: 0.000046 ***** Running evaluation checkpoint-57483 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-57483 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 89.904410, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.28355, "eval_total_loss": 902.33586, "eval_mae": 0.970594, "eval_mse": 1.283494, "eval_r2": 0.184127, "eval_sp_statistic": 0.386073, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.472778, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.150121, "test_total_loss": 577.360877, "test_mae": 0.768707, "test_mse": 1.150427, "test_r2": 0.257504, "test_sp_statistic": 0.428372, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.514712, "test_ps_pvalue": 0.0, "lr": 4.6436225699383594e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6441904820132793, "train_cur_epoch_loss": 2856.662774026394, "train_cur_epoch_avg_loss": 1.34178617849995, "train_cur_epoch_time": 89.90441012382507, "train_cur_epoch_avg_time": 0.0422284688228394, "epoch": 27, "step": 57483} ################################################## Training, Epoch: 0028, Batch: 000007, Sample Num: 112, Cur Loss: 1.06103349, Cur Avg Loss: 1.65347951, Log Avg loss: 1.33416014, Global Avg Loss: 1.64419161, Time: 0.0402 Steps: 57490, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000017, Sample Num: 272, Cur Loss: 0.49441445, Cur Avg Loss: 1.41046348, Log Avg loss: 1.24035227, Global Avg Loss: 1.64412138, Time: 0.0402 Steps: 57500, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000027, Sample Num: 432, Cur Loss: 1.44470251, Cur Avg Loss: 1.25765701, Log Avg loss: 0.99788602, Global Avg Loss: 1.64400901, Time: 0.0402 Steps: 57510, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000037, Sample Num: 592, Cur Loss: 1.05931818, Cur Avg Loss: 1.31012921, Log Avg loss: 1.45180412, Global Avg Loss: 1.64397560, Time: 0.0402 Steps: 57520, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000047, Sample Num: 752, Cur Loss: 1.37655854, Cur Avg Loss: 1.31043486, Log Avg loss: 1.31156579, Global Avg Loss: 1.64391782, Time: 0.0590 Steps: 57530, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000057, Sample Num: 912, Cur Loss: 1.43562245, Cur Avg Loss: 1.34091283, Log Avg loss: 1.48415928, Global Avg Loss: 1.64389005, Time: 0.0407 Steps: 57540, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000067, Sample Num: 1072, Cur Loss: 1.00783503, Cur Avg Loss: 1.40748345, Log Avg loss: 1.78693597, Global Avg Loss: 1.64391491, Time: 0.0427 Steps: 57550, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000077, Sample Num: 1232, Cur Loss: 1.07789207, Cur Avg Loss: 1.39727057, Log Avg loss: 1.32884429, Global Avg Loss: 1.64386017, Time: 0.0407 Steps: 57560, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000087, Sample Num: 1392, Cur Loss: 0.80390936, Cur Avg Loss: 1.33697600, Log Avg loss: 0.87270778, Global Avg Loss: 1.64372622, Time: 0.0496 Steps: 57570, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000097, Sample Num: 1552, Cur Loss: 1.58679271, Cur Avg Loss: 1.32721521, Log Avg loss: 1.24229638, Global Avg Loss: 1.64365650, Time: 0.0404 Steps: 57580, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000107, Sample Num: 1712, Cur Loss: 0.91925859, Cur Avg Loss: 1.34200715, Log Avg loss: 1.48548895, Global Avg Loss: 1.64362904, Time: 0.0402 Steps: 57590, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000117, Sample Num: 1872, Cur Loss: 1.63612831, Cur Avg Loss: 1.33421659, Log Avg loss: 1.25085762, Global Avg Loss: 1.64356085, Time: 0.0402 Steps: 57600, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000127, Sample Num: 2032, Cur Loss: 1.73462415, Cur Avg Loss: 1.35507204, Log Avg loss: 1.59908078, Global Avg Loss: 1.64355313, Time: 0.0402 Steps: 57610, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000137, Sample Num: 2192, Cur Loss: 0.46417627, Cur Avg Loss: 1.34303825, Log Avg loss: 1.19020912, Global Avg Loss: 1.64347445, Time: 0.0402 Steps: 57620, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000147, Sample Num: 2352, Cur Loss: 1.73715222, Cur Avg Loss: 1.34350863, Log Avg loss: 1.34995278, Global Avg Loss: 1.64342352, Time: 0.0402 Steps: 57630, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000157, Sample Num: 2512, Cur Loss: 1.27893221, Cur Avg Loss: 1.33680502, Log Avg loss: 1.23826202, Global Avg Loss: 1.64335322, Time: 0.0402 Steps: 57640, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000167, Sample Num: 2672, Cur Loss: 2.89942646, Cur Avg Loss: 1.34100845, Log Avg loss: 1.40700236, Global Avg Loss: 1.64331223, Time: 0.0403 Steps: 57650, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000177, Sample Num: 2832, Cur Loss: 1.15233302, Cur Avg Loss: 1.33923686, Log Avg loss: 1.30965121, Global Avg Loss: 1.64325436, Time: 0.0402 Steps: 57660, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000187, Sample Num: 2992, Cur Loss: 2.07550645, Cur Avg Loss: 1.34794573, Log Avg loss: 1.50209283, Global Avg Loss: 1.64322988, Time: 0.0402 Steps: 57670, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000197, Sample Num: 3152, Cur Loss: 0.62732822, Cur Avg Loss: 1.34095460, Log Avg loss: 1.21022049, Global Avg Loss: 1.64315481, Time: 0.0402 Steps: 57680, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000207, Sample Num: 3312, Cur Loss: 1.03669155, Cur Avg Loss: 1.33647928, Log Avg loss: 1.24831544, Global Avg Loss: 1.64308637, Time: 0.0402 Steps: 57690, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000217, Sample Num: 3472, Cur Loss: 0.59479195, Cur Avg Loss: 1.32805560, Log Avg loss: 1.15368541, Global Avg Loss: 1.64300155, Time: 0.0402 Steps: 57700, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000227, Sample Num: 3632, Cur Loss: 1.40970039, Cur Avg Loss: 1.32920682, Log Avg loss: 1.35418823, Global Avg Loss: 1.64295151, Time: 0.0402 Steps: 57710, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000237, Sample Num: 3792, Cur Loss: 1.58096516, Cur Avg Loss: 1.33577125, Log Avg loss: 1.48478389, Global Avg Loss: 1.64292410, Time: 0.0402 Steps: 57720, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000247, Sample Num: 3952, Cur Loss: 0.73617339, Cur Avg Loss: 1.32403733, Log Avg loss: 1.04594331, Global Avg Loss: 1.64282069, Time: 0.0402 Steps: 57730, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000257, Sample Num: 4112, Cur Loss: 2.24204731, Cur Avg Loss: 1.31961337, Log Avg loss: 1.21034162, Global Avg Loss: 1.64274579, Time: 0.0402 Steps: 57740, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000267, Sample Num: 4272, Cur Loss: 1.40227962, Cur Avg Loss: 1.31028560, Log Avg loss: 1.07056187, Global Avg Loss: 1.64264671, Time: 0.0402 Steps: 57750, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000277, Sample Num: 4432, Cur Loss: 1.53246248, Cur Avg Loss: 1.31650405, Log Avg loss: 1.48253677, Global Avg Loss: 1.64261899, Time: 0.0402 Steps: 57760, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000287, Sample Num: 4592, Cur Loss: 1.49616051, Cur Avg Loss: 1.31732154, Log Avg loss: 1.33996584, Global Avg Loss: 1.64256660, Time: 0.0402 Steps: 57770, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000297, Sample Num: 4752, Cur Loss: 1.46690154, Cur Avg Loss: 1.31464157, Log Avg loss: 1.23772660, Global Avg Loss: 1.64249654, Time: 0.0402 Steps: 57780, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000307, Sample Num: 4912, Cur Loss: 2.26432061, Cur Avg Loss: 1.32593377, Log Avg loss: 1.66131213, Global Avg Loss: 1.64249979, Time: 0.0402 Steps: 57790, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000317, Sample Num: 5072, Cur Loss: 0.77179474, Cur Avg Loss: 1.32827216, Log Avg loss: 1.40006071, Global Avg Loss: 1.64245785, Time: 0.0402 Steps: 57800, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000327, Sample Num: 5232, Cur Loss: 2.12197161, Cur Avg Loss: 1.33293920, Log Avg loss: 1.48088434, Global Avg Loss: 1.64242990, Time: 0.0402 Steps: 57810, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000337, Sample Num: 5392, Cur Loss: 2.64211655, Cur Avg Loss: 1.32870332, Log Avg loss: 1.19018991, Global Avg Loss: 1.64235169, Time: 0.0402 Steps: 57820, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000347, Sample Num: 5552, Cur Loss: 1.45806313, Cur Avg Loss: 1.32590660, Log Avg loss: 1.23165724, Global Avg Loss: 1.64228067, Time: 0.0537 Steps: 57830, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000357, Sample Num: 5712, Cur Loss: 1.45584095, Cur Avg Loss: 1.32550142, Log Avg loss: 1.31144158, Global Avg Loss: 1.64222347, Time: 0.0407 Steps: 57840, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000367, Sample Num: 5872, Cur Loss: 1.82904959, Cur Avg Loss: 1.33041179, Log Avg loss: 1.50571213, Global Avg Loss: 1.64219987, Time: 0.0557 Steps: 57850, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000377, Sample Num: 6032, Cur Loss: 0.81436688, Cur Avg Loss: 1.32994501, Log Avg loss: 1.31281419, Global Avg Loss: 1.64214294, Time: 0.0663 Steps: 57860, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000387, Sample Num: 6192, Cur Loss: 1.70592666, Cur Avg Loss: 1.33460882, Log Avg loss: 1.51043425, Global Avg Loss: 1.64212018, Time: 0.0520 Steps: 57870, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000397, Sample Num: 6352, Cur Loss: 0.63380730, Cur Avg Loss: 1.32488034, Log Avg loss: 0.94838820, Global Avg Loss: 1.64200033, Time: 0.0513 Steps: 57880, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000407, Sample Num: 6512, Cur Loss: 0.28853914, Cur Avg Loss: 1.32253004, Log Avg loss: 1.22922345, Global Avg Loss: 1.64192902, Time: 0.0457 Steps: 57890, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000417, Sample Num: 6672, Cur Loss: 0.67720193, Cur Avg Loss: 1.32327595, Log Avg loss: 1.35363439, Global Avg Loss: 1.64187923, Time: 0.0402 Steps: 57900, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000427, Sample Num: 6832, Cur Loss: 1.64333630, Cur Avg Loss: 1.31975964, Log Avg loss: 1.17312959, Global Avg Loss: 1.64179829, Time: 0.0402 Steps: 57910, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000437, Sample Num: 6992, Cur Loss: 1.19791627, Cur Avg Loss: 1.31492130, Log Avg loss: 1.10832390, Global Avg Loss: 1.64170618, Time: 0.0402 Steps: 57920, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000447, Sample Num: 7152, Cur Loss: 0.27844661, Cur Avg Loss: 1.31409884, Log Avg loss: 1.27815728, Global Avg Loss: 1.64164343, Time: 0.0402 Steps: 57930, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000457, Sample Num: 7312, Cur Loss: 2.12303829, Cur Avg Loss: 1.31823714, Log Avg loss: 1.50321916, Global Avg Loss: 1.64161953, Time: 0.0402 Steps: 57940, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000467, Sample Num: 7472, Cur Loss: 0.43855107, Cur Avg Loss: 1.31057457, Log Avg loss: 0.96039511, Global Avg Loss: 1.64150198, Time: 0.0402 Steps: 57950, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000477, Sample Num: 7632, Cur Loss: 1.37898815, Cur Avg Loss: 1.31711251, Log Avg loss: 1.62243437, Global Avg Loss: 1.64149869, Time: 0.0402 Steps: 57960, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000487, Sample Num: 7792, Cur Loss: 1.40448856, Cur Avg Loss: 1.31420338, Log Avg loss: 1.17543817, Global Avg Loss: 1.64141829, Time: 0.0401 Steps: 57970, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000497, Sample Num: 7952, Cur Loss: 0.63528252, Cur Avg Loss: 1.31575247, Log Avg loss: 1.39119323, Global Avg Loss: 1.64137514, Time: 0.0402 Steps: 57980, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000507, Sample Num: 8112, Cur Loss: 1.09078538, Cur Avg Loss: 1.32003413, Log Avg loss: 1.53283236, Global Avg Loss: 1.64135642, Time: 0.0402 Steps: 57990, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000517, Sample Num: 8272, Cur Loss: 0.87481558, Cur Avg Loss: 1.32291039, Log Avg loss: 1.46873685, Global Avg Loss: 1.64132666, Time: 0.0402 Steps: 58000, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000527, Sample Num: 8432, Cur Loss: 1.33108294, Cur Avg Loss: 1.31922516, Log Avg loss: 1.12869868, Global Avg Loss: 1.64123829, Time: 0.0402 Steps: 58010, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000537, Sample Num: 8592, Cur Loss: 0.53525746, Cur Avg Loss: 1.31713548, Log Avg loss: 1.20700927, Global Avg Loss: 1.64116345, Time: 0.0401 Steps: 58020, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000547, Sample Num: 8752, Cur Loss: 2.56904125, Cur Avg Loss: 1.31835041, Log Avg loss: 1.38359239, Global Avg Loss: 1.64111906, Time: 0.0402 Steps: 58030, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000557, Sample Num: 8912, Cur Loss: 2.11112046, Cur Avg Loss: 1.31889146, Log Avg loss: 1.34848652, Global Avg Loss: 1.64106864, Time: 0.0402 Steps: 58040, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000567, Sample Num: 9072, Cur Loss: 0.84153682, Cur Avg Loss: 1.31907576, Log Avg loss: 1.32934136, Global Avg Loss: 1.64101494, Time: 0.0402 Steps: 58050, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000577, Sample Num: 9232, Cur Loss: 1.04101026, Cur Avg Loss: 1.32143183, Log Avg loss: 1.45502122, Global Avg Loss: 1.64098291, Time: 0.0402 Steps: 58060, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000587, Sample Num: 9392, Cur Loss: 1.27853990, Cur Avg Loss: 1.32243651, Log Avg loss: 1.38040652, Global Avg Loss: 1.64093804, Time: 0.0402 Steps: 58070, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000597, Sample Num: 9552, Cur Loss: 1.55645037, Cur Avg Loss: 1.31764138, Log Avg loss: 1.03616715, Global Avg Loss: 1.64083391, Time: 0.0402 Steps: 58080, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000607, Sample Num: 9712, Cur Loss: 0.52950013, Cur Avg Loss: 1.31625609, Log Avg loss: 1.23355432, Global Avg Loss: 1.64076380, Time: 0.0402 Steps: 58090, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000617, Sample Num: 9872, Cur Loss: 0.73848248, Cur Avg Loss: 1.31645739, Log Avg loss: 1.32867649, Global Avg Loss: 1.64071008, Time: 0.0402 Steps: 58100, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000627, Sample Num: 10032, Cur Loss: 0.95273328, Cur Avg Loss: 1.31425225, Log Avg loss: 1.17819495, Global Avg Loss: 1.64063049, Time: 0.0402 Steps: 58110, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000637, Sample Num: 10192, Cur Loss: 1.18067896, Cur Avg Loss: 1.30861689, Log Avg loss: 0.95527971, Global Avg Loss: 1.64051257, Time: 0.0438 Steps: 58120, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000647, Sample Num: 10352, Cur Loss: 1.18485677, Cur Avg Loss: 1.31312000, Log Avg loss: 1.59996850, Global Avg Loss: 1.64050559, Time: 0.0540 Steps: 58130, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000657, Sample Num: 10512, Cur Loss: 1.82582510, Cur Avg Loss: 1.31044586, Log Avg loss: 1.13742883, Global Avg Loss: 1.64041906, Time: 0.0406 Steps: 58140, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000667, Sample Num: 10672, Cur Loss: 0.80147034, Cur Avg Loss: 1.30749993, Log Avg loss: 1.11395227, Global Avg Loss: 1.64032853, Time: 0.0407 Steps: 58150, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000677, Sample Num: 10832, Cur Loss: 2.12768245, Cur Avg Loss: 1.30872718, Log Avg loss: 1.39058458, Global Avg Loss: 1.64028559, Time: 0.0436 Steps: 58160, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000687, Sample Num: 10992, Cur Loss: 0.63862473, Cur Avg Loss: 1.30442157, Log Avg loss: 1.01293183, Global Avg Loss: 1.64017774, Time: 0.0482 Steps: 58170, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000697, Sample Num: 11152, Cur Loss: 1.61697316, Cur Avg Loss: 1.30593651, Log Avg loss: 1.41001308, Global Avg Loss: 1.64013818, Time: 0.0481 Steps: 58180, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000707, Sample Num: 11312, Cur Loss: 1.22370017, Cur Avg Loss: 1.30438241, Log Avg loss: 1.19606162, Global Avg Loss: 1.64006186, Time: 0.0403 Steps: 58190, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000717, Sample Num: 11472, Cur Loss: 0.87230259, Cur Avg Loss: 1.30479642, Log Avg loss: 1.33406683, Global Avg Loss: 1.64000929, Time: 0.0402 Steps: 58200, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000727, Sample Num: 11632, Cur Loss: 0.94340241, Cur Avg Loss: 1.30101551, Log Avg loss: 1.02992449, Global Avg Loss: 1.63990448, Time: 0.0401 Steps: 58210, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000737, Sample Num: 11792, Cur Loss: 0.98141688, Cur Avg Loss: 1.30161627, Log Avg loss: 1.34529129, Global Avg Loss: 1.63985388, Time: 0.0402 Steps: 58220, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000747, Sample Num: 11952, Cur Loss: 1.03900349, Cur Avg Loss: 1.29826115, Log Avg loss: 1.05098897, Global Avg Loss: 1.63975275, Time: 0.0401 Steps: 58230, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000757, Sample Num: 12112, Cur Loss: 1.11191094, Cur Avg Loss: 1.29621327, Log Avg loss: 1.14323649, Global Avg Loss: 1.63966750, Time: 0.0402 Steps: 58240, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000767, Sample Num: 12272, Cur Loss: 0.53597146, Cur Avg Loss: 1.29433914, Log Avg loss: 1.15246781, Global Avg Loss: 1.63958386, Time: 0.0402 Steps: 58250, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000777, Sample Num: 12432, Cur Loss: 1.49268007, Cur Avg Loss: 1.29740386, Log Avg loss: 1.53246753, Global Avg Loss: 1.63956547, Time: 0.0402 Steps: 58260, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000787, Sample Num: 12592, Cur Loss: 1.34571743, Cur Avg Loss: 1.30068606, Log Avg loss: 1.55571283, Global Avg Loss: 1.63955108, Time: 0.0402 Steps: 58270, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000797, Sample Num: 12752, Cur Loss: 1.84900904, Cur Avg Loss: 1.30174739, Log Avg loss: 1.38527399, Global Avg Loss: 1.63950745, Time: 0.0402 Steps: 58280, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000807, Sample Num: 12912, Cur Loss: 0.98851001, Cur Avg Loss: 1.30189206, Log Avg loss: 1.31342283, Global Avg Loss: 1.63945151, Time: 0.0402 Steps: 58290, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000817, Sample Num: 13072, Cur Loss: 0.79359722, Cur Avg Loss: 1.30278406, Log Avg loss: 1.37476850, Global Avg Loss: 1.63940611, Time: 0.0402 Steps: 58300, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000827, Sample Num: 13232, Cur Loss: 1.37953389, Cur Avg Loss: 1.30038679, Log Avg loss: 1.10452941, Global Avg Loss: 1.63931438, Time: 0.0405 Steps: 58310, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000837, Sample Num: 13392, Cur Loss: 1.16017842, Cur Avg Loss: 1.29815819, Log Avg loss: 1.11385345, Global Avg Loss: 1.63922428, Time: 0.0402 Steps: 58320, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000847, Sample Num: 13552, Cur Loss: 1.59926367, Cur Avg Loss: 1.30088822, Log Avg loss: 1.52939103, Global Avg Loss: 1.63920545, Time: 0.0402 Steps: 58330, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000857, Sample Num: 13712, Cur Loss: 1.29857826, Cur Avg Loss: 1.29986887, Log Avg loss: 1.21353037, Global Avg Loss: 1.63913248, Time: 0.0402 Steps: 58340, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000867, Sample Num: 13872, Cur Loss: 1.51076126, Cur Avg Loss: 1.30275376, Log Avg loss: 1.54998906, Global Avg Loss: 1.63911721, Time: 0.0401 Steps: 58350, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000877, Sample Num: 14032, Cur Loss: 0.55434787, Cur Avg Loss: 1.29888613, Log Avg loss: 0.96356190, Global Avg Loss: 1.63900145, Time: 0.0403 Steps: 58360, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000887, Sample Num: 14192, Cur Loss: 1.49002814, Cur Avg Loss: 1.29854611, Log Avg loss: 1.26872703, Global Avg Loss: 1.63893801, Time: 0.0403 Steps: 58370, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000897, Sample Num: 14352, Cur Loss: 1.72235894, Cur Avg Loss: 1.30167040, Log Avg loss: 1.57879419, Global Avg Loss: 1.63892771, Time: 0.0402 Steps: 58380, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000907, Sample Num: 14512, Cur Loss: 0.44724506, Cur Avg Loss: 1.29945478, Log Avg loss: 1.10071402, Global Avg Loss: 1.63883554, Time: 0.0402 Steps: 58390, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000917, Sample Num: 14672, Cur Loss: 0.65263003, Cur Avg Loss: 1.30093052, Log Avg loss: 1.43478064, Global Avg Loss: 1.63880060, Time: 0.0402 Steps: 58400, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000927, Sample Num: 14832, Cur Loss: 1.44533396, Cur Avg Loss: 1.29996806, Log Avg loss: 1.21170989, Global Avg Loss: 1.63872748, Time: 0.0402 Steps: 58410, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000937, Sample Num: 14992, Cur Loss: 1.77078223, Cur Avg Loss: 1.29879907, Log Avg loss: 1.19043406, Global Avg Loss: 1.63865074, Time: 0.0536 Steps: 58420, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000947, Sample Num: 15152, Cur Loss: 1.35551119, Cur Avg Loss: 1.29999672, Log Avg loss: 1.41221623, Global Avg Loss: 1.63861199, Time: 0.0406 Steps: 58430, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000957, Sample Num: 15312, Cur Loss: 2.31336522, Cur Avg Loss: 1.30245311, Log Avg loss: 1.53507302, Global Avg Loss: 1.63859427, Time: 0.0452 Steps: 58440, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000967, Sample Num: 15472, Cur Loss: 1.85184073, Cur Avg Loss: 1.30140854, Log Avg loss: 1.20144358, Global Avg Loss: 1.63851948, Time: 0.0431 Steps: 58450, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000977, Sample Num: 15632, Cur Loss: 1.44683743, Cur Avg Loss: 1.30531741, Log Avg loss: 1.68330472, Global Avg Loss: 1.63852714, Time: 0.0521 Steps: 58460, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000987, Sample Num: 15792, Cur Loss: 0.77906549, Cur Avg Loss: 1.30610427, Log Avg loss: 1.38298125, Global Avg Loss: 1.63848343, Time: 0.0481 Steps: 58470, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000997, Sample Num: 15952, Cur Loss: 0.99524105, Cur Avg Loss: 1.30362703, Log Avg loss: 1.05912268, Global Avg Loss: 1.63838436, Time: 0.0431 Steps: 58480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001007, Sample Num: 16112, Cur Loss: 2.14163876, Cur Avg Loss: 1.30618302, Log Avg loss: 1.56101511, Global Avg Loss: 1.63837114, Time: 0.0402 Steps: 58490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001017, Sample Num: 16272, Cur Loss: 1.44737363, Cur Avg Loss: 1.30694676, Log Avg loss: 1.38385565, Global Avg Loss: 1.63832763, Time: 0.0402 Steps: 58500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001027, Sample Num: 16432, Cur Loss: 0.56172341, Cur Avg Loss: 1.30372857, Log Avg loss: 0.97643833, Global Avg Loss: 1.63821451, Time: 0.0402 Steps: 58510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001037, Sample Num: 16592, Cur Loss: 0.76130819, Cur Avg Loss: 1.30103290, Log Avg loss: 1.02418817, Global Avg Loss: 1.63810958, Time: 0.0402 Steps: 58520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001047, Sample Num: 16752, Cur Loss: 0.86443126, Cur Avg Loss: 1.30418581, Log Avg loss: 1.63114221, Global Avg Loss: 1.63810839, Time: 0.0403 Steps: 58530, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001057, Sample Num: 16912, Cur Loss: 1.77626586, Cur Avg Loss: 1.30597099, Log Avg loss: 1.49287955, Global Avg Loss: 1.63808358, Time: 0.0403 Steps: 58540, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001067, Sample Num: 17072, Cur Loss: 1.00429296, Cur Avg Loss: 1.30567322, Log Avg loss: 1.27419904, Global Avg Loss: 1.63802143, Time: 0.0402 Steps: 58550, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001077, Sample Num: 17232, Cur Loss: 2.81409693, Cur Avg Loss: 1.30436201, Log Avg loss: 1.16445597, Global Avg Loss: 1.63794056, Time: 0.0403 Steps: 58560, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001087, Sample Num: 17392, Cur Loss: 0.98026615, Cur Avg Loss: 1.30558765, Log Avg loss: 1.43758907, Global Avg Loss: 1.63790636, Time: 0.0401 Steps: 58570, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001097, Sample Num: 17552, Cur Loss: 0.79850852, Cur Avg Loss: 1.30625946, Log Avg loss: 1.37928474, Global Avg Loss: 1.63786221, Time: 0.0402 Steps: 58580, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001107, Sample Num: 17712, Cur Loss: 2.12752008, Cur Avg Loss: 1.30607101, Log Avg loss: 1.28539855, Global Avg Loss: 1.63780205, Time: 0.0401 Steps: 58590, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001117, Sample Num: 17872, Cur Loss: 0.65137851, Cur Avg Loss: 1.30615400, Log Avg loss: 1.31534053, Global Avg Loss: 1.63774702, Time: 0.0403 Steps: 58600, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001127, Sample Num: 18032, Cur Loss: 1.17023814, Cur Avg Loss: 1.30807718, Log Avg loss: 1.52289659, Global Avg Loss: 1.63772743, Time: 0.0402 Steps: 58610, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001137, Sample Num: 18192, Cur Loss: 1.57427287, Cur Avg Loss: 1.30890588, Log Avg loss: 1.40230035, Global Avg Loss: 1.63768726, Time: 0.0403 Steps: 58620, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001147, Sample Num: 18352, Cur Loss: 1.27641046, Cur Avg Loss: 1.30970259, Log Avg loss: 1.40028838, Global Avg Loss: 1.63764677, Time: 0.0403 Steps: 58630, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001157, Sample Num: 18512, Cur Loss: 2.07877851, Cur Avg Loss: 1.30840255, Log Avg loss: 1.15928780, Global Avg Loss: 1.63756520, Time: 0.0402 Steps: 58640, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001167, Sample Num: 18672, Cur Loss: 0.72787613, Cur Avg Loss: 1.30651407, Log Avg loss: 1.08801752, Global Avg Loss: 1.63747150, Time: 0.0403 Steps: 58650, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001177, Sample Num: 18832, Cur Loss: 2.12543559, Cur Avg Loss: 1.30584648, Log Avg loss: 1.22793874, Global Avg Loss: 1.63740168, Time: 0.0402 Steps: 58660, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001187, Sample Num: 18992, Cur Loss: 1.26645207, Cur Avg Loss: 1.30694434, Log Avg loss: 1.43616208, Global Avg Loss: 1.63736738, Time: 0.0402 Steps: 58670, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001197, Sample Num: 19152, Cur Loss: 1.60879803, Cur Avg Loss: 1.30726768, Log Avg loss: 1.34564792, Global Avg Loss: 1.63731767, Time: 0.0402 Steps: 58680, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001207, Sample Num: 19312, Cur Loss: 0.68570173, Cur Avg Loss: 1.30487542, Log Avg loss: 1.01852190, Global Avg Loss: 1.63721224, Time: 0.0402 Steps: 58690, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001217, Sample Num: 19472, Cur Loss: 1.39094412, Cur Avg Loss: 1.30507736, Log Avg loss: 1.32945147, Global Avg Loss: 1.63715981, Time: 0.0404 Steps: 58700, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001227, Sample Num: 19632, Cur Loss: 1.55348063, Cur Avg Loss: 1.30661092, Log Avg loss: 1.49324618, Global Avg Loss: 1.63713529, Time: 0.0495 Steps: 58710, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001237, Sample Num: 19792, Cur Loss: 0.64314181, Cur Avg Loss: 1.30623056, Log Avg loss: 1.25955965, Global Avg Loss: 1.63707099, Time: 0.0525 Steps: 58720, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001247, Sample Num: 19952, Cur Loss: 0.42073637, Cur Avg Loss: 1.30574868, Log Avg loss: 1.24614072, Global Avg Loss: 1.63700443, Time: 0.0558 Steps: 58730, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001257, Sample Num: 20112, Cur Loss: 1.09479666, Cur Avg Loss: 1.30656947, Log Avg loss: 1.40892155, Global Avg Loss: 1.63696560, Time: 0.0611 Steps: 58740, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001267, Sample Num: 20272, Cur Loss: 1.63945341, Cur Avg Loss: 1.30711216, Log Avg loss: 1.37532792, Global Avg Loss: 1.63692107, Time: 0.0592 Steps: 58750, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001277, Sample Num: 20432, Cur Loss: 2.34250784, Cur Avg Loss: 1.30736800, Log Avg loss: 1.33978335, Global Avg Loss: 1.63687050, Time: 0.0407 Steps: 58760, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001287, Sample Num: 20592, Cur Loss: 2.59268498, Cur Avg Loss: 1.30740183, Log Avg loss: 1.31172226, Global Avg Loss: 1.63681517, Time: 0.0402 Steps: 58770, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001297, Sample Num: 20752, Cur Loss: 0.55072355, Cur Avg Loss: 1.30797019, Log Avg loss: 1.38111743, Global Avg Loss: 1.63677167, Time: 0.0402 Steps: 58780, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001307, Sample Num: 20912, Cur Loss: 1.08120251, Cur Avg Loss: 1.30763664, Log Avg loss: 1.26437548, Global Avg Loss: 1.63670833, Time: 0.0403 Steps: 58790, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001317, Sample Num: 21072, Cur Loss: 2.02549243, Cur Avg Loss: 1.31043425, Log Avg loss: 1.67608203, Global Avg Loss: 1.63671502, Time: 0.0402 Steps: 58800, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001327, Sample Num: 21232, Cur Loss: 1.30815411, Cur Avg Loss: 1.30820332, Log Avg loss: 1.01439011, Global Avg Loss: 1.63660920, Time: 0.0402 Steps: 58810, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001337, Sample Num: 21392, Cur Loss: 1.42605102, Cur Avg Loss: 1.30714925, Log Avg loss: 1.16727410, Global Avg Loss: 1.63652941, Time: 0.0402 Steps: 58820, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001347, Sample Num: 21552, Cur Loss: 1.42769957, Cur Avg Loss: 1.30652274, Log Avg loss: 1.22275854, Global Avg Loss: 1.63645908, Time: 0.0402 Steps: 58830, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001357, Sample Num: 21712, Cur Loss: 0.86224312, Cur Avg Loss: 1.30487471, Log Avg loss: 1.08288401, Global Avg Loss: 1.63636500, Time: 0.0402 Steps: 58840, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001367, Sample Num: 21872, Cur Loss: 0.67908037, Cur Avg Loss: 1.30316351, Log Avg loss: 1.07095412, Global Avg Loss: 1.63626892, Time: 0.0402 Steps: 58850, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001377, Sample Num: 22032, Cur Loss: 1.14103699, Cur Avg Loss: 1.30421573, Log Avg loss: 1.44805413, Global Avg Loss: 1.63623694, Time: 0.0403 Steps: 58860, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001387, Sample Num: 22192, Cur Loss: 1.92445397, Cur Avg Loss: 1.30585416, Log Avg loss: 1.53146599, Global Avg Loss: 1.63621915, Time: 0.0402 Steps: 58870, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001397, Sample Num: 22352, Cur Loss: 1.31421733, Cur Avg Loss: 1.30859535, Log Avg loss: 1.68879850, Global Avg Loss: 1.63622808, Time: 0.0402 Steps: 58880, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001407, Sample Num: 22512, Cur Loss: 1.15202045, Cur Avg Loss: 1.30742076, Log Avg loss: 1.14333050, Global Avg Loss: 1.63614438, Time: 0.0402 Steps: 58890, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001417, Sample Num: 22672, Cur Loss: 1.92213213, Cur Avg Loss: 1.30801914, Log Avg loss: 1.39221179, Global Avg Loss: 1.63610296, Time: 0.0402 Steps: 58900, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001427, Sample Num: 22832, Cur Loss: 2.64146662, Cur Avg Loss: 1.30860411, Log Avg loss: 1.39149401, Global Avg Loss: 1.63606144, Time: 0.0402 Steps: 58910, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001437, Sample Num: 22992, Cur Loss: 1.01155567, Cur Avg Loss: 1.30631963, Log Avg loss: 0.98032454, Global Avg Loss: 1.63595015, Time: 0.0402 Steps: 58920, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001447, Sample Num: 23152, Cur Loss: 2.23603702, Cur Avg Loss: 1.30723066, Log Avg loss: 1.43814498, Global Avg Loss: 1.63591658, Time: 0.0402 Steps: 58930, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001457, Sample Num: 23312, Cur Loss: 1.14360237, Cur Avg Loss: 1.30941341, Log Avg loss: 1.62525797, Global Avg Loss: 1.63591477, Time: 0.0402 Steps: 58940, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001467, Sample Num: 23472, Cur Loss: 2.07223511, Cur Avg Loss: 1.31036139, Log Avg loss: 1.44848194, Global Avg Loss: 1.63588298, Time: 0.0402 Steps: 58950, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001477, Sample Num: 23632, Cur Loss: 1.89929521, Cur Avg Loss: 1.31098254, Log Avg loss: 1.40210553, Global Avg Loss: 1.63584333, Time: 0.0402 Steps: 58960, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001487, Sample Num: 23792, Cur Loss: 0.51299191, Cur Avg Loss: 1.30836402, Log Avg loss: 0.92160850, Global Avg Loss: 1.63572221, Time: 0.0402 Steps: 58970, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001497, Sample Num: 23952, Cur Loss: 2.52584410, Cur Avg Loss: 1.30879944, Log Avg loss: 1.37354552, Global Avg Loss: 1.63567776, Time: 0.0402 Steps: 58980, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001507, Sample Num: 24112, Cur Loss: 1.83028054, Cur Avg Loss: 1.30798189, Log Avg loss: 1.18559484, Global Avg Loss: 1.63560146, Time: 0.0489 Steps: 58990, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001517, Sample Num: 24272, Cur Loss: 2.01023936, Cur Avg Loss: 1.31027866, Log Avg loss: 1.65640224, Global Avg Loss: 1.63560499, Time: 0.0510 Steps: 59000, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001527, Sample Num: 24432, Cur Loss: 1.11338389, Cur Avg Loss: 1.30958563, Log Avg loss: 1.20445335, Global Avg Loss: 1.63553192, Time: 0.0407 Steps: 59010, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001537, Sample Num: 24592, Cur Loss: 0.50278676, Cur Avg Loss: 1.30901353, Log Avg loss: 1.22165346, Global Avg Loss: 1.63546180, Time: 0.0407 Steps: 59020, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001547, Sample Num: 24752, Cur Loss: 1.44650924, Cur Avg Loss: 1.30844347, Log Avg loss: 1.22082542, Global Avg Loss: 1.63539156, Time: 0.0507 Steps: 59030, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001557, Sample Num: 24912, Cur Loss: 1.21578836, Cur Avg Loss: 1.30893393, Log Avg loss: 1.38480878, Global Avg Loss: 1.63534911, Time: 0.0503 Steps: 59040, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001567, Sample Num: 25072, Cur Loss: 1.07211506, Cur Avg Loss: 1.30937354, Log Avg loss: 1.37782066, Global Avg Loss: 1.63530550, Time: 0.0405 Steps: 59050, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001577, Sample Num: 25232, Cur Loss: 1.54453218, Cur Avg Loss: 1.31127978, Log Avg loss: 1.60998724, Global Avg Loss: 1.63530121, Time: 0.0402 Steps: 59060, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001587, Sample Num: 25392, Cur Loss: 0.49949625, Cur Avg Loss: 1.30840207, Log Avg loss: 0.85458665, Global Avg Loss: 1.63516905, Time: 0.0402 Steps: 59070, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001597, Sample Num: 25552, Cur Loss: 1.59698987, Cur Avg Loss: 1.30789291, Log Avg loss: 1.22708880, Global Avg Loss: 1.63509997, Time: 0.0402 Steps: 59080, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001607, Sample Num: 25712, Cur Loss: 1.72952044, Cur Avg Loss: 1.30872957, Log Avg loss: 1.44234544, Global Avg Loss: 1.63506735, Time: 0.0402 Steps: 59090, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001617, Sample Num: 25872, Cur Loss: 3.22804022, Cur Avg Loss: 1.30933082, Log Avg loss: 1.40595106, Global Avg Loss: 1.63502859, Time: 0.0403 Steps: 59100, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001627, Sample Num: 26032, Cur Loss: 2.28781629, Cur Avg Loss: 1.31069388, Log Avg loss: 1.53110095, Global Avg Loss: 1.63501100, Time: 0.0403 Steps: 59110, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001637, Sample Num: 26192, Cur Loss: 1.37575102, Cur Avg Loss: 1.31017398, Log Avg loss: 1.22558646, Global Avg Loss: 1.63494175, Time: 0.0403 Steps: 59120, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001647, Sample Num: 26352, Cur Loss: 0.78522319, Cur Avg Loss: 1.30988341, Log Avg loss: 1.26231638, Global Avg Loss: 1.63487873, Time: 0.0402 Steps: 59130, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001657, Sample Num: 26512, Cur Loss: 1.23217535, Cur Avg Loss: 1.30892180, Log Avg loss: 1.15054454, Global Avg Loss: 1.63479684, Time: 0.0403 Steps: 59140, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001667, Sample Num: 26672, Cur Loss: 1.58587432, Cur Avg Loss: 1.30916172, Log Avg loss: 1.34891770, Global Avg Loss: 1.63474851, Time: 0.0402 Steps: 59150, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001677, Sample Num: 26832, Cur Loss: 1.59162581, Cur Avg Loss: 1.30892302, Log Avg loss: 1.26913163, Global Avg Loss: 1.63468670, Time: 0.0403 Steps: 59160, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001687, Sample Num: 26992, Cur Loss: 0.54167193, Cur Avg Loss: 1.30842835, Log Avg loss: 1.22547218, Global Avg Loss: 1.63461754, Time: 0.0402 Steps: 59170, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001697, Sample Num: 27152, Cur Loss: 0.55312139, Cur Avg Loss: 1.30766703, Log Avg loss: 1.17923125, Global Avg Loss: 1.63454060, Time: 0.0403 Steps: 59180, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001707, Sample Num: 27312, Cur Loss: 1.19747639, Cur Avg Loss: 1.30790304, Log Avg loss: 1.34795443, Global Avg Loss: 1.63449218, Time: 0.0403 Steps: 59190, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001717, Sample Num: 27472, Cur Loss: 1.61268735, Cur Avg Loss: 1.30879994, Log Avg loss: 1.46190029, Global Avg Loss: 1.63446302, Time: 0.0402 Steps: 59200, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001727, Sample Num: 27632, Cur Loss: 1.98621893, Cur Avg Loss: 1.30939896, Log Avg loss: 1.41225109, Global Avg Loss: 1.63442549, Time: 0.0403 Steps: 59210, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001737, Sample Num: 27792, Cur Loss: 0.92068249, Cur Avg Loss: 1.31138993, Log Avg loss: 1.65523061, Global Avg Loss: 1.63442901, Time: 0.0402 Steps: 59220, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001747, Sample Num: 27952, Cur Loss: 1.19917643, Cur Avg Loss: 1.31086063, Log Avg loss: 1.21892082, Global Avg Loss: 1.63435886, Time: 0.0403 Steps: 59230, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001757, Sample Num: 28112, Cur Loss: 0.49830288, Cur Avg Loss: 1.30907803, Log Avg loss: 0.99765845, Global Avg Loss: 1.63425138, Time: 0.0403 Steps: 59240, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001767, Sample Num: 28272, Cur Loss: 0.95159090, Cur Avg Loss: 1.30707079, Log Avg loss: 0.95439818, Global Avg Loss: 1.63413663, Time: 0.0403 Steps: 59250, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001777, Sample Num: 28432, Cur Loss: 1.18563485, Cur Avg Loss: 1.30516451, Log Avg loss: 0.96832518, Global Avg Loss: 1.63402428, Time: 0.0402 Steps: 59260, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001787, Sample Num: 28592, Cur Loss: 1.89025342, Cur Avg Loss: 1.30461136, Log Avg loss: 1.20631615, Global Avg Loss: 1.63395212, Time: 0.0403 Steps: 59270, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001797, Sample Num: 28752, Cur Loss: 1.07217216, Cur Avg Loss: 1.30539936, Log Avg loss: 1.44621580, Global Avg Loss: 1.63392045, Time: 0.0403 Steps: 59280, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001807, Sample Num: 28912, Cur Loss: 0.39193293, Cur Avg Loss: 1.30639301, Log Avg loss: 1.48495133, Global Avg Loss: 1.63389532, Time: 0.0406 Steps: 59290, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001817, Sample Num: 29072, Cur Loss: 2.53329086, Cur Avg Loss: 1.30764470, Log Avg loss: 1.53382488, Global Avg Loss: 1.63387845, Time: 0.0518 Steps: 59300, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001827, Sample Num: 29232, Cur Loss: 1.10641634, Cur Avg Loss: 1.30776246, Log Avg loss: 1.32915904, Global Avg Loss: 1.63382707, Time: 0.0481 Steps: 59310, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001837, Sample Num: 29392, Cur Loss: 0.68601465, Cur Avg Loss: 1.30837278, Log Avg loss: 1.41987832, Global Avg Loss: 1.63379100, Time: 0.0540 Steps: 59320, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001847, Sample Num: 29552, Cur Loss: 1.52842093, Cur Avg Loss: 1.30932376, Log Avg loss: 1.48401858, Global Avg Loss: 1.63376576, Time: 0.0407 Steps: 59330, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001857, Sample Num: 29712, Cur Loss: 1.35874701, Cur Avg Loss: 1.30963917, Log Avg loss: 1.36789680, Global Avg Loss: 1.63372095, Time: 0.0406 Steps: 59340, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001867, Sample Num: 29872, Cur Loss: 1.02696252, Cur Avg Loss: 1.30903654, Log Avg loss: 1.19712802, Global Avg Loss: 1.63364739, Time: 0.0402 Steps: 59350, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001877, Sample Num: 30032, Cur Loss: 2.03514957, Cur Avg Loss: 1.30832103, Log Avg loss: 1.17473535, Global Avg Loss: 1.63357008, Time: 0.0402 Steps: 59360, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001887, Sample Num: 30192, Cur Loss: 0.57268429, Cur Avg Loss: 1.30892714, Log Avg loss: 1.42269394, Global Avg Loss: 1.63353456, Time: 0.0401 Steps: 59370, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001897, Sample Num: 30352, Cur Loss: 2.19231272, Cur Avg Loss: 1.30834704, Log Avg loss: 1.19888247, Global Avg Loss: 1.63346136, Time: 0.0403 Steps: 59380, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001907, Sample Num: 30512, Cur Loss: 1.54677892, Cur Avg Loss: 1.30662040, Log Avg loss: 0.97907575, Global Avg Loss: 1.63335118, Time: 0.0403 Steps: 59390, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001917, Sample Num: 30672, Cur Loss: 1.80723500, Cur Avg Loss: 1.30714079, Log Avg loss: 1.40638020, Global Avg Loss: 1.63331297, Time: 0.0402 Steps: 59400, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001927, Sample Num: 30832, Cur Loss: 0.76046753, Cur Avg Loss: 1.30803383, Log Avg loss: 1.47922845, Global Avg Loss: 1.63328703, Time: 0.0401 Steps: 59410, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001937, Sample Num: 30992, Cur Loss: 1.18486714, Cur Avg Loss: 1.30827423, Log Avg loss: 1.35459909, Global Avg Loss: 1.63324013, Time: 0.0402 Steps: 59420, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001947, Sample Num: 31152, Cur Loss: 2.24945831, Cur Avg Loss: 1.30832896, Log Avg loss: 1.31893078, Global Avg Loss: 1.63318724, Time: 0.0402 Steps: 59430, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001957, Sample Num: 31312, Cur Loss: 0.81497943, Cur Avg Loss: 1.30747607, Log Avg loss: 1.14141913, Global Avg Loss: 1.63310451, Time: 0.0403 Steps: 59440, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001967, Sample Num: 31472, Cur Loss: 0.91640329, Cur Avg Loss: 1.30755472, Log Avg loss: 1.32294652, Global Avg Loss: 1.63305234, Time: 0.0402 Steps: 59450, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001977, Sample Num: 31632, Cur Loss: 1.19699633, Cur Avg Loss: 1.30692725, Log Avg loss: 1.18350260, Global Avg Loss: 1.63297673, Time: 0.0402 Steps: 59460, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001987, Sample Num: 31792, Cur Loss: 1.92387843, Cur Avg Loss: 1.30635872, Log Avg loss: 1.19396182, Global Avg Loss: 1.63290291, Time: 0.0402 Steps: 59470, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001997, Sample Num: 31952, Cur Loss: 1.09965038, Cur Avg Loss: 1.30520965, Log Avg loss: 1.07688813, Global Avg Loss: 1.63280943, Time: 0.0402 Steps: 59480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002007, Sample Num: 32112, Cur Loss: 1.34439087, Cur Avg Loss: 1.30672953, Log Avg loss: 1.61025019, Global Avg Loss: 1.63280564, Time: 0.0402 Steps: 59490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002017, Sample Num: 32272, Cur Loss: 2.43089461, Cur Avg Loss: 1.30716630, Log Avg loss: 1.39482496, Global Avg Loss: 1.63276565, Time: 0.0402 Steps: 59500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002027, Sample Num: 32432, Cur Loss: 1.70858514, Cur Avg Loss: 1.30800555, Log Avg loss: 1.47728320, Global Avg Loss: 1.63273952, Time: 0.0402 Steps: 59510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002037, Sample Num: 32592, Cur Loss: 0.91394490, Cur Avg Loss: 1.30731539, Log Avg loss: 1.16742055, Global Avg Loss: 1.63266134, Time: 0.0402 Steps: 59520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002047, Sample Num: 32752, Cur Loss: 0.73130459, Cur Avg Loss: 1.30687637, Log Avg loss: 1.21744758, Global Avg Loss: 1.63259159, Time: 0.0402 Steps: 59530, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002057, Sample Num: 32912, Cur Loss: 0.55299973, Cur Avg Loss: 1.30454576, Log Avg loss: 0.82746889, Global Avg Loss: 1.63245637, Time: 0.0402 Steps: 59540, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002067, Sample Num: 33072, Cur Loss: 2.93706536, Cur Avg Loss: 1.30462637, Log Avg loss: 1.32120813, Global Avg Loss: 1.63240410, Time: 0.0403 Steps: 59550, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002077, Sample Num: 33232, Cur Loss: 1.06402302, Cur Avg Loss: 1.30493329, Log Avg loss: 1.36837496, Global Avg Loss: 1.63235977, Time: 0.0403 Steps: 59560, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002087, Sample Num: 33392, Cur Loss: 1.65911722, Cur Avg Loss: 1.30493564, Log Avg loss: 1.30542256, Global Avg Loss: 1.63230489, Time: 0.0403 Steps: 59570, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002097, Sample Num: 33552, Cur Loss: 2.52545905, Cur Avg Loss: 1.30470961, Log Avg loss: 1.25753798, Global Avg Loss: 1.63224199, Time: 0.0403 Steps: 59580, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002107, Sample Num: 33712, Cur Loss: 2.52561378, Cur Avg Loss: 1.30517341, Log Avg loss: 1.40243190, Global Avg Loss: 1.63220342, Time: 0.0555 Steps: 59590, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002117, Sample Num: 33872, Cur Loss: 1.19400191, Cur Avg Loss: 1.30442442, Log Avg loss: 1.14661126, Global Avg Loss: 1.63212195, Time: 0.0404 Steps: 59600, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002127, Sample Num: 34032, Cur Loss: 1.43578506, Cur Avg Loss: 1.30430690, Log Avg loss: 1.27942903, Global Avg Loss: 1.63206278, Time: 0.0511 Steps: 59610, Updated lr: 0.000044 ***** Running evaluation checkpoint-59612 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-59612 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 89.989491, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.268162, "eval_total_loss": 891.517626, "eval_mae": 0.952706, "eval_mse": 1.268161, "eval_r2": 0.193874, "eval_sp_statistic": 0.376333, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.492341, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.335348, "test_total_loss": 670.344741, "test_mae": 1.003374, "test_mse": 1.335505, "test_r2": 0.138053, "test_sp_statistic": 0.403166, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.529794, "test_ps_pvalue": 0.0, "lr": 4.441725936462779e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6321206198521432, "train_cur_epoch_loss": 2780.972913056612, "train_cur_epoch_avg_loss": 1.3062343415014617, "train_cur_epoch_time": 89.9894905090332, "train_cur_epoch_avg_time": 0.04226843142744632, "epoch": 28, "step": 59612} ################################################## Training, Epoch: 0029, Batch: 000008, Sample Num: 128, Cur Loss: 1.73417020, Cur Avg Loss: 1.44625394, Log Avg loss: 1.82821649, Global Avg Loss: 1.63209568, Time: 0.0402 Steps: 59620, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000018, Sample Num: 288, Cur Loss: 0.60075718, Cur Avg Loss: 1.51932835, Log Avg loss: 1.57778787, Global Avg Loss: 1.63208657, Time: 0.0402 Steps: 59630, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000028, Sample Num: 448, Cur Loss: 1.31015468, Cur Avg Loss: 1.29143400, Log Avg loss: 0.88122418, Global Avg Loss: 1.63196067, Time: 0.0402 Steps: 59640, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000038, Sample Num: 608, Cur Loss: 0.71213305, Cur Avg Loss: 1.31722619, Log Avg loss: 1.38944431, Global Avg Loss: 1.63192002, Time: 0.0402 Steps: 59650, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000048, Sample Num: 768, Cur Loss: 0.59594738, Cur Avg Loss: 1.30693904, Log Avg loss: 1.26784786, Global Avg Loss: 1.63185899, Time: 0.0402 Steps: 59660, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000058, Sample Num: 928, Cur Loss: 0.74755645, Cur Avg Loss: 1.27865373, Log Avg loss: 1.14288428, Global Avg Loss: 1.63177705, Time: 0.0402 Steps: 59670, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000068, Sample Num: 1088, Cur Loss: 1.37733972, Cur Avg Loss: 1.23071952, Log Avg loss: 0.95270106, Global Avg Loss: 1.63166326, Time: 0.0402 Steps: 59680, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000078, Sample Num: 1248, Cur Loss: 2.36354256, Cur Avg Loss: 1.25122983, Log Avg loss: 1.39069997, Global Avg Loss: 1.63162289, Time: 0.0401 Steps: 59690, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000088, Sample Num: 1408, Cur Loss: 0.61980379, Cur Avg Loss: 1.23923597, Log Avg loss: 1.14568385, Global Avg Loss: 1.63154149, Time: 0.0402 Steps: 59700, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000098, Sample Num: 1568, Cur Loss: 0.82549793, Cur Avg Loss: 1.25329473, Log Avg loss: 1.37701185, Global Avg Loss: 1.63149887, Time: 0.0402 Steps: 59710, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000108, Sample Num: 1728, Cur Loss: 0.71700567, Cur Avg Loss: 1.24528293, Log Avg loss: 1.16676729, Global Avg Loss: 1.63142105, Time: 0.0402 Steps: 59720, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000118, Sample Num: 1888, Cur Loss: 0.45350194, Cur Avg Loss: 1.23503411, Log Avg loss: 1.12434678, Global Avg Loss: 1.63133615, Time: 0.0402 Steps: 59730, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000128, Sample Num: 2048, Cur Loss: 1.53304672, Cur Avg Loss: 1.24866809, Log Avg loss: 1.40954909, Global Avg Loss: 1.63129903, Time: 0.0402 Steps: 59740, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000138, Sample Num: 2208, Cur Loss: 0.40933406, Cur Avg Loss: 1.24166850, Log Avg loss: 1.15207372, Global Avg Loss: 1.63121882, Time: 0.0403 Steps: 59750, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000148, Sample Num: 2368, Cur Loss: 2.38682699, Cur Avg Loss: 1.26085527, Log Avg loss: 1.52563277, Global Avg Loss: 1.63120115, Time: 0.0402 Steps: 59760, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000158, Sample Num: 2528, Cur Loss: 0.50919986, Cur Avg Loss: 1.26404482, Log Avg loss: 1.31125019, Global Avg Loss: 1.63114762, Time: 0.0401 Steps: 59770, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000168, Sample Num: 2688, Cur Loss: 0.86918348, Cur Avg Loss: 1.27573113, Log Avg loss: 1.46037474, Global Avg Loss: 1.63111906, Time: 0.0402 Steps: 59780, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000178, Sample Num: 2848, Cur Loss: 1.13821661, Cur Avg Loss: 1.26693499, Log Avg loss: 1.11915986, Global Avg Loss: 1.63103343, Time: 0.0402 Steps: 59790, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000188, Sample Num: 3008, Cur Loss: 1.95964658, Cur Avg Loss: 1.27128159, Log Avg loss: 1.34865100, Global Avg Loss: 1.63098621, Time: 0.0402 Steps: 59800, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000198, Sample Num: 3168, Cur Loss: 0.43180194, Cur Avg Loss: 1.27682498, Log Avg loss: 1.38104081, Global Avg Loss: 1.63094442, Time: 0.0403 Steps: 59810, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000208, Sample Num: 3328, Cur Loss: 2.52527094, Cur Avg Loss: 1.29117598, Log Avg loss: 1.57532566, Global Avg Loss: 1.63093512, Time: 0.0402 Steps: 59820, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000218, Sample Num: 3488, Cur Loss: 1.38667440, Cur Avg Loss: 1.30756865, Log Avg loss: 1.64853632, Global Avg Loss: 1.63093806, Time: 0.0507 Steps: 59830, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000228, Sample Num: 3648, Cur Loss: 1.48654103, Cur Avg Loss: 1.30094953, Log Avg loss: 1.15665266, Global Avg Loss: 1.63085880, Time: 0.0532 Steps: 59840, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000238, Sample Num: 3808, Cur Loss: 0.98121691, Cur Avg Loss: 1.30531661, Log Avg loss: 1.40488594, Global Avg Loss: 1.63082105, Time: 0.0405 Steps: 59850, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000248, Sample Num: 3968, Cur Loss: 0.58651936, Cur Avg Loss: 1.30766205, Log Avg loss: 1.36348350, Global Avg Loss: 1.63077639, Time: 0.0531 Steps: 59860, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000258, Sample Num: 4128, Cur Loss: 1.02595150, Cur Avg Loss: 1.29453734, Log Avg loss: 0.96904468, Global Avg Loss: 1.63066586, Time: 0.0504 Steps: 59870, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000268, Sample Num: 4288, Cur Loss: 0.73360455, Cur Avg Loss: 1.29345059, Log Avg loss: 1.26541234, Global Avg Loss: 1.63060486, Time: 0.0404 Steps: 59880, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000278, Sample Num: 4448, Cur Loss: 2.52582741, Cur Avg Loss: 1.29699969, Log Avg loss: 1.39211569, Global Avg Loss: 1.63056504, Time: 0.0405 Steps: 59890, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000288, Sample Num: 4608, Cur Loss: 0.58498275, Cur Avg Loss: 1.29575704, Log Avg loss: 1.26121123, Global Avg Loss: 1.63050338, Time: 0.0402 Steps: 59900, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000298, Sample Num: 4768, Cur Loss: 0.97171628, Cur Avg Loss: 1.28583153, Log Avg loss: 0.99997706, Global Avg Loss: 1.63039813, Time: 0.0402 Steps: 59910, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000308, Sample Num: 4928, Cur Loss: 0.41173178, Cur Avg Loss: 1.28813104, Log Avg loss: 1.35665618, Global Avg Loss: 1.63035245, Time: 0.0401 Steps: 59920, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000318, Sample Num: 5088, Cur Loss: 1.11223602, Cur Avg Loss: 1.28421549, Log Avg loss: 1.16361658, Global Avg Loss: 1.63027457, Time: 0.0402 Steps: 59930, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000328, Sample Num: 5248, Cur Loss: 0.99154675, Cur Avg Loss: 1.28085189, Log Avg loss: 1.17388951, Global Avg Loss: 1.63019843, Time: 0.0402 Steps: 59940, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000338, Sample Num: 5408, Cur Loss: 1.39726496, Cur Avg Loss: 1.28318555, Log Avg loss: 1.35972946, Global Avg Loss: 1.63015331, Time: 0.0402 Steps: 59950, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000348, Sample Num: 5568, Cur Loss: 0.88758665, Cur Avg Loss: 1.27995773, Log Avg loss: 1.17085741, Global Avg Loss: 1.63007671, Time: 0.0402 Steps: 59960, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000358, Sample Num: 5728, Cur Loss: 1.66127825, Cur Avg Loss: 1.28063255, Log Avg loss: 1.30411646, Global Avg Loss: 1.63002236, Time: 0.0401 Steps: 59970, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000368, Sample Num: 5888, Cur Loss: 1.91396856, Cur Avg Loss: 1.27611203, Log Avg loss: 1.11427733, Global Avg Loss: 1.62993637, Time: 0.0402 Steps: 59980, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000378, Sample Num: 6048, Cur Loss: 1.40399957, Cur Avg Loss: 1.27717192, Log Avg loss: 1.31617593, Global Avg Loss: 1.62988407, Time: 0.0402 Steps: 59990, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000388, Sample Num: 6208, Cur Loss: 2.32119465, Cur Avg Loss: 1.27966931, Log Avg loss: 1.37407065, Global Avg Loss: 1.62984143, Time: 0.0402 Steps: 60000, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000398, Sample Num: 6368, Cur Loss: 1.08193636, Cur Avg Loss: 1.27898207, Log Avg loss: 1.25231705, Global Avg Loss: 1.62977852, Time: 0.0402 Steps: 60010, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000408, Sample Num: 6528, Cur Loss: 2.80521154, Cur Avg Loss: 1.28127914, Log Avg loss: 1.37270262, Global Avg Loss: 1.62973569, Time: 0.0402 Steps: 60020, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000418, Sample Num: 6688, Cur Loss: 1.33073068, Cur Avg Loss: 1.27874847, Log Avg loss: 1.17549700, Global Avg Loss: 1.62966002, Time: 0.0402 Steps: 60030, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000428, Sample Num: 6848, Cur Loss: 1.76030207, Cur Avg Loss: 1.27823074, Log Avg loss: 1.25658971, Global Avg Loss: 1.62959789, Time: 0.0402 Steps: 60040, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000438, Sample Num: 7008, Cur Loss: 1.46760190, Cur Avg Loss: 1.28945769, Log Avg loss: 1.76997103, Global Avg Loss: 1.62962126, Time: 0.0402 Steps: 60050, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000448, Sample Num: 7168, Cur Loss: 0.78418076, Cur Avg Loss: 1.28624508, Log Avg loss: 1.14553303, Global Avg Loss: 1.62954066, Time: 0.0402 Steps: 60060, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000458, Sample Num: 7328, Cur Loss: 1.75537670, Cur Avg Loss: 1.28243712, Log Avg loss: 1.11184044, Global Avg Loss: 1.62945448, Time: 0.0402 Steps: 60070, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000468, Sample Num: 7488, Cur Loss: 3.36902380, Cur Avg Loss: 1.29534563, Log Avg loss: 1.88655546, Global Avg Loss: 1.62949727, Time: 0.0402 Steps: 60080, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000478, Sample Num: 7648, Cur Loss: 2.57080030, Cur Avg Loss: 1.29687571, Log Avg loss: 1.36848318, Global Avg Loss: 1.62945384, Time: 0.0402 Steps: 60090, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000488, Sample Num: 7808, Cur Loss: 1.76054597, Cur Avg Loss: 1.29942587, Log Avg loss: 1.42132355, Global Avg Loss: 1.62941920, Time: 0.0402 Steps: 60100, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000498, Sample Num: 7968, Cur Loss: 1.10047925, Cur Avg Loss: 1.29914750, Log Avg loss: 1.28556312, Global Avg Loss: 1.62936200, Time: 0.0403 Steps: 60110, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000508, Sample Num: 8128, Cur Loss: 1.28146648, Cur Avg Loss: 1.30373154, Log Avg loss: 1.53201703, Global Avg Loss: 1.62934581, Time: 0.0540 Steps: 60120, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000518, Sample Num: 8288, Cur Loss: 2.01583290, Cur Avg Loss: 1.30739536, Log Avg loss: 1.49351721, Global Avg Loss: 1.62932322, Time: 0.0505 Steps: 60130, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000528, Sample Num: 8448, Cur Loss: 0.86294776, Cur Avg Loss: 1.30553364, Log Avg loss: 1.20909669, Global Avg Loss: 1.62925334, Time: 0.0484 Steps: 60140, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000538, Sample Num: 8608, Cur Loss: 2.09102154, Cur Avg Loss: 1.29907342, Log Avg loss: 0.95797386, Global Avg Loss: 1.62914174, Time: 0.0456 Steps: 60150, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000548, Sample Num: 8768, Cur Loss: 1.22973800, Cur Avg Loss: 1.30381046, Log Avg loss: 1.55866301, Global Avg Loss: 1.62913003, Time: 0.0491 Steps: 60160, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000558, Sample Num: 8928, Cur Loss: 1.14543641, Cur Avg Loss: 1.30223162, Log Avg loss: 1.21571122, Global Avg Loss: 1.62906132, Time: 0.0517 Steps: 60170, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000568, Sample Num: 9088, Cur Loss: 0.54253876, Cur Avg Loss: 1.29818976, Log Avg loss: 1.07265398, Global Avg Loss: 1.62896886, Time: 0.0482 Steps: 60180, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000578, Sample Num: 9248, Cur Loss: 0.30953485, Cur Avg Loss: 1.29890392, Log Avg loss: 1.33946794, Global Avg Loss: 1.62892077, Time: 0.0402 Steps: 60190, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000588, Sample Num: 9408, Cur Loss: 2.78700614, Cur Avg Loss: 1.29801310, Log Avg loss: 1.24652368, Global Avg Loss: 1.62885724, Time: 0.0402 Steps: 60200, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000598, Sample Num: 9568, Cur Loss: 1.15283179, Cur Avg Loss: 1.29770109, Log Avg loss: 1.27935515, Global Avg Loss: 1.62879920, Time: 0.0401 Steps: 60210, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000608, Sample Num: 9728, Cur Loss: 0.35430223, Cur Avg Loss: 1.29096402, Log Avg loss: 0.88808718, Global Avg Loss: 1.62867620, Time: 0.0403 Steps: 60220, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000618, Sample Num: 9888, Cur Loss: 0.77986038, Cur Avg Loss: 1.29182632, Log Avg loss: 1.34425441, Global Avg Loss: 1.62862897, Time: 0.0402 Steps: 60230, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000628, Sample Num: 10048, Cur Loss: 1.18793833, Cur Avg Loss: 1.29189605, Log Avg loss: 1.29620488, Global Avg Loss: 1.62857379, Time: 0.0402 Steps: 60240, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000638, Sample Num: 10208, Cur Loss: 1.44021201, Cur Avg Loss: 1.29152182, Log Avg loss: 1.26802075, Global Avg Loss: 1.62851395, Time: 0.0402 Steps: 60250, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000648, Sample Num: 10368, Cur Loss: 1.02761006, Cur Avg Loss: 1.28980885, Log Avg loss: 1.18052086, Global Avg Loss: 1.62843960, Time: 0.0402 Steps: 60260, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000658, Sample Num: 10528, Cur Loss: 1.36199844, Cur Avg Loss: 1.28711338, Log Avg loss: 1.11244688, Global Avg Loss: 1.62835399, Time: 0.0402 Steps: 60270, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000668, Sample Num: 10688, Cur Loss: 0.81357741, Cur Avg Loss: 1.28760107, Log Avg loss: 1.31969105, Global Avg Loss: 1.62830279, Time: 0.0402 Steps: 60280, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000678, Sample Num: 10848, Cur Loss: 0.88648558, Cur Avg Loss: 1.28718042, Log Avg loss: 1.25908132, Global Avg Loss: 1.62824154, Time: 0.0402 Steps: 60290, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000688, Sample Num: 11008, Cur Loss: 2.43926334, Cur Avg Loss: 1.28955910, Log Avg loss: 1.45083327, Global Avg Loss: 1.62821212, Time: 0.0402 Steps: 60300, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000698, Sample Num: 11168, Cur Loss: 0.81744587, Cur Avg Loss: 1.28819166, Log Avg loss: 1.19411232, Global Avg Loss: 1.62814015, Time: 0.0402 Steps: 60310, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000708, Sample Num: 11328, Cur Loss: 0.53310978, Cur Avg Loss: 1.28536626, Log Avg loss: 1.08815331, Global Avg Loss: 1.62805063, Time: 0.0402 Steps: 60320, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000718, Sample Num: 11488, Cur Loss: 0.55748260, Cur Avg Loss: 1.28180444, Log Avg loss: 1.02962715, Global Avg Loss: 1.62795143, Time: 0.0402 Steps: 60330, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000728, Sample Num: 11648, Cur Loss: 1.06477654, Cur Avg Loss: 1.27947182, Log Avg loss: 1.11199018, Global Avg Loss: 1.62786592, Time: 0.0402 Steps: 60340, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000738, Sample Num: 11808, Cur Loss: 0.89509106, Cur Avg Loss: 1.27946598, Log Avg loss: 1.27904035, Global Avg Loss: 1.62780812, Time: 0.0402 Steps: 60350, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000748, Sample Num: 11968, Cur Loss: 0.81012350, Cur Avg Loss: 1.27413960, Log Avg loss: 0.88105329, Global Avg Loss: 1.62768441, Time: 0.0402 Steps: 60360, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000758, Sample Num: 12128, Cur Loss: 0.73987931, Cur Avg Loss: 1.27345837, Log Avg loss: 1.22250206, Global Avg Loss: 1.62761729, Time: 0.0402 Steps: 60370, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000768, Sample Num: 12288, Cur Loss: 1.49381781, Cur Avg Loss: 1.27111933, Log Avg loss: 1.09381976, Global Avg Loss: 1.62752888, Time: 0.0403 Steps: 60380, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000778, Sample Num: 12448, Cur Loss: 0.85852540, Cur Avg Loss: 1.27005239, Log Avg loss: 1.18811155, Global Avg Loss: 1.62745612, Time: 0.0402 Steps: 60390, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000788, Sample Num: 12608, Cur Loss: 1.23375583, Cur Avg Loss: 1.27053188, Log Avg loss: 1.30783641, Global Avg Loss: 1.62740320, Time: 0.0402 Steps: 60400, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000798, Sample Num: 12768, Cur Loss: 1.62000108, Cur Avg Loss: 1.27057910, Log Avg loss: 1.27430030, Global Avg Loss: 1.62734475, Time: 0.0402 Steps: 60410, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000808, Sample Num: 12928, Cur Loss: 1.59284890, Cur Avg Loss: 1.27197076, Log Avg loss: 1.38302492, Global Avg Loss: 1.62730432, Time: 0.0406 Steps: 60420, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000818, Sample Num: 13088, Cur Loss: 0.68437356, Cur Avg Loss: 1.27624487, Log Avg loss: 1.62159304, Global Avg Loss: 1.62730337, Time: 0.0539 Steps: 60430, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000828, Sample Num: 13248, Cur Loss: 1.35827708, Cur Avg Loss: 1.27731999, Log Avg loss: 1.36526461, Global Avg Loss: 1.62726002, Time: 0.0534 Steps: 60440, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000838, Sample Num: 13408, Cur Loss: 0.48896921, Cur Avg Loss: 1.27737684, Log Avg loss: 1.28208451, Global Avg Loss: 1.62720291, Time: 0.0553 Steps: 60450, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000848, Sample Num: 13568, Cur Loss: 1.36337113, Cur Avg Loss: 1.28125679, Log Avg loss: 1.60639617, Global Avg Loss: 1.62719947, Time: 0.0471 Steps: 60460, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000858, Sample Num: 13728, Cur Loss: 1.14947438, Cur Avg Loss: 1.27946007, Log Avg loss: 1.12709805, Global Avg Loss: 1.62711677, Time: 0.0431 Steps: 60470, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000868, Sample Num: 13888, Cur Loss: 0.64364123, Cur Avg Loss: 1.27916560, Log Avg loss: 1.25390046, Global Avg Loss: 1.62705506, Time: 0.0406 Steps: 60480, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000878, Sample Num: 14048, Cur Loss: 1.15707922, Cur Avg Loss: 1.27737654, Log Avg loss: 1.12208617, Global Avg Loss: 1.62697158, Time: 0.0402 Steps: 60490, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000888, Sample Num: 14208, Cur Loss: 2.25087309, Cur Avg Loss: 1.27953024, Log Avg loss: 1.46862462, Global Avg Loss: 1.62694541, Time: 0.0402 Steps: 60500, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000898, Sample Num: 14368, Cur Loss: 1.25345325, Cur Avg Loss: 1.27937915, Log Avg loss: 1.26596261, Global Avg Loss: 1.62688575, Time: 0.0401 Steps: 60510, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000908, Sample Num: 14528, Cur Loss: 0.58822215, Cur Avg Loss: 1.27879968, Log Avg loss: 1.22676326, Global Avg Loss: 1.62681964, Time: 0.0403 Steps: 60520, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000918, Sample Num: 14688, Cur Loss: 2.22009826, Cur Avg Loss: 1.28009116, Log Avg loss: 1.39735768, Global Avg Loss: 1.62678173, Time: 0.0402 Steps: 60530, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000928, Sample Num: 14848, Cur Loss: 2.24134302, Cur Avg Loss: 1.28286043, Log Avg loss: 1.53707893, Global Avg Loss: 1.62676691, Time: 0.0402 Steps: 60540, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000938, Sample Num: 15008, Cur Loss: 2.23202276, Cur Avg Loss: 1.28447514, Log Avg loss: 1.43432050, Global Avg Loss: 1.62673513, Time: 0.0402 Steps: 60550, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000948, Sample Num: 15168, Cur Loss: 1.42323768, Cur Avg Loss: 1.28431789, Log Avg loss: 1.26956815, Global Avg Loss: 1.62667615, Time: 0.0402 Steps: 60560, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000958, Sample Num: 15328, Cur Loss: 1.47791386, Cur Avg Loss: 1.28773069, Log Avg loss: 1.61126392, Global Avg Loss: 1.62667361, Time: 0.0402 Steps: 60570, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000968, Sample Num: 15488, Cur Loss: 1.66757953, Cur Avg Loss: 1.28980808, Log Avg loss: 1.48882189, Global Avg Loss: 1.62665085, Time: 0.0403 Steps: 60580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000978, Sample Num: 15648, Cur Loss: 1.15835917, Cur Avg Loss: 1.28814068, Log Avg loss: 1.12673681, Global Avg Loss: 1.62656834, Time: 0.0402 Steps: 60590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000988, Sample Num: 15808, Cur Loss: 0.73330396, Cur Avg Loss: 1.29136368, Log Avg loss: 1.60657299, Global Avg Loss: 1.62656504, Time: 0.0402 Steps: 60600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000998, Sample Num: 15968, Cur Loss: 0.67205262, Cur Avg Loss: 1.29089705, Log Avg loss: 1.24479333, Global Avg Loss: 1.62650206, Time: 0.0402 Steps: 60610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001008, Sample Num: 16128, Cur Loss: 1.55311906, Cur Avg Loss: 1.29007527, Log Avg loss: 1.20806233, Global Avg Loss: 1.62643303, Time: 0.0402 Steps: 60620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001018, Sample Num: 16288, Cur Loss: 0.76757658, Cur Avg Loss: 1.28948496, Log Avg loss: 1.22998150, Global Avg Loss: 1.62636764, Time: 0.0401 Steps: 60630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001028, Sample Num: 16448, Cur Loss: 1.83353043, Cur Avg Loss: 1.28666372, Log Avg loss: 0.99946154, Global Avg Loss: 1.62626426, Time: 0.0401 Steps: 60640, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001038, Sample Num: 16608, Cur Loss: 1.15723193, Cur Avg Loss: 1.28784031, Log Avg loss: 1.40879382, Global Avg Loss: 1.62622840, Time: 0.0402 Steps: 60650, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001048, Sample Num: 16768, Cur Loss: 0.65286303, Cur Avg Loss: 1.28612378, Log Avg loss: 1.10794814, Global Avg Loss: 1.62614296, Time: 0.0402 Steps: 60660, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001058, Sample Num: 16928, Cur Loss: 2.08611536, Cur Avg Loss: 1.28612831, Log Avg loss: 1.28660326, Global Avg Loss: 1.62608700, Time: 0.0403 Steps: 60670, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001068, Sample Num: 17088, Cur Loss: 1.45976126, Cur Avg Loss: 1.28583966, Log Avg loss: 1.25530001, Global Avg Loss: 1.62602589, Time: 0.0402 Steps: 60680, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001078, Sample Num: 17248, Cur Loss: 1.30011249, Cur Avg Loss: 1.28375493, Log Avg loss: 1.06110600, Global Avg Loss: 1.62593281, Time: 0.0403 Steps: 60690, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001088, Sample Num: 17408, Cur Loss: 1.28936696, Cur Avg Loss: 1.28468462, Log Avg loss: 1.38490494, Global Avg Loss: 1.62589310, Time: 0.0402 Steps: 60700, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001098, Sample Num: 17568, Cur Loss: 1.80964231, Cur Avg Loss: 1.28534359, Log Avg loss: 1.35703994, Global Avg Loss: 1.62584882, Time: 0.0406 Steps: 60710, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001108, Sample Num: 17728, Cur Loss: 0.30203137, Cur Avg Loss: 1.28274935, Log Avg loss: 0.99790174, Global Avg Loss: 1.62574540, Time: 0.0538 Steps: 60720, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001118, Sample Num: 17888, Cur Loss: 0.37666273, Cur Avg Loss: 1.28136881, Log Avg loss: 1.12840465, Global Avg Loss: 1.62566351, Time: 0.0511 Steps: 60730, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001128, Sample Num: 18048, Cur Loss: 0.43126673, Cur Avg Loss: 1.28034009, Log Avg loss: 1.16532968, Global Avg Loss: 1.62558772, Time: 0.0506 Steps: 60740, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001138, Sample Num: 18208, Cur Loss: 0.84957296, Cur Avg Loss: 1.28080169, Log Avg loss: 1.33286951, Global Avg Loss: 1.62553953, Time: 0.0409 Steps: 60750, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001148, Sample Num: 18368, Cur Loss: 1.40969241, Cur Avg Loss: 1.28131196, Log Avg loss: 1.33938075, Global Avg Loss: 1.62549244, Time: 0.0512 Steps: 60760, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001158, Sample Num: 18528, Cur Loss: 1.96446323, Cur Avg Loss: 1.27829337, Log Avg loss: 0.93175925, Global Avg Loss: 1.62537828, Time: 0.0554 Steps: 60770, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001168, Sample Num: 18688, Cur Loss: 0.69763660, Cur Avg Loss: 1.27959497, Log Avg loss: 1.43032061, Global Avg Loss: 1.62534619, Time: 0.0403 Steps: 60780, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001178, Sample Num: 18848, Cur Loss: 1.49794805, Cur Avg Loss: 1.27815246, Log Avg loss: 1.10966666, Global Avg Loss: 1.62526136, Time: 0.0402 Steps: 60790, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001188, Sample Num: 19008, Cur Loss: 1.18688416, Cur Avg Loss: 1.27720216, Log Avg loss: 1.16525676, Global Avg Loss: 1.62518570, Time: 0.0402 Steps: 60800, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001198, Sample Num: 19168, Cur Loss: 0.85469091, Cur Avg Loss: 1.27447125, Log Avg loss: 0.95003896, Global Avg Loss: 1.62507467, Time: 0.0403 Steps: 60810, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001208, Sample Num: 19328, Cur Loss: 2.39604950, Cur Avg Loss: 1.27654983, Log Avg loss: 1.52556377, Global Avg Loss: 1.62505831, Time: 0.0402 Steps: 60820, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001218, Sample Num: 19488, Cur Loss: 1.58907580, Cur Avg Loss: 1.27842434, Log Avg loss: 1.50486627, Global Avg Loss: 1.62503855, Time: 0.0403 Steps: 60830, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001228, Sample Num: 19648, Cur Loss: 1.39290273, Cur Avg Loss: 1.27911619, Log Avg loss: 1.36338303, Global Avg Loss: 1.62499555, Time: 0.0402 Steps: 60840, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001238, Sample Num: 19808, Cur Loss: 1.35004365, Cur Avg Loss: 1.28013076, Log Avg loss: 1.40471959, Global Avg Loss: 1.62495935, Time: 0.0404 Steps: 60850, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001248, Sample Num: 19968, Cur Loss: 1.61854124, Cur Avg Loss: 1.28253466, Log Avg loss: 1.58013747, Global Avg Loss: 1.62495198, Time: 0.0402 Steps: 60860, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001258, Sample Num: 20128, Cur Loss: 1.78877234, Cur Avg Loss: 1.28420240, Log Avg loss: 1.49233637, Global Avg Loss: 1.62493020, Time: 0.0402 Steps: 60870, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001268, Sample Num: 20288, Cur Loss: 0.79577690, Cur Avg Loss: 1.28383777, Log Avg loss: 1.23796760, Global Avg Loss: 1.62486663, Time: 0.0402 Steps: 60880, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001278, Sample Num: 20448, Cur Loss: 1.83447170, Cur Avg Loss: 1.28267182, Log Avg loss: 1.13482979, Global Avg Loss: 1.62478616, Time: 0.0402 Steps: 60890, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001288, Sample Num: 20608, Cur Loss: 1.40778720, Cur Avg Loss: 1.28419350, Log Avg loss: 1.47866312, Global Avg Loss: 1.62476216, Time: 0.0402 Steps: 60900, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001298, Sample Num: 20768, Cur Loss: 0.92909223, Cur Avg Loss: 1.28416815, Log Avg loss: 1.28090335, Global Avg Loss: 1.62470571, Time: 0.0402 Steps: 60910, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001308, Sample Num: 20928, Cur Loss: 0.98888648, Cur Avg Loss: 1.28530842, Log Avg loss: 1.43331606, Global Avg Loss: 1.62467429, Time: 0.0403 Steps: 60920, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001318, Sample Num: 21088, Cur Loss: 0.68751454, Cur Avg Loss: 1.28507715, Log Avg loss: 1.25482721, Global Avg Loss: 1.62461359, Time: 0.0401 Steps: 60930, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001328, Sample Num: 21248, Cur Loss: 0.28274661, Cur Avg Loss: 1.28262305, Log Avg loss: 0.95917269, Global Avg Loss: 1.62450439, Time: 0.0487 Steps: 60940, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001338, Sample Num: 21408, Cur Loss: 0.89059353, Cur Avg Loss: 1.28200789, Log Avg loss: 1.20031465, Global Avg Loss: 1.62443480, Time: 0.0507 Steps: 60950, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001348, Sample Num: 21568, Cur Loss: 0.93707281, Cur Avg Loss: 1.28417350, Log Avg loss: 1.57393102, Global Avg Loss: 1.62442651, Time: 0.0518 Steps: 60960, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001358, Sample Num: 21728, Cur Loss: 1.62966180, Cur Avg Loss: 1.28250706, Log Avg loss: 1.05787188, Global Avg Loss: 1.62433359, Time: 0.0500 Steps: 60970, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001368, Sample Num: 21888, Cur Loss: 0.46079063, Cur Avg Loss: 1.28087046, Log Avg loss: 1.05861940, Global Avg Loss: 1.62424082, Time: 0.0556 Steps: 60980, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001378, Sample Num: 22048, Cur Loss: 1.44710612, Cur Avg Loss: 1.28126963, Log Avg loss: 1.33587648, Global Avg Loss: 1.62419354, Time: 0.0407 Steps: 60990, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001388, Sample Num: 22208, Cur Loss: 1.58363390, Cur Avg Loss: 1.28121492, Log Avg loss: 1.27367593, Global Avg Loss: 1.62413608, Time: 0.0408 Steps: 61000, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001398, Sample Num: 22368, Cur Loss: 1.34615946, Cur Avg Loss: 1.28193499, Log Avg loss: 1.38188020, Global Avg Loss: 1.62409637, Time: 0.0560 Steps: 61010, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001408, Sample Num: 22528, Cur Loss: 0.83294821, Cur Avg Loss: 1.28132595, Log Avg loss: 1.19618322, Global Avg Loss: 1.62402624, Time: 0.0415 Steps: 61020, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001418, Sample Num: 22688, Cur Loss: 1.00552535, Cur Avg Loss: 1.28297697, Log Avg loss: 1.51544033, Global Avg Loss: 1.62400845, Time: 0.0853 Steps: 61030, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001428, Sample Num: 22848, Cur Loss: 1.04905534, Cur Avg Loss: 1.28388301, Log Avg loss: 1.41235939, Global Avg Loss: 1.62397378, Time: 0.0728 Steps: 61040, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001438, Sample Num: 23008, Cur Loss: 1.69230402, Cur Avg Loss: 1.28458964, Log Avg loss: 1.38549592, Global Avg Loss: 1.62393471, Time: 0.0877 Steps: 61050, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001448, Sample Num: 23168, Cur Loss: 0.44062641, Cur Avg Loss: 1.28477472, Log Avg loss: 1.31138968, Global Avg Loss: 1.62388353, Time: 0.0635 Steps: 61060, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001458, Sample Num: 23328, Cur Loss: 1.82371497, Cur Avg Loss: 1.28621424, Log Avg loss: 1.49465620, Global Avg Loss: 1.62386237, Time: 0.0883 Steps: 61070, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001468, Sample Num: 23488, Cur Loss: 1.96651566, Cur Avg Loss: 1.28878985, Log Avg loss: 1.66431417, Global Avg Loss: 1.62386899, Time: 0.0402 Steps: 61080, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001478, Sample Num: 23648, Cur Loss: 0.93793792, Cur Avg Loss: 1.28711492, Log Avg loss: 1.04123526, Global Avg Loss: 1.62377362, Time: 0.0402 Steps: 61090, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001488, Sample Num: 23808, Cur Loss: 1.48678637, Cur Avg Loss: 1.28627749, Log Avg loss: 1.16250492, Global Avg Loss: 1.62369812, Time: 0.0402 Steps: 61100, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001498, Sample Num: 23968, Cur Loss: 0.47291893, Cur Avg Loss: 1.28479446, Log Avg loss: 1.06411985, Global Avg Loss: 1.62360655, Time: 0.0403 Steps: 61110, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001508, Sample Num: 24128, Cur Loss: 0.63379073, Cur Avg Loss: 1.28256903, Log Avg loss: 0.94919980, Global Avg Loss: 1.62349621, Time: 0.0402 Steps: 61120, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001518, Sample Num: 24288, Cur Loss: 2.88879204, Cur Avg Loss: 1.28340832, Log Avg loss: 1.40997369, Global Avg Loss: 1.62346128, Time: 0.0402 Steps: 61130, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001528, Sample Num: 24448, Cur Loss: 0.85406482, Cur Avg Loss: 1.28300979, Log Avg loss: 1.22251188, Global Avg Loss: 1.62339570, Time: 0.0402 Steps: 61140, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001538, Sample Num: 24608, Cur Loss: 1.52853703, Cur Avg Loss: 1.28203349, Log Avg loss: 1.13285541, Global Avg Loss: 1.62331548, Time: 0.0402 Steps: 61150, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001548, Sample Num: 24768, Cur Loss: 1.09374547, Cur Avg Loss: 1.28245116, Log Avg loss: 1.34668933, Global Avg Loss: 1.62327025, Time: 0.0402 Steps: 61160, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001558, Sample Num: 24928, Cur Loss: 2.58259106, Cur Avg Loss: 1.28538633, Log Avg loss: 1.73975032, Global Avg Loss: 1.62328930, Time: 0.0403 Steps: 61170, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001568, Sample Num: 25088, Cur Loss: 0.39638782, Cur Avg Loss: 1.28310142, Log Avg loss: 0.92711297, Global Avg Loss: 1.62317551, Time: 0.0402 Steps: 61180, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001578, Sample Num: 25248, Cur Loss: 1.81369543, Cur Avg Loss: 1.28537577, Log Avg loss: 1.64199315, Global Avg Loss: 1.62317858, Time: 0.0402 Steps: 61190, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001588, Sample Num: 25408, Cur Loss: 2.69477940, Cur Avg Loss: 1.28661964, Log Avg loss: 1.48290293, Global Avg Loss: 1.62315566, Time: 0.0403 Steps: 61200, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001598, Sample Num: 25568, Cur Loss: 0.77418852, Cur Avg Loss: 1.28485369, Log Avg loss: 1.00442080, Global Avg Loss: 1.62305458, Time: 0.0401 Steps: 61210, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001608, Sample Num: 25728, Cur Loss: 0.89398557, Cur Avg Loss: 1.28440305, Log Avg loss: 1.21239005, Global Avg Loss: 1.62298750, Time: 0.0402 Steps: 61220, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001618, Sample Num: 25888, Cur Loss: 0.90343583, Cur Avg Loss: 1.28470111, Log Avg loss: 1.33262855, Global Avg Loss: 1.62294007, Time: 0.0403 Steps: 61230, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001628, Sample Num: 26048, Cur Loss: 1.63790941, Cur Avg Loss: 1.28287781, Log Avg loss: 0.98786804, Global Avg Loss: 1.62283637, Time: 0.0402 Steps: 61240, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001638, Sample Num: 26208, Cur Loss: 1.25225854, Cur Avg Loss: 1.28094519, Log Avg loss: 0.96631496, Global Avg Loss: 1.62272919, Time: 0.0402 Steps: 61250, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001648, Sample Num: 26368, Cur Loss: 2.22953868, Cur Avg Loss: 1.28041981, Log Avg loss: 1.19436357, Global Avg Loss: 1.62265926, Time: 0.0402 Steps: 61260, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001658, Sample Num: 26528, Cur Loss: 1.93709409, Cur Avg Loss: 1.28129642, Log Avg loss: 1.42576098, Global Avg Loss: 1.62262712, Time: 0.0402 Steps: 61270, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001668, Sample Num: 26688, Cur Loss: 2.19370770, Cur Avg Loss: 1.28248138, Log Avg loss: 1.47894737, Global Avg Loss: 1.62260368, Time: 0.0402 Steps: 61280, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001678, Sample Num: 26848, Cur Loss: 1.31973326, Cur Avg Loss: 1.28171462, Log Avg loss: 1.15381973, Global Avg Loss: 1.62252719, Time: 0.0402 Steps: 61290, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001688, Sample Num: 27008, Cur Loss: 0.70671964, Cur Avg Loss: 1.28225262, Log Avg loss: 1.37252872, Global Avg Loss: 1.62248641, Time: 0.0402 Steps: 61300, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001698, Sample Num: 27168, Cur Loss: 0.79194492, Cur Avg Loss: 1.28434828, Log Avg loss: 1.63809636, Global Avg Loss: 1.62248895, Time: 0.0402 Steps: 61310, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001708, Sample Num: 27328, Cur Loss: 0.44670045, Cur Avg Loss: 1.28349695, Log Avg loss: 1.13894124, Global Avg Loss: 1.62241010, Time: 0.0402 Steps: 61320, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001718, Sample Num: 27488, Cur Loss: 0.99869478, Cur Avg Loss: 1.28228346, Log Avg loss: 1.07501788, Global Avg Loss: 1.62232084, Time: 0.0403 Steps: 61330, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001728, Sample Num: 27648, Cur Loss: 0.87955344, Cur Avg Loss: 1.28257353, Log Avg loss: 1.33240882, Global Avg Loss: 1.62227358, Time: 0.0402 Steps: 61340, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001738, Sample Num: 27808, Cur Loss: 0.36196956, Cur Avg Loss: 1.28108067, Log Avg loss: 1.02311359, Global Avg Loss: 1.62217592, Time: 0.0407 Steps: 61350, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001748, Sample Num: 27968, Cur Loss: 1.88627386, Cur Avg Loss: 1.28192674, Log Avg loss: 1.42897337, Global Avg Loss: 1.62214443, Time: 0.0405 Steps: 61360, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001758, Sample Num: 28128, Cur Loss: 0.76072586, Cur Avg Loss: 1.28191417, Log Avg loss: 1.27971705, Global Avg Loss: 1.62208863, Time: 0.0642 Steps: 61370, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001768, Sample Num: 28288, Cur Loss: 0.65762889, Cur Avg Loss: 1.28055286, Log Avg loss: 1.04123549, Global Avg Loss: 1.62199400, Time: 0.0454 Steps: 61380, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001778, Sample Num: 28448, Cur Loss: 1.85345781, Cur Avg Loss: 1.28357690, Log Avg loss: 1.81822753, Global Avg Loss: 1.62202597, Time: 0.0717 Steps: 61390, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001788, Sample Num: 28608, Cur Loss: 0.47418195, Cur Avg Loss: 1.28306458, Log Avg loss: 1.19197403, Global Avg Loss: 1.62195593, Time: 0.0587 Steps: 61400, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001798, Sample Num: 28768, Cur Loss: 0.76519126, Cur Avg Loss: 1.28202982, Log Avg loss: 1.09701452, Global Avg Loss: 1.62187044, Time: 0.0535 Steps: 61410, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001808, Sample Num: 28928, Cur Loss: 3.27547956, Cur Avg Loss: 1.28245386, Log Avg loss: 1.35869641, Global Avg Loss: 1.62182760, Time: 0.0549 Steps: 61420, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001818, Sample Num: 29088, Cur Loss: 1.54762888, Cur Avg Loss: 1.28133297, Log Avg loss: 1.07867520, Global Avg Loss: 1.62173918, Time: 0.0403 Steps: 61430, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001828, Sample Num: 29248, Cur Loss: 0.72727603, Cur Avg Loss: 1.28265451, Log Avg loss: 1.52291130, Global Avg Loss: 1.62172309, Time: 0.0995 Steps: 61440, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001838, Sample Num: 29408, Cur Loss: 1.45144713, Cur Avg Loss: 1.28394000, Log Avg loss: 1.51892692, Global Avg Loss: 1.62170636, Time: 0.0402 Steps: 61450, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001848, Sample Num: 29568, Cur Loss: 2.82014537, Cur Avg Loss: 1.28437725, Log Avg loss: 1.36474424, Global Avg Loss: 1.62166456, Time: 0.0402 Steps: 61460, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001858, Sample Num: 29728, Cur Loss: 0.81554151, Cur Avg Loss: 1.28259767, Log Avg loss: 0.95373007, Global Avg Loss: 1.62155589, Time: 0.0403 Steps: 61470, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001868, Sample Num: 29888, Cur Loss: 1.05656791, Cur Avg Loss: 1.28159505, Log Avg loss: 1.09530901, Global Avg Loss: 1.62147030, Time: 0.0402 Steps: 61480, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001878, Sample Num: 30048, Cur Loss: 1.30324292, Cur Avg Loss: 1.28182849, Log Avg loss: 1.32543539, Global Avg Loss: 1.62142215, Time: 0.0402 Steps: 61490, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001888, Sample Num: 30208, Cur Loss: 0.62679964, Cur Avg Loss: 1.28121630, Log Avg loss: 1.16624589, Global Avg Loss: 1.62134814, Time: 0.0403 Steps: 61500, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001898, Sample Num: 30368, Cur Loss: 1.36682963, Cur Avg Loss: 1.28021124, Log Avg loss: 1.09045731, Global Avg Loss: 1.62126183, Time: 0.0403 Steps: 61510, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001908, Sample Num: 30528, Cur Loss: 1.45564806, Cur Avg Loss: 1.27923108, Log Avg loss: 1.09319594, Global Avg Loss: 1.62117600, Time: 0.0402 Steps: 61520, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001918, Sample Num: 30688, Cur Loss: 1.42880237, Cur Avg Loss: 1.28001076, Log Avg loss: 1.42877443, Global Avg Loss: 1.62114473, Time: 0.0402 Steps: 61530, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001928, Sample Num: 30848, Cur Loss: 2.66442919, Cur Avg Loss: 1.28080306, Log Avg loss: 1.43276603, Global Avg Loss: 1.62111412, Time: 0.0403 Steps: 61540, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001938, Sample Num: 31008, Cur Loss: 0.81813300, Cur Avg Loss: 1.28080679, Log Avg loss: 1.28152472, Global Avg Loss: 1.62105894, Time: 0.0402 Steps: 61550, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001948, Sample Num: 31168, Cur Loss: 2.47030830, Cur Avg Loss: 1.28251206, Log Avg loss: 1.61299370, Global Avg Loss: 1.62105763, Time: 0.0403 Steps: 61560, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001958, Sample Num: 31328, Cur Loss: 1.62551141, Cur Avg Loss: 1.28205023, Log Avg loss: 1.19208719, Global Avg Loss: 1.62098796, Time: 0.0403 Steps: 61570, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001968, Sample Num: 31488, Cur Loss: 0.32067689, Cur Avg Loss: 1.28169476, Log Avg loss: 1.21209295, Global Avg Loss: 1.62092156, Time: 0.0402 Steps: 61580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001978, Sample Num: 31648, Cur Loss: 1.08659506, Cur Avg Loss: 1.28298339, Log Avg loss: 1.53658637, Global Avg Loss: 1.62090787, Time: 0.0403 Steps: 61590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001988, Sample Num: 31808, Cur Loss: 0.83919728, Cur Avg Loss: 1.28366003, Log Avg loss: 1.41749831, Global Avg Loss: 1.62087485, Time: 0.0403 Steps: 61600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001998, Sample Num: 31968, Cur Loss: 1.08452439, Cur Avg Loss: 1.28225334, Log Avg loss: 1.00260304, Global Avg Loss: 1.62077449, Time: 0.0402 Steps: 61610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002008, Sample Num: 32128, Cur Loss: 0.46550399, Cur Avg Loss: 1.28156628, Log Avg loss: 1.14429328, Global Avg Loss: 1.62069717, Time: 0.0403 Steps: 61620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002018, Sample Num: 32288, Cur Loss: 1.51635277, Cur Avg Loss: 1.28187555, Log Avg loss: 1.34397663, Global Avg Loss: 1.62065227, Time: 0.0402 Steps: 61630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002028, Sample Num: 32448, Cur Loss: 2.01795101, Cur Avg Loss: 1.28411622, Log Avg loss: 1.73628363, Global Avg Loss: 1.62067103, Time: 0.0403 Steps: 61640, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002038, Sample Num: 32608, Cur Loss: 2.28788900, Cur Avg Loss: 1.28556101, Log Avg loss: 1.57856325, Global Avg Loss: 1.62066420, Time: 0.0402 Steps: 61650, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002048, Sample Num: 32768, Cur Loss: 1.42992747, Cur Avg Loss: 1.28481345, Log Avg loss: 1.13246118, Global Avg Loss: 1.62058502, Time: 0.0403 Steps: 61660, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002058, Sample Num: 32928, Cur Loss: 0.37162155, Cur Avg Loss: 1.28266746, Log Avg loss: 0.84316898, Global Avg Loss: 1.62045896, Time: 0.0402 Steps: 61670, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002068, Sample Num: 33088, Cur Loss: 3.85373068, Cur Avg Loss: 1.28402476, Log Avg loss: 1.56335685, Global Avg Loss: 1.62044970, Time: 0.0402 Steps: 61680, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002078, Sample Num: 33248, Cur Loss: 1.47453380, Cur Avg Loss: 1.28364521, Log Avg loss: 1.20515507, Global Avg Loss: 1.62038238, Time: 0.0402 Steps: 61690, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002088, Sample Num: 33408, Cur Loss: 0.54497015, Cur Avg Loss: 1.28260491, Log Avg loss: 1.06643006, Global Avg Loss: 1.62029260, Time: 0.0402 Steps: 61700, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002098, Sample Num: 33568, Cur Loss: 1.16229773, Cur Avg Loss: 1.28338607, Log Avg loss: 1.44649274, Global Avg Loss: 1.62026444, Time: 0.0568 Steps: 61710, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002108, Sample Num: 33728, Cur Loss: 0.63657641, Cur Avg Loss: 1.28437281, Log Avg loss: 1.49139104, Global Avg Loss: 1.62024356, Time: 0.0407 Steps: 61720, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002118, Sample Num: 33888, Cur Loss: 0.49076855, Cur Avg Loss: 1.28350952, Log Avg loss: 1.10152633, Global Avg Loss: 1.62015953, Time: 0.0602 Steps: 61730, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002128, Sample Num: 34048, Cur Loss: 1.28016043, Cur Avg Loss: 1.28421599, Log Avg loss: 1.43384675, Global Avg Loss: 1.62012935, Time: 0.0422 Steps: 61740, Updated lr: 0.000042 ***** Running evaluation checkpoint-61741 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-61741 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 92.712914, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.353087, "eval_total_loss": 951.220153, "eval_mae": 1.011591, "eval_mse": 1.352987, "eval_r2": 0.139953, "eval_sp_statistic": 0.386959, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.506323, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.372191, "test_total_loss": 688.83974, "test_mae": 1.025708, "test_mse": 1.372333, "test_r2": 0.114284, "test_sp_statistic": 0.405924, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.535261, "test_ps_pvalue": 0.0, "lr": 4.239829302987198e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6201207630855308, "train_cur_epoch_loss": 2733.901643037796, "train_cur_epoch_avg_loss": 1.2841247736203834, "train_cur_epoch_time": 92.71291375160217, "train_cur_epoch_avg_time": 0.043547634453547285, "epoch": 29, "step": 61741} ################################################## Training, Epoch: 0030, Batch: 000009, Sample Num: 144, Cur Loss: 0.71951759, Cur Avg Loss: 0.96145363, Log Avg loss: 0.97431025, Global Avg Loss: 1.62002476, Time: 0.0402 Steps: 61750, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000019, Sample Num: 304, Cur Loss: 0.95419288, Cur Avg Loss: 1.13694092, Log Avg loss: 1.29487949, Global Avg Loss: 1.61997212, Time: 0.0664 Steps: 61760, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000029, Sample Num: 464, Cur Loss: 2.03032470, Cur Avg Loss: 1.19665184, Log Avg loss: 1.31010258, Global Avg Loss: 1.61992195, Time: 0.0854 Steps: 61770, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000039, Sample Num: 624, Cur Loss: 2.54307508, Cur Avg Loss: 1.31559515, Log Avg loss: 1.66053075, Global Avg Loss: 1.61992852, Time: 0.0404 Steps: 61780, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000049, Sample Num: 784, Cur Loss: 1.04948568, Cur Avg Loss: 1.30183038, Log Avg loss: 1.24814777, Global Avg Loss: 1.61986836, Time: 0.0995 Steps: 61790, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000059, Sample Num: 944, Cur Loss: 1.16280484, Cur Avg Loss: 1.33257715, Log Avg loss: 1.48323634, Global Avg Loss: 1.61984625, Time: 0.0745 Steps: 61800, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000069, Sample Num: 1104, Cur Loss: 1.32832527, Cur Avg Loss: 1.33393002, Log Avg loss: 1.34191195, Global Avg Loss: 1.61980128, Time: 0.0542 Steps: 61810, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000079, Sample Num: 1264, Cur Loss: 0.50846052, Cur Avg Loss: 1.34990709, Log Avg loss: 1.46014886, Global Avg Loss: 1.61977546, Time: 0.0536 Steps: 61820, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000089, Sample Num: 1424, Cur Loss: 1.38898730, Cur Avg Loss: 1.34142263, Log Avg loss: 1.27439539, Global Avg Loss: 1.61971960, Time: 0.0922 Steps: 61830, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000099, Sample Num: 1584, Cur Loss: 1.88652778, Cur Avg Loss: 1.30714252, Log Avg loss: 1.00204951, Global Avg Loss: 1.61961971, Time: 0.0406 Steps: 61840, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000109, Sample Num: 1744, Cur Loss: 1.57165802, Cur Avg Loss: 1.30005592, Log Avg loss: 1.22989858, Global Avg Loss: 1.61955670, Time: 0.0403 Steps: 61850, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000119, Sample Num: 1904, Cur Loss: 0.34702787, Cur Avg Loss: 1.32835964, Log Avg loss: 1.63687021, Global Avg Loss: 1.61955950, Time: 0.0403 Steps: 61860, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000129, Sample Num: 2064, Cur Loss: 0.90774632, Cur Avg Loss: 1.31895583, Log Avg loss: 1.20705048, Global Avg Loss: 1.61949283, Time: 0.0403 Steps: 61870, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000139, Sample Num: 2224, Cur Loss: 1.66895413, Cur Avg Loss: 1.34887764, Log Avg loss: 1.73486903, Global Avg Loss: 1.61951147, Time: 0.0404 Steps: 61880, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000149, Sample Num: 2384, Cur Loss: 1.31162238, Cur Avg Loss: 1.33291641, Log Avg loss: 1.11105529, Global Avg Loss: 1.61942932, Time: 0.0402 Steps: 61890, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000159, Sample Num: 2544, Cur Loss: 0.90342450, Cur Avg Loss: 1.31372197, Log Avg loss: 1.02772490, Global Avg Loss: 1.61933373, Time: 0.0403 Steps: 61900, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000169, Sample Num: 2704, Cur Loss: 0.97862828, Cur Avg Loss: 1.33141267, Log Avg loss: 1.61269467, Global Avg Loss: 1.61933266, Time: 0.0403 Steps: 61910, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000179, Sample Num: 2864, Cur Loss: 1.43371403, Cur Avg Loss: 1.34091751, Log Avg loss: 1.50154936, Global Avg Loss: 1.61931363, Time: 0.0402 Steps: 61920, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000189, Sample Num: 3024, Cur Loss: 1.49738717, Cur Avg Loss: 1.33257948, Log Avg loss: 1.18332872, Global Avg Loss: 1.61924324, Time: 0.0402 Steps: 61930, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000199, Sample Num: 3184, Cur Loss: 1.22743726, Cur Avg Loss: 1.32471299, Log Avg loss: 1.17603641, Global Avg Loss: 1.61917168, Time: 0.0404 Steps: 61940, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000209, Sample Num: 3344, Cur Loss: 0.39038205, Cur Avg Loss: 1.30830774, Log Avg loss: 0.98184328, Global Avg Loss: 1.61906880, Time: 0.0402 Steps: 61950, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000219, Sample Num: 3504, Cur Loss: 0.72731936, Cur Avg Loss: 1.30047430, Log Avg loss: 1.13675544, Global Avg Loss: 1.61899096, Time: 0.0402 Steps: 61960, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000229, Sample Num: 3664, Cur Loss: 0.92871606, Cur Avg Loss: 1.29221274, Log Avg loss: 1.11128452, Global Avg Loss: 1.61890903, Time: 0.0403 Steps: 61970, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000239, Sample Num: 3824, Cur Loss: 2.57365990, Cur Avg Loss: 1.29536893, Log Avg loss: 1.36764576, Global Avg Loss: 1.61886849, Time: 0.0403 Steps: 61980, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000249, Sample Num: 3984, Cur Loss: 0.66510397, Cur Avg Loss: 1.28141149, Log Avg loss: 0.94782861, Global Avg Loss: 1.61876024, Time: 0.0402 Steps: 61990, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000259, Sample Num: 4144, Cur Loss: 1.69239128, Cur Avg Loss: 1.28208719, Log Avg loss: 1.29891213, Global Avg Loss: 1.61870866, Time: 0.0402 Steps: 62000, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000269, Sample Num: 4304, Cur Loss: 1.42146146, Cur Avg Loss: 1.27801364, Log Avg loss: 1.17250861, Global Avg Loss: 1.61863670, Time: 0.0402 Steps: 62010, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000279, Sample Num: 4464, Cur Loss: 1.90691447, Cur Avg Loss: 1.27925467, Log Avg loss: 1.31263846, Global Avg Loss: 1.61858736, Time: 0.0402 Steps: 62020, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000289, Sample Num: 4624, Cur Loss: 1.24127054, Cur Avg Loss: 1.28060814, Log Avg loss: 1.31836983, Global Avg Loss: 1.61853896, Time: 0.0402 Steps: 62030, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000299, Sample Num: 4784, Cur Loss: 0.85438013, Cur Avg Loss: 1.28355481, Log Avg loss: 1.36871359, Global Avg Loss: 1.61849869, Time: 0.0403 Steps: 62040, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000309, Sample Num: 4944, Cur Loss: 1.14430928, Cur Avg Loss: 1.29524812, Log Avg loss: 1.64487810, Global Avg Loss: 1.61850294, Time: 0.0403 Steps: 62050, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000319, Sample Num: 5104, Cur Loss: 1.14636075, Cur Avg Loss: 1.29797959, Log Avg loss: 1.38238207, Global Avg Loss: 1.61846490, Time: 0.0402 Steps: 62060, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000329, Sample Num: 5264, Cur Loss: 1.18844652, Cur Avg Loss: 1.30356221, Log Avg loss: 1.48164784, Global Avg Loss: 1.61844285, Time: 0.0402 Steps: 62070, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000339, Sample Num: 5424, Cur Loss: 1.65149903, Cur Avg Loss: 1.30427705, Log Avg loss: 1.32779525, Global Avg Loss: 1.61839604, Time: 0.0402 Steps: 62080, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000349, Sample Num: 5584, Cur Loss: 1.04982948, Cur Avg Loss: 1.29568842, Log Avg loss: 1.00453389, Global Avg Loss: 1.61829717, Time: 0.0402 Steps: 62090, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000359, Sample Num: 5744, Cur Loss: 0.82221931, Cur Avg Loss: 1.29311356, Log Avg loss: 1.20325089, Global Avg Loss: 1.61823033, Time: 0.0402 Steps: 62100, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000369, Sample Num: 5904, Cur Loss: 1.69799817, Cur Avg Loss: 1.29841824, Log Avg loss: 1.48885631, Global Avg Loss: 1.61820951, Time: 0.0402 Steps: 62110, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000379, Sample Num: 6064, Cur Loss: 0.54469800, Cur Avg Loss: 1.29018126, Log Avg loss: 0.98623658, Global Avg Loss: 1.61810777, Time: 0.0431 Steps: 62120, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000389, Sample Num: 6224, Cur Loss: 1.00585651, Cur Avg Loss: 1.28464939, Log Avg loss: 1.07499168, Global Avg Loss: 1.61802035, Time: 0.0403 Steps: 62130, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000399, Sample Num: 6384, Cur Loss: 1.30139041, Cur Avg Loss: 1.28351899, Log Avg loss: 1.23954628, Global Avg Loss: 1.61795945, Time: 0.0695 Steps: 62140, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000409, Sample Num: 6544, Cur Loss: 1.51385450, Cur Avg Loss: 1.28204760, Log Avg loss: 1.22333929, Global Avg Loss: 1.61789595, Time: 0.0677 Steps: 62150, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000419, Sample Num: 6704, Cur Loss: 0.81533802, Cur Avg Loss: 1.28590571, Log Avg loss: 1.44370222, Global Avg Loss: 1.61786793, Time: 0.0904 Steps: 62160, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000429, Sample Num: 6864, Cur Loss: 1.27467549, Cur Avg Loss: 1.28351109, Log Avg loss: 1.18317666, Global Avg Loss: 1.61779801, Time: 0.0529 Steps: 62170, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000439, Sample Num: 7024, Cur Loss: 1.50387311, Cur Avg Loss: 1.28217236, Log Avg loss: 1.22474068, Global Avg Loss: 1.61773480, Time: 0.0430 Steps: 62180, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000449, Sample Num: 7184, Cur Loss: 0.74181455, Cur Avg Loss: 1.27960027, Log Avg loss: 1.16668578, Global Avg Loss: 1.61766227, Time: 0.0910 Steps: 62190, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000459, Sample Num: 7344, Cur Loss: 0.39928615, Cur Avg Loss: 1.27380742, Log Avg loss: 1.01370828, Global Avg Loss: 1.61756517, Time: 0.0402 Steps: 62200, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000469, Sample Num: 7504, Cur Loss: 0.82177931, Cur Avg Loss: 1.27438551, Log Avg loss: 1.30091995, Global Avg Loss: 1.61751427, Time: 0.0403 Steps: 62210, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000479, Sample Num: 7664, Cur Loss: 0.95834148, Cur Avg Loss: 1.27018288, Log Avg loss: 1.07307933, Global Avg Loss: 1.61742677, Time: 0.0402 Steps: 62220, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000489, Sample Num: 7824, Cur Loss: 1.43987918, Cur Avg Loss: 1.26842094, Log Avg loss: 1.18402407, Global Avg Loss: 1.61735712, Time: 0.0402 Steps: 62230, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000499, Sample Num: 7984, Cur Loss: 2.22615290, Cur Avg Loss: 1.26520369, Log Avg loss: 1.10788033, Global Avg Loss: 1.61727527, Time: 0.0402 Steps: 62240, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000509, Sample Num: 8144, Cur Loss: 1.20877624, Cur Avg Loss: 1.26642270, Log Avg loss: 1.32725102, Global Avg Loss: 1.61722868, Time: 0.0403 Steps: 62250, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000519, Sample Num: 8304, Cur Loss: 0.36248836, Cur Avg Loss: 1.26776236, Log Avg loss: 1.33595118, Global Avg Loss: 1.61718350, Time: 0.0403 Steps: 62260, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000529, Sample Num: 8464, Cur Loss: 1.04061389, Cur Avg Loss: 1.26632069, Log Avg loss: 1.19149806, Global Avg Loss: 1.61711514, Time: 0.0403 Steps: 62270, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000539, Sample Num: 8624, Cur Loss: 1.19115090, Cur Avg Loss: 1.26677952, Log Avg loss: 1.29105178, Global Avg Loss: 1.61706278, Time: 0.0402 Steps: 62280, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000549, Sample Num: 8784, Cur Loss: 0.84482890, Cur Avg Loss: 1.26671659, Log Avg loss: 1.26332462, Global Avg Loss: 1.61700600, Time: 0.0403 Steps: 62290, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000559, Sample Num: 8944, Cur Loss: 0.66114444, Cur Avg Loss: 1.26138246, Log Avg loss: 0.96853887, Global Avg Loss: 1.61690191, Time: 0.0403 Steps: 62300, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000569, Sample Num: 9104, Cur Loss: 0.30787164, Cur Avg Loss: 1.26091721, Log Avg loss: 1.23490922, Global Avg Loss: 1.61684060, Time: 0.0403 Steps: 62310, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000579, Sample Num: 9264, Cur Loss: 2.21291447, Cur Avg Loss: 1.26133925, Log Avg loss: 1.28535355, Global Avg Loss: 1.61678741, Time: 0.0403 Steps: 62320, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000589, Sample Num: 9424, Cur Loss: 0.36141980, Cur Avg Loss: 1.25920949, Log Avg loss: 1.13589659, Global Avg Loss: 1.61671026, Time: 0.0402 Steps: 62330, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000599, Sample Num: 9584, Cur Loss: 2.87131548, Cur Avg Loss: 1.26228767, Log Avg loss: 1.44359219, Global Avg Loss: 1.61668249, Time: 0.0403 Steps: 62340, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000609, Sample Num: 9744, Cur Loss: 1.86413813, Cur Avg Loss: 1.26525891, Log Avg loss: 1.44323641, Global Avg Loss: 1.61665467, Time: 0.0403 Steps: 62350, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000619, Sample Num: 9904, Cur Loss: 1.29850912, Cur Avg Loss: 1.26260890, Log Avg loss: 1.10122337, Global Avg Loss: 1.61657202, Time: 0.0403 Steps: 62360, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000629, Sample Num: 10064, Cur Loss: 1.20244598, Cur Avg Loss: 1.26279054, Log Avg loss: 1.27403403, Global Avg Loss: 1.61651710, Time: 0.0403 Steps: 62370, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000639, Sample Num: 10224, Cur Loss: 1.73388219, Cur Avg Loss: 1.26037874, Log Avg loss: 1.10867619, Global Avg Loss: 1.61643569, Time: 0.0403 Steps: 62380, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000649, Sample Num: 10384, Cur Loss: 2.28924227, Cur Avg Loss: 1.26094170, Log Avg loss: 1.29691483, Global Avg Loss: 1.61638447, Time: 0.0402 Steps: 62390, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000659, Sample Num: 10544, Cur Loss: 2.21847486, Cur Avg Loss: 1.25642409, Log Avg loss: 0.96323161, Global Avg Loss: 1.61627980, Time: 0.0402 Steps: 62400, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000669, Sample Num: 10704, Cur Loss: 2.26255226, Cur Avg Loss: 1.25490450, Log Avg loss: 1.15476342, Global Avg Loss: 1.61620585, Time: 0.0402 Steps: 62410, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000679, Sample Num: 10864, Cur Loss: 0.62271184, Cur Avg Loss: 1.25455686, Log Avg loss: 1.23129967, Global Avg Loss: 1.61614419, Time: 0.0403 Steps: 62420, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000689, Sample Num: 11024, Cur Loss: 0.72497773, Cur Avg Loss: 1.25612307, Log Avg loss: 1.36246881, Global Avg Loss: 1.61610355, Time: 0.0402 Steps: 62430, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000699, Sample Num: 11184, Cur Loss: 1.99923992, Cur Avg Loss: 1.25938253, Log Avg loss: 1.48395897, Global Avg Loss: 1.61608239, Time: 0.0403 Steps: 62440, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000709, Sample Num: 11344, Cur Loss: 0.96469200, Cur Avg Loss: 1.25685786, Log Avg loss: 1.08038332, Global Avg Loss: 1.61599661, Time: 0.0402 Steps: 62450, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000719, Sample Num: 11504, Cur Loss: 0.69850689, Cur Avg Loss: 1.25442616, Log Avg loss: 1.08201897, Global Avg Loss: 1.61591112, Time: 0.0402 Steps: 62460, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000729, Sample Num: 11664, Cur Loss: 1.05725145, Cur Avg Loss: 1.25042880, Log Avg loss: 0.96301854, Global Avg Loss: 1.61580661, Time: 0.1002 Steps: 62470, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000739, Sample Num: 11824, Cur Loss: 1.65132070, Cur Avg Loss: 1.25141470, Log Avg loss: 1.32328663, Global Avg Loss: 1.61575979, Time: 0.0457 Steps: 62480, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000749, Sample Num: 11984, Cur Loss: 1.85319781, Cur Avg Loss: 1.24812952, Log Avg loss: 1.00535482, Global Avg Loss: 1.61566211, Time: 0.0409 Steps: 62490, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000759, Sample Num: 12144, Cur Loss: 1.63494861, Cur Avg Loss: 1.24763053, Log Avg loss: 1.21025644, Global Avg Loss: 1.61559724, Time: 0.0457 Steps: 62500, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000769, Sample Num: 12304, Cur Loss: 0.24380460, Cur Avg Loss: 1.24227698, Log Avg loss: 0.83594272, Global Avg Loss: 1.61547252, Time: 0.0508 Steps: 62510, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000779, Sample Num: 12464, Cur Loss: 0.80306649, Cur Avg Loss: 1.24754818, Log Avg loss: 1.65290314, Global Avg Loss: 1.61547850, Time: 0.0544 Steps: 62520, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000789, Sample Num: 12624, Cur Loss: 1.45209014, Cur Avg Loss: 1.24795769, Log Avg loss: 1.27985829, Global Avg Loss: 1.61542483, Time: 0.0609 Steps: 62530, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000799, Sample Num: 12784, Cur Loss: 1.16498244, Cur Avg Loss: 1.24779825, Log Avg loss: 1.23521849, Global Avg Loss: 1.61536404, Time: 0.0455 Steps: 62540, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000809, Sample Num: 12944, Cur Loss: 0.72671765, Cur Avg Loss: 1.24673269, Log Avg loss: 1.16159462, Global Avg Loss: 1.61529149, Time: 0.0405 Steps: 62550, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000819, Sample Num: 13104, Cur Loss: 1.74288201, Cur Avg Loss: 1.24622755, Log Avg loss: 1.20536168, Global Avg Loss: 1.61522597, Time: 0.0403 Steps: 62560, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000829, Sample Num: 13264, Cur Loss: 1.66596115, Cur Avg Loss: 1.24772301, Log Avg loss: 1.37020128, Global Avg Loss: 1.61518681, Time: 0.0403 Steps: 62570, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000839, Sample Num: 13424, Cur Loss: 1.43035555, Cur Avg Loss: 1.24829381, Log Avg loss: 1.29561285, Global Avg Loss: 1.61513574, Time: 0.0403 Steps: 62580, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000849, Sample Num: 13584, Cur Loss: 1.79028416, Cur Avg Loss: 1.25273704, Log Avg loss: 1.62552453, Global Avg Loss: 1.61513740, Time: 0.0403 Steps: 62590, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000859, Sample Num: 13744, Cur Loss: 3.13898063, Cur Avg Loss: 1.25167559, Log Avg loss: 1.16155843, Global Avg Loss: 1.61506494, Time: 0.0403 Steps: 62600, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000869, Sample Num: 13904, Cur Loss: 0.92431408, Cur Avg Loss: 1.25269321, Log Avg loss: 1.34010649, Global Avg Loss: 1.61502103, Time: 0.0402 Steps: 62610, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000879, Sample Num: 14064, Cur Loss: 1.56246376, Cur Avg Loss: 1.25810327, Log Avg loss: 1.72823719, Global Avg Loss: 1.61503911, Time: 0.0402 Steps: 62620, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000889, Sample Num: 14224, Cur Loss: 1.05724525, Cur Avg Loss: 1.26039472, Log Avg loss: 1.46181383, Global Avg Loss: 1.61501464, Time: 0.0403 Steps: 62630, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000899, Sample Num: 14384, Cur Loss: 1.51794314, Cur Avg Loss: 1.25976169, Log Avg loss: 1.20348463, Global Avg Loss: 1.61494894, Time: 0.0402 Steps: 62640, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000909, Sample Num: 14544, Cur Loss: 1.09110558, Cur Avg Loss: 1.26074383, Log Avg loss: 1.34903847, Global Avg Loss: 1.61490650, Time: 0.0403 Steps: 62650, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000919, Sample Num: 14704, Cur Loss: 0.73522741, Cur Avg Loss: 1.26063749, Log Avg loss: 1.25097102, Global Avg Loss: 1.61484842, Time: 0.0402 Steps: 62660, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000929, Sample Num: 14864, Cur Loss: 1.61457896, Cur Avg Loss: 1.26098036, Log Avg loss: 1.29249061, Global Avg Loss: 1.61479698, Time: 0.0402 Steps: 62670, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000939, Sample Num: 15024, Cur Loss: 0.94426727, Cur Avg Loss: 1.25930569, Log Avg loss: 1.10372881, Global Avg Loss: 1.61471544, Time: 0.0403 Steps: 62680, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000949, Sample Num: 15184, Cur Loss: 1.65944529, Cur Avg Loss: 1.25945771, Log Avg loss: 1.27373233, Global Avg Loss: 1.61466105, Time: 0.0403 Steps: 62690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000959, Sample Num: 15344, Cur Loss: 0.49181908, Cur Avg Loss: 1.25477778, Log Avg loss: 0.81065207, Global Avg Loss: 1.61453282, Time: 0.0402 Steps: 62700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000969, Sample Num: 15504, Cur Loss: 0.58246797, Cur Avg Loss: 1.25411661, Log Avg loss: 1.19071026, Global Avg Loss: 1.61446524, Time: 0.0402 Steps: 62710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000979, Sample Num: 15664, Cur Loss: 0.44676012, Cur Avg Loss: 1.25035491, Log Avg loss: 0.88584680, Global Avg Loss: 1.61434907, Time: 0.0402 Steps: 62720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000989, Sample Num: 15824, Cur Loss: 1.14979291, Cur Avg Loss: 1.25021783, Log Avg loss: 1.23679768, Global Avg Loss: 1.61428888, Time: 0.0403 Steps: 62730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000999, Sample Num: 15984, Cur Loss: 1.40020013, Cur Avg Loss: 1.25228695, Log Avg loss: 1.45692263, Global Avg Loss: 1.61426380, Time: 0.0402 Steps: 62740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001009, Sample Num: 16144, Cur Loss: 0.74907577, Cur Avg Loss: 1.25226000, Log Avg loss: 1.24956792, Global Avg Loss: 1.61420568, Time: 0.0402 Steps: 62750, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001019, Sample Num: 16304, Cur Loss: 1.57602918, Cur Avg Loss: 1.25460980, Log Avg loss: 1.49170462, Global Avg Loss: 1.61418616, Time: 0.0402 Steps: 62760, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001029, Sample Num: 16464, Cur Loss: 1.80262899, Cur Avg Loss: 1.25255912, Log Avg loss: 1.04359442, Global Avg Loss: 1.61409526, Time: 0.0402 Steps: 62770, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001039, Sample Num: 16624, Cur Loss: 1.05039704, Cur Avg Loss: 1.25270238, Log Avg loss: 1.26744441, Global Avg Loss: 1.61404004, Time: 0.0402 Steps: 62780, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001049, Sample Num: 16784, Cur Loss: 0.60837877, Cur Avg Loss: 1.25392896, Log Avg loss: 1.38136991, Global Avg Loss: 1.61400299, Time: 0.0402 Steps: 62790, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001059, Sample Num: 16944, Cur Loss: 1.55830145, Cur Avg Loss: 1.25774358, Log Avg loss: 1.65789731, Global Avg Loss: 1.61400998, Time: 0.0402 Steps: 62800, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001069, Sample Num: 17104, Cur Loss: 0.80296171, Cur Avg Loss: 1.25442402, Log Avg loss: 0.90288334, Global Avg Loss: 1.61389676, Time: 0.0402 Steps: 62810, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001079, Sample Num: 17264, Cur Loss: 1.02672517, Cur Avg Loss: 1.25303706, Log Avg loss: 1.10477073, Global Avg Loss: 1.61381571, Time: 0.0403 Steps: 62820, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001089, Sample Num: 17424, Cur Loss: 1.62356067, Cur Avg Loss: 1.25188608, Log Avg loss: 1.12769569, Global Avg Loss: 1.61373834, Time: 0.0405 Steps: 62830, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001099, Sample Num: 17584, Cur Loss: 1.31001425, Cur Avg Loss: 1.25410262, Log Avg loss: 1.49548310, Global Avg Loss: 1.61371952, Time: 0.0404 Steps: 62840, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001109, Sample Num: 17744, Cur Loss: 1.81585693, Cur Avg Loss: 1.25431707, Log Avg loss: 1.27788537, Global Avg Loss: 1.61366609, Time: 0.0995 Steps: 62850, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001119, Sample Num: 17904, Cur Loss: 1.99594033, Cur Avg Loss: 1.25637444, Log Avg loss: 1.48453693, Global Avg Loss: 1.61364555, Time: 0.0404 Steps: 62860, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001129, Sample Num: 18064, Cur Loss: 1.13441491, Cur Avg Loss: 1.25476571, Log Avg loss: 1.07474854, Global Avg Loss: 1.61355983, Time: 0.0508 Steps: 62870, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001139, Sample Num: 18224, Cur Loss: 1.48124504, Cur Avg Loss: 1.25421965, Log Avg loss: 1.19256986, Global Avg Loss: 1.61349288, Time: 0.0671 Steps: 62880, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001149, Sample Num: 18384, Cur Loss: 2.00149417, Cur Avg Loss: 1.25340742, Log Avg loss: 1.16089363, Global Avg Loss: 1.61342091, Time: 0.0518 Steps: 62890, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001159, Sample Num: 18544, Cur Loss: 1.54037380, Cur Avg Loss: 1.25719537, Log Avg loss: 1.69243111, Global Avg Loss: 1.61343347, Time: 0.0941 Steps: 62900, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001169, Sample Num: 18704, Cur Loss: 1.25076699, Cur Avg Loss: 1.25905468, Log Avg loss: 1.47454936, Global Avg Loss: 1.61341140, Time: 0.0403 Steps: 62910, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001179, Sample Num: 18864, Cur Loss: 1.75673890, Cur Avg Loss: 1.25815146, Log Avg loss: 1.15256512, Global Avg Loss: 1.61333815, Time: 0.0403 Steps: 62920, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001189, Sample Num: 19024, Cur Loss: 0.98877585, Cur Avg Loss: 1.25945285, Log Avg loss: 1.41288666, Global Avg Loss: 1.61330630, Time: 0.0403 Steps: 62930, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001199, Sample Num: 19184, Cur Loss: 0.89436644, Cur Avg Loss: 1.25983796, Log Avg loss: 1.30562758, Global Avg Loss: 1.61325742, Time: 0.0402 Steps: 62940, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001209, Sample Num: 19344, Cur Loss: 1.59589410, Cur Avg Loss: 1.25967562, Log Avg loss: 1.24021035, Global Avg Loss: 1.61319815, Time: 0.0402 Steps: 62950, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001219, Sample Num: 19504, Cur Loss: 1.18381178, Cur Avg Loss: 1.25715100, Log Avg loss: 0.95192503, Global Avg Loss: 1.61309312, Time: 0.0402 Steps: 62960, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001229, Sample Num: 19664, Cur Loss: 0.34587920, Cur Avg Loss: 1.25507938, Log Avg loss: 1.00254918, Global Avg Loss: 1.61299617, Time: 0.0402 Steps: 62970, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001239, Sample Num: 19824, Cur Loss: 1.22965395, Cur Avg Loss: 1.25610409, Log Avg loss: 1.38203978, Global Avg Loss: 1.61295950, Time: 0.0402 Steps: 62980, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001249, Sample Num: 19984, Cur Loss: 2.17762136, Cur Avg Loss: 1.25597910, Log Avg loss: 1.24049343, Global Avg Loss: 1.61290036, Time: 0.0403 Steps: 62990, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001259, Sample Num: 20144, Cur Loss: 0.81698769, Cur Avg Loss: 1.25836954, Log Avg loss: 1.55693526, Global Avg Loss: 1.61289148, Time: 0.0402 Steps: 63000, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001269, Sample Num: 20304, Cur Loss: 1.03316700, Cur Avg Loss: 1.25838112, Log Avg loss: 1.25983971, Global Avg Loss: 1.61283545, Time: 0.0403 Steps: 63010, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001279, Sample Num: 20464, Cur Loss: 0.26336002, Cur Avg Loss: 1.25702232, Log Avg loss: 1.08459032, Global Avg Loss: 1.61275163, Time: 0.0402 Steps: 63020, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001289, Sample Num: 20624, Cur Loss: 1.33874679, Cur Avg Loss: 1.25449894, Log Avg loss: 0.93175903, Global Avg Loss: 1.61264359, Time: 0.0403 Steps: 63030, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001299, Sample Num: 20784, Cur Loss: 0.76914442, Cur Avg Loss: 1.25411075, Log Avg loss: 1.20407316, Global Avg Loss: 1.61257877, Time: 0.0403 Steps: 63040, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001309, Sample Num: 20944, Cur Loss: 2.42736721, Cur Avg Loss: 1.25677404, Log Avg loss: 1.60273502, Global Avg Loss: 1.61257721, Time: 0.0402 Steps: 63050, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001319, Sample Num: 21104, Cur Loss: 3.00164795, Cur Avg Loss: 1.25692425, Log Avg loss: 1.27658622, Global Avg Loss: 1.61252393, Time: 0.0403 Steps: 63060, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001329, Sample Num: 21264, Cur Loss: 0.56465733, Cur Avg Loss: 1.25495792, Log Avg loss: 0.99559869, Global Avg Loss: 1.61242612, Time: 0.0403 Steps: 63070, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001339, Sample Num: 21424, Cur Loss: 1.13505554, Cur Avg Loss: 1.25367962, Log Avg loss: 1.08379474, Global Avg Loss: 1.61234231, Time: 0.0402 Steps: 63080, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001349, Sample Num: 21584, Cur Loss: 1.17416835, Cur Avg Loss: 1.25356230, Log Avg loss: 1.23785283, Global Avg Loss: 1.61228295, Time: 0.0403 Steps: 63090, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001359, Sample Num: 21744, Cur Loss: 0.60746390, Cur Avg Loss: 1.25387655, Log Avg loss: 1.29626852, Global Avg Loss: 1.61223287, Time: 0.0402 Steps: 63100, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001369, Sample Num: 21904, Cur Loss: 1.66357231, Cur Avg Loss: 1.25393651, Log Avg loss: 1.26208574, Global Avg Loss: 1.61217739, Time: 0.0403 Steps: 63110, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001379, Sample Num: 22064, Cur Loss: 0.69913292, Cur Avg Loss: 1.25448117, Log Avg loss: 1.32904398, Global Avg Loss: 1.61213253, Time: 0.0403 Steps: 63120, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001389, Sample Num: 22224, Cur Loss: 1.44861567, Cur Avg Loss: 1.25503604, Log Avg loss: 1.33155318, Global Avg Loss: 1.61208809, Time: 0.0402 Steps: 63130, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001399, Sample Num: 22384, Cur Loss: 1.23061132, Cur Avg Loss: 1.25763411, Log Avg loss: 1.61850615, Global Avg Loss: 1.61208911, Time: 0.0402 Steps: 63140, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001409, Sample Num: 22544, Cur Loss: 1.58063710, Cur Avg Loss: 1.25908200, Log Avg loss: 1.46164210, Global Avg Loss: 1.61206528, Time: 0.0402 Steps: 63150, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001419, Sample Num: 22704, Cur Loss: 1.41146660, Cur Avg Loss: 1.25955190, Log Avg loss: 1.32576053, Global Avg Loss: 1.61201995, Time: 0.0402 Steps: 63160, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001429, Sample Num: 22864, Cur Loss: 1.34915221, Cur Avg Loss: 1.26008462, Log Avg loss: 1.33567708, Global Avg Loss: 1.61197621, Time: 0.0402 Steps: 63170, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001439, Sample Num: 23024, Cur Loss: 1.19850028, Cur Avg Loss: 1.26117398, Log Avg loss: 1.41684354, Global Avg Loss: 1.61194532, Time: 0.0403 Steps: 63180, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001449, Sample Num: 23184, Cur Loss: 1.96508825, Cur Avg Loss: 1.26248125, Log Avg loss: 1.45059732, Global Avg Loss: 1.61191979, Time: 0.0537 Steps: 63190, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001459, Sample Num: 23344, Cur Loss: 1.33148098, Cur Avg Loss: 1.26244860, Log Avg loss: 1.25771812, Global Avg Loss: 1.61186374, Time: 0.0403 Steps: 63200, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001469, Sample Num: 23504, Cur Loss: 1.55346751, Cur Avg Loss: 1.26109040, Log Avg loss: 1.06292868, Global Avg Loss: 1.61177690, Time: 0.0828 Steps: 63210, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001479, Sample Num: 23664, Cur Loss: 0.94433761, Cur Avg Loss: 1.25966524, Log Avg loss: 1.05030997, Global Avg Loss: 1.61168809, Time: 0.0403 Steps: 63220, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001489, Sample Num: 23824, Cur Loss: 1.37348700, Cur Avg Loss: 1.26015107, Log Avg loss: 1.33200492, Global Avg Loss: 1.61164386, Time: 0.0855 Steps: 63230, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001499, Sample Num: 23984, Cur Loss: 1.69282460, Cur Avg Loss: 1.26074049, Log Avg loss: 1.34850511, Global Avg Loss: 1.61160225, Time: 0.0460 Steps: 63240, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001509, Sample Num: 24144, Cur Loss: 0.94777071, Cur Avg Loss: 1.25993291, Log Avg loss: 1.13887649, Global Avg Loss: 1.61152751, Time: 0.0593 Steps: 63250, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001519, Sample Num: 24304, Cur Loss: 0.66613227, Cur Avg Loss: 1.26121744, Log Avg loss: 1.45505303, Global Avg Loss: 1.61150277, Time: 0.0559 Steps: 63260, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001529, Sample Num: 24464, Cur Loss: 0.93784851, Cur Avg Loss: 1.26326413, Log Avg loss: 1.57415717, Global Avg Loss: 1.61149687, Time: 0.0485 Steps: 63270, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001539, Sample Num: 24624, Cur Loss: 2.98367977, Cur Avg Loss: 1.26461904, Log Avg loss: 1.47178416, Global Avg Loss: 1.61147479, Time: 0.0796 Steps: 63280, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001549, Sample Num: 24784, Cur Loss: 0.81455839, Cur Avg Loss: 1.26261410, Log Avg loss: 0.95405382, Global Avg Loss: 1.61137092, Time: 0.0403 Steps: 63290, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001559, Sample Num: 24944, Cur Loss: 1.52916479, Cur Avg Loss: 1.26399112, Log Avg loss: 1.47729169, Global Avg Loss: 1.61134973, Time: 0.0402 Steps: 63300, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001569, Sample Num: 25104, Cur Loss: 0.93940943, Cur Avg Loss: 1.26306863, Log Avg loss: 1.11925209, Global Avg Loss: 1.61127201, Time: 0.0402 Steps: 63310, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001579, Sample Num: 25264, Cur Loss: 1.28338492, Cur Avg Loss: 1.26407062, Log Avg loss: 1.42128231, Global Avg Loss: 1.61124200, Time: 0.0402 Steps: 63320, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001589, Sample Num: 25424, Cur Loss: 0.61775881, Cur Avg Loss: 1.26276561, Log Avg loss: 1.05670571, Global Avg Loss: 1.61115444, Time: 0.0403 Steps: 63330, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001599, Sample Num: 25584, Cur Loss: 1.17059207, Cur Avg Loss: 1.26312850, Log Avg loss: 1.32079096, Global Avg Loss: 1.61110860, Time: 0.0402 Steps: 63340, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001609, Sample Num: 25744, Cur Loss: 0.81201345, Cur Avg Loss: 1.26355635, Log Avg loss: 1.33197012, Global Avg Loss: 1.61106453, Time: 0.0403 Steps: 63350, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001619, Sample Num: 25904, Cur Loss: 1.55734086, Cur Avg Loss: 1.26400158, Log Avg loss: 1.33563909, Global Avg Loss: 1.61102106, Time: 0.0402 Steps: 63360, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001629, Sample Num: 26064, Cur Loss: 0.76927811, Cur Avg Loss: 1.26251786, Log Avg loss: 1.02230339, Global Avg Loss: 1.61092816, Time: 0.0402 Steps: 63370, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001639, Sample Num: 26224, Cur Loss: 1.02855253, Cur Avg Loss: 1.26329018, Log Avg loss: 1.38910109, Global Avg Loss: 1.61089316, Time: 0.0402 Steps: 63380, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001649, Sample Num: 26384, Cur Loss: 1.45773602, Cur Avg Loss: 1.26250577, Log Avg loss: 1.13394028, Global Avg Loss: 1.61081792, Time: 0.0403 Steps: 63390, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001659, Sample Num: 26544, Cur Loss: 0.79464281, Cur Avg Loss: 1.26451418, Log Avg loss: 1.59570092, Global Avg Loss: 1.61081554, Time: 0.0402 Steps: 63400, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001669, Sample Num: 26704, Cur Loss: 1.22265851, Cur Avg Loss: 1.26438277, Log Avg loss: 1.24258200, Global Avg Loss: 1.61075747, Time: 0.0403 Steps: 63410, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001679, Sample Num: 26864, Cur Loss: 1.75731254, Cur Avg Loss: 1.26399732, Log Avg loss: 1.19966621, Global Avg Loss: 1.61069264, Time: 0.0403 Steps: 63420, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001689, Sample Num: 27024, Cur Loss: 2.02479124, Cur Avg Loss: 1.26467343, Log Avg loss: 1.37819305, Global Avg Loss: 1.61065599, Time: 0.0402 Steps: 63430, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001699, Sample Num: 27184, Cur Loss: 0.88821346, Cur Avg Loss: 1.26452446, Log Avg loss: 1.23936213, Global Avg Loss: 1.61059746, Time: 0.0402 Steps: 63440, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001709, Sample Num: 27344, Cur Loss: 1.39242887, Cur Avg Loss: 1.26365809, Log Avg loss: 1.11646306, Global Avg Loss: 1.61051959, Time: 0.0402 Steps: 63450, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001719, Sample Num: 27504, Cur Loss: 0.65957230, Cur Avg Loss: 1.26228949, Log Avg loss: 1.02839432, Global Avg Loss: 1.61042785, Time: 0.0402 Steps: 63460, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001729, Sample Num: 27664, Cur Loss: 1.26348734, Cur Avg Loss: 1.26245637, Log Avg loss: 1.29114439, Global Avg Loss: 1.61037755, Time: 0.0402 Steps: 63470, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001739, Sample Num: 27824, Cur Loss: 0.77774429, Cur Avg Loss: 1.26249570, Log Avg loss: 1.26929556, Global Avg Loss: 1.61032382, Time: 0.0402 Steps: 63480, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001749, Sample Num: 27984, Cur Loss: 0.60247052, Cur Avg Loss: 1.26174700, Log Avg loss: 1.13154808, Global Avg Loss: 1.61024841, Time: 0.0402 Steps: 63490, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001759, Sample Num: 28144, Cur Loss: 1.68665028, Cur Avg Loss: 1.26097494, Log Avg loss: 1.12594153, Global Avg Loss: 1.61017214, Time: 0.0402 Steps: 63500, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001769, Sample Num: 28304, Cur Loss: 1.07919931, Cur Avg Loss: 1.26064411, Log Avg loss: 1.20245031, Global Avg Loss: 1.61010794, Time: 0.0402 Steps: 63510, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001779, Sample Num: 28464, Cur Loss: 2.40670300, Cur Avg Loss: 1.26004537, Log Avg loss: 1.15412956, Global Avg Loss: 1.61003616, Time: 0.0402 Steps: 63520, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001789, Sample Num: 28624, Cur Loss: 0.73954105, Cur Avg Loss: 1.25973098, Log Avg loss: 1.20380025, Global Avg Loss: 1.60997221, Time: 0.0402 Steps: 63530, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001799, Sample Num: 28784, Cur Loss: 0.99854606, Cur Avg Loss: 1.25881415, Log Avg loss: 1.09479283, Global Avg Loss: 1.60989113, Time: 0.0402 Steps: 63540, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001809, Sample Num: 28944, Cur Loss: 2.27964711, Cur Avg Loss: 1.25860442, Log Avg loss: 1.22087429, Global Avg Loss: 1.60982992, Time: 0.0794 Steps: 63550, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001819, Sample Num: 29104, Cur Loss: 0.64992946, Cur Avg Loss: 1.25877441, Log Avg loss: 1.28952585, Global Avg Loss: 1.60977953, Time: 0.0757 Steps: 63560, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001829, Sample Num: 29264, Cur Loss: 1.59146690, Cur Avg Loss: 1.26018527, Log Avg loss: 1.51682109, Global Avg Loss: 1.60976490, Time: 0.0752 Steps: 63570, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001839, Sample Num: 29424, Cur Loss: 1.07142746, Cur Avg Loss: 1.26114095, Log Avg loss: 1.43593442, Global Avg Loss: 1.60973756, Time: 0.0715 Steps: 63580, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001849, Sample Num: 29584, Cur Loss: 0.68561924, Cur Avg Loss: 1.26094431, Log Avg loss: 1.22478158, Global Avg Loss: 1.60967703, Time: 0.0512 Steps: 63590, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001859, Sample Num: 29744, Cur Loss: 1.02349424, Cur Avg Loss: 1.26038712, Log Avg loss: 1.15736248, Global Avg Loss: 1.60960591, Time: 0.0726 Steps: 63600, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001869, Sample Num: 29904, Cur Loss: 1.67982459, Cur Avg Loss: 1.26027432, Log Avg loss: 1.23930473, Global Avg Loss: 1.60954769, Time: 0.0455 Steps: 63610, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001879, Sample Num: 30064, Cur Loss: 1.03725040, Cur Avg Loss: 1.25984641, Log Avg loss: 1.17987004, Global Avg Loss: 1.60948015, Time: 0.0696 Steps: 63620, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001889, Sample Num: 30224, Cur Loss: 1.18205500, Cur Avg Loss: 1.26068671, Log Avg loss: 1.41857899, Global Avg Loss: 1.60945015, Time: 0.0406 Steps: 63630, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001899, Sample Num: 30384, Cur Loss: 1.03774381, Cur Avg Loss: 1.26116094, Log Avg loss: 1.35074331, Global Avg Loss: 1.60940950, Time: 0.0403 Steps: 63640, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001909, Sample Num: 30544, Cur Loss: 1.16987717, Cur Avg Loss: 1.26099666, Log Avg loss: 1.22980032, Global Avg Loss: 1.60934986, Time: 0.0402 Steps: 63650, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001919, Sample Num: 30704, Cur Loss: 1.25903738, Cur Avg Loss: 1.25915111, Log Avg loss: 0.90683584, Global Avg Loss: 1.60923951, Time: 0.0402 Steps: 63660, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001929, Sample Num: 30864, Cur Loss: 1.47380865, Cur Avg Loss: 1.25962071, Log Avg loss: 1.34973607, Global Avg Loss: 1.60919875, Time: 0.0402 Steps: 63670, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001939, Sample Num: 31024, Cur Loss: 1.17130387, Cur Avg Loss: 1.26044227, Log Avg loss: 1.41892145, Global Avg Loss: 1.60916887, Time: 0.0403 Steps: 63680, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001949, Sample Num: 31184, Cur Loss: 2.94581580, Cur Avg Loss: 1.26182136, Log Avg loss: 1.52922682, Global Avg Loss: 1.60915632, Time: 0.0402 Steps: 63690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001959, Sample Num: 31344, Cur Loss: 2.21598387, Cur Avg Loss: 1.26337808, Log Avg loss: 1.56678370, Global Avg Loss: 1.60914967, Time: 0.0402 Steps: 63700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001969, Sample Num: 31504, Cur Loss: 1.83801055, Cur Avg Loss: 1.26350278, Log Avg loss: 1.28793034, Global Avg Loss: 1.60909925, Time: 0.0402 Steps: 63710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001979, Sample Num: 31664, Cur Loss: 1.97755921, Cur Avg Loss: 1.26468514, Log Avg loss: 1.49749191, Global Avg Loss: 1.60908173, Time: 0.0402 Steps: 63720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001989, Sample Num: 31824, Cur Loss: 0.54397207, Cur Avg Loss: 1.26418973, Log Avg loss: 1.16614949, Global Avg Loss: 1.60901223, Time: 0.0402 Steps: 63730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001999, Sample Num: 31984, Cur Loss: 0.74601871, Cur Avg Loss: 1.26253913, Log Avg loss: 0.93423461, Global Avg Loss: 1.60890637, Time: 0.0402 Steps: 63740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 002009, Sample Num: 32144, Cur Loss: 0.87163067, Cur Avg Loss: 1.26241837, Log Avg loss: 1.23827884, Global Avg Loss: 1.60884823, Time: 0.0402 Steps: 63750, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002019, Sample Num: 32304, Cur Loss: 0.72340375, Cur Avg Loss: 1.26145888, Log Avg loss: 1.06869590, Global Avg Loss: 1.60876351, Time: 0.0402 Steps: 63760, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002029, Sample Num: 32464, Cur Loss: 1.23705983, Cur Avg Loss: 1.26175287, Log Avg loss: 1.32111000, Global Avg Loss: 1.60871840, Time: 0.0402 Steps: 63770, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002039, Sample Num: 32624, Cur Loss: 1.73246622, Cur Avg Loss: 1.26069491, Log Avg loss: 1.04603560, Global Avg Loss: 1.60863018, Time: 0.0402 Steps: 63780, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002049, Sample Num: 32784, Cur Loss: 2.00156164, Cur Avg Loss: 1.26109022, Log Avg loss: 1.34169406, Global Avg Loss: 1.60858834, Time: 0.0403 Steps: 63790, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002059, Sample Num: 32944, Cur Loss: 1.96585977, Cur Avg Loss: 1.26298331, Log Avg loss: 1.65087754, Global Avg Loss: 1.60859496, Time: 0.0402 Steps: 63800, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002069, Sample Num: 33104, Cur Loss: 1.46911800, Cur Avg Loss: 1.26227744, Log Avg loss: 1.11693834, Global Avg Loss: 1.60851791, Time: 0.0402 Steps: 63810, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002079, Sample Num: 33264, Cur Loss: 0.66224408, Cur Avg Loss: 1.26021425, Log Avg loss: 0.83333897, Global Avg Loss: 1.60839645, Time: 0.0402 Steps: 63820, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002089, Sample Num: 33424, Cur Loss: 1.31166840, Cur Avg Loss: 1.25969669, Log Avg loss: 1.15209702, Global Avg Loss: 1.60832496, Time: 0.0402 Steps: 63830, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002099, Sample Num: 33584, Cur Loss: 0.58992314, Cur Avg Loss: 1.26057787, Log Avg loss: 1.44465653, Global Avg Loss: 1.60829933, Time: 0.0402 Steps: 63840, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002109, Sample Num: 33744, Cur Loss: 0.80434197, Cur Avg Loss: 1.26024944, Log Avg loss: 1.19131079, Global Avg Loss: 1.60823402, Time: 0.0402 Steps: 63850, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002119, Sample Num: 33904, Cur Loss: 1.98444343, Cur Avg Loss: 1.25990845, Log Avg loss: 1.18799489, Global Avg Loss: 1.60816821, Time: 0.0402 Steps: 63860, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002129, Sample Num: 34055, Cur Loss: 0.37003163, Cur Avg Loss: 1.26049567, Log Avg loss: 1.38492825, Global Avg Loss: 1.60813326, Time: 0.0198 Steps: 63870, Updated lr: 0.000040 ***** Running evaluation checkpoint-63870 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-63870 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.954621, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.201572, "eval_total_loss": 844.705108, "eval_mae": 0.902044, "eval_mse": 1.201674, "eval_r2": 0.236137, "eval_sp_statistic": 0.390467, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.502269, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.186171, "test_total_loss": 595.457936, "test_mae": 0.897653, "test_mse": 1.186393, "test_r2": 0.234292, "test_sp_statistic": 0.418687, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.540006, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6081332601428369, "train_cur_epoch_loss": 2683.595291659236, "train_cur_epoch_avg_loss": 1.260495674804714, "train_cur_epoch_time": 95.95462131500244, "train_cur_epoch_avg_time": 0.04507027774307301, "epoch": 30, "step": 63870} ################################################## Training, Epoch: 0031, Batch: 000010, Sample Num: 160, Cur Loss: 1.25549412, Cur Avg Loss: 1.69254811, Log Avg loss: 1.69254811, Global Avg Loss: 1.60814647, Time: 0.0402 Steps: 63880, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000020, Sample Num: 320, Cur Loss: 1.28878379, Cur Avg Loss: 1.52793061, Log Avg loss: 1.36331311, Global Avg Loss: 1.60810815, Time: 0.0402 Steps: 63890, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000030, Sample Num: 480, Cur Loss: 1.50413966, Cur Avg Loss: 1.55609531, Log Avg loss: 1.61242471, Global Avg Loss: 1.60810883, Time: 0.0402 Steps: 63900, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000040, Sample Num: 640, Cur Loss: 2.17245507, Cur Avg Loss: 1.43912516, Log Avg loss: 1.08821471, Global Avg Loss: 1.60802748, Time: 0.0402 Steps: 63910, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000050, Sample Num: 800, Cur Loss: 0.84293532, Cur Avg Loss: 1.38326735, Log Avg loss: 1.15983609, Global Avg Loss: 1.60795736, Time: 0.0402 Steps: 63920, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000060, Sample Num: 960, Cur Loss: 0.72687948, Cur Avg Loss: 1.33536640, Log Avg loss: 1.09586165, Global Avg Loss: 1.60787726, Time: 0.0402 Steps: 63930, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000070, Sample Num: 1120, Cur Loss: 0.87479526, Cur Avg Loss: 1.33936421, Log Avg loss: 1.36335106, Global Avg Loss: 1.60783902, Time: 0.0402 Steps: 63940, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000080, Sample Num: 1280, Cur Loss: 0.66980296, Cur Avg Loss: 1.34154965, Log Avg loss: 1.35684773, Global Avg Loss: 1.60779977, Time: 0.0406 Steps: 63950, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000090, Sample Num: 1440, Cur Loss: 1.12556052, Cur Avg Loss: 1.33345080, Log Avg loss: 1.26865999, Global Avg Loss: 1.60774675, Time: 0.0779 Steps: 63960, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000100, Sample Num: 1600, Cur Loss: 1.13710380, Cur Avg Loss: 1.33466643, Log Avg loss: 1.34560713, Global Avg Loss: 1.60770577, Time: 0.0404 Steps: 63970, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000110, Sample Num: 1760, Cur Loss: 0.85295391, Cur Avg Loss: 1.30990151, Log Avg loss: 1.06225237, Global Avg Loss: 1.60762051, Time: 0.0782 Steps: 63980, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000120, Sample Num: 1920, Cur Loss: 0.97653776, Cur Avg Loss: 1.32684531, Log Avg loss: 1.51322706, Global Avg Loss: 1.60760576, Time: 0.0457 Steps: 63990, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000130, Sample Num: 2080, Cur Loss: 3.23343492, Cur Avg Loss: 1.34385703, Log Avg loss: 1.54799762, Global Avg Loss: 1.60759645, Time: 0.0429 Steps: 64000, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000140, Sample Num: 2240, Cur Loss: 0.95240974, Cur Avg Loss: 1.32039263, Log Avg loss: 1.01535542, Global Avg Loss: 1.60750393, Time: 0.0411 Steps: 64010, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000150, Sample Num: 2400, Cur Loss: 0.89327973, Cur Avg Loss: 1.30511774, Log Avg loss: 1.09126927, Global Avg Loss: 1.60742329, Time: 0.0404 Steps: 64020, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000160, Sample Num: 2560, Cur Loss: 1.67459726, Cur Avg Loss: 1.29381503, Log Avg loss: 1.12427448, Global Avg Loss: 1.60734783, Time: 0.0460 Steps: 64030, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000170, Sample Num: 2720, Cur Loss: 2.88997340, Cur Avg Loss: 1.28812767, Log Avg loss: 1.19712987, Global Avg Loss: 1.60728378, Time: 0.0402 Steps: 64040, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000180, Sample Num: 2880, Cur Loss: 2.12944460, Cur Avg Loss: 1.30263040, Log Avg loss: 1.54917673, Global Avg Loss: 1.60727470, Time: 0.0402 Steps: 64050, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000190, Sample Num: 3040, Cur Loss: 2.60182047, Cur Avg Loss: 1.28748595, Log Avg loss: 1.01488592, Global Avg Loss: 1.60718223, Time: 0.0402 Steps: 64060, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000200, Sample Num: 3200, Cur Loss: 1.91513634, Cur Avg Loss: 1.29771774, Log Avg loss: 1.49212177, Global Avg Loss: 1.60716427, Time: 0.0403 Steps: 64070, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000210, Sample Num: 3360, Cur Loss: 0.86063331, Cur Avg Loss: 1.29845302, Log Avg loss: 1.31315860, Global Avg Loss: 1.60711839, Time: 0.0402 Steps: 64080, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000220, Sample Num: 3520, Cur Loss: 2.24337053, Cur Avg Loss: 1.29326716, Log Avg loss: 1.18436406, Global Avg Loss: 1.60705243, Time: 0.0402 Steps: 64090, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000230, Sample Num: 3680, Cur Loss: 1.10772848, Cur Avg Loss: 1.29608448, Log Avg loss: 1.35806559, Global Avg Loss: 1.60701358, Time: 0.0402 Steps: 64100, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000240, Sample Num: 3840, Cur Loss: 1.57856655, Cur Avg Loss: 1.28510804, Log Avg loss: 1.03264996, Global Avg Loss: 1.60692399, Time: 0.0402 Steps: 64110, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000250, Sample Num: 4000, Cur Loss: 0.48261762, Cur Avg Loss: 1.27404722, Log Avg loss: 1.00858757, Global Avg Loss: 1.60683068, Time: 0.0403 Steps: 64120, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000260, Sample Num: 4160, Cur Loss: 0.88968575, Cur Avg Loss: 1.27282901, Log Avg loss: 1.24237359, Global Avg Loss: 1.60677385, Time: 0.0402 Steps: 64130, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000270, Sample Num: 4320, Cur Loss: 1.66657221, Cur Avg Loss: 1.27826204, Log Avg loss: 1.41952088, Global Avg Loss: 1.60674465, Time: 0.0403 Steps: 64140, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000280, Sample Num: 4480, Cur Loss: 1.69777775, Cur Avg Loss: 1.27360653, Log Avg loss: 1.14790791, Global Avg Loss: 1.60667313, Time: 0.0403 Steps: 64150, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000290, Sample Num: 4640, Cur Loss: 0.62825990, Cur Avg Loss: 1.27673343, Log Avg loss: 1.36428657, Global Avg Loss: 1.60663535, Time: 0.0402 Steps: 64160, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000300, Sample Num: 4800, Cur Loss: 1.47425699, Cur Avg Loss: 1.27572522, Log Avg loss: 1.24648697, Global Avg Loss: 1.60657923, Time: 0.0402 Steps: 64170, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000310, Sample Num: 4960, Cur Loss: 1.40343463, Cur Avg Loss: 1.27082025, Log Avg loss: 1.12367137, Global Avg Loss: 1.60650398, Time: 0.0402 Steps: 64180, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000320, Sample Num: 5120, Cur Loss: 0.74452078, Cur Avg Loss: 1.26279755, Log Avg loss: 1.01409364, Global Avg Loss: 1.60641169, Time: 0.0402 Steps: 64190, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000330, Sample Num: 5280, Cur Loss: 2.49270296, Cur Avg Loss: 1.27802189, Log Avg loss: 1.76520084, Global Avg Loss: 1.60643643, Time: 0.0402 Steps: 64200, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000340, Sample Num: 5440, Cur Loss: 1.60669208, Cur Avg Loss: 1.28145010, Log Avg loss: 1.39458107, Global Avg Loss: 1.60640343, Time: 0.0402 Steps: 64210, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000350, Sample Num: 5600, Cur Loss: 3.03832483, Cur Avg Loss: 1.28503767, Log Avg loss: 1.40701494, Global Avg Loss: 1.60637238, Time: 0.0402 Steps: 64220, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000360, Sample Num: 5760, Cur Loss: 1.46994758, Cur Avg Loss: 1.28981363, Log Avg loss: 1.45697243, Global Avg Loss: 1.60634912, Time: 0.0402 Steps: 64230, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000370, Sample Num: 5920, Cur Loss: 0.72651124, Cur Avg Loss: 1.28349451, Log Avg loss: 1.05600593, Global Avg Loss: 1.60626345, Time: 0.0402 Steps: 64240, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000380, Sample Num: 6080, Cur Loss: 1.47414947, Cur Avg Loss: 1.27936054, Log Avg loss: 1.12640381, Global Avg Loss: 1.60618877, Time: 0.0402 Steps: 64250, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000390, Sample Num: 6240, Cur Loss: 1.08282304, Cur Avg Loss: 1.28188001, Log Avg loss: 1.37762004, Global Avg Loss: 1.60615320, Time: 0.0402 Steps: 64260, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000400, Sample Num: 6400, Cur Loss: 1.97169125, Cur Avg Loss: 1.28783498, Log Avg loss: 1.52007874, Global Avg Loss: 1.60613981, Time: 0.0402 Steps: 64270, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000410, Sample Num: 6560, Cur Loss: 1.33896172, Cur Avg Loss: 1.29301373, Log Avg loss: 1.50016371, Global Avg Loss: 1.60612332, Time: 0.0402 Steps: 64280, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000420, Sample Num: 6720, Cur Loss: 1.24105525, Cur Avg Loss: 1.29009567, Log Avg loss: 1.17045506, Global Avg Loss: 1.60605555, Time: 0.0402 Steps: 64290, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000430, Sample Num: 6880, Cur Loss: 0.88711488, Cur Avg Loss: 1.28739210, Log Avg loss: 1.17384219, Global Avg Loss: 1.60598833, Time: 0.0435 Steps: 64300, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000440, Sample Num: 7040, Cur Loss: 0.93812865, Cur Avg Loss: 1.28688482, Log Avg loss: 1.26507190, Global Avg Loss: 1.60593532, Time: 0.0403 Steps: 64310, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000450, Sample Num: 7200, Cur Loss: 1.95182490, Cur Avg Loss: 1.28419540, Log Avg loss: 1.16586066, Global Avg Loss: 1.60586690, Time: 0.0590 Steps: 64320, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000460, Sample Num: 7360, Cur Loss: 1.26821375, Cur Avg Loss: 1.28326252, Log Avg loss: 1.24128290, Global Avg Loss: 1.60581023, Time: 0.0503 Steps: 64330, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000470, Sample Num: 7520, Cur Loss: 1.86090875, Cur Avg Loss: 1.28223449, Log Avg loss: 1.23494540, Global Avg Loss: 1.60575259, Time: 0.0609 Steps: 64340, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000480, Sample Num: 7680, Cur Loss: 0.61068058, Cur Avg Loss: 1.27971615, Log Avg loss: 1.16135394, Global Avg Loss: 1.60568353, Time: 0.0726 Steps: 64350, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000490, Sample Num: 7840, Cur Loss: 0.99251860, Cur Avg Loss: 1.28452536, Log Avg loss: 1.51536773, Global Avg Loss: 1.60566950, Time: 0.0480 Steps: 64360, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000500, Sample Num: 8000, Cur Loss: 0.76083183, Cur Avg Loss: 1.29286582, Log Avg loss: 1.70154806, Global Avg Loss: 1.60568439, Time: 0.0846 Steps: 64370, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000510, Sample Num: 8160, Cur Loss: 0.80503988, Cur Avg Loss: 1.28829222, Log Avg loss: 1.05961217, Global Avg Loss: 1.60559957, Time: 0.0426 Steps: 64380, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000520, Sample Num: 8320, Cur Loss: 1.71669066, Cur Avg Loss: 1.29111328, Log Avg loss: 1.43498739, Global Avg Loss: 1.60557307, Time: 0.0402 Steps: 64390, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000530, Sample Num: 8480, Cur Loss: 0.68865585, Cur Avg Loss: 1.28686540, Log Avg loss: 1.06597585, Global Avg Loss: 1.60548929, Time: 0.0402 Steps: 64400, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000540, Sample Num: 8640, Cur Loss: 1.59463310, Cur Avg Loss: 1.28535428, Log Avg loss: 1.20526488, Global Avg Loss: 1.60542715, Time: 0.0402 Steps: 64410, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000550, Sample Num: 8800, Cur Loss: 1.80486441, Cur Avg Loss: 1.28460246, Log Avg loss: 1.24400424, Global Avg Loss: 1.60537104, Time: 0.0402 Steps: 64420, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000560, Sample Num: 8960, Cur Loss: 0.49676406, Cur Avg Loss: 1.27769341, Log Avg loss: 0.89769576, Global Avg Loss: 1.60526121, Time: 0.0403 Steps: 64430, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000570, Sample Num: 9120, Cur Loss: 0.75766057, Cur Avg Loss: 1.27552563, Log Avg loss: 1.15412996, Global Avg Loss: 1.60519120, Time: 0.0403 Steps: 64440, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000580, Sample Num: 9280, Cur Loss: 1.40876555, Cur Avg Loss: 1.27268751, Log Avg loss: 1.11091452, Global Avg Loss: 1.60511451, Time: 0.0402 Steps: 64450, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000590, Sample Num: 9440, Cur Loss: 2.78578472, Cur Avg Loss: 1.27008633, Log Avg loss: 1.11921796, Global Avg Loss: 1.60503913, Time: 0.0401 Steps: 64460, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000600, Sample Num: 9600, Cur Loss: 2.42816067, Cur Avg Loss: 1.26898824, Log Avg loss: 1.20420063, Global Avg Loss: 1.60497695, Time: 0.0403 Steps: 64470, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000610, Sample Num: 9760, Cur Loss: 1.72859895, Cur Avg Loss: 1.27016250, Log Avg loss: 1.34061832, Global Avg Loss: 1.60493596, Time: 0.0402 Steps: 64480, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000620, Sample Num: 9920, Cur Loss: 1.46425962, Cur Avg Loss: 1.27002134, Log Avg loss: 1.26141033, Global Avg Loss: 1.60488269, Time: 0.0402 Steps: 64490, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000630, Sample Num: 10080, Cur Loss: 1.18654907, Cur Avg Loss: 1.26368143, Log Avg loss: 0.87060691, Global Avg Loss: 1.60476885, Time: 0.0402 Steps: 64500, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000640, Sample Num: 10240, Cur Loss: 0.81322497, Cur Avg Loss: 1.26028421, Log Avg loss: 1.04625933, Global Avg Loss: 1.60468227, Time: 0.0402 Steps: 64510, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000650, Sample Num: 10400, Cur Loss: 0.33054408, Cur Avg Loss: 1.25649724, Log Avg loss: 1.01413126, Global Avg Loss: 1.60459074, Time: 0.0403 Steps: 64520, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000660, Sample Num: 10560, Cur Loss: 0.70340312, Cur Avg Loss: 1.26111276, Log Avg loss: 1.56112189, Global Avg Loss: 1.60458400, Time: 0.0402 Steps: 64530, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000670, Sample Num: 10720, Cur Loss: 0.29123014, Cur Avg Loss: 1.25905183, Log Avg loss: 1.12303061, Global Avg Loss: 1.60450939, Time: 0.0402 Steps: 64540, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000680, Sample Num: 10880, Cur Loss: 0.76609439, Cur Avg Loss: 1.26221054, Log Avg loss: 1.47384366, Global Avg Loss: 1.60448915, Time: 0.0402 Steps: 64550, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000690, Sample Num: 11040, Cur Loss: 1.34058785, Cur Avg Loss: 1.25960842, Log Avg loss: 1.08266439, Global Avg Loss: 1.60440832, Time: 0.0402 Steps: 64560, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000700, Sample Num: 11200, Cur Loss: 1.29438353, Cur Avg Loss: 1.25829766, Log Avg loss: 1.16785533, Global Avg Loss: 1.60434071, Time: 0.0402 Steps: 64570, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000710, Sample Num: 11360, Cur Loss: 2.03537059, Cur Avg Loss: 1.25821099, Log Avg loss: 1.25214434, Global Avg Loss: 1.60428617, Time: 0.0402 Steps: 64580, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000720, Sample Num: 11520, Cur Loss: 0.40747333, Cur Avg Loss: 1.25586455, Log Avg loss: 1.08926663, Global Avg Loss: 1.60420644, Time: 0.0402 Steps: 64590, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000730, Sample Num: 11680, Cur Loss: 0.84395242, Cur Avg Loss: 1.25083263, Log Avg loss: 0.88853447, Global Avg Loss: 1.60409565, Time: 0.0402 Steps: 64600, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000740, Sample Num: 11840, Cur Loss: 2.52215767, Cur Avg Loss: 1.25359578, Log Avg loss: 1.45530625, Global Avg Loss: 1.60407262, Time: 0.0402 Steps: 64610, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000750, Sample Num: 12000, Cur Loss: 2.73621345, Cur Avg Loss: 1.25583632, Log Avg loss: 1.42163619, Global Avg Loss: 1.60404439, Time: 0.0402 Steps: 64620, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000760, Sample Num: 12160, Cur Loss: 0.62799907, Cur Avg Loss: 1.25822520, Log Avg loss: 1.43739116, Global Avg Loss: 1.60401861, Time: 0.0402 Steps: 64630, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000770, Sample Num: 12320, Cur Loss: 2.04966331, Cur Avg Loss: 1.25693302, Log Avg loss: 1.15872692, Global Avg Loss: 1.60394972, Time: 0.0402 Steps: 64640, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000780, Sample Num: 12480, Cur Loss: 0.87975252, Cur Avg Loss: 1.25469122, Log Avg loss: 1.08207256, Global Avg Loss: 1.60386899, Time: 0.0647 Steps: 64650, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000790, Sample Num: 12640, Cur Loss: 0.52691227, Cur Avg Loss: 1.24980501, Log Avg loss: 0.86868075, Global Avg Loss: 1.60375529, Time: 0.0404 Steps: 64660, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000800, Sample Num: 12800, Cur Loss: 1.10549021, Cur Avg Loss: 1.25176707, Log Avg loss: 1.40677022, Global Avg Loss: 1.60372483, Time: 0.0587 Steps: 64670, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000810, Sample Num: 12960, Cur Loss: 1.20520580, Cur Avg Loss: 1.25375478, Log Avg loss: 1.41277171, Global Avg Loss: 1.60369531, Time: 0.0404 Steps: 64680, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000820, Sample Num: 13120, Cur Loss: 1.45176888, Cur Avg Loss: 1.25423233, Log Avg loss: 1.29291398, Global Avg Loss: 1.60364727, Time: 0.0682 Steps: 64690, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000830, Sample Num: 13280, Cur Loss: 1.15437138, Cur Avg Loss: 1.25357198, Log Avg loss: 1.19942299, Global Avg Loss: 1.60358479, Time: 0.0503 Steps: 64700, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000840, Sample Num: 13440, Cur Loss: 1.46111131, Cur Avg Loss: 1.25670507, Log Avg loss: 1.51675124, Global Avg Loss: 1.60357137, Time: 0.0457 Steps: 64710, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000850, Sample Num: 13600, Cur Loss: 2.50323343, Cur Avg Loss: 1.25659963, Log Avg loss: 1.24774298, Global Avg Loss: 1.60351639, Time: 0.0609 Steps: 64720, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000860, Sample Num: 13760, Cur Loss: 1.48825407, Cur Avg Loss: 1.25679256, Log Avg loss: 1.27319112, Global Avg Loss: 1.60346536, Time: 0.0402 Steps: 64730, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000870, Sample Num: 13920, Cur Loss: 1.02950382, Cur Avg Loss: 1.25743661, Log Avg loss: 1.31282544, Global Avg Loss: 1.60342047, Time: 0.0402 Steps: 64740, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000880, Sample Num: 14080, Cur Loss: 0.62593102, Cur Avg Loss: 1.25962233, Log Avg loss: 1.44978004, Global Avg Loss: 1.60339674, Time: 0.0402 Steps: 64750, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000890, Sample Num: 14240, Cur Loss: 0.50909662, Cur Avg Loss: 1.26542954, Log Avg loss: 1.77646406, Global Avg Loss: 1.60342347, Time: 0.0402 Steps: 64760, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000900, Sample Num: 14400, Cur Loss: 0.38211814, Cur Avg Loss: 1.26585822, Log Avg loss: 1.30401028, Global Avg Loss: 1.60337724, Time: 0.0402 Steps: 64770, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000910, Sample Num: 14560, Cur Loss: 1.63565946, Cur Avg Loss: 1.26360767, Log Avg loss: 1.06105839, Global Avg Loss: 1.60329352, Time: 0.0402 Steps: 64780, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000920, Sample Num: 14720, Cur Loss: 1.68221200, Cur Avg Loss: 1.26364571, Log Avg loss: 1.26710746, Global Avg Loss: 1.60324163, Time: 0.0403 Steps: 64790, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000930, Sample Num: 14880, Cur Loss: 1.56648278, Cur Avg Loss: 1.26437484, Log Avg loss: 1.33145494, Global Avg Loss: 1.60319969, Time: 0.0403 Steps: 64800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000940, Sample Num: 15040, Cur Loss: 0.44247195, Cur Avg Loss: 1.26205395, Log Avg loss: 1.04621086, Global Avg Loss: 1.60311375, Time: 0.0402 Steps: 64810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000950, Sample Num: 15200, Cur Loss: 2.35936284, Cur Avg Loss: 1.26546276, Log Avg loss: 1.58589057, Global Avg Loss: 1.60311109, Time: 0.0402 Steps: 64820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000960, Sample Num: 15360, Cur Loss: 1.35618174, Cur Avg Loss: 1.26550293, Log Avg loss: 1.26931915, Global Avg Loss: 1.60305960, Time: 0.0402 Steps: 64830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000970, Sample Num: 15520, Cur Loss: 1.50395942, Cur Avg Loss: 1.26443730, Log Avg loss: 1.16213705, Global Avg Loss: 1.60299160, Time: 0.0402 Steps: 64840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000980, Sample Num: 15680, Cur Loss: 0.41637787, Cur Avg Loss: 1.26777716, Log Avg loss: 1.59174348, Global Avg Loss: 1.60298987, Time: 0.0402 Steps: 64850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000990, Sample Num: 15840, Cur Loss: 1.22125089, Cur Avg Loss: 1.26723386, Log Avg loss: 1.21399075, Global Avg Loss: 1.60292989, Time: 0.0402 Steps: 64860, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001000, Sample Num: 16000, Cur Loss: 2.32454252, Cur Avg Loss: 1.26554061, Log Avg loss: 1.09790878, Global Avg Loss: 1.60285204, Time: 0.0403 Steps: 64870, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001010, Sample Num: 16160, Cur Loss: 0.94856071, Cur Avg Loss: 1.26496373, Log Avg loss: 1.20727603, Global Avg Loss: 1.60279107, Time: 0.0402 Steps: 64880, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001020, Sample Num: 16320, Cur Loss: 1.76616633, Cur Avg Loss: 1.26747316, Log Avg loss: 1.52092559, Global Avg Loss: 1.60277846, Time: 0.0402 Steps: 64890, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001030, Sample Num: 16480, Cur Loss: 2.02973270, Cur Avg Loss: 1.26650597, Log Avg loss: 1.16785189, Global Avg Loss: 1.60271144, Time: 0.0402 Steps: 64900, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001040, Sample Num: 16640, Cur Loss: 2.00606847, Cur Avg Loss: 1.26845875, Log Avg loss: 1.46959533, Global Avg Loss: 1.60269093, Time: 0.0402 Steps: 64910, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001050, Sample Num: 16800, Cur Loss: 3.41595650, Cur Avg Loss: 1.26980135, Log Avg loss: 1.40943209, Global Avg Loss: 1.60266116, Time: 0.0402 Steps: 64920, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001060, Sample Num: 16960, Cur Loss: 1.24806809, Cur Avg Loss: 1.26804346, Log Avg loss: 1.08346449, Global Avg Loss: 1.60258120, Time: 0.0402 Steps: 64930, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001070, Sample Num: 17120, Cur Loss: 0.93207425, Cur Avg Loss: 1.26720439, Log Avg loss: 1.17826304, Global Avg Loss: 1.60251586, Time: 0.0402 Steps: 64940, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001080, Sample Num: 17280, Cur Loss: 0.64256132, Cur Avg Loss: 1.26524223, Log Avg loss: 1.05529149, Global Avg Loss: 1.60243161, Time: 0.0402 Steps: 64950, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001090, Sample Num: 17440, Cur Loss: 1.08542848, Cur Avg Loss: 1.26290478, Log Avg loss: 1.01045969, Global Avg Loss: 1.60234048, Time: 0.0402 Steps: 64960, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001100, Sample Num: 17600, Cur Loss: 1.61333990, Cur Avg Loss: 1.26193261, Log Avg loss: 1.15596589, Global Avg Loss: 1.60227177, Time: 0.0402 Steps: 64970, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001110, Sample Num: 17760, Cur Loss: 2.39590025, Cur Avg Loss: 1.26277979, Log Avg loss: 1.35596959, Global Avg Loss: 1.60223387, Time: 0.0402 Steps: 64980, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001120, Sample Num: 17920, Cur Loss: 0.47056723, Cur Avg Loss: 1.26269407, Log Avg loss: 1.25317966, Global Avg Loss: 1.60218016, Time: 0.0402 Steps: 64990, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001130, Sample Num: 18080, Cur Loss: 0.55805254, Cur Avg Loss: 1.26492948, Log Avg loss: 1.51529507, Global Avg Loss: 1.60216679, Time: 0.0404 Steps: 65000, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001140, Sample Num: 18240, Cur Loss: 2.73344326, Cur Avg Loss: 1.26566531, Log Avg loss: 1.34881455, Global Avg Loss: 1.60212782, Time: 0.0404 Steps: 65010, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001150, Sample Num: 18400, Cur Loss: 0.97069055, Cur Avg Loss: 1.26391684, Log Avg loss: 1.06459101, Global Avg Loss: 1.60204515, Time: 0.0917 Steps: 65020, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001160, Sample Num: 18560, Cur Loss: 2.82298851, Cur Avg Loss: 1.26458707, Log Avg loss: 1.34166363, Global Avg Loss: 1.60200511, Time: 0.0593 Steps: 65030, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001170, Sample Num: 18720, Cur Loss: 0.38487059, Cur Avg Loss: 1.26068611, Log Avg loss: 0.80817478, Global Avg Loss: 1.60188306, Time: 0.0404 Steps: 65040, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001180, Sample Num: 18880, Cur Loss: 0.79905891, Cur Avg Loss: 1.26085454, Log Avg loss: 1.28056082, Global Avg Loss: 1.60183366, Time: 0.0995 Steps: 65050, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001190, Sample Num: 19040, Cur Loss: 1.15067542, Cur Avg Loss: 1.26030033, Log Avg loss: 1.19490388, Global Avg Loss: 1.60177111, Time: 0.0870 Steps: 65060, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001200, Sample Num: 19200, Cur Loss: 1.10939503, Cur Avg Loss: 1.25958549, Log Avg loss: 1.17451980, Global Avg Loss: 1.60170545, Time: 0.0988 Steps: 65070, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001210, Sample Num: 19360, Cur Loss: 0.29923576, Cur Avg Loss: 1.25929758, Log Avg loss: 1.22474748, Global Avg Loss: 1.60164753, Time: 0.0404 Steps: 65080, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001220, Sample Num: 19520, Cur Loss: 0.69803452, Cur Avg Loss: 1.25945188, Log Avg loss: 1.27812299, Global Avg Loss: 1.60159783, Time: 0.0687 Steps: 65090, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001230, Sample Num: 19680, Cur Loss: 0.88923955, Cur Avg Loss: 1.25944713, Log Avg loss: 1.25886738, Global Avg Loss: 1.60154518, Time: 0.0402 Steps: 65100, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001240, Sample Num: 19840, Cur Loss: 2.19262576, Cur Avg Loss: 1.26068169, Log Avg loss: 1.41253180, Global Avg Loss: 1.60151615, Time: 0.0402 Steps: 65110, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001250, Sample Num: 20000, Cur Loss: 0.43736196, Cur Avg Loss: 1.25914469, Log Avg loss: 1.06855699, Global Avg Loss: 1.60143431, Time: 0.0403 Steps: 65120, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001260, Sample Num: 20160, Cur Loss: 2.21860838, Cur Avg Loss: 1.26224437, Log Avg loss: 1.64970429, Global Avg Loss: 1.60144172, Time: 0.0402 Steps: 65130, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001270, Sample Num: 20320, Cur Loss: 0.83356649, Cur Avg Loss: 1.26135741, Log Avg loss: 1.14960055, Global Avg Loss: 1.60137236, Time: 0.0402 Steps: 65140, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001280, Sample Num: 20480, Cur Loss: 2.70893621, Cur Avg Loss: 1.26483045, Log Avg loss: 1.70590739, Global Avg Loss: 1.60138840, Time: 0.0403 Steps: 65150, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001290, Sample Num: 20640, Cur Loss: 1.28200686, Cur Avg Loss: 1.26649434, Log Avg loss: 1.47947170, Global Avg Loss: 1.60136969, Time: 0.0403 Steps: 65160, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001300, Sample Num: 20800, Cur Loss: 1.41407800, Cur Avg Loss: 1.26494198, Log Avg loss: 1.06468777, Global Avg Loss: 1.60128734, Time: 0.0402 Steps: 65170, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001310, Sample Num: 20960, Cur Loss: 1.15225494, Cur Avg Loss: 1.26617420, Log Avg loss: 1.42636279, Global Avg Loss: 1.60126050, Time: 0.0403 Steps: 65180, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001320, Sample Num: 21120, Cur Loss: 0.96809685, Cur Avg Loss: 1.26339207, Log Avg loss: 0.89893302, Global Avg Loss: 1.60115277, Time: 0.0402 Steps: 65190, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001330, Sample Num: 21280, Cur Loss: 1.61827016, Cur Avg Loss: 1.26382017, Log Avg loss: 1.32032914, Global Avg Loss: 1.60110970, Time: 0.0402 Steps: 65200, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001340, Sample Num: 21440, Cur Loss: 2.30559540, Cur Avg Loss: 1.26328154, Log Avg loss: 1.19164430, Global Avg Loss: 1.60104690, Time: 0.0402 Steps: 65210, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001350, Sample Num: 21600, Cur Loss: 1.62939858, Cur Avg Loss: 1.26246181, Log Avg loss: 1.15261694, Global Avg Loss: 1.60097815, Time: 0.0402 Steps: 65220, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001360, Sample Num: 21760, Cur Loss: 0.92534590, Cur Avg Loss: 1.26003815, Log Avg loss: 0.93284494, Global Avg Loss: 1.60087572, Time: 0.0403 Steps: 65230, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001370, Sample Num: 21920, Cur Loss: 2.43720555, Cur Avg Loss: 1.26023206, Log Avg loss: 1.28660405, Global Avg Loss: 1.60082755, Time: 0.0402 Steps: 65240, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001380, Sample Num: 22080, Cur Loss: 1.52162647, Cur Avg Loss: 1.26254257, Log Avg loss: 1.57908193, Global Avg Loss: 1.60082422, Time: 0.0403 Steps: 65250, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001390, Sample Num: 22240, Cur Loss: 0.82116151, Cur Avg Loss: 1.26420516, Log Avg loss: 1.49364299, Global Avg Loss: 1.60080779, Time: 0.0402 Steps: 65260, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001400, Sample Num: 22400, Cur Loss: 1.36559391, Cur Avg Loss: 1.26656024, Log Avg loss: 1.59391637, Global Avg Loss: 1.60080674, Time: 0.0402 Steps: 65270, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001410, Sample Num: 22560, Cur Loss: 2.32279491, Cur Avg Loss: 1.26637584, Log Avg loss: 1.24055877, Global Avg Loss: 1.60075155, Time: 0.0402 Steps: 65280, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001420, Sample Num: 22720, Cur Loss: 0.56004226, Cur Avg Loss: 1.26606081, Log Avg loss: 1.22164217, Global Avg Loss: 1.60069349, Time: 0.0402 Steps: 65290, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001430, Sample Num: 22880, Cur Loss: 0.68119633, Cur Avg Loss: 1.26453258, Log Avg loss: 1.04752377, Global Avg Loss: 1.60060877, Time: 0.0402 Steps: 65300, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001440, Sample Num: 23040, Cur Loss: 1.84487331, Cur Avg Loss: 1.26556206, Log Avg loss: 1.41277787, Global Avg Loss: 1.60058001, Time: 0.0403 Steps: 65310, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001450, Sample Num: 23200, Cur Loss: 1.45732224, Cur Avg Loss: 1.26702085, Log Avg loss: 1.47708675, Global Avg Loss: 1.60056111, Time: 0.0402 Steps: 65320, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001460, Sample Num: 23360, Cur Loss: 1.77508378, Cur Avg Loss: 1.27081219, Log Avg loss: 1.82055628, Global Avg Loss: 1.60059478, Time: 0.0402 Steps: 65330, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001470, Sample Num: 23520, Cur Loss: 1.87066817, Cur Avg Loss: 1.26933589, Log Avg loss: 1.05379660, Global Avg Loss: 1.60051110, Time: 0.0402 Steps: 65340, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001480, Sample Num: 23680, Cur Loss: 1.21381795, Cur Avg Loss: 1.26892582, Log Avg loss: 1.20864505, Global Avg Loss: 1.60045113, Time: 0.0486 Steps: 65350, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001490, Sample Num: 23840, Cur Loss: 1.30420840, Cur Avg Loss: 1.26878740, Log Avg loss: 1.24830186, Global Avg Loss: 1.60039725, Time: 0.0587 Steps: 65360, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001500, Sample Num: 24000, Cur Loss: 1.44360685, Cur Avg Loss: 1.26738945, Log Avg loss: 1.05909455, Global Avg Loss: 1.60031445, Time: 0.0441 Steps: 65370, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001510, Sample Num: 24160, Cur Loss: 1.30328727, Cur Avg Loss: 1.26671440, Log Avg loss: 1.16545667, Global Avg Loss: 1.60024794, Time: 0.0405 Steps: 65380, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001520, Sample Num: 24320, Cur Loss: 1.49210560, Cur Avg Loss: 1.26674942, Log Avg loss: 1.27203810, Global Avg Loss: 1.60019774, Time: 0.0814 Steps: 65390, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001530, Sample Num: 24480, Cur Loss: 0.67227751, Cur Avg Loss: 1.26746750, Log Avg loss: 1.37661470, Global Avg Loss: 1.60016356, Time: 0.0730 Steps: 65400, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001540, Sample Num: 24640, Cur Loss: 1.55392158, Cur Avg Loss: 1.26716429, Log Avg loss: 1.22077371, Global Avg Loss: 1.60010555, Time: 0.0673 Steps: 65410, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001550, Sample Num: 24800, Cur Loss: 1.52363777, Cur Avg Loss: 1.26725979, Log Avg loss: 1.28196639, Global Avg Loss: 1.60005692, Time: 0.0519 Steps: 65420, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001560, Sample Num: 24960, Cur Loss: 0.70992422, Cur Avg Loss: 1.26668554, Log Avg loss: 1.17767618, Global Avg Loss: 1.59999237, Time: 0.0507 Steps: 65430, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001570, Sample Num: 25120, Cur Loss: 1.24287748, Cur Avg Loss: 1.26607801, Log Avg loss: 1.17130475, Global Avg Loss: 1.59992686, Time: 0.0402 Steps: 65440, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001580, Sample Num: 25280, Cur Loss: 1.15032828, Cur Avg Loss: 1.26526829, Log Avg loss: 1.13814169, Global Avg Loss: 1.59985631, Time: 0.0402 Steps: 65450, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001590, Sample Num: 25440, Cur Loss: 1.34568286, Cur Avg Loss: 1.26577967, Log Avg loss: 1.34657744, Global Avg Loss: 1.59981761, Time: 0.0403 Steps: 65460, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001600, Sample Num: 25600, Cur Loss: 1.36473536, Cur Avg Loss: 1.26454003, Log Avg loss: 1.06743736, Global Avg Loss: 1.59973630, Time: 0.0402 Steps: 65470, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001610, Sample Num: 25760, Cur Loss: 0.56362927, Cur Avg Loss: 1.26476786, Log Avg loss: 1.30122060, Global Avg Loss: 1.59969071, Time: 0.0403 Steps: 65480, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001620, Sample Num: 25920, Cur Loss: 1.57414007, Cur Avg Loss: 1.26439558, Log Avg loss: 1.20445845, Global Avg Loss: 1.59963036, Time: 0.0402 Steps: 65490, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001630, Sample Num: 26080, Cur Loss: 0.89038962, Cur Avg Loss: 1.26425329, Log Avg loss: 1.24120307, Global Avg Loss: 1.59957564, Time: 0.0403 Steps: 65500, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001640, Sample Num: 26240, Cur Loss: 1.61079967, Cur Avg Loss: 1.26410690, Log Avg loss: 1.24024393, Global Avg Loss: 1.59952079, Time: 0.0402 Steps: 65510, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001650, Sample Num: 26400, Cur Loss: 1.66385913, Cur Avg Loss: 1.26508306, Log Avg loss: 1.42517362, Global Avg Loss: 1.59949418, Time: 0.0403 Steps: 65520, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001660, Sample Num: 26560, Cur Loss: 1.45681298, Cur Avg Loss: 1.26578373, Log Avg loss: 1.38139433, Global Avg Loss: 1.59946089, Time: 0.0403 Steps: 65530, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001670, Sample Num: 26720, Cur Loss: 1.69340611, Cur Avg Loss: 1.26622009, Log Avg loss: 1.33865692, Global Avg Loss: 1.59942110, Time: 0.0402 Steps: 65540, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001680, Sample Num: 26880, Cur Loss: 1.40042198, Cur Avg Loss: 1.26542123, Log Avg loss: 1.13201136, Global Avg Loss: 1.59934979, Time: 0.0403 Steps: 65550, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001690, Sample Num: 27040, Cur Loss: 0.56588268, Cur Avg Loss: 1.26444897, Log Avg loss: 1.10110913, Global Avg Loss: 1.59927380, Time: 0.0403 Steps: 65560, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001700, Sample Num: 27200, Cur Loss: 1.82467175, Cur Avg Loss: 1.26549484, Log Avg loss: 1.44224608, Global Avg Loss: 1.59924985, Time: 0.0402 Steps: 65570, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001710, Sample Num: 27360, Cur Loss: 1.79106259, Cur Avg Loss: 1.26639946, Log Avg loss: 1.42018555, Global Avg Loss: 1.59922254, Time: 0.0402 Steps: 65580, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001720, Sample Num: 27520, Cur Loss: 0.55320591, Cur Avg Loss: 1.26598474, Log Avg loss: 1.19506681, Global Avg Loss: 1.59916093, Time: 0.0402 Steps: 65590, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001730, Sample Num: 27680, Cur Loss: 0.68990546, Cur Avg Loss: 1.26691286, Log Avg loss: 1.42654938, Global Avg Loss: 1.59913461, Time: 0.0402 Steps: 65600, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001740, Sample Num: 27840, Cur Loss: 1.44973946, Cur Avg Loss: 1.26703918, Log Avg loss: 1.28889273, Global Avg Loss: 1.59908733, Time: 0.0403 Steps: 65610, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001750, Sample Num: 28000, Cur Loss: 0.79289782, Cur Avg Loss: 1.26536702, Log Avg loss: 0.97441181, Global Avg Loss: 1.59899213, Time: 0.0402 Steps: 65620, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001760, Sample Num: 28160, Cur Loss: 0.93076116, Cur Avg Loss: 1.26722026, Log Avg loss: 1.59153696, Global Avg Loss: 1.59899099, Time: 0.0402 Steps: 65630, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001770, Sample Num: 28320, Cur Loss: 0.78261483, Cur Avg Loss: 1.26740600, Log Avg loss: 1.30009610, Global Avg Loss: 1.59894546, Time: 0.0402 Steps: 65640, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001780, Sample Num: 28480, Cur Loss: 0.82437378, Cur Avg Loss: 1.26687117, Log Avg loss: 1.17220607, Global Avg Loss: 1.59888046, Time: 0.0402 Steps: 65650, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001790, Sample Num: 28640, Cur Loss: 0.73027098, Cur Avg Loss: 1.26669700, Log Avg loss: 1.23569585, Global Avg Loss: 1.59882514, Time: 0.0402 Steps: 65660, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001800, Sample Num: 28800, Cur Loss: 1.13156462, Cur Avg Loss: 1.26695367, Log Avg loss: 1.31289786, Global Avg Loss: 1.59878160, Time: 0.0402 Steps: 65670, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001810, Sample Num: 28960, Cur Loss: 1.74030972, Cur Avg Loss: 1.26560750, Log Avg loss: 1.02329539, Global Avg Loss: 1.59869398, Time: 0.0402 Steps: 65680, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001820, Sample Num: 29120, Cur Loss: 2.05750036, Cur Avg Loss: 1.26746036, Log Avg loss: 1.60282851, Global Avg Loss: 1.59869461, Time: 0.0756 Steps: 65690, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001830, Sample Num: 29280, Cur Loss: 3.60923958, Cur Avg Loss: 1.26759608, Log Avg loss: 1.29229745, Global Avg Loss: 1.59864798, Time: 0.0807 Steps: 65700, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001840, Sample Num: 29440, Cur Loss: 0.86968261, Cur Avg Loss: 1.26835083, Log Avg loss: 1.40647085, Global Avg Loss: 1.59861873, Time: 0.0532 Steps: 65710, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001850, Sample Num: 29600, Cur Loss: 0.65350419, Cur Avg Loss: 1.26733924, Log Avg loss: 1.08120595, Global Avg Loss: 1.59854000, Time: 0.0530 Steps: 65720, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001860, Sample Num: 29760, Cur Loss: 0.96459085, Cur Avg Loss: 1.26721249, Log Avg loss: 1.24376273, Global Avg Loss: 1.59848603, Time: 0.0491 Steps: 65730, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001870, Sample Num: 29920, Cur Loss: 1.05581772, Cur Avg Loss: 1.26729436, Log Avg loss: 1.28252275, Global Avg Loss: 1.59843796, Time: 0.0566 Steps: 65740, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001880, Sample Num: 30080, Cur Loss: 1.48068070, Cur Avg Loss: 1.26787908, Log Avg loss: 1.37722122, Global Avg Loss: 1.59840432, Time: 0.0405 Steps: 65750, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001890, Sample Num: 30240, Cur Loss: 2.15946198, Cur Avg Loss: 1.26817937, Log Avg loss: 1.32463520, Global Avg Loss: 1.59836269, Time: 0.0586 Steps: 65760, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001900, Sample Num: 30400, Cur Loss: 2.02284980, Cur Avg Loss: 1.26902834, Log Avg loss: 1.42948376, Global Avg Loss: 1.59833701, Time: 0.0461 Steps: 65770, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001910, Sample Num: 30560, Cur Loss: 0.56619281, Cur Avg Loss: 1.26912128, Log Avg loss: 1.28677896, Global Avg Loss: 1.59828965, Time: 0.0403 Steps: 65780, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001920, Sample Num: 30720, Cur Loss: 1.45756221, Cur Avg Loss: 1.26922938, Log Avg loss: 1.28987586, Global Avg Loss: 1.59824277, Time: 0.0402 Steps: 65790, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001930, Sample Num: 30880, Cur Loss: 1.40307117, Cur Avg Loss: 1.26809436, Log Avg loss: 1.05017087, Global Avg Loss: 1.59815947, Time: 0.0403 Steps: 65800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001940, Sample Num: 31040, Cur Loss: 1.74100733, Cur Avg Loss: 1.26894392, Log Avg loss: 1.43291034, Global Avg Loss: 1.59813436, Time: 0.0402 Steps: 65810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001950, Sample Num: 31200, Cur Loss: 1.59071755, Cur Avg Loss: 1.26925581, Log Avg loss: 1.32976252, Global Avg Loss: 1.59809359, Time: 0.0403 Steps: 65820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001960, Sample Num: 31360, Cur Loss: 1.15288150, Cur Avg Loss: 1.27108775, Log Avg loss: 1.62831554, Global Avg Loss: 1.59809818, Time: 0.0403 Steps: 65830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001970, Sample Num: 31520, Cur Loss: 1.96685326, Cur Avg Loss: 1.26982714, Log Avg loss: 1.02274659, Global Avg Loss: 1.59801080, Time: 0.0403 Steps: 65840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001980, Sample Num: 31680, Cur Loss: 1.17619228, Cur Avg Loss: 1.27054287, Log Avg loss: 1.41154242, Global Avg Loss: 1.59798248, Time: 0.0403 Steps: 65850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001990, Sample Num: 31840, Cur Loss: 0.76689184, Cur Avg Loss: 1.27096907, Log Avg loss: 1.35535710, Global Avg Loss: 1.59794564, Time: 0.0404 Steps: 65860, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002000, Sample Num: 32000, Cur Loss: 0.66666198, Cur Avg Loss: 1.27018480, Log Avg loss: 1.11411455, Global Avg Loss: 1.59787219, Time: 0.0403 Steps: 65870, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002010, Sample Num: 32160, Cur Loss: 1.54117560, Cur Avg Loss: 1.27115072, Log Avg loss: 1.46433510, Global Avg Loss: 1.59785192, Time: 0.0403 Steps: 65880, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002020, Sample Num: 32320, Cur Loss: 1.89412665, Cur Avg Loss: 1.27087179, Log Avg loss: 1.21480691, Global Avg Loss: 1.59779378, Time: 0.0402 Steps: 65890, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002030, Sample Num: 32480, Cur Loss: 1.25377476, Cur Avg Loss: 1.27081564, Log Avg loss: 1.25947300, Global Avg Loss: 1.59774244, Time: 0.0402 Steps: 65900, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002040, Sample Num: 32640, Cur Loss: 0.51791918, Cur Avg Loss: 1.26946021, Log Avg loss: 0.99430695, Global Avg Loss: 1.59765089, Time: 0.0402 Steps: 65910, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002050, Sample Num: 32800, Cur Loss: 1.09334302, Cur Avg Loss: 1.26814413, Log Avg loss: 0.99966551, Global Avg Loss: 1.59756018, Time: 0.0403 Steps: 65920, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002060, Sample Num: 32960, Cur Loss: 0.64127785, Cur Avg Loss: 1.26736599, Log Avg loss: 1.10784601, Global Avg Loss: 1.59748590, Time: 0.0403 Steps: 65930, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002070, Sample Num: 33120, Cur Loss: 1.16729617, Cur Avg Loss: 1.26869832, Log Avg loss: 1.54315854, Global Avg Loss: 1.59747766, Time: 0.0402 Steps: 65940, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002080, Sample Num: 33280, Cur Loss: 0.55408418, Cur Avg Loss: 1.26863156, Log Avg loss: 1.25481256, Global Avg Loss: 1.59742570, Time: 0.0402 Steps: 65950, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002090, Sample Num: 33440, Cur Loss: 1.19742370, Cur Avg Loss: 1.27036906, Log Avg loss: 1.63176942, Global Avg Loss: 1.59743091, Time: 0.0402 Steps: 65960, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002100, Sample Num: 33600, Cur Loss: 1.59605789, Cur Avg Loss: 1.27044091, Log Avg loss: 1.28545649, Global Avg Loss: 1.59738362, Time: 0.0402 Steps: 65970, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002110, Sample Num: 33760, Cur Loss: 0.65404189, Cur Avg Loss: 1.26974743, Log Avg loss: 1.12411813, Global Avg Loss: 1.59731189, Time: 0.0402 Steps: 65980, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002120, Sample Num: 33920, Cur Loss: 0.37408450, Cur Avg Loss: 1.26882587, Log Avg loss: 1.07437585, Global Avg Loss: 1.59723264, Time: 0.0402 Steps: 65990, Updated lr: 0.000038 ***** Running evaluation checkpoint-65999 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-65999 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.206909, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.179662, "eval_total_loss": 829.302557, "eval_mae": 0.87048, "eval_mse": 1.179811, "eval_r2": 0.250034, "eval_sp_statistic": 0.401768, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.506315, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.126757, "test_total_loss": 565.631786, "test_mae": 0.832941, "test_mse": 1.127007, "test_r2": 0.27262, "test_sp_statistic": 0.428351, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.544919, "test_ps_pvalue": 0.0, "lr": 3.8360360360360366e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.59714702663968, "train_cur_epoch_loss": 2698.6352858692408, "train_cur_epoch_avg_loss": 1.2675600215449698, "train_cur_epoch_time": 95.20690894126892, "train_cur_epoch_avg_time": 0.04471907418565943, "epoch": 31, "step": 65999} ################################################## Training, Epoch: 0032, Batch: 000001, Sample Num: 16, Cur Loss: 2.57340264, Cur Avg Loss: 2.57340264, Log Avg loss: 1.12978441, Global Avg Loss: 1.59716182, Time: 0.0406 Steps: 66000, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000011, Sample Num: 176, Cur Loss: 1.15092850, Cur Avg Loss: 1.13882461, Log Avg loss: 0.99536680, Global Avg Loss: 1.59707065, Time: 0.0402 Steps: 66010, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000021, Sample Num: 336, Cur Loss: 1.51436675, Cur Avg Loss: 1.12796040, Log Avg loss: 1.11600977, Global Avg Loss: 1.59699779, Time: 0.0403 Steps: 66020, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000031, Sample Num: 496, Cur Loss: 0.63622129, Cur Avg Loss: 1.18289882, Log Avg loss: 1.29826950, Global Avg Loss: 1.59695254, Time: 0.0402 Steps: 66030, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000041, Sample Num: 656, Cur Loss: 0.88434768, Cur Avg Loss: 1.19706538, Log Avg loss: 1.24098171, Global Avg Loss: 1.59689864, Time: 0.0402 Steps: 66040, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000051, Sample Num: 816, Cur Loss: 1.17022204, Cur Avg Loss: 1.12394339, Log Avg loss: 0.82414322, Global Avg Loss: 1.59678165, Time: 0.0402 Steps: 66050, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000061, Sample Num: 976, Cur Loss: 0.93885690, Cur Avg Loss: 1.20014136, Log Avg loss: 1.58875101, Global Avg Loss: 1.59678043, Time: 0.0402 Steps: 66060, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000071, Sample Num: 1136, Cur Loss: 1.09573436, Cur Avg Loss: 1.18135188, Log Avg loss: 1.06673608, Global Avg Loss: 1.59670021, Time: 0.0402 Steps: 66070, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000081, Sample Num: 1296, Cur Loss: 0.81895095, Cur Avg Loss: 1.20400207, Log Avg loss: 1.36481839, Global Avg Loss: 1.59666511, Time: 0.0402 Steps: 66080, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000091, Sample Num: 1456, Cur Loss: 1.32061052, Cur Avg Loss: 1.21403924, Log Avg loss: 1.29534035, Global Avg Loss: 1.59661952, Time: 0.0585 Steps: 66090, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000101, Sample Num: 1616, Cur Loss: 0.83454353, Cur Avg Loss: 1.23784439, Log Avg loss: 1.45447129, Global Avg Loss: 1.59659802, Time: 0.0534 Steps: 66100, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000111, Sample Num: 1776, Cur Loss: 0.68212700, Cur Avg Loss: 1.26577244, Log Avg loss: 1.54784570, Global Avg Loss: 1.59659064, Time: 0.0557 Steps: 66110, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000121, Sample Num: 1936, Cur Loss: 0.24154785, Cur Avg Loss: 1.26510300, Log Avg loss: 1.25767224, Global Avg Loss: 1.59653938, Time: 0.0405 Steps: 66120, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000131, Sample Num: 2096, Cur Loss: 0.45586362, Cur Avg Loss: 1.24818773, Log Avg loss: 1.04351288, Global Avg Loss: 1.59645576, Time: 0.0500 Steps: 66130, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000141, Sample Num: 2256, Cur Loss: 1.07128215, Cur Avg Loss: 1.23730216, Log Avg loss: 1.09470128, Global Avg Loss: 1.59637989, Time: 0.0404 Steps: 66140, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000151, Sample Num: 2416, Cur Loss: 0.46295971, Cur Avg Loss: 1.23382007, Log Avg loss: 1.18472255, Global Avg Loss: 1.59631766, Time: 0.0795 Steps: 66150, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000161, Sample Num: 2576, Cur Loss: 2.41147065, Cur Avg Loss: 1.23964760, Log Avg loss: 1.32764339, Global Avg Loss: 1.59627705, Time: 0.0547 Steps: 66160, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000171, Sample Num: 2736, Cur Loss: 0.84714156, Cur Avg Loss: 1.23537664, Log Avg loss: 1.16661417, Global Avg Loss: 1.59621212, Time: 0.0404 Steps: 66170, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000181, Sample Num: 2896, Cur Loss: 1.98383355, Cur Avg Loss: 1.24207384, Log Avg loss: 1.35659586, Global Avg Loss: 1.59617591, Time: 0.0403 Steps: 66180, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000191, Sample Num: 3056, Cur Loss: 0.73937267, Cur Avg Loss: 1.24978054, Log Avg loss: 1.38927189, Global Avg Loss: 1.59614465, Time: 0.0403 Steps: 66190, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000201, Sample Num: 3216, Cur Loss: 1.37097061, Cur Avg Loss: 1.25019204, Log Avg loss: 1.25805170, Global Avg Loss: 1.59609358, Time: 0.0403 Steps: 66200, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000211, Sample Num: 3376, Cur Loss: 1.58782935, Cur Avg Loss: 1.25443149, Log Avg loss: 1.33964450, Global Avg Loss: 1.59605485, Time: 0.0402 Steps: 66210, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000221, Sample Num: 3536, Cur Loss: 0.61843914, Cur Avg Loss: 1.26787263, Log Avg loss: 1.55148064, Global Avg Loss: 1.59604812, Time: 0.0402 Steps: 66220, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000231, Sample Num: 3696, Cur Loss: 0.77369308, Cur Avg Loss: 1.25671186, Log Avg loss: 1.01005886, Global Avg Loss: 1.59595964, Time: 0.0403 Steps: 66230, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000241, Sample Num: 3856, Cur Loss: 0.80785036, Cur Avg Loss: 1.26327888, Log Avg loss: 1.41497709, Global Avg Loss: 1.59593232, Time: 0.0404 Steps: 66240, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000251, Sample Num: 4016, Cur Loss: 0.78791916, Cur Avg Loss: 1.25817452, Log Avg loss: 1.13515925, Global Avg Loss: 1.59586277, Time: 0.0402 Steps: 66250, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000261, Sample Num: 4176, Cur Loss: 0.32686803, Cur Avg Loss: 1.25719597, Log Avg loss: 1.23263452, Global Avg Loss: 1.59580795, Time: 0.0403 Steps: 66260, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000271, Sample Num: 4336, Cur Loss: 0.75156689, Cur Avg Loss: 1.25495489, Log Avg loss: 1.19646270, Global Avg Loss: 1.59574769, Time: 0.0402 Steps: 66270, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000281, Sample Num: 4496, Cur Loss: 1.26296675, Cur Avg Loss: 1.24857737, Log Avg loss: 1.07574656, Global Avg Loss: 1.59566923, Time: 0.0403 Steps: 66280, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000291, Sample Num: 4656, Cur Loss: 1.48013580, Cur Avg Loss: 1.24007806, Log Avg loss: 1.00124730, Global Avg Loss: 1.59557956, Time: 0.0402 Steps: 66290, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000301, Sample Num: 4816, Cur Loss: 1.10157084, Cur Avg Loss: 1.24572639, Log Avg loss: 1.41009291, Global Avg Loss: 1.59555159, Time: 0.0402 Steps: 66300, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000311, Sample Num: 4976, Cur Loss: 1.52570045, Cur Avg Loss: 1.24312944, Log Avg loss: 1.16496137, Global Avg Loss: 1.59548665, Time: 0.0403 Steps: 66310, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000321, Sample Num: 5136, Cur Loss: 0.52085459, Cur Avg Loss: 1.24342018, Log Avg loss: 1.25246204, Global Avg Loss: 1.59543493, Time: 0.0403 Steps: 66320, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000331, Sample Num: 5296, Cur Loss: 0.52191830, Cur Avg Loss: 1.23849034, Log Avg loss: 1.08024244, Global Avg Loss: 1.59535726, Time: 0.0402 Steps: 66330, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000341, Sample Num: 5456, Cur Loss: 0.83081567, Cur Avg Loss: 1.23739214, Log Avg loss: 1.20104182, Global Avg Loss: 1.59529782, Time: 0.0403 Steps: 66340, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000351, Sample Num: 5616, Cur Loss: 0.82636648, Cur Avg Loss: 1.23851680, Log Avg loss: 1.27686749, Global Avg Loss: 1.59524983, Time: 0.0402 Steps: 66350, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000361, Sample Num: 5776, Cur Loss: 0.85688138, Cur Avg Loss: 1.25162319, Log Avg loss: 1.71165773, Global Avg Loss: 1.59526737, Time: 0.0402 Steps: 66360, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000371, Sample Num: 5936, Cur Loss: 0.70424610, Cur Avg Loss: 1.25943603, Log Avg loss: 1.54147955, Global Avg Loss: 1.59525926, Time: 0.0402 Steps: 66370, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000381, Sample Num: 6096, Cur Loss: 1.72629809, Cur Avg Loss: 1.27794152, Log Avg loss: 1.96449500, Global Avg Loss: 1.59531489, Time: 0.0402 Steps: 66380, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000391, Sample Num: 6256, Cur Loss: 2.31336141, Cur Avg Loss: 1.28495541, Log Avg loss: 1.55218469, Global Avg Loss: 1.59530839, Time: 0.0403 Steps: 66390, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000401, Sample Num: 6416, Cur Loss: 1.20085585, Cur Avg Loss: 1.28527803, Log Avg loss: 1.29789264, Global Avg Loss: 1.59526360, Time: 0.0402 Steps: 66400, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000411, Sample Num: 6576, Cur Loss: 1.15721941, Cur Avg Loss: 1.27781061, Log Avg loss: 0.97836679, Global Avg Loss: 1.59517071, Time: 0.0402 Steps: 66410, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000421, Sample Num: 6736, Cur Loss: 2.32231617, Cur Avg Loss: 1.27209178, Log Avg loss: 1.03704804, Global Avg Loss: 1.59508668, Time: 0.0402 Steps: 66420, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000431, Sample Num: 6896, Cur Loss: 1.76063204, Cur Avg Loss: 1.27419628, Log Avg loss: 1.36279561, Global Avg Loss: 1.59505171, Time: 0.0483 Steps: 66430, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000441, Sample Num: 7056, Cur Loss: 0.49250013, Cur Avg Loss: 1.26646966, Log Avg loss: 0.93345256, Global Avg Loss: 1.59495213, Time: 0.0640 Steps: 66440, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000451, Sample Num: 7216, Cur Loss: 2.24946952, Cur Avg Loss: 1.27585478, Log Avg loss: 1.68973849, Global Avg Loss: 1.59496640, Time: 0.0482 Steps: 66450, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000461, Sample Num: 7376, Cur Loss: 0.51662493, Cur Avg Loss: 1.27767166, Log Avg loss: 1.35961288, Global Avg Loss: 1.59493098, Time: 0.0504 Steps: 66460, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000471, Sample Num: 7536, Cur Loss: 1.91718459, Cur Avg Loss: 1.27866123, Log Avg loss: 1.32428020, Global Avg Loss: 1.59489027, Time: 0.0484 Steps: 66470, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000481, Sample Num: 7696, Cur Loss: 1.55941558, Cur Avg Loss: 1.27118926, Log Avg loss: 0.91925949, Global Avg Loss: 1.59478864, Time: 0.0560 Steps: 66480, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000491, Sample Num: 7856, Cur Loss: 1.04725313, Cur Avg Loss: 1.27807869, Log Avg loss: 1.60946048, Global Avg Loss: 1.59479084, Time: 0.0452 Steps: 66490, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000501, Sample Num: 8016, Cur Loss: 1.47993279, Cur Avg Loss: 1.27295213, Log Avg loss: 1.02123825, Global Avg Loss: 1.59470460, Time: 0.0481 Steps: 66500, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000511, Sample Num: 8176, Cur Loss: 1.34243488, Cur Avg Loss: 1.26802443, Log Avg loss: 1.02114630, Global Avg Loss: 1.59461836, Time: 0.0402 Steps: 66510, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000521, Sample Num: 8336, Cur Loss: 1.39053309, Cur Avg Loss: 1.26849939, Log Avg loss: 1.29277003, Global Avg Loss: 1.59457298, Time: 0.0402 Steps: 66520, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000531, Sample Num: 8496, Cur Loss: 0.52299738, Cur Avg Loss: 1.26703020, Log Avg loss: 1.19048512, Global Avg Loss: 1.59451224, Time: 0.0402 Steps: 66530, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000541, Sample Num: 8656, Cur Loss: 0.70223486, Cur Avg Loss: 1.26280036, Log Avg loss: 1.03819587, Global Avg Loss: 1.59442864, Time: 0.0403 Steps: 66540, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000551, Sample Num: 8816, Cur Loss: 2.01310229, Cur Avg Loss: 1.26469420, Log Avg loss: 1.36715119, Global Avg Loss: 1.59439449, Time: 0.0403 Steps: 66550, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000561, Sample Num: 8976, Cur Loss: 2.24691176, Cur Avg Loss: 1.26496836, Log Avg loss: 1.28007481, Global Avg Loss: 1.59434726, Time: 0.0403 Steps: 66560, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000571, Sample Num: 9136, Cur Loss: 0.33819771, Cur Avg Loss: 1.25929518, Log Avg loss: 0.94102936, Global Avg Loss: 1.59424912, Time: 0.0403 Steps: 66570, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000581, Sample Num: 9296, Cur Loss: 1.19345725, Cur Avg Loss: 1.25851587, Log Avg loss: 1.21401739, Global Avg Loss: 1.59419201, Time: 0.0403 Steps: 66580, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000591, Sample Num: 9456, Cur Loss: 0.81657255, Cur Avg Loss: 1.25836524, Log Avg loss: 1.24961349, Global Avg Loss: 1.59414027, Time: 0.0402 Steps: 66590, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000601, Sample Num: 9616, Cur Loss: 0.44310993, Cur Avg Loss: 1.25369101, Log Avg loss: 0.97744408, Global Avg Loss: 1.59404767, Time: 0.0403 Steps: 66600, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000611, Sample Num: 9776, Cur Loss: 1.95061970, Cur Avg Loss: 1.25340836, Log Avg loss: 1.23642136, Global Avg Loss: 1.59399398, Time: 0.0402 Steps: 66610, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000621, Sample Num: 9936, Cur Loss: 0.42830348, Cur Avg Loss: 1.24682639, Log Avg loss: 0.84466784, Global Avg Loss: 1.59388150, Time: 0.0402 Steps: 66620, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000631, Sample Num: 10096, Cur Loss: 0.51200652, Cur Avg Loss: 1.24178904, Log Avg loss: 0.92896981, Global Avg Loss: 1.59378171, Time: 0.0403 Steps: 66630, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000641, Sample Num: 10256, Cur Loss: 2.00860548, Cur Avg Loss: 1.24271190, Log Avg loss: 1.30094442, Global Avg Loss: 1.59373777, Time: 0.0402 Steps: 66640, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000651, Sample Num: 10416, Cur Loss: 1.43854785, Cur Avg Loss: 1.24091434, Log Avg loss: 1.12569041, Global Avg Loss: 1.59366754, Time: 0.0402 Steps: 66650, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000661, Sample Num: 10576, Cur Loss: 0.80188644, Cur Avg Loss: 1.23400727, Log Avg loss: 0.78435700, Global Avg Loss: 1.59354614, Time: 0.0403 Steps: 66660, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000671, Sample Num: 10736, Cur Loss: 0.57846743, Cur Avg Loss: 1.23576983, Log Avg loss: 1.35227500, Global Avg Loss: 1.59350995, Time: 0.0402 Steps: 66670, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000681, Sample Num: 10896, Cur Loss: 1.20193219, Cur Avg Loss: 1.23370648, Log Avg loss: 1.09525597, Global Avg Loss: 1.59343522, Time: 0.0402 Steps: 66680, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000691, Sample Num: 11056, Cur Loss: 2.51166940, Cur Avg Loss: 1.23708244, Log Avg loss: 1.46698545, Global Avg Loss: 1.59341626, Time: 0.0402 Steps: 66690, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000701, Sample Num: 11216, Cur Loss: 2.51719236, Cur Avg Loss: 1.23750039, Log Avg loss: 1.26638052, Global Avg Loss: 1.59336723, Time: 0.0402 Steps: 66700, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000711, Sample Num: 11376, Cur Loss: 0.33799440, Cur Avg Loss: 1.23744744, Log Avg loss: 1.23373597, Global Avg Loss: 1.59331332, Time: 0.0401 Steps: 66710, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000721, Sample Num: 11536, Cur Loss: 1.35983312, Cur Avg Loss: 1.23504823, Log Avg loss: 1.06446438, Global Avg Loss: 1.59323406, Time: 0.0402 Steps: 66720, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000731, Sample Num: 11696, Cur Loss: 1.48792768, Cur Avg Loss: 1.23159486, Log Avg loss: 0.98260627, Global Avg Loss: 1.59314255, Time: 0.0402 Steps: 66730, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000741, Sample Num: 11856, Cur Loss: 0.74394268, Cur Avg Loss: 1.23431169, Log Avg loss: 1.43291250, Global Avg Loss: 1.59311854, Time: 0.0402 Steps: 66740, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000751, Sample Num: 12016, Cur Loss: 1.16046703, Cur Avg Loss: 1.23430754, Log Avg loss: 1.23399997, Global Avg Loss: 1.59306474, Time: 0.0403 Steps: 66750, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000761, Sample Num: 12176, Cur Loss: 2.25091362, Cur Avg Loss: 1.23413045, Log Avg loss: 1.22083094, Global Avg Loss: 1.59300899, Time: 0.0402 Steps: 66760, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000771, Sample Num: 12336, Cur Loss: 0.86386615, Cur Avg Loss: 1.23450635, Log Avg loss: 1.26311253, Global Avg Loss: 1.59295958, Time: 0.0635 Steps: 66770, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000781, Sample Num: 12496, Cur Loss: 1.23216176, Cur Avg Loss: 1.23681871, Log Avg loss: 1.41510102, Global Avg Loss: 1.59293294, Time: 0.0993 Steps: 66780, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000791, Sample Num: 12656, Cur Loss: 0.52024591, Cur Avg Loss: 1.23794306, Log Avg loss: 1.32575483, Global Avg Loss: 1.59289294, Time: 0.0637 Steps: 66790, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000801, Sample Num: 12816, Cur Loss: 1.30095553, Cur Avg Loss: 1.23682051, Log Avg loss: 1.14802713, Global Avg Loss: 1.59282634, Time: 0.0408 Steps: 66800, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000811, Sample Num: 12976, Cur Loss: 0.50099039, Cur Avg Loss: 1.23525347, Log Avg loss: 1.10973375, Global Avg Loss: 1.59275404, Time: 0.0500 Steps: 66810, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000821, Sample Num: 13136, Cur Loss: 1.27963161, Cur Avg Loss: 1.23477786, Log Avg loss: 1.19620595, Global Avg Loss: 1.59269469, Time: 0.0960 Steps: 66820, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000831, Sample Num: 13296, Cur Loss: 1.78115261, Cur Avg Loss: 1.24088788, Log Avg loss: 1.74252034, Global Avg Loss: 1.59271711, Time: 0.0404 Steps: 66830, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000841, Sample Num: 13456, Cur Loss: 1.76962364, Cur Avg Loss: 1.24121971, Log Avg loss: 1.26879460, Global Avg Loss: 1.59266865, Time: 0.0612 Steps: 66840, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000851, Sample Num: 13616, Cur Loss: 1.28528810, Cur Avg Loss: 1.23942957, Log Avg loss: 1.08887864, Global Avg Loss: 1.59259329, Time: 0.0522 Steps: 66850, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000861, Sample Num: 13776, Cur Loss: 1.21348715, Cur Avg Loss: 1.23987988, Log Avg loss: 1.27820158, Global Avg Loss: 1.59254626, Time: 0.0402 Steps: 66860, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000871, Sample Num: 13936, Cur Loss: 1.15280271, Cur Avg Loss: 1.23865703, Log Avg loss: 1.13336984, Global Avg Loss: 1.59247760, Time: 0.0402 Steps: 66870, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000881, Sample Num: 14096, Cur Loss: 0.52119732, Cur Avg Loss: 1.23622052, Log Avg loss: 1.02400056, Global Avg Loss: 1.59239260, Time: 0.0403 Steps: 66880, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000891, Sample Num: 14256, Cur Loss: 1.85360575, Cur Avg Loss: 1.23732377, Log Avg loss: 1.33451931, Global Avg Loss: 1.59235405, Time: 0.0402 Steps: 66890, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000901, Sample Num: 14416, Cur Loss: 0.77347732, Cur Avg Loss: 1.23660965, Log Avg loss: 1.17298171, Global Avg Loss: 1.59229136, Time: 0.0402 Steps: 66900, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000911, Sample Num: 14576, Cur Loss: 1.96004999, Cur Avg Loss: 1.23456431, Log Avg loss: 1.05027929, Global Avg Loss: 1.59221035, Time: 0.0402 Steps: 66910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000921, Sample Num: 14736, Cur Loss: 0.65788907, Cur Avg Loss: 1.23435939, Log Avg loss: 1.21569156, Global Avg Loss: 1.59215409, Time: 0.0402 Steps: 66920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000931, Sample Num: 14896, Cur Loss: 0.68013030, Cur Avg Loss: 1.23458381, Log Avg loss: 1.25525261, Global Avg Loss: 1.59210375, Time: 0.0402 Steps: 66930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000941, Sample Num: 15056, Cur Loss: 1.50931096, Cur Avg Loss: 1.23437350, Log Avg loss: 1.21479349, Global Avg Loss: 1.59204739, Time: 0.0403 Steps: 66940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000951, Sample Num: 15216, Cur Loss: 1.27693176, Cur Avg Loss: 1.23472923, Log Avg loss: 1.26820349, Global Avg Loss: 1.59199902, Time: 0.0403 Steps: 66950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000961, Sample Num: 15376, Cur Loss: 0.50451589, Cur Avg Loss: 1.23632340, Log Avg loss: 1.38792880, Global Avg Loss: 1.59196854, Time: 0.0403 Steps: 66960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000971, Sample Num: 15536, Cur Loss: 2.33857346, Cur Avg Loss: 1.23980237, Log Avg loss: 1.57413160, Global Avg Loss: 1.59196588, Time: 0.0402 Steps: 66970, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000981, Sample Num: 15696, Cur Loss: 1.15260577, Cur Avg Loss: 1.24001147, Log Avg loss: 1.26031538, Global Avg Loss: 1.59191636, Time: 0.0403 Steps: 66980, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000991, Sample Num: 15856, Cur Loss: 1.07164907, Cur Avg Loss: 1.24146912, Log Avg loss: 1.38446479, Global Avg Loss: 1.59188539, Time: 0.0402 Steps: 66990, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001001, Sample Num: 16016, Cur Loss: 2.14140153, Cur Avg Loss: 1.24432184, Log Avg loss: 1.52702579, Global Avg Loss: 1.59187571, Time: 0.0403 Steps: 67000, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001011, Sample Num: 16176, Cur Loss: 0.86057222, Cur Avg Loss: 1.24598864, Log Avg loss: 1.41283557, Global Avg Loss: 1.59184899, Time: 0.0403 Steps: 67010, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001021, Sample Num: 16336, Cur Loss: 1.23220158, Cur Avg Loss: 1.24718255, Log Avg loss: 1.36788722, Global Avg Loss: 1.59181558, Time: 0.0402 Steps: 67020, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001031, Sample Num: 16496, Cur Loss: 0.93413115, Cur Avg Loss: 1.24718903, Log Avg loss: 1.24785032, Global Avg Loss: 1.59176426, Time: 0.0402 Steps: 67030, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001041, Sample Num: 16656, Cur Loss: 1.97591913, Cur Avg Loss: 1.25119470, Log Avg loss: 1.66417939, Global Avg Loss: 1.59177506, Time: 0.0402 Steps: 67040, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001051, Sample Num: 16816, Cur Loss: 0.96547627, Cur Avg Loss: 1.24913128, Log Avg loss: 1.03432884, Global Avg Loss: 1.59169193, Time: 0.0402 Steps: 67050, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001061, Sample Num: 16976, Cur Loss: 1.46887088, Cur Avg Loss: 1.24888504, Log Avg loss: 1.22300591, Global Avg Loss: 1.59163695, Time: 0.0403 Steps: 67060, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001071, Sample Num: 17136, Cur Loss: 0.62358510, Cur Avg Loss: 1.24912032, Log Avg loss: 1.27408276, Global Avg Loss: 1.59158960, Time: 0.0402 Steps: 67070, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001081, Sample Num: 17296, Cur Loss: 1.86733747, Cur Avg Loss: 1.25012387, Log Avg loss: 1.35760479, Global Avg Loss: 1.59155472, Time: 0.0402 Steps: 67080, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001091, Sample Num: 17456, Cur Loss: 0.70538908, Cur Avg Loss: 1.25369472, Log Avg loss: 1.63970332, Global Avg Loss: 1.59156190, Time: 0.0402 Steps: 67090, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001101, Sample Num: 17616, Cur Loss: 1.78532195, Cur Avg Loss: 1.25441484, Log Avg loss: 1.33297999, Global Avg Loss: 1.59152336, Time: 0.0403 Steps: 67100, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001111, Sample Num: 17776, Cur Loss: 1.92524767, Cur Avg Loss: 1.25315053, Log Avg loss: 1.11394986, Global Avg Loss: 1.59145220, Time: 0.0754 Steps: 67110, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001121, Sample Num: 17936, Cur Loss: 1.26451802, Cur Avg Loss: 1.25187483, Log Avg loss: 1.11014488, Global Avg Loss: 1.59138049, Time: 0.0545 Steps: 67120, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001131, Sample Num: 18096, Cur Loss: 1.83745313, Cur Avg Loss: 1.25158059, Log Avg loss: 1.21859571, Global Avg Loss: 1.59132496, Time: 0.0566 Steps: 67130, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001141, Sample Num: 18256, Cur Loss: 1.13247371, Cur Avg Loss: 1.25202977, Log Avg loss: 1.30283222, Global Avg Loss: 1.59128199, Time: 0.0647 Steps: 67140, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001151, Sample Num: 18416, Cur Loss: 1.33563852, Cur Avg Loss: 1.25116769, Log Avg loss: 1.15280433, Global Avg Loss: 1.59121669, Time: 0.0555 Steps: 67150, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001161, Sample Num: 18576, Cur Loss: 0.73176265, Cur Avg Loss: 1.25043150, Log Avg loss: 1.16569641, Global Avg Loss: 1.59115333, Time: 0.0794 Steps: 67160, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001171, Sample Num: 18736, Cur Loss: 1.31908929, Cur Avg Loss: 1.25402388, Log Avg loss: 1.67109845, Global Avg Loss: 1.59116523, Time: 0.0886 Steps: 67170, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001181, Sample Num: 18896, Cur Loss: 0.44404992, Cur Avg Loss: 1.25314991, Log Avg loss: 1.15080838, Global Avg Loss: 1.59109968, Time: 0.0432 Steps: 67180, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001191, Sample Num: 19056, Cur Loss: 0.55760229, Cur Avg Loss: 1.25184008, Log Avg loss: 1.09714888, Global Avg Loss: 1.59102617, Time: 0.0673 Steps: 67190, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001201, Sample Num: 19216, Cur Loss: 0.50007659, Cur Avg Loss: 1.25216930, Log Avg loss: 1.29137953, Global Avg Loss: 1.59098158, Time: 0.0403 Steps: 67200, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001211, Sample Num: 19376, Cur Loss: 1.31074178, Cur Avg Loss: 1.25046084, Log Avg loss: 1.04527452, Global Avg Loss: 1.59090038, Time: 0.0403 Steps: 67210, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001221, Sample Num: 19536, Cur Loss: 0.96226496, Cur Avg Loss: 1.24754152, Log Avg loss: 0.89401274, Global Avg Loss: 1.59079671, Time: 0.0402 Steps: 67220, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001231, Sample Num: 19696, Cur Loss: 1.92390263, Cur Avg Loss: 1.24943975, Log Avg loss: 1.48121287, Global Avg Loss: 1.59078041, Time: 0.0402 Steps: 67230, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001241, Sample Num: 19856, Cur Loss: 1.32786191, Cur Avg Loss: 1.24849158, Log Avg loss: 1.13177252, Global Avg Loss: 1.59071215, Time: 0.0403 Steps: 67240, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001251, Sample Num: 20016, Cur Loss: 0.99016631, Cur Avg Loss: 1.25129919, Log Avg loss: 1.59972334, Global Avg Loss: 1.59071349, Time: 0.0403 Steps: 67250, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001261, Sample Num: 20176, Cur Loss: 1.25259912, Cur Avg Loss: 1.25277788, Log Avg loss: 1.43776143, Global Avg Loss: 1.59069075, Time: 0.0403 Steps: 67260, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001271, Sample Num: 20336, Cur Loss: 0.79880381, Cur Avg Loss: 1.25209685, Log Avg loss: 1.16621948, Global Avg Loss: 1.59062765, Time: 0.0403 Steps: 67270, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001281, Sample Num: 20496, Cur Loss: 1.79677725, Cur Avg Loss: 1.25203375, Log Avg loss: 1.24401374, Global Avg Loss: 1.59057613, Time: 0.0403 Steps: 67280, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001291, Sample Num: 20656, Cur Loss: 1.49614942, Cur Avg Loss: 1.25226494, Log Avg loss: 1.28188047, Global Avg Loss: 1.59053025, Time: 0.0403 Steps: 67290, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001301, Sample Num: 20816, Cur Loss: 1.55921638, Cur Avg Loss: 1.25224775, Log Avg loss: 1.25002882, Global Avg Loss: 1.59047966, Time: 0.0402 Steps: 67300, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001311, Sample Num: 20976, Cur Loss: 0.80062509, Cur Avg Loss: 1.25144325, Log Avg loss: 1.14677685, Global Avg Loss: 1.59041374, Time: 0.0402 Steps: 67310, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001321, Sample Num: 21136, Cur Loss: 0.51413137, Cur Avg Loss: 1.25110013, Log Avg loss: 1.20611738, Global Avg Loss: 1.59035665, Time: 0.0403 Steps: 67320, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001331, Sample Num: 21296, Cur Loss: 1.26782858, Cur Avg Loss: 1.25162910, Log Avg loss: 1.32150604, Global Avg Loss: 1.59031672, Time: 0.0402 Steps: 67330, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001341, Sample Num: 21456, Cur Loss: 1.43918192, Cur Avg Loss: 1.25265744, Log Avg loss: 1.38952933, Global Avg Loss: 1.59028691, Time: 0.0403 Steps: 67340, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001351, Sample Num: 21616, Cur Loss: 1.92712176, Cur Avg Loss: 1.25510044, Log Avg loss: 1.58270721, Global Avg Loss: 1.59028578, Time: 0.0402 Steps: 67350, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001361, Sample Num: 21776, Cur Loss: 1.71487427, Cur Avg Loss: 1.25583258, Log Avg loss: 1.35474416, Global Avg Loss: 1.59025081, Time: 0.0402 Steps: 67360, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001371, Sample Num: 21936, Cur Loss: 1.04522061, Cur Avg Loss: 1.25769423, Log Avg loss: 1.51106583, Global Avg Loss: 1.59023906, Time: 0.0402 Steps: 67370, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001381, Sample Num: 22096, Cur Loss: 1.04554093, Cur Avg Loss: 1.25751434, Log Avg loss: 1.23285065, Global Avg Loss: 1.59018602, Time: 0.0402 Steps: 67380, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001391, Sample Num: 22256, Cur Loss: 1.18757105, Cur Avg Loss: 1.25780586, Log Avg loss: 1.29806439, Global Avg Loss: 1.59014267, Time: 0.0402 Steps: 67390, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001401, Sample Num: 22416, Cur Loss: 0.70581931, Cur Avg Loss: 1.26008677, Log Avg loss: 1.57736143, Global Avg Loss: 1.59014077, Time: 0.0402 Steps: 67400, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001411, Sample Num: 22576, Cur Loss: 1.33140540, Cur Avg Loss: 1.25852667, Log Avg loss: 1.03995691, Global Avg Loss: 1.59005916, Time: 0.0402 Steps: 67410, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001421, Sample Num: 22736, Cur Loss: 0.66657299, Cur Avg Loss: 1.25784589, Log Avg loss: 1.16178874, Global Avg Loss: 1.58999563, Time: 0.0402 Steps: 67420, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001431, Sample Num: 22896, Cur Loss: 2.04013753, Cur Avg Loss: 1.25750934, Log Avg loss: 1.20968462, Global Avg Loss: 1.58993923, Time: 0.0402 Steps: 67430, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001441, Sample Num: 23056, Cur Loss: 1.72977722, Cur Avg Loss: 1.25663045, Log Avg loss: 1.13086231, Global Avg Loss: 1.58987116, Time: 0.0402 Steps: 67440, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001451, Sample Num: 23216, Cur Loss: 1.92500532, Cur Avg Loss: 1.25752257, Log Avg loss: 1.38607700, Global Avg Loss: 1.58984095, Time: 0.0402 Steps: 67450, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001461, Sample Num: 23376, Cur Loss: 0.24101239, Cur Avg Loss: 1.25657836, Log Avg loss: 1.11957354, Global Avg Loss: 1.58977124, Time: 0.0637 Steps: 67460, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001471, Sample Num: 23536, Cur Loss: 1.41202521, Cur Avg Loss: 1.25782708, Log Avg loss: 1.44026417, Global Avg Loss: 1.58974908, Time: 0.0405 Steps: 67470, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001481, Sample Num: 23696, Cur Loss: 0.94941062, Cur Avg Loss: 1.25927782, Log Avg loss: 1.47268167, Global Avg Loss: 1.58973173, Time: 0.0627 Steps: 67480, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001491, Sample Num: 23856, Cur Loss: 0.92981648, Cur Avg Loss: 1.26060019, Log Avg loss: 1.45644393, Global Avg Loss: 1.58971198, Time: 0.0495 Steps: 67490, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001501, Sample Num: 24016, Cur Loss: 1.50699079, Cur Avg Loss: 1.26002105, Log Avg loss: 1.17367122, Global Avg Loss: 1.58965034, Time: 0.0766 Steps: 67500, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001511, Sample Num: 24176, Cur Loss: 1.31554615, Cur Avg Loss: 1.26136824, Log Avg loss: 1.46358039, Global Avg Loss: 1.58963167, Time: 0.0648 Steps: 67510, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001521, Sample Num: 24336, Cur Loss: 0.85834062, Cur Avg Loss: 1.26028568, Log Avg loss: 1.09671154, Global Avg Loss: 1.58955867, Time: 0.0691 Steps: 67520, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001531, Sample Num: 24496, Cur Loss: 1.40774477, Cur Avg Loss: 1.26088474, Log Avg loss: 1.35200185, Global Avg Loss: 1.58952349, Time: 0.0848 Steps: 67530, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001541, Sample Num: 24656, Cur Loss: 0.63597977, Cur Avg Loss: 1.25851720, Log Avg loss: 0.89604613, Global Avg Loss: 1.58942081, Time: 0.0405 Steps: 67540, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001551, Sample Num: 24816, Cur Loss: 2.02170467, Cur Avg Loss: 1.25897663, Log Avg loss: 1.32977502, Global Avg Loss: 1.58938237, Time: 0.0560 Steps: 67550, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001561, Sample Num: 24976, Cur Loss: 0.70570993, Cur Avg Loss: 1.25972575, Log Avg loss: 1.37591511, Global Avg Loss: 1.58935078, Time: 0.0402 Steps: 67560, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001571, Sample Num: 25136, Cur Loss: 0.61395824, Cur Avg Loss: 1.25722415, Log Avg loss: 0.86672388, Global Avg Loss: 1.58924383, Time: 0.0403 Steps: 67570, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001581, Sample Num: 25296, Cur Loss: 1.57767403, Cur Avg Loss: 1.25625184, Log Avg loss: 1.10350234, Global Avg Loss: 1.58917196, Time: 0.0402 Steps: 67580, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001591, Sample Num: 25456, Cur Loss: 0.60954428, Cur Avg Loss: 1.25523268, Log Avg loss: 1.09410250, Global Avg Loss: 1.58909871, Time: 0.0402 Steps: 67590, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001601, Sample Num: 25616, Cur Loss: 1.24766564, Cur Avg Loss: 1.25559005, Log Avg loss: 1.31244882, Global Avg Loss: 1.58905779, Time: 0.0403 Steps: 67600, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001611, Sample Num: 25776, Cur Loss: 0.24228258, Cur Avg Loss: 1.25369054, Log Avg loss: 0.94957836, Global Avg Loss: 1.58896320, Time: 0.0402 Steps: 67610, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001621, Sample Num: 25936, Cur Loss: 1.91287875, Cur Avg Loss: 1.25571640, Log Avg loss: 1.58208276, Global Avg Loss: 1.58896218, Time: 0.0403 Steps: 67620, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001631, Sample Num: 26096, Cur Loss: 1.28887391, Cur Avg Loss: 1.25730754, Log Avg loss: 1.51523128, Global Avg Loss: 1.58895128, Time: 0.0402 Steps: 67630, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001641, Sample Num: 26256, Cur Loss: 1.99572635, Cur Avg Loss: 1.25793583, Log Avg loss: 1.36040894, Global Avg Loss: 1.58891749, Time: 0.0403 Steps: 67640, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001651, Sample Num: 26416, Cur Loss: 1.41271901, Cur Avg Loss: 1.25654364, Log Avg loss: 1.02808592, Global Avg Loss: 1.58883459, Time: 0.0402 Steps: 67650, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001661, Sample Num: 26576, Cur Loss: 0.80304611, Cur Avg Loss: 1.25750656, Log Avg loss: 1.41648418, Global Avg Loss: 1.58880912, Time: 0.0403 Steps: 67660, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001671, Sample Num: 26736, Cur Loss: 1.40452349, Cur Avg Loss: 1.26119353, Log Avg loss: 1.87359960, Global Avg Loss: 1.58885120, Time: 0.0403 Steps: 67670, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001681, Sample Num: 26896, Cur Loss: 1.70042861, Cur Avg Loss: 1.26294730, Log Avg loss: 1.55600175, Global Avg Loss: 1.58884635, Time: 0.0403 Steps: 67680, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001691, Sample Num: 27056, Cur Loss: 1.15286994, Cur Avg Loss: 1.26174348, Log Avg loss: 1.05938138, Global Avg Loss: 1.58876813, Time: 0.0402 Steps: 67690, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001701, Sample Num: 27216, Cur Loss: 1.01871216, Cur Avg Loss: 1.26393729, Log Avg loss: 1.63491172, Global Avg Loss: 1.58877495, Time: 0.0403 Steps: 67700, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001711, Sample Num: 27376, Cur Loss: 1.30884612, Cur Avg Loss: 1.26344104, Log Avg loss: 1.17902783, Global Avg Loss: 1.58871443, Time: 0.0402 Steps: 67710, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001721, Sample Num: 27536, Cur Loss: 1.41028202, Cur Avg Loss: 1.26333078, Log Avg loss: 1.24446616, Global Avg Loss: 1.58866360, Time: 0.0403 Steps: 67720, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001731, Sample Num: 27696, Cur Loss: 1.99571395, Cur Avg Loss: 1.26334782, Log Avg loss: 1.26628020, Global Avg Loss: 1.58861600, Time: 0.0402 Steps: 67730, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001741, Sample Num: 27856, Cur Loss: 1.17243791, Cur Avg Loss: 1.26318017, Log Avg loss: 1.23416017, Global Avg Loss: 1.58856367, Time: 0.0403 Steps: 67740, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001751, Sample Num: 28016, Cur Loss: 4.09508133, Cur Avg Loss: 1.26614694, Log Avg loss: 1.78266180, Global Avg Loss: 1.58859232, Time: 0.0402 Steps: 67750, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001761, Sample Num: 28176, Cur Loss: 0.68014753, Cur Avg Loss: 1.26428065, Log Avg loss: 0.93749337, Global Avg Loss: 1.58849623, Time: 0.0402 Steps: 67760, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001771, Sample Num: 28336, Cur Loss: 1.18552148, Cur Avg Loss: 1.26530916, Log Avg loss: 1.44642972, Global Avg Loss: 1.58847527, Time: 0.0402 Steps: 67770, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001781, Sample Num: 28496, Cur Loss: 0.46210325, Cur Avg Loss: 1.26388028, Log Avg loss: 1.01082522, Global Avg Loss: 1.58839005, Time: 0.0402 Steps: 67780, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001791, Sample Num: 28656, Cur Loss: 0.68306684, Cur Avg Loss: 1.26269959, Log Avg loss: 1.05241797, Global Avg Loss: 1.58831098, Time: 0.0402 Steps: 67790, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001801, Sample Num: 28816, Cur Loss: 1.19924033, Cur Avg Loss: 1.26286236, Log Avg loss: 1.29201443, Global Avg Loss: 1.58826728, Time: 0.0402 Steps: 67800, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001811, Sample Num: 28976, Cur Loss: 0.55932647, Cur Avg Loss: 1.26165970, Log Avg loss: 1.04506088, Global Avg Loss: 1.58818717, Time: 0.0402 Steps: 67810, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001821, Sample Num: 29136, Cur Loss: 2.28849339, Cur Avg Loss: 1.26320478, Log Avg loss: 1.54301875, Global Avg Loss: 1.58818051, Time: 0.0406 Steps: 67820, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001831, Sample Num: 29296, Cur Loss: 1.94675064, Cur Avg Loss: 1.26346373, Log Avg loss: 1.31061808, Global Avg Loss: 1.58813959, Time: 0.0871 Steps: 67830, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001841, Sample Num: 29456, Cur Loss: 0.83072448, Cur Avg Loss: 1.26396938, Log Avg loss: 1.35655466, Global Avg Loss: 1.58810546, Time: 0.0455 Steps: 67840, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001851, Sample Num: 29616, Cur Loss: 1.23004591, Cur Avg Loss: 1.26412907, Log Avg loss: 1.29352838, Global Avg Loss: 1.58806204, Time: 0.0405 Steps: 67850, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001861, Sample Num: 29776, Cur Loss: 1.52653432, Cur Avg Loss: 1.26526205, Log Avg loss: 1.47497607, Global Avg Loss: 1.58804538, Time: 0.0841 Steps: 67860, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001871, Sample Num: 29936, Cur Loss: 3.10054231, Cur Avg Loss: 1.26614749, Log Avg loss: 1.43092887, Global Avg Loss: 1.58802223, Time: 0.0876 Steps: 67870, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001881, Sample Num: 30096, Cur Loss: 1.64635503, Cur Avg Loss: 1.26642562, Log Avg loss: 1.31846397, Global Avg Loss: 1.58798252, Time: 0.0518 Steps: 67880, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001891, Sample Num: 30256, Cur Loss: 1.18993378, Cur Avg Loss: 1.26558898, Log Avg loss: 1.10821692, Global Avg Loss: 1.58791185, Time: 0.0711 Steps: 67890, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001901, Sample Num: 30416, Cur Loss: 0.90540898, Cur Avg Loss: 1.26543646, Log Avg loss: 1.23659344, Global Avg Loss: 1.58786011, Time: 0.0404 Steps: 67900, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001911, Sample Num: 30576, Cur Loss: 0.50129747, Cur Avg Loss: 1.26572953, Log Avg loss: 1.32144258, Global Avg Loss: 1.58782088, Time: 0.0404 Steps: 67910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001921, Sample Num: 30736, Cur Loss: 1.13209641, Cur Avg Loss: 1.26494513, Log Avg loss: 1.11504596, Global Avg Loss: 1.58775127, Time: 0.0403 Steps: 67920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001931, Sample Num: 30896, Cur Loss: 0.85288775, Cur Avg Loss: 1.26331892, Log Avg loss: 0.95092396, Global Avg Loss: 1.58765752, Time: 0.0403 Steps: 67930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001941, Sample Num: 31056, Cur Loss: 1.13189268, Cur Avg Loss: 1.26281883, Log Avg loss: 1.16625254, Global Avg Loss: 1.58759550, Time: 0.0403 Steps: 67940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001951, Sample Num: 31216, Cur Loss: 0.29198304, Cur Avg Loss: 1.26303368, Log Avg loss: 1.30473532, Global Avg Loss: 1.58755387, Time: 0.0402 Steps: 67950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001961, Sample Num: 31376, Cur Loss: 1.57516217, Cur Avg Loss: 1.26343203, Log Avg loss: 1.34115014, Global Avg Loss: 1.58751761, Time: 0.0403 Steps: 67960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001971, Sample Num: 31536, Cur Loss: 1.10083711, Cur Avg Loss: 1.26495914, Log Avg loss: 1.56442609, Global Avg Loss: 1.58751421, Time: 0.0403 Steps: 67970, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001981, Sample Num: 31696, Cur Loss: 0.67110884, Cur Avg Loss: 1.26440998, Log Avg loss: 1.15616963, Global Avg Loss: 1.58745076, Time: 0.0403 Steps: 67980, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001991, Sample Num: 31856, Cur Loss: 1.47346735, Cur Avg Loss: 1.26336822, Log Avg loss: 1.05699538, Global Avg Loss: 1.58737274, Time: 0.0403 Steps: 67990, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002001, Sample Num: 32016, Cur Loss: 0.47889245, Cur Avg Loss: 1.26332634, Log Avg loss: 1.25498861, Global Avg Loss: 1.58732386, Time: 0.0403 Steps: 68000, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002011, Sample Num: 32176, Cur Loss: 1.29423785, Cur Avg Loss: 1.26396907, Log Avg loss: 1.39257937, Global Avg Loss: 1.58729523, Time: 0.0402 Steps: 68010, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002021, Sample Num: 32336, Cur Loss: 0.98323369, Cur Avg Loss: 1.26297644, Log Avg loss: 1.06335899, Global Avg Loss: 1.58721820, Time: 0.0402 Steps: 68020, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002031, Sample Num: 32496, Cur Loss: 2.78293371, Cur Avg Loss: 1.26396361, Log Avg loss: 1.46346930, Global Avg Loss: 1.58720001, Time: 0.0402 Steps: 68030, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002041, Sample Num: 32656, Cur Loss: 1.18188989, Cur Avg Loss: 1.26429230, Log Avg loss: 1.33104988, Global Avg Loss: 1.58716236, Time: 0.0402 Steps: 68040, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002051, Sample Num: 32816, Cur Loss: 1.59138870, Cur Avg Loss: 1.26461310, Log Avg loss: 1.33008895, Global Avg Loss: 1.58712459, Time: 0.0402 Steps: 68050, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002061, Sample Num: 32976, Cur Loss: 0.91001761, Cur Avg Loss: 1.26431786, Log Avg loss: 1.20376354, Global Avg Loss: 1.58706826, Time: 0.0402 Steps: 68060, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002071, Sample Num: 33136, Cur Loss: 1.11225533, Cur Avg Loss: 1.26330414, Log Avg loss: 1.05437749, Global Avg Loss: 1.58699000, Time: 0.0402 Steps: 68070, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002081, Sample Num: 33296, Cur Loss: 1.05670023, Cur Avg Loss: 1.26286419, Log Avg loss: 1.17174968, Global Avg Loss: 1.58692901, Time: 0.0402 Steps: 68080, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002091, Sample Num: 33456, Cur Loss: 2.10866165, Cur Avg Loss: 1.26299162, Log Avg loss: 1.28951064, Global Avg Loss: 1.58688533, Time: 0.0402 Steps: 68090, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002101, Sample Num: 33616, Cur Loss: 1.07197618, Cur Avg Loss: 1.26197054, Log Avg loss: 1.04846280, Global Avg Loss: 1.58680627, Time: 0.0402 Steps: 68100, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002111, Sample Num: 33776, Cur Loss: 1.18974066, Cur Avg Loss: 1.26181229, Log Avg loss: 1.22856383, Global Avg Loss: 1.58675367, Time: 0.0402 Steps: 68110, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002121, Sample Num: 33936, Cur Loss: 1.12505007, Cur Avg Loss: 1.26158424, Log Avg loss: 1.21344161, Global Avg Loss: 1.58669887, Time: 0.0402 Steps: 68120, Updated lr: 0.000036 ***** Running evaluation checkpoint-68128 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-68128 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.348167, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.164086, "eval_total_loss": 818.352259, "eval_mae": 0.861252, "eval_mse": 1.164238, "eval_r2": 0.259934, "eval_sp_statistic": 0.419131, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.514778, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.132327, "test_total_loss": 568.428402, "test_mae": 0.851422, "test_mse": 1.132558, "test_r2": 0.269037, "test_sp_statistic": 0.443581, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.553291, "test_ps_pvalue": 0.0, "lr": 3.634139402560455e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5866193511412194, "train_cur_epoch_loss": 2683.0965433567762, "train_cur_epoch_avg_loss": 1.2602614106889507, "train_cur_epoch_time": 95.34816718101501, "train_cur_epoch_avg_time": 0.04478542375810945, "epoch": 32, "step": 68128} ################################################## Training, Epoch: 0033, Batch: 000002, Sample Num: 32, Cur Loss: 1.13972092, Cur Avg Loss: 1.16393787, Log Avg loss: 0.96042521, Global Avg Loss: 1.58660694, Time: 0.0403 Steps: 68130, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000012, Sample Num: 192, Cur Loss: 1.54878759, Cur Avg Loss: 1.46611076, Log Avg loss: 1.52654534, Global Avg Loss: 1.58659813, Time: 0.0402 Steps: 68140, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000022, Sample Num: 352, Cur Loss: 2.81344771, Cur Avg Loss: 1.49397678, Log Avg loss: 1.52741601, Global Avg Loss: 1.58658944, Time: 0.0402 Steps: 68150, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000032, Sample Num: 512, Cur Loss: 1.39220262, Cur Avg Loss: 1.40724095, Log Avg loss: 1.21642211, Global Avg Loss: 1.58653514, Time: 0.0402 Steps: 68160, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000042, Sample Num: 672, Cur Loss: 1.68148732, Cur Avg Loss: 1.39774620, Log Avg loss: 1.36736302, Global Avg Loss: 1.58650299, Time: 0.0402 Steps: 68170, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000052, Sample Num: 832, Cur Loss: 1.69129014, Cur Avg Loss: 1.41468671, Log Avg loss: 1.48583682, Global Avg Loss: 1.58648822, Time: 0.0402 Steps: 68180, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000062, Sample Num: 992, Cur Loss: 0.61245584, Cur Avg Loss: 1.41361251, Log Avg loss: 1.40802672, Global Avg Loss: 1.58646205, Time: 0.0402 Steps: 68190, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000072, Sample Num: 1152, Cur Loss: 0.76280200, Cur Avg Loss: 1.38409831, Log Avg loss: 1.20111026, Global Avg Loss: 1.58640555, Time: 0.0402 Steps: 68200, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000082, Sample Num: 1312, Cur Loss: 0.64802039, Cur Avg Loss: 1.34679567, Log Avg loss: 1.07821662, Global Avg Loss: 1.58633104, Time: 0.0402 Steps: 68210, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000092, Sample Num: 1472, Cur Loss: 0.86427498, Cur Avg Loss: 1.34097124, Log Avg loss: 1.29321095, Global Avg Loss: 1.58628808, Time: 0.0830 Steps: 68220, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000102, Sample Num: 1632, Cur Loss: 0.81464481, Cur Avg Loss: 1.29763559, Log Avg loss: 0.89894755, Global Avg Loss: 1.58618734, Time: 0.0452 Steps: 68230, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000112, Sample Num: 1792, Cur Loss: 2.08837223, Cur Avg Loss: 1.27761075, Log Avg loss: 1.07335739, Global Avg Loss: 1.58611219, Time: 0.0559 Steps: 68240, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000122, Sample Num: 1952, Cur Loss: 1.44705987, Cur Avg Loss: 1.28356983, Log Avg loss: 1.35031161, Global Avg Loss: 1.58607764, Time: 0.0851 Steps: 68250, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000132, Sample Num: 2112, Cur Loss: 0.68826628, Cur Avg Loss: 1.29037006, Log Avg loss: 1.37333285, Global Avg Loss: 1.58604647, Time: 0.0408 Steps: 68260, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000142, Sample Num: 2272, Cur Loss: 0.78520244, Cur Avg Loss: 1.28904413, Log Avg loss: 1.27154184, Global Avg Loss: 1.58600040, Time: 0.0403 Steps: 68270, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000152, Sample Num: 2432, Cur Loss: 0.97342277, Cur Avg Loss: 1.29825305, Log Avg loss: 1.42901964, Global Avg Loss: 1.58597741, Time: 0.0694 Steps: 68280, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000162, Sample Num: 2592, Cur Loss: 0.87510431, Cur Avg Loss: 1.31604163, Log Avg loss: 1.58642803, Global Avg Loss: 1.58597748, Time: 0.0430 Steps: 68290, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000172, Sample Num: 2752, Cur Loss: 0.95535028, Cur Avg Loss: 1.30001303, Log Avg loss: 1.04034980, Global Avg Loss: 1.58589759, Time: 0.0607 Steps: 68300, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000182, Sample Num: 2912, Cur Loss: 2.10549307, Cur Avg Loss: 1.28941634, Log Avg loss: 1.10715326, Global Avg Loss: 1.58582751, Time: 0.0406 Steps: 68310, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000192, Sample Num: 3072, Cur Loss: 2.63475561, Cur Avg Loss: 1.29793790, Log Avg loss: 1.45303035, Global Avg Loss: 1.58580807, Time: 0.0402 Steps: 68320, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000202, Sample Num: 3232, Cur Loss: 0.74882698, Cur Avg Loss: 1.28661287, Log Avg loss: 1.06917228, Global Avg Loss: 1.58573246, Time: 0.0403 Steps: 68330, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000212, Sample Num: 3392, Cur Loss: 1.45048094, Cur Avg Loss: 1.29199061, Log Avg loss: 1.40062096, Global Avg Loss: 1.58570537, Time: 0.0402 Steps: 68340, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000222, Sample Num: 3552, Cur Loss: 1.25410831, Cur Avg Loss: 1.28494591, Log Avg loss: 1.13559818, Global Avg Loss: 1.58563952, Time: 0.0402 Steps: 68350, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000232, Sample Num: 3712, Cur Loss: 0.67288297, Cur Avg Loss: 1.28130107, Log Avg loss: 1.20038562, Global Avg Loss: 1.58558316, Time: 0.0403 Steps: 68360, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000242, Sample Num: 3872, Cur Loss: 1.51779985, Cur Avg Loss: 1.28553917, Log Avg loss: 1.38386305, Global Avg Loss: 1.58555366, Time: 0.0402 Steps: 68370, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000252, Sample Num: 4032, Cur Loss: 1.57665062, Cur Avg Loss: 1.28080509, Log Avg loss: 1.16624040, Global Avg Loss: 1.58549234, Time: 0.0402 Steps: 68380, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000262, Sample Num: 4192, Cur Loss: 0.45000055, Cur Avg Loss: 1.27342872, Log Avg loss: 1.08754424, Global Avg Loss: 1.58541953, Time: 0.0402 Steps: 68390, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000272, Sample Num: 4352, Cur Loss: 1.16076899, Cur Avg Loss: 1.27463788, Log Avg loss: 1.30631773, Global Avg Loss: 1.58537872, Time: 0.0403 Steps: 68400, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000282, Sample Num: 4512, Cur Loss: 1.78345621, Cur Avg Loss: 1.27656493, Log Avg loss: 1.32898076, Global Avg Loss: 1.58534124, Time: 0.0402 Steps: 68410, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000292, Sample Num: 4672, Cur Loss: 1.56515968, Cur Avg Loss: 1.27446913, Log Avg loss: 1.21536769, Global Avg Loss: 1.58528717, Time: 0.0402 Steps: 68420, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000302, Sample Num: 4832, Cur Loss: 1.82791674, Cur Avg Loss: 1.28146588, Log Avg loss: 1.48577105, Global Avg Loss: 1.58527263, Time: 0.0403 Steps: 68430, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000312, Sample Num: 4992, Cur Loss: 1.52323294, Cur Avg Loss: 1.27694348, Log Avg loss: 1.14036701, Global Avg Loss: 1.58520762, Time: 0.0403 Steps: 68440, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000322, Sample Num: 5152, Cur Loss: 1.32718420, Cur Avg Loss: 1.26926234, Log Avg loss: 1.02961074, Global Avg Loss: 1.58512645, Time: 0.0402 Steps: 68450, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000332, Sample Num: 5312, Cur Loss: 2.22884417, Cur Avg Loss: 1.27987767, Log Avg loss: 1.62169123, Global Avg Loss: 1.58513179, Time: 0.0402 Steps: 68460, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000342, Sample Num: 5472, Cur Loss: 0.73183692, Cur Avg Loss: 1.27785064, Log Avg loss: 1.21055330, Global Avg Loss: 1.58507709, Time: 0.0403 Steps: 68470, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000352, Sample Num: 5632, Cur Loss: 1.44615149, Cur Avg Loss: 1.27789925, Log Avg loss: 1.27956156, Global Avg Loss: 1.58503247, Time: 0.0402 Steps: 68480, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000362, Sample Num: 5792, Cur Loss: 1.01530218, Cur Avg Loss: 1.28592287, Log Avg loss: 1.56835422, Global Avg Loss: 1.58503004, Time: 0.0402 Steps: 68490, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000372, Sample Num: 5952, Cur Loss: 0.95617282, Cur Avg Loss: 1.29160800, Log Avg loss: 1.49740984, Global Avg Loss: 1.58501725, Time: 0.0402 Steps: 68500, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000382, Sample Num: 6112, Cur Loss: 0.99891937, Cur Avg Loss: 1.28934097, Log Avg loss: 1.20500754, Global Avg Loss: 1.58496178, Time: 0.0402 Steps: 68510, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000392, Sample Num: 6272, Cur Loss: 1.21286166, Cur Avg Loss: 1.29182758, Log Avg loss: 1.38681583, Global Avg Loss: 1.58493286, Time: 0.0402 Steps: 68520, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000402, Sample Num: 6432, Cur Loss: 1.34701610, Cur Avg Loss: 1.29110190, Log Avg loss: 1.26265536, Global Avg Loss: 1.58488583, Time: 0.0402 Steps: 68530, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000412, Sample Num: 6592, Cur Loss: 1.04743218, Cur Avg Loss: 1.28236103, Log Avg loss: 0.93097815, Global Avg Loss: 1.58479043, Time: 0.0402 Steps: 68540, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000422, Sample Num: 6752, Cur Loss: 2.08259916, Cur Avg Loss: 1.28089872, Log Avg loss: 1.22065161, Global Avg Loss: 1.58473731, Time: 0.0647 Steps: 68550, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000432, Sample Num: 6912, Cur Loss: 3.02140188, Cur Avg Loss: 1.28398211, Log Avg loss: 1.41410123, Global Avg Loss: 1.58471242, Time: 0.0769 Steps: 68560, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000442, Sample Num: 7072, Cur Loss: 0.46304351, Cur Avg Loss: 1.28190794, Log Avg loss: 1.19230372, Global Avg Loss: 1.58465519, Time: 0.0404 Steps: 68570, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000452, Sample Num: 7232, Cur Loss: 1.10625803, Cur Avg Loss: 1.28514848, Log Avg loss: 1.42838023, Global Avg Loss: 1.58463240, Time: 0.0585 Steps: 68580, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000462, Sample Num: 7392, Cur Loss: 1.66600502, Cur Avg Loss: 1.28270654, Log Avg loss: 1.17233104, Global Avg Loss: 1.58457229, Time: 0.0482 Steps: 68590, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000472, Sample Num: 7552, Cur Loss: 1.07119787, Cur Avg Loss: 1.28007985, Log Avg loss: 1.15872669, Global Avg Loss: 1.58451022, Time: 0.0538 Steps: 68600, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000482, Sample Num: 7712, Cur Loss: 3.82326674, Cur Avg Loss: 1.28179926, Log Avg loss: 1.36295556, Global Avg Loss: 1.58447792, Time: 0.0675 Steps: 68610, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000492, Sample Num: 7872, Cur Loss: 1.75492978, Cur Avg Loss: 1.28290883, Log Avg loss: 1.33638980, Global Avg Loss: 1.58444177, Time: 0.0639 Steps: 68620, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000502, Sample Num: 8032, Cur Loss: 1.26843524, Cur Avg Loss: 1.28404242, Log Avg loss: 1.33981501, Global Avg Loss: 1.58440613, Time: 0.0403 Steps: 68630, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000512, Sample Num: 8192, Cur Loss: 0.28361607, Cur Avg Loss: 1.27629698, Log Avg loss: 0.88747617, Global Avg Loss: 1.58430459, Time: 0.0404 Steps: 68640, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000522, Sample Num: 8352, Cur Loss: 1.00697219, Cur Avg Loss: 1.27963599, Log Avg loss: 1.45059312, Global Avg Loss: 1.58428511, Time: 0.0403 Steps: 68650, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000532, Sample Num: 8512, Cur Loss: 1.57984996, Cur Avg Loss: 1.28073157, Log Avg loss: 1.33792082, Global Avg Loss: 1.58424923, Time: 0.0403 Steps: 68660, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000542, Sample Num: 8672, Cur Loss: 0.95844960, Cur Avg Loss: 1.27553403, Log Avg loss: 0.99902489, Global Avg Loss: 1.58416401, Time: 0.0402 Steps: 68670, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000552, Sample Num: 8832, Cur Loss: 1.24426222, Cur Avg Loss: 1.27938673, Log Avg loss: 1.48820319, Global Avg Loss: 1.58415004, Time: 0.0403 Steps: 68680, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000562, Sample Num: 8992, Cur Loss: 1.37206829, Cur Avg Loss: 1.27947698, Log Avg loss: 1.28445876, Global Avg Loss: 1.58410641, Time: 0.0403 Steps: 68690, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000572, Sample Num: 9152, Cur Loss: 0.91467345, Cur Avg Loss: 1.28219100, Log Avg loss: 1.43471910, Global Avg Loss: 1.58408466, Time: 0.0404 Steps: 68700, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000582, Sample Num: 9312, Cur Loss: 1.69729376, Cur Avg Loss: 1.28025001, Log Avg loss: 1.16922493, Global Avg Loss: 1.58402429, Time: 0.0402 Steps: 68710, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000592, Sample Num: 9472, Cur Loss: 1.05819952, Cur Avg Loss: 1.27757779, Log Avg loss: 1.12205479, Global Avg Loss: 1.58395706, Time: 0.0402 Steps: 68720, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000602, Sample Num: 9632, Cur Loss: 0.73164952, Cur Avg Loss: 1.27310915, Log Avg loss: 1.00856562, Global Avg Loss: 1.58387334, Time: 0.0403 Steps: 68730, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000612, Sample Num: 9792, Cur Loss: 1.03215694, Cur Avg Loss: 1.26901410, Log Avg loss: 1.02249241, Global Avg Loss: 1.58379168, Time: 0.0402 Steps: 68740, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000622, Sample Num: 9952, Cur Loss: 1.42583156, Cur Avg Loss: 1.26498034, Log Avg loss: 1.01811428, Global Avg Loss: 1.58370940, Time: 0.0403 Steps: 68750, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000632, Sample Num: 10112, Cur Loss: 0.40855262, Cur Avg Loss: 1.26214028, Log Avg loss: 1.08548811, Global Avg Loss: 1.58363694, Time: 0.0402 Steps: 68760, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000642, Sample Num: 10272, Cur Loss: 0.44668224, Cur Avg Loss: 1.26022595, Log Avg loss: 1.13924038, Global Avg Loss: 1.58357232, Time: 0.0402 Steps: 68770, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000652, Sample Num: 10432, Cur Loss: 1.11816609, Cur Avg Loss: 1.26329451, Log Avg loss: 1.46029613, Global Avg Loss: 1.58355439, Time: 0.0402 Steps: 68780, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000662, Sample Num: 10592, Cur Loss: 1.07840419, Cur Avg Loss: 1.26082815, Log Avg loss: 1.10002180, Global Avg Loss: 1.58348410, Time: 0.0402 Steps: 68790, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000672, Sample Num: 10752, Cur Loss: 1.09106874, Cur Avg Loss: 1.26211564, Log Avg loss: 1.34734693, Global Avg Loss: 1.58344978, Time: 0.0402 Steps: 68800, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000682, Sample Num: 10912, Cur Loss: 0.65916383, Cur Avg Loss: 1.25927383, Log Avg loss: 1.06830435, Global Avg Loss: 1.58337492, Time: 0.0402 Steps: 68810, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000692, Sample Num: 11072, Cur Loss: 1.54078889, Cur Avg Loss: 1.26211806, Log Avg loss: 1.45609442, Global Avg Loss: 1.58335642, Time: 0.0403 Steps: 68820, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000702, Sample Num: 11232, Cur Loss: 1.69959152, Cur Avg Loss: 1.25981489, Log Avg loss: 1.10043553, Global Avg Loss: 1.58328626, Time: 0.0402 Steps: 68830, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000712, Sample Num: 11392, Cur Loss: 1.10920191, Cur Avg Loss: 1.26108170, Log Avg loss: 1.35001225, Global Avg Loss: 1.58325237, Time: 0.0403 Steps: 68840, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000722, Sample Num: 11552, Cur Loss: 1.43911374, Cur Avg Loss: 1.25987303, Log Avg loss: 1.17381539, Global Avg Loss: 1.58319290, Time: 0.0402 Steps: 68850, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000732, Sample Num: 11712, Cur Loss: 0.45143872, Cur Avg Loss: 1.25959337, Log Avg loss: 1.23940205, Global Avg Loss: 1.58314298, Time: 0.0402 Steps: 68860, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000742, Sample Num: 11872, Cur Loss: 1.74761438, Cur Avg Loss: 1.25627530, Log Avg loss: 1.01339230, Global Avg Loss: 1.58306025, Time: 0.0401 Steps: 68870, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000752, Sample Num: 12032, Cur Loss: 1.12186170, Cur Avg Loss: 1.25366827, Log Avg loss: 1.06022685, Global Avg Loss: 1.58298435, Time: 0.0404 Steps: 68880, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000762, Sample Num: 12192, Cur Loss: 0.49750188, Cur Avg Loss: 1.25253289, Log Avg loss: 1.16715218, Global Avg Loss: 1.58292398, Time: 0.0409 Steps: 68890, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000772, Sample Num: 12352, Cur Loss: 1.69510531, Cur Avg Loss: 1.25161867, Log Avg loss: 1.18195502, Global Avg Loss: 1.58286579, Time: 0.0406 Steps: 68900, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000782, Sample Num: 12512, Cur Loss: 0.32563841, Cur Avg Loss: 1.24841762, Log Avg loss: 1.00129662, Global Avg Loss: 1.58278139, Time: 0.0456 Steps: 68910, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000792, Sample Num: 12672, Cur Loss: 2.20930123, Cur Avg Loss: 1.24849084, Log Avg loss: 1.25421649, Global Avg Loss: 1.58273372, Time: 0.0671 Steps: 68920, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000802, Sample Num: 12832, Cur Loss: 2.38604474, Cur Avg Loss: 1.25175108, Log Avg loss: 1.50996262, Global Avg Loss: 1.58272316, Time: 0.0612 Steps: 68930, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000812, Sample Num: 12992, Cur Loss: 0.68930984, Cur Avg Loss: 1.24907097, Log Avg loss: 1.03412620, Global Avg Loss: 1.58264359, Time: 0.0404 Steps: 68940, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000822, Sample Num: 13152, Cur Loss: 1.76779962, Cur Avg Loss: 1.25032995, Log Avg loss: 1.35255858, Global Avg Loss: 1.58261022, Time: 0.0623 Steps: 68950, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000832, Sample Num: 13312, Cur Loss: 0.75803596, Cur Avg Loss: 1.24770756, Log Avg loss: 1.03214750, Global Avg Loss: 1.58253039, Time: 0.0997 Steps: 68960, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000842, Sample Num: 13472, Cur Loss: 1.58145201, Cur Avg Loss: 1.25034176, Log Avg loss: 1.46950704, Global Avg Loss: 1.58251400, Time: 0.0404 Steps: 68970, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000852, Sample Num: 13632, Cur Loss: 1.07376456, Cur Avg Loss: 1.25193764, Log Avg loss: 1.38631071, Global Avg Loss: 1.58248556, Time: 0.0402 Steps: 68980, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000862, Sample Num: 13792, Cur Loss: 0.44363326, Cur Avg Loss: 1.25205742, Log Avg loss: 1.26226264, Global Avg Loss: 1.58243915, Time: 0.0403 Steps: 68990, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000872, Sample Num: 13952, Cur Loss: 0.58998781, Cur Avg Loss: 1.24879895, Log Avg loss: 0.96791922, Global Avg Loss: 1.58235008, Time: 0.0402 Steps: 69000, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000882, Sample Num: 14112, Cur Loss: 1.04169893, Cur Avg Loss: 1.25243091, Log Avg loss: 1.56913766, Global Avg Loss: 1.58234817, Time: 0.0403 Steps: 69010, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000892, Sample Num: 14272, Cur Loss: 0.77401447, Cur Avg Loss: 1.25226043, Log Avg loss: 1.23722420, Global Avg Loss: 1.58229817, Time: 0.0403 Steps: 69020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000902, Sample Num: 14432, Cur Loss: 1.64895642, Cur Avg Loss: 1.25645442, Log Avg loss: 1.63055781, Global Avg Loss: 1.58230516, Time: 0.0402 Steps: 69030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000912, Sample Num: 14592, Cur Loss: 0.32734433, Cur Avg Loss: 1.25712789, Log Avg loss: 1.31787538, Global Avg Loss: 1.58226686, Time: 0.0403 Steps: 69040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000922, Sample Num: 14752, Cur Loss: 0.78102797, Cur Avg Loss: 1.25770818, Log Avg loss: 1.31063015, Global Avg Loss: 1.58222752, Time: 0.0403 Steps: 69050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000932, Sample Num: 14912, Cur Loss: 0.98102415, Cur Avg Loss: 1.25704883, Log Avg loss: 1.19625677, Global Avg Loss: 1.58217163, Time: 0.0402 Steps: 69060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000942, Sample Num: 15072, Cur Loss: 2.18807459, Cur Avg Loss: 1.25592848, Log Avg loss: 1.15151249, Global Avg Loss: 1.58210928, Time: 0.0403 Steps: 69070, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000952, Sample Num: 15232, Cur Loss: 1.20959473, Cur Avg Loss: 1.25747280, Log Avg loss: 1.40294766, Global Avg Loss: 1.58208334, Time: 0.0403 Steps: 69080, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000962, Sample Num: 15392, Cur Loss: 1.47169852, Cur Avg Loss: 1.25780567, Log Avg loss: 1.28949432, Global Avg Loss: 1.58204099, Time: 0.0403 Steps: 69090, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000972, Sample Num: 15552, Cur Loss: 1.06680059, Cur Avg Loss: 1.25887494, Log Avg loss: 1.36173863, Global Avg Loss: 1.58200911, Time: 0.0403 Steps: 69100, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000982, Sample Num: 15712, Cur Loss: 2.14395142, Cur Avg Loss: 1.26003604, Log Avg loss: 1.37289515, Global Avg Loss: 1.58197885, Time: 0.0402 Steps: 69110, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000992, Sample Num: 15872, Cur Loss: 0.71958303, Cur Avg Loss: 1.26017806, Log Avg loss: 1.27412472, Global Avg Loss: 1.58193431, Time: 0.0403 Steps: 69120, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001002, Sample Num: 16032, Cur Loss: 0.87514472, Cur Avg Loss: 1.25995106, Log Avg loss: 1.23743267, Global Avg Loss: 1.58188448, Time: 0.0402 Steps: 69130, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001012, Sample Num: 16192, Cur Loss: 1.06346130, Cur Avg Loss: 1.25823836, Log Avg loss: 1.08662530, Global Avg Loss: 1.58181285, Time: 0.0402 Steps: 69140, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001022, Sample Num: 16352, Cur Loss: 0.69278073, Cur Avg Loss: 1.25448152, Log Avg loss: 0.87428972, Global Avg Loss: 1.58171053, Time: 0.0402 Steps: 69150, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001032, Sample Num: 16512, Cur Loss: 0.81845146, Cur Avg Loss: 1.25110834, Log Avg loss: 0.90636898, Global Avg Loss: 1.58161288, Time: 0.0402 Steps: 69160, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001042, Sample Num: 16672, Cur Loss: 1.55227435, Cur Avg Loss: 1.25340043, Log Avg loss: 1.48994486, Global Avg Loss: 1.58159963, Time: 0.0402 Steps: 69170, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001052, Sample Num: 16832, Cur Loss: 0.71865946, Cur Avg Loss: 1.25189017, Log Avg loss: 1.09452090, Global Avg Loss: 1.58152922, Time: 0.0402 Steps: 69180, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001062, Sample Num: 16992, Cur Loss: 0.57678598, Cur Avg Loss: 1.25055886, Log Avg loss: 1.11050511, Global Avg Loss: 1.58146115, Time: 0.0403 Steps: 69190, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001072, Sample Num: 17152, Cur Loss: 0.66748452, Cur Avg Loss: 1.25112961, Log Avg loss: 1.31174269, Global Avg Loss: 1.58142217, Time: 0.0402 Steps: 69200, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001082, Sample Num: 17312, Cur Loss: 0.65575647, Cur Avg Loss: 1.25110777, Log Avg loss: 1.24876711, Global Avg Loss: 1.58137410, Time: 0.0402 Steps: 69210, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001092, Sample Num: 17472, Cur Loss: 1.93325233, Cur Avg Loss: 1.24989998, Log Avg loss: 1.11921713, Global Avg Loss: 1.58130734, Time: 0.0404 Steps: 69220, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001102, Sample Num: 17632, Cur Loss: 1.16105163, Cur Avg Loss: 1.24934219, Log Avg loss: 1.18843153, Global Avg Loss: 1.58125059, Time: 0.0539 Steps: 69230, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001112, Sample Num: 17792, Cur Loss: 1.89737189, Cur Avg Loss: 1.24995287, Log Avg loss: 1.31724912, Global Avg Loss: 1.58121246, Time: 0.0404 Steps: 69240, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001122, Sample Num: 17952, Cur Loss: 1.38608825, Cur Avg Loss: 1.24965486, Log Avg loss: 1.21651656, Global Avg Loss: 1.58115980, Time: 0.0513 Steps: 69250, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001132, Sample Num: 18112, Cur Loss: 2.13576150, Cur Avg Loss: 1.25391570, Log Avg loss: 1.73198185, Global Avg Loss: 1.58118157, Time: 0.0438 Steps: 69260, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001142, Sample Num: 18272, Cur Loss: 1.46139741, Cur Avg Loss: 1.25326018, Log Avg loss: 1.17905560, Global Avg Loss: 1.58112352, Time: 0.0726 Steps: 69270, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001152, Sample Num: 18432, Cur Loss: 2.10472274, Cur Avg Loss: 1.25304426, Log Avg loss: 1.22838593, Global Avg Loss: 1.58107261, Time: 0.0600 Steps: 69280, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001162, Sample Num: 18592, Cur Loss: 0.29979113, Cur Avg Loss: 1.25158651, Log Avg loss: 1.08365402, Global Avg Loss: 1.58100082, Time: 0.0457 Steps: 69290, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001172, Sample Num: 18752, Cur Loss: 0.73094356, Cur Avg Loss: 1.25070676, Log Avg loss: 1.14847998, Global Avg Loss: 1.58093841, Time: 0.0402 Steps: 69300, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001182, Sample Num: 18912, Cur Loss: 0.69032592, Cur Avg Loss: 1.24853128, Log Avg loss: 0.99356478, Global Avg Loss: 1.58085366, Time: 0.0403 Steps: 69310, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001192, Sample Num: 19072, Cur Loss: 0.94264972, Cur Avg Loss: 1.24961091, Log Avg loss: 1.37722273, Global Avg Loss: 1.58082428, Time: 0.0403 Steps: 69320, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001202, Sample Num: 19232, Cur Loss: 0.85026562, Cur Avg Loss: 1.25090454, Log Avg loss: 1.40510572, Global Avg Loss: 1.58079894, Time: 0.0403 Steps: 69330, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001212, Sample Num: 19392, Cur Loss: 2.05439854, Cur Avg Loss: 1.25265913, Log Avg loss: 1.46356106, Global Avg Loss: 1.58078203, Time: 0.0402 Steps: 69340, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001222, Sample Num: 19552, Cur Loss: 1.47310066, Cur Avg Loss: 1.25474949, Log Avg loss: 1.50810086, Global Avg Loss: 1.58077155, Time: 0.0402 Steps: 69350, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001232, Sample Num: 19712, Cur Loss: 0.85892940, Cur Avg Loss: 1.25562953, Log Avg loss: 1.36316999, Global Avg Loss: 1.58074018, Time: 0.0403 Steps: 69360, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001242, Sample Num: 19872, Cur Loss: 1.75298762, Cur Avg Loss: 1.25408736, Log Avg loss: 1.06409192, Global Avg Loss: 1.58066570, Time: 0.0403 Steps: 69370, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001252, Sample Num: 20032, Cur Loss: 0.71618712, Cur Avg Loss: 1.25454752, Log Avg loss: 1.31169955, Global Avg Loss: 1.58062693, Time: 0.0403 Steps: 69380, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001262, Sample Num: 20192, Cur Loss: 1.74516928, Cur Avg Loss: 1.25421081, Log Avg loss: 1.21205533, Global Avg Loss: 1.58057382, Time: 0.0403 Steps: 69390, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001272, Sample Num: 20352, Cur Loss: 1.26035726, Cur Avg Loss: 1.25529426, Log Avg loss: 1.39202577, Global Avg Loss: 1.58054665, Time: 0.0403 Steps: 69400, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001282, Sample Num: 20512, Cur Loss: 0.81024539, Cur Avg Loss: 1.25450617, Log Avg loss: 1.15426006, Global Avg Loss: 1.58048523, Time: 0.0404 Steps: 69410, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001292, Sample Num: 20672, Cur Loss: 0.49538520, Cur Avg Loss: 1.25179210, Log Avg loss: 0.90384933, Global Avg Loss: 1.58038776, Time: 0.0403 Steps: 69420, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001302, Sample Num: 20832, Cur Loss: 0.28559941, Cur Avg Loss: 1.25230779, Log Avg loss: 1.31893448, Global Avg Loss: 1.58035011, Time: 0.0403 Steps: 69430, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001312, Sample Num: 20992, Cur Loss: 0.66770536, Cur Avg Loss: 1.25135498, Log Avg loss: 1.12729877, Global Avg Loss: 1.58028486, Time: 0.0402 Steps: 69440, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001322, Sample Num: 21152, Cur Loss: 1.71727431, Cur Avg Loss: 1.25179480, Log Avg loss: 1.30949946, Global Avg Loss: 1.58024587, Time: 0.0402 Steps: 69450, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001332, Sample Num: 21312, Cur Loss: 1.74239004, Cur Avg Loss: 1.25159739, Log Avg loss: 1.22549991, Global Avg Loss: 1.58019480, Time: 0.0402 Steps: 69460, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001342, Sample Num: 21472, Cur Loss: 0.63938105, Cur Avg Loss: 1.24954781, Log Avg loss: 0.97654322, Global Avg Loss: 1.58010791, Time: 0.0402 Steps: 69470, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001352, Sample Num: 21632, Cur Loss: 1.36965525, Cur Avg Loss: 1.25046269, Log Avg loss: 1.37324053, Global Avg Loss: 1.58007813, Time: 0.0402 Steps: 69480, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001362, Sample Num: 21792, Cur Loss: 0.80143237, Cur Avg Loss: 1.25293433, Log Avg loss: 1.58709911, Global Avg Loss: 1.58007914, Time: 0.0403 Steps: 69490, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001372, Sample Num: 21952, Cur Loss: 1.10864520, Cur Avg Loss: 1.25374139, Log Avg loss: 1.36366352, Global Avg Loss: 1.58004800, Time: 0.0402 Steps: 69500, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001382, Sample Num: 22112, Cur Loss: 0.66291887, Cur Avg Loss: 1.25104625, Log Avg loss: 0.88127302, Global Avg Loss: 1.57994748, Time: 0.0403 Steps: 69510, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001392, Sample Num: 22272, Cur Loss: 0.49730712, Cur Avg Loss: 1.25011766, Log Avg loss: 1.12178665, Global Avg Loss: 1.57988157, Time: 0.0402 Steps: 69520, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001402, Sample Num: 22432, Cur Loss: 1.46716249, Cur Avg Loss: 1.25132024, Log Avg loss: 1.41871911, Global Avg Loss: 1.57985839, Time: 0.0402 Steps: 69530, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001412, Sample Num: 22592, Cur Loss: 1.66151094, Cur Avg Loss: 1.25275931, Log Avg loss: 1.45451663, Global Avg Loss: 1.57984037, Time: 0.0403 Steps: 69540, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001422, Sample Num: 22752, Cur Loss: 1.32988882, Cur Avg Loss: 1.25366994, Log Avg loss: 1.38225139, Global Avg Loss: 1.57981196, Time: 0.0402 Steps: 69550, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001432, Sample Num: 22912, Cur Loss: 0.46416983, Cur Avg Loss: 1.25230823, Log Avg loss: 1.05867235, Global Avg Loss: 1.57973704, Time: 0.0646 Steps: 69560, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001442, Sample Num: 23072, Cur Loss: 2.23153377, Cur Avg Loss: 1.25303524, Log Avg loss: 1.35714318, Global Avg Loss: 1.57970504, Time: 0.0405 Steps: 69570, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001452, Sample Num: 23232, Cur Loss: 0.55187762, Cur Avg Loss: 1.25247667, Log Avg loss: 1.17193130, Global Avg Loss: 1.57964644, Time: 0.0769 Steps: 69580, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001462, Sample Num: 23392, Cur Loss: 2.12535405, Cur Avg Loss: 1.25238464, Log Avg loss: 1.23902207, Global Avg Loss: 1.57959749, Time: 0.0702 Steps: 69590, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001472, Sample Num: 23552, Cur Loss: 0.98350489, Cur Avg Loss: 1.25392611, Log Avg loss: 1.47928825, Global Avg Loss: 1.57958308, Time: 0.0403 Steps: 69600, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001482, Sample Num: 23712, Cur Loss: 1.18516636, Cur Avg Loss: 1.25673166, Log Avg loss: 1.66970913, Global Avg Loss: 1.57959603, Time: 0.0585 Steps: 69610, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001492, Sample Num: 23872, Cur Loss: 0.41316640, Cur Avg Loss: 1.25701463, Log Avg loss: 1.29895043, Global Avg Loss: 1.57955572, Time: 0.0480 Steps: 69620, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001502, Sample Num: 24032, Cur Loss: 1.21048224, Cur Avg Loss: 1.25725908, Log Avg loss: 1.29373218, Global Avg Loss: 1.57951467, Time: 0.0442 Steps: 69630, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001512, Sample Num: 24192, Cur Loss: 2.17821813, Cur Avg Loss: 1.25851631, Log Avg loss: 1.44735165, Global Avg Loss: 1.57949569, Time: 0.0479 Steps: 69640, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001522, Sample Num: 24352, Cur Loss: 0.52137816, Cur Avg Loss: 1.25655906, Log Avg loss: 0.96062340, Global Avg Loss: 1.57940683, Time: 0.0863 Steps: 69650, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001532, Sample Num: 24512, Cur Loss: 2.64237165, Cur Avg Loss: 1.25668392, Log Avg loss: 1.27568642, Global Avg Loss: 1.57936323, Time: 0.0403 Steps: 69660, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001542, Sample Num: 24672, Cur Loss: 1.27984881, Cur Avg Loss: 1.25611123, Log Avg loss: 1.16837628, Global Avg Loss: 1.57930424, Time: 0.0402 Steps: 69670, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001552, Sample Num: 24832, Cur Loss: 2.54884100, Cur Avg Loss: 1.25738152, Log Avg loss: 1.45325923, Global Avg Loss: 1.57928615, Time: 0.0403 Steps: 69680, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001562, Sample Num: 24992, Cur Loss: 1.61378515, Cur Avg Loss: 1.25785997, Log Avg loss: 1.33211503, Global Avg Loss: 1.57925069, Time: 0.0403 Steps: 69690, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001572, Sample Num: 25152, Cur Loss: 0.66438663, Cur Avg Loss: 1.25750947, Log Avg loss: 1.20276227, Global Avg Loss: 1.57919667, Time: 0.0403 Steps: 69700, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001582, Sample Num: 25312, Cur Loss: 1.76348507, Cur Avg Loss: 1.25808666, Log Avg loss: 1.34882082, Global Avg Loss: 1.57916362, Time: 0.0402 Steps: 69710, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001592, Sample Num: 25472, Cur Loss: 1.20444942, Cur Avg Loss: 1.25863484, Log Avg loss: 1.34535654, Global Avg Loss: 1.57913009, Time: 0.0402 Steps: 69720, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001602, Sample Num: 25632, Cur Loss: 0.30797988, Cur Avg Loss: 1.25758784, Log Avg loss: 1.09090644, Global Avg Loss: 1.57906007, Time: 0.0403 Steps: 69730, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001612, Sample Num: 25792, Cur Loss: 0.63780957, Cur Avg Loss: 1.25607745, Log Avg loss: 1.01411172, Global Avg Loss: 1.57897907, Time: 0.0402 Steps: 69740, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001622, Sample Num: 25952, Cur Loss: 1.28957534, Cur Avg Loss: 1.25756497, Log Avg loss: 1.49735450, Global Avg Loss: 1.57896736, Time: 0.0402 Steps: 69750, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001632, Sample Num: 26112, Cur Loss: 0.73357117, Cur Avg Loss: 1.25632764, Log Avg loss: 1.05563253, Global Avg Loss: 1.57889234, Time: 0.0403 Steps: 69760, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001642, Sample Num: 26272, Cur Loss: 0.88290614, Cur Avg Loss: 1.25660705, Log Avg loss: 1.30220690, Global Avg Loss: 1.57885269, Time: 0.0402 Steps: 69770, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001652, Sample Num: 26432, Cur Loss: 1.50869834, Cur Avg Loss: 1.25555053, Log Avg loss: 1.08206995, Global Avg Loss: 1.57878149, Time: 0.0402 Steps: 69780, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001662, Sample Num: 26592, Cur Loss: 1.67490041, Cur Avg Loss: 1.25633766, Log Avg loss: 1.38637146, Global Avg Loss: 1.57875392, Time: 0.0402 Steps: 69790, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001672, Sample Num: 26752, Cur Loss: 1.79650593, Cur Avg Loss: 1.25707162, Log Avg loss: 1.37905486, Global Avg Loss: 1.57872531, Time: 0.0402 Steps: 69800, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001682, Sample Num: 26912, Cur Loss: 1.11157036, Cur Avg Loss: 1.25577997, Log Avg loss: 1.03981636, Global Avg Loss: 1.57864812, Time: 0.0403 Steps: 69810, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001692, Sample Num: 27072, Cur Loss: 2.02577519, Cur Avg Loss: 1.25853078, Log Avg loss: 1.72121703, Global Avg Loss: 1.57866854, Time: 0.0402 Steps: 69820, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001702, Sample Num: 27232, Cur Loss: 1.48086488, Cur Avg Loss: 1.26021873, Log Avg loss: 1.54581996, Global Avg Loss: 1.57866383, Time: 0.0402 Steps: 69830, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001712, Sample Num: 27392, Cur Loss: 1.47695184, Cur Avg Loss: 1.25969939, Log Avg loss: 1.17130828, Global Avg Loss: 1.57860551, Time: 0.0402 Steps: 69840, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001722, Sample Num: 27552, Cur Loss: 1.09702384, Cur Avg Loss: 1.25958242, Log Avg loss: 1.23955612, Global Avg Loss: 1.57855697, Time: 0.0402 Steps: 69850, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001732, Sample Num: 27712, Cur Loss: 1.51592183, Cur Avg Loss: 1.25929380, Log Avg loss: 1.20959381, Global Avg Loss: 1.57850415, Time: 0.0402 Steps: 69860, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001742, Sample Num: 27872, Cur Loss: 0.63985252, Cur Avg Loss: 1.25750281, Log Avg loss: 0.94730304, Global Avg Loss: 1.57841381, Time: 0.0402 Steps: 69870, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001752, Sample Num: 28032, Cur Loss: 2.08739376, Cur Avg Loss: 1.25828876, Log Avg loss: 1.39520126, Global Avg Loss: 1.57838759, Time: 0.0402 Steps: 69880, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001762, Sample Num: 28192, Cur Loss: 0.65768367, Cur Avg Loss: 1.25852767, Log Avg loss: 1.30038589, Global Avg Loss: 1.57834782, Time: 0.0402 Steps: 69890, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001772, Sample Num: 28352, Cur Loss: 0.40398678, Cur Avg Loss: 1.25767688, Log Avg loss: 1.10776620, Global Avg Loss: 1.57828049, Time: 0.0402 Steps: 69900, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001782, Sample Num: 28512, Cur Loss: 1.11029041, Cur Avg Loss: 1.25622604, Log Avg loss: 0.99913860, Global Avg Loss: 1.57819765, Time: 0.0403 Steps: 69910, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001792, Sample Num: 28672, Cur Loss: 0.70639116, Cur Avg Loss: 1.25559929, Log Avg loss: 1.14391150, Global Avg Loss: 1.57813554, Time: 0.0743 Steps: 69920, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001802, Sample Num: 28832, Cur Loss: 1.84915650, Cur Avg Loss: 1.25365225, Log Avg loss: 0.90474350, Global Avg Loss: 1.57803925, Time: 0.0493 Steps: 69930, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001812, Sample Num: 28992, Cur Loss: 0.53153133, Cur Avg Loss: 1.25341855, Log Avg loss: 1.21130458, Global Avg Loss: 1.57798681, Time: 0.0593 Steps: 69940, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001822, Sample Num: 29152, Cur Loss: 1.07932878, Cur Avg Loss: 1.25385123, Log Avg loss: 1.33225326, Global Avg Loss: 1.57795168, Time: 0.0666 Steps: 69950, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001832, Sample Num: 29312, Cur Loss: 0.97752649, Cur Avg Loss: 1.25250934, Log Avg loss: 1.00801795, Global Avg Loss: 1.57787022, Time: 0.0602 Steps: 69960, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001842, Sample Num: 29472, Cur Loss: 2.19744253, Cur Avg Loss: 1.25218441, Log Avg loss: 1.19265616, Global Avg Loss: 1.57781516, Time: 0.0903 Steps: 69970, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001852, Sample Num: 29632, Cur Loss: 0.40084401, Cur Avg Loss: 1.25287546, Log Avg loss: 1.38016660, Global Avg Loss: 1.57778692, Time: 0.0410 Steps: 69980, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001862, Sample Num: 29792, Cur Loss: 1.56104302, Cur Avg Loss: 1.25406891, Log Avg loss: 1.47509617, Global Avg Loss: 1.57777225, Time: 0.0404 Steps: 69990, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001872, Sample Num: 29952, Cur Loss: 0.40718397, Cur Avg Loss: 1.25291883, Log Avg loss: 1.03877462, Global Avg Loss: 1.57769525, Time: 0.0403 Steps: 70000, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001882, Sample Num: 30112, Cur Loss: 1.27100885, Cur Avg Loss: 1.25425553, Log Avg loss: 1.50448472, Global Avg Loss: 1.57768479, Time: 0.0403 Steps: 70010, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001892, Sample Num: 30272, Cur Loss: 1.56805944, Cur Avg Loss: 1.25533996, Log Avg loss: 1.45943095, Global Avg Loss: 1.57766790, Time: 0.0403 Steps: 70020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001902, Sample Num: 30432, Cur Loss: 0.89827091, Cur Avg Loss: 1.25345612, Log Avg loss: 0.89703372, Global Avg Loss: 1.57757071, Time: 0.0402 Steps: 70030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001912, Sample Num: 30592, Cur Loss: 1.35705090, Cur Avg Loss: 1.25492129, Log Avg loss: 1.53359674, Global Avg Loss: 1.57756443, Time: 0.0403 Steps: 70040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001922, Sample Num: 30752, Cur Loss: 1.22234273, Cur Avg Loss: 1.25574429, Log Avg loss: 1.41310168, Global Avg Loss: 1.57754095, Time: 0.0403 Steps: 70050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001932, Sample Num: 30912, Cur Loss: 0.63152111, Cur Avg Loss: 1.25620608, Log Avg loss: 1.34496130, Global Avg Loss: 1.57750775, Time: 0.0402 Steps: 70060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001942, Sample Num: 31072, Cur Loss: 1.78890610, Cur Avg Loss: 1.25567114, Log Avg loss: 1.15232100, Global Avg Loss: 1.57744707, Time: 0.0405 Steps: 70070, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001952, Sample Num: 31232, Cur Loss: 0.74000573, Cur Avg Loss: 1.25439617, Log Avg loss: 1.00679716, Global Avg Loss: 1.57736565, Time: 0.0404 Steps: 70080, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001962, Sample Num: 31392, Cur Loss: 1.14130616, Cur Avg Loss: 1.25411714, Log Avg loss: 1.19964938, Global Avg Loss: 1.57731176, Time: 0.0402 Steps: 70090, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001972, Sample Num: 31552, Cur Loss: 0.32958877, Cur Avg Loss: 1.25381767, Log Avg loss: 1.19506271, Global Avg Loss: 1.57725723, Time: 0.0403 Steps: 70100, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001982, Sample Num: 31712, Cur Loss: 1.01268995, Cur Avg Loss: 1.25377949, Log Avg loss: 1.24625036, Global Avg Loss: 1.57721001, Time: 0.0403 Steps: 70110, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001992, Sample Num: 31872, Cur Loss: 0.42558706, Cur Avg Loss: 1.25347421, Log Avg loss: 1.19296702, Global Avg Loss: 1.57715522, Time: 0.0403 Steps: 70120, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002002, Sample Num: 32032, Cur Loss: 2.23088217, Cur Avg Loss: 1.25427048, Log Avg loss: 1.41288825, Global Avg Loss: 1.57713179, Time: 0.0403 Steps: 70130, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002012, Sample Num: 32192, Cur Loss: 1.29971528, Cur Avg Loss: 1.25335365, Log Avg loss: 1.06980471, Global Avg Loss: 1.57705946, Time: 0.0402 Steps: 70140, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002022, Sample Num: 32352, Cur Loss: 0.93681049, Cur Avg Loss: 1.25242398, Log Avg loss: 1.06537464, Global Avg Loss: 1.57698652, Time: 0.0402 Steps: 70150, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002032, Sample Num: 32512, Cur Loss: 0.96197689, Cur Avg Loss: 1.25113924, Log Avg loss: 0.99136471, Global Avg Loss: 1.57690305, Time: 0.0403 Steps: 70160, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002042, Sample Num: 32672, Cur Loss: 1.90721846, Cur Avg Loss: 1.25033320, Log Avg loss: 1.08654442, Global Avg Loss: 1.57683317, Time: 0.0402 Steps: 70170, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002052, Sample Num: 32832, Cur Loss: 1.87268448, Cur Avg Loss: 1.25256548, Log Avg loss: 1.70839847, Global Avg Loss: 1.57685192, Time: 0.0402 Steps: 70180, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002062, Sample Num: 32992, Cur Loss: 2.66091776, Cur Avg Loss: 1.25251632, Log Avg loss: 1.24242715, Global Avg Loss: 1.57680427, Time: 0.0403 Steps: 70190, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002072, Sample Num: 33152, Cur Loss: 1.84266639, Cur Avg Loss: 1.25275487, Log Avg loss: 1.30194489, Global Avg Loss: 1.57676512, Time: 0.0402 Steps: 70200, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002082, Sample Num: 33312, Cur Loss: 0.54656154, Cur Avg Loss: 1.25285969, Log Avg loss: 1.27457760, Global Avg Loss: 1.57672208, Time: 0.0402 Steps: 70210, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002092, Sample Num: 33472, Cur Loss: 1.16198862, Cur Avg Loss: 1.25357412, Log Avg loss: 1.40231907, Global Avg Loss: 1.57669724, Time: 0.0402 Steps: 70220, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002102, Sample Num: 33632, Cur Loss: 2.70544243, Cur Avg Loss: 1.25405283, Log Avg loss: 1.35419934, Global Avg Loss: 1.57666556, Time: 0.0403 Steps: 70230, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002112, Sample Num: 33792, Cur Loss: 1.45490944, Cur Avg Loss: 1.25286295, Log Avg loss: 1.00274998, Global Avg Loss: 1.57658385, Time: 0.0402 Steps: 70240, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002122, Sample Num: 33952, Cur Loss: 1.58043349, Cur Avg Loss: 1.25204050, Log Avg loss: 1.07833899, Global Avg Loss: 1.57651293, Time: 0.0402 Steps: 70250, Updated lr: 0.000034 ***** Running evaluation checkpoint-70257 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-70257 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.169591, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.203484, "eval_total_loss": 846.04908, "eval_mae": 0.921388, "eval_mse": 1.203503, "eval_r2": 0.234974, "eval_sp_statistic": 0.428927, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.530818, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.216085, "test_total_loss": 610.4748, "test_mae": 0.933011, "test_mse": 1.216275, "test_r2": 0.215005, "test_sp_statistic": 0.452785, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.563339, "test_ps_pvalue": 0.0, "lr": 3.432242769084875e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5764690729676987, "train_cur_epoch_loss": 2664.784504942596, "train_cur_epoch_avg_loss": 1.2516601714150286, "train_cur_epoch_time": 95.16959071159363, "train_cur_epoch_avg_time": 0.044701545660682775, "epoch": 33, "step": 70257} ################################################## Training, Epoch: 0034, Batch: 000003, Sample Num: 48, Cur Loss: 0.54076511, Cur Avg Loss: 0.94210571, Log Avg loss: 1.07808781, Global Avg Loss: 1.57644199, Time: 0.0402 Steps: 70260, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000013, Sample Num: 208, Cur Loss: 1.32387137, Cur Avg Loss: 1.10571904, Log Avg loss: 1.15480304, Global Avg Loss: 1.57638198, Time: 0.0402 Steps: 70270, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000023, Sample Num: 368, Cur Loss: 0.86255252, Cur Avg Loss: 1.19568487, Log Avg loss: 1.31264045, Global Avg Loss: 1.57634446, Time: 0.0402 Steps: 70280, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000033, Sample Num: 528, Cur Loss: 1.04881644, Cur Avg Loss: 1.23150129, Log Avg loss: 1.31387904, Global Avg Loss: 1.57630712, Time: 0.0403 Steps: 70290, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000043, Sample Num: 688, Cur Loss: 1.86285520, Cur Avg Loss: 1.23313517, Log Avg loss: 1.23852699, Global Avg Loss: 1.57625907, Time: 0.0402 Steps: 70300, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000053, Sample Num: 848, Cur Loss: 1.06845689, Cur Avg Loss: 1.25768619, Log Avg loss: 1.36325556, Global Avg Loss: 1.57622877, Time: 0.0401 Steps: 70310, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000063, Sample Num: 1008, Cur Loss: 2.83279371, Cur Avg Loss: 1.32299699, Log Avg loss: 1.66914427, Global Avg Loss: 1.57624199, Time: 0.0402 Steps: 70320, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000073, Sample Num: 1168, Cur Loss: 0.70521474, Cur Avg Loss: 1.33015294, Log Avg loss: 1.37523539, Global Avg Loss: 1.57621341, Time: 0.0806 Steps: 70330, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000083, Sample Num: 1328, Cur Loss: 1.71523571, Cur Avg Loss: 1.28688713, Log Avg loss: 0.97104670, Global Avg Loss: 1.57612737, Time: 0.0404 Steps: 70340, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000093, Sample Num: 1488, Cur Loss: 1.84203458, Cur Avg Loss: 1.31525102, Log Avg loss: 1.55067129, Global Avg Loss: 1.57612375, Time: 0.0533 Steps: 70350, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000103, Sample Num: 1648, Cur Loss: 2.51685119, Cur Avg Loss: 1.31072433, Log Avg loss: 1.26862617, Global Avg Loss: 1.57608005, Time: 0.0817 Steps: 70360, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000113, Sample Num: 1808, Cur Loss: 0.46553972, Cur Avg Loss: 1.30892806, Log Avg loss: 1.29042645, Global Avg Loss: 1.57603946, Time: 0.0410 Steps: 70370, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000123, Sample Num: 1968, Cur Loss: 1.29896355, Cur Avg Loss: 1.31096000, Log Avg loss: 1.33392097, Global Avg Loss: 1.57600505, Time: 0.0572 Steps: 70380, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000133, Sample Num: 2128, Cur Loss: 1.21413851, Cur Avg Loss: 1.28890189, Log Avg loss: 1.01758704, Global Avg Loss: 1.57592572, Time: 0.0653 Steps: 70390, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000143, Sample Num: 2288, Cur Loss: 1.93532801, Cur Avg Loss: 1.29814968, Log Avg loss: 1.42114533, Global Avg Loss: 1.57590374, Time: 0.0511 Steps: 70400, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000153, Sample Num: 2448, Cur Loss: 0.58853674, Cur Avg Loss: 1.29174761, Log Avg loss: 1.20019810, Global Avg Loss: 1.57585038, Time: 0.0990 Steps: 70410, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000163, Sample Num: 2608, Cur Loss: 1.27410054, Cur Avg Loss: 1.29528426, Log Avg loss: 1.34939495, Global Avg Loss: 1.57581822, Time: 0.0403 Steps: 70420, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000173, Sample Num: 2768, Cur Loss: 0.87912738, Cur Avg Loss: 1.28225200, Log Avg loss: 1.06982615, Global Avg Loss: 1.57574638, Time: 0.0402 Steps: 70430, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000183, Sample Num: 2928, Cur Loss: 0.96583903, Cur Avg Loss: 1.27089531, Log Avg loss: 1.07442459, Global Avg Loss: 1.57567521, Time: 0.0403 Steps: 70440, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000193, Sample Num: 3088, Cur Loss: 1.93631613, Cur Avg Loss: 1.27378574, Log Avg loss: 1.32668063, Global Avg Loss: 1.57563986, Time: 0.0403 Steps: 70450, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000203, Sample Num: 3248, Cur Loss: 0.92768180, Cur Avg Loss: 1.27737781, Log Avg loss: 1.34670473, Global Avg Loss: 1.57560737, Time: 0.0402 Steps: 70460, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000213, Sample Num: 3408, Cur Loss: 1.13649607, Cur Avg Loss: 1.27967644, Log Avg loss: 1.32633865, Global Avg Loss: 1.57557200, Time: 0.0403 Steps: 70470, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000223, Sample Num: 3568, Cur Loss: 1.43160689, Cur Avg Loss: 1.27732700, Log Avg loss: 1.22728395, Global Avg Loss: 1.57552258, Time: 0.0402 Steps: 70480, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000233, Sample Num: 3728, Cur Loss: 0.46993101, Cur Avg Loss: 1.27715981, Log Avg loss: 1.27343139, Global Avg Loss: 1.57547973, Time: 0.0402 Steps: 70490, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000243, Sample Num: 3888, Cur Loss: 1.08247316, Cur Avg Loss: 1.27058033, Log Avg loss: 1.11727837, Global Avg Loss: 1.57541473, Time: 0.0403 Steps: 70500, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000253, Sample Num: 4048, Cur Loss: 1.34119201, Cur Avg Loss: 1.25876157, Log Avg loss: 0.97156577, Global Avg Loss: 1.57532909, Time: 0.0403 Steps: 70510, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000263, Sample Num: 4208, Cur Loss: 1.44792581, Cur Avg Loss: 1.25238906, Log Avg loss: 1.09116458, Global Avg Loss: 1.57526044, Time: 0.0402 Steps: 70520, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000273, Sample Num: 4368, Cur Loss: 0.64340168, Cur Avg Loss: 1.26012770, Log Avg loss: 1.46365383, Global Avg Loss: 1.57524461, Time: 0.0402 Steps: 70530, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000283, Sample Num: 4528, Cur Loss: 0.38102126, Cur Avg Loss: 1.26211279, Log Avg loss: 1.31630586, Global Avg Loss: 1.57520790, Time: 0.0402 Steps: 70540, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000293, Sample Num: 4688, Cur Loss: 0.79340082, Cur Avg Loss: 1.26911467, Log Avg loss: 1.46726779, Global Avg Loss: 1.57519260, Time: 0.0402 Steps: 70550, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000303, Sample Num: 4848, Cur Loss: 1.99827003, Cur Avg Loss: 1.28270519, Log Avg loss: 1.68090752, Global Avg Loss: 1.57520759, Time: 0.0403 Steps: 70560, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000313, Sample Num: 5008, Cur Loss: 1.28216159, Cur Avg Loss: 1.28117785, Log Avg loss: 1.23489945, Global Avg Loss: 1.57515936, Time: 0.0402 Steps: 70570, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000323, Sample Num: 5168, Cur Loss: 1.13145435, Cur Avg Loss: 1.27143542, Log Avg loss: 0.96649729, Global Avg Loss: 1.57507313, Time: 0.0402 Steps: 70580, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000333, Sample Num: 5328, Cur Loss: 0.33338335, Cur Avg Loss: 1.26293344, Log Avg loss: 0.98831962, Global Avg Loss: 1.57499001, Time: 0.0402 Steps: 70590, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000343, Sample Num: 5488, Cur Loss: 1.21135736, Cur Avg Loss: 1.25807301, Log Avg loss: 1.09622056, Global Avg Loss: 1.57492219, Time: 0.0402 Steps: 70600, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000353, Sample Num: 5648, Cur Loss: 0.61081338, Cur Avg Loss: 1.25183612, Log Avg loss: 1.03791080, Global Avg Loss: 1.57484614, Time: 0.0403 Steps: 70610, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000363, Sample Num: 5808, Cur Loss: 0.52446657, Cur Avg Loss: 1.25528112, Log Avg loss: 1.37688956, Global Avg Loss: 1.57481811, Time: 0.0402 Steps: 70620, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000373, Sample Num: 5968, Cur Loss: 1.64879704, Cur Avg Loss: 1.24755096, Log Avg loss: 0.96694640, Global Avg Loss: 1.57473204, Time: 0.0402 Steps: 70630, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000383, Sample Num: 6128, Cur Loss: 1.36168718, Cur Avg Loss: 1.24971503, Log Avg loss: 1.33043483, Global Avg Loss: 1.57469746, Time: 0.0402 Steps: 70640, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000393, Sample Num: 6288, Cur Loss: 0.64774054, Cur Avg Loss: 1.24305118, Log Avg loss: 0.98782554, Global Avg Loss: 1.57461439, Time: 0.0402 Steps: 70650, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000403, Sample Num: 6448, Cur Loss: 1.94303942, Cur Avg Loss: 1.23667526, Log Avg loss: 0.98610163, Global Avg Loss: 1.57453110, Time: 0.0532 Steps: 70660, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000413, Sample Num: 6608, Cur Loss: 2.33116961, Cur Avg Loss: 1.23914574, Log Avg loss: 1.33870615, Global Avg Loss: 1.57449773, Time: 0.0635 Steps: 70670, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000423, Sample Num: 6768, Cur Loss: 1.65292490, Cur Avg Loss: 1.23649103, Log Avg loss: 1.12685149, Global Avg Loss: 1.57443440, Time: 0.0512 Steps: 70680, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000433, Sample Num: 6928, Cur Loss: 1.41431236, Cur Avg Loss: 1.24087617, Log Avg loss: 1.42636778, Global Avg Loss: 1.57441345, Time: 0.0669 Steps: 70690, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000443, Sample Num: 7088, Cur Loss: 1.37299180, Cur Avg Loss: 1.24164224, Log Avg loss: 1.27481271, Global Avg Loss: 1.57437108, Time: 0.0673 Steps: 70700, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000453, Sample Num: 7248, Cur Loss: 1.62481117, Cur Avg Loss: 1.24073956, Log Avg loss: 1.20075108, Global Avg Loss: 1.57431824, Time: 0.0561 Steps: 70710, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000463, Sample Num: 7408, Cur Loss: 0.40471914, Cur Avg Loss: 1.23739135, Log Avg loss: 1.08571756, Global Avg Loss: 1.57424915, Time: 0.0587 Steps: 70720, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000473, Sample Num: 7568, Cur Loss: 0.76023102, Cur Avg Loss: 1.24241735, Log Avg loss: 1.47512111, Global Avg Loss: 1.57423513, Time: 0.0686 Steps: 70730, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000483, Sample Num: 7728, Cur Loss: 1.35329676, Cur Avg Loss: 1.24142692, Log Avg loss: 1.19457934, Global Avg Loss: 1.57418147, Time: 0.0402 Steps: 70740, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000493, Sample Num: 7888, Cur Loss: 0.78792453, Cur Avg Loss: 1.23385356, Log Avg loss: 0.86806024, Global Avg Loss: 1.57408166, Time: 0.0402 Steps: 70750, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000503, Sample Num: 8048, Cur Loss: 1.54295909, Cur Avg Loss: 1.23689498, Log Avg loss: 1.38683695, Global Avg Loss: 1.57405520, Time: 0.0402 Steps: 70760, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000513, Sample Num: 8208, Cur Loss: 1.66941845, Cur Avg Loss: 1.24292781, Log Avg loss: 1.54637916, Global Avg Loss: 1.57405129, Time: 0.0402 Steps: 70770, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000523, Sample Num: 8368, Cur Loss: 1.45359671, Cur Avg Loss: 1.24020855, Log Avg loss: 1.10071063, Global Avg Loss: 1.57398441, Time: 0.0402 Steps: 70780, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000533, Sample Num: 8528, Cur Loss: 1.61595798, Cur Avg Loss: 1.23683410, Log Avg loss: 1.06035028, Global Avg Loss: 1.57391186, Time: 0.0403 Steps: 70790, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000543, Sample Num: 8688, Cur Loss: 1.32584381, Cur Avg Loss: 1.23679499, Log Avg loss: 1.23471065, Global Avg Loss: 1.57386395, Time: 0.0402 Steps: 70800, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000553, Sample Num: 8848, Cur Loss: 0.51308501, Cur Avg Loss: 1.23518951, Log Avg loss: 1.14801157, Global Avg Loss: 1.57380381, Time: 0.0403 Steps: 70810, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000563, Sample Num: 9008, Cur Loss: 2.34201241, Cur Avg Loss: 1.23321696, Log Avg loss: 1.12413517, Global Avg Loss: 1.57374031, Time: 0.0403 Steps: 70820, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000573, Sample Num: 9168, Cur Loss: 0.62016487, Cur Avg Loss: 1.23030532, Log Avg loss: 1.06638024, Global Avg Loss: 1.57366868, Time: 0.0402 Steps: 70830, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000583, Sample Num: 9328, Cur Loss: 2.51331210, Cur Avg Loss: 1.23183965, Log Avg loss: 1.31975640, Global Avg Loss: 1.57363284, Time: 0.0402 Steps: 70840, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000593, Sample Num: 9488, Cur Loss: 1.66070795, Cur Avg Loss: 1.23689849, Log Avg loss: 1.53182914, Global Avg Loss: 1.57362694, Time: 0.0402 Steps: 70850, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000603, Sample Num: 9648, Cur Loss: 1.17009783, Cur Avg Loss: 1.23465009, Log Avg loss: 1.10131982, Global Avg Loss: 1.57356028, Time: 0.0402 Steps: 70860, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000613, Sample Num: 9808, Cur Loss: 0.81713462, Cur Avg Loss: 1.23339808, Log Avg loss: 1.15790197, Global Avg Loss: 1.57350163, Time: 0.0402 Steps: 70870, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000623, Sample Num: 9968, Cur Loss: 0.90690053, Cur Avg Loss: 1.23114510, Log Avg loss: 1.09303718, Global Avg Loss: 1.57343385, Time: 0.0402 Steps: 70880, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000633, Sample Num: 10128, Cur Loss: 2.23650146, Cur Avg Loss: 1.23210727, Log Avg loss: 1.29205068, Global Avg Loss: 1.57339415, Time: 0.0403 Steps: 70890, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000643, Sample Num: 10288, Cur Loss: 0.65789580, Cur Avg Loss: 1.23273517, Log Avg loss: 1.27248116, Global Avg Loss: 1.57335171, Time: 0.0402 Steps: 70900, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000653, Sample Num: 10448, Cur Loss: 0.83171552, Cur Avg Loss: 1.22844227, Log Avg loss: 0.95240856, Global Avg Loss: 1.57326414, Time: 0.0402 Steps: 70910, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000663, Sample Num: 10608, Cur Loss: 0.65684128, Cur Avg Loss: 1.22593185, Log Avg loss: 1.06200170, Global Avg Loss: 1.57319205, Time: 0.0402 Steps: 70920, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000673, Sample Num: 10768, Cur Loss: 1.02485919, Cur Avg Loss: 1.22822613, Log Avg loss: 1.38033694, Global Avg Loss: 1.57316486, Time: 0.0403 Steps: 70930, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000683, Sample Num: 10928, Cur Loss: 1.81482959, Cur Avg Loss: 1.22775609, Log Avg loss: 1.19612251, Global Avg Loss: 1.57311172, Time: 0.0402 Steps: 70940, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000693, Sample Num: 11088, Cur Loss: 1.10994887, Cur Avg Loss: 1.22553727, Log Avg loss: 1.07399198, Global Avg Loss: 1.57304137, Time: 0.0402 Steps: 70950, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000703, Sample Num: 11248, Cur Loss: 0.67717159, Cur Avg Loss: 1.22773737, Log Avg loss: 1.38020423, Global Avg Loss: 1.57301419, Time: 0.0402 Steps: 70960, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000713, Sample Num: 11408, Cur Loss: 0.61373812, Cur Avg Loss: 1.22633387, Log Avg loss: 1.12766748, Global Avg Loss: 1.57295144, Time: 0.0402 Steps: 70970, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000723, Sample Num: 11568, Cur Loss: 1.27931094, Cur Avg Loss: 1.22905022, Log Avg loss: 1.42272593, Global Avg Loss: 1.57293028, Time: 0.0402 Steps: 70980, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000733, Sample Num: 11728, Cur Loss: 1.12536740, Cur Avg Loss: 1.22610204, Log Avg loss: 1.01294892, Global Avg Loss: 1.57285139, Time: 0.0403 Steps: 70990, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000743, Sample Num: 11888, Cur Loss: 1.41526151, Cur Avg Loss: 1.22990619, Log Avg loss: 1.50875003, Global Avg Loss: 1.57284237, Time: 0.0548 Steps: 71000, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000753, Sample Num: 12048, Cur Loss: 1.03106976, Cur Avg Loss: 1.22775459, Log Avg loss: 1.06789102, Global Avg Loss: 1.57277126, Time: 0.0882 Steps: 71010, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000763, Sample Num: 12208, Cur Loss: 0.65320027, Cur Avg Loss: 1.22670831, Log Avg loss: 1.14792336, Global Avg Loss: 1.57271143, Time: 0.0485 Steps: 71020, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000773, Sample Num: 12368, Cur Loss: 1.53523993, Cur Avg Loss: 1.22507097, Log Avg loss: 1.10014171, Global Avg Loss: 1.57264490, Time: 0.0416 Steps: 71030, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000783, Sample Num: 12528, Cur Loss: 0.80315959, Cur Avg Loss: 1.22513511, Log Avg loss: 1.23009305, Global Avg Loss: 1.57259668, Time: 0.0557 Steps: 71040, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000793, Sample Num: 12688, Cur Loss: 1.67998099, Cur Avg Loss: 1.22171324, Log Avg loss: 0.95378121, Global Avg Loss: 1.57250959, Time: 0.0819 Steps: 71050, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000803, Sample Num: 12848, Cur Loss: 0.36597598, Cur Avg Loss: 1.22435882, Log Avg loss: 1.43415326, Global Avg Loss: 1.57249012, Time: 0.0799 Steps: 71060, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000813, Sample Num: 13008, Cur Loss: 1.24909079, Cur Avg Loss: 1.22655601, Log Avg loss: 1.40299004, Global Avg Loss: 1.57246627, Time: 0.0672 Steps: 71070, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000823, Sample Num: 13168, Cur Loss: 2.02169752, Cur Avg Loss: 1.23132640, Log Avg loss: 1.61915952, Global Avg Loss: 1.57247284, Time: 0.0487 Steps: 71080, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000833, Sample Num: 13328, Cur Loss: 0.41188067, Cur Avg Loss: 1.23010893, Log Avg loss: 1.12991135, Global Avg Loss: 1.57241058, Time: 0.0404 Steps: 71090, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000843, Sample Num: 13488, Cur Loss: 2.01049137, Cur Avg Loss: 1.23041928, Log Avg loss: 1.25627077, Global Avg Loss: 1.57236612, Time: 0.0402 Steps: 71100, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000853, Sample Num: 13648, Cur Loss: 2.25140715, Cur Avg Loss: 1.22962377, Log Avg loss: 1.16256292, Global Avg Loss: 1.57230849, Time: 0.0403 Steps: 71110, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000863, Sample Num: 13808, Cur Loss: 0.99060035, Cur Avg Loss: 1.23029507, Log Avg loss: 1.28755687, Global Avg Loss: 1.57226845, Time: 0.0402 Steps: 71120, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000873, Sample Num: 13968, Cur Loss: 0.92458522, Cur Avg Loss: 1.23199895, Log Avg loss: 1.37904315, Global Avg Loss: 1.57224129, Time: 0.0403 Steps: 71130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000883, Sample Num: 14128, Cur Loss: 1.12803841, Cur Avg Loss: 1.23108522, Log Avg loss: 1.15131664, Global Avg Loss: 1.57218212, Time: 0.0403 Steps: 71140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000893, Sample Num: 14288, Cur Loss: 1.32845426, Cur Avg Loss: 1.22951207, Log Avg loss: 1.09060313, Global Avg Loss: 1.57211443, Time: 0.0402 Steps: 71150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000903, Sample Num: 14448, Cur Loss: 1.20273209, Cur Avg Loss: 1.22965781, Log Avg loss: 1.24267223, Global Avg Loss: 1.57206814, Time: 0.0402 Steps: 71160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000913, Sample Num: 14608, Cur Loss: 1.33516979, Cur Avg Loss: 1.22873719, Log Avg loss: 1.14560537, Global Avg Loss: 1.57200822, Time: 0.0403 Steps: 71170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000923, Sample Num: 14768, Cur Loss: 2.21778107, Cur Avg Loss: 1.23038730, Log Avg loss: 1.38104207, Global Avg Loss: 1.57198139, Time: 0.0402 Steps: 71180, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000933, Sample Num: 14928, Cur Loss: 1.65330505, Cur Avg Loss: 1.23300981, Log Avg loss: 1.47506814, Global Avg Loss: 1.57196777, Time: 0.0403 Steps: 71190, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000943, Sample Num: 15088, Cur Loss: 0.78497654, Cur Avg Loss: 1.23511021, Log Avg loss: 1.43107691, Global Avg Loss: 1.57194799, Time: 0.0402 Steps: 71200, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000953, Sample Num: 15248, Cur Loss: 0.85057867, Cur Avg Loss: 1.23061183, Log Avg loss: 0.80641537, Global Avg Loss: 1.57184048, Time: 0.0402 Steps: 71210, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000963, Sample Num: 15408, Cur Loss: 0.26753953, Cur Avg Loss: 1.22775631, Log Avg loss: 0.95562450, Global Avg Loss: 1.57175396, Time: 0.0403 Steps: 71220, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000973, Sample Num: 15568, Cur Loss: 0.31215173, Cur Avg Loss: 1.22509579, Log Avg loss: 0.96888819, Global Avg Loss: 1.57166932, Time: 0.0403 Steps: 71230, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000983, Sample Num: 15728, Cur Loss: 1.48782086, Cur Avg Loss: 1.22748278, Log Avg loss: 1.45973644, Global Avg Loss: 1.57165361, Time: 0.0402 Steps: 71240, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000993, Sample Num: 15888, Cur Loss: 2.21613932, Cur Avg Loss: 1.22687103, Log Avg loss: 1.16673675, Global Avg Loss: 1.57159678, Time: 0.0403 Steps: 71250, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001003, Sample Num: 16048, Cur Loss: 0.58543849, Cur Avg Loss: 1.22674963, Log Avg loss: 1.21469406, Global Avg Loss: 1.57154670, Time: 0.0402 Steps: 71260, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001013, Sample Num: 16208, Cur Loss: 1.03364229, Cur Avg Loss: 1.22515905, Log Avg loss: 1.06562358, Global Avg Loss: 1.57147571, Time: 0.0403 Steps: 71270, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001023, Sample Num: 16368, Cur Loss: 0.97240257, Cur Avg Loss: 1.22505654, Log Avg loss: 1.21467273, Global Avg Loss: 1.57142565, Time: 0.0403 Steps: 71280, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001033, Sample Num: 16528, Cur Loss: 1.12581491, Cur Avg Loss: 1.22445360, Log Avg loss: 1.16277283, Global Avg Loss: 1.57136833, Time: 0.0403 Steps: 71290, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001043, Sample Num: 16688, Cur Loss: 0.76982784, Cur Avg Loss: 1.22451792, Log Avg loss: 1.23116243, Global Avg Loss: 1.57132062, Time: 0.0402 Steps: 71300, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001053, Sample Num: 16848, Cur Loss: 0.54231495, Cur Avg Loss: 1.22374555, Log Avg loss: 1.14318708, Global Avg Loss: 1.57126058, Time: 0.0402 Steps: 71310, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001063, Sample Num: 17008, Cur Loss: 1.07494164, Cur Avg Loss: 1.22326894, Log Avg loss: 1.17308185, Global Avg Loss: 1.57120475, Time: 0.0402 Steps: 71320, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001073, Sample Num: 17168, Cur Loss: 1.20084107, Cur Avg Loss: 1.22245307, Log Avg loss: 1.13572649, Global Avg Loss: 1.57114370, Time: 0.0403 Steps: 71330, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001083, Sample Num: 17328, Cur Loss: 1.38203287, Cur Avg Loss: 1.22394363, Log Avg loss: 1.38388085, Global Avg Loss: 1.57111745, Time: 0.0402 Steps: 71340, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001093, Sample Num: 17488, Cur Loss: 1.21764708, Cur Avg Loss: 1.22322692, Log Avg loss: 1.14560667, Global Avg Loss: 1.57105781, Time: 0.0402 Steps: 71350, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001103, Sample Num: 17648, Cur Loss: 0.62854564, Cur Avg Loss: 1.22177448, Log Avg loss: 1.06302278, Global Avg Loss: 1.57098662, Time: 0.0403 Steps: 71360, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001113, Sample Num: 17808, Cur Loss: 1.94005322, Cur Avg Loss: 1.22073038, Log Avg loss: 1.10556661, Global Avg Loss: 1.57092140, Time: 0.0589 Steps: 71370, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001123, Sample Num: 17968, Cur Loss: 1.60257173, Cur Avg Loss: 1.22301156, Log Avg loss: 1.47690603, Global Avg Loss: 1.57090823, Time: 0.0480 Steps: 71380, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001133, Sample Num: 18128, Cur Loss: 1.42005396, Cur Avg Loss: 1.22531629, Log Avg loss: 1.48413808, Global Avg Loss: 1.57089608, Time: 0.0405 Steps: 71390, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001143, Sample Num: 18288, Cur Loss: 1.77494478, Cur Avg Loss: 1.22772745, Log Avg loss: 1.50091228, Global Avg Loss: 1.57088628, Time: 0.0404 Steps: 71400, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001153, Sample Num: 18448, Cur Loss: 0.84773201, Cur Avg Loss: 1.22845691, Log Avg loss: 1.31183322, Global Avg Loss: 1.57085000, Time: 0.0737 Steps: 71410, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001163, Sample Num: 18608, Cur Loss: 0.93379307, Cur Avg Loss: 1.22818812, Log Avg loss: 1.19719712, Global Avg Loss: 1.57079768, Time: 0.1000 Steps: 71420, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001173, Sample Num: 18768, Cur Loss: 1.88591313, Cur Avg Loss: 1.22929854, Log Avg loss: 1.35844071, Global Avg Loss: 1.57076795, Time: 0.0417 Steps: 71430, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001183, Sample Num: 18928, Cur Loss: 0.73421514, Cur Avg Loss: 1.22742529, Log Avg loss: 1.00769277, Global Avg Loss: 1.57068913, Time: 0.0499 Steps: 71440, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001193, Sample Num: 19088, Cur Loss: 0.87701327, Cur Avg Loss: 1.22975564, Log Avg loss: 1.50543647, Global Avg Loss: 1.57068000, Time: 0.0405 Steps: 71450, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001203, Sample Num: 19248, Cur Loss: 1.75363994, Cur Avg Loss: 1.22927852, Log Avg loss: 1.17235758, Global Avg Loss: 1.57062426, Time: 0.0405 Steps: 71460, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001213, Sample Num: 19408, Cur Loss: 1.38338780, Cur Avg Loss: 1.22892241, Log Avg loss: 1.18608242, Global Avg Loss: 1.57057046, Time: 0.0403 Steps: 71470, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001223, Sample Num: 19568, Cur Loss: 2.26712370, Cur Avg Loss: 1.22896010, Log Avg loss: 1.23353148, Global Avg Loss: 1.57052331, Time: 0.0402 Steps: 71480, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001233, Sample Num: 19728, Cur Loss: 1.07942700, Cur Avg Loss: 1.23095890, Log Avg loss: 1.47541262, Global Avg Loss: 1.57051000, Time: 0.0402 Steps: 71490, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001243, Sample Num: 19888, Cur Loss: 1.82238913, Cur Avg Loss: 1.23277031, Log Avg loss: 1.45611727, Global Avg Loss: 1.57049400, Time: 0.0402 Steps: 71500, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001253, Sample Num: 20048, Cur Loss: 0.94636494, Cur Avg Loss: 1.23121732, Log Avg loss: 1.03818057, Global Avg Loss: 1.57041956, Time: 0.0402 Steps: 71510, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001263, Sample Num: 20208, Cur Loss: 1.80367827, Cur Avg Loss: 1.23457236, Log Avg loss: 1.65495859, Global Avg Loss: 1.57043138, Time: 0.0403 Steps: 71520, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001273, Sample Num: 20368, Cur Loss: 0.86527288, Cur Avg Loss: 1.23261486, Log Avg loss: 0.98538239, Global Avg Loss: 1.57034959, Time: 0.0403 Steps: 71530, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001283, Sample Num: 20528, Cur Loss: 0.88903487, Cur Avg Loss: 1.23110204, Log Avg loss: 1.03852069, Global Avg Loss: 1.57027525, Time: 0.0402 Steps: 71540, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001293, Sample Num: 20688, Cur Loss: 1.16018677, Cur Avg Loss: 1.23255407, Log Avg loss: 1.41884921, Global Avg Loss: 1.57025409, Time: 0.0402 Steps: 71550, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001303, Sample Num: 20848, Cur Loss: 1.80138791, Cur Avg Loss: 1.23452799, Log Avg loss: 1.48975654, Global Avg Loss: 1.57024284, Time: 0.0402 Steps: 71560, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001313, Sample Num: 21008, Cur Loss: 0.88976389, Cur Avg Loss: 1.23393430, Log Avg loss: 1.15657606, Global Avg Loss: 1.57018504, Time: 0.0402 Steps: 71570, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001323, Sample Num: 21168, Cur Loss: 1.87994730, Cur Avg Loss: 1.23440371, Log Avg loss: 1.29603749, Global Avg Loss: 1.57014674, Time: 0.0402 Steps: 71580, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001333, Sample Num: 21328, Cur Loss: 1.29836893, Cur Avg Loss: 1.23397961, Log Avg loss: 1.17787035, Global Avg Loss: 1.57009195, Time: 0.0402 Steps: 71590, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001343, Sample Num: 21488, Cur Loss: 0.69856238, Cur Avg Loss: 1.23478769, Log Avg loss: 1.34250547, Global Avg Loss: 1.57006016, Time: 0.0402 Steps: 71600, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001353, Sample Num: 21648, Cur Loss: 1.47419143, Cur Avg Loss: 1.23313562, Log Avg loss: 1.01126179, Global Avg Loss: 1.56998213, Time: 0.0404 Steps: 71610, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001363, Sample Num: 21808, Cur Loss: 1.08418965, Cur Avg Loss: 1.23502325, Log Avg loss: 1.49042050, Global Avg Loss: 1.56997102, Time: 0.0402 Steps: 71620, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001373, Sample Num: 21968, Cur Loss: 1.54677749, Cur Avg Loss: 1.23542720, Log Avg loss: 1.29048460, Global Avg Loss: 1.56993200, Time: 0.0402 Steps: 71630, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001383, Sample Num: 22128, Cur Loss: 1.67988575, Cur Avg Loss: 1.23695385, Log Avg loss: 1.44656388, Global Avg Loss: 1.56991478, Time: 0.0402 Steps: 71640, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001393, Sample Num: 22288, Cur Loss: 0.57185161, Cur Avg Loss: 1.23596940, Log Avg loss: 1.09982018, Global Avg Loss: 1.56984917, Time: 0.0403 Steps: 71650, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001403, Sample Num: 22448, Cur Loss: 1.19827759, Cur Avg Loss: 1.23613509, Log Avg loss: 1.25921560, Global Avg Loss: 1.56980582, Time: 0.0402 Steps: 71660, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001413, Sample Num: 22608, Cur Loss: 0.59122133, Cur Avg Loss: 1.23699656, Log Avg loss: 1.35786000, Global Avg Loss: 1.56977625, Time: 0.0402 Steps: 71670, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001423, Sample Num: 22768, Cur Loss: 1.72604942, Cur Avg Loss: 1.23818020, Log Avg loss: 1.40542929, Global Avg Loss: 1.56975332, Time: 0.0402 Steps: 71680, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001433, Sample Num: 22928, Cur Loss: 1.29589045, Cur Avg Loss: 1.23676597, Log Avg loss: 1.03552021, Global Avg Loss: 1.56967880, Time: 0.0402 Steps: 71690, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001443, Sample Num: 23088, Cur Loss: 0.56214708, Cur Avg Loss: 1.23606943, Log Avg loss: 1.13625544, Global Avg Loss: 1.56961835, Time: 0.0402 Steps: 71700, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001453, Sample Num: 23248, Cur Loss: 1.01205993, Cur Avg Loss: 1.23528124, Log Avg loss: 1.12154558, Global Avg Loss: 1.56955587, Time: 0.0685 Steps: 71710, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001463, Sample Num: 23408, Cur Loss: 0.69132328, Cur Avg Loss: 1.23294396, Log Avg loss: 0.89333713, Global Avg Loss: 1.56946158, Time: 0.0407 Steps: 71720, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001473, Sample Num: 23568, Cur Loss: 1.01080704, Cur Avg Loss: 1.23418588, Log Avg loss: 1.41587927, Global Avg Loss: 1.56944017, Time: 0.0616 Steps: 71730, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001483, Sample Num: 23728, Cur Loss: 1.13989019, Cur Avg Loss: 1.23290924, Log Avg loss: 1.04485943, Global Avg Loss: 1.56936705, Time: 0.0620 Steps: 71740, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001493, Sample Num: 23888, Cur Loss: 1.65681934, Cur Avg Loss: 1.23453866, Log Avg loss: 1.47618236, Global Avg Loss: 1.56935406, Time: 0.0481 Steps: 71750, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001503, Sample Num: 24048, Cur Loss: 1.85943556, Cur Avg Loss: 1.23551386, Log Avg loss: 1.38111072, Global Avg Loss: 1.56932783, Time: 0.0427 Steps: 71760, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001513, Sample Num: 24208, Cur Loss: 1.51690936, Cur Avg Loss: 1.23511341, Log Avg loss: 1.17492639, Global Avg Loss: 1.56927288, Time: 0.0997 Steps: 71770, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001523, Sample Num: 24368, Cur Loss: 0.46555990, Cur Avg Loss: 1.23423588, Log Avg loss: 1.10146486, Global Avg Loss: 1.56920770, Time: 0.0643 Steps: 71780, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001533, Sample Num: 24528, Cur Loss: 1.69082248, Cur Avg Loss: 1.23243077, Log Avg loss: 0.95751280, Global Avg Loss: 1.56912250, Time: 0.0609 Steps: 71790, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001543, Sample Num: 24688, Cur Loss: 0.44515491, Cur Avg Loss: 1.22963581, Log Avg loss: 0.80116782, Global Avg Loss: 1.56901554, Time: 0.0403 Steps: 71800, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001553, Sample Num: 24848, Cur Loss: 1.40897000, Cur Avg Loss: 1.22974381, Log Avg loss: 1.24640872, Global Avg Loss: 1.56897061, Time: 0.0403 Steps: 71810, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001563, Sample Num: 25008, Cur Loss: 0.86834341, Cur Avg Loss: 1.22863164, Log Avg loss: 1.05591227, Global Avg Loss: 1.56889918, Time: 0.0403 Steps: 71820, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001573, Sample Num: 25168, Cur Loss: 1.70793521, Cur Avg Loss: 1.22881693, Log Avg loss: 1.25777709, Global Avg Loss: 1.56885586, Time: 0.0402 Steps: 71830, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001583, Sample Num: 25328, Cur Loss: 1.49341571, Cur Avg Loss: 1.22871046, Log Avg loss: 1.21196260, Global Avg Loss: 1.56880618, Time: 0.0402 Steps: 71840, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001593, Sample Num: 25488, Cur Loss: 1.40804625, Cur Avg Loss: 1.22809972, Log Avg loss: 1.13142056, Global Avg Loss: 1.56874531, Time: 0.0402 Steps: 71850, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001603, Sample Num: 25648, Cur Loss: 0.72337431, Cur Avg Loss: 1.22885121, Log Avg loss: 1.34856319, Global Avg Loss: 1.56871467, Time: 0.0402 Steps: 71860, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001613, Sample Num: 25808, Cur Loss: 0.66117322, Cur Avg Loss: 1.22717252, Log Avg loss: 0.95807818, Global Avg Loss: 1.56862971, Time: 0.0402 Steps: 71870, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001623, Sample Num: 25968, Cur Loss: 0.70496583, Cur Avg Loss: 1.22647821, Log Avg loss: 1.11448616, Global Avg Loss: 1.56856652, Time: 0.0402 Steps: 71880, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001633, Sample Num: 26128, Cur Loss: 1.50221217, Cur Avg Loss: 1.22532694, Log Avg loss: 1.03847595, Global Avg Loss: 1.56849279, Time: 0.0403 Steps: 71890, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001643, Sample Num: 26288, Cur Loss: 0.57726026, Cur Avg Loss: 1.22464314, Log Avg loss: 1.11297832, Global Avg Loss: 1.56842943, Time: 0.0402 Steps: 71900, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001653, Sample Num: 26448, Cur Loss: 1.31519365, Cur Avg Loss: 1.22504107, Log Avg loss: 1.29042154, Global Avg Loss: 1.56839077, Time: 0.0402 Steps: 71910, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001663, Sample Num: 26608, Cur Loss: 2.18387961, Cur Avg Loss: 1.22553050, Log Avg loss: 1.30643201, Global Avg Loss: 1.56835435, Time: 0.0403 Steps: 71920, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001673, Sample Num: 26768, Cur Loss: 0.57324862, Cur Avg Loss: 1.22563492, Log Avg loss: 1.24300048, Global Avg Loss: 1.56830912, Time: 0.0402 Steps: 71930, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001683, Sample Num: 26928, Cur Loss: 0.89472878, Cur Avg Loss: 1.22438689, Log Avg loss: 1.01559091, Global Avg Loss: 1.56823229, Time: 0.0402 Steps: 71940, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001693, Sample Num: 27088, Cur Loss: 0.54993302, Cur Avg Loss: 1.22348991, Log Avg loss: 1.07252862, Global Avg Loss: 1.56816339, Time: 0.0403 Steps: 71950, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001703, Sample Num: 27248, Cur Loss: 2.46808362, Cur Avg Loss: 1.22531940, Log Avg loss: 1.53505273, Global Avg Loss: 1.56815879, Time: 0.0402 Steps: 71960, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001713, Sample Num: 27408, Cur Loss: 1.75108838, Cur Avg Loss: 1.22528173, Log Avg loss: 1.21886557, Global Avg Loss: 1.56811026, Time: 0.0402 Steps: 71970, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001723, Sample Num: 27568, Cur Loss: 0.55982405, Cur Avg Loss: 1.22528637, Log Avg loss: 1.22608204, Global Avg Loss: 1.56806274, Time: 0.0402 Steps: 71980, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001733, Sample Num: 27728, Cur Loss: 1.70163810, Cur Avg Loss: 1.22540653, Log Avg loss: 1.24610918, Global Avg Loss: 1.56801802, Time: 0.0403 Steps: 71990, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001743, Sample Num: 27888, Cur Loss: 1.71553206, Cur Avg Loss: 1.22570308, Log Avg loss: 1.27709609, Global Avg Loss: 1.56797761, Time: 0.0402 Steps: 72000, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001753, Sample Num: 28048, Cur Loss: 0.78092945, Cur Avg Loss: 1.22475210, Log Avg loss: 1.05899572, Global Avg Loss: 1.56790693, Time: 0.0402 Steps: 72010, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001763, Sample Num: 28208, Cur Loss: 1.22313809, Cur Avg Loss: 1.22570259, Log Avg loss: 1.39232445, Global Avg Loss: 1.56788255, Time: 0.0402 Steps: 72020, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001773, Sample Num: 28368, Cur Loss: 0.74146396, Cur Avg Loss: 1.22482776, Log Avg loss: 1.07059471, Global Avg Loss: 1.56781351, Time: 0.0402 Steps: 72030, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001783, Sample Num: 28528, Cur Loss: 1.09128892, Cur Avg Loss: 1.22690510, Log Avg loss: 1.59521778, Global Avg Loss: 1.56781732, Time: 0.0403 Steps: 72040, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001793, Sample Num: 28688, Cur Loss: 1.12731135, Cur Avg Loss: 1.22764250, Log Avg loss: 1.35912088, Global Avg Loss: 1.56778835, Time: 0.0403 Steps: 72050, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001803, Sample Num: 28848, Cur Loss: 1.28006005, Cur Avg Loss: 1.22851205, Log Avg loss: 1.38442234, Global Avg Loss: 1.56776290, Time: 0.0402 Steps: 72060, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001813, Sample Num: 29008, Cur Loss: 1.34066117, Cur Avg Loss: 1.22969975, Log Avg loss: 1.44384204, Global Avg Loss: 1.56774571, Time: 0.0815 Steps: 72070, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001823, Sample Num: 29168, Cur Loss: 0.29085231, Cur Avg Loss: 1.22995666, Log Avg loss: 1.27653489, Global Avg Loss: 1.56770531, Time: 0.0625 Steps: 72080, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001833, Sample Num: 29328, Cur Loss: 2.88399863, Cur Avg Loss: 1.22932984, Log Avg loss: 1.11505932, Global Avg Loss: 1.56764252, Time: 0.0948 Steps: 72090, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001843, Sample Num: 29488, Cur Loss: 0.79443121, Cur Avg Loss: 1.22811701, Log Avg loss: 1.00580507, Global Avg Loss: 1.56756460, Time: 0.0786 Steps: 72100, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001853, Sample Num: 29648, Cur Loss: 1.59996426, Cur Avg Loss: 1.22809442, Log Avg loss: 1.22393251, Global Avg Loss: 1.56751694, Time: 0.0432 Steps: 72110, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001863, Sample Num: 29808, Cur Loss: 0.77678150, Cur Avg Loss: 1.22964950, Log Avg loss: 1.51780539, Global Avg Loss: 1.56751005, Time: 0.0659 Steps: 72120, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001873, Sample Num: 29968, Cur Loss: 0.63393044, Cur Avg Loss: 1.22777782, Log Avg loss: 0.87908361, Global Avg Loss: 1.56741461, Time: 0.0652 Steps: 72130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001883, Sample Num: 30128, Cur Loss: 0.50419295, Cur Avg Loss: 1.22657517, Log Avg loss: 1.00131855, Global Avg Loss: 1.56733613, Time: 0.0823 Steps: 72140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001893, Sample Num: 30288, Cur Loss: 1.40984535, Cur Avg Loss: 1.22727337, Log Avg loss: 1.35874385, Global Avg Loss: 1.56730722, Time: 0.0480 Steps: 72150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001903, Sample Num: 30448, Cur Loss: 0.88841856, Cur Avg Loss: 1.22681077, Log Avg loss: 1.13924153, Global Avg Loss: 1.56724790, Time: 0.0403 Steps: 72160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001913, Sample Num: 30608, Cur Loss: 0.45278615, Cur Avg Loss: 1.22752638, Log Avg loss: 1.36370590, Global Avg Loss: 1.56721970, Time: 0.0404 Steps: 72170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001923, Sample Num: 30768, Cur Loss: 0.95065504, Cur Avg Loss: 1.22884751, Log Avg loss: 1.48157988, Global Avg Loss: 1.56720783, Time: 0.0402 Steps: 72180, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001933, Sample Num: 30928, Cur Loss: 0.89190698, Cur Avg Loss: 1.22819510, Log Avg loss: 1.10273727, Global Avg Loss: 1.56714349, Time: 0.0402 Steps: 72190, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001943, Sample Num: 31088, Cur Loss: 0.87482655, Cur Avg Loss: 1.22781803, Log Avg loss: 1.15492954, Global Avg Loss: 1.56708640, Time: 0.0402 Steps: 72200, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001953, Sample Num: 31248, Cur Loss: 1.03287232, Cur Avg Loss: 1.22668906, Log Avg loss: 1.00733193, Global Avg Loss: 1.56700888, Time: 0.0404 Steps: 72210, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001963, Sample Num: 31408, Cur Loss: 0.75558430, Cur Avg Loss: 1.22704434, Log Avg loss: 1.29643008, Global Avg Loss: 1.56697142, Time: 0.0402 Steps: 72220, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001973, Sample Num: 31568, Cur Loss: 1.02828145, Cur Avg Loss: 1.22640635, Log Avg loss: 1.10116923, Global Avg Loss: 1.56690693, Time: 0.0403 Steps: 72230, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001983, Sample Num: 31728, Cur Loss: 2.55582714, Cur Avg Loss: 1.22815566, Log Avg loss: 1.57329340, Global Avg Loss: 1.56690781, Time: 0.0403 Steps: 72240, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001993, Sample Num: 31888, Cur Loss: 1.04773057, Cur Avg Loss: 1.22699927, Log Avg loss: 0.99768714, Global Avg Loss: 1.56682903, Time: 0.0402 Steps: 72250, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002003, Sample Num: 32048, Cur Loss: 2.05032635, Cur Avg Loss: 1.22662073, Log Avg loss: 1.15117842, Global Avg Loss: 1.56677151, Time: 0.0402 Steps: 72260, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002013, Sample Num: 32208, Cur Loss: 0.52741963, Cur Avg Loss: 1.22578671, Log Avg loss: 1.05873213, Global Avg Loss: 1.56670121, Time: 0.0403 Steps: 72270, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002023, Sample Num: 32368, Cur Loss: 1.22213221, Cur Avg Loss: 1.22569207, Log Avg loss: 1.20664107, Global Avg Loss: 1.56665139, Time: 0.0404 Steps: 72280, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002033, Sample Num: 32528, Cur Loss: 1.49838006, Cur Avg Loss: 1.22643487, Log Avg loss: 1.37670292, Global Avg Loss: 1.56662512, Time: 0.0402 Steps: 72290, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002043, Sample Num: 32688, Cur Loss: 2.05470657, Cur Avg Loss: 1.22718125, Log Avg loss: 1.37892109, Global Avg Loss: 1.56659916, Time: 0.0402 Steps: 72300, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002053, Sample Num: 32848, Cur Loss: 0.67842597, Cur Avg Loss: 1.22715502, Log Avg loss: 1.22179514, Global Avg Loss: 1.56655147, Time: 0.0403 Steps: 72310, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002063, Sample Num: 33008, Cur Loss: 0.65736639, Cur Avg Loss: 1.22796059, Log Avg loss: 1.39334471, Global Avg Loss: 1.56652752, Time: 0.0402 Steps: 72320, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002073, Sample Num: 33168, Cur Loss: 1.78350878, Cur Avg Loss: 1.22822852, Log Avg loss: 1.28350292, Global Avg Loss: 1.56648839, Time: 0.0402 Steps: 72330, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002083, Sample Num: 33328, Cur Loss: 0.79845083, Cur Avg Loss: 1.22889909, Log Avg loss: 1.36790882, Global Avg Loss: 1.56646094, Time: 0.0402 Steps: 72340, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002093, Sample Num: 33488, Cur Loss: 1.33665276, Cur Avg Loss: 1.22995828, Log Avg loss: 1.45058565, Global Avg Loss: 1.56644493, Time: 0.0402 Steps: 72350, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002103, Sample Num: 33648, Cur Loss: 1.62237239, Cur Avg Loss: 1.22996555, Log Avg loss: 1.23148890, Global Avg Loss: 1.56639863, Time: 0.0402 Steps: 72360, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002113, Sample Num: 33808, Cur Loss: 0.88995290, Cur Avg Loss: 1.22903294, Log Avg loss: 1.03290367, Global Avg Loss: 1.56632492, Time: 0.0402 Steps: 72370, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002123, Sample Num: 33968, Cur Loss: 1.39990568, Cur Avg Loss: 1.22914893, Log Avg loss: 1.25365842, Global Avg Loss: 1.56628172, Time: 0.0402 Steps: 72380, Updated lr: 0.000032 ***** Running evaluation checkpoint-72386 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-72386 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.067034, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.259782, "eval_total_loss": 885.626739, "eval_mae": 0.97018, "eval_mse": 1.259695, "eval_r2": 0.199255, "eval_sp_statistic": 0.442093, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.535699, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.294127, "test_total_loss": 649.651653, "test_mae": 0.999324, "test_mse": 1.294253, "test_r2": 0.164678, "test_sp_statistic": 0.460398, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.572346, "test_ps_pvalue": 0.0, "lr": 3.2303461356092936e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.56626740458733, "train_cur_epoch_loss": 2617.8446889668703, "train_cur_epoch_avg_loss": 1.2296123480351668, "train_cur_epoch_time": 95.06703400611877, "train_cur_epoch_avg_time": 0.044653374357030894, "epoch": 34, "step": 72386} ################################################## Training, Epoch: 0035, Batch: 000004, Sample Num: 64, Cur Loss: 1.86260378, Cur Avg Loss: 1.76168838, Log Avg loss: 1.54082630, Global Avg Loss: 1.56627820, Time: 0.0402 Steps: 72390, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000014, Sample Num: 224, Cur Loss: 0.84881914, Cur Avg Loss: 1.22653518, Log Avg loss: 1.01247391, Global Avg Loss: 1.56620171, Time: 0.0402 Steps: 72400, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000024, Sample Num: 384, Cur Loss: 0.65726817, Cur Avg Loss: 1.13539125, Log Avg loss: 1.00778973, Global Avg Loss: 1.56612459, Time: 0.0402 Steps: 72410, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000034, Sample Num: 544, Cur Loss: 0.72955430, Cur Avg Loss: 1.15181393, Log Avg loss: 1.19122838, Global Avg Loss: 1.56607283, Time: 0.0402 Steps: 72420, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000044, Sample Num: 704, Cur Loss: 1.90939212, Cur Avg Loss: 1.20229177, Log Avg loss: 1.37391642, Global Avg Loss: 1.56604630, Time: 0.0402 Steps: 72430, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000054, Sample Num: 864, Cur Loss: 0.60743797, Cur Avg Loss: 1.18639579, Log Avg loss: 1.11645346, Global Avg Loss: 1.56598423, Time: 0.0401 Steps: 72440, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000064, Sample Num: 1024, Cur Loss: 1.27274311, Cur Avg Loss: 1.24035749, Log Avg loss: 1.53175066, Global Avg Loss: 1.56597951, Time: 0.0402 Steps: 72450, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000074, Sample Num: 1184, Cur Loss: 1.51994300, Cur Avg Loss: 1.25210690, Log Avg loss: 1.32730312, Global Avg Loss: 1.56594657, Time: 0.0404 Steps: 72460, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000084, Sample Num: 1344, Cur Loss: 0.95329654, Cur Avg Loss: 1.26696483, Log Avg loss: 1.37691354, Global Avg Loss: 1.56592048, Time: 0.0407 Steps: 72470, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000094, Sample Num: 1504, Cur Loss: 0.36549175, Cur Avg Loss: 1.25297132, Log Avg loss: 1.13542588, Global Avg Loss: 1.56586109, Time: 0.0631 Steps: 72480, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000104, Sample Num: 1664, Cur Loss: 0.76126629, Cur Avg Loss: 1.23264235, Log Avg loss: 1.04154995, Global Avg Loss: 1.56578876, Time: 0.0404 Steps: 72490, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000114, Sample Num: 1824, Cur Loss: 1.53812790, Cur Avg Loss: 1.23029417, Log Avg loss: 1.20587308, Global Avg Loss: 1.56573912, Time: 0.0497 Steps: 72500, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000124, Sample Num: 1984, Cur Loss: 1.60890782, Cur Avg Loss: 1.21962533, Log Avg loss: 1.09800065, Global Avg Loss: 1.56567461, Time: 0.0617 Steps: 72510, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000134, Sample Num: 2144, Cur Loss: 1.38516998, Cur Avg Loss: 1.20371093, Log Avg loss: 1.00637239, Global Avg Loss: 1.56559749, Time: 0.0403 Steps: 72520, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000144, Sample Num: 2304, Cur Loss: 2.07461452, Cur Avg Loss: 1.22925648, Log Avg loss: 1.57156680, Global Avg Loss: 1.56559831, Time: 0.0583 Steps: 72530, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000154, Sample Num: 2464, Cur Loss: 0.78548229, Cur Avg Loss: 1.22225693, Log Avg loss: 1.12146335, Global Avg Loss: 1.56553708, Time: 0.0432 Steps: 72540, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000164, Sample Num: 2624, Cur Loss: 1.61308408, Cur Avg Loss: 1.21812038, Log Avg loss: 1.15441753, Global Avg Loss: 1.56548041, Time: 0.0402 Steps: 72550, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000174, Sample Num: 2784, Cur Loss: 1.48088396, Cur Avg Loss: 1.25000497, Log Avg loss: 1.77291226, Global Avg Loss: 1.56550900, Time: 0.0403 Steps: 72560, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000184, Sample Num: 2944, Cur Loss: 1.41996181, Cur Avg Loss: 1.24102583, Log Avg loss: 1.08478875, Global Avg Loss: 1.56544276, Time: 0.0402 Steps: 72570, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000194, Sample Num: 3104, Cur Loss: 0.69886112, Cur Avg Loss: 1.25357445, Log Avg loss: 1.48446906, Global Avg Loss: 1.56543160, Time: 0.0402 Steps: 72580, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000204, Sample Num: 3264, Cur Loss: 0.83025014, Cur Avg Loss: 1.25445387, Log Avg loss: 1.27151477, Global Avg Loss: 1.56539111, Time: 0.0403 Steps: 72590, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000214, Sample Num: 3424, Cur Loss: 0.42356342, Cur Avg Loss: 1.25274806, Log Avg loss: 1.21794938, Global Avg Loss: 1.56534326, Time: 0.0402 Steps: 72600, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000224, Sample Num: 3584, Cur Loss: 1.83110344, Cur Avg Loss: 1.24624274, Log Avg loss: 1.10702908, Global Avg Loss: 1.56528014, Time: 0.0402 Steps: 72610, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000234, Sample Num: 3744, Cur Loss: 2.07950211, Cur Avg Loss: 1.24397485, Log Avg loss: 1.19317407, Global Avg Loss: 1.56522890, Time: 0.0402 Steps: 72620, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000244, Sample Num: 3904, Cur Loss: 0.74295050, Cur Avg Loss: 1.24401384, Log Avg loss: 1.24492614, Global Avg Loss: 1.56518480, Time: 0.0402 Steps: 72630, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000254, Sample Num: 4064, Cur Loss: 0.72382259, Cur Avg Loss: 1.23844621, Log Avg loss: 1.10259592, Global Avg Loss: 1.56512111, Time: 0.0402 Steps: 72640, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000264, Sample Num: 4224, Cur Loss: 0.56247234, Cur Avg Loss: 1.25561256, Log Avg loss: 1.69163803, Global Avg Loss: 1.56513853, Time: 0.0402 Steps: 72650, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000274, Sample Num: 4384, Cur Loss: 1.11791492, Cur Avg Loss: 1.25041400, Log Avg loss: 1.11317200, Global Avg Loss: 1.56507633, Time: 0.0402 Steps: 72660, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000284, Sample Num: 4544, Cur Loss: 1.00350547, Cur Avg Loss: 1.25358860, Log Avg loss: 1.34057268, Global Avg Loss: 1.56504543, Time: 0.0402 Steps: 72670, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000294, Sample Num: 4704, Cur Loss: 1.25003552, Cur Avg Loss: 1.24244460, Log Avg loss: 0.92595501, Global Avg Loss: 1.56495750, Time: 0.0402 Steps: 72680, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000304, Sample Num: 4864, Cur Loss: 1.47555590, Cur Avg Loss: 1.24400590, Log Avg loss: 1.28990796, Global Avg Loss: 1.56491966, Time: 0.0402 Steps: 72690, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000314, Sample Num: 5024, Cur Loss: 1.43246579, Cur Avg Loss: 1.24364676, Log Avg loss: 1.23272898, Global Avg Loss: 1.56487397, Time: 0.0402 Steps: 72700, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000324, Sample Num: 5184, Cur Loss: 0.71513754, Cur Avg Loss: 1.24484624, Log Avg loss: 1.28250974, Global Avg Loss: 1.56483513, Time: 0.0402 Steps: 72710, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000334, Sample Num: 5344, Cur Loss: 1.17209947, Cur Avg Loss: 1.24653243, Log Avg loss: 1.30116507, Global Avg Loss: 1.56479887, Time: 0.0402 Steps: 72720, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000344, Sample Num: 5504, Cur Loss: 0.90805715, Cur Avg Loss: 1.24751727, Log Avg loss: 1.28041109, Global Avg Loss: 1.56475977, Time: 0.0402 Steps: 72730, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000354, Sample Num: 5664, Cur Loss: 1.84435225, Cur Avg Loss: 1.24644347, Log Avg loss: 1.20950460, Global Avg Loss: 1.56471093, Time: 0.0401 Steps: 72740, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000364, Sample Num: 5824, Cur Loss: 0.68065661, Cur Avg Loss: 1.24986409, Log Avg loss: 1.37095398, Global Avg Loss: 1.56468430, Time: 0.0402 Steps: 72750, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000374, Sample Num: 5984, Cur Loss: 0.37187433, Cur Avg Loss: 1.24337767, Log Avg loss: 1.00727201, Global Avg Loss: 1.56460769, Time: 0.0402 Steps: 72760, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000384, Sample Num: 6144, Cur Loss: 0.13858125, Cur Avg Loss: 1.23514475, Log Avg loss: 0.92723358, Global Avg Loss: 1.56452010, Time: 0.0401 Steps: 72770, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000394, Sample Num: 6304, Cur Loss: 0.31969225, Cur Avg Loss: 1.23880896, Log Avg loss: 1.37951458, Global Avg Loss: 1.56449468, Time: 0.0402 Steps: 72780, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000404, Sample Num: 6464, Cur Loss: 1.45207882, Cur Avg Loss: 1.23854370, Log Avg loss: 1.22809263, Global Avg Loss: 1.56444847, Time: 0.0402 Steps: 72790, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000414, Sample Num: 6624, Cur Loss: 0.99182999, Cur Avg Loss: 1.23618247, Log Avg loss: 1.14078876, Global Avg Loss: 1.56439027, Time: 0.0402 Steps: 72800, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000424, Sample Num: 6784, Cur Loss: 1.61884427, Cur Avg Loss: 1.24813948, Log Avg loss: 1.74315965, Global Avg Loss: 1.56441483, Time: 0.0401 Steps: 72810, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000434, Sample Num: 6944, Cur Loss: 0.84758151, Cur Avg Loss: 1.24448874, Log Avg loss: 1.08969738, Global Avg Loss: 1.56434964, Time: 0.0543 Steps: 72820, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000444, Sample Num: 7104, Cur Loss: 1.32350206, Cur Avg Loss: 1.24627070, Log Avg loss: 1.32360777, Global Avg Loss: 1.56431658, Time: 0.0696 Steps: 72830, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000454, Sample Num: 7264, Cur Loss: 1.58351719, Cur Avg Loss: 1.23970902, Log Avg loss: 0.94837027, Global Avg Loss: 1.56423202, Time: 0.0418 Steps: 72840, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000464, Sample Num: 7424, Cur Loss: 0.97488874, Cur Avg Loss: 1.24658904, Log Avg loss: 1.55894222, Global Avg Loss: 1.56423129, Time: 0.0736 Steps: 72850, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000474, Sample Num: 7584, Cur Loss: 0.20515859, Cur Avg Loss: 1.24593483, Log Avg loss: 1.21557913, Global Avg Loss: 1.56418344, Time: 0.0403 Steps: 72860, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000484, Sample Num: 7744, Cur Loss: 1.39174640, Cur Avg Loss: 1.24532645, Log Avg loss: 1.21648926, Global Avg Loss: 1.56413573, Time: 0.0848 Steps: 72870, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000494, Sample Num: 7904, Cur Loss: 0.58975357, Cur Avg Loss: 1.24334565, Log Avg loss: 1.14747527, Global Avg Loss: 1.56407856, Time: 0.0452 Steps: 72880, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000504, Sample Num: 8064, Cur Loss: 1.14118433, Cur Avg Loss: 1.24253929, Log Avg loss: 1.20270509, Global Avg Loss: 1.56402898, Time: 0.0593 Steps: 72890, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000514, Sample Num: 8224, Cur Loss: 0.84625530, Cur Avg Loss: 1.23845823, Log Avg loss: 1.03277277, Global Avg Loss: 1.56395610, Time: 0.0406 Steps: 72900, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000524, Sample Num: 8384, Cur Loss: 1.52201807, Cur Avg Loss: 1.23769455, Log Avg loss: 1.19844149, Global Avg Loss: 1.56390597, Time: 0.0403 Steps: 72910, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000534, Sample Num: 8544, Cur Loss: 2.08913445, Cur Avg Loss: 1.23901438, Log Avg loss: 1.30817343, Global Avg Loss: 1.56387090, Time: 0.0403 Steps: 72920, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000544, Sample Num: 8704, Cur Loss: 1.25902438, Cur Avg Loss: 1.24138067, Log Avg loss: 1.36774023, Global Avg Loss: 1.56384401, Time: 0.0402 Steps: 72930, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000554, Sample Num: 8864, Cur Loss: 0.90850490, Cur Avg Loss: 1.24009243, Log Avg loss: 1.17001221, Global Avg Loss: 1.56379001, Time: 0.0402 Steps: 72940, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000564, Sample Num: 9024, Cur Loss: 1.15517151, Cur Avg Loss: 1.23899406, Log Avg loss: 1.17814441, Global Avg Loss: 1.56373715, Time: 0.0403 Steps: 72950, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000574, Sample Num: 9184, Cur Loss: 0.73382157, Cur Avg Loss: 1.23674002, Log Avg loss: 1.10961220, Global Avg Loss: 1.56367491, Time: 0.0403 Steps: 72960, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000584, Sample Num: 9344, Cur Loss: 1.54528928, Cur Avg Loss: 1.23620406, Log Avg loss: 1.20543994, Global Avg Loss: 1.56362581, Time: 0.0403 Steps: 72970, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000594, Sample Num: 9504, Cur Loss: 0.29998517, Cur Avg Loss: 1.23202863, Log Avg loss: 0.98818377, Global Avg Loss: 1.56354696, Time: 0.0402 Steps: 72980, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000604, Sample Num: 9664, Cur Loss: 0.73361611, Cur Avg Loss: 1.23138577, Log Avg loss: 1.19319973, Global Avg Loss: 1.56349622, Time: 0.0402 Steps: 72990, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000614, Sample Num: 9824, Cur Loss: 0.93005359, Cur Avg Loss: 1.22776881, Log Avg loss: 1.00930429, Global Avg Loss: 1.56342031, Time: 0.0402 Steps: 73000, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000624, Sample Num: 9984, Cur Loss: 0.96416670, Cur Avg Loss: 1.22436445, Log Avg loss: 1.01533664, Global Avg Loss: 1.56334524, Time: 0.0402 Steps: 73010, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000634, Sample Num: 10144, Cur Loss: 2.49806380, Cur Avg Loss: 1.22204054, Log Avg loss: 1.07702902, Global Avg Loss: 1.56327864, Time: 0.0402 Steps: 73020, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000644, Sample Num: 10304, Cur Loss: 1.91395545, Cur Avg Loss: 1.22295566, Log Avg loss: 1.28097419, Global Avg Loss: 1.56323998, Time: 0.0403 Steps: 73030, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000654, Sample Num: 10464, Cur Loss: 1.19566607, Cur Avg Loss: 1.21632045, Log Avg loss: 0.78901306, Global Avg Loss: 1.56313398, Time: 0.0402 Steps: 73040, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000664, Sample Num: 10624, Cur Loss: 0.63789314, Cur Avg Loss: 1.21445011, Log Avg loss: 1.09212957, Global Avg Loss: 1.56306950, Time: 0.0402 Steps: 73050, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000674, Sample Num: 10784, Cur Loss: 2.53990507, Cur Avg Loss: 1.21605892, Log Avg loss: 1.32288397, Global Avg Loss: 1.56303663, Time: 0.0403 Steps: 73060, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000684, Sample Num: 10944, Cur Loss: 2.81374645, Cur Avg Loss: 1.21380094, Log Avg loss: 1.06161308, Global Avg Loss: 1.56296801, Time: 0.0402 Steps: 73070, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000694, Sample Num: 11104, Cur Loss: 0.25401476, Cur Avg Loss: 1.20966197, Log Avg loss: 0.92655672, Global Avg Loss: 1.56288092, Time: 0.0403 Steps: 73080, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000704, Sample Num: 11264, Cur Loss: 1.24545836, Cur Avg Loss: 1.21005081, Log Avg loss: 1.23703629, Global Avg Loss: 1.56283634, Time: 0.0402 Steps: 73090, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000714, Sample Num: 11424, Cur Loss: 2.44353914, Cur Avg Loss: 1.20836888, Log Avg loss: 1.08996080, Global Avg Loss: 1.56277165, Time: 0.0402 Steps: 73100, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000724, Sample Num: 11584, Cur Loss: 1.64489901, Cur Avg Loss: 1.20736695, Log Avg loss: 1.13582910, Global Avg Loss: 1.56271325, Time: 0.0402 Steps: 73110, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000734, Sample Num: 11744, Cur Loss: 0.53955710, Cur Avg Loss: 1.20607592, Log Avg loss: 1.11260552, Global Avg Loss: 1.56265170, Time: 0.0402 Steps: 73120, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000744, Sample Num: 11904, Cur Loss: 0.29225144, Cur Avg Loss: 1.20480322, Log Avg loss: 1.11138717, Global Avg Loss: 1.56258999, Time: 0.0402 Steps: 73130, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000754, Sample Num: 12064, Cur Loss: 1.07875991, Cur Avg Loss: 1.20528658, Log Avg loss: 1.24124784, Global Avg Loss: 1.56254605, Time: 0.0402 Steps: 73140, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000764, Sample Num: 12224, Cur Loss: 0.87687290, Cur Avg Loss: 1.20704104, Log Avg loss: 1.33932757, Global Avg Loss: 1.56251554, Time: 0.0402 Steps: 73150, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000774, Sample Num: 12384, Cur Loss: 1.42398274, Cur Avg Loss: 1.20496078, Log Avg loss: 1.04602885, Global Avg Loss: 1.56244494, Time: 0.0402 Steps: 73160, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000784, Sample Num: 12544, Cur Loss: 1.57426357, Cur Avg Loss: 1.20598772, Log Avg loss: 1.28547319, Global Avg Loss: 1.56240709, Time: 0.0402 Steps: 73170, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000794, Sample Num: 12704, Cur Loss: 2.32878733, Cur Avg Loss: 1.20543469, Log Avg loss: 1.16207679, Global Avg Loss: 1.56235238, Time: 0.0404 Steps: 73180, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000804, Sample Num: 12864, Cur Loss: 0.78363657, Cur Avg Loss: 1.20860619, Log Avg loss: 1.46042320, Global Avg Loss: 1.56233846, Time: 0.0604 Steps: 73190, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000814, Sample Num: 13024, Cur Loss: 0.75621581, Cur Avg Loss: 1.20603197, Log Avg loss: 0.99906490, Global Avg Loss: 1.56226151, Time: 0.0783 Steps: 73200, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000824, Sample Num: 13184, Cur Loss: 0.70725954, Cur Avg Loss: 1.20202624, Log Avg loss: 0.87595974, Global Avg Loss: 1.56216776, Time: 0.0448 Steps: 73210, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000834, Sample Num: 13344, Cur Loss: 1.26000023, Cur Avg Loss: 1.20163050, Log Avg loss: 1.16902192, Global Avg Loss: 1.56211407, Time: 0.0764 Steps: 73220, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000844, Sample Num: 13504, Cur Loss: 0.96714175, Cur Avg Loss: 1.20518542, Log Avg loss: 1.50166593, Global Avg Loss: 1.56210582, Time: 0.0501 Steps: 73230, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000854, Sample Num: 13664, Cur Loss: 1.47931969, Cur Avg Loss: 1.20810414, Log Avg loss: 1.45444379, Global Avg Loss: 1.56209112, Time: 0.0552 Steps: 73240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000864, Sample Num: 13824, Cur Loss: 1.08962977, Cur Avg Loss: 1.20565059, Log Avg loss: 0.99611768, Global Avg Loss: 1.56201385, Time: 0.0581 Steps: 73250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000874, Sample Num: 13984, Cur Loss: 3.16598344, Cur Avg Loss: 1.20901058, Log Avg loss: 1.49931320, Global Avg Loss: 1.56200529, Time: 0.0516 Steps: 73260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000884, Sample Num: 14144, Cur Loss: 1.25288785, Cur Avg Loss: 1.21067825, Log Avg loss: 1.35643276, Global Avg Loss: 1.56197723, Time: 0.0403 Steps: 73270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000894, Sample Num: 14304, Cur Loss: 0.40210611, Cur Avg Loss: 1.20800498, Log Avg loss: 0.97168831, Global Avg Loss: 1.56189668, Time: 0.0402 Steps: 73280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000904, Sample Num: 14464, Cur Loss: 1.41529560, Cur Avg Loss: 1.20865998, Log Avg loss: 1.26721645, Global Avg Loss: 1.56185647, Time: 0.0403 Steps: 73290, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000914, Sample Num: 14624, Cur Loss: 0.89207089, Cur Avg Loss: 1.20818722, Log Avg loss: 1.16544961, Global Avg Loss: 1.56180239, Time: 0.0403 Steps: 73300, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000924, Sample Num: 14784, Cur Loss: 1.92005348, Cur Avg Loss: 1.20806836, Log Avg loss: 1.19720475, Global Avg Loss: 1.56175266, Time: 0.0402 Steps: 73310, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000934, Sample Num: 14944, Cur Loss: 2.06670570, Cur Avg Loss: 1.20694905, Log Avg loss: 1.10352464, Global Avg Loss: 1.56169016, Time: 0.0403 Steps: 73320, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000944, Sample Num: 15104, Cur Loss: 0.90417892, Cur Avg Loss: 1.20598516, Log Avg loss: 1.11595842, Global Avg Loss: 1.56162938, Time: 0.0402 Steps: 73330, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000954, Sample Num: 15264, Cur Loss: 0.89996594, Cur Avg Loss: 1.20830975, Log Avg loss: 1.42775114, Global Avg Loss: 1.56161112, Time: 0.0403 Steps: 73340, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000964, Sample Num: 15424, Cur Loss: 2.22929716, Cur Avg Loss: 1.20991281, Log Avg loss: 1.36284404, Global Avg Loss: 1.56158403, Time: 0.0403 Steps: 73350, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000974, Sample Num: 15584, Cur Loss: 1.89479101, Cur Avg Loss: 1.21192781, Log Avg loss: 1.40617402, Global Avg Loss: 1.56156284, Time: 0.0402 Steps: 73360, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000984, Sample Num: 15744, Cur Loss: 3.33174634, Cur Avg Loss: 1.21481882, Log Avg loss: 1.49640360, Global Avg Loss: 1.56155396, Time: 0.0402 Steps: 73370, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000994, Sample Num: 15904, Cur Loss: 1.28340518, Cur Avg Loss: 1.21470359, Log Avg loss: 1.20336438, Global Avg Loss: 1.56150515, Time: 0.0402 Steps: 73380, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001004, Sample Num: 16064, Cur Loss: 0.89652371, Cur Avg Loss: 1.21361577, Log Avg loss: 1.10548636, Global Avg Loss: 1.56144301, Time: 0.0403 Steps: 73390, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001014, Sample Num: 16224, Cur Loss: 1.20226717, Cur Avg Loss: 1.21595563, Log Avg loss: 1.45087802, Global Avg Loss: 1.56142795, Time: 0.0403 Steps: 73400, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001024, Sample Num: 16384, Cur Loss: 1.03791201, Cur Avg Loss: 1.21292710, Log Avg loss: 0.90583391, Global Avg Loss: 1.56133864, Time: 0.0403 Steps: 73410, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001034, Sample Num: 16544, Cur Loss: 1.70930648, Cur Avg Loss: 1.21348032, Log Avg loss: 1.27013047, Global Avg Loss: 1.56129898, Time: 0.0402 Steps: 73420, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001044, Sample Num: 16704, Cur Loss: 0.72819066, Cur Avg Loss: 1.21513064, Log Avg loss: 1.38577319, Global Avg Loss: 1.56127507, Time: 0.0402 Steps: 73430, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001054, Sample Num: 16864, Cur Loss: 0.59551001, Cur Avg Loss: 1.21596449, Log Avg loss: 1.30301908, Global Avg Loss: 1.56123991, Time: 0.0402 Steps: 73440, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001064, Sample Num: 17024, Cur Loss: 0.37991151, Cur Avg Loss: 1.21489312, Log Avg loss: 1.10196994, Global Avg Loss: 1.56117738, Time: 0.0402 Steps: 73450, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001074, Sample Num: 17184, Cur Loss: 0.80988336, Cur Avg Loss: 1.21444430, Log Avg loss: 1.16669010, Global Avg Loss: 1.56112368, Time: 0.0402 Steps: 73460, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001084, Sample Num: 17344, Cur Loss: 2.06965780, Cur Avg Loss: 1.21470546, Log Avg loss: 1.24275448, Global Avg Loss: 1.56108035, Time: 0.0402 Steps: 73470, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001094, Sample Num: 17504, Cur Loss: 1.69956243, Cur Avg Loss: 1.21422818, Log Avg loss: 1.16249053, Global Avg Loss: 1.56102610, Time: 0.0402 Steps: 73480, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001104, Sample Num: 17664, Cur Loss: 1.27905726, Cur Avg Loss: 1.21563528, Log Avg loss: 1.36957266, Global Avg Loss: 1.56100005, Time: 0.0402 Steps: 73490, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001114, Sample Num: 17824, Cur Loss: 1.31582689, Cur Avg Loss: 1.21637393, Log Avg loss: 1.29792022, Global Avg Loss: 1.56096426, Time: 0.0402 Steps: 73500, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001124, Sample Num: 17984, Cur Loss: 2.49588847, Cur Avg Loss: 1.21673099, Log Avg loss: 1.25650754, Global Avg Loss: 1.56092284, Time: 0.0402 Steps: 73510, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001134, Sample Num: 18144, Cur Loss: 1.18479669, Cur Avg Loss: 1.21769138, Log Avg loss: 1.32563921, Global Avg Loss: 1.56089084, Time: 0.0403 Steps: 73520, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001144, Sample Num: 18304, Cur Loss: 1.78906047, Cur Avg Loss: 1.21879044, Log Avg loss: 1.34342344, Global Avg Loss: 1.56086126, Time: 0.0403 Steps: 73530, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001154, Sample Num: 18464, Cur Loss: 1.28691316, Cur Avg Loss: 1.21743617, Log Avg loss: 1.06250787, Global Avg Loss: 1.56079350, Time: 0.0992 Steps: 73540, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001164, Sample Num: 18624, Cur Loss: 1.54271102, Cur Avg Loss: 1.21815421, Log Avg loss: 1.30101597, Global Avg Loss: 1.56075818, Time: 0.0438 Steps: 73550, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001174, Sample Num: 18784, Cur Loss: 0.99181354, Cur Avg Loss: 1.21960221, Log Avg loss: 1.38815012, Global Avg Loss: 1.56073471, Time: 0.0844 Steps: 73560, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001184, Sample Num: 18944, Cur Loss: 0.58270323, Cur Avg Loss: 1.22078303, Log Avg loss: 1.35941053, Global Avg Loss: 1.56070735, Time: 0.0407 Steps: 73570, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001194, Sample Num: 19104, Cur Loss: 1.45787656, Cur Avg Loss: 1.21847037, Log Avg loss: 0.94465141, Global Avg Loss: 1.56062362, Time: 0.0643 Steps: 73580, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001204, Sample Num: 19264, Cur Loss: 1.82051182, Cur Avg Loss: 1.22062885, Log Avg loss: 1.47835226, Global Avg Loss: 1.56061244, Time: 0.0402 Steps: 73590, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001214, Sample Num: 19424, Cur Loss: 0.55387175, Cur Avg Loss: 1.22098231, Log Avg loss: 1.26353799, Global Avg Loss: 1.56057208, Time: 0.0560 Steps: 73600, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001224, Sample Num: 19584, Cur Loss: 1.38376379, Cur Avg Loss: 1.22068557, Log Avg loss: 1.18466202, Global Avg Loss: 1.56052101, Time: 0.0638 Steps: 73610, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001234, Sample Num: 19744, Cur Loss: 0.54402298, Cur Avg Loss: 1.21957624, Log Avg loss: 1.08379417, Global Avg Loss: 1.56045625, Time: 0.0754 Steps: 73620, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001244, Sample Num: 19904, Cur Loss: 1.43232250, Cur Avg Loss: 1.21904147, Log Avg loss: 1.15305104, Global Avg Loss: 1.56040092, Time: 0.0403 Steps: 73630, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001254, Sample Num: 20064, Cur Loss: 0.48656398, Cur Avg Loss: 1.21728032, Log Avg loss: 0.99819315, Global Avg Loss: 1.56032458, Time: 0.0402 Steps: 73640, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001264, Sample Num: 20224, Cur Loss: 0.62267172, Cur Avg Loss: 1.21835188, Log Avg loss: 1.35272501, Global Avg Loss: 1.56029639, Time: 0.0403 Steps: 73650, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001274, Sample Num: 20384, Cur Loss: 0.88405275, Cur Avg Loss: 1.21837180, Log Avg loss: 1.22088935, Global Avg Loss: 1.56025031, Time: 0.0402 Steps: 73660, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001284, Sample Num: 20544, Cur Loss: 1.81572759, Cur Avg Loss: 1.21708605, Log Avg loss: 1.05328233, Global Avg Loss: 1.56018150, Time: 0.0403 Steps: 73670, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001294, Sample Num: 20704, Cur Loss: 2.21734118, Cur Avg Loss: 1.21794195, Log Avg loss: 1.32783934, Global Avg Loss: 1.56014996, Time: 0.0403 Steps: 73680, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001304, Sample Num: 20864, Cur Loss: 1.43290019, Cur Avg Loss: 1.21586759, Log Avg loss: 0.94744472, Global Avg Loss: 1.56006682, Time: 0.0403 Steps: 73690, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001314, Sample Num: 21024, Cur Loss: 1.25334525, Cur Avg Loss: 1.21519054, Log Avg loss: 1.12690373, Global Avg Loss: 1.56000804, Time: 0.0403 Steps: 73700, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001324, Sample Num: 21184, Cur Loss: 2.37808681, Cur Avg Loss: 1.22025042, Log Avg loss: 1.88511882, Global Avg Loss: 1.56005215, Time: 0.0403 Steps: 73710, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001334, Sample Num: 21344, Cur Loss: 1.68969882, Cur Avg Loss: 1.22158599, Log Avg loss: 1.39841539, Global Avg Loss: 1.56003022, Time: 0.0403 Steps: 73720, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001344, Sample Num: 21504, Cur Loss: 0.91518712, Cur Avg Loss: 1.22248773, Log Avg loss: 1.34277998, Global Avg Loss: 1.56000076, Time: 0.0403 Steps: 73730, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001354, Sample Num: 21664, Cur Loss: 1.90712762, Cur Avg Loss: 1.22178285, Log Avg loss: 1.12704728, Global Avg Loss: 1.55994204, Time: 0.0404 Steps: 73740, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001364, Sample Num: 21824, Cur Loss: 1.02667105, Cur Avg Loss: 1.22307644, Log Avg loss: 1.39822853, Global Avg Loss: 1.55992012, Time: 0.0403 Steps: 73750, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001374, Sample Num: 21984, Cur Loss: 1.11749315, Cur Avg Loss: 1.22036707, Log Avg loss: 0.85080842, Global Avg Loss: 1.55982398, Time: 0.0403 Steps: 73760, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001384, Sample Num: 22144, Cur Loss: 1.72444284, Cur Avg Loss: 1.22240186, Log Avg loss: 1.50198275, Global Avg Loss: 1.55981614, Time: 0.0402 Steps: 73770, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001394, Sample Num: 22304, Cur Loss: 1.80956125, Cur Avg Loss: 1.22238385, Log Avg loss: 1.21989004, Global Avg Loss: 1.55977007, Time: 0.0403 Steps: 73780, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001404, Sample Num: 22464, Cur Loss: 1.18979383, Cur Avg Loss: 1.22066579, Log Avg loss: 0.98116958, Global Avg Loss: 1.55969165, Time: 0.0402 Steps: 73790, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001414, Sample Num: 22624, Cur Loss: 1.21948957, Cur Avg Loss: 1.22173841, Log Avg loss: 1.37233290, Global Avg Loss: 1.55966627, Time: 0.0402 Steps: 73800, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001424, Sample Num: 22784, Cur Loss: 0.31391865, Cur Avg Loss: 1.22052461, Log Avg loss: 1.04889373, Global Avg Loss: 1.55959707, Time: 0.0403 Steps: 73810, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001434, Sample Num: 22944, Cur Loss: 1.65493643, Cur Avg Loss: 1.21989708, Log Avg loss: 1.13053764, Global Avg Loss: 1.55953894, Time: 0.0403 Steps: 73820, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001444, Sample Num: 23104, Cur Loss: 1.10327053, Cur Avg Loss: 1.22092925, Log Avg loss: 1.36894233, Global Avg Loss: 1.55951313, Time: 0.0403 Steps: 73830, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001454, Sample Num: 23264, Cur Loss: 0.97730970, Cur Avg Loss: 1.22037977, Log Avg loss: 1.14103498, Global Avg Loss: 1.55945645, Time: 0.0402 Steps: 73840, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001464, Sample Num: 23424, Cur Loss: 1.13396525, Cur Avg Loss: 1.21966592, Log Avg loss: 1.11587130, Global Avg Loss: 1.55939639, Time: 0.0403 Steps: 73850, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001474, Sample Num: 23584, Cur Loss: 2.36829901, Cur Avg Loss: 1.22183091, Log Avg loss: 1.53878604, Global Avg Loss: 1.55939360, Time: 0.0402 Steps: 73860, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001484, Sample Num: 23744, Cur Loss: 0.81872022, Cur Avg Loss: 1.22011993, Log Avg loss: 0.96792133, Global Avg Loss: 1.55931353, Time: 0.0403 Steps: 73870, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001494, Sample Num: 23904, Cur Loss: 1.43957686, Cur Avg Loss: 1.22142808, Log Avg loss: 1.41555759, Global Avg Loss: 1.55929407, Time: 0.0402 Steps: 73880, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001504, Sample Num: 24064, Cur Loss: 1.40956426, Cur Avg Loss: 1.21982329, Log Avg loss: 0.98006727, Global Avg Loss: 1.55921568, Time: 0.0402 Steps: 73890, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001514, Sample Num: 24224, Cur Loss: 0.78135818, Cur Avg Loss: 1.21831563, Log Avg loss: 0.99156357, Global Avg Loss: 1.55913887, Time: 0.0705 Steps: 73900, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001524, Sample Num: 24384, Cur Loss: 0.86218703, Cur Avg Loss: 1.21841874, Log Avg loss: 1.23402961, Global Avg Loss: 1.55909488, Time: 0.0856 Steps: 73910, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001534, Sample Num: 24544, Cur Loss: 1.51630569, Cur Avg Loss: 1.21992892, Log Avg loss: 1.45008000, Global Avg Loss: 1.55908013, Time: 0.0578 Steps: 73920, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001544, Sample Num: 24704, Cur Loss: 0.58912504, Cur Avg Loss: 1.22086139, Log Avg loss: 1.36390234, Global Avg Loss: 1.55905373, Time: 0.0662 Steps: 73930, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001554, Sample Num: 24864, Cur Loss: 0.89244407, Cur Avg Loss: 1.22031045, Log Avg loss: 1.13524494, Global Avg Loss: 1.55899641, Time: 0.0462 Steps: 73940, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001564, Sample Num: 25024, Cur Loss: 0.28295150, Cur Avg Loss: 1.21763210, Log Avg loss: 0.80141747, Global Avg Loss: 1.55889397, Time: 0.0628 Steps: 73950, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001574, Sample Num: 25184, Cur Loss: 0.44756854, Cur Avg Loss: 1.21733238, Log Avg loss: 1.17045640, Global Avg Loss: 1.55884145, Time: 0.0586 Steps: 73960, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001584, Sample Num: 25344, Cur Loss: 1.03160143, Cur Avg Loss: 1.21703445, Log Avg loss: 1.17014013, Global Avg Loss: 1.55878890, Time: 0.0731 Steps: 73970, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001594, Sample Num: 25504, Cur Loss: 0.48269808, Cur Avg Loss: 1.21753744, Log Avg loss: 1.29721118, Global Avg Loss: 1.55875354, Time: 0.0598 Steps: 73980, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001604, Sample Num: 25664, Cur Loss: 1.01915312, Cur Avg Loss: 1.21676717, Log Avg loss: 1.09398506, Global Avg Loss: 1.55869073, Time: 0.0403 Steps: 73990, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001614, Sample Num: 25824, Cur Loss: 1.92818034, Cur Avg Loss: 1.21540028, Log Avg loss: 0.99615155, Global Avg Loss: 1.55861471, Time: 0.0403 Steps: 74000, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001624, Sample Num: 25984, Cur Loss: 1.87225580, Cur Avg Loss: 1.21504528, Log Avg loss: 1.15774917, Global Avg Loss: 1.55856054, Time: 0.0402 Steps: 74010, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001634, Sample Num: 26144, Cur Loss: 0.75363636, Cur Avg Loss: 1.21376116, Log Avg loss: 1.00521896, Global Avg Loss: 1.55848579, Time: 0.0403 Steps: 74020, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001644, Sample Num: 26304, Cur Loss: 1.85472429, Cur Avg Loss: 1.21391276, Log Avg loss: 1.23868507, Global Avg Loss: 1.55844259, Time: 0.0404 Steps: 74030, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001654, Sample Num: 26464, Cur Loss: 1.02990103, Cur Avg Loss: 1.21485155, Log Avg loss: 1.36918779, Global Avg Loss: 1.55841703, Time: 0.0404 Steps: 74040, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001664, Sample Num: 26624, Cur Loss: 1.69115353, Cur Avg Loss: 1.21605756, Log Avg loss: 1.41553174, Global Avg Loss: 1.55839773, Time: 0.0404 Steps: 74050, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001674, Sample Num: 26784, Cur Loss: 1.14766431, Cur Avg Loss: 1.21593745, Log Avg loss: 1.19595194, Global Avg Loss: 1.55834879, Time: 0.0404 Steps: 74060, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001684, Sample Num: 26944, Cur Loss: 1.75284696, Cur Avg Loss: 1.21607306, Log Avg loss: 1.23877409, Global Avg Loss: 1.55830565, Time: 0.0403 Steps: 74070, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001694, Sample Num: 27104, Cur Loss: 1.31709170, Cur Avg Loss: 1.21608648, Log Avg loss: 1.21834515, Global Avg Loss: 1.55825976, Time: 0.0402 Steps: 74080, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001704, Sample Num: 27264, Cur Loss: 0.59483635, Cur Avg Loss: 1.21353162, Log Avg loss: 0.78073970, Global Avg Loss: 1.55815481, Time: 0.0402 Steps: 74090, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001714, Sample Num: 27424, Cur Loss: 2.46191311, Cur Avg Loss: 1.21304348, Log Avg loss: 1.12986367, Global Avg Loss: 1.55809702, Time: 0.0402 Steps: 74100, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001724, Sample Num: 27584, Cur Loss: 0.73607522, Cur Avg Loss: 1.21247594, Log Avg loss: 1.11519955, Global Avg Loss: 1.55803725, Time: 0.0403 Steps: 74110, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001734, Sample Num: 27744, Cur Loss: 0.81451297, Cur Avg Loss: 1.21315926, Log Avg loss: 1.33096365, Global Avg Loss: 1.55800662, Time: 0.0403 Steps: 74120, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001744, Sample Num: 27904, Cur Loss: 1.21605599, Cur Avg Loss: 1.21412973, Log Avg loss: 1.38241000, Global Avg Loss: 1.55798293, Time: 0.0402 Steps: 74130, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001754, Sample Num: 28064, Cur Loss: 1.55568719, Cur Avg Loss: 1.21566448, Log Avg loss: 1.48332515, Global Avg Loss: 1.55797286, Time: 0.0404 Steps: 74140, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001764, Sample Num: 28224, Cur Loss: 1.19814026, Cur Avg Loss: 1.21551195, Log Avg loss: 1.18875818, Global Avg Loss: 1.55792307, Time: 0.0402 Steps: 74150, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001774, Sample Num: 28384, Cur Loss: 1.14266920, Cur Avg Loss: 1.21549101, Log Avg loss: 1.21179690, Global Avg Loss: 1.55787639, Time: 0.0404 Steps: 74160, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001784, Sample Num: 28544, Cur Loss: 1.75322127, Cur Avg Loss: 1.21476624, Log Avg loss: 1.08619120, Global Avg Loss: 1.55781280, Time: 0.0404 Steps: 74170, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001794, Sample Num: 28704, Cur Loss: 2.13528705, Cur Avg Loss: 1.21587677, Log Avg loss: 1.41399491, Global Avg Loss: 1.55779341, Time: 0.0402 Steps: 74180, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001804, Sample Num: 28864, Cur Loss: 1.24413741, Cur Avg Loss: 1.21555731, Log Avg loss: 1.15824611, Global Avg Loss: 1.55773956, Time: 0.0403 Steps: 74190, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001814, Sample Num: 29024, Cur Loss: 1.51938128, Cur Avg Loss: 1.21580927, Log Avg loss: 1.26126401, Global Avg Loss: 1.55769960, Time: 0.0402 Steps: 74200, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001824, Sample Num: 29184, Cur Loss: 0.81414926, Cur Avg Loss: 1.21669343, Log Avg loss: 1.37707892, Global Avg Loss: 1.55767526, Time: 0.0403 Steps: 74210, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001834, Sample Num: 29344, Cur Loss: 0.43045795, Cur Avg Loss: 1.21487117, Log Avg loss: 0.88249124, Global Avg Loss: 1.55758429, Time: 0.0402 Steps: 74220, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001844, Sample Num: 29504, Cur Loss: 1.16280758, Cur Avg Loss: 1.21760909, Log Avg loss: 1.71974415, Global Avg Loss: 1.55760614, Time: 0.0402 Steps: 74230, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001854, Sample Num: 29664, Cur Loss: 2.15259790, Cur Avg Loss: 1.21986707, Log Avg loss: 1.63623821, Global Avg Loss: 1.55761673, Time: 0.0402 Steps: 74240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001864, Sample Num: 29824, Cur Loss: 0.70326442, Cur Avg Loss: 1.21994994, Log Avg loss: 1.23531412, Global Avg Loss: 1.55757332, Time: 0.0402 Steps: 74250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001874, Sample Num: 29984, Cur Loss: 0.58455473, Cur Avg Loss: 1.22028214, Log Avg loss: 1.28220360, Global Avg Loss: 1.55753624, Time: 0.0636 Steps: 74260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001884, Sample Num: 30144, Cur Loss: 0.74731362, Cur Avg Loss: 1.22334290, Log Avg loss: 1.79692996, Global Avg Loss: 1.55756847, Time: 0.0403 Steps: 74270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001894, Sample Num: 30304, Cur Loss: 1.08374822, Cur Avg Loss: 1.22315162, Log Avg loss: 1.18711560, Global Avg Loss: 1.55751860, Time: 0.0403 Steps: 74280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001904, Sample Num: 30464, Cur Loss: 1.20715964, Cur Avg Loss: 1.22306427, Log Avg loss: 1.20651999, Global Avg Loss: 1.55747135, Time: 0.0415 Steps: 74290, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001914, Sample Num: 30624, Cur Loss: 0.38735753, Cur Avg Loss: 1.22072053, Log Avg loss: 0.77447214, Global Avg Loss: 1.55736597, Time: 0.0634 Steps: 74300, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001924, Sample Num: 30784, Cur Loss: 0.44388527, Cur Avg Loss: 1.21840889, Log Avg loss: 0.77596035, Global Avg Loss: 1.55726081, Time: 0.0402 Steps: 74310, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001934, Sample Num: 30944, Cur Loss: 1.19046593, Cur Avg Loss: 1.21833785, Log Avg loss: 1.20466916, Global Avg Loss: 1.55721337, Time: 0.0586 Steps: 74320, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001944, Sample Num: 31104, Cur Loss: 1.32481921, Cur Avg Loss: 1.21906084, Log Avg loss: 1.35888885, Global Avg Loss: 1.55718669, Time: 0.0405 Steps: 74330, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001954, Sample Num: 31264, Cur Loss: 0.59221178, Cur Avg Loss: 1.21995744, Log Avg loss: 1.39425600, Global Avg Loss: 1.55716477, Time: 0.0504 Steps: 74340, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001964, Sample Num: 31424, Cur Loss: 0.86810440, Cur Avg Loss: 1.21954576, Log Avg loss: 1.13910328, Global Avg Loss: 1.55710854, Time: 0.0654 Steps: 74350, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001974, Sample Num: 31584, Cur Loss: 1.19245064, Cur Avg Loss: 1.22064649, Log Avg loss: 1.43682911, Global Avg Loss: 1.55709237, Time: 0.0404 Steps: 74360, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001984, Sample Num: 31744, Cur Loss: 0.48030248, Cur Avg Loss: 1.22098764, Log Avg loss: 1.28833155, Global Avg Loss: 1.55705623, Time: 0.0402 Steps: 74370, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001994, Sample Num: 31904, Cur Loss: 0.92649698, Cur Avg Loss: 1.22033260, Log Avg loss: 1.09037254, Global Avg Loss: 1.55699349, Time: 0.0402 Steps: 74380, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002004, Sample Num: 32064, Cur Loss: 0.89361989, Cur Avg Loss: 1.22042731, Log Avg loss: 1.23931174, Global Avg Loss: 1.55695078, Time: 0.0403 Steps: 74390, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002014, Sample Num: 32224, Cur Loss: 1.26057589, Cur Avg Loss: 1.22035677, Log Avg loss: 1.20622166, Global Avg Loss: 1.55690364, Time: 0.0402 Steps: 74400, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002024, Sample Num: 32384, Cur Loss: 0.68070495, Cur Avg Loss: 1.22056575, Log Avg loss: 1.26265351, Global Avg Loss: 1.55686410, Time: 0.0403 Steps: 74410, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002034, Sample Num: 32544, Cur Loss: 0.27550405, Cur Avg Loss: 1.21944127, Log Avg loss: 0.99184768, Global Avg Loss: 1.55678817, Time: 0.0403 Steps: 74420, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002044, Sample Num: 32704, Cur Loss: 1.51152372, Cur Avg Loss: 1.21883581, Log Avg loss: 1.09568434, Global Avg Loss: 1.55672622, Time: 0.0403 Steps: 74430, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002054, Sample Num: 32864, Cur Loss: 0.44346729, Cur Avg Loss: 1.21739667, Log Avg loss: 0.92323661, Global Avg Loss: 1.55664112, Time: 0.0403 Steps: 74440, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002064, Sample Num: 33024, Cur Loss: 1.50036323, Cur Avg Loss: 1.21646020, Log Avg loss: 1.02410856, Global Avg Loss: 1.55656959, Time: 0.0403 Steps: 74450, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002074, Sample Num: 33184, Cur Loss: 1.18386269, Cur Avg Loss: 1.21599294, Log Avg loss: 1.11955094, Global Avg Loss: 1.55651090, Time: 0.0403 Steps: 74460, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002084, Sample Num: 33344, Cur Loss: 0.61358917, Cur Avg Loss: 1.21487640, Log Avg loss: 0.98330635, Global Avg Loss: 1.55643393, Time: 0.0402 Steps: 74470, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002094, Sample Num: 33504, Cur Loss: 1.14024878, Cur Avg Loss: 1.21606358, Log Avg loss: 1.46347245, Global Avg Loss: 1.55642145, Time: 0.0403 Steps: 74480, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002104, Sample Num: 33664, Cur Loss: 1.99690819, Cur Avg Loss: 1.21720677, Log Avg loss: 1.45658911, Global Avg Loss: 1.55640805, Time: 0.0403 Steps: 74490, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002114, Sample Num: 33824, Cur Loss: 1.03545249, Cur Avg Loss: 1.21639811, Log Avg loss: 1.04625589, Global Avg Loss: 1.55633957, Time: 0.0403 Steps: 74500, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002124, Sample Num: 33984, Cur Loss: 1.28682411, Cur Avg Loss: 1.21590297, Log Avg loss: 1.11123042, Global Avg Loss: 1.55627983, Time: 0.0403 Steps: 74510, Updated lr: 0.000030 ***** Running evaluation checkpoint-74515 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-74515 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.884638, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.358364, "eval_total_loss": 954.930051, "eval_mae": 1.014334, "eval_mse": 1.358241, "eval_r2": 0.136613, "eval_sp_statistic": 0.41228, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.528781, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.457791, "test_total_loss": 731.811324, "test_mae": 1.077378, "test_mse": 1.457893, "test_r2": 0.059063, "test_sp_statistic": 0.435879, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.566343, "test_ps_pvalue": 0.0, "lr": 3.0284495021337127e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.556276365145336, "train_cur_epoch_loss": 2590.1010003462434, "train_cur_epoch_avg_loss": 1.2165810241175403, "train_cur_epoch_time": 95.88463807106018, "train_cur_epoch_avg_time": 0.045037406327412016, "epoch": 35, "step": 74515} ################################################## Training, Epoch: 0036, Batch: 000005, Sample Num: 80, Cur Loss: 1.00166035, Cur Avg Loss: 1.15190712, Log Avg loss: 1.32826369, Global Avg Loss: 1.55624923, Time: 0.0404 Steps: 74520, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000015, Sample Num: 240, Cur Loss: 1.52374744, Cur Avg Loss: 1.31648352, Log Avg loss: 1.39877172, Global Avg Loss: 1.55622810, Time: 0.0404 Steps: 74530, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000025, Sample Num: 400, Cur Loss: 0.51510090, Cur Avg Loss: 1.37367241, Log Avg loss: 1.45945575, Global Avg Loss: 1.55621512, Time: 0.0402 Steps: 74540, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000035, Sample Num: 560, Cur Loss: 0.61273879, Cur Avg Loss: 1.31649149, Log Avg loss: 1.17353919, Global Avg Loss: 1.55616379, Time: 0.0403 Steps: 74550, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000045, Sample Num: 720, Cur Loss: 1.94930434, Cur Avg Loss: 1.27911362, Log Avg loss: 1.14829108, Global Avg Loss: 1.55610909, Time: 0.0403 Steps: 74560, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000055, Sample Num: 880, Cur Loss: 0.44419825, Cur Avg Loss: 1.26708140, Log Avg loss: 1.21293639, Global Avg Loss: 1.55606307, Time: 0.0402 Steps: 74570, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000065, Sample Num: 1040, Cur Loss: 1.72478104, Cur Avg Loss: 1.24694734, Log Avg loss: 1.13621006, Global Avg Loss: 1.55600677, Time: 0.0403 Steps: 74580, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000075, Sample Num: 1200, Cur Loss: 1.12038124, Cur Avg Loss: 1.22663174, Log Avg loss: 1.09458029, Global Avg Loss: 1.55594491, Time: 0.0403 Steps: 74590, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000085, Sample Num: 1360, Cur Loss: 1.05689931, Cur Avg Loss: 1.23708701, Log Avg loss: 1.31550152, Global Avg Loss: 1.55591268, Time: 0.0402 Steps: 74600, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000095, Sample Num: 1520, Cur Loss: 2.19442320, Cur Avg Loss: 1.25682602, Log Avg loss: 1.42460763, Global Avg Loss: 1.55589508, Time: 0.0403 Steps: 74610, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000105, Sample Num: 1680, Cur Loss: 1.29606938, Cur Avg Loss: 1.28018760, Log Avg loss: 1.50212259, Global Avg Loss: 1.55588787, Time: 0.0402 Steps: 74620, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000115, Sample Num: 1840, Cur Loss: 1.65448904, Cur Avg Loss: 1.27198342, Log Avg loss: 1.18583958, Global Avg Loss: 1.55583829, Time: 0.0402 Steps: 74630, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000125, Sample Num: 2000, Cur Loss: 0.95522946, Cur Avg Loss: 1.26599153, Log Avg loss: 1.19708473, Global Avg Loss: 1.55579022, Time: 0.0402 Steps: 74640, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000135, Sample Num: 2160, Cur Loss: 1.13602495, Cur Avg Loss: 1.26672255, Log Avg loss: 1.27586034, Global Avg Loss: 1.55575272, Time: 0.0402 Steps: 74650, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000145, Sample Num: 2320, Cur Loss: 0.74722576, Cur Avg Loss: 1.26229284, Log Avg loss: 1.20249173, Global Avg Loss: 1.55570541, Time: 0.0403 Steps: 74660, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000155, Sample Num: 2480, Cur Loss: 1.47600889, Cur Avg Loss: 1.26847748, Log Avg loss: 1.35815482, Global Avg Loss: 1.55567895, Time: 0.0623 Steps: 74670, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000165, Sample Num: 2640, Cur Loss: 0.94767308, Cur Avg Loss: 1.25166235, Log Avg loss: 0.99102784, Global Avg Loss: 1.55560334, Time: 0.0585 Steps: 74680, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000175, Sample Num: 2800, Cur Loss: 1.32519281, Cur Avg Loss: 1.22123725, Log Avg loss: 0.71922308, Global Avg Loss: 1.55549136, Time: 0.0755 Steps: 74690, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000185, Sample Num: 2960, Cur Loss: 1.07266772, Cur Avg Loss: 1.22401689, Log Avg loss: 1.27266061, Global Avg Loss: 1.55545350, Time: 0.0417 Steps: 74700, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000195, Sample Num: 3120, Cur Loss: 1.44169855, Cur Avg Loss: 1.21252838, Log Avg loss: 0.99999091, Global Avg Loss: 1.55537915, Time: 0.0483 Steps: 74710, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000205, Sample Num: 3280, Cur Loss: 1.35141277, Cur Avg Loss: 1.21022194, Log Avg loss: 1.16524633, Global Avg Loss: 1.55532694, Time: 0.0677 Steps: 74720, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000215, Sample Num: 3440, Cur Loss: 1.73329759, Cur Avg Loss: 1.20152839, Log Avg loss: 1.02331068, Global Avg Loss: 1.55525575, Time: 0.0607 Steps: 74730, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000225, Sample Num: 3600, Cur Loss: 0.63588643, Cur Avg Loss: 1.20098326, Log Avg loss: 1.18926297, Global Avg Loss: 1.55520678, Time: 0.0755 Steps: 74740, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000235, Sample Num: 3760, Cur Loss: 0.65965140, Cur Avg Loss: 1.20437980, Log Avg loss: 1.28080183, Global Avg Loss: 1.55517007, Time: 0.0632 Steps: 74750, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000245, Sample Num: 3920, Cur Loss: 1.08629930, Cur Avg Loss: 1.19651426, Log Avg loss: 1.01167413, Global Avg Loss: 1.55509737, Time: 0.0402 Steps: 74760, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000255, Sample Num: 4080, Cur Loss: 2.57244396, Cur Avg Loss: 1.21005150, Log Avg loss: 1.54171384, Global Avg Loss: 1.55509558, Time: 0.0403 Steps: 74770, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000265, Sample Num: 4240, Cur Loss: 0.44280517, Cur Avg Loss: 1.20396141, Log Avg loss: 1.04866420, Global Avg Loss: 1.55502786, Time: 0.0402 Steps: 74780, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000275, Sample Num: 4400, Cur Loss: 0.48281264, Cur Avg Loss: 1.19892846, Log Avg loss: 1.06555516, Global Avg Loss: 1.55496241, Time: 0.0402 Steps: 74790, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000285, Sample Num: 4560, Cur Loss: 0.34398267, Cur Avg Loss: 1.19911753, Log Avg loss: 1.20431695, Global Avg Loss: 1.55491553, Time: 0.0402 Steps: 74800, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000295, Sample Num: 4720, Cur Loss: 2.72463202, Cur Avg Loss: 1.20564128, Log Avg loss: 1.39156812, Global Avg Loss: 1.55489370, Time: 0.0402 Steps: 74810, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000305, Sample Num: 4880, Cur Loss: 0.97441328, Cur Avg Loss: 1.19508738, Log Avg loss: 0.88374754, Global Avg Loss: 1.55480400, Time: 0.0403 Steps: 74820, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000315, Sample Num: 5040, Cur Loss: 2.01103163, Cur Avg Loss: 1.19031705, Log Avg loss: 1.04482186, Global Avg Loss: 1.55473584, Time: 0.0402 Steps: 74830, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000325, Sample Num: 5200, Cur Loss: 0.17528829, Cur Avg Loss: 1.18184519, Log Avg loss: 0.91498177, Global Avg Loss: 1.55465036, Time: 0.0402 Steps: 74840, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000335, Sample Num: 5360, Cur Loss: 1.33514917, Cur Avg Loss: 1.18296839, Log Avg loss: 1.21947230, Global Avg Loss: 1.55460558, Time: 0.0402 Steps: 74850, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000345, Sample Num: 5520, Cur Loss: 0.61915278, Cur Avg Loss: 1.17902198, Log Avg loss: 1.04681725, Global Avg Loss: 1.55453775, Time: 0.0402 Steps: 74860, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000355, Sample Num: 5680, Cur Loss: 1.50158000, Cur Avg Loss: 1.18366629, Log Avg loss: 1.34389513, Global Avg Loss: 1.55450962, Time: 0.0402 Steps: 74870, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000365, Sample Num: 5840, Cur Loss: 0.63774729, Cur Avg Loss: 1.18601561, Log Avg loss: 1.26941627, Global Avg Loss: 1.55447154, Time: 0.0403 Steps: 74880, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000375, Sample Num: 6000, Cur Loss: 1.35231543, Cur Avg Loss: 1.18449135, Log Avg loss: 1.12885588, Global Avg Loss: 1.55441471, Time: 0.0402 Steps: 74890, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000385, Sample Num: 6160, Cur Loss: 1.19538856, Cur Avg Loss: 1.19260362, Log Avg loss: 1.49681360, Global Avg Loss: 1.55440702, Time: 0.0402 Steps: 74900, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000395, Sample Num: 6320, Cur Loss: 0.64692837, Cur Avg Loss: 1.20346435, Log Avg loss: 1.62160248, Global Avg Loss: 1.55441599, Time: 0.0402 Steps: 74910, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000405, Sample Num: 6480, Cur Loss: 1.30850816, Cur Avg Loss: 1.20756476, Log Avg loss: 1.36953088, Global Avg Loss: 1.55439131, Time: 0.0404 Steps: 74920, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000415, Sample Num: 6640, Cur Loss: 1.39945126, Cur Avg Loss: 1.20470935, Log Avg loss: 1.08906547, Global Avg Loss: 1.55432921, Time: 0.0402 Steps: 74930, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000425, Sample Num: 6800, Cur Loss: 1.30261981, Cur Avg Loss: 1.20985560, Log Avg loss: 1.42342484, Global Avg Loss: 1.55431174, Time: 0.0402 Steps: 74940, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000435, Sample Num: 6960, Cur Loss: 1.87338448, Cur Avg Loss: 1.21214899, Log Avg loss: 1.30961807, Global Avg Loss: 1.55427909, Time: 0.0402 Steps: 74950, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000445, Sample Num: 7120, Cur Loss: 1.85991514, Cur Avg Loss: 1.21518965, Log Avg loss: 1.34745837, Global Avg Loss: 1.55425150, Time: 0.0402 Steps: 74960, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000455, Sample Num: 7280, Cur Loss: 1.20310891, Cur Avg Loss: 1.22024904, Log Avg loss: 1.44539185, Global Avg Loss: 1.55423698, Time: 0.0402 Steps: 74970, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000465, Sample Num: 7440, Cur Loss: 1.33096111, Cur Avg Loss: 1.22172093, Log Avg loss: 1.28869213, Global Avg Loss: 1.55420157, Time: 0.0402 Steps: 74980, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000475, Sample Num: 7600, Cur Loss: 1.10805464, Cur Avg Loss: 1.22632910, Log Avg loss: 1.44060905, Global Avg Loss: 1.55418642, Time: 0.0402 Steps: 74990, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000485, Sample Num: 7760, Cur Loss: 1.13590670, Cur Avg Loss: 1.22902394, Log Avg loss: 1.35702862, Global Avg Loss: 1.55416013, Time: 0.0402 Steps: 75000, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000495, Sample Num: 7920, Cur Loss: 0.80929041, Cur Avg Loss: 1.23046807, Log Avg loss: 1.30050868, Global Avg Loss: 1.55412632, Time: 0.0402 Steps: 75010, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000505, Sample Num: 8080, Cur Loss: 0.38057575, Cur Avg Loss: 1.23028407, Log Avg loss: 1.22117589, Global Avg Loss: 1.55408194, Time: 0.0435 Steps: 75020, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000515, Sample Num: 8240, Cur Loss: 0.49967420, Cur Avg Loss: 1.22797576, Log Avg loss: 1.11140610, Global Avg Loss: 1.55402294, Time: 0.0835 Steps: 75030, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000525, Sample Num: 8400, Cur Loss: 0.47903687, Cur Avg Loss: 1.22418585, Log Avg loss: 1.02900524, Global Avg Loss: 1.55395297, Time: 0.0410 Steps: 75040, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000535, Sample Num: 8560, Cur Loss: 2.85358071, Cur Avg Loss: 1.22483174, Log Avg loss: 1.25874122, Global Avg Loss: 1.55391364, Time: 0.0480 Steps: 75050, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000545, Sample Num: 8720, Cur Loss: 0.92510927, Cur Avg Loss: 1.22535276, Log Avg loss: 1.25322711, Global Avg Loss: 1.55387358, Time: 0.0404 Steps: 75060, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000555, Sample Num: 8880, Cur Loss: 1.12928438, Cur Avg Loss: 1.22494325, Log Avg loss: 1.20262538, Global Avg Loss: 1.55382679, Time: 0.0693 Steps: 75070, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000565, Sample Num: 9040, Cur Loss: 1.75314391, Cur Avg Loss: 1.22672959, Log Avg loss: 1.32587098, Global Avg Loss: 1.55379642, Time: 0.0407 Steps: 75080, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000575, Sample Num: 9200, Cur Loss: 1.63765502, Cur Avg Loss: 1.22484620, Log Avg loss: 1.11843477, Global Avg Loss: 1.55373845, Time: 0.0469 Steps: 75090, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000585, Sample Num: 9360, Cur Loss: 0.78402948, Cur Avg Loss: 1.22097925, Log Avg loss: 0.99862947, Global Avg Loss: 1.55366453, Time: 0.0402 Steps: 75100, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000595, Sample Num: 9520, Cur Loss: 0.88428175, Cur Avg Loss: 1.22173184, Log Avg loss: 1.26575856, Global Avg Loss: 1.55362620, Time: 0.0402 Steps: 75110, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000605, Sample Num: 9680, Cur Loss: 1.30786586, Cur Avg Loss: 1.22882528, Log Avg loss: 1.65088494, Global Avg Loss: 1.55363915, Time: 0.0404 Steps: 75120, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000615, Sample Num: 9840, Cur Loss: 0.73397827, Cur Avg Loss: 1.22614219, Log Avg loss: 1.06381556, Global Avg Loss: 1.55357395, Time: 0.0402 Steps: 75130, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000625, Sample Num: 10000, Cur Loss: 2.92712212, Cur Avg Loss: 1.22503178, Log Avg loss: 1.15674146, Global Avg Loss: 1.55352114, Time: 0.0402 Steps: 75140, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000635, Sample Num: 10160, Cur Loss: 1.63985896, Cur Avg Loss: 1.22608775, Log Avg loss: 1.29208571, Global Avg Loss: 1.55348635, Time: 0.0402 Steps: 75150, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000645, Sample Num: 10320, Cur Loss: 1.67209315, Cur Avg Loss: 1.22359545, Log Avg loss: 1.06533449, Global Avg Loss: 1.55342140, Time: 0.0402 Steps: 75160, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000655, Sample Num: 10480, Cur Loss: 0.87200350, Cur Avg Loss: 1.22388860, Log Avg loss: 1.24279644, Global Avg Loss: 1.55338008, Time: 0.0403 Steps: 75170, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000665, Sample Num: 10640, Cur Loss: 0.83655214, Cur Avg Loss: 1.21970527, Log Avg loss: 0.94569721, Global Avg Loss: 1.55329925, Time: 0.0403 Steps: 75180, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000675, Sample Num: 10800, Cur Loss: 1.51330054, Cur Avg Loss: 1.22369807, Log Avg loss: 1.48921953, Global Avg Loss: 1.55329072, Time: 0.0402 Steps: 75190, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000685, Sample Num: 10960, Cur Loss: 2.04243517, Cur Avg Loss: 1.22340367, Log Avg loss: 1.20353153, Global Avg Loss: 1.55324421, Time: 0.0402 Steps: 75200, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000695, Sample Num: 11120, Cur Loss: 0.83038574, Cur Avg Loss: 1.21932397, Log Avg loss: 0.93986440, Global Avg Loss: 1.55316266, Time: 0.0402 Steps: 75210, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000705, Sample Num: 11280, Cur Loss: 0.35803014, Cur Avg Loss: 1.21880623, Log Avg loss: 1.18282334, Global Avg Loss: 1.55311342, Time: 0.0402 Steps: 75220, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000715, Sample Num: 11440, Cur Loss: 1.36572266, Cur Avg Loss: 1.21986954, Log Avg loss: 1.29483323, Global Avg Loss: 1.55307909, Time: 0.0402 Steps: 75230, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000725, Sample Num: 11600, Cur Loss: 1.21326566, Cur Avg Loss: 1.21729648, Log Avg loss: 1.03332230, Global Avg Loss: 1.55301001, Time: 0.0402 Steps: 75240, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000735, Sample Num: 11760, Cur Loss: 0.44899943, Cur Avg Loss: 1.21338018, Log Avg loss: 0.92944874, Global Avg Loss: 1.55292715, Time: 0.0402 Steps: 75250, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000745, Sample Num: 11920, Cur Loss: 1.37256026, Cur Avg Loss: 1.21267735, Log Avg loss: 1.16101923, Global Avg Loss: 1.55287507, Time: 0.0402 Steps: 75260, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000755, Sample Num: 12080, Cur Loss: 0.48005670, Cur Avg Loss: 1.21307859, Log Avg loss: 1.24297071, Global Avg Loss: 1.55283390, Time: 0.0402 Steps: 75270, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000765, Sample Num: 12240, Cur Loss: 1.68260527, Cur Avg Loss: 1.20821460, Log Avg loss: 0.84098334, Global Avg Loss: 1.55273934, Time: 0.0403 Steps: 75280, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000775, Sample Num: 12400, Cur Loss: 1.04709566, Cur Avg Loss: 1.21197359, Log Avg loss: 1.49953642, Global Avg Loss: 1.55273227, Time: 0.0403 Steps: 75290, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000785, Sample Num: 12560, Cur Loss: 0.97672033, Cur Avg Loss: 1.21051358, Log Avg loss: 1.09736307, Global Avg Loss: 1.55267180, Time: 0.0402 Steps: 75300, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000795, Sample Num: 12720, Cur Loss: 1.78304446, Cur Avg Loss: 1.21459530, Log Avg loss: 1.53501005, Global Avg Loss: 1.55266945, Time: 0.0403 Steps: 75310, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000805, Sample Num: 12880, Cur Loss: 1.05072999, Cur Avg Loss: 1.21229824, Log Avg loss: 1.02968258, Global Avg Loss: 1.55260002, Time: 0.0402 Steps: 75320, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000815, Sample Num: 13040, Cur Loss: 1.19021499, Cur Avg Loss: 1.21329109, Log Avg loss: 1.29321539, Global Avg Loss: 1.55256559, Time: 0.0402 Steps: 75330, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000825, Sample Num: 13200, Cur Loss: 1.01713657, Cur Avg Loss: 1.20911155, Log Avg loss: 0.86847914, Global Avg Loss: 1.55247479, Time: 0.0402 Steps: 75340, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000835, Sample Num: 13360, Cur Loss: 1.05202138, Cur Avg Loss: 1.20939414, Log Avg loss: 1.23270757, Global Avg Loss: 1.55243235, Time: 0.0403 Steps: 75350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000845, Sample Num: 13520, Cur Loss: 1.03375292, Cur Avg Loss: 1.21299953, Log Avg loss: 1.51404952, Global Avg Loss: 1.55242726, Time: 0.0402 Steps: 75360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000855, Sample Num: 13680, Cur Loss: 0.86972451, Cur Avg Loss: 1.21302127, Log Avg loss: 1.21485794, Global Avg Loss: 1.55238247, Time: 0.0926 Steps: 75370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000865, Sample Num: 13840, Cur Loss: 0.91585028, Cur Avg Loss: 1.21191401, Log Avg loss: 1.11724405, Global Avg Loss: 1.55232474, Time: 0.0780 Steps: 75380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000875, Sample Num: 14000, Cur Loss: 2.16641283, Cur Avg Loss: 1.21011279, Log Avg loss: 1.05430669, Global Avg Loss: 1.55225868, Time: 0.0513 Steps: 75390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000885, Sample Num: 14160, Cur Loss: 1.54901409, Cur Avg Loss: 1.21283901, Log Avg loss: 1.45138296, Global Avg Loss: 1.55224530, Time: 0.0403 Steps: 75400, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000895, Sample Num: 14320, Cur Loss: 0.76731241, Cur Avg Loss: 1.21557463, Log Avg loss: 1.45767704, Global Avg Loss: 1.55223276, Time: 0.0595 Steps: 75410, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000905, Sample Num: 14480, Cur Loss: 1.65822208, Cur Avg Loss: 1.21717280, Log Avg loss: 1.36020914, Global Avg Loss: 1.55220730, Time: 0.0591 Steps: 75420, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000915, Sample Num: 14640, Cur Loss: 0.47149956, Cur Avg Loss: 1.21892511, Log Avg loss: 1.37750905, Global Avg Loss: 1.55218414, Time: 0.0825 Steps: 75430, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000925, Sample Num: 14800, Cur Loss: 0.49813953, Cur Avg Loss: 1.21785861, Log Avg loss: 1.12027467, Global Avg Loss: 1.55212689, Time: 0.0401 Steps: 75440, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000935, Sample Num: 14960, Cur Loss: 1.38784933, Cur Avg Loss: 1.21613123, Log Avg loss: 1.05634849, Global Avg Loss: 1.55206118, Time: 0.0411 Steps: 75450, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000945, Sample Num: 15120, Cur Loss: 1.65029907, Cur Avg Loss: 1.21573713, Log Avg loss: 1.17888807, Global Avg Loss: 1.55201173, Time: 0.0403 Steps: 75460, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000955, Sample Num: 15280, Cur Loss: 0.98001593, Cur Avg Loss: 1.21481076, Log Avg loss: 1.12726948, Global Avg Loss: 1.55195545, Time: 0.0403 Steps: 75470, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000965, Sample Num: 15440, Cur Loss: 1.76171231, Cur Avg Loss: 1.21361769, Log Avg loss: 1.09967960, Global Avg Loss: 1.55189553, Time: 0.0402 Steps: 75480, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000975, Sample Num: 15600, Cur Loss: 1.45730507, Cur Avg Loss: 1.21464972, Log Avg loss: 1.31423997, Global Avg Loss: 1.55186405, Time: 0.0402 Steps: 75490, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000985, Sample Num: 15760, Cur Loss: 0.65631592, Cur Avg Loss: 1.21720820, Log Avg loss: 1.46666003, Global Avg Loss: 1.55185276, Time: 0.0402 Steps: 75500, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000995, Sample Num: 15920, Cur Loss: 0.47678289, Cur Avg Loss: 1.21480707, Log Avg loss: 0.97829643, Global Avg Loss: 1.55177680, Time: 0.0402 Steps: 75510, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001005, Sample Num: 16080, Cur Loss: 2.26856852, Cur Avg Loss: 1.21677401, Log Avg loss: 1.41248420, Global Avg Loss: 1.55175836, Time: 0.0403 Steps: 75520, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001015, Sample Num: 16240, Cur Loss: 0.26527476, Cur Avg Loss: 1.21326326, Log Avg loss: 0.86043255, Global Avg Loss: 1.55166683, Time: 0.0402 Steps: 75530, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001025, Sample Num: 16400, Cur Loss: 2.15500641, Cur Avg Loss: 1.21380881, Log Avg loss: 1.26918186, Global Avg Loss: 1.55162943, Time: 0.0403 Steps: 75540, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001035, Sample Num: 16560, Cur Loss: 0.74008775, Cur Avg Loss: 1.21359945, Log Avg loss: 1.19214025, Global Avg Loss: 1.55158185, Time: 0.0402 Steps: 75550, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001045, Sample Num: 16720, Cur Loss: 1.78830707, Cur Avg Loss: 1.21747543, Log Avg loss: 1.61863942, Global Avg Loss: 1.55159072, Time: 0.0403 Steps: 75560, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001055, Sample Num: 16880, Cur Loss: 0.76804698, Cur Avg Loss: 1.21685911, Log Avg loss: 1.15245349, Global Avg Loss: 1.55153791, Time: 0.0402 Steps: 75570, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001065, Sample Num: 17040, Cur Loss: 0.78162777, Cur Avg Loss: 1.21350677, Log Avg loss: 0.85983508, Global Avg Loss: 1.55144639, Time: 0.0403 Steps: 75580, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001075, Sample Num: 17200, Cur Loss: 0.80183798, Cur Avg Loss: 1.21296774, Log Avg loss: 1.15556083, Global Avg Loss: 1.55139402, Time: 0.0402 Steps: 75590, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001085, Sample Num: 17360, Cur Loss: 1.29673624, Cur Avg Loss: 1.21652491, Log Avg loss: 1.59892134, Global Avg Loss: 1.55140030, Time: 0.0404 Steps: 75600, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001095, Sample Num: 17520, Cur Loss: 0.94354093, Cur Avg Loss: 1.21882816, Log Avg loss: 1.46873040, Global Avg Loss: 1.55138937, Time: 0.0404 Steps: 75610, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001105, Sample Num: 17680, Cur Loss: 1.03550684, Cur Avg Loss: 1.21724626, Log Avg loss: 1.04402819, Global Avg Loss: 1.55132228, Time: 0.0403 Steps: 75620, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001115, Sample Num: 17840, Cur Loss: 1.10595989, Cur Avg Loss: 1.21791324, Log Avg loss: 1.29161471, Global Avg Loss: 1.55128794, Time: 0.0403 Steps: 75630, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001125, Sample Num: 18000, Cur Loss: 1.11182785, Cur Avg Loss: 1.21941624, Log Avg loss: 1.38700022, Global Avg Loss: 1.55126622, Time: 0.0402 Steps: 75640, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001135, Sample Num: 18160, Cur Loss: 0.62380856, Cur Avg Loss: 1.21834591, Log Avg loss: 1.09793451, Global Avg Loss: 1.55120629, Time: 0.0402 Steps: 75650, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001145, Sample Num: 18320, Cur Loss: 0.79807448, Cur Avg Loss: 1.22020971, Log Avg loss: 1.43175028, Global Avg Loss: 1.55119050, Time: 0.0402 Steps: 75660, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001155, Sample Num: 18480, Cur Loss: 1.22087455, Cur Avg Loss: 1.21829861, Log Avg loss: 0.99947804, Global Avg Loss: 1.55111759, Time: 0.0402 Steps: 75670, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001165, Sample Num: 18640, Cur Loss: 1.17121911, Cur Avg Loss: 1.21925619, Log Avg loss: 1.32985631, Global Avg Loss: 1.55108836, Time: 0.0402 Steps: 75680, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001175, Sample Num: 18800, Cur Loss: 0.92725265, Cur Avg Loss: 1.21857879, Log Avg loss: 1.13966274, Global Avg Loss: 1.55103400, Time: 0.0402 Steps: 75690, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001185, Sample Num: 18960, Cur Loss: 0.89123476, Cur Avg Loss: 1.21883073, Log Avg loss: 1.24843303, Global Avg Loss: 1.55099403, Time: 0.0402 Steps: 75700, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001195, Sample Num: 19120, Cur Loss: 1.23485541, Cur Avg Loss: 1.21865642, Log Avg loss: 1.19800112, Global Avg Loss: 1.55094740, Time: 0.0402 Steps: 75710, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001205, Sample Num: 19280, Cur Loss: 1.02197695, Cur Avg Loss: 1.21933125, Log Avg loss: 1.29997331, Global Avg Loss: 1.55091426, Time: 0.0402 Steps: 75720, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001215, Sample Num: 19440, Cur Loss: 0.55863923, Cur Avg Loss: 1.21719264, Log Avg loss: 0.95948976, Global Avg Loss: 1.55083616, Time: 0.0880 Steps: 75730, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001225, Sample Num: 19600, Cur Loss: 0.78246903, Cur Avg Loss: 1.21712133, Log Avg loss: 1.20845744, Global Avg Loss: 1.55079096, Time: 0.0413 Steps: 75740, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001235, Sample Num: 19760, Cur Loss: 1.99995279, Cur Avg Loss: 1.21744200, Log Avg loss: 1.25672410, Global Avg Loss: 1.55075213, Time: 0.0633 Steps: 75750, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001245, Sample Num: 19920, Cur Loss: 0.78224373, Cur Avg Loss: 1.21904404, Log Avg loss: 1.41689565, Global Avg Loss: 1.55073447, Time: 0.0688 Steps: 75760, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001255, Sample Num: 20080, Cur Loss: 0.61868590, Cur Avg Loss: 1.21938638, Log Avg loss: 1.26200802, Global Avg Loss: 1.55069636, Time: 0.0711 Steps: 75770, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001265, Sample Num: 20240, Cur Loss: 0.37835252, Cur Avg Loss: 1.22025320, Log Avg loss: 1.32903943, Global Avg Loss: 1.55066711, Time: 0.0427 Steps: 75780, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001275, Sample Num: 20400, Cur Loss: 1.21220207, Cur Avg Loss: 1.22110014, Log Avg loss: 1.32823737, Global Avg Loss: 1.55063776, Time: 0.0482 Steps: 75790, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001285, Sample Num: 20560, Cur Loss: 1.34095573, Cur Avg Loss: 1.21946667, Log Avg loss: 1.01119942, Global Avg Loss: 1.55056660, Time: 0.0406 Steps: 75800, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001295, Sample Num: 20720, Cur Loss: 2.00006223, Cur Avg Loss: 1.22124856, Log Avg loss: 1.45022123, Global Avg Loss: 1.55055336, Time: 0.0665 Steps: 75810, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001305, Sample Num: 20880, Cur Loss: 2.48177052, Cur Avg Loss: 1.22329783, Log Avg loss: 1.48867892, Global Avg Loss: 1.55054520, Time: 0.0403 Steps: 75820, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001315, Sample Num: 21040, Cur Loss: 1.06529629, Cur Avg Loss: 1.22395392, Log Avg loss: 1.30957317, Global Avg Loss: 1.55051342, Time: 0.0403 Steps: 75830, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001325, Sample Num: 21200, Cur Loss: 1.45203447, Cur Avg Loss: 1.22434612, Log Avg loss: 1.27591991, Global Avg Loss: 1.55047721, Time: 0.0402 Steps: 75840, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001335, Sample Num: 21360, Cur Loss: 2.33975554, Cur Avg Loss: 1.22340119, Log Avg loss: 1.09819842, Global Avg Loss: 1.55041759, Time: 0.0402 Steps: 75850, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001345, Sample Num: 21520, Cur Loss: 0.84375697, Cur Avg Loss: 1.22284479, Log Avg loss: 1.14856531, Global Avg Loss: 1.55036461, Time: 0.0403 Steps: 75860, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001355, Sample Num: 21680, Cur Loss: 1.15565777, Cur Avg Loss: 1.22374912, Log Avg loss: 1.34538223, Global Avg Loss: 1.55033760, Time: 0.0402 Steps: 75870, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001365, Sample Num: 21840, Cur Loss: 1.02434778, Cur Avg Loss: 1.22271449, Log Avg loss: 1.08252100, Global Avg Loss: 1.55027594, Time: 0.0403 Steps: 75880, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001375, Sample Num: 22000, Cur Loss: 0.90286922, Cur Avg Loss: 1.22337483, Log Avg loss: 1.31351129, Global Avg Loss: 1.55024475, Time: 0.0402 Steps: 75890, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001385, Sample Num: 22160, Cur Loss: 1.20420206, Cur Avg Loss: 1.22395469, Log Avg loss: 1.30368561, Global Avg Loss: 1.55021226, Time: 0.0402 Steps: 75900, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001395, Sample Num: 22320, Cur Loss: 0.58181190, Cur Avg Loss: 1.22343377, Log Avg loss: 1.15128709, Global Avg Loss: 1.55015971, Time: 0.0402 Steps: 75910, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001405, Sample Num: 22480, Cur Loss: 1.57250094, Cur Avg Loss: 1.22542680, Log Avg loss: 1.50345427, Global Avg Loss: 1.55015356, Time: 0.0403 Steps: 75920, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001415, Sample Num: 22640, Cur Loss: 1.08532071, Cur Avg Loss: 1.22536584, Log Avg loss: 1.21680018, Global Avg Loss: 1.55010965, Time: 0.0403 Steps: 75930, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001425, Sample Num: 22800, Cur Loss: 1.47798634, Cur Avg Loss: 1.22572671, Log Avg loss: 1.27679040, Global Avg Loss: 1.55007366, Time: 0.0403 Steps: 75940, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001435, Sample Num: 22960, Cur Loss: 1.75038481, Cur Avg Loss: 1.22938119, Log Avg loss: 1.75014504, Global Avg Loss: 1.55010000, Time: 0.0402 Steps: 75950, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001445, Sample Num: 23120, Cur Loss: 0.87522155, Cur Avg Loss: 1.22783404, Log Avg loss: 1.00581785, Global Avg Loss: 1.55002835, Time: 0.0403 Steps: 75960, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001455, Sample Num: 23280, Cur Loss: 1.90873432, Cur Avg Loss: 1.22830926, Log Avg loss: 1.29697800, Global Avg Loss: 1.54999504, Time: 0.0402 Steps: 75970, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001465, Sample Num: 23440, Cur Loss: 1.57855308, Cur Avg Loss: 1.22894098, Log Avg loss: 1.32085642, Global Avg Loss: 1.54996488, Time: 0.0402 Steps: 75980, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001475, Sample Num: 23600, Cur Loss: 0.98195565, Cur Avg Loss: 1.22881079, Log Avg loss: 1.20973873, Global Avg Loss: 1.54992011, Time: 0.0402 Steps: 75990, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001485, Sample Num: 23760, Cur Loss: 2.22138643, Cur Avg Loss: 1.22972108, Log Avg loss: 1.36398864, Global Avg Loss: 1.54989565, Time: 0.0402 Steps: 76000, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001495, Sample Num: 23920, Cur Loss: 1.04015434, Cur Avg Loss: 1.22854323, Log Avg loss: 1.05363266, Global Avg Loss: 1.54983036, Time: 0.0402 Steps: 76010, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001505, Sample Num: 24080, Cur Loss: 0.84564650, Cur Avg Loss: 1.22577919, Log Avg loss: 0.81255532, Global Avg Loss: 1.54973337, Time: 0.0402 Steps: 76020, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001515, Sample Num: 24240, Cur Loss: 1.25993633, Cur Avg Loss: 1.22622177, Log Avg loss: 1.29282951, Global Avg Loss: 1.54969958, Time: 0.0402 Steps: 76030, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001525, Sample Num: 24400, Cur Loss: 0.33778888, Cur Avg Loss: 1.22477978, Log Avg loss: 1.00631775, Global Avg Loss: 1.54962812, Time: 0.0402 Steps: 76040, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001535, Sample Num: 24560, Cur Loss: 2.24437499, Cur Avg Loss: 1.22289341, Log Avg loss: 0.93522207, Global Avg Loss: 1.54954733, Time: 0.0402 Steps: 76050, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001545, Sample Num: 24720, Cur Loss: 0.43356445, Cur Avg Loss: 1.22209832, Log Avg loss: 1.10005248, Global Avg Loss: 1.54948824, Time: 0.0402 Steps: 76060, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001555, Sample Num: 24880, Cur Loss: 0.29000744, Cur Avg Loss: 1.22097612, Log Avg loss: 1.04759578, Global Avg Loss: 1.54942226, Time: 0.0402 Steps: 76070, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001565, Sample Num: 25040, Cur Loss: 2.11602736, Cur Avg Loss: 1.22072220, Log Avg loss: 1.18123732, Global Avg Loss: 1.54937386, Time: 0.0617 Steps: 76080, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001575, Sample Num: 25200, Cur Loss: 0.33383334, Cur Avg Loss: 1.21895043, Log Avg loss: 0.94166972, Global Avg Loss: 1.54929400, Time: 0.0503 Steps: 76090, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001585, Sample Num: 25360, Cur Loss: 1.10837185, Cur Avg Loss: 1.21755456, Log Avg loss: 0.99770432, Global Avg Loss: 1.54922152, Time: 0.0919 Steps: 76100, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001595, Sample Num: 25520, Cur Loss: 1.17094600, Cur Avg Loss: 1.21847741, Log Avg loss: 1.36474994, Global Avg Loss: 1.54919728, Time: 0.0405 Steps: 76110, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001605, Sample Num: 25680, Cur Loss: 1.21506023, Cur Avg Loss: 1.21822721, Log Avg loss: 1.17832008, Global Avg Loss: 1.54914856, Time: 0.0524 Steps: 76120, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001615, Sample Num: 25840, Cur Loss: 0.34948564, Cur Avg Loss: 1.21722039, Log Avg loss: 1.05562516, Global Avg Loss: 1.54908373, Time: 0.0407 Steps: 76130, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001625, Sample Num: 26000, Cur Loss: 1.15654993, Cur Avg Loss: 1.21753511, Log Avg loss: 1.26836205, Global Avg Loss: 1.54904686, Time: 0.0403 Steps: 76140, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001635, Sample Num: 26160, Cur Loss: 1.23410094, Cur Avg Loss: 1.21712765, Log Avg loss: 1.15091608, Global Avg Loss: 1.54899458, Time: 0.0405 Steps: 76150, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001645, Sample Num: 26320, Cur Loss: 0.70523965, Cur Avg Loss: 1.21823867, Log Avg loss: 1.39988989, Global Avg Loss: 1.54897500, Time: 0.0403 Steps: 76160, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001655, Sample Num: 26480, Cur Loss: 1.76964796, Cur Avg Loss: 1.22055487, Log Avg loss: 1.60157005, Global Avg Loss: 1.54898190, Time: 0.0402 Steps: 76170, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001665, Sample Num: 26640, Cur Loss: 1.03378785, Cur Avg Loss: 1.22164528, Log Avg loss: 1.40210736, Global Avg Loss: 1.54896262, Time: 0.0402 Steps: 76180, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001675, Sample Num: 26800, Cur Loss: 0.54454958, Cur Avg Loss: 1.22138611, Log Avg loss: 1.17823505, Global Avg Loss: 1.54891397, Time: 0.0402 Steps: 76190, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001685, Sample Num: 26960, Cur Loss: 0.68259633, Cur Avg Loss: 1.21944945, Log Avg loss: 0.89505882, Global Avg Loss: 1.54882816, Time: 0.0403 Steps: 76200, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001695, Sample Num: 27120, Cur Loss: 2.27079701, Cur Avg Loss: 1.22040306, Log Avg loss: 1.38108680, Global Avg Loss: 1.54880615, Time: 0.0402 Steps: 76210, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001705, Sample Num: 27280, Cur Loss: 2.26140261, Cur Avg Loss: 1.22231745, Log Avg loss: 1.54680702, Global Avg Loss: 1.54880589, Time: 0.0402 Steps: 76220, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001715, Sample Num: 27440, Cur Loss: 1.75764656, Cur Avg Loss: 1.22364616, Log Avg loss: 1.45019022, Global Avg Loss: 1.54879295, Time: 0.0402 Steps: 76230, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001725, Sample Num: 27600, Cur Loss: 0.36664596, Cur Avg Loss: 1.22484502, Log Avg loss: 1.43044948, Global Avg Loss: 1.54877743, Time: 0.0402 Steps: 76240, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001735, Sample Num: 27760, Cur Loss: 0.56059575, Cur Avg Loss: 1.22303395, Log Avg loss: 0.91062524, Global Avg Loss: 1.54869373, Time: 0.0403 Steps: 76250, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001745, Sample Num: 27920, Cur Loss: 0.74470770, Cur Avg Loss: 1.22185582, Log Avg loss: 1.01744994, Global Avg Loss: 1.54862407, Time: 0.0403 Steps: 76260, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001755, Sample Num: 28080, Cur Loss: 2.88121510, Cur Avg Loss: 1.22294236, Log Avg loss: 1.41254377, Global Avg Loss: 1.54860623, Time: 0.0403 Steps: 76270, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001765, Sample Num: 28240, Cur Loss: 1.29809713, Cur Avg Loss: 1.22122156, Log Avg loss: 0.91922053, Global Avg Loss: 1.54852372, Time: 0.0403 Steps: 76280, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001775, Sample Num: 28400, Cur Loss: 1.02820194, Cur Avg Loss: 1.22056139, Log Avg loss: 1.10404126, Global Avg Loss: 1.54846546, Time: 0.0403 Steps: 76290, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001785, Sample Num: 28560, Cur Loss: 0.85595882, Cur Avg Loss: 1.21873665, Log Avg loss: 0.89484614, Global Avg Loss: 1.54837979, Time: 0.0402 Steps: 76300, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001795, Sample Num: 28720, Cur Loss: 1.16641033, Cur Avg Loss: 1.21807512, Log Avg loss: 1.09999080, Global Avg Loss: 1.54832104, Time: 0.0403 Steps: 76310, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001805, Sample Num: 28880, Cur Loss: 1.47035646, Cur Avg Loss: 1.21719937, Log Avg loss: 1.06000204, Global Avg Loss: 1.54825705, Time: 0.0403 Steps: 76320, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001815, Sample Num: 29040, Cur Loss: 1.47133994, Cur Avg Loss: 1.21621805, Log Avg loss: 1.03909113, Global Avg Loss: 1.54819035, Time: 0.0402 Steps: 76330, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001825, Sample Num: 29200, Cur Loss: 0.92373192, Cur Avg Loss: 1.21542411, Log Avg loss: 1.07132308, Global Avg Loss: 1.54812788, Time: 0.0403 Steps: 76340, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001835, Sample Num: 29360, Cur Loss: 0.52184236, Cur Avg Loss: 1.21649999, Log Avg loss: 1.41284871, Global Avg Loss: 1.54811016, Time: 0.0402 Steps: 76350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001845, Sample Num: 29520, Cur Loss: 0.98408949, Cur Avg Loss: 1.21664150, Log Avg loss: 1.24260864, Global Avg Loss: 1.54807015, Time: 0.0402 Steps: 76360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001855, Sample Num: 29680, Cur Loss: 0.77098328, Cur Avg Loss: 1.21610685, Log Avg loss: 1.11746342, Global Avg Loss: 1.54801377, Time: 0.0402 Steps: 76370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001865, Sample Num: 29840, Cur Loss: 1.09372997, Cur Avg Loss: 1.21719880, Log Avg loss: 1.41975623, Global Avg Loss: 1.54799698, Time: 0.0404 Steps: 76380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001875, Sample Num: 30000, Cur Loss: 0.36875984, Cur Avg Loss: 1.21679002, Log Avg loss: 1.14055262, Global Avg Loss: 1.54794364, Time: 0.0403 Steps: 76390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001885, Sample Num: 30160, Cur Loss: 1.60274446, Cur Avg Loss: 1.21744630, Log Avg loss: 1.34049818, Global Avg Loss: 1.54791649, Time: 0.0402 Steps: 76400, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001895, Sample Num: 30320, Cur Loss: 2.19320989, Cur Avg Loss: 1.21739608, Log Avg loss: 1.20792946, Global Avg Loss: 1.54787199, Time: 0.0402 Steps: 76410, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001905, Sample Num: 30480, Cur Loss: 0.80241865, Cur Avg Loss: 1.21647316, Log Avg loss: 1.04157946, Global Avg Loss: 1.54780574, Time: 0.0403 Steps: 76420, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001915, Sample Num: 30640, Cur Loss: 1.16593301, Cur Avg Loss: 1.21660773, Log Avg loss: 1.24224441, Global Avg Loss: 1.54776576, Time: 0.0403 Steps: 76430, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001925, Sample Num: 30800, Cur Loss: 1.82300973, Cur Avg Loss: 1.21733134, Log Avg loss: 1.35590159, Global Avg Loss: 1.54774066, Time: 0.1002 Steps: 76440, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001935, Sample Num: 30960, Cur Loss: 0.80162722, Cur Avg Loss: 1.21864317, Log Avg loss: 1.47117156, Global Avg Loss: 1.54773065, Time: 0.0798 Steps: 76450, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001945, Sample Num: 31120, Cur Loss: 0.93425685, Cur Avg Loss: 1.21749269, Log Avg loss: 0.99487519, Global Avg Loss: 1.54765834, Time: 0.0454 Steps: 76460, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001955, Sample Num: 31280, Cur Loss: 0.61986417, Cur Avg Loss: 1.21856572, Log Avg loss: 1.42726831, Global Avg Loss: 1.54764260, Time: 0.0411 Steps: 76470, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001965, Sample Num: 31440, Cur Loss: 0.95178342, Cur Avg Loss: 1.21800977, Log Avg loss: 1.10932253, Global Avg Loss: 1.54758528, Time: 0.0574 Steps: 76480, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001975, Sample Num: 31600, Cur Loss: 0.99974310, Cur Avg Loss: 1.21686758, Log Avg loss: 0.99242680, Global Avg Loss: 1.54751271, Time: 0.0409 Steps: 76490, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001985, Sample Num: 31760, Cur Loss: 1.69395685, Cur Avg Loss: 1.21936370, Log Avg loss: 1.71234847, Global Avg Loss: 1.54753425, Time: 0.0652 Steps: 76500, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001995, Sample Num: 31920, Cur Loss: 0.65300953, Cur Avg Loss: 1.21884143, Log Avg loss: 1.11516988, Global Avg Loss: 1.54747774, Time: 0.0417 Steps: 76510, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002005, Sample Num: 32080, Cur Loss: 1.80963874, Cur Avg Loss: 1.21976423, Log Avg loss: 1.40386236, Global Avg Loss: 1.54745897, Time: 0.0642 Steps: 76520, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002015, Sample Num: 32240, Cur Loss: 1.70025325, Cur Avg Loss: 1.21916221, Log Avg loss: 1.09845830, Global Avg Loss: 1.54740030, Time: 0.0402 Steps: 76530, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002025, Sample Num: 32400, Cur Loss: 1.24189091, Cur Avg Loss: 1.22051917, Log Avg loss: 1.49394527, Global Avg Loss: 1.54739332, Time: 0.0403 Steps: 76540, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002035, Sample Num: 32560, Cur Loss: 0.65308547, Cur Avg Loss: 1.22032546, Log Avg loss: 1.18110040, Global Avg Loss: 1.54734547, Time: 0.0402 Steps: 76550, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002045, Sample Num: 32720, Cur Loss: 1.57529879, Cur Avg Loss: 1.22121769, Log Avg loss: 1.40278681, Global Avg Loss: 1.54732659, Time: 0.0403 Steps: 76560, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002055, Sample Num: 32880, Cur Loss: 1.53831267, Cur Avg Loss: 1.22030426, Log Avg loss: 1.03350758, Global Avg Loss: 1.54725948, Time: 0.0403 Steps: 76570, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002065, Sample Num: 33040, Cur Loss: 1.03910089, Cur Avg Loss: 1.21899749, Log Avg loss: 0.95045684, Global Avg Loss: 1.54718155, Time: 0.0402 Steps: 76580, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002075, Sample Num: 33200, Cur Loss: 1.06119347, Cur Avg Loss: 1.21853122, Log Avg loss: 1.12224606, Global Avg Loss: 1.54712607, Time: 0.0402 Steps: 76590, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002085, Sample Num: 33360, Cur Loss: 0.60132682, Cur Avg Loss: 1.21959456, Log Avg loss: 1.44023768, Global Avg Loss: 1.54711212, Time: 0.0403 Steps: 76600, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002095, Sample Num: 33520, Cur Loss: 0.94953299, Cur Avg Loss: 1.21957738, Log Avg loss: 1.21599485, Global Avg Loss: 1.54706889, Time: 0.0403 Steps: 76610, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002105, Sample Num: 33680, Cur Loss: 0.56681734, Cur Avg Loss: 1.22140267, Log Avg loss: 1.60380004, Global Avg Loss: 1.54707630, Time: 0.0403 Steps: 76620, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002115, Sample Num: 33840, Cur Loss: 0.99474579, Cur Avg Loss: 1.22136905, Log Avg loss: 1.21429330, Global Avg Loss: 1.54703287, Time: 0.0402 Steps: 76630, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002125, Sample Num: 34000, Cur Loss: 0.25180995, Cur Avg Loss: 1.22035380, Log Avg loss: 1.00562842, Global Avg Loss: 1.54696223, Time: 0.0402 Steps: 76640, Updated lr: 0.000028 ***** Running evaluation checkpoint-76644 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-76644 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.313516, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.123336, "eval_total_loss": 789.705407, "eval_mae": 0.83519, "eval_mse": 1.123493, "eval_r2": 0.285834, "eval_sp_statistic": 0.434131, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.537984, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.088367, "test_total_loss": 546.360241, "test_mae": 0.822853, "test_mse": 1.088603, "test_r2": 0.297406, "test_sp_statistic": 0.457822, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.575288, "test_ps_pvalue": 0.0, "lr": 2.826552868658132e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.54693627473027, "train_cur_epoch_loss": 2597.4504916220903, "train_cur_epoch_avg_loss": 1.2200331102029547, "train_cur_epoch_time": 95.31351566314697, "train_cur_epoch_avg_time": 0.04476914779856598, "epoch": 36, "step": 76644} ################################################## Training, Epoch: 0037, Batch: 000006, Sample Num: 96, Cur Loss: 0.76332915, Cur Avg Loss: 1.29169751, Log Avg loss: 1.19488469, Global Avg Loss: 1.54691630, Time: 0.0403 Steps: 76650, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000016, Sample Num: 256, Cur Loss: 0.74765289, Cur Avg Loss: 1.22388606, Log Avg loss: 1.18319919, Global Avg Loss: 1.54686885, Time: 0.0403 Steps: 76660, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000026, Sample Num: 416, Cur Loss: 1.44209087, Cur Avg Loss: 1.16200619, Log Avg loss: 1.06299839, Global Avg Loss: 1.54680574, Time: 0.0402 Steps: 76670, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000036, Sample Num: 576, Cur Loss: 1.17651093, Cur Avg Loss: 1.19430012, Log Avg loss: 1.27826434, Global Avg Loss: 1.54677072, Time: 0.0402 Steps: 76680, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000046, Sample Num: 736, Cur Loss: 1.56413293, Cur Avg Loss: 1.28629672, Log Avg loss: 1.61748449, Global Avg Loss: 1.54677994, Time: 0.0402 Steps: 76690, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000056, Sample Num: 896, Cur Loss: 1.90614557, Cur Avg Loss: 1.29631751, Log Avg loss: 1.34241315, Global Avg Loss: 1.54675329, Time: 0.0402 Steps: 76700, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000066, Sample Num: 1056, Cur Loss: 1.18976963, Cur Avg Loss: 1.25968371, Log Avg loss: 1.05453444, Global Avg Loss: 1.54668913, Time: 0.0402 Steps: 76710, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000076, Sample Num: 1216, Cur Loss: 0.45422685, Cur Avg Loss: 1.26162612, Log Avg loss: 1.27444600, Global Avg Loss: 1.54665364, Time: 0.0403 Steps: 76720, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000086, Sample Num: 1376, Cur Loss: 0.85949183, Cur Avg Loss: 1.27758964, Log Avg loss: 1.39891237, Global Avg Loss: 1.54663439, Time: 0.0401 Steps: 76730, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000096, Sample Num: 1536, Cur Loss: 0.48832726, Cur Avg Loss: 1.27170677, Log Avg loss: 1.22111415, Global Avg Loss: 1.54659197, Time: 0.0402 Steps: 76740, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000106, Sample Num: 1696, Cur Loss: 1.00310934, Cur Avg Loss: 1.30274870, Log Avg loss: 1.60075119, Global Avg Loss: 1.54659903, Time: 0.0402 Steps: 76750, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000116, Sample Num: 1856, Cur Loss: 1.16341615, Cur Avg Loss: 1.28961180, Log Avg loss: 1.15036071, Global Avg Loss: 1.54654741, Time: 0.0402 Steps: 76760, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000126, Sample Num: 2016, Cur Loss: 0.95066684, Cur Avg Loss: 1.29259236, Log Avg loss: 1.32716678, Global Avg Loss: 1.54651883, Time: 0.0402 Steps: 76770, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000136, Sample Num: 2176, Cur Loss: 0.76362419, Cur Avg Loss: 1.29173657, Log Avg loss: 1.28095360, Global Avg Loss: 1.54648424, Time: 0.0402 Steps: 76780, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000146, Sample Num: 2336, Cur Loss: 0.66598248, Cur Avg Loss: 1.27958292, Log Avg loss: 1.11429331, Global Avg Loss: 1.54642796, Time: 0.0402 Steps: 76790, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000156, Sample Num: 2496, Cur Loss: 0.53743082, Cur Avg Loss: 1.25871582, Log Avg loss: 0.95405620, Global Avg Loss: 1.54635083, Time: 0.0402 Steps: 76800, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000166, Sample Num: 2656, Cur Loss: 1.37733829, Cur Avg Loss: 1.25737158, Log Avg loss: 1.23640137, Global Avg Loss: 1.54631047, Time: 0.0544 Steps: 76810, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000176, Sample Num: 2816, Cur Loss: 1.36572719, Cur Avg Loss: 1.25577087, Log Avg loss: 1.22919904, Global Avg Loss: 1.54626919, Time: 0.0996 Steps: 76820, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000186, Sample Num: 2976, Cur Loss: 1.27201879, Cur Avg Loss: 1.26544175, Log Avg loss: 1.43564929, Global Avg Loss: 1.54625480, Time: 0.0687 Steps: 76830, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000196, Sample Num: 3136, Cur Loss: 2.36663795, Cur Avg Loss: 1.27001926, Log Avg loss: 1.35516089, Global Avg Loss: 1.54622993, Time: 0.0644 Steps: 76840, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000206, Sample Num: 3296, Cur Loss: 1.91375709, Cur Avg Loss: 1.27495133, Log Avg loss: 1.37161999, Global Avg Loss: 1.54620721, Time: 0.0941 Steps: 76850, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000216, Sample Num: 3456, Cur Loss: 0.56788468, Cur Avg Loss: 1.26807386, Log Avg loss: 1.12639807, Global Avg Loss: 1.54615259, Time: 0.0733 Steps: 76860, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000226, Sample Num: 3616, Cur Loss: 0.75960851, Cur Avg Loss: 1.25526842, Log Avg loss: 0.97867082, Global Avg Loss: 1.54607876, Time: 0.0429 Steps: 76870, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000236, Sample Num: 3776, Cur Loss: 2.00816250, Cur Avg Loss: 1.25972731, Log Avg loss: 1.36049832, Global Avg Loss: 1.54605462, Time: 0.0639 Steps: 76880, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000246, Sample Num: 3936, Cur Loss: 0.45152655, Cur Avg Loss: 1.24338201, Log Avg loss: 0.85763288, Global Avg Loss: 1.54596509, Time: 0.0407 Steps: 76890, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000256, Sample Num: 4096, Cur Loss: 0.81234646, Cur Avg Loss: 1.23798167, Log Avg loss: 1.10513327, Global Avg Loss: 1.54590777, Time: 0.0404 Steps: 76900, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000266, Sample Num: 4256, Cur Loss: 0.76165277, Cur Avg Loss: 1.24444723, Log Avg loss: 1.40996550, Global Avg Loss: 1.54589009, Time: 0.0403 Steps: 76910, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000276, Sample Num: 4416, Cur Loss: 0.54661518, Cur Avg Loss: 1.24260877, Log Avg loss: 1.19370590, Global Avg Loss: 1.54584430, Time: 0.0403 Steps: 76920, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000286, Sample Num: 4576, Cur Loss: 1.96328449, Cur Avg Loss: 1.25207856, Log Avg loss: 1.51344462, Global Avg Loss: 1.54584009, Time: 0.0403 Steps: 76930, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000296, Sample Num: 4736, Cur Loss: 0.98520315, Cur Avg Loss: 1.25179421, Log Avg loss: 1.24366198, Global Avg Loss: 1.54580082, Time: 0.0402 Steps: 76940, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000306, Sample Num: 4896, Cur Loss: 1.92534947, Cur Avg Loss: 1.25780756, Log Avg loss: 1.43580259, Global Avg Loss: 1.54578652, Time: 0.0402 Steps: 76950, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000316, Sample Num: 5056, Cur Loss: 0.83740383, Cur Avg Loss: 1.25340712, Log Avg loss: 1.11875364, Global Avg Loss: 1.54573104, Time: 0.0402 Steps: 76960, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000326, Sample Num: 5216, Cur Loss: 1.17719984, Cur Avg Loss: 1.25541141, Log Avg loss: 1.31874714, Global Avg Loss: 1.54570155, Time: 0.0402 Steps: 76970, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000336, Sample Num: 5376, Cur Loss: 0.97161072, Cur Avg Loss: 1.25944304, Log Avg loss: 1.39087400, Global Avg Loss: 1.54568143, Time: 0.0402 Steps: 76980, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000346, Sample Num: 5536, Cur Loss: 1.04829597, Cur Avg Loss: 1.26516016, Log Avg loss: 1.45725543, Global Avg Loss: 1.54566995, Time: 0.0402 Steps: 76990, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000356, Sample Num: 5696, Cur Loss: 1.02976525, Cur Avg Loss: 1.25457820, Log Avg loss: 0.88844221, Global Avg Loss: 1.54558459, Time: 0.0402 Steps: 77000, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000366, Sample Num: 5856, Cur Loss: 1.58396232, Cur Avg Loss: 1.25735801, Log Avg loss: 1.35631927, Global Avg Loss: 1.54556002, Time: 0.0403 Steps: 77010, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000376, Sample Num: 6016, Cur Loss: 1.49435711, Cur Avg Loss: 1.25762866, Log Avg loss: 1.26753446, Global Avg Loss: 1.54552392, Time: 0.0403 Steps: 77020, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000386, Sample Num: 6176, Cur Loss: 1.01728010, Cur Avg Loss: 1.25441574, Log Avg loss: 1.13361015, Global Avg Loss: 1.54547044, Time: 0.0403 Steps: 77030, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000396, Sample Num: 6336, Cur Loss: 1.50349033, Cur Avg Loss: 1.25824577, Log Avg loss: 1.40608468, Global Avg Loss: 1.54545235, Time: 0.0402 Steps: 77040, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000406, Sample Num: 6496, Cur Loss: 0.64167506, Cur Avg Loss: 1.25865821, Log Avg loss: 1.27499090, Global Avg Loss: 1.54541725, Time: 0.0402 Steps: 77050, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000416, Sample Num: 6656, Cur Loss: 1.94553101, Cur Avg Loss: 1.26141067, Log Avg loss: 1.37316055, Global Avg Loss: 1.54539490, Time: 0.0402 Steps: 77060, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000426, Sample Num: 6816, Cur Loss: 2.08904004, Cur Avg Loss: 1.26242196, Log Avg loss: 1.30449180, Global Avg Loss: 1.54536364, Time: 0.0402 Steps: 77070, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000436, Sample Num: 6976, Cur Loss: 0.85608256, Cur Avg Loss: 1.25421173, Log Avg loss: 0.90445605, Global Avg Loss: 1.54528049, Time: 0.0401 Steps: 77080, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000446, Sample Num: 7136, Cur Loss: 1.40389371, Cur Avg Loss: 1.25507257, Log Avg loss: 1.29260512, Global Avg Loss: 1.54524771, Time: 0.0402 Steps: 77090, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000456, Sample Num: 7296, Cur Loss: 0.42681462, Cur Avg Loss: 1.25012037, Log Avg loss: 1.02925217, Global Avg Loss: 1.54518079, Time: 0.0402 Steps: 77100, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000466, Sample Num: 7456, Cur Loss: 0.91652840, Cur Avg Loss: 1.24748622, Log Avg loss: 1.12736876, Global Avg Loss: 1.54512660, Time: 0.0402 Steps: 77110, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000476, Sample Num: 7616, Cur Loss: 0.68498135, Cur Avg Loss: 1.24290942, Log Avg loss: 1.02963056, Global Avg Loss: 1.54505976, Time: 0.0402 Steps: 77120, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000486, Sample Num: 7776, Cur Loss: 0.46226159, Cur Avg Loss: 1.24901309, Log Avg loss: 1.53954808, Global Avg Loss: 1.54505905, Time: 0.0402 Steps: 77130, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000496, Sample Num: 7936, Cur Loss: 0.90541387, Cur Avg Loss: 1.25109853, Log Avg loss: 1.35245096, Global Avg Loss: 1.54503408, Time: 0.0402 Steps: 77140, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000506, Sample Num: 8096, Cur Loss: 0.80878770, Cur Avg Loss: 1.24831103, Log Avg loss: 1.11005066, Global Avg Loss: 1.54497770, Time: 0.0401 Steps: 77150, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000516, Sample Num: 8256, Cur Loss: 0.55444801, Cur Avg Loss: 1.24301301, Log Avg loss: 0.97493364, Global Avg Loss: 1.54490382, Time: 0.0402 Steps: 77160, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000526, Sample Num: 8416, Cur Loss: 1.62767732, Cur Avg Loss: 1.24304067, Log Avg loss: 1.24446791, Global Avg Loss: 1.54486489, Time: 0.0711 Steps: 77170, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000536, Sample Num: 8576, Cur Loss: 1.49549794, Cur Avg Loss: 1.23712373, Log Avg loss: 0.92589260, Global Avg Loss: 1.54478469, Time: 0.0404 Steps: 77180, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000546, Sample Num: 8736, Cur Loss: 0.45117670, Cur Avg Loss: 1.22640533, Log Avg loss: 0.65189869, Global Avg Loss: 1.54466901, Time: 0.0438 Steps: 77190, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000556, Sample Num: 8896, Cur Loss: 2.18756366, Cur Avg Loss: 1.22365699, Log Avg loss: 1.07359811, Global Avg Loss: 1.54460799, Time: 0.0430 Steps: 77200, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000566, Sample Num: 9056, Cur Loss: 2.23596430, Cur Avg Loss: 1.22492548, Log Avg loss: 1.29545344, Global Avg Loss: 1.54457572, Time: 0.0454 Steps: 77210, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000576, Sample Num: 9216, Cur Loss: 0.36286950, Cur Avg Loss: 1.22105199, Log Avg loss: 1.00181217, Global Avg Loss: 1.54450544, Time: 0.0680 Steps: 77220, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000586, Sample Num: 9376, Cur Loss: 2.09658241, Cur Avg Loss: 1.22215771, Log Avg loss: 1.28584715, Global Avg Loss: 1.54447194, Time: 0.0579 Steps: 77230, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000596, Sample Num: 9536, Cur Loss: 2.12679458, Cur Avg Loss: 1.21889432, Log Avg loss: 1.02765954, Global Avg Loss: 1.54440503, Time: 0.0774 Steps: 77240, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000606, Sample Num: 9696, Cur Loss: 1.27905083, Cur Avg Loss: 1.22017822, Log Avg loss: 1.29669875, Global Avg Loss: 1.54437297, Time: 0.0453 Steps: 77250, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000616, Sample Num: 9856, Cur Loss: 1.40966523, Cur Avg Loss: 1.22129712, Log Avg loss: 1.28910244, Global Avg Loss: 1.54433993, Time: 0.0403 Steps: 77260, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000626, Sample Num: 10016, Cur Loss: 1.78641891, Cur Avg Loss: 1.22450914, Log Avg loss: 1.42236999, Global Avg Loss: 1.54432414, Time: 0.0402 Steps: 77270, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000636, Sample Num: 10176, Cur Loss: 0.47175279, Cur Avg Loss: 1.22380303, Log Avg loss: 1.17960043, Global Avg Loss: 1.54427695, Time: 0.0402 Steps: 77280, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000646, Sample Num: 10336, Cur Loss: 3.76566887, Cur Avg Loss: 1.22604896, Log Avg loss: 1.36888996, Global Avg Loss: 1.54425426, Time: 0.0402 Steps: 77290, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000656, Sample Num: 10496, Cur Loss: 0.95965683, Cur Avg Loss: 1.22471647, Log Avg loss: 1.13863742, Global Avg Loss: 1.54420178, Time: 0.0403 Steps: 77300, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000666, Sample Num: 10656, Cur Loss: 1.01248312, Cur Avg Loss: 1.22556626, Log Avg loss: 1.28131301, Global Avg Loss: 1.54416778, Time: 0.0402 Steps: 77310, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000676, Sample Num: 10816, Cur Loss: 1.29733324, Cur Avg Loss: 1.22396709, Log Avg loss: 1.11746224, Global Avg Loss: 1.54411259, Time: 0.0402 Steps: 77320, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000686, Sample Num: 10976, Cur Loss: 0.43554872, Cur Avg Loss: 1.22190226, Log Avg loss: 1.08231992, Global Avg Loss: 1.54405287, Time: 0.0402 Steps: 77330, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000696, Sample Num: 11136, Cur Loss: 1.47723055, Cur Avg Loss: 1.21605801, Log Avg loss: 0.81514243, Global Avg Loss: 1.54395863, Time: 0.0403 Steps: 77340, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000706, Sample Num: 11296, Cur Loss: 0.84606183, Cur Avg Loss: 1.21978928, Log Avg loss: 1.47948554, Global Avg Loss: 1.54395029, Time: 0.0403 Steps: 77350, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000716, Sample Num: 11456, Cur Loss: 0.50021482, Cur Avg Loss: 1.21787713, Log Avg loss: 1.08287956, Global Avg Loss: 1.54389069, Time: 0.0403 Steps: 77360, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000726, Sample Num: 11616, Cur Loss: 2.68967772, Cur Avg Loss: 1.21872758, Log Avg loss: 1.27961932, Global Avg Loss: 1.54385653, Time: 0.0403 Steps: 77370, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000736, Sample Num: 11776, Cur Loss: 0.83807182, Cur Avg Loss: 1.21841857, Log Avg loss: 1.19598451, Global Avg Loss: 1.54381158, Time: 0.0402 Steps: 77380, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000746, Sample Num: 11936, Cur Loss: 0.57384551, Cur Avg Loss: 1.21775926, Log Avg loss: 1.16923374, Global Avg Loss: 1.54376318, Time: 0.0402 Steps: 77390, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000756, Sample Num: 12096, Cur Loss: 0.50593764, Cur Avg Loss: 1.21424333, Log Avg loss: 0.95195504, Global Avg Loss: 1.54368672, Time: 0.0403 Steps: 77400, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000766, Sample Num: 12256, Cur Loss: 1.07317412, Cur Avg Loss: 1.21615339, Log Avg loss: 1.36055440, Global Avg Loss: 1.54366306, Time: 0.0402 Steps: 77410, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000776, Sample Num: 12416, Cur Loss: 1.80657506, Cur Avg Loss: 1.21498604, Log Avg loss: 1.12556697, Global Avg Loss: 1.54360905, Time: 0.0402 Steps: 77420, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000786, Sample Num: 12576, Cur Loss: 1.36879063, Cur Avg Loss: 1.21200685, Log Avg loss: 0.98082196, Global Avg Loss: 1.54353637, Time: 0.0403 Steps: 77430, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000796, Sample Num: 12736, Cur Loss: 1.02528429, Cur Avg Loss: 1.21481104, Log Avg loss: 1.43521964, Global Avg Loss: 1.54352238, Time: 0.0402 Steps: 77440, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000806, Sample Num: 12896, Cur Loss: 0.60192990, Cur Avg Loss: 1.21410617, Log Avg loss: 1.15799882, Global Avg Loss: 1.54347261, Time: 0.0402 Steps: 77450, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000816, Sample Num: 13056, Cur Loss: 1.38622499, Cur Avg Loss: 1.21457999, Log Avg loss: 1.25277026, Global Avg Loss: 1.54343508, Time: 0.0403 Steps: 77460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000826, Sample Num: 13216, Cur Loss: 0.16910993, Cur Avg Loss: 1.21361470, Log Avg loss: 1.13484652, Global Avg Loss: 1.54338234, Time: 0.0403 Steps: 77470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000836, Sample Num: 13376, Cur Loss: 0.78036773, Cur Avg Loss: 1.21395655, Log Avg loss: 1.24219381, Global Avg Loss: 1.54334346, Time: 0.0403 Steps: 77480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000846, Sample Num: 13536, Cur Loss: 0.91932976, Cur Avg Loss: 1.21449391, Log Avg loss: 1.25941705, Global Avg Loss: 1.54330682, Time: 0.0403 Steps: 77490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000856, Sample Num: 13696, Cur Loss: 0.29055113, Cur Avg Loss: 1.21412142, Log Avg loss: 1.18260830, Global Avg Loss: 1.54326028, Time: 0.0403 Steps: 77500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000866, Sample Num: 13856, Cur Loss: 2.02645397, Cur Avg Loss: 1.21748330, Log Avg loss: 1.50526061, Global Avg Loss: 1.54325538, Time: 0.0403 Steps: 77510, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000876, Sample Num: 14016, Cur Loss: 0.80588430, Cur Avg Loss: 1.21802202, Log Avg loss: 1.26467483, Global Avg Loss: 1.54321944, Time: 0.0535 Steps: 77520, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000886, Sample Num: 14176, Cur Loss: 2.52780700, Cur Avg Loss: 1.21822528, Log Avg loss: 1.23603128, Global Avg Loss: 1.54317982, Time: 0.0431 Steps: 77530, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000896, Sample Num: 14336, Cur Loss: 1.46046114, Cur Avg Loss: 1.22056757, Log Avg loss: 1.42809395, Global Avg Loss: 1.54316498, Time: 0.0405 Steps: 77540, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000906, Sample Num: 14496, Cur Loss: 0.89917576, Cur Avg Loss: 1.21924366, Log Avg loss: 1.10062162, Global Avg Loss: 1.54310791, Time: 0.0405 Steps: 77550, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000916, Sample Num: 14656, Cur Loss: 1.28635335, Cur Avg Loss: 1.21827648, Log Avg loss: 1.13064967, Global Avg Loss: 1.54305473, Time: 0.1009 Steps: 77560, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000926, Sample Num: 14816, Cur Loss: 1.51904774, Cur Avg Loss: 1.21983195, Log Avg loss: 1.36231314, Global Avg Loss: 1.54303143, Time: 0.0763 Steps: 77570, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000936, Sample Num: 14976, Cur Loss: 1.20863068, Cur Avg Loss: 1.21985439, Log Avg loss: 1.22193277, Global Avg Loss: 1.54299004, Time: 0.0406 Steps: 77580, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000946, Sample Num: 15136, Cur Loss: 1.50535285, Cur Avg Loss: 1.21912182, Log Avg loss: 1.15055356, Global Avg Loss: 1.54293946, Time: 0.0620 Steps: 77590, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000956, Sample Num: 15296, Cur Loss: 0.78211010, Cur Avg Loss: 1.21957133, Log Avg loss: 1.26209407, Global Avg Loss: 1.54290327, Time: 0.0688 Steps: 77600, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000966, Sample Num: 15456, Cur Loss: 1.67721748, Cur Avg Loss: 1.22037535, Log Avg loss: 1.29724017, Global Avg Loss: 1.54287162, Time: 0.0432 Steps: 77610, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000976, Sample Num: 15616, Cur Loss: 0.73466671, Cur Avg Loss: 1.22237504, Log Avg loss: 1.41554507, Global Avg Loss: 1.54285522, Time: 0.0403 Steps: 77620, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000986, Sample Num: 15776, Cur Loss: 0.76305604, Cur Avg Loss: 1.22173025, Log Avg loss: 1.15879858, Global Avg Loss: 1.54280574, Time: 0.0403 Steps: 77630, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000996, Sample Num: 15936, Cur Loss: 1.22120261, Cur Avg Loss: 1.22079436, Log Avg loss: 1.12851535, Global Avg Loss: 1.54275238, Time: 0.0402 Steps: 77640, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001006, Sample Num: 16096, Cur Loss: 2.00105786, Cur Avg Loss: 1.22245608, Log Avg loss: 1.38796375, Global Avg Loss: 1.54273245, Time: 0.0402 Steps: 77650, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001016, Sample Num: 16256, Cur Loss: 0.60068476, Cur Avg Loss: 1.21886050, Log Avg loss: 0.85714543, Global Avg Loss: 1.54264417, Time: 0.0403 Steps: 77660, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001026, Sample Num: 16416, Cur Loss: 0.38097000, Cur Avg Loss: 1.21480797, Log Avg loss: 0.80307038, Global Avg Loss: 1.54254895, Time: 0.0403 Steps: 77670, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001036, Sample Num: 16576, Cur Loss: 0.34851575, Cur Avg Loss: 1.21564537, Log Avg loss: 1.30156325, Global Avg Loss: 1.54251793, Time: 0.0403 Steps: 77680, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001046, Sample Num: 16736, Cur Loss: 2.33902574, Cur Avg Loss: 1.21841957, Log Avg loss: 1.50582621, Global Avg Loss: 1.54251320, Time: 0.0402 Steps: 77690, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001056, Sample Num: 16896, Cur Loss: 1.72994506, Cur Avg Loss: 1.21818740, Log Avg loss: 1.19390211, Global Avg Loss: 1.54246834, Time: 0.0402 Steps: 77700, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001066, Sample Num: 17056, Cur Loss: 0.61227381, Cur Avg Loss: 1.21741056, Log Avg loss: 1.13537652, Global Avg Loss: 1.54241595, Time: 0.0402 Steps: 77710, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001076, Sample Num: 17216, Cur Loss: 1.60203505, Cur Avg Loss: 1.21673719, Log Avg loss: 1.14495612, Global Avg Loss: 1.54236481, Time: 0.0403 Steps: 77720, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001086, Sample Num: 17376, Cur Loss: 0.27217638, Cur Avg Loss: 1.21648161, Log Avg loss: 1.18898097, Global Avg Loss: 1.54231935, Time: 0.0402 Steps: 77730, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001096, Sample Num: 17536, Cur Loss: 1.91755819, Cur Avg Loss: 1.21968199, Log Avg loss: 1.56724383, Global Avg Loss: 1.54232255, Time: 0.0402 Steps: 77740, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001106, Sample Num: 17696, Cur Loss: 1.05214167, Cur Avg Loss: 1.22100161, Log Avg loss: 1.36563127, Global Avg Loss: 1.54229983, Time: 0.0402 Steps: 77750, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001116, Sample Num: 17856, Cur Loss: 0.72300208, Cur Avg Loss: 1.21826009, Log Avg loss: 0.91504813, Global Avg Loss: 1.54221916, Time: 0.0403 Steps: 77760, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001126, Sample Num: 18016, Cur Loss: 0.83305770, Cur Avg Loss: 1.21854485, Log Avg loss: 1.25032446, Global Avg Loss: 1.54218163, Time: 0.0402 Steps: 77770, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001136, Sample Num: 18176, Cur Loss: 1.25995266, Cur Avg Loss: 1.21885093, Log Avg loss: 1.25331562, Global Avg Loss: 1.54214449, Time: 0.0403 Steps: 77780, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001146, Sample Num: 18336, Cur Loss: 1.55406010, Cur Avg Loss: 1.21955747, Log Avg loss: 1.29982009, Global Avg Loss: 1.54211334, Time: 0.0402 Steps: 77790, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001156, Sample Num: 18496, Cur Loss: 0.67755914, Cur Avg Loss: 1.21742726, Log Avg loss: 0.97330536, Global Avg Loss: 1.54204023, Time: 0.0402 Steps: 77800, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001166, Sample Num: 18656, Cur Loss: 0.59359115, Cur Avg Loss: 1.21728935, Log Avg loss: 1.20134645, Global Avg Loss: 1.54199644, Time: 0.0403 Steps: 77810, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001176, Sample Num: 18816, Cur Loss: 1.00472367, Cur Avg Loss: 1.21848821, Log Avg loss: 1.35827563, Global Avg Loss: 1.54197283, Time: 0.0402 Steps: 77820, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001186, Sample Num: 18976, Cur Loss: 2.86275768, Cur Avg Loss: 1.21905025, Log Avg loss: 1.28514569, Global Avg Loss: 1.54193984, Time: 0.0402 Steps: 77830, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001196, Sample Num: 19136, Cur Loss: 1.01379132, Cur Avg Loss: 1.21756942, Log Avg loss: 1.04194352, Global Avg Loss: 1.54187560, Time: 0.0402 Steps: 77840, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001206, Sample Num: 19296, Cur Loss: 0.44577599, Cur Avg Loss: 1.21834107, Log Avg loss: 1.31063036, Global Avg Loss: 1.54184590, Time: 0.0402 Steps: 77850, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001216, Sample Num: 19456, Cur Loss: 1.83965576, Cur Avg Loss: 1.21927165, Log Avg loss: 1.33150010, Global Avg Loss: 1.54181888, Time: 0.0402 Steps: 77860, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001226, Sample Num: 19616, Cur Loss: 1.08882356, Cur Avg Loss: 1.21953140, Log Avg loss: 1.25111665, Global Avg Loss: 1.54178155, Time: 0.0736 Steps: 77870, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001236, Sample Num: 19776, Cur Loss: 1.90203094, Cur Avg Loss: 1.21892846, Log Avg loss: 1.14500812, Global Avg Loss: 1.54173060, Time: 0.0840 Steps: 77880, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001246, Sample Num: 19936, Cur Loss: 1.45408702, Cur Avg Loss: 1.21892362, Log Avg loss: 1.21832480, Global Avg Loss: 1.54168908, Time: 0.0408 Steps: 77890, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001256, Sample Num: 20096, Cur Loss: 0.75915980, Cur Avg Loss: 1.21819087, Log Avg loss: 1.12689064, Global Avg Loss: 1.54163584, Time: 0.0616 Steps: 77900, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001266, Sample Num: 20256, Cur Loss: 0.21724562, Cur Avg Loss: 1.21727919, Log Avg loss: 1.10277222, Global Avg Loss: 1.54157951, Time: 0.0949 Steps: 77910, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001276, Sample Num: 20416, Cur Loss: 1.17513156, Cur Avg Loss: 1.21780030, Log Avg loss: 1.28377291, Global Avg Loss: 1.54154642, Time: 0.0746 Steps: 77920, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001286, Sample Num: 20576, Cur Loss: 0.67553401, Cur Avg Loss: 1.21804225, Log Avg loss: 1.24891507, Global Avg Loss: 1.54150887, Time: 0.0434 Steps: 77930, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001296, Sample Num: 20736, Cur Loss: 1.76036620, Cur Avg Loss: 1.21739571, Log Avg loss: 1.13425100, Global Avg Loss: 1.54145662, Time: 0.0600 Steps: 77940, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001306, Sample Num: 20896, Cur Loss: 1.35049987, Cur Avg Loss: 1.21716706, Log Avg loss: 1.18753389, Global Avg Loss: 1.54141121, Time: 0.0610 Steps: 77950, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001316, Sample Num: 21056, Cur Loss: 0.80824125, Cur Avg Loss: 1.21709146, Log Avg loss: 1.20721800, Global Avg Loss: 1.54136835, Time: 0.0403 Steps: 77960, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001326, Sample Num: 21216, Cur Loss: 1.45919561, Cur Avg Loss: 1.21629774, Log Avg loss: 1.11184332, Global Avg Loss: 1.54131326, Time: 0.0402 Steps: 77970, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001336, Sample Num: 21376, Cur Loss: 1.09186041, Cur Avg Loss: 1.21465276, Log Avg loss: 0.99652836, Global Avg Loss: 1.54124339, Time: 0.0402 Steps: 77980, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001346, Sample Num: 21536, Cur Loss: 0.57660234, Cur Avg Loss: 1.21581855, Log Avg loss: 1.37156900, Global Avg Loss: 1.54122164, Time: 0.0402 Steps: 77990, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001356, Sample Num: 21696, Cur Loss: 0.92658341, Cur Avg Loss: 1.21580199, Log Avg loss: 1.21357230, Global Avg Loss: 1.54117963, Time: 0.0402 Steps: 78000, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001366, Sample Num: 21856, Cur Loss: 1.66601849, Cur Avg Loss: 1.21516553, Log Avg loss: 1.12886185, Global Avg Loss: 1.54112678, Time: 0.0402 Steps: 78010, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001376, Sample Num: 22016, Cur Loss: 0.53574705, Cur Avg Loss: 1.21550030, Log Avg loss: 1.26122994, Global Avg Loss: 1.54109090, Time: 0.0402 Steps: 78020, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001386, Sample Num: 22176, Cur Loss: 2.32836938, Cur Avg Loss: 1.21648973, Log Avg loss: 1.35263596, Global Avg Loss: 1.54106675, Time: 0.0402 Steps: 78030, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001396, Sample Num: 22336, Cur Loss: 0.95490557, Cur Avg Loss: 1.21612288, Log Avg loss: 1.16527703, Global Avg Loss: 1.54101860, Time: 0.0402 Steps: 78040, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001406, Sample Num: 22496, Cur Loss: 0.86754388, Cur Avg Loss: 1.21831874, Log Avg loss: 1.52486119, Global Avg Loss: 1.54101653, Time: 0.0402 Steps: 78050, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001416, Sample Num: 22656, Cur Loss: 1.06708169, Cur Avg Loss: 1.21837342, Log Avg loss: 1.22606030, Global Avg Loss: 1.54097618, Time: 0.0403 Steps: 78060, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001426, Sample Num: 22816, Cur Loss: 0.42498571, Cur Avg Loss: 1.21656974, Log Avg loss: 0.96116906, Global Avg Loss: 1.54090191, Time: 0.0403 Steps: 78070, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001436, Sample Num: 22976, Cur Loss: 0.79237163, Cur Avg Loss: 1.21528362, Log Avg loss: 1.03188358, Global Avg Loss: 1.54083672, Time: 0.0402 Steps: 78080, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001446, Sample Num: 23136, Cur Loss: 1.26096046, Cur Avg Loss: 1.21565480, Log Avg loss: 1.26895628, Global Avg Loss: 1.54080190, Time: 0.0402 Steps: 78090, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001456, Sample Num: 23296, Cur Loss: 1.28821778, Cur Avg Loss: 1.21465117, Log Avg loss: 1.06952612, Global Avg Loss: 1.54074156, Time: 0.0403 Steps: 78100, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001466, Sample Num: 23456, Cur Loss: 1.63380623, Cur Avg Loss: 1.21426901, Log Avg loss: 1.15862597, Global Avg Loss: 1.54069264, Time: 0.0403 Steps: 78110, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001476, Sample Num: 23616, Cur Loss: 2.47823310, Cur Avg Loss: 1.21442324, Log Avg loss: 1.23703325, Global Avg Loss: 1.54065377, Time: 0.0403 Steps: 78120, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001486, Sample Num: 23776, Cur Loss: 1.74167418, Cur Avg Loss: 1.21601918, Log Avg loss: 1.45158049, Global Avg Loss: 1.54064237, Time: 0.0402 Steps: 78130, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001496, Sample Num: 23936, Cur Loss: 0.89627433, Cur Avg Loss: 1.21608463, Log Avg loss: 1.22581027, Global Avg Loss: 1.54060208, Time: 0.0402 Steps: 78140, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001506, Sample Num: 24096, Cur Loss: 0.76379299, Cur Avg Loss: 1.21680010, Log Avg loss: 1.32383364, Global Avg Loss: 1.54057434, Time: 0.0402 Steps: 78150, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001516, Sample Num: 24256, Cur Loss: 1.01862621, Cur Avg Loss: 1.21672509, Log Avg loss: 1.20543006, Global Avg Loss: 1.54053146, Time: 0.0402 Steps: 78160, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001526, Sample Num: 24416, Cur Loss: 0.47726959, Cur Avg Loss: 1.21712054, Log Avg loss: 1.27707065, Global Avg Loss: 1.54049776, Time: 0.0402 Steps: 78170, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001536, Sample Num: 24576, Cur Loss: 0.85506427, Cur Avg Loss: 1.21669957, Log Avg loss: 1.15245906, Global Avg Loss: 1.54044812, Time: 0.0403 Steps: 78180, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001546, Sample Num: 24736, Cur Loss: 1.29787636, Cur Avg Loss: 1.21590408, Log Avg loss: 1.09371688, Global Avg Loss: 1.54039099, Time: 0.0403 Steps: 78190, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001556, Sample Num: 24896, Cur Loss: 1.66478944, Cur Avg Loss: 1.21526505, Log Avg loss: 1.11647017, Global Avg Loss: 1.54033678, Time: 0.0403 Steps: 78200, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001566, Sample Num: 25056, Cur Loss: 2.14499569, Cur Avg Loss: 1.21516631, Log Avg loss: 1.19980371, Global Avg Loss: 1.54029324, Time: 0.0403 Steps: 78210, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001576, Sample Num: 25216, Cur Loss: 0.57205188, Cur Avg Loss: 1.21378512, Log Avg loss: 0.99748968, Global Avg Loss: 1.54022385, Time: 0.0487 Steps: 78220, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001586, Sample Num: 25376, Cur Loss: 0.55455363, Cur Avg Loss: 1.21467899, Log Avg loss: 1.35555287, Global Avg Loss: 1.54020024, Time: 0.0652 Steps: 78230, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001596, Sample Num: 25536, Cur Loss: 0.91887456, Cur Avg Loss: 1.21340314, Log Avg loss: 1.01105393, Global Avg Loss: 1.54013261, Time: 0.0737 Steps: 78240, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001606, Sample Num: 25696, Cur Loss: 0.84329724, Cur Avg Loss: 1.21323232, Log Avg loss: 1.18596914, Global Avg Loss: 1.54008735, Time: 0.0413 Steps: 78250, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001616, Sample Num: 25856, Cur Loss: 1.54897463, Cur Avg Loss: 1.21342699, Log Avg loss: 1.24469041, Global Avg Loss: 1.54004960, Time: 0.0408 Steps: 78260, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001626, Sample Num: 26016, Cur Loss: 1.67555094, Cur Avg Loss: 1.21403367, Log Avg loss: 1.31207452, Global Avg Loss: 1.54002048, Time: 0.0406 Steps: 78270, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001636, Sample Num: 26176, Cur Loss: 1.86824250, Cur Avg Loss: 1.21504278, Log Avg loss: 1.37912325, Global Avg Loss: 1.53999992, Time: 0.0687 Steps: 78280, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001646, Sample Num: 26336, Cur Loss: 0.82592881, Cur Avg Loss: 1.21382483, Log Avg loss: 1.01456796, Global Avg Loss: 1.53993281, Time: 0.0481 Steps: 78290, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001656, Sample Num: 26496, Cur Loss: 1.70341361, Cur Avg Loss: 1.21298890, Log Avg loss: 1.07539480, Global Avg Loss: 1.53987348, Time: 0.0874 Steps: 78300, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001666, Sample Num: 26656, Cur Loss: 0.68715197, Cur Avg Loss: 1.21170485, Log Avg loss: 0.99906662, Global Avg Loss: 1.53980442, Time: 0.0402 Steps: 78310, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001676, Sample Num: 26816, Cur Loss: 0.84184110, Cur Avg Loss: 1.21180003, Log Avg loss: 1.22765617, Global Avg Loss: 1.53976456, Time: 0.0403 Steps: 78320, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001686, Sample Num: 26976, Cur Loss: 0.87946749, Cur Avg Loss: 1.21220166, Log Avg loss: 1.27951559, Global Avg Loss: 1.53973134, Time: 0.0403 Steps: 78330, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001696, Sample Num: 27136, Cur Loss: 1.36124372, Cur Avg Loss: 1.21214017, Log Avg loss: 1.20177355, Global Avg Loss: 1.53968820, Time: 0.0402 Steps: 78340, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001706, Sample Num: 27296, Cur Loss: 0.68654597, Cur Avg Loss: 1.21368164, Log Avg loss: 1.47511393, Global Avg Loss: 1.53967996, Time: 0.0402 Steps: 78350, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001716, Sample Num: 27456, Cur Loss: 1.31500936, Cur Avg Loss: 1.21440193, Log Avg loss: 1.33728374, Global Avg Loss: 1.53965413, Time: 0.0402 Steps: 78360, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001726, Sample Num: 27616, Cur Loss: 0.43354174, Cur Avg Loss: 1.21246002, Log Avg loss: 0.87922888, Global Avg Loss: 1.53956986, Time: 0.0403 Steps: 78370, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001736, Sample Num: 27776, Cur Loss: 1.57781065, Cur Avg Loss: 1.21213256, Log Avg loss: 1.15561330, Global Avg Loss: 1.53952087, Time: 0.0402 Steps: 78380, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001746, Sample Num: 27936, Cur Loss: 1.20180428, Cur Avg Loss: 1.21204451, Log Avg loss: 1.19675844, Global Avg Loss: 1.53947715, Time: 0.0402 Steps: 78390, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001756, Sample Num: 28096, Cur Loss: 0.75007141, Cur Avg Loss: 1.21351173, Log Avg loss: 1.46968767, Global Avg Loss: 1.53946825, Time: 0.0402 Steps: 78400, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001766, Sample Num: 28256, Cur Loss: 0.64343458, Cur Avg Loss: 1.21321363, Log Avg loss: 1.16086705, Global Avg Loss: 1.53941996, Time: 0.0402 Steps: 78410, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001776, Sample Num: 28416, Cur Loss: 0.88605821, Cur Avg Loss: 1.21400583, Log Avg loss: 1.35390950, Global Avg Loss: 1.53939630, Time: 0.0403 Steps: 78420, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001786, Sample Num: 28576, Cur Loss: 1.13277054, Cur Avg Loss: 1.21346844, Log Avg loss: 1.11802785, Global Avg Loss: 1.53934258, Time: 0.0402 Steps: 78430, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001796, Sample Num: 28736, Cur Loss: 0.56866038, Cur Avg Loss: 1.21222688, Log Avg loss: 0.99048369, Global Avg Loss: 1.53927261, Time: 0.0402 Steps: 78440, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001806, Sample Num: 28896, Cur Loss: 1.65984547, Cur Avg Loss: 1.21236312, Log Avg loss: 1.23683174, Global Avg Loss: 1.53923406, Time: 0.0402 Steps: 78450, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001816, Sample Num: 29056, Cur Loss: 1.10081136, Cur Avg Loss: 1.21252026, Log Avg loss: 1.24090041, Global Avg Loss: 1.53919603, Time: 0.0403 Steps: 78460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001826, Sample Num: 29216, Cur Loss: 1.41781545, Cur Avg Loss: 1.21333060, Log Avg loss: 1.36048811, Global Avg Loss: 1.53917326, Time: 0.0402 Steps: 78470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001836, Sample Num: 29376, Cur Loss: 1.55282950, Cur Avg Loss: 1.21230572, Log Avg loss: 1.02516295, Global Avg Loss: 1.53910776, Time: 0.0402 Steps: 78480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001846, Sample Num: 29536, Cur Loss: 1.56743371, Cur Avg Loss: 1.21306196, Log Avg loss: 1.35190747, Global Avg Loss: 1.53908391, Time: 0.0402 Steps: 78490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001856, Sample Num: 29696, Cur Loss: 1.02722049, Cur Avg Loss: 1.21259860, Log Avg loss: 1.12706233, Global Avg Loss: 1.53903142, Time: 0.0402 Steps: 78500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001866, Sample Num: 29856, Cur Loss: 0.78457838, Cur Avg Loss: 1.20944769, Log Avg loss: 0.62463828, Global Avg Loss: 1.53891496, Time: 0.0402 Steps: 78510, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001876, Sample Num: 30016, Cur Loss: 1.22145426, Cur Avg Loss: 1.21013549, Log Avg loss: 1.33847971, Global Avg Loss: 1.53888943, Time: 0.0402 Steps: 78520, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001886, Sample Num: 30176, Cur Loss: 0.99235475, Cur Avg Loss: 1.20939060, Log Avg loss: 1.06964822, Global Avg Loss: 1.53882968, Time: 0.0402 Steps: 78530, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001896, Sample Num: 30336, Cur Loss: 2.28896523, Cur Avg Loss: 1.21102666, Log Avg loss: 1.51958868, Global Avg Loss: 1.53882723, Time: 0.0402 Steps: 78540, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001906, Sample Num: 30496, Cur Loss: 0.87510455, Cur Avg Loss: 1.21084654, Log Avg loss: 1.17669548, Global Avg Loss: 1.53878112, Time: 0.0403 Steps: 78550, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001916, Sample Num: 30656, Cur Loss: 1.60815883, Cur Avg Loss: 1.21059886, Log Avg loss: 1.16339003, Global Avg Loss: 1.53873334, Time: 0.0401 Steps: 78560, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001926, Sample Num: 30816, Cur Loss: 1.73012114, Cur Avg Loss: 1.21174250, Log Avg loss: 1.43086476, Global Avg Loss: 1.53871961, Time: 0.0402 Steps: 78570, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001936, Sample Num: 30976, Cur Loss: 1.78140140, Cur Avg Loss: 1.21371572, Log Avg loss: 1.59375787, Global Avg Loss: 1.53872662, Time: 0.0455 Steps: 78580, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001946, Sample Num: 31136, Cur Loss: 1.05510402, Cur Avg Loss: 1.21470515, Log Avg loss: 1.40625938, Global Avg Loss: 1.53870976, Time: 0.0608 Steps: 78590, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001956, Sample Num: 31296, Cur Loss: 1.02345383, Cur Avg Loss: 1.21368958, Log Avg loss: 1.01605928, Global Avg Loss: 1.53864327, Time: 0.0569 Steps: 78600, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001966, Sample Num: 31456, Cur Loss: 0.97549617, Cur Avg Loss: 1.21225084, Log Avg loss: 0.93083263, Global Avg Loss: 1.53856595, Time: 0.0445 Steps: 78610, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001976, Sample Num: 31616, Cur Loss: 3.52556276, Cur Avg Loss: 1.21190318, Log Avg loss: 1.14355411, Global Avg Loss: 1.53851570, Time: 0.0580 Steps: 78620, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001986, Sample Num: 31776, Cur Loss: 1.27693701, Cur Avg Loss: 1.21151805, Log Avg loss: 1.13541633, Global Avg Loss: 1.53846444, Time: 0.0403 Steps: 78630, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001996, Sample Num: 31936, Cur Loss: 0.60057014, Cur Avg Loss: 1.21096960, Log Avg loss: 1.10204643, Global Avg Loss: 1.53840894, Time: 0.0406 Steps: 78640, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002006, Sample Num: 32096, Cur Loss: 1.49893796, Cur Avg Loss: 1.21180254, Log Avg loss: 1.37805713, Global Avg Loss: 1.53838855, Time: 0.0831 Steps: 78650, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002016, Sample Num: 32256, Cur Loss: 0.36755371, Cur Avg Loss: 1.21090915, Log Avg loss: 1.03169671, Global Avg Loss: 1.53832414, Time: 0.0650 Steps: 78660, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002026, Sample Num: 32416, Cur Loss: 1.80063522, Cur Avg Loss: 1.21365212, Log Avg loss: 1.76663379, Global Avg Loss: 1.53835316, Time: 0.0405 Steps: 78670, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002036, Sample Num: 32576, Cur Loss: 0.93727261, Cur Avg Loss: 1.21304824, Log Avg loss: 1.09070200, Global Avg Loss: 1.53829626, Time: 0.0403 Steps: 78680, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002046, Sample Num: 32736, Cur Loss: 2.27537918, Cur Avg Loss: 1.21233353, Log Avg loss: 1.06681873, Global Avg Loss: 1.53823635, Time: 0.0403 Steps: 78690, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002056, Sample Num: 32896, Cur Loss: 1.54348373, Cur Avg Loss: 1.21368401, Log Avg loss: 1.48999232, Global Avg Loss: 1.53823022, Time: 0.0403 Steps: 78700, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002066, Sample Num: 33056, Cur Loss: 2.00714374, Cur Avg Loss: 1.21280736, Log Avg loss: 1.03256864, Global Avg Loss: 1.53816597, Time: 0.0403 Steps: 78710, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002076, Sample Num: 33216, Cur Loss: 0.32301384, Cur Avg Loss: 1.21156459, Log Avg loss: 0.95480859, Global Avg Loss: 1.53809187, Time: 0.0403 Steps: 78720, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002086, Sample Num: 33376, Cur Loss: 1.00146735, Cur Avg Loss: 1.21130883, Log Avg loss: 1.15821320, Global Avg Loss: 1.53804362, Time: 0.0402 Steps: 78730, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002096, Sample Num: 33536, Cur Loss: 1.30396700, Cur Avg Loss: 1.21053136, Log Avg loss: 1.04835040, Global Avg Loss: 1.53798143, Time: 0.0403 Steps: 78740, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002106, Sample Num: 33696, Cur Loss: 1.79302049, Cur Avg Loss: 1.21099753, Log Avg loss: 1.30870597, Global Avg Loss: 1.53795231, Time: 0.0403 Steps: 78750, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002116, Sample Num: 33856, Cur Loss: 0.77205330, Cur Avg Loss: 1.21186244, Log Avg loss: 1.39401424, Global Avg Loss: 1.53793404, Time: 0.0403 Steps: 78760, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002126, Sample Num: 34016, Cur Loss: 1.06759286, Cur Avg Loss: 1.21081048, Log Avg loss: 0.98821519, Global Avg Loss: 1.53786425, Time: 0.0402 Steps: 78770, Updated lr: 0.000026 ***** Running evaluation checkpoint-78773 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-78773 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 96.037993, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.122831, "eval_total_loss": 789.350497, "eval_mae": 0.838828, "eval_mse": 1.122993, "eval_r2": 0.286152, "eval_sp_statistic": 0.434982, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.541666, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.098616, "test_total_loss": 551.505052, "test_mae": 0.842309, "test_mse": 1.098837, "test_r2": 0.290801, "test_sp_statistic": 0.458424, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.583099, "test_ps_pvalue": 0.0, "lr": 2.624656235182551e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.537854072444895, "train_cur_epoch_loss": 2577.995008274913, "train_cur_epoch_avg_loss": 1.2108947901714011, "train_cur_epoch_time": 96.03799343109131, "train_cur_epoch_avg_time": 0.0451094379666939, "epoch": 37, "step": 78773} ################################################## Training, Epoch: 0038, Batch: 000007, Sample Num: 112, Cur Loss: 0.60242081, Cur Avg Loss: 1.07172056, Log Avg loss: 1.13139669, Global Avg Loss: 1.53781265, Time: 0.0402 Steps: 78780, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000017, Sample Num: 272, Cur Loss: 1.01404667, Cur Avg Loss: 1.13377862, Log Avg loss: 1.17721927, Global Avg Loss: 1.53776689, Time: 0.0402 Steps: 78790, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000027, Sample Num: 432, Cur Loss: 0.71710038, Cur Avg Loss: 1.19742207, Log Avg loss: 1.30561593, Global Avg Loss: 1.53773743, Time: 0.0403 Steps: 78800, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000037, Sample Num: 592, Cur Loss: 1.15877843, Cur Avg Loss: 1.14240099, Log Avg loss: 0.99384408, Global Avg Loss: 1.53766841, Time: 0.0402 Steps: 78810, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000047, Sample Num: 752, Cur Loss: 0.63864207, Cur Avg Loss: 1.08424981, Log Avg loss: 0.86909044, Global Avg Loss: 1.53758359, Time: 0.0403 Steps: 78820, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000057, Sample Num: 912, Cur Loss: 0.95305926, Cur Avg Loss: 1.05412180, Log Avg loss: 0.91252014, Global Avg Loss: 1.53750430, Time: 0.0402 Steps: 78830, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000067, Sample Num: 1072, Cur Loss: 1.51600552, Cur Avg Loss: 1.07704221, Log Avg loss: 1.20768853, Global Avg Loss: 1.53746246, Time: 0.0403 Steps: 78840, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000077, Sample Num: 1232, Cur Loss: 2.00952435, Cur Avg Loss: 1.08749632, Log Avg loss: 1.15753891, Global Avg Loss: 1.53741428, Time: 0.0403 Steps: 78850, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000087, Sample Num: 1392, Cur Loss: 1.13694692, Cur Avg Loss: 1.08577622, Log Avg loss: 1.07253141, Global Avg Loss: 1.53735533, Time: 0.0402 Steps: 78860, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000097, Sample Num: 1552, Cur Loss: 2.99961710, Cur Avg Loss: 1.15392433, Log Avg loss: 1.74681294, Global Avg Loss: 1.53738189, Time: 0.0402 Steps: 78870, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000107, Sample Num: 1712, Cur Loss: 0.44911540, Cur Avg Loss: 1.13314566, Log Avg loss: 0.93159255, Global Avg Loss: 1.53730509, Time: 0.0402 Steps: 78880, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000117, Sample Num: 1872, Cur Loss: 0.91253191, Cur Avg Loss: 1.15760277, Log Avg loss: 1.41929387, Global Avg Loss: 1.53729013, Time: 0.0402 Steps: 78890, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000127, Sample Num: 2032, Cur Loss: 0.44948730, Cur Avg Loss: 1.16141833, Log Avg loss: 1.20606028, Global Avg Loss: 1.53724815, Time: 0.0403 Steps: 78900, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000137, Sample Num: 2192, Cur Loss: 2.00011539, Cur Avg Loss: 1.15881875, Log Avg loss: 1.12580419, Global Avg Loss: 1.53719601, Time: 0.0403 Steps: 78910, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000147, Sample Num: 2352, Cur Loss: 0.93222386, Cur Avg Loss: 1.16632957, Log Avg loss: 1.26922771, Global Avg Loss: 1.53716205, Time: 0.0403 Steps: 78920, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000157, Sample Num: 2512, Cur Loss: 1.63670945, Cur Avg Loss: 1.17519932, Log Avg loss: 1.30558470, Global Avg Loss: 1.53713271, Time: 0.0403 Steps: 78930, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000167, Sample Num: 2672, Cur Loss: 1.13303351, Cur Avg Loss: 1.17179773, Log Avg loss: 1.11839281, Global Avg Loss: 1.53707967, Time: 0.0402 Steps: 78940, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000177, Sample Num: 2832, Cur Loss: 0.93264771, Cur Avg Loss: 1.18853478, Log Avg loss: 1.46804352, Global Avg Loss: 1.53707092, Time: 0.0402 Steps: 78950, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000187, Sample Num: 2992, Cur Loss: 0.46707958, Cur Avg Loss: 1.18509737, Log Avg loss: 1.12425510, Global Avg Loss: 1.53701864, Time: 0.0402 Steps: 78960, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000197, Sample Num: 3152, Cur Loss: 0.83052474, Cur Avg Loss: 1.18053171, Log Avg loss: 1.09515394, Global Avg Loss: 1.53696269, Time: 0.0402 Steps: 78970, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000207, Sample Num: 3312, Cur Loss: 1.67528081, Cur Avg Loss: 1.19792794, Log Avg loss: 1.54063364, Global Avg Loss: 1.53696315, Time: 0.0402 Steps: 78980, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000217, Sample Num: 3472, Cur Loss: 1.18456388, Cur Avg Loss: 1.18395528, Log Avg loss: 0.89472131, Global Avg Loss: 1.53688185, Time: 0.0611 Steps: 78990, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000227, Sample Num: 3632, Cur Loss: 0.53145647, Cur Avg Loss: 1.18996749, Log Avg loss: 1.32043239, Global Avg Loss: 1.53685445, Time: 0.0921 Steps: 79000, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000237, Sample Num: 3792, Cur Loss: 1.58709216, Cur Avg Loss: 1.19534996, Log Avg loss: 1.31753199, Global Avg Loss: 1.53682669, Time: 0.0810 Steps: 79010, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000247, Sample Num: 3952, Cur Loss: 0.87143868, Cur Avg Loss: 1.20180602, Log Avg loss: 1.35481475, Global Avg Loss: 1.53680366, Time: 0.0982 Steps: 79020, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000257, Sample Num: 4112, Cur Loss: 1.02515495, Cur Avg Loss: 1.20133092, Log Avg loss: 1.18959590, Global Avg Loss: 1.53675972, Time: 0.0511 Steps: 79030, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000267, Sample Num: 4272, Cur Loss: 1.50185752, Cur Avg Loss: 1.20664221, Log Avg loss: 1.34314243, Global Avg Loss: 1.53673523, Time: 0.0720 Steps: 79040, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000277, Sample Num: 4432, Cur Loss: 1.24154687, Cur Avg Loss: 1.21446926, Log Avg loss: 1.42345142, Global Avg Loss: 1.53672090, Time: 0.0817 Steps: 79050, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000287, Sample Num: 4592, Cur Loss: 2.02208519, Cur Avg Loss: 1.20629399, Log Avg loss: 0.97983904, Global Avg Loss: 1.53665046, Time: 0.0404 Steps: 79060, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000297, Sample Num: 4752, Cur Loss: 0.85722238, Cur Avg Loss: 1.20844018, Log Avg loss: 1.27003576, Global Avg Loss: 1.53661674, Time: 0.0607 Steps: 79070, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000307, Sample Num: 4912, Cur Loss: 1.31060481, Cur Avg Loss: 1.19869926, Log Avg loss: 0.90939406, Global Avg Loss: 1.53653742, Time: 0.0403 Steps: 79080, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000317, Sample Num: 5072, Cur Loss: 2.05455399, Cur Avg Loss: 1.20367240, Log Avg loss: 1.35634766, Global Avg Loss: 1.53651464, Time: 0.0402 Steps: 79090, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000327, Sample Num: 5232, Cur Loss: 0.71167105, Cur Avg Loss: 1.19908082, Log Avg loss: 1.05352785, Global Avg Loss: 1.53645358, Time: 0.0402 Steps: 79100, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000337, Sample Num: 5392, Cur Loss: 0.81046575, Cur Avg Loss: 1.19473124, Log Avg loss: 1.05249995, Global Avg Loss: 1.53639241, Time: 0.0403 Steps: 79110, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000347, Sample Num: 5552, Cur Loss: 0.59450185, Cur Avg Loss: 1.20154530, Log Avg loss: 1.43117904, Global Avg Loss: 1.53637911, Time: 0.0403 Steps: 79120, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000357, Sample Num: 5712, Cur Loss: 1.17455149, Cur Avg Loss: 1.20066127, Log Avg loss: 1.16998537, Global Avg Loss: 1.53633281, Time: 0.0403 Steps: 79130, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000367, Sample Num: 5872, Cur Loss: 1.63208270, Cur Avg Loss: 1.20511439, Log Avg loss: 1.36409091, Global Avg Loss: 1.53631104, Time: 0.0402 Steps: 79140, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000377, Sample Num: 6032, Cur Loss: 1.23093367, Cur Avg Loss: 1.20631825, Log Avg loss: 1.25049996, Global Avg Loss: 1.53627493, Time: 0.0403 Steps: 79150, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000387, Sample Num: 6192, Cur Loss: 1.30229700, Cur Avg Loss: 1.20983660, Log Avg loss: 1.34247815, Global Avg Loss: 1.53625045, Time: 0.0403 Steps: 79160, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000397, Sample Num: 6352, Cur Loss: 1.03262472, Cur Avg Loss: 1.21186187, Log Avg loss: 1.29024001, Global Avg Loss: 1.53621938, Time: 0.0402 Steps: 79170, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000407, Sample Num: 6512, Cur Loss: 1.20511472, Cur Avg Loss: 1.21341710, Log Avg loss: 1.27515981, Global Avg Loss: 1.53618641, Time: 0.0404 Steps: 79180, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000417, Sample Num: 6672, Cur Loss: 0.82366675, Cur Avg Loss: 1.21072468, Log Avg loss: 1.10114291, Global Avg Loss: 1.53613147, Time: 0.0403 Steps: 79190, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000427, Sample Num: 6832, Cur Loss: 1.63582742, Cur Avg Loss: 1.21446969, Log Avg loss: 1.37063683, Global Avg Loss: 1.53611057, Time: 0.0402 Steps: 79200, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000437, Sample Num: 6992, Cur Loss: 1.25436831, Cur Avg Loss: 1.21874093, Log Avg loss: 1.40112271, Global Avg Loss: 1.53609353, Time: 0.0403 Steps: 79210, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000447, Sample Num: 7152, Cur Loss: 0.90755934, Cur Avg Loss: 1.21792908, Log Avg loss: 1.18245129, Global Avg Loss: 1.53604889, Time: 0.0403 Steps: 79220, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000457, Sample Num: 7312, Cur Loss: 1.51390743, Cur Avg Loss: 1.21529126, Log Avg loss: 1.09738083, Global Avg Loss: 1.53599352, Time: 0.0403 Steps: 79230, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000467, Sample Num: 7472, Cur Loss: 1.38821697, Cur Avg Loss: 1.21241651, Log Avg loss: 1.08104015, Global Avg Loss: 1.53593611, Time: 0.0403 Steps: 79240, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000477, Sample Num: 7632, Cur Loss: 0.90969270, Cur Avg Loss: 1.21658108, Log Avg loss: 1.41106665, Global Avg Loss: 1.53592035, Time: 0.0403 Steps: 79250, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000487, Sample Num: 7792, Cur Loss: 1.33409643, Cur Avg Loss: 1.21658695, Log Avg loss: 1.21686702, Global Avg Loss: 1.53588010, Time: 0.0403 Steps: 79260, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000497, Sample Num: 7952, Cur Loss: 1.02564597, Cur Avg Loss: 1.21260251, Log Avg loss: 1.01856013, Global Avg Loss: 1.53581484, Time: 0.0402 Steps: 79270, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000507, Sample Num: 8112, Cur Loss: 3.25047207, Cur Avg Loss: 1.21708950, Log Avg loss: 1.44009299, Global Avg Loss: 1.53580277, Time: 0.0402 Steps: 79280, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000517, Sample Num: 8272, Cur Loss: 1.75452578, Cur Avg Loss: 1.22079755, Log Avg loss: 1.40879562, Global Avg Loss: 1.53578675, Time: 0.0403 Steps: 79290, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000527, Sample Num: 8432, Cur Loss: 1.12841344, Cur Avg Loss: 1.21892663, Log Avg loss: 1.12220039, Global Avg Loss: 1.53573459, Time: 0.0402 Steps: 79300, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000537, Sample Num: 8592, Cur Loss: 0.84238207, Cur Avg Loss: 1.22108530, Log Avg loss: 1.33484722, Global Avg Loss: 1.53570926, Time: 0.0402 Steps: 79310, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000547, Sample Num: 8752, Cur Loss: 1.05669582, Cur Avg Loss: 1.21573792, Log Avg loss: 0.92858346, Global Avg Loss: 1.53563272, Time: 0.0402 Steps: 79320, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000557, Sample Num: 8912, Cur Loss: 0.88024628, Cur Avg Loss: 1.20947111, Log Avg loss: 0.86667639, Global Avg Loss: 1.53554840, Time: 0.0403 Steps: 79330, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000567, Sample Num: 9072, Cur Loss: 1.26909459, Cur Avg Loss: 1.20980152, Log Avg loss: 1.22820528, Global Avg Loss: 1.53550966, Time: 0.0409 Steps: 79340, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000577, Sample Num: 9232, Cur Loss: 0.42573500, Cur Avg Loss: 1.20195603, Log Avg loss: 0.75711688, Global Avg Loss: 1.53541156, Time: 0.0405 Steps: 79350, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000587, Sample Num: 9392, Cur Loss: 0.40140343, Cur Avg Loss: 1.19680433, Log Avg loss: 0.89955157, Global Avg Loss: 1.53533144, Time: 0.0557 Steps: 79360, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000597, Sample Num: 9552, Cur Loss: 1.66667998, Cur Avg Loss: 1.19705956, Log Avg loss: 1.21204106, Global Avg Loss: 1.53529071, Time: 0.0690 Steps: 79370, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000607, Sample Num: 9712, Cur Loss: 2.07715893, Cur Avg Loss: 1.19582339, Log Avg loss: 1.12202401, Global Avg Loss: 1.53523865, Time: 0.0405 Steps: 79380, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000617, Sample Num: 9872, Cur Loss: 0.90908319, Cur Avg Loss: 1.19207997, Log Avg loss: 0.96485476, Global Avg Loss: 1.53516680, Time: 0.0637 Steps: 79390, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000627, Sample Num: 10032, Cur Loss: 1.44071341, Cur Avg Loss: 1.19183558, Log Avg loss: 1.17675641, Global Avg Loss: 1.53512166, Time: 0.0406 Steps: 79400, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000637, Sample Num: 10192, Cur Loss: 1.08338368, Cur Avg Loss: 1.19159136, Log Avg loss: 1.17627877, Global Avg Loss: 1.53507647, Time: 0.0563 Steps: 79410, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000647, Sample Num: 10352, Cur Loss: 0.74772429, Cur Avg Loss: 1.18797549, Log Avg loss: 0.95764495, Global Avg Loss: 1.53500376, Time: 0.0598 Steps: 79420, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000657, Sample Num: 10512, Cur Loss: 1.58619380, Cur Avg Loss: 1.18694952, Log Avg loss: 1.12056921, Global Avg Loss: 1.53495159, Time: 0.0403 Steps: 79430, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000667, Sample Num: 10672, Cur Loss: 0.23597892, Cur Avg Loss: 1.18878674, Log Avg loss: 1.30949230, Global Avg Loss: 1.53492321, Time: 0.0403 Steps: 79440, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000677, Sample Num: 10832, Cur Loss: 0.42695355, Cur Avg Loss: 1.19053869, Log Avg loss: 1.30739371, Global Avg Loss: 1.53489457, Time: 0.0404 Steps: 79450, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000687, Sample Num: 10992, Cur Loss: 0.54374504, Cur Avg Loss: 1.18953585, Log Avg loss: 1.12164341, Global Avg Loss: 1.53484256, Time: 0.0402 Steps: 79460, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000697, Sample Num: 11152, Cur Loss: 1.00075758, Cur Avg Loss: 1.19370737, Log Avg loss: 1.48029062, Global Avg Loss: 1.53483570, Time: 0.0403 Steps: 79470, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000707, Sample Num: 11312, Cur Loss: 2.00640416, Cur Avg Loss: 1.19667523, Log Avg loss: 1.40353541, Global Avg Loss: 1.53481918, Time: 0.0403 Steps: 79480, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000717, Sample Num: 11472, Cur Loss: 0.56164455, Cur Avg Loss: 1.19596594, Log Avg loss: 1.14581875, Global Avg Loss: 1.53477024, Time: 0.0403 Steps: 79490, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000727, Sample Num: 11632, Cur Loss: 1.56206644, Cur Avg Loss: 1.19656776, Log Avg loss: 1.23971837, Global Avg Loss: 1.53473313, Time: 0.0403 Steps: 79500, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000737, Sample Num: 11792, Cur Loss: 1.12976682, Cur Avg Loss: 1.19573630, Log Avg loss: 1.13528921, Global Avg Loss: 1.53468289, Time: 0.0403 Steps: 79510, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000747, Sample Num: 11952, Cur Loss: 1.11488783, Cur Avg Loss: 1.19429097, Log Avg loss: 1.08777017, Global Avg Loss: 1.53462669, Time: 0.0403 Steps: 79520, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000757, Sample Num: 12112, Cur Loss: 0.76548529, Cur Avg Loss: 1.19061253, Log Avg loss: 0.91583293, Global Avg Loss: 1.53454888, Time: 0.0403 Steps: 79530, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000767, Sample Num: 12272, Cur Loss: 0.93242174, Cur Avg Loss: 1.19255250, Log Avg loss: 1.33940851, Global Avg Loss: 1.53452435, Time: 0.0403 Steps: 79540, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000777, Sample Num: 12432, Cur Loss: 0.97468245, Cur Avg Loss: 1.19559801, Log Avg loss: 1.42918804, Global Avg Loss: 1.53451111, Time: 0.0403 Steps: 79550, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000787, Sample Num: 12592, Cur Loss: 0.60735273, Cur Avg Loss: 1.19589719, Log Avg loss: 1.21914374, Global Avg Loss: 1.53447147, Time: 0.0403 Steps: 79560, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000797, Sample Num: 12752, Cur Loss: 1.22820938, Cur Avg Loss: 1.19520091, Log Avg loss: 1.14040404, Global Avg Loss: 1.53442194, Time: 0.0402 Steps: 79570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000807, Sample Num: 12912, Cur Loss: 0.43349338, Cur Avg Loss: 1.19452419, Log Avg loss: 1.14058969, Global Avg Loss: 1.53437245, Time: 0.0403 Steps: 79580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000817, Sample Num: 13072, Cur Loss: 1.62397623, Cur Avg Loss: 1.19671352, Log Avg loss: 1.37339195, Global Avg Loss: 1.53435223, Time: 0.0404 Steps: 79590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000827, Sample Num: 13232, Cur Loss: 0.45176095, Cur Avg Loss: 1.19465514, Log Avg loss: 1.02648535, Global Avg Loss: 1.53428843, Time: 0.0402 Steps: 79600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000837, Sample Num: 13392, Cur Loss: 1.48758900, Cur Avg Loss: 1.19336417, Log Avg loss: 1.08660107, Global Avg Loss: 1.53423219, Time: 0.0402 Steps: 79610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000847, Sample Num: 13552, Cur Loss: 0.55979097, Cur Avg Loss: 1.18886007, Log Avg loss: 0.81186722, Global Avg Loss: 1.53414146, Time: 0.0402 Steps: 79620, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000857, Sample Num: 13712, Cur Loss: 2.07515311, Cur Avg Loss: 1.18994816, Log Avg loss: 1.28210929, Global Avg Loss: 1.53410981, Time: 0.0403 Steps: 79630, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000867, Sample Num: 13872, Cur Loss: 0.36698213, Cur Avg Loss: 1.18560944, Log Avg loss: 0.81378124, Global Avg Loss: 1.53401937, Time: 0.0403 Steps: 79640, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000877, Sample Num: 14032, Cur Loss: 1.58174765, Cur Avg Loss: 1.18953505, Log Avg loss: 1.52988515, Global Avg Loss: 1.53401885, Time: 0.0402 Steps: 79650, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000887, Sample Num: 14192, Cur Loss: 2.43815517, Cur Avg Loss: 1.19071451, Log Avg loss: 1.29415358, Global Avg Loss: 1.53398873, Time: 0.0402 Steps: 79660, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000897, Sample Num: 14352, Cur Loss: 0.74228853, Cur Avg Loss: 1.19027165, Log Avg loss: 1.15098972, Global Avg Loss: 1.53394066, Time: 0.0402 Steps: 79670, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000907, Sample Num: 14512, Cur Loss: 1.14263320, Cur Avg Loss: 1.19044173, Log Avg loss: 1.20569748, Global Avg Loss: 1.53389947, Time: 0.0402 Steps: 79680, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000917, Sample Num: 14672, Cur Loss: 1.98598754, Cur Avg Loss: 1.19208253, Log Avg loss: 1.34090338, Global Avg Loss: 1.53387525, Time: 0.0675 Steps: 79690, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000927, Sample Num: 14832, Cur Loss: 0.75676274, Cur Avg Loss: 1.19339820, Log Avg loss: 1.31404555, Global Avg Loss: 1.53384767, Time: 0.0439 Steps: 79700, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000937, Sample Num: 14992, Cur Loss: 1.26651692, Cur Avg Loss: 1.19382323, Log Avg loss: 1.23322301, Global Avg Loss: 1.53380995, Time: 0.0779 Steps: 79710, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000947, Sample Num: 15152, Cur Loss: 1.54730701, Cur Avg Loss: 1.19450010, Log Avg loss: 1.25792325, Global Avg Loss: 1.53377534, Time: 0.0692 Steps: 79720, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000957, Sample Num: 15312, Cur Loss: 0.94821537, Cur Avg Loss: 1.19568916, Log Avg loss: 1.30829303, Global Avg Loss: 1.53374706, Time: 0.0463 Steps: 79730, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000967, Sample Num: 15472, Cur Loss: 0.93258274, Cur Avg Loss: 1.19524080, Log Avg loss: 1.15233237, Global Avg Loss: 1.53369923, Time: 0.0612 Steps: 79740, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000977, Sample Num: 15632, Cur Loss: 1.49646032, Cur Avg Loss: 1.19580129, Log Avg loss: 1.25000050, Global Avg Loss: 1.53366366, Time: 0.0452 Steps: 79750, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000987, Sample Num: 15792, Cur Loss: 1.95596230, Cur Avg Loss: 1.19425713, Log Avg loss: 1.04339273, Global Avg Loss: 1.53360219, Time: 0.0611 Steps: 79760, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000997, Sample Num: 15952, Cur Loss: 1.58733702, Cur Avg Loss: 1.19748681, Log Avg loss: 1.51625650, Global Avg Loss: 1.53360002, Time: 0.0661 Steps: 79770, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001007, Sample Num: 16112, Cur Loss: 3.10855055, Cur Avg Loss: 1.19830545, Log Avg loss: 1.27992437, Global Avg Loss: 1.53356822, Time: 0.0403 Steps: 79780, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001017, Sample Num: 16272, Cur Loss: 1.54373479, Cur Avg Loss: 1.19602077, Log Avg loss: 0.96595261, Global Avg Loss: 1.53349708, Time: 0.0402 Steps: 79790, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001027, Sample Num: 16432, Cur Loss: 0.55239069, Cur Avg Loss: 1.19614810, Log Avg loss: 1.20909835, Global Avg Loss: 1.53345643, Time: 0.0402 Steps: 79800, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001037, Sample Num: 16592, Cur Loss: 1.94230926, Cur Avg Loss: 1.19556886, Log Avg loss: 1.13608070, Global Avg Loss: 1.53340664, Time: 0.0403 Steps: 79810, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001047, Sample Num: 16752, Cur Loss: 1.09224713, Cur Avg Loss: 1.19378849, Log Avg loss: 1.00916405, Global Avg Loss: 1.53334096, Time: 0.0404 Steps: 79820, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001057, Sample Num: 16912, Cur Loss: 0.81552970, Cur Avg Loss: 1.19299395, Log Avg loss: 1.10980566, Global Avg Loss: 1.53328790, Time: 0.0402 Steps: 79830, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001067, Sample Num: 17072, Cur Loss: 1.23141766, Cur Avg Loss: 1.19505852, Log Avg loss: 1.41328337, Global Avg Loss: 1.53327287, Time: 0.0402 Steps: 79840, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001077, Sample Num: 17232, Cur Loss: 0.57825732, Cur Avg Loss: 1.19657716, Log Avg loss: 1.35861641, Global Avg Loss: 1.53325100, Time: 0.0403 Steps: 79850, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001087, Sample Num: 17392, Cur Loss: 2.56333518, Cur Avg Loss: 1.19850477, Log Avg loss: 1.40610830, Global Avg Loss: 1.53323508, Time: 0.0402 Steps: 79860, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001097, Sample Num: 17552, Cur Loss: 1.35780799, Cur Avg Loss: 1.20002661, Log Avg loss: 1.36545075, Global Avg Loss: 1.53321407, Time: 0.0402 Steps: 79870, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001107, Sample Num: 17712, Cur Loss: 0.40500081, Cur Avg Loss: 1.19914221, Log Avg loss: 1.10212315, Global Avg Loss: 1.53316011, Time: 0.0403 Steps: 79880, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001117, Sample Num: 17872, Cur Loss: 0.54184061, Cur Avg Loss: 1.19948018, Log Avg loss: 1.23689398, Global Avg Loss: 1.53312302, Time: 0.0403 Steps: 79890, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001127, Sample Num: 18032, Cur Loss: 1.31984019, Cur Avg Loss: 1.19938697, Log Avg loss: 1.18897503, Global Avg Loss: 1.53307995, Time: 0.0403 Steps: 79900, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001137, Sample Num: 18192, Cur Loss: 0.92571068, Cur Avg Loss: 1.19887294, Log Avg loss: 1.14094150, Global Avg Loss: 1.53303088, Time: 0.0403 Steps: 79910, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001147, Sample Num: 18352, Cur Loss: 0.59818578, Cur Avg Loss: 1.19736663, Log Avg loss: 1.02609984, Global Avg Loss: 1.53296745, Time: 0.0403 Steps: 79920, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001157, Sample Num: 18512, Cur Loss: 1.00239396, Cur Avg Loss: 1.19610748, Log Avg loss: 1.05168273, Global Avg Loss: 1.53290723, Time: 0.0402 Steps: 79930, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001167, Sample Num: 18672, Cur Loss: 1.27674162, Cur Avg Loss: 1.19429528, Log Avg loss: 0.98462312, Global Avg Loss: 1.53283865, Time: 0.0403 Steps: 79940, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001177, Sample Num: 18832, Cur Loss: 1.48754036, Cur Avg Loss: 1.19459609, Log Avg loss: 1.22970153, Global Avg Loss: 1.53280073, Time: 0.0402 Steps: 79950, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001187, Sample Num: 18992, Cur Loss: 1.95096552, Cur Avg Loss: 1.19447319, Log Avg loss: 1.18000749, Global Avg Loss: 1.53275661, Time: 0.0403 Steps: 79960, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001197, Sample Num: 19152, Cur Loss: 1.57470942, Cur Avg Loss: 1.19561571, Log Avg loss: 1.33123266, Global Avg Loss: 1.53273141, Time: 0.0403 Steps: 79970, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001207, Sample Num: 19312, Cur Loss: 1.94269490, Cur Avg Loss: 1.19588489, Log Avg loss: 1.22810535, Global Avg Loss: 1.53269332, Time: 0.0402 Steps: 79980, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001217, Sample Num: 19472, Cur Loss: 1.14920199, Cur Avg Loss: 1.19615847, Log Avg loss: 1.22918053, Global Avg Loss: 1.53265538, Time: 0.0402 Steps: 79990, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001227, Sample Num: 19632, Cur Loss: 0.82922572, Cur Avg Loss: 1.19758767, Log Avg loss: 1.37152020, Global Avg Loss: 1.53263524, Time: 0.0402 Steps: 80000, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001237, Sample Num: 19792, Cur Loss: 1.99211502, Cur Avg Loss: 1.19896231, Log Avg loss: 1.36763061, Global Avg Loss: 1.53261461, Time: 0.0404 Steps: 80010, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001247, Sample Num: 19952, Cur Loss: 2.38434935, Cur Avg Loss: 1.20174766, Log Avg loss: 1.54629657, Global Avg Loss: 1.53261632, Time: 0.0402 Steps: 80020, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001257, Sample Num: 20112, Cur Loss: 1.54668891, Cur Avg Loss: 1.20216367, Log Avg loss: 1.25403962, Global Avg Loss: 1.53258151, Time: 0.0402 Steps: 80030, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001267, Sample Num: 20272, Cur Loss: 1.13452768, Cur Avg Loss: 1.20146764, Log Avg loss: 1.11397688, Global Avg Loss: 1.53252921, Time: 0.0630 Steps: 80040, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001277, Sample Num: 20432, Cur Loss: 0.88889039, Cur Avg Loss: 1.20112587, Log Avg loss: 1.15782291, Global Avg Loss: 1.53248241, Time: 0.0480 Steps: 80050, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001287, Sample Num: 20592, Cur Loss: 0.73968339, Cur Avg Loss: 1.20109233, Log Avg loss: 1.19681005, Global Avg Loss: 1.53244048, Time: 0.0458 Steps: 80060, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001297, Sample Num: 20752, Cur Loss: 1.45678270, Cur Avg Loss: 1.20144736, Log Avg loss: 1.24713984, Global Avg Loss: 1.53240485, Time: 0.0454 Steps: 80070, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001307, Sample Num: 20912, Cur Loss: 1.45863485, Cur Avg Loss: 1.20544450, Log Avg loss: 1.72387363, Global Avg Loss: 1.53242876, Time: 0.0403 Steps: 80080, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001317, Sample Num: 21072, Cur Loss: 1.16066539, Cur Avg Loss: 1.20508146, Log Avg loss: 1.15763181, Global Avg Loss: 1.53238196, Time: 0.0643 Steps: 80090, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001327, Sample Num: 21232, Cur Loss: 0.52743542, Cur Avg Loss: 1.20602397, Log Avg loss: 1.33015280, Global Avg Loss: 1.53235671, Time: 0.0548 Steps: 80100, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001337, Sample Num: 21392, Cur Loss: 1.04970908, Cur Avg Loss: 1.20501467, Log Avg loss: 1.07107953, Global Avg Loss: 1.53229913, Time: 0.0407 Steps: 80110, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001347, Sample Num: 21552, Cur Loss: 0.72641212, Cur Avg Loss: 1.20559029, Log Avg loss: 1.28255093, Global Avg Loss: 1.53226796, Time: 0.0575 Steps: 80120, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001357, Sample Num: 21712, Cur Loss: 0.85297704, Cur Avg Loss: 1.20561992, Log Avg loss: 1.20961209, Global Avg Loss: 1.53222769, Time: 0.0403 Steps: 80130, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001367, Sample Num: 21872, Cur Loss: 0.46335977, Cur Avg Loss: 1.20540916, Log Avg loss: 1.17680830, Global Avg Loss: 1.53218334, Time: 0.0403 Steps: 80140, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001377, Sample Num: 22032, Cur Loss: 0.47331065, Cur Avg Loss: 1.20387857, Log Avg loss: 0.99464680, Global Avg Loss: 1.53211628, Time: 0.0402 Steps: 80150, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001387, Sample Num: 22192, Cur Loss: 1.71780074, Cur Avg Loss: 1.20232754, Log Avg loss: 0.98875037, Global Avg Loss: 1.53204849, Time: 0.0402 Steps: 80160, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001397, Sample Num: 22352, Cur Loss: 1.48521423, Cur Avg Loss: 1.20152564, Log Avg loss: 1.09030299, Global Avg Loss: 1.53199339, Time: 0.0402 Steps: 80170, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001407, Sample Num: 22512, Cur Loss: 1.58240223, Cur Avg Loss: 1.20036443, Log Avg loss: 1.03814338, Global Avg Loss: 1.53193180, Time: 0.0402 Steps: 80180, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001417, Sample Num: 22672, Cur Loss: 0.42900887, Cur Avg Loss: 1.19983469, Log Avg loss: 1.12530039, Global Avg Loss: 1.53188109, Time: 0.0403 Steps: 80190, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001427, Sample Num: 22832, Cur Loss: 1.03680110, Cur Avg Loss: 1.19942335, Log Avg loss: 1.14113631, Global Avg Loss: 1.53183237, Time: 0.0402 Steps: 80200, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001437, Sample Num: 22992, Cur Loss: 1.82647061, Cur Avg Loss: 1.19920399, Log Avg loss: 1.16790097, Global Avg Loss: 1.53178700, Time: 0.0402 Steps: 80210, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001447, Sample Num: 23152, Cur Loss: 0.90584588, Cur Avg Loss: 1.20016801, Log Avg loss: 1.33869737, Global Avg Loss: 1.53176293, Time: 0.0402 Steps: 80220, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001457, Sample Num: 23312, Cur Loss: 1.25611162, Cur Avg Loss: 1.20017696, Log Avg loss: 1.20147166, Global Avg Loss: 1.53172176, Time: 0.0402 Steps: 80230, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001467, Sample Num: 23472, Cur Loss: 1.27917123, Cur Avg Loss: 1.20051624, Log Avg loss: 1.24995059, Global Avg Loss: 1.53168664, Time: 0.0404 Steps: 80240, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001477, Sample Num: 23632, Cur Loss: 0.37006840, Cur Avg Loss: 1.19919388, Log Avg loss: 1.00520266, Global Avg Loss: 1.53162104, Time: 0.0403 Steps: 80250, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001487, Sample Num: 23792, Cur Loss: 0.44126230, Cur Avg Loss: 1.20099554, Log Avg loss: 1.46710055, Global Avg Loss: 1.53161300, Time: 0.0403 Steps: 80260, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001497, Sample Num: 23952, Cur Loss: 0.34103966, Cur Avg Loss: 1.20012373, Log Avg loss: 1.07048620, Global Avg Loss: 1.53155555, Time: 0.0402 Steps: 80270, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001507, Sample Num: 24112, Cur Loss: 0.66135532, Cur Avg Loss: 1.20159787, Log Avg loss: 1.42227614, Global Avg Loss: 1.53154194, Time: 0.0403 Steps: 80280, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001517, Sample Num: 24272, Cur Loss: 1.68782091, Cur Avg Loss: 1.20315174, Log Avg loss: 1.43732084, Global Avg Loss: 1.53153020, Time: 0.0402 Steps: 80290, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001527, Sample Num: 24432, Cur Loss: 2.61665058, Cur Avg Loss: 1.20575045, Log Avg loss: 1.59997442, Global Avg Loss: 1.53153873, Time: 0.0402 Steps: 80300, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001537, Sample Num: 24592, Cur Loss: 0.86535144, Cur Avg Loss: 1.20418595, Log Avg loss: 0.96528673, Global Avg Loss: 1.53146822, Time: 0.0402 Steps: 80310, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001547, Sample Num: 24752, Cur Loss: 0.93623054, Cur Avg Loss: 1.20411661, Log Avg loss: 1.19345950, Global Avg Loss: 1.53142614, Time: 0.0402 Steps: 80320, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001557, Sample Num: 24912, Cur Loss: 1.64936757, Cur Avg Loss: 1.20581341, Log Avg loss: 1.46830868, Global Avg Loss: 1.53141828, Time: 0.0402 Steps: 80330, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001567, Sample Num: 25072, Cur Loss: 1.09173524, Cur Avg Loss: 1.20575182, Log Avg loss: 1.19616173, Global Avg Loss: 1.53137655, Time: 0.0402 Steps: 80340, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001577, Sample Num: 25232, Cur Loss: 1.58248830, Cur Avg Loss: 1.20507924, Log Avg loss: 1.09968503, Global Avg Loss: 1.53132282, Time: 0.0402 Steps: 80350, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001587, Sample Num: 25392, Cur Loss: 0.69939685, Cur Avg Loss: 1.20391123, Log Avg loss: 1.01971660, Global Avg Loss: 1.53125916, Time: 0.0402 Steps: 80360, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001597, Sample Num: 25552, Cur Loss: 0.61484200, Cur Avg Loss: 1.20464981, Log Avg loss: 1.32186282, Global Avg Loss: 1.53123310, Time: 0.0402 Steps: 80370, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001607, Sample Num: 25712, Cur Loss: 1.79086876, Cur Avg Loss: 1.20485800, Log Avg loss: 1.23810583, Global Avg Loss: 1.53119664, Time: 0.0402 Steps: 80380, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001617, Sample Num: 25872, Cur Loss: 1.34658480, Cur Avg Loss: 1.20452185, Log Avg loss: 1.15050271, Global Avg Loss: 1.53114928, Time: 0.0402 Steps: 80390, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001627, Sample Num: 26032, Cur Loss: 1.39455104, Cur Avg Loss: 1.20589651, Log Avg loss: 1.42817885, Global Avg Loss: 1.53113647, Time: 0.0731 Steps: 80400, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001637, Sample Num: 26192, Cur Loss: 1.47555053, Cur Avg Loss: 1.20700930, Log Avg loss: 1.38806044, Global Avg Loss: 1.53111868, Time: 0.1007 Steps: 80410, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001647, Sample Num: 26352, Cur Loss: 0.60435134, Cur Avg Loss: 1.20499318, Log Avg loss: 0.87495357, Global Avg Loss: 1.53103709, Time: 0.0459 Steps: 80420, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001657, Sample Num: 26512, Cur Loss: 1.38866138, Cur Avg Loss: 1.20673786, Log Avg loss: 1.49408687, Global Avg Loss: 1.53103249, Time: 0.0406 Steps: 80430, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001667, Sample Num: 26672, Cur Loss: 0.36831766, Cur Avg Loss: 1.20681418, Log Avg loss: 1.21945989, Global Avg Loss: 1.53099376, Time: 0.0456 Steps: 80440, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001677, Sample Num: 26832, Cur Loss: 1.51880741, Cur Avg Loss: 1.20567738, Log Avg loss: 1.01617287, Global Avg Loss: 1.53092977, Time: 0.0821 Steps: 80450, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001687, Sample Num: 26992, Cur Loss: 2.52799463, Cur Avg Loss: 1.20650601, Log Avg loss: 1.34546813, Global Avg Loss: 1.53090672, Time: 0.0405 Steps: 80460, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001697, Sample Num: 27152, Cur Loss: 0.76481640, Cur Avg Loss: 1.20531248, Log Avg loss: 1.00396383, Global Avg Loss: 1.53084123, Time: 0.0432 Steps: 80470, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001707, Sample Num: 27312, Cur Loss: 1.04514468, Cur Avg Loss: 1.20443734, Log Avg loss: 1.05592676, Global Avg Loss: 1.53078222, Time: 0.0591 Steps: 80480, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001717, Sample Num: 27472, Cur Loss: 1.36717892, Cur Avg Loss: 1.20292618, Log Avg loss: 0.94496989, Global Avg Loss: 1.53070944, Time: 0.0402 Steps: 80490, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001727, Sample Num: 27632, Cur Loss: 2.46760011, Cur Avg Loss: 1.20462869, Log Avg loss: 1.49695092, Global Avg Loss: 1.53070525, Time: 0.0403 Steps: 80500, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001737, Sample Num: 27792, Cur Loss: 0.60251421, Cur Avg Loss: 1.20343023, Log Avg loss: 0.99645513, Global Avg Loss: 1.53063889, Time: 0.0403 Steps: 80510, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001747, Sample Num: 27952, Cur Loss: 1.20471144, Cur Avg Loss: 1.20310993, Log Avg loss: 1.14747501, Global Avg Loss: 1.53059131, Time: 0.0403 Steps: 80520, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001757, Sample Num: 28112, Cur Loss: 0.86739731, Cur Avg Loss: 1.20137228, Log Avg loss: 0.89780426, Global Avg Loss: 1.53051273, Time: 0.0402 Steps: 80530, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001767, Sample Num: 28272, Cur Loss: 0.96720231, Cur Avg Loss: 1.20168836, Log Avg loss: 1.25722348, Global Avg Loss: 1.53047880, Time: 0.0402 Steps: 80540, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001777, Sample Num: 28432, Cur Loss: 0.85655475, Cur Avg Loss: 1.20212682, Log Avg loss: 1.27960339, Global Avg Loss: 1.53044765, Time: 0.0402 Steps: 80550, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001787, Sample Num: 28592, Cur Loss: 1.34236825, Cur Avg Loss: 1.20258885, Log Avg loss: 1.28468996, Global Avg Loss: 1.53041714, Time: 0.0402 Steps: 80560, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001797, Sample Num: 28752, Cur Loss: 1.63374543, Cur Avg Loss: 1.20231390, Log Avg loss: 1.15318159, Global Avg Loss: 1.53037032, Time: 0.0403 Steps: 80570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001807, Sample Num: 28912, Cur Loss: 0.76980400, Cur Avg Loss: 1.20204540, Log Avg loss: 1.15379589, Global Avg Loss: 1.53032359, Time: 0.0402 Steps: 80580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001817, Sample Num: 29072, Cur Loss: 1.10110092, Cur Avg Loss: 1.20188990, Log Avg loss: 1.17379074, Global Avg Loss: 1.53027935, Time: 0.0402 Steps: 80590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001827, Sample Num: 29232, Cur Loss: 1.17921638, Cur Avg Loss: 1.19941045, Log Avg loss: 0.74889460, Global Avg Loss: 1.53018240, Time: 0.0403 Steps: 80600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001837, Sample Num: 29392, Cur Loss: 0.95354223, Cur Avg Loss: 1.19753070, Log Avg loss: 0.85410086, Global Avg Loss: 1.53009853, Time: 0.0403 Steps: 80610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001847, Sample Num: 29552, Cur Loss: 2.24148512, Cur Avg Loss: 1.19878287, Log Avg loss: 1.42880569, Global Avg Loss: 1.53008597, Time: 0.0403 Steps: 80620, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001857, Sample Num: 29712, Cur Loss: 0.37784886, Cur Avg Loss: 1.19842399, Log Avg loss: 1.13213948, Global Avg Loss: 1.53003661, Time: 0.0402 Steps: 80630, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001867, Sample Num: 29872, Cur Loss: 0.98370975, Cur Avg Loss: 1.19761719, Log Avg loss: 1.04779465, Global Avg Loss: 1.52997681, Time: 0.0402 Steps: 80640, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001877, Sample Num: 30032, Cur Loss: 1.20512211, Cur Avg Loss: 1.19718831, Log Avg loss: 1.11711656, Global Avg Loss: 1.52992562, Time: 0.0403 Steps: 80650, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001887, Sample Num: 30192, Cur Loss: 2.92554426, Cur Avg Loss: 1.19856215, Log Avg loss: 1.45643168, Global Avg Loss: 1.52991651, Time: 0.0402 Steps: 80660, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001897, Sample Num: 30352, Cur Loss: 0.91379333, Cur Avg Loss: 1.19767018, Log Avg loss: 1.02935389, Global Avg Loss: 1.52985446, Time: 0.0402 Steps: 80670, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001907, Sample Num: 30512, Cur Loss: 1.27184141, Cur Avg Loss: 1.19685970, Log Avg loss: 1.04311250, Global Avg Loss: 1.52979413, Time: 0.0401 Steps: 80680, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001917, Sample Num: 30672, Cur Loss: 0.85408962, Cur Avg Loss: 1.19899435, Log Avg loss: 1.60607284, Global Avg Loss: 1.52980358, Time: 0.0402 Steps: 80690, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001927, Sample Num: 30832, Cur Loss: 1.03144920, Cur Avg Loss: 1.19934314, Log Avg loss: 1.26620594, Global Avg Loss: 1.52977092, Time: 0.0402 Steps: 80700, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001937, Sample Num: 30992, Cur Loss: 1.64178419, Cur Avg Loss: 1.19861124, Log Avg loss: 1.05757369, Global Avg Loss: 1.52971241, Time: 0.0402 Steps: 80710, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001947, Sample Num: 31152, Cur Loss: 1.37499881, Cur Avg Loss: 1.19950590, Log Avg loss: 1.37280095, Global Avg Loss: 1.52969297, Time: 0.0402 Steps: 80720, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001957, Sample Num: 31312, Cur Loss: 1.07896733, Cur Avg Loss: 1.19987659, Log Avg loss: 1.27205032, Global Avg Loss: 1.52966106, Time: 0.0403 Steps: 80730, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001967, Sample Num: 31472, Cur Loss: 0.93249983, Cur Avg Loss: 1.20063874, Log Avg loss: 1.34979121, Global Avg Loss: 1.52963878, Time: 0.0723 Steps: 80740, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001977, Sample Num: 31632, Cur Loss: 0.78742921, Cur Avg Loss: 1.20214403, Log Avg loss: 1.49823537, Global Avg Loss: 1.52963489, Time: 0.0432 Steps: 80750, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001987, Sample Num: 31792, Cur Loss: 0.81682217, Cur Avg Loss: 1.20156591, Log Avg loss: 1.08727193, Global Avg Loss: 1.52958012, Time: 0.0611 Steps: 80760, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001997, Sample Num: 31952, Cur Loss: 0.29646370, Cur Avg Loss: 1.20162238, Log Avg loss: 1.21284161, Global Avg Loss: 1.52954090, Time: 0.0686 Steps: 80770, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002007, Sample Num: 32112, Cur Loss: 2.24352741, Cur Avg Loss: 1.20104892, Log Avg loss: 1.08653061, Global Avg Loss: 1.52948606, Time: 0.0415 Steps: 80780, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002017, Sample Num: 32272, Cur Loss: 1.84689331, Cur Avg Loss: 1.20080143, Log Avg loss: 1.15112848, Global Avg Loss: 1.52943923, Time: 0.0549 Steps: 80790, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002027, Sample Num: 32432, Cur Loss: 2.09289980, Cur Avg Loss: 1.20120272, Log Avg loss: 1.28214409, Global Avg Loss: 1.52940862, Time: 0.0510 Steps: 80800, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002037, Sample Num: 32592, Cur Loss: 1.12648654, Cur Avg Loss: 1.20158918, Log Avg loss: 1.27992331, Global Avg Loss: 1.52937775, Time: 0.0634 Steps: 80810, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002047, Sample Num: 32752, Cur Loss: 1.49999452, Cur Avg Loss: 1.20223918, Log Avg loss: 1.33464485, Global Avg Loss: 1.52935366, Time: 0.0874 Steps: 80820, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002057, Sample Num: 32912, Cur Loss: 0.73342645, Cur Avg Loss: 1.20210981, Log Avg loss: 1.17562861, Global Avg Loss: 1.52930989, Time: 0.0403 Steps: 80830, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002067, Sample Num: 33072, Cur Loss: 1.41379452, Cur Avg Loss: 1.20212470, Log Avg loss: 1.20518693, Global Avg Loss: 1.52926980, Time: 0.0404 Steps: 80840, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002077, Sample Num: 33232, Cur Loss: 0.94516063, Cur Avg Loss: 1.19988597, Log Avg loss: 0.73714076, Global Avg Loss: 1.52917182, Time: 0.0403 Steps: 80850, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002087, Sample Num: 33392, Cur Loss: 0.90575731, Cur Avg Loss: 1.19876685, Log Avg loss: 0.96632544, Global Avg Loss: 1.52910222, Time: 0.0403 Steps: 80860, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002097, Sample Num: 33552, Cur Loss: 1.95719028, Cur Avg Loss: 1.19957656, Log Avg loss: 1.36856332, Global Avg Loss: 1.52908237, Time: 0.0403 Steps: 80870, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002107, Sample Num: 33712, Cur Loss: 0.42365611, Cur Avg Loss: 1.19951382, Log Avg loss: 1.18635730, Global Avg Loss: 1.52903999, Time: 0.0402 Steps: 80880, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002117, Sample Num: 33872, Cur Loss: 2.20465851, Cur Avg Loss: 1.20075509, Log Avg loss: 1.46229026, Global Avg Loss: 1.52903174, Time: 0.0403 Steps: 80890, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002127, Sample Num: 34032, Cur Loss: 1.29031777, Cur Avg Loss: 1.20161637, Log Avg loss: 1.38394840, Global Avg Loss: 1.52901381, Time: 0.0403 Steps: 80900, Updated lr: 0.000024 ***** Running evaluation checkpoint-80902 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-80902 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.541551, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.193649, "eval_total_loss": 839.13541, "eval_mae": 0.931616, "eval_mse": 1.193601, "eval_r2": 0.241269, "eval_sp_statistic": 0.45132, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.555664, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.214488, "test_total_loss": 609.672985, "test_mae": 0.955897, "test_mse": 1.214624, "test_r2": 0.216071, "test_sp_statistic": 0.481199, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.596795, "test_ps_pvalue": 0.0, "lr": 2.4227596017069703e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5290100546030867, "train_cur_epoch_loss": 2558.5925887972116, "train_cur_epoch_avg_loss": 1.2017813944561822, "train_cur_epoch_time": 95.54155135154724, "train_cur_epoch_avg_time": 0.04487625709325845, "epoch": 38, "step": 80902} ################################################## Training, Epoch: 0039, Batch: 000008, Sample Num: 128, Cur Loss: 0.63553417, Cur Avg Loss: 1.04981407, Log Avg loss: 1.11530915, Global Avg Loss: 1.52896267, Time: 0.0402 Steps: 80910, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000018, Sample Num: 288, Cur Loss: 0.57154185, Cur Avg Loss: 1.05245318, Log Avg loss: 1.05456447, Global Avg Loss: 1.52890405, Time: 0.0403 Steps: 80920, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000028, Sample Num: 448, Cur Loss: 1.15876794, Cur Avg Loss: 1.17683829, Log Avg loss: 1.40073149, Global Avg Loss: 1.52888821, Time: 0.0403 Steps: 80930, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000038, Sample Num: 608, Cur Loss: 2.06181002, Cur Avg Loss: 1.15719896, Log Avg loss: 1.10220885, Global Avg Loss: 1.52883550, Time: 0.0402 Steps: 80940, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000048, Sample Num: 768, Cur Loss: 0.99690878, Cur Avg Loss: 1.19127526, Log Avg loss: 1.32076520, Global Avg Loss: 1.52880979, Time: 0.0402 Steps: 80950, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000058, Sample Num: 928, Cur Loss: 0.67532098, Cur Avg Loss: 1.21661091, Log Avg loss: 1.33822203, Global Avg Loss: 1.52878625, Time: 0.0403 Steps: 80960, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000068, Sample Num: 1088, Cur Loss: 0.55941546, Cur Avg Loss: 1.18758650, Log Avg loss: 1.01924490, Global Avg Loss: 1.52872332, Time: 0.0403 Steps: 80970, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000078, Sample Num: 1248, Cur Loss: 1.16556323, Cur Avg Loss: 1.17304446, Log Avg loss: 1.07415858, Global Avg Loss: 1.52866719, Time: 0.0403 Steps: 80980, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000088, Sample Num: 1408, Cur Loss: 1.46421885, Cur Avg Loss: 1.15917452, Log Avg loss: 1.05098903, Global Avg Loss: 1.52860821, Time: 0.0403 Steps: 80990, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000098, Sample Num: 1568, Cur Loss: 1.04528046, Cur Avg Loss: 1.17249203, Log Avg loss: 1.28968614, Global Avg Loss: 1.52857871, Time: 0.0403 Steps: 81000, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000108, Sample Num: 1728, Cur Loss: 0.73037219, Cur Avg Loss: 1.14974145, Log Avg loss: 0.92678571, Global Avg Loss: 1.52850443, Time: 0.0403 Steps: 81010, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000118, Sample Num: 1888, Cur Loss: 1.70367372, Cur Avg Loss: 1.13632753, Log Avg loss: 0.99145725, Global Avg Loss: 1.52843814, Time: 0.0402 Steps: 81020, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000128, Sample Num: 2048, Cur Loss: 1.29767346, Cur Avg Loss: 1.13660753, Log Avg loss: 1.13991152, Global Avg Loss: 1.52839019, Time: 0.0403 Steps: 81030, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000138, Sample Num: 2208, Cur Loss: 0.21615934, Cur Avg Loss: 1.15642552, Log Avg loss: 1.41009571, Global Avg Loss: 1.52837559, Time: 0.0403 Steps: 81040, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000148, Sample Num: 2368, Cur Loss: 1.06432498, Cur Avg Loss: 1.16517597, Log Avg loss: 1.28593221, Global Avg Loss: 1.52834568, Time: 0.0402 Steps: 81050, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000158, Sample Num: 2528, Cur Loss: 1.19900620, Cur Avg Loss: 1.17466835, Log Avg loss: 1.31515567, Global Avg Loss: 1.52831938, Time: 0.0403 Steps: 81060, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000168, Sample Num: 2688, Cur Loss: 1.28719211, Cur Avg Loss: 1.18403704, Log Avg loss: 1.33206222, Global Avg Loss: 1.52829517, Time: 0.0403 Steps: 81070, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000178, Sample Num: 2848, Cur Loss: 1.49811542, Cur Avg Loss: 1.19279832, Log Avg loss: 1.33998790, Global Avg Loss: 1.52827195, Time: 0.0402 Steps: 81080, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000188, Sample Num: 3008, Cur Loss: 0.93654823, Cur Avg Loss: 1.18389730, Log Avg loss: 1.02545910, Global Avg Loss: 1.52820994, Time: 0.0403 Steps: 81090, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000198, Sample Num: 3168, Cur Loss: 0.24186309, Cur Avg Loss: 1.17577932, Log Avg loss: 1.02316123, Global Avg Loss: 1.52814767, Time: 0.0402 Steps: 81100, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000208, Sample Num: 3328, Cur Loss: 1.43461275, Cur Avg Loss: 1.17058051, Log Avg loss: 1.06764417, Global Avg Loss: 1.52809089, Time: 0.0403 Steps: 81110, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000218, Sample Num: 3488, Cur Loss: 1.61376834, Cur Avg Loss: 1.16256027, Log Avg loss: 0.99573923, Global Avg Loss: 1.52802527, Time: 0.0403 Steps: 81120, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000228, Sample Num: 3648, Cur Loss: 0.91527486, Cur Avg Loss: 1.16977792, Log Avg loss: 1.32712275, Global Avg Loss: 1.52800050, Time: 0.0987 Steps: 81130, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000238, Sample Num: 3808, Cur Loss: 1.36685753, Cur Avg Loss: 1.17546751, Log Avg loss: 1.30519017, Global Avg Loss: 1.52797304, Time: 0.0510 Steps: 81140, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000248, Sample Num: 3968, Cur Loss: 0.50476849, Cur Avg Loss: 1.18067961, Log Avg loss: 1.30472758, Global Avg Loss: 1.52794553, Time: 0.0490 Steps: 81150, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000258, Sample Num: 4128, Cur Loss: 1.74501038, Cur Avg Loss: 1.18021698, Log Avg loss: 1.16874359, Global Avg Loss: 1.52790127, Time: 0.0534 Steps: 81160, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000268, Sample Num: 4288, Cur Loss: 0.44169766, Cur Avg Loss: 1.18268792, Log Avg loss: 1.24643816, Global Avg Loss: 1.52786660, Time: 0.0404 Steps: 81170, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000278, Sample Num: 4448, Cur Loss: 0.39579934, Cur Avg Loss: 1.17917311, Log Avg loss: 1.08497630, Global Avg Loss: 1.52781204, Time: 0.0583 Steps: 81180, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000288, Sample Num: 4608, Cur Loss: 1.50537264, Cur Avg Loss: 1.18511799, Log Avg loss: 1.35038580, Global Avg Loss: 1.52779019, Time: 0.0408 Steps: 81190, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000298, Sample Num: 4768, Cur Loss: 0.54765785, Cur Avg Loss: 1.17798415, Log Avg loss: 0.97252944, Global Avg Loss: 1.52772181, Time: 0.0471 Steps: 81200, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000308, Sample Num: 4928, Cur Loss: 0.35590333, Cur Avg Loss: 1.18076137, Log Avg loss: 1.26352247, Global Avg Loss: 1.52768927, Time: 0.0707 Steps: 81210, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000318, Sample Num: 5088, Cur Loss: 0.50988948, Cur Avg Loss: 1.17104884, Log Avg loss: 0.87190302, Global Avg Loss: 1.52760853, Time: 0.0403 Steps: 81220, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000328, Sample Num: 5248, Cur Loss: 1.00228965, Cur Avg Loss: 1.17376805, Log Avg loss: 1.26023874, Global Avg Loss: 1.52757562, Time: 0.0403 Steps: 81230, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000338, Sample Num: 5408, Cur Loss: 1.46299255, Cur Avg Loss: 1.17409265, Log Avg loss: 1.18473974, Global Avg Loss: 1.52753342, Time: 0.0403 Steps: 81240, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000348, Sample Num: 5568, Cur Loss: 1.06506467, Cur Avg Loss: 1.17094035, Log Avg loss: 1.06439257, Global Avg Loss: 1.52747641, Time: 0.0403 Steps: 81250, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000358, Sample Num: 5728, Cur Loss: 0.87949437, Cur Avg Loss: 1.17297266, Log Avg loss: 1.24369697, Global Avg Loss: 1.52744149, Time: 0.0402 Steps: 81260, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000368, Sample Num: 5888, Cur Loss: 1.42363286, Cur Avg Loss: 1.17897573, Log Avg loss: 1.39388565, Global Avg Loss: 1.52742506, Time: 0.0402 Steps: 81270, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000378, Sample Num: 6048, Cur Loss: 0.85320663, Cur Avg Loss: 1.17935237, Log Avg loss: 1.19321290, Global Avg Loss: 1.52738394, Time: 0.0402 Steps: 81280, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000388, Sample Num: 6208, Cur Loss: 1.37442303, Cur Avg Loss: 1.18399382, Log Avg loss: 1.35944054, Global Avg Loss: 1.52736328, Time: 0.0403 Steps: 81290, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000398, Sample Num: 6368, Cur Loss: 1.34556651, Cur Avg Loss: 1.17916159, Log Avg loss: 0.99167089, Global Avg Loss: 1.52729739, Time: 0.0403 Steps: 81300, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000408, Sample Num: 6528, Cur Loss: 1.28445089, Cur Avg Loss: 1.18149959, Log Avg loss: 1.27455224, Global Avg Loss: 1.52726631, Time: 0.0403 Steps: 81310, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000418, Sample Num: 6688, Cur Loss: 1.02986193, Cur Avg Loss: 1.18912726, Log Avg loss: 1.50033602, Global Avg Loss: 1.52726299, Time: 0.0402 Steps: 81320, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000428, Sample Num: 6848, Cur Loss: 1.29520965, Cur Avg Loss: 1.18510684, Log Avg loss: 1.01705353, Global Avg Loss: 1.52720026, Time: 0.0403 Steps: 81330, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000438, Sample Num: 7008, Cur Loss: 0.78595549, Cur Avg Loss: 1.18210509, Log Avg loss: 1.05363020, Global Avg Loss: 1.52714204, Time: 0.0402 Steps: 81340, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000448, Sample Num: 7168, Cur Loss: 0.97170252, Cur Avg Loss: 1.17800426, Log Avg loss: 0.99838750, Global Avg Loss: 1.52707704, Time: 0.0402 Steps: 81350, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000458, Sample Num: 7328, Cur Loss: 1.80124354, Cur Avg Loss: 1.17798212, Log Avg loss: 1.17699064, Global Avg Loss: 1.52703401, Time: 0.0402 Steps: 81360, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000468, Sample Num: 7488, Cur Loss: 1.42948747, Cur Avg Loss: 1.18103948, Log Avg loss: 1.32106618, Global Avg Loss: 1.52700870, Time: 0.0402 Steps: 81370, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000478, Sample Num: 7648, Cur Loss: 1.81550813, Cur Avg Loss: 1.18279395, Log Avg loss: 1.26490346, Global Avg Loss: 1.52697649, Time: 0.0403 Steps: 81380, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000488, Sample Num: 7808, Cur Loss: 2.67079353, Cur Avg Loss: 1.18300146, Log Avg loss: 1.19292018, Global Avg Loss: 1.52693545, Time: 0.0402 Steps: 81390, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000498, Sample Num: 7968, Cur Loss: 1.05585575, Cur Avg Loss: 1.18303857, Log Avg loss: 1.18484961, Global Avg Loss: 1.52689342, Time: 0.0402 Steps: 81400, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000508, Sample Num: 8128, Cur Loss: 2.81764030, Cur Avg Loss: 1.18108067, Log Avg loss: 1.08357754, Global Avg Loss: 1.52683897, Time: 0.0402 Steps: 81410, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000518, Sample Num: 8288, Cur Loss: 1.46257317, Cur Avg Loss: 1.18640893, Log Avg loss: 1.45708457, Global Avg Loss: 1.52683040, Time: 0.0402 Steps: 81420, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000528, Sample Num: 8448, Cur Loss: 1.08006454, Cur Avg Loss: 1.18250452, Log Avg loss: 0.98025602, Global Avg Loss: 1.52676328, Time: 0.0402 Steps: 81430, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000538, Sample Num: 8608, Cur Loss: 1.23742115, Cur Avg Loss: 1.18024903, Log Avg loss: 1.06115900, Global Avg Loss: 1.52670611, Time: 0.0402 Steps: 81440, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000548, Sample Num: 8768, Cur Loss: 1.94529486, Cur Avg Loss: 1.18544871, Log Avg loss: 1.46519130, Global Avg Loss: 1.52669856, Time: 0.0403 Steps: 81450, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000558, Sample Num: 8928, Cur Loss: 1.46593428, Cur Avg Loss: 1.19015492, Log Avg loss: 1.44805545, Global Avg Loss: 1.52668890, Time: 0.0402 Steps: 81460, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000568, Sample Num: 9088, Cur Loss: 0.64637756, Cur Avg Loss: 1.19327939, Log Avg loss: 1.36762488, Global Avg Loss: 1.52666938, Time: 0.0627 Steps: 81470, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000578, Sample Num: 9248, Cur Loss: 0.58487439, Cur Avg Loss: 1.19090741, Log Avg loss: 1.05617864, Global Avg Loss: 1.52661163, Time: 0.0589 Steps: 81480, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000588, Sample Num: 9408, Cur Loss: 1.32269669, Cur Avg Loss: 1.18927069, Log Avg loss: 1.09466841, Global Avg Loss: 1.52655863, Time: 0.0667 Steps: 81490, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000598, Sample Num: 9568, Cur Loss: 0.85064131, Cur Avg Loss: 1.18896662, Log Avg loss: 1.17108738, Global Avg Loss: 1.52651501, Time: 0.0479 Steps: 81500, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000608, Sample Num: 9728, Cur Loss: 1.73773479, Cur Avg Loss: 1.19147438, Log Avg loss: 1.34143869, Global Avg Loss: 1.52649231, Time: 0.0587 Steps: 81510, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000618, Sample Num: 9888, Cur Loss: 0.88872337, Cur Avg Loss: 1.18834938, Log Avg loss: 0.99834919, Global Avg Loss: 1.52642752, Time: 0.0418 Steps: 81520, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000628, Sample Num: 10048, Cur Loss: 1.60503578, Cur Avg Loss: 1.18765404, Log Avg loss: 1.14468170, Global Avg Loss: 1.52638070, Time: 0.0624 Steps: 81530, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000638, Sample Num: 10208, Cur Loss: 1.04766738, Cur Avg Loss: 1.18907168, Log Avg loss: 1.27809983, Global Avg Loss: 1.52635025, Time: 0.0428 Steps: 81540, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000648, Sample Num: 10368, Cur Loss: 1.50179636, Cur Avg Loss: 1.18793072, Log Avg loss: 1.11513763, Global Avg Loss: 1.52629982, Time: 0.0522 Steps: 81550, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000658, Sample Num: 10528, Cur Loss: 1.14600921, Cur Avg Loss: 1.18677395, Log Avg loss: 1.11181469, Global Avg Loss: 1.52624900, Time: 0.0402 Steps: 81560, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000668, Sample Num: 10688, Cur Loss: 0.74430019, Cur Avg Loss: 1.18089446, Log Avg loss: 0.79402439, Global Avg Loss: 1.52615924, Time: 0.0403 Steps: 81570, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000678, Sample Num: 10848, Cur Loss: 0.99935329, Cur Avg Loss: 1.18133586, Log Avg loss: 1.21082135, Global Avg Loss: 1.52612058, Time: 0.0402 Steps: 81580, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000688, Sample Num: 11008, Cur Loss: 0.84508032, Cur Avg Loss: 1.17925009, Log Avg loss: 1.03783477, Global Avg Loss: 1.52606074, Time: 0.0402 Steps: 81590, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000698, Sample Num: 11168, Cur Loss: 1.12718165, Cur Avg Loss: 1.17335989, Log Avg loss: 0.76811396, Global Avg Loss: 1.52596785, Time: 0.0403 Steps: 81600, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000708, Sample Num: 11328, Cur Loss: 1.48809576, Cur Avg Loss: 1.17048481, Log Avg loss: 0.96980435, Global Avg Loss: 1.52589970, Time: 0.0402 Steps: 81610, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000718, Sample Num: 11488, Cur Loss: 0.28523016, Cur Avg Loss: 1.17159002, Log Avg loss: 1.24983916, Global Avg Loss: 1.52586588, Time: 0.0402 Steps: 81620, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000728, Sample Num: 11648, Cur Loss: 1.36319530, Cur Avg Loss: 1.17350105, Log Avg loss: 1.31071277, Global Avg Loss: 1.52583952, Time: 0.0402 Steps: 81630, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000738, Sample Num: 11808, Cur Loss: 1.26225936, Cur Avg Loss: 1.17135464, Log Avg loss: 1.01509628, Global Avg Loss: 1.52577696, Time: 0.0402 Steps: 81640, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000748, Sample Num: 11968, Cur Loss: 1.49477148, Cur Avg Loss: 1.16894934, Log Avg loss: 0.99143817, Global Avg Loss: 1.52571152, Time: 0.0403 Steps: 81650, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000758, Sample Num: 12128, Cur Loss: 1.15816283, Cur Avg Loss: 1.16691964, Log Avg loss: 1.01509781, Global Avg Loss: 1.52564899, Time: 0.0402 Steps: 81660, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000768, Sample Num: 12288, Cur Loss: 0.39235422, Cur Avg Loss: 1.16140812, Log Avg loss: 0.74363488, Global Avg Loss: 1.52555324, Time: 0.0403 Steps: 81670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000778, Sample Num: 12448, Cur Loss: 0.96420616, Cur Avg Loss: 1.16502983, Log Avg loss: 1.44317710, Global Avg Loss: 1.52554315, Time: 0.0402 Steps: 81680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000788, Sample Num: 12608, Cur Loss: 0.47111204, Cur Avg Loss: 1.16060665, Log Avg loss: 0.81648312, Global Avg Loss: 1.52545635, Time: 0.0402 Steps: 81690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000798, Sample Num: 12768, Cur Loss: 0.16987164, Cur Avg Loss: 1.15984601, Log Avg loss: 1.09990746, Global Avg Loss: 1.52540427, Time: 0.0402 Steps: 81700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000808, Sample Num: 12928, Cur Loss: 0.42307547, Cur Avg Loss: 1.16037275, Log Avg loss: 1.20240729, Global Avg Loss: 1.52536474, Time: 0.0402 Steps: 81710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000818, Sample Num: 13088, Cur Loss: 0.93406010, Cur Avg Loss: 1.15974285, Log Avg loss: 1.10884644, Global Avg Loss: 1.52531377, Time: 0.0402 Steps: 81720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000828, Sample Num: 13248, Cur Loss: 2.34430075, Cur Avg Loss: 1.16332405, Log Avg loss: 1.45626664, Global Avg Loss: 1.52530532, Time: 0.0402 Steps: 81730, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000838, Sample Num: 13408, Cur Loss: 0.82195258, Cur Avg Loss: 1.16290064, Log Avg loss: 1.12784238, Global Avg Loss: 1.52525669, Time: 0.0402 Steps: 81740, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000848, Sample Num: 13568, Cur Loss: 1.00903809, Cur Avg Loss: 1.16517690, Log Avg loss: 1.35592702, Global Avg Loss: 1.52523598, Time: 0.0403 Steps: 81750, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000858, Sample Num: 13728, Cur Loss: 2.06502032, Cur Avg Loss: 1.16783809, Log Avg loss: 1.39350703, Global Avg Loss: 1.52521987, Time: 0.0403 Steps: 81760, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000868, Sample Num: 13888, Cur Loss: 1.25039554, Cur Avg Loss: 1.17077478, Log Avg loss: 1.42274316, Global Avg Loss: 1.52520734, Time: 0.0402 Steps: 81770, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000878, Sample Num: 14048, Cur Loss: 1.80142570, Cur Avg Loss: 1.16910927, Log Avg loss: 1.02454245, Global Avg Loss: 1.52514612, Time: 0.0402 Steps: 81780, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000888, Sample Num: 14208, Cur Loss: 0.55459672, Cur Avg Loss: 1.16809886, Log Avg loss: 1.07938481, Global Avg Loss: 1.52509162, Time: 0.0402 Steps: 81790, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000898, Sample Num: 14368, Cur Loss: 2.37359667, Cur Avg Loss: 1.16980902, Log Avg loss: 1.32167178, Global Avg Loss: 1.52506675, Time: 0.0402 Steps: 81800, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000908, Sample Num: 14528, Cur Loss: 0.23230156, Cur Avg Loss: 1.16997768, Log Avg loss: 1.18512271, Global Avg Loss: 1.52502519, Time: 0.0402 Steps: 81810, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000918, Sample Num: 14688, Cur Loss: 1.68554187, Cur Avg Loss: 1.17222404, Log Avg loss: 1.37619348, Global Avg Loss: 1.52500700, Time: 0.0402 Steps: 81820, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000928, Sample Num: 14848, Cur Loss: 0.65041268, Cur Avg Loss: 1.16949840, Log Avg loss: 0.91928495, Global Avg Loss: 1.52493298, Time: 0.0404 Steps: 81830, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000938, Sample Num: 15008, Cur Loss: 1.01315832, Cur Avg Loss: 1.17001246, Log Avg loss: 1.21771739, Global Avg Loss: 1.52489544, Time: 0.0484 Steps: 81840, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000948, Sample Num: 15168, Cur Loss: 1.28114629, Cur Avg Loss: 1.16865472, Log Avg loss: 1.04129837, Global Avg Loss: 1.52483636, Time: 0.0406 Steps: 81850, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000958, Sample Num: 15328, Cur Loss: 2.17726374, Cur Avg Loss: 1.16850661, Log Avg loss: 1.15446609, Global Avg Loss: 1.52479112, Time: 0.0947 Steps: 81860, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000968, Sample Num: 15488, Cur Loss: 0.72107494, Cur Avg Loss: 1.16850988, Log Avg loss: 1.16882351, Global Avg Loss: 1.52474764, Time: 0.1005 Steps: 81870, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000978, Sample Num: 15648, Cur Loss: 0.63923067, Cur Avg Loss: 1.16701191, Log Avg loss: 1.02200818, Global Avg Loss: 1.52468624, Time: 0.0404 Steps: 81880, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000988, Sample Num: 15808, Cur Loss: 1.44416952, Cur Avg Loss: 1.16740361, Log Avg loss: 1.20571151, Global Avg Loss: 1.52464729, Time: 0.0405 Steps: 81890, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000998, Sample Num: 15968, Cur Loss: 0.92436039, Cur Avg Loss: 1.16862148, Log Avg loss: 1.28894683, Global Avg Loss: 1.52461851, Time: 0.0406 Steps: 81900, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001008, Sample Num: 16128, Cur Loss: 1.60292006, Cur Avg Loss: 1.16881991, Log Avg loss: 1.18862369, Global Avg Loss: 1.52457749, Time: 0.0404 Steps: 81910, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001018, Sample Num: 16288, Cur Loss: 1.61717892, Cur Avg Loss: 1.16874544, Log Avg loss: 1.16123863, Global Avg Loss: 1.52453313, Time: 0.0523 Steps: 81920, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001028, Sample Num: 16448, Cur Loss: 2.51178837, Cur Avg Loss: 1.17011330, Log Avg loss: 1.30936189, Global Avg Loss: 1.52450687, Time: 0.0403 Steps: 81930, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001038, Sample Num: 16608, Cur Loss: 0.64306718, Cur Avg Loss: 1.16992264, Log Avg loss: 1.15032243, Global Avg Loss: 1.52446120, Time: 0.0403 Steps: 81940, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001048, Sample Num: 16768, Cur Loss: 1.09886920, Cur Avg Loss: 1.17193272, Log Avg loss: 1.38057937, Global Avg Loss: 1.52444365, Time: 0.0403 Steps: 81950, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001058, Sample Num: 16928, Cur Loss: 2.09447479, Cur Avg Loss: 1.17311788, Log Avg loss: 1.29732251, Global Avg Loss: 1.52441594, Time: 0.0403 Steps: 81960, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001068, Sample Num: 17088, Cur Loss: 0.52811313, Cur Avg Loss: 1.17265082, Log Avg loss: 1.12323578, Global Avg Loss: 1.52436699, Time: 0.0402 Steps: 81970, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001078, Sample Num: 17248, Cur Loss: 3.12404609, Cur Avg Loss: 1.17033838, Log Avg loss: 0.92336992, Global Avg Loss: 1.52429368, Time: 0.0403 Steps: 81980, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001088, Sample Num: 17408, Cur Loss: 1.00944901, Cur Avg Loss: 1.16968538, Log Avg loss: 1.09929144, Global Avg Loss: 1.52424185, Time: 0.0402 Steps: 81990, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001098, Sample Num: 17568, Cur Loss: 1.28906083, Cur Avg Loss: 1.16805635, Log Avg loss: 0.99081804, Global Avg Loss: 1.52417680, Time: 0.0402 Steps: 82000, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001108, Sample Num: 17728, Cur Loss: 1.06449604, Cur Avg Loss: 1.16647389, Log Avg loss: 0.99272049, Global Avg Loss: 1.52411199, Time: 0.0405 Steps: 82010, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001118, Sample Num: 17888, Cur Loss: 1.04336596, Cur Avg Loss: 1.16624876, Log Avg loss: 1.14130438, Global Avg Loss: 1.52406532, Time: 0.0402 Steps: 82020, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001128, Sample Num: 18048, Cur Loss: 1.00796473, Cur Avg Loss: 1.16623684, Log Avg loss: 1.16490320, Global Avg Loss: 1.52402154, Time: 0.0403 Steps: 82030, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001138, Sample Num: 18208, Cur Loss: 1.50018740, Cur Avg Loss: 1.16546712, Log Avg loss: 1.07864292, Global Avg Loss: 1.52396725, Time: 0.0402 Steps: 82040, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001148, Sample Num: 18368, Cur Loss: 0.64839041, Cur Avg Loss: 1.16576745, Log Avg loss: 1.19994569, Global Avg Loss: 1.52392776, Time: 0.0403 Steps: 82050, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001158, Sample Num: 18528, Cur Loss: 0.69779170, Cur Avg Loss: 1.16498377, Log Avg loss: 1.07501723, Global Avg Loss: 1.52387305, Time: 0.0402 Steps: 82060, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001168, Sample Num: 18688, Cur Loss: 1.50348687, Cur Avg Loss: 1.16700440, Log Avg loss: 1.40099356, Global Avg Loss: 1.52385808, Time: 0.0402 Steps: 82070, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001178, Sample Num: 18848, Cur Loss: 0.44576111, Cur Avg Loss: 1.16589723, Log Avg loss: 1.03657970, Global Avg Loss: 1.52379871, Time: 0.0402 Steps: 82080, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001188, Sample Num: 19008, Cur Loss: 1.05432546, Cur Avg Loss: 1.16960427, Log Avg loss: 1.60629361, Global Avg Loss: 1.52380876, Time: 0.0403 Steps: 82090, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001198, Sample Num: 19168, Cur Loss: 0.60804325, Cur Avg Loss: 1.17042823, Log Avg loss: 1.26831454, Global Avg Loss: 1.52377764, Time: 0.0402 Steps: 82100, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001208, Sample Num: 19328, Cur Loss: 0.71427423, Cur Avg Loss: 1.16761789, Log Avg loss: 0.83093924, Global Avg Loss: 1.52369326, Time: 0.0402 Steps: 82110, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001218, Sample Num: 19488, Cur Loss: 0.69373232, Cur Avg Loss: 1.16679373, Log Avg loss: 1.06723520, Global Avg Loss: 1.52363768, Time: 0.0402 Steps: 82120, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001228, Sample Num: 19648, Cur Loss: 1.23084950, Cur Avg Loss: 1.16991690, Log Avg loss: 1.55031924, Global Avg Loss: 1.52364093, Time: 0.0403 Steps: 82130, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001238, Sample Num: 19808, Cur Loss: 0.45751390, Cur Avg Loss: 1.16806306, Log Avg loss: 0.94041045, Global Avg Loss: 1.52356992, Time: 0.0402 Steps: 82140, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001248, Sample Num: 19968, Cur Loss: 0.38628817, Cur Avg Loss: 1.16880301, Log Avg loss: 1.26040969, Global Avg Loss: 1.52353789, Time: 0.0402 Steps: 82150, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001258, Sample Num: 20128, Cur Loss: 2.25524092, Cur Avg Loss: 1.16844994, Log Avg loss: 1.12438640, Global Avg Loss: 1.52348931, Time: 0.0402 Steps: 82160, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001268, Sample Num: 20288, Cur Loss: 0.45932883, Cur Avg Loss: 1.16860788, Log Avg loss: 1.18847713, Global Avg Loss: 1.52344854, Time: 0.0402 Steps: 82170, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001278, Sample Num: 20448, Cur Loss: 0.47338995, Cur Avg Loss: 1.16881204, Log Avg loss: 1.19469899, Global Avg Loss: 1.52340853, Time: 0.0403 Steps: 82180, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001288, Sample Num: 20608, Cur Loss: 1.23951888, Cur Avg Loss: 1.16878894, Log Avg loss: 1.16583739, Global Avg Loss: 1.52336503, Time: 0.0560 Steps: 82190, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001298, Sample Num: 20768, Cur Loss: 0.76114601, Cur Avg Loss: 1.16823238, Log Avg loss: 1.09654756, Global Avg Loss: 1.52331310, Time: 0.0794 Steps: 82200, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001308, Sample Num: 20928, Cur Loss: 0.65731102, Cur Avg Loss: 1.16616001, Log Avg loss: 0.89716636, Global Avg Loss: 1.52323694, Time: 0.0650 Steps: 82210, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001318, Sample Num: 21088, Cur Loss: 0.53552693, Cur Avg Loss: 1.16423955, Log Avg loss: 0.91304295, Global Avg Loss: 1.52316272, Time: 0.0564 Steps: 82220, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001328, Sample Num: 21248, Cur Loss: 3.44317865, Cur Avg Loss: 1.16599865, Log Avg loss: 1.39784762, Global Avg Loss: 1.52314748, Time: 0.0736 Steps: 82230, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001338, Sample Num: 21408, Cur Loss: 0.64719212, Cur Avg Loss: 1.16477193, Log Avg loss: 1.00186388, Global Avg Loss: 1.52308410, Time: 0.0881 Steps: 82240, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001348, Sample Num: 21568, Cur Loss: 1.29131484, Cur Avg Loss: 1.16503789, Log Avg loss: 1.20062336, Global Avg Loss: 1.52304489, Time: 0.0430 Steps: 82250, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001358, Sample Num: 21728, Cur Loss: 0.59031290, Cur Avg Loss: 1.16371652, Log Avg loss: 0.98559524, Global Avg Loss: 1.52297956, Time: 0.0568 Steps: 82260, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001368, Sample Num: 21888, Cur Loss: 0.93781382, Cur Avg Loss: 1.16300306, Log Avg loss: 1.06611624, Global Avg Loss: 1.52292403, Time: 0.0407 Steps: 82270, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001378, Sample Num: 22048, Cur Loss: 1.70349407, Cur Avg Loss: 1.16215532, Log Avg loss: 1.04618349, Global Avg Loss: 1.52286608, Time: 0.0403 Steps: 82280, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001388, Sample Num: 22208, Cur Loss: 0.90442157, Cur Avg Loss: 1.16240210, Log Avg loss: 1.19640894, Global Avg Loss: 1.52282641, Time: 0.0403 Steps: 82290, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001398, Sample Num: 22368, Cur Loss: 1.70914435, Cur Avg Loss: 1.16375778, Log Avg loss: 1.35192580, Global Avg Loss: 1.52280565, Time: 0.0403 Steps: 82300, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001408, Sample Num: 22528, Cur Loss: 0.81921029, Cur Avg Loss: 1.16611770, Log Avg loss: 1.49603445, Global Avg Loss: 1.52280240, Time: 0.0403 Steps: 82310, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001418, Sample Num: 22688, Cur Loss: 1.31844068, Cur Avg Loss: 1.16613002, Log Avg loss: 1.16786450, Global Avg Loss: 1.52275928, Time: 0.0403 Steps: 82320, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001428, Sample Num: 22848, Cur Loss: 1.60133338, Cur Avg Loss: 1.16696937, Log Avg loss: 1.28599016, Global Avg Loss: 1.52273052, Time: 0.0402 Steps: 82330, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001438, Sample Num: 23008, Cur Loss: 0.72248554, Cur Avg Loss: 1.16823286, Log Avg loss: 1.34865940, Global Avg Loss: 1.52270938, Time: 0.0402 Steps: 82340, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001448, Sample Num: 23168, Cur Loss: 1.44072783, Cur Avg Loss: 1.16870104, Log Avg loss: 1.23602492, Global Avg Loss: 1.52267457, Time: 0.0403 Steps: 82350, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001458, Sample Num: 23328, Cur Loss: 0.82354206, Cur Avg Loss: 1.16898694, Log Avg loss: 1.21038537, Global Avg Loss: 1.52263665, Time: 0.0403 Steps: 82360, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001468, Sample Num: 23488, Cur Loss: 1.41456151, Cur Avg Loss: 1.17130540, Log Avg loss: 1.50933683, Global Avg Loss: 1.52263503, Time: 0.0403 Steps: 82370, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001478, Sample Num: 23648, Cur Loss: 1.65430379, Cur Avg Loss: 1.17282279, Log Avg loss: 1.39557470, Global Avg Loss: 1.52261961, Time: 0.0402 Steps: 82380, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001488, Sample Num: 23808, Cur Loss: 0.74972332, Cur Avg Loss: 1.17249089, Log Avg loss: 1.12343677, Global Avg Loss: 1.52257116, Time: 0.0403 Steps: 82390, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001498, Sample Num: 23968, Cur Loss: 0.70974827, Cur Avg Loss: 1.17427456, Log Avg loss: 1.43968491, Global Avg Loss: 1.52256110, Time: 0.0403 Steps: 82400, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001508, Sample Num: 24128, Cur Loss: 1.10150266, Cur Avg Loss: 1.17288310, Log Avg loss: 0.96444175, Global Avg Loss: 1.52249338, Time: 0.0402 Steps: 82410, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001518, Sample Num: 24288, Cur Loss: 0.53495777, Cur Avg Loss: 1.17382366, Log Avg loss: 1.31565992, Global Avg Loss: 1.52246828, Time: 0.0403 Steps: 82420, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001528, Sample Num: 24448, Cur Loss: 2.15004802, Cur Avg Loss: 1.17373934, Log Avg loss: 1.16094074, Global Avg Loss: 1.52242442, Time: 0.0402 Steps: 82430, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001538, Sample Num: 24608, Cur Loss: 0.65451050, Cur Avg Loss: 1.17506249, Log Avg loss: 1.37723867, Global Avg Loss: 1.52240681, Time: 0.0402 Steps: 82440, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001548, Sample Num: 24768, Cur Loss: 1.80803990, Cur Avg Loss: 1.17394738, Log Avg loss: 1.00244382, Global Avg Loss: 1.52234375, Time: 0.0402 Steps: 82450, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001558, Sample Num: 24928, Cur Loss: 0.64362991, Cur Avg Loss: 1.17538603, Log Avg loss: 1.39808884, Global Avg Loss: 1.52232868, Time: 0.0402 Steps: 82460, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001568, Sample Num: 25088, Cur Loss: 1.02100551, Cur Avg Loss: 1.17642294, Log Avg loss: 1.33797452, Global Avg Loss: 1.52230632, Time: 0.0402 Steps: 82470, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001578, Sample Num: 25248, Cur Loss: 0.57958174, Cur Avg Loss: 1.17500404, Log Avg loss: 0.95251958, Global Avg Loss: 1.52223724, Time: 0.0402 Steps: 82480, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001588, Sample Num: 25408, Cur Loss: 0.51543170, Cur Avg Loss: 1.17374251, Log Avg loss: 0.97467379, Global Avg Loss: 1.52217086, Time: 0.0402 Steps: 82490, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001598, Sample Num: 25568, Cur Loss: 1.52195716, Cur Avg Loss: 1.17638467, Log Avg loss: 1.59595900, Global Avg Loss: 1.52217981, Time: 0.0402 Steps: 82500, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001608, Sample Num: 25728, Cur Loss: 1.77522600, Cur Avg Loss: 1.17680587, Log Avg loss: 1.24411427, Global Avg Loss: 1.52214611, Time: 0.0402 Steps: 82510, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001618, Sample Num: 25888, Cur Loss: 0.73513341, Cur Avg Loss: 1.17784483, Log Avg loss: 1.34490880, Global Avg Loss: 1.52212463, Time: 0.0402 Steps: 82520, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001628, Sample Num: 26048, Cur Loss: 2.31464672, Cur Avg Loss: 1.17574493, Log Avg loss: 0.83598172, Global Avg Loss: 1.52204149, Time: 0.0480 Steps: 82530, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001638, Sample Num: 26208, Cur Loss: 0.98281741, Cur Avg Loss: 1.17708445, Log Avg loss: 1.39515769, Global Avg Loss: 1.52202612, Time: 0.0416 Steps: 82540, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001648, Sample Num: 26368, Cur Loss: 1.75303447, Cur Avg Loss: 1.17680877, Log Avg loss: 1.13165211, Global Avg Loss: 1.52197883, Time: 0.0565 Steps: 82550, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001658, Sample Num: 26528, Cur Loss: 2.36969209, Cur Avg Loss: 1.17783859, Log Avg loss: 1.34755349, Global Avg Loss: 1.52195770, Time: 0.0829 Steps: 82560, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001668, Sample Num: 26688, Cur Loss: 0.47177476, Cur Avg Loss: 1.17703251, Log Avg loss: 1.04338519, Global Avg Loss: 1.52189974, Time: 0.0403 Steps: 82570, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001678, Sample Num: 26848, Cur Loss: 0.55012727, Cur Avg Loss: 1.17640532, Log Avg loss: 1.07178923, Global Avg Loss: 1.52184524, Time: 0.0536 Steps: 82580, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001688, Sample Num: 27008, Cur Loss: 1.28577936, Cur Avg Loss: 1.17517015, Log Avg loss: 0.96790830, Global Avg Loss: 1.52177816, Time: 0.0543 Steps: 82590, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001698, Sample Num: 27168, Cur Loss: 0.64336693, Cur Avg Loss: 1.17466407, Log Avg loss: 1.08923825, Global Avg Loss: 1.52172580, Time: 0.0402 Steps: 82600, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001708, Sample Num: 27328, Cur Loss: 1.41128778, Cur Avg Loss: 1.17462581, Log Avg loss: 1.16812965, Global Avg Loss: 1.52168300, Time: 0.0712 Steps: 82610, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001718, Sample Num: 27488, Cur Loss: 1.97364855, Cur Avg Loss: 1.17485447, Log Avg loss: 1.21390876, Global Avg Loss: 1.52164574, Time: 0.0402 Steps: 82620, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001728, Sample Num: 27648, Cur Loss: 1.12727904, Cur Avg Loss: 1.17440412, Log Avg loss: 1.09703484, Global Avg Loss: 1.52159436, Time: 0.0402 Steps: 82630, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001738, Sample Num: 27808, Cur Loss: 1.93679667, Cur Avg Loss: 1.17551993, Log Avg loss: 1.36833201, Global Avg Loss: 1.52157581, Time: 0.0402 Steps: 82640, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001748, Sample Num: 27968, Cur Loss: 1.47625995, Cur Avg Loss: 1.17900011, Log Avg loss: 1.78385497, Global Avg Loss: 1.52160755, Time: 0.0402 Steps: 82650, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001758, Sample Num: 28128, Cur Loss: 0.77511424, Cur Avg Loss: 1.17903267, Log Avg loss: 1.18472369, Global Avg Loss: 1.52156679, Time: 0.0402 Steps: 82660, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001768, Sample Num: 28288, Cur Loss: 1.84077156, Cur Avg Loss: 1.17995001, Log Avg loss: 1.34121813, Global Avg Loss: 1.52154497, Time: 0.0402 Steps: 82670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001778, Sample Num: 28448, Cur Loss: 0.84653610, Cur Avg Loss: 1.18114757, Log Avg loss: 1.39287609, Global Avg Loss: 1.52152941, Time: 0.0402 Steps: 82680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001788, Sample Num: 28608, Cur Loss: 0.68862766, Cur Avg Loss: 1.18006849, Log Avg loss: 0.98820953, Global Avg Loss: 1.52146492, Time: 0.0403 Steps: 82690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001798, Sample Num: 28768, Cur Loss: 0.58084923, Cur Avg Loss: 1.18055855, Log Avg loss: 1.26818110, Global Avg Loss: 1.52143429, Time: 0.0402 Steps: 82700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001808, Sample Num: 28928, Cur Loss: 1.23764944, Cur Avg Loss: 1.18076736, Log Avg loss: 1.21831122, Global Avg Loss: 1.52139764, Time: 0.0403 Steps: 82710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001818, Sample Num: 29088, Cur Loss: 0.92142820, Cur Avg Loss: 1.18063790, Log Avg loss: 1.15723193, Global Avg Loss: 1.52135362, Time: 0.0404 Steps: 82720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001828, Sample Num: 29248, Cur Loss: 2.28716826, Cur Avg Loss: 1.17968057, Log Avg loss: 1.00563772, Global Avg Loss: 1.52129128, Time: 0.0403 Steps: 82730, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001838, Sample Num: 29408, Cur Loss: 0.92723715, Cur Avg Loss: 1.18099515, Log Avg loss: 1.42129892, Global Avg Loss: 1.52127919, Time: 0.0404 Steps: 82740, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001848, Sample Num: 29568, Cur Loss: 2.16995907, Cur Avg Loss: 1.18126617, Log Avg loss: 1.23108125, Global Avg Loss: 1.52124412, Time: 0.0402 Steps: 82750, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001858, Sample Num: 29728, Cur Loss: 0.35801238, Cur Avg Loss: 1.17946293, Log Avg loss: 0.84622279, Global Avg Loss: 1.52116256, Time: 0.0401 Steps: 82760, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001868, Sample Num: 29888, Cur Loss: 0.47052425, Cur Avg Loss: 1.17796884, Log Avg loss: 0.90036736, Global Avg Loss: 1.52108756, Time: 0.0402 Steps: 82770, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001878, Sample Num: 30048, Cur Loss: 0.95698822, Cur Avg Loss: 1.17798806, Log Avg loss: 1.18157883, Global Avg Loss: 1.52104655, Time: 0.0402 Steps: 82780, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001888, Sample Num: 30208, Cur Loss: 0.41488969, Cur Avg Loss: 1.17712754, Log Avg loss: 1.01552117, Global Avg Loss: 1.52098548, Time: 0.0402 Steps: 82790, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001898, Sample Num: 30368, Cur Loss: 0.76252073, Cur Avg Loss: 1.17751519, Log Avg loss: 1.25070423, Global Avg Loss: 1.52095284, Time: 0.0403 Steps: 82800, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001908, Sample Num: 30528, Cur Loss: 1.20937300, Cur Avg Loss: 1.17796311, Log Avg loss: 1.26297821, Global Avg Loss: 1.52092169, Time: 0.0403 Steps: 82810, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001918, Sample Num: 30688, Cur Loss: 0.87559998, Cur Avg Loss: 1.17717307, Log Avg loss: 1.02643244, Global Avg Loss: 1.52086198, Time: 0.0402 Steps: 82820, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001928, Sample Num: 30848, Cur Loss: 1.44785047, Cur Avg Loss: 1.17640852, Log Avg loss: 1.02976928, Global Avg Loss: 1.52080269, Time: 0.0402 Steps: 82830, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001938, Sample Num: 31008, Cur Loss: 1.36690974, Cur Avg Loss: 1.17705418, Log Avg loss: 1.30153661, Global Avg Loss: 1.52077622, Time: 0.0403 Steps: 82840, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001948, Sample Num: 31168, Cur Loss: 0.38192558, Cur Avg Loss: 1.17686679, Log Avg loss: 1.14055091, Global Avg Loss: 1.52073033, Time: 0.0402 Steps: 82850, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001958, Sample Num: 31328, Cur Loss: 1.60520315, Cur Avg Loss: 1.17772867, Log Avg loss: 1.34562306, Global Avg Loss: 1.52070920, Time: 0.0402 Steps: 82860, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001968, Sample Num: 31488, Cur Loss: 1.44791782, Cur Avg Loss: 1.17771259, Log Avg loss: 1.17456447, Global Avg Loss: 1.52066743, Time: 0.0402 Steps: 82870, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001978, Sample Num: 31648, Cur Loss: 1.74275112, Cur Avg Loss: 1.17964058, Log Avg loss: 1.55906772, Global Avg Loss: 1.52067206, Time: 0.0402 Steps: 82880, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001988, Sample Num: 31808, Cur Loss: 1.08979368, Cur Avg Loss: 1.18006185, Log Avg loss: 1.26338979, Global Avg Loss: 1.52064102, Time: 0.0831 Steps: 82890, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001998, Sample Num: 31968, Cur Loss: 1.44931579, Cur Avg Loss: 1.18005584, Log Avg loss: 1.17886045, Global Avg Loss: 1.52059979, Time: 0.0639 Steps: 82900, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002008, Sample Num: 32128, Cur Loss: 1.32386696, Cur Avg Loss: 1.18008419, Log Avg loss: 1.18574820, Global Avg Loss: 1.52055941, Time: 0.0862 Steps: 82910, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002018, Sample Num: 32288, Cur Loss: 1.43559396, Cur Avg Loss: 1.17926961, Log Avg loss: 1.01570188, Global Avg Loss: 1.52049852, Time: 0.0524 Steps: 82920, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002028, Sample Num: 32448, Cur Loss: 2.20375824, Cur Avg Loss: 1.17897636, Log Avg loss: 1.11979874, Global Avg Loss: 1.52045020, Time: 0.0992 Steps: 82930, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002038, Sample Num: 32608, Cur Loss: 0.55453980, Cur Avg Loss: 1.17991584, Log Avg loss: 1.37044234, Global Avg Loss: 1.52043212, Time: 0.0455 Steps: 82940, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002048, Sample Num: 32768, Cur Loss: 1.04028964, Cur Avg Loss: 1.18069654, Log Avg loss: 1.33980479, Global Avg Loss: 1.52041034, Time: 0.0483 Steps: 82950, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002058, Sample Num: 32928, Cur Loss: 0.97267258, Cur Avg Loss: 1.18021015, Log Avg loss: 1.08059569, Global Avg Loss: 1.52035733, Time: 0.0560 Steps: 82960, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002068, Sample Num: 33088, Cur Loss: 0.82295126, Cur Avg Loss: 1.17919563, Log Avg loss: 0.97040734, Global Avg Loss: 1.52029104, Time: 0.0430 Steps: 82970, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002078, Sample Num: 33248, Cur Loss: 0.97798300, Cur Avg Loss: 1.18128778, Log Avg loss: 1.61394535, Global Avg Loss: 1.52030233, Time: 0.0402 Steps: 82980, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002088, Sample Num: 33408, Cur Loss: 1.37221670, Cur Avg Loss: 1.18052483, Log Avg loss: 1.02198354, Global Avg Loss: 1.52024229, Time: 0.0402 Steps: 82990, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002098, Sample Num: 33568, Cur Loss: 1.39927852, Cur Avg Loss: 1.18222721, Log Avg loss: 1.53768414, Global Avg Loss: 1.52024439, Time: 0.0403 Steps: 83000, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002108, Sample Num: 33728, Cur Loss: 1.07744074, Cur Avg Loss: 1.18234991, Log Avg loss: 1.20809306, Global Avg Loss: 1.52020678, Time: 0.0402 Steps: 83010, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002118, Sample Num: 33888, Cur Loss: 0.99467891, Cur Avg Loss: 1.18216659, Log Avg loss: 1.14352216, Global Avg Loss: 1.52016141, Time: 0.0403 Steps: 83020, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002128, Sample Num: 34048, Cur Loss: 1.00451839, Cur Avg Loss: 1.18163784, Log Avg loss: 1.06964875, Global Avg Loss: 1.52010715, Time: 0.0402 Steps: 83030, Updated lr: 0.000022 ***** Running evaluation checkpoint-83031 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-83031 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.706680, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.088736, "eval_total_loss": 765.381708, "eval_mae": 0.808557, "eval_mse": 1.088943, "eval_r2": 0.307796, "eval_sp_statistic": 0.457048, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.555303, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.041417, "test_total_loss": 522.791378, "test_mae": 0.800148, "test_mse": 1.041657, "test_r2": 0.327705, "test_sp_statistic": 0.485959, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.596587, "test_ps_pvalue": 0.0, "lr": 2.2208629682313894e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5200907144735407, "train_cur_epoch_loss": 2514.6806759536266, "train_cur_epoch_avg_loss": 1.1811557895507876, "train_cur_epoch_time": 95.70668029785156, "train_cur_epoch_avg_time": 0.044953818834124734, "epoch": 39, "step": 83031} ################################################## Training, Epoch: 0040, Batch: 000009, Sample Num: 144, Cur Loss: 1.18672335, Cur Avg Loss: 1.33463704, Log Avg loss: 1.21670864, Global Avg Loss: 1.52007061, Time: 0.0406 Steps: 83040, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000019, Sample Num: 304, Cur Loss: 1.01576042, Cur Avg Loss: 1.21852851, Log Avg loss: 1.11403084, Global Avg Loss: 1.52002172, Time: 0.0402 Steps: 83050, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000029, Sample Num: 464, Cur Loss: 0.48617202, Cur Avg Loss: 1.20706853, Log Avg loss: 1.18529456, Global Avg Loss: 1.51998142, Time: 0.0403 Steps: 83060, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000039, Sample Num: 624, Cur Loss: 0.43902284, Cur Avg Loss: 1.07598556, Log Avg loss: 0.69584494, Global Avg Loss: 1.51988221, Time: 0.0402 Steps: 83070, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000049, Sample Num: 784, Cur Loss: 0.40291455, Cur Avg Loss: 1.03652920, Log Avg loss: 0.88264941, Global Avg Loss: 1.51980551, Time: 0.0402 Steps: 83080, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000059, Sample Num: 944, Cur Loss: 1.43800211, Cur Avg Loss: 1.10978836, Log Avg loss: 1.46875824, Global Avg Loss: 1.51979937, Time: 0.0402 Steps: 83090, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000069, Sample Num: 1104, Cur Loss: 1.21942949, Cur Avg Loss: 1.15104353, Log Avg loss: 1.39444906, Global Avg Loss: 1.51978429, Time: 0.0402 Steps: 83100, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000079, Sample Num: 1264, Cur Loss: 0.48916659, Cur Avg Loss: 1.14824607, Log Avg loss: 1.12894361, Global Avg Loss: 1.51973726, Time: 0.0402 Steps: 83110, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000089, Sample Num: 1424, Cur Loss: 2.78418827, Cur Avg Loss: 1.18120310, Log Avg loss: 1.44156361, Global Avg Loss: 1.51972785, Time: 0.0403 Steps: 83120, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000099, Sample Num: 1584, Cur Loss: 1.51340604, Cur Avg Loss: 1.21019245, Log Avg loss: 1.46819761, Global Avg Loss: 1.51972165, Time: 0.0402 Steps: 83130, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000109, Sample Num: 1744, Cur Loss: 0.65448374, Cur Avg Loss: 1.20316902, Log Avg loss: 1.13363708, Global Avg Loss: 1.51967522, Time: 0.0402 Steps: 83140, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000119, Sample Num: 1904, Cur Loss: 1.36139560, Cur Avg Loss: 1.20669658, Log Avg loss: 1.24514707, Global Avg Loss: 1.51964220, Time: 0.0403 Steps: 83150, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000129, Sample Num: 2064, Cur Loss: 1.65988517, Cur Avg Loss: 1.21378638, Log Avg loss: 1.29815490, Global Avg Loss: 1.51961557, Time: 0.0403 Steps: 83160, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000139, Sample Num: 2224, Cur Loss: 0.90799141, Cur Avg Loss: 1.21194433, Log Avg loss: 1.18818196, Global Avg Loss: 1.51957572, Time: 0.0402 Steps: 83170, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000149, Sample Num: 2384, Cur Loss: 1.93150187, Cur Avg Loss: 1.20211123, Log Avg loss: 1.06543103, Global Avg Loss: 1.51952112, Time: 0.0403 Steps: 83180, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000159, Sample Num: 2544, Cur Loss: 0.99838030, Cur Avg Loss: 1.18905857, Log Avg loss: 0.99457407, Global Avg Loss: 1.51945802, Time: 0.0401 Steps: 83190, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000169, Sample Num: 2704, Cur Loss: 0.99847353, Cur Avg Loss: 1.18744838, Log Avg loss: 1.16184626, Global Avg Loss: 1.51941503, Time: 0.0403 Steps: 83200, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000179, Sample Num: 2864, Cur Loss: 1.04539680, Cur Avg Loss: 1.18461215, Log Avg loss: 1.13667993, Global Avg Loss: 1.51936904, Time: 0.0401 Steps: 83210, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000189, Sample Num: 3024, Cur Loss: 1.19582582, Cur Avg Loss: 1.17945138, Log Avg loss: 1.08707352, Global Avg Loss: 1.51931709, Time: 0.0402 Steps: 83220, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000199, Sample Num: 3184, Cur Loss: 1.12778449, Cur Avg Loss: 1.19664083, Log Avg loss: 1.52152147, Global Avg Loss: 1.51931736, Time: 0.0403 Steps: 83230, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000209, Sample Num: 3344, Cur Loss: 0.55118471, Cur Avg Loss: 1.20102864, Log Avg loss: 1.28834599, Global Avg Loss: 1.51928961, Time: 0.0402 Steps: 83240, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000219, Sample Num: 3504, Cur Loss: 0.35350168, Cur Avg Loss: 1.20214950, Log Avg loss: 1.22557566, Global Avg Loss: 1.51925433, Time: 0.0402 Steps: 83250, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000229, Sample Num: 3664, Cur Loss: 0.59297585, Cur Avg Loss: 1.19218073, Log Avg loss: 0.97386468, Global Avg Loss: 1.51918882, Time: 0.0402 Steps: 83260, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000239, Sample Num: 3824, Cur Loss: 0.38188118, Cur Avg Loss: 1.18157214, Log Avg loss: 0.93863536, Global Avg Loss: 1.51911910, Time: 0.0402 Steps: 83270, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000249, Sample Num: 3984, Cur Loss: 0.95454574, Cur Avg Loss: 1.18215908, Log Avg loss: 1.19618699, Global Avg Loss: 1.51908033, Time: 0.0402 Steps: 83280, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000259, Sample Num: 4144, Cur Loss: 0.38788086, Cur Avg Loss: 1.17482768, Log Avg loss: 0.99227580, Global Avg Loss: 1.51901708, Time: 0.0402 Steps: 83290, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000269, Sample Num: 4304, Cur Loss: 2.03225446, Cur Avg Loss: 1.17654694, Log Avg loss: 1.22107580, Global Avg Loss: 1.51898131, Time: 0.0406 Steps: 83300, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000279, Sample Num: 4464, Cur Loss: 0.84648919, Cur Avg Loss: 1.17271840, Log Avg loss: 1.06973069, Global Avg Loss: 1.51892739, Time: 0.0488 Steps: 83310, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000289, Sample Num: 4624, Cur Loss: 0.62542188, Cur Avg Loss: 1.16766442, Log Avg loss: 1.02665826, Global Avg Loss: 1.51886830, Time: 0.0675 Steps: 83320, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000299, Sample Num: 4784, Cur Loss: 2.35726929, Cur Avg Loss: 1.17980300, Log Avg loss: 1.53060787, Global Avg Loss: 1.51886971, Time: 0.0404 Steps: 83330, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000309, Sample Num: 4944, Cur Loss: 1.69136083, Cur Avg Loss: 1.18407613, Log Avg loss: 1.31184284, Global Avg Loss: 1.51884487, Time: 0.0502 Steps: 83340, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000319, Sample Num: 5104, Cur Loss: 0.32406193, Cur Avg Loss: 1.17912687, Log Avg loss: 1.02619479, Global Avg Loss: 1.51878577, Time: 0.0405 Steps: 83350, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000329, Sample Num: 5264, Cur Loss: 0.63592434, Cur Avg Loss: 1.17734484, Log Avg loss: 1.12049795, Global Avg Loss: 1.51873799, Time: 0.0456 Steps: 83360, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000339, Sample Num: 5424, Cur Loss: 1.54793382, Cur Avg Loss: 1.17587559, Log Avg loss: 1.12753719, Global Avg Loss: 1.51869106, Time: 0.0405 Steps: 83370, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000349, Sample Num: 5584, Cur Loss: 0.49521744, Cur Avg Loss: 1.18160730, Log Avg loss: 1.37591233, Global Avg Loss: 1.51867394, Time: 0.0635 Steps: 83380, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000359, Sample Num: 5744, Cur Loss: 1.17128193, Cur Avg Loss: 1.17151895, Log Avg loss: 0.81943560, Global Avg Loss: 1.51859009, Time: 0.0404 Steps: 83390, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000369, Sample Num: 5904, Cur Loss: 1.37890375, Cur Avg Loss: 1.17993006, Log Avg loss: 1.48188875, Global Avg Loss: 1.51858569, Time: 0.0403 Steps: 83400, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000379, Sample Num: 6064, Cur Loss: 1.96336114, Cur Avg Loss: 1.17681343, Log Avg loss: 1.06181003, Global Avg Loss: 1.51853092, Time: 0.0402 Steps: 83410, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000389, Sample Num: 6224, Cur Loss: 0.94212615, Cur Avg Loss: 1.17379394, Log Avg loss: 1.05935528, Global Avg Loss: 1.51847588, Time: 0.0402 Steps: 83420, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000399, Sample Num: 6384, Cur Loss: 1.34277010, Cur Avg Loss: 1.17273284, Log Avg loss: 1.13145605, Global Avg Loss: 1.51842949, Time: 0.0402 Steps: 83430, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000409, Sample Num: 6544, Cur Loss: 0.72194827, Cur Avg Loss: 1.16790886, Log Avg loss: 0.97543189, Global Avg Loss: 1.51836442, Time: 0.0403 Steps: 83440, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000419, Sample Num: 6704, Cur Loss: 1.54816878, Cur Avg Loss: 1.16528076, Log Avg loss: 1.05779168, Global Avg Loss: 1.51830922, Time: 0.0402 Steps: 83450, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000429, Sample Num: 6864, Cur Loss: 3.16386318, Cur Avg Loss: 1.16990174, Log Avg loss: 1.36352055, Global Avg Loss: 1.51829068, Time: 0.0402 Steps: 83460, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000439, Sample Num: 7024, Cur Loss: 1.04573059, Cur Avg Loss: 1.17165242, Log Avg loss: 1.24675653, Global Avg Loss: 1.51825815, Time: 0.0402 Steps: 83470, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000449, Sample Num: 7184, Cur Loss: 2.55305362, Cur Avg Loss: 1.17416110, Log Avg loss: 1.28429247, Global Avg Loss: 1.51823012, Time: 0.0402 Steps: 83480, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000459, Sample Num: 7344, Cur Loss: 1.05781507, Cur Avg Loss: 1.17453549, Log Avg loss: 1.19134539, Global Avg Loss: 1.51819097, Time: 0.0403 Steps: 83490, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000469, Sample Num: 7504, Cur Loss: 0.49935746, Cur Avg Loss: 1.17650957, Log Avg loss: 1.26711970, Global Avg Loss: 1.51816090, Time: 0.0402 Steps: 83500, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000479, Sample Num: 7664, Cur Loss: 2.01981282, Cur Avg Loss: 1.17931821, Log Avg loss: 1.31104350, Global Avg Loss: 1.51813610, Time: 0.0402 Steps: 83510, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000489, Sample Num: 7824, Cur Loss: 1.70116138, Cur Avg Loss: 1.18145119, Log Avg loss: 1.28362086, Global Avg Loss: 1.51810802, Time: 0.0403 Steps: 83520, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000499, Sample Num: 7984, Cur Loss: 0.75685769, Cur Avg Loss: 1.17740080, Log Avg loss: 0.97933714, Global Avg Loss: 1.51804352, Time: 0.0403 Steps: 83530, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000509, Sample Num: 8144, Cur Loss: 0.66979623, Cur Avg Loss: 1.17540719, Log Avg loss: 1.07592560, Global Avg Loss: 1.51799060, Time: 0.0403 Steps: 83540, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000519, Sample Num: 8304, Cur Loss: 0.52750599, Cur Avg Loss: 1.17377582, Log Avg loss: 1.09073916, Global Avg Loss: 1.51793946, Time: 0.0403 Steps: 83550, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000529, Sample Num: 8464, Cur Loss: 0.69262707, Cur Avg Loss: 1.17483264, Log Avg loss: 1.22968152, Global Avg Loss: 1.51790496, Time: 0.0403 Steps: 83560, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000539, Sample Num: 8624, Cur Loss: 0.97100341, Cur Avg Loss: 1.17706163, Log Avg loss: 1.29497561, Global Avg Loss: 1.51787829, Time: 0.0402 Steps: 83570, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000549, Sample Num: 8784, Cur Loss: 0.65493286, Cur Avg Loss: 1.17919118, Log Avg loss: 1.29397395, Global Avg Loss: 1.51785150, Time: 0.0402 Steps: 83580, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000559, Sample Num: 8944, Cur Loss: 0.43756092, Cur Avg Loss: 1.17603126, Log Avg loss: 1.00255116, Global Avg Loss: 1.51778985, Time: 0.0402 Steps: 83590, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000569, Sample Num: 9104, Cur Loss: 0.90434819, Cur Avg Loss: 1.17652240, Log Avg loss: 1.20397760, Global Avg Loss: 1.51775231, Time: 0.0402 Steps: 83600, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000579, Sample Num: 9264, Cur Loss: 1.45437551, Cur Avg Loss: 1.17188119, Log Avg loss: 0.90779592, Global Avg Loss: 1.51767936, Time: 0.0403 Steps: 83610, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000589, Sample Num: 9424, Cur Loss: 1.88948727, Cur Avg Loss: 1.16894991, Log Avg loss: 0.99922916, Global Avg Loss: 1.51761736, Time: 0.0402 Steps: 83620, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000599, Sample Num: 9584, Cur Loss: 2.41474986, Cur Avg Loss: 1.17327272, Log Avg loss: 1.42788614, Global Avg Loss: 1.51760663, Time: 0.0402 Steps: 83630, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000609, Sample Num: 9744, Cur Loss: 1.18694878, Cur Avg Loss: 1.17298500, Log Avg loss: 1.15575058, Global Avg Loss: 1.51756337, Time: 0.0402 Steps: 83640, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000619, Sample Num: 9904, Cur Loss: 1.48537135, Cur Avg Loss: 1.16797372, Log Avg loss: 0.86278699, Global Avg Loss: 1.51748509, Time: 0.0635 Steps: 83650, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000629, Sample Num: 10064, Cur Loss: 0.65961236, Cur Avg Loss: 1.16595938, Log Avg loss: 1.04127127, Global Avg Loss: 1.51742817, Time: 0.0614 Steps: 83660, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000639, Sample Num: 10224, Cur Loss: 0.66222829, Cur Avg Loss: 1.16560947, Log Avg loss: 1.14360009, Global Avg Loss: 1.51738349, Time: 0.0626 Steps: 83670, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000649, Sample Num: 10384, Cur Loss: 1.16060686, Cur Avg Loss: 1.16408093, Log Avg loss: 1.06640734, Global Avg Loss: 1.51732960, Time: 0.0405 Steps: 83680, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000659, Sample Num: 10544, Cur Loss: 0.58984923, Cur Avg Loss: 1.16375753, Log Avg loss: 1.14276870, Global Avg Loss: 1.51728484, Time: 0.0403 Steps: 83690, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000669, Sample Num: 10704, Cur Loss: 1.46176720, Cur Avg Loss: 1.16352449, Log Avg loss: 1.14816735, Global Avg Loss: 1.51724074, Time: 0.0457 Steps: 83700, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000679, Sample Num: 10864, Cur Loss: 1.21720481, Cur Avg Loss: 1.16262274, Log Avg loss: 1.10229561, Global Avg Loss: 1.51719117, Time: 0.0449 Steps: 83710, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000689, Sample Num: 11024, Cur Loss: 0.46472740, Cur Avg Loss: 1.16308226, Log Avg loss: 1.19428396, Global Avg Loss: 1.51715260, Time: 0.0429 Steps: 83720, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000699, Sample Num: 11184, Cur Loss: 1.28068590, Cur Avg Loss: 1.16266862, Log Avg loss: 1.13416839, Global Avg Loss: 1.51710686, Time: 0.0757 Steps: 83730, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000709, Sample Num: 11344, Cur Loss: 0.31058139, Cur Avg Loss: 1.15961868, Log Avg loss: 0.94642814, Global Avg Loss: 1.51703871, Time: 0.0768 Steps: 83740, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000719, Sample Num: 11504, Cur Loss: 1.22062814, Cur Avg Loss: 1.16080223, Log Avg loss: 1.24471626, Global Avg Loss: 1.51700620, Time: 0.0403 Steps: 83750, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000729, Sample Num: 11664, Cur Loss: 0.72878730, Cur Avg Loss: 1.15811242, Log Avg loss: 0.96471486, Global Avg Loss: 1.51694026, Time: 0.0402 Steps: 83760, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000739, Sample Num: 11824, Cur Loss: 0.70567107, Cur Avg Loss: 1.15661818, Log Avg loss: 1.04768791, Global Avg Loss: 1.51688424, Time: 0.0402 Steps: 83770, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000749, Sample Num: 11984, Cur Loss: 0.30218297, Cur Avg Loss: 1.15574535, Log Avg loss: 1.09124363, Global Avg Loss: 1.51683344, Time: 0.0402 Steps: 83780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000759, Sample Num: 12144, Cur Loss: 1.09140325, Cur Avg Loss: 1.15481333, Log Avg loss: 1.08500441, Global Avg Loss: 1.51678190, Time: 0.0403 Steps: 83790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000769, Sample Num: 12304, Cur Loss: 0.91925436, Cur Avg Loss: 1.15630226, Log Avg loss: 1.26931254, Global Avg Loss: 1.51675237, Time: 0.0402 Steps: 83800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000779, Sample Num: 12464, Cur Loss: 1.01478958, Cur Avg Loss: 1.15471296, Log Avg loss: 1.03249578, Global Avg Loss: 1.51669459, Time: 0.0402 Steps: 83810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000789, Sample Num: 12624, Cur Loss: 0.96494037, Cur Avg Loss: 1.15557329, Log Avg loss: 1.22259256, Global Avg Loss: 1.51665950, Time: 0.0404 Steps: 83820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000799, Sample Num: 12784, Cur Loss: 0.54219776, Cur Avg Loss: 1.15585921, Log Avg loss: 1.17841836, Global Avg Loss: 1.51661915, Time: 0.0403 Steps: 83830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000809, Sample Num: 12944, Cur Loss: 0.72697997, Cur Avg Loss: 1.15696944, Log Avg loss: 1.24567686, Global Avg Loss: 1.51658684, Time: 0.0403 Steps: 83840, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000819, Sample Num: 13104, Cur Loss: 0.78094268, Cur Avg Loss: 1.15753218, Log Avg loss: 1.20305827, Global Avg Loss: 1.51654945, Time: 0.0403 Steps: 83850, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000829, Sample Num: 13264, Cur Loss: 1.00997889, Cur Avg Loss: 1.15695611, Log Avg loss: 1.10977587, Global Avg Loss: 1.51650094, Time: 0.0403 Steps: 83860, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000839, Sample Num: 13424, Cur Loss: 3.26070118, Cur Avg Loss: 1.16288065, Log Avg loss: 1.65402498, Global Avg Loss: 1.51651734, Time: 0.0402 Steps: 83870, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000849, Sample Num: 13584, Cur Loss: 0.36533606, Cur Avg Loss: 1.16240734, Log Avg loss: 1.12269627, Global Avg Loss: 1.51647039, Time: 0.0403 Steps: 83880, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000859, Sample Num: 13744, Cur Loss: 0.86712670, Cur Avg Loss: 1.15995359, Log Avg loss: 0.95163030, Global Avg Loss: 1.51640305, Time: 0.0402 Steps: 83890, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000869, Sample Num: 13904, Cur Loss: 1.75060058, Cur Avg Loss: 1.16185872, Log Avg loss: 1.32550932, Global Avg Loss: 1.51638030, Time: 0.0403 Steps: 83900, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000879, Sample Num: 14064, Cur Loss: 0.85546237, Cur Avg Loss: 1.16565696, Log Avg loss: 1.49572409, Global Avg Loss: 1.51637784, Time: 0.0402 Steps: 83910, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000889, Sample Num: 14224, Cur Loss: 1.36470008, Cur Avg Loss: 1.16589850, Log Avg loss: 1.18712958, Global Avg Loss: 1.51633861, Time: 0.0402 Steps: 83920, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000899, Sample Num: 14384, Cur Loss: 3.15057898, Cur Avg Loss: 1.16682737, Log Avg loss: 1.24940435, Global Avg Loss: 1.51630680, Time: 0.0403 Steps: 83930, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000909, Sample Num: 14544, Cur Loss: 0.31565639, Cur Avg Loss: 1.16558331, Log Avg loss: 1.05374196, Global Avg Loss: 1.51625170, Time: 0.0402 Steps: 83940, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000919, Sample Num: 14704, Cur Loss: 2.14714861, Cur Avg Loss: 1.16703735, Log Avg loss: 1.29921018, Global Avg Loss: 1.51622584, Time: 0.0402 Steps: 83950, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000929, Sample Num: 14864, Cur Loss: 1.83554900, Cur Avg Loss: 1.16769663, Log Avg loss: 1.22828383, Global Avg Loss: 1.51619155, Time: 0.0403 Steps: 83960, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000939, Sample Num: 15024, Cur Loss: 1.77308941, Cur Avg Loss: 1.16670111, Log Avg loss: 1.07421803, Global Avg Loss: 1.51613891, Time: 0.0402 Steps: 83970, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000949, Sample Num: 15184, Cur Loss: 1.14730334, Cur Avg Loss: 1.16692027, Log Avg loss: 1.18749867, Global Avg Loss: 1.51609978, Time: 0.0402 Steps: 83980, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000959, Sample Num: 15344, Cur Loss: 0.35339391, Cur Avg Loss: 1.16823547, Log Avg loss: 1.29304832, Global Avg Loss: 1.51607322, Time: 0.0402 Steps: 83990, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000969, Sample Num: 15504, Cur Loss: 1.45383358, Cur Avg Loss: 1.16832503, Log Avg loss: 1.17691387, Global Avg Loss: 1.51603285, Time: 0.0403 Steps: 84000, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000979, Sample Num: 15664, Cur Loss: 1.20972359, Cur Avg Loss: 1.16793273, Log Avg loss: 1.12991905, Global Avg Loss: 1.51598689, Time: 0.0534 Steps: 84010, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000989, Sample Num: 15824, Cur Loss: 1.34142804, Cur Avg Loss: 1.16449069, Log Avg loss: 0.82751443, Global Avg Loss: 1.51590494, Time: 0.0454 Steps: 84020, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000999, Sample Num: 15984, Cur Loss: 0.41528243, Cur Avg Loss: 1.16461841, Log Avg loss: 1.17725045, Global Avg Loss: 1.51586464, Time: 0.0457 Steps: 84030, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001009, Sample Num: 16144, Cur Loss: 0.38969615, Cur Avg Loss: 1.16111103, Log Avg loss: 0.81072346, Global Avg Loss: 1.51578074, Time: 0.0613 Steps: 84040, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001019, Sample Num: 16304, Cur Loss: 0.70365751, Cur Avg Loss: 1.16002898, Log Avg loss: 1.05085021, Global Avg Loss: 1.51572542, Time: 0.0486 Steps: 84050, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001029, Sample Num: 16464, Cur Loss: 1.43773305, Cur Avg Loss: 1.16428797, Log Avg loss: 1.59827939, Global Avg Loss: 1.51573524, Time: 0.0655 Steps: 84060, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001039, Sample Num: 16624, Cur Loss: 1.14609110, Cur Avg Loss: 1.16182969, Log Avg loss: 0.90887217, Global Avg Loss: 1.51566306, Time: 0.0404 Steps: 84070, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001049, Sample Num: 16784, Cur Loss: 0.53430384, Cur Avg Loss: 1.15926646, Log Avg loss: 0.89294727, Global Avg Loss: 1.51558899, Time: 0.0872 Steps: 84080, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001059, Sample Num: 16944, Cur Loss: 0.90199548, Cur Avg Loss: 1.16048477, Log Avg loss: 1.28828506, Global Avg Loss: 1.51556196, Time: 0.0655 Steps: 84090, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001069, Sample Num: 17104, Cur Loss: 1.08207083, Cur Avg Loss: 1.16024289, Log Avg loss: 1.13462810, Global Avg Loss: 1.51551667, Time: 0.0403 Steps: 84100, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001079, Sample Num: 17264, Cur Loss: 1.02246785, Cur Avg Loss: 1.16009520, Log Avg loss: 1.14430732, Global Avg Loss: 1.51547253, Time: 0.0402 Steps: 84110, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001089, Sample Num: 17424, Cur Loss: 0.53382707, Cur Avg Loss: 1.16064372, Log Avg loss: 1.21982885, Global Avg Loss: 1.51543739, Time: 0.0402 Steps: 84120, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001099, Sample Num: 17584, Cur Loss: 0.66124189, Cur Avg Loss: 1.16084005, Log Avg loss: 1.18222019, Global Avg Loss: 1.51539778, Time: 0.0402 Steps: 84130, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001109, Sample Num: 17744, Cur Loss: 1.11013329, Cur Avg Loss: 1.15884554, Log Avg loss: 0.93964886, Global Avg Loss: 1.51532935, Time: 0.0403 Steps: 84140, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001119, Sample Num: 17904, Cur Loss: 1.07368851, Cur Avg Loss: 1.16057629, Log Avg loss: 1.35251634, Global Avg Loss: 1.51531001, Time: 0.0402 Steps: 84150, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001129, Sample Num: 18064, Cur Loss: 1.58785868, Cur Avg Loss: 1.16458372, Log Avg loss: 1.61301521, Global Avg Loss: 1.51532162, Time: 0.0403 Steps: 84160, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001139, Sample Num: 18224, Cur Loss: 0.94614017, Cur Avg Loss: 1.16453080, Log Avg loss: 1.15855580, Global Avg Loss: 1.51527923, Time: 0.0403 Steps: 84170, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001149, Sample Num: 18384, Cur Loss: 0.51130676, Cur Avg Loss: 1.16536730, Log Avg loss: 1.26064568, Global Avg Loss: 1.51524898, Time: 0.0403 Steps: 84180, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001159, Sample Num: 18544, Cur Loss: 1.65654993, Cur Avg Loss: 1.16458793, Log Avg loss: 1.07503821, Global Avg Loss: 1.51519669, Time: 0.0402 Steps: 84190, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001169, Sample Num: 18704, Cur Loss: 1.67401648, Cur Avg Loss: 1.16490003, Log Avg loss: 1.20107144, Global Avg Loss: 1.51515939, Time: 0.0403 Steps: 84200, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001179, Sample Num: 18864, Cur Loss: 1.37606144, Cur Avg Loss: 1.16417637, Log Avg loss: 1.07958071, Global Avg Loss: 1.51510766, Time: 0.0402 Steps: 84210, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001189, Sample Num: 19024, Cur Loss: 0.66771990, Cur Avg Loss: 1.16543100, Log Avg loss: 1.31335254, Global Avg Loss: 1.51508370, Time: 0.0403 Steps: 84220, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001199, Sample Num: 19184, Cur Loss: 1.45369339, Cur Avg Loss: 1.16710188, Log Avg loss: 1.36576912, Global Avg Loss: 1.51506598, Time: 0.0403 Steps: 84230, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001209, Sample Num: 19344, Cur Loss: 0.48945719, Cur Avg Loss: 1.16381008, Log Avg loss: 0.76912290, Global Avg Loss: 1.51497743, Time: 0.0403 Steps: 84240, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001219, Sample Num: 19504, Cur Loss: 1.23753452, Cur Avg Loss: 1.16329837, Log Avg loss: 1.10143309, Global Avg Loss: 1.51492834, Time: 0.0402 Steps: 84250, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001229, Sample Num: 19664, Cur Loss: 1.25471485, Cur Avg Loss: 1.16517182, Log Avg loss: 1.39354518, Global Avg Loss: 1.51491394, Time: 0.0402 Steps: 84260, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001239, Sample Num: 19824, Cur Loss: 0.90974092, Cur Avg Loss: 1.16594151, Log Avg loss: 1.26053706, Global Avg Loss: 1.51488375, Time: 0.0402 Steps: 84270, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001249, Sample Num: 19984, Cur Loss: 0.32971814, Cur Avg Loss: 1.16692493, Log Avg loss: 1.28877052, Global Avg Loss: 1.51485692, Time: 0.0402 Steps: 84280, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001259, Sample Num: 20144, Cur Loss: 1.20976198, Cur Avg Loss: 1.16740403, Log Avg loss: 1.22724391, Global Avg Loss: 1.51482280, Time: 0.0402 Steps: 84290, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001269, Sample Num: 20304, Cur Loss: 0.74661434, Cur Avg Loss: 1.16667444, Log Avg loss: 1.07481821, Global Avg Loss: 1.51477060, Time: 0.0402 Steps: 84300, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001279, Sample Num: 20464, Cur Loss: 2.08828712, Cur Avg Loss: 1.16857020, Log Avg loss: 1.40914260, Global Avg Loss: 1.51475808, Time: 0.0402 Steps: 84310, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001289, Sample Num: 20624, Cur Loss: 0.42006791, Cur Avg Loss: 1.16805290, Log Avg loss: 1.10189011, Global Avg Loss: 1.51470911, Time: 0.0402 Steps: 84320, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001299, Sample Num: 20784, Cur Loss: 0.62344050, Cur Avg Loss: 1.16907558, Log Avg loss: 1.30089917, Global Avg Loss: 1.51468376, Time: 0.0402 Steps: 84330, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001309, Sample Num: 20944, Cur Loss: 0.69615346, Cur Avg Loss: 1.16783032, Log Avg loss: 1.00607047, Global Avg Loss: 1.51462345, Time: 0.0637 Steps: 84340, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001319, Sample Num: 21104, Cur Loss: 0.51779759, Cur Avg Loss: 1.16697708, Log Avg loss: 1.05528826, Global Avg Loss: 1.51456900, Time: 0.0584 Steps: 84350, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001329, Sample Num: 21264, Cur Loss: 1.20923400, Cur Avg Loss: 1.16657447, Log Avg loss: 1.11347057, Global Avg Loss: 1.51452145, Time: 0.0431 Steps: 84360, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001339, Sample Num: 21424, Cur Loss: 0.80474305, Cur Avg Loss: 1.16553290, Log Avg loss: 1.02710782, Global Avg Loss: 1.51446368, Time: 0.0404 Steps: 84370, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001349, Sample Num: 21584, Cur Loss: 0.83704412, Cur Avg Loss: 1.16499779, Log Avg loss: 1.09334666, Global Avg Loss: 1.51441377, Time: 0.0405 Steps: 84380, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001359, Sample Num: 21744, Cur Loss: 0.61820120, Cur Avg Loss: 1.16588362, Log Avg loss: 1.28538208, Global Avg Loss: 1.51438663, Time: 0.0585 Steps: 84390, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001369, Sample Num: 21904, Cur Loss: 1.78279996, Cur Avg Loss: 1.16646942, Log Avg loss: 1.24608033, Global Avg Loss: 1.51435484, Time: 0.0404 Steps: 84400, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001379, Sample Num: 22064, Cur Loss: 0.96893406, Cur Avg Loss: 1.16583042, Log Avg loss: 1.07835011, Global Avg Loss: 1.51430319, Time: 0.0484 Steps: 84410, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001389, Sample Num: 22224, Cur Loss: 2.25384641, Cur Avg Loss: 1.16752270, Log Avg loss: 1.40088920, Global Avg Loss: 1.51428976, Time: 0.0863 Steps: 84420, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001399, Sample Num: 22384, Cur Loss: 1.61940837, Cur Avg Loss: 1.16910498, Log Avg loss: 1.38888254, Global Avg Loss: 1.51427490, Time: 0.0403 Steps: 84430, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001409, Sample Num: 22544, Cur Loss: 0.76294088, Cur Avg Loss: 1.16979421, Log Avg loss: 1.26621745, Global Avg Loss: 1.51424553, Time: 0.0402 Steps: 84440, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001419, Sample Num: 22704, Cur Loss: 0.90447795, Cur Avg Loss: 1.16996372, Log Avg loss: 1.19384875, Global Avg Loss: 1.51420759, Time: 0.0403 Steps: 84450, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001429, Sample Num: 22864, Cur Loss: 1.30253839, Cur Avg Loss: 1.17249972, Log Avg loss: 1.53235735, Global Avg Loss: 1.51420973, Time: 0.0402 Steps: 84460, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001439, Sample Num: 23024, Cur Loss: 1.38914871, Cur Avg Loss: 1.17274409, Log Avg loss: 1.20766466, Global Avg Loss: 1.51417344, Time: 0.0402 Steps: 84470, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001449, Sample Num: 23184, Cur Loss: 1.41322184, Cur Avg Loss: 1.17158854, Log Avg loss: 1.00530509, Global Avg Loss: 1.51411321, Time: 0.0403 Steps: 84480, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001459, Sample Num: 23344, Cur Loss: 0.91908348, Cur Avg Loss: 1.17023396, Log Avg loss: 0.97395505, Global Avg Loss: 1.51404928, Time: 0.0403 Steps: 84490, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001469, Sample Num: 23504, Cur Loss: 2.85916734, Cur Avg Loss: 1.17104854, Log Avg loss: 1.28989619, Global Avg Loss: 1.51402275, Time: 0.0402 Steps: 84500, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001479, Sample Num: 23664, Cur Loss: 1.88789654, Cur Avg Loss: 1.17134482, Log Avg loss: 1.21486886, Global Avg Loss: 1.51398735, Time: 0.0402 Steps: 84510, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001489, Sample Num: 23824, Cur Loss: 0.37144125, Cur Avg Loss: 1.17090373, Log Avg loss: 1.10566533, Global Avg Loss: 1.51393904, Time: 0.0403 Steps: 84520, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001499, Sample Num: 23984, Cur Loss: 1.64168382, Cur Avg Loss: 1.17273038, Log Avg loss: 1.44471969, Global Avg Loss: 1.51393085, Time: 0.0402 Steps: 84530, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001509, Sample Num: 24144, Cur Loss: 0.68343091, Cur Avg Loss: 1.17224299, Log Avg loss: 1.09918222, Global Avg Loss: 1.51388179, Time: 0.0403 Steps: 84540, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001519, Sample Num: 24304, Cur Loss: 1.17181933, Cur Avg Loss: 1.17142610, Log Avg loss: 1.04815754, Global Avg Loss: 1.51382671, Time: 0.0403 Steps: 84550, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001529, Sample Num: 24464, Cur Loss: 1.99731171, Cur Avg Loss: 1.17330053, Log Avg loss: 1.45802739, Global Avg Loss: 1.51382011, Time: 0.0402 Steps: 84560, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001539, Sample Num: 24624, Cur Loss: 0.62582153, Cur Avg Loss: 1.17237963, Log Avg loss: 1.03157291, Global Avg Loss: 1.51376309, Time: 0.0402 Steps: 84570, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001549, Sample Num: 24784, Cur Loss: 0.40845570, Cur Avg Loss: 1.17189560, Log Avg loss: 1.09740414, Global Avg Loss: 1.51371386, Time: 0.0402 Steps: 84580, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001559, Sample Num: 24944, Cur Loss: 0.43966547, Cur Avg Loss: 1.17007888, Log Avg loss: 0.88866818, Global Avg Loss: 1.51363997, Time: 0.0402 Steps: 84590, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001569, Sample Num: 25104, Cur Loss: 0.39904156, Cur Avg Loss: 1.17031570, Log Avg loss: 1.20723639, Global Avg Loss: 1.51360375, Time: 0.0402 Steps: 84600, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001579, Sample Num: 25264, Cur Loss: 0.96298337, Cur Avg Loss: 1.17113423, Log Avg loss: 1.29956173, Global Avg Loss: 1.51357845, Time: 0.0402 Steps: 84610, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001589, Sample Num: 25424, Cur Loss: 1.09588766, Cur Avg Loss: 1.16981317, Log Avg loss: 0.96121731, Global Avg Loss: 1.51351318, Time: 0.0403 Steps: 84620, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001599, Sample Num: 25584, Cur Loss: 0.42327470, Cur Avg Loss: 1.16785618, Log Avg loss: 0.85689132, Global Avg Loss: 1.51343559, Time: 0.0402 Steps: 84630, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001609, Sample Num: 25744, Cur Loss: 1.76751137, Cur Avg Loss: 1.16898315, Log Avg loss: 1.34918496, Global Avg Loss: 1.51341619, Time: 0.0402 Steps: 84640, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001619, Sample Num: 25904, Cur Loss: 2.40892816, Cur Avg Loss: 1.17247745, Log Avg loss: 1.73471053, Global Avg Loss: 1.51344233, Time: 0.0403 Steps: 84650, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001629, Sample Num: 26064, Cur Loss: 3.13255310, Cur Avg Loss: 1.17310789, Log Avg loss: 1.27517556, Global Avg Loss: 1.51341418, Time: 0.0403 Steps: 84660, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001639, Sample Num: 26224, Cur Loss: 1.64628267, Cur Avg Loss: 1.17212968, Log Avg loss: 1.01277913, Global Avg Loss: 1.51335506, Time: 0.0402 Steps: 84670, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001649, Sample Num: 26384, Cur Loss: 0.43546313, Cur Avg Loss: 1.17118905, Log Avg loss: 1.01702013, Global Avg Loss: 1.51329644, Time: 0.0402 Steps: 84680, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001659, Sample Num: 26544, Cur Loss: 0.65750158, Cur Avg Loss: 1.17124367, Log Avg loss: 1.18025010, Global Avg Loss: 1.51325712, Time: 0.0401 Steps: 84690, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001669, Sample Num: 26704, Cur Loss: 2.20168018, Cur Avg Loss: 1.17089873, Log Avg loss: 1.11367372, Global Avg Loss: 1.51320994, Time: 0.0480 Steps: 84700, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001679, Sample Num: 26864, Cur Loss: 1.51942587, Cur Avg Loss: 1.17108560, Log Avg loss: 1.20227511, Global Avg Loss: 1.51317324, Time: 0.0513 Steps: 84710, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001689, Sample Num: 27024, Cur Loss: 0.61968935, Cur Avg Loss: 1.17047223, Log Avg loss: 1.06748752, Global Avg Loss: 1.51312063, Time: 0.0403 Steps: 84720, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001699, Sample Num: 27184, Cur Loss: 0.77445817, Cur Avg Loss: 1.17085081, Log Avg loss: 1.23479285, Global Avg Loss: 1.51308778, Time: 0.0741 Steps: 84730, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001709, Sample Num: 27344, Cur Loss: 2.87487316, Cur Avg Loss: 1.17129767, Log Avg loss: 1.24721873, Global Avg Loss: 1.51305641, Time: 0.0750 Steps: 84740, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001719, Sample Num: 27504, Cur Loss: 0.68972319, Cur Avg Loss: 1.17157615, Log Avg loss: 1.21916751, Global Avg Loss: 1.51302173, Time: 0.0443 Steps: 84750, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001729, Sample Num: 27664, Cur Loss: 0.66815042, Cur Avg Loss: 1.17295887, Log Avg loss: 1.41064875, Global Avg Loss: 1.51300965, Time: 0.0828 Steps: 84760, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001739, Sample Num: 27824, Cur Loss: 0.57641113, Cur Avg Loss: 1.17233573, Log Avg loss: 1.06459534, Global Avg Loss: 1.51295675, Time: 0.1007 Steps: 84770, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001749, Sample Num: 27984, Cur Loss: 1.85117090, Cur Avg Loss: 1.17103869, Log Avg loss: 0.94548403, Global Avg Loss: 1.51288982, Time: 0.0588 Steps: 84780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001759, Sample Num: 28144, Cur Loss: 1.20232940, Cur Avg Loss: 1.17158978, Log Avg loss: 1.26797545, Global Avg Loss: 1.51286093, Time: 0.0402 Steps: 84790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001769, Sample Num: 28304, Cur Loss: 1.94440019, Cur Avg Loss: 1.17161386, Log Avg loss: 1.17584808, Global Avg Loss: 1.51282119, Time: 0.0402 Steps: 84800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001779, Sample Num: 28464, Cur Loss: 0.80626047, Cur Avg Loss: 1.17369982, Log Avg loss: 1.54270746, Global Avg Loss: 1.51282472, Time: 0.0402 Steps: 84810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001789, Sample Num: 28624, Cur Loss: 1.67271614, Cur Avg Loss: 1.17263358, Log Avg loss: 0.98294917, Global Avg Loss: 1.51276224, Time: 0.0403 Steps: 84820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001799, Sample Num: 28784, Cur Loss: 1.42813623, Cur Avg Loss: 1.17189237, Log Avg loss: 1.03929039, Global Avg Loss: 1.51270643, Time: 0.0402 Steps: 84830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001809, Sample Num: 28944, Cur Loss: 1.76707172, Cur Avg Loss: 1.17197628, Log Avg loss: 1.18707145, Global Avg Loss: 1.51266805, Time: 0.0403 Steps: 84840, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001819, Sample Num: 29104, Cur Loss: 0.58067226, Cur Avg Loss: 1.17272259, Log Avg loss: 1.30773020, Global Avg Loss: 1.51264390, Time: 0.0403 Steps: 84850, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001829, Sample Num: 29264, Cur Loss: 1.67326045, Cur Avg Loss: 1.17380358, Log Avg loss: 1.37043465, Global Avg Loss: 1.51262714, Time: 0.0403 Steps: 84860, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001839, Sample Num: 29424, Cur Loss: 0.66327006, Cur Avg Loss: 1.17226236, Log Avg loss: 0.89037322, Global Avg Loss: 1.51255382, Time: 0.0403 Steps: 84870, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001849, Sample Num: 29584, Cur Loss: 1.36513722, Cur Avg Loss: 1.17202927, Log Avg loss: 1.12916344, Global Avg Loss: 1.51250865, Time: 0.0402 Steps: 84880, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001859, Sample Num: 29744, Cur Loss: 0.76045775, Cur Avg Loss: 1.17181937, Log Avg loss: 1.13300889, Global Avg Loss: 1.51246395, Time: 0.0403 Steps: 84890, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001869, Sample Num: 29904, Cur Loss: 1.41054940, Cur Avg Loss: 1.17174811, Log Avg loss: 1.15850079, Global Avg Loss: 1.51242225, Time: 0.0403 Steps: 84900, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001879, Sample Num: 30064, Cur Loss: 2.06090307, Cur Avg Loss: 1.17166716, Log Avg loss: 1.15653858, Global Avg Loss: 1.51238034, Time: 0.0402 Steps: 84910, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001889, Sample Num: 30224, Cur Loss: 0.95483482, Cur Avg Loss: 1.17124337, Log Avg loss: 1.09161289, Global Avg Loss: 1.51233079, Time: 0.0403 Steps: 84920, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001899, Sample Num: 30384, Cur Loss: 0.40446627, Cur Avg Loss: 1.17131368, Log Avg loss: 1.18459638, Global Avg Loss: 1.51229220, Time: 0.0402 Steps: 84930, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001909, Sample Num: 30544, Cur Loss: 2.14402890, Cur Avg Loss: 1.17234168, Log Avg loss: 1.36755750, Global Avg Loss: 1.51227516, Time: 0.0402 Steps: 84940, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001919, Sample Num: 30704, Cur Loss: 0.36284214, Cur Avg Loss: 1.17302136, Log Avg loss: 1.30277336, Global Avg Loss: 1.51225050, Time: 0.0402 Steps: 84950, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001929, Sample Num: 30864, Cur Loss: 2.06993079, Cur Avg Loss: 1.17366236, Log Avg loss: 1.29666910, Global Avg Loss: 1.51222513, Time: 0.0402 Steps: 84960, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001939, Sample Num: 31024, Cur Loss: 1.55567753, Cur Avg Loss: 1.17390053, Log Avg loss: 1.21984347, Global Avg Loss: 1.51219072, Time: 0.0402 Steps: 84970, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001949, Sample Num: 31184, Cur Loss: 2.15178108, Cur Avg Loss: 1.17305760, Log Avg loss: 1.00961410, Global Avg Loss: 1.51213158, Time: 0.0402 Steps: 84980, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001959, Sample Num: 31344, Cur Loss: 0.70314789, Cur Avg Loss: 1.17330529, Log Avg loss: 1.22158002, Global Avg Loss: 1.51209739, Time: 0.0402 Steps: 84990, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001969, Sample Num: 31504, Cur Loss: 0.98599494, Cur Avg Loss: 1.17476878, Log Avg loss: 1.46146668, Global Avg Loss: 1.51209143, Time: 0.0402 Steps: 85000, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001979, Sample Num: 31664, Cur Loss: 2.24866438, Cur Avg Loss: 1.17377717, Log Avg loss: 0.97852813, Global Avg Loss: 1.51202867, Time: 0.0402 Steps: 85010, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001989, Sample Num: 31824, Cur Loss: 1.45789099, Cur Avg Loss: 1.17392689, Log Avg loss: 1.20355798, Global Avg Loss: 1.51199239, Time: 0.0402 Steps: 85020, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001999, Sample Num: 31984, Cur Loss: 1.38861275, Cur Avg Loss: 1.17475097, Log Avg loss: 1.33865903, Global Avg Loss: 1.51197200, Time: 0.0402 Steps: 85030, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002009, Sample Num: 32144, Cur Loss: 0.41755947, Cur Avg Loss: 1.17470387, Log Avg loss: 1.16528992, Global Avg Loss: 1.51193123, Time: 0.0402 Steps: 85040, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002019, Sample Num: 32304, Cur Loss: 0.38907659, Cur Avg Loss: 1.17321586, Log Avg loss: 0.87427445, Global Avg Loss: 1.51185626, Time: 0.0561 Steps: 85050, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002029, Sample Num: 32464, Cur Loss: 1.17553067, Cur Avg Loss: 1.17281648, Log Avg loss: 1.09218230, Global Avg Loss: 1.51180692, Time: 0.0776 Steps: 85060, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002039, Sample Num: 32624, Cur Loss: 0.99249780, Cur Avg Loss: 1.17275156, Log Avg loss: 1.15957843, Global Avg Loss: 1.51176552, Time: 0.0495 Steps: 85070, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002049, Sample Num: 32784, Cur Loss: 0.50487918, Cur Avg Loss: 1.17334531, Log Avg loss: 1.29441044, Global Avg Loss: 1.51173997, Time: 0.0491 Steps: 85080, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002059, Sample Num: 32944, Cur Loss: 1.88494611, Cur Avg Loss: 1.17502854, Log Avg loss: 1.51992231, Global Avg Loss: 1.51174093, Time: 0.0647 Steps: 85090, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002069, Sample Num: 33104, Cur Loss: 0.71735489, Cur Avg Loss: 1.17442729, Log Avg loss: 1.05063001, Global Avg Loss: 1.51168675, Time: 0.0958 Steps: 85100, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002079, Sample Num: 33264, Cur Loss: 1.37054181, Cur Avg Loss: 1.17559598, Log Avg loss: 1.41739803, Global Avg Loss: 1.51167567, Time: 0.0431 Steps: 85110, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002089, Sample Num: 33424, Cur Loss: 1.15083778, Cur Avg Loss: 1.17447795, Log Avg loss: 0.94203958, Global Avg Loss: 1.51160875, Time: 0.0560 Steps: 85120, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002099, Sample Num: 33584, Cur Loss: 0.55799437, Cur Avg Loss: 1.17363748, Log Avg loss: 0.99806320, Global Avg Loss: 1.51154842, Time: 0.0406 Steps: 85130, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002109, Sample Num: 33744, Cur Loss: 0.91686869, Cur Avg Loss: 1.17260897, Log Avg loss: 0.95672531, Global Avg Loss: 1.51148326, Time: 0.0403 Steps: 85140, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002119, Sample Num: 33904, Cur Loss: 0.69268423, Cur Avg Loss: 1.17147270, Log Avg loss: 0.93183395, Global Avg Loss: 1.51141518, Time: 0.0402 Steps: 85150, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002129, Sample Num: 34055, Cur Loss: 0.87720764, Cur Avg Loss: 1.17209776, Log Avg loss: 1.30454713, Global Avg Loss: 1.51139089, Time: 0.0198 Steps: 85160, Updated lr: 0.000020 ***** Running evaluation checkpoint-85160 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-85160 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.516239, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.104019, "eval_total_loss": 776.125251, "eval_mae": 0.842362, "eval_mse": 1.104173, "eval_r2": 0.298115, "eval_sp_statistic": 0.449545, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.554882, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.106317, "test_total_loss": 555.371223, "test_mae": 0.868841, "test_mse": 1.106537, "test_r2": 0.285831, "test_sp_statistic": 0.476423, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.596327, "test_ps_pvalue": 0.0, "lr": 2.0189663347558086e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5113908906122113, "train_cur_epoch_loss": 2495.3961310833693, "train_cur_epoch_avg_loss": 1.1720977600203708, "train_cur_epoch_time": 95.51623868942261, "train_cur_epoch_avg_time": 0.0448643676324202, "epoch": 40, "step": 85160} ################################################## Training, Epoch: 0041, Batch: 000010, Sample Num: 160, Cur Loss: 1.19527841, Cur Avg Loss: 1.10050185, Log Avg loss: 1.10050185, Global Avg Loss: 1.51134265, Time: 0.0638 Steps: 85170, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000020, Sample Num: 320, Cur Loss: 2.23215818, Cur Avg Loss: 1.12048955, Log Avg loss: 1.14047725, Global Avg Loss: 1.51129911, Time: 0.0635 Steps: 85180, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000030, Sample Num: 480, Cur Loss: 1.03442907, Cur Avg Loss: 1.15136318, Log Avg loss: 1.21311044, Global Avg Loss: 1.51126411, Time: 0.0402 Steps: 85190, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000040, Sample Num: 640, Cur Loss: 0.81377709, Cur Avg Loss: 1.21201605, Log Avg loss: 1.39397467, Global Avg Loss: 1.51125034, Time: 0.0403 Steps: 85200, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000050, Sample Num: 800, Cur Loss: 0.37718505, Cur Avg Loss: 1.17087341, Log Avg loss: 1.00630283, Global Avg Loss: 1.51119108, Time: 0.0402 Steps: 85210, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000060, Sample Num: 960, Cur Loss: 2.51324677, Cur Avg Loss: 1.14715855, Log Avg loss: 1.02858428, Global Avg Loss: 1.51113445, Time: 0.0402 Steps: 85220, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000070, Sample Num: 1120, Cur Loss: 2.55298758, Cur Avg Loss: 1.19646830, Log Avg loss: 1.49232675, Global Avg Loss: 1.51113224, Time: 0.0402 Steps: 85230, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000080, Sample Num: 1280, Cur Loss: 1.51364601, Cur Avg Loss: 1.20946837, Log Avg loss: 1.30046889, Global Avg Loss: 1.51110753, Time: 0.0403 Steps: 85240, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000090, Sample Num: 1440, Cur Loss: 1.75545716, Cur Avg Loss: 1.20618980, Log Avg loss: 1.17996123, Global Avg Loss: 1.51106868, Time: 0.0402 Steps: 85250, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000100, Sample Num: 1600, Cur Loss: 1.00502980, Cur Avg Loss: 1.19862877, Log Avg loss: 1.13057948, Global Avg Loss: 1.51102406, Time: 0.0402 Steps: 85260, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000110, Sample Num: 1760, Cur Loss: 0.86412704, Cur Avg Loss: 1.15717567, Log Avg loss: 0.74264469, Global Avg Loss: 1.51093395, Time: 0.0402 Steps: 85270, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000120, Sample Num: 1920, Cur Loss: 1.03343189, Cur Avg Loss: 1.18285985, Log Avg loss: 1.46538584, Global Avg Loss: 1.51092860, Time: 0.0402 Steps: 85280, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000130, Sample Num: 2080, Cur Loss: 0.86856753, Cur Avg Loss: 1.17725148, Log Avg loss: 1.10995101, Global Avg Loss: 1.51088159, Time: 0.0404 Steps: 85290, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000140, Sample Num: 2240, Cur Loss: 1.99707329, Cur Avg Loss: 1.18476153, Log Avg loss: 1.28239225, Global Avg Loss: 1.51085480, Time: 0.0402 Steps: 85300, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000150, Sample Num: 2400, Cur Loss: 1.42315340, Cur Avg Loss: 1.17505960, Log Avg loss: 1.03923249, Global Avg Loss: 1.51079952, Time: 0.0402 Steps: 85310, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000160, Sample Num: 2560, Cur Loss: 1.32459021, Cur Avg Loss: 1.16878629, Log Avg loss: 1.07468669, Global Avg Loss: 1.51074841, Time: 0.0403 Steps: 85320, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000170, Sample Num: 2720, Cur Loss: 0.72043175, Cur Avg Loss: 1.16449264, Log Avg loss: 1.09579422, Global Avg Loss: 1.51069978, Time: 0.0402 Steps: 85330, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000180, Sample Num: 2880, Cur Loss: 2.00061202, Cur Avg Loss: 1.16679403, Log Avg loss: 1.20591768, Global Avg Loss: 1.51066406, Time: 0.0403 Steps: 85340, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000190, Sample Num: 3040, Cur Loss: 0.89643568, Cur Avg Loss: 1.16964833, Log Avg loss: 1.22102575, Global Avg Loss: 1.51063013, Time: 0.0402 Steps: 85350, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000200, Sample Num: 3200, Cur Loss: 0.83790118, Cur Avg Loss: 1.16817488, Log Avg loss: 1.14017924, Global Avg Loss: 1.51058673, Time: 0.0402 Steps: 85360, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000210, Sample Num: 3360, Cur Loss: 0.25648952, Cur Avg Loss: 1.16664935, Log Avg loss: 1.13613878, Global Avg Loss: 1.51054287, Time: 0.0402 Steps: 85370, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000220, Sample Num: 3520, Cur Loss: 1.21584105, Cur Avg Loss: 1.16190820, Log Avg loss: 1.06234405, Global Avg Loss: 1.51049037, Time: 0.0402 Steps: 85380, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000230, Sample Num: 3680, Cur Loss: 0.57181638, Cur Avg Loss: 1.16227462, Log Avg loss: 1.17033600, Global Avg Loss: 1.51045054, Time: 0.0402 Steps: 85390, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000240, Sample Num: 3840, Cur Loss: 1.08965302, Cur Avg Loss: 1.16822064, Log Avg loss: 1.30497897, Global Avg Loss: 1.51042648, Time: 0.0402 Steps: 85400, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000250, Sample Num: 4000, Cur Loss: 0.78347760, Cur Avg Loss: 1.16511559, Log Avg loss: 1.09059449, Global Avg Loss: 1.51037732, Time: 0.0402 Steps: 85410, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000260, Sample Num: 4160, Cur Loss: 0.52203351, Cur Avg Loss: 1.15416470, Log Avg loss: 0.88039241, Global Avg Loss: 1.51030357, Time: 0.0402 Steps: 85420, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000270, Sample Num: 4320, Cur Loss: 1.17661309, Cur Avg Loss: 1.15792189, Log Avg loss: 1.25560880, Global Avg Loss: 1.51027376, Time: 0.0521 Steps: 85430, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000280, Sample Num: 4480, Cur Loss: 0.56468338, Cur Avg Loss: 1.14409818, Log Avg loss: 0.77085809, Global Avg Loss: 1.51018722, Time: 0.0926 Steps: 85440, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000290, Sample Num: 4640, Cur Loss: 2.48364091, Cur Avg Loss: 1.15505108, Log Avg loss: 1.46173209, Global Avg Loss: 1.51018155, Time: 0.0411 Steps: 85450, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000300, Sample Num: 4800, Cur Loss: 1.31247151, Cur Avg Loss: 1.16127957, Log Avg loss: 1.34190594, Global Avg Loss: 1.51016185, Time: 0.0588 Steps: 85460, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000310, Sample Num: 4960, Cur Loss: 1.11055779, Cur Avg Loss: 1.17070384, Log Avg loss: 1.45343184, Global Avg Loss: 1.51015522, Time: 0.0436 Steps: 85470, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000320, Sample Num: 5120, Cur Loss: 1.87496352, Cur Avg Loss: 1.17481497, Log Avg loss: 1.30226007, Global Avg Loss: 1.51013090, Time: 0.0403 Steps: 85480, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000330, Sample Num: 5280, Cur Loss: 1.03525317, Cur Avg Loss: 1.17315533, Log Avg loss: 1.12004690, Global Avg Loss: 1.51008527, Time: 0.0406 Steps: 85490, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000340, Sample Num: 5440, Cur Loss: 1.83160329, Cur Avg Loss: 1.17840557, Log Avg loss: 1.35166335, Global Avg Loss: 1.51006674, Time: 0.0795 Steps: 85500, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000350, Sample Num: 5600, Cur Loss: 0.91920292, Cur Avg Loss: 1.18007877, Log Avg loss: 1.23696774, Global Avg Loss: 1.51003480, Time: 0.0722 Steps: 85510, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000360, Sample Num: 5760, Cur Loss: 1.58050561, Cur Avg Loss: 1.17440738, Log Avg loss: 0.97590848, Global Avg Loss: 1.50997234, Time: 0.0402 Steps: 85520, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000370, Sample Num: 5920, Cur Loss: 0.42041942, Cur Avg Loss: 1.17785779, Log Avg loss: 1.30207262, Global Avg Loss: 1.50994804, Time: 0.0402 Steps: 85530, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000380, Sample Num: 6080, Cur Loss: 2.62270308, Cur Avg Loss: 1.18257720, Log Avg loss: 1.35719545, Global Avg Loss: 1.50993018, Time: 0.0402 Steps: 85540, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000390, Sample Num: 6240, Cur Loss: 0.49878109, Cur Avg Loss: 1.18380478, Log Avg loss: 1.23045282, Global Avg Loss: 1.50989751, Time: 0.0404 Steps: 85550, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000400, Sample Num: 6400, Cur Loss: 0.55802923, Cur Avg Loss: 1.17975963, Log Avg loss: 1.02199893, Global Avg Loss: 1.50984049, Time: 0.0403 Steps: 85560, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000410, Sample Num: 6560, Cur Loss: 0.91293931, Cur Avg Loss: 1.17391542, Log Avg loss: 0.94014673, Global Avg Loss: 1.50977391, Time: 0.0403 Steps: 85570, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000420, Sample Num: 6720, Cur Loss: 0.98824561, Cur Avg Loss: 1.16867872, Log Avg loss: 0.95397400, Global Avg Loss: 1.50970897, Time: 0.0403 Steps: 85580, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000430, Sample Num: 6880, Cur Loss: 0.83676326, Cur Avg Loss: 1.16362066, Log Avg loss: 0.95118225, Global Avg Loss: 1.50964371, Time: 0.0402 Steps: 85590, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000440, Sample Num: 7040, Cur Loss: 0.86805356, Cur Avg Loss: 1.16542608, Log Avg loss: 1.24305928, Global Avg Loss: 1.50961257, Time: 0.0403 Steps: 85600, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000450, Sample Num: 7200, Cur Loss: 0.58038759, Cur Avg Loss: 1.17155659, Log Avg loss: 1.44129914, Global Avg Loss: 1.50960459, Time: 0.0403 Steps: 85610, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000460, Sample Num: 7360, Cur Loss: 1.33212495, Cur Avg Loss: 1.17729068, Log Avg loss: 1.43532465, Global Avg Loss: 1.50959591, Time: 0.0407 Steps: 85620, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000470, Sample Num: 7520, Cur Loss: 1.27296996, Cur Avg Loss: 1.17804069, Log Avg loss: 1.21254095, Global Avg Loss: 1.50956122, Time: 0.0404 Steps: 85630, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000480, Sample Num: 7680, Cur Loss: 0.66229308, Cur Avg Loss: 1.17807348, Log Avg loss: 1.17961467, Global Avg Loss: 1.50952269, Time: 0.0404 Steps: 85640, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000490, Sample Num: 7840, Cur Loss: 1.31819105, Cur Avg Loss: 1.17616273, Log Avg loss: 1.08444684, Global Avg Loss: 1.50947306, Time: 0.0405 Steps: 85650, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000500, Sample Num: 8000, Cur Loss: 0.56410187, Cur Avg Loss: 1.17200249, Log Avg loss: 0.96815059, Global Avg Loss: 1.50940987, Time: 0.0404 Steps: 85660, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000510, Sample Num: 8160, Cur Loss: 0.69092572, Cur Avg Loss: 1.16924243, Log Avg loss: 1.03123951, Global Avg Loss: 1.50935405, Time: 0.0404 Steps: 85670, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000520, Sample Num: 8320, Cur Loss: 0.54896736, Cur Avg Loss: 1.16618331, Log Avg loss: 1.01016830, Global Avg Loss: 1.50929579, Time: 0.0404 Steps: 85680, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000530, Sample Num: 8480, Cur Loss: 0.74992740, Cur Avg Loss: 1.16351757, Log Avg loss: 1.02489890, Global Avg Loss: 1.50923926, Time: 0.0404 Steps: 85690, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000540, Sample Num: 8640, Cur Loss: 1.63834655, Cur Avg Loss: 1.16086491, Log Avg loss: 1.02027420, Global Avg Loss: 1.50918221, Time: 0.0404 Steps: 85700, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000550, Sample Num: 8800, Cur Loss: 2.31423855, Cur Avg Loss: 1.15744711, Log Avg loss: 0.97288560, Global Avg Loss: 1.50911964, Time: 0.0403 Steps: 85710, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000560, Sample Num: 8960, Cur Loss: 0.27645171, Cur Avg Loss: 1.15699805, Log Avg loss: 1.13230004, Global Avg Loss: 1.50907568, Time: 0.0404 Steps: 85720, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000570, Sample Num: 9120, Cur Loss: 0.36423436, Cur Avg Loss: 1.15285078, Log Avg loss: 0.92060322, Global Avg Loss: 1.50900704, Time: 0.0403 Steps: 85730, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000580, Sample Num: 9280, Cur Loss: 0.49732631, Cur Avg Loss: 1.15639472, Log Avg loss: 1.35839973, Global Avg Loss: 1.50898947, Time: 0.0404 Steps: 85740, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000590, Sample Num: 9440, Cur Loss: 1.39764166, Cur Avg Loss: 1.15685437, Log Avg loss: 1.18351393, Global Avg Loss: 1.50895151, Time: 0.0404 Steps: 85750, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000600, Sample Num: 9600, Cur Loss: 0.42070085, Cur Avg Loss: 1.15386656, Log Avg loss: 0.97758558, Global Avg Loss: 1.50888955, Time: 0.0404 Steps: 85760, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000610, Sample Num: 9760, Cur Loss: 1.60371578, Cur Avg Loss: 1.15053183, Log Avg loss: 0.95044802, Global Avg Loss: 1.50882445, Time: 0.0405 Steps: 85770, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000620, Sample Num: 9920, Cur Loss: 0.99266922, Cur Avg Loss: 1.14750128, Log Avg loss: 0.96263757, Global Avg Loss: 1.50876077, Time: 0.0410 Steps: 85780, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000630, Sample Num: 10080, Cur Loss: 0.71578336, Cur Avg Loss: 1.15291482, Log Avg loss: 1.48855445, Global Avg Loss: 1.50875842, Time: 0.0510 Steps: 85790, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000640, Sample Num: 10240, Cur Loss: 0.71723825, Cur Avg Loss: 1.15367657, Log Avg loss: 1.20166668, Global Avg Loss: 1.50872263, Time: 0.0669 Steps: 85800, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000650, Sample Num: 10400, Cur Loss: 0.25959945, Cur Avg Loss: 1.15107695, Log Avg loss: 0.98470164, Global Avg Loss: 1.50866156, Time: 0.0634 Steps: 85810, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000660, Sample Num: 10560, Cur Loss: 1.34146583, Cur Avg Loss: 1.14951972, Log Avg loss: 1.04829973, Global Avg Loss: 1.50860791, Time: 0.0507 Steps: 85820, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000670, Sample Num: 10720, Cur Loss: 1.19510460, Cur Avg Loss: 1.15225779, Log Avg loss: 1.33297011, Global Avg Loss: 1.50858745, Time: 0.0608 Steps: 85830, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000680, Sample Num: 10880, Cur Loss: 0.61583304, Cur Avg Loss: 1.14914965, Log Avg loss: 0.94090451, Global Avg Loss: 1.50852132, Time: 0.0686 Steps: 85840, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000690, Sample Num: 11040, Cur Loss: 0.50271499, Cur Avg Loss: 1.15002875, Log Avg loss: 1.20980730, Global Avg Loss: 1.50848652, Time: 0.0457 Steps: 85850, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000700, Sample Num: 11200, Cur Loss: 0.89848006, Cur Avg Loss: 1.15255971, Log Avg loss: 1.32719603, Global Avg Loss: 1.50846541, Time: 0.0779 Steps: 85860, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000710, Sample Num: 11360, Cur Loss: 1.09964895, Cur Avg Loss: 1.15234008, Log Avg loss: 1.13696609, Global Avg Loss: 1.50842215, Time: 0.0403 Steps: 85870, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000720, Sample Num: 11520, Cur Loss: 1.02479053, Cur Avg Loss: 1.15194672, Log Avg loss: 1.12401801, Global Avg Loss: 1.50837739, Time: 0.0402 Steps: 85880, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000730, Sample Num: 11680, Cur Loss: 0.94008249, Cur Avg Loss: 1.15277072, Log Avg loss: 1.21209916, Global Avg Loss: 1.50834289, Time: 0.0402 Steps: 85890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000740, Sample Num: 11840, Cur Loss: 0.42911530, Cur Avg Loss: 1.15261147, Log Avg loss: 1.14098626, Global Avg Loss: 1.50830012, Time: 0.0402 Steps: 85900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000750, Sample Num: 12000, Cur Loss: 0.70259815, Cur Avg Loss: 1.15549181, Log Avg loss: 1.36863645, Global Avg Loss: 1.50828387, Time: 0.0402 Steps: 85910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000760, Sample Num: 12160, Cur Loss: 0.70178318, Cur Avg Loss: 1.15370449, Log Avg loss: 1.01965594, Global Avg Loss: 1.50822700, Time: 0.0402 Steps: 85920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000770, Sample Num: 12320, Cur Loss: 0.70646065, Cur Avg Loss: 1.15275339, Log Avg loss: 1.08046957, Global Avg Loss: 1.50817722, Time: 0.0402 Steps: 85930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000780, Sample Num: 12480, Cur Loss: 1.38081717, Cur Avg Loss: 1.15214922, Log Avg loss: 1.10562829, Global Avg Loss: 1.50813038, Time: 0.0402 Steps: 85940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000790, Sample Num: 12640, Cur Loss: 2.28907967, Cur Avg Loss: 1.15549178, Log Avg loss: 1.41621111, Global Avg Loss: 1.50811968, Time: 0.0403 Steps: 85950, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000800, Sample Num: 12800, Cur Loss: 0.25914732, Cur Avg Loss: 1.15180114, Log Avg loss: 0.86024082, Global Avg Loss: 1.50804431, Time: 0.0402 Steps: 85960, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000810, Sample Num: 12960, Cur Loss: 1.63040173, Cur Avg Loss: 1.15205675, Log Avg loss: 1.17250580, Global Avg Loss: 1.50800528, Time: 0.0401 Steps: 85970, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000820, Sample Num: 13120, Cur Loss: 1.46163499, Cur Avg Loss: 1.14986180, Log Avg loss: 0.97207043, Global Avg Loss: 1.50794295, Time: 0.0402 Steps: 85980, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000830, Sample Num: 13280, Cur Loss: 0.89159030, Cur Avg Loss: 1.14849360, Log Avg loss: 1.03630149, Global Avg Loss: 1.50788810, Time: 0.0402 Steps: 85990, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000840, Sample Num: 13440, Cur Loss: 1.88221455, Cur Avg Loss: 1.15065774, Log Avg loss: 1.33028115, Global Avg Loss: 1.50786745, Time: 0.0402 Steps: 86000, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000850, Sample Num: 13600, Cur Loss: 1.97740412, Cur Avg Loss: 1.15427736, Log Avg loss: 1.45832561, Global Avg Loss: 1.50786169, Time: 0.0402 Steps: 86010, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000860, Sample Num: 13760, Cur Loss: 1.02062190, Cur Avg Loss: 1.15505910, Log Avg loss: 1.22150718, Global Avg Loss: 1.50782840, Time: 0.0403 Steps: 86020, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000870, Sample Num: 13920, Cur Loss: 1.01598597, Cur Avg Loss: 1.15088834, Log Avg loss: 0.79220303, Global Avg Loss: 1.50774522, Time: 0.0402 Steps: 86030, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000880, Sample Num: 14080, Cur Loss: 1.16076422, Cur Avg Loss: 1.15160055, Log Avg loss: 1.21356218, Global Avg Loss: 1.50771103, Time: 0.0402 Steps: 86040, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000890, Sample Num: 14240, Cur Loss: 1.83473158, Cur Avg Loss: 1.15333031, Log Avg loss: 1.30554952, Global Avg Loss: 1.50768753, Time: 0.0401 Steps: 86050, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000900, Sample Num: 14400, Cur Loss: 0.58002615, Cur Avg Loss: 1.15147507, Log Avg loss: 0.98635884, Global Avg Loss: 1.50762696, Time: 0.0402 Steps: 86060, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000910, Sample Num: 14560, Cur Loss: 1.04433429, Cur Avg Loss: 1.15236159, Log Avg loss: 1.23214828, Global Avg Loss: 1.50759495, Time: 0.0402 Steps: 86070, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000920, Sample Num: 14720, Cur Loss: 1.13556242, Cur Avg Loss: 1.15386330, Log Avg loss: 1.29051908, Global Avg Loss: 1.50756973, Time: 0.0402 Steps: 86080, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000930, Sample Num: 14880, Cur Loss: 1.38288260, Cur Avg Loss: 1.15439670, Log Avg loss: 1.20346905, Global Avg Loss: 1.50753441, Time: 0.0402 Steps: 86090, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000940, Sample Num: 15040, Cur Loss: 0.81048524, Cur Avg Loss: 1.15438380, Log Avg loss: 1.15318469, Global Avg Loss: 1.50749325, Time: 0.0402 Steps: 86100, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000950, Sample Num: 15200, Cur Loss: 1.96085334, Cur Avg Loss: 1.15406419, Log Avg loss: 1.12402030, Global Avg Loss: 1.50744872, Time: 0.0402 Steps: 86110, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000960, Sample Num: 15360, Cur Loss: 0.56724894, Cur Avg Loss: 1.15141977, Log Avg loss: 0.90019969, Global Avg Loss: 1.50737821, Time: 0.0482 Steps: 86120, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000970, Sample Num: 15520, Cur Loss: 1.06037688, Cur Avg Loss: 1.15492299, Log Avg loss: 1.49123200, Global Avg Loss: 1.50737633, Time: 0.0404 Steps: 86130, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000980, Sample Num: 15680, Cur Loss: 1.59002376, Cur Avg Loss: 1.15665193, Log Avg loss: 1.32436005, Global Avg Loss: 1.50735509, Time: 0.0407 Steps: 86140, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000990, Sample Num: 15840, Cur Loss: 2.77281475, Cur Avg Loss: 1.15850745, Log Avg loss: 1.34034804, Global Avg Loss: 1.50733570, Time: 0.0411 Steps: 86150, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001000, Sample Num: 16000, Cur Loss: 0.85075343, Cur Avg Loss: 1.15811279, Log Avg loss: 1.11904114, Global Avg Loss: 1.50729063, Time: 0.0406 Steps: 86160, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001010, Sample Num: 16160, Cur Loss: 0.87726647, Cur Avg Loss: 1.15965350, Log Avg loss: 1.31372478, Global Avg Loss: 1.50726817, Time: 0.0486 Steps: 86170, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001020, Sample Num: 16320, Cur Loss: 1.74031138, Cur Avg Loss: 1.16079906, Log Avg loss: 1.27650086, Global Avg Loss: 1.50724139, Time: 0.0489 Steps: 86180, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001030, Sample Num: 16480, Cur Loss: 0.89427882, Cur Avg Loss: 1.16276084, Log Avg loss: 1.36286180, Global Avg Loss: 1.50722464, Time: 0.0564 Steps: 86190, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001040, Sample Num: 16640, Cur Loss: 0.97275776, Cur Avg Loss: 1.16221901, Log Avg loss: 1.10641067, Global Avg Loss: 1.50717814, Time: 0.0406 Steps: 86200, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001050, Sample Num: 16800, Cur Loss: 1.10825598, Cur Avg Loss: 1.15921922, Log Avg loss: 0.84724127, Global Avg Loss: 1.50710159, Time: 0.0403 Steps: 86210, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001060, Sample Num: 16960, Cur Loss: 0.66506481, Cur Avg Loss: 1.15785596, Log Avg loss: 1.01471334, Global Avg Loss: 1.50704449, Time: 0.0402 Steps: 86220, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001070, Sample Num: 17120, Cur Loss: 0.57169235, Cur Avg Loss: 1.15962566, Log Avg loss: 1.34721412, Global Avg Loss: 1.50702595, Time: 0.0403 Steps: 86230, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001080, Sample Num: 17280, Cur Loss: 0.22787037, Cur Avg Loss: 1.15741993, Log Avg loss: 0.92140710, Global Avg Loss: 1.50695804, Time: 0.0403 Steps: 86240, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001090, Sample Num: 17440, Cur Loss: 1.39784670, Cur Avg Loss: 1.15912143, Log Avg loss: 1.34288261, Global Avg Loss: 1.50693902, Time: 0.0402 Steps: 86250, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001100, Sample Num: 17600, Cur Loss: 0.59738863, Cur Avg Loss: 1.15984211, Log Avg loss: 1.23839700, Global Avg Loss: 1.50690789, Time: 0.0403 Steps: 86260, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001110, Sample Num: 17760, Cur Loss: 0.83287519, Cur Avg Loss: 1.15924950, Log Avg loss: 1.09406200, Global Avg Loss: 1.50686003, Time: 0.0402 Steps: 86270, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001120, Sample Num: 17920, Cur Loss: 1.05928147, Cur Avg Loss: 1.15899550, Log Avg loss: 1.13080186, Global Avg Loss: 1.50681645, Time: 0.0402 Steps: 86280, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001130, Sample Num: 18080, Cur Loss: 1.69181502, Cur Avg Loss: 1.15869448, Log Avg loss: 1.12498060, Global Avg Loss: 1.50677220, Time: 0.0402 Steps: 86290, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001140, Sample Num: 18240, Cur Loss: 1.31960118, Cur Avg Loss: 1.16001055, Log Avg loss: 1.30872537, Global Avg Loss: 1.50674925, Time: 0.0403 Steps: 86300, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001150, Sample Num: 18400, Cur Loss: 0.89832819, Cur Avg Loss: 1.16127690, Log Avg loss: 1.30564086, Global Avg Loss: 1.50672595, Time: 0.0402 Steps: 86310, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001160, Sample Num: 18560, Cur Loss: 1.28211975, Cur Avg Loss: 1.16104066, Log Avg loss: 1.13387398, Global Avg Loss: 1.50668276, Time: 0.0403 Steps: 86320, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001170, Sample Num: 18720, Cur Loss: 1.82132268, Cur Avg Loss: 1.16358325, Log Avg loss: 1.45852338, Global Avg Loss: 1.50667718, Time: 0.0402 Steps: 86330, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001180, Sample Num: 18880, Cur Loss: 0.76055658, Cur Avg Loss: 1.16343939, Log Avg loss: 1.14660745, Global Avg Loss: 1.50663547, Time: 0.0402 Steps: 86340, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001190, Sample Num: 19040, Cur Loss: 0.59900439, Cur Avg Loss: 1.16352759, Log Avg loss: 1.17393531, Global Avg Loss: 1.50659694, Time: 0.0402 Steps: 86350, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001200, Sample Num: 19200, Cur Loss: 1.27937353, Cur Avg Loss: 1.16285650, Log Avg loss: 1.08299742, Global Avg Loss: 1.50654789, Time: 0.0402 Steps: 86360, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001210, Sample Num: 19360, Cur Loss: 0.99201339, Cur Avg Loss: 1.16381613, Log Avg loss: 1.27897125, Global Avg Loss: 1.50652154, Time: 0.0402 Steps: 86370, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001220, Sample Num: 19520, Cur Loss: 0.27645800, Cur Avg Loss: 1.16390660, Log Avg loss: 1.17485334, Global Avg Loss: 1.50648315, Time: 0.0402 Steps: 86380, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001230, Sample Num: 19680, Cur Loss: 0.79580384, Cur Avg Loss: 1.16366865, Log Avg loss: 1.13463866, Global Avg Loss: 1.50644011, Time: 0.0402 Steps: 86390, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001240, Sample Num: 19840, Cur Loss: 0.94835913, Cur Avg Loss: 1.16518220, Log Avg loss: 1.35134948, Global Avg Loss: 1.50642215, Time: 0.0402 Steps: 86400, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001250, Sample Num: 20000, Cur Loss: 0.89794314, Cur Avg Loss: 1.16573049, Log Avg loss: 1.23371810, Global Avg Loss: 1.50639060, Time: 0.0402 Steps: 86410, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001260, Sample Num: 20160, Cur Loss: 0.81868172, Cur Avg Loss: 1.16615654, Log Avg loss: 1.21941233, Global Avg Loss: 1.50635739, Time: 0.0402 Steps: 86420, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001270, Sample Num: 20320, Cur Loss: 1.44798136, Cur Avg Loss: 1.16717871, Log Avg loss: 1.29597254, Global Avg Loss: 1.50633305, Time: 0.0402 Steps: 86430, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001280, Sample Num: 20480, Cur Loss: 2.25683427, Cur Avg Loss: 1.16724115, Log Avg loss: 1.17517127, Global Avg Loss: 1.50629474, Time: 0.0403 Steps: 86440, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001290, Sample Num: 20640, Cur Loss: 1.38522577, Cur Avg Loss: 1.16936657, Log Avg loss: 1.44142047, Global Avg Loss: 1.50628723, Time: 0.0402 Steps: 86450, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001300, Sample Num: 20800, Cur Loss: 1.30677843, Cur Avg Loss: 1.16948519, Log Avg loss: 1.18478656, Global Avg Loss: 1.50625005, Time: 0.0402 Steps: 86460, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001310, Sample Num: 20960, Cur Loss: 1.38948894, Cur Avg Loss: 1.16825333, Log Avg loss: 1.00811124, Global Avg Loss: 1.50619244, Time: 0.0403 Steps: 86470, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001320, Sample Num: 21120, Cur Loss: 3.24768257, Cur Avg Loss: 1.17143960, Log Avg loss: 1.58884118, Global Avg Loss: 1.50620199, Time: 0.0481 Steps: 86480, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001330, Sample Num: 21280, Cur Loss: 0.66988450, Cur Avg Loss: 1.17153589, Log Avg loss: 1.18424601, Global Avg Loss: 1.50616477, Time: 0.0672 Steps: 86490, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001340, Sample Num: 21440, Cur Loss: 0.66522378, Cur Avg Loss: 1.16945778, Log Avg loss: 0.89306911, Global Avg Loss: 1.50609389, Time: 0.0460 Steps: 86500, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001350, Sample Num: 21600, Cur Loss: 0.93378210, Cur Avg Loss: 1.16830887, Log Avg loss: 1.01435558, Global Avg Loss: 1.50603705, Time: 0.0404 Steps: 86510, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001360, Sample Num: 21760, Cur Loss: 1.30387652, Cur Avg Loss: 1.16772363, Log Avg loss: 1.08871663, Global Avg Loss: 1.50598882, Time: 0.0404 Steps: 86520, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001370, Sample Num: 21920, Cur Loss: 0.27791181, Cur Avg Loss: 1.16695603, Log Avg loss: 1.06256181, Global Avg Loss: 1.50593757, Time: 0.0979 Steps: 86530, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001380, Sample Num: 22080, Cur Loss: 0.39024442, Cur Avg Loss: 1.16532302, Log Avg loss: 0.94160060, Global Avg Loss: 1.50587236, Time: 0.0560 Steps: 86540, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001390, Sample Num: 22240, Cur Loss: 2.91099787, Cur Avg Loss: 1.16462445, Log Avg loss: 1.06822139, Global Avg Loss: 1.50582179, Time: 0.0673 Steps: 86550, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001400, Sample Num: 22400, Cur Loss: 1.35430157, Cur Avg Loss: 1.16616046, Log Avg loss: 1.37966642, Global Avg Loss: 1.50580722, Time: 0.0570 Steps: 86560, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001410, Sample Num: 22560, Cur Loss: 0.50481778, Cur Avg Loss: 1.16564771, Log Avg loss: 1.09386301, Global Avg Loss: 1.50575963, Time: 0.0402 Steps: 86570, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001420, Sample Num: 22720, Cur Loss: 1.01990819, Cur Avg Loss: 1.16526574, Log Avg loss: 1.11140772, Global Avg Loss: 1.50571409, Time: 0.0403 Steps: 86580, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001430, Sample Num: 22880, Cur Loss: 0.97336364, Cur Avg Loss: 1.16589137, Log Avg loss: 1.25473053, Global Avg Loss: 1.50568510, Time: 0.0402 Steps: 86590, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001440, Sample Num: 23040, Cur Loss: 1.61534286, Cur Avg Loss: 1.16370238, Log Avg loss: 0.85067740, Global Avg Loss: 1.50560947, Time: 0.0403 Steps: 86600, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001450, Sample Num: 23200, Cur Loss: 3.00189829, Cur Avg Loss: 1.16381575, Log Avg loss: 1.18014111, Global Avg Loss: 1.50557189, Time: 0.0402 Steps: 86610, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001460, Sample Num: 23360, Cur Loss: 1.03348470, Cur Avg Loss: 1.16472125, Log Avg loss: 1.29601853, Global Avg Loss: 1.50554769, Time: 0.0403 Steps: 86620, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001470, Sample Num: 23520, Cur Loss: 0.97377932, Cur Avg Loss: 1.16572523, Log Avg loss: 1.31230655, Global Avg Loss: 1.50552539, Time: 0.0402 Steps: 86630, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001480, Sample Num: 23680, Cur Loss: 1.93747592, Cur Avg Loss: 1.16633807, Log Avg loss: 1.25642459, Global Avg Loss: 1.50549664, Time: 0.0402 Steps: 86640, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001490, Sample Num: 23840, Cur Loss: 1.66777706, Cur Avg Loss: 1.16672654, Log Avg loss: 1.22422113, Global Avg Loss: 1.50546418, Time: 0.0402 Steps: 86650, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001500, Sample Num: 24000, Cur Loss: 2.01418805, Cur Avg Loss: 1.16887233, Log Avg loss: 1.48859510, Global Avg Loss: 1.50546223, Time: 0.0403 Steps: 86660, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001510, Sample Num: 24160, Cur Loss: 1.20621908, Cur Avg Loss: 1.16980249, Log Avg loss: 1.30932646, Global Avg Loss: 1.50543960, Time: 0.0403 Steps: 86670, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001520, Sample Num: 24320, Cur Loss: 1.03686357, Cur Avg Loss: 1.16790075, Log Avg loss: 0.88073746, Global Avg Loss: 1.50536753, Time: 0.0402 Steps: 86680, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001530, Sample Num: 24480, Cur Loss: 1.13906610, Cur Avg Loss: 1.17050628, Log Avg loss: 1.56654595, Global Avg Loss: 1.50537459, Time: 0.0402 Steps: 86690, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001540, Sample Num: 24640, Cur Loss: 0.75250459, Cur Avg Loss: 1.16874021, Log Avg loss: 0.89853153, Global Avg Loss: 1.50530459, Time: 0.0402 Steps: 86700, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001550, Sample Num: 24800, Cur Loss: 0.87167245, Cur Avg Loss: 1.16788434, Log Avg loss: 1.03608156, Global Avg Loss: 1.50525048, Time: 0.0402 Steps: 86710, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001560, Sample Num: 24960, Cur Loss: 1.54886365, Cur Avg Loss: 1.16673878, Log Avg loss: 0.98917633, Global Avg Loss: 1.50519097, Time: 0.0402 Steps: 86720, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001570, Sample Num: 25120, Cur Loss: 1.16856480, Cur Avg Loss: 1.16634533, Log Avg loss: 1.10496789, Global Avg Loss: 1.50514482, Time: 0.0403 Steps: 86730, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001580, Sample Num: 25280, Cur Loss: 0.21834014, Cur Avg Loss: 1.16418468, Log Avg loss: 0.82496128, Global Avg Loss: 1.50506641, Time: 0.0402 Steps: 86740, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001590, Sample Num: 25440, Cur Loss: 0.21377575, Cur Avg Loss: 1.16420965, Log Avg loss: 1.16815596, Global Avg Loss: 1.50502757, Time: 0.0402 Steps: 86750, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001600, Sample Num: 25600, Cur Loss: 0.72828639, Cur Avg Loss: 1.16254352, Log Avg loss: 0.89762844, Global Avg Loss: 1.50495756, Time: 0.0402 Steps: 86760, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001610, Sample Num: 25760, Cur Loss: 1.66894722, Cur Avg Loss: 1.16200644, Log Avg loss: 1.07607338, Global Avg Loss: 1.50490813, Time: 0.0401 Steps: 86770, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001620, Sample Num: 25920, Cur Loss: 1.51184559, Cur Avg Loss: 1.16405987, Log Avg loss: 1.49466266, Global Avg Loss: 1.50490695, Time: 0.0402 Steps: 86780, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001630, Sample Num: 26080, Cur Loss: 2.18938041, Cur Avg Loss: 1.16594725, Log Avg loss: 1.47170237, Global Avg Loss: 1.50490313, Time: 0.0402 Steps: 86790, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001640, Sample Num: 26240, Cur Loss: 1.70844865, Cur Avg Loss: 1.16918922, Log Avg loss: 1.69763086, Global Avg Loss: 1.50492533, Time: 0.0402 Steps: 86800, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001650, Sample Num: 26400, Cur Loss: 0.89760852, Cur Avg Loss: 1.17016370, Log Avg loss: 1.32997881, Global Avg Loss: 1.50490518, Time: 0.0402 Steps: 86810, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001660, Sample Num: 26560, Cur Loss: 1.54401457, Cur Avg Loss: 1.16977206, Log Avg loss: 1.10515016, Global Avg Loss: 1.50485913, Time: 0.0402 Steps: 86820, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001670, Sample Num: 26720, Cur Loss: 0.82765681, Cur Avg Loss: 1.16999181, Log Avg loss: 1.20647146, Global Avg Loss: 1.50482477, Time: 0.0404 Steps: 86830, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001680, Sample Num: 26880, Cur Loss: 0.70782620, Cur Avg Loss: 1.16967821, Log Avg loss: 1.11730681, Global Avg Loss: 1.50478014, Time: 0.0404 Steps: 86840, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001690, Sample Num: 27040, Cur Loss: 0.75786507, Cur Avg Loss: 1.16923468, Log Avg loss: 1.09472203, Global Avg Loss: 1.50473293, Time: 0.0490 Steps: 86850, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001700, Sample Num: 27200, Cur Loss: 0.91311038, Cur Avg Loss: 1.17142082, Log Avg loss: 1.54087799, Global Avg Loss: 1.50473709, Time: 0.0618 Steps: 86860, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001710, Sample Num: 27360, Cur Loss: 2.18607473, Cur Avg Loss: 1.17204883, Log Avg loss: 1.27880950, Global Avg Loss: 1.50471108, Time: 0.0405 Steps: 86870, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001720, Sample Num: 27520, Cur Loss: 0.29042453, Cur Avg Loss: 1.17133361, Log Avg loss: 1.04903233, Global Avg Loss: 1.50465863, Time: 0.0803 Steps: 86880, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001730, Sample Num: 27680, Cur Loss: 0.55748522, Cur Avg Loss: 1.17184479, Log Avg loss: 1.25976686, Global Avg Loss: 1.50463045, Time: 0.0805 Steps: 86890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001740, Sample Num: 27840, Cur Loss: 1.79750788, Cur Avg Loss: 1.17080859, Log Avg loss: 0.99154610, Global Avg Loss: 1.50457141, Time: 0.0585 Steps: 86900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001750, Sample Num: 28000, Cur Loss: 0.65306133, Cur Avg Loss: 1.17149103, Log Avg loss: 1.29023641, Global Avg Loss: 1.50454674, Time: 0.0406 Steps: 86910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001760, Sample Num: 28160, Cur Loss: 1.11112702, Cur Avg Loss: 1.17086623, Log Avg loss: 1.06152550, Global Avg Loss: 1.50449578, Time: 0.0403 Steps: 86920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001770, Sample Num: 28320, Cur Loss: 2.16616964, Cur Avg Loss: 1.17155893, Log Avg loss: 1.29347403, Global Avg Loss: 1.50447150, Time: 0.0402 Steps: 86930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001780, Sample Num: 28480, Cur Loss: 0.81662238, Cur Avg Loss: 1.17187385, Log Avg loss: 1.22761417, Global Avg Loss: 1.50443966, Time: 0.0403 Steps: 86940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001790, Sample Num: 28640, Cur Loss: 2.03632045, Cur Avg Loss: 1.17193043, Log Avg loss: 1.18200253, Global Avg Loss: 1.50440257, Time: 0.0402 Steps: 86950, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001800, Sample Num: 28800, Cur Loss: 1.43855655, Cur Avg Loss: 1.17257521, Log Avg loss: 1.28799003, Global Avg Loss: 1.50437769, Time: 0.0403 Steps: 86960, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001810, Sample Num: 28960, Cur Loss: 0.94133604, Cur Avg Loss: 1.17298124, Log Avg loss: 1.24606724, Global Avg Loss: 1.50434799, Time: 0.0403 Steps: 86970, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001820, Sample Num: 29120, Cur Loss: 2.35826921, Cur Avg Loss: 1.17187819, Log Avg loss: 0.97222715, Global Avg Loss: 1.50428681, Time: 0.0403 Steps: 86980, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001830, Sample Num: 29280, Cur Loss: 0.46777534, Cur Avg Loss: 1.17128923, Log Avg loss: 1.06409716, Global Avg Loss: 1.50423621, Time: 0.0402 Steps: 86990, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001840, Sample Num: 29440, Cur Loss: 1.41778910, Cur Avg Loss: 1.17157330, Log Avg loss: 1.22355784, Global Avg Loss: 1.50420394, Time: 0.0402 Steps: 87000, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001850, Sample Num: 29600, Cur Loss: 1.14486194, Cur Avg Loss: 1.17055189, Log Avg loss: 0.98261339, Global Avg Loss: 1.50414400, Time: 0.0402 Steps: 87010, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001860, Sample Num: 29760, Cur Loss: 1.31344509, Cur Avg Loss: 1.17185313, Log Avg loss: 1.41258200, Global Avg Loss: 1.50413348, Time: 0.0402 Steps: 87020, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001870, Sample Num: 29920, Cur Loss: 0.92876565, Cur Avg Loss: 1.17193270, Log Avg loss: 1.18673396, Global Avg Loss: 1.50409701, Time: 0.0403 Steps: 87030, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001880, Sample Num: 30080, Cur Loss: 0.62259352, Cur Avg Loss: 1.17147543, Log Avg loss: 1.08596506, Global Avg Loss: 1.50404897, Time: 0.0402 Steps: 87040, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001890, Sample Num: 30240, Cur Loss: 0.54902142, Cur Avg Loss: 1.17088418, Log Avg loss: 1.05972957, Global Avg Loss: 1.50399792, Time: 0.0402 Steps: 87050, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001900, Sample Num: 30400, Cur Loss: 1.19480717, Cur Avg Loss: 1.17202258, Log Avg loss: 1.38717940, Global Avg Loss: 1.50398451, Time: 0.0402 Steps: 87060, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001910, Sample Num: 30560, Cur Loss: 2.67538190, Cur Avg Loss: 1.17191875, Log Avg loss: 1.15219103, Global Avg Loss: 1.50394410, Time: 0.0402 Steps: 87070, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001920, Sample Num: 30720, Cur Loss: 0.43914792, Cur Avg Loss: 1.17270688, Log Avg loss: 1.32324100, Global Avg Loss: 1.50392335, Time: 0.0402 Steps: 87080, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001930, Sample Num: 30880, Cur Loss: 0.85303038, Cur Avg Loss: 1.17359717, Log Avg loss: 1.34453190, Global Avg Loss: 1.50390505, Time: 0.0402 Steps: 87090, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001940, Sample Num: 31040, Cur Loss: 1.13879156, Cur Avg Loss: 1.17399666, Log Avg loss: 1.25109862, Global Avg Loss: 1.50387602, Time: 0.0402 Steps: 87100, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001950, Sample Num: 31200, Cur Loss: 2.25038338, Cur Avg Loss: 1.17426645, Log Avg loss: 1.22660522, Global Avg Loss: 1.50384419, Time: 0.0402 Steps: 87110, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001960, Sample Num: 31360, Cur Loss: 2.38748407, Cur Avg Loss: 1.17455515, Log Avg loss: 1.23085181, Global Avg Loss: 1.50381286, Time: 0.0402 Steps: 87120, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001970, Sample Num: 31520, Cur Loss: 1.48684263, Cur Avg Loss: 1.17454699, Log Avg loss: 1.17294830, Global Avg Loss: 1.50377489, Time: 0.0402 Steps: 87130, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001980, Sample Num: 31680, Cur Loss: 0.64452076, Cur Avg Loss: 1.17470847, Log Avg loss: 1.20651872, Global Avg Loss: 1.50374077, Time: 0.0402 Steps: 87140, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001990, Sample Num: 31840, Cur Loss: 0.99247932, Cur Avg Loss: 1.17545575, Log Avg loss: 1.32341735, Global Avg Loss: 1.50372008, Time: 0.0402 Steps: 87150, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002000, Sample Num: 32000, Cur Loss: 1.72149146, Cur Avg Loss: 1.17488626, Log Avg loss: 1.06155898, Global Avg Loss: 1.50366935, Time: 0.0403 Steps: 87160, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002010, Sample Num: 32160, Cur Loss: 1.09183037, Cur Avg Loss: 1.17406550, Log Avg loss: 1.00991362, Global Avg Loss: 1.50361271, Time: 0.0402 Steps: 87170, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002020, Sample Num: 32320, Cur Loss: 1.31053162, Cur Avg Loss: 1.17486133, Log Avg loss: 1.33482289, Global Avg Loss: 1.50359335, Time: 0.0427 Steps: 87180, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002030, Sample Num: 32480, Cur Loss: 0.83682203, Cur Avg Loss: 1.17589547, Log Avg loss: 1.38479105, Global Avg Loss: 1.50357972, Time: 0.0743 Steps: 87190, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002040, Sample Num: 32640, Cur Loss: 0.59864080, Cur Avg Loss: 1.17538998, Log Avg loss: 1.07277614, Global Avg Loss: 1.50353032, Time: 0.0739 Steps: 87200, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002050, Sample Num: 32800, Cur Loss: 0.36378658, Cur Avg Loss: 1.17457103, Log Avg loss: 1.00750439, Global Avg Loss: 1.50347344, Time: 0.0661 Steps: 87210, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002060, Sample Num: 32960, Cur Loss: 0.48935756, Cur Avg Loss: 1.17544592, Log Avg loss: 1.35479938, Global Avg Loss: 1.50345640, Time: 0.0404 Steps: 87220, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002070, Sample Num: 33120, Cur Loss: 1.05153513, Cur Avg Loss: 1.17599274, Log Avg loss: 1.28863795, Global Avg Loss: 1.50343177, Time: 0.0659 Steps: 87230, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002080, Sample Num: 33280, Cur Loss: 0.87601823, Cur Avg Loss: 1.17609884, Log Avg loss: 1.19806148, Global Avg Loss: 1.50339677, Time: 0.0826 Steps: 87240, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002090, Sample Num: 33440, Cur Loss: 2.31501460, Cur Avg Loss: 1.17631366, Log Avg loss: 1.22099612, Global Avg Loss: 1.50336440, Time: 0.0406 Steps: 87250, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002100, Sample Num: 33600, Cur Loss: 0.89300472, Cur Avg Loss: 1.17549426, Log Avg loss: 1.00423977, Global Avg Loss: 1.50330720, Time: 0.0504 Steps: 87260, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002110, Sample Num: 33760, Cur Loss: 1.48985052, Cur Avg Loss: 1.17505406, Log Avg loss: 1.08261171, Global Avg Loss: 1.50325899, Time: 0.0402 Steps: 87270, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002120, Sample Num: 33920, Cur Loss: 0.98736417, Cur Avg Loss: 1.17508404, Log Avg loss: 1.18141000, Global Avg Loss: 1.50322212, Time: 0.0402 Steps: 87280, Updated lr: 0.000018 ***** Running evaluation checkpoint-87289 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-87289 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.188094, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.07608, "eval_total_loss": 756.484308, "eval_mae": 0.792397, "eval_mse": 1.076307, "eval_r2": 0.315829, "eval_sp_statistic": 0.453845, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.562033, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.046499, "test_total_loss": 525.342378, "test_mae": 0.819674, "test_mse": 1.046709, "test_r2": 0.324445, "test_sp_statistic": 0.478865, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.604591, "test_ps_pvalue": 0.0, "lr": 1.8170697012802277e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.503141666282119, "train_cur_epoch_loss": 2497.6846635639668, "train_cur_epoch_avg_loss": 1.173172693078425, "train_cur_epoch_time": 95.1880943775177, "train_cur_epoch_avg_time": 0.04471023690818116, "epoch": 41, "step": 87289} ################################################## Training, Epoch: 0042, Batch: 000001, Sample Num: 16, Cur Loss: 0.61215401, Cur Avg Loss: 0.61215401, Log Avg loss: 0.71186448, Global Avg Loss: 1.50313146, Time: 0.0628 Steps: 87290, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000011, Sample Num: 176, Cur Loss: 1.43113732, Cur Avg Loss: 0.99054467, Log Avg loss: 1.02838373, Global Avg Loss: 1.50307708, Time: 0.0561 Steps: 87300, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000021, Sample Num: 336, Cur Loss: 1.66172671, Cur Avg Loss: 1.15291441, Log Avg loss: 1.33152113, Global Avg Loss: 1.50305743, Time: 0.0402 Steps: 87310, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000031, Sample Num: 496, Cur Loss: 1.73150063, Cur Avg Loss: 1.18728032, Log Avg loss: 1.25944872, Global Avg Loss: 1.50302953, Time: 0.0402 Steps: 87320, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000041, Sample Num: 656, Cur Loss: 1.93794942, Cur Avg Loss: 1.16203657, Log Avg loss: 1.08378094, Global Avg Loss: 1.50298152, Time: 0.0402 Steps: 87330, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000051, Sample Num: 816, Cur Loss: 1.43044448, Cur Avg Loss: 1.14980014, Log Avg loss: 1.09963080, Global Avg Loss: 1.50293534, Time: 0.0402 Steps: 87340, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000061, Sample Num: 976, Cur Loss: 0.54543316, Cur Avg Loss: 1.11900951, Log Avg loss: 0.96197731, Global Avg Loss: 1.50287341, Time: 0.0402 Steps: 87350, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000071, Sample Num: 1136, Cur Loss: 1.97906160, Cur Avg Loss: 1.15253945, Log Avg loss: 1.35707205, Global Avg Loss: 1.50285672, Time: 0.0402 Steps: 87360, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000081, Sample Num: 1296, Cur Loss: 1.22115076, Cur Avg Loss: 1.12376619, Log Avg loss: 0.91947605, Global Avg Loss: 1.50278995, Time: 0.0403 Steps: 87370, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000091, Sample Num: 1456, Cur Loss: 1.25786531, Cur Avg Loss: 1.12617959, Log Avg loss: 1.14572816, Global Avg Loss: 1.50274909, Time: 0.0402 Steps: 87380, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000101, Sample Num: 1616, Cur Loss: 1.05908215, Cur Avg Loss: 1.13775172, Log Avg loss: 1.24305808, Global Avg Loss: 1.50271937, Time: 0.0402 Steps: 87390, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000111, Sample Num: 1776, Cur Loss: 0.79524457, Cur Avg Loss: 1.15538830, Log Avg loss: 1.33351774, Global Avg Loss: 1.50270001, Time: 0.0402 Steps: 87400, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000121, Sample Num: 1936, Cur Loss: 0.79001862, Cur Avg Loss: 1.15312761, Log Avg loss: 1.12803396, Global Avg Loss: 1.50265715, Time: 0.0402 Steps: 87410, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000131, Sample Num: 2096, Cur Loss: 0.72136664, Cur Avg Loss: 1.16465109, Log Avg loss: 1.30408521, Global Avg Loss: 1.50263443, Time: 0.0403 Steps: 87420, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000141, Sample Num: 2256, Cur Loss: 1.38629413, Cur Avg Loss: 1.14357710, Log Avg loss: 0.86750782, Global Avg Loss: 1.50256179, Time: 0.0402 Steps: 87430, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000151, Sample Num: 2416, Cur Loss: 1.16480899, Cur Avg Loss: 1.14802086, Log Avg loss: 1.21067792, Global Avg Loss: 1.50252841, Time: 0.0402 Steps: 87440, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000161, Sample Num: 2576, Cur Loss: 1.18993592, Cur Avg Loss: 1.16732532, Log Avg loss: 1.45882268, Global Avg Loss: 1.50252341, Time: 0.0402 Steps: 87450, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000171, Sample Num: 2736, Cur Loss: 1.13778615, Cur Avg Loss: 1.16897876, Log Avg loss: 1.19559914, Global Avg Loss: 1.50248832, Time: 0.0402 Steps: 87460, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000181, Sample Num: 2896, Cur Loss: 1.63336968, Cur Avg Loss: 1.15985858, Log Avg loss: 1.00390341, Global Avg Loss: 1.50243132, Time: 0.0402 Steps: 87470, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000191, Sample Num: 3056, Cur Loss: 1.12791300, Cur Avg Loss: 1.16082277, Log Avg loss: 1.17827455, Global Avg Loss: 1.50239426, Time: 0.0402 Steps: 87480, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000201, Sample Num: 3216, Cur Loss: 1.50797784, Cur Avg Loss: 1.16501852, Log Avg loss: 1.24515753, Global Avg Loss: 1.50236486, Time: 0.0402 Steps: 87490, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000211, Sample Num: 3376, Cur Loss: 1.29474795, Cur Avg Loss: 1.16538022, Log Avg loss: 1.17265039, Global Avg Loss: 1.50232718, Time: 0.0402 Steps: 87500, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000221, Sample Num: 3536, Cur Loss: 1.22436213, Cur Avg Loss: 1.15498553, Log Avg loss: 0.93565753, Global Avg Loss: 1.50226242, Time: 0.0402 Steps: 87510, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000231, Sample Num: 3696, Cur Loss: 0.97599477, Cur Avg Loss: 1.15602553, Log Avg loss: 1.17900940, Global Avg Loss: 1.50222549, Time: 0.0402 Steps: 87520, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000241, Sample Num: 3856, Cur Loss: 2.09839964, Cur Avg Loss: 1.15532940, Log Avg loss: 1.13924887, Global Avg Loss: 1.50218402, Time: 0.0402 Steps: 87530, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000251, Sample Num: 4016, Cur Loss: 2.22927570, Cur Avg Loss: 1.15741653, Log Avg loss: 1.20771641, Global Avg Loss: 1.50215038, Time: 0.0402 Steps: 87540, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000261, Sample Num: 4176, Cur Loss: 0.64070332, Cur Avg Loss: 1.15858034, Log Avg loss: 1.18779206, Global Avg Loss: 1.50211448, Time: 0.0402 Steps: 87550, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000271, Sample Num: 4336, Cur Loss: 1.22955525, Cur Avg Loss: 1.16173039, Log Avg loss: 1.24394646, Global Avg Loss: 1.50208499, Time: 0.0402 Steps: 87560, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000281, Sample Num: 4496, Cur Loss: 0.83729345, Cur Avg Loss: 1.16111032, Log Avg loss: 1.14430654, Global Avg Loss: 1.50204414, Time: 0.0507 Steps: 87570, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000291, Sample Num: 4656, Cur Loss: 1.13146639, Cur Avg Loss: 1.16447264, Log Avg loss: 1.25895381, Global Avg Loss: 1.50201638, Time: 0.0965 Steps: 87580, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000301, Sample Num: 4816, Cur Loss: 1.08830714, Cur Avg Loss: 1.15982151, Log Avg loss: 1.02447360, Global Avg Loss: 1.50196186, Time: 0.0720 Steps: 87590, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000311, Sample Num: 4976, Cur Loss: 1.05310416, Cur Avg Loss: 1.16179013, Log Avg loss: 1.22104571, Global Avg Loss: 1.50192979, Time: 0.0405 Steps: 87600, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000321, Sample Num: 5136, Cur Loss: 0.52034765, Cur Avg Loss: 1.16157160, Log Avg loss: 1.15477523, Global Avg Loss: 1.50189017, Time: 0.0807 Steps: 87610, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000331, Sample Num: 5296, Cur Loss: 0.40678957, Cur Avg Loss: 1.16370941, Log Avg loss: 1.23233309, Global Avg Loss: 1.50185940, Time: 0.0629 Steps: 87620, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000341, Sample Num: 5456, Cur Loss: 1.71644378, Cur Avg Loss: 1.16536646, Log Avg loss: 1.22021477, Global Avg Loss: 1.50182726, Time: 0.0457 Steps: 87630, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000351, Sample Num: 5616, Cur Loss: 1.18478894, Cur Avg Loss: 1.16624447, Log Avg loss: 1.19618480, Global Avg Loss: 1.50179239, Time: 0.0577 Steps: 87640, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000361, Sample Num: 5776, Cur Loss: 2.69059467, Cur Avg Loss: 1.16804752, Log Avg loss: 1.23133430, Global Avg Loss: 1.50176153, Time: 0.0562 Steps: 87650, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000371, Sample Num: 5936, Cur Loss: 1.19784832, Cur Avg Loss: 1.16982707, Log Avg loss: 1.23406917, Global Avg Loss: 1.50173099, Time: 0.0404 Steps: 87660, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000381, Sample Num: 6096, Cur Loss: 1.95829201, Cur Avg Loss: 1.17556236, Log Avg loss: 1.38834151, Global Avg Loss: 1.50171806, Time: 0.0402 Steps: 87670, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000391, Sample Num: 6256, Cur Loss: 2.13740563, Cur Avg Loss: 1.17350892, Log Avg loss: 1.09527268, Global Avg Loss: 1.50167170, Time: 0.0402 Steps: 87680, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000401, Sample Num: 6416, Cur Loss: 1.28735232, Cur Avg Loss: 1.17731625, Log Avg loss: 1.32618281, Global Avg Loss: 1.50165169, Time: 0.0402 Steps: 87690, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000411, Sample Num: 6576, Cur Loss: 0.26999497, Cur Avg Loss: 1.17483358, Log Avg loss: 1.07527855, Global Avg Loss: 1.50160307, Time: 0.0402 Steps: 87700, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000421, Sample Num: 6736, Cur Loss: 0.33024523, Cur Avg Loss: 1.16586293, Log Avg loss: 0.79716945, Global Avg Loss: 1.50152276, Time: 0.0403 Steps: 87710, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000431, Sample Num: 6896, Cur Loss: 1.06721950, Cur Avg Loss: 1.15986946, Log Avg loss: 0.90754418, Global Avg Loss: 1.50145505, Time: 0.0402 Steps: 87720, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000441, Sample Num: 7056, Cur Loss: 0.44331187, Cur Avg Loss: 1.15643858, Log Avg loss: 1.00856794, Global Avg Loss: 1.50139886, Time: 0.0402 Steps: 87730, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000451, Sample Num: 7216, Cur Loss: 0.88601935, Cur Avg Loss: 1.15769811, Log Avg loss: 1.21324295, Global Avg Loss: 1.50136602, Time: 0.0403 Steps: 87740, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000461, Sample Num: 7376, Cur Loss: 0.56546396, Cur Avg Loss: 1.16059021, Log Avg loss: 1.29102398, Global Avg Loss: 1.50134205, Time: 0.0403 Steps: 87750, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000471, Sample Num: 7536, Cur Loss: 0.56423736, Cur Avg Loss: 1.15737681, Log Avg loss: 1.00923918, Global Avg Loss: 1.50128598, Time: 0.0402 Steps: 87760, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000481, Sample Num: 7696, Cur Loss: 1.70956576, Cur Avg Loss: 1.16535267, Log Avg loss: 1.54101594, Global Avg Loss: 1.50129050, Time: 0.0402 Steps: 87770, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000491, Sample Num: 7856, Cur Loss: 0.83867198, Cur Avg Loss: 1.16087856, Log Avg loss: 0.94567358, Global Avg Loss: 1.50122721, Time: 0.0402 Steps: 87780, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000501, Sample Num: 8016, Cur Loss: 2.14217639, Cur Avg Loss: 1.16793132, Log Avg loss: 1.51422192, Global Avg Loss: 1.50122869, Time: 0.0402 Steps: 87790, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000511, Sample Num: 8176, Cur Loss: 1.31246841, Cur Avg Loss: 1.16712656, Log Avg loss: 1.12680785, Global Avg Loss: 1.50118604, Time: 0.0403 Steps: 87800, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000521, Sample Num: 8336, Cur Loss: 1.57057452, Cur Avg Loss: 1.16606749, Log Avg loss: 1.11194899, Global Avg Loss: 1.50114172, Time: 0.0403 Steps: 87810, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000531, Sample Num: 8496, Cur Loss: 1.06330371, Cur Avg Loss: 1.16118093, Log Avg loss: 0.90659122, Global Avg Loss: 1.50107401, Time: 0.0403 Steps: 87820, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000541, Sample Num: 8656, Cur Loss: 1.46840167, Cur Avg Loss: 1.16765607, Log Avg loss: 1.51148614, Global Avg Loss: 1.50107520, Time: 0.0402 Steps: 87830, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000551, Sample Num: 8816, Cur Loss: 1.95006764, Cur Avg Loss: 1.16838614, Log Avg loss: 1.20788282, Global Avg Loss: 1.50104182, Time: 0.0403 Steps: 87840, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000561, Sample Num: 8976, Cur Loss: 1.24657404, Cur Avg Loss: 1.16462263, Log Avg loss: 0.95725346, Global Avg Loss: 1.50097992, Time: 0.0402 Steps: 87850, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000571, Sample Num: 9136, Cur Loss: 1.04438877, Cur Avg Loss: 1.16136431, Log Avg loss: 0.97857258, Global Avg Loss: 1.50092046, Time: 0.0402 Steps: 87860, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000581, Sample Num: 9296, Cur Loss: 1.61227262, Cur Avg Loss: 1.15784882, Log Avg loss: 0.95711440, Global Avg Loss: 1.50085858, Time: 0.0402 Steps: 87870, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000591, Sample Num: 9456, Cur Loss: 0.83791775, Cur Avg Loss: 1.15402658, Log Avg loss: 0.93195406, Global Avg Loss: 1.50079384, Time: 0.0402 Steps: 87880, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000601, Sample Num: 9616, Cur Loss: 1.96478951, Cur Avg Loss: 1.15512405, Log Avg loss: 1.21998461, Global Avg Loss: 1.50076189, Time: 0.0402 Steps: 87890, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000611, Sample Num: 9776, Cur Loss: 0.44393280, Cur Avg Loss: 1.15702926, Log Avg loss: 1.27153263, Global Avg Loss: 1.50073581, Time: 0.0402 Steps: 87900, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000621, Sample Num: 9936, Cur Loss: 2.07349372, Cur Avg Loss: 1.16059040, Log Avg loss: 1.37817614, Global Avg Loss: 1.50072187, Time: 0.0403 Steps: 87910, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000631, Sample Num: 10096, Cur Loss: 1.33385849, Cur Avg Loss: 1.16220922, Log Avg loss: 1.26273791, Global Avg Loss: 1.50069480, Time: 0.0409 Steps: 87920, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000641, Sample Num: 10256, Cur Loss: 1.06522322, Cur Avg Loss: 1.16367183, Log Avg loss: 1.25596267, Global Avg Loss: 1.50066697, Time: 0.0405 Steps: 87930, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000651, Sample Num: 10416, Cur Loss: 0.65949571, Cur Avg Loss: 1.16204076, Log Avg loss: 1.05748915, Global Avg Loss: 1.50061657, Time: 0.0535 Steps: 87940, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000661, Sample Num: 10576, Cur Loss: 1.03126335, Cur Avg Loss: 1.16055095, Log Avg loss: 1.06356393, Global Avg Loss: 1.50056688, Time: 0.0659 Steps: 87950, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000671, Sample Num: 10736, Cur Loss: 1.28396654, Cur Avg Loss: 1.15765465, Log Avg loss: 0.96620960, Global Avg Loss: 1.50050613, Time: 0.0819 Steps: 87960, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000681, Sample Num: 10896, Cur Loss: 0.38545835, Cur Avg Loss: 1.15633244, Log Avg loss: 1.06761206, Global Avg Loss: 1.50045692, Time: 0.0585 Steps: 87970, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000691, Sample Num: 11056, Cur Loss: 0.59161758, Cur Avg Loss: 1.15257409, Log Avg loss: 0.89663053, Global Avg Loss: 1.50038829, Time: 0.0470 Steps: 87980, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000701, Sample Num: 11216, Cur Loss: 0.41239858, Cur Avg Loss: 1.14881752, Log Avg loss: 0.88923818, Global Avg Loss: 1.50031883, Time: 0.0509 Steps: 87990, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000711, Sample Num: 11376, Cur Loss: 1.92946172, Cur Avg Loss: 1.14838555, Log Avg loss: 1.11810484, Global Avg Loss: 1.50027540, Time: 0.0402 Steps: 88000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000721, Sample Num: 11536, Cur Loss: 1.44066334, Cur Avg Loss: 1.14714622, Log Avg loss: 1.05902928, Global Avg Loss: 1.50022526, Time: 0.0402 Steps: 88010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000731, Sample Num: 11696, Cur Loss: 0.47593498, Cur Avg Loss: 1.14810570, Log Avg loss: 1.21728456, Global Avg Loss: 1.50019312, Time: 0.0402 Steps: 88020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000741, Sample Num: 11856, Cur Loss: 1.61345184, Cur Avg Loss: 1.14806205, Log Avg loss: 1.14487133, Global Avg Loss: 1.50015275, Time: 0.0402 Steps: 88030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000751, Sample Num: 12016, Cur Loss: 1.62339830, Cur Avg Loss: 1.14653538, Log Avg loss: 1.03340919, Global Avg Loss: 1.50009974, Time: 0.0402 Steps: 88040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000761, Sample Num: 12176, Cur Loss: 0.95169574, Cur Avg Loss: 1.14511110, Log Avg loss: 1.03814715, Global Avg Loss: 1.50004727, Time: 0.0403 Steps: 88050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000771, Sample Num: 12336, Cur Loss: 1.17102432, Cur Avg Loss: 1.14589304, Log Avg loss: 1.20539883, Global Avg Loss: 1.50001381, Time: 0.0402 Steps: 88060, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000781, Sample Num: 12496, Cur Loss: 0.53104019, Cur Avg Loss: 1.14423599, Log Avg loss: 1.01647785, Global Avg Loss: 1.49995891, Time: 0.0402 Steps: 88070, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000791, Sample Num: 12656, Cur Loss: 2.02039075, Cur Avg Loss: 1.14125350, Log Avg loss: 0.90832098, Global Avg Loss: 1.49989174, Time: 0.0402 Steps: 88080, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000801, Sample Num: 12816, Cur Loss: 1.07796311, Cur Avg Loss: 1.13931416, Log Avg loss: 0.98591224, Global Avg Loss: 1.49983339, Time: 0.0403 Steps: 88090, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000811, Sample Num: 12976, Cur Loss: 0.86588264, Cur Avg Loss: 1.13724830, Log Avg loss: 0.97177318, Global Avg Loss: 1.49977345, Time: 0.0403 Steps: 88100, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000821, Sample Num: 13136, Cur Loss: 2.09479141, Cur Avg Loss: 1.13865979, Log Avg loss: 1.25313145, Global Avg Loss: 1.49974546, Time: 0.0403 Steps: 88110, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000831, Sample Num: 13296, Cur Loss: 0.91764224, Cur Avg Loss: 1.13804736, Log Avg loss: 1.08776693, Global Avg Loss: 1.49969871, Time: 0.0402 Steps: 88120, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000841, Sample Num: 13456, Cur Loss: 1.67417932, Cur Avg Loss: 1.14204092, Log Avg loss: 1.47390591, Global Avg Loss: 1.49969578, Time: 0.0402 Steps: 88130, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000851, Sample Num: 13616, Cur Loss: 0.38982326, Cur Avg Loss: 1.14100213, Log Avg loss: 1.05363989, Global Avg Loss: 1.49964518, Time: 0.0402 Steps: 88140, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000861, Sample Num: 13776, Cur Loss: 0.60013634, Cur Avg Loss: 1.13988615, Log Avg loss: 1.04491576, Global Avg Loss: 1.49959359, Time: 0.0402 Steps: 88150, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000871, Sample Num: 13936, Cur Loss: 1.02348733, Cur Avg Loss: 1.14181405, Log Avg loss: 1.30780601, Global Avg Loss: 1.49957183, Time: 0.0403 Steps: 88160, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000881, Sample Num: 14096, Cur Loss: 2.36642122, Cur Avg Loss: 1.14374640, Log Avg loss: 1.31205418, Global Avg Loss: 1.49955057, Time: 0.0402 Steps: 88170, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000891, Sample Num: 14256, Cur Loss: 0.80299151, Cur Avg Loss: 1.14462577, Log Avg loss: 1.22209854, Global Avg Loss: 1.49951910, Time: 0.0402 Steps: 88180, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000901, Sample Num: 14416, Cur Loss: 0.69855201, Cur Avg Loss: 1.14453383, Log Avg loss: 1.13634223, Global Avg Loss: 1.49947792, Time: 0.0402 Steps: 88190, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000911, Sample Num: 14576, Cur Loss: 1.07714260, Cur Avg Loss: 1.14454649, Log Avg loss: 1.14568660, Global Avg Loss: 1.49943781, Time: 0.0402 Steps: 88200, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000921, Sample Num: 14736, Cur Loss: 0.38786575, Cur Avg Loss: 1.14408663, Log Avg loss: 1.10219320, Global Avg Loss: 1.49939278, Time: 0.0402 Steps: 88210, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000931, Sample Num: 14896, Cur Loss: 1.10289907, Cur Avg Loss: 1.14376230, Log Avg loss: 1.11389222, Global Avg Loss: 1.49934908, Time: 0.0402 Steps: 88220, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000941, Sample Num: 15056, Cur Loss: 0.30774420, Cur Avg Loss: 1.13997418, Log Avg loss: 0.78729979, Global Avg Loss: 1.49926837, Time: 0.0402 Steps: 88230, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000951, Sample Num: 15216, Cur Loss: 0.53505903, Cur Avg Loss: 1.14415372, Log Avg loss: 1.53744892, Global Avg Loss: 1.49927270, Time: 0.0402 Steps: 88240, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000961, Sample Num: 15376, Cur Loss: 1.66131008, Cur Avg Loss: 1.14399073, Log Avg loss: 1.12848986, Global Avg Loss: 1.49923069, Time: 0.0875 Steps: 88250, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000971, Sample Num: 15536, Cur Loss: 1.62381160, Cur Avg Loss: 1.14659715, Log Avg loss: 1.39707399, Global Avg Loss: 1.49921911, Time: 0.0613 Steps: 88260, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000981, Sample Num: 15696, Cur Loss: 1.20879519, Cur Avg Loss: 1.14824393, Log Avg loss: 1.30814663, Global Avg Loss: 1.49919746, Time: 0.0664 Steps: 88270, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000991, Sample Num: 15856, Cur Loss: 0.86193228, Cur Avg Loss: 1.14771473, Log Avg loss: 1.09579993, Global Avg Loss: 1.49915177, Time: 0.0521 Steps: 88280, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001001, Sample Num: 16016, Cur Loss: 0.65431440, Cur Avg Loss: 1.14792951, Log Avg loss: 1.16921462, Global Avg Loss: 1.49911440, Time: 0.0719 Steps: 88290, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001011, Sample Num: 16176, Cur Loss: 1.54606271, Cur Avg Loss: 1.14953874, Log Avg loss: 1.31062231, Global Avg Loss: 1.49909305, Time: 0.0690 Steps: 88300, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001021, Sample Num: 16336, Cur Loss: 0.58576936, Cur Avg Loss: 1.14781878, Log Avg loss: 0.97393136, Global Avg Loss: 1.49903358, Time: 0.0480 Steps: 88310, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001031, Sample Num: 16496, Cur Loss: 0.42200863, Cur Avg Loss: 1.14780556, Log Avg loss: 1.14645552, Global Avg Loss: 1.49899366, Time: 0.0871 Steps: 88320, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001041, Sample Num: 16656, Cur Loss: 0.85187197, Cur Avg Loss: 1.15001998, Log Avg loss: 1.37832696, Global Avg Loss: 1.49898000, Time: 0.0402 Steps: 88330, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001051, Sample Num: 16816, Cur Loss: 0.82032597, Cur Avg Loss: 1.15064380, Log Avg loss: 1.21558286, Global Avg Loss: 1.49894792, Time: 0.0402 Steps: 88340, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001061, Sample Num: 16976, Cur Loss: 1.16400230, Cur Avg Loss: 1.15011143, Log Avg loss: 1.09415973, Global Avg Loss: 1.49890211, Time: 0.0403 Steps: 88350, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001071, Sample Num: 17136, Cur Loss: 1.93817770, Cur Avg Loss: 1.15035210, Log Avg loss: 1.17588727, Global Avg Loss: 1.49886555, Time: 0.0403 Steps: 88360, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001081, Sample Num: 17296, Cur Loss: 0.43818080, Cur Avg Loss: 1.14822936, Log Avg loss: 0.92088376, Global Avg Loss: 1.49880015, Time: 0.0403 Steps: 88370, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001091, Sample Num: 17456, Cur Loss: 2.21280193, Cur Avg Loss: 1.14767228, Log Avg loss: 1.08745144, Global Avg Loss: 1.49875360, Time: 0.0402 Steps: 88380, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001101, Sample Num: 17616, Cur Loss: 1.20193744, Cur Avg Loss: 1.14766618, Log Avg loss: 1.14700074, Global Avg Loss: 1.49871381, Time: 0.0403 Steps: 88390, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001111, Sample Num: 17776, Cur Loss: 1.36996067, Cur Avg Loss: 1.14730206, Log Avg loss: 1.10721252, Global Avg Loss: 1.49866952, Time: 0.0402 Steps: 88400, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001121, Sample Num: 17936, Cur Loss: 0.96785736, Cur Avg Loss: 1.14714692, Log Avg loss: 1.12991084, Global Avg Loss: 1.49862781, Time: 0.0402 Steps: 88410, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001131, Sample Num: 18096, Cur Loss: 0.92971623, Cur Avg Loss: 1.14553549, Log Avg loss: 0.96489399, Global Avg Loss: 1.49856745, Time: 0.0402 Steps: 88420, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001141, Sample Num: 18256, Cur Loss: 0.54795456, Cur Avg Loss: 1.14406041, Log Avg loss: 0.97722947, Global Avg Loss: 1.49850849, Time: 0.0402 Steps: 88430, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001151, Sample Num: 18416, Cur Loss: 0.56527889, Cur Avg Loss: 1.14473762, Log Avg loss: 1.22200688, Global Avg Loss: 1.49847723, Time: 0.0402 Steps: 88440, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001161, Sample Num: 18576, Cur Loss: 0.96564984, Cur Avg Loss: 1.14626476, Log Avg loss: 1.32203875, Global Avg Loss: 1.49845728, Time: 0.0402 Steps: 88450, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001171, Sample Num: 18736, Cur Loss: 1.32795191, Cur Avg Loss: 1.14759049, Log Avg loss: 1.30150742, Global Avg Loss: 1.49843501, Time: 0.0402 Steps: 88460, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001181, Sample Num: 18896, Cur Loss: 1.38411212, Cur Avg Loss: 1.15034108, Log Avg loss: 1.47243606, Global Avg Loss: 1.49843208, Time: 0.0402 Steps: 88470, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001191, Sample Num: 19056, Cur Loss: 0.73700356, Cur Avg Loss: 1.15041822, Log Avg loss: 1.15952796, Global Avg Loss: 1.49839377, Time: 0.0402 Steps: 88480, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001201, Sample Num: 19216, Cur Loss: 2.19529057, Cur Avg Loss: 1.15113751, Log Avg loss: 1.23680440, Global Avg Loss: 1.49836421, Time: 0.0403 Steps: 88490, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001211, Sample Num: 19376, Cur Loss: 2.37654853, Cur Avg Loss: 1.15049039, Log Avg loss: 1.07277243, Global Avg Loss: 1.49831612, Time: 0.0402 Steps: 88500, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001221, Sample Num: 19536, Cur Loss: 0.84019053, Cur Avg Loss: 1.14902453, Log Avg loss: 0.97150880, Global Avg Loss: 1.49825660, Time: 0.0402 Steps: 88510, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001231, Sample Num: 19696, Cur Loss: 1.03390670, Cur Avg Loss: 1.15168326, Log Avg loss: 1.47631395, Global Avg Loss: 1.49825412, Time: 0.0402 Steps: 88520, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001241, Sample Num: 19856, Cur Loss: 0.36962366, Cur Avg Loss: 1.15032330, Log Avg loss: 0.98291224, Global Avg Loss: 1.49819591, Time: 0.0402 Steps: 88530, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001251, Sample Num: 20016, Cur Loss: 1.59274662, Cur Avg Loss: 1.15028467, Log Avg loss: 1.14549047, Global Avg Loss: 1.49815608, Time: 0.0401 Steps: 88540, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001261, Sample Num: 20176, Cur Loss: 1.28320467, Cur Avg Loss: 1.14930547, Log Avg loss: 1.02680790, Global Avg Loss: 1.49810285, Time: 0.0402 Steps: 88550, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001271, Sample Num: 20336, Cur Loss: 0.49888167, Cur Avg Loss: 1.14796697, Log Avg loss: 0.97918227, Global Avg Loss: 1.49804425, Time: 0.0402 Steps: 88560, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001281, Sample Num: 20496, Cur Loss: 0.67168516, Cur Avg Loss: 1.14782712, Log Avg loss: 1.13005195, Global Avg Loss: 1.49800270, Time: 0.0402 Steps: 88570, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001291, Sample Num: 20656, Cur Loss: 1.36024296, Cur Avg Loss: 1.14757812, Log Avg loss: 1.11568050, Global Avg Loss: 1.49795954, Time: 0.0401 Steps: 88580, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001301, Sample Num: 20816, Cur Loss: 1.34257305, Cur Avg Loss: 1.14768540, Log Avg loss: 1.16153528, Global Avg Loss: 1.49792157, Time: 0.0485 Steps: 88590, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001311, Sample Num: 20976, Cur Loss: 0.55931848, Cur Avg Loss: 1.14870191, Log Avg loss: 1.28094999, Global Avg Loss: 1.49789708, Time: 0.0482 Steps: 88600, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001321, Sample Num: 21136, Cur Loss: 1.28036821, Cur Avg Loss: 1.14894557, Log Avg loss: 1.18088952, Global Avg Loss: 1.49786130, Time: 0.0541 Steps: 88610, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001331, Sample Num: 21296, Cur Loss: 1.69813812, Cur Avg Loss: 1.14946124, Log Avg loss: 1.21758156, Global Avg Loss: 1.49782968, Time: 0.0514 Steps: 88620, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001341, Sample Num: 21456, Cur Loss: 1.25686789, Cur Avg Loss: 1.14963603, Log Avg loss: 1.17290044, Global Avg Loss: 1.49779301, Time: 0.0789 Steps: 88630, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001351, Sample Num: 21616, Cur Loss: 2.60221386, Cur Avg Loss: 1.15024929, Log Avg loss: 1.23248798, Global Avg Loss: 1.49776308, Time: 0.0867 Steps: 88640, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001361, Sample Num: 21776, Cur Loss: 1.02634084, Cur Avg Loss: 1.15121070, Log Avg loss: 1.28109661, Global Avg Loss: 1.49773864, Time: 0.0634 Steps: 88650, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001371, Sample Num: 21936, Cur Loss: 1.23749423, Cur Avg Loss: 1.15114482, Log Avg loss: 1.14217908, Global Avg Loss: 1.49769854, Time: 0.0986 Steps: 88660, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001381, Sample Num: 22096, Cur Loss: 1.57713270, Cur Avg Loss: 1.15237741, Log Avg loss: 1.32136533, Global Avg Loss: 1.49767865, Time: 0.0664 Steps: 88670, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001391, Sample Num: 22256, Cur Loss: 0.70453185, Cur Avg Loss: 1.15231307, Log Avg loss: 1.14342673, Global Avg Loss: 1.49763871, Time: 0.0402 Steps: 88680, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001401, Sample Num: 22416, Cur Loss: 0.40947306, Cur Avg Loss: 1.15106674, Log Avg loss: 0.97770342, Global Avg Loss: 1.49758008, Time: 0.0402 Steps: 88690, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001411, Sample Num: 22576, Cur Loss: 1.30086684, Cur Avg Loss: 1.15333246, Log Avg loss: 1.47075887, Global Avg Loss: 1.49757706, Time: 0.0402 Steps: 88700, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001421, Sample Num: 22736, Cur Loss: 0.66909146, Cur Avg Loss: 1.15342141, Log Avg loss: 1.16597327, Global Avg Loss: 1.49753968, Time: 0.0402 Steps: 88710, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001431, Sample Num: 22896, Cur Loss: 0.36146575, Cur Avg Loss: 1.15134689, Log Avg loss: 0.85655734, Global Avg Loss: 1.49746743, Time: 0.0402 Steps: 88720, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001441, Sample Num: 23056, Cur Loss: 0.91441190, Cur Avg Loss: 1.15304059, Log Avg loss: 1.39540809, Global Avg Loss: 1.49745593, Time: 0.0403 Steps: 88730, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001451, Sample Num: 23216, Cur Loss: 0.54692650, Cur Avg Loss: 1.15307525, Log Avg loss: 1.15807088, Global Avg Loss: 1.49741768, Time: 0.0403 Steps: 88740, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001461, Sample Num: 23376, Cur Loss: 0.96945047, Cur Avg Loss: 1.15397229, Log Avg loss: 1.28413285, Global Avg Loss: 1.49739365, Time: 0.0402 Steps: 88750, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001471, Sample Num: 23536, Cur Loss: 1.07265425, Cur Avg Loss: 1.15427618, Log Avg loss: 1.19867422, Global Avg Loss: 1.49736000, Time: 0.0403 Steps: 88760, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001481, Sample Num: 23696, Cur Loss: 1.46772790, Cur Avg Loss: 1.15537869, Log Avg loss: 1.31755710, Global Avg Loss: 1.49733974, Time: 0.0402 Steps: 88770, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001491, Sample Num: 23856, Cur Loss: 1.26575184, Cur Avg Loss: 1.15500925, Log Avg loss: 1.10029559, Global Avg Loss: 1.49729502, Time: 0.0403 Steps: 88780, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001501, Sample Num: 24016, Cur Loss: 0.57791591, Cur Avg Loss: 1.15493234, Log Avg loss: 1.14346448, Global Avg Loss: 1.49725517, Time: 0.0402 Steps: 88790, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001511, Sample Num: 24176, Cur Loss: 0.89484304, Cur Avg Loss: 1.15508108, Log Avg loss: 1.17740803, Global Avg Loss: 1.49721915, Time: 0.0403 Steps: 88800, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001521, Sample Num: 24336, Cur Loss: 1.29839778, Cur Avg Loss: 1.15468838, Log Avg loss: 1.09535042, Global Avg Loss: 1.49717390, Time: 0.0403 Steps: 88810, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001531, Sample Num: 24496, Cur Loss: 0.72744358, Cur Avg Loss: 1.15413412, Log Avg loss: 1.06983169, Global Avg Loss: 1.49712579, Time: 0.0403 Steps: 88820, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001541, Sample Num: 24656, Cur Loss: 1.18399107, Cur Avg Loss: 1.15642672, Log Avg loss: 1.50742404, Global Avg Loss: 1.49712694, Time: 0.0402 Steps: 88830, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001551, Sample Num: 24816, Cur Loss: 0.59561956, Cur Avg Loss: 1.15560388, Log Avg loss: 1.02880432, Global Avg Loss: 1.49707423, Time: 0.0402 Steps: 88840, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001561, Sample Num: 24976, Cur Loss: 1.13130879, Cur Avg Loss: 1.15467461, Log Avg loss: 1.01054453, Global Avg Loss: 1.49701947, Time: 0.0403 Steps: 88850, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001571, Sample Num: 25136, Cur Loss: 0.82257259, Cur Avg Loss: 1.15529709, Log Avg loss: 1.25246645, Global Avg Loss: 1.49699195, Time: 0.0402 Steps: 88860, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001581, Sample Num: 25296, Cur Loss: 0.87999332, Cur Avg Loss: 1.15309058, Log Avg loss: 0.80644694, Global Avg Loss: 1.49691425, Time: 0.0402 Steps: 88870, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001591, Sample Num: 25456, Cur Loss: 1.37364650, Cur Avg Loss: 1.15177953, Log Avg loss: 0.94450274, Global Avg Loss: 1.49685209, Time: 0.0402 Steps: 88880, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001601, Sample Num: 25616, Cur Loss: 0.62198853, Cur Avg Loss: 1.15218968, Log Avg loss: 1.21744512, Global Avg Loss: 1.49682066, Time: 0.0402 Steps: 88890, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001611, Sample Num: 25776, Cur Loss: 0.74645972, Cur Avg Loss: 1.15132408, Log Avg loss: 1.01274136, Global Avg Loss: 1.49676621, Time: 0.0402 Steps: 88900, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001621, Sample Num: 25936, Cur Loss: 0.71046126, Cur Avg Loss: 1.15065619, Log Avg loss: 1.04305955, Global Avg Loss: 1.49671518, Time: 0.0402 Steps: 88910, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001631, Sample Num: 26096, Cur Loss: 0.78084862, Cur Avg Loss: 1.15084598, Log Avg loss: 1.18161058, Global Avg Loss: 1.49667974, Time: 0.0403 Steps: 88920, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001641, Sample Num: 26256, Cur Loss: 1.56651485, Cur Avg Loss: 1.15112097, Log Avg loss: 1.19597176, Global Avg Loss: 1.49664593, Time: 0.0403 Steps: 88930, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001651, Sample Num: 26416, Cur Loss: 1.50878549, Cur Avg Loss: 1.15169670, Log Avg loss: 1.24617429, Global Avg Loss: 1.49661777, Time: 0.0782 Steps: 88940, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001661, Sample Num: 26576, Cur Loss: 1.43672121, Cur Avg Loss: 1.15356003, Log Avg loss: 1.46119628, Global Avg Loss: 1.49661378, Time: 0.0654 Steps: 88950, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001671, Sample Num: 26736, Cur Loss: 1.17433167, Cur Avg Loss: 1.15383984, Log Avg loss: 1.20031642, Global Avg Loss: 1.49658048, Time: 0.0429 Steps: 88960, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001681, Sample Num: 26896, Cur Loss: 0.68311656, Cur Avg Loss: 1.15215137, Log Avg loss: 0.87000757, Global Avg Loss: 1.49651005, Time: 0.0406 Steps: 88970, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001691, Sample Num: 27056, Cur Loss: 1.13298023, Cur Avg Loss: 1.15382503, Log Avg loss: 1.43516733, Global Avg Loss: 1.49650316, Time: 0.0571 Steps: 88980, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001701, Sample Num: 27216, Cur Loss: 0.89098322, Cur Avg Loss: 1.15364444, Log Avg loss: 1.12310553, Global Avg Loss: 1.49646120, Time: 0.0510 Steps: 88990, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001711, Sample Num: 27376, Cur Loss: 0.94107896, Cur Avg Loss: 1.15235409, Log Avg loss: 0.93286582, Global Avg Loss: 1.49639787, Time: 0.0404 Steps: 89000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001721, Sample Num: 27536, Cur Loss: 2.46517920, Cur Avg Loss: 1.15379113, Log Avg loss: 1.39966986, Global Avg Loss: 1.49638701, Time: 0.0692 Steps: 89010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001731, Sample Num: 27696, Cur Loss: 1.56435299, Cur Avg Loss: 1.15479163, Log Avg loss: 1.32697683, Global Avg Loss: 1.49636798, Time: 0.0404 Steps: 89020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001741, Sample Num: 27856, Cur Loss: 1.18905544, Cur Avg Loss: 1.15420006, Log Avg loss: 1.05179994, Global Avg Loss: 1.49631804, Time: 0.0402 Steps: 89030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001751, Sample Num: 28016, Cur Loss: 1.57988524, Cur Avg Loss: 1.15568290, Log Avg loss: 1.41384564, Global Avg Loss: 1.49630878, Time: 0.0402 Steps: 89040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001761, Sample Num: 28176, Cur Loss: 0.69504422, Cur Avg Loss: 1.15606431, Log Avg loss: 1.22284870, Global Avg Loss: 1.49627807, Time: 0.0402 Steps: 89050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001771, Sample Num: 28336, Cur Loss: 0.62387836, Cur Avg Loss: 1.15627909, Log Avg loss: 1.19410127, Global Avg Loss: 1.49624414, Time: 0.0402 Steps: 89060, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001781, Sample Num: 28496, Cur Loss: 1.28029954, Cur Avg Loss: 1.15571324, Log Avg loss: 1.05550179, Global Avg Loss: 1.49619466, Time: 0.0403 Steps: 89070, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001791, Sample Num: 28656, Cur Loss: 2.41689062, Cur Avg Loss: 1.15738859, Log Avg loss: 1.45576783, Global Avg Loss: 1.49619012, Time: 0.0402 Steps: 89080, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001801, Sample Num: 28816, Cur Loss: 0.55758351, Cur Avg Loss: 1.15897885, Log Avg loss: 1.44379454, Global Avg Loss: 1.49618424, Time: 0.0403 Steps: 89090, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001811, Sample Num: 28976, Cur Loss: 0.98368317, Cur Avg Loss: 1.15881911, Log Avg loss: 1.13005081, Global Avg Loss: 1.49614315, Time: 0.0403 Steps: 89100, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001821, Sample Num: 29136, Cur Loss: 1.32917094, Cur Avg Loss: 1.15852208, Log Avg loss: 1.10472952, Global Avg Loss: 1.49609922, Time: 0.0403 Steps: 89110, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001831, Sample Num: 29296, Cur Loss: 0.38742483, Cur Avg Loss: 1.15863679, Log Avg loss: 1.17952461, Global Avg Loss: 1.49606370, Time: 0.0402 Steps: 89120, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001841, Sample Num: 29456, Cur Loss: 1.83451414, Cur Avg Loss: 1.15945443, Log Avg loss: 1.30916413, Global Avg Loss: 1.49604273, Time: 0.0402 Steps: 89130, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001851, Sample Num: 29616, Cur Loss: 1.06543350, Cur Avg Loss: 1.16028077, Log Avg loss: 1.31241086, Global Avg Loss: 1.49602213, Time: 0.0403 Steps: 89140, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001861, Sample Num: 29776, Cur Loss: 3.43466306, Cur Avg Loss: 1.16061016, Log Avg loss: 1.22158071, Global Avg Loss: 1.49599135, Time: 0.0402 Steps: 89150, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001871, Sample Num: 29936, Cur Loss: 0.35172060, Cur Avg Loss: 1.16027379, Log Avg loss: 1.09767384, Global Avg Loss: 1.49594667, Time: 0.0402 Steps: 89160, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001881, Sample Num: 30096, Cur Loss: 1.85435152, Cur Avg Loss: 1.15980586, Log Avg loss: 1.07225795, Global Avg Loss: 1.49589916, Time: 0.0403 Steps: 89170, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001891, Sample Num: 30256, Cur Loss: 0.59706932, Cur Avg Loss: 1.15988969, Log Avg loss: 1.17565720, Global Avg Loss: 1.49586325, Time: 0.0402 Steps: 89180, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001901, Sample Num: 30416, Cur Loss: 0.55342269, Cur Avg Loss: 1.15949070, Log Avg loss: 1.08404136, Global Avg Loss: 1.49581707, Time: 0.0402 Steps: 89190, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001911, Sample Num: 30576, Cur Loss: 0.73206049, Cur Avg Loss: 1.15974059, Log Avg loss: 1.20724591, Global Avg Loss: 1.49578472, Time: 0.0402 Steps: 89200, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001921, Sample Num: 30736, Cur Loss: 0.38107151, Cur Avg Loss: 1.15756077, Log Avg loss: 0.74099590, Global Avg Loss: 1.49570011, Time: 0.0402 Steps: 89210, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001931, Sample Num: 30896, Cur Loss: 1.21053362, Cur Avg Loss: 1.15639290, Log Avg loss: 0.93204542, Global Avg Loss: 1.49563694, Time: 0.0402 Steps: 89220, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001941, Sample Num: 31056, Cur Loss: 0.48330346, Cur Avg Loss: 1.15619096, Log Avg loss: 1.11719579, Global Avg Loss: 1.49559453, Time: 0.0402 Steps: 89230, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001951, Sample Num: 31216, Cur Loss: 1.87168872, Cur Avg Loss: 1.15682568, Log Avg loss: 1.28002632, Global Avg Loss: 1.49557037, Time: 0.0402 Steps: 89240, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001961, Sample Num: 31376, Cur Loss: 1.39036536, Cur Avg Loss: 1.15620695, Log Avg loss: 1.03549247, Global Avg Loss: 1.49551882, Time: 0.0402 Steps: 89250, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001971, Sample Num: 31536, Cur Loss: 0.45072863, Cur Avg Loss: 1.15666191, Log Avg loss: 1.24587952, Global Avg Loss: 1.49549085, Time: 0.0402 Steps: 89260, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001981, Sample Num: 31696, Cur Loss: 0.96892864, Cur Avg Loss: 1.15621109, Log Avg loss: 1.06735445, Global Avg Loss: 1.49544289, Time: 0.0402 Steps: 89270, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001991, Sample Num: 31856, Cur Loss: 2.03549767, Cur Avg Loss: 1.15591757, Log Avg loss: 1.09777119, Global Avg Loss: 1.49539835, Time: 0.0402 Steps: 89280, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002001, Sample Num: 32016, Cur Loss: 0.98217779, Cur Avg Loss: 1.15587720, Log Avg loss: 1.14783927, Global Avg Loss: 1.49535943, Time: 0.0405 Steps: 89290, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002011, Sample Num: 32176, Cur Loss: 0.62140787, Cur Avg Loss: 1.15728073, Log Avg loss: 1.43812649, Global Avg Loss: 1.49535302, Time: 0.0679 Steps: 89300, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002021, Sample Num: 32336, Cur Loss: 1.40534067, Cur Avg Loss: 1.15870276, Log Avg loss: 1.44467357, Global Avg Loss: 1.49534734, Time: 0.0811 Steps: 89310, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002031, Sample Num: 32496, Cur Loss: 2.14795232, Cur Avg Loss: 1.15845530, Log Avg loss: 1.10844373, Global Avg Loss: 1.49530403, Time: 0.0816 Steps: 89320, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002041, Sample Num: 32656, Cur Loss: 0.62546921, Cur Avg Loss: 1.15818895, Log Avg loss: 1.10409321, Global Avg Loss: 1.49526023, Time: 0.0586 Steps: 89330, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002051, Sample Num: 32816, Cur Loss: 1.59210253, Cur Avg Loss: 1.15847324, Log Avg loss: 1.21649669, Global Avg Loss: 1.49522903, Time: 0.0452 Steps: 89340, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002061, Sample Num: 32976, Cur Loss: 0.38120735, Cur Avg Loss: 1.15873345, Log Avg loss: 1.21210221, Global Avg Loss: 1.49519734, Time: 0.0808 Steps: 89350, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002071, Sample Num: 33136, Cur Loss: 0.45484784, Cur Avg Loss: 1.15884581, Log Avg loss: 1.18200378, Global Avg Loss: 1.49516229, Time: 0.0813 Steps: 89360, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002081, Sample Num: 33296, Cur Loss: 0.81306922, Cur Avg Loss: 1.15837981, Log Avg loss: 1.06187135, Global Avg Loss: 1.49511381, Time: 0.0429 Steps: 89370, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002091, Sample Num: 33456, Cur Loss: 1.24246168, Cur Avg Loss: 1.15891482, Log Avg loss: 1.27024927, Global Avg Loss: 1.49508865, Time: 0.0402 Steps: 89380, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002101, Sample Num: 33616, Cur Loss: 1.61020398, Cur Avg Loss: 1.15917497, Log Avg loss: 1.21357292, Global Avg Loss: 1.49505716, Time: 0.0402 Steps: 89390, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002111, Sample Num: 33776, Cur Loss: 1.66635799, Cur Avg Loss: 1.15970026, Log Avg loss: 1.27006395, Global Avg Loss: 1.49503199, Time: 0.0402 Steps: 89400, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002121, Sample Num: 33936, Cur Loss: 1.29550278, Cur Avg Loss: 1.15961165, Log Avg loss: 1.14090599, Global Avg Loss: 1.49499239, Time: 0.0402 Steps: 89410, Updated lr: 0.000016 ***** Running evaluation checkpoint-89418 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-89418 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.296738, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.081965, "eval_total_loss": 760.621298, "eval_mae": 0.813562, "eval_mse": 1.08214, "eval_r2": 0.312121, "eval_sp_statistic": 0.44831, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.560365, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.040739, "test_total_loss": 522.450988, "test_mae": 0.807643, "test_mse": 1.040965, "test_r2": 0.328152, "test_sp_statistic": 0.47344, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.60221, "test_ps_pvalue": 0.0, "lr": 1.6151730678046468e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4949753221360464, "train_cur_epoch_loss": 2469.9704466611147, "train_cur_epoch_avg_loss": 1.1601552121470713, "train_cur_epoch_time": 95.29673790931702, "train_cur_epoch_avg_time": 0.04476126721903101, "epoch": 42, "step": 89418} ################################################## Training, Epoch: 0043, Batch: 000002, Sample Num: 32, Cur Loss: 1.26691532, Cur Avg Loss: 0.73786575, Log Avg loss: 1.19098672, Global Avg Loss: 1.49495839, Time: 0.0402 Steps: 89420, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000012, Sample Num: 192, Cur Loss: 1.72555494, Cur Avg Loss: 1.06723488, Log Avg loss: 1.13310871, Global Avg Loss: 1.49491793, Time: 0.0403 Steps: 89430, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000022, Sample Num: 352, Cur Loss: 1.44191027, Cur Avg Loss: 1.17161835, Log Avg loss: 1.29687852, Global Avg Loss: 1.49489578, Time: 0.0402 Steps: 89440, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000032, Sample Num: 512, Cur Loss: 1.55334449, Cur Avg Loss: 1.12232601, Log Avg loss: 1.01388287, Global Avg Loss: 1.49484201, Time: 0.0402 Steps: 89450, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000042, Sample Num: 672, Cur Loss: 0.99759096, Cur Avg Loss: 1.14574906, Log Avg loss: 1.22070279, Global Avg Loss: 1.49481137, Time: 0.0402 Steps: 89460, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000052, Sample Num: 832, Cur Loss: 0.91051084, Cur Avg Loss: 1.13457909, Log Avg loss: 1.08766524, Global Avg Loss: 1.49476586, Time: 0.0402 Steps: 89470, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000062, Sample Num: 992, Cur Loss: 0.79157728, Cur Avg Loss: 1.10544021, Log Avg loss: 0.95391803, Global Avg Loss: 1.49470542, Time: 0.0403 Steps: 89480, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000072, Sample Num: 1152, Cur Loss: 1.84825659, Cur Avg Loss: 1.08299225, Log Avg loss: 0.94381491, Global Avg Loss: 1.49464386, Time: 0.0402 Steps: 89490, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000082, Sample Num: 1312, Cur Loss: 1.94695795, Cur Avg Loss: 1.10423039, Log Avg loss: 1.25714497, Global Avg Loss: 1.49461732, Time: 0.0402 Steps: 89500, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000092, Sample Num: 1472, Cur Loss: 1.41973662, Cur Avg Loss: 1.10767476, Log Avg loss: 1.13591864, Global Avg Loss: 1.49457725, Time: 0.0402 Steps: 89510, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000102, Sample Num: 1632, Cur Loss: 0.54232252, Cur Avg Loss: 1.10602489, Log Avg loss: 1.09084609, Global Avg Loss: 1.49453215, Time: 0.0402 Steps: 89520, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000112, Sample Num: 1792, Cur Loss: 1.02860320, Cur Avg Loss: 1.10543767, Log Avg loss: 1.09944794, Global Avg Loss: 1.49448802, Time: 0.0403 Steps: 89530, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000122, Sample Num: 1952, Cur Loss: 1.48026049, Cur Avg Loss: 1.10775531, Log Avg loss: 1.13371296, Global Avg Loss: 1.49444773, Time: 0.0402 Steps: 89540, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000132, Sample Num: 2112, Cur Loss: 0.86659998, Cur Avg Loss: 1.13609202, Log Avg loss: 1.48179985, Global Avg Loss: 1.49444631, Time: 0.0402 Steps: 89550, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000142, Sample Num: 2272, Cur Loss: 0.70557976, Cur Avg Loss: 1.12755096, Log Avg loss: 1.01480902, Global Avg Loss: 1.49439276, Time: 0.0403 Steps: 89560, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000152, Sample Num: 2432, Cur Loss: 1.93858194, Cur Avg Loss: 1.15191401, Log Avg loss: 1.49786926, Global Avg Loss: 1.49439315, Time: 0.0402 Steps: 89570, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000162, Sample Num: 2592, Cur Loss: 0.86602467, Cur Avg Loss: 1.16636699, Log Avg loss: 1.38605235, Global Avg Loss: 1.49438105, Time: 0.0402 Steps: 89580, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000172, Sample Num: 2752, Cur Loss: 2.52790475, Cur Avg Loss: 1.15590557, Log Avg loss: 0.98643053, Global Avg Loss: 1.49432436, Time: 0.0402 Steps: 89590, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000182, Sample Num: 2912, Cur Loss: 0.60264695, Cur Avg Loss: 1.14579694, Log Avg loss: 0.97192855, Global Avg Loss: 1.49426605, Time: 0.0402 Steps: 89600, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000192, Sample Num: 3072, Cur Loss: 0.68733776, Cur Avg Loss: 1.15955761, Log Avg loss: 1.41000167, Global Avg Loss: 1.49425665, Time: 0.0402 Steps: 89610, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000202, Sample Num: 3232, Cur Loss: 0.61724985, Cur Avg Loss: 1.15937934, Log Avg loss: 1.15595669, Global Avg Loss: 1.49421890, Time: 0.0402 Steps: 89620, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000212, Sample Num: 3392, Cur Loss: 0.45618340, Cur Avg Loss: 1.14960503, Log Avg loss: 0.95216393, Global Avg Loss: 1.49415842, Time: 0.0402 Steps: 89630, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000222, Sample Num: 3552, Cur Loss: 1.34117734, Cur Avg Loss: 1.15448545, Log Avg loss: 1.25795025, Global Avg Loss: 1.49413207, Time: 0.0402 Steps: 89640, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000232, Sample Num: 3712, Cur Loss: 1.25659800, Cur Avg Loss: 1.15712191, Log Avg loss: 1.21565145, Global Avg Loss: 1.49410101, Time: 0.0402 Steps: 89650, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000242, Sample Num: 3872, Cur Loss: 1.11227298, Cur Avg Loss: 1.15056463, Log Avg loss: 0.99843574, Global Avg Loss: 1.49404573, Time: 0.0402 Steps: 89660, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000252, Sample Num: 4032, Cur Loss: 1.08473420, Cur Avg Loss: 1.15170329, Log Avg loss: 1.17925890, Global Avg Loss: 1.49401062, Time: 0.0402 Steps: 89670, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000262, Sample Num: 4192, Cur Loss: 1.15036249, Cur Avg Loss: 1.15279911, Log Avg loss: 1.18041378, Global Avg Loss: 1.49397565, Time: 0.0542 Steps: 89680, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000272, Sample Num: 4352, Cur Loss: 1.57850051, Cur Avg Loss: 1.15657565, Log Avg loss: 1.25552100, Global Avg Loss: 1.49394907, Time: 0.0404 Steps: 89690, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000282, Sample Num: 4512, Cur Loss: 1.03030956, Cur Avg Loss: 1.15306300, Log Avg loss: 1.05751870, Global Avg Loss: 1.49390041, Time: 0.0631 Steps: 89700, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000292, Sample Num: 4672, Cur Loss: 0.21724738, Cur Avg Loss: 1.14882519, Log Avg loss: 1.02931919, Global Avg Loss: 1.49384863, Time: 0.0822 Steps: 89710, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000302, Sample Num: 4832, Cur Loss: 0.56195992, Cur Avg Loss: 1.15947359, Log Avg loss: 1.47040665, Global Avg Loss: 1.49384601, Time: 0.0492 Steps: 89720, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000312, Sample Num: 4992, Cur Loss: 0.79894626, Cur Avg Loss: 1.16450331, Log Avg loss: 1.31640093, Global Avg Loss: 1.49382624, Time: 0.0441 Steps: 89730, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000322, Sample Num: 5152, Cur Loss: 0.64994621, Cur Avg Loss: 1.16103744, Log Avg loss: 1.05290223, Global Avg Loss: 1.49377711, Time: 0.0589 Steps: 89740, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000332, Sample Num: 5312, Cur Loss: 1.91836214, Cur Avg Loss: 1.16203032, Log Avg loss: 1.19400125, Global Avg Loss: 1.49374370, Time: 0.0670 Steps: 89750, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000342, Sample Num: 5472, Cur Loss: 1.90281653, Cur Avg Loss: 1.17182325, Log Avg loss: 1.49694859, Global Avg Loss: 1.49374406, Time: 0.0402 Steps: 89760, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000352, Sample Num: 5632, Cur Loss: 0.47939384, Cur Avg Loss: 1.17795297, Log Avg loss: 1.38758936, Global Avg Loss: 1.49373224, Time: 0.0402 Steps: 89770, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000362, Sample Num: 5792, Cur Loss: 0.84087825, Cur Avg Loss: 1.17026034, Log Avg loss: 0.89947971, Global Avg Loss: 1.49366605, Time: 0.0402 Steps: 89780, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000372, Sample Num: 5952, Cur Loss: 1.85837638, Cur Avg Loss: 1.17561196, Log Avg loss: 1.36934047, Global Avg Loss: 1.49365220, Time: 0.0402 Steps: 89790, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000382, Sample Num: 6112, Cur Loss: 1.51141524, Cur Avg Loss: 1.17503196, Log Avg loss: 1.15345593, Global Avg Loss: 1.49361432, Time: 0.0402 Steps: 89800, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000392, Sample Num: 6272, Cur Loss: 1.19635034, Cur Avg Loss: 1.17157592, Log Avg loss: 1.03955537, Global Avg Loss: 1.49356376, Time: 0.0402 Steps: 89810, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000402, Sample Num: 6432, Cur Loss: 1.28134096, Cur Avg Loss: 1.17316855, Log Avg loss: 1.23559962, Global Avg Loss: 1.49353504, Time: 0.0402 Steps: 89820, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000412, Sample Num: 6592, Cur Loss: 1.02056777, Cur Avg Loss: 1.17038871, Log Avg loss: 1.05863913, Global Avg Loss: 1.49348662, Time: 0.0402 Steps: 89830, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000422, Sample Num: 6752, Cur Loss: 1.03852105, Cur Avg Loss: 1.16059768, Log Avg loss: 0.75720704, Global Avg Loss: 1.49340467, Time: 0.0403 Steps: 89840, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000432, Sample Num: 6912, Cur Loss: 0.37706864, Cur Avg Loss: 1.15393101, Log Avg loss: 0.87259777, Global Avg Loss: 1.49333558, Time: 0.0402 Steps: 89850, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000442, Sample Num: 7072, Cur Loss: 2.11234236, Cur Avg Loss: 1.15474640, Log Avg loss: 1.18997127, Global Avg Loss: 1.49330182, Time: 0.0402 Steps: 89860, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000452, Sample Num: 7232, Cur Loss: 1.19266868, Cur Avg Loss: 1.15975992, Log Avg loss: 1.38135751, Global Avg Loss: 1.49328936, Time: 0.0402 Steps: 89870, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000462, Sample Num: 7392, Cur Loss: 1.57966948, Cur Avg Loss: 1.16271445, Log Avg loss: 1.29625916, Global Avg Loss: 1.49326744, Time: 0.0402 Steps: 89880, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000472, Sample Num: 7552, Cur Loss: 0.58090425, Cur Avg Loss: 1.16051185, Log Avg loss: 1.05875175, Global Avg Loss: 1.49321910, Time: 0.0402 Steps: 89890, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000482, Sample Num: 7712, Cur Loss: 1.24038720, Cur Avg Loss: 1.15890609, Log Avg loss: 1.08311416, Global Avg Loss: 1.49317348, Time: 0.0403 Steps: 89900, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000492, Sample Num: 7872, Cur Loss: 0.44676620, Cur Avg Loss: 1.15987012, Log Avg loss: 1.20633641, Global Avg Loss: 1.49314158, Time: 0.0403 Steps: 89910, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000502, Sample Num: 8032, Cur Loss: 1.20590842, Cur Avg Loss: 1.15505237, Log Avg loss: 0.91801892, Global Avg Loss: 1.49307762, Time: 0.0403 Steps: 89920, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000512, Sample Num: 8192, Cur Loss: 0.46519595, Cur Avg Loss: 1.15144719, Log Avg loss: 0.97046701, Global Avg Loss: 1.49301951, Time: 0.0403 Steps: 89930, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000522, Sample Num: 8352, Cur Loss: 0.92986596, Cur Avg Loss: 1.14670249, Log Avg loss: 0.90377397, Global Avg Loss: 1.49295399, Time: 0.0402 Steps: 89940, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000532, Sample Num: 8512, Cur Loss: 0.97737437, Cur Avg Loss: 1.14333067, Log Avg loss: 0.96732157, Global Avg Loss: 1.49289556, Time: 0.0402 Steps: 89950, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000542, Sample Num: 8672, Cur Loss: 1.89831352, Cur Avg Loss: 1.14430942, Log Avg loss: 1.19637893, Global Avg Loss: 1.49286260, Time: 0.0403 Steps: 89960, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000552, Sample Num: 8832, Cur Loss: 1.92004752, Cur Avg Loss: 1.15288458, Log Avg loss: 1.61765837, Global Avg Loss: 1.49287647, Time: 0.0403 Steps: 89970, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000562, Sample Num: 8992, Cur Loss: 2.33110929, Cur Avg Loss: 1.15739380, Log Avg loss: 1.40630264, Global Avg Loss: 1.49286684, Time: 0.0402 Steps: 89980, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000572, Sample Num: 9152, Cur Loss: 0.86757821, Cur Avg Loss: 1.15755339, Log Avg loss: 1.16652267, Global Avg Loss: 1.49283058, Time: 0.0402 Steps: 89990, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000582, Sample Num: 9312, Cur Loss: 1.38070488, Cur Avg Loss: 1.16061707, Log Avg loss: 1.33585962, Global Avg Loss: 1.49281314, Time: 0.0403 Steps: 90000, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000592, Sample Num: 9472, Cur Loss: 0.66629994, Cur Avg Loss: 1.16015653, Log Avg loss: 1.13335311, Global Avg Loss: 1.49277320, Time: 0.0403 Steps: 90010, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000602, Sample Num: 9632, Cur Loss: 1.02739131, Cur Avg Loss: 1.16180620, Log Avg loss: 1.25946633, Global Avg Loss: 1.49274729, Time: 0.0404 Steps: 90020, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000612, Sample Num: 9792, Cur Loss: 0.38157427, Cur Avg Loss: 1.16021306, Log Avg loss: 1.06430600, Global Avg Loss: 1.49269970, Time: 0.0406 Steps: 90030, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000622, Sample Num: 9952, Cur Loss: 0.60372460, Cur Avg Loss: 1.15400058, Log Avg loss: 0.77379710, Global Avg Loss: 1.49261985, Time: 0.0403 Steps: 90040, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000632, Sample Num: 10112, Cur Loss: 0.45392862, Cur Avg Loss: 1.15472623, Log Avg loss: 1.19986127, Global Avg Loss: 1.49258734, Time: 0.0480 Steps: 90050, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000642, Sample Num: 10272, Cur Loss: 1.38646948, Cur Avg Loss: 1.15334725, Log Avg loss: 1.06619625, Global Avg Loss: 1.49254000, Time: 0.0705 Steps: 90060, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000652, Sample Num: 10432, Cur Loss: 1.12707043, Cur Avg Loss: 1.15138131, Log Avg loss: 1.02516796, Global Avg Loss: 1.49248811, Time: 0.0404 Steps: 90070, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000662, Sample Num: 10592, Cur Loss: 0.64489609, Cur Avg Loss: 1.15224614, Log Avg loss: 1.20863251, Global Avg Loss: 1.49245660, Time: 0.0435 Steps: 90080, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000672, Sample Num: 10752, Cur Loss: 0.67419970, Cur Avg Loss: 1.14714069, Log Avg loss: 0.80916018, Global Avg Loss: 1.49238075, Time: 0.0403 Steps: 90090, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000682, Sample Num: 10912, Cur Loss: 0.20717695, Cur Avg Loss: 1.14112206, Log Avg loss: 0.73667032, Global Avg Loss: 1.49229688, Time: 0.0685 Steps: 90100, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000692, Sample Num: 11072, Cur Loss: 0.99812126, Cur Avg Loss: 1.13781182, Log Avg loss: 0.91205347, Global Avg Loss: 1.49223248, Time: 0.0402 Steps: 90110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000702, Sample Num: 11232, Cur Loss: 1.02254450, Cur Avg Loss: 1.13624879, Log Avg loss: 1.02808701, Global Avg Loss: 1.49218098, Time: 0.0402 Steps: 90120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000712, Sample Num: 11392, Cur Loss: 0.96361518, Cur Avg Loss: 1.13527087, Log Avg loss: 1.06662058, Global Avg Loss: 1.49213376, Time: 0.0404 Steps: 90130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000722, Sample Num: 11552, Cur Loss: 0.47983146, Cur Avg Loss: 1.13329371, Log Avg loss: 0.99252009, Global Avg Loss: 1.49207834, Time: 0.0402 Steps: 90140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000732, Sample Num: 11712, Cur Loss: 0.70845628, Cur Avg Loss: 1.13357579, Log Avg loss: 1.15394233, Global Avg Loss: 1.49204083, Time: 0.0403 Steps: 90150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000742, Sample Num: 11872, Cur Loss: 0.33051497, Cur Avg Loss: 1.13116920, Log Avg loss: 0.95500661, Global Avg Loss: 1.49198127, Time: 0.0402 Steps: 90160, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000752, Sample Num: 12032, Cur Loss: 1.42339015, Cur Avg Loss: 1.13009471, Log Avg loss: 1.05036740, Global Avg Loss: 1.49193229, Time: 0.0402 Steps: 90170, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000762, Sample Num: 12192, Cur Loss: 0.73729956, Cur Avg Loss: 1.12876233, Log Avg loss: 1.02856752, Global Avg Loss: 1.49188091, Time: 0.0402 Steps: 90180, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000772, Sample Num: 12352, Cur Loss: 0.55487293, Cur Avg Loss: 1.12965073, Log Avg loss: 1.19734704, Global Avg Loss: 1.49184825, Time: 0.0404 Steps: 90190, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000782, Sample Num: 12512, Cur Loss: 1.06808496, Cur Avg Loss: 1.12675816, Log Avg loss: 0.90345104, Global Avg Loss: 1.49178302, Time: 0.0402 Steps: 90200, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000792, Sample Num: 12672, Cur Loss: 0.48940098, Cur Avg Loss: 1.12534122, Log Avg loss: 1.01453654, Global Avg Loss: 1.49173011, Time: 0.0403 Steps: 90210, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000802, Sample Num: 12832, Cur Loss: 3.04042411, Cur Avg Loss: 1.12802690, Log Avg loss: 1.34073340, Global Avg Loss: 1.49171338, Time: 0.0403 Steps: 90220, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000812, Sample Num: 12992, Cur Loss: 0.34500033, Cur Avg Loss: 1.12479988, Log Avg loss: 0.86599264, Global Avg Loss: 1.49164403, Time: 0.0403 Steps: 90230, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000822, Sample Num: 13152, Cur Loss: 1.46998858, Cur Avg Loss: 1.12388912, Log Avg loss: 1.04993549, Global Avg Loss: 1.49159508, Time: 0.0402 Steps: 90240, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000832, Sample Num: 13312, Cur Loss: 1.27199268, Cur Avg Loss: 1.12565010, Log Avg loss: 1.27040237, Global Avg Loss: 1.49157057, Time: 0.0402 Steps: 90250, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000842, Sample Num: 13472, Cur Loss: 0.25047481, Cur Avg Loss: 1.12305593, Log Avg loss: 0.90722097, Global Avg Loss: 1.49150583, Time: 0.0402 Steps: 90260, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000852, Sample Num: 13632, Cur Loss: 2.02168679, Cur Avg Loss: 1.12542195, Log Avg loss: 1.32464061, Global Avg Loss: 1.49148735, Time: 0.0402 Steps: 90270, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000862, Sample Num: 13792, Cur Loss: 0.73509240, Cur Avg Loss: 1.12068816, Log Avg loss: 0.71736947, Global Avg Loss: 1.49140160, Time: 0.0402 Steps: 90280, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000872, Sample Num: 13952, Cur Loss: 0.70909131, Cur Avg Loss: 1.12031406, Log Avg loss: 1.08806651, Global Avg Loss: 1.49135693, Time: 0.0403 Steps: 90290, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000882, Sample Num: 14112, Cur Loss: 1.13176847, Cur Avg Loss: 1.11919080, Log Avg loss: 1.02124287, Global Avg Loss: 1.49130487, Time: 0.0402 Steps: 90300, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000892, Sample Num: 14272, Cur Loss: 1.91160715, Cur Avg Loss: 1.11985082, Log Avg loss: 1.17806431, Global Avg Loss: 1.49127018, Time: 0.0402 Steps: 90310, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000902, Sample Num: 14432, Cur Loss: 2.03581524, Cur Avg Loss: 1.12261399, Log Avg loss: 1.36908926, Global Avg Loss: 1.49125666, Time: 0.0402 Steps: 90320, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000912, Sample Num: 14592, Cur Loss: 1.32926166, Cur Avg Loss: 1.12191616, Log Avg loss: 1.05897198, Global Avg Loss: 1.49120880, Time: 0.0402 Steps: 90330, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000922, Sample Num: 14752, Cur Loss: 2.28842330, Cur Avg Loss: 1.12277385, Log Avg loss: 1.20099519, Global Avg Loss: 1.49117668, Time: 0.0402 Steps: 90340, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000932, Sample Num: 14912, Cur Loss: 1.06972849, Cur Avg Loss: 1.12484133, Log Avg loss: 1.31546302, Global Avg Loss: 1.49115723, Time: 0.0820 Steps: 90350, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000942, Sample Num: 15072, Cur Loss: 1.42578483, Cur Avg Loss: 1.12495310, Log Avg loss: 1.13536930, Global Avg Loss: 1.49111785, Time: 0.0661 Steps: 90360, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000952, Sample Num: 15232, Cur Loss: 1.64245534, Cur Avg Loss: 1.12595680, Log Avg loss: 1.22050589, Global Avg Loss: 1.49108791, Time: 0.0643 Steps: 90370, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000962, Sample Num: 15392, Cur Loss: 2.18049097, Cur Avg Loss: 1.13036635, Log Avg loss: 1.55015483, Global Avg Loss: 1.49109444, Time: 0.0838 Steps: 90380, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000972, Sample Num: 15552, Cur Loss: 0.84719110, Cur Avg Loss: 1.13328713, Log Avg loss: 1.41426702, Global Avg Loss: 1.49108594, Time: 0.0831 Steps: 90390, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000982, Sample Num: 15712, Cur Loss: 1.53647983, Cur Avg Loss: 1.13362234, Log Avg loss: 1.16620401, Global Avg Loss: 1.49105001, Time: 0.0684 Steps: 90400, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000992, Sample Num: 15872, Cur Loss: 0.94960821, Cur Avg Loss: 1.12894009, Log Avg loss: 0.66914397, Global Avg Loss: 1.49095910, Time: 0.0544 Steps: 90410, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001002, Sample Num: 16032, Cur Loss: 1.01565826, Cur Avg Loss: 1.13234672, Log Avg loss: 1.47028437, Global Avg Loss: 1.49095681, Time: 0.0846 Steps: 90420, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001012, Sample Num: 16192, Cur Loss: 0.49915564, Cur Avg Loss: 1.13451200, Log Avg loss: 1.35147260, Global Avg Loss: 1.49094139, Time: 0.0505 Steps: 90430, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001022, Sample Num: 16352, Cur Loss: 1.00035191, Cur Avg Loss: 1.13409582, Log Avg loss: 1.09197832, Global Avg Loss: 1.49089727, Time: 0.0403 Steps: 90440, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001032, Sample Num: 16512, Cur Loss: 0.45210767, Cur Avg Loss: 1.13131709, Log Avg loss: 0.84733121, Global Avg Loss: 1.49082612, Time: 0.0402 Steps: 90450, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001042, Sample Num: 16672, Cur Loss: 1.48766589, Cur Avg Loss: 1.13173116, Log Avg loss: 1.17446318, Global Avg Loss: 1.49079115, Time: 0.0402 Steps: 90460, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001052, Sample Num: 16832, Cur Loss: 1.21486390, Cur Avg Loss: 1.13138716, Log Avg loss: 1.09554182, Global Avg Loss: 1.49074746, Time: 0.0402 Steps: 90470, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001062, Sample Num: 16992, Cur Loss: 0.77606452, Cur Avg Loss: 1.13119235, Log Avg loss: 1.11069881, Global Avg Loss: 1.49070546, Time: 0.0403 Steps: 90480, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001072, Sample Num: 17152, Cur Loss: 1.30870676, Cur Avg Loss: 1.12958606, Log Avg loss: 0.95899782, Global Avg Loss: 1.49064670, Time: 0.0403 Steps: 90490, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001082, Sample Num: 17312, Cur Loss: 1.67971539, Cur Avg Loss: 1.13207372, Log Avg loss: 1.39875103, Global Avg Loss: 1.49063654, Time: 0.0402 Steps: 90500, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001092, Sample Num: 17472, Cur Loss: 0.66552210, Cur Avg Loss: 1.13119913, Log Avg loss: 1.03656845, Global Avg Loss: 1.49058638, Time: 0.0403 Steps: 90510, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001102, Sample Num: 17632, Cur Loss: 0.41813010, Cur Avg Loss: 1.13195866, Log Avg loss: 1.21489976, Global Avg Loss: 1.49055592, Time: 0.0403 Steps: 90520, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001112, Sample Num: 17792, Cur Loss: 2.11744642, Cur Avg Loss: 1.13275705, Log Avg loss: 1.22073937, Global Avg Loss: 1.49052612, Time: 0.0403 Steps: 90530, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001122, Sample Num: 17952, Cur Loss: 1.20166719, Cur Avg Loss: 1.13147103, Log Avg loss: 0.98846575, Global Avg Loss: 1.49047066, Time: 0.0402 Steps: 90540, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001132, Sample Num: 18112, Cur Loss: 2.23598647, Cur Avg Loss: 1.13244378, Log Avg loss: 1.24158584, Global Avg Loss: 1.49044318, Time: 0.0403 Steps: 90550, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001142, Sample Num: 18272, Cur Loss: 1.37251329, Cur Avg Loss: 1.13372639, Log Avg loss: 1.27891826, Global Avg Loss: 1.49041982, Time: 0.0403 Steps: 90560, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001152, Sample Num: 18432, Cur Loss: 2.22646904, Cur Avg Loss: 1.13394217, Log Avg loss: 1.15858423, Global Avg Loss: 1.49038318, Time: 0.0403 Steps: 90570, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001162, Sample Num: 18592, Cur Loss: 1.68784356, Cur Avg Loss: 1.13425253, Log Avg loss: 1.17000548, Global Avg Loss: 1.49034781, Time: 0.0403 Steps: 90580, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001172, Sample Num: 18752, Cur Loss: 1.44188190, Cur Avg Loss: 1.13538952, Log Avg loss: 1.26750827, Global Avg Loss: 1.49032321, Time: 0.0403 Steps: 90590, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001182, Sample Num: 18912, Cur Loss: 0.47690880, Cur Avg Loss: 1.13335034, Log Avg loss: 0.89435768, Global Avg Loss: 1.49025743, Time: 0.0403 Steps: 90600, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001192, Sample Num: 19072, Cur Loss: 0.96718770, Cur Avg Loss: 1.13147727, Log Avg loss: 0.91008071, Global Avg Loss: 1.49019340, Time: 0.0402 Steps: 90610, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001202, Sample Num: 19232, Cur Loss: 1.56149709, Cur Avg Loss: 1.13338157, Log Avg loss: 1.36037439, Global Avg Loss: 1.49017908, Time: 0.0402 Steps: 90620, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001212, Sample Num: 19392, Cur Loss: 0.43204671, Cur Avg Loss: 1.13242609, Log Avg loss: 1.01757755, Global Avg Loss: 1.49012693, Time: 0.0403 Steps: 90630, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001222, Sample Num: 19552, Cur Loss: 1.46008515, Cur Avg Loss: 1.13165998, Log Avg loss: 1.03880782, Global Avg Loss: 1.49007714, Time: 0.0402 Steps: 90640, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001232, Sample Num: 19712, Cur Loss: 1.28675795, Cur Avg Loss: 1.12992010, Log Avg loss: 0.91730654, Global Avg Loss: 1.49001395, Time: 0.0403 Steps: 90650, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001242, Sample Num: 19872, Cur Loss: 0.71748191, Cur Avg Loss: 1.13195781, Log Avg loss: 1.38300394, Global Avg Loss: 1.49000215, Time: 0.0402 Steps: 90660, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001252, Sample Num: 20032, Cur Loss: 0.96681130, Cur Avg Loss: 1.13292550, Log Avg loss: 1.25311245, Global Avg Loss: 1.48997602, Time: 0.0402 Steps: 90670, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001262, Sample Num: 20192, Cur Loss: 1.69877875, Cur Avg Loss: 1.13509402, Log Avg loss: 1.40659208, Global Avg Loss: 1.48996683, Time: 0.0402 Steps: 90680, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001272, Sample Num: 20352, Cur Loss: 0.57952976, Cur Avg Loss: 1.13528312, Log Avg loss: 1.15914732, Global Avg Loss: 1.48993035, Time: 0.0690 Steps: 90690, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001282, Sample Num: 20512, Cur Loss: 1.45766461, Cur Avg Loss: 1.13420722, Log Avg loss: 0.99735291, Global Avg Loss: 1.48987604, Time: 0.0607 Steps: 90700, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001292, Sample Num: 20672, Cur Loss: 0.81928635, Cur Avg Loss: 1.13422975, Log Avg loss: 1.13711791, Global Avg Loss: 1.48983715, Time: 0.0481 Steps: 90710, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001302, Sample Num: 20832, Cur Loss: 0.53190696, Cur Avg Loss: 1.13310780, Log Avg loss: 0.98815270, Global Avg Loss: 1.48978185, Time: 0.0768 Steps: 90720, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001312, Sample Num: 20992, Cur Loss: 1.18789792, Cur Avg Loss: 1.13229714, Log Avg loss: 1.02674929, Global Avg Loss: 1.48973082, Time: 0.0926 Steps: 90730, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001322, Sample Num: 21152, Cur Loss: 0.78520954, Cur Avg Loss: 1.12967103, Log Avg loss: 0.78512438, Global Avg Loss: 1.48965317, Time: 0.0661 Steps: 90740, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001332, Sample Num: 21312, Cur Loss: 1.65007544, Cur Avg Loss: 1.12856471, Log Avg loss: 0.98230962, Global Avg Loss: 1.48959726, Time: 0.0603 Steps: 90750, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001342, Sample Num: 21472, Cur Loss: 1.47702932, Cur Avg Loss: 1.12911368, Log Avg loss: 1.20223660, Global Avg Loss: 1.48956560, Time: 0.0506 Steps: 90760, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001352, Sample Num: 21632, Cur Loss: 0.97425562, Cur Avg Loss: 1.12783165, Log Avg loss: 0.95578353, Global Avg Loss: 1.48950679, Time: 0.0642 Steps: 90770, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001362, Sample Num: 21792, Cur Loss: 1.99239612, Cur Avg Loss: 1.12988430, Log Avg loss: 1.40740265, Global Avg Loss: 1.48949775, Time: 0.0402 Steps: 90780, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001372, Sample Num: 21952, Cur Loss: 0.99793589, Cur Avg Loss: 1.12979578, Log Avg loss: 1.11773942, Global Avg Loss: 1.48945680, Time: 0.0403 Steps: 90790, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001382, Sample Num: 22112, Cur Loss: 2.22742701, Cur Avg Loss: 1.13020914, Log Avg loss: 1.18692115, Global Avg Loss: 1.48942348, Time: 0.0403 Steps: 90800, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001392, Sample Num: 22272, Cur Loss: 0.56167340, Cur Avg Loss: 1.13088043, Log Avg loss: 1.22365318, Global Avg Loss: 1.48939422, Time: 0.0402 Steps: 90810, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001402, Sample Num: 22432, Cur Loss: 0.97630376, Cur Avg Loss: 1.13097807, Log Avg loss: 1.14456898, Global Avg Loss: 1.48935625, Time: 0.0403 Steps: 90820, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001412, Sample Num: 22592, Cur Loss: 1.62496328, Cur Avg Loss: 1.13070199, Log Avg loss: 1.09199643, Global Avg Loss: 1.48931250, Time: 0.0403 Steps: 90830, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001422, Sample Num: 22752, Cur Loss: 0.35363719, Cur Avg Loss: 1.13118820, Log Avg loss: 1.19984037, Global Avg Loss: 1.48928064, Time: 0.0402 Steps: 90840, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001432, Sample Num: 22912, Cur Loss: 0.31268010, Cur Avg Loss: 1.12923468, Log Avg loss: 0.85144494, Global Avg Loss: 1.48921043, Time: 0.0403 Steps: 90850, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001442, Sample Num: 23072, Cur Loss: 1.23494625, Cur Avg Loss: 1.12743783, Log Avg loss: 0.87012876, Global Avg Loss: 1.48914229, Time: 0.0403 Steps: 90860, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001452, Sample Num: 23232, Cur Loss: 2.43697309, Cur Avg Loss: 1.13035393, Log Avg loss: 1.55085549, Global Avg Loss: 1.48914908, Time: 0.0403 Steps: 90870, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001462, Sample Num: 23392, Cur Loss: 0.66815525, Cur Avg Loss: 1.13292949, Log Avg loss: 1.50690080, Global Avg Loss: 1.48915104, Time: 0.0401 Steps: 90880, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001472, Sample Num: 23552, Cur Loss: 2.09578848, Cur Avg Loss: 1.13136923, Log Avg loss: 0.90325890, Global Avg Loss: 1.48908658, Time: 0.0402 Steps: 90890, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001482, Sample Num: 23712, Cur Loss: 0.99980211, Cur Avg Loss: 1.13301618, Log Avg loss: 1.37544677, Global Avg Loss: 1.48907407, Time: 0.0403 Steps: 90900, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001492, Sample Num: 23872, Cur Loss: 1.31432736, Cur Avg Loss: 1.13730217, Log Avg loss: 1.77248608, Global Avg Loss: 1.48910525, Time: 0.0403 Steps: 90910, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001502, Sample Num: 24032, Cur Loss: 0.87247312, Cur Avg Loss: 1.13830533, Log Avg loss: 1.28797749, Global Avg Loss: 1.48908313, Time: 0.0403 Steps: 90920, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001512, Sample Num: 24192, Cur Loss: 0.85094619, Cur Avg Loss: 1.13884240, Log Avg loss: 1.21950958, Global Avg Loss: 1.48905348, Time: 0.0402 Steps: 90930, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001522, Sample Num: 24352, Cur Loss: 1.66151440, Cur Avg Loss: 1.13978971, Log Avg loss: 1.28302277, Global Avg Loss: 1.48903083, Time: 0.0403 Steps: 90940, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001532, Sample Num: 24512, Cur Loss: 1.03257227, Cur Avg Loss: 1.13953677, Log Avg loss: 1.10104007, Global Avg Loss: 1.48898817, Time: 0.0402 Steps: 90950, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001542, Sample Num: 24672, Cur Loss: 1.35845315, Cur Avg Loss: 1.13861038, Log Avg loss: 0.99668750, Global Avg Loss: 1.48893404, Time: 0.0402 Steps: 90960, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001552, Sample Num: 24832, Cur Loss: 1.09518421, Cur Avg Loss: 1.13838861, Log Avg loss: 1.10419144, Global Avg Loss: 1.48889175, Time: 0.0403 Steps: 90970, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001562, Sample Num: 24992, Cur Loss: 0.54504687, Cur Avg Loss: 1.13659926, Log Avg loss: 0.85889244, Global Avg Loss: 1.48882250, Time: 0.0402 Steps: 90980, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001572, Sample Num: 25152, Cur Loss: 3.57568240, Cur Avg Loss: 1.13767562, Log Avg loss: 1.30580261, Global Avg Loss: 1.48880239, Time: 0.0402 Steps: 90990, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001582, Sample Num: 25312, Cur Loss: 1.37860429, Cur Avg Loss: 1.13579033, Log Avg loss: 0.83942313, Global Avg Loss: 1.48873103, Time: 0.0403 Steps: 91000, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001592, Sample Num: 25472, Cur Loss: 1.91147447, Cur Avg Loss: 1.13580331, Log Avg loss: 1.13785622, Global Avg Loss: 1.48869248, Time: 0.0402 Steps: 91010, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001602, Sample Num: 25632, Cur Loss: 1.01521158, Cur Avg Loss: 1.13633752, Log Avg loss: 1.22138468, Global Avg Loss: 1.48866311, Time: 0.0403 Steps: 91020, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001612, Sample Num: 25792, Cur Loss: 1.27965140, Cur Avg Loss: 1.13644375, Log Avg loss: 1.15346071, Global Avg Loss: 1.48862628, Time: 0.0402 Steps: 91030, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001622, Sample Num: 25952, Cur Loss: 1.99931455, Cur Avg Loss: 1.13714752, Log Avg loss: 1.25059545, Global Avg Loss: 1.48860014, Time: 0.0406 Steps: 91040, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001632, Sample Num: 26112, Cur Loss: 1.01026678, Cur Avg Loss: 1.13853309, Log Avg loss: 1.36327230, Global Avg Loss: 1.48858637, Time: 0.1011 Steps: 91050, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001642, Sample Num: 26272, Cur Loss: 2.02767038, Cur Avg Loss: 1.13911410, Log Avg loss: 1.23393529, Global Avg Loss: 1.48855841, Time: 0.0506 Steps: 91060, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001652, Sample Num: 26432, Cur Loss: 1.41049719, Cur Avg Loss: 1.14096222, Log Avg loss: 1.44442436, Global Avg Loss: 1.48855356, Time: 0.0931 Steps: 91070, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001662, Sample Num: 26592, Cur Loss: 0.94919848, Cur Avg Loss: 1.14099365, Log Avg loss: 1.14618485, Global Avg Loss: 1.48851597, Time: 0.0508 Steps: 91080, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001672, Sample Num: 26752, Cur Loss: 0.85562098, Cur Avg Loss: 1.14105034, Log Avg loss: 1.15047301, Global Avg Loss: 1.48847886, Time: 0.0410 Steps: 91090, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001682, Sample Num: 26912, Cur Loss: 1.52210772, Cur Avg Loss: 1.14057404, Log Avg loss: 1.06093564, Global Avg Loss: 1.48843193, Time: 0.0754 Steps: 91100, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001692, Sample Num: 27072, Cur Loss: 0.63873798, Cur Avg Loss: 1.13927934, Log Avg loss: 0.92151199, Global Avg Loss: 1.48836971, Time: 0.0584 Steps: 91110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001702, Sample Num: 27232, Cur Loss: 1.95494795, Cur Avg Loss: 1.13867589, Log Avg loss: 1.03657202, Global Avg Loss: 1.48832012, Time: 0.0479 Steps: 91120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001712, Sample Num: 27392, Cur Loss: 0.91979617, Cur Avg Loss: 1.14032918, Log Avg loss: 1.42171790, Global Avg Loss: 1.48831282, Time: 0.0402 Steps: 91130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001722, Sample Num: 27552, Cur Loss: 1.30642533, Cur Avg Loss: 1.14042996, Log Avg loss: 1.15768461, Global Avg Loss: 1.48827654, Time: 0.0402 Steps: 91140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001732, Sample Num: 27712, Cur Loss: 0.39085203, Cur Avg Loss: 1.13979984, Log Avg loss: 1.03129316, Global Avg Loss: 1.48822640, Time: 0.0403 Steps: 91150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001742, Sample Num: 27872, Cur Loss: 0.95391452, Cur Avg Loss: 1.13965508, Log Avg loss: 1.11458231, Global Avg Loss: 1.48818542, Time: 0.0402 Steps: 91160, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001752, Sample Num: 28032, Cur Loss: 1.67673910, Cur Avg Loss: 1.14002293, Log Avg loss: 1.20410198, Global Avg Loss: 1.48815426, Time: 0.0402 Steps: 91170, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001762, Sample Num: 28192, Cur Loss: 1.54456902, Cur Avg Loss: 1.14351736, Log Avg loss: 1.75574134, Global Avg Loss: 1.48818360, Time: 0.0402 Steps: 91180, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001772, Sample Num: 28352, Cur Loss: 1.66709208, Cur Avg Loss: 1.14274725, Log Avg loss: 1.00705449, Global Avg Loss: 1.48813084, Time: 0.0402 Steps: 91190, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001782, Sample Num: 28512, Cur Loss: 1.28005266, Cur Avg Loss: 1.14280563, Log Avg loss: 1.15314971, Global Avg Loss: 1.48809411, Time: 0.0403 Steps: 91200, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001792, Sample Num: 28672, Cur Loss: 1.68624151, Cur Avg Loss: 1.14301307, Log Avg loss: 1.17998003, Global Avg Loss: 1.48806033, Time: 0.0404 Steps: 91210, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001802, Sample Num: 28832, Cur Loss: 0.33203268, Cur Avg Loss: 1.14300113, Log Avg loss: 1.14086074, Global Avg Loss: 1.48802227, Time: 0.0403 Steps: 91220, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001812, Sample Num: 28992, Cur Loss: 1.26328778, Cur Avg Loss: 1.14490580, Log Avg loss: 1.48812830, Global Avg Loss: 1.48802228, Time: 0.0402 Steps: 91230, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001822, Sample Num: 29152, Cur Loss: 0.39257815, Cur Avg Loss: 1.14357905, Log Avg loss: 0.90317177, Global Avg Loss: 1.48795818, Time: 0.0402 Steps: 91240, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001832, Sample Num: 29312, Cur Loss: 0.57250369, Cur Avg Loss: 1.14417034, Log Avg loss: 1.25190380, Global Avg Loss: 1.48793231, Time: 0.0402 Steps: 91250, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001842, Sample Num: 29472, Cur Loss: 3.31717062, Cur Avg Loss: 1.14569348, Log Avg loss: 1.42473137, Global Avg Loss: 1.48792539, Time: 0.0402 Steps: 91260, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001852, Sample Num: 29632, Cur Loss: 1.65806794, Cur Avg Loss: 1.14530148, Log Avg loss: 1.07309537, Global Avg Loss: 1.48787994, Time: 0.0403 Steps: 91270, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001862, Sample Num: 29792, Cur Loss: 0.68901658, Cur Avg Loss: 1.14604526, Log Avg loss: 1.28379449, Global Avg Loss: 1.48785758, Time: 0.0402 Steps: 91280, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001872, Sample Num: 29952, Cur Loss: 1.69983959, Cur Avg Loss: 1.14665174, Log Avg loss: 1.25957785, Global Avg Loss: 1.48783257, Time: 0.0402 Steps: 91290, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001882, Sample Num: 30112, Cur Loss: 1.38559031, Cur Avg Loss: 1.14730917, Log Avg loss: 1.27037903, Global Avg Loss: 1.48780875, Time: 0.0402 Steps: 91300, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001892, Sample Num: 30272, Cur Loss: 0.36443254, Cur Avg Loss: 1.14689456, Log Avg loss: 1.06886508, Global Avg Loss: 1.48776287, Time: 0.0402 Steps: 91310, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001902, Sample Num: 30432, Cur Loss: 1.56716585, Cur Avg Loss: 1.14800412, Log Avg loss: 1.35793402, Global Avg Loss: 1.48774866, Time: 0.0402 Steps: 91320, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001912, Sample Num: 30592, Cur Loss: 0.84637135, Cur Avg Loss: 1.14694918, Log Avg loss: 0.94629911, Global Avg Loss: 1.48768937, Time: 0.0402 Steps: 91330, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001922, Sample Num: 30752, Cur Loss: 0.46433178, Cur Avg Loss: 1.14670066, Log Avg loss: 1.09918321, Global Avg Loss: 1.48764684, Time: 0.0402 Steps: 91340, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001932, Sample Num: 30912, Cur Loss: 1.74658465, Cur Avg Loss: 1.14683658, Log Avg loss: 1.17296157, Global Avg Loss: 1.48761239, Time: 0.0402 Steps: 91350, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001942, Sample Num: 31072, Cur Loss: 1.23202562, Cur Avg Loss: 1.14532132, Log Avg loss: 0.85257287, Global Avg Loss: 1.48754288, Time: 0.0402 Steps: 91360, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001952, Sample Num: 31232, Cur Loss: 2.41396475, Cur Avg Loss: 1.14492627, Log Avg loss: 1.06820623, Global Avg Loss: 1.48749698, Time: 0.0402 Steps: 91370, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001962, Sample Num: 31392, Cur Loss: 2.00259709, Cur Avg Loss: 1.14431528, Log Avg loss: 1.02505074, Global Avg Loss: 1.48744638, Time: 0.0402 Steps: 91380, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001972, Sample Num: 31552, Cur Loss: 0.52558792, Cur Avg Loss: 1.14431478, Log Avg loss: 1.14421626, Global Avg Loss: 1.48740882, Time: 0.0402 Steps: 91390, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001982, Sample Num: 31712, Cur Loss: 1.41352713, Cur Avg Loss: 1.14412350, Log Avg loss: 1.10640340, Global Avg Loss: 1.48736713, Time: 0.0780 Steps: 91400, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001992, Sample Num: 31872, Cur Loss: 0.49430749, Cur Avg Loss: 1.14357106, Log Avg loss: 1.03407755, Global Avg Loss: 1.48731755, Time: 0.0484 Steps: 91410, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002002, Sample Num: 32032, Cur Loss: 1.80128551, Cur Avg Loss: 1.14490064, Log Avg loss: 1.40975323, Global Avg Loss: 1.48730906, Time: 0.0439 Steps: 91420, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002012, Sample Num: 32192, Cur Loss: 1.63439274, Cur Avg Loss: 1.14522888, Log Avg loss: 1.21094294, Global Avg Loss: 1.48727883, Time: 0.0479 Steps: 91430, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002022, Sample Num: 32352, Cur Loss: 1.36913848, Cur Avg Loss: 1.14575330, Log Avg loss: 1.25126527, Global Avg Loss: 1.48725302, Time: 0.0607 Steps: 91440, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002032, Sample Num: 32512, Cur Loss: 1.51185060, Cur Avg Loss: 1.14611289, Log Avg loss: 1.21882314, Global Avg Loss: 1.48722367, Time: 0.0640 Steps: 91450, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002042, Sample Num: 32672, Cur Loss: 0.41027206, Cur Avg Loss: 1.14498627, Log Avg loss: 0.91605684, Global Avg Loss: 1.48716122, Time: 0.0485 Steps: 91460, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002052, Sample Num: 32832, Cur Loss: 2.52931952, Cur Avg Loss: 1.14454402, Log Avg loss: 1.05423582, Global Avg Loss: 1.48711389, Time: 0.0406 Steps: 91470, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002062, Sample Num: 32992, Cur Loss: 0.88986349, Cur Avg Loss: 1.14608081, Log Avg loss: 1.46142960, Global Avg Loss: 1.48711108, Time: 0.0843 Steps: 91480, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002072, Sample Num: 33152, Cur Loss: 1.83878851, Cur Avg Loss: 1.14706303, Log Avg loss: 1.34959800, Global Avg Loss: 1.48709605, Time: 0.0402 Steps: 91490, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002082, Sample Num: 33312, Cur Loss: 0.86020464, Cur Avg Loss: 1.14741389, Log Avg loss: 1.22011089, Global Avg Loss: 1.48706688, Time: 0.0402 Steps: 91500, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002092, Sample Num: 33472, Cur Loss: 1.03806996, Cur Avg Loss: 1.14698387, Log Avg loss: 1.05745499, Global Avg Loss: 1.48701993, Time: 0.0404 Steps: 91510, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002102, Sample Num: 33632, Cur Loss: 0.37976727, Cur Avg Loss: 1.14664293, Log Avg loss: 1.07531860, Global Avg Loss: 1.48697494, Time: 0.0404 Steps: 91520, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002112, Sample Num: 33792, Cur Loss: 0.87146133, Cur Avg Loss: 1.14636664, Log Avg loss: 1.08829086, Global Avg Loss: 1.48693139, Time: 0.0404 Steps: 91530, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002122, Sample Num: 33952, Cur Loss: 0.97576523, Cur Avg Loss: 1.14633906, Log Avg loss: 1.14051334, Global Avg Loss: 1.48689354, Time: 0.0404 Steps: 91540, Updated lr: 0.000014 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.282430, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.158757, "eval_total_loss": 814.606017, "eval_mae": 0.911261, "eval_mse": 1.158785, "eval_r2": 0.2634, "eval_sp_statistic": 0.456688, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.569591, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.202899, "test_total_loss": 603.855121, "test_mae": 0.951313, "test_mse": 1.203055, "test_r2": 0.223537, "test_sp_statistic": 0.479119, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.608791, "test_ps_pvalue": 0.0, "lr": 1.413276434329066e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4868485682306007, "train_cur_epoch_loss": 2438.822521045804, "train_cur_epoch_avg_loss": 1.1455249042018807, "train_cur_epoch_time": 95.28243041038513, "train_cur_epoch_avg_time": 0.04475454692831617, "epoch": 43, "step": 91547} ################################################## Training, Epoch: 0044, Batch: 000003, Sample Num: 48, Cur Loss: 0.82415056, Cur Avg Loss: 1.25603471, Log Avg loss: 1.00591382, Global Avg Loss: 1.48684100, Time: 0.0404 Steps: 91550, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000013, Sample Num: 208, Cur Loss: 2.13015676, Cur Avg Loss: 1.30467580, Log Avg loss: 1.31926813, Global Avg Loss: 1.48682270, Time: 0.0403 Steps: 91560, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000023, Sample Num: 368, Cur Loss: 0.82716548, Cur Avg Loss: 1.21873036, Log Avg loss: 1.10700129, Global Avg Loss: 1.48678122, Time: 0.0403 Steps: 91570, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000033, Sample Num: 528, Cur Loss: 0.85290152, Cur Avg Loss: 1.17334719, Log Avg loss: 1.06896591, Global Avg Loss: 1.48673560, Time: 0.0402 Steps: 91580, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000043, Sample Num: 688, Cur Loss: 0.52375835, Cur Avg Loss: 1.16392222, Log Avg loss: 1.13281980, Global Avg Loss: 1.48669696, Time: 0.0402 Steps: 91590, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000053, Sample Num: 848, Cur Loss: 0.89333582, Cur Avg Loss: 1.16982151, Log Avg loss: 1.19518850, Global Avg Loss: 1.48666514, Time: 0.0402 Steps: 91600, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000063, Sample Num: 1008, Cur Loss: 0.37045988, Cur Avg Loss: 1.15433049, Log Avg loss: 1.07222805, Global Avg Loss: 1.48661990, Time: 0.0402 Steps: 91610, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000073, Sample Num: 1168, Cur Loss: 1.27569139, Cur Avg Loss: 1.16486247, Log Avg loss: 1.23121391, Global Avg Loss: 1.48659202, Time: 0.0402 Steps: 91620, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000083, Sample Num: 1328, Cur Loss: 0.56605542, Cur Avg Loss: 1.14462971, Log Avg loss: 0.99693059, Global Avg Loss: 1.48653858, Time: 0.0402 Steps: 91630, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000093, Sample Num: 1488, Cur Loss: 0.56829095, Cur Avg Loss: 1.13122357, Log Avg loss: 1.01995258, Global Avg Loss: 1.48648767, Time: 0.0402 Steps: 91640, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000103, Sample Num: 1648, Cur Loss: 0.45285362, Cur Avg Loss: 1.12276847, Log Avg loss: 1.04413609, Global Avg Loss: 1.48643940, Time: 0.0403 Steps: 91650, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000113, Sample Num: 1808, Cur Loss: 0.52290905, Cur Avg Loss: 1.12952984, Log Avg loss: 1.19917195, Global Avg Loss: 1.48640806, Time: 0.0402 Steps: 91660, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000123, Sample Num: 1968, Cur Loss: 0.96443415, Cur Avg Loss: 1.12477343, Log Avg loss: 1.07102602, Global Avg Loss: 1.48636275, Time: 0.0402 Steps: 91670, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000133, Sample Num: 2128, Cur Loss: 1.41997945, Cur Avg Loss: 1.12786580, Log Avg loss: 1.16590192, Global Avg Loss: 1.48632779, Time: 0.0403 Steps: 91680, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000143, Sample Num: 2288, Cur Loss: 0.86448884, Cur Avg Loss: 1.15144024, Log Avg loss: 1.46498030, Global Avg Loss: 1.48632546, Time: 0.0402 Steps: 91690, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000153, Sample Num: 2448, Cur Loss: 1.45912302, Cur Avg Loss: 1.18144454, Log Avg loss: 1.61050597, Global Avg Loss: 1.48633901, Time: 0.0402 Steps: 91700, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000163, Sample Num: 2608, Cur Loss: 0.90891182, Cur Avg Loss: 1.15862439, Log Avg loss: 0.80947613, Global Avg Loss: 1.48626520, Time: 0.0402 Steps: 91710, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000173, Sample Num: 2768, Cur Loss: 0.44498548, Cur Avg Loss: 1.16042518, Log Avg loss: 1.18977816, Global Avg Loss: 1.48623288, Time: 0.0402 Steps: 91720, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000183, Sample Num: 2928, Cur Loss: 0.22170755, Cur Avg Loss: 1.15236391, Log Avg loss: 1.01290394, Global Avg Loss: 1.48618128, Time: 0.0402 Steps: 91730, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000193, Sample Num: 3088, Cur Loss: 0.47262704, Cur Avg Loss: 1.14517755, Log Avg loss: 1.01366713, Global Avg Loss: 1.48612977, Time: 0.0402 Steps: 91740, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000203, Sample Num: 3248, Cur Loss: 0.44591132, Cur Avg Loss: 1.13360645, Log Avg loss: 0.91028412, Global Avg Loss: 1.48606701, Time: 0.0402 Steps: 91750, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000213, Sample Num: 3408, Cur Loss: 0.62197626, Cur Avg Loss: 1.13653260, Log Avg loss: 1.19593357, Global Avg Loss: 1.48603539, Time: 0.0402 Steps: 91760, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000223, Sample Num: 3568, Cur Loss: 0.65080732, Cur Avg Loss: 1.14109631, Log Avg loss: 1.23830333, Global Avg Loss: 1.48600839, Time: 0.0402 Steps: 91770, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000233, Sample Num: 3728, Cur Loss: 0.42702997, Cur Avg Loss: 1.14062383, Log Avg loss: 1.13008745, Global Avg Loss: 1.48596961, Time: 0.0402 Steps: 91780, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000243, Sample Num: 3888, Cur Loss: 2.54137993, Cur Avg Loss: 1.15068105, Log Avg loss: 1.38501436, Global Avg Loss: 1.48595862, Time: 0.0403 Steps: 91790, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000253, Sample Num: 4048, Cur Loss: 1.76973569, Cur Avg Loss: 1.15928305, Log Avg loss: 1.36831168, Global Avg Loss: 1.48594580, Time: 0.0402 Steps: 91800, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000263, Sample Num: 4208, Cur Loss: 2.39066792, Cur Avg Loss: 1.15843527, Log Avg loss: 1.13698630, Global Avg Loss: 1.48590779, Time: 0.0445 Steps: 91810, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000273, Sample Num: 4368, Cur Loss: 0.67253315, Cur Avg Loss: 1.15430066, Log Avg loss: 1.04556049, Global Avg Loss: 1.48585983, Time: 0.0404 Steps: 91820, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000283, Sample Num: 4528, Cur Loss: 0.94033784, Cur Avg Loss: 1.14707731, Log Avg loss: 0.94987988, Global Avg Loss: 1.48580147, Time: 0.0754 Steps: 91830, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000293, Sample Num: 4688, Cur Loss: 1.18269622, Cur Avg Loss: 1.14446518, Log Avg loss: 1.07054184, Global Avg Loss: 1.48575625, Time: 0.1004 Steps: 91840, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000303, Sample Num: 4848, Cur Loss: 1.44091582, Cur Avg Loss: 1.14792899, Log Avg loss: 1.24941853, Global Avg Loss: 1.48573052, Time: 0.0403 Steps: 91850, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000313, Sample Num: 5008, Cur Loss: 0.44067219, Cur Avg Loss: 1.15116576, Log Avg loss: 1.24924001, Global Avg Loss: 1.48570478, Time: 0.0532 Steps: 91860, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000323, Sample Num: 5168, Cur Loss: 2.49706960, Cur Avg Loss: 1.14983053, Log Avg loss: 1.10803776, Global Avg Loss: 1.48566367, Time: 0.0586 Steps: 91870, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000333, Sample Num: 5328, Cur Loss: 1.67190957, Cur Avg Loss: 1.15345268, Log Avg loss: 1.27044809, Global Avg Loss: 1.48564024, Time: 0.0713 Steps: 91880, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000343, Sample Num: 5488, Cur Loss: 1.04419875, Cur Avg Loss: 1.15055548, Log Avg loss: 1.05407870, Global Avg Loss: 1.48559328, Time: 0.0528 Steps: 91890, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000353, Sample Num: 5648, Cur Loss: 1.01940513, Cur Avg Loss: 1.15080821, Log Avg loss: 1.15947689, Global Avg Loss: 1.48555779, Time: 0.0404 Steps: 91900, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000363, Sample Num: 5808, Cur Loss: 0.43961883, Cur Avg Loss: 1.14941658, Log Avg loss: 1.10029192, Global Avg Loss: 1.48551588, Time: 0.0403 Steps: 91910, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000373, Sample Num: 5968, Cur Loss: 0.60096681, Cur Avg Loss: 1.15605050, Log Avg loss: 1.39686196, Global Avg Loss: 1.48550623, Time: 0.0403 Steps: 91920, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000383, Sample Num: 6128, Cur Loss: 1.05365992, Cur Avg Loss: 1.15816004, Log Avg loss: 1.23684601, Global Avg Loss: 1.48547918, Time: 0.0402 Steps: 91930, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000393, Sample Num: 6288, Cur Loss: 0.47848159, Cur Avg Loss: 1.15217692, Log Avg loss: 0.92302313, Global Avg Loss: 1.48541801, Time: 0.0403 Steps: 91940, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000403, Sample Num: 6448, Cur Loss: 1.10786331, Cur Avg Loss: 1.15467786, Log Avg loss: 1.25296512, Global Avg Loss: 1.48539272, Time: 0.0402 Steps: 91950, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000413, Sample Num: 6608, Cur Loss: 1.42084658, Cur Avg Loss: 1.15615238, Log Avg loss: 1.21557545, Global Avg Loss: 1.48536338, Time: 0.0403 Steps: 91960, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000423, Sample Num: 6768, Cur Loss: 0.37775648, Cur Avg Loss: 1.15287984, Log Avg loss: 1.01772376, Global Avg Loss: 1.48531254, Time: 0.0403 Steps: 91970, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000433, Sample Num: 6928, Cur Loss: 0.74955517, Cur Avg Loss: 1.15213429, Log Avg loss: 1.12059762, Global Avg Loss: 1.48527289, Time: 0.0403 Steps: 91980, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000443, Sample Num: 7088, Cur Loss: 0.29843694, Cur Avg Loss: 1.15119061, Log Avg loss: 1.11032908, Global Avg Loss: 1.48523213, Time: 0.0402 Steps: 91990, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000453, Sample Num: 7248, Cur Loss: 2.44222021, Cur Avg Loss: 1.14974890, Log Avg loss: 1.08588117, Global Avg Loss: 1.48518872, Time: 0.0403 Steps: 92000, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000463, Sample Num: 7408, Cur Loss: 1.49762988, Cur Avg Loss: 1.15277167, Log Avg loss: 1.28970336, Global Avg Loss: 1.48516747, Time: 0.0403 Steps: 92010, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000473, Sample Num: 7568, Cur Loss: 1.18724251, Cur Avg Loss: 1.16423256, Log Avg loss: 1.69487166, Global Avg Loss: 1.48519026, Time: 0.0402 Steps: 92020, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000483, Sample Num: 7728, Cur Loss: 1.18335199, Cur Avg Loss: 1.16442712, Log Avg loss: 1.17362996, Global Avg Loss: 1.48515641, Time: 0.0402 Steps: 92030, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000493, Sample Num: 7888, Cur Loss: 1.24621260, Cur Avg Loss: 1.16188099, Log Avg loss: 1.03890289, Global Avg Loss: 1.48510792, Time: 0.0403 Steps: 92040, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000503, Sample Num: 8048, Cur Loss: 0.56928313, Cur Avg Loss: 1.16539382, Log Avg loss: 1.33857650, Global Avg Loss: 1.48509200, Time: 0.0402 Steps: 92050, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000513, Sample Num: 8208, Cur Loss: 1.17245221, Cur Avg Loss: 1.16737378, Log Avg loss: 1.26696581, Global Avg Loss: 1.48506831, Time: 0.0403 Steps: 92060, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000523, Sample Num: 8368, Cur Loss: 0.45041773, Cur Avg Loss: 1.16349424, Log Avg loss: 0.96447352, Global Avg Loss: 1.48501177, Time: 0.0402 Steps: 92070, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000533, Sample Num: 8528, Cur Loss: 0.31043836, Cur Avg Loss: 1.16421143, Log Avg loss: 1.20172041, Global Avg Loss: 1.48498100, Time: 0.0402 Steps: 92080, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000543, Sample Num: 8688, Cur Loss: 1.57353389, Cur Avg Loss: 1.16609269, Log Avg loss: 1.26636385, Global Avg Loss: 1.48495726, Time: 0.0403 Steps: 92090, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000553, Sample Num: 8848, Cur Loss: 1.07868540, Cur Avg Loss: 1.16530317, Log Avg loss: 1.12243244, Global Avg Loss: 1.48491790, Time: 0.0402 Steps: 92100, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000563, Sample Num: 9008, Cur Loss: 1.46356750, Cur Avg Loss: 1.16132170, Log Avg loss: 0.94114639, Global Avg Loss: 1.48485886, Time: 0.0402 Steps: 92110, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000573, Sample Num: 9168, Cur Loss: 0.48144078, Cur Avg Loss: 1.15822873, Log Avg loss: 0.98409449, Global Avg Loss: 1.48480450, Time: 0.0402 Steps: 92120, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000583, Sample Num: 9328, Cur Loss: 0.38208362, Cur Avg Loss: 1.16110677, Log Avg loss: 1.32601833, Global Avg Loss: 1.48478727, Time: 0.0402 Steps: 92130, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000593, Sample Num: 9488, Cur Loss: 0.54305410, Cur Avg Loss: 1.15840173, Log Avg loss: 1.00069792, Global Avg Loss: 1.48473473, Time: 0.0402 Steps: 92140, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000603, Sample Num: 9648, Cur Loss: 1.12384796, Cur Avg Loss: 1.15519405, Log Avg loss: 0.96497852, Global Avg Loss: 1.48467833, Time: 0.0403 Steps: 92150, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000613, Sample Num: 9808, Cur Loss: 1.06232190, Cur Avg Loss: 1.15489505, Log Avg loss: 1.13686584, Global Avg Loss: 1.48464059, Time: 0.0466 Steps: 92160, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000623, Sample Num: 9968, Cur Loss: 2.10877228, Cur Avg Loss: 1.15741054, Log Avg loss: 1.31160986, Global Avg Loss: 1.48462181, Time: 0.0405 Steps: 92170, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000633, Sample Num: 10128, Cur Loss: 1.49656558, Cur Avg Loss: 1.15586953, Log Avg loss: 1.05986468, Global Avg Loss: 1.48457574, Time: 0.0629 Steps: 92180, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000643, Sample Num: 10288, Cur Loss: 0.54115051, Cur Avg Loss: 1.15642475, Log Avg loss: 1.19156991, Global Avg Loss: 1.48454395, Time: 0.0968 Steps: 92190, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000653, Sample Num: 10448, Cur Loss: 1.32963097, Cur Avg Loss: 1.15509298, Log Avg loss: 1.06946051, Global Avg Loss: 1.48449893, Time: 0.0480 Steps: 92200, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000663, Sample Num: 10608, Cur Loss: 2.13197660, Cur Avg Loss: 1.14963202, Log Avg loss: 0.79303134, Global Avg Loss: 1.48442394, Time: 0.0758 Steps: 92210, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000673, Sample Num: 10768, Cur Loss: 1.64353681, Cur Avg Loss: 1.15196494, Log Avg loss: 1.30663708, Global Avg Loss: 1.48440467, Time: 0.0444 Steps: 92220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000683, Sample Num: 10928, Cur Loss: 0.28409380, Cur Avg Loss: 1.15473854, Log Avg loss: 1.34140210, Global Avg Loss: 1.48438916, Time: 0.0482 Steps: 92230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000693, Sample Num: 11088, Cur Loss: 2.10343242, Cur Avg Loss: 1.15257219, Log Avg loss: 1.00461044, Global Avg Loss: 1.48433715, Time: 0.0402 Steps: 92240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000703, Sample Num: 11248, Cur Loss: 0.74676973, Cur Avg Loss: 1.15213645, Log Avg loss: 1.12193997, Global Avg Loss: 1.48429786, Time: 0.0403 Steps: 92250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000713, Sample Num: 11408, Cur Loss: 0.55269307, Cur Avg Loss: 1.15081780, Log Avg loss: 1.05811625, Global Avg Loss: 1.48425167, Time: 0.0402 Steps: 92260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000723, Sample Num: 11568, Cur Loss: 1.04311371, Cur Avg Loss: 1.14907927, Log Avg loss: 1.02512232, Global Avg Loss: 1.48420191, Time: 0.0402 Steps: 92270, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000733, Sample Num: 11728, Cur Loss: 0.34501705, Cur Avg Loss: 1.14533334, Log Avg loss: 0.87450230, Global Avg Loss: 1.48413584, Time: 0.0403 Steps: 92280, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000743, Sample Num: 11888, Cur Loss: 0.93459332, Cur Avg Loss: 1.14542504, Log Avg loss: 1.15214704, Global Avg Loss: 1.48409987, Time: 0.0402 Steps: 92290, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000753, Sample Num: 12048, Cur Loss: 1.25105512, Cur Avg Loss: 1.13976391, Log Avg loss: 0.71914185, Global Avg Loss: 1.48401699, Time: 0.0402 Steps: 92300, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000763, Sample Num: 12208, Cur Loss: 1.00770056, Cur Avg Loss: 1.14453953, Log Avg loss: 1.50414358, Global Avg Loss: 1.48401917, Time: 0.0402 Steps: 92310, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000773, Sample Num: 12368, Cur Loss: 2.25076652, Cur Avg Loss: 1.14412692, Log Avg loss: 1.11264504, Global Avg Loss: 1.48397894, Time: 0.0403 Steps: 92320, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000783, Sample Num: 12528, Cur Loss: 1.58877158, Cur Avg Loss: 1.14019888, Log Avg loss: 0.83656090, Global Avg Loss: 1.48390882, Time: 0.0403 Steps: 92330, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000793, Sample Num: 12688, Cur Loss: 1.43595743, Cur Avg Loss: 1.13836135, Log Avg loss: 0.99448317, Global Avg Loss: 1.48385582, Time: 0.0402 Steps: 92340, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000803, Sample Num: 12848, Cur Loss: 0.98779756, Cur Avg Loss: 1.14035782, Log Avg loss: 1.29867743, Global Avg Loss: 1.48383577, Time: 0.0402 Steps: 92350, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000813, Sample Num: 13008, Cur Loss: 1.44670439, Cur Avg Loss: 1.14096767, Log Avg loss: 1.18993891, Global Avg Loss: 1.48380395, Time: 0.0403 Steps: 92360, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000823, Sample Num: 13168, Cur Loss: 1.38399088, Cur Avg Loss: 1.14175163, Log Avg loss: 1.20548726, Global Avg Loss: 1.48377382, Time: 0.0402 Steps: 92370, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000833, Sample Num: 13328, Cur Loss: 0.62225497, Cur Avg Loss: 1.13909334, Log Avg loss: 0.92031617, Global Avg Loss: 1.48371282, Time: 0.0402 Steps: 92380, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000843, Sample Num: 13488, Cur Loss: 0.89653051, Cur Avg Loss: 1.13922907, Log Avg loss: 1.15053602, Global Avg Loss: 1.48367676, Time: 0.0402 Steps: 92390, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000853, Sample Num: 13648, Cur Loss: 1.12912130, Cur Avg Loss: 1.13973183, Log Avg loss: 1.18211437, Global Avg Loss: 1.48364412, Time: 0.0402 Steps: 92400, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000863, Sample Num: 13808, Cur Loss: 0.58672249, Cur Avg Loss: 1.13939357, Log Avg loss: 1.11053981, Global Avg Loss: 1.48360375, Time: 0.0402 Steps: 92410, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000873, Sample Num: 13968, Cur Loss: 1.99031639, Cur Avg Loss: 1.14033332, Log Avg loss: 1.22143374, Global Avg Loss: 1.48357538, Time: 0.0401 Steps: 92420, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000883, Sample Num: 14128, Cur Loss: 1.21392286, Cur Avg Loss: 1.14139591, Log Avg loss: 1.23415976, Global Avg Loss: 1.48354840, Time: 0.0402 Steps: 92430, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000893, Sample Num: 14288, Cur Loss: 1.07805717, Cur Avg Loss: 1.14472659, Log Avg loss: 1.43882573, Global Avg Loss: 1.48354356, Time: 0.0402 Steps: 92440, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000903, Sample Num: 14448, Cur Loss: 0.51455104, Cur Avg Loss: 1.14652396, Log Avg loss: 1.30702883, Global Avg Loss: 1.48352447, Time: 0.0403 Steps: 92450, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000913, Sample Num: 14608, Cur Loss: 0.55088484, Cur Avg Loss: 1.15040700, Log Avg loss: 1.50104588, Global Avg Loss: 1.48352636, Time: 0.0402 Steps: 92460, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000923, Sample Num: 14768, Cur Loss: 1.19301224, Cur Avg Loss: 1.14953951, Log Avg loss: 1.07033790, Global Avg Loss: 1.48348168, Time: 0.0402 Steps: 92470, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000933, Sample Num: 14928, Cur Loss: 1.43732333, Cur Avg Loss: 1.15071585, Log Avg loss: 1.25929163, Global Avg Loss: 1.48345744, Time: 0.0403 Steps: 92480, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000943, Sample Num: 15088, Cur Loss: 1.02700031, Cur Avg Loss: 1.15243768, Log Avg loss: 1.31308440, Global Avg Loss: 1.48343902, Time: 0.0402 Steps: 92490, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000953, Sample Num: 15248, Cur Loss: 1.40967512, Cur Avg Loss: 1.15252654, Log Avg loss: 1.16090609, Global Avg Loss: 1.48340415, Time: 0.0871 Steps: 92500, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000963, Sample Num: 15408, Cur Loss: 1.08795261, Cur Avg Loss: 1.15074627, Log Avg loss: 0.98108620, Global Avg Loss: 1.48334985, Time: 0.0431 Steps: 92510, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000973, Sample Num: 15568, Cur Loss: 1.45235014, Cur Avg Loss: 1.15094633, Log Avg loss: 1.17021240, Global Avg Loss: 1.48331600, Time: 0.0676 Steps: 92520, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000983, Sample Num: 15728, Cur Loss: 1.24622393, Cur Avg Loss: 1.15160826, Log Avg loss: 1.21601443, Global Avg Loss: 1.48328712, Time: 0.0593 Steps: 92530, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000993, Sample Num: 15888, Cur Loss: 1.23113704, Cur Avg Loss: 1.15154972, Log Avg loss: 1.14579509, Global Avg Loss: 1.48325065, Time: 0.0765 Steps: 92540, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001003, Sample Num: 16048, Cur Loss: 1.45688891, Cur Avg Loss: 1.15073076, Log Avg loss: 1.06940779, Global Avg Loss: 1.48320593, Time: 0.0792 Steps: 92550, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001013, Sample Num: 16208, Cur Loss: 0.49307650, Cur Avg Loss: 1.15045982, Log Avg loss: 1.12328456, Global Avg Loss: 1.48316704, Time: 0.0405 Steps: 92560, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001023, Sample Num: 16368, Cur Loss: 1.14324057, Cur Avg Loss: 1.15077467, Log Avg loss: 1.18266913, Global Avg Loss: 1.48313458, Time: 0.0452 Steps: 92570, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001033, Sample Num: 16528, Cur Loss: 1.27831578, Cur Avg Loss: 1.14858524, Log Avg loss: 0.92460693, Global Avg Loss: 1.48307425, Time: 0.0577 Steps: 92580, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001043, Sample Num: 16688, Cur Loss: 0.91318393, Cur Avg Loss: 1.14896327, Log Avg loss: 1.18801298, Global Avg Loss: 1.48304239, Time: 0.0404 Steps: 92590, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001053, Sample Num: 16848, Cur Loss: 1.59026241, Cur Avg Loss: 1.14734224, Log Avg loss: 0.97826928, Global Avg Loss: 1.48298788, Time: 0.0403 Steps: 92600, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001063, Sample Num: 17008, Cur Loss: 1.66594672, Cur Avg Loss: 1.15080463, Log Avg loss: 1.51539437, Global Avg Loss: 1.48299137, Time: 0.0403 Steps: 92610, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001073, Sample Num: 17168, Cur Loss: 1.97638941, Cur Avg Loss: 1.15262487, Log Avg loss: 1.34611605, Global Avg Loss: 1.48297660, Time: 0.0403 Steps: 92620, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001083, Sample Num: 17328, Cur Loss: 0.74480212, Cur Avg Loss: 1.15217967, Log Avg loss: 1.10441043, Global Avg Loss: 1.48293573, Time: 0.0403 Steps: 92630, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001093, Sample Num: 17488, Cur Loss: 1.12960613, Cur Avg Loss: 1.15212910, Log Avg loss: 1.14665193, Global Avg Loss: 1.48289943, Time: 0.0403 Steps: 92640, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001103, Sample Num: 17648, Cur Loss: 1.13683987, Cur Avg Loss: 1.14944759, Log Avg loss: 0.85635852, Global Avg Loss: 1.48283180, Time: 0.0402 Steps: 92650, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001113, Sample Num: 17808, Cur Loss: 1.56981230, Cur Avg Loss: 1.14967019, Log Avg loss: 1.17422297, Global Avg Loss: 1.48279850, Time: 0.0404 Steps: 92660, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001123, Sample Num: 17968, Cur Loss: 0.50967622, Cur Avg Loss: 1.14765294, Log Avg loss: 0.92313342, Global Avg Loss: 1.48273810, Time: 0.0403 Steps: 92670, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001133, Sample Num: 18128, Cur Loss: 1.20481813, Cur Avg Loss: 1.14678272, Log Avg loss: 1.04905698, Global Avg Loss: 1.48269131, Time: 0.0402 Steps: 92680, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001143, Sample Num: 18288, Cur Loss: 1.33252239, Cur Avg Loss: 1.14624334, Log Avg loss: 1.08513129, Global Avg Loss: 1.48264842, Time: 0.0403 Steps: 92690, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001153, Sample Num: 18448, Cur Loss: 2.20378304, Cur Avg Loss: 1.14743242, Log Avg loss: 1.28334439, Global Avg Loss: 1.48262692, Time: 0.0403 Steps: 92700, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001163, Sample Num: 18608, Cur Loss: 1.85337543, Cur Avg Loss: 1.14661711, Log Avg loss: 1.05261161, Global Avg Loss: 1.48258054, Time: 0.0403 Steps: 92710, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001173, Sample Num: 18768, Cur Loss: 1.04375458, Cur Avg Loss: 1.14781795, Log Avg loss: 1.28747530, Global Avg Loss: 1.48255949, Time: 0.0402 Steps: 92720, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001183, Sample Num: 18928, Cur Loss: 0.66648465, Cur Avg Loss: 1.14783160, Log Avg loss: 1.14943349, Global Avg Loss: 1.48252357, Time: 0.0402 Steps: 92730, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001193, Sample Num: 19088, Cur Loss: 0.41555929, Cur Avg Loss: 1.14883925, Log Avg loss: 1.26804332, Global Avg Loss: 1.48250044, Time: 0.0403 Steps: 92740, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001203, Sample Num: 19248, Cur Loss: 2.74343872, Cur Avg Loss: 1.14986428, Log Avg loss: 1.27215058, Global Avg Loss: 1.48247776, Time: 0.0402 Steps: 92750, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001213, Sample Num: 19408, Cur Loss: 1.75563943, Cur Avg Loss: 1.15236627, Log Avg loss: 1.45335561, Global Avg Loss: 1.48247462, Time: 0.0402 Steps: 92760, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001223, Sample Num: 19568, Cur Loss: 0.85311902, Cur Avg Loss: 1.15164295, Log Avg loss: 1.06390498, Global Avg Loss: 1.48242951, Time: 0.0402 Steps: 92770, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001233, Sample Num: 19728, Cur Loss: 1.41498065, Cur Avg Loss: 1.15284098, Log Avg loss: 1.29936007, Global Avg Loss: 1.48240977, Time: 0.0402 Steps: 92780, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001243, Sample Num: 19888, Cur Loss: 0.34378719, Cur Avg Loss: 1.15085546, Log Avg loss: 0.90603993, Global Avg Loss: 1.48234766, Time: 0.0402 Steps: 92790, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001253, Sample Num: 20048, Cur Loss: 1.76884341, Cur Avg Loss: 1.15145725, Log Avg loss: 1.22626066, Global Avg Loss: 1.48232006, Time: 0.0402 Steps: 92800, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001263, Sample Num: 20208, Cur Loss: 0.51740640, Cur Avg Loss: 1.14837254, Log Avg loss: 0.76185850, Global Avg Loss: 1.48224244, Time: 0.0402 Steps: 92810, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001273, Sample Num: 20368, Cur Loss: 0.82752895, Cur Avg Loss: 1.14922500, Log Avg loss: 1.25689014, Global Avg Loss: 1.48221816, Time: 0.0402 Steps: 92820, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001283, Sample Num: 20528, Cur Loss: 0.79489064, Cur Avg Loss: 1.14858979, Log Avg loss: 1.06772767, Global Avg Loss: 1.48217351, Time: 0.0402 Steps: 92830, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001293, Sample Num: 20688, Cur Loss: 1.63150764, Cur Avg Loss: 1.14689677, Log Avg loss: 0.92968178, Global Avg Loss: 1.48211400, Time: 0.0401 Steps: 92840, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001303, Sample Num: 20848, Cur Loss: 1.92218769, Cur Avg Loss: 1.14633334, Log Avg loss: 1.07348239, Global Avg Loss: 1.48206999, Time: 0.0402 Steps: 92850, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001313, Sample Num: 21008, Cur Loss: 0.36905083, Cur Avg Loss: 1.14726336, Log Avg loss: 1.26844491, Global Avg Loss: 1.48204698, Time: 0.0402 Steps: 92860, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001323, Sample Num: 21168, Cur Loss: 2.41858387, Cur Avg Loss: 1.14823997, Log Avg loss: 1.27646819, Global Avg Loss: 1.48202484, Time: 0.0442 Steps: 92870, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001333, Sample Num: 21328, Cur Loss: 1.28601074, Cur Avg Loss: 1.14662859, Log Avg loss: 0.93344428, Global Avg Loss: 1.48196578, Time: 0.0510 Steps: 92880, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001343, Sample Num: 21488, Cur Loss: 1.13429427, Cur Avg Loss: 1.14502055, Log Avg loss: 0.93066768, Global Avg Loss: 1.48190643, Time: 0.0741 Steps: 92890, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001353, Sample Num: 21648, Cur Loss: 2.06132102, Cur Avg Loss: 1.14498073, Log Avg loss: 1.13963337, Global Avg Loss: 1.48186959, Time: 0.0461 Steps: 92900, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001363, Sample Num: 21808, Cur Loss: 0.95318735, Cur Avg Loss: 1.14688961, Log Avg loss: 1.40516145, Global Avg Loss: 1.48186133, Time: 0.0822 Steps: 92910, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001373, Sample Num: 21968, Cur Loss: 2.38527536, Cur Avg Loss: 1.14762677, Log Avg loss: 1.24810116, Global Avg Loss: 1.48183618, Time: 0.0597 Steps: 92920, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001383, Sample Num: 22128, Cur Loss: 1.57184350, Cur Avg Loss: 1.14810085, Log Avg loss: 1.21319246, Global Avg Loss: 1.48180727, Time: 0.0405 Steps: 92930, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001393, Sample Num: 22288, Cur Loss: 1.23714745, Cur Avg Loss: 1.14633282, Log Avg loss: 0.90181359, Global Avg Loss: 1.48174486, Time: 0.0486 Steps: 92940, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001403, Sample Num: 22448, Cur Loss: 2.02099991, Cur Avg Loss: 1.14604666, Log Avg loss: 1.10618530, Global Avg Loss: 1.48170446, Time: 0.0405 Steps: 92950, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001413, Sample Num: 22608, Cur Loss: 1.61097956, Cur Avg Loss: 1.14364911, Log Avg loss: 0.80727229, Global Avg Loss: 1.48163191, Time: 0.0718 Steps: 92960, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001423, Sample Num: 22768, Cur Loss: 1.41104841, Cur Avg Loss: 1.14493430, Log Avg loss: 1.32653121, Global Avg Loss: 1.48161522, Time: 0.0402 Steps: 92970, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001433, Sample Num: 22928, Cur Loss: 2.30636978, Cur Avg Loss: 1.14733894, Log Avg loss: 1.48951958, Global Avg Loss: 1.48161607, Time: 0.0402 Steps: 92980, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001443, Sample Num: 23088, Cur Loss: 0.90505266, Cur Avg Loss: 1.14606287, Log Avg loss: 0.96320252, Global Avg Loss: 1.48156032, Time: 0.0402 Steps: 92990, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001453, Sample Num: 23248, Cur Loss: 1.21014428, Cur Avg Loss: 1.14711078, Log Avg loss: 1.29832433, Global Avg Loss: 1.48154062, Time: 0.0403 Steps: 93000, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001463, Sample Num: 23408, Cur Loss: 1.85272503, Cur Avg Loss: 1.15019615, Log Avg loss: 1.59850034, Global Avg Loss: 1.48155320, Time: 0.0404 Steps: 93010, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001473, Sample Num: 23568, Cur Loss: 0.92703843, Cur Avg Loss: 1.15080804, Log Avg loss: 1.24032736, Global Avg Loss: 1.48152726, Time: 0.0402 Steps: 93020, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001483, Sample Num: 23728, Cur Loss: 1.18597221, Cur Avg Loss: 1.15042588, Log Avg loss: 1.09413419, Global Avg Loss: 1.48148562, Time: 0.0402 Steps: 93030, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001493, Sample Num: 23888, Cur Loss: 0.75906658, Cur Avg Loss: 1.14916845, Log Avg loss: 0.96269042, Global Avg Loss: 1.48142986, Time: 0.0403 Steps: 93040, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001503, Sample Num: 24048, Cur Loss: 1.68604624, Cur Avg Loss: 1.14988155, Log Avg loss: 1.25634772, Global Avg Loss: 1.48140567, Time: 0.0402 Steps: 93050, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001513, Sample Num: 24208, Cur Loss: 0.50244629, Cur Avg Loss: 1.14944523, Log Avg loss: 1.08386685, Global Avg Loss: 1.48136295, Time: 0.0402 Steps: 93060, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001523, Sample Num: 24368, Cur Loss: 0.94037735, Cur Avg Loss: 1.14720081, Log Avg loss: 0.80762047, Global Avg Loss: 1.48129056, Time: 0.0402 Steps: 93070, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001533, Sample Num: 24528, Cur Loss: 2.14122128, Cur Avg Loss: 1.14833128, Log Avg loss: 1.32050194, Global Avg Loss: 1.48127329, Time: 0.0403 Steps: 93080, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001543, Sample Num: 24688, Cur Loss: 0.92847264, Cur Avg Loss: 1.14944043, Log Avg loss: 1.31947299, Global Avg Loss: 1.48125591, Time: 0.0402 Steps: 93090, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001553, Sample Num: 24848, Cur Loss: 1.25019920, Cur Avg Loss: 1.14928939, Log Avg loss: 1.12598293, Global Avg Loss: 1.48121775, Time: 0.0402 Steps: 93100, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001563, Sample Num: 25008, Cur Loss: 1.21421206, Cur Avg Loss: 1.14951200, Log Avg loss: 1.18408318, Global Avg Loss: 1.48118584, Time: 0.0402 Steps: 93110, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001573, Sample Num: 25168, Cur Loss: 1.07291722, Cur Avg Loss: 1.15122598, Log Avg loss: 1.41912158, Global Avg Loss: 1.48117917, Time: 0.0402 Steps: 93120, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001583, Sample Num: 25328, Cur Loss: 1.32726169, Cur Avg Loss: 1.15141759, Log Avg loss: 1.18155755, Global Avg Loss: 1.48114700, Time: 0.0402 Steps: 93130, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001593, Sample Num: 25488, Cur Loss: 0.61861962, Cur Avg Loss: 1.15083056, Log Avg loss: 1.05790485, Global Avg Loss: 1.48110156, Time: 0.0402 Steps: 93140, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001603, Sample Num: 25648, Cur Loss: 2.45006633, Cur Avg Loss: 1.15336698, Log Avg loss: 1.55741744, Global Avg Loss: 1.48110975, Time: 0.0402 Steps: 93150, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001613, Sample Num: 25808, Cur Loss: 0.65942502, Cur Avg Loss: 1.15277442, Log Avg loss: 1.05778770, Global Avg Loss: 1.48106431, Time: 0.0403 Steps: 93160, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001623, Sample Num: 25968, Cur Loss: 0.58810419, Cur Avg Loss: 1.15161172, Log Avg loss: 0.96406778, Global Avg Loss: 1.48100882, Time: 0.0402 Steps: 93170, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001633, Sample Num: 26128, Cur Loss: 1.37802649, Cur Avg Loss: 1.15124247, Log Avg loss: 1.09131346, Global Avg Loss: 1.48096700, Time: 0.0402 Steps: 93180, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001643, Sample Num: 26288, Cur Loss: 0.94665557, Cur Avg Loss: 1.15046480, Log Avg loss: 1.02347150, Global Avg Loss: 1.48091790, Time: 0.0403 Steps: 93190, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001653, Sample Num: 26448, Cur Loss: 1.54330540, Cur Avg Loss: 1.15143703, Log Avg loss: 1.31117358, Global Avg Loss: 1.48089969, Time: 0.0403 Steps: 93200, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001663, Sample Num: 26608, Cur Loss: 0.61027551, Cur Avg Loss: 1.15118445, Log Avg loss: 1.10943373, Global Avg Loss: 1.48085984, Time: 0.0402 Steps: 93210, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001673, Sample Num: 26768, Cur Loss: 1.16881144, Cur Avg Loss: 1.15065493, Log Avg loss: 1.06259512, Global Avg Loss: 1.48081497, Time: 0.0402 Steps: 93220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001683, Sample Num: 26928, Cur Loss: 2.01242352, Cur Avg Loss: 1.15114100, Log Avg loss: 1.23246036, Global Avg Loss: 1.48078833, Time: 0.0574 Steps: 93230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001693, Sample Num: 27088, Cur Loss: 0.86217624, Cur Avg Loss: 1.15099104, Log Avg loss: 1.12575284, Global Avg Loss: 1.48075025, Time: 0.0656 Steps: 93240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001703, Sample Num: 27248, Cur Loss: 0.75659710, Cur Avg Loss: 1.15069137, Log Avg loss: 1.09995772, Global Avg Loss: 1.48070942, Time: 0.0406 Steps: 93250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001713, Sample Num: 27408, Cur Loss: 0.79337096, Cur Avg Loss: 1.15088118, Log Avg loss: 1.18320585, Global Avg Loss: 1.48067752, Time: 0.0499 Steps: 93260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001723, Sample Num: 27568, Cur Loss: 0.93130761, Cur Avg Loss: 1.15034865, Log Avg loss: 1.05912628, Global Avg Loss: 1.48063232, Time: 0.0689 Steps: 93270, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001733, Sample Num: 27728, Cur Loss: 0.66799080, Cur Avg Loss: 1.15059103, Log Avg loss: 1.19235328, Global Avg Loss: 1.48060142, Time: 0.0840 Steps: 93280, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001743, Sample Num: 27888, Cur Loss: 0.35972664, Cur Avg Loss: 1.14918041, Log Avg loss: 0.90471922, Global Avg Loss: 1.48053969, Time: 0.0459 Steps: 93290, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001753, Sample Num: 28048, Cur Loss: 0.98779190, Cur Avg Loss: 1.14820464, Log Avg loss: 0.97812911, Global Avg Loss: 1.48048584, Time: 0.0554 Steps: 93300, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001763, Sample Num: 28208, Cur Loss: 1.14534903, Cur Avg Loss: 1.14655354, Log Avg loss: 0.85711578, Global Avg Loss: 1.48041903, Time: 0.0614 Steps: 93310, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001773, Sample Num: 28368, Cur Loss: 1.60514271, Cur Avg Loss: 1.14612431, Log Avg loss: 1.07045102, Global Avg Loss: 1.48037510, Time: 0.0470 Steps: 93320, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001783, Sample Num: 28528, Cur Loss: 1.50094128, Cur Avg Loss: 1.14687973, Log Avg loss: 1.28081511, Global Avg Loss: 1.48035372, Time: 0.0402 Steps: 93330, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001793, Sample Num: 28688, Cur Loss: 0.96984971, Cur Avg Loss: 1.14709987, Log Avg loss: 1.18635037, Global Avg Loss: 1.48032222, Time: 0.0402 Steps: 93340, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001803, Sample Num: 28848, Cur Loss: 1.06417775, Cur Avg Loss: 1.14830963, Log Avg loss: 1.36522061, Global Avg Loss: 1.48030989, Time: 0.0402 Steps: 93350, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001813, Sample Num: 29008, Cur Loss: 2.15080476, Cur Avg Loss: 1.15006751, Log Avg loss: 1.46701199, Global Avg Loss: 1.48030846, Time: 0.0403 Steps: 93360, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001823, Sample Num: 29168, Cur Loss: 0.73495692, Cur Avg Loss: 1.15023855, Log Avg loss: 1.18124840, Global Avg Loss: 1.48027644, Time: 0.0403 Steps: 93370, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001833, Sample Num: 29328, Cur Loss: 1.11586988, Cur Avg Loss: 1.14909567, Log Avg loss: 0.94074942, Global Avg Loss: 1.48021866, Time: 0.0403 Steps: 93380, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001843, Sample Num: 29488, Cur Loss: 1.35318601, Cur Avg Loss: 1.14779587, Log Avg loss: 0.90954228, Global Avg Loss: 1.48015755, Time: 0.0402 Steps: 93390, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001853, Sample Num: 29648, Cur Loss: 0.79955846, Cur Avg Loss: 1.14844836, Log Avg loss: 1.26870287, Global Avg Loss: 1.48013491, Time: 0.0403 Steps: 93400, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001863, Sample Num: 29808, Cur Loss: 0.77883679, Cur Avg Loss: 1.14901981, Log Avg loss: 1.25490891, Global Avg Loss: 1.48011080, Time: 0.0403 Steps: 93410, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001873, Sample Num: 29968, Cur Loss: 1.63803554, Cur Avg Loss: 1.14731411, Log Avg loss: 0.82954125, Global Avg Loss: 1.48004116, Time: 0.0403 Steps: 93420, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001883, Sample Num: 30128, Cur Loss: 1.05034041, Cur Avg Loss: 1.14610162, Log Avg loss: 0.91900219, Global Avg Loss: 1.47998111, Time: 0.0404 Steps: 93430, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001893, Sample Num: 30288, Cur Loss: 1.27493882, Cur Avg Loss: 1.14538987, Log Avg loss: 1.01136846, Global Avg Loss: 1.47993096, Time: 0.0403 Steps: 93440, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001903, Sample Num: 30448, Cur Loss: 0.88888437, Cur Avg Loss: 1.14518344, Log Avg loss: 1.10610546, Global Avg Loss: 1.47989096, Time: 0.0403 Steps: 93450, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001913, Sample Num: 30608, Cur Loss: 1.22303700, Cur Avg Loss: 1.14499351, Log Avg loss: 1.10885107, Global Avg Loss: 1.47985126, Time: 0.0402 Steps: 93460, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001923, Sample Num: 30768, Cur Loss: 0.75946200, Cur Avg Loss: 1.14500343, Log Avg loss: 1.14690099, Global Avg Loss: 1.47981564, Time: 0.0403 Steps: 93470, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001933, Sample Num: 30928, Cur Loss: 1.99496031, Cur Avg Loss: 1.14535965, Log Avg loss: 1.21385927, Global Avg Loss: 1.47978719, Time: 0.0402 Steps: 93480, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001943, Sample Num: 31088, Cur Loss: 1.18611479, Cur Avg Loss: 1.14616274, Log Avg loss: 1.30140129, Global Avg Loss: 1.47976810, Time: 0.0402 Steps: 93490, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001953, Sample Num: 31248, Cur Loss: 1.57166135, Cur Avg Loss: 1.14532488, Log Avg loss: 0.98252868, Global Avg Loss: 1.47971492, Time: 0.0402 Steps: 93500, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001963, Sample Num: 31408, Cur Loss: 0.58137655, Cur Avg Loss: 1.14498734, Log Avg loss: 1.07906575, Global Avg Loss: 1.47967208, Time: 0.0403 Steps: 93510, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001973, Sample Num: 31568, Cur Loss: 1.10903883, Cur Avg Loss: 1.14537519, Log Avg loss: 1.22150936, Global Avg Loss: 1.47964447, Time: 0.0403 Steps: 93520, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001983, Sample Num: 31728, Cur Loss: 0.65597504, Cur Avg Loss: 1.14488261, Log Avg loss: 1.04769772, Global Avg Loss: 1.47959829, Time: 0.0402 Steps: 93530, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001993, Sample Num: 31888, Cur Loss: 2.21421480, Cur Avg Loss: 1.14573037, Log Avg loss: 1.31383978, Global Avg Loss: 1.47958057, Time: 0.0402 Steps: 93540, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002003, Sample Num: 32048, Cur Loss: 1.05720747, Cur Avg Loss: 1.14609047, Log Avg loss: 1.21785921, Global Avg Loss: 1.47955259, Time: 0.0401 Steps: 93550, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002013, Sample Num: 32208, Cur Loss: 1.14993000, Cur Avg Loss: 1.14795206, Log Avg loss: 1.52082783, Global Avg Loss: 1.47955700, Time: 0.0403 Steps: 93560, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002023, Sample Num: 32368, Cur Loss: 0.79659534, Cur Avg Loss: 1.14706087, Log Avg loss: 0.96766396, Global Avg Loss: 1.47950230, Time: 0.0466 Steps: 93570, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002033, Sample Num: 32528, Cur Loss: 1.12536299, Cur Avg Loss: 1.14693529, Log Avg loss: 1.12153205, Global Avg Loss: 1.47946404, Time: 0.0403 Steps: 93580, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002043, Sample Num: 32688, Cur Loss: 1.74857998, Cur Avg Loss: 1.14657905, Log Avg loss: 1.07415549, Global Avg Loss: 1.47942074, Time: 0.0490 Steps: 93590, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002053, Sample Num: 32848, Cur Loss: 1.73662043, Cur Avg Loss: 1.14723921, Log Avg loss: 1.28210850, Global Avg Loss: 1.47939966, Time: 0.0405 Steps: 93600, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002063, Sample Num: 33008, Cur Loss: 0.92948008, Cur Avg Loss: 1.14601607, Log Avg loss: 0.89490648, Global Avg Loss: 1.47933722, Time: 0.0523 Steps: 93610, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002073, Sample Num: 33168, Cur Loss: 1.69120741, Cur Avg Loss: 1.14511399, Log Avg loss: 0.95901502, Global Avg Loss: 1.47928164, Time: 0.0811 Steps: 93620, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002083, Sample Num: 33328, Cur Loss: 2.08153677, Cur Avg Loss: 1.14668991, Log Avg loss: 1.47337714, Global Avg Loss: 1.47928101, Time: 0.0674 Steps: 93630, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002093, Sample Num: 33488, Cur Loss: 1.83662188, Cur Avg Loss: 1.14665709, Log Avg loss: 1.13982177, Global Avg Loss: 1.47924476, Time: 0.0815 Steps: 93640, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002103, Sample Num: 33648, Cur Loss: 1.22972834, Cur Avg Loss: 1.14560021, Log Avg loss: 0.92439546, Global Avg Loss: 1.47918551, Time: 0.0402 Steps: 93650, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002113, Sample Num: 33808, Cur Loss: 1.32392764, Cur Avg Loss: 1.14623160, Log Avg loss: 1.27901239, Global Avg Loss: 1.47916414, Time: 0.0403 Steps: 93660, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002123, Sample Num: 33968, Cur Loss: 0.95239943, Cur Avg Loss: 1.14664554, Log Avg loss: 1.23411146, Global Avg Loss: 1.47913798, Time: 0.0403 Steps: 93670, Updated lr: 0.000012 ***** Running evaluation checkpoint-93676 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-93676 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.852034, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.164744, "eval_total_loss": 818.815323, "eval_mae": 0.914847, "eval_mse": 1.164749, "eval_r2": 0.259609, "eval_sp_statistic": 0.457152, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.568194, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.194484, "test_total_loss": 599.630825, "test_mae": 0.946124, "test_mse": 1.194626, "test_r2": 0.228978, "test_sp_statistic": 0.478575, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.611372, "test_ps_pvalue": 0.0, "lr": 1.2113798008534852e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4791203970702378, "train_cur_epoch_loss": 2441.5564401447773, "train_cur_epoch_avg_loss": 1.1468090371746253, "train_cur_epoch_time": 95.8520336151123, "train_cur_epoch_avg_time": 0.04502209188121761, "epoch": 44, "step": 93676} ################################################## Training, Epoch: 0045, Batch: 000004, Sample Num: 64, Cur Loss: 1.27121294, Cur Avg Loss: 1.41669437, Log Avg loss: 1.28947300, Global Avg Loss: 1.47911773, Time: 0.0637 Steps: 93680, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000014, Sample Num: 224, Cur Loss: 1.13591814, Cur Avg Loss: 1.15520709, Log Avg loss: 1.05061218, Global Avg Loss: 1.47907200, Time: 0.0402 Steps: 93690, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000024, Sample Num: 384, Cur Loss: 0.67234886, Cur Avg Loss: 1.05872467, Log Avg loss: 0.92364929, Global Avg Loss: 1.47901272, Time: 0.0402 Steps: 93700, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000034, Sample Num: 544, Cur Loss: 1.59162390, Cur Avg Loss: 1.02499532, Log Avg loss: 0.94404488, Global Avg Loss: 1.47895563, Time: 0.0402 Steps: 93710, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000044, Sample Num: 704, Cur Loss: 1.51019466, Cur Avg Loss: 1.07108614, Log Avg loss: 1.22779493, Global Avg Loss: 1.47892883, Time: 0.0402 Steps: 93720, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000054, Sample Num: 864, Cur Loss: 1.26739049, Cur Avg Loss: 1.08815416, Log Avg loss: 1.16325344, Global Avg Loss: 1.47889515, Time: 0.0402 Steps: 93730, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000064, Sample Num: 1024, Cur Loss: 0.66645169, Cur Avg Loss: 1.07267720, Log Avg loss: 0.98910160, Global Avg Loss: 1.47884290, Time: 0.0402 Steps: 93740, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000074, Sample Num: 1184, Cur Loss: 0.52175069, Cur Avg Loss: 1.08566544, Log Avg loss: 1.16879017, Global Avg Loss: 1.47880983, Time: 0.0402 Steps: 93750, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000084, Sample Num: 1344, Cur Loss: 0.51728702, Cur Avg Loss: 1.07843947, Log Avg loss: 1.02496734, Global Avg Loss: 1.47876143, Time: 0.0402 Steps: 93760, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000094, Sample Num: 1504, Cur Loss: 0.65130067, Cur Avg Loss: 1.06467232, Log Avg loss: 0.94902820, Global Avg Loss: 1.47870493, Time: 0.0403 Steps: 93770, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000104, Sample Num: 1664, Cur Loss: 1.41108906, Cur Avg Loss: 1.09432170, Log Avg loss: 1.37302589, Global Avg Loss: 1.47869366, Time: 0.0403 Steps: 93780, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000114, Sample Num: 1824, Cur Loss: 0.45890725, Cur Avg Loss: 1.09904940, Log Avg loss: 1.14821747, Global Avg Loss: 1.47865843, Time: 0.0402 Steps: 93790, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000124, Sample Num: 1984, Cur Loss: 0.43091524, Cur Avg Loss: 1.10742438, Log Avg loss: 1.20289921, Global Avg Loss: 1.47862903, Time: 0.0403 Steps: 93800, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000134, Sample Num: 2144, Cur Loss: 1.31307888, Cur Avg Loss: 1.11914371, Log Avg loss: 1.26446336, Global Avg Loss: 1.47860620, Time: 0.0403 Steps: 93810, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000144, Sample Num: 2304, Cur Loss: 0.40475845, Cur Avg Loss: 1.11543704, Log Avg loss: 1.06576768, Global Avg Loss: 1.47856220, Time: 0.0403 Steps: 93820, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000154, Sample Num: 2464, Cur Loss: 0.74731910, Cur Avg Loss: 1.10296534, Log Avg loss: 0.92337283, Global Avg Loss: 1.47850303, Time: 0.0402 Steps: 93830, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000164, Sample Num: 2624, Cur Loss: 1.49255300, Cur Avg Loss: 1.11691977, Log Avg loss: 1.33181801, Global Avg Loss: 1.47848740, Time: 0.0402 Steps: 93840, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000174, Sample Num: 2784, Cur Loss: 1.74101782, Cur Avg Loss: 1.12022940, Log Avg loss: 1.17450736, Global Avg Loss: 1.47845501, Time: 0.0403 Steps: 93850, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000184, Sample Num: 2944, Cur Loss: 1.68537247, Cur Avg Loss: 1.11561479, Log Avg loss: 1.03532055, Global Avg Loss: 1.47840779, Time: 0.0402 Steps: 93860, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000194, Sample Num: 3104, Cur Loss: 0.77314228, Cur Avg Loss: 1.12019349, Log Avg loss: 1.20444160, Global Avg Loss: 1.47837861, Time: 0.0402 Steps: 93870, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000204, Sample Num: 3264, Cur Loss: 2.03635669, Cur Avg Loss: 1.12133542, Log Avg loss: 1.14348881, Global Avg Loss: 1.47834294, Time: 0.0402 Steps: 93880, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000214, Sample Num: 3424, Cur Loss: 1.50448465, Cur Avg Loss: 1.12640217, Log Avg loss: 1.22976397, Global Avg Loss: 1.47831646, Time: 0.0402 Steps: 93890, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000224, Sample Num: 3584, Cur Loss: 1.23949802, Cur Avg Loss: 1.12743450, Log Avg loss: 1.14952636, Global Avg Loss: 1.47828144, Time: 0.0403 Steps: 93900, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000234, Sample Num: 3744, Cur Loss: 0.91893542, Cur Avg Loss: 1.13677937, Log Avg loss: 1.34610428, Global Avg Loss: 1.47826737, Time: 0.0402 Steps: 93910, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000244, Sample Num: 3904, Cur Loss: 0.48179737, Cur Avg Loss: 1.13952300, Log Avg loss: 1.20372393, Global Avg Loss: 1.47823814, Time: 0.0402 Steps: 93920, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000254, Sample Num: 4064, Cur Loss: 0.85450739, Cur Avg Loss: 1.14361526, Log Avg loss: 1.24346662, Global Avg Loss: 1.47821314, Time: 0.0402 Steps: 93930, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000264, Sample Num: 4224, Cur Loss: 0.43174449, Cur Avg Loss: 1.14659602, Log Avg loss: 1.22230724, Global Avg Loss: 1.47818590, Time: 0.0402 Steps: 93940, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000274, Sample Num: 4384, Cur Loss: 0.54603195, Cur Avg Loss: 1.13911971, Log Avg loss: 0.94174498, Global Avg Loss: 1.47812880, Time: 0.0401 Steps: 93950, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000284, Sample Num: 4544, Cur Loss: 0.52011156, Cur Avg Loss: 1.13421824, Log Avg loss: 0.99991806, Global Avg Loss: 1.47807791, Time: 0.0469 Steps: 93960, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000294, Sample Num: 4704, Cur Loss: 1.80642354, Cur Avg Loss: 1.12869519, Log Avg loss: 0.97184067, Global Avg Loss: 1.47802404, Time: 0.0853 Steps: 93970, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000304, Sample Num: 4864, Cur Loss: 0.53040457, Cur Avg Loss: 1.11724355, Log Avg loss: 0.78056513, Global Avg Loss: 1.47794982, Time: 0.0541 Steps: 93980, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000314, Sample Num: 5024, Cur Loss: 0.67547345, Cur Avg Loss: 1.11672136, Log Avg loss: 1.10084683, Global Avg Loss: 1.47790970, Time: 0.0572 Steps: 93990, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000324, Sample Num: 5184, Cur Loss: 1.09699309, Cur Avg Loss: 1.11872730, Log Avg loss: 1.18171390, Global Avg Loss: 1.47787819, Time: 0.0552 Steps: 94000, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000334, Sample Num: 5344, Cur Loss: 1.30557847, Cur Avg Loss: 1.13056093, Log Avg loss: 1.51397040, Global Avg Loss: 1.47788203, Time: 0.0403 Steps: 94010, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000344, Sample Num: 5504, Cur Loss: 0.59049833, Cur Avg Loss: 1.13439064, Log Avg loss: 1.26230309, Global Avg Loss: 1.47785910, Time: 0.0780 Steps: 94020, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000354, Sample Num: 5664, Cur Loss: 1.42583215, Cur Avg Loss: 1.13680501, Log Avg loss: 1.21985940, Global Avg Loss: 1.47783166, Time: 0.0590 Steps: 94030, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000364, Sample Num: 5824, Cur Loss: 1.22780800, Cur Avg Loss: 1.13453938, Log Avg loss: 1.05433590, Global Avg Loss: 1.47778663, Time: 0.0403 Steps: 94040, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000374, Sample Num: 5984, Cur Loss: 1.21277380, Cur Avg Loss: 1.12664777, Log Avg loss: 0.83939331, Global Avg Loss: 1.47771875, Time: 0.0403 Steps: 94050, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000384, Sample Num: 6144, Cur Loss: 1.43741083, Cur Avg Loss: 1.13216041, Log Avg loss: 1.33833312, Global Avg Loss: 1.47770393, Time: 0.0403 Steps: 94060, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000394, Sample Num: 6304, Cur Loss: 2.42952538, Cur Avg Loss: 1.12866261, Log Avg loss: 0.99434695, Global Avg Loss: 1.47765255, Time: 0.0402 Steps: 94070, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000404, Sample Num: 6464, Cur Loss: 1.14846921, Cur Avg Loss: 1.12690266, Log Avg loss: 1.05756085, Global Avg Loss: 1.47760790, Time: 0.0403 Steps: 94080, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000414, Sample Num: 6624, Cur Loss: 0.98659098, Cur Avg Loss: 1.12563821, Log Avg loss: 1.07455419, Global Avg Loss: 1.47756506, Time: 0.0404 Steps: 94090, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000424, Sample Num: 6784, Cur Loss: 0.53187680, Cur Avg Loss: 1.12119838, Log Avg loss: 0.93738942, Global Avg Loss: 1.47750766, Time: 0.0402 Steps: 94100, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000434, Sample Num: 6944, Cur Loss: 0.63469356, Cur Avg Loss: 1.11865134, Log Avg loss: 1.01065695, Global Avg Loss: 1.47745805, Time: 0.0403 Steps: 94110, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000444, Sample Num: 7104, Cur Loss: 1.23382354, Cur Avg Loss: 1.12238892, Log Avg loss: 1.28459979, Global Avg Loss: 1.47743756, Time: 0.0402 Steps: 94120, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000454, Sample Num: 7264, Cur Loss: 1.27611828, Cur Avg Loss: 1.12276172, Log Avg loss: 1.13931432, Global Avg Loss: 1.47740164, Time: 0.0402 Steps: 94130, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000464, Sample Num: 7424, Cur Loss: 1.22636914, Cur Avg Loss: 1.13121889, Log Avg loss: 1.51517434, Global Avg Loss: 1.47740565, Time: 0.0402 Steps: 94140, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000474, Sample Num: 7584, Cur Loss: 0.82162130, Cur Avg Loss: 1.13745159, Log Avg loss: 1.42664892, Global Avg Loss: 1.47740026, Time: 0.0402 Steps: 94150, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000484, Sample Num: 7744, Cur Loss: 0.75349957, Cur Avg Loss: 1.13972464, Log Avg loss: 1.24746712, Global Avg Loss: 1.47737584, Time: 0.0403 Steps: 94160, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000494, Sample Num: 7904, Cur Loss: 1.16204262, Cur Avg Loss: 1.14017392, Log Avg loss: 1.16191927, Global Avg Loss: 1.47734234, Time: 0.0402 Steps: 94170, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000504, Sample Num: 8064, Cur Loss: 1.71423483, Cur Avg Loss: 1.14481434, Log Avg loss: 1.37405069, Global Avg Loss: 1.47733137, Time: 0.0403 Steps: 94180, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000514, Sample Num: 8224, Cur Loss: 1.00717258, Cur Avg Loss: 1.14285132, Log Avg loss: 1.04391516, Global Avg Loss: 1.47728536, Time: 0.0402 Steps: 94190, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000524, Sample Num: 8384, Cur Loss: 0.93499798, Cur Avg Loss: 1.13741200, Log Avg loss: 0.85783105, Global Avg Loss: 1.47721960, Time: 0.0402 Steps: 94200, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000534, Sample Num: 8544, Cur Loss: 0.40662676, Cur Avg Loss: 1.13502690, Log Avg loss: 1.01004773, Global Avg Loss: 1.47717001, Time: 0.0402 Steps: 94210, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000544, Sample Num: 8704, Cur Loss: 1.88110137, Cur Avg Loss: 1.13795929, Log Avg loss: 1.29454908, Global Avg Loss: 1.47715063, Time: 0.0402 Steps: 94220, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000554, Sample Num: 8864, Cur Loss: 0.79837930, Cur Avg Loss: 1.13638257, Log Avg loss: 1.05060873, Global Avg Loss: 1.47710536, Time: 0.0402 Steps: 94230, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000564, Sample Num: 9024, Cur Loss: 0.96866429, Cur Avg Loss: 1.13186381, Log Avg loss: 0.88152471, Global Avg Loss: 1.47704216, Time: 0.0402 Steps: 94240, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000574, Sample Num: 9184, Cur Loss: 0.95830250, Cur Avg Loss: 1.12973315, Log Avg loss: 1.00956379, Global Avg Loss: 1.47699256, Time: 0.0402 Steps: 94250, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000584, Sample Num: 9344, Cur Loss: 1.07359457, Cur Avg Loss: 1.13034381, Log Avg loss: 1.16539567, Global Avg Loss: 1.47695951, Time: 0.0402 Steps: 94260, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000594, Sample Num: 9504, Cur Loss: 0.55949962, Cur Avg Loss: 1.12912139, Log Avg loss: 1.05773207, Global Avg Loss: 1.47691504, Time: 0.0403 Steps: 94270, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000604, Sample Num: 9664, Cur Loss: 0.91165268, Cur Avg Loss: 1.13120868, Log Avg loss: 1.25519347, Global Avg Loss: 1.47689152, Time: 0.0402 Steps: 94280, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000614, Sample Num: 9824, Cur Loss: 0.64909482, Cur Avg Loss: 1.12889471, Log Avg loss: 0.98913141, Global Avg Loss: 1.47683979, Time: 0.0403 Steps: 94290, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000624, Sample Num: 9984, Cur Loss: 1.29628956, Cur Avg Loss: 1.12931915, Log Avg loss: 1.15537959, Global Avg Loss: 1.47680570, Time: 0.0825 Steps: 94300, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000634, Sample Num: 10144, Cur Loss: 0.95582330, Cur Avg Loss: 1.12335532, Log Avg loss: 0.75121197, Global Avg Loss: 1.47672876, Time: 0.1005 Steps: 94310, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000644, Sample Num: 10304, Cur Loss: 1.02671337, Cur Avg Loss: 1.12254948, Log Avg loss: 1.07145932, Global Avg Loss: 1.47668579, Time: 0.0613 Steps: 94320, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000654, Sample Num: 10464, Cur Loss: 0.92446971, Cur Avg Loss: 1.11898720, Log Avg loss: 0.88957632, Global Avg Loss: 1.47662355, Time: 0.0548 Steps: 94330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000664, Sample Num: 10624, Cur Loss: 0.60170037, Cur Avg Loss: 1.11469130, Log Avg loss: 0.83373997, Global Avg Loss: 1.47655541, Time: 0.0829 Steps: 94340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000674, Sample Num: 10784, Cur Loss: 1.22968054, Cur Avg Loss: 1.11430105, Log Avg loss: 1.08838818, Global Avg Loss: 1.47651427, Time: 0.0589 Steps: 94350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000684, Sample Num: 10944, Cur Loss: 3.44641471, Cur Avg Loss: 1.11629455, Log Avg loss: 1.25065629, Global Avg Loss: 1.47649033, Time: 0.0594 Steps: 94360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000694, Sample Num: 11104, Cur Loss: 0.50788021, Cur Avg Loss: 1.11363673, Log Avg loss: 0.93184222, Global Avg Loss: 1.47643262, Time: 0.0406 Steps: 94370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000704, Sample Num: 11264, Cur Loss: 2.01742363, Cur Avg Loss: 1.11510997, Log Avg loss: 1.21735263, Global Avg Loss: 1.47640517, Time: 0.0403 Steps: 94380, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000714, Sample Num: 11424, Cur Loss: 1.02599227, Cur Avg Loss: 1.11601683, Log Avg loss: 1.17985944, Global Avg Loss: 1.47637375, Time: 0.0402 Steps: 94390, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000724, Sample Num: 11584, Cur Loss: 0.92750311, Cur Avg Loss: 1.11572154, Log Avg loss: 1.09463821, Global Avg Loss: 1.47633331, Time: 0.0403 Steps: 94400, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000734, Sample Num: 11744, Cur Loss: 1.24914885, Cur Avg Loss: 1.11608277, Log Avg loss: 1.14223606, Global Avg Loss: 1.47629792, Time: 0.0402 Steps: 94410, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000744, Sample Num: 11904, Cur Loss: 1.71462786, Cur Avg Loss: 1.11895300, Log Avg loss: 1.32962793, Global Avg Loss: 1.47628239, Time: 0.0403 Steps: 94420, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000754, Sample Num: 12064, Cur Loss: 0.44733429, Cur Avg Loss: 1.11349740, Log Avg loss: 0.70760044, Global Avg Loss: 1.47620099, Time: 0.0403 Steps: 94430, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000764, Sample Num: 12224, Cur Loss: 1.16394389, Cur Avg Loss: 1.11389441, Log Avg loss: 1.14382873, Global Avg Loss: 1.47616579, Time: 0.0402 Steps: 94440, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000774, Sample Num: 12384, Cur Loss: 0.47457173, Cur Avg Loss: 1.11356720, Log Avg loss: 1.08856875, Global Avg Loss: 1.47612476, Time: 0.0402 Steps: 94450, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000784, Sample Num: 12544, Cur Loss: 1.35820186, Cur Avg Loss: 1.11204821, Log Avg loss: 0.99447788, Global Avg Loss: 1.47607377, Time: 0.0403 Steps: 94460, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000794, Sample Num: 12704, Cur Loss: 1.11596251, Cur Avg Loss: 1.11400691, Log Avg loss: 1.26756923, Global Avg Loss: 1.47605170, Time: 0.0402 Steps: 94470, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000804, Sample Num: 12864, Cur Loss: 1.37738264, Cur Avg Loss: 1.11409135, Log Avg loss: 1.12079583, Global Avg Loss: 1.47601410, Time: 0.0402 Steps: 94480, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000814, Sample Num: 13024, Cur Loss: 0.53601086, Cur Avg Loss: 1.11557520, Log Avg loss: 1.23487672, Global Avg Loss: 1.47598858, Time: 0.0402 Steps: 94490, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000824, Sample Num: 13184, Cur Loss: 1.82227314, Cur Avg Loss: 1.11564741, Log Avg loss: 1.12152530, Global Avg Loss: 1.47595107, Time: 0.0402 Steps: 94500, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000834, Sample Num: 13344, Cur Loss: 0.78829825, Cur Avg Loss: 1.11685580, Log Avg loss: 1.21642701, Global Avg Loss: 1.47592361, Time: 0.0402 Steps: 94510, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000844, Sample Num: 13504, Cur Loss: 0.82895118, Cur Avg Loss: 1.11867032, Log Avg loss: 1.27000170, Global Avg Loss: 1.47590182, Time: 0.0402 Steps: 94520, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000854, Sample Num: 13664, Cur Loss: 0.38222957, Cur Avg Loss: 1.11710954, Log Avg loss: 0.98537977, Global Avg Loss: 1.47584993, Time: 0.0402 Steps: 94530, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000864, Sample Num: 13824, Cur Loss: 1.17716312, Cur Avg Loss: 1.11648659, Log Avg loss: 1.06328608, Global Avg Loss: 1.47580629, Time: 0.0402 Steps: 94540, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000874, Sample Num: 13984, Cur Loss: 0.85676008, Cur Avg Loss: 1.11591999, Log Avg loss: 1.06696589, Global Avg Loss: 1.47576305, Time: 0.0403 Steps: 94550, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000884, Sample Num: 14144, Cur Loss: 0.32549882, Cur Avg Loss: 1.12363590, Log Avg loss: 1.79800716, Global Avg Loss: 1.47579713, Time: 0.0402 Steps: 94560, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000894, Sample Num: 14304, Cur Loss: 1.53171229, Cur Avg Loss: 1.12366527, Log Avg loss: 1.12626112, Global Avg Loss: 1.47576017, Time: 0.0402 Steps: 94570, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000904, Sample Num: 14464, Cur Loss: 0.68630159, Cur Avg Loss: 1.12391109, Log Avg loss: 1.14588736, Global Avg Loss: 1.47572529, Time: 0.0402 Steps: 94580, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000914, Sample Num: 14624, Cur Loss: 1.95226431, Cur Avg Loss: 1.12419118, Log Avg loss: 1.14951098, Global Avg Loss: 1.47569080, Time: 0.0402 Steps: 94590, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000924, Sample Num: 14784, Cur Loss: 2.15919876, Cur Avg Loss: 1.12255082, Log Avg loss: 0.97262224, Global Avg Loss: 1.47563762, Time: 0.0402 Steps: 94600, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000934, Sample Num: 14944, Cur Loss: 2.52209902, Cur Avg Loss: 1.12372674, Log Avg loss: 1.23238203, Global Avg Loss: 1.47561191, Time: 0.0402 Steps: 94610, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000944, Sample Num: 15104, Cur Loss: 1.21352386, Cur Avg Loss: 1.12521385, Log Avg loss: 1.26410928, Global Avg Loss: 1.47558956, Time: 0.0403 Steps: 94620, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000954, Sample Num: 15264, Cur Loss: 0.24521631, Cur Avg Loss: 1.12456606, Log Avg loss: 1.06341552, Global Avg Loss: 1.47554600, Time: 0.0402 Steps: 94630, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000964, Sample Num: 15424, Cur Loss: 1.21134162, Cur Avg Loss: 1.12218372, Log Avg loss: 0.89490829, Global Avg Loss: 1.47548465, Time: 0.0432 Steps: 94640, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000974, Sample Num: 15584, Cur Loss: 0.97472727, Cur Avg Loss: 1.12033815, Log Avg loss: 0.94242540, Global Avg Loss: 1.47542833, Time: 0.0857 Steps: 94650, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000984, Sample Num: 15744, Cur Loss: 0.97402561, Cur Avg Loss: 1.11957064, Log Avg loss: 1.04481466, Global Avg Loss: 1.47538284, Time: 0.0625 Steps: 94660, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000994, Sample Num: 15904, Cur Loss: 0.31684682, Cur Avg Loss: 1.11981833, Log Avg loss: 1.14419104, Global Avg Loss: 1.47534786, Time: 0.0898 Steps: 94670, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001004, Sample Num: 16064, Cur Loss: 0.76136035, Cur Avg Loss: 1.12194827, Log Avg loss: 1.33366440, Global Avg Loss: 1.47533289, Time: 0.0452 Steps: 94680, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001014, Sample Num: 16224, Cur Loss: 1.41631925, Cur Avg Loss: 1.12295956, Log Avg loss: 1.22449277, Global Avg Loss: 1.47530640, Time: 0.0863 Steps: 94690, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001024, Sample Num: 16384, Cur Loss: 1.16077840, Cur Avg Loss: 1.12007612, Log Avg loss: 0.82769524, Global Avg Loss: 1.47523802, Time: 0.0469 Steps: 94700, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001034, Sample Num: 16544, Cur Loss: 1.13421929, Cur Avg Loss: 1.12299210, Log Avg loss: 1.42158922, Global Avg Loss: 1.47523235, Time: 0.0817 Steps: 94710, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001044, Sample Num: 16704, Cur Loss: 1.05550194, Cur Avg Loss: 1.12234831, Log Avg loss: 1.05577988, Global Avg Loss: 1.47518807, Time: 0.0880 Steps: 94720, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001054, Sample Num: 16864, Cur Loss: 1.04764903, Cur Avg Loss: 1.12420090, Log Avg loss: 1.31761184, Global Avg Loss: 1.47517144, Time: 0.0403 Steps: 94730, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001064, Sample Num: 17024, Cur Loss: 0.72943991, Cur Avg Loss: 1.12384678, Log Avg loss: 1.08652244, Global Avg Loss: 1.47513041, Time: 0.0403 Steps: 94740, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001074, Sample Num: 17184, Cur Loss: 0.48257807, Cur Avg Loss: 1.12250369, Log Avg loss: 0.97959807, Global Avg Loss: 1.47507811, Time: 0.0402 Steps: 94750, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001084, Sample Num: 17344, Cur Loss: 0.87721437, Cur Avg Loss: 1.12078786, Log Avg loss: 0.93650853, Global Avg Loss: 1.47502128, Time: 0.0402 Steps: 94760, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001094, Sample Num: 17504, Cur Loss: 2.73918080, Cur Avg Loss: 1.12449798, Log Avg loss: 1.52667441, Global Avg Loss: 1.47502673, Time: 0.0403 Steps: 94770, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001104, Sample Num: 17664, Cur Loss: 0.57914752, Cur Avg Loss: 1.12490870, Log Avg loss: 1.16984156, Global Avg Loss: 1.47499453, Time: 0.0402 Steps: 94780, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001114, Sample Num: 17824, Cur Loss: 1.00369525, Cur Avg Loss: 1.12587357, Log Avg loss: 1.23239509, Global Avg Loss: 1.47496894, Time: 0.0403 Steps: 94790, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001124, Sample Num: 17984, Cur Loss: 1.67260325, Cur Avg Loss: 1.12552914, Log Avg loss: 1.08716016, Global Avg Loss: 1.47492803, Time: 0.0402 Steps: 94800, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001134, Sample Num: 18144, Cur Loss: 1.06614447, Cur Avg Loss: 1.12640772, Log Avg loss: 1.22516029, Global Avg Loss: 1.47490168, Time: 0.0403 Steps: 94810, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001144, Sample Num: 18304, Cur Loss: 1.09599495, Cur Avg Loss: 1.12545940, Log Avg loss: 1.01791992, Global Avg Loss: 1.47485349, Time: 0.0403 Steps: 94820, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001154, Sample Num: 18464, Cur Loss: 0.76638544, Cur Avg Loss: 1.12559332, Log Avg loss: 1.14091351, Global Avg Loss: 1.47481827, Time: 0.0402 Steps: 94830, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001164, Sample Num: 18624, Cur Loss: 1.28251505, Cur Avg Loss: 1.12510941, Log Avg loss: 1.06926604, Global Avg Loss: 1.47477551, Time: 0.0402 Steps: 94840, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001174, Sample Num: 18784, Cur Loss: 1.00945246, Cur Avg Loss: 1.12656448, Log Avg loss: 1.29593492, Global Avg Loss: 1.47475666, Time: 0.0403 Steps: 94850, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001184, Sample Num: 18944, Cur Loss: 1.46783912, Cur Avg Loss: 1.12655858, Log Avg loss: 1.12586576, Global Avg Loss: 1.47471988, Time: 0.0403 Steps: 94860, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001194, Sample Num: 19104, Cur Loss: 0.71067375, Cur Avg Loss: 1.12519899, Log Avg loss: 0.96422339, Global Avg Loss: 1.47466607, Time: 0.0403 Steps: 94870, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001204, Sample Num: 19264, Cur Loss: 1.04192531, Cur Avg Loss: 1.12538201, Log Avg loss: 1.14723423, Global Avg Loss: 1.47463156, Time: 0.0403 Steps: 94880, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001214, Sample Num: 19424, Cur Loss: 2.10942197, Cur Avg Loss: 1.12591256, Log Avg loss: 1.18979179, Global Avg Loss: 1.47460154, Time: 0.0403 Steps: 94890, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001224, Sample Num: 19584, Cur Loss: 0.80892479, Cur Avg Loss: 1.12593741, Log Avg loss: 1.12895324, Global Avg Loss: 1.47456512, Time: 0.0402 Steps: 94900, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001234, Sample Num: 19744, Cur Loss: 0.60779536, Cur Avg Loss: 1.12483769, Log Avg loss: 0.99023180, Global Avg Loss: 1.47451409, Time: 0.0403 Steps: 94910, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001244, Sample Num: 19904, Cur Loss: 0.51176763, Cur Avg Loss: 1.12189825, Log Avg loss: 0.75917179, Global Avg Loss: 1.47443872, Time: 0.0403 Steps: 94920, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001254, Sample Num: 20064, Cur Loss: 3.53017688, Cur Avg Loss: 1.12253031, Log Avg loss: 1.20115859, Global Avg Loss: 1.47440994, Time: 0.0402 Steps: 94930, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001264, Sample Num: 20224, Cur Loss: 1.23371899, Cur Avg Loss: 1.12176667, Log Avg loss: 1.02600623, Global Avg Loss: 1.47436271, Time: 0.0402 Steps: 94940, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001274, Sample Num: 20384, Cur Loss: 1.76970398, Cur Avg Loss: 1.12194354, Log Avg loss: 1.14429976, Global Avg Loss: 1.47432795, Time: 0.0402 Steps: 94950, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001284, Sample Num: 20544, Cur Loss: 1.58687317, Cur Avg Loss: 1.12282633, Log Avg loss: 1.23529437, Global Avg Loss: 1.47430277, Time: 0.0402 Steps: 94960, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001294, Sample Num: 20704, Cur Loss: 0.80238783, Cur Avg Loss: 1.12293053, Log Avg loss: 1.13630888, Global Avg Loss: 1.47426718, Time: 0.0403 Steps: 94970, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001304, Sample Num: 20864, Cur Loss: 1.03469479, Cur Avg Loss: 1.12598241, Log Avg loss: 1.52089581, Global Avg Loss: 1.47427209, Time: 0.0403 Steps: 94980, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001314, Sample Num: 21024, Cur Loss: 0.58845001, Cur Avg Loss: 1.12562540, Log Avg loss: 1.07907134, Global Avg Loss: 1.47423049, Time: 0.0402 Steps: 94990, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001324, Sample Num: 21184, Cur Loss: 1.88397002, Cur Avg Loss: 1.12517082, Log Avg loss: 1.06543982, Global Avg Loss: 1.47418746, Time: 0.0628 Steps: 95000, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001334, Sample Num: 21344, Cur Loss: 0.64205229, Cur Avg Loss: 1.12605498, Log Avg loss: 1.24311730, Global Avg Loss: 1.47416314, Time: 0.1002 Steps: 95010, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001344, Sample Num: 21504, Cur Loss: 1.19191372, Cur Avg Loss: 1.12662293, Log Avg loss: 1.20238717, Global Avg Loss: 1.47413454, Time: 0.0704 Steps: 95020, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001354, Sample Num: 21664, Cur Loss: 1.36275077, Cur Avg Loss: 1.12933778, Log Avg loss: 1.49421327, Global Avg Loss: 1.47413665, Time: 0.0493 Steps: 95030, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001364, Sample Num: 21824, Cur Loss: 1.09624171, Cur Avg Loss: 1.12876399, Log Avg loss: 1.05107319, Global Avg Loss: 1.47409213, Time: 0.0593 Steps: 95040, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001374, Sample Num: 21984, Cur Loss: 1.55424643, Cur Avg Loss: 1.13066223, Log Avg loss: 1.38958251, Global Avg Loss: 1.47408324, Time: 0.0718 Steps: 95050, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001384, Sample Num: 22144, Cur Loss: 0.71966803, Cur Avg Loss: 1.13003387, Log Avg loss: 1.04369754, Global Avg Loss: 1.47403797, Time: 0.0619 Steps: 95060, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001394, Sample Num: 22304, Cur Loss: 1.24327588, Cur Avg Loss: 1.12863679, Log Avg loss: 0.93528004, Global Avg Loss: 1.47398130, Time: 0.0406 Steps: 95070, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001404, Sample Num: 22464, Cur Loss: 0.28867906, Cur Avg Loss: 1.12663486, Log Avg loss: 0.84756652, Global Avg Loss: 1.47391542, Time: 0.0604 Steps: 95080, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001414, Sample Num: 22624, Cur Loss: 0.51565099, Cur Avg Loss: 1.12680371, Log Avg loss: 1.15051041, Global Avg Loss: 1.47388140, Time: 0.0403 Steps: 95090, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001424, Sample Num: 22784, Cur Loss: 1.28566670, Cur Avg Loss: 1.12878103, Log Avg loss: 1.40837385, Global Avg Loss: 1.47387452, Time: 0.0403 Steps: 95100, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001434, Sample Num: 22944, Cur Loss: 0.78665549, Cur Avg Loss: 1.12677331, Log Avg loss: 0.84087451, Global Avg Loss: 1.47380796, Time: 0.0403 Steps: 95110, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001444, Sample Num: 23104, Cur Loss: 2.23010707, Cur Avg Loss: 1.12928658, Log Avg loss: 1.48968835, Global Avg Loss: 1.47380963, Time: 0.0403 Steps: 95120, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001454, Sample Num: 23264, Cur Loss: 1.03548551, Cur Avg Loss: 1.12924877, Log Avg loss: 1.12378913, Global Avg Loss: 1.47377284, Time: 0.0403 Steps: 95130, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001464, Sample Num: 23424, Cur Loss: 0.88372099, Cur Avg Loss: 1.12879431, Log Avg loss: 1.06271547, Global Avg Loss: 1.47372963, Time: 0.0403 Steps: 95140, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001474, Sample Num: 23584, Cur Loss: 1.52763486, Cur Avg Loss: 1.13040853, Log Avg loss: 1.36673069, Global Avg Loss: 1.47371839, Time: 0.0403 Steps: 95150, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001484, Sample Num: 23744, Cur Loss: 2.07187462, Cur Avg Loss: 1.12964222, Log Avg loss: 1.01668887, Global Avg Loss: 1.47367036, Time: 0.0403 Steps: 95160, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001494, Sample Num: 23904, Cur Loss: 1.42266655, Cur Avg Loss: 1.13252139, Log Avg loss: 1.55979028, Global Avg Loss: 1.47367941, Time: 0.0402 Steps: 95170, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001504, Sample Num: 24064, Cur Loss: 1.35464191, Cur Avg Loss: 1.13041631, Log Avg loss: 0.81591756, Global Avg Loss: 1.47361030, Time: 0.0402 Steps: 95180, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001514, Sample Num: 24224, Cur Loss: 0.99763560, Cur Avg Loss: 1.13136288, Log Avg loss: 1.27372695, Global Avg Loss: 1.47358930, Time: 0.0401 Steps: 95190, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001524, Sample Num: 24384, Cur Loss: 0.71977592, Cur Avg Loss: 1.12889728, Log Avg loss: 0.75560467, Global Avg Loss: 1.47351388, Time: 0.0402 Steps: 95200, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001534, Sample Num: 24544, Cur Loss: 1.27922916, Cur Avg Loss: 1.12886588, Log Avg loss: 1.12408012, Global Avg Loss: 1.47347718, Time: 0.0403 Steps: 95210, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001544, Sample Num: 24704, Cur Loss: 1.12511730, Cur Avg Loss: 1.12860541, Log Avg loss: 1.08864950, Global Avg Loss: 1.47343677, Time: 0.0403 Steps: 95220, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001554, Sample Num: 24864, Cur Loss: 2.49757957, Cur Avg Loss: 1.13004012, Log Avg loss: 1.35155947, Global Avg Loss: 1.47342397, Time: 0.0402 Steps: 95230, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001564, Sample Num: 25024, Cur Loss: 0.33772683, Cur Avg Loss: 1.12963647, Log Avg loss: 1.06690877, Global Avg Loss: 1.47338129, Time: 0.0402 Steps: 95240, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001574, Sample Num: 25184, Cur Loss: 0.80508602, Cur Avg Loss: 1.12843276, Log Avg loss: 0.94017284, Global Avg Loss: 1.47332531, Time: 0.0402 Steps: 95250, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001584, Sample Num: 25344, Cur Loss: 1.48160338, Cur Avg Loss: 1.12904408, Log Avg loss: 1.22526692, Global Avg Loss: 1.47329927, Time: 0.0402 Steps: 95260, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001594, Sample Num: 25504, Cur Loss: 1.78319597, Cur Avg Loss: 1.13192088, Log Avg loss: 1.58760550, Global Avg Loss: 1.47331126, Time: 0.0402 Steps: 95270, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001604, Sample Num: 25664, Cur Loss: 1.86813807, Cur Avg Loss: 1.13433202, Log Avg loss: 1.51866697, Global Avg Loss: 1.47331603, Time: 0.0402 Steps: 95280, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001614, Sample Num: 25824, Cur Loss: 1.85213590, Cur Avg Loss: 1.13519802, Log Avg loss: 1.27410527, Global Avg Loss: 1.47329512, Time: 0.0403 Steps: 95290, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001624, Sample Num: 25984, Cur Loss: 0.28459650, Cur Avg Loss: 1.13586976, Log Avg loss: 1.24428874, Global Avg Loss: 1.47327109, Time: 0.0403 Steps: 95300, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001634, Sample Num: 26144, Cur Loss: 1.87112045, Cur Avg Loss: 1.13756982, Log Avg loss: 1.41365871, Global Avg Loss: 1.47326483, Time: 0.0402 Steps: 95310, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001644, Sample Num: 26304, Cur Loss: 2.36639023, Cur Avg Loss: 1.13683092, Log Avg loss: 1.01609564, Global Avg Loss: 1.47321687, Time: 0.0403 Steps: 95320, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001654, Sample Num: 26464, Cur Loss: 0.16113691, Cur Avg Loss: 1.13592362, Log Avg loss: 0.98676350, Global Avg Loss: 1.47316584, Time: 0.0402 Steps: 95330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001664, Sample Num: 26624, Cur Loss: 1.06736028, Cur Avg Loss: 1.13709906, Log Avg loss: 1.33151585, Global Avg Loss: 1.47315099, Time: 0.0402 Steps: 95340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001674, Sample Num: 26784, Cur Loss: 1.16526747, Cur Avg Loss: 1.13621607, Log Avg loss: 0.98928661, Global Avg Loss: 1.47310024, Time: 0.0679 Steps: 95350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001684, Sample Num: 26944, Cur Loss: 0.76914454, Cur Avg Loss: 1.13489826, Log Avg loss: 0.91429785, Global Avg Loss: 1.47304164, Time: 0.0425 Steps: 95360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001694, Sample Num: 27104, Cur Loss: 1.15494752, Cur Avg Loss: 1.13459896, Log Avg loss: 1.08419589, Global Avg Loss: 1.47300087, Time: 0.0806 Steps: 95370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001704, Sample Num: 27264, Cur Loss: 0.40262422, Cur Avg Loss: 1.13316739, Log Avg loss: 0.89066059, Global Avg Loss: 1.47293982, Time: 0.0405 Steps: 95380, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001714, Sample Num: 27424, Cur Loss: 0.58351934, Cur Avg Loss: 1.13253550, Log Avg loss: 1.02486101, Global Avg Loss: 1.47289284, Time: 0.0405 Steps: 95390, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001724, Sample Num: 27584, Cur Loss: 0.51273322, Cur Avg Loss: 1.13119972, Log Avg loss: 0.90224701, Global Avg Loss: 1.47283303, Time: 0.0461 Steps: 95400, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001734, Sample Num: 27744, Cur Loss: 1.43510878, Cur Avg Loss: 1.13130910, Log Avg loss: 1.15016567, Global Avg Loss: 1.47279921, Time: 0.0849 Steps: 95410, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001744, Sample Num: 27904, Cur Loss: 1.49918950, Cur Avg Loss: 1.13027524, Log Avg loss: 0.95100413, Global Avg Loss: 1.47274452, Time: 0.0680 Steps: 95420, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001754, Sample Num: 28064, Cur Loss: 0.85392940, Cur Avg Loss: 1.12803473, Log Avg loss: 0.73729043, Global Avg Loss: 1.47266746, Time: 0.1005 Steps: 95430, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001764, Sample Num: 28224, Cur Loss: 0.83511662, Cur Avg Loss: 1.12817140, Log Avg loss: 1.15214292, Global Avg Loss: 1.47263387, Time: 0.0402 Steps: 95440, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001774, Sample Num: 28384, Cur Loss: 1.01965916, Cur Avg Loss: 1.12822907, Log Avg loss: 1.13840235, Global Avg Loss: 1.47259885, Time: 0.0402 Steps: 95450, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001784, Sample Num: 28544, Cur Loss: 0.80165994, Cur Avg Loss: 1.12785390, Log Avg loss: 1.06129780, Global Avg Loss: 1.47255577, Time: 0.0403 Steps: 95460, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001794, Sample Num: 28704, Cur Loss: 0.51911938, Cur Avg Loss: 1.12856516, Log Avg loss: 1.25545372, Global Avg Loss: 1.47253303, Time: 0.0403 Steps: 95470, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001804, Sample Num: 28864, Cur Loss: 1.45817173, Cur Avg Loss: 1.12828569, Log Avg loss: 1.07815012, Global Avg Loss: 1.47249172, Time: 0.0402 Steps: 95480, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001814, Sample Num: 29024, Cur Loss: 1.12311196, Cur Avg Loss: 1.12837341, Log Avg loss: 1.14419734, Global Avg Loss: 1.47245734, Time: 0.0404 Steps: 95490, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001824, Sample Num: 29184, Cur Loss: 0.96896136, Cur Avg Loss: 1.12811224, Log Avg loss: 1.08073625, Global Avg Loss: 1.47241633, Time: 0.0402 Steps: 95500, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001834, Sample Num: 29344, Cur Loss: 0.91712308, Cur Avg Loss: 1.12724001, Log Avg loss: 0.96814586, Global Avg Loss: 1.47236353, Time: 0.0402 Steps: 95510, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001844, Sample Num: 29504, Cur Loss: 1.14188004, Cur Avg Loss: 1.12712827, Log Avg loss: 1.10663435, Global Avg Loss: 1.47232524, Time: 0.0402 Steps: 95520, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001854, Sample Num: 29664, Cur Loss: 0.41789848, Cur Avg Loss: 1.12571655, Log Avg loss: 0.86539609, Global Avg Loss: 1.47226171, Time: 0.0404 Steps: 95530, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001864, Sample Num: 29824, Cur Loss: 0.96746004, Cur Avg Loss: 1.12587971, Log Avg loss: 1.15612905, Global Avg Loss: 1.47222862, Time: 0.0403 Steps: 95540, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001874, Sample Num: 29984, Cur Loss: 3.00110674, Cur Avg Loss: 1.12705113, Log Avg loss: 1.34540434, Global Avg Loss: 1.47221534, Time: 0.0403 Steps: 95550, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001884, Sample Num: 30144, Cur Loss: 0.51888824, Cur Avg Loss: 1.12642230, Log Avg loss: 1.00857899, Global Avg Loss: 1.47216683, Time: 0.0402 Steps: 95560, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001894, Sample Num: 30304, Cur Loss: 0.82794595, Cur Avg Loss: 1.12657107, Log Avg loss: 1.15460008, Global Avg Loss: 1.47213360, Time: 0.0403 Steps: 95570, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001904, Sample Num: 30464, Cur Loss: 1.04952133, Cur Avg Loss: 1.12552073, Log Avg loss: 0.92658551, Global Avg Loss: 1.47207652, Time: 0.0403 Steps: 95580, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001914, Sample Num: 30624, Cur Loss: 0.66159868, Cur Avg Loss: 1.12560538, Log Avg loss: 1.14172183, Global Avg Loss: 1.47204196, Time: 0.0403 Steps: 95590, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001924, Sample Num: 30784, Cur Loss: 0.91845644, Cur Avg Loss: 1.12713653, Log Avg loss: 1.42019855, Global Avg Loss: 1.47203654, Time: 0.0402 Steps: 95600, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001934, Sample Num: 30944, Cur Loss: 1.28424788, Cur Avg Loss: 1.12690966, Log Avg loss: 1.08326053, Global Avg Loss: 1.47199587, Time: 0.0402 Steps: 95610, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001944, Sample Num: 31104, Cur Loss: 1.24679244, Cur Avg Loss: 1.12711220, Log Avg loss: 1.16628379, Global Avg Loss: 1.47196390, Time: 0.0402 Steps: 95620, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001954, Sample Num: 31264, Cur Loss: 1.54745674, Cur Avg Loss: 1.12733705, Log Avg loss: 1.17104794, Global Avg Loss: 1.47193244, Time: 0.0402 Steps: 95630, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001964, Sample Num: 31424, Cur Loss: 1.69906545, Cur Avg Loss: 1.12773787, Log Avg loss: 1.20605810, Global Avg Loss: 1.47190464, Time: 0.0402 Steps: 95640, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001974, Sample Num: 31584, Cur Loss: 0.48251542, Cur Avg Loss: 1.12662983, Log Avg loss: 0.90900986, Global Avg Loss: 1.47184579, Time: 0.0403 Steps: 95650, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001984, Sample Num: 31744, Cur Loss: 0.58132350, Cur Avg Loss: 1.12566577, Log Avg loss: 0.93536156, Global Avg Loss: 1.47178971, Time: 0.0403 Steps: 95660, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001994, Sample Num: 31904, Cur Loss: 2.25580907, Cur Avg Loss: 1.12682927, Log Avg loss: 1.35766695, Global Avg Loss: 1.47177778, Time: 0.0402 Steps: 95670, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002004, Sample Num: 32064, Cur Loss: 0.87050807, Cur Avg Loss: 1.12550176, Log Avg loss: 0.86079717, Global Avg Loss: 1.47171392, Time: 0.0402 Steps: 95680, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002014, Sample Num: 32224, Cur Loss: 0.90242326, Cur Avg Loss: 1.12701294, Log Avg loss: 1.42985300, Global Avg Loss: 1.47170955, Time: 0.0402 Steps: 95690, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002024, Sample Num: 32384, Cur Loss: 0.94119602, Cur Avg Loss: 1.12625896, Log Avg loss: 0.97440748, Global Avg Loss: 1.47165758, Time: 0.0626 Steps: 95700, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002034, Sample Num: 32544, Cur Loss: 0.48792756, Cur Avg Loss: 1.12507797, Log Avg loss: 0.88604462, Global Avg Loss: 1.47159639, Time: 0.0723 Steps: 95710, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002044, Sample Num: 32704, Cur Loss: 1.49124539, Cur Avg Loss: 1.12786987, Log Avg loss: 1.69574354, Global Avg Loss: 1.47161981, Time: 0.0572 Steps: 95720, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002054, Sample Num: 32864, Cur Loss: 0.88588464, Cur Avg Loss: 1.12842407, Log Avg loss: 1.24170216, Global Avg Loss: 1.47159579, Time: 0.0602 Steps: 95730, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002064, Sample Num: 33024, Cur Loss: 1.69691658, Cur Avg Loss: 1.12880532, Log Avg loss: 1.20711325, Global Avg Loss: 1.47156817, Time: 0.0406 Steps: 95740, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002074, Sample Num: 33184, Cur Loss: 0.69605625, Cur Avg Loss: 1.12850933, Log Avg loss: 1.06741730, Global Avg Loss: 1.47152596, Time: 0.0861 Steps: 95750, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002084, Sample Num: 33344, Cur Loss: 0.65104288, Cur Avg Loss: 1.12800665, Log Avg loss: 1.02375127, Global Avg Loss: 1.47147920, Time: 0.0996 Steps: 95760, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002094, Sample Num: 33504, Cur Loss: 1.86791158, Cur Avg Loss: 1.12844801, Log Avg loss: 1.22042644, Global Avg Loss: 1.47145299, Time: 0.0887 Steps: 95770, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002104, Sample Num: 33664, Cur Loss: 2.38400340, Cur Avg Loss: 1.12928372, Log Avg loss: 1.30428126, Global Avg Loss: 1.47143553, Time: 0.0449 Steps: 95780, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002114, Sample Num: 33824, Cur Loss: 0.31220171, Cur Avg Loss: 1.12852571, Log Avg loss: 0.96904183, Global Avg Loss: 1.47138308, Time: 0.0404 Steps: 95790, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002124, Sample Num: 33984, Cur Loss: 0.60758734, Cur Avg Loss: 1.12734193, Log Avg loss: 0.87708969, Global Avg Loss: 1.47132105, Time: 0.0404 Steps: 95800, Updated lr: 0.000010 ***** Running evaluation checkpoint-95805 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-95805 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.632313, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.072179, "eval_total_loss": 753.741824, "eval_mae": 0.830734, "eval_mse": 1.072329, "eval_r2": 0.318357, "eval_sp_statistic": 0.468212, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.57339, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.049457, "test_total_loss": 526.827174, "test_mae": 0.837903, "test_mse": 1.049671, "test_r2": 0.322533, "test_sp_statistic": 0.489802, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.619063, "test_ps_pvalue": 0.0, "lr": 1.0094831673779043e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4713217153887828, "train_cur_epoch_loss": 2401.8946268707514, "train_cur_epoch_avg_loss": 1.1281797214047682, "train_cur_epoch_time": 95.6323127746582, "train_cur_epoch_avg_time": 0.04491888810458347, "epoch": 45, "step": 95805} ################################################## Training, Epoch: 0046, Batch: 000005, Sample Num: 80, Cur Loss: 1.05658722, Cur Avg Loss: 1.21421883, Log Avg loss: 1.34914694, Global Avg Loss: 1.47130830, Time: 0.0460 Steps: 95810, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000015, Sample Num: 240, Cur Loss: 0.92264968, Cur Avg Loss: 1.28993309, Log Avg loss: 1.32779021, Global Avg Loss: 1.47129332, Time: 0.0652 Steps: 95820, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000025, Sample Num: 400, Cur Loss: 1.73216593, Cur Avg Loss: 1.21035691, Log Avg loss: 1.09099264, Global Avg Loss: 1.47125364, Time: 0.0435 Steps: 95830, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000035, Sample Num: 560, Cur Loss: 1.87422371, Cur Avg Loss: 1.21440683, Log Avg loss: 1.22453165, Global Avg Loss: 1.47122789, Time: 0.0523 Steps: 95840, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000045, Sample Num: 720, Cur Loss: 1.80498970, Cur Avg Loss: 1.19254077, Log Avg loss: 1.11600954, Global Avg Loss: 1.47119083, Time: 0.0402 Steps: 95850, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000055, Sample Num: 880, Cur Loss: 1.53324533, Cur Avg Loss: 1.19361247, Log Avg loss: 1.19843512, Global Avg Loss: 1.47116238, Time: 0.0403 Steps: 95860, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000065, Sample Num: 1040, Cur Loss: 0.96110260, Cur Avg Loss: 1.16161784, Log Avg loss: 0.98564738, Global Avg Loss: 1.47111174, Time: 0.0404 Steps: 95870, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000075, Sample Num: 1200, Cur Loss: 0.38939083, Cur Avg Loss: 1.10887330, Log Avg loss: 0.76603383, Global Avg Loss: 1.47103820, Time: 0.0404 Steps: 95880, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000085, Sample Num: 1360, Cur Loss: 1.66945636, Cur Avg Loss: 1.12639450, Log Avg loss: 1.25780345, Global Avg Loss: 1.47101596, Time: 0.0402 Steps: 95890, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000095, Sample Num: 1520, Cur Loss: 0.51200235, Cur Avg Loss: 1.12287793, Log Avg loss: 1.09298710, Global Avg Loss: 1.47097654, Time: 0.0403 Steps: 95900, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000105, Sample Num: 1680, Cur Loss: 1.02220261, Cur Avg Loss: 1.10719129, Log Avg loss: 0.95816820, Global Avg Loss: 1.47092307, Time: 0.0402 Steps: 95910, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000115, Sample Num: 1840, Cur Loss: 0.92049450, Cur Avg Loss: 1.11248171, Log Avg loss: 1.16803110, Global Avg Loss: 1.47089150, Time: 0.0402 Steps: 95920, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000125, Sample Num: 2000, Cur Loss: 1.84496498, Cur Avg Loss: 1.11596085, Log Avg loss: 1.15597102, Global Avg Loss: 1.47085867, Time: 0.0402 Steps: 95930, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000135, Sample Num: 2160, Cur Loss: 0.54888058, Cur Avg Loss: 1.12948388, Log Avg loss: 1.29852175, Global Avg Loss: 1.47084071, Time: 0.0402 Steps: 95940, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000145, Sample Num: 2320, Cur Loss: 1.29573560, Cur Avg Loss: 1.12246495, Log Avg loss: 1.02770934, Global Avg Loss: 1.47079452, Time: 0.0402 Steps: 95950, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000155, Sample Num: 2480, Cur Loss: 1.83405757, Cur Avg Loss: 1.14491538, Log Avg loss: 1.47044671, Global Avg Loss: 1.47079449, Time: 0.0403 Steps: 95960, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000165, Sample Num: 2640, Cur Loss: 0.79963005, Cur Avg Loss: 1.13914098, Log Avg loss: 1.04963771, Global Avg Loss: 1.47075060, Time: 0.0402 Steps: 95970, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000175, Sample Num: 2800, Cur Loss: 1.72798157, Cur Avg Loss: 1.12413185, Log Avg loss: 0.87648121, Global Avg Loss: 1.47068869, Time: 0.0402 Steps: 95980, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000185, Sample Num: 2960, Cur Loss: 0.61320978, Cur Avg Loss: 1.13887164, Log Avg loss: 1.39681805, Global Avg Loss: 1.47068099, Time: 0.0402 Steps: 95990, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000195, Sample Num: 3120, Cur Loss: 0.31508899, Cur Avg Loss: 1.12641091, Log Avg loss: 0.89588730, Global Avg Loss: 1.47062112, Time: 0.0402 Steps: 96000, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000205, Sample Num: 3280, Cur Loss: 1.17763436, Cur Avg Loss: 1.13925746, Log Avg loss: 1.38976529, Global Avg Loss: 1.47061269, Time: 0.0402 Steps: 96010, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000215, Sample Num: 3440, Cur Loss: 2.76275206, Cur Avg Loss: 1.14688669, Log Avg loss: 1.30328584, Global Avg Loss: 1.47059527, Time: 0.0402 Steps: 96020, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000225, Sample Num: 3600, Cur Loss: 1.01929557, Cur Avg Loss: 1.15192959, Log Avg loss: 1.26035203, Global Avg Loss: 1.47057337, Time: 0.0402 Steps: 96030, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000235, Sample Num: 3760, Cur Loss: 0.45719889, Cur Avg Loss: 1.13725718, Log Avg loss: 0.80712793, Global Avg Loss: 1.47050429, Time: 0.0402 Steps: 96040, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000245, Sample Num: 3920, Cur Loss: 2.10134101, Cur Avg Loss: 1.13018507, Log Avg loss: 0.96399029, Global Avg Loss: 1.47045156, Time: 0.0403 Steps: 96050, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000255, Sample Num: 4080, Cur Loss: 0.71681023, Cur Avg Loss: 1.14017184, Log Avg loss: 1.38484771, Global Avg Loss: 1.47044265, Time: 0.0402 Steps: 96060, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000265, Sample Num: 4240, Cur Loss: 1.35586476, Cur Avg Loss: 1.13129762, Log Avg loss: 0.90500505, Global Avg Loss: 1.47038379, Time: 0.0402 Steps: 96070, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000275, Sample Num: 4400, Cur Loss: 0.96126759, Cur Avg Loss: 1.14148407, Log Avg loss: 1.41142505, Global Avg Loss: 1.47037765, Time: 0.0402 Steps: 96080, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000285, Sample Num: 4560, Cur Loss: 1.29911661, Cur Avg Loss: 1.13900062, Log Avg loss: 1.07070577, Global Avg Loss: 1.47033606, Time: 0.0402 Steps: 96090, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000295, Sample Num: 4720, Cur Loss: 0.40271300, Cur Avg Loss: 1.14446791, Log Avg loss: 1.30028578, Global Avg Loss: 1.47031837, Time: 0.0406 Steps: 96100, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000305, Sample Num: 4880, Cur Loss: 1.05239773, Cur Avg Loss: 1.14439967, Log Avg loss: 1.14238653, Global Avg Loss: 1.47028425, Time: 0.0979 Steps: 96110, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000315, Sample Num: 5040, Cur Loss: 2.35684681, Cur Avg Loss: 1.15310183, Log Avg loss: 1.41851752, Global Avg Loss: 1.47027886, Time: 0.0404 Steps: 96120, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000325, Sample Num: 5200, Cur Loss: 1.50669408, Cur Avg Loss: 1.15869307, Log Avg loss: 1.33481727, Global Avg Loss: 1.47026477, Time: 0.0661 Steps: 96130, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000335, Sample Num: 5360, Cur Loss: 0.72124809, Cur Avg Loss: 1.15340737, Log Avg loss: 0.98162208, Global Avg Loss: 1.47021394, Time: 0.0499 Steps: 96140, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000345, Sample Num: 5520, Cur Loss: 1.09766030, Cur Avg Loss: 1.14797671, Log Avg loss: 0.96604958, Global Avg Loss: 1.47016151, Time: 0.0509 Steps: 96150, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000355, Sample Num: 5680, Cur Loss: 1.42729735, Cur Avg Loss: 1.14568109, Log Avg loss: 1.06648241, Global Avg Loss: 1.47011953, Time: 0.0887 Steps: 96160, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000365, Sample Num: 5840, Cur Loss: 1.06329036, Cur Avg Loss: 1.14049207, Log Avg loss: 0.95628158, Global Avg Loss: 1.47006610, Time: 0.0712 Steps: 96170, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000375, Sample Num: 6000, Cur Loss: 0.87348181, Cur Avg Loss: 1.13882503, Log Avg loss: 1.07797822, Global Avg Loss: 1.47002533, Time: 0.0403 Steps: 96180, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000385, Sample Num: 6160, Cur Loss: 1.33054101, Cur Avg Loss: 1.13851489, Log Avg loss: 1.12688453, Global Avg Loss: 1.46998966, Time: 0.0403 Steps: 96190, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000395, Sample Num: 6320, Cur Loss: 0.38405660, Cur Avg Loss: 1.13624642, Log Avg loss: 1.04891059, Global Avg Loss: 1.46994589, Time: 0.0402 Steps: 96200, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000405, Sample Num: 6480, Cur Loss: 0.31360471, Cur Avg Loss: 1.13737414, Log Avg loss: 1.18191880, Global Avg Loss: 1.46991595, Time: 0.0403 Steps: 96210, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000415, Sample Num: 6640, Cur Loss: 1.18263888, Cur Avg Loss: 1.13828343, Log Avg loss: 1.17510989, Global Avg Loss: 1.46988531, Time: 0.0404 Steps: 96220, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000425, Sample Num: 6800, Cur Loss: 0.70261562, Cur Avg Loss: 1.13406990, Log Avg loss: 0.95920822, Global Avg Loss: 1.46983224, Time: 0.0403 Steps: 96230, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000435, Sample Num: 6960, Cur Loss: 0.27957493, Cur Avg Loss: 1.13157740, Log Avg loss: 1.02564637, Global Avg Loss: 1.46978609, Time: 0.0403 Steps: 96240, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000445, Sample Num: 7120, Cur Loss: 0.94462651, Cur Avg Loss: 1.12552691, Log Avg loss: 0.86233052, Global Avg Loss: 1.46972298, Time: 0.0402 Steps: 96250, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000455, Sample Num: 7280, Cur Loss: 1.56030262, Cur Avg Loss: 1.13038245, Log Avg loss: 1.34645395, Global Avg Loss: 1.46971017, Time: 0.0402 Steps: 96260, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000465, Sample Num: 7440, Cur Loss: 1.70887363, Cur Avg Loss: 1.13029093, Log Avg loss: 1.12612689, Global Avg Loss: 1.46967448, Time: 0.0403 Steps: 96270, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000475, Sample Num: 7600, Cur Loss: 1.02696490, Cur Avg Loss: 1.13033005, Log Avg loss: 1.13214880, Global Avg Loss: 1.46963942, Time: 0.0402 Steps: 96280, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000485, Sample Num: 7760, Cur Loss: 1.13990939, Cur Avg Loss: 1.13467039, Log Avg loss: 1.34083677, Global Avg Loss: 1.46962605, Time: 0.0402 Steps: 96290, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000495, Sample Num: 7920, Cur Loss: 0.47619829, Cur Avg Loss: 1.13746769, Log Avg loss: 1.27313666, Global Avg Loss: 1.46960564, Time: 0.0403 Steps: 96300, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000505, Sample Num: 8080, Cur Loss: 1.76664567, Cur Avg Loss: 1.13912293, Log Avg loss: 1.22105739, Global Avg Loss: 1.46957984, Time: 0.0402 Steps: 96310, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000515, Sample Num: 8240, Cur Loss: 0.87903857, Cur Avg Loss: 1.13798048, Log Avg loss: 1.08028688, Global Avg Loss: 1.46953942, Time: 0.0402 Steps: 96320, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000525, Sample Num: 8400, Cur Loss: 0.68123853, Cur Avg Loss: 1.13658933, Log Avg loss: 1.06494494, Global Avg Loss: 1.46949742, Time: 0.0403 Steps: 96330, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000535, Sample Num: 8560, Cur Loss: 0.79735982, Cur Avg Loss: 1.13581221, Log Avg loss: 1.09501320, Global Avg Loss: 1.46945855, Time: 0.0402 Steps: 96340, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000545, Sample Num: 8720, Cur Loss: 0.86504161, Cur Avg Loss: 1.13523901, Log Avg loss: 1.10457284, Global Avg Loss: 1.46942068, Time: 0.0403 Steps: 96350, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000555, Sample Num: 8880, Cur Loss: 0.61585075, Cur Avg Loss: 1.13039842, Log Avg loss: 0.86658647, Global Avg Loss: 1.46935812, Time: 0.0402 Steps: 96360, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000565, Sample Num: 9040, Cur Loss: 1.79397130, Cur Avg Loss: 1.12762179, Log Avg loss: 0.97351886, Global Avg Loss: 1.46930666, Time: 0.0402 Steps: 96370, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000575, Sample Num: 9200, Cur Loss: 0.36084896, Cur Avg Loss: 1.12590339, Log Avg loss: 1.02881388, Global Avg Loss: 1.46926096, Time: 0.0402 Steps: 96380, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000585, Sample Num: 9360, Cur Loss: 0.90124190, Cur Avg Loss: 1.12563765, Log Avg loss: 1.11035731, Global Avg Loss: 1.46922373, Time: 0.0402 Steps: 96390, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000595, Sample Num: 9520, Cur Loss: 0.99592978, Cur Avg Loss: 1.12137987, Log Avg loss: 0.87229964, Global Avg Loss: 1.46916180, Time: 0.0402 Steps: 96400, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000605, Sample Num: 9680, Cur Loss: 0.73817599, Cur Avg Loss: 1.12450015, Log Avg loss: 1.31015697, Global Avg Loss: 1.46914531, Time: 0.0402 Steps: 96410, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000615, Sample Num: 9840, Cur Loss: 0.69530302, Cur Avg Loss: 1.12674515, Log Avg loss: 1.26256788, Global Avg Loss: 1.46912389, Time: 0.0402 Steps: 96420, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000625, Sample Num: 10000, Cur Loss: 2.30991292, Cur Avg Loss: 1.12526460, Log Avg loss: 1.03421085, Global Avg Loss: 1.46907879, Time: 0.0402 Steps: 96430, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000635, Sample Num: 10160, Cur Loss: 2.12264323, Cur Avg Loss: 1.12442942, Log Avg loss: 1.07223008, Global Avg Loss: 1.46903764, Time: 0.0607 Steps: 96440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000645, Sample Num: 10320, Cur Loss: 1.52840734, Cur Avg Loss: 1.12334654, Log Avg loss: 1.05458383, Global Avg Loss: 1.46899467, Time: 0.0531 Steps: 96450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000655, Sample Num: 10480, Cur Loss: 0.51720136, Cur Avg Loss: 1.12159863, Log Avg loss: 1.00885878, Global Avg Loss: 1.46894696, Time: 0.0434 Steps: 96460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000665, Sample Num: 10640, Cur Loss: 1.81966376, Cur Avg Loss: 1.12302437, Log Avg loss: 1.21641040, Global Avg Loss: 1.46892079, Time: 0.0460 Steps: 96470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000675, Sample Num: 10800, Cur Loss: 1.14199364, Cur Avg Loss: 1.12124823, Log Avg loss: 1.00313495, Global Avg Loss: 1.46887251, Time: 0.0467 Steps: 96480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000685, Sample Num: 10960, Cur Loss: 0.40514022, Cur Avg Loss: 1.12104860, Log Avg loss: 1.10757363, Global Avg Loss: 1.46883506, Time: 0.0427 Steps: 96490, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000695, Sample Num: 11120, Cur Loss: 0.95130122, Cur Avg Loss: 1.11516807, Log Avg loss: 0.71235146, Global Avg Loss: 1.46875667, Time: 0.0505 Steps: 96500, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000705, Sample Num: 11280, Cur Loss: 1.13611579, Cur Avg Loss: 1.11375083, Log Avg loss: 1.01525267, Global Avg Loss: 1.46870968, Time: 0.0679 Steps: 96510, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000715, Sample Num: 11440, Cur Loss: 1.21308148, Cur Avg Loss: 1.11215804, Log Avg loss: 0.99986605, Global Avg Loss: 1.46866111, Time: 0.0402 Steps: 96520, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000725, Sample Num: 11600, Cur Loss: 0.97781169, Cur Avg Loss: 1.10930098, Log Avg loss: 0.90502148, Global Avg Loss: 1.46860272, Time: 0.0402 Steps: 96530, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000735, Sample Num: 11760, Cur Loss: 0.33794457, Cur Avg Loss: 1.10557983, Log Avg loss: 0.83579670, Global Avg Loss: 1.46853717, Time: 0.0402 Steps: 96540, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000745, Sample Num: 11920, Cur Loss: 0.41201147, Cur Avg Loss: 1.10613061, Log Avg loss: 1.14661259, Global Avg Loss: 1.46850382, Time: 0.0403 Steps: 96550, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000755, Sample Num: 12080, Cur Loss: 0.73377019, Cur Avg Loss: 1.10955763, Log Avg loss: 1.36487045, Global Avg Loss: 1.46849309, Time: 0.0403 Steps: 96560, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000765, Sample Num: 12240, Cur Loss: 0.83752978, Cur Avg Loss: 1.10815778, Log Avg loss: 1.00246928, Global Avg Loss: 1.46844483, Time: 0.0403 Steps: 96570, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000775, Sample Num: 12400, Cur Loss: 0.90760881, Cur Avg Loss: 1.11156516, Log Avg loss: 1.37222989, Global Avg Loss: 1.46843487, Time: 0.0402 Steps: 96580, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000785, Sample Num: 12560, Cur Loss: 0.97972542, Cur Avg Loss: 1.11656777, Log Avg loss: 1.50426986, Global Avg Loss: 1.46843858, Time: 0.0403 Steps: 96590, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000795, Sample Num: 12720, Cur Loss: 0.75211012, Cur Avg Loss: 1.11317559, Log Avg loss: 0.84688937, Global Avg Loss: 1.46837424, Time: 0.0402 Steps: 96600, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000805, Sample Num: 12880, Cur Loss: 0.33501533, Cur Avg Loss: 1.11511731, Log Avg loss: 1.26948389, Global Avg Loss: 1.46835365, Time: 0.0403 Steps: 96610, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000815, Sample Num: 13040, Cur Loss: 1.18754971, Cur Avg Loss: 1.11171938, Log Avg loss: 0.83818598, Global Avg Loss: 1.46828843, Time: 0.0403 Steps: 96620, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000825, Sample Num: 13200, Cur Loss: 1.54660189, Cur Avg Loss: 1.11233340, Log Avg loss: 1.16237679, Global Avg Loss: 1.46825677, Time: 0.0403 Steps: 96630, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000835, Sample Num: 13360, Cur Loss: 1.04700482, Cur Avg Loss: 1.11017387, Log Avg loss: 0.93201209, Global Avg Loss: 1.46820128, Time: 0.0403 Steps: 96640, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000845, Sample Num: 13520, Cur Loss: 0.96971065, Cur Avg Loss: 1.10802478, Log Avg loss: 0.92857611, Global Avg Loss: 1.46814545, Time: 0.0402 Steps: 96650, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000855, Sample Num: 13680, Cur Loss: 0.70234495, Cur Avg Loss: 1.10560158, Log Avg loss: 0.90084064, Global Avg Loss: 1.46808676, Time: 0.0403 Steps: 96660, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000865, Sample Num: 13840, Cur Loss: 1.39919138, Cur Avg Loss: 1.10541049, Log Avg loss: 1.08907244, Global Avg Loss: 1.46804755, Time: 0.0403 Steps: 96670, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000875, Sample Num: 14000, Cur Loss: 0.44223988, Cur Avg Loss: 1.10482953, Log Avg loss: 1.05457645, Global Avg Loss: 1.46800479, Time: 0.0402 Steps: 96680, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000885, Sample Num: 14160, Cur Loss: 0.72472084, Cur Avg Loss: 1.10561767, Log Avg loss: 1.17458019, Global Avg Loss: 1.46797444, Time: 0.0402 Steps: 96690, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000895, Sample Num: 14320, Cur Loss: 1.25242591, Cur Avg Loss: 1.10710269, Log Avg loss: 1.23852722, Global Avg Loss: 1.46795071, Time: 0.0401 Steps: 96700, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000905, Sample Num: 14480, Cur Loss: 0.67166746, Cur Avg Loss: 1.10622997, Log Avg loss: 1.02812120, Global Avg Loss: 1.46790523, Time: 0.0404 Steps: 96710, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000915, Sample Num: 14640, Cur Loss: 1.03728342, Cur Avg Loss: 1.10839152, Log Avg loss: 1.30401180, Global Avg Loss: 1.46788829, Time: 0.0403 Steps: 96720, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000925, Sample Num: 14800, Cur Loss: 1.83252454, Cur Avg Loss: 1.11033333, Log Avg loss: 1.28800943, Global Avg Loss: 1.46786969, Time: 0.0403 Steps: 96730, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000935, Sample Num: 14960, Cur Loss: 0.91011894, Cur Avg Loss: 1.10984516, Log Avg loss: 1.06468911, Global Avg Loss: 1.46782801, Time: 0.0403 Steps: 96740, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000945, Sample Num: 15120, Cur Loss: 0.33268824, Cur Avg Loss: 1.11040436, Log Avg loss: 1.16268965, Global Avg Loss: 1.46779648, Time: 0.0402 Steps: 96750, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000955, Sample Num: 15280, Cur Loss: 0.85663813, Cur Avg Loss: 1.10734052, Log Avg loss: 0.81780747, Global Avg Loss: 1.46772930, Time: 0.0402 Steps: 96760, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000965, Sample Num: 15440, Cur Loss: 0.63918656, Cur Avg Loss: 1.11025403, Log Avg loss: 1.38849446, Global Avg Loss: 1.46772111, Time: 0.0403 Steps: 96770, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000975, Sample Num: 15600, Cur Loss: 1.62923086, Cur Avg Loss: 1.11048529, Log Avg loss: 1.13280196, Global Avg Loss: 1.46768651, Time: 0.0471 Steps: 96780, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000985, Sample Num: 15760, Cur Loss: 0.60088229, Cur Avg Loss: 1.10813658, Log Avg loss: 0.87913655, Global Avg Loss: 1.46762570, Time: 0.0652 Steps: 96790, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000995, Sample Num: 15920, Cur Loss: 0.55722970, Cur Avg Loss: 1.10870305, Log Avg loss: 1.16450082, Global Avg Loss: 1.46759439, Time: 0.0405 Steps: 96800, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001005, Sample Num: 16080, Cur Loss: 0.89985663, Cur Avg Loss: 1.10788851, Log Avg loss: 1.02684130, Global Avg Loss: 1.46754886, Time: 0.0502 Steps: 96810, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001015, Sample Num: 16240, Cur Loss: 0.64648628, Cur Avg Loss: 1.11026462, Log Avg loss: 1.34906400, Global Avg Loss: 1.46753662, Time: 0.0506 Steps: 96820, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001025, Sample Num: 16400, Cur Loss: 1.30334735, Cur Avg Loss: 1.11178901, Log Avg loss: 1.26651503, Global Avg Loss: 1.46751586, Time: 0.0407 Steps: 96830, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001035, Sample Num: 16560, Cur Loss: 0.81565714, Cur Avg Loss: 1.11007760, Log Avg loss: 0.93465808, Global Avg Loss: 1.46746084, Time: 0.0833 Steps: 96840, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001045, Sample Num: 16720, Cur Loss: 1.25576174, Cur Avg Loss: 1.11295732, Log Avg loss: 1.41100760, Global Avg Loss: 1.46745501, Time: 0.0404 Steps: 96850, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001055, Sample Num: 16880, Cur Loss: 2.06167746, Cur Avg Loss: 1.11532120, Log Avg loss: 1.36234660, Global Avg Loss: 1.46744415, Time: 0.0403 Steps: 96860, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001065, Sample Num: 17040, Cur Loss: 0.65248328, Cur Avg Loss: 1.11316922, Log Avg loss: 0.88613596, Global Avg Loss: 1.46738415, Time: 0.0402 Steps: 96870, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001075, Sample Num: 17200, Cur Loss: 0.53267026, Cur Avg Loss: 1.11156444, Log Avg loss: 0.94065481, Global Avg Loss: 1.46732978, Time: 0.0403 Steps: 96880, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001085, Sample Num: 17360, Cur Loss: 0.90255928, Cur Avg Loss: 1.11070049, Log Avg loss: 1.01782603, Global Avg Loss: 1.46728338, Time: 0.0404 Steps: 96890, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001095, Sample Num: 17520, Cur Loss: 1.04901838, Cur Avg Loss: 1.11173628, Log Avg loss: 1.22411968, Global Avg Loss: 1.46725829, Time: 0.0404 Steps: 96900, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001105, Sample Num: 17680, Cur Loss: 1.40446377, Cur Avg Loss: 1.11084802, Log Avg loss: 1.01358389, Global Avg Loss: 1.46721147, Time: 0.0403 Steps: 96910, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001115, Sample Num: 17840, Cur Loss: 1.08659065, Cur Avg Loss: 1.11056405, Log Avg loss: 1.07918545, Global Avg Loss: 1.46717144, Time: 0.0402 Steps: 96920, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001125, Sample Num: 18000, Cur Loss: 1.83508444, Cur Avg Loss: 1.11158702, Log Avg loss: 1.22564803, Global Avg Loss: 1.46714652, Time: 0.0404 Steps: 96930, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001135, Sample Num: 18160, Cur Loss: 0.59884471, Cur Avg Loss: 1.11100998, Log Avg loss: 1.04609295, Global Avg Loss: 1.46710309, Time: 0.0404 Steps: 96940, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001145, Sample Num: 18320, Cur Loss: 0.81242490, Cur Avg Loss: 1.11321513, Log Avg loss: 1.36349940, Global Avg Loss: 1.46709240, Time: 0.0403 Steps: 96950, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001155, Sample Num: 18480, Cur Loss: 0.63430882, Cur Avg Loss: 1.11215441, Log Avg loss: 0.99070219, Global Avg Loss: 1.46704327, Time: 0.0403 Steps: 96960, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001165, Sample Num: 18640, Cur Loss: 1.20468354, Cur Avg Loss: 1.11222549, Log Avg loss: 1.12043512, Global Avg Loss: 1.46700752, Time: 0.0404 Steps: 96970, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001175, Sample Num: 18800, Cur Loss: 0.51132441, Cur Avg Loss: 1.11077818, Log Avg loss: 0.94216675, Global Avg Loss: 1.46695341, Time: 0.0402 Steps: 96980, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001185, Sample Num: 18960, Cur Loss: 0.81228900, Cur Avg Loss: 1.11183781, Log Avg loss: 1.23634425, Global Avg Loss: 1.46692963, Time: 0.0403 Steps: 96990, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001195, Sample Num: 19120, Cur Loss: 0.94095397, Cur Avg Loss: 1.11416650, Log Avg loss: 1.39011606, Global Avg Loss: 1.46692171, Time: 0.0402 Steps: 97000, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001205, Sample Num: 19280, Cur Loss: 0.79657257, Cur Avg Loss: 1.11133456, Log Avg loss: 0.77291834, Global Avg Loss: 1.46685017, Time: 0.0402 Steps: 97010, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001215, Sample Num: 19440, Cur Loss: 0.49767563, Cur Avg Loss: 1.11178821, Log Avg loss: 1.16645263, Global Avg Loss: 1.46681921, Time: 0.0402 Steps: 97020, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001225, Sample Num: 19600, Cur Loss: 1.62460494, Cur Avg Loss: 1.11302149, Log Avg loss: 1.26286487, Global Avg Loss: 1.46679819, Time: 0.0402 Steps: 97030, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001235, Sample Num: 19760, Cur Loss: 0.95763242, Cur Avg Loss: 1.11156355, Log Avg loss: 0.93296636, Global Avg Loss: 1.46674318, Time: 0.0402 Steps: 97040, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001245, Sample Num: 19920, Cur Loss: 0.28815651, Cur Avg Loss: 1.11090327, Log Avg loss: 1.02935875, Global Avg Loss: 1.46669811, Time: 0.0403 Steps: 97050, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001255, Sample Num: 20080, Cur Loss: 2.52578688, Cur Avg Loss: 1.11369745, Log Avg loss: 1.46157226, Global Avg Loss: 1.46669758, Time: 0.0403 Steps: 97060, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001265, Sample Num: 20240, Cur Loss: 0.59223539, Cur Avg Loss: 1.11387066, Log Avg loss: 1.13560848, Global Avg Loss: 1.46666347, Time: 0.0403 Steps: 97070, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001275, Sample Num: 20400, Cur Loss: 0.59774554, Cur Avg Loss: 1.11311791, Log Avg loss: 1.01789486, Global Avg Loss: 1.46661725, Time: 0.0402 Steps: 97080, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001285, Sample Num: 20560, Cur Loss: 1.30079031, Cur Avg Loss: 1.11408338, Log Avg loss: 1.23718110, Global Avg Loss: 1.46659362, Time: 0.0403 Steps: 97090, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001295, Sample Num: 20720, Cur Loss: 1.15671325, Cur Avg Loss: 1.11602529, Log Avg loss: 1.36556070, Global Avg Loss: 1.46658321, Time: 0.0403 Steps: 97100, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001305, Sample Num: 20880, Cur Loss: 0.68362772, Cur Avg Loss: 1.11605478, Log Avg loss: 1.11987314, Global Avg Loss: 1.46654751, Time: 0.0403 Steps: 97110, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001315, Sample Num: 21040, Cur Loss: 0.38004082, Cur Avg Loss: 1.11498526, Log Avg loss: 0.97541306, Global Avg Loss: 1.46649694, Time: 0.0698 Steps: 97120, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001325, Sample Num: 21200, Cur Loss: 2.17962646, Cur Avg Loss: 1.11662001, Log Avg loss: 1.33158983, Global Avg Loss: 1.46648305, Time: 0.1002 Steps: 97130, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001335, Sample Num: 21360, Cur Loss: 1.25252926, Cur Avg Loss: 1.11536193, Log Avg loss: 0.94866684, Global Avg Loss: 1.46642974, Time: 0.0406 Steps: 97140, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001345, Sample Num: 21520, Cur Loss: 0.93819356, Cur Avg Loss: 1.11438669, Log Avg loss: 0.98419202, Global Avg Loss: 1.46638010, Time: 0.0624 Steps: 97150, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001355, Sample Num: 21680, Cur Loss: 1.18009043, Cur Avg Loss: 1.11366560, Log Avg loss: 1.01667918, Global Avg Loss: 1.46633382, Time: 0.0528 Steps: 97160, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001365, Sample Num: 21840, Cur Loss: 1.31761670, Cur Avg Loss: 1.11586033, Log Avg loss: 1.41324651, Global Avg Loss: 1.46632836, Time: 0.1002 Steps: 97170, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001375, Sample Num: 22000, Cur Loss: 0.75023973, Cur Avg Loss: 1.11432108, Log Avg loss: 0.90421352, Global Avg Loss: 1.46627051, Time: 0.0589 Steps: 97180, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001385, Sample Num: 22160, Cur Loss: 0.54695052, Cur Avg Loss: 1.11402480, Log Avg loss: 1.07328556, Global Avg Loss: 1.46623008, Time: 0.0844 Steps: 97190, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001395, Sample Num: 22320, Cur Loss: 0.28442198, Cur Avg Loss: 1.11645393, Log Avg loss: 1.45288916, Global Avg Loss: 1.46622871, Time: 0.0404 Steps: 97200, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001405, Sample Num: 22480, Cur Loss: 1.27041388, Cur Avg Loss: 1.11682037, Log Avg loss: 1.16793805, Global Avg Loss: 1.46619802, Time: 0.0402 Steps: 97210, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001415, Sample Num: 22640, Cur Loss: 1.34460402, Cur Avg Loss: 1.11532304, Log Avg loss: 0.90494868, Global Avg Loss: 1.46614029, Time: 0.0404 Steps: 97220, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001425, Sample Num: 22800, Cur Loss: 1.24048436, Cur Avg Loss: 1.11480605, Log Avg loss: 1.04165215, Global Avg Loss: 1.46609663, Time: 0.0404 Steps: 97230, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001435, Sample Num: 22960, Cur Loss: 1.90633452, Cur Avg Loss: 1.11413979, Log Avg loss: 1.01919715, Global Avg Loss: 1.46605067, Time: 0.0403 Steps: 97240, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001445, Sample Num: 23120, Cur Loss: 1.37509608, Cur Avg Loss: 1.11593293, Log Avg loss: 1.37324787, Global Avg Loss: 1.46604113, Time: 0.0403 Steps: 97250, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001455, Sample Num: 23280, Cur Loss: 1.43598747, Cur Avg Loss: 1.11536069, Log Avg loss: 1.03267336, Global Avg Loss: 1.46599657, Time: 0.0403 Steps: 97260, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001465, Sample Num: 23440, Cur Loss: 0.31647658, Cur Avg Loss: 1.11536533, Log Avg loss: 1.11604015, Global Avg Loss: 1.46596060, Time: 0.0404 Steps: 97270, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001475, Sample Num: 23600, Cur Loss: 0.83784419, Cur Avg Loss: 1.11565276, Log Avg loss: 1.15776087, Global Avg Loss: 1.46592891, Time: 0.0402 Steps: 97280, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001485, Sample Num: 23760, Cur Loss: 2.15131187, Cur Avg Loss: 1.11618142, Log Avg loss: 1.19415891, Global Avg Loss: 1.46590098, Time: 0.0403 Steps: 97290, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001495, Sample Num: 23920, Cur Loss: 0.40457591, Cur Avg Loss: 1.11585531, Log Avg loss: 1.06742758, Global Avg Loss: 1.46586003, Time: 0.0403 Steps: 97300, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001505, Sample Num: 24080, Cur Loss: 1.35425675, Cur Avg Loss: 1.11692611, Log Avg loss: 1.27701149, Global Avg Loss: 1.46584062, Time: 0.0403 Steps: 97310, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001515, Sample Num: 24240, Cur Loss: 1.09343851, Cur Avg Loss: 1.11579198, Log Avg loss: 0.94510495, Global Avg Loss: 1.46578711, Time: 0.0403 Steps: 97320, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001525, Sample Num: 24400, Cur Loss: 1.22246480, Cur Avg Loss: 1.11563540, Log Avg loss: 1.09191378, Global Avg Loss: 1.46574870, Time: 0.0403 Steps: 97330, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001535, Sample Num: 24560, Cur Loss: 0.47045290, Cur Avg Loss: 1.11595466, Log Avg loss: 1.16464171, Global Avg Loss: 1.46571777, Time: 0.0403 Steps: 97340, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001545, Sample Num: 24720, Cur Loss: 0.98575068, Cur Avg Loss: 1.11490335, Log Avg loss: 0.95352769, Global Avg Loss: 1.46566515, Time: 0.0403 Steps: 97350, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001555, Sample Num: 24880, Cur Loss: 2.29731894, Cur Avg Loss: 1.11467205, Log Avg loss: 1.07893511, Global Avg Loss: 1.46562543, Time: 0.0402 Steps: 97360, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001565, Sample Num: 25040, Cur Loss: 1.60311532, Cur Avg Loss: 1.11434370, Log Avg loss: 1.06328549, Global Avg Loss: 1.46558411, Time: 0.0403 Steps: 97370, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001575, Sample Num: 25200, Cur Loss: 1.17805886, Cur Avg Loss: 1.11495592, Log Avg loss: 1.21076854, Global Avg Loss: 1.46555794, Time: 0.0402 Steps: 97380, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001585, Sample Num: 25360, Cur Loss: 1.29611778, Cur Avg Loss: 1.11655235, Log Avg loss: 1.36799019, Global Avg Loss: 1.46554793, Time: 0.0403 Steps: 97390, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001595, Sample Num: 25520, Cur Loss: 0.53575408, Cur Avg Loss: 1.11506020, Log Avg loss: 0.87855472, Global Avg Loss: 1.46548766, Time: 0.0403 Steps: 97400, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001605, Sample Num: 25680, Cur Loss: 1.21509624, Cur Avg Loss: 1.11565496, Log Avg loss: 1.21051953, Global Avg Loss: 1.46546148, Time: 0.0403 Steps: 97410, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001615, Sample Num: 25840, Cur Loss: 1.03291583, Cur Avg Loss: 1.11530690, Log Avg loss: 1.05944200, Global Avg Loss: 1.46541981, Time: 0.0403 Steps: 97420, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001625, Sample Num: 26000, Cur Loss: 0.67411447, Cur Avg Loss: 1.11618898, Log Avg loss: 1.25864551, Global Avg Loss: 1.46539858, Time: 0.0402 Steps: 97430, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001635, Sample Num: 26160, Cur Loss: 0.78555095, Cur Avg Loss: 1.11710357, Log Avg loss: 1.26572448, Global Avg Loss: 1.46537809, Time: 0.0402 Steps: 97440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001645, Sample Num: 26320, Cur Loss: 1.79826379, Cur Avg Loss: 1.11673540, Log Avg loss: 1.05653977, Global Avg Loss: 1.46533614, Time: 0.0402 Steps: 97450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001655, Sample Num: 26480, Cur Loss: 0.74242961, Cur Avg Loss: 1.11546826, Log Avg loss: 0.90702422, Global Avg Loss: 1.46527885, Time: 0.0624 Steps: 97460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001665, Sample Num: 26640, Cur Loss: 0.23605292, Cur Avg Loss: 1.11550390, Log Avg loss: 1.12140228, Global Avg Loss: 1.46524357, Time: 0.0565 Steps: 97470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001675, Sample Num: 26800, Cur Loss: 2.75488472, Cur Avg Loss: 1.11614447, Log Avg loss: 1.22279786, Global Avg Loss: 1.46521870, Time: 0.0539 Steps: 97480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001685, Sample Num: 26960, Cur Loss: 0.23129143, Cur Avg Loss: 1.11756485, Log Avg loss: 1.35547988, Global Avg Loss: 1.46520744, Time: 0.0780 Steps: 97490, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001695, Sample Num: 27120, Cur Loss: 0.90159416, Cur Avg Loss: 1.11634323, Log Avg loss: 0.91050011, Global Avg Loss: 1.46515055, Time: 0.0849 Steps: 97500, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001705, Sample Num: 27280, Cur Loss: 1.54726243, Cur Avg Loss: 1.11573475, Log Avg loss: 1.01259657, Global Avg Loss: 1.46510414, Time: 0.0518 Steps: 97510, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001715, Sample Num: 27440, Cur Loss: 1.42324972, Cur Avg Loss: 1.11572881, Log Avg loss: 1.11471641, Global Avg Loss: 1.46506821, Time: 0.0586 Steps: 97520, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001725, Sample Num: 27600, Cur Loss: 1.00281382, Cur Avg Loss: 1.11592912, Log Avg loss: 1.15028146, Global Avg Loss: 1.46503593, Time: 0.0402 Steps: 97530, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001735, Sample Num: 27760, Cur Loss: 1.46328187, Cur Avg Loss: 1.11796891, Log Avg loss: 1.46983267, Global Avg Loss: 1.46503643, Time: 0.0406 Steps: 97540, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001745, Sample Num: 27920, Cur Loss: 0.79158568, Cur Avg Loss: 1.11848997, Log Avg loss: 1.20889443, Global Avg Loss: 1.46501017, Time: 0.0403 Steps: 97550, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001755, Sample Num: 28080, Cur Loss: 0.69297922, Cur Avg Loss: 1.11865507, Log Avg loss: 1.14746544, Global Avg Loss: 1.46497762, Time: 0.0404 Steps: 97560, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001765, Sample Num: 28240, Cur Loss: 0.41614649, Cur Avg Loss: 1.11890940, Log Avg loss: 1.16354491, Global Avg Loss: 1.46494673, Time: 0.0403 Steps: 97570, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001775, Sample Num: 28400, Cur Loss: 0.47097588, Cur Avg Loss: 1.11761103, Log Avg loss: 0.88844823, Global Avg Loss: 1.46488765, Time: 0.0403 Steps: 97580, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001785, Sample Num: 28560, Cur Loss: 1.14169908, Cur Avg Loss: 1.11723894, Log Avg loss: 1.05119298, Global Avg Loss: 1.46484526, Time: 0.0403 Steps: 97590, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001795, Sample Num: 28720, Cur Loss: 1.44749081, Cur Avg Loss: 1.11705438, Log Avg loss: 1.08411041, Global Avg Loss: 1.46480625, Time: 0.0403 Steps: 97600, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001805, Sample Num: 28880, Cur Loss: 0.91997051, Cur Avg Loss: 1.11758205, Log Avg loss: 1.21229814, Global Avg Loss: 1.46478038, Time: 0.0403 Steps: 97610, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001815, Sample Num: 29040, Cur Loss: 1.73362339, Cur Avg Loss: 1.11756486, Log Avg loss: 1.11446336, Global Avg Loss: 1.46474449, Time: 0.0403 Steps: 97620, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001825, Sample Num: 29200, Cur Loss: 0.93143499, Cur Avg Loss: 1.11732777, Log Avg loss: 1.07429508, Global Avg Loss: 1.46470450, Time: 0.0404 Steps: 97630, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001835, Sample Num: 29360, Cur Loss: 1.06117880, Cur Avg Loss: 1.11745825, Log Avg loss: 1.14127174, Global Avg Loss: 1.46467137, Time: 0.0403 Steps: 97640, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001845, Sample Num: 29520, Cur Loss: 0.48052546, Cur Avg Loss: 1.11675068, Log Avg loss: 0.98691095, Global Avg Loss: 1.46462245, Time: 0.0403 Steps: 97650, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001855, Sample Num: 29680, Cur Loss: 1.36889470, Cur Avg Loss: 1.11659067, Log Avg loss: 1.08706894, Global Avg Loss: 1.46458379, Time: 0.0403 Steps: 97660, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001865, Sample Num: 29840, Cur Loss: 1.55607045, Cur Avg Loss: 1.11750442, Log Avg loss: 1.28700441, Global Avg Loss: 1.46456561, Time: 0.0403 Steps: 97670, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001875, Sample Num: 30000, Cur Loss: 1.31981504, Cur Avg Loss: 1.11751974, Log Avg loss: 1.12037767, Global Avg Loss: 1.46453037, Time: 0.0403 Steps: 97680, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001885, Sample Num: 30160, Cur Loss: 0.52207410, Cur Avg Loss: 1.11571229, Log Avg loss: 0.77681587, Global Avg Loss: 1.46445997, Time: 0.0402 Steps: 97690, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001895, Sample Num: 30320, Cur Loss: 1.70039749, Cur Avg Loss: 1.11768739, Log Avg loss: 1.48999219, Global Avg Loss: 1.46446258, Time: 0.0402 Steps: 97700, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001905, Sample Num: 30480, Cur Loss: 2.90498734, Cur Avg Loss: 1.11922131, Log Avg loss: 1.40989965, Global Avg Loss: 1.46445700, Time: 0.0403 Steps: 97710, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001915, Sample Num: 30640, Cur Loss: 0.84199631, Cur Avg Loss: 1.11686477, Log Avg loss: 0.66794383, Global Avg Loss: 1.46437549, Time: 0.0404 Steps: 97720, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001925, Sample Num: 30800, Cur Loss: 1.88740611, Cur Avg Loss: 1.11697306, Log Avg loss: 1.13771127, Global Avg Loss: 1.46434207, Time: 0.0403 Steps: 97730, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001935, Sample Num: 30960, Cur Loss: 0.95385009, Cur Avg Loss: 1.11681691, Log Avg loss: 1.08675785, Global Avg Loss: 1.46430343, Time: 0.0403 Steps: 97740, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001945, Sample Num: 31120, Cur Loss: 0.59987271, Cur Avg Loss: 1.11697123, Log Avg loss: 1.14683130, Global Avg Loss: 1.46427096, Time: 0.0402 Steps: 97750, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001955, Sample Num: 31280, Cur Loss: 1.08503056, Cur Avg Loss: 1.11731424, Log Avg loss: 1.18402964, Global Avg Loss: 1.46424229, Time: 0.0402 Steps: 97760, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001965, Sample Num: 31440, Cur Loss: 2.14111018, Cur Avg Loss: 1.11675122, Log Avg loss: 1.00668082, Global Avg Loss: 1.46419549, Time: 0.0402 Steps: 97770, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001975, Sample Num: 31600, Cur Loss: 1.72416413, Cur Avg Loss: 1.11758050, Log Avg loss: 1.28053392, Global Avg Loss: 1.46417671, Time: 0.0402 Steps: 97780, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001985, Sample Num: 31760, Cur Loss: 0.96746159, Cur Avg Loss: 1.11827402, Log Avg loss: 1.25524477, Global Avg Loss: 1.46415534, Time: 0.0402 Steps: 97790, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001995, Sample Num: 31920, Cur Loss: 0.72321665, Cur Avg Loss: 1.11808977, Log Avg loss: 1.08151639, Global Avg Loss: 1.46411622, Time: 0.0402 Steps: 97800, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002005, Sample Num: 32080, Cur Loss: 0.83966529, Cur Avg Loss: 1.11884030, Log Avg loss: 1.26856996, Global Avg Loss: 1.46409622, Time: 0.0405 Steps: 97810, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002015, Sample Num: 32240, Cur Loss: 0.46683994, Cur Avg Loss: 1.11944527, Log Avg loss: 1.24074203, Global Avg Loss: 1.46407339, Time: 0.0436 Steps: 97820, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002025, Sample Num: 32400, Cur Loss: 1.32484531, Cur Avg Loss: 1.11932688, Log Avg loss: 1.09547119, Global Avg Loss: 1.46403571, Time: 0.0588 Steps: 97830, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002035, Sample Num: 32560, Cur Loss: 1.27217197, Cur Avg Loss: 1.11817021, Log Avg loss: 0.88394532, Global Avg Loss: 1.46397642, Time: 0.0407 Steps: 97840, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002045, Sample Num: 32720, Cur Loss: 0.49486059, Cur Avg Loss: 1.11663174, Log Avg loss: 0.80355310, Global Avg Loss: 1.46390893, Time: 0.0533 Steps: 97850, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002055, Sample Num: 32880, Cur Loss: 1.26837587, Cur Avg Loss: 1.11536794, Log Avg loss: 0.85692068, Global Avg Loss: 1.46384690, Time: 0.0618 Steps: 97860, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002065, Sample Num: 33040, Cur Loss: 1.18379247, Cur Avg Loss: 1.11494042, Log Avg loss: 1.02708549, Global Avg Loss: 1.46380228, Time: 0.0671 Steps: 97870, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002075, Sample Num: 33200, Cur Loss: 0.36576751, Cur Avg Loss: 1.11523160, Log Avg loss: 1.17536036, Global Avg Loss: 1.46377281, Time: 0.0476 Steps: 97880, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002085, Sample Num: 33360, Cur Loss: 1.04109037, Cur Avg Loss: 1.11470519, Log Avg loss: 1.00547487, Global Avg Loss: 1.46372599, Time: 0.0402 Steps: 97890, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002095, Sample Num: 33520, Cur Loss: 1.86153531, Cur Avg Loss: 1.11558952, Log Avg loss: 1.29997129, Global Avg Loss: 1.46370926, Time: 0.0402 Steps: 97900, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002105, Sample Num: 33680, Cur Loss: 1.25214100, Cur Avg Loss: 1.11495860, Log Avg loss: 0.98278201, Global Avg Loss: 1.46366015, Time: 0.0403 Steps: 97910, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002115, Sample Num: 33840, Cur Loss: 2.37671924, Cur Avg Loss: 1.11657613, Log Avg loss: 1.45706500, Global Avg Loss: 1.46365947, Time: 0.0402 Steps: 97920, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002125, Sample Num: 34000, Cur Loss: 1.00654817, Cur Avg Loss: 1.11600566, Log Avg loss: 0.99535233, Global Avg Loss: 1.46361165, Time: 0.0404 Steps: 97930, Updated lr: 0.000008 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.351364, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.079558, "eval_total_loss": 758.929467, "eval_mae": 0.847219, "eval_mse": 1.079673, "eval_r2": 0.313689, "eval_sp_statistic": 0.47168, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.577888, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.071927, "test_total_loss": 538.107171, "test_mae": 0.860217, "test_mse": 1.072121, "test_r2": 0.308044, "test_sp_statistic": 0.491911, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.621708, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4635954279053383, "train_cur_epoch_loss": 2375.777693659067, "train_cur_epoch_avg_loss": 1.1159124911503369, "train_cur_epoch_time": 95.35136413574219, "train_cur_epoch_avg_time": 0.04478692538080892, "epoch": 46, "step": 97934} ################################################## Training, Epoch: 0047, Batch: 000006, Sample Num: 96, Cur Loss: 1.42326725, Cur Avg Loss: 1.31898839, Log Avg loss: 1.21795955, Global Avg Loss: 1.46358657, Time: 0.0403 Steps: 97940, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000016, Sample Num: 256, Cur Loss: 0.39371008, Cur Avg Loss: 0.97260994, Log Avg loss: 0.76478287, Global Avg Loss: 1.46351523, Time: 0.0403 Steps: 97950, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000026, Sample Num: 416, Cur Loss: 2.07375050, Cur Avg Loss: 1.02335493, Log Avg loss: 1.10454692, Global Avg Loss: 1.46347858, Time: 0.0402 Steps: 97960, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000036, Sample Num: 576, Cur Loss: 1.00000656, Cur Avg Loss: 1.07332011, Log Avg loss: 1.20322957, Global Avg Loss: 1.46345202, Time: 0.0402 Steps: 97970, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000046, Sample Num: 736, Cur Loss: 0.91321635, Cur Avg Loss: 1.01753983, Log Avg loss: 0.81673084, Global Avg Loss: 1.46338601, Time: 0.0403 Steps: 97980, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000056, Sample Num: 896, Cur Loss: 0.65989733, Cur Avg Loss: 1.05741864, Log Avg loss: 1.24086114, Global Avg Loss: 1.46336330, Time: 0.0403 Steps: 97990, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000066, Sample Num: 1056, Cur Loss: 2.06978512, Cur Avg Loss: 1.04073360, Log Avg loss: 0.94729742, Global Avg Loss: 1.46331064, Time: 0.0403 Steps: 98000, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000076, Sample Num: 1216, Cur Loss: 0.29098621, Cur Avg Loss: 1.04429075, Log Avg loss: 1.06776793, Global Avg Loss: 1.46327029, Time: 0.0403 Steps: 98010, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000086, Sample Num: 1376, Cur Loss: 2.07292986, Cur Avg Loss: 1.04404969, Log Avg loss: 1.04221764, Global Avg Loss: 1.46322733, Time: 0.0403 Steps: 98020, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000096, Sample Num: 1536, Cur Loss: 0.37525052, Cur Avg Loss: 1.04094968, Log Avg loss: 1.01428958, Global Avg Loss: 1.46318153, Time: 0.0402 Steps: 98030, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000106, Sample Num: 1696, Cur Loss: 1.08819914, Cur Avg Loss: 1.05570996, Log Avg loss: 1.19740859, Global Avg Loss: 1.46315443, Time: 0.0402 Steps: 98040, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000116, Sample Num: 1856, Cur Loss: 2.35768747, Cur Avg Loss: 1.07124383, Log Avg loss: 1.23590292, Global Avg Loss: 1.46313125, Time: 0.0402 Steps: 98050, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000126, Sample Num: 2016, Cur Loss: 0.66992527, Cur Avg Loss: 1.05237220, Log Avg loss: 0.83346125, Global Avg Loss: 1.46306704, Time: 0.0403 Steps: 98060, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000136, Sample Num: 2176, Cur Loss: 2.04840922, Cur Avg Loss: 1.06449937, Log Avg loss: 1.21730178, Global Avg Loss: 1.46304198, Time: 0.0402 Steps: 98070, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000146, Sample Num: 2336, Cur Loss: 1.22077787, Cur Avg Loss: 1.08208813, Log Avg loss: 1.32129521, Global Avg Loss: 1.46302752, Time: 0.0402 Steps: 98080, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000156, Sample Num: 2496, Cur Loss: 2.23427773, Cur Avg Loss: 1.07837723, Log Avg loss: 1.02419816, Global Avg Loss: 1.46298279, Time: 0.0402 Steps: 98090, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000166, Sample Num: 2656, Cur Loss: 0.61104912, Cur Avg Loss: 1.06543884, Log Avg loss: 0.86359988, Global Avg Loss: 1.46292169, Time: 0.0402 Steps: 98100, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000176, Sample Num: 2816, Cur Loss: 2.86671877, Cur Avg Loss: 1.07347313, Log Avg loss: 1.20684227, Global Avg Loss: 1.46289559, Time: 0.0402 Steps: 98110, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000186, Sample Num: 2976, Cur Loss: 0.34422424, Cur Avg Loss: 1.07514257, Log Avg loss: 1.10452489, Global Avg Loss: 1.46285906, Time: 0.0402 Steps: 98120, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000196, Sample Num: 3136, Cur Loss: 0.84296405, Cur Avg Loss: 1.07129413, Log Avg loss: 0.99971298, Global Avg Loss: 1.46281186, Time: 0.0403 Steps: 98130, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000206, Sample Num: 3296, Cur Loss: 0.82579547, Cur Avg Loss: 1.08794535, Log Avg loss: 1.41430925, Global Avg Loss: 1.46280692, Time: 0.0402 Steps: 98140, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000216, Sample Num: 3456, Cur Loss: 1.20547903, Cur Avg Loss: 1.10300453, Log Avg loss: 1.41322367, Global Avg Loss: 1.46280187, Time: 0.0403 Steps: 98150, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000226, Sample Num: 3616, Cur Loss: 1.17677152, Cur Avg Loss: 1.09044450, Log Avg loss: 0.81914781, Global Avg Loss: 1.46273630, Time: 0.0402 Steps: 98160, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000236, Sample Num: 3776, Cur Loss: 0.33906093, Cur Avg Loss: 1.08733745, Log Avg loss: 1.01711822, Global Avg Loss: 1.46269091, Time: 0.0402 Steps: 98170, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000246, Sample Num: 3936, Cur Loss: 0.58700103, Cur Avg Loss: 1.09917266, Log Avg loss: 1.37848369, Global Avg Loss: 1.46268233, Time: 0.0402 Steps: 98180, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000256, Sample Num: 4096, Cur Loss: 0.92425483, Cur Avg Loss: 1.10154537, Log Avg loss: 1.15991405, Global Avg Loss: 1.46265149, Time: 0.0402 Steps: 98190, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000266, Sample Num: 4256, Cur Loss: 1.05228496, Cur Avg Loss: 1.09210546, Log Avg loss: 0.85044355, Global Avg Loss: 1.46258915, Time: 0.0708 Steps: 98200, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000276, Sample Num: 4416, Cur Loss: 0.74220300, Cur Avg Loss: 1.08930683, Log Avg loss: 1.01486341, Global Avg Loss: 1.46254356, Time: 0.0637 Steps: 98210, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000286, Sample Num: 4576, Cur Loss: 0.75185132, Cur Avg Loss: 1.10035397, Log Avg loss: 1.40525507, Global Avg Loss: 1.46253773, Time: 0.0740 Steps: 98220, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000296, Sample Num: 4736, Cur Loss: 0.82758433, Cur Avg Loss: 1.09593631, Log Avg loss: 0.96959122, Global Avg Loss: 1.46248755, Time: 0.0638 Steps: 98230, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000306, Sample Num: 4896, Cur Loss: 1.12286949, Cur Avg Loss: 1.09686120, Log Avg loss: 1.12423801, Global Avg Loss: 1.46245312, Time: 0.0408 Steps: 98240, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000316, Sample Num: 5056, Cur Loss: 1.45844316, Cur Avg Loss: 1.10969403, Log Avg loss: 1.50237852, Global Avg Loss: 1.46245718, Time: 0.0407 Steps: 98250, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000326, Sample Num: 5216, Cur Loss: 0.61109006, Cur Avg Loss: 1.10924517, Log Avg loss: 1.09506110, Global Avg Loss: 1.46241979, Time: 0.0405 Steps: 98260, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000336, Sample Num: 5376, Cur Loss: 0.60781580, Cur Avg Loss: 1.11182594, Log Avg loss: 1.19595900, Global Avg Loss: 1.46239267, Time: 0.0405 Steps: 98270, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000346, Sample Num: 5536, Cur Loss: 2.89783263, Cur Avg Loss: 1.12274254, Log Avg loss: 1.48954045, Global Avg Loss: 1.46239544, Time: 0.0402 Steps: 98280, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000356, Sample Num: 5696, Cur Loss: 0.61611199, Cur Avg Loss: 1.12227446, Log Avg loss: 1.10607897, Global Avg Loss: 1.46235919, Time: 0.0404 Steps: 98290, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000366, Sample Num: 5856, Cur Loss: 1.10492289, Cur Avg Loss: 1.12291369, Log Avg loss: 1.14567025, Global Avg Loss: 1.46232697, Time: 0.0403 Steps: 98300, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000376, Sample Num: 6016, Cur Loss: 1.02143312, Cur Avg Loss: 1.11937181, Log Avg loss: 0.98973913, Global Avg Loss: 1.46227890, Time: 0.0402 Steps: 98310, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000386, Sample Num: 6176, Cur Loss: 1.22346961, Cur Avg Loss: 1.11659195, Log Avg loss: 1.01206914, Global Avg Loss: 1.46223311, Time: 0.0403 Steps: 98320, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000396, Sample Num: 6336, Cur Loss: 0.99736106, Cur Avg Loss: 1.11377596, Log Avg loss: 1.00507884, Global Avg Loss: 1.46218662, Time: 0.0402 Steps: 98330, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000406, Sample Num: 6496, Cur Loss: 0.64705616, Cur Avg Loss: 1.12166407, Log Avg loss: 1.43403316, Global Avg Loss: 1.46218375, Time: 0.0403 Steps: 98340, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000416, Sample Num: 6656, Cur Loss: 0.57489228, Cur Avg Loss: 1.12010475, Log Avg loss: 1.05679605, Global Avg Loss: 1.46214253, Time: 0.0403 Steps: 98350, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000426, Sample Num: 6816, Cur Loss: 0.77378178, Cur Avg Loss: 1.11880850, Log Avg loss: 1.06488471, Global Avg Loss: 1.46210215, Time: 0.0402 Steps: 98360, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000436, Sample Num: 6976, Cur Loss: 0.82466561, Cur Avg Loss: 1.12281966, Log Avg loss: 1.29369509, Global Avg Loss: 1.46208503, Time: 0.0403 Steps: 98370, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000446, Sample Num: 7136, Cur Loss: 2.08681583, Cur Avg Loss: 1.12414816, Log Avg loss: 1.18207083, Global Avg Loss: 1.46205656, Time: 0.0403 Steps: 98380, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000456, Sample Num: 7296, Cur Loss: 0.73593020, Cur Avg Loss: 1.12122297, Log Avg loss: 0.99075954, Global Avg Loss: 1.46200866, Time: 0.0403 Steps: 98390, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000466, Sample Num: 7456, Cur Loss: 1.76457477, Cur Avg Loss: 1.12465298, Log Avg loss: 1.28106127, Global Avg Loss: 1.46199027, Time: 0.0403 Steps: 98400, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000476, Sample Num: 7616, Cur Loss: 1.40049362, Cur Avg Loss: 1.12335166, Log Avg loss: 1.06271009, Global Avg Loss: 1.46194970, Time: 0.0402 Steps: 98410, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000486, Sample Num: 7776, Cur Loss: 2.89824247, Cur Avg Loss: 1.12608998, Log Avg loss: 1.25643393, Global Avg Loss: 1.46192882, Time: 0.0402 Steps: 98420, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000496, Sample Num: 7936, Cur Loss: 1.59705877, Cur Avg Loss: 1.12973590, Log Avg loss: 1.30692764, Global Avg Loss: 1.46191307, Time: 0.0402 Steps: 98430, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000506, Sample Num: 8096, Cur Loss: 0.51966000, Cur Avg Loss: 1.12777946, Log Avg loss: 1.03074031, Global Avg Loss: 1.46186927, Time: 0.0402 Steps: 98440, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000516, Sample Num: 8256, Cur Loss: 1.59519029, Cur Avg Loss: 1.12499295, Log Avg loss: 0.98399515, Global Avg Loss: 1.46182073, Time: 0.0402 Steps: 98450, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000526, Sample Num: 8416, Cur Loss: 1.11391187, Cur Avg Loss: 1.12681598, Log Avg loss: 1.22088448, Global Avg Loss: 1.46179626, Time: 0.0402 Steps: 98460, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000536, Sample Num: 8576, Cur Loss: 1.49600029, Cur Avg Loss: 1.12676548, Log Avg loss: 1.12410903, Global Avg Loss: 1.46176197, Time: 0.0402 Steps: 98470, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000546, Sample Num: 8736, Cur Loss: 0.68535191, Cur Avg Loss: 1.12296566, Log Avg loss: 0.91929560, Global Avg Loss: 1.46170688, Time: 0.0402 Steps: 98480, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000556, Sample Num: 8896, Cur Loss: 1.52795863, Cur Avg Loss: 1.12279099, Log Avg loss: 1.11325413, Global Avg Loss: 1.46167150, Time: 0.0402 Steps: 98490, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000566, Sample Num: 9056, Cur Loss: 1.30362964, Cur Avg Loss: 1.11853527, Log Avg loss: 0.88191692, Global Avg Loss: 1.46161265, Time: 0.0402 Steps: 98500, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000576, Sample Num: 9216, Cur Loss: 1.94608450, Cur Avg Loss: 1.12029020, Log Avg loss: 1.21961919, Global Avg Loss: 1.46158808, Time: 0.0402 Steps: 98510, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000586, Sample Num: 9376, Cur Loss: 0.54873294, Cur Avg Loss: 1.11312319, Log Avg loss: 0.70030387, Global Avg Loss: 1.46151081, Time: 0.0402 Steps: 98520, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000596, Sample Num: 9536, Cur Loss: 1.19285822, Cur Avg Loss: 1.10973500, Log Avg loss: 0.91118671, Global Avg Loss: 1.46145495, Time: 0.0677 Steps: 98530, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000606, Sample Num: 9696, Cur Loss: 1.05294991, Cur Avg Loss: 1.10497966, Log Avg loss: 0.82156135, Global Avg Loss: 1.46139002, Time: 0.0592 Steps: 98540, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000616, Sample Num: 9856, Cur Loss: 0.71434301, Cur Avg Loss: 1.10342235, Log Avg loss: 1.00904959, Global Avg Loss: 1.46134412, Time: 0.0432 Steps: 98550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000626, Sample Num: 10016, Cur Loss: 1.84843385, Cur Avg Loss: 1.10996746, Log Avg loss: 1.51314592, Global Avg Loss: 1.46134937, Time: 0.0561 Steps: 98560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000636, Sample Num: 10176, Cur Loss: 1.20531416, Cur Avg Loss: 1.10951758, Log Avg loss: 1.08135540, Global Avg Loss: 1.46131082, Time: 0.0610 Steps: 98570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000646, Sample Num: 10336, Cur Loss: 0.32991049, Cur Avg Loss: 1.10582632, Log Avg loss: 0.87106230, Global Avg Loss: 1.46125095, Time: 0.0560 Steps: 98580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000656, Sample Num: 10496, Cur Loss: 1.93821609, Cur Avg Loss: 1.10232637, Log Avg loss: 0.87622944, Global Avg Loss: 1.46119161, Time: 0.0522 Steps: 98590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000666, Sample Num: 10656, Cur Loss: 0.93673587, Cur Avg Loss: 1.10234483, Log Avg loss: 1.10355592, Global Avg Loss: 1.46115534, Time: 0.0728 Steps: 98600, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000676, Sample Num: 10816, Cur Loss: 0.50564170, Cur Avg Loss: 1.09911817, Log Avg loss: 0.88422227, Global Avg Loss: 1.46109683, Time: 0.0402 Steps: 98610, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000686, Sample Num: 10976, Cur Loss: 0.47288716, Cur Avg Loss: 1.10228425, Log Avg loss: 1.31631174, Global Avg Loss: 1.46108215, Time: 0.0402 Steps: 98620, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000696, Sample Num: 11136, Cur Loss: 0.99312472, Cur Avg Loss: 1.10160230, Log Avg loss: 1.05482024, Global Avg Loss: 1.46104096, Time: 0.0403 Steps: 98630, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000706, Sample Num: 11296, Cur Loss: 0.68596977, Cur Avg Loss: 1.10338234, Log Avg loss: 1.22727310, Global Avg Loss: 1.46101726, Time: 0.0402 Steps: 98640, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000716, Sample Num: 11456, Cur Loss: 0.74329180, Cur Avg Loss: 1.10094615, Log Avg loss: 0.92895100, Global Avg Loss: 1.46096333, Time: 0.0402 Steps: 98650, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000726, Sample Num: 11616, Cur Loss: 0.56648988, Cur Avg Loss: 1.09816939, Log Avg loss: 0.89935387, Global Avg Loss: 1.46090640, Time: 0.0402 Steps: 98660, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000736, Sample Num: 11776, Cur Loss: 0.67906356, Cur Avg Loss: 1.09593349, Log Avg loss: 0.93360697, Global Avg Loss: 1.46085296, Time: 0.0402 Steps: 98670, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000746, Sample Num: 11936, Cur Loss: 1.88223267, Cur Avg Loss: 1.09726173, Log Avg loss: 1.19501970, Global Avg Loss: 1.46082602, Time: 0.0403 Steps: 98680, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000756, Sample Num: 12096, Cur Loss: 0.42091459, Cur Avg Loss: 1.10064735, Log Avg loss: 1.35321484, Global Avg Loss: 1.46081512, Time: 0.0403 Steps: 98690, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000766, Sample Num: 12256, Cur Loss: 1.58244121, Cur Avg Loss: 1.10302538, Log Avg loss: 1.28280433, Global Avg Loss: 1.46079708, Time: 0.0403 Steps: 98700, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000776, Sample Num: 12416, Cur Loss: 2.41761255, Cur Avg Loss: 1.10719895, Log Avg loss: 1.42689466, Global Avg Loss: 1.46079365, Time: 0.0403 Steps: 98710, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000786, Sample Num: 12576, Cur Loss: 1.25526643, Cur Avg Loss: 1.10697912, Log Avg loss: 1.08992015, Global Avg Loss: 1.46075608, Time: 0.0402 Steps: 98720, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000796, Sample Num: 12736, Cur Loss: 1.11068273, Cur Avg Loss: 1.10937773, Log Avg loss: 1.29790862, Global Avg Loss: 1.46073959, Time: 0.0403 Steps: 98730, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000806, Sample Num: 12896, Cur Loss: 1.55332422, Cur Avg Loss: 1.11144722, Log Avg loss: 1.27617823, Global Avg Loss: 1.46072089, Time: 0.0402 Steps: 98740, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000816, Sample Num: 13056, Cur Loss: 0.80653375, Cur Avg Loss: 1.10773534, Log Avg loss: 0.80855828, Global Avg Loss: 1.46065485, Time: 0.0402 Steps: 98750, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000826, Sample Num: 13216, Cur Loss: 1.79925442, Cur Avg Loss: 1.10898186, Log Avg loss: 1.21069789, Global Avg Loss: 1.46062954, Time: 0.0403 Steps: 98760, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000836, Sample Num: 13376, Cur Loss: 0.90902489, Cur Avg Loss: 1.10739516, Log Avg loss: 0.97633399, Global Avg Loss: 1.46058051, Time: 0.0404 Steps: 98770, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000846, Sample Num: 13536, Cur Loss: 0.38552341, Cur Avg Loss: 1.10439776, Log Avg loss: 0.85381500, Global Avg Loss: 1.46051908, Time: 0.0402 Steps: 98780, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000856, Sample Num: 13696, Cur Loss: 1.22932649, Cur Avg Loss: 1.11031644, Log Avg loss: 1.61103680, Global Avg Loss: 1.46053432, Time: 0.0402 Steps: 98790, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000866, Sample Num: 13856, Cur Loss: 1.33073258, Cur Avg Loss: 1.11022561, Log Avg loss: 1.10245004, Global Avg Loss: 1.46049808, Time: 0.0403 Steps: 98800, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000876, Sample Num: 14016, Cur Loss: 1.63940561, Cur Avg Loss: 1.11256146, Log Avg loss: 1.31484593, Global Avg Loss: 1.46048334, Time: 0.0402 Steps: 98810, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000886, Sample Num: 14176, Cur Loss: 0.52508652, Cur Avg Loss: 1.11079866, Log Avg loss: 0.95637768, Global Avg Loss: 1.46043232, Time: 0.0402 Steps: 98820, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000896, Sample Num: 14336, Cur Loss: 1.03192151, Cur Avg Loss: 1.11438599, Log Avg loss: 1.43222380, Global Avg Loss: 1.46042947, Time: 0.0403 Steps: 98830, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000906, Sample Num: 14496, Cur Loss: 0.58835125, Cur Avg Loss: 1.11549395, Log Avg loss: 1.21476686, Global Avg Loss: 1.46040462, Time: 0.0402 Steps: 98840, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000916, Sample Num: 14656, Cur Loss: 1.26541269, Cur Avg Loss: 1.11786266, Log Avg loss: 1.33246770, Global Avg Loss: 1.46039167, Time: 0.0402 Steps: 98850, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000926, Sample Num: 14816, Cur Loss: 1.02025056, Cur Avg Loss: 1.11882584, Log Avg loss: 1.20705293, Global Avg Loss: 1.46036605, Time: 0.0402 Steps: 98860, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000936, Sample Num: 14976, Cur Loss: 0.49172491, Cur Avg Loss: 1.11470053, Log Avg loss: 0.73269751, Global Avg Loss: 1.46029245, Time: 0.0402 Steps: 98870, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000946, Sample Num: 15136, Cur Loss: 1.22010994, Cur Avg Loss: 1.11213558, Log Avg loss: 0.87205618, Global Avg Loss: 1.46023296, Time: 0.0403 Steps: 98880, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000956, Sample Num: 15296, Cur Loss: 0.63334113, Cur Avg Loss: 1.11395189, Log Avg loss: 1.28577439, Global Avg Loss: 1.46021532, Time: 0.0759 Steps: 98890, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000966, Sample Num: 15456, Cur Loss: 1.07213485, Cur Avg Loss: 1.11484176, Log Avg loss: 1.19991380, Global Avg Loss: 1.46018900, Time: 0.0404 Steps: 98900, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000976, Sample Num: 15616, Cur Loss: 0.88637376, Cur Avg Loss: 1.11457360, Log Avg loss: 1.08866951, Global Avg Loss: 1.46015144, Time: 0.0670 Steps: 98910, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000986, Sample Num: 15776, Cur Loss: 1.86673903, Cur Avg Loss: 1.11258828, Log Avg loss: 0.91882072, Global Avg Loss: 1.46009671, Time: 0.0726 Steps: 98920, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000996, Sample Num: 15936, Cur Loss: 0.42906803, Cur Avg Loss: 1.10935133, Log Avg loss: 0.79018781, Global Avg Loss: 1.46002900, Time: 0.0585 Steps: 98930, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001006, Sample Num: 16096, Cur Loss: 2.31462336, Cur Avg Loss: 1.10762023, Log Avg loss: 0.93520268, Global Avg Loss: 1.45997595, Time: 0.0976 Steps: 98940, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001016, Sample Num: 16256, Cur Loss: 1.56215382, Cur Avg Loss: 1.10723167, Log Avg loss: 1.06814243, Global Avg Loss: 1.45993635, Time: 0.0403 Steps: 98950, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001026, Sample Num: 16416, Cur Loss: 1.85535598, Cur Avg Loss: 1.10576876, Log Avg loss: 0.95713698, Global Avg Loss: 1.45988554, Time: 0.0794 Steps: 98960, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001036, Sample Num: 16576, Cur Loss: 0.74418342, Cur Avg Loss: 1.10503559, Log Avg loss: 1.02981272, Global Avg Loss: 1.45984209, Time: 0.0402 Steps: 98970, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001046, Sample Num: 16736, Cur Loss: 0.66095805, Cur Avg Loss: 1.10478528, Log Avg loss: 1.07885264, Global Avg Loss: 1.45980360, Time: 0.0402 Steps: 98980, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001056, Sample Num: 16896, Cur Loss: 1.49982047, Cur Avg Loss: 1.10764177, Log Avg loss: 1.40643107, Global Avg Loss: 1.45979821, Time: 0.0402 Steps: 98990, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001066, Sample Num: 17056, Cur Loss: 1.45407808, Cur Avg Loss: 1.10543614, Log Avg loss: 0.87252125, Global Avg Loss: 1.45973888, Time: 0.0403 Steps: 99000, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001076, Sample Num: 17216, Cur Loss: 2.47052360, Cur Avg Loss: 1.10767241, Log Avg loss: 1.34605893, Global Avg Loss: 1.45972740, Time: 0.0402 Steps: 99010, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001086, Sample Num: 17376, Cur Loss: 1.02633369, Cur Avg Loss: 1.10676074, Log Avg loss: 1.00866528, Global Avg Loss: 1.45968185, Time: 0.0402 Steps: 99020, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001096, Sample Num: 17536, Cur Loss: 0.77319890, Cur Avg Loss: 1.10951231, Log Avg loss: 1.40833277, Global Avg Loss: 1.45967666, Time: 0.0403 Steps: 99030, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001106, Sample Num: 17696, Cur Loss: 1.20743120, Cur Avg Loss: 1.10989614, Log Avg loss: 1.15196356, Global Avg Loss: 1.45964560, Time: 0.0403 Steps: 99040, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001116, Sample Num: 17856, Cur Loss: 0.61467129, Cur Avg Loss: 1.10973887, Log Avg loss: 1.09234526, Global Avg Loss: 1.45960851, Time: 0.0403 Steps: 99050, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001126, Sample Num: 18016, Cur Loss: 1.00418615, Cur Avg Loss: 1.10981835, Log Avg loss: 1.11868787, Global Avg Loss: 1.45957410, Time: 0.0402 Steps: 99060, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001136, Sample Num: 18176, Cur Loss: 0.81335634, Cur Avg Loss: 1.10979340, Log Avg loss: 1.10698459, Global Avg Loss: 1.45953851, Time: 0.0403 Steps: 99070, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001146, Sample Num: 18336, Cur Loss: 1.26616025, Cur Avg Loss: 1.11131790, Log Avg loss: 1.28450150, Global Avg Loss: 1.45952084, Time: 0.0403 Steps: 99080, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001156, Sample Num: 18496, Cur Loss: 1.67202663, Cur Avg Loss: 1.11120891, Log Avg loss: 1.09871811, Global Avg Loss: 1.45948443, Time: 0.0402 Steps: 99090, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001166, Sample Num: 18656, Cur Loss: 1.37400198, Cur Avg Loss: 1.11278473, Log Avg loss: 1.29494913, Global Avg Loss: 1.45946783, Time: 0.0402 Steps: 99100, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001176, Sample Num: 18816, Cur Loss: 1.10187352, Cur Avg Loss: 1.11407675, Log Avg loss: 1.26472644, Global Avg Loss: 1.45944818, Time: 0.0402 Steps: 99110, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001186, Sample Num: 18976, Cur Loss: 1.78434300, Cur Avg Loss: 1.11410343, Log Avg loss: 1.11724135, Global Avg Loss: 1.45941365, Time: 0.0402 Steps: 99120, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001196, Sample Num: 19136, Cur Loss: 0.76478893, Cur Avg Loss: 1.11400450, Log Avg loss: 1.10227189, Global Avg Loss: 1.45937763, Time: 0.0403 Steps: 99130, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001206, Sample Num: 19296, Cur Loss: 1.10897064, Cur Avg Loss: 1.11377816, Log Avg loss: 1.08670680, Global Avg Loss: 1.45934004, Time: 0.0402 Steps: 99140, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001216, Sample Num: 19456, Cur Loss: 1.74222100, Cur Avg Loss: 1.11307284, Log Avg loss: 1.02801144, Global Avg Loss: 1.45929653, Time: 0.0402 Steps: 99150, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001226, Sample Num: 19616, Cur Loss: 1.44849145, Cur Avg Loss: 1.11230511, Log Avg loss: 1.01894951, Global Avg Loss: 1.45925212, Time: 0.0402 Steps: 99160, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001236, Sample Num: 19776, Cur Loss: 1.63817620, Cur Avg Loss: 1.11280975, Log Avg loss: 1.17467812, Global Avg Loss: 1.45922343, Time: 0.0402 Steps: 99170, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001246, Sample Num: 19936, Cur Loss: 0.70515633, Cur Avg Loss: 1.11112226, Log Avg loss: 0.90254846, Global Avg Loss: 1.45916730, Time: 0.0402 Steps: 99180, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001256, Sample Num: 20096, Cur Loss: 0.56326801, Cur Avg Loss: 1.11030538, Log Avg loss: 1.00852202, Global Avg Loss: 1.45912187, Time: 0.0403 Steps: 99190, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001266, Sample Num: 20256, Cur Loss: 0.67659694, Cur Avg Loss: 1.10913781, Log Avg loss: 0.96249222, Global Avg Loss: 1.45907181, Time: 0.0403 Steps: 99200, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001276, Sample Num: 20416, Cur Loss: 0.48906788, Cur Avg Loss: 1.11108870, Log Avg loss: 1.35807125, Global Avg Loss: 1.45906163, Time: 0.0403 Steps: 99210, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001286, Sample Num: 20576, Cur Loss: 1.80885649, Cur Avg Loss: 1.11241292, Log Avg loss: 1.28138309, Global Avg Loss: 1.45904372, Time: 0.0403 Steps: 99220, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001296, Sample Num: 20736, Cur Loss: 1.70472217, Cur Avg Loss: 1.11387718, Log Avg loss: 1.30218034, Global Avg Loss: 1.45902791, Time: 0.0402 Steps: 99230, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001306, Sample Num: 20896, Cur Loss: 0.54881346, Cur Avg Loss: 1.11257906, Log Avg loss: 0.94434290, Global Avg Loss: 1.45897605, Time: 0.0577 Steps: 99240, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001316, Sample Num: 21056, Cur Loss: 0.91368490, Cur Avg Loss: 1.11470139, Log Avg loss: 1.39187816, Global Avg Loss: 1.45896929, Time: 0.0853 Steps: 99250, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001326, Sample Num: 21216, Cur Loss: 0.58628190, Cur Avg Loss: 1.11445270, Log Avg loss: 1.08172489, Global Avg Loss: 1.45893128, Time: 0.0731 Steps: 99260, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001336, Sample Num: 21376, Cur Loss: 2.02931929, Cur Avg Loss: 1.11608815, Log Avg loss: 1.33294902, Global Avg Loss: 1.45891859, Time: 0.0483 Steps: 99270, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001346, Sample Num: 21536, Cur Loss: 1.13258207, Cur Avg Loss: 1.11574825, Log Avg loss: 1.07033686, Global Avg Loss: 1.45887945, Time: 0.0818 Steps: 99280, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001356, Sample Num: 21696, Cur Loss: 0.58724517, Cur Avg Loss: 1.11413779, Log Avg loss: 0.89737049, Global Avg Loss: 1.45882290, Time: 0.0624 Steps: 99290, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001366, Sample Num: 21856, Cur Loss: 1.56560361, Cur Avg Loss: 1.11627000, Log Avg loss: 1.40539776, Global Avg Loss: 1.45881752, Time: 0.0681 Steps: 99300, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001376, Sample Num: 22016, Cur Loss: 0.66535616, Cur Avg Loss: 1.11789440, Log Avg loss: 1.33978753, Global Avg Loss: 1.45880553, Time: 0.0493 Steps: 99310, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001386, Sample Num: 22176, Cur Loss: 0.45929030, Cur Avg Loss: 1.11752045, Log Avg loss: 1.06606483, Global Avg Loss: 1.45876599, Time: 0.0626 Steps: 99320, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001396, Sample Num: 22336, Cur Loss: 0.39162034, Cur Avg Loss: 1.11948797, Log Avg loss: 1.39218655, Global Avg Loss: 1.45875929, Time: 0.0403 Steps: 99330, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001406, Sample Num: 22496, Cur Loss: 0.40871051, Cur Avg Loss: 1.12000811, Log Avg loss: 1.19261925, Global Avg Loss: 1.45873249, Time: 0.0403 Steps: 99340, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001416, Sample Num: 22656, Cur Loss: 0.77662575, Cur Avg Loss: 1.11929632, Log Avg loss: 1.01921924, Global Avg Loss: 1.45868826, Time: 0.0403 Steps: 99350, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001426, Sample Num: 22816, Cur Loss: 1.35490131, Cur Avg Loss: 1.11959536, Log Avg loss: 1.16193834, Global Avg Loss: 1.45865839, Time: 0.0404 Steps: 99360, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001436, Sample Num: 22976, Cur Loss: 2.05194974, Cur Avg Loss: 1.11897655, Log Avg loss: 1.03073473, Global Avg Loss: 1.45861533, Time: 0.0403 Steps: 99370, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001446, Sample Num: 23136, Cur Loss: 1.03774762, Cur Avg Loss: 1.11907792, Log Avg loss: 1.13363478, Global Avg Loss: 1.45858263, Time: 0.0404 Steps: 99380, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001456, Sample Num: 23296, Cur Loss: 0.44598180, Cur Avg Loss: 1.11788461, Log Avg loss: 0.94533199, Global Avg Loss: 1.45853099, Time: 0.0403 Steps: 99390, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001466, Sample Num: 23456, Cur Loss: 1.33185983, Cur Avg Loss: 1.11752919, Log Avg loss: 1.06577933, Global Avg Loss: 1.45849147, Time: 0.0403 Steps: 99400, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001476, Sample Num: 23616, Cur Loss: 0.30464315, Cur Avg Loss: 1.11580464, Log Avg loss: 0.86298670, Global Avg Loss: 1.45843157, Time: 0.0403 Steps: 99410, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001486, Sample Num: 23776, Cur Loss: 0.94270325, Cur Avg Loss: 1.11377228, Log Avg loss: 0.81379476, Global Avg Loss: 1.45836673, Time: 0.0403 Steps: 99420, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001496, Sample Num: 23936, Cur Loss: 1.96378195, Cur Avg Loss: 1.11248574, Log Avg loss: 0.92130700, Global Avg Loss: 1.45831272, Time: 0.0403 Steps: 99430, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001506, Sample Num: 24096, Cur Loss: 0.94562542, Cur Avg Loss: 1.11149794, Log Avg loss: 0.96372257, Global Avg Loss: 1.45826298, Time: 0.0403 Steps: 99440, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001516, Sample Num: 24256, Cur Loss: 0.95641780, Cur Avg Loss: 1.11039674, Log Avg loss: 0.94455597, Global Avg Loss: 1.45821132, Time: 0.0402 Steps: 99450, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001526, Sample Num: 24416, Cur Loss: 2.18055630, Cur Avg Loss: 1.11149888, Log Avg loss: 1.27858394, Global Avg Loss: 1.45819326, Time: 0.0402 Steps: 99460, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001536, Sample Num: 24576, Cur Loss: 1.21514559, Cur Avg Loss: 1.11138340, Log Avg loss: 1.09376061, Global Avg Loss: 1.45815663, Time: 0.0404 Steps: 99470, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001546, Sample Num: 24736, Cur Loss: 1.66690564, Cur Avg Loss: 1.11399519, Log Avg loss: 1.51516597, Global Avg Loss: 1.45816236, Time: 0.0403 Steps: 99480, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001556, Sample Num: 24896, Cur Loss: 0.47008628, Cur Avg Loss: 1.11338737, Log Avg loss: 1.01941788, Global Avg Loss: 1.45811826, Time: 0.0403 Steps: 99490, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001566, Sample Num: 25056, Cur Loss: 1.25434136, Cur Avg Loss: 1.11259989, Log Avg loss: 0.99006932, Global Avg Loss: 1.45807122, Time: 0.0402 Steps: 99500, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001576, Sample Num: 25216, Cur Loss: 1.31918490, Cur Avg Loss: 1.11194996, Log Avg loss: 1.01017048, Global Avg Loss: 1.45802621, Time: 0.0402 Steps: 99510, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001586, Sample Num: 25376, Cur Loss: 1.16736388, Cur Avg Loss: 1.11088162, Log Avg loss: 0.94251051, Global Avg Loss: 1.45797441, Time: 0.0402 Steps: 99520, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001596, Sample Num: 25536, Cur Loss: 0.33586898, Cur Avg Loss: 1.11123717, Log Avg loss: 1.16762864, Global Avg Loss: 1.45794523, Time: 0.0402 Steps: 99530, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001606, Sample Num: 25696, Cur Loss: 0.41641062, Cur Avg Loss: 1.11162897, Log Avg loss: 1.17415917, Global Avg Loss: 1.45791672, Time: 0.0402 Steps: 99540, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001616, Sample Num: 25856, Cur Loss: 1.67285991, Cur Avg Loss: 1.11278550, Log Avg loss: 1.29852384, Global Avg Loss: 1.45790071, Time: 0.0402 Steps: 99550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001626, Sample Num: 26016, Cur Loss: 0.56831741, Cur Avg Loss: 1.11211763, Log Avg loss: 1.00419008, Global Avg Loss: 1.45785514, Time: 0.0403 Steps: 99560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001636, Sample Num: 26176, Cur Loss: 1.07811534, Cur Avg Loss: 1.11254571, Log Avg loss: 1.18215177, Global Avg Loss: 1.45782745, Time: 0.0402 Steps: 99570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001646, Sample Num: 26336, Cur Loss: 0.41136175, Cur Avg Loss: 1.11169465, Log Avg loss: 0.97246140, Global Avg Loss: 1.45777871, Time: 0.0402 Steps: 99580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001656, Sample Num: 26496, Cur Loss: 0.51352930, Cur Avg Loss: 1.11168681, Log Avg loss: 1.11039592, Global Avg Loss: 1.45774383, Time: 0.0700 Steps: 99590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001666, Sample Num: 26656, Cur Loss: 0.82292616, Cur Avg Loss: 1.11250203, Log Avg loss: 1.24750219, Global Avg Loss: 1.45772272, Time: 0.0403 Steps: 99600, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001676, Sample Num: 26816, Cur Loss: 0.82717979, Cur Avg Loss: 1.11233785, Log Avg loss: 1.08498670, Global Avg Loss: 1.45768530, Time: 0.0489 Steps: 99610, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001686, Sample Num: 26976, Cur Loss: 1.82607770, Cur Avg Loss: 1.11366853, Log Avg loss: 1.33668963, Global Avg Loss: 1.45767316, Time: 0.0551 Steps: 99620, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001696, Sample Num: 27136, Cur Loss: 0.72595525, Cur Avg Loss: 1.11233596, Log Avg loss: 0.88766590, Global Avg Loss: 1.45761594, Time: 0.0531 Steps: 99630, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001706, Sample Num: 27296, Cur Loss: 0.31898403, Cur Avg Loss: 1.11273515, Log Avg loss: 1.18043715, Global Avg Loss: 1.45758813, Time: 0.0681 Steps: 99640, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001716, Sample Num: 27456, Cur Loss: 0.80475318, Cur Avg Loss: 1.11165658, Log Avg loss: 0.92765192, Global Avg Loss: 1.45753495, Time: 0.0963 Steps: 99650, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001726, Sample Num: 27616, Cur Loss: 1.85484993, Cur Avg Loss: 1.11273667, Log Avg loss: 1.29808020, Global Avg Loss: 1.45751895, Time: 0.0559 Steps: 99660, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001736, Sample Num: 27776, Cur Loss: 1.77683115, Cur Avg Loss: 1.11257057, Log Avg loss: 1.08390175, Global Avg Loss: 1.45748146, Time: 0.0877 Steps: 99670, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001746, Sample Num: 27936, Cur Loss: 0.99554521, Cur Avg Loss: 1.11198413, Log Avg loss: 1.01017775, Global Avg Loss: 1.45743659, Time: 0.0517 Steps: 99680, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001756, Sample Num: 28096, Cur Loss: 2.31860900, Cur Avg Loss: 1.11245399, Log Avg loss: 1.19449281, Global Avg Loss: 1.45741021, Time: 0.0403 Steps: 99690, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001766, Sample Num: 28256, Cur Loss: 1.29332030, Cur Avg Loss: 1.11163424, Log Avg loss: 0.96768556, Global Avg Loss: 1.45736109, Time: 0.0403 Steps: 99700, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001776, Sample Num: 28416, Cur Loss: 1.63460493, Cur Avg Loss: 1.11266772, Log Avg loss: 1.29517958, Global Avg Loss: 1.45734482, Time: 0.0402 Steps: 99710, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001786, Sample Num: 28576, Cur Loss: 1.48102069, Cur Avg Loss: 1.11186759, Log Avg loss: 0.96976582, Global Avg Loss: 1.45729593, Time: 0.0402 Steps: 99720, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001796, Sample Num: 28736, Cur Loss: 1.54949546, Cur Avg Loss: 1.11143724, Log Avg loss: 1.03457529, Global Avg Loss: 1.45725354, Time: 0.0403 Steps: 99730, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001806, Sample Num: 28896, Cur Loss: 0.92897618, Cur Avg Loss: 1.11248567, Log Avg loss: 1.30078463, Global Avg Loss: 1.45723786, Time: 0.0403 Steps: 99740, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001816, Sample Num: 29056, Cur Loss: 1.47471988, Cur Avg Loss: 1.11170689, Log Avg loss: 0.97105898, Global Avg Loss: 1.45718912, Time: 0.0403 Steps: 99750, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001826, Sample Num: 29216, Cur Loss: 0.63211668, Cur Avg Loss: 1.11000000, Log Avg loss: 0.80002932, Global Avg Loss: 1.45712324, Time: 0.0403 Steps: 99760, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001836, Sample Num: 29376, Cur Loss: 0.33871877, Cur Avg Loss: 1.11009794, Log Avg loss: 1.12798105, Global Avg Loss: 1.45709025, Time: 0.0403 Steps: 99770, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001846, Sample Num: 29536, Cur Loss: 0.78944033, Cur Avg Loss: 1.10985995, Log Avg loss: 1.06616536, Global Avg Loss: 1.45705107, Time: 0.0402 Steps: 99780, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001856, Sample Num: 29696, Cur Loss: 2.19500184, Cur Avg Loss: 1.11019268, Log Avg loss: 1.17161496, Global Avg Loss: 1.45702247, Time: 0.0403 Steps: 99790, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001866, Sample Num: 29856, Cur Loss: 0.77403212, Cur Avg Loss: 1.10979641, Log Avg loss: 1.03624914, Global Avg Loss: 1.45698031, Time: 0.0403 Steps: 99800, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001876, Sample Num: 30016, Cur Loss: 0.67871153, Cur Avg Loss: 1.10924876, Log Avg loss: 1.00705715, Global Avg Loss: 1.45693523, Time: 0.0402 Steps: 99810, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001886, Sample Num: 30176, Cur Loss: 0.26541695, Cur Avg Loss: 1.10938560, Log Avg loss: 1.13505675, Global Avg Loss: 1.45690298, Time: 0.0402 Steps: 99820, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001896, Sample Num: 30336, Cur Loss: 0.70950997, Cur Avg Loss: 1.10871084, Log Avg loss: 0.98145039, Global Avg Loss: 1.45685536, Time: 0.0403 Steps: 99830, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001906, Sample Num: 30496, Cur Loss: 0.52891487, Cur Avg Loss: 1.10915491, Log Avg loss: 1.19334998, Global Avg Loss: 1.45682897, Time: 0.0403 Steps: 99840, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001916, Sample Num: 30656, Cur Loss: 1.80137873, Cur Avg Loss: 1.10806038, Log Avg loss: 0.89944417, Global Avg Loss: 1.45677314, Time: 0.0402 Steps: 99850, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001926, Sample Num: 30816, Cur Loss: 1.64465559, Cur Avg Loss: 1.10885150, Log Avg loss: 1.26043040, Global Avg Loss: 1.45675348, Time: 0.0403 Steps: 99860, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001936, Sample Num: 30976, Cur Loss: 1.14187324, Cur Avg Loss: 1.10837742, Log Avg loss: 1.01706860, Global Avg Loss: 1.45670946, Time: 0.0404 Steps: 99870, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001946, Sample Num: 31136, Cur Loss: 0.67133129, Cur Avg Loss: 1.10839187, Log Avg loss: 1.11118881, Global Avg Loss: 1.45667486, Time: 0.0404 Steps: 99880, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001956, Sample Num: 31296, Cur Loss: 0.44974741, Cur Avg Loss: 1.10935418, Log Avg loss: 1.29662103, Global Avg Loss: 1.45665884, Time: 0.0402 Steps: 99890, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001966, Sample Num: 31456, Cur Loss: 0.53950268, Cur Avg Loss: 1.11062951, Log Avg loss: 1.36008283, Global Avg Loss: 1.45664917, Time: 0.0403 Steps: 99900, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001976, Sample Num: 31616, Cur Loss: 0.76449049, Cur Avg Loss: 1.11197209, Log Avg loss: 1.37592300, Global Avg Loss: 1.45664109, Time: 0.0402 Steps: 99910, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001986, Sample Num: 31776, Cur Loss: 1.51924753, Cur Avg Loss: 1.11188929, Log Avg loss: 1.09552857, Global Avg Loss: 1.45660495, Time: 0.0402 Steps: 99920, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001996, Sample Num: 31936, Cur Loss: 1.72262430, Cur Avg Loss: 1.11109148, Log Avg loss: 0.95264602, Global Avg Loss: 1.45655452, Time: 0.0402 Steps: 99930, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002006, Sample Num: 32096, Cur Loss: 0.49473426, Cur Avg Loss: 1.11122940, Log Avg loss: 1.13875881, Global Avg Loss: 1.45652272, Time: 0.0648 Steps: 99940, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002016, Sample Num: 32256, Cur Loss: 1.20739877, Cur Avg Loss: 1.11011065, Log Avg loss: 0.88568942, Global Avg Loss: 1.45646561, Time: 0.0529 Steps: 99950, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002026, Sample Num: 32416, Cur Loss: 0.96867144, Cur Avg Loss: 1.11083309, Log Avg loss: 1.25647767, Global Avg Loss: 1.45644560, Time: 0.0606 Steps: 99960, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002036, Sample Num: 32576, Cur Loss: 0.91817462, Cur Avg Loss: 1.11050551, Log Avg loss: 1.04413794, Global Avg Loss: 1.45640436, Time: 0.0651 Steps: 99970, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002046, Sample Num: 32736, Cur Loss: 2.10343957, Cur Avg Loss: 1.11125103, Log Avg loss: 1.26303873, Global Avg Loss: 1.45638502, Time: 0.0514 Steps: 99980, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002056, Sample Num: 32896, Cur Loss: 1.08399379, Cur Avg Loss: 1.11016366, Log Avg loss: 0.88768647, Global Avg Loss: 1.45632814, Time: 0.0406 Steps: 99990, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002066, Sample Num: 33056, Cur Loss: 0.49998927, Cur Avg Loss: 1.10885015, Log Avg loss: 0.83879391, Global Avg Loss: 1.45626639, Time: 0.0992 Steps: 100000, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002076, Sample Num: 33216, Cur Loss: 2.20498443, Cur Avg Loss: 1.11019972, Log Avg loss: 1.38902045, Global Avg Loss: 1.45625967, Time: 0.0432 Steps: 100010, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002086, Sample Num: 33376, Cur Loss: 1.13250458, Cur Avg Loss: 1.11056269, Log Avg loss: 1.18591524, Global Avg Loss: 1.45623264, Time: 0.0619 Steps: 100020, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002096, Sample Num: 33536, Cur Loss: 0.84932441, Cur Avg Loss: 1.11095612, Log Avg loss: 1.19302622, Global Avg Loss: 1.45620632, Time: 0.0404 Steps: 100030, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002106, Sample Num: 33696, Cur Loss: 1.33340621, Cur Avg Loss: 1.11056559, Log Avg loss: 1.02870904, Global Avg Loss: 1.45616359, Time: 0.0403 Steps: 100040, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002116, Sample Num: 33856, Cur Loss: 0.99726397, Cur Avg Loss: 1.11030564, Log Avg loss: 1.05556099, Global Avg Loss: 1.45612355, Time: 0.0403 Steps: 100050, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002126, Sample Num: 34016, Cur Loss: 0.84938341, Cur Avg Loss: 1.11051904, Log Avg loss: 1.15567484, Global Avg Loss: 1.45609353, Time: 0.0404 Steps: 100060, Updated lr: 0.000006 ***** Running evaluation checkpoint-100063 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-100063 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.218417, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.095281, "eval_total_loss": 769.982393, "eval_mae": 0.866647, "eval_mse": 1.095354, "eval_r2": 0.303721, "eval_sp_statistic": 0.484085, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.584134, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.105154, "test_total_loss": 554.787372, "test_mae": 0.887676, "test_mse": 1.105336, "test_r2": 0.286607, "test_sp_statistic": 0.503155, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.626193, "test_ps_pvalue": 0.0, "lr": 6.056899004267426e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4560754189901344, "train_cur_epoch_loss": 2363.5200139284134, "train_cur_epoch_avg_loss": 1.110155008890753, "train_cur_epoch_time": 95.218416929245, "train_cur_epoch_avg_time": 0.044724479534638324, "epoch": 47, "step": 100063} ################################################## Training, Epoch: 0048, Batch: 000007, Sample Num: 112, Cur Loss: 0.82958609, Cur Avg Loss: 0.93903827, Log Avg loss: 0.91297984, Global Avg Loss: 1.45603925, Time: 0.0804 Steps: 100070, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000017, Sample Num: 272, Cur Loss: 0.59771472, Cur Avg Loss: 0.98855006, Log Avg loss: 1.02320831, Global Avg Loss: 1.45599600, Time: 0.0402 Steps: 100080, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000027, Sample Num: 432, Cur Loss: 0.58877945, Cur Avg Loss: 1.05578185, Log Avg loss: 1.17007590, Global Avg Loss: 1.45596744, Time: 0.0403 Steps: 100090, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000037, Sample Num: 592, Cur Loss: 2.64682508, Cur Avg Loss: 1.13116104, Log Avg loss: 1.33468485, Global Avg Loss: 1.45595532, Time: 0.0404 Steps: 100100, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000047, Sample Num: 752, Cur Loss: 1.23625171, Cur Avg Loss: 1.13958514, Log Avg loss: 1.17075431, Global Avg Loss: 1.45592683, Time: 0.0402 Steps: 100110, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000057, Sample Num: 912, Cur Loss: 1.04376149, Cur Avg Loss: 1.08598627, Log Avg loss: 0.83407158, Global Avg Loss: 1.45586472, Time: 0.0403 Steps: 100120, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000067, Sample Num: 1072, Cur Loss: 0.40983209, Cur Avg Loss: 1.08204578, Log Avg loss: 1.05958497, Global Avg Loss: 1.45582514, Time: 0.0403 Steps: 100130, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000077, Sample Num: 1232, Cur Loss: 1.32056451, Cur Avg Loss: 1.10245043, Log Avg loss: 1.23916159, Global Avg Loss: 1.45580351, Time: 0.0403 Steps: 100140, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000087, Sample Num: 1392, Cur Loss: 0.47270525, Cur Avg Loss: 1.10931422, Log Avg loss: 1.16216545, Global Avg Loss: 1.45577419, Time: 0.0403 Steps: 100150, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000097, Sample Num: 1552, Cur Loss: 3.02873516, Cur Avg Loss: 1.14708199, Log Avg loss: 1.47566157, Global Avg Loss: 1.45577617, Time: 0.0403 Steps: 100160, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000107, Sample Num: 1712, Cur Loss: 0.93544650, Cur Avg Loss: 1.15379946, Log Avg loss: 1.21895890, Global Avg Loss: 1.45575253, Time: 0.0402 Steps: 100170, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000117, Sample Num: 1872, Cur Loss: 0.90991217, Cur Avg Loss: 1.12850188, Log Avg loss: 0.85781776, Global Avg Loss: 1.45569285, Time: 0.0402 Steps: 100180, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000127, Sample Num: 2032, Cur Loss: 0.68745005, Cur Avg Loss: 1.12708436, Log Avg loss: 1.11049937, Global Avg Loss: 1.45565839, Time: 0.0402 Steps: 100190, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000137, Sample Num: 2192, Cur Loss: 1.16327047, Cur Avg Loss: 1.12791449, Log Avg loss: 1.13845714, Global Avg Loss: 1.45562674, Time: 0.0403 Steps: 100200, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000147, Sample Num: 2352, Cur Loss: 0.91310436, Cur Avg Loss: 1.11952497, Log Avg loss: 1.00458860, Global Avg Loss: 1.45558173, Time: 0.0403 Steps: 100210, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000157, Sample Num: 2512, Cur Loss: 0.46860299, Cur Avg Loss: 1.12142986, Log Avg loss: 1.14943165, Global Avg Loss: 1.45555118, Time: 0.0403 Steps: 100220, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000167, Sample Num: 2672, Cur Loss: 1.07328987, Cur Avg Loss: 1.11615463, Log Avg loss: 1.03333359, Global Avg Loss: 1.45550905, Time: 0.0403 Steps: 100230, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000177, Sample Num: 2832, Cur Loss: 0.84599388, Cur Avg Loss: 1.11277926, Log Avg loss: 1.05641050, Global Avg Loss: 1.45546924, Time: 0.0403 Steps: 100240, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000187, Sample Num: 2992, Cur Loss: 0.28315997, Cur Avg Loss: 1.09459719, Log Avg loss: 0.77277463, Global Avg Loss: 1.45540114, Time: 0.0402 Steps: 100250, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000197, Sample Num: 3152, Cur Loss: 0.96381921, Cur Avg Loss: 1.08836365, Log Avg loss: 0.97179646, Global Avg Loss: 1.45535291, Time: 0.0403 Steps: 100260, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000207, Sample Num: 3312, Cur Loss: 2.27240229, Cur Avg Loss: 1.08962592, Log Avg loss: 1.11449251, Global Avg Loss: 1.45531891, Time: 0.0402 Steps: 100270, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000217, Sample Num: 3472, Cur Loss: 1.24347365, Cur Avg Loss: 1.10782033, Log Avg loss: 1.48444466, Global Avg Loss: 1.45532182, Time: 0.0402 Steps: 100280, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000227, Sample Num: 3632, Cur Loss: 1.31572044, Cur Avg Loss: 1.11410070, Log Avg loss: 1.25038490, Global Avg Loss: 1.45530138, Time: 0.0403 Steps: 100290, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000237, Sample Num: 3792, Cur Loss: 1.04354286, Cur Avg Loss: 1.11013625, Log Avg loss: 1.02014309, Global Avg Loss: 1.45525800, Time: 0.0403 Steps: 100300, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000247, Sample Num: 3952, Cur Loss: 1.94579184, Cur Avg Loss: 1.13020366, Log Avg loss: 1.60580121, Global Avg Loss: 1.45527300, Time: 0.0402 Steps: 100310, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000257, Sample Num: 4112, Cur Loss: 1.27369034, Cur Avg Loss: 1.13261066, Log Avg loss: 1.19206368, Global Avg Loss: 1.45524677, Time: 0.0402 Steps: 100320, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000267, Sample Num: 4272, Cur Loss: 0.77959782, Cur Avg Loss: 1.13214888, Log Avg loss: 1.12028113, Global Avg Loss: 1.45521338, Time: 0.0403 Steps: 100330, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000277, Sample Num: 4432, Cur Loss: 0.39623123, Cur Avg Loss: 1.12802269, Log Avg loss: 1.01785352, Global Avg Loss: 1.45516979, Time: 0.0407 Steps: 100340, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000287, Sample Num: 4592, Cur Loss: 0.77669919, Cur Avg Loss: 1.12794742, Log Avg loss: 1.12586229, Global Avg Loss: 1.45513698, Time: 0.0407 Steps: 100350, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000297, Sample Num: 4752, Cur Loss: 1.40530169, Cur Avg Loss: 1.11957190, Log Avg loss: 0.87919455, Global Avg Loss: 1.45507959, Time: 0.0680 Steps: 100360, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000307, Sample Num: 4912, Cur Loss: 0.92527395, Cur Avg Loss: 1.11902407, Log Avg loss: 1.10275337, Global Avg Loss: 1.45504449, Time: 0.0404 Steps: 100370, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000317, Sample Num: 5072, Cur Loss: 0.52067572, Cur Avg Loss: 1.12481203, Log Avg loss: 1.30250247, Global Avg Loss: 1.45502929, Time: 0.0710 Steps: 100380, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000327, Sample Num: 5232, Cur Loss: 1.06374288, Cur Avg Loss: 1.11568211, Log Avg loss: 0.82626385, Global Avg Loss: 1.45496666, Time: 0.0463 Steps: 100390, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000337, Sample Num: 5392, Cur Loss: 1.27750909, Cur Avg Loss: 1.10819316, Log Avg loss: 0.86330430, Global Avg Loss: 1.45490773, Time: 0.0408 Steps: 100400, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000347, Sample Num: 5552, Cur Loss: 0.49617010, Cur Avg Loss: 1.11667789, Log Avg loss: 1.40261340, Global Avg Loss: 1.45490252, Time: 0.0407 Steps: 100410, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000357, Sample Num: 5712, Cur Loss: 1.27944279, Cur Avg Loss: 1.12000122, Log Avg loss: 1.23532086, Global Avg Loss: 1.45488065, Time: 0.0404 Steps: 100420, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000367, Sample Num: 5872, Cur Loss: 1.34882116, Cur Avg Loss: 1.11730394, Log Avg loss: 1.02101077, Global Avg Loss: 1.45483745, Time: 0.0402 Steps: 100430, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000377, Sample Num: 6032, Cur Loss: 1.31447303, Cur Avg Loss: 1.11606636, Log Avg loss: 1.07064726, Global Avg Loss: 1.45479920, Time: 0.0402 Steps: 100440, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000387, Sample Num: 6192, Cur Loss: 1.03933597, Cur Avg Loss: 1.11204589, Log Avg loss: 0.96047415, Global Avg Loss: 1.45474999, Time: 0.0402 Steps: 100450, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000397, Sample Num: 6352, Cur Loss: 2.64775705, Cur Avg Loss: 1.11952134, Log Avg loss: 1.40882121, Global Avg Loss: 1.45474542, Time: 0.0402 Steps: 100460, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000407, Sample Num: 6512, Cur Loss: 1.41431928, Cur Avg Loss: 1.11303901, Log Avg loss: 0.85569057, Global Avg Loss: 1.45468579, Time: 0.0403 Steps: 100470, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000417, Sample Num: 6672, Cur Loss: 1.46947920, Cur Avg Loss: 1.11473149, Log Avg loss: 1.18361526, Global Avg Loss: 1.45465881, Time: 0.0404 Steps: 100480, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000427, Sample Num: 6832, Cur Loss: 0.55985457, Cur Avg Loss: 1.11266993, Log Avg loss: 1.02670290, Global Avg Loss: 1.45461623, Time: 0.0402 Steps: 100490, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000437, Sample Num: 6992, Cur Loss: 0.90681839, Cur Avg Loss: 1.10657341, Log Avg loss: 0.84625237, Global Avg Loss: 1.45455569, Time: 0.0402 Steps: 100500, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000447, Sample Num: 7152, Cur Loss: 0.71725166, Cur Avg Loss: 1.11086543, Log Avg loss: 1.29842639, Global Avg Loss: 1.45454016, Time: 0.0402 Steps: 100510, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000457, Sample Num: 7312, Cur Loss: 1.22649825, Cur Avg Loss: 1.11241906, Log Avg loss: 1.18186660, Global Avg Loss: 1.45451303, Time: 0.0402 Steps: 100520, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000467, Sample Num: 7472, Cur Loss: 0.36549324, Cur Avg Loss: 1.11057231, Log Avg loss: 1.02617589, Global Avg Loss: 1.45447043, Time: 0.0403 Steps: 100530, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000477, Sample Num: 7632, Cur Loss: 0.51965505, Cur Avg Loss: 1.11505594, Log Avg loss: 1.32444104, Global Avg Loss: 1.45445749, Time: 0.0403 Steps: 100540, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000487, Sample Num: 7792, Cur Loss: 0.89825964, Cur Avg Loss: 1.11739347, Log Avg loss: 1.22889392, Global Avg Loss: 1.45443506, Time: 0.0402 Steps: 100550, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000497, Sample Num: 7952, Cur Loss: 1.24725163, Cur Avg Loss: 1.11866466, Log Avg loss: 1.18057166, Global Avg Loss: 1.45440783, Time: 0.0402 Steps: 100560, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000507, Sample Num: 8112, Cur Loss: 1.19597900, Cur Avg Loss: 1.12629222, Log Avg loss: 1.50538182, Global Avg Loss: 1.45441289, Time: 0.0403 Steps: 100570, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000517, Sample Num: 8272, Cur Loss: 1.15826356, Cur Avg Loss: 1.12515996, Log Avg loss: 1.06775444, Global Avg Loss: 1.45437445, Time: 0.0402 Steps: 100580, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000527, Sample Num: 8432, Cur Loss: 0.46410364, Cur Avg Loss: 1.12842698, Log Avg loss: 1.29733174, Global Avg Loss: 1.45435884, Time: 0.0402 Steps: 100590, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000537, Sample Num: 8592, Cur Loss: 0.74904901, Cur Avg Loss: 1.12370131, Log Avg loss: 0.87465869, Global Avg Loss: 1.45430122, Time: 0.0403 Steps: 100600, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000547, Sample Num: 8752, Cur Loss: 1.28814507, Cur Avg Loss: 1.13095680, Log Avg loss: 1.52057647, Global Avg Loss: 1.45430780, Time: 0.0402 Steps: 100610, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000557, Sample Num: 8912, Cur Loss: 0.51231116, Cur Avg Loss: 1.12640544, Log Avg loss: 0.87744620, Global Avg Loss: 1.45425047, Time: 0.0402 Steps: 100620, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000567, Sample Num: 9072, Cur Loss: 1.00424600, Cur Avg Loss: 1.12195864, Log Avg loss: 0.87427205, Global Avg Loss: 1.45419284, Time: 0.0402 Steps: 100630, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000577, Sample Num: 9232, Cur Loss: 1.21503413, Cur Avg Loss: 1.11764143, Log Avg loss: 0.87285553, Global Avg Loss: 1.45413507, Time: 0.0402 Steps: 100640, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000587, Sample Num: 9392, Cur Loss: 0.74596965, Cur Avg Loss: 1.11253284, Log Avg loss: 0.81776703, Global Avg Loss: 1.45407185, Time: 0.0402 Steps: 100650, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000597, Sample Num: 9552, Cur Loss: 1.47381306, Cur Avg Loss: 1.11082904, Log Avg loss: 1.01081627, Global Avg Loss: 1.45402781, Time: 0.0402 Steps: 100660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000607, Sample Num: 9712, Cur Loss: 0.38236403, Cur Avg Loss: 1.10753198, Log Avg loss: 0.91069754, Global Avg Loss: 1.45397384, Time: 0.0402 Steps: 100670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000617, Sample Num: 9872, Cur Loss: 2.02222705, Cur Avg Loss: 1.11331337, Log Avg loss: 1.46424329, Global Avg Loss: 1.45397486, Time: 0.0403 Steps: 100680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000627, Sample Num: 10032, Cur Loss: 0.53187728, Cur Avg Loss: 1.11198184, Log Avg loss: 1.02982674, Global Avg Loss: 1.45393274, Time: 0.0402 Steps: 100690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000637, Sample Num: 10192, Cur Loss: 1.48009443, Cur Avg Loss: 1.11317573, Log Avg loss: 1.18803280, Global Avg Loss: 1.45390633, Time: 0.0404 Steps: 100700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000647, Sample Num: 10352, Cur Loss: 1.00604224, Cur Avg Loss: 1.11424824, Log Avg loss: 1.18256668, Global Avg Loss: 1.45387939, Time: 0.0530 Steps: 100710, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000657, Sample Num: 10512, Cur Loss: 1.41230285, Cur Avg Loss: 1.11061636, Log Avg loss: 0.87563408, Global Avg Loss: 1.45382198, Time: 0.0696 Steps: 100720, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000667, Sample Num: 10672, Cur Loss: 0.76650429, Cur Avg Loss: 1.10761900, Log Avg loss: 0.91069204, Global Avg Loss: 1.45376806, Time: 0.0802 Steps: 100730, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000677, Sample Num: 10832, Cur Loss: 2.30044699, Cur Avg Loss: 1.10487752, Log Avg loss: 0.92202122, Global Avg Loss: 1.45371527, Time: 0.0790 Steps: 100740, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000687, Sample Num: 10992, Cur Loss: 0.44591618, Cur Avg Loss: 1.10352474, Log Avg loss: 1.01194104, Global Avg Loss: 1.45367143, Time: 0.0404 Steps: 100750, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000697, Sample Num: 11152, Cur Loss: 1.25756836, Cur Avg Loss: 1.10591949, Log Avg loss: 1.27043877, Global Avg Loss: 1.45365324, Time: 0.0432 Steps: 100760, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000707, Sample Num: 11312, Cur Loss: 0.61882710, Cur Avg Loss: 1.10746091, Log Avg loss: 1.21489844, Global Avg Loss: 1.45362955, Time: 0.0403 Steps: 100770, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000717, Sample Num: 11472, Cur Loss: 1.21096897, Cur Avg Loss: 1.10584228, Log Avg loss: 0.99140458, Global Avg Loss: 1.45358368, Time: 0.0403 Steps: 100780, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000727, Sample Num: 11632, Cur Loss: 0.71711677, Cur Avg Loss: 1.10752155, Log Avg loss: 1.22792582, Global Avg Loss: 1.45356129, Time: 0.0403 Steps: 100790, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000737, Sample Num: 11792, Cur Loss: 1.91014719, Cur Avg Loss: 1.10849057, Log Avg loss: 1.17893796, Global Avg Loss: 1.45353405, Time: 0.0403 Steps: 100800, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000747, Sample Num: 11952, Cur Loss: 0.72666645, Cur Avg Loss: 1.11064105, Log Avg loss: 1.26913132, Global Avg Loss: 1.45351576, Time: 0.0403 Steps: 100810, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000757, Sample Num: 12112, Cur Loss: 1.25647926, Cur Avg Loss: 1.10921208, Log Avg loss: 1.00246782, Global Avg Loss: 1.45347102, Time: 0.0404 Steps: 100820, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000767, Sample Num: 12272, Cur Loss: 0.70546031, Cur Avg Loss: 1.10711094, Log Avg loss: 0.94805511, Global Avg Loss: 1.45342089, Time: 0.0403 Steps: 100830, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000777, Sample Num: 12432, Cur Loss: 1.13378811, Cur Avg Loss: 1.10728150, Log Avg loss: 1.12036311, Global Avg Loss: 1.45338787, Time: 0.0402 Steps: 100840, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000787, Sample Num: 12592, Cur Loss: 1.21154523, Cur Avg Loss: 1.10747691, Log Avg loss: 1.12266073, Global Avg Loss: 1.45335507, Time: 0.0403 Steps: 100850, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000797, Sample Num: 12752, Cur Loss: 2.11989474, Cur Avg Loss: 1.11015702, Log Avg loss: 1.32108177, Global Avg Loss: 1.45334196, Time: 0.0403 Steps: 100860, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000807, Sample Num: 12912, Cur Loss: 0.92809880, Cur Avg Loss: 1.10862561, Log Avg loss: 0.98657198, Global Avg Loss: 1.45329568, Time: 0.0402 Steps: 100870, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000817, Sample Num: 13072, Cur Loss: 1.10412109, Cur Avg Loss: 1.10735165, Log Avg loss: 1.00454296, Global Avg Loss: 1.45325120, Time: 0.0402 Steps: 100880, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000827, Sample Num: 13232, Cur Loss: 0.39930996, Cur Avg Loss: 1.10342810, Log Avg loss: 0.78287415, Global Avg Loss: 1.45318475, Time: 0.0403 Steps: 100890, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000837, Sample Num: 13392, Cur Loss: 0.79683405, Cur Avg Loss: 1.10170938, Log Avg loss: 0.95957098, Global Avg Loss: 1.45313583, Time: 0.0402 Steps: 100900, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000847, Sample Num: 13552, Cur Loss: 2.98054218, Cur Avg Loss: 1.10331170, Log Avg loss: 1.23742577, Global Avg Loss: 1.45311446, Time: 0.0403 Steps: 100910, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000857, Sample Num: 13712, Cur Loss: 2.85083461, Cur Avg Loss: 1.10143037, Log Avg loss: 0.94208235, Global Avg Loss: 1.45306382, Time: 0.0403 Steps: 100920, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000867, Sample Num: 13872, Cur Loss: 2.36254787, Cur Avg Loss: 1.10335764, Log Avg loss: 1.26852405, Global Avg Loss: 1.45304553, Time: 0.0402 Steps: 100930, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000877, Sample Num: 14032, Cur Loss: 1.07973242, Cur Avg Loss: 1.10293841, Log Avg loss: 1.06659169, Global Avg Loss: 1.45300725, Time: 0.0402 Steps: 100940, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000887, Sample Num: 14192, Cur Loss: 1.03184414, Cur Avg Loss: 1.10606467, Log Avg loss: 1.38023780, Global Avg Loss: 1.45300004, Time: 0.0402 Steps: 100950, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000897, Sample Num: 14352, Cur Loss: 1.48672974, Cur Avg Loss: 1.11181303, Log Avg loss: 1.62169223, Global Avg Loss: 1.45301675, Time: 0.0402 Steps: 100960, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000907, Sample Num: 14512, Cur Loss: 1.03964615, Cur Avg Loss: 1.11074729, Log Avg loss: 1.01515020, Global Avg Loss: 1.45297338, Time: 0.0402 Steps: 100970, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000917, Sample Num: 14672, Cur Loss: 0.38733119, Cur Avg Loss: 1.11159239, Log Avg loss: 1.18824356, Global Avg Loss: 1.45294717, Time: 0.0402 Steps: 100980, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000927, Sample Num: 14832, Cur Loss: 0.48713157, Cur Avg Loss: 1.10955518, Log Avg loss: 0.92274275, Global Avg Loss: 1.45289467, Time: 0.0403 Steps: 100990, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000937, Sample Num: 14992, Cur Loss: 0.77807719, Cur Avg Loss: 1.10845461, Log Avg loss: 1.00643178, Global Avg Loss: 1.45285046, Time: 0.0403 Steps: 101000, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000947, Sample Num: 15152, Cur Loss: 0.30272689, Cur Avg Loss: 1.10678053, Log Avg loss: 0.94991905, Global Avg Loss: 1.45280067, Time: 0.0402 Steps: 101010, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000957, Sample Num: 15312, Cur Loss: 1.06295991, Cur Avg Loss: 1.10646818, Log Avg loss: 1.07688853, Global Avg Loss: 1.45276346, Time: 0.0402 Steps: 101020, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000967, Sample Num: 15472, Cur Loss: 0.30161846, Cur Avg Loss: 1.10597526, Log Avg loss: 1.05880264, Global Avg Loss: 1.45272447, Time: 0.0402 Steps: 101030, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000977, Sample Num: 15632, Cur Loss: 2.02127886, Cur Avg Loss: 1.11045097, Log Avg loss: 1.54325209, Global Avg Loss: 1.45273342, Time: 0.0679 Steps: 101040, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000987, Sample Num: 15792, Cur Loss: 1.76164198, Cur Avg Loss: 1.11044858, Log Avg loss: 1.11021545, Global Avg Loss: 1.45269953, Time: 0.0509 Steps: 101050, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000997, Sample Num: 15952, Cur Loss: 0.58441150, Cur Avg Loss: 1.10724908, Log Avg loss: 0.79145854, Global Avg Loss: 1.45263410, Time: 0.0678 Steps: 101060, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001007, Sample Num: 16112, Cur Loss: 0.43060684, Cur Avg Loss: 1.10548571, Log Avg loss: 0.92967732, Global Avg Loss: 1.45258236, Time: 0.0406 Steps: 101070, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001017, Sample Num: 16272, Cur Loss: 1.04516935, Cur Avg Loss: 1.10550530, Log Avg loss: 1.10747856, Global Avg Loss: 1.45254821, Time: 0.0920 Steps: 101080, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001027, Sample Num: 16432, Cur Loss: 0.50318837, Cur Avg Loss: 1.10324935, Log Avg loss: 0.87381863, Global Avg Loss: 1.45249097, Time: 0.0505 Steps: 101090, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001037, Sample Num: 16592, Cur Loss: 1.24379003, Cur Avg Loss: 1.10038284, Log Avg loss: 0.80599253, Global Avg Loss: 1.45242702, Time: 0.0469 Steps: 101100, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001047, Sample Num: 16752, Cur Loss: 0.70605642, Cur Avg Loss: 1.10207445, Log Avg loss: 1.27749407, Global Avg Loss: 1.45240972, Time: 0.0435 Steps: 101110, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001057, Sample Num: 16912, Cur Loss: 1.01337147, Cur Avg Loss: 1.10222370, Log Avg loss: 1.11785012, Global Avg Loss: 1.45237663, Time: 0.0936 Steps: 101120, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001067, Sample Num: 17072, Cur Loss: 0.59736288, Cur Avg Loss: 1.10199963, Log Avg loss: 1.07831591, Global Avg Loss: 1.45233964, Time: 0.0403 Steps: 101130, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001077, Sample Num: 17232, Cur Loss: 1.32464755, Cur Avg Loss: 1.10447439, Log Avg loss: 1.36853166, Global Avg Loss: 1.45233136, Time: 0.0403 Steps: 101140, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001087, Sample Num: 17392, Cur Loss: 0.84280527, Cur Avg Loss: 1.10388096, Log Avg loss: 1.03996792, Global Avg Loss: 1.45229059, Time: 0.0403 Steps: 101150, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001097, Sample Num: 17552, Cur Loss: 1.48221123, Cur Avg Loss: 1.10690334, Log Avg loss: 1.43543627, Global Avg Loss: 1.45228892, Time: 0.0403 Steps: 101160, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001107, Sample Num: 17712, Cur Loss: 1.22493410, Cur Avg Loss: 1.10668598, Log Avg loss: 1.08284153, Global Avg Loss: 1.45225241, Time: 0.0403 Steps: 101170, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001117, Sample Num: 17872, Cur Loss: 0.61910313, Cur Avg Loss: 1.10507938, Log Avg loss: 0.92722928, Global Avg Loss: 1.45220052, Time: 0.0404 Steps: 101180, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001127, Sample Num: 18032, Cur Loss: 1.55640793, Cur Avg Loss: 1.10677313, Log Avg loss: 1.29596478, Global Avg Loss: 1.45218508, Time: 0.0402 Steps: 101190, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001137, Sample Num: 18192, Cur Loss: 0.87046760, Cur Avg Loss: 1.10576891, Log Avg loss: 0.99259270, Global Avg Loss: 1.45213966, Time: 0.0403 Steps: 101200, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001147, Sample Num: 18352, Cur Loss: 1.00890720, Cur Avg Loss: 1.10898304, Log Avg loss: 1.47443022, Global Avg Loss: 1.45214187, Time: 0.0403 Steps: 101210, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001157, Sample Num: 18512, Cur Loss: 0.87651879, Cur Avg Loss: 1.10675527, Log Avg loss: 0.85123030, Global Avg Loss: 1.45208250, Time: 0.0403 Steps: 101220, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001167, Sample Num: 18672, Cur Loss: 0.81957328, Cur Avg Loss: 1.10531806, Log Avg loss: 0.93903221, Global Avg Loss: 1.45203182, Time: 0.0402 Steps: 101230, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001177, Sample Num: 18832, Cur Loss: 0.92573744, Cur Avg Loss: 1.10482959, Log Avg loss: 1.04782565, Global Avg Loss: 1.45199189, Time: 0.0402 Steps: 101240, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001187, Sample Num: 18992, Cur Loss: 2.16518068, Cur Avg Loss: 1.10629892, Log Avg loss: 1.27923931, Global Avg Loss: 1.45197483, Time: 0.0403 Steps: 101250, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001197, Sample Num: 19152, Cur Loss: 0.69605958, Cur Avg Loss: 1.10543792, Log Avg loss: 1.00323700, Global Avg Loss: 1.45193051, Time: 0.0403 Steps: 101260, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001207, Sample Num: 19312, Cur Loss: 1.30411887, Cur Avg Loss: 1.10295430, Log Avg loss: 0.80566419, Global Avg Loss: 1.45186670, Time: 0.0403 Steps: 101270, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001217, Sample Num: 19472, Cur Loss: 1.03954041, Cur Avg Loss: 1.10529572, Log Avg loss: 1.38790567, Global Avg Loss: 1.45186038, Time: 0.0403 Steps: 101280, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001227, Sample Num: 19632, Cur Loss: 1.29536080, Cur Avg Loss: 1.10739551, Log Avg loss: 1.36294026, Global Avg Loss: 1.45185160, Time: 0.0403 Steps: 101290, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001237, Sample Num: 19792, Cur Loss: 1.91700983, Cur Avg Loss: 1.10645483, Log Avg loss: 0.99103259, Global Avg Loss: 1.45180611, Time: 0.0402 Steps: 101300, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001247, Sample Num: 19952, Cur Loss: 1.59466577, Cur Avg Loss: 1.10704977, Log Avg loss: 1.18064396, Global Avg Loss: 1.45177935, Time: 0.0402 Steps: 101310, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001257, Sample Num: 20112, Cur Loss: 1.32797313, Cur Avg Loss: 1.10639535, Log Avg loss: 1.02478950, Global Avg Loss: 1.45173720, Time: 0.0402 Steps: 101320, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001267, Sample Num: 20272, Cur Loss: 0.21544990, Cur Avg Loss: 1.10458514, Log Avg loss: 0.87704223, Global Avg Loss: 1.45168049, Time: 0.0403 Steps: 101330, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001277, Sample Num: 20432, Cur Loss: 1.39867151, Cur Avg Loss: 1.10415286, Log Avg loss: 1.04938292, Global Avg Loss: 1.45164079, Time: 0.0402 Steps: 101340, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001287, Sample Num: 20592, Cur Loss: 2.03395796, Cur Avg Loss: 1.10591948, Log Avg loss: 1.33151691, Global Avg Loss: 1.45162894, Time: 0.0404 Steps: 101350, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001297, Sample Num: 20752, Cur Loss: 1.35687387, Cur Avg Loss: 1.10549322, Log Avg loss: 1.05063295, Global Avg Loss: 1.45158938, Time: 0.0403 Steps: 101360, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001307, Sample Num: 20912, Cur Loss: 1.03305686, Cur Avg Loss: 1.10585712, Log Avg loss: 1.15305517, Global Avg Loss: 1.45155993, Time: 0.0403 Steps: 101370, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001317, Sample Num: 21072, Cur Loss: 1.95952535, Cur Avg Loss: 1.10685459, Log Avg loss: 1.23722320, Global Avg Loss: 1.45153879, Time: 0.0402 Steps: 101380, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001327, Sample Num: 21232, Cur Loss: 0.94353259, Cur Avg Loss: 1.10573540, Log Avg loss: 0.95833907, Global Avg Loss: 1.45149014, Time: 0.1007 Steps: 101390, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001337, Sample Num: 21392, Cur Loss: 0.97773957, Cur Avg Loss: 1.10438431, Log Avg loss: 0.92509387, Global Avg Loss: 1.45143823, Time: 0.0458 Steps: 101400, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001347, Sample Num: 21552, Cur Loss: 1.02960968, Cur Avg Loss: 1.10741167, Log Avg loss: 1.51217015, Global Avg Loss: 1.45144422, Time: 0.0958 Steps: 101410, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001357, Sample Num: 21712, Cur Loss: 0.65322047, Cur Avg Loss: 1.10556126, Log Avg loss: 0.85631049, Global Avg Loss: 1.45138554, Time: 0.0404 Steps: 101420, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001367, Sample Num: 21872, Cur Loss: 1.59414589, Cur Avg Loss: 1.10563884, Log Avg loss: 1.11616758, Global Avg Loss: 1.45135249, Time: 0.1001 Steps: 101430, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001377, Sample Num: 22032, Cur Loss: 1.53794658, Cur Avg Loss: 1.10958058, Log Avg loss: 1.64841613, Global Avg Loss: 1.45137192, Time: 0.0607 Steps: 101440, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001387, Sample Num: 22192, Cur Loss: 0.76173002, Cur Avg Loss: 1.11050646, Log Avg loss: 1.23799966, Global Avg Loss: 1.45135088, Time: 0.0405 Steps: 101450, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001397, Sample Num: 22352, Cur Loss: 0.66515249, Cur Avg Loss: 1.11070271, Log Avg loss: 1.13792339, Global Avg Loss: 1.45131999, Time: 0.0754 Steps: 101460, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001407, Sample Num: 22512, Cur Loss: 0.90373409, Cur Avg Loss: 1.11142577, Log Avg loss: 1.21243707, Global Avg Loss: 1.45129645, Time: 0.0484 Steps: 101470, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001417, Sample Num: 22672, Cur Loss: 3.31682062, Cur Avg Loss: 1.11302066, Log Avg loss: 1.33742150, Global Avg Loss: 1.45128523, Time: 0.0850 Steps: 101480, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001427, Sample Num: 22832, Cur Loss: 0.63714087, Cur Avg Loss: 1.11179497, Log Avg loss: 0.93811392, Global Avg Loss: 1.45123466, Time: 0.0402 Steps: 101490, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001437, Sample Num: 22992, Cur Loss: 1.18178844, Cur Avg Loss: 1.11161764, Log Avg loss: 1.08631278, Global Avg Loss: 1.45119871, Time: 0.0402 Steps: 101500, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001447, Sample Num: 23152, Cur Loss: 0.96361744, Cur Avg Loss: 1.11075944, Log Avg loss: 0.98743596, Global Avg Loss: 1.45115302, Time: 0.0403 Steps: 101510, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001457, Sample Num: 23312, Cur Loss: 1.51940227, Cur Avg Loss: 1.11098393, Log Avg loss: 1.14346775, Global Avg Loss: 1.45112272, Time: 0.0403 Steps: 101520, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001467, Sample Num: 23472, Cur Loss: 1.05813038, Cur Avg Loss: 1.11161927, Log Avg loss: 1.20418921, Global Avg Loss: 1.45109840, Time: 0.0402 Steps: 101530, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001477, Sample Num: 23632, Cur Loss: 1.23244739, Cur Avg Loss: 1.10989631, Log Avg loss: 0.85713718, Global Avg Loss: 1.45103990, Time: 0.0402 Steps: 101540, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001487, Sample Num: 23792, Cur Loss: 0.57479113, Cur Avg Loss: 1.10871307, Log Avg loss: 0.93394944, Global Avg Loss: 1.45098898, Time: 0.0402 Steps: 101550, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001497, Sample Num: 23952, Cur Loss: 0.40602875, Cur Avg Loss: 1.10707200, Log Avg loss: 0.86304368, Global Avg Loss: 1.45093109, Time: 0.0403 Steps: 101560, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001507, Sample Num: 24112, Cur Loss: 0.51291680, Cur Avg Loss: 1.10633354, Log Avg loss: 0.99578722, Global Avg Loss: 1.45088628, Time: 0.0402 Steps: 101570, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001517, Sample Num: 24272, Cur Loss: 0.68301058, Cur Avg Loss: 1.10525657, Log Avg loss: 0.94295630, Global Avg Loss: 1.45083628, Time: 0.0402 Steps: 101580, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001527, Sample Num: 24432, Cur Loss: 0.69583994, Cur Avg Loss: 1.10427956, Log Avg loss: 0.95606811, Global Avg Loss: 1.45078757, Time: 0.0402 Steps: 101590, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001537, Sample Num: 24592, Cur Loss: 1.11757565, Cur Avg Loss: 1.10534466, Log Avg loss: 1.26798414, Global Avg Loss: 1.45076958, Time: 0.0403 Steps: 101600, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001547, Sample Num: 24752, Cur Loss: 0.59329450, Cur Avg Loss: 1.10442599, Log Avg loss: 0.96322780, Global Avg Loss: 1.45072160, Time: 0.0402 Steps: 101610, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001557, Sample Num: 24912, Cur Loss: 0.47103617, Cur Avg Loss: 1.10320799, Log Avg loss: 0.91478275, Global Avg Loss: 1.45066886, Time: 0.0402 Steps: 101620, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001567, Sample Num: 25072, Cur Loss: 1.19109297, Cur Avg Loss: 1.10152860, Log Avg loss: 0.84004813, Global Avg Loss: 1.45060878, Time: 0.0402 Steps: 101630, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001577, Sample Num: 25232, Cur Loss: 0.42702273, Cur Avg Loss: 1.10115963, Log Avg loss: 1.04334092, Global Avg Loss: 1.45056871, Time: 0.0403 Steps: 101640, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001587, Sample Num: 25392, Cur Loss: 0.99525940, Cur Avg Loss: 1.10064471, Log Avg loss: 1.01944177, Global Avg Loss: 1.45052629, Time: 0.0404 Steps: 101650, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001597, Sample Num: 25552, Cur Loss: 1.74958825, Cur Avg Loss: 1.10154621, Log Avg loss: 1.24461565, Global Avg Loss: 1.45050604, Time: 0.0403 Steps: 101660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001607, Sample Num: 25712, Cur Loss: 1.75848663, Cur Avg Loss: 1.10247241, Log Avg loss: 1.25038651, Global Avg Loss: 1.45048636, Time: 0.0403 Steps: 101670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001617, Sample Num: 25872, Cur Loss: 2.11984539, Cur Avg Loss: 1.10165081, Log Avg loss: 0.96961861, Global Avg Loss: 1.45043906, Time: 0.0403 Steps: 101680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001627, Sample Num: 26032, Cur Loss: 0.31494004, Cur Avg Loss: 1.10245799, Log Avg loss: 1.23297854, Global Avg Loss: 1.45041768, Time: 0.0402 Steps: 101690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001637, Sample Num: 26192, Cur Loss: 1.48238146, Cur Avg Loss: 1.10300262, Log Avg loss: 1.19161472, Global Avg Loss: 1.45039223, Time: 0.0402 Steps: 101700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001647, Sample Num: 26352, Cur Loss: 1.87372243, Cur Avg Loss: 1.10288368, Log Avg loss: 1.08341248, Global Avg Loss: 1.45035615, Time: 0.0402 Steps: 101710, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001657, Sample Num: 26512, Cur Loss: 0.80506754, Cur Avg Loss: 1.10274481, Log Avg loss: 1.07987396, Global Avg Loss: 1.45031973, Time: 0.0402 Steps: 101720, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001667, Sample Num: 26672, Cur Loss: 1.11758518, Cur Avg Loss: 1.10270274, Log Avg loss: 1.09573141, Global Avg Loss: 1.45028487, Time: 0.0402 Steps: 101730, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001677, Sample Num: 26832, Cur Loss: 1.89884603, Cur Avg Loss: 1.10389207, Log Avg loss: 1.30215413, Global Avg Loss: 1.45027031, Time: 0.0402 Steps: 101740, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001687, Sample Num: 26992, Cur Loss: 1.80568075, Cur Avg Loss: 1.10305077, Log Avg loss: 0.96196344, Global Avg Loss: 1.45022232, Time: 0.0459 Steps: 101750, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001697, Sample Num: 27152, Cur Loss: 0.26972327, Cur Avg Loss: 1.10038893, Log Avg loss: 0.65133761, Global Avg Loss: 1.45014382, Time: 0.0446 Steps: 101760, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001707, Sample Num: 27312, Cur Loss: 2.48832655, Cur Avg Loss: 1.10130642, Log Avg loss: 1.25700343, Global Avg Loss: 1.45012484, Time: 0.0509 Steps: 101770, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001717, Sample Num: 27472, Cur Loss: 0.32868427, Cur Avg Loss: 1.10254018, Log Avg loss: 1.31314273, Global Avg Loss: 1.45011138, Time: 0.0421 Steps: 101780, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001727, Sample Num: 27632, Cur Loss: 1.88866293, Cur Avg Loss: 1.10323730, Log Avg loss: 1.22293431, Global Avg Loss: 1.45008906, Time: 0.0899 Steps: 101790, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001737, Sample Num: 27792, Cur Loss: 1.49496996, Cur Avg Loss: 1.10246398, Log Avg loss: 0.96891095, Global Avg Loss: 1.45004179, Time: 0.0519 Steps: 101800, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001747, Sample Num: 27952, Cur Loss: 0.47992456, Cur Avg Loss: 1.10218707, Log Avg loss: 1.05408775, Global Avg Loss: 1.45000290, Time: 0.0403 Steps: 101810, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001757, Sample Num: 28112, Cur Loss: 2.73027635, Cur Avg Loss: 1.10339864, Log Avg loss: 1.31505964, Global Avg Loss: 1.44998965, Time: 0.0483 Steps: 101820, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001767, Sample Num: 28272, Cur Loss: 1.85745037, Cur Avg Loss: 1.10357833, Log Avg loss: 1.13515100, Global Avg Loss: 1.44995873, Time: 0.0686 Steps: 101830, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001777, Sample Num: 28432, Cur Loss: 2.19021726, Cur Avg Loss: 1.10465637, Log Avg loss: 1.29514533, Global Avg Loss: 1.44994353, Time: 0.0404 Steps: 101840, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001787, Sample Num: 28592, Cur Loss: 1.41043818, Cur Avg Loss: 1.10626037, Log Avg loss: 1.39129154, Global Avg Loss: 1.44993777, Time: 0.0403 Steps: 101850, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001797, Sample Num: 28752, Cur Loss: 1.33819509, Cur Avg Loss: 1.10650974, Log Avg loss: 1.15107210, Global Avg Loss: 1.44990843, Time: 0.0403 Steps: 101860, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001807, Sample Num: 28912, Cur Loss: 1.31067109, Cur Avg Loss: 1.10685489, Log Avg loss: 1.16887843, Global Avg Loss: 1.44988084, Time: 0.0403 Steps: 101870, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001817, Sample Num: 29072, Cur Loss: 0.45956406, Cur Avg Loss: 1.10601025, Log Avg loss: 0.95338390, Global Avg Loss: 1.44983211, Time: 0.0404 Steps: 101880, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001827, Sample Num: 29232, Cur Loss: 1.40531850, Cur Avg Loss: 1.10641126, Log Avg loss: 1.17927352, Global Avg Loss: 1.44980556, Time: 0.0403 Steps: 101890, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001837, Sample Num: 29392, Cur Loss: 0.77509981, Cur Avg Loss: 1.10643058, Log Avg loss: 1.10996052, Global Avg Loss: 1.44977220, Time: 0.0403 Steps: 101900, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001847, Sample Num: 29552, Cur Loss: 0.89460081, Cur Avg Loss: 1.10486214, Log Avg loss: 0.81674074, Global Avg Loss: 1.44971009, Time: 0.0402 Steps: 101910, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001857, Sample Num: 29712, Cur Loss: 0.61353946, Cur Avg Loss: 1.10372899, Log Avg loss: 0.89443512, Global Avg Loss: 1.44965561, Time: 0.0403 Steps: 101920, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001867, Sample Num: 29872, Cur Loss: 0.64011014, Cur Avg Loss: 1.10343653, Log Avg loss: 1.04912786, Global Avg Loss: 1.44961631, Time: 0.0403 Steps: 101930, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001877, Sample Num: 30032, Cur Loss: 0.47643438, Cur Avg Loss: 1.10302106, Log Avg loss: 1.02545274, Global Avg Loss: 1.44957470, Time: 0.0403 Steps: 101940, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001887, Sample Num: 30192, Cur Loss: 0.87069052, Cur Avg Loss: 1.10445212, Log Avg loss: 1.37306096, Global Avg Loss: 1.44956720, Time: 0.0402 Steps: 101950, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001897, Sample Num: 30352, Cur Loss: 1.17917967, Cur Avg Loss: 1.10524228, Log Avg loss: 1.25434596, Global Avg Loss: 1.44954805, Time: 0.0403 Steps: 101960, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001907, Sample Num: 30512, Cur Loss: 0.68126082, Cur Avg Loss: 1.10609990, Log Avg loss: 1.26879110, Global Avg Loss: 1.44953032, Time: 0.0403 Steps: 101970, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001917, Sample Num: 30672, Cur Loss: 1.29063606, Cur Avg Loss: 1.10664658, Log Avg loss: 1.21089725, Global Avg Loss: 1.44950692, Time: 0.0403 Steps: 101980, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001927, Sample Num: 30832, Cur Loss: 0.73667699, Cur Avg Loss: 1.10559820, Log Avg loss: 0.90462521, Global Avg Loss: 1.44945350, Time: 0.0402 Steps: 101990, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001937, Sample Num: 30992, Cur Loss: 1.20393348, Cur Avg Loss: 1.10527230, Log Avg loss: 1.04247046, Global Avg Loss: 1.44941360, Time: 0.0402 Steps: 102000, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001947, Sample Num: 31152, Cur Loss: 1.42420888, Cur Avg Loss: 1.10407257, Log Avg loss: 0.87168424, Global Avg Loss: 1.44935696, Time: 0.0402 Steps: 102010, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001957, Sample Num: 31312, Cur Loss: 0.26863694, Cur Avg Loss: 1.10367730, Log Avg loss: 1.02671819, Global Avg Loss: 1.44931554, Time: 0.0403 Steps: 102020, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001967, Sample Num: 31472, Cur Loss: 0.91771460, Cur Avg Loss: 1.10361376, Log Avg loss: 1.09118008, Global Avg Loss: 1.44928044, Time: 0.0403 Steps: 102030, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001977, Sample Num: 31632, Cur Loss: 1.19857025, Cur Avg Loss: 1.10347204, Log Avg loss: 1.07559481, Global Avg Loss: 1.44924381, Time: 0.0402 Steps: 102040, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001987, Sample Num: 31792, Cur Loss: 0.75579983, Cur Avg Loss: 1.10392592, Log Avg loss: 1.19365944, Global Avg Loss: 1.44921877, Time: 0.0403 Steps: 102050, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001997, Sample Num: 31952, Cur Loss: 1.76232600, Cur Avg Loss: 1.10572408, Log Avg loss: 1.46301658, Global Avg Loss: 1.44922012, Time: 0.0403 Steps: 102060, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002007, Sample Num: 32112, Cur Loss: 1.07687211, Cur Avg Loss: 1.10566561, Log Avg loss: 1.09398926, Global Avg Loss: 1.44918532, Time: 0.0402 Steps: 102070, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002017, Sample Num: 32272, Cur Loss: 0.79474008, Cur Avg Loss: 1.10680575, Log Avg loss: 1.33563318, Global Avg Loss: 1.44917420, Time: 0.0402 Steps: 102080, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002027, Sample Num: 32432, Cur Loss: 2.05312872, Cur Avg Loss: 1.10564047, Log Avg loss: 0.87060299, Global Avg Loss: 1.44911752, Time: 0.0403 Steps: 102090, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002037, Sample Num: 32592, Cur Loss: 0.84019125, Cur Avg Loss: 1.10593133, Log Avg loss: 1.16488851, Global Avg Loss: 1.44908968, Time: 0.0404 Steps: 102100, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002047, Sample Num: 32752, Cur Loss: 0.52226406, Cur Avg Loss: 1.10412179, Log Avg loss: 0.73551775, Global Avg Loss: 1.44901980, Time: 0.0616 Steps: 102110, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002057, Sample Num: 32912, Cur Loss: 0.41787571, Cur Avg Loss: 1.10375762, Log Avg loss: 1.02921391, Global Avg Loss: 1.44897869, Time: 0.0448 Steps: 102120, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002067, Sample Num: 33072, Cur Loss: 3.33756971, Cur Avg Loss: 1.10389121, Log Avg loss: 1.13136878, Global Avg Loss: 1.44894759, Time: 0.0406 Steps: 102130, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002077, Sample Num: 33232, Cur Loss: 2.38258696, Cur Avg Loss: 1.10316081, Log Avg loss: 0.95218809, Global Avg Loss: 1.44889896, Time: 0.0581 Steps: 102140, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002087, Sample Num: 33392, Cur Loss: 1.61719728, Cur Avg Loss: 1.10322302, Log Avg loss: 1.11614323, Global Avg Loss: 1.44886638, Time: 0.0593 Steps: 102150, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002097, Sample Num: 33552, Cur Loss: 1.82075703, Cur Avg Loss: 1.10582045, Log Avg loss: 1.64790539, Global Avg Loss: 1.44888587, Time: 0.0404 Steps: 102160, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002107, Sample Num: 33712, Cur Loss: 1.99351621, Cur Avg Loss: 1.10645252, Log Avg loss: 1.23899776, Global Avg Loss: 1.44886532, Time: 0.0531 Steps: 102170, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002117, Sample Num: 33872, Cur Loss: 0.56539571, Cur Avg Loss: 1.10553163, Log Avg loss: 0.91149985, Global Avg Loss: 1.44881273, Time: 0.0560 Steps: 102180, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002127, Sample Num: 34032, Cur Loss: 0.60012633, Cur Avg Loss: 1.10408768, Log Avg loss: 0.79840400, Global Avg Loss: 1.44874909, Time: 0.0406 Steps: 102190, Updated lr: 0.000004 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.977917, Avg time per batch (s): 0.050000 {"eval_avg_loss": 1.095306, "eval_total_loss": 770.000325, "eval_mae": 0.868845, "eval_mse": 1.095359, "eval_r2": 0.303718, "eval_sp_statistic": 0.486543, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.586352, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.092433, "test_total_loss": 548.401274, "test_mae": 0.879922, "test_mse": 1.092608, "test_r2": 0.294821, "test_sp_statistic": 0.503909, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.628846, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4487384317058698, "train_cur_epoch_loss": 2350.2031624764204, "train_cur_epoch_avg_loss": 1.10390002934543, "train_cur_epoch_time": 95.97791695594788, "train_cur_epoch_avg_time": 0.045081219800820985, "epoch": 48, "step": 102192} ################################################## Training, Epoch: 0049, Batch: 000008, Sample Num: 128, Cur Loss: 0.75596046, Cur Avg Loss: 0.92300713, Log Avg loss: 0.91927147, Global Avg Loss: 1.44869728, Time: 0.0586 Steps: 102200, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000018, Sample Num: 288, Cur Loss: 1.13527846, Cur Avg Loss: 1.01703677, Log Avg loss: 1.09226049, Global Avg Loss: 1.44866241, Time: 0.0659 Steps: 102210, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000028, Sample Num: 448, Cur Loss: 1.56121123, Cur Avg Loss: 1.06569847, Log Avg loss: 1.15328953, Global Avg Loss: 1.44863351, Time: 0.0731 Steps: 102220, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000038, Sample Num: 608, Cur Loss: 0.36271459, Cur Avg Loss: 1.02114156, Log Avg loss: 0.89638222, Global Avg Loss: 1.44857949, Time: 0.0405 Steps: 102230, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000048, Sample Num: 768, Cur Loss: 0.92714614, Cur Avg Loss: 1.07493466, Log Avg loss: 1.27934841, Global Avg Loss: 1.44856294, Time: 0.0538 Steps: 102240, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000058, Sample Num: 928, Cur Loss: 0.66533411, Cur Avg Loss: 1.01893165, Log Avg loss: 0.75011720, Global Avg Loss: 1.44849463, Time: 0.0780 Steps: 102250, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000068, Sample Num: 1088, Cur Loss: 0.82088923, Cur Avg Loss: 1.05491738, Log Avg loss: 1.26363465, Global Avg Loss: 1.44847655, Time: 0.0479 Steps: 102260, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000078, Sample Num: 1248, Cur Loss: 0.41738647, Cur Avg Loss: 1.03806136, Log Avg loss: 0.92344040, Global Avg Loss: 1.44842521, Time: 0.0403 Steps: 102270, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000088, Sample Num: 1408, Cur Loss: 1.74865234, Cur Avg Loss: 1.07547282, Log Avg loss: 1.36728218, Global Avg Loss: 1.44841728, Time: 0.0402 Steps: 102280, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000098, Sample Num: 1568, Cur Loss: 1.60580087, Cur Avg Loss: 1.08687175, Log Avg loss: 1.18718239, Global Avg Loss: 1.44839174, Time: 0.0402 Steps: 102290, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000108, Sample Num: 1728, Cur Loss: 1.24268854, Cur Avg Loss: 1.10086529, Log Avg loss: 1.23800193, Global Avg Loss: 1.44837118, Time: 0.0402 Steps: 102300, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000118, Sample Num: 1888, Cur Loss: 0.92982304, Cur Avg Loss: 1.10887081, Log Avg loss: 1.19533042, Global Avg Loss: 1.44834644, Time: 0.0403 Steps: 102310, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000128, Sample Num: 2048, Cur Loss: 1.16730082, Cur Avg Loss: 1.10787368, Log Avg loss: 1.09610754, Global Avg Loss: 1.44831202, Time: 0.0404 Steps: 102320, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000138, Sample Num: 2208, Cur Loss: 0.10019098, Cur Avg Loss: 1.10931861, Log Avg loss: 1.12781381, Global Avg Loss: 1.44828070, Time: 0.0402 Steps: 102330, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000148, Sample Num: 2368, Cur Loss: 0.58161467, Cur Avg Loss: 1.10765457, Log Avg loss: 1.08469079, Global Avg Loss: 1.44824517, Time: 0.0403 Steps: 102340, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000158, Sample Num: 2528, Cur Loss: 1.08116138, Cur Avg Loss: 1.10527230, Log Avg loss: 1.07001469, Global Avg Loss: 1.44820822, Time: 0.0403 Steps: 102350, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000168, Sample Num: 2688, Cur Loss: 1.29197764, Cur Avg Loss: 1.10612060, Log Avg loss: 1.11952370, Global Avg Loss: 1.44817610, Time: 0.0403 Steps: 102360, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000178, Sample Num: 2848, Cur Loss: 2.70975137, Cur Avg Loss: 1.10906920, Log Avg loss: 1.15860569, Global Avg Loss: 1.44814782, Time: 0.0403 Steps: 102370, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000188, Sample Num: 3008, Cur Loss: 2.28960323, Cur Avg Loss: 1.12693942, Log Avg loss: 1.44502936, Global Avg Loss: 1.44814751, Time: 0.0402 Steps: 102380, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000198, Sample Num: 3168, Cur Loss: 0.69038814, Cur Avg Loss: 1.13884355, Log Avg loss: 1.36264127, Global Avg Loss: 1.44813916, Time: 0.0402 Steps: 102390, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000208, Sample Num: 3328, Cur Loss: 1.41075873, Cur Avg Loss: 1.14529205, Log Avg loss: 1.27297234, Global Avg Loss: 1.44812206, Time: 0.0402 Steps: 102400, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000218, Sample Num: 3488, Cur Loss: 1.64286947, Cur Avg Loss: 1.14095994, Log Avg loss: 1.05085192, Global Avg Loss: 1.44808326, Time: 0.0403 Steps: 102410, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000228, Sample Num: 3648, Cur Loss: 1.21064723, Cur Avg Loss: 1.13601748, Log Avg loss: 1.02827193, Global Avg Loss: 1.44804227, Time: 0.0403 Steps: 102420, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000238, Sample Num: 3808, Cur Loss: 0.58424306, Cur Avg Loss: 1.12625610, Log Avg loss: 0.90369667, Global Avg Loss: 1.44798913, Time: 0.0402 Steps: 102430, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000248, Sample Num: 3968, Cur Loss: 0.73607635, Cur Avg Loss: 1.11649364, Log Avg loss: 0.88414710, Global Avg Loss: 1.44793409, Time: 0.0402 Steps: 102440, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000258, Sample Num: 4128, Cur Loss: 1.21523774, Cur Avg Loss: 1.11212890, Log Avg loss: 1.00388325, Global Avg Loss: 1.44789075, Time: 0.0403 Steps: 102450, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000268, Sample Num: 4288, Cur Loss: 1.98149943, Cur Avg Loss: 1.11271329, Log Avg loss: 1.12779048, Global Avg Loss: 1.44785951, Time: 0.0402 Steps: 102460, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000278, Sample Num: 4448, Cur Loss: 0.52736443, Cur Avg Loss: 1.10745795, Log Avg loss: 0.96661490, Global Avg Loss: 1.44781254, Time: 0.0402 Steps: 102470, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000288, Sample Num: 4608, Cur Loss: 1.81649423, Cur Avg Loss: 1.10676295, Log Avg loss: 1.08744193, Global Avg Loss: 1.44777738, Time: 0.0403 Steps: 102480, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000298, Sample Num: 4768, Cur Loss: 0.58082378, Cur Avg Loss: 1.10981739, Log Avg loss: 1.19778521, Global Avg Loss: 1.44775298, Time: 0.0402 Steps: 102490, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000308, Sample Num: 4928, Cur Loss: 2.71143150, Cur Avg Loss: 1.10827246, Log Avg loss: 1.06223376, Global Avg Loss: 1.44771537, Time: 0.0402 Steps: 102500, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000318, Sample Num: 5088, Cur Loss: 1.11822319, Cur Avg Loss: 1.10444940, Log Avg loss: 0.98669892, Global Avg Loss: 1.44767040, Time: 0.0402 Steps: 102510, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000328, Sample Num: 5248, Cur Loss: 1.52557707, Cur Avg Loss: 1.10548698, Log Avg loss: 1.13848230, Global Avg Loss: 1.44764024, Time: 0.0403 Steps: 102520, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000338, Sample Num: 5408, Cur Loss: 0.53825986, Cur Avg Loss: 1.10716009, Log Avg loss: 1.16203804, Global Avg Loss: 1.44761239, Time: 0.0402 Steps: 102530, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000348, Sample Num: 5568, Cur Loss: 0.64573383, Cur Avg Loss: 1.11692174, Log Avg loss: 1.44686539, Global Avg Loss: 1.44761231, Time: 0.0761 Steps: 102540, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000358, Sample Num: 5728, Cur Loss: 0.59085530, Cur Avg Loss: 1.11762442, Log Avg loss: 1.14207790, Global Avg Loss: 1.44758252, Time: 0.0409 Steps: 102550, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000368, Sample Num: 5888, Cur Loss: 0.64704829, Cur Avg Loss: 1.11205028, Log Avg loss: 0.91249573, Global Avg Loss: 1.44753035, Time: 0.0429 Steps: 102560, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000378, Sample Num: 6048, Cur Loss: 1.87292826, Cur Avg Loss: 1.11182888, Log Avg loss: 1.10368147, Global Avg Loss: 1.44749682, Time: 0.0603 Steps: 102570, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000388, Sample Num: 6208, Cur Loss: 1.51223087, Cur Avg Loss: 1.11676502, Log Avg loss: 1.30335113, Global Avg Loss: 1.44748277, Time: 0.0689 Steps: 102580, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000398, Sample Num: 6368, Cur Loss: 1.26341355, Cur Avg Loss: 1.11988326, Log Avg loss: 1.24087103, Global Avg Loss: 1.44746263, Time: 0.0829 Steps: 102590, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000408, Sample Num: 6528, Cur Loss: 1.29950833, Cur Avg Loss: 1.11981182, Log Avg loss: 1.11696864, Global Avg Loss: 1.44743042, Time: 0.0709 Steps: 102600, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000418, Sample Num: 6688, Cur Loss: 0.85046279, Cur Avg Loss: 1.11696574, Log Avg loss: 1.00084550, Global Avg Loss: 1.44738690, Time: 0.0890 Steps: 102610, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000428, Sample Num: 6848, Cur Loss: 1.81542754, Cur Avg Loss: 1.11847092, Log Avg loss: 1.18138731, Global Avg Loss: 1.44736098, Time: 0.0992 Steps: 102620, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000438, Sample Num: 7008, Cur Loss: 0.37447804, Cur Avg Loss: 1.12115429, Log Avg loss: 1.23600283, Global Avg Loss: 1.44734038, Time: 0.0562 Steps: 102630, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000448, Sample Num: 7168, Cur Loss: 0.41967243, Cur Avg Loss: 1.11245829, Log Avg loss: 0.73157316, Global Avg Loss: 1.44727065, Time: 0.0402 Steps: 102640, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000458, Sample Num: 7328, Cur Loss: 0.92042220, Cur Avg Loss: 1.10682168, Log Avg loss: 0.85430194, Global Avg Loss: 1.44721288, Time: 0.0402 Steps: 102650, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000468, Sample Num: 7488, Cur Loss: 0.55901074, Cur Avg Loss: 1.10656376, Log Avg loss: 1.09475071, Global Avg Loss: 1.44717855, Time: 0.0403 Steps: 102660, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000478, Sample Num: 7648, Cur Loss: 2.29755592, Cur Avg Loss: 1.10910747, Log Avg loss: 1.22815297, Global Avg Loss: 1.44715721, Time: 0.0402 Steps: 102670, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000488, Sample Num: 7808, Cur Loss: 0.97929186, Cur Avg Loss: 1.11294128, Log Avg loss: 1.29619764, Global Avg Loss: 1.44714251, Time: 0.0402 Steps: 102680, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000498, Sample Num: 7968, Cur Loss: 0.98407590, Cur Avg Loss: 1.11058404, Log Avg loss: 0.99555051, Global Avg Loss: 1.44709854, Time: 0.0402 Steps: 102690, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000508, Sample Num: 8128, Cur Loss: 1.76867056, Cur Avg Loss: 1.11039784, Log Avg loss: 1.10112532, Global Avg Loss: 1.44706485, Time: 0.0403 Steps: 102700, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000518, Sample Num: 8288, Cur Loss: 0.52436566, Cur Avg Loss: 1.11255724, Log Avg loss: 1.22225457, Global Avg Loss: 1.44704296, Time: 0.0403 Steps: 102710, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000528, Sample Num: 8448, Cur Loss: 0.65078491, Cur Avg Loss: 1.10980296, Log Avg loss: 0.96713126, Global Avg Loss: 1.44699624, Time: 0.0403 Steps: 102720, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000538, Sample Num: 8608, Cur Loss: 0.45173547, Cur Avg Loss: 1.10517952, Log Avg loss: 0.86106188, Global Avg Loss: 1.44693920, Time: 0.0403 Steps: 102730, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000548, Sample Num: 8768, Cur Loss: 1.02220809, Cur Avg Loss: 1.10498017, Log Avg loss: 1.09425517, Global Avg Loss: 1.44690488, Time: 0.0402 Steps: 102740, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000558, Sample Num: 8928, Cur Loss: 0.45871657, Cur Avg Loss: 1.10207452, Log Avg loss: 0.94284500, Global Avg Loss: 1.44685582, Time: 0.0403 Steps: 102750, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000568, Sample Num: 9088, Cur Loss: 0.71371198, Cur Avg Loss: 1.10201255, Log Avg loss: 1.09855478, Global Avg Loss: 1.44682192, Time: 0.0403 Steps: 102760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000578, Sample Num: 9248, Cur Loss: 1.09596443, Cur Avg Loss: 1.10406862, Log Avg loss: 1.22085311, Global Avg Loss: 1.44679994, Time: 0.0402 Steps: 102770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000588, Sample Num: 9408, Cur Loss: 2.49887991, Cur Avg Loss: 1.10488641, Log Avg loss: 1.15215455, Global Avg Loss: 1.44677127, Time: 0.0402 Steps: 102780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000598, Sample Num: 9568, Cur Loss: 0.66611373, Cur Avg Loss: 1.10083693, Log Avg loss: 0.86272805, Global Avg Loss: 1.44671445, Time: 0.0402 Steps: 102790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000608, Sample Num: 9728, Cur Loss: 1.56733179, Cur Avg Loss: 1.09701364, Log Avg loss: 0.86838093, Global Avg Loss: 1.44665819, Time: 0.0403 Steps: 102800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000618, Sample Num: 9888, Cur Loss: 1.84988749, Cur Avg Loss: 1.09299903, Log Avg loss: 0.84891073, Global Avg Loss: 1.44660005, Time: 0.0402 Steps: 102810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000628, Sample Num: 10048, Cur Loss: 0.25775093, Cur Avg Loss: 1.09167587, Log Avg loss: 1.00990408, Global Avg Loss: 1.44655758, Time: 0.0402 Steps: 102820, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000638, Sample Num: 10208, Cur Loss: 0.82810360, Cur Avg Loss: 1.09260041, Log Avg loss: 1.15066179, Global Avg Loss: 1.44652880, Time: 0.0402 Steps: 102830, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000648, Sample Num: 10368, Cur Loss: 0.80407387, Cur Avg Loss: 1.09201822, Log Avg loss: 1.05487421, Global Avg Loss: 1.44649072, Time: 0.0402 Steps: 102840, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000658, Sample Num: 10528, Cur Loss: 1.05697739, Cur Avg Loss: 1.09404591, Log Avg loss: 1.22544044, Global Avg Loss: 1.44646923, Time: 0.0402 Steps: 102850, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000668, Sample Num: 10688, Cur Loss: 0.93519199, Cur Avg Loss: 1.09340484, Log Avg loss: 1.05122225, Global Avg Loss: 1.44643080, Time: 0.0402 Steps: 102860, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000678, Sample Num: 10848, Cur Loss: 0.46579349, Cur Avg Loss: 1.09649949, Log Avg loss: 1.30322235, Global Avg Loss: 1.44641688, Time: 0.0402 Steps: 102870, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000688, Sample Num: 11008, Cur Loss: 1.50793576, Cur Avg Loss: 1.09341286, Log Avg loss: 0.88413917, Global Avg Loss: 1.44636223, Time: 0.0402 Steps: 102880, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000698, Sample Num: 11168, Cur Loss: 1.27659106, Cur Avg Loss: 1.09082793, Log Avg loss: 0.91298503, Global Avg Loss: 1.44631039, Time: 0.0403 Steps: 102890, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000708, Sample Num: 11328, Cur Loss: 1.49519050, Cur Avg Loss: 1.09050508, Log Avg loss: 1.06796992, Global Avg Loss: 1.44627362, Time: 0.0403 Steps: 102900, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000718, Sample Num: 11488, Cur Loss: 1.19954491, Cur Avg Loss: 1.09463780, Log Avg loss: 1.38723477, Global Avg Loss: 1.44626788, Time: 0.0836 Steps: 102910, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000728, Sample Num: 11648, Cur Loss: 0.59365344, Cur Avg Loss: 1.09320158, Log Avg loss: 0.99008035, Global Avg Loss: 1.44622356, Time: 0.0780 Steps: 102920, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000738, Sample Num: 11808, Cur Loss: 1.73240447, Cur Avg Loss: 1.09194803, Log Avg loss: 1.00068964, Global Avg Loss: 1.44618027, Time: 0.0431 Steps: 102930, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000748, Sample Num: 11968, Cur Loss: 1.60712266, Cur Avg Loss: 1.09387696, Log Avg loss: 1.23623198, Global Avg Loss: 1.44615988, Time: 0.0510 Steps: 102940, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000758, Sample Num: 12128, Cur Loss: 0.94248319, Cur Avg Loss: 1.09326315, Log Avg loss: 1.04735047, Global Avg Loss: 1.44612114, Time: 0.0412 Steps: 102950, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000768, Sample Num: 12288, Cur Loss: 0.77642518, Cur Avg Loss: 1.09030970, Log Avg loss: 0.86643786, Global Avg Loss: 1.44606484, Time: 0.0407 Steps: 102960, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000778, Sample Num: 12448, Cur Loss: 1.60905647, Cur Avg Loss: 1.09186953, Log Avg loss: 1.21166509, Global Avg Loss: 1.44604207, Time: 0.0494 Steps: 102970, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000788, Sample Num: 12608, Cur Loss: 0.80862987, Cur Avg Loss: 1.09116208, Log Avg loss: 1.03612181, Global Avg Loss: 1.44600227, Time: 0.0415 Steps: 102980, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000798, Sample Num: 12768, Cur Loss: 1.36328459, Cur Avg Loss: 1.09108970, Log Avg loss: 1.08538657, Global Avg Loss: 1.44596725, Time: 0.0403 Steps: 102990, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000808, Sample Num: 12928, Cur Loss: 1.20379925, Cur Avg Loss: 1.09387643, Log Avg loss: 1.31625720, Global Avg Loss: 1.44595466, Time: 0.0402 Steps: 103000, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000818, Sample Num: 13088, Cur Loss: 0.94106752, Cur Avg Loss: 1.09358237, Log Avg loss: 1.06982293, Global Avg Loss: 1.44591815, Time: 0.0403 Steps: 103010, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000828, Sample Num: 13248, Cur Loss: 1.39525688, Cur Avg Loss: 1.09532995, Log Avg loss: 1.23828156, Global Avg Loss: 1.44589799, Time: 0.0403 Steps: 103020, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000838, Sample Num: 13408, Cur Loss: 1.59792840, Cur Avg Loss: 1.09516913, Log Avg loss: 1.08185298, Global Avg Loss: 1.44586266, Time: 0.0402 Steps: 103030, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000848, Sample Num: 13568, Cur Loss: 1.71702182, Cur Avg Loss: 1.09792314, Log Avg loss: 1.32870981, Global Avg Loss: 1.44585129, Time: 0.0403 Steps: 103040, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000858, Sample Num: 13728, Cur Loss: 0.86445689, Cur Avg Loss: 1.10036238, Log Avg loss: 1.30720961, Global Avg Loss: 1.44583783, Time: 0.0402 Steps: 103050, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000868, Sample Num: 13888, Cur Loss: 0.97979218, Cur Avg Loss: 1.10009608, Log Avg loss: 1.07724772, Global Avg Loss: 1.44580207, Time: 0.0403 Steps: 103060, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000878, Sample Num: 14048, Cur Loss: 1.06471026, Cur Avg Loss: 1.09925737, Log Avg loss: 1.02645706, Global Avg Loss: 1.44576138, Time: 0.0403 Steps: 103070, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000888, Sample Num: 14208, Cur Loss: 1.38238692, Cur Avg Loss: 1.09990167, Log Avg loss: 1.15647140, Global Avg Loss: 1.44573332, Time: 0.0403 Steps: 103080, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000898, Sample Num: 14368, Cur Loss: 0.73583436, Cur Avg Loss: 1.09644324, Log Avg loss: 0.78933487, Global Avg Loss: 1.44566965, Time: 0.0403 Steps: 103090, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000908, Sample Num: 14528, Cur Loss: 1.61041117, Cur Avg Loss: 1.09856186, Log Avg loss: 1.28881388, Global Avg Loss: 1.44565443, Time: 0.0403 Steps: 103100, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000918, Sample Num: 14688, Cur Loss: 0.40734148, Cur Avg Loss: 1.09756837, Log Avg loss: 1.00735955, Global Avg Loss: 1.44561192, Time: 0.0402 Steps: 103110, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000928, Sample Num: 14848, Cur Loss: 0.59538978, Cur Avg Loss: 1.09758335, Log Avg loss: 1.09895807, Global Avg Loss: 1.44557831, Time: 0.0403 Steps: 103120, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000938, Sample Num: 15008, Cur Loss: 0.24588710, Cur Avg Loss: 1.09646510, Log Avg loss: 0.99269188, Global Avg Loss: 1.44553439, Time: 0.0403 Steps: 103130, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000948, Sample Num: 15168, Cur Loss: 1.12096024, Cur Avg Loss: 1.09769630, Log Avg loss: 1.21318250, Global Avg Loss: 1.44551187, Time: 0.0403 Steps: 103140, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000958, Sample Num: 15328, Cur Loss: 1.21379018, Cur Avg Loss: 1.09729325, Log Avg loss: 1.05908391, Global Avg Loss: 1.44547440, Time: 0.0403 Steps: 103150, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000968, Sample Num: 15488, Cur Loss: 0.80921382, Cur Avg Loss: 1.09570811, Log Avg loss: 0.94385230, Global Avg Loss: 1.44542578, Time: 0.0403 Steps: 103160, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000978, Sample Num: 15648, Cur Loss: 2.58187366, Cur Avg Loss: 1.09668171, Log Avg loss: 1.19092636, Global Avg Loss: 1.44540111, Time: 0.0403 Steps: 103170, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000988, Sample Num: 15808, Cur Loss: 1.36062694, Cur Avg Loss: 1.09622881, Log Avg loss: 1.05193515, Global Avg Loss: 1.44536298, Time: 0.0403 Steps: 103180, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000998, Sample Num: 15968, Cur Loss: 0.88058245, Cur Avg Loss: 1.09530363, Log Avg loss: 1.00389577, Global Avg Loss: 1.44532019, Time: 0.0402 Steps: 103190, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001008, Sample Num: 16128, Cur Loss: 0.61123151, Cur Avg Loss: 1.09577025, Log Avg loss: 1.14233852, Global Avg Loss: 1.44529084, Time: 0.0402 Steps: 103200, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001018, Sample Num: 16288, Cur Loss: 1.01735747, Cur Avg Loss: 1.09461790, Log Avg loss: 0.97846149, Global Avg Loss: 1.44524560, Time: 0.0402 Steps: 103210, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001028, Sample Num: 16448, Cur Loss: 0.61495113, Cur Avg Loss: 1.09428901, Log Avg loss: 1.06080737, Global Avg Loss: 1.44520836, Time: 0.0402 Steps: 103220, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001038, Sample Num: 16608, Cur Loss: 0.51080585, Cur Avg Loss: 1.09385136, Log Avg loss: 1.04886119, Global Avg Loss: 1.44516997, Time: 0.0402 Steps: 103230, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001048, Sample Num: 16768, Cur Loss: 0.64741504, Cur Avg Loss: 1.09253097, Log Avg loss: 0.95547496, Global Avg Loss: 1.44512253, Time: 0.0430 Steps: 103240, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001058, Sample Num: 16928, Cur Loss: 2.64658809, Cur Avg Loss: 1.09501343, Log Avg loss: 1.35517453, Global Avg Loss: 1.44511382, Time: 0.0845 Steps: 103250, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001068, Sample Num: 17088, Cur Loss: 0.59449804, Cur Avg Loss: 1.09276545, Log Avg loss: 0.85492904, Global Avg Loss: 1.44505667, Time: 0.0510 Steps: 103260, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001078, Sample Num: 17248, Cur Loss: 0.97222716, Cur Avg Loss: 1.09172097, Log Avg loss: 0.98017095, Global Avg Loss: 1.44501165, Time: 0.0406 Steps: 103270, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001088, Sample Num: 17408, Cur Loss: 1.28846002, Cur Avg Loss: 1.09192240, Log Avg loss: 1.11363680, Global Avg Loss: 1.44497956, Time: 0.0403 Steps: 103280, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001098, Sample Num: 17568, Cur Loss: 1.69677508, Cur Avg Loss: 1.09489461, Log Avg loss: 1.41827089, Global Avg Loss: 1.44497698, Time: 0.0573 Steps: 103290, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001108, Sample Num: 17728, Cur Loss: 0.32312024, Cur Avg Loss: 1.09393926, Log Avg loss: 0.98904127, Global Avg Loss: 1.44493284, Time: 0.0403 Steps: 103300, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001118, Sample Num: 17888, Cur Loss: 0.63928092, Cur Avg Loss: 1.09335312, Log Avg loss: 1.02840862, Global Avg Loss: 1.44489252, Time: 0.0473 Steps: 103310, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001128, Sample Num: 18048, Cur Loss: 1.26509798, Cur Avg Loss: 1.09359259, Log Avg loss: 1.12036533, Global Avg Loss: 1.44486111, Time: 0.0571 Steps: 103320, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001138, Sample Num: 18208, Cur Loss: 1.20452642, Cur Avg Loss: 1.09308713, Log Avg loss: 1.03607126, Global Avg Loss: 1.44482155, Time: 0.0404 Steps: 103330, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001148, Sample Num: 18368, Cur Loss: 1.06350648, Cur Avg Loss: 1.09339954, Log Avg loss: 1.12895251, Global Avg Loss: 1.44479099, Time: 0.0402 Steps: 103340, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001158, Sample Num: 18528, Cur Loss: 0.90834093, Cur Avg Loss: 1.09483314, Log Avg loss: 1.25941004, Global Avg Loss: 1.44477305, Time: 0.0402 Steps: 103350, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001168, Sample Num: 18688, Cur Loss: 1.11614907, Cur Avg Loss: 1.09459992, Log Avg loss: 1.06759278, Global Avg Loss: 1.44473656, Time: 0.0403 Steps: 103360, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001178, Sample Num: 18848, Cur Loss: 2.34397578, Cur Avg Loss: 1.09483238, Log Avg loss: 1.12198386, Global Avg Loss: 1.44470533, Time: 0.0402 Steps: 103370, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001188, Sample Num: 19008, Cur Loss: 0.39984435, Cur Avg Loss: 1.09385760, Log Avg loss: 0.97902912, Global Avg Loss: 1.44466029, Time: 0.0403 Steps: 103380, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001198, Sample Num: 19168, Cur Loss: 0.78933734, Cur Avg Loss: 1.09113829, Log Avg loss: 0.76808422, Global Avg Loss: 1.44459485, Time: 0.0402 Steps: 103390, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001208, Sample Num: 19328, Cur Loss: 1.24167192, Cur Avg Loss: 1.09374525, Log Avg loss: 1.40605845, Global Avg Loss: 1.44459112, Time: 0.0402 Steps: 103400, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001218, Sample Num: 19488, Cur Loss: 1.78240538, Cur Avg Loss: 1.09365309, Log Avg loss: 1.08252101, Global Avg Loss: 1.44455611, Time: 0.0402 Steps: 103410, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001228, Sample Num: 19648, Cur Loss: 1.24003482, Cur Avg Loss: 1.09247522, Log Avg loss: 0.94900969, Global Avg Loss: 1.44450819, Time: 0.0403 Steps: 103420, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001238, Sample Num: 19808, Cur Loss: 1.08258748, Cur Avg Loss: 1.09332886, Log Avg loss: 1.19815654, Global Avg Loss: 1.44448438, Time: 0.0402 Steps: 103430, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001248, Sample Num: 19968, Cur Loss: 0.46975511, Cur Avg Loss: 1.09393658, Log Avg loss: 1.16917184, Global Avg Loss: 1.44445776, Time: 0.0403 Steps: 103440, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001258, Sample Num: 20128, Cur Loss: 1.12215066, Cur Avg Loss: 1.09601412, Log Avg loss: 1.35529078, Global Avg Loss: 1.44444914, Time: 0.0403 Steps: 103450, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001268, Sample Num: 20288, Cur Loss: 0.38357276, Cur Avg Loss: 1.09571667, Log Avg loss: 1.05829861, Global Avg Loss: 1.44441182, Time: 0.0403 Steps: 103460, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001278, Sample Num: 20448, Cur Loss: 1.48480880, Cur Avg Loss: 1.09817510, Log Avg loss: 1.40990379, Global Avg Loss: 1.44440848, Time: 0.0402 Steps: 103470, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001288, Sample Num: 20608, Cur Loss: 0.97358763, Cur Avg Loss: 1.09905994, Log Avg loss: 1.21214194, Global Avg Loss: 1.44438604, Time: 0.0402 Steps: 103480, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001298, Sample Num: 20768, Cur Loss: 0.60726273, Cur Avg Loss: 1.09742725, Log Avg loss: 0.88713697, Global Avg Loss: 1.44433219, Time: 0.0402 Steps: 103490, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001308, Sample Num: 20928, Cur Loss: 1.01316535, Cur Avg Loss: 1.09783702, Log Avg loss: 1.15102527, Global Avg Loss: 1.44430385, Time: 0.0402 Steps: 103500, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001318, Sample Num: 21088, Cur Loss: 1.71845365, Cur Avg Loss: 1.09878619, Log Avg loss: 1.22293775, Global Avg Loss: 1.44428247, Time: 0.0403 Steps: 103510, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001328, Sample Num: 21248, Cur Loss: 1.55057287, Cur Avg Loss: 1.09943294, Log Avg loss: 1.18467504, Global Avg Loss: 1.44425739, Time: 0.0402 Steps: 103520, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001338, Sample Num: 21408, Cur Loss: 0.44131953, Cur Avg Loss: 1.10021307, Log Avg loss: 1.20381359, Global Avg Loss: 1.44423416, Time: 0.0402 Steps: 103530, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001348, Sample Num: 21568, Cur Loss: 0.98867774, Cur Avg Loss: 1.10114949, Log Avg loss: 1.22644249, Global Avg Loss: 1.44421313, Time: 0.0402 Steps: 103540, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001358, Sample Num: 21728, Cur Loss: 2.09583497, Cur Avg Loss: 1.10151135, Log Avg loss: 1.15029082, Global Avg Loss: 1.44418474, Time: 0.0402 Steps: 103550, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001368, Sample Num: 21888, Cur Loss: 0.56192309, Cur Avg Loss: 1.10081120, Log Avg loss: 1.00572951, Global Avg Loss: 1.44414241, Time: 0.0402 Steps: 103560, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001378, Sample Num: 22048, Cur Loss: 0.41518337, Cur Avg Loss: 1.10182410, Log Avg loss: 1.24038937, Global Avg Loss: 1.44412273, Time: 0.0402 Steps: 103570, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001388, Sample Num: 22208, Cur Loss: 1.39462435, Cur Avg Loss: 1.10097224, Log Avg loss: 0.98358588, Global Avg Loss: 1.44407827, Time: 0.0402 Steps: 103580, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001398, Sample Num: 22368, Cur Loss: 0.46394122, Cur Avg Loss: 1.10066618, Log Avg loss: 1.05818592, Global Avg Loss: 1.44404102, Time: 0.0505 Steps: 103590, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001408, Sample Num: 22528, Cur Loss: 0.78371662, Cur Avg Loss: 1.10393803, Log Avg loss: 1.56134183, Global Avg Loss: 1.44405234, Time: 0.0620 Steps: 103600, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001418, Sample Num: 22688, Cur Loss: 1.60745633, Cur Avg Loss: 1.10351110, Log Avg loss: 1.04339892, Global Avg Loss: 1.44401367, Time: 0.0432 Steps: 103610, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001428, Sample Num: 22848, Cur Loss: 1.45993853, Cur Avg Loss: 1.10281187, Log Avg loss: 1.00366187, Global Avg Loss: 1.44397118, Time: 0.0484 Steps: 103620, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001438, Sample Num: 23008, Cur Loss: 1.71473110, Cur Avg Loss: 1.10644149, Log Avg loss: 1.62475111, Global Avg Loss: 1.44398862, Time: 0.0662 Steps: 103630, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001448, Sample Num: 23168, Cur Loss: 0.57826340, Cur Avg Loss: 1.10452046, Log Avg loss: 0.82827570, Global Avg Loss: 1.44392921, Time: 0.0630 Steps: 103640, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001458, Sample Num: 23328, Cur Loss: 0.33231157, Cur Avg Loss: 1.10224715, Log Avg loss: 0.77307186, Global Avg Loss: 1.44386449, Time: 0.0523 Steps: 103650, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001468, Sample Num: 23488, Cur Loss: 0.26136053, Cur Avg Loss: 1.10074450, Log Avg loss: 0.88165830, Global Avg Loss: 1.44381025, Time: 0.0554 Steps: 103660, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001478, Sample Num: 23648, Cur Loss: 0.46940115, Cur Avg Loss: 1.10088477, Log Avg loss: 1.12147708, Global Avg Loss: 1.44377916, Time: 0.0409 Steps: 103670, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001488, Sample Num: 23808, Cur Loss: 1.73986578, Cur Avg Loss: 1.10191934, Log Avg loss: 1.25482867, Global Avg Loss: 1.44376094, Time: 0.0403 Steps: 103680, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001498, Sample Num: 23968, Cur Loss: 0.96039093, Cur Avg Loss: 1.10174337, Log Avg loss: 1.07555891, Global Avg Loss: 1.44372543, Time: 0.0402 Steps: 103690, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001508, Sample Num: 24128, Cur Loss: 0.43111941, Cur Avg Loss: 1.10047715, Log Avg loss: 0.91079682, Global Avg Loss: 1.44367403, Time: 0.0403 Steps: 103700, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001518, Sample Num: 24288, Cur Loss: 0.60632980, Cur Avg Loss: 1.10024026, Log Avg loss: 1.06451761, Global Avg Loss: 1.44363747, Time: 0.0402 Steps: 103710, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001528, Sample Num: 24448, Cur Loss: 0.45922092, Cur Avg Loss: 1.10183861, Log Avg loss: 1.34446835, Global Avg Loss: 1.44362791, Time: 0.0403 Steps: 103720, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001538, Sample Num: 24608, Cur Loss: 1.32324445, Cur Avg Loss: 1.10106827, Log Avg loss: 0.98335984, Global Avg Loss: 1.44358354, Time: 0.0403 Steps: 103730, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001548, Sample Num: 24768, Cur Loss: 1.11234689, Cur Avg Loss: 1.10044097, Log Avg loss: 1.00396278, Global Avg Loss: 1.44354116, Time: 0.0402 Steps: 103740, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001558, Sample Num: 24928, Cur Loss: 0.43926960, Cur Avg Loss: 1.09930139, Log Avg loss: 0.92289388, Global Avg Loss: 1.44349098, Time: 0.0402 Steps: 103750, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001568, Sample Num: 25088, Cur Loss: 0.57172799, Cur Avg Loss: 1.10173245, Log Avg loss: 1.48049240, Global Avg Loss: 1.44349455, Time: 0.0403 Steps: 103760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001578, Sample Num: 25248, Cur Loss: 1.15577459, Cur Avg Loss: 1.09979736, Log Avg loss: 0.79637505, Global Avg Loss: 1.44343219, Time: 0.0403 Steps: 103770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001588, Sample Num: 25408, Cur Loss: 0.71614766, Cur Avg Loss: 1.09924845, Log Avg loss: 1.01263021, Global Avg Loss: 1.44339068, Time: 0.0403 Steps: 103780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001598, Sample Num: 25568, Cur Loss: 1.27591443, Cur Avg Loss: 1.09811128, Log Avg loss: 0.91752929, Global Avg Loss: 1.44334001, Time: 0.0404 Steps: 103790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001608, Sample Num: 25728, Cur Loss: 1.19363856, Cur Avg Loss: 1.09786150, Log Avg loss: 1.05794557, Global Avg Loss: 1.44330288, Time: 0.0403 Steps: 103800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001618, Sample Num: 25888, Cur Loss: 1.90975332, Cur Avg Loss: 1.09953488, Log Avg loss: 1.36861477, Global Avg Loss: 1.44329569, Time: 0.0403 Steps: 103810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001628, Sample Num: 26048, Cur Loss: 1.28026295, Cur Avg Loss: 1.10146573, Log Avg loss: 1.41387669, Global Avg Loss: 1.44329285, Time: 0.0403 Steps: 103820, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001638, Sample Num: 26208, Cur Loss: 2.39665532, Cur Avg Loss: 1.10185816, Log Avg loss: 1.16574666, Global Avg Loss: 1.44326612, Time: 0.0402 Steps: 103830, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001648, Sample Num: 26368, Cur Loss: 1.61590016, Cur Avg Loss: 1.10231578, Log Avg loss: 1.17727293, Global Avg Loss: 1.44324051, Time: 0.0402 Steps: 103840, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001658, Sample Num: 26528, Cur Loss: 1.28915668, Cur Avg Loss: 1.10353958, Log Avg loss: 1.30522280, Global Avg Loss: 1.44322722, Time: 0.0402 Steps: 103850, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001668, Sample Num: 26688, Cur Loss: 1.29615569, Cur Avg Loss: 1.10210492, Log Avg loss: 0.86423801, Global Avg Loss: 1.44317147, Time: 0.0403 Steps: 103860, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001678, Sample Num: 26848, Cur Loss: 1.39453554, Cur Avg Loss: 1.10064871, Log Avg loss: 0.85775358, Global Avg Loss: 1.44311511, Time: 0.0402 Steps: 103870, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001688, Sample Num: 27008, Cur Loss: 1.77878809, Cur Avg Loss: 1.10328010, Log Avg loss: 1.54482591, Global Avg Loss: 1.44312490, Time: 0.0402 Steps: 103880, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001698, Sample Num: 27168, Cur Loss: 0.56866831, Cur Avg Loss: 1.10229256, Log Avg loss: 0.93559735, Global Avg Loss: 1.44307605, Time: 0.0403 Steps: 103890, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001708, Sample Num: 27328, Cur Loss: 1.00664139, Cur Avg Loss: 1.10141247, Log Avg loss: 0.95197200, Global Avg Loss: 1.44302878, Time: 0.0402 Steps: 103900, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001718, Sample Num: 27488, Cur Loss: 1.34311330, Cur Avg Loss: 1.09975362, Log Avg loss: 0.81642209, Global Avg Loss: 1.44296848, Time: 0.0402 Steps: 103910, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001728, Sample Num: 27648, Cur Loss: 1.47667205, Cur Avg Loss: 1.09964149, Log Avg loss: 1.08037724, Global Avg Loss: 1.44293359, Time: 0.0402 Steps: 103920, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001738, Sample Num: 27808, Cur Loss: 0.89477974, Cur Avg Loss: 1.09976364, Log Avg loss: 1.12087224, Global Avg Loss: 1.44290260, Time: 0.0402 Steps: 103930, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001748, Sample Num: 27968, Cur Loss: 2.24717903, Cur Avg Loss: 1.10181253, Log Avg loss: 1.45790859, Global Avg Loss: 1.44290404, Time: 0.0402 Steps: 103940, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001758, Sample Num: 28128, Cur Loss: 1.00497580, Cur Avg Loss: 1.10265656, Log Avg loss: 1.25019323, Global Avg Loss: 1.44288550, Time: 0.0593 Steps: 103950, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001768, Sample Num: 28288, Cur Loss: 1.17463183, Cur Avg Loss: 1.10375933, Log Avg loss: 1.29762747, Global Avg Loss: 1.44287153, Time: 0.0519 Steps: 103960, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001778, Sample Num: 28448, Cur Loss: 0.40496549, Cur Avg Loss: 1.10145168, Log Avg loss: 0.69345768, Global Avg Loss: 1.44279945, Time: 0.0642 Steps: 103970, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001788, Sample Num: 28608, Cur Loss: 1.17140484, Cur Avg Loss: 1.10129989, Log Avg loss: 1.07431157, Global Avg Loss: 1.44276401, Time: 0.0982 Steps: 103980, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001798, Sample Num: 28768, Cur Loss: 1.27057266, Cur Avg Loss: 1.10070361, Log Avg loss: 0.99409018, Global Avg Loss: 1.44272087, Time: 0.0407 Steps: 103990, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001808, Sample Num: 28928, Cur Loss: 0.58926189, Cur Avg Loss: 1.10188140, Log Avg loss: 1.31364673, Global Avg Loss: 1.44270846, Time: 0.0461 Steps: 104000, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001818, Sample Num: 29088, Cur Loss: 1.08104229, Cur Avg Loss: 1.10324410, Log Avg loss: 1.34962005, Global Avg Loss: 1.44269951, Time: 0.0515 Steps: 104010, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001828, Sample Num: 29248, Cur Loss: 1.13253701, Cur Avg Loss: 1.10190661, Log Avg loss: 0.85875093, Global Avg Loss: 1.44264337, Time: 0.0482 Steps: 104020, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001838, Sample Num: 29408, Cur Loss: 2.80654430, Cur Avg Loss: 1.10208167, Log Avg loss: 1.13408291, Global Avg Loss: 1.44261371, Time: 0.0603 Steps: 104030, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001848, Sample Num: 29568, Cur Loss: 0.25307748, Cur Avg Loss: 1.10033393, Log Avg loss: 0.77910003, Global Avg Loss: 1.44254993, Time: 0.0403 Steps: 104040, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001858, Sample Num: 29728, Cur Loss: 1.14004278, Cur Avg Loss: 1.10124049, Log Avg loss: 1.26877267, Global Avg Loss: 1.44253323, Time: 0.0403 Steps: 104050, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001868, Sample Num: 29888, Cur Loss: 1.05477285, Cur Avg Loss: 1.10250085, Log Avg loss: 1.33667607, Global Avg Loss: 1.44252306, Time: 0.0403 Steps: 104060, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001878, Sample Num: 30048, Cur Loss: 1.84006107, Cur Avg Loss: 1.10395395, Log Avg loss: 1.37539365, Global Avg Loss: 1.44251661, Time: 0.0402 Steps: 104070, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001888, Sample Num: 30208, Cur Loss: 0.49652475, Cur Avg Loss: 1.10441567, Log Avg loss: 1.19112566, Global Avg Loss: 1.44249245, Time: 0.0403 Steps: 104080, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001898, Sample Num: 30368, Cur Loss: 1.39625239, Cur Avg Loss: 1.10432417, Log Avg loss: 1.08704871, Global Avg Loss: 1.44245831, Time: 0.0402 Steps: 104090, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001908, Sample Num: 30528, Cur Loss: 0.92568350, Cur Avg Loss: 1.10376493, Log Avg loss: 0.99762211, Global Avg Loss: 1.44241557, Time: 0.0402 Steps: 104100, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001918, Sample Num: 30688, Cur Loss: 0.58092958, Cur Avg Loss: 1.10207401, Log Avg loss: 0.77944612, Global Avg Loss: 1.44235189, Time: 0.0403 Steps: 104110, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001928, Sample Num: 30848, Cur Loss: 0.52247274, Cur Avg Loss: 1.10113054, Log Avg loss: 0.92017244, Global Avg Loss: 1.44230174, Time: 0.0403 Steps: 104120, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001938, Sample Num: 31008, Cur Loss: 1.00897419, Cur Avg Loss: 1.10215860, Log Avg loss: 1.30036913, Global Avg Loss: 1.44228811, Time: 0.0404 Steps: 104130, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001948, Sample Num: 31168, Cur Loss: 0.61713922, Cur Avg Loss: 1.10119562, Log Avg loss: 0.91457075, Global Avg Loss: 1.44223744, Time: 0.0403 Steps: 104140, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001958, Sample Num: 31328, Cur Loss: 1.69118190, Cur Avg Loss: 1.10185308, Log Avg loss: 1.22992584, Global Avg Loss: 1.44221705, Time: 0.0403 Steps: 104150, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001968, Sample Num: 31488, Cur Loss: 1.91750443, Cur Avg Loss: 1.10207645, Log Avg loss: 1.14581145, Global Avg Loss: 1.44218860, Time: 0.0403 Steps: 104160, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001978, Sample Num: 31648, Cur Loss: 1.71825600, Cur Avg Loss: 1.10210547, Log Avg loss: 1.10781660, Global Avg Loss: 1.44215650, Time: 0.0403 Steps: 104170, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001988, Sample Num: 31808, Cur Loss: 1.12604415, Cur Avg Loss: 1.10193659, Log Avg loss: 1.06853275, Global Avg Loss: 1.44212064, Time: 0.0402 Steps: 104180, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001998, Sample Num: 31968, Cur Loss: 0.79949927, Cur Avg Loss: 1.10103259, Log Avg loss: 0.92131689, Global Avg Loss: 1.44207065, Time: 0.0402 Steps: 104190, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002008, Sample Num: 32128, Cur Loss: 0.59519005, Cur Avg Loss: 1.10088258, Log Avg loss: 1.07091081, Global Avg Loss: 1.44203503, Time: 0.0402 Steps: 104200, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002018, Sample Num: 32288, Cur Loss: 1.19403291, Cur Avg Loss: 1.10225011, Log Avg loss: 1.37685010, Global Avg Loss: 1.44202877, Time: 0.0402 Steps: 104210, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002028, Sample Num: 32448, Cur Loss: 0.67329818, Cur Avg Loss: 1.10286389, Log Avg loss: 1.22672474, Global Avg Loss: 1.44200812, Time: 0.0402 Steps: 104220, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002038, Sample Num: 32608, Cur Loss: 0.88978094, Cur Avg Loss: 1.10270506, Log Avg loss: 1.07049366, Global Avg Loss: 1.44197247, Time: 0.0402 Steps: 104230, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002048, Sample Num: 32768, Cur Loss: 1.58653975, Cur Avg Loss: 1.10285857, Log Avg loss: 1.13414409, Global Avg Loss: 1.44194294, Time: 0.0402 Steps: 104240, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002058, Sample Num: 32928, Cur Loss: 2.05621243, Cur Avg Loss: 1.10346441, Log Avg loss: 1.22754043, Global Avg Loss: 1.44192237, Time: 0.0402 Steps: 104250, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002068, Sample Num: 33088, Cur Loss: 0.53870839, Cur Avg Loss: 1.10334866, Log Avg loss: 1.07952746, Global Avg Loss: 1.44188762, Time: 0.0403 Steps: 104260, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002078, Sample Num: 33248, Cur Loss: 1.09879601, Cur Avg Loss: 1.10243108, Log Avg loss: 0.91267647, Global Avg Loss: 1.44183686, Time: 0.0402 Steps: 104270, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002088, Sample Num: 33408, Cur Loss: 0.93056709, Cur Avg Loss: 1.10239804, Log Avg loss: 1.09553132, Global Avg Loss: 1.44180365, Time: 0.0402 Steps: 104280, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002098, Sample Num: 33568, Cur Loss: 0.34812617, Cur Avg Loss: 1.10118830, Log Avg loss: 0.84859430, Global Avg Loss: 1.44174677, Time: 0.0402 Steps: 104290, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002108, Sample Num: 33728, Cur Loss: 1.16673636, Cur Avg Loss: 1.10169426, Log Avg loss: 1.20784505, Global Avg Loss: 1.44172435, Time: 0.0403 Steps: 104300, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002118, Sample Num: 33888, Cur Loss: 0.98293447, Cur Avg Loss: 1.10082334, Log Avg loss: 0.91723379, Global Avg Loss: 1.44167406, Time: 0.0413 Steps: 104310, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002128, Sample Num: 34048, Cur Loss: 1.42604709, Cur Avg Loss: 1.10107611, Log Avg loss: 1.15461328, Global Avg Loss: 1.44164655, Time: 0.0822 Steps: 104320, Updated lr: 0.000002 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.459944, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.076821, "eval_total_loss": 757.005072, "eval_mae": 0.85399, "eval_mse": 1.076891, "eval_r2": 0.315457, "eval_sp_statistic": 0.489068, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.587699, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.067234, "test_total_loss": 535.751679, "test_mae": 0.861554, "test_mse": 1.067413, "test_r2": 0.311082, "test_sp_statistic": 0.506214, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.63068, "test_ps_pvalue": 0.0, "lr": 2.0189663347558085e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4416361261008164, "train_cur_epoch_loss": 2343.444498077035, "train_cur_epoch_avg_loss": 1.1007254570582596, "train_cur_epoch_time": 95.45994400978088, "train_cur_epoch_avg_time": 0.04483792579134847, "epoch": 49, "step": 104321} ################################################## Training, Epoch: 0050, Batch: 000009, Sample Num: 144, Cur Loss: 0.54565430, Cur Avg Loss: 0.85754300, Log Avg loss: 0.80724180, Global Avg Loss: 1.44158574, Time: 0.0402 Steps: 104330, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000019, Sample Num: 304, Cur Loss: 1.84616315, Cur Avg Loss: 1.01340333, Log Avg loss: 1.15367762, Global Avg Loss: 1.44155815, Time: 0.0625 Steps: 104340, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000029, Sample Num: 464, Cur Loss: 1.18633842, Cur Avg Loss: 0.97305359, Log Avg loss: 0.89638909, Global Avg Loss: 1.44150590, Time: 0.0485 Steps: 104350, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000039, Sample Num: 624, Cur Loss: 1.04701674, Cur Avg Loss: 1.00263855, Log Avg loss: 1.08843494, Global Avg Loss: 1.44147207, Time: 0.0401 Steps: 104360, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000049, Sample Num: 784, Cur Loss: 0.69114041, Cur Avg Loss: 0.93559691, Log Avg loss: 0.67413450, Global Avg Loss: 1.44139855, Time: 0.0540 Steps: 104370, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000059, Sample Num: 944, Cur Loss: 0.61574948, Cur Avg Loss: 0.98087200, Log Avg loss: 1.20271992, Global Avg Loss: 1.44137568, Time: 0.0469 Steps: 104380, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000069, Sample Num: 1104, Cur Loss: 1.05240679, Cur Avg Loss: 0.99111875, Log Avg loss: 1.05157460, Global Avg Loss: 1.44133834, Time: 0.0850 Steps: 104390, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000079, Sample Num: 1264, Cur Loss: 0.49008113, Cur Avg Loss: 1.00367097, Log Avg loss: 1.09028126, Global Avg Loss: 1.44130472, Time: 0.0776 Steps: 104400, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000089, Sample Num: 1424, Cur Loss: 1.03841043, Cur Avg Loss: 1.04319295, Log Avg loss: 1.35541661, Global Avg Loss: 1.44129649, Time: 0.0711 Steps: 104410, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000099, Sample Num: 1584, Cur Loss: 1.22861481, Cur Avg Loss: 1.02637450, Log Avg loss: 0.87669030, Global Avg Loss: 1.44124242, Time: 0.0744 Steps: 104420, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000109, Sample Num: 1744, Cur Loss: 0.94412857, Cur Avg Loss: 1.06285677, Log Avg loss: 1.42403123, Global Avg Loss: 1.44124077, Time: 0.0402 Steps: 104430, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000119, Sample Num: 1904, Cur Loss: 0.93959510, Cur Avg Loss: 1.06943402, Log Avg loss: 1.14112609, Global Avg Loss: 1.44121204, Time: 0.0402 Steps: 104440, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000129, Sample Num: 2064, Cur Loss: 0.50470388, Cur Avg Loss: 1.07488780, Log Avg loss: 1.13978779, Global Avg Loss: 1.44118318, Time: 0.0402 Steps: 104450, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000139, Sample Num: 2224, Cur Loss: 1.43604803, Cur Avg Loss: 1.09912237, Log Avg loss: 1.41174826, Global Avg Loss: 1.44118036, Time: 0.0402 Steps: 104460, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000149, Sample Num: 2384, Cur Loss: 0.88609958, Cur Avg Loss: 1.10688927, Log Avg loss: 1.21484927, Global Avg Loss: 1.44115869, Time: 0.0403 Steps: 104470, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000159, Sample Num: 2544, Cur Loss: 1.03708243, Cur Avg Loss: 1.09285691, Log Avg loss: 0.88377473, Global Avg Loss: 1.44110535, Time: 0.0402 Steps: 104480, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000169, Sample Num: 2704, Cur Loss: 0.97225094, Cur Avg Loss: 1.09342157, Log Avg loss: 1.10239957, Global Avg Loss: 1.44107293, Time: 0.0402 Steps: 104490, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000179, Sample Num: 2864, Cur Loss: 1.24422646, Cur Avg Loss: 1.08541290, Log Avg loss: 0.95006634, Global Avg Loss: 1.44102594, Time: 0.0403 Steps: 104500, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000189, Sample Num: 3024, Cur Loss: 0.70441818, Cur Avg Loss: 1.10167480, Log Avg loss: 1.39276297, Global Avg Loss: 1.44102133, Time: 0.0403 Steps: 104510, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000199, Sample Num: 3184, Cur Loss: 0.62249655, Cur Avg Loss: 1.09555132, Log Avg loss: 0.97981752, Global Avg Loss: 1.44097720, Time: 0.0403 Steps: 104520, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000209, Sample Num: 3344, Cur Loss: 0.40664977, Cur Avg Loss: 1.08328843, Log Avg loss: 0.83925693, Global Avg Loss: 1.44091964, Time: 0.0402 Steps: 104530, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000219, Sample Num: 3504, Cur Loss: 1.68017256, Cur Avg Loss: 1.08378606, Log Avg loss: 1.09418652, Global Avg Loss: 1.44088647, Time: 0.0403 Steps: 104540, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000229, Sample Num: 3664, Cur Loss: 1.12178862, Cur Avg Loss: 1.07878525, Log Avg loss: 0.96926738, Global Avg Loss: 1.44084136, Time: 0.0403 Steps: 104550, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000239, Sample Num: 3824, Cur Loss: 0.73426735, Cur Avg Loss: 1.08739675, Log Avg loss: 1.28460008, Global Avg Loss: 1.44082642, Time: 0.0403 Steps: 104560, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000249, Sample Num: 3984, Cur Loss: 0.99882388, Cur Avg Loss: 1.09510041, Log Avg loss: 1.27921792, Global Avg Loss: 1.44081096, Time: 0.0402 Steps: 104570, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000259, Sample Num: 4144, Cur Loss: 0.32696363, Cur Avg Loss: 1.09203502, Log Avg loss: 1.01570697, Global Avg Loss: 1.44077031, Time: 0.0403 Steps: 104580, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000269, Sample Num: 4304, Cur Loss: 0.69621813, Cur Avg Loss: 1.08685582, Log Avg loss: 0.95271443, Global Avg Loss: 1.44072365, Time: 0.0402 Steps: 104590, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000279, Sample Num: 4464, Cur Loss: 0.54873878, Cur Avg Loss: 1.08286298, Log Avg loss: 0.97545573, Global Avg Loss: 1.44067917, Time: 0.0402 Steps: 104600, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000289, Sample Num: 4624, Cur Loss: 1.09952736, Cur Avg Loss: 1.08072829, Log Avg loss: 1.02117025, Global Avg Loss: 1.44063907, Time: 0.0402 Steps: 104610, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000299, Sample Num: 4784, Cur Loss: 1.95392692, Cur Avg Loss: 1.08418614, Log Avg loss: 1.18411797, Global Avg Loss: 1.44061455, Time: 0.0402 Steps: 104620, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000309, Sample Num: 4944, Cur Loss: 0.66292119, Cur Avg Loss: 1.08202286, Log Avg loss: 1.01734077, Global Avg Loss: 1.44057409, Time: 0.0402 Steps: 104630, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000319, Sample Num: 5104, Cur Loss: 0.77051771, Cur Avg Loss: 1.07593657, Log Avg loss: 0.88787028, Global Avg Loss: 1.44052127, Time: 0.0402 Steps: 104640, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000329, Sample Num: 5264, Cur Loss: 1.07988334, Cur Avg Loss: 1.07612436, Log Avg loss: 1.08211488, Global Avg Loss: 1.44048703, Time: 0.0403 Steps: 104650, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000339, Sample Num: 5424, Cur Loss: 1.52557480, Cur Avg Loss: 1.08140549, Log Avg loss: 1.25515453, Global Avg Loss: 1.44046932, Time: 0.0402 Steps: 104660, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000349, Sample Num: 5584, Cur Loss: 1.31573331, Cur Avg Loss: 1.10044337, Log Avg loss: 1.74582763, Global Avg Loss: 1.44049849, Time: 0.0402 Steps: 104670, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000359, Sample Num: 5744, Cur Loss: 0.51752770, Cur Avg Loss: 1.09681012, Log Avg loss: 0.97000954, Global Avg Loss: 1.44045355, Time: 0.0402 Steps: 104680, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000369, Sample Num: 5904, Cur Loss: 0.63659245, Cur Avg Loss: 1.09281613, Log Avg loss: 0.94943209, Global Avg Loss: 1.44040664, Time: 0.0416 Steps: 104690, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000379, Sample Num: 6064, Cur Loss: 1.14106739, Cur Avg Loss: 1.09527094, Log Avg loss: 1.18585328, Global Avg Loss: 1.44038233, Time: 0.0404 Steps: 104700, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000389, Sample Num: 6224, Cur Loss: 2.44932008, Cur Avg Loss: 1.09433180, Log Avg loss: 1.05873838, Global Avg Loss: 1.44034588, Time: 0.0995 Steps: 104710, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000399, Sample Num: 6384, Cur Loss: 1.40103734, Cur Avg Loss: 1.09267215, Log Avg loss: 1.02811196, Global Avg Loss: 1.44030652, Time: 0.0455 Steps: 104720, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000409, Sample Num: 6544, Cur Loss: 1.71754849, Cur Avg Loss: 1.09623646, Log Avg loss: 1.23845246, Global Avg Loss: 1.44028724, Time: 0.0794 Steps: 104730, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000419, Sample Num: 6704, Cur Loss: 0.56574374, Cur Avg Loss: 1.09449512, Log Avg loss: 1.02327405, Global Avg Loss: 1.44024743, Time: 0.0404 Steps: 104740, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000429, Sample Num: 6864, Cur Loss: 0.58071017, Cur Avg Loss: 1.09178771, Log Avg loss: 0.97834737, Global Avg Loss: 1.44020333, Time: 0.0854 Steps: 104750, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000439, Sample Num: 7024, Cur Loss: 1.77707124, Cur Avg Loss: 1.09776946, Log Avg loss: 1.35438634, Global Avg Loss: 1.44019514, Time: 0.0455 Steps: 104760, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000449, Sample Num: 7184, Cur Loss: 1.66252518, Cur Avg Loss: 1.10017492, Log Avg loss: 1.20577489, Global Avg Loss: 1.44017277, Time: 0.0996 Steps: 104770, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000459, Sample Num: 7344, Cur Loss: 0.63859433, Cur Avg Loss: 1.10083012, Log Avg loss: 1.13024835, Global Avg Loss: 1.44014319, Time: 0.0402 Steps: 104780, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000469, Sample Num: 7504, Cur Loss: 0.84194118, Cur Avg Loss: 1.10582490, Log Avg loss: 1.33508539, Global Avg Loss: 1.44013316, Time: 0.0403 Steps: 104790, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000479, Sample Num: 7664, Cur Loss: 0.92770559, Cur Avg Loss: 1.10814569, Log Avg loss: 1.21699066, Global Avg Loss: 1.44011187, Time: 0.0403 Steps: 104800, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000489, Sample Num: 7824, Cur Loss: 1.83759546, Cur Avg Loss: 1.10643670, Log Avg loss: 1.02457621, Global Avg Loss: 1.44007222, Time: 0.0402 Steps: 104810, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000499, Sample Num: 7984, Cur Loss: 0.47904199, Cur Avg Loss: 1.10549710, Log Avg loss: 1.05955094, Global Avg Loss: 1.44003592, Time: 0.0402 Steps: 104820, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000509, Sample Num: 8144, Cur Loss: 1.00107884, Cur Avg Loss: 1.10344408, Log Avg loss: 1.00099835, Global Avg Loss: 1.43999404, Time: 0.0403 Steps: 104830, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000519, Sample Num: 8304, Cur Loss: 1.00510037, Cur Avg Loss: 1.09994310, Log Avg loss: 0.92174300, Global Avg Loss: 1.43994461, Time: 0.0403 Steps: 104840, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000529, Sample Num: 8464, Cur Loss: 0.26321644, Cur Avg Loss: 1.10264370, Log Avg loss: 1.24280508, Global Avg Loss: 1.43992581, Time: 0.0403 Steps: 104850, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000539, Sample Num: 8624, Cur Loss: 1.22671556, Cur Avg Loss: 1.10192608, Log Avg loss: 1.06396357, Global Avg Loss: 1.43988995, Time: 0.0402 Steps: 104860, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000549, Sample Num: 8784, Cur Loss: 0.74641132, Cur Avg Loss: 1.10321922, Log Avg loss: 1.17291943, Global Avg Loss: 1.43986450, Time: 0.0402 Steps: 104870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000559, Sample Num: 8944, Cur Loss: 1.76379204, Cur Avg Loss: 1.10196429, Log Avg loss: 1.03306899, Global Avg Loss: 1.43982571, Time: 0.0403 Steps: 104880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000569, Sample Num: 9104, Cur Loss: 1.31548429, Cur Avg Loss: 1.09633573, Log Avg loss: 0.78169930, Global Avg Loss: 1.43976296, Time: 0.0404 Steps: 104890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000579, Sample Num: 9264, Cur Loss: 0.65067339, Cur Avg Loss: 1.09462600, Log Avg loss: 0.99734235, Global Avg Loss: 1.43972079, Time: 0.0402 Steps: 104900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000589, Sample Num: 9424, Cur Loss: 1.45768142, Cur Avg Loss: 1.09148479, Log Avg loss: 0.90960847, Global Avg Loss: 1.43967026, Time: 0.0402 Steps: 104910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000599, Sample Num: 9584, Cur Loss: 0.73908943, Cur Avg Loss: 1.08964459, Log Avg loss: 0.98125685, Global Avg Loss: 1.43962657, Time: 0.0402 Steps: 104920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000609, Sample Num: 9744, Cur Loss: 0.35175014, Cur Avg Loss: 1.09020611, Log Avg loss: 1.12384093, Global Avg Loss: 1.43959647, Time: 0.0402 Steps: 104930, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000619, Sample Num: 9904, Cur Loss: 1.14038861, Cur Avg Loss: 1.09338252, Log Avg loss: 1.28682621, Global Avg Loss: 1.43958191, Time: 0.0402 Steps: 104940, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000629, Sample Num: 10064, Cur Loss: 0.80211091, Cur Avg Loss: 1.08713980, Log Avg loss: 0.70071544, Global Avg Loss: 1.43951151, Time: 0.0402 Steps: 104950, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000639, Sample Num: 10224, Cur Loss: 1.95807815, Cur Avg Loss: 1.09088650, Log Avg loss: 1.32655402, Global Avg Loss: 1.43950075, Time: 0.0402 Steps: 104960, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000649, Sample Num: 10384, Cur Loss: 0.29518309, Cur Avg Loss: 1.09172867, Log Avg loss: 1.14554326, Global Avg Loss: 1.43947275, Time: 0.0402 Steps: 104970, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000659, Sample Num: 10544, Cur Loss: 1.23442113, Cur Avg Loss: 1.09303883, Log Avg loss: 1.17806790, Global Avg Loss: 1.43944785, Time: 0.0402 Steps: 104980, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000669, Sample Num: 10704, Cur Loss: 0.82091677, Cur Avg Loss: 1.09060960, Log Avg loss: 0.93052328, Global Avg Loss: 1.43939937, Time: 0.0403 Steps: 104990, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000679, Sample Num: 10864, Cur Loss: 0.55077678, Cur Avg Loss: 1.08824730, Log Avg loss: 0.93020962, Global Avg Loss: 1.43935088, Time: 0.0402 Steps: 105000, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000689, Sample Num: 11024, Cur Loss: 0.93489337, Cur Avg Loss: 1.09104523, Log Avg loss: 1.28102501, Global Avg Loss: 1.43933580, Time: 0.0402 Steps: 105010, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000699, Sample Num: 11184, Cur Loss: 1.13526940, Cur Avg Loss: 1.08607397, Log Avg loss: 0.74355415, Global Avg Loss: 1.43926955, Time: 0.0403 Steps: 105020, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000709, Sample Num: 11344, Cur Loss: 1.00726902, Cur Avg Loss: 1.08288736, Log Avg loss: 0.86014316, Global Avg Loss: 1.43921441, Time: 0.0403 Steps: 105030, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000719, Sample Num: 11504, Cur Loss: 0.86130613, Cur Avg Loss: 1.08203634, Log Avg loss: 1.02169927, Global Avg Loss: 1.43917466, Time: 0.0455 Steps: 105040, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000729, Sample Num: 11664, Cur Loss: 1.14448714, Cur Avg Loss: 1.08406590, Log Avg loss: 1.22999129, Global Avg Loss: 1.43915475, Time: 0.0406 Steps: 105050, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000739, Sample Num: 11824, Cur Loss: 2.18423939, Cur Avg Loss: 1.08673945, Log Avg loss: 1.28164108, Global Avg Loss: 1.43913976, Time: 0.0431 Steps: 105060, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000749, Sample Num: 11984, Cur Loss: 2.02836871, Cur Avg Loss: 1.08720441, Log Avg loss: 1.12156500, Global Avg Loss: 1.43910953, Time: 0.0403 Steps: 105070, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000759, Sample Num: 12144, Cur Loss: 1.43724334, Cur Avg Loss: 1.08406897, Log Avg loss: 0.84922429, Global Avg Loss: 1.43905339, Time: 0.0748 Steps: 105080, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000769, Sample Num: 12304, Cur Loss: 0.37013531, Cur Avg Loss: 1.08448559, Log Avg loss: 1.11610678, Global Avg Loss: 1.43902266, Time: 0.0404 Steps: 105090, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000779, Sample Num: 12464, Cur Loss: 0.77026320, Cur Avg Loss: 1.08638603, Log Avg loss: 1.23252978, Global Avg Loss: 1.43900302, Time: 0.0405 Steps: 105100, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000789, Sample Num: 12624, Cur Loss: 0.49796599, Cur Avg Loss: 1.08451707, Log Avg loss: 0.93892525, Global Avg Loss: 1.43895544, Time: 0.0503 Steps: 105110, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000799, Sample Num: 12784, Cur Loss: 0.68934333, Cur Avg Loss: 1.08215621, Log Avg loss: 0.89588464, Global Avg Loss: 1.43890378, Time: 0.0501 Steps: 105120, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000809, Sample Num: 12944, Cur Loss: 1.28552401, Cur Avg Loss: 1.08513493, Log Avg loss: 1.32313496, Global Avg Loss: 1.43889277, Time: 0.0403 Steps: 105130, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000819, Sample Num: 13104, Cur Loss: 0.78043073, Cur Avg Loss: 1.08657531, Log Avg loss: 1.20310131, Global Avg Loss: 1.43887034, Time: 0.0402 Steps: 105140, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000829, Sample Num: 13264, Cur Loss: 0.94560564, Cur Avg Loss: 1.08954668, Log Avg loss: 1.33290223, Global Avg Loss: 1.43886026, Time: 0.0402 Steps: 105150, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000839, Sample Num: 13424, Cur Loss: 0.45029995, Cur Avg Loss: 1.09268113, Log Avg loss: 1.35252691, Global Avg Loss: 1.43885205, Time: 0.0402 Steps: 105160, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000849, Sample Num: 13584, Cur Loss: 2.87529612, Cur Avg Loss: 1.09280810, Log Avg loss: 1.10346128, Global Avg Loss: 1.43882016, Time: 0.0403 Steps: 105170, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000859, Sample Num: 13744, Cur Loss: 0.92066938, Cur Avg Loss: 1.09137380, Log Avg loss: 0.96960118, Global Avg Loss: 1.43877555, Time: 0.0402 Steps: 105180, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000869, Sample Num: 13904, Cur Loss: 0.54681039, Cur Avg Loss: 1.09002707, Log Avg loss: 0.97434314, Global Avg Loss: 1.43873140, Time: 0.0402 Steps: 105190, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000879, Sample Num: 14064, Cur Loss: 0.47979546, Cur Avg Loss: 1.09016674, Log Avg loss: 1.10230454, Global Avg Loss: 1.43869942, Time: 0.0402 Steps: 105200, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000889, Sample Num: 14224, Cur Loss: 0.87943006, Cur Avg Loss: 1.08774258, Log Avg loss: 0.87465888, Global Avg Loss: 1.43864581, Time: 0.0402 Steps: 105210, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000899, Sample Num: 14384, Cur Loss: 1.23127937, Cur Avg Loss: 1.08742072, Log Avg loss: 1.05880738, Global Avg Loss: 1.43860971, Time: 0.0402 Steps: 105220, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000909, Sample Num: 14544, Cur Loss: 0.52653241, Cur Avg Loss: 1.08731228, Log Avg loss: 1.07756318, Global Avg Loss: 1.43857540, Time: 0.0403 Steps: 105230, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000919, Sample Num: 14704, Cur Loss: 1.92426515, Cur Avg Loss: 1.08552028, Log Avg loss: 0.92262712, Global Avg Loss: 1.43852637, Time: 0.0402 Steps: 105240, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000929, Sample Num: 14864, Cur Loss: 0.92411190, Cur Avg Loss: 1.08459540, Log Avg loss: 0.99959929, Global Avg Loss: 1.43848467, Time: 0.0404 Steps: 105250, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000939, Sample Num: 15024, Cur Loss: 0.63468182, Cur Avg Loss: 1.08434628, Log Avg loss: 1.06120308, Global Avg Loss: 1.43844883, Time: 0.0402 Steps: 105260, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000949, Sample Num: 15184, Cur Loss: 1.81865358, Cur Avg Loss: 1.08722686, Log Avg loss: 1.35771365, Global Avg Loss: 1.43844116, Time: 0.0403 Steps: 105270, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000959, Sample Num: 15344, Cur Loss: 0.70026040, Cur Avg Loss: 1.08927355, Log Avg loss: 1.28350445, Global Avg Loss: 1.43842644, Time: 0.0402 Steps: 105280, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000969, Sample Num: 15504, Cur Loss: 1.33217847, Cur Avg Loss: 1.08546806, Log Avg loss: 0.72052073, Global Avg Loss: 1.43835826, Time: 0.0402 Steps: 105290, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000979, Sample Num: 15664, Cur Loss: 1.06714499, Cur Avg Loss: 1.08668235, Log Avg loss: 1.20434795, Global Avg Loss: 1.43833603, Time: 0.0402 Steps: 105300, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000989, Sample Num: 15824, Cur Loss: 1.05656600, Cur Avg Loss: 1.08773514, Log Avg loss: 1.19080308, Global Avg Loss: 1.43831253, Time: 0.0402 Steps: 105310, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000999, Sample Num: 15984, Cur Loss: 1.62248230, Cur Avg Loss: 1.08600183, Log Avg loss: 0.91457691, Global Avg Loss: 1.43826280, Time: 0.0402 Steps: 105320, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001009, Sample Num: 16144, Cur Loss: 1.74555182, Cur Avg Loss: 1.08583620, Log Avg loss: 1.06928960, Global Avg Loss: 1.43822777, Time: 0.0403 Steps: 105330, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001019, Sample Num: 16304, Cur Loss: 0.80115241, Cur Avg Loss: 1.08408406, Log Avg loss: 0.90729411, Global Avg Loss: 1.43817737, Time: 0.0402 Steps: 105340, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001029, Sample Num: 16464, Cur Loss: 0.78679729, Cur Avg Loss: 1.08705862, Log Avg loss: 1.39016595, Global Avg Loss: 1.43817281, Time: 0.0402 Steps: 105350, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001039, Sample Num: 16624, Cur Loss: 0.98975742, Cur Avg Loss: 1.08873009, Log Avg loss: 1.26072437, Global Avg Loss: 1.43815597, Time: 0.0402 Steps: 105360, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001049, Sample Num: 16784, Cur Loss: 0.45808467, Cur Avg Loss: 1.08898951, Log Avg loss: 1.11594308, Global Avg Loss: 1.43812539, Time: 0.0403 Steps: 105370, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001059, Sample Num: 16944, Cur Loss: 0.62489069, Cur Avg Loss: 1.08677985, Log Avg loss: 0.85498680, Global Avg Loss: 1.43807005, Time: 0.0402 Steps: 105380, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001069, Sample Num: 17104, Cur Loss: 1.23027849, Cur Avg Loss: 1.08651663, Log Avg loss: 1.05864146, Global Avg Loss: 1.43803405, Time: 0.0406 Steps: 105390, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001079, Sample Num: 17264, Cur Loss: 0.49684125, Cur Avg Loss: 1.08591381, Log Avg loss: 1.02147207, Global Avg Loss: 1.43799453, Time: 0.0571 Steps: 105400, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001089, Sample Num: 17424, Cur Loss: 0.39191985, Cur Avg Loss: 1.08458091, Log Avg loss: 0.94076092, Global Avg Loss: 1.43794736, Time: 0.0588 Steps: 105410, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001099, Sample Num: 17584, Cur Loss: 1.26145673, Cur Avg Loss: 1.08604011, Log Avg loss: 1.24494772, Global Avg Loss: 1.43792905, Time: 0.0660 Steps: 105420, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001109, Sample Num: 17744, Cur Loss: 0.59826005, Cur Avg Loss: 1.08583936, Log Avg loss: 1.06377699, Global Avg Loss: 1.43789356, Time: 0.0437 Steps: 105430, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001119, Sample Num: 17904, Cur Loss: 0.88729382, Cur Avg Loss: 1.08605054, Log Avg loss: 1.10947042, Global Avg Loss: 1.43786241, Time: 0.0405 Steps: 105440, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001129, Sample Num: 18064, Cur Loss: 1.00584543, Cur Avg Loss: 1.08534787, Log Avg loss: 1.00671850, Global Avg Loss: 1.43782153, Time: 0.0410 Steps: 105450, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001139, Sample Num: 18224, Cur Loss: 0.94827586, Cur Avg Loss: 1.08518518, Log Avg loss: 1.06681759, Global Avg Loss: 1.43778635, Time: 0.0404 Steps: 105460, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001149, Sample Num: 18384, Cur Loss: 1.57265282, Cur Avg Loss: 1.08589997, Log Avg loss: 1.16731460, Global Avg Loss: 1.43776070, Time: 0.0976 Steps: 105470, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001159, Sample Num: 18544, Cur Loss: 1.60346484, Cur Avg Loss: 1.08634392, Log Avg loss: 1.13735381, Global Avg Loss: 1.43773222, Time: 0.0402 Steps: 105480, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001169, Sample Num: 18704, Cur Loss: 0.89329469, Cur Avg Loss: 1.08513685, Log Avg loss: 0.94523690, Global Avg Loss: 1.43768554, Time: 0.0402 Steps: 105490, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001179, Sample Num: 18864, Cur Loss: 0.74312812, Cur Avg Loss: 1.08395134, Log Avg loss: 0.94536557, Global Avg Loss: 1.43763887, Time: 0.0403 Steps: 105500, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001189, Sample Num: 19024, Cur Loss: 1.78726256, Cur Avg Loss: 1.08504099, Log Avg loss: 1.21351097, Global Avg Loss: 1.43761763, Time: 0.0403 Steps: 105510, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001199, Sample Num: 19184, Cur Loss: 1.55777192, Cur Avg Loss: 1.08405929, Log Avg loss: 0.96733502, Global Avg Loss: 1.43757306, Time: 0.0402 Steps: 105520, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001209, Sample Num: 19344, Cur Loss: 0.72894460, Cur Avg Loss: 1.08622953, Log Avg loss: 1.34644076, Global Avg Loss: 1.43756443, Time: 0.0402 Steps: 105530, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001219, Sample Num: 19504, Cur Loss: 0.67206919, Cur Avg Loss: 1.08605348, Log Avg loss: 1.06476957, Global Avg Loss: 1.43752910, Time: 0.0404 Steps: 105540, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001229, Sample Num: 19664, Cur Loss: 0.74311709, Cur Avg Loss: 1.08667548, Log Avg loss: 1.16249703, Global Avg Loss: 1.43750305, Time: 0.0404 Steps: 105550, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001239, Sample Num: 19824, Cur Loss: 0.98550427, Cur Avg Loss: 1.08804223, Log Avg loss: 1.25601541, Global Avg Loss: 1.43748585, Time: 0.0403 Steps: 105560, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001249, Sample Num: 19984, Cur Loss: 2.21824098, Cur Avg Loss: 1.08942478, Log Avg loss: 1.26072362, Global Avg Loss: 1.43746911, Time: 0.0403 Steps: 105570, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001259, Sample Num: 20144, Cur Loss: 1.64542413, Cur Avg Loss: 1.08867645, Log Avg loss: 0.99520935, Global Avg Loss: 1.43742722, Time: 0.0403 Steps: 105580, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001269, Sample Num: 20304, Cur Loss: 2.12870741, Cur Avg Loss: 1.09132571, Log Avg loss: 1.42486778, Global Avg Loss: 1.43742603, Time: 0.0402 Steps: 105590, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001279, Sample Num: 20464, Cur Loss: 0.73415554, Cur Avg Loss: 1.09086729, Log Avg loss: 1.03269424, Global Avg Loss: 1.43738770, Time: 0.0402 Steps: 105600, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001289, Sample Num: 20624, Cur Loss: 0.91850895, Cur Avg Loss: 1.09134993, Log Avg loss: 1.15307975, Global Avg Loss: 1.43736078, Time: 0.0403 Steps: 105610, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001299, Sample Num: 20784, Cur Loss: 0.87344265, Cur Avg Loss: 1.09041097, Log Avg loss: 0.96937898, Global Avg Loss: 1.43731648, Time: 0.0402 Steps: 105620, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001309, Sample Num: 20944, Cur Loss: 1.45886302, Cur Avg Loss: 1.09007793, Log Avg loss: 1.04681589, Global Avg Loss: 1.43727951, Time: 0.0403 Steps: 105630, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001319, Sample Num: 21104, Cur Loss: 0.90031552, Cur Avg Loss: 1.08826141, Log Avg loss: 0.85047810, Global Avg Loss: 1.43722396, Time: 0.0402 Steps: 105640, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001329, Sample Num: 21264, Cur Loss: 0.60961163, Cur Avg Loss: 1.08926120, Log Avg loss: 1.22113387, Global Avg Loss: 1.43720351, Time: 0.0403 Steps: 105650, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001339, Sample Num: 21424, Cur Loss: 2.62426066, Cur Avg Loss: 1.09400778, Log Avg loss: 1.72482786, Global Avg Loss: 1.43723073, Time: 0.0403 Steps: 105660, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001349, Sample Num: 21584, Cur Loss: 0.42556590, Cur Avg Loss: 1.09585760, Log Avg loss: 1.34354970, Global Avg Loss: 1.43722186, Time: 0.0403 Steps: 105670, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001359, Sample Num: 21744, Cur Loss: 2.37350702, Cur Avg Loss: 1.09508480, Log Avg loss: 0.99083347, Global Avg Loss: 1.43717962, Time: 0.0402 Steps: 105680, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001369, Sample Num: 21904, Cur Loss: 0.70491540, Cur Avg Loss: 1.09289872, Log Avg loss: 0.79581045, Global Avg Loss: 1.43711894, Time: 0.0403 Steps: 105690, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001379, Sample Num: 22064, Cur Loss: 0.59343207, Cur Avg Loss: 1.09287218, Log Avg loss: 1.08923834, Global Avg Loss: 1.43708603, Time: 0.0402 Steps: 105700, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001389, Sample Num: 22224, Cur Loss: 1.23584270, Cur Avg Loss: 1.09251999, Log Avg loss: 1.04395396, Global Avg Loss: 1.43704884, Time: 0.0403 Steps: 105710, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001399, Sample Num: 22384, Cur Loss: 1.80482411, Cur Avg Loss: 1.09370811, Log Avg loss: 1.25873713, Global Avg Loss: 1.43703197, Time: 0.0403 Steps: 105720, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001409, Sample Num: 22544, Cur Loss: 1.82454038, Cur Avg Loss: 1.09409474, Log Avg loss: 1.14818408, Global Avg Loss: 1.43700465, Time: 0.0456 Steps: 105730, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001419, Sample Num: 22704, Cur Loss: 2.01784039, Cur Avg Loss: 1.09683266, Log Avg loss: 1.48260644, Global Avg Loss: 1.43700896, Time: 0.0573 Steps: 105740, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001429, Sample Num: 22864, Cur Loss: 0.65903717, Cur Avg Loss: 1.09583396, Log Avg loss: 0.95411834, Global Avg Loss: 1.43696330, Time: 0.0447 Steps: 105750, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001439, Sample Num: 23024, Cur Loss: 1.05429351, Cur Avg Loss: 1.09496465, Log Avg loss: 0.97073991, Global Avg Loss: 1.43691922, Time: 0.0875 Steps: 105760, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001449, Sample Num: 23184, Cur Loss: 0.92664313, Cur Avg Loss: 1.09358129, Log Avg loss: 0.89451612, Global Avg Loss: 1.43686794, Time: 0.0518 Steps: 105770, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001459, Sample Num: 23344, Cur Loss: 0.75931108, Cur Avg Loss: 1.09205821, Log Avg loss: 0.87136401, Global Avg Loss: 1.43681448, Time: 0.0533 Steps: 105780, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001469, Sample Num: 23504, Cur Loss: 0.43091646, Cur Avg Loss: 1.09195559, Log Avg loss: 1.07698330, Global Avg Loss: 1.43678046, Time: 0.0471 Steps: 105790, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001479, Sample Num: 23664, Cur Loss: 1.98736143, Cur Avg Loss: 1.09153407, Log Avg loss: 1.02961330, Global Avg Loss: 1.43674198, Time: 0.0533 Steps: 105800, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001489, Sample Num: 23824, Cur Loss: 1.64406872, Cur Avg Loss: 1.09233998, Log Avg loss: 1.21153376, Global Avg Loss: 1.43672069, Time: 0.0805 Steps: 105810, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001499, Sample Num: 23984, Cur Loss: 1.24282002, Cur Avg Loss: 1.09217401, Log Avg loss: 1.06746007, Global Avg Loss: 1.43668580, Time: 0.0403 Steps: 105820, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001509, Sample Num: 24144, Cur Loss: 0.45319676, Cur Avg Loss: 1.09169620, Log Avg loss: 1.02007339, Global Avg Loss: 1.43664643, Time: 0.0402 Steps: 105830, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001519, Sample Num: 24304, Cur Loss: 1.47439194, Cur Avg Loss: 1.09097604, Log Avg loss: 0.98230386, Global Avg Loss: 1.43660350, Time: 0.0402 Steps: 105840, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001529, Sample Num: 24464, Cur Loss: 0.24483491, Cur Avg Loss: 1.09305479, Log Avg loss: 1.40881624, Global Avg Loss: 1.43660088, Time: 0.0403 Steps: 105850, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001539, Sample Num: 24624, Cur Loss: 0.81472683, Cur Avg Loss: 1.09396153, Log Avg loss: 1.23260251, Global Avg Loss: 1.43658161, Time: 0.0402 Steps: 105860, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001549, Sample Num: 24784, Cur Loss: 0.44991687, Cur Avg Loss: 1.09632597, Log Avg loss: 1.46021253, Global Avg Loss: 1.43658384, Time: 0.0403 Steps: 105870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001559, Sample Num: 24944, Cur Loss: 1.71462011, Cur Avg Loss: 1.09800217, Log Avg loss: 1.35764675, Global Avg Loss: 1.43657639, Time: 0.0402 Steps: 105880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001569, Sample Num: 25104, Cur Loss: 2.05545330, Cur Avg Loss: 1.09842984, Log Avg loss: 1.16510365, Global Avg Loss: 1.43655075, Time: 0.0403 Steps: 105890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001579, Sample Num: 25264, Cur Loss: 0.44066924, Cur Avg Loss: 1.09714613, Log Avg loss: 0.89573132, Global Avg Loss: 1.43649968, Time: 0.0402 Steps: 105900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001589, Sample Num: 25424, Cur Loss: 0.30318671, Cur Avg Loss: 1.09707291, Log Avg loss: 1.08551137, Global Avg Loss: 1.43646654, Time: 0.0403 Steps: 105910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001599, Sample Num: 25584, Cur Loss: 1.10330558, Cur Avg Loss: 1.09695776, Log Avg loss: 1.07866141, Global Avg Loss: 1.43643276, Time: 0.0403 Steps: 105920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001609, Sample Num: 25744, Cur Loss: 0.22036399, Cur Avg Loss: 1.09525017, Log Avg loss: 0.82220593, Global Avg Loss: 1.43637477, Time: 0.0403 Steps: 105930, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001619, Sample Num: 25904, Cur Loss: 0.85670507, Cur Avg Loss: 1.09508006, Log Avg loss: 1.06770883, Global Avg Loss: 1.43633997, Time: 0.0402 Steps: 105940, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001629, Sample Num: 26064, Cur Loss: 0.79854363, Cur Avg Loss: 1.09411263, Log Avg loss: 0.93748560, Global Avg Loss: 1.43629289, Time: 0.0403 Steps: 105950, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001639, Sample Num: 26224, Cur Loss: 0.43510634, Cur Avg Loss: 1.09419145, Log Avg loss: 1.10703158, Global Avg Loss: 1.43626182, Time: 0.0403 Steps: 105960, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001649, Sample Num: 26384, Cur Loss: 1.88175583, Cur Avg Loss: 1.09518649, Log Avg loss: 1.25827329, Global Avg Loss: 1.43624502, Time: 0.0402 Steps: 105970, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001659, Sample Num: 26544, Cur Loss: 0.65318984, Cur Avg Loss: 1.09482096, Log Avg loss: 1.03454525, Global Avg Loss: 1.43620712, Time: 0.0402 Steps: 105980, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001669, Sample Num: 26704, Cur Loss: 1.65608084, Cur Avg Loss: 1.09645971, Log Avg loss: 1.36832858, Global Avg Loss: 1.43620071, Time: 0.0403 Steps: 105990, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001679, Sample Num: 26864, Cur Loss: 0.41315034, Cur Avg Loss: 1.09564118, Log Avg loss: 0.95902801, Global Avg Loss: 1.43615570, Time: 0.0402 Steps: 106000, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001689, Sample Num: 27024, Cur Loss: 0.80695981, Cur Avg Loss: 1.09535807, Log Avg loss: 1.04782486, Global Avg Loss: 1.43611907, Time: 0.0402 Steps: 106010, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001699, Sample Num: 27184, Cur Loss: 0.93157005, Cur Avg Loss: 1.09439449, Log Avg loss: 0.93164516, Global Avg Loss: 1.43607148, Time: 0.0402 Steps: 106020, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001709, Sample Num: 27344, Cur Loss: 0.75978976, Cur Avg Loss: 1.09337932, Log Avg loss: 0.92090285, Global Avg Loss: 1.43602290, Time: 0.0402 Steps: 106030, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001719, Sample Num: 27504, Cur Loss: 0.77553999, Cur Avg Loss: 1.09331570, Log Avg loss: 1.08244310, Global Avg Loss: 1.43598955, Time: 0.0402 Steps: 106040, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001729, Sample Num: 27664, Cur Loss: 0.90954494, Cur Avg Loss: 1.09346615, Log Avg loss: 1.11932789, Global Avg Loss: 1.43595969, Time: 0.0402 Steps: 106050, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001739, Sample Num: 27824, Cur Loss: 0.76419961, Cur Avg Loss: 1.09441249, Log Avg loss: 1.25803441, Global Avg Loss: 1.43594292, Time: 0.0402 Steps: 106060, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001749, Sample Num: 27984, Cur Loss: 1.06142187, Cur Avg Loss: 1.09377673, Log Avg loss: 0.98321809, Global Avg Loss: 1.43590023, Time: 0.0746 Steps: 106070, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001759, Sample Num: 28144, Cur Loss: 0.90759099, Cur Avg Loss: 1.09337631, Log Avg loss: 1.02334334, Global Avg Loss: 1.43586134, Time: 0.0456 Steps: 106080, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001769, Sample Num: 28304, Cur Loss: 1.42132986, Cur Avg Loss: 1.09375476, Log Avg loss: 1.16032324, Global Avg Loss: 1.43583537, Time: 0.0670 Steps: 106090, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001779, Sample Num: 28464, Cur Loss: 1.18011403, Cur Avg Loss: 1.09418680, Log Avg loss: 1.17061583, Global Avg Loss: 1.43581037, Time: 0.0408 Steps: 106100, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001789, Sample Num: 28624, Cur Loss: 0.71375746, Cur Avg Loss: 1.09432915, Log Avg loss: 1.11965167, Global Avg Loss: 1.43578058, Time: 0.0617 Steps: 106110, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001799, Sample Num: 28784, Cur Loss: 0.82357693, Cur Avg Loss: 1.09342273, Log Avg loss: 0.93126585, Global Avg Loss: 1.43573304, Time: 0.0497 Steps: 106120, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001809, Sample Num: 28944, Cur Loss: 0.78311235, Cur Avg Loss: 1.09315285, Log Avg loss: 1.04460034, Global Avg Loss: 1.43569618, Time: 0.0403 Steps: 106130, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001819, Sample Num: 29104, Cur Loss: 1.26902437, Cur Avg Loss: 1.09215722, Log Avg loss: 0.91204863, Global Avg Loss: 1.43564685, Time: 0.0405 Steps: 106140, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001829, Sample Num: 29264, Cur Loss: 1.12480640, Cur Avg Loss: 1.09204135, Log Avg loss: 1.07096475, Global Avg Loss: 1.43561249, Time: 0.0403 Steps: 106150, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001839, Sample Num: 29424, Cur Loss: 0.59922481, Cur Avg Loss: 1.09074212, Log Avg loss: 0.85311286, Global Avg Loss: 1.43555762, Time: 0.0403 Steps: 106160, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001849, Sample Num: 29584, Cur Loss: 0.89220941, Cur Avg Loss: 1.08971176, Log Avg loss: 0.90022767, Global Avg Loss: 1.43550720, Time: 0.0405 Steps: 106170, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001859, Sample Num: 29744, Cur Loss: 0.92659646, Cur Avg Loss: 1.08906540, Log Avg loss: 0.96955432, Global Avg Loss: 1.43546332, Time: 0.0402 Steps: 106180, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001869, Sample Num: 29904, Cur Loss: 1.58592749, Cur Avg Loss: 1.08765847, Log Avg loss: 0.82611000, Global Avg Loss: 1.43540593, Time: 0.0402 Steps: 106190, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001879, Sample Num: 30064, Cur Loss: 1.07841635, Cur Avg Loss: 1.08919905, Log Avg loss: 1.37713370, Global Avg Loss: 1.43540045, Time: 0.0403 Steps: 106200, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001889, Sample Num: 30224, Cur Loss: 1.41543102, Cur Avg Loss: 1.09047777, Log Avg loss: 1.33074943, Global Avg Loss: 1.43539059, Time: 0.0403 Steps: 106210, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001899, Sample Num: 30384, Cur Loss: 1.10309768, Cur Avg Loss: 1.09014612, Log Avg loss: 1.02749667, Global Avg Loss: 1.43535219, Time: 0.0403 Steps: 106220, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001909, Sample Num: 30544, Cur Loss: 0.65373337, Cur Avg Loss: 1.08945476, Log Avg loss: 0.95816472, Global Avg Loss: 1.43530727, Time: 0.0403 Steps: 106230, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001919, Sample Num: 30704, Cur Loss: 2.69149208, Cur Avg Loss: 1.09019943, Log Avg loss: 1.23235774, Global Avg Loss: 1.43528817, Time: 0.0403 Steps: 106240, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001929, Sample Num: 30864, Cur Loss: 1.39577079, Cur Avg Loss: 1.09163927, Log Avg loss: 1.36794388, Global Avg Loss: 1.43528183, Time: 0.0403 Steps: 106250, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001939, Sample Num: 31024, Cur Loss: 0.44514304, Cur Avg Loss: 1.09098115, Log Avg loss: 0.96403088, Global Avg Loss: 1.43523748, Time: 0.0403 Steps: 106260, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001949, Sample Num: 31184, Cur Loss: 1.35846472, Cur Avg Loss: 1.09248490, Log Avg loss: 1.38406200, Global Avg Loss: 1.43523267, Time: 0.0403 Steps: 106270, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001959, Sample Num: 31344, Cur Loss: 1.87005401, Cur Avg Loss: 1.09365591, Log Avg loss: 1.32188517, Global Avg Loss: 1.43522200, Time: 0.0403 Steps: 106280, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001969, Sample Num: 31504, Cur Loss: 0.47362244, Cur Avg Loss: 1.09360548, Log Avg loss: 1.08372537, Global Avg Loss: 1.43518893, Time: 0.0402 Steps: 106290, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001979, Sample Num: 31664, Cur Loss: 1.71605968, Cur Avg Loss: 1.09398488, Log Avg loss: 1.16869036, Global Avg Loss: 1.43516386, Time: 0.0403 Steps: 106300, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001989, Sample Num: 31824, Cur Loss: 1.39376462, Cur Avg Loss: 1.09498711, Log Avg loss: 1.29332854, Global Avg Loss: 1.43515052, Time: 0.0402 Steps: 106310, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001999, Sample Num: 31984, Cur Loss: 1.74272192, Cur Avg Loss: 1.09565605, Log Avg loss: 1.22870695, Global Avg Loss: 1.43513110, Time: 0.0402 Steps: 106320, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002009, Sample Num: 32144, Cur Loss: 0.82120109, Cur Avg Loss: 1.09607153, Log Avg loss: 1.17912675, Global Avg Loss: 1.43510703, Time: 0.0402 Steps: 106330, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002019, Sample Num: 32304, Cur Loss: 1.72782218, Cur Avg Loss: 1.09589565, Log Avg loss: 1.06056101, Global Avg Loss: 1.43507180, Time: 0.0402 Steps: 106340, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002029, Sample Num: 32464, Cur Loss: 0.60959864, Cur Avg Loss: 1.09647985, Log Avg loss: 1.21443013, Global Avg Loss: 1.43505106, Time: 0.0402 Steps: 106350, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002039, Sample Num: 32624, Cur Loss: 0.50005913, Cur Avg Loss: 1.09593713, Log Avg loss: 0.98581858, Global Avg Loss: 1.43500882, Time: 0.0402 Steps: 106360, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002049, Sample Num: 32784, Cur Loss: 0.82950068, Cur Avg Loss: 1.09599412, Log Avg loss: 1.10761458, Global Avg Loss: 1.43497804, Time: 0.0402 Steps: 106370, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002059, Sample Num: 32944, Cur Loss: 0.32046711, Cur Avg Loss: 1.09586539, Log Avg loss: 1.06948910, Global Avg Loss: 1.43494368, Time: 0.0403 Steps: 106380, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002069, Sample Num: 33104, Cur Loss: 0.80620944, Cur Avg Loss: 1.09499147, Log Avg loss: 0.91505018, Global Avg Loss: 1.43489482, Time: 0.0402 Steps: 106390, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002079, Sample Num: 33264, Cur Loss: 1.20541203, Cur Avg Loss: 1.09407125, Log Avg loss: 0.90367828, Global Avg Loss: 1.43484489, Time: 0.0402 Steps: 106400, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002089, Sample Num: 33424, Cur Loss: 1.11648846, Cur Avg Loss: 1.09219516, Log Avg loss: 0.70215675, Global Avg Loss: 1.43477604, Time: 0.0402 Steps: 106410, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002099, Sample Num: 33584, Cur Loss: 1.94425237, Cur Avg Loss: 1.09275012, Log Avg loss: 1.20868195, Global Avg Loss: 1.43475479, Time: 0.0402 Steps: 106420, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002109, Sample Num: 33744, Cur Loss: 0.86915696, Cur Avg Loss: 1.09234945, Log Avg loss: 1.00824717, Global Avg Loss: 1.43471472, Time: 0.0726 Steps: 106430, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002119, Sample Num: 33904, Cur Loss: 1.22991645, Cur Avg Loss: 1.09274442, Log Avg loss: 1.17604447, Global Avg Loss: 1.43469041, Time: 0.0403 Steps: 106440, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002129, Sample Num: 34055, Cur Loss: 3.38724661, Cur Avg Loss: 1.09519322, Log Avg loss: 1.61409403, Global Avg Loss: 1.43470727, Time: 0.0516 Steps: 106450, Updated lr: 0.000000 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 95.715481, Avg time per batch (s): 0.040000 {"eval_avg_loss": 1.072385, "eval_total_loss": 753.886779, "eval_mae": 0.850235, "eval_mse": 1.072465, "eval_r2": 0.318271, "eval_sp_statistic": 0.489295, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.588, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.058563, "test_total_loss": 531.398437, "test_mae": 0.854947, "test_mse": 1.058747, "test_r2": 0.316676, "test_sp_statistic": 0.506499, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.631258, "test_ps_pvalue": 0.0, "lr": 0.0, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4347072680104838, "train_cur_epoch_loss": 2331.666368752718, "train_cur_epoch_avg_loss": 1.0951932215841793, "train_cur_epoch_time": 95.71548080444336, "train_cur_epoch_avg_time": 0.04495795246803352, "epoch": 50, "step": 106450} ################################################## #########################Best Metric######################### {"epoch": 50, "global_step": 106450, "eval_avg_loss": 1.072385, "eval_total_loss": 753.886779, "eval_mae": 0.850235, "eval_mse": 1.072465, "eval_r2": 0.318271, "eval_sp_statistic": 0.489295, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.588, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.058563, "test_total_loss": 531.398437, "test_mae": 0.854947, "test_mse": 1.058747, "test_r2": 0.316676, "test_sp_statistic": 0.506499, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.631258, "test_ps_pvalue": 0.0} ################################################## Total Time: 9463.850394, Avg time per epoch(50 epochs): 189.280000 ++++++++++++Validation+++++++++++++ best sp_statistic global step: 106450 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/seq/20250410011916/checkpoint-106450 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## {"evaluation_avg_loss_106450": 1.072385, "evaluation_total_loss_106450": 753.886779, "evaluation_mae_106450": 0.850235, "evaluation_mse_106450": 1.072465, "evaluation_r2_106450": 0.318271, "evaluation_sp_statistic_106450": 0.489295, "evaluation_sp_pvalue_106450": 0.0, "evaluation_ps_statistic_106450": 0.588, "evaluation_ps_pvalue_106450": 0.0} ++++++++++++Testing+++++++++++++ best sp_statistic global step: 106450 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/seq/20250410011916/checkpoint-106450 ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## {"evaluation_avg_loss_106450": 1.058563, "evaluation_total_loss_106450": 531.398437, "evaluation_mae_106450": 0.854947, "evaluation_mse_106450": 1.058747, "evaluation_r2_106450": 0.316676, "evaluation_sp_statistic_106450": 0.506499, "evaluation_sp_pvalue_106450": 0.0, "evaluation_ps_statistic_106450": 0.631258, "evaluation_ps_pvalue_106450": 0.0}