{"adam_epsilon": 1e-08, "alphabet": "gene_prot", "append_eos": true, "asl_gamma_neg": 4.0, "asl_gamma_pos": 1.0, "best_metric_type": "sp_statistic", "beta1": 0.9, "beta2": 0.98, "buffer_size": 4096, "cache_dir": null, "classifier_activate_func": "gelu", "classifier_size": 128, "codes_file": null, "config_path": "../config/luca_base/luca_base_config.json", "cross_atten": false, "dataset_name": "DMS_Bind_Reps_Strain", "dataset_type": "protein", "delete_old": true, "dev_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/dev/", "device": "cuda", "do_eval": true, "do_lower_case": false, "do_metrics": true, "do_predict": true, "do_train": true, "dropout_prob": 0.1, "early_stop_epoch": -1, "emb_activate_func": "gelu", "embedding_complete": true, "embedding_complete_seg_overlap": true, "embedding_fixed_len_a_time": 3072, "embedding_input_size": 2560, "embedding_input_size_a": null, "embedding_input_size_b": null, "eval_all_checkpoints": false, "evaluate_during_training": true, "evaluate_steps": 2000, "evaluate_strategy": "epoch", "fc_activate_func": "gelu", "focal_loss_alpha": 0.7, "focal_loss_gamma": 2.0, "focal_loss_reduce": false, "fp16": false, "fp16_embedding": false, "fp16_opt_level": "O1", "fusion_type": "concat", "gradient_accumulation_steps": 1, "hidden_size": 1024, "ignore_index": -100, "input_mode": "single", "input_type": "matrix", "intermediate_size": 4096, "label_filepath": "../dataset/DMS_Bind_Reps_Strain/protein/regression/label.txt", "label_size": 1, "label_type": "DMS_Bind_Reps_Strain", "learning_rate": 0.0001, "llm_dir": "..", "llm_dirpath": "../llm/models/lucagplm/v2.0/token_level,span_level,seq_level,structure_level/lucaone_gplm/20231125113045/checkpoint-step17600000", "llm_step": "17600000", "llm_task_level": "token_level,span_level,seq_level,structure_level", "llm_time_str": "20231125113045", "llm_type": "lucaone_gplm", "llm_version": "v2.0", "lmdb_path": null, "local_rank": -1, "log_dir": "../logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226035651", "logging_steps": 10, "loss_reduction": "mean", "loss_type": "l2", "lr_decay_rate": 0.9, "lr_update_strategy": "step", "matrix_add_special_token": false, "matrix_dirpath": "../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v2.0/lucaone_gplm/20231125113045/17600000", "matrix_embedding_exists": false, "matrix_encoder": false, "matrix_encoder_act": false, "matrix_fc_size": "128", "matrix_max_length": 100000, "matrix_max_length_a": null, "matrix_max_length_b": null, "matrix_pooling_type": "value_attention", "max_grad_norm": 1.0, "max_sentence_length": null, "max_sentences": null, "max_steps": -1, "model_dirpath": null, "model_type": "luca_base", "n_gpu": 1, "no_cuda": false, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "non_ignore": false, "not_append_eos": false, "not_matrix_encoder_shared": false, "not_prepend_bos": false, "not_save_emb_to_disk": false, "not_seq_encoder_shared": false, "num_attention_heads": 4, "num_hidden_layers": 2, "num_train_epochs": 50, "output_dir": "../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226035651", "output_mode": "regression", "overwrite_cache": false, "overwrite_output_dir": true, "per_gpu_eval_batch_size": 16, "per_gpu_train_batch_size": 16, "pos_weight": 1.0, "position_embedding_type": "absolute", "prepend_bos": true, "save_all": false, "save_steps": -1, "seed": 1221, "self_atten": false, "seq_fc_size": "null", "seq_max_length": 100000, "seq_max_length_a": null, "seq_max_length_b": null, "seq_pooling_type": "value_attention", "seq_subword": false, "seq_vocab_path": "gene_prot", "sigmoid": false, "task_level_type": "seq_level", "task_type": "regression", "tb_log_dir": "../tb-logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226035651", "test_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/test/", "time_str": "20250226035655", "train_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/train/", "trunc_type": "right", "vector_dirpath": "../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v2.0/lucaone_gplm/20231125113045/17600000", "vector_fc_size": "null", "vocab_size": 39, "warmup_steps": 1000, "weight": null, "weight_decay": 0.01, "worker_num": 0} ################################################## n_gpu: 1 ################################################## Inputs: Input Name List: protein,embedding_matrix ################################################## Encoder Config: {'llm_type': 'lucaone_gplm', 'llm_version': 'v2.0', 'llm_step': '17600000', 'llm_dirpath': '../llm/models/lucagplm/v2.0/token_level,span_level,seq_level,structure_level/lucaone_gplm/20231125113045/checkpoint-step17600000', 'input_type': 'matrix', 'trunc_type': 'right', 'seq_max_length': 100000, 'atom_seq_max_length': None, 'vector_dirpath': '../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v2.0/lucaone_gplm/20231125113045/17600000', 'matrix_dirpath': '../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v2.0/lucaone_gplm/20231125113045/17600000', 'local_rank': -1, 'max_sentence_length': None, 'max_sentences': None, 'matrix_add_special_token': False, 'embedding_complete': True, 'embedding_complete_seg_overlap': True, 'embedding_fixed_len_a_time': 3072, 'matrix_embedding_exists': False, 'save_emb_to_disk': True, 'fp16_embedding': False} ################################################## Model Config: LucaConfig { "alphabet": "gene_prot", "attention_probs_dropout_prob": 0.1, "classifier_activate_func": "gelu", "classifier_dropout_prob": 0.1, "classifier_size": 128, "cls_token_id": 2, "cross_atten": false, "directionality": "bidi", "emb_activate_func": "gelu", "embedding_input_size": 2560, "fc_activate_func": "gelu", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "LABEL_0" }, "ignore_index": -100, "initializer_range": 0.02, "intermediate_size": 4096, "kernel_size": 7, "label2id": { "LABEL_0": 0 }, "layer_norm_eps": 1e-12, "loss_reduction": "mean", "matrix_fc_size": [ 128 ], "matrix_max_length": 100000, "matrix_pooling_type": "value_attention", "max_position_embeddings": 100002, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "num_attention_heads": 4, "num_hidden_layers": 2, "pad_token_id": 0, "pos_weight": 1.0, "position_embedding_type": "absolute", "self_atten": false, "sep_token_id": 3, "seq_fc_size": null, "seq_max_length": 100000, "seq_pooling_type": "value_attention", "token_dropout": null, "transformers_version": "4.29.0", "type_vocab_size": 2, "use_luca_layer_norm_v2": true, "vector_fc_size": null, "vocab_size": 39 } ################################################## Mode Architecture: LucaBase( (matrix_pooler): GlobalMaskValueAttentionPooling1D (2560 -> 2560) (linear): ModuleList( (0): ModuleList( (0): Linear(in_features=2560, out_features=128, bias=True) (1): GELU(approximate='none') ) ) (dropout): Dropout(p=0.1, inplace=False) (hidden_layer): Linear(in_features=128, out_features=128, bias=True) (hidden_act): GELU(approximate='none') (classifier): Linear(in_features=128, out_features=1, bias=True) (loss_fct): MaskedMSELoss( (criterion): MSELoss() ) ) ################################################## Model parameters: 20005249 ################################################## {"total_num": "19.080000M", "total_size": "76.310000MB", "param_sum": "19.080000M", "param_size": "76.310000MB", "buffer_sum": "0.000000M", "buffer_size": "0.000000MB", "trainable_num": "19.078492M", "trainable_size": "76.313969MB"} ################################################## Train dataset len: 34055, batch size: 16, batch num: 2129 Train dataset t_total: 106450, max_steps: -1 ***** Running training ***** Train Dataset Num examples = 34055 Train Dataset Num Epochs = 50 Logging Steps = 10 Saving Steps = -1 Evaluating Strategy = epoch Train Dataset Instantaneous batch size per GPU = 16 Train Dataset Total train batch size (w. parallel, distributed & accumulation) = 16 Train Dataset Gradient Accumulation steps = 1 Train Dataset Total optimization steps = 106450 ################################################## Training, Epoch: 0001, Batch: 000010, Sample Num: 160, Cur Loss: 70.90951538, Cur Avg Loss: 71.97157822, Log Avg loss: 71.97157822, Global Avg Loss: 71.97157822, Time: 0.0209 Steps: 10, Updated lr: 0.000001 Training, Epoch: 0001, Batch: 000020, Sample Num: 320, Cur Loss: 73.38737488, Cur Avg Loss: 71.44826431, Log Avg loss: 70.92495041, Global Avg Loss: 71.44826431, Time: 0.0208 Steps: 20, Updated lr: 0.000002 Training, Epoch: 0001, Batch: 000030, Sample Num: 480, Cur Loss: 73.06079102, Cur Avg Loss: 72.76414312, Log Avg loss: 75.39590073, Global Avg Loss: 72.76414312, Time: 0.0209 Steps: 30, Updated lr: 0.000003 Training, Epoch: 0001, Batch: 000040, Sample Num: 640, Cur Loss: 67.75593567, Cur Avg Loss: 72.24549026, Log Avg loss: 70.68953171, Global Avg Loss: 72.24549026, Time: 0.0208 Steps: 40, Updated lr: 0.000004 Training, Epoch: 0001, Batch: 000050, Sample Num: 800, Cur Loss: 63.26806259, Cur Avg Loss: 72.05173531, Log Avg loss: 71.27671547, Global Avg Loss: 72.05173531, Time: 0.0208 Steps: 50, Updated lr: 0.000005 Training, Epoch: 0001, Batch: 000060, Sample Num: 960, Cur Loss: 74.06801605, Cur Avg Loss: 71.98779812, Log Avg loss: 71.66811218, Global Avg Loss: 71.98779812, Time: 0.0208 Steps: 60, Updated lr: 0.000006 Training, Epoch: 0001, Batch: 000070, Sample Num: 1120, Cur Loss: 63.26382446, Cur Avg Loss: 71.65383524, Log Avg loss: 69.65005798, Global Avg Loss: 71.65383524, Time: 0.0209 Steps: 70, Updated lr: 0.000007 Training, Epoch: 0001, Batch: 000080, Sample Num: 1280, Cur Loss: 74.89678192, Cur Avg Loss: 71.26885509, Log Avg loss: 68.57399406, Global Avg Loss: 71.26885509, Time: 0.0208 Steps: 80, Updated lr: 0.000008 Training, Epoch: 0001, Batch: 000090, Sample Num: 1440, Cur Loss: 73.61959076, Cur Avg Loss: 70.92178531, Log Avg loss: 68.14522705, Global Avg Loss: 70.92178531, Time: 0.0209 Steps: 90, Updated lr: 0.000009 Training, Epoch: 0001, Batch: 000100, Sample Num: 1600, Cur Loss: 77.03057861, Cur Avg Loss: 71.14635666, Log Avg loss: 73.16749878, Global Avg Loss: 71.14635666, Time: 0.0208 Steps: 100, Updated lr: 0.000010 Training, Epoch: 0001, Batch: 000110, Sample Num: 1760, Cur Loss: 68.08878326, Cur Avg Loss: 71.15902925, Log Avg loss: 71.28575516, Global Avg Loss: 71.15902925, Time: 0.0208 Steps: 110, Updated lr: 0.000011 Training, Epoch: 0001, Batch: 000120, Sample Num: 1920, Cur Loss: 71.50331116, Cur Avg Loss: 71.29881376, Log Avg loss: 72.83644333, Global Avg Loss: 71.29881376, Time: 0.0208 Steps: 120, Updated lr: 0.000012 Training, Epoch: 0001, Batch: 000130, Sample Num: 2080, Cur Loss: 77.49301147, Cur Avg Loss: 71.43821998, Log Avg loss: 73.11109467, Global Avg Loss: 71.43821998, Time: 0.0208 Steps: 130, Updated lr: 0.000013 Training, Epoch: 0001, Batch: 000140, Sample Num: 2240, Cur Loss: 77.38601685, Cur Avg Loss: 71.55339475, Log Avg loss: 73.05066681, Global Avg Loss: 71.55339475, Time: 0.0209 Steps: 140, Updated lr: 0.000014 Training, Epoch: 0001, Batch: 000150, Sample Num: 2400, Cur Loss: 73.16833496, Cur Avg Loss: 71.80470047, Log Avg loss: 75.32298050, Global Avg Loss: 71.80470047, Time: 0.0207 Steps: 150, Updated lr: 0.000015 Training, Epoch: 0001, Batch: 000160, Sample Num: 2560, Cur Loss: 65.29322052, Cur Avg Loss: 71.69389179, Log Avg loss: 70.03176155, Global Avg Loss: 71.69389179, Time: 0.0208 Steps: 160, Updated lr: 0.000016 Training, Epoch: 0001, Batch: 000170, Sample Num: 2720, Cur Loss: 70.90133667, Cur Avg Loss: 71.58245349, Log Avg loss: 69.79944077, Global Avg Loss: 71.58245349, Time: 0.0207 Steps: 170, Updated lr: 0.000017 Training, Epoch: 0001, Batch: 000180, Sample Num: 2880, Cur Loss: 62.54254532, Cur Avg Loss: 71.55022697, Log Avg loss: 71.00237617, Global Avg Loss: 71.55022697, Time: 0.0209 Steps: 180, Updated lr: 0.000018 Training, Epoch: 0001, Batch: 000190, Sample Num: 3040, Cur Loss: 77.74044037, Cur Avg Loss: 71.59516361, Log Avg loss: 72.40402298, Global Avg Loss: 71.59516361, Time: 0.0209 Steps: 190, Updated lr: 0.000019 Training, Epoch: 0001, Batch: 000200, Sample Num: 3200, Cur Loss: 64.84249878, Cur Avg Loss: 71.58985394, Log Avg loss: 71.48897018, Global Avg Loss: 71.58985394, Time: 0.0208 Steps: 200, Updated lr: 0.000020 Training, Epoch: 0001, Batch: 000210, Sample Num: 3360, Cur Loss: 68.44723511, Cur Avg Loss: 71.51799424, Log Avg loss: 70.08080025, Global Avg Loss: 71.51799424, Time: 0.0209 Steps: 210, Updated lr: 0.000021 Training, Epoch: 0001, Batch: 000220, Sample Num: 3520, Cur Loss: 63.29708481, Cur Avg Loss: 71.39962812, Log Avg loss: 68.91393967, Global Avg Loss: 71.39962812, Time: 0.0211 Steps: 220, Updated lr: 0.000022 Training, Epoch: 0001, Batch: 000230, Sample Num: 3680, Cur Loss: 76.90761566, Cur Avg Loss: 71.24403416, Log Avg loss: 67.82096710, Global Avg Loss: 71.24403416, Time: 0.0208 Steps: 230, Updated lr: 0.000023 Training, Epoch: 0001, Batch: 000240, Sample Num: 3840, Cur Loss: 69.95604706, Cur Avg Loss: 71.29823365, Log Avg loss: 72.54482193, Global Avg Loss: 71.29823365, Time: 0.0208 Steps: 240, Updated lr: 0.000024 Training, Epoch: 0001, Batch: 000250, Sample Num: 4000, Cur Loss: 71.23264313, Cur Avg Loss: 71.23045607, Log Avg loss: 69.60379410, Global Avg Loss: 71.23045607, Time: 0.0208 Steps: 250, Updated lr: 0.000025 Training, Epoch: 0001, Batch: 000260, Sample Num: 4160, Cur Loss: 72.00950623, Cur Avg Loss: 71.13731685, Log Avg loss: 68.80883636, Global Avg Loss: 71.13731685, Time: 0.0208 Steps: 260, Updated lr: 0.000026 Training, Epoch: 0001, Batch: 000270, Sample Num: 4320, Cur Loss: 58.43429947, Cur Avg Loss: 70.97957759, Log Avg loss: 66.87835693, Global Avg Loss: 70.97957759, Time: 0.0208 Steps: 270, Updated lr: 0.000027 Training, Epoch: 0001, Batch: 000280, Sample Num: 4480, Cur Loss: 64.54907227, Cur Avg Loss: 70.86935049, Log Avg loss: 67.89321861, Global Avg Loss: 70.86935049, Time: 0.0208 Steps: 280, Updated lr: 0.000028 Training, Epoch: 0001, Batch: 000290, Sample Num: 4640, Cur Loss: 64.32907867, Cur Avg Loss: 70.75050028, Log Avg loss: 67.42269440, Global Avg Loss: 70.75050028, Time: 0.0208 Steps: 290, Updated lr: 0.000029 Training, Epoch: 0001, Batch: 000300, Sample Num: 4800, Cur Loss: 67.38747406, Cur Avg Loss: 70.58220507, Log Avg loss: 65.70164413, Global Avg Loss: 70.58220507, Time: 0.0208 Steps: 300, Updated lr: 0.000030 Training, Epoch: 0001, Batch: 000310, Sample Num: 4960, Cur Loss: 58.04085541, Cur Avg Loss: 70.33992623, Log Avg loss: 63.07156105, Global Avg Loss: 70.33992623, Time: 0.0208 Steps: 310, Updated lr: 0.000031 Training, Epoch: 0001, Batch: 000320, Sample Num: 5120, Cur Loss: 66.36784363, Cur Avg Loss: 70.08196191, Log Avg loss: 62.08506775, Global Avg Loss: 70.08196191, Time: 0.0208 Steps: 320, Updated lr: 0.000032 Training, Epoch: 0001, Batch: 000330, Sample Num: 5280, Cur Loss: 61.39367294, Cur Avg Loss: 69.73367149, Log Avg loss: 58.58837814, Global Avg Loss: 69.73367149, Time: 0.0208 Steps: 330, Updated lr: 0.000033 Training, Epoch: 0001, Batch: 000340, Sample Num: 5440, Cur Loss: 55.32677460, Cur Avg Loss: 69.37252874, Log Avg loss: 57.45481796, Global Avg Loss: 69.37252874, Time: 0.0207 Steps: 340, Updated lr: 0.000034 Training, Epoch: 0001, Batch: 000350, Sample Num: 5600, Cur Loss: 65.16910553, Cur Avg Loss: 69.06835692, Log Avg loss: 58.72651520, Global Avg Loss: 69.06835692, Time: 0.0209 Steps: 350, Updated lr: 0.000035 Training, Epoch: 0001, Batch: 000360, Sample Num: 5760, Cur Loss: 51.23231888, Cur Avg Loss: 68.66861230, Log Avg loss: 54.67755051, Global Avg Loss: 68.66861230, Time: 0.0208 Steps: 360, Updated lr: 0.000036 Training, Epoch: 0001, Batch: 000370, Sample Num: 5920, Cur Loss: 46.70240402, Cur Avg Loss: 68.14888950, Log Avg loss: 49.43886871, Global Avg Loss: 68.14888950, Time: 0.0208 Steps: 370, Updated lr: 0.000037 Training, Epoch: 0001, Batch: 000380, Sample Num: 6080, Cur Loss: 45.58461380, Cur Avg Loss: 67.59011840, Log Avg loss: 46.91558762, Global Avg Loss: 67.59011840, Time: 0.0208 Steps: 380, Updated lr: 0.000038 Training, Epoch: 0001, Batch: 000390, Sample Num: 6240, Cur Loss: 35.08296204, Cur Avg Loss: 66.97057596, Log Avg loss: 43.42796326, Global Avg Loss: 66.97057596, Time: 0.0208 Steps: 390, Updated lr: 0.000039 Training, Epoch: 0001, Batch: 000400, Sample Num: 6400, Cur Loss: 39.25920105, Cur Avg Loss: 66.31044791, Log Avg loss: 40.56545410, Global Avg Loss: 66.31044791, Time: 0.0208 Steps: 400, Updated lr: 0.000040 Training, Epoch: 0001, Batch: 000410, Sample Num: 6560, Cur Loss: 38.13664246, Cur Avg Loss: 65.61638452, Log Avg loss: 37.85384884, Global Avg Loss: 65.61638452, Time: 0.0208 Steps: 410, Updated lr: 0.000041 Training, Epoch: 0001, Batch: 000420, Sample Num: 6720, Cur Loss: 32.00386429, Cur Avg Loss: 64.85209804, Log Avg loss: 33.51635227, Global Avg Loss: 64.85209804, Time: 0.0208 Steps: 420, Updated lr: 0.000042 Training, Epoch: 0001, Batch: 000430, Sample Num: 6880, Cur Loss: 24.73770523, Cur Avg Loss: 64.01670542, Log Avg loss: 28.93021526, Global Avg Loss: 64.01670542, Time: 0.0208 Steps: 430, Updated lr: 0.000043 Training, Epoch: 0001, Batch: 000440, Sample Num: 7040, Cur Loss: 26.33440208, Cur Avg Loss: 63.13484366, Log Avg loss: 25.21478806, Global Avg Loss: 63.13484366, Time: 0.0208 Steps: 440, Updated lr: 0.000044 Training, Epoch: 0001, Batch: 000450, Sample Num: 7200, Cur Loss: 18.04591751, Cur Avg Loss: 62.21250813, Log Avg loss: 21.62974510, Global Avg Loss: 62.21250813, Time: 0.0208 Steps: 450, Updated lr: 0.000045 Training, Epoch: 0001, Batch: 000460, Sample Num: 7360, Cur Loss: 16.97378540, Cur Avg Loss: 61.24216814, Log Avg loss: 17.57686825, Global Avg Loss: 61.24216814, Time: 0.0208 Steps: 460, Updated lr: 0.000046 Training, Epoch: 0001, Batch: 000470, Sample Num: 7520, Cur Loss: 16.98702621, Cur Avg Loss: 60.24498903, Log Avg loss: 14.37475004, Global Avg Loss: 60.24498903, Time: 0.0209 Steps: 470, Updated lr: 0.000047 Training, Epoch: 0001, Batch: 000480, Sample Num: 7680, Cur Loss: 6.04378319, Cur Avg Loss: 59.20572080, Log Avg loss: 10.36011429, Global Avg Loss: 59.20572080, Time: 0.0208 Steps: 480, Updated lr: 0.000048 Training, Epoch: 0001, Batch: 000490, Sample Num: 7840, Cur Loss: 6.14756584, Cur Avg Loss: 58.15283508, Log Avg loss: 7.61432052, Global Avg Loss: 58.15283508, Time: 0.0208 Steps: 490, Updated lr: 0.000049 Training, Epoch: 0001, Batch: 000500, Sample Num: 8000, Cur Loss: 3.97294259, Cur Avg Loss: 57.08794447, Log Avg loss: 4.90830431, Global Avg Loss: 57.08794447, Time: 0.0208 Steps: 500, Updated lr: 0.000050 Training, Epoch: 0001, Batch: 000510, Sample Num: 8160, Cur Loss: 2.07843256, Cur Avg Loss: 56.02971967, Log Avg loss: 3.11847966, Global Avg Loss: 56.02971967, Time: 0.0209 Steps: 510, Updated lr: 0.000051 Training, Epoch: 0001, Batch: 000520, Sample Num: 8320, Cur Loss: 2.21813393, Cur Avg Loss: 54.98870200, Log Avg loss: 1.89680083, Global Avg Loss: 54.98870200, Time: 0.0210 Steps: 520, Updated lr: 0.000052 Training, Epoch: 0001, Batch: 000530, Sample Num: 8480, Cur Loss: 1.39256763, Cur Avg Loss: 53.98682440, Log Avg loss: 1.88918903, Global Avg Loss: 53.98682440, Time: 0.0210 Steps: 530, Updated lr: 0.000053 Training, Epoch: 0001, Batch: 000540, Sample Num: 8640, Cur Loss: 1.12017047, Cur Avg Loss: 53.01587627, Log Avg loss: 1.55562573, Global Avg Loss: 53.01587627, Time: 0.0210 Steps: 540, Updated lr: 0.000054 Training, Epoch: 0001, Batch: 000550, Sample Num: 8800, Cur Loss: 1.95544624, Cur Avg Loss: 52.08808995, Log Avg loss: 1.98762883, Global Avg Loss: 52.08808995, Time: 0.0211 Steps: 550, Updated lr: 0.000055 Training, Epoch: 0001, Batch: 000560, Sample Num: 8960, Cur Loss: 1.67597461, Cur Avg Loss: 51.18579424, Log Avg loss: 1.55952967, Global Avg Loss: 51.18579424, Time: 0.0210 Steps: 560, Updated lr: 0.000056 Training, Epoch: 0001, Batch: 000570, Sample Num: 9120, Cur Loss: 1.05855823, Cur Avg Loss: 50.32214287, Log Avg loss: 1.95766623, Global Avg Loss: 50.32214287, Time: 0.0210 Steps: 570, Updated lr: 0.000057 Training, Epoch: 0001, Batch: 000580, Sample Num: 9280, Cur Loss: 1.05566263, Cur Avg Loss: 49.48161169, Log Avg loss: 1.57133459, Global Avg Loss: 49.48161169, Time: 0.0211 Steps: 580, Updated lr: 0.000058 Training, Epoch: 0001, Batch: 000590, Sample Num: 9440, Cur Loss: 0.67639273, Cur Avg Loss: 48.67065598, Log Avg loss: 1.63522508, Global Avg Loss: 48.67065598, Time: 0.0211 Steps: 590, Updated lr: 0.000059 Training, Epoch: 0001, Batch: 000600, Sample Num: 9600, Cur Loss: 2.78337455, Cur Avg Loss: 47.89210655, Log Avg loss: 1.95768970, Global Avg Loss: 47.89210655, Time: 0.0211 Steps: 600, Updated lr: 0.000060 Training, Epoch: 0001, Batch: 000610, Sample Num: 9760, Cur Loss: 1.61834371, Cur Avg Loss: 47.12982870, Log Avg loss: 1.39315802, Global Avg Loss: 47.12982870, Time: 0.0210 Steps: 610, Updated lr: 0.000061 Training, Epoch: 0001, Batch: 000620, Sample Num: 9920, Cur Loss: 1.00718355, Cur Avg Loss: 46.39865115, Log Avg loss: 1.79682081, Global Avg Loss: 46.39865115, Time: 0.0210 Steps: 620, Updated lr: 0.000062 Training, Epoch: 0001, Batch: 000630, Sample Num: 10080, Cur Loss: 1.01368439, Cur Avg Loss: 45.68956270, Log Avg loss: 1.72607878, Global Avg Loss: 45.68956270, Time: 0.0209 Steps: 630, Updated lr: 0.000063 Training, Epoch: 0001, Batch: 000640, Sample Num: 10240, Cur Loss: 1.52800417, Cur Avg Loss: 45.00284556, Log Avg loss: 1.73966555, Global Avg Loss: 45.00284556, Time: 0.0210 Steps: 640, Updated lr: 0.000064 Training, Epoch: 0001, Batch: 000650, Sample Num: 10400, Cur Loss: 1.58764505, Cur Avg Loss: 44.33940860, Log Avg loss: 1.87944303, Global Avg Loss: 44.33940860, Time: 0.0210 Steps: 650, Updated lr: 0.000065 Training, Epoch: 0001, Batch: 000660, Sample Num: 10560, Cur Loss: 1.49836981, Cur Avg Loss: 43.69634808, Log Avg loss: 1.89741427, Global Avg Loss: 43.69634808, Time: 0.0210 Steps: 660, Updated lr: 0.000066 Training, Epoch: 0001, Batch: 000670, Sample Num: 10720, Cur Loss: 1.14921820, Cur Avg Loss: 43.06530059, Log Avg loss: 1.41616600, Global Avg Loss: 43.06530059, Time: 0.0210 Steps: 670, Updated lr: 0.000067 Training, Epoch: 0001, Batch: 000680, Sample Num: 10880, Cur Loss: 1.91261458, Cur Avg Loss: 42.46058515, Log Avg loss: 1.94465126, Global Avg Loss: 42.46058515, Time: 0.0210 Steps: 680, Updated lr: 0.000068 Training, Epoch: 0001, Batch: 000690, Sample Num: 11040, Cur Loss: 0.72542989, Cur Avg Loss: 41.87021787, Log Avg loss: 1.72524261, Global Avg Loss: 41.87021787, Time: 0.0210 Steps: 690, Updated lr: 0.000069 Training, Epoch: 0001, Batch: 000700, Sample Num: 11200, Cur Loss: 1.98509014, Cur Avg Loss: 41.29818159, Log Avg loss: 1.82767793, Global Avg Loss: 41.29818159, Time: 0.0210 Steps: 700, Updated lr: 0.000070 Training, Epoch: 0001, Batch: 000710, Sample Num: 11360, Cur Loss: 2.23463345, Cur Avg Loss: 40.74343283, Log Avg loss: 1.91102005, Global Avg Loss: 40.74343283, Time: 0.0210 Steps: 710, Updated lr: 0.000071 Training, Epoch: 0001, Batch: 000720, Sample Num: 11520, Cur Loss: 1.62306678, Cur Avg Loss: 40.19929075, Log Avg loss: 1.56520280, Global Avg Loss: 40.19929075, Time: 0.0210 Steps: 720, Updated lr: 0.000072 Training, Epoch: 0001, Batch: 000730, Sample Num: 11680, Cur Loss: 2.74335289, Cur Avg Loss: 39.67138938, Log Avg loss: 1.66249079, Global Avg Loss: 39.67138938, Time: 0.0210 Steps: 730, Updated lr: 0.000073 Training, Epoch: 0001, Batch: 000740, Sample Num: 11840, Cur Loss: 1.70948219, Cur Avg Loss: 39.15911983, Log Avg loss: 1.76344263, Global Avg Loss: 39.15911983, Time: 0.0210 Steps: 740, Updated lr: 0.000074 Training, Epoch: 0001, Batch: 000750, Sample Num: 12000, Cur Loss: 3.12766933, Cur Avg Loss: 38.66173079, Log Avg loss: 1.85494202, Global Avg Loss: 38.66173079, Time: 0.0210 Steps: 750, Updated lr: 0.000075 Training, Epoch: 0001, Batch: 000760, Sample Num: 12160, Cur Loss: 0.97303689, Cur Avg Loss: 38.17735076, Log Avg loss: 1.84884827, Global Avg Loss: 38.17735076, Time: 0.0210 Steps: 760, Updated lr: 0.000076 Training, Epoch: 0001, Batch: 000770, Sample Num: 12320, Cur Loss: 1.66447020, Cur Avg Loss: 37.70645587, Log Avg loss: 1.91844469, Global Avg Loss: 37.70645587, Time: 0.0209 Steps: 770, Updated lr: 0.000077 Training, Epoch: 0001, Batch: 000780, Sample Num: 12480, Cur Loss: 1.49392223, Cur Avg Loss: 37.24884330, Log Avg loss: 2.01267549, Global Avg Loss: 37.24884330, Time: 0.0208 Steps: 780, Updated lr: 0.000078 Training, Epoch: 0001, Batch: 000790, Sample Num: 12640, Cur Loss: 1.13783669, Cur Avg Loss: 36.79733037, Log Avg loss: 1.57932148, Global Avg Loss: 36.79733037, Time: 0.0208 Steps: 790, Updated lr: 0.000079 Training, Epoch: 0001, Batch: 000800, Sample Num: 12800, Cur Loss: 1.28982687, Cur Avg Loss: 36.35906971, Log Avg loss: 1.73647760, Global Avg Loss: 36.35906971, Time: 0.0208 Steps: 800, Updated lr: 0.000080 Training, Epoch: 0001, Batch: 000810, Sample Num: 12960, Cur Loss: 2.04872036, Cur Avg Loss: 35.93118365, Log Avg loss: 1.70029843, Global Avg Loss: 35.93118365, Time: 0.0209 Steps: 810, Updated lr: 0.000081 Training, Epoch: 0001, Batch: 000820, Sample Num: 13120, Cur Loss: 1.20649576, Cur Avg Loss: 35.51587221, Log Avg loss: 1.87564593, Global Avg Loss: 35.51587221, Time: 0.0208 Steps: 820, Updated lr: 0.000082 Training, Epoch: 0001, Batch: 000830, Sample Num: 13280, Cur Loss: 1.02980208, Cur Avg Loss: 35.10937179, Log Avg loss: 1.77633761, Global Avg Loss: 35.10937179, Time: 0.0209 Steps: 830, Updated lr: 0.000083 Training, Epoch: 0001, Batch: 000840, Sample Num: 13440, Cur Loss: 1.25788331, Cur Avg Loss: 34.71229522, Log Avg loss: 1.75493932, Global Avg Loss: 34.71229522, Time: 0.0208 Steps: 840, Updated lr: 0.000084 Training, Epoch: 0001, Batch: 000850, Sample Num: 13600, Cur Loss: 3.20693064, Cur Avg Loss: 34.32418688, Log Avg loss: 1.72308624, Global Avg Loss: 34.32418688, Time: 0.0208 Steps: 850, Updated lr: 0.000085 Training, Epoch: 0001, Batch: 000860, Sample Num: 13760, Cur Loss: 1.43662238, Cur Avg Loss: 33.94376354, Log Avg loss: 1.60777972, Global Avg Loss: 33.94376354, Time: 0.0209 Steps: 860, Updated lr: 0.000086 Training, Epoch: 0001, Batch: 000870, Sample Num: 13920, Cur Loss: 1.25190055, Cur Avg Loss: 33.57493427, Log Avg loss: 1.85561739, Global Avg Loss: 33.57493427, Time: 0.0208 Steps: 870, Updated lr: 0.000087 Training, Epoch: 0001, Batch: 000880, Sample Num: 14080, Cur Loss: 1.17266011, Cur Avg Loss: 33.21435138, Log Avg loss: 1.84363999, Global Avg Loss: 33.21435138, Time: 0.0208 Steps: 880, Updated lr: 0.000088 Training, Epoch: 0001, Batch: 000890, Sample Num: 14240, Cur Loss: 1.17084789, Cur Avg Loss: 32.86059191, Log Avg loss: 1.72975827, Global Avg Loss: 32.86059191, Time: 0.0208 Steps: 890, Updated lr: 0.000089 Training, Epoch: 0001, Batch: 000900, Sample Num: 14400, Cur Loss: 1.07807088, Cur Avg Loss: 32.51344659, Log Avg loss: 1.61751316, Global Avg Loss: 32.51344659, Time: 0.0208 Steps: 900, Updated lr: 0.000090 Training, Epoch: 0001, Batch: 000910, Sample Num: 14560, Cur Loss: 2.16501617, Cur Avg Loss: 32.17763357, Log Avg loss: 1.95446194, Global Avg Loss: 32.17763357, Time: 0.0208 Steps: 910, Updated lr: 0.000091 Training, Epoch: 0001, Batch: 000920, Sample Num: 14720, Cur Loss: 2.12338662, Cur Avg Loss: 31.84754801, Log Avg loss: 1.80976247, Global Avg Loss: 31.84754801, Time: 0.0209 Steps: 920, Updated lr: 0.000092 Training, Epoch: 0001, Batch: 000930, Sample Num: 14880, Cur Loss: 2.34390068, Cur Avg Loss: 31.52385616, Log Avg loss: 1.74420568, Global Avg Loss: 31.52385616, Time: 0.0209 Steps: 930, Updated lr: 0.000093 Training, Epoch: 0001, Batch: 000940, Sample Num: 15040, Cur Loss: 2.33668351, Cur Avg Loss: 31.20951559, Log Avg loss: 1.97584199, Global Avg Loss: 31.20951559, Time: 0.0209 Steps: 940, Updated lr: 0.000094 Training, Epoch: 0001, Batch: 000950, Sample Num: 15200, Cur Loss: 2.37254643, Cur Avg Loss: 30.89940607, Log Avg loss: 1.74911183, Global Avg Loss: 30.89940607, Time: 0.0209 Steps: 950, Updated lr: 0.000095 Training, Epoch: 0001, Batch: 000960, Sample Num: 15360, Cur Loss: 2.73151731, Cur Avg Loss: 30.59596372, Log Avg loss: 1.76894062, Global Avg Loss: 30.59596372, Time: 0.0208 Steps: 960, Updated lr: 0.000096 Training, Epoch: 0001, Batch: 000970, Sample Num: 15520, Cur Loss: 1.56146216, Cur Avg Loss: 30.30033774, Log Avg loss: 1.92024378, Global Avg Loss: 30.30033774, Time: 0.0208 Steps: 970, Updated lr: 0.000097 Training, Epoch: 0001, Batch: 000980, Sample Num: 15680, Cur Loss: 1.67005587, Cur Avg Loss: 30.01052551, Log Avg loss: 1.89873834, Global Avg Loss: 30.01052551, Time: 0.0208 Steps: 980, Updated lr: 0.000098 Training, Epoch: 0001, Batch: 000990, Sample Num: 15840, Cur Loss: 2.19963074, Cur Avg Loss: 29.72580415, Log Avg loss: 1.82311133, Global Avg Loss: 29.72580415, Time: 0.0208 Steps: 990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001000, Sample Num: 16000, Cur Loss: 1.15697908, Cur Avg Loss: 29.44368161, Log Avg loss: 1.51354981, Global Avg Loss: 29.44368161, Time: 0.0208 Steps: 1000, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001010, Sample Num: 16160, Cur Loss: 1.27867341, Cur Avg Loss: 29.16881896, Log Avg loss: 1.68255457, Global Avg Loss: 29.16881896, Time: 0.0208 Steps: 1010, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001020, Sample Num: 16320, Cur Loss: 1.26083636, Cur Avg Loss: 28.89686633, Log Avg loss: 1.42965088, Global Avg Loss: 28.89686633, Time: 0.0208 Steps: 1020, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001030, Sample Num: 16480, Cur Loss: 1.29948819, Cur Avg Loss: 28.62938333, Log Avg loss: 1.34611709, Global Avg Loss: 28.62938333, Time: 0.0208 Steps: 1030, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001040, Sample Num: 16640, Cur Loss: 1.31076944, Cur Avg Loss: 28.36640225, Log Avg loss: 1.27935103, Global Avg Loss: 28.36640225, Time: 0.0209 Steps: 1040, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001050, Sample Num: 16800, Cur Loss: 1.47584701, Cur Avg Loss: 28.11274875, Log Avg loss: 1.73278455, Global Avg Loss: 28.11274875, Time: 0.0209 Steps: 1050, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001060, Sample Num: 16960, Cur Loss: 2.09165525, Cur Avg Loss: 27.86112363, Log Avg loss: 1.44048583, Global Avg Loss: 27.86112363, Time: 0.0209 Steps: 1060, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001070, Sample Num: 17120, Cur Loss: 3.04572082, Cur Avg Loss: 27.62186802, Log Avg loss: 2.26077347, Global Avg Loss: 27.62186802, Time: 0.0209 Steps: 1070, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001080, Sample Num: 17280, Cur Loss: 3.03870988, Cur Avg Loss: 27.38439714, Log Avg loss: 1.97501329, Global Avg Loss: 27.38439714, Time: 0.0209 Steps: 1080, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001090, Sample Num: 17440, Cur Loss: 2.59722733, Cur Avg Loss: 27.15072399, Log Avg loss: 1.91402307, Global Avg Loss: 27.15072399, Time: 0.0209 Steps: 1090, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001100, Sample Num: 17600, Cur Loss: 1.82216549, Cur Avg Loss: 26.92017198, Log Avg loss: 1.79000306, Global Avg Loss: 26.92017198, Time: 0.0209 Steps: 1100, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001110, Sample Num: 17760, Cur Loss: 0.90059566, Cur Avg Loss: 26.69157803, Log Avg loss: 1.54624428, Global Avg Loss: 26.69157803, Time: 0.0209 Steps: 1110, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001120, Sample Num: 17920, Cur Loss: 0.93404162, Cur Avg Loss: 26.46942939, Log Avg loss: 1.81092998, Global Avg Loss: 26.46942939, Time: 0.0209 Steps: 1120, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001130, Sample Num: 18080, Cur Loss: 1.67308283, Cur Avg Loss: 26.25026101, Log Avg loss: 1.70340177, Global Avg Loss: 26.25026101, Time: 0.0208 Steps: 1130, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001140, Sample Num: 18240, Cur Loss: 2.63046908, Cur Avg Loss: 26.03210563, Log Avg loss: 1.38054774, Global Avg Loss: 26.03210563, Time: 0.0209 Steps: 1140, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001150, Sample Num: 18400, Cur Loss: 1.86818290, Cur Avg Loss: 25.82045790, Log Avg loss: 1.69261676, Global Avg Loss: 25.82045790, Time: 0.0208 Steps: 1150, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001160, Sample Num: 18560, Cur Loss: 2.01221275, Cur Avg Loss: 25.61182786, Log Avg loss: 1.61937393, Global Avg Loss: 25.61182786, Time: 0.0209 Steps: 1160, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001170, Sample Num: 18720, Cur Loss: 1.36737943, Cur Avg Loss: 25.40439158, Log Avg loss: 1.34178250, Global Avg Loss: 25.40439158, Time: 0.0209 Steps: 1170, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001180, Sample Num: 18880, Cur Loss: 1.13491154, Cur Avg Loss: 25.20470184, Log Avg loss: 1.84100268, Global Avg Loss: 25.20470184, Time: 0.0209 Steps: 1180, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001190, Sample Num: 19040, Cur Loss: 1.13066483, Cur Avg Loss: 25.00451731, Log Avg loss: 1.38274326, Global Avg Loss: 25.00451731, Time: 0.0208 Steps: 1190, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001200, Sample Num: 19200, Cur Loss: 1.72852659, Cur Avg Loss: 24.80851985, Log Avg loss: 1.48482113, Global Avg Loss: 24.80851985, Time: 0.0209 Steps: 1200, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001210, Sample Num: 19360, Cur Loss: 0.50360060, Cur Avg Loss: 24.61410866, Log Avg loss: 1.28476683, Global Avg Loss: 24.61410866, Time: 0.0209 Steps: 1210, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001220, Sample Num: 19520, Cur Loss: 0.99967039, Cur Avg Loss: 24.42490070, Log Avg loss: 1.53073678, Global Avg Loss: 24.42490070, Time: 0.0209 Steps: 1220, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001230, Sample Num: 19680, Cur Loss: 1.42558074, Cur Avg Loss: 24.23885707, Log Avg loss: 1.54153488, Global Avg Loss: 24.23885707, Time: 0.0209 Steps: 1230, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001240, Sample Num: 19840, Cur Loss: 1.38042009, Cur Avg Loss: 24.05649807, Log Avg loss: 1.62634063, Global Avg Loss: 24.05649807, Time: 0.0208 Steps: 1240, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001250, Sample Num: 20000, Cur Loss: 1.45466316, Cur Avg Loss: 23.87419867, Log Avg loss: 1.26907319, Global Avg Loss: 23.87419867, Time: 0.0209 Steps: 1250, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001260, Sample Num: 20160, Cur Loss: 0.93682259, Cur Avg Loss: 23.69288998, Log Avg loss: 1.02930341, Global Avg Loss: 23.69288998, Time: 0.0209 Steps: 1260, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001270, Sample Num: 20320, Cur Loss: 2.05364466, Cur Avg Loss: 23.51683839, Log Avg loss: 1.33433887, Global Avg Loss: 23.51683839, Time: 0.0209 Steps: 1270, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001280, Sample Num: 20480, Cur Loss: 1.49225390, Cur Avg Loss: 23.34720703, Log Avg loss: 1.80402408, Global Avg Loss: 23.34720703, Time: 0.0216 Steps: 1280, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001290, Sample Num: 20640, Cur Loss: 0.82483667, Cur Avg Loss: 23.17603104, Log Avg loss: 1.26550390, Global Avg Loss: 23.17603104, Time: 0.0209 Steps: 1290, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001300, Sample Num: 20800, Cur Loss: 2.16935849, Cur Avg Loss: 23.00991562, Log Avg loss: 1.58102634, Global Avg Loss: 23.00991562, Time: 0.0209 Steps: 1300, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001310, Sample Num: 20960, Cur Loss: 0.74942482, Cur Avg Loss: 22.84236015, Log Avg loss: 1.06014883, Global Avg Loss: 22.84236015, Time: 0.0208 Steps: 1310, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001320, Sample Num: 21120, Cur Loss: 1.56846833, Cur Avg Loss: 22.68196076, Log Avg loss: 1.66964143, Global Avg Loss: 22.68196076, Time: 0.0208 Steps: 1320, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001330, Sample Num: 21280, Cur Loss: 0.54681849, Cur Avg Loss: 22.52375712, Log Avg loss: 1.64087626, Global Avg Loss: 22.52375712, Time: 0.0208 Steps: 1330, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001340, Sample Num: 21440, Cur Loss: 1.56666636, Cur Avg Loss: 22.36512889, Log Avg loss: 1.26757389, Global Avg Loss: 22.36512889, Time: 0.0208 Steps: 1340, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001350, Sample Num: 21600, Cur Loss: 1.08392239, Cur Avg Loss: 22.21033835, Log Avg loss: 1.46840659, Global Avg Loss: 22.21033835, Time: 0.0208 Steps: 1350, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001360, Sample Num: 21760, Cur Loss: 1.66378391, Cur Avg Loss: 22.06029287, Log Avg loss: 1.80415338, Global Avg Loss: 22.06029287, Time: 0.0208 Steps: 1360, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001370, Sample Num: 21920, Cur Loss: 1.12991214, Cur Avg Loss: 21.91183688, Log Avg loss: 1.72182131, Global Avg Loss: 21.91183688, Time: 0.0211 Steps: 1370, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001380, Sample Num: 22080, Cur Loss: 1.66844201, Cur Avg Loss: 21.76590948, Log Avg loss: 1.77385658, Global Avg Loss: 21.76590948, Time: 0.0208 Steps: 1380, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001390, Sample Num: 22240, Cur Loss: 1.98988199, Cur Avg Loss: 21.62185620, Log Avg loss: 1.74250252, Global Avg Loss: 21.62185620, Time: 0.0208 Steps: 1390, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001400, Sample Num: 22400, Cur Loss: 2.20607853, Cur Avg Loss: 21.47950334, Log Avg loss: 1.69245647, Global Avg Loss: 21.47950334, Time: 0.0208 Steps: 1400, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001410, Sample Num: 22560, Cur Loss: 1.10070562, Cur Avg Loss: 21.33654407, Log Avg loss: 1.32224665, Global Avg Loss: 21.33654407, Time: 0.0208 Steps: 1410, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001420, Sample Num: 22720, Cur Loss: 1.27701735, Cur Avg Loss: 21.19667195, Log Avg loss: 1.47470186, Global Avg Loss: 21.19667195, Time: 0.0211 Steps: 1420, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001430, Sample Num: 22880, Cur Loss: 1.42393279, Cur Avg Loss: 21.06190688, Log Avg loss: 1.92526709, Global Avg Loss: 21.06190688, Time: 0.0208 Steps: 1430, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001440, Sample Num: 23040, Cur Loss: 1.13162732, Cur Avg Loss: 20.92679435, Log Avg loss: 1.60570245, Global Avg Loss: 20.92679435, Time: 0.0208 Steps: 1440, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001450, Sample Num: 23200, Cur Loss: 0.45979196, Cur Avg Loss: 20.79289945, Log Avg loss: 1.51203417, Global Avg Loss: 20.79289945, Time: 0.0208 Steps: 1450, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001460, Sample Num: 23360, Cur Loss: 1.93753648, Cur Avg Loss: 20.66180221, Log Avg loss: 1.65270266, Global Avg Loss: 20.66180221, Time: 0.0209 Steps: 1460, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001470, Sample Num: 23520, Cur Loss: 2.65495682, Cur Avg Loss: 20.52963166, Log Avg loss: 1.23273090, Global Avg Loss: 20.52963166, Time: 0.0209 Steps: 1470, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001480, Sample Num: 23680, Cur Loss: 0.90870464, Cur Avg Loss: 20.40009005, Log Avg loss: 1.35747340, Global Avg Loss: 20.40009005, Time: 0.0208 Steps: 1480, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001490, Sample Num: 23840, Cur Loss: 1.68104696, Cur Avg Loss: 20.27289135, Log Avg loss: 1.44748482, Global Avg Loss: 20.27289135, Time: 0.0209 Steps: 1490, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001500, Sample Num: 24000, Cur Loss: 1.83267927, Cur Avg Loss: 20.14898387, Log Avg loss: 1.68676919, Global Avg Loss: 20.14898387, Time: 0.0209 Steps: 1500, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001510, Sample Num: 24160, Cur Loss: 0.84980619, Cur Avg Loss: 20.02586287, Log Avg loss: 1.55771245, Global Avg Loss: 20.02586287, Time: 0.0208 Steps: 1510, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001520, Sample Num: 24320, Cur Loss: 1.04702616, Cur Avg Loss: 19.90463512, Log Avg loss: 1.59924530, Global Avg Loss: 19.90463512, Time: 0.0209 Steps: 1520, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001530, Sample Num: 24480, Cur Loss: 0.73890305, Cur Avg Loss: 19.78279960, Log Avg loss: 1.26380066, Global Avg Loss: 19.78279960, Time: 0.0208 Steps: 1530, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001540, Sample Num: 24640, Cur Loss: 2.19670057, Cur Avg Loss: 19.66632131, Log Avg loss: 1.84514188, Global Avg Loss: 19.66632131, Time: 0.0210 Steps: 1540, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001550, Sample Num: 24800, Cur Loss: 1.98421788, Cur Avg Loss: 19.54921391, Log Avg loss: 1.51467525, Global Avg Loss: 19.54921391, Time: 0.0209 Steps: 1550, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001560, Sample Num: 24960, Cur Loss: 0.66709214, Cur Avg Loss: 19.43298078, Log Avg loss: 1.41684508, Global Avg Loss: 19.43298078, Time: 0.0209 Steps: 1560, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001570, Sample Num: 25120, Cur Loss: 1.40973306, Cur Avg Loss: 19.31773254, Log Avg loss: 1.33900656, Global Avg Loss: 19.31773254, Time: 0.0210 Steps: 1570, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001580, Sample Num: 25280, Cur Loss: 1.07689846, Cur Avg Loss: 19.20422634, Log Avg loss: 1.38375341, Global Avg Loss: 19.20422634, Time: 0.0209 Steps: 1580, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001590, Sample Num: 25440, Cur Loss: 1.13733280, Cur Avg Loss: 19.09152459, Log Avg loss: 1.28464811, Global Avg Loss: 19.09152459, Time: 0.0210 Steps: 1590, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001600, Sample Num: 25600, Cur Loss: 1.39553869, Cur Avg Loss: 18.97992099, Log Avg loss: 1.23494913, Global Avg Loss: 18.97992099, Time: 0.0209 Steps: 1600, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001610, Sample Num: 25760, Cur Loss: 1.71658754, Cur Avg Loss: 18.87093556, Log Avg loss: 1.43326578, Global Avg Loss: 18.87093556, Time: 0.0209 Steps: 1610, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001620, Sample Num: 25920, Cur Loss: 1.11835575, Cur Avg Loss: 18.76408826, Log Avg loss: 1.56167433, Global Avg Loss: 18.76408826, Time: 0.0209 Steps: 1620, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001630, Sample Num: 26080, Cur Loss: 1.08374894, Cur Avg Loss: 18.65542795, Log Avg loss: 1.05245720, Global Avg Loss: 18.65542795, Time: 0.0209 Steps: 1630, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001640, Sample Num: 26240, Cur Loss: 1.49111688, Cur Avg Loss: 18.54952572, Log Avg loss: 1.28746256, Global Avg Loss: 18.54952572, Time: 0.0208 Steps: 1640, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001650, Sample Num: 26400, Cur Loss: 1.03977168, Cur Avg Loss: 18.44410617, Log Avg loss: 1.15529862, Global Avg Loss: 18.44410617, Time: 0.0208 Steps: 1650, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001660, Sample Num: 26560, Cur Loss: 1.81056511, Cur Avg Loss: 18.34237581, Log Avg loss: 1.55686641, Global Avg Loss: 18.34237581, Time: 0.0209 Steps: 1660, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001670, Sample Num: 26720, Cur Loss: 1.24357569, Cur Avg Loss: 18.24074603, Log Avg loss: 1.37020332, Global Avg Loss: 18.24074603, Time: 0.0208 Steps: 1670, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001680, Sample Num: 26880, Cur Loss: 1.85424531, Cur Avg Loss: 18.14055011, Log Avg loss: 1.40783140, Global Avg Loss: 18.14055011, Time: 0.0209 Steps: 1680, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001690, Sample Num: 27040, Cur Loss: 1.62286782, Cur Avg Loss: 18.04345651, Log Avg loss: 1.73173184, Global Avg Loss: 18.04345651, Time: 0.0210 Steps: 1690, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001700, Sample Num: 27200, Cur Loss: 1.50360179, Cur Avg Loss: 17.94780689, Log Avg loss: 1.78302038, Global Avg Loss: 17.94780689, Time: 0.0208 Steps: 1700, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001710, Sample Num: 27360, Cur Loss: 1.02114201, Cur Avg Loss: 17.85122677, Log Avg loss: 1.43260742, Global Avg Loss: 17.85122677, Time: 0.0209 Steps: 1710, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001720, Sample Num: 27520, Cur Loss: 0.79776102, Cur Avg Loss: 17.75415528, Log Avg loss: 1.15493023, Global Avg Loss: 17.75415528, Time: 0.0210 Steps: 1720, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001730, Sample Num: 27680, Cur Loss: 1.43924010, Cur Avg Loss: 17.65973160, Log Avg loss: 1.41885938, Global Avg Loss: 17.65973160, Time: 0.0209 Steps: 1730, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001740, Sample Num: 27840, Cur Loss: 1.42183495, Cur Avg Loss: 17.56479482, Log Avg loss: 1.14073076, Global Avg Loss: 17.56479482, Time: 0.0209 Steps: 1740, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001750, Sample Num: 28000, Cur Loss: 1.42849517, Cur Avg Loss: 17.47233935, Log Avg loss: 1.38508872, Global Avg Loss: 17.47233935, Time: 0.0209 Steps: 1750, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001760, Sample Num: 28160, Cur Loss: 1.76912332, Cur Avg Loss: 17.38168753, Log Avg loss: 1.51761830, Global Avg Loss: 17.38168753, Time: 0.0209 Steps: 1760, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001770, Sample Num: 28320, Cur Loss: 1.89753127, Cur Avg Loss: 17.29067935, Log Avg loss: 1.27324026, Global Avg Loss: 17.29067935, Time: 0.0209 Steps: 1770, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001780, Sample Num: 28480, Cur Loss: 1.07605410, Cur Avg Loss: 17.20001361, Log Avg loss: 1.15217701, Global Avg Loss: 17.20001361, Time: 0.0208 Steps: 1780, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001790, Sample Num: 28640, Cur Loss: 1.17262137, Cur Avg Loss: 17.11126868, Log Avg loss: 1.31467130, Global Avg Loss: 17.11126868, Time: 0.0210 Steps: 1790, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001800, Sample Num: 28800, Cur Loss: 1.28502965, Cur Avg Loss: 17.02238856, Log Avg loss: 1.11284709, Global Avg Loss: 17.02238856, Time: 0.0209 Steps: 1800, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001810, Sample Num: 28960, Cur Loss: 1.37997413, Cur Avg Loss: 16.93603876, Log Avg loss: 1.39307539, Global Avg Loss: 16.93603876, Time: 0.0209 Steps: 1810, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001820, Sample Num: 29120, Cur Loss: 1.44723141, Cur Avg Loss: 16.85011125, Log Avg loss: 1.29723115, Global Avg Loss: 16.85011125, Time: 0.0209 Steps: 1820, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001830, Sample Num: 29280, Cur Loss: 0.88447982, Cur Avg Loss: 16.76487014, Log Avg loss: 1.25098831, Global Avg Loss: 16.76487014, Time: 0.0209 Steps: 1830, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001840, Sample Num: 29440, Cur Loss: 1.13566387, Cur Avg Loss: 16.68114386, Log Avg loss: 1.35923370, Global Avg Loss: 16.68114386, Time: 0.0209 Steps: 1840, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001850, Sample Num: 29600, Cur Loss: 2.11960077, Cur Avg Loss: 16.59852442, Log Avg loss: 1.39654756, Global Avg Loss: 16.59852442, Time: 0.0210 Steps: 1850, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001860, Sample Num: 29760, Cur Loss: 1.38011551, Cur Avg Loss: 16.51670412, Log Avg loss: 1.37994952, Global Avg Loss: 16.51670412, Time: 0.0209 Steps: 1860, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001870, Sample Num: 29920, Cur Loss: 1.40423656, Cur Avg Loss: 16.43627025, Log Avg loss: 1.47556916, Global Avg Loss: 16.43627025, Time: 0.0209 Steps: 1870, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001880, Sample Num: 30080, Cur Loss: 1.43085265, Cur Avg Loss: 16.35547247, Log Avg loss: 1.24628846, Global Avg Loss: 16.35547247, Time: 0.0209 Steps: 1880, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001890, Sample Num: 30240, Cur Loss: 0.65381712, Cur Avg Loss: 16.27544249, Log Avg loss: 1.22980724, Global Avg Loss: 16.27544249, Time: 0.0209 Steps: 1890, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001900, Sample Num: 30400, Cur Loss: 1.28969836, Cur Avg Loss: 16.19646887, Log Avg loss: 1.27045301, Global Avg Loss: 16.19646887, Time: 0.0209 Steps: 1900, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001910, Sample Num: 30560, Cur Loss: 1.27622294, Cur Avg Loss: 16.11992175, Log Avg loss: 1.57597067, Global Avg Loss: 16.11992175, Time: 0.0209 Steps: 1910, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001920, Sample Num: 30720, Cur Loss: 0.59249133, Cur Avg Loss: 16.04649523, Log Avg loss: 2.02202918, Global Avg Loss: 16.04649523, Time: 0.0209 Steps: 1920, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001930, Sample Num: 30880, Cur Loss: 0.81801653, Cur Avg Loss: 15.97215188, Log Avg loss: 1.69822857, Global Avg Loss: 15.97215188, Time: 0.0209 Steps: 1930, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001940, Sample Num: 31040, Cur Loss: 0.80289245, Cur Avg Loss: 15.89594297, Log Avg loss: 1.18762255, Global Avg Loss: 15.89594297, Time: 0.0209 Steps: 1940, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001950, Sample Num: 31200, Cur Loss: 1.20318818, Cur Avg Loss: 15.82185231, Log Avg loss: 1.44826581, Global Avg Loss: 15.82185231, Time: 0.0210 Steps: 1950, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001960, Sample Num: 31360, Cur Loss: 1.31866097, Cur Avg Loss: 15.74854374, Log Avg loss: 1.45337090, Global Avg Loss: 15.74854374, Time: 0.0209 Steps: 1960, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001970, Sample Num: 31520, Cur Loss: 1.32100725, Cur Avg Loss: 15.67521784, Log Avg loss: 1.30334246, Global Avg Loss: 15.67521784, Time: 0.0209 Steps: 1970, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001980, Sample Num: 31680, Cur Loss: 1.24990702, Cur Avg Loss: 15.60243959, Log Avg loss: 1.26512368, Global Avg Loss: 15.60243959, Time: 0.0209 Steps: 1980, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001990, Sample Num: 31840, Cur Loss: 1.24389887, Cur Avg Loss: 15.53133255, Log Avg loss: 1.45213883, Global Avg Loss: 15.53133255, Time: 0.0209 Steps: 1990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002000, Sample Num: 32000, Cur Loss: 0.83788699, Cur Avg Loss: 15.46058046, Log Avg loss: 1.38091490, Global Avg Loss: 15.46058046, Time: 0.0209 Steps: 2000, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002010, Sample Num: 32160, Cur Loss: 1.58568060, Cur Avg Loss: 15.39016605, Log Avg loss: 1.30728487, Global Avg Loss: 15.39016605, Time: 0.0210 Steps: 2010, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002020, Sample Num: 32320, Cur Loss: 1.33368206, Cur Avg Loss: 15.31975305, Log Avg loss: 1.16673937, Global Avg Loss: 15.31975305, Time: 0.0209 Steps: 2020, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002030, Sample Num: 32480, Cur Loss: 0.92358458, Cur Avg Loss: 15.25080824, Log Avg loss: 1.32395592, Global Avg Loss: 15.25080824, Time: 0.0209 Steps: 2030, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002040, Sample Num: 32640, Cur Loss: 1.28570032, Cur Avg Loss: 15.18460893, Log Avg loss: 1.74615000, Global Avg Loss: 15.18460893, Time: 0.0209 Steps: 2040, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002050, Sample Num: 32800, Cur Loss: 2.55289245, Cur Avg Loss: 15.11921904, Log Avg loss: 1.77968179, Global Avg Loss: 15.11921904, Time: 0.0248 Steps: 2050, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002060, Sample Num: 32960, Cur Loss: 1.38438272, Cur Avg Loss: 15.05246322, Log Avg loss: 1.36751990, Global Avg Loss: 15.05246322, Time: 0.0209 Steps: 2060, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002070, Sample Num: 33120, Cur Loss: 1.41595733, Cur Avg Loss: 14.98617373, Log Avg loss: 1.33053846, Global Avg Loss: 14.98617373, Time: 0.0209 Steps: 2070, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002080, Sample Num: 33280, Cur Loss: 0.88645989, Cur Avg Loss: 14.91964665, Log Avg loss: 1.14854153, Global Avg Loss: 14.91964665, Time: 0.0209 Steps: 2080, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002090, Sample Num: 33440, Cur Loss: 0.93306959, Cur Avg Loss: 14.85315566, Log Avg loss: 1.02303002, Global Avg Loss: 14.85315566, Time: 0.0209 Steps: 2090, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002100, Sample Num: 33600, Cur Loss: 0.70076561, Cur Avg Loss: 14.78791563, Log Avg loss: 1.15274770, Global Avg Loss: 14.78791563, Time: 0.0209 Steps: 2100, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002110, Sample Num: 33760, Cur Loss: 0.83419645, Cur Avg Loss: 14.72440625, Log Avg loss: 1.38743658, Global Avg Loss: 14.72440625, Time: 0.0210 Steps: 2110, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002120, Sample Num: 33920, Cur Loss: 1.57326794, Cur Avg Loss: 14.66155677, Log Avg loss: 1.40031674, Global Avg Loss: 14.66155677, Time: 0.0208 Steps: 2120, Updated lr: 0.000099 ***** Running evaluation checkpoint-2129 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-2129 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.494986, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.112756, "eval_total_loss": 782.267326, "eval_mae": 0.878302, "eval_mse": 1.113, "eval_r2": 0.292504, "eval_sp_statistic": 0.572896, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.580883, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.935873, "test_total_loss": 971.808235, "test_mae": 1.313614, "test_mse": 1.936021, "test_r2": -0.249525, "test_sp_statistic": 0.557734, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.584579, "test_ps_pvalue": 0.0, "lr": 9.892935040303462e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 14.605059143787486, "train_cur_epoch_loss": 31094.170917123556, "train_cur_epoch_avg_loss": 14.605059143787486, "train_cur_epoch_time": 44.49498629570007, "train_cur_epoch_avg_time": 0.020899476888539256, "epoch": 1, "step": 2129} ################################################## Training, Epoch: 0002, Batch: 000001, Sample Num: 16, Cur Loss: 1.28996432, Cur Avg Loss: 1.28996432, Log Avg loss: 1.29605317, Global Avg Loss: 14.59880793, Time: 0.0247 Steps: 2130, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000011, Sample Num: 176, Cur Loss: 1.26184273, Cur Avg Loss: 1.24693578, Log Avg loss: 1.24263293, Global Avg Loss: 14.53639589, Time: 0.0209 Steps: 2140, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000021, Sample Num: 336, Cur Loss: 2.60817957, Cur Avg Loss: 1.22903412, Log Avg loss: 1.20934230, Global Avg Loss: 14.47440960, Time: 0.0208 Steps: 2150, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000031, Sample Num: 496, Cur Loss: 1.81722271, Cur Avg Loss: 1.24263665, Log Avg loss: 1.27120194, Global Avg Loss: 14.41328364, Time: 0.0208 Steps: 2160, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000041, Sample Num: 656, Cur Loss: 1.54999340, Cur Avg Loss: 1.27579614, Log Avg loss: 1.37859056, Global Avg Loss: 14.35321593, Time: 0.0209 Steps: 2170, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000051, Sample Num: 816, Cur Loss: 0.41372967, Cur Avg Loss: 1.20549126, Log Avg loss: 0.91724125, Global Avg Loss: 14.29158301, Time: 0.0208 Steps: 2180, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000061, Sample Num: 976, Cur Loss: 1.13470042, Cur Avg Loss: 1.20218195, Log Avg loss: 1.18530448, Global Avg Loss: 14.23173699, Time: 0.0209 Steps: 2190, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000071, Sample Num: 1136, Cur Loss: 1.09591877, Cur Avg Loss: 1.21059436, Log Avg loss: 1.26191006, Global Avg Loss: 14.17278323, Time: 0.0208 Steps: 2200, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000081, Sample Num: 1296, Cur Loss: 1.51232290, Cur Avg Loss: 1.23680014, Log Avg loss: 1.42286116, Global Avg Loss: 14.11509128, Time: 0.0208 Steps: 2210, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000091, Sample Num: 1456, Cur Loss: 1.55121756, Cur Avg Loss: 1.29155017, Log Avg loss: 1.73502545, Global Avg Loss: 14.05932522, Time: 0.0209 Steps: 2220, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000101, Sample Num: 1616, Cur Loss: 1.17206264, Cur Avg Loss: 1.29365588, Log Avg loss: 1.31281784, Global Avg Loss: 14.00216599, Time: 0.0208 Steps: 2230, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000111, Sample Num: 1776, Cur Loss: 1.21441364, Cur Avg Loss: 1.29970710, Log Avg loss: 1.36082439, Global Avg Loss: 13.94573143, Time: 0.0209 Steps: 2240, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000121, Sample Num: 1936, Cur Loss: 0.67291462, Cur Avg Loss: 1.27245766, Log Avg loss: 0.96998895, Global Avg Loss: 13.88806146, Time: 0.0209 Steps: 2250, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000131, Sample Num: 2096, Cur Loss: 1.24011564, Cur Avg Loss: 1.28157702, Log Avg loss: 1.39192125, Global Avg Loss: 13.83276881, Time: 0.0208 Steps: 2260, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000141, Sample Num: 2256, Cur Loss: 1.75667453, Cur Avg Loss: 1.27328681, Log Avg loss: 1.16468506, Global Avg Loss: 13.77696227, Time: 0.0208 Steps: 2270, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000151, Sample Num: 2416, Cur Loss: 1.52306592, Cur Avg Loss: 1.28209923, Log Avg loss: 1.40635433, Global Avg Loss: 13.72270522, Time: 0.0209 Steps: 2280, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000161, Sample Num: 2576, Cur Loss: 1.55095434, Cur Avg Loss: 1.28111644, Log Avg loss: 1.26627625, Global Avg Loss: 13.66831033, Time: 0.0209 Steps: 2290, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000171, Sample Num: 2736, Cur Loss: 2.26639438, Cur Avg Loss: 1.28777137, Log Avg loss: 1.39491577, Global Avg Loss: 13.61494775, Time: 0.0209 Steps: 2300, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000181, Sample Num: 2896, Cur Loss: 1.08727729, Cur Avg Loss: 1.28881839, Log Avg loss: 1.30672244, Global Avg Loss: 13.56166539, Time: 0.0208 Steps: 2310, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000191, Sample Num: 3056, Cur Loss: 1.30870128, Cur Avg Loss: 1.28000480, Log Avg loss: 1.12047890, Global Avg Loss: 13.50803958, Time: 0.0208 Steps: 2320, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000201, Sample Num: 3216, Cur Loss: 1.20891929, Cur Avg Loss: 1.28967269, Log Avg loss: 1.47432943, Global Avg Loss: 13.45639276, Time: 0.0208 Steps: 2330, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000211, Sample Num: 3376, Cur Loss: 1.26978505, Cur Avg Loss: 1.28142352, Log Avg loss: 1.11561519, Global Avg Loss: 13.40365439, Time: 0.0209 Steps: 2340, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000221, Sample Num: 3536, Cur Loss: 1.18662155, Cur Avg Loss: 1.27710626, Log Avg loss: 1.18601206, Global Avg Loss: 13.35166443, Time: 0.0208 Steps: 2350, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000231, Sample Num: 3696, Cur Loss: 0.35454008, Cur Avg Loss: 1.27133957, Log Avg loss: 1.14389555, Global Avg Loss: 13.29993659, Time: 0.0208 Steps: 2360, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000241, Sample Num: 3856, Cur Loss: 1.20790386, Cur Avg Loss: 1.26706427, Log Avg loss: 1.16830492, Global Avg Loss: 13.24874827, Time: 0.0209 Steps: 2370, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000251, Sample Num: 4016, Cur Loss: 0.45222917, Cur Avg Loss: 1.26177908, Log Avg loss: 1.13440599, Global Avg Loss: 13.19784767, Time: 0.0208 Steps: 2380, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000261, Sample Num: 4176, Cur Loss: 0.94046438, Cur Avg Loss: 1.25954907, Log Avg loss: 1.20357590, Global Avg Loss: 13.14766244, Time: 0.0210 Steps: 2390, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000271, Sample Num: 4336, Cur Loss: 2.25240374, Cur Avg Loss: 1.25716570, Log Avg loss: 1.19495967, Global Avg Loss: 13.09785951, Time: 0.0209 Steps: 2400, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000281, Sample Num: 4496, Cur Loss: 0.62161756, Cur Avg Loss: 1.25757153, Log Avg loss: 1.26856954, Global Avg Loss: 13.04877532, Time: 0.0208 Steps: 2410, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000291, Sample Num: 4656, Cur Loss: 2.14446068, Cur Avg Loss: 1.26619801, Log Avg loss: 1.50860209, Global Avg Loss: 13.00108865, Time: 0.0209 Steps: 2420, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000301, Sample Num: 4816, Cur Loss: 0.79105240, Cur Avg Loss: 1.25828456, Log Avg loss: 1.02800331, Global Avg Loss: 12.95181670, Time: 0.0208 Steps: 2430, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000311, Sample Num: 4976, Cur Loss: 1.64723349, Cur Avg Loss: 1.26068336, Log Avg loss: 1.33288715, Global Avg Loss: 12.90419813, Time: 0.0209 Steps: 2440, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000321, Sample Num: 5136, Cur Loss: 0.93741465, Cur Avg Loss: 1.25601678, Log Avg loss: 1.11088598, Global Avg Loss: 12.85606216, Time: 0.0209 Steps: 2450, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000331, Sample Num: 5296, Cur Loss: 2.24440145, Cur Avg Loss: 1.26333750, Log Avg loss: 1.49833272, Global Avg Loss: 12.80989253, Time: 0.0209 Steps: 2460, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000341, Sample Num: 5456, Cur Loss: 1.29561186, Cur Avg Loss: 1.26471829, Log Avg loss: 1.31042238, Global Avg Loss: 12.76333597, Time: 0.0209 Steps: 2470, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000351, Sample Num: 5616, Cur Loss: 1.04878759, Cur Avg Loss: 1.26359335, Log Avg loss: 1.22523301, Global Avg Loss: 12.71681136, Time: 0.0208 Steps: 2480, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000361, Sample Num: 5776, Cur Loss: 1.01336610, Cur Avg Loss: 1.25304267, Log Avg loss: 0.88271361, Global Avg Loss: 12.66928487, Time: 0.0208 Steps: 2490, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000371, Sample Num: 5936, Cur Loss: 1.09599042, Cur Avg Loss: 1.25985084, Log Avg loss: 1.50562600, Global Avg Loss: 12.62463023, Time: 0.0208 Steps: 2500, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000381, Sample Num: 6096, Cur Loss: 1.34178460, Cur Avg Loss: 1.25490632, Log Avg loss: 1.07146466, Global Avg Loss: 12.57860168, Time: 0.0208 Steps: 2510, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000391, Sample Num: 6256, Cur Loss: 1.13817871, Cur Avg Loss: 1.27438703, Log Avg loss: 2.01660206, Global Avg Loss: 12.53668899, Time: 0.0208 Steps: 2520, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000401, Sample Num: 6416, Cur Loss: 0.77098882, Cur Avg Loss: 1.26775606, Log Avg loss: 1.00848495, Global Avg Loss: 12.49112296, Time: 0.0208 Steps: 2530, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000411, Sample Num: 6576, Cur Loss: 1.31880307, Cur Avg Loss: 1.26354626, Log Avg loss: 1.09473339, Global Avg Loss: 12.44625529, Time: 0.0208 Steps: 2540, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000421, Sample Num: 6736, Cur Loss: 0.94790977, Cur Avg Loss: 1.26209507, Log Avg loss: 1.20245124, Global Avg Loss: 12.40216194, Time: 0.0208 Steps: 2550, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000431, Sample Num: 6896, Cur Loss: 0.85898685, Cur Avg Loss: 1.25608791, Log Avg loss: 1.00318633, Global Avg Loss: 12.35763469, Time: 0.0209 Steps: 2560, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000441, Sample Num: 7056, Cur Loss: 0.96058935, Cur Avg Loss: 1.25610593, Log Avg loss: 1.25688258, Global Avg Loss: 12.31444110, Time: 0.0209 Steps: 2570, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000451, Sample Num: 7216, Cur Loss: 1.73976803, Cur Avg Loss: 1.25179836, Log Avg loss: 1.06183472, Global Avg Loss: 12.27082635, Time: 0.0209 Steps: 2580, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000461, Sample Num: 7376, Cur Loss: 0.48652542, Cur Avg Loss: 1.24799716, Log Avg loss: 1.07656300, Global Avg Loss: 12.22760525, Time: 0.0208 Steps: 2590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000471, Sample Num: 7536, Cur Loss: 1.19870389, Cur Avg Loss: 1.24497315, Log Avg loss: 1.10556640, Global Avg Loss: 12.18482818, Time: 0.0208 Steps: 2600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000481, Sample Num: 7696, Cur Loss: 2.21996117, Cur Avg Loss: 1.24386398, Log Avg loss: 1.19162197, Global Avg Loss: 12.14270862, Time: 0.0208 Steps: 2610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000491, Sample Num: 7856, Cur Loss: 1.19613314, Cur Avg Loss: 1.24590079, Log Avg loss: 1.34387127, Global Avg Loss: 12.10149168, Time: 0.0209 Steps: 2620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000501, Sample Num: 8016, Cur Loss: 0.89792931, Cur Avg Loss: 1.24369080, Log Avg loss: 1.13518004, Global Avg Loss: 12.05979468, Time: 0.0208 Steps: 2630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000511, Sample Num: 8176, Cur Loss: 1.34764481, Cur Avg Loss: 1.23681814, Log Avg loss: 0.89249825, Global Avg Loss: 12.01749431, Time: 0.0208 Steps: 2640, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000521, Sample Num: 8336, Cur Loss: 1.56638956, Cur Avg Loss: 1.23687357, Log Avg loss: 1.23970582, Global Avg Loss: 11.97682341, Time: 0.0210 Steps: 2650, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000531, Sample Num: 8496, Cur Loss: 2.16093826, Cur Avg Loss: 1.24119572, Log Avg loss: 1.46637974, Global Avg Loss: 11.93731047, Time: 0.0209 Steps: 2660, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000541, Sample Num: 8656, Cur Loss: 0.90903604, Cur Avg Loss: 1.24178603, Log Avg loss: 1.27313130, Global Avg Loss: 11.89736972, Time: 0.0209 Steps: 2670, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000551, Sample Num: 8816, Cur Loss: 0.85761482, Cur Avg Loss: 1.23557597, Log Avg loss: 0.89961207, Global Avg Loss: 11.85633331, Time: 0.0209 Steps: 2680, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000561, Sample Num: 8976, Cur Loss: 1.96869481, Cur Avg Loss: 1.23701414, Log Avg loss: 1.31625747, Global Avg Loss: 11.81715087, Time: 0.0209 Steps: 2690, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000571, Sample Num: 9136, Cur Loss: 0.86654472, Cur Avg Loss: 1.23720886, Log Avg loss: 1.24813212, Global Avg Loss: 11.77800636, Time: 0.0209 Steps: 2700, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000581, Sample Num: 9296, Cur Loss: 0.86971474, Cur Avg Loss: 1.23917196, Log Avg loss: 1.35126529, Global Avg Loss: 11.73953130, Time: 0.0209 Steps: 2710, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000591, Sample Num: 9456, Cur Loss: 1.30496037, Cur Avg Loss: 1.23570298, Log Avg loss: 1.03415492, Global Avg Loss: 11.70017330, Time: 0.0209 Steps: 2720, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000601, Sample Num: 9616, Cur Loss: 1.13323724, Cur Avg Loss: 1.23432039, Log Avg loss: 1.15260939, Global Avg Loss: 11.66153753, Time: 0.0209 Steps: 2730, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000611, Sample Num: 9776, Cur Loss: 0.60001260, Cur Avg Loss: 1.22837824, Log Avg loss: 0.87125533, Global Avg Loss: 11.62215694, Time: 0.0209 Steps: 2740, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000621, Sample Num: 9936, Cur Loss: 0.50838107, Cur Avg Loss: 1.22461982, Log Avg loss: 0.99498039, Global Avg Loss: 11.58351266, Time: 0.0209 Steps: 2750, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000631, Sample Num: 10096, Cur Loss: 1.09501946, Cur Avg Loss: 1.22893681, Log Avg loss: 1.49702186, Global Avg Loss: 11.54696741, Time: 0.0209 Steps: 2760, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000641, Sample Num: 10256, Cur Loss: 0.17581433, Cur Avg Loss: 1.22553880, Log Avg loss: 1.01112440, Global Avg Loss: 11.50893187, Time: 0.0209 Steps: 2770, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000651, Sample Num: 10416, Cur Loss: 1.01176190, Cur Avg Loss: 1.22290075, Log Avg loss: 1.05380135, Global Avg Loss: 11.47132349, Time: 0.0209 Steps: 2780, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000661, Sample Num: 10576, Cur Loss: 0.82514238, Cur Avg Loss: 1.22110933, Log Avg loss: 1.10448843, Global Avg Loss: 11.43416638, Time: 0.0209 Steps: 2790, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000671, Sample Num: 10736, Cur Loss: 1.63157225, Cur Avg Loss: 1.22769432, Log Avg loss: 1.66296217, Global Avg Loss: 11.39926922, Time: 0.0209 Steps: 2800, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000681, Sample Num: 10896, Cur Loss: 2.09768391, Cur Avg Loss: 1.23238315, Log Avg loss: 1.54700345, Global Avg Loss: 11.36420777, Time: 0.0210 Steps: 2810, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000691, Sample Num: 11056, Cur Loss: 0.75932276, Cur Avg Loss: 1.23877210, Log Avg loss: 1.67385978, Global Avg Loss: 11.32984484, Time: 0.0209 Steps: 2820, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000701, Sample Num: 11216, Cur Loss: 1.33246136, Cur Avg Loss: 1.23901708, Log Avg loss: 1.25594510, Global Avg Loss: 11.29424802, Time: 0.0209 Steps: 2830, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000711, Sample Num: 11376, Cur Loss: 1.38806701, Cur Avg Loss: 1.24016991, Log Avg loss: 1.32098320, Global Avg Loss: 11.25913089, Time: 0.0209 Steps: 2840, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000721, Sample Num: 11536, Cur Loss: 1.28130484, Cur Avg Loss: 1.23856058, Log Avg loss: 1.12413702, Global Avg Loss: 11.22356951, Time: 0.0209 Steps: 2850, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000731, Sample Num: 11696, Cur Loss: 0.83425915, Cur Avg Loss: 1.23351071, Log Avg loss: 0.86941510, Global Avg Loss: 11.18736617, Time: 0.0209 Steps: 2860, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000741, Sample Num: 11856, Cur Loss: 0.72646469, Cur Avg Loss: 1.22866795, Log Avg loss: 0.87466210, Global Avg Loss: 11.15143340, Time: 0.0209 Steps: 2870, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000751, Sample Num: 12016, Cur Loss: 1.22995853, Cur Avg Loss: 1.22767753, Log Avg loss: 1.15428766, Global Avg Loss: 11.11672109, Time: 0.0209 Steps: 2880, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000761, Sample Num: 12176, Cur Loss: 0.46082544, Cur Avg Loss: 1.22271009, Log Avg loss: 0.84965514, Global Avg Loss: 11.08119491, Time: 0.0209 Steps: 2890, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000771, Sample Num: 12336, Cur Loss: 0.97129458, Cur Avg Loss: 1.21926681, Log Avg loss: 0.95723372, Global Avg Loss: 11.04628470, Time: 0.0246 Steps: 2900, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000781, Sample Num: 12496, Cur Loss: 0.99148303, Cur Avg Loss: 1.22085718, Log Avg loss: 1.34347398, Global Avg Loss: 11.01294171, Time: 0.0209 Steps: 2910, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000791, Sample Num: 12656, Cur Loss: 1.10176134, Cur Avg Loss: 1.22023753, Log Avg loss: 1.17184351, Global Avg Loss: 10.97923932, Time: 0.0209 Steps: 2920, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000801, Sample Num: 12816, Cur Loss: 0.41072339, Cur Avg Loss: 1.22155887, Log Avg loss: 1.32607645, Global Avg Loss: 10.94629337, Time: 0.0209 Steps: 2930, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000811, Sample Num: 12976, Cur Loss: 0.75836349, Cur Avg Loss: 1.22123526, Log Avg loss: 1.19531443, Global Avg Loss: 10.91312677, Time: 0.0209 Steps: 2940, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000821, Sample Num: 13136, Cur Loss: 1.00230408, Cur Avg Loss: 1.22625349, Log Avg loss: 1.63323179, Global Avg Loss: 10.88166950, Time: 0.0209 Steps: 2950, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000831, Sample Num: 13296, Cur Loss: 2.56596112, Cur Avg Loss: 1.22844539, Log Avg loss: 1.40840023, Global Avg Loss: 10.84966521, Time: 0.0209 Steps: 2960, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000841, Sample Num: 13456, Cur Loss: 0.76553428, Cur Avg Loss: 1.22402989, Log Avg loss: 0.85710191, Global Avg Loss: 10.81602022, Time: 0.0209 Steps: 2970, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000851, Sample Num: 13616, Cur Loss: 1.35895908, Cur Avg Loss: 1.22235359, Log Avg loss: 1.08137690, Global Avg Loss: 10.78335363, Time: 0.0209 Steps: 2980, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000861, Sample Num: 13776, Cur Loss: 0.69334137, Cur Avg Loss: 1.22051127, Log Avg loss: 1.06372960, Global Avg Loss: 10.75084653, Time: 0.0209 Steps: 2990, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000871, Sample Num: 13936, Cur Loss: 0.92227274, Cur Avg Loss: 1.21336892, Log Avg loss: 0.59841314, Global Avg Loss: 10.71700508, Time: 0.0209 Steps: 3000, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000881, Sample Num: 14096, Cur Loss: 0.98014230, Cur Avg Loss: 1.21388401, Log Avg loss: 1.25874788, Global Avg Loss: 10.68558230, Time: 0.0209 Steps: 3010, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000891, Sample Num: 14256, Cur Loss: 1.31777573, Cur Avg Loss: 1.21236687, Log Avg loss: 1.07870723, Global Avg Loss: 10.65377146, Time: 0.0209 Steps: 3020, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000901, Sample Num: 14416, Cur Loss: 0.45480055, Cur Avg Loss: 1.20704789, Log Avg loss: 0.73312653, Global Avg Loss: 10.62103006, Time: 0.0208 Steps: 3030, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000911, Sample Num: 14576, Cur Loss: 0.97428542, Cur Avg Loss: 1.20739162, Log Avg loss: 1.23836201, Global Avg Loss: 10.59016602, Time: 0.0209 Steps: 3040, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000921, Sample Num: 14736, Cur Loss: 1.30544221, Cur Avg Loss: 1.20760822, Log Avg loss: 1.22734044, Global Avg Loss: 10.55946823, Time: 0.0209 Steps: 3050, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000931, Sample Num: 14896, Cur Loss: 1.43982816, Cur Avg Loss: 1.20614900, Log Avg loss: 1.07175442, Global Avg Loss: 10.52846263, Time: 0.0210 Steps: 3060, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000941, Sample Num: 15056, Cur Loss: 0.55045164, Cur Avg Loss: 1.20391131, Log Avg loss: 0.99558219, Global Avg Loss: 10.49741090, Time: 0.0209 Steps: 3070, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000951, Sample Num: 15216, Cur Loss: 0.87811136, Cur Avg Loss: 1.20286494, Log Avg loss: 1.10440185, Global Avg Loss: 10.46691412, Time: 0.0209 Steps: 3080, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000961, Sample Num: 15376, Cur Loss: 1.36421835, Cur Avg Loss: 1.20239046, Log Avg loss: 1.15726692, Global Avg Loss: 10.43678581, Time: 0.0209 Steps: 3090, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000971, Sample Num: 15536, Cur Loss: 0.89501381, Cur Avg Loss: 1.20033583, Log Avg loss: 1.00288641, Global Avg Loss: 10.40635387, Time: 0.0208 Steps: 3100, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000981, Sample Num: 15696, Cur Loss: 0.62385106, Cur Avg Loss: 1.19956400, Log Avg loss: 1.12461902, Global Avg Loss: 10.37650907, Time: 0.0209 Steps: 3110, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000991, Sample Num: 15856, Cur Loss: 1.75619769, Cur Avg Loss: 1.20205943, Log Avg loss: 1.44686130, Global Avg Loss: 10.34788840, Time: 0.0209 Steps: 3120, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001001, Sample Num: 16016, Cur Loss: 0.59548891, Cur Avg Loss: 1.20181683, Log Avg loss: 1.17777485, Global Avg Loss: 10.31859091, Time: 0.0209 Steps: 3130, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001011, Sample Num: 16176, Cur Loss: 0.97637570, Cur Avg Loss: 1.20158711, Log Avg loss: 1.17859262, Global Avg Loss: 10.28948264, Time: 0.0208 Steps: 3140, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001021, Sample Num: 16336, Cur Loss: 0.87196440, Cur Avg Loss: 1.19995855, Log Avg loss: 1.03531053, Global Avg Loss: 10.26010432, Time: 0.0209 Steps: 3150, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001031, Sample Num: 16496, Cur Loss: 0.87928951, Cur Avg Loss: 1.19826193, Log Avg loss: 1.02503740, Global Avg Loss: 10.23087942, Time: 0.0210 Steps: 3160, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001041, Sample Num: 16656, Cur Loss: 0.93909431, Cur Avg Loss: 1.19551579, Log Avg loss: 0.91238907, Global Avg Loss: 10.20148355, Time: 0.0209 Steps: 3170, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001051, Sample Num: 16816, Cur Loss: 1.59425819, Cur Avg Loss: 1.19647307, Log Avg loss: 1.29612542, Global Avg Loss: 10.17347928, Time: 0.0209 Steps: 3180, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001061, Sample Num: 16976, Cur Loss: 1.12564528, Cur Avg Loss: 1.19498230, Log Avg loss: 1.03830273, Global Avg Loss: 10.14484236, Time: 0.0209 Steps: 3190, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001071, Sample Num: 17136, Cur Loss: 1.48194361, Cur Avg Loss: 1.19185933, Log Avg loss: 0.86051222, Global Avg Loss: 10.11582883, Time: 0.0209 Steps: 3200, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001081, Sample Num: 17296, Cur Loss: 1.39249659, Cur Avg Loss: 1.19099910, Log Avg loss: 1.09886777, Global Avg Loss: 10.08773861, Time: 0.0208 Steps: 3210, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001091, Sample Num: 17456, Cur Loss: 0.72721255, Cur Avg Loss: 1.19104070, Log Avg loss: 1.19553766, Global Avg Loss: 10.06012308, Time: 0.0209 Steps: 3220, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001101, Sample Num: 17616, Cur Loss: 0.98309112, Cur Avg Loss: 1.18922691, Log Avg loss: 0.99134305, Global Avg Loss: 10.03204636, Time: 0.0209 Steps: 3230, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001111, Sample Num: 17776, Cur Loss: 0.56285107, Cur Avg Loss: 1.18833777, Log Avg loss: 1.09044287, Global Avg Loss: 10.00444882, Time: 0.0209 Steps: 3240, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001121, Sample Num: 17936, Cur Loss: 1.14993250, Cur Avg Loss: 1.18632109, Log Avg loss: 0.96226789, Global Avg Loss: 9.97662672, Time: 0.0209 Steps: 3250, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001131, Sample Num: 18096, Cur Loss: 0.23257694, Cur Avg Loss: 1.18625322, Log Avg loss: 1.17864585, Global Avg Loss: 9.94963905, Time: 0.0209 Steps: 3260, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001141, Sample Num: 18256, Cur Loss: 1.01350689, Cur Avg Loss: 1.18457966, Log Avg loss: 0.99529999, Global Avg Loss: 9.92225575, Time: 0.0209 Steps: 3270, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001151, Sample Num: 18416, Cur Loss: 1.38558578, Cur Avg Loss: 1.18475704, Log Avg loss: 1.20499603, Global Avg Loss: 9.89567874, Time: 0.0209 Steps: 3280, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001161, Sample Num: 18576, Cur Loss: 0.88949472, Cur Avg Loss: 1.18449969, Log Avg loss: 1.15487884, Global Avg Loss: 9.86911096, Time: 0.0209 Steps: 3290, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001171, Sample Num: 18736, Cur Loss: 0.77356440, Cur Avg Loss: 1.18449636, Log Avg loss: 1.18410930, Global Avg Loss: 9.84279277, Time: 0.0209 Steps: 3300, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001181, Sample Num: 18896, Cur Loss: 0.81422150, Cur Avg Loss: 1.18373299, Log Avg loss: 1.09434233, Global Avg Loss: 9.81636241, Time: 0.0213 Steps: 3310, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001191, Sample Num: 19056, Cur Loss: 0.58424401, Cur Avg Loss: 1.18340667, Log Avg loss: 1.14486784, Global Avg Loss: 9.79024345, Time: 0.0209 Steps: 3320, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001201, Sample Num: 19216, Cur Loss: 0.76428795, Cur Avg Loss: 1.18255939, Log Avg loss: 1.08164843, Global Avg Loss: 9.76409151, Time: 0.0209 Steps: 3330, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001211, Sample Num: 19376, Cur Loss: 2.16187286, Cur Avg Loss: 1.18345913, Log Avg loss: 1.29151845, Global Avg Loss: 9.73872453, Time: 0.0209 Steps: 3340, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001221, Sample Num: 19536, Cur Loss: 0.77090561, Cur Avg Loss: 1.18312093, Log Avg loss: 1.14216444, Global Avg Loss: 9.71306315, Time: 0.0209 Steps: 3350, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001231, Sample Num: 19696, Cur Loss: 0.88045204, Cur Avg Loss: 1.18253893, Log Avg loss: 1.11147722, Global Avg Loss: 9.68746320, Time: 0.0209 Steps: 3360, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001241, Sample Num: 19856, Cur Loss: 0.89011967, Cur Avg Loss: 1.17952245, Log Avg loss: 0.80819308, Global Avg Loss: 9.66111521, Time: 0.0209 Steps: 3370, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001251, Sample Num: 20016, Cur Loss: 1.09707856, Cur Avg Loss: 1.17702496, Log Avg loss: 0.86708698, Global Avg Loss: 9.63509738, Time: 0.0209 Steps: 3380, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001261, Sample Num: 20176, Cur Loss: 0.88418788, Cur Avg Loss: 1.17677630, Log Avg loss: 1.14566908, Global Avg Loss: 9.61005482, Time: 0.0209 Steps: 3390, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001271, Sample Num: 20336, Cur Loss: 1.91338432, Cur Avg Loss: 1.17567000, Log Avg loss: 1.03616562, Global Avg Loss: 9.58483750, Time: 0.0209 Steps: 3400, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001281, Sample Num: 20496, Cur Loss: 0.81320196, Cur Avg Loss: 1.17242579, Log Avg loss: 0.76008644, Global Avg Loss: 9.55895846, Time: 0.0245 Steps: 3410, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001291, Sample Num: 20656, Cur Loss: 2.06346893, Cur Avg Loss: 1.17246055, Log Avg loss: 1.17691287, Global Avg Loss: 9.53444956, Time: 0.0209 Steps: 3420, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001301, Sample Num: 20816, Cur Loss: 1.12662601, Cur Avg Loss: 1.17123047, Log Avg loss: 1.01242732, Global Avg Loss: 9.50960401, Time: 0.0209 Steps: 3430, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001311, Sample Num: 20976, Cur Loss: 2.78144073, Cur Avg Loss: 1.17243267, Log Avg loss: 1.32883916, Global Avg Loss: 9.48582272, Time: 0.0209 Steps: 3440, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001321, Sample Num: 21136, Cur Loss: 0.93957460, Cur Avg Loss: 1.17328948, Log Avg loss: 1.28561740, Global Avg Loss: 9.46205401, Time: 0.0209 Steps: 3450, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001331, Sample Num: 21296, Cur Loss: 0.77784514, Cur Avg Loss: 1.17180097, Log Avg loss: 0.97516818, Global Avg Loss: 9.43752543, Time: 0.0208 Steps: 3460, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001341, Sample Num: 21456, Cur Loss: 1.17183828, Cur Avg Loss: 1.17190888, Log Avg loss: 1.18627248, Global Avg Loss: 9.41374661, Time: 0.0208 Steps: 3470, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001351, Sample Num: 21616, Cur Loss: 0.94419146, Cur Avg Loss: 1.16970384, Log Avg loss: 0.87400826, Global Avg Loss: 9.38920713, Time: 0.0209 Steps: 3480, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001361, Sample Num: 21776, Cur Loss: 0.72952610, Cur Avg Loss: 1.16692605, Log Avg loss: 0.79164606, Global Avg Loss: 9.36457228, Time: 0.0208 Steps: 3490, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001371, Sample Num: 21936, Cur Loss: 2.74274683, Cur Avg Loss: 1.16863945, Log Avg loss: 1.40183295, Global Avg Loss: 9.34182160, Time: 0.0208 Steps: 3500, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001381, Sample Num: 22096, Cur Loss: 1.64285111, Cur Avg Loss: 1.16790922, Log Avg loss: 1.06779478, Global Avg Loss: 9.31824887, Time: 0.0209 Steps: 3510, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001391, Sample Num: 22256, Cur Loss: 0.99021441, Cur Avg Loss: 1.16885482, Log Avg loss: 1.29944270, Global Avg Loss: 9.29546817, Time: 0.0209 Steps: 3520, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001401, Sample Num: 22416, Cur Loss: 1.81273055, Cur Avg Loss: 1.17033532, Log Avg loss: 1.37627299, Global Avg Loss: 9.27303419, Time: 0.0209 Steps: 3530, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001411, Sample Num: 22576, Cur Loss: 1.74917161, Cur Avg Loss: 1.17265371, Log Avg loss: 1.49746001, Global Avg Loss: 9.25106930, Time: 0.0208 Steps: 3540, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001421, Sample Num: 22736, Cur Loss: 0.80831707, Cur Avg Loss: 1.17173623, Log Avg loss: 1.04228014, Global Avg Loss: 9.22794595, Time: 0.0208 Steps: 3550, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001431, Sample Num: 22896, Cur Loss: 0.84326768, Cur Avg Loss: 1.17094228, Log Avg loss: 1.05812107, Global Avg Loss: 9.20499700, Time: 0.0208 Steps: 3560, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001441, Sample Num: 23056, Cur Loss: 0.46400470, Cur Avg Loss: 1.16865100, Log Avg loss: 0.84076867, Global Avg Loss: 9.18156779, Time: 0.0208 Steps: 3570, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001451, Sample Num: 23216, Cur Loss: 1.68059397, Cur Avg Loss: 1.16626436, Log Avg loss: 0.82234972, Global Avg Loss: 9.15821802, Time: 0.0209 Steps: 3580, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001461, Sample Num: 23376, Cur Loss: 0.80446827, Cur Avg Loss: 1.16403919, Log Avg loss: 0.84116670, Global Avg Loss: 9.13505074, Time: 0.0209 Steps: 3590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001471, Sample Num: 23536, Cur Loss: 0.18518296, Cur Avg Loss: 1.16172515, Log Avg loss: 0.82364489, Global Avg Loss: 9.11196350, Time: 0.0208 Steps: 3600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001481, Sample Num: 23696, Cur Loss: 1.15281034, Cur Avg Loss: 1.15976680, Log Avg loss: 0.87169332, Global Avg Loss: 9.08913727, Time: 0.0208 Steps: 3610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001491, Sample Num: 23856, Cur Loss: 0.47457713, Cur Avg Loss: 1.15913304, Log Avg loss: 1.06527265, Global Avg Loss: 9.06697190, Time: 0.0209 Steps: 3620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001501, Sample Num: 24016, Cur Loss: 0.71750790, Cur Avg Loss: 1.15742221, Log Avg loss: 0.90233759, Global Avg Loss: 9.04447979, Time: 0.0209 Steps: 3630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001511, Sample Num: 24176, Cur Loss: 0.59512621, Cur Avg Loss: 1.15503436, Log Avg loss: 0.79661821, Global Avg Loss: 9.02182083, Time: 0.0208 Steps: 3640, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001521, Sample Num: 24336, Cur Loss: 1.41719937, Cur Avg Loss: 1.15911661, Log Avg loss: 1.77594517, Global Avg Loss: 9.00196912, Time: 0.0209 Steps: 3650, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001531, Sample Num: 24496, Cur Loss: 0.73491490, Cur Avg Loss: 1.15902603, Log Avg loss: 1.14524763, Global Avg Loss: 8.98050267, Time: 0.0208 Steps: 3660, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001541, Sample Num: 24656, Cur Loss: 0.58036584, Cur Avg Loss: 1.15681918, Log Avg loss: 0.81895149, Global Avg Loss: 8.95826411, Time: 0.0246 Steps: 3670, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001551, Sample Num: 24816, Cur Loss: 1.01762688, Cur Avg Loss: 1.15502252, Log Avg loss: 0.87815681, Global Avg Loss: 8.93630729, Time: 0.0209 Steps: 3680, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001561, Sample Num: 24976, Cur Loss: 0.96723270, Cur Avg Loss: 1.15357390, Log Avg loss: 0.92889300, Global Avg Loss: 8.91460699, Time: 0.0208 Steps: 3690, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001571, Sample Num: 25136, Cur Loss: 1.85762656, Cur Avg Loss: 1.15296231, Log Avg loss: 1.05749287, Global Avg Loss: 8.89337154, Time: 0.0209 Steps: 3700, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001581, Sample Num: 25296, Cur Loss: 0.90163136, Cur Avg Loss: 1.15338375, Log Avg loss: 1.21959269, Global Avg Loss: 8.87268750, Time: 0.0208 Steps: 3710, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001591, Sample Num: 25456, Cur Loss: 1.83295369, Cur Avg Loss: 1.15401378, Log Avg loss: 1.25362062, Global Avg Loss: 8.85220614, Time: 0.0208 Steps: 3720, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001601, Sample Num: 25616, Cur Loss: 0.31821507, Cur Avg Loss: 1.15425976, Log Avg loss: 1.19339641, Global Avg Loss: 8.83167314, Time: 0.0209 Steps: 3730, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001611, Sample Num: 25776, Cur Loss: 1.05639791, Cur Avg Loss: 1.15265502, Log Avg loss: 0.89573590, Global Avg Loss: 8.81045405, Time: 0.0209 Steps: 3740, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001621, Sample Num: 25936, Cur Loss: 0.79666537, Cur Avg Loss: 1.15059975, Log Avg loss: 0.81949591, Global Avg Loss: 8.78914483, Time: 0.0209 Steps: 3750, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001631, Sample Num: 26096, Cur Loss: 0.98661727, Cur Avg Loss: 1.15090509, Log Avg loss: 1.20040052, Global Avg Loss: 8.76896200, Time: 0.0209 Steps: 3760, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001641, Sample Num: 26256, Cur Loss: 0.60368919, Cur Avg Loss: 1.14851109, Log Avg loss: 0.75804918, Global Avg Loss: 8.74771290, Time: 0.0209 Steps: 3770, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001651, Sample Num: 26416, Cur Loss: 0.92022491, Cur Avg Loss: 1.14713858, Log Avg loss: 0.92191005, Global Avg Loss: 8.72700971, Time: 0.0209 Steps: 3780, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001661, Sample Num: 26576, Cur Loss: 0.68327028, Cur Avg Loss: 1.14413892, Log Avg loss: 0.64889532, Global Avg Loss: 8.70569543, Time: 0.0209 Steps: 3790, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001671, Sample Num: 26736, Cur Loss: 0.59206390, Cur Avg Loss: 1.14154588, Log Avg loss: 0.71084118, Global Avg Loss: 8.68465634, Time: 0.0208 Steps: 3800, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001681, Sample Num: 26896, Cur Loss: 0.84456635, Cur Avg Loss: 1.13974925, Log Avg loss: 0.83953216, Global Avg Loss: 8.66406546, Time: 0.0209 Steps: 3810, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001691, Sample Num: 27056, Cur Loss: 0.62087345, Cur Avg Loss: 1.13837877, Log Avg loss: 0.90800122, Global Avg Loss: 8.64376163, Time: 0.0209 Steps: 3820, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001701, Sample Num: 27216, Cur Loss: 1.23697221, Cur Avg Loss: 1.13682909, Log Avg loss: 0.87477822, Global Avg Loss: 8.62347708, Time: 0.0209 Steps: 3830, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001711, Sample Num: 27376, Cur Loss: 0.96032047, Cur Avg Loss: 1.13710648, Log Avg loss: 1.18429129, Global Avg Loss: 8.60410420, Time: 0.0209 Steps: 3840, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001721, Sample Num: 27536, Cur Loss: 1.40319419, Cur Avg Loss: 1.13904062, Log Avg loss: 1.46997136, Global Avg Loss: 8.58557398, Time: 0.0209 Steps: 3850, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001731, Sample Num: 27696, Cur Loss: 1.11943221, Cur Avg Loss: 1.13757272, Log Avg loss: 0.88494678, Global Avg Loss: 8.56562417, Time: 0.0209 Steps: 3860, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001741, Sample Num: 27856, Cur Loss: 0.72812414, Cur Avg Loss: 1.13612603, Log Avg loss: 0.88570397, Global Avg Loss: 8.54577941, Time: 0.0209 Steps: 3870, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001751, Sample Num: 28016, Cur Loss: 0.77294886, Cur Avg Loss: 1.13471699, Log Avg loss: 0.88940345, Global Avg Loss: 8.52604649, Time: 0.0209 Steps: 3880, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001761, Sample Num: 28176, Cur Loss: 0.95471442, Cur Avg Loss: 1.13512872, Log Avg loss: 1.20722211, Global Avg Loss: 8.50723203, Time: 0.0208 Steps: 3890, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001771, Sample Num: 28336, Cur Loss: 2.05311489, Cur Avg Loss: 1.13360996, Log Avg loss: 0.86615618, Global Avg Loss: 8.48763953, Time: 0.0209 Steps: 3900, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001781, Sample Num: 28496, Cur Loss: 1.23554790, Cur Avg Loss: 1.13231863, Log Avg loss: 0.90362552, Global Avg Loss: 8.46824307, Time: 0.0209 Steps: 3910, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001791, Sample Num: 28656, Cur Loss: 1.18339121, Cur Avg Loss: 1.13078115, Log Avg loss: 0.85695563, Global Avg Loss: 8.44882652, Time: 0.0209 Steps: 3920, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001801, Sample Num: 28816, Cur Loss: 0.17600456, Cur Avg Loss: 1.12796672, Log Avg loss: 0.62390223, Global Avg Loss: 8.42891577, Time: 0.0209 Steps: 3930, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001811, Sample Num: 28976, Cur Loss: 1.33826220, Cur Avg Loss: 1.12773602, Log Avg loss: 1.08618737, Global Avg Loss: 8.41027941, Time: 0.0209 Steps: 3940, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001821, Sample Num: 29136, Cur Loss: 0.72920907, Cur Avg Loss: 1.12597499, Log Avg loss: 0.80705099, Global Avg Loss: 8.39103073, Time: 0.0209 Steps: 3950, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001831, Sample Num: 29296, Cur Loss: 1.28665137, Cur Avg Loss: 1.12654195, Log Avg loss: 1.22978590, Global Avg Loss: 8.37294677, Time: 0.0209 Steps: 3960, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001841, Sample Num: 29456, Cur Loss: 0.81291640, Cur Avg Loss: 1.12400959, Log Avg loss: 0.66033478, Global Avg Loss: 8.35351954, Time: 0.0209 Steps: 3970, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001851, Sample Num: 29616, Cur Loss: 0.19671151, Cur Avg Loss: 1.12208213, Log Avg loss: 0.76723744, Global Avg Loss: 8.33445853, Time: 0.0209 Steps: 3980, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001861, Sample Num: 29776, Cur Loss: 0.47289506, Cur Avg Loss: 1.12074581, Log Avg loss: 0.87339148, Global Avg Loss: 8.31575911, Time: 0.0209 Steps: 3990, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001871, Sample Num: 29936, Cur Loss: 0.81309235, Cur Avg Loss: 1.11987003, Log Avg loss: 0.95688780, Global Avg Loss: 8.29736194, Time: 0.0209 Steps: 4000, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001881, Sample Num: 30096, Cur Loss: 1.44209480, Cur Avg Loss: 1.11931482, Log Avg loss: 1.01543606, Global Avg Loss: 8.27920252, Time: 0.0209 Steps: 4010, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001891, Sample Num: 30256, Cur Loss: 0.86321181, Cur Avg Loss: 1.11833271, Log Avg loss: 0.93359762, Global Avg Loss: 8.26092987, Time: 0.0209 Steps: 4020, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001901, Sample Num: 30416, Cur Loss: 0.64294732, Cur Avg Loss: 1.11688183, Log Avg loss: 0.84251964, Global Avg Loss: 8.24252190, Time: 0.0209 Steps: 4030, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001911, Sample Num: 30576, Cur Loss: 0.99652195, Cur Avg Loss: 1.11532228, Log Avg loss: 0.81885265, Global Avg Loss: 8.22414649, Time: 0.0209 Steps: 4040, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001921, Sample Num: 30736, Cur Loss: 0.59585285, Cur Avg Loss: 1.11427606, Log Avg loss: 0.91434321, Global Avg Loss: 8.20609759, Time: 0.0209 Steps: 4050, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001931, Sample Num: 30896, Cur Loss: 0.62279922, Cur Avg Loss: 1.11357090, Log Avg loss: 0.97810848, Global Avg Loss: 8.18829466, Time: 0.0209 Steps: 4060, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001941, Sample Num: 31056, Cur Loss: 0.44387096, Cur Avg Loss: 1.11273958, Log Avg loss: 0.95221333, Global Avg Loss: 8.17051559, Time: 0.0209 Steps: 4070, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001951, Sample Num: 31216, Cur Loss: 1.03182423, Cur Avg Loss: 1.11132681, Log Avg loss: 0.83710762, Global Avg Loss: 8.15254155, Time: 0.0209 Steps: 4080, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001961, Sample Num: 31376, Cur Loss: 0.95818400, Cur Avg Loss: 1.11100884, Log Avg loss: 1.04897173, Global Avg Loss: 8.13517341, Time: 0.0209 Steps: 4090, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001971, Sample Num: 31536, Cur Loss: 1.11056948, Cur Avg Loss: 1.10991739, Log Avg loss: 0.89588419, Global Avg Loss: 8.11751661, Time: 0.0209 Steps: 4100, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001981, Sample Num: 31696, Cur Loss: 1.15389371, Cur Avg Loss: 1.10855100, Log Avg loss: 0.83923575, Global Avg Loss: 8.09980789, Time: 0.0209 Steps: 4110, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001991, Sample Num: 31856, Cur Loss: 1.75801098, Cur Avg Loss: 1.10653825, Log Avg loss: 0.70781350, Global Avg Loss: 8.08186616, Time: 0.0209 Steps: 4120, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002001, Sample Num: 32016, Cur Loss: 2.25613594, Cur Avg Loss: 1.10665945, Log Avg loss: 1.13078954, Global Avg Loss: 8.06503547, Time: 0.0209 Steps: 4130, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002011, Sample Num: 32176, Cur Loss: 0.64964008, Cur Avg Loss: 1.10549595, Log Avg loss: 0.87267974, Global Avg Loss: 8.04766263, Time: 0.0209 Steps: 4140, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002021, Sample Num: 32336, Cur Loss: 0.78013945, Cur Avg Loss: 1.10385788, Log Avg loss: 0.77444285, Global Avg Loss: 8.03013679, Time: 0.0209 Steps: 4150, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002031, Sample Num: 32496, Cur Loss: 1.02320242, Cur Avg Loss: 1.10316363, Log Avg loss: 0.96285571, Global Avg Loss: 8.01314814, Time: 0.0208 Steps: 4160, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002041, Sample Num: 32656, Cur Loss: 0.89431643, Cur Avg Loss: 1.10243457, Log Avg loss: 0.95436231, Global Avg Loss: 7.99622059, Time: 0.0209 Steps: 4170, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002051, Sample Num: 32816, Cur Loss: 1.08270085, Cur Avg Loss: 1.10123183, Log Avg loss: 0.85575228, Global Avg Loss: 7.97913813, Time: 0.0248 Steps: 4180, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002061, Sample Num: 32976, Cur Loss: 2.09144068, Cur Avg Loss: 1.10217253, Log Avg loss: 1.29511007, Global Avg Loss: 7.96318580, Time: 0.0209 Steps: 4190, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002071, Sample Num: 33136, Cur Loss: 1.45914376, Cur Avg Loss: 1.10379450, Log Avg loss: 1.43808205, Global Avg Loss: 7.94764984, Time: 0.0211 Steps: 4200, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002081, Sample Num: 33296, Cur Loss: 1.50097775, Cur Avg Loss: 1.10549944, Log Avg loss: 1.45859349, Global Avg Loss: 7.93223640, Time: 0.0211 Steps: 4210, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002091, Sample Num: 33456, Cur Loss: 1.26530600, Cur Avg Loss: 1.10521293, Log Avg loss: 1.04558904, Global Avg Loss: 7.91591733, Time: 0.0210 Steps: 4220, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002101, Sample Num: 33616, Cur Loss: 0.78029329, Cur Avg Loss: 1.10362331, Log Avg loss: 0.77123335, Global Avg Loss: 7.89902683, Time: 0.0210 Steps: 4230, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002111, Sample Num: 33776, Cur Loss: 1.06131089, Cur Avg Loss: 1.10317675, Log Avg loss: 1.00935552, Global Avg Loss: 7.88277760, Time: 0.0211 Steps: 4240, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002121, Sample Num: 33936, Cur Loss: 0.88373721, Cur Avg Loss: 1.10350191, Log Avg loss: 1.17214246, Global Avg Loss: 7.86698787, Time: 0.0210 Steps: 4250, Updated lr: 0.000097 ***** Running evaluation checkpoint-4258 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-4258 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.804857, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.026151, "eval_total_loss": 721.384099, "eval_mae": 0.861124, "eval_mse": 1.026287, "eval_r2": 0.347624, "eval_sp_statistic": 0.670707, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.731706, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 4.287597, "test_total_loss": 2152.373466, "test_mae": 1.92604, "test_mse": 4.28789, "test_r2": -1.767442, "test_sp_statistic": 0.65537, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.730837, "test_ps_pvalue": 0.0, "lr": 9.691038406827881e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 7.853464492563003, "train_cur_epoch_loss": 2345.8808922097087, "train_cur_epoch_avg_loss": 1.1018698413385197, "train_cur_epoch_time": 44.80485725402832, "train_cur_epoch_avg_time": 0.021045024543930636, "epoch": 2, "step": 4258} ################################################## Training, Epoch: 0003, Batch: 000002, Sample Num: 32, Cur Loss: 0.62216866, Cur Avg Loss: 0.61459592, Log Avg loss: 0.65825391, Global Avg Loss: 7.85006596, Time: 0.0248 Steps: 4260, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000012, Sample Num: 192, Cur Loss: 1.23357964, Cur Avg Loss: 0.84368793, Log Avg loss: 0.88950633, Global Avg Loss: 7.83376489, Time: 0.0210 Steps: 4270, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000022, Sample Num: 352, Cur Loss: 1.36990952, Cur Avg Loss: 0.86150762, Log Avg loss: 0.88289125, Global Avg Loss: 7.81752453, Time: 0.0210 Steps: 4280, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000032, Sample Num: 512, Cur Loss: 0.82901812, Cur Avg Loss: 0.82389408, Log Avg loss: 0.74114430, Global Avg Loss: 7.80102947, Time: 0.0210 Steps: 4290, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000042, Sample Num: 672, Cur Loss: 0.98555529, Cur Avg Loss: 0.84836983, Log Avg loss: 0.92669222, Global Avg Loss: 7.78504264, Time: 0.0210 Steps: 4300, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000052, Sample Num: 832, Cur Loss: 1.03657889, Cur Avg Loss: 0.96835109, Log Avg loss: 1.47227236, Global Avg Loss: 7.77039584, Time: 0.0210 Steps: 4310, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000062, Sample Num: 992, Cur Loss: 0.83201194, Cur Avg Loss: 0.94128395, Log Avg loss: 0.80053482, Global Avg Loss: 7.75426190, Time: 0.0210 Steps: 4320, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000072, Sample Num: 1152, Cur Loss: 0.60831511, Cur Avg Loss: 0.91307083, Log Avg loss: 0.73814952, Global Avg Loss: 7.73805841, Time: 0.0210 Steps: 4330, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000082, Sample Num: 1312, Cur Loss: 1.43024445, Cur Avg Loss: 0.90708302, Log Avg loss: 0.86397077, Global Avg Loss: 7.72221950, Time: 0.0210 Steps: 4340, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000092, Sample Num: 1472, Cur Loss: 1.18986654, Cur Avg Loss: 0.91431278, Log Avg loss: 0.97359686, Global Avg Loss: 7.70670542, Time: 0.0210 Steps: 4350, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000102, Sample Num: 1632, Cur Loss: 1.09931087, Cur Avg Loss: 0.93227983, Log Avg loss: 1.09757662, Global Avg Loss: 7.69154687, Time: 0.0210 Steps: 4360, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000112, Sample Num: 1792, Cur Loss: 1.19844711, Cur Avg Loss: 0.94396389, Log Avg loss: 1.06314134, Global Avg Loss: 7.67637889, Time: 0.0210 Steps: 4370, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000122, Sample Num: 1952, Cur Loss: 0.62960124, Cur Avg Loss: 0.97231433, Log Avg loss: 1.28983921, Global Avg Loss: 7.66179775, Time: 0.0208 Steps: 4380, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000132, Sample Num: 2112, Cur Loss: 1.41710377, Cur Avg Loss: 0.98998406, Log Avg loss: 1.20555484, Global Avg Loss: 7.64709105, Time: 0.0211 Steps: 4390, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000142, Sample Num: 2272, Cur Loss: 0.66437185, Cur Avg Loss: 0.97869605, Log Avg loss: 0.82969424, Global Avg Loss: 7.63159697, Time: 0.0209 Steps: 4400, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000152, Sample Num: 2432, Cur Loss: 1.23277509, Cur Avg Loss: 0.97061239, Log Avg loss: 0.85582442, Global Avg Loss: 7.61623240, Time: 0.0210 Steps: 4410, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000162, Sample Num: 2592, Cur Loss: 1.78550327, Cur Avg Loss: 0.99156673, Log Avg loss: 1.31007279, Global Avg Loss: 7.60196507, Time: 0.0208 Steps: 4420, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000172, Sample Num: 2752, Cur Loss: 0.70856428, Cur Avg Loss: 0.97956336, Log Avg loss: 0.78510878, Global Avg Loss: 7.58657713, Time: 0.0209 Steps: 4430, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000182, Sample Num: 2912, Cur Loss: 0.28915185, Cur Avg Loss: 0.96981673, Log Avg loss: 0.80217471, Global Avg Loss: 7.57129695, Time: 0.0210 Steps: 4440, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000192, Sample Num: 3072, Cur Loss: 0.93967462, Cur Avg Loss: 0.96044548, Log Avg loss: 0.78988867, Global Avg Loss: 7.55605783, Time: 0.0210 Steps: 4450, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000202, Sample Num: 3232, Cur Loss: 0.62998599, Cur Avg Loss: 0.96066443, Log Avg loss: 0.96486822, Global Avg Loss: 7.54127938, Time: 0.0210 Steps: 4460, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000212, Sample Num: 3392, Cur Loss: 0.53715461, Cur Avg Loss: 0.94555046, Log Avg loss: 0.64024821, Global Avg Loss: 7.52584083, Time: 0.0210 Steps: 4470, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000222, Sample Num: 3552, Cur Loss: 1.84652829, Cur Avg Loss: 0.96625952, Log Avg loss: 1.40529158, Global Avg Loss: 7.51217889, Time: 0.0211 Steps: 4480, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000232, Sample Num: 3712, Cur Loss: 2.72288656, Cur Avg Loss: 0.97575758, Log Avg loss: 1.18661450, Global Avg Loss: 7.49809077, Time: 0.0210 Steps: 4490, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000242, Sample Num: 3872, Cur Loss: 0.72939324, Cur Avg Loss: 0.96441079, Log Avg loss: 0.70116545, Global Avg Loss: 7.48298649, Time: 0.0209 Steps: 4500, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000252, Sample Num: 4032, Cur Loss: 0.32036993, Cur Avg Loss: 0.95811803, Log Avg loss: 0.80583315, Global Avg Loss: 7.46818128, Time: 0.0208 Steps: 4510, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000262, Sample Num: 4192, Cur Loss: 0.36717033, Cur Avg Loss: 0.95707736, Log Avg loss: 0.93085239, Global Avg Loss: 7.45371816, Time: 0.0246 Steps: 4520, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000272, Sample Num: 4352, Cur Loss: 0.48046160, Cur Avg Loss: 0.96775330, Log Avg loss: 1.24746293, Global Avg Loss: 7.44001782, Time: 0.0206 Steps: 4530, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000282, Sample Num: 4512, Cur Loss: 0.49054834, Cur Avg Loss: 0.97356195, Log Avg loss: 1.13155718, Global Avg Loss: 7.42612253, Time: 0.0208 Steps: 4540, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000292, Sample Num: 4672, Cur Loss: 0.35108924, Cur Avg Loss: 0.96395200, Log Avg loss: 0.69295155, Global Avg Loss: 7.41132435, Time: 0.0207 Steps: 4550, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000302, Sample Num: 4832, Cur Loss: 1.49452460, Cur Avg Loss: 0.96445098, Log Avg loss: 0.97902119, Global Avg Loss: 7.39721842, Time: 0.0208 Steps: 4560, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000312, Sample Num: 4992, Cur Loss: 0.57395566, Cur Avg Loss: 0.96458452, Log Avg loss: 0.96861742, Global Avg Loss: 7.38315146, Time: 0.0208 Steps: 4570, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000322, Sample Num: 5152, Cur Loss: 0.35337996, Cur Avg Loss: 0.95572663, Log Avg loss: 0.67936044, Global Avg Loss: 7.36851436, Time: 0.0208 Steps: 4580, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000332, Sample Num: 5312, Cur Loss: 1.67641926, Cur Avg Loss: 0.95505352, Log Avg loss: 0.93337951, Global Avg Loss: 7.35449446, Time: 0.0208 Steps: 4590, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000342, Sample Num: 5472, Cur Loss: 0.45903260, Cur Avg Loss: 0.95415181, Log Avg loss: 0.92421501, Global Avg Loss: 7.34051559, Time: 0.0208 Steps: 4600, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000352, Sample Num: 5632, Cur Loss: 0.81346726, Cur Avg Loss: 0.94942801, Log Avg loss: 0.78787400, Global Avg Loss: 7.32630162, Time: 0.0208 Steps: 4610, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000362, Sample Num: 5792, Cur Loss: 0.55277586, Cur Avg Loss: 0.94348210, Log Avg loss: 0.73418592, Global Avg Loss: 7.31203297, Time: 0.0208 Steps: 4620, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000372, Sample Num: 5952, Cur Loss: 0.76340759, Cur Avg Loss: 0.94203042, Log Avg loss: 0.88947985, Global Avg Loss: 7.29816137, Time: 0.0208 Steps: 4630, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000382, Sample Num: 6112, Cur Loss: 0.95729232, Cur Avg Loss: 0.94785001, Log Avg loss: 1.16433883, Global Avg Loss: 7.28494192, Time: 0.0208 Steps: 4640, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000392, Sample Num: 6272, Cur Loss: 1.18269420, Cur Avg Loss: 0.94638414, Log Avg loss: 0.89038779, Global Avg Loss: 7.27119019, Time: 0.0208 Steps: 4650, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000402, Sample Num: 6432, Cur Loss: 0.98020685, Cur Avg Loss: 0.94321163, Log Avg loss: 0.81884911, Global Avg Loss: 7.25734397, Time: 0.0209 Steps: 4660, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000412, Sample Num: 6592, Cur Loss: 0.95960701, Cur Avg Loss: 0.94391091, Log Avg loss: 0.97202219, Global Avg Loss: 7.24388503, Time: 0.0208 Steps: 4670, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000422, Sample Num: 6752, Cur Loss: 0.46192479, Cur Avg Loss: 0.94186556, Log Avg loss: 0.85759714, Global Avg Loss: 7.23023912, Time: 0.0208 Steps: 4680, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000432, Sample Num: 6912, Cur Loss: 0.45199662, Cur Avg Loss: 0.94187453, Log Avg loss: 0.94225268, Global Avg Loss: 7.21683190, Time: 0.0208 Steps: 4690, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000442, Sample Num: 7072, Cur Loss: 0.65725589, Cur Avg Loss: 0.94612779, Log Avg loss: 1.12986904, Global Avg Loss: 7.20388091, Time: 0.0208 Steps: 4700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000452, Sample Num: 7232, Cur Loss: 0.51931095, Cur Avg Loss: 0.95400407, Log Avg loss: 1.30213564, Global Avg Loss: 7.19135067, Time: 0.0208 Steps: 4710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000462, Sample Num: 7392, Cur Loss: 0.73155010, Cur Avg Loss: 0.94904918, Log Avg loss: 0.72508807, Global Avg Loss: 7.17765096, Time: 0.0208 Steps: 4720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000472, Sample Num: 7552, Cur Loss: 0.36670995, Cur Avg Loss: 0.94490081, Log Avg loss: 0.75324613, Global Avg Loss: 7.16406871, Time: 0.0208 Steps: 4730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000482, Sample Num: 7712, Cur Loss: 0.85205632, Cur Avg Loss: 0.94759908, Log Avg loss: 1.07495716, Global Avg Loss: 7.15122248, Time: 0.0208 Steps: 4740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000492, Sample Num: 7872, Cur Loss: 0.20799898, Cur Avg Loss: 0.94416305, Log Avg loss: 0.77854636, Global Avg Loss: 7.13780632, Time: 0.0208 Steps: 4750, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000502, Sample Num: 8032, Cur Loss: 1.08603811, Cur Avg Loss: 0.93751566, Log Avg loss: 0.61046425, Global Avg Loss: 7.12409342, Time: 0.0208 Steps: 4760, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000512, Sample Num: 8192, Cur Loss: 0.29111323, Cur Avg Loss: 0.93374738, Log Avg loss: 0.74457974, Global Avg Loss: 7.11071918, Time: 0.0253 Steps: 4770, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000522, Sample Num: 8352, Cur Loss: 0.90139842, Cur Avg Loss: 0.93172900, Log Avg loss: 0.82838814, Global Avg Loss: 7.09757622, Time: 0.0209 Steps: 4780, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000532, Sample Num: 8512, Cur Loss: 1.08189011, Cur Avg Loss: 0.92943223, Log Avg loss: 0.80954051, Global Avg Loss: 7.08444880, Time: 0.0207 Steps: 4790, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000542, Sample Num: 8672, Cur Loss: 1.58836460, Cur Avg Loss: 0.93475617, Log Avg loss: 1.21798998, Global Avg Loss: 7.07222701, Time: 0.0207 Steps: 4800, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000552, Sample Num: 8832, Cur Loss: 0.89464754, Cur Avg Loss: 0.93349794, Log Avg loss: 0.86530198, Global Avg Loss: 7.05932280, Time: 0.0207 Steps: 4810, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000562, Sample Num: 8992, Cur Loss: 0.63320231, Cur Avg Loss: 0.93234081, Log Avg loss: 0.86846677, Global Avg Loss: 7.04647870, Time: 0.0207 Steps: 4820, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000572, Sample Num: 9152, Cur Loss: 0.49476886, Cur Avg Loss: 0.92739528, Log Avg loss: 0.64945675, Global Avg Loss: 7.03323435, Time: 0.0207 Steps: 4830, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000582, Sample Num: 9312, Cur Loss: 0.79959685, Cur Avg Loss: 0.92260305, Log Avg loss: 0.64848760, Global Avg Loss: 7.02004272, Time: 0.0207 Steps: 4840, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000592, Sample Num: 9472, Cur Loss: 1.01620388, Cur Avg Loss: 0.91795625, Log Avg loss: 0.64751260, Global Avg Loss: 7.00690349, Time: 0.0208 Steps: 4850, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000602, Sample Num: 9632, Cur Loss: 0.41963568, Cur Avg Loss: 0.91410422, Log Avg loss: 0.68606374, Global Avg Loss: 6.99389764, Time: 0.0207 Steps: 4860, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000612, Sample Num: 9792, Cur Loss: 1.03716278, Cur Avg Loss: 0.91243168, Log Avg loss: 0.81174466, Global Avg Loss: 6.98120328, Time: 0.0207 Steps: 4870, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000622, Sample Num: 9952, Cur Loss: 0.32923740, Cur Avg Loss: 0.91202448, Log Avg loss: 0.88710375, Global Avg Loss: 6.96871538, Time: 0.0208 Steps: 4880, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000632, Sample Num: 10112, Cur Loss: 0.91691267, Cur Avg Loss: 0.91049049, Log Avg loss: 0.81507646, Global Avg Loss: 6.95613125, Time: 0.0207 Steps: 4890, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000642, Sample Num: 10272, Cur Loss: 1.00644588, Cur Avg Loss: 0.90928699, Log Avg loss: 0.83322595, Global Avg Loss: 6.94363552, Time: 0.0207 Steps: 4900, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000652, Sample Num: 10432, Cur Loss: 0.77980220, Cur Avg Loss: 0.90998321, Log Avg loss: 0.95468064, Global Avg Loss: 6.93143806, Time: 0.0208 Steps: 4910, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000662, Sample Num: 10592, Cur Loss: 0.78899908, Cur Avg Loss: 0.90723152, Log Avg loss: 0.72782108, Global Avg Loss: 6.91882908, Time: 0.0207 Steps: 4920, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000672, Sample Num: 10752, Cur Loss: 0.86720592, Cur Avg Loss: 0.90835566, Log Avg loss: 0.98277390, Global Avg Loss: 6.90678840, Time: 0.0207 Steps: 4930, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000682, Sample Num: 10912, Cur Loss: 0.48262471, Cur Avg Loss: 0.90610422, Log Avg loss: 0.75480751, Global Avg Loss: 6.89433500, Time: 0.0207 Steps: 4940, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000692, Sample Num: 11072, Cur Loss: 0.83792585, Cur Avg Loss: 0.90284101, Log Avg loss: 0.68028970, Global Avg Loss: 6.88178137, Time: 0.0207 Steps: 4950, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000702, Sample Num: 11232, Cur Loss: 0.91463006, Cur Avg Loss: 0.89836075, Log Avg loss: 0.58832735, Global Avg Loss: 6.86909296, Time: 0.0208 Steps: 4960, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000712, Sample Num: 11392, Cur Loss: 0.65584797, Cur Avg Loss: 0.89466481, Log Avg loss: 0.63520944, Global Avg Loss: 6.85654993, Time: 0.0207 Steps: 4970, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000722, Sample Num: 11552, Cur Loss: 0.76512915, Cur Avg Loss: 0.89325064, Log Avg loss: 0.79256214, Global Avg Loss: 6.84437325, Time: 0.0207 Steps: 4980, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000732, Sample Num: 11712, Cur Loss: 2.44540286, Cur Avg Loss: 0.89754753, Log Avg loss: 1.20778265, Global Avg Loss: 6.83307748, Time: 0.0208 Steps: 4990, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000742, Sample Num: 11872, Cur Loss: 0.57535791, Cur Avg Loss: 0.89880150, Log Avg loss: 0.99059243, Global Avg Loss: 6.82139251, Time: 0.0208 Steps: 5000, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000752, Sample Num: 12032, Cur Loss: 0.97448558, Cur Avg Loss: 0.89810614, Log Avg loss: 0.84651045, Global Avg Loss: 6.80946659, Time: 0.0207 Steps: 5010, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000762, Sample Num: 12192, Cur Loss: 0.97937900, Cur Avg Loss: 0.89493083, Log Avg loss: 0.65614722, Global Avg Loss: 6.79720898, Time: 0.0209 Steps: 5020, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000772, Sample Num: 12352, Cur Loss: 1.02503610, Cur Avg Loss: 0.89379149, Log Avg loss: 0.80697388, Global Avg Loss: 6.78529997, Time: 0.0245 Steps: 5030, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000782, Sample Num: 12512, Cur Loss: 1.56313407, Cur Avg Loss: 0.89478385, Log Avg loss: 0.97139418, Global Avg Loss: 6.77376444, Time: 0.0208 Steps: 5040, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000792, Sample Num: 12672, Cur Loss: 1.05132258, Cur Avg Loss: 0.89784606, Log Avg loss: 1.13731092, Global Avg Loss: 6.76260315, Time: 0.0208 Steps: 5050, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000802, Sample Num: 12832, Cur Loss: 0.59142035, Cur Avg Loss: 0.89641870, Log Avg loss: 0.78337179, Global Avg Loss: 6.75078648, Time: 0.0207 Steps: 5060, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000812, Sample Num: 12992, Cur Loss: 1.68336630, Cur Avg Loss: 0.89665428, Log Avg loss: 0.91554732, Global Avg Loss: 6.73927714, Time: 0.0208 Steps: 5070, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000822, Sample Num: 13152, Cur Loss: 0.80210018, Cur Avg Loss: 0.89645003, Log Avg loss: 0.87986531, Global Avg Loss: 6.72774286, Time: 0.0207 Steps: 5080, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000832, Sample Num: 13312, Cur Loss: 0.32622355, Cur Avg Loss: 0.89417412, Log Avg loss: 0.70709444, Global Avg Loss: 6.71591448, Time: 0.0208 Steps: 5090, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000842, Sample Num: 13472, Cur Loss: 0.57191181, Cur Avg Loss: 0.89311535, Log Avg loss: 0.80502577, Global Avg Loss: 6.70432450, Time: 0.0207 Steps: 5100, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000852, Sample Num: 13632, Cur Loss: 0.45605621, Cur Avg Loss: 0.89457622, Log Avg loss: 1.01758134, Global Avg Loss: 6.69319584, Time: 0.0208 Steps: 5110, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000862, Sample Num: 13792, Cur Loss: 0.41209587, Cur Avg Loss: 0.89312484, Log Avg loss: 0.76946726, Global Avg Loss: 6.68162606, Time: 0.0207 Steps: 5120, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000872, Sample Num: 13952, Cur Loss: 2.00251174, Cur Avg Loss: 0.90128862, Log Avg loss: 1.60500641, Global Avg Loss: 6.67173011, Time: 0.0207 Steps: 5130, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000882, Sample Num: 14112, Cur Loss: 0.97985917, Cur Avg Loss: 0.90698853, Log Avg loss: 1.40402061, Global Avg Loss: 6.66148165, Time: 0.0208 Steps: 5140, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000892, Sample Num: 14272, Cur Loss: 0.80767477, Cur Avg Loss: 0.91109313, Log Avg loss: 1.27311873, Global Avg Loss: 6.65101881, Time: 0.0208 Steps: 5150, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000902, Sample Num: 14432, Cur Loss: 0.53351879, Cur Avg Loss: 0.91103624, Log Avg loss: 0.90596144, Global Avg Loss: 6.63988498, Time: 0.0208 Steps: 5160, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000912, Sample Num: 14592, Cur Loss: 0.42984664, Cur Avg Loss: 0.91117625, Log Avg loss: 0.92380492, Global Avg Loss: 6.62882873, Time: 0.0207 Steps: 5170, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000922, Sample Num: 14752, Cur Loss: 0.44157982, Cur Avg Loss: 0.91116241, Log Avg loss: 0.90990077, Global Avg Loss: 6.61778833, Time: 0.0208 Steps: 5180, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000932, Sample Num: 14912, Cur Loss: 0.48860812, Cur Avg Loss: 0.91078426, Log Avg loss: 0.87591831, Global Avg Loss: 6.60672500, Time: 0.0207 Steps: 5190, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000942, Sample Num: 15072, Cur Loss: 1.43368340, Cur Avg Loss: 0.90900312, Log Avg loss: 0.74300142, Global Avg Loss: 6.59544861, Time: 0.0207 Steps: 5200, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000952, Sample Num: 15232, Cur Loss: 0.72261798, Cur Avg Loss: 0.90626515, Log Avg loss: 0.64834825, Global Avg Loss: 6.58403383, Time: 0.0207 Steps: 5210, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000962, Sample Num: 15392, Cur Loss: 0.99529707, Cur Avg Loss: 0.90641380, Log Avg loss: 0.92056480, Global Avg Loss: 6.57318427, Time: 0.0208 Steps: 5220, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000972, Sample Num: 15552, Cur Loss: 0.51156449, Cur Avg Loss: 0.90502499, Log Avg loss: 0.77142164, Global Avg Loss: 6.56209103, Time: 0.0208 Steps: 5230, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000982, Sample Num: 15712, Cur Loss: 1.23090053, Cur Avg Loss: 0.90441056, Log Avg loss: 0.84468870, Global Avg Loss: 6.55117996, Time: 0.0207 Steps: 5240, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000992, Sample Num: 15872, Cur Loss: 1.04128563, Cur Avg Loss: 0.90584819, Log Avg loss: 1.04702280, Global Avg Loss: 6.54069585, Time: 0.0207 Steps: 5250, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001002, Sample Num: 16032, Cur Loss: 0.52746880, Cur Avg Loss: 0.90509432, Log Avg loss: 0.83031049, Global Avg Loss: 6.52983960, Time: 0.0208 Steps: 5260, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001012, Sample Num: 16192, Cur Loss: 0.65459144, Cur Avg Loss: 0.90266293, Log Avg loss: 0.65903772, Global Avg Loss: 6.51869956, Time: 0.0208 Steps: 5270, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001022, Sample Num: 16352, Cur Loss: 0.59470773, Cur Avg Loss: 0.90136978, Log Avg loss: 0.77050340, Global Avg Loss: 6.50781283, Time: 0.0208 Steps: 5280, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001032, Sample Num: 16512, Cur Loss: 0.54993343, Cur Avg Loss: 0.90196241, Log Avg loss: 0.96252931, Global Avg Loss: 6.49733025, Time: 0.0208 Steps: 5290, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001042, Sample Num: 16672, Cur Loss: 1.37604797, Cur Avg Loss: 0.90134073, Log Avg loss: 0.83718328, Global Avg Loss: 6.48665073, Time: 0.0208 Steps: 5300, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001052, Sample Num: 16832, Cur Loss: 0.83055258, Cur Avg Loss: 0.90657256, Log Avg loss: 1.45172881, Global Avg Loss: 6.47716877, Time: 0.0207 Steps: 5310, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001062, Sample Num: 16992, Cur Loss: 1.96119809, Cur Avg Loss: 0.90936637, Log Avg loss: 1.20327473, Global Avg Loss: 6.46725543, Time: 0.0207 Steps: 5320, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001072, Sample Num: 17152, Cur Loss: 0.66470718, Cur Avg Loss: 0.90796625, Log Avg loss: 0.75927392, Global Avg Loss: 6.45654627, Time: 0.0208 Steps: 5330, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001082, Sample Num: 17312, Cur Loss: 0.85531223, Cur Avg Loss: 0.91187997, Log Avg loss: 1.33143062, Global Avg Loss: 6.44694868, Time: 0.0207 Steps: 5340, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001092, Sample Num: 17472, Cur Loss: 1.24188113, Cur Avg Loss: 0.91085094, Log Avg loss: 0.79951046, Global Avg Loss: 6.43639272, Time: 0.0208 Steps: 5350, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001102, Sample Num: 17632, Cur Loss: 1.31057310, Cur Avg Loss: 0.91026174, Log Avg loss: 0.84592120, Global Avg Loss: 6.42596273, Time: 0.0208 Steps: 5360, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001112, Sample Num: 17792, Cur Loss: 0.67687535, Cur Avg Loss: 0.90883023, Log Avg loss: 0.75107777, Global Avg Loss: 6.41539498, Time: 0.0207 Steps: 5370, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001122, Sample Num: 17952, Cur Loss: 0.24730736, Cur Avg Loss: 0.90736901, Log Avg loss: 0.74488087, Global Avg Loss: 6.40485499, Time: 0.0207 Steps: 5380, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001132, Sample Num: 18112, Cur Loss: 1.56172693, Cur Avg Loss: 0.91066619, Log Avg loss: 1.28061002, Global Avg Loss: 6.39534804, Time: 0.0207 Steps: 5390, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001142, Sample Num: 18272, Cur Loss: 0.45839232, Cur Avg Loss: 0.91116219, Log Avg loss: 0.96730917, Global Avg Loss: 6.38529612, Time: 0.0208 Steps: 5400, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001152, Sample Num: 18432, Cur Loss: 1.18973100, Cur Avg Loss: 0.90979350, Log Avg loss: 0.75348879, Global Avg Loss: 6.37488612, Time: 0.0207 Steps: 5410, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001162, Sample Num: 18592, Cur Loss: 0.70686400, Cur Avg Loss: 0.91148792, Log Avg loss: 1.10668493, Global Avg Loss: 6.36516619, Time: 0.0207 Steps: 5420, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001172, Sample Num: 18752, Cur Loss: 0.66830140, Cur Avg Loss: 0.90890917, Log Avg loss: 0.60925849, Global Avg Loss: 6.35456599, Time: 0.0208 Steps: 5430, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001182, Sample Num: 18912, Cur Loss: 0.47419667, Cur Avg Loss: 0.90754744, Log Avg loss: 0.74795267, Global Avg Loss: 6.34425972, Time: 0.0207 Steps: 5440, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001192, Sample Num: 19072, Cur Loss: 1.35287666, Cur Avg Loss: 0.90730155, Log Avg loss: 0.87823797, Global Avg Loss: 6.33423032, Time: 0.0207 Steps: 5450, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001202, Sample Num: 19232, Cur Loss: 0.26504305, Cur Avg Loss: 0.90830589, Log Avg loss: 1.02802288, Global Avg Loss: 6.32451199, Time: 0.0208 Steps: 5460, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001212, Sample Num: 19392, Cur Loss: 1.46639967, Cur Avg Loss: 0.90716683, Log Avg loss: 0.77025195, Global Avg Loss: 6.31435795, Time: 0.0207 Steps: 5470, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001222, Sample Num: 19552, Cur Loss: 0.55388516, Cur Avg Loss: 0.90525723, Log Avg loss: 0.67381368, Global Avg Loss: 6.30406499, Time: 0.0207 Steps: 5480, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001232, Sample Num: 19712, Cur Loss: 0.36692673, Cur Avg Loss: 0.90492545, Log Avg loss: 0.86438205, Global Avg Loss: 6.29415664, Time: 0.0207 Steps: 5490, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001242, Sample Num: 19872, Cur Loss: 1.46403015, Cur Avg Loss: 0.90357413, Log Avg loss: 0.73709112, Global Avg Loss: 6.28405289, Time: 0.0208 Steps: 5500, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001252, Sample Num: 20032, Cur Loss: 0.75812113, Cur Avg Loss: 0.90243297, Log Avg loss: 0.76070182, Global Avg Loss: 6.27402866, Time: 0.0208 Steps: 5510, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001262, Sample Num: 20192, Cur Loss: 0.62482405, Cur Avg Loss: 0.90178883, Log Avg loss: 0.82114251, Global Avg Loss: 6.26415024, Time: 0.0208 Steps: 5520, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001272, Sample Num: 20352, Cur Loss: 0.99806660, Cur Avg Loss: 0.89954782, Log Avg loss: 0.61673126, Global Avg Loss: 6.25393791, Time: 0.0207 Steps: 5530, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001282, Sample Num: 20512, Cur Loss: 0.56870210, Cur Avg Loss: 0.89847371, Log Avg loss: 0.76184717, Global Avg Loss: 6.24402439, Time: 0.0245 Steps: 5540, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001292, Sample Num: 20672, Cur Loss: 0.18292439, Cur Avg Loss: 0.89623422, Log Avg loss: 0.60913218, Global Avg Loss: 6.23387143, Time: 0.0207 Steps: 5550, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001302, Sample Num: 20832, Cur Loss: 0.49001199, Cur Avg Loss: 0.89445501, Log Avg loss: 0.66458038, Global Avg Loss: 6.22385472, Time: 0.0208 Steps: 5560, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001312, Sample Num: 20992, Cur Loss: 0.64795226, Cur Avg Loss: 0.89345540, Log Avg loss: 0.76330643, Global Avg Loss: 6.21405122, Time: 0.0208 Steps: 5570, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001322, Sample Num: 21152, Cur Loss: 0.42973930, Cur Avg Loss: 0.89111536, Log Avg loss: 0.58410289, Global Avg Loss: 6.20396171, Time: 0.0208 Steps: 5580, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001332, Sample Num: 21312, Cur Loss: 1.96537256, Cur Avg Loss: 0.89417099, Log Avg loss: 1.29812508, Global Avg Loss: 6.19518561, Time: 0.0208 Steps: 5590, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001342, Sample Num: 21472, Cur Loss: 1.39418197, Cur Avg Loss: 0.89391370, Log Avg loss: 0.85964178, Global Avg Loss: 6.18565786, Time: 0.0209 Steps: 5600, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001352, Sample Num: 21632, Cur Loss: 0.48746994, Cur Avg Loss: 0.89329005, Log Avg loss: 0.80959709, Global Avg Loss: 6.17607486, Time: 0.0209 Steps: 5610, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001362, Sample Num: 21792, Cur Loss: 0.77095956, Cur Avg Loss: 0.89276920, Log Avg loss: 0.82235011, Global Avg Loss: 6.16654866, Time: 0.0209 Steps: 5620, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001372, Sample Num: 21952, Cur Loss: 0.38615280, Cur Avg Loss: 0.89106741, Log Avg loss: 0.65928398, Global Avg Loss: 6.15676666, Time: 0.0208 Steps: 5630, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001382, Sample Num: 22112, Cur Loss: 1.36050248, Cur Avg Loss: 0.89075377, Log Avg loss: 0.84772207, Global Avg Loss: 6.14735346, Time: 0.0209 Steps: 5640, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001392, Sample Num: 22272, Cur Loss: 1.78841829, Cur Avg Loss: 0.89303015, Log Avg loss: 1.20762555, Global Avg Loss: 6.13861058, Time: 0.0210 Steps: 5650, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001402, Sample Num: 22432, Cur Loss: 0.89015234, Cur Avg Loss: 0.89429319, Log Avg loss: 1.07010873, Global Avg Loss: 6.12965563, Time: 0.0209 Steps: 5660, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001412, Sample Num: 22592, Cur Loss: 0.48976159, Cur Avg Loss: 0.89614404, Log Avg loss: 1.15563222, Global Avg Loss: 6.12088310, Time: 0.0209 Steps: 5670, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001422, Sample Num: 22752, Cur Loss: 0.73545021, Cur Avg Loss: 0.89613325, Log Avg loss: 0.89461052, Global Avg Loss: 6.11168192, Time: 0.0209 Steps: 5680, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001432, Sample Num: 22912, Cur Loss: 0.46028444, Cur Avg Loss: 0.89467638, Log Avg loss: 0.68750897, Global Avg Loss: 6.10214910, Time: 0.0209 Steps: 5690, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001442, Sample Num: 23072, Cur Loss: 0.49679708, Cur Avg Loss: 0.89327058, Log Avg loss: 0.69195988, Global Avg Loss: 6.09265754, Time: 0.0208 Steps: 5700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001452, Sample Num: 23232, Cur Loss: 0.21725643, Cur Avg Loss: 0.89273022, Log Avg loss: 0.81481036, Global Avg Loss: 6.08341438, Time: 0.0209 Steps: 5710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001462, Sample Num: 23392, Cur Loss: 0.90225244, Cur Avg Loss: 0.89088567, Log Avg loss: 0.62305689, Global Avg Loss: 6.07386830, Time: 0.0209 Steps: 5720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001472, Sample Num: 23552, Cur Loss: 3.06310320, Cur Avg Loss: 0.89401862, Log Avg loss: 1.35205597, Global Avg Loss: 6.06562779, Time: 0.0209 Steps: 5730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001482, Sample Num: 23712, Cur Loss: 0.52666521, Cur Avg Loss: 0.89141633, Log Avg loss: 0.50836020, Global Avg Loss: 6.05594614, Time: 0.0209 Steps: 5740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001492, Sample Num: 23872, Cur Loss: 0.92181444, Cur Avg Loss: 0.88999846, Log Avg loss: 0.67986951, Global Avg Loss: 6.04659644, Time: 0.0209 Steps: 5750, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001502, Sample Num: 24032, Cur Loss: 0.77334762, Cur Avg Loss: 0.88944841, Log Avg loss: 0.80738142, Global Avg Loss: 6.03750058, Time: 0.0209 Steps: 5760, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001512, Sample Num: 24192, Cur Loss: 0.92035544, Cur Avg Loss: 0.88995074, Log Avg loss: 0.96540087, Global Avg Loss: 6.02871011, Time: 0.0209 Steps: 5770, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001522, Sample Num: 24352, Cur Loss: 1.05250621, Cur Avg Loss: 0.89168846, Log Avg loss: 1.15443162, Global Avg Loss: 6.02027710, Time: 0.0208 Steps: 5780, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001532, Sample Num: 24512, Cur Loss: 1.72060835, Cur Avg Loss: 0.89044885, Log Avg loss: 0.70177916, Global Avg Loss: 6.01109144, Time: 0.0209 Steps: 5790, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001542, Sample Num: 24672, Cur Loss: 1.23344684, Cur Avg Loss: 0.89281925, Log Avg loss: 1.25596547, Global Avg Loss: 6.00289295, Time: 0.0208 Steps: 5800, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001552, Sample Num: 24832, Cur Loss: 0.88268089, Cur Avg Loss: 0.89170252, Log Avg loss: 0.71950285, Global Avg Loss: 5.99379933, Time: 0.0208 Steps: 5810, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001562, Sample Num: 24992, Cur Loss: 0.46186373, Cur Avg Loss: 0.89212160, Log Avg loss: 0.95716297, Global Avg Loss: 5.98514532, Time: 0.0208 Steps: 5820, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001572, Sample Num: 25152, Cur Loss: 0.57062894, Cur Avg Loss: 0.89154098, Log Avg loss: 0.80084790, Global Avg Loss: 5.97625287, Time: 0.0208 Steps: 5830, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001582, Sample Num: 25312, Cur Loss: 0.48972389, Cur Avg Loss: 0.89087241, Log Avg loss: 0.78577248, Global Avg Loss: 5.96736506, Time: 0.0207 Steps: 5840, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001592, Sample Num: 25472, Cur Loss: 0.89759040, Cur Avg Loss: 0.89103801, Log Avg loss: 0.91723582, Global Avg Loss: 5.95873236, Time: 0.0208 Steps: 5850, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001602, Sample Num: 25632, Cur Loss: 0.76410937, Cur Avg Loss: 0.89094867, Log Avg loss: 0.87672678, Global Avg Loss: 5.95006000, Time: 0.0208 Steps: 5860, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001612, Sample Num: 25792, Cur Loss: 1.68346131, Cur Avg Loss: 0.88972987, Log Avg loss: 0.69447837, Global Avg Loss: 5.94110671, Time: 0.0208 Steps: 5870, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001622, Sample Num: 25952, Cur Loss: 0.73313642, Cur Avg Loss: 0.88793373, Log Avg loss: 0.59839582, Global Avg Loss: 5.93202046, Time: 0.0207 Steps: 5880, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001632, Sample Num: 26112, Cur Loss: 0.37845528, Cur Avg Loss: 0.88650905, Log Avg loss: 0.65542508, Global Avg Loss: 5.92306190, Time: 0.0207 Steps: 5890, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001642, Sample Num: 26272, Cur Loss: 0.70002753, Cur Avg Loss: 0.88602450, Log Avg loss: 0.80694605, Global Avg Loss: 5.91439051, Time: 0.0208 Steps: 5900, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001652, Sample Num: 26432, Cur Loss: 0.69911897, Cur Avg Loss: 0.88484010, Log Avg loss: 0.69036104, Global Avg Loss: 5.90555121, Time: 0.0208 Steps: 5910, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001662, Sample Num: 26592, Cur Loss: 0.89831555, Cur Avg Loss: 0.88397189, Log Avg loss: 0.74054381, Global Avg Loss: 5.89682653, Time: 0.0207 Steps: 5920, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001672, Sample Num: 26752, Cur Loss: 0.60578835, Cur Avg Loss: 0.88279998, Log Avg loss: 0.68802837, Global Avg Loss: 5.88804273, Time: 0.0207 Steps: 5930, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001682, Sample Num: 26912, Cur Loss: 0.85302341, Cur Avg Loss: 0.88370699, Log Avg loss: 1.03536048, Global Avg Loss: 5.87987323, Time: 0.0208 Steps: 5940, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001692, Sample Num: 27072, Cur Loss: 0.19028652, Cur Avg Loss: 0.88443120, Log Avg loss: 1.00624244, Global Avg Loss: 5.87168225, Time: 0.0208 Steps: 5950, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001702, Sample Num: 27232, Cur Loss: 0.75518191, Cur Avg Loss: 0.88437489, Log Avg loss: 0.87484816, Global Avg Loss: 5.86329830, Time: 0.0208 Steps: 5960, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001712, Sample Num: 27392, Cur Loss: 0.96394122, Cur Avg Loss: 0.88301224, Log Avg loss: 0.65108785, Global Avg Loss: 5.85456763, Time: 0.0208 Steps: 5970, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001722, Sample Num: 27552, Cur Loss: 0.72448027, Cur Avg Loss: 0.88239106, Log Avg loss: 0.77604491, Global Avg Loss: 5.84607512, Time: 0.0208 Steps: 5980, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001732, Sample Num: 27712, Cur Loss: 0.59146774, Cur Avg Loss: 0.88084886, Log Avg loss: 0.61528252, Global Avg Loss: 5.83734258, Time: 0.0208 Steps: 5990, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001742, Sample Num: 27872, Cur Loss: 1.32637799, Cur Avg Loss: 0.88055030, Log Avg loss: 0.82884003, Global Avg Loss: 5.82899507, Time: 0.0208 Steps: 6000, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001752, Sample Num: 28032, Cur Loss: 0.45867896, Cur Avg Loss: 0.87879899, Log Avg loss: 0.57372096, Global Avg Loss: 5.82025086, Time: 0.0208 Steps: 6010, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001762, Sample Num: 28192, Cur Loss: 1.05841649, Cur Avg Loss: 0.87865076, Log Avg loss: 0.85268110, Global Avg Loss: 5.81199908, Time: 0.0208 Steps: 6020, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001772, Sample Num: 28352, Cur Loss: 0.62923729, Cur Avg Loss: 0.87704741, Log Avg loss: 0.59453667, Global Avg Loss: 5.80334657, Time: 0.0208 Steps: 6030, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001782, Sample Num: 28512, Cur Loss: 0.79032683, Cur Avg Loss: 0.87550734, Log Avg loss: 0.60260680, Global Avg Loss: 5.79473607, Time: 0.0208 Steps: 6040, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001792, Sample Num: 28672, Cur Loss: 0.60376000, Cur Avg Loss: 0.87388766, Log Avg loss: 0.58526095, Global Avg Loss: 5.78612537, Time: 0.0253 Steps: 6050, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001802, Sample Num: 28832, Cur Loss: 0.78687823, Cur Avg Loss: 0.87296138, Log Avg loss: 0.70697148, Global Avg Loss: 5.77774393, Time: 0.0208 Steps: 6060, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001812, Sample Num: 28992, Cur Loss: 1.01189637, Cur Avg Loss: 0.87305027, Log Avg loss: 0.88906829, Global Avg Loss: 5.76969010, Time: 0.0208 Steps: 6070, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001822, Sample Num: 29152, Cur Loss: 0.73026532, Cur Avg Loss: 0.87244671, Log Avg loss: 0.76308160, Global Avg Loss: 5.76145554, Time: 0.0208 Steps: 6080, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001832, Sample Num: 29312, Cur Loss: 0.73293072, Cur Avg Loss: 0.87124028, Log Avg loss: 0.65142960, Global Avg Loss: 5.75306470, Time: 0.0208 Steps: 6090, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001842, Sample Num: 29472, Cur Loss: 1.21539378, Cur Avg Loss: 0.87054710, Log Avg loss: 0.74355526, Global Avg Loss: 5.74485239, Time: 0.0208 Steps: 6100, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001852, Sample Num: 29632, Cur Loss: 0.44256446, Cur Avg Loss: 0.86944364, Log Avg loss: 0.66618745, Global Avg Loss: 5.73654033, Time: 0.0208 Steps: 6110, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001862, Sample Num: 29792, Cur Loss: 0.86164397, Cur Avg Loss: 0.86731343, Log Avg loss: 0.47279745, Global Avg Loss: 5.72793945, Time: 0.0208 Steps: 6120, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001872, Sample Num: 29952, Cur Loss: 0.24478811, Cur Avg Loss: 0.86678419, Log Avg loss: 0.76824113, Global Avg Loss: 5.71984858, Time: 0.0208 Steps: 6130, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001882, Sample Num: 30112, Cur Loss: 1.35375655, Cur Avg Loss: 0.86648989, Log Avg loss: 0.81139613, Global Avg Loss: 5.71185436, Time: 0.0208 Steps: 6140, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001892, Sample Num: 30272, Cur Loss: 0.66897899, Cur Avg Loss: 0.86530030, Log Avg loss: 0.64141870, Global Avg Loss: 5.70360975, Time: 0.0208 Steps: 6150, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001902, Sample Num: 30432, Cur Loss: 0.43088180, Cur Avg Loss: 0.86492376, Log Avg loss: 0.79368263, Global Avg Loss: 5.69563909, Time: 0.0208 Steps: 6160, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001912, Sample Num: 30592, Cur Loss: 0.96576601, Cur Avg Loss: 0.86541625, Log Avg loss: 0.95908816, Global Avg Loss: 5.68796235, Time: 0.0208 Steps: 6170, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001922, Sample Num: 30752, Cur Loss: 1.01301372, Cur Avg Loss: 0.86474790, Log Avg loss: 0.73695982, Global Avg Loss: 5.67995102, Time: 0.0208 Steps: 6180, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001932, Sample Num: 30912, Cur Loss: 0.72935033, Cur Avg Loss: 0.86392336, Log Avg loss: 0.70544594, Global Avg Loss: 5.67191466, Time: 0.0209 Steps: 6190, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001942, Sample Num: 31072, Cur Loss: 0.61146903, Cur Avg Loss: 0.86433240, Log Avg loss: 0.94336049, Global Avg Loss: 5.66428796, Time: 0.0208 Steps: 6200, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001952, Sample Num: 31232, Cur Loss: 0.78041238, Cur Avg Loss: 0.86495122, Log Avg loss: 0.98512490, Global Avg Loss: 5.65675307, Time: 0.0208 Steps: 6210, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001962, Sample Num: 31392, Cur Loss: 0.91365010, Cur Avg Loss: 0.86445333, Log Avg loss: 0.76726464, Global Avg Loss: 5.64889216, Time: 0.0208 Steps: 6220, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001972, Sample Num: 31552, Cur Loss: 0.79164207, Cur Avg Loss: 0.86520686, Log Avg loss: 1.01305017, Global Avg Loss: 5.64145100, Time: 0.0208 Steps: 6230, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001982, Sample Num: 31712, Cur Loss: 0.80880654, Cur Avg Loss: 0.86707081, Log Avg loss: 1.23464279, Global Avg Loss: 5.63438881, Time: 0.0208 Steps: 6240, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001992, Sample Num: 31872, Cur Loss: 0.36425620, Cur Avg Loss: 0.86728464, Log Avg loss: 0.90966547, Global Avg Loss: 5.62682925, Time: 0.0207 Steps: 6250, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002002, Sample Num: 32032, Cur Loss: 0.72653985, Cur Avg Loss: 0.86711188, Log Avg loss: 0.83269645, Global Avg Loss: 5.61917089, Time: 0.0208 Steps: 6260, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002012, Sample Num: 32192, Cur Loss: 0.31765586, Cur Avg Loss: 0.86611092, Log Avg loss: 0.66572023, Global Avg Loss: 5.61127065, Time: 0.0208 Steps: 6270, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002022, Sample Num: 32352, Cur Loss: 0.50574666, Cur Avg Loss: 0.86439016, Log Avg loss: 0.51817345, Global Avg Loss: 5.60316062, Time: 0.0208 Steps: 6280, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002032, Sample Num: 32512, Cur Loss: 1.25548661, Cur Avg Loss: 0.86360167, Log Avg loss: 0.70416800, Global Avg Loss: 5.59537208, Time: 0.0209 Steps: 6290, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002042, Sample Num: 32672, Cur Loss: 0.55340546, Cur Avg Loss: 0.86364352, Log Avg loss: 0.87214694, Global Avg Loss: 5.58787490, Time: 0.0208 Steps: 6300, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002052, Sample Num: 32832, Cur Loss: 1.00091195, Cur Avg Loss: 0.86269641, Log Avg loss: 0.66929672, Global Avg Loss: 5.58008001, Time: 0.0246 Steps: 6310, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002062, Sample Num: 32992, Cur Loss: 0.45247999, Cur Avg Loss: 0.86122708, Log Avg loss: 0.55972112, Global Avg Loss: 5.57213640, Time: 0.0208 Steps: 6320, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002072, Sample Num: 33152, Cur Loss: 0.27086109, Cur Avg Loss: 0.86058359, Log Avg loss: 0.72789690, Global Avg Loss: 5.56448357, Time: 0.0209 Steps: 6330, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002082, Sample Num: 33312, Cur Loss: 0.70786178, Cur Avg Loss: 0.86037883, Log Avg loss: 0.81795116, Global Avg Loss: 5.55699693, Time: 0.0209 Steps: 6340, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002092, Sample Num: 33472, Cur Loss: 0.55288506, Cur Avg Loss: 0.85986066, Log Avg loss: 0.75197912, Global Avg Loss: 5.54942997, Time: 0.0209 Steps: 6350, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002102, Sample Num: 33632, Cur Loss: 0.34741023, Cur Avg Loss: 0.85819534, Log Avg loss: 0.50980842, Global Avg Loss: 5.54150604, Time: 0.0209 Steps: 6360, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002112, Sample Num: 33792, Cur Loss: 0.55341673, Cur Avg Loss: 0.85759893, Log Avg loss: 0.73223557, Global Avg Loss: 5.53395616, Time: 0.0208 Steps: 6370, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002122, Sample Num: 33952, Cur Loss: 0.94569528, Cur Avg Loss: 0.85705263, Log Avg loss: 0.74167323, Global Avg Loss: 5.52644475, Time: 0.0208 Steps: 6380, Updated lr: 0.000095 ***** Running evaluation checkpoint-6387 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-6387 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.513521, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.649431, "eval_total_loss": 456.550302, "eval_mae": 0.562986, "eval_mse": 0.649681, "eval_r2": 0.58702, "eval_sp_statistic": 0.745396, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.788021, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.971269, "test_total_loss": 487.576812, "test_mae": 0.879423, "test_mse": 0.971476, "test_r2": 0.373001, "test_sp_statistic": 0.737306, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.812686, "test_ps_pvalue": 0.0, "lr": 9.489141773352301e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 5.5210974838322775, "train_cur_epoch_loss": 1823.197819903493, "train_cur_epoch_avg_loss": 0.856363466370828, "train_cur_epoch_time": 44.51352095603943, "train_cur_epoch_avg_time": 0.02090818269424116, "epoch": 3, "step": 6387} ################################################## Training, Epoch: 0004, Batch: 000003, Sample Num: 48, Cur Loss: 0.31219125, Cur Avg Loss: 1.28832793, Log Avg loss: 0.83971210, Global Avg Loss: 5.51911027, Time: 0.0244 Steps: 6390, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000013, Sample Num: 208, Cur Loss: 0.67854828, Cur Avg Loss: 1.02724131, Log Avg loss: 0.94891532, Global Avg Loss: 5.51196934, Time: 0.0208 Steps: 6400, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000023, Sample Num: 368, Cur Loss: 1.05441928, Cur Avg Loss: 0.91841845, Log Avg loss: 0.77694873, Global Avg Loss: 5.50458241, Time: 0.0208 Steps: 6410, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000033, Sample Num: 528, Cur Loss: 0.36109608, Cur Avg Loss: 0.85104243, Log Avg loss: 0.69607757, Global Avg Loss: 5.49709253, Time: 0.0208 Steps: 6420, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000043, Sample Num: 688, Cur Loss: 0.76969969, Cur Avg Loss: 0.84279668, Log Avg loss: 0.81558573, Global Avg Loss: 5.48981180, Time: 0.0208 Steps: 6430, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000053, Sample Num: 848, Cur Loss: 1.73110580, Cur Avg Loss: 0.85467460, Log Avg loss: 0.90574967, Global Avg Loss: 5.48269369, Time: 0.0207 Steps: 6440, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000063, Sample Num: 1008, Cur Loss: 1.00353062, Cur Avg Loss: 0.86074725, Log Avg loss: 0.89293227, Global Avg Loss: 5.47557778, Time: 0.0208 Steps: 6450, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000073, Sample Num: 1168, Cur Loss: 0.54705125, Cur Avg Loss: 0.84650011, Log Avg loss: 0.75674314, Global Avg Loss: 5.46827309, Time: 0.0207 Steps: 6460, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000083, Sample Num: 1328, Cur Loss: 0.33930746, Cur Avg Loss: 0.79948146, Log Avg loss: 0.45624527, Global Avg Loss: 5.46052652, Time: 0.0209 Steps: 6470, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000093, Sample Num: 1488, Cur Loss: 0.32669955, Cur Avg Loss: 0.79007474, Log Avg loss: 0.71199900, Global Avg Loss: 5.45319855, Time: 0.0209 Steps: 6480, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000103, Sample Num: 1648, Cur Loss: 0.58881706, Cur Avg Loss: 0.77827629, Log Avg loss: 0.66855071, Global Avg Loss: 5.44582621, Time: 0.0208 Steps: 6490, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000113, Sample Num: 1808, Cur Loss: 0.68778038, Cur Avg Loss: 0.77668325, Log Avg loss: 0.76027490, Global Avg Loss: 5.43861767, Time: 0.0208 Steps: 6500, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000123, Sample Num: 1968, Cur Loss: 0.95256937, Cur Avg Loss: 0.76487225, Log Avg loss: 0.63140801, Global Avg Loss: 5.43123332, Time: 0.0210 Steps: 6510, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000133, Sample Num: 2128, Cur Loss: 0.83065081, Cur Avg Loss: 0.76588497, Log Avg loss: 0.77834137, Global Avg Loss: 5.42409698, Time: 0.0209 Steps: 6520, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000143, Sample Num: 2288, Cur Loss: 0.53475189, Cur Avg Loss: 0.76692518, Log Avg loss: 0.78075997, Global Avg Loss: 5.41698621, Time: 0.0208 Steps: 6530, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000153, Sample Num: 2448, Cur Loss: 0.56720763, Cur Avg Loss: 0.76502837, Log Avg loss: 0.73790407, Global Avg Loss: 5.40983165, Time: 0.0209 Steps: 6540, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000163, Sample Num: 2608, Cur Loss: 0.62948024, Cur Avg Loss: 0.76639224, Log Avg loss: 0.78725941, Global Avg Loss: 5.40277428, Time: 0.0209 Steps: 6550, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000173, Sample Num: 2768, Cur Loss: 0.93295807, Cur Avg Loss: 0.76182780, Log Avg loss: 0.68742747, Global Avg Loss: 5.39558626, Time: 0.0209 Steps: 6560, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000183, Sample Num: 2928, Cur Loss: 0.95177215, Cur Avg Loss: 0.76657943, Log Avg loss: 0.84878258, Global Avg Loss: 5.38866570, Time: 0.0209 Steps: 6570, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000193, Sample Num: 3088, Cur Loss: 0.37397701, Cur Avg Loss: 0.76560180, Log Avg loss: 0.74771118, Global Avg Loss: 5.38161258, Time: 0.0209 Steps: 6580, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000203, Sample Num: 3248, Cur Loss: 0.38159680, Cur Avg Loss: 0.75526930, Log Avg loss: 0.55585203, Global Avg Loss: 5.37428973, Time: 0.0208 Steps: 6590, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000213, Sample Num: 3408, Cur Loss: 0.10529943, Cur Avg Loss: 0.74261720, Log Avg loss: 0.48577958, Global Avg Loss: 5.36688289, Time: 0.0209 Steps: 6600, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000223, Sample Num: 3568, Cur Loss: 0.25998607, Cur Avg Loss: 0.73255395, Log Avg loss: 0.51820678, Global Avg Loss: 5.35954753, Time: 0.0208 Steps: 6610, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000233, Sample Num: 3728, Cur Loss: 1.26991367, Cur Avg Loss: 0.73313252, Log Avg loss: 0.74603448, Global Avg Loss: 5.35257848, Time: 0.0209 Steps: 6620, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000243, Sample Num: 3888, Cur Loss: 0.70027769, Cur Avg Loss: 0.72722979, Log Avg loss: 0.58969634, Global Avg Loss: 5.34539464, Time: 0.0208 Steps: 6630, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000253, Sample Num: 4048, Cur Loss: 0.94938505, Cur Avg Loss: 0.73584085, Log Avg loss: 0.94508956, Global Avg Loss: 5.33876768, Time: 0.0209 Steps: 6640, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000263, Sample Num: 4208, Cur Loss: 1.76051116, Cur Avg Loss: 0.74103811, Log Avg loss: 0.87252887, Global Avg Loss: 5.33205153, Time: 0.0210 Steps: 6650, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000273, Sample Num: 4368, Cur Loss: 0.94200224, Cur Avg Loss: 0.74754208, Log Avg loss: 0.91859636, Global Avg Loss: 5.32542472, Time: 0.0207 Steps: 6660, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000283, Sample Num: 4528, Cur Loss: 0.47506946, Cur Avg Loss: 0.74302673, Log Avg loss: 0.61975774, Global Avg Loss: 5.31836974, Time: 0.0208 Steps: 6670, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000293, Sample Num: 4688, Cur Loss: 1.01262879, Cur Avg Loss: 0.73806603, Log Avg loss: 0.59767833, Global Avg Loss: 5.31130284, Time: 0.0207 Steps: 6680, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000303, Sample Num: 4848, Cur Loss: 0.89174110, Cur Avg Loss: 0.73544811, Log Avg loss: 0.65874278, Global Avg Loss: 5.30434834, Time: 0.0207 Steps: 6690, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000313, Sample Num: 5008, Cur Loss: 2.13905525, Cur Avg Loss: 0.74232372, Log Avg loss: 0.95065493, Global Avg Loss: 5.29785029, Time: 0.0208 Steps: 6700, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000323, Sample Num: 5168, Cur Loss: 0.18489106, Cur Avg Loss: 0.75875763, Log Avg loss: 1.27313900, Global Avg Loss: 5.29185221, Time: 0.0207 Steps: 6710, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000333, Sample Num: 5328, Cur Loss: 0.57469994, Cur Avg Loss: 0.75513109, Log Avg loss: 0.63799382, Global Avg Loss: 5.28492683, Time: 0.0207 Steps: 6720, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000343, Sample Num: 5488, Cur Loss: 0.77297777, Cur Avg Loss: 0.75270436, Log Avg loss: 0.67189429, Global Avg Loss: 5.27807240, Time: 0.0208 Steps: 6730, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000353, Sample Num: 5648, Cur Loss: 0.38557366, Cur Avg Loss: 0.74641274, Log Avg loss: 0.53061026, Global Avg Loss: 5.27102868, Time: 0.0207 Steps: 6740, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000363, Sample Num: 5808, Cur Loss: 0.56807745, Cur Avg Loss: 0.74568250, Log Avg loss: 0.71990489, Global Avg Loss: 5.26428628, Time: 0.0208 Steps: 6750, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000373, Sample Num: 5968, Cur Loss: 0.62600499, Cur Avg Loss: 0.74979847, Log Avg loss: 0.89920830, Global Avg Loss: 5.25782906, Time: 0.0209 Steps: 6760, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000383, Sample Num: 6128, Cur Loss: 0.31414574, Cur Avg Loss: 0.74637703, Log Avg loss: 0.61875705, Global Avg Loss: 5.25097667, Time: 0.0209 Steps: 6770, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000393, Sample Num: 6288, Cur Loss: 0.88870931, Cur Avg Loss: 0.74497330, Log Avg loss: 0.69121062, Global Avg Loss: 5.24425135, Time: 0.0208 Steps: 6780, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000403, Sample Num: 6448, Cur Loss: 0.29621542, Cur Avg Loss: 0.74507542, Log Avg loss: 0.74908850, Global Avg Loss: 5.23763108, Time: 0.0209 Steps: 6790, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000413, Sample Num: 6608, Cur Loss: 0.58823884, Cur Avg Loss: 0.73904616, Log Avg loss: 0.49606725, Global Avg Loss: 5.23065819, Time: 0.0207 Steps: 6800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000423, Sample Num: 6768, Cur Loss: 2.36867356, Cur Avg Loss: 0.76106094, Log Avg loss: 1.67027123, Global Avg Loss: 5.22543002, Time: 0.0208 Steps: 6810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000433, Sample Num: 6928, Cur Loss: 0.77835834, Cur Avg Loss: 0.75922981, Log Avg loss: 0.68177311, Global Avg Loss: 5.21876776, Time: 0.0208 Steps: 6820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000443, Sample Num: 7088, Cur Loss: 0.56857049, Cur Avg Loss: 0.75533490, Log Avg loss: 0.58668517, Global Avg Loss: 5.21198580, Time: 0.0208 Steps: 6830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000453, Sample Num: 7248, Cur Loss: 0.15921597, Cur Avg Loss: 0.74833985, Log Avg loss: 0.43845907, Global Avg Loss: 5.20500696, Time: 0.0208 Steps: 6840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000463, Sample Num: 7408, Cur Loss: 1.01694238, Cur Avg Loss: 0.75002921, Log Avg loss: 0.82655713, Global Avg Loss: 5.19861506, Time: 0.0208 Steps: 6850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000473, Sample Num: 7568, Cur Loss: 0.24747676, Cur Avg Loss: 0.74768955, Log Avg loss: 0.63936362, Global Avg Loss: 5.19196892, Time: 0.0208 Steps: 6860, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000483, Sample Num: 7728, Cur Loss: 0.51802576, Cur Avg Loss: 0.74328843, Log Avg loss: 0.53511539, Global Avg Loss: 5.18519038, Time: 0.0209 Steps: 6870, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000493, Sample Num: 7888, Cur Loss: 0.94139576, Cur Avg Loss: 0.74323885, Log Avg loss: 0.74084408, Global Avg Loss: 5.17873058, Time: 0.0209 Steps: 6880, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000503, Sample Num: 8048, Cur Loss: 0.81963122, Cur Avg Loss: 0.74210239, Log Avg loss: 0.68607500, Global Avg Loss: 5.17221003, Time: 0.0209 Steps: 6890, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000513, Sample Num: 8208, Cur Loss: 0.78931153, Cur Avg Loss: 0.74387744, Log Avg loss: 0.83316248, Global Avg Loss: 5.16592156, Time: 0.0244 Steps: 6900, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000523, Sample Num: 8368, Cur Loss: 0.52313292, Cur Avg Loss: 0.74400092, Log Avg loss: 0.75033512, Global Avg Loss: 5.15953142, Time: 0.0208 Steps: 6910, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000533, Sample Num: 8528, Cur Loss: 0.40665844, Cur Avg Loss: 0.75309171, Log Avg loss: 1.22854024, Global Avg Loss: 5.15385080, Time: 0.0208 Steps: 6920, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000543, Sample Num: 8688, Cur Loss: 0.56890601, Cur Avg Loss: 0.75737205, Log Avg loss: 0.98551408, Global Avg Loss: 5.14783588, Time: 0.0208 Steps: 6930, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000553, Sample Num: 8848, Cur Loss: 1.02508342, Cur Avg Loss: 0.76453505, Log Avg loss: 1.15348593, Global Avg Loss: 5.14208033, Time: 0.0208 Steps: 6940, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000563, Sample Num: 9008, Cur Loss: 0.32338914, Cur Avg Loss: 0.76686946, Log Avg loss: 0.89596251, Global Avg Loss: 5.13597081, Time: 0.0208 Steps: 6950, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000573, Sample Num: 9168, Cur Loss: 0.67118609, Cur Avg Loss: 0.76665343, Log Avg loss: 0.75449090, Global Avg Loss: 5.12967558, Time: 0.0208 Steps: 6960, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000583, Sample Num: 9328, Cur Loss: 1.52819085, Cur Avg Loss: 0.76789184, Log Avg loss: 0.83885268, Global Avg Loss: 5.12351945, Time: 0.0208 Steps: 6970, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000593, Sample Num: 9488, Cur Loss: 0.44254744, Cur Avg Loss: 0.76612673, Log Avg loss: 0.66322068, Global Avg Loss: 5.11712934, Time: 0.0208 Steps: 6980, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000603, Sample Num: 9648, Cur Loss: 0.61249363, Cur Avg Loss: 0.76877609, Log Avg loss: 0.92588360, Global Avg Loss: 5.11113328, Time: 0.0209 Steps: 6990, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000613, Sample Num: 9808, Cur Loss: 0.62106317, Cur Avg Loss: 0.77017102, Log Avg loss: 0.85428512, Global Avg Loss: 5.10505207, Time: 0.0208 Steps: 7000, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000623, Sample Num: 9968, Cur Loss: 0.62645751, Cur Avg Loss: 0.76727728, Log Avg loss: 0.58989107, Global Avg Loss: 5.09861104, Time: 0.0208 Steps: 7010, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000633, Sample Num: 10128, Cur Loss: 0.44896281, Cur Avg Loss: 0.76949372, Log Avg loss: 0.90757807, Global Avg Loss: 5.09264091, Time: 0.0208 Steps: 7020, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000643, Sample Num: 10288, Cur Loss: 0.45536429, Cur Avg Loss: 0.76722591, Log Avg loss: 0.62367316, Global Avg Loss: 5.08628391, Time: 0.0208 Steps: 7030, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000653, Sample Num: 10448, Cur Loss: 0.50917888, Cur Avg Loss: 0.76630407, Log Avg loss: 0.70702961, Global Avg Loss: 5.08006338, Time: 0.0208 Steps: 7040, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000663, Sample Num: 10608, Cur Loss: 1.82172692, Cur Avg Loss: 0.76622534, Log Avg loss: 0.76108454, Global Avg Loss: 5.07393717, Time: 0.0208 Steps: 7050, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000673, Sample Num: 10768, Cur Loss: 0.56555086, Cur Avg Loss: 0.76164177, Log Avg loss: 0.45775073, Global Avg Loss: 5.06739866, Time: 0.0208 Steps: 7060, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000683, Sample Num: 10928, Cur Loss: 0.13250633, Cur Avg Loss: 0.75822974, Log Avg loss: 0.52860043, Global Avg Loss: 5.06097886, Time: 0.0208 Steps: 7070, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000693, Sample Num: 11088, Cur Loss: 0.31597877, Cur Avg Loss: 0.75685410, Log Avg loss: 0.66289789, Global Avg Loss: 5.05476688, Time: 0.0208 Steps: 7080, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000703, Sample Num: 11248, Cur Loss: 0.86431086, Cur Avg Loss: 0.75414154, Log Avg loss: 0.56616147, Global Avg Loss: 5.04843599, Time: 0.0209 Steps: 7090, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000713, Sample Num: 11408, Cur Loss: 0.84018970, Cur Avg Loss: 0.76037987, Log Avg loss: 1.19893397, Global Avg Loss: 5.04301415, Time: 0.0208 Steps: 7100, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000723, Sample Num: 11568, Cur Loss: 0.25971124, Cur Avg Loss: 0.75741455, Log Avg loss: 0.54598756, Global Avg Loss: 5.03668922, Time: 0.0209 Steps: 7110, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000733, Sample Num: 11728, Cur Loss: 0.83385646, Cur Avg Loss: 0.75567394, Log Avg loss: 0.62982781, Global Avg Loss: 5.03049981, Time: 0.0209 Steps: 7120, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000743, Sample Num: 11888, Cur Loss: 0.61237979, Cur Avg Loss: 0.75444180, Log Avg loss: 0.66412578, Global Avg Loss: 5.02437586, Time: 0.0209 Steps: 7130, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000753, Sample Num: 12048, Cur Loss: 1.03306746, Cur Avg Loss: 0.75445227, Log Avg loss: 0.75523015, Global Avg Loss: 5.01839666, Time: 0.0208 Steps: 7140, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000763, Sample Num: 12208, Cur Loss: 0.23759240, Cur Avg Loss: 0.75223298, Log Avg loss: 0.58512041, Global Avg Loss: 5.01219628, Time: 0.0208 Steps: 7150, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000773, Sample Num: 12368, Cur Loss: 1.01260018, Cur Avg Loss: 0.75681480, Log Avg loss: 1.10640738, Global Avg Loss: 5.00674127, Time: 0.0209 Steps: 7160, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000783, Sample Num: 12528, Cur Loss: 0.46974194, Cur Avg Loss: 0.75804335, Log Avg loss: 0.85301034, Global Avg Loss: 5.00094806, Time: 0.0209 Steps: 7170, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000793, Sample Num: 12688, Cur Loss: 0.96901453, Cur Avg Loss: 0.75705419, Log Avg loss: 0.67960361, Global Avg Loss: 4.99492947, Time: 0.0207 Steps: 7180, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000803, Sample Num: 12848, Cur Loss: 1.14858413, Cur Avg Loss: 0.76403653, Log Avg loss: 1.31773544, Global Avg Loss: 4.98981515, Time: 0.0208 Steps: 7190, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000813, Sample Num: 13008, Cur Loss: 0.44161615, Cur Avg Loss: 0.76668063, Log Avg loss: 0.97900250, Global Avg Loss: 4.98424458, Time: 0.0210 Steps: 7200, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000823, Sample Num: 13168, Cur Loss: 0.24555296, Cur Avg Loss: 0.76586256, Log Avg loss: 0.69935291, Global Avg Loss: 4.97830160, Time: 0.0209 Steps: 7210, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000833, Sample Num: 13328, Cur Loss: 1.46147716, Cur Avg Loss: 0.76420605, Log Avg loss: 0.62787541, Global Avg Loss: 4.97227608, Time: 0.0208 Steps: 7220, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000843, Sample Num: 13488, Cur Loss: 1.06162238, Cur Avg Loss: 0.76579736, Log Avg loss: 0.89835344, Global Avg Loss: 4.96664133, Time: 0.0208 Steps: 7230, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000853, Sample Num: 13648, Cur Loss: 0.91590142, Cur Avg Loss: 0.76585001, Log Avg loss: 0.77028863, Global Avg Loss: 4.96084526, Time: 0.0209 Steps: 7240, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000863, Sample Num: 13808, Cur Loss: 2.24950862, Cur Avg Loss: 0.76762417, Log Avg loss: 0.91896024, Global Avg Loss: 4.95527025, Time: 0.0208 Steps: 7250, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000873, Sample Num: 13968, Cur Loss: 0.79375648, Cur Avg Loss: 0.77138499, Log Avg loss: 1.09594309, Global Avg Loss: 4.94995437, Time: 0.0209 Steps: 7260, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000883, Sample Num: 14128, Cur Loss: 0.99261796, Cur Avg Loss: 0.76950937, Log Avg loss: 0.60576789, Global Avg Loss: 4.94397887, Time: 0.0209 Steps: 7270, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000893, Sample Num: 14288, Cur Loss: 0.59843546, Cur Avg Loss: 0.76765050, Log Avg loss: 0.60351232, Global Avg Loss: 4.93801669, Time: 0.0209 Steps: 7280, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000903, Sample Num: 14448, Cur Loss: 0.86535978, Cur Avg Loss: 0.76817242, Log Avg loss: 0.81477983, Global Avg Loss: 4.93236068, Time: 0.0210 Steps: 7290, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000913, Sample Num: 14608, Cur Loss: 1.44741988, Cur Avg Loss: 0.76837234, Log Avg loss: 0.78642520, Global Avg Loss: 4.92668131, Time: 0.0208 Steps: 7300, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000923, Sample Num: 14768, Cur Loss: 2.56302762, Cur Avg Loss: 0.77581733, Log Avg loss: 1.45554498, Global Avg Loss: 4.92193284, Time: 0.0208 Steps: 7310, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000933, Sample Num: 14928, Cur Loss: 1.17163599, Cur Avg Loss: 0.78120615, Log Avg loss: 1.27859392, Global Avg Loss: 4.91695560, Time: 0.0208 Steps: 7320, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000943, Sample Num: 15088, Cur Loss: 3.25470638, Cur Avg Loss: 0.78837550, Log Avg loss: 1.45727605, Global Avg Loss: 4.91223571, Time: 0.0209 Steps: 7330, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000953, Sample Num: 15248, Cur Loss: 1.28429532, Cur Avg Loss: 0.78914742, Log Avg loss: 0.86193996, Global Avg Loss: 4.90671759, Time: 0.0209 Steps: 7340, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000963, Sample Num: 15408, Cur Loss: 0.53358901, Cur Avg Loss: 0.79372771, Log Avg loss: 1.23022943, Global Avg Loss: 4.90171557, Time: 0.0208 Steps: 7350, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000973, Sample Num: 15568, Cur Loss: 0.75821757, Cur Avg Loss: 0.79263156, Log Avg loss: 0.68707161, Global Avg Loss: 4.89598915, Time: 0.0208 Steps: 7360, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000983, Sample Num: 15728, Cur Loss: 0.45322654, Cur Avg Loss: 0.79066110, Log Avg loss: 0.59893568, Global Avg Loss: 4.89015868, Time: 0.0209 Steps: 7370, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000993, Sample Num: 15888, Cur Loss: 0.43666667, Cur Avg Loss: 0.78930649, Log Avg loss: 0.65614818, Global Avg Loss: 4.88442154, Time: 0.0208 Steps: 7380, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001003, Sample Num: 16048, Cur Loss: 0.77341568, Cur Avg Loss: 0.78777126, Log Avg loss: 0.63532340, Global Avg Loss: 4.87867175, Time: 0.0208 Steps: 7390, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001013, Sample Num: 16208, Cur Loss: 0.32451257, Cur Avg Loss: 0.78536741, Log Avg loss: 0.54426059, Global Avg Loss: 4.87281443, Time: 0.0209 Steps: 7400, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001023, Sample Num: 16368, Cur Loss: 0.63992566, Cur Avg Loss: 0.78781750, Log Avg loss: 1.03601225, Global Avg Loss: 4.86763656, Time: 0.0208 Steps: 7410, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001033, Sample Num: 16528, Cur Loss: 0.62236911, Cur Avg Loss: 0.78825569, Log Avg loss: 0.83308245, Global Avg Loss: 4.86219916, Time: 0.0207 Steps: 7420, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001043, Sample Num: 16688, Cur Loss: 1.26205814, Cur Avg Loss: 0.78607448, Log Avg loss: 0.56075513, Global Avg Loss: 4.85640987, Time: 0.0207 Steps: 7430, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001053, Sample Num: 16848, Cur Loss: 0.27325040, Cur Avg Loss: 0.78390651, Log Avg loss: 0.55778748, Global Avg Loss: 4.85063215, Time: 0.0208 Steps: 7440, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001063, Sample Num: 17008, Cur Loss: 0.67476493, Cur Avg Loss: 0.78097066, Log Avg loss: 0.47182536, Global Avg Loss: 4.84475456, Time: 0.0207 Steps: 7450, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001073, Sample Num: 17168, Cur Loss: 0.90892076, Cur Avg Loss: 0.77994566, Log Avg loss: 0.67098874, Global Avg Loss: 4.83915970, Time: 0.0208 Steps: 7460, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001083, Sample Num: 17328, Cur Loss: 0.73618132, Cur Avg Loss: 0.77805117, Log Avg loss: 0.57477183, Global Avg Loss: 4.83345101, Time: 0.0207 Steps: 7470, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001093, Sample Num: 17488, Cur Loss: 0.72577751, Cur Avg Loss: 0.77635186, Log Avg loss: 0.59231716, Global Avg Loss: 4.82778104, Time: 0.0207 Steps: 7480, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001103, Sample Num: 17648, Cur Loss: 0.81484097, Cur Avg Loss: 0.77498251, Log Avg loss: 0.62531165, Global Avg Loss: 4.82217027, Time: 0.0208 Steps: 7490, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001113, Sample Num: 17808, Cur Loss: 0.39240068, Cur Avg Loss: 0.77389810, Log Avg loss: 0.65428775, Global Avg Loss: 4.81661309, Time: 0.0208 Steps: 7500, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001123, Sample Num: 17968, Cur Loss: 0.58014154, Cur Avg Loss: 0.77415460, Log Avg loss: 0.80270309, Global Avg Loss: 4.81126834, Time: 0.0207 Steps: 7510, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001133, Sample Num: 18128, Cur Loss: 0.80465597, Cur Avg Loss: 0.77297999, Log Avg loss: 0.64107156, Global Avg Loss: 4.80572287, Time: 0.0207 Steps: 7520, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001143, Sample Num: 18288, Cur Loss: 0.62153649, Cur Avg Loss: 0.77287231, Log Avg loss: 0.76067246, Global Avg Loss: 4.80035095, Time: 0.0207 Steps: 7530, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001153, Sample Num: 18448, Cur Loss: 0.41618508, Cur Avg Loss: 0.77036210, Log Avg loss: 0.48344447, Global Avg Loss: 4.79462561, Time: 0.0207 Steps: 7540, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001163, Sample Num: 18608, Cur Loss: 0.55034524, Cur Avg Loss: 0.76854064, Log Avg loss: 0.55852668, Global Avg Loss: 4.78901489, Time: 0.0207 Steps: 7550, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001173, Sample Num: 18768, Cur Loss: 0.59884596, Cur Avg Loss: 0.76751207, Log Avg loss: 0.64788956, Global Avg Loss: 4.78353721, Time: 0.0208 Steps: 7560, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001183, Sample Num: 18928, Cur Loss: 0.84220409, Cur Avg Loss: 0.76864503, Log Avg loss: 0.90154077, Global Avg Loss: 4.77840907, Time: 0.0207 Steps: 7570, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001193, Sample Num: 19088, Cur Loss: 0.44292399, Cur Avg Loss: 0.77014139, Log Avg loss: 0.94716101, Global Avg Loss: 4.77335466, Time: 0.0207 Steps: 7580, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001203, Sample Num: 19248, Cur Loss: 0.87834227, Cur Avg Loss: 0.76918575, Log Avg loss: 0.65517744, Global Avg Loss: 4.76792886, Time: 0.0207 Steps: 7590, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001213, Sample Num: 19408, Cur Loss: 1.11940026, Cur Avg Loss: 0.77122844, Log Avg loss: 1.01696433, Global Avg Loss: 4.76299338, Time: 0.0207 Steps: 7600, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001223, Sample Num: 19568, Cur Loss: 0.40559447, Cur Avg Loss: 0.77197525, Log Avg loss: 0.86256329, Global Avg Loss: 4.75786798, Time: 0.0208 Steps: 7610, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001233, Sample Num: 19728, Cur Loss: 0.94326293, Cur Avg Loss: 0.77008461, Log Avg loss: 0.53885997, Global Avg Loss: 4.75233123, Time: 0.0208 Steps: 7620, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001243, Sample Num: 19888, Cur Loss: 1.29913735, Cur Avg Loss: 0.77017927, Log Avg loss: 0.78185045, Global Avg Loss: 4.74712745, Time: 0.0208 Steps: 7630, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001253, Sample Num: 20048, Cur Loss: 0.28512335, Cur Avg Loss: 0.77010734, Log Avg loss: 0.76116632, Global Avg Loss: 4.74191023, Time: 0.0207 Steps: 7640, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001263, Sample Num: 20208, Cur Loss: 1.06188512, Cur Avg Loss: 0.77194940, Log Avg loss: 1.00275973, Global Avg Loss: 4.73702245, Time: 0.0208 Steps: 7650, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001273, Sample Num: 20368, Cur Loss: 0.87320787, Cur Avg Loss: 0.77197590, Log Avg loss: 0.77532270, Global Avg Loss: 4.73185052, Time: 0.0207 Steps: 7660, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001283, Sample Num: 20528, Cur Loss: 0.29088911, Cur Avg Loss: 0.77170785, Log Avg loss: 0.73758539, Global Avg Loss: 4.72664287, Time: 0.0244 Steps: 7670, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001293, Sample Num: 20688, Cur Loss: 0.62278634, Cur Avg Loss: 0.77239883, Log Avg loss: 0.86105172, Global Avg Loss: 4.72160955, Time: 0.0209 Steps: 7680, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001303, Sample Num: 20848, Cur Loss: 0.86868376, Cur Avg Loss: 0.77402386, Log Avg loss: 0.98414047, Global Avg Loss: 4.71674938, Time: 0.0208 Steps: 7690, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001313, Sample Num: 21008, Cur Loss: 0.52228785, Cur Avg Loss: 0.77496936, Log Avg loss: 0.89816770, Global Avg Loss: 4.71179018, Time: 0.0208 Steps: 7700, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001323, Sample Num: 21168, Cur Loss: 0.55381483, Cur Avg Loss: 0.77334889, Log Avg loss: 0.56058130, Global Avg Loss: 4.70640599, Time: 0.0208 Steps: 7710, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001333, Sample Num: 21328, Cur Loss: 1.05485058, Cur Avg Loss: 0.77298218, Log Avg loss: 0.72446649, Global Avg Loss: 4.70124804, Time: 0.0208 Steps: 7720, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001343, Sample Num: 21488, Cur Loss: 0.35349554, Cur Avg Loss: 0.77183784, Log Avg loss: 0.61929655, Global Avg Loss: 4.69596738, Time: 0.0208 Steps: 7730, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001353, Sample Num: 21648, Cur Loss: 0.35834387, Cur Avg Loss: 0.77118669, Log Avg loss: 0.68373793, Global Avg Loss: 4.69078362, Time: 0.0208 Steps: 7740, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001363, Sample Num: 21808, Cur Loss: 0.62841946, Cur Avg Loss: 0.77084355, Log Avg loss: 0.72441571, Global Avg Loss: 4.68566573, Time: 0.0207 Steps: 7750, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001373, Sample Num: 21968, Cur Loss: 0.55352288, Cur Avg Loss: 0.76896554, Log Avg loss: 0.51299395, Global Avg Loss: 4.68028857, Time: 0.0208 Steps: 7760, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001383, Sample Num: 22128, Cur Loss: 0.80652028, Cur Avg Loss: 0.76819655, Log Avg loss: 0.66261419, Global Avg Loss: 4.67511782, Time: 0.0207 Steps: 7770, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001393, Sample Num: 22288, Cur Loss: 0.57954216, Cur Avg Loss: 0.76662677, Log Avg loss: 0.54952509, Global Avg Loss: 4.66981500, Time: 0.0207 Steps: 7780, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001403, Sample Num: 22448, Cur Loss: 0.78848159, Cur Avg Loss: 0.76655890, Log Avg loss: 0.75710521, Global Avg Loss: 4.66479227, Time: 0.0208 Steps: 7790, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001413, Sample Num: 22608, Cur Loss: 0.17741387, Cur Avg Loss: 0.76505419, Log Avg loss: 0.55394393, Global Avg Loss: 4.65952195, Time: 0.0208 Steps: 7800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001423, Sample Num: 22768, Cur Loss: 0.55239826, Cur Avg Loss: 0.76475440, Log Avg loss: 0.72239387, Global Avg Loss: 4.65448081, Time: 0.0208 Steps: 7810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001433, Sample Num: 22928, Cur Loss: 0.99747574, Cur Avg Loss: 0.76457252, Log Avg loss: 0.73869057, Global Avg Loss: 4.64947341, Time: 0.0207 Steps: 7820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001443, Sample Num: 23088, Cur Loss: 0.48552257, Cur Avg Loss: 0.76388477, Log Avg loss: 0.66532975, Global Avg Loss: 4.64438510, Time: 0.0208 Steps: 7830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001453, Sample Num: 23248, Cur Loss: 0.88249123, Cur Avg Loss: 0.76597706, Log Avg loss: 1.06789490, Global Avg Loss: 4.63982325, Time: 0.0209 Steps: 7840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001463, Sample Num: 23408, Cur Loss: 0.33880138, Cur Avg Loss: 0.76781845, Log Avg loss: 1.03537297, Global Avg Loss: 4.63523160, Time: 0.0208 Steps: 7850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001473, Sample Num: 23568, Cur Loss: 0.74516213, Cur Avg Loss: 0.76588731, Log Avg loss: 0.48336056, Global Avg Loss: 4.62994932, Time: 0.0207 Steps: 7860, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001483, Sample Num: 23728, Cur Loss: 1.04196644, Cur Avg Loss: 0.76722039, Log Avg loss: 0.96358368, Global Avg Loss: 4.62529066, Time: 0.0208 Steps: 7870, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001493, Sample Num: 23888, Cur Loss: 0.27668527, Cur Avg Loss: 0.76572103, Log Avg loss: 0.54336656, Global Avg Loss: 4.62011055, Time: 0.0208 Steps: 7880, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001503, Sample Num: 24048, Cur Loss: 0.62624371, Cur Avg Loss: 0.76431123, Log Avg loss: 0.55382758, Global Avg Loss: 4.61495683, Time: 0.0208 Steps: 7890, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001513, Sample Num: 24208, Cur Loss: 0.65019649, Cur Avg Loss: 0.76344775, Log Avg loss: 0.63366704, Global Avg Loss: 4.60991723, Time: 0.0208 Steps: 7900, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001523, Sample Num: 24368, Cur Loss: 0.77227134, Cur Avg Loss: 0.76274978, Log Avg loss: 0.65714579, Global Avg Loss: 4.60492004, Time: 0.0208 Steps: 7910, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001533, Sample Num: 24528, Cur Loss: 0.56422579, Cur Avg Loss: 0.76386800, Log Avg loss: 0.93417287, Global Avg Loss: 4.60028526, Time: 0.0207 Steps: 7920, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001543, Sample Num: 24688, Cur Loss: 0.79739577, Cur Avg Loss: 0.76321547, Log Avg loss: 0.66318315, Global Avg Loss: 4.59532044, Time: 0.0213 Steps: 7930, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001553, Sample Num: 24848, Cur Loss: 0.95494145, Cur Avg Loss: 0.76388089, Log Avg loss: 0.86655542, Global Avg Loss: 4.59062426, Time: 0.0208 Steps: 7940, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001563, Sample Num: 25008, Cur Loss: 0.43260491, Cur Avg Loss: 0.76436800, Log Avg loss: 0.84001559, Global Avg Loss: 4.58590652, Time: 0.0208 Steps: 7950, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001573, Sample Num: 25168, Cur Loss: 0.55243719, Cur Avg Loss: 0.76373193, Log Avg loss: 0.66431404, Global Avg Loss: 4.58097989, Time: 0.0208 Steps: 7960, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001583, Sample Num: 25328, Cur Loss: 0.98296094, Cur Avg Loss: 0.76284100, Log Avg loss: 0.62269884, Global Avg Loss: 4.57601342, Time: 0.0208 Steps: 7970, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001593, Sample Num: 25488, Cur Loss: 0.94922119, Cur Avg Loss: 0.76211294, Log Avg loss: 0.64686099, Global Avg Loss: 4.57108967, Time: 0.0208 Steps: 7980, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001603, Sample Num: 25648, Cur Loss: 0.64993125, Cur Avg Loss: 0.76059722, Log Avg loss: 0.51914228, Global Avg Loss: 4.56601839, Time: 0.0207 Steps: 7990, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001613, Sample Num: 25808, Cur Loss: 0.93840575, Cur Avg Loss: 0.76071469, Log Avg loss: 0.77954523, Global Avg Loss: 4.56128530, Time: 0.0207 Steps: 8000, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001623, Sample Num: 25968, Cur Loss: 0.71764702, Cur Avg Loss: 0.75930950, Log Avg loss: 0.53265291, Global Avg Loss: 4.55625580, Time: 0.0209 Steps: 8010, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001633, Sample Num: 26128, Cur Loss: 0.23290724, Cur Avg Loss: 0.75820592, Log Avg loss: 0.57909416, Global Avg Loss: 4.55129674, Time: 0.0207 Steps: 8020, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001643, Sample Num: 26288, Cur Loss: 0.45177281, Cur Avg Loss: 0.75845073, Log Avg loss: 0.79842904, Global Avg Loss: 4.54662319, Time: 0.0208 Steps: 8030, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001653, Sample Num: 26448, Cur Loss: 0.72176099, Cur Avg Loss: 0.75718315, Log Avg loss: 0.54891914, Global Avg Loss: 4.54165092, Time: 0.0208 Steps: 8040, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001663, Sample Num: 26608, Cur Loss: 0.75215483, Cur Avg Loss: 0.75711843, Log Avg loss: 0.74642009, Global Avg Loss: 4.53693634, Time: 0.0208 Steps: 8050, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001673, Sample Num: 26768, Cur Loss: 0.69687170, Cur Avg Loss: 0.75771839, Log Avg loss: 0.85749163, Global Avg Loss: 4.53237128, Time: 0.0208 Steps: 8060, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001683, Sample Num: 26928, Cur Loss: 0.45472521, Cur Avg Loss: 0.75653348, Log Avg loss: 0.55829916, Global Avg Loss: 4.52744678, Time: 0.0208 Steps: 8070, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001693, Sample Num: 27088, Cur Loss: 1.07026827, Cur Avg Loss: 0.75667256, Log Avg loss: 0.78007888, Global Avg Loss: 4.52280894, Time: 0.0208 Steps: 8080, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001703, Sample Num: 27248, Cur Loss: 0.53545904, Cur Avg Loss: 0.75564090, Log Avg loss: 0.58098162, Global Avg Loss: 4.51793648, Time: 0.0208 Steps: 8090, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001713, Sample Num: 27408, Cur Loss: 0.64661294, Cur Avg Loss: 0.75483754, Log Avg loss: 0.61802434, Global Avg Loss: 4.51312177, Time: 0.0208 Steps: 8100, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001723, Sample Num: 27568, Cur Loss: 0.75463027, Cur Avg Loss: 0.75566716, Log Avg loss: 0.89778102, Global Avg Loss: 4.50866389, Time: 0.0208 Steps: 8110, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001733, Sample Num: 27728, Cur Loss: 0.62698197, Cur Avg Loss: 0.75478493, Log Avg loss: 0.60277776, Global Avg Loss: 4.50385368, Time: 0.0208 Steps: 8120, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001743, Sample Num: 27888, Cur Loss: 0.97820890, Cur Avg Loss: 0.75519170, Log Avg loss: 0.82568365, Global Avg Loss: 4.49932949, Time: 0.0208 Steps: 8130, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001753, Sample Num: 28048, Cur Loss: 1.36341226, Cur Avg Loss: 0.75752083, Log Avg loss: 1.16348802, Global Avg Loss: 4.49523140, Time: 0.0208 Steps: 8140, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001763, Sample Num: 28208, Cur Loss: 0.46431163, Cur Avg Loss: 0.75669150, Log Avg loss: 0.61130992, Global Avg Loss: 4.49046586, Time: 0.0208 Steps: 8150, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001773, Sample Num: 28368, Cur Loss: 0.32000399, Cur Avg Loss: 0.75623194, Log Avg loss: 0.67521229, Global Avg Loss: 4.48579030, Time: 0.0208 Steps: 8160, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001783, Sample Num: 28528, Cur Loss: 0.55236840, Cur Avg Loss: 0.75550507, Log Avg loss: 0.62663161, Global Avg Loss: 4.48106673, Time: 0.0208 Steps: 8170, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001793, Sample Num: 28688, Cur Loss: 0.74624699, Cur Avg Loss: 0.75584666, Log Avg loss: 0.81675164, Global Avg Loss: 4.47658713, Time: 0.0244 Steps: 8180, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001803, Sample Num: 28848, Cur Loss: 0.46535355, Cur Avg Loss: 0.75553373, Log Avg loss: 0.69942619, Global Avg Loss: 4.47197521, Time: 0.0208 Steps: 8190, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001813, Sample Num: 29008, Cur Loss: 0.49378154, Cur Avg Loss: 0.75554185, Log Avg loss: 0.75700573, Global Avg Loss: 4.46744476, Time: 0.0209 Steps: 8200, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001823, Sample Num: 29168, Cur Loss: 0.30915552, Cur Avg Loss: 0.75635820, Log Avg loss: 0.90436099, Global Avg Loss: 4.46310483, Time: 0.0208 Steps: 8210, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001833, Sample Num: 29328, Cur Loss: 0.42497864, Cur Avg Loss: 0.75888754, Log Avg loss: 1.21998613, Global Avg Loss: 4.45915943, Time: 0.0208 Steps: 8220, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001843, Sample Num: 29488, Cur Loss: 0.72304547, Cur Avg Loss: 0.75852698, Log Avg loss: 0.69243685, Global Avg Loss: 4.45458261, Time: 0.0208 Steps: 8230, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001853, Sample Num: 29648, Cur Loss: 0.88692641, Cur Avg Loss: 0.75902737, Log Avg loss: 0.85124967, Global Avg Loss: 4.45020963, Time: 0.0208 Steps: 8240, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001863, Sample Num: 29808, Cur Loss: 0.36035192, Cur Avg Loss: 0.75873194, Log Avg loss: 0.70398795, Global Avg Loss: 4.44566875, Time: 0.0208 Steps: 8250, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001873, Sample Num: 29968, Cur Loss: 0.71351373, Cur Avg Loss: 0.75908696, Log Avg loss: 0.82522778, Global Avg Loss: 4.44128565, Time: 0.0208 Steps: 8260, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001883, Sample Num: 30128, Cur Loss: 0.44362369, Cur Avg Loss: 0.75776321, Log Avg loss: 0.50982544, Global Avg Loss: 4.43653177, Time: 0.0208 Steps: 8270, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001893, Sample Num: 30288, Cur Loss: 0.37931395, Cur Avg Loss: 0.75649221, Log Avg loss: 0.51716289, Global Avg Loss: 4.43179824, Time: 0.0208 Steps: 8280, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001903, Sample Num: 30448, Cur Loss: 0.77437270, Cur Avg Loss: 0.75591187, Log Avg loss: 0.64605389, Global Avg Loss: 4.42723160, Time: 0.0208 Steps: 8290, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001913, Sample Num: 30608, Cur Loss: 0.95638382, Cur Avg Loss: 0.75568578, Log Avg loss: 0.71265980, Global Avg Loss: 4.42275621, Time: 0.0208 Steps: 8300, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001923, Sample Num: 30768, Cur Loss: 0.12533766, Cur Avg Loss: 0.75475496, Log Avg loss: 0.57668997, Global Avg Loss: 4.41812797, Time: 0.0208 Steps: 8310, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001933, Sample Num: 30928, Cur Loss: 0.48433283, Cur Avg Loss: 0.75269706, Log Avg loss: 0.35696246, Global Avg Loss: 4.41324676, Time: 0.0208 Steps: 8320, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001943, Sample Num: 31088, Cur Loss: 0.56347686, Cur Avg Loss: 0.75262139, Log Avg loss: 0.73799375, Global Avg Loss: 4.40883469, Time: 0.0208 Steps: 8330, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001953, Sample Num: 31248, Cur Loss: 0.29281572, Cur Avg Loss: 0.75234889, Log Avg loss: 0.69940353, Global Avg Loss: 4.40438693, Time: 0.0208 Steps: 8340, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001963, Sample Num: 31408, Cur Loss: 0.76144820, Cur Avg Loss: 0.75130709, Log Avg loss: 0.54784233, Global Avg Loss: 4.39976832, Time: 0.0207 Steps: 8350, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001973, Sample Num: 31568, Cur Loss: 0.35906827, Cur Avg Loss: 0.74994660, Log Avg loss: 0.48288328, Global Avg Loss: 4.39508305, Time: 0.0208 Steps: 8360, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001983, Sample Num: 31728, Cur Loss: 1.09063005, Cur Avg Loss: 0.74932110, Log Avg loss: 0.62590911, Global Avg Loss: 4.39057985, Time: 0.0208 Steps: 8370, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001993, Sample Num: 31888, Cur Loss: 0.37543771, Cur Avg Loss: 0.74902514, Log Avg loss: 0.69033565, Global Avg Loss: 4.38616429, Time: 0.0208 Steps: 8380, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002003, Sample Num: 32048, Cur Loss: 3.76427794, Cur Avg Loss: 0.75252671, Log Avg loss: 1.45039065, Global Avg Loss: 4.38266515, Time: 0.0208 Steps: 8390, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002013, Sample Num: 32208, Cur Loss: 0.84599096, Cur Avg Loss: 0.75432808, Log Avg loss: 1.11514299, Global Avg Loss: 4.37877525, Time: 0.0208 Steps: 8400, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002023, Sample Num: 32368, Cur Loss: 0.90418446, Cur Avg Loss: 0.75413941, Log Avg loss: 0.71615856, Global Avg Loss: 4.37442017, Time: 0.0208 Steps: 8410, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002033, Sample Num: 32528, Cur Loss: 0.73689282, Cur Avg Loss: 0.75456376, Log Avg loss: 0.84041140, Global Avg Loss: 4.37022301, Time: 0.0208 Steps: 8420, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002043, Sample Num: 32688, Cur Loss: 0.25051859, Cur Avg Loss: 0.75535012, Log Avg loss: 0.91521723, Global Avg Loss: 4.36612455, Time: 0.0208 Steps: 8430, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002053, Sample Num: 32848, Cur Loss: 0.40029284, Cur Avg Loss: 0.75408833, Log Avg loss: 0.49630309, Global Avg Loss: 4.36153945, Time: 0.0245 Steps: 8440, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002063, Sample Num: 33008, Cur Loss: 0.20033649, Cur Avg Loss: 0.75363352, Log Avg loss: 0.66026137, Global Avg Loss: 4.35715924, Time: 0.0208 Steps: 8450, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002073, Sample Num: 33168, Cur Loss: 0.43158722, Cur Avg Loss: 0.75332903, Log Avg loss: 0.69051223, Global Avg Loss: 4.35282514, Time: 0.0208 Steps: 8460, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002083, Sample Num: 33328, Cur Loss: 0.22304061, Cur Avg Loss: 0.75268583, Log Avg loss: 0.61935043, Global Avg Loss: 4.34841726, Time: 0.0209 Steps: 8470, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002093, Sample Num: 33488, Cur Loss: 0.15395051, Cur Avg Loss: 0.75169825, Log Avg loss: 0.54598548, Global Avg Loss: 4.34393326, Time: 0.0208 Steps: 8480, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002103, Sample Num: 33648, Cur Loss: 0.13305594, Cur Avg Loss: 0.75106414, Log Avg loss: 0.61834472, Global Avg Loss: 4.33954505, Time: 0.0208 Steps: 8490, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002113, Sample Num: 33808, Cur Loss: 0.50461364, Cur Avg Loss: 0.74955631, Log Avg loss: 0.43246025, Global Avg Loss: 4.33494848, Time: 0.0209 Steps: 8500, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002123, Sample Num: 33968, Cur Loss: 1.32601106, Cur Avg Loss: 0.75019883, Log Avg loss: 0.88596290, Global Avg Loss: 4.33089562, Time: 0.0208 Steps: 8510, Updated lr: 0.000093 ***** Running evaluation checkpoint-8516 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-8516 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.479027, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.575294, "eval_total_loss": 404.431372, "eval_mae": 0.594102, "eval_mse": 0.575465, "eval_r2": 0.634197, "eval_sp_statistic": 0.77546, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.817659, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.311979, "test_total_loss": 658.613453, "test_mae": 1.043134, "test_mse": 1.312249, "test_r2": 0.153063, "test_sp_statistic": 0.774446, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.845793, "test_ps_pvalue": 0.0, "lr": 9.287245139876719e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 4.328263714225826, "train_cur_epoch_loss": 1596.2441611103714, "train_cur_epoch_avg_loss": 0.7497624054064684, "train_cur_epoch_time": 44.479026556015015, "train_cur_epoch_avg_time": 0.020891980533590895, "epoch": 4, "step": 8516} ################################################## Training, Epoch: 0005, Batch: 000004, Sample Num: 64, Cur Loss: 0.46938875, Cur Avg Loss: 0.73271685, Log Avg loss: 0.65029202, Global Avg Loss: 4.32657566, Time: 0.0246 Steps: 8520, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000014, Sample Num: 224, Cur Loss: 0.10221770, Cur Avg Loss: 0.51913166, Log Avg loss: 0.43369759, Global Avg Loss: 4.32201191, Time: 0.0208 Steps: 8530, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000024, Sample Num: 384, Cur Loss: 1.34348559, Cur Avg Loss: 0.63816397, Log Avg loss: 0.80480919, Global Avg Loss: 4.31789341, Time: 0.0207 Steps: 8540, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000034, Sample Num: 544, Cur Loss: 0.64349127, Cur Avg Loss: 0.62477686, Log Avg loss: 0.59264781, Global Avg Loss: 4.31353640, Time: 0.0208 Steps: 8550, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000044, Sample Num: 704, Cur Loss: 0.33546862, Cur Avg Loss: 0.59757585, Log Avg loss: 0.50509239, Global Avg Loss: 4.30908728, Time: 0.0208 Steps: 8560, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000054, Sample Num: 864, Cur Loss: 0.90584254, Cur Avg Loss: 0.60243611, Log Avg loss: 0.62382126, Global Avg Loss: 4.30478709, Time: 0.0208 Steps: 8570, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000064, Sample Num: 1024, Cur Loss: 0.92706174, Cur Avg Loss: 0.61042337, Log Avg loss: 0.65355459, Global Avg Loss: 4.30053157, Time: 0.0208 Steps: 8580, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000074, Sample Num: 1184, Cur Loss: 0.67706710, Cur Avg Loss: 0.62895556, Log Avg loss: 0.74756161, Global Avg Loss: 4.29639540, Time: 0.0208 Steps: 8590, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000084, Sample Num: 1344, Cur Loss: 0.31885478, Cur Avg Loss: 0.62893068, Log Avg loss: 0.62874657, Global Avg Loss: 4.29213069, Time: 0.0208 Steps: 8600, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000094, Sample Num: 1504, Cur Loss: 0.60522908, Cur Avg Loss: 0.63659815, Log Avg loss: 0.70100489, Global Avg Loss: 4.28795982, Time: 0.0209 Steps: 8610, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000104, Sample Num: 1664, Cur Loss: 0.33746874, Cur Avg Loss: 0.61340994, Log Avg loss: 0.39544074, Global Avg Loss: 4.28344413, Time: 0.0209 Steps: 8620, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000114, Sample Num: 1824, Cur Loss: 0.22916415, Cur Avg Loss: 0.60623772, Log Avg loss: 0.53164667, Global Avg Loss: 4.27909674, Time: 0.0209 Steps: 8630, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000124, Sample Num: 1984, Cur Loss: 0.64680159, Cur Avg Loss: 0.62523418, Log Avg loss: 0.84179377, Global Avg Loss: 4.27511838, Time: 0.0209 Steps: 8640, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000134, Sample Num: 2144, Cur Loss: 0.86364591, Cur Avg Loss: 0.62477170, Log Avg loss: 0.61903700, Global Avg Loss: 4.27089170, Time: 0.0209 Steps: 8650, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000144, Sample Num: 2304, Cur Loss: 0.26991075, Cur Avg Loss: 0.62442699, Log Avg loss: 0.61980786, Global Avg Loss: 4.26667567, Time: 0.0209 Steps: 8660, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000154, Sample Num: 2464, Cur Loss: 0.53616887, Cur Avg Loss: 0.62454826, Log Avg loss: 0.62629455, Global Avg Loss: 4.26247684, Time: 0.0209 Steps: 8670, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000164, Sample Num: 2624, Cur Loss: 0.22024785, Cur Avg Loss: 0.62099981, Log Avg loss: 0.56635368, Global Avg Loss: 4.25821864, Time: 0.0209 Steps: 8680, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000174, Sample Num: 2784, Cur Loss: 0.36368707, Cur Avg Loss: 0.61809686, Log Avg loss: 0.57048839, Global Avg Loss: 4.25397499, Time: 0.0209 Steps: 8690, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000184, Sample Num: 2944, Cur Loss: 0.42835283, Cur Avg Loss: 0.62445336, Log Avg loss: 0.73505656, Global Avg Loss: 4.24993025, Time: 0.0209 Steps: 8700, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000194, Sample Num: 3104, Cur Loss: 0.46300292, Cur Avg Loss: 0.63056992, Log Avg loss: 0.74311457, Global Avg Loss: 4.24590406, Time: 0.0209 Steps: 8710, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000204, Sample Num: 3264, Cur Loss: 0.34676111, Cur Avg Loss: 0.62664707, Log Avg loss: 0.55054376, Global Avg Loss: 4.24166626, Time: 0.0209 Steps: 8720, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000214, Sample Num: 3424, Cur Loss: 1.15134811, Cur Avg Loss: 0.64105694, Log Avg loss: 0.93501830, Global Avg Loss: 4.23787858, Time: 0.0209 Steps: 8730, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000224, Sample Num: 3584, Cur Loss: 0.69497240, Cur Avg Loss: 0.65198035, Log Avg loss: 0.88574126, Global Avg Loss: 4.23404318, Time: 0.0209 Steps: 8740, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000234, Sample Num: 3744, Cur Loss: 0.49618918, Cur Avg Loss: 0.64980690, Log Avg loss: 0.60112162, Global Avg Loss: 4.22989127, Time: 0.0209 Steps: 8750, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000244, Sample Num: 3904, Cur Loss: 0.55772108, Cur Avg Loss: 0.64766000, Log Avg loss: 0.59742265, Global Avg Loss: 4.22574462, Time: 0.0209 Steps: 8760, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000254, Sample Num: 4064, Cur Loss: 0.54794037, Cur Avg Loss: 0.64788011, Log Avg loss: 0.65325091, Global Avg Loss: 4.22167108, Time: 0.0209 Steps: 8770, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000264, Sample Num: 4224, Cur Loss: 0.17066301, Cur Avg Loss: 0.64880290, Log Avg loss: 0.67224175, Global Avg Loss: 4.21762845, Time: 0.0208 Steps: 8780, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000274, Sample Num: 4384, Cur Loss: 0.18453431, Cur Avg Loss: 0.64159325, Log Avg loss: 0.45125838, Global Avg Loss: 4.21334361, Time: 0.0209 Steps: 8790, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000284, Sample Num: 4544, Cur Loss: 0.72425699, Cur Avg Loss: 0.64503513, Log Avg loss: 0.73934256, Global Avg Loss: 4.20939588, Time: 0.0208 Steps: 8800, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000294, Sample Num: 4704, Cur Loss: 0.11932529, Cur Avg Loss: 0.64202375, Log Avg loss: 0.55650067, Global Avg Loss: 4.20524958, Time: 0.0208 Steps: 8810, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000304, Sample Num: 4864, Cur Loss: 0.27430335, Cur Avg Loss: 0.63936180, Log Avg loss: 0.56110055, Global Avg Loss: 4.20111789, Time: 0.0208 Steps: 8820, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000314, Sample Num: 5024, Cur Loss: 0.46484029, Cur Avg Loss: 0.64677066, Log Avg loss: 0.87199984, Global Avg Loss: 4.19734765, Time: 0.0208 Steps: 8830, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000324, Sample Num: 5184, Cur Loss: 0.52442622, Cur Avg Loss: 0.64386288, Log Avg loss: 0.55255858, Global Avg Loss: 4.19322459, Time: 0.0208 Steps: 8840, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000334, Sample Num: 5344, Cur Loss: 0.42590910, Cur Avg Loss: 0.64437428, Log Avg loss: 0.66094375, Global Avg Loss: 4.18923331, Time: 0.0209 Steps: 8850, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000344, Sample Num: 5504, Cur Loss: 0.65413195, Cur Avg Loss: 0.64070761, Log Avg loss: 0.51824062, Global Avg Loss: 4.18508998, Time: 0.0209 Steps: 8860, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000354, Sample Num: 5664, Cur Loss: 0.70091379, Cur Avg Loss: 0.63703659, Log Avg loss: 0.51075379, Global Avg Loss: 4.18094755, Time: 0.0208 Steps: 8870, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000364, Sample Num: 5824, Cur Loss: 0.70809960, Cur Avg Loss: 0.64413292, Log Avg loss: 0.89534299, Global Avg Loss: 4.17724754, Time: 0.0208 Steps: 8880, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000374, Sample Num: 5984, Cur Loss: 1.51462412, Cur Avg Loss: 0.65514020, Log Avg loss: 1.05580505, Global Avg Loss: 4.17373636, Time: 0.0209 Steps: 8890, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000384, Sample Num: 6144, Cur Loss: 0.43513006, Cur Avg Loss: 0.66227979, Log Avg loss: 0.92930065, Global Avg Loss: 4.17009092, Time: 0.0209 Steps: 8900, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000394, Sample Num: 6304, Cur Loss: 0.83332849, Cur Avg Loss: 0.66817939, Log Avg loss: 0.89472406, Global Avg Loss: 4.16641487, Time: 0.0208 Steps: 8910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000404, Sample Num: 6464, Cur Loss: 0.53582442, Cur Avg Loss: 0.66529560, Log Avg loss: 0.55167425, Global Avg Loss: 4.16236247, Time: 0.0207 Steps: 8920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000414, Sample Num: 6624, Cur Loss: 0.52805161, Cur Avg Loss: 0.66131804, Log Avg loss: 0.50062426, Global Avg Loss: 4.15826198, Time: 0.0208 Steps: 8930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000424, Sample Num: 6784, Cur Loss: 0.96406817, Cur Avg Loss: 0.65887238, Log Avg loss: 0.55762241, Global Avg Loss: 4.15423442, Time: 0.0209 Steps: 8940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000434, Sample Num: 6944, Cur Loss: 0.23950973, Cur Avg Loss: 0.65769408, Log Avg loss: 0.60773410, Global Avg Loss: 4.15027185, Time: 0.0208 Steps: 8950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000444, Sample Num: 7104, Cur Loss: 0.45370749, Cur Avg Loss: 0.65226577, Log Avg loss: 0.41667700, Global Avg Loss: 4.14610489, Time: 0.0208 Steps: 8960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000454, Sample Num: 7264, Cur Loss: 0.17759115, Cur Avg Loss: 0.65994460, Log Avg loss: 1.00088480, Global Avg Loss: 4.14259851, Time: 0.0208 Steps: 8970, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000464, Sample Num: 7424, Cur Loss: 0.48345804, Cur Avg Loss: 0.66234803, Log Avg loss: 0.77146344, Global Avg Loss: 4.13884446, Time: 0.0208 Steps: 8980, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000474, Sample Num: 7584, Cur Loss: 0.76877356, Cur Avg Loss: 0.66286304, Log Avg loss: 0.68675989, Global Avg Loss: 4.13500455, Time: 0.0208 Steps: 8990, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000484, Sample Num: 7744, Cur Loss: 0.74328709, Cur Avg Loss: 0.66025939, Log Avg loss: 0.53684639, Global Avg Loss: 4.13100659, Time: 0.0208 Steps: 9000, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000494, Sample Num: 7904, Cur Loss: 1.18983960, Cur Avg Loss: 0.67272263, Log Avg loss: 1.27594329, Global Avg Loss: 4.12783782, Time: 0.0208 Steps: 9010, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000504, Sample Num: 8064, Cur Loss: 1.09716344, Cur Avg Loss: 0.68011268, Log Avg loss: 1.04518126, Global Avg Loss: 4.12442024, Time: 0.0208 Steps: 9020, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000514, Sample Num: 8224, Cur Loss: 0.47219497, Cur Avg Loss: 0.67742840, Log Avg loss: 0.54214056, Global Avg Loss: 4.12045315, Time: 0.0246 Steps: 9030, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000524, Sample Num: 8384, Cur Loss: 0.60754800, Cur Avg Loss: 0.68005851, Log Avg loss: 0.81524622, Global Avg Loss: 4.11679695, Time: 0.0207 Steps: 9040, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000534, Sample Num: 8544, Cur Loss: 0.54997027, Cur Avg Loss: 0.67940960, Log Avg loss: 0.64540666, Global Avg Loss: 4.11296116, Time: 0.0208 Steps: 9050, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000544, Sample Num: 8704, Cur Loss: 0.70848423, Cur Avg Loss: 0.67714837, Log Avg loss: 0.55639857, Global Avg Loss: 4.10903560, Time: 0.0208 Steps: 9060, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000554, Sample Num: 8864, Cur Loss: 0.41416207, Cur Avg Loss: 0.67729529, Log Avg loss: 0.68528766, Global Avg Loss: 4.10526079, Time: 0.0208 Steps: 9070, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000564, Sample Num: 9024, Cur Loss: 0.37878317, Cur Avg Loss: 0.67408642, Log Avg loss: 0.49631507, Global Avg Loss: 4.10128618, Time: 0.0208 Steps: 9080, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000574, Sample Num: 9184, Cur Loss: 0.84549415, Cur Avg Loss: 0.67335687, Log Avg loss: 0.63221013, Global Avg Loss: 4.09746982, Time: 0.0207 Steps: 9090, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000584, Sample Num: 9344, Cur Loss: 0.29905224, Cur Avg Loss: 0.67256542, Log Avg loss: 0.62713662, Global Avg Loss: 4.09365626, Time: 0.0208 Steps: 9100, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000594, Sample Num: 9504, Cur Loss: 0.34129590, Cur Avg Loss: 0.66995914, Log Avg loss: 0.51775241, Global Avg Loss: 4.08973101, Time: 0.0207 Steps: 9110, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000604, Sample Num: 9664, Cur Loss: 1.09760475, Cur Avg Loss: 0.67117045, Log Avg loss: 0.74312211, Global Avg Loss: 4.08606148, Time: 0.0208 Steps: 9120, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000614, Sample Num: 9824, Cur Loss: 0.85694295, Cur Avg Loss: 0.68058874, Log Avg loss: 1.24945359, Global Avg Loss: 4.08295458, Time: 0.0207 Steps: 9130, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000624, Sample Num: 9984, Cur Loss: 1.26958942, Cur Avg Loss: 0.68557277, Log Avg loss: 0.99159227, Global Avg Loss: 4.07957234, Time: 0.0208 Steps: 9140, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000634, Sample Num: 10144, Cur Loss: 0.95854461, Cur Avg Loss: 0.68384320, Log Avg loss: 0.57591810, Global Avg Loss: 4.07574321, Time: 0.0208 Steps: 9150, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000644, Sample Num: 10304, Cur Loss: 0.25572351, Cur Avg Loss: 0.67860573, Log Avg loss: 0.34654962, Global Avg Loss: 4.07167204, Time: 0.0208 Steps: 9160, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000654, Sample Num: 10464, Cur Loss: 0.84277415, Cur Avg Loss: 0.67904779, Log Avg loss: 0.70751660, Global Avg Loss: 4.06800339, Time: 0.0211 Steps: 9170, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000664, Sample Num: 10624, Cur Loss: 0.41394180, Cur Avg Loss: 0.67711085, Log Avg loss: 0.55043509, Global Avg Loss: 4.06417161, Time: 0.0210 Steps: 9180, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000674, Sample Num: 10784, Cur Loss: 0.56671768, Cur Avg Loss: 0.67480817, Log Avg loss: 0.52190994, Global Avg Loss: 4.06031714, Time: 0.0211 Steps: 9190, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000684, Sample Num: 10944, Cur Loss: 0.94660604, Cur Avg Loss: 0.67588587, Log Avg loss: 0.74852334, Global Avg Loss: 4.05671736, Time: 0.0211 Steps: 9200, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000694, Sample Num: 11104, Cur Loss: 0.58337992, Cur Avg Loss: 0.67556578, Log Avg loss: 0.65367169, Global Avg Loss: 4.05302242, Time: 0.0211 Steps: 9210, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000704, Sample Num: 11264, Cur Loss: 0.45709974, Cur Avg Loss: 0.67658601, Log Avg loss: 0.74738995, Global Avg Loss: 4.04943713, Time: 0.0210 Steps: 9220, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000714, Sample Num: 11424, Cur Loss: 0.75284636, Cur Avg Loss: 0.67519472, Log Avg loss: 0.57724746, Global Avg Loss: 4.04567528, Time: 0.0211 Steps: 9230, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000724, Sample Num: 11584, Cur Loss: 0.22163963, Cur Avg Loss: 0.67541963, Log Avg loss: 0.69147865, Global Avg Loss: 4.04204520, Time: 0.0210 Steps: 9240, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000734, Sample Num: 11744, Cur Loss: 0.72694486, Cur Avg Loss: 0.67790718, Log Avg loss: 0.85800571, Global Avg Loss: 4.03860299, Time: 0.0211 Steps: 9250, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000744, Sample Num: 11904, Cur Loss: 0.34812585, Cur Avg Loss: 0.67936933, Log Avg loss: 0.78669070, Global Avg Loss: 4.03509121, Time: 0.0210 Steps: 9260, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000754, Sample Num: 12064, Cur Loss: 0.28433520, Cur Avg Loss: 0.67652783, Log Avg loss: 0.46512029, Global Avg Loss: 4.03124010, Time: 0.0210 Steps: 9270, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000764, Sample Num: 12224, Cur Loss: 0.71279514, Cur Avg Loss: 0.67464146, Log Avg loss: 0.53240900, Global Avg Loss: 4.02746981, Time: 0.0211 Steps: 9280, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000774, Sample Num: 12384, Cur Loss: 0.88873875, Cur Avg Loss: 0.67447958, Log Avg loss: 0.66211252, Global Avg Loss: 4.02384725, Time: 0.0211 Steps: 9290, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000784, Sample Num: 12544, Cur Loss: 0.41775915, Cur Avg Loss: 0.67309842, Log Avg loss: 0.56619639, Global Avg Loss: 4.02012935, Time: 0.0210 Steps: 9300, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000794, Sample Num: 12704, Cur Loss: 0.61105722, Cur Avg Loss: 0.67117358, Log Avg loss: 0.52026626, Global Avg Loss: 4.01637010, Time: 0.0210 Steps: 9310, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000804, Sample Num: 12864, Cur Loss: 0.63207430, Cur Avg Loss: 0.66978236, Log Avg loss: 0.55931947, Global Avg Loss: 4.01266082, Time: 0.0210 Steps: 9320, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000814, Sample Num: 13024, Cur Loss: 0.24907900, Cur Avg Loss: 0.66663711, Log Avg loss: 0.41375860, Global Avg Loss: 4.00880347, Time: 0.0210 Steps: 9330, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000824, Sample Num: 13184, Cur Loss: 0.68173158, Cur Avg Loss: 0.66620569, Log Avg loss: 0.63108819, Global Avg Loss: 4.00518707, Time: 0.0210 Steps: 9340, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000834, Sample Num: 13344, Cur Loss: 1.73093438, Cur Avg Loss: 0.66772996, Log Avg loss: 0.79333041, Global Avg Loss: 4.00175193, Time: 0.0210 Steps: 9350, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000844, Sample Num: 13504, Cur Loss: 0.77551413, Cur Avg Loss: 0.66817460, Log Avg loss: 0.70525682, Global Avg Loss: 3.99823004, Time: 0.0211 Steps: 9360, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000854, Sample Num: 13664, Cur Loss: 0.62446028, Cur Avg Loss: 0.66911784, Log Avg loss: 0.74872797, Global Avg Loss: 3.99476205, Time: 0.0211 Steps: 9370, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000864, Sample Num: 13824, Cur Loss: 0.48113155, Cur Avg Loss: 0.66830160, Log Avg loss: 0.59859460, Global Avg Loss: 3.99114140, Time: 0.0210 Steps: 9380, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000874, Sample Num: 13984, Cur Loss: 0.38750511, Cur Avg Loss: 0.66701609, Log Avg loss: 0.55594744, Global Avg Loss: 3.98748305, Time: 0.0210 Steps: 9390, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000884, Sample Num: 14144, Cur Loss: 0.35134274, Cur Avg Loss: 0.66687138, Log Avg loss: 0.65422409, Global Avg Loss: 3.98393703, Time: 0.0210 Steps: 9400, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000894, Sample Num: 14304, Cur Loss: 0.93886840, Cur Avg Loss: 0.66594452, Log Avg loss: 0.58401028, Global Avg Loss: 3.98032393, Time: 0.0210 Steps: 9410, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000904, Sample Num: 14464, Cur Loss: 0.74203920, Cur Avg Loss: 0.66419032, Log Avg loss: 0.50736498, Global Avg Loss: 3.97663714, Time: 0.0211 Steps: 9420, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000914, Sample Num: 14624, Cur Loss: 0.58351529, Cur Avg Loss: 0.66379429, Log Avg loss: 0.62799323, Global Avg Loss: 3.97308608, Time: 0.0210 Steps: 9430, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000924, Sample Num: 14784, Cur Loss: 0.58418489, Cur Avg Loss: 0.66272911, Log Avg loss: 0.56537150, Global Avg Loss: 3.96947622, Time: 0.0210 Steps: 9440, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000934, Sample Num: 14944, Cur Loss: 0.74301237, Cur Avg Loss: 0.66123332, Log Avg loss: 0.52302230, Global Avg Loss: 3.96582918, Time: 0.0210 Steps: 9450, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000944, Sample Num: 15104, Cur Loss: 0.41821566, Cur Avg Loss: 0.66241496, Log Avg loss: 0.77278036, Global Avg Loss: 3.96245386, Time: 0.0212 Steps: 9460, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000954, Sample Num: 15264, Cur Loss: 0.43905762, Cur Avg Loss: 0.66225227, Log Avg loss: 0.64689353, Global Avg Loss: 3.95895274, Time: 0.0210 Steps: 9470, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000964, Sample Num: 15424, Cur Loss: 0.67636371, Cur Avg Loss: 0.66386449, Log Avg loss: 0.81767028, Global Avg Loss: 3.95563915, Time: 0.0210 Steps: 9480, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000974, Sample Num: 15584, Cur Loss: 0.40387785, Cur Avg Loss: 0.66523484, Log Avg loss: 0.79733689, Global Avg Loss: 3.95231112, Time: 0.0210 Steps: 9490, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000984, Sample Num: 15744, Cur Loss: 0.72125250, Cur Avg Loss: 0.66597858, Log Avg loss: 0.73841850, Global Avg Loss: 3.94892807, Time: 0.0210 Steps: 9500, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000994, Sample Num: 15904, Cur Loss: 0.59739798, Cur Avg Loss: 0.66416628, Log Avg loss: 0.48583613, Global Avg Loss: 3.94528655, Time: 0.0211 Steps: 9510, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001004, Sample Num: 16064, Cur Loss: 0.77539051, Cur Avg Loss: 0.66492754, Log Avg loss: 0.74059662, Global Avg Loss: 3.94192028, Time: 0.0210 Steps: 9520, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001014, Sample Num: 16224, Cur Loss: 1.03530741, Cur Avg Loss: 0.66703938, Log Avg loss: 0.87906818, Global Avg Loss: 3.93870637, Time: 0.0210 Steps: 9530, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001024, Sample Num: 16384, Cur Loss: 1.30488050, Cur Avg Loss: 0.66908277, Log Avg loss: 0.87628262, Global Avg Loss: 3.93549628, Time: 0.0256 Steps: 9540, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001034, Sample Num: 16544, Cur Loss: 0.38065177, Cur Avg Loss: 0.66652806, Log Avg loss: 0.40492571, Global Avg Loss: 3.93179935, Time: 0.0210 Steps: 9550, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001044, Sample Num: 16704, Cur Loss: 0.95266944, Cur Avg Loss: 0.66617884, Log Avg loss: 0.63006934, Global Avg Loss: 3.92834566, Time: 0.0210 Steps: 9560, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001054, Sample Num: 16864, Cur Loss: 1.62497282, Cur Avg Loss: 0.66694343, Log Avg loss: 0.74676729, Global Avg Loss: 3.92502113, Time: 0.0211 Steps: 9570, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001064, Sample Num: 17024, Cur Loss: 0.52600437, Cur Avg Loss: 0.66656696, Log Avg loss: 0.62688640, Global Avg Loss: 3.92157840, Time: 0.0211 Steps: 9580, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001074, Sample Num: 17184, Cur Loss: 1.30238926, Cur Avg Loss: 0.66731608, Log Avg loss: 0.74702281, Global Avg Loss: 3.91826812, Time: 0.0211 Steps: 9590, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001084, Sample Num: 17344, Cur Loss: 0.56098360, Cur Avg Loss: 0.66527676, Log Avg loss: 0.44625389, Global Avg Loss: 3.91465144, Time: 0.0210 Steps: 9600, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001094, Sample Num: 17504, Cur Loss: 1.19134402, Cur Avg Loss: 0.66538009, Log Avg loss: 0.67658152, Global Avg Loss: 3.91128196, Time: 0.0210 Steps: 9610, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001104, Sample Num: 17664, Cur Loss: 0.86470175, Cur Avg Loss: 0.66501980, Log Avg loss: 0.62560324, Global Avg Loss: 3.90786649, Time: 0.0210 Steps: 9620, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001114, Sample Num: 17824, Cur Loss: 0.70742601, Cur Avg Loss: 0.66546847, Log Avg loss: 0.71500188, Global Avg Loss: 3.90455095, Time: 0.0211 Steps: 9630, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001124, Sample Num: 17984, Cur Loss: 0.42400086, Cur Avg Loss: 0.66559377, Log Avg loss: 0.67955212, Global Avg Loss: 3.90120552, Time: 0.0210 Steps: 9640, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001134, Sample Num: 18144, Cur Loss: 0.74858606, Cur Avg Loss: 0.66458067, Log Avg loss: 0.55070851, Global Avg Loss: 3.89773350, Time: 0.0210 Steps: 9650, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001144, Sample Num: 18304, Cur Loss: 0.22369626, Cur Avg Loss: 0.66424497, Log Avg loss: 0.62617700, Global Avg Loss: 3.89434680, Time: 0.0210 Steps: 9660, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001154, Sample Num: 18464, Cur Loss: 1.02813423, Cur Avg Loss: 0.66534523, Log Avg loss: 0.79121450, Global Avg Loss: 3.89113776, Time: 0.0210 Steps: 9670, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001164, Sample Num: 18624, Cur Loss: 1.16146219, Cur Avg Loss: 0.66681161, Log Avg loss: 0.83603136, Global Avg Loss: 3.88798166, Time: 0.0211 Steps: 9680, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001174, Sample Num: 18784, Cur Loss: 0.49677306, Cur Avg Loss: 0.66738456, Log Avg loss: 0.73407644, Global Avg Loss: 3.88472686, Time: 0.0210 Steps: 9690, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001184, Sample Num: 18944, Cur Loss: 0.54520762, Cur Avg Loss: 0.66674077, Log Avg loss: 0.59115984, Global Avg Loss: 3.88133143, Time: 0.0210 Steps: 9700, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001194, Sample Num: 19104, Cur Loss: 0.55227995, Cur Avg Loss: 0.66624529, Log Avg loss: 0.60758044, Global Avg Loss: 3.87795990, Time: 0.0210 Steps: 9710, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001204, Sample Num: 19264, Cur Loss: 0.81698799, Cur Avg Loss: 0.66515895, Log Avg loss: 0.53545017, Global Avg Loss: 3.87452111, Time: 0.0211 Steps: 9720, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001214, Sample Num: 19424, Cur Loss: 0.93688095, Cur Avg Loss: 0.66731446, Log Avg loss: 0.92683746, Global Avg Loss: 3.87149163, Time: 0.0211 Steps: 9730, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001224, Sample Num: 19584, Cur Loss: 0.19251072, Cur Avg Loss: 0.66621807, Log Avg loss: 0.53311665, Global Avg Loss: 3.86806414, Time: 0.0211 Steps: 9740, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001234, Sample Num: 19744, Cur Loss: 0.86643809, Cur Avg Loss: 0.66621356, Log Avg loss: 0.66566076, Global Avg Loss: 3.86477962, Time: 0.0211 Steps: 9750, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001244, Sample Num: 19904, Cur Loss: 0.90099567, Cur Avg Loss: 0.66797640, Log Avg loss: 0.88551084, Global Avg Loss: 3.86172709, Time: 0.0210 Steps: 9760, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001254, Sample Num: 20064, Cur Loss: 0.38893348, Cur Avg Loss: 0.66755444, Log Avg loss: 0.61506322, Global Avg Loss: 3.85840400, Time: 0.0210 Steps: 9770, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001264, Sample Num: 20224, Cur Loss: 0.31738082, Cur Avg Loss: 0.66646394, Log Avg loss: 0.52971464, Global Avg Loss: 3.85500043, Time: 0.0210 Steps: 9780, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001274, Sample Num: 20384, Cur Loss: 0.42920297, Cur Avg Loss: 0.66545249, Log Avg loss: 0.53760522, Global Avg Loss: 3.85161188, Time: 0.0210 Steps: 9790, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001284, Sample Num: 20544, Cur Loss: 0.45027938, Cur Avg Loss: 0.66365987, Log Avg loss: 0.43528129, Global Avg Loss: 3.84812582, Time: 0.0243 Steps: 9800, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001294, Sample Num: 20704, Cur Loss: 0.82060254, Cur Avg Loss: 0.66292548, Log Avg loss: 0.56862893, Global Avg Loss: 3.84478281, Time: 0.0211 Steps: 9810, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001304, Sample Num: 20864, Cur Loss: 0.49899763, Cur Avg Loss: 0.66249002, Log Avg loss: 0.60614133, Global Avg Loss: 3.84148480, Time: 0.0210 Steps: 9820, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001314, Sample Num: 21024, Cur Loss: 0.36085564, Cur Avg Loss: 0.66204533, Log Avg loss: 0.60405776, Global Avg Loss: 3.83819139, Time: 0.0210 Steps: 9830, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001324, Sample Num: 21184, Cur Loss: 0.19443138, Cur Avg Loss: 0.66078330, Log Avg loss: 0.49495285, Global Avg Loss: 3.83479379, Time: 0.0210 Steps: 9840, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001334, Sample Num: 21344, Cur Loss: 0.14819446, Cur Avg Loss: 0.65968782, Log Avg loss: 0.51464585, Global Avg Loss: 3.83142308, Time: 0.0210 Steps: 9850, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001344, Sample Num: 21504, Cur Loss: 1.78757763, Cur Avg Loss: 0.66015779, Log Avg loss: 0.72285269, Global Avg Loss: 3.82827037, Time: 0.0211 Steps: 9860, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001354, Sample Num: 21664, Cur Loss: 0.64421111, Cur Avg Loss: 0.66016211, Log Avg loss: 0.66074238, Global Avg Loss: 3.82506112, Time: 0.0211 Steps: 9870, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001364, Sample Num: 21824, Cur Loss: 0.37130564, Cur Avg Loss: 0.65991701, Log Avg loss: 0.62673033, Global Avg Loss: 3.82182395, Time: 0.0211 Steps: 9880, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001374, Sample Num: 21984, Cur Loss: 0.58539671, Cur Avg Loss: 0.66096773, Log Avg loss: 0.80428661, Global Avg Loss: 3.81877285, Time: 0.0211 Steps: 9890, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001384, Sample Num: 22144, Cur Loss: 0.50508785, Cur Avg Loss: 0.65982667, Log Avg loss: 0.50304420, Global Avg Loss: 3.81542363, Time: 0.0210 Steps: 9900, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001394, Sample Num: 22304, Cur Loss: 0.31798977, Cur Avg Loss: 0.65970873, Log Avg loss: 0.64338676, Global Avg Loss: 3.81222278, Time: 0.0210 Steps: 9910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001404, Sample Num: 22464, Cur Loss: 0.82765329, Cur Avg Loss: 0.65979464, Log Avg loss: 0.67176969, Global Avg Loss: 3.80905700, Time: 0.0210 Steps: 9920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001414, Sample Num: 22624, Cur Loss: 0.34399283, Cur Avg Loss: 0.65925547, Log Avg loss: 0.58355618, Global Avg Loss: 3.80580876, Time: 0.0210 Steps: 9930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001424, Sample Num: 22784, Cur Loss: 0.49130511, Cur Avg Loss: 0.65969824, Log Avg loss: 0.72230610, Global Avg Loss: 3.80270665, Time: 0.0210 Steps: 9940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001434, Sample Num: 22944, Cur Loss: 0.37651253, Cur Avg Loss: 0.65819522, Log Avg loss: 0.44416443, Global Avg Loss: 3.79933123, Time: 0.0210 Steps: 9950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001444, Sample Num: 23104, Cur Loss: 0.49439645, Cur Avg Loss: 0.65802806, Log Avg loss: 0.63405748, Global Avg Loss: 3.79615324, Time: 0.0210 Steps: 9960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001454, Sample Num: 23264, Cur Loss: 0.23807341, Cur Avg Loss: 0.65822912, Log Avg loss: 0.68726221, Global Avg Loss: 3.79303500, Time: 0.0210 Steps: 9970, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001464, Sample Num: 23424, Cur Loss: 0.47550675, Cur Avg Loss: 0.65773441, Log Avg loss: 0.58580330, Global Avg Loss: 3.78982134, Time: 0.0210 Steps: 9980, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001474, Sample Num: 23584, Cur Loss: 0.41171837, Cur Avg Loss: 0.65695538, Log Avg loss: 0.54290581, Global Avg Loss: 3.78657117, Time: 0.0210 Steps: 9990, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001484, Sample Num: 23744, Cur Loss: 1.04894185, Cur Avg Loss: 0.65733279, Log Avg loss: 0.71296372, Global Avg Loss: 3.78349757, Time: 0.0210 Steps: 10000, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001494, Sample Num: 23904, Cur Loss: 0.72672457, Cur Avg Loss: 0.65695371, Log Avg loss: 0.60069826, Global Avg Loss: 3.78031795, Time: 0.0210 Steps: 10010, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001504, Sample Num: 24064, Cur Loss: 0.25997743, Cur Avg Loss: 0.65660468, Log Avg loss: 0.60445981, Global Avg Loss: 3.77714843, Time: 0.0211 Steps: 10020, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001514, Sample Num: 24224, Cur Loss: 1.76530814, Cur Avg Loss: 0.65684807, Log Avg loss: 0.69345370, Global Avg Loss: 3.77407396, Time: 0.0210 Steps: 10030, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001524, Sample Num: 24384, Cur Loss: 0.64565307, Cur Avg Loss: 0.65633136, Log Avg loss: 0.57810146, Global Avg Loss: 3.77089072, Time: 0.0210 Steps: 10040, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001534, Sample Num: 24544, Cur Loss: 0.66457194, Cur Avg Loss: 0.65658744, Log Avg loss: 0.69561348, Global Avg Loss: 3.76783074, Time: 0.0210 Steps: 10050, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001544, Sample Num: 24704, Cur Loss: 0.32142478, Cur Avg Loss: 0.65596556, Log Avg loss: 0.56056926, Global Avg Loss: 3.76464261, Time: 0.0211 Steps: 10060, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001554, Sample Num: 24864, Cur Loss: 1.03470182, Cur Avg Loss: 0.65568870, Log Avg loss: 0.61294123, Global Avg Loss: 3.76151281, Time: 0.0211 Steps: 10070, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001564, Sample Num: 25024, Cur Loss: 0.55515742, Cur Avg Loss: 0.65672563, Log Avg loss: 0.81786547, Global Avg Loss: 3.75859253, Time: 0.0211 Steps: 10080, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001574, Sample Num: 25184, Cur Loss: 0.47643286, Cur Avg Loss: 0.65494452, Log Avg loss: 0.37637875, Global Avg Loss: 3.75524048, Time: 0.0210 Steps: 10090, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001584, Sample Num: 25344, Cur Loss: 0.23080043, Cur Avg Loss: 0.65326308, Log Avg loss: 0.38860464, Global Avg Loss: 3.75190718, Time: 0.0210 Steps: 10100, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001594, Sample Num: 25504, Cur Loss: 0.36281091, Cur Avg Loss: 0.65219939, Log Avg loss: 0.48371009, Global Avg Loss: 3.74867454, Time: 0.0210 Steps: 10110, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001604, Sample Num: 25664, Cur Loss: 0.69631153, Cur Avg Loss: 0.65162957, Log Avg loss: 0.56079988, Global Avg Loss: 3.74552447, Time: 0.0211 Steps: 10120, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001614, Sample Num: 25824, Cur Loss: 0.74095237, Cur Avg Loss: 0.65177041, Log Avg loss: 0.67436198, Global Avg Loss: 3.74249272, Time: 0.0210 Steps: 10130, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001624, Sample Num: 25984, Cur Loss: 0.32740095, Cur Avg Loss: 0.65101909, Log Avg loss: 0.52975546, Global Avg Loss: 3.73932434, Time: 0.0210 Steps: 10140, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001634, Sample Num: 26144, Cur Loss: 0.50227761, Cur Avg Loss: 0.65079395, Log Avg loss: 0.61423110, Global Avg Loss: 3.73624543, Time: 0.0210 Steps: 10150, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001644, Sample Num: 26304, Cur Loss: 0.51327270, Cur Avg Loss: 0.65002912, Log Avg loss: 0.52505627, Global Avg Loss: 3.73308481, Time: 0.0210 Steps: 10160, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001654, Sample Num: 26464, Cur Loss: 0.70549536, Cur Avg Loss: 0.65544861, Log Avg loss: 1.54641314, Global Avg Loss: 3.73093469, Time: 0.0210 Steps: 10170, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001664, Sample Num: 26624, Cur Loss: 0.26794529, Cur Avg Loss: 0.65745182, Log Avg loss: 0.98878300, Global Avg Loss: 3.72824102, Time: 0.0211 Steps: 10180, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001674, Sample Num: 26784, Cur Loss: 1.74798393, Cur Avg Loss: 0.65821822, Log Avg loss: 0.78574602, Global Avg Loss: 3.72535339, Time: 0.0211 Steps: 10190, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001684, Sample Num: 26944, Cur Loss: 1.36324430, Cur Avg Loss: 0.65964059, Log Avg loss: 0.89774649, Global Avg Loss: 3.72258123, Time: 0.0211 Steps: 10200, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001694, Sample Num: 27104, Cur Loss: 0.23297076, Cur Avg Loss: 0.65889016, Log Avg loss: 0.53251696, Global Avg Loss: 3.71945678, Time: 0.0210 Steps: 10210, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001704, Sample Num: 27264, Cur Loss: 0.63293540, Cur Avg Loss: 0.65837999, Log Avg loss: 0.57195803, Global Avg Loss: 3.71637704, Time: 0.0211 Steps: 10220, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001714, Sample Num: 27424, Cur Loss: 0.31057030, Cur Avg Loss: 0.65815576, Log Avg loss: 0.61994654, Global Avg Loss: 3.71335022, Time: 0.0210 Steps: 10230, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001724, Sample Num: 27584, Cur Loss: 0.74859810, Cur Avg Loss: 0.65767357, Log Avg loss: 0.57502621, Global Avg Loss: 3.71028545, Time: 0.0211 Steps: 10240, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001734, Sample Num: 27744, Cur Loss: 1.27713323, Cur Avg Loss: 0.65799030, Log Avg loss: 0.71259429, Global Avg Loss: 3.70736088, Time: 0.0210 Steps: 10250, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001744, Sample Num: 27904, Cur Loss: 0.56095141, Cur Avg Loss: 0.65900309, Log Avg loss: 0.83462047, Global Avg Loss: 3.70456093, Time: 0.0210 Steps: 10260, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001754, Sample Num: 28064, Cur Loss: 0.55470252, Cur Avg Loss: 0.65906458, Log Avg loss: 0.66978802, Global Avg Loss: 3.70160595, Time: 0.0211 Steps: 10270, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001764, Sample Num: 28224, Cur Loss: 0.32160306, Cur Avg Loss: 0.65775469, Log Avg loss: 0.42800116, Global Avg Loss: 3.69842150, Time: 0.0210 Steps: 10280, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001774, Sample Num: 28384, Cur Loss: 0.73277432, Cur Avg Loss: 0.65853534, Log Avg loss: 0.79624187, Global Avg Loss: 3.69560112, Time: 0.0210 Steps: 10290, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001784, Sample Num: 28544, Cur Loss: 0.63000107, Cur Avg Loss: 0.65886719, Log Avg loss: 0.71773675, Global Avg Loss: 3.69270999, Time: 0.0210 Steps: 10300, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001794, Sample Num: 28704, Cur Loss: 0.39692730, Cur Avg Loss: 0.65790004, Log Avg loss: 0.48536069, Global Avg Loss: 3.68959907, Time: 0.0249 Steps: 10310, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001804, Sample Num: 28864, Cur Loss: 0.55331051, Cur Avg Loss: 0.65697457, Log Avg loss: 0.49094534, Global Avg Loss: 3.68649960, Time: 0.0210 Steps: 10320, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001814, Sample Num: 29024, Cur Loss: 0.92789060, Cur Avg Loss: 0.65695543, Log Avg loss: 0.65350218, Global Avg Loss: 3.68356350, Time: 0.0211 Steps: 10330, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001824, Sample Num: 29184, Cur Loss: 1.18582392, Cur Avg Loss: 0.65838430, Log Avg loss: 0.91758254, Global Avg Loss: 3.68088847, Time: 0.0210 Steps: 10340, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001834, Sample Num: 29344, Cur Loss: 0.89718008, Cur Avg Loss: 0.65919387, Log Avg loss: 0.80685834, Global Avg Loss: 3.67811163, Time: 0.0210 Steps: 10350, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001844, Sample Num: 29504, Cur Loss: 0.82667887, Cur Avg Loss: 0.65906759, Log Avg loss: 0.63590873, Global Avg Loss: 3.67517514, Time: 0.0210 Steps: 10360, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001854, Sample Num: 29664, Cur Loss: 0.30553606, Cur Avg Loss: 0.66007426, Log Avg loss: 0.84570310, Global Avg Loss: 3.67244662, Time: 0.0210 Steps: 10370, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001864, Sample Num: 29824, Cur Loss: 1.21616066, Cur Avg Loss: 0.66051875, Log Avg loss: 0.74292762, Global Avg Loss: 3.66962435, Time: 0.0210 Steps: 10380, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001874, Sample Num: 29984, Cur Loss: 1.93973589, Cur Avg Loss: 0.66166899, Log Avg loss: 0.87607462, Global Avg Loss: 3.66693566, Time: 0.0211 Steps: 10390, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001884, Sample Num: 30144, Cur Loss: 0.95592141, Cur Avg Loss: 0.66317201, Log Avg loss: 0.94483668, Global Avg Loss: 3.66431826, Time: 0.0210 Steps: 10400, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001894, Sample Num: 30304, Cur Loss: 0.75190592, Cur Avg Loss: 0.66234590, Log Avg loss: 0.50670805, Global Avg Loss: 3.66128501, Time: 0.0210 Steps: 10410, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001904, Sample Num: 30464, Cur Loss: 0.30345207, Cur Avg Loss: 0.66183982, Log Avg loss: 0.56598707, Global Avg Loss: 3.65831447, Time: 0.0210 Steps: 10420, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001914, Sample Num: 30624, Cur Loss: 0.10913180, Cur Avg Loss: 0.66050099, Log Avg loss: 0.40558847, Global Avg Loss: 3.65519585, Time: 0.0210 Steps: 10430, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001924, Sample Num: 30784, Cur Loss: 0.87412965, Cur Avg Loss: 0.66004409, Log Avg loss: 0.57259273, Global Avg Loss: 3.65224316, Time: 0.0211 Steps: 10440, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001934, Sample Num: 30944, Cur Loss: 0.86049271, Cur Avg Loss: 0.66062183, Log Avg loss: 0.77177942, Global Avg Loss: 3.64948674, Time: 0.0210 Steps: 10450, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001944, Sample Num: 31104, Cur Loss: 0.76925182, Cur Avg Loss: 0.65998176, Log Avg loss: 0.53619311, Global Avg Loss: 3.64651036, Time: 0.0211 Steps: 10460, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001954, Sample Num: 31264, Cur Loss: 0.36191887, Cur Avg Loss: 0.65997364, Log Avg loss: 0.65839480, Global Avg Loss: 3.64365638, Time: 0.0211 Steps: 10470, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001964, Sample Num: 31424, Cur Loss: 0.58575189, Cur Avg Loss: 0.65966003, Log Avg loss: 0.59837908, Global Avg Loss: 3.64075058, Time: 0.0210 Steps: 10480, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001974, Sample Num: 31584, Cur Loss: 0.31204808, Cur Avg Loss: 0.65854687, Log Avg loss: 0.43992389, Global Avg Loss: 3.63769927, Time: 0.0210 Steps: 10490, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001984, Sample Num: 31744, Cur Loss: 0.32319465, Cur Avg Loss: 0.65818985, Log Avg loss: 0.58771252, Global Avg Loss: 3.63479452, Time: 0.0210 Steps: 10500, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001994, Sample Num: 31904, Cur Loss: 0.53443491, Cur Avg Loss: 0.65772019, Log Avg loss: 0.56454037, Global Avg Loss: 3.63187325, Time: 0.0210 Steps: 10510, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002004, Sample Num: 32064, Cur Loss: 0.68741083, Cur Avg Loss: 0.65755897, Log Avg loss: 0.62541289, Global Avg Loss: 3.62901540, Time: 0.0210 Steps: 10520, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002014, Sample Num: 32224, Cur Loss: 0.31223482, Cur Avg Loss: 0.65897785, Log Avg loss: 0.94332017, Global Avg Loss: 3.62646488, Time: 0.0210 Steps: 10530, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002024, Sample Num: 32384, Cur Loss: 0.67139161, Cur Avg Loss: 0.65743727, Log Avg loss: 0.34716454, Global Avg Loss: 3.62335359, Time: 0.0210 Steps: 10540, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002034, Sample Num: 32544, Cur Loss: 1.23416543, Cur Avg Loss: 0.65741821, Log Avg loss: 0.65356054, Global Avg Loss: 3.62053862, Time: 0.0210 Steps: 10550, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002044, Sample Num: 32704, Cur Loss: 0.36453360, Cur Avg Loss: 0.65723531, Log Avg loss: 0.62003380, Global Avg Loss: 3.61769723, Time: 0.0210 Steps: 10560, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002054, Sample Num: 32864, Cur Loss: 0.61580485, Cur Avg Loss: 0.65687422, Log Avg loss: 0.58306713, Global Avg Loss: 3.61482625, Time: 0.0213 Steps: 10570, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002064, Sample Num: 33024, Cur Loss: 0.42183846, Cur Avg Loss: 0.65564322, Log Avg loss: 0.40279671, Global Avg Loss: 3.61179030, Time: 0.0211 Steps: 10580, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002074, Sample Num: 33184, Cur Loss: 0.26255596, Cur Avg Loss: 0.65470775, Log Avg loss: 0.46162614, Global Avg Loss: 3.60881564, Time: 0.0211 Steps: 10590, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002084, Sample Num: 33344, Cur Loss: 1.23471427, Cur Avg Loss: 0.65411519, Log Avg loss: 0.53121703, Global Avg Loss: 3.60591225, Time: 0.0211 Steps: 10600, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002094, Sample Num: 33504, Cur Loss: 0.46124530, Cur Avg Loss: 0.65333338, Log Avg loss: 0.49040424, Global Avg Loss: 3.60297586, Time: 0.0211 Steps: 10610, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002104, Sample Num: 33664, Cur Loss: 1.42667007, Cur Avg Loss: 0.65325460, Log Avg loss: 0.63675849, Global Avg Loss: 3.60018281, Time: 0.0212 Steps: 10620, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002114, Sample Num: 33824, Cur Loss: 1.26992834, Cur Avg Loss: 0.65422623, Log Avg loss: 0.85865791, Global Avg Loss: 3.59760377, Time: 0.0211 Steps: 10630, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002124, Sample Num: 33984, Cur Loss: 0.22515054, Cur Avg Loss: 0.65409535, Log Avg loss: 0.62642720, Global Avg Loss: 3.59481131, Time: 0.0210 Steps: 10640, Updated lr: 0.000091 ***** Running evaluation checkpoint-10645 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-10645 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.022552, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.520118, "eval_total_loss": 365.643147, "eval_mae": 0.547694, "eval_mse": 0.52028, "eval_r2": 0.669276, "eval_sp_statistic": 0.801483, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.826696, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.111904, "test_total_loss": 558.175992, "test_mae": 0.974194, "test_mse": 1.112088, "test_r2": 0.282249, "test_sp_statistic": 0.795144, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.85355, "test_ps_pvalue": 0.0, "lr": 9.085348506401138e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.593304255402119, "train_cur_epoch_loss": 1391.2300084084272, "train_cur_epoch_avg_loss": 0.6534664201072932, "train_cur_epoch_time": 45.02255201339722, "train_cur_epoch_avg_time": 0.021147276661999632, "epoch": 5, "step": 10645} ################################################## Training, Epoch: 0006, Batch: 000005, Sample Num: 80, Cur Loss: 0.26465654, Cur Avg Loss: 0.51787963, Log Avg loss: 0.45208817, Global Avg Loss: 3.59186039, Time: 0.0247 Steps: 10650, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000015, Sample Num: 240, Cur Loss: 0.24312031, Cur Avg Loss: 0.42310250, Log Avg loss: 0.37571394, Global Avg Loss: 3.58884337, Time: 0.0211 Steps: 10660, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000025, Sample Num: 400, Cur Loss: 0.72823894, Cur Avg Loss: 0.46261277, Log Avg loss: 0.52187818, Global Avg Loss: 3.58596899, Time: 0.0211 Steps: 10670, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000035, Sample Num: 560, Cur Loss: 0.45154113, Cur Avg Loss: 0.52786328, Log Avg loss: 0.69098955, Global Avg Loss: 3.58325833, Time: 0.0210 Steps: 10680, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000045, Sample Num: 720, Cur Loss: 0.72013974, Cur Avg Loss: 0.53816411, Log Avg loss: 0.57421703, Global Avg Loss: 3.58044352, Time: 0.0210 Steps: 10690, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000055, Sample Num: 880, Cur Loss: 0.34588477, Cur Avg Loss: 0.52116298, Log Avg loss: 0.44465789, Global Avg Loss: 3.57751288, Time: 0.0210 Steps: 10700, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000065, Sample Num: 1040, Cur Loss: 0.62821883, Cur Avg Loss: 0.49710922, Log Avg loss: 0.36481353, Global Avg Loss: 3.57451316, Time: 0.0211 Steps: 10710, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000075, Sample Num: 1200, Cur Loss: 0.36586410, Cur Avg Loss: 0.49296157, Log Avg loss: 0.46600184, Global Avg Loss: 3.57161343, Time: 0.0210 Steps: 10720, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000085, Sample Num: 1360, Cur Loss: 0.43584970, Cur Avg Loss: 0.53600775, Log Avg loss: 0.85885414, Global Avg Loss: 3.56908522, Time: 0.0212 Steps: 10730, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000095, Sample Num: 1520, Cur Loss: 0.29216152, Cur Avg Loss: 0.55208836, Log Avg loss: 0.68877347, Global Avg Loss: 3.56640337, Time: 0.0211 Steps: 10740, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000105, Sample Num: 1680, Cur Loss: 0.83057123, Cur Avg Loss: 0.61385388, Log Avg loss: 1.20062638, Global Avg Loss: 3.56420265, Time: 0.0211 Steps: 10750, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000115, Sample Num: 1840, Cur Loss: 0.29652515, Cur Avg Loss: 0.61294981, Log Avg loss: 0.60345707, Global Avg Loss: 3.56145102, Time: 0.0212 Steps: 10760, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000125, Sample Num: 2000, Cur Loss: 0.77363694, Cur Avg Loss: 0.61807048, Log Avg loss: 0.67695813, Global Avg Loss: 3.55877276, Time: 0.0211 Steps: 10770, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000135, Sample Num: 2160, Cur Loss: 0.29641926, Cur Avg Loss: 0.60770584, Log Avg loss: 0.47814788, Global Avg Loss: 3.55591504, Time: 0.0212 Steps: 10780, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000145, Sample Num: 2320, Cur Loss: 0.28382492, Cur Avg Loss: 0.60479552, Log Avg loss: 0.56550617, Global Avg Loss: 3.55314357, Time: 0.0211 Steps: 10790, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000155, Sample Num: 2480, Cur Loss: 0.56033063, Cur Avg Loss: 0.60253078, Log Avg loss: 0.56969200, Global Avg Loss: 3.55038112, Time: 0.0210 Steps: 10800, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000165, Sample Num: 2640, Cur Loss: 0.40357795, Cur Avg Loss: 0.59757201, Log Avg loss: 0.52071115, Global Avg Loss: 3.54757846, Time: 0.0212 Steps: 10810, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000175, Sample Num: 2800, Cur Loss: 0.42476958, Cur Avg Loss: 0.60243461, Log Avg loss: 0.68266755, Global Avg Loss: 3.54493067, Time: 0.0210 Steps: 10820, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000185, Sample Num: 2960, Cur Loss: 0.42143172, Cur Avg Loss: 0.59548811, Log Avg loss: 0.47392438, Global Avg Loss: 3.54209502, Time: 0.0211 Steps: 10830, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000195, Sample Num: 3120, Cur Loss: 0.28850991, Cur Avg Loss: 0.58712264, Log Avg loss: 0.43236145, Global Avg Loss: 3.53922627, Time: 0.0211 Steps: 10840, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000205, Sample Num: 3280, Cur Loss: 0.71696985, Cur Avg Loss: 0.58257844, Log Avg loss: 0.49396648, Global Avg Loss: 3.53641957, Time: 0.0211 Steps: 10850, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000215, Sample Num: 3440, Cur Loss: 0.40660274, Cur Avg Loss: 0.58397770, Log Avg loss: 0.61266253, Global Avg Loss: 3.53372735, Time: 0.0211 Steps: 10860, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000225, Sample Num: 3600, Cur Loss: 0.95075697, Cur Avg Loss: 0.58309775, Log Avg loss: 0.56417885, Global Avg Loss: 3.53099547, Time: 0.0212 Steps: 10870, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000235, Sample Num: 3760, Cur Loss: 0.60666549, Cur Avg Loss: 0.57890201, Log Avg loss: 0.48449783, Global Avg Loss: 3.52819538, Time: 0.0210 Steps: 10880, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000245, Sample Num: 3920, Cur Loss: 0.71656400, Cur Avg Loss: 0.57692571, Log Avg loss: 0.53048276, Global Avg Loss: 3.52544266, Time: 0.0211 Steps: 10890, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000255, Sample Num: 4080, Cur Loss: 0.80886757, Cur Avg Loss: 0.58053912, Log Avg loss: 0.66906768, Global Avg Loss: 3.52282214, Time: 0.0211 Steps: 10900, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000265, Sample Num: 4240, Cur Loss: 0.20507324, Cur Avg Loss: 0.58153548, Log Avg loss: 0.60694245, Global Avg Loss: 3.52014947, Time: 0.0212 Steps: 10910, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000275, Sample Num: 4400, Cur Loss: 0.33080432, Cur Avg Loss: 0.58255708, Log Avg loss: 0.60962972, Global Avg Loss: 3.51748416, Time: 0.0211 Steps: 10920, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000285, Sample Num: 4560, Cur Loss: 0.97236180, Cur Avg Loss: 0.59221449, Log Avg loss: 0.85779309, Global Avg Loss: 3.51505077, Time: 0.0211 Steps: 10930, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000295, Sample Num: 4720, Cur Loss: 1.14481640, Cur Avg Loss: 0.60484399, Log Avg loss: 0.96478472, Global Avg Loss: 3.51271963, Time: 0.0211 Steps: 10940, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000305, Sample Num: 4880, Cur Loss: 0.59055847, Cur Avg Loss: 0.60235752, Log Avg loss: 0.52900669, Global Avg Loss: 3.50999478, Time: 0.0211 Steps: 10950, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000315, Sample Num: 5040, Cur Loss: 0.68437040, Cur Avg Loss: 0.59954716, Log Avg loss: 0.51383106, Global Avg Loss: 3.50726105, Time: 0.0211 Steps: 10960, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000325, Sample Num: 5200, Cur Loss: 0.46802339, Cur Avg Loss: 0.59615627, Log Avg loss: 0.48934333, Global Avg Loss: 3.50450999, Time: 0.0211 Steps: 10970, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000335, Sample Num: 5360, Cur Loss: 0.36075827, Cur Avg Loss: 0.59231517, Log Avg loss: 0.46747942, Global Avg Loss: 3.50174402, Time: 0.0211 Steps: 10980, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000345, Sample Num: 5520, Cur Loss: 0.16511837, Cur Avg Loss: 0.58998054, Log Avg loss: 0.51177039, Global Avg Loss: 3.49902339, Time: 0.0211 Steps: 10990, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000355, Sample Num: 5680, Cur Loss: 0.34827733, Cur Avg Loss: 0.59615539, Log Avg loss: 0.80918776, Global Avg Loss: 3.49657809, Time: 0.0211 Steps: 11000, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000365, Sample Num: 5840, Cur Loss: 0.42851597, Cur Avg Loss: 0.60340552, Log Avg loss: 0.86078504, Global Avg Loss: 3.49418409, Time: 0.0211 Steps: 11010, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000375, Sample Num: 6000, Cur Loss: 0.46571478, Cur Avg Loss: 0.60174389, Log Avg loss: 0.54109436, Global Avg Loss: 3.49150433, Time: 0.0211 Steps: 11020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000385, Sample Num: 6160, Cur Loss: 0.56178623, Cur Avg Loss: 0.60052013, Log Avg loss: 0.55462912, Global Avg Loss: 3.48884171, Time: 0.0211 Steps: 11030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000395, Sample Num: 6320, Cur Loss: 1.54776096, Cur Avg Loss: 0.60283134, Log Avg loss: 0.69181314, Global Avg Loss: 3.48630817, Time: 0.0211 Steps: 11040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000405, Sample Num: 6480, Cur Loss: 0.41530323, Cur Avg Loss: 0.60519156, Log Avg loss: 0.69842007, Global Avg Loss: 3.48378519, Time: 0.0211 Steps: 11050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000415, Sample Num: 6640, Cur Loss: 0.36315796, Cur Avg Loss: 0.59979313, Log Avg loss: 0.38115684, Global Avg Loss: 3.48097992, Time: 0.0211 Steps: 11060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000425, Sample Num: 6800, Cur Loss: 0.62601852, Cur Avg Loss: 0.59879378, Log Avg loss: 0.55732058, Global Avg Loss: 3.47833886, Time: 0.0211 Steps: 11070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000435, Sample Num: 6960, Cur Loss: 0.82877839, Cur Avg Loss: 0.60512550, Log Avg loss: 0.87422395, Global Avg Loss: 3.47598857, Time: 0.0211 Steps: 11080, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000445, Sample Num: 7120, Cur Loss: 0.12565123, Cur Avg Loss: 0.60847843, Log Avg loss: 0.75433095, Global Avg Loss: 3.47353442, Time: 0.0211 Steps: 11090, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000455, Sample Num: 7280, Cur Loss: 0.67491341, Cur Avg Loss: 0.61004578, Log Avg loss: 0.67979255, Global Avg Loss: 3.47101753, Time: 0.0211 Steps: 11100, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000465, Sample Num: 7440, Cur Loss: 0.92142415, Cur Avg Loss: 0.60961499, Log Avg loss: 0.59001412, Global Avg Loss: 3.46842437, Time: 0.0211 Steps: 11110, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000475, Sample Num: 7600, Cur Loss: 1.29152799, Cur Avg Loss: 0.60857997, Log Avg loss: 0.56045138, Global Avg Loss: 3.46580929, Time: 0.0211 Steps: 11120, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000485, Sample Num: 7760, Cur Loss: 0.42321494, Cur Avg Loss: 0.60609851, Log Avg loss: 0.48822915, Global Avg Loss: 3.46313401, Time: 0.0211 Steps: 11130, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000495, Sample Num: 7920, Cur Loss: 0.27202439, Cur Avg Loss: 0.60633633, Log Avg loss: 0.61787062, Global Avg Loss: 3.46057992, Time: 0.0211 Steps: 11140, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000505, Sample Num: 8080, Cur Loss: 1.25830269, Cur Avg Loss: 0.60506634, Log Avg loss: 0.54220201, Global Avg Loss: 3.45796254, Time: 0.0210 Steps: 11150, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000515, Sample Num: 8240, Cur Loss: 0.79715002, Cur Avg Loss: 0.60480338, Log Avg loss: 0.59152393, Global Avg Loss: 3.45539404, Time: 0.0249 Steps: 11160, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000525, Sample Num: 8400, Cur Loss: 1.12422323, Cur Avg Loss: 0.60613933, Log Avg loss: 0.67494092, Global Avg Loss: 3.45290483, Time: 0.0212 Steps: 11170, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000535, Sample Num: 8560, Cur Loss: 0.35958159, Cur Avg Loss: 0.60437960, Log Avg loss: 0.51199339, Global Avg Loss: 3.45027432, Time: 0.0212 Steps: 11180, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000545, Sample Num: 8720, Cur Loss: 0.36354196, Cur Avg Loss: 0.60158843, Log Avg loss: 0.45226082, Global Avg Loss: 3.44759513, Time: 0.0212 Steps: 11190, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000555, Sample Num: 8880, Cur Loss: 0.27986580, Cur Avg Loss: 0.59845865, Log Avg loss: 0.42788585, Global Avg Loss: 3.44489896, Time: 0.0212 Steps: 11200, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000565, Sample Num: 9040, Cur Loss: 0.22058764, Cur Avg Loss: 0.59451802, Log Avg loss: 0.37581330, Global Avg Loss: 3.44216115, Time: 0.0212 Steps: 11210, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000575, Sample Num: 9200, Cur Loss: 0.15599571, Cur Avg Loss: 0.59002570, Log Avg loss: 0.33620955, Global Avg Loss: 3.43939292, Time: 0.0212 Steps: 11220, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000585, Sample Num: 9360, Cur Loss: 0.78006947, Cur Avg Loss: 0.59342484, Log Avg loss: 0.78887541, Global Avg Loss: 3.43703271, Time: 0.0212 Steps: 11230, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000595, Sample Num: 9520, Cur Loss: 0.48250288, Cur Avg Loss: 0.58975938, Log Avg loss: 0.37532998, Global Avg Loss: 3.43430878, Time: 0.0212 Steps: 11240, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000605, Sample Num: 9680, Cur Loss: 0.21229883, Cur Avg Loss: 0.59102246, Log Avg loss: 0.66617574, Global Avg Loss: 3.43184821, Time: 0.0212 Steps: 11250, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000615, Sample Num: 9840, Cur Loss: 0.76636034, Cur Avg Loss: 0.58987142, Log Avg loss: 0.52023334, Global Avg Loss: 3.42926241, Time: 0.0212 Steps: 11260, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000625, Sample Num: 10000, Cur Loss: 0.20937309, Cur Avg Loss: 0.58766951, Log Avg loss: 0.45225219, Global Avg Loss: 3.42662087, Time: 0.0213 Steps: 11270, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000635, Sample Num: 10160, Cur Loss: 0.24342874, Cur Avg Loss: 0.58951612, Log Avg loss: 0.70492879, Global Avg Loss: 3.42420803, Time: 0.0212 Steps: 11280, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000645, Sample Num: 10320, Cur Loss: 1.61521089, Cur Avg Loss: 0.58977054, Log Avg loss: 0.60592623, Global Avg Loss: 3.42171176, Time: 0.0212 Steps: 11290, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000655, Sample Num: 10480, Cur Loss: 0.52707040, Cur Avg Loss: 0.58867841, Log Avg loss: 0.51823596, Global Avg Loss: 3.41914231, Time: 0.0212 Steps: 11300, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000665, Sample Num: 10640, Cur Loss: 0.29686511, Cur Avg Loss: 0.58818415, Log Avg loss: 0.55581021, Global Avg Loss: 3.41661063, Time: 0.0211 Steps: 11310, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000675, Sample Num: 10800, Cur Loss: 1.19579470, Cur Avg Loss: 0.59029328, Log Avg loss: 0.73055050, Global Avg Loss: 3.41423779, Time: 0.0212 Steps: 11320, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000685, Sample Num: 10960, Cur Loss: 1.20787060, Cur Avg Loss: 0.59447812, Log Avg loss: 0.87695525, Global Avg Loss: 3.41199835, Time: 0.0211 Steps: 11330, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000695, Sample Num: 11120, Cur Loss: 0.40235168, Cur Avg Loss: 0.59478832, Log Avg loss: 0.61603653, Global Avg Loss: 3.40953278, Time: 0.0211 Steps: 11340, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000705, Sample Num: 11280, Cur Loss: 0.63465631, Cur Avg Loss: 0.59692158, Log Avg loss: 0.74518333, Global Avg Loss: 3.40718533, Time: 0.0212 Steps: 11350, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000715, Sample Num: 11440, Cur Loss: 0.74490017, Cur Avg Loss: 0.59723658, Log Avg loss: 0.61944393, Global Avg Loss: 3.40473133, Time: 0.0212 Steps: 11360, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000725, Sample Num: 11600, Cur Loss: 0.26657516, Cur Avg Loss: 0.59504924, Log Avg loss: 0.43865456, Global Avg Loss: 3.40212265, Time: 0.0212 Steps: 11370, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000735, Sample Num: 11760, Cur Loss: 1.15038741, Cur Avg Loss: 0.59380706, Log Avg loss: 0.50374891, Global Avg Loss: 3.39957575, Time: 0.0212 Steps: 11380, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000745, Sample Num: 11920, Cur Loss: 1.34950566, Cur Avg Loss: 0.59330602, Log Avg loss: 0.55647962, Global Avg Loss: 3.39707961, Time: 0.0212 Steps: 11390, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000755, Sample Num: 12080, Cur Loss: 0.07571021, Cur Avg Loss: 0.59170965, Log Avg loss: 0.47277985, Global Avg Loss: 3.39451444, Time: 0.0213 Steps: 11400, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000765, Sample Num: 12240, Cur Loss: 0.26127586, Cur Avg Loss: 0.59117124, Log Avg loss: 0.55052179, Global Avg Loss: 3.39202189, Time: 0.0212 Steps: 11410, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000775, Sample Num: 12400, Cur Loss: 0.31328398, Cur Avg Loss: 0.58935991, Log Avg loss: 0.45079312, Global Avg Loss: 3.38944639, Time: 0.0210 Steps: 11420, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000785, Sample Num: 12560, Cur Loss: 0.19061190, Cur Avg Loss: 0.58762575, Log Avg loss: 0.45322834, Global Avg Loss: 3.38687752, Time: 0.0210 Steps: 11430, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000795, Sample Num: 12720, Cur Loss: 0.42626423, Cur Avg Loss: 0.58578111, Log Avg loss: 0.44097694, Global Avg Loss: 3.38430243, Time: 0.0210 Steps: 11440, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000805, Sample Num: 12880, Cur Loss: 0.30592769, Cur Avg Loss: 0.58610302, Log Avg loss: 0.61169449, Global Avg Loss: 3.38188094, Time: 0.0211 Steps: 11450, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000815, Sample Num: 13040, Cur Loss: 0.49301931, Cur Avg Loss: 0.58547598, Log Avg loss: 0.53499972, Global Avg Loss: 3.37939675, Time: 0.0210 Steps: 11460, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000825, Sample Num: 13200, Cur Loss: 0.21095879, Cur Avg Loss: 0.58588378, Log Avg loss: 0.61911882, Global Avg Loss: 3.37699023, Time: 0.0210 Steps: 11470, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000835, Sample Num: 13360, Cur Loss: 0.55495715, Cur Avg Loss: 0.58468637, Log Avg loss: 0.48590012, Global Avg Loss: 3.37447186, Time: 0.0211 Steps: 11480, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000845, Sample Num: 13520, Cur Loss: 0.39830875, Cur Avg Loss: 0.58604885, Log Avg loss: 0.69981642, Global Avg Loss: 3.37214405, Time: 0.0210 Steps: 11490, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000855, Sample Num: 13680, Cur Loss: 1.11489975, Cur Avg Loss: 0.58752082, Log Avg loss: 0.71190174, Global Avg Loss: 3.36983079, Time: 0.0210 Steps: 11500, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000865, Sample Num: 13840, Cur Loss: 0.55119348, Cur Avg Loss: 0.58785757, Log Avg loss: 0.61664975, Global Avg Loss: 3.36743880, Time: 0.0210 Steps: 11510, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000875, Sample Num: 14000, Cur Loss: 0.64223778, Cur Avg Loss: 0.58789811, Log Avg loss: 0.59140490, Global Avg Loss: 3.36502905, Time: 0.0210 Steps: 11520, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000885, Sample Num: 14160, Cur Loss: 0.46469542, Cur Avg Loss: 0.58559393, Log Avg loss: 0.38397883, Global Avg Loss: 3.36244358, Time: 0.0210 Steps: 11530, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000895, Sample Num: 14320, Cur Loss: 0.36687827, Cur Avg Loss: 0.58286041, Log Avg loss: 0.34094325, Global Avg Loss: 3.35982529, Time: 0.0210 Steps: 11540, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000905, Sample Num: 14480, Cur Loss: 0.81407261, Cur Avg Loss: 0.58670018, Log Avg loss: 0.93035937, Global Avg Loss: 3.35772186, Time: 0.0211 Steps: 11550, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000915, Sample Num: 14640, Cur Loss: 1.04588985, Cur Avg Loss: 0.58879081, Log Avg loss: 0.77799317, Global Avg Loss: 3.35549026, Time: 0.0210 Steps: 11560, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000925, Sample Num: 14800, Cur Loss: 0.66861045, Cur Avg Loss: 0.58771812, Log Avg loss: 0.48956690, Global Avg Loss: 3.35301323, Time: 0.0211 Steps: 11570, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000935, Sample Num: 14960, Cur Loss: 0.79639661, Cur Avg Loss: 0.58749101, Log Avg loss: 0.56648390, Global Avg Loss: 3.35060690, Time: 0.0210 Steps: 11580, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000945, Sample Num: 15120, Cur Loss: 0.22861925, Cur Avg Loss: 0.58655107, Log Avg loss: 0.49866588, Global Avg Loss: 3.34814621, Time: 0.0210 Steps: 11590, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000955, Sample Num: 15280, Cur Loss: 1.32352579, Cur Avg Loss: 0.58658651, Log Avg loss: 0.58993583, Global Avg Loss: 3.34576844, Time: 0.0210 Steps: 11600, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000965, Sample Num: 15440, Cur Loss: 0.52407765, Cur Avg Loss: 0.58569310, Log Avg loss: 0.50037232, Global Avg Loss: 3.34331763, Time: 0.0210 Steps: 11610, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000975, Sample Num: 15600, Cur Loss: 0.29774895, Cur Avg Loss: 0.58740997, Log Avg loss: 0.75308859, Global Avg Loss: 3.34108851, Time: 0.0210 Steps: 11620, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000985, Sample Num: 15760, Cur Loss: 0.89689738, Cur Avg Loss: 0.58810331, Log Avg loss: 0.65570373, Global Avg Loss: 3.33877950, Time: 0.0211 Steps: 11630, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000995, Sample Num: 15920, Cur Loss: 0.46711585, Cur Avg Loss: 0.58964773, Log Avg loss: 0.74177299, Global Avg Loss: 3.33654839, Time: 0.0211 Steps: 11640, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001005, Sample Num: 16080, Cur Loss: 0.43288296, Cur Avg Loss: 0.59116665, Log Avg loss: 0.74229946, Global Avg Loss: 3.33432157, Time: 0.0211 Steps: 11650, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001015, Sample Num: 16240, Cur Loss: 0.58761871, Cur Avg Loss: 0.59074635, Log Avg loss: 0.54850580, Global Avg Loss: 3.33193236, Time: 0.0211 Steps: 11660, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001025, Sample Num: 16400, Cur Loss: 0.29863036, Cur Avg Loss: 0.59043151, Log Avg loss: 0.55847539, Global Avg Loss: 3.32955579, Time: 0.0247 Steps: 11670, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001035, Sample Num: 16560, Cur Loss: 0.37040541, Cur Avg Loss: 0.58959091, Log Avg loss: 0.50342932, Global Avg Loss: 3.32713616, Time: 0.0210 Steps: 11680, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001045, Sample Num: 16720, Cur Loss: 0.25589865, Cur Avg Loss: 0.59104971, Log Avg loss: 0.74203595, Global Avg Loss: 3.32492479, Time: 0.0210 Steps: 11690, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001055, Sample Num: 16880, Cur Loss: 0.48951924, Cur Avg Loss: 0.58970430, Log Avg loss: 0.44910834, Global Avg Loss: 3.32246682, Time: 0.0210 Steps: 11700, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001065, Sample Num: 17040, Cur Loss: 0.98157692, Cur Avg Loss: 0.59077569, Log Avg loss: 0.70380798, Global Avg Loss: 3.32023056, Time: 0.0210 Steps: 11710, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001075, Sample Num: 17200, Cur Loss: 0.27572900, Cur Avg Loss: 0.58925184, Log Avg loss: 0.42696155, Global Avg Loss: 3.31776191, Time: 0.0210 Steps: 11720, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001085, Sample Num: 17360, Cur Loss: 0.84018600, Cur Avg Loss: 0.58787583, Log Avg loss: 0.43995516, Global Avg Loss: 3.31530853, Time: 0.0210 Steps: 11730, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001095, Sample Num: 17520, Cur Loss: 1.25816178, Cur Avg Loss: 0.58754789, Log Avg loss: 0.55196620, Global Avg Loss: 3.31295475, Time: 0.0210 Steps: 11740, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001105, Sample Num: 17680, Cur Loss: 0.19933587, Cur Avg Loss: 0.58583111, Log Avg loss: 0.39784293, Global Avg Loss: 3.31047380, Time: 0.0210 Steps: 11750, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001115, Sample Num: 17840, Cur Loss: 0.24678485, Cur Avg Loss: 0.58591654, Log Avg loss: 0.59535746, Global Avg Loss: 3.30816503, Time: 0.0211 Steps: 11760, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001125, Sample Num: 18000, Cur Loss: 0.49763730, Cur Avg Loss: 0.58512943, Log Avg loss: 0.49736573, Global Avg Loss: 3.30577692, Time: 0.0210 Steps: 11770, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001135, Sample Num: 18160, Cur Loss: 0.31914282, Cur Avg Loss: 0.58470713, Log Avg loss: 0.53719885, Global Avg Loss: 3.30342669, Time: 0.0210 Steps: 11780, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001145, Sample Num: 18320, Cur Loss: 0.65967596, Cur Avg Loss: 0.58499468, Log Avg loss: 0.61763155, Global Avg Loss: 3.30114866, Time: 0.0210 Steps: 11790, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001155, Sample Num: 18480, Cur Loss: 0.84218693, Cur Avg Loss: 0.58496951, Log Avg loss: 0.58208798, Global Avg Loss: 3.29884437, Time: 0.0210 Steps: 11800, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001165, Sample Num: 18640, Cur Loss: 0.61653256, Cur Avg Loss: 0.58415983, Log Avg loss: 0.49064160, Global Avg Loss: 3.29646655, Time: 0.0210 Steps: 11810, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001175, Sample Num: 18800, Cur Loss: 0.27272469, Cur Avg Loss: 0.58325442, Log Avg loss: 0.47777395, Global Avg Loss: 3.29408187, Time: 0.0210 Steps: 11820, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001185, Sample Num: 18960, Cur Loss: 0.35405511, Cur Avg Loss: 0.58226667, Log Avg loss: 0.46620556, Global Avg Loss: 3.29169145, Time: 0.0210 Steps: 11830, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001195, Sample Num: 19120, Cur Loss: 0.73550975, Cur Avg Loss: 0.58161445, Log Avg loss: 0.50432728, Global Avg Loss: 3.28933725, Time: 0.0210 Steps: 11840, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001205, Sample Num: 19280, Cur Loss: 0.54548645, Cur Avg Loss: 0.58141367, Log Avg loss: 0.55742018, Global Avg Loss: 3.28703184, Time: 0.0210 Steps: 11850, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001215, Sample Num: 19440, Cur Loss: 0.45910829, Cur Avg Loss: 0.58223412, Log Avg loss: 0.68109794, Global Avg Loss: 3.28483459, Time: 0.0210 Steps: 11860, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001225, Sample Num: 19600, Cur Loss: 0.56567383, Cur Avg Loss: 0.58433070, Log Avg loss: 0.83906506, Global Avg Loss: 3.28277413, Time: 0.0210 Steps: 11870, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001235, Sample Num: 19760, Cur Loss: 0.36528760, Cur Avg Loss: 0.58425646, Log Avg loss: 0.57516196, Global Avg Loss: 3.28049499, Time: 0.0210 Steps: 11880, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001245, Sample Num: 19920, Cur Loss: 0.47473729, Cur Avg Loss: 0.58400212, Log Avg loss: 0.55259118, Global Avg Loss: 3.27820071, Time: 0.0210 Steps: 11890, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001255, Sample Num: 20080, Cur Loss: 0.41756943, Cur Avg Loss: 0.58412514, Log Avg loss: 0.59944201, Global Avg Loss: 3.27594965, Time: 0.0211 Steps: 11900, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001265, Sample Num: 20240, Cur Loss: 0.26037121, Cur Avg Loss: 0.58231227, Log Avg loss: 0.35479641, Global Avg Loss: 3.27349696, Time: 0.0210 Steps: 11910, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001275, Sample Num: 20400, Cur Loss: 0.51041126, Cur Avg Loss: 0.58248674, Log Avg loss: 0.60455688, Global Avg Loss: 3.27125792, Time: 0.0211 Steps: 11920, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001285, Sample Num: 20560, Cur Loss: 0.29527315, Cur Avg Loss: 0.58092101, Log Avg loss: 0.38129098, Global Avg Loss: 3.26883548, Time: 0.0248 Steps: 11930, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001295, Sample Num: 20720, Cur Loss: 0.66723025, Cur Avg Loss: 0.58025771, Log Avg loss: 0.49502325, Global Avg Loss: 3.26651236, Time: 0.0210 Steps: 11940, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001305, Sample Num: 20880, Cur Loss: 0.84972048, Cur Avg Loss: 0.58192291, Log Avg loss: 0.79756629, Global Avg Loss: 3.26444629, Time: 0.0210 Steps: 11950, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001315, Sample Num: 21040, Cur Loss: 0.72689378, Cur Avg Loss: 0.58178692, Log Avg loss: 0.56404005, Global Avg Loss: 3.26218843, Time: 0.0210 Steps: 11960, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001325, Sample Num: 21200, Cur Loss: 0.33522832, Cur Avg Loss: 0.58168044, Log Avg loss: 0.56767858, Global Avg Loss: 3.25993737, Time: 0.0211 Steps: 11970, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001335, Sample Num: 21360, Cur Loss: 0.19941288, Cur Avg Loss: 0.58119248, Log Avg loss: 0.51653821, Global Avg Loss: 3.25764739, Time: 0.0211 Steps: 11980, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001345, Sample Num: 21520, Cur Loss: 0.46570146, Cur Avg Loss: 0.58073380, Log Avg loss: 0.51950028, Global Avg Loss: 3.25536370, Time: 0.0211 Steps: 11990, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001355, Sample Num: 21680, Cur Loss: 0.72407019, Cur Avg Loss: 0.58038888, Log Avg loss: 0.53399628, Global Avg Loss: 3.25309589, Time: 0.0210 Steps: 12000, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001365, Sample Num: 21840, Cur Loss: 0.51014656, Cur Avg Loss: 0.57977244, Log Avg loss: 0.49624519, Global Avg Loss: 3.25080043, Time: 0.0211 Steps: 12010, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001375, Sample Num: 22000, Cur Loss: 0.40986130, Cur Avg Loss: 0.57934872, Log Avg loss: 0.52151130, Global Avg Loss: 3.24852981, Time: 0.0211 Steps: 12020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001385, Sample Num: 22160, Cur Loss: 0.55890775, Cur Avg Loss: 0.58084210, Log Avg loss: 0.78618104, Global Avg Loss: 3.24648297, Time: 0.0211 Steps: 12030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001395, Sample Num: 22320, Cur Loss: 0.37443697, Cur Avg Loss: 0.58129607, Log Avg loss: 0.64417152, Global Avg Loss: 3.24432158, Time: 0.0211 Steps: 12040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001405, Sample Num: 22480, Cur Loss: 0.16177349, Cur Avg Loss: 0.58033500, Log Avg loss: 0.44626627, Global Avg Loss: 3.24199954, Time: 0.0211 Steps: 12050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001415, Sample Num: 22640, Cur Loss: 0.21762829, Cur Avg Loss: 0.57899881, Log Avg loss: 0.39126396, Global Avg Loss: 3.23963575, Time: 0.0211 Steps: 12060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001425, Sample Num: 22800, Cur Loss: 0.54921800, Cur Avg Loss: 0.57974943, Log Avg loss: 0.68596110, Global Avg Loss: 3.23752003, Time: 0.0211 Steps: 12070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001435, Sample Num: 22960, Cur Loss: 0.36872464, Cur Avg Loss: 0.58059875, Log Avg loss: 0.70162705, Global Avg Loss: 3.23542078, Time: 0.0211 Steps: 12080, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001445, Sample Num: 23120, Cur Loss: 0.70564151, Cur Avg Loss: 0.58216558, Log Avg loss: 0.80700679, Global Avg Loss: 3.23341216, Time: 0.0211 Steps: 12090, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001455, Sample Num: 23280, Cur Loss: 1.47740579, Cur Avg Loss: 0.58424164, Log Avg loss: 0.88423212, Global Avg Loss: 3.23147069, Time: 0.0211 Steps: 12100, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001465, Sample Num: 23440, Cur Loss: 0.94436502, Cur Avg Loss: 0.58534676, Log Avg loss: 0.74614186, Global Avg Loss: 3.22941840, Time: 0.0211 Steps: 12110, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001475, Sample Num: 23600, Cur Loss: 0.31375486, Cur Avg Loss: 0.58608989, Log Avg loss: 0.69495761, Global Avg Loss: 3.22732726, Time: 0.0211 Steps: 12120, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001485, Sample Num: 23760, Cur Loss: 0.89015090, Cur Avg Loss: 0.58656977, Log Avg loss: 0.65735227, Global Avg Loss: 3.22520857, Time: 0.0211 Steps: 12130, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001495, Sample Num: 23920, Cur Loss: 0.63527107, Cur Avg Loss: 0.58741345, Log Avg loss: 0.71269992, Global Avg Loss: 3.22313895, Time: 0.0211 Steps: 12140, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001505, Sample Num: 24080, Cur Loss: 0.39987552, Cur Avg Loss: 0.58762531, Log Avg loss: 0.61929883, Global Avg Loss: 3.22099588, Time: 0.0211 Steps: 12150, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001515, Sample Num: 24240, Cur Loss: 0.25035915, Cur Avg Loss: 0.58684876, Log Avg loss: 0.46997831, Global Avg Loss: 3.21873353, Time: 0.0211 Steps: 12160, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001525, Sample Num: 24400, Cur Loss: 0.38148010, Cur Avg Loss: 0.58544455, Log Avg loss: 0.37270596, Global Avg Loss: 3.21639497, Time: 0.0211 Steps: 12170, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001535, Sample Num: 24560, Cur Loss: 0.25808305, Cur Avg Loss: 0.58359463, Log Avg loss: 0.30148165, Global Avg Loss: 3.21400177, Time: 0.0211 Steps: 12180, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001545, Sample Num: 24720, Cur Loss: 0.66900009, Cur Avg Loss: 0.58316722, Log Avg loss: 0.51756053, Global Avg Loss: 3.21178976, Time: 0.0212 Steps: 12190, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001555, Sample Num: 24880, Cur Loss: 0.55100578, Cur Avg Loss: 0.58331577, Log Avg loss: 0.60626580, Global Avg Loss: 3.20965408, Time: 0.0211 Steps: 12200, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001565, Sample Num: 25040, Cur Loss: 0.22971752, Cur Avg Loss: 0.58166352, Log Avg loss: 0.32473913, Global Avg Loss: 3.20729134, Time: 0.0210 Steps: 12210, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001575, Sample Num: 25200, Cur Loss: 0.55345857, Cur Avg Loss: 0.58146761, Log Avg loss: 0.55080783, Global Avg Loss: 3.20511745, Time: 0.0211 Steps: 12220, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001585, Sample Num: 25360, Cur Loss: 0.18079534, Cur Avg Loss: 0.57993113, Log Avg loss: 0.33793563, Global Avg Loss: 3.20277307, Time: 0.0211 Steps: 12230, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001595, Sample Num: 25520, Cur Loss: 0.82522428, Cur Avg Loss: 0.57945140, Log Avg loss: 0.50341462, Global Avg Loss: 3.20056771, Time: 0.0211 Steps: 12240, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001605, Sample Num: 25680, Cur Loss: 0.38970983, Cur Avg Loss: 0.57828799, Log Avg loss: 0.39272366, Global Avg Loss: 3.19827559, Time: 0.0210 Steps: 12250, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001615, Sample Num: 25840, Cur Loss: 1.59205890, Cur Avg Loss: 0.57908004, Log Avg loss: 0.70620345, Global Avg Loss: 3.19624291, Time: 0.0211 Steps: 12260, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001625, Sample Num: 26000, Cur Loss: 0.27801824, Cur Avg Loss: 0.57944880, Log Avg loss: 0.63900379, Global Avg Loss: 3.19415877, Time: 0.0211 Steps: 12270, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001635, Sample Num: 26160, Cur Loss: 0.68000603, Cur Avg Loss: 0.57912829, Log Avg loss: 0.52704577, Global Avg Loss: 3.19198685, Time: 0.0211 Steps: 12280, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001645, Sample Num: 26320, Cur Loss: 0.78338116, Cur Avg Loss: 0.57791802, Log Avg loss: 0.38003902, Global Avg Loss: 3.18969886, Time: 0.0211 Steps: 12290, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001655, Sample Num: 26480, Cur Loss: 0.74549639, Cur Avg Loss: 0.57770388, Log Avg loss: 0.54247688, Global Avg Loss: 3.18754664, Time: 0.0210 Steps: 12300, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001665, Sample Num: 26640, Cur Loss: 0.34245056, Cur Avg Loss: 0.57822423, Log Avg loss: 0.66434194, Global Avg Loss: 3.18549692, Time: 0.0211 Steps: 12310, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001675, Sample Num: 26800, Cur Loss: 1.03875113, Cur Avg Loss: 0.57778228, Log Avg loss: 0.50419870, Global Avg Loss: 3.18332055, Time: 0.0211 Steps: 12320, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001685, Sample Num: 26960, Cur Loss: 0.12409135, Cur Avg Loss: 0.57789292, Log Avg loss: 0.59642509, Global Avg Loss: 3.18122250, Time: 0.0211 Steps: 12330, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001695, Sample Num: 27120, Cur Loss: 0.28264087, Cur Avg Loss: 0.57958832, Log Avg loss: 0.86526296, Global Avg Loss: 3.17934571, Time: 0.0211 Steps: 12340, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001705, Sample Num: 27280, Cur Loss: 0.43291998, Cur Avg Loss: 0.58013100, Log Avg loss: 0.67211532, Global Avg Loss: 3.17731556, Time: 0.0211 Steps: 12350, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001715, Sample Num: 27440, Cur Loss: 0.30972809, Cur Avg Loss: 0.57972838, Log Avg loss: 0.51108184, Global Avg Loss: 3.17515841, Time: 0.0211 Steps: 12360, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001725, Sample Num: 27600, Cur Loss: 0.52576607, Cur Avg Loss: 0.57960863, Log Avg loss: 0.55907053, Global Avg Loss: 3.17304355, Time: 0.0211 Steps: 12370, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001735, Sample Num: 27760, Cur Loss: 0.68089998, Cur Avg Loss: 0.57918142, Log Avg loss: 0.50548816, Global Avg Loss: 3.17088882, Time: 0.0211 Steps: 12380, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001745, Sample Num: 27920, Cur Loss: 0.93270999, Cur Avg Loss: 0.57973138, Log Avg loss: 0.67514890, Global Avg Loss: 3.16887450, Time: 0.0211 Steps: 12390, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001755, Sample Num: 28080, Cur Loss: 0.17929418, Cur Avg Loss: 0.57908738, Log Avg loss: 0.46671040, Global Avg Loss: 3.16669533, Time: 0.0210 Steps: 12400, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001765, Sample Num: 28240, Cur Loss: 0.44914055, Cur Avg Loss: 0.57943378, Log Avg loss: 0.64022751, Global Avg Loss: 3.16465950, Time: 0.0211 Steps: 12410, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001775, Sample Num: 28400, Cur Loss: 0.40821379, Cur Avg Loss: 0.57967794, Log Avg loss: 0.62277090, Global Avg Loss: 3.16261289, Time: 0.0211 Steps: 12420, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001785, Sample Num: 28560, Cur Loss: 0.16149926, Cur Avg Loss: 0.57900163, Log Avg loss: 0.45895726, Global Avg Loss: 3.16043779, Time: 0.0211 Steps: 12430, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001795, Sample Num: 28720, Cur Loss: 0.43858680, Cur Avg Loss: 0.57854475, Log Avg loss: 0.49699070, Global Avg Loss: 3.15829675, Time: 0.0249 Steps: 12440, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001805, Sample Num: 28880, Cur Loss: 0.65120530, Cur Avg Loss: 0.57826941, Log Avg loss: 0.52884727, Global Avg Loss: 3.15618475, Time: 0.0213 Steps: 12450, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001815, Sample Num: 29040, Cur Loss: 0.17661232, Cur Avg Loss: 0.57739642, Log Avg loss: 0.41982035, Global Avg Loss: 3.15398863, Time: 0.0212 Steps: 12460, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001825, Sample Num: 29200, Cur Loss: 0.82691443, Cur Avg Loss: 0.57716105, Log Avg loss: 0.53444188, Global Avg Loss: 3.15188795, Time: 0.0212 Steps: 12470, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001835, Sample Num: 29360, Cur Loss: 0.55502188, Cur Avg Loss: 0.57796747, Log Avg loss: 0.72513874, Global Avg Loss: 3.14994344, Time: 0.0212 Steps: 12480, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001845, Sample Num: 29520, Cur Loss: 1.49392557, Cur Avg Loss: 0.58351259, Log Avg loss: 1.60104222, Global Avg Loss: 3.14870332, Time: 0.0212 Steps: 12490, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001855, Sample Num: 29680, Cur Loss: 0.69374120, Cur Avg Loss: 0.58322709, Log Avg loss: 0.53055302, Global Avg Loss: 3.14660880, Time: 0.0212 Steps: 12500, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001865, Sample Num: 29840, Cur Loss: 1.36764681, Cur Avg Loss: 0.58336830, Log Avg loss: 0.60956265, Global Avg Loss: 3.14458079, Time: 0.0212 Steps: 12510, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001875, Sample Num: 30000, Cur Loss: 0.62019539, Cur Avg Loss: 0.58505154, Log Avg loss: 0.89897586, Global Avg Loss: 3.14278718, Time: 0.0212 Steps: 12520, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001885, Sample Num: 30160, Cur Loss: 0.65478384, Cur Avg Loss: 0.58515753, Log Avg loss: 0.60503065, Global Avg Loss: 3.14076183, Time: 0.0212 Steps: 12530, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001895, Sample Num: 30320, Cur Loss: 0.50010991, Cur Avg Loss: 0.58543310, Log Avg loss: 0.63737735, Global Avg Loss: 3.13876551, Time: 0.0212 Steps: 12540, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001905, Sample Num: 30480, Cur Loss: 0.49148351, Cur Avg Loss: 0.58450680, Log Avg loss: 0.40897302, Global Avg Loss: 3.13659038, Time: 0.0212 Steps: 12550, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001915, Sample Num: 30640, Cur Loss: 0.66175991, Cur Avg Loss: 0.58542719, Log Avg loss: 0.76076276, Global Avg Loss: 3.13469880, Time: 0.0212 Steps: 12560, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001925, Sample Num: 30800, Cur Loss: 0.48569959, Cur Avg Loss: 0.58651807, Log Avg loss: 0.79542052, Global Avg Loss: 3.13283779, Time: 0.0212 Steps: 12570, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001935, Sample Num: 30960, Cur Loss: 0.75702345, Cur Avg Loss: 0.58691695, Log Avg loss: 0.66370101, Global Avg Loss: 3.13087505, Time: 0.0212 Steps: 12580, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001945, Sample Num: 31120, Cur Loss: 0.53597307, Cur Avg Loss: 0.58663105, Log Avg loss: 0.53130975, Global Avg Loss: 3.12881026, Time: 0.0212 Steps: 12590, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001955, Sample Num: 31280, Cur Loss: 0.48392078, Cur Avg Loss: 0.58554173, Log Avg loss: 0.37366926, Global Avg Loss: 3.12662364, Time: 0.0212 Steps: 12600, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001965, Sample Num: 31440, Cur Loss: 0.45711735, Cur Avg Loss: 0.58538322, Log Avg loss: 0.55439535, Global Avg Loss: 3.12458381, Time: 0.0211 Steps: 12610, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001975, Sample Num: 31600, Cur Loss: 0.13643591, Cur Avg Loss: 0.58463707, Log Avg loss: 0.43801742, Global Avg Loss: 3.12245499, Time: 0.0212 Steps: 12620, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001985, Sample Num: 31760, Cur Loss: 0.55711949, Cur Avg Loss: 0.58474770, Log Avg loss: 0.60659747, Global Avg Loss: 3.12046302, Time: 0.0212 Steps: 12630, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001995, Sample Num: 31920, Cur Loss: 0.49252844, Cur Avg Loss: 0.58490525, Log Avg loss: 0.61617973, Global Avg Loss: 3.11848179, Time: 0.0212 Steps: 12640, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002005, Sample Num: 32080, Cur Loss: 0.77413547, Cur Avg Loss: 0.58469886, Log Avg loss: 0.54352415, Global Avg Loss: 3.11644625, Time: 0.0212 Steps: 12650, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002015, Sample Num: 32240, Cur Loss: 1.14436483, Cur Avg Loss: 0.58563988, Log Avg loss: 0.77431293, Global Avg Loss: 3.11459622, Time: 0.0212 Steps: 12660, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002025, Sample Num: 32400, Cur Loss: 0.27012950, Cur Avg Loss: 0.58648589, Log Avg loss: 0.75695758, Global Avg Loss: 3.11273542, Time: 0.0212 Steps: 12670, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002035, Sample Num: 32560, Cur Loss: 0.18299344, Cur Avg Loss: 0.58601576, Log Avg loss: 0.49081434, Global Avg Loss: 3.11066766, Time: 0.0212 Steps: 12680, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002045, Sample Num: 32720, Cur Loss: 0.15955542, Cur Avg Loss: 0.58559275, Log Avg loss: 0.49951097, Global Avg Loss: 3.10861001, Time: 0.0212 Steps: 12690, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002055, Sample Num: 32880, Cur Loss: 0.93002856, Cur Avg Loss: 0.58783573, Log Avg loss: 1.04652438, Global Avg Loss: 3.10698632, Time: 0.0211 Steps: 12700, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002065, Sample Num: 33040, Cur Loss: 0.10225828, Cur Avg Loss: 0.58648419, Log Avg loss: 0.30874272, Global Avg Loss: 3.10478471, Time: 0.0211 Steps: 12710, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002075, Sample Num: 33200, Cur Loss: 0.26740146, Cur Avg Loss: 0.58584279, Log Avg loss: 0.45339344, Global Avg Loss: 3.10270028, Time: 0.0211 Steps: 12720, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002085, Sample Num: 33360, Cur Loss: 1.56108475, Cur Avg Loss: 0.58620339, Log Avg loss: 0.66102764, Global Avg Loss: 3.10078224, Time: 0.0211 Steps: 12730, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002095, Sample Num: 33520, Cur Loss: 0.28284979, Cur Avg Loss: 0.58592643, Log Avg loss: 0.52818014, Global Avg Loss: 3.09876292, Time: 0.0211 Steps: 12740, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002105, Sample Num: 33680, Cur Loss: 0.48271683, Cur Avg Loss: 0.58503012, Log Avg loss: 0.39725371, Global Avg Loss: 3.09664409, Time: 0.0211 Steps: 12750, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002115, Sample Num: 33840, Cur Loss: 0.41036272, Cur Avg Loss: 0.58417404, Log Avg loss: 0.40397012, Global Avg Loss: 3.09453385, Time: 0.0211 Steps: 12760, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002125, Sample Num: 34000, Cur Loss: 0.48783416, Cur Avg Loss: 0.58417103, Log Avg loss: 0.58353291, Global Avg Loss: 3.09256752, Time: 0.0211 Steps: 12770, Updated lr: 0.000089 ***** Running evaluation checkpoint-12774 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-12774 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.173042, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.456811, "eval_total_loss": 321.137989, "eval_mae": 0.504709, "eval_mse": 0.456951, "eval_r2": 0.709532, "eval_sp_statistic": 0.803632, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.847471, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.574651, "test_total_loss": 288.474806, "test_mae": 0.645761, "test_mse": 0.574812, "test_r2": 0.629011, "test_sp_statistic": 0.80488, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.865954, "test_ps_pvalue": 0.0, "lr": 8.883451872925558e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.0917292960312994, "train_cur_epoch_loss": 1243.026228748262, "train_cur_epoch_avg_loss": 0.5838544991772014, "train_cur_epoch_time": 45.173041582107544, "train_cur_epoch_avg_time": 0.021217962227387293, "epoch": 6, "step": 12774} ################################################## Training, Epoch: 0007, Batch: 000006, Sample Num: 96, Cur Loss: 0.32655135, Cur Avg Loss: 0.31424051, Log Avg loss: 0.35482408, Global Avg Loss: 3.09042531, Time: 0.0213 Steps: 12780, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000016, Sample Num: 256, Cur Loss: 0.86476588, Cur Avg Loss: 0.43319131, Log Avg loss: 0.50456180, Global Avg Loss: 3.08840353, Time: 0.0213 Steps: 12790, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000026, Sample Num: 416, Cur Loss: 0.32350957, Cur Avg Loss: 0.47323052, Log Avg loss: 0.53729325, Global Avg Loss: 3.08641047, Time: 0.0211 Steps: 12800, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000036, Sample Num: 576, Cur Loss: 0.52182573, Cur Avg Loss: 0.43027748, Log Avg loss: 0.31859958, Global Avg Loss: 3.08424981, Time: 0.0212 Steps: 12810, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000046, Sample Num: 736, Cur Loss: 0.93011504, Cur Avg Loss: 0.46342525, Log Avg loss: 0.58275722, Global Avg Loss: 3.08229856, Time: 0.0211 Steps: 12820, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000056, Sample Num: 896, Cur Loss: 1.16828954, Cur Avg Loss: 0.49213963, Log Avg loss: 0.62422579, Global Avg Loss: 3.08038268, Time: 0.0211 Steps: 12830, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000066, Sample Num: 1056, Cur Loss: 0.75093186, Cur Avg Loss: 0.49947343, Log Avg loss: 0.54054273, Global Avg Loss: 3.07840462, Time: 0.0211 Steps: 12840, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000076, Sample Num: 1216, Cur Loss: 0.86282778, Cur Avg Loss: 0.51667645, Log Avg loss: 0.63021638, Global Avg Loss: 3.07649941, Time: 0.0212 Steps: 12850, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000086, Sample Num: 1376, Cur Loss: 0.59958059, Cur Avg Loss: 0.53486234, Log Avg loss: 0.67307504, Global Avg Loss: 3.07463050, Time: 0.0211 Steps: 12860, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000096, Sample Num: 1536, Cur Loss: 0.36951768, Cur Avg Loss: 0.54934113, Log Avg loss: 0.67385876, Global Avg Loss: 3.07276510, Time: 0.0211 Steps: 12870, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000106, Sample Num: 1696, Cur Loss: 0.30854389, Cur Avg Loss: 0.56304818, Log Avg loss: 0.69463588, Global Avg Loss: 3.07091872, Time: 0.0212 Steps: 12880, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000116, Sample Num: 1856, Cur Loss: 0.55289710, Cur Avg Loss: 0.57609875, Log Avg loss: 0.71443481, Global Avg Loss: 3.06909057, Time: 0.0211 Steps: 12890, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000126, Sample Num: 2016, Cur Loss: 0.83296347, Cur Avg Loss: 0.57019022, Log Avg loss: 0.50165127, Global Avg Loss: 3.06710031, Time: 0.0212 Steps: 12900, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000136, Sample Num: 2176, Cur Loss: 0.67869627, Cur Avg Loss: 0.56291543, Log Avg loss: 0.47125301, Global Avg Loss: 3.06508958, Time: 0.0212 Steps: 12910, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000146, Sample Num: 2336, Cur Loss: 0.63693494, Cur Avg Loss: 0.55753898, Log Avg loss: 0.48441924, Global Avg Loss: 3.06309216, Time: 0.0211 Steps: 12920, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000156, Sample Num: 2496, Cur Loss: 0.25613442, Cur Avg Loss: 0.55495774, Log Avg loss: 0.51727174, Global Avg Loss: 3.06112324, Time: 0.0211 Steps: 12930, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000166, Sample Num: 2656, Cur Loss: 0.12676120, Cur Avg Loss: 0.54276121, Log Avg loss: 0.35249533, Global Avg Loss: 3.05903001, Time: 0.0212 Steps: 12940, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000176, Sample Num: 2816, Cur Loss: 0.91429234, Cur Avg Loss: 0.54074665, Log Avg loss: 0.50730494, Global Avg Loss: 3.05705957, Time: 0.0212 Steps: 12950, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000186, Sample Num: 2976, Cur Loss: 0.63636076, Cur Avg Loss: 0.54117515, Log Avg loss: 0.54871666, Global Avg Loss: 3.05512412, Time: 0.0212 Steps: 12960, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000196, Sample Num: 3136, Cur Loss: 0.46026039, Cur Avg Loss: 0.54822391, Log Avg loss: 0.67933091, Global Avg Loss: 3.05329236, Time: 0.0211 Steps: 12970, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000206, Sample Num: 3296, Cur Loss: 0.68456024, Cur Avg Loss: 0.54877493, Log Avg loss: 0.55957498, Global Avg Loss: 3.05137116, Time: 0.0211 Steps: 12980, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000216, Sample Num: 3456, Cur Loss: 1.30809069, Cur Avg Loss: 0.54933113, Log Avg loss: 0.56078870, Global Avg Loss: 3.04945385, Time: 0.0212 Steps: 12990, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000226, Sample Num: 3616, Cur Loss: 0.17813590, Cur Avg Loss: 0.54683117, Log Avg loss: 0.49283212, Global Avg Loss: 3.04748722, Time: 0.0211 Steps: 13000, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000236, Sample Num: 3776, Cur Loss: 0.32423860, Cur Avg Loss: 0.54207875, Log Avg loss: 0.43467414, Global Avg Loss: 3.04547891, Time: 0.0211 Steps: 13010, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000246, Sample Num: 3936, Cur Loss: 0.84568971, Cur Avg Loss: 0.54317484, Log Avg loss: 0.56904257, Global Avg Loss: 3.04357688, Time: 0.0211 Steps: 13020, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000256, Sample Num: 4096, Cur Loss: 1.40608323, Cur Avg Loss: 0.54638548, Log Avg loss: 0.62536704, Global Avg Loss: 3.04172101, Time: 0.0256 Steps: 13030, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000266, Sample Num: 4256, Cur Loss: 0.30808306, Cur Avg Loss: 0.54447436, Log Avg loss: 0.49554987, Global Avg Loss: 3.03976842, Time: 0.0213 Steps: 13040, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000276, Sample Num: 4416, Cur Loss: 0.21382117, Cur Avg Loss: 0.54244190, Log Avg loss: 0.48837844, Global Avg Loss: 3.03781333, Time: 0.0213 Steps: 13050, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000286, Sample Num: 4576, Cur Loss: 0.19350582, Cur Avg Loss: 0.53904104, Log Avg loss: 0.44517720, Global Avg Loss: 3.03582816, Time: 0.0213 Steps: 13060, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000296, Sample Num: 4736, Cur Loss: 0.65056121, Cur Avg Loss: 0.54383418, Log Avg loss: 0.68091796, Global Avg Loss: 3.03402639, Time: 0.0213 Steps: 13070, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000306, Sample Num: 4896, Cur Loss: 0.29251891, Cur Avg Loss: 0.53879548, Log Avg loss: 0.38965006, Global Avg Loss: 3.03200470, Time: 0.0213 Steps: 13080, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000316, Sample Num: 5056, Cur Loss: 0.27733386, Cur Avg Loss: 0.53680894, Log Avg loss: 0.47602071, Global Avg Loss: 3.03005207, Time: 0.0213 Steps: 13090, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000326, Sample Num: 5216, Cur Loss: 0.25148180, Cur Avg Loss: 0.53401031, Log Avg loss: 0.44557358, Global Avg Loss: 3.02807919, Time: 0.0213 Steps: 13100, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000336, Sample Num: 5376, Cur Loss: 0.55265385, Cur Avg Loss: 0.53132899, Log Avg loss: 0.44391805, Global Avg Loss: 3.02610805, Time: 0.0213 Steps: 13110, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000346, Sample Num: 5536, Cur Loss: 0.38283861, Cur Avg Loss: 0.52850221, Log Avg loss: 0.43352235, Global Avg Loss: 3.02413200, Time: 0.0213 Steps: 13120, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000356, Sample Num: 5696, Cur Loss: 0.34471178, Cur Avg Loss: 0.52694364, Log Avg loss: 0.47301730, Global Avg Loss: 3.02218903, Time: 0.0213 Steps: 13130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000366, Sample Num: 5856, Cur Loss: 0.61485595, Cur Avg Loss: 0.52773681, Log Avg loss: 0.55597346, Global Avg Loss: 3.02031215, Time: 0.0213 Steps: 13140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000376, Sample Num: 6016, Cur Loss: 0.38921082, Cur Avg Loss: 0.52852979, Log Avg loss: 0.55755295, Global Avg Loss: 3.01843933, Time: 0.0213 Steps: 13150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000386, Sample Num: 6176, Cur Loss: 0.37904540, Cur Avg Loss: 0.52624158, Log Avg loss: 0.44020499, Global Avg Loss: 3.01648019, Time: 0.0213 Steps: 13160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000396, Sample Num: 6336, Cur Loss: 0.45840672, Cur Avg Loss: 0.52769071, Log Avg loss: 0.58362697, Global Avg Loss: 3.01463292, Time: 0.0210 Steps: 13170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000406, Sample Num: 6496, Cur Loss: 0.38517579, Cur Avg Loss: 0.52730426, Log Avg loss: 0.51200080, Global Avg Loss: 3.01273411, Time: 0.0211 Steps: 13180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000416, Sample Num: 6656, Cur Loss: 0.15554933, Cur Avg Loss: 0.52479676, Log Avg loss: 0.42299233, Global Avg Loss: 3.01077070, Time: 0.0211 Steps: 13190, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000426, Sample Num: 6816, Cur Loss: 0.52129394, Cur Avg Loss: 0.52742368, Log Avg loss: 0.63670377, Global Avg Loss: 3.00897216, Time: 0.0210 Steps: 13200, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000436, Sample Num: 6976, Cur Loss: 0.54418683, Cur Avg Loss: 0.52678568, Log Avg loss: 0.49960649, Global Avg Loss: 3.00707256, Time: 0.0211 Steps: 13210, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000446, Sample Num: 7136, Cur Loss: 0.59167635, Cur Avg Loss: 0.52625149, Log Avg loss: 0.50296115, Global Avg Loss: 3.00517838, Time: 0.0210 Steps: 13220, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000456, Sample Num: 7296, Cur Loss: 0.30450189, Cur Avg Loss: 0.52593945, Log Avg loss: 0.51202213, Global Avg Loss: 3.00329391, Time: 0.0210 Steps: 13230, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000466, Sample Num: 7456, Cur Loss: 0.61256659, Cur Avg Loss: 0.53074235, Log Avg loss: 0.74975471, Global Avg Loss: 3.00159184, Time: 0.0210 Steps: 13240, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000476, Sample Num: 7616, Cur Loss: 0.29367495, Cur Avg Loss: 0.52762604, Log Avg loss: 0.38240618, Global Avg Loss: 2.99961510, Time: 0.0210 Steps: 13250, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000486, Sample Num: 7776, Cur Loss: 0.98476678, Cur Avg Loss: 0.52888400, Log Avg loss: 0.58876281, Global Avg Loss: 2.99779696, Time: 0.0211 Steps: 13260, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000496, Sample Num: 7936, Cur Loss: 0.29138392, Cur Avg Loss: 0.53014386, Log Avg loss: 0.59137288, Global Avg Loss: 2.99598353, Time: 0.0210 Steps: 13270, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000506, Sample Num: 8096, Cur Loss: 0.28084081, Cur Avg Loss: 0.52970931, Log Avg loss: 0.50815555, Global Avg Loss: 2.99411016, Time: 0.0210 Steps: 13280, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000516, Sample Num: 8256, Cur Loss: 0.36574531, Cur Avg Loss: 0.53715232, Log Avg loss: 0.91376909, Global Avg Loss: 2.99254482, Time: 0.0248 Steps: 13290, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000526, Sample Num: 8416, Cur Loss: 0.45598692, Cur Avg Loss: 0.53712877, Log Avg loss: 0.53591340, Global Avg Loss: 2.99069773, Time: 0.0210 Steps: 13300, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000536, Sample Num: 8576, Cur Loss: 0.60089624, Cur Avg Loss: 0.53921435, Log Avg loss: 0.64891603, Global Avg Loss: 2.98893831, Time: 0.0210 Steps: 13310, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000546, Sample Num: 8736, Cur Loss: 1.30604386, Cur Avg Loss: 0.53965193, Log Avg loss: 0.56310576, Global Avg Loss: 2.98711712, Time: 0.0210 Steps: 13320, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000556, Sample Num: 8896, Cur Loss: 0.49839154, Cur Avg Loss: 0.53875964, Log Avg loss: 0.49004076, Global Avg Loss: 2.98524384, Time: 0.0210 Steps: 13330, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000566, Sample Num: 9056, Cur Loss: 0.72310686, Cur Avg Loss: 0.53846498, Log Avg loss: 0.52208177, Global Avg Loss: 2.98339739, Time: 0.0210 Steps: 13340, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000576, Sample Num: 9216, Cur Loss: 0.28714499, Cur Avg Loss: 0.53874707, Log Avg loss: 0.55471352, Global Avg Loss: 2.98157815, Time: 0.0210 Steps: 13350, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000586, Sample Num: 9376, Cur Loss: 0.36238045, Cur Avg Loss: 0.54099891, Log Avg loss: 0.67070497, Global Avg Loss: 2.97984846, Time: 0.0211 Steps: 13360, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000596, Sample Num: 9536, Cur Loss: 0.28630573, Cur Avg Loss: 0.53748669, Log Avg loss: 0.33167032, Global Avg Loss: 2.97786777, Time: 0.0210 Steps: 13370, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000606, Sample Num: 9696, Cur Loss: 0.39736700, Cur Avg Loss: 0.53883688, Log Avg loss: 0.61930862, Global Avg Loss: 2.97610502, Time: 0.0210 Steps: 13380, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000616, Sample Num: 9856, Cur Loss: 0.73599744, Cur Avg Loss: 0.53766934, Log Avg loss: 0.46691629, Global Avg Loss: 2.97423109, Time: 0.0210 Steps: 13390, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000626, Sample Num: 10016, Cur Loss: 0.26889449, Cur Avg Loss: 0.53808544, Log Avg loss: 0.56371742, Global Avg Loss: 2.97243220, Time: 0.0210 Steps: 13400, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000636, Sample Num: 10176, Cur Loss: 1.03536022, Cur Avg Loss: 0.54063637, Log Avg loss: 0.70032461, Global Avg Loss: 2.97073786, Time: 0.0210 Steps: 13410, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000646, Sample Num: 10336, Cur Loss: 0.25238353, Cur Avg Loss: 0.54016979, Log Avg loss: 0.51049519, Global Avg Loss: 2.96890460, Time: 0.0211 Steps: 13420, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000656, Sample Num: 10496, Cur Loss: 0.46315983, Cur Avg Loss: 0.53928613, Log Avg loss: 0.48220171, Global Avg Loss: 2.96705300, Time: 0.0210 Steps: 13430, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000666, Sample Num: 10656, Cur Loss: 0.55436879, Cur Avg Loss: 0.53538025, Log Avg loss: 0.27915429, Global Avg Loss: 2.96505307, Time: 0.0211 Steps: 13440, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000676, Sample Num: 10816, Cur Loss: 0.33499530, Cur Avg Loss: 0.53472704, Log Avg loss: 0.49122342, Global Avg Loss: 2.96321379, Time: 0.0210 Steps: 13450, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000686, Sample Num: 10976, Cur Loss: 1.36627543, Cur Avg Loss: 0.53483961, Log Avg loss: 0.54244908, Global Avg Loss: 2.96141530, Time: 0.0210 Steps: 13460, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000696, Sample Num: 11136, Cur Loss: 0.28749135, Cur Avg Loss: 0.53372803, Log Avg loss: 0.45747359, Global Avg Loss: 2.95955640, Time: 0.0210 Steps: 13470, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000706, Sample Num: 11296, Cur Loss: 0.21711789, Cur Avg Loss: 0.53343432, Log Avg loss: 0.51299200, Global Avg Loss: 2.95774144, Time: 0.0210 Steps: 13480, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000716, Sample Num: 11456, Cur Loss: 0.26357716, Cur Avg Loss: 0.53406304, Log Avg loss: 0.57845084, Global Avg Loss: 2.95597770, Time: 0.0211 Steps: 13490, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000726, Sample Num: 11616, Cur Loss: 1.05096817, Cur Avg Loss: 0.53446655, Log Avg loss: 0.56335777, Global Avg Loss: 2.95420539, Time: 0.0212 Steps: 13500, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000736, Sample Num: 11776, Cur Loss: 0.68473786, Cur Avg Loss: 0.53373972, Log Avg loss: 0.48097241, Global Avg Loss: 2.95237472, Time: 0.0212 Steps: 13510, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000746, Sample Num: 11936, Cur Loss: 0.56079316, Cur Avg Loss: 0.53131113, Log Avg loss: 0.35256684, Global Avg Loss: 2.95045178, Time: 0.0212 Steps: 13520, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000756, Sample Num: 12096, Cur Loss: 0.83113050, Cur Avg Loss: 0.53362753, Log Avg loss: 0.70643058, Global Avg Loss: 2.94879323, Time: 0.0212 Steps: 13530, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000766, Sample Num: 12256, Cur Loss: 0.23704924, Cur Avg Loss: 0.53144137, Log Avg loss: 0.36616782, Global Avg Loss: 2.94688583, Time: 0.0212 Steps: 13540, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000776, Sample Num: 12416, Cur Loss: 0.20898339, Cur Avg Loss: 0.53014436, Log Avg loss: 0.43079337, Global Avg Loss: 2.94502893, Time: 0.0210 Steps: 13550, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000786, Sample Num: 12576, Cur Loss: 0.14793204, Cur Avg Loss: 0.52784213, Log Avg loss: 0.34918917, Global Avg Loss: 2.94311460, Time: 0.0211 Steps: 13560, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000796, Sample Num: 12736, Cur Loss: 0.66926241, Cur Avg Loss: 0.52748583, Log Avg loss: 0.49948022, Global Avg Loss: 2.94131384, Time: 0.0211 Steps: 13570, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000806, Sample Num: 12896, Cur Loss: 0.27948612, Cur Avg Loss: 0.52720472, Log Avg loss: 0.50482852, Global Avg Loss: 2.93951966, Time: 0.0210 Steps: 13580, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000816, Sample Num: 13056, Cur Loss: 1.00182724, Cur Avg Loss: 0.52619175, Log Avg loss: 0.44454675, Global Avg Loss: 2.93768377, Time: 0.0210 Steps: 13590, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000826, Sample Num: 13216, Cur Loss: 1.13634896, Cur Avg Loss: 0.53063554, Log Avg loss: 0.89324888, Global Avg Loss: 2.93618051, Time: 0.0210 Steps: 13600, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000836, Sample Num: 13376, Cur Loss: 1.21633101, Cur Avg Loss: 0.53392429, Log Avg loss: 0.80557512, Global Avg Loss: 2.93461504, Time: 0.0210 Steps: 13610, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000846, Sample Num: 13536, Cur Loss: 0.63359213, Cur Avg Loss: 0.53703921, Log Avg loss: 0.79744588, Global Avg Loss: 2.93304590, Time: 0.0210 Steps: 13620, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000856, Sample Num: 13696, Cur Loss: 0.30607444, Cur Avg Loss: 0.53532879, Log Avg loss: 0.39062750, Global Avg Loss: 2.93118059, Time: 0.0210 Steps: 13630, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000866, Sample Num: 13856, Cur Loss: 0.44607311, Cur Avg Loss: 0.53520813, Log Avg loss: 0.52487987, Global Avg Loss: 2.92941644, Time: 0.0210 Steps: 13640, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000876, Sample Num: 14016, Cur Loss: 0.65450877, Cur Avg Loss: 0.53555480, Log Avg loss: 0.56557612, Global Avg Loss: 2.92768469, Time: 0.0210 Steps: 13650, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000886, Sample Num: 14176, Cur Loss: 0.41090679, Cur Avg Loss: 0.53582676, Log Avg loss: 0.55965080, Global Avg Loss: 2.92595114, Time: 0.0210 Steps: 13660, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000896, Sample Num: 14336, Cur Loss: 0.26749790, Cur Avg Loss: 0.53671648, Log Avg loss: 0.61554574, Global Avg Loss: 2.92426101, Time: 0.0210 Steps: 13670, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000906, Sample Num: 14496, Cur Loss: 0.24710302, Cur Avg Loss: 0.53562642, Log Avg loss: 0.43795717, Global Avg Loss: 2.92244354, Time: 0.0210 Steps: 13680, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000916, Sample Num: 14656, Cur Loss: 0.10338791, Cur Avg Loss: 0.53402156, Log Avg loss: 0.38862107, Global Avg Loss: 2.92059268, Time: 0.0210 Steps: 13690, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000926, Sample Num: 14816, Cur Loss: 0.36729762, Cur Avg Loss: 0.53217302, Log Avg loss: 0.36284668, Global Avg Loss: 2.91872571, Time: 0.0210 Steps: 13700, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000936, Sample Num: 14976, Cur Loss: 0.29043612, Cur Avg Loss: 0.53317377, Log Avg loss: 0.62584275, Global Avg Loss: 2.91705329, Time: 0.0210 Steps: 13710, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000946, Sample Num: 15136, Cur Loss: 0.65055311, Cur Avg Loss: 0.53468866, Log Avg loss: 0.67648274, Global Avg Loss: 2.91542023, Time: 0.0210 Steps: 13720, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000956, Sample Num: 15296, Cur Loss: 0.34020767, Cur Avg Loss: 0.53600051, Log Avg loss: 0.66010160, Global Avg Loss: 2.91377760, Time: 0.0211 Steps: 13730, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000966, Sample Num: 15456, Cur Loss: 0.33436048, Cur Avg Loss: 0.53647608, Log Avg loss: 0.58194045, Global Avg Loss: 2.91208049, Time: 0.0210 Steps: 13740, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000976, Sample Num: 15616, Cur Loss: 0.43254143, Cur Avg Loss: 0.53741506, Log Avg loss: 0.62812074, Global Avg Loss: 2.91041943, Time: 0.0210 Steps: 13750, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000986, Sample Num: 15776, Cur Loss: 0.23908767, Cur Avg Loss: 0.53730423, Log Avg loss: 0.52648728, Global Avg Loss: 2.90868692, Time: 0.0210 Steps: 13760, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000996, Sample Num: 15936, Cur Loss: 0.82925498, Cur Avg Loss: 0.54040252, Log Avg loss: 0.84589371, Global Avg Loss: 2.90718888, Time: 0.0210 Steps: 13770, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001006, Sample Num: 16096, Cur Loss: 0.75629658, Cur Avg Loss: 0.54017530, Log Avg loss: 0.51754413, Global Avg Loss: 2.90545474, Time: 0.0210 Steps: 13780, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001016, Sample Num: 16256, Cur Loss: 1.31404483, Cur Avg Loss: 0.54037266, Log Avg loss: 0.56022672, Global Avg Loss: 2.90375407, Time: 0.0211 Steps: 13790, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001026, Sample Num: 16416, Cur Loss: 0.80938447, Cur Avg Loss: 0.54035177, Log Avg loss: 0.53822943, Global Avg Loss: 2.90203992, Time: 0.0249 Steps: 13800, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001036, Sample Num: 16576, Cur Loss: 0.30317891, Cur Avg Loss: 0.53917740, Log Avg loss: 0.41868692, Global Avg Loss: 2.90024170, Time: 0.0212 Steps: 13810, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001046, Sample Num: 16736, Cur Loss: 0.34826672, Cur Avg Loss: 0.53768748, Log Avg loss: 0.38333163, Global Avg Loss: 2.89842049, Time: 0.0212 Steps: 13820, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001056, Sample Num: 16896, Cur Loss: 0.28879285, Cur Avg Loss: 0.53707681, Log Avg loss: 0.47320099, Global Avg Loss: 2.89666689, Time: 0.0212 Steps: 13830, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001066, Sample Num: 17056, Cur Loss: 0.47614336, Cur Avg Loss: 0.53573095, Log Avg loss: 0.39360836, Global Avg Loss: 2.89485833, Time: 0.0212 Steps: 13840, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001076, Sample Num: 17216, Cur Loss: 0.50010401, Cur Avg Loss: 0.53441228, Log Avg loss: 0.39384190, Global Avg Loss: 2.89305254, Time: 0.0211 Steps: 13850, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001086, Sample Num: 17376, Cur Loss: 0.21650049, Cur Avg Loss: 0.53381845, Log Avg loss: 0.46992198, Global Avg Loss: 2.89130425, Time: 0.0211 Steps: 13860, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001096, Sample Num: 17536, Cur Loss: 0.26222801, Cur Avg Loss: 0.53345303, Log Avg loss: 0.49376912, Global Avg Loss: 2.88957567, Time: 0.0212 Steps: 13870, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001106, Sample Num: 17696, Cur Loss: 0.40354112, Cur Avg Loss: 0.53226145, Log Avg loss: 0.40166401, Global Avg Loss: 2.88778323, Time: 0.0211 Steps: 13880, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001116, Sample Num: 17856, Cur Loss: 0.47687238, Cur Avg Loss: 0.53146978, Log Avg loss: 0.44391127, Global Avg Loss: 2.88602378, Time: 0.0211 Steps: 13890, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001126, Sample Num: 18016, Cur Loss: 0.49995229, Cur Avg Loss: 0.53124558, Log Avg loss: 0.50622419, Global Avg Loss: 2.88431169, Time: 0.0211 Steps: 13900, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001136, Sample Num: 18176, Cur Loss: 1.22028136, Cur Avg Loss: 0.53157803, Log Avg loss: 0.56901194, Global Avg Loss: 2.88264721, Time: 0.0211 Steps: 13910, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001146, Sample Num: 18336, Cur Loss: 0.29548666, Cur Avg Loss: 0.53092893, Log Avg loss: 0.45719184, Global Avg Loss: 2.88090478, Time: 0.0212 Steps: 13920, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001156, Sample Num: 18496, Cur Loss: 0.41101563, Cur Avg Loss: 0.53130372, Log Avg loss: 0.57425391, Global Avg Loss: 2.87924890, Time: 0.0212 Steps: 13930, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001166, Sample Num: 18656, Cur Loss: 0.41389382, Cur Avg Loss: 0.53109278, Log Avg loss: 0.50670832, Global Avg Loss: 2.87754693, Time: 0.0211 Steps: 13940, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001176, Sample Num: 18816, Cur Loss: 0.56691563, Cur Avg Loss: 0.53000794, Log Avg loss: 0.40351636, Global Avg Loss: 2.87577343, Time: 0.0212 Steps: 13950, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001186, Sample Num: 18976, Cur Loss: 0.60066897, Cur Avg Loss: 0.52923648, Log Avg loss: 0.43851275, Global Avg Loss: 2.87402754, Time: 0.0211 Steps: 13960, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001196, Sample Num: 19136, Cur Loss: 1.54157209, Cur Avg Loss: 0.52995266, Log Avg loss: 0.61489140, Global Avg Loss: 2.87241041, Time: 0.0212 Steps: 13970, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001206, Sample Num: 19296, Cur Loss: 0.17493159, Cur Avg Loss: 0.53022698, Log Avg loss: 0.56303591, Global Avg Loss: 2.87075850, Time: 0.0212 Steps: 13980, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001216, Sample Num: 19456, Cur Loss: 0.29936990, Cur Avg Loss: 0.53046310, Log Avg loss: 0.55893924, Global Avg Loss: 2.86910602, Time: 0.0213 Steps: 13990, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001226, Sample Num: 19616, Cur Loss: 0.31210768, Cur Avg Loss: 0.53113102, Log Avg loss: 0.61234935, Global Avg Loss: 2.86749405, Time: 0.0212 Steps: 14000, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001236, Sample Num: 19776, Cur Loss: 0.53891480, Cur Avg Loss: 0.52968393, Log Avg loss: 0.35227074, Global Avg Loss: 2.86569874, Time: 0.0211 Steps: 14010, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001246, Sample Num: 19936, Cur Loss: 0.89831233, Cur Avg Loss: 0.53058933, Log Avg loss: 0.64249699, Global Avg Loss: 2.86411301, Time: 0.0212 Steps: 14020, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001256, Sample Num: 20096, Cur Loss: 0.22972915, Cur Avg Loss: 0.53228725, Log Avg loss: 0.74384798, Global Avg Loss: 2.86260177, Time: 0.0211 Steps: 14030, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001266, Sample Num: 20256, Cur Loss: 0.47986621, Cur Avg Loss: 0.53291211, Log Avg loss: 0.61139507, Global Avg Loss: 2.86099834, Time: 0.0212 Steps: 14040, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001276, Sample Num: 20416, Cur Loss: 0.45007658, Cur Avg Loss: 0.53212066, Log Avg loss: 0.43192206, Global Avg Loss: 2.85926947, Time: 0.0211 Steps: 14050, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001286, Sample Num: 20576, Cur Loss: 0.14870979, Cur Avg Loss: 0.53128469, Log Avg loss: 0.42461614, Global Avg Loss: 2.85753785, Time: 0.0213 Steps: 14060, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001296, Sample Num: 20736, Cur Loss: 0.99219555, Cur Avg Loss: 0.53038401, Log Avg loss: 0.41455600, Global Avg Loss: 2.85580154, Time: 0.0211 Steps: 14070, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001306, Sample Num: 20896, Cur Loss: 0.34241652, Cur Avg Loss: 0.53038128, Log Avg loss: 0.53002701, Global Avg Loss: 2.85414971, Time: 0.0210 Steps: 14080, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001316, Sample Num: 21056, Cur Loss: 0.69732147, Cur Avg Loss: 0.53247938, Log Avg loss: 0.80649144, Global Avg Loss: 2.85269644, Time: 0.0210 Steps: 14090, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001326, Sample Num: 21216, Cur Loss: 0.31580392, Cur Avg Loss: 0.53368590, Log Avg loss: 0.69246391, Global Avg Loss: 2.85116436, Time: 0.0210 Steps: 14100, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001336, Sample Num: 21376, Cur Loss: 0.34936774, Cur Avg Loss: 0.53407375, Log Avg loss: 0.58550273, Global Avg Loss: 2.84955865, Time: 0.0210 Steps: 14110, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001346, Sample Num: 21536, Cur Loss: 0.23316315, Cur Avg Loss: 0.53361149, Log Avg loss: 0.47185346, Global Avg Loss: 2.84787472, Time: 0.0210 Steps: 14120, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001356, Sample Num: 21696, Cur Loss: 1.12014580, Cur Avg Loss: 0.53319867, Log Avg loss: 0.47763354, Global Avg Loss: 2.84619727, Time: 0.0210 Steps: 14130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001366, Sample Num: 21856, Cur Loss: 0.60746855, Cur Avg Loss: 0.53348758, Log Avg loss: 0.57266401, Global Avg Loss: 2.84458940, Time: 0.0210 Steps: 14140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001376, Sample Num: 22016, Cur Loss: 0.20000675, Cur Avg Loss: 0.53357658, Log Avg loss: 0.54573346, Global Avg Loss: 2.84296476, Time: 0.0210 Steps: 14150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001386, Sample Num: 22176, Cur Loss: 0.47330463, Cur Avg Loss: 0.53244096, Log Avg loss: 0.37617954, Global Avg Loss: 2.84122268, Time: 0.0210 Steps: 14160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001396, Sample Num: 22336, Cur Loss: 0.93372428, Cur Avg Loss: 0.53202584, Log Avg loss: 0.47448982, Global Avg Loss: 2.83955244, Time: 0.0210 Steps: 14170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001406, Sample Num: 22496, Cur Loss: 0.70564353, Cur Avg Loss: 0.53512984, Log Avg loss: 0.96844833, Global Avg Loss: 2.83823290, Time: 0.0210 Steps: 14180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001416, Sample Num: 22656, Cur Loss: 0.62826282, Cur Avg Loss: 0.53900186, Log Avg loss: 1.08340850, Global Avg Loss: 2.83699624, Time: 0.0210 Steps: 14190, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001426, Sample Num: 22816, Cur Loss: 0.55311966, Cur Avg Loss: 0.54132161, Log Avg loss: 0.86979863, Global Avg Loss: 2.83561089, Time: 0.0210 Steps: 14200, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001436, Sample Num: 22976, Cur Loss: 0.14395858, Cur Avg Loss: 0.54331748, Log Avg loss: 0.82792822, Global Avg Loss: 2.83419802, Time: 0.0210 Steps: 14210, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001446, Sample Num: 23136, Cur Loss: 0.47602159, Cur Avg Loss: 0.54236494, Log Avg loss: 0.40558050, Global Avg Loss: 2.83249014, Time: 0.0210 Steps: 14220, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001456, Sample Num: 23296, Cur Loss: 0.70673358, Cur Avg Loss: 0.54190322, Log Avg loss: 0.47513779, Global Avg Loss: 2.83083353, Time: 0.0210 Steps: 14230, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001466, Sample Num: 23456, Cur Loss: 0.27050832, Cur Avg Loss: 0.54134695, Log Avg loss: 0.46035372, Global Avg Loss: 2.82916887, Time: 0.0210 Steps: 14240, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001476, Sample Num: 23616, Cur Loss: 0.81358910, Cur Avg Loss: 0.54072101, Log Avg loss: 0.44895926, Global Avg Loss: 2.82749854, Time: 0.0211 Steps: 14250, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001486, Sample Num: 23776, Cur Loss: 0.49895251, Cur Avg Loss: 0.54056532, Log Avg loss: 0.51758554, Global Avg Loss: 2.82587869, Time: 0.0210 Steps: 14260, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001496, Sample Num: 23936, Cur Loss: 0.55839813, Cur Avg Loss: 0.54051194, Log Avg loss: 0.53257837, Global Avg Loss: 2.82427161, Time: 0.0210 Steps: 14270, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001506, Sample Num: 24096, Cur Loss: 0.44766244, Cur Avg Loss: 0.54036816, Log Avg loss: 0.51885986, Global Avg Loss: 2.82265718, Time: 0.0210 Steps: 14280, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001516, Sample Num: 24256, Cur Loss: 0.50549555, Cur Avg Loss: 0.54309313, Log Avg loss: 0.95347311, Global Avg Loss: 2.82134914, Time: 0.0210 Steps: 14290, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001526, Sample Num: 24416, Cur Loss: 0.60752720, Cur Avg Loss: 0.54332427, Log Avg loss: 0.57836536, Global Avg Loss: 2.81978062, Time: 0.0210 Steps: 14300, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001536, Sample Num: 24576, Cur Loss: 0.55867195, Cur Avg Loss: 0.54378837, Log Avg loss: 0.61461004, Global Avg Loss: 2.81823962, Time: 0.0255 Steps: 14310, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001546, Sample Num: 24736, Cur Loss: 0.66152400, Cur Avg Loss: 0.54265512, Log Avg loss: 0.36858778, Global Avg Loss: 2.81652897, Time: 0.0211 Steps: 14320, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001556, Sample Num: 24896, Cur Loss: 0.85535437, Cur Avg Loss: 0.54283259, Log Avg loss: 0.57026895, Global Avg Loss: 2.81496145, Time: 0.0211 Steps: 14330, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001566, Sample Num: 25056, Cur Loss: 0.49849588, Cur Avg Loss: 0.54127986, Log Avg loss: 0.29967483, Global Avg Loss: 2.81320741, Time: 0.0210 Steps: 14340, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001576, Sample Num: 25216, Cur Loss: 0.82446170, Cur Avg Loss: 0.54204248, Log Avg loss: 0.66147011, Global Avg Loss: 2.81170794, Time: 0.0210 Steps: 14350, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001586, Sample Num: 25376, Cur Loss: 1.10186815, Cur Avg Loss: 0.54275027, Log Avg loss: 0.65429720, Global Avg Loss: 2.81020557, Time: 0.0211 Steps: 14360, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001596, Sample Num: 25536, Cur Loss: 0.51026124, Cur Avg Loss: 0.54294792, Log Avg loss: 0.57429446, Global Avg Loss: 2.80864961, Time: 0.0210 Steps: 14370, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001606, Sample Num: 25696, Cur Loss: 0.43783465, Cur Avg Loss: 0.54387017, Log Avg loss: 0.69106231, Global Avg Loss: 2.80717702, Time: 0.0211 Steps: 14380, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001616, Sample Num: 25856, Cur Loss: 0.71963489, Cur Avg Loss: 0.54441901, Log Avg loss: 0.63256273, Global Avg Loss: 2.80566582, Time: 0.0211 Steps: 14390, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001626, Sample Num: 26016, Cur Loss: 0.48303437, Cur Avg Loss: 0.54466237, Log Avg loss: 0.58398984, Global Avg Loss: 2.80412299, Time: 0.0212 Steps: 14400, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001636, Sample Num: 26176, Cur Loss: 0.47709316, Cur Avg Loss: 0.54395625, Log Avg loss: 0.42914084, Global Avg Loss: 2.80247484, Time: 0.0210 Steps: 14410, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001646, Sample Num: 26336, Cur Loss: 0.25748393, Cur Avg Loss: 0.54289396, Log Avg loss: 0.36910298, Global Avg Loss: 2.80078734, Time: 0.0211 Steps: 14420, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001656, Sample Num: 26496, Cur Loss: 0.35844487, Cur Avg Loss: 0.54288672, Log Avg loss: 0.54169446, Global Avg Loss: 2.79922179, Time: 0.0210 Steps: 14430, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001666, Sample Num: 26656, Cur Loss: 1.16889548, Cur Avg Loss: 0.54271218, Log Avg loss: 0.51380807, Global Avg Loss: 2.79763909, Time: 0.0211 Steps: 14440, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001676, Sample Num: 26816, Cur Loss: 0.18214796, Cur Avg Loss: 0.54532924, Log Avg loss: 0.98133153, Global Avg Loss: 2.79638213, Time: 0.0210 Steps: 14450, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001686, Sample Num: 26976, Cur Loss: 1.13094854, Cur Avg Loss: 0.54487366, Log Avg loss: 0.46851966, Global Avg Loss: 2.79477227, Time: 0.0210 Steps: 14460, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001696, Sample Num: 27136, Cur Loss: 0.38856104, Cur Avg Loss: 0.54446366, Log Avg loss: 0.47533762, Global Avg Loss: 2.79316934, Time: 0.0210 Steps: 14470, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001706, Sample Num: 27296, Cur Loss: 0.55709964, Cur Avg Loss: 0.54425309, Log Avg loss: 0.50854013, Global Avg Loss: 2.79159156, Time: 0.0210 Steps: 14480, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001716, Sample Num: 27456, Cur Loss: 0.39731762, Cur Avg Loss: 0.54358021, Log Avg loss: 0.42878634, Global Avg Loss: 2.78996092, Time: 0.0210 Steps: 14490, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001726, Sample Num: 27616, Cur Loss: 0.48007289, Cur Avg Loss: 0.54313136, Log Avg loss: 0.46610874, Global Avg Loss: 2.78835826, Time: 0.0210 Steps: 14500, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001736, Sample Num: 27776, Cur Loss: 1.21162319, Cur Avg Loss: 0.54488434, Log Avg loss: 0.84744979, Global Avg Loss: 2.78702062, Time: 0.0210 Steps: 14510, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001746, Sample Num: 27936, Cur Loss: 0.74541342, Cur Avg Loss: 0.54549990, Log Avg loss: 0.65236112, Global Avg Loss: 2.78555047, Time: 0.0210 Steps: 14520, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001756, Sample Num: 28096, Cur Loss: 0.35426903, Cur Avg Loss: 0.54513587, Log Avg loss: 0.48157582, Global Avg Loss: 2.78396481, Time: 0.0211 Steps: 14530, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001766, Sample Num: 28256, Cur Loss: 0.71671343, Cur Avg Loss: 0.54536677, Log Avg loss: 0.58591305, Global Avg Loss: 2.78245308, Time: 0.0211 Steps: 14540, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001776, Sample Num: 28416, Cur Loss: 0.19767840, Cur Avg Loss: 0.54514247, Log Avg loss: 0.50553052, Global Avg Loss: 2.78088818, Time: 0.0210 Steps: 14550, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001786, Sample Num: 28576, Cur Loss: 0.29761204, Cur Avg Loss: 0.54441041, Log Avg loss: 0.41439587, Global Avg Loss: 2.77926284, Time: 0.0211 Steps: 14560, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001796, Sample Num: 28736, Cur Loss: 0.15887985, Cur Avg Loss: 0.54396635, Log Avg loss: 0.46465794, Global Avg Loss: 2.77767423, Time: 0.0248 Steps: 14570, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001806, Sample Num: 28896, Cur Loss: 0.63400060, Cur Avg Loss: 0.54353422, Log Avg loss: 0.46592413, Global Avg Loss: 2.77608867, Time: 0.0210 Steps: 14580, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001816, Sample Num: 29056, Cur Loss: 1.19427562, Cur Avg Loss: 0.54471040, Log Avg loss: 0.75712799, Global Avg Loss: 2.77470487, Time: 0.0210 Steps: 14590, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001826, Sample Num: 29216, Cur Loss: 0.37652233, Cur Avg Loss: 0.54410072, Log Avg loss: 0.43338223, Global Avg Loss: 2.77310123, Time: 0.0210 Steps: 14600, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001836, Sample Num: 29376, Cur Loss: 0.36974710, Cur Avg Loss: 0.54459629, Log Avg loss: 0.63508759, Global Avg Loss: 2.77163784, Time: 0.0212 Steps: 14610, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001846, Sample Num: 29536, Cur Loss: 0.52503639, Cur Avg Loss: 0.54412903, Log Avg loss: 0.45834078, Global Avg Loss: 2.77005556, Time: 0.0210 Steps: 14620, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001856, Sample Num: 29696, Cur Loss: 0.15026984, Cur Avg Loss: 0.54512981, Log Avg loss: 0.72987301, Global Avg Loss: 2.76866104, Time: 0.0210 Steps: 14630, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001866, Sample Num: 29856, Cur Loss: 0.50621009, Cur Avg Loss: 0.54436918, Log Avg loss: 0.40319626, Global Avg Loss: 2.76704528, Time: 0.0212 Steps: 14640, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001876, Sample Num: 30016, Cur Loss: 0.21909374, Cur Avg Loss: 0.54407283, Log Avg loss: 0.48877523, Global Avg Loss: 2.76549015, Time: 0.0212 Steps: 14650, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001886, Sample Num: 30176, Cur Loss: 0.43363911, Cur Avg Loss: 0.54427739, Log Avg loss: 0.58265202, Global Avg Loss: 2.76400117, Time: 0.0211 Steps: 14660, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001896, Sample Num: 30336, Cur Loss: 0.25174677, Cur Avg Loss: 0.54355790, Log Avg loss: 0.40786189, Global Avg Loss: 2.76239508, Time: 0.0211 Steps: 14670, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001906, Sample Num: 30496, Cur Loss: 0.87926590, Cur Avg Loss: 0.54352591, Log Avg loss: 0.53746149, Global Avg Loss: 2.76087946, Time: 0.0211 Steps: 14680, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001916, Sample Num: 30656, Cur Loss: 0.76300043, Cur Avg Loss: 0.54324138, Log Avg loss: 0.48900836, Global Avg Loss: 2.75933291, Time: 0.0211 Steps: 14690, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001926, Sample Num: 30816, Cur Loss: 0.61845297, Cur Avg Loss: 0.54388934, Log Avg loss: 0.66804015, Global Avg Loss: 2.75791027, Time: 0.0210 Steps: 14700, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001936, Sample Num: 30976, Cur Loss: 0.25017393, Cur Avg Loss: 0.54324628, Log Avg loss: 0.41939175, Global Avg Loss: 2.75632052, Time: 0.0210 Steps: 14710, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001946, Sample Num: 31136, Cur Loss: 0.74299985, Cur Avg Loss: 0.54503132, Log Avg loss: 0.89061495, Global Avg Loss: 2.75505306, Time: 0.0213 Steps: 14720, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001956, Sample Num: 31296, Cur Loss: 0.18172029, Cur Avg Loss: 0.54446531, Log Avg loss: 0.43431939, Global Avg Loss: 2.75347754, Time: 0.0211 Steps: 14730, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001966, Sample Num: 31456, Cur Loss: 0.42918882, Cur Avg Loss: 0.54446903, Log Avg loss: 0.54519810, Global Avg Loss: 2.75197939, Time: 0.0212 Steps: 14740, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001976, Sample Num: 31616, Cur Loss: 0.49868873, Cur Avg Loss: 0.54463088, Log Avg loss: 0.57644992, Global Avg Loss: 2.75050445, Time: 0.0210 Steps: 14750, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001986, Sample Num: 31776, Cur Loss: 0.33720931, Cur Avg Loss: 0.54434503, Log Avg loss: 0.48786083, Global Avg Loss: 2.74897149, Time: 0.0211 Steps: 14760, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001996, Sample Num: 31936, Cur Loss: 0.23685834, Cur Avg Loss: 0.54373447, Log Avg loss: 0.42247773, Global Avg Loss: 2.74739635, Time: 0.0212 Steps: 14770, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002006, Sample Num: 32096, Cur Loss: 0.40705678, Cur Avg Loss: 0.54360251, Log Avg loss: 0.51726387, Global Avg Loss: 2.74588746, Time: 0.0211 Steps: 14780, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002016, Sample Num: 32256, Cur Loss: 0.76005733, Cur Avg Loss: 0.54395671, Log Avg loss: 0.61500774, Global Avg Loss: 2.74444670, Time: 0.0210 Steps: 14790, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002026, Sample Num: 32416, Cur Loss: 0.19127913, Cur Avg Loss: 0.54401588, Log Avg loss: 0.55594562, Global Avg Loss: 2.74296799, Time: 0.0210 Steps: 14800, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002036, Sample Num: 32576, Cur Loss: 0.91820717, Cur Avg Loss: 0.54364571, Log Avg loss: 0.46864960, Global Avg Loss: 2.74143232, Time: 0.0211 Steps: 14810, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002046, Sample Num: 32736, Cur Loss: 0.15496302, Cur Avg Loss: 0.54259098, Log Avg loss: 0.32784733, Global Avg Loss: 2.73980372, Time: 0.0212 Steps: 14820, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002056, Sample Num: 32896, Cur Loss: 0.44382328, Cur Avg Loss: 0.54178386, Log Avg loss: 0.37664603, Global Avg Loss: 2.73821022, Time: 0.0212 Steps: 14830, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002066, Sample Num: 33056, Cur Loss: 0.33413324, Cur Avg Loss: 0.54186877, Log Avg loss: 0.55932779, Global Avg Loss: 2.73674198, Time: 0.0211 Steps: 14840, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002076, Sample Num: 33216, Cur Loss: 0.83075690, Cur Avg Loss: 0.54155301, Log Avg loss: 0.47631576, Global Avg Loss: 2.73521980, Time: 0.0210 Steps: 14850, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002086, Sample Num: 33376, Cur Loss: 1.10634387, Cur Avg Loss: 0.54089856, Log Avg loss: 0.40503605, Global Avg Loss: 2.73365171, Time: 0.0210 Steps: 14860, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002096, Sample Num: 33536, Cur Loss: 0.08980026, Cur Avg Loss: 0.54070482, Log Avg loss: 0.50029039, Global Avg Loss: 2.73214979, Time: 0.0211 Steps: 14870, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002106, Sample Num: 33696, Cur Loss: 0.34121883, Cur Avg Loss: 0.54007748, Log Avg loss: 0.40858762, Global Avg Loss: 2.73058825, Time: 0.0211 Steps: 14880, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002116, Sample Num: 33856, Cur Loss: 0.48196527, Cur Avg Loss: 0.53951941, Log Avg loss: 0.42198897, Global Avg Loss: 2.72903782, Time: 0.0211 Steps: 14890, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002126, Sample Num: 34016, Cur Loss: 1.06048632, Cur Avg Loss: 0.54062185, Log Avg loss: 0.77389769, Global Avg Loss: 2.72772564, Time: 0.0212 Steps: 14900, Updated lr: 0.000087 ***** Running evaluation checkpoint-14903 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-14903 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.115666, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.425706, "eval_total_loss": 299.271449, "eval_mae": 0.445516, "eval_mse": 0.425867, "eval_r2": 0.729291, "eval_sp_statistic": 0.822418, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.861287, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.422459, "test_total_loss": 212.074309, "test_mae": 0.477386, "test_mse": 0.422603, "test_r2": 0.727248, "test_sp_statistic": 0.819959, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.877473, "test_ps_pvalue": 0.0, "lr": 8.681555239449977e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.727416188989865, "train_cur_epoch_loss": 1152.933437012136, "train_cur_epoch_avg_loss": 0.5415375467412569, "train_cur_epoch_time": 45.11566638946533, "train_cur_epoch_avg_time": 0.021191012864943792, "epoch": 7, "step": 14903} ################################################## Training, Epoch: 0008, Batch: 000007, Sample Num: 112, Cur Loss: 1.23746693, Cur Avg Loss: 0.92973294, Log Avg loss: 1.00795187, Global Avg Loss: 2.72657221, Time: 0.0212 Steps: 14910, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000017, Sample Num: 272, Cur Loss: 0.29042420, Cur Avg Loss: 0.62763834, Log Avg loss: 0.41617212, Global Avg Loss: 2.72502368, Time: 0.0211 Steps: 14920, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000027, Sample Num: 432, Cur Loss: 0.31949088, Cur Avg Loss: 0.58605623, Log Avg loss: 0.51536664, Global Avg Loss: 2.72354367, Time: 0.0211 Steps: 14930, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000037, Sample Num: 592, Cur Loss: 0.74934494, Cur Avg Loss: 0.57175995, Log Avg loss: 0.53316000, Global Avg Loss: 2.72207755, Time: 0.0211 Steps: 14940, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000047, Sample Num: 752, Cur Loss: 0.81027651, Cur Avg Loss: 0.57089413, Log Avg loss: 0.56769061, Global Avg Loss: 2.72063649, Time: 0.0211 Steps: 14950, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000057, Sample Num: 912, Cur Loss: 0.42336851, Cur Avg Loss: 0.53348483, Log Avg loss: 0.35766107, Global Avg Loss: 2.71905696, Time: 0.0211 Steps: 14960, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000067, Sample Num: 1072, Cur Loss: 0.32206070, Cur Avg Loss: 0.52575431, Log Avg loss: 0.48169037, Global Avg Loss: 2.71756239, Time: 0.0211 Steps: 14970, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000077, Sample Num: 1232, Cur Loss: 0.17926082, Cur Avg Loss: 0.53217017, Log Avg loss: 0.57515643, Global Avg Loss: 2.71613221, Time: 0.0211 Steps: 14980, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000087, Sample Num: 1392, Cur Loss: 0.31161755, Cur Avg Loss: 0.55418148, Log Avg loss: 0.72366855, Global Avg Loss: 2.71480302, Time: 0.0211 Steps: 14990, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000097, Sample Num: 1552, Cur Loss: 0.24868757, Cur Avg Loss: 0.58904081, Log Avg loss: 0.89231697, Global Avg Loss: 2.71358803, Time: 0.0211 Steps: 15000, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000107, Sample Num: 1712, Cur Loss: 0.75985605, Cur Avg Loss: 0.58229611, Log Avg loss: 0.51687255, Global Avg Loss: 2.71212453, Time: 0.0211 Steps: 15010, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000117, Sample Num: 1872, Cur Loss: 0.74606109, Cur Avg Loss: 0.60561557, Log Avg loss: 0.85513384, Global Avg Loss: 2.71088818, Time: 0.0211 Steps: 15020, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000127, Sample Num: 2032, Cur Loss: 0.91724575, Cur Avg Loss: 0.60800519, Log Avg loss: 0.63596366, Global Avg Loss: 2.70950766, Time: 0.0211 Steps: 15030, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000137, Sample Num: 2192, Cur Loss: 0.49089283, Cur Avg Loss: 0.59327408, Log Avg loss: 0.40618907, Global Avg Loss: 2.70797620, Time: 0.0211 Steps: 15040, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000147, Sample Num: 2352, Cur Loss: 0.23944342, Cur Avg Loss: 0.58384682, Log Avg loss: 0.45469332, Global Avg Loss: 2.70647900, Time: 0.0211 Steps: 15050, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000157, Sample Num: 2512, Cur Loss: 0.44526398, Cur Avg Loss: 0.57567878, Log Avg loss: 0.45560850, Global Avg Loss: 2.70498440, Time: 0.0211 Steps: 15060, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000167, Sample Num: 2672, Cur Loss: 0.27663073, Cur Avg Loss: 0.56184211, Log Avg loss: 0.34460653, Global Avg Loss: 2.70341812, Time: 0.0211 Steps: 15070, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000177, Sample Num: 2832, Cur Loss: 0.41107014, Cur Avg Loss: 0.56310495, Log Avg loss: 0.58419437, Global Avg Loss: 2.70201280, Time: 0.0211 Steps: 15080, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000187, Sample Num: 2992, Cur Loss: 0.59260404, Cur Avg Loss: 0.54857892, Log Avg loss: 0.29146812, Global Avg Loss: 2.70041536, Time: 0.0211 Steps: 15090, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000197, Sample Num: 3152, Cur Loss: 0.21282655, Cur Avg Loss: 0.53913643, Log Avg loss: 0.36256195, Global Avg Loss: 2.69886711, Time: 0.0211 Steps: 15100, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000207, Sample Num: 3312, Cur Loss: 0.34677026, Cur Avg Loss: 0.53486286, Log Avg loss: 0.45067355, Global Avg Loss: 2.69737922, Time: 0.0211 Steps: 15110, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000217, Sample Num: 3472, Cur Loss: 0.38686028, Cur Avg Loss: 0.52815620, Log Avg loss: 0.38932817, Global Avg Loss: 2.69585274, Time: 0.0211 Steps: 15120, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000227, Sample Num: 3632, Cur Loss: 0.29669252, Cur Avg Loss: 0.52849314, Log Avg loss: 0.53580484, Global Avg Loss: 2.69442508, Time: 0.0211 Steps: 15130, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000237, Sample Num: 3792, Cur Loss: 0.66236436, Cur Avg Loss: 0.52694249, Log Avg loss: 0.49174276, Global Avg Loss: 2.69297020, Time: 0.0211 Steps: 15140, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000247, Sample Num: 3952, Cur Loss: 1.13068295, Cur Avg Loss: 0.53560336, Log Avg loss: 0.74086586, Global Avg Loss: 2.69168168, Time: 0.0211 Steps: 15150, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000257, Sample Num: 4112, Cur Loss: 0.91562408, Cur Avg Loss: 0.54448597, Log Avg loss: 0.76388644, Global Avg Loss: 2.69041005, Time: 0.0247 Steps: 15160, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000267, Sample Num: 4272, Cur Loss: 0.65288377, Cur Avg Loss: 0.55028441, Log Avg loss: 0.69930427, Global Avg Loss: 2.68909752, Time: 0.0212 Steps: 15170, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000277, Sample Num: 4432, Cur Loss: 0.27445009, Cur Avg Loss: 0.54497312, Log Avg loss: 0.40316168, Global Avg Loss: 2.68759163, Time: 0.0211 Steps: 15180, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000287, Sample Num: 4592, Cur Loss: 0.24309584, Cur Avg Loss: 0.54271320, Log Avg loss: 0.48011349, Global Avg Loss: 2.68613839, Time: 0.0213 Steps: 15190, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000297, Sample Num: 4752, Cur Loss: 0.59052354, Cur Avg Loss: 0.53932371, Log Avg loss: 0.44204534, Global Avg Loss: 2.68466201, Time: 0.0212 Steps: 15200, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000307, Sample Num: 4912, Cur Loss: 0.39405426, Cur Avg Loss: 0.53960297, Log Avg loss: 0.54789703, Global Avg Loss: 2.68325717, Time: 0.0213 Steps: 15210, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000317, Sample Num: 5072, Cur Loss: 0.44461513, Cur Avg Loss: 0.53380346, Log Avg loss: 0.35575849, Global Avg Loss: 2.68172793, Time: 0.0213 Steps: 15220, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000327, Sample Num: 5232, Cur Loss: 0.25595158, Cur Avg Loss: 0.53114323, Log Avg loss: 0.44681394, Global Avg Loss: 2.68026049, Time: 0.0211 Steps: 15230, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000337, Sample Num: 5392, Cur Loss: 0.26779905, Cur Avg Loss: 0.52563716, Log Avg loss: 0.34558854, Global Avg Loss: 2.67872856, Time: 0.0211 Steps: 15240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000347, Sample Num: 5552, Cur Loss: 0.27870142, Cur Avg Loss: 0.52067864, Log Avg loss: 0.35357665, Global Avg Loss: 2.67720387, Time: 0.0212 Steps: 15250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000357, Sample Num: 5712, Cur Loss: 0.27968889, Cur Avg Loss: 0.51824369, Log Avg loss: 0.43375107, Global Avg Loss: 2.67573371, Time: 0.0211 Steps: 15260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000367, Sample Num: 5872, Cur Loss: 0.58836359, Cur Avg Loss: 0.51704175, Log Avg loss: 0.47413240, Global Avg Loss: 2.67429193, Time: 0.0211 Steps: 15270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000377, Sample Num: 6032, Cur Loss: 0.77534366, Cur Avg Loss: 0.53034142, Log Avg loss: 1.01843920, Global Avg Loss: 2.67320826, Time: 0.0212 Steps: 15280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000387, Sample Num: 6192, Cur Loss: 0.63718843, Cur Avg Loss: 0.53476467, Log Avg loss: 0.70152137, Global Avg Loss: 2.67191873, Time: 0.0212 Steps: 15290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000397, Sample Num: 6352, Cur Loss: 0.50578511, Cur Avg Loss: 0.54538487, Log Avg loss: 0.95638638, Global Avg Loss: 2.67079747, Time: 0.0212 Steps: 15300, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000407, Sample Num: 6512, Cur Loss: 1.37061071, Cur Avg Loss: 0.55553900, Log Avg loss: 0.95865826, Global Avg Loss: 2.66967915, Time: 0.0212 Steps: 15310, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000417, Sample Num: 6672, Cur Loss: 1.05350375, Cur Avg Loss: 0.56015856, Log Avg loss: 0.74817462, Global Avg Loss: 2.66842491, Time: 0.0213 Steps: 15320, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000427, Sample Num: 6832, Cur Loss: 0.35096040, Cur Avg Loss: 0.55911772, Log Avg loss: 0.51571472, Global Avg Loss: 2.66702066, Time: 0.0212 Steps: 15330, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000437, Sample Num: 6992, Cur Loss: 0.39404523, Cur Avg Loss: 0.55897813, Log Avg loss: 0.55301743, Global Avg Loss: 2.66564256, Time: 0.0211 Steps: 15340, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000447, Sample Num: 7152, Cur Loss: 1.03935111, Cur Avg Loss: 0.56325594, Log Avg loss: 0.75019622, Global Avg Loss: 2.66439471, Time: 0.0211 Steps: 15350, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000457, Sample Num: 7312, Cur Loss: 0.57640362, Cur Avg Loss: 0.56337395, Log Avg loss: 0.56864889, Global Avg Loss: 2.66303030, Time: 0.0212 Steps: 15360, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000467, Sample Num: 7472, Cur Loss: 0.52569503, Cur Avg Loss: 0.55936992, Log Avg loss: 0.37638598, Global Avg Loss: 2.66154256, Time: 0.0212 Steps: 15370, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000477, Sample Num: 7632, Cur Loss: 2.67111683, Cur Avg Loss: 0.56639023, Log Avg loss: 0.89423864, Global Avg Loss: 2.66039347, Time: 0.0211 Steps: 15380, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000487, Sample Num: 7792, Cur Loss: 0.62526822, Cur Avg Loss: 0.56471701, Log Avg loss: 0.48490437, Global Avg Loss: 2.65897990, Time: 0.0212 Steps: 15390, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000497, Sample Num: 7952, Cur Loss: 0.55406964, Cur Avg Loss: 0.56026780, Log Avg loss: 0.34359155, Global Avg Loss: 2.65747640, Time: 0.0211 Steps: 15400, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000507, Sample Num: 8112, Cur Loss: 0.40215942, Cur Avg Loss: 0.55838378, Log Avg loss: 0.46474798, Global Avg Loss: 2.65605347, Time: 0.0212 Steps: 15410, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000517, Sample Num: 8272, Cur Loss: 0.43063715, Cur Avg Loss: 0.55589919, Log Avg loss: 0.42993009, Global Avg Loss: 2.65460981, Time: 0.0212 Steps: 15420, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000527, Sample Num: 8432, Cur Loss: 0.34060919, Cur Avg Loss: 0.55293782, Log Avg loss: 0.39983504, Global Avg Loss: 2.65314852, Time: 0.0211 Steps: 15430, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000537, Sample Num: 8592, Cur Loss: 0.18445271, Cur Avg Loss: 0.55091262, Log Avg loss: 0.44418470, Global Avg Loss: 2.65171785, Time: 0.0210 Steps: 15440, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000547, Sample Num: 8752, Cur Loss: 0.15514168, Cur Avg Loss: 0.54565275, Log Avg loss: 0.26319789, Global Avg Loss: 2.65017188, Time: 0.0211 Steps: 15450, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000557, Sample Num: 8912, Cur Loss: 0.57329577, Cur Avg Loss: 0.54524221, Log Avg loss: 0.52278561, Global Avg Loss: 2.64879582, Time: 0.0211 Steps: 15460, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000567, Sample Num: 9072, Cur Loss: 0.19206414, Cur Avg Loss: 0.54562191, Log Avg loss: 0.56677091, Global Avg Loss: 2.64744997, Time: 0.0210 Steps: 15470, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000577, Sample Num: 9232, Cur Loss: 0.89721453, Cur Avg Loss: 0.54493913, Log Avg loss: 0.50622573, Global Avg Loss: 2.64606675, Time: 0.0211 Steps: 15480, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000587, Sample Num: 9392, Cur Loss: 1.03308010, Cur Avg Loss: 0.54499338, Log Avg loss: 0.54812372, Global Avg Loss: 2.64471237, Time: 0.0211 Steps: 15490, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000597, Sample Num: 9552, Cur Loss: 0.44302946, Cur Avg Loss: 0.54916774, Log Avg loss: 0.79420258, Global Avg Loss: 2.64351849, Time: 0.0210 Steps: 15500, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000607, Sample Num: 9712, Cur Loss: 0.38237870, Cur Avg Loss: 0.55317214, Log Avg loss: 0.79223501, Global Avg Loss: 2.64232488, Time: 0.0211 Steps: 15510, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000617, Sample Num: 9872, Cur Loss: 0.33338049, Cur Avg Loss: 0.55182050, Log Avg loss: 0.46977593, Global Avg Loss: 2.64092505, Time: 0.0211 Steps: 15520, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000627, Sample Num: 10032, Cur Loss: 0.38178512, Cur Avg Loss: 0.55332999, Log Avg loss: 0.64646545, Global Avg Loss: 2.63964078, Time: 0.0210 Steps: 15530, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000637, Sample Num: 10192, Cur Loss: 0.46124354, Cur Avg Loss: 0.55346998, Log Avg loss: 0.56224720, Global Avg Loss: 2.63830398, Time: 0.0211 Steps: 15540, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000647, Sample Num: 10352, Cur Loss: 0.53010762, Cur Avg Loss: 0.55174339, Log Avg loss: 0.44175993, Global Avg Loss: 2.63689141, Time: 0.0211 Steps: 15550, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000657, Sample Num: 10512, Cur Loss: 1.24776065, Cur Avg Loss: 0.55417844, Log Avg loss: 0.71172599, Global Avg Loss: 2.63565416, Time: 0.0211 Steps: 15560, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000667, Sample Num: 10672, Cur Loss: 0.23150045, Cur Avg Loss: 0.55154597, Log Avg loss: 0.37859250, Global Avg Loss: 2.63420454, Time: 0.0211 Steps: 15570, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000677, Sample Num: 10832, Cur Loss: 0.46788943, Cur Avg Loss: 0.54840219, Log Avg loss: 0.33871205, Global Avg Loss: 2.63273118, Time: 0.0211 Steps: 15580, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000687, Sample Num: 10992, Cur Loss: 0.48423812, Cur Avg Loss: 0.54903206, Log Avg loss: 0.59167453, Global Avg Loss: 2.63142197, Time: 0.0211 Steps: 15590, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000697, Sample Num: 11152, Cur Loss: 0.28681439, Cur Avg Loss: 0.54926423, Log Avg loss: 0.56521444, Global Avg Loss: 2.63009748, Time: 0.0210 Steps: 15600, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000707, Sample Num: 11312, Cur Loss: 0.45052743, Cur Avg Loss: 0.54771588, Log Avg loss: 0.43979526, Global Avg Loss: 2.62869434, Time: 0.0210 Steps: 15610, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000717, Sample Num: 11472, Cur Loss: 0.54036939, Cur Avg Loss: 0.54466852, Log Avg loss: 0.32922077, Global Avg Loss: 2.62722220, Time: 0.0210 Steps: 15620, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000727, Sample Num: 11632, Cur Loss: 0.36385763, Cur Avg Loss: 0.54369955, Log Avg loss: 0.47422395, Global Avg Loss: 2.62584472, Time: 0.0211 Steps: 15630, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000737, Sample Num: 11792, Cur Loss: 1.18347263, Cur Avg Loss: 0.54296179, Log Avg loss: 0.48932677, Global Avg Loss: 2.62447866, Time: 0.0211 Steps: 15640, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000747, Sample Num: 11952, Cur Loss: 0.34890527, Cur Avg Loss: 0.54160508, Log Avg loss: 0.44161593, Global Avg Loss: 2.62308386, Time: 0.0211 Steps: 15650, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000757, Sample Num: 12112, Cur Loss: 0.77118909, Cur Avg Loss: 0.54051618, Log Avg loss: 0.45917476, Global Avg Loss: 2.62170206, Time: 0.0210 Steps: 15660, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000767, Sample Num: 12272, Cur Loss: 0.23968054, Cur Avg Loss: 0.53800153, Log Avg loss: 0.34764273, Global Avg Loss: 2.62025084, Time: 0.0210 Steps: 15670, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000777, Sample Num: 12432, Cur Loss: 0.33409932, Cur Avg Loss: 0.53493098, Log Avg loss: 0.29941957, Global Avg Loss: 2.61877072, Time: 0.0211 Steps: 15680, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000787, Sample Num: 12592, Cur Loss: 0.46862113, Cur Avg Loss: 0.53345566, Log Avg loss: 0.41882367, Global Avg Loss: 2.61736858, Time: 0.0210 Steps: 15690, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000797, Sample Num: 12752, Cur Loss: 0.45647985, Cur Avg Loss: 0.53195367, Log Avg loss: 0.41374703, Global Avg Loss: 2.61596500, Time: 0.0210 Steps: 15700, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000807, Sample Num: 12912, Cur Loss: 0.51374614, Cur Avg Loss: 0.53567779, Log Avg loss: 0.83248984, Global Avg Loss: 2.61482975, Time: 0.0210 Steps: 15710, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000817, Sample Num: 13072, Cur Loss: 0.86383152, Cur Avg Loss: 0.53687885, Log Avg loss: 0.63380459, Global Avg Loss: 2.61356956, Time: 0.0211 Steps: 15720, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000827, Sample Num: 13232, Cur Loss: 0.55493337, Cur Avg Loss: 0.53791834, Log Avg loss: 0.62284505, Global Avg Loss: 2.61230400, Time: 0.0210 Steps: 15730, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000837, Sample Num: 13392, Cur Loss: 0.25336918, Cur Avg Loss: 0.53765375, Log Avg loss: 0.51577200, Global Avg Loss: 2.61097202, Time: 0.0210 Steps: 15740, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000847, Sample Num: 13552, Cur Loss: 0.65206140, Cur Avg Loss: 0.53622191, Log Avg loss: 0.41637678, Global Avg Loss: 2.60957863, Time: 0.0210 Steps: 15750, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000857, Sample Num: 13712, Cur Loss: 0.15000826, Cur Avg Loss: 0.53507092, Log Avg loss: 0.43758165, Global Avg Loss: 2.60820046, Time: 0.0210 Steps: 15760, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000867, Sample Num: 13872, Cur Loss: 0.84211934, Cur Avg Loss: 0.53433487, Log Avg loss: 0.47125590, Global Avg Loss: 2.60684539, Time: 0.0211 Steps: 15770, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000877, Sample Num: 14032, Cur Loss: 0.52180332, Cur Avg Loss: 0.53849661, Log Avg loss: 0.89931931, Global Avg Loss: 2.60576331, Time: 0.0210 Steps: 15780, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000887, Sample Num: 14192, Cur Loss: 0.39484876, Cur Avg Loss: 0.53788086, Log Avg loss: 0.48387934, Global Avg Loss: 2.60441949, Time: 0.0211 Steps: 15790, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000897, Sample Num: 14352, Cur Loss: 0.58874136, Cur Avg Loss: 0.53806476, Log Avg loss: 0.55437737, Global Avg Loss: 2.60312200, Time: 0.0210 Steps: 15800, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000907, Sample Num: 14512, Cur Loss: 0.36133778, Cur Avg Loss: 0.53708389, Log Avg loss: 0.44909967, Global Avg Loss: 2.60175955, Time: 0.0210 Steps: 15810, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000917, Sample Num: 14672, Cur Loss: 0.73020065, Cur Avg Loss: 0.53716523, Log Avg loss: 0.54454245, Global Avg Loss: 2.60045916, Time: 0.0211 Steps: 15820, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000927, Sample Num: 14832, Cur Loss: 0.45138347, Cur Avg Loss: 0.53627203, Log Avg loss: 0.45436542, Global Avg Loss: 2.59910345, Time: 0.0211 Steps: 15830, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000937, Sample Num: 14992, Cur Loss: 0.15900344, Cur Avg Loss: 0.53552281, Log Avg loss: 0.46607029, Global Avg Loss: 2.59775684, Time: 0.0210 Steps: 15840, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000947, Sample Num: 15152, Cur Loss: 0.87694043, Cur Avg Loss: 0.53362375, Log Avg loss: 0.35568220, Global Avg Loss: 2.59634228, Time: 0.0210 Steps: 15850, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000957, Sample Num: 15312, Cur Loss: 0.43123302, Cur Avg Loss: 0.53199082, Log Avg loss: 0.37735251, Global Avg Loss: 2.59494317, Time: 0.0211 Steps: 15860, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000967, Sample Num: 15472, Cur Loss: 0.69243348, Cur Avg Loss: 0.53117153, Log Avg loss: 0.45276536, Global Avg Loss: 2.59359334, Time: 0.0212 Steps: 15870, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000977, Sample Num: 15632, Cur Loss: 1.11380005, Cur Avg Loss: 0.53594686, Log Avg loss: 0.99772090, Global Avg Loss: 2.59258838, Time: 0.0211 Steps: 15880, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000987, Sample Num: 15792, Cur Loss: 0.60240072, Cur Avg Loss: 0.53411571, Log Avg loss: 0.35521215, Global Avg Loss: 2.59118034, Time: 0.0211 Steps: 15890, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000997, Sample Num: 15952, Cur Loss: 0.66608387, Cur Avg Loss: 0.53511595, Log Avg loss: 0.63383987, Global Avg Loss: 2.58994931, Time: 0.0211 Steps: 15900, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001007, Sample Num: 16112, Cur Loss: 0.33620706, Cur Avg Loss: 0.53448016, Log Avg loss: 0.47109233, Global Avg Loss: 2.58861754, Time: 0.0210 Steps: 15910, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001017, Sample Num: 16272, Cur Loss: 0.29357630, Cur Avg Loss: 0.53316524, Log Avg loss: 0.40075285, Global Avg Loss: 2.58724325, Time: 0.0210 Steps: 15920, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001027, Sample Num: 16432, Cur Loss: 0.72004473, Cur Avg Loss: 0.53444728, Log Avg loss: 0.66483044, Global Avg Loss: 2.58603646, Time: 0.0248 Steps: 15930, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001037, Sample Num: 16592, Cur Loss: 0.53495109, Cur Avg Loss: 0.53481774, Log Avg loss: 0.57286386, Global Avg Loss: 2.58477349, Time: 0.0210 Steps: 15940, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001047, Sample Num: 16752, Cur Loss: 0.36636209, Cur Avg Loss: 0.53346483, Log Avg loss: 0.39316831, Global Avg Loss: 2.58339944, Time: 0.0211 Steps: 15950, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001057, Sample Num: 16912, Cur Loss: 0.09661473, Cur Avg Loss: 0.53360528, Log Avg loss: 0.54831007, Global Avg Loss: 2.58212433, Time: 0.0210 Steps: 15960, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001067, Sample Num: 17072, Cur Loss: 0.47802728, Cur Avg Loss: 0.53237374, Log Avg loss: 0.40219945, Global Avg Loss: 2.58075931, Time: 0.0212 Steps: 15970, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001077, Sample Num: 17232, Cur Loss: 0.46708265, Cur Avg Loss: 0.53115398, Log Avg loss: 0.40100602, Global Avg Loss: 2.57939526, Time: 0.0210 Steps: 15980, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001087, Sample Num: 17392, Cur Loss: 0.20471282, Cur Avg Loss: 0.53047396, Log Avg loss: 0.45723553, Global Avg Loss: 2.57806808, Time: 0.0211 Steps: 15990, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001097, Sample Num: 17552, Cur Loss: 0.59984910, Cur Avg Loss: 0.53036086, Log Avg loss: 0.51806755, Global Avg Loss: 2.57678058, Time: 0.0211 Steps: 16000, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001107, Sample Num: 17712, Cur Loss: 0.60355031, Cur Avg Loss: 0.53060989, Log Avg loss: 0.55792800, Global Avg Loss: 2.57551959, Time: 0.0210 Steps: 16010, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001117, Sample Num: 17872, Cur Loss: 0.40455556, Cur Avg Loss: 0.53132234, Log Avg loss: 0.61019052, Global Avg Loss: 2.57429279, Time: 0.0211 Steps: 16020, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001127, Sample Num: 18032, Cur Loss: 0.58427948, Cur Avg Loss: 0.53286168, Log Avg loss: 0.70480665, Global Avg Loss: 2.57312655, Time: 0.0211 Steps: 16030, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001137, Sample Num: 18192, Cur Loss: 0.50154948, Cur Avg Loss: 0.53185942, Log Avg loss: 0.41890481, Global Avg Loss: 2.57178352, Time: 0.0211 Steps: 16040, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001147, Sample Num: 18352, Cur Loss: 0.45455801, Cur Avg Loss: 0.53108309, Log Avg loss: 0.44281349, Global Avg Loss: 2.57045706, Time: 0.0212 Steps: 16050, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001157, Sample Num: 18512, Cur Loss: 0.42588025, Cur Avg Loss: 0.53165639, Log Avg loss: 0.59741473, Global Avg Loss: 2.56922851, Time: 0.0210 Steps: 16060, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001167, Sample Num: 18672, Cur Loss: 0.35294941, Cur Avg Loss: 0.53107377, Log Avg loss: 0.46366361, Global Avg Loss: 2.56791827, Time: 0.0211 Steps: 16070, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001177, Sample Num: 18832, Cur Loss: 0.43158373, Cur Avg Loss: 0.53039003, Log Avg loss: 0.45059848, Global Avg Loss: 2.56660153, Time: 0.0210 Steps: 16080, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001187, Sample Num: 18992, Cur Loss: 0.80449021, Cur Avg Loss: 0.53129065, Log Avg loss: 0.63729356, Global Avg Loss: 2.56540245, Time: 0.0211 Steps: 16090, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001197, Sample Num: 19152, Cur Loss: 0.40790266, Cur Avg Loss: 0.53003086, Log Avg loss: 0.38049304, Global Avg Loss: 2.56404537, Time: 0.0210 Steps: 16100, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001207, Sample Num: 19312, Cur Loss: 0.37079632, Cur Avg Loss: 0.52831901, Log Avg loss: 0.32341111, Global Avg Loss: 2.56265453, Time: 0.0210 Steps: 16110, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001217, Sample Num: 19472, Cur Loss: 0.31129679, Cur Avg Loss: 0.52685418, Log Avg loss: 0.35004870, Global Avg Loss: 2.56128195, Time: 0.0210 Steps: 16120, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001227, Sample Num: 19632, Cur Loss: 0.49609822, Cur Avg Loss: 0.52633186, Log Avg loss: 0.46276560, Global Avg Loss: 2.55998095, Time: 0.0211 Steps: 16130, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001237, Sample Num: 19792, Cur Loss: 0.31176588, Cur Avg Loss: 0.52530061, Log Avg loss: 0.39876645, Global Avg Loss: 2.55864190, Time: 0.0211 Steps: 16140, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001247, Sample Num: 19952, Cur Loss: 0.63706833, Cur Avg Loss: 0.52439585, Log Avg loss: 0.41247678, Global Avg Loss: 2.55731301, Time: 0.0212 Steps: 16150, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001257, Sample Num: 20112, Cur Loss: 0.66537726, Cur Avg Loss: 0.52335383, Log Avg loss: 0.39341451, Global Avg Loss: 2.55597396, Time: 0.0211 Steps: 16160, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001267, Sample Num: 20272, Cur Loss: 1.08099830, Cur Avg Loss: 0.52472532, Log Avg loss: 0.69712084, Global Avg Loss: 2.55482439, Time: 0.0211 Steps: 16170, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001277, Sample Num: 20432, Cur Loss: 0.40571398, Cur Avg Loss: 0.52405843, Log Avg loss: 0.43956458, Global Avg Loss: 2.55351706, Time: 0.0211 Steps: 16180, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001287, Sample Num: 20592, Cur Loss: 0.87099332, Cur Avg Loss: 0.52471567, Log Avg loss: 0.60864468, Global Avg Loss: 2.55231578, Time: 0.0212 Steps: 16190, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001297, Sample Num: 20752, Cur Loss: 0.43405807, Cur Avg Loss: 0.52630881, Log Avg loss: 0.73134533, Global Avg Loss: 2.55119173, Time: 0.0210 Steps: 16200, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001307, Sample Num: 20912, Cur Loss: 0.69607699, Cur Avg Loss: 0.52660347, Log Avg loss: 0.56482187, Global Avg Loss: 2.54996633, Time: 0.0210 Steps: 16210, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001317, Sample Num: 21072, Cur Loss: 0.39664912, Cur Avg Loss: 0.52581328, Log Avg loss: 0.42253487, Global Avg Loss: 2.54865472, Time: 0.0210 Steps: 16220, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001327, Sample Num: 21232, Cur Loss: 0.21577990, Cur Avg Loss: 0.52665927, Log Avg loss: 0.63807566, Global Avg Loss: 2.54747753, Time: 0.0211 Steps: 16230, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001337, Sample Num: 21392, Cur Loss: 0.71851033, Cur Avg Loss: 0.52561079, Log Avg loss: 0.38647840, Global Avg Loss: 2.54614687, Time: 0.0210 Steps: 16240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001347, Sample Num: 21552, Cur Loss: 0.43693334, Cur Avg Loss: 0.52509957, Log Avg loss: 0.45674858, Global Avg Loss: 2.54486108, Time: 0.0211 Steps: 16250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001357, Sample Num: 21712, Cur Loss: 0.22803922, Cur Avg Loss: 0.52387530, Log Avg loss: 0.35896692, Global Avg Loss: 2.54351674, Time: 0.0211 Steps: 16260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001367, Sample Num: 21872, Cur Loss: 0.62074441, Cur Avg Loss: 0.52326643, Log Avg loss: 0.44064289, Global Avg Loss: 2.54222426, Time: 0.0210 Steps: 16270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001377, Sample Num: 22032, Cur Loss: 0.38474959, Cur Avg Loss: 0.52332066, Log Avg loss: 0.53073385, Global Avg Loss: 2.54098870, Time: 0.0211 Steps: 16280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001387, Sample Num: 22192, Cur Loss: 0.61059356, Cur Avg Loss: 0.52763010, Log Avg loss: 1.12103937, Global Avg Loss: 2.54011703, Time: 0.0211 Steps: 16290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001397, Sample Num: 22352, Cur Loss: 0.55050159, Cur Avg Loss: 0.52788079, Log Avg loss: 0.56265160, Global Avg Loss: 2.53890386, Time: 0.0211 Steps: 16300, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001407, Sample Num: 22512, Cur Loss: 0.35282540, Cur Avg Loss: 0.52728651, Log Avg loss: 0.44426574, Global Avg Loss: 2.53761959, Time: 0.0211 Steps: 16310, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001417, Sample Num: 22672, Cur Loss: 0.25887996, Cur Avg Loss: 0.52742483, Log Avg loss: 0.54688705, Global Avg Loss: 2.53639978, Time: 0.0210 Steps: 16320, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001427, Sample Num: 22832, Cur Loss: 0.69769728, Cur Avg Loss: 0.52668226, Log Avg loss: 0.42145960, Global Avg Loss: 2.53510466, Time: 0.0211 Steps: 16330, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001437, Sample Num: 22992, Cur Loss: 0.13159209, Cur Avg Loss: 0.52674397, Log Avg loss: 0.53554972, Global Avg Loss: 2.53388094, Time: 0.0211 Steps: 16340, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001447, Sample Num: 23152, Cur Loss: 0.70203114, Cur Avg Loss: 0.52665504, Log Avg loss: 0.51387540, Global Avg Loss: 2.53264546, Time: 0.0211 Steps: 16350, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001457, Sample Num: 23312, Cur Loss: 0.30567843, Cur Avg Loss: 0.52612397, Log Avg loss: 0.44927814, Global Avg Loss: 2.53137201, Time: 0.0211 Steps: 16360, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001467, Sample Num: 23472, Cur Loss: 0.94392467, Cur Avg Loss: 0.52816805, Log Avg loss: 0.82599047, Global Avg Loss: 2.53033024, Time: 0.0211 Steps: 16370, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001477, Sample Num: 23632, Cur Loss: 2.24334598, Cur Avg Loss: 0.53037579, Log Avg loss: 0.85425201, Global Avg Loss: 2.52930699, Time: 0.0210 Steps: 16380, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001487, Sample Num: 23792, Cur Loss: 0.51362145, Cur Avg Loss: 0.52990541, Log Avg loss: 0.46043055, Global Avg Loss: 2.52804471, Time: 0.0210 Steps: 16390, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001497, Sample Num: 23952, Cur Loss: 0.25491047, Cur Avg Loss: 0.53060649, Log Avg loss: 0.63485736, Global Avg Loss: 2.52689033, Time: 0.0211 Steps: 16400, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001507, Sample Num: 24112, Cur Loss: 0.60088634, Cur Avg Loss: 0.52964754, Log Avg loss: 0.38609161, Global Avg Loss: 2.52558576, Time: 0.0210 Steps: 16410, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001517, Sample Num: 24272, Cur Loss: 0.22206402, Cur Avg Loss: 0.52981387, Log Avg loss: 0.55488056, Global Avg Loss: 2.52438557, Time: 0.0209 Steps: 16420, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001527, Sample Num: 24432, Cur Loss: 0.56555963, Cur Avg Loss: 0.52944088, Log Avg loss: 0.47285757, Global Avg Loss: 2.52313693, Time: 0.0209 Steps: 16430, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001537, Sample Num: 24592, Cur Loss: 0.21730064, Cur Avg Loss: 0.53358404, Log Avg loss: 1.16624487, Global Avg Loss: 2.52231157, Time: 0.0247 Steps: 16440, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001547, Sample Num: 24752, Cur Loss: 0.45907775, Cur Avg Loss: 0.53407172, Log Avg loss: 0.60902883, Global Avg Loss: 2.52114848, Time: 0.0210 Steps: 16450, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001557, Sample Num: 24912, Cur Loss: 0.56555688, Cur Avg Loss: 0.53507684, Log Avg loss: 0.69056855, Global Avg Loss: 2.52003634, Time: 0.0210 Steps: 16460, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001567, Sample Num: 25072, Cur Loss: 0.59962785, Cur Avg Loss: 0.53432015, Log Avg loss: 0.41650261, Global Avg Loss: 2.51875915, Time: 0.0210 Steps: 16470, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001577, Sample Num: 25232, Cur Loss: 0.58934808, Cur Avg Loss: 0.53380227, Log Avg loss: 0.45265049, Global Avg Loss: 2.51750544, Time: 0.0210 Steps: 16480, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001587, Sample Num: 25392, Cur Loss: 0.43280193, Cur Avg Loss: 0.53338329, Log Avg loss: 0.46731106, Global Avg Loss: 2.51626214, Time: 0.0210 Steps: 16490, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001597, Sample Num: 25552, Cur Loss: 0.36797288, Cur Avg Loss: 0.53226292, Log Avg loss: 0.35446068, Global Avg Loss: 2.51495196, Time: 0.0211 Steps: 16500, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001607, Sample Num: 25712, Cur Loss: 0.17070137, Cur Avg Loss: 0.53180903, Log Avg loss: 0.45932158, Global Avg Loss: 2.51370688, Time: 0.0210 Steps: 16510, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001617, Sample Num: 25872, Cur Loss: 0.66658926, Cur Avg Loss: 0.53092523, Log Avg loss: 0.38889851, Global Avg Loss: 2.51242068, Time: 0.0210 Steps: 16520, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001627, Sample Num: 26032, Cur Loss: 0.20631799, Cur Avg Loss: 0.53084513, Log Avg loss: 0.51789365, Global Avg Loss: 2.51121406, Time: 0.0210 Steps: 16530, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001637, Sample Num: 26192, Cur Loss: 0.38451701, Cur Avg Loss: 0.53022543, Log Avg loss: 0.42940039, Global Avg Loss: 2.50995541, Time: 0.0210 Steps: 16540, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001647, Sample Num: 26352, Cur Loss: 0.90057480, Cur Avg Loss: 0.53063449, Log Avg loss: 0.59759810, Global Avg Loss: 2.50879991, Time: 0.0211 Steps: 16550, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001657, Sample Num: 26512, Cur Loss: 0.34692565, Cur Avg Loss: 0.53055312, Log Avg loss: 0.51715049, Global Avg Loss: 2.50759722, Time: 0.0210 Steps: 16560, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001667, Sample Num: 26672, Cur Loss: 0.07865666, Cur Avg Loss: 0.52989979, Log Avg loss: 0.42164318, Global Avg Loss: 2.50633835, Time: 0.0210 Steps: 16570, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001677, Sample Num: 26832, Cur Loss: 0.51043999, Cur Avg Loss: 0.52919785, Log Avg loss: 0.41218521, Global Avg Loss: 2.50507529, Time: 0.0210 Steps: 16580, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001687, Sample Num: 26992, Cur Loss: 0.49019501, Cur Avg Loss: 0.52915919, Log Avg loss: 0.52267583, Global Avg Loss: 2.50388035, Time: 0.0210 Steps: 16590, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001697, Sample Num: 27152, Cur Loss: 0.12030697, Cur Avg Loss: 0.52779565, Log Avg loss: 0.29776592, Global Avg Loss: 2.50255137, Time: 0.0210 Steps: 16600, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001707, Sample Num: 27312, Cur Loss: 0.25271636, Cur Avg Loss: 0.52825087, Log Avg loss: 0.60550110, Global Avg Loss: 2.50140925, Time: 0.0211 Steps: 16610, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001717, Sample Num: 27472, Cur Loss: 0.44460279, Cur Avg Loss: 0.52875277, Log Avg loss: 0.61442744, Global Avg Loss: 2.50027388, Time: 0.0210 Steps: 16620, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001727, Sample Num: 27632, Cur Loss: 0.42489088, Cur Avg Loss: 0.52974203, Log Avg loss: 0.69959864, Global Avg Loss: 2.49919110, Time: 0.0210 Steps: 16630, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001737, Sample Num: 27792, Cur Loss: 0.46200478, Cur Avg Loss: 0.52915509, Log Avg loss: 0.42779077, Global Avg Loss: 2.49794627, Time: 0.0211 Steps: 16640, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001747, Sample Num: 27952, Cur Loss: 0.50120127, Cur Avg Loss: 0.52929977, Log Avg loss: 0.55442977, Global Avg Loss: 2.49677899, Time: 0.0210 Steps: 16650, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001757, Sample Num: 28112, Cur Loss: 0.55327022, Cur Avg Loss: 0.52942550, Log Avg loss: 0.55139083, Global Avg Loss: 2.49561129, Time: 0.0210 Steps: 16660, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001767, Sample Num: 28272, Cur Loss: 1.01740098, Cur Avg Loss: 0.53039629, Log Avg loss: 0.70096406, Global Avg Loss: 2.49453472, Time: 0.0210 Steps: 16670, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001777, Sample Num: 28432, Cur Loss: 0.40235046, Cur Avg Loss: 0.53081058, Log Avg loss: 0.60401511, Global Avg Loss: 2.49340131, Time: 0.0210 Steps: 16680, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001787, Sample Num: 28592, Cur Loss: 0.60340959, Cur Avg Loss: 0.53014389, Log Avg loss: 0.41167278, Global Avg Loss: 2.49215402, Time: 0.0210 Steps: 16690, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001797, Sample Num: 28752, Cur Loss: 0.54550946, Cur Avg Loss: 0.53117962, Log Avg loss: 0.71626586, Global Avg Loss: 2.49109061, Time: 0.0246 Steps: 16700, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001807, Sample Num: 28912, Cur Loss: 0.85677731, Cur Avg Loss: 0.53064965, Log Avg loss: 0.43541275, Global Avg Loss: 2.48986041, Time: 0.0211 Steps: 16710, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001817, Sample Num: 29072, Cur Loss: 0.51897717, Cur Avg Loss: 0.53071813, Log Avg loss: 0.54309350, Global Avg Loss: 2.48869607, Time: 0.0211 Steps: 16720, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001827, Sample Num: 29232, Cur Loss: 0.15404418, Cur Avg Loss: 0.53021153, Log Avg loss: 0.43816223, Global Avg Loss: 2.48747041, Time: 0.0211 Steps: 16730, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001837, Sample Num: 29392, Cur Loss: 0.56794393, Cur Avg Loss: 0.52968589, Log Avg loss: 0.43365131, Global Avg Loss: 2.48624352, Time: 0.0210 Steps: 16740, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001847, Sample Num: 29552, Cur Loss: 0.43107367, Cur Avg Loss: 0.52994419, Log Avg loss: 0.57739403, Global Avg Loss: 2.48510390, Time: 0.0210 Steps: 16750, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001857, Sample Num: 29712, Cur Loss: 0.26884845, Cur Avg Loss: 0.52974563, Log Avg loss: 0.49307198, Global Avg Loss: 2.48391534, Time: 0.0208 Steps: 16760, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001867, Sample Num: 29872, Cur Loss: 0.93832266, Cur Avg Loss: 0.53053625, Log Avg loss: 0.67735323, Global Avg Loss: 2.48283808, Time: 0.0208 Steps: 16770, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001877, Sample Num: 30032, Cur Loss: 0.29446903, Cur Avg Loss: 0.52986398, Log Avg loss: 0.40435163, Global Avg Loss: 2.48159941, Time: 0.0208 Steps: 16780, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001887, Sample Num: 30192, Cur Loss: 0.33233619, Cur Avg Loss: 0.52960619, Log Avg loss: 0.48121995, Global Avg Loss: 2.48040800, Time: 0.0208 Steps: 16790, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001897, Sample Num: 30352, Cur Loss: 0.61435103, Cur Avg Loss: 0.52847640, Log Avg loss: 0.31528479, Global Avg Loss: 2.47911924, Time: 0.0208 Steps: 16800, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001907, Sample Num: 30512, Cur Loss: 0.57665145, Cur Avg Loss: 0.52784974, Log Avg loss: 0.40897191, Global Avg Loss: 2.47788774, Time: 0.0208 Steps: 16810, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001917, Sample Num: 30672, Cur Loss: 0.19303074, Cur Avg Loss: 0.52716801, Log Avg loss: 0.39716253, Global Avg Loss: 2.47665069, Time: 0.0209 Steps: 16820, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001927, Sample Num: 30832, Cur Loss: 0.90485126, Cur Avg Loss: 0.52702903, Log Avg loss: 0.50038600, Global Avg Loss: 2.47547644, Time: 0.0208 Steps: 16830, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001937, Sample Num: 30992, Cur Loss: 0.28184086, Cur Avg Loss: 0.52715545, Log Avg loss: 0.55151690, Global Avg Loss: 2.47433394, Time: 0.0208 Steps: 16840, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001947, Sample Num: 31152, Cur Loss: 1.76440382, Cur Avg Loss: 0.52948556, Log Avg loss: 0.98082773, Global Avg Loss: 2.47344759, Time: 0.0208 Steps: 16850, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001957, Sample Num: 31312, Cur Loss: 0.50253963, Cur Avg Loss: 0.53048842, Log Avg loss: 0.72574556, Global Avg Loss: 2.47241099, Time: 0.0210 Steps: 16860, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001967, Sample Num: 31472, Cur Loss: 0.74072766, Cur Avg Loss: 0.53072764, Log Avg loss: 0.57754288, Global Avg Loss: 2.47128777, Time: 0.0208 Steps: 16870, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001977, Sample Num: 31632, Cur Loss: 0.22910708, Cur Avg Loss: 0.52968808, Log Avg loss: 0.32520712, Global Avg Loss: 2.47001640, Time: 0.0208 Steps: 16880, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001987, Sample Num: 31792, Cur Loss: 0.61668074, Cur Avg Loss: 0.52950230, Log Avg loss: 0.49277333, Global Avg Loss: 2.46884574, Time: 0.0208 Steps: 16890, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001997, Sample Num: 31952, Cur Loss: 1.05830240, Cur Avg Loss: 0.53051472, Log Avg loss: 0.73168184, Global Avg Loss: 2.46781783, Time: 0.0208 Steps: 16900, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002007, Sample Num: 32112, Cur Loss: 0.28588250, Cur Avg Loss: 0.52948852, Log Avg loss: 0.32455669, Global Avg Loss: 2.46655038, Time: 0.0208 Steps: 16910, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002017, Sample Num: 32272, Cur Loss: 0.30924878, Cur Avg Loss: 0.52852053, Log Avg loss: 0.33424549, Global Avg Loss: 2.46529015, Time: 0.0208 Steps: 16920, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002027, Sample Num: 32432, Cur Loss: 0.70894074, Cur Avg Loss: 0.52820585, Log Avg loss: 0.46473419, Global Avg Loss: 2.46410849, Time: 0.0208 Steps: 16930, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002037, Sample Num: 32592, Cur Loss: 0.43173337, Cur Avg Loss: 0.52777401, Log Avg loss: 0.44024071, Global Avg Loss: 2.46291376, Time: 0.0208 Steps: 16940, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002047, Sample Num: 32752, Cur Loss: 1.10174346, Cur Avg Loss: 0.52776401, Log Avg loss: 0.52572672, Global Avg Loss: 2.46177088, Time: 0.0208 Steps: 16950, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002057, Sample Num: 32912, Cur Loss: 0.32832196, Cur Avg Loss: 0.52699492, Log Avg loss: 0.36956263, Global Avg Loss: 2.46053727, Time: 0.0207 Steps: 16960, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002067, Sample Num: 33072, Cur Loss: 0.75535876, Cur Avg Loss: 0.52649704, Log Avg loss: 0.42408227, Global Avg Loss: 2.45933723, Time: 0.0207 Steps: 16970, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002077, Sample Num: 33232, Cur Loss: 0.23642802, Cur Avg Loss: 0.52540565, Log Avg loss: 0.29981551, Global Avg Loss: 2.45806543, Time: 0.0207 Steps: 16980, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002087, Sample Num: 33392, Cur Loss: 0.19440314, Cur Avg Loss: 0.52471373, Log Avg loss: 0.38100204, Global Avg Loss: 2.45684291, Time: 0.0207 Steps: 16990, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002097, Sample Num: 33552, Cur Loss: 0.32792938, Cur Avg Loss: 0.52375792, Log Avg loss: 0.32428104, Global Avg Loss: 2.45558846, Time: 0.0207 Steps: 17000, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002107, Sample Num: 33712, Cur Loss: 0.73891592, Cur Avg Loss: 0.52352940, Log Avg loss: 0.47560890, Global Avg Loss: 2.45442445, Time: 0.0207 Steps: 17010, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002117, Sample Num: 33872, Cur Loss: 0.34782833, Cur Avg Loss: 0.52260565, Log Avg loss: 0.32797001, Global Avg Loss: 2.45317507, Time: 0.0207 Steps: 17020, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002127, Sample Num: 34032, Cur Loss: 0.23417071, Cur Avg Loss: 0.52157379, Log Avg loss: 0.30312954, Global Avg Loss: 2.45191256, Time: 0.0207 Steps: 17030, Updated lr: 0.000085 ***** Running evaluation checkpoint-17032 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-17032 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.995881, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.45195, "eval_total_loss": 317.720723, "eval_mae": 0.529396, "eval_mse": 0.452061, "eval_r2": 0.712641, "eval_sp_statistic": 0.823985, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.860747, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.565467, "test_total_loss": 283.86449, "test_mae": 0.64638, "test_mse": 0.56562, "test_r2": 0.634944, "test_sp_statistic": 0.813869, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.864743, "test_ps_pvalue": 0.0, "lr": 8.479658605974395e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.4516712040404194, "train_cur_epoch_loss": 1110.180482700467, "train_cur_epoch_avg_loss": 0.5214563093943011, "train_cur_epoch_time": 44.99588108062744, "train_cur_epoch_avg_time": 0.021134749215888888, "epoch": 8, "step": 17032} ################################################## Training, Epoch: 0009, Batch: 000008, Sample Num: 128, Cur Loss: 0.68648314, Cur Avg Loss: 0.51074537, Log Avg loss: 0.48789943, Global Avg Loss: 2.45075997, Time: 0.0210 Steps: 17040, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000018, Sample Num: 288, Cur Loss: 0.21379134, Cur Avg Loss: 0.48800711, Log Avg loss: 0.46981650, Global Avg Loss: 2.44959813, Time: 0.0209 Steps: 17050, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000028, Sample Num: 448, Cur Loss: 0.46603763, Cur Avg Loss: 0.48877966, Log Avg loss: 0.49017024, Global Avg Loss: 2.44844958, Time: 0.0209 Steps: 17060, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000038, Sample Num: 608, Cur Loss: 0.63424993, Cur Avg Loss: 0.47548967, Log Avg loss: 0.43827771, Global Avg Loss: 2.44727197, Time: 0.0209 Steps: 17070, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000048, Sample Num: 768, Cur Loss: 0.46421683, Cur Avg Loss: 0.48260645, Log Avg loss: 0.50965022, Global Avg Loss: 2.44613753, Time: 0.0209 Steps: 17080, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000058, Sample Num: 928, Cur Loss: 0.18151659, Cur Avg Loss: 0.46612582, Log Avg loss: 0.38701878, Global Avg Loss: 2.44493266, Time: 0.0208 Steps: 17090, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000068, Sample Num: 1088, Cur Loss: 0.27638280, Cur Avg Loss: 0.44836301, Log Avg loss: 0.34533874, Global Avg Loss: 2.44370483, Time: 0.0209 Steps: 17100, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000078, Sample Num: 1248, Cur Loss: 0.21961603, Cur Avg Loss: 0.44626047, Log Avg loss: 0.43196319, Global Avg Loss: 2.44252906, Time: 0.0208 Steps: 17110, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000088, Sample Num: 1408, Cur Loss: 0.74514568, Cur Avg Loss: 0.46448156, Log Avg loss: 0.60660604, Global Avg Loss: 2.44145668, Time: 0.0208 Steps: 17120, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000098, Sample Num: 1568, Cur Loss: 0.17558527, Cur Avg Loss: 0.45212400, Log Avg loss: 0.34337745, Global Avg Loss: 2.44023188, Time: 0.0209 Steps: 17130, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000108, Sample Num: 1728, Cur Loss: 0.48332965, Cur Avg Loss: 0.45091805, Log Avg loss: 0.43909980, Global Avg Loss: 2.43906436, Time: 0.0209 Steps: 17140, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000118, Sample Num: 1888, Cur Loss: 0.33717334, Cur Avg Loss: 0.43988786, Log Avg loss: 0.32076178, Global Avg Loss: 2.43782920, Time: 0.0209 Steps: 17150, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000128, Sample Num: 2048, Cur Loss: 0.46021885, Cur Avg Loss: 0.43946669, Log Avg loss: 0.43449686, Global Avg Loss: 2.43666175, Time: 0.0209 Steps: 17160, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000138, Sample Num: 2208, Cur Loss: 0.23833174, Cur Avg Loss: 0.43261782, Log Avg loss: 0.34495230, Global Avg Loss: 2.43544352, Time: 0.0208 Steps: 17170, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000148, Sample Num: 2368, Cur Loss: 0.29746890, Cur Avg Loss: 0.42338373, Log Avg loss: 0.29595334, Global Avg Loss: 2.43419818, Time: 0.0209 Steps: 17180, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000158, Sample Num: 2528, Cur Loss: 1.07556605, Cur Avg Loss: 0.44075390, Log Avg loss: 0.69783244, Global Avg Loss: 2.43318808, Time: 0.0209 Steps: 17190, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000168, Sample Num: 2688, Cur Loss: 0.36455396, Cur Avg Loss: 0.46425906, Log Avg loss: 0.83564045, Global Avg Loss: 2.43225927, Time: 0.0209 Steps: 17200, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000178, Sample Num: 2848, Cur Loss: 0.58482230, Cur Avg Loss: 0.46762791, Log Avg loss: 0.52422470, Global Avg Loss: 2.43115059, Time: 0.0209 Steps: 17210, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000188, Sample Num: 3008, Cur Loss: 0.74317801, Cur Avg Loss: 0.46162732, Log Avg loss: 0.35481670, Global Avg Loss: 2.42994482, Time: 0.0209 Steps: 17220, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000198, Sample Num: 3168, Cur Loss: 0.78277212, Cur Avg Loss: 0.46120105, Log Avg loss: 0.45318729, Global Avg Loss: 2.42879755, Time: 0.0209 Steps: 17230, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000208, Sample Num: 3328, Cur Loss: 0.61058855, Cur Avg Loss: 0.46251939, Log Avg loss: 0.48862248, Global Avg Loss: 2.42767216, Time: 0.0209 Steps: 17240, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000218, Sample Num: 3488, Cur Loss: 0.94797111, Cur Avg Loss: 0.46468871, Log Avg loss: 0.50981065, Global Avg Loss: 2.42656035, Time: 0.0208 Steps: 17250, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000228, Sample Num: 3648, Cur Loss: 0.36546800, Cur Avg Loss: 0.46301740, Log Avg loss: 0.42658266, Global Avg Loss: 2.42540162, Time: 0.0209 Steps: 17260, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000238, Sample Num: 3808, Cur Loss: 0.28366625, Cur Avg Loss: 0.46853886, Log Avg loss: 0.59442816, Global Avg Loss: 2.42434141, Time: 0.0209 Steps: 17270, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000248, Sample Num: 3968, Cur Loss: 0.64597452, Cur Avg Loss: 0.47290911, Log Avg loss: 0.57692111, Global Avg Loss: 2.42327230, Time: 0.0209 Steps: 17280, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000258, Sample Num: 4128, Cur Loss: 0.63829279, Cur Avg Loss: 0.47391698, Log Avg loss: 0.49891220, Global Avg Loss: 2.42215931, Time: 0.0246 Steps: 17290, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000268, Sample Num: 4288, Cur Loss: 0.25368965, Cur Avg Loss: 0.47029593, Log Avg loss: 0.37687291, Global Avg Loss: 2.42097707, Time: 0.0209 Steps: 17300, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000278, Sample Num: 4448, Cur Loss: 0.40703386, Cur Avg Loss: 0.46622365, Log Avg loss: 0.35708653, Global Avg Loss: 2.41978476, Time: 0.0209 Steps: 17310, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000288, Sample Num: 4608, Cur Loss: 0.45647988, Cur Avg Loss: 0.46471524, Log Avg loss: 0.42278146, Global Avg Loss: 2.41863175, Time: 0.0208 Steps: 17320, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000298, Sample Num: 4768, Cur Loss: 0.62437165, Cur Avg Loss: 0.47156159, Log Avg loss: 0.66873628, Global Avg Loss: 2.41762200, Time: 0.0209 Steps: 17330, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000308, Sample Num: 4928, Cur Loss: 0.08087452, Cur Avg Loss: 0.47038556, Log Avg loss: 0.43533984, Global Avg Loss: 2.41647882, Time: 0.0208 Steps: 17340, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000318, Sample Num: 5088, Cur Loss: 0.48759189, Cur Avg Loss: 0.46793073, Log Avg loss: 0.39232220, Global Avg Loss: 2.41531216, Time: 0.0208 Steps: 17350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000328, Sample Num: 5248, Cur Loss: 0.69118160, Cur Avg Loss: 0.46536629, Log Avg loss: 0.38381694, Global Avg Loss: 2.41414194, Time: 0.0208 Steps: 17360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000338, Sample Num: 5408, Cur Loss: 0.41391826, Cur Avg Loss: 0.46376344, Log Avg loss: 0.41119000, Global Avg Loss: 2.41298883, Time: 0.0208 Steps: 17370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000348, Sample Num: 5568, Cur Loss: 0.67581034, Cur Avg Loss: 0.46950615, Log Avg loss: 0.66360984, Global Avg Loss: 2.41198228, Time: 0.0208 Steps: 17380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000358, Sample Num: 5728, Cur Loss: 0.17556781, Cur Avg Loss: 0.46668178, Log Avg loss: 0.36839346, Global Avg Loss: 2.41080713, Time: 0.0208 Steps: 17390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000368, Sample Num: 5888, Cur Loss: 0.22784264, Cur Avg Loss: 0.46610602, Log Avg loss: 0.44549393, Global Avg Loss: 2.40967764, Time: 0.0209 Steps: 17400, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000378, Sample Num: 6048, Cur Loss: 0.10576189, Cur Avg Loss: 0.46213341, Log Avg loss: 0.31594133, Global Avg Loss: 2.40847504, Time: 0.0208 Steps: 17410, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000388, Sample Num: 6208, Cur Loss: 0.50781727, Cur Avg Loss: 0.46276243, Log Avg loss: 0.48653963, Global Avg Loss: 2.40737174, Time: 0.0208 Steps: 17420, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000398, Sample Num: 6368, Cur Loss: 0.13386989, Cur Avg Loss: 0.46313584, Log Avg loss: 0.47762412, Global Avg Loss: 2.40626460, Time: 0.0208 Steps: 17430, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000408, Sample Num: 6528, Cur Loss: 0.47246212, Cur Avg Loss: 0.46347552, Log Avg loss: 0.47699461, Global Avg Loss: 2.40515837, Time: 0.0208 Steps: 17440, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000418, Sample Num: 6688, Cur Loss: 0.44702402, Cur Avg Loss: 0.46654672, Log Avg loss: 0.59185181, Global Avg Loss: 2.40411923, Time: 0.0208 Steps: 17450, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000428, Sample Num: 6848, Cur Loss: 0.32430202, Cur Avg Loss: 0.47247619, Log Avg loss: 0.72032804, Global Avg Loss: 2.40315485, Time: 0.0208 Steps: 17460, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000438, Sample Num: 7008, Cur Loss: 0.24299753, Cur Avg Loss: 0.46935569, Log Avg loss: 0.33579817, Global Avg Loss: 2.40197148, Time: 0.0208 Steps: 17470, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000448, Sample Num: 7168, Cur Loss: 0.21167003, Cur Avg Loss: 0.46797681, Log Avg loss: 0.40758173, Global Avg Loss: 2.40083052, Time: 0.0208 Steps: 17480, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000458, Sample Num: 7328, Cur Loss: 0.57286257, Cur Avg Loss: 0.46779963, Log Avg loss: 0.45986198, Global Avg Loss: 2.39972076, Time: 0.0208 Steps: 17490, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000468, Sample Num: 7488, Cur Loss: 1.04948807, Cur Avg Loss: 0.46740554, Log Avg loss: 0.44935659, Global Avg Loss: 2.39860627, Time: 0.0208 Steps: 17500, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000478, Sample Num: 7648, Cur Loss: 0.37024724, Cur Avg Loss: 0.47184568, Log Avg loss: 0.67964384, Global Avg Loss: 2.39762457, Time: 0.0208 Steps: 17510, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000488, Sample Num: 7808, Cur Loss: 0.49962923, Cur Avg Loss: 0.47036888, Log Avg loss: 0.39977813, Global Avg Loss: 2.39648424, Time: 0.0208 Steps: 17520, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000498, Sample Num: 7968, Cur Loss: 0.61374223, Cur Avg Loss: 0.47045782, Log Avg loss: 0.47479821, Global Avg Loss: 2.39538802, Time: 0.0208 Steps: 17530, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000508, Sample Num: 8128, Cur Loss: 0.36550573, Cur Avg Loss: 0.47112016, Log Avg loss: 0.50410443, Global Avg Loss: 2.39430975, Time: 0.0208 Steps: 17540, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000518, Sample Num: 8288, Cur Loss: 0.34517944, Cur Avg Loss: 0.47564519, Log Avg loss: 0.70551687, Global Avg Loss: 2.39334747, Time: 0.0210 Steps: 17550, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000528, Sample Num: 8448, Cur Loss: 0.44968554, Cur Avg Loss: 0.47702177, Log Avg loss: 0.54832874, Global Avg Loss: 2.39229678, Time: 0.0208 Steps: 17560, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000538, Sample Num: 8608, Cur Loss: 0.32039613, Cur Avg Loss: 0.48048038, Log Avg loss: 0.66309461, Global Avg Loss: 2.39131260, Time: 0.0208 Steps: 17570, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000548, Sample Num: 8768, Cur Loss: 1.12436342, Cur Avg Loss: 0.48222288, Log Avg loss: 0.57596968, Global Avg Loss: 2.39027998, Time: 0.0208 Steps: 17580, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000558, Sample Num: 8928, Cur Loss: 0.43252981, Cur Avg Loss: 0.48155877, Log Avg loss: 0.44516557, Global Avg Loss: 2.38917418, Time: 0.0208 Steps: 17590, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000568, Sample Num: 9088, Cur Loss: 0.31549555, Cur Avg Loss: 0.47880022, Log Avg loss: 0.32487266, Global Avg Loss: 2.38800128, Time: 0.0208 Steps: 17600, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000578, Sample Num: 9248, Cur Loss: 0.33015773, Cur Avg Loss: 0.47801053, Log Avg loss: 0.43315613, Global Avg Loss: 2.38689120, Time: 0.0208 Steps: 17610, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000588, Sample Num: 9408, Cur Loss: 0.39886069, Cur Avg Loss: 0.47575484, Log Avg loss: 0.34537653, Global Avg Loss: 2.38573257, Time: 0.0208 Steps: 17620, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000598, Sample Num: 9568, Cur Loss: 0.71897364, Cur Avg Loss: 0.47747178, Log Avg loss: 0.57842778, Global Avg Loss: 2.38470743, Time: 0.0208 Steps: 17630, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000608, Sample Num: 9728, Cur Loss: 0.26465046, Cur Avg Loss: 0.47656187, Log Avg loss: 0.42214924, Global Avg Loss: 2.38359487, Time: 0.0209 Steps: 17640, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000618, Sample Num: 9888, Cur Loss: 0.22666323, Cur Avg Loss: 0.47500441, Log Avg loss: 0.38031064, Global Avg Loss: 2.38245987, Time: 0.0208 Steps: 17650, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000628, Sample Num: 10048, Cur Loss: 0.69396591, Cur Avg Loss: 0.47500521, Log Avg loss: 0.47505472, Global Avg Loss: 2.38137980, Time: 0.0208 Steps: 17660, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000638, Sample Num: 10208, Cur Loss: 0.40018588, Cur Avg Loss: 0.47497364, Log Avg loss: 0.47299128, Global Avg Loss: 2.38029978, Time: 0.0208 Steps: 17670, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000648, Sample Num: 10368, Cur Loss: 0.24262536, Cur Avg Loss: 0.47431873, Log Avg loss: 0.43253507, Global Avg Loss: 2.37919810, Time: 0.0208 Steps: 17680, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000658, Sample Num: 10528, Cur Loss: 1.12473702, Cur Avg Loss: 0.47728559, Log Avg loss: 0.66953789, Global Avg Loss: 2.37823165, Time: 0.0208 Steps: 17690, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000668, Sample Num: 10688, Cur Loss: 0.39140940, Cur Avg Loss: 0.48154485, Log Avg loss: 0.76180462, Global Avg Loss: 2.37731841, Time: 0.0208 Steps: 17700, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000678, Sample Num: 10848, Cur Loss: 0.08285207, Cur Avg Loss: 0.48035046, Log Avg loss: 0.40056519, Global Avg Loss: 2.37620223, Time: 0.0208 Steps: 17710, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000688, Sample Num: 11008, Cur Loss: 0.43782055, Cur Avg Loss: 0.48138575, Log Avg loss: 0.55157836, Global Avg Loss: 2.37517254, Time: 0.0208 Steps: 17720, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000698, Sample Num: 11168, Cur Loss: 1.00971079, Cur Avg Loss: 0.48121809, Log Avg loss: 0.46968298, Global Avg Loss: 2.37409781, Time: 0.0208 Steps: 17730, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000708, Sample Num: 11328, Cur Loss: 0.16619122, Cur Avg Loss: 0.48005305, Log Avg loss: 0.39873355, Global Avg Loss: 2.37298430, Time: 0.0208 Steps: 17740, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000718, Sample Num: 11488, Cur Loss: 0.20840818, Cur Avg Loss: 0.48101724, Log Avg loss: 0.54928160, Global Avg Loss: 2.37195686, Time: 0.0208 Steps: 17750, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000728, Sample Num: 11648, Cur Loss: 0.64444512, Cur Avg Loss: 0.48070193, Log Avg loss: 0.45806275, Global Avg Loss: 2.37087922, Time: 0.0208 Steps: 17760, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000738, Sample Num: 11808, Cur Loss: 0.17723744, Cur Avg Loss: 0.48180392, Log Avg loss: 0.56202905, Global Avg Loss: 2.36986130, Time: 0.0208 Steps: 17770, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000748, Sample Num: 11968, Cur Loss: 0.26103786, Cur Avg Loss: 0.48215897, Log Avg loss: 0.50836143, Global Avg Loss: 2.36881433, Time: 0.0208 Steps: 17780, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000758, Sample Num: 12128, Cur Loss: 0.48868871, Cur Avg Loss: 0.48126346, Log Avg loss: 0.41427959, Global Avg Loss: 2.36771566, Time: 0.0208 Steps: 17790, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000768, Sample Num: 12288, Cur Loss: 0.19547127, Cur Avg Loss: 0.47993935, Log Avg loss: 0.37957144, Global Avg Loss: 2.36659873, Time: 0.0253 Steps: 17800, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000778, Sample Num: 12448, Cur Loss: 0.44868729, Cur Avg Loss: 0.47942836, Log Avg loss: 0.44018436, Global Avg Loss: 2.36551708, Time: 0.0210 Steps: 17810, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000788, Sample Num: 12608, Cur Loss: 0.84994721, Cur Avg Loss: 0.48007500, Log Avg loss: 0.53038357, Global Avg Loss: 2.36448726, Time: 0.0208 Steps: 17820, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000798, Sample Num: 12768, Cur Loss: 0.88075221, Cur Avg Loss: 0.48192547, Log Avg loss: 0.62774213, Global Avg Loss: 2.36351321, Time: 0.0207 Steps: 17830, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000808, Sample Num: 12928, Cur Loss: 0.43422416, Cur Avg Loss: 0.48392360, Log Avg loss: 0.64337482, Global Avg Loss: 2.36254900, Time: 0.0208 Steps: 17840, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000818, Sample Num: 13088, Cur Loss: 0.15734959, Cur Avg Loss: 0.48359034, Log Avg loss: 0.45666327, Global Avg Loss: 2.36148128, Time: 0.0208 Steps: 17850, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000828, Sample Num: 13248, Cur Loss: 0.85502040, Cur Avg Loss: 0.48180205, Log Avg loss: 0.33551956, Global Avg Loss: 2.36034692, Time: 0.0208 Steps: 17860, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000838, Sample Num: 13408, Cur Loss: 0.32249993, Cur Avg Loss: 0.48252249, Log Avg loss: 0.54217460, Global Avg Loss: 2.35932948, Time: 0.0208 Steps: 17870, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000848, Sample Num: 13568, Cur Loss: 0.29639709, Cur Avg Loss: 0.48108281, Log Avg loss: 0.36043806, Global Avg Loss: 2.35821153, Time: 0.0208 Steps: 17880, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000858, Sample Num: 13728, Cur Loss: 0.56882668, Cur Avg Loss: 0.48207300, Log Avg loss: 0.56604126, Global Avg Loss: 2.35720976, Time: 0.0208 Steps: 17890, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000868, Sample Num: 13888, Cur Loss: 0.99052209, Cur Avg Loss: 0.48451166, Log Avg loss: 0.69374861, Global Avg Loss: 2.35628045, Time: 0.0208 Steps: 17900, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000878, Sample Num: 14048, Cur Loss: 0.71815252, Cur Avg Loss: 0.48894526, Log Avg loss: 0.87378164, Global Avg Loss: 2.35545270, Time: 0.0208 Steps: 17910, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000888, Sample Num: 14208, Cur Loss: 0.90201402, Cur Avg Loss: 0.48914030, Log Avg loss: 0.50626504, Global Avg Loss: 2.35442079, Time: 0.0208 Steps: 17920, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000898, Sample Num: 14368, Cur Loss: 0.52437204, Cur Avg Loss: 0.48805165, Log Avg loss: 0.39137892, Global Avg Loss: 2.35332595, Time: 0.0208 Steps: 17930, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000908, Sample Num: 14528, Cur Loss: 0.87655151, Cur Avg Loss: 0.48718369, Log Avg loss: 0.40924101, Global Avg Loss: 2.35224229, Time: 0.0208 Steps: 17940, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000918, Sample Num: 14688, Cur Loss: 0.55381429, Cur Avg Loss: 0.48606444, Log Avg loss: 0.38443685, Global Avg Loss: 2.35114602, Time: 0.0208 Steps: 17950, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000928, Sample Num: 14848, Cur Loss: 0.59953475, Cur Avg Loss: 0.48810042, Log Avg loss: 0.67500293, Global Avg Loss: 2.35021276, Time: 0.0208 Steps: 17960, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000938, Sample Num: 15008, Cur Loss: 0.84034288, Cur Avg Loss: 0.48864274, Log Avg loss: 0.53897002, Global Avg Loss: 2.34920483, Time: 0.0208 Steps: 17970, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000948, Sample Num: 15168, Cur Loss: 0.24604592, Cur Avg Loss: 0.48801116, Log Avg loss: 0.42876942, Global Avg Loss: 2.34813674, Time: 0.0208 Steps: 17980, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000958, Sample Num: 15328, Cur Loss: 0.38049504, Cur Avg Loss: 0.48661025, Log Avg loss: 0.35380377, Global Avg Loss: 2.34702816, Time: 0.0208 Steps: 17990, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000968, Sample Num: 15488, Cur Loss: 0.43735200, Cur Avg Loss: 0.48585403, Log Avg loss: 0.41340796, Global Avg Loss: 2.34595392, Time: 0.0208 Steps: 18000, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000978, Sample Num: 15648, Cur Loss: 0.30086976, Cur Avg Loss: 0.48472917, Log Avg loss: 0.37584302, Global Avg Loss: 2.34486003, Time: 0.0209 Steps: 18010, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000988, Sample Num: 15808, Cur Loss: 0.47283676, Cur Avg Loss: 0.48370199, Log Avg loss: 0.38324356, Global Avg Loss: 2.34377145, Time: 0.0208 Steps: 18020, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000998, Sample Num: 15968, Cur Loss: 0.63637388, Cur Avg Loss: 0.48321985, Log Avg loss: 0.43558495, Global Avg Loss: 2.34271311, Time: 0.0208 Steps: 18030, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001008, Sample Num: 16128, Cur Loss: 0.61708164, Cur Avg Loss: 0.48313113, Log Avg loss: 0.47427629, Global Avg Loss: 2.34167739, Time: 0.0209 Steps: 18040, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001018, Sample Num: 16288, Cur Loss: 0.18583190, Cur Avg Loss: 0.48228778, Log Avg loss: 0.39727781, Global Avg Loss: 2.34060016, Time: 0.0208 Steps: 18050, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001028, Sample Num: 16448, Cur Loss: 0.28352529, Cur Avg Loss: 0.48205750, Log Avg loss: 0.45861510, Global Avg Loss: 2.33955809, Time: 0.0246 Steps: 18060, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001038, Sample Num: 16608, Cur Loss: 1.50338626, Cur Avg Loss: 0.48217876, Log Avg loss: 0.49464418, Global Avg Loss: 2.33853711, Time: 0.0208 Steps: 18070, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001048, Sample Num: 16768, Cur Loss: 0.27435347, Cur Avg Loss: 0.48178651, Log Avg loss: 0.44107147, Global Avg Loss: 2.33748762, Time: 0.0208 Steps: 18080, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001058, Sample Num: 16928, Cur Loss: 0.59830517, Cur Avg Loss: 0.48203599, Log Avg loss: 0.50818196, Global Avg Loss: 2.33647640, Time: 0.0208 Steps: 18090, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001068, Sample Num: 17088, Cur Loss: 0.38661748, Cur Avg Loss: 0.48863734, Log Avg loss: 1.18706011, Global Avg Loss: 2.33584136, Time: 0.0208 Steps: 18100, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001078, Sample Num: 17248, Cur Loss: 0.38144600, Cur Avg Loss: 0.49047043, Log Avg loss: 0.68624374, Global Avg Loss: 2.33493048, Time: 0.0208 Steps: 18110, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001088, Sample Num: 17408, Cur Loss: 0.74176037, Cur Avg Loss: 0.48955757, Log Avg loss: 0.39115150, Global Avg Loss: 2.33385776, Time: 0.0208 Steps: 18120, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001098, Sample Num: 17568, Cur Loss: 0.58535731, Cur Avg Loss: 0.48942965, Log Avg loss: 0.47551213, Global Avg Loss: 2.33283275, Time: 0.0209 Steps: 18130, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001108, Sample Num: 17728, Cur Loss: 1.07247090, Cur Avg Loss: 0.49007963, Log Avg loss: 0.56144778, Global Avg Loss: 2.33185624, Time: 0.0208 Steps: 18140, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001118, Sample Num: 17888, Cur Loss: 0.43105668, Cur Avg Loss: 0.48965717, Log Avg loss: 0.44284825, Global Avg Loss: 2.33081546, Time: 0.0208 Steps: 18150, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001128, Sample Num: 18048, Cur Loss: 0.67156380, Cur Avg Loss: 0.48857578, Log Avg loss: 0.36767628, Global Avg Loss: 2.32973444, Time: 0.0208 Steps: 18160, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001138, Sample Num: 18208, Cur Loss: 0.29413259, Cur Avg Loss: 0.48828564, Log Avg loss: 0.45555781, Global Avg Loss: 2.32870297, Time: 0.0208 Steps: 18170, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001148, Sample Num: 18368, Cur Loss: 0.61545610, Cur Avg Loss: 0.48805420, Log Avg loss: 0.46171591, Global Avg Loss: 2.32767603, Time: 0.0208 Steps: 18180, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001158, Sample Num: 18528, Cur Loss: 0.61302441, Cur Avg Loss: 0.48816203, Log Avg loss: 0.50054113, Global Avg Loss: 2.32667155, Time: 0.0208 Steps: 18190, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001168, Sample Num: 18688, Cur Loss: 0.17791945, Cur Avg Loss: 0.48775372, Log Avg loss: 0.44047134, Global Avg Loss: 2.32563518, Time: 0.0208 Steps: 18200, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001178, Sample Num: 18848, Cur Loss: 0.37086046, Cur Avg Loss: 0.48697755, Log Avg loss: 0.39632161, Global Avg Loss: 2.32457570, Time: 0.0208 Steps: 18210, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001188, Sample Num: 19008, Cur Loss: 0.15679501, Cur Avg Loss: 0.48581295, Log Avg loss: 0.34862214, Global Avg Loss: 2.32349120, Time: 0.0208 Steps: 18220, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001198, Sample Num: 19168, Cur Loss: 0.16048072, Cur Avg Loss: 0.48722224, Log Avg loss: 0.65464697, Global Avg Loss: 2.32257577, Time: 0.0207 Steps: 18230, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001208, Sample Num: 19328, Cur Loss: 0.11301175, Cur Avg Loss: 0.48623672, Log Avg loss: 0.36817080, Global Avg Loss: 2.32150427, Time: 0.0208 Steps: 18240, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001218, Sample Num: 19488, Cur Loss: 0.56304872, Cur Avg Loss: 0.48598581, Log Avg loss: 0.45567651, Global Avg Loss: 2.32048190, Time: 0.0208 Steps: 18250, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001228, Sample Num: 19648, Cur Loss: 0.12153777, Cur Avg Loss: 0.48575056, Log Avg loss: 0.45709649, Global Avg Loss: 2.31946143, Time: 0.0208 Steps: 18260, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001238, Sample Num: 19808, Cur Loss: 0.73826194, Cur Avg Loss: 0.48459856, Log Avg loss: 0.34313342, Global Avg Loss: 2.31837969, Time: 0.0208 Steps: 18270, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001248, Sample Num: 19968, Cur Loss: 0.46986791, Cur Avg Loss: 0.48456307, Log Avg loss: 0.48016853, Global Avg Loss: 2.31737411, Time: 0.0208 Steps: 18280, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001258, Sample Num: 20128, Cur Loss: 0.99709487, Cur Avg Loss: 0.48476940, Log Avg loss: 0.51051980, Global Avg Loss: 2.31638621, Time: 0.0208 Steps: 18290, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001268, Sample Num: 20288, Cur Loss: 0.46102467, Cur Avg Loss: 0.48402129, Log Avg loss: 0.38990974, Global Avg Loss: 2.31533349, Time: 0.0209 Steps: 18300, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001278, Sample Num: 20448, Cur Loss: 0.17115167, Cur Avg Loss: 0.48385088, Log Avg loss: 0.46224235, Global Avg Loss: 2.31432143, Time: 0.0208 Steps: 18310, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001288, Sample Num: 20608, Cur Loss: 0.55147666, Cur Avg Loss: 0.48352534, Log Avg loss: 0.44192171, Global Avg Loss: 2.31329938, Time: 0.0209 Steps: 18320, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001298, Sample Num: 20768, Cur Loss: 0.52068585, Cur Avg Loss: 0.48316544, Log Avg loss: 0.43681019, Global Avg Loss: 2.31227565, Time: 0.0209 Steps: 18330, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001308, Sample Num: 20928, Cur Loss: 0.62624520, Cur Avg Loss: 0.48346271, Log Avg loss: 0.52204804, Global Avg Loss: 2.31129952, Time: 0.0208 Steps: 18340, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001318, Sample Num: 21088, Cur Loss: 0.25018889, Cur Avg Loss: 0.48621468, Log Avg loss: 0.84617238, Global Avg Loss: 2.31050108, Time: 0.0208 Steps: 18350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001328, Sample Num: 21248, Cur Loss: 0.64921075, Cur Avg Loss: 0.49067624, Log Avg loss: 1.07870999, Global Avg Loss: 2.30983017, Time: 0.0208 Steps: 18360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001338, Sample Num: 21408, Cur Loss: 0.29720879, Cur Avg Loss: 0.49310612, Log Avg loss: 0.81579426, Global Avg Loss: 2.30901687, Time: 0.0208 Steps: 18370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001348, Sample Num: 21568, Cur Loss: 0.46845269, Cur Avg Loss: 0.49356586, Log Avg loss: 0.55507944, Global Avg Loss: 2.30806261, Time: 0.0208 Steps: 18380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001358, Sample Num: 21728, Cur Loss: 0.34607136, Cur Avg Loss: 0.49366515, Log Avg loss: 0.50704904, Global Avg Loss: 2.30708326, Time: 0.0208 Steps: 18390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001368, Sample Num: 21888, Cur Loss: 0.56280607, Cur Avg Loss: 0.49400200, Log Avg loss: 0.53974612, Global Avg Loss: 2.30612275, Time: 0.0208 Steps: 18400, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001378, Sample Num: 22048, Cur Loss: 0.92216825, Cur Avg Loss: 0.49518905, Log Avg loss: 0.65757704, Global Avg Loss: 2.30522729, Time: 0.0208 Steps: 18410, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001388, Sample Num: 22208, Cur Loss: 1.41565466, Cur Avg Loss: 0.49606567, Log Avg loss: 0.61686405, Global Avg Loss: 2.30431070, Time: 0.0208 Steps: 18420, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001398, Sample Num: 22368, Cur Loss: 0.29495963, Cur Avg Loss: 0.49648050, Log Avg loss: 0.55405914, Global Avg Loss: 2.30336102, Time: 0.0208 Steps: 18430, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001408, Sample Num: 22528, Cur Loss: 0.59109437, Cur Avg Loss: 0.49625620, Log Avg loss: 0.46489882, Global Avg Loss: 2.30236403, Time: 0.0208 Steps: 18440, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001418, Sample Num: 22688, Cur Loss: 0.14376375, Cur Avg Loss: 0.49695337, Log Avg loss: 0.59511492, Global Avg Loss: 2.30143869, Time: 0.0208 Steps: 18450, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001428, Sample Num: 22848, Cur Loss: 0.50008279, Cur Avg Loss: 0.49778437, Log Avg loss: 0.61562035, Global Avg Loss: 2.30052546, Time: 0.0208 Steps: 18460, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001438, Sample Num: 23008, Cur Loss: 0.71476126, Cur Avg Loss: 0.49712196, Log Avg loss: 0.40252972, Global Avg Loss: 2.29949785, Time: 0.0208 Steps: 18470, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001448, Sample Num: 23168, Cur Loss: 0.83193409, Cur Avg Loss: 0.49717217, Log Avg loss: 0.50439218, Global Avg Loss: 2.29852647, Time: 0.0208 Steps: 18480, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001458, Sample Num: 23328, Cur Loss: 0.41304994, Cur Avg Loss: 0.49645855, Log Avg loss: 0.39312672, Global Avg Loss: 2.29749597, Time: 0.0208 Steps: 18490, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001468, Sample Num: 23488, Cur Loss: 0.39008546, Cur Avg Loss: 0.49566030, Log Avg loss: 0.37927488, Global Avg Loss: 2.29645910, Time: 0.0208 Steps: 18500, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001478, Sample Num: 23648, Cur Loss: 0.34723917, Cur Avg Loss: 0.49457126, Log Avg loss: 0.33470011, Global Avg Loss: 2.29539926, Time: 0.0208 Steps: 18510, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001488, Sample Num: 23808, Cur Loss: 0.41237175, Cur Avg Loss: 0.49495307, Log Avg loss: 0.55138481, Global Avg Loss: 2.29445757, Time: 0.0208 Steps: 18520, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001498, Sample Num: 23968, Cur Loss: 0.38325363, Cur Avg Loss: 0.49555752, Log Avg loss: 0.58549991, Global Avg Loss: 2.29353530, Time: 0.0208 Steps: 18530, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001508, Sample Num: 24128, Cur Loss: 0.51055557, Cur Avg Loss: 0.49407215, Log Avg loss: 0.27156395, Global Avg Loss: 2.29244470, Time: 0.0208 Steps: 18540, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001518, Sample Num: 24288, Cur Loss: 0.09586066, Cur Avg Loss: 0.49384690, Log Avg loss: 0.45987860, Global Avg Loss: 2.29145679, Time: 0.0208 Steps: 18550, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001528, Sample Num: 24448, Cur Loss: 0.46623346, Cur Avg Loss: 0.49285593, Log Avg loss: 0.34242778, Global Avg Loss: 2.29040667, Time: 0.0208 Steps: 18560, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001538, Sample Num: 24608, Cur Loss: 0.27058581, Cur Avg Loss: 0.49147086, Log Avg loss: 0.27983150, Global Avg Loss: 2.28932397, Time: 0.0245 Steps: 18570, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001548, Sample Num: 24768, Cur Loss: 0.85965759, Cur Avg Loss: 0.49272533, Log Avg loss: 0.68566217, Global Avg Loss: 2.28846086, Time: 0.0207 Steps: 18580, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001558, Sample Num: 24928, Cur Loss: 1.47169912, Cur Avg Loss: 0.49451161, Log Avg loss: 0.77102896, Global Avg Loss: 2.28764460, Time: 0.0207 Steps: 18590, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001568, Sample Num: 25088, Cur Loss: 0.84019256, Cur Avg Loss: 0.49541213, Log Avg loss: 0.63571222, Global Avg Loss: 2.28675646, Time: 0.0207 Steps: 18600, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001578, Sample Num: 25248, Cur Loss: 0.82883394, Cur Avg Loss: 0.49573579, Log Avg loss: 0.54648612, Global Avg Loss: 2.28582133, Time: 0.0207 Steps: 18610, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001588, Sample Num: 25408, Cur Loss: 0.35872987, Cur Avg Loss: 0.49563848, Log Avg loss: 0.48028287, Global Avg Loss: 2.28485166, Time: 0.0208 Steps: 18620, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001598, Sample Num: 25568, Cur Loss: 0.42049125, Cur Avg Loss: 0.49492608, Log Avg loss: 0.38179781, Global Avg Loss: 2.28383016, Time: 0.0207 Steps: 18630, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001608, Sample Num: 25728, Cur Loss: 0.15501110, Cur Avg Loss: 0.49519865, Log Avg loss: 0.53875403, Global Avg Loss: 2.28289396, Time: 0.0207 Steps: 18640, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001618, Sample Num: 25888, Cur Loss: 1.29819584, Cur Avg Loss: 0.49470084, Log Avg loss: 0.41465375, Global Avg Loss: 2.28189222, Time: 0.0207 Steps: 18650, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001628, Sample Num: 26048, Cur Loss: 0.40984255, Cur Avg Loss: 0.49434551, Log Avg loss: 0.43685250, Global Avg Loss: 2.28090345, Time: 0.0207 Steps: 18660, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001638, Sample Num: 26208, Cur Loss: 0.68015075, Cur Avg Loss: 0.49534522, Log Avg loss: 0.65809767, Global Avg Loss: 2.28003425, Time: 0.0207 Steps: 18670, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001648, Sample Num: 26368, Cur Loss: 0.45719892, Cur Avg Loss: 0.49611273, Log Avg loss: 0.62183189, Global Avg Loss: 2.27914656, Time: 0.0207 Steps: 18680, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001658, Sample Num: 26528, Cur Loss: 0.14976531, Cur Avg Loss: 0.49597156, Log Avg loss: 0.47270583, Global Avg Loss: 2.27818003, Time: 0.0207 Steps: 18690, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001668, Sample Num: 26688, Cur Loss: 0.81301820, Cur Avg Loss: 0.49566442, Log Avg loss: 0.44474147, Global Avg Loss: 2.27719958, Time: 0.0207 Steps: 18700, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001678, Sample Num: 26848, Cur Loss: 0.29064977, Cur Avg Loss: 0.49562041, Log Avg loss: 0.48827963, Global Avg Loss: 2.27624345, Time: 0.0208 Steps: 18710, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001688, Sample Num: 27008, Cur Loss: 0.72801840, Cur Avg Loss: 0.49529288, Log Avg loss: 0.44033239, Global Avg Loss: 2.27526273, Time: 0.0208 Steps: 18720, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001698, Sample Num: 27168, Cur Loss: 0.21939126, Cur Avg Loss: 0.49472173, Log Avg loss: 0.39831209, Global Avg Loss: 2.27426062, Time: 0.0210 Steps: 18730, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001708, Sample Num: 27328, Cur Loss: 0.18263558, Cur Avg Loss: 0.49349531, Log Avg loss: 0.28524955, Global Avg Loss: 2.27319925, Time: 0.0208 Steps: 18740, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001718, Sample Num: 27488, Cur Loss: 0.50570703, Cur Avg Loss: 0.49264739, Log Avg loss: 0.34782283, Global Avg Loss: 2.27217238, Time: 0.0208 Steps: 18750, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001728, Sample Num: 27648, Cur Loss: 0.57483160, Cur Avg Loss: 0.49296431, Log Avg loss: 0.54741092, Global Avg Loss: 2.27125300, Time: 0.0210 Steps: 18760, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001738, Sample Num: 27808, Cur Loss: 0.85744369, Cur Avg Loss: 0.49280411, Log Avg loss: 0.46512226, Global Avg Loss: 2.27029076, Time: 0.0208 Steps: 18770, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001748, Sample Num: 27968, Cur Loss: 0.49956211, Cur Avg Loss: 0.49311427, Log Avg loss: 0.54701846, Global Avg Loss: 2.26937315, Time: 0.0210 Steps: 18780, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001758, Sample Num: 28128, Cur Loss: 0.19308633, Cur Avg Loss: 0.49195936, Log Avg loss: 0.29008104, Global Avg Loss: 2.26831977, Time: 0.0208 Steps: 18790, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001768, Sample Num: 28288, Cur Loss: 0.54653329, Cur Avg Loss: 0.49161182, Log Avg loss: 0.43051457, Global Avg Loss: 2.26734221, Time: 0.0209 Steps: 18800, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001778, Sample Num: 28448, Cur Loss: 0.27740061, Cur Avg Loss: 0.49097121, Log Avg loss: 0.37771222, Global Avg Loss: 2.26633763, Time: 0.0210 Steps: 18810, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001788, Sample Num: 28608, Cur Loss: 0.65957290, Cur Avg Loss: 0.49076051, Log Avg loss: 0.45329730, Global Avg Loss: 2.26537427, Time: 0.0212 Steps: 18820, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001798, Sample Num: 28768, Cur Loss: 0.76183796, Cur Avg Loss: 0.49034270, Log Avg loss: 0.41563796, Global Avg Loss: 2.26439193, Time: 0.0209 Steps: 18830, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001808, Sample Num: 28928, Cur Loss: 0.44615906, Cur Avg Loss: 0.49034847, Log Avg loss: 0.49138579, Global Avg Loss: 2.26345085, Time: 0.0207 Steps: 18840, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001818, Sample Num: 29088, Cur Loss: 0.58319753, Cur Avg Loss: 0.49164345, Log Avg loss: 0.72577622, Global Avg Loss: 2.26263511, Time: 0.0208 Steps: 18850, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001828, Sample Num: 29248, Cur Loss: 0.52453727, Cur Avg Loss: 0.49148044, Log Avg loss: 0.46184519, Global Avg Loss: 2.26168029, Time: 0.0208 Steps: 18860, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001838, Sample Num: 29408, Cur Loss: 0.38259703, Cur Avg Loss: 0.49012577, Log Avg loss: 0.24249265, Global Avg Loss: 2.26061023, Time: 0.0207 Steps: 18870, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001848, Sample Num: 29568, Cur Loss: 0.45058733, Cur Avg Loss: 0.49012345, Log Avg loss: 0.48969662, Global Avg Loss: 2.25967225, Time: 0.0208 Steps: 18880, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001858, Sample Num: 29728, Cur Loss: 0.18160252, Cur Avg Loss: 0.49029227, Log Avg loss: 0.52149002, Global Avg Loss: 2.25875209, Time: 0.0208 Steps: 18890, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001868, Sample Num: 29888, Cur Loss: 0.14781633, Cur Avg Loss: 0.48938536, Log Avg loss: 0.32088220, Global Avg Loss: 2.25772676, Time: 0.0208 Steps: 18900, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001878, Sample Num: 30048, Cur Loss: 0.16523364, Cur Avg Loss: 0.48877491, Log Avg loss: 0.37474353, Global Avg Loss: 2.25673100, Time: 0.0208 Steps: 18910, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001888, Sample Num: 30208, Cur Loss: 0.56284428, Cur Avg Loss: 0.48901107, Log Avg loss: 0.53336155, Global Avg Loss: 2.25582013, Time: 0.0208 Steps: 18920, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001898, Sample Num: 30368, Cur Loss: 0.19209895, Cur Avg Loss: 0.48846880, Log Avg loss: 0.38608802, Global Avg Loss: 2.25483242, Time: 0.0208 Steps: 18930, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001908, Sample Num: 30528, Cur Loss: 0.17096454, Cur Avg Loss: 0.48780593, Log Avg loss: 0.36199302, Global Avg Loss: 2.25383303, Time: 0.0208 Steps: 18940, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001918, Sample Num: 30688, Cur Loss: 0.21134487, Cur Avg Loss: 0.48766444, Log Avg loss: 0.46066749, Global Avg Loss: 2.25288677, Time: 0.0208 Steps: 18950, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001928, Sample Num: 30848, Cur Loss: 0.81165284, Cur Avg Loss: 0.48775394, Log Avg loss: 0.50492001, Global Avg Loss: 2.25196485, Time: 0.0207 Steps: 18960, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001938, Sample Num: 31008, Cur Loss: 0.54115939, Cur Avg Loss: 0.48712804, Log Avg loss: 0.36645567, Global Avg Loss: 2.25097091, Time: 0.0207 Steps: 18970, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001948, Sample Num: 31168, Cur Loss: 0.28853175, Cur Avg Loss: 0.48607595, Log Avg loss: 0.28218063, Global Avg Loss: 2.24993361, Time: 0.0208 Steps: 18980, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001958, Sample Num: 31328, Cur Loss: 0.26316473, Cur Avg Loss: 0.48629315, Log Avg loss: 0.52860369, Global Avg Loss: 2.24902717, Time: 0.0207 Steps: 18990, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001968, Sample Num: 31488, Cur Loss: 0.31348574, Cur Avg Loss: 0.48572411, Log Avg loss: 0.37430604, Global Avg Loss: 2.24804047, Time: 0.0208 Steps: 19000, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001978, Sample Num: 31648, Cur Loss: 0.10869388, Cur Avg Loss: 0.48499326, Log Avg loss: 0.34116179, Global Avg Loss: 2.24703738, Time: 0.0207 Steps: 19010, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001988, Sample Num: 31808, Cur Loss: 0.41647068, Cur Avg Loss: 0.48564557, Log Avg loss: 0.61467315, Global Avg Loss: 2.24617915, Time: 0.0207 Steps: 19020, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001998, Sample Num: 31968, Cur Loss: 0.47204655, Cur Avg Loss: 0.48695875, Log Avg loss: 0.74801799, Global Avg Loss: 2.24539188, Time: 0.0208 Steps: 19030, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002008, Sample Num: 32128, Cur Loss: 0.51918483, Cur Avg Loss: 0.48660408, Log Avg loss: 0.41574131, Global Avg Loss: 2.24443093, Time: 0.0208 Steps: 19040, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002018, Sample Num: 32288, Cur Loss: 1.15064311, Cur Avg Loss: 0.48640579, Log Avg loss: 0.44658980, Global Avg Loss: 2.24348718, Time: 0.0207 Steps: 19050, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002028, Sample Num: 32448, Cur Loss: 0.20164049, Cur Avg Loss: 0.48612064, Log Avg loss: 0.42857758, Global Avg Loss: 2.24253497, Time: 0.0207 Steps: 19060, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002038, Sample Num: 32608, Cur Loss: 0.30129230, Cur Avg Loss: 0.48596242, Log Avg loss: 0.45387436, Global Avg Loss: 2.24159703, Time: 0.0208 Steps: 19070, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002048, Sample Num: 32768, Cur Loss: 0.21512210, Cur Avg Loss: 0.48570838, Log Avg loss: 0.43393605, Global Avg Loss: 2.24064962, Time: 0.0253 Steps: 19080, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002058, Sample Num: 32928, Cur Loss: 0.59802926, Cur Avg Loss: 0.48504397, Log Avg loss: 0.34897243, Global Avg Loss: 2.23965869, Time: 0.0207 Steps: 19090, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002068, Sample Num: 33088, Cur Loss: 0.58337182, Cur Avg Loss: 0.48465597, Log Avg loss: 0.40480411, Global Avg Loss: 2.23869804, Time: 0.0208 Steps: 19100, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002078, Sample Num: 33248, Cur Loss: 0.20173281, Cur Avg Loss: 0.48434879, Log Avg loss: 0.42082465, Global Avg Loss: 2.23774677, Time: 0.0207 Steps: 19110, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002088, Sample Num: 33408, Cur Loss: 0.15241857, Cur Avg Loss: 0.48351240, Log Avg loss: 0.30971053, Global Avg Loss: 2.23673838, Time: 0.0207 Steps: 19120, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002098, Sample Num: 33568, Cur Loss: 0.25076663, Cur Avg Loss: 0.48360890, Log Avg loss: 0.50375737, Global Avg Loss: 2.23583248, Time: 0.0207 Steps: 19130, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002108, Sample Num: 33728, Cur Loss: 0.35936615, Cur Avg Loss: 0.48331783, Log Avg loss: 0.42225258, Global Avg Loss: 2.23488495, Time: 0.0207 Steps: 19140, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002118, Sample Num: 33888, Cur Loss: 0.32565224, Cur Avg Loss: 0.48280892, Log Avg loss: 0.37553083, Global Avg Loss: 2.23391401, Time: 0.0207 Steps: 19150, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002128, Sample Num: 34048, Cur Loss: 0.24186112, Cur Avg Loss: 0.48224272, Log Avg loss: 0.36232095, Global Avg Loss: 2.23293718, Time: 0.0208 Steps: 19160, Updated lr: 0.000083 ***** Running evaluation checkpoint-19161 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-19161 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.520437, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.3835, "eval_total_loss": 269.600848, "eval_mae": 0.454102, "eval_mse": 0.383629, "eval_r2": 0.75614, "eval_sp_statistic": 0.828911, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.872733, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.394509, "test_total_loss": 198.043713, "test_mae": 0.430747, "test_mse": 0.394644, "test_r2": 0.745293, "test_sp_statistic": 0.827169, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.882385, "test_ps_pvalue": 0.0, "lr": 8.277761972498815e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.232834449283171, "train_cur_epoch_loss": 1026.4769354984164, "train_cur_epoch_avg_loss": 0.48214041122518386, "train_cur_epoch_time": 44.520437240600586, "train_cur_epoch_avg_time": 0.020911431301362415, "epoch": 9, "step": 19161} ################################################## Training, Epoch: 0010, Batch: 000009, Sample Num: 144, Cur Loss: 0.15976916, Cur Avg Loss: 0.42680221, Log Avg loss: 0.41056498, Global Avg Loss: 2.23198655, Time: 0.0209 Steps: 19170, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000019, Sample Num: 304, Cur Loss: 0.32001144, Cur Avg Loss: 0.58017920, Log Avg loss: 0.71821850, Global Avg Loss: 2.23119730, Time: 0.0208 Steps: 19180, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000029, Sample Num: 464, Cur Loss: 0.47411340, Cur Avg Loss: 0.52334383, Log Avg loss: 0.41535663, Global Avg Loss: 2.23025106, Time: 0.0209 Steps: 19190, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000039, Sample Num: 624, Cur Loss: 0.19554198, Cur Avg Loss: 0.48527257, Log Avg loss: 0.37486592, Global Avg Loss: 2.22928471, Time: 0.0209 Steps: 19200, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000049, Sample Num: 784, Cur Loss: 0.22856465, Cur Avg Loss: 0.46539335, Log Avg loss: 0.38786436, Global Avg Loss: 2.22832614, Time: 0.0208 Steps: 19210, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000059, Sample Num: 944, Cur Loss: 0.50778210, Cur Avg Loss: 0.45953495, Log Avg loss: 0.43082878, Global Avg Loss: 2.22739092, Time: 0.0209 Steps: 19220, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000069, Sample Num: 1104, Cur Loss: 0.49796918, Cur Avg Loss: 0.47997393, Log Avg loss: 0.60056393, Global Avg Loss: 2.22654493, Time: 0.0209 Steps: 19230, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000079, Sample Num: 1264, Cur Loss: 0.18491054, Cur Avg Loss: 0.47168210, Log Avg loss: 0.41446850, Global Avg Loss: 2.22560311, Time: 0.0209 Steps: 19240, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000089, Sample Num: 1424, Cur Loss: 0.93422616, Cur Avg Loss: 0.46247829, Log Avg loss: 0.38976819, Global Avg Loss: 2.22464943, Time: 0.0208 Steps: 19250, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000099, Sample Num: 1584, Cur Loss: 0.42705646, Cur Avg Loss: 0.47125333, Log Avg loss: 0.54935117, Global Avg Loss: 2.22377959, Time: 0.0208 Steps: 19260, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000109, Sample Num: 1744, Cur Loss: 0.47335917, Cur Avg Loss: 0.46418221, Log Avg loss: 0.39417814, Global Avg Loss: 2.22283014, Time: 0.0209 Steps: 19270, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000119, Sample Num: 1904, Cur Loss: 0.26038134, Cur Avg Loss: 0.44976510, Log Avg loss: 0.29261858, Global Avg Loss: 2.22182899, Time: 0.0209 Steps: 19280, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000129, Sample Num: 2064, Cur Loss: 0.21425846, Cur Avg Loss: 0.44662777, Log Avg loss: 0.40929358, Global Avg Loss: 2.22088937, Time: 0.0209 Steps: 19290, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000139, Sample Num: 2224, Cur Loss: 0.20721868, Cur Avg Loss: 0.44017102, Log Avg loss: 0.35687888, Global Avg Loss: 2.21992356, Time: 0.0209 Steps: 19300, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000149, Sample Num: 2384, Cur Loss: 0.20934364, Cur Avg Loss: 0.43859995, Log Avg loss: 0.41676211, Global Avg Loss: 2.21898976, Time: 0.0208 Steps: 19310, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000159, Sample Num: 2544, Cur Loss: 0.52771449, Cur Avg Loss: 0.43745797, Log Avg loss: 0.42044251, Global Avg Loss: 2.21805884, Time: 0.0209 Steps: 19320, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000169, Sample Num: 2704, Cur Loss: 0.29358873, Cur Avg Loss: 0.43699108, Log Avg loss: 0.42956753, Global Avg Loss: 2.21713359, Time: 0.0209 Steps: 19330, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000179, Sample Num: 2864, Cur Loss: 0.10997809, Cur Avg Loss: 0.43440324, Log Avg loss: 0.39066862, Global Avg Loss: 2.21618920, Time: 0.0209 Steps: 19340, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000189, Sample Num: 3024, Cur Loss: 0.33412713, Cur Avg Loss: 0.43400301, Log Avg loss: 0.42683895, Global Avg Loss: 2.21526447, Time: 0.0208 Steps: 19350, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000199, Sample Num: 3184, Cur Loss: 0.37806696, Cur Avg Loss: 0.43835759, Log Avg loss: 0.52065909, Global Avg Loss: 2.21438916, Time: 0.0209 Steps: 19360, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000209, Sample Num: 3344, Cur Loss: 0.42837414, Cur Avg Loss: 0.43470167, Log Avg loss: 0.36194900, Global Avg Loss: 2.21343281, Time: 0.0209 Steps: 19370, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000219, Sample Num: 3504, Cur Loss: 0.39474833, Cur Avg Loss: 0.42991940, Log Avg loss: 0.32996981, Global Avg Loss: 2.21246095, Time: 0.0208 Steps: 19380, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000229, Sample Num: 3664, Cur Loss: 0.43581933, Cur Avg Loss: 0.42673757, Log Avg loss: 0.35705565, Global Avg Loss: 2.21150406, Time: 0.0208 Steps: 19390, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000239, Sample Num: 3824, Cur Loss: 0.86020339, Cur Avg Loss: 0.42655741, Log Avg loss: 0.42243169, Global Avg Loss: 2.21058186, Time: 0.0209 Steps: 19400, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000249, Sample Num: 3984, Cur Loss: 0.42177856, Cur Avg Loss: 0.43851063, Log Avg loss: 0.72419264, Global Avg Loss: 2.20981608, Time: 0.0209 Steps: 19410, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000259, Sample Num: 4144, Cur Loss: 0.13332169, Cur Avg Loss: 0.43694913, Log Avg loss: 0.39806767, Global Avg Loss: 2.20888315, Time: 0.0242 Steps: 19420, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000269, Sample Num: 4304, Cur Loss: 0.08324652, Cur Avg Loss: 0.43610649, Log Avg loss: 0.41428224, Global Avg Loss: 2.20795952, Time: 0.0208 Steps: 19430, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000279, Sample Num: 4464, Cur Loss: 0.36969912, Cur Avg Loss: 0.43072783, Log Avg loss: 0.28604173, Global Avg Loss: 2.20697088, Time: 0.0208 Steps: 19440, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000289, Sample Num: 4624, Cur Loss: 0.49028739, Cur Avg Loss: 0.42876815, Log Avg loss: 0.37409321, Global Avg Loss: 2.20602853, Time: 0.0209 Steps: 19450, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000299, Sample Num: 4784, Cur Loss: 0.24606661, Cur Avg Loss: 0.42615524, Log Avg loss: 0.35064218, Global Avg Loss: 2.20507509, Time: 0.0208 Steps: 19460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000309, Sample Num: 4944, Cur Loss: 0.29820371, Cur Avg Loss: 0.43516911, Log Avg loss: 0.70468363, Global Avg Loss: 2.20430448, Time: 0.0207 Steps: 19470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000319, Sample Num: 5104, Cur Loss: 0.15714395, Cur Avg Loss: 0.43109026, Log Avg loss: 0.30505403, Global Avg Loss: 2.20332950, Time: 0.0208 Steps: 19480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000329, Sample Num: 5264, Cur Loss: 0.21684270, Cur Avg Loss: 0.43470233, Log Avg loss: 0.54992733, Global Avg Loss: 2.20248117, Time: 0.0208 Steps: 19490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000339, Sample Num: 5424, Cur Loss: 0.11419784, Cur Avg Loss: 0.43836416, Log Avg loss: 0.55883831, Global Avg Loss: 2.20163827, Time: 0.0207 Steps: 19500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000349, Sample Num: 5584, Cur Loss: 0.32726216, Cur Avg Loss: 0.44149781, Log Avg loss: 0.54772849, Global Avg Loss: 2.20079055, Time: 0.0208 Steps: 19510, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000359, Sample Num: 5744, Cur Loss: 1.06668401, Cur Avg Loss: 0.44294561, Log Avg loss: 0.49347368, Global Avg Loss: 2.19991590, Time: 0.0208 Steps: 19520, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000369, Sample Num: 5904, Cur Loss: 0.47527403, Cur Avg Loss: 0.44582316, Log Avg loss: 0.54912732, Global Avg Loss: 2.19907064, Time: 0.0207 Steps: 19530, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000379, Sample Num: 6064, Cur Loss: 0.42592430, Cur Avg Loss: 0.44448071, Log Avg loss: 0.39494425, Global Avg Loss: 2.19814734, Time: 0.0208 Steps: 19540, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000389, Sample Num: 6224, Cur Loss: 0.29089063, Cur Avg Loss: 0.44469467, Log Avg loss: 0.45280389, Global Avg Loss: 2.19725458, Time: 0.0208 Steps: 19550, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000399, Sample Num: 6384, Cur Loss: 0.62404877, Cur Avg Loss: 0.44301593, Log Avg loss: 0.37771295, Global Avg Loss: 2.19632435, Time: 0.0208 Steps: 19560, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000409, Sample Num: 6544, Cur Loss: 0.45438462, Cur Avg Loss: 0.44595915, Log Avg loss: 0.56339352, Global Avg Loss: 2.19548994, Time: 0.0209 Steps: 19570, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000419, Sample Num: 6704, Cur Loss: 0.47071514, Cur Avg Loss: 0.44487399, Log Avg loss: 0.40049102, Global Avg Loss: 2.19457319, Time: 0.0208 Steps: 19580, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000429, Sample Num: 6864, Cur Loss: 0.24428488, Cur Avg Loss: 0.44455856, Log Avg loss: 0.43134207, Global Avg Loss: 2.19367312, Time: 0.0208 Steps: 19590, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000439, Sample Num: 7024, Cur Loss: 1.07411087, Cur Avg Loss: 0.44970365, Log Avg loss: 0.67042797, Global Avg Loss: 2.19289596, Time: 0.0208 Steps: 19600, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000449, Sample Num: 7184, Cur Loss: 0.31869745, Cur Avg Loss: 0.45032932, Log Avg loss: 0.47779635, Global Avg Loss: 2.19202135, Time: 0.0208 Steps: 19610, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000459, Sample Num: 7344, Cur Loss: 0.12458280, Cur Avg Loss: 0.44876191, Log Avg loss: 0.37838510, Global Avg Loss: 2.19109697, Time: 0.0209 Steps: 19620, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000469, Sample Num: 7504, Cur Loss: 0.36016476, Cur Avg Loss: 0.44551362, Log Avg loss: 0.29641721, Global Avg Loss: 2.19013178, Time: 0.0208 Steps: 19630, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000479, Sample Num: 7664, Cur Loss: 0.44725212, Cur Avg Loss: 0.44581860, Log Avg loss: 0.46012212, Global Avg Loss: 2.18925092, Time: 0.0208 Steps: 19640, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000489, Sample Num: 7824, Cur Loss: 0.82268822, Cur Avg Loss: 0.44483887, Log Avg loss: 0.39790968, Global Avg Loss: 2.18833929, Time: 0.0208 Steps: 19650, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000499, Sample Num: 7984, Cur Loss: 0.58377093, Cur Avg Loss: 0.44308770, Log Avg loss: 0.35745549, Global Avg Loss: 2.18740802, Time: 0.0208 Steps: 19660, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000509, Sample Num: 8144, Cur Loss: 1.60369253, Cur Avg Loss: 0.45196775, Log Avg loss: 0.89508218, Global Avg Loss: 2.18675102, Time: 0.0208 Steps: 19670, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000519, Sample Num: 8304, Cur Loss: 0.37922579, Cur Avg Loss: 0.45245208, Log Avg loss: 0.47710446, Global Avg Loss: 2.18588229, Time: 0.0209 Steps: 19680, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000529, Sample Num: 8464, Cur Loss: 0.74719197, Cur Avg Loss: 0.45371732, Log Avg loss: 0.51938354, Global Avg Loss: 2.18503592, Time: 0.0208 Steps: 19690, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000539, Sample Num: 8624, Cur Loss: 0.54179931, Cur Avg Loss: 0.45419860, Log Avg loss: 0.47965822, Global Avg Loss: 2.18417025, Time: 0.0208 Steps: 19700, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000549, Sample Num: 8784, Cur Loss: 0.13438536, Cur Avg Loss: 0.45345264, Log Avg loss: 0.41324557, Global Avg Loss: 2.18327176, Time: 0.0208 Steps: 19710, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000559, Sample Num: 8944, Cur Loss: 0.61006939, Cur Avg Loss: 0.45269971, Log Avg loss: 0.41136379, Global Avg Loss: 2.18237323, Time: 0.0208 Steps: 19720, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000569, Sample Num: 9104, Cur Loss: 1.09092760, Cur Avg Loss: 0.45496752, Log Avg loss: 0.58173790, Global Avg Loss: 2.18156196, Time: 0.0207 Steps: 19730, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000579, Sample Num: 9264, Cur Loss: 1.82778692, Cur Avg Loss: 0.45905718, Log Avg loss: 0.69175881, Global Avg Loss: 2.18080724, Time: 0.0208 Steps: 19740, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000589, Sample Num: 9424, Cur Loss: 0.68932545, Cur Avg Loss: 0.45924806, Log Avg loss: 0.47029996, Global Avg Loss: 2.17994116, Time: 0.0208 Steps: 19750, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000599, Sample Num: 9584, Cur Loss: 0.50402164, Cur Avg Loss: 0.45744778, Log Avg loss: 0.35141162, Global Avg Loss: 2.17901579, Time: 0.0208 Steps: 19760, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000609, Sample Num: 9744, Cur Loss: 0.58015066, Cur Avg Loss: 0.45468895, Log Avg loss: 0.28943478, Global Avg Loss: 2.17806001, Time: 0.0210 Steps: 19770, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000619, Sample Num: 9904, Cur Loss: 0.23343363, Cur Avg Loss: 0.45238838, Log Avg loss: 0.31228342, Global Avg Loss: 2.17711675, Time: 0.0209 Steps: 19780, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000629, Sample Num: 10064, Cur Loss: 0.17018083, Cur Avg Loss: 0.45019676, Log Avg loss: 0.31453590, Global Avg Loss: 2.17617558, Time: 0.0209 Steps: 19790, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000639, Sample Num: 10224, Cur Loss: 0.55378354, Cur Avg Loss: 0.45160811, Log Avg loss: 0.54038184, Global Avg Loss: 2.17534942, Time: 0.0209 Steps: 19800, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000649, Sample Num: 10384, Cur Loss: 0.35222232, Cur Avg Loss: 0.45115445, Log Avg loss: 0.42216581, Global Avg Loss: 2.17446442, Time: 0.0209 Steps: 19810, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000659, Sample Num: 10544, Cur Loss: 0.48891455, Cur Avg Loss: 0.45006794, Log Avg loss: 0.37955342, Global Avg Loss: 2.17355881, Time: 0.0209 Steps: 19820, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000669, Sample Num: 10704, Cur Loss: 0.88752449, Cur Avg Loss: 0.44980682, Log Avg loss: 0.43259853, Global Avg Loss: 2.17268087, Time: 0.0209 Steps: 19830, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000679, Sample Num: 10864, Cur Loss: 0.27534717, Cur Avg Loss: 0.44921367, Log Avg loss: 0.40953241, Global Avg Loss: 2.17179219, Time: 0.0209 Steps: 19840, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000689, Sample Num: 11024, Cur Loss: 0.29286578, Cur Avg Loss: 0.44789982, Log Avg loss: 0.35868904, Global Avg Loss: 2.17087878, Time: 0.0209 Steps: 19850, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000699, Sample Num: 11184, Cur Loss: 0.40677413, Cur Avg Loss: 0.44621301, Log Avg loss: 0.32999177, Global Avg Loss: 2.16995185, Time: 0.0209 Steps: 19860, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000709, Sample Num: 11344, Cur Loss: 0.19634636, Cur Avg Loss: 0.44515670, Log Avg loss: 0.37132076, Global Avg Loss: 2.16904665, Time: 0.0209 Steps: 19870, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000719, Sample Num: 11504, Cur Loss: 0.48857686, Cur Avg Loss: 0.44725113, Log Avg loss: 0.59574642, Global Avg Loss: 2.16825525, Time: 0.0209 Steps: 19880, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000729, Sample Num: 11664, Cur Loss: 0.16888231, Cur Avg Loss: 0.44789526, Log Avg loss: 0.49420795, Global Avg Loss: 2.16741360, Time: 0.0209 Steps: 19890, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000739, Sample Num: 11824, Cur Loss: 0.88907719, Cur Avg Loss: 0.44986557, Log Avg loss: 0.59350108, Global Avg Loss: 2.16662269, Time: 0.0209 Steps: 19900, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000749, Sample Num: 11984, Cur Loss: 0.36014855, Cur Avg Loss: 0.44738240, Log Avg loss: 0.26387618, Global Avg Loss: 2.16566702, Time: 0.0209 Steps: 19910, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000759, Sample Num: 12144, Cur Loss: 0.50897163, Cur Avg Loss: 0.44591202, Log Avg loss: 0.33578084, Global Avg Loss: 2.16474840, Time: 0.0209 Steps: 19920, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000769, Sample Num: 12304, Cur Loss: 0.44746137, Cur Avg Loss: 0.44698797, Log Avg loss: 0.52865252, Global Avg Loss: 2.16392748, Time: 0.0246 Steps: 19930, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000779, Sample Num: 12464, Cur Loss: 0.95482540, Cur Avg Loss: 0.44864283, Log Avg loss: 0.57590187, Global Avg Loss: 2.16313108, Time: 0.0209 Steps: 19940, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000789, Sample Num: 12624, Cur Loss: 0.41795480, Cur Avg Loss: 0.45052356, Log Avg loss: 0.59703216, Global Avg Loss: 2.16234606, Time: 0.0209 Steps: 19950, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000799, Sample Num: 12784, Cur Loss: 0.25383168, Cur Avg Loss: 0.44907158, Log Avg loss: 0.33451018, Global Avg Loss: 2.16143031, Time: 0.0209 Steps: 19960, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000809, Sample Num: 12944, Cur Loss: 0.17205578, Cur Avg Loss: 0.44635949, Log Avg loss: 0.22966390, Global Avg Loss: 2.16046298, Time: 0.0209 Steps: 19970, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000819, Sample Num: 13104, Cur Loss: 0.21332566, Cur Avg Loss: 0.44678299, Log Avg loss: 0.48104364, Global Avg Loss: 2.15962243, Time: 0.0209 Steps: 19980, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000829, Sample Num: 13264, Cur Loss: 0.60219246, Cur Avg Loss: 0.44859979, Log Avg loss: 0.59739603, Global Avg Loss: 2.15884093, Time: 0.0209 Steps: 19990, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000839, Sample Num: 13424, Cur Loss: 0.65613794, Cur Avg Loss: 0.44838798, Log Avg loss: 0.43082883, Global Avg Loss: 2.15797692, Time: 0.0209 Steps: 20000, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000849, Sample Num: 13584, Cur Loss: 0.59988922, Cur Avg Loss: 0.44770578, Log Avg loss: 0.39046899, Global Avg Loss: 2.15709361, Time: 0.0209 Steps: 20010, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000859, Sample Num: 13744, Cur Loss: 0.21397407, Cur Avg Loss: 0.44633386, Log Avg loss: 0.32985794, Global Avg Loss: 2.15618090, Time: 0.0209 Steps: 20020, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000869, Sample Num: 13904, Cur Loss: 0.42726672, Cur Avg Loss: 0.44682948, Log Avg loss: 0.48940296, Global Avg Loss: 2.15534876, Time: 0.0209 Steps: 20030, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000879, Sample Num: 14064, Cur Loss: 0.34192729, Cur Avg Loss: 0.44466889, Log Avg loss: 0.25691418, Global Avg Loss: 2.15440144, Time: 0.0209 Steps: 20040, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000889, Sample Num: 14224, Cur Loss: 0.14968459, Cur Avg Loss: 0.44435216, Log Avg loss: 0.41651152, Global Avg Loss: 2.15353466, Time: 0.0209 Steps: 20050, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000899, Sample Num: 14384, Cur Loss: 0.24861184, Cur Avg Loss: 0.44462215, Log Avg loss: 0.46862372, Global Avg Loss: 2.15269473, Time: 0.0209 Steps: 20060, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000909, Sample Num: 14544, Cur Loss: 0.13359222, Cur Avg Loss: 0.44430986, Log Avg loss: 0.41623581, Global Avg Loss: 2.15182952, Time: 0.0209 Steps: 20070, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000919, Sample Num: 14704, Cur Loss: 0.35163122, Cur Avg Loss: 0.44455954, Log Avg loss: 0.46725537, Global Avg Loss: 2.15099059, Time: 0.0209 Steps: 20080, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000929, Sample Num: 14864, Cur Loss: 0.57845682, Cur Avg Loss: 0.44524792, Log Avg loss: 0.50850935, Global Avg Loss: 2.15017303, Time: 0.0209 Steps: 20090, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000939, Sample Num: 15024, Cur Loss: 0.36115050, Cur Avg Loss: 0.44406015, Log Avg loss: 0.33371626, Global Avg Loss: 2.14926932, Time: 0.0209 Steps: 20100, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000949, Sample Num: 15184, Cur Loss: 0.29630500, Cur Avg Loss: 0.44386901, Log Avg loss: 0.42592182, Global Avg Loss: 2.14841236, Time: 0.0209 Steps: 20110, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000959, Sample Num: 15344, Cur Loss: 0.16935998, Cur Avg Loss: 0.44357338, Log Avg loss: 0.41551811, Global Avg Loss: 2.14755108, Time: 0.0210 Steps: 20120, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000969, Sample Num: 15504, Cur Loss: 0.35905370, Cur Avg Loss: 0.44324418, Log Avg loss: 0.41167342, Global Avg Loss: 2.14668875, Time: 0.0209 Steps: 20130, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000979, Sample Num: 15664, Cur Loss: 0.85014904, Cur Avg Loss: 0.44684010, Log Avg loss: 0.79528472, Global Avg Loss: 2.14601774, Time: 0.0209 Steps: 20140, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000989, Sample Num: 15824, Cur Loss: 0.93099087, Cur Avg Loss: 0.44770875, Log Avg loss: 0.53274999, Global Avg Loss: 2.14521711, Time: 0.0209 Steps: 20150, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000999, Sample Num: 15984, Cur Loss: 0.90201175, Cur Avg Loss: 0.44912106, Log Avg loss: 0.58879792, Global Avg Loss: 2.14444508, Time: 0.0209 Steps: 20160, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001009, Sample Num: 16144, Cur Loss: 0.87636608, Cur Avg Loss: 0.44941453, Log Avg loss: 0.47873251, Global Avg Loss: 2.14361924, Time: 0.0209 Steps: 20170, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001019, Sample Num: 16304, Cur Loss: 0.77574241, Cur Avg Loss: 0.45640910, Log Avg loss: 1.16216085, Global Avg Loss: 2.14313289, Time: 0.0209 Steps: 20180, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001029, Sample Num: 16464, Cur Loss: 0.87787843, Cur Avg Loss: 0.45949815, Log Avg loss: 0.77427231, Global Avg Loss: 2.14245490, Time: 0.0210 Steps: 20190, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001039, Sample Num: 16624, Cur Loss: 0.88498956, Cur Avg Loss: 0.46136723, Log Avg loss: 0.65369595, Global Avg Loss: 2.14171789, Time: 0.0209 Steps: 20200, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001049, Sample Num: 16784, Cur Loss: 0.28088936, Cur Avg Loss: 0.46064421, Log Avg loss: 0.38552245, Global Avg Loss: 2.14084892, Time: 0.0209 Steps: 20210, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001059, Sample Num: 16944, Cur Loss: 0.82962739, Cur Avg Loss: 0.46071491, Log Avg loss: 0.46813101, Global Avg Loss: 2.14002166, Time: 0.0208 Steps: 20220, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001069, Sample Num: 17104, Cur Loss: 0.18079400, Cur Avg Loss: 0.45946004, Log Avg loss: 0.32656971, Global Avg Loss: 2.13912524, Time: 0.0209 Steps: 20230, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001079, Sample Num: 17264, Cur Loss: 0.50032628, Cur Avg Loss: 0.45786928, Log Avg loss: 0.28781670, Global Avg Loss: 2.13821056, Time: 0.0209 Steps: 20240, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001089, Sample Num: 17424, Cur Loss: 0.19672985, Cur Avg Loss: 0.45652653, Log Avg loss: 0.31164348, Global Avg Loss: 2.13730856, Time: 0.0209 Steps: 20250, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001099, Sample Num: 17584, Cur Loss: 0.19328627, Cur Avg Loss: 0.45528338, Log Avg loss: 0.31990518, Global Avg Loss: 2.13641152, Time: 0.0209 Steps: 20260, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001109, Sample Num: 17744, Cur Loss: 0.46655115, Cur Avg Loss: 0.45643940, Log Avg loss: 0.58348593, Global Avg Loss: 2.13564540, Time: 0.0209 Steps: 20270, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001119, Sample Num: 17904, Cur Loss: 0.78790891, Cur Avg Loss: 0.45622833, Log Avg loss: 0.43282017, Global Avg Loss: 2.13480574, Time: 0.0211 Steps: 20280, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001129, Sample Num: 18064, Cur Loss: 0.75456250, Cur Avg Loss: 0.45558009, Log Avg loss: 0.38304246, Global Avg Loss: 2.13394238, Time: 0.0209 Steps: 20290, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001139, Sample Num: 18224, Cur Loss: 0.44033682, Cur Avg Loss: 0.46019590, Log Avg loss: 0.98132094, Global Avg Loss: 2.13337458, Time: 0.0209 Steps: 20300, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001149, Sample Num: 18384, Cur Loss: 0.24688217, Cur Avg Loss: 0.46188777, Log Avg loss: 0.65459111, Global Avg Loss: 2.13264648, Time: 0.0208 Steps: 20310, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001159, Sample Num: 18544, Cur Loss: 0.60048735, Cur Avg Loss: 0.46231561, Log Avg loss: 0.51147493, Global Avg Loss: 2.13184866, Time: 0.0209 Steps: 20320, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001169, Sample Num: 18704, Cur Loss: 0.88244355, Cur Avg Loss: 0.46282867, Log Avg loss: 0.52229240, Global Avg Loss: 2.13105694, Time: 0.0209 Steps: 20330, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001179, Sample Num: 18864, Cur Loss: 0.43979496, Cur Avg Loss: 0.46156152, Log Avg loss: 0.31343135, Global Avg Loss: 2.13016332, Time: 0.0209 Steps: 20340, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001189, Sample Num: 19024, Cur Loss: 0.41851348, Cur Avg Loss: 0.46094791, Log Avg loss: 0.38860317, Global Avg Loss: 2.12930752, Time: 0.0209 Steps: 20350, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001199, Sample Num: 19184, Cur Loss: 0.38728821, Cur Avg Loss: 0.46007086, Log Avg loss: 0.35578979, Global Avg Loss: 2.12843644, Time: 0.0209 Steps: 20360, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001209, Sample Num: 19344, Cur Loss: 0.82109213, Cur Avg Loss: 0.45917524, Log Avg loss: 0.35179041, Global Avg Loss: 2.12756425, Time: 0.0209 Steps: 20370, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001219, Sample Num: 19504, Cur Loss: 0.91613448, Cur Avg Loss: 0.45963205, Log Avg loss: 0.51485986, Global Avg Loss: 2.12677293, Time: 0.0209 Steps: 20380, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001229, Sample Num: 19664, Cur Loss: 0.20331895, Cur Avg Loss: 0.45976403, Log Avg loss: 0.47585245, Global Avg Loss: 2.12596326, Time: 0.0209 Steps: 20390, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001239, Sample Num: 19824, Cur Loss: 0.22943573, Cur Avg Loss: 0.45970203, Log Avg loss: 0.45208245, Global Avg Loss: 2.12514273, Time: 0.0209 Steps: 20400, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001249, Sample Num: 19984, Cur Loss: 0.52352870, Cur Avg Loss: 0.45900106, Log Avg loss: 0.37215070, Global Avg Loss: 2.12428384, Time: 0.0209 Steps: 20410, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001259, Sample Num: 20144, Cur Loss: 0.25111443, Cur Avg Loss: 0.45820209, Log Avg loss: 0.35841058, Global Avg Loss: 2.12341907, Time: 0.0209 Steps: 20420, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001269, Sample Num: 20304, Cur Loss: 0.22305204, Cur Avg Loss: 0.45812213, Log Avg loss: 0.44805583, Global Avg Loss: 2.12259901, Time: 0.0209 Steps: 20430, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001279, Sample Num: 20464, Cur Loss: 0.76914018, Cur Avg Loss: 0.45856784, Log Avg loss: 0.51512875, Global Avg Loss: 2.12181258, Time: 0.0209 Steps: 20440, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001289, Sample Num: 20624, Cur Loss: 0.17993678, Cur Avg Loss: 0.45846963, Log Avg loss: 0.44590809, Global Avg Loss: 2.12099307, Time: 0.0208 Steps: 20450, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001299, Sample Num: 20784, Cur Loss: 0.22858101, Cur Avg Loss: 0.45864422, Log Avg loss: 0.48114943, Global Avg Loss: 2.12019158, Time: 0.0210 Steps: 20460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001309, Sample Num: 20944, Cur Loss: 1.00067437, Cur Avg Loss: 0.45829818, Log Avg loss: 0.41334759, Global Avg Loss: 2.11935775, Time: 0.0210 Steps: 20470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001319, Sample Num: 21104, Cur Loss: 0.23715901, Cur Avg Loss: 0.46025318, Log Avg loss: 0.71616287, Global Avg Loss: 2.11867260, Time: 0.0209 Steps: 20480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001329, Sample Num: 21264, Cur Loss: 0.15116601, Cur Avg Loss: 0.46141451, Log Avg loss: 0.61459382, Global Avg Loss: 2.11793854, Time: 0.0209 Steps: 20490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001339, Sample Num: 21424, Cur Loss: 0.87733513, Cur Avg Loss: 0.46080716, Log Avg loss: 0.38008982, Global Avg Loss: 2.11709081, Time: 0.0210 Steps: 20500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001349, Sample Num: 21584, Cur Loss: 0.42498583, Cur Avg Loss: 0.46131635, Log Avg loss: 0.52949668, Global Avg Loss: 2.11631675, Time: 0.0210 Steps: 20510, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001359, Sample Num: 21744, Cur Loss: 0.51629579, Cur Avg Loss: 0.46147128, Log Avg loss: 0.48237131, Global Avg Loss: 2.11552048, Time: 0.0210 Steps: 20520, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001369, Sample Num: 21904, Cur Loss: 0.58510065, Cur Avg Loss: 0.46309562, Log Avg loss: 0.68384402, Global Avg Loss: 2.11482313, Time: 0.0210 Steps: 20530, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001379, Sample Num: 22064, Cur Loss: 0.65861005, Cur Avg Loss: 0.46280652, Log Avg loss: 0.42322802, Global Avg Loss: 2.11399957, Time: 0.0210 Steps: 20540, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001389, Sample Num: 22224, Cur Loss: 0.19598246, Cur Avg Loss: 0.46367780, Log Avg loss: 0.58382772, Global Avg Loss: 2.11325496, Time: 0.0210 Steps: 20550, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001399, Sample Num: 22384, Cur Loss: 0.78499967, Cur Avg Loss: 0.46370070, Log Avg loss: 0.46688101, Global Avg Loss: 2.11245419, Time: 0.0209 Steps: 20560, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001409, Sample Num: 22544, Cur Loss: 0.29088792, Cur Avg Loss: 0.46305794, Log Avg loss: 0.37313693, Global Avg Loss: 2.11160863, Time: 0.0209 Steps: 20570, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001419, Sample Num: 22704, Cur Loss: 0.32406718, Cur Avg Loss: 0.46179170, Log Avg loss: 0.28337821, Global Avg Loss: 2.11072028, Time: 0.0209 Steps: 20580, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001429, Sample Num: 22864, Cur Loss: 0.75518727, Cur Avg Loss: 0.46159852, Log Avg loss: 0.43418655, Global Avg Loss: 2.10990603, Time: 0.0209 Steps: 20590, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001439, Sample Num: 23024, Cur Loss: 0.20876764, Cur Avg Loss: 0.46054693, Log Avg loss: 0.31027362, Global Avg Loss: 2.10903242, Time: 0.0210 Steps: 20600, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001449, Sample Num: 23184, Cur Loss: 0.64292753, Cur Avg Loss: 0.46062037, Log Avg loss: 0.47118929, Global Avg Loss: 2.10823774, Time: 0.0210 Steps: 20610, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001459, Sample Num: 23344, Cur Loss: 0.21988738, Cur Avg Loss: 0.46015894, Log Avg loss: 0.39329763, Global Avg Loss: 2.10740605, Time: 0.0210 Steps: 20620, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001469, Sample Num: 23504, Cur Loss: 0.41426027, Cur Avg Loss: 0.45951501, Log Avg loss: 0.36556590, Global Avg Loss: 2.10656173, Time: 0.0210 Steps: 20630, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001479, Sample Num: 23664, Cur Loss: 0.25304756, Cur Avg Loss: 0.45936114, Log Avg loss: 0.43675652, Global Avg Loss: 2.10575271, Time: 0.0209 Steps: 20640, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001489, Sample Num: 23824, Cur Loss: 0.14308088, Cur Avg Loss: 0.45912701, Log Avg loss: 0.42449932, Global Avg Loss: 2.10493855, Time: 0.0209 Steps: 20650, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001499, Sample Num: 23984, Cur Loss: 0.66861731, Cur Avg Loss: 0.45827854, Log Avg loss: 0.33194155, Global Avg Loss: 2.10408037, Time: 0.0210 Steps: 20660, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001509, Sample Num: 24144, Cur Loss: 0.21248442, Cur Avg Loss: 0.45834167, Log Avg loss: 0.46780531, Global Avg Loss: 2.10328875, Time: 0.0209 Steps: 20670, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001519, Sample Num: 24304, Cur Loss: 0.21370760, Cur Avg Loss: 0.45777419, Log Avg loss: 0.37214141, Global Avg Loss: 2.10245164, Time: 0.0209 Steps: 20680, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001529, Sample Num: 24464, Cur Loss: 0.41739863, Cur Avg Loss: 0.45770638, Log Avg loss: 0.44740623, Global Avg Loss: 2.10165171, Time: 0.0209 Steps: 20690, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001539, Sample Num: 24624, Cur Loss: 0.46448377, Cur Avg Loss: 0.45735254, Log Avg loss: 0.40325039, Global Avg Loss: 2.10083123, Time: 0.0246 Steps: 20700, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001549, Sample Num: 24784, Cur Loss: 0.25626117, Cur Avg Loss: 0.45667752, Log Avg loss: 0.35279104, Global Avg Loss: 2.09998717, Time: 0.0210 Steps: 20710, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001559, Sample Num: 24944, Cur Loss: 0.18963787, Cur Avg Loss: 0.45570434, Log Avg loss: 0.30495970, Global Avg Loss: 2.09912085, Time: 0.0209 Steps: 20720, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001569, Sample Num: 25104, Cur Loss: 0.23000506, Cur Avg Loss: 0.45533729, Log Avg loss: 0.39811346, Global Avg Loss: 2.09830029, Time: 0.0210 Steps: 20730, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001579, Sample Num: 25264, Cur Loss: 0.22419837, Cur Avg Loss: 0.45522805, Log Avg loss: 0.43808848, Global Avg Loss: 2.09749981, Time: 0.0209 Steps: 20740, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001589, Sample Num: 25424, Cur Loss: 0.36908373, Cur Avg Loss: 0.45452130, Log Avg loss: 0.34292523, Global Avg Loss: 2.09665423, Time: 0.0209 Steps: 20750, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001599, Sample Num: 25584, Cur Loss: 0.94998097, Cur Avg Loss: 0.45455885, Log Avg loss: 0.46052627, Global Avg Loss: 2.09586611, Time: 0.0209 Steps: 20760, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001609, Sample Num: 25744, Cur Loss: 0.10826319, Cur Avg Loss: 0.45463172, Log Avg loss: 0.46628343, Global Avg Loss: 2.09508153, Time: 0.0209 Steps: 20770, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001619, Sample Num: 25904, Cur Loss: 0.19729626, Cur Avg Loss: 0.45536002, Log Avg loss: 0.57254307, Global Avg Loss: 2.09434883, Time: 0.0209 Steps: 20780, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001629, Sample Num: 26064, Cur Loss: 1.08386111, Cur Avg Loss: 0.45584668, Log Avg loss: 0.53463673, Global Avg Loss: 2.09359861, Time: 0.0209 Steps: 20790, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001639, Sample Num: 26224, Cur Loss: 0.14033073, Cur Avg Loss: 0.45519674, Log Avg loss: 0.34932229, Global Avg Loss: 2.09276002, Time: 0.0209 Steps: 20800, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001649, Sample Num: 26384, Cur Loss: 0.92290950, Cur Avg Loss: 0.45548871, Log Avg loss: 0.50334247, Global Avg Loss: 2.09199624, Time: 0.0209 Steps: 20810, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001659, Sample Num: 26544, Cur Loss: 0.32973236, Cur Avg Loss: 0.45520607, Log Avg loss: 0.40859800, Global Avg Loss: 2.09118769, Time: 0.0209 Steps: 20820, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001669, Sample Num: 26704, Cur Loss: 0.28662089, Cur Avg Loss: 0.45454713, Log Avg loss: 0.34523030, Global Avg Loss: 2.09034950, Time: 0.0209 Steps: 20830, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001679, Sample Num: 26864, Cur Loss: 0.17232677, Cur Avg Loss: 0.45418673, Log Avg loss: 0.39403439, Global Avg Loss: 2.08953553, Time: 0.0209 Steps: 20840, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001689, Sample Num: 27024, Cur Loss: 0.44883430, Cur Avg Loss: 0.45346092, Log Avg loss: 0.33159802, Global Avg Loss: 2.08869239, Time: 0.0210 Steps: 20850, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001699, Sample Num: 27184, Cur Loss: 0.23437545, Cur Avg Loss: 0.45344554, Log Avg loss: 0.45084720, Global Avg Loss: 2.08790723, Time: 0.0209 Steps: 20860, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001709, Sample Num: 27344, Cur Loss: 0.05688015, Cur Avg Loss: 0.45299195, Log Avg loss: 0.37592749, Global Avg Loss: 2.08708692, Time: 0.0209 Steps: 20870, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001719, Sample Num: 27504, Cur Loss: 0.21473548, Cur Avg Loss: 0.45262879, Log Avg loss: 0.39056434, Global Avg Loss: 2.08627441, Time: 0.0209 Steps: 20880, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001729, Sample Num: 27664, Cur Loss: 0.92048174, Cur Avg Loss: 0.45329568, Log Avg loss: 0.56793528, Global Avg Loss: 2.08554759, Time: 0.0210 Steps: 20890, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001739, Sample Num: 27824, Cur Loss: 0.18101385, Cur Avg Loss: 0.45422067, Log Avg loss: 0.61415172, Global Avg Loss: 2.08484357, Time: 0.0209 Steps: 20900, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001749, Sample Num: 27984, Cur Loss: 0.16553693, Cur Avg Loss: 0.45447394, Log Avg loss: 0.49851710, Global Avg Loss: 2.08408493, Time: 0.0209 Steps: 20910, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001759, Sample Num: 28144, Cur Loss: 0.10629004, Cur Avg Loss: 0.45406246, Log Avg loss: 0.38209353, Global Avg Loss: 2.08327135, Time: 0.0209 Steps: 20920, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001769, Sample Num: 28304, Cur Loss: 0.30667567, Cur Avg Loss: 0.45336916, Log Avg loss: 0.33141925, Global Avg Loss: 2.08243435, Time: 0.0209 Steps: 20930, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001779, Sample Num: 28464, Cur Loss: 0.34570801, Cur Avg Loss: 0.45334769, Log Avg loss: 0.44954922, Global Avg Loss: 2.08165456, Time: 0.0209 Steps: 20940, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001789, Sample Num: 28624, Cur Loss: 0.60279745, Cur Avg Loss: 0.45421486, Log Avg loss: 0.60848489, Global Avg Loss: 2.08095137, Time: 0.0209 Steps: 20950, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001799, Sample Num: 28784, Cur Loss: 0.61254305, Cur Avg Loss: 0.45580989, Log Avg loss: 0.74115976, Global Avg Loss: 2.08031216, Time: 0.0209 Steps: 20960, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001809, Sample Num: 28944, Cur Loss: 0.33886701, Cur Avg Loss: 0.45591822, Log Avg loss: 0.47540775, Global Avg Loss: 2.07954683, Time: 0.0208 Steps: 20970, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001819, Sample Num: 29104, Cur Loss: 1.41570210, Cur Avg Loss: 0.45702196, Log Avg loss: 0.65668697, Global Avg Loss: 2.07886863, Time: 0.0209 Steps: 20980, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001829, Sample Num: 29264, Cur Loss: 0.42225176, Cur Avg Loss: 0.45785358, Log Avg loss: 0.60912584, Global Avg Loss: 2.07816842, Time: 0.0208 Steps: 20990, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001839, Sample Num: 29424, Cur Loss: 0.48565111, Cur Avg Loss: 0.45773327, Log Avg loss: 0.43572805, Global Avg Loss: 2.07738630, Time: 0.0209 Steps: 21000, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001849, Sample Num: 29584, Cur Loss: 0.49576306, Cur Avg Loss: 0.45695164, Log Avg loss: 0.31321051, Global Avg Loss: 2.07654662, Time: 0.0209 Steps: 21010, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001859, Sample Num: 29744, Cur Loss: 2.74829698, Cur Avg Loss: 0.46015453, Log Avg loss: 1.05236965, Global Avg Loss: 2.07605938, Time: 0.0208 Steps: 21020, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001869, Sample Num: 29904, Cur Loss: 0.44559124, Cur Avg Loss: 0.46067168, Log Avg loss: 0.55680995, Global Avg Loss: 2.07533696, Time: 0.0209 Steps: 21030, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001879, Sample Num: 30064, Cur Loss: 0.84388065, Cur Avg Loss: 0.46317216, Log Avg loss: 0.93051118, Global Avg Loss: 2.07479284, Time: 0.0209 Steps: 21040, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001889, Sample Num: 30224, Cur Loss: 1.47404146, Cur Avg Loss: 0.46585709, Log Avg loss: 0.97035553, Global Avg Loss: 2.07426817, Time: 0.0209 Steps: 21050, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001899, Sample Num: 30384, Cur Loss: 0.17893380, Cur Avg Loss: 0.46629996, Log Avg loss: 0.54995806, Global Avg Loss: 2.07354437, Time: 0.0208 Steps: 21060, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001909, Sample Num: 30544, Cur Loss: 0.19439542, Cur Avg Loss: 0.46549172, Log Avg loss: 0.31200780, Global Avg Loss: 2.07270833, Time: 0.0209 Steps: 21070, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001919, Sample Num: 30704, Cur Loss: 0.33180416, Cur Avg Loss: 0.46540443, Log Avg loss: 0.44873930, Global Avg Loss: 2.07193795, Time: 0.0208 Steps: 21080, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001929, Sample Num: 30864, Cur Loss: 0.19759060, Cur Avg Loss: 0.46461216, Log Avg loss: 0.31257648, Global Avg Loss: 2.07110373, Time: 0.0209 Steps: 21090, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001939, Sample Num: 31024, Cur Loss: 0.42526305, Cur Avg Loss: 0.46443179, Log Avg loss: 0.42963745, Global Avg Loss: 2.07032579, Time: 0.0209 Steps: 21100, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001949, Sample Num: 31184, Cur Loss: 0.69256574, Cur Avg Loss: 0.46575884, Log Avg loss: 0.72307514, Global Avg Loss: 2.06968758, Time: 0.0210 Steps: 21110, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001959, Sample Num: 31344, Cur Loss: 0.29318264, Cur Avg Loss: 0.46564954, Log Avg loss: 0.44434619, Global Avg Loss: 2.06891801, Time: 0.0208 Steps: 21120, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001969, Sample Num: 31504, Cur Loss: 0.29104856, Cur Avg Loss: 0.46569539, Log Avg loss: 0.47467813, Global Avg Loss: 2.06816352, Time: 0.0209 Steps: 21130, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001979, Sample Num: 31664, Cur Loss: 0.80335927, Cur Avg Loss: 0.46616347, Log Avg loss: 0.55832822, Global Avg Loss: 2.06744931, Time: 0.0209 Steps: 21140, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001989, Sample Num: 31824, Cur Loss: 0.46771407, Cur Avg Loss: 0.46594909, Log Avg loss: 0.42352292, Global Avg Loss: 2.06667204, Time: 0.0210 Steps: 21150, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001999, Sample Num: 31984, Cur Loss: 0.39892101, Cur Avg Loss: 0.46558424, Log Avg loss: 0.39301636, Global Avg Loss: 2.06588109, Time: 0.0208 Steps: 21160, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002009, Sample Num: 32144, Cur Loss: 0.34155309, Cur Avg Loss: 0.46506341, Log Avg loss: 0.36094880, Global Avg Loss: 2.06507573, Time: 0.0209 Steps: 21170, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002019, Sample Num: 32304, Cur Loss: 0.34097251, Cur Avg Loss: 0.46454496, Log Avg loss: 0.36038886, Global Avg Loss: 2.06427088, Time: 0.0209 Steps: 21180, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002029, Sample Num: 32464, Cur Loss: 0.30408740, Cur Avg Loss: 0.46446444, Log Avg loss: 0.44820642, Global Avg Loss: 2.06350822, Time: 0.0208 Steps: 21190, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002039, Sample Num: 32624, Cur Loss: 0.18487950, Cur Avg Loss: 0.46360240, Log Avg loss: 0.28869569, Global Avg Loss: 2.06267105, Time: 0.0209 Steps: 21200, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002049, Sample Num: 32784, Cur Loss: 0.78471434, Cur Avg Loss: 0.46408761, Log Avg loss: 0.56302027, Global Avg Loss: 2.06196400, Time: 0.0246 Steps: 21210, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002059, Sample Num: 32944, Cur Loss: 0.15895972, Cur Avg Loss: 0.46385857, Log Avg loss: 0.41692885, Global Avg Loss: 2.06118877, Time: 0.0209 Steps: 21220, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002069, Sample Num: 33104, Cur Loss: 0.23132893, Cur Avg Loss: 0.46497036, Log Avg loss: 0.69388811, Global Avg Loss: 2.06054473, Time: 0.0209 Steps: 21230, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002079, Sample Num: 33264, Cur Loss: 0.51903880, Cur Avg Loss: 0.46525679, Log Avg loss: 0.52451927, Global Avg Loss: 2.05982155, Time: 0.0208 Steps: 21240, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002089, Sample Num: 33424, Cur Loss: 0.19211799, Cur Avg Loss: 0.46459923, Log Avg loss: 0.32789216, Global Avg Loss: 2.05900653, Time: 0.0209 Steps: 21250, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002099, Sample Num: 33584, Cur Loss: 0.58504063, Cur Avg Loss: 0.46437102, Log Avg loss: 0.41669773, Global Avg Loss: 2.05823404, Time: 0.0209 Steps: 21260, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002109, Sample Num: 33744, Cur Loss: 0.17732312, Cur Avg Loss: 0.46492641, Log Avg loss: 0.58150401, Global Avg Loss: 2.05753976, Time: 0.0209 Steps: 21270, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002119, Sample Num: 33904, Cur Loss: 0.61332405, Cur Avg Loss: 0.46405519, Log Avg loss: 0.28031511, Global Avg Loss: 2.05670460, Time: 0.0209 Steps: 21280, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002129, Sample Num: 34055, Cur Loss: 0.27718905, Cur Avg Loss: 0.46379622, Log Avg loss: 0.40891996, Global Avg Loss: 2.05593063, Time: 0.0101 Steps: 21290, Updated lr: 0.000081 ***** Running evaluation checkpoint-21290 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-21290 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.603790, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.79113, "eval_total_loss": 556.16437, "eval_mae": 0.782335, "eval_mse": 0.79119, "eval_r2": 0.497068, "eval_sp_statistic": 0.841244, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.877069, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.661152, "test_total_loss": 331.898226, "test_mae": 0.735144, "test_mse": 0.661295, "test_r2": 0.573194, "test_sp_statistic": 0.831163, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.880754, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.055930626572417, "train_cur_epoch_loss": 987.4221570119262, "train_cur_epoch_avg_loss": 0.46379622217563465, "train_cur_epoch_time": 44.60378980636597, "train_cur_epoch_avg_time": 0.02095058234211647, "epoch": 10, "step": 21290} ################################################## Training, Epoch: 0011, Batch: 000010, Sample Num: 160, Cur Loss: 0.63832760, Cur Avg Loss: 0.38806633, Log Avg loss: 0.38806633, Global Avg Loss: 2.05514759, Time: 0.0208 Steps: 21300, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000020, Sample Num: 320, Cur Loss: 0.82027233, Cur Avg Loss: 0.54692617, Log Avg loss: 0.70578601, Global Avg Loss: 2.05451439, Time: 0.0209 Steps: 21310, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000030, Sample Num: 480, Cur Loss: 0.33021691, Cur Avg Loss: 0.68393568, Log Avg loss: 0.95795470, Global Avg Loss: 2.05400005, Time: 0.0210 Steps: 21320, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000040, Sample Num: 640, Cur Loss: 0.53075844, Cur Avg Loss: 0.64288016, Log Avg loss: 0.51971359, Global Avg Loss: 2.05328074, Time: 0.0209 Steps: 21330, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000050, Sample Num: 800, Cur Loss: 0.16800614, Cur Avg Loss: 0.59950443, Log Avg loss: 0.42600153, Global Avg Loss: 2.05251819, Time: 0.0210 Steps: 21340, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000060, Sample Num: 960, Cur Loss: 0.64671361, Cur Avg Loss: 0.57526742, Log Avg loss: 0.45408234, Global Avg Loss: 2.05176951, Time: 0.0209 Steps: 21350, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000070, Sample Num: 1120, Cur Loss: 1.02452576, Cur Avg Loss: 0.62178260, Log Avg loss: 0.90087372, Global Avg Loss: 2.05123070, Time: 0.0209 Steps: 21360, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000080, Sample Num: 1280, Cur Loss: 0.19218239, Cur Avg Loss: 0.66807063, Log Avg loss: 0.99208678, Global Avg Loss: 2.05073508, Time: 0.0208 Steps: 21370, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000090, Sample Num: 1440, Cur Loss: 0.33418900, Cur Avg Loss: 0.65058307, Log Avg loss: 0.51068266, Global Avg Loss: 2.05001476, Time: 0.0210 Steps: 21380, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000100, Sample Num: 1600, Cur Loss: 0.35135749, Cur Avg Loss: 0.61758069, Log Avg loss: 0.32055920, Global Avg Loss: 2.04920622, Time: 0.0208 Steps: 21390, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000110, Sample Num: 1760, Cur Loss: 0.39985764, Cur Avg Loss: 0.58825058, Log Avg loss: 0.29494953, Global Avg Loss: 2.04838648, Time: 0.0210 Steps: 21400, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000120, Sample Num: 1920, Cur Loss: 0.24510430, Cur Avg Loss: 0.59214749, Log Avg loss: 0.63501347, Global Avg Loss: 2.04772633, Time: 0.0208 Steps: 21410, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000130, Sample Num: 2080, Cur Loss: 0.32160068, Cur Avg Loss: 0.57771727, Log Avg loss: 0.40455463, Global Avg Loss: 2.04695921, Time: 0.0210 Steps: 21420, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000140, Sample Num: 2240, Cur Loss: 0.14733103, Cur Avg Loss: 0.56276858, Log Avg loss: 0.36843566, Global Avg Loss: 2.04617595, Time: 0.0209 Steps: 21430, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000150, Sample Num: 2400, Cur Loss: 0.29333493, Cur Avg Loss: 0.55161679, Log Avg loss: 0.39549163, Global Avg Loss: 2.04540604, Time: 0.0210 Steps: 21440, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000160, Sample Num: 2560, Cur Loss: 0.25232568, Cur Avg Loss: 0.54435346, Log Avg loss: 0.43540352, Global Avg Loss: 2.04465546, Time: 0.0209 Steps: 21450, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000170, Sample Num: 2720, Cur Loss: 0.47489458, Cur Avg Loss: 0.53810617, Log Avg loss: 0.43814953, Global Avg Loss: 2.04390685, Time: 0.0210 Steps: 21460, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000180, Sample Num: 2880, Cur Loss: 0.27170089, Cur Avg Loss: 0.53124275, Log Avg loss: 0.41456474, Global Avg Loss: 2.04314796, Time: 0.0209 Steps: 21470, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000190, Sample Num: 3040, Cur Loss: 0.59293872, Cur Avg Loss: 0.53095383, Log Avg loss: 0.52575317, Global Avg Loss: 2.04244154, Time: 0.0208 Steps: 21480, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000200, Sample Num: 3200, Cur Loss: 0.31483009, Cur Avg Loss: 0.52105096, Log Avg loss: 0.33289642, Global Avg Loss: 2.04164603, Time: 0.0209 Steps: 21490, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000210, Sample Num: 3360, Cur Loss: 0.50104207, Cur Avg Loss: 0.51632748, Log Avg loss: 0.42185797, Global Avg Loss: 2.04089264, Time: 0.0208 Steps: 21500, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000220, Sample Num: 3520, Cur Loss: 0.39426714, Cur Avg Loss: 0.51723316, Log Avg loss: 0.53625233, Global Avg Loss: 2.04019314, Time: 0.0209 Steps: 21510, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000230, Sample Num: 3680, Cur Loss: 0.34512615, Cur Avg Loss: 0.51184331, Log Avg loss: 0.39326655, Global Avg Loss: 2.03942783, Time: 0.0210 Steps: 21520, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000240, Sample Num: 3840, Cur Loss: 0.55726969, Cur Avg Loss: 0.50819689, Log Avg loss: 0.42432945, Global Avg Loss: 2.03867767, Time: 0.0210 Steps: 21530, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000250, Sample Num: 4000, Cur Loss: 0.38883376, Cur Avg Loss: 0.50330576, Log Avg loss: 0.38591850, Global Avg Loss: 2.03791038, Time: 0.0209 Steps: 21540, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000260, Sample Num: 4160, Cur Loss: 0.41575316, Cur Avg Loss: 0.49446432, Log Avg loss: 0.27342825, Global Avg Loss: 2.03709159, Time: 0.0246 Steps: 21550, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000270, Sample Num: 4320, Cur Loss: 0.30556926, Cur Avg Loss: 0.48804728, Log Avg loss: 0.32120441, Global Avg Loss: 2.03629572, Time: 0.0209 Steps: 21560, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000280, Sample Num: 4480, Cur Loss: 0.77704000, Cur Avg Loss: 0.49085903, Log Avg loss: 0.56677625, Global Avg Loss: 2.03561444, Time: 0.0209 Steps: 21570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000290, Sample Num: 4640, Cur Loss: 0.20232412, Cur Avg Loss: 0.48894566, Log Avg loss: 0.43537121, Global Avg Loss: 2.03487290, Time: 0.0209 Steps: 21580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000300, Sample Num: 4800, Cur Loss: 0.44616070, Cur Avg Loss: 0.48839536, Log Avg loss: 0.47243672, Global Avg Loss: 2.03414922, Time: 0.0211 Steps: 21590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000310, Sample Num: 4960, Cur Loss: 0.82495356, Cur Avg Loss: 0.49930233, Log Avg loss: 0.82651126, Global Avg Loss: 2.03359013, Time: 0.0209 Steps: 21600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000320, Sample Num: 5120, Cur Loss: 0.56839937, Cur Avg Loss: 0.50542637, Log Avg loss: 0.69527175, Global Avg Loss: 2.03297082, Time: 0.0209 Steps: 21610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000330, Sample Num: 5280, Cur Loss: 0.45918342, Cur Avg Loss: 0.50343860, Log Avg loss: 0.43982990, Global Avg Loss: 2.03223394, Time: 0.0210 Steps: 21620, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000340, Sample Num: 5440, Cur Loss: 0.31387287, Cur Avg Loss: 0.50076053, Log Avg loss: 0.41238427, Global Avg Loss: 2.03148505, Time: 0.0209 Steps: 21630, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000350, Sample Num: 5600, Cur Loss: 0.21265157, Cur Avg Loss: 0.50178917, Log Avg loss: 0.53676280, Global Avg Loss: 2.03079433, Time: 0.0209 Steps: 21640, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000360, Sample Num: 5760, Cur Loss: 0.46803215, Cur Avg Loss: 0.49780418, Log Avg loss: 0.35832953, Global Avg Loss: 2.03002183, Time: 0.0210 Steps: 21650, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000370, Sample Num: 5920, Cur Loss: 0.46981078, Cur Avg Loss: 0.49624061, Log Avg loss: 0.43995236, Global Avg Loss: 2.02928772, Time: 0.0209 Steps: 21660, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000380, Sample Num: 6080, Cur Loss: 0.33056030, Cur Avg Loss: 0.49462151, Log Avg loss: 0.43471475, Global Avg Loss: 2.02855188, Time: 0.0209 Steps: 21670, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000390, Sample Num: 6240, Cur Loss: 1.22480786, Cur Avg Loss: 0.49505237, Log Avg loss: 0.51142503, Global Avg Loss: 2.02785210, Time: 0.0209 Steps: 21680, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000400, Sample Num: 6400, Cur Loss: 0.73489338, Cur Avg Loss: 0.49161093, Log Avg loss: 0.35739454, Global Avg Loss: 2.02708195, Time: 0.0209 Steps: 21690, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000410, Sample Num: 6560, Cur Loss: 0.16187480, Cur Avg Loss: 0.49034192, Log Avg loss: 0.43958178, Global Avg Loss: 2.02635038, Time: 0.0210 Steps: 21700, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000420, Sample Num: 6720, Cur Loss: 0.36390603, Cur Avg Loss: 0.48688788, Log Avg loss: 0.34527202, Global Avg Loss: 2.02557605, Time: 0.0209 Steps: 21710, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000430, Sample Num: 6880, Cur Loss: 0.20230842, Cur Avg Loss: 0.48488711, Log Avg loss: 0.40085480, Global Avg Loss: 2.02482802, Time: 0.0209 Steps: 21720, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000440, Sample Num: 7040, Cur Loss: 0.18716621, Cur Avg Loss: 0.48294551, Log Avg loss: 0.39945669, Global Avg Loss: 2.02408003, Time: 0.0209 Steps: 21730, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000450, Sample Num: 7200, Cur Loss: 0.32013407, Cur Avg Loss: 0.48000892, Log Avg loss: 0.35079931, Global Avg Loss: 2.02331035, Time: 0.0209 Steps: 21740, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000460, Sample Num: 7360, Cur Loss: 0.53692663, Cur Avg Loss: 0.47824889, Log Avg loss: 0.39904729, Global Avg Loss: 2.02256356, Time: 0.0210 Steps: 21750, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000470, Sample Num: 7520, Cur Loss: 0.34114873, Cur Avg Loss: 0.47681784, Log Avg loss: 0.41098947, Global Avg Loss: 2.02182295, Time: 0.0210 Steps: 21760, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000480, Sample Num: 7680, Cur Loss: 0.63495392, Cur Avg Loss: 0.47456385, Log Avg loss: 0.36862631, Global Avg Loss: 2.02106356, Time: 0.0210 Steps: 21770, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000490, Sample Num: 7840, Cur Loss: 0.64398819, Cur Avg Loss: 0.47340478, Log Avg loss: 0.41776950, Global Avg Loss: 2.02032743, Time: 0.0209 Steps: 21780, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000500, Sample Num: 8000, Cur Loss: 0.19374767, Cur Avg Loss: 0.47201542, Log Avg loss: 0.40393678, Global Avg Loss: 2.01958562, Time: 0.0209 Steps: 21790, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000510, Sample Num: 8160, Cur Loss: 0.06761423, Cur Avg Loss: 0.46870248, Log Avg loss: 0.30305556, Global Avg Loss: 2.01879822, Time: 0.0209 Steps: 21800, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000520, Sample Num: 8320, Cur Loss: 0.10111938, Cur Avg Loss: 0.46707451, Log Avg loss: 0.38404806, Global Avg Loss: 2.01804868, Time: 0.0210 Steps: 21810, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000530, Sample Num: 8480, Cur Loss: 0.59918404, Cur Avg Loss: 0.46716759, Log Avg loss: 0.47200779, Global Avg Loss: 2.01734014, Time: 0.0209 Steps: 21820, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000540, Sample Num: 8640, Cur Loss: 0.91270888, Cur Avg Loss: 0.47025239, Log Avg loss: 0.63374686, Global Avg Loss: 2.01670634, Time: 0.0209 Steps: 21830, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000550, Sample Num: 8800, Cur Loss: 0.68366253, Cur Avg Loss: 0.46829695, Log Avg loss: 0.36270283, Global Avg Loss: 2.01594901, Time: 0.0209 Steps: 21840, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000560, Sample Num: 8960, Cur Loss: 0.38144714, Cur Avg Loss: 0.46569752, Log Avg loss: 0.32272895, Global Avg Loss: 2.01517408, Time: 0.0209 Steps: 21850, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000570, Sample Num: 9120, Cur Loss: 0.50543135, Cur Avg Loss: 0.46468080, Log Avg loss: 0.40774453, Global Avg Loss: 2.01443875, Time: 0.0209 Steps: 21860, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000580, Sample Num: 9280, Cur Loss: 0.62496388, Cur Avg Loss: 0.46238281, Log Avg loss: 0.33139760, Global Avg Loss: 2.01366918, Time: 0.0209 Steps: 21870, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000590, Sample Num: 9440, Cur Loss: 0.44847006, Cur Avg Loss: 0.46191580, Log Avg loss: 0.43482899, Global Avg Loss: 2.01294759, Time: 0.0210 Steps: 21880, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000600, Sample Num: 9600, Cur Loss: 0.32042229, Cur Avg Loss: 0.45982792, Log Avg loss: 0.33664329, Global Avg Loss: 2.01218181, Time: 0.0209 Steps: 21890, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000610, Sample Num: 9760, Cur Loss: 0.20579201, Cur Avg Loss: 0.45724155, Log Avg loss: 0.30205894, Global Avg Loss: 2.01140093, Time: 0.0210 Steps: 21900, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000620, Sample Num: 9920, Cur Loss: 0.46979332, Cur Avg Loss: 0.45569762, Log Avg loss: 0.36151823, Global Avg Loss: 2.01064790, Time: 0.0209 Steps: 21910, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000630, Sample Num: 10080, Cur Loss: 1.17825854, Cur Avg Loss: 0.45713509, Log Avg loss: 0.54625827, Global Avg Loss: 2.00997984, Time: 0.0209 Steps: 21920, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000640, Sample Num: 10240, Cur Loss: 0.22425303, Cur Avg Loss: 0.45381998, Log Avg loss: 0.24496808, Global Avg Loss: 2.00917500, Time: 0.0209 Steps: 21930, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000650, Sample Num: 10400, Cur Loss: 0.31332529, Cur Avg Loss: 0.45060132, Log Avg loss: 0.24460688, Global Avg Loss: 2.00837073, Time: 0.0209 Steps: 21940, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000660, Sample Num: 10560, Cur Loss: 0.21484622, Cur Avg Loss: 0.45049132, Log Avg loss: 0.44334129, Global Avg Loss: 2.00765774, Time: 0.0210 Steps: 21950, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000670, Sample Num: 10720, Cur Loss: 0.69191056, Cur Avg Loss: 0.45240823, Log Avg loss: 0.57892412, Global Avg Loss: 2.00700713, Time: 0.0209 Steps: 21960, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000680, Sample Num: 10880, Cur Loss: 0.24281010, Cur Avg Loss: 0.45207360, Log Avg loss: 0.42965391, Global Avg Loss: 2.00628917, Time: 0.0209 Steps: 21970, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000690, Sample Num: 11040, Cur Loss: 1.46996331, Cur Avg Loss: 0.45644288, Log Avg loss: 0.75355390, Global Avg Loss: 2.00571923, Time: 0.0209 Steps: 21980, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000700, Sample Num: 11200, Cur Loss: 0.13425061, Cur Avg Loss: 0.46387742, Log Avg loss: 0.97686005, Global Avg Loss: 2.00525135, Time: 0.0209 Steps: 21990, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000710, Sample Num: 11360, Cur Loss: 0.24884143, Cur Avg Loss: 0.46124165, Log Avg loss: 0.27673801, Global Avg Loss: 2.00446566, Time: 0.0209 Steps: 22000, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000720, Sample Num: 11520, Cur Loss: 0.06321575, Cur Avg Loss: 0.45850365, Log Avg loss: 0.26410552, Global Avg Loss: 2.00367495, Time: 0.0209 Steps: 22010, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000730, Sample Num: 11680, Cur Loss: 0.45234609, Cur Avg Loss: 0.45878751, Log Avg loss: 0.47922564, Global Avg Loss: 2.00298265, Time: 0.0209 Steps: 22020, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000740, Sample Num: 11840, Cur Loss: 0.23881429, Cur Avg Loss: 0.45767279, Log Avg loss: 0.37629853, Global Avg Loss: 2.00224425, Time: 0.0209 Steps: 22030, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000750, Sample Num: 12000, Cur Loss: 1.17163563, Cur Avg Loss: 0.45989757, Log Avg loss: 0.62453098, Global Avg Loss: 2.00161916, Time: 0.0209 Steps: 22040, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000760, Sample Num: 12160, Cur Loss: 0.16006306, Cur Avg Loss: 0.45873702, Log Avg loss: 0.37169555, Global Avg Loss: 2.00087996, Time: 0.0209 Steps: 22050, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000770, Sample Num: 12320, Cur Loss: 0.20996383, Cur Avg Loss: 0.45856168, Log Avg loss: 0.44523620, Global Avg Loss: 2.00017477, Time: 0.0245 Steps: 22060, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000780, Sample Num: 12480, Cur Loss: 0.35885298, Cur Avg Loss: 0.45695251, Log Avg loss: 0.33304599, Global Avg Loss: 1.99941939, Time: 0.0208 Steps: 22070, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000790, Sample Num: 12640, Cur Loss: 0.38420609, Cur Avg Loss: 0.45644712, Log Avg loss: 0.41702672, Global Avg Loss: 1.99870273, Time: 0.0208 Steps: 22080, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000800, Sample Num: 12800, Cur Loss: 0.19054830, Cur Avg Loss: 0.45552468, Log Avg loss: 0.38265244, Global Avg Loss: 1.99797115, Time: 0.0208 Steps: 22090, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000810, Sample Num: 12960, Cur Loss: 0.44637674, Cur Avg Loss: 0.45388882, Log Avg loss: 0.32301966, Global Avg Loss: 1.99721326, Time: 0.0208 Steps: 22100, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000820, Sample Num: 13120, Cur Loss: 0.11565422, Cur Avg Loss: 0.45377229, Log Avg loss: 0.44433368, Global Avg Loss: 1.99651091, Time: 0.0208 Steps: 22110, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000830, Sample Num: 13280, Cur Loss: 0.23038045, Cur Avg Loss: 0.46043968, Log Avg loss: 1.00716548, Global Avg Loss: 1.99606365, Time: 0.0207 Steps: 22120, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000840, Sample Num: 13440, Cur Loss: 0.36921334, Cur Avg Loss: 0.46136960, Log Avg loss: 0.53855265, Global Avg Loss: 1.99540504, Time: 0.0207 Steps: 22130, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000850, Sample Num: 13600, Cur Loss: 0.63134623, Cur Avg Loss: 0.46434852, Log Avg loss: 0.71457834, Global Avg Loss: 1.99482653, Time: 0.0207 Steps: 22140, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000860, Sample Num: 13760, Cur Loss: 0.49233332, Cur Avg Loss: 0.46545092, Log Avg loss: 0.55915473, Global Avg Loss: 1.99417837, Time: 0.0208 Steps: 22150, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000870, Sample Num: 13920, Cur Loss: 0.68736815, Cur Avg Loss: 0.46842811, Log Avg loss: 0.72446670, Global Avg Loss: 1.99360539, Time: 0.0207 Steps: 22160, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000880, Sample Num: 14080, Cur Loss: 0.57911474, Cur Avg Loss: 0.46779995, Log Avg loss: 0.41314949, Global Avg Loss: 1.99289251, Time: 0.0207 Steps: 22170, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000890, Sample Num: 14240, Cur Loss: 0.45193228, Cur Avg Loss: 0.46976195, Log Avg loss: 0.64241844, Global Avg Loss: 1.99228364, Time: 0.0207 Steps: 22180, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000900, Sample Num: 14400, Cur Loss: 0.40793484, Cur Avg Loss: 0.47119824, Log Avg loss: 0.59902795, Global Avg Loss: 1.99165577, Time: 0.0207 Steps: 22190, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000910, Sample Num: 14560, Cur Loss: 1.06001592, Cur Avg Loss: 0.47248764, Log Avg loss: 0.58853356, Global Avg Loss: 1.99102373, Time: 0.0208 Steps: 22200, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000920, Sample Num: 14720, Cur Loss: 0.80166358, Cur Avg Loss: 0.47237389, Log Avg loss: 0.46202254, Global Avg Loss: 1.99033530, Time: 0.0208 Steps: 22210, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000930, Sample Num: 14880, Cur Loss: 0.59885848, Cur Avg Loss: 0.47149248, Log Avg loss: 0.39040263, Global Avg Loss: 1.98961526, Time: 0.0208 Steps: 22220, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000940, Sample Num: 15040, Cur Loss: 0.17879543, Cur Avg Loss: 0.46950993, Log Avg loss: 0.28513281, Global Avg Loss: 1.98884851, Time: 0.0208 Steps: 22230, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000950, Sample Num: 15200, Cur Loss: 0.13113302, Cur Avg Loss: 0.46782754, Log Avg loss: 0.30968316, Global Avg Loss: 1.98809349, Time: 0.0208 Steps: 22240, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000960, Sample Num: 15360, Cur Loss: 0.58914948, Cur Avg Loss: 0.46737615, Log Avg loss: 0.42449357, Global Avg Loss: 1.98739075, Time: 0.0208 Steps: 22250, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000970, Sample Num: 15520, Cur Loss: 0.36993754, Cur Avg Loss: 0.46566263, Log Avg loss: 0.30116544, Global Avg Loss: 1.98663323, Time: 0.0208 Steps: 22260, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000980, Sample Num: 15680, Cur Loss: 0.44801003, Cur Avg Loss: 0.46596574, Log Avg loss: 0.49536709, Global Avg Loss: 1.98596360, Time: 0.0207 Steps: 22270, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000990, Sample Num: 15840, Cur Loss: 0.47953489, Cur Avg Loss: 0.46916583, Log Avg loss: 0.78277463, Global Avg Loss: 1.98542357, Time: 0.0207 Steps: 22280, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001000, Sample Num: 16000, Cur Loss: 0.74325395, Cur Avg Loss: 0.46871400, Log Avg loss: 0.42398275, Global Avg Loss: 1.98472306, Time: 0.0209 Steps: 22290, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001010, Sample Num: 16160, Cur Loss: 0.16012427, Cur Avg Loss: 0.46763793, Log Avg loss: 0.36003121, Global Avg Loss: 1.98399450, Time: 0.0209 Steps: 22300, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001020, Sample Num: 16320, Cur Loss: 0.49004719, Cur Avg Loss: 0.46806501, Log Avg loss: 0.51119938, Global Avg Loss: 1.98333435, Time: 0.0208 Steps: 22310, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001030, Sample Num: 16480, Cur Loss: 0.95169729, Cur Avg Loss: 0.46725253, Log Avg loss: 0.38437983, Global Avg Loss: 1.98261797, Time: 0.0246 Steps: 22320, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001040, Sample Num: 16640, Cur Loss: 0.50729650, Cur Avg Loss: 0.46562480, Log Avg loss: 0.29796880, Global Avg Loss: 1.98186354, Time: 0.0210 Steps: 22330, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001050, Sample Num: 16800, Cur Loss: 0.54490292, Cur Avg Loss: 0.46590832, Log Avg loss: 0.49539444, Global Avg Loss: 1.98119815, Time: 0.0209 Steps: 22340, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001060, Sample Num: 16960, Cur Loss: 0.36050126, Cur Avg Loss: 0.46586360, Log Avg loss: 0.46116801, Global Avg Loss: 1.98051805, Time: 0.0210 Steps: 22350, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001070, Sample Num: 17120, Cur Loss: 0.31942403, Cur Avg Loss: 0.46613294, Log Avg loss: 0.49468311, Global Avg Loss: 1.97985355, Time: 0.0210 Steps: 22360, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001080, Sample Num: 17280, Cur Loss: 0.40209910, Cur Avg Loss: 0.46475293, Log Avg loss: 0.31709223, Global Avg Loss: 1.97911025, Time: 0.0210 Steps: 22370, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001090, Sample Num: 17440, Cur Loss: 0.25743371, Cur Avg Loss: 0.46332358, Log Avg loss: 0.30895350, Global Avg Loss: 1.97836397, Time: 0.0210 Steps: 22380, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001100, Sample Num: 17600, Cur Loss: 0.35353515, Cur Avg Loss: 0.46250020, Log Avg loss: 0.37275187, Global Avg Loss: 1.97764686, Time: 0.0210 Steps: 22390, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001110, Sample Num: 17760, Cur Loss: 0.44974595, Cur Avg Loss: 0.46274126, Log Avg loss: 0.48925794, Global Avg Loss: 1.97698240, Time: 0.0210 Steps: 22400, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001120, Sample Num: 17920, Cur Loss: 0.34385112, Cur Avg Loss: 0.46165081, Log Avg loss: 0.34061008, Global Avg Loss: 1.97625221, Time: 0.0211 Steps: 22410, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001130, Sample Num: 18080, Cur Loss: 0.45042211, Cur Avg Loss: 0.46157371, Log Avg loss: 0.45293889, Global Avg Loss: 1.97557276, Time: 0.0208 Steps: 22420, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001140, Sample Num: 18240, Cur Loss: 0.22715066, Cur Avg Loss: 0.46145005, Log Avg loss: 0.44747693, Global Avg Loss: 1.97489149, Time: 0.0210 Steps: 22430, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001150, Sample Num: 18400, Cur Loss: 0.28729701, Cur Avg Loss: 0.46188289, Log Avg loss: 0.51122624, Global Avg Loss: 1.97423923, Time: 0.0210 Steps: 22440, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001160, Sample Num: 18560, Cur Loss: 0.40944874, Cur Avg Loss: 0.46075909, Log Avg loss: 0.33152247, Global Avg Loss: 1.97350751, Time: 0.0210 Steps: 22450, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001170, Sample Num: 18720, Cur Loss: 0.15863797, Cur Avg Loss: 0.46003638, Log Avg loss: 0.37620205, Global Avg Loss: 1.97279633, Time: 0.0211 Steps: 22460, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001180, Sample Num: 18880, Cur Loss: 0.28713030, Cur Avg Loss: 0.45936972, Log Avg loss: 0.38137018, Global Avg Loss: 1.97208809, Time: 0.0209 Steps: 22470, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001190, Sample Num: 19040, Cur Loss: 0.19273543, Cur Avg Loss: 0.45942608, Log Avg loss: 0.46607655, Global Avg Loss: 1.97141815, Time: 0.0211 Steps: 22480, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001200, Sample Num: 19200, Cur Loss: 0.32893944, Cur Avg Loss: 0.45914023, Log Avg loss: 0.42512443, Global Avg Loss: 1.97073061, Time: 0.0209 Steps: 22490, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001210, Sample Num: 19360, Cur Loss: 0.45755082, Cur Avg Loss: 0.45905710, Log Avg loss: 0.44908071, Global Avg Loss: 1.97005432, Time: 0.0210 Steps: 22500, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001220, Sample Num: 19520, Cur Loss: 0.28667796, Cur Avg Loss: 0.45847601, Log Avg loss: 0.38816420, Global Avg Loss: 1.96935157, Time: 0.0219 Steps: 22510, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001230, Sample Num: 19680, Cur Loss: 0.83534133, Cur Avg Loss: 0.45916832, Log Avg loss: 0.54363093, Global Avg Loss: 1.96871848, Time: 0.0210 Steps: 22520, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001240, Sample Num: 19840, Cur Loss: 0.13263261, Cur Avg Loss: 0.46029470, Log Avg loss: 0.59883873, Global Avg Loss: 1.96811045, Time: 0.0211 Steps: 22530, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001250, Sample Num: 20000, Cur Loss: 0.19532259, Cur Avg Loss: 0.46033543, Log Avg loss: 0.46538615, Global Avg Loss: 1.96744376, Time: 0.0210 Steps: 22540, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001260, Sample Num: 20160, Cur Loss: 0.31531051, Cur Avg Loss: 0.46141951, Log Avg loss: 0.59692922, Global Avg Loss: 1.96683599, Time: 0.0211 Steps: 22550, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001270, Sample Num: 20320, Cur Loss: 0.43466800, Cur Avg Loss: 0.46116087, Log Avg loss: 0.42857296, Global Avg Loss: 1.96615414, Time: 0.0210 Steps: 22560, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001280, Sample Num: 20480, Cur Loss: 1.02389169, Cur Avg Loss: 0.45987361, Log Avg loss: 0.29639092, Global Avg Loss: 1.96541432, Time: 0.0254 Steps: 22570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001290, Sample Num: 20640, Cur Loss: 0.73545235, Cur Avg Loss: 0.45965791, Log Avg loss: 0.43204819, Global Avg Loss: 1.96473524, Time: 0.0208 Steps: 22580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001300, Sample Num: 20800, Cur Loss: 0.53111148, Cur Avg Loss: 0.46268132, Log Avg loss: 0.85270100, Global Avg Loss: 1.96424297, Time: 0.0207 Steps: 22590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001310, Sample Num: 20960, Cur Loss: 0.43231720, Cur Avg Loss: 0.46309159, Log Avg loss: 0.51642755, Global Avg Loss: 1.96360235, Time: 0.0208 Steps: 22600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001320, Sample Num: 21120, Cur Loss: 0.55123436, Cur Avg Loss: 0.46240305, Log Avg loss: 0.37220400, Global Avg Loss: 1.96289850, Time: 0.0207 Steps: 22610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001330, Sample Num: 21280, Cur Loss: 1.00847304, Cur Avg Loss: 0.46288092, Log Avg loss: 0.52596014, Global Avg Loss: 1.96226325, Time: 0.0208 Steps: 22620, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001340, Sample Num: 21440, Cur Loss: 0.36565122, Cur Avg Loss: 0.46387550, Log Avg loss: 0.59615440, Global Avg Loss: 1.96165958, Time: 0.0209 Steps: 22630, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001350, Sample Num: 21600, Cur Loss: 0.35329497, Cur Avg Loss: 0.46346911, Log Avg loss: 0.40901291, Global Avg Loss: 1.96097378, Time: 0.0207 Steps: 22640, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001360, Sample Num: 21760, Cur Loss: 0.18162239, Cur Avg Loss: 0.46363466, Log Avg loss: 0.48598352, Global Avg Loss: 1.96032257, Time: 0.0208 Steps: 22650, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001370, Sample Num: 21920, Cur Loss: 0.49588999, Cur Avg Loss: 0.46309902, Log Avg loss: 0.39025165, Global Avg Loss: 1.95962969, Time: 0.0207 Steps: 22660, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001380, Sample Num: 22080, Cur Loss: 0.20458932, Cur Avg Loss: 0.46283451, Log Avg loss: 0.42659670, Global Avg Loss: 1.95895345, Time: 0.0208 Steps: 22670, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001390, Sample Num: 22240, Cur Loss: 0.11058740, Cur Avg Loss: 0.46139129, Log Avg loss: 0.26222787, Global Avg Loss: 1.95820533, Time: 0.0207 Steps: 22680, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001400, Sample Num: 22400, Cur Loss: 0.50235701, Cur Avg Loss: 0.46082513, Log Avg loss: 0.38212814, Global Avg Loss: 1.95751072, Time: 0.0207 Steps: 22690, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001410, Sample Num: 22560, Cur Loss: 0.14318547, Cur Avg Loss: 0.46096124, Log Avg loss: 0.48001641, Global Avg Loss: 1.95685984, Time: 0.0208 Steps: 22700, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001420, Sample Num: 22720, Cur Loss: 0.98787296, Cur Avg Loss: 0.46152535, Log Avg loss: 0.54106510, Global Avg Loss: 1.95623642, Time: 0.0208 Steps: 22710, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001430, Sample Num: 22880, Cur Loss: 0.16903672, Cur Avg Loss: 0.46098495, Log Avg loss: 0.38424896, Global Avg Loss: 1.95554452, Time: 0.0207 Steps: 22720, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001440, Sample Num: 23040, Cur Loss: 0.69057250, Cur Avg Loss: 0.46220293, Log Avg loss: 0.63637368, Global Avg Loss: 1.95496416, Time: 0.0208 Steps: 22730, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001450, Sample Num: 23200, Cur Loss: 0.06930983, Cur Avg Loss: 0.46175572, Log Avg loss: 0.39735718, Global Avg Loss: 1.95427919, Time: 0.0208 Steps: 22740, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001460, Sample Num: 23360, Cur Loss: 0.52793169, Cur Avg Loss: 0.46175068, Log Avg loss: 0.46101999, Global Avg Loss: 1.95362281, Time: 0.0208 Steps: 22750, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001470, Sample Num: 23520, Cur Loss: 0.66879523, Cur Avg Loss: 0.46280006, Log Avg loss: 0.61600890, Global Avg Loss: 1.95303511, Time: 0.0208 Steps: 22760, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001480, Sample Num: 23680, Cur Loss: 0.41703355, Cur Avg Loss: 0.46373150, Log Avg loss: 0.60065381, Global Avg Loss: 1.95244118, Time: 0.0208 Steps: 22770, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001490, Sample Num: 23840, Cur Loss: 0.94400299, Cur Avg Loss: 0.46521851, Log Avg loss: 0.68529613, Global Avg Loss: 1.95188493, Time: 0.0207 Steps: 22780, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001500, Sample Num: 24000, Cur Loss: 0.62328297, Cur Avg Loss: 0.46743036, Log Avg loss: 0.79699592, Global Avg Loss: 1.95137817, Time: 0.0207 Steps: 22790, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001510, Sample Num: 24160, Cur Loss: 0.96279907, Cur Avg Loss: 0.46944789, Log Avg loss: 0.77207745, Global Avg Loss: 1.95086094, Time: 0.0209 Steps: 22800, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001520, Sample Num: 24320, Cur Loss: 0.43554914, Cur Avg Loss: 0.46918099, Log Avg loss: 0.42887948, Global Avg Loss: 1.95019369, Time: 0.0208 Steps: 22810, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001530, Sample Num: 24480, Cur Loss: 0.34012914, Cur Avg Loss: 0.46864988, Log Avg loss: 0.38792100, Global Avg Loss: 1.94950909, Time: 0.0207 Steps: 22820, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001540, Sample Num: 24640, Cur Loss: 0.13152531, Cur Avg Loss: 0.46788017, Log Avg loss: 0.35011419, Global Avg Loss: 1.94880852, Time: 0.0236 Steps: 22830, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001550, Sample Num: 24800, Cur Loss: 0.56163335, Cur Avg Loss: 0.46814297, Log Avg loss: 0.50861464, Global Avg Loss: 1.94817796, Time: 0.0207 Steps: 22840, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001560, Sample Num: 24960, Cur Loss: 0.17211586, Cur Avg Loss: 0.46745235, Log Avg loss: 0.36040610, Global Avg Loss: 1.94748309, Time: 0.0207 Steps: 22850, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001570, Sample Num: 25120, Cur Loss: 0.23410019, Cur Avg Loss: 0.46619277, Log Avg loss: 0.26969822, Global Avg Loss: 1.94674916, Time: 0.0207 Steps: 22860, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001580, Sample Num: 25280, Cur Loss: 0.28728026, Cur Avg Loss: 0.46576655, Log Avg loss: 0.39885039, Global Avg Loss: 1.94607233, Time: 0.0207 Steps: 22870, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001590, Sample Num: 25440, Cur Loss: 0.18178056, Cur Avg Loss: 0.46470434, Log Avg loss: 0.29687493, Global Avg Loss: 1.94535153, Time: 0.0207 Steps: 22880, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001600, Sample Num: 25600, Cur Loss: 0.37227136, Cur Avg Loss: 0.46379754, Log Avg loss: 0.31961514, Global Avg Loss: 1.94464129, Time: 0.0207 Steps: 22890, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001610, Sample Num: 25760, Cur Loss: 0.37059787, Cur Avg Loss: 0.46283312, Log Avg loss: 0.30852624, Global Avg Loss: 1.94392683, Time: 0.0208 Steps: 22900, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001620, Sample Num: 25920, Cur Loss: 0.60795414, Cur Avg Loss: 0.46345278, Log Avg loss: 0.56321893, Global Avg Loss: 1.94332416, Time: 0.0207 Steps: 22910, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001630, Sample Num: 26080, Cur Loss: 0.83478665, Cur Avg Loss: 0.46596092, Log Avg loss: 0.87227888, Global Avg Loss: 1.94285686, Time: 0.0208 Steps: 22920, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001640, Sample Num: 26240, Cur Loss: 0.39531770, Cur Avg Loss: 0.46540871, Log Avg loss: 0.37539955, Global Avg Loss: 1.94217328, Time: 0.0207 Steps: 22930, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001650, Sample Num: 26400, Cur Loss: 0.30760819, Cur Avg Loss: 0.46534438, Log Avg loss: 0.45479399, Global Avg Loss: 1.94152490, Time: 0.0208 Steps: 22940, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001660, Sample Num: 26560, Cur Loss: 0.13952272, Cur Avg Loss: 0.46516457, Log Avg loss: 0.43549604, Global Avg Loss: 1.94086868, Time: 0.0207 Steps: 22950, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001670, Sample Num: 26720, Cur Loss: 0.77069956, Cur Avg Loss: 0.46472712, Log Avg loss: 0.39210990, Global Avg Loss: 1.94019413, Time: 0.0207 Steps: 22960, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001680, Sample Num: 26880, Cur Loss: 0.99136400, Cur Avg Loss: 0.46632355, Log Avg loss: 0.73292706, Global Avg Loss: 1.93966855, Time: 0.0207 Steps: 22970, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001690, Sample Num: 27040, Cur Loss: 0.22821105, Cur Avg Loss: 0.46750549, Log Avg loss: 0.66607110, Global Avg Loss: 1.93911433, Time: 0.0207 Steps: 22980, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001700, Sample Num: 27200, Cur Loss: 0.17355409, Cur Avg Loss: 0.46796190, Log Avg loss: 0.54509605, Global Avg Loss: 1.93850797, Time: 0.0207 Steps: 22990, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001710, Sample Num: 27360, Cur Loss: 0.16299647, Cur Avg Loss: 0.46772062, Log Avg loss: 0.42670312, Global Avg Loss: 1.93785067, Time: 0.0208 Steps: 23000, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001720, Sample Num: 27520, Cur Loss: 0.59170872, Cur Avg Loss: 0.46814977, Log Avg loss: 0.54153384, Global Avg Loss: 1.93724383, Time: 0.0207 Steps: 23010, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001730, Sample Num: 27680, Cur Loss: 0.56797802, Cur Avg Loss: 0.46815778, Log Avg loss: 0.46953620, Global Avg Loss: 1.93660626, Time: 0.0212 Steps: 23020, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001740, Sample Num: 27840, Cur Loss: 0.62793404, Cur Avg Loss: 0.46793369, Log Avg loss: 0.42916488, Global Avg Loss: 1.93595170, Time: 0.0208 Steps: 23030, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001750, Sample Num: 28000, Cur Loss: 0.13787876, Cur Avg Loss: 0.46714985, Log Avg loss: 0.33076279, Global Avg Loss: 1.93525500, Time: 0.0208 Steps: 23040, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001760, Sample Num: 28160, Cur Loss: 0.74098223, Cur Avg Loss: 0.46782281, Log Avg loss: 0.58559057, Global Avg Loss: 1.93466947, Time: 0.0208 Steps: 23050, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001770, Sample Num: 28320, Cur Loss: 0.44465381, Cur Avg Loss: 0.46833486, Log Avg loss: 0.55845552, Global Avg Loss: 1.93407267, Time: 0.0207 Steps: 23060, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001780, Sample Num: 28480, Cur Loss: 0.22729754, Cur Avg Loss: 0.46744359, Log Avg loss: 0.30968854, Global Avg Loss: 1.93336856, Time: 0.0207 Steps: 23070, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001790, Sample Num: 28640, Cur Loss: 0.36585754, Cur Avg Loss: 0.46685832, Log Avg loss: 0.36268001, Global Avg Loss: 1.93268802, Time: 0.0208 Steps: 23080, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001800, Sample Num: 28800, Cur Loss: 0.52849448, Cur Avg Loss: 0.46713349, Log Avg loss: 0.51638897, Global Avg Loss: 1.93207463, Time: 0.0209 Steps: 23090, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001810, Sample Num: 28960, Cur Loss: 0.63611060, Cur Avg Loss: 0.46637007, Log Avg loss: 0.32895562, Global Avg Loss: 1.93138064, Time: 0.0208 Steps: 23100, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001820, Sample Num: 29120, Cur Loss: 0.79350960, Cur Avg Loss: 0.46587327, Log Avg loss: 0.37595115, Global Avg Loss: 1.93070759, Time: 0.0208 Steps: 23110, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001830, Sample Num: 29280, Cur Loss: 0.21361545, Cur Avg Loss: 0.46565102, Log Avg loss: 0.42520153, Global Avg Loss: 1.93005642, Time: 0.0208 Steps: 23120, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001840, Sample Num: 29440, Cur Loss: 0.54726112, Cur Avg Loss: 0.46517868, Log Avg loss: 0.37874159, Global Avg Loss: 1.92938572, Time: 0.0207 Steps: 23130, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001850, Sample Num: 29600, Cur Loss: 0.58098501, Cur Avg Loss: 0.46616895, Log Avg loss: 0.64837719, Global Avg Loss: 1.92883213, Time: 0.0208 Steps: 23140, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001860, Sample Num: 29760, Cur Loss: 0.84178722, Cur Avg Loss: 0.46585318, Log Avg loss: 0.40743713, Global Avg Loss: 1.92817494, Time: 0.0208 Steps: 23150, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001870, Sample Num: 29920, Cur Loss: 0.36790198, Cur Avg Loss: 0.46564415, Log Avg loss: 0.42676400, Global Avg Loss: 1.92752667, Time: 0.0208 Steps: 23160, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001880, Sample Num: 30080, Cur Loss: 0.13841687, Cur Avg Loss: 0.46519456, Log Avg loss: 0.38112047, Global Avg Loss: 1.92685925, Time: 0.0208 Steps: 23170, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001890, Sample Num: 30240, Cur Loss: 0.14731529, Cur Avg Loss: 0.46470392, Log Avg loss: 0.37246377, Global Avg Loss: 1.92618867, Time: 0.0207 Steps: 23180, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001900, Sample Num: 30400, Cur Loss: 0.14780611, Cur Avg Loss: 0.46385202, Log Avg loss: 0.30284304, Global Avg Loss: 1.92548865, Time: 0.0207 Steps: 23190, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001910, Sample Num: 30560, Cur Loss: 0.98708099, Cur Avg Loss: 0.46348592, Log Avg loss: 0.39392739, Global Avg Loss: 1.92482850, Time: 0.0207 Steps: 23200, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001920, Sample Num: 30720, Cur Loss: 0.48580146, Cur Avg Loss: 0.46352900, Log Avg loss: 0.47175739, Global Avg Loss: 1.92420244, Time: 0.0207 Steps: 23210, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001930, Sample Num: 30880, Cur Loss: 0.40875185, Cur Avg Loss: 0.46402022, Log Avg loss: 0.55833444, Global Avg Loss: 1.92361421, Time: 0.0207 Steps: 23220, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001940, Sample Num: 31040, Cur Loss: 0.10162872, Cur Avg Loss: 0.46383734, Log Avg loss: 0.42854109, Global Avg Loss: 1.92297062, Time: 0.0208 Steps: 23230, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001950, Sample Num: 31200, Cur Loss: 0.40708417, Cur Avg Loss: 0.46295663, Log Avg loss: 0.29209902, Global Avg Loss: 1.92226887, Time: 0.0207 Steps: 23240, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001960, Sample Num: 31360, Cur Loss: 0.71729088, Cur Avg Loss: 0.46248311, Log Avg loss: 0.37014653, Global Avg Loss: 1.92160129, Time: 0.0207 Steps: 23250, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001970, Sample Num: 31520, Cur Loss: 0.14426532, Cur Avg Loss: 0.46144896, Log Avg loss: 0.25875536, Global Avg Loss: 1.92088639, Time: 0.0207 Steps: 23260, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001980, Sample Num: 31680, Cur Loss: 0.62998271, Cur Avg Loss: 0.46122771, Log Avg loss: 0.41764188, Global Avg Loss: 1.92024039, Time: 0.0207 Steps: 23270, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001990, Sample Num: 31840, Cur Loss: 0.39102021, Cur Avg Loss: 0.46036952, Log Avg loss: 0.29044758, Global Avg Loss: 1.91954031, Time: 0.0208 Steps: 23280, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002000, Sample Num: 32000, Cur Loss: 0.23279409, Cur Avg Loss: 0.45948566, Log Avg loss: 0.28359880, Global Avg Loss: 1.91883789, Time: 0.0207 Steps: 23290, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002010, Sample Num: 32160, Cur Loss: 0.52946478, Cur Avg Loss: 0.45955848, Log Avg loss: 0.47412132, Global Avg Loss: 1.91821784, Time: 0.0207 Steps: 23300, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002020, Sample Num: 32320, Cur Loss: 0.32444772, Cur Avg Loss: 0.45881780, Log Avg loss: 0.30994228, Global Avg Loss: 1.91752789, Time: 0.0211 Steps: 23310, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002030, Sample Num: 32480, Cur Loss: 0.86428666, Cur Avg Loss: 0.45877663, Log Avg loss: 0.45045873, Global Avg Loss: 1.91689878, Time: 0.0211 Steps: 23320, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002040, Sample Num: 32640, Cur Loss: 0.25034058, Cur Avg Loss: 0.45828494, Log Avg loss: 0.35847172, Global Avg Loss: 1.91623079, Time: 0.0209 Steps: 23330, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002050, Sample Num: 32800, Cur Loss: 0.12817815, Cur Avg Loss: 0.45796343, Log Avg loss: 0.39237563, Global Avg Loss: 1.91557789, Time: 0.0247 Steps: 23340, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002060, Sample Num: 32960, Cur Loss: 0.09978180, Cur Avg Loss: 0.45799742, Log Avg loss: 0.46496527, Global Avg Loss: 1.91495665, Time: 0.0209 Steps: 23350, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002070, Sample Num: 33120, Cur Loss: 0.09957901, Cur Avg Loss: 0.45756325, Log Avg loss: 0.36812442, Global Avg Loss: 1.91429448, Time: 0.0208 Steps: 23360, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002080, Sample Num: 33280, Cur Loss: 0.74164730, Cur Avg Loss: 0.45759619, Log Avg loss: 0.46441472, Global Avg Loss: 1.91367407, Time: 0.0209 Steps: 23370, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002090, Sample Num: 33440, Cur Loss: 0.84782380, Cur Avg Loss: 0.45724276, Log Avg loss: 0.38373022, Global Avg Loss: 1.91301969, Time: 0.0209 Steps: 23380, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002100, Sample Num: 33600, Cur Loss: 0.17519173, Cur Avg Loss: 0.45658994, Log Avg loss: 0.32014967, Global Avg Loss: 1.91233869, Time: 0.0210 Steps: 23390, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002110, Sample Num: 33760, Cur Loss: 0.12643172, Cur Avg Loss: 0.45571511, Log Avg loss: 0.27200067, Global Avg Loss: 1.91163769, Time: 0.0209 Steps: 23400, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002120, Sample Num: 33920, Cur Loss: 0.08869380, Cur Avg Loss: 0.45515606, Log Avg loss: 0.33719663, Global Avg Loss: 1.91096514, Time: 0.0209 Steps: 23410, Updated lr: 0.000079 ***** Running evaluation checkpoint-23419 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-23419 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.738445, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.440881, "eval_total_loss": 309.939391, "eval_mae": 0.452945, "eval_mse": 0.44105, "eval_r2": 0.71964, "eval_sp_statistic": 0.849175, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.883288, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.60264, "test_total_loss": 302.525293, "test_mae": 0.488411, "test_mse": 0.602817, "test_r2": 0.610937, "test_sp_statistic": 0.834768, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.884032, "test_ps_pvalue": 0.0, "lr": 7.873968705547652e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.9104276320235425, "train_cur_epoch_loss": 969.5416746325791, "train_cur_epoch_avg_loss": 0.45539768653479523, "train_cur_epoch_time": 44.73844528198242, "train_cur_epoch_avg_time": 0.021013830569273095, "epoch": 11, "step": 23419} ################################################## Training, Epoch: 0012, Batch: 000001, Sample Num: 16, Cur Loss: 0.29135397, Cur Avg Loss: 0.29135397, Log Avg loss: 0.49021882, Global Avg Loss: 1.91035850, Time: 0.0248 Steps: 23420, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000011, Sample Num: 176, Cur Loss: 0.29359302, Cur Avg Loss: 0.51999602, Log Avg loss: 0.54286023, Global Avg Loss: 1.90977485, Time: 0.0210 Steps: 23430, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000021, Sample Num: 336, Cur Loss: 0.18609668, Cur Avg Loss: 0.44115747, Log Avg loss: 0.35443507, Global Avg Loss: 1.90911131, Time: 0.0209 Steps: 23440, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000031, Sample Num: 496, Cur Loss: 0.64567965, Cur Avg Loss: 0.43432295, Log Avg loss: 0.41997046, Global Avg Loss: 1.90847628, Time: 0.0209 Steps: 23450, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000041, Sample Num: 656, Cur Loss: 0.77782482, Cur Avg Loss: 0.42258215, Log Avg loss: 0.38618567, Global Avg Loss: 1.90782739, Time: 0.0210 Steps: 23460, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000051, Sample Num: 816, Cur Loss: 0.31270295, Cur Avg Loss: 0.39738284, Log Avg loss: 0.29406564, Global Avg Loss: 1.90713981, Time: 0.0209 Steps: 23470, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000061, Sample Num: 976, Cur Loss: 0.36104256, Cur Avg Loss: 0.44374699, Log Avg loss: 0.68020416, Global Avg Loss: 1.90661726, Time: 0.0209 Steps: 23480, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000071, Sample Num: 1136, Cur Loss: 0.11547698, Cur Avg Loss: 0.47089192, Log Avg loss: 0.63647600, Global Avg Loss: 1.90607654, Time: 0.0210 Steps: 23490, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000081, Sample Num: 1296, Cur Loss: 0.64536119, Cur Avg Loss: 0.46353874, Log Avg loss: 0.41133116, Global Avg Loss: 1.90544048, Time: 0.0209 Steps: 23500, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000091, Sample Num: 1456, Cur Loss: 0.42385402, Cur Avg Loss: 0.47200811, Log Avg loss: 0.54061005, Global Avg Loss: 1.90485995, Time: 0.0209 Steps: 23510, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000101, Sample Num: 1616, Cur Loss: 0.11070925, Cur Avg Loss: 0.45651850, Log Avg loss: 0.31556305, Global Avg Loss: 1.90418423, Time: 0.0209 Steps: 23520, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000111, Sample Num: 1776, Cur Loss: 0.30861413, Cur Avg Loss: 0.45353769, Log Avg loss: 0.42343144, Global Avg Loss: 1.90355493, Time: 0.0209 Steps: 23530, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000121, Sample Num: 1936, Cur Loss: 0.21645525, Cur Avg Loss: 0.45610538, Log Avg loss: 0.48460679, Global Avg Loss: 1.90295214, Time: 0.0210 Steps: 23540, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000131, Sample Num: 2096, Cur Loss: 0.19751228, Cur Avg Loss: 0.44714549, Log Avg loss: 0.33873087, Global Avg Loss: 1.90228793, Time: 0.0210 Steps: 23550, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000141, Sample Num: 2256, Cur Loss: 0.23489162, Cur Avg Loss: 0.43919633, Log Avg loss: 0.33506234, Global Avg Loss: 1.90162272, Time: 0.0209 Steps: 23560, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000151, Sample Num: 2416, Cur Loss: 0.17350024, Cur Avg Loss: 0.42788968, Log Avg loss: 0.26846580, Global Avg Loss: 1.90092983, Time: 0.0210 Steps: 23570, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000161, Sample Num: 2576, Cur Loss: 0.33504009, Cur Avg Loss: 0.42385942, Log Avg loss: 0.36300257, Global Avg Loss: 1.90027761, Time: 0.0209 Steps: 23580, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000171, Sample Num: 2736, Cur Loss: 0.22001919, Cur Avg Loss: 0.42530673, Log Avg loss: 0.44860839, Global Avg Loss: 1.89966224, Time: 0.0209 Steps: 23590, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000181, Sample Num: 2896, Cur Loss: 0.48047212, Cur Avg Loss: 0.42379144, Log Avg loss: 0.39788003, Global Avg Loss: 1.89902589, Time: 0.0209 Steps: 23600, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000191, Sample Num: 3056, Cur Loss: 0.13812575, Cur Avg Loss: 0.41580641, Log Avg loss: 0.27127727, Global Avg Loss: 1.89833646, Time: 0.0209 Steps: 23610, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000201, Sample Num: 3216, Cur Loss: 0.36240381, Cur Avg Loss: 0.41803580, Log Avg loss: 0.46061725, Global Avg Loss: 1.89772777, Time: 0.0209 Steps: 23620, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000211, Sample Num: 3376, Cur Loss: 0.27382094, Cur Avg Loss: 0.41712263, Log Avg loss: 0.39876783, Global Avg Loss: 1.89709342, Time: 0.0209 Steps: 23630, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000221, Sample Num: 3536, Cur Loss: 0.26236415, Cur Avg Loss: 0.41739579, Log Avg loss: 0.42315957, Global Avg Loss: 1.89646993, Time: 0.0209 Steps: 23640, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000231, Sample Num: 3696, Cur Loss: 0.43975437, Cur Avg Loss: 0.41667458, Log Avg loss: 0.40073570, Global Avg Loss: 1.89583749, Time: 0.0209 Steps: 23650, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000241, Sample Num: 3856, Cur Loss: 0.46728706, Cur Avg Loss: 0.41375180, Log Avg loss: 0.34623564, Global Avg Loss: 1.89518254, Time: 0.0210 Steps: 23660, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000251, Sample Num: 4016, Cur Loss: 0.47158727, Cur Avg Loss: 0.41427924, Log Avg loss: 0.42699057, Global Avg Loss: 1.89456226, Time: 0.0210 Steps: 23670, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000261, Sample Num: 4176, Cur Loss: 0.29097003, Cur Avg Loss: 0.41375338, Log Avg loss: 0.40055425, Global Avg Loss: 1.89393135, Time: 0.0209 Steps: 23680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000271, Sample Num: 4336, Cur Loss: 0.23241422, Cur Avg Loss: 0.40781154, Log Avg loss: 0.25272946, Global Avg Loss: 1.89323857, Time: 0.0209 Steps: 23690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000281, Sample Num: 4496, Cur Loss: 0.34523153, Cur Avg Loss: 0.40530328, Log Avg loss: 0.33732946, Global Avg Loss: 1.89258206, Time: 0.0209 Steps: 23700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000291, Sample Num: 4656, Cur Loss: 0.47305915, Cur Avg Loss: 0.40273612, Log Avg loss: 0.33059887, Global Avg Loss: 1.89192328, Time: 0.0209 Steps: 23710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000301, Sample Num: 4816, Cur Loss: 0.25080156, Cur Avg Loss: 0.40033910, Log Avg loss: 0.33058605, Global Avg Loss: 1.89126504, Time: 0.0210 Steps: 23720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000311, Sample Num: 4976, Cur Loss: 0.72759163, Cur Avg Loss: 0.40446969, Log Avg loss: 0.52880045, Global Avg Loss: 1.89069089, Time: 0.0209 Steps: 23730, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000321, Sample Num: 5136, Cur Loss: 0.19534810, Cur Avg Loss: 0.40345559, Log Avg loss: 0.37191709, Global Avg Loss: 1.89005114, Time: 0.0209 Steps: 23740, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000331, Sample Num: 5296, Cur Loss: 0.63327110, Cur Avg Loss: 0.40436631, Log Avg loss: 0.43360041, Global Avg Loss: 1.88943789, Time: 0.0209 Steps: 23750, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000341, Sample Num: 5456, Cur Loss: 0.24009827, Cur Avg Loss: 0.40747521, Log Avg loss: 0.51037963, Global Avg Loss: 1.88885748, Time: 0.0209 Steps: 23760, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000351, Sample Num: 5616, Cur Loss: 0.34474421, Cur Avg Loss: 0.40847772, Log Avg loss: 0.44266354, Global Avg Loss: 1.88824907, Time: 0.0208 Steps: 23770, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000361, Sample Num: 5776, Cur Loss: 0.42144012, Cur Avg Loss: 0.40798443, Log Avg loss: 0.39066996, Global Avg Loss: 1.88761931, Time: 0.0209 Steps: 23780, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000371, Sample Num: 5936, Cur Loss: 0.30565178, Cur Avg Loss: 0.40705947, Log Avg loss: 0.37366813, Global Avg Loss: 1.88698292, Time: 0.0208 Steps: 23790, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000381, Sample Num: 6096, Cur Loss: 0.73149627, Cur Avg Loss: 0.41986087, Log Avg loss: 0.89479310, Global Avg Loss: 1.88656604, Time: 0.0208 Steps: 23800, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000391, Sample Num: 6256, Cur Loss: 0.25947040, Cur Avg Loss: 0.42038299, Log Avg loss: 0.44027561, Global Avg Loss: 1.88595861, Time: 0.0208 Steps: 23810, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000401, Sample Num: 6416, Cur Loss: 0.38511461, Cur Avg Loss: 0.41973293, Log Avg loss: 0.39431575, Global Avg Loss: 1.88533239, Time: 0.0208 Steps: 23820, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000411, Sample Num: 6576, Cur Loss: 0.28506106, Cur Avg Loss: 0.41690782, Log Avg loss: 0.30362070, Global Avg Loss: 1.88466865, Time: 0.0208 Steps: 23830, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000421, Sample Num: 6736, Cur Loss: 0.41961181, Cur Avg Loss: 0.41540545, Log Avg loss: 0.35365825, Global Avg Loss: 1.88402644, Time: 0.0208 Steps: 23840, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000431, Sample Num: 6896, Cur Loss: 0.23203230, Cur Avg Loss: 0.41665529, Log Avg loss: 0.46927348, Global Avg Loss: 1.88343326, Time: 0.0207 Steps: 23850, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000441, Sample Num: 7056, Cur Loss: 0.29541174, Cur Avg Loss: 0.41829972, Log Avg loss: 0.48917438, Global Avg Loss: 1.88284891, Time: 0.0208 Steps: 23860, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000451, Sample Num: 7216, Cur Loss: 0.68684012, Cur Avg Loss: 0.42084286, Log Avg loss: 0.53299546, Global Avg Loss: 1.88228340, Time: 0.0208 Steps: 23870, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000461, Sample Num: 7376, Cur Loss: 0.10057562, Cur Avg Loss: 0.42254161, Log Avg loss: 0.49915519, Global Avg Loss: 1.88170420, Time: 0.0208 Steps: 23880, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000471, Sample Num: 7536, Cur Loss: 0.16595086, Cur Avg Loss: 0.42122873, Log Avg loss: 0.36070497, Global Avg Loss: 1.88106754, Time: 0.0208 Steps: 23890, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000481, Sample Num: 7696, Cur Loss: 0.24274921, Cur Avg Loss: 0.42143255, Log Avg loss: 0.43103241, Global Avg Loss: 1.88046083, Time: 0.0208 Steps: 23900, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000491, Sample Num: 7856, Cur Loss: 0.27583152, Cur Avg Loss: 0.41849734, Log Avg loss: 0.27731402, Global Avg Loss: 1.87979033, Time: 0.0208 Steps: 23910, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000501, Sample Num: 8016, Cur Loss: 0.45475066, Cur Avg Loss: 0.41598414, Log Avg loss: 0.29258605, Global Avg Loss: 1.87912679, Time: 0.0207 Steps: 23920, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000511, Sample Num: 8176, Cur Loss: 0.06911773, Cur Avg Loss: 0.41585847, Log Avg loss: 0.40956211, Global Avg Loss: 1.87851268, Time: 0.0208 Steps: 23930, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000521, Sample Num: 8336, Cur Loss: 0.45310396, Cur Avg Loss: 0.41414669, Log Avg loss: 0.32667467, Global Avg Loss: 1.87786446, Time: 0.0209 Steps: 23940, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000531, Sample Num: 8496, Cur Loss: 0.94801199, Cur Avg Loss: 0.41531640, Log Avg loss: 0.47625866, Global Avg Loss: 1.87727924, Time: 0.0208 Steps: 23950, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000541, Sample Num: 8656, Cur Loss: 0.44199416, Cur Avg Loss: 0.41799363, Log Avg loss: 0.56015441, Global Avg Loss: 1.87672952, Time: 0.0208 Steps: 23960, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000551, Sample Num: 8816, Cur Loss: 0.57313067, Cur Avg Loss: 0.41759148, Log Avg loss: 0.39583521, Global Avg Loss: 1.87611171, Time: 0.0208 Steps: 23970, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000561, Sample Num: 8976, Cur Loss: 0.28436702, Cur Avg Loss: 0.41569406, Log Avg loss: 0.31114632, Global Avg Loss: 1.87545909, Time: 0.0209 Steps: 23980, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000571, Sample Num: 9136, Cur Loss: 0.21107574, Cur Avg Loss: 0.41685874, Log Avg loss: 0.48219733, Global Avg Loss: 1.87487833, Time: 0.0207 Steps: 23990, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000581, Sample Num: 9296, Cur Loss: 0.21801651, Cur Avg Loss: 0.42215460, Log Avg loss: 0.72454772, Global Avg Loss: 1.87439902, Time: 0.0208 Steps: 24000, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000591, Sample Num: 9456, Cur Loss: 0.20918925, Cur Avg Loss: 0.42015165, Log Avg loss: 0.30378070, Global Avg Loss: 1.87374487, Time: 0.0207 Steps: 24010, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000601, Sample Num: 9616, Cur Loss: 0.66716081, Cur Avg Loss: 0.41799181, Log Avg loss: 0.29034518, Global Avg Loss: 1.87308567, Time: 0.0207 Steps: 24020, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000611, Sample Num: 9776, Cur Loss: 0.48345321, Cur Avg Loss: 0.41643092, Log Avg loss: 0.32262112, Global Avg Loss: 1.87244045, Time: 0.0209 Steps: 24030, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000621, Sample Num: 9936, Cur Loss: 0.23272291, Cur Avg Loss: 0.41759921, Log Avg loss: 0.48898214, Global Avg Loss: 1.87186497, Time: 0.0207 Steps: 24040, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000631, Sample Num: 10096, Cur Loss: 0.21174076, Cur Avg Loss: 0.41904501, Log Avg loss: 0.50882917, Global Avg Loss: 1.87129822, Time: 0.0209 Steps: 24050, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000641, Sample Num: 10256, Cur Loss: 0.74717033, Cur Avg Loss: 0.41810251, Log Avg loss: 0.35863065, Global Avg Loss: 1.87066951, Time: 0.0208 Steps: 24060, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000651, Sample Num: 10416, Cur Loss: 0.46676326, Cur Avg Loss: 0.41790990, Log Avg loss: 0.40556334, Global Avg Loss: 1.87006082, Time: 0.0208 Steps: 24070, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000661, Sample Num: 10576, Cur Loss: 0.17040206, Cur Avg Loss: 0.41602911, Log Avg loss: 0.29358972, Global Avg Loss: 1.86940614, Time: 0.0208 Steps: 24080, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000671, Sample Num: 10736, Cur Loss: 0.13527364, Cur Avg Loss: 0.41390577, Log Avg loss: 0.27355346, Global Avg Loss: 1.86874369, Time: 0.0209 Steps: 24090, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000681, Sample Num: 10896, Cur Loss: 0.49069732, Cur Avg Loss: 0.41411211, Log Avg loss: 0.42795726, Global Avg Loss: 1.86814585, Time: 0.0208 Steps: 24100, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000691, Sample Num: 11056, Cur Loss: 0.33452305, Cur Avg Loss: 0.41704352, Log Avg loss: 0.61667230, Global Avg Loss: 1.86762678, Time: 0.0208 Steps: 24110, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000701, Sample Num: 11216, Cur Loss: 0.24417476, Cur Avg Loss: 0.41687707, Log Avg loss: 0.40537541, Global Avg Loss: 1.86702054, Time: 0.0209 Steps: 24120, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000711, Sample Num: 11376, Cur Loss: 0.41281211, Cur Avg Loss: 0.41493093, Log Avg loss: 0.27850693, Global Avg Loss: 1.86636223, Time: 0.0208 Steps: 24130, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000721, Sample Num: 11536, Cur Loss: 0.13786364, Cur Avg Loss: 0.41322215, Log Avg loss: 0.29172796, Global Avg Loss: 1.86570994, Time: 0.0209 Steps: 24140, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000731, Sample Num: 11696, Cur Loss: 0.42400074, Cur Avg Loss: 0.41365401, Log Avg loss: 0.44479082, Global Avg Loss: 1.86512157, Time: 0.0207 Steps: 24150, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000741, Sample Num: 11856, Cur Loss: 0.27711236, Cur Avg Loss: 0.41197671, Log Avg loss: 0.28936637, Global Avg Loss: 1.86446935, Time: 0.0207 Steps: 24160, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000751, Sample Num: 12016, Cur Loss: 0.32710898, Cur Avg Loss: 0.41284542, Log Avg loss: 0.47721685, Global Avg Loss: 1.86389539, Time: 0.0209 Steps: 24170, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000761, Sample Num: 12176, Cur Loss: 0.38441175, Cur Avg Loss: 0.41315777, Log Avg loss: 0.43661495, Global Avg Loss: 1.86330512, Time: 0.0208 Steps: 24180, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000771, Sample Num: 12336, Cur Loss: 0.45303863, Cur Avg Loss: 0.41166608, Log Avg loss: 0.29814826, Global Avg Loss: 1.86265809, Time: 0.0245 Steps: 24190, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000781, Sample Num: 12496, Cur Loss: 1.69772446, Cur Avg Loss: 0.41745870, Log Avg loss: 0.86406969, Global Avg Loss: 1.86224545, Time: 0.0207 Steps: 24200, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000791, Sample Num: 12656, Cur Loss: 1.37612474, Cur Avg Loss: 0.42254267, Log Avg loss: 0.81960061, Global Avg Loss: 1.86181479, Time: 0.0208 Steps: 24210, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000801, Sample Num: 12816, Cur Loss: 0.21951272, Cur Avg Loss: 0.42349399, Log Avg loss: 0.49874385, Global Avg Loss: 1.86125200, Time: 0.0207 Steps: 24220, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000811, Sample Num: 12976, Cur Loss: 0.25697738, Cur Avg Loss: 0.42210837, Log Avg loss: 0.31112023, Global Avg Loss: 1.86061224, Time: 0.0207 Steps: 24230, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000821, Sample Num: 13136, Cur Loss: 0.40853202, Cur Avg Loss: 0.42165696, Log Avg loss: 0.38504720, Global Avg Loss: 1.86000351, Time: 0.0207 Steps: 24240, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000831, Sample Num: 13296, Cur Loss: 0.47560480, Cur Avg Loss: 0.42105874, Log Avg loss: 0.37194493, Global Avg Loss: 1.85938988, Time: 0.0207 Steps: 24250, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000841, Sample Num: 13456, Cur Loss: 0.76563847, Cur Avg Loss: 0.42060134, Log Avg loss: 0.38259140, Global Avg Loss: 1.85878114, Time: 0.0207 Steps: 24260, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000851, Sample Num: 13616, Cur Loss: 0.67259037, Cur Avg Loss: 0.41987975, Log Avg loss: 0.35919461, Global Avg Loss: 1.85816326, Time: 0.0208 Steps: 24270, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000861, Sample Num: 13776, Cur Loss: 0.53827882, Cur Avg Loss: 0.42287244, Log Avg loss: 0.67754985, Global Avg Loss: 1.85767701, Time: 0.0208 Steps: 24280, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000871, Sample Num: 13936, Cur Loss: 0.50949574, Cur Avg Loss: 0.42245594, Log Avg loss: 0.38659572, Global Avg Loss: 1.85707138, Time: 0.0208 Steps: 24290, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000881, Sample Num: 14096, Cur Loss: 0.28866482, Cur Avg Loss: 0.42182154, Log Avg loss: 0.36656498, Global Avg Loss: 1.85645800, Time: 0.0207 Steps: 24300, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000891, Sample Num: 14256, Cur Loss: 0.66435331, Cur Avg Loss: 0.42285014, Log Avg loss: 0.51346975, Global Avg Loss: 1.85590556, Time: 0.0207 Steps: 24310, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000901, Sample Num: 14416, Cur Loss: 0.67177582, Cur Avg Loss: 0.42441881, Log Avg loss: 0.56418709, Global Avg Loss: 1.85537443, Time: 0.0207 Steps: 24320, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000911, Sample Num: 14576, Cur Loss: 0.50544751, Cur Avg Loss: 0.42769229, Log Avg loss: 0.72263329, Global Avg Loss: 1.85490885, Time: 0.0208 Steps: 24330, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000921, Sample Num: 14736, Cur Loss: 0.27770770, Cur Avg Loss: 0.42823491, Log Avg loss: 0.47766775, Global Avg Loss: 1.85434302, Time: 0.0207 Steps: 24340, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000931, Sample Num: 14896, Cur Loss: 0.64272219, Cur Avg Loss: 0.42788010, Log Avg loss: 0.39520175, Global Avg Loss: 1.85374378, Time: 0.0207 Steps: 24350, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000941, Sample Num: 15056, Cur Loss: 0.51978636, Cur Avg Loss: 0.42726905, Log Avg loss: 0.37038046, Global Avg Loss: 1.85313485, Time: 0.0207 Steps: 24360, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000951, Sample Num: 15216, Cur Loss: 0.10201359, Cur Avg Loss: 0.42745615, Log Avg loss: 0.44506163, Global Avg Loss: 1.85255706, Time: 0.0208 Steps: 24370, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000961, Sample Num: 15376, Cur Loss: 0.28195712, Cur Avg Loss: 0.42685343, Log Avg loss: 0.36953563, Global Avg Loss: 1.85194876, Time: 0.0208 Steps: 24380, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000971, Sample Num: 15536, Cur Loss: 0.30030307, Cur Avg Loss: 0.42703734, Log Avg loss: 0.44471098, Global Avg Loss: 1.85137179, Time: 0.0207 Steps: 24390, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000981, Sample Num: 15696, Cur Loss: 1.24369788, Cur Avg Loss: 0.43013248, Log Avg loss: 0.73067001, Global Avg Loss: 1.85091249, Time: 0.0208 Steps: 24400, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000991, Sample Num: 15856, Cur Loss: 0.09735067, Cur Avg Loss: 0.43094761, Log Avg loss: 0.51091238, Global Avg Loss: 1.85036353, Time: 0.0207 Steps: 24410, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001001, Sample Num: 16016, Cur Loss: 0.33836019, Cur Avg Loss: 0.43332916, Log Avg loss: 0.66934012, Global Avg Loss: 1.84987990, Time: 0.0207 Steps: 24420, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001011, Sample Num: 16176, Cur Loss: 0.10734141, Cur Avg Loss: 0.43331968, Log Avg loss: 0.43237118, Global Avg Loss: 1.84929967, Time: 0.0207 Steps: 24430, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001021, Sample Num: 16336, Cur Loss: 0.22291937, Cur Avg Loss: 0.43303568, Log Avg loss: 0.40432277, Global Avg Loss: 1.84870843, Time: 0.0207 Steps: 24440, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001031, Sample Num: 16496, Cur Loss: 0.70364231, Cur Avg Loss: 0.43356151, Log Avg loss: 0.48724965, Global Avg Loss: 1.84815160, Time: 0.0228 Steps: 24450, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001041, Sample Num: 16656, Cur Loss: 0.37041828, Cur Avg Loss: 0.43258011, Log Avg loss: 0.33139747, Global Avg Loss: 1.84753150, Time: 0.0207 Steps: 24460, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001051, Sample Num: 16816, Cur Loss: 0.84342217, Cur Avg Loss: 0.43188995, Log Avg loss: 0.36004391, Global Avg Loss: 1.84692362, Time: 0.0209 Steps: 24470, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001061, Sample Num: 16976, Cur Loss: 0.63184202, Cur Avg Loss: 0.43229474, Log Avg loss: 0.47483878, Global Avg Loss: 1.84636313, Time: 0.0210 Steps: 24480, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001071, Sample Num: 17136, Cur Loss: 0.64180487, Cur Avg Loss: 0.43144477, Log Avg loss: 0.34126239, Global Avg Loss: 1.84574855, Time: 0.0209 Steps: 24490, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001081, Sample Num: 17296, Cur Loss: 0.33020392, Cur Avg Loss: 0.43110017, Log Avg loss: 0.39419365, Global Avg Loss: 1.84515608, Time: 0.0209 Steps: 24500, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001091, Sample Num: 17456, Cur Loss: 0.42693675, Cur Avg Loss: 0.42957286, Log Avg loss: 0.26447097, Global Avg Loss: 1.84451117, Time: 0.0209 Steps: 24510, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001101, Sample Num: 17616, Cur Loss: 0.25423148, Cur Avg Loss: 0.42984470, Log Avg loss: 0.45950267, Global Avg Loss: 1.84394632, Time: 0.0209 Steps: 24520, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001111, Sample Num: 17776, Cur Loss: 0.95221126, Cur Avg Loss: 0.43099163, Log Avg loss: 0.55726851, Global Avg Loss: 1.84342179, Time: 0.0209 Steps: 24530, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001121, Sample Num: 17936, Cur Loss: 0.33043396, Cur Avg Loss: 0.43080585, Log Avg loss: 0.41016492, Global Avg Loss: 1.84283774, Time: 0.0209 Steps: 24540, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001131, Sample Num: 18096, Cur Loss: 0.33221903, Cur Avg Loss: 0.43089283, Log Avg loss: 0.44064371, Global Avg Loss: 1.84226658, Time: 0.0208 Steps: 24550, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001141, Sample Num: 18256, Cur Loss: 0.65237945, Cur Avg Loss: 0.43027930, Log Avg loss: 0.36088939, Global Avg Loss: 1.84166341, Time: 0.0209 Steps: 24560, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001151, Sample Num: 18416, Cur Loss: 0.17875671, Cur Avg Loss: 0.43036168, Log Avg loss: 0.43976116, Global Avg Loss: 1.84109284, Time: 0.0212 Steps: 24570, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001161, Sample Num: 18576, Cur Loss: 0.21343611, Cur Avg Loss: 0.42919310, Log Avg loss: 0.29468921, Global Avg Loss: 1.84046371, Time: 0.0210 Steps: 24580, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001171, Sample Num: 18736, Cur Loss: 0.30077437, Cur Avg Loss: 0.42955210, Log Avg loss: 0.47123254, Global Avg Loss: 1.83990688, Time: 0.0210 Steps: 24590, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001181, Sample Num: 18896, Cur Loss: 0.05901808, Cur Avg Loss: 0.42884283, Log Avg loss: 0.34578736, Global Avg Loss: 1.83929952, Time: 0.0211 Steps: 24600, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001191, Sample Num: 19056, Cur Loss: 0.57030523, Cur Avg Loss: 0.42965340, Log Avg loss: 0.52538098, Global Avg Loss: 1.83876562, Time: 0.0210 Steps: 24610, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001201, Sample Num: 19216, Cur Loss: 0.37551433, Cur Avg Loss: 0.42859456, Log Avg loss: 0.30248735, Global Avg Loss: 1.83814162, Time: 0.0210 Steps: 24620, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001211, Sample Num: 19376, Cur Loss: 0.23286806, Cur Avg Loss: 0.43067213, Log Avg loss: 0.68018775, Global Avg Loss: 1.83767148, Time: 0.0210 Steps: 24630, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001221, Sample Num: 19536, Cur Loss: 0.55997968, Cur Avg Loss: 0.43318381, Log Avg loss: 0.73734789, Global Avg Loss: 1.83722492, Time: 0.0210 Steps: 24640, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001231, Sample Num: 19696, Cur Loss: 0.06654605, Cur Avg Loss: 0.43309822, Log Avg loss: 0.42264872, Global Avg Loss: 1.83665106, Time: 0.0210 Steps: 24650, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001241, Sample Num: 19856, Cur Loss: 0.18352161, Cur Avg Loss: 0.43180372, Log Avg loss: 0.27245020, Global Avg Loss: 1.83601675, Time: 0.0210 Steps: 24660, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001251, Sample Num: 20016, Cur Loss: 0.28254434, Cur Avg Loss: 0.43121208, Log Avg loss: 0.35778982, Global Avg Loss: 1.83541755, Time: 0.0210 Steps: 24670, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001261, Sample Num: 20176, Cur Loss: 0.54359061, Cur Avg Loss: 0.43117071, Log Avg loss: 0.42599565, Global Avg Loss: 1.83484647, Time: 0.0210 Steps: 24680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001271, Sample Num: 20336, Cur Loss: 0.78149009, Cur Avg Loss: 0.43075627, Log Avg loss: 0.37849428, Global Avg Loss: 1.83425662, Time: 0.0210 Steps: 24690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001281, Sample Num: 20496, Cur Loss: 0.19441542, Cur Avg Loss: 0.43078101, Log Avg loss: 0.43392562, Global Avg Loss: 1.83368968, Time: 0.0247 Steps: 24700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001291, Sample Num: 20656, Cur Loss: 0.55676770, Cur Avg Loss: 0.43034364, Log Avg loss: 0.37431727, Global Avg Loss: 1.83309908, Time: 0.0211 Steps: 24710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001301, Sample Num: 20816, Cur Loss: 0.22442576, Cur Avg Loss: 0.42950348, Log Avg loss: 0.32103826, Global Avg Loss: 1.83248741, Time: 0.0210 Steps: 24720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001311, Sample Num: 20976, Cur Loss: 0.15981939, Cur Avg Loss: 0.42992175, Log Avg loss: 0.48433846, Global Avg Loss: 1.83194226, Time: 0.0210 Steps: 24730, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001321, Sample Num: 21136, Cur Loss: 0.86880189, Cur Avg Loss: 0.43089385, Log Avg loss: 0.55833691, Global Avg Loss: 1.83142747, Time: 0.0211 Steps: 24740, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001331, Sample Num: 21296, Cur Loss: 0.24151802, Cur Avg Loss: 0.43030800, Log Avg loss: 0.35291750, Global Avg Loss: 1.83083009, Time: 0.0210 Steps: 24750, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001341, Sample Num: 21456, Cur Loss: 1.03909421, Cur Avg Loss: 0.43032029, Log Avg loss: 0.43195520, Global Avg Loss: 1.83026511, Time: 0.0210 Steps: 24760, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001351, Sample Num: 21616, Cur Loss: 0.39356846, Cur Avg Loss: 0.42963305, Log Avg loss: 0.33747452, Global Avg Loss: 1.82966245, Time: 0.0210 Steps: 24770, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001361, Sample Num: 21776, Cur Loss: 0.76668572, Cur Avg Loss: 0.42906842, Log Avg loss: 0.35278677, Global Avg Loss: 1.82906646, Time: 0.0210 Steps: 24780, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001371, Sample Num: 21936, Cur Loss: 0.20024300, Cur Avg Loss: 0.42849787, Log Avg loss: 0.35084581, Global Avg Loss: 1.82847016, Time: 0.0211 Steps: 24790, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001381, Sample Num: 22096, Cur Loss: 1.03002119, Cur Avg Loss: 0.42836848, Log Avg loss: 0.41062954, Global Avg Loss: 1.82789845, Time: 0.0210 Steps: 24800, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001391, Sample Num: 22256, Cur Loss: 0.31499198, Cur Avg Loss: 0.42814757, Log Avg loss: 0.39764022, Global Avg Loss: 1.82732197, Time: 0.0210 Steps: 24810, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001401, Sample Num: 22416, Cur Loss: 0.55959564, Cur Avg Loss: 0.42771727, Log Avg loss: 0.36786244, Global Avg Loss: 1.82673395, Time: 0.0210 Steps: 24820, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001411, Sample Num: 22576, Cur Loss: 0.34120840, Cur Avg Loss: 0.42792752, Log Avg loss: 0.45738299, Global Avg Loss: 1.82618246, Time: 0.0211 Steps: 24830, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001421, Sample Num: 22736, Cur Loss: 0.53691828, Cur Avg Loss: 0.42781654, Log Avg loss: 0.41215679, Global Avg Loss: 1.82561320, Time: 0.0210 Steps: 24840, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001431, Sample Num: 22896, Cur Loss: 0.23342144, Cur Avg Loss: 0.42729797, Log Avg loss: 0.35360919, Global Avg Loss: 1.82502085, Time: 0.0211 Steps: 24850, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001441, Sample Num: 23056, Cur Loss: 0.11161370, Cur Avg Loss: 0.42631427, Log Avg loss: 0.28554692, Global Avg Loss: 1.82440159, Time: 0.0210 Steps: 24860, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001451, Sample Num: 23216, Cur Loss: 0.26128766, Cur Avg Loss: 0.42550508, Log Avg loss: 0.30890148, Global Avg Loss: 1.82379222, Time: 0.0210 Steps: 24870, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001461, Sample Num: 23376, Cur Loss: 0.22609819, Cur Avg Loss: 0.42541089, Log Avg loss: 0.41174406, Global Avg Loss: 1.82322468, Time: 0.0210 Steps: 24880, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001471, Sample Num: 23536, Cur Loss: 0.14771351, Cur Avg Loss: 0.42473648, Log Avg loss: 0.32620412, Global Avg Loss: 1.82262322, Time: 0.0210 Steps: 24890, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001481, Sample Num: 23696, Cur Loss: 0.45754763, Cur Avg Loss: 0.42420296, Log Avg loss: 0.34572298, Global Avg Loss: 1.82203009, Time: 0.0210 Steps: 24900, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001491, Sample Num: 23856, Cur Loss: 0.26848772, Cur Avg Loss: 0.42404288, Log Avg loss: 0.40033539, Global Avg Loss: 1.82145936, Time: 0.0210 Steps: 24910, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001501, Sample Num: 24016, Cur Loss: 0.61808050, Cur Avg Loss: 0.42327810, Log Avg loss: 0.30924886, Global Avg Loss: 1.82085253, Time: 0.0211 Steps: 24920, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001511, Sample Num: 24176, Cur Loss: 0.97436059, Cur Avg Loss: 0.42330815, Log Avg loss: 0.42781897, Global Avg Loss: 1.82029376, Time: 0.0210 Steps: 24930, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001521, Sample Num: 24336, Cur Loss: 0.46145874, Cur Avg Loss: 0.42360408, Log Avg loss: 0.46831847, Global Avg Loss: 1.81975166, Time: 0.0210 Steps: 24940, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001531, Sample Num: 24496, Cur Loss: 0.36683920, Cur Avg Loss: 0.42324579, Log Avg loss: 0.36875017, Global Avg Loss: 1.81917010, Time: 0.0210 Steps: 24950, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001541, Sample Num: 24656, Cur Loss: 0.55833423, Cur Avg Loss: 0.42250177, Log Avg loss: 0.30859217, Global Avg Loss: 1.81856490, Time: 0.0227 Steps: 24960, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001551, Sample Num: 24816, Cur Loss: 0.24114093, Cur Avg Loss: 0.42238832, Log Avg loss: 0.40490518, Global Avg Loss: 1.81799876, Time: 0.0210 Steps: 24970, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001561, Sample Num: 24976, Cur Loss: 0.47400713, Cur Avg Loss: 0.42200355, Log Avg loss: 0.36232605, Global Avg Loss: 1.81741602, Time: 0.0212 Steps: 24980, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001571, Sample Num: 25136, Cur Loss: 0.76718211, Cur Avg Loss: 0.42249830, Log Avg loss: 0.49972894, Global Avg Loss: 1.81688874, Time: 0.0211 Steps: 24990, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001581, Sample Num: 25296, Cur Loss: 0.05636337, Cur Avg Loss: 0.42291743, Log Avg loss: 0.48876223, Global Avg Loss: 1.81635749, Time: 0.0211 Steps: 25000, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001591, Sample Num: 25456, Cur Loss: 0.97102326, Cur Avg Loss: 0.42353336, Log Avg loss: 0.52091327, Global Avg Loss: 1.81583952, Time: 0.0210 Steps: 25010, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001601, Sample Num: 25616, Cur Loss: 0.20295396, Cur Avg Loss: 0.42328906, Log Avg loss: 0.38441953, Global Avg Loss: 1.81526741, Time: 0.0213 Steps: 25020, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001611, Sample Num: 25776, Cur Loss: 0.18304011, Cur Avg Loss: 0.42244015, Log Avg loss: 0.28653099, Global Avg Loss: 1.81465664, Time: 0.0213 Steps: 25030, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001621, Sample Num: 25936, Cur Loss: 0.52959776, Cur Avg Loss: 0.42165055, Log Avg loss: 0.29444621, Global Avg Loss: 1.81404953, Time: 0.0211 Steps: 25040, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001631, Sample Num: 26096, Cur Loss: 0.37412310, Cur Avg Loss: 0.42105033, Log Avg loss: 0.32375418, Global Avg Loss: 1.81345460, Time: 0.0213 Steps: 25050, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001641, Sample Num: 26256, Cur Loss: 0.43980318, Cur Avg Loss: 0.42074068, Log Avg loss: 0.37023593, Global Avg Loss: 1.81287870, Time: 0.0212 Steps: 25060, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001651, Sample Num: 26416, Cur Loss: 0.40734753, Cur Avg Loss: 0.42059335, Log Avg loss: 0.39641776, Global Avg Loss: 1.81231370, Time: 0.0211 Steps: 25070, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001661, Sample Num: 26576, Cur Loss: 0.14726666, Cur Avg Loss: 0.42001377, Log Avg loss: 0.32432422, Global Avg Loss: 1.81172040, Time: 0.0213 Steps: 25080, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001671, Sample Num: 26736, Cur Loss: 0.19713396, Cur Avg Loss: 0.41943351, Log Avg loss: 0.32305195, Global Avg Loss: 1.81112707, Time: 0.0210 Steps: 25090, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001681, Sample Num: 26896, Cur Loss: 0.28033623, Cur Avg Loss: 0.41959146, Log Avg loss: 0.44598474, Global Avg Loss: 1.81058319, Time: 0.0211 Steps: 25100, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001691, Sample Num: 27056, Cur Loss: 0.39707583, Cur Avg Loss: 0.41891835, Log Avg loss: 0.30576902, Global Avg Loss: 1.80998390, Time: 0.0210 Steps: 25110, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001701, Sample Num: 27216, Cur Loss: 0.30016240, Cur Avg Loss: 0.41877064, Log Avg loss: 0.39379292, Global Avg Loss: 1.80942013, Time: 0.0210 Steps: 25120, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001711, Sample Num: 27376, Cur Loss: 0.17954130, Cur Avg Loss: 0.41755692, Log Avg loss: 0.21110324, Global Avg Loss: 1.80878411, Time: 0.0210 Steps: 25130, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001721, Sample Num: 27536, Cur Loss: 0.36241755, Cur Avg Loss: 0.41694093, Log Avg loss: 0.31154603, Global Avg Loss: 1.80818855, Time: 0.0212 Steps: 25140, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001731, Sample Num: 27696, Cur Loss: 0.12195334, Cur Avg Loss: 0.41627176, Log Avg loss: 0.30110647, Global Avg Loss: 1.80758931, Time: 0.0213 Steps: 25150, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001741, Sample Num: 27856, Cur Loss: 0.74040860, Cur Avg Loss: 0.41629617, Log Avg loss: 0.42052178, Global Avg Loss: 1.80703801, Time: 0.0210 Steps: 25160, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001751, Sample Num: 28016, Cur Loss: 0.14580539, Cur Avg Loss: 0.41637069, Log Avg loss: 0.42934514, Global Avg Loss: 1.80649066, Time: 0.0213 Steps: 25170, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001761, Sample Num: 28176, Cur Loss: 0.25850379, Cur Avg Loss: 0.41562977, Log Avg loss: 0.28589506, Global Avg Loss: 1.80588677, Time: 0.0210 Steps: 25180, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001771, Sample Num: 28336, Cur Loss: 0.54952568, Cur Avg Loss: 0.41574714, Log Avg loss: 0.43641596, Global Avg Loss: 1.80534311, Time: 0.0210 Steps: 25190, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001781, Sample Num: 28496, Cur Loss: 0.38907707, Cur Avg Loss: 0.41605217, Log Avg loss: 0.47007146, Global Avg Loss: 1.80481324, Time: 0.0210 Steps: 25200, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001791, Sample Num: 28656, Cur Loss: 0.14680333, Cur Avg Loss: 0.41534704, Log Avg loss: 0.28976385, Global Avg Loss: 1.80421227, Time: 0.0213 Steps: 25210, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001801, Sample Num: 28816, Cur Loss: 0.24505076, Cur Avg Loss: 0.41531101, Log Avg loss: 0.40885832, Global Avg Loss: 1.80365899, Time: 0.0211 Steps: 25220, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001811, Sample Num: 28976, Cur Loss: 0.46273828, Cur Avg Loss: 0.41490735, Log Avg loss: 0.34220755, Global Avg Loss: 1.80307974, Time: 0.0211 Steps: 25230, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001821, Sample Num: 29136, Cur Loss: 0.34873143, Cur Avg Loss: 0.41505362, Log Avg loss: 0.44154325, Global Avg Loss: 1.80254031, Time: 0.0211 Steps: 25240, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001831, Sample Num: 29296, Cur Loss: 0.63473558, Cur Avg Loss: 0.41458759, Log Avg loss: 0.32972487, Global Avg Loss: 1.80195701, Time: 0.0211 Steps: 25250, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001841, Sample Num: 29456, Cur Loss: 0.30787018, Cur Avg Loss: 0.41537035, Log Avg loss: 0.55869298, Global Avg Loss: 1.80146483, Time: 0.0211 Steps: 25260, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001851, Sample Num: 29616, Cur Loss: 0.47938982, Cur Avg Loss: 0.41623944, Log Avg loss: 0.57623883, Global Avg Loss: 1.80097997, Time: 0.0211 Steps: 25270, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001861, Sample Num: 29776, Cur Loss: 1.05088305, Cur Avg Loss: 0.41688903, Log Avg loss: 0.53712749, Global Avg Loss: 1.80048003, Time: 0.0211 Steps: 25280, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001871, Sample Num: 29936, Cur Loss: 0.95116431, Cur Avg Loss: 0.41741520, Log Avg loss: 0.51533688, Global Avg Loss: 1.79997187, Time: 0.0210 Steps: 25290, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001881, Sample Num: 30096, Cur Loss: 0.66992325, Cur Avg Loss: 0.41815488, Log Avg loss: 0.55654815, Global Avg Loss: 1.79948040, Time: 0.0211 Steps: 25300, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001891, Sample Num: 30256, Cur Loss: 0.39666480, Cur Avg Loss: 0.41796190, Log Avg loss: 0.38166285, Global Avg Loss: 1.79892022, Time: 0.0211 Steps: 25310, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001901, Sample Num: 30416, Cur Loss: 0.40690818, Cur Avg Loss: 0.41802131, Log Avg loss: 0.42925569, Global Avg Loss: 1.79837927, Time: 0.0211 Steps: 25320, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001911, Sample Num: 30576, Cur Loss: 0.63860840, Cur Avg Loss: 0.41782963, Log Avg loss: 0.38138988, Global Avg Loss: 1.79781986, Time: 0.0212 Steps: 25330, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001921, Sample Num: 30736, Cur Loss: 0.19578534, Cur Avg Loss: 0.41715090, Log Avg loss: 0.28744714, Global Avg Loss: 1.79722382, Time: 0.0211 Steps: 25340, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001931, Sample Num: 30896, Cur Loss: 0.49291569, Cur Avg Loss: 0.41626568, Log Avg loss: 0.24621407, Global Avg Loss: 1.79661198, Time: 0.0211 Steps: 25350, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001941, Sample Num: 31056, Cur Loss: 0.70436394, Cur Avg Loss: 0.41672014, Log Avg loss: 0.50447756, Global Avg Loss: 1.79610247, Time: 0.0211 Steps: 25360, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001951, Sample Num: 31216, Cur Loss: 0.30396640, Cur Avg Loss: 0.41611827, Log Avg loss: 0.29929499, Global Avg Loss: 1.79551247, Time: 0.0211 Steps: 25370, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001961, Sample Num: 31376, Cur Loss: 0.25742111, Cur Avg Loss: 0.41651805, Log Avg loss: 0.49451504, Global Avg Loss: 1.79499987, Time: 0.0211 Steps: 25380, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001971, Sample Num: 31536, Cur Loss: 0.14980869, Cur Avg Loss: 0.41696052, Log Avg loss: 0.50372863, Global Avg Loss: 1.79449129, Time: 0.0211 Steps: 25390, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001981, Sample Num: 31696, Cur Loss: 0.18821129, Cur Avg Loss: 0.41652975, Log Avg loss: 0.33162556, Global Avg Loss: 1.79391536, Time: 0.0211 Steps: 25400, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001991, Sample Num: 31856, Cur Loss: 0.21831521, Cur Avg Loss: 0.41667714, Log Avg loss: 0.44587509, Global Avg Loss: 1.79338484, Time: 0.0210 Steps: 25410, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002001, Sample Num: 32016, Cur Loss: 0.18596098, Cur Avg Loss: 0.41625306, Log Avg loss: 0.33181760, Global Avg Loss: 1.79280988, Time: 0.0211 Steps: 25420, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002011, Sample Num: 32176, Cur Loss: 0.30165821, Cur Avg Loss: 0.41616820, Log Avg loss: 0.39918844, Global Avg Loss: 1.79226185, Time: 0.0211 Steps: 25430, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002021, Sample Num: 32336, Cur Loss: 0.37717783, Cur Avg Loss: 0.41555220, Log Avg loss: 0.29167480, Global Avg Loss: 1.79167200, Time: 0.0211 Steps: 25440, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002031, Sample Num: 32496, Cur Loss: 0.18025061, Cur Avg Loss: 0.41546732, Log Avg loss: 0.39831321, Global Avg Loss: 1.79112451, Time: 0.0211 Steps: 25450, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002041, Sample Num: 32656, Cur Loss: 0.70016623, Cur Avg Loss: 0.41648724, Log Avg loss: 0.62363207, Global Avg Loss: 1.79066595, Time: 0.0212 Steps: 25460, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002051, Sample Num: 32816, Cur Loss: 0.38604552, Cur Avg Loss: 0.41639355, Log Avg loss: 0.39727233, Global Avg Loss: 1.79011888, Time: 0.0248 Steps: 25470, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002061, Sample Num: 32976, Cur Loss: 0.38593027, Cur Avg Loss: 0.41629988, Log Avg loss: 0.39708699, Global Avg Loss: 1.78957216, Time: 0.0211 Steps: 25480, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002071, Sample Num: 33136, Cur Loss: 0.77639067, Cur Avg Loss: 0.41615920, Log Avg loss: 0.38716537, Global Avg Loss: 1.78902199, Time: 0.0211 Steps: 25490, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002081, Sample Num: 33296, Cur Loss: 0.56560397, Cur Avg Loss: 0.41610044, Log Avg loss: 0.40393036, Global Avg Loss: 1.78847881, Time: 0.0212 Steps: 25500, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002091, Sample Num: 33456, Cur Loss: 0.32910645, Cur Avg Loss: 0.41583996, Log Avg loss: 0.36163554, Global Avg Loss: 1.78791949, Time: 0.0211 Steps: 25510, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002101, Sample Num: 33616, Cur Loss: 0.41729540, Cur Avg Loss: 0.41531790, Log Avg loss: 0.30615503, Global Avg Loss: 1.78733886, Time: 0.0212 Steps: 25520, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002111, Sample Num: 33776, Cur Loss: 0.25492880, Cur Avg Loss: 0.41496066, Log Avg loss: 0.33990482, Global Avg Loss: 1.78677190, Time: 0.0211 Steps: 25530, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002121, Sample Num: 33936, Cur Loss: 0.53497845, Cur Avg Loss: 0.41454514, Log Avg loss: 0.32682794, Global Avg Loss: 1.78620027, Time: 0.0211 Steps: 25540, Updated lr: 0.000077 ***** Running evaluation checkpoint-25548 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-25548 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.840370, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.5432, "eval_total_loss": 381.869388, "eval_mae": 0.613822, "eval_mse": 0.543274, "eval_r2": 0.654659, "eval_sp_statistic": 0.847793, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.886885, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.394577, "test_total_loss": 198.077836, "test_mae": 0.480096, "test_mse": 0.39471, "test_r2": 0.745251, "test_sp_statistic": 0.840122, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.893316, "test_ps_pvalue": 0.0, "lr": 7.672072072072073e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7857608194157473, "train_cur_epoch_loss": 882.3127000741661, "train_cur_epoch_avg_loss": 0.41442588072999814, "train_cur_epoch_time": 44.840370178222656, "train_cur_epoch_avg_time": 0.021061705109545633, "epoch": 12, "step": 25548} ################################################## Training, Epoch: 0013, Batch: 000002, Sample Num: 32, Cur Loss: 0.32715750, Cur Avg Loss: 0.33115782, Log Avg loss: 0.37247762, Global Avg Loss: 1.78564696, Time: 0.0249 Steps: 25550, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000012, Sample Num: 192, Cur Loss: 0.15402883, Cur Avg Loss: 0.43816622, Log Avg loss: 0.45956791, Global Avg Loss: 1.78512815, Time: 0.0210 Steps: 25560, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000022, Sample Num: 352, Cur Loss: 0.51443344, Cur Avg Loss: 0.40490541, Log Avg loss: 0.36499244, Global Avg Loss: 1.78457275, Time: 0.0211 Steps: 25570, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000032, Sample Num: 512, Cur Loss: 0.21873707, Cur Avg Loss: 0.42032249, Log Avg loss: 0.45424005, Global Avg Loss: 1.78405269, Time: 0.0210 Steps: 25580, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000042, Sample Num: 672, Cur Loss: 0.15140669, Cur Avg Loss: 0.48881480, Log Avg loss: 0.70799020, Global Avg Loss: 1.78363219, Time: 0.0210 Steps: 25590, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000052, Sample Num: 832, Cur Loss: 0.32995236, Cur Avg Loss: 0.47348861, Log Avg loss: 0.40911864, Global Avg Loss: 1.78309527, Time: 0.0211 Steps: 25600, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000062, Sample Num: 992, Cur Loss: 0.31113404, Cur Avg Loss: 0.44809114, Log Avg loss: 0.31602429, Global Avg Loss: 1.78252242, Time: 0.0211 Steps: 25610, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000072, Sample Num: 1152, Cur Loss: 0.22091347, Cur Avg Loss: 0.45452966, Log Avg loss: 0.49444850, Global Avg Loss: 1.78201965, Time: 0.0211 Steps: 25620, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000082, Sample Num: 1312, Cur Loss: 0.23307405, Cur Avg Loss: 0.45177981, Log Avg loss: 0.43198089, Global Avg Loss: 1.78149291, Time: 0.0211 Steps: 25630, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000092, Sample Num: 1472, Cur Loss: 0.17068174, Cur Avg Loss: 0.45627779, Log Avg loss: 0.49316115, Global Avg Loss: 1.78099044, Time: 0.0211 Steps: 25640, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000102, Sample Num: 1632, Cur Loss: 0.23800239, Cur Avg Loss: 0.46247187, Log Avg loss: 0.51945749, Global Avg Loss: 1.78049862, Time: 0.0211 Steps: 25650, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000112, Sample Num: 1792, Cur Loss: 0.21854061, Cur Avg Loss: 0.44879740, Log Avg loss: 0.30931779, Global Avg Loss: 1.77992528, Time: 0.0211 Steps: 25660, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000122, Sample Num: 1952, Cur Loss: 0.54262829, Cur Avg Loss: 0.43992483, Log Avg loss: 0.34055198, Global Avg Loss: 1.77936456, Time: 0.0211 Steps: 25670, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000132, Sample Num: 2112, Cur Loss: 0.27731583, Cur Avg Loss: 0.43359063, Log Avg loss: 0.35631342, Global Avg Loss: 1.77881041, Time: 0.0211 Steps: 25680, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000142, Sample Num: 2272, Cur Loss: 0.23030989, Cur Avg Loss: 0.42229841, Log Avg loss: 0.27324108, Global Avg Loss: 1.77822436, Time: 0.0211 Steps: 25690, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000152, Sample Num: 2432, Cur Loss: 0.91621208, Cur Avg Loss: 0.42126857, Log Avg loss: 0.40664492, Global Avg Loss: 1.77769067, Time: 0.0212 Steps: 25700, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000162, Sample Num: 2592, Cur Loss: 0.46623874, Cur Avg Loss: 0.41970194, Log Avg loss: 0.39588909, Global Avg Loss: 1.77715321, Time: 0.0211 Steps: 25710, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000172, Sample Num: 2752, Cur Loss: 0.25518245, Cur Avg Loss: 0.41983068, Log Avg loss: 0.42191627, Global Avg Loss: 1.77662629, Time: 0.0211 Steps: 25720, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000182, Sample Num: 2912, Cur Loss: 0.31889924, Cur Avg Loss: 0.41386362, Log Avg loss: 0.31123031, Global Avg Loss: 1.77605677, Time: 0.0211 Steps: 25730, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000192, Sample Num: 3072, Cur Loss: 0.22366972, Cur Avg Loss: 0.40713883, Log Avg loss: 0.28474749, Global Avg Loss: 1.77547739, Time: 0.0211 Steps: 25740, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000202, Sample Num: 3232, Cur Loss: 0.12288389, Cur Avg Loss: 0.40040067, Log Avg loss: 0.27102807, Global Avg Loss: 1.77489314, Time: 0.0211 Steps: 25750, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000212, Sample Num: 3392, Cur Loss: 0.27052796, Cur Avg Loss: 0.39884774, Log Avg loss: 0.36747863, Global Avg Loss: 1.77434678, Time: 0.0211 Steps: 25760, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000222, Sample Num: 3552, Cur Loss: 0.28389832, Cur Avg Loss: 0.39819419, Log Avg loss: 0.38433881, Global Avg Loss: 1.77380739, Time: 0.0211 Steps: 25770, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000232, Sample Num: 3712, Cur Loss: 0.11058041, Cur Avg Loss: 0.39392387, Log Avg loss: 0.29912288, Global Avg Loss: 1.77323537, Time: 0.0211 Steps: 25780, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000242, Sample Num: 3872, Cur Loss: 0.24862668, Cur Avg Loss: 0.39369363, Log Avg loss: 0.38835205, Global Avg Loss: 1.77269838, Time: 0.0211 Steps: 25790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000252, Sample Num: 4032, Cur Loss: 0.76853865, Cur Avg Loss: 0.39655094, Log Avg loss: 0.46569781, Global Avg Loss: 1.77219179, Time: 0.0211 Steps: 25800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000262, Sample Num: 4192, Cur Loss: 0.49244174, Cur Avg Loss: 0.39639992, Log Avg loss: 0.39259416, Global Avg Loss: 1.77165727, Time: 0.0231 Steps: 25810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000272, Sample Num: 4352, Cur Loss: 0.65426683, Cur Avg Loss: 0.40116295, Log Avg loss: 0.52595442, Global Avg Loss: 1.77117482, Time: 0.0211 Steps: 25820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000282, Sample Num: 4512, Cur Loss: 1.08910835, Cur Avg Loss: 0.40514164, Log Avg loss: 0.51336199, Global Avg Loss: 1.77068786, Time: 0.0211 Steps: 25830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000292, Sample Num: 4672, Cur Loss: 0.47641459, Cur Avg Loss: 0.40339015, Log Avg loss: 0.35399812, Global Avg Loss: 1.77013960, Time: 0.0211 Steps: 25840, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000302, Sample Num: 4832, Cur Loss: 1.09965062, Cur Avg Loss: 0.40989307, Log Avg loss: 0.59977844, Global Avg Loss: 1.76968685, Time: 0.0212 Steps: 25850, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000312, Sample Num: 4992, Cur Loss: 0.16622674, Cur Avg Loss: 0.40791339, Log Avg loss: 0.34812702, Global Avg Loss: 1.76913714, Time: 0.0210 Steps: 25860, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000322, Sample Num: 5152, Cur Loss: 0.40988612, Cur Avg Loss: 0.40606191, Log Avg loss: 0.34829560, Global Avg Loss: 1.76858791, Time: 0.0211 Steps: 25870, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000332, Sample Num: 5312, Cur Loss: 0.22290626, Cur Avg Loss: 0.40816456, Log Avg loss: 0.47587005, Global Avg Loss: 1.76808841, Time: 0.0211 Steps: 25880, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000342, Sample Num: 5472, Cur Loss: 0.30863851, Cur Avg Loss: 0.40671982, Log Avg loss: 0.35875445, Global Avg Loss: 1.76754406, Time: 0.0211 Steps: 25890, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000352, Sample Num: 5632, Cur Loss: 1.04997301, Cur Avg Loss: 0.40539043, Log Avg loss: 0.35992517, Global Avg Loss: 1.76700057, Time: 0.0211 Steps: 25900, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000362, Sample Num: 5792, Cur Loss: 0.39002594, Cur Avg Loss: 0.41290401, Log Avg loss: 0.67738216, Global Avg Loss: 1.76658003, Time: 0.0211 Steps: 25910, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000372, Sample Num: 5952, Cur Loss: 0.17806605, Cur Avg Loss: 0.41305932, Log Avg loss: 0.41868158, Global Avg Loss: 1.76606001, Time: 0.0211 Steps: 25920, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000382, Sample Num: 6112, Cur Loss: 0.35260737, Cur Avg Loss: 0.41227157, Log Avg loss: 0.38296696, Global Avg Loss: 1.76552662, Time: 0.0211 Steps: 25930, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000392, Sample Num: 6272, Cur Loss: 0.40804294, Cur Avg Loss: 0.41434195, Log Avg loss: 0.49343055, Global Avg Loss: 1.76503622, Time: 0.0211 Steps: 25940, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000402, Sample Num: 6432, Cur Loss: 0.15301487, Cur Avg Loss: 0.41068899, Log Avg loss: 0.26749291, Global Avg Loss: 1.76445913, Time: 0.0211 Steps: 25950, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000412, Sample Num: 6592, Cur Loss: 0.43878448, Cur Avg Loss: 0.40862537, Log Avg loss: 0.32566805, Global Avg Loss: 1.76390489, Time: 0.0212 Steps: 25960, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000422, Sample Num: 6752, Cur Loss: 0.34998506, Cur Avg Loss: 0.40561954, Log Avg loss: 0.28177940, Global Avg Loss: 1.76333419, Time: 0.0211 Steps: 25970, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000432, Sample Num: 6912, Cur Loss: 0.44789785, Cur Avg Loss: 0.40349433, Log Avg loss: 0.31381044, Global Avg Loss: 1.76277625, Time: 0.0211 Steps: 25980, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000442, Sample Num: 7072, Cur Loss: 0.54674685, Cur Avg Loss: 0.40516946, Log Avg loss: 0.47753517, Global Avg Loss: 1.76228174, Time: 0.0211 Steps: 25990, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000452, Sample Num: 7232, Cur Loss: 0.38405263, Cur Avg Loss: 0.40328611, Log Avg loss: 0.32004170, Global Avg Loss: 1.76172703, Time: 0.0211 Steps: 26000, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000462, Sample Num: 7392, Cur Loss: 0.35407162, Cur Avg Loss: 0.40355876, Log Avg loss: 0.41588263, Global Avg Loss: 1.76120959, Time: 0.0210 Steps: 26010, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000472, Sample Num: 7552, Cur Loss: 0.38410473, Cur Avg Loss: 0.40236158, Log Avg loss: 0.34705205, Global Avg Loss: 1.76066611, Time: 0.0211 Steps: 26020, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000482, Sample Num: 7712, Cur Loss: 0.36855257, Cur Avg Loss: 0.40404822, Log Avg loss: 0.48365763, Global Avg Loss: 1.76017552, Time: 0.0211 Steps: 26030, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000492, Sample Num: 7872, Cur Loss: 0.37458587, Cur Avg Loss: 0.40257562, Log Avg loss: 0.33159592, Global Avg Loss: 1.75962691, Time: 0.0211 Steps: 26040, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000502, Sample Num: 8032, Cur Loss: 0.49898216, Cur Avg Loss: 0.40259584, Log Avg loss: 0.40359067, Global Avg Loss: 1.75910635, Time: 0.0211 Steps: 26050, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000512, Sample Num: 8192, Cur Loss: 0.50366807, Cur Avg Loss: 0.40214331, Log Avg loss: 0.37942628, Global Avg Loss: 1.75857693, Time: 0.0255 Steps: 26060, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000522, Sample Num: 8352, Cur Loss: 0.28002775, Cur Avg Loss: 0.40002516, Log Avg loss: 0.29157606, Global Avg Loss: 1.75801421, Time: 0.0211 Steps: 26070, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000532, Sample Num: 8512, Cur Loss: 0.31622437, Cur Avg Loss: 0.40072347, Log Avg loss: 0.43717515, Global Avg Loss: 1.75750776, Time: 0.0211 Steps: 26080, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000542, Sample Num: 8672, Cur Loss: 0.47988272, Cur Avg Loss: 0.40381416, Log Avg loss: 0.56823904, Global Avg Loss: 1.75705192, Time: 0.0210 Steps: 26090, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000552, Sample Num: 8832, Cur Loss: 0.27612019, Cur Avg Loss: 0.40174759, Log Avg loss: 0.28973946, Global Avg Loss: 1.75648974, Time: 0.0211 Steps: 26100, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000562, Sample Num: 8992, Cur Loss: 0.45105156, Cur Avg Loss: 0.40094722, Log Avg loss: 0.35676711, Global Avg Loss: 1.75595365, Time: 0.0211 Steps: 26110, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000572, Sample Num: 9152, Cur Loss: 0.10611142, Cur Avg Loss: 0.39907613, Log Avg loss: 0.29392067, Global Avg Loss: 1.75539391, Time: 0.0211 Steps: 26120, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000582, Sample Num: 9312, Cur Loss: 0.60416579, Cur Avg Loss: 0.39761701, Log Avg loss: 0.31415506, Global Avg Loss: 1.75484235, Time: 0.0210 Steps: 26130, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000592, Sample Num: 9472, Cur Loss: 0.48255086, Cur Avg Loss: 0.40002384, Log Avg loss: 0.54010150, Global Avg Loss: 1.75437764, Time: 0.0211 Steps: 26140, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000602, Sample Num: 9632, Cur Loss: 0.50929111, Cur Avg Loss: 0.39942113, Log Avg loss: 0.36374062, Global Avg Loss: 1.75384585, Time: 0.0211 Steps: 26150, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000612, Sample Num: 9792, Cur Loss: 0.83850449, Cur Avg Loss: 0.40052913, Log Avg loss: 0.46723111, Global Avg Loss: 1.75335402, Time: 0.0211 Steps: 26160, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000622, Sample Num: 9952, Cur Loss: 0.25861195, Cur Avg Loss: 0.39946086, Log Avg loss: 0.33408249, Global Avg Loss: 1.75281170, Time: 0.0211 Steps: 26170, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000632, Sample Num: 10112, Cur Loss: 0.26921141, Cur Avg Loss: 0.39918307, Log Avg loss: 0.38190478, Global Avg Loss: 1.75228805, Time: 0.0211 Steps: 26180, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000642, Sample Num: 10272, Cur Loss: 0.35755184, Cur Avg Loss: 0.39951018, Log Avg loss: 0.42018356, Global Avg Loss: 1.75177942, Time: 0.0211 Steps: 26190, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000652, Sample Num: 10432, Cur Loss: 0.60609382, Cur Avg Loss: 0.40433151, Log Avg loss: 0.71386063, Global Avg Loss: 1.75138327, Time: 0.0211 Steps: 26200, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000662, Sample Num: 10592, Cur Loss: 0.23731893, Cur Avg Loss: 0.40497107, Log Avg loss: 0.44667027, Global Avg Loss: 1.75088547, Time: 0.0211 Steps: 26210, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000672, Sample Num: 10752, Cur Loss: 0.11731618, Cur Avg Loss: 0.40303637, Log Avg loss: 0.27495934, Global Avg Loss: 1.75032257, Time: 0.0211 Steps: 26220, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000682, Sample Num: 10912, Cur Loss: 1.57964814, Cur Avg Loss: 0.40500096, Log Avg loss: 0.53702117, Global Avg Loss: 1.74986001, Time: 0.0211 Steps: 26230, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000692, Sample Num: 11072, Cur Loss: 0.40323269, Cur Avg Loss: 0.40444842, Log Avg loss: 0.36676552, Global Avg Loss: 1.74933292, Time: 0.0211 Steps: 26240, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000702, Sample Num: 11232, Cur Loss: 0.49207348, Cur Avg Loss: 0.40418223, Log Avg loss: 0.38576191, Global Avg Loss: 1.74881346, Time: 0.0212 Steps: 26250, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000712, Sample Num: 11392, Cur Loss: 0.32213479, Cur Avg Loss: 0.40344373, Log Avg loss: 0.35160083, Global Avg Loss: 1.74828139, Time: 0.0212 Steps: 26260, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000722, Sample Num: 11552, Cur Loss: 0.22689128, Cur Avg Loss: 0.40231020, Log Avg loss: 0.32160280, Global Avg Loss: 1.74773831, Time: 0.0211 Steps: 26270, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000732, Sample Num: 11712, Cur Loss: 0.44084024, Cur Avg Loss: 0.40220134, Log Avg loss: 0.39434198, Global Avg Loss: 1.74722332, Time: 0.0211 Steps: 26280, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000742, Sample Num: 11872, Cur Loss: 0.40379336, Cur Avg Loss: 0.40303746, Log Avg loss: 0.46424091, Global Avg Loss: 1.74673531, Time: 0.0211 Steps: 26290, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000752, Sample Num: 12032, Cur Loss: 0.46954742, Cur Avg Loss: 0.40255793, Log Avg loss: 0.36697720, Global Avg Loss: 1.74621068, Time: 0.0211 Steps: 26300, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000762, Sample Num: 12192, Cur Loss: 0.24832241, Cur Avg Loss: 0.40331785, Log Avg loss: 0.46046398, Global Avg Loss: 1.74572199, Time: 0.0211 Steps: 26310, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000772, Sample Num: 12352, Cur Loss: 0.95533943, Cur Avg Loss: 0.40390191, Log Avg loss: 0.44840688, Global Avg Loss: 1.74522909, Time: 0.0247 Steps: 26320, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000782, Sample Num: 12512, Cur Loss: 0.82827383, Cur Avg Loss: 0.40348334, Log Avg loss: 0.37117022, Global Avg Loss: 1.74470723, Time: 0.0211 Steps: 26330, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000792, Sample Num: 12672, Cur Loss: 0.20522642, Cur Avg Loss: 0.40262278, Log Avg loss: 0.33532670, Global Avg Loss: 1.74417216, Time: 0.0210 Steps: 26340, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000802, Sample Num: 12832, Cur Loss: 0.24633642, Cur Avg Loss: 0.40267888, Log Avg loss: 0.40712212, Global Avg Loss: 1.74366474, Time: 0.0210 Steps: 26350, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000812, Sample Num: 12992, Cur Loss: 0.32217979, Cur Avg Loss: 0.40148461, Log Avg loss: 0.30570376, Global Avg Loss: 1.74311923, Time: 0.0210 Steps: 26360, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000822, Sample Num: 13152, Cur Loss: 0.33554703, Cur Avg Loss: 0.40211152, Log Avg loss: 0.45301708, Global Avg Loss: 1.74263000, Time: 0.0210 Steps: 26370, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000832, Sample Num: 13312, Cur Loss: 0.53197002, Cur Avg Loss: 0.40287054, Log Avg loss: 0.46526169, Global Avg Loss: 1.74214578, Time: 0.0209 Steps: 26380, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000842, Sample Num: 13472, Cur Loss: 0.25864452, Cur Avg Loss: 0.40188125, Log Avg loss: 0.31957213, Global Avg Loss: 1.74160672, Time: 0.0211 Steps: 26390, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000852, Sample Num: 13632, Cur Loss: 0.56826764, Cur Avg Loss: 0.40056217, Log Avg loss: 0.28949570, Global Avg Loss: 1.74105668, Time: 0.0210 Steps: 26400, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000862, Sample Num: 13792, Cur Loss: 0.88887966, Cur Avg Loss: 0.40105626, Log Avg loss: 0.44315327, Global Avg Loss: 1.74056524, Time: 0.0210 Steps: 26410, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000872, Sample Num: 13952, Cur Loss: 0.70765865, Cur Avg Loss: 0.40121651, Log Avg loss: 0.41502955, Global Avg Loss: 1.74006352, Time: 0.0210 Steps: 26420, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000882, Sample Num: 14112, Cur Loss: 0.36284870, Cur Avg Loss: 0.40216839, Log Avg loss: 0.48517259, Global Avg Loss: 1.73958872, Time: 0.0211 Steps: 26430, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000892, Sample Num: 14272, Cur Loss: 0.28073186, Cur Avg Loss: 0.40170543, Log Avg loss: 0.36087200, Global Avg Loss: 1.73906727, Time: 0.0210 Steps: 26440, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000902, Sample Num: 14432, Cur Loss: 0.32811895, Cur Avg Loss: 0.40044562, Log Avg loss: 0.28807091, Global Avg Loss: 1.73851869, Time: 0.0210 Steps: 26450, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000912, Sample Num: 14592, Cur Loss: 0.11531049, Cur Avg Loss: 0.39952042, Log Avg loss: 0.31606762, Global Avg Loss: 1.73798111, Time: 0.0210 Steps: 26460, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000922, Sample Num: 14752, Cur Loss: 1.54942560, Cur Avg Loss: 0.40115890, Log Avg loss: 0.55058824, Global Avg Loss: 1.73753252, Time: 0.0210 Steps: 26470, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000932, Sample Num: 14912, Cur Loss: 0.39928919, Cur Avg Loss: 0.40221737, Log Avg loss: 0.49980812, Global Avg Loss: 1.73706511, Time: 0.0210 Steps: 26480, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000942, Sample Num: 15072, Cur Loss: 0.48970115, Cur Avg Loss: 0.40181485, Log Avg loss: 0.36429970, Global Avg Loss: 1.73654689, Time: 0.0210 Steps: 26490, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000952, Sample Num: 15232, Cur Loss: 0.42918938, Cur Avg Loss: 0.40103891, Log Avg loss: 0.32794570, Global Avg Loss: 1.73601534, Time: 0.0210 Steps: 26500, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000962, Sample Num: 15392, Cur Loss: 0.12641728, Cur Avg Loss: 0.40016755, Log Avg loss: 0.31721383, Global Avg Loss: 1.73548014, Time: 0.0211 Steps: 26510, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000972, Sample Num: 15552, Cur Loss: 0.68739897, Cur Avg Loss: 0.39940739, Log Avg loss: 0.32628033, Global Avg Loss: 1.73494877, Time: 0.0210 Steps: 26520, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000982, Sample Num: 15712, Cur Loss: 0.16518919, Cur Avg Loss: 0.39960666, Log Avg loss: 0.41897523, Global Avg Loss: 1.73445274, Time: 0.0211 Steps: 26530, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000992, Sample Num: 15872, Cur Loss: 0.38628450, Cur Avg Loss: 0.39957577, Log Avg loss: 0.39654286, Global Avg Loss: 1.73394863, Time: 0.0210 Steps: 26540, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001002, Sample Num: 16032, Cur Loss: 1.16078198, Cur Avg Loss: 0.39963033, Log Avg loss: 0.40504213, Global Avg Loss: 1.73344810, Time: 0.0210 Steps: 26550, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001012, Sample Num: 16192, Cur Loss: 1.23009300, Cur Avg Loss: 0.40281755, Log Avg loss: 0.72217694, Global Avg Loss: 1.73306735, Time: 0.0210 Steps: 26560, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001022, Sample Num: 16352, Cur Loss: 0.11304163, Cur Avg Loss: 0.40331450, Log Avg loss: 0.45360606, Global Avg Loss: 1.73258580, Time: 0.0211 Steps: 26570, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001032, Sample Num: 16512, Cur Loss: 0.66394192, Cur Avg Loss: 0.40582723, Log Avg loss: 0.66262886, Global Avg Loss: 1.73218326, Time: 0.0212 Steps: 26580, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001042, Sample Num: 16672, Cur Loss: 0.45609534, Cur Avg Loss: 0.40886464, Log Avg loss: 0.72232513, Global Avg Loss: 1.73180347, Time: 0.0211 Steps: 26590, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001052, Sample Num: 16832, Cur Loss: 0.43989107, Cur Avg Loss: 0.40961278, Log Avg loss: 0.48756865, Global Avg Loss: 1.73133572, Time: 0.0211 Steps: 26600, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001062, Sample Num: 16992, Cur Loss: 0.54531717, Cur Avg Loss: 0.40960777, Log Avg loss: 0.40908124, Global Avg Loss: 1.73083882, Time: 0.0211 Steps: 26610, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001072, Sample Num: 17152, Cur Loss: 1.21935058, Cur Avg Loss: 0.41050676, Log Avg loss: 0.50597949, Global Avg Loss: 1.73037869, Time: 0.0211 Steps: 26620, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001082, Sample Num: 17312, Cur Loss: 0.52251536, Cur Avg Loss: 0.41285429, Log Avg loss: 0.66450872, Global Avg Loss: 1.72997844, Time: 0.0210 Steps: 26630, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001092, Sample Num: 17472, Cur Loss: 0.13967392, Cur Avg Loss: 0.41328688, Log Avg loss: 0.46009294, Global Avg Loss: 1.72950175, Time: 0.0211 Steps: 26640, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001102, Sample Num: 17632, Cur Loss: 0.77629906, Cur Avg Loss: 0.41303998, Log Avg loss: 0.38607865, Global Avg Loss: 1.72899765, Time: 0.0211 Steps: 26650, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001112, Sample Num: 17792, Cur Loss: 0.12161525, Cur Avg Loss: 0.41240374, Log Avg loss: 0.34229064, Global Avg Loss: 1.72847751, Time: 0.0211 Steps: 26660, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001122, Sample Num: 17952, Cur Loss: 0.20743120, Cur Avg Loss: 0.41222456, Log Avg loss: 0.39229947, Global Avg Loss: 1.72797650, Time: 0.0210 Steps: 26670, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001132, Sample Num: 18112, Cur Loss: 0.24924791, Cur Avg Loss: 0.41092185, Log Avg loss: 0.26475799, Global Avg Loss: 1.72742807, Time: 0.0211 Steps: 26680, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001142, Sample Num: 18272, Cur Loss: 0.22631997, Cur Avg Loss: 0.41052011, Log Avg loss: 0.36504355, Global Avg Loss: 1.72691762, Time: 0.0210 Steps: 26690, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001152, Sample Num: 18432, Cur Loss: 0.15271638, Cur Avg Loss: 0.41035296, Log Avg loss: 0.39126351, Global Avg Loss: 1.72641738, Time: 0.0210 Steps: 26700, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001162, Sample Num: 18592, Cur Loss: 0.75467867, Cur Avg Loss: 0.41097931, Log Avg loss: 0.48313553, Global Avg Loss: 1.72595190, Time: 0.0210 Steps: 26710, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001172, Sample Num: 18752, Cur Loss: 0.34622961, Cur Avg Loss: 0.41090811, Log Avg loss: 0.40263478, Global Avg Loss: 1.72545665, Time: 0.0211 Steps: 26720, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001182, Sample Num: 18912, Cur Loss: 0.32486433, Cur Avg Loss: 0.41027257, Log Avg loss: 0.33578694, Global Avg Loss: 1.72493676, Time: 0.0211 Steps: 26730, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001192, Sample Num: 19072, Cur Loss: 0.08363773, Cur Avg Loss: 0.40871762, Log Avg loss: 0.22492298, Global Avg Loss: 1.72437580, Time: 0.0211 Steps: 26740, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001202, Sample Num: 19232, Cur Loss: 0.13422559, Cur Avg Loss: 0.40821367, Log Avg loss: 0.34814231, Global Avg Loss: 1.72386132, Time: 0.0210 Steps: 26750, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001212, Sample Num: 19392, Cur Loss: 0.11763245, Cur Avg Loss: 0.40776350, Log Avg loss: 0.35365259, Global Avg Loss: 1.72334928, Time: 0.0211 Steps: 26760, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001222, Sample Num: 19552, Cur Loss: 0.32012743, Cur Avg Loss: 0.40683356, Log Avg loss: 0.29412510, Global Avg Loss: 1.72281539, Time: 0.0211 Steps: 26770, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001232, Sample Num: 19712, Cur Loss: 0.31770426, Cur Avg Loss: 0.40562133, Log Avg loss: 0.25748661, Global Avg Loss: 1.72226822, Time: 0.0211 Steps: 26780, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001242, Sample Num: 19872, Cur Loss: 0.49092564, Cur Avg Loss: 0.40571283, Log Avg loss: 0.41698591, Global Avg Loss: 1.72178099, Time: 0.0211 Steps: 26790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001252, Sample Num: 20032, Cur Loss: 0.62706167, Cur Avg Loss: 0.40550147, Log Avg loss: 0.37925053, Global Avg Loss: 1.72128005, Time: 0.0210 Steps: 26800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001262, Sample Num: 20192, Cur Loss: 0.37364513, Cur Avg Loss: 0.40552645, Log Avg loss: 0.40865422, Global Avg Loss: 1.72079044, Time: 0.0210 Steps: 26810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001272, Sample Num: 20352, Cur Loss: 0.19790626, Cur Avg Loss: 0.40507372, Log Avg loss: 0.34793946, Global Avg Loss: 1.72027857, Time: 0.0211 Steps: 26820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001282, Sample Num: 20512, Cur Loss: 0.19025835, Cur Avg Loss: 0.40405776, Log Avg loss: 0.27482688, Global Avg Loss: 1.71973982, Time: 0.0248 Steps: 26830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001292, Sample Num: 20672, Cur Loss: 0.41994423, Cur Avg Loss: 0.40395177, Log Avg loss: 0.39036438, Global Avg Loss: 1.71924453, Time: 0.0211 Steps: 26840, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001302, Sample Num: 20832, Cur Loss: 0.16294831, Cur Avg Loss: 0.40332817, Log Avg loss: 0.32275929, Global Avg Loss: 1.71872442, Time: 0.0212 Steps: 26850, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001312, Sample Num: 20992, Cur Loss: 0.59832835, Cur Avg Loss: 0.40335568, Log Avg loss: 0.40693753, Global Avg Loss: 1.71823604, Time: 0.0211 Steps: 26860, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001322, Sample Num: 21152, Cur Loss: 0.38833553, Cur Avg Loss: 0.40366586, Log Avg loss: 0.44436134, Global Avg Loss: 1.71776195, Time: 0.0211 Steps: 26870, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001332, Sample Num: 21312, Cur Loss: 0.17348847, Cur Avg Loss: 0.40325503, Log Avg loss: 0.34894252, Global Avg Loss: 1.71725272, Time: 0.0212 Steps: 26880, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001342, Sample Num: 21472, Cur Loss: 0.35211027, Cur Avg Loss: 0.40295438, Log Avg loss: 0.36290807, Global Avg Loss: 1.71674906, Time: 0.0212 Steps: 26890, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001352, Sample Num: 21632, Cur Loss: 1.05189323, Cur Avg Loss: 0.40355475, Log Avg loss: 0.48412525, Global Avg Loss: 1.71629083, Time: 0.0212 Steps: 26900, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001362, Sample Num: 21792, Cur Loss: 0.24043253, Cur Avg Loss: 0.40282479, Log Avg loss: 0.30413293, Global Avg Loss: 1.71576606, Time: 0.0213 Steps: 26910, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001372, Sample Num: 21952, Cur Loss: 0.26703840, Cur Avg Loss: 0.40304091, Log Avg loss: 0.43247683, Global Avg Loss: 1.71528936, Time: 0.0212 Steps: 26920, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001382, Sample Num: 22112, Cur Loss: 0.20000565, Cur Avg Loss: 0.40255157, Log Avg loss: 0.33541495, Global Avg Loss: 1.71477697, Time: 0.0211 Steps: 26930, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001392, Sample Num: 22272, Cur Loss: 0.85482311, Cur Avg Loss: 0.40307155, Log Avg loss: 0.47493157, Global Avg Loss: 1.71431674, Time: 0.0211 Steps: 26940, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001402, Sample Num: 22432, Cur Loss: 0.27956396, Cur Avg Loss: 0.40303211, Log Avg loss: 0.39754225, Global Avg Loss: 1.71382814, Time: 0.0211 Steps: 26950, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001412, Sample Num: 22592, Cur Loss: 0.13751452, Cur Avg Loss: 0.40223820, Log Avg loss: 0.29093282, Global Avg Loss: 1.71330036, Time: 0.0212 Steps: 26960, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001422, Sample Num: 22752, Cur Loss: 0.23919046, Cur Avg Loss: 0.40159951, Log Avg loss: 0.31141589, Global Avg Loss: 1.71278057, Time: 0.0211 Steps: 26970, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001432, Sample Num: 22912, Cur Loss: 0.40798151, Cur Avg Loss: 0.40173202, Log Avg loss: 0.42057535, Global Avg Loss: 1.71230162, Time: 0.0212 Steps: 26980, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001442, Sample Num: 23072, Cur Loss: 0.28527290, Cur Avg Loss: 0.40071586, Log Avg loss: 0.25520178, Global Avg Loss: 1.71176175, Time: 0.0211 Steps: 26990, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001452, Sample Num: 23232, Cur Loss: 0.94494772, Cur Avg Loss: 0.40076623, Log Avg loss: 0.40802880, Global Avg Loss: 1.71127889, Time: 0.0212 Steps: 27000, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001462, Sample Num: 23392, Cur Loss: 0.11316384, Cur Avg Loss: 0.40088957, Log Avg loss: 0.41879923, Global Avg Loss: 1.71080037, Time: 0.0211 Steps: 27010, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001472, Sample Num: 23552, Cur Loss: 0.41807830, Cur Avg Loss: 0.40062696, Log Avg loss: 0.36223347, Global Avg Loss: 1.71030127, Time: 0.0211 Steps: 27020, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001482, Sample Num: 23712, Cur Loss: 0.23021987, Cur Avg Loss: 0.40025789, Log Avg loss: 0.34593059, Global Avg Loss: 1.70979651, Time: 0.0211 Steps: 27030, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001492, Sample Num: 23872, Cur Loss: 0.36788952, Cur Avg Loss: 0.39959936, Log Avg loss: 0.30200550, Global Avg Loss: 1.70927588, Time: 0.0212 Steps: 27040, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001502, Sample Num: 24032, Cur Loss: 0.28579706, Cur Avg Loss: 0.40042036, Log Avg loss: 0.52291323, Global Avg Loss: 1.70883729, Time: 0.0212 Steps: 27050, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001512, Sample Num: 24192, Cur Loss: 0.33933315, Cur Avg Loss: 0.40055451, Log Avg loss: 0.42070423, Global Avg Loss: 1.70836127, Time: 0.0211 Steps: 27060, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001522, Sample Num: 24352, Cur Loss: 0.32281071, Cur Avg Loss: 0.40077807, Log Avg loss: 0.43457938, Global Avg Loss: 1.70789071, Time: 0.0213 Steps: 27070, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001532, Sample Num: 24512, Cur Loss: 0.51402736, Cur Avg Loss: 0.40149570, Log Avg loss: 0.51071923, Global Avg Loss: 1.70744863, Time: 0.0211 Steps: 27080, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001542, Sample Num: 24672, Cur Loss: 0.75916684, Cur Avg Loss: 0.40174836, Log Avg loss: 0.44045560, Global Avg Loss: 1.70698093, Time: 0.0212 Steps: 27090, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001552, Sample Num: 24832, Cur Loss: 0.24108733, Cur Avg Loss: 0.40290062, Log Avg loss: 0.58057956, Global Avg Loss: 1.70656528, Time: 0.0210 Steps: 27100, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001562, Sample Num: 24992, Cur Loss: 0.12499645, Cur Avg Loss: 0.40331766, Log Avg loss: 0.46804180, Global Avg Loss: 1.70610843, Time: 0.0210 Steps: 27110, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001572, Sample Num: 25152, Cur Loss: 0.47571781, Cur Avg Loss: 0.40311158, Log Avg loss: 0.37092319, Global Avg Loss: 1.70561611, Time: 0.0210 Steps: 27120, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001582, Sample Num: 25312, Cur Loss: 0.37446657, Cur Avg Loss: 0.40365763, Log Avg loss: 0.48949656, Global Avg Loss: 1.70516785, Time: 0.0211 Steps: 27130, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001592, Sample Num: 25472, Cur Loss: 0.24280736, Cur Avg Loss: 0.40425582, Log Avg loss: 0.49888899, Global Avg Loss: 1.70472339, Time: 0.0211 Steps: 27140, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001602, Sample Num: 25632, Cur Loss: 0.22735384, Cur Avg Loss: 0.40453990, Log Avg loss: 0.44976526, Global Avg Loss: 1.70426115, Time: 0.0210 Steps: 27150, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001612, Sample Num: 25792, Cur Loss: 0.37192142, Cur Avg Loss: 0.40378286, Log Avg loss: 0.28250441, Global Avg Loss: 1.70373768, Time: 0.0210 Steps: 27160, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001622, Sample Num: 25952, Cur Loss: 0.21810845, Cur Avg Loss: 0.40427389, Log Avg loss: 0.48342872, Global Avg Loss: 1.70328854, Time: 0.0210 Steps: 27170, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001632, Sample Num: 26112, Cur Loss: 0.30347580, Cur Avg Loss: 0.40343472, Log Avg loss: 0.26732091, Global Avg Loss: 1.70276022, Time: 0.0210 Steps: 27180, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001642, Sample Num: 26272, Cur Loss: 0.51937407, Cur Avg Loss: 0.40332639, Log Avg loss: 0.38564732, Global Avg Loss: 1.70227581, Time: 0.0210 Steps: 27190, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001652, Sample Num: 26432, Cur Loss: 0.54321414, Cur Avg Loss: 0.40287195, Log Avg loss: 0.32825295, Global Avg Loss: 1.70177066, Time: 0.0210 Steps: 27200, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001662, Sample Num: 26592, Cur Loss: 0.23951054, Cur Avg Loss: 0.40305498, Log Avg loss: 0.43329161, Global Avg Loss: 1.70130448, Time: 0.0210 Steps: 27210, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001672, Sample Num: 26752, Cur Loss: 0.26619607, Cur Avg Loss: 0.40273017, Log Avg loss: 0.34874666, Global Avg Loss: 1.70080758, Time: 0.0210 Steps: 27220, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001682, Sample Num: 26912, Cur Loss: 0.12870374, Cur Avg Loss: 0.40244357, Log Avg loss: 0.35452378, Global Avg Loss: 1.70031317, Time: 0.0210 Steps: 27230, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001692, Sample Num: 27072, Cur Loss: 0.33113915, Cur Avg Loss: 0.40321022, Log Avg loss: 0.53216106, Global Avg Loss: 1.69988433, Time: 0.0210 Steps: 27240, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001702, Sample Num: 27232, Cur Loss: 0.84793723, Cur Avg Loss: 0.40331446, Log Avg loss: 0.42095177, Global Avg Loss: 1.69941500, Time: 0.0211 Steps: 27250, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001712, Sample Num: 27392, Cur Loss: 0.28803301, Cur Avg Loss: 0.40286420, Log Avg loss: 0.32622956, Global Avg Loss: 1.69891126, Time: 0.0210 Steps: 27260, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001722, Sample Num: 27552, Cur Loss: 0.12617941, Cur Avg Loss: 0.40253322, Log Avg loss: 0.34587020, Global Avg Loss: 1.69841509, Time: 0.0210 Steps: 27270, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001732, Sample Num: 27712, Cur Loss: 0.54993993, Cur Avg Loss: 0.40242474, Log Avg loss: 0.38374483, Global Avg Loss: 1.69793318, Time: 0.0210 Steps: 27280, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001742, Sample Num: 27872, Cur Loss: 0.53406507, Cur Avg Loss: 0.40298706, Log Avg loss: 0.50038098, Global Avg Loss: 1.69749435, Time: 0.0210 Steps: 27290, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001752, Sample Num: 28032, Cur Loss: 0.29163605, Cur Avg Loss: 0.40227515, Log Avg loss: 0.27825902, Global Avg Loss: 1.69697449, Time: 0.0210 Steps: 27300, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001762, Sample Num: 28192, Cur Loss: 0.26365173, Cur Avg Loss: 0.40219484, Log Avg loss: 0.38812438, Global Avg Loss: 1.69649523, Time: 0.0211 Steps: 27310, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001772, Sample Num: 28352, Cur Loss: 0.44195819, Cur Avg Loss: 0.40177497, Log Avg loss: 0.32779507, Global Avg Loss: 1.69599424, Time: 0.0210 Steps: 27320, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001782, Sample Num: 28512, Cur Loss: 0.09922757, Cur Avg Loss: 0.40146988, Log Avg loss: 0.34740749, Global Avg Loss: 1.69550080, Time: 0.0210 Steps: 27330, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001792, Sample Num: 28672, Cur Loss: 0.24166560, Cur Avg Loss: 0.40097380, Log Avg loss: 0.31257230, Global Avg Loss: 1.69499497, Time: 0.0254 Steps: 27340, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001802, Sample Num: 28832, Cur Loss: 0.39027470, Cur Avg Loss: 0.40125803, Log Avg loss: 0.45219145, Global Avg Loss: 1.69454056, Time: 0.0210 Steps: 27350, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001812, Sample Num: 28992, Cur Loss: 0.04606592, Cur Avg Loss: 0.40144394, Log Avg loss: 0.43494649, Global Avg Loss: 1.69408018, Time: 0.0209 Steps: 27360, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001822, Sample Num: 29152, Cur Loss: 0.95470405, Cur Avg Loss: 0.40157295, Log Avg loss: 0.42494833, Global Avg Loss: 1.69361649, Time: 0.0208 Steps: 27370, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001832, Sample Num: 29312, Cur Loss: 0.07978845, Cur Avg Loss: 0.40403905, Log Avg loss: 0.85336355, Global Avg Loss: 1.69330960, Time: 0.0208 Steps: 27380, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001842, Sample Num: 29472, Cur Loss: 0.61127448, Cur Avg Loss: 0.40408126, Log Avg loss: 0.41181308, Global Avg Loss: 1.69284173, Time: 0.0209 Steps: 27390, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001852, Sample Num: 29632, Cur Loss: 0.28692302, Cur Avg Loss: 0.40405642, Log Avg loss: 0.39948046, Global Avg Loss: 1.69236970, Time: 0.0208 Steps: 27400, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001862, Sample Num: 29792, Cur Loss: 0.67194223, Cur Avg Loss: 0.40493463, Log Avg loss: 0.56757952, Global Avg Loss: 1.69195935, Time: 0.0208 Steps: 27410, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001872, Sample Num: 29952, Cur Loss: 0.25567079, Cur Avg Loss: 0.40541205, Log Avg loss: 0.49430721, Global Avg Loss: 1.69152257, Time: 0.0209 Steps: 27420, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001882, Sample Num: 30112, Cur Loss: 0.25510222, Cur Avg Loss: 0.40555055, Log Avg loss: 0.43147783, Global Avg Loss: 1.69106320, Time: 0.0208 Steps: 27430, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001892, Sample Num: 30272, Cur Loss: 0.32947239, Cur Avg Loss: 0.40547161, Log Avg loss: 0.39061558, Global Avg Loss: 1.69058927, Time: 0.0208 Steps: 27440, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001902, Sample Num: 30432, Cur Loss: 0.53755075, Cur Avg Loss: 0.40530227, Log Avg loss: 0.37326391, Global Avg Loss: 1.69010937, Time: 0.0208 Steps: 27450, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001912, Sample Num: 30592, Cur Loss: 0.53002948, Cur Avg Loss: 0.40594044, Log Avg loss: 0.52731900, Global Avg Loss: 1.68968593, Time: 0.0208 Steps: 27460, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001922, Sample Num: 30752, Cur Loss: 0.38729450, Cur Avg Loss: 0.40627503, Log Avg loss: 0.47024908, Global Avg Loss: 1.68924201, Time: 0.0208 Steps: 27470, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001932, Sample Num: 30912, Cur Loss: 0.16411820, Cur Avg Loss: 0.40586337, Log Avg loss: 0.32674192, Global Avg Loss: 1.68874619, Time: 0.0208 Steps: 27480, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001942, Sample Num: 31072, Cur Loss: 0.18033272, Cur Avg Loss: 0.40565556, Log Avg loss: 0.36550756, Global Avg Loss: 1.68826484, Time: 0.0208 Steps: 27490, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001952, Sample Num: 31232, Cur Loss: 0.09490157, Cur Avg Loss: 0.40515916, Log Avg loss: 0.30875838, Global Avg Loss: 1.68776320, Time: 0.0208 Steps: 27500, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001962, Sample Num: 31392, Cur Loss: 0.41518471, Cur Avg Loss: 0.40489433, Log Avg loss: 0.35319847, Global Avg Loss: 1.68727808, Time: 0.0209 Steps: 27510, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001972, Sample Num: 31552, Cur Loss: 0.05948849, Cur Avg Loss: 0.40416675, Log Avg loss: 0.26141672, Global Avg Loss: 1.68675997, Time: 0.0209 Steps: 27520, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001982, Sample Num: 31712, Cur Loss: 0.57800877, Cur Avg Loss: 0.40441042, Log Avg loss: 0.45246276, Global Avg Loss: 1.68631162, Time: 0.0209 Steps: 27530, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001992, Sample Num: 31872, Cur Loss: 0.22780780, Cur Avg Loss: 0.40392711, Log Avg loss: 0.30813360, Global Avg Loss: 1.68581119, Time: 0.0209 Steps: 27540, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002002, Sample Num: 32032, Cur Loss: 0.58562267, Cur Avg Loss: 0.40398877, Log Avg loss: 0.41627264, Global Avg Loss: 1.68535038, Time: 0.0208 Steps: 27550, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002012, Sample Num: 32192, Cur Loss: 0.60556602, Cur Avg Loss: 0.40435653, Log Avg loss: 0.47798235, Global Avg Loss: 1.68491229, Time: 0.0208 Steps: 27560, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002022, Sample Num: 32352, Cur Loss: 0.26785448, Cur Avg Loss: 0.40416891, Log Avg loss: 0.36641967, Global Avg Loss: 1.68443406, Time: 0.0209 Steps: 27570, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002032, Sample Num: 32512, Cur Loss: 0.61723375, Cur Avg Loss: 0.40463239, Log Avg loss: 0.49834742, Global Avg Loss: 1.68400400, Time: 0.0209 Steps: 27580, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002042, Sample Num: 32672, Cur Loss: 0.38887024, Cur Avg Loss: 0.40487123, Log Avg loss: 0.45340367, Global Avg Loss: 1.68355797, Time: 0.0208 Steps: 27590, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002052, Sample Num: 32832, Cur Loss: 0.34328496, Cur Avg Loss: 0.40460617, Log Avg loss: 0.35048076, Global Avg Loss: 1.68307497, Time: 0.0247 Steps: 27600, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002062, Sample Num: 32992, Cur Loss: 0.17715363, Cur Avg Loss: 0.40389657, Log Avg loss: 0.25828743, Global Avg Loss: 1.68255893, Time: 0.0210 Steps: 27610, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002072, Sample Num: 33152, Cur Loss: 0.64234376, Cur Avg Loss: 0.40371963, Log Avg loss: 0.36723383, Global Avg Loss: 1.68208271, Time: 0.0210 Steps: 27620, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002082, Sample Num: 33312, Cur Loss: 0.10268837, Cur Avg Loss: 0.40414936, Log Avg loss: 0.49318991, Global Avg Loss: 1.68165242, Time: 0.0210 Steps: 27630, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002092, Sample Num: 33472, Cur Loss: 0.31400657, Cur Avg Loss: 0.40343104, Log Avg loss: 0.25387642, Global Avg Loss: 1.68113586, Time: 0.0210 Steps: 27640, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002102, Sample Num: 33632, Cur Loss: 0.17874093, Cur Avg Loss: 0.40279702, Log Avg loss: 0.27015889, Global Avg Loss: 1.68062556, Time: 0.0210 Steps: 27650, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002112, Sample Num: 33792, Cur Loss: 0.35676050, Cur Avg Loss: 0.40229695, Log Avg loss: 0.29718320, Global Avg Loss: 1.68012540, Time: 0.0210 Steps: 27660, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002122, Sample Num: 33952, Cur Loss: 0.69779557, Cur Avg Loss: 0.40275238, Log Avg loss: 0.49893972, Global Avg Loss: 1.67969852, Time: 0.0210 Steps: 27670, Updated lr: 0.000075 ***** Running evaluation checkpoint-27677 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-27677 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.007442, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.351345, "eval_total_loss": 246.995327, "eval_mae": 0.446179, "eval_mse": 0.351442, "eval_r2": 0.776601, "eval_sp_statistic": 0.850731, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.88846, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.458945, "test_total_loss": 230.39044, "test_mae": 0.416016, "test_mse": 0.459107, "test_r2": 0.703688, "test_sp_statistic": 0.839593, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.8874, "test_ps_pvalue": 0.0, "lr": 7.470175438596491e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6793624027250085, "train_cur_epoch_loss": 857.09580578655, "train_cur_epoch_avg_loss": 0.40258140243614376, "train_cur_epoch_time": 45.007441997528076, "train_cur_epoch_avg_time": 0.021140179425799942, "epoch": 13, "step": 27677} ################################################## Training, Epoch: 0014, Batch: 000003, Sample Num: 48, Cur Loss: 0.10788265, Cur Avg Loss: 0.19310205, Log Avg loss: 0.30345554, Global Avg Loss: 1.67920132, Time: 0.0245 Steps: 27680, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000013, Sample Num: 208, Cur Loss: 0.42387089, Cur Avg Loss: 0.35263302, Log Avg loss: 0.40049231, Global Avg Loss: 1.67873953, Time: 0.0210 Steps: 27690, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000023, Sample Num: 368, Cur Loss: 0.31419569, Cur Avg Loss: 0.32543065, Log Avg loss: 0.29006757, Global Avg Loss: 1.67823820, Time: 0.0210 Steps: 27700, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000033, Sample Num: 528, Cur Loss: 0.27308404, Cur Avg Loss: 0.34093227, Log Avg loss: 0.37658599, Global Avg Loss: 1.67776846, Time: 0.0210 Steps: 27710, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000043, Sample Num: 688, Cur Loss: 0.19348991, Cur Avg Loss: 0.35200106, Log Avg loss: 0.38852810, Global Avg Loss: 1.67730336, Time: 0.0210 Steps: 27720, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000053, Sample Num: 848, Cur Loss: 0.28233588, Cur Avg Loss: 0.36672509, Log Avg loss: 0.43003841, Global Avg Loss: 1.67685358, Time: 0.0209 Steps: 27730, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000063, Sample Num: 1008, Cur Loss: 0.27313501, Cur Avg Loss: 0.36613216, Log Avg loss: 0.36298961, Global Avg Loss: 1.67637994, Time: 0.0209 Steps: 27740, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000073, Sample Num: 1168, Cur Loss: 0.17557409, Cur Avg Loss: 0.36393610, Log Avg loss: 0.35010095, Global Avg Loss: 1.67590200, Time: 0.0209 Steps: 27750, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000083, Sample Num: 1328, Cur Loss: 0.43020779, Cur Avg Loss: 0.38407605, Log Avg loss: 0.53109765, Global Avg Loss: 1.67548961, Time: 0.0209 Steps: 27760, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000093, Sample Num: 1488, Cur Loss: 0.59749448, Cur Avg Loss: 0.38082512, Log Avg loss: 0.35384242, Global Avg Loss: 1.67501368, Time: 0.0209 Steps: 27770, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000103, Sample Num: 1648, Cur Loss: 0.13907346, Cur Avg Loss: 0.38467870, Log Avg loss: 0.42051697, Global Avg Loss: 1.67456210, Time: 0.0209 Steps: 27780, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000113, Sample Num: 1808, Cur Loss: 0.27565691, Cur Avg Loss: 0.37697911, Log Avg loss: 0.29767335, Global Avg Loss: 1.67406664, Time: 0.0210 Steps: 27790, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000123, Sample Num: 1968, Cur Loss: 0.13630328, Cur Avg Loss: 0.36987596, Log Avg loss: 0.28961040, Global Avg Loss: 1.67356863, Time: 0.0209 Steps: 27800, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000133, Sample Num: 2128, Cur Loss: 1.09125972, Cur Avg Loss: 0.38056980, Log Avg loss: 0.51210394, Global Avg Loss: 1.67315099, Time: 0.0210 Steps: 27810, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000143, Sample Num: 2288, Cur Loss: 0.35290265, Cur Avg Loss: 0.37944304, Log Avg loss: 0.36445712, Global Avg Loss: 1.67268057, Time: 0.0209 Steps: 27820, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000153, Sample Num: 2448, Cur Loss: 0.34949917, Cur Avg Loss: 0.37286699, Log Avg loss: 0.27882960, Global Avg Loss: 1.67217973, Time: 0.0209 Steps: 27830, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000163, Sample Num: 2608, Cur Loss: 0.17438816, Cur Avg Loss: 0.37205746, Log Avg loss: 0.35967167, Global Avg Loss: 1.67170828, Time: 0.0209 Steps: 27840, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000173, Sample Num: 2768, Cur Loss: 0.79855144, Cur Avg Loss: 0.37193792, Log Avg loss: 0.36998937, Global Avg Loss: 1.67124088, Time: 0.0209 Steps: 27850, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000183, Sample Num: 2928, Cur Loss: 0.29023039, Cur Avg Loss: 0.37502687, Log Avg loss: 0.42846564, Global Avg Loss: 1.67079480, Time: 0.0209 Steps: 27860, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000193, Sample Num: 3088, Cur Loss: 0.11720678, Cur Avg Loss: 0.37606375, Log Avg loss: 0.39503860, Global Avg Loss: 1.67033705, Time: 0.0209 Steps: 27870, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000203, Sample Num: 3248, Cur Loss: 0.17022690, Cur Avg Loss: 0.37003907, Log Avg loss: 0.25376276, Global Avg Loss: 1.66982895, Time: 0.0210 Steps: 27880, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000213, Sample Num: 3408, Cur Loss: 0.17025352, Cur Avg Loss: 0.36690002, Log Avg loss: 0.30317739, Global Avg Loss: 1.66933894, Time: 0.0209 Steps: 27890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000223, Sample Num: 3568, Cur Loss: 0.80278182, Cur Avg Loss: 0.37695757, Log Avg loss: 0.59118345, Global Avg Loss: 1.66895250, Time: 0.0210 Steps: 27900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000233, Sample Num: 3728, Cur Loss: 0.85793591, Cur Avg Loss: 0.38426074, Log Avg loss: 0.54712136, Global Avg Loss: 1.66855055, Time: 0.0209 Steps: 27910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000243, Sample Num: 3888, Cur Loss: 0.31884995, Cur Avg Loss: 0.38298164, Log Avg loss: 0.35317863, Global Avg Loss: 1.66807943, Time: 0.0209 Steps: 27920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000253, Sample Num: 4048, Cur Loss: 0.27361649, Cur Avg Loss: 0.37801126, Log Avg loss: 0.25723086, Global Avg Loss: 1.66757430, Time: 0.0210 Steps: 27930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000263, Sample Num: 4208, Cur Loss: 0.30864778, Cur Avg Loss: 0.37314903, Log Avg loss: 0.25013483, Global Avg Loss: 1.66706698, Time: 0.0210 Steps: 27940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000273, Sample Num: 4368, Cur Loss: 0.28831446, Cur Avg Loss: 0.36985295, Log Avg loss: 0.28316599, Global Avg Loss: 1.66657185, Time: 0.0208 Steps: 27950, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000283, Sample Num: 4528, Cur Loss: 0.14350265, Cur Avg Loss: 0.36572917, Log Avg loss: 0.25314999, Global Avg Loss: 1.66606633, Time: 0.0209 Steps: 27960, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000293, Sample Num: 4688, Cur Loss: 1.07724380, Cur Avg Loss: 0.36794571, Log Avg loss: 0.43067384, Global Avg Loss: 1.66562464, Time: 0.0210 Steps: 27970, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000303, Sample Num: 4848, Cur Loss: 0.19053988, Cur Avg Loss: 0.36662360, Log Avg loss: 0.32788564, Global Avg Loss: 1.66514654, Time: 0.0209 Steps: 27980, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000313, Sample Num: 5008, Cur Loss: 0.45493758, Cur Avg Loss: 0.36817826, Log Avg loss: 0.41528463, Global Avg Loss: 1.66470000, Time: 0.0209 Steps: 27990, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000323, Sample Num: 5168, Cur Loss: 0.30164847, Cur Avg Loss: 0.36545683, Log Avg loss: 0.28027610, Global Avg Loss: 1.66420556, Time: 0.0209 Steps: 28000, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000333, Sample Num: 5328, Cur Loss: 0.13452779, Cur Avg Loss: 0.36516877, Log Avg loss: 0.35586426, Global Avg Loss: 1.66373847, Time: 0.0209 Steps: 28010, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000343, Sample Num: 5488, Cur Loss: 1.06573582, Cur Avg Loss: 0.37001373, Log Avg loss: 0.53135100, Global Avg Loss: 1.66333433, Time: 0.0208 Steps: 28020, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000353, Sample Num: 5648, Cur Loss: 0.37758875, Cur Avg Loss: 0.37166972, Log Avg loss: 0.42846992, Global Avg Loss: 1.66289378, Time: 0.0209 Steps: 28030, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000363, Sample Num: 5808, Cur Loss: 0.41995731, Cur Avg Loss: 0.37322934, Log Avg loss: 0.42828420, Global Avg Loss: 1.66245348, Time: 0.0209 Steps: 28040, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000373, Sample Num: 5968, Cur Loss: 0.46218541, Cur Avg Loss: 0.37363088, Log Avg loss: 0.38820652, Global Avg Loss: 1.66199920, Time: 0.0209 Steps: 28050, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000383, Sample Num: 6128, Cur Loss: 0.31105843, Cur Avg Loss: 0.37451745, Log Avg loss: 0.40758656, Global Avg Loss: 1.66155215, Time: 0.0208 Steps: 28060, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000393, Sample Num: 6288, Cur Loss: 0.66944063, Cur Avg Loss: 0.37341474, Log Avg loss: 0.33118112, Global Avg Loss: 1.66107820, Time: 0.0209 Steps: 28070, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000403, Sample Num: 6448, Cur Loss: 0.92887425, Cur Avg Loss: 0.38034567, Log Avg loss: 0.65273120, Global Avg Loss: 1.66071911, Time: 0.0209 Steps: 28080, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000413, Sample Num: 6608, Cur Loss: 0.34572560, Cur Avg Loss: 0.38019022, Log Avg loss: 0.37392563, Global Avg Loss: 1.66026101, Time: 0.0208 Steps: 28090, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000423, Sample Num: 6768, Cur Loss: 0.57559812, Cur Avg Loss: 0.38153365, Log Avg loss: 0.43701703, Global Avg Loss: 1.65982569, Time: 0.0209 Steps: 28100, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000433, Sample Num: 6928, Cur Loss: 1.20143056, Cur Avg Loss: 0.38954316, Log Avg loss: 0.72834574, Global Avg Loss: 1.65949432, Time: 0.0209 Steps: 28110, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000443, Sample Num: 7088, Cur Loss: 0.39114752, Cur Avg Loss: 0.38975283, Log Avg loss: 0.39883143, Global Avg Loss: 1.65904601, Time: 0.0209 Steps: 28120, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000453, Sample Num: 7248, Cur Loss: 0.21068551, Cur Avg Loss: 0.38910085, Log Avg loss: 0.36021792, Global Avg Loss: 1.65858428, Time: 0.0208 Steps: 28130, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000463, Sample Num: 7408, Cur Loss: 0.40343407, Cur Avg Loss: 0.38981754, Log Avg loss: 0.42228389, Global Avg Loss: 1.65814494, Time: 0.0208 Steps: 28140, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000473, Sample Num: 7568, Cur Loss: 0.32205296, Cur Avg Loss: 0.39067811, Log Avg loss: 0.43052242, Global Avg Loss: 1.65770884, Time: 0.0209 Steps: 28150, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000483, Sample Num: 7728, Cur Loss: 0.50590867, Cur Avg Loss: 0.39215585, Log Avg loss: 0.46205279, Global Avg Loss: 1.65728425, Time: 0.0208 Steps: 28160, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000493, Sample Num: 7888, Cur Loss: 0.13572714, Cur Avg Loss: 0.39248820, Log Avg loss: 0.40854076, Global Avg Loss: 1.65684096, Time: 0.0209 Steps: 28170, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000503, Sample Num: 8048, Cur Loss: 0.50360179, Cur Avg Loss: 0.39300037, Log Avg loss: 0.41825021, Global Avg Loss: 1.65640143, Time: 0.0208 Steps: 28180, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000513, Sample Num: 8208, Cur Loss: 0.18691099, Cur Avg Loss: 0.39031439, Log Avg loss: 0.25520963, Global Avg Loss: 1.65590438, Time: 0.0247 Steps: 28190, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000523, Sample Num: 8368, Cur Loss: 0.16255534, Cur Avg Loss: 0.39007353, Log Avg loss: 0.37771766, Global Avg Loss: 1.65545112, Time: 0.0209 Steps: 28200, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000533, Sample Num: 8528, Cur Loss: 0.21327423, Cur Avg Loss: 0.39151778, Log Avg loss: 0.46705202, Global Avg Loss: 1.65502985, Time: 0.0209 Steps: 28210, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000543, Sample Num: 8688, Cur Loss: 0.51504529, Cur Avg Loss: 0.38982441, Log Avg loss: 0.29956788, Global Avg Loss: 1.65454953, Time: 0.0209 Steps: 28220, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000553, Sample Num: 8848, Cur Loss: 0.39675874, Cur Avg Loss: 0.38875304, Log Avg loss: 0.33057779, Global Avg Loss: 1.65408054, Time: 0.0208 Steps: 28230, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000563, Sample Num: 9008, Cur Loss: 0.39182559, Cur Avg Loss: 0.38876787, Log Avg loss: 0.38958760, Global Avg Loss: 1.65363277, Time: 0.0208 Steps: 28240, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000573, Sample Num: 9168, Cur Loss: 0.21101123, Cur Avg Loss: 0.38897532, Log Avg loss: 0.40065498, Global Avg Loss: 1.65318924, Time: 0.0209 Steps: 28250, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000583, Sample Num: 9328, Cur Loss: 0.44220382, Cur Avg Loss: 0.38742968, Log Avg loss: 0.29886462, Global Avg Loss: 1.65271000, Time: 0.0209 Steps: 28260, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000593, Sample Num: 9488, Cur Loss: 0.21518159, Cur Avg Loss: 0.38638432, Log Avg loss: 0.32543968, Global Avg Loss: 1.65224051, Time: 0.0209 Steps: 28270, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000603, Sample Num: 9648, Cur Loss: 0.09683797, Cur Avg Loss: 0.38502585, Log Avg loss: 0.30446863, Global Avg Loss: 1.65176393, Time: 0.0209 Steps: 28280, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000613, Sample Num: 9808, Cur Loss: 0.44068804, Cur Avg Loss: 0.38432166, Log Avg loss: 0.34185882, Global Avg Loss: 1.65130090, Time: 0.0209 Steps: 28290, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000623, Sample Num: 9968, Cur Loss: 0.11874427, Cur Avg Loss: 0.38343072, Log Avg loss: 0.32881602, Global Avg Loss: 1.65083359, Time: 0.0209 Steps: 28300, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000633, Sample Num: 10128, Cur Loss: 0.99879038, Cur Avg Loss: 0.38460582, Log Avg loss: 0.45781501, Global Avg Loss: 1.65041218, Time: 0.0209 Steps: 28310, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000643, Sample Num: 10288, Cur Loss: 0.45749569, Cur Avg Loss: 0.38277911, Log Avg loss: 0.26714798, Global Avg Loss: 1.64992374, Time: 0.0209 Steps: 28320, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000653, Sample Num: 10448, Cur Loss: 0.29596394, Cur Avg Loss: 0.38502021, Log Avg loss: 0.52912273, Global Avg Loss: 1.64952811, Time: 0.0209 Steps: 28330, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000663, Sample Num: 10608, Cur Loss: 0.30595356, Cur Avg Loss: 0.38443631, Log Avg loss: 0.34630809, Global Avg Loss: 1.64906826, Time: 0.0209 Steps: 28340, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000673, Sample Num: 10768, Cur Loss: 0.23954836, Cur Avg Loss: 0.38427618, Log Avg loss: 0.37365916, Global Avg Loss: 1.64861838, Time: 0.0208 Steps: 28350, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000683, Sample Num: 10928, Cur Loss: 0.28997946, Cur Avg Loss: 0.38199886, Log Avg loss: 0.22873528, Global Avg Loss: 1.64811772, Time: 0.0209 Steps: 28360, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000693, Sample Num: 11088, Cur Loss: 0.43849525, Cur Avg Loss: 0.38321253, Log Avg loss: 0.46610647, Global Avg Loss: 1.64770108, Time: 0.0209 Steps: 28370, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000703, Sample Num: 11248, Cur Loss: 0.27159131, Cur Avg Loss: 0.38253786, Log Avg loss: 0.33578329, Global Avg Loss: 1.64723881, Time: 0.0209 Steps: 28380, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000713, Sample Num: 11408, Cur Loss: 0.88614112, Cur Avg Loss: 0.38232181, Log Avg loss: 0.36713356, Global Avg Loss: 1.64678791, Time: 0.0209 Steps: 28390, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000723, Sample Num: 11568, Cur Loss: 0.31519493, Cur Avg Loss: 0.38041255, Log Avg loss: 0.24428187, Global Avg Loss: 1.64629407, Time: 0.0209 Steps: 28400, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000733, Sample Num: 11728, Cur Loss: 0.28934246, Cur Avg Loss: 0.37934996, Log Avg loss: 0.30252485, Global Avg Loss: 1.64582107, Time: 0.0209 Steps: 28410, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000743, Sample Num: 11888, Cur Loss: 0.18995792, Cur Avg Loss: 0.37953517, Log Avg loss: 0.39311120, Global Avg Loss: 1.64538029, Time: 0.0209 Steps: 28420, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000753, Sample Num: 12048, Cur Loss: 0.22668129, Cur Avg Loss: 0.37865533, Log Avg loss: 0.31328309, Global Avg Loss: 1.64491174, Time: 0.0209 Steps: 28430, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000763, Sample Num: 12208, Cur Loss: 0.72168535, Cur Avg Loss: 0.37828938, Log Avg loss: 0.35073322, Global Avg Loss: 1.64445668, Time: 0.0209 Steps: 28440, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000773, Sample Num: 12368, Cur Loss: 0.18653908, Cur Avg Loss: 0.37795524, Log Avg loss: 0.35246040, Global Avg Loss: 1.64400255, Time: 0.0248 Steps: 28450, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000783, Sample Num: 12528, Cur Loss: 0.35604158, Cur Avg Loss: 0.37747160, Log Avg loss: 0.34008617, Global Avg Loss: 1.64354439, Time: 0.0211 Steps: 28460, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000793, Sample Num: 12688, Cur Loss: 0.78241217, Cur Avg Loss: 0.37810150, Log Avg loss: 0.42742302, Global Avg Loss: 1.64311724, Time: 0.0210 Steps: 28470, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000803, Sample Num: 12848, Cur Loss: 0.55985749, Cur Avg Loss: 0.37716843, Log Avg loss: 0.30317619, Global Avg Loss: 1.64264675, Time: 0.0210 Steps: 28480, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000813, Sample Num: 13008, Cur Loss: 0.33562326, Cur Avg Loss: 0.37709913, Log Avg loss: 0.37153367, Global Avg Loss: 1.64220059, Time: 0.0210 Steps: 28490, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000823, Sample Num: 13168, Cur Loss: 0.42627949, Cur Avg Loss: 0.37706256, Log Avg loss: 0.37409010, Global Avg Loss: 1.64175564, Time: 0.0210 Steps: 28500, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000833, Sample Num: 13328, Cur Loss: 0.61175799, Cur Avg Loss: 0.38061716, Log Avg loss: 0.67316051, Global Avg Loss: 1.64141590, Time: 0.0210 Steps: 28510, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000843, Sample Num: 13488, Cur Loss: 0.28092098, Cur Avg Loss: 0.38017605, Log Avg loss: 0.34343107, Global Avg Loss: 1.64096079, Time: 0.0210 Steps: 28520, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000853, Sample Num: 13648, Cur Loss: 0.42492670, Cur Avg Loss: 0.38236443, Log Avg loss: 0.56684492, Global Avg Loss: 1.64058430, Time: 0.0210 Steps: 28530, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000863, Sample Num: 13808, Cur Loss: 0.24377036, Cur Avg Loss: 0.38259037, Log Avg loss: 0.40186330, Global Avg Loss: 1.64015027, Time: 0.0210 Steps: 28540, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000873, Sample Num: 13968, Cur Loss: 0.24211234, Cur Avg Loss: 0.38297206, Log Avg loss: 0.41591198, Global Avg Loss: 1.63972147, Time: 0.0210 Steps: 28550, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000883, Sample Num: 14128, Cur Loss: 0.57317746, Cur Avg Loss: 0.38274636, Log Avg loss: 0.36304307, Global Avg Loss: 1.63927445, Time: 0.0210 Steps: 28560, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000893, Sample Num: 14288, Cur Loss: 0.40083313, Cur Avg Loss: 0.38234539, Log Avg loss: 0.34693933, Global Avg Loss: 1.63882211, Time: 0.0210 Steps: 28570, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000903, Sample Num: 14448, Cur Loss: 0.13199113, Cur Avg Loss: 0.38167394, Log Avg loss: 0.32171308, Global Avg Loss: 1.63836126, Time: 0.0210 Steps: 28580, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000913, Sample Num: 14608, Cur Loss: 0.39697537, Cur Avg Loss: 0.38270299, Log Avg loss: 0.47562652, Global Avg Loss: 1.63795457, Time: 0.0210 Steps: 28590, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000923, Sample Num: 14768, Cur Loss: 0.66323924, Cur Avg Loss: 0.38214923, Log Avg loss: 0.33159114, Global Avg Loss: 1.63749780, Time: 0.0210 Steps: 28600, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000933, Sample Num: 14928, Cur Loss: 0.29428783, Cur Avg Loss: 0.38178423, Log Avg loss: 0.34809434, Global Avg Loss: 1.63704711, Time: 0.0210 Steps: 28610, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000943, Sample Num: 15088, Cur Loss: 0.21051064, Cur Avg Loss: 0.38202771, Log Avg loss: 0.40474459, Global Avg Loss: 1.63661654, Time: 0.0210 Steps: 28620, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000953, Sample Num: 15248, Cur Loss: 0.18357939, Cur Avg Loss: 0.38288354, Log Avg loss: 0.46358855, Global Avg Loss: 1.63620682, Time: 0.0210 Steps: 28630, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000963, Sample Num: 15408, Cur Loss: 0.78535086, Cur Avg Loss: 0.38349604, Log Avg loss: 0.44186721, Global Avg Loss: 1.63578980, Time: 0.0210 Steps: 28640, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000973, Sample Num: 15568, Cur Loss: 0.33211491, Cur Avg Loss: 0.38572484, Log Avg loss: 0.60035807, Global Avg Loss: 1.63542839, Time: 0.0211 Steps: 28650, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000983, Sample Num: 15728, Cur Loss: 0.28821650, Cur Avg Loss: 0.38471853, Log Avg loss: 0.28680517, Global Avg Loss: 1.63495783, Time: 0.0210 Steps: 28660, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000993, Sample Num: 15888, Cur Loss: 0.11454122, Cur Avg Loss: 0.38400317, Log Avg loss: 0.31368287, Global Avg Loss: 1.63449698, Time: 0.0210 Steps: 28670, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001003, Sample Num: 16048, Cur Loss: 0.35410932, Cur Avg Loss: 0.38282679, Log Avg loss: 0.26601212, Global Avg Loss: 1.63401982, Time: 0.0210 Steps: 28680, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001013, Sample Num: 16208, Cur Loss: 0.66909057, Cur Avg Loss: 0.38271099, Log Avg loss: 0.37109667, Global Avg Loss: 1.63357963, Time: 0.0210 Steps: 28690, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001023, Sample Num: 16368, Cur Loss: 0.37452590, Cur Avg Loss: 0.38668485, Log Avg loss: 0.78923692, Global Avg Loss: 1.63328543, Time: 0.0210 Steps: 28700, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001033, Sample Num: 16528, Cur Loss: 0.83631283, Cur Avg Loss: 0.39080726, Log Avg loss: 0.81252903, Global Avg Loss: 1.63299955, Time: 0.0208 Steps: 28710, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001043, Sample Num: 16688, Cur Loss: 0.16435510, Cur Avg Loss: 0.39030913, Log Avg loss: 0.33885284, Global Avg Loss: 1.63254894, Time: 0.0208 Steps: 28720, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001053, Sample Num: 16848, Cur Loss: 0.21120727, Cur Avg Loss: 0.38941200, Log Avg loss: 0.29584103, Global Avg Loss: 1.63208368, Time: 0.0208 Steps: 28730, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001063, Sample Num: 17008, Cur Loss: 0.43869388, Cur Avg Loss: 0.39008166, Log Avg loss: 0.46059685, Global Avg Loss: 1.63167606, Time: 0.0209 Steps: 28740, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001073, Sample Num: 17168, Cur Loss: 0.65780848, Cur Avg Loss: 0.39017914, Log Avg loss: 0.40054128, Global Avg Loss: 1.63124784, Time: 0.0208 Steps: 28750, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001083, Sample Num: 17328, Cur Loss: 0.12878409, Cur Avg Loss: 0.38968997, Log Avg loss: 0.33720252, Global Avg Loss: 1.63079790, Time: 0.0208 Steps: 28760, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001093, Sample Num: 17488, Cur Loss: 0.20584609, Cur Avg Loss: 0.38859069, Log Avg loss: 0.26953859, Global Avg Loss: 1.63032474, Time: 0.0209 Steps: 28770, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001103, Sample Num: 17648, Cur Loss: 0.31617555, Cur Avg Loss: 0.38847168, Log Avg loss: 0.37546369, Global Avg Loss: 1.62988872, Time: 0.0209 Steps: 28780, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001113, Sample Num: 17808, Cur Loss: 0.24672455, Cur Avg Loss: 0.38804406, Log Avg loss: 0.34087774, Global Avg Loss: 1.62944100, Time: 0.0208 Steps: 28790, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001123, Sample Num: 17968, Cur Loss: 0.13858855, Cur Avg Loss: 0.38768385, Log Avg loss: 0.34759201, Global Avg Loss: 1.62899591, Time: 0.0208 Steps: 28800, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001133, Sample Num: 18128, Cur Loss: 0.72447836, Cur Avg Loss: 0.38834430, Log Avg loss: 0.46251348, Global Avg Loss: 1.62859102, Time: 0.0209 Steps: 28810, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001143, Sample Num: 18288, Cur Loss: 0.32671389, Cur Avg Loss: 0.38926923, Log Avg loss: 0.49406381, Global Avg Loss: 1.62819736, Time: 0.0208 Steps: 28820, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001153, Sample Num: 18448, Cur Loss: 0.31124055, Cur Avg Loss: 0.38841688, Log Avg loss: 0.29099324, Global Avg Loss: 1.62773354, Time: 0.0209 Steps: 28830, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001163, Sample Num: 18608, Cur Loss: 0.18750919, Cur Avg Loss: 0.38776426, Log Avg loss: 0.31251733, Global Avg Loss: 1.62727750, Time: 0.0208 Steps: 28840, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001173, Sample Num: 18768, Cur Loss: 0.08995286, Cur Avg Loss: 0.38697216, Log Avg loss: 0.29485039, Global Avg Loss: 1.62681565, Time: 0.0209 Steps: 28850, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001183, Sample Num: 18928, Cur Loss: 0.33092779, Cur Avg Loss: 0.38649916, Log Avg loss: 0.33101678, Global Avg Loss: 1.62636666, Time: 0.0208 Steps: 28860, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001193, Sample Num: 19088, Cur Loss: 0.33808994, Cur Avg Loss: 0.38662192, Log Avg loss: 0.40114403, Global Avg Loss: 1.62594226, Time: 0.0209 Steps: 28870, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001203, Sample Num: 19248, Cur Loss: 0.25304559, Cur Avg Loss: 0.38635380, Log Avg loss: 0.35436733, Global Avg Loss: 1.62550197, Time: 0.0208 Steps: 28880, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001213, Sample Num: 19408, Cur Loss: 0.45984745, Cur Avg Loss: 0.38593836, Log Avg loss: 0.33596010, Global Avg Loss: 1.62505561, Time: 0.0208 Steps: 28890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001223, Sample Num: 19568, Cur Loss: 0.33372158, Cur Avg Loss: 0.38629557, Log Avg loss: 0.42962517, Global Avg Loss: 1.62464196, Time: 0.0208 Steps: 28900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001233, Sample Num: 19728, Cur Loss: 0.26511356, Cur Avg Loss: 0.38563392, Log Avg loss: 0.30471479, Global Avg Loss: 1.62418540, Time: 0.0209 Steps: 28910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001243, Sample Num: 19888, Cur Loss: 0.25574902, Cur Avg Loss: 0.38562195, Log Avg loss: 0.38414540, Global Avg Loss: 1.62375661, Time: 0.0209 Steps: 28920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001253, Sample Num: 20048, Cur Loss: 0.17044824, Cur Avg Loss: 0.38544298, Log Avg loss: 0.36319767, Global Avg Loss: 1.62332089, Time: 0.0209 Steps: 28930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001263, Sample Num: 20208, Cur Loss: 0.91056085, Cur Avg Loss: 0.38537626, Log Avg loss: 0.37701626, Global Avg Loss: 1.62289024, Time: 0.0209 Steps: 28940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001273, Sample Num: 20368, Cur Loss: 0.96726096, Cur Avg Loss: 0.38458723, Log Avg loss: 0.28493279, Global Avg Loss: 1.62242807, Time: 0.0208 Steps: 28950, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001283, Sample Num: 20528, Cur Loss: 0.09097306, Cur Avg Loss: 0.38474802, Log Avg loss: 0.40521622, Global Avg Loss: 1.62200777, Time: 0.0246 Steps: 28960, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001293, Sample Num: 20688, Cur Loss: 0.50180584, Cur Avg Loss: 0.38512976, Log Avg loss: 0.43410652, Global Avg Loss: 1.62159772, Time: 0.0209 Steps: 28970, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001303, Sample Num: 20848, Cur Loss: 0.35834160, Cur Avg Loss: 0.38547032, Log Avg loss: 0.42950510, Global Avg Loss: 1.62118637, Time: 0.0209 Steps: 28980, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001313, Sample Num: 21008, Cur Loss: 0.37858513, Cur Avg Loss: 0.38576915, Log Avg loss: 0.42470632, Global Avg Loss: 1.62077365, Time: 0.0209 Steps: 28990, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001323, Sample Num: 21168, Cur Loss: 0.09859020, Cur Avg Loss: 0.38592356, Log Avg loss: 0.40619825, Global Avg Loss: 1.62035483, Time: 0.0209 Steps: 29000, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001333, Sample Num: 21328, Cur Loss: 0.63753474, Cur Avg Loss: 0.38580414, Log Avg loss: 0.37000438, Global Avg Loss: 1.61992382, Time: 0.0209 Steps: 29010, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001343, Sample Num: 21488, Cur Loss: 0.41873503, Cur Avg Loss: 0.38672796, Log Avg loss: 0.50987404, Global Avg Loss: 1.61954131, Time: 0.0208 Steps: 29020, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001353, Sample Num: 21648, Cur Loss: 0.44332129, Cur Avg Loss: 0.38699937, Log Avg loss: 0.42344963, Global Avg Loss: 1.61912929, Time: 0.0209 Steps: 29030, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001363, Sample Num: 21808, Cur Loss: 0.37239099, Cur Avg Loss: 0.38667022, Log Avg loss: 0.34213623, Global Avg Loss: 1.61868956, Time: 0.0209 Steps: 29040, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001373, Sample Num: 21968, Cur Loss: 1.85158575, Cur Avg Loss: 0.38774972, Log Avg loss: 0.53488507, Global Avg Loss: 1.61831647, Time: 0.0209 Steps: 29050, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001383, Sample Num: 22128, Cur Loss: 0.28490317, Cur Avg Loss: 0.38805881, Log Avg loss: 0.43049654, Global Avg Loss: 1.61790773, Time: 0.0209 Steps: 29060, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001393, Sample Num: 22288, Cur Loss: 0.32629931, Cur Avg Loss: 0.38766057, Log Avg loss: 0.33258456, Global Avg Loss: 1.61746558, Time: 0.0209 Steps: 29070, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001403, Sample Num: 22448, Cur Loss: 0.27852589, Cur Avg Loss: 0.38729228, Log Avg loss: 0.33598905, Global Avg Loss: 1.61702491, Time: 0.0209 Steps: 29080, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001413, Sample Num: 22608, Cur Loss: 0.16559958, Cur Avg Loss: 0.38713777, Log Avg loss: 0.36545996, Global Avg Loss: 1.61659467, Time: 0.0209 Steps: 29090, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001423, Sample Num: 22768, Cur Loss: 0.05531511, Cur Avg Loss: 0.38645674, Log Avg loss: 0.29022813, Global Avg Loss: 1.61613887, Time: 0.0209 Steps: 29100, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001433, Sample Num: 22928, Cur Loss: 0.05457876, Cur Avg Loss: 0.38566050, Log Avg loss: 0.27235497, Global Avg Loss: 1.61567725, Time: 0.0209 Steps: 29110, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001443, Sample Num: 23088, Cur Loss: 0.38572085, Cur Avg Loss: 0.38486643, Log Avg loss: 0.27107651, Global Avg Loss: 1.61521550, Time: 0.0209 Steps: 29120, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001453, Sample Num: 23248, Cur Loss: 0.43687779, Cur Avg Loss: 0.38554125, Log Avg loss: 0.48291686, Global Avg Loss: 1.61482680, Time: 0.0209 Steps: 29130, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001463, Sample Num: 23408, Cur Loss: 0.12189817, Cur Avg Loss: 0.38687580, Log Avg loss: 0.58078602, Global Avg Loss: 1.61447195, Time: 0.0209 Steps: 29140, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001473, Sample Num: 23568, Cur Loss: 1.24811172, Cur Avg Loss: 0.38655208, Log Avg loss: 0.33919171, Global Avg Loss: 1.61403446, Time: 0.0208 Steps: 29150, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001483, Sample Num: 23728, Cur Loss: 0.19148491, Cur Avg Loss: 0.38798433, Log Avg loss: 0.59895602, Global Avg Loss: 1.61368635, Time: 0.0209 Steps: 29160, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001493, Sample Num: 23888, Cur Loss: 0.56274825, Cur Avg Loss: 0.38979590, Log Avg loss: 0.65845093, Global Avg Loss: 1.61335888, Time: 0.0208 Steps: 29170, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001503, Sample Num: 24048, Cur Loss: 0.32887763, Cur Avg Loss: 0.38903358, Log Avg loss: 0.27521911, Global Avg Loss: 1.61290030, Time: 0.0209 Steps: 29180, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001513, Sample Num: 24208, Cur Loss: 0.34285378, Cur Avg Loss: 0.38808791, Log Avg loss: 0.24595457, Global Avg Loss: 1.61243201, Time: 0.0209 Steps: 29190, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001523, Sample Num: 24368, Cur Loss: 0.39307970, Cur Avg Loss: 0.38763056, Log Avg loss: 0.31843303, Global Avg Loss: 1.61198885, Time: 0.0209 Steps: 29200, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001533, Sample Num: 24528, Cur Loss: 0.25595471, Cur Avg Loss: 0.38754073, Log Avg loss: 0.37385933, Global Avg Loss: 1.61156498, Time: 0.0209 Steps: 29210, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001543, Sample Num: 24688, Cur Loss: 0.29701072, Cur Avg Loss: 0.38678194, Log Avg loss: 0.27045970, Global Avg Loss: 1.61110601, Time: 0.0210 Steps: 29220, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001553, Sample Num: 24848, Cur Loss: 0.28995419, Cur Avg Loss: 0.38635365, Log Avg loss: 0.32026772, Global Avg Loss: 1.61066440, Time: 0.0209 Steps: 29230, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001563, Sample Num: 25008, Cur Loss: 0.59006119, Cur Avg Loss: 0.38630631, Log Avg loss: 0.37895454, Global Avg Loss: 1.61024316, Time: 0.0209 Steps: 29240, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001573, Sample Num: 25168, Cur Loss: 0.18491876, Cur Avg Loss: 0.38637060, Log Avg loss: 0.39641917, Global Avg Loss: 1.60982818, Time: 0.0211 Steps: 29250, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001583, Sample Num: 25328, Cur Loss: 0.33880365, Cur Avg Loss: 0.38606244, Log Avg loss: 0.33758939, Global Avg Loss: 1.60939337, Time: 0.0209 Steps: 29260, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001593, Sample Num: 25488, Cur Loss: 0.14693350, Cur Avg Loss: 0.38532919, Log Avg loss: 0.26925499, Global Avg Loss: 1.60893552, Time: 0.0209 Steps: 29270, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001603, Sample Num: 25648, Cur Loss: 0.11345793, Cur Avg Loss: 0.38509761, Log Avg loss: 0.34820773, Global Avg Loss: 1.60850494, Time: 0.0209 Steps: 29280, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001613, Sample Num: 25808, Cur Loss: 0.30463520, Cur Avg Loss: 0.38449126, Log Avg loss: 0.28729319, Global Avg Loss: 1.60805386, Time: 0.0209 Steps: 29290, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001623, Sample Num: 25968, Cur Loss: 0.55871558, Cur Avg Loss: 0.38470985, Log Avg loss: 0.41996903, Global Avg Loss: 1.60764837, Time: 0.0209 Steps: 29300, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001633, Sample Num: 26128, Cur Loss: 0.09701502, Cur Avg Loss: 0.38512889, Log Avg loss: 0.45313870, Global Avg Loss: 1.60725448, Time: 0.0209 Steps: 29310, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001643, Sample Num: 26288, Cur Loss: 0.17379293, Cur Avg Loss: 0.38499513, Log Avg loss: 0.36315205, Global Avg Loss: 1.60683016, Time: 0.0209 Steps: 29320, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001653, Sample Num: 26448, Cur Loss: 0.44495192, Cur Avg Loss: 0.38439476, Log Avg loss: 0.28575449, Global Avg Loss: 1.60637974, Time: 0.0209 Steps: 29330, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001663, Sample Num: 26608, Cur Loss: 0.16739842, Cur Avg Loss: 0.38403241, Log Avg loss: 0.32413537, Global Avg Loss: 1.60594271, Time: 0.0209 Steps: 29340, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001673, Sample Num: 26768, Cur Loss: 0.89488786, Cur Avg Loss: 0.38427218, Log Avg loss: 0.42414555, Global Avg Loss: 1.60554005, Time: 0.0209 Steps: 29350, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001683, Sample Num: 26928, Cur Loss: 0.46357098, Cur Avg Loss: 0.38472714, Log Avg loss: 0.46084295, Global Avg Loss: 1.60515017, Time: 0.0209 Steps: 29360, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001693, Sample Num: 27088, Cur Loss: 0.18058583, Cur Avg Loss: 0.38538820, Log Avg loss: 0.49664418, Global Avg Loss: 1.60477274, Time: 0.0209 Steps: 29370, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001703, Sample Num: 27248, Cur Loss: 0.77406549, Cur Avg Loss: 0.38531211, Log Avg loss: 0.37243055, Global Avg Loss: 1.60435329, Time: 0.0209 Steps: 29380, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001713, Sample Num: 27408, Cur Loss: 0.25140619, Cur Avg Loss: 0.38505145, Log Avg loss: 0.34066006, Global Avg Loss: 1.60392332, Time: 0.0209 Steps: 29390, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001723, Sample Num: 27568, Cur Loss: 0.51466894, Cur Avg Loss: 0.38466090, Log Avg loss: 0.31776012, Global Avg Loss: 1.60348585, Time: 0.0209 Steps: 29400, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001733, Sample Num: 27728, Cur Loss: 0.45694393, Cur Avg Loss: 0.38500496, Log Avg loss: 0.44428672, Global Avg Loss: 1.60309170, Time: 0.0209 Steps: 29410, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001743, Sample Num: 27888, Cur Loss: 0.26958114, Cur Avg Loss: 0.38465527, Log Avg loss: 0.32405357, Global Avg Loss: 1.60265695, Time: 0.0208 Steps: 29420, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001753, Sample Num: 28048, Cur Loss: 0.78972411, Cur Avg Loss: 0.38478520, Log Avg loss: 0.40743295, Global Avg Loss: 1.60225082, Time: 0.0209 Steps: 29430, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001763, Sample Num: 28208, Cur Loss: 0.58224726, Cur Avg Loss: 0.38521703, Log Avg loss: 0.46091610, Global Avg Loss: 1.60186314, Time: 0.0209 Steps: 29440, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001773, Sample Num: 28368, Cur Loss: 0.57437104, Cur Avg Loss: 0.38552266, Log Avg loss: 0.43940559, Global Avg Loss: 1.60146842, Time: 0.0209 Steps: 29450, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001783, Sample Num: 28528, Cur Loss: 0.26582509, Cur Avg Loss: 0.38504561, Log Avg loss: 0.30046397, Global Avg Loss: 1.60102680, Time: 0.0209 Steps: 29460, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001793, Sample Num: 28688, Cur Loss: 0.27480781, Cur Avg Loss: 0.38500863, Log Avg loss: 0.37841487, Global Avg Loss: 1.60061193, Time: 0.0247 Steps: 29470, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001803, Sample Num: 28848, Cur Loss: 0.70137793, Cur Avg Loss: 0.38518895, Log Avg loss: 0.41752129, Global Avg Loss: 1.60021061, Time: 0.0209 Steps: 29480, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001813, Sample Num: 29008, Cur Loss: 0.11185641, Cur Avg Loss: 0.38427158, Log Avg loss: 0.21886836, Global Avg Loss: 1.59974220, Time: 0.0209 Steps: 29490, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001823, Sample Num: 29168, Cur Loss: 0.24709654, Cur Avg Loss: 0.38404807, Log Avg loss: 0.34352738, Global Avg Loss: 1.59931637, Time: 0.0209 Steps: 29500, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001833, Sample Num: 29328, Cur Loss: 0.08317214, Cur Avg Loss: 0.38349114, Log Avg loss: 0.28196264, Global Avg Loss: 1.59886996, Time: 0.0209 Steps: 29510, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001843, Sample Num: 29488, Cur Loss: 0.21633197, Cur Avg Loss: 0.38360713, Log Avg loss: 0.40486700, Global Avg Loss: 1.59846549, Time: 0.0209 Steps: 29520, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001853, Sample Num: 29648, Cur Loss: 0.20444584, Cur Avg Loss: 0.38384143, Log Avg loss: 0.42702351, Global Avg Loss: 1.59806879, Time: 0.0209 Steps: 29530, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001863, Sample Num: 29808, Cur Loss: 1.10028577, Cur Avg Loss: 0.38449455, Log Avg loss: 0.50551666, Global Avg Loss: 1.59769894, Time: 0.0209 Steps: 29540, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001873, Sample Num: 29968, Cur Loss: 0.39302105, Cur Avg Loss: 0.38427008, Log Avg loss: 0.34245234, Global Avg Loss: 1.59727415, Time: 0.0209 Steps: 29550, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001883, Sample Num: 30128, Cur Loss: 0.21857424, Cur Avg Loss: 0.38374615, Log Avg loss: 0.28561409, Global Avg Loss: 1.59683042, Time: 0.0210 Steps: 29560, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001893, Sample Num: 30288, Cur Loss: 0.76497561, Cur Avg Loss: 0.38395365, Log Avg loss: 0.42302500, Global Avg Loss: 1.59643346, Time: 0.0210 Steps: 29570, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001903, Sample Num: 30448, Cur Loss: 0.55079097, Cur Avg Loss: 0.38356972, Log Avg loss: 0.31089206, Global Avg Loss: 1.59599886, Time: 0.0210 Steps: 29580, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001913, Sample Num: 30608, Cur Loss: 0.30660439, Cur Avg Loss: 0.38276064, Log Avg loss: 0.22879282, Global Avg Loss: 1.59553681, Time: 0.0210 Steps: 29590, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001923, Sample Num: 30768, Cur Loss: 0.52655208, Cur Avg Loss: 0.38237248, Log Avg loss: 0.30811874, Global Avg Loss: 1.59510188, Time: 0.0210 Steps: 29600, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001933, Sample Num: 30928, Cur Loss: 0.45437595, Cur Avg Loss: 0.38210889, Log Avg loss: 0.33142041, Global Avg Loss: 1.59467510, Time: 0.0210 Steps: 29610, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001943, Sample Num: 31088, Cur Loss: 0.46586305, Cur Avg Loss: 0.38195119, Log Avg loss: 0.35146672, Global Avg Loss: 1.59425538, Time: 0.0210 Steps: 29620, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001953, Sample Num: 31248, Cur Loss: 0.22233692, Cur Avg Loss: 0.38156068, Log Avg loss: 0.30568475, Global Avg Loss: 1.59382049, Time: 0.0210 Steps: 29630, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001963, Sample Num: 31408, Cur Loss: 0.68704748, Cur Avg Loss: 0.38142737, Log Avg loss: 0.35539270, Global Avg Loss: 1.59340267, Time: 0.0210 Steps: 29640, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001973, Sample Num: 31568, Cur Loss: 0.33852735, Cur Avg Loss: 0.38209405, Log Avg loss: 0.51296203, Global Avg Loss: 1.59303827, Time: 0.0210 Steps: 29650, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001983, Sample Num: 31728, Cur Loss: 0.18990701, Cur Avg Loss: 0.38192545, Log Avg loss: 0.34866180, Global Avg Loss: 1.59261873, Time: 0.0210 Steps: 29660, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001993, Sample Num: 31888, Cur Loss: 0.71401489, Cur Avg Loss: 0.38228846, Log Avg loss: 0.45427337, Global Avg Loss: 1.59223506, Time: 0.0210 Steps: 29670, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002003, Sample Num: 32048, Cur Loss: 0.24708098, Cur Avg Loss: 0.38240469, Log Avg loss: 0.40556907, Global Avg Loss: 1.59183524, Time: 0.0210 Steps: 29680, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002013, Sample Num: 32208, Cur Loss: 0.36992654, Cur Avg Loss: 0.38253731, Log Avg loss: 0.40910041, Global Avg Loss: 1.59143688, Time: 0.0210 Steps: 29690, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002023, Sample Num: 32368, Cur Loss: 0.39313543, Cur Avg Loss: 0.38272145, Log Avg loss: 0.41979007, Global Avg Loss: 1.59104238, Time: 0.0210 Steps: 29700, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002033, Sample Num: 32528, Cur Loss: 0.26826501, Cur Avg Loss: 0.38331293, Log Avg loss: 0.50296873, Global Avg Loss: 1.59067615, Time: 0.0210 Steps: 29710, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002043, Sample Num: 32688, Cur Loss: 0.16743241, Cur Avg Loss: 0.38436546, Log Avg loss: 0.59834488, Global Avg Loss: 1.59034226, Time: 0.0210 Steps: 29720, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002053, Sample Num: 32848, Cur Loss: 0.44118753, Cur Avg Loss: 0.38598778, Log Avg loss: 0.71742798, Global Avg Loss: 1.59004864, Time: 0.0232 Steps: 29730, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002063, Sample Num: 33008, Cur Loss: 0.40526956, Cur Avg Loss: 0.38575303, Log Avg loss: 0.33755760, Global Avg Loss: 1.58962750, Time: 0.0211 Steps: 29740, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002073, Sample Num: 33168, Cur Loss: 0.45483714, Cur Avg Loss: 0.38568644, Log Avg loss: 0.37195040, Global Avg Loss: 1.58921819, Time: 0.0211 Steps: 29750, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002083, Sample Num: 33328, Cur Loss: 0.43989295, Cur Avg Loss: 0.38521546, Log Avg loss: 0.28757967, Global Avg Loss: 1.58878081, Time: 0.0212 Steps: 29760, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002093, Sample Num: 33488, Cur Loss: 0.33256596, Cur Avg Loss: 0.38533397, Log Avg loss: 0.41001953, Global Avg Loss: 1.58838486, Time: 0.0211 Steps: 29770, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002103, Sample Num: 33648, Cur Loss: 0.38026470, Cur Avg Loss: 0.38570113, Log Avg loss: 0.46254790, Global Avg Loss: 1.58800681, Time: 0.0211 Steps: 29780, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002113, Sample Num: 33808, Cur Loss: 0.48638108, Cur Avg Loss: 0.38526922, Log Avg loss: 0.29443902, Global Avg Loss: 1.58757258, Time: 0.0211 Steps: 29790, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002123, Sample Num: 33968, Cur Loss: 0.09144866, Cur Avg Loss: 0.38479580, Log Avg loss: 0.28476210, Global Avg Loss: 1.58713539, Time: 0.0211 Steps: 29800, Updated lr: 0.000073 ***** Running evaluation checkpoint-29806 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-29806 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.731666, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.328874, "eval_total_loss": 231.198414, "eval_mae": 0.421522, "eval_mse": 0.328976, "eval_r2": 0.790881, "eval_sp_statistic": 0.85943, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.894229, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.559328, "test_total_loss": 280.782618, "test_mae": 0.478404, "test_mse": 0.559492, "test_r2": 0.638899, "test_sp_statistic": 0.84448, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.890485, "test_ps_pvalue": 0.0, "lr": 7.26827880512091e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5868891800873695, "train_cur_epoch_loss": 819.105681464076, "train_cur_epoch_avg_loss": 0.38473728579806293, "train_cur_epoch_time": 44.73166584968567, "train_cur_epoch_avg_time": 0.021010646242219665, "epoch": 14, "step": 29806} ################################################## Training, Epoch: 0015, Batch: 000004, Sample Num: 64, Cur Loss: 0.17005286, Cur Avg Loss: 0.32929466, Log Avg loss: 0.35013799, Global Avg Loss: 1.58672043, Time: 0.0248 Steps: 29810, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000014, Sample Num: 224, Cur Loss: 0.23185919, Cur Avg Loss: 0.25468248, Log Avg loss: 0.22483761, Global Avg Loss: 1.58626373, Time: 0.0211 Steps: 29820, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000024, Sample Num: 384, Cur Loss: 0.09894770, Cur Avg Loss: 0.25687030, Log Avg loss: 0.25993324, Global Avg Loss: 1.58581910, Time: 0.0210 Steps: 29830, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000034, Sample Num: 544, Cur Loss: 0.32625335, Cur Avg Loss: 0.33526755, Log Avg loss: 0.52342095, Global Avg Loss: 1.58546307, Time: 0.0211 Steps: 29840, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000044, Sample Num: 704, Cur Loss: 0.41034156, Cur Avg Loss: 0.32967738, Log Avg loss: 0.31067079, Global Avg Loss: 1.58503600, Time: 0.0210 Steps: 29850, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000054, Sample Num: 864, Cur Loss: 0.26981273, Cur Avg Loss: 0.33035031, Log Avg loss: 0.33331121, Global Avg Loss: 1.58461681, Time: 0.0210 Steps: 29860, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000064, Sample Num: 1024, Cur Loss: 0.53039110, Cur Avg Loss: 0.33121476, Log Avg loss: 0.33588279, Global Avg Loss: 1.58419875, Time: 0.0211 Steps: 29870, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000074, Sample Num: 1184, Cur Loss: 0.30263466, Cur Avg Loss: 0.34294974, Log Avg loss: 0.41805363, Global Avg Loss: 1.58380847, Time: 0.0211 Steps: 29880, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000084, Sample Num: 1344, Cur Loss: 0.21904597, Cur Avg Loss: 0.33261023, Log Avg loss: 0.25609782, Global Avg Loss: 1.58336427, Time: 0.0211 Steps: 29890, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000094, Sample Num: 1504, Cur Loss: 0.19924690, Cur Avg Loss: 0.34086506, Log Avg loss: 0.41020567, Global Avg Loss: 1.58297191, Time: 0.0212 Steps: 29900, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000104, Sample Num: 1664, Cur Loss: 0.19553021, Cur Avg Loss: 0.33742054, Log Avg loss: 0.30504200, Global Avg Loss: 1.58254466, Time: 0.0210 Steps: 29910, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000114, Sample Num: 1824, Cur Loss: 0.85519254, Cur Avg Loss: 0.34502769, Log Avg loss: 0.42414205, Global Avg Loss: 1.58215749, Time: 0.0210 Steps: 29920, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000124, Sample Num: 1984, Cur Loss: 0.63664293, Cur Avg Loss: 0.37350355, Log Avg loss: 0.69812841, Global Avg Loss: 1.58186212, Time: 0.0211 Steps: 29930, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000134, Sample Num: 2144, Cur Loss: 0.12374855, Cur Avg Loss: 0.37761673, Log Avg loss: 0.42862010, Global Avg Loss: 1.58147694, Time: 0.0212 Steps: 29940, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000144, Sample Num: 2304, Cur Loss: 0.07848576, Cur Avg Loss: 0.39415906, Log Avg loss: 0.61582628, Global Avg Loss: 1.58115452, Time: 0.0211 Steps: 29950, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000154, Sample Num: 2464, Cur Loss: 0.35488355, Cur Avg Loss: 0.38558413, Log Avg loss: 0.26210514, Global Avg Loss: 1.58071425, Time: 0.0211 Steps: 29960, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000164, Sample Num: 2624, Cur Loss: 0.52902353, Cur Avg Loss: 0.38158235, Log Avg loss: 0.31995505, Global Avg Loss: 1.58029357, Time: 0.0210 Steps: 29970, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000174, Sample Num: 2784, Cur Loss: 0.11359344, Cur Avg Loss: 0.37302356, Log Avg loss: 0.23265937, Global Avg Loss: 1.57984406, Time: 0.0211 Steps: 29980, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000184, Sample Num: 2944, Cur Loss: 0.37289152, Cur Avg Loss: 0.37501963, Log Avg loss: 0.40975121, Global Avg Loss: 1.57945390, Time: 0.0212 Steps: 29990, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000194, Sample Num: 3104, Cur Loss: 0.74072832, Cur Avg Loss: 0.37629293, Log Avg loss: 0.39972160, Global Avg Loss: 1.57906066, Time: 0.0210 Steps: 30000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000204, Sample Num: 3264, Cur Loss: 0.22695401, Cur Avg Loss: 0.37095272, Log Avg loss: 0.26735276, Global Avg Loss: 1.57862357, Time: 0.0210 Steps: 30010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000214, Sample Num: 3424, Cur Loss: 0.22108816, Cur Avg Loss: 0.36878156, Log Avg loss: 0.32448992, Global Avg Loss: 1.57820580, Time: 0.0211 Steps: 30020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000224, Sample Num: 3584, Cur Loss: 0.32540199, Cur Avg Loss: 0.36592395, Log Avg loss: 0.30477109, Global Avg Loss: 1.57778175, Time: 0.0210 Steps: 30030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000234, Sample Num: 3744, Cur Loss: 0.56298345, Cur Avg Loss: 0.36427058, Log Avg loss: 0.32723503, Global Avg Loss: 1.57736545, Time: 0.0210 Steps: 30040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000244, Sample Num: 3904, Cur Loss: 0.17057857, Cur Avg Loss: 0.36047914, Log Avg loss: 0.27175939, Global Avg Loss: 1.57693098, Time: 0.0210 Steps: 30050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000254, Sample Num: 4064, Cur Loss: 0.73705631, Cur Avg Loss: 0.36021387, Log Avg loss: 0.35374126, Global Avg Loss: 1.57652406, Time: 0.0211 Steps: 30060, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000264, Sample Num: 4224, Cur Loss: 0.22888470, Cur Avg Loss: 0.36395702, Log Avg loss: 0.45903320, Global Avg Loss: 1.57615243, Time: 0.0211 Steps: 30070, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000274, Sample Num: 4384, Cur Loss: 0.38150328, Cur Avg Loss: 0.36136894, Log Avg loss: 0.29304361, Global Avg Loss: 1.57572586, Time: 0.0210 Steps: 30080, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000284, Sample Num: 4544, Cur Loss: 0.44982433, Cur Avg Loss: 0.36132001, Log Avg loss: 0.35997938, Global Avg Loss: 1.57532183, Time: 0.0211 Steps: 30090, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000294, Sample Num: 4704, Cur Loss: 0.47277796, Cur Avg Loss: 0.36588214, Log Avg loss: 0.49544647, Global Avg Loss: 1.57496306, Time: 0.0210 Steps: 30100, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000304, Sample Num: 4864, Cur Loss: 0.24440083, Cur Avg Loss: 0.36477341, Log Avg loss: 0.33217669, Global Avg Loss: 1.57455032, Time: 0.0210 Steps: 30110, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000314, Sample Num: 5024, Cur Loss: 0.64071757, Cur Avg Loss: 0.36710831, Log Avg loss: 0.43808925, Global Avg Loss: 1.57417300, Time: 0.0210 Steps: 30120, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000324, Sample Num: 5184, Cur Loss: 0.34451938, Cur Avg Loss: 0.36900442, Log Avg loss: 0.42854232, Global Avg Loss: 1.57379278, Time: 0.0210 Steps: 30130, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000334, Sample Num: 5344, Cur Loss: 0.31766412, Cur Avg Loss: 0.37308322, Log Avg loss: 0.50523635, Global Avg Loss: 1.57343824, Time: 0.0210 Steps: 30140, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000344, Sample Num: 5504, Cur Loss: 0.24808642, Cur Avg Loss: 0.37173529, Log Avg loss: 0.32671441, Global Avg Loss: 1.57302474, Time: 0.0210 Steps: 30150, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000354, Sample Num: 5664, Cur Loss: 0.22046086, Cur Avg Loss: 0.36913652, Log Avg loss: 0.27973893, Global Avg Loss: 1.57259593, Time: 0.0210 Steps: 30160, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000364, Sample Num: 5824, Cur Loss: 0.38810673, Cur Avg Loss: 0.36777766, Log Avg loss: 0.31967386, Global Avg Loss: 1.57218064, Time: 0.0211 Steps: 30170, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000374, Sample Num: 5984, Cur Loss: 0.04499829, Cur Avg Loss: 0.36404705, Log Avg loss: 0.22825287, Global Avg Loss: 1.57173534, Time: 0.0211 Steps: 30180, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000384, Sample Num: 6144, Cur Loss: 0.35572684, Cur Avg Loss: 0.36519327, Log Avg loss: 0.40806201, Global Avg Loss: 1.57134989, Time: 0.0211 Steps: 30190, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000394, Sample Num: 6304, Cur Loss: 0.65613204, Cur Avg Loss: 0.36595303, Log Avg loss: 0.39512798, Global Avg Loss: 1.57096041, Time: 0.0210 Steps: 30200, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000404, Sample Num: 6464, Cur Loss: 0.61734384, Cur Avg Loss: 0.36856989, Log Avg loss: 0.47167393, Global Avg Loss: 1.57059653, Time: 0.0210 Steps: 30210, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000414, Sample Num: 6624, Cur Loss: 0.10150830, Cur Avg Loss: 0.36791744, Log Avg loss: 0.34155836, Global Avg Loss: 1.57018983, Time: 0.0210 Steps: 30220, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000424, Sample Num: 6784, Cur Loss: 0.39782938, Cur Avg Loss: 0.36516685, Log Avg loss: 0.25129247, Global Avg Loss: 1.56975354, Time: 0.0210 Steps: 30230, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000434, Sample Num: 6944, Cur Loss: 0.11758655, Cur Avg Loss: 0.36582409, Log Avg loss: 0.39369133, Global Avg Loss: 1.56936463, Time: 0.0210 Steps: 30240, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000444, Sample Num: 7104, Cur Loss: 0.27803177, Cur Avg Loss: 0.36431381, Log Avg loss: 0.29876748, Global Avg Loss: 1.56894460, Time: 0.0211 Steps: 30250, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000454, Sample Num: 7264, Cur Loss: 0.44322902, Cur Avg Loss: 0.36255071, Log Avg loss: 0.28426908, Global Avg Loss: 1.56852006, Time: 0.0210 Steps: 30260, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000464, Sample Num: 7424, Cur Loss: 0.29122651, Cur Avg Loss: 0.36130088, Log Avg loss: 0.30455862, Global Avg Loss: 1.56810249, Time: 0.0210 Steps: 30270, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000474, Sample Num: 7584, Cur Loss: 0.31988746, Cur Avg Loss: 0.36045200, Log Avg loss: 0.32106390, Global Avg Loss: 1.56769066, Time: 0.0210 Steps: 30280, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000484, Sample Num: 7744, Cur Loss: 0.10584387, Cur Avg Loss: 0.35786110, Log Avg loss: 0.23505232, Global Avg Loss: 1.56725070, Time: 0.0210 Steps: 30290, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000494, Sample Num: 7904, Cur Loss: 0.22619820, Cur Avg Loss: 0.35843029, Log Avg loss: 0.38597920, Global Avg Loss: 1.56686084, Time: 0.0210 Steps: 30300, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000504, Sample Num: 8064, Cur Loss: 0.39801455, Cur Avg Loss: 0.35790124, Log Avg loss: 0.33176620, Global Avg Loss: 1.56645335, Time: 0.0210 Steps: 30310, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000514, Sample Num: 8224, Cur Loss: 0.27567628, Cur Avg Loss: 0.36106300, Log Avg loss: 0.52041558, Global Avg Loss: 1.56610835, Time: 0.0250 Steps: 30320, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000524, Sample Num: 8384, Cur Loss: 0.53766418, Cur Avg Loss: 0.36263881, Log Avg loss: 0.44363578, Global Avg Loss: 1.56573827, Time: 0.0211 Steps: 30330, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000534, Sample Num: 8544, Cur Loss: 0.47323245, Cur Avg Loss: 0.36363707, Log Avg loss: 0.41594562, Global Avg Loss: 1.56535930, Time: 0.0211 Steps: 30340, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000544, Sample Num: 8704, Cur Loss: 0.27333325, Cur Avg Loss: 0.36176762, Log Avg loss: 0.26193897, Global Avg Loss: 1.56492983, Time: 0.0211 Steps: 30350, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000554, Sample Num: 8864, Cur Loss: 0.36377531, Cur Avg Loss: 0.36077278, Log Avg loss: 0.30665386, Global Avg Loss: 1.56451538, Time: 0.0211 Steps: 30360, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000564, Sample Num: 9024, Cur Loss: 0.25497022, Cur Avg Loss: 0.35793191, Log Avg loss: 0.20054736, Global Avg Loss: 1.56406627, Time: 0.0212 Steps: 30370, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000574, Sample Num: 9184, Cur Loss: 0.26100230, Cur Avg Loss: 0.35908861, Log Avg loss: 0.42432662, Global Avg Loss: 1.56369110, Time: 0.0211 Steps: 30380, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000584, Sample Num: 9344, Cur Loss: 0.33198798, Cur Avg Loss: 0.35893784, Log Avg loss: 0.35028362, Global Avg Loss: 1.56329183, Time: 0.0211 Steps: 30390, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000594, Sample Num: 9504, Cur Loss: 0.06398787, Cur Avg Loss: 0.35639333, Log Avg loss: 0.20779409, Global Avg Loss: 1.56284594, Time: 0.0211 Steps: 30400, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000604, Sample Num: 9664, Cur Loss: 0.10416642, Cur Avg Loss: 0.35626889, Log Avg loss: 0.34887686, Global Avg Loss: 1.56244674, Time: 0.0211 Steps: 30410, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000614, Sample Num: 9824, Cur Loss: 0.19816169, Cur Avg Loss: 0.35519030, Log Avg loss: 0.29004384, Global Avg Loss: 1.56202846, Time: 0.0212 Steps: 30420, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000624, Sample Num: 9984, Cur Loss: 0.16095898, Cur Avg Loss: 0.35408735, Log Avg loss: 0.28636624, Global Avg Loss: 1.56160925, Time: 0.0211 Steps: 30430, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000634, Sample Num: 10144, Cur Loss: 0.20660736, Cur Avg Loss: 0.35393076, Log Avg loss: 0.34415905, Global Avg Loss: 1.56120930, Time: 0.0211 Steps: 30440, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000644, Sample Num: 10304, Cur Loss: 0.48895794, Cur Avg Loss: 0.35204897, Log Avg loss: 0.23274397, Global Avg Loss: 1.56077302, Time: 0.0212 Steps: 30450, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000654, Sample Num: 10464, Cur Loss: 0.28371826, Cur Avg Loss: 0.35168697, Log Avg loss: 0.32837415, Global Avg Loss: 1.56036842, Time: 0.0212 Steps: 30460, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000664, Sample Num: 10624, Cur Loss: 0.09695788, Cur Avg Loss: 0.35445771, Log Avg loss: 0.53566402, Global Avg Loss: 1.56003212, Time: 0.0211 Steps: 30470, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000674, Sample Num: 10784, Cur Loss: 0.16315182, Cur Avg Loss: 0.35596967, Log Avg loss: 0.45636349, Global Avg Loss: 1.55967003, Time: 0.0212 Steps: 30480, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000684, Sample Num: 10944, Cur Loss: 0.40668297, Cur Avg Loss: 0.35611013, Log Avg loss: 0.36557711, Global Avg Loss: 1.55927839, Time: 0.0211 Steps: 30490, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000694, Sample Num: 11104, Cur Loss: 0.20572874, Cur Avg Loss: 0.35542995, Log Avg loss: 0.30890586, Global Avg Loss: 1.55886844, Time: 0.0212 Steps: 30500, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000704, Sample Num: 11264, Cur Loss: 0.74569094, Cur Avg Loss: 0.35710297, Log Avg loss: 0.47321079, Global Avg Loss: 1.55851260, Time: 0.0211 Steps: 30510, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000714, Sample Num: 11424, Cur Loss: 0.29619029, Cur Avg Loss: 0.35672828, Log Avg loss: 0.33034952, Global Avg Loss: 1.55811019, Time: 0.0211 Steps: 30520, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000724, Sample Num: 11584, Cur Loss: 0.16515401, Cur Avg Loss: 0.35526131, Log Avg loss: 0.25051998, Global Avg Loss: 1.55768189, Time: 0.0210 Steps: 30530, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000734, Sample Num: 11744, Cur Loss: 0.35266080, Cur Avg Loss: 0.35558177, Log Avg loss: 0.37878283, Global Avg Loss: 1.55729587, Time: 0.0211 Steps: 30540, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000744, Sample Num: 11904, Cur Loss: 0.23218164, Cur Avg Loss: 0.35618410, Log Avg loss: 0.40039519, Global Avg Loss: 1.55691718, Time: 0.0210 Steps: 30550, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000754, Sample Num: 12064, Cur Loss: 0.12718022, Cur Avg Loss: 0.35421588, Log Avg loss: 0.20778029, Global Avg Loss: 1.55647571, Time: 0.0210 Steps: 30560, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000764, Sample Num: 12224, Cur Loss: 0.22552574, Cur Avg Loss: 0.35324864, Log Avg loss: 0.28031901, Global Avg Loss: 1.55605826, Time: 0.0211 Steps: 30570, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000774, Sample Num: 12384, Cur Loss: 0.21197031, Cur Avg Loss: 0.35313281, Log Avg loss: 0.34428370, Global Avg Loss: 1.55566199, Time: 0.0211 Steps: 30580, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000784, Sample Num: 12544, Cur Loss: 0.46523035, Cur Avg Loss: 0.35616680, Log Avg loss: 0.59099708, Global Avg Loss: 1.55534664, Time: 0.0211 Steps: 30590, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000794, Sample Num: 12704, Cur Loss: 0.79882926, Cur Avg Loss: 0.35781632, Log Avg loss: 0.48713900, Global Avg Loss: 1.55499755, Time: 0.0210 Steps: 30600, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000804, Sample Num: 12864, Cur Loss: 0.36249048, Cur Avg Loss: 0.35785339, Log Avg loss: 0.36079661, Global Avg Loss: 1.55460742, Time: 0.0211 Steps: 30610, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000814, Sample Num: 13024, Cur Loss: 0.16708469, Cur Avg Loss: 0.35835525, Log Avg loss: 0.39870448, Global Avg Loss: 1.55422992, Time: 0.0211 Steps: 30620, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000824, Sample Num: 13184, Cur Loss: 0.07300003, Cur Avg Loss: 0.35688457, Log Avg loss: 0.23717149, Global Avg Loss: 1.55379993, Time: 0.0211 Steps: 30630, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000834, Sample Num: 13344, Cur Loss: 0.33438498, Cur Avg Loss: 0.35657832, Log Avg loss: 0.33134321, Global Avg Loss: 1.55340095, Time: 0.0211 Steps: 30640, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000844, Sample Num: 13504, Cur Loss: 0.06996014, Cur Avg Loss: 0.35893705, Log Avg loss: 0.55565507, Global Avg Loss: 1.55307542, Time: 0.0211 Steps: 30650, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000854, Sample Num: 13664, Cur Loss: 0.36604047, Cur Avg Loss: 0.35864003, Log Avg loss: 0.33357161, Global Avg Loss: 1.55267767, Time: 0.0211 Steps: 30660, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000864, Sample Num: 13824, Cur Loss: 0.31050909, Cur Avg Loss: 0.35814801, Log Avg loss: 0.31612941, Global Avg Loss: 1.55227450, Time: 0.0210 Steps: 30670, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000874, Sample Num: 13984, Cur Loss: 0.29650921, Cur Avg Loss: 0.35730905, Log Avg loss: 0.28482331, Global Avg Loss: 1.55186138, Time: 0.0211 Steps: 30680, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000884, Sample Num: 14144, Cur Loss: 0.11283045, Cur Avg Loss: 0.35632344, Log Avg loss: 0.27018134, Global Avg Loss: 1.55144375, Time: 0.0211 Steps: 30690, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000894, Sample Num: 14304, Cur Loss: 0.32762176, Cur Avg Loss: 0.35518252, Log Avg loss: 0.25432437, Global Avg Loss: 1.55102124, Time: 0.0212 Steps: 30700, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000904, Sample Num: 14464, Cur Loss: 0.29025736, Cur Avg Loss: 0.35542927, Log Avg loss: 0.37748948, Global Avg Loss: 1.55063911, Time: 0.0211 Steps: 30710, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000914, Sample Num: 14624, Cur Loss: 0.63885808, Cur Avg Loss: 0.35651670, Log Avg loss: 0.45482010, Global Avg Loss: 1.55028239, Time: 0.0211 Steps: 30720, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000924, Sample Num: 14784, Cur Loss: 0.10006630, Cur Avg Loss: 0.35634340, Log Avg loss: 0.34050418, Global Avg Loss: 1.54988871, Time: 0.0210 Steps: 30730, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000934, Sample Num: 14944, Cur Loss: 0.21963552, Cur Avg Loss: 0.35590206, Log Avg loss: 0.31512184, Global Avg Loss: 1.54948703, Time: 0.0210 Steps: 30740, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000944, Sample Num: 15104, Cur Loss: 0.52710342, Cur Avg Loss: 0.35530409, Log Avg loss: 0.29945382, Global Avg Loss: 1.54908052, Time: 0.0211 Steps: 30750, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000954, Sample Num: 15264, Cur Loss: 0.26412949, Cur Avg Loss: 0.35488435, Log Avg loss: 0.31526107, Global Avg Loss: 1.54867941, Time: 0.0211 Steps: 30760, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000964, Sample Num: 15424, Cur Loss: 0.54978585, Cur Avg Loss: 0.35699379, Log Avg loss: 0.55823440, Global Avg Loss: 1.54835752, Time: 0.0210 Steps: 30770, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000974, Sample Num: 15584, Cur Loss: 0.28629997, Cur Avg Loss: 0.35713896, Log Avg loss: 0.37113274, Global Avg Loss: 1.54797506, Time: 0.0211 Steps: 30780, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000984, Sample Num: 15744, Cur Loss: 0.27524081, Cur Avg Loss: 0.35718550, Log Avg loss: 0.36171868, Global Avg Loss: 1.54758978, Time: 0.0211 Steps: 30790, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000994, Sample Num: 15904, Cur Loss: 0.38707986, Cur Avg Loss: 0.35780689, Log Avg loss: 0.41895206, Global Avg Loss: 1.54722334, Time: 0.0211 Steps: 30800, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001004, Sample Num: 16064, Cur Loss: 0.25601912, Cur Avg Loss: 0.35680200, Log Avg loss: 0.25691548, Global Avg Loss: 1.54680455, Time: 0.0210 Steps: 30810, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001014, Sample Num: 16224, Cur Loss: 0.28029189, Cur Avg Loss: 0.35672351, Log Avg loss: 0.34884346, Global Avg Loss: 1.54641585, Time: 0.0211 Steps: 30820, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001024, Sample Num: 16384, Cur Loss: 0.25835484, Cur Avg Loss: 0.35624129, Log Avg loss: 0.30734368, Global Avg Loss: 1.54601395, Time: 0.0256 Steps: 30830, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001034, Sample Num: 16544, Cur Loss: 0.36908871, Cur Avg Loss: 0.35556590, Log Avg loss: 0.28640642, Global Avg Loss: 1.54560551, Time: 0.0211 Steps: 30840, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001044, Sample Num: 16704, Cur Loss: 0.62059689, Cur Avg Loss: 0.35596376, Log Avg loss: 0.39710246, Global Avg Loss: 1.54523323, Time: 0.0211 Steps: 30850, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001054, Sample Num: 16864, Cur Loss: 0.40001494, Cur Avg Loss: 0.35572321, Log Avg loss: 0.33060952, Global Avg Loss: 1.54483964, Time: 0.0210 Steps: 30860, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001064, Sample Num: 17024, Cur Loss: 0.40381432, Cur Avg Loss: 0.35554538, Log Avg loss: 0.33680259, Global Avg Loss: 1.54444831, Time: 0.0211 Steps: 30870, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001074, Sample Num: 17184, Cur Loss: 0.78241318, Cur Avg Loss: 0.35623164, Log Avg loss: 0.42924969, Global Avg Loss: 1.54408717, Time: 0.0210 Steps: 30880, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001084, Sample Num: 17344, Cur Loss: 0.53097570, Cur Avg Loss: 0.35680538, Log Avg loss: 0.41842505, Global Avg Loss: 1.54372276, Time: 0.0211 Steps: 30890, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001094, Sample Num: 17504, Cur Loss: 0.23629418, Cur Avg Loss: 0.35694716, Log Avg loss: 0.37231544, Global Avg Loss: 1.54334366, Time: 0.0210 Steps: 30900, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001104, Sample Num: 17664, Cur Loss: 0.18877131, Cur Avg Loss: 0.35712655, Log Avg loss: 0.37675195, Global Avg Loss: 1.54296624, Time: 0.0210 Steps: 30910, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001114, Sample Num: 17824, Cur Loss: 0.15192299, Cur Avg Loss: 0.35644695, Log Avg loss: 0.28141958, Global Avg Loss: 1.54255824, Time: 0.0211 Steps: 30920, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001124, Sample Num: 17984, Cur Loss: 0.22925386, Cur Avg Loss: 0.35582353, Log Avg loss: 0.28637368, Global Avg Loss: 1.54215210, Time: 0.0211 Steps: 30930, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001134, Sample Num: 18144, Cur Loss: 0.38233089, Cur Avg Loss: 0.35524734, Log Avg loss: 0.29048458, Global Avg Loss: 1.54174756, Time: 0.0211 Steps: 30940, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001144, Sample Num: 18304, Cur Loss: 0.24776292, Cur Avg Loss: 0.35604552, Log Avg loss: 0.44655822, Global Avg Loss: 1.54139370, Time: 0.0211 Steps: 30950, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001154, Sample Num: 18464, Cur Loss: 0.41694310, Cur Avg Loss: 0.35634781, Log Avg loss: 0.39092997, Global Avg Loss: 1.54102210, Time: 0.0211 Steps: 30960, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001164, Sample Num: 18624, Cur Loss: 0.17574580, Cur Avg Loss: 0.35614463, Log Avg loss: 0.33269847, Global Avg Loss: 1.54063194, Time: 0.0211 Steps: 30970, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001174, Sample Num: 18784, Cur Loss: 1.33291328, Cur Avg Loss: 0.35659559, Log Avg loss: 0.40908705, Global Avg Loss: 1.54026669, Time: 0.0210 Steps: 30980, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001184, Sample Num: 18944, Cur Loss: 0.33919203, Cur Avg Loss: 0.35692267, Log Avg loss: 0.39532129, Global Avg Loss: 1.53989724, Time: 0.0211 Steps: 30990, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001194, Sample Num: 19104, Cur Loss: 0.40677607, Cur Avg Loss: 0.35663639, Log Avg loss: 0.32274082, Global Avg Loss: 1.53950460, Time: 0.0210 Steps: 31000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001204, Sample Num: 19264, Cur Loss: 0.78301448, Cur Avg Loss: 0.35720561, Log Avg loss: 0.42517129, Global Avg Loss: 1.53914526, Time: 0.0210 Steps: 31010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001214, Sample Num: 19424, Cur Loss: 0.20931432, Cur Avg Loss: 0.35898544, Log Avg loss: 0.57327673, Global Avg Loss: 1.53883389, Time: 0.0210 Steps: 31020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001224, Sample Num: 19584, Cur Loss: 0.50089908, Cur Avg Loss: 0.35936418, Log Avg loss: 0.40534351, Global Avg Loss: 1.53846860, Time: 0.0211 Steps: 31030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001234, Sample Num: 19744, Cur Loss: 0.11258977, Cur Avg Loss: 0.35855328, Log Avg loss: 0.25929808, Global Avg Loss: 1.53805650, Time: 0.0210 Steps: 31040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001244, Sample Num: 19904, Cur Loss: 0.29898524, Cur Avg Loss: 0.35824729, Log Avg loss: 0.32048861, Global Avg Loss: 1.53766436, Time: 0.0211 Steps: 31050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001254, Sample Num: 20064, Cur Loss: 0.26285702, Cur Avg Loss: 0.35915567, Log Avg loss: 0.47215788, Global Avg Loss: 1.53732132, Time: 0.0210 Steps: 31060, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001264, Sample Num: 20224, Cur Loss: 0.19515496, Cur Avg Loss: 0.35950222, Log Avg loss: 0.40296031, Global Avg Loss: 1.53695622, Time: 0.0211 Steps: 31070, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001274, Sample Num: 20384, Cur Loss: 0.68546033, Cur Avg Loss: 0.35991965, Log Avg loss: 0.41268232, Global Avg Loss: 1.53659448, Time: 0.0211 Steps: 31080, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001284, Sample Num: 20544, Cur Loss: 0.76661003, Cur Avg Loss: 0.36025574, Log Avg loss: 0.40307408, Global Avg Loss: 1.53622989, Time: 0.0248 Steps: 31090, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001294, Sample Num: 20704, Cur Loss: 1.28084540, Cur Avg Loss: 0.36192124, Log Avg loss: 0.57577151, Global Avg Loss: 1.53592106, Time: 0.0211 Steps: 31100, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001304, Sample Num: 20864, Cur Loss: 0.31205142, Cur Avg Loss: 0.36351162, Log Avg loss: 0.56930682, Global Avg Loss: 1.53561035, Time: 0.0210 Steps: 31110, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001314, Sample Num: 21024, Cur Loss: 0.65063655, Cur Avg Loss: 0.36495447, Log Avg loss: 0.55310157, Global Avg Loss: 1.53529464, Time: 0.0210 Steps: 31120, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001324, Sample Num: 21184, Cur Loss: 0.24236353, Cur Avg Loss: 0.36472746, Log Avg loss: 0.33489811, Global Avg Loss: 1.53490903, Time: 0.0210 Steps: 31130, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001334, Sample Num: 21344, Cur Loss: 0.12837812, Cur Avg Loss: 0.36507238, Log Avg loss: 0.41073964, Global Avg Loss: 1.53454802, Time: 0.0210 Steps: 31140, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001344, Sample Num: 21504, Cur Loss: 0.14989983, Cur Avg Loss: 0.36493853, Log Avg loss: 0.34708297, Global Avg Loss: 1.53416681, Time: 0.0211 Steps: 31150, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001354, Sample Num: 21664, Cur Loss: 0.14586049, Cur Avg Loss: 0.36406316, Log Avg loss: 0.24641360, Global Avg Loss: 1.53375354, Time: 0.0211 Steps: 31160, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001364, Sample Num: 21824, Cur Loss: 0.47732431, Cur Avg Loss: 0.36441963, Log Avg loss: 0.41268612, Global Avg Loss: 1.53339388, Time: 0.0210 Steps: 31170, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001374, Sample Num: 21984, Cur Loss: 0.29645047, Cur Avg Loss: 0.36406199, Log Avg loss: 0.31528016, Global Avg Loss: 1.53300321, Time: 0.0210 Steps: 31180, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001384, Sample Num: 22144, Cur Loss: 0.29075497, Cur Avg Loss: 0.36444988, Log Avg loss: 0.41774573, Global Avg Loss: 1.53264564, Time: 0.0210 Steps: 31190, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001394, Sample Num: 22304, Cur Loss: 0.42768168, Cur Avg Loss: 0.36477241, Log Avg loss: 0.40941084, Global Avg Loss: 1.53228563, Time: 0.0210 Steps: 31200, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001404, Sample Num: 22464, Cur Loss: 0.11427961, Cur Avg Loss: 0.36387822, Log Avg loss: 0.23922761, Global Avg Loss: 1.53187132, Time: 0.0211 Steps: 31210, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001414, Sample Num: 22624, Cur Loss: 0.43646586, Cur Avg Loss: 0.36392307, Log Avg loss: 0.37021937, Global Avg Loss: 1.53149923, Time: 0.0211 Steps: 31220, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001424, Sample Num: 22784, Cur Loss: 0.17119150, Cur Avg Loss: 0.36689260, Log Avg loss: 0.78678535, Global Avg Loss: 1.53126077, Time: 0.0211 Steps: 31230, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001434, Sample Num: 22944, Cur Loss: 0.23438036, Cur Avg Loss: 0.36637700, Log Avg loss: 0.29295424, Global Avg Loss: 1.53086439, Time: 0.0211 Steps: 31240, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001444, Sample Num: 23104, Cur Loss: 0.48469198, Cur Avg Loss: 0.36708793, Log Avg loss: 0.46903660, Global Avg Loss: 1.53052460, Time: 0.0210 Steps: 31250, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001454, Sample Num: 23264, Cur Loss: 0.97640473, Cur Avg Loss: 0.36675886, Log Avg loss: 0.31923985, Global Avg Loss: 1.53013712, Time: 0.0210 Steps: 31260, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001464, Sample Num: 23424, Cur Loss: 0.47201908, Cur Avg Loss: 0.36681883, Log Avg loss: 0.37553985, Global Avg Loss: 1.52976788, Time: 0.0211 Steps: 31270, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001474, Sample Num: 23584, Cur Loss: 0.27583230, Cur Avg Loss: 0.36593074, Log Avg loss: 0.23591421, Global Avg Loss: 1.52935425, Time: 0.0210 Steps: 31280, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001484, Sample Num: 23744, Cur Loss: 0.09668154, Cur Avg Loss: 0.36524236, Log Avg loss: 0.26377444, Global Avg Loss: 1.52894978, Time: 0.0210 Steps: 31290, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001494, Sample Num: 23904, Cur Loss: 0.41220623, Cur Avg Loss: 0.36582500, Log Avg loss: 0.45228870, Global Avg Loss: 1.52860580, Time: 0.0211 Steps: 31300, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001504, Sample Num: 24064, Cur Loss: 0.25165552, Cur Avg Loss: 0.36536881, Log Avg loss: 0.29721419, Global Avg Loss: 1.52821251, Time: 0.0211 Steps: 31310, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001514, Sample Num: 24224, Cur Loss: 1.04899192, Cur Avg Loss: 0.36557763, Log Avg loss: 0.39698427, Global Avg Loss: 1.52785132, Time: 0.0211 Steps: 31320, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001524, Sample Num: 24384, Cur Loss: 0.12919147, Cur Avg Loss: 0.36454859, Log Avg loss: 0.20875180, Global Avg Loss: 1.52743029, Time: 0.0211 Steps: 31330, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001534, Sample Num: 24544, Cur Loss: 0.15360841, Cur Avg Loss: 0.36396257, Log Avg loss: 0.27465284, Global Avg Loss: 1.52703055, Time: 0.0211 Steps: 31340, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001544, Sample Num: 24704, Cur Loss: 0.16532302, Cur Avg Loss: 0.36336698, Log Avg loss: 0.27200438, Global Avg Loss: 1.52663022, Time: 0.0211 Steps: 31350, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001554, Sample Num: 24864, Cur Loss: 0.03759593, Cur Avg Loss: 0.36275430, Log Avg loss: 0.26815528, Global Avg Loss: 1.52622892, Time: 0.0211 Steps: 31360, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001564, Sample Num: 25024, Cur Loss: 0.26946938, Cur Avg Loss: 0.36378849, Log Avg loss: 0.52450195, Global Avg Loss: 1.52590960, Time: 0.0211 Steps: 31370, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001574, Sample Num: 25184, Cur Loss: 0.12845355, Cur Avg Loss: 0.36464980, Log Avg loss: 0.49935954, Global Avg Loss: 1.52558246, Time: 0.0211 Steps: 31380, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001584, Sample Num: 25344, Cur Loss: 0.32214680, Cur Avg Loss: 0.36454992, Log Avg loss: 0.34882899, Global Avg Loss: 1.52520758, Time: 0.0212 Steps: 31390, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001594, Sample Num: 25504, Cur Loss: 0.19616546, Cur Avg Loss: 0.36482944, Log Avg loss: 0.40910481, Global Avg Loss: 1.52485213, Time: 0.0211 Steps: 31400, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001604, Sample Num: 25664, Cur Loss: 0.13020639, Cur Avg Loss: 0.36511788, Log Avg loss: 0.41109457, Global Avg Loss: 1.52449755, Time: 0.0211 Steps: 31410, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001614, Sample Num: 25824, Cur Loss: 0.24876252, Cur Avg Loss: 0.36437025, Log Avg loss: 0.24445119, Global Avg Loss: 1.52409015, Time: 0.0211 Steps: 31420, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001624, Sample Num: 25984, Cur Loss: 0.45616725, Cur Avg Loss: 0.36510717, Log Avg loss: 0.48404604, Global Avg Loss: 1.52375924, Time: 0.0211 Steps: 31430, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001634, Sample Num: 26144, Cur Loss: 0.32331142, Cur Avg Loss: 0.36587287, Log Avg loss: 0.49022167, Global Avg Loss: 1.52343051, Time: 0.0211 Steps: 31440, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001644, Sample Num: 26304, Cur Loss: 0.21883652, Cur Avg Loss: 0.36653567, Log Avg loss: 0.47483796, Global Avg Loss: 1.52309709, Time: 0.0211 Steps: 31450, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001654, Sample Num: 26464, Cur Loss: 0.49412942, Cur Avg Loss: 0.36695550, Log Avg loss: 0.43597536, Global Avg Loss: 1.52275154, Time: 0.0211 Steps: 31460, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001664, Sample Num: 26624, Cur Loss: 0.18422353, Cur Avg Loss: 0.36704745, Log Avg loss: 0.38225530, Global Avg Loss: 1.52238913, Time: 0.0211 Steps: 31470, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001674, Sample Num: 26784, Cur Loss: 0.46590996, Cur Avg Loss: 0.36712357, Log Avg loss: 0.37979120, Global Avg Loss: 1.52202617, Time: 0.0211 Steps: 31480, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001684, Sample Num: 26944, Cur Loss: 0.06771605, Cur Avg Loss: 0.36682782, Log Avg loss: 0.31731902, Global Avg Loss: 1.52164360, Time: 0.0211 Steps: 31490, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001694, Sample Num: 27104, Cur Loss: 0.24996938, Cur Avg Loss: 0.36679202, Log Avg loss: 0.36076258, Global Avg Loss: 1.52127507, Time: 0.0211 Steps: 31500, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001704, Sample Num: 27264, Cur Loss: 0.29701844, Cur Avg Loss: 0.36623103, Log Avg loss: 0.27120021, Global Avg Loss: 1.52087834, Time: 0.0211 Steps: 31510, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001714, Sample Num: 27424, Cur Loss: 0.23260106, Cur Avg Loss: 0.36620449, Log Avg loss: 0.36168168, Global Avg Loss: 1.52051058, Time: 0.0211 Steps: 31520, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001724, Sample Num: 27584, Cur Loss: 0.16665533, Cur Avg Loss: 0.36577499, Log Avg loss: 0.29215894, Global Avg Loss: 1.52012100, Time: 0.0211 Steps: 31530, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001734, Sample Num: 27744, Cur Loss: 0.29618299, Cur Avg Loss: 0.36533392, Log Avg loss: 0.28929394, Global Avg Loss: 1.51973075, Time: 0.0211 Steps: 31540, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001744, Sample Num: 27904, Cur Loss: 0.83967280, Cur Avg Loss: 0.36523036, Log Avg loss: 0.34727149, Global Avg Loss: 1.51935913, Time: 0.0211 Steps: 31550, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001754, Sample Num: 28064, Cur Loss: 0.09327743, Cur Avg Loss: 0.36539490, Log Avg loss: 0.39409068, Global Avg Loss: 1.51900258, Time: 0.0211 Steps: 31560, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001764, Sample Num: 28224, Cur Loss: 0.09557129, Cur Avg Loss: 0.36521071, Log Avg loss: 0.33290504, Global Avg Loss: 1.51862688, Time: 0.0212 Steps: 31570, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001774, Sample Num: 28384, Cur Loss: 0.27333158, Cur Avg Loss: 0.36448229, Log Avg loss: 0.23598794, Global Avg Loss: 1.51822072, Time: 0.0211 Steps: 31580, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001784, Sample Num: 28544, Cur Loss: 0.19418797, Cur Avg Loss: 0.36476195, Log Avg loss: 0.41437386, Global Avg Loss: 1.51787130, Time: 0.0211 Steps: 31590, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001794, Sample Num: 28704, Cur Loss: 0.44093001, Cur Avg Loss: 0.36453165, Log Avg loss: 0.32344608, Global Avg Loss: 1.51749331, Time: 0.0250 Steps: 31600, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001804, Sample Num: 28864, Cur Loss: 0.28758827, Cur Avg Loss: 0.36491426, Log Avg loss: 0.43355520, Global Avg Loss: 1.51715040, Time: 0.0212 Steps: 31610, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001814, Sample Num: 29024, Cur Loss: 0.42033052, Cur Avg Loss: 0.36470604, Log Avg loss: 0.32714321, Global Avg Loss: 1.51677406, Time: 0.0212 Steps: 31620, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001824, Sample Num: 29184, Cur Loss: 0.62853992, Cur Avg Loss: 0.36497186, Log Avg loss: 0.41319097, Global Avg Loss: 1.51642515, Time: 0.0212 Steps: 31630, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001834, Sample Num: 29344, Cur Loss: 0.49689946, Cur Avg Loss: 0.36460407, Log Avg loss: 0.29751924, Global Avg Loss: 1.51603991, Time: 0.0212 Steps: 31640, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001844, Sample Num: 29504, Cur Loss: 0.25850126, Cur Avg Loss: 0.36424561, Log Avg loss: 0.29850363, Global Avg Loss: 1.51565522, Time: 0.0212 Steps: 31650, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001854, Sample Num: 29664, Cur Loss: 0.56307054, Cur Avg Loss: 0.36437592, Log Avg loss: 0.38840541, Global Avg Loss: 1.51529917, Time: 0.0212 Steps: 31660, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001864, Sample Num: 29824, Cur Loss: 0.24209879, Cur Avg Loss: 0.36388950, Log Avg loss: 0.27370832, Global Avg Loss: 1.51490713, Time: 0.0212 Steps: 31670, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001874, Sample Num: 29984, Cur Loss: 0.51811206, Cur Avg Loss: 0.36377052, Log Avg loss: 0.34159142, Global Avg Loss: 1.51453677, Time: 0.0212 Steps: 31680, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001884, Sample Num: 30144, Cur Loss: 0.67984998, Cur Avg Loss: 0.36357651, Log Avg loss: 0.32721958, Global Avg Loss: 1.51416210, Time: 0.0212 Steps: 31690, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001894, Sample Num: 30304, Cur Loss: 0.52239990, Cur Avg Loss: 0.36361279, Log Avg loss: 0.37044827, Global Avg Loss: 1.51380131, Time: 0.0212 Steps: 31700, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001904, Sample Num: 30464, Cur Loss: 0.71381712, Cur Avg Loss: 0.36296452, Log Avg loss: 0.24018208, Global Avg Loss: 1.51339966, Time: 0.0212 Steps: 31710, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001914, Sample Num: 30624, Cur Loss: 0.12468477, Cur Avg Loss: 0.36253674, Log Avg loss: 0.28108806, Global Avg Loss: 1.51301117, Time: 0.0212 Steps: 31720, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001924, Sample Num: 30784, Cur Loss: 0.29198128, Cur Avg Loss: 0.36214300, Log Avg loss: 0.28677964, Global Avg Loss: 1.51262471, Time: 0.0212 Steps: 31730, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001934, Sample Num: 30944, Cur Loss: 0.69712740, Cur Avg Loss: 0.36228520, Log Avg loss: 0.38964421, Global Avg Loss: 1.51227090, Time: 0.0213 Steps: 31740, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001944, Sample Num: 31104, Cur Loss: 0.73696482, Cur Avg Loss: 0.36280006, Log Avg loss: 0.46237419, Global Avg Loss: 1.51194023, Time: 0.0212 Steps: 31750, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001954, Sample Num: 31264, Cur Loss: 0.17632040, Cur Avg Loss: 0.36220069, Log Avg loss: 0.24568436, Global Avg Loss: 1.51154153, Time: 0.0212 Steps: 31760, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001964, Sample Num: 31424, Cur Loss: 0.19263442, Cur Avg Loss: 0.36244779, Log Avg loss: 0.41073145, Global Avg Loss: 1.51119504, Time: 0.0212 Steps: 31770, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001974, Sample Num: 31584, Cur Loss: 0.27445054, Cur Avg Loss: 0.36314559, Log Avg loss: 0.50019270, Global Avg Loss: 1.51087691, Time: 0.0212 Steps: 31780, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001984, Sample Num: 31744, Cur Loss: 0.42320383, Cur Avg Loss: 0.36356110, Log Avg loss: 0.44558321, Global Avg Loss: 1.51054181, Time: 0.0212 Steps: 31790, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001994, Sample Num: 31904, Cur Loss: 0.08832195, Cur Avg Loss: 0.36352436, Log Avg loss: 0.35623474, Global Avg Loss: 1.51017882, Time: 0.0212 Steps: 31800, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002004, Sample Num: 32064, Cur Loss: 0.15188920, Cur Avg Loss: 0.36337144, Log Avg loss: 0.33287916, Global Avg Loss: 1.50980872, Time: 0.0212 Steps: 31810, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002014, Sample Num: 32224, Cur Loss: 0.16136096, Cur Avg Loss: 0.36296871, Log Avg loss: 0.28226200, Global Avg Loss: 1.50942294, Time: 0.0212 Steps: 31820, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002024, Sample Num: 32384, Cur Loss: 0.14085075, Cur Avg Loss: 0.36250168, Log Avg loss: 0.26844065, Global Avg Loss: 1.50903306, Time: 0.0213 Steps: 31830, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002034, Sample Num: 32544, Cur Loss: 0.19054334, Cur Avg Loss: 0.36333210, Log Avg loss: 0.53140971, Global Avg Loss: 1.50872602, Time: 0.0212 Steps: 31840, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002044, Sample Num: 32704, Cur Loss: 0.47100371, Cur Avg Loss: 0.36317206, Log Avg loss: 0.33061979, Global Avg Loss: 1.50835613, Time: 0.0212 Steps: 31850, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002054, Sample Num: 32864, Cur Loss: 0.32446581, Cur Avg Loss: 0.36330621, Log Avg loss: 0.39072592, Global Avg Loss: 1.50800533, Time: 0.0211 Steps: 31860, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002064, Sample Num: 33024, Cur Loss: 0.33662704, Cur Avg Loss: 0.36290257, Log Avg loss: 0.27999566, Global Avg Loss: 1.50762001, Time: 0.0211 Steps: 31870, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002074, Sample Num: 33184, Cur Loss: 0.26981318, Cur Avg Loss: 0.36235314, Log Avg loss: 0.24895108, Global Avg Loss: 1.50722520, Time: 0.0211 Steps: 31880, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002084, Sample Num: 33344, Cur Loss: 0.23835565, Cur Avg Loss: 0.36169042, Log Avg loss: 0.22424302, Global Avg Loss: 1.50682288, Time: 0.0211 Steps: 31890, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002094, Sample Num: 33504, Cur Loss: 0.23995496, Cur Avg Loss: 0.36141273, Log Avg loss: 0.30354179, Global Avg Loss: 1.50644568, Time: 0.0211 Steps: 31900, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002104, Sample Num: 33664, Cur Loss: 0.08482137, Cur Avg Loss: 0.36063217, Log Avg loss: 0.19718251, Global Avg Loss: 1.50603538, Time: 0.0211 Steps: 31910, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002114, Sample Num: 33824, Cur Loss: 0.28136843, Cur Avg Loss: 0.36048034, Log Avg loss: 0.32853454, Global Avg Loss: 1.50566649, Time: 0.0211 Steps: 31920, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002124, Sample Num: 33984, Cur Loss: 0.13559705, Cur Avg Loss: 0.36047713, Log Avg loss: 0.35979829, Global Avg Loss: 1.50530762, Time: 0.0211 Steps: 31930, Updated lr: 0.000071 ***** Running evaluation checkpoint-31935 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-31935 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.081549, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.341584, "eval_total_loss": 240.133666, "eval_mae": 0.410969, "eval_mse": 0.341701, "eval_r2": 0.782792, "eval_sp_statistic": 0.863632, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.89253, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.482486, "test_total_loss": 242.208169, "test_mae": 0.41654, "test_mse": 0.482663, "test_r2": 0.688485, "test_sp_statistic": 0.849118, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.89592, "test_ps_pvalue": 0.0, "lr": 7.066382171645329e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5051293540618822, "train_cur_epoch_loss": 767.4870202820748, "train_cur_epoch_avg_loss": 0.36049178970506096, "train_cur_epoch_time": 45.08154916763306, "train_cur_epoch_avg_time": 0.021174987866431686, "epoch": 15, "step": 31935} ################################################## Training, Epoch: 0016, Batch: 000005, Sample Num: 80, Cur Loss: 0.44616085, Cur Avg Loss: 0.37874021, Log Avg loss: 0.37273058, Global Avg Loss: 1.50495303, Time: 0.0223 Steps: 31940, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000015, Sample Num: 240, Cur Loss: 0.64071763, Cur Avg Loss: 0.41250656, Log Avg loss: 0.42938974, Global Avg Loss: 1.50461639, Time: 0.0210 Steps: 31950, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000025, Sample Num: 400, Cur Loss: 0.88698065, Cur Avg Loss: 0.41120505, Log Avg loss: 0.40925279, Global Avg Loss: 1.50427366, Time: 0.0210 Steps: 31960, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000035, Sample Num: 560, Cur Loss: 0.26677370, Cur Avg Loss: 0.39903048, Log Avg loss: 0.36859403, Global Avg Loss: 1.50391842, Time: 0.0210 Steps: 31970, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000045, Sample Num: 720, Cur Loss: 0.08963180, Cur Avg Loss: 0.37165942, Log Avg loss: 0.27586071, Global Avg Loss: 1.50353442, Time: 0.0210 Steps: 31980, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000055, Sample Num: 880, Cur Loss: 0.41452441, Cur Avg Loss: 0.36433165, Log Avg loss: 0.33135673, Global Avg Loss: 1.50316800, Time: 0.0211 Steps: 31990, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000065, Sample Num: 1040, Cur Loss: 0.22344676, Cur Avg Loss: 0.36770175, Log Avg loss: 0.38623729, Global Avg Loss: 1.50281895, Time: 0.0210 Steps: 32000, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000075, Sample Num: 1200, Cur Loss: 0.18818974, Cur Avg Loss: 0.36887419, Log Avg loss: 0.37649506, Global Avg Loss: 1.50246709, Time: 0.0210 Steps: 32010, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000085, Sample Num: 1360, Cur Loss: 0.13071713, Cur Avg Loss: 0.36632674, Log Avg loss: 0.34722085, Global Avg Loss: 1.50210630, Time: 0.0211 Steps: 32020, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000095, Sample Num: 1520, Cur Loss: 0.52147532, Cur Avg Loss: 0.36007320, Log Avg loss: 0.30691808, Global Avg Loss: 1.50173315, Time: 0.0210 Steps: 32030, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000105, Sample Num: 1680, Cur Loss: 0.25763744, Cur Avg Loss: 0.35336287, Log Avg loss: 0.28961475, Global Avg Loss: 1.50135484, Time: 0.0210 Steps: 32040, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000115, Sample Num: 1840, Cur Loss: 0.31475806, Cur Avg Loss: 0.34492862, Log Avg loss: 0.25636904, Global Avg Loss: 1.50096639, Time: 0.0211 Steps: 32050, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000125, Sample Num: 2000, Cur Loss: 0.47169703, Cur Avg Loss: 0.34021804, Log Avg loss: 0.28604627, Global Avg Loss: 1.50058744, Time: 0.0210 Steps: 32060, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000135, Sample Num: 2160, Cur Loss: 0.14478937, Cur Avg Loss: 0.34522454, Log Avg loss: 0.40780581, Global Avg Loss: 1.50024669, Time: 0.0210 Steps: 32070, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000145, Sample Num: 2320, Cur Loss: 0.43765727, Cur Avg Loss: 0.33775033, Log Avg loss: 0.23684851, Global Avg Loss: 1.49985286, Time: 0.0210 Steps: 32080, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000155, Sample Num: 2480, Cur Loss: 0.23332158, Cur Avg Loss: 0.33665226, Log Avg loss: 0.32073033, Global Avg Loss: 1.49948542, Time: 0.0210 Steps: 32090, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000165, Sample Num: 2640, Cur Loss: 0.31006983, Cur Avg Loss: 0.33451618, Log Avg loss: 0.30140686, Global Avg Loss: 1.49911218, Time: 0.0210 Steps: 32100, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000175, Sample Num: 2800, Cur Loss: 0.10988301, Cur Avg Loss: 0.33163489, Log Avg loss: 0.28409357, Global Avg Loss: 1.49873379, Time: 0.0210 Steps: 32110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000185, Sample Num: 2960, Cur Loss: 0.40456831, Cur Avg Loss: 0.34771189, Log Avg loss: 0.62905934, Global Avg Loss: 1.49846303, Time: 0.0210 Steps: 32120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000195, Sample Num: 3120, Cur Loss: 0.19602507, Cur Avg Loss: 0.34417084, Log Avg loss: 0.27866144, Global Avg Loss: 1.49808339, Time: 0.0210 Steps: 32130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000205, Sample Num: 3280, Cur Loss: 0.11054203, Cur Avg Loss: 0.34176701, Log Avg loss: 0.29489241, Global Avg Loss: 1.49770903, Time: 0.0210 Steps: 32140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000215, Sample Num: 3440, Cur Loss: 0.10154872, Cur Avg Loss: 0.34695898, Log Avg loss: 0.45339444, Global Avg Loss: 1.49738420, Time: 0.0210 Steps: 32150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000225, Sample Num: 3600, Cur Loss: 0.82785451, Cur Avg Loss: 0.35119513, Log Avg loss: 0.44227236, Global Avg Loss: 1.49705612, Time: 0.0210 Steps: 32160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000235, Sample Num: 3760, Cur Loss: 0.32825765, Cur Avg Loss: 0.35621996, Log Avg loss: 0.46927851, Global Avg Loss: 1.49673664, Time: 0.0211 Steps: 32170, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000245, Sample Num: 3920, Cur Loss: 0.12094459, Cur Avg Loss: 0.35164429, Log Avg loss: 0.24411615, Global Avg Loss: 1.49634738, Time: 0.0210 Steps: 32180, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000255, Sample Num: 4080, Cur Loss: 0.23390830, Cur Avg Loss: 0.35021531, Log Avg loss: 0.31520533, Global Avg Loss: 1.49598045, Time: 0.0210 Steps: 32190, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000265, Sample Num: 4240, Cur Loss: 0.60976696, Cur Avg Loss: 0.34966796, Log Avg loss: 0.33571037, Global Avg Loss: 1.49562012, Time: 0.0211 Steps: 32200, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000275, Sample Num: 4400, Cur Loss: 0.93230647, Cur Avg Loss: 0.35522698, Log Avg loss: 0.50254095, Global Avg Loss: 1.49531181, Time: 0.0210 Steps: 32210, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000285, Sample Num: 4560, Cur Loss: 0.47478288, Cur Avg Loss: 0.35677892, Log Avg loss: 0.39945750, Global Avg Loss: 1.49497169, Time: 0.0210 Steps: 32220, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000295, Sample Num: 4720, Cur Loss: 0.30693141, Cur Avg Loss: 0.35812786, Log Avg loss: 0.39657255, Global Avg Loss: 1.49463089, Time: 0.0210 Steps: 32230, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000305, Sample Num: 4880, Cur Loss: 0.44941056, Cur Avg Loss: 0.35717090, Log Avg loss: 0.32894060, Global Avg Loss: 1.49426933, Time: 0.0211 Steps: 32240, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000315, Sample Num: 5040, Cur Loss: 0.47251958, Cur Avg Loss: 0.35793379, Log Avg loss: 0.38120177, Global Avg Loss: 1.49392419, Time: 0.0210 Steps: 32250, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000325, Sample Num: 5200, Cur Loss: 0.22475965, Cur Avg Loss: 0.35822048, Log Avg loss: 0.36725120, Global Avg Loss: 1.49357494, Time: 0.0210 Steps: 32260, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000335, Sample Num: 5360, Cur Loss: 0.07267900, Cur Avg Loss: 0.35604139, Log Avg loss: 0.28522106, Global Avg Loss: 1.49320049, Time: 0.0210 Steps: 32270, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000345, Sample Num: 5520, Cur Loss: 0.31174123, Cur Avg Loss: 0.35411012, Log Avg loss: 0.28941268, Global Avg Loss: 1.49282757, Time: 0.0211 Steps: 32280, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000355, Sample Num: 5680, Cur Loss: 0.14967743, Cur Avg Loss: 0.35422582, Log Avg loss: 0.35821735, Global Avg Loss: 1.49247619, Time: 0.0210 Steps: 32290, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000365, Sample Num: 5840, Cur Loss: 0.41601700, Cur Avg Loss: 0.35237371, Log Avg loss: 0.28662399, Global Avg Loss: 1.49210286, Time: 0.0210 Steps: 32300, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000375, Sample Num: 6000, Cur Loss: 0.64903426, Cur Avg Loss: 0.34958671, Log Avg loss: 0.24786125, Global Avg Loss: 1.49171776, Time: 0.0211 Steps: 32310, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000385, Sample Num: 6160, Cur Loss: 0.40852356, Cur Avg Loss: 0.35508251, Log Avg loss: 0.56117498, Global Avg Loss: 1.49142985, Time: 0.0210 Steps: 32320, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000395, Sample Num: 6320, Cur Loss: 0.39721680, Cur Avg Loss: 0.35520146, Log Avg loss: 0.35978094, Global Avg Loss: 1.49107982, Time: 0.0211 Steps: 32330, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000405, Sample Num: 6480, Cur Loss: 0.23142526, Cur Avg Loss: 0.35471427, Log Avg loss: 0.33547003, Global Avg Loss: 1.49072249, Time: 0.0211 Steps: 32340, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000415, Sample Num: 6640, Cur Loss: 0.17977150, Cur Avg Loss: 0.35508988, Log Avg loss: 0.37030228, Global Avg Loss: 1.49037614, Time: 0.0211 Steps: 32350, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000425, Sample Num: 6800, Cur Loss: 0.93469501, Cur Avg Loss: 0.35668992, Log Avg loss: 0.42309149, Global Avg Loss: 1.49004633, Time: 0.0212 Steps: 32360, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000435, Sample Num: 6960, Cur Loss: 0.37363017, Cur Avg Loss: 0.35968487, Log Avg loss: 0.48697022, Global Avg Loss: 1.48973645, Time: 0.0211 Steps: 32370, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000445, Sample Num: 7120, Cur Loss: 0.41875452, Cur Avg Loss: 0.36253416, Log Avg loss: 0.48647834, Global Avg Loss: 1.48942661, Time: 0.0210 Steps: 32380, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000455, Sample Num: 7280, Cur Loss: 0.28304514, Cur Avg Loss: 0.36261918, Log Avg loss: 0.36640258, Global Avg Loss: 1.48907989, Time: 0.0211 Steps: 32390, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000465, Sample Num: 7440, Cur Loss: 0.59772193, Cur Avg Loss: 0.36195177, Log Avg loss: 0.33158470, Global Avg Loss: 1.48872264, Time: 0.0211 Steps: 32400, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000475, Sample Num: 7600, Cur Loss: 0.65009463, Cur Avg Loss: 0.36239649, Log Avg loss: 0.38307602, Global Avg Loss: 1.48838150, Time: 0.0210 Steps: 32410, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000485, Sample Num: 7760, Cur Loss: 0.44134945, Cur Avg Loss: 0.36264906, Log Avg loss: 0.37464592, Global Avg Loss: 1.48803796, Time: 0.0211 Steps: 32420, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000495, Sample Num: 7920, Cur Loss: 0.72065586, Cur Avg Loss: 0.36736397, Log Avg loss: 0.59603738, Global Avg Loss: 1.48776291, Time: 0.0211 Steps: 32430, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000505, Sample Num: 8080, Cur Loss: 0.30816805, Cur Avg Loss: 0.36712722, Log Avg loss: 0.35540779, Global Avg Loss: 1.48741385, Time: 0.0210 Steps: 32440, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000515, Sample Num: 8240, Cur Loss: 0.61152852, Cur Avg Loss: 0.36627233, Log Avg loss: 0.32310067, Global Avg Loss: 1.48705504, Time: 0.0249 Steps: 32450, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000525, Sample Num: 8400, Cur Loss: 0.11376013, Cur Avg Loss: 0.36711715, Log Avg loss: 0.41062548, Global Avg Loss: 1.48672343, Time: 0.0212 Steps: 32460, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000535, Sample Num: 8560, Cur Loss: 0.20243452, Cur Avg Loss: 0.36670577, Log Avg loss: 0.34510785, Global Avg Loss: 1.48637184, Time: 0.0211 Steps: 32470, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000545, Sample Num: 8720, Cur Loss: 0.10583623, Cur Avg Loss: 0.36553773, Log Avg loss: 0.30304797, Global Avg Loss: 1.48600751, Time: 0.0211 Steps: 32480, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000555, Sample Num: 8880, Cur Loss: 0.22346097, Cur Avg Loss: 0.36498549, Log Avg loss: 0.33488804, Global Avg Loss: 1.48565321, Time: 0.0212 Steps: 32490, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000565, Sample Num: 9040, Cur Loss: 0.16569307, Cur Avg Loss: 0.36425473, Log Avg loss: 0.32369758, Global Avg Loss: 1.48529569, Time: 0.0212 Steps: 32500, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000575, Sample Num: 9200, Cur Loss: 0.27821362, Cur Avg Loss: 0.36515431, Log Avg loss: 0.41598063, Global Avg Loss: 1.48496677, Time: 0.0212 Steps: 32510, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000585, Sample Num: 9360, Cur Loss: 0.17829704, Cur Avg Loss: 0.36605211, Log Avg loss: 0.41767589, Global Avg Loss: 1.48463857, Time: 0.0212 Steps: 32520, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000595, Sample Num: 9520, Cur Loss: 0.13210401, Cur Avg Loss: 0.36497606, Log Avg loss: 0.30202729, Global Avg Loss: 1.48427503, Time: 0.0212 Steps: 32530, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000605, Sample Num: 9680, Cur Loss: 0.04151619, Cur Avg Loss: 0.36332497, Log Avg loss: 0.26508500, Global Avg Loss: 1.48390035, Time: 0.0212 Steps: 32540, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000615, Sample Num: 9840, Cur Loss: 0.16430065, Cur Avg Loss: 0.36235618, Log Avg loss: 0.30374437, Global Avg Loss: 1.48353779, Time: 0.0212 Steps: 32550, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000625, Sample Num: 10000, Cur Loss: 0.21974213, Cur Avg Loss: 0.36083717, Log Avg loss: 0.26741766, Global Avg Loss: 1.48316429, Time: 0.0212 Steps: 32560, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000635, Sample Num: 10160, Cur Loss: 0.18849143, Cur Avg Loss: 0.35950818, Log Avg loss: 0.27644680, Global Avg Loss: 1.48279379, Time: 0.0211 Steps: 32570, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000645, Sample Num: 10320, Cur Loss: 0.30686224, Cur Avg Loss: 0.35876866, Log Avg loss: 0.31180883, Global Avg Loss: 1.48243437, Time: 0.0212 Steps: 32580, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000655, Sample Num: 10480, Cur Loss: 0.13218546, Cur Avg Loss: 0.35726528, Log Avg loss: 0.26029720, Global Avg Loss: 1.48205936, Time: 0.0211 Steps: 32590, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000665, Sample Num: 10640, Cur Loss: 0.20161930, Cur Avg Loss: 0.35505110, Log Avg loss: 0.21002261, Global Avg Loss: 1.48166917, Time: 0.0211 Steps: 32600, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000675, Sample Num: 10800, Cur Loss: 0.38939017, Cur Avg Loss: 0.35583111, Log Avg loss: 0.40770149, Global Avg Loss: 1.48133983, Time: 0.0212 Steps: 32610, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000685, Sample Num: 10960, Cur Loss: 0.26527047, Cur Avg Loss: 0.35568546, Log Avg loss: 0.34585395, Global Avg Loss: 1.48099174, Time: 0.0212 Steps: 32620, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000695, Sample Num: 11120, Cur Loss: 0.22324322, Cur Avg Loss: 0.35496573, Log Avg loss: 0.30566475, Global Avg Loss: 1.48063154, Time: 0.0212 Steps: 32630, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000705, Sample Num: 11280, Cur Loss: 0.26156345, Cur Avg Loss: 0.35482202, Log Avg loss: 0.34483412, Global Avg Loss: 1.48028356, Time: 0.0212 Steps: 32640, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000715, Sample Num: 11440, Cur Loss: 0.52661717, Cur Avg Loss: 0.35467283, Log Avg loss: 0.34415440, Global Avg Loss: 1.47993559, Time: 0.0212 Steps: 32650, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000725, Sample Num: 11600, Cur Loss: 0.30532765, Cur Avg Loss: 0.35490692, Log Avg loss: 0.37164446, Global Avg Loss: 1.47959625, Time: 0.0212 Steps: 32660, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000735, Sample Num: 11760, Cur Loss: 0.77900887, Cur Avg Loss: 0.35721953, Log Avg loss: 0.52488390, Global Avg Loss: 1.47930402, Time: 0.0211 Steps: 32670, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000745, Sample Num: 11920, Cur Loss: 0.21195689, Cur Avg Loss: 0.35873326, Log Avg loss: 0.46999279, Global Avg Loss: 1.47899517, Time: 0.0212 Steps: 32680, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000755, Sample Num: 12080, Cur Loss: 0.51188570, Cur Avg Loss: 0.35859377, Log Avg loss: 0.34820161, Global Avg Loss: 1.47864926, Time: 0.0211 Steps: 32690, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000765, Sample Num: 12240, Cur Loss: 0.16442697, Cur Avg Loss: 0.35671720, Log Avg loss: 0.21503576, Global Avg Loss: 1.47826283, Time: 0.0212 Steps: 32700, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000775, Sample Num: 12400, Cur Loss: 0.18129846, Cur Avg Loss: 0.35511061, Log Avg loss: 0.23220645, Global Avg Loss: 1.47788189, Time: 0.0212 Steps: 32710, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000785, Sample Num: 12560, Cur Loss: 0.15851624, Cur Avg Loss: 0.35497879, Log Avg loss: 0.34476305, Global Avg Loss: 1.47753558, Time: 0.0212 Steps: 32720, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000795, Sample Num: 12720, Cur Loss: 0.16538122, Cur Avg Loss: 0.35457444, Log Avg loss: 0.32283310, Global Avg Loss: 1.47718279, Time: 0.0212 Steps: 32730, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000805, Sample Num: 12880, Cur Loss: 0.17539541, Cur Avg Loss: 0.35398650, Log Avg loss: 0.30724521, Global Avg Loss: 1.47682544, Time: 0.0211 Steps: 32740, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000815, Sample Num: 13040, Cur Loss: 0.28983837, Cur Avg Loss: 0.35303135, Log Avg loss: 0.27614204, Global Avg Loss: 1.47645882, Time: 0.0212 Steps: 32750, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000825, Sample Num: 13200, Cur Loss: 0.52996445, Cur Avg Loss: 0.35271473, Log Avg loss: 0.32690960, Global Avg Loss: 1.47610792, Time: 0.0212 Steps: 32760, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000835, Sample Num: 13360, Cur Loss: 0.35063034, Cur Avg Loss: 0.35282888, Log Avg loss: 0.36224613, Global Avg Loss: 1.47576802, Time: 0.0211 Steps: 32770, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000845, Sample Num: 13520, Cur Loss: 0.20190758, Cur Avg Loss: 0.35155530, Log Avg loss: 0.24521135, Global Avg Loss: 1.47539262, Time: 0.0211 Steps: 32780, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000855, Sample Num: 13680, Cur Loss: 0.54775649, Cur Avg Loss: 0.35169194, Log Avg loss: 0.36323832, Global Avg Loss: 1.47505345, Time: 0.0212 Steps: 32790, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000865, Sample Num: 13840, Cur Loss: 0.50805193, Cur Avg Loss: 0.35350417, Log Avg loss: 0.50845022, Global Avg Loss: 1.47475875, Time: 0.0212 Steps: 32800, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000875, Sample Num: 14000, Cur Loss: 0.21370089, Cur Avg Loss: 0.35394652, Log Avg loss: 0.39220950, Global Avg Loss: 1.47442881, Time: 0.0213 Steps: 32810, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000885, Sample Num: 14160, Cur Loss: 0.33882982, Cur Avg Loss: 0.35370153, Log Avg loss: 0.33226506, Global Avg Loss: 1.47408080, Time: 0.0212 Steps: 32820, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000895, Sample Num: 14320, Cur Loss: 0.15851538, Cur Avg Loss: 0.35359995, Log Avg loss: 0.34461019, Global Avg Loss: 1.47373676, Time: 0.0211 Steps: 32830, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000905, Sample Num: 14480, Cur Loss: 0.30341232, Cur Avg Loss: 0.35303296, Log Avg loss: 0.30228727, Global Avg Loss: 1.47338005, Time: 0.0211 Steps: 32840, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000915, Sample Num: 14640, Cur Loss: 0.56783956, Cur Avg Loss: 0.35454164, Log Avg loss: 0.49107696, Global Avg Loss: 1.47308102, Time: 0.0211 Steps: 32850, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000925, Sample Num: 14800, Cur Loss: 0.87428415, Cur Avg Loss: 0.35507114, Log Avg loss: 0.40352056, Global Avg Loss: 1.47275553, Time: 0.0212 Steps: 32860, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000935, Sample Num: 14960, Cur Loss: 0.60510498, Cur Avg Loss: 0.35574605, Log Avg loss: 0.41817542, Global Avg Loss: 1.47243470, Time: 0.0212 Steps: 32870, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000945, Sample Num: 15120, Cur Loss: 0.14464280, Cur Avg Loss: 0.35483906, Log Avg loss: 0.27003521, Global Avg Loss: 1.47206900, Time: 0.0212 Steps: 32880, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000955, Sample Num: 15280, Cur Loss: 0.11689661, Cur Avg Loss: 0.35547640, Log Avg loss: 0.41570480, Global Avg Loss: 1.47174782, Time: 0.0212 Steps: 32890, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000965, Sample Num: 15440, Cur Loss: 0.50634336, Cur Avg Loss: 0.35486634, Log Avg loss: 0.29660595, Global Avg Loss: 1.47139064, Time: 0.0212 Steps: 32900, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000975, Sample Num: 15600, Cur Loss: 0.21791624, Cur Avg Loss: 0.35439236, Log Avg loss: 0.30865318, Global Avg Loss: 1.47103733, Time: 0.0211 Steps: 32910, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000985, Sample Num: 15760, Cur Loss: 0.37762558, Cur Avg Loss: 0.35413054, Log Avg loss: 0.32860281, Global Avg Loss: 1.47069029, Time: 0.0212 Steps: 32920, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000995, Sample Num: 15920, Cur Loss: 0.10631216, Cur Avg Loss: 0.35332339, Log Avg loss: 0.27381949, Global Avg Loss: 1.47032684, Time: 0.0212 Steps: 32930, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001005, Sample Num: 16080, Cur Loss: 0.26988316, Cur Avg Loss: 0.35313412, Log Avg loss: 0.33430156, Global Avg Loss: 1.46998196, Time: 0.0211 Steps: 32940, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001015, Sample Num: 16240, Cur Loss: 0.22819655, Cur Avg Loss: 0.35326499, Log Avg loss: 0.36641757, Global Avg Loss: 1.46964704, Time: 0.0211 Steps: 32950, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001025, Sample Num: 16400, Cur Loss: 0.17008251, Cur Avg Loss: 0.35289836, Log Avg loss: 0.31568570, Global Avg Loss: 1.46929693, Time: 0.0247 Steps: 32960, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001035, Sample Num: 16560, Cur Loss: 0.37175938, Cur Avg Loss: 0.35260700, Log Avg loss: 0.32274180, Global Avg Loss: 1.46894917, Time: 0.0211 Steps: 32970, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001045, Sample Num: 16720, Cur Loss: 0.81133604, Cur Avg Loss: 0.35286939, Log Avg loss: 0.38002751, Global Avg Loss: 1.46861899, Time: 0.0211 Steps: 32980, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001055, Sample Num: 16880, Cur Loss: 0.44259527, Cur Avg Loss: 0.35384822, Log Avg loss: 0.45613617, Global Avg Loss: 1.46831209, Time: 0.0210 Steps: 32990, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001065, Sample Num: 17040, Cur Loss: 0.20912012, Cur Avg Loss: 0.35368943, Log Avg loss: 0.33693701, Global Avg Loss: 1.46796925, Time: 0.0210 Steps: 33000, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001075, Sample Num: 17200, Cur Loss: 0.35942501, Cur Avg Loss: 0.35331253, Log Avg loss: 0.31317235, Global Avg Loss: 1.46761942, Time: 0.0211 Steps: 33010, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001085, Sample Num: 17360, Cur Loss: 0.97519362, Cur Avg Loss: 0.35394262, Log Avg loss: 0.42167733, Global Avg Loss: 1.46730265, Time: 0.0211 Steps: 33020, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001095, Sample Num: 17520, Cur Loss: 0.24005361, Cur Avg Loss: 0.35418511, Log Avg loss: 0.38049510, Global Avg Loss: 1.46697362, Time: 0.0211 Steps: 33030, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001105, Sample Num: 17680, Cur Loss: 0.60843456, Cur Avg Loss: 0.35460048, Log Avg loss: 0.40008350, Global Avg Loss: 1.46665071, Time: 0.0211 Steps: 33040, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001115, Sample Num: 17840, Cur Loss: 0.62657022, Cur Avg Loss: 0.35756524, Log Avg loss: 0.68517070, Global Avg Loss: 1.46641426, Time: 0.0210 Steps: 33050, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001125, Sample Num: 18000, Cur Loss: 0.16442519, Cur Avg Loss: 0.36165795, Log Avg loss: 0.81799580, Global Avg Loss: 1.46621812, Time: 0.0210 Steps: 33060, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001135, Sample Num: 18160, Cur Loss: 0.32595927, Cur Avg Loss: 0.36241070, Log Avg loss: 0.44709493, Global Avg Loss: 1.46590995, Time: 0.0213 Steps: 33070, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001145, Sample Num: 18320, Cur Loss: 0.52998281, Cur Avg Loss: 0.36192050, Log Avg loss: 0.30628328, Global Avg Loss: 1.46555940, Time: 0.0211 Steps: 33080, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001155, Sample Num: 18480, Cur Loss: 0.21082920, Cur Avg Loss: 0.36161279, Log Avg loss: 0.32637946, Global Avg Loss: 1.46521513, Time: 0.0210 Steps: 33090, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001165, Sample Num: 18640, Cur Loss: 0.61662632, Cur Avg Loss: 0.36102513, Log Avg loss: 0.29315104, Global Avg Loss: 1.46486103, Time: 0.0212 Steps: 33100, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001175, Sample Num: 18800, Cur Loss: 0.16133845, Cur Avg Loss: 0.36056944, Log Avg loss: 0.30748079, Global Avg Loss: 1.46451148, Time: 0.0211 Steps: 33110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001185, Sample Num: 18960, Cur Loss: 0.14601065, Cur Avg Loss: 0.36062320, Log Avg loss: 0.36693962, Global Avg Loss: 1.46418008, Time: 0.0212 Steps: 33120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001195, Sample Num: 19120, Cur Loss: 0.22522366, Cur Avg Loss: 0.36024584, Log Avg loss: 0.31552941, Global Avg Loss: 1.46383337, Time: 0.0211 Steps: 33130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001205, Sample Num: 19280, Cur Loss: 0.24543813, Cur Avg Loss: 0.36003652, Log Avg loss: 0.33502212, Global Avg Loss: 1.46349276, Time: 0.0212 Steps: 33140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001215, Sample Num: 19440, Cur Loss: 0.27643636, Cur Avg Loss: 0.35920919, Log Avg loss: 0.25951646, Global Avg Loss: 1.46312957, Time: 0.0211 Steps: 33150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001225, Sample Num: 19600, Cur Loss: 0.43065274, Cur Avg Loss: 0.35907445, Log Avg loss: 0.34270341, Global Avg Loss: 1.46279168, Time: 0.0210 Steps: 33160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001235, Sample Num: 19760, Cur Loss: 0.57475889, Cur Avg Loss: 0.35964272, Log Avg loss: 0.42925599, Global Avg Loss: 1.46248009, Time: 0.0211 Steps: 33170, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001245, Sample Num: 19920, Cur Loss: 0.73819762, Cur Avg Loss: 0.36066783, Log Avg loss: 0.48726897, Global Avg Loss: 1.46218618, Time: 0.0211 Steps: 33180, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001255, Sample Num: 20080, Cur Loss: 0.36541483, Cur Avg Loss: 0.36060299, Log Avg loss: 0.35253048, Global Avg Loss: 1.46185184, Time: 0.0210 Steps: 33190, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001265, Sample Num: 20240, Cur Loss: 0.36017990, Cur Avg Loss: 0.36183254, Log Avg loss: 0.51614063, Global Avg Loss: 1.46156699, Time: 0.0211 Steps: 33200, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001275, Sample Num: 20400, Cur Loss: 1.09861982, Cur Avg Loss: 0.36277081, Log Avg loss: 0.48146254, Global Avg Loss: 1.46127187, Time: 0.0211 Steps: 33210, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001285, Sample Num: 20560, Cur Loss: 0.07241301, Cur Avg Loss: 0.36417701, Log Avg loss: 0.54346657, Global Avg Loss: 1.46099559, Time: 0.0221 Steps: 33220, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001295, Sample Num: 20720, Cur Loss: 0.29509833, Cur Avg Loss: 0.36596356, Log Avg loss: 0.59553610, Global Avg Loss: 1.46073514, Time: 0.0210 Steps: 33230, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001305, Sample Num: 20880, Cur Loss: 0.83011693, Cur Avg Loss: 0.36579778, Log Avg loss: 0.34432948, Global Avg Loss: 1.46039928, Time: 0.0210 Steps: 33240, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001315, Sample Num: 21040, Cur Loss: 0.25731042, Cur Avg Loss: 0.36584384, Log Avg loss: 0.37185465, Global Avg Loss: 1.46007190, Time: 0.0210 Steps: 33250, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001325, Sample Num: 21200, Cur Loss: 0.15686309, Cur Avg Loss: 0.36540238, Log Avg loss: 0.30734970, Global Avg Loss: 1.45972532, Time: 0.0211 Steps: 33260, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001335, Sample Num: 21360, Cur Loss: 1.00617719, Cur Avg Loss: 0.36529735, Log Avg loss: 0.35138136, Global Avg Loss: 1.45939218, Time: 0.0210 Steps: 33270, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001345, Sample Num: 21520, Cur Loss: 0.30629843, Cur Avg Loss: 0.36517086, Log Avg loss: 0.34828381, Global Avg Loss: 1.45905832, Time: 0.0210 Steps: 33280, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001355, Sample Num: 21680, Cur Loss: 0.98178899, Cur Avg Loss: 0.36570509, Log Avg loss: 0.43755980, Global Avg Loss: 1.45875147, Time: 0.0210 Steps: 33290, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001365, Sample Num: 21840, Cur Loss: 0.31717330, Cur Avg Loss: 0.36669114, Log Avg loss: 0.50030064, Global Avg Loss: 1.45846364, Time: 0.0211 Steps: 33300, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001375, Sample Num: 22000, Cur Loss: 1.45898843, Cur Avg Loss: 0.37055452, Log Avg loss: 0.89790575, Global Avg Loss: 1.45829536, Time: 0.0210 Steps: 33310, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001385, Sample Num: 22160, Cur Loss: 0.32705900, Cur Avg Loss: 0.37076261, Log Avg loss: 0.39937519, Global Avg Loss: 1.45797756, Time: 0.0211 Steps: 33320, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001395, Sample Num: 22320, Cur Loss: 0.26272509, Cur Avg Loss: 0.37026542, Log Avg loss: 0.30140460, Global Avg Loss: 1.45763055, Time: 0.0210 Steps: 33330, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001405, Sample Num: 22480, Cur Loss: 0.60788268, Cur Avg Loss: 0.37017616, Log Avg loss: 0.35772442, Global Avg Loss: 1.45730064, Time: 0.0210 Steps: 33340, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001415, Sample Num: 22640, Cur Loss: 0.77990878, Cur Avg Loss: 0.36984877, Log Avg loss: 0.32385006, Global Avg Loss: 1.45696078, Time: 0.0211 Steps: 33350, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001425, Sample Num: 22800, Cur Loss: 0.67376089, Cur Avg Loss: 0.37039011, Log Avg loss: 0.44698989, Global Avg Loss: 1.45665803, Time: 0.0210 Steps: 33360, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001435, Sample Num: 22960, Cur Loss: 0.20369889, Cur Avg Loss: 0.36997426, Log Avg loss: 0.31071567, Global Avg Loss: 1.45631462, Time: 0.0210 Steps: 33370, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001445, Sample Num: 23120, Cur Loss: 0.16449173, Cur Avg Loss: 0.36962516, Log Avg loss: 0.31952859, Global Avg Loss: 1.45597406, Time: 0.0210 Steps: 33380, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001455, Sample Num: 23280, Cur Loss: 0.20033512, Cur Avg Loss: 0.36914565, Log Avg loss: 0.29985752, Global Avg Loss: 1.45562782, Time: 0.0210 Steps: 33390, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001465, Sample Num: 23440, Cur Loss: 0.30236003, Cur Avg Loss: 0.36930811, Log Avg loss: 0.39294517, Global Avg Loss: 1.45530965, Time: 0.0210 Steps: 33400, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001475, Sample Num: 23600, Cur Loss: 0.52338535, Cur Avg Loss: 0.36927284, Log Avg loss: 0.36410701, Global Avg Loss: 1.45498304, Time: 0.0210 Steps: 33410, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001485, Sample Num: 23760, Cur Loss: 0.40891021, Cur Avg Loss: 0.36873479, Log Avg loss: 0.28937238, Global Avg Loss: 1.45463426, Time: 0.0210 Steps: 33420, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001495, Sample Num: 23920, Cur Loss: 0.32634935, Cur Avg Loss: 0.36818176, Log Avg loss: 0.28605541, Global Avg Loss: 1.45428470, Time: 0.0210 Steps: 33430, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001505, Sample Num: 24080, Cur Loss: 0.34778419, Cur Avg Loss: 0.36770974, Log Avg loss: 0.29714313, Global Avg Loss: 1.45393867, Time: 0.0211 Steps: 33440, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001515, Sample Num: 24240, Cur Loss: 0.27864522, Cur Avg Loss: 0.36746508, Log Avg loss: 0.33064483, Global Avg Loss: 1.45360286, Time: 0.0210 Steps: 33450, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001525, Sample Num: 24400, Cur Loss: 0.30055135, Cur Avg Loss: 0.36650323, Log Avg loss: 0.22078289, Global Avg Loss: 1.45323441, Time: 0.0211 Steps: 33460, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001535, Sample Num: 24560, Cur Loss: 0.45635477, Cur Avg Loss: 0.36680347, Log Avg loss: 0.41258978, Global Avg Loss: 1.45292349, Time: 0.0211 Steps: 33470, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001545, Sample Num: 24720, Cur Loss: 0.08569771, Cur Avg Loss: 0.36624872, Log Avg loss: 0.28109484, Global Avg Loss: 1.45257348, Time: 0.0212 Steps: 33480, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001555, Sample Num: 24880, Cur Loss: 0.12915778, Cur Avg Loss: 0.36676389, Log Avg loss: 0.44635712, Global Avg Loss: 1.45227303, Time: 0.0211 Steps: 33490, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001565, Sample Num: 25040, Cur Loss: 0.26172343, Cur Avg Loss: 0.36677436, Log Avg loss: 0.36840261, Global Avg Loss: 1.45194949, Time: 0.0211 Steps: 33500, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001575, Sample Num: 25200, Cur Loss: 0.36338329, Cur Avg Loss: 0.36653212, Log Avg loss: 0.32862055, Global Avg Loss: 1.45161426, Time: 0.0211 Steps: 33510, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001585, Sample Num: 25360, Cur Loss: 0.15219933, Cur Avg Loss: 0.36593687, Log Avg loss: 0.27218585, Global Avg Loss: 1.45126241, Time: 0.0211 Steps: 33520, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001595, Sample Num: 25520, Cur Loss: 0.24685282, Cur Avg Loss: 0.36544312, Log Avg loss: 0.28718421, Global Avg Loss: 1.45091523, Time: 0.0210 Steps: 33530, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001605, Sample Num: 25680, Cur Loss: 0.40412739, Cur Avg Loss: 0.36677082, Log Avg loss: 0.57853801, Global Avg Loss: 1.45065513, Time: 0.0211 Steps: 33540, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001615, Sample Num: 25840, Cur Loss: 0.51429302, Cur Avg Loss: 0.36691845, Log Avg loss: 0.39061302, Global Avg Loss: 1.45033917, Time: 0.0211 Steps: 33550, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001625, Sample Num: 26000, Cur Loss: 0.68073678, Cur Avg Loss: 0.36695968, Log Avg loss: 0.37361907, Global Avg Loss: 1.45001834, Time: 0.0210 Steps: 33560, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001635, Sample Num: 26160, Cur Loss: 0.32229775, Cur Avg Loss: 0.36592107, Log Avg loss: 0.19714591, Global Avg Loss: 1.44964513, Time: 0.0210 Steps: 33570, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001645, Sample Num: 26320, Cur Loss: 0.66414881, Cur Avg Loss: 0.36639967, Log Avg loss: 0.44465124, Global Avg Loss: 1.44934584, Time: 0.0210 Steps: 33580, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001655, Sample Num: 26480, Cur Loss: 0.14028841, Cur Avg Loss: 0.36708104, Log Avg loss: 0.47916632, Global Avg Loss: 1.44905701, Time: 0.0210 Steps: 33590, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001665, Sample Num: 26640, Cur Loss: 0.83767247, Cur Avg Loss: 0.36736368, Log Avg loss: 0.41414022, Global Avg Loss: 1.44874900, Time: 0.0211 Steps: 33600, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001675, Sample Num: 26800, Cur Loss: 0.39148021, Cur Avg Loss: 0.36873622, Log Avg loss: 0.59726501, Global Avg Loss: 1.44849566, Time: 0.0211 Steps: 33610, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001685, Sample Num: 26960, Cur Loss: 0.21787521, Cur Avg Loss: 0.36846003, Log Avg loss: 0.32219888, Global Avg Loss: 1.44816065, Time: 0.0211 Steps: 33620, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001695, Sample Num: 27120, Cur Loss: 0.76706529, Cur Avg Loss: 0.36793832, Log Avg loss: 0.28002930, Global Avg Loss: 1.44781330, Time: 0.0210 Steps: 33630, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001705, Sample Num: 27280, Cur Loss: 0.66043770, Cur Avg Loss: 0.36906587, Log Avg loss: 0.56018649, Global Avg Loss: 1.44754944, Time: 0.0210 Steps: 33640, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001715, Sample Num: 27440, Cur Loss: 0.57013226, Cur Avg Loss: 0.36939055, Log Avg loss: 0.42474813, Global Avg Loss: 1.44724549, Time: 0.0210 Steps: 33650, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001725, Sample Num: 27600, Cur Loss: 0.13399342, Cur Avg Loss: 0.36867829, Log Avg loss: 0.24652583, Global Avg Loss: 1.44688877, Time: 0.0211 Steps: 33660, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001735, Sample Num: 27760, Cur Loss: 0.24647246, Cur Avg Loss: 0.36812937, Log Avg loss: 0.27343932, Global Avg Loss: 1.44654025, Time: 0.0211 Steps: 33670, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001745, Sample Num: 27920, Cur Loss: 0.39368218, Cur Avg Loss: 0.36792371, Log Avg loss: 0.33224286, Global Avg Loss: 1.44620941, Time: 0.0210 Steps: 33680, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001755, Sample Num: 28080, Cur Loss: 0.22833914, Cur Avg Loss: 0.36719513, Log Avg loss: 0.24005715, Global Avg Loss: 1.44585139, Time: 0.0210 Steps: 33690, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001765, Sample Num: 28240, Cur Loss: 0.25512189, Cur Avg Loss: 0.36702595, Log Avg loss: 0.33733553, Global Avg Loss: 1.44552245, Time: 0.0211 Steps: 33700, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001775, Sample Num: 28400, Cur Loss: 0.44687650, Cur Avg Loss: 0.36698507, Log Avg loss: 0.35976901, Global Avg Loss: 1.44520037, Time: 0.0211 Steps: 33710, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001785, Sample Num: 28560, Cur Loss: 0.43859875, Cur Avg Loss: 0.36781655, Log Avg loss: 0.51540444, Global Avg Loss: 1.44492463, Time: 0.0211 Steps: 33720, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001795, Sample Num: 28720, Cur Loss: 0.48699704, Cur Avg Loss: 0.36810131, Log Avg loss: 0.41893168, Global Avg Loss: 1.44462045, Time: 0.0248 Steps: 33730, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001805, Sample Num: 28880, Cur Loss: 0.30593446, Cur Avg Loss: 0.36806725, Log Avg loss: 0.36195348, Global Avg Loss: 1.44429956, Time: 0.0210 Steps: 33740, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001815, Sample Num: 29040, Cur Loss: 0.46572885, Cur Avg Loss: 0.36963843, Log Avg loss: 0.65323660, Global Avg Loss: 1.44406518, Time: 0.0211 Steps: 33750, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001825, Sample Num: 29200, Cur Loss: 0.59430146, Cur Avg Loss: 0.36950293, Log Avg loss: 0.34490963, Global Avg Loss: 1.44373960, Time: 0.0210 Steps: 33760, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001835, Sample Num: 29360, Cur Loss: 0.50524235, Cur Avg Loss: 0.36913854, Log Avg loss: 0.30263705, Global Avg Loss: 1.44340169, Time: 0.0210 Steps: 33770, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001845, Sample Num: 29520, Cur Loss: 0.30701452, Cur Avg Loss: 0.36842148, Log Avg loss: 0.23684011, Global Avg Loss: 1.44304451, Time: 0.0210 Steps: 33780, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001855, Sample Num: 29680, Cur Loss: 0.33709493, Cur Avg Loss: 0.36803794, Log Avg loss: 0.29727489, Global Avg Loss: 1.44270542, Time: 0.0210 Steps: 33790, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001865, Sample Num: 29840, Cur Loss: 0.22232413, Cur Avg Loss: 0.36796150, Log Avg loss: 0.35378323, Global Avg Loss: 1.44238326, Time: 0.0211 Steps: 33800, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001875, Sample Num: 30000, Cur Loss: 0.16366605, Cur Avg Loss: 0.36765921, Log Avg loss: 0.31128135, Global Avg Loss: 1.44204871, Time: 0.0211 Steps: 33810, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001885, Sample Num: 30160, Cur Loss: 0.13445470, Cur Avg Loss: 0.36748452, Log Avg loss: 0.33473061, Global Avg Loss: 1.44172130, Time: 0.0211 Steps: 33820, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001895, Sample Num: 30320, Cur Loss: 0.20079416, Cur Avg Loss: 0.36756386, Log Avg loss: 0.38251888, Global Avg Loss: 1.44140820, Time: 0.0210 Steps: 33830, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001905, Sample Num: 30480, Cur Loss: 0.44305867, Cur Avg Loss: 0.36730887, Log Avg loss: 0.31898836, Global Avg Loss: 1.44107652, Time: 0.0210 Steps: 33840, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001915, Sample Num: 30640, Cur Loss: 0.42312521, Cur Avg Loss: 0.36684160, Log Avg loss: 0.27782570, Global Avg Loss: 1.44073287, Time: 0.0210 Steps: 33850, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001925, Sample Num: 30800, Cur Loss: 0.80779129, Cur Avg Loss: 0.36746151, Log Avg loss: 0.48617514, Global Avg Loss: 1.44045095, Time: 0.0210 Steps: 33860, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001935, Sample Num: 30960, Cur Loss: 0.26206747, Cur Avg Loss: 0.36811874, Log Avg loss: 0.49463604, Global Avg Loss: 1.44017171, Time: 0.0210 Steps: 33870, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001945, Sample Num: 31120, Cur Loss: 0.35506332, Cur Avg Loss: 0.36825503, Log Avg loss: 0.39462703, Global Avg Loss: 1.43986310, Time: 0.0211 Steps: 33880, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001955, Sample Num: 31280, Cur Loss: 0.68671727, Cur Avg Loss: 0.36803721, Log Avg loss: 0.32567177, Global Avg Loss: 1.43953434, Time: 0.0210 Steps: 33890, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001965, Sample Num: 31440, Cur Loss: 0.35982776, Cur Avg Loss: 0.36915943, Log Avg loss: 0.58855283, Global Avg Loss: 1.43928331, Time: 0.0211 Steps: 33900, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001975, Sample Num: 31600, Cur Loss: 0.08719787, Cur Avg Loss: 0.36931744, Log Avg loss: 0.40036702, Global Avg Loss: 1.43897694, Time: 0.0211 Steps: 33910, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001985, Sample Num: 31760, Cur Loss: 0.22885185, Cur Avg Loss: 0.36866788, Log Avg loss: 0.24037888, Global Avg Loss: 1.43862358, Time: 0.0210 Steps: 33920, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001995, Sample Num: 31920, Cur Loss: 0.43903524, Cur Avg Loss: 0.36886116, Log Avg loss: 0.40722679, Global Avg Loss: 1.43831960, Time: 0.0210 Steps: 33930, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002005, Sample Num: 32080, Cur Loss: 0.29375571, Cur Avg Loss: 0.36903994, Log Avg loss: 0.40470782, Global Avg Loss: 1.43801506, Time: 0.0210 Steps: 33940, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002015, Sample Num: 32240, Cur Loss: 0.41044447, Cur Avg Loss: 0.36908926, Log Avg loss: 0.37897633, Global Avg Loss: 1.43770312, Time: 0.0210 Steps: 33950, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002025, Sample Num: 32400, Cur Loss: 0.28050053, Cur Avg Loss: 0.37066451, Log Avg loss: 0.68807739, Global Avg Loss: 1.43748238, Time: 0.0210 Steps: 33960, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002035, Sample Num: 32560, Cur Loss: 0.42373675, Cur Avg Loss: 0.37077117, Log Avg loss: 0.39237087, Global Avg Loss: 1.43717472, Time: 0.0210 Steps: 33970, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002045, Sample Num: 32720, Cur Loss: 0.46418577, Cur Avg Loss: 0.37031462, Log Avg loss: 0.27740710, Global Avg Loss: 1.43683341, Time: 0.0210 Steps: 33980, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002055, Sample Num: 32880, Cur Loss: 0.19863746, Cur Avg Loss: 0.37022155, Log Avg loss: 0.35118736, Global Avg Loss: 1.43651401, Time: 0.0212 Steps: 33990, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002065, Sample Num: 33040, Cur Loss: 0.35198843, Cur Avg Loss: 0.36953452, Log Avg loss: 0.22834967, Global Avg Loss: 1.43615867, Time: 0.0210 Steps: 34000, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002075, Sample Num: 33200, Cur Loss: 0.07560950, Cur Avg Loss: 0.37014460, Log Avg loss: 0.49612603, Global Avg Loss: 1.43588227, Time: 0.0210 Steps: 34010, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002085, Sample Num: 33360, Cur Loss: 0.18749310, Cur Avg Loss: 0.36971243, Log Avg loss: 0.28003831, Global Avg Loss: 1.43554251, Time: 0.0210 Steps: 34020, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002095, Sample Num: 33520, Cur Loss: 0.17560333, Cur Avg Loss: 0.36896088, Log Avg loss: 0.21226276, Global Avg Loss: 1.43518304, Time: 0.0210 Steps: 34030, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002105, Sample Num: 33680, Cur Loss: 0.36235467, Cur Avg Loss: 0.36887445, Log Avg loss: 0.35076812, Global Avg Loss: 1.43486447, Time: 0.0210 Steps: 34040, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002115, Sample Num: 33840, Cur Loss: 0.28610358, Cur Avg Loss: 0.36852012, Log Avg loss: 0.29393337, Global Avg Loss: 1.43452940, Time: 0.0211 Steps: 34050, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002125, Sample Num: 34000, Cur Loss: 0.66382837, Cur Avg Loss: 0.36803334, Log Avg loss: 0.26507877, Global Avg Loss: 1.43418605, Time: 0.0210 Steps: 34060, Updated lr: 0.000069 ***** Running evaluation checkpoint-34064 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-34064 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.156960, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.87761, "eval_total_loss": 616.959661, "eval_mae": 0.831733, "eval_mse": 0.877675, "eval_r2": 0.442092, "eval_sp_statistic": 0.864154, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.90245, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.432172, "test_total_loss": 216.95038, "test_mae": 0.563182, "test_mse": 0.432288, "test_r2": 0.720997, "test_sp_statistic": 0.850515, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.903165, "test_ps_pvalue": 0.0, "lr": 6.86448553816975e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4340620912916493, "train_cur_epoch_loss": 783.5851557925344, "train_cur_epoch_avg_loss": 0.3680531497381561, "train_cur_epoch_time": 45.156959533691406, "train_cur_epoch_avg_time": 0.02121040842352814, "epoch": 16, "step": 34064} ################################################## Training, Epoch: 0017, Batch: 000006, Sample Num: 96, Cur Loss: 0.18362136, Cur Avg Loss: 0.41425312, Log Avg loss: 0.39998254, Global Avg Loss: 1.43388249, Time: 0.0247 Steps: 34070, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000016, Sample Num: 256, Cur Loss: 0.20819336, Cur Avg Loss: 0.26741834, Log Avg loss: 0.17931747, Global Avg Loss: 1.43351437, Time: 0.0210 Steps: 34080, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000026, Sample Num: 416, Cur Loss: 0.14813995, Cur Avg Loss: 0.28277280, Log Avg loss: 0.30733995, Global Avg Loss: 1.43318402, Time: 0.0210 Steps: 34090, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000036, Sample Num: 576, Cur Loss: 0.39059949, Cur Avg Loss: 0.28492266, Log Avg loss: 0.29051230, Global Avg Loss: 1.43284892, Time: 0.0210 Steps: 34100, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000046, Sample Num: 736, Cur Loss: 0.30508587, Cur Avg Loss: 0.28544679, Log Avg loss: 0.28733366, Global Avg Loss: 1.43251309, Time: 0.0211 Steps: 34110, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000056, Sample Num: 896, Cur Loss: 0.31559527, Cur Avg Loss: 0.32422187, Log Avg loss: 0.50258722, Global Avg Loss: 1.43224055, Time: 0.0210 Steps: 34120, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000066, Sample Num: 1056, Cur Loss: 0.26136044, Cur Avg Loss: 0.32106385, Log Avg loss: 0.30337894, Global Avg Loss: 1.43190979, Time: 0.0209 Steps: 34130, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000076, Sample Num: 1216, Cur Loss: 1.25503504, Cur Avg Loss: 0.35241934, Log Avg loss: 0.55936555, Global Avg Loss: 1.43165422, Time: 0.0210 Steps: 34140, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000086, Sample Num: 1376, Cur Loss: 0.77810645, Cur Avg Loss: 0.37638036, Log Avg loss: 0.55848418, Global Avg Loss: 1.43139853, Time: 0.0210 Steps: 34150, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000096, Sample Num: 1536, Cur Loss: 0.74182951, Cur Avg Loss: 0.38482409, Log Avg loss: 0.45744011, Global Avg Loss: 1.43111341, Time: 0.0210 Steps: 34160, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000106, Sample Num: 1696, Cur Loss: 0.07544835, Cur Avg Loss: 0.37165694, Log Avg loss: 0.24525232, Global Avg Loss: 1.43076637, Time: 0.0210 Steps: 34170, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000116, Sample Num: 1856, Cur Loss: 0.09486692, Cur Avg Loss: 0.36205396, Log Avg loss: 0.26026239, Global Avg Loss: 1.43042391, Time: 0.0210 Steps: 34180, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000126, Sample Num: 2016, Cur Loss: 0.16208848, Cur Avg Loss: 0.36409737, Log Avg loss: 0.38780096, Global Avg Loss: 1.43011896, Time: 0.0210 Steps: 34190, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000136, Sample Num: 2176, Cur Loss: 0.19988132, Cur Avg Loss: 0.35631799, Log Avg loss: 0.25829770, Global Avg Loss: 1.42977633, Time: 0.0209 Steps: 34200, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000146, Sample Num: 2336, Cur Loss: 0.16257048, Cur Avg Loss: 0.34819111, Log Avg loss: 0.23766555, Global Avg Loss: 1.42942786, Time: 0.0210 Steps: 34210, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000156, Sample Num: 2496, Cur Loss: 0.43772453, Cur Avg Loss: 0.34370113, Log Avg loss: 0.27814748, Global Avg Loss: 1.42909142, Time: 0.0210 Steps: 34220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000166, Sample Num: 2656, Cur Loss: 0.31558788, Cur Avg Loss: 0.34151482, Log Avg loss: 0.30740832, Global Avg Loss: 1.42876373, Time: 0.0210 Steps: 34230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000176, Sample Num: 2816, Cur Loss: 0.35586634, Cur Avg Loss: 0.33909771, Log Avg loss: 0.29897368, Global Avg Loss: 1.42843377, Time: 0.0209 Steps: 34240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000186, Sample Num: 2976, Cur Loss: 0.71423864, Cur Avg Loss: 0.33963625, Log Avg loss: 0.34911459, Global Avg Loss: 1.42811864, Time: 0.0210 Steps: 34250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000196, Sample Num: 3136, Cur Loss: 0.14066154, Cur Avg Loss: 0.33698035, Log Avg loss: 0.28758067, Global Avg Loss: 1.42778573, Time: 0.0210 Steps: 34260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000206, Sample Num: 3296, Cur Loss: 0.33745414, Cur Avg Loss: 0.33524363, Log Avg loss: 0.30120385, Global Avg Loss: 1.42745700, Time: 0.0210 Steps: 34270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000216, Sample Num: 3456, Cur Loss: 0.27789861, Cur Avg Loss: 0.33346560, Log Avg loss: 0.29683819, Global Avg Loss: 1.42712718, Time: 0.0210 Steps: 34280, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000226, Sample Num: 3616, Cur Loss: 0.39441657, Cur Avg Loss: 0.33909563, Log Avg loss: 0.46070439, Global Avg Loss: 1.42684534, Time: 0.0209 Steps: 34290, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000236, Sample Num: 3776, Cur Loss: 0.52486455, Cur Avg Loss: 0.33979488, Log Avg loss: 0.35559780, Global Avg Loss: 1.42653302, Time: 0.0210 Steps: 34300, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000246, Sample Num: 3936, Cur Loss: 0.08760096, Cur Avg Loss: 0.33898268, Log Avg loss: 0.31981489, Global Avg Loss: 1.42621046, Time: 0.0209 Steps: 34310, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000256, Sample Num: 4096, Cur Loss: 0.57500732, Cur Avg Loss: 0.33738568, Log Avg loss: 0.29809927, Global Avg Loss: 1.42588175, Time: 0.0254 Steps: 34320, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000266, Sample Num: 4256, Cur Loss: 0.34655502, Cur Avg Loss: 0.33785234, Log Avg loss: 0.34979906, Global Avg Loss: 1.42556830, Time: 0.0210 Steps: 34330, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000276, Sample Num: 4416, Cur Loss: 0.18317032, Cur Avg Loss: 0.33865426, Log Avg loss: 0.35998534, Global Avg Loss: 1.42525800, Time: 0.0210 Steps: 34340, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000286, Sample Num: 4576, Cur Loss: 0.05851781, Cur Avg Loss: 0.33984805, Log Avg loss: 0.37279662, Global Avg Loss: 1.42495160, Time: 0.0209 Steps: 34350, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000296, Sample Num: 4736, Cur Loss: 0.27519369, Cur Avg Loss: 0.33524369, Log Avg loss: 0.20355903, Global Avg Loss: 1.42459614, Time: 0.0209 Steps: 34360, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000306, Sample Num: 4896, Cur Loss: 0.88095796, Cur Avg Loss: 0.33581273, Log Avg loss: 0.35265607, Global Avg Loss: 1.42428425, Time: 0.0209 Steps: 34370, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000316, Sample Num: 5056, Cur Loss: 0.15449615, Cur Avg Loss: 0.33423302, Log Avg loss: 0.28589391, Global Avg Loss: 1.42395313, Time: 0.0209 Steps: 34380, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000326, Sample Num: 5216, Cur Loss: 0.13299614, Cur Avg Loss: 0.33376202, Log Avg loss: 0.31887844, Global Avg Loss: 1.42363180, Time: 0.0209 Steps: 34390, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000336, Sample Num: 5376, Cur Loss: 0.12149718, Cur Avg Loss: 0.33575528, Log Avg loss: 0.40073552, Global Avg Loss: 1.42333444, Time: 0.0209 Steps: 34400, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000346, Sample Num: 5536, Cur Loss: 0.47312313, Cur Avg Loss: 0.33454092, Log Avg loss: 0.29373857, Global Avg Loss: 1.42300617, Time: 0.0209 Steps: 34410, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000356, Sample Num: 5696, Cur Loss: 0.26782495, Cur Avg Loss: 0.33487781, Log Avg loss: 0.34653421, Global Avg Loss: 1.42269342, Time: 0.0210 Steps: 34420, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000366, Sample Num: 5856, Cur Loss: 0.81410897, Cur Avg Loss: 0.33590477, Log Avg loss: 0.37246443, Global Avg Loss: 1.42238839, Time: 0.0209 Steps: 34430, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000376, Sample Num: 6016, Cur Loss: 0.34293884, Cur Avg Loss: 0.33596802, Log Avg loss: 0.33828293, Global Avg Loss: 1.42207361, Time: 0.0209 Steps: 34440, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000386, Sample Num: 6176, Cur Loss: 0.19291097, Cur Avg Loss: 0.33470978, Log Avg loss: 0.28740014, Global Avg Loss: 1.42174424, Time: 0.0209 Steps: 34450, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000396, Sample Num: 6336, Cur Loss: 0.06221502, Cur Avg Loss: 0.33955661, Log Avg loss: 0.52664406, Global Avg Loss: 1.42148449, Time: 0.0209 Steps: 34460, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000406, Sample Num: 6496, Cur Loss: 0.31033307, Cur Avg Loss: 0.33992218, Log Avg loss: 0.35439883, Global Avg Loss: 1.42117492, Time: 0.0209 Steps: 34470, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000416, Sample Num: 6656, Cur Loss: 0.52907050, Cur Avg Loss: 0.34094807, Log Avg loss: 0.38259931, Global Avg Loss: 1.42087371, Time: 0.0209 Steps: 34480, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000426, Sample Num: 6816, Cur Loss: 0.17480206, Cur Avg Loss: 0.34021854, Log Avg loss: 0.30986992, Global Avg Loss: 1.42055159, Time: 0.0209 Steps: 34490, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000436, Sample Num: 6976, Cur Loss: 0.73494506, Cur Avg Loss: 0.34200711, Log Avg loss: 0.41820023, Global Avg Loss: 1.42026105, Time: 0.0209 Steps: 34500, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000446, Sample Num: 7136, Cur Loss: 0.27523607, Cur Avg Loss: 0.34294533, Log Avg loss: 0.38385193, Global Avg Loss: 1.41996073, Time: 0.0209 Steps: 34510, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000456, Sample Num: 7296, Cur Loss: 0.55336320, Cur Avg Loss: 0.34299259, Log Avg loss: 0.34510014, Global Avg Loss: 1.41964935, Time: 0.0210 Steps: 34520, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000466, Sample Num: 7456, Cur Loss: 0.29132348, Cur Avg Loss: 0.34476562, Log Avg loss: 0.42561574, Global Avg Loss: 1.41936148, Time: 0.0209 Steps: 34530, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000476, Sample Num: 7616, Cur Loss: 0.31359127, Cur Avg Loss: 0.34622212, Log Avg loss: 0.41409535, Global Avg Loss: 1.41907043, Time: 0.0209 Steps: 34540, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000486, Sample Num: 7776, Cur Loss: 0.36389986, Cur Avg Loss: 0.34497587, Log Avg loss: 0.28565416, Global Avg Loss: 1.41874238, Time: 0.0209 Steps: 34550, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000496, Sample Num: 7936, Cur Loss: 0.23979425, Cur Avg Loss: 0.34308294, Log Avg loss: 0.25108640, Global Avg Loss: 1.41840452, Time: 0.0209 Steps: 34560, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000506, Sample Num: 8096, Cur Loss: 0.34668350, Cur Avg Loss: 0.34420703, Log Avg loss: 0.39996186, Global Avg Loss: 1.41810992, Time: 0.0210 Steps: 34570, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000516, Sample Num: 8256, Cur Loss: 0.36511919, Cur Avg Loss: 0.34701025, Log Avg loss: 0.48885364, Global Avg Loss: 1.41784119, Time: 0.0247 Steps: 34580, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000526, Sample Num: 8416, Cur Loss: 0.62336218, Cur Avg Loss: 0.34977633, Log Avg loss: 0.49250585, Global Avg Loss: 1.41757368, Time: 0.0208 Steps: 34590, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000536, Sample Num: 8576, Cur Loss: 0.27782062, Cur Avg Loss: 0.35062360, Log Avg loss: 0.39518982, Global Avg Loss: 1.41727819, Time: 0.0209 Steps: 34600, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000546, Sample Num: 8736, Cur Loss: 0.46954477, Cur Avg Loss: 0.35328647, Log Avg loss: 0.49601671, Global Avg Loss: 1.41701200, Time: 0.0208 Steps: 34610, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000556, Sample Num: 8896, Cur Loss: 0.25331089, Cur Avg Loss: 0.35265408, Log Avg loss: 0.31812551, Global Avg Loss: 1.41669459, Time: 0.0208 Steps: 34620, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000566, Sample Num: 9056, Cur Loss: 0.32196125, Cur Avg Loss: 0.35390277, Log Avg loss: 0.42332949, Global Avg Loss: 1.41640774, Time: 0.0209 Steps: 34630, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000576, Sample Num: 9216, Cur Loss: 0.36350891, Cur Avg Loss: 0.35402861, Log Avg loss: 0.36115162, Global Avg Loss: 1.41610311, Time: 0.0208 Steps: 34640, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000586, Sample Num: 9376, Cur Loss: 0.47686470, Cur Avg Loss: 0.35448981, Log Avg loss: 0.38105462, Global Avg Loss: 1.41580439, Time: 0.0209 Steps: 34650, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000596, Sample Num: 9536, Cur Loss: 0.10282566, Cur Avg Loss: 0.35391746, Log Avg loss: 0.32037798, Global Avg Loss: 1.41548834, Time: 0.0209 Steps: 34660, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000606, Sample Num: 9696, Cur Loss: 0.16294986, Cur Avg Loss: 0.35443148, Log Avg loss: 0.38506692, Global Avg Loss: 1.41519113, Time: 0.0209 Steps: 34670, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000616, Sample Num: 9856, Cur Loss: 0.12933455, Cur Avg Loss: 0.35217827, Log Avg loss: 0.21563372, Global Avg Loss: 1.41484524, Time: 0.0209 Steps: 34680, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000626, Sample Num: 10016, Cur Loss: 0.32853127, Cur Avg Loss: 0.35101869, Log Avg loss: 0.27958850, Global Avg Loss: 1.41451798, Time: 0.0209 Steps: 34690, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000636, Sample Num: 10176, Cur Loss: 0.20623103, Cur Avg Loss: 0.35031782, Log Avg loss: 0.30644325, Global Avg Loss: 1.41419865, Time: 0.0209 Steps: 34700, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000646, Sample Num: 10336, Cur Loss: 0.29371867, Cur Avg Loss: 0.35011644, Log Avg loss: 0.33730900, Global Avg Loss: 1.41388840, Time: 0.0209 Steps: 34710, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000656, Sample Num: 10496, Cur Loss: 0.26118845, Cur Avg Loss: 0.35100083, Log Avg loss: 0.40813215, Global Avg Loss: 1.41359872, Time: 0.0209 Steps: 34720, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000666, Sample Num: 10656, Cur Loss: 0.41539049, Cur Avg Loss: 0.34935639, Log Avg loss: 0.24148118, Global Avg Loss: 1.41326123, Time: 0.0209 Steps: 34730, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000676, Sample Num: 10816, Cur Loss: 0.50872386, Cur Avg Loss: 0.34817199, Log Avg loss: 0.26929115, Global Avg Loss: 1.41293193, Time: 0.0208 Steps: 34740, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000686, Sample Num: 10976, Cur Loss: 0.35318148, Cur Avg Loss: 0.34710431, Log Avg loss: 0.27492881, Global Avg Loss: 1.41260445, Time: 0.0209 Steps: 34750, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000696, Sample Num: 11136, Cur Loss: 0.19908458, Cur Avg Loss: 0.34631368, Log Avg loss: 0.29207653, Global Avg Loss: 1.41228209, Time: 0.0209 Steps: 34760, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000706, Sample Num: 11296, Cur Loss: 0.20186998, Cur Avg Loss: 0.34627909, Log Avg loss: 0.34387209, Global Avg Loss: 1.41197481, Time: 0.0209 Steps: 34770, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000716, Sample Num: 11456, Cur Loss: 0.16995478, Cur Avg Loss: 0.34829040, Log Avg loss: 0.49028887, Global Avg Loss: 1.41170980, Time: 0.0209 Steps: 34780, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000726, Sample Num: 11616, Cur Loss: 0.32192343, Cur Avg Loss: 0.35063197, Log Avg loss: 0.51828797, Global Avg Loss: 1.41145300, Time: 0.0209 Steps: 34790, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000736, Sample Num: 11776, Cur Loss: 0.17231509, Cur Avg Loss: 0.35099171, Log Avg loss: 0.37710891, Global Avg Loss: 1.41115578, Time: 0.0209 Steps: 34800, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000746, Sample Num: 11936, Cur Loss: 0.36345497, Cur Avg Loss: 0.35021534, Log Avg loss: 0.29307480, Global Avg Loss: 1.41083458, Time: 0.0209 Steps: 34810, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000756, Sample Num: 12096, Cur Loss: 0.44921389, Cur Avg Loss: 0.35113550, Log Avg loss: 0.41977957, Global Avg Loss: 1.41054996, Time: 0.0208 Steps: 34820, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000766, Sample Num: 12256, Cur Loss: 0.76929033, Cur Avg Loss: 0.35484849, Log Avg loss: 0.63555048, Global Avg Loss: 1.41032745, Time: 0.0209 Steps: 34830, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000776, Sample Num: 12416, Cur Loss: 0.47417182, Cur Avg Loss: 0.35505452, Log Avg loss: 0.37083612, Global Avg Loss: 1.41002909, Time: 0.0210 Steps: 34840, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000786, Sample Num: 12576, Cur Loss: 0.13837743, Cur Avg Loss: 0.35437945, Log Avg loss: 0.30199401, Global Avg Loss: 1.40971114, Time: 0.0210 Steps: 34850, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000796, Sample Num: 12736, Cur Loss: 0.20067072, Cur Avg Loss: 0.35354736, Log Avg loss: 0.28814469, Global Avg Loss: 1.40938941, Time: 0.0210 Steps: 34860, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000806, Sample Num: 12896, Cur Loss: 0.48344824, Cur Avg Loss: 0.35284846, Log Avg loss: 0.29721649, Global Avg Loss: 1.40907046, Time: 0.0210 Steps: 34870, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000816, Sample Num: 13056, Cur Loss: 0.23655069, Cur Avg Loss: 0.35245866, Log Avg loss: 0.32104046, Global Avg Loss: 1.40875852, Time: 0.0210 Steps: 34880, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000826, Sample Num: 13216, Cur Loss: 0.15274087, Cur Avg Loss: 0.35263398, Log Avg loss: 0.36694015, Global Avg Loss: 1.40845992, Time: 0.0210 Steps: 34890, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000836, Sample Num: 13376, Cur Loss: 0.21009874, Cur Avg Loss: 0.35205780, Log Avg loss: 0.30446569, Global Avg Loss: 1.40814359, Time: 0.0210 Steps: 34900, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000846, Sample Num: 13536, Cur Loss: 0.39749652, Cur Avg Loss: 0.35180774, Log Avg loss: 0.33090242, Global Avg Loss: 1.40783502, Time: 0.0210 Steps: 34910, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000856, Sample Num: 13696, Cur Loss: 0.14675054, Cur Avg Loss: 0.35046502, Log Avg loss: 0.23687109, Global Avg Loss: 1.40749969, Time: 0.0211 Steps: 34920, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000866, Sample Num: 13856, Cur Loss: 0.45931354, Cur Avg Loss: 0.35013363, Log Avg loss: 0.32176641, Global Avg Loss: 1.40718886, Time: 0.0210 Steps: 34930, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000876, Sample Num: 14016, Cur Loss: 1.25072134, Cur Avg Loss: 0.35193000, Log Avg loss: 0.50749574, Global Avg Loss: 1.40693136, Time: 0.0210 Steps: 34940, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000886, Sample Num: 14176, Cur Loss: 0.40658861, Cur Avg Loss: 0.35154038, Log Avg loss: 0.31740965, Global Avg Loss: 1.40661962, Time: 0.0211 Steps: 34950, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000896, Sample Num: 14336, Cur Loss: 0.29433429, Cur Avg Loss: 0.35188558, Log Avg loss: 0.38247073, Global Avg Loss: 1.40632668, Time: 0.0210 Steps: 34960, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000906, Sample Num: 14496, Cur Loss: 0.15384305, Cur Avg Loss: 0.35093356, Log Avg loss: 0.26563247, Global Avg Loss: 1.40600048, Time: 0.0210 Steps: 34970, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000916, Sample Num: 14656, Cur Loss: 0.50622970, Cur Avg Loss: 0.35027922, Log Avg loss: 0.29099564, Global Avg Loss: 1.40568173, Time: 0.0210 Steps: 34980, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000926, Sample Num: 14816, Cur Loss: 0.28168148, Cur Avg Loss: 0.34988588, Log Avg loss: 0.31385600, Global Avg Loss: 1.40536969, Time: 0.0210 Steps: 34990, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000936, Sample Num: 14976, Cur Loss: 0.19930074, Cur Avg Loss: 0.34995942, Log Avg loss: 0.35676967, Global Avg Loss: 1.40507009, Time: 0.0210 Steps: 35000, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000946, Sample Num: 15136, Cur Loss: 0.31924790, Cur Avg Loss: 0.35000185, Log Avg loss: 0.35397249, Global Avg Loss: 1.40476986, Time: 0.0210 Steps: 35010, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000956, Sample Num: 15296, Cur Loss: 0.19286147, Cur Avg Loss: 0.34913099, Log Avg loss: 0.26674770, Global Avg Loss: 1.40444490, Time: 0.0210 Steps: 35020, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000966, Sample Num: 15456, Cur Loss: 0.25776395, Cur Avg Loss: 0.34891280, Log Avg loss: 0.32805426, Global Avg Loss: 1.40413762, Time: 0.0210 Steps: 35030, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000976, Sample Num: 15616, Cur Loss: 0.30009326, Cur Avg Loss: 0.34758520, Log Avg loss: 0.21933926, Global Avg Loss: 1.40379949, Time: 0.0210 Steps: 35040, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000986, Sample Num: 15776, Cur Loss: 0.23195474, Cur Avg Loss: 0.34733622, Log Avg loss: 0.32303568, Global Avg Loss: 1.40349114, Time: 0.0210 Steps: 35050, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000996, Sample Num: 15936, Cur Loss: 0.34442937, Cur Avg Loss: 0.34655993, Log Avg loss: 0.27001722, Global Avg Loss: 1.40316785, Time: 0.0210 Steps: 35060, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001006, Sample Num: 16096, Cur Loss: 0.15083821, Cur Avg Loss: 0.34530003, Log Avg loss: 0.21981404, Global Avg Loss: 1.40283042, Time: 0.0210 Steps: 35070, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001016, Sample Num: 16256, Cur Loss: 0.56750762, Cur Avg Loss: 0.34529358, Log Avg loss: 0.34464467, Global Avg Loss: 1.40252877, Time: 0.0210 Steps: 35080, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001026, Sample Num: 16416, Cur Loss: 0.18360996, Cur Avg Loss: 0.34492908, Log Avg loss: 0.30789598, Global Avg Loss: 1.40221682, Time: 0.0247 Steps: 35090, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001036, Sample Num: 16576, Cur Loss: 0.44848230, Cur Avg Loss: 0.34398722, Log Avg loss: 0.24735217, Global Avg Loss: 1.40188780, Time: 0.0209 Steps: 35100, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001046, Sample Num: 16736, Cur Loss: 0.35365742, Cur Avg Loss: 0.34333326, Log Avg loss: 0.27558321, Global Avg Loss: 1.40156701, Time: 0.0209 Steps: 35110, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001056, Sample Num: 16896, Cur Loss: 0.17609602, Cur Avg Loss: 0.34398385, Log Avg loss: 0.41203623, Global Avg Loss: 1.40128525, Time: 0.0210 Steps: 35120, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001066, Sample Num: 17056, Cur Loss: 0.36813861, Cur Avg Loss: 0.34532537, Log Avg loss: 0.48698932, Global Avg Loss: 1.40102499, Time: 0.0210 Steps: 35130, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001076, Sample Num: 17216, Cur Loss: 0.37631136, Cur Avg Loss: 0.34423132, Log Avg loss: 0.22760602, Global Avg Loss: 1.40069106, Time: 0.0209 Steps: 35140, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001086, Sample Num: 17376, Cur Loss: 0.25304279, Cur Avg Loss: 0.34354219, Log Avg loss: 0.26939127, Global Avg Loss: 1.40036921, Time: 0.0210 Steps: 35150, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001096, Sample Num: 17536, Cur Loss: 0.29766619, Cur Avg Loss: 0.34314081, Log Avg loss: 0.29955150, Global Avg Loss: 1.40005613, Time: 0.0208 Steps: 35160, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001106, Sample Num: 17696, Cur Loss: 0.81881267, Cur Avg Loss: 0.34632744, Log Avg loss: 0.69558200, Global Avg Loss: 1.39985582, Time: 0.0210 Steps: 35170, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001116, Sample Num: 17856, Cur Loss: 0.51901370, Cur Avg Loss: 0.34772712, Log Avg loss: 0.50253191, Global Avg Loss: 1.39960075, Time: 0.0210 Steps: 35180, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001126, Sample Num: 18016, Cur Loss: 0.11499986, Cur Avg Loss: 0.34814380, Log Avg loss: 0.39464491, Global Avg Loss: 1.39931517, Time: 0.0209 Steps: 35190, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001136, Sample Num: 18176, Cur Loss: 0.21291620, Cur Avg Loss: 0.34795316, Log Avg loss: 0.32648692, Global Avg Loss: 1.39901039, Time: 0.0209 Steps: 35200, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001146, Sample Num: 18336, Cur Loss: 0.43653131, Cur Avg Loss: 0.34719249, Log Avg loss: 0.26078098, Global Avg Loss: 1.39868713, Time: 0.0210 Steps: 35210, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001156, Sample Num: 18496, Cur Loss: 0.19332191, Cur Avg Loss: 0.34657062, Log Avg loss: 0.27530401, Global Avg Loss: 1.39836816, Time: 0.0208 Steps: 35220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001166, Sample Num: 18656, Cur Loss: 0.10253940, Cur Avg Loss: 0.34571833, Log Avg loss: 0.24719338, Global Avg Loss: 1.39804140, Time: 0.0209 Steps: 35230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001176, Sample Num: 18816, Cur Loss: 0.10547896, Cur Avg Loss: 0.34546354, Log Avg loss: 0.31575545, Global Avg Loss: 1.39773429, Time: 0.0210 Steps: 35240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001186, Sample Num: 18976, Cur Loss: 0.64670497, Cur Avg Loss: 0.34469005, Log Avg loss: 0.25372723, Global Avg Loss: 1.39740974, Time: 0.0210 Steps: 35250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001196, Sample Num: 19136, Cur Loss: 0.20091888, Cur Avg Loss: 0.34445254, Log Avg loss: 0.31628441, Global Avg Loss: 1.39710313, Time: 0.0210 Steps: 35260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001206, Sample Num: 19296, Cur Loss: 0.68908232, Cur Avg Loss: 0.34698686, Log Avg loss: 0.65009124, Global Avg Loss: 1.39689133, Time: 0.0209 Steps: 35270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001216, Sample Num: 19456, Cur Loss: 0.24508898, Cur Avg Loss: 0.34715763, Log Avg loss: 0.36775219, Global Avg Loss: 1.39659962, Time: 0.0210 Steps: 35280, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001226, Sample Num: 19616, Cur Loss: 0.24061683, Cur Avg Loss: 0.34659148, Log Avg loss: 0.27774800, Global Avg Loss: 1.39628258, Time: 0.0209 Steps: 35290, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001236, Sample Num: 19776, Cur Loss: 0.18867382, Cur Avg Loss: 0.34543859, Log Avg loss: 0.20409368, Global Avg Loss: 1.39594485, Time: 0.0210 Steps: 35300, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001246, Sample Num: 19936, Cur Loss: 0.69017512, Cur Avg Loss: 0.34636463, Log Avg loss: 0.46082353, Global Avg Loss: 1.39568002, Time: 0.0210 Steps: 35310, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001256, Sample Num: 20096, Cur Loss: 0.16597106, Cur Avg Loss: 0.34619274, Log Avg loss: 0.32477457, Global Avg Loss: 1.39537682, Time: 0.0210 Steps: 35320, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001266, Sample Num: 20256, Cur Loss: 0.71146417, Cur Avg Loss: 0.34719238, Log Avg loss: 0.47274796, Global Avg Loss: 1.39511567, Time: 0.0210 Steps: 35330, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001276, Sample Num: 20416, Cur Loss: 0.44867471, Cur Avg Loss: 0.34790224, Log Avg loss: 0.43777014, Global Avg Loss: 1.39484477, Time: 0.0210 Steps: 35340, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001286, Sample Num: 20576, Cur Loss: 0.10613157, Cur Avg Loss: 0.34756562, Log Avg loss: 0.30461345, Global Avg Loss: 1.39453636, Time: 0.0212 Steps: 35350, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001296, Sample Num: 20736, Cur Loss: 0.37369961, Cur Avg Loss: 0.34791976, Log Avg loss: 0.39346157, Global Avg Loss: 1.39425325, Time: 0.0209 Steps: 35360, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001306, Sample Num: 20896, Cur Loss: 0.28947762, Cur Avg Loss: 0.34782654, Log Avg loss: 0.33574540, Global Avg Loss: 1.39395399, Time: 0.0209 Steps: 35370, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001316, Sample Num: 21056, Cur Loss: 0.73594606, Cur Avg Loss: 0.34748182, Log Avg loss: 0.30246136, Global Avg Loss: 1.39364548, Time: 0.0209 Steps: 35380, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001326, Sample Num: 21216, Cur Loss: 0.35636872, Cur Avg Loss: 0.34723622, Log Avg loss: 0.31491538, Global Avg Loss: 1.39334067, Time: 0.0210 Steps: 35390, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001336, Sample Num: 21376, Cur Loss: 0.25225323, Cur Avg Loss: 0.34710551, Log Avg loss: 0.32977349, Global Avg Loss: 1.39304023, Time: 0.0209 Steps: 35400, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001346, Sample Num: 21536, Cur Loss: 0.25119671, Cur Avg Loss: 0.34633150, Log Avg loss: 0.24292388, Global Avg Loss: 1.39271543, Time: 0.0209 Steps: 35410, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001356, Sample Num: 21696, Cur Loss: 0.09479211, Cur Avg Loss: 0.34648613, Log Avg loss: 0.36729968, Global Avg Loss: 1.39242593, Time: 0.0209 Steps: 35420, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001366, Sample Num: 21856, Cur Loss: 0.41486034, Cur Avg Loss: 0.34588409, Log Avg loss: 0.26424724, Global Avg Loss: 1.39210750, Time: 0.0210 Steps: 35430, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001376, Sample Num: 22016, Cur Loss: 0.09569818, Cur Avg Loss: 0.34580982, Log Avg loss: 0.33566399, Global Avg Loss: 1.39180941, Time: 0.0209 Steps: 35440, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001386, Sample Num: 22176, Cur Loss: 0.11953931, Cur Avg Loss: 0.34559552, Log Avg loss: 0.31610729, Global Avg Loss: 1.39150597, Time: 0.0209 Steps: 35450, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001396, Sample Num: 22336, Cur Loss: 0.41668063, Cur Avg Loss: 0.34634250, Log Avg loss: 0.44987435, Global Avg Loss: 1.39124042, Time: 0.0209 Steps: 35460, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001406, Sample Num: 22496, Cur Loss: 0.59157151, Cur Avg Loss: 0.34623578, Log Avg loss: 0.33133815, Global Avg Loss: 1.39094160, Time: 0.0209 Steps: 35470, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001416, Sample Num: 22656, Cur Loss: 0.40951079, Cur Avg Loss: 0.34717603, Log Avg loss: 0.47937480, Global Avg Loss: 1.39068468, Time: 0.0210 Steps: 35480, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001426, Sample Num: 22816, Cur Loss: 0.37026790, Cur Avg Loss: 0.34694395, Log Avg loss: 0.31408190, Global Avg Loss: 1.39038132, Time: 0.0209 Steps: 35490, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001436, Sample Num: 22976, Cur Loss: 0.25618806, Cur Avg Loss: 0.34717953, Log Avg loss: 0.38077273, Global Avg Loss: 1.39009693, Time: 0.0209 Steps: 35500, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001446, Sample Num: 23136, Cur Loss: 0.54817045, Cur Avg Loss: 0.34684909, Log Avg loss: 0.29939869, Global Avg Loss: 1.38978977, Time: 0.0209 Steps: 35510, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001456, Sample Num: 23296, Cur Loss: 0.15921059, Cur Avg Loss: 0.34671241, Log Avg loss: 0.32694717, Global Avg Loss: 1.38949055, Time: 0.0209 Steps: 35520, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001466, Sample Num: 23456, Cur Loss: 0.20699617, Cur Avg Loss: 0.34637969, Log Avg loss: 0.29793678, Global Avg Loss: 1.38918333, Time: 0.0209 Steps: 35530, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001476, Sample Num: 23616, Cur Loss: 0.15784955, Cur Avg Loss: 0.34669436, Log Avg loss: 0.39282487, Global Avg Loss: 1.38890298, Time: 0.0209 Steps: 35540, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001486, Sample Num: 23776, Cur Loss: 0.24109498, Cur Avg Loss: 0.34658751, Log Avg loss: 0.33081622, Global Avg Loss: 1.38860535, Time: 0.0210 Steps: 35550, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001496, Sample Num: 23936, Cur Loss: 0.33563492, Cur Avg Loss: 0.34588622, Log Avg loss: 0.24167411, Global Avg Loss: 1.38828281, Time: 0.0209 Steps: 35560, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001506, Sample Num: 24096, Cur Loss: 0.56705010, Cur Avg Loss: 0.34631297, Log Avg loss: 0.41015507, Global Avg Loss: 1.38800783, Time: 0.0210 Steps: 35570, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001516, Sample Num: 24256, Cur Loss: 0.44546938, Cur Avg Loss: 0.34591130, Log Avg loss: 0.28542046, Global Avg Loss: 1.38769794, Time: 0.0210 Steps: 35580, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001526, Sample Num: 24416, Cur Loss: 0.13102253, Cur Avg Loss: 0.34544490, Log Avg loss: 0.27473798, Global Avg Loss: 1.38738522, Time: 0.0209 Steps: 35590, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001536, Sample Num: 24576, Cur Loss: 0.13248794, Cur Avg Loss: 0.34490442, Log Avg loss: 0.26242689, Global Avg Loss: 1.38706922, Time: 0.0254 Steps: 35600, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001546, Sample Num: 24736, Cur Loss: 0.62956291, Cur Avg Loss: 0.34499023, Log Avg loss: 0.35817032, Global Avg Loss: 1.38678029, Time: 0.0211 Steps: 35610, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001556, Sample Num: 24896, Cur Loss: 0.08205307, Cur Avg Loss: 0.34462679, Log Avg loss: 0.28844007, Global Avg Loss: 1.38647194, Time: 0.0211 Steps: 35620, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001566, Sample Num: 25056, Cur Loss: 0.22193289, Cur Avg Loss: 0.34507780, Log Avg loss: 0.41525442, Global Avg Loss: 1.38619935, Time: 0.0212 Steps: 35630, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001576, Sample Num: 25216, Cur Loss: 0.29405296, Cur Avg Loss: 0.34560058, Log Avg loss: 0.42746748, Global Avg Loss: 1.38593035, Time: 0.0212 Steps: 35640, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001586, Sample Num: 25376, Cur Loss: 0.22334164, Cur Avg Loss: 0.34634514, Log Avg loss: 0.46368899, Global Avg Loss: 1.38567165, Time: 0.0211 Steps: 35650, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001596, Sample Num: 25536, Cur Loss: 0.19396043, Cur Avg Loss: 0.34567793, Log Avg loss: 0.23985845, Global Avg Loss: 1.38535034, Time: 0.0211 Steps: 35660, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001606, Sample Num: 25696, Cur Loss: 0.13299794, Cur Avg Loss: 0.34534650, Log Avg loss: 0.29244926, Global Avg Loss: 1.38504395, Time: 0.0219 Steps: 35670, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001616, Sample Num: 25856, Cur Loss: 0.35909534, Cur Avg Loss: 0.34483175, Log Avg loss: 0.26216323, Global Avg Loss: 1.38472924, Time: 0.0219 Steps: 35680, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001626, Sample Num: 26016, Cur Loss: 0.31401366, Cur Avg Loss: 0.34458769, Log Avg loss: 0.30514801, Global Avg Loss: 1.38442675, Time: 0.0219 Steps: 35690, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001636, Sample Num: 26176, Cur Loss: 0.22909401, Cur Avg Loss: 0.34447213, Log Avg loss: 0.32568138, Global Avg Loss: 1.38413018, Time: 0.0211 Steps: 35700, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001646, Sample Num: 26336, Cur Loss: 0.14180136, Cur Avg Loss: 0.34408543, Log Avg loss: 0.28082080, Global Avg Loss: 1.38382122, Time: 0.0210 Steps: 35710, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001656, Sample Num: 26496, Cur Loss: 0.28992385, Cur Avg Loss: 0.34368162, Log Avg loss: 0.27721551, Global Avg Loss: 1.38351142, Time: 0.0219 Steps: 35720, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001666, Sample Num: 26656, Cur Loss: 0.19716962, Cur Avg Loss: 0.34329721, Log Avg loss: 0.27963905, Global Avg Loss: 1.38320247, Time: 0.0210 Steps: 35730, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001676, Sample Num: 26816, Cur Loss: 0.12472147, Cur Avg Loss: 0.34309565, Log Avg loss: 0.30951592, Global Avg Loss: 1.38290205, Time: 0.0211 Steps: 35740, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001686, Sample Num: 26976, Cur Loss: 0.09123165, Cur Avg Loss: 0.34248541, Log Avg loss: 0.24020947, Global Avg Loss: 1.38258242, Time: 0.0211 Steps: 35750, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001696, Sample Num: 27136, Cur Loss: 0.14711539, Cur Avg Loss: 0.34203205, Log Avg loss: 0.26559487, Global Avg Loss: 1.38227006, Time: 0.0220 Steps: 35760, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001706, Sample Num: 27296, Cur Loss: 0.19973329, Cur Avg Loss: 0.34190925, Log Avg loss: 0.32108261, Global Avg Loss: 1.38197339, Time: 0.0211 Steps: 35770, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001716, Sample Num: 27456, Cur Loss: 0.53968114, Cur Avg Loss: 0.34210809, Log Avg loss: 0.37603053, Global Avg Loss: 1.38169225, Time: 0.0212 Steps: 35780, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001726, Sample Num: 27616, Cur Loss: 0.22189672, Cur Avg Loss: 0.34148964, Log Avg loss: 0.23536238, Global Avg Loss: 1.38137195, Time: 0.0210 Steps: 35790, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001736, Sample Num: 27776, Cur Loss: 0.17880522, Cur Avg Loss: 0.34125544, Log Avg loss: 0.30083235, Global Avg Loss: 1.38107013, Time: 0.0220 Steps: 35800, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001746, Sample Num: 27936, Cur Loss: 0.18343686, Cur Avg Loss: 0.34156969, Log Avg loss: 0.39612360, Global Avg Loss: 1.38079508, Time: 0.0211 Steps: 35810, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001756, Sample Num: 28096, Cur Loss: 0.37979138, Cur Avg Loss: 0.34101612, Log Avg loss: 0.24436314, Global Avg Loss: 1.38047782, Time: 0.0212 Steps: 35820, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001766, Sample Num: 28256, Cur Loss: 0.60508335, Cur Avg Loss: 0.34189515, Log Avg loss: 0.49625369, Global Avg Loss: 1.38023103, Time: 0.0211 Steps: 35830, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001776, Sample Num: 28416, Cur Loss: 0.36655822, Cur Avg Loss: 0.34197674, Log Avg loss: 0.35638561, Global Avg Loss: 1.37994536, Time: 0.0211 Steps: 35840, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001786, Sample Num: 28576, Cur Loss: 0.55907106, Cur Avg Loss: 0.34258561, Log Avg loss: 0.45071955, Global Avg Loss: 1.37968616, Time: 0.0211 Steps: 35850, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001796, Sample Num: 28736, Cur Loss: 0.20837021, Cur Avg Loss: 0.34243622, Log Avg loss: 0.31575528, Global Avg Loss: 1.37938947, Time: 0.0247 Steps: 35860, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001806, Sample Num: 28896, Cur Loss: 0.03683859, Cur Avg Loss: 0.34206774, Log Avg loss: 0.27588929, Global Avg Loss: 1.37908183, Time: 0.0210 Steps: 35870, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001816, Sample Num: 29056, Cur Loss: 0.78335607, Cur Avg Loss: 0.34205628, Log Avg loss: 0.33998655, Global Avg Loss: 1.37879223, Time: 0.0209 Steps: 35880, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001826, Sample Num: 29216, Cur Loss: 0.34195325, Cur Avg Loss: 0.34153104, Log Avg loss: 0.24614829, Global Avg Loss: 1.37847664, Time: 0.0210 Steps: 35890, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001836, Sample Num: 29376, Cur Loss: 0.41299096, Cur Avg Loss: 0.34162311, Log Avg loss: 0.35843450, Global Avg Loss: 1.37819251, Time: 0.0211 Steps: 35900, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001846, Sample Num: 29536, Cur Loss: 0.56351078, Cur Avg Loss: 0.34171483, Log Avg loss: 0.35855394, Global Avg Loss: 1.37790857, Time: 0.0210 Steps: 35910, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001856, Sample Num: 29696, Cur Loss: 1.25713873, Cur Avg Loss: 0.34288214, Log Avg loss: 0.55836717, Global Avg Loss: 1.37768041, Time: 0.0211 Steps: 35920, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001866, Sample Num: 29856, Cur Loss: 0.36931920, Cur Avg Loss: 0.34275765, Log Avg loss: 0.31965412, Global Avg Loss: 1.37738594, Time: 0.0210 Steps: 35930, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001876, Sample Num: 30016, Cur Loss: 0.90798628, Cur Avg Loss: 0.34320868, Log Avg loss: 0.42737002, Global Avg Loss: 1.37712161, Time: 0.0210 Steps: 35940, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001886, Sample Num: 30176, Cur Loss: 0.28224537, Cur Avg Loss: 0.34409487, Log Avg loss: 0.51034499, Global Avg Loss: 1.37688050, Time: 0.0210 Steps: 35950, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001896, Sample Num: 30336, Cur Loss: 0.20448634, Cur Avg Loss: 0.34365189, Log Avg loss: 0.26010523, Global Avg Loss: 1.37656994, Time: 0.0211 Steps: 35960, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001906, Sample Num: 30496, Cur Loss: 0.19299863, Cur Avg Loss: 0.34410616, Log Avg loss: 0.43023574, Global Avg Loss: 1.37630685, Time: 0.0210 Steps: 35970, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001916, Sample Num: 30656, Cur Loss: 0.12257008, Cur Avg Loss: 0.34386627, Log Avg loss: 0.29814343, Global Avg Loss: 1.37600719, Time: 0.0209 Steps: 35980, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001926, Sample Num: 30816, Cur Loss: 0.39307827, Cur Avg Loss: 0.34394044, Log Avg loss: 0.35815068, Global Avg Loss: 1.37572438, Time: 0.0211 Steps: 35990, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001936, Sample Num: 30976, Cur Loss: 0.40401334, Cur Avg Loss: 0.34379438, Log Avg loss: 0.31566357, Global Avg Loss: 1.37542992, Time: 0.0210 Steps: 36000, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001946, Sample Num: 31136, Cur Loss: 0.23119462, Cur Avg Loss: 0.34438532, Log Avg loss: 0.45879073, Global Avg Loss: 1.37517537, Time: 0.0211 Steps: 36010, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001956, Sample Num: 31296, Cur Loss: 0.27715945, Cur Avg Loss: 0.34417118, Log Avg loss: 0.30249944, Global Avg Loss: 1.37487757, Time: 0.0210 Steps: 36020, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001966, Sample Num: 31456, Cur Loss: 0.63066214, Cur Avg Loss: 0.34424252, Log Avg loss: 0.35819669, Global Avg Loss: 1.37459539, Time: 0.0211 Steps: 36030, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001976, Sample Num: 31616, Cur Loss: 0.20049983, Cur Avg Loss: 0.34413394, Log Avg loss: 0.32278863, Global Avg Loss: 1.37430354, Time: 0.0210 Steps: 36040, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001986, Sample Num: 31776, Cur Loss: 0.29275513, Cur Avg Loss: 0.34407170, Log Avg loss: 0.33177286, Global Avg Loss: 1.37401435, Time: 0.0211 Steps: 36050, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001996, Sample Num: 31936, Cur Loss: 0.46182901, Cur Avg Loss: 0.34382746, Log Avg loss: 0.29532048, Global Avg Loss: 1.37371522, Time: 0.0210 Steps: 36060, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002006, Sample Num: 32096, Cur Loss: 0.22649562, Cur Avg Loss: 0.34385520, Log Avg loss: 0.34939294, Global Avg Loss: 1.37343123, Time: 0.0210 Steps: 36070, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002016, Sample Num: 32256, Cur Loss: 0.56621635, Cur Avg Loss: 0.34412462, Log Avg loss: 0.39816873, Global Avg Loss: 1.37316093, Time: 0.0211 Steps: 36080, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002026, Sample Num: 32416, Cur Loss: 0.18136381, Cur Avg Loss: 0.34373706, Log Avg loss: 0.26560595, Global Avg Loss: 1.37285404, Time: 0.0210 Steps: 36090, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002036, Sample Num: 32576, Cur Loss: 0.06290618, Cur Avg Loss: 0.34337345, Log Avg loss: 0.26970691, Global Avg Loss: 1.37254846, Time: 0.0210 Steps: 36100, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002046, Sample Num: 32736, Cur Loss: 0.47635627, Cur Avg Loss: 0.34323893, Log Avg loss: 0.31584929, Global Avg Loss: 1.37225583, Time: 0.0210 Steps: 36110, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002056, Sample Num: 32896, Cur Loss: 0.11520660, Cur Avg Loss: 0.34255572, Log Avg loss: 0.20277125, Global Avg Loss: 1.37193205, Time: 0.0209 Steps: 36120, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002066, Sample Num: 33056, Cur Loss: 0.54142600, Cur Avg Loss: 0.34245446, Log Avg loss: 0.32163523, Global Avg Loss: 1.37164135, Time: 0.0209 Steps: 36130, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002076, Sample Num: 33216, Cur Loss: 0.18720868, Cur Avg Loss: 0.34248336, Log Avg loss: 0.34845401, Global Avg Loss: 1.37135823, Time: 0.0209 Steps: 36140, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002086, Sample Num: 33376, Cur Loss: 0.20184831, Cur Avg Loss: 0.34203985, Log Avg loss: 0.24996779, Global Avg Loss: 1.37104803, Time: 0.0209 Steps: 36150, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002096, Sample Num: 33536, Cur Loss: 0.22400877, Cur Avg Loss: 0.34140629, Log Avg loss: 0.20924641, Global Avg Loss: 1.37072673, Time: 0.0209 Steps: 36160, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002106, Sample Num: 33696, Cur Loss: 0.19487822, Cur Avg Loss: 0.34139111, Log Avg loss: 0.33820930, Global Avg Loss: 1.37044127, Time: 0.0209 Steps: 36170, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002116, Sample Num: 33856, Cur Loss: 0.14864758, Cur Avg Loss: 0.34122454, Log Avg loss: 0.30614401, Global Avg Loss: 1.37014710, Time: 0.0209 Steps: 36180, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002126, Sample Num: 34016, Cur Loss: 0.56408381, Cur Avg Loss: 0.34147660, Log Avg loss: 0.39481242, Global Avg Loss: 1.36987760, Time: 0.0209 Steps: 36190, Updated lr: 0.000067 ***** Running evaluation checkpoint-36193 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-36193 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.841479, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.052625, "eval_total_loss": 739.995409, "eval_mae": 0.928528, "eval_mse": 1.052618, "eval_r2": 0.330887, "eval_sp_statistic": 0.864168, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.900936, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.576514, "test_total_loss": 289.409809, "test_mae": 0.686672, "test_mse": 0.576616, "test_r2": 0.627847, "test_sp_statistic": 0.847759, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.899213, "test_ps_pvalue": 0.0, "lr": 6.662588904694168e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3698051299023792, "train_cur_epoch_loss": 727.465988798067, "train_cur_epoch_avg_loss": 0.3416937476740568, "train_cur_epoch_time": 44.8414785861969, "train_cur_epoch_avg_time": 0.021062225733300563, "epoch": 17, "step": 36193} ################################################## Training, Epoch: 0018, Batch: 000007, Sample Num: 112, Cur Loss: 0.52847230, Cur Avg Loss: 0.60789870, Log Avg loss: 0.57420287, Global Avg Loss: 1.36965780, Time: 0.0210 Steps: 36200, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000017, Sample Num: 272, Cur Loss: 0.34426081, Cur Avg Loss: 0.41827506, Log Avg loss: 0.28553851, Global Avg Loss: 1.36935840, Time: 0.0209 Steps: 36210, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000027, Sample Num: 432, Cur Loss: 1.11457801, Cur Avg Loss: 0.40961549, Log Avg loss: 0.39489422, Global Avg Loss: 1.36908936, Time: 0.0209 Steps: 36220, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000037, Sample Num: 592, Cur Loss: 0.26784185, Cur Avg Loss: 0.39900719, Log Avg loss: 0.37036479, Global Avg Loss: 1.36881370, Time: 0.0209 Steps: 36230, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000047, Sample Num: 752, Cur Loss: 1.56716084, Cur Avg Loss: 0.40391911, Log Avg loss: 0.42209321, Global Avg Loss: 1.36855246, Time: 0.0209 Steps: 36240, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000057, Sample Num: 912, Cur Loss: 0.31216249, Cur Avg Loss: 0.40032548, Log Avg loss: 0.38343545, Global Avg Loss: 1.36828071, Time: 0.0209 Steps: 36250, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000067, Sample Num: 1072, Cur Loss: 0.11725293, Cur Avg Loss: 0.39660087, Log Avg loss: 0.37537060, Global Avg Loss: 1.36800688, Time: 0.0209 Steps: 36260, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000077, Sample Num: 1232, Cur Loss: 0.11268299, Cur Avg Loss: 0.37456961, Log Avg loss: 0.22696018, Global Avg Loss: 1.36769228, Time: 0.0209 Steps: 36270, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000087, Sample Num: 1392, Cur Loss: 0.42842579, Cur Avg Loss: 0.36570006, Log Avg loss: 0.29740448, Global Avg Loss: 1.36739727, Time: 0.0209 Steps: 36280, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000097, Sample Num: 1552, Cur Loss: 0.25859690, Cur Avg Loss: 0.35807164, Log Avg loss: 0.29170434, Global Avg Loss: 1.36710085, Time: 0.0209 Steps: 36290, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000107, Sample Num: 1712, Cur Loss: 0.23053068, Cur Avg Loss: 0.34711380, Log Avg loss: 0.24082284, Global Avg Loss: 1.36679059, Time: 0.0209 Steps: 36300, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000117, Sample Num: 1872, Cur Loss: 0.27819902, Cur Avg Loss: 0.34482388, Log Avg loss: 0.32032171, Global Avg Loss: 1.36650238, Time: 0.0209 Steps: 36310, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000127, Sample Num: 2032, Cur Loss: 0.17367193, Cur Avg Loss: 0.33284414, Log Avg loss: 0.19268120, Global Avg Loss: 1.36617919, Time: 0.0209 Steps: 36320, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000137, Sample Num: 2192, Cur Loss: 0.26092839, Cur Avg Loss: 0.32618398, Log Avg loss: 0.24159993, Global Avg Loss: 1.36586965, Time: 0.0210 Steps: 36330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000147, Sample Num: 2352, Cur Loss: 0.21544942, Cur Avg Loss: 0.32466227, Log Avg loss: 0.30381485, Global Avg Loss: 1.36557739, Time: 0.0210 Steps: 36340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000157, Sample Num: 2512, Cur Loss: 0.43990684, Cur Avg Loss: 0.33016601, Log Avg loss: 0.41107101, Global Avg Loss: 1.36531480, Time: 0.0209 Steps: 36350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000167, Sample Num: 2672, Cur Loss: 0.15900846, Cur Avg Loss: 0.32272421, Log Avg loss: 0.20588799, Global Avg Loss: 1.36499593, Time: 0.0209 Steps: 36360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000177, Sample Num: 2832, Cur Loss: 0.23504759, Cur Avg Loss: 0.32933247, Log Avg loss: 0.43969030, Global Avg Loss: 1.36474152, Time: 0.0209 Steps: 36370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000187, Sample Num: 2992, Cur Loss: 0.58776700, Cur Avg Loss: 0.32492448, Log Avg loss: 0.24690317, Global Avg Loss: 1.36443425, Time: 0.0209 Steps: 36380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000197, Sample Num: 3152, Cur Loss: 0.28994116, Cur Avg Loss: 0.32367875, Log Avg loss: 0.30038348, Global Avg Loss: 1.36414185, Time: 0.0209 Steps: 36390, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000207, Sample Num: 3312, Cur Loss: 0.07880452, Cur Avg Loss: 0.31870147, Log Avg loss: 0.22064902, Global Avg Loss: 1.36382770, Time: 0.0209 Steps: 36400, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000217, Sample Num: 3472, Cur Loss: 0.43230766, Cur Avg Loss: 0.32089703, Log Avg loss: 0.36634515, Global Avg Loss: 1.36355374, Time: 0.0209 Steps: 36410, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000227, Sample Num: 3632, Cur Loss: 0.58789134, Cur Avg Loss: 0.32235629, Log Avg loss: 0.35402228, Global Avg Loss: 1.36327655, Time: 0.0209 Steps: 36420, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000237, Sample Num: 3792, Cur Loss: 0.30588955, Cur Avg Loss: 0.32933686, Log Avg loss: 0.48779574, Global Avg Loss: 1.36303623, Time: 0.0209 Steps: 36430, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000247, Sample Num: 3952, Cur Loss: 0.37091255, Cur Avg Loss: 0.34873731, Log Avg loss: 0.80852810, Global Avg Loss: 1.36288406, Time: 0.0209 Steps: 36440, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000257, Sample Num: 4112, Cur Loss: 0.28615814, Cur Avg Loss: 0.35115336, Log Avg loss: 0.41082978, Global Avg Loss: 1.36262287, Time: 0.0247 Steps: 36450, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000267, Sample Num: 4272, Cur Loss: 0.17439714, Cur Avg Loss: 0.35068237, Log Avg loss: 0.33857782, Global Avg Loss: 1.36234200, Time: 0.0210 Steps: 36460, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000277, Sample Num: 4432, Cur Loss: 0.24175289, Cur Avg Loss: 0.35035213, Log Avg loss: 0.34153477, Global Avg Loss: 1.36206210, Time: 0.0210 Steps: 36470, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000287, Sample Num: 4592, Cur Loss: 0.76371264, Cur Avg Loss: 0.35171301, Log Avg loss: 0.38940945, Global Avg Loss: 1.36179547, Time: 0.0209 Steps: 36480, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000297, Sample Num: 4752, Cur Loss: 0.14493635, Cur Avg Loss: 0.34859357, Log Avg loss: 0.25906565, Global Avg Loss: 1.36149327, Time: 0.0211 Steps: 36490, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000307, Sample Num: 4912, Cur Loss: 0.41219515, Cur Avg Loss: 0.34620461, Log Avg loss: 0.27525245, Global Avg Loss: 1.36119567, Time: 0.0212 Steps: 36500, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000317, Sample Num: 5072, Cur Loss: 0.13265857, Cur Avg Loss: 0.34220547, Log Avg loss: 0.21943180, Global Avg Loss: 1.36088294, Time: 0.0212 Steps: 36510, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000327, Sample Num: 5232, Cur Loss: 0.30873778, Cur Avg Loss: 0.34292571, Log Avg loss: 0.36575731, Global Avg Loss: 1.36061045, Time: 0.0212 Steps: 36520, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000337, Sample Num: 5392, Cur Loss: 0.24407731, Cur Avg Loss: 0.34180480, Log Avg loss: 0.30515108, Global Avg Loss: 1.36032152, Time: 0.0211 Steps: 36530, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000347, Sample Num: 5552, Cur Loss: 0.26401424, Cur Avg Loss: 0.34052656, Log Avg loss: 0.29745005, Global Avg Loss: 1.36003065, Time: 0.0210 Steps: 36540, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000357, Sample Num: 5712, Cur Loss: 0.24847156, Cur Avg Loss: 0.33757192, Log Avg loss: 0.23504563, Global Avg Loss: 1.35972285, Time: 0.0212 Steps: 36550, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000367, Sample Num: 5872, Cur Loss: 0.22197776, Cur Avg Loss: 0.33609508, Log Avg loss: 0.28337208, Global Avg Loss: 1.35942845, Time: 0.0209 Steps: 36560, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000377, Sample Num: 6032, Cur Loss: 0.32531008, Cur Avg Loss: 0.33957701, Log Avg loss: 0.46736384, Global Avg Loss: 1.35918451, Time: 0.0210 Steps: 36570, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000387, Sample Num: 6192, Cur Loss: 0.51394069, Cur Avg Loss: 0.33966493, Log Avg loss: 0.34297938, Global Avg Loss: 1.35890671, Time: 0.0209 Steps: 36580, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000397, Sample Num: 6352, Cur Loss: 0.26266468, Cur Avg Loss: 0.33905513, Log Avg loss: 0.31545606, Global Avg Loss: 1.35862153, Time: 0.0208 Steps: 36590, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000407, Sample Num: 6512, Cur Loss: 1.29116821, Cur Avg Loss: 0.34256517, Log Avg loss: 0.48191381, Global Avg Loss: 1.35838200, Time: 0.0211 Steps: 36600, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000417, Sample Num: 6672, Cur Loss: 0.51267201, Cur Avg Loss: 0.34412495, Log Avg loss: 0.40760766, Global Avg Loss: 1.35812229, Time: 0.0211 Steps: 36610, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000427, Sample Num: 6832, Cur Loss: 1.07197022, Cur Avg Loss: 0.34868216, Log Avg loss: 0.53871778, Global Avg Loss: 1.35789853, Time: 0.0210 Steps: 36620, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000437, Sample Num: 6992, Cur Loss: 0.18091357, Cur Avg Loss: 0.34902493, Log Avg loss: 0.36366126, Global Avg Loss: 1.35762711, Time: 0.0211 Steps: 36630, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000447, Sample Num: 7152, Cur Loss: 0.11631497, Cur Avg Loss: 0.34595592, Log Avg loss: 0.21184042, Global Avg Loss: 1.35731439, Time: 0.0211 Steps: 36640, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000457, Sample Num: 7312, Cur Loss: 0.10319829, Cur Avg Loss: 0.34412407, Log Avg loss: 0.26224035, Global Avg Loss: 1.35701560, Time: 0.0211 Steps: 36650, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000467, Sample Num: 7472, Cur Loss: 0.23914722, Cur Avg Loss: 0.34165148, Log Avg loss: 0.22865407, Global Avg Loss: 1.35670781, Time: 0.0210 Steps: 36660, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000477, Sample Num: 7632, Cur Loss: 0.26480526, Cur Avg Loss: 0.34282979, Log Avg loss: 0.39785705, Global Avg Loss: 1.35644633, Time: 0.0209 Steps: 36670, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000487, Sample Num: 7792, Cur Loss: 1.27731204, Cur Avg Loss: 0.34856861, Log Avg loss: 0.62231018, Global Avg Loss: 1.35624618, Time: 0.0209 Steps: 36680, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000497, Sample Num: 7952, Cur Loss: 0.32525253, Cur Avg Loss: 0.34832499, Log Avg loss: 0.33646065, Global Avg Loss: 1.35596824, Time: 0.0209 Steps: 36690, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000507, Sample Num: 8112, Cur Loss: 0.35143590, Cur Avg Loss: 0.34604971, Log Avg loss: 0.23296831, Global Avg Loss: 1.35566224, Time: 0.0209 Steps: 36700, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000517, Sample Num: 8272, Cur Loss: 0.14612560, Cur Avg Loss: 0.34541118, Log Avg loss: 0.31303765, Global Avg Loss: 1.35537823, Time: 0.0212 Steps: 36710, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000527, Sample Num: 8432, Cur Loss: 0.10517067, Cur Avg Loss: 0.34207839, Log Avg loss: 0.16977324, Global Avg Loss: 1.35505535, Time: 0.0212 Steps: 36720, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000537, Sample Num: 8592, Cur Loss: 0.42421594, Cur Avg Loss: 0.34005761, Log Avg loss: 0.23356229, Global Avg Loss: 1.35475001, Time: 0.0211 Steps: 36730, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000547, Sample Num: 8752, Cur Loss: 0.41034228, Cur Avg Loss: 0.33717652, Log Avg loss: 0.18246190, Global Avg Loss: 1.35443094, Time: 0.0212 Steps: 36740, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000557, Sample Num: 8912, Cur Loss: 0.95299351, Cur Avg Loss: 0.33769397, Log Avg loss: 0.36599855, Global Avg Loss: 1.35416198, Time: 0.0212 Steps: 36750, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000567, Sample Num: 9072, Cur Loss: 0.16057992, Cur Avg Loss: 0.33547324, Log Avg loss: 0.21177886, Global Avg Loss: 1.35385121, Time: 0.0211 Steps: 36760, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000577, Sample Num: 9232, Cur Loss: 0.14051758, Cur Avg Loss: 0.33369086, Log Avg loss: 0.23262963, Global Avg Loss: 1.35354628, Time: 0.0219 Steps: 36770, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000587, Sample Num: 9392, Cur Loss: 0.49351728, Cur Avg Loss: 0.33603232, Log Avg loss: 0.47113480, Global Avg Loss: 1.35330636, Time: 0.0212 Steps: 36780, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000597, Sample Num: 9552, Cur Loss: 0.31757683, Cur Avg Loss: 0.33416622, Log Avg loss: 0.22462606, Global Avg Loss: 1.35299957, Time: 0.0212 Steps: 36790, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000607, Sample Num: 9712, Cur Loss: 0.46800750, Cur Avg Loss: 0.33340050, Log Avg loss: 0.28768716, Global Avg Loss: 1.35271009, Time: 0.0212 Steps: 36800, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000617, Sample Num: 9872, Cur Loss: 0.87744498, Cur Avg Loss: 0.33710108, Log Avg loss: 0.56172613, Global Avg Loss: 1.35249520, Time: 0.0212 Steps: 36810, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000627, Sample Num: 10032, Cur Loss: 0.63560766, Cur Avg Loss: 0.33710213, Log Avg loss: 0.33716726, Global Avg Loss: 1.35221945, Time: 0.0211 Steps: 36820, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000637, Sample Num: 10192, Cur Loss: 0.44765043, Cur Avg Loss: 0.33691222, Log Avg loss: 0.32500475, Global Avg Loss: 1.35194054, Time: 0.0212 Steps: 36830, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000647, Sample Num: 10352, Cur Loss: 0.42769325, Cur Avg Loss: 0.33747596, Log Avg loss: 0.37338634, Global Avg Loss: 1.35167492, Time: 0.0220 Steps: 36840, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000657, Sample Num: 10512, Cur Loss: 0.29521209, Cur Avg Loss: 0.33781768, Log Avg loss: 0.35992668, Global Avg Loss: 1.35140579, Time: 0.0211 Steps: 36850, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000667, Sample Num: 10672, Cur Loss: 0.63934231, Cur Avg Loss: 0.34155839, Log Avg loss: 0.58732273, Global Avg Loss: 1.35119849, Time: 0.0220 Steps: 36860, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000677, Sample Num: 10832, Cur Loss: 0.66354358, Cur Avg Loss: 0.34551745, Log Avg loss: 0.60958686, Global Avg Loss: 1.35099735, Time: 0.0219 Steps: 36870, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000687, Sample Num: 10992, Cur Loss: 0.62487125, Cur Avg Loss: 0.34779321, Log Avg loss: 0.50186241, Global Avg Loss: 1.35076711, Time: 0.0220 Steps: 36880, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000697, Sample Num: 11152, Cur Loss: 0.31262180, Cur Avg Loss: 0.34958136, Log Avg loss: 0.47242712, Global Avg Loss: 1.35052901, Time: 0.0211 Steps: 36890, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000707, Sample Num: 11312, Cur Loss: 0.55621141, Cur Avg Loss: 0.34924132, Log Avg loss: 0.32554096, Global Avg Loss: 1.35025124, Time: 0.0220 Steps: 36900, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000717, Sample Num: 11472, Cur Loss: 0.23645234, Cur Avg Loss: 0.34809069, Log Avg loss: 0.26674104, Global Avg Loss: 1.34995768, Time: 0.0212 Steps: 36910, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000727, Sample Num: 11632, Cur Loss: 0.34683177, Cur Avg Loss: 0.34714406, Log Avg loss: 0.27927064, Global Avg Loss: 1.34966768, Time: 0.0220 Steps: 36920, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000737, Sample Num: 11792, Cur Loss: 0.24199155, Cur Avg Loss: 0.34577590, Log Avg loss: 0.24631059, Global Avg Loss: 1.34936891, Time: 0.0211 Steps: 36930, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000747, Sample Num: 11952, Cur Loss: 0.15733737, Cur Avg Loss: 0.34448994, Log Avg loss: 0.24971438, Global Avg Loss: 1.34907122, Time: 0.0215 Steps: 36940, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000757, Sample Num: 12112, Cur Loss: 0.22290011, Cur Avg Loss: 0.34302058, Log Avg loss: 0.23325973, Global Avg Loss: 1.34876925, Time: 0.0223 Steps: 36950, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000767, Sample Num: 12272, Cur Loss: 0.17645858, Cur Avg Loss: 0.34308755, Log Avg loss: 0.34815733, Global Avg Loss: 1.34849852, Time: 0.0214 Steps: 36960, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000777, Sample Num: 12432, Cur Loss: 0.16470571, Cur Avg Loss: 0.34298806, Log Avg loss: 0.33535673, Global Avg Loss: 1.34822447, Time: 0.0213 Steps: 36970, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000787, Sample Num: 12592, Cur Loss: 0.44905776, Cur Avg Loss: 0.34250499, Log Avg loss: 0.30497093, Global Avg Loss: 1.34794236, Time: 0.0212 Steps: 36980, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000797, Sample Num: 12752, Cur Loss: 0.45391822, Cur Avg Loss: 0.34111761, Log Avg loss: 0.23193081, Global Avg Loss: 1.34764065, Time: 0.0210 Steps: 36990, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000807, Sample Num: 12912, Cur Loss: 0.34876955, Cur Avg Loss: 0.34086165, Log Avg loss: 0.32046136, Global Avg Loss: 1.34736304, Time: 0.0210 Steps: 37000, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000817, Sample Num: 13072, Cur Loss: 0.30358607, Cur Avg Loss: 0.33960346, Log Avg loss: 0.23806731, Global Avg Loss: 1.34706331, Time: 0.0208 Steps: 37010, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000827, Sample Num: 13232, Cur Loss: 0.75976419, Cur Avg Loss: 0.34056548, Log Avg loss: 0.41916297, Global Avg Loss: 1.34681266, Time: 0.0209 Steps: 37020, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000837, Sample Num: 13392, Cur Loss: 0.39969885, Cur Avg Loss: 0.34113067, Log Avg loss: 0.38787182, Global Avg Loss: 1.34655370, Time: 0.0209 Steps: 37030, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000847, Sample Num: 13552, Cur Loss: 0.56651938, Cur Avg Loss: 0.34047162, Log Avg loss: 0.28530855, Global Avg Loss: 1.34626718, Time: 0.0208 Steps: 37040, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000857, Sample Num: 13712, Cur Loss: 0.43998992, Cur Avg Loss: 0.33948606, Log Avg loss: 0.25600919, Global Avg Loss: 1.34597292, Time: 0.0208 Steps: 37050, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000867, Sample Num: 13872, Cur Loss: 0.20948890, Cur Avg Loss: 0.34046645, Log Avg loss: 0.42448641, Global Avg Loss: 1.34572427, Time: 0.0209 Steps: 37060, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000877, Sample Num: 14032, Cur Loss: 0.28558260, Cur Avg Loss: 0.34026893, Log Avg loss: 0.32314368, Global Avg Loss: 1.34544842, Time: 0.0210 Steps: 37070, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000887, Sample Num: 14192, Cur Loss: 0.76588255, Cur Avg Loss: 0.34047166, Log Avg loss: 0.35825117, Global Avg Loss: 1.34518219, Time: 0.0208 Steps: 37080, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000897, Sample Num: 14352, Cur Loss: 0.46896428, Cur Avg Loss: 0.34051133, Log Avg loss: 0.34403007, Global Avg Loss: 1.34491226, Time: 0.0208 Steps: 37090, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000907, Sample Num: 14512, Cur Loss: 0.31046832, Cur Avg Loss: 0.34045100, Log Avg loss: 0.33503959, Global Avg Loss: 1.34464006, Time: 0.0208 Steps: 37100, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000917, Sample Num: 14672, Cur Loss: 0.22557160, Cur Avg Loss: 0.33989022, Log Avg loss: 0.28902697, Global Avg Loss: 1.34435560, Time: 0.0209 Steps: 37110, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000927, Sample Num: 14832, Cur Loss: 0.38080189, Cur Avg Loss: 0.34043320, Log Avg loss: 0.39022446, Global Avg Loss: 1.34409856, Time: 0.0208 Steps: 37120, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000937, Sample Num: 14992, Cur Loss: 0.41554224, Cur Avg Loss: 0.34139365, Log Avg loss: 0.43042715, Global Avg Loss: 1.34385249, Time: 0.0209 Steps: 37130, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000947, Sample Num: 15152, Cur Loss: 0.20366010, Cur Avg Loss: 0.34500789, Log Avg loss: 0.68366222, Global Avg Loss: 1.34367473, Time: 0.0208 Steps: 37140, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000957, Sample Num: 15312, Cur Loss: 0.31999385, Cur Avg Loss: 0.34642708, Log Avg loss: 0.48082506, Global Avg Loss: 1.34344247, Time: 0.0209 Steps: 37150, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000967, Sample Num: 15472, Cur Loss: 0.39797026, Cur Avg Loss: 0.34621841, Log Avg loss: 0.32624811, Global Avg Loss: 1.34316874, Time: 0.0208 Steps: 37160, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000977, Sample Num: 15632, Cur Loss: 0.24517757, Cur Avg Loss: 0.34687833, Log Avg loss: 0.41069314, Global Avg Loss: 1.34291787, Time: 0.0208 Steps: 37170, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000987, Sample Num: 15792, Cur Loss: 0.50348115, Cur Avg Loss: 0.34768483, Log Avg loss: 0.42647990, Global Avg Loss: 1.34267138, Time: 0.0209 Steps: 37180, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000997, Sample Num: 15952, Cur Loss: 0.18431671, Cur Avg Loss: 0.34651853, Log Avg loss: 0.23140405, Global Avg Loss: 1.34237257, Time: 0.0208 Steps: 37190, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001007, Sample Num: 16112, Cur Loss: 0.24246043, Cur Avg Loss: 0.34630438, Log Avg loss: 0.32495429, Global Avg Loss: 1.34209907, Time: 0.0208 Steps: 37200, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001017, Sample Num: 16272, Cur Loss: 0.48457539, Cur Avg Loss: 0.34832271, Log Avg loss: 0.55156803, Global Avg Loss: 1.34188662, Time: 0.0208 Steps: 37210, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001027, Sample Num: 16432, Cur Loss: 0.48550478, Cur Avg Loss: 0.34854238, Log Avg loss: 0.37088261, Global Avg Loss: 1.34162574, Time: 0.0246 Steps: 37220, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001037, Sample Num: 16592, Cur Loss: 0.67127240, Cur Avg Loss: 0.34800350, Log Avg loss: 0.29266054, Global Avg Loss: 1.34134399, Time: 0.0209 Steps: 37230, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001047, Sample Num: 16752, Cur Loss: 0.74017835, Cur Avg Loss: 0.34976438, Log Avg loss: 0.53236811, Global Avg Loss: 1.34112676, Time: 0.0209 Steps: 37240, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001057, Sample Num: 16912, Cur Loss: 0.52003509, Cur Avg Loss: 0.35087571, Log Avg loss: 0.46723160, Global Avg Loss: 1.34089215, Time: 0.0209 Steps: 37250, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001067, Sample Num: 17072, Cur Loss: 0.17267461, Cur Avg Loss: 0.34929597, Log Avg loss: 0.18231793, Global Avg Loss: 1.34058121, Time: 0.0209 Steps: 37260, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001077, Sample Num: 17232, Cur Loss: 0.09119476, Cur Avg Loss: 0.34843016, Log Avg loss: 0.25604828, Global Avg Loss: 1.34029022, Time: 0.0209 Steps: 37270, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001087, Sample Num: 17392, Cur Loss: 0.26196870, Cur Avg Loss: 0.34810469, Log Avg loss: 0.31305186, Global Avg Loss: 1.34001467, Time: 0.0209 Steps: 37280, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001097, Sample Num: 17552, Cur Loss: 0.16915056, Cur Avg Loss: 0.34690222, Log Avg loss: 0.21619285, Global Avg Loss: 1.33971330, Time: 0.0209 Steps: 37290, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001107, Sample Num: 17712, Cur Loss: 0.32051575, Cur Avg Loss: 0.34699009, Log Avg loss: 0.35662965, Global Avg Loss: 1.33944973, Time: 0.0209 Steps: 37300, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001117, Sample Num: 17872, Cur Loss: 0.48673448, Cur Avg Loss: 0.34840000, Log Avg loss: 0.50447762, Global Avg Loss: 1.33922594, Time: 0.0209 Steps: 37310, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001127, Sample Num: 18032, Cur Loss: 0.54596955, Cur Avg Loss: 0.34970394, Log Avg loss: 0.49535358, Global Avg Loss: 1.33899982, Time: 0.0209 Steps: 37320, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001137, Sample Num: 18192, Cur Loss: 0.28284925, Cur Avg Loss: 0.35258385, Log Avg loss: 0.67714992, Global Avg Loss: 1.33882253, Time: 0.0209 Steps: 37330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001147, Sample Num: 18352, Cur Loss: 0.20325018, Cur Avg Loss: 0.35203498, Log Avg loss: 0.28962828, Global Avg Loss: 1.33854154, Time: 0.0208 Steps: 37340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001157, Sample Num: 18512, Cur Loss: 0.42190069, Cur Avg Loss: 0.35204853, Log Avg loss: 0.35360297, Global Avg Loss: 1.33827784, Time: 0.0209 Steps: 37350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001167, Sample Num: 18672, Cur Loss: 1.07024252, Cur Avg Loss: 0.35203428, Log Avg loss: 0.35038544, Global Avg Loss: 1.33801341, Time: 0.0209 Steps: 37360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001177, Sample Num: 18832, Cur Loss: 0.58165103, Cur Avg Loss: 0.35117084, Log Avg loss: 0.25040683, Global Avg Loss: 1.33772237, Time: 0.0209 Steps: 37370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001187, Sample Num: 18992, Cur Loss: 0.27965096, Cur Avg Loss: 0.35173739, Log Avg loss: 0.41842025, Global Avg Loss: 1.33747644, Time: 0.0209 Steps: 37380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001197, Sample Num: 19152, Cur Loss: 0.46680498, Cur Avg Loss: 0.35198956, Log Avg loss: 0.38192227, Global Avg Loss: 1.33722088, Time: 0.0209 Steps: 37390, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001207, Sample Num: 19312, Cur Loss: 0.18071580, Cur Avg Loss: 0.35162790, Log Avg loss: 0.30833798, Global Avg Loss: 1.33694577, Time: 0.0209 Steps: 37400, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001217, Sample Num: 19472, Cur Loss: 0.14507823, Cur Avg Loss: 0.35138159, Log Avg loss: 0.32165152, Global Avg Loss: 1.33667438, Time: 0.0209 Steps: 37410, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001227, Sample Num: 19632, Cur Loss: 0.24377638, Cur Avg Loss: 0.35100863, Log Avg loss: 0.30561910, Global Avg Loss: 1.33639884, Time: 0.0208 Steps: 37420, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001237, Sample Num: 19792, Cur Loss: 0.19192533, Cur Avg Loss: 0.35120763, Log Avg loss: 0.37562485, Global Avg Loss: 1.33614216, Time: 0.0209 Steps: 37430, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001247, Sample Num: 19952, Cur Loss: 0.09808803, Cur Avg Loss: 0.35189582, Log Avg loss: 0.43702562, Global Avg Loss: 1.33590201, Time: 0.0209 Steps: 37440, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001257, Sample Num: 20112, Cur Loss: 0.25040725, Cur Avg Loss: 0.35125257, Log Avg loss: 0.27103866, Global Avg Loss: 1.33561766, Time: 0.0209 Steps: 37450, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001267, Sample Num: 20272, Cur Loss: 0.33212358, Cur Avg Loss: 0.35027226, Log Avg loss: 0.22704811, Global Avg Loss: 1.33532173, Time: 0.0209 Steps: 37460, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001277, Sample Num: 20432, Cur Loss: 0.40497267, Cur Avg Loss: 0.35054139, Log Avg loss: 0.38463960, Global Avg Loss: 1.33506801, Time: 0.0209 Steps: 37470, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001287, Sample Num: 20592, Cur Loss: 0.68038023, Cur Avg Loss: 0.35052203, Log Avg loss: 0.34804922, Global Avg Loss: 1.33480467, Time: 0.0209 Steps: 37480, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001297, Sample Num: 20752, Cur Loss: 0.48558369, Cur Avg Loss: 0.34978634, Log Avg loss: 0.25510322, Global Avg Loss: 1.33451667, Time: 0.0209 Steps: 37490, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001307, Sample Num: 20912, Cur Loss: 0.26974469, Cur Avg Loss: 0.34958888, Log Avg loss: 0.32397836, Global Avg Loss: 1.33424719, Time: 0.0209 Steps: 37500, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001317, Sample Num: 21072, Cur Loss: 0.31573412, Cur Avg Loss: 0.34983990, Log Avg loss: 0.38264825, Global Avg Loss: 1.33399350, Time: 0.0208 Steps: 37510, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001327, Sample Num: 21232, Cur Loss: 0.51521808, Cur Avg Loss: 0.35055141, Log Avg loss: 0.44425765, Global Avg Loss: 1.33375636, Time: 0.0219 Steps: 37520, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001337, Sample Num: 21392, Cur Loss: 0.63439226, Cur Avg Loss: 0.35022002, Log Avg loss: 0.30624490, Global Avg Loss: 1.33348258, Time: 0.0209 Steps: 37530, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001347, Sample Num: 21552, Cur Loss: 0.14269999, Cur Avg Loss: 0.35013402, Log Avg loss: 0.33863578, Global Avg Loss: 1.33321757, Time: 0.0210 Steps: 37540, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001357, Sample Num: 21712, Cur Loss: 0.20552212, Cur Avg Loss: 0.34971252, Log Avg loss: 0.29293601, Global Avg Loss: 1.33294053, Time: 0.0210 Steps: 37550, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001367, Sample Num: 21872, Cur Loss: 0.14362189, Cur Avg Loss: 0.34956837, Log Avg loss: 0.33000749, Global Avg Loss: 1.33267351, Time: 0.0209 Steps: 37560, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001377, Sample Num: 22032, Cur Loss: 0.07357669, Cur Avg Loss: 0.34895285, Log Avg loss: 0.26481118, Global Avg Loss: 1.33238928, Time: 0.0211 Steps: 37570, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001387, Sample Num: 22192, Cur Loss: 0.21911860, Cur Avg Loss: 0.34893071, Log Avg loss: 0.34588132, Global Avg Loss: 1.33212677, Time: 0.0211 Steps: 37580, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001397, Sample Num: 22352, Cur Loss: 0.15568542, Cur Avg Loss: 0.34859048, Log Avg loss: 0.30140105, Global Avg Loss: 1.33185257, Time: 0.0210 Steps: 37590, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001407, Sample Num: 22512, Cur Loss: 0.10746756, Cur Avg Loss: 0.34799452, Log Avg loss: 0.26473977, Global Avg Loss: 1.33156876, Time: 0.0209 Steps: 37600, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001417, Sample Num: 22672, Cur Loss: 0.18729332, Cur Avg Loss: 0.34812169, Log Avg loss: 0.36601318, Global Avg Loss: 1.33131203, Time: 0.0209 Steps: 37610, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001427, Sample Num: 22832, Cur Loss: 0.04667547, Cur Avg Loss: 0.34842632, Log Avg loss: 0.39159246, Global Avg Loss: 1.33106224, Time: 0.0209 Steps: 37620, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001437, Sample Num: 22992, Cur Loss: 0.21579316, Cur Avg Loss: 0.34823637, Log Avg loss: 0.32113129, Global Avg Loss: 1.33079385, Time: 0.0210 Steps: 37630, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001447, Sample Num: 23152, Cur Loss: 0.20093125, Cur Avg Loss: 0.34711099, Log Avg loss: 0.18539393, Global Avg Loss: 1.33048955, Time: 0.0209 Steps: 37640, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001457, Sample Num: 23312, Cur Loss: 0.23109899, Cur Avg Loss: 0.34711826, Log Avg loss: 0.34817028, Global Avg Loss: 1.33022864, Time: 0.0210 Steps: 37650, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001467, Sample Num: 23472, Cur Loss: 0.28192642, Cur Avg Loss: 0.34683858, Log Avg loss: 0.30608898, Global Avg Loss: 1.32995670, Time: 0.0210 Steps: 37660, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001477, Sample Num: 23632, Cur Loss: 0.41356456, Cur Avg Loss: 0.34712792, Log Avg loss: 0.38957357, Global Avg Loss: 1.32970706, Time: 0.0209 Steps: 37670, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001487, Sample Num: 23792, Cur Loss: 0.57645965, Cur Avg Loss: 0.34738387, Log Avg loss: 0.38518780, Global Avg Loss: 1.32945639, Time: 0.0208 Steps: 37680, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001497, Sample Num: 23952, Cur Loss: 0.52482355, Cur Avg Loss: 0.34857600, Log Avg loss: 0.52584611, Global Avg Loss: 1.32924318, Time: 0.0209 Steps: 37690, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001507, Sample Num: 24112, Cur Loss: 0.13238966, Cur Avg Loss: 0.34813954, Log Avg loss: 0.28280078, Global Avg Loss: 1.32896561, Time: 0.0208 Steps: 37700, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001517, Sample Num: 24272, Cur Loss: 0.14476602, Cur Avg Loss: 0.34776468, Log Avg loss: 0.29127350, Global Avg Loss: 1.32869043, Time: 0.0209 Steps: 37710, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001527, Sample Num: 24432, Cur Loss: 0.24414833, Cur Avg Loss: 0.34755380, Log Avg loss: 0.31556320, Global Avg Loss: 1.32842184, Time: 0.0209 Steps: 37720, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001537, Sample Num: 24592, Cur Loss: 0.12368959, Cur Avg Loss: 0.34740458, Log Avg loss: 0.32461949, Global Avg Loss: 1.32815579, Time: 0.0245 Steps: 37730, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001547, Sample Num: 24752, Cur Loss: 0.14231899, Cur Avg Loss: 0.34705072, Log Avg loss: 0.29266185, Global Avg Loss: 1.32788141, Time: 0.0209 Steps: 37740, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001557, Sample Num: 24912, Cur Loss: 0.21632341, Cur Avg Loss: 0.34754583, Log Avg loss: 0.42413996, Global Avg Loss: 1.32764201, Time: 0.0209 Steps: 37750, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001567, Sample Num: 25072, Cur Loss: 0.13730407, Cur Avg Loss: 0.34766014, Log Avg loss: 0.36545769, Global Avg Loss: 1.32738720, Time: 0.0209 Steps: 37760, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001577, Sample Num: 25232, Cur Loss: 0.39480555, Cur Avg Loss: 0.34818736, Log Avg loss: 0.43080251, Global Avg Loss: 1.32714982, Time: 0.0208 Steps: 37770, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001587, Sample Num: 25392, Cur Loss: 0.19953683, Cur Avg Loss: 0.34771718, Log Avg loss: 0.27357067, Global Avg Loss: 1.32687094, Time: 0.0209 Steps: 37780, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001597, Sample Num: 25552, Cur Loss: 0.46372458, Cur Avg Loss: 0.34729259, Log Avg loss: 0.27990992, Global Avg Loss: 1.32659390, Time: 0.0208 Steps: 37790, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001607, Sample Num: 25712, Cur Loss: 0.24313229, Cur Avg Loss: 0.34745779, Log Avg loss: 0.37384046, Global Avg Loss: 1.32634184, Time: 0.0209 Steps: 37800, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001617, Sample Num: 25872, Cur Loss: 0.07582779, Cur Avg Loss: 0.34722624, Log Avg loss: 0.31001626, Global Avg Loss: 1.32607305, Time: 0.0209 Steps: 37810, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001627, Sample Num: 26032, Cur Loss: 0.14237285, Cur Avg Loss: 0.34677639, Log Avg loss: 0.27403573, Global Avg Loss: 1.32579488, Time: 0.0209 Steps: 37820, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001637, Sample Num: 26192, Cur Loss: 0.31546164, Cur Avg Loss: 0.34673919, Log Avg loss: 0.34068668, Global Avg Loss: 1.32553447, Time: 0.0208 Steps: 37830, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001647, Sample Num: 26352, Cur Loss: 0.43960202, Cur Avg Loss: 0.34593514, Log Avg loss: 0.21431138, Global Avg Loss: 1.32524081, Time: 0.0210 Steps: 37840, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001657, Sample Num: 26512, Cur Loss: 0.12630937, Cur Avg Loss: 0.34545988, Log Avg loss: 0.26718404, Global Avg Loss: 1.32496127, Time: 0.0209 Steps: 37850, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001667, Sample Num: 26672, Cur Loss: 0.15106352, Cur Avg Loss: 0.34468993, Log Avg loss: 0.21710995, Global Avg Loss: 1.32466865, Time: 0.0210 Steps: 37860, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001677, Sample Num: 26832, Cur Loss: 0.09012720, Cur Avg Loss: 0.34398670, Log Avg loss: 0.22675872, Global Avg Loss: 1.32437874, Time: 0.0209 Steps: 37870, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001687, Sample Num: 26992, Cur Loss: 0.15712596, Cur Avg Loss: 0.34388651, Log Avg loss: 0.32708364, Global Avg Loss: 1.32411546, Time: 0.0210 Steps: 37880, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001697, Sample Num: 27152, Cur Loss: 0.43554699, Cur Avg Loss: 0.34391126, Log Avg loss: 0.34808773, Global Avg Loss: 1.32385786, Time: 0.0208 Steps: 37890, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001707, Sample Num: 27312, Cur Loss: 0.28743795, Cur Avg Loss: 0.34391641, Log Avg loss: 0.34478937, Global Avg Loss: 1.32359953, Time: 0.0209 Steps: 37900, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001717, Sample Num: 27472, Cur Loss: 0.11989179, Cur Avg Loss: 0.34384568, Log Avg loss: 0.33177269, Global Avg Loss: 1.32333791, Time: 0.0209 Steps: 37910, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001727, Sample Num: 27632, Cur Loss: 0.71329719, Cur Avg Loss: 0.34405164, Log Avg loss: 0.37941523, Global Avg Loss: 1.32308898, Time: 0.0209 Steps: 37920, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001737, Sample Num: 27792, Cur Loss: 0.20376107, Cur Avg Loss: 0.34382748, Log Avg loss: 0.30511449, Global Avg Loss: 1.32282060, Time: 0.0210 Steps: 37930, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001747, Sample Num: 27952, Cur Loss: 0.25782850, Cur Avg Loss: 0.34330059, Log Avg loss: 0.25177994, Global Avg Loss: 1.32253830, Time: 0.0207 Steps: 37940, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001757, Sample Num: 28112, Cur Loss: 0.22787985, Cur Avg Loss: 0.34291356, Log Avg loss: 0.27529927, Global Avg Loss: 1.32226235, Time: 0.0208 Steps: 37950, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001767, Sample Num: 28272, Cur Loss: 0.26915380, Cur Avg Loss: 0.34275649, Log Avg loss: 0.31515855, Global Avg Loss: 1.32199704, Time: 0.0211 Steps: 37960, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001777, Sample Num: 28432, Cur Loss: 0.13594599, Cur Avg Loss: 0.34273637, Log Avg loss: 0.33918146, Global Avg Loss: 1.32173820, Time: 0.0210 Steps: 37970, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001787, Sample Num: 28592, Cur Loss: 0.27511936, Cur Avg Loss: 0.34349214, Log Avg loss: 0.47779299, Global Avg Loss: 1.32151600, Time: 0.0208 Steps: 37980, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001797, Sample Num: 28752, Cur Loss: 0.53740025, Cur Avg Loss: 0.34412417, Log Avg loss: 0.45706809, Global Avg Loss: 1.32128845, Time: 0.0216 Steps: 37990, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001807, Sample Num: 28912, Cur Loss: 0.48719695, Cur Avg Loss: 0.34490464, Log Avg loss: 0.48515560, Global Avg Loss: 1.32106841, Time: 0.0208 Steps: 38000, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001817, Sample Num: 29072, Cur Loss: 0.30907100, Cur Avg Loss: 0.34521093, Log Avg loss: 0.40055700, Global Avg Loss: 1.32082624, Time: 0.0208 Steps: 38010, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001827, Sample Num: 29232, Cur Loss: 0.12688628, Cur Avg Loss: 0.34476814, Log Avg loss: 0.26431305, Global Avg Loss: 1.32054836, Time: 0.0207 Steps: 38020, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001837, Sample Num: 29392, Cur Loss: 0.63166648, Cur Avg Loss: 0.34514888, Log Avg loss: 0.41471093, Global Avg Loss: 1.32031016, Time: 0.0208 Steps: 38030, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001847, Sample Num: 29552, Cur Loss: 0.47931242, Cur Avg Loss: 0.34533912, Log Avg loss: 0.38028582, Global Avg Loss: 1.32006305, Time: 0.0208 Steps: 38040, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001857, Sample Num: 29712, Cur Loss: 0.34219754, Cur Avg Loss: 0.34507707, Log Avg loss: 0.29667542, Global Avg Loss: 1.31979409, Time: 0.0208 Steps: 38050, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001867, Sample Num: 29872, Cur Loss: 0.07600731, Cur Avg Loss: 0.34461945, Log Avg loss: 0.25963962, Global Avg Loss: 1.31951554, Time: 0.0208 Steps: 38060, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001877, Sample Num: 30032, Cur Loss: 0.14684644, Cur Avg Loss: 0.34484819, Log Avg loss: 0.38755489, Global Avg Loss: 1.31927074, Time: 0.0208 Steps: 38070, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001887, Sample Num: 30192, Cur Loss: 0.59769326, Cur Avg Loss: 0.34609735, Log Avg loss: 0.58056483, Global Avg Loss: 1.31907675, Time: 0.0207 Steps: 38080, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001897, Sample Num: 30352, Cur Loss: 0.51543587, Cur Avg Loss: 0.34682675, Log Avg loss: 0.48446291, Global Avg Loss: 1.31885764, Time: 0.0208 Steps: 38090, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001907, Sample Num: 30512, Cur Loss: 0.55771405, Cur Avg Loss: 0.34725079, Log Avg loss: 0.42769236, Global Avg Loss: 1.31862374, Time: 0.0208 Steps: 38100, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001917, Sample Num: 30672, Cur Loss: 0.35494310, Cur Avg Loss: 0.34716135, Log Avg loss: 0.33010514, Global Avg Loss: 1.31836435, Time: 0.0208 Steps: 38110, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001927, Sample Num: 30832, Cur Loss: 0.53284580, Cur Avg Loss: 0.34758334, Log Avg loss: 0.42847824, Global Avg Loss: 1.31813091, Time: 0.0208 Steps: 38120, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001937, Sample Num: 30992, Cur Loss: 0.26824781, Cur Avg Loss: 0.34724465, Log Avg loss: 0.28197869, Global Avg Loss: 1.31785916, Time: 0.0209 Steps: 38130, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001947, Sample Num: 31152, Cur Loss: 0.49802780, Cur Avg Loss: 0.34699965, Log Avg loss: 0.29954443, Global Avg Loss: 1.31759217, Time: 0.0208 Steps: 38140, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001957, Sample Num: 31312, Cur Loss: 0.08671757, Cur Avg Loss: 0.34875817, Log Avg loss: 0.69114174, Global Avg Loss: 1.31742796, Time: 0.0208 Steps: 38150, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001967, Sample Num: 31472, Cur Loss: 0.19781220, Cur Avg Loss: 0.34845810, Log Avg loss: 0.28973373, Global Avg Loss: 1.31715865, Time: 0.0208 Steps: 38160, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001977, Sample Num: 31632, Cur Loss: 0.27981237, Cur Avg Loss: 0.34812174, Log Avg loss: 0.28196080, Global Avg Loss: 1.31688744, Time: 0.0208 Steps: 38170, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001987, Sample Num: 31792, Cur Loss: 0.10302997, Cur Avg Loss: 0.34829035, Log Avg loss: 0.38162438, Global Avg Loss: 1.31664248, Time: 0.0208 Steps: 38180, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001997, Sample Num: 31952, Cur Loss: 0.41247335, Cur Avg Loss: 0.34914224, Log Avg loss: 0.51841235, Global Avg Loss: 1.31643347, Time: 0.0208 Steps: 38190, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002007, Sample Num: 32112, Cur Loss: 0.31895581, Cur Avg Loss: 0.34910826, Log Avg loss: 0.34232187, Global Avg Loss: 1.31617846, Time: 0.0208 Steps: 38200, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002017, Sample Num: 32272, Cur Loss: 0.32780603, Cur Avg Loss: 0.34905538, Log Avg loss: 0.33844202, Global Avg Loss: 1.31592258, Time: 0.0208 Steps: 38210, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002027, Sample Num: 32432, Cur Loss: 0.58708298, Cur Avg Loss: 0.34883103, Log Avg loss: 0.30357981, Global Avg Loss: 1.31565771, Time: 0.0208 Steps: 38220, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002037, Sample Num: 32592, Cur Loss: 0.23489547, Cur Avg Loss: 0.34853559, Log Avg loss: 0.28865056, Global Avg Loss: 1.31538907, Time: 0.0208 Steps: 38230, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002047, Sample Num: 32752, Cur Loss: 0.35871950, Cur Avg Loss: 0.34849548, Log Avg loss: 0.34032438, Global Avg Loss: 1.31513408, Time: 0.0208 Steps: 38240, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002057, Sample Num: 32912, Cur Loss: 0.10150543, Cur Avg Loss: 0.34826956, Log Avg loss: 0.30202425, Global Avg Loss: 1.31486922, Time: 0.0209 Steps: 38250, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002067, Sample Num: 33072, Cur Loss: 0.32689834, Cur Avg Loss: 0.34855101, Log Avg loss: 0.40644482, Global Avg Loss: 1.31463178, Time: 0.0209 Steps: 38260, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002077, Sample Num: 33232, Cur Loss: 0.09742871, Cur Avg Loss: 0.34819779, Log Avg loss: 0.27518851, Global Avg Loss: 1.31436017, Time: 0.0209 Steps: 38270, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002087, Sample Num: 33392, Cur Loss: 1.35038054, Cur Avg Loss: 0.34824093, Log Avg loss: 0.35720044, Global Avg Loss: 1.31411013, Time: 0.0209 Steps: 38280, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002097, Sample Num: 33552, Cur Loss: 0.18838742, Cur Avg Loss: 0.34785475, Log Avg loss: 0.26725922, Global Avg Loss: 1.31383673, Time: 0.0209 Steps: 38290, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002107, Sample Num: 33712, Cur Loss: 0.68553627, Cur Avg Loss: 0.34737037, Log Avg loss: 0.24579623, Global Avg Loss: 1.31355787, Time: 0.0209 Steps: 38300, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002117, Sample Num: 33872, Cur Loss: 0.05757078, Cur Avg Loss: 0.34703373, Log Avg loss: 0.27610400, Global Avg Loss: 1.31328707, Time: 0.0209 Steps: 38310, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002127, Sample Num: 34032, Cur Loss: 0.34596854, Cur Avg Loss: 0.34660179, Log Avg loss: 0.25515878, Global Avg Loss: 1.31301094, Time: 0.0209 Steps: 38320, Updated lr: 0.000065 ***** Running evaluation checkpoint-38322 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-38322 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.852397, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.304359, "eval_total_loss": 213.964501, "eval_mae": 0.397094, "eval_mse": 0.304454, "eval_r2": 0.806469, "eval_sp_statistic": 0.859565, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.899381, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.447933, "test_total_loss": 224.862503, "test_mae": 0.441079, "test_mse": 0.448074, "test_r2": 0.710809, "test_sp_statistic": 0.857711, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.911049, "test_ps_pvalue": 0.0, "lr": 6.460692271218587e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3129534030203969, "train_cur_epoch_loss": 737.6432439908385, "train_cur_epoch_avg_loss": 0.3464740460266973, "train_cur_epoch_time": 44.852396965026855, "train_cur_epoch_avg_time": 0.021067354140454134, "epoch": 18, "step": 38322} ################################################## Training, Epoch: 0019, Batch: 000008, Sample Num: 128, Cur Loss: 0.24117169, Cur Avg Loss: 0.33457515, Log Avg loss: 0.30978424, Global Avg Loss: 1.31274920, Time: 0.0210 Steps: 38330, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000018, Sample Num: 288, Cur Loss: 0.31716633, Cur Avg Loss: 0.46935364, Log Avg loss: 0.57717642, Global Avg Loss: 1.31255735, Time: 0.0209 Steps: 38340, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000028, Sample Num: 448, Cur Loss: 0.26984963, Cur Avg Loss: 0.45484071, Log Avg loss: 0.42871743, Global Avg Loss: 1.31232688, Time: 0.0209 Steps: 38350, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000038, Sample Num: 608, Cur Loss: 0.27884889, Cur Avg Loss: 0.48153505, Log Avg loss: 0.55627921, Global Avg Loss: 1.31212979, Time: 0.0209 Steps: 38360, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000048, Sample Num: 768, Cur Loss: 0.27463204, Cur Avg Loss: 0.45022334, Log Avg loss: 0.33123884, Global Avg Loss: 1.31187415, Time: 0.0208 Steps: 38370, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000058, Sample Num: 928, Cur Loss: 0.33044282, Cur Avg Loss: 0.44686518, Log Avg loss: 0.43074604, Global Avg Loss: 1.31164457, Time: 0.0208 Steps: 38380, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000068, Sample Num: 1088, Cur Loss: 0.31360444, Cur Avg Loss: 0.41659753, Log Avg loss: 0.24104517, Global Avg Loss: 1.31136569, Time: 0.0209 Steps: 38390, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000078, Sample Num: 1248, Cur Loss: 0.21855867, Cur Avg Loss: 0.39452122, Log Avg loss: 0.24440232, Global Avg Loss: 1.31108784, Time: 0.0209 Steps: 38400, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000088, Sample Num: 1408, Cur Loss: 0.23188742, Cur Avg Loss: 0.37430463, Log Avg loss: 0.21661517, Global Avg Loss: 1.31080289, Time: 0.0209 Steps: 38410, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000098, Sample Num: 1568, Cur Loss: 0.24628404, Cur Avg Loss: 0.36559541, Log Avg loss: 0.28895425, Global Avg Loss: 1.31053693, Time: 0.0208 Steps: 38420, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000108, Sample Num: 1728, Cur Loss: 0.34553990, Cur Avg Loss: 0.36928708, Log Avg loss: 0.40546548, Global Avg Loss: 1.31030141, Time: 0.0208 Steps: 38430, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000118, Sample Num: 1888, Cur Loss: 0.69777578, Cur Avg Loss: 0.38065781, Log Avg loss: 0.50346174, Global Avg Loss: 1.31009152, Time: 0.0209 Steps: 38440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000128, Sample Num: 2048, Cur Loss: 0.10368393, Cur Avg Loss: 0.37092334, Log Avg loss: 0.25605655, Global Avg Loss: 1.30981739, Time: 0.0208 Steps: 38450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000138, Sample Num: 2208, Cur Loss: 0.03180131, Cur Avg Loss: 0.35771972, Log Avg loss: 0.18871336, Global Avg Loss: 1.30952589, Time: 0.0209 Steps: 38460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000148, Sample Num: 2368, Cur Loss: 0.29788002, Cur Avg Loss: 0.35369935, Log Avg loss: 0.29821827, Global Avg Loss: 1.30926301, Time: 0.0209 Steps: 38470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000158, Sample Num: 2528, Cur Loss: 0.60010868, Cur Avg Loss: 0.35179764, Log Avg loss: 0.32365228, Global Avg Loss: 1.30900687, Time: 0.0208 Steps: 38480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000168, Sample Num: 2688, Cur Loss: 0.33580270, Cur Avg Loss: 0.34984909, Log Avg loss: 0.31906200, Global Avg Loss: 1.30874967, Time: 0.0208 Steps: 38490, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000178, Sample Num: 2848, Cur Loss: 0.19348828, Cur Avg Loss: 0.34434769, Log Avg loss: 0.25192419, Global Avg Loss: 1.30847517, Time: 0.0209 Steps: 38500, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000188, Sample Num: 3008, Cur Loss: 0.08097367, Cur Avg Loss: 0.33756695, Log Avg loss: 0.21686981, Global Avg Loss: 1.30819171, Time: 0.0209 Steps: 38510, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000198, Sample Num: 3168, Cur Loss: 0.17725521, Cur Avg Loss: 0.33182436, Log Avg loss: 0.22386365, Global Avg Loss: 1.30791022, Time: 0.0208 Steps: 38520, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000208, Sample Num: 3328, Cur Loss: 0.18384053, Cur Avg Loss: 0.32767395, Log Avg loss: 0.24549587, Global Avg Loss: 1.30763448, Time: 0.0209 Steps: 38530, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000218, Sample Num: 3488, Cur Loss: 0.42460829, Cur Avg Loss: 0.32921762, Log Avg loss: 0.36132603, Global Avg Loss: 1.30738894, Time: 0.0208 Steps: 38540, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000228, Sample Num: 3648, Cur Loss: 0.37223583, Cur Avg Loss: 0.32827007, Log Avg loss: 0.30761328, Global Avg Loss: 1.30712959, Time: 0.0209 Steps: 38550, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000238, Sample Num: 3808, Cur Loss: 0.41872346, Cur Avg Loss: 0.32747021, Log Avg loss: 0.30923349, Global Avg Loss: 1.30687080, Time: 0.0209 Steps: 38560, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000248, Sample Num: 3968, Cur Loss: 0.40925220, Cur Avg Loss: 0.32685388, Log Avg loss: 0.31218526, Global Avg Loss: 1.30661291, Time: 0.0208 Steps: 38570, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000258, Sample Num: 4128, Cur Loss: 0.21338406, Cur Avg Loss: 0.32823404, Log Avg loss: 0.36246205, Global Avg Loss: 1.30636819, Time: 0.0246 Steps: 38580, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000268, Sample Num: 4288, Cur Loss: 0.39247143, Cur Avg Loss: 0.32542018, Log Avg loss: 0.25282259, Global Avg Loss: 1.30609518, Time: 0.0209 Steps: 38590, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000278, Sample Num: 4448, Cur Loss: 0.68032694, Cur Avg Loss: 0.32751558, Log Avg loss: 0.38367219, Global Avg Loss: 1.30585621, Time: 0.0208 Steps: 38600, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000288, Sample Num: 4608, Cur Loss: 0.17915192, Cur Avg Loss: 0.32776252, Log Avg loss: 0.33462762, Global Avg Loss: 1.30560466, Time: 0.0209 Steps: 38610, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000298, Sample Num: 4768, Cur Loss: 0.39925271, Cur Avg Loss: 0.33079362, Log Avg loss: 0.41808935, Global Avg Loss: 1.30537485, Time: 0.0209 Steps: 38620, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000308, Sample Num: 4928, Cur Loss: 0.29898769, Cur Avg Loss: 0.33159594, Log Avg loss: 0.35550493, Global Avg Loss: 1.30512896, Time: 0.0208 Steps: 38630, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000318, Sample Num: 5088, Cur Loss: 0.43550840, Cur Avg Loss: 0.33769088, Log Avg loss: 0.52541496, Global Avg Loss: 1.30492717, Time: 0.0208 Steps: 38640, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000328, Sample Num: 5248, Cur Loss: 0.16931874, Cur Avg Loss: 0.33537412, Log Avg loss: 0.26170108, Global Avg Loss: 1.30465726, Time: 0.0208 Steps: 38650, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000338, Sample Num: 5408, Cur Loss: 0.23415962, Cur Avg Loss: 0.33174795, Log Avg loss: 0.21280973, Global Avg Loss: 1.30437483, Time: 0.0209 Steps: 38660, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000348, Sample Num: 5568, Cur Loss: 0.67435265, Cur Avg Loss: 0.33159411, Log Avg loss: 0.32639421, Global Avg Loss: 1.30412193, Time: 0.0208 Steps: 38670, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000358, Sample Num: 5728, Cur Loss: 0.34063959, Cur Avg Loss: 0.32912963, Log Avg loss: 0.24336589, Global Avg Loss: 1.30384769, Time: 0.0208 Steps: 38680, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000368, Sample Num: 5888, Cur Loss: 0.06513173, Cur Avg Loss: 0.32703225, Log Avg loss: 0.25194591, Global Avg Loss: 1.30357581, Time: 0.0208 Steps: 38690, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000378, Sample Num: 6048, Cur Loss: 0.40784115, Cur Avg Loss: 0.32695705, Log Avg loss: 0.32418968, Global Avg Loss: 1.30332274, Time: 0.0208 Steps: 38700, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000388, Sample Num: 6208, Cur Loss: 0.32654759, Cur Avg Loss: 0.32791591, Log Avg loss: 0.36416078, Global Avg Loss: 1.30308013, Time: 0.0208 Steps: 38710, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000398, Sample Num: 6368, Cur Loss: 0.39666736, Cur Avg Loss: 0.32812885, Log Avg loss: 0.33639088, Global Avg Loss: 1.30283046, Time: 0.0208 Steps: 38720, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000408, Sample Num: 6528, Cur Loss: 0.38823450, Cur Avg Loss: 0.32946315, Log Avg loss: 0.38256852, Global Avg Loss: 1.30259286, Time: 0.0209 Steps: 38730, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000418, Sample Num: 6688, Cur Loss: 0.23125122, Cur Avg Loss: 0.32895751, Log Avg loss: 0.30832728, Global Avg Loss: 1.30233620, Time: 0.0208 Steps: 38740, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000428, Sample Num: 6848, Cur Loss: 0.57211250, Cur Avg Loss: 0.32926624, Log Avg loss: 0.34217117, Global Avg Loss: 1.30208842, Time: 0.0208 Steps: 38750, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000438, Sample Num: 7008, Cur Loss: 0.24947575, Cur Avg Loss: 0.33160332, Log Avg loss: 0.43163020, Global Avg Loss: 1.30186384, Time: 0.0208 Steps: 38760, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000448, Sample Num: 7168, Cur Loss: 0.53963280, Cur Avg Loss: 0.33190263, Log Avg loss: 0.34501249, Global Avg Loss: 1.30161704, Time: 0.0208 Steps: 38770, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000458, Sample Num: 7328, Cur Loss: 0.56405777, Cur Avg Loss: 0.33229863, Log Avg loss: 0.35003962, Global Avg Loss: 1.30137166, Time: 0.0208 Steps: 38780, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000468, Sample Num: 7488, Cur Loss: 0.30299789, Cur Avg Loss: 0.33258102, Log Avg loss: 0.34551442, Global Avg Loss: 1.30112524, Time: 0.0209 Steps: 38790, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000478, Sample Num: 7648, Cur Loss: 0.15706103, Cur Avg Loss: 0.33120798, Log Avg loss: 0.26694989, Global Avg Loss: 1.30085870, Time: 0.0208 Steps: 38800, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000488, Sample Num: 7808, Cur Loss: 0.15305495, Cur Avg Loss: 0.33015324, Log Avg loss: 0.27973633, Global Avg Loss: 1.30059560, Time: 0.0208 Steps: 38810, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000498, Sample Num: 7968, Cur Loss: 0.20757109, Cur Avg Loss: 0.33189360, Log Avg loss: 0.41682331, Global Avg Loss: 1.30036794, Time: 0.0209 Steps: 38820, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000508, Sample Num: 8128, Cur Loss: 0.15389049, Cur Avg Loss: 0.32951235, Log Avg loss: 0.21092612, Global Avg Loss: 1.30008737, Time: 0.0208 Steps: 38830, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000518, Sample Num: 8288, Cur Loss: 0.36642689, Cur Avg Loss: 0.32795088, Log Avg loss: 0.24862815, Global Avg Loss: 1.29981665, Time: 0.0209 Steps: 38840, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000528, Sample Num: 8448, Cur Loss: 0.17873007, Cur Avg Loss: 0.32956836, Log Avg loss: 0.41335384, Global Avg Loss: 1.29958848, Time: 0.0208 Steps: 38850, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000538, Sample Num: 8608, Cur Loss: 0.39944273, Cur Avg Loss: 0.33109699, Log Avg loss: 0.41180840, Global Avg Loss: 1.29936002, Time: 0.0208 Steps: 38860, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000548, Sample Num: 8768, Cur Loss: 0.33238465, Cur Avg Loss: 0.33123714, Log Avg loss: 0.33877760, Global Avg Loss: 1.29911290, Time: 0.0208 Steps: 38870, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000558, Sample Num: 8928, Cur Loss: 0.21972038, Cur Avg Loss: 0.33116530, Log Avg loss: 0.32722832, Global Avg Loss: 1.29886293, Time: 0.0209 Steps: 38880, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000568, Sample Num: 9088, Cur Loss: 0.41957808, Cur Avg Loss: 0.33457284, Log Avg loss: 0.52471353, Global Avg Loss: 1.29866386, Time: 0.0209 Steps: 38890, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000578, Sample Num: 9248, Cur Loss: 0.23065095, Cur Avg Loss: 0.33393432, Log Avg loss: 0.29766640, Global Avg Loss: 1.29840654, Time: 0.0208 Steps: 38900, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000588, Sample Num: 9408, Cur Loss: 0.55803692, Cur Avg Loss: 0.33251017, Log Avg loss: 0.25019450, Global Avg Loss: 1.29813714, Time: 0.0209 Steps: 38910, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000598, Sample Num: 9568, Cur Loss: 0.31843883, Cur Avg Loss: 0.33183614, Log Avg loss: 0.29220292, Global Avg Loss: 1.29787868, Time: 0.0210 Steps: 38920, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000608, Sample Num: 9728, Cur Loss: 0.11657766, Cur Avg Loss: 0.33176006, Log Avg loss: 0.32721076, Global Avg Loss: 1.29762935, Time: 0.0208 Steps: 38930, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000618, Sample Num: 9888, Cur Loss: 0.09720188, Cur Avg Loss: 0.33062200, Log Avg loss: 0.26142792, Global Avg Loss: 1.29736324, Time: 0.0209 Steps: 38940, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000628, Sample Num: 10048, Cur Loss: 0.38466567, Cur Avg Loss: 0.33086287, Log Avg loss: 0.34574834, Global Avg Loss: 1.29711893, Time: 0.0209 Steps: 38950, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000638, Sample Num: 10208, Cur Loss: 0.12284951, Cur Avg Loss: 0.32995581, Log Avg loss: 0.27299245, Global Avg Loss: 1.29685606, Time: 0.0208 Steps: 38960, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000648, Sample Num: 10368, Cur Loss: 0.39957920, Cur Avg Loss: 0.32813336, Log Avg loss: 0.21186147, Global Avg Loss: 1.29657764, Time: 0.0208 Steps: 38970, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000658, Sample Num: 10528, Cur Loss: 0.24597169, Cur Avg Loss: 0.32734680, Log Avg loss: 0.27637721, Global Avg Loss: 1.29631592, Time: 0.0208 Steps: 38980, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000668, Sample Num: 10688, Cur Loss: 0.22252314, Cur Avg Loss: 0.32614428, Log Avg loss: 0.24701840, Global Avg Loss: 1.29604680, Time: 0.0208 Steps: 38990, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000678, Sample Num: 10848, Cur Loss: 0.30999231, Cur Avg Loss: 0.32442916, Log Avg loss: 0.20985952, Global Avg Loss: 1.29576829, Time: 0.0209 Steps: 39000, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000688, Sample Num: 11008, Cur Loss: 0.51375878, Cur Avg Loss: 0.32352854, Log Avg loss: 0.26246616, Global Avg Loss: 1.29550341, Time: 0.0209 Steps: 39010, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000698, Sample Num: 11168, Cur Loss: 0.19461165, Cur Avg Loss: 0.32596141, Log Avg loss: 0.49334308, Global Avg Loss: 1.29529783, Time: 0.0209 Steps: 39020, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000708, Sample Num: 11328, Cur Loss: 0.40678123, Cur Avg Loss: 0.32704199, Log Avg loss: 0.40246682, Global Avg Loss: 1.29506908, Time: 0.0208 Steps: 39030, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000718, Sample Num: 11488, Cur Loss: 0.15233514, Cur Avg Loss: 0.32728868, Log Avg loss: 0.34475392, Global Avg Loss: 1.29482566, Time: 0.0208 Steps: 39040, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000728, Sample Num: 11648, Cur Loss: 0.23473912, Cur Avg Loss: 0.32708110, Log Avg loss: 0.31217711, Global Avg Loss: 1.29457402, Time: 0.0209 Steps: 39050, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000738, Sample Num: 11808, Cur Loss: 0.64709121, Cur Avg Loss: 0.32770422, Log Avg loss: 0.37306720, Global Avg Loss: 1.29433810, Time: 0.0208 Steps: 39060, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000748, Sample Num: 11968, Cur Loss: 0.09922604, Cur Avg Loss: 0.32791921, Log Avg loss: 0.34378563, Global Avg Loss: 1.29409480, Time: 0.0209 Steps: 39070, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000758, Sample Num: 12128, Cur Loss: 0.26349512, Cur Avg Loss: 0.32865946, Log Avg loss: 0.38403023, Global Avg Loss: 1.29386193, Time: 0.0208 Steps: 39080, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000768, Sample Num: 12288, Cur Loss: 0.19354555, Cur Avg Loss: 0.32838828, Log Avg loss: 0.30783242, Global Avg Loss: 1.29360968, Time: 0.0255 Steps: 39090, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000778, Sample Num: 12448, Cur Loss: 0.15939549, Cur Avg Loss: 0.32978939, Log Avg loss: 0.43739510, Global Avg Loss: 1.29339070, Time: 0.0210 Steps: 39100, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000788, Sample Num: 12608, Cur Loss: 0.38723600, Cur Avg Loss: 0.32932728, Log Avg loss: 0.29337492, Global Avg Loss: 1.29313501, Time: 0.0211 Steps: 39110, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000798, Sample Num: 12768, Cur Loss: 0.27845150, Cur Avg Loss: 0.32807538, Log Avg loss: 0.22942563, Global Avg Loss: 1.29286310, Time: 0.0219 Steps: 39120, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000808, Sample Num: 12928, Cur Loss: 0.42763484, Cur Avg Loss: 0.32746279, Log Avg loss: 0.27857777, Global Avg Loss: 1.29260389, Time: 0.0210 Steps: 39130, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000818, Sample Num: 13088, Cur Loss: 0.21974617, Cur Avg Loss: 0.32669053, Log Avg loss: 0.26429194, Global Avg Loss: 1.29234116, Time: 0.0211 Steps: 39140, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000828, Sample Num: 13248, Cur Loss: 0.29667407, Cur Avg Loss: 0.32641123, Log Avg loss: 0.30356501, Global Avg Loss: 1.29208860, Time: 0.0220 Steps: 39150, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000838, Sample Num: 13408, Cur Loss: 0.36784145, Cur Avg Loss: 0.32660558, Log Avg loss: 0.34269746, Global Avg Loss: 1.29184616, Time: 0.0210 Steps: 39160, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000848, Sample Num: 13568, Cur Loss: 0.41545698, Cur Avg Loss: 0.32841080, Log Avg loss: 0.47968821, Global Avg Loss: 1.29163882, Time: 0.0219 Steps: 39170, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000858, Sample Num: 13728, Cur Loss: 0.17926285, Cur Avg Loss: 0.32825462, Log Avg loss: 0.31501073, Global Avg Loss: 1.29138956, Time: 0.0210 Steps: 39180, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000868, Sample Num: 13888, Cur Loss: 0.55552489, Cur Avg Loss: 0.32994790, Log Avg loss: 0.47523124, Global Avg Loss: 1.29118130, Time: 0.0219 Steps: 39190, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000878, Sample Num: 14048, Cur Loss: 0.56659031, Cur Avg Loss: 0.33034202, Log Avg loss: 0.36455157, Global Avg Loss: 1.29094491, Time: 0.0210 Steps: 39200, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000888, Sample Num: 14208, Cur Loss: 0.81608891, Cur Avg Loss: 0.32946792, Log Avg loss: 0.25272195, Global Avg Loss: 1.29068013, Time: 0.0220 Steps: 39210, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000898, Sample Num: 14368, Cur Loss: 0.24577513, Cur Avg Loss: 0.32994218, Log Avg loss: 0.37205679, Global Avg Loss: 1.29044590, Time: 0.0210 Steps: 39220, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000908, Sample Num: 14528, Cur Loss: 0.40979236, Cur Avg Loss: 0.32940660, Log Avg loss: 0.28131138, Global Avg Loss: 1.29018867, Time: 0.0211 Steps: 39230, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000918, Sample Num: 14688, Cur Loss: 0.30716103, Cur Avg Loss: 0.32912734, Log Avg loss: 0.30377081, Global Avg Loss: 1.28993729, Time: 0.0219 Steps: 39240, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000928, Sample Num: 14848, Cur Loss: 0.33043253, Cur Avg Loss: 0.32865895, Log Avg loss: 0.28566011, Global Avg Loss: 1.28968142, Time: 0.0210 Steps: 39250, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000938, Sample Num: 15008, Cur Loss: 0.41555133, Cur Avg Loss: 0.32904598, Log Avg loss: 0.36496236, Global Avg Loss: 1.28944588, Time: 0.0212 Steps: 39260, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000948, Sample Num: 15168, Cur Loss: 0.59565699, Cur Avg Loss: 0.32990630, Log Avg loss: 0.41060442, Global Avg Loss: 1.28922209, Time: 0.0210 Steps: 39270, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000958, Sample Num: 15328, Cur Loss: 0.21711183, Cur Avg Loss: 0.33073012, Log Avg loss: 0.40882835, Global Avg Loss: 1.28899796, Time: 0.0211 Steps: 39280, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000968, Sample Num: 15488, Cur Loss: 0.35642412, Cur Avg Loss: 0.32995235, Log Avg loss: 0.25544230, Global Avg Loss: 1.28873490, Time: 0.0209 Steps: 39290, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000978, Sample Num: 15648, Cur Loss: 0.27187014, Cur Avg Loss: 0.33108069, Log Avg loss: 0.44030408, Global Avg Loss: 1.28851901, Time: 0.0211 Steps: 39300, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000988, Sample Num: 15808, Cur Loss: 0.23931780, Cur Avg Loss: 0.33119438, Log Avg loss: 0.34231249, Global Avg Loss: 1.28827831, Time: 0.0219 Steps: 39310, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000998, Sample Num: 15968, Cur Loss: 0.32148007, Cur Avg Loss: 0.33025679, Log Avg loss: 0.23762346, Global Avg Loss: 1.28801110, Time: 0.0210 Steps: 39320, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001008, Sample Num: 16128, Cur Loss: 0.27772659, Cur Avg Loss: 0.32973455, Log Avg loss: 0.27761450, Global Avg Loss: 1.28775420, Time: 0.0212 Steps: 39330, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001018, Sample Num: 16288, Cur Loss: 0.30727541, Cur Avg Loss: 0.32889976, Log Avg loss: 0.24475313, Global Avg Loss: 1.28748908, Time: 0.0219 Steps: 39340, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001028, Sample Num: 16448, Cur Loss: 0.29501757, Cur Avg Loss: 0.32947861, Log Avg loss: 0.38840577, Global Avg Loss: 1.28726059, Time: 0.0212 Steps: 39350, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001038, Sample Num: 16608, Cur Loss: 0.33586439, Cur Avg Loss: 0.32854962, Log Avg loss: 0.23304946, Global Avg Loss: 1.28699275, Time: 0.0209 Steps: 39360, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001048, Sample Num: 16768, Cur Loss: 0.34082860, Cur Avg Loss: 0.32850892, Log Avg loss: 0.32428428, Global Avg Loss: 1.28674823, Time: 0.0209 Steps: 39370, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001058, Sample Num: 16928, Cur Loss: 0.76263058, Cur Avg Loss: 0.32929120, Log Avg loss: 0.41127364, Global Avg Loss: 1.28652591, Time: 0.0209 Steps: 39380, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001068, Sample Num: 17088, Cur Loss: 0.35429806, Cur Avg Loss: 0.32888706, Log Avg loss: 0.28612916, Global Avg Loss: 1.28627194, Time: 0.0209 Steps: 39390, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001078, Sample Num: 17248, Cur Loss: 0.16271703, Cur Avg Loss: 0.32858689, Log Avg loss: 0.29652938, Global Avg Loss: 1.28602074, Time: 0.0209 Steps: 39400, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001088, Sample Num: 17408, Cur Loss: 0.30167720, Cur Avg Loss: 0.32740211, Log Avg loss: 0.19968263, Global Avg Loss: 1.28574509, Time: 0.0209 Steps: 39410, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001098, Sample Num: 17568, Cur Loss: 0.51361531, Cur Avg Loss: 0.32762251, Log Avg loss: 0.35160131, Global Avg Loss: 1.28550811, Time: 0.0209 Steps: 39420, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001108, Sample Num: 17728, Cur Loss: 0.32683131, Cur Avg Loss: 0.32859246, Log Avg loss: 0.43509393, Global Avg Loss: 1.28529244, Time: 0.0209 Steps: 39430, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001118, Sample Num: 17888, Cur Loss: 0.46146259, Cur Avg Loss: 0.32883323, Log Avg loss: 0.35551008, Global Avg Loss: 1.28505669, Time: 0.0209 Steps: 39440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001128, Sample Num: 18048, Cur Loss: 0.15200838, Cur Avg Loss: 0.32947883, Log Avg loss: 0.40165728, Global Avg Loss: 1.28483276, Time: 0.0209 Steps: 39450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001138, Sample Num: 18208, Cur Loss: 0.48809955, Cur Avg Loss: 0.32874858, Log Avg loss: 0.24637593, Global Avg Loss: 1.28456959, Time: 0.0209 Steps: 39460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001148, Sample Num: 18368, Cur Loss: 0.19448721, Cur Avg Loss: 0.32896414, Log Avg loss: 0.35349507, Global Avg Loss: 1.28433370, Time: 0.0209 Steps: 39470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001158, Sample Num: 18528, Cur Loss: 0.25871682, Cur Avg Loss: 0.32942217, Log Avg loss: 0.38200378, Global Avg Loss: 1.28410515, Time: 0.0209 Steps: 39480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001168, Sample Num: 18688, Cur Loss: 0.13973445, Cur Avg Loss: 0.32975854, Log Avg loss: 0.36870997, Global Avg Loss: 1.28387334, Time: 0.0209 Steps: 39490, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001178, Sample Num: 18848, Cur Loss: 0.38054061, Cur Avg Loss: 0.33221394, Log Avg loss: 0.61900481, Global Avg Loss: 1.28370502, Time: 0.0209 Steps: 39500, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001188, Sample Num: 19008, Cur Loss: 0.14163256, Cur Avg Loss: 0.33254733, Log Avg loss: 0.37182127, Global Avg Loss: 1.28347422, Time: 0.0209 Steps: 39510, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001198, Sample Num: 19168, Cur Loss: 0.15858874, Cur Avg Loss: 0.33188717, Log Avg loss: 0.25346005, Global Avg Loss: 1.28321359, Time: 0.0209 Steps: 39520, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001208, Sample Num: 19328, Cur Loss: 0.40093407, Cur Avg Loss: 0.33162723, Log Avg loss: 0.30048597, Global Avg Loss: 1.28296499, Time: 0.0209 Steps: 39530, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001218, Sample Num: 19488, Cur Loss: 0.54922545, Cur Avg Loss: 0.33123603, Log Avg loss: 0.28397909, Global Avg Loss: 1.28271234, Time: 0.0209 Steps: 39540, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001228, Sample Num: 19648, Cur Loss: 0.70307016, Cur Avg Loss: 0.33104783, Log Avg loss: 0.30812477, Global Avg Loss: 1.28246592, Time: 0.0209 Steps: 39550, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001238, Sample Num: 19808, Cur Loss: 0.12087677, Cur Avg Loss: 0.32964564, Log Avg loss: 0.15745672, Global Avg Loss: 1.28218154, Time: 0.0209 Steps: 39560, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001248, Sample Num: 19968, Cur Loss: 0.14892790, Cur Avg Loss: 0.32863700, Log Avg loss: 0.20376766, Global Avg Loss: 1.28190900, Time: 0.0208 Steps: 39570, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001258, Sample Num: 20128, Cur Loss: 0.22084533, Cur Avg Loss: 0.32894701, Log Avg loss: 0.36763611, Global Avg Loss: 1.28167801, Time: 0.0208 Steps: 39580, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001268, Sample Num: 20288, Cur Loss: 0.18277285, Cur Avg Loss: 0.32834961, Log Avg loss: 0.25319751, Global Avg Loss: 1.28141823, Time: 0.0209 Steps: 39590, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001278, Sample Num: 20448, Cur Loss: 0.50161362, Cur Avg Loss: 0.32971264, Log Avg loss: 0.50254449, Global Avg Loss: 1.28122154, Time: 0.0209 Steps: 39600, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001288, Sample Num: 20608, Cur Loss: 0.38103834, Cur Avg Loss: 0.33080762, Log Avg loss: 0.47074590, Global Avg Loss: 1.28101693, Time: 0.0209 Steps: 39610, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001298, Sample Num: 20768, Cur Loss: 0.15650350, Cur Avg Loss: 0.33074861, Log Avg loss: 0.32314790, Global Avg Loss: 1.28077516, Time: 0.0219 Steps: 39620, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001308, Sample Num: 20928, Cur Loss: 0.83113623, Cur Avg Loss: 0.33058347, Log Avg loss: 0.30914799, Global Avg Loss: 1.28052999, Time: 0.0209 Steps: 39630, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001318, Sample Num: 21088, Cur Loss: 0.13549784, Cur Avg Loss: 0.33086360, Log Avg loss: 0.36750536, Global Avg Loss: 1.28029966, Time: 0.0219 Steps: 39640, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001328, Sample Num: 21248, Cur Loss: 0.27663147, Cur Avg Loss: 0.33028490, Log Avg loss: 0.25401139, Global Avg Loss: 1.28004082, Time: 0.0209 Steps: 39650, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001338, Sample Num: 21408, Cur Loss: 0.27165878, Cur Avg Loss: 0.32968886, Log Avg loss: 0.25053570, Global Avg Loss: 1.27978124, Time: 0.0219 Steps: 39660, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001348, Sample Num: 21568, Cur Loss: 0.19448911, Cur Avg Loss: 0.32941810, Log Avg loss: 0.29318967, Global Avg Loss: 1.27953254, Time: 0.0208 Steps: 39670, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001358, Sample Num: 21728, Cur Loss: 0.09003917, Cur Avg Loss: 0.32982015, Log Avg loss: 0.38401694, Global Avg Loss: 1.27930686, Time: 0.0209 Steps: 39680, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001368, Sample Num: 21888, Cur Loss: 0.19606546, Cur Avg Loss: 0.33011846, Log Avg loss: 0.37062856, Global Avg Loss: 1.27907791, Time: 0.0209 Steps: 39690, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001378, Sample Num: 22048, Cur Loss: 0.54727221, Cur Avg Loss: 0.32997736, Log Avg loss: 0.31067538, Global Avg Loss: 1.27883398, Time: 0.0208 Steps: 39700, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001388, Sample Num: 22208, Cur Loss: 0.26518103, Cur Avg Loss: 0.32982275, Log Avg loss: 0.30851760, Global Avg Loss: 1.27858963, Time: 0.0208 Steps: 39710, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001398, Sample Num: 22368, Cur Loss: 0.41303104, Cur Avg Loss: 0.32888855, Log Avg loss: 0.19922136, Global Avg Loss: 1.27831789, Time: 0.0208 Steps: 39720, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001408, Sample Num: 22528, Cur Loss: 0.16183043, Cur Avg Loss: 0.32896929, Log Avg loss: 0.34025687, Global Avg Loss: 1.27808178, Time: 0.0208 Steps: 39730, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001418, Sample Num: 22688, Cur Loss: 0.24076775, Cur Avg Loss: 0.32852442, Log Avg loss: 0.26588594, Global Avg Loss: 1.27782707, Time: 0.0208 Steps: 39740, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001428, Sample Num: 22848, Cur Loss: 0.41870809, Cur Avg Loss: 0.32848847, Log Avg loss: 0.32339204, Global Avg Loss: 1.27758696, Time: 0.0208 Steps: 39750, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001438, Sample Num: 23008, Cur Loss: 0.35016492, Cur Avg Loss: 0.32940690, Log Avg loss: 0.46055747, Global Avg Loss: 1.27738147, Time: 0.0208 Steps: 39760, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001448, Sample Num: 23168, Cur Loss: 0.63535112, Cur Avg Loss: 0.33010220, Log Avg loss: 0.43008710, Global Avg Loss: 1.27716843, Time: 0.0208 Steps: 39770, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001458, Sample Num: 23328, Cur Loss: 0.12161808, Cur Avg Loss: 0.32977679, Log Avg loss: 0.28265774, Global Avg Loss: 1.27691842, Time: 0.0209 Steps: 39780, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001468, Sample Num: 23488, Cur Loss: 0.39467973, Cur Avg Loss: 0.32995542, Log Avg loss: 0.35599952, Global Avg Loss: 1.27668698, Time: 0.0209 Steps: 39790, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001478, Sample Num: 23648, Cur Loss: 0.29583135, Cur Avg Loss: 0.32993503, Log Avg loss: 0.32694124, Global Avg Loss: 1.27644835, Time: 0.0219 Steps: 39800, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001488, Sample Num: 23808, Cur Loss: 0.11773132, Cur Avg Loss: 0.33004819, Log Avg loss: 0.34677314, Global Avg Loss: 1.27621482, Time: 0.0209 Steps: 39810, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001498, Sample Num: 23968, Cur Loss: 0.13561213, Cur Avg Loss: 0.32972440, Log Avg loss: 0.28154454, Global Avg Loss: 1.27596503, Time: 0.0209 Steps: 39820, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001508, Sample Num: 24128, Cur Loss: 0.33477727, Cur Avg Loss: 0.32970849, Log Avg loss: 0.32732536, Global Avg Loss: 1.27572686, Time: 0.0219 Steps: 39830, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001518, Sample Num: 24288, Cur Loss: 0.23441917, Cur Avg Loss: 0.32995719, Log Avg loss: 0.36746191, Global Avg Loss: 1.27549888, Time: 0.0209 Steps: 39840, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001528, Sample Num: 24448, Cur Loss: 0.78133994, Cur Avg Loss: 0.33069740, Log Avg loss: 0.44306098, Global Avg Loss: 1.27528999, Time: 0.0209 Steps: 39850, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001538, Sample Num: 24608, Cur Loss: 0.11967821, Cur Avg Loss: 0.33059920, Log Avg loss: 0.31559334, Global Avg Loss: 1.27504922, Time: 0.0247 Steps: 39860, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001548, Sample Num: 24768, Cur Loss: 0.37082452, Cur Avg Loss: 0.33042588, Log Avg loss: 0.30377021, Global Avg Loss: 1.27480561, Time: 0.0208 Steps: 39870, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001558, Sample Num: 24928, Cur Loss: 0.31199220, Cur Avg Loss: 0.32973327, Log Avg loss: 0.22251706, Global Avg Loss: 1.27454174, Time: 0.0208 Steps: 39880, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001568, Sample Num: 25088, Cur Loss: 0.05865017, Cur Avg Loss: 0.32882063, Log Avg loss: 0.18663127, Global Avg Loss: 1.27426902, Time: 0.0209 Steps: 39890, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001578, Sample Num: 25248, Cur Loss: 0.40873796, Cur Avg Loss: 0.32941409, Log Avg loss: 0.42246821, Global Avg Loss: 1.27405553, Time: 0.0208 Steps: 39900, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001588, Sample Num: 25408, Cur Loss: 0.08860256, Cur Avg Loss: 0.32868715, Log Avg loss: 0.21397688, Global Avg Loss: 1.27378992, Time: 0.0209 Steps: 39910, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001598, Sample Num: 25568, Cur Loss: 0.14880779, Cur Avg Loss: 0.32841778, Log Avg loss: 0.28564135, Global Avg Loss: 1.27354238, Time: 0.0208 Steps: 39920, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001608, Sample Num: 25728, Cur Loss: 0.30933273, Cur Avg Loss: 0.32851877, Log Avg loss: 0.34465648, Global Avg Loss: 1.27330975, Time: 0.0209 Steps: 39930, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001618, Sample Num: 25888, Cur Loss: 0.19769853, Cur Avg Loss: 0.32834883, Log Avg loss: 0.30102329, Global Avg Loss: 1.27306632, Time: 0.0208 Steps: 39940, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001628, Sample Num: 26048, Cur Loss: 0.67116034, Cur Avg Loss: 0.32884553, Log Avg loss: 0.40921118, Global Avg Loss: 1.27285008, Time: 0.0208 Steps: 39950, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001638, Sample Num: 26208, Cur Loss: 0.17381881, Cur Avg Loss: 0.32895868, Log Avg loss: 0.34737964, Global Avg Loss: 1.27261848, Time: 0.0208 Steps: 39960, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001648, Sample Num: 26368, Cur Loss: 0.39452177, Cur Avg Loss: 0.32888723, Log Avg loss: 0.31718332, Global Avg Loss: 1.27237945, Time: 0.0208 Steps: 39970, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001658, Sample Num: 26528, Cur Loss: 0.18452963, Cur Avg Loss: 0.32938373, Log Avg loss: 0.41120656, Global Avg Loss: 1.27216405, Time: 0.0209 Steps: 39980, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001668, Sample Num: 26688, Cur Loss: 0.63147092, Cur Avg Loss: 0.32982819, Log Avg loss: 0.40351962, Global Avg Loss: 1.27194683, Time: 0.0209 Steps: 39990, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001678, Sample Num: 26848, Cur Loss: 0.15488151, Cur Avg Loss: 0.32999032, Log Avg loss: 0.35703337, Global Avg Loss: 1.27171810, Time: 0.0208 Steps: 40000, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001688, Sample Num: 27008, Cur Loss: 0.30277327, Cur Avg Loss: 0.33043166, Log Avg loss: 0.40448948, Global Avg Loss: 1.27150135, Time: 0.0209 Steps: 40010, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001698, Sample Num: 27168, Cur Loss: 0.40110588, Cur Avg Loss: 0.33034657, Log Avg loss: 0.31598308, Global Avg Loss: 1.27126259, Time: 0.0209 Steps: 40020, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001708, Sample Num: 27328, Cur Loss: 0.41286173, Cur Avg Loss: 0.33070218, Log Avg loss: 0.39108468, Global Avg Loss: 1.27104271, Time: 0.0208 Steps: 40030, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001718, Sample Num: 27488, Cur Loss: 0.63489974, Cur Avg Loss: 0.33041975, Log Avg loss: 0.28218122, Global Avg Loss: 1.27079574, Time: 0.0209 Steps: 40040, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001728, Sample Num: 27648, Cur Loss: 0.31380200, Cur Avg Loss: 0.32974131, Log Avg loss: 0.21318528, Global Avg Loss: 1.27053167, Time: 0.0208 Steps: 40050, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001738, Sample Num: 27808, Cur Loss: 0.34645197, Cur Avg Loss: 0.33004983, Log Avg loss: 0.38336112, Global Avg Loss: 1.27031021, Time: 0.0209 Steps: 40060, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001748, Sample Num: 27968, Cur Loss: 0.10108911, Cur Avg Loss: 0.33017194, Log Avg loss: 0.35139513, Global Avg Loss: 1.27008088, Time: 0.0209 Steps: 40070, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001758, Sample Num: 28128, Cur Loss: 0.14616783, Cur Avg Loss: 0.32922641, Log Avg loss: 0.16394835, Global Avg Loss: 1.26980490, Time: 0.0208 Steps: 40080, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001768, Sample Num: 28288, Cur Loss: 0.37817484, Cur Avg Loss: 0.32913530, Log Avg loss: 0.31311769, Global Avg Loss: 1.26956626, Time: 0.0208 Steps: 40090, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001778, Sample Num: 28448, Cur Loss: 0.58435619, Cur Avg Loss: 0.32918176, Log Avg loss: 0.33739598, Global Avg Loss: 1.26933380, Time: 0.0208 Steps: 40100, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001788, Sample Num: 28608, Cur Loss: 0.39735249, Cur Avg Loss: 0.32950716, Log Avg loss: 0.38736258, Global Avg Loss: 1.26911391, Time: 0.0209 Steps: 40110, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001798, Sample Num: 28768, Cur Loss: 0.37891978, Cur Avg Loss: 0.33001376, Log Avg loss: 0.42059521, Global Avg Loss: 1.26890242, Time: 0.0210 Steps: 40120, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001808, Sample Num: 28928, Cur Loss: 0.35928264, Cur Avg Loss: 0.33026814, Log Avg loss: 0.37600499, Global Avg Loss: 1.26867992, Time: 0.0209 Steps: 40130, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001818, Sample Num: 29088, Cur Loss: 0.71621406, Cur Avg Loss: 0.33101695, Log Avg loss: 0.46640222, Global Avg Loss: 1.26848005, Time: 0.0209 Steps: 40140, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001828, Sample Num: 29248, Cur Loss: 0.40873533, Cur Avg Loss: 0.33194761, Log Avg loss: 0.50114060, Global Avg Loss: 1.26828893, Time: 0.0209 Steps: 40150, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001838, Sample Num: 29408, Cur Loss: 0.43904987, Cur Avg Loss: 0.33186897, Log Avg loss: 0.31749359, Global Avg Loss: 1.26805218, Time: 0.0209 Steps: 40160, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001848, Sample Num: 29568, Cur Loss: 0.13851513, Cur Avg Loss: 0.33172136, Log Avg loss: 0.30459034, Global Avg Loss: 1.26781233, Time: 0.0209 Steps: 40170, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001858, Sample Num: 29728, Cur Loss: 0.86985314, Cur Avg Loss: 0.33229686, Log Avg loss: 0.43865031, Global Avg Loss: 1.26760597, Time: 0.0209 Steps: 40180, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001868, Sample Num: 29888, Cur Loss: 1.47726560, Cur Avg Loss: 0.33362361, Log Avg loss: 0.58013280, Global Avg Loss: 1.26743491, Time: 0.0209 Steps: 40190, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001878, Sample Num: 30048, Cur Loss: 0.35164350, Cur Avg Loss: 0.33387421, Log Avg loss: 0.38068728, Global Avg Loss: 1.26721433, Time: 0.0209 Steps: 40200, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001888, Sample Num: 30208, Cur Loss: 0.21287933, Cur Avg Loss: 0.33395250, Log Avg loss: 0.34865597, Global Avg Loss: 1.26698589, Time: 0.0208 Steps: 40210, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001898, Sample Num: 30368, Cur Loss: 1.14746392, Cur Avg Loss: 0.33460896, Log Avg loss: 0.45854795, Global Avg Loss: 1.26678489, Time: 0.0209 Steps: 40220, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001908, Sample Num: 30528, Cur Loss: 0.30430105, Cur Avg Loss: 0.33546703, Log Avg loss: 0.49832899, Global Avg Loss: 1.26659387, Time: 0.0209 Steps: 40230, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001918, Sample Num: 30688, Cur Loss: 0.55705839, Cur Avg Loss: 0.33572520, Log Avg loss: 0.38498437, Global Avg Loss: 1.26637478, Time: 0.0208 Steps: 40240, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001928, Sample Num: 30848, Cur Loss: 0.56739104, Cur Avg Loss: 0.33607137, Log Avg loss: 0.40246641, Global Avg Loss: 1.26616015, Time: 0.0209 Steps: 40250, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001938, Sample Num: 31008, Cur Loss: 0.41085333, Cur Avg Loss: 0.33631448, Log Avg loss: 0.38318472, Global Avg Loss: 1.26594083, Time: 0.0208 Steps: 40260, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001948, Sample Num: 31168, Cur Loss: 0.24934462, Cur Avg Loss: 0.33634404, Log Avg loss: 0.34207341, Global Avg Loss: 1.26571141, Time: 0.0208 Steps: 40270, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001958, Sample Num: 31328, Cur Loss: 0.42077181, Cur Avg Loss: 0.33619657, Log Avg loss: 0.30747010, Global Avg Loss: 1.26547352, Time: 0.0210 Steps: 40280, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001968, Sample Num: 31488, Cur Loss: 0.62018132, Cur Avg Loss: 0.33674292, Log Avg loss: 0.44371741, Global Avg Loss: 1.26526956, Time: 0.0208 Steps: 40290, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001978, Sample Num: 31648, Cur Loss: 0.25860223, Cur Avg Loss: 0.33703380, Log Avg loss: 0.39427925, Global Avg Loss: 1.26505343, Time: 0.0210 Steps: 40300, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001988, Sample Num: 31808, Cur Loss: 0.30237859, Cur Avg Loss: 0.33641203, Log Avg loss: 0.21342591, Global Avg Loss: 1.26479254, Time: 0.0209 Steps: 40310, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001998, Sample Num: 31968, Cur Loss: 0.41592082, Cur Avg Loss: 0.33642280, Log Avg loss: 0.33856314, Global Avg Loss: 1.26456282, Time: 0.0209 Steps: 40320, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002008, Sample Num: 32128, Cur Loss: 0.16302481, Cur Avg Loss: 0.33632267, Log Avg loss: 0.31631735, Global Avg Loss: 1.26432770, Time: 0.0210 Steps: 40330, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002018, Sample Num: 32288, Cur Loss: 0.31149974, Cur Avg Loss: 0.33663631, Log Avg loss: 0.39961618, Global Avg Loss: 1.26411335, Time: 0.0209 Steps: 40340, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002028, Sample Num: 32448, Cur Loss: 0.28906906, Cur Avg Loss: 0.33660721, Log Avg loss: 0.33073462, Global Avg Loss: 1.26388203, Time: 0.0208 Steps: 40350, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002038, Sample Num: 32608, Cur Loss: 0.16416065, Cur Avg Loss: 0.33608181, Log Avg loss: 0.22952972, Global Avg Loss: 1.26362574, Time: 0.0208 Steps: 40360, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002048, Sample Num: 32768, Cur Loss: 0.39939314, Cur Avg Loss: 0.33618550, Log Avg loss: 0.35731895, Global Avg Loss: 1.26340124, Time: 0.0254 Steps: 40370, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002058, Sample Num: 32928, Cur Loss: 0.28813460, Cur Avg Loss: 0.33566309, Log Avg loss: 0.22867242, Global Avg Loss: 1.26314500, Time: 0.0209 Steps: 40380, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002068, Sample Num: 33088, Cur Loss: 0.45677826, Cur Avg Loss: 0.33540979, Log Avg loss: 0.28328149, Global Avg Loss: 1.26290240, Time: 0.0208 Steps: 40390, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002078, Sample Num: 33248, Cur Loss: 0.26137960, Cur Avg Loss: 0.33602173, Log Avg loss: 0.46256975, Global Avg Loss: 1.26270429, Time: 0.0208 Steps: 40400, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002088, Sample Num: 33408, Cur Loss: 0.13460104, Cur Avg Loss: 0.33758781, Log Avg loss: 0.66302076, Global Avg Loss: 1.26255589, Time: 0.0208 Steps: 40410, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002098, Sample Num: 33568, Cur Loss: 0.09514254, Cur Avg Loss: 0.33867001, Log Avg loss: 0.56463217, Global Avg Loss: 1.26238323, Time: 0.0209 Steps: 40420, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002108, Sample Num: 33728, Cur Loss: 0.27403736, Cur Avg Loss: 0.33835724, Log Avg loss: 0.27273888, Global Avg Loss: 1.26213845, Time: 0.0208 Steps: 40430, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002118, Sample Num: 33888, Cur Loss: 0.17879373, Cur Avg Loss: 0.33833346, Log Avg loss: 0.33332043, Global Avg Loss: 1.26190877, Time: 0.0209 Steps: 40440, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002128, Sample Num: 34048, Cur Loss: 0.41247147, Cur Avg Loss: 0.33861844, Log Avg loss: 0.39897613, Global Avg Loss: 1.26169543, Time: 0.0208 Steps: 40450, Updated lr: 0.000063 ***** Running evaluation checkpoint-40451 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-40451 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.831010, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.690286, "eval_total_loss": 485.2712, "eval_mae": 0.732118, "eval_mse": 0.690299, "eval_r2": 0.561201, "eval_sp_statistic": 0.871891, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.903729, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.393935, "test_total_loss": 197.755367, "test_mae": 0.511628, "test_mse": 0.394055, "test_r2": 0.745674, "test_sp_statistic": 0.856591, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.903424, "test_ps_pvalue": 0.0, "lr": 6.258795637743007e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2616701695969232, "train_cur_epoch_loss": 720.8197198174894, "train_cur_epoch_avg_loss": 0.33857196797439615, "train_cur_epoch_time": 44.83101034164429, "train_cur_epoch_avg_time": 0.0210573087560565, "epoch": 19, "step": 40451} ################################################## Training, Epoch: 0020, Batch: 000009, Sample Num: 144, Cur Loss: 0.39429832, Cur Avg Loss: 0.38737544, Log Avg loss: 0.37260656, Global Avg Loss: 1.26147569, Time: 0.0220 Steps: 40460, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000019, Sample Num: 304, Cur Loss: 0.64051646, Cur Avg Loss: 0.45466219, Log Avg loss: 0.51522025, Global Avg Loss: 1.26129129, Time: 0.0219 Steps: 40470, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000029, Sample Num: 464, Cur Loss: 0.56536132, Cur Avg Loss: 0.46585498, Log Avg loss: 0.48712130, Global Avg Loss: 1.26110005, Time: 0.0219 Steps: 40480, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000039, Sample Num: 624, Cur Loss: 0.35597515, Cur Avg Loss: 0.46152146, Log Avg loss: 0.44895425, Global Avg Loss: 1.26089947, Time: 0.0219 Steps: 40490, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000049, Sample Num: 784, Cur Loss: 0.51129562, Cur Avg Loss: 0.46787382, Log Avg loss: 0.49264803, Global Avg Loss: 1.26070977, Time: 0.0219 Steps: 40500, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000059, Sample Num: 944, Cur Loss: 0.05518064, Cur Avg Loss: 0.45670374, Log Avg loss: 0.40197032, Global Avg Loss: 1.26049779, Time: 0.0219 Steps: 40510, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000069, Sample Num: 1104, Cur Loss: 0.23149954, Cur Avg Loss: 0.42318858, Log Avg loss: 0.22544916, Global Avg Loss: 1.26024235, Time: 0.0219 Steps: 40520, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000079, Sample Num: 1264, Cur Loss: 0.33618376, Cur Avg Loss: 0.39254942, Log Avg loss: 0.18113919, Global Avg Loss: 1.25997610, Time: 0.0219 Steps: 40530, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000089, Sample Num: 1424, Cur Loss: 0.50178742, Cur Avg Loss: 0.43299473, Log Avg loss: 0.75251272, Global Avg Loss: 1.25985093, Time: 0.0219 Steps: 40540, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000099, Sample Num: 1584, Cur Loss: 1.35057712, Cur Avg Loss: 0.48133753, Log Avg loss: 0.91158842, Global Avg Loss: 1.25976504, Time: 0.0219 Steps: 40550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000109, Sample Num: 1744, Cur Loss: 0.77927327, Cur Avg Loss: 0.49761830, Log Avg loss: 0.65879794, Global Avg Loss: 1.25961687, Time: 0.0219 Steps: 40560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000119, Sample Num: 1904, Cur Loss: 0.05047435, Cur Avg Loss: 0.47578887, Log Avg loss: 0.23784806, Global Avg Loss: 1.25936502, Time: 0.0219 Steps: 40570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000129, Sample Num: 2064, Cur Loss: 0.28669012, Cur Avg Loss: 0.46312825, Log Avg loss: 0.31246684, Global Avg Loss: 1.25913168, Time: 0.0219 Steps: 40580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000139, Sample Num: 2224, Cur Loss: 0.27614683, Cur Avg Loss: 0.45096760, Log Avg loss: 0.29409532, Global Avg Loss: 1.25889393, Time: 0.0219 Steps: 40590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000149, Sample Num: 2384, Cur Loss: 0.28838885, Cur Avg Loss: 0.44335870, Log Avg loss: 0.33759492, Global Avg Loss: 1.25866701, Time: 0.0219 Steps: 40600, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000159, Sample Num: 2544, Cur Loss: 0.46107626, Cur Avg Loss: 0.43096709, Log Avg loss: 0.24633207, Global Avg Loss: 1.25841772, Time: 0.0219 Steps: 40610, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000169, Sample Num: 2704, Cur Loss: 0.22418898, Cur Avg Loss: 0.42275974, Log Avg loss: 0.29226289, Global Avg Loss: 1.25817987, Time: 0.0219 Steps: 40620, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000179, Sample Num: 2864, Cur Loss: 0.23444280, Cur Avg Loss: 0.41621432, Log Avg loss: 0.30559681, Global Avg Loss: 1.25794542, Time: 0.0219 Steps: 40630, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000189, Sample Num: 3024, Cur Loss: 0.16261628, Cur Avg Loss: 0.41003957, Log Avg loss: 0.29951144, Global Avg Loss: 1.25770958, Time: 0.0219 Steps: 40640, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000199, Sample Num: 3184, Cur Loss: 0.46147379, Cur Avg Loss: 0.40725862, Log Avg loss: 0.35469871, Global Avg Loss: 1.25748744, Time: 0.0219 Steps: 40650, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000209, Sample Num: 3344, Cur Loss: 0.27976453, Cur Avg Loss: 0.39851692, Log Avg loss: 0.22455702, Global Avg Loss: 1.25723340, Time: 0.0219 Steps: 40660, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000219, Sample Num: 3504, Cur Loss: 0.18861884, Cur Avg Loss: 0.39233183, Log Avg loss: 0.26306361, Global Avg Loss: 1.25698895, Time: 0.0219 Steps: 40670, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000229, Sample Num: 3664, Cur Loss: 0.34459502, Cur Avg Loss: 0.38455220, Log Avg loss: 0.21417831, Global Avg Loss: 1.25673261, Time: 0.0219 Steps: 40680, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000239, Sample Num: 3824, Cur Loss: 0.52347463, Cur Avg Loss: 0.37977143, Log Avg loss: 0.27029177, Global Avg Loss: 1.25649018, Time: 0.0219 Steps: 40690, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000249, Sample Num: 3984, Cur Loss: 0.13349630, Cur Avg Loss: 0.37347704, Log Avg loss: 0.22304115, Global Avg Loss: 1.25623626, Time: 0.0219 Steps: 40700, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000259, Sample Num: 4144, Cur Loss: 0.11673966, Cur Avg Loss: 0.36762074, Log Avg loss: 0.22179879, Global Avg Loss: 1.25598216, Time: 0.0245 Steps: 40710, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000269, Sample Num: 4304, Cur Loss: 0.31731603, Cur Avg Loss: 0.36253529, Log Avg loss: 0.23082220, Global Avg Loss: 1.25573040, Time: 0.0208 Steps: 40720, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000279, Sample Num: 4464, Cur Loss: 0.28179091, Cur Avg Loss: 0.35897008, Log Avg loss: 0.26306585, Global Avg Loss: 1.25548669, Time: 0.0208 Steps: 40730, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000289, Sample Num: 4624, Cur Loss: 0.61376905, Cur Avg Loss: 0.35868643, Log Avg loss: 0.35077246, Global Avg Loss: 1.25526461, Time: 0.0208 Steps: 40740, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000299, Sample Num: 4784, Cur Loss: 0.32326123, Cur Avg Loss: 0.35617929, Log Avg loss: 0.28372305, Global Avg Loss: 1.25502620, Time: 0.0208 Steps: 40750, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000309, Sample Num: 4944, Cur Loss: 0.39206219, Cur Avg Loss: 0.36113855, Log Avg loss: 0.50942043, Global Avg Loss: 1.25484327, Time: 0.0208 Steps: 40760, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000319, Sample Num: 5104, Cur Loss: 0.68877119, Cur Avg Loss: 0.36990548, Log Avg loss: 0.64080348, Global Avg Loss: 1.25469266, Time: 0.0208 Steps: 40770, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000329, Sample Num: 5264, Cur Loss: 0.41887921, Cur Avg Loss: 0.36739108, Log Avg loss: 0.28718180, Global Avg Loss: 1.25445541, Time: 0.0208 Steps: 40780, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000339, Sample Num: 5424, Cur Loss: 0.09899575, Cur Avg Loss: 0.36616983, Log Avg loss: 0.32599086, Global Avg Loss: 1.25422779, Time: 0.0209 Steps: 40790, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000349, Sample Num: 5584, Cur Loss: 0.90138948, Cur Avg Loss: 0.36980174, Log Avg loss: 0.49292326, Global Avg Loss: 1.25404120, Time: 0.0208 Steps: 40800, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000359, Sample Num: 5744, Cur Loss: 0.12629896, Cur Avg Loss: 0.36575688, Log Avg loss: 0.22459144, Global Avg Loss: 1.25378894, Time: 0.0208 Steps: 40810, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000369, Sample Num: 5904, Cur Loss: 0.43657321, Cur Avg Loss: 0.36229586, Log Avg loss: 0.23804515, Global Avg Loss: 1.25354011, Time: 0.0208 Steps: 40820, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000379, Sample Num: 6064, Cur Loss: 0.47338617, Cur Avg Loss: 0.36548604, Log Avg loss: 0.48320359, Global Avg Loss: 1.25335144, Time: 0.0208 Steps: 40830, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000389, Sample Num: 6224, Cur Loss: 0.93925345, Cur Avg Loss: 0.36622902, Log Avg loss: 0.39438795, Global Avg Loss: 1.25314111, Time: 0.0208 Steps: 40840, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000399, Sample Num: 6384, Cur Loss: 0.17960732, Cur Avg Loss: 0.36380675, Log Avg loss: 0.26958044, Global Avg Loss: 1.25290034, Time: 0.0208 Steps: 40850, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000409, Sample Num: 6544, Cur Loss: 0.08071530, Cur Avg Loss: 0.36296642, Log Avg loss: 0.32943726, Global Avg Loss: 1.25267433, Time: 0.0208 Steps: 40860, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000419, Sample Num: 6704, Cur Loss: 0.34634781, Cur Avg Loss: 0.36060910, Log Avg loss: 0.26419487, Global Avg Loss: 1.25243247, Time: 0.0208 Steps: 40870, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000429, Sample Num: 6864, Cur Loss: 0.48787171, Cur Avg Loss: 0.35952995, Log Avg loss: 0.31431357, Global Avg Loss: 1.25220299, Time: 0.0208 Steps: 40880, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000439, Sample Num: 7024, Cur Loss: 0.30803594, Cur Avg Loss: 0.35841443, Log Avg loss: 0.31055851, Global Avg Loss: 1.25197271, Time: 0.0208 Steps: 40890, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000449, Sample Num: 7184, Cur Loss: 0.44839394, Cur Avg Loss: 0.35924058, Log Avg loss: 0.39550861, Global Avg Loss: 1.25176330, Time: 0.0209 Steps: 40900, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000459, Sample Num: 7344, Cur Loss: 0.18713894, Cur Avg Loss: 0.35787318, Log Avg loss: 0.29647707, Global Avg Loss: 1.25152979, Time: 0.0208 Steps: 40910, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000469, Sample Num: 7504, Cur Loss: 0.18522215, Cur Avg Loss: 0.35579998, Log Avg loss: 0.26064016, Global Avg Loss: 1.25128764, Time: 0.0208 Steps: 40920, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000479, Sample Num: 7664, Cur Loss: 0.98595303, Cur Avg Loss: 0.35569420, Log Avg loss: 0.35073296, Global Avg Loss: 1.25106762, Time: 0.0208 Steps: 40930, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000489, Sample Num: 7824, Cur Loss: 0.59562874, Cur Avg Loss: 0.35453406, Log Avg loss: 0.29896349, Global Avg Loss: 1.25083506, Time: 0.0208 Steps: 40940, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000499, Sample Num: 7984, Cur Loss: 0.21028945, Cur Avg Loss: 0.35460788, Log Avg loss: 0.35821763, Global Avg Loss: 1.25061708, Time: 0.0208 Steps: 40950, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000509, Sample Num: 8144, Cur Loss: 0.48552090, Cur Avg Loss: 0.35491595, Log Avg loss: 0.37028841, Global Avg Loss: 1.25040215, Time: 0.0208 Steps: 40960, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000519, Sample Num: 8304, Cur Loss: 0.48260888, Cur Avg Loss: 0.35728493, Log Avg loss: 0.47786616, Global Avg Loss: 1.25021359, Time: 0.0209 Steps: 40970, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000529, Sample Num: 8464, Cur Loss: 0.11305172, Cur Avg Loss: 0.35481097, Log Avg loss: 0.22641243, Global Avg Loss: 1.24996376, Time: 0.0208 Steps: 40980, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000539, Sample Num: 8624, Cur Loss: 0.68912029, Cur Avg Loss: 0.35446909, Log Avg loss: 0.33638343, Global Avg Loss: 1.24974088, Time: 0.0208 Steps: 40990, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000549, Sample Num: 8784, Cur Loss: 0.14244792, Cur Avg Loss: 0.35222784, Log Avg loss: 0.23142451, Global Avg Loss: 1.24949251, Time: 0.0208 Steps: 41000, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000559, Sample Num: 8944, Cur Loss: 0.13818184, Cur Avg Loss: 0.35190975, Log Avg loss: 0.33444660, Global Avg Loss: 1.24926939, Time: 0.0208 Steps: 41010, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000569, Sample Num: 9104, Cur Loss: 0.51960468, Cur Avg Loss: 0.35268154, Log Avg loss: 0.39582471, Global Avg Loss: 1.24906133, Time: 0.0208 Steps: 41020, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000579, Sample Num: 9264, Cur Loss: 0.45628345, Cur Avg Loss: 0.35130171, Log Avg loss: 0.27278933, Global Avg Loss: 1.24882339, Time: 0.0208 Steps: 41030, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000589, Sample Num: 9424, Cur Loss: 0.60967880, Cur Avg Loss: 0.35035110, Log Avg loss: 0.29531083, Global Avg Loss: 1.24859105, Time: 0.0208 Steps: 41040, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000599, Sample Num: 9584, Cur Loss: 0.44013000, Cur Avg Loss: 0.35048299, Log Avg loss: 0.35825162, Global Avg Loss: 1.24837416, Time: 0.0208 Steps: 41050, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000609, Sample Num: 9744, Cur Loss: 0.20097288, Cur Avg Loss: 0.34928653, Log Avg loss: 0.27761820, Global Avg Loss: 1.24813774, Time: 0.0208 Steps: 41060, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000619, Sample Num: 9904, Cur Loss: 0.44591475, Cur Avg Loss: 0.34962687, Log Avg loss: 0.37035374, Global Avg Loss: 1.24792401, Time: 0.0208 Steps: 41070, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000629, Sample Num: 10064, Cur Loss: 0.40216780, Cur Avg Loss: 0.34812333, Log Avg loss: 0.25505424, Global Avg Loss: 1.24768232, Time: 0.0208 Steps: 41080, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000639, Sample Num: 10224, Cur Loss: 0.16166906, Cur Avg Loss: 0.34727051, Log Avg loss: 0.29362804, Global Avg Loss: 1.24745013, Time: 0.0208 Steps: 41090, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000649, Sample Num: 10384, Cur Loss: 0.06375254, Cur Avg Loss: 0.34456182, Log Avg loss: 0.17147629, Global Avg Loss: 1.24718834, Time: 0.0208 Steps: 41100, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000659, Sample Num: 10544, Cur Loss: 0.19976746, Cur Avg Loss: 0.34330169, Log Avg loss: 0.26151937, Global Avg Loss: 1.24694857, Time: 0.0208 Steps: 41110, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000669, Sample Num: 10704, Cur Loss: 0.23469377, Cur Avg Loss: 0.34284256, Log Avg loss: 0.31258587, Global Avg Loss: 1.24672134, Time: 0.0208 Steps: 41120, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000679, Sample Num: 10864, Cur Loss: 0.27401924, Cur Avg Loss: 0.34191115, Log Avg loss: 0.27960011, Global Avg Loss: 1.24648621, Time: 0.0208 Steps: 41130, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000689, Sample Num: 11024, Cur Loss: 0.81542814, Cur Avg Loss: 0.34100227, Log Avg loss: 0.27928941, Global Avg Loss: 1.24625111, Time: 0.0208 Steps: 41140, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000699, Sample Num: 11184, Cur Loss: 0.34297541, Cur Avg Loss: 0.34027255, Log Avg loss: 0.28999483, Global Avg Loss: 1.24601873, Time: 0.0208 Steps: 41150, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000709, Sample Num: 11344, Cur Loss: 0.28058618, Cur Avg Loss: 0.34042932, Log Avg loss: 0.35138708, Global Avg Loss: 1.24580137, Time: 0.0208 Steps: 41160, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000719, Sample Num: 11504, Cur Loss: 0.06308825, Cur Avg Loss: 0.34034642, Log Avg loss: 0.33446913, Global Avg Loss: 1.24558001, Time: 0.0210 Steps: 41170, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000729, Sample Num: 11664, Cur Loss: 0.20030157, Cur Avg Loss: 0.33925025, Log Avg loss: 0.26043540, Global Avg Loss: 1.24534078, Time: 0.0208 Steps: 41180, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000739, Sample Num: 11824, Cur Loss: 0.13631843, Cur Avg Loss: 0.33959590, Log Avg loss: 0.36479372, Global Avg Loss: 1.24512701, Time: 0.0208 Steps: 41190, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000749, Sample Num: 11984, Cur Loss: 0.42854881, Cur Avg Loss: 0.33963289, Log Avg loss: 0.34236638, Global Avg Loss: 1.24490789, Time: 0.0208 Steps: 41200, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000759, Sample Num: 12144, Cur Loss: 0.20299667, Cur Avg Loss: 0.33838264, Log Avg loss: 0.24473933, Global Avg Loss: 1.24466519, Time: 0.0208 Steps: 41210, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000769, Sample Num: 12304, Cur Loss: 0.16462868, Cur Avg Loss: 0.33723918, Log Avg loss: 0.25045066, Global Avg Loss: 1.24442399, Time: 0.0245 Steps: 41220, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000779, Sample Num: 12464, Cur Loss: 0.37511441, Cur Avg Loss: 0.33596936, Log Avg loss: 0.23831965, Global Avg Loss: 1.24417997, Time: 0.0209 Steps: 41230, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000789, Sample Num: 12624, Cur Loss: 0.14478977, Cur Avg Loss: 0.33547667, Log Avg loss: 0.29709620, Global Avg Loss: 1.24395032, Time: 0.0209 Steps: 41240, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000799, Sample Num: 12784, Cur Loss: 0.35499495, Cur Avg Loss: 0.33505591, Log Avg loss: 0.30185834, Global Avg Loss: 1.24372193, Time: 0.0209 Steps: 41250, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000809, Sample Num: 12944, Cur Loss: 0.29679930, Cur Avg Loss: 0.33432223, Log Avg loss: 0.27570147, Global Avg Loss: 1.24348732, Time: 0.0209 Steps: 41260, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000819, Sample Num: 13104, Cur Loss: 0.51586157, Cur Avg Loss: 0.33378023, Log Avg loss: 0.28993211, Global Avg Loss: 1.24325626, Time: 0.0209 Steps: 41270, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000829, Sample Num: 13264, Cur Loss: 0.27676055, Cur Avg Loss: 0.33261226, Log Avg loss: 0.23695550, Global Avg Loss: 1.24301249, Time: 0.0209 Steps: 41280, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000839, Sample Num: 13424, Cur Loss: 0.51809007, Cur Avg Loss: 0.33252670, Log Avg loss: 0.32543354, Global Avg Loss: 1.24279026, Time: 0.0209 Steps: 41290, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000849, Sample Num: 13584, Cur Loss: 0.13097735, Cur Avg Loss: 0.33174057, Log Avg loss: 0.26578438, Global Avg Loss: 1.24255370, Time: 0.0209 Steps: 41300, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000859, Sample Num: 13744, Cur Loss: 0.17332003, Cur Avg Loss: 0.33032393, Log Avg loss: 0.21005161, Global Avg Loss: 1.24230376, Time: 0.0209 Steps: 41310, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000869, Sample Num: 13904, Cur Loss: 0.41794845, Cur Avg Loss: 0.33028434, Log Avg loss: 0.32688331, Global Avg Loss: 1.24208221, Time: 0.0209 Steps: 41320, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000879, Sample Num: 14064, Cur Loss: 0.14805624, Cur Avg Loss: 0.32943640, Log Avg loss: 0.25575044, Global Avg Loss: 1.24184357, Time: 0.0209 Steps: 41330, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000889, Sample Num: 14224, Cur Loss: 0.04610890, Cur Avg Loss: 0.32818180, Log Avg loss: 0.21790266, Global Avg Loss: 1.24159588, Time: 0.0209 Steps: 41340, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000899, Sample Num: 14384, Cur Loss: 0.11353780, Cur Avg Loss: 0.32769907, Log Avg loss: 0.28478367, Global Avg Loss: 1.24136449, Time: 0.0209 Steps: 41350, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000909, Sample Num: 14544, Cur Loss: 0.16011621, Cur Avg Loss: 0.32794174, Log Avg loss: 0.34975859, Global Avg Loss: 1.24114891, Time: 0.0209 Steps: 41360, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000919, Sample Num: 14704, Cur Loss: 0.19822724, Cur Avg Loss: 0.32749490, Log Avg loss: 0.28687651, Global Avg Loss: 1.24091825, Time: 0.0209 Steps: 41370, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000929, Sample Num: 14864, Cur Loss: 0.31673288, Cur Avg Loss: 0.32733047, Log Avg loss: 0.31221987, Global Avg Loss: 1.24069381, Time: 0.0209 Steps: 41380, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000939, Sample Num: 15024, Cur Loss: 0.09499842, Cur Avg Loss: 0.32594803, Log Avg loss: 0.19751931, Global Avg Loss: 1.24044178, Time: 0.0209 Steps: 41390, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000949, Sample Num: 15184, Cur Loss: 0.11292298, Cur Avg Loss: 0.32615361, Log Avg loss: 0.34545742, Global Avg Loss: 1.24022560, Time: 0.0209 Steps: 41400, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000959, Sample Num: 15344, Cur Loss: 0.25407153, Cur Avg Loss: 0.32578801, Log Avg loss: 0.29109206, Global Avg Loss: 1.23999640, Time: 0.0209 Steps: 41410, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000969, Sample Num: 15504, Cur Loss: 0.23499373, Cur Avg Loss: 0.32496528, Log Avg loss: 0.24606634, Global Avg Loss: 1.23975643, Time: 0.0209 Steps: 41420, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000979, Sample Num: 15664, Cur Loss: 0.65563148, Cur Avg Loss: 0.32538578, Log Avg loss: 0.36613135, Global Avg Loss: 1.23954556, Time: 0.0209 Steps: 41430, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000989, Sample Num: 15824, Cur Loss: 0.73293138, Cur Avg Loss: 0.32675065, Log Avg loss: 0.46037154, Global Avg Loss: 1.23935754, Time: 0.0208 Steps: 41440, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000999, Sample Num: 15984, Cur Loss: 0.73851478, Cur Avg Loss: 0.32734322, Log Avg loss: 0.38594863, Global Avg Loss: 1.23915165, Time: 0.0209 Steps: 41450, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001009, Sample Num: 16144, Cur Loss: 0.45303637, Cur Avg Loss: 0.32823989, Log Avg loss: 0.41781681, Global Avg Loss: 1.23895355, Time: 0.0209 Steps: 41460, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001019, Sample Num: 16304, Cur Loss: 0.54509753, Cur Avg Loss: 0.32793229, Log Avg loss: 0.29689561, Global Avg Loss: 1.23872638, Time: 0.0209 Steps: 41470, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001029, Sample Num: 16464, Cur Loss: 0.68379462, Cur Avg Loss: 0.32930720, Log Avg loss: 0.46941123, Global Avg Loss: 1.23854091, Time: 0.0246 Steps: 41480, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001039, Sample Num: 16624, Cur Loss: 0.16337426, Cur Avg Loss: 0.32801884, Log Avg loss: 0.19544605, Global Avg Loss: 1.23828951, Time: 0.0209 Steps: 41490, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001049, Sample Num: 16784, Cur Loss: 0.25143078, Cur Avg Loss: 0.32672572, Log Avg loss: 0.19237036, Global Avg Loss: 1.23803748, Time: 0.0210 Steps: 41500, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001059, Sample Num: 16944, Cur Loss: 0.27429068, Cur Avg Loss: 0.32616733, Log Avg loss: 0.26759232, Global Avg Loss: 1.23780369, Time: 0.0209 Steps: 41510, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001069, Sample Num: 17104, Cur Loss: 0.29904303, Cur Avg Loss: 0.32636357, Log Avg loss: 0.34714557, Global Avg Loss: 1.23758918, Time: 0.0209 Steps: 41520, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001079, Sample Num: 17264, Cur Loss: 0.10178015, Cur Avg Loss: 0.32659718, Log Avg loss: 0.35156986, Global Avg Loss: 1.23737583, Time: 0.0209 Steps: 41530, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001089, Sample Num: 17424, Cur Loss: 0.16506937, Cur Avg Loss: 0.32722301, Log Avg loss: 0.39475087, Global Avg Loss: 1.23717299, Time: 0.0209 Steps: 41540, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001099, Sample Num: 17584, Cur Loss: 0.18657143, Cur Avg Loss: 0.32656441, Log Avg loss: 0.25484252, Global Avg Loss: 1.23693657, Time: 0.0209 Steps: 41550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001109, Sample Num: 17744, Cur Loss: 0.56306994, Cur Avg Loss: 0.32595193, Log Avg loss: 0.25863995, Global Avg Loss: 1.23670117, Time: 0.0209 Steps: 41560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001119, Sample Num: 17904, Cur Loss: 0.18739918, Cur Avg Loss: 0.32482072, Log Avg loss: 0.19937000, Global Avg Loss: 1.23645163, Time: 0.0209 Steps: 41570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001129, Sample Num: 18064, Cur Loss: 0.75698227, Cur Avg Loss: 0.32605857, Log Avg loss: 0.46457334, Global Avg Loss: 1.23626600, Time: 0.0209 Steps: 41580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001139, Sample Num: 18224, Cur Loss: 0.08831121, Cur Avg Loss: 0.32505721, Log Avg loss: 0.21200422, Global Avg Loss: 1.23601972, Time: 0.0209 Steps: 41590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001149, Sample Num: 18384, Cur Loss: 0.21894509, Cur Avg Loss: 0.32478594, Log Avg loss: 0.29388829, Global Avg Loss: 1.23579325, Time: 0.0209 Steps: 41600, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001159, Sample Num: 18544, Cur Loss: 0.30842805, Cur Avg Loss: 0.32486602, Log Avg loss: 0.33406739, Global Avg Loss: 1.23557654, Time: 0.0209 Steps: 41610, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001169, Sample Num: 18704, Cur Loss: 0.75688213, Cur Avg Loss: 0.32470486, Log Avg loss: 0.30602601, Global Avg Loss: 1.23535320, Time: 0.0209 Steps: 41620, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001179, Sample Num: 18864, Cur Loss: 0.31879193, Cur Avg Loss: 0.32473306, Log Avg loss: 0.32803014, Global Avg Loss: 1.23513525, Time: 0.0209 Steps: 41630, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001189, Sample Num: 19024, Cur Loss: 0.51028836, Cur Avg Loss: 0.32760026, Log Avg loss: 0.66564326, Global Avg Loss: 1.23499848, Time: 0.0209 Steps: 41640, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001199, Sample Num: 19184, Cur Loss: 0.32648677, Cur Avg Loss: 0.32675016, Log Avg loss: 0.22567262, Global Avg Loss: 1.23475615, Time: 0.0209 Steps: 41650, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001209, Sample Num: 19344, Cur Loss: 0.26605135, Cur Avg Loss: 0.32643546, Log Avg loss: 0.28870262, Global Avg Loss: 1.23452906, Time: 0.0209 Steps: 41660, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001219, Sample Num: 19504, Cur Loss: 0.28794360, Cur Avg Loss: 0.32649967, Log Avg loss: 0.33426341, Global Avg Loss: 1.23431301, Time: 0.0209 Steps: 41670, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001229, Sample Num: 19664, Cur Loss: 0.04490963, Cur Avg Loss: 0.32587181, Log Avg loss: 0.24933556, Global Avg Loss: 1.23407669, Time: 0.0209 Steps: 41680, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001239, Sample Num: 19824, Cur Loss: 0.28182095, Cur Avg Loss: 0.32589251, Log Avg loss: 0.32843614, Global Avg Loss: 1.23385946, Time: 0.0209 Steps: 41690, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001249, Sample Num: 19984, Cur Loss: 0.23020692, Cur Avg Loss: 0.32550479, Log Avg loss: 0.27746587, Global Avg Loss: 1.23363011, Time: 0.0209 Steps: 41700, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001259, Sample Num: 20144, Cur Loss: 0.26705545, Cur Avg Loss: 0.32467517, Log Avg loss: 0.22105586, Global Avg Loss: 1.23338734, Time: 0.0209 Steps: 41710, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001269, Sample Num: 20304, Cur Loss: 0.35360870, Cur Avg Loss: 0.32445898, Log Avg loss: 0.29724052, Global Avg Loss: 1.23316295, Time: 0.0209 Steps: 41720, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001279, Sample Num: 20464, Cur Loss: 0.25473547, Cur Avg Loss: 0.32438993, Log Avg loss: 0.31562838, Global Avg Loss: 1.23294308, Time: 0.0209 Steps: 41730, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001289, Sample Num: 20624, Cur Loss: 0.12890916, Cur Avg Loss: 0.32372185, Log Avg loss: 0.23827467, Global Avg Loss: 1.23270478, Time: 0.0211 Steps: 41740, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001299, Sample Num: 20784, Cur Loss: 0.07123163, Cur Avg Loss: 0.32328585, Log Avg loss: 0.26708535, Global Avg Loss: 1.23247349, Time: 0.0210 Steps: 41750, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001309, Sample Num: 20944, Cur Loss: 0.32128730, Cur Avg Loss: 0.32299435, Log Avg loss: 0.28512834, Global Avg Loss: 1.23224664, Time: 0.0210 Steps: 41760, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001319, Sample Num: 21104, Cur Loss: 0.44409978, Cur Avg Loss: 0.32283161, Log Avg loss: 0.30152800, Global Avg Loss: 1.23202382, Time: 0.0209 Steps: 41770, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001329, Sample Num: 21264, Cur Loss: 0.41219184, Cur Avg Loss: 0.32214660, Log Avg loss: 0.23179469, Global Avg Loss: 1.23178442, Time: 0.0210 Steps: 41780, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001339, Sample Num: 21424, Cur Loss: 0.18447304, Cur Avg Loss: 0.32198582, Log Avg loss: 0.30061722, Global Avg Loss: 1.23156159, Time: 0.0210 Steps: 41790, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001349, Sample Num: 21584, Cur Loss: 0.10641942, Cur Avg Loss: 0.32238871, Log Avg loss: 0.37633674, Global Avg Loss: 1.23135700, Time: 0.0210 Steps: 41800, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001359, Sample Num: 21744, Cur Loss: 0.20877257, Cur Avg Loss: 0.32361544, Log Avg loss: 0.48910112, Global Avg Loss: 1.23117946, Time: 0.0210 Steps: 41810, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001369, Sample Num: 21904, Cur Loss: 0.35695073, Cur Avg Loss: 0.32333795, Log Avg loss: 0.28562739, Global Avg Loss: 1.23095336, Time: 0.0210 Steps: 41820, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001379, Sample Num: 22064, Cur Loss: 0.19914402, Cur Avg Loss: 0.32238808, Log Avg loss: 0.19235020, Global Avg Loss: 1.23070507, Time: 0.0210 Steps: 41830, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001389, Sample Num: 22224, Cur Loss: 0.21287805, Cur Avg Loss: 0.32254906, Log Avg loss: 0.34474870, Global Avg Loss: 1.23049332, Time: 0.0210 Steps: 41840, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001399, Sample Num: 22384, Cur Loss: 0.07512668, Cur Avg Loss: 0.32312371, Log Avg loss: 0.40294154, Global Avg Loss: 1.23029558, Time: 0.0210 Steps: 41850, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001409, Sample Num: 22544, Cur Loss: 0.22312202, Cur Avg Loss: 0.32336525, Log Avg loss: 0.35715788, Global Avg Loss: 1.23008700, Time: 0.0210 Steps: 41860, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001419, Sample Num: 22704, Cur Loss: 0.62930346, Cur Avg Loss: 0.32388111, Log Avg loss: 0.39656509, Global Avg Loss: 1.22988792, Time: 0.0210 Steps: 41870, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001429, Sample Num: 22864, Cur Loss: 0.21393445, Cur Avg Loss: 0.32374359, Log Avg loss: 0.30422935, Global Avg Loss: 1.22966690, Time: 0.0209 Steps: 41880, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001439, Sample Num: 23024, Cur Loss: 0.35927737, Cur Avg Loss: 0.32410865, Log Avg loss: 0.37627664, Global Avg Loss: 1.22946317, Time: 0.0210 Steps: 41890, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001449, Sample Num: 23184, Cur Loss: 0.09541951, Cur Avg Loss: 0.32404794, Log Avg loss: 0.31531172, Global Avg Loss: 1.22924500, Time: 0.0210 Steps: 41900, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001459, Sample Num: 23344, Cur Loss: 0.15764642, Cur Avg Loss: 0.32350281, Log Avg loss: 0.24451233, Global Avg Loss: 1.22901004, Time: 0.0209 Steps: 41910, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001469, Sample Num: 23504, Cur Loss: 0.27505580, Cur Avg Loss: 0.32296651, Log Avg loss: 0.24472160, Global Avg Loss: 1.22877523, Time: 0.0210 Steps: 41920, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001479, Sample Num: 23664, Cur Loss: 0.66892910, Cur Avg Loss: 0.32288921, Log Avg loss: 0.31153369, Global Avg Loss: 1.22855648, Time: 0.0210 Steps: 41930, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001489, Sample Num: 23824, Cur Loss: 0.66539502, Cur Avg Loss: 0.32373346, Log Avg loss: 0.44859766, Global Avg Loss: 1.22837051, Time: 0.0210 Steps: 41940, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001499, Sample Num: 23984, Cur Loss: 0.25340062, Cur Avg Loss: 0.32389920, Log Avg loss: 0.34857746, Global Avg Loss: 1.22816078, Time: 0.0209 Steps: 41950, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001509, Sample Num: 24144, Cur Loss: 0.08815596, Cur Avg Loss: 0.32440989, Log Avg loss: 0.40096274, Global Avg Loss: 1.22796365, Time: 0.0210 Steps: 41960, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001519, Sample Num: 24304, Cur Loss: 0.33912045, Cur Avg Loss: 0.32396384, Log Avg loss: 0.25665427, Global Avg Loss: 1.22773222, Time: 0.0210 Steps: 41970, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001529, Sample Num: 24464, Cur Loss: 0.24949260, Cur Avg Loss: 0.32340782, Log Avg loss: 0.23894869, Global Avg Loss: 1.22749668, Time: 0.0210 Steps: 41980, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001539, Sample Num: 24624, Cur Loss: 0.22532576, Cur Avg Loss: 0.32344961, Log Avg loss: 0.32984007, Global Avg Loss: 1.22728290, Time: 0.0247 Steps: 41990, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001549, Sample Num: 24784, Cur Loss: 0.16161075, Cur Avg Loss: 0.32304949, Log Avg loss: 0.26147088, Global Avg Loss: 1.22705295, Time: 0.0209 Steps: 42000, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001559, Sample Num: 24944, Cur Loss: 0.40424797, Cur Avg Loss: 0.32298367, Log Avg loss: 0.31278739, Global Avg Loss: 1.22683531, Time: 0.0210 Steps: 42010, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001569, Sample Num: 25104, Cur Loss: 0.08088782, Cur Avg Loss: 0.32247092, Log Avg loss: 0.24253371, Global Avg Loss: 1.22660107, Time: 0.0209 Steps: 42020, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001579, Sample Num: 25264, Cur Loss: 0.34534520, Cur Avg Loss: 0.32252964, Log Avg loss: 0.33174285, Global Avg Loss: 1.22638816, Time: 0.0210 Steps: 42030, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001589, Sample Num: 25424, Cur Loss: 0.46006307, Cur Avg Loss: 0.32251810, Log Avg loss: 0.32069597, Global Avg Loss: 1.22617272, Time: 0.0210 Steps: 42040, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001599, Sample Num: 25584, Cur Loss: 0.17446536, Cur Avg Loss: 0.32167087, Log Avg loss: 0.18704635, Global Avg Loss: 1.22592561, Time: 0.0209 Steps: 42050, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001609, Sample Num: 25744, Cur Loss: 0.10334297, Cur Avg Loss: 0.32095684, Log Avg loss: 0.20678338, Global Avg Loss: 1.22568330, Time: 0.0209 Steps: 42060, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001619, Sample Num: 25904, Cur Loss: 0.21359205, Cur Avg Loss: 0.32058764, Log Avg loss: 0.26118321, Global Avg Loss: 1.22545404, Time: 0.0209 Steps: 42070, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001629, Sample Num: 26064, Cur Loss: 0.15255155, Cur Avg Loss: 0.31995387, Log Avg loss: 0.21734621, Global Avg Loss: 1.22521447, Time: 0.0210 Steps: 42080, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001639, Sample Num: 26224, Cur Loss: 0.14013390, Cur Avg Loss: 0.31962344, Log Avg loss: 0.26579665, Global Avg Loss: 1.22498653, Time: 0.0210 Steps: 42090, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001649, Sample Num: 26384, Cur Loss: 0.54465944, Cur Avg Loss: 0.31980778, Log Avg loss: 0.35002056, Global Avg Loss: 1.22477869, Time: 0.0209 Steps: 42100, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001659, Sample Num: 26544, Cur Loss: 0.28212678, Cur Avg Loss: 0.31947088, Log Avg loss: 0.26391688, Global Avg Loss: 1.22455052, Time: 0.0209 Steps: 42110, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001669, Sample Num: 26704, Cur Loss: 0.55402708, Cur Avg Loss: 0.31944055, Log Avg loss: 0.31440760, Global Avg Loss: 1.22433443, Time: 0.0209 Steps: 42120, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001679, Sample Num: 26864, Cur Loss: 0.21056685, Cur Avg Loss: 0.31953037, Log Avg loss: 0.33452168, Global Avg Loss: 1.22412323, Time: 0.0209 Steps: 42130, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001689, Sample Num: 27024, Cur Loss: 0.21170801, Cur Avg Loss: 0.31913922, Log Avg loss: 0.25346585, Global Avg Loss: 1.22389289, Time: 0.0210 Steps: 42140, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001699, Sample Num: 27184, Cur Loss: 0.48396632, Cur Avg Loss: 0.32039341, Log Avg loss: 0.53222520, Global Avg Loss: 1.22372879, Time: 0.0210 Steps: 42150, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001709, Sample Num: 27344, Cur Loss: 0.24738094, Cur Avg Loss: 0.32094539, Log Avg loss: 0.41472710, Global Avg Loss: 1.22353690, Time: 0.0209 Steps: 42160, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001719, Sample Num: 27504, Cur Loss: 0.25735709, Cur Avg Loss: 0.32095154, Log Avg loss: 0.32200334, Global Avg Loss: 1.22332311, Time: 0.0210 Steps: 42170, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001729, Sample Num: 27664, Cur Loss: 0.35515481, Cur Avg Loss: 0.32090080, Log Avg loss: 0.31217811, Global Avg Loss: 1.22310710, Time: 0.0209 Steps: 42180, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001739, Sample Num: 27824, Cur Loss: 0.55586207, Cur Avg Loss: 0.32107617, Log Avg loss: 0.35139794, Global Avg Loss: 1.22290049, Time: 0.0209 Steps: 42190, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001749, Sample Num: 27984, Cur Loss: 0.35303980, Cur Avg Loss: 0.32074766, Log Avg loss: 0.26361848, Global Avg Loss: 1.22267317, Time: 0.0209 Steps: 42200, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001759, Sample Num: 28144, Cur Loss: 0.20497683, Cur Avg Loss: 0.32037403, Log Avg loss: 0.25502691, Global Avg Loss: 1.22244392, Time: 0.0209 Steps: 42210, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001769, Sample Num: 28304, Cur Loss: 0.27388147, Cur Avg Loss: 0.31974624, Log Avg loss: 0.20931836, Global Avg Loss: 1.22220396, Time: 0.0209 Steps: 42220, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001779, Sample Num: 28464, Cur Loss: 0.12700081, Cur Avg Loss: 0.32034079, Log Avg loss: 0.42551554, Global Avg Loss: 1.22201530, Time: 0.0209 Steps: 42230, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001789, Sample Num: 28624, Cur Loss: 0.76800871, Cur Avg Loss: 0.32034389, Log Avg loss: 0.32089616, Global Avg Loss: 1.22180197, Time: 0.0209 Steps: 42240, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001799, Sample Num: 28784, Cur Loss: 0.80937481, Cur Avg Loss: 0.32210196, Log Avg loss: 0.63662109, Global Avg Loss: 1.22166347, Time: 0.0210 Steps: 42250, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001809, Sample Num: 28944, Cur Loss: 0.34762669, Cur Avg Loss: 0.32302492, Log Avg loss: 0.48906561, Global Avg Loss: 1.22149011, Time: 0.0209 Steps: 42260, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001819, Sample Num: 29104, Cur Loss: 0.17060840, Cur Avg Loss: 0.32301280, Log Avg loss: 0.32081974, Global Avg Loss: 1.22127704, Time: 0.0209 Steps: 42270, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001829, Sample Num: 29264, Cur Loss: 0.14001626, Cur Avg Loss: 0.32278167, Log Avg loss: 0.28073981, Global Avg Loss: 1.22105458, Time: 0.0210 Steps: 42280, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001839, Sample Num: 29424, Cur Loss: 0.65094799, Cur Avg Loss: 0.32276371, Log Avg loss: 0.31947777, Global Avg Loss: 1.22084139, Time: 0.0209 Steps: 42290, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001849, Sample Num: 29584, Cur Loss: 0.25356466, Cur Avg Loss: 0.32276203, Log Avg loss: 0.32245308, Global Avg Loss: 1.22062901, Time: 0.0210 Steps: 42300, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001859, Sample Num: 29744, Cur Loss: 0.88385314, Cur Avg Loss: 0.32309219, Log Avg loss: 0.38413872, Global Avg Loss: 1.22043130, Time: 0.0209 Steps: 42310, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001869, Sample Num: 29904, Cur Loss: 0.53739625, Cur Avg Loss: 0.32289305, Log Avg loss: 0.28587382, Global Avg Loss: 1.22021047, Time: 0.0210 Steps: 42320, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001879, Sample Num: 30064, Cur Loss: 0.35099742, Cur Avg Loss: 0.32307390, Log Avg loss: 0.35687373, Global Avg Loss: 1.22000652, Time: 0.0210 Steps: 42330, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001889, Sample Num: 30224, Cur Loss: 0.11591147, Cur Avg Loss: 0.32302475, Log Avg loss: 0.31379049, Global Avg Loss: 1.21979248, Time: 0.0209 Steps: 42340, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001899, Sample Num: 30384, Cur Loss: 0.19654553, Cur Avg Loss: 0.32236142, Log Avg loss: 0.19705712, Global Avg Loss: 1.21955099, Time: 0.0209 Steps: 42350, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001909, Sample Num: 30544, Cur Loss: 0.22554100, Cur Avg Loss: 0.32251408, Log Avg loss: 0.35150449, Global Avg Loss: 1.21934607, Time: 0.0209 Steps: 42360, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001919, Sample Num: 30704, Cur Loss: 0.20960021, Cur Avg Loss: 0.32183398, Log Avg loss: 0.19200276, Global Avg Loss: 1.21910360, Time: 0.0209 Steps: 42370, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001929, Sample Num: 30864, Cur Loss: 0.11006224, Cur Avg Loss: 0.32158332, Log Avg loss: 0.27348276, Global Avg Loss: 1.21888047, Time: 0.0209 Steps: 42380, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001939, Sample Num: 31024, Cur Loss: 0.51194942, Cur Avg Loss: 0.32091904, Log Avg loss: 0.19277948, Global Avg Loss: 1.21863841, Time: 0.0209 Steps: 42390, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001949, Sample Num: 31184, Cur Loss: 0.35214162, Cur Avg Loss: 0.32081532, Log Avg loss: 0.30070284, Global Avg Loss: 1.21842191, Time: 0.0209 Steps: 42400, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001959, Sample Num: 31344, Cur Loss: 0.23977078, Cur Avg Loss: 0.32029690, Log Avg loss: 0.21925833, Global Avg Loss: 1.21818632, Time: 0.0209 Steps: 42410, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001969, Sample Num: 31504, Cur Loss: 0.17720339, Cur Avg Loss: 0.32029814, Log Avg loss: 0.32053957, Global Avg Loss: 1.21797471, Time: 0.0209 Steps: 42420, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001979, Sample Num: 31664, Cur Loss: 0.09053652, Cur Avg Loss: 0.32029298, Log Avg loss: 0.31927855, Global Avg Loss: 1.21776290, Time: 0.0209 Steps: 42430, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001989, Sample Num: 31824, Cur Loss: 0.32280892, Cur Avg Loss: 0.32039058, Log Avg loss: 0.33970508, Global Avg Loss: 1.21755601, Time: 0.0209 Steps: 42440, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001999, Sample Num: 31984, Cur Loss: 0.17815503, Cur Avg Loss: 0.32128931, Log Avg loss: 0.50004549, Global Avg Loss: 1.21738698, Time: 0.0209 Steps: 42450, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002009, Sample Num: 32144, Cur Loss: 0.27033544, Cur Avg Loss: 0.32127599, Log Avg loss: 0.31861346, Global Avg Loss: 1.21717531, Time: 0.0209 Steps: 42460, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002019, Sample Num: 32304, Cur Loss: 0.14437270, Cur Avg Loss: 0.32160559, Log Avg loss: 0.38782381, Global Avg Loss: 1.21698003, Time: 0.0209 Steps: 42470, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002029, Sample Num: 32464, Cur Loss: 0.08859125, Cur Avg Loss: 0.32166933, Log Avg loss: 0.33453842, Global Avg Loss: 1.21677230, Time: 0.0209 Steps: 42480, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002039, Sample Num: 32624, Cur Loss: 0.22669546, Cur Avg Loss: 0.32210446, Log Avg loss: 0.41039239, Global Avg Loss: 1.21658251, Time: 0.0209 Steps: 42490, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002049, Sample Num: 32784, Cur Loss: 0.17494386, Cur Avg Loss: 0.32188669, Log Avg loss: 0.27748292, Global Avg Loss: 1.21636155, Time: 0.0246 Steps: 42500, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002059, Sample Num: 32944, Cur Loss: 0.39495146, Cur Avg Loss: 0.32167498, Log Avg loss: 0.27829618, Global Avg Loss: 1.21614088, Time: 0.0209 Steps: 42510, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002069, Sample Num: 33104, Cur Loss: 0.95484483, Cur Avg Loss: 0.32182001, Log Avg loss: 0.35168060, Global Avg Loss: 1.21593757, Time: 0.0209 Steps: 42520, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002079, Sample Num: 33264, Cur Loss: 0.35848156, Cur Avg Loss: 0.32210550, Log Avg loss: 0.38117376, Global Avg Loss: 1.21574130, Time: 0.0209 Steps: 42530, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002089, Sample Num: 33424, Cur Loss: 0.14995895, Cur Avg Loss: 0.32163723, Log Avg loss: 0.22428360, Global Avg Loss: 1.21550823, Time: 0.0209 Steps: 42540, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002099, Sample Num: 33584, Cur Loss: 0.18843874, Cur Avg Loss: 0.32108242, Log Avg loss: 0.20518209, Global Avg Loss: 1.21527079, Time: 0.0209 Steps: 42550, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002109, Sample Num: 33744, Cur Loss: 0.20867184, Cur Avg Loss: 0.32150806, Log Avg loss: 0.41084943, Global Avg Loss: 1.21508178, Time: 0.0209 Steps: 42560, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002119, Sample Num: 33904, Cur Loss: 0.52008963, Cur Avg Loss: 0.32115772, Log Avg loss: 0.24727256, Global Avg Loss: 1.21485443, Time: 0.0209 Steps: 42570, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002129, Sample Num: 34055, Cur Loss: 0.02902981, Cur Avg Loss: 0.32096124, Log Avg loss: 0.27932737, Global Avg Loss: 1.21463472, Time: 0.0101 Steps: 42580, Updated lr: 0.000061 ***** Running evaluation checkpoint-42580 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-42580 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.843097, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.433716, "eval_total_loss": 304.902228, "eval_mae": 0.546154, "eval_mse": 0.433745, "eval_r2": 0.724283, "eval_sp_statistic": 0.874768, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.906416, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.342671, "test_total_loss": 172.020904, "test_mae": 0.428873, "test_mse": 0.342787, "test_r2": 0.778762, "test_sp_statistic": 0.861918, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.909984, "test_ps_pvalue": 0.0, "lr": 6.056899004267425e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.214634723297879, "train_cur_epoch_loss": 683.3264876585454, "train_cur_epoch_avg_loss": 0.3209612436160382, "train_cur_epoch_time": 44.84309720993042, "train_cur_epoch_avg_time": 0.021062986007482583, "epoch": 20, "step": 42580} ################################################## Training, Epoch: 0021, Batch: 000010, Sample Num: 160, Cur Loss: 0.16941455, Cur Avg Loss: 0.23747434, Log Avg loss: 0.23747434, Global Avg Loss: 1.21440529, Time: 0.0212 Steps: 42590, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000020, Sample Num: 320, Cur Loss: 0.08268222, Cur Avg Loss: 0.24448455, Log Avg loss: 0.25149476, Global Avg Loss: 1.21417925, Time: 0.0210 Steps: 42600, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000030, Sample Num: 480, Cur Loss: 0.22244190, Cur Avg Loss: 0.24277170, Log Avg loss: 0.23934599, Global Avg Loss: 1.21395047, Time: 0.0211 Steps: 42610, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000040, Sample Num: 640, Cur Loss: 0.10865172, Cur Avg Loss: 0.27075468, Log Avg loss: 0.35470363, Global Avg Loss: 1.21374887, Time: 0.0210 Steps: 42620, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000050, Sample Num: 800, Cur Loss: 0.39282686, Cur Avg Loss: 0.27164480, Log Avg loss: 0.27520530, Global Avg Loss: 1.21352871, Time: 0.0211 Steps: 42630, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000060, Sample Num: 960, Cur Loss: 0.16849186, Cur Avg Loss: 0.26578361, Log Avg loss: 0.23647763, Global Avg Loss: 1.21329957, Time: 0.0211 Steps: 42640, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000070, Sample Num: 1120, Cur Loss: 0.26557493, Cur Avg Loss: 0.25875336, Log Avg loss: 0.21657185, Global Avg Loss: 1.21306587, Time: 0.0211 Steps: 42650, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000080, Sample Num: 1280, Cur Loss: 0.14737847, Cur Avg Loss: 0.25751106, Log Avg loss: 0.24881500, Global Avg Loss: 1.21283984, Time: 0.0211 Steps: 42660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000090, Sample Num: 1440, Cur Loss: 0.10178818, Cur Avg Loss: 0.26602984, Log Avg loss: 0.33418009, Global Avg Loss: 1.21263392, Time: 0.0210 Steps: 42670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000100, Sample Num: 1600, Cur Loss: 0.18054566, Cur Avg Loss: 0.26780595, Log Avg loss: 0.28379091, Global Avg Loss: 1.21241629, Time: 0.0211 Steps: 42680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000110, Sample Num: 1760, Cur Loss: 0.53198564, Cur Avg Loss: 0.28324551, Log Avg loss: 0.43764117, Global Avg Loss: 1.21223480, Time: 0.0211 Steps: 42690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000120, Sample Num: 1920, Cur Loss: 0.34294832, Cur Avg Loss: 0.29796447, Log Avg loss: 0.45987300, Global Avg Loss: 1.21205860, Time: 0.0211 Steps: 42700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000130, Sample Num: 2080, Cur Loss: 0.62803006, Cur Avg Loss: 0.30172803, Log Avg loss: 0.34689073, Global Avg Loss: 1.21185603, Time: 0.0211 Steps: 42710, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000140, Sample Num: 2240, Cur Loss: 0.46922001, Cur Avg Loss: 0.31045765, Log Avg loss: 0.42394264, Global Avg Loss: 1.21167160, Time: 0.0211 Steps: 42720, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000150, Sample Num: 2400, Cur Loss: 0.16542818, Cur Avg Loss: 0.31000150, Log Avg loss: 0.30361552, Global Avg Loss: 1.21145909, Time: 0.0212 Steps: 42730, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000160, Sample Num: 2560, Cur Loss: 0.30533063, Cur Avg Loss: 0.31117582, Log Avg loss: 0.32879051, Global Avg Loss: 1.21125257, Time: 0.0210 Steps: 42740, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000170, Sample Num: 2720, Cur Loss: 0.19362533, Cur Avg Loss: 0.30682023, Log Avg loss: 0.23713093, Global Avg Loss: 1.21102470, Time: 0.0211 Steps: 42750, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000180, Sample Num: 2880, Cur Loss: 0.35199612, Cur Avg Loss: 0.30490896, Log Avg loss: 0.27241735, Global Avg Loss: 1.21080519, Time: 0.0212 Steps: 42760, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000190, Sample Num: 3040, Cur Loss: 0.39495301, Cur Avg Loss: 0.30179955, Log Avg loss: 0.24583003, Global Avg Loss: 1.21057958, Time: 0.0211 Steps: 42770, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000200, Sample Num: 3200, Cur Loss: 0.25330532, Cur Avg Loss: 0.30475446, Log Avg loss: 0.36089786, Global Avg Loss: 1.21038096, Time: 0.0211 Steps: 42780, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000210, Sample Num: 3360, Cur Loss: 0.70057774, Cur Avg Loss: 0.30396840, Log Avg loss: 0.28824713, Global Avg Loss: 1.21016546, Time: 0.0211 Steps: 42790, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000220, Sample Num: 3520, Cur Loss: 0.25305972, Cur Avg Loss: 0.30259228, Log Avg loss: 0.27369387, Global Avg Loss: 1.20994665, Time: 0.0211 Steps: 42800, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000230, Sample Num: 3680, Cur Loss: 0.22946754, Cur Avg Loss: 0.30908773, Log Avg loss: 0.45198754, Global Avg Loss: 1.20976960, Time: 0.0210 Steps: 42810, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000240, Sample Num: 3840, Cur Loss: 0.47290403, Cur Avg Loss: 0.31366353, Log Avg loss: 0.41890706, Global Avg Loss: 1.20958491, Time: 0.0210 Steps: 42820, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000250, Sample Num: 4000, Cur Loss: 0.30910721, Cur Avg Loss: 0.31282448, Log Avg loss: 0.29268710, Global Avg Loss: 1.20937083, Time: 0.0210 Steps: 42830, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000260, Sample Num: 4160, Cur Loss: 0.31915066, Cur Avg Loss: 0.30865190, Log Avg loss: 0.20433752, Global Avg Loss: 1.20913623, Time: 0.0247 Steps: 42840, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000270, Sample Num: 4320, Cur Loss: 0.21316822, Cur Avg Loss: 0.31170015, Log Avg loss: 0.39095452, Global Avg Loss: 1.20894529, Time: 0.0210 Steps: 42850, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000280, Sample Num: 4480, Cur Loss: 0.32644510, Cur Avg Loss: 0.31177042, Log Avg loss: 0.31366786, Global Avg Loss: 1.20873640, Time: 0.0210 Steps: 42860, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000290, Sample Num: 4640, Cur Loss: 0.24421631, Cur Avg Loss: 0.31114567, Log Avg loss: 0.29365259, Global Avg Loss: 1.20852295, Time: 0.0210 Steps: 42870, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000300, Sample Num: 4800, Cur Loss: 0.12890913, Cur Avg Loss: 0.31230203, Log Avg loss: 0.34583637, Global Avg Loss: 1.20832176, Time: 0.0210 Steps: 42880, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000310, Sample Num: 4960, Cur Loss: 0.45006299, Cur Avg Loss: 0.31056493, Log Avg loss: 0.25845215, Global Avg Loss: 1.20810029, Time: 0.0209 Steps: 42890, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000320, Sample Num: 5120, Cur Loss: 0.25670221, Cur Avg Loss: 0.30975416, Log Avg loss: 0.28462024, Global Avg Loss: 1.20788503, Time: 0.0210 Steps: 42900, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000330, Sample Num: 5280, Cur Loss: 0.54428601, Cur Avg Loss: 0.30952126, Log Avg loss: 0.30206826, Global Avg Loss: 1.20767393, Time: 0.0210 Steps: 42910, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000340, Sample Num: 5440, Cur Loss: 0.21003076, Cur Avg Loss: 0.30727312, Log Avg loss: 0.23308457, Global Avg Loss: 1.20744686, Time: 0.0209 Steps: 42920, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000350, Sample Num: 5600, Cur Loss: 0.56244862, Cur Avg Loss: 0.30710556, Log Avg loss: 0.30140854, Global Avg Loss: 1.20723581, Time: 0.0209 Steps: 42930, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000360, Sample Num: 5760, Cur Loss: 0.30915204, Cur Avg Loss: 0.30895581, Log Avg loss: 0.37371461, Global Avg Loss: 1.20704170, Time: 0.0209 Steps: 42940, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000370, Sample Num: 5920, Cur Loss: 0.19920732, Cur Avg Loss: 0.30628683, Log Avg loss: 0.21020343, Global Avg Loss: 1.20680961, Time: 0.0209 Steps: 42950, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000380, Sample Num: 6080, Cur Loss: 0.16788760, Cur Avg Loss: 0.30484591, Log Avg loss: 0.25153192, Global Avg Loss: 1.20658724, Time: 0.0210 Steps: 42960, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000390, Sample Num: 6240, Cur Loss: 0.35230702, Cur Avg Loss: 0.30382092, Log Avg loss: 0.26487134, Global Avg Loss: 1.20636809, Time: 0.0211 Steps: 42970, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000400, Sample Num: 6400, Cur Loss: 0.20041747, Cur Avg Loss: 0.30247755, Log Avg loss: 0.25008634, Global Avg Loss: 1.20614559, Time: 0.0209 Steps: 42980, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000410, Sample Num: 6560, Cur Loss: 0.34475371, Cur Avg Loss: 0.30443818, Log Avg loss: 0.38286299, Global Avg Loss: 1.20595409, Time: 0.0209 Steps: 42990, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000420, Sample Num: 6720, Cur Loss: 0.08927128, Cur Avg Loss: 0.30389302, Log Avg loss: 0.28154170, Global Avg Loss: 1.20573911, Time: 0.0209 Steps: 43000, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000430, Sample Num: 6880, Cur Loss: 0.29874223, Cur Avg Loss: 0.30237074, Log Avg loss: 0.23843496, Global Avg Loss: 1.20551420, Time: 0.0211 Steps: 43010, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000440, Sample Num: 7040, Cur Loss: 0.41682148, Cur Avg Loss: 0.30127449, Log Avg loss: 0.25413580, Global Avg Loss: 1.20529306, Time: 0.0210 Steps: 43020, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000450, Sample Num: 7200, Cur Loss: 0.07018048, Cur Avg Loss: 0.30117906, Log Avg loss: 0.29698028, Global Avg Loss: 1.20508197, Time: 0.0210 Steps: 43030, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000460, Sample Num: 7360, Cur Loss: 0.74982017, Cur Avg Loss: 0.30163140, Log Avg loss: 0.32198659, Global Avg Loss: 1.20487679, Time: 0.0209 Steps: 43040, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000470, Sample Num: 7520, Cur Loss: 0.26512763, Cur Avg Loss: 0.29964022, Log Avg loss: 0.20804599, Global Avg Loss: 1.20464524, Time: 0.0209 Steps: 43050, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000480, Sample Num: 7680, Cur Loss: 0.13634150, Cur Avg Loss: 0.29737205, Log Avg loss: 0.19076777, Global Avg Loss: 1.20440978, Time: 0.0209 Steps: 43060, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000490, Sample Num: 7840, Cur Loss: 0.30306616, Cur Avg Loss: 0.29561689, Log Avg loss: 0.21136946, Global Avg Loss: 1.20417922, Time: 0.0209 Steps: 43070, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000500, Sample Num: 8000, Cur Loss: 0.22635029, Cur Avg Loss: 0.29461913, Log Avg loss: 0.24572866, Global Avg Loss: 1.20395673, Time: 0.0211 Steps: 43080, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000510, Sample Num: 8160, Cur Loss: 0.26181158, Cur Avg Loss: 0.29720338, Log Avg loss: 0.42641575, Global Avg Loss: 1.20377629, Time: 0.0210 Steps: 43090, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000520, Sample Num: 8320, Cur Loss: 0.09984665, Cur Avg Loss: 0.29738669, Log Avg loss: 0.30673563, Global Avg Loss: 1.20356816, Time: 0.0210 Steps: 43100, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000530, Sample Num: 8480, Cur Loss: 0.28044933, Cur Avg Loss: 0.29580628, Log Avg loss: 0.21362528, Global Avg Loss: 1.20333853, Time: 0.0209 Steps: 43110, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000540, Sample Num: 8640, Cur Loss: 0.23097067, Cur Avg Loss: 0.29427613, Log Avg loss: 0.21317774, Global Avg Loss: 1.20310890, Time: 0.0209 Steps: 43120, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000550, Sample Num: 8800, Cur Loss: 0.12469092, Cur Avg Loss: 0.29346177, Log Avg loss: 0.24948676, Global Avg Loss: 1.20288779, Time: 0.0210 Steps: 43130, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000560, Sample Num: 8960, Cur Loss: 0.21719782, Cur Avg Loss: 0.29330987, Log Avg loss: 0.28495502, Global Avg Loss: 1.20267501, Time: 0.0210 Steps: 43140, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000570, Sample Num: 9120, Cur Loss: 0.51265025, Cur Avg Loss: 0.29737634, Log Avg loss: 0.52509852, Global Avg Loss: 1.20251798, Time: 0.0209 Steps: 43150, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000580, Sample Num: 9280, Cur Loss: 0.09414801, Cur Avg Loss: 0.30048319, Log Avg loss: 0.47757378, Global Avg Loss: 1.20235002, Time: 0.0210 Steps: 43160, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000590, Sample Num: 9440, Cur Loss: 0.48557603, Cur Avg Loss: 0.30120261, Log Avg loss: 0.34292904, Global Avg Loss: 1.20215094, Time: 0.0210 Steps: 43170, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000600, Sample Num: 9600, Cur Loss: 0.08505353, Cur Avg Loss: 0.30061229, Log Avg loss: 0.26578346, Global Avg Loss: 1.20193409, Time: 0.0210 Steps: 43180, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000610, Sample Num: 9760, Cur Loss: 0.09094200, Cur Avg Loss: 0.29964738, Log Avg loss: 0.24175296, Global Avg Loss: 1.20171177, Time: 0.0210 Steps: 43190, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000620, Sample Num: 9920, Cur Loss: 0.30044433, Cur Avg Loss: 0.30028050, Log Avg loss: 0.33890080, Global Avg Loss: 1.20151205, Time: 0.0210 Steps: 43200, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000630, Sample Num: 10080, Cur Loss: 0.39452201, Cur Avg Loss: 0.30017147, Log Avg loss: 0.29341131, Global Avg Loss: 1.20130189, Time: 0.0210 Steps: 43210, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000640, Sample Num: 10240, Cur Loss: 0.18609180, Cur Avg Loss: 0.29971021, Log Avg loss: 0.27065080, Global Avg Loss: 1.20108656, Time: 0.0210 Steps: 43220, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000650, Sample Num: 10400, Cur Loss: 1.32727253, Cur Avg Loss: 0.30345342, Log Avg loss: 0.54301930, Global Avg Loss: 1.20093433, Time: 0.0210 Steps: 43230, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000660, Sample Num: 10560, Cur Loss: 0.56491709, Cur Avg Loss: 0.30525115, Log Avg loss: 0.42210361, Global Avg Loss: 1.20075422, Time: 0.0210 Steps: 43240, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000670, Sample Num: 10720, Cur Loss: 0.26439324, Cur Avg Loss: 0.30548823, Log Avg loss: 0.32113491, Global Avg Loss: 1.20055084, Time: 0.0210 Steps: 43250, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000680, Sample Num: 10880, Cur Loss: 0.16724604, Cur Avg Loss: 0.30594020, Log Avg loss: 0.33622251, Global Avg Loss: 1.20035104, Time: 0.0210 Steps: 43260, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000690, Sample Num: 11040, Cur Loss: 0.21424736, Cur Avg Loss: 0.30577757, Log Avg loss: 0.29471865, Global Avg Loss: 1.20014174, Time: 0.0209 Steps: 43270, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000700, Sample Num: 11200, Cur Loss: 0.16323680, Cur Avg Loss: 0.30474860, Log Avg loss: 0.23374970, Global Avg Loss: 1.19991845, Time: 0.0210 Steps: 43280, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000710, Sample Num: 11360, Cur Loss: 0.21036991, Cur Avg Loss: 0.30324216, Log Avg loss: 0.19779130, Global Avg Loss: 1.19968696, Time: 0.0209 Steps: 43290, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000720, Sample Num: 11520, Cur Loss: 0.35954094, Cur Avg Loss: 0.30375121, Log Avg loss: 0.33989354, Global Avg Loss: 1.19948839, Time: 0.0209 Steps: 43300, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000730, Sample Num: 11680, Cur Loss: 0.28825495, Cur Avg Loss: 0.30347206, Log Avg loss: 0.28337390, Global Avg Loss: 1.19927687, Time: 0.0209 Steps: 43310, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000740, Sample Num: 11840, Cur Loss: 0.09621373, Cur Avg Loss: 0.30437652, Log Avg loss: 0.37040194, Global Avg Loss: 1.19908553, Time: 0.0211 Steps: 43320, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000750, Sample Num: 12000, Cur Loss: 0.29942325, Cur Avg Loss: 0.30425428, Log Avg loss: 0.29520820, Global Avg Loss: 1.19887693, Time: 0.0211 Steps: 43330, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000760, Sample Num: 12160, Cur Loss: 0.39114690, Cur Avg Loss: 0.30439721, Log Avg loss: 0.31511718, Global Avg Loss: 1.19867301, Time: 0.0212 Steps: 43340, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000770, Sample Num: 12320, Cur Loss: 0.48217764, Cur Avg Loss: 0.30400890, Log Avg loss: 0.27449757, Global Avg Loss: 1.19845982, Time: 0.0248 Steps: 43350, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000780, Sample Num: 12480, Cur Loss: 0.20516740, Cur Avg Loss: 0.30475645, Log Avg loss: 0.36231789, Global Avg Loss: 1.19826699, Time: 0.0211 Steps: 43360, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000790, Sample Num: 12640, Cur Loss: 0.49216008, Cur Avg Loss: 0.30654886, Log Avg loss: 0.44635682, Global Avg Loss: 1.19809362, Time: 0.0211 Steps: 43370, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000800, Sample Num: 12800, Cur Loss: 0.41993207, Cur Avg Loss: 0.30738987, Log Avg loss: 0.37382894, Global Avg Loss: 1.19790361, Time: 0.0210 Steps: 43380, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000810, Sample Num: 12960, Cur Loss: 0.59575474, Cur Avg Loss: 0.30744054, Log Avg loss: 0.31149424, Global Avg Loss: 1.19769932, Time: 0.0211 Steps: 43390, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000820, Sample Num: 13120, Cur Loss: 0.11921700, Cur Avg Loss: 0.30661233, Log Avg loss: 0.23952799, Global Avg Loss: 1.19747854, Time: 0.0211 Steps: 43400, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000830, Sample Num: 13280, Cur Loss: 0.55963260, Cur Avg Loss: 0.30615845, Log Avg loss: 0.26893960, Global Avg Loss: 1.19726464, Time: 0.0211 Steps: 43410, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000840, Sample Num: 13440, Cur Loss: 0.24819307, Cur Avg Loss: 0.30539566, Log Avg loss: 0.24208464, Global Avg Loss: 1.19704465, Time: 0.0211 Steps: 43420, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000850, Sample Num: 13600, Cur Loss: 0.39508656, Cur Avg Loss: 0.30542738, Log Avg loss: 0.30809178, Global Avg Loss: 1.19683997, Time: 0.0211 Steps: 43430, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000860, Sample Num: 13760, Cur Loss: 0.26147991, Cur Avg Loss: 0.30577858, Log Avg loss: 0.33563017, Global Avg Loss: 1.19664171, Time: 0.0210 Steps: 43440, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000870, Sample Num: 13920, Cur Loss: 0.35030061, Cur Avg Loss: 0.30563406, Log Avg loss: 0.29320599, Global Avg Loss: 1.19643379, Time: 0.0211 Steps: 43450, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000880, Sample Num: 14080, Cur Loss: 0.22790316, Cur Avg Loss: 0.30503298, Log Avg loss: 0.25273821, Global Avg Loss: 1.19621665, Time: 0.0211 Steps: 43460, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000890, Sample Num: 14240, Cur Loss: 0.40999505, Cur Avg Loss: 0.30484387, Log Avg loss: 0.28820236, Global Avg Loss: 1.19600777, Time: 0.0211 Steps: 43470, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000900, Sample Num: 14400, Cur Loss: 0.22335875, Cur Avg Loss: 0.30399116, Log Avg loss: 0.22809986, Global Avg Loss: 1.19578516, Time: 0.0211 Steps: 43480, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000910, Sample Num: 14560, Cur Loss: 0.28440440, Cur Avg Loss: 0.30446772, Log Avg loss: 0.34735805, Global Avg Loss: 1.19559007, Time: 0.0210 Steps: 43490, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000920, Sample Num: 14720, Cur Loss: 0.41928369, Cur Avg Loss: 0.30453029, Log Avg loss: 0.31022473, Global Avg Loss: 1.19538654, Time: 0.0210 Steps: 43500, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000930, Sample Num: 14880, Cur Loss: 0.15926884, Cur Avg Loss: 0.30426184, Log Avg loss: 0.27956419, Global Avg Loss: 1.19517605, Time: 0.0210 Steps: 43510, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000940, Sample Num: 15040, Cur Loss: 0.24572654, Cur Avg Loss: 0.30377198, Log Avg loss: 0.25821517, Global Avg Loss: 1.19496076, Time: 0.0211 Steps: 43520, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000950, Sample Num: 15200, Cur Loss: 0.16070065, Cur Avg Loss: 0.30307117, Log Avg loss: 0.23719467, Global Avg Loss: 1.19474073, Time: 0.0211 Steps: 43530, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000960, Sample Num: 15360, Cur Loss: 0.19202538, Cur Avg Loss: 0.30161380, Log Avg loss: 0.16316409, Global Avg Loss: 1.19450381, Time: 0.0210 Steps: 43540, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000970, Sample Num: 15520, Cur Loss: 0.43029061, Cur Avg Loss: 0.30153827, Log Avg loss: 0.29428675, Global Avg Loss: 1.19429710, Time: 0.0210 Steps: 43550, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000980, Sample Num: 15680, Cur Loss: 0.50140631, Cur Avg Loss: 0.30261830, Log Avg loss: 0.40738151, Global Avg Loss: 1.19411645, Time: 0.0211 Steps: 43560, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000990, Sample Num: 15840, Cur Loss: 0.22080882, Cur Avg Loss: 0.30244589, Log Avg loss: 0.28554957, Global Avg Loss: 1.19390792, Time: 0.0211 Steps: 43570, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001000, Sample Num: 16000, Cur Loss: 0.11017402, Cur Avg Loss: 0.30127683, Log Avg loss: 0.18553971, Global Avg Loss: 1.19367653, Time: 0.0212 Steps: 43580, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001010, Sample Num: 16160, Cur Loss: 0.50812417, Cur Avg Loss: 0.30262358, Log Avg loss: 0.43729921, Global Avg Loss: 1.19350301, Time: 0.0211 Steps: 43590, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001020, Sample Num: 16320, Cur Loss: 0.19136000, Cur Avg Loss: 0.30292933, Log Avg loss: 0.33380957, Global Avg Loss: 1.19330584, Time: 0.0209 Steps: 43600, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001030, Sample Num: 16480, Cur Loss: 0.23258919, Cur Avg Loss: 0.30334615, Log Avg loss: 0.34586234, Global Avg Loss: 1.19311151, Time: 0.0213 Steps: 43610, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001040, Sample Num: 16640, Cur Loss: 0.22030123, Cur Avg Loss: 0.30426661, Log Avg loss: 0.39907359, Global Avg Loss: 1.19292948, Time: 0.0210 Steps: 43620, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001050, Sample Num: 16800, Cur Loss: 0.38407651, Cur Avg Loss: 0.30413937, Log Avg loss: 0.29090697, Global Avg Loss: 1.19272273, Time: 0.0210 Steps: 43630, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001060, Sample Num: 16960, Cur Loss: 0.34433505, Cur Avg Loss: 0.30356725, Log Avg loss: 0.24349393, Global Avg Loss: 1.19250522, Time: 0.0210 Steps: 43640, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001070, Sample Num: 17120, Cur Loss: 0.30315745, Cur Avg Loss: 0.30331896, Log Avg loss: 0.27700103, Global Avg Loss: 1.19229548, Time: 0.0210 Steps: 43650, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001080, Sample Num: 17280, Cur Loss: 0.10636690, Cur Avg Loss: 0.30314196, Log Avg loss: 0.28420296, Global Avg Loss: 1.19208749, Time: 0.0210 Steps: 43660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001090, Sample Num: 17440, Cur Loss: 0.07894182, Cur Avg Loss: 0.30246377, Log Avg loss: 0.22921891, Global Avg Loss: 1.19186700, Time: 0.0210 Steps: 43670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001100, Sample Num: 17600, Cur Loss: 0.08764748, Cur Avg Loss: 0.30186271, Log Avg loss: 0.23634691, Global Avg Loss: 1.19164825, Time: 0.0210 Steps: 43680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001110, Sample Num: 17760, Cur Loss: 0.34257531, Cur Avg Loss: 0.30112065, Log Avg loss: 0.21949427, Global Avg Loss: 1.19142574, Time: 0.0210 Steps: 43690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001120, Sample Num: 17920, Cur Loss: 0.08130948, Cur Avg Loss: 0.30217270, Log Avg loss: 0.41895044, Global Avg Loss: 1.19124897, Time: 0.0210 Steps: 43700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001130, Sample Num: 18080, Cur Loss: 0.20766000, Cur Avg Loss: 0.30143537, Log Avg loss: 0.21885461, Global Avg Loss: 1.19102650, Time: 0.0211 Steps: 43710, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001140, Sample Num: 18240, Cur Loss: 0.91365302, Cur Avg Loss: 0.30291143, Log Avg loss: 0.46970600, Global Avg Loss: 1.19086152, Time: 0.0210 Steps: 43720, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001150, Sample Num: 18400, Cur Loss: 0.39083400, Cur Avg Loss: 0.30447175, Log Avg loss: 0.48234826, Global Avg Loss: 1.19069950, Time: 0.0210 Steps: 43730, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001160, Sample Num: 18560, Cur Loss: 0.38003355, Cur Avg Loss: 0.30442132, Log Avg loss: 0.29862131, Global Avg Loss: 1.19049555, Time: 0.0210 Steps: 43740, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001170, Sample Num: 18720, Cur Loss: 0.27654901, Cur Avg Loss: 0.30428964, Log Avg loss: 0.28901562, Global Avg Loss: 1.19028949, Time: 0.0209 Steps: 43750, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001180, Sample Num: 18880, Cur Loss: 0.22963285, Cur Avg Loss: 0.30417423, Log Avg loss: 0.29067107, Global Avg Loss: 1.19008391, Time: 0.0209 Steps: 43760, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001190, Sample Num: 19040, Cur Loss: 0.24919304, Cur Avg Loss: 0.30387698, Log Avg loss: 0.26880089, Global Avg Loss: 1.18987343, Time: 0.0210 Steps: 43770, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001200, Sample Num: 19200, Cur Loss: 0.18121120, Cur Avg Loss: 0.30447541, Log Avg loss: 0.37568831, Global Avg Loss: 1.18968746, Time: 0.0209 Steps: 43780, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001210, Sample Num: 19360, Cur Loss: 1.06795847, Cur Avg Loss: 0.30754601, Log Avg loss: 0.67601838, Global Avg Loss: 1.18957016, Time: 0.0210 Steps: 43790, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001220, Sample Num: 19520, Cur Loss: 0.39716363, Cur Avg Loss: 0.30786981, Log Avg loss: 0.34704998, Global Avg Loss: 1.18937780, Time: 0.0210 Steps: 43800, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001230, Sample Num: 19680, Cur Loss: 0.27521813, Cur Avg Loss: 0.30815797, Log Avg loss: 0.34331314, Global Avg Loss: 1.18918468, Time: 0.0210 Steps: 43810, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001240, Sample Num: 19840, Cur Loss: 1.00120807, Cur Avg Loss: 0.30988015, Log Avg loss: 0.52170879, Global Avg Loss: 1.18903236, Time: 0.0209 Steps: 43820, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001250, Sample Num: 20000, Cur Loss: 0.28127936, Cur Avg Loss: 0.31047841, Log Avg loss: 0.38466195, Global Avg Loss: 1.18884884, Time: 0.0210 Steps: 43830, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001260, Sample Num: 20160, Cur Loss: 0.51465833, Cur Avg Loss: 0.31076574, Log Avg loss: 0.34668284, Global Avg Loss: 1.18865674, Time: 0.0210 Steps: 43840, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001270, Sample Num: 20320, Cur Loss: 0.14132464, Cur Avg Loss: 0.31106727, Log Avg loss: 0.34905983, Global Avg Loss: 1.18846527, Time: 0.0210 Steps: 43850, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001280, Sample Num: 20480, Cur Loss: 0.62494951, Cur Avg Loss: 0.31138909, Log Avg loss: 0.35225927, Global Avg Loss: 1.18827461, Time: 0.0255 Steps: 43860, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001290, Sample Num: 20640, Cur Loss: 0.23926163, Cur Avg Loss: 0.31167132, Log Avg loss: 0.34779709, Global Avg Loss: 1.18808303, Time: 0.0210 Steps: 43870, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001300, Sample Num: 20800, Cur Loss: 0.61383688, Cur Avg Loss: 0.31221004, Log Avg loss: 0.38170563, Global Avg Loss: 1.18789926, Time: 0.0211 Steps: 43880, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001310, Sample Num: 20960, Cur Loss: 0.23948959, Cur Avg Loss: 0.31242436, Log Avg loss: 0.34028560, Global Avg Loss: 1.18770614, Time: 0.0210 Steps: 43890, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001320, Sample Num: 21120, Cur Loss: 0.14677000, Cur Avg Loss: 0.31204240, Log Avg loss: 0.26200583, Global Avg Loss: 1.18749527, Time: 0.0210 Steps: 43900, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001330, Sample Num: 21280, Cur Loss: 0.32708567, Cur Avg Loss: 0.31238173, Log Avg loss: 0.35717331, Global Avg Loss: 1.18730618, Time: 0.0210 Steps: 43910, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001340, Sample Num: 21440, Cur Loss: 0.15961187, Cur Avg Loss: 0.31247762, Log Avg loss: 0.32523073, Global Avg Loss: 1.18710989, Time: 0.0210 Steps: 43920, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001350, Sample Num: 21600, Cur Loss: 0.17048424, Cur Avg Loss: 0.31227246, Log Avg loss: 0.28478086, Global Avg Loss: 1.18690449, Time: 0.0210 Steps: 43930, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001360, Sample Num: 21760, Cur Loss: 0.57760787, Cur Avg Loss: 0.31242074, Log Avg loss: 0.33243830, Global Avg Loss: 1.18671003, Time: 0.0210 Steps: 43940, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001370, Sample Num: 21920, Cur Loss: 0.09278860, Cur Avg Loss: 0.31274178, Log Avg loss: 0.35640311, Global Avg Loss: 1.18652111, Time: 0.0211 Steps: 43950, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001380, Sample Num: 22080, Cur Loss: 0.24923927, Cur Avg Loss: 0.31296412, Log Avg loss: 0.34342524, Global Avg Loss: 1.18632932, Time: 0.0211 Steps: 43960, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001390, Sample Num: 22240, Cur Loss: 0.22294976, Cur Avg Loss: 0.31363782, Log Avg loss: 0.40660824, Global Avg Loss: 1.18615199, Time: 0.0210 Steps: 43970, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001400, Sample Num: 22400, Cur Loss: 0.10853039, Cur Avg Loss: 0.31433747, Log Avg loss: 0.41158909, Global Avg Loss: 1.18597587, Time: 0.0210 Steps: 43980, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001410, Sample Num: 22560, Cur Loss: 0.25926337, Cur Avg Loss: 0.31464273, Log Avg loss: 0.35737901, Global Avg Loss: 1.18578751, Time: 0.0210 Steps: 43990, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001420, Sample Num: 22720, Cur Loss: 0.64535475, Cur Avg Loss: 0.31528460, Log Avg loss: 0.40578868, Global Avg Loss: 1.18561024, Time: 0.0211 Steps: 44000, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001430, Sample Num: 22880, Cur Loss: 0.11191659, Cur Avg Loss: 0.31492230, Log Avg loss: 0.26347560, Global Avg Loss: 1.18540071, Time: 0.0210 Steps: 44010, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001440, Sample Num: 23040, Cur Loss: 0.41774458, Cur Avg Loss: 0.31492678, Log Avg loss: 0.31556750, Global Avg Loss: 1.18520311, Time: 0.0209 Steps: 44020, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001450, Sample Num: 23200, Cur Loss: 0.22466290, Cur Avg Loss: 0.31480701, Log Avg loss: 0.29756005, Global Avg Loss: 1.18500151, Time: 0.0211 Steps: 44030, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001460, Sample Num: 23360, Cur Loss: 0.22426310, Cur Avg Loss: 0.31514766, Log Avg loss: 0.36454217, Global Avg Loss: 1.18481522, Time: 0.0210 Steps: 44040, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001470, Sample Num: 23520, Cur Loss: 0.20796400, Cur Avg Loss: 0.31479451, Log Avg loss: 0.26323348, Global Avg Loss: 1.18460600, Time: 0.0211 Steps: 44050, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001480, Sample Num: 23680, Cur Loss: 0.21939144, Cur Avg Loss: 0.31486193, Log Avg loss: 0.32477317, Global Avg Loss: 1.18441085, Time: 0.0212 Steps: 44060, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001490, Sample Num: 23840, Cur Loss: 0.54977715, Cur Avg Loss: 0.31439265, Log Avg loss: 0.24493924, Global Avg Loss: 1.18419768, Time: 0.0211 Steps: 44070, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001500, Sample Num: 24000, Cur Loss: 0.16053671, Cur Avg Loss: 0.31345173, Log Avg loss: 0.17325408, Global Avg Loss: 1.18396833, Time: 0.0212 Steps: 44080, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001510, Sample Num: 24160, Cur Loss: 0.08595501, Cur Avg Loss: 0.31326876, Log Avg loss: 0.28582311, Global Avg Loss: 1.18376463, Time: 0.0211 Steps: 44090, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001520, Sample Num: 24320, Cur Loss: 0.44229597, Cur Avg Loss: 0.31290812, Log Avg loss: 0.25845271, Global Avg Loss: 1.18355480, Time: 0.0211 Steps: 44100, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001530, Sample Num: 24480, Cur Loss: 0.11542404, Cur Avg Loss: 0.31310553, Log Avg loss: 0.34311205, Global Avg Loss: 1.18336427, Time: 0.0210 Steps: 44110, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001540, Sample Num: 24640, Cur Loss: 0.69073808, Cur Avg Loss: 0.31380858, Log Avg loss: 0.42137431, Global Avg Loss: 1.18319156, Time: 0.0248 Steps: 44120, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001550, Sample Num: 24800, Cur Loss: 0.40166086, Cur Avg Loss: 0.31396407, Log Avg loss: 0.33790954, Global Avg Loss: 1.18300002, Time: 0.0210 Steps: 44130, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001560, Sample Num: 24960, Cur Loss: 0.28642890, Cur Avg Loss: 0.31405112, Log Avg loss: 0.32754449, Global Avg Loss: 1.18280621, Time: 0.0211 Steps: 44140, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001570, Sample Num: 25120, Cur Loss: 0.38972282, Cur Avg Loss: 0.31410782, Log Avg loss: 0.32295186, Global Avg Loss: 1.18261146, Time: 0.0211 Steps: 44150, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001580, Sample Num: 25280, Cur Loss: 0.25404769, Cur Avg Loss: 0.31370065, Log Avg loss: 0.24977630, Global Avg Loss: 1.18240022, Time: 0.0211 Steps: 44160, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001590, Sample Num: 25440, Cur Loss: 0.28214154, Cur Avg Loss: 0.31310129, Log Avg loss: 0.21840245, Global Avg Loss: 1.18218197, Time: 0.0211 Steps: 44170, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001600, Sample Num: 25600, Cur Loss: 0.40580371, Cur Avg Loss: 0.31313374, Log Avg loss: 0.31829336, Global Avg Loss: 1.18198643, Time: 0.0211 Steps: 44180, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001610, Sample Num: 25760, Cur Loss: 0.40017956, Cur Avg Loss: 0.31272397, Log Avg loss: 0.24716083, Global Avg Loss: 1.18177488, Time: 0.0211 Steps: 44190, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001620, Sample Num: 25920, Cur Loss: 0.25060105, Cur Avg Loss: 0.31216201, Log Avg loss: 0.22168638, Global Avg Loss: 1.18155767, Time: 0.0210 Steps: 44200, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001630, Sample Num: 26080, Cur Loss: 0.36931401, Cur Avg Loss: 0.31218196, Log Avg loss: 0.31541284, Global Avg Loss: 1.18136175, Time: 0.0211 Steps: 44210, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001640, Sample Num: 26240, Cur Loss: 0.39941797, Cur Avg Loss: 0.31210134, Log Avg loss: 0.29896116, Global Avg Loss: 1.18116221, Time: 0.0211 Steps: 44220, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001650, Sample Num: 26400, Cur Loss: 0.35618019, Cur Avg Loss: 0.31227382, Log Avg loss: 0.34056030, Global Avg Loss: 1.18097215, Time: 0.0210 Steps: 44230, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001660, Sample Num: 26560, Cur Loss: 0.22011814, Cur Avg Loss: 0.31199765, Log Avg loss: 0.26642905, Global Avg Loss: 1.18076543, Time: 0.0211 Steps: 44240, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001670, Sample Num: 26720, Cur Loss: 0.77810729, Cur Avg Loss: 0.31173035, Log Avg loss: 0.26735838, Global Avg Loss: 1.18055901, Time: 0.0211 Steps: 44250, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001680, Sample Num: 26880, Cur Loss: 0.25854653, Cur Avg Loss: 0.31200832, Log Avg loss: 0.35843035, Global Avg Loss: 1.18037326, Time: 0.0211 Steps: 44260, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001690, Sample Num: 27040, Cur Loss: 0.04552710, Cur Avg Loss: 0.31173046, Log Avg loss: 0.26504969, Global Avg Loss: 1.18016650, Time: 0.0211 Steps: 44270, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001700, Sample Num: 27200, Cur Loss: 0.35192567, Cur Avg Loss: 0.31168252, Log Avg loss: 0.30358077, Global Avg Loss: 1.17996854, Time: 0.0210 Steps: 44280, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001710, Sample Num: 27360, Cur Loss: 0.21512939, Cur Avg Loss: 0.31144417, Log Avg loss: 0.27092502, Global Avg Loss: 1.17976329, Time: 0.0210 Steps: 44290, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001720, Sample Num: 27520, Cur Loss: 0.10553432, Cur Avg Loss: 0.31096138, Log Avg loss: 0.22840370, Global Avg Loss: 1.17954853, Time: 0.0211 Steps: 44300, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001730, Sample Num: 27680, Cur Loss: 0.11717405, Cur Avg Loss: 0.31104823, Log Avg loss: 0.32598699, Global Avg Loss: 1.17935590, Time: 0.0211 Steps: 44310, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001740, Sample Num: 27840, Cur Loss: 0.52676487, Cur Avg Loss: 0.31088234, Log Avg loss: 0.28218262, Global Avg Loss: 1.17915347, Time: 0.0211 Steps: 44320, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001750, Sample Num: 28000, Cur Loss: 0.35341322, Cur Avg Loss: 0.31046267, Log Avg loss: 0.23743980, Global Avg Loss: 1.17894104, Time: 0.0211 Steps: 44330, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001760, Sample Num: 28160, Cur Loss: 0.68895715, Cur Avg Loss: 0.31018135, Log Avg loss: 0.26095112, Global Avg Loss: 1.17873400, Time: 0.0211 Steps: 44340, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001770, Sample Num: 28320, Cur Loss: 0.88692838, Cur Avg Loss: 0.31032756, Log Avg loss: 0.33605982, Global Avg Loss: 1.17854400, Time: 0.0211 Steps: 44350, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001780, Sample Num: 28480, Cur Loss: 0.24306883, Cur Avg Loss: 0.31040161, Log Avg loss: 0.32350816, Global Avg Loss: 1.17835125, Time: 0.0211 Steps: 44360, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001790, Sample Num: 28640, Cur Loss: 0.39404750, Cur Avg Loss: 0.31091238, Log Avg loss: 0.40183029, Global Avg Loss: 1.17817624, Time: 0.0211 Steps: 44370, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001800, Sample Num: 28800, Cur Loss: 0.25385499, Cur Avg Loss: 0.31088426, Log Avg loss: 0.30585042, Global Avg Loss: 1.17797968, Time: 0.0212 Steps: 44380, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001810, Sample Num: 28960, Cur Loss: 0.08687271, Cur Avg Loss: 0.31102027, Log Avg loss: 0.33550130, Global Avg Loss: 1.17778989, Time: 0.0211 Steps: 44390, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001820, Sample Num: 29120, Cur Loss: 0.17601530, Cur Avg Loss: 0.31121629, Log Avg loss: 0.34669744, Global Avg Loss: 1.17760271, Time: 0.0210 Steps: 44400, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001830, Sample Num: 29280, Cur Loss: 0.16314916, Cur Avg Loss: 0.31102050, Log Avg loss: 0.27538657, Global Avg Loss: 1.17739955, Time: 0.0211 Steps: 44410, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001840, Sample Num: 29440, Cur Loss: 0.27855417, Cur Avg Loss: 0.31151451, Log Avg loss: 0.40191756, Global Avg Loss: 1.17722497, Time: 0.0211 Steps: 44420, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001850, Sample Num: 29600, Cur Loss: 0.18095309, Cur Avg Loss: 0.31155666, Log Avg loss: 0.31931169, Global Avg Loss: 1.17703188, Time: 0.0211 Steps: 44430, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001860, Sample Num: 29760, Cur Loss: 0.10464767, Cur Avg Loss: 0.31123149, Log Avg loss: 0.25107538, Global Avg Loss: 1.17682352, Time: 0.0211 Steps: 44440, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001870, Sample Num: 29920, Cur Loss: 0.24761024, Cur Avg Loss: 0.31087954, Log Avg loss: 0.24541722, Global Avg Loss: 1.17661398, Time: 0.0211 Steps: 44450, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001880, Sample Num: 30080, Cur Loss: 0.42987823, Cur Avg Loss: 0.31090537, Log Avg loss: 0.31573480, Global Avg Loss: 1.17642035, Time: 0.0210 Steps: 44460, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001890, Sample Num: 30240, Cur Loss: 0.06432235, Cur Avg Loss: 0.31030491, Log Avg loss: 0.19741959, Global Avg Loss: 1.17620020, Time: 0.0211 Steps: 44470, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001900, Sample Num: 30400, Cur Loss: 0.27279699, Cur Avg Loss: 0.31033938, Log Avg loss: 0.31685474, Global Avg Loss: 1.17600700, Time: 0.0210 Steps: 44480, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001910, Sample Num: 30560, Cur Loss: 0.72217214, Cur Avg Loss: 0.31066693, Log Avg loss: 0.37290118, Global Avg Loss: 1.17582649, Time: 0.0211 Steps: 44490, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001920, Sample Num: 30720, Cur Loss: 0.28524646, Cur Avg Loss: 0.31046238, Log Avg loss: 0.27139300, Global Avg Loss: 1.17562324, Time: 0.0211 Steps: 44500, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001930, Sample Num: 30880, Cur Loss: 0.45735329, Cur Avg Loss: 0.31021916, Log Avg loss: 0.26352092, Global Avg Loss: 1.17541832, Time: 0.0210 Steps: 44510, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001940, Sample Num: 31040, Cur Loss: 0.11108199, Cur Avg Loss: 0.31040337, Log Avg loss: 0.34595470, Global Avg Loss: 1.17523201, Time: 0.0210 Steps: 44520, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001950, Sample Num: 31200, Cur Loss: 0.16289257, Cur Avg Loss: 0.30981402, Log Avg loss: 0.19548092, Global Avg Loss: 1.17501199, Time: 0.0211 Steps: 44530, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001960, Sample Num: 31360, Cur Loss: 0.40849394, Cur Avg Loss: 0.30967030, Log Avg loss: 0.28164561, Global Avg Loss: 1.17481141, Time: 0.0211 Steps: 44540, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001970, Sample Num: 31520, Cur Loss: 1.09795403, Cur Avg Loss: 0.31051490, Log Avg loss: 0.47605686, Global Avg Loss: 1.17465457, Time: 0.0211 Steps: 44550, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001980, Sample Num: 31680, Cur Loss: 0.53277552, Cur Avg Loss: 0.31112655, Log Avg loss: 0.43162025, Global Avg Loss: 1.17448782, Time: 0.0211 Steps: 44560, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001990, Sample Num: 31840, Cur Loss: 0.19453289, Cur Avg Loss: 0.31132556, Log Avg loss: 0.35073043, Global Avg Loss: 1.17430299, Time: 0.0211 Steps: 44570, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002000, Sample Num: 32000, Cur Loss: 0.62198627, Cur Avg Loss: 0.31184275, Log Avg loss: 0.41476252, Global Avg Loss: 1.17413262, Time: 0.0210 Steps: 44580, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002010, Sample Num: 32160, Cur Loss: 0.31858188, Cur Avg Loss: 0.31211337, Log Avg loss: 0.36623835, Global Avg Loss: 1.17395143, Time: 0.0211 Steps: 44590, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002020, Sample Num: 32320, Cur Loss: 0.12971021, Cur Avg Loss: 0.31171074, Log Avg loss: 0.23078208, Global Avg Loss: 1.17373996, Time: 0.0210 Steps: 44600, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002030, Sample Num: 32480, Cur Loss: 0.39389455, Cur Avg Loss: 0.31156738, Log Avg loss: 0.28260783, Global Avg Loss: 1.17354020, Time: 0.0210 Steps: 44610, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002040, Sample Num: 32640, Cur Loss: 0.29932743, Cur Avg Loss: 0.31156008, Log Avg loss: 0.31007941, Global Avg Loss: 1.17334669, Time: 0.0210 Steps: 44620, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002050, Sample Num: 32800, Cur Loss: 0.21067865, Cur Avg Loss: 0.31141966, Log Avg loss: 0.28277327, Global Avg Loss: 1.17314714, Time: 0.0248 Steps: 44630, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002060, Sample Num: 32960, Cur Loss: 0.59812778, Cur Avg Loss: 0.31143642, Log Avg loss: 0.31487209, Global Avg Loss: 1.17295487, Time: 0.0210 Steps: 44640, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002070, Sample Num: 33120, Cur Loss: 0.65048897, Cur Avg Loss: 0.31191094, Log Avg loss: 0.40966284, Global Avg Loss: 1.17278392, Time: 0.0210 Steps: 44650, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002080, Sample Num: 33280, Cur Loss: 0.50311959, Cur Avg Loss: 0.31175312, Log Avg loss: 0.27908436, Global Avg Loss: 1.17258381, Time: 0.0210 Steps: 44660, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002090, Sample Num: 33440, Cur Loss: 0.30061409, Cur Avg Loss: 0.31113439, Log Avg loss: 0.18243740, Global Avg Loss: 1.17236215, Time: 0.0210 Steps: 44670, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002100, Sample Num: 33600, Cur Loss: 0.30844587, Cur Avg Loss: 0.31120678, Log Avg loss: 0.32633703, Global Avg Loss: 1.17217280, Time: 0.0210 Steps: 44680, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002110, Sample Num: 33760, Cur Loss: 1.21335387, Cur Avg Loss: 0.31257517, Log Avg loss: 0.59993697, Global Avg Loss: 1.17204476, Time: 0.0210 Steps: 44690, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002120, Sample Num: 33920, Cur Loss: 0.22292259, Cur Avg Loss: 0.31259552, Log Avg loss: 0.31688972, Global Avg Loss: 1.17185345, Time: 0.0210 Steps: 44700, Updated lr: 0.000059 ***** Running evaluation checkpoint-44709 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-44709 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.989951, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.312031, "eval_total_loss": 219.358125, "eval_mae": 0.427299, "eval_mse": 0.31209, "eval_r2": 0.801615, "eval_sp_statistic": 0.883462, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.906294, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.406197, "test_total_loss": 203.910807, "test_mae": 0.393558, "test_mse": 0.406346, "test_r2": 0.737741, "test_sp_statistic": 0.86505, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.908819, "test_ps_pvalue": 0.0, "lr": 5.855002370791844e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1716730151620929, "train_cur_epoch_loss": 665.1823168583214, "train_cur_epoch_avg_loss": 0.3124388524463699, "train_cur_epoch_time": 44.98995113372803, "train_cur_epoch_avg_time": 0.021131963895597947, "epoch": 21, "step": 44709} ################################################## Training, Epoch: 0022, Batch: 000001, Sample Num: 16, Cur Loss: 0.23394853, Cur Avg Loss: 0.23394853, Log Avg loss: 0.27137573, Global Avg Loss: 1.17165204, Time: 0.0250 Steps: 44710, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000011, Sample Num: 176, Cur Loss: 0.11010715, Cur Avg Loss: 0.23167785, Log Avg loss: 0.23145078, Global Avg Loss: 1.17144180, Time: 0.0212 Steps: 44720, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000021, Sample Num: 336, Cur Loss: 0.20796293, Cur Avg Loss: 0.31717065, Log Avg loss: 0.41121274, Global Avg Loss: 1.17127184, Time: 0.0212 Steps: 44730, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000031, Sample Num: 496, Cur Loss: 0.13495532, Cur Avg Loss: 0.31040675, Log Avg loss: 0.29620256, Global Avg Loss: 1.17107625, Time: 0.0211 Steps: 44740, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000041, Sample Num: 656, Cur Loss: 0.14370750, Cur Avg Loss: 0.28118042, Log Avg loss: 0.19057877, Global Avg Loss: 1.17085714, Time: 0.0211 Steps: 44750, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000051, Sample Num: 816, Cur Loss: 0.34711233, Cur Avg Loss: 0.28441835, Log Avg loss: 0.29769389, Global Avg Loss: 1.17066207, Time: 0.0211 Steps: 44760, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000061, Sample Num: 976, Cur Loss: 0.06444843, Cur Avg Loss: 0.27912621, Log Avg loss: 0.25213630, Global Avg Loss: 1.17045690, Time: 0.0211 Steps: 44770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000071, Sample Num: 1136, Cur Loss: 0.32052135, Cur Avg Loss: 0.28918424, Log Avg loss: 0.35053820, Global Avg Loss: 1.17027380, Time: 0.0212 Steps: 44780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000081, Sample Num: 1296, Cur Loss: 0.19303533, Cur Avg Loss: 0.27518585, Log Avg loss: 0.17579730, Global Avg Loss: 1.17005177, Time: 0.0211 Steps: 44790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000091, Sample Num: 1456, Cur Loss: 0.22487122, Cur Avg Loss: 0.27105619, Log Avg loss: 0.23760593, Global Avg Loss: 1.16984364, Time: 0.0212 Steps: 44800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000101, Sample Num: 1616, Cur Loss: 0.36379352, Cur Avg Loss: 0.26860364, Log Avg loss: 0.24628549, Global Avg Loss: 1.16963753, Time: 0.0211 Steps: 44810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000111, Sample Num: 1776, Cur Loss: 0.29658026, Cur Avg Loss: 0.26960439, Log Avg loss: 0.27971187, Global Avg Loss: 1.16943898, Time: 0.0211 Steps: 44820, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000121, Sample Num: 1936, Cur Loss: 0.14324066, Cur Avg Loss: 0.26885607, Log Avg loss: 0.26054970, Global Avg Loss: 1.16923624, Time: 0.0211 Steps: 44830, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000131, Sample Num: 2096, Cur Loss: 0.21986619, Cur Avg Loss: 0.26742599, Log Avg loss: 0.25012204, Global Avg Loss: 1.16903126, Time: 0.0211 Steps: 44840, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000141, Sample Num: 2256, Cur Loss: 0.31022939, Cur Avg Loss: 0.26939538, Log Avg loss: 0.29519440, Global Avg Loss: 1.16883642, Time: 0.0211 Steps: 44850, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000151, Sample Num: 2416, Cur Loss: 0.15770864, Cur Avg Loss: 0.27915696, Log Avg loss: 0.41679533, Global Avg Loss: 1.16866878, Time: 0.0211 Steps: 44860, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000161, Sample Num: 2576, Cur Loss: 0.36760062, Cur Avg Loss: 0.28439837, Log Avg loss: 0.36354355, Global Avg Loss: 1.16848935, Time: 0.0213 Steps: 44870, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000171, Sample Num: 2736, Cur Loss: 0.14526886, Cur Avg Loss: 0.28818873, Log Avg loss: 0.34921360, Global Avg Loss: 1.16830680, Time: 0.0211 Steps: 44880, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000181, Sample Num: 2896, Cur Loss: 0.37853080, Cur Avg Loss: 0.29067943, Log Avg loss: 0.33327035, Global Avg Loss: 1.16812078, Time: 0.0211 Steps: 44890, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000191, Sample Num: 3056, Cur Loss: 0.26728576, Cur Avg Loss: 0.29218408, Log Avg loss: 0.31941834, Global Avg Loss: 1.16793176, Time: 0.0211 Steps: 44900, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000201, Sample Num: 3216, Cur Loss: 0.36346933, Cur Avg Loss: 0.28742208, Log Avg loss: 0.19646786, Global Avg Loss: 1.16771545, Time: 0.0212 Steps: 44910, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000211, Sample Num: 3376, Cur Loss: 0.19620010, Cur Avg Loss: 0.28413847, Log Avg loss: 0.21813794, Global Avg Loss: 1.16750405, Time: 0.0211 Steps: 44920, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000221, Sample Num: 3536, Cur Loss: 0.30795577, Cur Avg Loss: 0.28305917, Log Avg loss: 0.26028592, Global Avg Loss: 1.16730213, Time: 0.0211 Steps: 44930, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000231, Sample Num: 3696, Cur Loss: 0.31822890, Cur Avg Loss: 0.28047403, Log Avg loss: 0.22334246, Global Avg Loss: 1.16709209, Time: 0.0211 Steps: 44940, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000241, Sample Num: 3856, Cur Loss: 0.79169047, Cur Avg Loss: 0.27873263, Log Avg loss: 0.23850619, Global Avg Loss: 1.16688550, Time: 0.0211 Steps: 44950, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000251, Sample Num: 4016, Cur Loss: 0.16669628, Cur Avg Loss: 0.27718526, Log Avg loss: 0.23989373, Global Avg Loss: 1.16667932, Time: 0.0211 Steps: 44960, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000261, Sample Num: 4176, Cur Loss: 0.12231924, Cur Avg Loss: 0.27819910, Log Avg loss: 0.30364632, Global Avg Loss: 1.16648741, Time: 0.0213 Steps: 44970, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000271, Sample Num: 4336, Cur Loss: 0.25964028, Cur Avg Loss: 0.27833360, Log Avg loss: 0.28184412, Global Avg Loss: 1.16629073, Time: 0.0211 Steps: 44980, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000281, Sample Num: 4496, Cur Loss: 0.29818541, Cur Avg Loss: 0.27823741, Log Avg loss: 0.27563082, Global Avg Loss: 1.16609277, Time: 0.0212 Steps: 44990, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000291, Sample Num: 4656, Cur Loss: 0.04494156, Cur Avg Loss: 0.27578256, Log Avg loss: 0.20680122, Global Avg Loss: 1.16587959, Time: 0.0211 Steps: 45000, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000301, Sample Num: 4816, Cur Loss: 0.33904117, Cur Avg Loss: 0.27561358, Log Avg loss: 0.27069619, Global Avg Loss: 1.16568070, Time: 0.0211 Steps: 45010, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000311, Sample Num: 4976, Cur Loss: 0.10168044, Cur Avg Loss: 0.27702743, Log Avg loss: 0.31958417, Global Avg Loss: 1.16549277, Time: 0.0211 Steps: 45020, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000321, Sample Num: 5136, Cur Loss: 0.37258774, Cur Avg Loss: 0.27772415, Log Avg loss: 0.29939239, Global Avg Loss: 1.16530043, Time: 0.0211 Steps: 45030, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000331, Sample Num: 5296, Cur Loss: 0.57328725, Cur Avg Loss: 0.27888452, Log Avg loss: 0.31613235, Global Avg Loss: 1.16511189, Time: 0.0212 Steps: 45040, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000341, Sample Num: 5456, Cur Loss: 0.21114445, Cur Avg Loss: 0.27665605, Log Avg loss: 0.20289368, Global Avg Loss: 1.16489830, Time: 0.0211 Steps: 45050, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000351, Sample Num: 5616, Cur Loss: 0.25177115, Cur Avg Loss: 0.27785173, Log Avg loss: 0.31862445, Global Avg Loss: 1.16471049, Time: 0.0211 Steps: 45060, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000361, Sample Num: 5776, Cur Loss: 0.09905800, Cur Avg Loss: 0.28263621, Log Avg loss: 0.45057127, Global Avg Loss: 1.16455204, Time: 0.0211 Steps: 45070, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000371, Sample Num: 5936, Cur Loss: 0.08009785, Cur Avg Loss: 0.28496122, Log Avg loss: 0.36889422, Global Avg Loss: 1.16437554, Time: 0.0211 Steps: 45080, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000381, Sample Num: 6096, Cur Loss: 0.12473769, Cur Avg Loss: 0.28688600, Log Avg loss: 0.35829522, Global Avg Loss: 1.16419677, Time: 0.0211 Steps: 45090, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000391, Sample Num: 6256, Cur Loss: 0.31895193, Cur Avg Loss: 0.29284430, Log Avg loss: 0.51985552, Global Avg Loss: 1.16405390, Time: 0.0211 Steps: 45100, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000401, Sample Num: 6416, Cur Loss: 0.23359314, Cur Avg Loss: 0.29470495, Log Avg loss: 0.36745652, Global Avg Loss: 1.16387731, Time: 0.0211 Steps: 45110, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000411, Sample Num: 6576, Cur Loss: 0.28189510, Cur Avg Loss: 0.29544305, Log Avg loss: 0.32504079, Global Avg Loss: 1.16369140, Time: 0.0211 Steps: 45120, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000421, Sample Num: 6736, Cur Loss: 0.32721978, Cur Avg Loss: 0.29448681, Log Avg loss: 0.25518517, Global Avg Loss: 1.16349009, Time: 0.0212 Steps: 45130, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000431, Sample Num: 6896, Cur Loss: 0.55681658, Cur Avg Loss: 0.29476305, Log Avg loss: 0.30639307, Global Avg Loss: 1.16330022, Time: 0.0211 Steps: 45140, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000441, Sample Num: 7056, Cur Loss: 0.23724626, Cur Avg Loss: 0.29430035, Log Avg loss: 0.27435792, Global Avg Loss: 1.16310333, Time: 0.0211 Steps: 45150, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000451, Sample Num: 7216, Cur Loss: 0.04957167, Cur Avg Loss: 0.29207607, Log Avg loss: 0.19398534, Global Avg Loss: 1.16288873, Time: 0.0211 Steps: 45160, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000461, Sample Num: 7376, Cur Loss: 0.08425275, Cur Avg Loss: 0.29134987, Log Avg loss: 0.25859818, Global Avg Loss: 1.16268853, Time: 0.0212 Steps: 45170, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000471, Sample Num: 7536, Cur Loss: 0.23920910, Cur Avg Loss: 0.29039473, Log Avg loss: 0.24636301, Global Avg Loss: 1.16248572, Time: 0.0211 Steps: 45180, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000481, Sample Num: 7696, Cur Loss: 0.16402793, Cur Avg Loss: 0.28824071, Log Avg loss: 0.18678622, Global Avg Loss: 1.16226981, Time: 0.0211 Steps: 45190, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000491, Sample Num: 7856, Cur Loss: 0.08458459, Cur Avg Loss: 0.28696740, Log Avg loss: 0.22572134, Global Avg Loss: 1.16206261, Time: 0.0211 Steps: 45200, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000501, Sample Num: 8016, Cur Loss: 0.37707376, Cur Avg Loss: 0.28890048, Log Avg loss: 0.38381426, Global Avg Loss: 1.16189047, Time: 0.0212 Steps: 45210, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000511, Sample Num: 8176, Cur Loss: 0.11088449, Cur Avg Loss: 0.28818575, Log Avg loss: 0.25237791, Global Avg Loss: 1.16168934, Time: 0.0211 Steps: 45220, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000521, Sample Num: 8336, Cur Loss: 0.16447277, Cur Avg Loss: 0.28953952, Log Avg loss: 0.35871735, Global Avg Loss: 1.16151180, Time: 0.0213 Steps: 45230, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000531, Sample Num: 8496, Cur Loss: 0.27748221, Cur Avg Loss: 0.29089888, Log Avg loss: 0.36172143, Global Avg Loss: 1.16133502, Time: 0.0210 Steps: 45240, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000541, Sample Num: 8656, Cur Loss: 0.16770144, Cur Avg Loss: 0.29209955, Log Avg loss: 0.35585487, Global Avg Loss: 1.16115701, Time: 0.0210 Steps: 45250, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000551, Sample Num: 8816, Cur Loss: 0.11875254, Cur Avg Loss: 0.29155026, Log Avg loss: 0.26183419, Global Avg Loss: 1.16095831, Time: 0.0210 Steps: 45260, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000561, Sample Num: 8976, Cur Loss: 0.26706868, Cur Avg Loss: 0.29137427, Log Avg loss: 0.28167712, Global Avg Loss: 1.16076408, Time: 0.0210 Steps: 45270, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000571, Sample Num: 9136, Cur Loss: 0.42833060, Cur Avg Loss: 0.29094199, Log Avg loss: 0.26669115, Global Avg Loss: 1.16056662, Time: 0.0210 Steps: 45280, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000581, Sample Num: 9296, Cur Loss: 0.30782166, Cur Avg Loss: 0.29261235, Log Avg loss: 0.38798961, Global Avg Loss: 1.16039604, Time: 0.0210 Steps: 45290, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000591, Sample Num: 9456, Cur Loss: 0.07618918, Cur Avg Loss: 0.29115318, Log Avg loss: 0.20637577, Global Avg Loss: 1.16018544, Time: 0.0210 Steps: 45300, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000601, Sample Num: 9616, Cur Loss: 0.07517861, Cur Avg Loss: 0.29108461, Log Avg loss: 0.28703178, Global Avg Loss: 1.15999273, Time: 0.0210 Steps: 45310, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000611, Sample Num: 9776, Cur Loss: 0.32238480, Cur Avg Loss: 0.29049118, Log Avg loss: 0.25482640, Global Avg Loss: 1.15979300, Time: 0.0210 Steps: 45320, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000621, Sample Num: 9936, Cur Loss: 0.17214048, Cur Avg Loss: 0.28909186, Log Avg loss: 0.20359301, Global Avg Loss: 1.15958206, Time: 0.0210 Steps: 45330, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000631, Sample Num: 10096, Cur Loss: 0.46238300, Cur Avg Loss: 0.28866339, Log Avg loss: 0.26205545, Global Avg Loss: 1.15938411, Time: 0.0210 Steps: 45340, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000641, Sample Num: 10256, Cur Loss: 0.50309473, Cur Avg Loss: 0.28898708, Log Avg loss: 0.30941207, Global Avg Loss: 1.15919668, Time: 0.0210 Steps: 45350, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000651, Sample Num: 10416, Cur Loss: 0.20065458, Cur Avg Loss: 0.28958586, Log Avg loss: 0.32796788, Global Avg Loss: 1.15901343, Time: 0.0210 Steps: 45360, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000661, Sample Num: 10576, Cur Loss: 0.16432802, Cur Avg Loss: 0.28896252, Log Avg loss: 0.24838261, Global Avg Loss: 1.15881272, Time: 0.0210 Steps: 45370, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000671, Sample Num: 10736, Cur Loss: 0.10150983, Cur Avg Loss: 0.28826211, Log Avg loss: 0.24196500, Global Avg Loss: 1.15861068, Time: 0.0210 Steps: 45380, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000681, Sample Num: 10896, Cur Loss: 0.25217634, Cur Avg Loss: 0.28715512, Log Avg loss: 0.21287596, Global Avg Loss: 1.15840232, Time: 0.0210 Steps: 45390, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000691, Sample Num: 11056, Cur Loss: 0.13094848, Cur Avg Loss: 0.28654065, Log Avg loss: 0.24469533, Global Avg Loss: 1.15820107, Time: 0.0210 Steps: 45400, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000701, Sample Num: 11216, Cur Loss: 0.41692138, Cur Avg Loss: 0.28758168, Log Avg loss: 0.35951732, Global Avg Loss: 1.15802518, Time: 0.0210 Steps: 45410, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000711, Sample Num: 11376, Cur Loss: 0.24733841, Cur Avg Loss: 0.28866285, Log Avg loss: 0.36445292, Global Avg Loss: 1.15785047, Time: 0.0210 Steps: 45420, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000721, Sample Num: 11536, Cur Loss: 0.08368183, Cur Avg Loss: 0.28781036, Log Avg loss: 0.22719773, Global Avg Loss: 1.15764561, Time: 0.0210 Steps: 45430, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000731, Sample Num: 11696, Cur Loss: 0.14498673, Cur Avg Loss: 0.28728623, Log Avg loss: 0.24949707, Global Avg Loss: 1.15744575, Time: 0.0210 Steps: 45440, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000741, Sample Num: 11856, Cur Loss: 0.22378165, Cur Avg Loss: 0.28598389, Log Avg loss: 0.19078246, Global Avg Loss: 1.15723307, Time: 0.0210 Steps: 45450, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000751, Sample Num: 12016, Cur Loss: 0.33482140, Cur Avg Loss: 0.28653721, Log Avg loss: 0.32753797, Global Avg Loss: 1.15705056, Time: 0.0210 Steps: 45460, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000761, Sample Num: 12176, Cur Loss: 0.41158059, Cur Avg Loss: 0.28821276, Log Avg loss: 0.41404654, Global Avg Loss: 1.15688715, Time: 0.0210 Steps: 45470, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000771, Sample Num: 12336, Cur Loss: 0.47983551, Cur Avg Loss: 0.28887258, Log Avg loss: 0.33908484, Global Avg Loss: 1.15670733, Time: 0.0247 Steps: 45480, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000781, Sample Num: 12496, Cur Loss: 0.22247706, Cur Avg Loss: 0.29032935, Log Avg loss: 0.40264682, Global Avg Loss: 1.15654157, Time: 0.0211 Steps: 45490, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000791, Sample Num: 12656, Cur Loss: 0.14545672, Cur Avg Loss: 0.29393101, Log Avg loss: 0.57522057, Global Avg Loss: 1.15641381, Time: 0.0210 Steps: 45500, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000801, Sample Num: 12816, Cur Loss: 0.41216332, Cur Avg Loss: 0.29339572, Log Avg loss: 0.25105428, Global Avg Loss: 1.15621487, Time: 0.0211 Steps: 45510, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000811, Sample Num: 12976, Cur Loss: 0.36614764, Cur Avg Loss: 0.29268593, Log Avg loss: 0.23583178, Global Avg Loss: 1.15601268, Time: 0.0211 Steps: 45520, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000821, Sample Num: 13136, Cur Loss: 0.14901888, Cur Avg Loss: 0.29178423, Log Avg loss: 0.21865617, Global Avg Loss: 1.15580680, Time: 0.0211 Steps: 45530, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000831, Sample Num: 13296, Cur Loss: 0.45926008, Cur Avg Loss: 0.29185689, Log Avg loss: 0.29782202, Global Avg Loss: 1.15561840, Time: 0.0211 Steps: 45540, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000841, Sample Num: 13456, Cur Loss: 0.22815999, Cur Avg Loss: 0.29254985, Log Avg loss: 0.35013501, Global Avg Loss: 1.15544156, Time: 0.0211 Steps: 45550, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000851, Sample Num: 13616, Cur Loss: 0.08980476, Cur Avg Loss: 0.29223164, Log Avg loss: 0.26547005, Global Avg Loss: 1.15524622, Time: 0.0210 Steps: 45560, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000861, Sample Num: 13776, Cur Loss: 0.20865831, Cur Avg Loss: 0.29242718, Log Avg loss: 0.30906830, Global Avg Loss: 1.15506054, Time: 0.0211 Steps: 45570, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000871, Sample Num: 13936, Cur Loss: 0.41670120, Cur Avg Loss: 0.29208367, Log Avg loss: 0.26250694, Global Avg Loss: 1.15486472, Time: 0.0211 Steps: 45580, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000881, Sample Num: 14096, Cur Loss: 0.28887469, Cur Avg Loss: 0.29298654, Log Avg loss: 0.37162700, Global Avg Loss: 1.15469291, Time: 0.0211 Steps: 45590, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000891, Sample Num: 14256, Cur Loss: 0.30680627, Cur Avg Loss: 0.29351463, Log Avg loss: 0.34003863, Global Avg Loss: 1.15451426, Time: 0.0210 Steps: 45600, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000901, Sample Num: 14416, Cur Loss: 0.33975297, Cur Avg Loss: 0.29365825, Log Avg loss: 0.30645494, Global Avg Loss: 1.15432833, Time: 0.0211 Steps: 45610, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000911, Sample Num: 14576, Cur Loss: 0.31125051, Cur Avg Loss: 0.29249480, Log Avg loss: 0.18766854, Global Avg Loss: 1.15411643, Time: 0.0210 Steps: 45620, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000921, Sample Num: 14736, Cur Loss: 0.10345808, Cur Avg Loss: 0.29273929, Log Avg loss: 0.31501190, Global Avg Loss: 1.15393254, Time: 0.0211 Steps: 45630, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000931, Sample Num: 14896, Cur Loss: 0.24628040, Cur Avg Loss: 0.29274637, Log Avg loss: 0.29339858, Global Avg Loss: 1.15374399, Time: 0.0210 Steps: 45640, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000941, Sample Num: 15056, Cur Loss: 0.09321415, Cur Avg Loss: 0.29233615, Log Avg loss: 0.25414430, Global Avg Loss: 1.15354693, Time: 0.0211 Steps: 45650, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000951, Sample Num: 15216, Cur Loss: 0.27487123, Cur Avg Loss: 0.29223304, Log Avg loss: 0.28253076, Global Avg Loss: 1.15335616, Time: 0.0211 Steps: 45660, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000961, Sample Num: 15376, Cur Loss: 0.31532937, Cur Avg Loss: 0.29186705, Log Avg loss: 0.25706129, Global Avg Loss: 1.15315991, Time: 0.0210 Steps: 45670, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000971, Sample Num: 15536, Cur Loss: 0.13455424, Cur Avg Loss: 0.29306385, Log Avg loss: 0.40807671, Global Avg Loss: 1.15299680, Time: 0.0211 Steps: 45680, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000981, Sample Num: 15696, Cur Loss: 0.27787095, Cur Avg Loss: 0.29385058, Log Avg loss: 0.37024130, Global Avg Loss: 1.15282548, Time: 0.0211 Steps: 45690, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000991, Sample Num: 15856, Cur Loss: 0.20468518, Cur Avg Loss: 0.29500214, Log Avg loss: 0.40797108, Global Avg Loss: 1.15266249, Time: 0.0210 Steps: 45700, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001001, Sample Num: 16016, Cur Loss: 0.60451686, Cur Avg Loss: 0.29685260, Log Avg loss: 0.48023286, Global Avg Loss: 1.15251539, Time: 0.0210 Steps: 45710, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001011, Sample Num: 16176, Cur Loss: 0.22001681, Cur Avg Loss: 0.29807366, Log Avg loss: 0.42030158, Global Avg Loss: 1.15235523, Time: 0.0211 Steps: 45720, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001021, Sample Num: 16336, Cur Loss: 0.37481540, Cur Avg Loss: 0.29793463, Log Avg loss: 0.28387920, Global Avg Loss: 1.15216532, Time: 0.0210 Steps: 45730, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001031, Sample Num: 16496, Cur Loss: 0.23814517, Cur Avg Loss: 0.29759875, Log Avg loss: 0.26330481, Global Avg Loss: 1.15197099, Time: 0.0210 Steps: 45740, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001041, Sample Num: 16656, Cur Loss: 0.16328162, Cur Avg Loss: 0.29658744, Log Avg loss: 0.19232146, Global Avg Loss: 1.15176123, Time: 0.0209 Steps: 45750, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001051, Sample Num: 16816, Cur Loss: 0.28707916, Cur Avg Loss: 0.29557456, Log Avg loss: 0.19013378, Global Avg Loss: 1.15155109, Time: 0.0209 Steps: 45760, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001061, Sample Num: 16976, Cur Loss: 0.22762722, Cur Avg Loss: 0.29568154, Log Avg loss: 0.30692493, Global Avg Loss: 1.15136655, Time: 0.0209 Steps: 45770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001071, Sample Num: 17136, Cur Loss: 0.52977312, Cur Avg Loss: 0.29539193, Log Avg loss: 0.26466497, Global Avg Loss: 1.15117286, Time: 0.0209 Steps: 45780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001081, Sample Num: 17296, Cur Loss: 0.44289422, Cur Avg Loss: 0.29579234, Log Avg loss: 0.33867619, Global Avg Loss: 1.15099542, Time: 0.0209 Steps: 45790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001091, Sample Num: 17456, Cur Loss: 0.56586099, Cur Avg Loss: 0.29661961, Log Avg loss: 0.38604723, Global Avg Loss: 1.15082840, Time: 0.0209 Steps: 45800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001101, Sample Num: 17616, Cur Loss: 0.13590756, Cur Avg Loss: 0.29625710, Log Avg loss: 0.25670748, Global Avg Loss: 1.15063322, Time: 0.0209 Steps: 45810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001111, Sample Num: 17776, Cur Loss: 0.30136243, Cur Avg Loss: 0.29573190, Log Avg loss: 0.23790672, Global Avg Loss: 1.15043402, Time: 0.0209 Steps: 45820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001121, Sample Num: 17936, Cur Loss: 0.26053762, Cur Avg Loss: 0.29565576, Log Avg loss: 0.28719695, Global Avg Loss: 1.15024567, Time: 0.0209 Steps: 45830, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001131, Sample Num: 18096, Cur Loss: 0.09787185, Cur Avg Loss: 0.29493972, Log Avg loss: 0.21467163, Global Avg Loss: 1.15004157, Time: 0.0209 Steps: 45840, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001141, Sample Num: 18256, Cur Loss: 0.13575846, Cur Avg Loss: 0.29462362, Log Avg loss: 0.25887314, Global Avg Loss: 1.14984721, Time: 0.0209 Steps: 45850, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001151, Sample Num: 18416, Cur Loss: 0.08903395, Cur Avg Loss: 0.29462741, Log Avg loss: 0.29505986, Global Avg Loss: 1.14966082, Time: 0.0208 Steps: 45860, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001161, Sample Num: 18576, Cur Loss: 0.34060040, Cur Avg Loss: 0.29377693, Log Avg loss: 0.19588580, Global Avg Loss: 1.14945289, Time: 0.0209 Steps: 45870, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001171, Sample Num: 18736, Cur Loss: 0.83782864, Cur Avg Loss: 0.29451362, Log Avg loss: 0.38004360, Global Avg Loss: 1.14928518, Time: 0.0209 Steps: 45880, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001181, Sample Num: 18896, Cur Loss: 0.33689404, Cur Avg Loss: 0.29413925, Log Avg loss: 0.25030029, Global Avg Loss: 1.14908928, Time: 0.0208 Steps: 45890, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001191, Sample Num: 19056, Cur Loss: 0.38249296, Cur Avg Loss: 0.29370188, Log Avg loss: 0.24204856, Global Avg Loss: 1.14889167, Time: 0.0209 Steps: 45900, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001201, Sample Num: 19216, Cur Loss: 0.23665465, Cur Avg Loss: 0.29374339, Log Avg loss: 0.29868763, Global Avg Loss: 1.14870648, Time: 0.0209 Steps: 45910, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001211, Sample Num: 19376, Cur Loss: 0.46222404, Cur Avg Loss: 0.29375765, Log Avg loss: 0.29546975, Global Avg Loss: 1.14852067, Time: 0.0209 Steps: 45920, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001221, Sample Num: 19536, Cur Loss: 0.17035028, Cur Avg Loss: 0.29389915, Log Avg loss: 0.31103508, Global Avg Loss: 1.14833833, Time: 0.0209 Steps: 45930, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001231, Sample Num: 19696, Cur Loss: 0.32995823, Cur Avg Loss: 0.29414345, Log Avg loss: 0.32397241, Global Avg Loss: 1.14815889, Time: 0.0209 Steps: 45940, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001241, Sample Num: 19856, Cur Loss: 0.20598677, Cur Avg Loss: 0.29418097, Log Avg loss: 0.29880006, Global Avg Loss: 1.14797405, Time: 0.0209 Steps: 45950, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001251, Sample Num: 20016, Cur Loss: 0.69535208, Cur Avg Loss: 0.29450282, Log Avg loss: 0.33444375, Global Avg Loss: 1.14779704, Time: 0.0211 Steps: 45960, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001261, Sample Num: 20176, Cur Loss: 0.19250157, Cur Avg Loss: 0.29470941, Log Avg loss: 0.32055445, Global Avg Loss: 1.14761709, Time: 0.0208 Steps: 45970, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001271, Sample Num: 20336, Cur Loss: 0.43971533, Cur Avg Loss: 0.29440170, Log Avg loss: 0.25559964, Global Avg Loss: 1.14742308, Time: 0.0208 Steps: 45980, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001281, Sample Num: 20496, Cur Loss: 0.19830637, Cur Avg Loss: 0.29409239, Log Avg loss: 0.25477887, Global Avg Loss: 1.14722899, Time: 0.0245 Steps: 45990, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001291, Sample Num: 20656, Cur Loss: 0.23177800, Cur Avg Loss: 0.29405967, Log Avg loss: 0.28986810, Global Avg Loss: 1.14704261, Time: 0.0207 Steps: 46000, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001301, Sample Num: 20816, Cur Loss: 0.69435686, Cur Avg Loss: 0.29394395, Log Avg loss: 0.27900445, Global Avg Loss: 1.14685394, Time: 0.0207 Steps: 46010, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001311, Sample Num: 20976, Cur Loss: 0.29679009, Cur Avg Loss: 0.29338200, Log Avg loss: 0.22027278, Global Avg Loss: 1.14665260, Time: 0.0207 Steps: 46020, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001321, Sample Num: 21136, Cur Loss: 0.23395303, Cur Avg Loss: 0.29335933, Log Avg loss: 0.29038703, Global Avg Loss: 1.14646658, Time: 0.0210 Steps: 46030, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001331, Sample Num: 21296, Cur Loss: 0.19498652, Cur Avg Loss: 0.29328102, Log Avg loss: 0.28293555, Global Avg Loss: 1.14627902, Time: 0.0208 Steps: 46040, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001341, Sample Num: 21456, Cur Loss: 0.36946714, Cur Avg Loss: 0.29354420, Log Avg loss: 0.32857406, Global Avg Loss: 1.14610145, Time: 0.0207 Steps: 46050, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001351, Sample Num: 21616, Cur Loss: 0.28151271, Cur Avg Loss: 0.29375382, Log Avg loss: 0.32186327, Global Avg Loss: 1.14592250, Time: 0.0209 Steps: 46060, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001361, Sample Num: 21776, Cur Loss: 0.30355608, Cur Avg Loss: 0.29317920, Log Avg loss: 0.21554801, Global Avg Loss: 1.14572055, Time: 0.0209 Steps: 46070, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001371, Sample Num: 21936, Cur Loss: 0.39055461, Cur Avg Loss: 0.29335309, Log Avg loss: 0.31702003, Global Avg Loss: 1.14554071, Time: 0.0210 Steps: 46080, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001381, Sample Num: 22096, Cur Loss: 0.48666337, Cur Avg Loss: 0.29306222, Log Avg loss: 0.25318369, Global Avg Loss: 1.14534710, Time: 0.0210 Steps: 46090, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001391, Sample Num: 22256, Cur Loss: 0.57953262, Cur Avg Loss: 0.29361429, Log Avg loss: 0.36985507, Global Avg Loss: 1.14517888, Time: 0.0209 Steps: 46100, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001401, Sample Num: 22416, Cur Loss: 0.45005733, Cur Avg Loss: 0.29428278, Log Avg loss: 0.38726978, Global Avg Loss: 1.14501451, Time: 0.0210 Steps: 46110, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001411, Sample Num: 22576, Cur Loss: 0.28297436, Cur Avg Loss: 0.29419041, Log Avg loss: 0.28124965, Global Avg Loss: 1.14482722, Time: 0.0209 Steps: 46120, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001421, Sample Num: 22736, Cur Loss: 0.23406738, Cur Avg Loss: 0.29357495, Log Avg loss: 0.20673407, Global Avg Loss: 1.14462386, Time: 0.0208 Steps: 46130, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001431, Sample Num: 22896, Cur Loss: 0.26779270, Cur Avg Loss: 0.29312292, Log Avg loss: 0.22888968, Global Avg Loss: 1.14442540, Time: 0.0209 Steps: 46140, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001441, Sample Num: 23056, Cur Loss: 0.73877895, Cur Avg Loss: 0.29358303, Log Avg loss: 0.35942377, Global Avg Loss: 1.14425530, Time: 0.0210 Steps: 46150, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001451, Sample Num: 23216, Cur Loss: 0.64118218, Cur Avg Loss: 0.29477541, Log Avg loss: 0.46659791, Global Avg Loss: 1.14410849, Time: 0.0210 Steps: 46160, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001461, Sample Num: 23376, Cur Loss: 0.10314969, Cur Avg Loss: 0.29444260, Log Avg loss: 0.24615241, Global Avg Loss: 1.14391400, Time: 0.0207 Steps: 46170, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001471, Sample Num: 23536, Cur Loss: 0.17538917, Cur Avg Loss: 0.29401474, Log Avg loss: 0.23150299, Global Avg Loss: 1.14371643, Time: 0.0210 Steps: 46180, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001481, Sample Num: 23696, Cur Loss: 0.30048668, Cur Avg Loss: 0.29397434, Log Avg loss: 0.28803153, Global Avg Loss: 1.14353117, Time: 0.0210 Steps: 46190, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001491, Sample Num: 23856, Cur Loss: 0.29332083, Cur Avg Loss: 0.29346300, Log Avg loss: 0.21773369, Global Avg Loss: 1.14333078, Time: 0.0209 Steps: 46200, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001501, Sample Num: 24016, Cur Loss: 0.18455219, Cur Avg Loss: 0.29462083, Log Avg loss: 0.46725356, Global Avg Loss: 1.14318448, Time: 0.0210 Steps: 46210, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001511, Sample Num: 24176, Cur Loss: 0.12782672, Cur Avg Loss: 0.29417474, Log Avg loss: 0.22721678, Global Avg Loss: 1.14298630, Time: 0.0208 Steps: 46220, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001521, Sample Num: 24336, Cur Loss: 0.14287138, Cur Avg Loss: 0.29399413, Log Avg loss: 0.26670484, Global Avg Loss: 1.14279675, Time: 0.0210 Steps: 46230, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001531, Sample Num: 24496, Cur Loss: 0.57043952, Cur Avg Loss: 0.29459108, Log Avg loss: 0.38538618, Global Avg Loss: 1.14263295, Time: 0.0209 Steps: 46240, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001541, Sample Num: 24656, Cur Loss: 0.26992023, Cur Avg Loss: 0.29484862, Log Avg loss: 0.33427855, Global Avg Loss: 1.14245817, Time: 0.0247 Steps: 46250, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001551, Sample Num: 24816, Cur Loss: 0.35153961, Cur Avg Loss: 0.29523422, Log Avg loss: 0.35465481, Global Avg Loss: 1.14228788, Time: 0.0210 Steps: 46260, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001561, Sample Num: 24976, Cur Loss: 0.40770328, Cur Avg Loss: 0.29534710, Log Avg loss: 0.31285535, Global Avg Loss: 1.14210862, Time: 0.0210 Steps: 46270, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001571, Sample Num: 25136, Cur Loss: 0.31115124, Cur Avg Loss: 0.29572289, Log Avg loss: 0.35438335, Global Avg Loss: 1.14193841, Time: 0.0209 Steps: 46280, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001581, Sample Num: 25296, Cur Loss: 0.28857675, Cur Avg Loss: 0.29604963, Log Avg loss: 0.34738056, Global Avg Loss: 1.14176676, Time: 0.0210 Steps: 46290, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001591, Sample Num: 25456, Cur Loss: 0.13392970, Cur Avg Loss: 0.29557030, Log Avg loss: 0.21978764, Global Avg Loss: 1.14156763, Time: 0.0210 Steps: 46300, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001601, Sample Num: 25616, Cur Loss: 1.06771374, Cur Avg Loss: 0.29698183, Log Avg loss: 0.52155608, Global Avg Loss: 1.14143375, Time: 0.0210 Steps: 46310, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001611, Sample Num: 25776, Cur Loss: 0.28124315, Cur Avg Loss: 0.29937798, Log Avg loss: 0.68300285, Global Avg Loss: 1.14133477, Time: 0.0210 Steps: 46320, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001621, Sample Num: 25936, Cur Loss: 0.18397455, Cur Avg Loss: 0.29921048, Log Avg loss: 0.27222586, Global Avg Loss: 1.14114718, Time: 0.0210 Steps: 46330, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001631, Sample Num: 26096, Cur Loss: 0.35496086, Cur Avg Loss: 0.29875468, Log Avg loss: 0.22486857, Global Avg Loss: 1.14094945, Time: 0.0210 Steps: 46340, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001641, Sample Num: 26256, Cur Loss: 0.22143267, Cur Avg Loss: 0.29864512, Log Avg loss: 0.28077636, Global Avg Loss: 1.14076387, Time: 0.0210 Steps: 46350, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001651, Sample Num: 26416, Cur Loss: 0.29140979, Cur Avg Loss: 0.29861344, Log Avg loss: 0.29341519, Global Avg Loss: 1.14058110, Time: 0.0210 Steps: 46360, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001661, Sample Num: 26576, Cur Loss: 0.14226449, Cur Avg Loss: 0.29798451, Log Avg loss: 0.19414741, Global Avg Loss: 1.14037699, Time: 0.0210 Steps: 46370, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001671, Sample Num: 26736, Cur Loss: 0.07228436, Cur Avg Loss: 0.29734803, Log Avg loss: 0.19162996, Global Avg Loss: 1.14017243, Time: 0.0210 Steps: 46380, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001681, Sample Num: 26896, Cur Loss: 0.21436381, Cur Avg Loss: 0.29737274, Log Avg loss: 0.30150080, Global Avg Loss: 1.13999164, Time: 0.0210 Steps: 46390, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001691, Sample Num: 27056, Cur Loss: 0.27350813, Cur Avg Loss: 0.29828266, Log Avg loss: 0.45123961, Global Avg Loss: 1.13984321, Time: 0.0210 Steps: 46400, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001701, Sample Num: 27216, Cur Loss: 0.07612514, Cur Avg Loss: 0.29902065, Log Avg loss: 0.42381570, Global Avg Loss: 1.13968892, Time: 0.0210 Steps: 46410, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001711, Sample Num: 27376, Cur Loss: 0.20065264, Cur Avg Loss: 0.29874473, Log Avg loss: 0.25181065, Global Avg Loss: 1.13949765, Time: 0.0209 Steps: 46420, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001721, Sample Num: 27536, Cur Loss: 0.20323025, Cur Avg Loss: 0.29822911, Log Avg loss: 0.21000561, Global Avg Loss: 1.13929746, Time: 0.0210 Steps: 46430, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001731, Sample Num: 27696, Cur Loss: 0.19442590, Cur Avg Loss: 0.29763977, Log Avg loss: 0.19621553, Global Avg Loss: 1.13909439, Time: 0.0210 Steps: 46440, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001741, Sample Num: 27856, Cur Loss: 0.13804081, Cur Avg Loss: 0.29749251, Log Avg loss: 0.27200061, Global Avg Loss: 1.13890771, Time: 0.0210 Steps: 46450, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001751, Sample Num: 28016, Cur Loss: 0.08018544, Cur Avg Loss: 0.29712766, Log Avg loss: 0.23360889, Global Avg Loss: 1.13871286, Time: 0.0210 Steps: 46460, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001761, Sample Num: 28176, Cur Loss: 0.10475553, Cur Avg Loss: 0.29736622, Log Avg loss: 0.33913644, Global Avg Loss: 1.13854079, Time: 0.0210 Steps: 46470, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001771, Sample Num: 28336, Cur Loss: 0.26437485, Cur Avg Loss: 0.29677820, Log Avg loss: 0.19322829, Global Avg Loss: 1.13833741, Time: 0.0210 Steps: 46480, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001781, Sample Num: 28496, Cur Loss: 0.16535373, Cur Avg Loss: 0.29678640, Log Avg loss: 0.29823916, Global Avg Loss: 1.13815671, Time: 0.0210 Steps: 46490, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001791, Sample Num: 28656, Cur Loss: 0.07822365, Cur Avg Loss: 0.29646448, Log Avg loss: 0.23912997, Global Avg Loss: 1.13796337, Time: 0.0210 Steps: 46500, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001801, Sample Num: 28816, Cur Loss: 0.26332653, Cur Avg Loss: 0.29649485, Log Avg loss: 0.30193505, Global Avg Loss: 1.13778362, Time: 0.0209 Steps: 46510, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001811, Sample Num: 28976, Cur Loss: 0.24374306, Cur Avg Loss: 0.29653667, Log Avg loss: 0.30406744, Global Avg Loss: 1.13760440, Time: 0.0210 Steps: 46520, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001821, Sample Num: 29136, Cur Loss: 0.54613447, Cur Avg Loss: 0.29687491, Log Avg loss: 0.35813145, Global Avg Loss: 1.13743688, Time: 0.0209 Steps: 46530, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001831, Sample Num: 29296, Cur Loss: 0.36861044, Cur Avg Loss: 0.29702080, Log Avg loss: 0.32358641, Global Avg Loss: 1.13726201, Time: 0.0210 Steps: 46540, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001841, Sample Num: 29456, Cur Loss: 0.45532554, Cur Avg Loss: 0.29766291, Log Avg loss: 0.41523268, Global Avg Loss: 1.13710690, Time: 0.0210 Steps: 46550, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001851, Sample Num: 29616, Cur Loss: 0.47108284, Cur Avg Loss: 0.29773839, Log Avg loss: 0.31163592, Global Avg Loss: 1.13692961, Time: 0.0210 Steps: 46560, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001861, Sample Num: 29776, Cur Loss: 0.45585620, Cur Avg Loss: 0.29839839, Log Avg loss: 0.42056406, Global Avg Loss: 1.13677578, Time: 0.0209 Steps: 46570, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001871, Sample Num: 29936, Cur Loss: 0.22666770, Cur Avg Loss: 0.29834346, Log Avg loss: 0.28812037, Global Avg Loss: 1.13659359, Time: 0.0210 Steps: 46580, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001881, Sample Num: 30096, Cur Loss: 0.14319009, Cur Avg Loss: 0.29800256, Log Avg loss: 0.23421971, Global Avg Loss: 1.13639991, Time: 0.0209 Steps: 46590, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001891, Sample Num: 30256, Cur Loss: 0.11844304, Cur Avg Loss: 0.29827247, Log Avg loss: 0.34904350, Global Avg Loss: 1.13623095, Time: 0.0211 Steps: 46600, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001901, Sample Num: 30416, Cur Loss: 0.20344859, Cur Avg Loss: 0.29820880, Log Avg loss: 0.28616759, Global Avg Loss: 1.13604857, Time: 0.0210 Steps: 46610, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001911, Sample Num: 30576, Cur Loss: 0.57059002, Cur Avg Loss: 0.29936947, Log Avg loss: 0.52001366, Global Avg Loss: 1.13591643, Time: 0.0211 Steps: 46620, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001921, Sample Num: 30736, Cur Loss: 0.32230806, Cur Avg Loss: 0.30047099, Log Avg loss: 0.51097104, Global Avg Loss: 1.13578241, Time: 0.0210 Steps: 46630, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001931, Sample Num: 30896, Cur Loss: 0.13659367, Cur Avg Loss: 0.29997577, Log Avg loss: 0.20484393, Global Avg Loss: 1.13558281, Time: 0.0211 Steps: 46640, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001941, Sample Num: 31056, Cur Loss: 0.70561582, Cur Avg Loss: 0.30014778, Log Avg loss: 0.33336427, Global Avg Loss: 1.13541084, Time: 0.0210 Steps: 46650, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001951, Sample Num: 31216, Cur Loss: 0.36957961, Cur Avg Loss: 0.30126387, Log Avg loss: 0.51789536, Global Avg Loss: 1.13527850, Time: 0.0209 Steps: 46660, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001961, Sample Num: 31376, Cur Loss: 0.38652742, Cur Avg Loss: 0.30234939, Log Avg loss: 0.51413415, Global Avg Loss: 1.13514540, Time: 0.0209 Steps: 46670, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001971, Sample Num: 31536, Cur Loss: 0.30255985, Cur Avg Loss: 0.30205123, Log Avg loss: 0.24358207, Global Avg Loss: 1.13495441, Time: 0.0209 Steps: 46680, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001981, Sample Num: 31696, Cur Loss: 0.22133189, Cur Avg Loss: 0.30188409, Log Avg loss: 0.26894106, Global Avg Loss: 1.13476893, Time: 0.0209 Steps: 46690, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001991, Sample Num: 31856, Cur Loss: 0.22018081, Cur Avg Loss: 0.30165471, Log Avg loss: 0.25621436, Global Avg Loss: 1.13458080, Time: 0.0210 Steps: 46700, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002001, Sample Num: 32016, Cur Loss: 0.18770488, Cur Avg Loss: 0.30176038, Log Avg loss: 0.32280077, Global Avg Loss: 1.13440701, Time: 0.0211 Steps: 46710, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002011, Sample Num: 32176, Cur Loss: 0.21905029, Cur Avg Loss: 0.30158621, Log Avg loss: 0.26673475, Global Avg Loss: 1.13422129, Time: 0.0211 Steps: 46720, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002021, Sample Num: 32336, Cur Loss: 0.09073647, Cur Avg Loss: 0.30127382, Log Avg loss: 0.23845228, Global Avg Loss: 1.13402960, Time: 0.0210 Steps: 46730, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002031, Sample Num: 32496, Cur Loss: 0.30458456, Cur Avg Loss: 0.30124426, Log Avg loss: 0.29527015, Global Avg Loss: 1.13385015, Time: 0.0209 Steps: 46740, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002041, Sample Num: 32656, Cur Loss: 0.30992725, Cur Avg Loss: 0.30082616, Log Avg loss: 0.21590944, Global Avg Loss: 1.13365380, Time: 0.0210 Steps: 46750, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002051, Sample Num: 32816, Cur Loss: 0.51993650, Cur Avg Loss: 0.30084353, Log Avg loss: 0.30438955, Global Avg Loss: 1.13347645, Time: 0.0247 Steps: 46760, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002061, Sample Num: 32976, Cur Loss: 0.12565988, Cur Avg Loss: 0.30041210, Log Avg loss: 0.21192517, Global Avg Loss: 1.13327941, Time: 0.0210 Steps: 46770, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002071, Sample Num: 33136, Cur Loss: 0.42376980, Cur Avg Loss: 0.30017263, Log Avg loss: 0.25081777, Global Avg Loss: 1.13309077, Time: 0.0210 Steps: 46780, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002081, Sample Num: 33296, Cur Loss: 0.07966764, Cur Avg Loss: 0.29956556, Log Avg loss: 0.17384060, Global Avg Loss: 1.13288576, Time: 0.0209 Steps: 46790, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002091, Sample Num: 33456, Cur Loss: 0.36484778, Cur Avg Loss: 0.29925421, Log Avg loss: 0.23446382, Global Avg Loss: 1.13269379, Time: 0.0210 Steps: 46800, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002101, Sample Num: 33616, Cur Loss: 0.14221492, Cur Avg Loss: 0.29938286, Log Avg loss: 0.32628213, Global Avg Loss: 1.13252152, Time: 0.0209 Steps: 46810, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002111, Sample Num: 33776, Cur Loss: 0.27306452, Cur Avg Loss: 0.29936283, Log Avg loss: 0.29515498, Global Avg Loss: 1.13234267, Time: 0.0210 Steps: 46820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002121, Sample Num: 33936, Cur Loss: 0.21510802, Cur Avg Loss: 0.29933268, Log Avg loss: 0.29296752, Global Avg Loss: 1.13216343, Time: 0.0209 Steps: 46830, Updated lr: 0.000057 ***** Running evaluation checkpoint-46838 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-46838 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.916526, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.281463, "eval_total_loss": 197.868618, "eval_mae": 0.38273, "eval_mse": 0.281544, "eval_r2": 0.821032, "eval_sp_statistic": 0.879336, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.908737, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.433624, "test_total_loss": 217.679357, "test_mae": 0.404189, "test_mse": 0.433773, "test_r2": 0.720039, "test_sp_statistic": 0.861748, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.910478, "test_ps_pvalue": 0.0, "lr": 5.653105737316264e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1320073384770297, "train_cur_epoch_loss": 636.6308847051114, "train_cur_epoch_avg_loss": 0.2990281280907052, "train_cur_epoch_time": 44.916526079177856, "train_cur_epoch_avg_time": 0.02109747584742971, "epoch": 22, "step": 46838} ################################################## Training, Epoch: 0023, Batch: 000002, Sample Num: 32, Cur Loss: 1.04453468, Cur Avg Loss: 0.61296609, Log Avg loss: 0.29722078, Global Avg Loss: 1.13198518, Time: 0.0249 Steps: 46840, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000012, Sample Num: 192, Cur Loss: 0.17105606, Cur Avg Loss: 0.40439975, Log Avg loss: 0.36268648, Global Avg Loss: 1.13182097, Time: 0.0211 Steps: 46850, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000022, Sample Num: 352, Cur Loss: 0.12633918, Cur Avg Loss: 0.32966316, Log Avg loss: 0.23997925, Global Avg Loss: 1.13163065, Time: 0.0212 Steps: 46860, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000032, Sample Num: 512, Cur Loss: 0.50829226, Cur Avg Loss: 0.31818991, Log Avg loss: 0.29294877, Global Avg Loss: 1.13145171, Time: 0.0211 Steps: 46870, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000042, Sample Num: 672, Cur Loss: 0.05428711, Cur Avg Loss: 0.31341635, Log Avg loss: 0.29814098, Global Avg Loss: 1.13127396, Time: 0.0211 Steps: 46880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000052, Sample Num: 832, Cur Loss: 0.11529255, Cur Avg Loss: 0.30163299, Log Avg loss: 0.25214287, Global Avg Loss: 1.13108647, Time: 0.0210 Steps: 46890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000062, Sample Num: 992, Cur Loss: 0.42838013, Cur Avg Loss: 0.30877755, Log Avg loss: 0.34592924, Global Avg Loss: 1.13091906, Time: 0.0211 Steps: 46900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000072, Sample Num: 1152, Cur Loss: 0.07708088, Cur Avg Loss: 0.30997183, Log Avg loss: 0.31737639, Global Avg Loss: 1.13074563, Time: 0.0211 Steps: 46910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000082, Sample Num: 1312, Cur Loss: 0.11833381, Cur Avg Loss: 0.29670999, Log Avg loss: 0.20122469, Global Avg Loss: 1.13054753, Time: 0.0211 Steps: 46920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000092, Sample Num: 1472, Cur Loss: 0.50208282, Cur Avg Loss: 0.29648721, Log Avg loss: 0.29466048, Global Avg Loss: 1.13036941, Time: 0.0211 Steps: 46930, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000102, Sample Num: 1632, Cur Loss: 0.21235028, Cur Avg Loss: 0.29011497, Log Avg loss: 0.23149036, Global Avg Loss: 1.13017792, Time: 0.0211 Steps: 46940, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000112, Sample Num: 1792, Cur Loss: 0.15254895, Cur Avg Loss: 0.28082358, Log Avg loss: 0.18605136, Global Avg Loss: 1.12997683, Time: 0.0213 Steps: 46950, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000122, Sample Num: 1952, Cur Loss: 0.06580659, Cur Avg Loss: 0.28138846, Log Avg loss: 0.28771513, Global Avg Loss: 1.12979747, Time: 0.0211 Steps: 46960, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000132, Sample Num: 2112, Cur Loss: 0.34190744, Cur Avg Loss: 0.28532483, Log Avg loss: 0.33334853, Global Avg Loss: 1.12962790, Time: 0.0211 Steps: 46970, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000142, Sample Num: 2272, Cur Loss: 0.16958295, Cur Avg Loss: 0.28307368, Log Avg loss: 0.25335854, Global Avg Loss: 1.12944138, Time: 0.0211 Steps: 46980, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000152, Sample Num: 2432, Cur Loss: 0.14117461, Cur Avg Loss: 0.28436660, Log Avg loss: 0.30272609, Global Avg Loss: 1.12926545, Time: 0.0211 Steps: 46990, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000162, Sample Num: 2592, Cur Loss: 0.12514530, Cur Avg Loss: 0.27911045, Log Avg loss: 0.19921691, Global Avg Loss: 1.12906757, Time: 0.0210 Steps: 47000, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000172, Sample Num: 2752, Cur Loss: 0.27405506, Cur Avg Loss: 0.27696906, Log Avg loss: 0.24227864, Global Avg Loss: 1.12887893, Time: 0.0211 Steps: 47010, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000182, Sample Num: 2912, Cur Loss: 0.06365776, Cur Avg Loss: 0.27545298, Log Avg loss: 0.24937633, Global Avg Loss: 1.12869188, Time: 0.0211 Steps: 47020, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000192, Sample Num: 3072, Cur Loss: 0.65940678, Cur Avg Loss: 0.27760787, Log Avg loss: 0.31682689, Global Avg Loss: 1.12851925, Time: 0.0211 Steps: 47030, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000202, Sample Num: 3232, Cur Loss: 0.18917394, Cur Avg Loss: 0.27958366, Log Avg loss: 0.31751876, Global Avg Loss: 1.12834685, Time: 0.0211 Steps: 47040, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000212, Sample Num: 3392, Cur Loss: 0.23170325, Cur Avg Loss: 0.28331410, Log Avg loss: 0.35866902, Global Avg Loss: 1.12818326, Time: 0.0210 Steps: 47050, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000222, Sample Num: 3552, Cur Loss: 0.18029611, Cur Avg Loss: 0.27965175, Log Avg loss: 0.20200991, Global Avg Loss: 1.12798645, Time: 0.0211 Steps: 47060, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000232, Sample Num: 3712, Cur Loss: 0.22115761, Cur Avg Loss: 0.27734970, Log Avg loss: 0.22624426, Global Avg Loss: 1.12779488, Time: 0.0210 Steps: 47070, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000242, Sample Num: 3872, Cur Loss: 0.36318743, Cur Avg Loss: 0.27658442, Log Avg loss: 0.25882982, Global Avg Loss: 1.12761030, Time: 0.0211 Steps: 47080, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000252, Sample Num: 4032, Cur Loss: 0.36458102, Cur Avg Loss: 0.27992026, Log Avg loss: 0.36064777, Global Avg Loss: 1.12744743, Time: 0.0211 Steps: 47090, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000262, Sample Num: 4192, Cur Loss: 0.14028749, Cur Avg Loss: 0.27787635, Log Avg loss: 0.22636962, Global Avg Loss: 1.12725612, Time: 0.0246 Steps: 47100, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000272, Sample Num: 4352, Cur Loss: 0.18274599, Cur Avg Loss: 0.27785658, Log Avg loss: 0.27733866, Global Avg Loss: 1.12707571, Time: 0.0209 Steps: 47110, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000282, Sample Num: 4512, Cur Loss: 0.08158591, Cur Avg Loss: 0.28056247, Log Avg loss: 0.35416282, Global Avg Loss: 1.12691168, Time: 0.0209 Steps: 47120, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000292, Sample Num: 4672, Cur Loss: 0.07258321, Cur Avg Loss: 0.28221543, Log Avg loss: 0.32882871, Global Avg Loss: 1.12674234, Time: 0.0208 Steps: 47130, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000302, Sample Num: 4832, Cur Loss: 0.21747719, Cur Avg Loss: 0.28133825, Log Avg loss: 0.25572467, Global Avg Loss: 1.12655757, Time: 0.0209 Steps: 47140, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000312, Sample Num: 4992, Cur Loss: 0.26868662, Cur Avg Loss: 0.28140498, Log Avg loss: 0.28342010, Global Avg Loss: 1.12637875, Time: 0.0209 Steps: 47150, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000322, Sample Num: 5152, Cur Loss: 0.23005998, Cur Avg Loss: 0.28280095, Log Avg loss: 0.32635518, Global Avg Loss: 1.12620911, Time: 0.0209 Steps: 47160, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000332, Sample Num: 5312, Cur Loss: 0.23032556, Cur Avg Loss: 0.28400776, Log Avg loss: 0.32286725, Global Avg Loss: 1.12603880, Time: 0.0209 Steps: 47170, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000342, Sample Num: 5472, Cur Loss: 0.09232417, Cur Avg Loss: 0.28415991, Log Avg loss: 0.28921140, Global Avg Loss: 1.12586143, Time: 0.0209 Steps: 47180, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000352, Sample Num: 5632, Cur Loss: 0.32744449, Cur Avg Loss: 0.28353164, Log Avg loss: 0.26204472, Global Avg Loss: 1.12567838, Time: 0.0209 Steps: 47190, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000362, Sample Num: 5792, Cur Loss: 0.42200288, Cur Avg Loss: 0.28410968, Log Avg loss: 0.30445665, Global Avg Loss: 1.12550439, Time: 0.0209 Steps: 47200, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000372, Sample Num: 5952, Cur Loss: 0.17814149, Cur Avg Loss: 0.28227980, Log Avg loss: 0.21603825, Global Avg Loss: 1.12531175, Time: 0.0209 Steps: 47210, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000382, Sample Num: 6112, Cur Loss: 0.14347887, Cur Avg Loss: 0.28179282, Log Avg loss: 0.26367698, Global Avg Loss: 1.12512928, Time: 0.0209 Steps: 47220, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000392, Sample Num: 6272, Cur Loss: 0.17431062, Cur Avg Loss: 0.28243926, Log Avg loss: 0.30713338, Global Avg Loss: 1.12495609, Time: 0.0209 Steps: 47230, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000402, Sample Num: 6432, Cur Loss: 0.28418130, Cur Avg Loss: 0.28297466, Log Avg loss: 0.30396238, Global Avg Loss: 1.12478229, Time: 0.0209 Steps: 47240, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000412, Sample Num: 6592, Cur Loss: 0.15211003, Cur Avg Loss: 0.28097451, Log Avg loss: 0.20056818, Global Avg Loss: 1.12458669, Time: 0.0209 Steps: 47250, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000422, Sample Num: 6752, Cur Loss: 0.13970959, Cur Avg Loss: 0.27938180, Log Avg loss: 0.21376228, Global Avg Loss: 1.12439397, Time: 0.0209 Steps: 47260, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000432, Sample Num: 6912, Cur Loss: 0.21826495, Cur Avg Loss: 0.27763897, Log Avg loss: 0.20409175, Global Avg Loss: 1.12419928, Time: 0.0209 Steps: 47270, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000442, Sample Num: 7072, Cur Loss: 0.42495766, Cur Avg Loss: 0.27845000, Log Avg loss: 0.31348657, Global Avg Loss: 1.12402781, Time: 0.0209 Steps: 47280, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000452, Sample Num: 7232, Cur Loss: 0.18793230, Cur Avg Loss: 0.27926617, Log Avg loss: 0.31534068, Global Avg Loss: 1.12385680, Time: 0.0209 Steps: 47290, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000462, Sample Num: 7392, Cur Loss: 0.38625717, Cur Avg Loss: 0.28198380, Log Avg loss: 0.40482059, Global Avg Loss: 1.12370478, Time: 0.0210 Steps: 47300, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000472, Sample Num: 7552, Cur Loss: 0.09851845, Cur Avg Loss: 0.28193378, Log Avg loss: 0.27962293, Global Avg Loss: 1.12352637, Time: 0.0209 Steps: 47310, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000482, Sample Num: 7712, Cur Loss: 0.15861395, Cur Avg Loss: 0.28055510, Log Avg loss: 0.21548128, Global Avg Loss: 1.12333447, Time: 0.0209 Steps: 47320, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000492, Sample Num: 7872, Cur Loss: 0.39256117, Cur Avg Loss: 0.28089592, Log Avg loss: 0.29732338, Global Avg Loss: 1.12315995, Time: 0.0209 Steps: 47330, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000502, Sample Num: 8032, Cur Loss: 0.18081754, Cur Avg Loss: 0.27988376, Log Avg loss: 0.23008575, Global Avg Loss: 1.12297130, Time: 0.0209 Steps: 47340, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000512, Sample Num: 8192, Cur Loss: 0.05486339, Cur Avg Loss: 0.27865621, Log Avg loss: 0.21703287, Global Avg Loss: 1.12277997, Time: 0.0255 Steps: 47350, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000522, Sample Num: 8352, Cur Loss: 0.12031855, Cur Avg Loss: 0.27599216, Log Avg loss: 0.13959310, Global Avg Loss: 1.12257237, Time: 0.0209 Steps: 47360, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000532, Sample Num: 8512, Cur Loss: 0.09924248, Cur Avg Loss: 0.27452759, Log Avg loss: 0.19807700, Global Avg Loss: 1.12237721, Time: 0.0209 Steps: 47370, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000542, Sample Num: 8672, Cur Loss: 0.09999828, Cur Avg Loss: 0.27583176, Log Avg loss: 0.34521380, Global Avg Loss: 1.12221318, Time: 0.0210 Steps: 47380, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000552, Sample Num: 8832, Cur Loss: 0.25769234, Cur Avg Loss: 0.27727371, Log Avg loss: 0.35542739, Global Avg Loss: 1.12205138, Time: 0.0209 Steps: 47390, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000562, Sample Num: 8992, Cur Loss: 0.14847270, Cur Avg Loss: 0.28105217, Log Avg loss: 0.48962274, Global Avg Loss: 1.12191795, Time: 0.0210 Steps: 47400, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000572, Sample Num: 9152, Cur Loss: 0.15504459, Cur Avg Loss: 0.28160309, Log Avg loss: 0.31256518, Global Avg Loss: 1.12174724, Time: 0.0210 Steps: 47410, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000582, Sample Num: 9312, Cur Loss: 0.28938913, Cur Avg Loss: 0.28373497, Log Avg loss: 0.40567849, Global Avg Loss: 1.12159624, Time: 0.0211 Steps: 47420, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000592, Sample Num: 9472, Cur Loss: 0.27254528, Cur Avg Loss: 0.28563005, Log Avg loss: 0.39592382, Global Avg Loss: 1.12144324, Time: 0.0211 Steps: 47430, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000602, Sample Num: 9632, Cur Loss: 0.11855472, Cur Avg Loss: 0.28576499, Log Avg loss: 0.29375314, Global Avg Loss: 1.12126877, Time: 0.0209 Steps: 47440, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000612, Sample Num: 9792, Cur Loss: 0.12995371, Cur Avg Loss: 0.28496566, Log Avg loss: 0.23684583, Global Avg Loss: 1.12108238, Time: 0.0209 Steps: 47450, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000622, Sample Num: 9952, Cur Loss: 0.28829727, Cur Avg Loss: 0.28378789, Log Avg loss: 0.21170841, Global Avg Loss: 1.12089077, Time: 0.0209 Steps: 47460, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000632, Sample Num: 10112, Cur Loss: 0.12473208, Cur Avg Loss: 0.28355513, Log Avg loss: 0.26907744, Global Avg Loss: 1.12071132, Time: 0.0210 Steps: 47470, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000642, Sample Num: 10272, Cur Loss: 0.40907758, Cur Avg Loss: 0.28518165, Log Avg loss: 0.38797781, Global Avg Loss: 1.12055700, Time: 0.0209 Steps: 47480, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000652, Sample Num: 10432, Cur Loss: 0.18483210, Cur Avg Loss: 0.28455653, Log Avg loss: 0.24442418, Global Avg Loss: 1.12037251, Time: 0.0209 Steps: 47490, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000662, Sample Num: 10592, Cur Loss: 0.45616698, Cur Avg Loss: 0.28445170, Log Avg loss: 0.27761630, Global Avg Loss: 1.12019509, Time: 0.0210 Steps: 47500, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000672, Sample Num: 10752, Cur Loss: 0.41072345, Cur Avg Loss: 0.28700383, Log Avg loss: 0.45595476, Global Avg Loss: 1.12005528, Time: 0.0209 Steps: 47510, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000682, Sample Num: 10912, Cur Loss: 0.24705486, Cur Avg Loss: 0.28786825, Log Avg loss: 0.34595768, Global Avg Loss: 1.11989238, Time: 0.0209 Steps: 47520, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000692, Sample Num: 11072, Cur Loss: 0.37146637, Cur Avg Loss: 0.28918863, Log Avg loss: 0.37923859, Global Avg Loss: 1.11973655, Time: 0.0209 Steps: 47530, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000702, Sample Num: 11232, Cur Loss: 0.12086912, Cur Avg Loss: 0.28817841, Log Avg loss: 0.21827134, Global Avg Loss: 1.11954693, Time: 0.0209 Steps: 47540, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000712, Sample Num: 11392, Cur Loss: 0.18010789, Cur Avg Loss: 0.28846501, Log Avg loss: 0.30858417, Global Avg Loss: 1.11937638, Time: 0.0209 Steps: 47550, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000722, Sample Num: 11552, Cur Loss: 0.32668945, Cur Avg Loss: 0.28834107, Log Avg loss: 0.27951623, Global Avg Loss: 1.11919979, Time: 0.0210 Steps: 47560, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000732, Sample Num: 11712, Cur Loss: 0.14459571, Cur Avg Loss: 0.28853398, Log Avg loss: 0.30246185, Global Avg Loss: 1.11902810, Time: 0.0210 Steps: 47570, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000742, Sample Num: 11872, Cur Loss: 0.48887193, Cur Avg Loss: 0.28920137, Log Avg loss: 0.33805446, Global Avg Loss: 1.11886396, Time: 0.0210 Steps: 47580, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000752, Sample Num: 12032, Cur Loss: 0.15471071, Cur Avg Loss: 0.28865491, Log Avg loss: 0.24810747, Global Avg Loss: 1.11868099, Time: 0.0210 Steps: 47590, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000762, Sample Num: 12192, Cur Loss: 0.30387419, Cur Avg Loss: 0.28735098, Log Avg loss: 0.18929543, Global Avg Loss: 1.11848574, Time: 0.0209 Steps: 47600, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000772, Sample Num: 12352, Cur Loss: 0.41277003, Cur Avg Loss: 0.28688201, Log Avg loss: 0.25114715, Global Avg Loss: 1.11830356, Time: 0.0247 Steps: 47610, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000782, Sample Num: 12512, Cur Loss: 0.21671280, Cur Avg Loss: 0.28839621, Log Avg loss: 0.40529243, Global Avg Loss: 1.11815383, Time: 0.0210 Steps: 47620, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000792, Sample Num: 12672, Cur Loss: 0.41994381, Cur Avg Loss: 0.29102768, Log Avg loss: 0.49680871, Global Avg Loss: 1.11802338, Time: 0.0209 Steps: 47630, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000802, Sample Num: 12832, Cur Loss: 0.21364999, Cur Avg Loss: 0.29183488, Log Avg loss: 0.35576454, Global Avg Loss: 1.11786338, Time: 0.0210 Steps: 47640, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000812, Sample Num: 12992, Cur Loss: 0.53755713, Cur Avg Loss: 0.29233239, Log Avg loss: 0.33223281, Global Avg Loss: 1.11769850, Time: 0.0209 Steps: 47650, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000822, Sample Num: 13152, Cur Loss: 0.16418374, Cur Avg Loss: 0.29231162, Log Avg loss: 0.29062507, Global Avg Loss: 1.11752497, Time: 0.0210 Steps: 47660, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000832, Sample Num: 13312, Cur Loss: 0.40268379, Cur Avg Loss: 0.29207803, Log Avg loss: 0.27287670, Global Avg Loss: 1.11734778, Time: 0.0209 Steps: 47670, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000842, Sample Num: 13472, Cur Loss: 0.17310326, Cur Avg Loss: 0.29117507, Log Avg loss: 0.21604923, Global Avg Loss: 1.11715875, Time: 0.0209 Steps: 47680, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000852, Sample Num: 13632, Cur Loss: 0.16955395, Cur Avg Loss: 0.29003201, Log Avg loss: 0.19378668, Global Avg Loss: 1.11696513, Time: 0.0209 Steps: 47690, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000862, Sample Num: 13792, Cur Loss: 0.51272410, Cur Avg Loss: 0.28932793, Log Avg loss: 0.22934016, Global Avg Loss: 1.11677904, Time: 0.0209 Steps: 47700, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000872, Sample Num: 13952, Cur Loss: 0.23487434, Cur Avg Loss: 0.28911267, Log Avg loss: 0.27055741, Global Avg Loss: 1.11660168, Time: 0.0209 Steps: 47710, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000882, Sample Num: 14112, Cur Loss: 0.28797078, Cur Avg Loss: 0.28926778, Log Avg loss: 0.30279323, Global Avg Loss: 1.11643114, Time: 0.0209 Steps: 47720, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000892, Sample Num: 14272, Cur Loss: 0.48496410, Cur Avg Loss: 0.29025152, Log Avg loss: 0.37701690, Global Avg Loss: 1.11627622, Time: 0.0211 Steps: 47730, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000902, Sample Num: 14432, Cur Loss: 0.09196146, Cur Avg Loss: 0.29114477, Log Avg loss: 0.37082286, Global Avg Loss: 1.11612007, Time: 0.0211 Steps: 47740, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000912, Sample Num: 14592, Cur Loss: 0.11184083, Cur Avg Loss: 0.28990490, Log Avg loss: 0.17806855, Global Avg Loss: 1.11592362, Time: 0.0211 Steps: 47750, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000922, Sample Num: 14752, Cur Loss: 0.17352179, Cur Avg Loss: 0.28995234, Log Avg loss: 0.29427868, Global Avg Loss: 1.11575159, Time: 0.0211 Steps: 47760, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000932, Sample Num: 14912, Cur Loss: 0.74125761, Cur Avg Loss: 0.29101208, Log Avg loss: 0.38872022, Global Avg Loss: 1.11559939, Time: 0.0212 Steps: 47770, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000942, Sample Num: 15072, Cur Loss: 0.46445602, Cur Avg Loss: 0.29115460, Log Avg loss: 0.30443759, Global Avg Loss: 1.11542962, Time: 0.0211 Steps: 47780, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000952, Sample Num: 15232, Cur Loss: 0.19161692, Cur Avg Loss: 0.29084268, Log Avg loss: 0.26146013, Global Avg Loss: 1.11525093, Time: 0.0211 Steps: 47790, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000962, Sample Num: 15392, Cur Loss: 0.59887505, Cur Avg Loss: 0.29076737, Log Avg loss: 0.28359809, Global Avg Loss: 1.11507694, Time: 0.0211 Steps: 47800, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000972, Sample Num: 15552, Cur Loss: 0.17250332, Cur Avg Loss: 0.29012446, Log Avg loss: 0.22827639, Global Avg Loss: 1.11489146, Time: 0.0211 Steps: 47810, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000982, Sample Num: 15712, Cur Loss: 0.23584566, Cur Avg Loss: 0.28974665, Log Avg loss: 0.25302284, Global Avg Loss: 1.11471123, Time: 0.0210 Steps: 47820, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000992, Sample Num: 15872, Cur Loss: 0.22745860, Cur Avg Loss: 0.28861971, Log Avg loss: 0.17795470, Global Avg Loss: 1.11451538, Time: 0.0211 Steps: 47830, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001002, Sample Num: 16032, Cur Loss: 0.07255495, Cur Avg Loss: 0.28869468, Log Avg loss: 0.29613178, Global Avg Loss: 1.11434431, Time: 0.0211 Steps: 47840, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001012, Sample Num: 16192, Cur Loss: 0.35099298, Cur Avg Loss: 0.28880560, Log Avg loss: 0.29991938, Global Avg Loss: 1.11417411, Time: 0.0208 Steps: 47850, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001022, Sample Num: 16352, Cur Loss: 0.18777317, Cur Avg Loss: 0.28970450, Log Avg loss: 0.38067353, Global Avg Loss: 1.11402085, Time: 0.0209 Steps: 47860, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001032, Sample Num: 16512, Cur Loss: 0.30861199, Cur Avg Loss: 0.28981597, Log Avg loss: 0.30120836, Global Avg Loss: 1.11385105, Time: 0.0209 Steps: 47870, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001042, Sample Num: 16672, Cur Loss: 0.18972781, Cur Avg Loss: 0.28962788, Log Avg loss: 0.27021677, Global Avg Loss: 1.11367485, Time: 0.0209 Steps: 47880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001052, Sample Num: 16832, Cur Loss: 0.08577201, Cur Avg Loss: 0.28885076, Log Avg loss: 0.20787477, Global Avg Loss: 1.11348571, Time: 0.0209 Steps: 47890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001062, Sample Num: 16992, Cur Loss: 0.09130950, Cur Avg Loss: 0.28867689, Log Avg loss: 0.27038569, Global Avg Loss: 1.11330970, Time: 0.0207 Steps: 47900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001072, Sample Num: 17152, Cur Loss: 0.25773564, Cur Avg Loss: 0.28957805, Log Avg loss: 0.38528091, Global Avg Loss: 1.11315774, Time: 0.0208 Steps: 47910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001082, Sample Num: 17312, Cur Loss: 0.28624278, Cur Avg Loss: 0.28863818, Log Avg loss: 0.18788491, Global Avg Loss: 1.11296465, Time: 0.0209 Steps: 47920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001092, Sample Num: 17472, Cur Loss: 0.05657206, Cur Avg Loss: 0.28823412, Log Avg loss: 0.24451488, Global Avg Loss: 1.11278346, Time: 0.0210 Steps: 47930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001102, Sample Num: 17632, Cur Loss: 0.21222395, Cur Avg Loss: 0.28773713, Log Avg loss: 0.23346498, Global Avg Loss: 1.11260004, Time: 0.0209 Steps: 47940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001112, Sample Num: 17792, Cur Loss: 0.32454473, Cur Avg Loss: 0.28772099, Log Avg loss: 0.28594232, Global Avg Loss: 1.11242764, Time: 0.0208 Steps: 47950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001122, Sample Num: 17952, Cur Loss: 0.07856879, Cur Avg Loss: 0.28824039, Log Avg loss: 0.34599779, Global Avg Loss: 1.11226784, Time: 0.0209 Steps: 47960, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001132, Sample Num: 18112, Cur Loss: 0.13851014, Cur Avg Loss: 0.28714361, Log Avg loss: 0.16408530, Global Avg Loss: 1.11207017, Time: 0.0211 Steps: 47970, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001142, Sample Num: 18272, Cur Loss: 0.17846313, Cur Avg Loss: 0.28629747, Log Avg loss: 0.19051480, Global Avg Loss: 1.11187810, Time: 0.0211 Steps: 47980, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001152, Sample Num: 18432, Cur Loss: 0.18359999, Cur Avg Loss: 0.28717046, Log Avg loss: 0.38686542, Global Avg Loss: 1.11172703, Time: 0.0210 Steps: 47990, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001162, Sample Num: 18592, Cur Loss: 0.31539673, Cur Avg Loss: 0.28651708, Log Avg loss: 0.21124799, Global Avg Loss: 1.11153943, Time: 0.0210 Steps: 48000, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001172, Sample Num: 18752, Cur Loss: 0.13492660, Cur Avg Loss: 0.28560287, Log Avg loss: 0.17937120, Global Avg Loss: 1.11134527, Time: 0.0210 Steps: 48010, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001182, Sample Num: 18912, Cur Loss: 0.07495186, Cur Avg Loss: 0.28542652, Log Avg loss: 0.26475857, Global Avg Loss: 1.11116897, Time: 0.0209 Steps: 48020, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001192, Sample Num: 19072, Cur Loss: 0.15879065, Cur Avg Loss: 0.28495266, Log Avg loss: 0.22894252, Global Avg Loss: 1.11098529, Time: 0.0210 Steps: 48030, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001202, Sample Num: 19232, Cur Loss: 0.16154748, Cur Avg Loss: 0.28522922, Log Avg loss: 0.31819463, Global Avg Loss: 1.11082026, Time: 0.0209 Steps: 48040, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001212, Sample Num: 19392, Cur Loss: 0.51462561, Cur Avg Loss: 0.28536051, Log Avg loss: 0.30114152, Global Avg Loss: 1.11065175, Time: 0.0211 Steps: 48050, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001222, Sample Num: 19552, Cur Loss: 0.52127969, Cur Avg Loss: 0.28637038, Log Avg loss: 0.40876648, Global Avg Loss: 1.11050571, Time: 0.0209 Steps: 48060, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001232, Sample Num: 19712, Cur Loss: 0.41702807, Cur Avg Loss: 0.28722909, Log Avg loss: 0.39216383, Global Avg Loss: 1.11035627, Time: 0.0210 Steps: 48070, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001242, Sample Num: 19872, Cur Loss: 0.20414278, Cur Avg Loss: 0.28724473, Log Avg loss: 0.28917143, Global Avg Loss: 1.11018548, Time: 0.0208 Steps: 48080, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001252, Sample Num: 20032, Cur Loss: 0.29089439, Cur Avg Loss: 0.28734690, Log Avg loss: 0.30003697, Global Avg Loss: 1.11001701, Time: 0.0209 Steps: 48090, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001262, Sample Num: 20192, Cur Loss: 0.22271848, Cur Avg Loss: 0.28693665, Log Avg loss: 0.23557344, Global Avg Loss: 1.10983521, Time: 0.0209 Steps: 48100, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001272, Sample Num: 20352, Cur Loss: 0.26791680, Cur Avg Loss: 0.28680320, Log Avg loss: 0.26996097, Global Avg Loss: 1.10966064, Time: 0.0209 Steps: 48110, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001282, Sample Num: 20512, Cur Loss: 0.32204396, Cur Avg Loss: 0.28693481, Log Avg loss: 0.30367646, Global Avg Loss: 1.10949315, Time: 0.0246 Steps: 48120, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001292, Sample Num: 20672, Cur Loss: 0.14790826, Cur Avg Loss: 0.28668554, Log Avg loss: 0.25472851, Global Avg Loss: 1.10931555, Time: 0.0208 Steps: 48130, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001302, Sample Num: 20832, Cur Loss: 0.17006266, Cur Avg Loss: 0.28612347, Log Avg loss: 0.21350486, Global Avg Loss: 1.10912947, Time: 0.0208 Steps: 48140, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001312, Sample Num: 20992, Cur Loss: 0.42514169, Cur Avg Loss: 0.28625111, Log Avg loss: 0.30286923, Global Avg Loss: 1.10896202, Time: 0.0208 Steps: 48150, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001322, Sample Num: 21152, Cur Loss: 0.33528355, Cur Avg Loss: 0.28711507, Log Avg loss: 0.40046683, Global Avg Loss: 1.10881491, Time: 0.0209 Steps: 48160, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001332, Sample Num: 21312, Cur Loss: 0.37100282, Cur Avg Loss: 0.28847874, Log Avg loss: 0.46875566, Global Avg Loss: 1.10868203, Time: 0.0208 Steps: 48170, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001342, Sample Num: 21472, Cur Loss: 0.35311836, Cur Avg Loss: 0.28873764, Log Avg loss: 0.32322268, Global Avg Loss: 1.10851900, Time: 0.0208 Steps: 48180, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001352, Sample Num: 21632, Cur Loss: 0.34024158, Cur Avg Loss: 0.28843823, Log Avg loss: 0.24825756, Global Avg Loss: 1.10834049, Time: 0.0208 Steps: 48190, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001362, Sample Num: 21792, Cur Loss: 0.28280956, Cur Avg Loss: 0.28827145, Log Avg loss: 0.26572311, Global Avg Loss: 1.10816567, Time: 0.0208 Steps: 48200, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001372, Sample Num: 21952, Cur Loss: 0.21544123, Cur Avg Loss: 0.28745317, Log Avg loss: 0.17600356, Global Avg Loss: 1.10797232, Time: 0.0208 Steps: 48210, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001382, Sample Num: 22112, Cur Loss: 0.52189004, Cur Avg Loss: 0.28754380, Log Avg loss: 0.29997870, Global Avg Loss: 1.10780475, Time: 0.0208 Steps: 48220, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001392, Sample Num: 22272, Cur Loss: 0.10234427, Cur Avg Loss: 0.28726875, Log Avg loss: 0.24925567, Global Avg Loss: 1.10762674, Time: 0.0208 Steps: 48230, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001402, Sample Num: 22432, Cur Loss: 0.44647989, Cur Avg Loss: 0.28710559, Log Avg loss: 0.26439400, Global Avg Loss: 1.10745194, Time: 0.0209 Steps: 48240, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001412, Sample Num: 22592, Cur Loss: 0.35822132, Cur Avg Loss: 0.28737493, Log Avg loss: 0.32513738, Global Avg Loss: 1.10728981, Time: 0.0208 Steps: 48250, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001422, Sample Num: 22752, Cur Loss: 0.41333342, Cur Avg Loss: 0.28766509, Log Avg loss: 0.32863460, Global Avg Loss: 1.10712846, Time: 0.0208 Steps: 48260, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001432, Sample Num: 22912, Cur Loss: 0.08984335, Cur Avg Loss: 0.28731289, Log Avg loss: 0.23723073, Global Avg Loss: 1.10694824, Time: 0.0209 Steps: 48270, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001442, Sample Num: 23072, Cur Loss: 0.30703962, Cur Avg Loss: 0.28728595, Log Avg loss: 0.28342748, Global Avg Loss: 1.10677767, Time: 0.0208 Steps: 48280, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001452, Sample Num: 23232, Cur Loss: 0.24692535, Cur Avg Loss: 0.28712828, Log Avg loss: 0.26439305, Global Avg Loss: 1.10660323, Time: 0.0208 Steps: 48290, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001462, Sample Num: 23392, Cur Loss: 0.12438656, Cur Avg Loss: 0.28742381, Log Avg loss: 0.33033421, Global Avg Loss: 1.10644251, Time: 0.0208 Steps: 48300, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001472, Sample Num: 23552, Cur Loss: 0.13675871, Cur Avg Loss: 0.28751160, Log Avg loss: 0.30034722, Global Avg Loss: 1.10627565, Time: 0.0208 Steps: 48310, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001482, Sample Num: 23712, Cur Loss: 0.75079823, Cur Avg Loss: 0.28833934, Log Avg loss: 0.41018259, Global Avg Loss: 1.10613159, Time: 0.0208 Steps: 48320, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001492, Sample Num: 23872, Cur Loss: 0.17143305, Cur Avg Loss: 0.28853359, Log Avg loss: 0.31732128, Global Avg Loss: 1.10596838, Time: 0.0209 Steps: 48330, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001502, Sample Num: 24032, Cur Loss: 0.15753998, Cur Avg Loss: 0.28838768, Log Avg loss: 0.26661818, Global Avg Loss: 1.10579475, Time: 0.0208 Steps: 48340, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001512, Sample Num: 24192, Cur Loss: 0.15985806, Cur Avg Loss: 0.28840175, Log Avg loss: 0.29051394, Global Avg Loss: 1.10562613, Time: 0.0208 Steps: 48350, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001522, Sample Num: 24352, Cur Loss: 0.10239519, Cur Avg Loss: 0.28820389, Log Avg loss: 0.25828799, Global Avg Loss: 1.10545091, Time: 0.0209 Steps: 48360, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001532, Sample Num: 24512, Cur Loss: 0.31992078, Cur Avg Loss: 0.28773430, Log Avg loss: 0.21626205, Global Avg Loss: 1.10526708, Time: 0.0207 Steps: 48370, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001542, Sample Num: 24672, Cur Loss: 0.24039145, Cur Avg Loss: 0.28808868, Log Avg loss: 0.34238050, Global Avg Loss: 1.10510939, Time: 0.0235 Steps: 48380, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001552, Sample Num: 24832, Cur Loss: 0.46164310, Cur Avg Loss: 0.28850656, Log Avg loss: 0.35294337, Global Avg Loss: 1.10495396, Time: 0.0208 Steps: 48390, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001562, Sample Num: 24992, Cur Loss: 0.31061989, Cur Avg Loss: 0.28850228, Log Avg loss: 0.28783847, Global Avg Loss: 1.10478513, Time: 0.0208 Steps: 48400, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001572, Sample Num: 25152, Cur Loss: 0.31719947, Cur Avg Loss: 0.28868804, Log Avg loss: 0.31770387, Global Avg Loss: 1.10462254, Time: 0.0208 Steps: 48410, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001582, Sample Num: 25312, Cur Loss: 0.17535681, Cur Avg Loss: 0.28847112, Log Avg loss: 0.25437093, Global Avg Loss: 1.10444694, Time: 0.0207 Steps: 48420, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001592, Sample Num: 25472, Cur Loss: 0.27685016, Cur Avg Loss: 0.28828639, Log Avg loss: 0.25906166, Global Avg Loss: 1.10427239, Time: 0.0207 Steps: 48430, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001602, Sample Num: 25632, Cur Loss: 0.59773785, Cur Avg Loss: 0.28866457, Log Avg loss: 0.34887058, Global Avg Loss: 1.10411644, Time: 0.0207 Steps: 48440, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001612, Sample Num: 25792, Cur Loss: 0.18109946, Cur Avg Loss: 0.28858500, Log Avg loss: 0.27583790, Global Avg Loss: 1.10394548, Time: 0.0207 Steps: 48450, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001622, Sample Num: 25952, Cur Loss: 0.23582695, Cur Avg Loss: 0.28839312, Log Avg loss: 0.25746255, Global Avg Loss: 1.10377081, Time: 0.0208 Steps: 48460, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001632, Sample Num: 26112, Cur Loss: 0.34091622, Cur Avg Loss: 0.28872427, Log Avg loss: 0.34243629, Global Avg Loss: 1.10361373, Time: 0.0207 Steps: 48470, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001642, Sample Num: 26272, Cur Loss: 0.27796990, Cur Avg Loss: 0.28891995, Log Avg loss: 0.32085553, Global Avg Loss: 1.10345227, Time: 0.0207 Steps: 48480, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001652, Sample Num: 26432, Cur Loss: 0.37935856, Cur Avg Loss: 0.28864132, Log Avg loss: 0.24289084, Global Avg Loss: 1.10327480, Time: 0.0207 Steps: 48490, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001662, Sample Num: 26592, Cur Loss: 0.14832281, Cur Avg Loss: 0.28871955, Log Avg loss: 0.30164322, Global Avg Loss: 1.10310952, Time: 0.0207 Steps: 48500, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001672, Sample Num: 26752, Cur Loss: 0.38012177, Cur Avg Loss: 0.28841826, Log Avg loss: 0.23834299, Global Avg Loss: 1.10293125, Time: 0.0207 Steps: 48510, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001682, Sample Num: 26912, Cur Loss: 0.14469112, Cur Avg Loss: 0.28788546, Log Avg loss: 0.19880199, Global Avg Loss: 1.10274491, Time: 0.0207 Steps: 48520, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001692, Sample Num: 27072, Cur Loss: 0.08760066, Cur Avg Loss: 0.28767910, Log Avg loss: 0.25296899, Global Avg Loss: 1.10256981, Time: 0.0207 Steps: 48530, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001702, Sample Num: 27232, Cur Loss: 0.21178332, Cur Avg Loss: 0.28713008, Log Avg loss: 0.19423584, Global Avg Loss: 1.10238268, Time: 0.0207 Steps: 48540, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001712, Sample Num: 27392, Cur Loss: 0.16379200, Cur Avg Loss: 0.28706641, Log Avg loss: 0.27623027, Global Avg Loss: 1.10221251, Time: 0.0207 Steps: 48550, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001722, Sample Num: 27552, Cur Loss: 0.32382122, Cur Avg Loss: 0.28695445, Log Avg loss: 0.26778703, Global Avg Loss: 1.10204068, Time: 0.0207 Steps: 48560, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001732, Sample Num: 27712, Cur Loss: 0.19103485, Cur Avg Loss: 0.28674748, Log Avg loss: 0.25110730, Global Avg Loss: 1.10186548, Time: 0.0207 Steps: 48570, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001742, Sample Num: 27872, Cur Loss: 0.17828068, Cur Avg Loss: 0.28657645, Log Avg loss: 0.25695404, Global Avg Loss: 1.10169156, Time: 0.0208 Steps: 48580, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001752, Sample Num: 28032, Cur Loss: 0.17185527, Cur Avg Loss: 0.28637605, Log Avg loss: 0.25146557, Global Avg Loss: 1.10151658, Time: 0.0207 Steps: 48590, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001762, Sample Num: 28192, Cur Loss: 0.33495176, Cur Avg Loss: 0.28617065, Log Avg loss: 0.25018504, Global Avg Loss: 1.10134141, Time: 0.0207 Steps: 48600, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001772, Sample Num: 28352, Cur Loss: 0.42232871, Cur Avg Loss: 0.28634251, Log Avg loss: 0.31662321, Global Avg Loss: 1.10117998, Time: 0.0207 Steps: 48610, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001782, Sample Num: 28512, Cur Loss: 0.30752766, Cur Avg Loss: 0.28612277, Log Avg loss: 0.24718569, Global Avg Loss: 1.10100433, Time: 0.0207 Steps: 48620, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001792, Sample Num: 28672, Cur Loss: 0.44536319, Cur Avg Loss: 0.28630374, Log Avg loss: 0.31855262, Global Avg Loss: 1.10084343, Time: 0.0253 Steps: 48630, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001802, Sample Num: 28832, Cur Loss: 0.12215661, Cur Avg Loss: 0.28583940, Log Avg loss: 0.20262966, Global Avg Loss: 1.10065876, Time: 0.0209 Steps: 48640, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001812, Sample Num: 28992, Cur Loss: 0.10900775, Cur Avg Loss: 0.28590522, Log Avg loss: 0.29776670, Global Avg Loss: 1.10049373, Time: 0.0208 Steps: 48650, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001822, Sample Num: 29152, Cur Loss: 0.34883586, Cur Avg Loss: 0.28575709, Log Avg loss: 0.25891561, Global Avg Loss: 1.10032078, Time: 0.0209 Steps: 48660, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001832, Sample Num: 29312, Cur Loss: 0.13057032, Cur Avg Loss: 0.28568137, Log Avg loss: 0.27188437, Global Avg Loss: 1.10015056, Time: 0.0208 Steps: 48670, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001842, Sample Num: 29472, Cur Loss: 0.08114484, Cur Avg Loss: 0.28495285, Log Avg loss: 0.15148810, Global Avg Loss: 1.09995569, Time: 0.0208 Steps: 48680, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001852, Sample Num: 29632, Cur Loss: 0.39457950, Cur Avg Loss: 0.28515080, Log Avg loss: 0.32161433, Global Avg Loss: 1.09979583, Time: 0.0208 Steps: 48690, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001862, Sample Num: 29792, Cur Loss: 0.80674952, Cur Avg Loss: 0.28525176, Log Avg loss: 0.30394823, Global Avg Loss: 1.09963241, Time: 0.0208 Steps: 48700, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001872, Sample Num: 29952, Cur Loss: 0.24217010, Cur Avg Loss: 0.28533100, Log Avg loss: 0.30008634, Global Avg Loss: 1.09946827, Time: 0.0208 Steps: 48710, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001882, Sample Num: 30112, Cur Loss: 0.20276222, Cur Avg Loss: 0.28488831, Log Avg loss: 0.20201631, Global Avg Loss: 1.09928406, Time: 0.0209 Steps: 48720, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001892, Sample Num: 30272, Cur Loss: 0.22046658, Cur Avg Loss: 0.28458283, Log Avg loss: 0.22709205, Global Avg Loss: 1.09910508, Time: 0.0208 Steps: 48730, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001902, Sample Num: 30432, Cur Loss: 0.54139388, Cur Avg Loss: 0.28477796, Log Avg loss: 0.32169570, Global Avg Loss: 1.09894558, Time: 0.0208 Steps: 48740, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001912, Sample Num: 30592, Cur Loss: 0.48784357, Cur Avg Loss: 0.28482494, Log Avg loss: 0.29376050, Global Avg Loss: 1.09878041, Time: 0.0208 Steps: 48750, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001922, Sample Num: 30752, Cur Loss: 0.15164170, Cur Avg Loss: 0.28460456, Log Avg loss: 0.24246785, Global Avg Loss: 1.09860479, Time: 0.0208 Steps: 48760, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001932, Sample Num: 30912, Cur Loss: 0.34836113, Cur Avg Loss: 0.28475342, Log Avg loss: 0.31336388, Global Avg Loss: 1.09844378, Time: 0.0209 Steps: 48770, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001942, Sample Num: 31072, Cur Loss: 0.38668883, Cur Avg Loss: 0.28562934, Log Avg loss: 0.45485717, Global Avg Loss: 1.09831185, Time: 0.0210 Steps: 48780, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001952, Sample Num: 31232, Cur Loss: 0.22439784, Cur Avg Loss: 0.28565560, Log Avg loss: 0.29075551, Global Avg Loss: 1.09814633, Time: 0.0208 Steps: 48790, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001962, Sample Num: 31392, Cur Loss: 0.10848823, Cur Avg Loss: 0.28573834, Log Avg loss: 0.30188925, Global Avg Loss: 1.09798316, Time: 0.0208 Steps: 48800, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001972, Sample Num: 31552, Cur Loss: 0.03946522, Cur Avg Loss: 0.28614723, Log Avg loss: 0.36637157, Global Avg Loss: 1.09783327, Time: 0.0209 Steps: 48810, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001982, Sample Num: 31712, Cur Loss: 0.43798947, Cur Avg Loss: 0.28608112, Log Avg loss: 0.27304498, Global Avg Loss: 1.09766433, Time: 0.0209 Steps: 48820, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001992, Sample Num: 31872, Cur Loss: 0.67771661, Cur Avg Loss: 0.28594629, Log Avg loss: 0.25922172, Global Avg Loss: 1.09749262, Time: 0.0209 Steps: 48830, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002002, Sample Num: 32032, Cur Loss: 0.10871740, Cur Avg Loss: 0.28602191, Log Avg loss: 0.30108569, Global Avg Loss: 1.09732956, Time: 0.0209 Steps: 48840, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002012, Sample Num: 32192, Cur Loss: 0.18027847, Cur Avg Loss: 0.28589072, Log Avg loss: 0.25962710, Global Avg Loss: 1.09715807, Time: 0.0209 Steps: 48850, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002022, Sample Num: 32352, Cur Loss: 0.49034315, Cur Avg Loss: 0.28596295, Log Avg loss: 0.30049642, Global Avg Loss: 1.09699502, Time: 0.0209 Steps: 48860, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002032, Sample Num: 32512, Cur Loss: 0.06233006, Cur Avg Loss: 0.28581397, Log Avg loss: 0.25568917, Global Avg Loss: 1.09682287, Time: 0.0208 Steps: 48870, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002042, Sample Num: 32672, Cur Loss: 0.37862661, Cur Avg Loss: 0.28558003, Log Avg loss: 0.23804280, Global Avg Loss: 1.09664718, Time: 0.0209 Steps: 48880, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002052, Sample Num: 32832, Cur Loss: 0.26315975, Cur Avg Loss: 0.28548455, Log Avg loss: 0.26598881, Global Avg Loss: 1.09647728, Time: 0.0232 Steps: 48890, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002062, Sample Num: 32992, Cur Loss: 0.20156097, Cur Avg Loss: 0.28557640, Log Avg loss: 0.30442278, Global Avg Loss: 1.09631530, Time: 0.0207 Steps: 48900, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002072, Sample Num: 33152, Cur Loss: 0.21579590, Cur Avg Loss: 0.28534491, Log Avg loss: 0.23761337, Global Avg Loss: 1.09613973, Time: 0.0207 Steps: 48910, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002082, Sample Num: 33312, Cur Loss: 0.23327854, Cur Avg Loss: 0.28457508, Log Avg loss: 0.12506534, Global Avg Loss: 1.09594123, Time: 0.0207 Steps: 48920, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002092, Sample Num: 33472, Cur Loss: 0.46531358, Cur Avg Loss: 0.28438479, Log Avg loss: 0.24476675, Global Avg Loss: 1.09576727, Time: 0.0208 Steps: 48930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002102, Sample Num: 33632, Cur Loss: 0.09519650, Cur Avg Loss: 0.28434284, Log Avg loss: 0.27556716, Global Avg Loss: 1.09559968, Time: 0.0207 Steps: 48940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002112, Sample Num: 33792, Cur Loss: 1.02132201, Cur Avg Loss: 0.28463450, Log Avg loss: 0.34594114, Global Avg Loss: 1.09544653, Time: 0.0208 Steps: 48950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002122, Sample Num: 33952, Cur Loss: 0.09805422, Cur Avg Loss: 0.28420832, Log Avg loss: 0.19419961, Global Avg Loss: 1.09526245, Time: 0.0208 Steps: 48960, Updated lr: 0.000055 ***** Running evaluation checkpoint-48967 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-48967 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.705652, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.475015, "eval_total_loss": 333.935874, "eval_mae": 0.585784, "eval_mse": 0.475002, "eval_r2": 0.698057, "eval_sp_statistic": 0.880222, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.909736, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.342303, "test_total_loss": 171.835919, "test_mae": 0.440425, "test_mse": 0.342409, "test_r2": 0.779006, "test_sp_statistic": 0.863263, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.910545, "test_ps_pvalue": 0.0, "lr": 5.451209103840683e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0951334103143255, "train_cur_epoch_loss": 604.4379832744598, "train_cur_epoch_avg_loss": 0.28390699073483316, "train_cur_epoch_time": 44.70565223693848, "train_cur_epoch_avg_time": 0.020998427542009618, "epoch": 23, "step": 48967} ################################################## Training, Epoch: 0024, Batch: 000003, Sample Num: 48, Cur Loss: 0.28287798, Cur Avg Loss: 0.20992422, Log Avg loss: 0.19776933, Global Avg Loss: 1.09507918, Time: 0.0248 Steps: 48970, Updated lr: 0.000055 Training, Epoch: 0024, Batch: 000013, Sample Num: 208, Cur Loss: 0.15480730, Cur Avg Loss: 0.27112079, Log Avg loss: 0.28947976, Global Avg Loss: 1.09491471, Time: 0.0210 Steps: 48980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000023, Sample Num: 368, Cur Loss: 0.20102718, Cur Avg Loss: 0.23989422, Log Avg loss: 0.19929968, Global Avg Loss: 1.09473189, Time: 0.0209 Steps: 48990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000033, Sample Num: 528, Cur Loss: 0.19911084, Cur Avg Loss: 0.27960545, Log Avg loss: 0.37094130, Global Avg Loss: 1.09458418, Time: 0.0209 Steps: 49000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000043, Sample Num: 688, Cur Loss: 0.43690065, Cur Avg Loss: 0.28775003, Log Avg loss: 0.31462713, Global Avg Loss: 1.09442503, Time: 0.0209 Steps: 49010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000053, Sample Num: 848, Cur Loss: 0.34662035, Cur Avg Loss: 0.27752712, Log Avg loss: 0.23356859, Global Avg Loss: 1.09424942, Time: 0.0209 Steps: 49020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000063, Sample Num: 1008, Cur Loss: 0.12976517, Cur Avg Loss: 0.26455426, Log Avg loss: 0.19579812, Global Avg Loss: 1.09406618, Time: 0.0209 Steps: 49030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000073, Sample Num: 1168, Cur Loss: 0.08634800, Cur Avg Loss: 0.26782869, Log Avg loss: 0.28845761, Global Avg Loss: 1.09390190, Time: 0.0209 Steps: 49040, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000083, Sample Num: 1328, Cur Loss: 0.10686661, Cur Avg Loss: 0.26871137, Log Avg loss: 0.27515496, Global Avg Loss: 1.09373498, Time: 0.0210 Steps: 49050, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000093, Sample Num: 1488, Cur Loss: 0.16250364, Cur Avg Loss: 0.27167726, Log Avg loss: 0.29629414, Global Avg Loss: 1.09357244, Time: 0.0208 Steps: 49060, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000103, Sample Num: 1648, Cur Loss: 0.53993928, Cur Avg Loss: 0.27034588, Log Avg loss: 0.25796405, Global Avg Loss: 1.09340215, Time: 0.0207 Steps: 49070, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000113, Sample Num: 1808, Cur Loss: 0.36463976, Cur Avg Loss: 0.26592790, Log Avg loss: 0.22042266, Global Avg Loss: 1.09322428, Time: 0.0207 Steps: 49080, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000123, Sample Num: 1968, Cur Loss: 0.06880747, Cur Avg Loss: 0.26495385, Log Avg loss: 0.25394704, Global Avg Loss: 1.09305331, Time: 0.0208 Steps: 49090, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000133, Sample Num: 2128, Cur Loss: 0.29910105, Cur Avg Loss: 0.26372443, Log Avg loss: 0.24860256, Global Avg Loss: 1.09288132, Time: 0.0207 Steps: 49100, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000143, Sample Num: 2288, Cur Loss: 0.12960756, Cur Avg Loss: 0.26170495, Log Avg loss: 0.23484591, Global Avg Loss: 1.09270661, Time: 0.0208 Steps: 49110, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000153, Sample Num: 2448, Cur Loss: 0.18843186, Cur Avg Loss: 0.26636313, Log Avg loss: 0.33297506, Global Avg Loss: 1.09255194, Time: 0.0207 Steps: 49120, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000163, Sample Num: 2608, Cur Loss: 0.15368819, Cur Avg Loss: 0.26850693, Log Avg loss: 0.30130706, Global Avg Loss: 1.09239089, Time: 0.0208 Steps: 49130, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000173, Sample Num: 2768, Cur Loss: 0.56295907, Cur Avg Loss: 0.27140525, Log Avg loss: 0.31864796, Global Avg Loss: 1.09223343, Time: 0.0208 Steps: 49140, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000183, Sample Num: 2928, Cur Loss: 0.92470998, Cur Avg Loss: 0.28156500, Log Avg loss: 0.45732871, Global Avg Loss: 1.09210425, Time: 0.0207 Steps: 49150, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000193, Sample Num: 3088, Cur Loss: 0.30787566, Cur Avg Loss: 0.27978737, Log Avg loss: 0.24725665, Global Avg Loss: 1.09193240, Time: 0.0207 Steps: 49160, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000203, Sample Num: 3248, Cur Loss: 0.19217370, Cur Avg Loss: 0.27520725, Log Avg loss: 0.18681099, Global Avg Loss: 1.09174832, Time: 0.0207 Steps: 49170, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000213, Sample Num: 3408, Cur Loss: 0.51777744, Cur Avg Loss: 0.27327220, Log Avg loss: 0.23399071, Global Avg Loss: 1.09157391, Time: 0.0207 Steps: 49180, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000223, Sample Num: 3568, Cur Loss: 0.18719593, Cur Avg Loss: 0.27046847, Log Avg loss: 0.21074908, Global Avg Loss: 1.09139484, Time: 0.0207 Steps: 49190, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000233, Sample Num: 3728, Cur Loss: 0.33608463, Cur Avg Loss: 0.27046946, Log Avg loss: 0.27049145, Global Avg Loss: 1.09122799, Time: 0.0208 Steps: 49200, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000243, Sample Num: 3888, Cur Loss: 0.27180979, Cur Avg Loss: 0.27651660, Log Avg loss: 0.41741502, Global Avg Loss: 1.09109106, Time: 0.0208 Steps: 49210, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000253, Sample Num: 4048, Cur Loss: 0.44694623, Cur Avg Loss: 0.28531910, Log Avg loss: 0.49921966, Global Avg Loss: 1.09097081, Time: 0.0208 Steps: 49220, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000263, Sample Num: 4208, Cur Loss: 0.27573299, Cur Avg Loss: 0.28350167, Log Avg loss: 0.23752073, Global Avg Loss: 1.09079745, Time: 0.0209 Steps: 49230, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000273, Sample Num: 4368, Cur Loss: 0.08951727, Cur Avg Loss: 0.28309537, Log Avg loss: 0.27240977, Global Avg Loss: 1.09063125, Time: 0.0208 Steps: 49240, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000283, Sample Num: 4528, Cur Loss: 0.17038959, Cur Avg Loss: 0.28088769, Log Avg loss: 0.22061794, Global Avg Loss: 1.09045460, Time: 0.0207 Steps: 49250, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000293, Sample Num: 4688, Cur Loss: 0.10642433, Cur Avg Loss: 0.27686578, Log Avg loss: 0.16304566, Global Avg Loss: 1.09026633, Time: 0.0208 Steps: 49260, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000303, Sample Num: 4848, Cur Loss: 0.10627134, Cur Avg Loss: 0.27360103, Log Avg loss: 0.17794413, Global Avg Loss: 1.09008116, Time: 0.0208 Steps: 49270, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000313, Sample Num: 5008, Cur Loss: 0.96966708, Cur Avg Loss: 0.27527256, Log Avg loss: 0.32591976, Global Avg Loss: 1.08992610, Time: 0.0207 Steps: 49280, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000323, Sample Num: 5168, Cur Loss: 0.28483444, Cur Avg Loss: 0.27540256, Log Avg loss: 0.27947153, Global Avg Loss: 1.08976167, Time: 0.0207 Steps: 49290, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000333, Sample Num: 5328, Cur Loss: 0.18785933, Cur Avg Loss: 0.27489790, Log Avg loss: 0.25859730, Global Avg Loss: 1.08959308, Time: 0.0207 Steps: 49300, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000343, Sample Num: 5488, Cur Loss: 0.13991430, Cur Avg Loss: 0.27282306, Log Avg loss: 0.20373095, Global Avg Loss: 1.08941343, Time: 0.0207 Steps: 49310, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000353, Sample Num: 5648, Cur Loss: 0.19107017, Cur Avg Loss: 0.27437441, Log Avg loss: 0.32758590, Global Avg Loss: 1.08925896, Time: 0.0207 Steps: 49320, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000363, Sample Num: 5808, Cur Loss: 0.45606363, Cur Avg Loss: 0.27598875, Log Avg loss: 0.33297466, Global Avg Loss: 1.08910565, Time: 0.0207 Steps: 49330, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000373, Sample Num: 5968, Cur Loss: 0.46315435, Cur Avg Loss: 0.28081112, Log Avg loss: 0.45586344, Global Avg Loss: 1.08897731, Time: 0.0208 Steps: 49340, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000383, Sample Num: 6128, Cur Loss: 0.67649579, Cur Avg Loss: 0.28390498, Log Avg loss: 0.39930593, Global Avg Loss: 1.08883755, Time: 0.0208 Steps: 49350, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000393, Sample Num: 6288, Cur Loss: 0.22275789, Cur Avg Loss: 0.28534374, Log Avg loss: 0.34044803, Global Avg Loss: 1.08868594, Time: 0.0207 Steps: 49360, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000403, Sample Num: 6448, Cur Loss: 0.19233447, Cur Avg Loss: 0.28533417, Log Avg loss: 0.28495826, Global Avg Loss: 1.08852314, Time: 0.0207 Steps: 49370, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000413, Sample Num: 6608, Cur Loss: 0.18852276, Cur Avg Loss: 0.28439960, Log Avg loss: 0.24673636, Global Avg Loss: 1.08835267, Time: 0.0207 Steps: 49380, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000423, Sample Num: 6768, Cur Loss: 0.23079480, Cur Avg Loss: 0.28395177, Log Avg loss: 0.26545626, Global Avg Loss: 1.08818606, Time: 0.0207 Steps: 49390, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000433, Sample Num: 6928, Cur Loss: 0.14960457, Cur Avg Loss: 0.28198691, Log Avg loss: 0.19887353, Global Avg Loss: 1.08800603, Time: 0.0207 Steps: 49400, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000443, Sample Num: 7088, Cur Loss: 0.26191652, Cur Avg Loss: 0.28070830, Log Avg loss: 0.22534432, Global Avg Loss: 1.08783144, Time: 0.0207 Steps: 49410, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000453, Sample Num: 7248, Cur Loss: 0.12075628, Cur Avg Loss: 0.28049787, Log Avg loss: 0.27117609, Global Avg Loss: 1.08766619, Time: 0.0208 Steps: 49420, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000463, Sample Num: 7408, Cur Loss: 0.41263664, Cur Avg Loss: 0.28084227, Log Avg loss: 0.29644360, Global Avg Loss: 1.08750612, Time: 0.0208 Steps: 49430, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000473, Sample Num: 7568, Cur Loss: 0.08909737, Cur Avg Loss: 0.28151166, Log Avg loss: 0.31250417, Global Avg Loss: 1.08734937, Time: 0.0208 Steps: 49440, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000483, Sample Num: 7728, Cur Loss: 0.03532929, Cur Avg Loss: 0.28078263, Log Avg loss: 0.24629948, Global Avg Loss: 1.08717929, Time: 0.0207 Steps: 49450, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000493, Sample Num: 7888, Cur Loss: 0.28202492, Cur Avg Loss: 0.27961032, Log Avg loss: 0.22298762, Global Avg Loss: 1.08700456, Time: 0.0208 Steps: 49460, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000503, Sample Num: 8048, Cur Loss: 0.20666446, Cur Avg Loss: 0.28209124, Log Avg loss: 0.40440093, Global Avg Loss: 1.08686658, Time: 0.0208 Steps: 49470, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000513, Sample Num: 8208, Cur Loss: 0.22410414, Cur Avg Loss: 0.28148568, Log Avg loss: 0.25102573, Global Avg Loss: 1.08669765, Time: 0.0245 Steps: 49480, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000523, Sample Num: 8368, Cur Loss: 0.27166939, Cur Avg Loss: 0.27972120, Log Avg loss: 0.18920340, Global Avg Loss: 1.08651630, Time: 0.0208 Steps: 49490, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000533, Sample Num: 8528, Cur Loss: 0.41502959, Cur Avg Loss: 0.27924749, Log Avg loss: 0.25447243, Global Avg Loss: 1.08634821, Time: 0.0208 Steps: 49500, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000543, Sample Num: 8688, Cur Loss: 0.25901476, Cur Avg Loss: 0.27839350, Log Avg loss: 0.23287610, Global Avg Loss: 1.08617583, Time: 0.0208 Steps: 49510, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000553, Sample Num: 8848, Cur Loss: 0.35771060, Cur Avg Loss: 0.27935131, Log Avg loss: 0.33136025, Global Avg Loss: 1.08602340, Time: 0.0208 Steps: 49520, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000563, Sample Num: 9008, Cur Loss: 0.07840623, Cur Avg Loss: 0.27882843, Log Avg loss: 0.24991335, Global Avg Loss: 1.08585460, Time: 0.0208 Steps: 49530, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000573, Sample Num: 9168, Cur Loss: 0.32468569, Cur Avg Loss: 0.27928239, Log Avg loss: 0.30484039, Global Avg Loss: 1.08569694, Time: 0.0208 Steps: 49540, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000583, Sample Num: 9328, Cur Loss: 1.18447936, Cur Avg Loss: 0.28237919, Log Avg loss: 0.45982589, Global Avg Loss: 1.08557063, Time: 0.0208 Steps: 49550, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000593, Sample Num: 9488, Cur Loss: 0.70416129, Cur Avg Loss: 0.28382261, Log Avg loss: 0.36797401, Global Avg Loss: 1.08542584, Time: 0.0207 Steps: 49560, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000603, Sample Num: 9648, Cur Loss: 0.09719223, Cur Avg Loss: 0.28307782, Log Avg loss: 0.23891131, Global Avg Loss: 1.08525507, Time: 0.0208 Steps: 49570, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000613, Sample Num: 9808, Cur Loss: 0.35970059, Cur Avg Loss: 0.28236490, Log Avg loss: 0.23937621, Global Avg Loss: 1.08508446, Time: 0.0208 Steps: 49580, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000623, Sample Num: 9968, Cur Loss: 0.21456788, Cur Avg Loss: 0.28295834, Log Avg loss: 0.31933610, Global Avg Loss: 1.08493004, Time: 0.0207 Steps: 49590, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000633, Sample Num: 10128, Cur Loss: 0.24134688, Cur Avg Loss: 0.28419034, Log Avg loss: 0.36094418, Global Avg Loss: 1.08478408, Time: 0.0208 Steps: 49600, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000643, Sample Num: 10288, Cur Loss: 0.69094384, Cur Avg Loss: 0.28539524, Log Avg loss: 0.36166512, Global Avg Loss: 1.08463832, Time: 0.0208 Steps: 49610, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000653, Sample Num: 10448, Cur Loss: 0.25195688, Cur Avg Loss: 0.28389827, Log Avg loss: 0.18764330, Global Avg Loss: 1.08445754, Time: 0.0207 Steps: 49620, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000663, Sample Num: 10608, Cur Loss: 0.17571992, Cur Avg Loss: 0.28240039, Log Avg loss: 0.18458852, Global Avg Loss: 1.08427623, Time: 0.0208 Steps: 49630, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000673, Sample Num: 10768, Cur Loss: 0.29496020, Cur Avg Loss: 0.28273734, Log Avg loss: 0.30507703, Global Avg Loss: 1.08411926, Time: 0.0208 Steps: 49640, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000683, Sample Num: 10928, Cur Loss: 0.30955055, Cur Avg Loss: 0.28233781, Log Avg loss: 0.25544946, Global Avg Loss: 1.08395235, Time: 0.0208 Steps: 49650, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000693, Sample Num: 11088, Cur Loss: 0.22595850, Cur Avg Loss: 0.28226377, Log Avg loss: 0.27720718, Global Avg Loss: 1.08378990, Time: 0.0208 Steps: 49660, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000703, Sample Num: 11248, Cur Loss: 0.47584391, Cur Avg Loss: 0.28243884, Log Avg loss: 0.29457081, Global Avg Loss: 1.08363101, Time: 0.0207 Steps: 49670, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000713, Sample Num: 11408, Cur Loss: 0.17324847, Cur Avg Loss: 0.28214235, Log Avg loss: 0.26129950, Global Avg Loss: 1.08346548, Time: 0.0207 Steps: 49680, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000723, Sample Num: 11568, Cur Loss: 0.34277007, Cur Avg Loss: 0.28235922, Log Avg loss: 0.29782185, Global Avg Loss: 1.08330737, Time: 0.0207 Steps: 49690, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000733, Sample Num: 11728, Cur Loss: 0.14176455, Cur Avg Loss: 0.28143334, Log Avg loss: 0.21449255, Global Avg Loss: 1.08313256, Time: 0.0207 Steps: 49700, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000743, Sample Num: 11888, Cur Loss: 0.29393128, Cur Avg Loss: 0.28449176, Log Avg loss: 0.50867365, Global Avg Loss: 1.08301700, Time: 0.0208 Steps: 49710, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000753, Sample Num: 12048, Cur Loss: 0.74594194, Cur Avg Loss: 0.28703475, Log Avg loss: 0.47597903, Global Avg Loss: 1.08289491, Time: 0.0208 Steps: 49720, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000763, Sample Num: 12208, Cur Loss: 0.29531506, Cur Avg Loss: 0.28822029, Log Avg loss: 0.37749104, Global Avg Loss: 1.08275306, Time: 0.0207 Steps: 49730, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000773, Sample Num: 12368, Cur Loss: 0.37856421, Cur Avg Loss: 0.28831574, Log Avg loss: 0.29559906, Global Avg Loss: 1.08259481, Time: 0.0208 Steps: 49740, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000783, Sample Num: 12528, Cur Loss: 0.30862197, Cur Avg Loss: 0.28803312, Log Avg loss: 0.26618654, Global Avg Loss: 1.08243071, Time: 0.0208 Steps: 49750, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000793, Sample Num: 12688, Cur Loss: 0.23030205, Cur Avg Loss: 0.28643493, Log Avg loss: 0.16129647, Global Avg Loss: 1.08224559, Time: 0.0208 Steps: 49760, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000803, Sample Num: 12848, Cur Loss: 0.29381531, Cur Avg Loss: 0.28752501, Log Avg loss: 0.37396839, Global Avg Loss: 1.08210328, Time: 0.0207 Steps: 49770, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000813, Sample Num: 13008, Cur Loss: 0.07890763, Cur Avg Loss: 0.28640987, Log Avg loss: 0.19686404, Global Avg Loss: 1.08192545, Time: 0.0208 Steps: 49780, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000823, Sample Num: 13168, Cur Loss: 0.39199585, Cur Avg Loss: 0.28717621, Log Avg loss: 0.34947995, Global Avg Loss: 1.08177834, Time: 0.0207 Steps: 49790, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000833, Sample Num: 13328, Cur Loss: 0.35585058, Cur Avg Loss: 0.28721252, Log Avg loss: 0.29020033, Global Avg Loss: 1.08161939, Time: 0.0207 Steps: 49800, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000843, Sample Num: 13488, Cur Loss: 0.25947523, Cur Avg Loss: 0.28665026, Log Avg loss: 0.23981469, Global Avg Loss: 1.08145039, Time: 0.0207 Steps: 49810, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000853, Sample Num: 13648, Cur Loss: 0.36149648, Cur Avg Loss: 0.28628105, Log Avg loss: 0.25515662, Global Avg Loss: 1.08128453, Time: 0.0207 Steps: 49820, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000863, Sample Num: 13808, Cur Loss: 0.17349391, Cur Avg Loss: 0.28625578, Log Avg loss: 0.28409995, Global Avg Loss: 1.08112455, Time: 0.0207 Steps: 49830, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000873, Sample Num: 13968, Cur Loss: 0.17930800, Cur Avg Loss: 0.28490143, Log Avg loss: 0.16802074, Global Avg Loss: 1.08094135, Time: 0.0207 Steps: 49840, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000883, Sample Num: 14128, Cur Loss: 0.46305048, Cur Avg Loss: 0.28455860, Log Avg loss: 0.25462994, Global Avg Loss: 1.08077559, Time: 0.0209 Steps: 49850, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000893, Sample Num: 14288, Cur Loss: 0.15798061, Cur Avg Loss: 0.28594813, Log Avg loss: 0.40864330, Global Avg Loss: 1.08064078, Time: 0.0207 Steps: 49860, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000903, Sample Num: 14448, Cur Loss: 0.25422987, Cur Avg Loss: 0.28557537, Log Avg loss: 0.25228825, Global Avg Loss: 1.08047468, Time: 0.0208 Steps: 49870, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000913, Sample Num: 14608, Cur Loss: 0.39400414, Cur Avg Loss: 0.28511274, Log Avg loss: 0.24333682, Global Avg Loss: 1.08030685, Time: 0.0207 Steps: 49880, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000923, Sample Num: 14768, Cur Loss: 0.40696415, Cur Avg Loss: 0.28547070, Log Avg loss: 0.31815230, Global Avg Loss: 1.08015408, Time: 0.0207 Steps: 49890, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000933, Sample Num: 14928, Cur Loss: 0.25841531, Cur Avg Loss: 0.28593402, Log Avg loss: 0.32869900, Global Avg Loss: 1.08000349, Time: 0.0207 Steps: 49900, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000943, Sample Num: 15088, Cur Loss: 0.16723616, Cur Avg Loss: 0.28504120, Log Avg loss: 0.20174066, Global Avg Loss: 1.07982752, Time: 0.0208 Steps: 49910, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000953, Sample Num: 15248, Cur Loss: 0.34260094, Cur Avg Loss: 0.28485674, Log Avg loss: 0.26746227, Global Avg Loss: 1.07966479, Time: 0.0207 Steps: 49920, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000963, Sample Num: 15408, Cur Loss: 0.25223741, Cur Avg Loss: 0.28432814, Log Avg loss: 0.23395234, Global Avg Loss: 1.07949541, Time: 0.0207 Steps: 49930, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000973, Sample Num: 15568, Cur Loss: 0.22463684, Cur Avg Loss: 0.28481728, Log Avg loss: 0.33192186, Global Avg Loss: 1.07934571, Time: 0.0207 Steps: 49940, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000983, Sample Num: 15728, Cur Loss: 0.25426131, Cur Avg Loss: 0.28442315, Log Avg loss: 0.24607427, Global Avg Loss: 1.07917889, Time: 0.0207 Steps: 49950, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000993, Sample Num: 15888, Cur Loss: 0.32272777, Cur Avg Loss: 0.28570222, Log Avg loss: 0.41143454, Global Avg Loss: 1.07904524, Time: 0.0207 Steps: 49960, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001003, Sample Num: 16048, Cur Loss: 0.29176965, Cur Avg Loss: 0.28512178, Log Avg loss: 0.22748467, Global Avg Loss: 1.07887482, Time: 0.0207 Steps: 49970, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001013, Sample Num: 16208, Cur Loss: 0.13815176, Cur Avg Loss: 0.28385067, Log Avg loss: 0.15635792, Global Avg Loss: 1.07869024, Time: 0.0207 Steps: 49980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001023, Sample Num: 16368, Cur Loss: 0.25632137, Cur Avg Loss: 0.28321413, Log Avg loss: 0.21873278, Global Avg Loss: 1.07851822, Time: 0.0209 Steps: 49990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001033, Sample Num: 16528, Cur Loss: 0.75883794, Cur Avg Loss: 0.28362542, Log Avg loss: 0.32569991, Global Avg Loss: 1.07836766, Time: 0.0208 Steps: 50000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001043, Sample Num: 16688, Cur Loss: 0.44775033, Cur Avg Loss: 0.28395987, Log Avg loss: 0.31850855, Global Avg Loss: 1.07821571, Time: 0.0208 Steps: 50010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001053, Sample Num: 16848, Cur Loss: 0.36063045, Cur Avg Loss: 0.28377745, Log Avg loss: 0.26475158, Global Avg Loss: 1.07805309, Time: 0.0208 Steps: 50020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001063, Sample Num: 17008, Cur Loss: 0.12474366, Cur Avg Loss: 0.28345410, Log Avg loss: 0.24940517, Global Avg Loss: 1.07788746, Time: 0.0207 Steps: 50030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001073, Sample Num: 17168, Cur Loss: 0.98113692, Cur Avg Loss: 0.28489920, Log Avg loss: 0.43851347, Global Avg Loss: 1.07775968, Time: 0.0207 Steps: 50040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001083, Sample Num: 17328, Cur Loss: 0.47468379, Cur Avg Loss: 0.28594664, Log Avg loss: 0.39833642, Global Avg Loss: 1.07762393, Time: 0.0208 Steps: 50050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001093, Sample Num: 17488, Cur Loss: 1.29605269, Cur Avg Loss: 0.28855497, Log Avg loss: 0.57103712, Global Avg Loss: 1.07752274, Time: 0.0209 Steps: 50060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001103, Sample Num: 17648, Cur Loss: 0.30892175, Cur Avg Loss: 0.28837130, Log Avg loss: 0.26829705, Global Avg Loss: 1.07736112, Time: 0.0210 Steps: 50070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001113, Sample Num: 17808, Cur Loss: 0.34896365, Cur Avg Loss: 0.28802563, Log Avg loss: 0.24989792, Global Avg Loss: 1.07719589, Time: 0.0209 Steps: 50080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001123, Sample Num: 17968, Cur Loss: 0.32087687, Cur Avg Loss: 0.28830318, Log Avg loss: 0.31919487, Global Avg Loss: 1.07704456, Time: 0.0209 Steps: 50090, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001133, Sample Num: 18128, Cur Loss: 0.21165785, Cur Avg Loss: 0.28781697, Log Avg loss: 0.23321512, Global Avg Loss: 1.07687613, Time: 0.0209 Steps: 50100, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001143, Sample Num: 18288, Cur Loss: 0.51612878, Cur Avg Loss: 0.28977641, Log Avg loss: 0.51178140, Global Avg Loss: 1.07676336, Time: 0.0209 Steps: 50110, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001153, Sample Num: 18448, Cur Loss: 0.33717310, Cur Avg Loss: 0.29123473, Log Avg loss: 0.45792069, Global Avg Loss: 1.07663989, Time: 0.0209 Steps: 50120, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001163, Sample Num: 18608, Cur Loss: 0.38630953, Cur Avg Loss: 0.29139883, Log Avg loss: 0.31031945, Global Avg Loss: 1.07648702, Time: 0.0209 Steps: 50130, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001173, Sample Num: 18768, Cur Loss: 0.40254152, Cur Avg Loss: 0.29134919, Log Avg loss: 0.28557564, Global Avg Loss: 1.07632928, Time: 0.0209 Steps: 50140, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001183, Sample Num: 18928, Cur Loss: 0.20823796, Cur Avg Loss: 0.29256794, Log Avg loss: 0.43552783, Global Avg Loss: 1.07620151, Time: 0.0209 Steps: 50150, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001193, Sample Num: 19088, Cur Loss: 0.15731323, Cur Avg Loss: 0.29216132, Log Avg loss: 0.24405817, Global Avg Loss: 1.07603561, Time: 0.0209 Steps: 50160, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001203, Sample Num: 19248, Cur Loss: 0.06949405, Cur Avg Loss: 0.29162791, Log Avg loss: 0.22799184, Global Avg Loss: 1.07586658, Time: 0.0211 Steps: 50170, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001213, Sample Num: 19408, Cur Loss: 0.35551170, Cur Avg Loss: 0.29104678, Log Avg loss: 0.22113708, Global Avg Loss: 1.07569624, Time: 0.0209 Steps: 50180, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001223, Sample Num: 19568, Cur Loss: 0.11243088, Cur Avg Loss: 0.29055432, Log Avg loss: 0.23081797, Global Avg Loss: 1.07552791, Time: 0.0209 Steps: 50190, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001233, Sample Num: 19728, Cur Loss: 0.30545238, Cur Avg Loss: 0.29028369, Log Avg loss: 0.25718629, Global Avg Loss: 1.07536489, Time: 0.0209 Steps: 50200, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001243, Sample Num: 19888, Cur Loss: 0.27914029, Cur Avg Loss: 0.28932130, Log Avg loss: 0.17065820, Global Avg Loss: 1.07518471, Time: 0.0209 Steps: 50210, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001253, Sample Num: 20048, Cur Loss: 0.11944576, Cur Avg Loss: 0.28944525, Log Avg loss: 0.30485217, Global Avg Loss: 1.07503131, Time: 0.0209 Steps: 50220, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001263, Sample Num: 20208, Cur Loss: 0.74753582, Cur Avg Loss: 0.28969650, Log Avg loss: 0.32117903, Global Avg Loss: 1.07488123, Time: 0.0209 Steps: 50230, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001273, Sample Num: 20368, Cur Loss: 0.17719646, Cur Avg Loss: 0.28950311, Log Avg loss: 0.26507733, Global Avg Loss: 1.07472005, Time: 0.0209 Steps: 50240, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001283, Sample Num: 20528, Cur Loss: 0.13715234, Cur Avg Loss: 0.28891209, Log Avg loss: 0.21367514, Global Avg Loss: 1.07454869, Time: 0.0245 Steps: 50250, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001293, Sample Num: 20688, Cur Loss: 0.21484759, Cur Avg Loss: 0.28892457, Log Avg loss: 0.29052640, Global Avg Loss: 1.07439270, Time: 0.0208 Steps: 50260, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001303, Sample Num: 20848, Cur Loss: 0.25752035, Cur Avg Loss: 0.28867038, Log Avg loss: 0.25580321, Global Avg Loss: 1.07422986, Time: 0.0208 Steps: 50270, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001313, Sample Num: 21008, Cur Loss: 1.13936162, Cur Avg Loss: 0.28968764, Log Avg loss: 0.42223717, Global Avg Loss: 1.07410019, Time: 0.0208 Steps: 50280, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001323, Sample Num: 21168, Cur Loss: 0.22014582, Cur Avg Loss: 0.28893075, Log Avg loss: 0.18955073, Global Avg Loss: 1.07392430, Time: 0.0208 Steps: 50290, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001333, Sample Num: 21328, Cur Loss: 0.14284515, Cur Avg Loss: 0.28902706, Log Avg loss: 0.30176906, Global Avg Loss: 1.07377079, Time: 0.0208 Steps: 50300, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001343, Sample Num: 21488, Cur Loss: 0.26971206, Cur Avg Loss: 0.28895982, Log Avg loss: 0.27999582, Global Avg Loss: 1.07361301, Time: 0.0208 Steps: 50310, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001353, Sample Num: 21648, Cur Loss: 0.10347790, Cur Avg Loss: 0.28825214, Log Avg loss: 0.19321057, Global Avg Loss: 1.07343805, Time: 0.0207 Steps: 50320, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001363, Sample Num: 21808, Cur Loss: 0.14887723, Cur Avg Loss: 0.28768263, Log Avg loss: 0.21062835, Global Avg Loss: 1.07326662, Time: 0.0207 Steps: 50330, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001373, Sample Num: 21968, Cur Loss: 0.08559875, Cur Avg Loss: 0.28739056, Log Avg loss: 0.24758144, Global Avg Loss: 1.07310260, Time: 0.0207 Steps: 50340, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001383, Sample Num: 22128, Cur Loss: 0.06372707, Cur Avg Loss: 0.28740365, Log Avg loss: 0.28920140, Global Avg Loss: 1.07294691, Time: 0.0208 Steps: 50350, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001393, Sample Num: 22288, Cur Loss: 0.25305110, Cur Avg Loss: 0.28694396, Log Avg loss: 0.22336796, Global Avg Loss: 1.07277821, Time: 0.0207 Steps: 50360, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001403, Sample Num: 22448, Cur Loss: 0.62763226, Cur Avg Loss: 0.28685758, Log Avg loss: 0.27482500, Global Avg Loss: 1.07261979, Time: 0.0208 Steps: 50370, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001413, Sample Num: 22608, Cur Loss: 0.15703985, Cur Avg Loss: 0.28662220, Log Avg loss: 0.25359946, Global Avg Loss: 1.07245722, Time: 0.0207 Steps: 50380, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001423, Sample Num: 22768, Cur Loss: 0.10216053, Cur Avg Loss: 0.28664004, Log Avg loss: 0.28915999, Global Avg Loss: 1.07230178, Time: 0.0207 Steps: 50390, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001433, Sample Num: 22928, Cur Loss: 0.27527368, Cur Avg Loss: 0.28652410, Log Avg loss: 0.27002536, Global Avg Loss: 1.07214259, Time: 0.0207 Steps: 50400, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001443, Sample Num: 23088, Cur Loss: 0.58764535, Cur Avg Loss: 0.28673418, Log Avg loss: 0.31683933, Global Avg Loss: 1.07199276, Time: 0.0207 Steps: 50410, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001453, Sample Num: 23248, Cur Loss: 0.12153384, Cur Avg Loss: 0.28642337, Log Avg loss: 0.24157344, Global Avg Loss: 1.07182806, Time: 0.0207 Steps: 50420, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001463, Sample Num: 23408, Cur Loss: 0.28299582, Cur Avg Loss: 0.28669127, Log Avg loss: 0.32561779, Global Avg Loss: 1.07168009, Time: 0.0207 Steps: 50430, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001473, Sample Num: 23568, Cur Loss: 0.23302913, Cur Avg Loss: 0.28674686, Log Avg loss: 0.29487871, Global Avg Loss: 1.07152609, Time: 0.0207 Steps: 50440, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001483, Sample Num: 23728, Cur Loss: 0.16064306, Cur Avg Loss: 0.28657995, Log Avg loss: 0.26199487, Global Avg Loss: 1.07136562, Time: 0.0207 Steps: 50450, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001493, Sample Num: 23888, Cur Loss: 0.62587243, Cur Avg Loss: 0.28640071, Log Avg loss: 0.25981845, Global Avg Loss: 1.07120480, Time: 0.0207 Steps: 50460, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001503, Sample Num: 24048, Cur Loss: 0.21418566, Cur Avg Loss: 0.28674332, Log Avg loss: 0.33789548, Global Avg Loss: 1.07105950, Time: 0.0208 Steps: 50470, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001513, Sample Num: 24208, Cur Loss: 0.76318204, Cur Avg Loss: 0.28661752, Log Avg loss: 0.26770999, Global Avg Loss: 1.07090036, Time: 0.0207 Steps: 50480, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001523, Sample Num: 24368, Cur Loss: 0.58300734, Cur Avg Loss: 0.28746925, Log Avg loss: 0.41633511, Global Avg Loss: 1.07077071, Time: 0.0207 Steps: 50490, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001533, Sample Num: 24528, Cur Loss: 0.19240691, Cur Avg Loss: 0.28731247, Log Avg loss: 0.26343617, Global Avg Loss: 1.07061085, Time: 0.0208 Steps: 50500, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001543, Sample Num: 24688, Cur Loss: 0.19699843, Cur Avg Loss: 0.28687436, Log Avg loss: 0.21971125, Global Avg Loss: 1.07044238, Time: 0.0224 Steps: 50510, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001553, Sample Num: 24848, Cur Loss: 0.19744855, Cur Avg Loss: 0.28684375, Log Avg loss: 0.28212097, Global Avg Loss: 1.07028634, Time: 0.0208 Steps: 50520, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001563, Sample Num: 25008, Cur Loss: 0.18476561, Cur Avg Loss: 0.28638952, Log Avg loss: 0.21584728, Global Avg Loss: 1.07011725, Time: 0.0208 Steps: 50530, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001573, Sample Num: 25168, Cur Loss: 0.35754484, Cur Avg Loss: 0.28610758, Log Avg loss: 0.24204012, Global Avg Loss: 1.06995340, Time: 0.0208 Steps: 50540, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001583, Sample Num: 25328, Cur Loss: 0.20762868, Cur Avg Loss: 0.28647512, Log Avg loss: 0.34428920, Global Avg Loss: 1.06980985, Time: 0.0207 Steps: 50550, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001593, Sample Num: 25488, Cur Loss: 0.38559225, Cur Avg Loss: 0.28704178, Log Avg loss: 0.37674384, Global Avg Loss: 1.06967277, Time: 0.0208 Steps: 50560, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001603, Sample Num: 25648, Cur Loss: 0.42399061, Cur Avg Loss: 0.28707854, Log Avg loss: 0.29293564, Global Avg Loss: 1.06951917, Time: 0.0207 Steps: 50570, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001613, Sample Num: 25808, Cur Loss: 0.40450549, Cur Avg Loss: 0.28740431, Log Avg loss: 0.33962523, Global Avg Loss: 1.06937487, Time: 0.0207 Steps: 50580, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001623, Sample Num: 25968, Cur Loss: 0.33102691, Cur Avg Loss: 0.28827164, Log Avg loss: 0.42817137, Global Avg Loss: 1.06924812, Time: 0.0208 Steps: 50590, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001633, Sample Num: 26128, Cur Loss: 0.52396041, Cur Avg Loss: 0.28865327, Log Avg loss: 0.35059258, Global Avg Loss: 1.06910610, Time: 0.0208 Steps: 50600, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001643, Sample Num: 26288, Cur Loss: 0.22482048, Cur Avg Loss: 0.28861886, Log Avg loss: 0.28299957, Global Avg Loss: 1.06895077, Time: 0.0207 Steps: 50610, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001653, Sample Num: 26448, Cur Loss: 0.18949589, Cur Avg Loss: 0.28818322, Log Avg loss: 0.21660641, Global Avg Loss: 1.06878239, Time: 0.0208 Steps: 50620, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001663, Sample Num: 26608, Cur Loss: 0.13683972, Cur Avg Loss: 0.28786871, Log Avg loss: 0.23588131, Global Avg Loss: 1.06861788, Time: 0.0208 Steps: 50630, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001673, Sample Num: 26768, Cur Loss: 0.03004452, Cur Avg Loss: 0.28767179, Log Avg loss: 0.25492309, Global Avg Loss: 1.06845720, Time: 0.0208 Steps: 50640, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001683, Sample Num: 26928, Cur Loss: 0.20582765, Cur Avg Loss: 0.28758600, Log Avg loss: 0.27323422, Global Avg Loss: 1.06830020, Time: 0.0207 Steps: 50650, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001693, Sample Num: 27088, Cur Loss: 0.28515014, Cur Avg Loss: 0.28747516, Log Avg loss: 0.26882004, Global Avg Loss: 1.06814238, Time: 0.0208 Steps: 50660, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001703, Sample Num: 27248, Cur Loss: 0.31592524, Cur Avg Loss: 0.28728444, Log Avg loss: 0.25499525, Global Avg Loss: 1.06798190, Time: 0.0208 Steps: 50670, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001713, Sample Num: 27408, Cur Loss: 0.12376375, Cur Avg Loss: 0.28723046, Log Avg loss: 0.27803761, Global Avg Loss: 1.06782604, Time: 0.0208 Steps: 50680, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001723, Sample Num: 27568, Cur Loss: 0.16790560, Cur Avg Loss: 0.28756826, Log Avg loss: 0.34543405, Global Avg Loss: 1.06768352, Time: 0.0208 Steps: 50690, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001733, Sample Num: 27728, Cur Loss: 0.16273037, Cur Avg Loss: 0.28704334, Log Avg loss: 0.19659886, Global Avg Loss: 1.06751171, Time: 0.0208 Steps: 50700, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001743, Sample Num: 27888, Cur Loss: 0.29891592, Cur Avg Loss: 0.28690948, Log Avg loss: 0.26371201, Global Avg Loss: 1.06735320, Time: 0.0208 Steps: 50710, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001753, Sample Num: 28048, Cur Loss: 0.26713014, Cur Avg Loss: 0.28637775, Log Avg loss: 0.19369699, Global Avg Loss: 1.06718095, Time: 0.0208 Steps: 50720, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001763, Sample Num: 28208, Cur Loss: 0.32028604, Cur Avg Loss: 0.28597775, Log Avg loss: 0.21585780, Global Avg Loss: 1.06701314, Time: 0.0207 Steps: 50730, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001773, Sample Num: 28368, Cur Loss: 0.08051057, Cur Avg Loss: 0.28561108, Log Avg loss: 0.22096684, Global Avg Loss: 1.06684640, Time: 0.0208 Steps: 50740, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001783, Sample Num: 28528, Cur Loss: 0.27359152, Cur Avg Loss: 0.28499903, Log Avg loss: 0.17648369, Global Avg Loss: 1.06667096, Time: 0.0208 Steps: 50750, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001793, Sample Num: 28688, Cur Loss: 0.22739907, Cur Avg Loss: 0.28479758, Log Avg loss: 0.24887853, Global Avg Loss: 1.06650985, Time: 0.0244 Steps: 50760, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001803, Sample Num: 28848, Cur Loss: 0.37996662, Cur Avg Loss: 0.28543617, Log Avg loss: 0.39993546, Global Avg Loss: 1.06637855, Time: 0.0209 Steps: 50770, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001813, Sample Num: 29008, Cur Loss: 0.53627205, Cur Avg Loss: 0.28565668, Log Avg loss: 0.32541476, Global Avg Loss: 1.06623264, Time: 0.0209 Steps: 50780, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001823, Sample Num: 29168, Cur Loss: 0.28560454, Cur Avg Loss: 0.28591898, Log Avg loss: 0.33347391, Global Avg Loss: 1.06608836, Time: 0.0208 Steps: 50790, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001833, Sample Num: 29328, Cur Loss: 0.21053192, Cur Avg Loss: 0.28569599, Log Avg loss: 0.24504507, Global Avg Loss: 1.06592674, Time: 0.0209 Steps: 50800, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001843, Sample Num: 29488, Cur Loss: 0.48748082, Cur Avg Loss: 0.28568038, Log Avg loss: 0.28281955, Global Avg Loss: 1.06577262, Time: 0.0207 Steps: 50810, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001853, Sample Num: 29648, Cur Loss: 0.29265851, Cur Avg Loss: 0.28554197, Log Avg loss: 0.26003155, Global Avg Loss: 1.06561407, Time: 0.0209 Steps: 50820, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001863, Sample Num: 29808, Cur Loss: 0.21288607, Cur Avg Loss: 0.28531934, Log Avg loss: 0.24406613, Global Avg Loss: 1.06545244, Time: 0.0209 Steps: 50830, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001873, Sample Num: 29968, Cur Loss: 0.10076694, Cur Avg Loss: 0.28551975, Log Avg loss: 0.32285653, Global Avg Loss: 1.06530638, Time: 0.0209 Steps: 50840, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001883, Sample Num: 30128, Cur Loss: 0.19851768, Cur Avg Loss: 0.28516955, Log Avg loss: 0.21957799, Global Avg Loss: 1.06514006, Time: 0.0208 Steps: 50850, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001893, Sample Num: 30288, Cur Loss: 0.36894259, Cur Avg Loss: 0.28546289, Log Avg loss: 0.34069764, Global Avg Loss: 1.06499762, Time: 0.0208 Steps: 50860, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001903, Sample Num: 30448, Cur Loss: 0.15102980, Cur Avg Loss: 0.28573997, Log Avg loss: 0.33819257, Global Avg Loss: 1.06485474, Time: 0.0210 Steps: 50870, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001913, Sample Num: 30608, Cur Loss: 0.11752655, Cur Avg Loss: 0.28514887, Log Avg loss: 0.17266075, Global Avg Loss: 1.06467939, Time: 0.0208 Steps: 50880, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001923, Sample Num: 30768, Cur Loss: 0.15082525, Cur Avg Loss: 0.28490403, Log Avg loss: 0.23806800, Global Avg Loss: 1.06451696, Time: 0.0208 Steps: 50890, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001933, Sample Num: 30928, Cur Loss: 0.11809440, Cur Avg Loss: 0.28484543, Log Avg loss: 0.27357605, Global Avg Loss: 1.06436157, Time: 0.0209 Steps: 50900, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001943, Sample Num: 31088, Cur Loss: 0.12666972, Cur Avg Loss: 0.28453385, Log Avg loss: 0.22430506, Global Avg Loss: 1.06419656, Time: 0.0209 Steps: 50910, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001953, Sample Num: 31248, Cur Loss: 0.25049013, Cur Avg Loss: 0.28504470, Log Avg loss: 0.38430385, Global Avg Loss: 1.06406304, Time: 0.0208 Steps: 50920, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001963, Sample Num: 31408, Cur Loss: 0.38519052, Cur Avg Loss: 0.28585588, Log Avg loss: 0.44427740, Global Avg Loss: 1.06394135, Time: 0.0208 Steps: 50930, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001973, Sample Num: 31568, Cur Loss: 0.04850713, Cur Avg Loss: 0.28561903, Log Avg loss: 0.23912704, Global Avg Loss: 1.06377943, Time: 0.0209 Steps: 50940, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001983, Sample Num: 31728, Cur Loss: 0.27220902, Cur Avg Loss: 0.28537468, Log Avg loss: 0.23716333, Global Avg Loss: 1.06361719, Time: 0.0208 Steps: 50950, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001993, Sample Num: 31888, Cur Loss: 0.16519266, Cur Avg Loss: 0.28521639, Log Avg loss: 0.25382840, Global Avg Loss: 1.06345828, Time: 0.0209 Steps: 50960, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002003, Sample Num: 32048, Cur Loss: 0.22221459, Cur Avg Loss: 0.28485400, Log Avg loss: 0.21262999, Global Avg Loss: 1.06329135, Time: 0.0208 Steps: 50970, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002013, Sample Num: 32208, Cur Loss: 0.23336619, Cur Avg Loss: 0.28489063, Log Avg loss: 0.29222613, Global Avg Loss: 1.06314010, Time: 0.0208 Steps: 50980, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002023, Sample Num: 32368, Cur Loss: 0.08304422, Cur Avg Loss: 0.28444345, Log Avg loss: 0.19442730, Global Avg Loss: 1.06296974, Time: 0.0208 Steps: 50990, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002033, Sample Num: 32528, Cur Loss: 0.26296642, Cur Avg Loss: 0.28465642, Log Avg loss: 0.32774063, Global Avg Loss: 1.06282557, Time: 0.0210 Steps: 51000, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002043, Sample Num: 32688, Cur Loss: 0.17301235, Cur Avg Loss: 0.28473403, Log Avg loss: 0.30051046, Global Avg Loss: 1.06267613, Time: 0.0208 Steps: 51010, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002053, Sample Num: 32848, Cur Loss: 0.09826569, Cur Avg Loss: 0.28478033, Log Avg loss: 0.29424025, Global Avg Loss: 1.06252551, Time: 0.0245 Steps: 51020, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002063, Sample Num: 33008, Cur Loss: 0.19638444, Cur Avg Loss: 0.28451056, Log Avg loss: 0.22912616, Global Avg Loss: 1.06236220, Time: 0.0209 Steps: 51030, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002073, Sample Num: 33168, Cur Loss: 0.12174486, Cur Avg Loss: 0.28438390, Log Avg loss: 0.25825475, Global Avg Loss: 1.06220465, Time: 0.0208 Steps: 51040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002083, Sample Num: 33328, Cur Loss: 0.13201405, Cur Avg Loss: 0.28417070, Log Avg loss: 0.23997400, Global Avg Loss: 1.06204359, Time: 0.0208 Steps: 51050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002093, Sample Num: 33488, Cur Loss: 0.07235851, Cur Avg Loss: 0.28410350, Log Avg loss: 0.27010628, Global Avg Loss: 1.06188849, Time: 0.0208 Steps: 51060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002103, Sample Num: 33648, Cur Loss: 0.03118221, Cur Avg Loss: 0.28383494, Log Avg loss: 0.22762589, Global Avg Loss: 1.06172513, Time: 0.0208 Steps: 51070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002113, Sample Num: 33808, Cur Loss: 0.31815219, Cur Avg Loss: 0.28361170, Log Avg loss: 0.23666267, Global Avg Loss: 1.06156361, Time: 0.0208 Steps: 51080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002123, Sample Num: 33968, Cur Loss: 0.26987347, Cur Avg Loss: 0.28398670, Log Avg loss: 0.36322571, Global Avg Loss: 1.06142692, Time: 0.0209 Steps: 51090, Updated lr: 0.000052 ***** Running evaluation checkpoint-51096 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-51096 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.507257, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.276238, "eval_total_loss": 194.195231, "eval_mae": 0.373409, "eval_mse": 0.27632, "eval_r2": 0.824353, "eval_sp_statistic": 0.880938, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.909919, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.441752, "test_total_loss": 221.759278, "test_mae": 0.411139, "test_mse": 0.441903, "test_r2": 0.714792, "test_sp_statistic": 0.862852, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.914074, "test_ps_pvalue": 0.0, "lr": 5.249312470365102e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.061332444418786, "train_cur_epoch_loss": 604.4448771607131, "train_cur_epoch_avg_loss": 0.2839102288213777, "train_cur_epoch_time": 44.507256746292114, "train_cur_epoch_avg_time": 0.020905240369324618, "epoch": 24, "step": 51096} ################################################## Training, Epoch: 0025, Batch: 000004, Sample Num: 64, Cur Loss: 0.14619304, Cur Avg Loss: 0.33314205, Log Avg loss: 0.28736728, Global Avg Loss: 1.06127544, Time: 0.0245 Steps: 51100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000014, Sample Num: 224, Cur Loss: 0.29455176, Cur Avg Loss: 0.25457495, Log Avg loss: 0.22314811, Global Avg Loss: 1.06111146, Time: 0.0208 Steps: 51110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000024, Sample Num: 384, Cur Loss: 0.24655554, Cur Avg Loss: 0.28616844, Log Avg loss: 0.33039932, Global Avg Loss: 1.06096852, Time: 0.0208 Steps: 51120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000034, Sample Num: 544, Cur Loss: 0.15069355, Cur Avg Loss: 0.27683019, Log Avg loss: 0.25441840, Global Avg Loss: 1.06081077, Time: 0.0208 Steps: 51130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000044, Sample Num: 704, Cur Loss: 0.83813232, Cur Avg Loss: 0.28115437, Log Avg loss: 0.29585658, Global Avg Loss: 1.06066119, Time: 0.0208 Steps: 51140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000054, Sample Num: 864, Cur Loss: 0.29196179, Cur Avg Loss: 0.27884472, Log Avg loss: 0.26868225, Global Avg Loss: 1.06050636, Time: 0.0208 Steps: 51150, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000064, Sample Num: 1024, Cur Loss: 0.08621334, Cur Avg Loss: 0.27426657, Log Avg loss: 0.24954454, Global Avg Loss: 1.06034784, Time: 0.0208 Steps: 51160, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000074, Sample Num: 1184, Cur Loss: 0.64727217, Cur Avg Loss: 0.28886638, Log Avg loss: 0.38230521, Global Avg Loss: 1.06021534, Time: 0.0208 Steps: 51170, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000084, Sample Num: 1344, Cur Loss: 0.26636735, Cur Avg Loss: 0.27785242, Log Avg loss: 0.19634911, Global Avg Loss: 1.06004655, Time: 0.0208 Steps: 51180, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000094, Sample Num: 1504, Cur Loss: 0.33031037, Cur Avg Loss: 0.27416081, Log Avg loss: 0.24315127, Global Avg Loss: 1.05988696, Time: 0.0208 Steps: 51190, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000104, Sample Num: 1664, Cur Loss: 0.27467307, Cur Avg Loss: 0.26908478, Log Avg loss: 0.22137006, Global Avg Loss: 1.05972319, Time: 0.0208 Steps: 51200, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000114, Sample Num: 1824, Cur Loss: 0.41201481, Cur Avg Loss: 0.27181195, Log Avg loss: 0.30017459, Global Avg Loss: 1.05957487, Time: 0.0207 Steps: 51210, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000124, Sample Num: 1984, Cur Loss: 0.10806640, Cur Avg Loss: 0.27338654, Log Avg loss: 0.29133678, Global Avg Loss: 1.05942488, Time: 0.0208 Steps: 51220, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000134, Sample Num: 2144, Cur Loss: 0.30590999, Cur Avg Loss: 0.26892660, Log Avg loss: 0.21362342, Global Avg Loss: 1.05925978, Time: 0.0207 Steps: 51230, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000144, Sample Num: 2304, Cur Loss: 0.28318653, Cur Avg Loss: 0.27024975, Log Avg loss: 0.28797993, Global Avg Loss: 1.05910926, Time: 0.0208 Steps: 51240, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000154, Sample Num: 2464, Cur Loss: 0.49728203, Cur Avg Loss: 0.26970019, Log Avg loss: 0.26178650, Global Avg Loss: 1.05895369, Time: 0.0208 Steps: 51250, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000164, Sample Num: 2624, Cur Loss: 0.11887403, Cur Avg Loss: 0.27171394, Log Avg loss: 0.30272564, Global Avg Loss: 1.05880616, Time: 0.0208 Steps: 51260, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000174, Sample Num: 2784, Cur Loss: 0.22334678, Cur Avg Loss: 0.27189244, Log Avg loss: 0.27481994, Global Avg Loss: 1.05865324, Time: 0.0208 Steps: 51270, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000184, Sample Num: 2944, Cur Loss: 0.10344343, Cur Avg Loss: 0.27217218, Log Avg loss: 0.27703955, Global Avg Loss: 1.05850082, Time: 0.0208 Steps: 51280, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000194, Sample Num: 3104, Cur Loss: 0.21096656, Cur Avg Loss: 0.26665918, Log Avg loss: 0.16522005, Global Avg Loss: 1.05832666, Time: 0.0208 Steps: 51290, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000204, Sample Num: 3264, Cur Loss: 0.33579609, Cur Avg Loss: 0.26825532, Log Avg loss: 0.29922040, Global Avg Loss: 1.05817869, Time: 0.0208 Steps: 51300, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000214, Sample Num: 3424, Cur Loss: 0.27953178, Cur Avg Loss: 0.26967714, Log Avg loss: 0.29868227, Global Avg Loss: 1.05803067, Time: 0.0209 Steps: 51310, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000224, Sample Num: 3584, Cur Loss: 0.09171314, Cur Avg Loss: 0.26924440, Log Avg loss: 0.25998385, Global Avg Loss: 1.05787516, Time: 0.0208 Steps: 51320, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000234, Sample Num: 3744, Cur Loss: 0.54440260, Cur Avg Loss: 0.27049052, Log Avg loss: 0.29840349, Global Avg Loss: 1.05772720, Time: 0.0208 Steps: 51330, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000244, Sample Num: 3904, Cur Loss: 0.05684440, Cur Avg Loss: 0.26797084, Log Avg loss: 0.20901034, Global Avg Loss: 1.05756189, Time: 0.0208 Steps: 51340, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000254, Sample Num: 4064, Cur Loss: 0.17749372, Cur Avg Loss: 0.26817164, Log Avg loss: 0.27307121, Global Avg Loss: 1.05740912, Time: 0.0208 Steps: 51350, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000264, Sample Num: 4224, Cur Loss: 0.21130428, Cur Avg Loss: 0.27060836, Log Avg loss: 0.33250109, Global Avg Loss: 1.05726797, Time: 0.0208 Steps: 51360, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000274, Sample Num: 4384, Cur Loss: 0.07444311, Cur Avg Loss: 0.26911975, Log Avg loss: 0.22982044, Global Avg Loss: 1.05710690, Time: 0.0208 Steps: 51370, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000284, Sample Num: 4544, Cur Loss: 0.23363125, Cur Avg Loss: 0.27133544, Log Avg loss: 0.33204536, Global Avg Loss: 1.05696578, Time: 0.0208 Steps: 51380, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000294, Sample Num: 4704, Cur Loss: 0.06879412, Cur Avg Loss: 0.27129335, Log Avg loss: 0.27009783, Global Avg Loss: 1.05681266, Time: 0.0208 Steps: 51390, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000304, Sample Num: 4864, Cur Loss: 0.22137964, Cur Avg Loss: 0.27081198, Log Avg loss: 0.25665972, Global Avg Loss: 1.05665699, Time: 0.0212 Steps: 51400, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000314, Sample Num: 5024, Cur Loss: 0.75421739, Cur Avg Loss: 0.28033854, Log Avg loss: 0.56994610, Global Avg Loss: 1.05656232, Time: 0.0208 Steps: 51410, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000324, Sample Num: 5184, Cur Loss: 0.18550900, Cur Avg Loss: 0.28033370, Log Avg loss: 0.28018181, Global Avg Loss: 1.05641133, Time: 0.0208 Steps: 51420, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000334, Sample Num: 5344, Cur Loss: 0.33037579, Cur Avg Loss: 0.28076938, Log Avg loss: 0.29488515, Global Avg Loss: 1.05626326, Time: 0.0208 Steps: 51430, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000344, Sample Num: 5504, Cur Loss: 0.26946282, Cur Avg Loss: 0.28091700, Log Avg loss: 0.28584758, Global Avg Loss: 1.05611349, Time: 0.0208 Steps: 51440, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000354, Sample Num: 5664, Cur Loss: 0.94551635, Cur Avg Loss: 0.28782417, Log Avg loss: 0.52543101, Global Avg Loss: 1.05601035, Time: 0.0207 Steps: 51450, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000364, Sample Num: 5824, Cur Loss: 0.32779461, Cur Avg Loss: 0.28675281, Log Avg loss: 0.24882668, Global Avg Loss: 1.05585349, Time: 0.0208 Steps: 51460, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000374, Sample Num: 5984, Cur Loss: 0.38212514, Cur Avg Loss: 0.28720339, Log Avg loss: 0.30360429, Global Avg Loss: 1.05570734, Time: 0.0209 Steps: 51470, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000384, Sample Num: 6144, Cur Loss: 0.77200550, Cur Avg Loss: 0.28556159, Log Avg loss: 0.22415820, Global Avg Loss: 1.05554581, Time: 0.0211 Steps: 51480, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000394, Sample Num: 6304, Cur Loss: 0.25330794, Cur Avg Loss: 0.28427152, Log Avg loss: 0.23473300, Global Avg Loss: 1.05538640, Time: 0.0208 Steps: 51490, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000404, Sample Num: 6464, Cur Loss: 0.29168752, Cur Avg Loss: 0.28404486, Log Avg loss: 0.27511461, Global Avg Loss: 1.05523489, Time: 0.0208 Steps: 51500, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000414, Sample Num: 6624, Cur Loss: 0.39576000, Cur Avg Loss: 0.28301703, Log Avg loss: 0.24149257, Global Avg Loss: 1.05507691, Time: 0.0208 Steps: 51510, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000424, Sample Num: 6784, Cur Loss: 0.10668494, Cur Avg Loss: 0.28216080, Log Avg loss: 0.24671284, Global Avg Loss: 1.05492001, Time: 0.0208 Steps: 51520, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000434, Sample Num: 6944, Cur Loss: 0.34572491, Cur Avg Loss: 0.28248767, Log Avg loss: 0.29634711, Global Avg Loss: 1.05477280, Time: 0.0208 Steps: 51530, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000444, Sample Num: 7104, Cur Loss: 0.25434434, Cur Avg Loss: 0.28185742, Log Avg loss: 0.25450422, Global Avg Loss: 1.05461753, Time: 0.0208 Steps: 51540, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000454, Sample Num: 7264, Cur Loss: 0.27357081, Cur Avg Loss: 0.28271797, Log Avg loss: 0.32092666, Global Avg Loss: 1.05447520, Time: 0.0208 Steps: 51550, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000464, Sample Num: 7424, Cur Loss: 0.60759652, Cur Avg Loss: 0.28364046, Log Avg loss: 0.32552165, Global Avg Loss: 1.05433382, Time: 0.0208 Steps: 51560, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000474, Sample Num: 7584, Cur Loss: 0.04678675, Cur Avg Loss: 0.28223778, Log Avg loss: 0.21715314, Global Avg Loss: 1.05417148, Time: 0.0208 Steps: 51570, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000484, Sample Num: 7744, Cur Loss: 0.15470988, Cur Avg Loss: 0.28058534, Log Avg loss: 0.20225975, Global Avg Loss: 1.05400632, Time: 0.0208 Steps: 51580, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000494, Sample Num: 7904, Cur Loss: 0.22635415, Cur Avg Loss: 0.27970812, Log Avg loss: 0.23725084, Global Avg Loss: 1.05384800, Time: 0.0208 Steps: 51590, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000504, Sample Num: 8064, Cur Loss: 0.17938703, Cur Avg Loss: 0.27953551, Log Avg loss: 0.27100820, Global Avg Loss: 1.05369629, Time: 0.0208 Steps: 51600, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000514, Sample Num: 8224, Cur Loss: 0.13618505, Cur Avg Loss: 0.28248129, Log Avg loss: 0.43094876, Global Avg Loss: 1.05357562, Time: 0.0246 Steps: 51610, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000524, Sample Num: 8384, Cur Loss: 0.23910590, Cur Avg Loss: 0.28126763, Log Avg loss: 0.21888565, Global Avg Loss: 1.05341393, Time: 0.0208 Steps: 51620, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000534, Sample Num: 8544, Cur Loss: 0.78299004, Cur Avg Loss: 0.28126315, Log Avg loss: 0.28102818, Global Avg Loss: 1.05326433, Time: 0.0208 Steps: 51630, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000544, Sample Num: 8704, Cur Loss: 0.43506932, Cur Avg Loss: 0.28132089, Log Avg loss: 0.28440411, Global Avg Loss: 1.05311544, Time: 0.0208 Steps: 51640, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000554, Sample Num: 8864, Cur Loss: 0.22432908, Cur Avg Loss: 0.28003967, Log Avg loss: 0.21034159, Global Avg Loss: 1.05295227, Time: 0.0208 Steps: 51650, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000564, Sample Num: 9024, Cur Loss: 0.14025587, Cur Avg Loss: 0.28021017, Log Avg loss: 0.28965603, Global Avg Loss: 1.05280451, Time: 0.0208 Steps: 51660, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000574, Sample Num: 9184, Cur Loss: 0.48150533, Cur Avg Loss: 0.28060791, Log Avg loss: 0.30304044, Global Avg Loss: 1.05265941, Time: 0.0208 Steps: 51670, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000584, Sample Num: 9344, Cur Loss: 1.15983319, Cur Avg Loss: 0.28508009, Log Avg loss: 0.54178316, Global Avg Loss: 1.05256055, Time: 0.0208 Steps: 51680, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000594, Sample Num: 9504, Cur Loss: 0.57355750, Cur Avg Loss: 0.28707045, Log Avg loss: 0.40330724, Global Avg Loss: 1.05243495, Time: 0.0208 Steps: 51690, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000604, Sample Num: 9664, Cur Loss: 0.14776438, Cur Avg Loss: 0.28541589, Log Avg loss: 0.18713526, Global Avg Loss: 1.05226758, Time: 0.0208 Steps: 51700, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000614, Sample Num: 9824, Cur Loss: 0.21794409, Cur Avg Loss: 0.28603589, Log Avg loss: 0.32348400, Global Avg Loss: 1.05212664, Time: 0.0208 Steps: 51710, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000624, Sample Num: 9984, Cur Loss: 0.10402259, Cur Avg Loss: 0.28618253, Log Avg loss: 0.29518583, Global Avg Loss: 1.05198029, Time: 0.0207 Steps: 51720, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000634, Sample Num: 10144, Cur Loss: 0.11187877, Cur Avg Loss: 0.28465141, Log Avg loss: 0.18910954, Global Avg Loss: 1.05181348, Time: 0.0208 Steps: 51730, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000644, Sample Num: 10304, Cur Loss: 0.59152091, Cur Avg Loss: 0.28409595, Log Avg loss: 0.24887998, Global Avg Loss: 1.05165830, Time: 0.0208 Steps: 51740, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000654, Sample Num: 10464, Cur Loss: 0.16630302, Cur Avg Loss: 0.28368813, Log Avg loss: 0.25742439, Global Avg Loss: 1.05150482, Time: 0.0209 Steps: 51750, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000664, Sample Num: 10624, Cur Loss: 0.26954719, Cur Avg Loss: 0.28311698, Log Avg loss: 0.24576401, Global Avg Loss: 1.05134915, Time: 0.0210 Steps: 51760, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000674, Sample Num: 10784, Cur Loss: 0.32649493, Cur Avg Loss: 0.28230610, Log Avg loss: 0.22846346, Global Avg Loss: 1.05119020, Time: 0.0208 Steps: 51770, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000684, Sample Num: 10944, Cur Loss: 0.43378013, Cur Avg Loss: 0.28136212, Log Avg loss: 0.21773762, Global Avg Loss: 1.05102924, Time: 0.0208 Steps: 51780, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000694, Sample Num: 11104, Cur Loss: 0.19042887, Cur Avg Loss: 0.28162665, Log Avg loss: 0.29972102, Global Avg Loss: 1.05088418, Time: 0.0208 Steps: 51790, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000704, Sample Num: 11264, Cur Loss: 0.13107373, Cur Avg Loss: 0.28028517, Log Avg loss: 0.18718594, Global Avg Loss: 1.05071744, Time: 0.0207 Steps: 51800, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000714, Sample Num: 11424, Cur Loss: 0.12708835, Cur Avg Loss: 0.27988081, Log Avg loss: 0.25141385, Global Avg Loss: 1.05056316, Time: 0.0208 Steps: 51810, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000724, Sample Num: 11584, Cur Loss: 0.19171044, Cur Avg Loss: 0.27869287, Log Avg loss: 0.19387423, Global Avg Loss: 1.05039784, Time: 0.0207 Steps: 51820, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000734, Sample Num: 11744, Cur Loss: 0.35819069, Cur Avg Loss: 0.27890068, Log Avg loss: 0.29394610, Global Avg Loss: 1.05025189, Time: 0.0208 Steps: 51830, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000744, Sample Num: 11904, Cur Loss: 0.30757052, Cur Avg Loss: 0.27791966, Log Avg loss: 0.20591283, Global Avg Loss: 1.05008902, Time: 0.0208 Steps: 51840, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000754, Sample Num: 12064, Cur Loss: 0.35947955, Cur Avg Loss: 0.27752025, Log Avg loss: 0.24780413, Global Avg Loss: 1.04993429, Time: 0.0208 Steps: 51850, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000764, Sample Num: 12224, Cur Loss: 0.04949003, Cur Avg Loss: 0.27826550, Log Avg loss: 0.33445737, Global Avg Loss: 1.04979633, Time: 0.0208 Steps: 51860, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000774, Sample Num: 12384, Cur Loss: 0.12655176, Cur Avg Loss: 0.27704619, Log Avg loss: 0.18389094, Global Avg Loss: 1.04962939, Time: 0.0208 Steps: 51870, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000784, Sample Num: 12544, Cur Loss: 0.17668566, Cur Avg Loss: 0.27664457, Log Avg loss: 0.24555932, Global Avg Loss: 1.04947440, Time: 0.0208 Steps: 51880, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000794, Sample Num: 12704, Cur Loss: 0.35462594, Cur Avg Loss: 0.27799807, Log Avg loss: 0.38411232, Global Avg Loss: 1.04934618, Time: 0.0208 Steps: 51890, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000804, Sample Num: 12864, Cur Loss: 0.09423070, Cur Avg Loss: 0.27816850, Log Avg loss: 0.29170060, Global Avg Loss: 1.04920019, Time: 0.0208 Steps: 51900, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000814, Sample Num: 13024, Cur Loss: 0.55713570, Cur Avg Loss: 0.27859587, Log Avg loss: 0.31295618, Global Avg Loss: 1.04905836, Time: 0.0208 Steps: 51910, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000824, Sample Num: 13184, Cur Loss: 0.08713377, Cur Avg Loss: 0.27866176, Log Avg loss: 0.28402513, Global Avg Loss: 1.04891101, Time: 0.0208 Steps: 51920, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000834, Sample Num: 13344, Cur Loss: 0.09388252, Cur Avg Loss: 0.27811928, Log Avg loss: 0.23341897, Global Avg Loss: 1.04875398, Time: 0.0208 Steps: 51930, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000844, Sample Num: 13504, Cur Loss: 0.16491260, Cur Avg Loss: 0.27690174, Log Avg loss: 0.17535896, Global Avg Loss: 1.04858582, Time: 0.0207 Steps: 51940, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000854, Sample Num: 13664, Cur Loss: 0.29170254, Cur Avg Loss: 0.27658400, Log Avg loss: 0.24976672, Global Avg Loss: 1.04843206, Time: 0.0207 Steps: 51950, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000864, Sample Num: 13824, Cur Loss: 0.41258162, Cur Avg Loss: 0.27776000, Log Avg loss: 0.37819103, Global Avg Loss: 1.04830306, Time: 0.0207 Steps: 51960, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000874, Sample Num: 13984, Cur Loss: 0.49456882, Cur Avg Loss: 0.27900946, Log Avg loss: 0.38696241, Global Avg Loss: 1.04817581, Time: 0.0207 Steps: 51970, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000884, Sample Num: 14144, Cur Loss: 0.15127476, Cur Avg Loss: 0.27838000, Log Avg loss: 0.22336513, Global Avg Loss: 1.04801713, Time: 0.0208 Steps: 51980, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000894, Sample Num: 14304, Cur Loss: 0.23206830, Cur Avg Loss: 0.27852495, Log Avg loss: 0.29133864, Global Avg Loss: 1.04787159, Time: 0.0208 Steps: 51990, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000904, Sample Num: 14464, Cur Loss: 0.53799534, Cur Avg Loss: 0.27846596, Log Avg loss: 0.27319237, Global Avg Loss: 1.04772261, Time: 0.0208 Steps: 52000, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000914, Sample Num: 14624, Cur Loss: 0.15660816, Cur Avg Loss: 0.27951413, Log Avg loss: 0.37426893, Global Avg Loss: 1.04759313, Time: 0.0208 Steps: 52010, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000924, Sample Num: 14784, Cur Loss: 0.49334320, Cur Avg Loss: 0.27921842, Log Avg loss: 0.25219029, Global Avg Loss: 1.04744022, Time: 0.0208 Steps: 52020, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000934, Sample Num: 14944, Cur Loss: 0.17911318, Cur Avg Loss: 0.27921921, Log Avg loss: 0.27929209, Global Avg Loss: 1.04729259, Time: 0.0208 Steps: 52030, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000944, Sample Num: 15104, Cur Loss: 0.12885439, Cur Avg Loss: 0.28064917, Log Avg loss: 0.41420776, Global Avg Loss: 1.04717093, Time: 0.0208 Steps: 52040, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000954, Sample Num: 15264, Cur Loss: 0.23373063, Cur Avg Loss: 0.28068216, Log Avg loss: 0.28379575, Global Avg Loss: 1.04702427, Time: 0.0208 Steps: 52050, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000964, Sample Num: 15424, Cur Loss: 0.21360189, Cur Avg Loss: 0.28215035, Log Avg loss: 0.42221560, Global Avg Loss: 1.04690425, Time: 0.0208 Steps: 52060, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000974, Sample Num: 15584, Cur Loss: 0.16657251, Cur Avg Loss: 0.28141393, Log Avg loss: 0.21042350, Global Avg Loss: 1.04674361, Time: 0.0208 Steps: 52070, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000984, Sample Num: 15744, Cur Loss: 0.24929532, Cur Avg Loss: 0.28130628, Log Avg loss: 0.27082103, Global Avg Loss: 1.04659462, Time: 0.0207 Steps: 52080, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000994, Sample Num: 15904, Cur Loss: 0.15192556, Cur Avg Loss: 0.28099353, Log Avg loss: 0.25021860, Global Avg Loss: 1.04644174, Time: 0.0208 Steps: 52090, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001004, Sample Num: 16064, Cur Loss: 0.14166884, Cur Avg Loss: 0.28008599, Log Avg loss: 0.18987711, Global Avg Loss: 1.04627733, Time: 0.0208 Steps: 52100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001014, Sample Num: 16224, Cur Loss: 0.38261583, Cur Avg Loss: 0.28094666, Log Avg loss: 0.36735799, Global Avg Loss: 1.04614704, Time: 0.0207 Steps: 52110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001024, Sample Num: 16384, Cur Loss: 0.33942685, Cur Avg Loss: 0.28134597, Log Avg loss: 0.32183551, Global Avg Loss: 1.04600807, Time: 0.0253 Steps: 52120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001034, Sample Num: 16544, Cur Loss: 0.47762814, Cur Avg Loss: 0.28081588, Log Avg loss: 0.22653468, Global Avg Loss: 1.04585088, Time: 0.0208 Steps: 52130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001044, Sample Num: 16704, Cur Loss: 0.15432143, Cur Avg Loss: 0.28088467, Log Avg loss: 0.28799799, Global Avg Loss: 1.04570553, Time: 0.0208 Steps: 52140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001054, Sample Num: 16864, Cur Loss: 0.33930868, Cur Avg Loss: 0.28054923, Log Avg loss: 0.24552932, Global Avg Loss: 1.04555209, Time: 0.0207 Steps: 52150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001064, Sample Num: 17024, Cur Loss: 0.13559079, Cur Avg Loss: 0.28018077, Log Avg loss: 0.24134459, Global Avg Loss: 1.04539791, Time: 0.0208 Steps: 52160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001074, Sample Num: 17184, Cur Loss: 0.11607724, Cur Avg Loss: 0.27984669, Log Avg loss: 0.24430114, Global Avg Loss: 1.04524435, Time: 0.0208 Steps: 52170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001084, Sample Num: 17344, Cur Loss: 0.16007084, Cur Avg Loss: 0.27904174, Log Avg loss: 0.19258926, Global Avg Loss: 1.04508095, Time: 0.0207 Steps: 52180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001094, Sample Num: 17504, Cur Loss: 0.24224824, Cur Avg Loss: 0.27868008, Log Avg loss: 0.23947698, Global Avg Loss: 1.04492659, Time: 0.0208 Steps: 52190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001104, Sample Num: 17664, Cur Loss: 0.15761515, Cur Avg Loss: 0.27766404, Log Avg loss: 0.16650944, Global Avg Loss: 1.04475831, Time: 0.0208 Steps: 52200, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001114, Sample Num: 17824, Cur Loss: 0.13359642, Cur Avg Loss: 0.27769515, Log Avg loss: 0.28112866, Global Avg Loss: 1.04461205, Time: 0.0209 Steps: 52210, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001124, Sample Num: 17984, Cur Loss: 0.17997412, Cur Avg Loss: 0.27667140, Log Avg loss: 0.16262639, Global Avg Loss: 1.04444315, Time: 0.0208 Steps: 52220, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001134, Sample Num: 18144, Cur Loss: 0.32462195, Cur Avg Loss: 0.27601080, Log Avg loss: 0.20175903, Global Avg Loss: 1.04428181, Time: 0.0208 Steps: 52230, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001144, Sample Num: 18304, Cur Loss: 0.13114944, Cur Avg Loss: 0.27574264, Log Avg loss: 0.24533340, Global Avg Loss: 1.04412887, Time: 0.0208 Steps: 52240, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001154, Sample Num: 18464, Cur Loss: 0.82389808, Cur Avg Loss: 0.27602521, Log Avg loss: 0.30835115, Global Avg Loss: 1.04398805, Time: 0.0207 Steps: 52250, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001164, Sample Num: 18624, Cur Loss: 0.07493706, Cur Avg Loss: 0.27553740, Log Avg loss: 0.21924472, Global Avg Loss: 1.04383024, Time: 0.0208 Steps: 52260, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001174, Sample Num: 18784, Cur Loss: 0.18464988, Cur Avg Loss: 0.27506024, Log Avg loss: 0.21951863, Global Avg Loss: 1.04367253, Time: 0.0208 Steps: 52270, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001184, Sample Num: 18944, Cur Loss: 0.06579995, Cur Avg Loss: 0.27576077, Log Avg loss: 0.35800318, Global Avg Loss: 1.04354138, Time: 0.0207 Steps: 52280, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001194, Sample Num: 19104, Cur Loss: 0.39113945, Cur Avg Loss: 0.27559696, Log Avg loss: 0.25620190, Global Avg Loss: 1.04339081, Time: 0.0207 Steps: 52290, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001204, Sample Num: 19264, Cur Loss: 0.34359115, Cur Avg Loss: 0.27549595, Log Avg loss: 0.26343460, Global Avg Loss: 1.04324168, Time: 0.0208 Steps: 52300, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001214, Sample Num: 19424, Cur Loss: 0.30863595, Cur Avg Loss: 0.27496365, Log Avg loss: 0.21087508, Global Avg Loss: 1.04308256, Time: 0.0208 Steps: 52310, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001224, Sample Num: 19584, Cur Loss: 0.18302378, Cur Avg Loss: 0.27498066, Log Avg loss: 0.27704583, Global Avg Loss: 1.04293614, Time: 0.0207 Steps: 52320, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001234, Sample Num: 19744, Cur Loss: 0.05899931, Cur Avg Loss: 0.27463312, Log Avg loss: 0.23209421, Global Avg Loss: 1.04278119, Time: 0.0208 Steps: 52330, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001244, Sample Num: 19904, Cur Loss: 0.62853867, Cur Avg Loss: 0.27721653, Log Avg loss: 0.59600896, Global Avg Loss: 1.04269583, Time: 0.0207 Steps: 52340, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001254, Sample Num: 20064, Cur Loss: 0.20010683, Cur Avg Loss: 0.27751450, Log Avg loss: 0.31458162, Global Avg Loss: 1.04255675, Time: 0.0208 Steps: 52350, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001264, Sample Num: 20224, Cur Loss: 0.57510471, Cur Avg Loss: 0.27923617, Log Avg loss: 0.49513408, Global Avg Loss: 1.04245220, Time: 0.0207 Steps: 52360, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001274, Sample Num: 20384, Cur Loss: 0.51896411, Cur Avg Loss: 0.28012377, Log Avg loss: 0.39231590, Global Avg Loss: 1.04232806, Time: 0.0208 Steps: 52370, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001284, Sample Num: 20544, Cur Loss: 0.33834180, Cur Avg Loss: 0.28032808, Log Avg loss: 0.30635803, Global Avg Loss: 1.04218755, Time: 0.0244 Steps: 52380, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001294, Sample Num: 20704, Cur Loss: 0.57954216, Cur Avg Loss: 0.28055257, Log Avg loss: 0.30937624, Global Avg Loss: 1.04204767, Time: 0.0207 Steps: 52390, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001304, Sample Num: 20864, Cur Loss: 0.10064209, Cur Avg Loss: 0.28081748, Log Avg loss: 0.31509667, Global Avg Loss: 1.04190894, Time: 0.0207 Steps: 52400, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001314, Sample Num: 21024, Cur Loss: 0.44530636, Cur Avg Loss: 0.28166786, Log Avg loss: 0.39255828, Global Avg Loss: 1.04178504, Time: 0.0208 Steps: 52410, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001324, Sample Num: 21184, Cur Loss: 0.31409523, Cur Avg Loss: 0.28174855, Log Avg loss: 0.29235122, Global Avg Loss: 1.04164208, Time: 0.0208 Steps: 52420, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001334, Sample Num: 21344, Cur Loss: 0.10509525, Cur Avg Loss: 0.28232657, Log Avg loss: 0.35885621, Global Avg Loss: 1.04151185, Time: 0.0207 Steps: 52430, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001344, Sample Num: 21504, Cur Loss: 0.24297193, Cur Avg Loss: 0.28165304, Log Avg loss: 0.19180423, Global Avg Loss: 1.04134981, Time: 0.0208 Steps: 52440, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001354, Sample Num: 21664, Cur Loss: 0.47701061, Cur Avg Loss: 0.28265079, Log Avg loss: 0.41674824, Global Avg Loss: 1.04123073, Time: 0.0207 Steps: 52450, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001364, Sample Num: 21824, Cur Loss: 0.46745023, Cur Avg Loss: 0.28403299, Log Avg loss: 0.47118267, Global Avg Loss: 1.04112207, Time: 0.0208 Steps: 52460, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001374, Sample Num: 21984, Cur Loss: 0.14504741, Cur Avg Loss: 0.28403586, Log Avg loss: 0.28442685, Global Avg Loss: 1.04097785, Time: 0.0208 Steps: 52470, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001384, Sample Num: 22144, Cur Loss: 0.51897568, Cur Avg Loss: 0.28376559, Log Avg loss: 0.24663060, Global Avg Loss: 1.04082649, Time: 0.0208 Steps: 52480, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001394, Sample Num: 22304, Cur Loss: 0.17746890, Cur Avg Loss: 0.28306872, Log Avg loss: 0.18662238, Global Avg Loss: 1.04066375, Time: 0.0208 Steps: 52490, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001404, Sample Num: 22464, Cur Loss: 0.14494140, Cur Avg Loss: 0.28229669, Log Avg loss: 0.17467623, Global Avg Loss: 1.04049880, Time: 0.0208 Steps: 52500, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001414, Sample Num: 22624, Cur Loss: 0.86372113, Cur Avg Loss: 0.28317597, Log Avg loss: 0.40662677, Global Avg Loss: 1.04037809, Time: 0.0208 Steps: 52510, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001424, Sample Num: 22784, Cur Loss: 0.34071305, Cur Avg Loss: 0.28335138, Log Avg loss: 0.30815388, Global Avg Loss: 1.04023867, Time: 0.0208 Steps: 52520, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001434, Sample Num: 22944, Cur Loss: 0.07992817, Cur Avg Loss: 0.28308488, Log Avg loss: 0.24513540, Global Avg Loss: 1.04008731, Time: 0.0208 Steps: 52530, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001444, Sample Num: 23104, Cur Loss: 0.37425578, Cur Avg Loss: 0.28238190, Log Avg loss: 0.18157389, Global Avg Loss: 1.03992391, Time: 0.0208 Steps: 52540, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001454, Sample Num: 23264, Cur Loss: 0.41555321, Cur Avg Loss: 0.28225537, Log Avg loss: 0.26398557, Global Avg Loss: 1.03977625, Time: 0.0208 Steps: 52550, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001464, Sample Num: 23424, Cur Loss: 0.16792616, Cur Avg Loss: 0.28132007, Log Avg loss: 0.14532632, Global Avg Loss: 1.03960607, Time: 0.0208 Steps: 52560, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001474, Sample Num: 23584, Cur Loss: 0.48206466, Cur Avg Loss: 0.28149526, Log Avg loss: 0.30714385, Global Avg Loss: 1.03946674, Time: 0.0208 Steps: 52570, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001484, Sample Num: 23744, Cur Loss: 0.16407239, Cur Avg Loss: 0.28152177, Log Avg loss: 0.28542966, Global Avg Loss: 1.03932333, Time: 0.0208 Steps: 52580, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001494, Sample Num: 23904, Cur Loss: 0.20410550, Cur Avg Loss: 0.28179870, Log Avg loss: 0.32289459, Global Avg Loss: 1.03918710, Time: 0.0207 Steps: 52590, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001504, Sample Num: 24064, Cur Loss: 0.24599385, Cur Avg Loss: 0.28193814, Log Avg loss: 0.30277073, Global Avg Loss: 1.03904710, Time: 0.0208 Steps: 52600, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001514, Sample Num: 24224, Cur Loss: 0.03830266, Cur Avg Loss: 0.28134650, Log Avg loss: 0.19236305, Global Avg Loss: 1.03888617, Time: 0.0209 Steps: 52610, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001524, Sample Num: 24384, Cur Loss: 0.37129480, Cur Avg Loss: 0.28082458, Log Avg loss: 0.20180659, Global Avg Loss: 1.03872709, Time: 0.0208 Steps: 52620, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001534, Sample Num: 24544, Cur Loss: 0.11084059, Cur Avg Loss: 0.28055008, Log Avg loss: 0.23871563, Global Avg Loss: 1.03857508, Time: 0.0208 Steps: 52630, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001544, Sample Num: 24704, Cur Loss: 0.39069232, Cur Avg Loss: 0.28013874, Log Avg loss: 0.21703958, Global Avg Loss: 1.03841901, Time: 0.0208 Steps: 52640, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001554, Sample Num: 24864, Cur Loss: 0.15494932, Cur Avg Loss: 0.27989153, Log Avg loss: 0.24172181, Global Avg Loss: 1.03826769, Time: 0.0208 Steps: 52650, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001564, Sample Num: 25024, Cur Loss: 0.07832809, Cur Avg Loss: 0.27939125, Log Avg loss: 0.20164816, Global Avg Loss: 1.03810882, Time: 0.0208 Steps: 52660, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001574, Sample Num: 25184, Cur Loss: 0.10221608, Cur Avg Loss: 0.27902855, Log Avg loss: 0.22230196, Global Avg Loss: 1.03795393, Time: 0.0208 Steps: 52670, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001584, Sample Num: 25344, Cur Loss: 0.46312928, Cur Avg Loss: 0.27933148, Log Avg loss: 0.32701381, Global Avg Loss: 1.03781898, Time: 0.0207 Steps: 52680, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001594, Sample Num: 25504, Cur Loss: 0.05876851, Cur Avg Loss: 0.27915973, Log Avg loss: 0.25195392, Global Avg Loss: 1.03766983, Time: 0.0208 Steps: 52690, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001604, Sample Num: 25664, Cur Loss: 0.29799339, Cur Avg Loss: 0.27882994, Log Avg loss: 0.22626203, Global Avg Loss: 1.03751586, Time: 0.0208 Steps: 52700, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001614, Sample Num: 25824, Cur Loss: 0.41374195, Cur Avg Loss: 0.27865445, Log Avg loss: 0.25050484, Global Avg Loss: 1.03736655, Time: 0.0209 Steps: 52710, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001624, Sample Num: 25984, Cur Loss: 0.31787270, Cur Avg Loss: 0.27847739, Log Avg loss: 0.24990031, Global Avg Loss: 1.03721718, Time: 0.0209 Steps: 52720, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001634, Sample Num: 26144, Cur Loss: 0.11662416, Cur Avg Loss: 0.27830223, Log Avg loss: 0.24985649, Global Avg Loss: 1.03706786, Time: 0.0208 Steps: 52730, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001644, Sample Num: 26304, Cur Loss: 0.24516624, Cur Avg Loss: 0.27803024, Log Avg loss: 0.23358749, Global Avg Loss: 1.03691552, Time: 0.0208 Steps: 52740, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001654, Sample Num: 26464, Cur Loss: 0.18859091, Cur Avg Loss: 0.27753072, Log Avg loss: 0.19540859, Global Avg Loss: 1.03675599, Time: 0.0207 Steps: 52750, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001664, Sample Num: 26624, Cur Loss: 0.11412435, Cur Avg Loss: 0.27740209, Log Avg loss: 0.25612760, Global Avg Loss: 1.03660803, Time: 0.0208 Steps: 52760, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001674, Sample Num: 26784, Cur Loss: 0.38218409, Cur Avg Loss: 0.27782087, Log Avg loss: 0.34750537, Global Avg Loss: 1.03647744, Time: 0.0209 Steps: 52770, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001684, Sample Num: 26944, Cur Loss: 0.20470355, Cur Avg Loss: 0.27763431, Log Avg loss: 0.24640379, Global Avg Loss: 1.03632775, Time: 0.0207 Steps: 52780, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001694, Sample Num: 27104, Cur Loss: 0.20585111, Cur Avg Loss: 0.27717983, Log Avg loss: 0.20064654, Global Avg Loss: 1.03616945, Time: 0.0209 Steps: 52790, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001704, Sample Num: 27264, Cur Loss: 0.31778041, Cur Avg Loss: 0.27699243, Log Avg loss: 0.24524593, Global Avg Loss: 1.03601965, Time: 0.0208 Steps: 52800, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001714, Sample Num: 27424, Cur Loss: 0.56323761, Cur Avg Loss: 0.27684402, Log Avg loss: 0.25155548, Global Avg Loss: 1.03587111, Time: 0.0207 Steps: 52810, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001724, Sample Num: 27584, Cur Loss: 0.08349273, Cur Avg Loss: 0.27661671, Log Avg loss: 0.23765548, Global Avg Loss: 1.03571999, Time: 0.0208 Steps: 52820, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001734, Sample Num: 27744, Cur Loss: 0.08001122, Cur Avg Loss: 0.27633617, Log Avg loss: 0.22797175, Global Avg Loss: 1.03556709, Time: 0.0207 Steps: 52830, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001744, Sample Num: 27904, Cur Loss: 0.05861081, Cur Avg Loss: 0.27630503, Log Avg loss: 0.27090531, Global Avg Loss: 1.03542238, Time: 0.0207 Steps: 52840, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001754, Sample Num: 28064, Cur Loss: 0.25694263, Cur Avg Loss: 0.27643596, Log Avg loss: 0.29927010, Global Avg Loss: 1.03528309, Time: 0.0207 Steps: 52850, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001764, Sample Num: 28224, Cur Loss: 0.07724969, Cur Avg Loss: 0.27718349, Log Avg loss: 0.40829955, Global Avg Loss: 1.03516448, Time: 0.0208 Steps: 52860, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001774, Sample Num: 28384, Cur Loss: 0.23317483, Cur Avg Loss: 0.27746271, Log Avg loss: 0.32671646, Global Avg Loss: 1.03503048, Time: 0.0207 Steps: 52870, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001784, Sample Num: 28544, Cur Loss: 0.47105384, Cur Avg Loss: 0.27731657, Log Avg loss: 0.25139122, Global Avg Loss: 1.03488229, Time: 0.0208 Steps: 52880, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001794, Sample Num: 28704, Cur Loss: 0.70346439, Cur Avg Loss: 0.27818514, Log Avg loss: 0.43313842, Global Avg Loss: 1.03476851, Time: 0.0247 Steps: 52890, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001804, Sample Num: 28864, Cur Loss: 0.66484058, Cur Avg Loss: 0.27862134, Log Avg loss: 0.35687543, Global Avg Loss: 1.03464037, Time: 0.0208 Steps: 52900, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001814, Sample Num: 29024, Cur Loss: 0.13769315, Cur Avg Loss: 0.27879437, Log Avg loss: 0.31000946, Global Avg Loss: 1.03450341, Time: 0.0208 Steps: 52910, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001824, Sample Num: 29184, Cur Loss: 0.19937384, Cur Avg Loss: 0.27882538, Log Avg loss: 0.28445088, Global Avg Loss: 1.03436168, Time: 0.0208 Steps: 52920, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001834, Sample Num: 29344, Cur Loss: 0.80669332, Cur Avg Loss: 0.27931875, Log Avg loss: 0.36930895, Global Avg Loss: 1.03423603, Time: 0.0208 Steps: 52930, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001844, Sample Num: 29504, Cur Loss: 0.62325418, Cur Avg Loss: 0.27962472, Log Avg loss: 0.33574062, Global Avg Loss: 1.03410409, Time: 0.0208 Steps: 52940, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001854, Sample Num: 29664, Cur Loss: 0.30034149, Cur Avg Loss: 0.27922047, Log Avg loss: 0.20467533, Global Avg Loss: 1.03394745, Time: 0.0208 Steps: 52950, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001864, Sample Num: 29824, Cur Loss: 0.37630928, Cur Avg Loss: 0.27913511, Log Avg loss: 0.26331113, Global Avg Loss: 1.03380193, Time: 0.0208 Steps: 52960, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001874, Sample Num: 29984, Cur Loss: 0.17812854, Cur Avg Loss: 0.27969531, Log Avg loss: 0.38411515, Global Avg Loss: 1.03367928, Time: 0.0208 Steps: 52970, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001884, Sample Num: 30144, Cur Loss: 0.37207919, Cur Avg Loss: 0.28033530, Log Avg loss: 0.40027080, Global Avg Loss: 1.03355973, Time: 0.0208 Steps: 52980, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001894, Sample Num: 30304, Cur Loss: 0.37436724, Cur Avg Loss: 0.28031580, Log Avg loss: 0.27664087, Global Avg Loss: 1.03341688, Time: 0.0208 Steps: 52990, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001904, Sample Num: 30464, Cur Loss: 0.24328257, Cur Avg Loss: 0.28016068, Log Avg loss: 0.25078054, Global Avg Loss: 1.03326922, Time: 0.0208 Steps: 53000, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001914, Sample Num: 30624, Cur Loss: 0.03458142, Cur Avg Loss: 0.27963413, Log Avg loss: 0.17937893, Global Avg Loss: 1.03310814, Time: 0.0207 Steps: 53010, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001924, Sample Num: 30784, Cur Loss: 0.32157147, Cur Avg Loss: 0.27956768, Log Avg loss: 0.26684968, Global Avg Loss: 1.03296361, Time: 0.0208 Steps: 53020, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001934, Sample Num: 30944, Cur Loss: 0.27818969, Cur Avg Loss: 0.27987513, Log Avg loss: 0.33902967, Global Avg Loss: 1.03283276, Time: 0.0208 Steps: 53030, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001944, Sample Num: 31104, Cur Loss: 0.44416711, Cur Avg Loss: 0.28010278, Log Avg loss: 0.32413004, Global Avg Loss: 1.03269914, Time: 0.0208 Steps: 53040, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001954, Sample Num: 31264, Cur Loss: 0.51190841, Cur Avg Loss: 0.27990486, Log Avg loss: 0.24142775, Global Avg Loss: 1.03254998, Time: 0.0208 Steps: 53050, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001964, Sample Num: 31424, Cur Loss: 0.20080572, Cur Avg Loss: 0.27976990, Log Avg loss: 0.25340021, Global Avg Loss: 1.03240314, Time: 0.0208 Steps: 53060, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001974, Sample Num: 31584, Cur Loss: 0.17008239, Cur Avg Loss: 0.27945780, Log Avg loss: 0.21816099, Global Avg Loss: 1.03224971, Time: 0.0208 Steps: 53070, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001984, Sample Num: 31744, Cur Loss: 0.20923823, Cur Avg Loss: 0.27917528, Log Avg loss: 0.22340559, Global Avg Loss: 1.03209733, Time: 0.0208 Steps: 53080, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001994, Sample Num: 31904, Cur Loss: 0.10001891, Cur Avg Loss: 0.27871599, Log Avg loss: 0.18759289, Global Avg Loss: 1.03193826, Time: 0.0208 Steps: 53090, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002004, Sample Num: 32064, Cur Loss: 0.29353419, Cur Avg Loss: 0.27878731, Log Avg loss: 0.29300835, Global Avg Loss: 1.03179910, Time: 0.0208 Steps: 53100, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002014, Sample Num: 32224, Cur Loss: 0.30536410, Cur Avg Loss: 0.27888572, Log Avg loss: 0.29860642, Global Avg Loss: 1.03166105, Time: 0.0207 Steps: 53110, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002024, Sample Num: 32384, Cur Loss: 0.07013134, Cur Avg Loss: 0.27839302, Log Avg loss: 0.17916451, Global Avg Loss: 1.03150057, Time: 0.0208 Steps: 53120, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002034, Sample Num: 32544, Cur Loss: 0.47665238, Cur Avg Loss: 0.27843802, Log Avg loss: 0.28754495, Global Avg Loss: 1.03136054, Time: 0.0208 Steps: 53130, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002044, Sample Num: 32704, Cur Loss: 0.32267249, Cur Avg Loss: 0.27847046, Log Avg loss: 0.28506923, Global Avg Loss: 1.03122010, Time: 0.0208 Steps: 53140, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002054, Sample Num: 32864, Cur Loss: 0.05706085, Cur Avg Loss: 0.27832117, Log Avg loss: 0.24780657, Global Avg Loss: 1.03107270, Time: 0.0246 Steps: 53150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002064, Sample Num: 33024, Cur Loss: 0.08270396, Cur Avg Loss: 0.27822524, Log Avg loss: 0.25852188, Global Avg Loss: 1.03092738, Time: 0.0209 Steps: 53160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002074, Sample Num: 33184, Cur Loss: 0.14626838, Cur Avg Loss: 0.27816374, Log Avg loss: 0.26546916, Global Avg Loss: 1.03078342, Time: 0.0208 Steps: 53170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002084, Sample Num: 33344, Cur Loss: 0.06867128, Cur Avg Loss: 0.27802871, Log Avg loss: 0.25002339, Global Avg Loss: 1.03063660, Time: 0.0209 Steps: 53180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002094, Sample Num: 33504, Cur Loss: 0.47895870, Cur Avg Loss: 0.27865026, Log Avg loss: 0.40818103, Global Avg Loss: 1.03051958, Time: 0.0210 Steps: 53190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002104, Sample Num: 33664, Cur Loss: 0.26552245, Cur Avg Loss: 0.27887381, Log Avg loss: 0.32568644, Global Avg Loss: 1.03038709, Time: 0.0209 Steps: 53200, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002114, Sample Num: 33824, Cur Loss: 0.20444301, Cur Avg Loss: 0.27859689, Log Avg loss: 0.22033222, Global Avg Loss: 1.03023485, Time: 0.0209 Steps: 53210, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002124, Sample Num: 33984, Cur Loss: 0.07996951, Cur Avg Loss: 0.27862168, Log Avg loss: 0.28386155, Global Avg Loss: 1.03009461, Time: 0.0209 Steps: 53220, Updated lr: 0.000050 ***** Running evaluation checkpoint-53225 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-53225 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.594689, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.388928, "eval_total_loss": 273.416616, "eval_mae": 0.515225, "eval_mse": 0.388961, "eval_r2": 0.752751, "eval_sp_statistic": 0.887074, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.913702, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.333291, "test_total_loss": 167.312159, "test_mae": 0.377322, "test_mse": 0.333412, "test_r2": 0.784813, "test_sp_statistic": 0.866436, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.914939, "test_ps_pvalue": 0.0, "lr": 5.0474158368895205e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0300116273686954, "train_cur_epoch_loss": 592.5262866765261, "train_cur_epoch_avg_loss": 0.27831201816652235, "train_cur_epoch_time": 44.5946888923645, "train_cur_epoch_avg_time": 0.020946307605619774, "epoch": 25, "step": 53225} ################################################## Training, Epoch: 0026, Batch: 000005, Sample Num: 80, Cur Loss: 0.56858063, Cur Avg Loss: 0.31320615, Log Avg loss: 0.22998751, Global Avg Loss: 1.02994430, Time: 0.0257 Steps: 53230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000015, Sample Num: 240, Cur Loss: 0.13330930, Cur Avg Loss: 0.25860161, Log Avg loss: 0.23129933, Global Avg Loss: 1.02979429, Time: 0.0220 Steps: 53240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000025, Sample Num: 400, Cur Loss: 0.19575961, Cur Avg Loss: 0.29801609, Log Avg loss: 0.35713782, Global Avg Loss: 1.02966797, Time: 0.0219 Steps: 53250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000035, Sample Num: 560, Cur Loss: 0.27999377, Cur Avg Loss: 0.29782224, Log Avg loss: 0.29733760, Global Avg Loss: 1.02953047, Time: 0.0219 Steps: 53260, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000045, Sample Num: 720, Cur Loss: 0.18977782, Cur Avg Loss: 0.29152676, Log Avg loss: 0.26949260, Global Avg Loss: 1.02938779, Time: 0.0220 Steps: 53270, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000055, Sample Num: 880, Cur Loss: 0.63831079, Cur Avg Loss: 0.31638054, Log Avg loss: 0.42822253, Global Avg Loss: 1.02927496, Time: 0.0219 Steps: 53280, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000065, Sample Num: 1040, Cur Loss: 0.24665818, Cur Avg Loss: 0.31049241, Log Avg loss: 0.27810768, Global Avg Loss: 1.02913400, Time: 0.0219 Steps: 53290, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000075, Sample Num: 1200, Cur Loss: 0.04210826, Cur Avg Loss: 0.29711468, Log Avg loss: 0.21015948, Global Avg Loss: 1.02898035, Time: 0.0219 Steps: 53300, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000085, Sample Num: 1360, Cur Loss: 0.60400909, Cur Avg Loss: 0.30087621, Log Avg loss: 0.32908768, Global Avg Loss: 1.02884906, Time: 0.0209 Steps: 53310, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000095, Sample Num: 1520, Cur Loss: 0.14909777, Cur Avg Loss: 0.29376709, Log Avg loss: 0.23333958, Global Avg Loss: 1.02869986, Time: 0.0208 Steps: 53320, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000105, Sample Num: 1680, Cur Loss: 0.42363450, Cur Avg Loss: 0.28586752, Log Avg loss: 0.21082154, Global Avg Loss: 1.02854650, Time: 0.0208 Steps: 53330, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000115, Sample Num: 1840, Cur Loss: 0.37284601, Cur Avg Loss: 0.29206084, Log Avg loss: 0.35709078, Global Avg Loss: 1.02842062, Time: 0.0209 Steps: 53340, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000125, Sample Num: 2000, Cur Loss: 0.23142962, Cur Avg Loss: 0.28296587, Log Avg loss: 0.17837366, Global Avg Loss: 1.02826129, Time: 0.0209 Steps: 53350, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000135, Sample Num: 2160, Cur Loss: 0.08823669, Cur Avg Loss: 0.27981651, Log Avg loss: 0.24044951, Global Avg Loss: 1.02811364, Time: 0.0208 Steps: 53360, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000145, Sample Num: 2320, Cur Loss: 0.34523052, Cur Avg Loss: 0.28158048, Log Avg loss: 0.30539402, Global Avg Loss: 1.02797823, Time: 0.0208 Steps: 53370, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000155, Sample Num: 2480, Cur Loss: 0.06955948, Cur Avg Loss: 0.28215160, Log Avg loss: 0.29043288, Global Avg Loss: 1.02784006, Time: 0.0209 Steps: 53380, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000165, Sample Num: 2640, Cur Loss: 0.32709235, Cur Avg Loss: 0.28808246, Log Avg loss: 0.38001083, Global Avg Loss: 1.02771872, Time: 0.0208 Steps: 53390, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000175, Sample Num: 2800, Cur Loss: 0.21184413, Cur Avg Loss: 0.28073145, Log Avg loss: 0.15943973, Global Avg Loss: 1.02755612, Time: 0.0208 Steps: 53400, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000185, Sample Num: 2960, Cur Loss: 0.22324494, Cur Avg Loss: 0.27850521, Log Avg loss: 0.23954612, Global Avg Loss: 1.02740858, Time: 0.0208 Steps: 53410, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000195, Sample Num: 3120, Cur Loss: 0.20863411, Cur Avg Loss: 0.27696809, Log Avg loss: 0.24853133, Global Avg Loss: 1.02726278, Time: 0.0209 Steps: 53420, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000205, Sample Num: 3280, Cur Loss: 0.18135720, Cur Avg Loss: 0.27761806, Log Avg loss: 0.29029240, Global Avg Loss: 1.02712485, Time: 0.0209 Steps: 53430, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000215, Sample Num: 3440, Cur Loss: 0.14231443, Cur Avg Loss: 0.27905969, Log Avg loss: 0.30861316, Global Avg Loss: 1.02699039, Time: 0.0208 Steps: 53440, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000225, Sample Num: 3600, Cur Loss: 0.72888196, Cur Avg Loss: 0.28264480, Log Avg loss: 0.35972455, Global Avg Loss: 1.02686556, Time: 0.0209 Steps: 53450, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000235, Sample Num: 3760, Cur Loss: 0.16037729, Cur Avg Loss: 0.28017227, Log Avg loss: 0.22454054, Global Avg Loss: 1.02671548, Time: 0.0208 Steps: 53460, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000245, Sample Num: 3920, Cur Loss: 0.44581091, Cur Avg Loss: 0.27940176, Log Avg loss: 0.26129460, Global Avg Loss: 1.02657233, Time: 0.0209 Steps: 53470, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000255, Sample Num: 4080, Cur Loss: 0.24281734, Cur Avg Loss: 0.27475081, Log Avg loss: 0.16080265, Global Avg Loss: 1.02641044, Time: 0.0208 Steps: 53480, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000265, Sample Num: 4240, Cur Loss: 0.51375103, Cur Avg Loss: 0.27436559, Log Avg loss: 0.26454245, Global Avg Loss: 1.02626801, Time: 0.0209 Steps: 53490, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000275, Sample Num: 4400, Cur Loss: 0.16537064, Cur Avg Loss: 0.27057121, Log Avg loss: 0.17002016, Global Avg Loss: 1.02610796, Time: 0.0209 Steps: 53500, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000285, Sample Num: 4560, Cur Loss: 0.58321679, Cur Avg Loss: 0.27016455, Log Avg loss: 0.25898133, Global Avg Loss: 1.02596460, Time: 0.0209 Steps: 53510, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000295, Sample Num: 4720, Cur Loss: 0.50306654, Cur Avg Loss: 0.27436940, Log Avg loss: 0.39420759, Global Avg Loss: 1.02584656, Time: 0.0208 Steps: 53520, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000305, Sample Num: 4880, Cur Loss: 0.15584809, Cur Avg Loss: 0.27358855, Log Avg loss: 0.25055350, Global Avg Loss: 1.02570173, Time: 0.0208 Steps: 53530, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000315, Sample Num: 5040, Cur Loss: 0.23844600, Cur Avg Loss: 0.27395338, Log Avg loss: 0.28508085, Global Avg Loss: 1.02556340, Time: 0.0209 Steps: 53540, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000325, Sample Num: 5200, Cur Loss: 0.29185560, Cur Avg Loss: 0.27228971, Log Avg loss: 0.21988405, Global Avg Loss: 1.02541294, Time: 0.0209 Steps: 53550, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000335, Sample Num: 5360, Cur Loss: 0.28033245, Cur Avg Loss: 0.27191109, Log Avg loss: 0.25960585, Global Avg Loss: 1.02526996, Time: 0.0209 Steps: 53560, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000345, Sample Num: 5520, Cur Loss: 0.26446277, Cur Avg Loss: 0.27038875, Log Avg loss: 0.21939036, Global Avg Loss: 1.02511953, Time: 0.0209 Steps: 53570, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000355, Sample Num: 5680, Cur Loss: 0.09038591, Cur Avg Loss: 0.26756498, Log Avg loss: 0.17014488, Global Avg Loss: 1.02495996, Time: 0.0209 Steps: 53580, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000365, Sample Num: 5840, Cur Loss: 0.20957702, Cur Avg Loss: 0.26752411, Log Avg loss: 0.26607323, Global Avg Loss: 1.02481835, Time: 0.0209 Steps: 53590, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000375, Sample Num: 6000, Cur Loss: 0.26307330, Cur Avg Loss: 0.26851419, Log Avg loss: 0.30465237, Global Avg Loss: 1.02468399, Time: 0.0209 Steps: 53600, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000385, Sample Num: 6160, Cur Loss: 0.17082217, Cur Avg Loss: 0.26662578, Log Avg loss: 0.19581041, Global Avg Loss: 1.02452938, Time: 0.0209 Steps: 53610, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000395, Sample Num: 6320, Cur Loss: 0.15632252, Cur Avg Loss: 0.26603567, Log Avg loss: 0.24331619, Global Avg Loss: 1.02438368, Time: 0.0209 Steps: 53620, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000405, Sample Num: 6480, Cur Loss: 0.30402318, Cur Avg Loss: 0.26546346, Log Avg loss: 0.24286123, Global Avg Loss: 1.02423796, Time: 0.0209 Steps: 53630, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000415, Sample Num: 6640, Cur Loss: 0.69200814, Cur Avg Loss: 0.26633230, Log Avg loss: 0.30152052, Global Avg Loss: 1.02410322, Time: 0.0208 Steps: 53640, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000425, Sample Num: 6800, Cur Loss: 0.14185140, Cur Avg Loss: 0.27472061, Log Avg loss: 0.62283518, Global Avg Loss: 1.02402843, Time: 0.0209 Steps: 53650, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000435, Sample Num: 6960, Cur Loss: 0.42709595, Cur Avg Loss: 0.28040674, Log Avg loss: 0.52206738, Global Avg Loss: 1.02393488, Time: 0.0209 Steps: 53660, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000445, Sample Num: 7120, Cur Loss: 0.13591564, Cur Avg Loss: 0.28098493, Log Avg loss: 0.30613607, Global Avg Loss: 1.02380114, Time: 0.0208 Steps: 53670, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000455, Sample Num: 7280, Cur Loss: 0.29014388, Cur Avg Loss: 0.28150152, Log Avg loss: 0.30448982, Global Avg Loss: 1.02366714, Time: 0.0209 Steps: 53680, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000465, Sample Num: 7440, Cur Loss: 0.12591869, Cur Avg Loss: 0.28110079, Log Avg loss: 0.26286771, Global Avg Loss: 1.02352544, Time: 0.0209 Steps: 53690, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000475, Sample Num: 7600, Cur Loss: 0.41938189, Cur Avg Loss: 0.28101012, Log Avg loss: 0.27679379, Global Avg Loss: 1.02338638, Time: 0.0209 Steps: 53700, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000485, Sample Num: 7760, Cur Loss: 0.23631419, Cur Avg Loss: 0.27978542, Log Avg loss: 0.22161212, Global Avg Loss: 1.02323710, Time: 0.0209 Steps: 53710, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000495, Sample Num: 7920, Cur Loss: 0.22300757, Cur Avg Loss: 0.27932800, Log Avg loss: 0.25714329, Global Avg Loss: 1.02309449, Time: 0.0209 Steps: 53720, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000505, Sample Num: 8080, Cur Loss: 0.27649701, Cur Avg Loss: 0.27846047, Log Avg loss: 0.23551785, Global Avg Loss: 1.02294791, Time: 0.0209 Steps: 53730, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000515, Sample Num: 8240, Cur Loss: 0.17586274, Cur Avg Loss: 0.27835660, Log Avg loss: 0.27311103, Global Avg Loss: 1.02280838, Time: 0.0245 Steps: 53740, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000525, Sample Num: 8400, Cur Loss: 0.50202727, Cur Avg Loss: 0.27763197, Log Avg loss: 0.24031369, Global Avg Loss: 1.02266280, Time: 0.0208 Steps: 53750, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000535, Sample Num: 8560, Cur Loss: 0.31146392, Cur Avg Loss: 0.27733328, Log Avg loss: 0.26165183, Global Avg Loss: 1.02252125, Time: 0.0208 Steps: 53760, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000545, Sample Num: 8720, Cur Loss: 0.21170098, Cur Avg Loss: 0.27621151, Log Avg loss: 0.21619677, Global Avg Loss: 1.02237129, Time: 0.0207 Steps: 53770, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000555, Sample Num: 8880, Cur Loss: 0.46828067, Cur Avg Loss: 0.27551567, Log Avg loss: 0.23759274, Global Avg Loss: 1.02222536, Time: 0.0208 Steps: 53780, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000565, Sample Num: 9040, Cur Loss: 0.11226071, Cur Avg Loss: 0.27395521, Log Avg loss: 0.18734969, Global Avg Loss: 1.02207015, Time: 0.0208 Steps: 53790, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000575, Sample Num: 9200, Cur Loss: 0.45649949, Cur Avg Loss: 0.27350294, Log Avg loss: 0.24794973, Global Avg Loss: 1.02192627, Time: 0.0207 Steps: 53800, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000585, Sample Num: 9360, Cur Loss: 0.14850886, Cur Avg Loss: 0.27144386, Log Avg loss: 0.15304629, Global Avg Loss: 1.02176479, Time: 0.0208 Steps: 53810, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000595, Sample Num: 9520, Cur Loss: 0.27485397, Cur Avg Loss: 0.27157810, Log Avg loss: 0.27943108, Global Avg Loss: 1.02162686, Time: 0.0208 Steps: 53820, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000605, Sample Num: 9680, Cur Loss: 0.13281687, Cur Avg Loss: 0.27151834, Log Avg loss: 0.26796316, Global Avg Loss: 1.02148686, Time: 0.0208 Steps: 53830, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000615, Sample Num: 9840, Cur Loss: 0.13363317, Cur Avg Loss: 0.27092631, Log Avg loss: 0.23510849, Global Avg Loss: 1.02134080, Time: 0.0208 Steps: 53840, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000625, Sample Num: 10000, Cur Loss: 0.25843874, Cur Avg Loss: 0.27191874, Log Avg loss: 0.33295298, Global Avg Loss: 1.02121296, Time: 0.0207 Steps: 53850, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000635, Sample Num: 10160, Cur Loss: 0.36396858, Cur Avg Loss: 0.27145477, Log Avg loss: 0.24245681, Global Avg Loss: 1.02106837, Time: 0.0208 Steps: 53860, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000645, Sample Num: 10320, Cur Loss: 0.19776875, Cur Avg Loss: 0.27043333, Log Avg loss: 0.20557192, Global Avg Loss: 1.02091699, Time: 0.0207 Steps: 53870, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000655, Sample Num: 10480, Cur Loss: 0.03411585, Cur Avg Loss: 0.26972242, Log Avg loss: 0.22386872, Global Avg Loss: 1.02076906, Time: 0.0207 Steps: 53880, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000665, Sample Num: 10640, Cur Loss: 0.28630674, Cur Avg Loss: 0.26941288, Log Avg loss: 0.24913752, Global Avg Loss: 1.02062588, Time: 0.0208 Steps: 53890, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000675, Sample Num: 10800, Cur Loss: 0.30772820, Cur Avg Loss: 0.26872763, Log Avg loss: 0.22315889, Global Avg Loss: 1.02047792, Time: 0.0207 Steps: 53900, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000685, Sample Num: 10960, Cur Loss: 0.14066152, Cur Avg Loss: 0.26800143, Log Avg loss: 0.21898250, Global Avg Loss: 1.02032925, Time: 0.0208 Steps: 53910, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000695, Sample Num: 11120, Cur Loss: 0.60835159, Cur Avg Loss: 0.26931925, Log Avg loss: 0.35958984, Global Avg Loss: 1.02020671, Time: 0.0208 Steps: 53920, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000705, Sample Num: 11280, Cur Loss: 0.20301574, Cur Avg Loss: 0.26862375, Log Avg loss: 0.22028660, Global Avg Loss: 1.02005838, Time: 0.0209 Steps: 53930, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000715, Sample Num: 11440, Cur Loss: 0.08382376, Cur Avg Loss: 0.26723486, Log Avg loss: 0.16931867, Global Avg Loss: 1.01990066, Time: 0.0208 Steps: 53940, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000725, Sample Num: 11600, Cur Loss: 0.15568702, Cur Avg Loss: 0.26569524, Log Avg loss: 0.15561246, Global Avg Loss: 1.01974046, Time: 0.0208 Steps: 53950, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000735, Sample Num: 11760, Cur Loss: 0.33189648, Cur Avg Loss: 0.26631030, Log Avg loss: 0.31090213, Global Avg Loss: 1.01960910, Time: 0.0208 Steps: 53960, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000745, Sample Num: 11920, Cur Loss: 0.25189960, Cur Avg Loss: 0.26571648, Log Avg loss: 0.22207068, Global Avg Loss: 1.01946132, Time: 0.0208 Steps: 53970, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000755, Sample Num: 12080, Cur Loss: 0.30983332, Cur Avg Loss: 0.26612532, Log Avg loss: 0.29658342, Global Avg Loss: 1.01932741, Time: 0.0208 Steps: 53980, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000765, Sample Num: 12240, Cur Loss: 0.14521220, Cur Avg Loss: 0.26518677, Log Avg loss: 0.19432628, Global Avg Loss: 1.01917460, Time: 0.0208 Steps: 53990, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000775, Sample Num: 12400, Cur Loss: 0.13372323, Cur Avg Loss: 0.26448759, Log Avg loss: 0.21100039, Global Avg Loss: 1.01902494, Time: 0.0208 Steps: 54000, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000785, Sample Num: 12560, Cur Loss: 0.14408560, Cur Avg Loss: 0.26346055, Log Avg loss: 0.18386515, Global Avg Loss: 1.01887031, Time: 0.0207 Steps: 54010, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000795, Sample Num: 12720, Cur Loss: 0.14258492, Cur Avg Loss: 0.26355510, Log Avg loss: 0.27097696, Global Avg Loss: 1.01873186, Time: 0.0208 Steps: 54020, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000805, Sample Num: 12880, Cur Loss: 0.19187789, Cur Avg Loss: 0.26358993, Log Avg loss: 0.26635942, Global Avg Loss: 1.01859261, Time: 0.0209 Steps: 54030, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000815, Sample Num: 13040, Cur Loss: 0.51167053, Cur Avg Loss: 0.26792503, Log Avg loss: 0.61690039, Global Avg Loss: 1.01851828, Time: 0.0208 Steps: 54040, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000825, Sample Num: 13200, Cur Loss: 0.16229773, Cur Avg Loss: 0.26921108, Log Avg loss: 0.37402381, Global Avg Loss: 1.01839904, Time: 0.0208 Steps: 54050, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000835, Sample Num: 13360, Cur Loss: 0.19104965, Cur Avg Loss: 0.26905874, Log Avg loss: 0.25649109, Global Avg Loss: 1.01825810, Time: 0.0208 Steps: 54060, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000845, Sample Num: 13520, Cur Loss: 0.16747987, Cur Avg Loss: 0.26840154, Log Avg loss: 0.21352495, Global Avg Loss: 1.01810927, Time: 0.0207 Steps: 54070, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000855, Sample Num: 13680, Cur Loss: 0.30274165, Cur Avg Loss: 0.26848962, Log Avg loss: 0.27593233, Global Avg Loss: 1.01797203, Time: 0.0207 Steps: 54080, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000865, Sample Num: 13840, Cur Loss: 0.15515178, Cur Avg Loss: 0.26831639, Log Avg loss: 0.25350581, Global Avg Loss: 1.01783070, Time: 0.0208 Steps: 54090, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000875, Sample Num: 14000, Cur Loss: 0.31990331, Cur Avg Loss: 0.26845264, Log Avg loss: 0.28023823, Global Avg Loss: 1.01769436, Time: 0.0208 Steps: 54100, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000885, Sample Num: 14160, Cur Loss: 0.20608912, Cur Avg Loss: 0.26761961, Log Avg loss: 0.19472887, Global Avg Loss: 1.01754227, Time: 0.0207 Steps: 54110, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000895, Sample Num: 14320, Cur Loss: 0.65402949, Cur Avg Loss: 0.26686251, Log Avg loss: 0.19985927, Global Avg Loss: 1.01739118, Time: 0.0207 Steps: 54120, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000905, Sample Num: 14480, Cur Loss: 0.08477601, Cur Avg Loss: 0.26644508, Log Avg loss: 0.22908517, Global Avg Loss: 1.01724555, Time: 0.0207 Steps: 54130, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000915, Sample Num: 14640, Cur Loss: 0.14898795, Cur Avg Loss: 0.26683004, Log Avg loss: 0.30166871, Global Avg Loss: 1.01711338, Time: 0.0208 Steps: 54140, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000925, Sample Num: 14800, Cur Loss: 0.26744956, Cur Avg Loss: 0.26714587, Log Avg loss: 0.29604482, Global Avg Loss: 1.01698022, Time: 0.0207 Steps: 54150, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000935, Sample Num: 14960, Cur Loss: 0.11790510, Cur Avg Loss: 0.26699469, Log Avg loss: 0.25301000, Global Avg Loss: 1.01683916, Time: 0.0207 Steps: 54160, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000945, Sample Num: 15120, Cur Loss: 0.07856075, Cur Avg Loss: 0.26683238, Log Avg loss: 0.25165649, Global Avg Loss: 1.01669790, Time: 0.0207 Steps: 54170, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000955, Sample Num: 15280, Cur Loss: 0.22258008, Cur Avg Loss: 0.26613407, Log Avg loss: 0.20014387, Global Avg Loss: 1.01654719, Time: 0.0208 Steps: 54180, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000965, Sample Num: 15440, Cur Loss: 0.34645146, Cur Avg Loss: 0.26525488, Log Avg loss: 0.18129275, Global Avg Loss: 1.01639306, Time: 0.0208 Steps: 54190, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000975, Sample Num: 15600, Cur Loss: 0.40091252, Cur Avg Loss: 0.26532379, Log Avg loss: 0.27197317, Global Avg Loss: 1.01625571, Time: 0.0208 Steps: 54200, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000985, Sample Num: 15760, Cur Loss: 0.20403361, Cur Avg Loss: 0.26755234, Log Avg loss: 0.48483601, Global Avg Loss: 1.01615768, Time: 0.0207 Steps: 54210, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000995, Sample Num: 15920, Cur Loss: 0.51919770, Cur Avg Loss: 0.26859588, Log Avg loss: 0.37138471, Global Avg Loss: 1.01603876, Time: 0.0208 Steps: 54220, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001005, Sample Num: 16080, Cur Loss: 0.31735042, Cur Avg Loss: 0.26876340, Log Avg loss: 0.28543113, Global Avg Loss: 1.01590404, Time: 0.0207 Steps: 54230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001015, Sample Num: 16240, Cur Loss: 0.27848619, Cur Avg Loss: 0.26829073, Log Avg loss: 0.22078773, Global Avg Loss: 1.01575745, Time: 0.0208 Steps: 54240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001025, Sample Num: 16400, Cur Loss: 0.32847035, Cur Avg Loss: 0.26844244, Log Avg loss: 0.28384145, Global Avg Loss: 1.01562253, Time: 0.0245 Steps: 54250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001035, Sample Num: 16560, Cur Loss: 0.49634212, Cur Avg Loss: 0.26862484, Log Avg loss: 0.28732012, Global Avg Loss: 1.01548831, Time: 0.0208 Steps: 54260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001045, Sample Num: 16720, Cur Loss: 0.54920763, Cur Avg Loss: 0.26862424, Log Avg loss: 0.26856281, Global Avg Loss: 1.01535068, Time: 0.0207 Steps: 54270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001055, Sample Num: 16880, Cur Loss: 0.19261678, Cur Avg Loss: 0.26922974, Log Avg loss: 0.33250454, Global Avg Loss: 1.01522488, Time: 0.0207 Steps: 54280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001065, Sample Num: 17040, Cur Loss: 0.14460464, Cur Avg Loss: 0.26854499, Log Avg loss: 0.19630402, Global Avg Loss: 1.01507403, Time: 0.0208 Steps: 54290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001075, Sample Num: 17200, Cur Loss: 0.09684510, Cur Avg Loss: 0.26811836, Log Avg loss: 0.22268220, Global Avg Loss: 1.01492811, Time: 0.0208 Steps: 54300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001085, Sample Num: 17360, Cur Loss: 0.13121289, Cur Avg Loss: 0.26861334, Log Avg loss: 0.32182279, Global Avg Loss: 1.01480048, Time: 0.0207 Steps: 54310, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001095, Sample Num: 17520, Cur Loss: 0.08268182, Cur Avg Loss: 0.26796118, Log Avg loss: 0.19720235, Global Avg Loss: 1.01464997, Time: 0.0208 Steps: 54320, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001105, Sample Num: 17680, Cur Loss: 0.43414143, Cur Avg Loss: 0.26828294, Log Avg loss: 0.30351568, Global Avg Loss: 1.01451908, Time: 0.0208 Steps: 54330, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001115, Sample Num: 17840, Cur Loss: 0.48464331, Cur Avg Loss: 0.26825669, Log Avg loss: 0.26535560, Global Avg Loss: 1.01438121, Time: 0.0207 Steps: 54340, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001125, Sample Num: 18000, Cur Loss: 0.15957697, Cur Avg Loss: 0.26768155, Log Avg loss: 0.20355340, Global Avg Loss: 1.01423203, Time: 0.0207 Steps: 54350, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001135, Sample Num: 18160, Cur Loss: 0.09978314, Cur Avg Loss: 0.26713906, Log Avg loss: 0.20610898, Global Avg Loss: 1.01408336, Time: 0.0207 Steps: 54360, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001145, Sample Num: 18320, Cur Loss: 0.17626928, Cur Avg Loss: 0.26687400, Log Avg loss: 0.23678998, Global Avg Loss: 1.01394040, Time: 0.0207 Steps: 54370, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001155, Sample Num: 18480, Cur Loss: 0.16694766, Cur Avg Loss: 0.26649159, Log Avg loss: 0.22270549, Global Avg Loss: 1.01379490, Time: 0.0207 Steps: 54380, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001165, Sample Num: 18640, Cur Loss: 0.16447450, Cur Avg Loss: 0.26656670, Log Avg loss: 0.27524163, Global Avg Loss: 1.01365911, Time: 0.0208 Steps: 54390, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001175, Sample Num: 18800, Cur Loss: 0.45550042, Cur Avg Loss: 0.26691582, Log Avg loss: 0.30758845, Global Avg Loss: 1.01352932, Time: 0.0208 Steps: 54400, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001185, Sample Num: 18960, Cur Loss: 0.16447960, Cur Avg Loss: 0.26650630, Log Avg loss: 0.21838808, Global Avg Loss: 1.01338318, Time: 0.0207 Steps: 54410, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001195, Sample Num: 19120, Cur Loss: 0.28251913, Cur Avg Loss: 0.26666078, Log Avg loss: 0.28496619, Global Avg Loss: 1.01324933, Time: 0.0207 Steps: 54420, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001205, Sample Num: 19280, Cur Loss: 0.19799753, Cur Avg Loss: 0.26669286, Log Avg loss: 0.27052687, Global Avg Loss: 1.01311287, Time: 0.0207 Steps: 54430, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001215, Sample Num: 19440, Cur Loss: 0.15076141, Cur Avg Loss: 0.26805372, Log Avg loss: 0.43203737, Global Avg Loss: 1.01300614, Time: 0.0208 Steps: 54440, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001225, Sample Num: 19600, Cur Loss: 0.21427903, Cur Avg Loss: 0.26925966, Log Avg loss: 0.41578193, Global Avg Loss: 1.01289645, Time: 0.0207 Steps: 54450, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001235, Sample Num: 19760, Cur Loss: 0.38998586, Cur Avg Loss: 0.27041250, Log Avg loss: 0.41163538, Global Avg Loss: 1.01278605, Time: 0.0207 Steps: 54460, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001245, Sample Num: 19920, Cur Loss: 0.41728696, Cur Avg Loss: 0.27068818, Log Avg loss: 0.30473397, Global Avg Loss: 1.01265606, Time: 0.0209 Steps: 54470, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001255, Sample Num: 20080, Cur Loss: 0.15170619, Cur Avg Loss: 0.27062951, Log Avg loss: 0.26332506, Global Avg Loss: 1.01251852, Time: 0.0207 Steps: 54480, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001265, Sample Num: 20240, Cur Loss: 0.19642621, Cur Avg Loss: 0.27074097, Log Avg loss: 0.28472967, Global Avg Loss: 1.01238495, Time: 0.0208 Steps: 54490, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001275, Sample Num: 20400, Cur Loss: 0.16363904, Cur Avg Loss: 0.27079183, Log Avg loss: 0.27722492, Global Avg Loss: 1.01225006, Time: 0.0207 Steps: 54500, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001285, Sample Num: 20560, Cur Loss: 0.41515565, Cur Avg Loss: 0.27329916, Log Avg loss: 0.59298379, Global Avg Loss: 1.01217315, Time: 0.0245 Steps: 54510, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001295, Sample Num: 20720, Cur Loss: 0.24039097, Cur Avg Loss: 0.27331261, Log Avg loss: 0.27504087, Global Avg Loss: 1.01203794, Time: 0.0207 Steps: 54520, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001305, Sample Num: 20880, Cur Loss: 0.07331853, Cur Avg Loss: 0.27306171, Log Avg loss: 0.24057009, Global Avg Loss: 1.01189647, Time: 0.0207 Steps: 54530, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001315, Sample Num: 21040, Cur Loss: 0.16682619, Cur Avg Loss: 0.27255606, Log Avg loss: 0.20656981, Global Avg Loss: 1.01174881, Time: 0.0207 Steps: 54540, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001325, Sample Num: 21200, Cur Loss: 0.15192747, Cur Avg Loss: 0.27245443, Log Avg loss: 0.25908915, Global Avg Loss: 1.01161083, Time: 0.0208 Steps: 54550, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001335, Sample Num: 21360, Cur Loss: 0.18169045, Cur Avg Loss: 0.27238673, Log Avg loss: 0.26341669, Global Avg Loss: 1.01147370, Time: 0.0208 Steps: 54560, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001345, Sample Num: 21520, Cur Loss: 0.16838804, Cur Avg Loss: 0.27247495, Log Avg loss: 0.28425225, Global Avg Loss: 1.01134044, Time: 0.0207 Steps: 54570, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001355, Sample Num: 21680, Cur Loss: 0.21983707, Cur Avg Loss: 0.27196668, Log Avg loss: 0.20360513, Global Avg Loss: 1.01119245, Time: 0.0208 Steps: 54580, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001365, Sample Num: 21840, Cur Loss: 0.31536648, Cur Avg Loss: 0.27159464, Log Avg loss: 0.22118284, Global Avg Loss: 1.01104773, Time: 0.0208 Steps: 54590, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001375, Sample Num: 22000, Cur Loss: 0.09584025, Cur Avg Loss: 0.27164567, Log Avg loss: 0.27861052, Global Avg Loss: 1.01091358, Time: 0.0207 Steps: 54600, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001385, Sample Num: 22160, Cur Loss: 0.49875101, Cur Avg Loss: 0.27225768, Log Avg loss: 0.35641004, Global Avg Loss: 1.01079373, Time: 0.0208 Steps: 54610, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001395, Sample Num: 22320, Cur Loss: 0.27308953, Cur Avg Loss: 0.27268112, Log Avg loss: 0.33132764, Global Avg Loss: 1.01066933, Time: 0.0207 Steps: 54620, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001405, Sample Num: 22480, Cur Loss: 0.14644998, Cur Avg Loss: 0.27176111, Log Avg loss: 0.14341929, Global Avg Loss: 1.01051058, Time: 0.0208 Steps: 54630, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001415, Sample Num: 22640, Cur Loss: 0.17249259, Cur Avg Loss: 0.27157914, Log Avg loss: 0.24601250, Global Avg Loss: 1.01037067, Time: 0.0208 Steps: 54640, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001425, Sample Num: 22800, Cur Loss: 0.23662320, Cur Avg Loss: 0.27166048, Log Avg loss: 0.28316925, Global Avg Loss: 1.01023760, Time: 0.0207 Steps: 54650, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001435, Sample Num: 22960, Cur Loss: 0.09347497, Cur Avg Loss: 0.27148924, Log Avg loss: 0.24708756, Global Avg Loss: 1.01009799, Time: 0.0207 Steps: 54660, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001445, Sample Num: 23120, Cur Loss: 0.25365669, Cur Avg Loss: 0.27087946, Log Avg loss: 0.18337642, Global Avg Loss: 1.00994677, Time: 0.0208 Steps: 54670, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001455, Sample Num: 23280, Cur Loss: 0.34058914, Cur Avg Loss: 0.27154872, Log Avg loss: 0.36825633, Global Avg Loss: 1.00982941, Time: 0.0207 Steps: 54680, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001465, Sample Num: 23440, Cur Loss: 0.38540667, Cur Avg Loss: 0.27147481, Log Avg loss: 0.26072216, Global Avg Loss: 1.00969244, Time: 0.0208 Steps: 54690, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001475, Sample Num: 23600, Cur Loss: 0.22371054, Cur Avg Loss: 0.27175938, Log Avg loss: 0.31344784, Global Avg Loss: 1.00956515, Time: 0.0208 Steps: 54700, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001485, Sample Num: 23760, Cur Loss: 0.20470312, Cur Avg Loss: 0.27122573, Log Avg loss: 0.19251225, Global Avg Loss: 1.00941581, Time: 0.0207 Steps: 54710, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001495, Sample Num: 23920, Cur Loss: 0.10440452, Cur Avg Loss: 0.27126443, Log Avg loss: 0.27701257, Global Avg Loss: 1.00928197, Time: 0.0207 Steps: 54720, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001505, Sample Num: 24080, Cur Loss: 0.38581446, Cur Avg Loss: 0.27151864, Log Avg loss: 0.30952291, Global Avg Loss: 1.00915411, Time: 0.0207 Steps: 54730, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001515, Sample Num: 24240, Cur Loss: 0.77558649, Cur Avg Loss: 0.27244555, Log Avg loss: 0.41194512, Global Avg Loss: 1.00904501, Time: 0.0207 Steps: 54740, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001525, Sample Num: 24400, Cur Loss: 0.48791659, Cur Avg Loss: 0.27328483, Log Avg loss: 0.40043551, Global Avg Loss: 1.00893385, Time: 0.0208 Steps: 54750, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001535, Sample Num: 24560, Cur Loss: 0.12414028, Cur Avg Loss: 0.27343723, Log Avg loss: 0.29667834, Global Avg Loss: 1.00880378, Time: 0.0208 Steps: 54760, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001545, Sample Num: 24720, Cur Loss: 0.23973399, Cur Avg Loss: 0.27368752, Log Avg loss: 0.31210668, Global Avg Loss: 1.00867658, Time: 0.0209 Steps: 54770, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001555, Sample Num: 24880, Cur Loss: 0.09814334, Cur Avg Loss: 0.27342339, Log Avg loss: 0.23261571, Global Avg Loss: 1.00853491, Time: 0.0207 Steps: 54780, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001565, Sample Num: 25040, Cur Loss: 0.40302315, Cur Avg Loss: 0.27276894, Log Avg loss: 0.17100221, Global Avg Loss: 1.00838205, Time: 0.0208 Steps: 54790, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001575, Sample Num: 25200, Cur Loss: 0.26252818, Cur Avg Loss: 0.27235334, Log Avg loss: 0.20731227, Global Avg Loss: 1.00823586, Time: 0.0208 Steps: 54800, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001585, Sample Num: 25360, Cur Loss: 0.41257250, Cur Avg Loss: 0.27214742, Log Avg loss: 0.23971369, Global Avg Loss: 1.00809565, Time: 0.0211 Steps: 54810, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001595, Sample Num: 25520, Cur Loss: 0.38038498, Cur Avg Loss: 0.27181048, Log Avg loss: 0.21840691, Global Avg Loss: 1.00795160, Time: 0.0212 Steps: 54820, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001605, Sample Num: 25680, Cur Loss: 0.32477918, Cur Avg Loss: 0.27229569, Log Avg loss: 0.34968517, Global Avg Loss: 1.00783154, Time: 0.0211 Steps: 54830, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001615, Sample Num: 25840, Cur Loss: 0.26850685, Cur Avg Loss: 0.27303959, Log Avg loss: 0.39243611, Global Avg Loss: 1.00771933, Time: 0.0211 Steps: 54840, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001625, Sample Num: 26000, Cur Loss: 0.26100212, Cur Avg Loss: 0.27270646, Log Avg loss: 0.21890687, Global Avg Loss: 1.00757551, Time: 0.0211 Steps: 54850, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001635, Sample Num: 26160, Cur Loss: 0.60717237, Cur Avg Loss: 0.27281767, Log Avg loss: 0.29088940, Global Avg Loss: 1.00744487, Time: 0.0211 Steps: 54860, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001645, Sample Num: 26320, Cur Loss: 0.04518273, Cur Avg Loss: 0.27272579, Log Avg loss: 0.25770246, Global Avg Loss: 1.00730823, Time: 0.0211 Steps: 54870, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001655, Sample Num: 26480, Cur Loss: 0.30920136, Cur Avg Loss: 0.27289682, Log Avg loss: 0.30103108, Global Avg Loss: 1.00717954, Time: 0.0210 Steps: 54880, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001665, Sample Num: 26640, Cur Loss: 0.41634667, Cur Avg Loss: 0.27281834, Log Avg loss: 0.25982944, Global Avg Loss: 1.00704338, Time: 0.0211 Steps: 54890, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001675, Sample Num: 26800, Cur Loss: 0.29009116, Cur Avg Loss: 0.27245050, Log Avg loss: 0.21120659, Global Avg Loss: 1.00689842, Time: 0.0208 Steps: 54900, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001685, Sample Num: 26960, Cur Loss: 0.25499833, Cur Avg Loss: 0.27240487, Log Avg loss: 0.26476165, Global Avg Loss: 1.00676327, Time: 0.0208 Steps: 54910, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001695, Sample Num: 27120, Cur Loss: 0.49734789, Cur Avg Loss: 0.27232167, Log Avg loss: 0.25830237, Global Avg Loss: 1.00662699, Time: 0.0208 Steps: 54920, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001705, Sample Num: 27280, Cur Loss: 0.31755275, Cur Avg Loss: 0.27191912, Log Avg loss: 0.20368576, Global Avg Loss: 1.00648081, Time: 0.0208 Steps: 54930, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001715, Sample Num: 27440, Cur Loss: 0.23014277, Cur Avg Loss: 0.27168504, Log Avg loss: 0.23177452, Global Avg Loss: 1.00633980, Time: 0.0207 Steps: 54940, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001725, Sample Num: 27600, Cur Loss: 0.17755771, Cur Avg Loss: 0.27135394, Log Avg loss: 0.21457019, Global Avg Loss: 1.00619571, Time: 0.0208 Steps: 54950, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001735, Sample Num: 27760, Cur Loss: 0.62811708, Cur Avg Loss: 0.27159883, Log Avg loss: 0.31384290, Global Avg Loss: 1.00606974, Time: 0.0211 Steps: 54960, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001745, Sample Num: 27920, Cur Loss: 0.21647300, Cur Avg Loss: 0.27170926, Log Avg loss: 0.29086843, Global Avg Loss: 1.00593963, Time: 0.0210 Steps: 54970, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001755, Sample Num: 28080, Cur Loss: 0.24436259, Cur Avg Loss: 0.27140076, Log Avg loss: 0.21756849, Global Avg Loss: 1.00579624, Time: 0.0211 Steps: 54980, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001765, Sample Num: 28240, Cur Loss: 0.37266707, Cur Avg Loss: 0.27157810, Log Avg loss: 0.30270044, Global Avg Loss: 1.00566838, Time: 0.0211 Steps: 54990, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001775, Sample Num: 28400, Cur Loss: 0.24092242, Cur Avg Loss: 0.27130103, Log Avg loss: 0.22239781, Global Avg Loss: 1.00552597, Time: 0.0211 Steps: 55000, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001785, Sample Num: 28560, Cur Loss: 0.31536883, Cur Avg Loss: 0.27154511, Log Avg loss: 0.31487061, Global Avg Loss: 1.00540042, Time: 0.0211 Steps: 55010, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001795, Sample Num: 28720, Cur Loss: 0.08590972, Cur Avg Loss: 0.27085266, Log Avg loss: 0.14725069, Global Avg Loss: 1.00524445, Time: 0.0248 Steps: 55020, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001805, Sample Num: 28880, Cur Loss: 0.08136348, Cur Avg Loss: 0.27008220, Log Avg loss: 0.13178461, Global Avg Loss: 1.00508572, Time: 0.0208 Steps: 55030, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001815, Sample Num: 29040, Cur Loss: 0.24511123, Cur Avg Loss: 0.27047074, Log Avg loss: 0.34060112, Global Avg Loss: 1.00496499, Time: 0.0211 Steps: 55040, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001825, Sample Num: 29200, Cur Loss: 0.51175767, Cur Avg Loss: 0.27028695, Log Avg loss: 0.23692899, Global Avg Loss: 1.00482548, Time: 0.0208 Steps: 55050, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001835, Sample Num: 29360, Cur Loss: 0.27571937, Cur Avg Loss: 0.27043687, Log Avg loss: 0.29779747, Global Avg Loss: 1.00469707, Time: 0.0208 Steps: 55060, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001845, Sample Num: 29520, Cur Loss: 0.23095223, Cur Avg Loss: 0.27064197, Log Avg loss: 0.30827819, Global Avg Loss: 1.00457061, Time: 0.0211 Steps: 55070, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001855, Sample Num: 29680, Cur Loss: 0.15169464, Cur Avg Loss: 0.27030246, Log Avg loss: 0.20766235, Global Avg Loss: 1.00442592, Time: 0.0208 Steps: 55080, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001865, Sample Num: 29840, Cur Loss: 0.29390907, Cur Avg Loss: 0.27019594, Log Avg loss: 0.25043650, Global Avg Loss: 1.00428906, Time: 0.0212 Steps: 55090, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001875, Sample Num: 30000, Cur Loss: 0.26645750, Cur Avg Loss: 0.27032426, Log Avg loss: 0.29425558, Global Avg Loss: 1.00416020, Time: 0.0211 Steps: 55100, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001885, Sample Num: 30160, Cur Loss: 0.43747392, Cur Avg Loss: 0.27070815, Log Avg loss: 0.34268889, Global Avg Loss: 1.00404017, Time: 0.0210 Steps: 55110, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001895, Sample Num: 30320, Cur Loss: 0.31403556, Cur Avg Loss: 0.27136401, Log Avg loss: 0.39499370, Global Avg Loss: 1.00392967, Time: 0.0208 Steps: 55120, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001905, Sample Num: 30480, Cur Loss: 0.32564273, Cur Avg Loss: 0.27214499, Log Avg loss: 0.42013945, Global Avg Loss: 1.00382378, Time: 0.0207 Steps: 55130, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001915, Sample Num: 30640, Cur Loss: 0.16669652, Cur Avg Loss: 0.27204421, Log Avg loss: 0.25284642, Global Avg Loss: 1.00368759, Time: 0.0208 Steps: 55140, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001925, Sample Num: 30800, Cur Loss: 0.18291727, Cur Avg Loss: 0.27189212, Log Avg loss: 0.24276604, Global Avg Loss: 1.00354961, Time: 0.0209 Steps: 55150, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001935, Sample Num: 30960, Cur Loss: 0.20437433, Cur Avg Loss: 0.27181757, Log Avg loss: 0.25746788, Global Avg Loss: 1.00341436, Time: 0.0207 Steps: 55160, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001945, Sample Num: 31120, Cur Loss: 0.05074201, Cur Avg Loss: 0.27182272, Log Avg loss: 0.27281881, Global Avg Loss: 1.00328193, Time: 0.0207 Steps: 55170, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001955, Sample Num: 31280, Cur Loss: 0.25267747, Cur Avg Loss: 0.27207066, Log Avg loss: 0.32029536, Global Avg Loss: 1.00315816, Time: 0.0208 Steps: 55180, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001965, Sample Num: 31440, Cur Loss: 0.14357635, Cur Avg Loss: 0.27213885, Log Avg loss: 0.28547007, Global Avg Loss: 1.00302812, Time: 0.0207 Steps: 55190, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001975, Sample Num: 31600, Cur Loss: 0.26564324, Cur Avg Loss: 0.27181506, Log Avg loss: 0.20819010, Global Avg Loss: 1.00288412, Time: 0.0207 Steps: 55200, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001985, Sample Num: 31760, Cur Loss: 0.16542026, Cur Avg Loss: 0.27152010, Log Avg loss: 0.21326582, Global Avg Loss: 1.00274110, Time: 0.0211 Steps: 55210, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001995, Sample Num: 31920, Cur Loss: 0.14789568, Cur Avg Loss: 0.27096700, Log Avg loss: 0.16117651, Global Avg Loss: 1.00258870, Time: 0.0211 Steps: 55220, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002005, Sample Num: 32080, Cur Loss: 0.21513090, Cur Avg Loss: 0.27123691, Log Avg loss: 0.32508247, Global Avg Loss: 1.00246603, Time: 0.0208 Steps: 55230, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002015, Sample Num: 32240, Cur Loss: 0.42088923, Cur Avg Loss: 0.27164219, Log Avg loss: 0.35290249, Global Avg Loss: 1.00234844, Time: 0.0211 Steps: 55240, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002025, Sample Num: 32400, Cur Loss: 0.45511848, Cur Avg Loss: 0.27188135, Log Avg loss: 0.32007215, Global Avg Loss: 1.00222495, Time: 0.0211 Steps: 55250, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002035, Sample Num: 32560, Cur Loss: 0.71113646, Cur Avg Loss: 0.27199276, Log Avg loss: 0.29455173, Global Avg Loss: 1.00209689, Time: 0.0208 Steps: 55260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002045, Sample Num: 32720, Cur Loss: 0.11514673, Cur Avg Loss: 0.27264340, Log Avg loss: 0.40504867, Global Avg Loss: 1.00198887, Time: 0.0208 Steps: 55270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002055, Sample Num: 32880, Cur Loss: 0.16000250, Cur Avg Loss: 0.27295138, Log Avg loss: 0.33593492, Global Avg Loss: 1.00186838, Time: 0.0208 Steps: 55280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002065, Sample Num: 33040, Cur Loss: 0.24371332, Cur Avg Loss: 0.27244259, Log Avg loss: 0.16788612, Global Avg Loss: 1.00171754, Time: 0.0207 Steps: 55290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002075, Sample Num: 33200, Cur Loss: 0.15725270, Cur Avg Loss: 0.27199420, Log Avg loss: 0.17940149, Global Avg Loss: 1.00156884, Time: 0.0208 Steps: 55300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002085, Sample Num: 33360, Cur Loss: 0.19207293, Cur Avg Loss: 0.27184588, Log Avg loss: 0.24106790, Global Avg Loss: 1.00143134, Time: 0.0208 Steps: 55310, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002095, Sample Num: 33520, Cur Loss: 0.33884469, Cur Avg Loss: 0.27158846, Log Avg loss: 0.21791753, Global Avg Loss: 1.00128971, Time: 0.0207 Steps: 55320, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002105, Sample Num: 33680, Cur Loss: 0.16300347, Cur Avg Loss: 0.27179339, Log Avg loss: 0.31472632, Global Avg Loss: 1.00116562, Time: 0.0207 Steps: 55330, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002115, Sample Num: 33840, Cur Loss: 0.20594832, Cur Avg Loss: 0.27187667, Log Avg loss: 0.28940737, Global Avg Loss: 1.00103701, Time: 0.0208 Steps: 55340, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002125, Sample Num: 34000, Cur Loss: 0.47581553, Cur Avg Loss: 0.27164031, Log Avg loss: 0.22164947, Global Avg Loss: 1.00089620, Time: 0.0208 Steps: 55350, Updated lr: 0.000048 ***** Running evaluation checkpoint-55354 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-55354 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.553354, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.422039, "eval_total_loss": 296.69351, "eval_mae": 0.535537, "eval_mse": 0.422064, "eval_r2": 0.731708, "eval_sp_statistic": 0.880055, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.910882, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.321216, "test_total_loss": 161.250487, "test_mae": 0.403831, "test_mse": 0.321327, "test_r2": 0.792613, "test_sp_statistic": 0.860546, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.911877, "test_ps_pvalue": 0.0, "lr": 4.845519203413941e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0008370800863418, "train_cur_epoch_loss": 577.9668644005433, "train_cur_epoch_avg_loss": 0.27147339802749804, "train_cur_epoch_time": 44.55335354804993, "train_cur_epoch_avg_time": 0.020926892225481414, "epoch": 26, "step": 55354} ################################################## Training, Epoch: 0027, Batch: 000006, Sample Num: 96, Cur Loss: 0.29683799, Cur Avg Loss: 0.16301422, Log Avg loss: 0.17092925, Global Avg Loss: 1.00074628, Time: 0.0247 Steps: 55360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000016, Sample Num: 256, Cur Loss: 0.26426354, Cur Avg Loss: 0.23444215, Log Avg loss: 0.27729891, Global Avg Loss: 1.00061562, Time: 0.0208 Steps: 55370, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000026, Sample Num: 416, Cur Loss: 0.19087696, Cur Avg Loss: 0.23266048, Log Avg loss: 0.22980981, Global Avg Loss: 1.00047643, Time: 0.0209 Steps: 55380, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000036, Sample Num: 576, Cur Loss: 0.13028595, Cur Avg Loss: 0.22402184, Log Avg loss: 0.20156136, Global Avg Loss: 1.00033220, Time: 0.0208 Steps: 55390, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000046, Sample Num: 736, Cur Loss: 0.24096459, Cur Avg Loss: 0.23804427, Log Avg loss: 0.28852504, Global Avg Loss: 1.00020371, Time: 0.0208 Steps: 55400, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000056, Sample Num: 896, Cur Loss: 0.21262860, Cur Avg Loss: 0.23712505, Log Avg loss: 0.23289663, Global Avg Loss: 1.00006524, Time: 0.0208 Steps: 55410, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000066, Sample Num: 1056, Cur Loss: 0.66318995, Cur Avg Loss: 0.25088466, Log Avg loss: 0.32793849, Global Avg Loss: 0.99994396, Time: 0.0209 Steps: 55420, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000076, Sample Num: 1216, Cur Loss: 0.12501197, Cur Avg Loss: 0.23929920, Log Avg loss: 0.16283514, Global Avg Loss: 0.99979294, Time: 0.0209 Steps: 55430, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000086, Sample Num: 1376, Cur Loss: 0.36747086, Cur Avg Loss: 0.24182465, Log Avg loss: 0.26101811, Global Avg Loss: 0.99965968, Time: 0.0208 Steps: 55440, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000096, Sample Num: 1536, Cur Loss: 0.29288542, Cur Avg Loss: 0.24271581, Log Avg loss: 0.25037971, Global Avg Loss: 0.99952455, Time: 0.0208 Steps: 55450, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000106, Sample Num: 1696, Cur Loss: 0.28450391, Cur Avg Loss: 0.24734779, Log Avg loss: 0.29181480, Global Avg Loss: 0.99939695, Time: 0.0208 Steps: 55460, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000116, Sample Num: 1856, Cur Loss: 0.14403343, Cur Avg Loss: 0.25249416, Log Avg loss: 0.30704568, Global Avg Loss: 0.99927213, Time: 0.0208 Steps: 55470, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000126, Sample Num: 2016, Cur Loss: 0.29319543, Cur Avg Loss: 0.24923611, Log Avg loss: 0.21144280, Global Avg Loss: 0.99913013, Time: 0.0208 Steps: 55480, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000136, Sample Num: 2176, Cur Loss: 0.25421795, Cur Avg Loss: 0.25515989, Log Avg loss: 0.32979945, Global Avg Loss: 0.99900951, Time: 0.0208 Steps: 55490, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000146, Sample Num: 2336, Cur Loss: 0.40989304, Cur Avg Loss: 0.25495986, Log Avg loss: 0.25223950, Global Avg Loss: 0.99887495, Time: 0.0208 Steps: 55500, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000156, Sample Num: 2496, Cur Loss: 0.14520560, Cur Avg Loss: 0.25570926, Log Avg loss: 0.26665051, Global Avg Loss: 0.99874304, Time: 0.0208 Steps: 55510, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000166, Sample Num: 2656, Cur Loss: 0.14072604, Cur Avg Loss: 0.25418299, Log Avg loss: 0.23037320, Global Avg Loss: 0.99860465, Time: 0.0208 Steps: 55520, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000176, Sample Num: 2816, Cur Loss: 0.09302773, Cur Avg Loss: 0.25024496, Log Avg loss: 0.18487359, Global Avg Loss: 0.99845811, Time: 0.0208 Steps: 55530, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000186, Sample Num: 2976, Cur Loss: 0.15038252, Cur Avg Loss: 0.25007201, Log Avg loss: 0.24702822, Global Avg Loss: 0.99832281, Time: 0.0208 Steps: 55540, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000196, Sample Num: 3136, Cur Loss: 0.23722354, Cur Avg Loss: 0.25124156, Log Avg loss: 0.27299507, Global Avg Loss: 0.99819224, Time: 0.0208 Steps: 55550, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000206, Sample Num: 3296, Cur Loss: 0.20686497, Cur Avg Loss: 0.24985153, Log Avg loss: 0.22260699, Global Avg Loss: 0.99805265, Time: 0.0208 Steps: 55560, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000216, Sample Num: 3456, Cur Loss: 0.44439831, Cur Avg Loss: 0.25273327, Log Avg loss: 0.31209708, Global Avg Loss: 0.99792921, Time: 0.0208 Steps: 55570, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000226, Sample Num: 3616, Cur Loss: 0.11524098, Cur Avg Loss: 0.24994743, Log Avg loss: 0.18977335, Global Avg Loss: 0.99778380, Time: 0.0209 Steps: 55580, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000236, Sample Num: 3776, Cur Loss: 0.20386165, Cur Avg Loss: 0.24737053, Log Avg loss: 0.18913260, Global Avg Loss: 0.99763834, Time: 0.0208 Steps: 55590, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000246, Sample Num: 3936, Cur Loss: 0.20755394, Cur Avg Loss: 0.24874382, Log Avg loss: 0.28115330, Global Avg Loss: 0.99750947, Time: 0.0208 Steps: 55600, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000256, Sample Num: 4096, Cur Loss: 0.15577114, Cur Avg Loss: 0.24944654, Log Avg loss: 0.26673364, Global Avg Loss: 0.99737806, Time: 0.0253 Steps: 55610, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000266, Sample Num: 4256, Cur Loss: 0.15496027, Cur Avg Loss: 0.24918333, Log Avg loss: 0.24244517, Global Avg Loss: 0.99724233, Time: 0.0208 Steps: 55620, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000276, Sample Num: 4416, Cur Loss: 0.19171539, Cur Avg Loss: 0.24854079, Log Avg loss: 0.23144899, Global Avg Loss: 0.99710467, Time: 0.0208 Steps: 55630, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000286, Sample Num: 4576, Cur Loss: 0.22598493, Cur Avg Loss: 0.24704114, Log Avg loss: 0.20565086, Global Avg Loss: 0.99696243, Time: 0.0208 Steps: 55640, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000296, Sample Num: 4736, Cur Loss: 0.23138511, Cur Avg Loss: 0.24902175, Log Avg loss: 0.30566728, Global Avg Loss: 0.99683821, Time: 0.0208 Steps: 55650, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000306, Sample Num: 4896, Cur Loss: 0.18083239, Cur Avg Loss: 0.24940091, Log Avg loss: 0.26062415, Global Avg Loss: 0.99670594, Time: 0.0208 Steps: 55660, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000316, Sample Num: 5056, Cur Loss: 0.30733490, Cur Avg Loss: 0.24884036, Log Avg loss: 0.23168751, Global Avg Loss: 0.99656852, Time: 0.0208 Steps: 55670, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000326, Sample Num: 5216, Cur Loss: 0.27004343, Cur Avg Loss: 0.25126427, Log Avg loss: 0.32785960, Global Avg Loss: 0.99644842, Time: 0.0208 Steps: 55680, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000336, Sample Num: 5376, Cur Loss: 0.21943161, Cur Avg Loss: 0.25099522, Log Avg loss: 0.24222435, Global Avg Loss: 0.99631298, Time: 0.0208 Steps: 55690, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000346, Sample Num: 5536, Cur Loss: 0.37320358, Cur Avg Loss: 0.25236935, Log Avg loss: 0.29853996, Global Avg Loss: 0.99618771, Time: 0.0208 Steps: 55700, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000356, Sample Num: 5696, Cur Loss: 0.16780491, Cur Avg Loss: 0.25326209, Log Avg loss: 0.28415101, Global Avg Loss: 0.99605990, Time: 0.0209 Steps: 55710, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000366, Sample Num: 5856, Cur Loss: 0.18934041, Cur Avg Loss: 0.25454674, Log Avg loss: 0.30028020, Global Avg Loss: 0.99593503, Time: 0.0208 Steps: 55720, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000376, Sample Num: 6016, Cur Loss: 0.33378792, Cur Avg Loss: 0.25416930, Log Avg loss: 0.24035496, Global Avg Loss: 0.99579945, Time: 0.0208 Steps: 55730, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000386, Sample Num: 6176, Cur Loss: 0.33682597, Cur Avg Loss: 0.25296231, Log Avg loss: 0.20757949, Global Avg Loss: 0.99565804, Time: 0.0208 Steps: 55740, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000396, Sample Num: 6336, Cur Loss: 0.22648528, Cur Avg Loss: 0.25221221, Log Avg loss: 0.22325830, Global Avg Loss: 0.99551949, Time: 0.0208 Steps: 55750, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000406, Sample Num: 6496, Cur Loss: 0.29084924, Cur Avg Loss: 0.25057788, Log Avg loss: 0.18585877, Global Avg Loss: 0.99537429, Time: 0.0208 Steps: 55760, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000416, Sample Num: 6656, Cur Loss: 0.71069807, Cur Avg Loss: 0.25004457, Log Avg loss: 0.22839181, Global Avg Loss: 0.99523676, Time: 0.0207 Steps: 55770, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000426, Sample Num: 6816, Cur Loss: 0.16085032, Cur Avg Loss: 0.24953750, Log Avg loss: 0.22844350, Global Avg Loss: 0.99509930, Time: 0.0208 Steps: 55780, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000436, Sample Num: 6976, Cur Loss: 0.04705523, Cur Avg Loss: 0.25234903, Log Avg loss: 0.37212020, Global Avg Loss: 0.99498763, Time: 0.0208 Steps: 55790, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000446, Sample Num: 7136, Cur Loss: 0.21129936, Cur Avg Loss: 0.25566266, Log Avg loss: 0.40013697, Global Avg Loss: 0.99488103, Time: 0.0208 Steps: 55800, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000456, Sample Num: 7296, Cur Loss: 0.13033669, Cur Avg Loss: 0.25596335, Log Avg loss: 0.26937400, Global Avg Loss: 0.99475103, Time: 0.0208 Steps: 55810, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000466, Sample Num: 7456, Cur Loss: 0.07927640, Cur Avg Loss: 0.25553456, Log Avg loss: 0.23598169, Global Avg Loss: 0.99461510, Time: 0.0208 Steps: 55820, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000476, Sample Num: 7616, Cur Loss: 0.07599272, Cur Avg Loss: 0.25388950, Log Avg loss: 0.17722984, Global Avg Loss: 0.99446869, Time: 0.0208 Steps: 55830, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000486, Sample Num: 7776, Cur Loss: 0.23591542, Cur Avg Loss: 0.25317831, Log Avg loss: 0.21932566, Global Avg Loss: 0.99432988, Time: 0.0208 Steps: 55840, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000496, Sample Num: 7936, Cur Loss: 0.61841255, Cur Avg Loss: 0.25545513, Log Avg loss: 0.36610876, Global Avg Loss: 0.99421739, Time: 0.0208 Steps: 55850, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000506, Sample Num: 8096, Cur Loss: 0.09300785, Cur Avg Loss: 0.25661004, Log Avg loss: 0.31389357, Global Avg Loss: 0.99409560, Time: 0.0208 Steps: 55860, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000516, Sample Num: 8256, Cur Loss: 0.11833245, Cur Avg Loss: 0.25599023, Log Avg loss: 0.22462782, Global Avg Loss: 0.99395788, Time: 0.0245 Steps: 55870, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000526, Sample Num: 8416, Cur Loss: 0.39342713, Cur Avg Loss: 0.25605320, Log Avg loss: 0.25930214, Global Avg Loss: 0.99382641, Time: 0.0208 Steps: 55880, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000536, Sample Num: 8576, Cur Loss: 0.30007926, Cur Avg Loss: 0.25677882, Log Avg loss: 0.29494663, Global Avg Loss: 0.99370136, Time: 0.0207 Steps: 55890, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000546, Sample Num: 8736, Cur Loss: 0.07886577, Cur Avg Loss: 0.25527107, Log Avg loss: 0.17445568, Global Avg Loss: 0.99355481, Time: 0.0208 Steps: 55900, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000556, Sample Num: 8896, Cur Loss: 0.07886209, Cur Avg Loss: 0.25426374, Log Avg loss: 0.19926377, Global Avg Loss: 0.99341274, Time: 0.0207 Steps: 55910, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000566, Sample Num: 9056, Cur Loss: 0.07493903, Cur Avg Loss: 0.25361390, Log Avg loss: 0.21748244, Global Avg Loss: 0.99327398, Time: 0.0208 Steps: 55920, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000576, Sample Num: 9216, Cur Loss: 0.15192589, Cur Avg Loss: 0.25329177, Log Avg loss: 0.23505929, Global Avg Loss: 0.99313842, Time: 0.0207 Steps: 55930, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000586, Sample Num: 9376, Cur Loss: 0.26497382, Cur Avg Loss: 0.25198828, Log Avg loss: 0.17690706, Global Avg Loss: 0.99299251, Time: 0.0208 Steps: 55940, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000596, Sample Num: 9536, Cur Loss: 0.30548596, Cur Avg Loss: 0.25242230, Log Avg loss: 0.27785605, Global Avg Loss: 0.99286469, Time: 0.0208 Steps: 55950, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000606, Sample Num: 9696, Cur Loss: 0.10129772, Cur Avg Loss: 0.25287053, Log Avg loss: 0.27958530, Global Avg Loss: 0.99273723, Time: 0.0207 Steps: 55960, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000616, Sample Num: 9856, Cur Loss: 0.28479952, Cur Avg Loss: 0.25298614, Log Avg loss: 0.25999185, Global Avg Loss: 0.99260631, Time: 0.0208 Steps: 55970, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000626, Sample Num: 10016, Cur Loss: 0.12633076, Cur Avg Loss: 0.25182683, Log Avg loss: 0.18041316, Global Avg Loss: 0.99246122, Time: 0.0208 Steps: 55980, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000636, Sample Num: 10176, Cur Loss: 0.29701155, Cur Avg Loss: 0.25358355, Log Avg loss: 0.36355461, Global Avg Loss: 0.99234890, Time: 0.0208 Steps: 55990, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000646, Sample Num: 10336, Cur Loss: 0.55004472, Cur Avg Loss: 0.25511153, Log Avg loss: 0.35229091, Global Avg Loss: 0.99223460, Time: 0.0208 Steps: 56000, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000656, Sample Num: 10496, Cur Loss: 0.22185098, Cur Avg Loss: 0.25535286, Log Avg loss: 0.27094293, Global Avg Loss: 0.99210582, Time: 0.0207 Steps: 56010, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000666, Sample Num: 10656, Cur Loss: 0.36428258, Cur Avg Loss: 0.25497529, Log Avg loss: 0.23020639, Global Avg Loss: 0.99196982, Time: 0.0208 Steps: 56020, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000676, Sample Num: 10816, Cur Loss: 0.42205054, Cur Avg Loss: 0.25545985, Log Avg loss: 0.28773206, Global Avg Loss: 0.99184413, Time: 0.0208 Steps: 56030, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000686, Sample Num: 10976, Cur Loss: 0.12557389, Cur Avg Loss: 0.25547786, Log Avg loss: 0.25669527, Global Avg Loss: 0.99171295, Time: 0.0208 Steps: 56040, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000696, Sample Num: 11136, Cur Loss: 0.51328826, Cur Avg Loss: 0.25536109, Log Avg loss: 0.24735037, Global Avg Loss: 0.99158014, Time: 0.0208 Steps: 56050, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000706, Sample Num: 11296, Cur Loss: 0.97163999, Cur Avg Loss: 0.25556938, Log Avg loss: 0.27006660, Global Avg Loss: 0.99145144, Time: 0.0208 Steps: 56060, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000716, Sample Num: 11456, Cur Loss: 0.16053820, Cur Avg Loss: 0.25570177, Log Avg loss: 0.26504840, Global Avg Loss: 0.99132189, Time: 0.0208 Steps: 56070, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000726, Sample Num: 11616, Cur Loss: 0.17801069, Cur Avg Loss: 0.25506297, Log Avg loss: 0.20932452, Global Avg Loss: 0.99118244, Time: 0.0208 Steps: 56080, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000736, Sample Num: 11776, Cur Loss: 0.14351130, Cur Avg Loss: 0.25575785, Log Avg loss: 0.30620677, Global Avg Loss: 0.99106032, Time: 0.0208 Steps: 56090, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000746, Sample Num: 11936, Cur Loss: 0.27609554, Cur Avg Loss: 0.25542501, Log Avg loss: 0.23092747, Global Avg Loss: 0.99092483, Time: 0.0208 Steps: 56100, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000756, Sample Num: 12096, Cur Loss: 0.27111438, Cur Avg Loss: 0.25399418, Log Avg loss: 0.14725449, Global Avg Loss: 0.99077447, Time: 0.0208 Steps: 56110, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000766, Sample Num: 12256, Cur Loss: 0.22456905, Cur Avg Loss: 0.25410789, Log Avg loss: 0.26270432, Global Avg Loss: 0.99064473, Time: 0.0208 Steps: 56120, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000776, Sample Num: 12416, Cur Loss: 0.25139460, Cur Avg Loss: 0.25507180, Log Avg loss: 0.32890728, Global Avg Loss: 0.99052684, Time: 0.0211 Steps: 56130, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000786, Sample Num: 12576, Cur Loss: 0.32570207, Cur Avg Loss: 0.25500275, Log Avg loss: 0.24964458, Global Avg Loss: 0.99039487, Time: 0.0210 Steps: 56140, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000796, Sample Num: 12736, Cur Loss: 0.25080562, Cur Avg Loss: 0.25659485, Log Avg loss: 0.38173386, Global Avg Loss: 0.99028647, Time: 0.0210 Steps: 56150, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000806, Sample Num: 12896, Cur Loss: 0.27188149, Cur Avg Loss: 0.25763162, Log Avg loss: 0.34015879, Global Avg Loss: 0.99017071, Time: 0.0210 Steps: 56160, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000816, Sample Num: 13056, Cur Loss: 0.27136600, Cur Avg Loss: 0.25767333, Log Avg loss: 0.26103485, Global Avg Loss: 0.99004090, Time: 0.0210 Steps: 56170, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000826, Sample Num: 13216, Cur Loss: 0.12473252, Cur Avg Loss: 0.25862969, Log Avg loss: 0.33666858, Global Avg Loss: 0.98992460, Time: 0.0210 Steps: 56180, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000836, Sample Num: 13376, Cur Loss: 0.20171019, Cur Avg Loss: 0.25964408, Log Avg loss: 0.34343313, Global Avg Loss: 0.98980954, Time: 0.0211 Steps: 56190, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000846, Sample Num: 13536, Cur Loss: 0.26881844, Cur Avg Loss: 0.25915843, Log Avg loss: 0.21855739, Global Avg Loss: 0.98967231, Time: 0.0210 Steps: 56200, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000856, Sample Num: 13696, Cur Loss: 0.15422748, Cur Avg Loss: 0.25874741, Log Avg loss: 0.22397512, Global Avg Loss: 0.98953609, Time: 0.0210 Steps: 56210, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000866, Sample Num: 13856, Cur Loss: 0.19342653, Cur Avg Loss: 0.25825077, Log Avg loss: 0.21573848, Global Avg Loss: 0.98939845, Time: 0.0210 Steps: 56220, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000876, Sample Num: 14016, Cur Loss: 0.48619759, Cur Avg Loss: 0.25753083, Log Avg loss: 0.19518398, Global Avg Loss: 0.98925721, Time: 0.0210 Steps: 56230, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000886, Sample Num: 14176, Cur Loss: 0.21468072, Cur Avg Loss: 0.25797506, Log Avg loss: 0.29688946, Global Avg Loss: 0.98913410, Time: 0.0210 Steps: 56240, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000896, Sample Num: 14336, Cur Loss: 0.13890937, Cur Avg Loss: 0.25772200, Log Avg loss: 0.23530137, Global Avg Loss: 0.98900008, Time: 0.0210 Steps: 56250, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000906, Sample Num: 14496, Cur Loss: 0.18572670, Cur Avg Loss: 0.25736741, Log Avg loss: 0.22559598, Global Avg Loss: 0.98886439, Time: 0.0210 Steps: 56260, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000916, Sample Num: 14656, Cur Loss: 0.60684800, Cur Avg Loss: 0.25766264, Log Avg loss: 0.28441065, Global Avg Loss: 0.98873920, Time: 0.0210 Steps: 56270, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000926, Sample Num: 14816, Cur Loss: 0.63877875, Cur Avg Loss: 0.25788430, Log Avg loss: 0.27818781, Global Avg Loss: 0.98861295, Time: 0.0210 Steps: 56280, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000936, Sample Num: 14976, Cur Loss: 0.13144898, Cur Avg Loss: 0.25805026, Log Avg loss: 0.27341853, Global Avg Loss: 0.98848589, Time: 0.0212 Steps: 56290, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000946, Sample Num: 15136, Cur Loss: 0.23352861, Cur Avg Loss: 0.25874921, Log Avg loss: 0.32417129, Global Avg Loss: 0.98836789, Time: 0.0210 Steps: 56300, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000956, Sample Num: 15296, Cur Loss: 0.55027497, Cur Avg Loss: 0.25876919, Log Avg loss: 0.26065927, Global Avg Loss: 0.98823866, Time: 0.0210 Steps: 56310, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000966, Sample Num: 15456, Cur Loss: 0.18395713, Cur Avg Loss: 0.25855744, Log Avg loss: 0.23831358, Global Avg Loss: 0.98810551, Time: 0.0210 Steps: 56320, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000976, Sample Num: 15616, Cur Loss: 0.43867522, Cur Avg Loss: 0.25956296, Log Avg loss: 0.35669667, Global Avg Loss: 0.98799342, Time: 0.0210 Steps: 56330, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000986, Sample Num: 15776, Cur Loss: 0.71564096, Cur Avg Loss: 0.26359182, Log Avg loss: 0.65680839, Global Avg Loss: 0.98793463, Time: 0.0210 Steps: 56340, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000996, Sample Num: 15936, Cur Loss: 0.48807403, Cur Avg Loss: 0.26545101, Log Avg loss: 0.44876699, Global Avg Loss: 0.98783895, Time: 0.0210 Steps: 56350, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001006, Sample Num: 16096, Cur Loss: 0.48250806, Cur Avg Loss: 0.26610544, Log Avg loss: 0.33128655, Global Avg Loss: 0.98772246, Time: 0.0210 Steps: 56360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001016, Sample Num: 16256, Cur Loss: 0.44357055, Cur Avg Loss: 0.26670325, Log Avg loss: 0.32684277, Global Avg Loss: 0.98760522, Time: 0.0210 Steps: 56370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001026, Sample Num: 16416, Cur Loss: 0.24380340, Cur Avg Loss: 0.26735722, Log Avg loss: 0.33380142, Global Avg Loss: 0.98748926, Time: 0.0246 Steps: 56380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001036, Sample Num: 16576, Cur Loss: 0.38658032, Cur Avg Loss: 0.26794660, Log Avg loss: 0.32841615, Global Avg Loss: 0.98737238, Time: 0.0208 Steps: 56390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001046, Sample Num: 16736, Cur Loss: 0.04723914, Cur Avg Loss: 0.26765859, Log Avg loss: 0.23782106, Global Avg Loss: 0.98723948, Time: 0.0208 Steps: 56400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001056, Sample Num: 16896, Cur Loss: 0.09934303, Cur Avg Loss: 0.26755397, Log Avg loss: 0.25661076, Global Avg Loss: 0.98710996, Time: 0.0208 Steps: 56410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001066, Sample Num: 17056, Cur Loss: 0.13541341, Cur Avg Loss: 0.26684064, Log Avg loss: 0.19151357, Global Avg Loss: 0.98696894, Time: 0.0208 Steps: 56420, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001076, Sample Num: 17216, Cur Loss: 0.17454010, Cur Avg Loss: 0.26727058, Log Avg loss: 0.31310155, Global Avg Loss: 0.98684953, Time: 0.0208 Steps: 56430, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001086, Sample Num: 17376, Cur Loss: 0.62851262, Cur Avg Loss: 0.26758621, Log Avg loss: 0.30154862, Global Avg Loss: 0.98672811, Time: 0.0208 Steps: 56440, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001096, Sample Num: 17536, Cur Loss: 0.53972697, Cur Avg Loss: 0.26765462, Log Avg loss: 0.27508355, Global Avg Loss: 0.98660204, Time: 0.0208 Steps: 56450, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001106, Sample Num: 17696, Cur Loss: 0.15275431, Cur Avg Loss: 0.26872815, Log Avg loss: 0.38638732, Global Avg Loss: 0.98649573, Time: 0.0208 Steps: 56460, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001116, Sample Num: 17856, Cur Loss: 0.22153804, Cur Avg Loss: 0.26836180, Log Avg loss: 0.22784277, Global Avg Loss: 0.98636139, Time: 0.0208 Steps: 56470, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001126, Sample Num: 18016, Cur Loss: 0.20148438, Cur Avg Loss: 0.26849920, Log Avg loss: 0.28383336, Global Avg Loss: 0.98623700, Time: 0.0208 Steps: 56480, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001136, Sample Num: 18176, Cur Loss: 0.17309800, Cur Avg Loss: 0.26786919, Log Avg loss: 0.19692966, Global Avg Loss: 0.98609728, Time: 0.0208 Steps: 56490, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001146, Sample Num: 18336, Cur Loss: 0.10147734, Cur Avg Loss: 0.26770243, Log Avg loss: 0.24875944, Global Avg Loss: 0.98596677, Time: 0.0208 Steps: 56500, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001156, Sample Num: 18496, Cur Loss: 0.23073721, Cur Avg Loss: 0.26718138, Log Avg loss: 0.20746818, Global Avg Loss: 0.98582901, Time: 0.0208 Steps: 56510, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001166, Sample Num: 18656, Cur Loss: 0.15645936, Cur Avg Loss: 0.26663996, Log Avg loss: 0.20405250, Global Avg Loss: 0.98569069, Time: 0.0208 Steps: 56520, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001176, Sample Num: 18816, Cur Loss: 0.31190997, Cur Avg Loss: 0.26776756, Log Avg loss: 0.39924521, Global Avg Loss: 0.98558695, Time: 0.0208 Steps: 56530, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001186, Sample Num: 18976, Cur Loss: 0.21305941, Cur Avg Loss: 0.26825584, Log Avg loss: 0.32567715, Global Avg Loss: 0.98547024, Time: 0.0208 Steps: 56540, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001196, Sample Num: 19136, Cur Loss: 0.06010055, Cur Avg Loss: 0.26888013, Log Avg loss: 0.34292129, Global Avg Loss: 0.98535661, Time: 0.0209 Steps: 56550, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001206, Sample Num: 19296, Cur Loss: 0.34744567, Cur Avg Loss: 0.26895723, Log Avg loss: 0.27817813, Global Avg Loss: 0.98523158, Time: 0.0208 Steps: 56560, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001216, Sample Num: 19456, Cur Loss: 0.27591625, Cur Avg Loss: 0.26893746, Log Avg loss: 0.26655314, Global Avg Loss: 0.98510454, Time: 0.0208 Steps: 56570, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001226, Sample Num: 19616, Cur Loss: 0.30606750, Cur Avg Loss: 0.26895758, Log Avg loss: 0.27140519, Global Avg Loss: 0.98497840, Time: 0.0208 Steps: 56580, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001236, Sample Num: 19776, Cur Loss: 0.11573094, Cur Avg Loss: 0.26885501, Log Avg loss: 0.25627930, Global Avg Loss: 0.98484963, Time: 0.0208 Steps: 56590, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001246, Sample Num: 19936, Cur Loss: 0.18686996, Cur Avg Loss: 0.26929121, Log Avg loss: 0.32320524, Global Avg Loss: 0.98473273, Time: 0.0208 Steps: 56600, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001256, Sample Num: 20096, Cur Loss: 0.11856925, Cur Avg Loss: 0.26935905, Log Avg loss: 0.27781222, Global Avg Loss: 0.98460786, Time: 0.0208 Steps: 56610, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001266, Sample Num: 20256, Cur Loss: 0.30138591, Cur Avg Loss: 0.26999120, Log Avg loss: 0.34938940, Global Avg Loss: 0.98449567, Time: 0.0208 Steps: 56620, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001276, Sample Num: 20416, Cur Loss: 0.19826938, Cur Avg Loss: 0.26917864, Log Avg loss: 0.16630788, Global Avg Loss: 0.98435119, Time: 0.0208 Steps: 56630, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001286, Sample Num: 20576, Cur Loss: 0.20250776, Cur Avg Loss: 0.26866127, Log Avg loss: 0.20264609, Global Avg Loss: 0.98421317, Time: 0.0245 Steps: 56640, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001296, Sample Num: 20736, Cur Loss: 0.14436336, Cur Avg Loss: 0.26830932, Log Avg loss: 0.22304746, Global Avg Loss: 0.98407881, Time: 0.0208 Steps: 56650, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001306, Sample Num: 20896, Cur Loss: 0.11783148, Cur Avg Loss: 0.26782134, Log Avg loss: 0.20458010, Global Avg Loss: 0.98394124, Time: 0.0207 Steps: 56660, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001316, Sample Num: 21056, Cur Loss: 0.41045415, Cur Avg Loss: 0.26860887, Log Avg loss: 0.37145939, Global Avg Loss: 0.98383316, Time: 0.0209 Steps: 56670, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001326, Sample Num: 21216, Cur Loss: 0.26566079, Cur Avg Loss: 0.26997648, Log Avg loss: 0.44995480, Global Avg Loss: 0.98373897, Time: 0.0207 Steps: 56680, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001336, Sample Num: 21376, Cur Loss: 0.51595795, Cur Avg Loss: 0.27071198, Log Avg loss: 0.36823941, Global Avg Loss: 0.98363039, Time: 0.0208 Steps: 56690, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001346, Sample Num: 21536, Cur Loss: 0.11862348, Cur Avg Loss: 0.27036846, Log Avg loss: 0.22447325, Global Avg Loss: 0.98349650, Time: 0.0208 Steps: 56700, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001356, Sample Num: 21696, Cur Loss: 0.09579670, Cur Avg Loss: 0.26986441, Log Avg loss: 0.20201998, Global Avg Loss: 0.98335870, Time: 0.0208 Steps: 56710, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001366, Sample Num: 21856, Cur Loss: 0.18389621, Cur Avg Loss: 0.26981771, Log Avg loss: 0.26348516, Global Avg Loss: 0.98323178, Time: 0.0207 Steps: 56720, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001376, Sample Num: 22016, Cur Loss: 0.16130382, Cur Avg Loss: 0.26997876, Log Avg loss: 0.29197734, Global Avg Loss: 0.98310993, Time: 0.0208 Steps: 56730, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001386, Sample Num: 22176, Cur Loss: 0.25102696, Cur Avg Loss: 0.27002415, Log Avg loss: 0.27627070, Global Avg Loss: 0.98298536, Time: 0.0208 Steps: 56740, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001396, Sample Num: 22336, Cur Loss: 0.15701108, Cur Avg Loss: 0.27007782, Log Avg loss: 0.27751601, Global Avg Loss: 0.98286105, Time: 0.0208 Steps: 56750, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001406, Sample Num: 22496, Cur Loss: 0.21270679, Cur Avg Loss: 0.26999486, Log Avg loss: 0.25841333, Global Avg Loss: 0.98273341, Time: 0.0207 Steps: 56760, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001416, Sample Num: 22656, Cur Loss: 0.13514403, Cur Avg Loss: 0.26973847, Log Avg loss: 0.23369023, Global Avg Loss: 0.98260147, Time: 0.0207 Steps: 56770, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001426, Sample Num: 22816, Cur Loss: 0.20044516, Cur Avg Loss: 0.27033794, Log Avg loss: 0.35522293, Global Avg Loss: 0.98249098, Time: 0.0208 Steps: 56780, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001436, Sample Num: 22976, Cur Loss: 0.14378856, Cur Avg Loss: 0.27041262, Log Avg loss: 0.28106205, Global Avg Loss: 0.98236746, Time: 0.0207 Steps: 56790, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001446, Sample Num: 23136, Cur Loss: 1.05618703, Cur Avg Loss: 0.27274309, Log Avg loss: 0.60739940, Global Avg Loss: 0.98230145, Time: 0.0208 Steps: 56800, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001456, Sample Num: 23296, Cur Loss: 0.13599873, Cur Avg Loss: 0.27308365, Log Avg loss: 0.32232787, Global Avg Loss: 0.98218528, Time: 0.0207 Steps: 56810, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001466, Sample Num: 23456, Cur Loss: 0.34790039, Cur Avg Loss: 0.27292843, Log Avg loss: 0.25032885, Global Avg Loss: 0.98205647, Time: 0.0208 Steps: 56820, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001476, Sample Num: 23616, Cur Loss: 0.08822083, Cur Avg Loss: 0.27256130, Log Avg loss: 0.21873923, Global Avg Loss: 0.98192216, Time: 0.0208 Steps: 56830, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001486, Sample Num: 23776, Cur Loss: 0.14069243, Cur Avg Loss: 0.27216214, Log Avg loss: 0.21324598, Global Avg Loss: 0.98178692, Time: 0.0208 Steps: 56840, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001496, Sample Num: 23936, Cur Loss: 0.26469582, Cur Avg Loss: 0.27202296, Log Avg loss: 0.25134164, Global Avg Loss: 0.98165844, Time: 0.0208 Steps: 56850, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001506, Sample Num: 24096, Cur Loss: 0.26546353, Cur Avg Loss: 0.27184451, Log Avg loss: 0.24514804, Global Avg Loss: 0.98152891, Time: 0.0208 Steps: 56860, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001516, Sample Num: 24256, Cur Loss: 0.31830311, Cur Avg Loss: 0.27179422, Log Avg loss: 0.26422095, Global Avg Loss: 0.98140277, Time: 0.0208 Steps: 56870, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001526, Sample Num: 24416, Cur Loss: 0.32180214, Cur Avg Loss: 0.27178399, Log Avg loss: 0.27023334, Global Avg Loss: 0.98127774, Time: 0.0208 Steps: 56880, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001536, Sample Num: 24576, Cur Loss: 0.39564830, Cur Avg Loss: 0.27276128, Log Avg loss: 0.42189541, Global Avg Loss: 0.98117942, Time: 0.0253 Steps: 56890, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001546, Sample Num: 24736, Cur Loss: 0.24531111, Cur Avg Loss: 0.27267120, Log Avg loss: 0.25883517, Global Avg Loss: 0.98105247, Time: 0.0209 Steps: 56900, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001556, Sample Num: 24896, Cur Loss: 0.13087408, Cur Avg Loss: 0.27240133, Log Avg loss: 0.23067831, Global Avg Loss: 0.98092061, Time: 0.0209 Steps: 56910, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001566, Sample Num: 25056, Cur Loss: 0.10277981, Cur Avg Loss: 0.27182770, Log Avg loss: 0.18257216, Global Avg Loss: 0.98078036, Time: 0.0209 Steps: 56920, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001576, Sample Num: 25216, Cur Loss: 0.20591676, Cur Avg Loss: 0.27134935, Log Avg loss: 0.19643900, Global Avg Loss: 0.98064258, Time: 0.0209 Steps: 56930, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001586, Sample Num: 25376, Cur Loss: 0.39588562, Cur Avg Loss: 0.27144873, Log Avg loss: 0.28711154, Global Avg Loss: 0.98052078, Time: 0.0209 Steps: 56940, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001596, Sample Num: 25536, Cur Loss: 0.28273207, Cur Avg Loss: 0.27132085, Log Avg loss: 0.25103861, Global Avg Loss: 0.98039269, Time: 0.0209 Steps: 56950, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001606, Sample Num: 25696, Cur Loss: 0.50912458, Cur Avg Loss: 0.27151259, Log Avg loss: 0.30211492, Global Avg Loss: 0.98027361, Time: 0.0209 Steps: 56960, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001616, Sample Num: 25856, Cur Loss: 0.34996754, Cur Avg Loss: 0.27143862, Log Avg loss: 0.25955801, Global Avg Loss: 0.98014710, Time: 0.0208 Steps: 56970, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001626, Sample Num: 26016, Cur Loss: 0.16716626, Cur Avg Loss: 0.27127559, Log Avg loss: 0.24492961, Global Avg Loss: 0.98001807, Time: 0.0209 Steps: 56980, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001636, Sample Num: 26176, Cur Loss: 0.30176759, Cur Avg Loss: 0.27097850, Log Avg loss: 0.22267254, Global Avg Loss: 0.97988518, Time: 0.0208 Steps: 56990, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001646, Sample Num: 26336, Cur Loss: 0.17857032, Cur Avg Loss: 0.27060723, Log Avg loss: 0.20986692, Global Avg Loss: 0.97975009, Time: 0.0209 Steps: 57000, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001656, Sample Num: 26496, Cur Loss: 0.18219696, Cur Avg Loss: 0.27082261, Log Avg loss: 0.30627426, Global Avg Loss: 0.97963196, Time: 0.0209 Steps: 57010, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001666, Sample Num: 26656, Cur Loss: 0.21186748, Cur Avg Loss: 0.27088358, Log Avg loss: 0.28097994, Global Avg Loss: 0.97950943, Time: 0.0209 Steps: 57020, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001676, Sample Num: 26816, Cur Loss: 0.22969580, Cur Avg Loss: 0.27031478, Log Avg loss: 0.17555374, Global Avg Loss: 0.97936846, Time: 0.0209 Steps: 57030, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001686, Sample Num: 26976, Cur Loss: 0.06682809, Cur Avg Loss: 0.26994597, Log Avg loss: 0.20813252, Global Avg Loss: 0.97923325, Time: 0.0209 Steps: 57040, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001696, Sample Num: 27136, Cur Loss: 0.18550475, Cur Avg Loss: 0.26964493, Log Avg loss: 0.21888930, Global Avg Loss: 0.97909997, Time: 0.0209 Steps: 57050, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001706, Sample Num: 27296, Cur Loss: 0.21190336, Cur Avg Loss: 0.26961408, Log Avg loss: 0.26438262, Global Avg Loss: 0.97897472, Time: 0.0208 Steps: 57060, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001716, Sample Num: 27456, Cur Loss: 0.41457373, Cur Avg Loss: 0.27014736, Log Avg loss: 0.36112414, Global Avg Loss: 0.97886646, Time: 0.0208 Steps: 57070, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001726, Sample Num: 27616, Cur Loss: 0.36658666, Cur Avg Loss: 0.27042104, Log Avg loss: 0.31738501, Global Avg Loss: 0.97875057, Time: 0.0208 Steps: 57080, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001736, Sample Num: 27776, Cur Loss: 0.58062619, Cur Avg Loss: 0.27112829, Log Avg loss: 0.39320030, Global Avg Loss: 0.97864800, Time: 0.0208 Steps: 57090, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001746, Sample Num: 27936, Cur Loss: 0.25831389, Cur Avg Loss: 0.27166607, Log Avg loss: 0.36502468, Global Avg Loss: 0.97854054, Time: 0.0208 Steps: 57100, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001756, Sample Num: 28096, Cur Loss: 0.08522969, Cur Avg Loss: 0.27141212, Log Avg loss: 0.22707251, Global Avg Loss: 0.97840895, Time: 0.0209 Steps: 57110, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001766, Sample Num: 28256, Cur Loss: 0.24878542, Cur Avg Loss: 0.27155534, Log Avg loss: 0.29670371, Global Avg Loss: 0.97828961, Time: 0.0208 Steps: 57120, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001776, Sample Num: 28416, Cur Loss: 0.18943299, Cur Avg Loss: 0.27114289, Log Avg loss: 0.19830536, Global Avg Loss: 0.97815308, Time: 0.0209 Steps: 57130, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001786, Sample Num: 28576, Cur Loss: 0.09003510, Cur Avg Loss: 0.27113357, Log Avg loss: 0.26947839, Global Avg Loss: 0.97802906, Time: 0.0209 Steps: 57140, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001796, Sample Num: 28736, Cur Loss: 0.15338193, Cur Avg Loss: 0.27114296, Log Avg loss: 0.27281975, Global Avg Loss: 0.97790566, Time: 0.0246 Steps: 57150, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001806, Sample Num: 28896, Cur Loss: 0.43406573, Cur Avg Loss: 0.27098524, Log Avg loss: 0.24265886, Global Avg Loss: 0.97777703, Time: 0.0209 Steps: 57160, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001816, Sample Num: 29056, Cur Loss: 0.19072427, Cur Avg Loss: 0.27086688, Log Avg loss: 0.24948996, Global Avg Loss: 0.97764964, Time: 0.0219 Steps: 57170, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001826, Sample Num: 29216, Cur Loss: 0.40064609, Cur Avg Loss: 0.27140531, Log Avg loss: 0.36918541, Global Avg Loss: 0.97754323, Time: 0.0219 Steps: 57180, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001836, Sample Num: 29376, Cur Loss: 0.26906604, Cur Avg Loss: 0.27189079, Log Avg loss: 0.36053923, Global Avg Loss: 0.97743534, Time: 0.0219 Steps: 57190, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001846, Sample Num: 29536, Cur Loss: 0.08638114, Cur Avg Loss: 0.27165574, Log Avg loss: 0.22850030, Global Avg Loss: 0.97730441, Time: 0.0219 Steps: 57200, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001856, Sample Num: 29696, Cur Loss: 0.36082605, Cur Avg Loss: 0.27143220, Log Avg loss: 0.23016735, Global Avg Loss: 0.97717381, Time: 0.0219 Steps: 57210, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001866, Sample Num: 29856, Cur Loss: 0.22169185, Cur Avg Loss: 0.27144155, Log Avg loss: 0.27317571, Global Avg Loss: 0.97705078, Time: 0.0219 Steps: 57220, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001876, Sample Num: 30016, Cur Loss: 0.09903250, Cur Avg Loss: 0.27159116, Log Avg loss: 0.29950892, Global Avg Loss: 0.97693239, Time: 0.0208 Steps: 57230, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001886, Sample Num: 30176, Cur Loss: 0.62366915, Cur Avg Loss: 0.27155320, Log Avg loss: 0.26443272, Global Avg Loss: 0.97680792, Time: 0.0209 Steps: 57240, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001896, Sample Num: 30336, Cur Loss: 0.50720060, Cur Avg Loss: 0.27216572, Log Avg loss: 0.38768691, Global Avg Loss: 0.97670501, Time: 0.0210 Steps: 57250, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001906, Sample Num: 30496, Cur Loss: 0.95883441, Cur Avg Loss: 0.27253722, Log Avg loss: 0.34297243, Global Avg Loss: 0.97659434, Time: 0.0208 Steps: 57260, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001916, Sample Num: 30656, Cur Loss: 0.17267917, Cur Avg Loss: 0.27235351, Log Avg loss: 0.23733893, Global Avg Loss: 0.97646525, Time: 0.0210 Steps: 57270, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001926, Sample Num: 30816, Cur Loss: 0.19796444, Cur Avg Loss: 0.27215295, Log Avg loss: 0.23372642, Global Avg Loss: 0.97633559, Time: 0.0210 Steps: 57280, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001936, Sample Num: 30976, Cur Loss: 0.21395503, Cur Avg Loss: 0.27200313, Log Avg loss: 0.24314786, Global Avg Loss: 0.97620761, Time: 0.0208 Steps: 57290, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001946, Sample Num: 31136, Cur Loss: 0.38732910, Cur Avg Loss: 0.27210061, Log Avg loss: 0.29097277, Global Avg Loss: 0.97608802, Time: 0.0208 Steps: 57300, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001956, Sample Num: 31296, Cur Loss: 0.16934755, Cur Avg Loss: 0.27188976, Log Avg loss: 0.23085699, Global Avg Loss: 0.97595798, Time: 0.0208 Steps: 57310, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001966, Sample Num: 31456, Cur Loss: 0.19688229, Cur Avg Loss: 0.27160273, Log Avg loss: 0.21546044, Global Avg Loss: 0.97582531, Time: 0.0219 Steps: 57320, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001976, Sample Num: 31616, Cur Loss: 0.24770410, Cur Avg Loss: 0.27150443, Log Avg loss: 0.25217788, Global Avg Loss: 0.97569908, Time: 0.0219 Steps: 57330, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001986, Sample Num: 31776, Cur Loss: 0.13944817, Cur Avg Loss: 0.27180174, Log Avg loss: 0.33055143, Global Avg Loss: 0.97558657, Time: 0.0219 Steps: 57340, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001996, Sample Num: 31936, Cur Loss: 0.28291571, Cur Avg Loss: 0.27150310, Log Avg loss: 0.21219296, Global Avg Loss: 0.97545346, Time: 0.0219 Steps: 57350, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002006, Sample Num: 32096, Cur Loss: 0.12184028, Cur Avg Loss: 0.27125535, Log Avg loss: 0.22180305, Global Avg Loss: 0.97532207, Time: 0.0220 Steps: 57360, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002016, Sample Num: 32256, Cur Loss: 0.03412033, Cur Avg Loss: 0.27086873, Log Avg loss: 0.19331457, Global Avg Loss: 0.97518576, Time: 0.0208 Steps: 57370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002026, Sample Num: 32416, Cur Loss: 0.17465103, Cur Avg Loss: 0.27089487, Log Avg loss: 0.27616302, Global Avg Loss: 0.97506394, Time: 0.0210 Steps: 57380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002036, Sample Num: 32576, Cur Loss: 0.33343899, Cur Avg Loss: 0.27130006, Log Avg loss: 0.35339204, Global Avg Loss: 0.97495561, Time: 0.0209 Steps: 57390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002046, Sample Num: 32736, Cur Loss: 0.08661738, Cur Avg Loss: 0.27131035, Log Avg loss: 0.27340583, Global Avg Loss: 0.97483339, Time: 0.0219 Steps: 57400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002056, Sample Num: 32896, Cur Loss: 0.17357951, Cur Avg Loss: 0.27083540, Log Avg loss: 0.17365983, Global Avg Loss: 0.97469384, Time: 0.0209 Steps: 57410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002066, Sample Num: 33056, Cur Loss: 0.11829130, Cur Avg Loss: 0.27071283, Log Avg loss: 0.24551271, Global Avg Loss: 0.97456685, Time: 0.0208 Steps: 57420, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002076, Sample Num: 33216, Cur Loss: 0.29482284, Cur Avg Loss: 0.27070154, Log Avg loss: 0.26836844, Global Avg Loss: 0.97444388, Time: 0.0208 Steps: 57430, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002086, Sample Num: 33376, Cur Loss: 0.19114330, Cur Avg Loss: 0.27055147, Log Avg loss: 0.23939814, Global Avg Loss: 0.97431591, Time: 0.0208 Steps: 57440, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002096, Sample Num: 33536, Cur Loss: 0.07488772, Cur Avg Loss: 0.27036172, Log Avg loss: 0.23077877, Global Avg Loss: 0.97418649, Time: 0.0208 Steps: 57450, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002106, Sample Num: 33696, Cur Loss: 0.32655010, Cur Avg Loss: 0.27042532, Log Avg loss: 0.28375661, Global Avg Loss: 0.97406633, Time: 0.0208 Steps: 57460, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002116, Sample Num: 33856, Cur Loss: 0.12962650, Cur Avg Loss: 0.27041568, Log Avg loss: 0.26838600, Global Avg Loss: 0.97394354, Time: 0.0207 Steps: 57470, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002126, Sample Num: 34016, Cur Loss: 0.20299765, Cur Avg Loss: 0.27051662, Log Avg loss: 0.29187604, Global Avg Loss: 0.97382488, Time: 0.0207 Steps: 57480, Updated lr: 0.000046 ***** Running evaluation checkpoint-57483 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-57483 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.815315, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.464779, "eval_total_loss": 326.73935, "eval_mae": 0.465715, "eval_mse": 0.464969, "eval_r2": 0.704435, "eval_sp_statistic": 0.88792, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.903058, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.689534, "test_total_loss": 346.146199, "test_mae": 0.560698, "test_mse": 0.689753, "test_r2": 0.554827, "test_sp_statistic": 0.857668, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.898301, "test_ps_pvalue": 0.0, "lr": 4.6436225699383594e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.973794013959813, "train_cur_epoch_loss": 576.2655733525753, "train_cur_epoch_avg_loss": 0.27067429467006826, "train_cur_epoch_time": 44.81531524658203, "train_cur_epoch_avg_time": 0.02104993670576892, "epoch": 27, "step": 57483} ################################################## Training, Epoch: 0028, Batch: 000007, Sample Num: 112, Cur Loss: 0.47438052, Cur Avg Loss: 0.42513791, Log Avg loss: 0.41231951, Global Avg Loss: 0.97372721, Time: 0.0211 Steps: 57490, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000017, Sample Num: 272, Cur Loss: 0.10800740, Cur Avg Loss: 0.36434861, Log Avg loss: 0.32179611, Global Avg Loss: 0.97361383, Time: 0.0208 Steps: 57500, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000027, Sample Num: 432, Cur Loss: 0.05725016, Cur Avg Loss: 0.33987732, Log Avg loss: 0.29827611, Global Avg Loss: 0.97349640, Time: 0.0208 Steps: 57510, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000037, Sample Num: 592, Cur Loss: 0.23927680, Cur Avg Loss: 0.31025939, Log Avg loss: 0.23029099, Global Avg Loss: 0.97336719, Time: 0.0208 Steps: 57520, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000047, Sample Num: 752, Cur Loss: 0.19474876, Cur Avg Loss: 0.29176968, Log Avg loss: 0.22335773, Global Avg Loss: 0.97323682, Time: 0.0207 Steps: 57530, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000057, Sample Num: 912, Cur Loss: 0.21458188, Cur Avg Loss: 0.27862789, Log Avg loss: 0.21686152, Global Avg Loss: 0.97310537, Time: 0.0207 Steps: 57540, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000067, Sample Num: 1072, Cur Loss: 0.63542777, Cur Avg Loss: 0.30310977, Log Avg loss: 0.44265644, Global Avg Loss: 0.97301320, Time: 0.0207 Steps: 57550, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000077, Sample Num: 1232, Cur Loss: 0.12904228, Cur Avg Loss: 0.28826273, Log Avg loss: 0.18878758, Global Avg Loss: 0.97287696, Time: 0.0207 Steps: 57560, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000087, Sample Num: 1392, Cur Loss: 0.25427598, Cur Avg Loss: 0.27556272, Log Avg loss: 0.17777261, Global Avg Loss: 0.97273884, Time: 0.0207 Steps: 57570, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000097, Sample Num: 1552, Cur Loss: 0.16555326, Cur Avg Loss: 0.26643648, Log Avg loss: 0.18703819, Global Avg Loss: 0.97260239, Time: 0.0207 Steps: 57580, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000107, Sample Num: 1712, Cur Loss: 0.18330732, Cur Avg Loss: 0.25918305, Log Avg loss: 0.18882482, Global Avg Loss: 0.97246629, Time: 0.0208 Steps: 57590, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000117, Sample Num: 1872, Cur Loss: 0.37273479, Cur Avg Loss: 0.25330348, Log Avg loss: 0.19039209, Global Avg Loss: 0.97233052, Time: 0.0208 Steps: 57600, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000127, Sample Num: 2032, Cur Loss: 0.45147118, Cur Avg Loss: 0.26630243, Log Avg loss: 0.41839018, Global Avg Loss: 0.97223436, Time: 0.0208 Steps: 57610, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000137, Sample Num: 2192, Cur Loss: 0.10647193, Cur Avg Loss: 0.26813314, Log Avg loss: 0.29138313, Global Avg Loss: 0.97211620, Time: 0.0207 Steps: 57620, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000147, Sample Num: 2352, Cur Loss: 0.58770311, Cur Avg Loss: 0.26925593, Log Avg loss: 0.28463814, Global Avg Loss: 0.97199691, Time: 0.0207 Steps: 57630, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000157, Sample Num: 2512, Cur Loss: 0.77352631, Cur Avg Loss: 0.27806986, Log Avg loss: 0.40763463, Global Avg Loss: 0.97189900, Time: 0.0207 Steps: 57640, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000167, Sample Num: 2672, Cur Loss: 0.53166401, Cur Avg Loss: 0.27940607, Log Avg loss: 0.30038449, Global Avg Loss: 0.97178252, Time: 0.0208 Steps: 57650, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000177, Sample Num: 2832, Cur Loss: 0.19007450, Cur Avg Loss: 0.27949097, Log Avg loss: 0.28090884, Global Avg Loss: 0.97166270, Time: 0.0208 Steps: 57660, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000187, Sample Num: 2992, Cur Loss: 0.30076879, Cur Avg Loss: 0.28071333, Log Avg loss: 0.30234906, Global Avg Loss: 0.97154664, Time: 0.0208 Steps: 57670, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000197, Sample Num: 3152, Cur Loss: 0.27634344, Cur Avg Loss: 0.28209936, Log Avg loss: 0.30801827, Global Avg Loss: 0.97143160, Time: 0.0208 Steps: 57680, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000207, Sample Num: 3312, Cur Loss: 0.69294238, Cur Avg Loss: 0.28443488, Log Avg loss: 0.33044445, Global Avg Loss: 0.97132049, Time: 0.0207 Steps: 57690, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000217, Sample Num: 3472, Cur Loss: 0.17717418, Cur Avg Loss: 0.28028767, Log Avg loss: 0.19444056, Global Avg Loss: 0.97118585, Time: 0.0207 Steps: 57700, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000227, Sample Num: 3632, Cur Loss: 0.12487605, Cur Avg Loss: 0.27681185, Log Avg loss: 0.20138658, Global Avg Loss: 0.97105246, Time: 0.0208 Steps: 57710, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000237, Sample Num: 3792, Cur Loss: 0.19736955, Cur Avg Loss: 0.27530276, Log Avg loss: 0.24104643, Global Avg Loss: 0.97092599, Time: 0.0208 Steps: 57720, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000247, Sample Num: 3952, Cur Loss: 0.12828827, Cur Avg Loss: 0.27348782, Log Avg loss: 0.23047361, Global Avg Loss: 0.97079773, Time: 0.0208 Steps: 57730, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000257, Sample Num: 4112, Cur Loss: 0.50051844, Cur Avg Loss: 0.27074493, Log Avg loss: 0.20299562, Global Avg Loss: 0.97066475, Time: 0.0244 Steps: 57740, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000267, Sample Num: 4272, Cur Loss: 0.15942788, Cur Avg Loss: 0.27233809, Log Avg loss: 0.31328225, Global Avg Loss: 0.97055092, Time: 0.0208 Steps: 57750, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000277, Sample Num: 4432, Cur Loss: 0.46559289, Cur Avg Loss: 0.27557347, Log Avg loss: 0.36195803, Global Avg Loss: 0.97044555, Time: 0.0208 Steps: 57760, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000287, Sample Num: 4592, Cur Loss: 0.19596705, Cur Avg Loss: 0.27233082, Log Avg loss: 0.18250964, Global Avg Loss: 0.97030916, Time: 0.0209 Steps: 57770, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000297, Sample Num: 4752, Cur Loss: 0.10543418, Cur Avg Loss: 0.26965632, Log Avg loss: 0.19289799, Global Avg Loss: 0.97017461, Time: 0.0209 Steps: 57780, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000307, Sample Num: 4912, Cur Loss: 0.30676144, Cur Avg Loss: 0.26933257, Log Avg loss: 0.25971733, Global Avg Loss: 0.97005168, Time: 0.0209 Steps: 57790, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000317, Sample Num: 5072, Cur Loss: 0.11212733, Cur Avg Loss: 0.27154378, Log Avg loss: 0.33942772, Global Avg Loss: 0.96994257, Time: 0.0208 Steps: 57800, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000327, Sample Num: 5232, Cur Loss: 0.19720021, Cur Avg Loss: 0.27220672, Log Avg loss: 0.29322190, Global Avg Loss: 0.96982551, Time: 0.0208 Steps: 57810, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000337, Sample Num: 5392, Cur Loss: 0.35862172, Cur Avg Loss: 0.27314780, Log Avg loss: 0.30392112, Global Avg Loss: 0.96971034, Time: 0.0208 Steps: 57820, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000347, Sample Num: 5552, Cur Loss: 0.37990266, Cur Avg Loss: 0.27404327, Log Avg loss: 0.30422073, Global Avg Loss: 0.96959527, Time: 0.0207 Steps: 57830, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000357, Sample Num: 5712, Cur Loss: 0.48175055, Cur Avg Loss: 0.27635050, Log Avg loss: 0.35641149, Global Avg Loss: 0.96948925, Time: 0.0208 Steps: 57840, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000367, Sample Num: 5872, Cur Loss: 0.56483811, Cur Avg Loss: 0.27648074, Log Avg loss: 0.28113034, Global Avg Loss: 0.96937026, Time: 0.0208 Steps: 57850, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000377, Sample Num: 6032, Cur Loss: 0.24856937, Cur Avg Loss: 0.27489868, Log Avg loss: 0.21683694, Global Avg Loss: 0.96924020, Time: 0.0208 Steps: 57860, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000387, Sample Num: 6192, Cur Loss: 0.21825026, Cur Avg Loss: 0.27374246, Log Avg loss: 0.23015306, Global Avg Loss: 0.96911249, Time: 0.0209 Steps: 57870, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000397, Sample Num: 6352, Cur Loss: 0.03576187, Cur Avg Loss: 0.27124628, Log Avg loss: 0.17464394, Global Avg Loss: 0.96897523, Time: 0.0208 Steps: 57880, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000407, Sample Num: 6512, Cur Loss: 0.05709140, Cur Avg Loss: 0.26921945, Log Avg loss: 0.18875451, Global Avg Loss: 0.96884045, Time: 0.0208 Steps: 57890, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000417, Sample Num: 6672, Cur Loss: 0.43064108, Cur Avg Loss: 0.26891908, Log Avg loss: 0.25669406, Global Avg Loss: 0.96871745, Time: 0.0208 Steps: 57900, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000427, Sample Num: 6832, Cur Loss: 0.28923351, Cur Avg Loss: 0.26798918, Log Avg loss: 0.22921201, Global Avg Loss: 0.96858975, Time: 0.0208 Steps: 57910, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000437, Sample Num: 6992, Cur Loss: 0.08020780, Cur Avg Loss: 0.26698022, Log Avg loss: 0.22389771, Global Avg Loss: 0.96846118, Time: 0.0208 Steps: 57920, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000447, Sample Num: 7152, Cur Loss: 0.14176211, Cur Avg Loss: 0.26532625, Log Avg loss: 0.19304778, Global Avg Loss: 0.96832733, Time: 0.0208 Steps: 57930, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000457, Sample Num: 7312, Cur Loss: 0.09107264, Cur Avg Loss: 0.26571910, Log Avg loss: 0.28327943, Global Avg Loss: 0.96820909, Time: 0.0209 Steps: 57940, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000467, Sample Num: 7472, Cur Loss: 0.09216644, Cur Avg Loss: 0.26662125, Log Avg loss: 0.30784966, Global Avg Loss: 0.96809514, Time: 0.0208 Steps: 57950, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000477, Sample Num: 7632, Cur Loss: 0.17420051, Cur Avg Loss: 0.26770862, Log Avg loss: 0.31848902, Global Avg Loss: 0.96798306, Time: 0.0208 Steps: 57960, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000487, Sample Num: 7792, Cur Loss: 0.14434130, Cur Avg Loss: 0.26622574, Log Avg loss: 0.19549202, Global Avg Loss: 0.96784981, Time: 0.0208 Steps: 57970, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000497, Sample Num: 7952, Cur Loss: 0.16167624, Cur Avg Loss: 0.26473120, Log Avg loss: 0.19194743, Global Avg Loss: 0.96771598, Time: 0.0208 Steps: 57980, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000507, Sample Num: 8112, Cur Loss: 0.19102103, Cur Avg Loss: 0.26471114, Log Avg loss: 0.26371416, Global Avg Loss: 0.96759458, Time: 0.0208 Steps: 57990, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000517, Sample Num: 8272, Cur Loss: 0.15893175, Cur Avg Loss: 0.26344125, Log Avg loss: 0.19905753, Global Avg Loss: 0.96746208, Time: 0.0247 Steps: 58000, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000527, Sample Num: 8432, Cur Loss: 0.27393502, Cur Avg Loss: 0.26407748, Log Avg loss: 0.29697051, Global Avg Loss: 0.96734649, Time: 0.0209 Steps: 58010, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000537, Sample Num: 8592, Cur Loss: 0.46948034, Cur Avg Loss: 0.26485013, Log Avg loss: 0.30556915, Global Avg Loss: 0.96723243, Time: 0.0209 Steps: 58020, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000547, Sample Num: 8752, Cur Loss: 0.39613724, Cur Avg Loss: 0.26448971, Log Avg loss: 0.24513467, Global Avg Loss: 0.96710800, Time: 0.0209 Steps: 58030, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000557, Sample Num: 8912, Cur Loss: 0.24123222, Cur Avg Loss: 0.26407447, Log Avg loss: 0.24136098, Global Avg Loss: 0.96698296, Time: 0.0209 Steps: 58040, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000567, Sample Num: 9072, Cur Loss: 0.05443268, Cur Avg Loss: 0.26413054, Log Avg loss: 0.26725402, Global Avg Loss: 0.96686242, Time: 0.0209 Steps: 58050, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000577, Sample Num: 9232, Cur Loss: 0.12765348, Cur Avg Loss: 0.26429648, Log Avg loss: 0.27370476, Global Avg Loss: 0.96674303, Time: 0.0209 Steps: 58060, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000587, Sample Num: 9392, Cur Loss: 0.11790846, Cur Avg Loss: 0.26484987, Log Avg loss: 0.29678055, Global Avg Loss: 0.96662766, Time: 0.0209 Steps: 58070, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000597, Sample Num: 9552, Cur Loss: 0.30026364, Cur Avg Loss: 0.26406720, Log Avg loss: 0.21812440, Global Avg Loss: 0.96649878, Time: 0.0209 Steps: 58080, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000607, Sample Num: 9712, Cur Loss: 0.07697463, Cur Avg Loss: 0.26234429, Log Avg loss: 0.15948678, Global Avg Loss: 0.96635986, Time: 0.0209 Steps: 58090, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000617, Sample Num: 9872, Cur Loss: 0.16517103, Cur Avg Loss: 0.26172204, Log Avg loss: 0.22395126, Global Avg Loss: 0.96623208, Time: 0.0209 Steps: 58100, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000627, Sample Num: 10032, Cur Loss: 0.20867401, Cur Avg Loss: 0.26247082, Log Avg loss: 0.30867095, Global Avg Loss: 0.96611892, Time: 0.0209 Steps: 58110, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000637, Sample Num: 10192, Cur Loss: 0.16144881, Cur Avg Loss: 0.26066888, Log Avg loss: 0.14768703, Global Avg Loss: 0.96597810, Time: 0.0209 Steps: 58120, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000647, Sample Num: 10352, Cur Loss: 0.19447373, Cur Avg Loss: 0.26032565, Log Avg loss: 0.23846222, Global Avg Loss: 0.96585295, Time: 0.0209 Steps: 58130, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000657, Sample Num: 10512, Cur Loss: 0.16121139, Cur Avg Loss: 0.26098719, Log Avg loss: 0.30378843, Global Avg Loss: 0.96573908, Time: 0.0209 Steps: 58140, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000667, Sample Num: 10672, Cur Loss: 0.09851220, Cur Avg Loss: 0.26032646, Log Avg loss: 0.21691679, Global Avg Loss: 0.96561030, Time: 0.0209 Steps: 58150, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000677, Sample Num: 10832, Cur Loss: 0.46845749, Cur Avg Loss: 0.25945661, Log Avg loss: 0.20143774, Global Avg Loss: 0.96547891, Time: 0.0209 Steps: 58160, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000687, Sample Num: 10992, Cur Loss: 0.18443145, Cur Avg Loss: 0.25808164, Log Avg loss: 0.16499574, Global Avg Loss: 0.96534130, Time: 0.0209 Steps: 58170, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000697, Sample Num: 11152, Cur Loss: 0.31394216, Cur Avg Loss: 0.25762232, Log Avg loss: 0.22606732, Global Avg Loss: 0.96521423, Time: 0.0209 Steps: 58180, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000707, Sample Num: 11312, Cur Loss: 0.21729726, Cur Avg Loss: 0.25723168, Log Avg loss: 0.23000374, Global Avg Loss: 0.96508789, Time: 0.0209 Steps: 58190, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000717, Sample Num: 11472, Cur Loss: 0.08779539, Cur Avg Loss: 0.25696249, Log Avg loss: 0.23793065, Global Avg Loss: 0.96496295, Time: 0.0209 Steps: 58200, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000727, Sample Num: 11632, Cur Loss: 0.19698764, Cur Avg Loss: 0.25690241, Log Avg loss: 0.25259518, Global Avg Loss: 0.96484057, Time: 0.0209 Steps: 58210, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000737, Sample Num: 11792, Cur Loss: 0.30642179, Cur Avg Loss: 0.25694870, Log Avg loss: 0.26031343, Global Avg Loss: 0.96471956, Time: 0.0209 Steps: 58220, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000747, Sample Num: 11952, Cur Loss: 0.09349661, Cur Avg Loss: 0.25684856, Log Avg loss: 0.24946869, Global Avg Loss: 0.96459672, Time: 0.0209 Steps: 58230, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000757, Sample Num: 12112, Cur Loss: 0.25018391, Cur Avg Loss: 0.25606524, Log Avg loss: 0.19755144, Global Avg Loss: 0.96446502, Time: 0.0209 Steps: 58240, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000767, Sample Num: 12272, Cur Loss: 0.25762212, Cur Avg Loss: 0.25577389, Log Avg loss: 0.23371851, Global Avg Loss: 0.96433957, Time: 0.0209 Steps: 58250, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000777, Sample Num: 12432, Cur Loss: 0.27486694, Cur Avg Loss: 0.25780978, Log Avg loss: 0.41396216, Global Avg Loss: 0.96424510, Time: 0.0209 Steps: 58260, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000787, Sample Num: 12592, Cur Loss: 0.12602496, Cur Avg Loss: 0.25758405, Log Avg loss: 0.24004528, Global Avg Loss: 0.96412082, Time: 0.0210 Steps: 58270, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000797, Sample Num: 12752, Cur Loss: 0.29300877, Cur Avg Loss: 0.25668032, Log Avg loss: 0.18555652, Global Avg Loss: 0.96398723, Time: 0.0209 Steps: 58280, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000807, Sample Num: 12912, Cur Loss: 0.34126291, Cur Avg Loss: 0.25608976, Log Avg loss: 0.20902203, Global Avg Loss: 0.96385771, Time: 0.0209 Steps: 58290, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000817, Sample Num: 13072, Cur Loss: 0.34620100, Cur Avg Loss: 0.25627218, Log Avg loss: 0.27099328, Global Avg Loss: 0.96373886, Time: 0.0208 Steps: 58300, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000827, Sample Num: 13232, Cur Loss: 0.23484392, Cur Avg Loss: 0.25710771, Log Avg loss: 0.32537071, Global Avg Loss: 0.96362938, Time: 0.0209 Steps: 58310, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000837, Sample Num: 13392, Cur Loss: 0.25869069, Cur Avg Loss: 0.25704128, Log Avg loss: 0.25154793, Global Avg Loss: 0.96350728, Time: 0.0209 Steps: 58320, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000847, Sample Num: 13552, Cur Loss: 0.17179725, Cur Avg Loss: 0.25717643, Log Avg loss: 0.26848824, Global Avg Loss: 0.96338813, Time: 0.0209 Steps: 58330, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000857, Sample Num: 13712, Cur Loss: 0.20972723, Cur Avg Loss: 0.25787978, Log Avg loss: 0.31745314, Global Avg Loss: 0.96327741, Time: 0.0209 Steps: 58340, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000867, Sample Num: 13872, Cur Loss: 0.18280697, Cur Avg Loss: 0.25740587, Log Avg loss: 0.21679236, Global Avg Loss: 0.96314948, Time: 0.0209 Steps: 58350, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000877, Sample Num: 14032, Cur Loss: 0.39814872, Cur Avg Loss: 0.25717700, Log Avg loss: 0.23733363, Global Avg Loss: 0.96302511, Time: 0.0208 Steps: 58360, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000887, Sample Num: 14192, Cur Loss: 0.16885489, Cur Avg Loss: 0.25669883, Log Avg loss: 0.21476329, Global Avg Loss: 0.96289692, Time: 0.0210 Steps: 58370, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000897, Sample Num: 14352, Cur Loss: 0.21999183, Cur Avg Loss: 0.25632175, Log Avg loss: 0.22287462, Global Avg Loss: 0.96277016, Time: 0.0209 Steps: 58380, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000907, Sample Num: 14512, Cur Loss: 0.18999097, Cur Avg Loss: 0.25553178, Log Avg loss: 0.18467180, Global Avg Loss: 0.96263690, Time: 0.0209 Steps: 58390, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000917, Sample Num: 14672, Cur Loss: 0.21303479, Cur Avg Loss: 0.25557119, Log Avg loss: 0.25914562, Global Avg Loss: 0.96251644, Time: 0.0208 Steps: 58400, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000927, Sample Num: 14832, Cur Loss: 0.13444151, Cur Avg Loss: 0.25632588, Log Avg loss: 0.32553079, Global Avg Loss: 0.96240739, Time: 0.0209 Steps: 58410, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000937, Sample Num: 14992, Cur Loss: 0.15668938, Cur Avg Loss: 0.25578801, Log Avg loss: 0.20592762, Global Avg Loss: 0.96227790, Time: 0.0209 Steps: 58420, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000947, Sample Num: 15152, Cur Loss: 0.24852604, Cur Avg Loss: 0.25660919, Log Avg loss: 0.33355409, Global Avg Loss: 0.96217029, Time: 0.0208 Steps: 58430, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000957, Sample Num: 15312, Cur Loss: 0.50871545, Cur Avg Loss: 0.25628259, Log Avg loss: 0.22535369, Global Avg Loss: 0.96204421, Time: 0.0208 Steps: 58440, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000967, Sample Num: 15472, Cur Loss: 0.22355612, Cur Avg Loss: 0.25585050, Log Avg loss: 0.21449910, Global Avg Loss: 0.96191632, Time: 0.0208 Steps: 58450, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000977, Sample Num: 15632, Cur Loss: 0.07517657, Cur Avg Loss: 0.25587291, Log Avg loss: 0.25804024, Global Avg Loss: 0.96179591, Time: 0.0209 Steps: 58460, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000987, Sample Num: 15792, Cur Loss: 0.22048962, Cur Avg Loss: 0.25575659, Log Avg loss: 0.24439213, Global Avg Loss: 0.96167322, Time: 0.0209 Steps: 58470, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000997, Sample Num: 15952, Cur Loss: 0.05752055, Cur Avg Loss: 0.25497000, Log Avg loss: 0.17733347, Global Avg Loss: 0.96153910, Time: 0.0208 Steps: 58480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001007, Sample Num: 16112, Cur Loss: 0.15672234, Cur Avg Loss: 0.25573919, Log Avg loss: 0.33242752, Global Avg Loss: 0.96143154, Time: 0.0208 Steps: 58490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001017, Sample Num: 16272, Cur Loss: 0.12788197, Cur Avg Loss: 0.25606618, Log Avg loss: 0.28899367, Global Avg Loss: 0.96131659, Time: 0.0209 Steps: 58500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001027, Sample Num: 16432, Cur Loss: 0.09836562, Cur Avg Loss: 0.25541832, Log Avg loss: 0.18953134, Global Avg Loss: 0.96118469, Time: 0.0245 Steps: 58510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001037, Sample Num: 16592, Cur Loss: 0.10529513, Cur Avg Loss: 0.25464433, Log Avg loss: 0.17515567, Global Avg Loss: 0.96105037, Time: 0.0208 Steps: 58520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001047, Sample Num: 16752, Cur Loss: 0.22633949, Cur Avg Loss: 0.25485630, Log Avg loss: 0.27683686, Global Avg Loss: 0.96093347, Time: 0.0208 Steps: 58530, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001057, Sample Num: 16912, Cur Loss: 0.23413780, Cur Avg Loss: 0.25591240, Log Avg loss: 0.36648613, Global Avg Loss: 0.96083192, Time: 0.0209 Steps: 58540, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001067, Sample Num: 17072, Cur Loss: 0.09630431, Cur Avg Loss: 0.25677789, Log Avg loss: 0.34826034, Global Avg Loss: 0.96072730, Time: 0.0208 Steps: 58550, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001077, Sample Num: 17232, Cur Loss: 0.37433705, Cur Avg Loss: 0.25592853, Log Avg loss: 0.16530171, Global Avg Loss: 0.96059147, Time: 0.0208 Steps: 58560, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001087, Sample Num: 17392, Cur Loss: 0.09660280, Cur Avg Loss: 0.25526509, Log Avg loss: 0.18381301, Global Avg Loss: 0.96045884, Time: 0.0209 Steps: 58570, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001097, Sample Num: 17552, Cur Loss: 0.18560407, Cur Avg Loss: 0.25559487, Log Avg loss: 0.29144185, Global Avg Loss: 0.96034464, Time: 0.0209 Steps: 58580, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001107, Sample Num: 17712, Cur Loss: 0.28945228, Cur Avg Loss: 0.25648722, Log Avg loss: 0.35437761, Global Avg Loss: 0.96024121, Time: 0.0209 Steps: 58590, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001117, Sample Num: 17872, Cur Loss: 0.44112337, Cur Avg Loss: 0.25834474, Log Avg loss: 0.46397301, Global Avg Loss: 0.96015653, Time: 0.0209 Steps: 58600, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001127, Sample Num: 18032, Cur Loss: 0.07065180, Cur Avg Loss: 0.25907147, Log Avg loss: 0.34024716, Global Avg Loss: 0.96005076, Time: 0.0209 Steps: 58610, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001137, Sample Num: 18192, Cur Loss: 0.16753761, Cur Avg Loss: 0.25816907, Log Avg loss: 0.15646816, Global Avg Loss: 0.95991367, Time: 0.0209 Steps: 58620, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001147, Sample Num: 18352, Cur Loss: 0.38442603, Cur Avg Loss: 0.25835926, Log Avg loss: 0.27998386, Global Avg Loss: 0.95979770, Time: 0.0208 Steps: 58630, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001157, Sample Num: 18512, Cur Loss: 1.06166792, Cur Avg Loss: 0.25885626, Log Avg loss: 0.31586241, Global Avg Loss: 0.95968789, Time: 0.0209 Steps: 58640, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001167, Sample Num: 18672, Cur Loss: 0.69740766, Cur Avg Loss: 0.26003980, Log Avg loss: 0.39697516, Global Avg Loss: 0.95959195, Time: 0.0208 Steps: 58650, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001177, Sample Num: 18832, Cur Loss: 0.19833320, Cur Avg Loss: 0.26001263, Log Avg loss: 0.25684181, Global Avg Loss: 0.95947215, Time: 0.0208 Steps: 58660, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001187, Sample Num: 18992, Cur Loss: 0.19336170, Cur Avg Loss: 0.25987762, Log Avg loss: 0.24398654, Global Avg Loss: 0.95935020, Time: 0.0209 Steps: 58670, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001197, Sample Num: 19152, Cur Loss: 0.95353460, Cur Avg Loss: 0.26021540, Log Avg loss: 0.30031005, Global Avg Loss: 0.95923789, Time: 0.0208 Steps: 58680, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001207, Sample Num: 19312, Cur Loss: 0.19156173, Cur Avg Loss: 0.25975931, Log Avg loss: 0.20516538, Global Avg Loss: 0.95910940, Time: 0.0209 Steps: 58690, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001217, Sample Num: 19472, Cur Loss: 0.33115596, Cur Avg Loss: 0.25975917, Log Avg loss: 0.25974202, Global Avg Loss: 0.95899026, Time: 0.0209 Steps: 58700, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001227, Sample Num: 19632, Cur Loss: 0.35098970, Cur Avg Loss: 0.25977576, Log Avg loss: 0.26179481, Global Avg Loss: 0.95887151, Time: 0.0209 Steps: 58710, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001237, Sample Num: 19792, Cur Loss: 0.19755167, Cur Avg Loss: 0.25955337, Log Avg loss: 0.23226661, Global Avg Loss: 0.95874777, Time: 0.0208 Steps: 58720, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001247, Sample Num: 19952, Cur Loss: 0.14508399, Cur Avg Loss: 0.26030410, Log Avg loss: 0.35316943, Global Avg Loss: 0.95864465, Time: 0.0208 Steps: 58730, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001257, Sample Num: 20112, Cur Loss: 0.21563619, Cur Avg Loss: 0.26034346, Log Avg loss: 0.26525163, Global Avg Loss: 0.95852661, Time: 0.0209 Steps: 58740, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001267, Sample Num: 20272, Cur Loss: 0.28555700, Cur Avg Loss: 0.26018287, Log Avg loss: 0.23999640, Global Avg Loss: 0.95840431, Time: 0.0208 Steps: 58750, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001277, Sample Num: 20432, Cur Loss: 0.30642512, Cur Avg Loss: 0.26014963, Log Avg loss: 0.25593890, Global Avg Loss: 0.95828476, Time: 0.0208 Steps: 58760, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001287, Sample Num: 20592, Cur Loss: 0.24371660, Cur Avg Loss: 0.26047437, Log Avg loss: 0.30194268, Global Avg Loss: 0.95817308, Time: 0.0208 Steps: 58770, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001297, Sample Num: 20752, Cur Loss: 0.24573693, Cur Avg Loss: 0.26057119, Log Avg loss: 0.27303251, Global Avg Loss: 0.95805652, Time: 0.0207 Steps: 58780, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001307, Sample Num: 20912, Cur Loss: 0.09039278, Cur Avg Loss: 0.26027418, Log Avg loss: 0.22175126, Global Avg Loss: 0.95793127, Time: 0.0208 Steps: 58790, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001317, Sample Num: 21072, Cur Loss: 0.14320847, Cur Avg Loss: 0.26014651, Log Avg loss: 0.24346124, Global Avg Loss: 0.95780977, Time: 0.0208 Steps: 58800, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001327, Sample Num: 21232, Cur Loss: 0.34652460, Cur Avg Loss: 0.25978491, Log Avg loss: 0.21216206, Global Avg Loss: 0.95768298, Time: 0.0207 Steps: 58810, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001337, Sample Num: 21392, Cur Loss: 0.32974738, Cur Avg Loss: 0.25936532, Log Avg loss: 0.20368526, Global Avg Loss: 0.95755479, Time: 0.0207 Steps: 58820, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001347, Sample Num: 21552, Cur Loss: 0.25791103, Cur Avg Loss: 0.25880963, Log Avg loss: 0.18451438, Global Avg Loss: 0.95742339, Time: 0.0208 Steps: 58830, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001357, Sample Num: 21712, Cur Loss: 0.15817231, Cur Avg Loss: 0.25834947, Log Avg loss: 0.19636489, Global Avg Loss: 0.95729404, Time: 0.0208 Steps: 58840, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001367, Sample Num: 21872, Cur Loss: 0.15748726, Cur Avg Loss: 0.25808794, Log Avg loss: 0.22259948, Global Avg Loss: 0.95716920, Time: 0.0207 Steps: 58850, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001377, Sample Num: 22032, Cur Loss: 0.28657702, Cur Avg Loss: 0.25796264, Log Avg loss: 0.24083360, Global Avg Loss: 0.95704750, Time: 0.0208 Steps: 58860, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001387, Sample Num: 22192, Cur Loss: 0.36805689, Cur Avg Loss: 0.25773075, Log Avg loss: 0.22579985, Global Avg Loss: 0.95692329, Time: 0.0209 Steps: 58870, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001397, Sample Num: 22352, Cur Loss: 0.77597749, Cur Avg Loss: 0.25786997, Log Avg loss: 0.27717905, Global Avg Loss: 0.95680784, Time: 0.0208 Steps: 58880, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001407, Sample Num: 22512, Cur Loss: 0.17383182, Cur Avg Loss: 0.25799908, Log Avg loss: 0.27603559, Global Avg Loss: 0.95669224, Time: 0.0208 Steps: 58890, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001417, Sample Num: 22672, Cur Loss: 0.15962401, Cur Avg Loss: 0.25909385, Log Avg loss: 0.41312908, Global Avg Loss: 0.95659995, Time: 0.0207 Steps: 58900, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001427, Sample Num: 22832, Cur Loss: 0.37150091, Cur Avg Loss: 0.26077253, Log Avg loss: 0.49864017, Global Avg Loss: 0.95652222, Time: 0.0208 Steps: 58910, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001437, Sample Num: 22992, Cur Loss: 0.27828151, Cur Avg Loss: 0.26002837, Log Avg loss: 0.15383735, Global Avg Loss: 0.95638598, Time: 0.0208 Steps: 58920, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001447, Sample Num: 23152, Cur Loss: 0.19832672, Cur Avg Loss: 0.25949873, Log Avg loss: 0.18338943, Global Avg Loss: 0.95625481, Time: 0.0208 Steps: 58930, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001457, Sample Num: 23312, Cur Loss: 0.24909621, Cur Avg Loss: 0.25954415, Log Avg loss: 0.26611686, Global Avg Loss: 0.95613772, Time: 0.0208 Steps: 58940, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001467, Sample Num: 23472, Cur Loss: 0.23119652, Cur Avg Loss: 0.25983535, Log Avg loss: 0.30226237, Global Avg Loss: 0.95602680, Time: 0.0208 Steps: 58950, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001477, Sample Num: 23632, Cur Loss: 0.48661399, Cur Avg Loss: 0.25968328, Log Avg loss: 0.23737476, Global Avg Loss: 0.95590491, Time: 0.0208 Steps: 58960, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001487, Sample Num: 23792, Cur Loss: 0.09129760, Cur Avg Loss: 0.25907520, Log Avg loss: 0.16926257, Global Avg Loss: 0.95577151, Time: 0.0208 Steps: 58970, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001497, Sample Num: 23952, Cur Loss: 0.30274814, Cur Avg Loss: 0.25869262, Log Avg loss: 0.20180195, Global Avg Loss: 0.95564368, Time: 0.0208 Steps: 58980, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001507, Sample Num: 24112, Cur Loss: 0.59554315, Cur Avg Loss: 0.25867672, Log Avg loss: 0.25629724, Global Avg Loss: 0.95552512, Time: 0.0208 Steps: 58990, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001517, Sample Num: 24272, Cur Loss: 0.30041432, Cur Avg Loss: 0.26010054, Log Avg loss: 0.47466962, Global Avg Loss: 0.95544362, Time: 0.0208 Steps: 59000, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001527, Sample Num: 24432, Cur Loss: 0.10892344, Cur Avg Loss: 0.25970565, Log Avg loss: 0.19980124, Global Avg Loss: 0.95531557, Time: 0.0208 Steps: 59010, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001537, Sample Num: 24592, Cur Loss: 0.24841273, Cur Avg Loss: 0.25955469, Log Avg loss: 0.23650377, Global Avg Loss: 0.95519378, Time: 0.0244 Steps: 59020, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001547, Sample Num: 24752, Cur Loss: 0.21399473, Cur Avg Loss: 0.25922769, Log Avg loss: 0.20896739, Global Avg Loss: 0.95506736, Time: 0.0207 Steps: 59030, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001557, Sample Num: 24912, Cur Loss: 0.24890390, Cur Avg Loss: 0.25986094, Log Avg loss: 0.35782418, Global Avg Loss: 0.95496621, Time: 0.0208 Steps: 59040, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001567, Sample Num: 25072, Cur Loss: 0.17073080, Cur Avg Loss: 0.26007226, Log Avg loss: 0.29297521, Global Avg Loss: 0.95485410, Time: 0.0207 Steps: 59050, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001577, Sample Num: 25232, Cur Loss: 0.18331754, Cur Avg Loss: 0.25987926, Log Avg loss: 0.22963530, Global Avg Loss: 0.95473131, Time: 0.0208 Steps: 59060, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001587, Sample Num: 25392, Cur Loss: 0.11811286, Cur Avg Loss: 0.25913744, Log Avg loss: 0.14215367, Global Avg Loss: 0.95459374, Time: 0.0208 Steps: 59070, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001597, Sample Num: 25552, Cur Loss: 0.30559033, Cur Avg Loss: 0.25868616, Log Avg loss: 0.18706695, Global Avg Loss: 0.95446383, Time: 0.0207 Steps: 59080, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001607, Sample Num: 25712, Cur Loss: 0.19399181, Cur Avg Loss: 0.25878094, Log Avg loss: 0.27391857, Global Avg Loss: 0.95434866, Time: 0.0207 Steps: 59090, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001617, Sample Num: 25872, Cur Loss: 0.77168572, Cur Avg Loss: 0.25868109, Log Avg loss: 0.24263523, Global Avg Loss: 0.95422823, Time: 0.0207 Steps: 59100, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001627, Sample Num: 26032, Cur Loss: 0.53777528, Cur Avg Loss: 0.25902314, Log Avg loss: 0.31433213, Global Avg Loss: 0.95411998, Time: 0.0207 Steps: 59110, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001637, Sample Num: 26192, Cur Loss: 0.11676109, Cur Avg Loss: 0.25824484, Log Avg loss: 0.13161490, Global Avg Loss: 0.95398085, Time: 0.0208 Steps: 59120, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001647, Sample Num: 26352, Cur Loss: 0.16339985, Cur Avg Loss: 0.25871334, Log Avg loss: 0.33540680, Global Avg Loss: 0.95387624, Time: 0.0208 Steps: 59130, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001657, Sample Num: 26512, Cur Loss: 0.58473063, Cur Avg Loss: 0.25989770, Log Avg loss: 0.45496150, Global Avg Loss: 0.95379188, Time: 0.0208 Steps: 59140, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001667, Sample Num: 26672, Cur Loss: 0.44607794, Cur Avg Loss: 0.26015461, Log Avg loss: 0.30272568, Global Avg Loss: 0.95368181, Time: 0.0207 Steps: 59150, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001677, Sample Num: 26832, Cur Loss: 0.22422999, Cur Avg Loss: 0.26016018, Log Avg loss: 0.26108861, Global Avg Loss: 0.95356474, Time: 0.0208 Steps: 59160, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001687, Sample Num: 26992, Cur Loss: 0.31596112, Cur Avg Loss: 0.26028754, Log Avg loss: 0.28164605, Global Avg Loss: 0.95345118, Time: 0.0207 Steps: 59170, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001697, Sample Num: 27152, Cur Loss: 0.18699154, Cur Avg Loss: 0.26037904, Log Avg loss: 0.27581434, Global Avg Loss: 0.95333668, Time: 0.0208 Steps: 59180, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001707, Sample Num: 27312, Cur Loss: 0.08808772, Cur Avg Loss: 0.26006041, Log Avg loss: 0.20598862, Global Avg Loss: 0.95321041, Time: 0.0207 Steps: 59190, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001717, Sample Num: 27472, Cur Loss: 0.12009826, Cur Avg Loss: 0.26001717, Log Avg loss: 0.25263656, Global Avg Loss: 0.95309207, Time: 0.0207 Steps: 59200, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001727, Sample Num: 27632, Cur Loss: 0.11263778, Cur Avg Loss: 0.26086177, Log Avg loss: 0.40587864, Global Avg Loss: 0.95299966, Time: 0.0207 Steps: 59210, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001737, Sample Num: 27792, Cur Loss: 0.20796056, Cur Avg Loss: 0.26126892, Log Avg loss: 0.33158417, Global Avg Loss: 0.95289472, Time: 0.0207 Steps: 59220, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001747, Sample Num: 27952, Cur Loss: 0.22786486, Cur Avg Loss: 0.26203010, Log Avg loss: 0.39424672, Global Avg Loss: 0.95280040, Time: 0.0208 Steps: 59230, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001757, Sample Num: 28112, Cur Loss: 0.24861121, Cur Avg Loss: 0.26194783, Log Avg loss: 0.24757556, Global Avg Loss: 0.95268136, Time: 0.0208 Steps: 59240, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001767, Sample Num: 28272, Cur Loss: 0.13328001, Cur Avg Loss: 0.26162758, Log Avg loss: 0.20535931, Global Avg Loss: 0.95255523, Time: 0.0207 Steps: 59250, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001777, Sample Num: 28432, Cur Loss: 0.23691067, Cur Avg Loss: 0.26111335, Log Avg loss: 0.17024958, Global Avg Loss: 0.95242322, Time: 0.0208 Steps: 59260, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001787, Sample Num: 28592, Cur Loss: 0.30617571, Cur Avg Loss: 0.26086011, Log Avg loss: 0.21585990, Global Avg Loss: 0.95229894, Time: 0.0208 Steps: 59270, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001797, Sample Num: 28752, Cur Loss: 0.22322333, Cur Avg Loss: 0.26100532, Log Avg loss: 0.28695470, Global Avg Loss: 0.95218670, Time: 0.0209 Steps: 59280, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001807, Sample Num: 28912, Cur Loss: 0.14823189, Cur Avg Loss: 0.26122759, Log Avg loss: 0.30116909, Global Avg Loss: 0.95207690, Time: 0.0208 Steps: 59290, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001817, Sample Num: 29072, Cur Loss: 0.16365123, Cur Avg Loss: 0.26138104, Log Avg loss: 0.28910854, Global Avg Loss: 0.95196510, Time: 0.0208 Steps: 59300, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001827, Sample Num: 29232, Cur Loss: 0.31677052, Cur Avg Loss: 0.26110801, Log Avg loss: 0.21149875, Global Avg Loss: 0.95184026, Time: 0.0208 Steps: 59310, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001837, Sample Num: 29392, Cur Loss: 0.17062669, Cur Avg Loss: 0.26100348, Log Avg loss: 0.24190689, Global Avg Loss: 0.95172058, Time: 0.0208 Steps: 59320, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001847, Sample Num: 29552, Cur Loss: 0.15997803, Cur Avg Loss: 0.26091677, Log Avg loss: 0.24498682, Global Avg Loss: 0.95160146, Time: 0.0208 Steps: 59330, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001857, Sample Num: 29712, Cur Loss: 0.14618875, Cur Avg Loss: 0.26091462, Log Avg loss: 0.26051862, Global Avg Loss: 0.95148500, Time: 0.0208 Steps: 59340, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001867, Sample Num: 29872, Cur Loss: 0.09132001, Cur Avg Loss: 0.26086616, Log Avg loss: 0.25186730, Global Avg Loss: 0.95136712, Time: 0.0208 Steps: 59350, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001877, Sample Num: 30032, Cur Loss: 0.24843836, Cur Avg Loss: 0.26071613, Log Avg loss: 0.23270495, Global Avg Loss: 0.95124605, Time: 0.0208 Steps: 59360, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001887, Sample Num: 30192, Cur Loss: 0.09633107, Cur Avg Loss: 0.26026751, Log Avg loss: 0.17606219, Global Avg Loss: 0.95111548, Time: 0.0208 Steps: 59370, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001897, Sample Num: 30352, Cur Loss: 0.15257147, Cur Avg Loss: 0.25998750, Log Avg loss: 0.20714939, Global Avg Loss: 0.95099019, Time: 0.0207 Steps: 59380, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001907, Sample Num: 30512, Cur Loss: 0.16725773, Cur Avg Loss: 0.26017682, Log Avg loss: 0.29609052, Global Avg Loss: 0.95087992, Time: 0.0208 Steps: 59390, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001917, Sample Num: 30672, Cur Loss: 0.33457160, Cur Avg Loss: 0.26049573, Log Avg loss: 0.32131086, Global Avg Loss: 0.95077393, Time: 0.0207 Steps: 59400, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001927, Sample Num: 30832, Cur Loss: 0.20227516, Cur Avg Loss: 0.26063461, Log Avg loss: 0.28725912, Global Avg Loss: 0.95066225, Time: 0.0208 Steps: 59410, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001937, Sample Num: 30992, Cur Loss: 0.18796319, Cur Avg Loss: 0.26091376, Log Avg loss: 0.31470623, Global Avg Loss: 0.95055522, Time: 0.0208 Steps: 59420, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001947, Sample Num: 31152, Cur Loss: 0.22896877, Cur Avg Loss: 0.26115446, Log Avg loss: 0.30777688, Global Avg Loss: 0.95044706, Time: 0.0207 Steps: 59430, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001957, Sample Num: 31312, Cur Loss: 0.14292166, Cur Avg Loss: 0.26101503, Log Avg loss: 0.23386909, Global Avg Loss: 0.95032651, Time: 0.0208 Steps: 59440, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001967, Sample Num: 31472, Cur Loss: 0.23840895, Cur Avg Loss: 0.26079605, Log Avg loss: 0.21794028, Global Avg Loss: 0.95020332, Time: 0.0207 Steps: 59450, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001977, Sample Num: 31632, Cur Loss: 0.33022439, Cur Avg Loss: 0.26075960, Log Avg loss: 0.25359101, Global Avg Loss: 0.95008616, Time: 0.0208 Steps: 59460, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001987, Sample Num: 31792, Cur Loss: 0.41925532, Cur Avg Loss: 0.26077195, Log Avg loss: 0.26321375, Global Avg Loss: 0.94997066, Time: 0.0208 Steps: 59470, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001997, Sample Num: 31952, Cur Loss: 0.24095434, Cur Avg Loss: 0.26085531, Log Avg loss: 0.27741907, Global Avg Loss: 0.94985759, Time: 0.0208 Steps: 59480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002007, Sample Num: 32112, Cur Loss: 0.49001038, Cur Avg Loss: 0.26097812, Log Avg loss: 0.28550245, Global Avg Loss: 0.94974591, Time: 0.0208 Steps: 59490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002017, Sample Num: 32272, Cur Loss: 0.27554482, Cur Avg Loss: 0.26119108, Log Avg loss: 0.30393179, Global Avg Loss: 0.94963737, Time: 0.0208 Steps: 59500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002027, Sample Num: 32432, Cur Loss: 0.21739486, Cur Avg Loss: 0.26125834, Log Avg loss: 0.27482482, Global Avg Loss: 0.94952398, Time: 0.0208 Steps: 59510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002037, Sample Num: 32592, Cur Loss: 0.18826509, Cur Avg Loss: 0.26110885, Log Avg loss: 0.23080760, Global Avg Loss: 0.94940323, Time: 0.0207 Steps: 59520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002047, Sample Num: 32752, Cur Loss: 0.19000070, Cur Avg Loss: 0.26108819, Log Avg loss: 0.25687970, Global Avg Loss: 0.94928689, Time: 0.0208 Steps: 59530, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002057, Sample Num: 32912, Cur Loss: 0.07761750, Cur Avg Loss: 0.26084343, Log Avg loss: 0.21074135, Global Avg Loss: 0.94916285, Time: 0.0208 Steps: 59540, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002067, Sample Num: 33072, Cur Loss: 0.24720675, Cur Avg Loss: 0.26073559, Log Avg loss: 0.23855306, Global Avg Loss: 0.94904352, Time: 0.0208 Steps: 59550, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002077, Sample Num: 33232, Cur Loss: 0.28585458, Cur Avg Loss: 0.26058397, Log Avg loss: 0.22924365, Global Avg Loss: 0.94892267, Time: 0.0208 Steps: 59560, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002087, Sample Num: 33392, Cur Loss: 0.66270983, Cur Avg Loss: 0.26037424, Log Avg loss: 0.21681433, Global Avg Loss: 0.94879977, Time: 0.0208 Steps: 59570, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002097, Sample Num: 33552, Cur Loss: 0.18819514, Cur Avg Loss: 0.26009149, Log Avg loss: 0.20108134, Global Avg Loss: 0.94867427, Time: 0.0208 Steps: 59580, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002107, Sample Num: 33712, Cur Loss: 0.47075307, Cur Avg Loss: 0.26011416, Log Avg loss: 0.26486808, Global Avg Loss: 0.94855952, Time: 0.0208 Steps: 59590, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002117, Sample Num: 33872, Cur Loss: 0.11491901, Cur Avg Loss: 0.26039247, Log Avg loss: 0.31903231, Global Avg Loss: 0.94845390, Time: 0.0208 Steps: 59600, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002127, Sample Num: 34032, Cur Loss: 0.20577636, Cur Avg Loss: 0.26005656, Log Avg loss: 0.18894364, Global Avg Loss: 0.94832648, Time: 0.0208 Steps: 59610, Updated lr: 0.000044 ***** Running evaluation checkpoint-59612 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-59612 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.543325, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.292528, "eval_total_loss": 205.646947, "eval_mae": 0.416492, "eval_mse": 0.292575, "eval_r2": 0.81402, "eval_sp_statistic": 0.890012, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.916297, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.327365, "test_total_loss": 164.337028, "test_mae": 0.361086, "test_mse": 0.327484, "test_r2": 0.788639, "test_sp_statistic": 0.867788, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.91797, "test_ps_pvalue": 0.0, "lr": 4.441725936462779e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9483027052632839, "train_cur_epoch_loss": 553.6195617029443, "train_cur_epoch_avg_loss": 0.26003737045699593, "train_cur_epoch_time": 44.543325424194336, "train_cur_epoch_avg_time": 0.02092218197472726, "epoch": 28, "step": 59612} ################################################## Training, Epoch: 0029, Batch: 000008, Sample Num: 128, Cur Loss: 0.13901608, Cur Avg Loss: 0.13527479, Log Avg loss: 0.15614574, Global Avg Loss: 0.94819361, Time: 0.0209 Steps: 59620, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000018, Sample Num: 288, Cur Loss: 0.10078448, Cur Avg Loss: 0.28601379, Log Avg loss: 0.40660500, Global Avg Loss: 0.94810279, Time: 0.0207 Steps: 59630, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000028, Sample Num: 448, Cur Loss: 0.15348887, Cur Avg Loss: 0.29529933, Log Avg loss: 0.31201328, Global Avg Loss: 0.94799613, Time: 0.0208 Steps: 59640, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000038, Sample Num: 608, Cur Loss: 0.15353480, Cur Avg Loss: 0.28215701, Log Avg loss: 0.24535853, Global Avg Loss: 0.94787834, Time: 0.0207 Steps: 59650, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000048, Sample Num: 768, Cur Loss: 0.14265962, Cur Avg Loss: 0.26641065, Log Avg loss: 0.20657449, Global Avg Loss: 0.94775408, Time: 0.0208 Steps: 59660, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000058, Sample Num: 928, Cur Loss: 0.38023779, Cur Avg Loss: 0.25485192, Log Avg loss: 0.19937003, Global Avg Loss: 0.94762866, Time: 0.0208 Steps: 59670, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000068, Sample Num: 1088, Cur Loss: 0.10787788, Cur Avg Loss: 0.25739394, Log Avg loss: 0.27213766, Global Avg Loss: 0.94751548, Time: 0.0208 Steps: 59680, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000078, Sample Num: 1248, Cur Loss: 0.33612072, Cur Avg Loss: 0.24646647, Log Avg loss: 0.17215968, Global Avg Loss: 0.94738558, Time: 0.0207 Steps: 59690, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000088, Sample Num: 1408, Cur Loss: 0.10297432, Cur Avg Loss: 0.23905767, Log Avg loss: 0.18126898, Global Avg Loss: 0.94725725, Time: 0.0208 Steps: 59700, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000098, Sample Num: 1568, Cur Loss: 0.55779737, Cur Avg Loss: 0.25278794, Log Avg loss: 0.37361434, Global Avg Loss: 0.94716118, Time: 0.0208 Steps: 59710, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000108, Sample Num: 1728, Cur Loss: 0.08901414, Cur Avg Loss: 0.25259048, Log Avg loss: 0.25065540, Global Avg Loss: 0.94704455, Time: 0.0207 Steps: 59720, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000118, Sample Num: 1888, Cur Loss: 0.22929221, Cur Avg Loss: 0.24805314, Log Avg loss: 0.19904989, Global Avg Loss: 0.94691932, Time: 0.0208 Steps: 59730, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000128, Sample Num: 2048, Cur Loss: 0.22661498, Cur Avg Loss: 0.24954606, Log Avg loss: 0.26716243, Global Avg Loss: 0.94680554, Time: 0.0208 Steps: 59740, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000138, Sample Num: 2208, Cur Loss: 0.10312695, Cur Avg Loss: 0.24621751, Log Avg loss: 0.20361208, Global Avg Loss: 0.94668115, Time: 0.0208 Steps: 59750, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000148, Sample Num: 2368, Cur Loss: 0.34047407, Cur Avg Loss: 0.24992788, Log Avg loss: 0.30113102, Global Avg Loss: 0.94657313, Time: 0.0208 Steps: 59760, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000158, Sample Num: 2528, Cur Loss: 0.09887431, Cur Avg Loss: 0.24734209, Log Avg loss: 0.20907238, Global Avg Loss: 0.94644974, Time: 0.0207 Steps: 59770, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000168, Sample Num: 2688, Cur Loss: 0.08055528, Cur Avg Loss: 0.24491431, Log Avg loss: 0.20655543, Global Avg Loss: 0.94632597, Time: 0.0207 Steps: 59780, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000178, Sample Num: 2848, Cur Loss: 0.35311294, Cur Avg Loss: 0.24383584, Log Avg loss: 0.22571751, Global Avg Loss: 0.94620545, Time: 0.0208 Steps: 59790, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000188, Sample Num: 3008, Cur Loss: 0.21892013, Cur Avg Loss: 0.24223451, Log Avg loss: 0.21373075, Global Avg Loss: 0.94608296, Time: 0.0208 Steps: 59800, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000198, Sample Num: 3168, Cur Loss: 1.53924716, Cur Avg Loss: 0.25101465, Log Avg loss: 0.41608130, Global Avg Loss: 0.94599434, Time: 0.0208 Steps: 59810, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000208, Sample Num: 3328, Cur Loss: 0.70847130, Cur Avg Loss: 0.25535614, Log Avg loss: 0.34131763, Global Avg Loss: 0.94589326, Time: 0.0208 Steps: 59820, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000218, Sample Num: 3488, Cur Loss: 0.28347608, Cur Avg Loss: 0.25702542, Log Avg loss: 0.29174661, Global Avg Loss: 0.94578393, Time: 0.0208 Steps: 59830, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000228, Sample Num: 3648, Cur Loss: 0.11670288, Cur Avg Loss: 0.25375165, Log Avg loss: 0.18238330, Global Avg Loss: 0.94565635, Time: 0.0208 Steps: 59840, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000238, Sample Num: 3808, Cur Loss: 0.07283438, Cur Avg Loss: 0.25275203, Log Avg loss: 0.22996085, Global Avg Loss: 0.94553677, Time: 0.0208 Steps: 59850, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000248, Sample Num: 3968, Cur Loss: 0.13413793, Cur Avg Loss: 0.25147011, Log Avg loss: 0.22096046, Global Avg Loss: 0.94541573, Time: 0.0208 Steps: 59860, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000258, Sample Num: 4128, Cur Loss: 0.16146864, Cur Avg Loss: 0.24929771, Log Avg loss: 0.19542196, Global Avg Loss: 0.94529046, Time: 0.0247 Steps: 59870, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000268, Sample Num: 4288, Cur Loss: 0.11786500, Cur Avg Loss: 0.25048210, Log Avg loss: 0.28103947, Global Avg Loss: 0.94517953, Time: 0.0209 Steps: 59880, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000278, Sample Num: 4448, Cur Loss: 0.45259064, Cur Avg Loss: 0.25512235, Log Avg loss: 0.37948098, Global Avg Loss: 0.94508507, Time: 0.0209 Steps: 59890, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000288, Sample Num: 4608, Cur Loss: 0.12507409, Cur Avg Loss: 0.25767811, Log Avg loss: 0.32872824, Global Avg Loss: 0.94498217, Time: 0.0208 Steps: 59900, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000298, Sample Num: 4768, Cur Loss: 0.02255441, Cur Avg Loss: 0.25527191, Log Avg loss: 0.18597349, Global Avg Loss: 0.94485548, Time: 0.0209 Steps: 59910, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000308, Sample Num: 4928, Cur Loss: 0.07037435, Cur Avg Loss: 0.25279020, Log Avg loss: 0.17883523, Global Avg Loss: 0.94472764, Time: 0.0208 Steps: 59920, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000318, Sample Num: 5088, Cur Loss: 0.24754238, Cur Avg Loss: 0.25107679, Log Avg loss: 0.19830362, Global Avg Loss: 0.94460309, Time: 0.0209 Steps: 59930, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000328, Sample Num: 5248, Cur Loss: 0.05162073, Cur Avg Loss: 0.25056130, Log Avg loss: 0.23416893, Global Avg Loss: 0.94448457, Time: 0.0209 Steps: 59940, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000338, Sample Num: 5408, Cur Loss: 0.34973156, Cur Avg Loss: 0.25029009, Log Avg loss: 0.24139426, Global Avg Loss: 0.94436729, Time: 0.0210 Steps: 59950, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000348, Sample Num: 5568, Cur Loss: 0.75297016, Cur Avg Loss: 0.25317525, Log Avg loss: 0.35069375, Global Avg Loss: 0.94426828, Time: 0.0209 Steps: 59960, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000358, Sample Num: 5728, Cur Loss: 0.70489609, Cur Avg Loss: 0.25827609, Log Avg loss: 0.43578511, Global Avg Loss: 0.94418349, Time: 0.0209 Steps: 59970, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000368, Sample Num: 5888, Cur Loss: 0.52947307, Cur Avg Loss: 0.25816153, Log Avg loss: 0.25406057, Global Avg Loss: 0.94406843, Time: 0.0209 Steps: 59980, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000378, Sample Num: 6048, Cur Loss: 0.32998106, Cur Avg Loss: 0.25742487, Log Avg loss: 0.23031555, Global Avg Loss: 0.94394945, Time: 0.0209 Steps: 59990, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000388, Sample Num: 6208, Cur Loss: 0.34750444, Cur Avg Loss: 0.25734418, Log Avg loss: 0.25429421, Global Avg Loss: 0.94383451, Time: 0.0209 Steps: 60000, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000398, Sample Num: 6368, Cur Loss: 0.41519603, Cur Avg Loss: 0.25843494, Log Avg loss: 0.30075654, Global Avg Loss: 0.94372735, Time: 0.0209 Steps: 60010, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000408, Sample Num: 6528, Cur Loss: 0.53766769, Cur Avg Loss: 0.25697826, Log Avg loss: 0.19900223, Global Avg Loss: 0.94360327, Time: 0.0210 Steps: 60020, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000418, Sample Num: 6688, Cur Loss: 0.16052446, Cur Avg Loss: 0.25635633, Log Avg loss: 0.23098158, Global Avg Loss: 0.94348455, Time: 0.0210 Steps: 60030, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000428, Sample Num: 6848, Cur Loss: 0.35788193, Cur Avg Loss: 0.25622356, Log Avg loss: 0.25067396, Global Avg Loss: 0.94336916, Time: 0.0208 Steps: 60040, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000438, Sample Num: 7008, Cur Loss: 0.23246959, Cur Avg Loss: 0.25604499, Log Avg loss: 0.24840201, Global Avg Loss: 0.94325343, Time: 0.0209 Steps: 60050, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000448, Sample Num: 7168, Cur Loss: 0.13274440, Cur Avg Loss: 0.25475799, Log Avg loss: 0.19838744, Global Avg Loss: 0.94312941, Time: 0.0209 Steps: 60060, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000458, Sample Num: 7328, Cur Loss: 0.22511071, Cur Avg Loss: 0.25459243, Log Avg loss: 0.24717530, Global Avg Loss: 0.94301355, Time: 0.0209 Steps: 60070, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000468, Sample Num: 7488, Cur Loss: 0.33325619, Cur Avg Loss: 0.25781926, Log Avg loss: 0.40560786, Global Avg Loss: 0.94292411, Time: 0.0208 Steps: 60080, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000478, Sample Num: 7648, Cur Loss: 0.28944975, Cur Avg Loss: 0.25795350, Log Avg loss: 0.26423628, Global Avg Loss: 0.94281116, Time: 0.0209 Steps: 60090, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000488, Sample Num: 7808, Cur Loss: 0.45561767, Cur Avg Loss: 0.25876425, Log Avg loss: 0.29751796, Global Avg Loss: 0.94270379, Time: 0.0209 Steps: 60100, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000498, Sample Num: 7968, Cur Loss: 0.12608215, Cur Avg Loss: 0.25840995, Log Avg loss: 0.24112013, Global Avg Loss: 0.94258707, Time: 0.0209 Steps: 60110, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000508, Sample Num: 8128, Cur Loss: 0.28549072, Cur Avg Loss: 0.26023043, Log Avg loss: 0.35089052, Global Avg Loss: 0.94248865, Time: 0.0208 Steps: 60120, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000518, Sample Num: 8288, Cur Loss: 0.36575553, Cur Avg Loss: 0.25987937, Log Avg loss: 0.24204540, Global Avg Loss: 0.94237217, Time: 0.0246 Steps: 60130, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000528, Sample Num: 8448, Cur Loss: 0.16732909, Cur Avg Loss: 0.25909169, Log Avg loss: 0.21828961, Global Avg Loss: 0.94225177, Time: 0.0209 Steps: 60140, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000538, Sample Num: 8608, Cur Loss: 0.35441464, Cur Avg Loss: 0.25876755, Log Avg loss: 0.24165309, Global Avg Loss: 0.94213529, Time: 0.0208 Steps: 60150, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000548, Sample Num: 8768, Cur Loss: 0.39027411, Cur Avg Loss: 0.26004938, Log Avg loss: 0.32901176, Global Avg Loss: 0.94203338, Time: 0.0208 Steps: 60160, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000558, Sample Num: 8928, Cur Loss: 0.33259296, Cur Avg Loss: 0.25943653, Log Avg loss: 0.22585245, Global Avg Loss: 0.94191435, Time: 0.0208 Steps: 60170, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000568, Sample Num: 9088, Cur Loss: 0.18547276, Cur Avg Loss: 0.25860479, Log Avg loss: 0.21219351, Global Avg Loss: 0.94179309, Time: 0.0208 Steps: 60180, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000578, Sample Num: 9248, Cur Loss: 0.07702459, Cur Avg Loss: 0.25897310, Log Avg loss: 0.27989362, Global Avg Loss: 0.94168313, Time: 0.0209 Steps: 60190, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000588, Sample Num: 9408, Cur Loss: 0.13360892, Cur Avg Loss: 0.25717891, Log Avg loss: 0.15347422, Global Avg Loss: 0.94155219, Time: 0.0208 Steps: 60200, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000598, Sample Num: 9568, Cur Loss: 0.19933611, Cur Avg Loss: 0.25574935, Log Avg loss: 0.17169136, Global Avg Loss: 0.94142433, Time: 0.0209 Steps: 60210, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000608, Sample Num: 9728, Cur Loss: 0.10768311, Cur Avg Loss: 0.25452803, Log Avg loss: 0.18149315, Global Avg Loss: 0.94129814, Time: 0.0208 Steps: 60220, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000618, Sample Num: 9888, Cur Loss: 0.13087313, Cur Avg Loss: 0.25593994, Log Avg loss: 0.34178384, Global Avg Loss: 0.94119860, Time: 0.0207 Steps: 60230, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000628, Sample Num: 10048, Cur Loss: 0.07857526, Cur Avg Loss: 0.25479398, Log Avg loss: 0.18397363, Global Avg Loss: 0.94107290, Time: 0.0208 Steps: 60240, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000638, Sample Num: 10208, Cur Loss: 0.19116876, Cur Avg Loss: 0.25388707, Log Avg loss: 0.19693363, Global Avg Loss: 0.94094939, Time: 0.0208 Steps: 60250, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000648, Sample Num: 10368, Cur Loss: 0.12107114, Cur Avg Loss: 0.25361263, Log Avg loss: 0.23610295, Global Avg Loss: 0.94083242, Time: 0.0208 Steps: 60260, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000658, Sample Num: 10528, Cur Loss: 0.20150822, Cur Avg Loss: 0.25344763, Log Avg loss: 0.24275588, Global Avg Loss: 0.94071660, Time: 0.0208 Steps: 60270, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000668, Sample Num: 10688, Cur Loss: 0.25535598, Cur Avg Loss: 0.25296856, Log Avg loss: 0.22144550, Global Avg Loss: 0.94059728, Time: 0.0208 Steps: 60280, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000678, Sample Num: 10848, Cur Loss: 0.38125789, Cur Avg Loss: 0.25259552, Log Avg loss: 0.22767663, Global Avg Loss: 0.94047903, Time: 0.0208 Steps: 60290, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000688, Sample Num: 11008, Cur Loss: 0.20802198, Cur Avg Loss: 0.25213188, Log Avg loss: 0.22069730, Global Avg Loss: 0.94035966, Time: 0.0208 Steps: 60300, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000698, Sample Num: 11168, Cur Loss: 0.10960190, Cur Avg Loss: 0.25092101, Log Avg loss: 0.16761275, Global Avg Loss: 0.94023153, Time: 0.0208 Steps: 60310, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000708, Sample Num: 11328, Cur Loss: 0.07716544, Cur Avg Loss: 0.25040736, Log Avg loss: 0.21455493, Global Avg Loss: 0.94011123, Time: 0.0208 Steps: 60320, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000718, Sample Num: 11488, Cur Loss: 0.10782649, Cur Avg Loss: 0.25209004, Log Avg loss: 0.37122377, Global Avg Loss: 0.94001693, Time: 0.0208 Steps: 60330, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000728, Sample Num: 11648, Cur Loss: 0.23284535, Cur Avg Loss: 0.25163702, Log Avg loss: 0.21911030, Global Avg Loss: 0.93989746, Time: 0.0208 Steps: 60340, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000738, Sample Num: 11808, Cur Loss: 0.20701319, Cur Avg Loss: 0.25129662, Log Avg loss: 0.22651526, Global Avg Loss: 0.93977925, Time: 0.0208 Steps: 60350, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000748, Sample Num: 11968, Cur Loss: 0.11164584, Cur Avg Loss: 0.25116108, Log Avg loss: 0.24115830, Global Avg Loss: 0.93966351, Time: 0.0208 Steps: 60360, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000758, Sample Num: 12128, Cur Loss: 0.09013432, Cur Avg Loss: 0.25044123, Log Avg loss: 0.19659631, Global Avg Loss: 0.93954042, Time: 0.0208 Steps: 60370, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000768, Sample Num: 12288, Cur Loss: 0.23397277, Cur Avg Loss: 0.24985300, Log Avg loss: 0.20526493, Global Avg Loss: 0.93941881, Time: 0.0253 Steps: 60380, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000778, Sample Num: 12448, Cur Loss: 0.07206848, Cur Avg Loss: 0.24884045, Log Avg loss: 0.17107662, Global Avg Loss: 0.93929158, Time: 0.0207 Steps: 60390, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000788, Sample Num: 12608, Cur Loss: 0.34088731, Cur Avg Loss: 0.24991475, Log Avg loss: 0.33349568, Global Avg Loss: 0.93919129, Time: 0.0207 Steps: 60400, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000798, Sample Num: 12768, Cur Loss: 0.34909981, Cur Avg Loss: 0.24948162, Log Avg loss: 0.21535079, Global Avg Loss: 0.93907146, Time: 0.0208 Steps: 60410, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000808, Sample Num: 12928, Cur Loss: 0.85907179, Cur Avg Loss: 0.24983025, Log Avg loss: 0.27765069, Global Avg Loss: 0.93896199, Time: 0.0208 Steps: 60420, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000818, Sample Num: 13088, Cur Loss: 0.11174929, Cur Avg Loss: 0.24948389, Log Avg loss: 0.22149813, Global Avg Loss: 0.93884327, Time: 0.0207 Steps: 60430, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000828, Sample Num: 13248, Cur Loss: 0.61360145, Cur Avg Loss: 0.25013034, Log Avg loss: 0.30300995, Global Avg Loss: 0.93873807, Time: 0.0208 Steps: 60440, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000838, Sample Num: 13408, Cur Loss: 0.30847704, Cur Avg Loss: 0.25002845, Log Avg loss: 0.24159255, Global Avg Loss: 0.93862274, Time: 0.0208 Steps: 60450, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000848, Sample Num: 13568, Cur Loss: 0.19860435, Cur Avg Loss: 0.25040021, Log Avg loss: 0.28155313, Global Avg Loss: 0.93851406, Time: 0.0208 Steps: 60460, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000858, Sample Num: 13728, Cur Loss: 0.32565999, Cur Avg Loss: 0.24996176, Log Avg loss: 0.21278117, Global Avg Loss: 0.93839405, Time: 0.0208 Steps: 60470, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000868, Sample Num: 13888, Cur Loss: 0.09032092, Cur Avg Loss: 0.24892993, Log Avg loss: 0.16039890, Global Avg Loss: 0.93826541, Time: 0.0207 Steps: 60480, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000878, Sample Num: 14048, Cur Loss: 0.37264684, Cur Avg Loss: 0.24911101, Log Avg loss: 0.26482926, Global Avg Loss: 0.93815408, Time: 0.0208 Steps: 60490, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000888, Sample Num: 14208, Cur Loss: 0.49766010, Cur Avg Loss: 0.24983564, Log Avg loss: 0.31345770, Global Avg Loss: 0.93805083, Time: 0.0208 Steps: 60500, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000898, Sample Num: 14368, Cur Loss: 0.21818370, Cur Avg Loss: 0.25056331, Log Avg loss: 0.31518085, Global Avg Loss: 0.93794789, Time: 0.0208 Steps: 60510, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000908, Sample Num: 14528, Cur Loss: 0.08653991, Cur Avg Loss: 0.25032219, Log Avg loss: 0.22866955, Global Avg Loss: 0.93783069, Time: 0.0207 Steps: 60520, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000918, Sample Num: 14688, Cur Loss: 0.33099306, Cur Avg Loss: 0.25024718, Log Avg loss: 0.24343569, Global Avg Loss: 0.93771597, Time: 0.0208 Steps: 60530, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000928, Sample Num: 14848, Cur Loss: 0.25826469, Cur Avg Loss: 0.25129796, Log Avg loss: 0.34775949, Global Avg Loss: 0.93761852, Time: 0.0208 Steps: 60540, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000938, Sample Num: 15008, Cur Loss: 0.77001286, Cur Avg Loss: 0.25186859, Log Avg loss: 0.30482364, Global Avg Loss: 0.93751401, Time: 0.0208 Steps: 60550, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000948, Sample Num: 15168, Cur Loss: 0.07319830, Cur Avg Loss: 0.25136832, Log Avg loss: 0.20444274, Global Avg Loss: 0.93739297, Time: 0.0208 Steps: 60560, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000958, Sample Num: 15328, Cur Loss: 0.33335847, Cur Avg Loss: 0.25101088, Log Avg loss: 0.21712533, Global Avg Loss: 0.93727405, Time: 0.0208 Steps: 60570, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000968, Sample Num: 15488, Cur Loss: 0.76232988, Cur Avg Loss: 0.25108490, Log Avg loss: 0.25817620, Global Avg Loss: 0.93716195, Time: 0.0207 Steps: 60580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000978, Sample Num: 15648, Cur Loss: 0.16487338, Cur Avg Loss: 0.25127294, Log Avg loss: 0.26947570, Global Avg Loss: 0.93705175, Time: 0.0208 Steps: 60590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000988, Sample Num: 15808, Cur Loss: 0.33654803, Cur Avg Loss: 0.25091739, Log Avg loss: 0.21614472, Global Avg Loss: 0.93693279, Time: 0.0208 Steps: 60600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000998, Sample Num: 15968, Cur Loss: 0.08514786, Cur Avg Loss: 0.25035751, Log Avg loss: 0.19504046, Global Avg Loss: 0.93681039, Time: 0.0207 Steps: 60610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001008, Sample Num: 16128, Cur Loss: 0.15557772, Cur Avg Loss: 0.25022975, Log Avg loss: 0.23747944, Global Avg Loss: 0.93669503, Time: 0.0207 Steps: 60620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001018, Sample Num: 16288, Cur Loss: 0.44123492, Cur Avg Loss: 0.25071411, Log Avg loss: 0.29953803, Global Avg Loss: 0.93658994, Time: 0.0208 Steps: 60630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001028, Sample Num: 16448, Cur Loss: 0.62002468, Cur Avg Loss: 0.25122657, Log Avg loss: 0.30339463, Global Avg Loss: 0.93648552, Time: 0.0235 Steps: 60640, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001038, Sample Num: 16608, Cur Loss: 0.11915348, Cur Avg Loss: 0.25273164, Log Avg loss: 0.40745263, Global Avg Loss: 0.93639829, Time: 0.0208 Steps: 60650, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001048, Sample Num: 16768, Cur Loss: 0.17178851, Cur Avg Loss: 0.25333673, Log Avg loss: 0.31614509, Global Avg Loss: 0.93629604, Time: 0.0208 Steps: 60660, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001058, Sample Num: 16928, Cur Loss: 0.10872874, Cur Avg Loss: 0.25520952, Log Avg loss: 0.45147845, Global Avg Loss: 0.93621613, Time: 0.0208 Steps: 60670, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001068, Sample Num: 17088, Cur Loss: 0.29469723, Cur Avg Loss: 0.25457736, Log Avg loss: 0.18769432, Global Avg Loss: 0.93609277, Time: 0.0208 Steps: 60680, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001078, Sample Num: 17248, Cur Loss: 0.16925630, Cur Avg Loss: 0.25476754, Log Avg loss: 0.27507893, Global Avg Loss: 0.93598386, Time: 0.0207 Steps: 60690, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001088, Sample Num: 17408, Cur Loss: 0.21893361, Cur Avg Loss: 0.25484134, Log Avg loss: 0.26279740, Global Avg Loss: 0.93587295, Time: 0.0208 Steps: 60700, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001098, Sample Num: 17568, Cur Loss: 0.23761883, Cur Avg Loss: 0.25424503, Log Avg loss: 0.18936679, Global Avg Loss: 0.93574999, Time: 0.0208 Steps: 60710, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001108, Sample Num: 17728, Cur Loss: 0.08139519, Cur Avg Loss: 0.25342796, Log Avg loss: 0.16371371, Global Avg Loss: 0.93562284, Time: 0.0208 Steps: 60720, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001118, Sample Num: 17888, Cur Loss: 0.10126652, Cur Avg Loss: 0.25264292, Log Avg loss: 0.16565961, Global Avg Loss: 0.93549606, Time: 0.0208 Steps: 60730, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001128, Sample Num: 18048, Cur Loss: 0.11809365, Cur Avg Loss: 0.25180566, Log Avg loss: 0.15820044, Global Avg Loss: 0.93536809, Time: 0.0208 Steps: 60740, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001138, Sample Num: 18208, Cur Loss: 0.16095437, Cur Avg Loss: 0.25171460, Log Avg loss: 0.24144299, Global Avg Loss: 0.93525386, Time: 0.0208 Steps: 60750, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001148, Sample Num: 18368, Cur Loss: 0.34886560, Cur Avg Loss: 0.25154406, Log Avg loss: 0.23213672, Global Avg Loss: 0.93513814, Time: 0.0208 Steps: 60760, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001158, Sample Num: 18528, Cur Loss: 0.17272817, Cur Avg Loss: 0.25126925, Log Avg loss: 0.21972127, Global Avg Loss: 0.93502042, Time: 0.0208 Steps: 60770, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001168, Sample Num: 18688, Cur Loss: 0.13295892, Cur Avg Loss: 0.25261883, Log Avg loss: 0.40890034, Global Avg Loss: 0.93493385, Time: 0.0208 Steps: 60780, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001178, Sample Num: 18848, Cur Loss: 0.37356627, Cur Avg Loss: 0.25219759, Log Avg loss: 0.20299649, Global Avg Loss: 0.93481345, Time: 0.0208 Steps: 60790, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001188, Sample Num: 19008, Cur Loss: 0.43432331, Cur Avg Loss: 0.25223800, Log Avg loss: 0.25699776, Global Avg Loss: 0.93470197, Time: 0.0208 Steps: 60800, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001198, Sample Num: 19168, Cur Loss: 0.07971823, Cur Avg Loss: 0.25140836, Log Avg loss: 0.15284700, Global Avg Loss: 0.93457339, Time: 0.0208 Steps: 60810, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001208, Sample Num: 19328, Cur Loss: 0.21518491, Cur Avg Loss: 0.25093367, Log Avg loss: 0.19406686, Global Avg Loss: 0.93445164, Time: 0.0208 Steps: 60820, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001218, Sample Num: 19488, Cur Loss: 0.26497325, Cur Avg Loss: 0.25132133, Log Avg loss: 0.29814988, Global Avg Loss: 0.93434704, Time: 0.0208 Steps: 60830, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001228, Sample Num: 19648, Cur Loss: 0.39003253, Cur Avg Loss: 0.25123281, Log Avg loss: 0.24045151, Global Avg Loss: 0.93423298, Time: 0.0208 Steps: 60840, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001238, Sample Num: 19808, Cur Loss: 0.14540771, Cur Avg Loss: 0.25205253, Log Avg loss: 0.35271355, Global Avg Loss: 0.93413742, Time: 0.0208 Steps: 60850, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001248, Sample Num: 19968, Cur Loss: 0.17791179, Cur Avg Loss: 0.25236103, Log Avg loss: 0.29055368, Global Avg Loss: 0.93403167, Time: 0.0208 Steps: 60860, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001258, Sample Num: 20128, Cur Loss: 0.30155796, Cur Avg Loss: 0.25236554, Log Avg loss: 0.25292810, Global Avg Loss: 0.93391978, Time: 0.0208 Steps: 60870, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001268, Sample Num: 20288, Cur Loss: 0.15898170, Cur Avg Loss: 0.25231712, Log Avg loss: 0.24622628, Global Avg Loss: 0.93380682, Time: 0.0207 Steps: 60880, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001278, Sample Num: 20448, Cur Loss: 0.14838262, Cur Avg Loss: 0.25204633, Log Avg loss: 0.21771001, Global Avg Loss: 0.93368921, Time: 0.0208 Steps: 60890, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001288, Sample Num: 20608, Cur Loss: 0.24882230, Cur Avg Loss: 0.25190426, Log Avg loss: 0.23374718, Global Avg Loss: 0.93357428, Time: 0.0209 Steps: 60900, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001298, Sample Num: 20768, Cur Loss: 0.08850591, Cur Avg Loss: 0.25219818, Log Avg loss: 0.29005543, Global Avg Loss: 0.93346863, Time: 0.0208 Steps: 60910, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001308, Sample Num: 20928, Cur Loss: 0.13092735, Cur Avg Loss: 0.25262366, Log Avg loss: 0.30785176, Global Avg Loss: 0.93336593, Time: 0.0208 Steps: 60920, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001318, Sample Num: 21088, Cur Loss: 0.32517093, Cur Avg Loss: 0.25282066, Log Avg loss: 0.27858757, Global Avg Loss: 0.93325847, Time: 0.0208 Steps: 60930, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001328, Sample Num: 21248, Cur Loss: 0.15146518, Cur Avg Loss: 0.25238206, Log Avg loss: 0.19457488, Global Avg Loss: 0.93313725, Time: 0.0208 Steps: 60940, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001338, Sample Num: 21408, Cur Loss: 0.21777485, Cur Avg Loss: 0.25230517, Log Avg loss: 0.24209373, Global Avg Loss: 0.93302387, Time: 0.0208 Steps: 60950, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001348, Sample Num: 21568, Cur Loss: 0.20413882, Cur Avg Loss: 0.25279164, Log Avg loss: 0.31788186, Global Avg Loss: 0.93292297, Time: 0.0209 Steps: 60960, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001358, Sample Num: 21728, Cur Loss: 0.19392315, Cur Avg Loss: 0.25274354, Log Avg loss: 0.24625991, Global Avg Loss: 0.93281034, Time: 0.0208 Steps: 60970, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001368, Sample Num: 21888, Cur Loss: 0.37796044, Cur Avg Loss: 0.25273607, Log Avg loss: 0.25172162, Global Avg Loss: 0.93269865, Time: 0.0208 Steps: 60980, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001378, Sample Num: 22048, Cur Loss: 0.20788325, Cur Avg Loss: 0.25242568, Log Avg loss: 0.20996363, Global Avg Loss: 0.93258015, Time: 0.0208 Steps: 60990, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001388, Sample Num: 22208, Cur Loss: 0.21861047, Cur Avg Loss: 0.25247597, Log Avg loss: 0.25940661, Global Avg Loss: 0.93246980, Time: 0.0208 Steps: 61000, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001398, Sample Num: 22368, Cur Loss: 0.18247785, Cur Avg Loss: 0.25255194, Log Avg loss: 0.26309553, Global Avg Loss: 0.93236008, Time: 0.0208 Steps: 61010, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001408, Sample Num: 22528, Cur Loss: 0.20858023, Cur Avg Loss: 0.25278450, Log Avg loss: 0.28529743, Global Avg Loss: 0.93225404, Time: 0.0208 Steps: 61020, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001418, Sample Num: 22688, Cur Loss: 0.22532457, Cur Avg Loss: 0.25278520, Log Avg loss: 0.25288319, Global Avg Loss: 0.93214272, Time: 0.0208 Steps: 61030, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001428, Sample Num: 22848, Cur Loss: 0.14376219, Cur Avg Loss: 0.25345309, Log Avg loss: 0.34815997, Global Avg Loss: 0.93204705, Time: 0.0208 Steps: 61040, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001438, Sample Num: 23008, Cur Loss: 0.15891051, Cur Avg Loss: 0.25386286, Log Avg loss: 0.31237758, Global Avg Loss: 0.93194555, Time: 0.0207 Steps: 61050, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001448, Sample Num: 23168, Cur Loss: 0.12305783, Cur Avg Loss: 0.25396619, Log Avg loss: 0.26882518, Global Avg Loss: 0.93183695, Time: 0.0211 Steps: 61060, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001458, Sample Num: 23328, Cur Loss: 0.51126152, Cur Avg Loss: 0.25442384, Log Avg loss: 0.32069152, Global Avg Loss: 0.93173687, Time: 0.0208 Steps: 61070, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001468, Sample Num: 23488, Cur Loss: 0.26694041, Cur Avg Loss: 0.25494172, Log Avg loss: 0.33044834, Global Avg Loss: 0.93163843, Time: 0.0208 Steps: 61080, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001478, Sample Num: 23648, Cur Loss: 0.22548614, Cur Avg Loss: 0.25419193, Log Avg loss: 0.14412312, Global Avg Loss: 0.93150952, Time: 0.0208 Steps: 61090, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001488, Sample Num: 23808, Cur Loss: 0.23982310, Cur Avg Loss: 0.25418610, Log Avg loss: 0.25332512, Global Avg Loss: 0.93139852, Time: 0.0208 Steps: 61100, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001498, Sample Num: 23968, Cur Loss: 0.14411257, Cur Avg Loss: 0.25386163, Log Avg loss: 0.20558005, Global Avg Loss: 0.93127975, Time: 0.0208 Steps: 61110, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001508, Sample Num: 24128, Cur Loss: 0.23960236, Cur Avg Loss: 0.25400430, Log Avg loss: 0.27537565, Global Avg Loss: 0.93117244, Time: 0.0208 Steps: 61120, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001518, Sample Num: 24288, Cur Loss: 0.38679874, Cur Avg Loss: 0.25409707, Log Avg loss: 0.26808703, Global Avg Loss: 0.93106397, Time: 0.0208 Steps: 61130, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001528, Sample Num: 24448, Cur Loss: 0.33352029, Cur Avg Loss: 0.25406548, Log Avg loss: 0.24927112, Global Avg Loss: 0.93095245, Time: 0.0208 Steps: 61140, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001538, Sample Num: 24608, Cur Loss: 0.21964784, Cur Avg Loss: 0.25346302, Log Avg loss: 0.16140653, Global Avg Loss: 0.93082661, Time: 0.0246 Steps: 61150, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001548, Sample Num: 24768, Cur Loss: 0.16290836, Cur Avg Loss: 0.25303377, Log Avg loss: 0.18701557, Global Avg Loss: 0.93070499, Time: 0.0208 Steps: 61160, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001558, Sample Num: 24928, Cur Loss: 0.67108834, Cur Avg Loss: 0.25315293, Log Avg loss: 0.27159846, Global Avg Loss: 0.93059724, Time: 0.0208 Steps: 61170, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001568, Sample Num: 25088, Cur Loss: 0.03983531, Cur Avg Loss: 0.25312046, Log Avg loss: 0.24806180, Global Avg Loss: 0.93048568, Time: 0.0208 Steps: 61180, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001578, Sample Num: 25248, Cur Loss: 0.19779429, Cur Avg Loss: 0.25360736, Log Avg loss: 0.32995292, Global Avg Loss: 0.93038754, Time: 0.0208 Steps: 61190, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001588, Sample Num: 25408, Cur Loss: 0.44899118, Cur Avg Loss: 0.25421194, Log Avg loss: 0.34961515, Global Avg Loss: 0.93029264, Time: 0.0208 Steps: 61200, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001598, Sample Num: 25568, Cur Loss: 0.12386091, Cur Avg Loss: 0.25489614, Log Avg loss: 0.36354752, Global Avg Loss: 0.93020005, Time: 0.0208 Steps: 61210, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001608, Sample Num: 25728, Cur Loss: 0.19282967, Cur Avg Loss: 0.25525549, Log Avg loss: 0.31267936, Global Avg Loss: 0.93009918, Time: 0.0209 Steps: 61220, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001618, Sample Num: 25888, Cur Loss: 0.07292031, Cur Avg Loss: 0.25534452, Log Avg loss: 0.26966026, Global Avg Loss: 0.92999132, Time: 0.0209 Steps: 61230, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001628, Sample Num: 26048, Cur Loss: 0.11310390, Cur Avg Loss: 0.25486543, Log Avg loss: 0.17734885, Global Avg Loss: 0.92986842, Time: 0.0209 Steps: 61240, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001638, Sample Num: 26208, Cur Loss: 0.41021019, Cur Avg Loss: 0.25484754, Log Avg loss: 0.25193538, Global Avg Loss: 0.92975773, Time: 0.0209 Steps: 61250, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001648, Sample Num: 26368, Cur Loss: 0.29572132, Cur Avg Loss: 0.25514792, Log Avg loss: 0.30435004, Global Avg Loss: 0.92965564, Time: 0.0210 Steps: 61260, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001658, Sample Num: 26528, Cur Loss: 0.09160535, Cur Avg Loss: 0.25520152, Log Avg loss: 0.26403402, Global Avg Loss: 0.92954700, Time: 0.0209 Steps: 61270, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001668, Sample Num: 26688, Cur Loss: 0.26250991, Cur Avg Loss: 0.25514226, Log Avg loss: 0.24531788, Global Avg Loss: 0.92943535, Time: 0.0209 Steps: 61280, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001678, Sample Num: 26848, Cur Loss: 0.26877394, Cur Avg Loss: 0.25555864, Log Avg loss: 0.32500990, Global Avg Loss: 0.92933673, Time: 0.0209 Steps: 61290, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001688, Sample Num: 27008, Cur Loss: 0.15738161, Cur Avg Loss: 0.25580153, Log Avg loss: 0.29655910, Global Avg Loss: 0.92923350, Time: 0.0209 Steps: 61300, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001698, Sample Num: 27168, Cur Loss: 0.43996689, Cur Avg Loss: 0.25622257, Log Avg loss: 0.32729345, Global Avg Loss: 0.92913533, Time: 0.0209 Steps: 61310, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001708, Sample Num: 27328, Cur Loss: 0.09587348, Cur Avg Loss: 0.25601377, Log Avg loss: 0.22056027, Global Avg Loss: 0.92901977, Time: 0.0209 Steps: 61320, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001718, Sample Num: 27488, Cur Loss: 0.11375631, Cur Avg Loss: 0.25568266, Log Avg loss: 0.19912857, Global Avg Loss: 0.92890076, Time: 0.0209 Steps: 61330, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001728, Sample Num: 27648, Cur Loss: 0.30413765, Cur Avg Loss: 0.25533248, Log Avg loss: 0.19517111, Global Avg Loss: 0.92878114, Time: 0.0209 Steps: 61340, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001738, Sample Num: 27808, Cur Loss: 0.11190894, Cur Avg Loss: 0.25477544, Log Avg loss: 0.15851993, Global Avg Loss: 0.92865559, Time: 0.0209 Steps: 61350, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001748, Sample Num: 27968, Cur Loss: 0.22868866, Cur Avg Loss: 0.25439737, Log Avg loss: 0.18868864, Global Avg Loss: 0.92853500, Time: 0.0209 Steps: 61360, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001758, Sample Num: 28128, Cur Loss: 0.22038162, Cur Avg Loss: 0.25421083, Log Avg loss: 0.22160403, Global Avg Loss: 0.92841981, Time: 0.0209 Steps: 61370, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001768, Sample Num: 28288, Cur Loss: 0.56740785, Cur Avg Loss: 0.25444452, Log Avg loss: 0.29552681, Global Avg Loss: 0.92831670, Time: 0.0209 Steps: 61380, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001778, Sample Num: 28448, Cur Loss: 0.26966634, Cur Avg Loss: 0.25476269, Log Avg loss: 0.31101403, Global Avg Loss: 0.92821614, Time: 0.0209 Steps: 61390, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001788, Sample Num: 28608, Cur Loss: 0.14755541, Cur Avg Loss: 0.25487079, Log Avg loss: 0.27409168, Global Avg Loss: 0.92810961, Time: 0.0209 Steps: 61400, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001798, Sample Num: 28768, Cur Loss: 0.18954644, Cur Avg Loss: 0.25463701, Log Avg loss: 0.21283741, Global Avg Loss: 0.92799313, Time: 0.0213 Steps: 61410, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001808, Sample Num: 28928, Cur Loss: 0.17194241, Cur Avg Loss: 0.25449142, Log Avg loss: 0.22831352, Global Avg Loss: 0.92787921, Time: 0.0212 Steps: 61420, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001818, Sample Num: 29088, Cur Loss: 0.24722564, Cur Avg Loss: 0.25437897, Log Avg loss: 0.23404861, Global Avg Loss: 0.92776627, Time: 0.0209 Steps: 61430, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001828, Sample Num: 29248, Cur Loss: 0.14870682, Cur Avg Loss: 0.25428873, Log Avg loss: 0.23788347, Global Avg Loss: 0.92765398, Time: 0.0209 Steps: 61440, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001838, Sample Num: 29408, Cur Loss: 0.16109759, Cur Avg Loss: 0.25468771, Log Avg loss: 0.32762004, Global Avg Loss: 0.92755634, Time: 0.0209 Steps: 61450, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001848, Sample Num: 29568, Cur Loss: 0.21600729, Cur Avg Loss: 0.25441734, Log Avg loss: 0.20472442, Global Avg Loss: 0.92743873, Time: 0.0209 Steps: 61460, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001858, Sample Num: 29728, Cur Loss: 0.07574111, Cur Avg Loss: 0.25395207, Log Avg loss: 0.16796910, Global Avg Loss: 0.92731517, Time: 0.0209 Steps: 61470, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001868, Sample Num: 29888, Cur Loss: 0.30099791, Cur Avg Loss: 0.25393416, Log Avg loss: 0.25060728, Global Avg Loss: 0.92720511, Time: 0.0209 Steps: 61480, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001878, Sample Num: 30048, Cur Loss: 0.41685167, Cur Avg Loss: 0.25407897, Log Avg loss: 0.28112883, Global Avg Loss: 0.92710004, Time: 0.0208 Steps: 61490, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001888, Sample Num: 30208, Cur Loss: 0.13406923, Cur Avg Loss: 0.25437204, Log Avg loss: 0.30941221, Global Avg Loss: 0.92699960, Time: 0.0210 Steps: 61500, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001898, Sample Num: 30368, Cur Loss: 0.50736111, Cur Avg Loss: 0.25495270, Log Avg loss: 0.36457945, Global Avg Loss: 0.92690816, Time: 0.0209 Steps: 61510, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001908, Sample Num: 30528, Cur Loss: 0.27665925, Cur Avg Loss: 0.25476068, Log Avg loss: 0.21831634, Global Avg Loss: 0.92679298, Time: 0.0210 Steps: 61520, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001918, Sample Num: 30688, Cur Loss: 0.13403192, Cur Avg Loss: 0.25467474, Log Avg loss: 0.23827673, Global Avg Loss: 0.92668108, Time: 0.0209 Steps: 61530, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001928, Sample Num: 30848, Cur Loss: 0.31942278, Cur Avg Loss: 0.25454036, Log Avg loss: 0.22876729, Global Avg Loss: 0.92656767, Time: 0.0209 Steps: 61540, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001938, Sample Num: 31008, Cur Loss: 0.18491340, Cur Avg Loss: 0.25420037, Log Avg loss: 0.18864911, Global Avg Loss: 0.92644779, Time: 0.0209 Steps: 61550, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001948, Sample Num: 31168, Cur Loss: 0.40643322, Cur Avg Loss: 0.25422802, Log Avg loss: 0.25958703, Global Avg Loss: 0.92633946, Time: 0.0209 Steps: 61560, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001958, Sample Num: 31328, Cur Loss: 0.30866075, Cur Avg Loss: 0.25447613, Log Avg loss: 0.30280742, Global Avg Loss: 0.92623819, Time: 0.0209 Steps: 61570, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001968, Sample Num: 31488, Cur Loss: 0.13948773, Cur Avg Loss: 0.25436739, Log Avg loss: 0.23307668, Global Avg Loss: 0.92612562, Time: 0.0208 Steps: 61580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001978, Sample Num: 31648, Cur Loss: 0.36192405, Cur Avg Loss: 0.25505599, Log Avg loss: 0.39057197, Global Avg Loss: 0.92603867, Time: 0.0209 Steps: 61590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001988, Sample Num: 31808, Cur Loss: 0.13482161, Cur Avg Loss: 0.25528164, Log Avg loss: 0.29991693, Global Avg Loss: 0.92593703, Time: 0.0209 Steps: 61600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001998, Sample Num: 31968, Cur Loss: 0.06902070, Cur Avg Loss: 0.25481450, Log Avg loss: 0.16194574, Global Avg Loss: 0.92581302, Time: 0.0209 Steps: 61610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002008, Sample Num: 32128, Cur Loss: 0.13277169, Cur Avg Loss: 0.25469750, Log Avg loss: 0.23132147, Global Avg Loss: 0.92570032, Time: 0.0209 Steps: 61620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002018, Sample Num: 32288, Cur Loss: 0.26568222, Cur Avg Loss: 0.25486025, Log Avg loss: 0.28754115, Global Avg Loss: 0.92559677, Time: 0.0208 Steps: 61630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002028, Sample Num: 32448, Cur Loss: 0.57146776, Cur Avg Loss: 0.25529894, Log Avg loss: 0.34382581, Global Avg Loss: 0.92550239, Time: 0.0209 Steps: 61640, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002038, Sample Num: 32608, Cur Loss: 0.29561940, Cur Avg Loss: 0.25603677, Log Avg loss: 0.40566784, Global Avg Loss: 0.92541807, Time: 0.0210 Steps: 61650, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002048, Sample Num: 32768, Cur Loss: 0.15662232, Cur Avg Loss: 0.25640687, Log Avg loss: 0.33183467, Global Avg Loss: 0.92532180, Time: 0.0255 Steps: 61660, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002058, Sample Num: 32928, Cur Loss: 0.10597911, Cur Avg Loss: 0.25636109, Log Avg loss: 0.24698370, Global Avg Loss: 0.92521180, Time: 0.0210 Steps: 61670, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002068, Sample Num: 33088, Cur Loss: 0.83895224, Cur Avg Loss: 0.25665399, Log Avg loss: 0.31693362, Global Avg Loss: 0.92511319, Time: 0.0209 Steps: 61680, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002078, Sample Num: 33248, Cur Loss: 0.48674116, Cur Avg Loss: 0.25654278, Log Avg loss: 0.23354533, Global Avg Loss: 0.92500108, Time: 0.0209 Steps: 61690, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002088, Sample Num: 33408, Cur Loss: 0.03975538, Cur Avg Loss: 0.25632864, Log Avg loss: 0.21182967, Global Avg Loss: 0.92488550, Time: 0.0209 Steps: 61700, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002098, Sample Num: 33568, Cur Loss: 0.10411857, Cur Avg Loss: 0.25587873, Log Avg loss: 0.16193694, Global Avg Loss: 0.92476186, Time: 0.0209 Steps: 61710, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002108, Sample Num: 33728, Cur Loss: 0.05207738, Cur Avg Loss: 0.25588948, Log Avg loss: 0.25814618, Global Avg Loss: 0.92465385, Time: 0.0209 Steps: 61720, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002118, Sample Num: 33888, Cur Loss: 0.12057293, Cur Avg Loss: 0.25545479, Log Avg loss: 0.16382119, Global Avg Loss: 0.92453060, Time: 0.0210 Steps: 61730, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002128, Sample Num: 34048, Cur Loss: 0.09572653, Cur Avg Loss: 0.25540090, Log Avg loss: 0.24398634, Global Avg Loss: 0.92442038, Time: 0.0210 Steps: 61740, Updated lr: 0.000042 ***** Running evaluation checkpoint-61741 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-61741 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.574178, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.296637, "eval_total_loss": 208.535835, "eval_mae": 0.426344, "eval_mse": 0.296668, "eval_r2": 0.811418, "eval_sp_statistic": 0.887583, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.91589, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.339636, "test_total_loss": 170.497278, "test_mae": 0.374944, "test_mse": 0.339759, "test_r2": 0.780717, "test_sp_statistic": 0.865536, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.912428, "test_ps_pvalue": 0.0, "lr": 4.239829302987198e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9244122022727266, "train_cur_epoch_loss": 543.91291436553, "train_cur_epoch_avg_loss": 0.2554781185371207, "train_cur_epoch_time": 44.574177742004395, "train_cur_epoch_avg_time": 0.020936673434478345, "epoch": 29, "step": 61741} ################################################## Training, Epoch: 0030, Batch: 000009, Sample Num: 144, Cur Loss: 0.36380443, Cur Avg Loss: 0.28683302, Log Avg loss: 0.30013050, Global Avg Loss: 0.92431928, Time: 0.0210 Steps: 61750, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000019, Sample Num: 304, Cur Loss: 0.10434040, Cur Avg Loss: 0.24820733, Log Avg loss: 0.21344421, Global Avg Loss: 0.92420417, Time: 0.0209 Steps: 61760, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000029, Sample Num: 464, Cur Loss: 0.43177322, Cur Avg Loss: 0.21928671, Log Avg loss: 0.16433754, Global Avg Loss: 0.92408116, Time: 0.0209 Steps: 61770, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000039, Sample Num: 624, Cur Loss: 0.37785497, Cur Avg Loss: 0.22792579, Log Avg loss: 0.25297911, Global Avg Loss: 0.92397253, Time: 0.0209 Steps: 61780, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000049, Sample Num: 784, Cur Loss: 0.24223240, Cur Avg Loss: 0.23919387, Log Avg loss: 0.28313939, Global Avg Loss: 0.92386882, Time: 0.0209 Steps: 61790, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000059, Sample Num: 944, Cur Loss: 0.54415822, Cur Avg Loss: 0.23418636, Log Avg loss: 0.20964959, Global Avg Loss: 0.92375325, Time: 0.0209 Steps: 61800, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000069, Sample Num: 1104, Cur Loss: 0.12745363, Cur Avg Loss: 0.23248052, Log Avg loss: 0.22241601, Global Avg Loss: 0.92363978, Time: 0.0209 Steps: 61810, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000079, Sample Num: 1264, Cur Loss: 0.19042669, Cur Avg Loss: 0.23487017, Log Avg loss: 0.25135876, Global Avg Loss: 0.92353103, Time: 0.0209 Steps: 61820, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000089, Sample Num: 1424, Cur Loss: 0.30325580, Cur Avg Loss: 0.23233823, Log Avg loss: 0.21233590, Global Avg Loss: 0.92341601, Time: 0.0209 Steps: 61830, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000099, Sample Num: 1584, Cur Loss: 0.17195907, Cur Avg Loss: 0.23203716, Log Avg loss: 0.22935768, Global Avg Loss: 0.92330378, Time: 0.0209 Steps: 61840, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000109, Sample Num: 1744, Cur Loss: 0.28858095, Cur Avg Loss: 0.23181774, Log Avg loss: 0.22964543, Global Avg Loss: 0.92319162, Time: 0.0209 Steps: 61850, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000119, Sample Num: 1904, Cur Loss: 0.58449906, Cur Avg Loss: 0.23774706, Log Avg loss: 0.30237673, Global Avg Loss: 0.92309127, Time: 0.0209 Steps: 61860, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000129, Sample Num: 2064, Cur Loss: 0.66716522, Cur Avg Loss: 0.25165296, Log Avg loss: 0.41713317, Global Avg Loss: 0.92300949, Time: 0.0209 Steps: 61870, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000139, Sample Num: 2224, Cur Loss: 0.50680256, Cur Avg Loss: 0.26045682, Log Avg loss: 0.37402663, Global Avg Loss: 0.92292077, Time: 0.0209 Steps: 61880, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000149, Sample Num: 2384, Cur Loss: 0.24338621, Cur Avg Loss: 0.26349158, Log Avg loss: 0.30567474, Global Avg Loss: 0.92282104, Time: 0.0209 Steps: 61890, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000159, Sample Num: 2544, Cur Loss: 0.16974685, Cur Avg Loss: 0.26191408, Log Avg loss: 0.23840931, Global Avg Loss: 0.92271047, Time: 0.0209 Steps: 61900, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000169, Sample Num: 2704, Cur Loss: 0.21636905, Cur Avg Loss: 0.26290639, Log Avg loss: 0.27868402, Global Avg Loss: 0.92260644, Time: 0.0210 Steps: 61910, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000179, Sample Num: 2864, Cur Loss: 0.34893936, Cur Avg Loss: 0.26753301, Log Avg loss: 0.34572291, Global Avg Loss: 0.92251328, Time: 0.0209 Steps: 61920, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000189, Sample Num: 3024, Cur Loss: 0.37787190, Cur Avg Loss: 0.26593068, Log Avg loss: 0.23724906, Global Avg Loss: 0.92240263, Time: 0.0209 Steps: 61930, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000199, Sample Num: 3184, Cur Loss: 0.26235968, Cur Avg Loss: 0.26580457, Log Avg loss: 0.26342106, Global Avg Loss: 0.92229624, Time: 0.0209 Steps: 61940, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000209, Sample Num: 3344, Cur Loss: 0.11439912, Cur Avg Loss: 0.26037871, Log Avg loss: 0.15240399, Global Avg Loss: 0.92217196, Time: 0.0209 Steps: 61950, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000219, Sample Num: 3504, Cur Loss: 0.09198348, Cur Avg Loss: 0.25742507, Log Avg loss: 0.19569417, Global Avg Loss: 0.92205471, Time: 0.0209 Steps: 61960, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000229, Sample Num: 3664, Cur Loss: 0.07367378, Cur Avg Loss: 0.25630194, Log Avg loss: 0.23170539, Global Avg Loss: 0.92194331, Time: 0.0209 Steps: 61970, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000239, Sample Num: 3824, Cur Loss: 0.14969474, Cur Avg Loss: 0.25715219, Log Avg loss: 0.27662280, Global Avg Loss: 0.92183919, Time: 0.0209 Steps: 61980, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000249, Sample Num: 3984, Cur Loss: 0.20263653, Cur Avg Loss: 0.25330606, Log Avg loss: 0.16138358, Global Avg Loss: 0.92171652, Time: 0.0209 Steps: 61990, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000259, Sample Num: 4144, Cur Loss: 0.21555635, Cur Avg Loss: 0.25225020, Log Avg loss: 0.22595931, Global Avg Loss: 0.92160430, Time: 0.0248 Steps: 62000, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000269, Sample Num: 4304, Cur Loss: 0.29052079, Cur Avg Loss: 0.25165802, Log Avg loss: 0.23632055, Global Avg Loss: 0.92149379, Time: 0.0211 Steps: 62010, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000279, Sample Num: 4464, Cur Loss: 0.14950898, Cur Avg Loss: 0.24909775, Log Avg loss: 0.18022661, Global Avg Loss: 0.92137427, Time: 0.0210 Steps: 62020, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000289, Sample Num: 4624, Cur Loss: 0.20649302, Cur Avg Loss: 0.24698613, Log Avg loss: 0.18807172, Global Avg Loss: 0.92125605, Time: 0.0210 Steps: 62030, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000299, Sample Num: 4784, Cur Loss: 0.17944637, Cur Avg Loss: 0.24385032, Log Avg loss: 0.15322551, Global Avg Loss: 0.92113225, Time: 0.0211 Steps: 62040, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000309, Sample Num: 4944, Cur Loss: 0.23383740, Cur Avg Loss: 0.24412430, Log Avg loss: 0.25231617, Global Avg Loss: 0.92102447, Time: 0.0210 Steps: 62050, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000319, Sample Num: 5104, Cur Loss: 0.28111359, Cur Avg Loss: 0.24436108, Log Avg loss: 0.25167766, Global Avg Loss: 0.92091661, Time: 0.0210 Steps: 62060, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000329, Sample Num: 5264, Cur Loss: 0.13564946, Cur Avg Loss: 0.24646767, Log Avg loss: 0.31366779, Global Avg Loss: 0.92081878, Time: 0.0211 Steps: 62070, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000339, Sample Num: 5424, Cur Loss: 0.19370396, Cur Avg Loss: 0.24810812, Log Avg loss: 0.30207903, Global Avg Loss: 0.92071911, Time: 0.0210 Steps: 62080, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000349, Sample Num: 5584, Cur Loss: 0.16914955, Cur Avg Loss: 0.24789209, Log Avg loss: 0.24056856, Global Avg Loss: 0.92060957, Time: 0.0211 Steps: 62090, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000359, Sample Num: 5744, Cur Loss: 0.07181863, Cur Avg Loss: 0.24867435, Log Avg loss: 0.27597545, Global Avg Loss: 0.92050576, Time: 0.0210 Steps: 62100, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000369, Sample Num: 5904, Cur Loss: 0.22592899, Cur Avg Loss: 0.24671835, Log Avg loss: 0.17649789, Global Avg Loss: 0.92038597, Time: 0.0210 Steps: 62110, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000379, Sample Num: 6064, Cur Loss: 0.24749710, Cur Avg Loss: 0.24643355, Log Avg loss: 0.23592424, Global Avg Loss: 0.92027579, Time: 0.0211 Steps: 62120, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000389, Sample Num: 6224, Cur Loss: 0.07733841, Cur Avg Loss: 0.24684651, Log Avg loss: 0.26249788, Global Avg Loss: 0.92016992, Time: 0.0210 Steps: 62130, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000399, Sample Num: 6384, Cur Loss: 0.50502181, Cur Avg Loss: 0.24651229, Log Avg loss: 0.23351111, Global Avg Loss: 0.92005942, Time: 0.0210 Steps: 62140, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000409, Sample Num: 6544, Cur Loss: 0.25820622, Cur Avg Loss: 0.24520500, Log Avg loss: 0.19304398, Global Avg Loss: 0.91994244, Time: 0.0211 Steps: 62150, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000419, Sample Num: 6704, Cur Loss: 0.16426331, Cur Avg Loss: 0.24960840, Log Avg loss: 0.42970781, Global Avg Loss: 0.91986357, Time: 0.0211 Steps: 62160, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000429, Sample Num: 6864, Cur Loss: 0.08539125, Cur Avg Loss: 0.24896027, Log Avg loss: 0.22180348, Global Avg Loss: 0.91975129, Time: 0.0211 Steps: 62170, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000439, Sample Num: 7024, Cur Loss: 0.11575176, Cur Avg Loss: 0.24823633, Log Avg loss: 0.21717923, Global Avg Loss: 0.91963830, Time: 0.0210 Steps: 62180, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000449, Sample Num: 7184, Cur Loss: 0.28658697, Cur Avg Loss: 0.24676446, Log Avg loss: 0.18214921, Global Avg Loss: 0.91951971, Time: 0.0210 Steps: 62190, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000459, Sample Num: 7344, Cur Loss: 0.10819676, Cur Avg Loss: 0.24499574, Log Avg loss: 0.16558046, Global Avg Loss: 0.91939850, Time: 0.0211 Steps: 62200, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000469, Sample Num: 7504, Cur Loss: 0.57190526, Cur Avg Loss: 0.24675509, Log Avg loss: 0.32750932, Global Avg Loss: 0.91930336, Time: 0.0210 Steps: 62210, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000479, Sample Num: 7664, Cur Loss: 0.56484628, Cur Avg Loss: 0.24877759, Log Avg loss: 0.34363267, Global Avg Loss: 0.91921084, Time: 0.0211 Steps: 62220, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000489, Sample Num: 7824, Cur Loss: 0.29781154, Cur Avg Loss: 0.24921956, Log Avg loss: 0.27038989, Global Avg Loss: 0.91910657, Time: 0.0210 Steps: 62230, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000499, Sample Num: 7984, Cur Loss: 0.24074054, Cur Avg Loss: 0.24757612, Log Avg loss: 0.16721193, Global Avg Loss: 0.91898577, Time: 0.0210 Steps: 62240, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000509, Sample Num: 8144, Cur Loss: 0.38232177, Cur Avg Loss: 0.24888392, Log Avg loss: 0.31414333, Global Avg Loss: 0.91888861, Time: 0.0210 Steps: 62250, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000519, Sample Num: 8304, Cur Loss: 0.12329748, Cur Avg Loss: 0.24677584, Log Avg loss: 0.13947450, Global Avg Loss: 0.91876342, Time: 0.0211 Steps: 62260, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000529, Sample Num: 8464, Cur Loss: 0.22467761, Cur Avg Loss: 0.24491990, Log Avg loss: 0.14859668, Global Avg Loss: 0.91863974, Time: 0.0210 Steps: 62270, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000539, Sample Num: 8624, Cur Loss: 0.20585768, Cur Avg Loss: 0.24450876, Log Avg loss: 0.22275950, Global Avg Loss: 0.91852800, Time: 0.0210 Steps: 62280, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000549, Sample Num: 8784, Cur Loss: 0.23296325, Cur Avg Loss: 0.24390935, Log Avg loss: 0.21160096, Global Avg Loss: 0.91841451, Time: 0.0210 Steps: 62290, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000559, Sample Num: 8944, Cur Loss: 0.07939665, Cur Avg Loss: 0.24284149, Log Avg loss: 0.18421600, Global Avg Loss: 0.91829666, Time: 0.0210 Steps: 62300, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000569, Sample Num: 9104, Cur Loss: 0.43555635, Cur Avg Loss: 0.24528418, Log Avg loss: 0.38183076, Global Avg Loss: 0.91821057, Time: 0.0211 Steps: 62310, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000579, Sample Num: 9264, Cur Loss: 0.40866554, Cur Avg Loss: 0.24699180, Log Avg loss: 0.34415532, Global Avg Loss: 0.91811845, Time: 0.0211 Steps: 62320, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000589, Sample Num: 9424, Cur Loss: 0.38255930, Cur Avg Loss: 0.24789147, Log Avg loss: 0.29998215, Global Avg Loss: 0.91801928, Time: 0.0211 Steps: 62330, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000599, Sample Num: 9584, Cur Loss: 0.22459209, Cur Avg Loss: 0.24827335, Log Avg loss: 0.27076596, Global Avg Loss: 0.91791546, Time: 0.0211 Steps: 62340, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000609, Sample Num: 9744, Cur Loss: 0.13650122, Cur Avg Loss: 0.24791727, Log Avg loss: 0.22658854, Global Avg Loss: 0.91780458, Time: 0.0210 Steps: 62350, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000619, Sample Num: 9904, Cur Loss: 0.42932963, Cur Avg Loss: 0.24639652, Log Avg loss: 0.15378255, Global Avg Loss: 0.91768206, Time: 0.0210 Steps: 62360, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000629, Sample Num: 10064, Cur Loss: 0.29169777, Cur Avg Loss: 0.24561839, Log Avg loss: 0.19745197, Global Avg Loss: 0.91756658, Time: 0.0211 Steps: 62370, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000639, Sample Num: 10224, Cur Loss: 0.12000203, Cur Avg Loss: 0.24461007, Log Avg loss: 0.18118701, Global Avg Loss: 0.91744854, Time: 0.0210 Steps: 62380, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000649, Sample Num: 10384, Cur Loss: 0.54811603, Cur Avg Loss: 0.24678586, Log Avg loss: 0.38581886, Global Avg Loss: 0.91736332, Time: 0.0209 Steps: 62390, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000659, Sample Num: 10544, Cur Loss: 0.42227429, Cur Avg Loss: 0.24712100, Log Avg loss: 0.26887149, Global Avg Loss: 0.91725940, Time: 0.0210 Steps: 62400, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000669, Sample Num: 10704, Cur Loss: 0.14026415, Cur Avg Loss: 0.24626173, Log Avg loss: 0.18963568, Global Avg Loss: 0.91714281, Time: 0.0210 Steps: 62410, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000679, Sample Num: 10864, Cur Loss: 0.09154078, Cur Avg Loss: 0.24592073, Log Avg loss: 0.22310768, Global Avg Loss: 0.91703162, Time: 0.0210 Steps: 62420, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000689, Sample Num: 11024, Cur Loss: 0.18711674, Cur Avg Loss: 0.24547104, Log Avg loss: 0.21493745, Global Avg Loss: 0.91691916, Time: 0.0210 Steps: 62430, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000699, Sample Num: 11184, Cur Loss: 0.09625927, Cur Avg Loss: 0.24552336, Log Avg loss: 0.24912803, Global Avg Loss: 0.91681221, Time: 0.0210 Steps: 62440, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000709, Sample Num: 11344, Cur Loss: 0.38590735, Cur Avg Loss: 0.24517763, Log Avg loss: 0.22101126, Global Avg Loss: 0.91670080, Time: 0.0210 Steps: 62450, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000719, Sample Num: 11504, Cur Loss: 0.08666184, Cur Avg Loss: 0.24490883, Log Avg loss: 0.22585121, Global Avg Loss: 0.91659019, Time: 0.0210 Steps: 62460, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000729, Sample Num: 11664, Cur Loss: 0.23328799, Cur Avg Loss: 0.24472003, Log Avg loss: 0.23114521, Global Avg Loss: 0.91648047, Time: 0.0210 Steps: 62470, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000739, Sample Num: 11824, Cur Loss: 0.29639009, Cur Avg Loss: 0.24420273, Log Avg loss: 0.20649154, Global Avg Loss: 0.91636683, Time: 0.0210 Steps: 62480, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000749, Sample Num: 11984, Cur Loss: 0.34136397, Cur Avg Loss: 0.24466220, Log Avg loss: 0.27861663, Global Avg Loss: 0.91626477, Time: 0.0210 Steps: 62490, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000759, Sample Num: 12144, Cur Loss: 0.40642089, Cur Avg Loss: 0.24610906, Log Avg loss: 0.35447891, Global Avg Loss: 0.91617489, Time: 0.0210 Steps: 62500, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000769, Sample Num: 12304, Cur Loss: 0.12448081, Cur Avg Loss: 0.24615151, Log Avg loss: 0.24937360, Global Avg Loss: 0.91606822, Time: 0.0247 Steps: 62510, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000779, Sample Num: 12464, Cur Loss: 0.26580536, Cur Avg Loss: 0.24641695, Log Avg loss: 0.26682958, Global Avg Loss: 0.91596437, Time: 0.0211 Steps: 62520, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000789, Sample Num: 12624, Cur Loss: 0.42570841, Cur Avg Loss: 0.24636098, Log Avg loss: 0.24200067, Global Avg Loss: 0.91585659, Time: 0.0210 Steps: 62530, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000799, Sample Num: 12784, Cur Loss: 0.18587747, Cur Avg Loss: 0.24548460, Log Avg loss: 0.17633821, Global Avg Loss: 0.91573834, Time: 0.0210 Steps: 62540, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000809, Sample Num: 12944, Cur Loss: 0.09006348, Cur Avg Loss: 0.24551626, Log Avg loss: 0.24804578, Global Avg Loss: 0.91563160, Time: 0.0210 Steps: 62550, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000819, Sample Num: 13104, Cur Loss: 0.24998245, Cur Avg Loss: 0.24513211, Log Avg loss: 0.21405413, Global Avg Loss: 0.91551945, Time: 0.0210 Steps: 62560, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000829, Sample Num: 13264, Cur Loss: 0.36263636, Cur Avg Loss: 0.24614987, Log Avg loss: 0.32950488, Global Avg Loss: 0.91542580, Time: 0.0210 Steps: 62570, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000839, Sample Num: 13424, Cur Loss: 0.26832864, Cur Avg Loss: 0.24603115, Log Avg loss: 0.23618873, Global Avg Loss: 0.91531726, Time: 0.0210 Steps: 62580, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000849, Sample Num: 13584, Cur Loss: 0.41402462, Cur Avg Loss: 0.24591169, Log Avg loss: 0.23588905, Global Avg Loss: 0.91520870, Time: 0.0211 Steps: 62590, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000859, Sample Num: 13744, Cur Loss: 1.29714870, Cur Avg Loss: 0.24718519, Log Avg loss: 0.35530598, Global Avg Loss: 0.91511926, Time: 0.0210 Steps: 62600, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000869, Sample Num: 13904, Cur Loss: 0.32612100, Cur Avg Loss: 0.24753672, Log Avg loss: 0.27773273, Global Avg Loss: 0.91501746, Time: 0.0210 Steps: 62610, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000879, Sample Num: 14064, Cur Loss: 0.21275602, Cur Avg Loss: 0.24797290, Log Avg loss: 0.28587746, Global Avg Loss: 0.91491699, Time: 0.0210 Steps: 62620, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000889, Sample Num: 14224, Cur Loss: 0.37180573, Cur Avg Loss: 0.24908518, Log Avg loss: 0.34685387, Global Avg Loss: 0.91482629, Time: 0.0210 Steps: 62630, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000899, Sample Num: 14384, Cur Loss: 0.09764972, Cur Avg Loss: 0.24844398, Log Avg loss: 0.19144152, Global Avg Loss: 0.91471081, Time: 0.0210 Steps: 62640, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000909, Sample Num: 14544, Cur Loss: 0.28292382, Cur Avg Loss: 0.24819191, Log Avg loss: 0.22553080, Global Avg Loss: 0.91460080, Time: 0.0210 Steps: 62650, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000919, Sample Num: 14704, Cur Loss: 0.16124845, Cur Avg Loss: 0.24844473, Log Avg loss: 0.27142591, Global Avg Loss: 0.91449816, Time: 0.0210 Steps: 62660, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000929, Sample Num: 14864, Cur Loss: 0.47016954, Cur Avg Loss: 0.24850571, Log Avg loss: 0.25410993, Global Avg Loss: 0.91439278, Time: 0.0211 Steps: 62670, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000939, Sample Num: 15024, Cur Loss: 0.13751817, Cur Avg Loss: 0.24773211, Log Avg loss: 0.17586505, Global Avg Loss: 0.91427496, Time: 0.0210 Steps: 62680, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000949, Sample Num: 15184, Cur Loss: 0.05856050, Cur Avg Loss: 0.24711820, Log Avg loss: 0.18947190, Global Avg Loss: 0.91415934, Time: 0.0210 Steps: 62690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000959, Sample Num: 15344, Cur Loss: 0.08554827, Cur Avg Loss: 0.24638483, Log Avg loss: 0.17678764, Global Avg Loss: 0.91404174, Time: 0.0210 Steps: 62700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000969, Sample Num: 15504, Cur Loss: 0.15329607, Cur Avg Loss: 0.24611080, Log Avg loss: 0.21983164, Global Avg Loss: 0.91393103, Time: 0.0210 Steps: 62710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000979, Sample Num: 15664, Cur Loss: 0.07551736, Cur Avg Loss: 0.24561631, Log Avg loss: 0.19769971, Global Avg Loss: 0.91381684, Time: 0.0210 Steps: 62720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000989, Sample Num: 15824, Cur Loss: 0.29449731, Cur Avg Loss: 0.24625139, Log Avg loss: 0.30842646, Global Avg Loss: 0.91372033, Time: 0.0210 Steps: 62730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000999, Sample Num: 15984, Cur Loss: 0.24112028, Cur Avg Loss: 0.24641672, Log Avg loss: 0.26276763, Global Avg Loss: 0.91361658, Time: 0.0210 Steps: 62740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001009, Sample Num: 16144, Cur Loss: 0.06609085, Cur Avg Loss: 0.24619040, Log Avg loss: 0.22358086, Global Avg Loss: 0.91350661, Time: 0.0210 Steps: 62750, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001019, Sample Num: 16304, Cur Loss: 0.41524348, Cur Avg Loss: 0.24672058, Log Avg loss: 0.30021585, Global Avg Loss: 0.91340889, Time: 0.0210 Steps: 62760, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001029, Sample Num: 16464, Cur Loss: 0.35277146, Cur Avg Loss: 0.24611913, Log Avg loss: 0.18483141, Global Avg Loss: 0.91329282, Time: 0.0211 Steps: 62770, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001039, Sample Num: 16624, Cur Loss: 0.10151899, Cur Avg Loss: 0.24600358, Log Avg loss: 0.23411386, Global Avg Loss: 0.91318464, Time: 0.0210 Steps: 62780, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001049, Sample Num: 16784, Cur Loss: 0.10740864, Cur Avg Loss: 0.24612199, Log Avg loss: 0.25842392, Global Avg Loss: 0.91308036, Time: 0.0210 Steps: 62790, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001059, Sample Num: 16944, Cur Loss: 0.21403779, Cur Avg Loss: 0.24697701, Log Avg loss: 0.33666919, Global Avg Loss: 0.91298857, Time: 0.0210 Steps: 62800, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001069, Sample Num: 17104, Cur Loss: 0.18899815, Cur Avg Loss: 0.24645171, Log Avg loss: 0.19082262, Global Avg Loss: 0.91287360, Time: 0.0210 Steps: 62810, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001079, Sample Num: 17264, Cur Loss: 0.32862133, Cur Avg Loss: 0.24720162, Log Avg loss: 0.32736631, Global Avg Loss: 0.91278039, Time: 0.0210 Steps: 62820, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001089, Sample Num: 17424, Cur Loss: 0.43435007, Cur Avg Loss: 0.24710327, Log Avg loss: 0.23649171, Global Avg Loss: 0.91267276, Time: 0.0210 Steps: 62830, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001099, Sample Num: 17584, Cur Loss: 0.17149508, Cur Avg Loss: 0.24739494, Log Avg loss: 0.27915768, Global Avg Loss: 0.91257194, Time: 0.0211 Steps: 62840, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001109, Sample Num: 17744, Cur Loss: 0.25517142, Cur Avg Loss: 0.24685442, Log Avg loss: 0.18745178, Global Avg Loss: 0.91245657, Time: 0.0210 Steps: 62850, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001119, Sample Num: 17904, Cur Loss: 0.14157632, Cur Avg Loss: 0.24705574, Log Avg loss: 0.26938160, Global Avg Loss: 0.91235427, Time: 0.0210 Steps: 62860, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001129, Sample Num: 18064, Cur Loss: 0.31134409, Cur Avg Loss: 0.24701266, Log Avg loss: 0.24219222, Global Avg Loss: 0.91224767, Time: 0.0210 Steps: 62870, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001139, Sample Num: 18224, Cur Loss: 0.24912901, Cur Avg Loss: 0.24666836, Log Avg loss: 0.20779699, Global Avg Loss: 0.91213564, Time: 0.0210 Steps: 62880, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001149, Sample Num: 18384, Cur Loss: 0.31876153, Cur Avg Loss: 0.24629117, Log Avg loss: 0.20332917, Global Avg Loss: 0.91202293, Time: 0.0210 Steps: 62890, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001159, Sample Num: 18544, Cur Loss: 0.33048177, Cur Avg Loss: 0.24625265, Log Avg loss: 0.24182647, Global Avg Loss: 0.91191638, Time: 0.0210 Steps: 62900, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001169, Sample Num: 18704, Cur Loss: 0.07185566, Cur Avg Loss: 0.24640415, Log Avg loss: 0.26396345, Global Avg Loss: 0.91181339, Time: 0.0210 Steps: 62910, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001179, Sample Num: 18864, Cur Loss: 0.50255942, Cur Avg Loss: 0.24620566, Log Avg loss: 0.22300205, Global Avg Loss: 0.91170391, Time: 0.0210 Steps: 62920, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001189, Sample Num: 19024, Cur Loss: 0.09416445, Cur Avg Loss: 0.24631418, Log Avg loss: 0.25910809, Global Avg Loss: 0.91160021, Time: 0.0210 Steps: 62930, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001199, Sample Num: 19184, Cur Loss: 0.29943520, Cur Avg Loss: 0.24617468, Log Avg loss: 0.22958853, Global Avg Loss: 0.91149185, Time: 0.0210 Steps: 62940, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001209, Sample Num: 19344, Cur Loss: 0.47935838, Cur Avg Loss: 0.24663221, Log Avg loss: 0.30148974, Global Avg Loss: 0.91139495, Time: 0.0210 Steps: 62950, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001219, Sample Num: 19504, Cur Loss: 0.26869333, Cur Avg Loss: 0.24592504, Log Avg loss: 0.16042859, Global Avg Loss: 0.91127567, Time: 0.0210 Steps: 62960, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001229, Sample Num: 19664, Cur Loss: 0.07362454, Cur Avg Loss: 0.24600517, Log Avg loss: 0.25577346, Global Avg Loss: 0.91117158, Time: 0.0210 Steps: 62970, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001239, Sample Num: 19824, Cur Loss: 0.18626586, Cur Avg Loss: 0.24569844, Log Avg loss: 0.20800133, Global Avg Loss: 0.91105993, Time: 0.0210 Steps: 62980, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001249, Sample Num: 19984, Cur Loss: 0.27631652, Cur Avg Loss: 0.24497252, Log Avg loss: 0.15503031, Global Avg Loss: 0.91093990, Time: 0.0209 Steps: 62990, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001259, Sample Num: 20144, Cur Loss: 0.37784529, Cur Avg Loss: 0.24529246, Log Avg loss: 0.28525276, Global Avg Loss: 0.91084059, Time: 0.0210 Steps: 63000, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001269, Sample Num: 20304, Cur Loss: 0.21770717, Cur Avg Loss: 0.24534185, Log Avg loss: 0.25156001, Global Avg Loss: 0.91073596, Time: 0.0210 Steps: 63010, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001279, Sample Num: 20464, Cur Loss: 0.06033427, Cur Avg Loss: 0.24519147, Log Avg loss: 0.22610845, Global Avg Loss: 0.91062732, Time: 0.0210 Steps: 63020, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001289, Sample Num: 20624, Cur Loss: 0.18589616, Cur Avg Loss: 0.24469454, Log Avg loss: 0.18113683, Global Avg Loss: 0.91051158, Time: 0.0210 Steps: 63030, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001299, Sample Num: 20784, Cur Loss: 0.10105065, Cur Avg Loss: 0.24428439, Log Avg loss: 0.19141600, Global Avg Loss: 0.91039751, Time: 0.0210 Steps: 63040, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001309, Sample Num: 20944, Cur Loss: 0.73216420, Cur Avg Loss: 0.24483821, Log Avg loss: 0.31678015, Global Avg Loss: 0.91030336, Time: 0.0210 Steps: 63050, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001319, Sample Num: 21104, Cur Loss: 0.10856006, Cur Avg Loss: 0.24461028, Log Avg loss: 0.21477370, Global Avg Loss: 0.91019307, Time: 0.0210 Steps: 63060, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001329, Sample Num: 21264, Cur Loss: 0.11841492, Cur Avg Loss: 0.24375853, Log Avg loss: 0.13141301, Global Avg Loss: 0.91006959, Time: 0.0210 Steps: 63070, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001339, Sample Num: 21424, Cur Loss: 0.19256403, Cur Avg Loss: 0.24371300, Log Avg loss: 0.23766231, Global Avg Loss: 0.90996299, Time: 0.0209 Steps: 63080, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001349, Sample Num: 21584, Cur Loss: 0.08806302, Cur Avg Loss: 0.24407462, Log Avg loss: 0.29249546, Global Avg Loss: 0.90986512, Time: 0.0210 Steps: 63090, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001359, Sample Num: 21744, Cur Loss: 0.11004218, Cur Avg Loss: 0.24359310, Log Avg loss: 0.17863649, Global Avg Loss: 0.90974924, Time: 0.0211 Steps: 63100, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001369, Sample Num: 21904, Cur Loss: 0.28107837, Cur Avg Loss: 0.24373216, Log Avg loss: 0.26262933, Global Avg Loss: 0.90964670, Time: 0.0210 Steps: 63110, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001379, Sample Num: 22064, Cur Loss: 0.12530905, Cur Avg Loss: 0.24317083, Log Avg loss: 0.16632578, Global Avg Loss: 0.90952893, Time: 0.0210 Steps: 63120, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001389, Sample Num: 22224, Cur Loss: 0.39221400, Cur Avg Loss: 0.24307359, Log Avg loss: 0.22966419, Global Avg Loss: 0.90942124, Time: 0.0210 Steps: 63130, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001399, Sample Num: 22384, Cur Loss: 0.39539742, Cur Avg Loss: 0.24342089, Log Avg loss: 0.29166012, Global Avg Loss: 0.90932340, Time: 0.0210 Steps: 63140, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001409, Sample Num: 22544, Cur Loss: 0.09740259, Cur Avg Loss: 0.24310273, Log Avg loss: 0.19859278, Global Avg Loss: 0.90921086, Time: 0.0209 Steps: 63150, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001419, Sample Num: 22704, Cur Loss: 0.43394029, Cur Avg Loss: 0.24293314, Log Avg loss: 0.21903722, Global Avg Loss: 0.90910158, Time: 0.0210 Steps: 63160, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001429, Sample Num: 22864, Cur Loss: 0.07519796, Cur Avg Loss: 0.24322699, Log Avg loss: 0.28492473, Global Avg Loss: 0.90900277, Time: 0.0209 Steps: 63170, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001439, Sample Num: 23024, Cur Loss: 0.02831518, Cur Avg Loss: 0.24322019, Log Avg loss: 0.24224916, Global Avg Loss: 0.90889724, Time: 0.0210 Steps: 63180, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001449, Sample Num: 23184, Cur Loss: 0.46883756, Cur Avg Loss: 0.24365873, Log Avg loss: 0.30676430, Global Avg Loss: 0.90880195, Time: 0.0210 Steps: 63190, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001459, Sample Num: 23344, Cur Loss: 0.26269346, Cur Avg Loss: 0.24410996, Log Avg loss: 0.30949354, Global Avg Loss: 0.90870712, Time: 0.0210 Steps: 63200, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001469, Sample Num: 23504, Cur Loss: 0.42406657, Cur Avg Loss: 0.24401745, Log Avg loss: 0.23052002, Global Avg Loss: 0.90859983, Time: 0.0209 Steps: 63210, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001479, Sample Num: 23664, Cur Loss: 0.08903541, Cur Avg Loss: 0.24430340, Log Avg loss: 0.28630834, Global Avg Loss: 0.90850140, Time: 0.0210 Steps: 63220, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001489, Sample Num: 23824, Cur Loss: 0.13419142, Cur Avg Loss: 0.24377500, Log Avg loss: 0.16562563, Global Avg Loss: 0.90838391, Time: 0.0210 Steps: 63230, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001499, Sample Num: 23984, Cur Loss: 0.25225419, Cur Avg Loss: 0.24474051, Log Avg loss: 0.38850529, Global Avg Loss: 0.90830170, Time: 0.0209 Steps: 63240, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001509, Sample Num: 24144, Cur Loss: 0.28876281, Cur Avg Loss: 0.24455420, Log Avg loss: 0.21662508, Global Avg Loss: 0.90819235, Time: 0.0210 Steps: 63250, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001519, Sample Num: 24304, Cur Loss: 0.47547695, Cur Avg Loss: 0.24529293, Log Avg loss: 0.35676840, Global Avg Loss: 0.90810518, Time: 0.0210 Steps: 63260, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001529, Sample Num: 24464, Cur Loss: 0.08328485, Cur Avg Loss: 0.24604172, Log Avg loss: 0.35978294, Global Avg Loss: 0.90801852, Time: 0.0209 Steps: 63270, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001539, Sample Num: 24624, Cur Loss: 0.40196896, Cur Avg Loss: 0.24661417, Log Avg loss: 0.33414078, Global Avg Loss: 0.90792783, Time: 0.0247 Steps: 63280, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001549, Sample Num: 24784, Cur Loss: 0.48625034, Cur Avg Loss: 0.24725232, Log Avg loss: 0.34546349, Global Avg Loss: 0.90783896, Time: 0.0210 Steps: 63290, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001559, Sample Num: 24944, Cur Loss: 0.13446195, Cur Avg Loss: 0.24773625, Log Avg loss: 0.32269723, Global Avg Loss: 0.90774652, Time: 0.0210 Steps: 63300, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001569, Sample Num: 25104, Cur Loss: 0.06453487, Cur Avg Loss: 0.24749118, Log Avg loss: 0.20928541, Global Avg Loss: 0.90763619, Time: 0.0209 Steps: 63310, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001579, Sample Num: 25264, Cur Loss: 0.35772595, Cur Avg Loss: 0.24732258, Log Avg loss: 0.22086866, Global Avg Loss: 0.90752773, Time: 0.0210 Steps: 63320, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001589, Sample Num: 25424, Cur Loss: 0.08377628, Cur Avg Loss: 0.24720221, Log Avg loss: 0.22819626, Global Avg Loss: 0.90742047, Time: 0.0210 Steps: 63330, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001599, Sample Num: 25584, Cur Loss: 0.28904337, Cur Avg Loss: 0.24744773, Log Avg loss: 0.28646126, Global Avg Loss: 0.90732243, Time: 0.0210 Steps: 63340, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001609, Sample Num: 25744, Cur Loss: 0.20268446, Cur Avg Loss: 0.24729043, Log Avg loss: 0.22213788, Global Avg Loss: 0.90721427, Time: 0.0210 Steps: 63350, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001619, Sample Num: 25904, Cur Loss: 0.14745112, Cur Avg Loss: 0.24704211, Log Avg loss: 0.20708775, Global Avg Loss: 0.90710377, Time: 0.0209 Steps: 63360, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001629, Sample Num: 26064, Cur Loss: 0.10299736, Cur Avg Loss: 0.24730769, Log Avg loss: 0.29030448, Global Avg Loss: 0.90700644, Time: 0.0211 Steps: 63370, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001639, Sample Num: 26224, Cur Loss: 0.31476116, Cur Avg Loss: 0.24730327, Log Avg loss: 0.24658262, Global Avg Loss: 0.90690224, Time: 0.0210 Steps: 63380, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001649, Sample Num: 26384, Cur Loss: 0.52031273, Cur Avg Loss: 0.24734844, Log Avg loss: 0.25475204, Global Avg Loss: 0.90679936, Time: 0.0210 Steps: 63390, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001659, Sample Num: 26544, Cur Loss: 0.09476480, Cur Avg Loss: 0.24710928, Log Avg loss: 0.20767137, Global Avg Loss: 0.90668909, Time: 0.0210 Steps: 63400, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001669, Sample Num: 26704, Cur Loss: 0.24390981, Cur Avg Loss: 0.24715472, Log Avg loss: 0.25469402, Global Avg Loss: 0.90658626, Time: 0.0210 Steps: 63410, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001679, Sample Num: 26864, Cur Loss: 0.25420931, Cur Avg Loss: 0.24701822, Log Avg loss: 0.22423628, Global Avg Loss: 0.90647867, Time: 0.0210 Steps: 63420, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001689, Sample Num: 27024, Cur Loss: 0.22371201, Cur Avg Loss: 0.24701897, Log Avg loss: 0.24714488, Global Avg Loss: 0.90637473, Time: 0.0210 Steps: 63430, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001699, Sample Num: 27184, Cur Loss: 0.38289753, Cur Avg Loss: 0.24686843, Log Avg loss: 0.22144167, Global Avg Loss: 0.90626676, Time: 0.0210 Steps: 63440, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001709, Sample Num: 27344, Cur Loss: 0.28122228, Cur Avg Loss: 0.24654335, Log Avg loss: 0.19131265, Global Avg Loss: 0.90615408, Time: 0.0210 Steps: 63450, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001719, Sample Num: 27504, Cur Loss: 0.19295737, Cur Avg Loss: 0.24646441, Log Avg loss: 0.23297436, Global Avg Loss: 0.90604800, Time: 0.0210 Steps: 63460, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001729, Sample Num: 27664, Cur Loss: 0.44159281, Cur Avg Loss: 0.24616399, Log Avg loss: 0.19452102, Global Avg Loss: 0.90593590, Time: 0.0210 Steps: 63470, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001739, Sample Num: 27824, Cur Loss: 0.15644076, Cur Avg Loss: 0.24598260, Log Avg loss: 0.21462030, Global Avg Loss: 0.90582699, Time: 0.0210 Steps: 63480, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001749, Sample Num: 27984, Cur Loss: 0.20342508, Cur Avg Loss: 0.24595466, Log Avg loss: 0.24109647, Global Avg Loss: 0.90572229, Time: 0.0210 Steps: 63490, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001759, Sample Num: 28144, Cur Loss: 0.18873748, Cur Avg Loss: 0.24551730, Log Avg loss: 0.16902236, Global Avg Loss: 0.90560628, Time: 0.0210 Steps: 63500, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001769, Sample Num: 28304, Cur Loss: 0.06758660, Cur Avg Loss: 0.24578646, Log Avg loss: 0.29313244, Global Avg Loss: 0.90550984, Time: 0.0210 Steps: 63510, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001779, Sample Num: 28464, Cur Loss: 0.43412319, Cur Avg Loss: 0.24648124, Log Avg loss: 0.36938671, Global Avg Loss: 0.90542544, Time: 0.0210 Steps: 63520, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001789, Sample Num: 28624, Cur Loss: 0.13981953, Cur Avg Loss: 0.24639322, Log Avg loss: 0.23073517, Global Avg Loss: 0.90531924, Time: 0.0210 Steps: 63530, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001799, Sample Num: 28784, Cur Loss: 0.43128151, Cur Avg Loss: 0.24622935, Log Avg loss: 0.21691297, Global Avg Loss: 0.90521090, Time: 0.0210 Steps: 63540, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001809, Sample Num: 28944, Cur Loss: 0.19334660, Cur Avg Loss: 0.24627404, Log Avg loss: 0.25431380, Global Avg Loss: 0.90510847, Time: 0.0209 Steps: 63550, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001819, Sample Num: 29104, Cur Loss: 0.07708035, Cur Avg Loss: 0.24602672, Log Avg loss: 0.20128676, Global Avg Loss: 0.90499774, Time: 0.0209 Steps: 63560, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001829, Sample Num: 29264, Cur Loss: 0.15969078, Cur Avg Loss: 0.24580852, Log Avg loss: 0.20611698, Global Avg Loss: 0.90488780, Time: 0.0209 Steps: 63570, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001839, Sample Num: 29424, Cur Loss: 0.27834204, Cur Avg Loss: 0.24622628, Log Avg loss: 0.32263477, Global Avg Loss: 0.90479622, Time: 0.0210 Steps: 63580, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001849, Sample Num: 29584, Cur Loss: 0.23079538, Cur Avg Loss: 0.24668734, Log Avg loss: 0.33147628, Global Avg Loss: 0.90470606, Time: 0.0210 Steps: 63590, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001859, Sample Num: 29744, Cur Loss: 0.13266692, Cur Avg Loss: 0.24659786, Log Avg loss: 0.23005273, Global Avg Loss: 0.90459999, Time: 0.0209 Steps: 63600, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001869, Sample Num: 29904, Cur Loss: 0.52971840, Cur Avg Loss: 0.24643466, Log Avg loss: 0.21609626, Global Avg Loss: 0.90449175, Time: 0.0209 Steps: 63610, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001879, Sample Num: 30064, Cur Loss: 0.27688217, Cur Avg Loss: 0.24644256, Log Avg loss: 0.24791951, Global Avg Loss: 0.90438855, Time: 0.0210 Steps: 63620, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001889, Sample Num: 30224, Cur Loss: 0.18638058, Cur Avg Loss: 0.24650224, Log Avg loss: 0.25771680, Global Avg Loss: 0.90428692, Time: 0.0209 Steps: 63630, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001899, Sample Num: 30384, Cur Loss: 0.11332723, Cur Avg Loss: 0.24597615, Log Avg loss: 0.14659765, Global Avg Loss: 0.90416786, Time: 0.0209 Steps: 63640, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001909, Sample Num: 30544, Cur Loss: 0.09074594, Cur Avg Loss: 0.24570537, Log Avg loss: 0.19428310, Global Avg Loss: 0.90405633, Time: 0.0209 Steps: 63650, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001919, Sample Num: 30704, Cur Loss: 0.19171558, Cur Avg Loss: 0.24526001, Log Avg loss: 0.16024182, Global Avg Loss: 0.90393949, Time: 0.0209 Steps: 63660, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001929, Sample Num: 30864, Cur Loss: 0.52240121, Cur Avg Loss: 0.24580805, Log Avg loss: 0.35097665, Global Avg Loss: 0.90385264, Time: 0.0209 Steps: 63670, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001939, Sample Num: 31024, Cur Loss: 0.16433993, Cur Avg Loss: 0.24557374, Log Avg loss: 0.20037510, Global Avg Loss: 0.90374217, Time: 0.0209 Steps: 63680, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001949, Sample Num: 31184, Cur Loss: 0.28532624, Cur Avg Loss: 0.24536507, Log Avg loss: 0.20490399, Global Avg Loss: 0.90363244, Time: 0.0209 Steps: 63690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001959, Sample Num: 31344, Cur Loss: 0.12628657, Cur Avg Loss: 0.24521759, Log Avg loss: 0.21647374, Global Avg Loss: 0.90352457, Time: 0.0209 Steps: 63700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001969, Sample Num: 31504, Cur Loss: 0.25596213, Cur Avg Loss: 0.24535878, Log Avg loss: 0.27301786, Global Avg Loss: 0.90342560, Time: 0.0209 Steps: 63710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001979, Sample Num: 31664, Cur Loss: 0.50518054, Cur Avg Loss: 0.24557339, Log Avg loss: 0.28783011, Global Avg Loss: 0.90332899, Time: 0.0209 Steps: 63720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001989, Sample Num: 31824, Cur Loss: 0.12720633, Cur Avg Loss: 0.24526730, Log Avg loss: 0.18469198, Global Avg Loss: 0.90321623, Time: 0.0209 Steps: 63730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001999, Sample Num: 31984, Cur Loss: 0.06931152, Cur Avg Loss: 0.24524594, Log Avg loss: 0.24099715, Global Avg Loss: 0.90311234, Time: 0.0209 Steps: 63740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 002009, Sample Num: 32144, Cur Loss: 0.26023984, Cur Avg Loss: 0.24547661, Log Avg loss: 0.29158779, Global Avg Loss: 0.90301641, Time: 0.0209 Steps: 63750, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002019, Sample Num: 32304, Cur Loss: 0.20739968, Cur Avg Loss: 0.24548066, Log Avg loss: 0.24629356, Global Avg Loss: 0.90291341, Time: 0.0209 Steps: 63760, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002029, Sample Num: 32464, Cur Loss: 0.16541074, Cur Avg Loss: 0.24567381, Log Avg loss: 0.28467093, Global Avg Loss: 0.90281646, Time: 0.0209 Steps: 63770, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002039, Sample Num: 32624, Cur Loss: 0.28924409, Cur Avg Loss: 0.24601148, Log Avg loss: 0.31452522, Global Avg Loss: 0.90272423, Time: 0.0209 Steps: 63780, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002049, Sample Num: 32784, Cur Loss: 0.24391225, Cur Avg Loss: 0.24595751, Log Avg loss: 0.23495368, Global Avg Loss: 0.90261954, Time: 0.0245 Steps: 63790, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002059, Sample Num: 32944, Cur Loss: 0.16307430, Cur Avg Loss: 0.24554354, Log Avg loss: 0.16072112, Global Avg Loss: 0.90250326, Time: 0.0208 Steps: 63800, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002069, Sample Num: 33104, Cur Loss: 0.19483328, Cur Avg Loss: 0.24525002, Log Avg loss: 0.18481418, Global Avg Loss: 0.90239079, Time: 0.0208 Steps: 63810, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002079, Sample Num: 33264, Cur Loss: 0.36116156, Cur Avg Loss: 0.24512652, Log Avg loss: 0.21957310, Global Avg Loss: 0.90228380, Time: 0.0208 Steps: 63820, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002089, Sample Num: 33424, Cur Loss: 0.07498626, Cur Avg Loss: 0.24500461, Log Avg loss: 0.21965950, Global Avg Loss: 0.90217685, Time: 0.0209 Steps: 63830, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002099, Sample Num: 33584, Cur Loss: 0.17831966, Cur Avg Loss: 0.24543543, Log Avg loss: 0.33543551, Global Avg Loss: 0.90208808, Time: 0.0208 Steps: 63840, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002109, Sample Num: 33744, Cur Loss: 0.03585545, Cur Avg Loss: 0.24530266, Log Avg loss: 0.21743272, Global Avg Loss: 0.90198085, Time: 0.0208 Steps: 63850, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002119, Sample Num: 33904, Cur Loss: 0.22597185, Cur Avg Loss: 0.24496279, Log Avg loss: 0.17328512, Global Avg Loss: 0.90186674, Time: 0.0208 Steps: 63860, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002129, Sample Num: 34055, Cur Loss: 0.35605648, Cur Avg Loss: 0.24493894, Log Avg loss: 0.23988451, Global Avg Loss: 0.90176309, Time: 0.0101 Steps: 63870, Updated lr: 0.000040 ***** Running evaluation checkpoint-63870 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-63870 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.809761, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.398145, "eval_total_loss": 279.896177, "eval_mae": 0.533667, "eval_mse": 0.398129, "eval_r2": 0.746923, "eval_sp_statistic": 0.891403, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.918116, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.330067, "test_total_loss": 165.693406, "test_mae": 0.410174, "test_mse": 0.330176, "test_r2": 0.786902, "test_sp_statistic": 0.866071, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.91519, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9017630934897248, "train_cur_epoch_loss": 521.4750006683171, "train_cur_epoch_avg_loss": 0.24493893878267595, "train_cur_epoch_time": 44.8097608089447, "train_cur_epoch_avg_time": 0.021047327763712872, "epoch": 30, "step": 63870} ################################################## Training, Epoch: 0031, Batch: 000010, Sample Num: 160, Cur Loss: 0.11136656, Cur Avg Loss: 0.33860965, Log Avg loss: 0.33860965, Global Avg Loss: 0.90167494, Time: 0.0209 Steps: 63880, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000020, Sample Num: 320, Cur Loss: 0.15852644, Cur Avg Loss: 0.30186743, Log Avg loss: 0.26512521, Global Avg Loss: 0.90157530, Time: 0.0209 Steps: 63890, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000030, Sample Num: 480, Cur Loss: 0.46288007, Cur Avg Loss: 0.29237308, Log Avg loss: 0.27338437, Global Avg Loss: 0.90147699, Time: 0.0210 Steps: 63900, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000040, Sample Num: 640, Cur Loss: 0.25927207, Cur Avg Loss: 0.26190914, Log Avg loss: 0.17051731, Global Avg Loss: 0.90136262, Time: 0.0210 Steps: 63910, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000050, Sample Num: 800, Cur Loss: 0.17899168, Cur Avg Loss: 0.24360327, Log Avg loss: 0.17037979, Global Avg Loss: 0.90124826, Time: 0.0209 Steps: 63920, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000060, Sample Num: 960, Cur Loss: 0.08317721, Cur Avg Loss: 0.23209103, Log Avg loss: 0.17452984, Global Avg Loss: 0.90113459, Time: 0.0209 Steps: 63930, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000070, Sample Num: 1120, Cur Loss: 0.31483114, Cur Avg Loss: 0.23671156, Log Avg loss: 0.26443475, Global Avg Loss: 0.90103501, Time: 0.0210 Steps: 63940, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000080, Sample Num: 1280, Cur Loss: 0.07824191, Cur Avg Loss: 0.23270183, Log Avg loss: 0.20463371, Global Avg Loss: 0.90092611, Time: 0.0210 Steps: 63950, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000090, Sample Num: 1440, Cur Loss: 0.21385787, Cur Avg Loss: 0.22606625, Log Avg loss: 0.17298159, Global Avg Loss: 0.90081230, Time: 0.0210 Steps: 63960, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000100, Sample Num: 1600, Cur Loss: 0.22947118, Cur Avg Loss: 0.22616088, Log Avg loss: 0.22701256, Global Avg Loss: 0.90070697, Time: 0.0209 Steps: 63970, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000110, Sample Num: 1760, Cur Loss: 0.08634181, Cur Avg Loss: 0.22049004, Log Avg loss: 0.16378167, Global Avg Loss: 0.90059179, Time: 0.0209 Steps: 63980, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000120, Sample Num: 1920, Cur Loss: 0.30468899, Cur Avg Loss: 0.22113408, Log Avg loss: 0.22821848, Global Avg Loss: 0.90048671, Time: 0.0210 Steps: 63990, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000130, Sample Num: 2080, Cur Loss: 0.21707945, Cur Avg Loss: 0.22001824, Log Avg loss: 0.20662814, Global Avg Loss: 0.90037830, Time: 0.0209 Steps: 64000, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000140, Sample Num: 2240, Cur Loss: 0.08924606, Cur Avg Loss: 0.21440225, Log Avg loss: 0.14139438, Global Avg Loss: 0.90025973, Time: 0.0210 Steps: 64010, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000150, Sample Num: 2400, Cur Loss: 0.55045044, Cur Avg Loss: 0.21644547, Log Avg loss: 0.24505061, Global Avg Loss: 0.90015738, Time: 0.0209 Steps: 64020, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000160, Sample Num: 2560, Cur Loss: 0.17918149, Cur Avg Loss: 0.21445787, Log Avg loss: 0.18464384, Global Avg Loss: 0.90004564, Time: 0.0210 Steps: 64030, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000170, Sample Num: 2720, Cur Loss: 0.21979266, Cur Avg Loss: 0.21475817, Log Avg loss: 0.21956303, Global Avg Loss: 0.89993938, Time: 0.0211 Steps: 64040, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000180, Sample Num: 2880, Cur Loss: 0.12639633, Cur Avg Loss: 0.21497586, Log Avg loss: 0.21867654, Global Avg Loss: 0.89983301, Time: 0.0210 Steps: 64050, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000190, Sample Num: 3040, Cur Loss: 0.50872916, Cur Avg Loss: 0.21498312, Log Avg loss: 0.21511378, Global Avg Loss: 0.89972613, Time: 0.0209 Steps: 64060, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000200, Sample Num: 3200, Cur Loss: 0.10250582, Cur Avg Loss: 0.21906403, Log Avg loss: 0.29660142, Global Avg Loss: 0.89963199, Time: 0.0209 Steps: 64070, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000210, Sample Num: 3360, Cur Loss: 0.17466152, Cur Avg Loss: 0.21994303, Log Avg loss: 0.23752299, Global Avg Loss: 0.89952866, Time: 0.0209 Steps: 64080, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000220, Sample Num: 3520, Cur Loss: 0.54916906, Cur Avg Loss: 0.22174320, Log Avg loss: 0.25954666, Global Avg Loss: 0.89942881, Time: 0.0209 Steps: 64090, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000230, Sample Num: 3680, Cur Loss: 0.05747963, Cur Avg Loss: 0.21849678, Log Avg loss: 0.14707552, Global Avg Loss: 0.89931144, Time: 0.0210 Steps: 64100, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000240, Sample Num: 3840, Cur Loss: 0.34081429, Cur Avg Loss: 0.21825094, Log Avg loss: 0.21259679, Global Avg Loss: 0.89920432, Time: 0.0209 Steps: 64110, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000250, Sample Num: 4000, Cur Loss: 0.14718550, Cur Avg Loss: 0.22118466, Log Avg loss: 0.29159381, Global Avg Loss: 0.89910956, Time: 0.0210 Steps: 64120, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000260, Sample Num: 4160, Cur Loss: 0.11624826, Cur Avg Loss: 0.22101391, Log Avg loss: 0.21674510, Global Avg Loss: 0.89900316, Time: 0.0247 Steps: 64130, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000270, Sample Num: 4320, Cur Loss: 0.21849233, Cur Avg Loss: 0.22079669, Log Avg loss: 0.21514910, Global Avg Loss: 0.89889654, Time: 0.0209 Steps: 64140, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000280, Sample Num: 4480, Cur Loss: 0.36168462, Cur Avg Loss: 0.21923580, Log Avg loss: 0.17709182, Global Avg Loss: 0.89878402, Time: 0.0209 Steps: 64150, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000290, Sample Num: 4640, Cur Loss: 0.28910908, Cur Avg Loss: 0.22045197, Log Avg loss: 0.25450458, Global Avg Loss: 0.89868360, Time: 0.0210 Steps: 64160, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000300, Sample Num: 4800, Cur Loss: 0.21822533, Cur Avg Loss: 0.21896077, Log Avg loss: 0.17571600, Global Avg Loss: 0.89857094, Time: 0.0209 Steps: 64170, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000310, Sample Num: 4960, Cur Loss: 0.18526903, Cur Avg Loss: 0.21710176, Log Avg loss: 0.16133136, Global Avg Loss: 0.89845607, Time: 0.0209 Steps: 64180, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000320, Sample Num: 5120, Cur Loss: 0.20795581, Cur Avg Loss: 0.21629377, Log Avg loss: 0.19124616, Global Avg Loss: 0.89834589, Time: 0.0209 Steps: 64190, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000330, Sample Num: 5280, Cur Loss: 0.46478555, Cur Avg Loss: 0.21715918, Log Avg loss: 0.24485230, Global Avg Loss: 0.89824410, Time: 0.0209 Steps: 64200, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000340, Sample Num: 5440, Cur Loss: 0.07534413, Cur Avg Loss: 0.22026850, Log Avg loss: 0.32287610, Global Avg Loss: 0.89815449, Time: 0.0209 Steps: 64210, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000350, Sample Num: 5600, Cur Loss: 0.35381445, Cur Avg Loss: 0.22107693, Log Avg loss: 0.24856364, Global Avg Loss: 0.89805334, Time: 0.0209 Steps: 64220, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000360, Sample Num: 5760, Cur Loss: 0.04444026, Cur Avg Loss: 0.22086701, Log Avg loss: 0.21351977, Global Avg Loss: 0.89794677, Time: 0.0209 Steps: 64230, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000370, Sample Num: 5920, Cur Loss: 0.11025164, Cur Avg Loss: 0.22037593, Log Avg loss: 0.20269697, Global Avg Loss: 0.89783854, Time: 0.0209 Steps: 64240, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000380, Sample Num: 6080, Cur Loss: 0.17715883, Cur Avg Loss: 0.21978038, Log Avg loss: 0.19774517, Global Avg Loss: 0.89772958, Time: 0.0210 Steps: 64250, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000390, Sample Num: 6240, Cur Loss: 0.21636930, Cur Avg Loss: 0.22083655, Log Avg loss: 0.26097108, Global Avg Loss: 0.89763049, Time: 0.0209 Steps: 64260, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000400, Sample Num: 6400, Cur Loss: 0.42247924, Cur Avg Loss: 0.22880333, Log Avg loss: 0.53950751, Global Avg Loss: 0.89757476, Time: 0.0209 Steps: 64270, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000410, Sample Num: 6560, Cur Loss: 0.31857330, Cur Avg Loss: 0.22978421, Log Avg loss: 0.26901968, Global Avg Loss: 0.89747698, Time: 0.0209 Steps: 64280, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000420, Sample Num: 6720, Cur Loss: 0.07526122, Cur Avg Loss: 0.23166963, Log Avg loss: 0.30897152, Global Avg Loss: 0.89738544, Time: 0.0209 Steps: 64290, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000430, Sample Num: 6880, Cur Loss: 0.28433543, Cur Avg Loss: 0.23263139, Log Avg loss: 0.27302525, Global Avg Loss: 0.89728834, Time: 0.0209 Steps: 64300, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000440, Sample Num: 7040, Cur Loss: 0.38882396, Cur Avg Loss: 0.23151460, Log Avg loss: 0.18349297, Global Avg Loss: 0.89717735, Time: 0.0209 Steps: 64310, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000450, Sample Num: 7200, Cur Loss: 0.32278946, Cur Avg Loss: 0.23244957, Log Avg loss: 0.27358801, Global Avg Loss: 0.89708040, Time: 0.0209 Steps: 64320, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000460, Sample Num: 7360, Cur Loss: 0.22819924, Cur Avg Loss: 0.23168360, Log Avg loss: 0.19721515, Global Avg Loss: 0.89697160, Time: 0.0209 Steps: 64330, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000470, Sample Num: 7520, Cur Loss: 0.29561755, Cur Avg Loss: 0.23001370, Log Avg loss: 0.15319800, Global Avg Loss: 0.89685600, Time: 0.0210 Steps: 64340, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000480, Sample Num: 7680, Cur Loss: 0.07412650, Cur Avg Loss: 0.22843791, Log Avg loss: 0.15437602, Global Avg Loss: 0.89674062, Time: 0.0209 Steps: 64350, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000490, Sample Num: 7840, Cur Loss: 0.33563647, Cur Avg Loss: 0.22844781, Log Avg loss: 0.22892279, Global Avg Loss: 0.89663686, Time: 0.0209 Steps: 64360, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000500, Sample Num: 8000, Cur Loss: 0.30153498, Cur Avg Loss: 0.22926113, Log Avg loss: 0.26911374, Global Avg Loss: 0.89653937, Time: 0.0209 Steps: 64370, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000510, Sample Num: 8160, Cur Loss: 0.12895158, Cur Avg Loss: 0.22838418, Log Avg loss: 0.18453694, Global Avg Loss: 0.89642878, Time: 0.0209 Steps: 64380, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000520, Sample Num: 8320, Cur Loss: 0.33356068, Cur Avg Loss: 0.22932278, Log Avg loss: 0.27719141, Global Avg Loss: 0.89633261, Time: 0.0211 Steps: 64390, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000530, Sample Num: 8480, Cur Loss: 0.08825186, Cur Avg Loss: 0.22776418, Log Avg loss: 0.14671679, Global Avg Loss: 0.89621621, Time: 0.0210 Steps: 64400, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000540, Sample Num: 8640, Cur Loss: 0.34778053, Cur Avg Loss: 0.22683749, Log Avg loss: 0.17772332, Global Avg Loss: 0.89610466, Time: 0.0211 Steps: 64410, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000550, Sample Num: 8800, Cur Loss: 0.15822667, Cur Avg Loss: 0.22688205, Log Avg loss: 0.22928783, Global Avg Loss: 0.89600115, Time: 0.0210 Steps: 64420, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000560, Sample Num: 8960, Cur Loss: 0.27743405, Cur Avg Loss: 0.22755571, Log Avg loss: 0.26460699, Global Avg Loss: 0.89590315, Time: 0.0209 Steps: 64430, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000570, Sample Num: 9120, Cur Loss: 0.09792931, Cur Avg Loss: 0.22733509, Log Avg loss: 0.21498061, Global Avg Loss: 0.89579748, Time: 0.0210 Steps: 64440, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000580, Sample Num: 9280, Cur Loss: 0.36644024, Cur Avg Loss: 0.22772340, Log Avg loss: 0.24985730, Global Avg Loss: 0.89569726, Time: 0.0210 Steps: 64450, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000590, Sample Num: 9440, Cur Loss: 0.45549595, Cur Avg Loss: 0.22756431, Log Avg loss: 0.21833697, Global Avg Loss: 0.89559218, Time: 0.0209 Steps: 64460, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000600, Sample Num: 9600, Cur Loss: 0.39323008, Cur Avg Loss: 0.22696248, Log Avg loss: 0.19145424, Global Avg Loss: 0.89548296, Time: 0.0210 Steps: 64470, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000610, Sample Num: 9760, Cur Loss: 0.31752527, Cur Avg Loss: 0.22598620, Log Avg loss: 0.16740946, Global Avg Loss: 0.89537004, Time: 0.0209 Steps: 64480, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000620, Sample Num: 9920, Cur Loss: 0.13628383, Cur Avg Loss: 0.22478206, Log Avg loss: 0.15132928, Global Avg Loss: 0.89525467, Time: 0.0210 Steps: 64490, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000630, Sample Num: 10080, Cur Loss: 0.10870039, Cur Avg Loss: 0.22471191, Log Avg loss: 0.22036282, Global Avg Loss: 0.89515004, Time: 0.0210 Steps: 64500, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000640, Sample Num: 10240, Cur Loss: 0.06520040, Cur Avg Loss: 0.22470627, Log Avg loss: 0.22435091, Global Avg Loss: 0.89504605, Time: 0.0211 Steps: 64510, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000650, Sample Num: 10400, Cur Loss: 0.14053044, Cur Avg Loss: 0.22425635, Log Avg loss: 0.19546146, Global Avg Loss: 0.89493762, Time: 0.0210 Steps: 64520, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000660, Sample Num: 10560, Cur Loss: 0.05938699, Cur Avg Loss: 0.22467278, Log Avg loss: 0.25174071, Global Avg Loss: 0.89483795, Time: 0.0210 Steps: 64530, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000670, Sample Num: 10720, Cur Loss: 0.11250624, Cur Avg Loss: 0.22404420, Log Avg loss: 0.18255791, Global Avg Loss: 0.89472759, Time: 0.0211 Steps: 64540, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000680, Sample Num: 10880, Cur Loss: 0.15504736, Cur Avg Loss: 0.22489275, Log Avg loss: 0.28174558, Global Avg Loss: 0.89463262, Time: 0.0209 Steps: 64550, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000690, Sample Num: 11040, Cur Loss: 0.18459530, Cur Avg Loss: 0.22454839, Log Avg loss: 0.20113220, Global Avg Loss: 0.89452520, Time: 0.0209 Steps: 64560, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000700, Sample Num: 11200, Cur Loss: 0.11984158, Cur Avg Loss: 0.22479236, Log Avg loss: 0.24162613, Global Avg Loss: 0.89442409, Time: 0.0210 Steps: 64570, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000710, Sample Num: 11360, Cur Loss: 0.32051355, Cur Avg Loss: 0.22533392, Log Avg loss: 0.26324324, Global Avg Loss: 0.89432635, Time: 0.0211 Steps: 64580, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000720, Sample Num: 11520, Cur Loss: 0.06238814, Cur Avg Loss: 0.22645739, Log Avg loss: 0.30622351, Global Avg Loss: 0.89423530, Time: 0.0210 Steps: 64590, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000730, Sample Num: 11680, Cur Loss: 0.17810293, Cur Avg Loss: 0.22679964, Log Avg loss: 0.25144175, Global Avg Loss: 0.89413580, Time: 0.0209 Steps: 64600, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000740, Sample Num: 11840, Cur Loss: 0.53856194, Cur Avg Loss: 0.22722133, Log Avg loss: 0.25800486, Global Avg Loss: 0.89403734, Time: 0.0210 Steps: 64610, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000750, Sample Num: 12000, Cur Loss: 0.29822484, Cur Avg Loss: 0.22860330, Log Avg loss: 0.33086885, Global Avg Loss: 0.89395019, Time: 0.0211 Steps: 64620, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000760, Sample Num: 12160, Cur Loss: 0.11814258, Cur Avg Loss: 0.22986398, Log Avg loss: 0.32441514, Global Avg Loss: 0.89386207, Time: 0.0210 Steps: 64630, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000770, Sample Num: 12320, Cur Loss: 0.17948487, Cur Avg Loss: 0.22994787, Log Avg loss: 0.23632320, Global Avg Loss: 0.89376034, Time: 0.0248 Steps: 64640, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000780, Sample Num: 12480, Cur Loss: 0.19570097, Cur Avg Loss: 0.22986586, Log Avg loss: 0.22355134, Global Avg Loss: 0.89365668, Time: 0.0210 Steps: 64650, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000790, Sample Num: 12640, Cur Loss: 0.11780167, Cur Avg Loss: 0.22907025, Log Avg loss: 0.16701242, Global Avg Loss: 0.89354430, Time: 0.0209 Steps: 64660, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000800, Sample Num: 12800, Cur Loss: 0.19522947, Cur Avg Loss: 0.22945587, Log Avg loss: 0.25992016, Global Avg Loss: 0.89344632, Time: 0.0210 Steps: 64670, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000810, Sample Num: 12960, Cur Loss: 0.27343079, Cur Avg Loss: 0.22977697, Log Avg loss: 0.25546473, Global Avg Loss: 0.89334768, Time: 0.0209 Steps: 64680, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000820, Sample Num: 13120, Cur Loss: 0.18720479, Cur Avg Loss: 0.22918287, Log Avg loss: 0.18106094, Global Avg Loss: 0.89323758, Time: 0.0209 Steps: 64690, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000830, Sample Num: 13280, Cur Loss: 0.32205036, Cur Avg Loss: 0.22904464, Log Avg loss: 0.21770967, Global Avg Loss: 0.89313317, Time: 0.0209 Steps: 64700, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000840, Sample Num: 13440, Cur Loss: 0.09094169, Cur Avg Loss: 0.22932315, Log Avg loss: 0.25243936, Global Avg Loss: 0.89303416, Time: 0.0209 Steps: 64710, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000850, Sample Num: 13600, Cur Loss: 0.42832810, Cur Avg Loss: 0.22930339, Log Avg loss: 0.22764425, Global Avg Loss: 0.89293135, Time: 0.0210 Steps: 64720, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000860, Sample Num: 13760, Cur Loss: 0.38056764, Cur Avg Loss: 0.22971643, Log Avg loss: 0.26482451, Global Avg Loss: 0.89283431, Time: 0.0209 Steps: 64730, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000870, Sample Num: 13920, Cur Loss: 0.14125279, Cur Avg Loss: 0.23035215, Log Avg loss: 0.28502393, Global Avg Loss: 0.89274043, Time: 0.0210 Steps: 64740, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000880, Sample Num: 14080, Cur Loss: 0.17617840, Cur Avg Loss: 0.23009311, Log Avg loss: 0.20755712, Global Avg Loss: 0.89263461, Time: 0.0210 Steps: 64750, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000890, Sample Num: 14240, Cur Loss: 0.13307214, Cur Avg Loss: 0.23023779, Log Avg loss: 0.24296961, Global Avg Loss: 0.89253429, Time: 0.0209 Steps: 64760, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000900, Sample Num: 14400, Cur Loss: 0.32227662, Cur Avg Loss: 0.23049672, Log Avg loss: 0.25354076, Global Avg Loss: 0.89243563, Time: 0.0210 Steps: 64770, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000910, Sample Num: 14560, Cur Loss: 0.22402179, Cur Avg Loss: 0.23076996, Log Avg loss: 0.25536145, Global Avg Loss: 0.89233729, Time: 0.0211 Steps: 64780, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000920, Sample Num: 14720, Cur Loss: 0.30577102, Cur Avg Loss: 0.23087623, Log Avg loss: 0.24054741, Global Avg Loss: 0.89223669, Time: 0.0209 Steps: 64790, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000930, Sample Num: 14880, Cur Loss: 0.16156231, Cur Avg Loss: 0.23026890, Log Avg loss: 0.17439405, Global Avg Loss: 0.89212591, Time: 0.0210 Steps: 64800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000940, Sample Num: 15040, Cur Loss: 0.21735781, Cur Avg Loss: 0.22970883, Log Avg loss: 0.17762244, Global Avg Loss: 0.89201566, Time: 0.0209 Steps: 64810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000950, Sample Num: 15200, Cur Loss: 0.53246319, Cur Avg Loss: 0.23064339, Log Avg loss: 0.31849228, Global Avg Loss: 0.89192718, Time: 0.0209 Steps: 64820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000960, Sample Num: 15360, Cur Loss: 0.09155242, Cur Avg Loss: 0.23155696, Log Avg loss: 0.31834580, Global Avg Loss: 0.89183871, Time: 0.0210 Steps: 64830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000970, Sample Num: 15520, Cur Loss: 0.15223172, Cur Avg Loss: 0.23177699, Log Avg loss: 0.25290033, Global Avg Loss: 0.89174017, Time: 0.0210 Steps: 64840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000980, Sample Num: 15680, Cur Loss: 0.12122692, Cur Avg Loss: 0.23210005, Log Avg loss: 0.26343647, Global Avg Loss: 0.89164328, Time: 0.0209 Steps: 64850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000990, Sample Num: 15840, Cur Loss: 0.14892259, Cur Avg Loss: 0.23219840, Log Avg loss: 0.24183664, Global Avg Loss: 0.89154310, Time: 0.0210 Steps: 64860, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001000, Sample Num: 16000, Cur Loss: 0.23623537, Cur Avg Loss: 0.23219108, Log Avg loss: 0.23146664, Global Avg Loss: 0.89144134, Time: 0.0209 Steps: 64870, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001010, Sample Num: 16160, Cur Loss: 0.41833651, Cur Avg Loss: 0.23168049, Log Avg loss: 0.18062154, Global Avg Loss: 0.89133178, Time: 0.0211 Steps: 64880, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001020, Sample Num: 16320, Cur Loss: 0.31953099, Cur Avg Loss: 0.23153349, Log Avg loss: 0.21668641, Global Avg Loss: 0.89122782, Time: 0.0211 Steps: 64890, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001030, Sample Num: 16480, Cur Loss: 0.45446050, Cur Avg Loss: 0.23129054, Log Avg loss: 0.20650995, Global Avg Loss: 0.89112231, Time: 0.0211 Steps: 64900, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001040, Sample Num: 16640, Cur Loss: 0.40694165, Cur Avg Loss: 0.23199237, Log Avg loss: 0.30428097, Global Avg Loss: 0.89103190, Time: 0.0210 Steps: 64910, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001050, Sample Num: 16800, Cur Loss: 0.67815912, Cur Avg Loss: 0.23215164, Log Avg loss: 0.24871560, Global Avg Loss: 0.89093296, Time: 0.0210 Steps: 64920, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001060, Sample Num: 16960, Cur Loss: 0.12809020, Cur Avg Loss: 0.23219766, Log Avg loss: 0.23702947, Global Avg Loss: 0.89083225, Time: 0.0210 Steps: 64930, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001070, Sample Num: 17120, Cur Loss: 0.21314360, Cur Avg Loss: 0.23254876, Log Avg loss: 0.26976580, Global Avg Loss: 0.89073662, Time: 0.0210 Steps: 64940, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001080, Sample Num: 17280, Cur Loss: 0.18138173, Cur Avg Loss: 0.23299162, Log Avg loss: 0.28037700, Global Avg Loss: 0.89064264, Time: 0.0210 Steps: 64950, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001090, Sample Num: 17440, Cur Loss: 0.41195238, Cur Avg Loss: 0.23252807, Log Avg loss: 0.18246446, Global Avg Loss: 0.89053363, Time: 0.0210 Steps: 64960, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001100, Sample Num: 17600, Cur Loss: 0.78077477, Cur Avg Loss: 0.23292173, Log Avg loss: 0.27583084, Global Avg Loss: 0.89043901, Time: 0.0210 Steps: 64970, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001110, Sample Num: 17760, Cur Loss: 0.17770636, Cur Avg Loss: 0.23288349, Log Avg loss: 0.22867787, Global Avg Loss: 0.89033717, Time: 0.0210 Steps: 64980, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001120, Sample Num: 17920, Cur Loss: 0.11773957, Cur Avg Loss: 0.23282259, Log Avg loss: 0.22606222, Global Avg Loss: 0.89023496, Time: 0.0211 Steps: 64990, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001130, Sample Num: 18080, Cur Loss: 0.23459038, Cur Avg Loss: 0.23272137, Log Avg loss: 0.22138454, Global Avg Loss: 0.89013206, Time: 0.0210 Steps: 65000, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001140, Sample Num: 18240, Cur Loss: 0.12079386, Cur Avg Loss: 0.23285189, Log Avg loss: 0.24760045, Global Avg Loss: 0.89003322, Time: 0.0210 Steps: 65010, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001150, Sample Num: 18400, Cur Loss: 0.43096024, Cur Avg Loss: 0.23289028, Log Avg loss: 0.23726676, Global Avg Loss: 0.88993283, Time: 0.0210 Steps: 65020, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001160, Sample Num: 18560, Cur Loss: 0.33058339, Cur Avg Loss: 0.23298261, Log Avg loss: 0.24360050, Global Avg Loss: 0.88983344, Time: 0.0210 Steps: 65030, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001170, Sample Num: 18720, Cur Loss: 0.24753754, Cur Avg Loss: 0.23258561, Log Avg loss: 0.18653395, Global Avg Loss: 0.88972531, Time: 0.0210 Steps: 65040, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001180, Sample Num: 18880, Cur Loss: 0.37883806, Cur Avg Loss: 0.23285399, Log Avg loss: 0.26425454, Global Avg Loss: 0.88962915, Time: 0.0210 Steps: 65050, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001190, Sample Num: 19040, Cur Loss: 0.10379813, Cur Avg Loss: 0.23268696, Log Avg loss: 0.21297799, Global Avg Loss: 0.88952515, Time: 0.0211 Steps: 65060, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001200, Sample Num: 19200, Cur Loss: 0.29583746, Cur Avg Loss: 0.23282342, Log Avg loss: 0.24906126, Global Avg Loss: 0.88942672, Time: 0.0210 Steps: 65070, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001210, Sample Num: 19360, Cur Loss: 0.23738760, Cur Avg Loss: 0.23288871, Log Avg loss: 0.24072400, Global Avg Loss: 0.88932705, Time: 0.0210 Steps: 65080, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001220, Sample Num: 19520, Cur Loss: 0.06600504, Cur Avg Loss: 0.23296202, Log Avg loss: 0.24183213, Global Avg Loss: 0.88922757, Time: 0.0210 Steps: 65090, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001230, Sample Num: 19680, Cur Loss: 0.24003516, Cur Avg Loss: 0.23266448, Log Avg loss: 0.19636455, Global Avg Loss: 0.88912114, Time: 0.0210 Steps: 65100, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001240, Sample Num: 19840, Cur Loss: 0.56623292, Cur Avg Loss: 0.23288405, Log Avg loss: 0.25989152, Global Avg Loss: 0.88902450, Time: 0.0210 Steps: 65110, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001250, Sample Num: 20000, Cur Loss: 0.46833837, Cur Avg Loss: 0.23266322, Log Avg loss: 0.20527961, Global Avg Loss: 0.88891950, Time: 0.0210 Steps: 65120, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001260, Sample Num: 20160, Cur Loss: 0.30361027, Cur Avg Loss: 0.23368428, Log Avg loss: 0.36131712, Global Avg Loss: 0.88883849, Time: 0.0211 Steps: 65130, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001270, Sample Num: 20320, Cur Loss: 0.32999939, Cur Avg Loss: 0.23386376, Log Avg loss: 0.25647808, Global Avg Loss: 0.88874141, Time: 0.0210 Steps: 65140, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001280, Sample Num: 20480, Cur Loss: 0.49645773, Cur Avg Loss: 0.23436970, Log Avg loss: 0.29862399, Global Avg Loss: 0.88865084, Time: 0.0255 Steps: 65150, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001290, Sample Num: 20640, Cur Loss: 0.26932159, Cur Avg Loss: 0.23592551, Log Avg loss: 0.43506937, Global Avg Loss: 0.88858123, Time: 0.0209 Steps: 65160, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001300, Sample Num: 20800, Cur Loss: 0.63639063, Cur Avg Loss: 0.23651973, Log Avg loss: 0.31317506, Global Avg Loss: 0.88849293, Time: 0.0210 Steps: 65170, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001310, Sample Num: 20960, Cur Loss: 0.14260265, Cur Avg Loss: 0.23644132, Log Avg loss: 0.22624730, Global Avg Loss: 0.88839133, Time: 0.0210 Steps: 65180, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001320, Sample Num: 21120, Cur Loss: 0.51178622, Cur Avg Loss: 0.23674225, Log Avg loss: 0.27616384, Global Avg Loss: 0.88829742, Time: 0.0210 Steps: 65190, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001330, Sample Num: 21280, Cur Loss: 0.14993215, Cur Avg Loss: 0.23599475, Log Avg loss: 0.13732521, Global Avg Loss: 0.88818224, Time: 0.0209 Steps: 65200, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001340, Sample Num: 21440, Cur Loss: 0.73495770, Cur Avg Loss: 0.23565641, Log Avg loss: 0.19065766, Global Avg Loss: 0.88807527, Time: 0.0209 Steps: 65210, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001350, Sample Num: 21600, Cur Loss: 0.16589600, Cur Avg Loss: 0.23554407, Log Avg loss: 0.22049044, Global Avg Loss: 0.88797291, Time: 0.0209 Steps: 65220, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001360, Sample Num: 21760, Cur Loss: 0.19760841, Cur Avg Loss: 0.23497841, Log Avg loss: 0.15861414, Global Avg Loss: 0.88786110, Time: 0.0210 Steps: 65230, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001370, Sample Num: 21920, Cur Loss: 0.35590258, Cur Avg Loss: 0.23463505, Log Avg loss: 0.18793832, Global Avg Loss: 0.88775381, Time: 0.0210 Steps: 65240, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001380, Sample Num: 22080, Cur Loss: 0.54283202, Cur Avg Loss: 0.23478245, Log Avg loss: 0.25497559, Global Avg Loss: 0.88765684, Time: 0.0209 Steps: 65250, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001390, Sample Num: 22240, Cur Loss: 0.16026437, Cur Avg Loss: 0.23438225, Log Avg loss: 0.17915500, Global Avg Loss: 0.88754827, Time: 0.0209 Steps: 65260, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001400, Sample Num: 22400, Cur Loss: 0.42538813, Cur Avg Loss: 0.23424182, Log Avg loss: 0.21472194, Global Avg Loss: 0.88744519, Time: 0.0210 Steps: 65270, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001410, Sample Num: 22560, Cur Loss: 0.18043393, Cur Avg Loss: 0.23377205, Log Avg loss: 0.16800389, Global Avg Loss: 0.88733498, Time: 0.0210 Steps: 65280, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001420, Sample Num: 22720, Cur Loss: 0.30728340, Cur Avg Loss: 0.23368934, Log Avg loss: 0.22202679, Global Avg Loss: 0.88723308, Time: 0.0209 Steps: 65290, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001430, Sample Num: 22880, Cur Loss: 0.26044360, Cur Avg Loss: 0.23401748, Log Avg loss: 0.28061404, Global Avg Loss: 0.88714018, Time: 0.0209 Steps: 65300, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001440, Sample Num: 23040, Cur Loss: 0.39081505, Cur Avg Loss: 0.23442033, Log Avg loss: 0.29202736, Global Avg Loss: 0.88704906, Time: 0.0210 Steps: 65310, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001450, Sample Num: 23200, Cur Loss: 0.08009861, Cur Avg Loss: 0.23486210, Log Avg loss: 0.29847788, Global Avg Loss: 0.88695895, Time: 0.0210 Steps: 65320, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001460, Sample Num: 23360, Cur Loss: 0.40211213, Cur Avg Loss: 0.23501766, Log Avg loss: 0.25757287, Global Avg Loss: 0.88686261, Time: 0.0209 Steps: 65330, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001470, Sample Num: 23520, Cur Loss: 0.21144888, Cur Avg Loss: 0.23467991, Log Avg loss: 0.18536964, Global Avg Loss: 0.88675525, Time: 0.0209 Steps: 65340, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001480, Sample Num: 23680, Cur Loss: 0.18294330, Cur Avg Loss: 0.23459862, Log Avg loss: 0.22264829, Global Avg Loss: 0.88665363, Time: 0.0210 Steps: 65350, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001490, Sample Num: 23840, Cur Loss: 0.11423923, Cur Avg Loss: 0.23480356, Log Avg loss: 0.26513456, Global Avg Loss: 0.88655854, Time: 0.0209 Steps: 65360, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001500, Sample Num: 24000, Cur Loss: 0.09355478, Cur Avg Loss: 0.23441857, Log Avg loss: 0.17705592, Global Avg Loss: 0.88645000, Time: 0.0210 Steps: 65370, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001510, Sample Num: 24160, Cur Loss: 0.14945412, Cur Avg Loss: 0.23442978, Log Avg loss: 0.23611097, Global Avg Loss: 0.88635053, Time: 0.0210 Steps: 65380, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001520, Sample Num: 24320, Cur Loss: 0.44002011, Cur Avg Loss: 0.23444462, Log Avg loss: 0.23668524, Global Avg Loss: 0.88625118, Time: 0.0209 Steps: 65390, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001530, Sample Num: 24480, Cur Loss: 0.11762143, Cur Avg Loss: 0.23427025, Log Avg loss: 0.20776629, Global Avg Loss: 0.88614744, Time: 0.0209 Steps: 65400, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001540, Sample Num: 24640, Cur Loss: 0.28461248, Cur Avg Loss: 0.23445897, Log Avg loss: 0.26333343, Global Avg Loss: 0.88605222, Time: 0.0247 Steps: 65410, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001550, Sample Num: 24800, Cur Loss: 0.19664471, Cur Avg Loss: 0.23415335, Log Avg loss: 0.18708790, Global Avg Loss: 0.88594538, Time: 0.0209 Steps: 65420, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001560, Sample Num: 24960, Cur Loss: 0.33374882, Cur Avg Loss: 0.23457541, Log Avg loss: 0.29999437, Global Avg Loss: 0.88585582, Time: 0.0209 Steps: 65430, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001570, Sample Num: 25120, Cur Loss: 0.22961365, Cur Avg Loss: 0.23407057, Log Avg loss: 0.15531457, Global Avg Loss: 0.88574419, Time: 0.0210 Steps: 65440, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001580, Sample Num: 25280, Cur Loss: 0.13092552, Cur Avg Loss: 0.23415054, Log Avg loss: 0.24670561, Global Avg Loss: 0.88564655, Time: 0.0209 Steps: 65450, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001590, Sample Num: 25440, Cur Loss: 0.23400024, Cur Avg Loss: 0.23366769, Log Avg loss: 0.15737869, Global Avg Loss: 0.88553530, Time: 0.0209 Steps: 65460, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001600, Sample Num: 25600, Cur Loss: 0.32143545, Cur Avg Loss: 0.23367736, Log Avg loss: 0.23521430, Global Avg Loss: 0.88543596, Time: 0.0209 Steps: 65470, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001610, Sample Num: 25760, Cur Loss: 0.18209943, Cur Avg Loss: 0.23389806, Log Avg loss: 0.26921088, Global Avg Loss: 0.88534186, Time: 0.0209 Steps: 65480, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001620, Sample Num: 25920, Cur Loss: 0.28466207, Cur Avg Loss: 0.23412808, Log Avg loss: 0.27116128, Global Avg Loss: 0.88524807, Time: 0.0209 Steps: 65490, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001630, Sample Num: 26080, Cur Loss: 0.04447486, Cur Avg Loss: 0.23372076, Log Avg loss: 0.16773494, Global Avg Loss: 0.88513853, Time: 0.0210 Steps: 65500, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001640, Sample Num: 26240, Cur Loss: 0.24909261, Cur Avg Loss: 0.23361340, Log Avg loss: 0.21611260, Global Avg Loss: 0.88503640, Time: 0.0209 Steps: 65510, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001650, Sample Num: 26400, Cur Loss: 0.10029487, Cur Avg Loss: 0.23341435, Log Avg loss: 0.20077055, Global Avg Loss: 0.88493197, Time: 0.0210 Steps: 65520, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001660, Sample Num: 26560, Cur Loss: 0.47182810, Cur Avg Loss: 0.23410295, Log Avg loss: 0.34772125, Global Avg Loss: 0.88484999, Time: 0.0209 Steps: 65530, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001670, Sample Num: 26720, Cur Loss: 0.31979647, Cur Avg Loss: 0.23426147, Log Avg loss: 0.26057637, Global Avg Loss: 0.88475474, Time: 0.0209 Steps: 65540, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001680, Sample Num: 26880, Cur Loss: 0.12215549, Cur Avg Loss: 0.23410590, Log Avg loss: 0.20812540, Global Avg Loss: 0.88465151, Time: 0.0209 Steps: 65550, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001690, Sample Num: 27040, Cur Loss: 0.12745196, Cur Avg Loss: 0.23423963, Log Avg loss: 0.25670649, Global Avg Loss: 0.88455573, Time: 0.0209 Steps: 65560, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001700, Sample Num: 27200, Cur Loss: 0.25897780, Cur Avg Loss: 0.23534597, Log Avg loss: 0.42231678, Global Avg Loss: 0.88448524, Time: 0.0210 Steps: 65570, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001710, Sample Num: 27360, Cur Loss: 0.33306700, Cur Avg Loss: 0.23594834, Log Avg loss: 0.33835273, Global Avg Loss: 0.88440196, Time: 0.0210 Steps: 65580, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001720, Sample Num: 27520, Cur Loss: 0.12158537, Cur Avg Loss: 0.23568968, Log Avg loss: 0.19145868, Global Avg Loss: 0.88429631, Time: 0.0209 Steps: 65590, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001730, Sample Num: 27680, Cur Loss: 0.19595276, Cur Avg Loss: 0.23616425, Log Avg loss: 0.31778920, Global Avg Loss: 0.88420995, Time: 0.0209 Steps: 65600, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001740, Sample Num: 27840, Cur Loss: 0.12871717, Cur Avg Loss: 0.23648115, Log Avg loss: 0.29130513, Global Avg Loss: 0.88411959, Time: 0.0209 Steps: 65610, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001750, Sample Num: 28000, Cur Loss: 0.15129782, Cur Avg Loss: 0.23655711, Log Avg loss: 0.24977424, Global Avg Loss: 0.88402292, Time: 0.0209 Steps: 65620, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001760, Sample Num: 28160, Cur Loss: 0.22645962, Cur Avg Loss: 0.23696274, Log Avg loss: 0.30794748, Global Avg Loss: 0.88393514, Time: 0.0209 Steps: 65630, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001770, Sample Num: 28320, Cur Loss: 0.13977154, Cur Avg Loss: 0.23701458, Log Avg loss: 0.24613979, Global Avg Loss: 0.88383797, Time: 0.0209 Steps: 65640, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001780, Sample Num: 28480, Cur Loss: 0.29295257, Cur Avg Loss: 0.23690167, Log Avg loss: 0.21691635, Global Avg Loss: 0.88373639, Time: 0.0209 Steps: 65650, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001790, Sample Num: 28640, Cur Loss: 0.13635151, Cur Avg Loss: 0.23656583, Log Avg loss: 0.17678570, Global Avg Loss: 0.88362872, Time: 0.0210 Steps: 65660, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001800, Sample Num: 28800, Cur Loss: 0.11145405, Cur Avg Loss: 0.23648955, Log Avg loss: 0.22283626, Global Avg Loss: 0.88352809, Time: 0.0211 Steps: 65670, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001810, Sample Num: 28960, Cur Loss: 0.32863969, Cur Avg Loss: 0.23613299, Log Avg loss: 0.17195095, Global Avg Loss: 0.88341975, Time: 0.0209 Steps: 65680, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001820, Sample Num: 29120, Cur Loss: 0.14585012, Cur Avg Loss: 0.23603529, Log Avg loss: 0.21835226, Global Avg Loss: 0.88331851, Time: 0.0209 Steps: 65690, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001830, Sample Num: 29280, Cur Loss: 0.32106018, Cur Avg Loss: 0.23594887, Log Avg loss: 0.22022075, Global Avg Loss: 0.88321758, Time: 0.0209 Steps: 65700, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001840, Sample Num: 29440, Cur Loss: 0.27205443, Cur Avg Loss: 0.23610733, Log Avg loss: 0.26510526, Global Avg Loss: 0.88312352, Time: 0.0209 Steps: 65710, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001850, Sample Num: 29600, Cur Loss: 0.28648868, Cur Avg Loss: 0.23614141, Log Avg loss: 0.24241118, Global Avg Loss: 0.88302603, Time: 0.0210 Steps: 65720, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001860, Sample Num: 29760, Cur Loss: 0.09029541, Cur Avg Loss: 0.23624347, Log Avg loss: 0.25512546, Global Avg Loss: 0.88293050, Time: 0.0210 Steps: 65730, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001870, Sample Num: 29920, Cur Loss: 0.24217863, Cur Avg Loss: 0.23640803, Log Avg loss: 0.26701628, Global Avg Loss: 0.88283681, Time: 0.0210 Steps: 65740, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001880, Sample Num: 30080, Cur Loss: 0.23151951, Cur Avg Loss: 0.23613911, Log Avg loss: 0.18585089, Global Avg Loss: 0.88273080, Time: 0.0210 Steps: 65750, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001890, Sample Num: 30240, Cur Loss: 0.11368988, Cur Avg Loss: 0.23600004, Log Avg loss: 0.20985470, Global Avg Loss: 0.88262848, Time: 0.0210 Steps: 65760, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001900, Sample Num: 30400, Cur Loss: 0.12511927, Cur Avg Loss: 0.23627217, Log Avg loss: 0.28770413, Global Avg Loss: 0.88253802, Time: 0.0210 Steps: 65770, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001910, Sample Num: 30560, Cur Loss: 0.19991028, Cur Avg Loss: 0.23661617, Log Avg loss: 0.30197611, Global Avg Loss: 0.88244977, Time: 0.0210 Steps: 65780, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001920, Sample Num: 30720, Cur Loss: 0.12216020, Cur Avg Loss: 0.23668703, Log Avg loss: 0.25022284, Global Avg Loss: 0.88235367, Time: 0.0209 Steps: 65790, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001930, Sample Num: 30880, Cur Loss: 0.22790056, Cur Avg Loss: 0.23646392, Log Avg loss: 0.19362658, Global Avg Loss: 0.88224900, Time: 0.0209 Steps: 65800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001940, Sample Num: 31040, Cur Loss: 0.17965430, Cur Avg Loss: 0.23682307, Log Avg loss: 0.30613920, Global Avg Loss: 0.88216146, Time: 0.0210 Steps: 65810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001950, Sample Num: 31200, Cur Loss: 0.19115394, Cur Avg Loss: 0.23706776, Log Avg loss: 0.28453692, Global Avg Loss: 0.88207066, Time: 0.0209 Steps: 65820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001960, Sample Num: 31360, Cur Loss: 0.29834598, Cur Avg Loss: 0.23717591, Log Avg loss: 0.25826512, Global Avg Loss: 0.88197590, Time: 0.0210 Steps: 65830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001970, Sample Num: 31520, Cur Loss: 0.33817369, Cur Avg Loss: 0.23696833, Log Avg loss: 0.19628262, Global Avg Loss: 0.88187176, Time: 0.0210 Steps: 65840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001980, Sample Num: 31680, Cur Loss: 0.10927988, Cur Avg Loss: 0.23715346, Log Avg loss: 0.27362365, Global Avg Loss: 0.88177939, Time: 0.0210 Steps: 65850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001990, Sample Num: 31840, Cur Loss: 0.20674369, Cur Avg Loss: 0.23771023, Log Avg loss: 0.34795140, Global Avg Loss: 0.88169833, Time: 0.0210 Steps: 65860, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002000, Sample Num: 32000, Cur Loss: 0.41514656, Cur Avg Loss: 0.23763642, Log Avg loss: 0.22294774, Global Avg Loss: 0.88159832, Time: 0.0210 Steps: 65870, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002010, Sample Num: 32160, Cur Loss: 0.18756372, Cur Avg Loss: 0.23757367, Log Avg loss: 0.22502406, Global Avg Loss: 0.88149866, Time: 0.0210 Steps: 65880, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002020, Sample Num: 32320, Cur Loss: 0.42712793, Cur Avg Loss: 0.23756011, Log Avg loss: 0.23483540, Global Avg Loss: 0.88140052, Time: 0.0209 Steps: 65890, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002030, Sample Num: 32480, Cur Loss: 0.81125462, Cur Avg Loss: 0.23830095, Log Avg loss: 0.38794927, Global Avg Loss: 0.88132564, Time: 0.0210 Steps: 65900, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002040, Sample Num: 32640, Cur Loss: 0.08808064, Cur Avg Loss: 0.23848335, Log Avg loss: 0.27551003, Global Avg Loss: 0.88123372, Time: 0.0210 Steps: 65910, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002050, Sample Num: 32800, Cur Loss: 0.11858355, Cur Avg Loss: 0.23804266, Log Avg loss: 0.14814260, Global Avg Loss: 0.88112252, Time: 0.0247 Steps: 65920, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002060, Sample Num: 32960, Cur Loss: 0.17985453, Cur Avg Loss: 0.23777987, Log Avg loss: 0.18390875, Global Avg Loss: 0.88101677, Time: 0.0209 Steps: 65930, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002070, Sample Num: 33120, Cur Loss: 0.21569650, Cur Avg Loss: 0.23789645, Log Avg loss: 0.26191114, Global Avg Loss: 0.88092288, Time: 0.0209 Steps: 65940, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002080, Sample Num: 33280, Cur Loss: 0.10705016, Cur Avg Loss: 0.23755424, Log Avg loss: 0.16671716, Global Avg Loss: 0.88081458, Time: 0.0208 Steps: 65950, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002090, Sample Num: 33440, Cur Loss: 0.19374292, Cur Avg Loss: 0.23756728, Log Avg loss: 0.24027913, Global Avg Loss: 0.88071747, Time: 0.0208 Steps: 65960, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002100, Sample Num: 33600, Cur Loss: 0.19822632, Cur Avg Loss: 0.23785903, Log Avg loss: 0.29883424, Global Avg Loss: 0.88062927, Time: 0.0209 Steps: 65970, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002110, Sample Num: 33760, Cur Loss: 0.18852112, Cur Avg Loss: 0.23768651, Log Avg loss: 0.20145818, Global Avg Loss: 0.88052633, Time: 0.0209 Steps: 65980, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002120, Sample Num: 33920, Cur Loss: 0.21154860, Cur Avg Loss: 0.23744348, Log Avg loss: 0.18616486, Global Avg Loss: 0.88042111, Time: 0.0209 Steps: 65990, Updated lr: 0.000038 ***** Running evaluation checkpoint-65999 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-65999 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.845527, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.296544, "eval_total_loss": 208.47054, "eval_mae": 0.424924, "eval_mse": 0.296576, "eval_r2": 0.811477, "eval_sp_statistic": 0.893652, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.917864, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.350031, "test_total_loss": 175.715536, "test_mae": 0.358728, "test_mse": 0.350159, "test_r2": 0.774004, "test_sp_statistic": 0.868863, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.91534, "test_ps_pvalue": 0.0, "lr": 3.8360360360360366e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.880325250257624, "train_cur_epoch_loss": 504.9774105641991, "train_cur_epoch_avg_loss": 0.23718995329459797, "train_cur_epoch_time": 44.8455274105072, "train_cur_epoch_avg_time": 0.02106412748262433, "epoch": 31, "step": 65999} ################################################## Training, Epoch: 0032, Batch: 000001, Sample Num: 16, Cur Loss: 0.80983657, Cur Avg Loss: 0.80983657, Log Avg loss: 0.24070622, Global Avg Loss: 0.88032418, Time: 0.0247 Steps: 66000, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000011, Sample Num: 176, Cur Loss: 0.17935920, Cur Avg Loss: 0.24551322, Log Avg loss: 0.18908088, Global Avg Loss: 0.88021946, Time: 0.0210 Steps: 66010, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000021, Sample Num: 336, Cur Loss: 0.15447766, Cur Avg Loss: 0.19974411, Log Avg loss: 0.14939809, Global Avg Loss: 0.88010877, Time: 0.0210 Steps: 66020, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000031, Sample Num: 496, Cur Loss: 0.10028715, Cur Avg Loss: 0.20057742, Log Avg loss: 0.20232738, Global Avg Loss: 0.88000612, Time: 0.0209 Steps: 66030, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000041, Sample Num: 656, Cur Loss: 0.05254362, Cur Avg Loss: 0.19626643, Log Avg loss: 0.18290235, Global Avg Loss: 0.87990056, Time: 0.0209 Steps: 66040, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000051, Sample Num: 816, Cur Loss: 0.08075035, Cur Avg Loss: 0.17812322, Log Avg loss: 0.10373608, Global Avg Loss: 0.87978305, Time: 0.0209 Steps: 66050, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000061, Sample Num: 976, Cur Loss: 0.32593405, Cur Avg Loss: 0.19376437, Log Avg loss: 0.27353422, Global Avg Loss: 0.87969128, Time: 0.0210 Steps: 66060, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000071, Sample Num: 1136, Cur Loss: 0.18476215, Cur Avg Loss: 0.19704941, Log Avg loss: 0.21708816, Global Avg Loss: 0.87959099, Time: 0.0209 Steps: 66070, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000081, Sample Num: 1296, Cur Loss: 0.39825505, Cur Avg Loss: 0.20322827, Log Avg loss: 0.24709818, Global Avg Loss: 0.87949527, Time: 0.0210 Steps: 66080, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000091, Sample Num: 1456, Cur Loss: 0.14681304, Cur Avg Loss: 0.20638816, Log Avg loss: 0.23198324, Global Avg Loss: 0.87939730, Time: 0.0209 Steps: 66090, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000101, Sample Num: 1616, Cur Loss: 0.22412756, Cur Avg Loss: 0.20636489, Log Avg loss: 0.20615317, Global Avg Loss: 0.87929545, Time: 0.0210 Steps: 66100, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000111, Sample Num: 1776, Cur Loss: 0.46196246, Cur Avg Loss: 0.21346555, Log Avg loss: 0.28518216, Global Avg Loss: 0.87920558, Time: 0.0209 Steps: 66110, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000121, Sample Num: 1936, Cur Loss: 0.23108169, Cur Avg Loss: 0.22212659, Log Avg loss: 0.31826415, Global Avg Loss: 0.87912074, Time: 0.0209 Steps: 66120, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000131, Sample Num: 2096, Cur Loss: 0.09306726, Cur Avg Loss: 0.22099478, Log Avg loss: 0.20729985, Global Avg Loss: 0.87901915, Time: 0.0210 Steps: 66130, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000141, Sample Num: 2256, Cur Loss: 0.18965110, Cur Avg Loss: 0.21732674, Log Avg loss: 0.16927543, Global Avg Loss: 0.87891184, Time: 0.0209 Steps: 66140, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000151, Sample Num: 2416, Cur Loss: 0.17021233, Cur Avg Loss: 0.21563121, Log Avg loss: 0.19172435, Global Avg Loss: 0.87880796, Time: 0.0209 Steps: 66150, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000161, Sample Num: 2576, Cur Loss: 0.09095669, Cur Avg Loss: 0.21335096, Log Avg loss: 0.17891906, Global Avg Loss: 0.87870217, Time: 0.0210 Steps: 66160, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000171, Sample Num: 2736, Cur Loss: 0.15700306, Cur Avg Loss: 0.21575414, Log Avg loss: 0.25444546, Global Avg Loss: 0.87860783, Time: 0.0209 Steps: 66170, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000181, Sample Num: 2896, Cur Loss: 0.39047259, Cur Avg Loss: 0.21538861, Log Avg loss: 0.20913802, Global Avg Loss: 0.87850667, Time: 0.0209 Steps: 66180, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000191, Sample Num: 3056, Cur Loss: 0.09264082, Cur Avg Loss: 0.22368410, Log Avg loss: 0.37383250, Global Avg Loss: 0.87843043, Time: 0.0209 Steps: 66190, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000201, Sample Num: 3216, Cur Loss: 0.17878734, Cur Avg Loss: 0.22434278, Log Avg loss: 0.23692346, Global Avg Loss: 0.87833352, Time: 0.0210 Steps: 66200, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000211, Sample Num: 3376, Cur Loss: 0.21573114, Cur Avg Loss: 0.22370338, Log Avg loss: 0.21085138, Global Avg Loss: 0.87823271, Time: 0.0209 Steps: 66210, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000221, Sample Num: 3536, Cur Loss: 0.19703619, Cur Avg Loss: 0.22277680, Log Avg loss: 0.20322611, Global Avg Loss: 0.87813077, Time: 0.0209 Steps: 66220, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000231, Sample Num: 3696, Cur Loss: 0.55362910, Cur Avg Loss: 0.22585637, Log Avg loss: 0.29391475, Global Avg Loss: 0.87804256, Time: 0.0209 Steps: 66230, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000241, Sample Num: 3856, Cur Loss: 0.05918781, Cur Avg Loss: 0.22275756, Log Avg loss: 0.15117510, Global Avg Loss: 0.87793283, Time: 0.0209 Steps: 66240, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000251, Sample Num: 4016, Cur Loss: 0.14131254, Cur Avg Loss: 0.22280055, Log Avg loss: 0.22383673, Global Avg Loss: 0.87783410, Time: 0.0210 Steps: 66250, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000261, Sample Num: 4176, Cur Loss: 0.26631257, Cur Avg Loss: 0.22202871, Log Avg loss: 0.20265544, Global Avg Loss: 0.87773220, Time: 0.0210 Steps: 66260, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000271, Sample Num: 4336, Cur Loss: 0.15084517, Cur Avg Loss: 0.22202465, Log Avg loss: 0.22191866, Global Avg Loss: 0.87763324, Time: 0.0209 Steps: 66270, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000281, Sample Num: 4496, Cur Loss: 0.30484349, Cur Avg Loss: 0.22156282, Log Avg loss: 0.20904735, Global Avg Loss: 0.87753237, Time: 0.0211 Steps: 66280, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000291, Sample Num: 4656, Cur Loss: 0.12793387, Cur Avg Loss: 0.21885821, Log Avg loss: 0.14285860, Global Avg Loss: 0.87742154, Time: 0.0209 Steps: 66290, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000301, Sample Num: 4816, Cur Loss: 0.12952501, Cur Avg Loss: 0.22206406, Log Avg loss: 0.31535425, Global Avg Loss: 0.87733676, Time: 0.0209 Steps: 66300, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000311, Sample Num: 4976, Cur Loss: 0.19573343, Cur Avg Loss: 0.22142736, Log Avg loss: 0.20226258, Global Avg Loss: 0.87723496, Time: 0.0210 Steps: 66310, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000321, Sample Num: 5136, Cur Loss: 0.09441203, Cur Avg Loss: 0.22152710, Log Avg loss: 0.22462898, Global Avg Loss: 0.87713656, Time: 0.0209 Steps: 66320, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000331, Sample Num: 5296, Cur Loss: 0.14698532, Cur Avg Loss: 0.21912791, Log Avg loss: 0.14211418, Global Avg Loss: 0.87702574, Time: 0.0210 Steps: 66330, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000341, Sample Num: 5456, Cur Loss: 0.04828316, Cur Avg Loss: 0.21939618, Log Avg loss: 0.22827589, Global Avg Loss: 0.87692795, Time: 0.0209 Steps: 66340, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000351, Sample Num: 5616, Cur Loss: 0.05766790, Cur Avg Loss: 0.22032772, Log Avg loss: 0.25209317, Global Avg Loss: 0.87683378, Time: 0.0210 Steps: 66350, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000361, Sample Num: 5776, Cur Loss: 0.25551254, Cur Avg Loss: 0.22086044, Log Avg loss: 0.23955890, Global Avg Loss: 0.87673775, Time: 0.0208 Steps: 66360, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000371, Sample Num: 5936, Cur Loss: 0.16775148, Cur Avg Loss: 0.22060471, Log Avg loss: 0.21137270, Global Avg Loss: 0.87663749, Time: 0.0209 Steps: 66370, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000381, Sample Num: 6096, Cur Loss: 0.18076210, Cur Avg Loss: 0.22016139, Log Avg loss: 0.20371425, Global Avg Loss: 0.87653612, Time: 0.0209 Steps: 66380, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000391, Sample Num: 6256, Cur Loss: 0.36341947, Cur Avg Loss: 0.22208171, Log Avg loss: 0.29524616, Global Avg Loss: 0.87644856, Time: 0.0210 Steps: 66390, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000401, Sample Num: 6416, Cur Loss: 0.17424977, Cur Avg Loss: 0.22216647, Log Avg loss: 0.22548032, Global Avg Loss: 0.87635053, Time: 0.0209 Steps: 66400, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000411, Sample Num: 6576, Cur Loss: 0.28574729, Cur Avg Loss: 0.22202419, Log Avg loss: 0.21631904, Global Avg Loss: 0.87625114, Time: 0.0209 Steps: 66410, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000421, Sample Num: 6736, Cur Loss: 0.16929807, Cur Avg Loss: 0.22217826, Log Avg loss: 0.22851029, Global Avg Loss: 0.87615362, Time: 0.0210 Steps: 66420, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000431, Sample Num: 6896, Cur Loss: 0.05949081, Cur Avg Loss: 0.22294540, Log Avg loss: 0.25524229, Global Avg Loss: 0.87606015, Time: 0.0209 Steps: 66430, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000441, Sample Num: 7056, Cur Loss: 0.09938097, Cur Avg Loss: 0.22288313, Log Avg loss: 0.22019907, Global Avg Loss: 0.87596143, Time: 0.0210 Steps: 66440, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000451, Sample Num: 7216, Cur Loss: 0.23208061, Cur Avg Loss: 0.22190188, Log Avg loss: 0.17862888, Global Avg Loss: 0.87585649, Time: 0.0209 Steps: 66450, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000461, Sample Num: 7376, Cur Loss: 0.53626376, Cur Avg Loss: 0.22261937, Log Avg loss: 0.25497819, Global Avg Loss: 0.87576307, Time: 0.0209 Steps: 66460, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000471, Sample Num: 7536, Cur Loss: 0.17824835, Cur Avg Loss: 0.22397322, Log Avg loss: 0.28638537, Global Avg Loss: 0.87567440, Time: 0.0211 Steps: 66470, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000481, Sample Num: 7696, Cur Loss: 0.33380044, Cur Avg Loss: 0.22356839, Log Avg loss: 0.20450091, Global Avg Loss: 0.87557344, Time: 0.0208 Steps: 66480, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000491, Sample Num: 7856, Cur Loss: 0.09585330, Cur Avg Loss: 0.22338420, Log Avg loss: 0.21452476, Global Avg Loss: 0.87547402, Time: 0.0211 Steps: 66490, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000501, Sample Num: 8016, Cur Loss: 0.09725848, Cur Avg Loss: 0.22303409, Log Avg loss: 0.20584370, Global Avg Loss: 0.87537333, Time: 0.0209 Steps: 66500, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000511, Sample Num: 8176, Cur Loss: 0.27951312, Cur Avg Loss: 0.22327208, Log Avg loss: 0.23519551, Global Avg Loss: 0.87527707, Time: 0.0209 Steps: 66510, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000521, Sample Num: 8336, Cur Loss: 0.10590921, Cur Avg Loss: 0.22360482, Log Avg loss: 0.24060773, Global Avg Loss: 0.87518166, Time: 0.0208 Steps: 66520, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000531, Sample Num: 8496, Cur Loss: 0.21104455, Cur Avg Loss: 0.22421028, Log Avg loss: 0.25575479, Global Avg Loss: 0.87508856, Time: 0.0209 Steps: 66530, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000541, Sample Num: 8656, Cur Loss: 0.16701268, Cur Avg Loss: 0.22419326, Log Avg loss: 0.22328960, Global Avg Loss: 0.87499060, Time: 0.0208 Steps: 66540, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000551, Sample Num: 8816, Cur Loss: 0.28346413, Cur Avg Loss: 0.22588002, Log Avg loss: 0.31713354, Global Avg Loss: 0.87490678, Time: 0.0209 Steps: 66550, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000561, Sample Num: 8976, Cur Loss: 0.48082530, Cur Avg Loss: 0.22655319, Log Avg loss: 0.26364489, Global Avg Loss: 0.87481494, Time: 0.0208 Steps: 66560, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000571, Sample Num: 9136, Cur Loss: 0.04268814, Cur Avg Loss: 0.22710082, Log Avg loss: 0.25782295, Global Avg Loss: 0.87472226, Time: 0.0208 Steps: 66570, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000581, Sample Num: 9296, Cur Loss: 0.24139410, Cur Avg Loss: 0.22669354, Log Avg loss: 0.20343798, Global Avg Loss: 0.87462143, Time: 0.0209 Steps: 66580, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000591, Sample Num: 9456, Cur Loss: 0.06881833, Cur Avg Loss: 0.22620050, Log Avg loss: 0.19755455, Global Avg Loss: 0.87451976, Time: 0.0208 Steps: 66590, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000601, Sample Num: 9616, Cur Loss: 0.16577852, Cur Avg Loss: 0.22639496, Log Avg loss: 0.23788755, Global Avg Loss: 0.87442417, Time: 0.0209 Steps: 66600, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000611, Sample Num: 9776, Cur Loss: 0.38140374, Cur Avg Loss: 0.22696203, Log Avg loss: 0.26104327, Global Avg Loss: 0.87433208, Time: 0.0208 Steps: 66610, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000621, Sample Num: 9936, Cur Loss: 0.04359141, Cur Avg Loss: 0.22685067, Log Avg loss: 0.22004632, Global Avg Loss: 0.87423387, Time: 0.0209 Steps: 66620, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000631, Sample Num: 10096, Cur Loss: 0.02213854, Cur Avg Loss: 0.22505638, Log Avg loss: 0.11363114, Global Avg Loss: 0.87411972, Time: 0.0208 Steps: 66630, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000641, Sample Num: 10256, Cur Loss: 0.06939767, Cur Avg Loss: 0.22558590, Log Avg loss: 0.25899864, Global Avg Loss: 0.87402741, Time: 0.0208 Steps: 66640, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000651, Sample Num: 10416, Cur Loss: 0.27930048, Cur Avg Loss: 0.22484021, Log Avg loss: 0.17704163, Global Avg Loss: 0.87392284, Time: 0.0209 Steps: 66650, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000661, Sample Num: 10576, Cur Loss: 0.12175211, Cur Avg Loss: 0.22324515, Log Avg loss: 0.11940666, Global Avg Loss: 0.87380965, Time: 0.0208 Steps: 66660, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000671, Sample Num: 10736, Cur Loss: 0.09615791, Cur Avg Loss: 0.22283269, Log Avg loss: 0.19556923, Global Avg Loss: 0.87370792, Time: 0.0209 Steps: 66670, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000681, Sample Num: 10896, Cur Loss: 0.11070062, Cur Avg Loss: 0.22205780, Log Avg loss: 0.17006211, Global Avg Loss: 0.87360239, Time: 0.0211 Steps: 66680, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000691, Sample Num: 11056, Cur Loss: 0.18187317, Cur Avg Loss: 0.22258681, Log Avg loss: 0.25861251, Global Avg Loss: 0.87351018, Time: 0.0208 Steps: 66690, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000701, Sample Num: 11216, Cur Loss: 0.54583788, Cur Avg Loss: 0.22340714, Log Avg loss: 0.28009234, Global Avg Loss: 0.87342121, Time: 0.0208 Steps: 66700, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000711, Sample Num: 11376, Cur Loss: 0.04499751, Cur Avg Loss: 0.22243497, Log Avg loss: 0.15428551, Global Avg Loss: 0.87331341, Time: 0.0208 Steps: 66710, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000721, Sample Num: 11536, Cur Loss: 0.10229595, Cur Avg Loss: 0.22176709, Log Avg loss: 0.17428057, Global Avg Loss: 0.87320864, Time: 0.0208 Steps: 66720, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000731, Sample Num: 11696, Cur Loss: 0.26145789, Cur Avg Loss: 0.22117397, Log Avg loss: 0.17841045, Global Avg Loss: 0.87310452, Time: 0.0208 Steps: 66730, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000741, Sample Num: 11856, Cur Loss: 0.26949036, Cur Avg Loss: 0.22256941, Log Avg loss: 0.32457560, Global Avg Loss: 0.87302233, Time: 0.0208 Steps: 66740, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000751, Sample Num: 12016, Cur Loss: 0.52218956, Cur Avg Loss: 0.22291629, Log Avg loss: 0.24862065, Global Avg Loss: 0.87292878, Time: 0.0208 Steps: 66750, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000761, Sample Num: 12176, Cur Loss: 0.19881804, Cur Avg Loss: 0.22304663, Log Avg loss: 0.23283498, Global Avg Loss: 0.87283290, Time: 0.0208 Steps: 66760, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000771, Sample Num: 12336, Cur Loss: 0.06145700, Cur Avg Loss: 0.22306209, Log Avg loss: 0.22423878, Global Avg Loss: 0.87273577, Time: 0.0245 Steps: 66770, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000781, Sample Num: 12496, Cur Loss: 0.10739194, Cur Avg Loss: 0.22393282, Log Avg loss: 0.29106573, Global Avg Loss: 0.87264866, Time: 0.0208 Steps: 66780, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000791, Sample Num: 12656, Cur Loss: 0.23700389, Cur Avg Loss: 0.22329353, Log Avg loss: 0.17336523, Global Avg Loss: 0.87254396, Time: 0.0208 Steps: 66790, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000801, Sample Num: 12816, Cur Loss: 0.31462488, Cur Avg Loss: 0.22335542, Log Avg loss: 0.22825062, Global Avg Loss: 0.87244751, Time: 0.0209 Steps: 66800, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000811, Sample Num: 12976, Cur Loss: 0.13092090, Cur Avg Loss: 0.22325046, Log Avg loss: 0.21484307, Global Avg Loss: 0.87234908, Time: 0.0208 Steps: 66810, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000821, Sample Num: 13136, Cur Loss: 0.18502030, Cur Avg Loss: 0.22305752, Log Avg loss: 0.20741082, Global Avg Loss: 0.87224957, Time: 0.0208 Steps: 66820, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000831, Sample Num: 13296, Cur Loss: 0.27381176, Cur Avg Loss: 0.22361555, Log Avg loss: 0.26942908, Global Avg Loss: 0.87215937, Time: 0.0209 Steps: 66830, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000841, Sample Num: 13456, Cur Loss: 0.18338251, Cur Avg Loss: 0.22428433, Log Avg loss: 0.27986012, Global Avg Loss: 0.87207076, Time: 0.0208 Steps: 66840, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000851, Sample Num: 13616, Cur Loss: 0.13095842, Cur Avg Loss: 0.22397271, Log Avg loss: 0.19776580, Global Avg Loss: 0.87196989, Time: 0.0208 Steps: 66850, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000861, Sample Num: 13776, Cur Loss: 0.11331113, Cur Avg Loss: 0.22413679, Log Avg loss: 0.23809977, Global Avg Loss: 0.87187508, Time: 0.0209 Steps: 66860, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000871, Sample Num: 13936, Cur Loss: 0.13702065, Cur Avg Loss: 0.22347736, Log Avg loss: 0.16670015, Global Avg Loss: 0.87176963, Time: 0.0208 Steps: 66870, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000881, Sample Num: 14096, Cur Loss: 0.11391937, Cur Avg Loss: 0.22268429, Log Avg loss: 0.15360784, Global Avg Loss: 0.87166225, Time: 0.0208 Steps: 66880, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000891, Sample Num: 14256, Cur Loss: 0.11717743, Cur Avg Loss: 0.22243243, Log Avg loss: 0.20024356, Global Avg Loss: 0.87156187, Time: 0.0209 Steps: 66890, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000901, Sample Num: 14416, Cur Loss: 0.28120697, Cur Avg Loss: 0.22267709, Log Avg loss: 0.24447674, Global Avg Loss: 0.87146814, Time: 0.0209 Steps: 66900, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000911, Sample Num: 14576, Cur Loss: 0.37714642, Cur Avg Loss: 0.22359681, Log Avg loss: 0.30646375, Global Avg Loss: 0.87138369, Time: 0.0210 Steps: 66910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000921, Sample Num: 14736, Cur Loss: 0.13557464, Cur Avg Loss: 0.22369605, Log Avg loss: 0.23273687, Global Avg Loss: 0.87128826, Time: 0.0209 Steps: 66920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000931, Sample Num: 14896, Cur Loss: 0.21920350, Cur Avg Loss: 0.22320773, Log Avg loss: 0.17823364, Global Avg Loss: 0.87118471, Time: 0.0208 Steps: 66930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000941, Sample Num: 15056, Cur Loss: 0.30694753, Cur Avg Loss: 0.22358724, Log Avg loss: 0.25891882, Global Avg Loss: 0.87109324, Time: 0.0209 Steps: 66940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000951, Sample Num: 15216, Cur Loss: 0.10788972, Cur Avg Loss: 0.22349933, Log Avg loss: 0.21522757, Global Avg Loss: 0.87099528, Time: 0.0208 Steps: 66950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000961, Sample Num: 15376, Cur Loss: 0.39285964, Cur Avg Loss: 0.22426839, Log Avg loss: 0.29740568, Global Avg Loss: 0.87090962, Time: 0.0209 Steps: 66960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000971, Sample Num: 15536, Cur Loss: 0.51512909, Cur Avg Loss: 0.22502458, Log Avg loss: 0.29769466, Global Avg Loss: 0.87082403, Time: 0.0208 Steps: 66970, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000981, Sample Num: 15696, Cur Loss: 0.40824139, Cur Avg Loss: 0.22503844, Log Avg loss: 0.22638385, Global Avg Loss: 0.87072781, Time: 0.0208 Steps: 66980, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000991, Sample Num: 15856, Cur Loss: 0.13326146, Cur Avg Loss: 0.22472303, Log Avg loss: 0.19378203, Global Avg Loss: 0.87062676, Time: 0.0209 Steps: 66990, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001001, Sample Num: 16016, Cur Loss: 0.36098599, Cur Avg Loss: 0.22619812, Log Avg loss: 0.37237895, Global Avg Loss: 0.87055240, Time: 0.0209 Steps: 67000, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001011, Sample Num: 16176, Cur Loss: 0.25853807, Cur Avg Loss: 0.22791327, Log Avg loss: 0.39960032, Global Avg Loss: 0.87048211, Time: 0.0208 Steps: 67010, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001021, Sample Num: 16336, Cur Loss: 0.57335949, Cur Avg Loss: 0.22920784, Log Avg loss: 0.36008847, Global Avg Loss: 0.87040596, Time: 0.0209 Steps: 67020, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001031, Sample Num: 16496, Cur Loss: 0.19564338, Cur Avg Loss: 0.22863551, Log Avg loss: 0.17020011, Global Avg Loss: 0.87030150, Time: 0.0213 Steps: 67030, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001041, Sample Num: 16656, Cur Loss: 0.10976765, Cur Avg Loss: 0.22891584, Log Avg loss: 0.25781819, Global Avg Loss: 0.87021014, Time: 0.0209 Steps: 67040, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001051, Sample Num: 16816, Cur Loss: 0.04834422, Cur Avg Loss: 0.22855170, Log Avg loss: 0.19064449, Global Avg Loss: 0.87010878, Time: 0.0208 Steps: 67050, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001061, Sample Num: 16976, Cur Loss: 0.44564104, Cur Avg Loss: 0.22884667, Log Avg loss: 0.25984823, Global Avg Loss: 0.87001778, Time: 0.0208 Steps: 67060, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001071, Sample Num: 17136, Cur Loss: 0.21202116, Cur Avg Loss: 0.22899492, Log Avg loss: 0.24472476, Global Avg Loss: 0.86992455, Time: 0.0209 Steps: 67070, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001081, Sample Num: 17296, Cur Loss: 0.13920233, Cur Avg Loss: 0.22882421, Log Avg loss: 0.21054132, Global Avg Loss: 0.86982625, Time: 0.0210 Steps: 67080, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001091, Sample Num: 17456, Cur Loss: 0.18285617, Cur Avg Loss: 0.22910820, Log Avg loss: 0.25980722, Global Avg Loss: 0.86973533, Time: 0.0209 Steps: 67090, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001101, Sample Num: 17616, Cur Loss: 0.31287500, Cur Avg Loss: 0.22925168, Log Avg loss: 0.24490466, Global Avg Loss: 0.86964221, Time: 0.0209 Steps: 67100, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001111, Sample Num: 17776, Cur Loss: 0.12761487, Cur Avg Loss: 0.22959079, Log Avg loss: 0.26692780, Global Avg Loss: 0.86955240, Time: 0.0210 Steps: 67110, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001121, Sample Num: 17936, Cur Loss: 0.11510845, Cur Avg Loss: 0.22958248, Log Avg loss: 0.22865890, Global Avg Loss: 0.86945692, Time: 0.0210 Steps: 67120, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001131, Sample Num: 18096, Cur Loss: 0.20698667, Cur Avg Loss: 0.22908905, Log Avg loss: 0.17377569, Global Avg Loss: 0.86935328, Time: 0.0209 Steps: 67130, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001141, Sample Num: 18256, Cur Loss: 0.59334767, Cur Avg Loss: 0.22972812, Log Avg loss: 0.30200652, Global Avg Loss: 0.86926878, Time: 0.0209 Steps: 67140, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001151, Sample Num: 18416, Cur Loss: 0.33717519, Cur Avg Loss: 0.23010647, Log Avg loss: 0.27327687, Global Avg Loss: 0.86918003, Time: 0.0209 Steps: 67150, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001161, Sample Num: 18576, Cur Loss: 0.21067527, Cur Avg Loss: 0.23260267, Log Avg loss: 0.51991514, Global Avg Loss: 0.86912802, Time: 0.0209 Steps: 67160, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001171, Sample Num: 18736, Cur Loss: 0.29050952, Cur Avg Loss: 0.23324223, Log Avg loss: 0.30749491, Global Avg Loss: 0.86904441, Time: 0.0209 Steps: 67170, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001181, Sample Num: 18896, Cur Loss: 0.07109046, Cur Avg Loss: 0.23255943, Log Avg loss: 0.15260347, Global Avg Loss: 0.86893776, Time: 0.0209 Steps: 67180, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001191, Sample Num: 19056, Cur Loss: 0.35787940, Cur Avg Loss: 0.23194890, Log Avg loss: 0.15984540, Global Avg Loss: 0.86883223, Time: 0.0209 Steps: 67190, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001201, Sample Num: 19216, Cur Loss: 0.20142502, Cur Avg Loss: 0.23260132, Log Avg loss: 0.31030437, Global Avg Loss: 0.86874911, Time: 0.0209 Steps: 67200, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001211, Sample Num: 19376, Cur Loss: 0.17893188, Cur Avg Loss: 0.23291423, Log Avg loss: 0.27049430, Global Avg Loss: 0.86866010, Time: 0.0209 Steps: 67210, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001221, Sample Num: 19536, Cur Loss: 0.29545251, Cur Avg Loss: 0.23283328, Log Avg loss: 0.22303009, Global Avg Loss: 0.86856405, Time: 0.0209 Steps: 67220, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001231, Sample Num: 19696, Cur Loss: 0.13442744, Cur Avg Loss: 0.23275140, Log Avg loss: 0.22275449, Global Avg Loss: 0.86846799, Time: 0.0209 Steps: 67230, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001241, Sample Num: 19856, Cur Loss: 0.32549900, Cur Avg Loss: 0.23282777, Log Avg loss: 0.24222903, Global Avg Loss: 0.86837486, Time: 0.0209 Steps: 67240, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001251, Sample Num: 20016, Cur Loss: 0.40233275, Cur Avg Loss: 0.23314809, Log Avg loss: 0.27289990, Global Avg Loss: 0.86828631, Time: 0.0209 Steps: 67250, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001261, Sample Num: 20176, Cur Loss: 0.28903204, Cur Avg Loss: 0.23359195, Log Avg loss: 0.28911809, Global Avg Loss: 0.86820020, Time: 0.0209 Steps: 67260, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001271, Sample Num: 20336, Cur Loss: 0.05807672, Cur Avg Loss: 0.23300844, Log Avg loss: 0.15942797, Global Avg Loss: 0.86809484, Time: 0.0210 Steps: 67270, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001281, Sample Num: 20496, Cur Loss: 0.07836516, Cur Avg Loss: 0.23255677, Log Avg loss: 0.17514992, Global Avg Loss: 0.86799185, Time: 0.0245 Steps: 67280, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001291, Sample Num: 20656, Cur Loss: 0.24241346, Cur Avg Loss: 0.23218127, Log Avg loss: 0.18407908, Global Avg Loss: 0.86789021, Time: 0.0208 Steps: 67290, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001301, Sample Num: 20816, Cur Loss: 0.21460640, Cur Avg Loss: 0.23150011, Log Avg loss: 0.14356267, Global Avg Loss: 0.86778258, Time: 0.0209 Steps: 67300, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001311, Sample Num: 20976, Cur Loss: 0.20888111, Cur Avg Loss: 0.23134386, Log Avg loss: 0.21101556, Global Avg Loss: 0.86768501, Time: 0.0209 Steps: 67310, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001321, Sample Num: 21136, Cur Loss: 0.11030198, Cur Avg Loss: 0.23153907, Log Avg loss: 0.25713190, Global Avg Loss: 0.86759432, Time: 0.0210 Steps: 67320, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001331, Sample Num: 21296, Cur Loss: 0.11866683, Cur Avg Loss: 0.23178898, Log Avg loss: 0.26480174, Global Avg Loss: 0.86750479, Time: 0.0209 Steps: 67330, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001341, Sample Num: 21456, Cur Loss: 0.42777696, Cur Avg Loss: 0.23166461, Log Avg loss: 0.21511048, Global Avg Loss: 0.86740791, Time: 0.0208 Steps: 67340, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001351, Sample Num: 21616, Cur Loss: 0.46806097, Cur Avg Loss: 0.23178002, Log Avg loss: 0.24725665, Global Avg Loss: 0.86731583, Time: 0.0208 Steps: 67350, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001361, Sample Num: 21776, Cur Loss: 0.45914555, Cur Avg Loss: 0.23204778, Log Avg loss: 0.26822305, Global Avg Loss: 0.86722689, Time: 0.0209 Steps: 67360, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001371, Sample Num: 21936, Cur Loss: 0.15206584, Cur Avg Loss: 0.23188804, Log Avg loss: 0.21014704, Global Avg Loss: 0.86712936, Time: 0.0210 Steps: 67370, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001381, Sample Num: 22096, Cur Loss: 0.17803046, Cur Avg Loss: 0.23203649, Log Avg loss: 0.25238918, Global Avg Loss: 0.86703812, Time: 0.0211 Steps: 67380, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001391, Sample Num: 22256, Cur Loss: 0.33548409, Cur Avg Loss: 0.23230202, Log Avg loss: 0.26897091, Global Avg Loss: 0.86694937, Time: 0.0209 Steps: 67390, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001401, Sample Num: 22416, Cur Loss: 0.36262131, Cur Avg Loss: 0.23293802, Log Avg loss: 0.32140581, Global Avg Loss: 0.86686843, Time: 0.0209 Steps: 67400, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001411, Sample Num: 22576, Cur Loss: 0.96943730, Cur Avg Loss: 0.23398070, Log Avg loss: 0.38006046, Global Avg Loss: 0.86679622, Time: 0.0208 Steps: 67410, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001421, Sample Num: 22736, Cur Loss: 0.35151404, Cur Avg Loss: 0.23480822, Log Avg loss: 0.35157115, Global Avg Loss: 0.86671980, Time: 0.0209 Steps: 67420, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001431, Sample Num: 22896, Cur Loss: 0.59900481, Cur Avg Loss: 0.23542380, Log Avg loss: 0.32289775, Global Avg Loss: 0.86663915, Time: 0.0209 Steps: 67430, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001441, Sample Num: 23056, Cur Loss: 0.17819726, Cur Avg Loss: 0.23499068, Log Avg loss: 0.17301124, Global Avg Loss: 0.86653630, Time: 0.0210 Steps: 67440, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001451, Sample Num: 23216, Cur Loss: 0.27765647, Cur Avg Loss: 0.23534022, Log Avg loss: 0.28570832, Global Avg Loss: 0.86645018, Time: 0.0208 Steps: 67450, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001461, Sample Num: 23376, Cur Loss: 0.08923224, Cur Avg Loss: 0.23586855, Log Avg loss: 0.31252988, Global Avg Loss: 0.86636807, Time: 0.0209 Steps: 67460, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001471, Sample Num: 23536, Cur Loss: 0.21159932, Cur Avg Loss: 0.23553879, Log Avg loss: 0.18736081, Global Avg Loss: 0.86626743, Time: 0.0209 Steps: 67470, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001481, Sample Num: 23696, Cur Loss: 0.16873117, Cur Avg Loss: 0.23517019, Log Avg loss: 0.18094874, Global Avg Loss: 0.86616587, Time: 0.0208 Steps: 67480, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001491, Sample Num: 23856, Cur Loss: 0.62105668, Cur Avg Loss: 0.23546408, Log Avg loss: 0.27898948, Global Avg Loss: 0.86607887, Time: 0.0210 Steps: 67490, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001501, Sample Num: 24016, Cur Loss: 0.17635708, Cur Avg Loss: 0.23522986, Log Avg loss: 0.20030839, Global Avg Loss: 0.86598024, Time: 0.0210 Steps: 67500, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001511, Sample Num: 24176, Cur Loss: 0.07478930, Cur Avg Loss: 0.23506464, Log Avg loss: 0.21026421, Global Avg Loss: 0.86588311, Time: 0.0211 Steps: 67510, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001521, Sample Num: 24336, Cur Loss: 0.28044027, Cur Avg Loss: 0.23516766, Log Avg loss: 0.25073389, Global Avg Loss: 0.86579201, Time: 0.0210 Steps: 67520, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001531, Sample Num: 24496, Cur Loss: 0.25087994, Cur Avg Loss: 0.23515811, Log Avg loss: 0.23370523, Global Avg Loss: 0.86569840, Time: 0.0208 Steps: 67530, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001541, Sample Num: 24656, Cur Loss: 0.14164622, Cur Avg Loss: 0.23526846, Log Avg loss: 0.25216395, Global Avg Loss: 0.86560756, Time: 0.0221 Steps: 67540, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001551, Sample Num: 24816, Cur Loss: 0.32052183, Cur Avg Loss: 0.23525014, Log Avg loss: 0.23242749, Global Avg Loss: 0.86551383, Time: 0.0209 Steps: 67550, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001561, Sample Num: 24976, Cur Loss: 0.31853041, Cur Avg Loss: 0.23505743, Log Avg loss: 0.20516707, Global Avg Loss: 0.86541609, Time: 0.0209 Steps: 67560, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001571, Sample Num: 25136, Cur Loss: 0.12499001, Cur Avg Loss: 0.23430122, Log Avg loss: 0.11625776, Global Avg Loss: 0.86530522, Time: 0.0210 Steps: 67570, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001581, Sample Num: 25296, Cur Loss: 0.18699540, Cur Avg Loss: 0.23392605, Log Avg loss: 0.17498612, Global Avg Loss: 0.86520307, Time: 0.0209 Steps: 67580, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001591, Sample Num: 25456, Cur Loss: 0.18829006, Cur Avg Loss: 0.23405017, Log Avg loss: 0.25367395, Global Avg Loss: 0.86511259, Time: 0.0209 Steps: 67590, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001601, Sample Num: 25616, Cur Loss: 0.16517079, Cur Avg Loss: 0.23377268, Log Avg loss: 0.18962386, Global Avg Loss: 0.86501267, Time: 0.0209 Steps: 67600, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001611, Sample Num: 25776, Cur Loss: 0.06561159, Cur Avg Loss: 0.23306370, Log Avg loss: 0.11955655, Global Avg Loss: 0.86490241, Time: 0.0209 Steps: 67610, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001621, Sample Num: 25936, Cur Loss: 0.71838957, Cur Avg Loss: 0.23364240, Log Avg loss: 0.32686963, Global Avg Loss: 0.86482284, Time: 0.0209 Steps: 67620, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001631, Sample Num: 26096, Cur Loss: 0.17374274, Cur Avg Loss: 0.23366734, Log Avg loss: 0.23771028, Global Avg Loss: 0.86473011, Time: 0.0209 Steps: 67630, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001641, Sample Num: 26256, Cur Loss: 0.25244308, Cur Avg Loss: 0.23375555, Log Avg loss: 0.24814333, Global Avg Loss: 0.86463896, Time: 0.0209 Steps: 67640, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001651, Sample Num: 26416, Cur Loss: 0.17742166, Cur Avg Loss: 0.23362068, Log Avg loss: 0.21148858, Global Avg Loss: 0.86454241, Time: 0.0208 Steps: 67650, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001661, Sample Num: 26576, Cur Loss: 0.13530859, Cur Avg Loss: 0.23351902, Log Avg loss: 0.21673449, Global Avg Loss: 0.86444666, Time: 0.0209 Steps: 67660, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001671, Sample Num: 26736, Cur Loss: 0.26575276, Cur Avg Loss: 0.23385687, Log Avg loss: 0.28997365, Global Avg Loss: 0.86436177, Time: 0.0209 Steps: 67670, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001681, Sample Num: 26896, Cur Loss: 0.17860301, Cur Avg Loss: 0.23388093, Log Avg loss: 0.23790199, Global Avg Loss: 0.86426921, Time: 0.0209 Steps: 67680, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001691, Sample Num: 27056, Cur Loss: 0.23578343, Cur Avg Loss: 0.23372370, Log Avg loss: 0.20729237, Global Avg Loss: 0.86417215, Time: 0.0209 Steps: 67690, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001701, Sample Num: 27216, Cur Loss: 0.11098766, Cur Avg Loss: 0.23424671, Log Avg loss: 0.32268775, Global Avg Loss: 0.86409217, Time: 0.0209 Steps: 67700, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001711, Sample Num: 27376, Cur Loss: 0.48795390, Cur Avg Loss: 0.23402834, Log Avg loss: 0.19688468, Global Avg Loss: 0.86399363, Time: 0.0210 Steps: 67710, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001721, Sample Num: 27536, Cur Loss: 0.19018681, Cur Avg Loss: 0.23381647, Log Avg loss: 0.19756584, Global Avg Loss: 0.86389522, Time: 0.0209 Steps: 67720, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001731, Sample Num: 27696, Cur Loss: 0.13500445, Cur Avg Loss: 0.23369947, Log Avg loss: 0.21356364, Global Avg Loss: 0.86379920, Time: 0.0209 Steps: 67730, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001741, Sample Num: 27856, Cur Loss: 0.29389632, Cur Avg Loss: 0.23358140, Log Avg loss: 0.21314312, Global Avg Loss: 0.86370315, Time: 0.0209 Steps: 67740, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001751, Sample Num: 28016, Cur Loss: 0.34353009, Cur Avg Loss: 0.23385996, Log Avg loss: 0.28235705, Global Avg Loss: 0.86361734, Time: 0.0209 Steps: 67750, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001761, Sample Num: 28176, Cur Loss: 0.12085286, Cur Avg Loss: 0.23343252, Log Avg loss: 0.15858736, Global Avg Loss: 0.86351329, Time: 0.0209 Steps: 67760, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001771, Sample Num: 28336, Cur Loss: 0.18573505, Cur Avg Loss: 0.23339713, Log Avg loss: 0.22716610, Global Avg Loss: 0.86341940, Time: 0.0209 Steps: 67770, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001781, Sample Num: 28496, Cur Loss: 0.15828991, Cur Avg Loss: 0.23291525, Log Avg loss: 0.14757357, Global Avg Loss: 0.86331378, Time: 0.0209 Steps: 67780, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001791, Sample Num: 28656, Cur Loss: 0.16008046, Cur Avg Loss: 0.23246212, Log Avg loss: 0.15176013, Global Avg Loss: 0.86320882, Time: 0.0209 Steps: 67790, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001801, Sample Num: 28816, Cur Loss: 0.07223934, Cur Avg Loss: 0.23222798, Log Avg loss: 0.19029356, Global Avg Loss: 0.86310957, Time: 0.0209 Steps: 67800, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001811, Sample Num: 28976, Cur Loss: 0.28775156, Cur Avg Loss: 0.23210680, Log Avg loss: 0.21028117, Global Avg Loss: 0.86301330, Time: 0.0209 Steps: 67810, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001821, Sample Num: 29136, Cur Loss: 0.97993541, Cur Avg Loss: 0.23309506, Log Avg loss: 0.41207006, Global Avg Loss: 0.86294680, Time: 0.0209 Steps: 67820, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001831, Sample Num: 29296, Cur Loss: 0.40528703, Cur Avg Loss: 0.23413333, Log Avg loss: 0.42320153, Global Avg Loss: 0.86288197, Time: 0.0209 Steps: 67830, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001841, Sample Num: 29456, Cur Loss: 0.23080379, Cur Avg Loss: 0.23443685, Log Avg loss: 0.29001192, Global Avg Loss: 0.86279753, Time: 0.0210 Steps: 67840, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001851, Sample Num: 29616, Cur Loss: 0.20184834, Cur Avg Loss: 0.23414468, Log Avg loss: 0.18035530, Global Avg Loss: 0.86269695, Time: 0.0209 Steps: 67850, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001861, Sample Num: 29776, Cur Loss: 0.19293177, Cur Avg Loss: 0.23445442, Log Avg loss: 0.29178816, Global Avg Loss: 0.86261282, Time: 0.0209 Steps: 67860, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001871, Sample Num: 29936, Cur Loss: 0.66189885, Cur Avg Loss: 0.23459285, Log Avg loss: 0.26035521, Global Avg Loss: 0.86252408, Time: 0.0209 Steps: 67870, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001881, Sample Num: 30096, Cur Loss: 0.31526357, Cur Avg Loss: 0.23465837, Log Avg loss: 0.24691542, Global Avg Loss: 0.86243339, Time: 0.0209 Steps: 67880, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001891, Sample Num: 30256, Cur Loss: 0.16197154, Cur Avg Loss: 0.23483640, Log Avg loss: 0.26832402, Global Avg Loss: 0.86234588, Time: 0.0209 Steps: 67890, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001901, Sample Num: 30416, Cur Loss: 0.16523902, Cur Avg Loss: 0.23459146, Log Avg loss: 0.18827307, Global Avg Loss: 0.86224661, Time: 0.0211 Steps: 67900, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001911, Sample Num: 30576, Cur Loss: 0.05008809, Cur Avg Loss: 0.23436919, Log Avg loss: 0.19211701, Global Avg Loss: 0.86214793, Time: 0.0209 Steps: 67910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001921, Sample Num: 30736, Cur Loss: 0.14082260, Cur Avg Loss: 0.23432252, Log Avg loss: 0.22540432, Global Avg Loss: 0.86205418, Time: 0.0210 Steps: 67920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001931, Sample Num: 30896, Cur Loss: 0.16556242, Cur Avg Loss: 0.23443806, Log Avg loss: 0.25663277, Global Avg Loss: 0.86196505, Time: 0.0210 Steps: 67930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001941, Sample Num: 31056, Cur Loss: 0.11835094, Cur Avg Loss: 0.23440544, Log Avg loss: 0.22810693, Global Avg Loss: 0.86187176, Time: 0.0209 Steps: 67940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001951, Sample Num: 31216, Cur Loss: 0.25770381, Cur Avg Loss: 0.23463831, Log Avg loss: 0.27983755, Global Avg Loss: 0.86178610, Time: 0.0211 Steps: 67950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001961, Sample Num: 31376, Cur Loss: 0.09889474, Cur Avg Loss: 0.23487729, Log Avg loss: 0.28150166, Global Avg Loss: 0.86170071, Time: 0.0209 Steps: 67960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001971, Sample Num: 31536, Cur Loss: 0.40121388, Cur Avg Loss: 0.23503355, Log Avg loss: 0.26567699, Global Avg Loss: 0.86161303, Time: 0.0211 Steps: 67970, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001981, Sample Num: 31696, Cur Loss: 0.18912716, Cur Avg Loss: 0.23484427, Log Avg loss: 0.19753660, Global Avg Loss: 0.86151534, Time: 0.0211 Steps: 67980, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001991, Sample Num: 31856, Cur Loss: 0.16708583, Cur Avg Loss: 0.23484020, Log Avg loss: 0.23403533, Global Avg Loss: 0.86142305, Time: 0.0211 Steps: 67990, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002001, Sample Num: 32016, Cur Loss: 0.13141873, Cur Avg Loss: 0.23529682, Log Avg loss: 0.32620983, Global Avg Loss: 0.86134434, Time: 0.0209 Steps: 68000, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002011, Sample Num: 32176, Cur Loss: 0.11355143, Cur Avg Loss: 0.23504767, Log Avg loss: 0.18519171, Global Avg Loss: 0.86124492, Time: 0.0211 Steps: 68010, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002021, Sample Num: 32336, Cur Loss: 0.24008511, Cur Avg Loss: 0.23495050, Log Avg loss: 0.21540990, Global Avg Loss: 0.86114997, Time: 0.0209 Steps: 68020, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002031, Sample Num: 32496, Cur Loss: 0.28541154, Cur Avg Loss: 0.23529948, Log Avg loss: 0.30582722, Global Avg Loss: 0.86106834, Time: 0.0211 Steps: 68030, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002041, Sample Num: 32656, Cur Loss: 0.48397353, Cur Avg Loss: 0.23536548, Log Avg loss: 0.24877077, Global Avg Loss: 0.86097835, Time: 0.0209 Steps: 68040, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002051, Sample Num: 32816, Cur Loss: 0.33392492, Cur Avg Loss: 0.23551953, Log Avg loss: 0.26696205, Global Avg Loss: 0.86089106, Time: 0.0246 Steps: 68050, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002061, Sample Num: 32976, Cur Loss: 0.18385778, Cur Avg Loss: 0.23538819, Log Avg loss: 0.20845008, Global Avg Loss: 0.86079520, Time: 0.0209 Steps: 68060, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002071, Sample Num: 33136, Cur Loss: 0.37436551, Cur Avg Loss: 0.23560277, Log Avg loss: 0.27982635, Global Avg Loss: 0.86070985, Time: 0.0209 Steps: 68070, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002081, Sample Num: 33296, Cur Loss: 0.28313184, Cur Avg Loss: 0.23556740, Log Avg loss: 0.22824373, Global Avg Loss: 0.86061695, Time: 0.0209 Steps: 68080, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002091, Sample Num: 33456, Cur Loss: 0.16192392, Cur Avg Loss: 0.23573156, Log Avg loss: 0.26989229, Global Avg Loss: 0.86053019, Time: 0.0209 Steps: 68090, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002101, Sample Num: 33616, Cur Loss: 0.28394121, Cur Avg Loss: 0.23566546, Log Avg loss: 0.22184429, Global Avg Loss: 0.86043641, Time: 0.0209 Steps: 68100, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002111, Sample Num: 33776, Cur Loss: 0.28583550, Cur Avg Loss: 0.23551552, Log Avg loss: 0.20401221, Global Avg Loss: 0.86034003, Time: 0.0209 Steps: 68110, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002121, Sample Num: 33936, Cur Loss: 0.18071839, Cur Avg Loss: 0.23525547, Log Avg loss: 0.18036050, Global Avg Loss: 0.86024021, Time: 0.0209 Steps: 68120, Updated lr: 0.000036 ***** Running evaluation checkpoint-68128 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-68128 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.702297, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.271824, "eval_total_loss": 191.092498, "eval_mae": 0.361091, "eval_mse": 0.271898, "eval_r2": 0.827163, "eval_sp_statistic": 0.894622, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.914772, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.438148, "test_total_loss": 219.950061, "test_mae": 0.390191, "test_mse": 0.438305, "test_r2": 0.717114, "test_sp_statistic": 0.864519, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.907393, "test_ps_pvalue": 0.0, "lr": 3.634139402560455e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8601635701399301, "train_cur_epoch_loss": 500.63751474022865, "train_cur_epoch_avg_loss": 0.23515148649141787, "train_cur_epoch_time": 44.70229721069336, "train_cur_epoch_avg_time": 0.02099685167247222, "epoch": 32, "step": 68128} ################################################## Training, Epoch: 0033, Batch: 000002, Sample Num: 32, Cur Loss: 0.14662960, Cur Avg Loss: 0.11025236, Log Avg loss: 0.18811622, Global Avg Loss: 0.86014156, Time: 0.0248 Steps: 68130, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000012, Sample Num: 192, Cur Loss: 0.40775669, Cur Avg Loss: 0.29097411, Log Avg loss: 0.32711846, Global Avg Loss: 0.86006333, Time: 0.0210 Steps: 68140, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000022, Sample Num: 352, Cur Loss: 0.17158991, Cur Avg Loss: 0.29128529, Log Avg loss: 0.29165871, Global Avg Loss: 0.85997993, Time: 0.0210 Steps: 68150, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000032, Sample Num: 512, Cur Loss: 0.24631125, Cur Avg Loss: 0.32357670, Log Avg loss: 0.39461778, Global Avg Loss: 0.85991165, Time: 0.0210 Steps: 68160, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000042, Sample Num: 672, Cur Loss: 0.37837672, Cur Avg Loss: 0.29991269, Log Avg loss: 0.22418786, Global Avg Loss: 0.85981840, Time: 0.0210 Steps: 68170, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000052, Sample Num: 832, Cur Loss: 0.31202704, Cur Avg Loss: 0.28406574, Log Avg loss: 0.21750859, Global Avg Loss: 0.85972419, Time: 0.0210 Steps: 68180, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000062, Sample Num: 992, Cur Loss: 0.38457346, Cur Avg Loss: 0.28720072, Log Avg loss: 0.30350262, Global Avg Loss: 0.85964262, Time: 0.0210 Steps: 68190, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000072, Sample Num: 1152, Cur Loss: 0.19907419, Cur Avg Loss: 0.27784017, Log Avg loss: 0.21980473, Global Avg Loss: 0.85954880, Time: 0.0210 Steps: 68200, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000082, Sample Num: 1312, Cur Loss: 0.13502721, Cur Avg Loss: 0.27347354, Log Avg loss: 0.24203381, Global Avg Loss: 0.85945827, Time: 0.0210 Steps: 68210, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000092, Sample Num: 1472, Cur Loss: 0.25149670, Cur Avg Loss: 0.26777261, Log Avg loss: 0.22102502, Global Avg Loss: 0.85936468, Time: 0.0210 Steps: 68220, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000102, Sample Num: 1632, Cur Loss: 0.29111409, Cur Avg Loss: 0.25959196, Log Avg loss: 0.18432991, Global Avg Loss: 0.85926575, Time: 0.0209 Steps: 68230, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000112, Sample Num: 1792, Cur Loss: 0.17905408, Cur Avg Loss: 0.25562199, Log Avg loss: 0.21512835, Global Avg Loss: 0.85917136, Time: 0.0210 Steps: 68240, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000122, Sample Num: 1952, Cur Loss: 0.17192012, Cur Avg Loss: 0.25148591, Log Avg loss: 0.20516178, Global Avg Loss: 0.85907553, Time: 0.0209 Steps: 68250, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000132, Sample Num: 2112, Cur Loss: 0.25502324, Cur Avg Loss: 0.24951733, Log Avg loss: 0.22550065, Global Avg Loss: 0.85898271, Time: 0.0209 Steps: 68260, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000142, Sample Num: 2272, Cur Loss: 0.21062061, Cur Avg Loss: 0.24270055, Log Avg loss: 0.15271912, Global Avg Loss: 0.85887926, Time: 0.0210 Steps: 68270, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000152, Sample Num: 2432, Cur Loss: 0.30542451, Cur Avg Loss: 0.24165006, Log Avg loss: 0.22673311, Global Avg Loss: 0.85878668, Time: 0.0209 Steps: 68280, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000162, Sample Num: 2592, Cur Loss: 0.21132548, Cur Avg Loss: 0.24165437, Log Avg loss: 0.24171981, Global Avg Loss: 0.85869632, Time: 0.0210 Steps: 68290, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000172, Sample Num: 2752, Cur Loss: 0.07929195, Cur Avg Loss: 0.23874699, Log Avg loss: 0.19164749, Global Avg Loss: 0.85859866, Time: 0.0209 Steps: 68300, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000182, Sample Num: 2912, Cur Loss: 0.14358976, Cur Avg Loss: 0.23499720, Log Avg loss: 0.17050086, Global Avg Loss: 0.85849792, Time: 0.0209 Steps: 68310, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000192, Sample Num: 3072, Cur Loss: 0.69906890, Cur Avg Loss: 0.23257628, Log Avg loss: 0.18851539, Global Avg Loss: 0.85839986, Time: 0.0209 Steps: 68320, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000202, Sample Num: 3232, Cur Loss: 0.14559817, Cur Avg Loss: 0.22948719, Log Avg loss: 0.17017665, Global Avg Loss: 0.85829914, Time: 0.0209 Steps: 68330, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000212, Sample Num: 3392, Cur Loss: 0.12431566, Cur Avg Loss: 0.22725153, Log Avg loss: 0.18209123, Global Avg Loss: 0.85820019, Time: 0.0209 Steps: 68340, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000222, Sample Num: 3552, Cur Loss: 0.23344208, Cur Avg Loss: 0.22362377, Log Avg loss: 0.14671532, Global Avg Loss: 0.85809610, Time: 0.0209 Steps: 68350, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000232, Sample Num: 3712, Cur Loss: 0.18534011, Cur Avg Loss: 0.22248887, Log Avg loss: 0.19729399, Global Avg Loss: 0.85799943, Time: 0.0209 Steps: 68360, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000242, Sample Num: 3872, Cur Loss: 0.08765082, Cur Avg Loss: 0.22220166, Log Avg loss: 0.21553846, Global Avg Loss: 0.85790546, Time: 0.0210 Steps: 68370, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000252, Sample Num: 4032, Cur Loss: 0.31722507, Cur Avg Loss: 0.22524192, Log Avg loss: 0.29881622, Global Avg Loss: 0.85782370, Time: 0.0210 Steps: 68380, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000262, Sample Num: 4192, Cur Loss: 0.25309685, Cur Avg Loss: 0.22734965, Log Avg loss: 0.28046452, Global Avg Loss: 0.85773928, Time: 0.0212 Steps: 68390, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000272, Sample Num: 4352, Cur Loss: 0.42087191, Cur Avg Loss: 0.22982093, Log Avg loss: 0.29456840, Global Avg Loss: 0.85765694, Time: 0.0210 Steps: 68400, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000282, Sample Num: 4512, Cur Loss: 0.46100026, Cur Avg Loss: 0.23114993, Log Avg loss: 0.26729876, Global Avg Loss: 0.85757065, Time: 0.0210 Steps: 68410, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000292, Sample Num: 4672, Cur Loss: 0.30524606, Cur Avg Loss: 0.23008372, Log Avg loss: 0.20001666, Global Avg Loss: 0.85747454, Time: 0.0210 Steps: 68420, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000302, Sample Num: 4832, Cur Loss: 0.22511919, Cur Avg Loss: 0.23318873, Log Avg loss: 0.32385483, Global Avg Loss: 0.85739656, Time: 0.0211 Steps: 68430, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000312, Sample Num: 4992, Cur Loss: 0.16649997, Cur Avg Loss: 0.23052677, Log Avg loss: 0.15013560, Global Avg Loss: 0.85729322, Time: 0.0210 Steps: 68440, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000322, Sample Num: 5152, Cur Loss: 0.27158523, Cur Avg Loss: 0.23260071, Log Avg loss: 0.29730784, Global Avg Loss: 0.85721141, Time: 0.0210 Steps: 68450, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000332, Sample Num: 5312, Cur Loss: 0.55747956, Cur Avg Loss: 0.23559664, Log Avg loss: 0.33206556, Global Avg Loss: 0.85713470, Time: 0.0209 Steps: 68460, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000342, Sample Num: 5472, Cur Loss: 0.20150927, Cur Avg Loss: 0.23720174, Log Avg loss: 0.29049091, Global Avg Loss: 0.85705195, Time: 0.0210 Steps: 68470, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000352, Sample Num: 5632, Cur Loss: 0.46565866, Cur Avg Loss: 0.24070147, Log Avg loss: 0.36039228, Global Avg Loss: 0.85697942, Time: 0.0210 Steps: 68480, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000362, Sample Num: 5792, Cur Loss: 0.08815207, Cur Avg Loss: 0.24204599, Log Avg loss: 0.28937303, Global Avg Loss: 0.85689654, Time: 0.0211 Steps: 68490, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000372, Sample Num: 5952, Cur Loss: 0.23352462, Cur Avg Loss: 0.24279124, Log Avg loss: 0.26976949, Global Avg Loss: 0.85681083, Time: 0.0210 Steps: 68500, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000382, Sample Num: 6112, Cur Loss: 0.23947588, Cur Avg Loss: 0.24303631, Log Avg loss: 0.25215271, Global Avg Loss: 0.85672257, Time: 0.0211 Steps: 68510, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000392, Sample Num: 6272, Cur Loss: 0.32162568, Cur Avg Loss: 0.24240561, Log Avg loss: 0.21831310, Global Avg Loss: 0.85662940, Time: 0.0211 Steps: 68520, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000402, Sample Num: 6432, Cur Loss: 0.25566694, Cur Avg Loss: 0.24120631, Log Avg loss: 0.19419343, Global Avg Loss: 0.85653274, Time: 0.0210 Steps: 68530, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000412, Sample Num: 6592, Cur Loss: 0.08291216, Cur Avg Loss: 0.23974503, Log Avg loss: 0.18100168, Global Avg Loss: 0.85643418, Time: 0.0210 Steps: 68540, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000422, Sample Num: 6752, Cur Loss: 0.34731132, Cur Avg Loss: 0.23910508, Log Avg loss: 0.21273918, Global Avg Loss: 0.85634028, Time: 0.0210 Steps: 68550, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000432, Sample Num: 6912, Cur Loss: 0.88992190, Cur Avg Loss: 0.24201686, Log Avg loss: 0.36489420, Global Avg Loss: 0.85626860, Time: 0.0210 Steps: 68560, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000442, Sample Num: 7072, Cur Loss: 0.09765290, Cur Avg Loss: 0.24180185, Log Avg loss: 0.23251305, Global Avg Loss: 0.85617763, Time: 0.0210 Steps: 68570, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000452, Sample Num: 7232, Cur Loss: 0.15704998, Cur Avg Loss: 0.24039928, Log Avg loss: 0.17840567, Global Avg Loss: 0.85607880, Time: 0.0210 Steps: 68580, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000462, Sample Num: 7392, Cur Loss: 0.19145204, Cur Avg Loss: 0.23853731, Log Avg loss: 0.15437648, Global Avg Loss: 0.85597650, Time: 0.0210 Steps: 68590, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000472, Sample Num: 7552, Cur Loss: 0.16886012, Cur Avg Loss: 0.23724574, Log Avg loss: 0.17757540, Global Avg Loss: 0.85587760, Time: 0.0210 Steps: 68600, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000482, Sample Num: 7712, Cur Loss: 0.33647475, Cur Avg Loss: 0.23715443, Log Avg loss: 0.23284433, Global Avg Loss: 0.85578680, Time: 0.0210 Steps: 68610, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000492, Sample Num: 7872, Cur Loss: 0.07866851, Cur Avg Loss: 0.23532510, Log Avg loss: 0.14715165, Global Avg Loss: 0.85568353, Time: 0.0211 Steps: 68620, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000502, Sample Num: 8032, Cur Loss: 0.46039727, Cur Avg Loss: 0.23455976, Log Avg loss: 0.19690492, Global Avg Loss: 0.85558754, Time: 0.0210 Steps: 68630, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000512, Sample Num: 8192, Cur Loss: 0.12941521, Cur Avg Loss: 0.23295427, Log Avg loss: 0.15235879, Global Avg Loss: 0.85548509, Time: 0.0256 Steps: 68640, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000522, Sample Num: 8352, Cur Loss: 0.26147714, Cur Avg Loss: 0.23377777, Log Avg loss: 0.27594076, Global Avg Loss: 0.85540067, Time: 0.0209 Steps: 68650, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000532, Sample Num: 8512, Cur Loss: 0.21673901, Cur Avg Loss: 0.23499638, Log Avg loss: 0.29860807, Global Avg Loss: 0.85531957, Time: 0.0210 Steps: 68660, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000542, Sample Num: 8672, Cur Loss: 0.31023282, Cur Avg Loss: 0.23414213, Log Avg loss: 0.18869573, Global Avg Loss: 0.85522250, Time: 0.0210 Steps: 68670, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000552, Sample Num: 8832, Cur Loss: 0.13228802, Cur Avg Loss: 0.23448579, Log Avg loss: 0.25311206, Global Avg Loss: 0.85513483, Time: 0.0210 Steps: 68680, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000562, Sample Num: 8992, Cur Loss: 0.57940310, Cur Avg Loss: 0.23484243, Log Avg loss: 0.25452935, Global Avg Loss: 0.85504739, Time: 0.0210 Steps: 68690, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000572, Sample Num: 9152, Cur Loss: 0.06478328, Cur Avg Loss: 0.23457652, Log Avg loss: 0.21963219, Global Avg Loss: 0.85495490, Time: 0.0210 Steps: 68700, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000582, Sample Num: 9312, Cur Loss: 0.37102523, Cur Avg Loss: 0.23556725, Log Avg loss: 0.29223685, Global Avg Loss: 0.85487300, Time: 0.0210 Steps: 68710, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000592, Sample Num: 9472, Cur Loss: 0.03277905, Cur Avg Loss: 0.23465452, Log Avg loss: 0.18153361, Global Avg Loss: 0.85477502, Time: 0.0210 Steps: 68720, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000602, Sample Num: 9632, Cur Loss: 0.08172572, Cur Avg Loss: 0.23356619, Log Avg loss: 0.16913692, Global Avg Loss: 0.85467526, Time: 0.0210 Steps: 68730, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000612, Sample Num: 9792, Cur Loss: 0.24134704, Cur Avg Loss: 0.23321390, Log Avg loss: 0.21200616, Global Avg Loss: 0.85458177, Time: 0.0209 Steps: 68740, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000622, Sample Num: 9952, Cur Loss: 0.21721199, Cur Avg Loss: 0.23374284, Log Avg loss: 0.26611403, Global Avg Loss: 0.85449617, Time: 0.0209 Steps: 68750, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000632, Sample Num: 10112, Cur Loss: 0.17281245, Cur Avg Loss: 0.23328176, Log Avg loss: 0.20460257, Global Avg Loss: 0.85440165, Time: 0.0210 Steps: 68760, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000642, Sample Num: 10272, Cur Loss: 0.06993298, Cur Avg Loss: 0.23172753, Log Avg loss: 0.13350041, Global Avg Loss: 0.85429683, Time: 0.0209 Steps: 68770, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000652, Sample Num: 10432, Cur Loss: 0.29056376, Cur Avg Loss: 0.23220557, Log Avg loss: 0.26289553, Global Avg Loss: 0.85421084, Time: 0.0210 Steps: 68780, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000662, Sample Num: 10592, Cur Loss: 0.10851517, Cur Avg Loss: 0.23106731, Log Avg loss: 0.15685304, Global Avg Loss: 0.85410947, Time: 0.0210 Steps: 68790, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000672, Sample Num: 10752, Cur Loss: 0.16133639, Cur Avg Loss: 0.23050899, Log Avg loss: 0.19354823, Global Avg Loss: 0.85401346, Time: 0.0210 Steps: 68800, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000682, Sample Num: 10912, Cur Loss: 0.04827788, Cur Avg Loss: 0.23107177, Log Avg loss: 0.26889016, Global Avg Loss: 0.85392842, Time: 0.0210 Steps: 68810, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000692, Sample Num: 11072, Cur Loss: 0.15078028, Cur Avg Loss: 0.23048262, Log Avg loss: 0.19030300, Global Avg Loss: 0.85383199, Time: 0.0209 Steps: 68820, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000702, Sample Num: 11232, Cur Loss: 0.13187350, Cur Avg Loss: 0.22951114, Log Avg loss: 0.16228423, Global Avg Loss: 0.85373152, Time: 0.0209 Steps: 68830, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000712, Sample Num: 11392, Cur Loss: 0.14217851, Cur Avg Loss: 0.22880360, Log Avg loss: 0.17913484, Global Avg Loss: 0.85363353, Time: 0.0209 Steps: 68840, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000722, Sample Num: 11552, Cur Loss: 0.08447738, Cur Avg Loss: 0.22817639, Log Avg loss: 0.18351845, Global Avg Loss: 0.85353620, Time: 0.0210 Steps: 68850, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000732, Sample Num: 11712, Cur Loss: 0.11024076, Cur Avg Loss: 0.22827834, Log Avg loss: 0.23563975, Global Avg Loss: 0.85344646, Time: 0.0210 Steps: 68860, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000742, Sample Num: 11872, Cur Loss: 0.29812801, Cur Avg Loss: 0.22681981, Log Avg loss: 0.12005518, Global Avg Loss: 0.85333997, Time: 0.0210 Steps: 68870, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000752, Sample Num: 12032, Cur Loss: 0.08428020, Cur Avg Loss: 0.22576473, Log Avg loss: 0.14747808, Global Avg Loss: 0.85323750, Time: 0.0210 Steps: 68880, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000762, Sample Num: 12192, Cur Loss: 0.34177133, Cur Avg Loss: 0.22582800, Log Avg loss: 0.23058551, Global Avg Loss: 0.85314711, Time: 0.0210 Steps: 68890, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000772, Sample Num: 12352, Cur Loss: 0.31139717, Cur Avg Loss: 0.22605066, Log Avg loss: 0.24301705, Global Avg Loss: 0.85305856, Time: 0.0246 Steps: 68900, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000782, Sample Num: 12512, Cur Loss: 0.11043711, Cur Avg Loss: 0.22575382, Log Avg loss: 0.20283841, Global Avg Loss: 0.85296420, Time: 0.0210 Steps: 68910, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000792, Sample Num: 12672, Cur Loss: 0.47281474, Cur Avg Loss: 0.22624997, Log Avg loss: 0.26504835, Global Avg Loss: 0.85287890, Time: 0.0210 Steps: 68920, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000802, Sample Num: 12832, Cur Loss: 0.19043759, Cur Avg Loss: 0.22647253, Log Avg loss: 0.24409928, Global Avg Loss: 0.85279058, Time: 0.0210 Steps: 68930, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000812, Sample Num: 12992, Cur Loss: 0.07301384, Cur Avg Loss: 0.22659142, Log Avg loss: 0.23612659, Global Avg Loss: 0.85270113, Time: 0.0210 Steps: 68940, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000822, Sample Num: 13152, Cur Loss: 0.65350163, Cur Avg Loss: 0.22753742, Log Avg loss: 0.30435279, Global Avg Loss: 0.85262160, Time: 0.0210 Steps: 68950, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000832, Sample Num: 13312, Cur Loss: 0.29297623, Cur Avg Loss: 0.22792839, Log Avg loss: 0.26006586, Global Avg Loss: 0.85253567, Time: 0.0210 Steps: 68960, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000842, Sample Num: 13472, Cur Loss: 0.14334929, Cur Avg Loss: 0.22816983, Log Avg loss: 0.24825778, Global Avg Loss: 0.85244806, Time: 0.0210 Steps: 68970, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000852, Sample Num: 13632, Cur Loss: 0.05067421, Cur Avg Loss: 0.22703159, Log Avg loss: 0.13119213, Global Avg Loss: 0.85234350, Time: 0.0210 Steps: 68980, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000862, Sample Num: 13792, Cur Loss: 0.10398960, Cur Avg Loss: 0.22725062, Log Avg loss: 0.24591165, Global Avg Loss: 0.85225560, Time: 0.0210 Steps: 68990, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000872, Sample Num: 13952, Cur Loss: 0.14601830, Cur Avg Loss: 0.22785604, Log Avg loss: 0.28004352, Global Avg Loss: 0.85217267, Time: 0.0212 Steps: 69000, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000882, Sample Num: 14112, Cur Loss: 0.32296151, Cur Avg Loss: 0.22914506, Log Avg loss: 0.34154741, Global Avg Loss: 0.85209868, Time: 0.0210 Steps: 69010, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000892, Sample Num: 14272, Cur Loss: 0.24334927, Cur Avg Loss: 0.22982497, Log Avg loss: 0.28979318, Global Avg Loss: 0.85201721, Time: 0.0210 Steps: 69020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000902, Sample Num: 14432, Cur Loss: 0.29354173, Cur Avg Loss: 0.23026642, Log Avg loss: 0.26964313, Global Avg Loss: 0.85193284, Time: 0.0210 Steps: 69030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000912, Sample Num: 14592, Cur Loss: 0.11540032, Cur Avg Loss: 0.23026338, Log Avg loss: 0.22998919, Global Avg Loss: 0.85184276, Time: 0.0210 Steps: 69040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000922, Sample Num: 14752, Cur Loss: 0.08887419, Cur Avg Loss: 0.23098951, Log Avg loss: 0.29721259, Global Avg Loss: 0.85176243, Time: 0.0210 Steps: 69050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000932, Sample Num: 14912, Cur Loss: 0.12335636, Cur Avg Loss: 0.23045916, Log Avg loss: 0.18156096, Global Avg Loss: 0.85166539, Time: 0.0210 Steps: 69060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000942, Sample Num: 15072, Cur Loss: 0.37221754, Cur Avg Loss: 0.23067761, Log Avg loss: 0.25103761, Global Avg Loss: 0.85157843, Time: 0.0210 Steps: 69070, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000952, Sample Num: 15232, Cur Loss: 0.33614254, Cur Avg Loss: 0.23029106, Log Avg loss: 0.19387837, Global Avg Loss: 0.85148322, Time: 0.0210 Steps: 69080, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000962, Sample Num: 15392, Cur Loss: 0.24518178, Cur Avg Loss: 0.23048129, Log Avg loss: 0.24859056, Global Avg Loss: 0.85139596, Time: 0.0210 Steps: 69090, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000972, Sample Num: 15552, Cur Loss: 0.28096142, Cur Avg Loss: 0.23014904, Log Avg loss: 0.19818664, Global Avg Loss: 0.85130143, Time: 0.0210 Steps: 69100, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000982, Sample Num: 15712, Cur Loss: 0.38113660, Cur Avg Loss: 0.22996064, Log Avg loss: 0.21164858, Global Avg Loss: 0.85120887, Time: 0.0210 Steps: 69110, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000992, Sample Num: 15872, Cur Loss: 0.28729966, Cur Avg Loss: 0.23030115, Log Avg loss: 0.26373894, Global Avg Loss: 0.85112388, Time: 0.0210 Steps: 69120, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001002, Sample Num: 16032, Cur Loss: 0.04568926, Cur Avg Loss: 0.22972286, Log Avg loss: 0.17235649, Global Avg Loss: 0.85102569, Time: 0.0210 Steps: 69130, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001012, Sample Num: 16192, Cur Loss: 0.15099710, Cur Avg Loss: 0.22941669, Log Avg loss: 0.19873864, Global Avg Loss: 0.85093135, Time: 0.0210 Steps: 69140, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001022, Sample Num: 16352, Cur Loss: 0.06077619, Cur Avg Loss: 0.22809040, Log Avg loss: 0.09386931, Global Avg Loss: 0.85082187, Time: 0.0210 Steps: 69150, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001032, Sample Num: 16512, Cur Loss: 0.09004009, Cur Avg Loss: 0.22720070, Log Avg loss: 0.13627397, Global Avg Loss: 0.85071855, Time: 0.0210 Steps: 69160, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001042, Sample Num: 16672, Cur Loss: 0.26600984, Cur Avg Loss: 0.22762506, Log Avg loss: 0.27141900, Global Avg Loss: 0.85063480, Time: 0.0210 Steps: 69170, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001052, Sample Num: 16832, Cur Loss: 0.26590258, Cur Avg Loss: 0.22762276, Log Avg loss: 0.22738333, Global Avg Loss: 0.85054471, Time: 0.0210 Steps: 69180, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001062, Sample Num: 16992, Cur Loss: 0.13112313, Cur Avg Loss: 0.22779850, Log Avg loss: 0.24628604, Global Avg Loss: 0.85045737, Time: 0.0210 Steps: 69190, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001072, Sample Num: 17152, Cur Loss: 0.14204277, Cur Avg Loss: 0.22735970, Log Avg loss: 0.18075880, Global Avg Loss: 0.85036060, Time: 0.0210 Steps: 69200, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001082, Sample Num: 17312, Cur Loss: 0.34413481, Cur Avg Loss: 0.22714031, Log Avg loss: 0.20362206, Global Avg Loss: 0.85026715, Time: 0.0210 Steps: 69210, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001092, Sample Num: 17472, Cur Loss: 0.14639449, Cur Avg Loss: 0.22677446, Log Avg loss: 0.18718911, Global Avg Loss: 0.85017136, Time: 0.0212 Steps: 69220, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001102, Sample Num: 17632, Cur Loss: 0.22559142, Cur Avg Loss: 0.22640506, Log Avg loss: 0.18606731, Global Avg Loss: 0.85007543, Time: 0.0210 Steps: 69230, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001112, Sample Num: 17792, Cur Loss: 0.26295701, Cur Avg Loss: 0.22622634, Log Avg loss: 0.20653080, Global Avg Loss: 0.84998249, Time: 0.0210 Steps: 69240, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001122, Sample Num: 17952, Cur Loss: 0.15570937, Cur Avg Loss: 0.22589661, Log Avg loss: 0.18923118, Global Avg Loss: 0.84988707, Time: 0.0210 Steps: 69250, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001132, Sample Num: 18112, Cur Loss: 0.45313728, Cur Avg Loss: 0.22637634, Log Avg loss: 0.28020139, Global Avg Loss: 0.84980482, Time: 0.0210 Steps: 69260, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001142, Sample Num: 18272, Cur Loss: 0.14756957, Cur Avg Loss: 0.22569794, Log Avg loss: 0.14890349, Global Avg Loss: 0.84970363, Time: 0.0210 Steps: 69270, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001152, Sample Num: 18432, Cur Loss: 0.41203910, Cur Avg Loss: 0.22585504, Log Avg loss: 0.24379510, Global Avg Loss: 0.84961618, Time: 0.0210 Steps: 69280, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001162, Sample Num: 18592, Cur Loss: 0.11309278, Cur Avg Loss: 0.22588822, Log Avg loss: 0.22971080, Global Avg Loss: 0.84952671, Time: 0.0210 Steps: 69290, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001172, Sample Num: 18752, Cur Loss: 0.09905070, Cur Avg Loss: 0.22535452, Log Avg loss: 0.16333939, Global Avg Loss: 0.84942769, Time: 0.0210 Steps: 69300, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001182, Sample Num: 18912, Cur Loss: 0.09913944, Cur Avg Loss: 0.22467277, Log Avg loss: 0.14477158, Global Avg Loss: 0.84932603, Time: 0.0210 Steps: 69310, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001192, Sample Num: 19072, Cur Loss: 0.12931944, Cur Avg Loss: 0.22481306, Log Avg loss: 0.24139437, Global Avg Loss: 0.84923833, Time: 0.0210 Steps: 69320, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001202, Sample Num: 19232, Cur Loss: 0.02641041, Cur Avg Loss: 0.22536214, Log Avg loss: 0.29081343, Global Avg Loss: 0.84915778, Time: 0.0210 Steps: 69330, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001212, Sample Num: 19392, Cur Loss: 0.64216101, Cur Avg Loss: 0.22660730, Log Avg loss: 0.37627561, Global Avg Loss: 0.84908958, Time: 0.0210 Steps: 69340, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001222, Sample Num: 19552, Cur Loss: 1.04652917, Cur Avg Loss: 0.22931248, Log Avg loss: 0.55717989, Global Avg Loss: 0.84904749, Time: 0.0210 Steps: 69350, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001232, Sample Num: 19712, Cur Loss: 0.15293035, Cur Avg Loss: 0.23016815, Log Avg loss: 0.33473056, Global Avg Loss: 0.84897334, Time: 0.0209 Steps: 69360, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001242, Sample Num: 19872, Cur Loss: 0.12816852, Cur Avg Loss: 0.23046228, Log Avg loss: 0.26669914, Global Avg Loss: 0.84888940, Time: 0.0208 Steps: 69370, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001252, Sample Num: 20032, Cur Loss: 0.18282229, Cur Avg Loss: 0.23043639, Log Avg loss: 0.22722156, Global Avg Loss: 0.84879980, Time: 0.0208 Steps: 69380, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001262, Sample Num: 20192, Cur Loss: 0.28081578, Cur Avg Loss: 0.23066446, Log Avg loss: 0.25921804, Global Avg Loss: 0.84871483, Time: 0.0208 Steps: 69390, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001272, Sample Num: 20352, Cur Loss: 0.08338126, Cur Avg Loss: 0.23059658, Log Avg loss: 0.22203014, Global Avg Loss: 0.84862453, Time: 0.0208 Steps: 69400, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001282, Sample Num: 20512, Cur Loss: 0.28544712, Cur Avg Loss: 0.23046022, Log Avg loss: 0.21311539, Global Avg Loss: 0.84853297, Time: 0.0247 Steps: 69410, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001292, Sample Num: 20672, Cur Loss: 0.14355226, Cur Avg Loss: 0.23038140, Log Avg loss: 0.22027665, Global Avg Loss: 0.84844247, Time: 0.0209 Steps: 69420, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001302, Sample Num: 20832, Cur Loss: 0.31257394, Cur Avg Loss: 0.23003975, Log Avg loss: 0.18589891, Global Avg Loss: 0.84834705, Time: 0.0209 Steps: 69430, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001312, Sample Num: 20992, Cur Loss: 0.14513338, Cur Avg Loss: 0.23010559, Log Avg loss: 0.23867724, Global Avg Loss: 0.84825925, Time: 0.0209 Steps: 69440, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001322, Sample Num: 21152, Cur Loss: 0.22901192, Cur Avg Loss: 0.22997793, Log Avg loss: 0.21322953, Global Avg Loss: 0.84816781, Time: 0.0209 Steps: 69450, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001332, Sample Num: 21312, Cur Loss: 0.64859176, Cur Avg Loss: 0.23048607, Log Avg loss: 0.29766209, Global Avg Loss: 0.84808856, Time: 0.0209 Steps: 69460, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001342, Sample Num: 21472, Cur Loss: 0.17888281, Cur Avg Loss: 0.23002888, Log Avg loss: 0.16913146, Global Avg Loss: 0.84799082, Time: 0.0208 Steps: 69470, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001352, Sample Num: 21632, Cur Loss: 0.10031185, Cur Avg Loss: 0.22977396, Log Avg loss: 0.19556339, Global Avg Loss: 0.84789692, Time: 0.0209 Steps: 69480, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001362, Sample Num: 21792, Cur Loss: 0.20737365, Cur Avg Loss: 0.22942427, Log Avg loss: 0.18214671, Global Avg Loss: 0.84780112, Time: 0.0209 Steps: 69490, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001372, Sample Num: 21952, Cur Loss: 0.15344077, Cur Avg Loss: 0.22925034, Log Avg loss: 0.20556092, Global Avg Loss: 0.84770871, Time: 0.0209 Steps: 69500, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001382, Sample Num: 22112, Cur Loss: 0.08037505, Cur Avg Loss: 0.22872692, Log Avg loss: 0.15691353, Global Avg Loss: 0.84760933, Time: 0.0208 Steps: 69510, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001392, Sample Num: 22272, Cur Loss: 0.06030980, Cur Avg Loss: 0.22865655, Log Avg loss: 0.21893087, Global Avg Loss: 0.84751890, Time: 0.0209 Steps: 69520, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001402, Sample Num: 22432, Cur Loss: 0.29241428, Cur Avg Loss: 0.22850360, Log Avg loss: 0.20721296, Global Avg Loss: 0.84742680, Time: 0.0209 Steps: 69530, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001412, Sample Num: 22592, Cur Loss: 0.14673819, Cur Avg Loss: 0.22825666, Log Avg loss: 0.19363597, Global Avg Loss: 0.84733279, Time: 0.0209 Steps: 69540, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001422, Sample Num: 22752, Cur Loss: 0.13372840, Cur Avg Loss: 0.22861261, Log Avg loss: 0.27887260, Global Avg Loss: 0.84725105, Time: 0.0209 Steps: 69550, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001432, Sample Num: 22912, Cur Loss: 0.12705009, Cur Avg Loss: 0.22845641, Log Avg loss: 0.20624518, Global Avg Loss: 0.84715890, Time: 0.0208 Steps: 69560, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001442, Sample Num: 23072, Cur Loss: 0.25778538, Cur Avg Loss: 0.22855148, Log Avg loss: 0.24216508, Global Avg Loss: 0.84707194, Time: 0.0208 Steps: 69570, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001452, Sample Num: 23232, Cur Loss: 0.11503448, Cur Avg Loss: 0.22843880, Log Avg loss: 0.21219039, Global Avg Loss: 0.84698070, Time: 0.0209 Steps: 69580, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001462, Sample Num: 23392, Cur Loss: 0.33070913, Cur Avg Loss: 0.22836277, Log Avg loss: 0.21732294, Global Avg Loss: 0.84689022, Time: 0.0209 Steps: 69590, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001472, Sample Num: 23552, Cur Loss: 0.22463158, Cur Avg Loss: 0.22877477, Log Avg loss: 0.28900984, Global Avg Loss: 0.84681006, Time: 0.0209 Steps: 69600, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001482, Sample Num: 23712, Cur Loss: 0.09157695, Cur Avg Loss: 0.22880190, Log Avg loss: 0.23279466, Global Avg Loss: 0.84672185, Time: 0.0209 Steps: 69610, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001492, Sample Num: 23872, Cur Loss: 0.13865408, Cur Avg Loss: 0.22908189, Log Avg loss: 0.27057687, Global Avg Loss: 0.84663910, Time: 0.0209 Steps: 69620, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001502, Sample Num: 24032, Cur Loss: 0.08149065, Cur Avg Loss: 0.22893911, Log Avg loss: 0.20763608, Global Avg Loss: 0.84654733, Time: 0.0208 Steps: 69630, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001512, Sample Num: 24192, Cur Loss: 0.19060999, Cur Avg Loss: 0.22900267, Log Avg loss: 0.23855048, Global Avg Loss: 0.84646002, Time: 0.0208 Steps: 69640, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001522, Sample Num: 24352, Cur Loss: 0.08013989, Cur Avg Loss: 0.22852653, Log Avg loss: 0.15653352, Global Avg Loss: 0.84636096, Time: 0.0209 Steps: 69650, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001532, Sample Num: 24512, Cur Loss: 0.38001776, Cur Avg Loss: 0.22882591, Log Avg loss: 0.27439160, Global Avg Loss: 0.84627885, Time: 0.0209 Steps: 69660, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001542, Sample Num: 24672, Cur Loss: 0.13520356, Cur Avg Loss: 0.22857249, Log Avg loss: 0.18974853, Global Avg Loss: 0.84618462, Time: 0.0210 Steps: 69670, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001552, Sample Num: 24832, Cur Loss: 0.54487097, Cur Avg Loss: 0.22890695, Log Avg loss: 0.28048048, Global Avg Loss: 0.84610343, Time: 0.0209 Steps: 69680, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001562, Sample Num: 24992, Cur Loss: 0.11772029, Cur Avg Loss: 0.22888917, Log Avg loss: 0.22612967, Global Avg Loss: 0.84601447, Time: 0.0209 Steps: 69690, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001572, Sample Num: 25152, Cur Loss: 0.21055627, Cur Avg Loss: 0.22869968, Log Avg loss: 0.19910117, Global Avg Loss: 0.84592166, Time: 0.0208 Steps: 69700, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001582, Sample Num: 25312, Cur Loss: 0.24965826, Cur Avg Loss: 0.22892781, Log Avg loss: 0.26478993, Global Avg Loss: 0.84583829, Time: 0.0209 Steps: 69710, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001592, Sample Num: 25472, Cur Loss: 0.21828520, Cur Avg Loss: 0.22868673, Log Avg loss: 0.19054831, Global Avg Loss: 0.84574431, Time: 0.0209 Steps: 69720, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001602, Sample Num: 25632, Cur Loss: 0.01784189, Cur Avg Loss: 0.22850198, Log Avg loss: 0.19909003, Global Avg Loss: 0.84565157, Time: 0.0208 Steps: 69730, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001612, Sample Num: 25792, Cur Loss: 0.16825418, Cur Avg Loss: 0.22792498, Log Avg loss: 0.13548888, Global Avg Loss: 0.84554974, Time: 0.0208 Steps: 69740, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001622, Sample Num: 25952, Cur Loss: 0.37415111, Cur Avg Loss: 0.22827398, Log Avg loss: 0.28453384, Global Avg Loss: 0.84546931, Time: 0.0209 Steps: 69750, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001632, Sample Num: 26112, Cur Loss: 0.09176905, Cur Avg Loss: 0.22802497, Log Avg loss: 0.18763466, Global Avg Loss: 0.84537501, Time: 0.0209 Steps: 69760, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001642, Sample Num: 26272, Cur Loss: 0.09151633, Cur Avg Loss: 0.22773812, Log Avg loss: 0.18092389, Global Avg Loss: 0.84527977, Time: 0.0208 Steps: 69770, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001652, Sample Num: 26432, Cur Loss: 0.15994471, Cur Avg Loss: 0.22782495, Log Avg loss: 0.24208288, Global Avg Loss: 0.84519333, Time: 0.0208 Steps: 69780, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001662, Sample Num: 26592, Cur Loss: 0.21456844, Cur Avg Loss: 0.22774875, Log Avg loss: 0.21516085, Global Avg Loss: 0.84510305, Time: 0.0208 Steps: 69790, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001672, Sample Num: 26752, Cur Loss: 0.24211149, Cur Avg Loss: 0.22752862, Log Avg loss: 0.19094238, Global Avg Loss: 0.84500933, Time: 0.0209 Steps: 69800, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001682, Sample Num: 26912, Cur Loss: 0.17321780, Cur Avg Loss: 0.22734510, Log Avg loss: 0.19666073, Global Avg Loss: 0.84491646, Time: 0.0208 Steps: 69810, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001692, Sample Num: 27072, Cur Loss: 0.24835779, Cur Avg Loss: 0.22785008, Log Avg loss: 0.31278735, Global Avg Loss: 0.84484025, Time: 0.0208 Steps: 69820, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001702, Sample Num: 27232, Cur Loss: 0.11286645, Cur Avg Loss: 0.22855469, Log Avg loss: 0.34777557, Global Avg Loss: 0.84476906, Time: 0.0209 Steps: 69830, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001712, Sample Num: 27392, Cur Loss: 0.28930229, Cur Avg Loss: 0.22925853, Log Avg loss: 0.34905092, Global Avg Loss: 0.84469809, Time: 0.0209 Steps: 69840, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001722, Sample Num: 27552, Cur Loss: 0.34188247, Cur Avg Loss: 0.22925339, Log Avg loss: 0.22837430, Global Avg Loss: 0.84460985, Time: 0.0209 Steps: 69850, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001732, Sample Num: 27712, Cur Loss: 0.24258173, Cur Avg Loss: 0.22896212, Log Avg loss: 0.17880573, Global Avg Loss: 0.84451454, Time: 0.0208 Steps: 69860, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001742, Sample Num: 27872, Cur Loss: 0.09127516, Cur Avg Loss: 0.22874115, Log Avg loss: 0.19046857, Global Avg Loss: 0.84442094, Time: 0.0209 Steps: 69870, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001752, Sample Num: 28032, Cur Loss: 0.29761982, Cur Avg Loss: 0.22912622, Log Avg loss: 0.29620533, Global Avg Loss: 0.84434248, Time: 0.0209 Steps: 69880, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001762, Sample Num: 28192, Cur Loss: 0.09293881, Cur Avg Loss: 0.22887846, Log Avg loss: 0.18547083, Global Avg Loss: 0.84424821, Time: 0.0209 Steps: 69890, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001772, Sample Num: 28352, Cur Loss: 0.18773621, Cur Avg Loss: 0.22864667, Log Avg loss: 0.18780646, Global Avg Loss: 0.84415430, Time: 0.0208 Steps: 69900, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001782, Sample Num: 28512, Cur Loss: 0.20299214, Cur Avg Loss: 0.22884444, Log Avg loss: 0.26388808, Global Avg Loss: 0.84407130, Time: 0.0208 Steps: 69910, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001792, Sample Num: 28672, Cur Loss: 0.06665897, Cur Avg Loss: 0.22838271, Log Avg loss: 0.14610218, Global Avg Loss: 0.84397147, Time: 0.0254 Steps: 69920, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001802, Sample Num: 28832, Cur Loss: 0.22488299, Cur Avg Loss: 0.22819152, Log Avg loss: 0.19393044, Global Avg Loss: 0.84387852, Time: 0.0209 Steps: 69930, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001812, Sample Num: 28992, Cur Loss: 0.09597953, Cur Avg Loss: 0.22812043, Log Avg loss: 0.21531122, Global Avg Loss: 0.84378865, Time: 0.0209 Steps: 69940, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001822, Sample Num: 29152, Cur Loss: 0.11634319, Cur Avg Loss: 0.22858580, Log Avg loss: 0.31290961, Global Avg Loss: 0.84371275, Time: 0.0209 Steps: 69950, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001832, Sample Num: 29312, Cur Loss: 0.19011599, Cur Avg Loss: 0.22855000, Log Avg loss: 0.22202745, Global Avg Loss: 0.84362389, Time: 0.0209 Steps: 69960, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001842, Sample Num: 29472, Cur Loss: 0.20349789, Cur Avg Loss: 0.22912879, Log Avg loss: 0.33516339, Global Avg Loss: 0.84355122, Time: 0.0209 Steps: 69970, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001852, Sample Num: 29632, Cur Loss: 0.10106274, Cur Avg Loss: 0.22902406, Log Avg loss: 0.20973272, Global Avg Loss: 0.84346065, Time: 0.0209 Steps: 69980, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001862, Sample Num: 29792, Cur Loss: 0.22193173, Cur Avg Loss: 0.22900749, Log Avg loss: 0.22593922, Global Avg Loss: 0.84337242, Time: 0.0209 Steps: 69990, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001872, Sample Num: 29952, Cur Loss: 0.12472456, Cur Avg Loss: 0.22855151, Log Avg loss: 0.14364825, Global Avg Loss: 0.84327246, Time: 0.0209 Steps: 70000, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001882, Sample Num: 30112, Cur Loss: 0.17569990, Cur Avg Loss: 0.22850944, Log Avg loss: 0.22063407, Global Avg Loss: 0.84318352, Time: 0.0209 Steps: 70010, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001892, Sample Num: 30272, Cur Loss: 0.43672329, Cur Avg Loss: 0.22857109, Log Avg loss: 0.24017233, Global Avg Loss: 0.84309740, Time: 0.0209 Steps: 70020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001902, Sample Num: 30432, Cur Loss: 0.08610103, Cur Avg Loss: 0.22831653, Log Avg loss: 0.18015368, Global Avg Loss: 0.84300274, Time: 0.0209 Steps: 70030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001912, Sample Num: 30592, Cur Loss: 0.14513989, Cur Avg Loss: 0.22852252, Log Avg loss: 0.26770225, Global Avg Loss: 0.84292060, Time: 0.0208 Steps: 70040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001922, Sample Num: 30752, Cur Loss: 0.39995605, Cur Avg Loss: 0.22855145, Log Avg loss: 0.23408373, Global Avg Loss: 0.84283368, Time: 0.0208 Steps: 70050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001932, Sample Num: 30912, Cur Loss: 0.20013265, Cur Avg Loss: 0.22827370, Log Avg loss: 0.17488859, Global Avg Loss: 0.84273835, Time: 0.0209 Steps: 70060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001942, Sample Num: 31072, Cur Loss: 0.16112429, Cur Avg Loss: 0.22829878, Log Avg loss: 0.23314444, Global Avg Loss: 0.84265135, Time: 0.0208 Steps: 70070, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001952, Sample Num: 31232, Cur Loss: 0.17161258, Cur Avg Loss: 0.22811782, Log Avg loss: 0.19297636, Global Avg Loss: 0.84255864, Time: 0.0209 Steps: 70080, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001962, Sample Num: 31392, Cur Loss: 0.06098454, Cur Avg Loss: 0.22796758, Log Avg loss: 0.19863976, Global Avg Loss: 0.84246677, Time: 0.0209 Steps: 70090, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001972, Sample Num: 31552, Cur Loss: 0.21451834, Cur Avg Loss: 0.22815993, Log Avg loss: 0.26589930, Global Avg Loss: 0.84238452, Time: 0.0209 Steps: 70100, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001982, Sample Num: 31712, Cur Loss: 0.14883555, Cur Avg Loss: 0.22862943, Log Avg loss: 0.32121423, Global Avg Loss: 0.84231019, Time: 0.0209 Steps: 70110, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001992, Sample Num: 31872, Cur Loss: 0.08513400, Cur Avg Loss: 0.22882998, Log Avg loss: 0.26857939, Global Avg Loss: 0.84222837, Time: 0.0209 Steps: 70120, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002002, Sample Num: 32032, Cur Loss: 0.20548326, Cur Avg Loss: 0.22893775, Log Avg loss: 0.25040536, Global Avg Loss: 0.84214398, Time: 0.0209 Steps: 70130, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002012, Sample Num: 32192, Cur Loss: 0.36628103, Cur Avg Loss: 0.22891657, Log Avg loss: 0.22467632, Global Avg Loss: 0.84205594, Time: 0.0209 Steps: 70140, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002022, Sample Num: 32352, Cur Loss: 0.17095993, Cur Avg Loss: 0.22851186, Log Avg loss: 0.14708538, Global Avg Loss: 0.84195687, Time: 0.0209 Steps: 70150, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002032, Sample Num: 32512, Cur Loss: 0.12838604, Cur Avg Loss: 0.22827867, Log Avg loss: 0.18112622, Global Avg Loss: 0.84186268, Time: 0.0209 Steps: 70160, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002042, Sample Num: 32672, Cur Loss: 0.59614480, Cur Avg Loss: 0.22862820, Log Avg loss: 0.29965404, Global Avg Loss: 0.84178541, Time: 0.0209 Steps: 70170, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002052, Sample Num: 32832, Cur Loss: 0.14835477, Cur Avg Loss: 0.22924484, Log Avg loss: 0.35516214, Global Avg Loss: 0.84171607, Time: 0.0246 Steps: 70180, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002062, Sample Num: 32992, Cur Loss: 0.33024043, Cur Avg Loss: 0.22923162, Log Avg loss: 0.22652012, Global Avg Loss: 0.84162843, Time: 0.0209 Steps: 70190, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002072, Sample Num: 33152, Cur Loss: 0.14913327, Cur Avg Loss: 0.22930105, Log Avg loss: 0.24361578, Global Avg Loss: 0.84154324, Time: 0.0209 Steps: 70200, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002082, Sample Num: 33312, Cur Loss: 0.19825837, Cur Avg Loss: 0.22949435, Log Avg loss: 0.26954740, Global Avg Loss: 0.84146177, Time: 0.0209 Steps: 70210, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002092, Sample Num: 33472, Cur Loss: 0.02577712, Cur Avg Loss: 0.22924423, Log Avg loss: 0.17716943, Global Avg Loss: 0.84136717, Time: 0.0209 Steps: 70220, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002102, Sample Num: 33632, Cur Loss: 0.19426063, Cur Avg Loss: 0.22894577, Log Avg loss: 0.16650701, Global Avg Loss: 0.84127108, Time: 0.0209 Steps: 70230, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002112, Sample Num: 33792, Cur Loss: 0.72060299, Cur Avg Loss: 0.22892978, Log Avg loss: 0.22556849, Global Avg Loss: 0.84118342, Time: 0.0209 Steps: 70240, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002122, Sample Num: 33952, Cur Loss: 0.11476544, Cur Avg Loss: 0.22878577, Log Avg loss: 0.19837067, Global Avg Loss: 0.84109192, Time: 0.0209 Steps: 70250, Updated lr: 0.000034 ***** Running evaluation checkpoint-70257 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-70257 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.779571, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.303903, "eval_total_loss": 213.644089, "eval_mae": 0.438186, "eval_mse": 0.303931, "eval_r2": 0.806802, "eval_sp_statistic": 0.894274, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.91955, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.32904, "test_total_loss": 165.177873, "test_mae": 0.362415, "test_mse": 0.329158, "test_r2": 0.787558, "test_sp_statistic": 0.868668, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.915461, "test_ps_pvalue": 0.0, "lr": 3.432242769084875e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8410245374632821, "train_cur_epoch_loss": 486.637222064659, "train_cur_epoch_avg_loss": 0.2285754918105491, "train_cur_epoch_time": 44.77957081794739, "train_cur_epoch_avg_time": 0.02103314740157228, "epoch": 33, "step": 70257} ################################################## Training, Epoch: 0034, Batch: 000003, Sample Num: 48, Cur Loss: 0.05710193, Cur Avg Loss: 0.06308676, Log Avg loss: 0.13430840, Global Avg Loss: 0.84099132, Time: 0.0245 Steps: 70260, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000013, Sample Num: 208, Cur Loss: 0.09913968, Cur Avg Loss: 0.15007230, Log Avg loss: 0.17616796, Global Avg Loss: 0.84089671, Time: 0.0209 Steps: 70270, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000023, Sample Num: 368, Cur Loss: 0.18210058, Cur Avg Loss: 0.16882767, Log Avg loss: 0.19320967, Global Avg Loss: 0.84080455, Time: 0.0209 Steps: 70280, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000033, Sample Num: 528, Cur Loss: 0.10009319, Cur Avg Loss: 0.18598593, Log Avg loss: 0.22544991, Global Avg Loss: 0.84071701, Time: 0.0209 Steps: 70290, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000043, Sample Num: 688, Cur Loss: 0.14324182, Cur Avg Loss: 0.17427333, Log Avg loss: 0.13562176, Global Avg Loss: 0.84061671, Time: 0.0209 Steps: 70300, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000053, Sample Num: 848, Cur Loss: 0.14694074, Cur Avg Loss: 0.17976820, Log Avg loss: 0.20339616, Global Avg Loss: 0.84052608, Time: 0.0209 Steps: 70310, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000063, Sample Num: 1008, Cur Loss: 0.15348154, Cur Avg Loss: 0.17912751, Log Avg loss: 0.17573183, Global Avg Loss: 0.84043154, Time: 0.0209 Steps: 70320, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000073, Sample Num: 1168, Cur Loss: 0.16047475, Cur Avg Loss: 0.18133831, Log Avg loss: 0.19526638, Global Avg Loss: 0.84033981, Time: 0.0209 Steps: 70330, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000083, Sample Num: 1328, Cur Loss: 0.19776541, Cur Avg Loss: 0.18080929, Log Avg loss: 0.17694739, Global Avg Loss: 0.84024549, Time: 0.0209 Steps: 70340, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000093, Sample Num: 1488, Cur Loss: 0.56584692, Cur Avg Loss: 0.18530067, Log Avg loss: 0.22257916, Global Avg Loss: 0.84015770, Time: 0.0209 Steps: 70350, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000103, Sample Num: 1648, Cur Loss: 0.35933164, Cur Avg Loss: 0.19228199, Log Avg loss: 0.25720828, Global Avg Loss: 0.84007484, Time: 0.0209 Steps: 70360, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000113, Sample Num: 1808, Cur Loss: 0.34077489, Cur Avg Loss: 0.20472006, Log Avg loss: 0.33283212, Global Avg Loss: 0.84000276, Time: 0.0209 Steps: 70370, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000123, Sample Num: 1968, Cur Loss: 0.61636043, Cur Avg Loss: 0.21012862, Log Avg loss: 0.27124538, Global Avg Loss: 0.83992195, Time: 0.0210 Steps: 70380, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000133, Sample Num: 2128, Cur Loss: 0.31927735, Cur Avg Loss: 0.20469155, Log Avg loss: 0.13781563, Global Avg Loss: 0.83982220, Time: 0.0212 Steps: 70390, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000143, Sample Num: 2288, Cur Loss: 0.40918040, Cur Avg Loss: 0.20482192, Log Avg loss: 0.20655576, Global Avg Loss: 0.83973225, Time: 0.0211 Steps: 70400, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000153, Sample Num: 2448, Cur Loss: 0.10359949, Cur Avg Loss: 0.20369472, Log Avg loss: 0.18757582, Global Avg Loss: 0.83963963, Time: 0.0211 Steps: 70410, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000163, Sample Num: 2608, Cur Loss: 0.57101488, Cur Avg Loss: 0.20753408, Log Avg loss: 0.26627626, Global Avg Loss: 0.83955821, Time: 0.0211 Steps: 70420, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000173, Sample Num: 2768, Cur Loss: 0.04993201, Cur Avg Loss: 0.20274114, Log Avg loss: 0.12461617, Global Avg Loss: 0.83945670, Time: 0.0210 Steps: 70430, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000183, Sample Num: 2928, Cur Loss: 0.12173486, Cur Avg Loss: 0.20334358, Log Avg loss: 0.21376580, Global Avg Loss: 0.83936787, Time: 0.0211 Steps: 70440, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000193, Sample Num: 3088, Cur Loss: 0.29176289, Cur Avg Loss: 0.20598686, Log Avg loss: 0.25435895, Global Avg Loss: 0.83928483, Time: 0.0211 Steps: 70450, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000203, Sample Num: 3248, Cur Loss: 0.17948520, Cur Avg Loss: 0.20532633, Log Avg loss: 0.19257813, Global Avg Loss: 0.83919305, Time: 0.0211 Steps: 70460, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000213, Sample Num: 3408, Cur Loss: 0.28979856, Cur Avg Loss: 0.20991858, Log Avg loss: 0.30314119, Global Avg Loss: 0.83911698, Time: 0.0211 Steps: 70470, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000223, Sample Num: 3568, Cur Loss: 0.34640878, Cur Avg Loss: 0.20964751, Log Avg loss: 0.20387375, Global Avg Loss: 0.83902685, Time: 0.0210 Steps: 70480, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000233, Sample Num: 3728, Cur Loss: 0.18990305, Cur Avg Loss: 0.20917099, Log Avg loss: 0.19854468, Global Avg Loss: 0.83893599, Time: 0.0210 Steps: 70490, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000243, Sample Num: 3888, Cur Loss: 0.07763294, Cur Avg Loss: 0.20870050, Log Avg loss: 0.19773789, Global Avg Loss: 0.83884504, Time: 0.0210 Steps: 70500, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000253, Sample Num: 4048, Cur Loss: 0.15329568, Cur Avg Loss: 0.20845530, Log Avg loss: 0.20249696, Global Avg Loss: 0.83875479, Time: 0.0210 Steps: 70510, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000263, Sample Num: 4208, Cur Loss: 0.35013753, Cur Avg Loss: 0.21379857, Log Avg loss: 0.34898334, Global Avg Loss: 0.83868534, Time: 0.0231 Steps: 70520, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000273, Sample Num: 4368, Cur Loss: 0.18418244, Cur Avg Loss: 0.21859682, Log Avg loss: 0.34479077, Global Avg Loss: 0.83861531, Time: 0.0210 Steps: 70530, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000283, Sample Num: 4528, Cur Loss: 0.12959303, Cur Avg Loss: 0.21851105, Log Avg loss: 0.21616961, Global Avg Loss: 0.83852707, Time: 0.0210 Steps: 70540, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000293, Sample Num: 4688, Cur Loss: 0.16964024, Cur Avg Loss: 0.21819652, Log Avg loss: 0.20929536, Global Avg Loss: 0.83843788, Time: 0.0210 Steps: 70550, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000303, Sample Num: 4848, Cur Loss: 0.33099273, Cur Avg Loss: 0.21863250, Log Avg loss: 0.23140659, Global Avg Loss: 0.83835185, Time: 0.0210 Steps: 70560, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000313, Sample Num: 5008, Cur Loss: 0.61247230, Cur Avg Loss: 0.21844677, Log Avg loss: 0.21281930, Global Avg Loss: 0.83826321, Time: 0.0210 Steps: 70570, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000323, Sample Num: 5168, Cur Loss: 0.10952143, Cur Avg Loss: 0.21827096, Log Avg loss: 0.21276788, Global Avg Loss: 0.83817459, Time: 0.0210 Steps: 70580, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000333, Sample Num: 5328, Cur Loss: 0.10364006, Cur Avg Loss: 0.21523659, Log Avg loss: 0.11722647, Global Avg Loss: 0.83807246, Time: 0.0210 Steps: 70590, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000343, Sample Num: 5488, Cur Loss: 0.36167502, Cur Avg Loss: 0.21665689, Log Avg loss: 0.26395283, Global Avg Loss: 0.83799114, Time: 0.0210 Steps: 70600, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000353, Sample Num: 5648, Cur Loss: 0.22300841, Cur Avg Loss: 0.21605972, Log Avg loss: 0.19557711, Global Avg Loss: 0.83790016, Time: 0.0210 Steps: 70610, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000363, Sample Num: 5808, Cur Loss: 0.28139514, Cur Avg Loss: 0.21567617, Log Avg loss: 0.20213677, Global Avg Loss: 0.83781013, Time: 0.0211 Steps: 70620, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000373, Sample Num: 5968, Cur Loss: 0.39191169, Cur Avg Loss: 0.21431922, Log Avg loss: 0.16506173, Global Avg Loss: 0.83771488, Time: 0.0210 Steps: 70630, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000383, Sample Num: 6128, Cur Loss: 0.10611270, Cur Avg Loss: 0.21346462, Log Avg loss: 0.18158809, Global Avg Loss: 0.83762200, Time: 0.0210 Steps: 70640, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000393, Sample Num: 6288, Cur Loss: 0.29245242, Cur Avg Loss: 0.21308498, Log Avg loss: 0.19854500, Global Avg Loss: 0.83753154, Time: 0.0210 Steps: 70650, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000403, Sample Num: 6448, Cur Loss: 0.26511794, Cur Avg Loss: 0.21120237, Log Avg loss: 0.13721548, Global Avg Loss: 0.83743243, Time: 0.0211 Steps: 70660, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000413, Sample Num: 6608, Cur Loss: 0.17035645, Cur Avg Loss: 0.21134038, Log Avg loss: 0.21690232, Global Avg Loss: 0.83734462, Time: 0.0210 Steps: 70670, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000423, Sample Num: 6768, Cur Loss: 0.17514101, Cur Avg Loss: 0.21057758, Log Avg loss: 0.17907417, Global Avg Loss: 0.83725149, Time: 0.0211 Steps: 70680, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000433, Sample Num: 6928, Cur Loss: 0.29422456, Cur Avg Loss: 0.21090650, Log Avg loss: 0.22481962, Global Avg Loss: 0.83716485, Time: 0.0210 Steps: 70690, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000443, Sample Num: 7088, Cur Loss: 0.24551871, Cur Avg Loss: 0.21240882, Log Avg loss: 0.27745928, Global Avg Loss: 0.83708569, Time: 0.0211 Steps: 70700, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000453, Sample Num: 7248, Cur Loss: 0.37056571, Cur Avg Loss: 0.21308334, Log Avg loss: 0.24296472, Global Avg Loss: 0.83700166, Time: 0.0211 Steps: 70710, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000463, Sample Num: 7408, Cur Loss: 0.07945003, Cur Avg Loss: 0.21422146, Log Avg loss: 0.26577811, Global Avg Loss: 0.83692089, Time: 0.0211 Steps: 70720, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000473, Sample Num: 7568, Cur Loss: 0.11576796, Cur Avg Loss: 0.21525251, Log Avg loss: 0.26299011, Global Avg Loss: 0.83683975, Time: 0.0210 Steps: 70730, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000483, Sample Num: 7728, Cur Loss: 0.06364896, Cur Avg Loss: 0.21383820, Log Avg loss: 0.14694157, Global Avg Loss: 0.83674222, Time: 0.0210 Steps: 70740, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000493, Sample Num: 7888, Cur Loss: 0.27900422, Cur Avg Loss: 0.21406186, Log Avg loss: 0.22486436, Global Avg Loss: 0.83665574, Time: 0.0211 Steps: 70750, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000503, Sample Num: 8048, Cur Loss: 0.30086100, Cur Avg Loss: 0.21751139, Log Avg loss: 0.38757309, Global Avg Loss: 0.83659227, Time: 0.0210 Steps: 70760, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000513, Sample Num: 8208, Cur Loss: 0.18317223, Cur Avg Loss: 0.21801611, Log Avg loss: 0.24340394, Global Avg Loss: 0.83650845, Time: 0.0248 Steps: 70770, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000523, Sample Num: 8368, Cur Loss: 0.51512283, Cur Avg Loss: 0.21737556, Log Avg loss: 0.18451502, Global Avg Loss: 0.83641634, Time: 0.0210 Steps: 70780, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000533, Sample Num: 8528, Cur Loss: 0.17129341, Cur Avg Loss: 0.21659093, Log Avg loss: 0.17555497, Global Avg Loss: 0.83632298, Time: 0.0211 Steps: 70790, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000543, Sample Num: 8688, Cur Loss: 0.10819505, Cur Avg Loss: 0.21573876, Log Avg loss: 0.17031830, Global Avg Loss: 0.83622891, Time: 0.0210 Steps: 70800, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000553, Sample Num: 8848, Cur Loss: 0.22602466, Cur Avg Loss: 0.21593192, Log Avg loss: 0.22642054, Global Avg Loss: 0.83614279, Time: 0.0210 Steps: 70810, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000563, Sample Num: 9008, Cur Loss: 0.89860982, Cur Avg Loss: 0.21901842, Log Avg loss: 0.38970169, Global Avg Loss: 0.83607976, Time: 0.0210 Steps: 70820, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000573, Sample Num: 9168, Cur Loss: 0.09066907, Cur Avg Loss: 0.21964736, Log Avg loss: 0.25505631, Global Avg Loss: 0.83599773, Time: 0.0210 Steps: 70830, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000583, Sample Num: 9328, Cur Loss: 0.31269604, Cur Avg Loss: 0.22158973, Log Avg loss: 0.33288760, Global Avg Loss: 0.83592670, Time: 0.0210 Steps: 70840, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000593, Sample Num: 9488, Cur Loss: 0.34221447, Cur Avg Loss: 0.22182679, Log Avg loss: 0.23564774, Global Avg Loss: 0.83584198, Time: 0.0210 Steps: 70850, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000603, Sample Num: 9648, Cur Loss: 0.15961824, Cur Avg Loss: 0.22077326, Log Avg loss: 0.15829904, Global Avg Loss: 0.83574636, Time: 0.0210 Steps: 70860, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000613, Sample Num: 9808, Cur Loss: 0.10801896, Cur Avg Loss: 0.21995827, Log Avg loss: 0.17081417, Global Avg Loss: 0.83565254, Time: 0.0210 Steps: 70870, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000623, Sample Num: 9968, Cur Loss: 0.15599713, Cur Avg Loss: 0.21966449, Log Avg loss: 0.20165547, Global Avg Loss: 0.83556309, Time: 0.0210 Steps: 70880, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000633, Sample Num: 10128, Cur Loss: 0.11458525, Cur Avg Loss: 0.22082256, Log Avg loss: 0.29297056, Global Avg Loss: 0.83548655, Time: 0.0210 Steps: 70890, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000643, Sample Num: 10288, Cur Loss: 0.10094424, Cur Avg Loss: 0.22038111, Log Avg loss: 0.19243725, Global Avg Loss: 0.83539585, Time: 0.0210 Steps: 70900, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000653, Sample Num: 10448, Cur Loss: 0.15844244, Cur Avg Loss: 0.21944054, Log Avg loss: 0.15896195, Global Avg Loss: 0.83530046, Time: 0.0210 Steps: 70910, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000663, Sample Num: 10608, Cur Loss: 0.26249710, Cur Avg Loss: 0.21935981, Log Avg loss: 0.21408840, Global Avg Loss: 0.83521287, Time: 0.0210 Steps: 70920, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000673, Sample Num: 10768, Cur Loss: 0.05227455, Cur Avg Loss: 0.21918128, Log Avg loss: 0.20734465, Global Avg Loss: 0.83512435, Time: 0.0210 Steps: 70930, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000683, Sample Num: 10928, Cur Loss: 0.09574611, Cur Avg Loss: 0.21853200, Log Avg loss: 0.17483513, Global Avg Loss: 0.83503127, Time: 0.0210 Steps: 70940, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000693, Sample Num: 11088, Cur Loss: 0.12968312, Cur Avg Loss: 0.21826511, Log Avg loss: 0.20003692, Global Avg Loss: 0.83494177, Time: 0.0210 Steps: 70950, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000703, Sample Num: 11248, Cur Loss: 0.13551559, Cur Avg Loss: 0.21887082, Log Avg loss: 0.26084657, Global Avg Loss: 0.83486087, Time: 0.0211 Steps: 70960, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000713, Sample Num: 11408, Cur Loss: 0.72371024, Cur Avg Loss: 0.22052381, Log Avg loss: 0.33672876, Global Avg Loss: 0.83479068, Time: 0.0210 Steps: 70970, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000723, Sample Num: 11568, Cur Loss: 0.26029235, Cur Avg Loss: 0.22251508, Log Avg loss: 0.36449256, Global Avg Loss: 0.83472442, Time: 0.0209 Steps: 70980, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000733, Sample Num: 11728, Cur Loss: 0.16338459, Cur Avg Loss: 0.22442009, Log Avg loss: 0.36215246, Global Avg Loss: 0.83465785, Time: 0.0210 Steps: 70990, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000743, Sample Num: 11888, Cur Loss: 0.13909569, Cur Avg Loss: 0.22543348, Log Avg loss: 0.29971462, Global Avg Loss: 0.83458251, Time: 0.0210 Steps: 71000, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000753, Sample Num: 12048, Cur Loss: 0.06859591, Cur Avg Loss: 0.22524635, Log Avg loss: 0.21134310, Global Avg Loss: 0.83449474, Time: 0.0210 Steps: 71010, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000763, Sample Num: 12208, Cur Loss: 0.19539279, Cur Avg Loss: 0.22453657, Log Avg loss: 0.17108953, Global Avg Loss: 0.83440133, Time: 0.0210 Steps: 71020, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000773, Sample Num: 12368, Cur Loss: 0.09818508, Cur Avg Loss: 0.22351339, Log Avg loss: 0.14544547, Global Avg Loss: 0.83430433, Time: 0.0211 Steps: 71030, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000783, Sample Num: 12528, Cur Loss: 0.08008491, Cur Avg Loss: 0.22278267, Log Avg loss: 0.16629787, Global Avg Loss: 0.83421030, Time: 0.0210 Steps: 71040, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000793, Sample Num: 12688, Cur Loss: 0.24141702, Cur Avg Loss: 0.22277448, Log Avg loss: 0.22213302, Global Avg Loss: 0.83412415, Time: 0.0210 Steps: 71050, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000803, Sample Num: 12848, Cur Loss: 0.11030728, Cur Avg Loss: 0.22242580, Log Avg loss: 0.19477535, Global Avg Loss: 0.83403418, Time: 0.0210 Steps: 71060, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000813, Sample Num: 13008, Cur Loss: 0.12585296, Cur Avg Loss: 0.22371019, Log Avg loss: 0.32684642, Global Avg Loss: 0.83396282, Time: 0.0210 Steps: 71070, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000823, Sample Num: 13168, Cur Loss: 0.22908646, Cur Avg Loss: 0.22506628, Log Avg loss: 0.33531666, Global Avg Loss: 0.83389266, Time: 0.0210 Steps: 71080, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000833, Sample Num: 13328, Cur Loss: 0.06007037, Cur Avg Loss: 0.22457168, Log Avg loss: 0.18386646, Global Avg Loss: 0.83380123, Time: 0.0211 Steps: 71090, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000843, Sample Num: 13488, Cur Loss: 0.05980179, Cur Avg Loss: 0.22337713, Log Avg loss: 0.12387051, Global Avg Loss: 0.83370138, Time: 0.0210 Steps: 71100, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000853, Sample Num: 13648, Cur Loss: 0.16444126, Cur Avg Loss: 0.22335649, Log Avg loss: 0.22161712, Global Avg Loss: 0.83361530, Time: 0.0210 Steps: 71110, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000863, Sample Num: 13808, Cur Loss: 0.10355978, Cur Avg Loss: 0.22285190, Log Avg loss: 0.17981004, Global Avg Loss: 0.83352337, Time: 0.0210 Steps: 71120, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000873, Sample Num: 13968, Cur Loss: 0.16789545, Cur Avg Loss: 0.22227007, Log Avg loss: 0.17205836, Global Avg Loss: 0.83343038, Time: 0.0210 Steps: 71130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000883, Sample Num: 14128, Cur Loss: 0.25272083, Cur Avg Loss: 0.22239229, Log Avg loss: 0.23306234, Global Avg Loss: 0.83334598, Time: 0.0210 Steps: 71140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000893, Sample Num: 14288, Cur Loss: 0.27803648, Cur Avg Loss: 0.22239417, Log Avg loss: 0.22256012, Global Avg Loss: 0.83326014, Time: 0.0210 Steps: 71150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000903, Sample Num: 14448, Cur Loss: 0.30033445, Cur Avg Loss: 0.22296852, Log Avg loss: 0.27425792, Global Avg Loss: 0.83318158, Time: 0.0210 Steps: 71160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000913, Sample Num: 14608, Cur Loss: 0.41219875, Cur Avg Loss: 0.22340860, Log Avg loss: 0.26314741, Global Avg Loss: 0.83310149, Time: 0.0210 Steps: 71170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000923, Sample Num: 14768, Cur Loss: 0.21626525, Cur Avg Loss: 0.22266440, Log Avg loss: 0.15471920, Global Avg Loss: 0.83300618, Time: 0.0210 Steps: 71180, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000933, Sample Num: 14928, Cur Loss: 0.20096540, Cur Avg Loss: 0.22367399, Log Avg loss: 0.31685893, Global Avg Loss: 0.83293368, Time: 0.0210 Steps: 71190, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000943, Sample Num: 15088, Cur Loss: 0.08423063, Cur Avg Loss: 0.22392828, Log Avg loss: 0.24765364, Global Avg Loss: 0.83285148, Time: 0.0210 Steps: 71200, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000953, Sample Num: 15248, Cur Loss: 0.07127318, Cur Avg Loss: 0.22365033, Log Avg loss: 0.19743947, Global Avg Loss: 0.83276225, Time: 0.0210 Steps: 71210, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000963, Sample Num: 15408, Cur Loss: 0.09981509, Cur Avg Loss: 0.22308814, Log Avg loss: 0.16951149, Global Avg Loss: 0.83266912, Time: 0.0210 Steps: 71220, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000973, Sample Num: 15568, Cur Loss: 0.12413601, Cur Avg Loss: 0.22265881, Log Avg loss: 0.18131481, Global Avg Loss: 0.83257768, Time: 0.0210 Steps: 71230, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000983, Sample Num: 15728, Cur Loss: 0.17810699, Cur Avg Loss: 0.22237479, Log Avg loss: 0.19473937, Global Avg Loss: 0.83248814, Time: 0.0211 Steps: 71240, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000993, Sample Num: 15888, Cur Loss: 0.54356706, Cur Avg Loss: 0.22286610, Log Avg loss: 0.27116221, Global Avg Loss: 0.83240936, Time: 0.0211 Steps: 71250, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001003, Sample Num: 16048, Cur Loss: 0.19829549, Cur Avg Loss: 0.22335692, Log Avg loss: 0.27209502, Global Avg Loss: 0.83233073, Time: 0.0210 Steps: 71260, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001013, Sample Num: 16208, Cur Loss: 0.26972261, Cur Avg Loss: 0.22326490, Log Avg loss: 0.21403524, Global Avg Loss: 0.83224398, Time: 0.0211 Steps: 71270, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001023, Sample Num: 16368, Cur Loss: 0.12849218, Cur Avg Loss: 0.22278730, Log Avg loss: 0.17440645, Global Avg Loss: 0.83215169, Time: 0.0211 Steps: 71280, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001033, Sample Num: 16528, Cur Loss: 0.16502275, Cur Avg Loss: 0.22298436, Log Avg loss: 0.24314400, Global Avg Loss: 0.83206907, Time: 0.0211 Steps: 71290, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001043, Sample Num: 16688, Cur Loss: 0.26366344, Cur Avg Loss: 0.22278489, Log Avg loss: 0.20217929, Global Avg Loss: 0.83198072, Time: 0.0210 Steps: 71300, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001053, Sample Num: 16848, Cur Loss: 0.07532500, Cur Avg Loss: 0.22272721, Log Avg loss: 0.21671143, Global Avg Loss: 0.83189444, Time: 0.0211 Steps: 71310, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001063, Sample Num: 17008, Cur Loss: 0.10147600, Cur Avg Loss: 0.22265067, Log Avg loss: 0.21459044, Global Avg Loss: 0.83180789, Time: 0.0210 Steps: 71320, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001073, Sample Num: 17168, Cur Loss: 0.33153695, Cur Avg Loss: 0.22271586, Log Avg loss: 0.22964606, Global Avg Loss: 0.83172347, Time: 0.0210 Steps: 71330, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001083, Sample Num: 17328, Cur Loss: 0.28276369, Cur Avg Loss: 0.22341201, Log Avg loss: 0.29810843, Global Avg Loss: 0.83164867, Time: 0.0210 Steps: 71340, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001093, Sample Num: 17488, Cur Loss: 0.18556340, Cur Avg Loss: 0.22407301, Log Avg loss: 0.29565932, Global Avg Loss: 0.83157355, Time: 0.0210 Steps: 71350, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001103, Sample Num: 17648, Cur Loss: 0.13588835, Cur Avg Loss: 0.22353854, Log Avg loss: 0.16512146, Global Avg Loss: 0.83148016, Time: 0.0210 Steps: 71360, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001113, Sample Num: 17808, Cur Loss: 0.49717864, Cur Avg Loss: 0.22312690, Log Avg loss: 0.17772274, Global Avg Loss: 0.83138855, Time: 0.0211 Steps: 71370, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001123, Sample Num: 17968, Cur Loss: 0.17390826, Cur Avg Loss: 0.22297221, Log Avg loss: 0.20575534, Global Avg Loss: 0.83130091, Time: 0.0210 Steps: 71380, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001133, Sample Num: 18128, Cur Loss: 0.10199666, Cur Avg Loss: 0.22248006, Log Avg loss: 0.16721136, Global Avg Loss: 0.83120788, Time: 0.0210 Steps: 71390, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001143, Sample Num: 18288, Cur Loss: 0.37407595, Cur Avg Loss: 0.22333594, Log Avg loss: 0.32030693, Global Avg Loss: 0.83113633, Time: 0.0210 Steps: 71400, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001153, Sample Num: 18448, Cur Loss: 0.19097105, Cur Avg Loss: 0.22386303, Log Avg loss: 0.28410925, Global Avg Loss: 0.83105973, Time: 0.0211 Steps: 71410, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001163, Sample Num: 18608, Cur Loss: 0.08013503, Cur Avg Loss: 0.22383438, Log Avg loss: 0.22053206, Global Avg Loss: 0.83097424, Time: 0.0210 Steps: 71420, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001173, Sample Num: 18768, Cur Loss: 0.32629937, Cur Avg Loss: 0.22400094, Log Avg loss: 0.24337090, Global Avg Loss: 0.83089198, Time: 0.0210 Steps: 71430, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001183, Sample Num: 18928, Cur Loss: 0.15250333, Cur Avg Loss: 0.22364892, Log Avg loss: 0.18235775, Global Avg Loss: 0.83080120, Time: 0.0210 Steps: 71440, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001193, Sample Num: 19088, Cur Loss: 0.33066449, Cur Avg Loss: 0.22352436, Log Avg loss: 0.20878804, Global Avg Loss: 0.83071414, Time: 0.0210 Steps: 71450, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001203, Sample Num: 19248, Cur Loss: 0.39495045, Cur Avg Loss: 0.22383030, Log Avg loss: 0.26032904, Global Avg Loss: 0.83063432, Time: 0.0210 Steps: 71460, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001213, Sample Num: 19408, Cur Loss: 0.19758028, Cur Avg Loss: 0.22338639, Log Avg loss: 0.16998494, Global Avg Loss: 0.83054189, Time: 0.0210 Steps: 71470, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001223, Sample Num: 19568, Cur Loss: 0.24439354, Cur Avg Loss: 0.22296806, Log Avg loss: 0.17222357, Global Avg Loss: 0.83044979, Time: 0.0210 Steps: 71480, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001233, Sample Num: 19728, Cur Loss: 0.15809840, Cur Avg Loss: 0.22279080, Log Avg loss: 0.20111202, Global Avg Loss: 0.83036176, Time: 0.0212 Steps: 71490, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001243, Sample Num: 19888, Cur Loss: 0.19220017, Cur Avg Loss: 0.22369333, Log Avg loss: 0.33497606, Global Avg Loss: 0.83029247, Time: 0.0211 Steps: 71500, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001253, Sample Num: 20048, Cur Loss: 0.11588483, Cur Avg Loss: 0.22319407, Log Avg loss: 0.16113566, Global Avg Loss: 0.83019890, Time: 0.0211 Steps: 71510, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001263, Sample Num: 20208, Cur Loss: 0.21449231, Cur Avg Loss: 0.22342809, Log Avg loss: 0.25275128, Global Avg Loss: 0.83011816, Time: 0.0210 Steps: 71520, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001273, Sample Num: 20368, Cur Loss: 0.10906988, Cur Avg Loss: 0.22342491, Log Avg loss: 0.22302261, Global Avg Loss: 0.83003328, Time: 0.0210 Steps: 71530, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001283, Sample Num: 20528, Cur Loss: 0.10877243, Cur Avg Loss: 0.22408571, Log Avg loss: 0.30820620, Global Avg Loss: 0.82996034, Time: 0.0249 Steps: 71540, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001293, Sample Num: 20688, Cur Loss: 0.38750502, Cur Avg Loss: 0.22656871, Log Avg loss: 0.54513697, Global Avg Loss: 0.82992053, Time: 0.0212 Steps: 71550, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001303, Sample Num: 20848, Cur Loss: 0.11868837, Cur Avg Loss: 0.22700429, Log Avg loss: 0.28332442, Global Avg Loss: 0.82984415, Time: 0.0211 Steps: 71560, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001313, Sample Num: 21008, Cur Loss: 0.10515388, Cur Avg Loss: 0.22741509, Log Avg loss: 0.28094280, Global Avg Loss: 0.82976746, Time: 0.0211 Steps: 71570, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001323, Sample Num: 21168, Cur Loss: 0.31061199, Cur Avg Loss: 0.22717093, Log Avg loss: 0.19511347, Global Avg Loss: 0.82967879, Time: 0.0211 Steps: 71580, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001333, Sample Num: 21328, Cur Loss: 0.10455339, Cur Avg Loss: 0.22721909, Log Avg loss: 0.23359000, Global Avg Loss: 0.82959553, Time: 0.0211 Steps: 71590, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001343, Sample Num: 21488, Cur Loss: 0.27267981, Cur Avg Loss: 0.22719074, Log Avg loss: 0.22341111, Global Avg Loss: 0.82951087, Time: 0.0211 Steps: 71600, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001353, Sample Num: 21648, Cur Loss: 0.18727836, Cur Avg Loss: 0.22718159, Log Avg loss: 0.22595306, Global Avg Loss: 0.82942658, Time: 0.0211 Steps: 71610, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001363, Sample Num: 21808, Cur Loss: 0.11034110, Cur Avg Loss: 0.22690541, Log Avg loss: 0.18953894, Global Avg Loss: 0.82933724, Time: 0.0211 Steps: 71620, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001373, Sample Num: 21968, Cur Loss: 0.19917166, Cur Avg Loss: 0.22684734, Log Avg loss: 0.21893211, Global Avg Loss: 0.82925202, Time: 0.0212 Steps: 71630, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001383, Sample Num: 22128, Cur Loss: 0.22904688, Cur Avg Loss: 0.22705349, Log Avg loss: 0.25535745, Global Avg Loss: 0.82917191, Time: 0.0211 Steps: 71640, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001393, Sample Num: 22288, Cur Loss: 0.23262450, Cur Avg Loss: 0.22681527, Log Avg loss: 0.19386996, Global Avg Loss: 0.82908325, Time: 0.0211 Steps: 71650, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001403, Sample Num: 22448, Cur Loss: 0.39425561, Cur Avg Loss: 0.22699840, Log Avg loss: 0.25250784, Global Avg Loss: 0.82900279, Time: 0.0211 Steps: 71660, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001413, Sample Num: 22608, Cur Loss: 0.12916158, Cur Avg Loss: 0.22721524, Log Avg loss: 0.25763764, Global Avg Loss: 0.82892306, Time: 0.0210 Steps: 71670, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001423, Sample Num: 22768, Cur Loss: 0.33465979, Cur Avg Loss: 0.22665656, Log Avg loss: 0.14771582, Global Avg Loss: 0.82882803, Time: 0.0211 Steps: 71680, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001433, Sample Num: 22928, Cur Loss: 0.24581625, Cur Avg Loss: 0.22704887, Log Avg loss: 0.28287467, Global Avg Loss: 0.82875188, Time: 0.0211 Steps: 71690, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001443, Sample Num: 23088, Cur Loss: 0.05664296, Cur Avg Loss: 0.22743653, Log Avg loss: 0.28298852, Global Avg Loss: 0.82867576, Time: 0.0211 Steps: 71700, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001453, Sample Num: 23248, Cur Loss: 0.06008348, Cur Avg Loss: 0.22788545, Log Avg loss: 0.29266340, Global Avg Loss: 0.82860101, Time: 0.0211 Steps: 71710, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001463, Sample Num: 23408, Cur Loss: 0.11273378, Cur Avg Loss: 0.22754142, Log Avg loss: 0.17755421, Global Avg Loss: 0.82851023, Time: 0.0211 Steps: 71720, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001473, Sample Num: 23568, Cur Loss: 0.10450287, Cur Avg Loss: 0.22712144, Log Avg loss: 0.16567840, Global Avg Loss: 0.82841783, Time: 0.0211 Steps: 71730, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001483, Sample Num: 23728, Cur Loss: 0.13570789, Cur Avg Loss: 0.22639589, Log Avg loss: 0.11952236, Global Avg Loss: 0.82831901, Time: 0.0211 Steps: 71740, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001493, Sample Num: 23888, Cur Loss: 0.15984233, Cur Avg Loss: 0.22659057, Log Avg loss: 0.25546262, Global Avg Loss: 0.82823917, Time: 0.0211 Steps: 71750, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001503, Sample Num: 24048, Cur Loss: 0.33906561, Cur Avg Loss: 0.22658457, Log Avg loss: 0.22568743, Global Avg Loss: 0.82815521, Time: 0.0212 Steps: 71760, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001513, Sample Num: 24208, Cur Loss: 0.58551961, Cur Avg Loss: 0.22697534, Log Avg loss: 0.28570858, Global Avg Loss: 0.82807962, Time: 0.0211 Steps: 71770, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001523, Sample Num: 24368, Cur Loss: 0.05461710, Cur Avg Loss: 0.22662758, Log Avg loss: 0.17401153, Global Avg Loss: 0.82798850, Time: 0.0211 Steps: 71780, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001533, Sample Num: 24528, Cur Loss: 0.50539702, Cur Avg Loss: 0.22638516, Log Avg loss: 0.18946454, Global Avg Loss: 0.82789956, Time: 0.0211 Steps: 71790, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001543, Sample Num: 24688, Cur Loss: 0.29046556, Cur Avg Loss: 0.22600530, Log Avg loss: 0.16777298, Global Avg Loss: 0.82780762, Time: 0.0214 Steps: 71800, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001553, Sample Num: 24848, Cur Loss: 0.29068989, Cur Avg Loss: 0.22611497, Log Avg loss: 0.24303731, Global Avg Loss: 0.82772619, Time: 0.0210 Steps: 71810, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001563, Sample Num: 25008, Cur Loss: 0.30162302, Cur Avg Loss: 0.22602468, Log Avg loss: 0.21200259, Global Avg Loss: 0.82764046, Time: 0.0211 Steps: 71820, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001573, Sample Num: 25168, Cur Loss: 0.17698438, Cur Avg Loss: 0.22613519, Log Avg loss: 0.24340691, Global Avg Loss: 0.82755912, Time: 0.0211 Steps: 71830, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001583, Sample Num: 25328, Cur Loss: 0.33651170, Cur Avg Loss: 0.22629378, Log Avg loss: 0.25124009, Global Avg Loss: 0.82747890, Time: 0.0211 Steps: 71840, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001593, Sample Num: 25488, Cur Loss: 0.16082622, Cur Avg Loss: 0.22615488, Log Avg loss: 0.20416685, Global Avg Loss: 0.82739215, Time: 0.0211 Steps: 71850, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001603, Sample Num: 25648, Cur Loss: 0.08587547, Cur Avg Loss: 0.22599107, Log Avg loss: 0.19989717, Global Avg Loss: 0.82730482, Time: 0.0212 Steps: 71860, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001613, Sample Num: 25808, Cur Loss: 0.04194402, Cur Avg Loss: 0.22568677, Log Avg loss: 0.17690646, Global Avg Loss: 0.82721433, Time: 0.0211 Steps: 71870, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001623, Sample Num: 25968, Cur Loss: 0.24993971, Cur Avg Loss: 0.22537168, Log Avg loss: 0.17454774, Global Avg Loss: 0.82712353, Time: 0.0212 Steps: 71880, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001633, Sample Num: 26128, Cur Loss: 0.39954096, Cur Avg Loss: 0.22545341, Log Avg loss: 0.23871890, Global Avg Loss: 0.82704168, Time: 0.0211 Steps: 71890, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001643, Sample Num: 26288, Cur Loss: 0.15064472, Cur Avg Loss: 0.22542841, Log Avg loss: 0.22134667, Global Avg Loss: 0.82695744, Time: 0.0210 Steps: 71900, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001653, Sample Num: 26448, Cur Loss: 0.22791006, Cur Avg Loss: 0.22513198, Log Avg loss: 0.17642815, Global Avg Loss: 0.82686697, Time: 0.0211 Steps: 71910, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001663, Sample Num: 26608, Cur Loss: 0.27452457, Cur Avg Loss: 0.22492944, Log Avg loss: 0.19144883, Global Avg Loss: 0.82677862, Time: 0.0210 Steps: 71920, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001673, Sample Num: 26768, Cur Loss: 0.06887145, Cur Avg Loss: 0.22453148, Log Avg loss: 0.15835128, Global Avg Loss: 0.82668570, Time: 0.0210 Steps: 71930, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001683, Sample Num: 26928, Cur Loss: 0.04259571, Cur Avg Loss: 0.22415362, Log Avg loss: 0.16093790, Global Avg Loss: 0.82659315, Time: 0.0211 Steps: 71940, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001693, Sample Num: 27088, Cur Loss: 0.34336960, Cur Avg Loss: 0.22399087, Log Avg loss: 0.19659966, Global Avg Loss: 0.82650559, Time: 0.0211 Steps: 71950, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001703, Sample Num: 27248, Cur Loss: 0.67040920, Cur Avg Loss: 0.22424350, Log Avg loss: 0.26701291, Global Avg Loss: 0.82642784, Time: 0.0210 Steps: 71960, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001713, Sample Num: 27408, Cur Loss: 0.12427889, Cur Avg Loss: 0.22434524, Log Avg loss: 0.24167233, Global Avg Loss: 0.82634659, Time: 0.0210 Steps: 71970, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001723, Sample Num: 27568, Cur Loss: 0.12304940, Cur Avg Loss: 0.22422833, Log Avg loss: 0.20420097, Global Avg Loss: 0.82626016, Time: 0.0211 Steps: 71980, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001733, Sample Num: 27728, Cur Loss: 0.75820500, Cur Avg Loss: 0.22481863, Log Avg loss: 0.32652752, Global Avg Loss: 0.82619074, Time: 0.0211 Steps: 71990, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001743, Sample Num: 27888, Cur Loss: 0.12516528, Cur Avg Loss: 0.22444958, Log Avg loss: 0.16049362, Global Avg Loss: 0.82609829, Time: 0.0210 Steps: 72000, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001753, Sample Num: 28048, Cur Loss: 0.27570325, Cur Avg Loss: 0.22435784, Log Avg loss: 0.20836778, Global Avg Loss: 0.82601250, Time: 0.0211 Steps: 72010, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001763, Sample Num: 28208, Cur Loss: 0.16775557, Cur Avg Loss: 0.22435051, Log Avg loss: 0.22306610, Global Avg Loss: 0.82592878, Time: 0.0211 Steps: 72020, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001773, Sample Num: 28368, Cur Loss: 0.09100786, Cur Avg Loss: 0.22429188, Log Avg loss: 0.21395390, Global Avg Loss: 0.82584382, Time: 0.0210 Steps: 72030, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001783, Sample Num: 28528, Cur Loss: 0.17550251, Cur Avg Loss: 0.22427805, Log Avg loss: 0.22182590, Global Avg Loss: 0.82575998, Time: 0.0211 Steps: 72040, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001793, Sample Num: 28688, Cur Loss: 0.46625388, Cur Avg Loss: 0.22478826, Log Avg loss: 0.31575875, Global Avg Loss: 0.82568919, Time: 0.0247 Steps: 72050, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001803, Sample Num: 28848, Cur Loss: 0.29357767, Cur Avg Loss: 0.22494367, Log Avg loss: 0.25280968, Global Avg Loss: 0.82560969, Time: 0.0210 Steps: 72060, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001813, Sample Num: 29008, Cur Loss: 0.32617345, Cur Avg Loss: 0.22532617, Log Avg loss: 0.29428994, Global Avg Loss: 0.82553597, Time: 0.0211 Steps: 72070, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001823, Sample Num: 29168, Cur Loss: 0.17394084, Cur Avg Loss: 0.22528942, Log Avg loss: 0.21862789, Global Avg Loss: 0.82545177, Time: 0.0210 Steps: 72080, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001833, Sample Num: 29328, Cur Loss: 0.44356367, Cur Avg Loss: 0.22543839, Log Avg loss: 0.25259503, Global Avg Loss: 0.82537231, Time: 0.0210 Steps: 72090, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001843, Sample Num: 29488, Cur Loss: 0.29618108, Cur Avg Loss: 0.22506289, Log Avg loss: 0.15623459, Global Avg Loss: 0.82527950, Time: 0.0210 Steps: 72100, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001853, Sample Num: 29648, Cur Loss: 0.20312780, Cur Avg Loss: 0.22534745, Log Avg loss: 0.27779135, Global Avg Loss: 0.82520357, Time: 0.0211 Steps: 72110, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001863, Sample Num: 29808, Cur Loss: 0.20052615, Cur Avg Loss: 0.22548377, Log Avg loss: 0.25074338, Global Avg Loss: 0.82512392, Time: 0.0210 Steps: 72120, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001873, Sample Num: 29968, Cur Loss: 0.08725847, Cur Avg Loss: 0.22541376, Log Avg loss: 0.21237087, Global Avg Loss: 0.82503897, Time: 0.0210 Steps: 72130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001883, Sample Num: 30128, Cur Loss: 0.14708742, Cur Avg Loss: 0.22589686, Log Avg loss: 0.31638181, Global Avg Loss: 0.82496846, Time: 0.0210 Steps: 72140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001893, Sample Num: 30288, Cur Loss: 0.21759349, Cur Avg Loss: 0.22574908, Log Avg loss: 0.19792261, Global Avg Loss: 0.82488155, Time: 0.0210 Steps: 72150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001903, Sample Num: 30448, Cur Loss: 0.17549068, Cur Avg Loss: 0.22530671, Log Avg loss: 0.14156581, Global Avg Loss: 0.82478686, Time: 0.0211 Steps: 72160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001913, Sample Num: 30608, Cur Loss: 0.11150065, Cur Avg Loss: 0.22564138, Log Avg loss: 0.28932978, Global Avg Loss: 0.82471266, Time: 0.0211 Steps: 72170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001923, Sample Num: 30768, Cur Loss: 0.20340526, Cur Avg Loss: 0.22546648, Log Avg loss: 0.19200776, Global Avg Loss: 0.82462501, Time: 0.0210 Steps: 72180, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001933, Sample Num: 30928, Cur Loss: 0.12922242, Cur Avg Loss: 0.22588319, Log Avg loss: 0.30601583, Global Avg Loss: 0.82455317, Time: 0.0210 Steps: 72190, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001943, Sample Num: 31088, Cur Loss: 0.04726063, Cur Avg Loss: 0.22615062, Log Avg loss: 0.27784492, Global Avg Loss: 0.82447745, Time: 0.0210 Steps: 72200, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001953, Sample Num: 31248, Cur Loss: 0.49614415, Cur Avg Loss: 0.22587570, Log Avg loss: 0.17245798, Global Avg Loss: 0.82438715, Time: 0.0210 Steps: 72210, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001963, Sample Num: 31408, Cur Loss: 0.11526170, Cur Avg Loss: 0.22546856, Log Avg loss: 0.14595511, Global Avg Loss: 0.82429321, Time: 0.0210 Steps: 72220, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001973, Sample Num: 31568, Cur Loss: 0.38764310, Cur Avg Loss: 0.22561348, Log Avg loss: 0.25406092, Global Avg Loss: 0.82421426, Time: 0.0211 Steps: 72230, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001983, Sample Num: 31728, Cur Loss: 0.11247525, Cur Avg Loss: 0.22573834, Log Avg loss: 0.25037418, Global Avg Loss: 0.82413483, Time: 0.0210 Steps: 72240, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001993, Sample Num: 31888, Cur Loss: 0.22710963, Cur Avg Loss: 0.22600689, Log Avg loss: 0.27925927, Global Avg Loss: 0.82405941, Time: 0.0211 Steps: 72250, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002003, Sample Num: 32048, Cur Loss: 0.58969164, Cur Avg Loss: 0.22599918, Log Avg loss: 0.22446370, Global Avg Loss: 0.82397644, Time: 0.0210 Steps: 72260, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002013, Sample Num: 32208, Cur Loss: 0.08452511, Cur Avg Loss: 0.22588083, Log Avg loss: 0.20217538, Global Avg Loss: 0.82389040, Time: 0.0210 Steps: 72270, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002023, Sample Num: 32368, Cur Loss: 0.15426856, Cur Avg Loss: 0.22564612, Log Avg loss: 0.17839874, Global Avg Loss: 0.82380109, Time: 0.0210 Steps: 72280, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002033, Sample Num: 32528, Cur Loss: 0.36909112, Cur Avg Loss: 0.22586244, Log Avg loss: 0.26962254, Global Avg Loss: 0.82372443, Time: 0.0211 Steps: 72290, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002043, Sample Num: 32688, Cur Loss: 0.41234660, Cur Avg Loss: 0.22597578, Log Avg loss: 0.24901780, Global Avg Loss: 0.82364494, Time: 0.0210 Steps: 72300, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002053, Sample Num: 32848, Cur Loss: 0.13305619, Cur Avg Loss: 0.22587471, Log Avg loss: 0.20522740, Global Avg Loss: 0.82355942, Time: 0.0248 Steps: 72310, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002063, Sample Num: 33008, Cur Loss: 0.33857986, Cur Avg Loss: 0.22554142, Log Avg loss: 0.15711666, Global Avg Loss: 0.82346727, Time: 0.0211 Steps: 72320, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002073, Sample Num: 33168, Cur Loss: 0.23678866, Cur Avg Loss: 0.22541322, Log Avg loss: 0.19896466, Global Avg Loss: 0.82338093, Time: 0.0211 Steps: 72330, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002083, Sample Num: 33328, Cur Loss: 0.21935120, Cur Avg Loss: 0.22560484, Log Avg loss: 0.26532913, Global Avg Loss: 0.82330379, Time: 0.0210 Steps: 72340, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002093, Sample Num: 33488, Cur Loss: 0.33952570, Cur Avg Loss: 0.22580149, Log Avg loss: 0.26676198, Global Avg Loss: 0.82322686, Time: 0.0210 Steps: 72350, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002103, Sample Num: 33648, Cur Loss: 0.18921298, Cur Avg Loss: 0.22539382, Log Avg loss: 0.14006909, Global Avg Loss: 0.82313245, Time: 0.0211 Steps: 72360, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002113, Sample Num: 33808, Cur Loss: 0.27806029, Cur Avg Loss: 0.22517199, Log Avg loss: 0.17852162, Global Avg Loss: 0.82304338, Time: 0.0211 Steps: 72370, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002123, Sample Num: 33968, Cur Loss: 0.17404762, Cur Avg Loss: 0.22488753, Log Avg loss: 0.16478113, Global Avg Loss: 0.82295243, Time: 0.0211 Steps: 72380, Updated lr: 0.000032 ***** Running evaluation checkpoint-72386 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-72386 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.021755, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.278421, "eval_total_loss": 195.729923, "eval_mae": 0.407763, "eval_mse": 0.278452, "eval_r2": 0.822998, "eval_sp_statistic": 0.896042, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.920376, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.317613, "test_total_loss": 159.44193, "test_mae": 0.363585, "test_mse": 0.317727, "test_r2": 0.794936, "test_sp_statistic": 0.871229, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.920755, "test_ps_pvalue": 0.0, "lr": 3.2303461356092936e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.822905027336776, "train_cur_epoch_loss": 478.9423802420497, "train_cur_epoch_avg_loss": 0.22496119316207125, "train_cur_epoch_time": 45.02175498008728, "train_cur_epoch_avg_time": 0.02114690229219694, "epoch": 34, "step": 72386} ################################################## Training, Epoch: 0035, Batch: 000004, Sample Num: 64, Cur Loss: 0.23139450, Cur Avg Loss: 0.35238644, Log Avg loss: 0.29156989, Global Avg Loss: 0.82287903, Time: 0.0248 Steps: 72390, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000014, Sample Num: 224, Cur Loss: 0.16100766, Cur Avg Loss: 0.19539365, Log Avg loss: 0.13259653, Global Avg Loss: 0.82278369, Time: 0.0210 Steps: 72400, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000024, Sample Num: 384, Cur Loss: 0.02340014, Cur Avg Loss: 0.15835012, Log Avg loss: 0.10648917, Global Avg Loss: 0.82268476, Time: 0.0211 Steps: 72410, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000034, Sample Num: 544, Cur Loss: 0.18903090, Cur Avg Loss: 0.17138124, Log Avg loss: 0.20265594, Global Avg Loss: 0.82259915, Time: 0.0210 Steps: 72420, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000044, Sample Num: 704, Cur Loss: 0.13631327, Cur Avg Loss: 0.17866051, Log Avg loss: 0.20341003, Global Avg Loss: 0.82251366, Time: 0.0210 Steps: 72430, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000054, Sample Num: 864, Cur Loss: 0.11811713, Cur Avg Loss: 0.19413390, Log Avg loss: 0.26221679, Global Avg Loss: 0.82243631, Time: 0.0210 Steps: 72440, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000064, Sample Num: 1024, Cur Loss: 0.11375728, Cur Avg Loss: 0.22354868, Log Avg loss: 0.38238852, Global Avg Loss: 0.82237558, Time: 0.0211 Steps: 72450, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000074, Sample Num: 1184, Cur Loss: 0.25385958, Cur Avg Loss: 0.24633069, Log Avg loss: 0.39213557, Global Avg Loss: 0.82231620, Time: 0.0211 Steps: 72460, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000084, Sample Num: 1344, Cur Loss: 0.46386611, Cur Avg Loss: 0.24588447, Log Avg loss: 0.24258241, Global Avg Loss: 0.82223620, Time: 0.0210 Steps: 72470, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000094, Sample Num: 1504, Cur Loss: 0.29069808, Cur Avg Loss: 0.23871277, Log Avg loss: 0.17847054, Global Avg Loss: 0.82214738, Time: 0.0211 Steps: 72480, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000104, Sample Num: 1664, Cur Loss: 0.22607848, Cur Avg Loss: 0.24906035, Log Avg loss: 0.34632760, Global Avg Loss: 0.82208174, Time: 0.0210 Steps: 72490, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000114, Sample Num: 1824, Cur Loss: 0.70161903, Cur Avg Loss: 0.24739268, Log Avg loss: 0.23004889, Global Avg Loss: 0.82200008, Time: 0.0210 Steps: 72500, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000124, Sample Num: 1984, Cur Loss: 0.21786402, Cur Avg Loss: 0.24097087, Log Avg loss: 0.16776218, Global Avg Loss: 0.82190986, Time: 0.0211 Steps: 72510, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000134, Sample Num: 2144, Cur Loss: 0.41597003, Cur Avg Loss: 0.23580690, Log Avg loss: 0.17177377, Global Avg Loss: 0.82182021, Time: 0.0211 Steps: 72520, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000144, Sample Num: 2304, Cur Loss: 0.33319074, Cur Avg Loss: 0.23779654, Log Avg loss: 0.26445773, Global Avg Loss: 0.82174336, Time: 0.0210 Steps: 72530, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000154, Sample Num: 2464, Cur Loss: 0.04939046, Cur Avg Loss: 0.23720135, Log Avg loss: 0.22863055, Global Avg Loss: 0.82166160, Time: 0.0211 Steps: 72540, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000164, Sample Num: 2624, Cur Loss: 0.09814499, Cur Avg Loss: 0.23150992, Log Avg loss: 0.14386188, Global Avg Loss: 0.82156817, Time: 0.0210 Steps: 72550, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000174, Sample Num: 2784, Cur Loss: 0.26090592, Cur Avg Loss: 0.23165993, Log Avg loss: 0.23412018, Global Avg Loss: 0.82148721, Time: 0.0210 Steps: 72560, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000184, Sample Num: 2944, Cur Loss: 1.40851641, Cur Avg Loss: 0.23537564, Log Avg loss: 0.30002897, Global Avg Loss: 0.82141536, Time: 0.0211 Steps: 72570, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000194, Sample Num: 3104, Cur Loss: 0.06632698, Cur Avg Loss: 0.24062237, Log Avg loss: 0.33716214, Global Avg Loss: 0.82134864, Time: 0.0210 Steps: 72580, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000204, Sample Num: 3264, Cur Loss: 0.04165302, Cur Avg Loss: 0.23747988, Log Avg loss: 0.17651568, Global Avg Loss: 0.82125980, Time: 0.0211 Steps: 72590, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000214, Sample Num: 3424, Cur Loss: 0.04468570, Cur Avg Loss: 0.23610851, Log Avg loss: 0.20813247, Global Avg Loss: 0.82117535, Time: 0.0211 Steps: 72600, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000224, Sample Num: 3584, Cur Loss: 0.30393758, Cur Avg Loss: 0.23680155, Log Avg loss: 0.25163268, Global Avg Loss: 0.82109691, Time: 0.0211 Steps: 72610, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000234, Sample Num: 3744, Cur Loss: 0.15295391, Cur Avg Loss: 0.23331839, Log Avg loss: 0.15529553, Global Avg Loss: 0.82100523, Time: 0.0210 Steps: 72620, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000244, Sample Num: 3904, Cur Loss: 0.18688558, Cur Avg Loss: 0.23074904, Log Avg loss: 0.17062632, Global Avg Loss: 0.82091568, Time: 0.0210 Steps: 72630, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000254, Sample Num: 4064, Cur Loss: 0.15172727, Cur Avg Loss: 0.22990469, Log Avg loss: 0.20930247, Global Avg Loss: 0.82083149, Time: 0.0210 Steps: 72640, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000264, Sample Num: 4224, Cur Loss: 0.09098071, Cur Avg Loss: 0.22988317, Log Avg loss: 0.22933647, Global Avg Loss: 0.82075007, Time: 0.0215 Steps: 72650, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000274, Sample Num: 4384, Cur Loss: 0.20985952, Cur Avg Loss: 0.22745382, Log Avg loss: 0.16331904, Global Avg Loss: 0.82065959, Time: 0.0212 Steps: 72660, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000284, Sample Num: 4544, Cur Loss: 0.15764713, Cur Avg Loss: 0.22783164, Log Avg loss: 0.23818406, Global Avg Loss: 0.82057943, Time: 0.0210 Steps: 72670, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000294, Sample Num: 4704, Cur Loss: 0.35283026, Cur Avg Loss: 0.22616147, Log Avg loss: 0.17872858, Global Avg Loss: 0.82049112, Time: 0.0211 Steps: 72680, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000304, Sample Num: 4864, Cur Loss: 0.18143909, Cur Avg Loss: 0.22440774, Log Avg loss: 0.17284795, Global Avg Loss: 0.82040203, Time: 0.0210 Steps: 72690, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000314, Sample Num: 5024, Cur Loss: 0.24782416, Cur Avg Loss: 0.22403813, Log Avg loss: 0.21280192, Global Avg Loss: 0.82031845, Time: 0.0208 Steps: 72700, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000324, Sample Num: 5184, Cur Loss: 0.22817111, Cur Avg Loss: 0.22446504, Log Avg loss: 0.23787009, Global Avg Loss: 0.82023834, Time: 0.0208 Steps: 72710, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000334, Sample Num: 5344, Cur Loss: 0.25535971, Cur Avg Loss: 0.22534299, Log Avg loss: 0.25378872, Global Avg Loss: 0.82016045, Time: 0.0209 Steps: 72720, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000344, Sample Num: 5504, Cur Loss: 0.10038104, Cur Avg Loss: 0.22368075, Log Avg loss: 0.16816200, Global Avg Loss: 0.82007080, Time: 0.0210 Steps: 72730, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000354, Sample Num: 5664, Cur Loss: 0.17514652, Cur Avg Loss: 0.22204513, Log Avg loss: 0.16577975, Global Avg Loss: 0.81998085, Time: 0.0210 Steps: 72740, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000364, Sample Num: 5824, Cur Loss: 0.11103680, Cur Avg Loss: 0.22222166, Log Avg loss: 0.22847076, Global Avg Loss: 0.81989955, Time: 0.0210 Steps: 72750, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000374, Sample Num: 5984, Cur Loss: 0.02500838, Cur Avg Loss: 0.22172094, Log Avg loss: 0.20349484, Global Avg Loss: 0.81981483, Time: 0.0210 Steps: 72760, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000384, Sample Num: 6144, Cur Loss: 0.12502839, Cur Avg Loss: 0.22092588, Log Avg loss: 0.19119039, Global Avg Loss: 0.81972844, Time: 0.0209 Steps: 72770, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000394, Sample Num: 6304, Cur Loss: 0.06100830, Cur Avg Loss: 0.21987607, Log Avg loss: 0.17956345, Global Avg Loss: 0.81964048, Time: 0.0208 Steps: 72780, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000404, Sample Num: 6464, Cur Loss: 0.27297938, Cur Avg Loss: 0.21981252, Log Avg loss: 0.21730881, Global Avg Loss: 0.81955774, Time: 0.0208 Steps: 72790, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000414, Sample Num: 6624, Cur Loss: 0.11127473, Cur Avg Loss: 0.21958019, Log Avg loss: 0.21019393, Global Avg Loss: 0.81947403, Time: 0.0208 Steps: 72800, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000424, Sample Num: 6784, Cur Loss: 0.28597772, Cur Avg Loss: 0.22204416, Log Avg loss: 0.32405261, Global Avg Loss: 0.81940599, Time: 0.0208 Steps: 72810, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000434, Sample Num: 6944, Cur Loss: 0.15129440, Cur Avg Loss: 0.22191270, Log Avg loss: 0.21633867, Global Avg Loss: 0.81932317, Time: 0.0208 Steps: 72820, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000444, Sample Num: 7104, Cur Loss: 0.32823730, Cur Avg Loss: 0.22091674, Log Avg loss: 0.17769207, Global Avg Loss: 0.81923507, Time: 0.0210 Steps: 72830, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000454, Sample Num: 7264, Cur Loss: 0.24535052, Cur Avg Loss: 0.22063355, Log Avg loss: 0.20805980, Global Avg Loss: 0.81915117, Time: 0.0210 Steps: 72840, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000464, Sample Num: 7424, Cur Loss: 0.09289958, Cur Avg Loss: 0.21997045, Log Avg loss: 0.18986605, Global Avg Loss: 0.81906479, Time: 0.0210 Steps: 72850, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000474, Sample Num: 7584, Cur Loss: 0.15187591, Cur Avg Loss: 0.21880335, Log Avg loss: 0.16464980, Global Avg Loss: 0.81897497, Time: 0.0211 Steps: 72860, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000484, Sample Num: 7744, Cur Loss: 0.05125546, Cur Avg Loss: 0.22014919, Log Avg loss: 0.28394177, Global Avg Loss: 0.81890154, Time: 0.0211 Steps: 72870, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000494, Sample Num: 7904, Cur Loss: 0.26530576, Cur Avg Loss: 0.22065965, Log Avg loss: 0.24536619, Global Avg Loss: 0.81882285, Time: 0.0210 Steps: 72880, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000504, Sample Num: 8064, Cur Loss: 0.04300071, Cur Avg Loss: 0.22078323, Log Avg loss: 0.22688780, Global Avg Loss: 0.81874164, Time: 0.0211 Steps: 72890, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000514, Sample Num: 8224, Cur Loss: 0.05928392, Cur Avg Loss: 0.22043476, Log Avg loss: 0.20287200, Global Avg Loss: 0.81865716, Time: 0.0248 Steps: 72900, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000524, Sample Num: 8384, Cur Loss: 0.13596117, Cur Avg Loss: 0.21914422, Log Avg loss: 0.15281031, Global Avg Loss: 0.81856583, Time: 0.0211 Steps: 72910, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000534, Sample Num: 8544, Cur Loss: 0.18121408, Cur Avg Loss: 0.22033844, Log Avg loss: 0.28291558, Global Avg Loss: 0.81849238, Time: 0.0211 Steps: 72920, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000544, Sample Num: 8704, Cur Loss: 0.17593773, Cur Avg Loss: 0.22042833, Log Avg loss: 0.22522857, Global Avg Loss: 0.81841103, Time: 0.0211 Steps: 72930, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000554, Sample Num: 8864, Cur Loss: 0.20349953, Cur Avg Loss: 0.21990323, Log Avg loss: 0.19133788, Global Avg Loss: 0.81832506, Time: 0.0211 Steps: 72940, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000564, Sample Num: 9024, Cur Loss: 0.20223071, Cur Avg Loss: 0.21983089, Log Avg loss: 0.21582342, Global Avg Loss: 0.81824247, Time: 0.0210 Steps: 72950, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000574, Sample Num: 9184, Cur Loss: 0.19973184, Cur Avg Loss: 0.21967899, Log Avg loss: 0.21111190, Global Avg Loss: 0.81815925, Time: 0.0211 Steps: 72960, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000584, Sample Num: 9344, Cur Loss: 0.14965054, Cur Avg Loss: 0.21910789, Log Avg loss: 0.18632632, Global Avg Loss: 0.81807266, Time: 0.0211 Steps: 72970, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000594, Sample Num: 9504, Cur Loss: 0.11363574, Cur Avg Loss: 0.21860803, Log Avg loss: 0.18941652, Global Avg Loss: 0.81798652, Time: 0.0211 Steps: 72980, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000604, Sample Num: 9664, Cur Loss: 0.07925063, Cur Avg Loss: 0.21859045, Log Avg loss: 0.21754588, Global Avg Loss: 0.81790426, Time: 0.0211 Steps: 72990, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000614, Sample Num: 9824, Cur Loss: 0.08504540, Cur Avg Loss: 0.21776369, Log Avg loss: 0.16782760, Global Avg Loss: 0.81781521, Time: 0.0209 Steps: 73000, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000624, Sample Num: 9984, Cur Loss: 0.10803778, Cur Avg Loss: 0.21719309, Log Avg loss: 0.18215844, Global Avg Loss: 0.81772814, Time: 0.0211 Steps: 73010, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000634, Sample Num: 10144, Cur Loss: 0.54125273, Cur Avg Loss: 0.21754672, Log Avg loss: 0.23961298, Global Avg Loss: 0.81764897, Time: 0.0211 Steps: 73020, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000644, Sample Num: 10304, Cur Loss: 0.38994047, Cur Avg Loss: 0.21729907, Log Avg loss: 0.20159826, Global Avg Loss: 0.81756462, Time: 0.0210 Steps: 73030, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000654, Sample Num: 10464, Cur Loss: 0.35869381, Cur Avg Loss: 0.21692187, Log Avg loss: 0.19262968, Global Avg Loss: 0.81747906, Time: 0.0211 Steps: 73040, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000664, Sample Num: 10624, Cur Loss: 0.19034779, Cur Avg Loss: 0.21687263, Log Avg loss: 0.21365252, Global Avg Loss: 0.81739640, Time: 0.0210 Steps: 73050, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000674, Sample Num: 10784, Cur Loss: 0.11311138, Cur Avg Loss: 0.21594079, Log Avg loss: 0.15406681, Global Avg Loss: 0.81730560, Time: 0.0210 Steps: 73060, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000684, Sample Num: 10944, Cur Loss: 0.25479746, Cur Avg Loss: 0.21534418, Log Avg loss: 0.17513230, Global Avg Loss: 0.81721772, Time: 0.0211 Steps: 73070, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000694, Sample Num: 11104, Cur Loss: 0.31677073, Cur Avg Loss: 0.21509289, Log Avg loss: 0.19790487, Global Avg Loss: 0.81713297, Time: 0.0211 Steps: 73080, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000704, Sample Num: 11264, Cur Loss: 0.15235136, Cur Avg Loss: 0.21589597, Log Avg loss: 0.27162966, Global Avg Loss: 0.81705834, Time: 0.0210 Steps: 73090, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000714, Sample Num: 11424, Cur Loss: 0.32386535, Cur Avg Loss: 0.21678235, Log Avg loss: 0.27918400, Global Avg Loss: 0.81698476, Time: 0.0211 Steps: 73100, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000724, Sample Num: 11584, Cur Loss: 0.52297813, Cur Avg Loss: 0.21697139, Log Avg loss: 0.23046873, Global Avg Loss: 0.81690454, Time: 0.0211 Steps: 73110, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000734, Sample Num: 11744, Cur Loss: 0.14003567, Cur Avg Loss: 0.21732031, Log Avg loss: 0.24258175, Global Avg Loss: 0.81682599, Time: 0.0210 Steps: 73120, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000744, Sample Num: 11904, Cur Loss: 0.15166369, Cur Avg Loss: 0.21704251, Log Avg loss: 0.19665222, Global Avg Loss: 0.81674119, Time: 0.0211 Steps: 73130, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000754, Sample Num: 12064, Cur Loss: 0.17162058, Cur Avg Loss: 0.21674849, Log Avg loss: 0.19487329, Global Avg Loss: 0.81665616, Time: 0.0211 Steps: 73140, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000764, Sample Num: 12224, Cur Loss: 0.13044490, Cur Avg Loss: 0.21802609, Log Avg loss: 0.31435734, Global Avg Loss: 0.81658749, Time: 0.0211 Steps: 73150, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000774, Sample Num: 12384, Cur Loss: 0.33003145, Cur Avg Loss: 0.21811969, Log Avg loss: 0.22527080, Global Avg Loss: 0.81650667, Time: 0.0212 Steps: 73160, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000784, Sample Num: 12544, Cur Loss: 0.32051474, Cur Avg Loss: 0.21832298, Log Avg loss: 0.23405714, Global Avg Loss: 0.81642707, Time: 0.0209 Steps: 73170, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000794, Sample Num: 12704, Cur Loss: 0.32582662, Cur Avg Loss: 0.21970210, Log Avg loss: 0.32782568, Global Avg Loss: 0.81636030, Time: 0.0210 Steps: 73180, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000804, Sample Num: 12864, Cur Loss: 0.11743142, Cur Avg Loss: 0.21954051, Log Avg loss: 0.20670964, Global Avg Loss: 0.81627700, Time: 0.0209 Steps: 73190, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000814, Sample Num: 13024, Cur Loss: 0.16216427, Cur Avg Loss: 0.21929037, Log Avg loss: 0.19917902, Global Avg Loss: 0.81619270, Time: 0.0210 Steps: 73200, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000824, Sample Num: 13184, Cur Loss: 0.12984230, Cur Avg Loss: 0.21812383, Log Avg loss: 0.12316750, Global Avg Loss: 0.81609804, Time: 0.0210 Steps: 73210, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000834, Sample Num: 13344, Cur Loss: 0.49164873, Cur Avg Loss: 0.21893049, Log Avg loss: 0.28539941, Global Avg Loss: 0.81602556, Time: 0.0208 Steps: 73220, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000844, Sample Num: 13504, Cur Loss: 0.16396640, Cur Avg Loss: 0.21954171, Log Avg loss: 0.27051742, Global Avg Loss: 0.81595107, Time: 0.0209 Steps: 73230, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000854, Sample Num: 13664, Cur Loss: 0.40910307, Cur Avg Loss: 0.22034954, Log Avg loss: 0.28853059, Global Avg Loss: 0.81587905, Time: 0.0208 Steps: 73240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000864, Sample Num: 13824, Cur Loss: 0.11497340, Cur Avg Loss: 0.22007832, Log Avg loss: 0.19691602, Global Avg Loss: 0.81579455, Time: 0.0208 Steps: 73250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000874, Sample Num: 13984, Cur Loss: 0.24935484, Cur Avg Loss: 0.21985291, Log Avg loss: 0.20037764, Global Avg Loss: 0.81571055, Time: 0.0209 Steps: 73260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000884, Sample Num: 14144, Cur Loss: 0.13143115, Cur Avg Loss: 0.21958280, Log Avg loss: 0.19597565, Global Avg Loss: 0.81562597, Time: 0.0210 Steps: 73270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000894, Sample Num: 14304, Cur Loss: 0.39952308, Cur Avg Loss: 0.21988741, Log Avg loss: 0.24681424, Global Avg Loss: 0.81554834, Time: 0.0209 Steps: 73280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000904, Sample Num: 14464, Cur Loss: 0.23652983, Cur Avg Loss: 0.22058369, Log Avg loss: 0.28283166, Global Avg Loss: 0.81547566, Time: 0.0210 Steps: 73290, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000914, Sample Num: 14624, Cur Loss: 0.19460745, Cur Avg Loss: 0.22038798, Log Avg loss: 0.20269583, Global Avg Loss: 0.81539206, Time: 0.0208 Steps: 73300, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000924, Sample Num: 14784, Cur Loss: 0.15097070, Cur Avg Loss: 0.21979602, Log Avg loss: 0.16569096, Global Avg Loss: 0.81530344, Time: 0.0210 Steps: 73310, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000934, Sample Num: 14944, Cur Loss: 0.58105129, Cur Avg Loss: 0.21979457, Log Avg loss: 0.21966059, Global Avg Loss: 0.81522220, Time: 0.0209 Steps: 73320, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000944, Sample Num: 15104, Cur Loss: 0.16190605, Cur Avg Loss: 0.21959046, Log Avg loss: 0.20052600, Global Avg Loss: 0.81513837, Time: 0.0208 Steps: 73330, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000954, Sample Num: 15264, Cur Loss: 0.06322595, Cur Avg Loss: 0.21877462, Log Avg loss: 0.14175961, Global Avg Loss: 0.81504655, Time: 0.0209 Steps: 73340, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000964, Sample Num: 15424, Cur Loss: 0.26281974, Cur Avg Loss: 0.21862612, Log Avg loss: 0.20445875, Global Avg Loss: 0.81496331, Time: 0.0209 Steps: 73350, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000974, Sample Num: 15584, Cur Loss: 0.13521332, Cur Avg Loss: 0.21834853, Log Avg loss: 0.19158877, Global Avg Loss: 0.81487834, Time: 0.0208 Steps: 73360, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000984, Sample Num: 15744, Cur Loss: 0.56190169, Cur Avg Loss: 0.21914369, Log Avg loss: 0.29659257, Global Avg Loss: 0.81480770, Time: 0.0209 Steps: 73370, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000994, Sample Num: 15904, Cur Loss: 0.18968579, Cur Avg Loss: 0.21895251, Log Avg loss: 0.20014036, Global Avg Loss: 0.81472393, Time: 0.0208 Steps: 73380, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001004, Sample Num: 16064, Cur Loss: 0.31682616, Cur Avg Loss: 0.21885871, Log Avg loss: 0.20953544, Global Avg Loss: 0.81464147, Time: 0.0209 Steps: 73390, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001014, Sample Num: 16224, Cur Loss: 0.27236643, Cur Avg Loss: 0.21883134, Log Avg loss: 0.21608270, Global Avg Loss: 0.81455992, Time: 0.0208 Steps: 73400, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001024, Sample Num: 16384, Cur Loss: 0.15156837, Cur Avg Loss: 0.21881161, Log Avg loss: 0.21681130, Global Avg Loss: 0.81447850, Time: 0.0254 Steps: 73410, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001034, Sample Num: 16544, Cur Loss: 0.14448754, Cur Avg Loss: 0.21858690, Log Avg loss: 0.19557695, Global Avg Loss: 0.81439420, Time: 0.0209 Steps: 73420, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001044, Sample Num: 16704, Cur Loss: 0.19864334, Cur Avg Loss: 0.21890540, Log Avg loss: 0.25183787, Global Avg Loss: 0.81431759, Time: 0.0210 Steps: 73430, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001054, Sample Num: 16864, Cur Loss: 0.07757609, Cur Avg Loss: 0.21868321, Log Avg loss: 0.19548719, Global Avg Loss: 0.81423333, Time: 0.0210 Steps: 73440, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001064, Sample Num: 17024, Cur Loss: 0.05211028, Cur Avg Loss: 0.21895353, Log Avg loss: 0.24744492, Global Avg Loss: 0.81415616, Time: 0.0208 Steps: 73450, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001074, Sample Num: 17184, Cur Loss: 0.08288790, Cur Avg Loss: 0.21848722, Log Avg loss: 0.16887193, Global Avg Loss: 0.81406832, Time: 0.0210 Steps: 73460, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001084, Sample Num: 17344, Cur Loss: 0.17066026, Cur Avg Loss: 0.21807291, Log Avg loss: 0.17357602, Global Avg Loss: 0.81398114, Time: 0.0209 Steps: 73470, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001094, Sample Num: 17504, Cur Loss: 0.29692203, Cur Avg Loss: 0.21853512, Log Avg loss: 0.26863865, Global Avg Loss: 0.81390692, Time: 0.0208 Steps: 73480, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001104, Sample Num: 17664, Cur Loss: 0.12100003, Cur Avg Loss: 0.21918012, Log Avg loss: 0.28974295, Global Avg Loss: 0.81383560, Time: 0.0209 Steps: 73490, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001114, Sample Num: 17824, Cur Loss: 0.15761028, Cur Avg Loss: 0.21901440, Log Avg loss: 0.20071922, Global Avg Loss: 0.81375218, Time: 0.0209 Steps: 73500, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001124, Sample Num: 17984, Cur Loss: 0.40406829, Cur Avg Loss: 0.21869507, Log Avg loss: 0.18312154, Global Avg Loss: 0.81366639, Time: 0.0209 Steps: 73510, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001134, Sample Num: 18144, Cur Loss: 0.14461702, Cur Avg Loss: 0.21917701, Log Avg loss: 0.27334726, Global Avg Loss: 0.81359290, Time: 0.0210 Steps: 73520, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001144, Sample Num: 18304, Cur Loss: 0.49850950, Cur Avg Loss: 0.21965996, Log Avg loss: 0.27442620, Global Avg Loss: 0.81351957, Time: 0.0209 Steps: 73530, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001154, Sample Num: 18464, Cur Loss: 0.08672068, Cur Avg Loss: 0.21903701, Log Avg loss: 0.14777139, Global Avg Loss: 0.81342905, Time: 0.0209 Steps: 73540, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001164, Sample Num: 18624, Cur Loss: 0.26110223, Cur Avg Loss: 0.21883714, Log Avg loss: 0.19577246, Global Avg Loss: 0.81334507, Time: 0.0209 Steps: 73550, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001174, Sample Num: 18784, Cur Loss: 0.25269535, Cur Avg Loss: 0.21883855, Log Avg loss: 0.21900297, Global Avg Loss: 0.81326427, Time: 0.0209 Steps: 73560, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001184, Sample Num: 18944, Cur Loss: 0.23900810, Cur Avg Loss: 0.21878130, Log Avg loss: 0.21205914, Global Avg Loss: 0.81318255, Time: 0.0210 Steps: 73570, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001194, Sample Num: 19104, Cur Loss: 0.44668925, Cur Avg Loss: 0.21910276, Log Avg loss: 0.25716400, Global Avg Loss: 0.81310699, Time: 0.0209 Steps: 73580, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001204, Sample Num: 19264, Cur Loss: 0.29469651, Cur Avg Loss: 0.22004222, Log Avg loss: 0.33221437, Global Avg Loss: 0.81304164, Time: 0.0211 Steps: 73590, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001214, Sample Num: 19424, Cur Loss: 0.13563117, Cur Avg Loss: 0.22000821, Log Avg loss: 0.21591268, Global Avg Loss: 0.81296051, Time: 0.0211 Steps: 73600, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001224, Sample Num: 19584, Cur Loss: 0.22185895, Cur Avg Loss: 0.21975378, Log Avg loss: 0.18886650, Global Avg Loss: 0.81287572, Time: 0.0209 Steps: 73610, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001234, Sample Num: 19744, Cur Loss: 0.05171580, Cur Avg Loss: 0.22028550, Log Avg loss: 0.28536756, Global Avg Loss: 0.81280407, Time: 0.0210 Steps: 73620, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001244, Sample Num: 19904, Cur Loss: 0.16038680, Cur Avg Loss: 0.22083597, Log Avg loss: 0.28876450, Global Avg Loss: 0.81273290, Time: 0.0210 Steps: 73630, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001254, Sample Num: 20064, Cur Loss: 0.15462953, Cur Avg Loss: 0.22064300, Log Avg loss: 0.19663735, Global Avg Loss: 0.81264923, Time: 0.0209 Steps: 73640, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001264, Sample Num: 20224, Cur Loss: 0.11141080, Cur Avg Loss: 0.22033291, Log Avg loss: 0.18144796, Global Avg Loss: 0.81256353, Time: 0.0209 Steps: 73650, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001274, Sample Num: 20384, Cur Loss: 0.05841199, Cur Avg Loss: 0.21988737, Log Avg loss: 0.16357016, Global Avg Loss: 0.81247543, Time: 0.0209 Steps: 73660, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001284, Sample Num: 20544, Cur Loss: 0.40298164, Cur Avg Loss: 0.21977893, Log Avg loss: 0.20596391, Global Avg Loss: 0.81239310, Time: 0.0247 Steps: 73670, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001294, Sample Num: 20704, Cur Loss: 0.67556548, Cur Avg Loss: 0.22038317, Log Avg loss: 0.29796826, Global Avg Loss: 0.81232328, Time: 0.0209 Steps: 73680, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001304, Sample Num: 20864, Cur Loss: 0.45770064, Cur Avg Loss: 0.22052656, Log Avg loss: 0.23908115, Global Avg Loss: 0.81224549, Time: 0.0210 Steps: 73690, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001314, Sample Num: 21024, Cur Loss: 0.21513274, Cur Avg Loss: 0.22054032, Log Avg loss: 0.22233430, Global Avg Loss: 0.81216544, Time: 0.0209 Steps: 73700, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001324, Sample Num: 21184, Cur Loss: 0.19733727, Cur Avg Loss: 0.22082884, Log Avg loss: 0.25873966, Global Avg Loss: 0.81209036, Time: 0.0208 Steps: 73710, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001334, Sample Num: 21344, Cur Loss: 0.15207475, Cur Avg Loss: 0.22109487, Log Avg loss: 0.25631750, Global Avg Loss: 0.81201497, Time: 0.0209 Steps: 73720, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001344, Sample Num: 21504, Cur Loss: 0.13592355, Cur Avg Loss: 0.22093763, Log Avg loss: 0.19996184, Global Avg Loss: 0.81193196, Time: 0.0210 Steps: 73730, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001354, Sample Num: 21664, Cur Loss: 0.21135294, Cur Avg Loss: 0.22145170, Log Avg loss: 0.29054320, Global Avg Loss: 0.81186125, Time: 0.0209 Steps: 73740, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001364, Sample Num: 21824, Cur Loss: 0.20058885, Cur Avg Loss: 0.22177528, Log Avg loss: 0.26558740, Global Avg Loss: 0.81178718, Time: 0.0210 Steps: 73750, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001374, Sample Num: 21984, Cur Loss: 0.16977759, Cur Avg Loss: 0.22155064, Log Avg loss: 0.19091050, Global Avg Loss: 0.81170301, Time: 0.0209 Steps: 73760, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001384, Sample Num: 22144, Cur Loss: 0.30094957, Cur Avg Loss: 0.22152149, Log Avg loss: 0.21751604, Global Avg Loss: 0.81162246, Time: 0.0209 Steps: 73770, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001394, Sample Num: 22304, Cur Loss: 0.14056149, Cur Avg Loss: 0.22137523, Log Avg loss: 0.20113239, Global Avg Loss: 0.81153972, Time: 0.0209 Steps: 73780, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001404, Sample Num: 22464, Cur Loss: 0.37542337, Cur Avg Loss: 0.22158342, Log Avg loss: 0.25060488, Global Avg Loss: 0.81146370, Time: 0.0209 Steps: 73790, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001414, Sample Num: 22624, Cur Loss: 0.36216143, Cur Avg Loss: 0.22186058, Log Avg loss: 0.26077430, Global Avg Loss: 0.81138908, Time: 0.0210 Steps: 73800, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001424, Sample Num: 22784, Cur Loss: 0.06799174, Cur Avg Loss: 0.22184670, Log Avg loss: 0.21988472, Global Avg Loss: 0.81130894, Time: 0.0210 Steps: 73810, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001434, Sample Num: 22944, Cur Loss: 0.19741811, Cur Avg Loss: 0.22186819, Log Avg loss: 0.22492822, Global Avg Loss: 0.81122951, Time: 0.0209 Steps: 73820, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001444, Sample Num: 23104, Cur Loss: 0.49419504, Cur Avg Loss: 0.22162260, Log Avg loss: 0.18640516, Global Avg Loss: 0.81114488, Time: 0.0209 Steps: 73830, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001454, Sample Num: 23264, Cur Loss: 0.10761185, Cur Avg Loss: 0.22109294, Log Avg loss: 0.14460911, Global Avg Loss: 0.81105461, Time: 0.0210 Steps: 73840, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001464, Sample Num: 23424, Cur Loss: 0.11488670, Cur Avg Loss: 0.22106821, Log Avg loss: 0.21747349, Global Avg Loss: 0.81097423, Time: 0.0210 Steps: 73850, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001474, Sample Num: 23584, Cur Loss: 0.20970827, Cur Avg Loss: 0.22077031, Log Avg loss: 0.17715655, Global Avg Loss: 0.81088842, Time: 0.0210 Steps: 73860, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001484, Sample Num: 23744, Cur Loss: 0.06445370, Cur Avg Loss: 0.22020852, Log Avg loss: 0.13740145, Global Avg Loss: 0.81079725, Time: 0.0208 Steps: 73870, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001494, Sample Num: 23904, Cur Loss: 0.17620200, Cur Avg Loss: 0.22045464, Log Avg loss: 0.25697907, Global Avg Loss: 0.81072229, Time: 0.0210 Steps: 73880, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001504, Sample Num: 24064, Cur Loss: 0.16329136, Cur Avg Loss: 0.22033580, Log Avg loss: 0.20258134, Global Avg Loss: 0.81063998, Time: 0.0209 Steps: 73890, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001514, Sample Num: 24224, Cur Loss: 0.13302696, Cur Avg Loss: 0.21986161, Log Avg loss: 0.14854263, Global Avg Loss: 0.81055039, Time: 0.0208 Steps: 73900, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001524, Sample Num: 24384, Cur Loss: 0.10174662, Cur Avg Loss: 0.21998135, Log Avg loss: 0.23810968, Global Avg Loss: 0.81047294, Time: 0.0209 Steps: 73910, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001534, Sample Num: 24544, Cur Loss: 0.27577221, Cur Avg Loss: 0.22023252, Log Avg loss: 0.25851192, Global Avg Loss: 0.81039827, Time: 0.0209 Steps: 73920, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001544, Sample Num: 24704, Cur Loss: 0.20400494, Cur Avg Loss: 0.22099955, Log Avg loss: 0.33866205, Global Avg Loss: 0.81033446, Time: 0.0209 Steps: 73930, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001554, Sample Num: 24864, Cur Loss: 0.19021456, Cur Avg Loss: 0.22075235, Log Avg loss: 0.18258367, Global Avg Loss: 0.81024956, Time: 0.0209 Steps: 73940, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001564, Sample Num: 25024, Cur Loss: 0.29031509, Cur Avg Loss: 0.22043883, Log Avg loss: 0.17171851, Global Avg Loss: 0.81016321, Time: 0.0210 Steps: 73950, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001574, Sample Num: 25184, Cur Loss: 0.07654247, Cur Avg Loss: 0.22057668, Log Avg loss: 0.24213635, Global Avg Loss: 0.81008641, Time: 0.0209 Steps: 73960, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001584, Sample Num: 25344, Cur Loss: 0.22307242, Cur Avg Loss: 0.22043272, Log Avg loss: 0.19777266, Global Avg Loss: 0.81000363, Time: 0.0208 Steps: 73970, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001594, Sample Num: 25504, Cur Loss: 0.11359033, Cur Avg Loss: 0.22046667, Log Avg loss: 0.22584513, Global Avg Loss: 0.80992467, Time: 0.0209 Steps: 73980, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001604, Sample Num: 25664, Cur Loss: 0.13171509, Cur Avg Loss: 0.22022852, Log Avg loss: 0.18226713, Global Avg Loss: 0.80983984, Time: 0.0210 Steps: 73990, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001614, Sample Num: 25824, Cur Loss: 0.35195577, Cur Avg Loss: 0.22000811, Log Avg loss: 0.18465433, Global Avg Loss: 0.80975536, Time: 0.0209 Steps: 74000, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001624, Sample Num: 25984, Cur Loss: 0.50039256, Cur Avg Loss: 0.22036791, Log Avg loss: 0.27843941, Global Avg Loss: 0.80968357, Time: 0.0208 Steps: 74010, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001634, Sample Num: 26144, Cur Loss: 0.24769998, Cur Avg Loss: 0.22010214, Log Avg loss: 0.17694181, Global Avg Loss: 0.80959808, Time: 0.0209 Steps: 74020, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001644, Sample Num: 26304, Cur Loss: 0.05243360, Cur Avg Loss: 0.21959996, Log Avg loss: 0.13754368, Global Avg Loss: 0.80950730, Time: 0.0208 Steps: 74030, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001654, Sample Num: 26464, Cur Loss: 0.23545158, Cur Avg Loss: 0.21963301, Log Avg loss: 0.22506545, Global Avg Loss: 0.80942837, Time: 0.0208 Steps: 74040, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001664, Sample Num: 26624, Cur Loss: 0.27740574, Cur Avg Loss: 0.22005701, Log Avg loss: 0.29018672, Global Avg Loss: 0.80935825, Time: 0.0209 Steps: 74050, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001674, Sample Num: 26784, Cur Loss: 0.13667879, Cur Avg Loss: 0.21995297, Log Avg loss: 0.20264059, Global Avg Loss: 0.80927632, Time: 0.0208 Steps: 74060, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001684, Sample Num: 26944, Cur Loss: 0.18676838, Cur Avg Loss: 0.21952237, Log Avg loss: 0.14743979, Global Avg Loss: 0.80918697, Time: 0.0208 Steps: 74070, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001694, Sample Num: 27104, Cur Loss: 0.13528931, Cur Avg Loss: 0.21920441, Log Avg loss: 0.16566055, Global Avg Loss: 0.80910010, Time: 0.0208 Steps: 74080, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001704, Sample Num: 27264, Cur Loss: 0.12007377, Cur Avg Loss: 0.21901683, Log Avg loss: 0.18724147, Global Avg Loss: 0.80901617, Time: 0.0209 Steps: 74090, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001714, Sample Num: 27424, Cur Loss: 0.46891612, Cur Avg Loss: 0.21932868, Log Avg loss: 0.27246774, Global Avg Loss: 0.80894376, Time: 0.0209 Steps: 74100, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001724, Sample Num: 27584, Cur Loss: 0.35094494, Cur Avg Loss: 0.21931496, Log Avg loss: 0.21696299, Global Avg Loss: 0.80886388, Time: 0.0208 Steps: 74110, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001734, Sample Num: 27744, Cur Loss: 0.41246527, Cur Avg Loss: 0.21962993, Log Avg loss: 0.27393004, Global Avg Loss: 0.80879171, Time: 0.0209 Steps: 74120, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001744, Sample Num: 27904, Cur Loss: 0.25676721, Cur Avg Loss: 0.22009579, Log Avg loss: 0.30087574, Global Avg Loss: 0.80872319, Time: 0.0209 Steps: 74130, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001754, Sample Num: 28064, Cur Loss: 0.24274644, Cur Avg Loss: 0.21983832, Log Avg loss: 0.17493654, Global Avg Loss: 0.80863771, Time: 0.0209 Steps: 74140, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001764, Sample Num: 28224, Cur Loss: 0.12808211, Cur Avg Loss: 0.21973220, Log Avg loss: 0.20111897, Global Avg Loss: 0.80855578, Time: 0.0208 Steps: 74150, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001774, Sample Num: 28384, Cur Loss: 0.26766491, Cur Avg Loss: 0.21946732, Log Avg loss: 0.17274157, Global Avg Loss: 0.80847004, Time: 0.0209 Steps: 74160, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001784, Sample Num: 28544, Cur Loss: 0.18702728, Cur Avg Loss: 0.21932093, Log Avg loss: 0.19335151, Global Avg Loss: 0.80838711, Time: 0.0209 Steps: 74170, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001794, Sample Num: 28704, Cur Loss: 0.39406365, Cur Avg Loss: 0.21968010, Log Avg loss: 0.28375633, Global Avg Loss: 0.80831638, Time: 0.0248 Steps: 74180, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001804, Sample Num: 28864, Cur Loss: 0.31951433, Cur Avg Loss: 0.22000389, Log Avg loss: 0.27809233, Global Avg Loss: 0.80824492, Time: 0.0209 Steps: 74190, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001814, Sample Num: 29024, Cur Loss: 0.30409038, Cur Avg Loss: 0.22005797, Log Avg loss: 0.22981274, Global Avg Loss: 0.80816696, Time: 0.0209 Steps: 74200, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001824, Sample Num: 29184, Cur Loss: 0.30021369, Cur Avg Loss: 0.22039447, Log Avg loss: 0.28143626, Global Avg Loss: 0.80809598, Time: 0.0209 Steps: 74210, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001834, Sample Num: 29344, Cur Loss: 0.17757162, Cur Avg Loss: 0.22048045, Log Avg loss: 0.23616327, Global Avg Loss: 0.80801892, Time: 0.0209 Steps: 74220, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001844, Sample Num: 29504, Cur Loss: 0.42975044, Cur Avg Loss: 0.22077197, Log Avg loss: 0.27423767, Global Avg Loss: 0.80794701, Time: 0.0208 Steps: 74230, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001854, Sample Num: 29664, Cur Loss: 0.69188100, Cur Avg Loss: 0.22136067, Log Avg loss: 0.32991511, Global Avg Loss: 0.80788262, Time: 0.0209 Steps: 74240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001864, Sample Num: 29824, Cur Loss: 0.27390543, Cur Avg Loss: 0.22196442, Log Avg loss: 0.33390005, Global Avg Loss: 0.80781879, Time: 0.0209 Steps: 74250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001874, Sample Num: 29984, Cur Loss: 0.17258185, Cur Avg Loss: 0.22216740, Log Avg loss: 0.26000291, Global Avg Loss: 0.80774502, Time: 0.0209 Steps: 74260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001884, Sample Num: 30144, Cur Loss: 0.27488440, Cur Avg Loss: 0.22268714, Log Avg loss: 0.32008613, Global Avg Loss: 0.80767936, Time: 0.0209 Steps: 74270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001894, Sample Num: 30304, Cur Loss: 0.05131187, Cur Avg Loss: 0.22294223, Log Avg loss: 0.27100205, Global Avg Loss: 0.80760711, Time: 0.0209 Steps: 74280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001904, Sample Num: 30464, Cur Loss: 0.24304070, Cur Avg Loss: 0.22316855, Log Avg loss: 0.26603404, Global Avg Loss: 0.80753421, Time: 0.0208 Steps: 74290, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001914, Sample Num: 30624, Cur Loss: 0.09924763, Cur Avg Loss: 0.22326193, Log Avg loss: 0.24104092, Global Avg Loss: 0.80745796, Time: 0.0209 Steps: 74300, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001924, Sample Num: 30784, Cur Loss: 0.16721076, Cur Avg Loss: 0.22288815, Log Avg loss: 0.15134592, Global Avg Loss: 0.80736967, Time: 0.0209 Steps: 74310, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001934, Sample Num: 30944, Cur Loss: 0.18288045, Cur Avg Loss: 0.22298171, Log Avg loss: 0.24098396, Global Avg Loss: 0.80729346, Time: 0.0209 Steps: 74320, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001944, Sample Num: 31104, Cur Loss: 0.11916861, Cur Avg Loss: 0.22329179, Log Avg loss: 0.28326088, Global Avg Loss: 0.80722296, Time: 0.0209 Steps: 74330, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001954, Sample Num: 31264, Cur Loss: 0.35108292, Cur Avg Loss: 0.22352209, Log Avg loss: 0.26829264, Global Avg Loss: 0.80715046, Time: 0.0209 Steps: 74340, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001964, Sample Num: 31424, Cur Loss: 0.09133397, Cur Avg Loss: 0.22400743, Log Avg loss: 0.31884237, Global Avg Loss: 0.80708479, Time: 0.0209 Steps: 74350, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001974, Sample Num: 31584, Cur Loss: 0.10959187, Cur Avg Loss: 0.22416167, Log Avg loss: 0.25445464, Global Avg Loss: 0.80701047, Time: 0.0209 Steps: 74360, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001984, Sample Num: 31744, Cur Loss: 0.19789031, Cur Avg Loss: 0.22480520, Log Avg loss: 0.35183860, Global Avg Loss: 0.80694926, Time: 0.0209 Steps: 74370, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001994, Sample Num: 31904, Cur Loss: 0.38361350, Cur Avg Loss: 0.22506898, Log Avg loss: 0.27740276, Global Avg Loss: 0.80687807, Time: 0.0209 Steps: 74380, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002004, Sample Num: 32064, Cur Loss: 0.17390272, Cur Avg Loss: 0.22454940, Log Avg loss: 0.12094500, Global Avg Loss: 0.80678586, Time: 0.0209 Steps: 74390, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002014, Sample Num: 32224, Cur Loss: 0.23989199, Cur Avg Loss: 0.22446006, Log Avg loss: 0.20655595, Global Avg Loss: 0.80670519, Time: 0.0209 Steps: 74400, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002024, Sample Num: 32384, Cur Loss: 0.24434419, Cur Avg Loss: 0.22405165, Log Avg loss: 0.14179754, Global Avg Loss: 0.80661583, Time: 0.0208 Steps: 74410, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002034, Sample Num: 32544, Cur Loss: 0.07509702, Cur Avg Loss: 0.22388761, Log Avg loss: 0.19068624, Global Avg Loss: 0.80653307, Time: 0.0209 Steps: 74420, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002044, Sample Num: 32704, Cur Loss: 0.19111443, Cur Avg Loss: 0.22364352, Log Avg loss: 0.17399585, Global Avg Loss: 0.80644808, Time: 0.0209 Steps: 74430, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002054, Sample Num: 32864, Cur Loss: 0.07597462, Cur Avg Loss: 0.22377423, Log Avg loss: 0.25049077, Global Avg Loss: 0.80637340, Time: 0.0227 Steps: 74440, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002064, Sample Num: 33024, Cur Loss: 0.18631953, Cur Avg Loss: 0.22359558, Log Avg loss: 0.18690126, Global Avg Loss: 0.80629019, Time: 0.0208 Steps: 74450, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002074, Sample Num: 33184, Cur Loss: 0.12123463, Cur Avg Loss: 0.22374861, Log Avg loss: 0.25533354, Global Avg Loss: 0.80621620, Time: 0.0209 Steps: 74460, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002084, Sample Num: 33344, Cur Loss: 0.24887556, Cur Avg Loss: 0.22394978, Log Avg loss: 0.26567291, Global Avg Loss: 0.80614361, Time: 0.0209 Steps: 74470, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002094, Sample Num: 33504, Cur Loss: 0.38973129, Cur Avg Loss: 0.22398518, Log Avg loss: 0.23136193, Global Avg Loss: 0.80606644, Time: 0.0209 Steps: 74480, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002104, Sample Num: 33664, Cur Loss: 0.59756672, Cur Avg Loss: 0.22417551, Log Avg loss: 0.26403158, Global Avg Loss: 0.80599367, Time: 0.0209 Steps: 74490, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002114, Sample Num: 33824, Cur Loss: 0.06198169, Cur Avg Loss: 0.22394045, Log Avg loss: 0.17448263, Global Avg Loss: 0.80590890, Time: 0.0209 Steps: 74500, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002124, Sample Num: 33984, Cur Loss: 0.11698861, Cur Avg Loss: 0.22379832, Log Avg loss: 0.19375345, Global Avg Loss: 0.80582675, Time: 0.0209 Steps: 74510, Updated lr: 0.000030 ***** Running evaluation checkpoint-74515 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-74515 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.750621, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.30902, "eval_total_loss": 217.241115, "eval_mae": 0.446867, "eval_mse": 0.30902, "eval_r2": 0.803567, "eval_sp_statistic": 0.894194, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.920433, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.320637, "test_total_loss": 160.959803, "test_mae": 0.383497, "test_mse": 0.320745, "test_r2": 0.792989, "test_sp_statistic": 0.870592, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.918829, "test_ps_pvalue": 0.0, "lr": 3.0284495021337127e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8057846411417725, "train_cur_epoch_loss": 476.23922587931156, "train_cur_epoch_avg_loss": 0.2236915105116541, "train_cur_epoch_time": 44.7506206035614, "train_cur_epoch_avg_time": 0.021019549367572287, "epoch": 35, "step": 74515} ################################################## Training, Epoch: 0036, Batch: 000005, Sample Num: 80, Cur Loss: 0.13648832, Cur Avg Loss: 0.19991430, Log Avg loss: 0.18911591, Global Avg Loss: 0.80574399, Time: 0.0246 Steps: 74520, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000015, Sample Num: 240, Cur Loss: 0.10634002, Cur Avg Loss: 0.17366090, Log Avg loss: 0.16053420, Global Avg Loss: 0.80565742, Time: 0.0209 Steps: 74530, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000025, Sample Num: 400, Cur Loss: 0.20649433, Cur Avg Loss: 0.19825940, Log Avg loss: 0.23515716, Global Avg Loss: 0.80558088, Time: 0.0209 Steps: 74540, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000035, Sample Num: 560, Cur Loss: 0.20814160, Cur Avg Loss: 0.20169770, Log Avg loss: 0.21029346, Global Avg Loss: 0.80550103, Time: 0.0209 Steps: 74550, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000045, Sample Num: 720, Cur Loss: 0.19335258, Cur Avg Loss: 0.19868787, Log Avg loss: 0.18815347, Global Avg Loss: 0.80541823, Time: 0.0209 Steps: 74560, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000055, Sample Num: 880, Cur Loss: 0.08150877, Cur Avg Loss: 0.23770824, Log Avg loss: 0.41329991, Global Avg Loss: 0.80536565, Time: 0.0208 Steps: 74570, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000065, Sample Num: 1040, Cur Loss: 0.18356334, Cur Avg Loss: 0.22903726, Log Avg loss: 0.18134685, Global Avg Loss: 0.80528198, Time: 0.0209 Steps: 74580, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000075, Sample Num: 1200, Cur Loss: 0.21016228, Cur Avg Loss: 0.24670802, Log Avg loss: 0.36156799, Global Avg Loss: 0.80522249, Time: 0.0209 Steps: 74590, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000085, Sample Num: 1360, Cur Loss: 0.16593885, Cur Avg Loss: 0.25155773, Log Avg loss: 0.28793054, Global Avg Loss: 0.80515315, Time: 0.0209 Steps: 74600, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000095, Sample Num: 1520, Cur Loss: 0.23030685, Cur Avg Loss: 0.24476344, Log Avg loss: 0.18701197, Global Avg Loss: 0.80507030, Time: 0.0208 Steps: 74610, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000105, Sample Num: 1680, Cur Loss: 0.24371871, Cur Avg Loss: 0.24233634, Log Avg loss: 0.21927891, Global Avg Loss: 0.80499180, Time: 0.0209 Steps: 74620, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000115, Sample Num: 1840, Cur Loss: 0.10696539, Cur Avg Loss: 0.23527912, Log Avg loss: 0.16117823, Global Avg Loss: 0.80490553, Time: 0.0209 Steps: 74630, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000125, Sample Num: 2000, Cur Loss: 0.16095480, Cur Avg Loss: 0.23812588, Log Avg loss: 0.27086364, Global Avg Loss: 0.80483398, Time: 0.0209 Steps: 74640, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000135, Sample Num: 2160, Cur Loss: 0.11930212, Cur Avg Loss: 0.23448718, Log Avg loss: 0.18900345, Global Avg Loss: 0.80475148, Time: 0.0209 Steps: 74650, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000145, Sample Num: 2320, Cur Loss: 0.08852975, Cur Avg Loss: 0.23169268, Log Avg loss: 0.19396700, Global Avg Loss: 0.80466968, Time: 0.0208 Steps: 74660, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000155, Sample Num: 2480, Cur Loss: 0.09275699, Cur Avg Loss: 0.22778088, Log Avg loss: 0.17105972, Global Avg Loss: 0.80458482, Time: 0.0208 Steps: 74670, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000165, Sample Num: 2640, Cur Loss: 0.23986962, Cur Avg Loss: 0.22630210, Log Avg loss: 0.20338100, Global Avg Loss: 0.80450432, Time: 0.0209 Steps: 74680, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000175, Sample Num: 2800, Cur Loss: 0.12148469, Cur Avg Loss: 0.22167262, Log Avg loss: 0.14528631, Global Avg Loss: 0.80441606, Time: 0.0208 Steps: 74690, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000185, Sample Num: 2960, Cur Loss: 0.07889970, Cur Avg Loss: 0.21847847, Log Avg loss: 0.16258080, Global Avg Loss: 0.80433013, Time: 0.0209 Steps: 74700, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000195, Sample Num: 3120, Cur Loss: 0.17583580, Cur Avg Loss: 0.21587192, Log Avg loss: 0.16765076, Global Avg Loss: 0.80424491, Time: 0.0209 Steps: 74710, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000205, Sample Num: 3280, Cur Loss: 0.16628301, Cur Avg Loss: 0.21236287, Log Avg loss: 0.14393643, Global Avg Loss: 0.80415654, Time: 0.0209 Steps: 74720, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000215, Sample Num: 3440, Cur Loss: 0.26072994, Cur Avg Loss: 0.21416546, Log Avg loss: 0.25111849, Global Avg Loss: 0.80408254, Time: 0.0210 Steps: 74730, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000225, Sample Num: 3600, Cur Loss: 0.07994266, Cur Avg Loss: 0.21490501, Log Avg loss: 0.23080535, Global Avg Loss: 0.80400584, Time: 0.0210 Steps: 74740, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000235, Sample Num: 3760, Cur Loss: 0.36981106, Cur Avg Loss: 0.21789607, Log Avg loss: 0.28519488, Global Avg Loss: 0.80393643, Time: 0.0210 Steps: 74750, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000245, Sample Num: 3920, Cur Loss: 0.24838196, Cur Avg Loss: 0.21677072, Log Avg loss: 0.19032501, Global Avg Loss: 0.80385435, Time: 0.0210 Steps: 74760, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000255, Sample Num: 4080, Cur Loss: 0.29744214, Cur Avg Loss: 0.21646362, Log Avg loss: 0.20893961, Global Avg Loss: 0.80377479, Time: 0.0208 Steps: 74770, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000265, Sample Num: 4240, Cur Loss: 0.58407992, Cur Avg Loss: 0.21549848, Log Avg loss: 0.19088739, Global Avg Loss: 0.80369283, Time: 0.0211 Steps: 74780, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000275, Sample Num: 4400, Cur Loss: 0.25558478, Cur Avg Loss: 0.21443862, Log Avg loss: 0.18635233, Global Avg Loss: 0.80361028, Time: 0.0219 Steps: 74790, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000285, Sample Num: 4560, Cur Loss: 0.08061800, Cur Avg Loss: 0.21265983, Log Avg loss: 0.16374331, Global Avg Loss: 0.80352474, Time: 0.0219 Steps: 74800, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000295, Sample Num: 4720, Cur Loss: 0.24531232, Cur Avg Loss: 0.21163993, Log Avg loss: 0.18257262, Global Avg Loss: 0.80344174, Time: 0.0220 Steps: 74810, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000305, Sample Num: 4880, Cur Loss: 0.09475088, Cur Avg Loss: 0.20932054, Log Avg loss: 0.14089869, Global Avg Loss: 0.80335318, Time: 0.0219 Steps: 74820, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000315, Sample Num: 5040, Cur Loss: 0.37909803, Cur Avg Loss: 0.20825243, Log Avg loss: 0.17567501, Global Avg Loss: 0.80326930, Time: 0.0212 Steps: 74830, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000325, Sample Num: 5200, Cur Loss: 0.06422819, Cur Avg Loss: 0.20790632, Log Avg loss: 0.19700369, Global Avg Loss: 0.80318830, Time: 0.0211 Steps: 74840, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000335, Sample Num: 5360, Cur Loss: 0.15910515, Cur Avg Loss: 0.20864490, Log Avg loss: 0.23264888, Global Avg Loss: 0.80311207, Time: 0.0207 Steps: 74850, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000345, Sample Num: 5520, Cur Loss: 0.09162774, Cur Avg Loss: 0.20828631, Log Avg loss: 0.19627338, Global Avg Loss: 0.80303101, Time: 0.0211 Steps: 74860, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000355, Sample Num: 5680, Cur Loss: 0.20131657, Cur Avg Loss: 0.20851022, Log Avg loss: 0.21623523, Global Avg Loss: 0.80295263, Time: 0.0220 Steps: 74870, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000365, Sample Num: 5840, Cur Loss: 0.13622579, Cur Avg Loss: 0.20755605, Log Avg loss: 0.17368297, Global Avg Loss: 0.80286860, Time: 0.0208 Steps: 74880, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000375, Sample Num: 6000, Cur Loss: 0.13996579, Cur Avg Loss: 0.20762777, Log Avg loss: 0.21024571, Global Avg Loss: 0.80278946, Time: 0.0208 Steps: 74890, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000385, Sample Num: 6160, Cur Loss: 0.20357046, Cur Avg Loss: 0.20802518, Log Avg loss: 0.22292811, Global Avg Loss: 0.80271205, Time: 0.0211 Steps: 74900, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000395, Sample Num: 6320, Cur Loss: 0.20622912, Cur Avg Loss: 0.21108763, Log Avg loss: 0.32899165, Global Avg Loss: 0.80264881, Time: 0.0211 Steps: 74910, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000405, Sample Num: 6480, Cur Loss: 0.15935451, Cur Avg Loss: 0.21013472, Log Avg loss: 0.17249498, Global Avg Loss: 0.80256470, Time: 0.0219 Steps: 74920, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000415, Sample Num: 6640, Cur Loss: 0.13033247, Cur Avg Loss: 0.20945825, Log Avg loss: 0.18206110, Global Avg Loss: 0.80248189, Time: 0.0208 Steps: 74930, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000425, Sample Num: 6800, Cur Loss: 0.31217369, Cur Avg Loss: 0.20946320, Log Avg loss: 0.20966884, Global Avg Loss: 0.80240278, Time: 0.0208 Steps: 74940, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000435, Sample Num: 6960, Cur Loss: 0.25603923, Cur Avg Loss: 0.20888625, Log Avg loss: 0.18436558, Global Avg Loss: 0.80232032, Time: 0.0208 Steps: 74950, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000445, Sample Num: 7120, Cur Loss: 0.32298642, Cur Avg Loss: 0.20794608, Log Avg loss: 0.16704872, Global Avg Loss: 0.80223557, Time: 0.0208 Steps: 74960, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000455, Sample Num: 7280, Cur Loss: 0.26055562, Cur Avg Loss: 0.20838633, Log Avg loss: 0.22797767, Global Avg Loss: 0.80215897, Time: 0.0220 Steps: 74970, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000465, Sample Num: 7440, Cur Loss: 0.05351534, Cur Avg Loss: 0.20790385, Log Avg loss: 0.18595105, Global Avg Loss: 0.80207679, Time: 0.0208 Steps: 74980, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000475, Sample Num: 7600, Cur Loss: 0.10872129, Cur Avg Loss: 0.20955601, Log Avg loss: 0.28638116, Global Avg Loss: 0.80200802, Time: 0.0208 Steps: 74990, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000485, Sample Num: 7760, Cur Loss: 0.26043350, Cur Avg Loss: 0.21015225, Log Avg loss: 0.23847362, Global Avg Loss: 0.80193288, Time: 0.0208 Steps: 75000, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000495, Sample Num: 7920, Cur Loss: 0.35574660, Cur Avg Loss: 0.21120844, Log Avg loss: 0.26243391, Global Avg Loss: 0.80186096, Time: 0.0208 Steps: 75010, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000505, Sample Num: 8080, Cur Loss: 0.21091180, Cur Avg Loss: 0.21097617, Log Avg loss: 0.19947870, Global Avg Loss: 0.80178067, Time: 0.0212 Steps: 75020, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000515, Sample Num: 8240, Cur Loss: 0.27647066, Cur Avg Loss: 0.21200768, Log Avg loss: 0.26409891, Global Avg Loss: 0.80170900, Time: 0.0244 Steps: 75030, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000525, Sample Num: 8400, Cur Loss: 0.31702909, Cur Avg Loss: 0.21162276, Log Avg loss: 0.19179917, Global Avg Loss: 0.80162772, Time: 0.0208 Steps: 75040, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000535, Sample Num: 8560, Cur Loss: 0.30930516, Cur Avg Loss: 0.21058821, Log Avg loss: 0.15627477, Global Avg Loss: 0.80154174, Time: 0.0208 Steps: 75050, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000545, Sample Num: 8720, Cur Loss: 0.30005127, Cur Avg Loss: 0.21062022, Log Avg loss: 0.21233283, Global Avg Loss: 0.80146324, Time: 0.0209 Steps: 75060, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000555, Sample Num: 8880, Cur Loss: 0.13626975, Cur Avg Loss: 0.21028714, Log Avg loss: 0.19213423, Global Avg Loss: 0.80138207, Time: 0.0208 Steps: 75070, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000565, Sample Num: 9040, Cur Loss: 0.23650782, Cur Avg Loss: 0.20957232, Log Avg loss: 0.16989982, Global Avg Loss: 0.80129796, Time: 0.0208 Steps: 75080, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000575, Sample Num: 9200, Cur Loss: 0.09192340, Cur Avg Loss: 0.20859149, Log Avg loss: 0.15317421, Global Avg Loss: 0.80121165, Time: 0.0208 Steps: 75090, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000585, Sample Num: 9360, Cur Loss: 0.08817218, Cur Avg Loss: 0.20893583, Log Avg loss: 0.22873548, Global Avg Loss: 0.80113542, Time: 0.0209 Steps: 75100, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000595, Sample Num: 9520, Cur Loss: 0.14783143, Cur Avg Loss: 0.20904871, Log Avg loss: 0.21565197, Global Avg Loss: 0.80105747, Time: 0.0208 Steps: 75110, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000605, Sample Num: 9680, Cur Loss: 0.09975655, Cur Avg Loss: 0.20905023, Log Avg loss: 0.20914074, Global Avg Loss: 0.80097867, Time: 0.0208 Steps: 75120, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000615, Sample Num: 9840, Cur Loss: 0.20921561, Cur Avg Loss: 0.20853121, Log Avg loss: 0.17713062, Global Avg Loss: 0.80089564, Time: 0.0208 Steps: 75130, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000625, Sample Num: 10000, Cur Loss: 0.33636996, Cur Avg Loss: 0.20904044, Log Avg loss: 0.24035813, Global Avg Loss: 0.80082104, Time: 0.0208 Steps: 75140, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000635, Sample Num: 10160, Cur Loss: 0.06336594, Cur Avg Loss: 0.20822876, Log Avg loss: 0.15749887, Global Avg Loss: 0.80073543, Time: 0.0208 Steps: 75150, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000645, Sample Num: 10320, Cur Loss: 0.26461983, Cur Avg Loss: 0.20855249, Log Avg loss: 0.22910938, Global Avg Loss: 0.80065938, Time: 0.0209 Steps: 75160, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000655, Sample Num: 10480, Cur Loss: 0.54264688, Cur Avg Loss: 0.20921258, Log Avg loss: 0.25178829, Global Avg Loss: 0.80058636, Time: 0.0208 Steps: 75170, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000665, Sample Num: 10640, Cur Loss: 0.31792748, Cur Avg Loss: 0.20935800, Log Avg loss: 0.21888267, Global Avg Loss: 0.80050899, Time: 0.0208 Steps: 75180, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000675, Sample Num: 10800, Cur Loss: 0.39696932, Cur Avg Loss: 0.21053920, Log Avg loss: 0.28908957, Global Avg Loss: 0.80044097, Time: 0.0208 Steps: 75190, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000685, Sample Num: 10960, Cur Loss: 0.18641955, Cur Avg Loss: 0.21061667, Log Avg loss: 0.21584594, Global Avg Loss: 0.80036323, Time: 0.0208 Steps: 75200, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000695, Sample Num: 11120, Cur Loss: 0.24151325, Cur Avg Loss: 0.20970110, Log Avg loss: 0.14698408, Global Avg Loss: 0.80027636, Time: 0.0208 Steps: 75210, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000705, Sample Num: 11280, Cur Loss: 0.05765194, Cur Avg Loss: 0.20946221, Log Avg loss: 0.19285957, Global Avg Loss: 0.80019560, Time: 0.0208 Steps: 75220, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000715, Sample Num: 11440, Cur Loss: 0.14647508, Cur Avg Loss: 0.20946065, Log Avg loss: 0.20935090, Global Avg Loss: 0.80011707, Time: 0.0209 Steps: 75230, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000725, Sample Num: 11600, Cur Loss: 0.57921976, Cur Avg Loss: 0.20952864, Log Avg loss: 0.21438954, Global Avg Loss: 0.80003922, Time: 0.0209 Steps: 75240, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000735, Sample Num: 11760, Cur Loss: 0.09189847, Cur Avg Loss: 0.20842084, Log Avg loss: 0.12810544, Global Avg Loss: 0.79994992, Time: 0.0208 Steps: 75250, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000745, Sample Num: 11920, Cur Loss: 0.13767076, Cur Avg Loss: 0.20905997, Log Avg loss: 0.25603627, Global Avg Loss: 0.79987765, Time: 0.0208 Steps: 75260, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000755, Sample Num: 12080, Cur Loss: 0.07981750, Cur Avg Loss: 0.20874326, Log Avg loss: 0.18514840, Global Avg Loss: 0.79979598, Time: 0.0208 Steps: 75270, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000765, Sample Num: 12240, Cur Loss: 0.17212272, Cur Avg Loss: 0.20821826, Log Avg loss: 0.16858050, Global Avg Loss: 0.79971213, Time: 0.0207 Steps: 75280, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000775, Sample Num: 12400, Cur Loss: 0.18941870, Cur Avg Loss: 0.20858685, Log Avg loss: 0.23678387, Global Avg Loss: 0.79963737, Time: 0.0227 Steps: 75290, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000785, Sample Num: 12560, Cur Loss: 0.23572791, Cur Avg Loss: 0.20830112, Log Avg loss: 0.18615698, Global Avg Loss: 0.79955590, Time: 0.0210 Steps: 75300, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000795, Sample Num: 12720, Cur Loss: 0.30579823, Cur Avg Loss: 0.20902470, Log Avg loss: 0.26582558, Global Avg Loss: 0.79948502, Time: 0.0209 Steps: 75310, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000805, Sample Num: 12880, Cur Loss: 0.18135308, Cur Avg Loss: 0.21044734, Log Avg loss: 0.32354774, Global Avg Loss: 0.79942184, Time: 0.0209 Steps: 75320, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000815, Sample Num: 13040, Cur Loss: 0.53964520, Cur Avg Loss: 0.21171653, Log Avg loss: 0.31388606, Global Avg Loss: 0.79935738, Time: 0.0209 Steps: 75330, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000825, Sample Num: 13200, Cur Loss: 0.29232800, Cur Avg Loss: 0.21117122, Log Avg loss: 0.16672815, Global Avg Loss: 0.79927341, Time: 0.0209 Steps: 75340, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000835, Sample Num: 13360, Cur Loss: 0.17527188, Cur Avg Loss: 0.21267913, Log Avg loss: 0.33708218, Global Avg Loss: 0.79921207, Time: 0.0209 Steps: 75350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000845, Sample Num: 13520, Cur Loss: 0.13255204, Cur Avg Loss: 0.21256987, Log Avg loss: 0.20344675, Global Avg Loss: 0.79913302, Time: 0.0209 Steps: 75360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000855, Sample Num: 13680, Cur Loss: 0.17210624, Cur Avg Loss: 0.21206484, Log Avg loss: 0.16938958, Global Avg Loss: 0.79904946, Time: 0.0209 Steps: 75370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000865, Sample Num: 13840, Cur Loss: 0.30148441, Cur Avg Loss: 0.21190472, Log Avg loss: 0.19821426, Global Avg Loss: 0.79896975, Time: 0.0209 Steps: 75380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000875, Sample Num: 14000, Cur Loss: 0.15428337, Cur Avg Loss: 0.21092808, Log Avg loss: 0.12644932, Global Avg Loss: 0.79888055, Time: 0.0209 Steps: 75390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000885, Sample Num: 14160, Cur Loss: 0.42102444, Cur Avg Loss: 0.21169975, Log Avg loss: 0.27922091, Global Avg Loss: 0.79881163, Time: 0.0209 Steps: 75400, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000895, Sample Num: 14320, Cur Loss: 0.18833753, Cur Avg Loss: 0.21216328, Log Avg loss: 0.25318530, Global Avg Loss: 0.79873927, Time: 0.0209 Steps: 75410, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000905, Sample Num: 14480, Cur Loss: 0.18684831, Cur Avg Loss: 0.21159334, Log Avg loss: 0.16058373, Global Avg Loss: 0.79865466, Time: 0.0209 Steps: 75420, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000915, Sample Num: 14640, Cur Loss: 0.24702728, Cur Avg Loss: 0.21140544, Log Avg loss: 0.19440047, Global Avg Loss: 0.79857455, Time: 0.0209 Steps: 75430, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000925, Sample Num: 14800, Cur Loss: 0.13055536, Cur Avg Loss: 0.21089817, Log Avg loss: 0.16448262, Global Avg Loss: 0.79849050, Time: 0.0209 Steps: 75440, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000935, Sample Num: 14960, Cur Loss: 0.28863770, Cur Avg Loss: 0.21106172, Log Avg loss: 0.22619076, Global Avg Loss: 0.79841465, Time: 0.0209 Steps: 75450, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000945, Sample Num: 15120, Cur Loss: 0.52151740, Cur Avg Loss: 0.21116847, Log Avg loss: 0.22114885, Global Avg Loss: 0.79833815, Time: 0.0208 Steps: 75460, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000955, Sample Num: 15280, Cur Loss: 0.22282666, Cur Avg Loss: 0.21132046, Log Avg loss: 0.22568432, Global Avg Loss: 0.79826227, Time: 0.0209 Steps: 75470, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000965, Sample Num: 15440, Cur Loss: 0.30670339, Cur Avg Loss: 0.21128455, Log Avg loss: 0.20785466, Global Avg Loss: 0.79818405, Time: 0.0209 Steps: 75480, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000975, Sample Num: 15600, Cur Loss: 0.69368595, Cur Avg Loss: 0.21201519, Log Avg loss: 0.28252159, Global Avg Loss: 0.79811574, Time: 0.0209 Steps: 75490, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000985, Sample Num: 15760, Cur Loss: 0.27012795, Cur Avg Loss: 0.21249219, Log Avg loss: 0.25900007, Global Avg Loss: 0.79804434, Time: 0.0209 Steps: 75500, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000995, Sample Num: 15920, Cur Loss: 0.27355966, Cur Avg Loss: 0.21344752, Log Avg loss: 0.30754810, Global Avg Loss: 0.79797938, Time: 0.0209 Steps: 75510, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001005, Sample Num: 16080, Cur Loss: 0.80796868, Cur Avg Loss: 0.21454960, Log Avg loss: 0.32420649, Global Avg Loss: 0.79791664, Time: 0.0209 Steps: 75520, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001015, Sample Num: 16240, Cur Loss: 0.19865838, Cur Avg Loss: 0.21432251, Log Avg loss: 0.19149953, Global Avg Loss: 0.79783635, Time: 0.0208 Steps: 75530, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001025, Sample Num: 16400, Cur Loss: 0.25164768, Cur Avg Loss: 0.21464115, Log Avg loss: 0.24698318, Global Avg Loss: 0.79776343, Time: 0.0246 Steps: 75540, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001035, Sample Num: 16560, Cur Loss: 0.23703808, Cur Avg Loss: 0.21456684, Log Avg loss: 0.20694990, Global Avg Loss: 0.79768523, Time: 0.0209 Steps: 75550, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001045, Sample Num: 16720, Cur Loss: 0.22589153, Cur Avg Loss: 0.21510418, Log Avg loss: 0.27071903, Global Avg Loss: 0.79761549, Time: 0.0209 Steps: 75560, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001055, Sample Num: 16880, Cur Loss: 0.17184877, Cur Avg Loss: 0.21490318, Log Avg loss: 0.19389811, Global Avg Loss: 0.79753560, Time: 0.0209 Steps: 75570, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001065, Sample Num: 17040, Cur Loss: 0.26486623, Cur Avg Loss: 0.21421186, Log Avg loss: 0.14127844, Global Avg Loss: 0.79744877, Time: 0.0209 Steps: 75580, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001075, Sample Num: 17200, Cur Loss: 0.15205446, Cur Avg Loss: 0.21456334, Log Avg loss: 0.25199516, Global Avg Loss: 0.79737661, Time: 0.0208 Steps: 75590, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001085, Sample Num: 17360, Cur Loss: 0.20078550, Cur Avg Loss: 0.21495250, Log Avg loss: 0.25678744, Global Avg Loss: 0.79730511, Time: 0.0209 Steps: 75600, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001095, Sample Num: 17520, Cur Loss: 0.08834893, Cur Avg Loss: 0.21511094, Log Avg loss: 0.23230223, Global Avg Loss: 0.79723038, Time: 0.0209 Steps: 75610, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001105, Sample Num: 17680, Cur Loss: 0.25180584, Cur Avg Loss: 0.21528903, Log Avg loss: 0.23478994, Global Avg Loss: 0.79715600, Time: 0.0209 Steps: 75620, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001115, Sample Num: 17840, Cur Loss: 0.16951998, Cur Avg Loss: 0.21529467, Log Avg loss: 0.21591800, Global Avg Loss: 0.79707915, Time: 0.0209 Steps: 75630, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001125, Sample Num: 18000, Cur Loss: 0.27076635, Cur Avg Loss: 0.21531390, Log Avg loss: 0.21745774, Global Avg Loss: 0.79700252, Time: 0.0209 Steps: 75640, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001135, Sample Num: 18160, Cur Loss: 0.07984310, Cur Avg Loss: 0.21512077, Log Avg loss: 0.19339349, Global Avg Loss: 0.79692273, Time: 0.0209 Steps: 75650, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001145, Sample Num: 18320, Cur Loss: 0.27491364, Cur Avg Loss: 0.21605158, Log Avg loss: 0.32169841, Global Avg Loss: 0.79685992, Time: 0.0209 Steps: 75660, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001155, Sample Num: 18480, Cur Loss: 0.19977316, Cur Avg Loss: 0.21540881, Log Avg loss: 0.14181122, Global Avg Loss: 0.79677335, Time: 0.0209 Steps: 75670, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001165, Sample Num: 18640, Cur Loss: 0.03825035, Cur Avg Loss: 0.21515145, Log Avg loss: 0.18542696, Global Avg Loss: 0.79669257, Time: 0.0209 Steps: 75680, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001175, Sample Num: 18800, Cur Loss: 0.20091447, Cur Avg Loss: 0.21461249, Log Avg loss: 0.15182342, Global Avg Loss: 0.79660737, Time: 0.0209 Steps: 75690, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001185, Sample Num: 18960, Cur Loss: 0.23166263, Cur Avg Loss: 0.21437020, Log Avg loss: 0.18590162, Global Avg Loss: 0.79652670, Time: 0.0209 Steps: 75700, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001195, Sample Num: 19120, Cur Loss: 0.09113315, Cur Avg Loss: 0.21378231, Log Avg loss: 0.14411695, Global Avg Loss: 0.79644053, Time: 0.0209 Steps: 75710, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001205, Sample Num: 19280, Cur Loss: 0.27474427, Cur Avg Loss: 0.21371896, Log Avg loss: 0.20614886, Global Avg Loss: 0.79636257, Time: 0.0209 Steps: 75720, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001215, Sample Num: 19440, Cur Loss: 0.08477950, Cur Avg Loss: 0.21349381, Log Avg loss: 0.18636257, Global Avg Loss: 0.79628202, Time: 0.0209 Steps: 75730, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001225, Sample Num: 19600, Cur Loss: 0.16633973, Cur Avg Loss: 0.21403721, Log Avg loss: 0.28006038, Global Avg Loss: 0.79621386, Time: 0.0209 Steps: 75740, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001235, Sample Num: 19760, Cur Loss: 0.26139897, Cur Avg Loss: 0.21434236, Log Avg loss: 0.25172389, Global Avg Loss: 0.79614198, Time: 0.0209 Steps: 75750, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001245, Sample Num: 19920, Cur Loss: 0.10649827, Cur Avg Loss: 0.21412885, Log Avg loss: 0.18776051, Global Avg Loss: 0.79606168, Time: 0.0209 Steps: 75760, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001255, Sample Num: 20080, Cur Loss: 0.08774883, Cur Avg Loss: 0.21455133, Log Avg loss: 0.26715022, Global Avg Loss: 0.79599188, Time: 0.0209 Steps: 75770, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001265, Sample Num: 20240, Cur Loss: 0.24252662, Cur Avg Loss: 0.21484171, Log Avg loss: 0.25128403, Global Avg Loss: 0.79592000, Time: 0.0209 Steps: 75780, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001275, Sample Num: 20400, Cur Loss: 0.86848021, Cur Avg Loss: 0.21558339, Log Avg loss: 0.30940628, Global Avg Loss: 0.79585580, Time: 0.0209 Steps: 75790, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001285, Sample Num: 20560, Cur Loss: 0.46226084, Cur Avg Loss: 0.21622621, Log Avg loss: 0.29818580, Global Avg Loss: 0.79579015, Time: 0.0246 Steps: 75800, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001295, Sample Num: 20720, Cur Loss: 0.18308696, Cur Avg Loss: 0.21675957, Log Avg loss: 0.28529595, Global Avg Loss: 0.79572281, Time: 0.0209 Steps: 75810, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001305, Sample Num: 20880, Cur Loss: 0.32583994, Cur Avg Loss: 0.21679301, Log Avg loss: 0.22112392, Global Avg Loss: 0.79564702, Time: 0.0209 Steps: 75820, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001315, Sample Num: 21040, Cur Loss: 0.29777151, Cur Avg Loss: 0.21700337, Log Avg loss: 0.24445508, Global Avg Loss: 0.79557434, Time: 0.0209 Steps: 75830, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001325, Sample Num: 21200, Cur Loss: 0.08849412, Cur Avg Loss: 0.21667410, Log Avg loss: 0.17337416, Global Avg Loss: 0.79549230, Time: 0.0208 Steps: 75840, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001335, Sample Num: 21360, Cur Loss: 0.83589667, Cur Avg Loss: 0.21675732, Log Avg loss: 0.22778428, Global Avg Loss: 0.79541745, Time: 0.0209 Steps: 75850, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001345, Sample Num: 21520, Cur Loss: 0.02310077, Cur Avg Loss: 0.21644586, Log Avg loss: 0.17486639, Global Avg Loss: 0.79533565, Time: 0.0209 Steps: 75860, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001355, Sample Num: 21680, Cur Loss: 0.09161223, Cur Avg Loss: 0.21631520, Log Avg loss: 0.19874122, Global Avg Loss: 0.79525701, Time: 0.0209 Steps: 75870, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001365, Sample Num: 21840, Cur Loss: 0.08613364, Cur Avg Loss: 0.21572743, Log Avg loss: 0.13608455, Global Avg Loss: 0.79517014, Time: 0.0210 Steps: 75880, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001375, Sample Num: 22000, Cur Loss: 0.06899953, Cur Avg Loss: 0.21544552, Log Avg loss: 0.17696429, Global Avg Loss: 0.79508868, Time: 0.0209 Steps: 75890, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001385, Sample Num: 22160, Cur Loss: 0.11852249, Cur Avg Loss: 0.21571629, Log Avg loss: 0.25294773, Global Avg Loss: 0.79501725, Time: 0.0209 Steps: 75900, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001395, Sample Num: 22320, Cur Loss: 0.19921336, Cur Avg Loss: 0.21550203, Log Avg loss: 0.18582657, Global Avg Loss: 0.79493700, Time: 0.0209 Steps: 75910, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001405, Sample Num: 22480, Cur Loss: 0.12426206, Cur Avg Loss: 0.21614935, Log Avg loss: 0.30645050, Global Avg Loss: 0.79487266, Time: 0.0209 Steps: 75920, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001415, Sample Num: 22640, Cur Loss: 0.17055367, Cur Avg Loss: 0.21621097, Log Avg loss: 0.22486903, Global Avg Loss: 0.79479759, Time: 0.0208 Steps: 75930, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001425, Sample Num: 22800, Cur Loss: 0.17496507, Cur Avg Loss: 0.21594208, Log Avg loss: 0.17789484, Global Avg Loss: 0.79471636, Time: 0.0209 Steps: 75940, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001435, Sample Num: 22960, Cur Loss: 0.42423406, Cur Avg Loss: 0.21634861, Log Avg loss: 0.27427830, Global Avg Loss: 0.79464783, Time: 0.0209 Steps: 75950, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001445, Sample Num: 23120, Cur Loss: 0.11191517, Cur Avg Loss: 0.21695389, Log Avg loss: 0.30381136, Global Avg Loss: 0.79458321, Time: 0.0208 Steps: 75960, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001455, Sample Num: 23280, Cur Loss: 0.15955308, Cur Avg Loss: 0.21732568, Log Avg loss: 0.27104998, Global Avg Loss: 0.79451430, Time: 0.0209 Steps: 75970, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001465, Sample Num: 23440, Cur Loss: 0.44568196, Cur Avg Loss: 0.21710240, Log Avg loss: 0.18461497, Global Avg Loss: 0.79443403, Time: 0.0209 Steps: 75980, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001475, Sample Num: 23600, Cur Loss: 0.17224509, Cur Avg Loss: 0.21750347, Log Avg loss: 0.27625975, Global Avg Loss: 0.79436584, Time: 0.0209 Steps: 75990, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001485, Sample Num: 23760, Cur Loss: 0.32782078, Cur Avg Loss: 0.21735359, Log Avg loss: 0.19524665, Global Avg Loss: 0.79428701, Time: 0.0209 Steps: 76000, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001495, Sample Num: 23920, Cur Loss: 0.34201035, Cur Avg Loss: 0.21713797, Log Avg loss: 0.18511876, Global Avg Loss: 0.79420686, Time: 0.0208 Steps: 76010, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001505, Sample Num: 24080, Cur Loss: 0.07248344, Cur Avg Loss: 0.21673018, Log Avg loss: 0.15576519, Global Avg Loss: 0.79412288, Time: 0.0209 Steps: 76020, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001515, Sample Num: 24240, Cur Loss: 0.11146824, Cur Avg Loss: 0.21672029, Log Avg loss: 0.21523172, Global Avg Loss: 0.79404674, Time: 0.0209 Steps: 76030, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001525, Sample Num: 24400, Cur Loss: 0.14784321, Cur Avg Loss: 0.21681099, Log Avg loss: 0.23055260, Global Avg Loss: 0.79397264, Time: 0.0209 Steps: 76040, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001535, Sample Num: 24560, Cur Loss: 0.34601831, Cur Avg Loss: 0.21640329, Log Avg loss: 0.15422930, Global Avg Loss: 0.79388852, Time: 0.0209 Steps: 76050, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001545, Sample Num: 24720, Cur Loss: 0.19304460, Cur Avg Loss: 0.21613718, Log Avg loss: 0.17528849, Global Avg Loss: 0.79380718, Time: 0.0209 Steps: 76060, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001555, Sample Num: 24880, Cur Loss: 0.18092629, Cur Avg Loss: 0.21590588, Log Avg loss: 0.18016970, Global Avg Loss: 0.79372652, Time: 0.0209 Steps: 76070, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001565, Sample Num: 25040, Cur Loss: 0.12340476, Cur Avg Loss: 0.21570688, Log Avg loss: 0.18476214, Global Avg Loss: 0.79364647, Time: 0.0209 Steps: 76080, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001575, Sample Num: 25200, Cur Loss: 0.09709343, Cur Avg Loss: 0.21501649, Log Avg loss: 0.10697190, Global Avg Loss: 0.79355623, Time: 0.0209 Steps: 76090, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001585, Sample Num: 25360, Cur Loss: 0.13128144, Cur Avg Loss: 0.21470928, Log Avg loss: 0.16632252, Global Avg Loss: 0.79347381, Time: 0.0209 Steps: 76100, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001595, Sample Num: 25520, Cur Loss: 0.24058981, Cur Avg Loss: 0.21488233, Log Avg loss: 0.24231157, Global Avg Loss: 0.79340139, Time: 0.0209 Steps: 76110, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001605, Sample Num: 25680, Cur Loss: 0.42150968, Cur Avg Loss: 0.21541689, Log Avg loss: 0.30067849, Global Avg Loss: 0.79333666, Time: 0.0209 Steps: 76120, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001615, Sample Num: 25840, Cur Loss: 0.09160943, Cur Avg Loss: 0.21507756, Log Avg loss: 0.16061555, Global Avg Loss: 0.79325355, Time: 0.0209 Steps: 76130, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001625, Sample Num: 26000, Cur Loss: 0.36979216, Cur Avg Loss: 0.21500049, Log Avg loss: 0.20255295, Global Avg Loss: 0.79317597, Time: 0.0209 Steps: 76140, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001635, Sample Num: 26160, Cur Loss: 0.14095899, Cur Avg Loss: 0.21465016, Log Avg loss: 0.15772233, Global Avg Loss: 0.79309252, Time: 0.0209 Steps: 76150, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001645, Sample Num: 26320, Cur Loss: 0.07985563, Cur Avg Loss: 0.21456492, Log Avg loss: 0.20062731, Global Avg Loss: 0.79301473, Time: 0.0209 Steps: 76160, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001655, Sample Num: 26480, Cur Loss: 0.19278139, Cur Avg Loss: 0.21453921, Log Avg loss: 0.21031057, Global Avg Loss: 0.79293823, Time: 0.0209 Steps: 76170, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001665, Sample Num: 26640, Cur Loss: 0.31505984, Cur Avg Loss: 0.21468740, Log Avg loss: 0.23921273, Global Avg Loss: 0.79286554, Time: 0.0209 Steps: 76180, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001675, Sample Num: 26800, Cur Loss: 0.17122003, Cur Avg Loss: 0.21482524, Log Avg loss: 0.23777658, Global Avg Loss: 0.79279269, Time: 0.0209 Steps: 76190, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001685, Sample Num: 26960, Cur Loss: 0.27421141, Cur Avg Loss: 0.21472103, Log Avg loss: 0.19726487, Global Avg Loss: 0.79271453, Time: 0.0209 Steps: 76200, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001695, Sample Num: 27120, Cur Loss: 0.32574862, Cur Avg Loss: 0.21487402, Log Avg loss: 0.24065357, Global Avg Loss: 0.79264209, Time: 0.0209 Steps: 76210, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001705, Sample Num: 27280, Cur Loss: 0.58219379, Cur Avg Loss: 0.21510185, Log Avg loss: 0.25371797, Global Avg Loss: 0.79257139, Time: 0.0209 Steps: 76220, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001715, Sample Num: 27440, Cur Loss: 0.12394824, Cur Avg Loss: 0.21550147, Log Avg loss: 0.28363638, Global Avg Loss: 0.79250462, Time: 0.0209 Steps: 76230, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001725, Sample Num: 27600, Cur Loss: 0.27619380, Cur Avg Loss: 0.21559486, Log Avg loss: 0.23161258, Global Avg Loss: 0.79243106, Time: 0.0209 Steps: 76240, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001735, Sample Num: 27760, Cur Loss: 0.11732603, Cur Avg Loss: 0.21536606, Log Avg loss: 0.17589818, Global Avg Loss: 0.79235020, Time: 0.0209 Steps: 76250, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001745, Sample Num: 27920, Cur Loss: 0.08855185, Cur Avg Loss: 0.21524088, Log Avg loss: 0.19352095, Global Avg Loss: 0.79227167, Time: 0.0209 Steps: 76260, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001755, Sample Num: 28080, Cur Loss: 0.33048841, Cur Avg Loss: 0.21511020, Log Avg loss: 0.19230681, Global Avg Loss: 0.79219301, Time: 0.0210 Steps: 76270, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001765, Sample Num: 28240, Cur Loss: 0.12658529, Cur Avg Loss: 0.21492397, Log Avg loss: 0.18224060, Global Avg Loss: 0.79211305, Time: 0.0209 Steps: 76280, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001775, Sample Num: 28400, Cur Loss: 0.09324838, Cur Avg Loss: 0.21495090, Log Avg loss: 0.21970450, Global Avg Loss: 0.79203802, Time: 0.0209 Steps: 76290, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001785, Sample Num: 28560, Cur Loss: 0.11534292, Cur Avg Loss: 0.21468065, Log Avg loss: 0.16671082, Global Avg Loss: 0.79195606, Time: 0.0209 Steps: 76300, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001795, Sample Num: 28720, Cur Loss: 0.15581337, Cur Avg Loss: 0.21461950, Log Avg loss: 0.20370404, Global Avg Loss: 0.79187897, Time: 0.0244 Steps: 76310, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001805, Sample Num: 28880, Cur Loss: 0.19489071, Cur Avg Loss: 0.21533413, Log Avg loss: 0.34361076, Global Avg Loss: 0.79182024, Time: 0.0208 Steps: 76320, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001815, Sample Num: 29040, Cur Loss: 0.12576245, Cur Avg Loss: 0.21511292, Log Avg loss: 0.17518374, Global Avg Loss: 0.79173945, Time: 0.0207 Steps: 76330, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001825, Sample Num: 29200, Cur Loss: 0.12429313, Cur Avg Loss: 0.21539927, Log Avg loss: 0.26737230, Global Avg Loss: 0.79167076, Time: 0.0207 Steps: 76340, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001835, Sample Num: 29360, Cur Loss: 0.17860232, Cur Avg Loss: 0.21581302, Log Avg loss: 0.29132197, Global Avg Loss: 0.79160523, Time: 0.0207 Steps: 76350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001845, Sample Num: 29520, Cur Loss: 0.25185904, Cur Avg Loss: 0.21635284, Log Avg loss: 0.31541040, Global Avg Loss: 0.79154287, Time: 0.0207 Steps: 76360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001855, Sample Num: 29680, Cur Loss: 0.22388442, Cur Avg Loss: 0.21721264, Log Avg loss: 0.37584555, Global Avg Loss: 0.79148844, Time: 0.0207 Steps: 76370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001865, Sample Num: 29840, Cur Loss: 0.22918421, Cur Avg Loss: 0.21767433, Log Avg loss: 0.30331887, Global Avg Loss: 0.79142452, Time: 0.0208 Steps: 76380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001875, Sample Num: 30000, Cur Loss: 0.13819718, Cur Avg Loss: 0.21789168, Log Avg loss: 0.25842653, Global Avg Loss: 0.79135475, Time: 0.0208 Steps: 76390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001885, Sample Num: 30160, Cur Loss: 0.27743384, Cur Avg Loss: 0.21812573, Log Avg loss: 0.26201062, Global Avg Loss: 0.79128547, Time: 0.0207 Steps: 76400, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001895, Sample Num: 30320, Cur Loss: 0.32591310, Cur Avg Loss: 0.21805207, Log Avg loss: 0.20416652, Global Avg Loss: 0.79120863, Time: 0.0208 Steps: 76410, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001905, Sample Num: 30480, Cur Loss: 0.12792774, Cur Avg Loss: 0.21791435, Log Avg loss: 0.19181668, Global Avg Loss: 0.79113019, Time: 0.0207 Steps: 76420, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001915, Sample Num: 30640, Cur Loss: 0.14098081, Cur Avg Loss: 0.21765263, Log Avg loss: 0.16779491, Global Avg Loss: 0.79104864, Time: 0.0208 Steps: 76430, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001925, Sample Num: 30800, Cur Loss: 0.35328797, Cur Avg Loss: 0.21769467, Log Avg loss: 0.22574612, Global Avg Loss: 0.79097468, Time: 0.0208 Steps: 76440, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001935, Sample Num: 30960, Cur Loss: 0.41120684, Cur Avg Loss: 0.21776386, Log Avg loss: 0.23108313, Global Avg Loss: 0.79090145, Time: 0.0208 Steps: 76450, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001945, Sample Num: 31120, Cur Loss: 0.15171427, Cur Avg Loss: 0.21756492, Log Avg loss: 0.17906986, Global Avg Loss: 0.79082143, Time: 0.0208 Steps: 76460, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001955, Sample Num: 31280, Cur Loss: 0.43553710, Cur Avg Loss: 0.21804016, Log Avg loss: 0.31047333, Global Avg Loss: 0.79075861, Time: 0.0207 Steps: 76470, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001965, Sample Num: 31440, Cur Loss: 0.52421296, Cur Avg Loss: 0.21873610, Log Avg loss: 0.35479261, Global Avg Loss: 0.79070161, Time: 0.0207 Steps: 76480, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001975, Sample Num: 31600, Cur Loss: 0.19294403, Cur Avg Loss: 0.21888268, Log Avg loss: 0.24768569, Global Avg Loss: 0.79063062, Time: 0.0208 Steps: 76490, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001985, Sample Num: 31760, Cur Loss: 0.21311709, Cur Avg Loss: 0.21924586, Log Avg loss: 0.29097314, Global Avg Loss: 0.79056530, Time: 0.0208 Steps: 76500, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001995, Sample Num: 31920, Cur Loss: 0.05391614, Cur Avg Loss: 0.21881216, Log Avg loss: 0.13272281, Global Avg Loss: 0.79047932, Time: 0.0207 Steps: 76510, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002005, Sample Num: 32080, Cur Loss: 0.23905425, Cur Avg Loss: 0.21888401, Log Avg loss: 0.23321865, Global Avg Loss: 0.79040649, Time: 0.0208 Steps: 76520, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002015, Sample Num: 32240, Cur Loss: 0.10077301, Cur Avg Loss: 0.21871274, Log Avg loss: 0.18437377, Global Avg Loss: 0.79032731, Time: 0.0208 Steps: 76530, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002025, Sample Num: 32400, Cur Loss: 0.19554895, Cur Avg Loss: 0.21865371, Log Avg loss: 0.20675844, Global Avg Loss: 0.79025106, Time: 0.0208 Steps: 76540, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002035, Sample Num: 32560, Cur Loss: 0.10121568, Cur Avg Loss: 0.21853406, Log Avg loss: 0.19430455, Global Avg Loss: 0.79017321, Time: 0.0207 Steps: 76550, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002045, Sample Num: 32720, Cur Loss: 0.19149581, Cur Avg Loss: 0.21839794, Log Avg loss: 0.19069875, Global Avg Loss: 0.79009491, Time: 0.0208 Steps: 76560, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002055, Sample Num: 32880, Cur Loss: 0.34591013, Cur Avg Loss: 0.21801845, Log Avg loss: 0.14041191, Global Avg Loss: 0.79001006, Time: 0.0209 Steps: 76570, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002065, Sample Num: 33040, Cur Loss: 0.15272641, Cur Avg Loss: 0.21787846, Log Avg loss: 0.18911121, Global Avg Loss: 0.78993160, Time: 0.0209 Steps: 76580, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002075, Sample Num: 33200, Cur Loss: 0.22729203, Cur Avg Loss: 0.21828357, Log Avg loss: 0.30193850, Global Avg Loss: 0.78986788, Time: 0.0208 Steps: 76590, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002085, Sample Num: 33360, Cur Loss: 0.07433031, Cur Avg Loss: 0.21823348, Log Avg loss: 0.20783935, Global Avg Loss: 0.78979190, Time: 0.0209 Steps: 76600, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002095, Sample Num: 33520, Cur Loss: 0.08074838, Cur Avg Loss: 0.21791733, Log Avg loss: 0.15199996, Global Avg Loss: 0.78970865, Time: 0.0208 Steps: 76610, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002105, Sample Num: 33680, Cur Loss: 0.26950967, Cur Avg Loss: 0.21805957, Log Avg loss: 0.24785825, Global Avg Loss: 0.78963793, Time: 0.0208 Steps: 76620, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002115, Sample Num: 33840, Cur Loss: 0.18407875, Cur Avg Loss: 0.21774748, Log Avg loss: 0.15205308, Global Avg Loss: 0.78955472, Time: 0.0209 Steps: 76630, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002125, Sample Num: 34000, Cur Loss: 0.01874008, Cur Avg Loss: 0.21733525, Log Avg loss: 0.13014918, Global Avg Loss: 0.78946868, Time: 0.0209 Steps: 76640, Updated lr: 0.000028 ***** Running evaluation checkpoint-76644 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-76644 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.730258, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.261531, "eval_total_loss": 183.856299, "eval_mae": 0.376311, "eval_mse": 0.261578, "eval_r2": 0.833724, "eval_sp_statistic": 0.890563, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.919987, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.298235, "test_total_loss": 149.713848, "test_mae": 0.346591, "test_mse": 0.298342, "test_r2": 0.807447, "test_sp_statistic": 0.870978, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.922217, "test_ps_pvalue": 0.0, "lr": 2.826552868658132e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7894357493479519, "train_cur_epoch_loss": 462.4710383452475, "train_cur_epoch_avg_loss": 0.21722453656423085, "train_cur_epoch_time": 44.730257749557495, "train_cur_epoch_avg_time": 0.021009984851835365, "epoch": 36, "step": 76644} ################################################## Training, Epoch: 0037, Batch: 000006, Sample Num: 96, Cur Loss: 0.09374920, Cur Avg Loss: 0.20819035, Log Avg loss: 0.18827720, Global Avg Loss: 0.78939025, Time: 0.0245 Steps: 76650, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000016, Sample Num: 256, Cur Loss: 0.18256572, Cur Avg Loss: 0.23995355, Log Avg loss: 0.25901147, Global Avg Loss: 0.78932106, Time: 0.0207 Steps: 76660, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000026, Sample Num: 416, Cur Loss: 0.13353731, Cur Avg Loss: 0.21708891, Log Avg loss: 0.18050549, Global Avg Loss: 0.78924166, Time: 0.0208 Steps: 76670, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000036, Sample Num: 576, Cur Loss: 0.20189781, Cur Avg Loss: 0.22286901, Log Avg loss: 0.23789726, Global Avg Loss: 0.78916976, Time: 0.0207 Steps: 76680, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000046, Sample Num: 736, Cur Loss: 0.16303098, Cur Avg Loss: 0.22735914, Log Avg loss: 0.24352362, Global Avg Loss: 0.78909861, Time: 0.0208 Steps: 76690, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000056, Sample Num: 896, Cur Loss: 0.11438845, Cur Avg Loss: 0.22498308, Log Avg loss: 0.21405319, Global Avg Loss: 0.78902363, Time: 0.0207 Steps: 76700, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000066, Sample Num: 1056, Cur Loss: 0.10109994, Cur Avg Loss: 0.21780278, Log Avg loss: 0.17759308, Global Avg Loss: 0.78894393, Time: 0.0207 Steps: 76710, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000076, Sample Num: 1216, Cur Loss: 0.27608976, Cur Avg Loss: 0.20555296, Log Avg loss: 0.12470416, Global Avg Loss: 0.78885735, Time: 0.0207 Steps: 76720, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000086, Sample Num: 1376, Cur Loss: 0.08573548, Cur Avg Loss: 0.20037675, Log Avg loss: 0.16103752, Global Avg Loss: 0.78877552, Time: 0.0207 Steps: 76730, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000096, Sample Num: 1536, Cur Loss: 0.04615390, Cur Avg Loss: 0.19583660, Log Avg loss: 0.15679137, Global Avg Loss: 0.78869317, Time: 0.0208 Steps: 76740, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000106, Sample Num: 1696, Cur Loss: 0.11112110, Cur Avg Loss: 0.19691984, Log Avg loss: 0.20731895, Global Avg Loss: 0.78861742, Time: 0.0207 Steps: 76750, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000116, Sample Num: 1856, Cur Loss: 0.15851377, Cur Avg Loss: 0.19878555, Log Avg loss: 0.21856210, Global Avg Loss: 0.78854316, Time: 0.0208 Steps: 76760, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000126, Sample Num: 2016, Cur Loss: 0.33639494, Cur Avg Loss: 0.20321368, Log Avg loss: 0.25457996, Global Avg Loss: 0.78847360, Time: 0.0207 Steps: 76770, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000136, Sample Num: 2176, Cur Loss: 0.32195330, Cur Avg Loss: 0.20352344, Log Avg loss: 0.20742634, Global Avg Loss: 0.78839793, Time: 0.0207 Steps: 76780, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000146, Sample Num: 2336, Cur Loss: 0.52253544, Cur Avg Loss: 0.20428922, Log Avg loss: 0.21470394, Global Avg Loss: 0.78832322, Time: 0.0208 Steps: 76790, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000156, Sample Num: 2496, Cur Loss: 0.16782497, Cur Avg Loss: 0.20095340, Log Avg loss: 0.15225033, Global Avg Loss: 0.78824039, Time: 0.0207 Steps: 76800, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000166, Sample Num: 2656, Cur Loss: 0.33212045, Cur Avg Loss: 0.20356280, Log Avg loss: 0.24426955, Global Avg Loss: 0.78816957, Time: 0.0207 Steps: 76810, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000176, Sample Num: 2816, Cur Loss: 0.17678490, Cur Avg Loss: 0.20471001, Log Avg loss: 0.22375370, Global Avg Loss: 0.78809610, Time: 0.0207 Steps: 76820, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000186, Sample Num: 2976, Cur Loss: 0.29047388, Cur Avg Loss: 0.20893130, Log Avg loss: 0.28322597, Global Avg Loss: 0.78803039, Time: 0.0208 Steps: 76830, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000196, Sample Num: 3136, Cur Loss: 0.31267262, Cur Avg Loss: 0.20861588, Log Avg loss: 0.20274897, Global Avg Loss: 0.78795422, Time: 0.0207 Steps: 76840, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000206, Sample Num: 3296, Cur Loss: 0.18910801, Cur Avg Loss: 0.20920449, Log Avg loss: 0.22074130, Global Avg Loss: 0.78788041, Time: 0.0207 Steps: 76850, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000216, Sample Num: 3456, Cur Loss: 0.09951852, Cur Avg Loss: 0.21025619, Log Avg loss: 0.23192122, Global Avg Loss: 0.78780808, Time: 0.0208 Steps: 76860, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000226, Sample Num: 3616, Cur Loss: 0.59567887, Cur Avg Loss: 0.21697456, Log Avg loss: 0.36209136, Global Avg Loss: 0.78775270, Time: 0.0208 Steps: 76870, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000236, Sample Num: 3776, Cur Loss: 0.27776843, Cur Avg Loss: 0.21816748, Log Avg loss: 0.24512759, Global Avg Loss: 0.78768212, Time: 0.0208 Steps: 76880, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000246, Sample Num: 3936, Cur Loss: 0.10614237, Cur Avg Loss: 0.21483430, Log Avg loss: 0.13617114, Global Avg Loss: 0.78759738, Time: 0.0207 Steps: 76890, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000256, Sample Num: 4096, Cur Loss: 0.22005722, Cur Avg Loss: 0.21429698, Log Avg loss: 0.20107904, Global Avg Loss: 0.78752111, Time: 0.0253 Steps: 76900, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000266, Sample Num: 4256, Cur Loss: 0.12814312, Cur Avg Loss: 0.21473917, Log Avg loss: 0.22605910, Global Avg Loss: 0.78744811, Time: 0.0208 Steps: 76910, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000276, Sample Num: 4416, Cur Loss: 0.48012602, Cur Avg Loss: 0.21653983, Log Avg loss: 0.26443752, Global Avg Loss: 0.78738012, Time: 0.0207 Steps: 76920, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000286, Sample Num: 4576, Cur Loss: 0.19503802, Cur Avg Loss: 0.21599277, Log Avg loss: 0.20089370, Global Avg Loss: 0.78730388, Time: 0.0208 Steps: 76930, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000296, Sample Num: 4736, Cur Loss: 0.03595287, Cur Avg Loss: 0.21380226, Log Avg loss: 0.15115387, Global Avg Loss: 0.78722120, Time: 0.0208 Steps: 76940, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000306, Sample Num: 4896, Cur Loss: 0.20187341, Cur Avg Loss: 0.21607930, Log Avg loss: 0.28347955, Global Avg Loss: 0.78715574, Time: 0.0209 Steps: 76950, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000316, Sample Num: 5056, Cur Loss: 0.08771636, Cur Avg Loss: 0.21553401, Log Avg loss: 0.19884813, Global Avg Loss: 0.78707929, Time: 0.0208 Steps: 76960, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000326, Sample Num: 5216, Cur Loss: 0.18063682, Cur Avg Loss: 0.21673568, Log Avg loss: 0.25470857, Global Avg Loss: 0.78701013, Time: 0.0209 Steps: 76970, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000336, Sample Num: 5376, Cur Loss: 0.24429330, Cur Avg Loss: 0.21651733, Log Avg loss: 0.20939899, Global Avg Loss: 0.78693509, Time: 0.0207 Steps: 76980, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000346, Sample Num: 5536, Cur Loss: 0.54001999, Cur Avg Loss: 0.21941891, Log Avg loss: 0.31691214, Global Avg Loss: 0.78687404, Time: 0.0208 Steps: 76990, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000356, Sample Num: 5696, Cur Loss: 0.06721992, Cur Avg Loss: 0.21763132, Log Avg loss: 0.15578069, Global Avg Loss: 0.78679208, Time: 0.0207 Steps: 77000, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000366, Sample Num: 5856, Cur Loss: 0.08039804, Cur Avg Loss: 0.21555686, Log Avg loss: 0.14170596, Global Avg Loss: 0.78670832, Time: 0.0207 Steps: 77010, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000376, Sample Num: 6016, Cur Loss: 0.16572519, Cur Avg Loss: 0.21763847, Log Avg loss: 0.29382551, Global Avg Loss: 0.78664432, Time: 0.0209 Steps: 77020, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000386, Sample Num: 6176, Cur Loss: 0.27473849, Cur Avg Loss: 0.21934008, Log Avg loss: 0.28332067, Global Avg Loss: 0.78657898, Time: 0.0208 Steps: 77030, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000396, Sample Num: 6336, Cur Loss: 0.24303941, Cur Avg Loss: 0.21817005, Log Avg loss: 0.17300677, Global Avg Loss: 0.78649934, Time: 0.0207 Steps: 77040, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000406, Sample Num: 6496, Cur Loss: 0.07599061, Cur Avg Loss: 0.22011541, Log Avg loss: 0.29715153, Global Avg Loss: 0.78643583, Time: 0.0207 Steps: 77050, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000416, Sample Num: 6656, Cur Loss: 0.34834152, Cur Avg Loss: 0.22387171, Log Avg loss: 0.37637777, Global Avg Loss: 0.78638261, Time: 0.0208 Steps: 77060, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000426, Sample Num: 6816, Cur Loss: 0.65390956, Cur Avg Loss: 0.22646137, Log Avg loss: 0.33419126, Global Avg Loss: 0.78632394, Time: 0.0208 Steps: 77070, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000436, Sample Num: 6976, Cur Loss: 0.22981516, Cur Avg Loss: 0.22586406, Log Avg loss: 0.20041848, Global Avg Loss: 0.78624793, Time: 0.0208 Steps: 77080, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000446, Sample Num: 7136, Cur Loss: 0.80493999, Cur Avg Loss: 0.22618851, Log Avg loss: 0.24033472, Global Avg Loss: 0.78617711, Time: 0.0208 Steps: 77090, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000456, Sample Num: 7296, Cur Loss: 0.04591532, Cur Avg Loss: 0.22443253, Log Avg loss: 0.14611543, Global Avg Loss: 0.78609410, Time: 0.0207 Steps: 77100, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000466, Sample Num: 7456, Cur Loss: 0.19917713, Cur Avg Loss: 0.22434250, Log Avg loss: 0.22023748, Global Avg Loss: 0.78602071, Time: 0.0208 Steps: 77110, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000476, Sample Num: 7616, Cur Loss: 0.16666451, Cur Avg Loss: 0.22275955, Log Avg loss: 0.14899378, Global Avg Loss: 0.78593811, Time: 0.0207 Steps: 77120, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000486, Sample Num: 7776, Cur Loss: 0.15479687, Cur Avg Loss: 0.22219104, Log Avg loss: 0.19513013, Global Avg Loss: 0.78586151, Time: 0.0209 Steps: 77130, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000496, Sample Num: 7936, Cur Loss: 0.07445732, Cur Avg Loss: 0.22171556, Log Avg loss: 0.19860706, Global Avg Loss: 0.78578538, Time: 0.0207 Steps: 77140, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000506, Sample Num: 8096, Cur Loss: 0.22607070, Cur Avg Loss: 0.22204882, Log Avg loss: 0.23857841, Global Avg Loss: 0.78571446, Time: 0.0207 Steps: 77150, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000516, Sample Num: 8256, Cur Loss: 0.17444944, Cur Avg Loss: 0.22239618, Log Avg loss: 0.23997294, Global Avg Loss: 0.78564373, Time: 0.0246 Steps: 77160, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000526, Sample Num: 8416, Cur Loss: 0.20597030, Cur Avg Loss: 0.22160128, Log Avg loss: 0.18058449, Global Avg Loss: 0.78556532, Time: 0.0208 Steps: 77170, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000536, Sample Num: 8576, Cur Loss: 0.11022100, Cur Avg Loss: 0.22055252, Log Avg loss: 0.16538743, Global Avg Loss: 0.78548497, Time: 0.0208 Steps: 77180, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000546, Sample Num: 8736, Cur Loss: 0.13044769, Cur Avg Loss: 0.21873998, Log Avg loss: 0.12158826, Global Avg Loss: 0.78539896, Time: 0.0208 Steps: 77190, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000556, Sample Num: 8896, Cur Loss: 0.16653499, Cur Avg Loss: 0.21769222, Log Avg loss: 0.16048414, Global Avg Loss: 0.78531801, Time: 0.0208 Steps: 77200, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000566, Sample Num: 9056, Cur Loss: 0.26358715, Cur Avg Loss: 0.21682341, Log Avg loss: 0.16851773, Global Avg Loss: 0.78523812, Time: 0.0207 Steps: 77210, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000576, Sample Num: 9216, Cur Loss: 0.06940895, Cur Avg Loss: 0.21669372, Log Avg loss: 0.20935348, Global Avg Loss: 0.78516355, Time: 0.0208 Steps: 77220, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000586, Sample Num: 9376, Cur Loss: 0.33392605, Cur Avg Loss: 0.21697311, Log Avg loss: 0.23306594, Global Avg Loss: 0.78509206, Time: 0.0208 Steps: 77230, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000596, Sample Num: 9536, Cur Loss: 0.23992559, Cur Avg Loss: 0.21656142, Log Avg loss: 0.19243629, Global Avg Loss: 0.78501533, Time: 0.0208 Steps: 77240, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000606, Sample Num: 9696, Cur Loss: 0.13905801, Cur Avg Loss: 0.21652576, Log Avg loss: 0.21440045, Global Avg Loss: 0.78494147, Time: 0.0209 Steps: 77250, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000616, Sample Num: 9856, Cur Loss: 0.21024992, Cur Avg Loss: 0.21600557, Log Avg loss: 0.18448166, Global Avg Loss: 0.78486375, Time: 0.0208 Steps: 77260, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000626, Sample Num: 10016, Cur Loss: 0.33639824, Cur Avg Loss: 0.21491090, Log Avg loss: 0.14747979, Global Avg Loss: 0.78478126, Time: 0.0208 Steps: 77270, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000636, Sample Num: 10176, Cur Loss: 0.33696243, Cur Avg Loss: 0.21533755, Log Avg loss: 0.24204554, Global Avg Loss: 0.78471103, Time: 0.0208 Steps: 77280, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000646, Sample Num: 10336, Cur Loss: 0.11633506, Cur Avg Loss: 0.21438202, Log Avg loss: 0.15361020, Global Avg Loss: 0.78462937, Time: 0.0208 Steps: 77290, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000656, Sample Num: 10496, Cur Loss: 0.35709283, Cur Avg Loss: 0.21473724, Log Avg loss: 0.23768447, Global Avg Loss: 0.78455862, Time: 0.0208 Steps: 77300, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000666, Sample Num: 10656, Cur Loss: 0.12223458, Cur Avg Loss: 0.21561287, Log Avg loss: 0.27305415, Global Avg Loss: 0.78449246, Time: 0.0208 Steps: 77310, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000676, Sample Num: 10816, Cur Loss: 0.14655720, Cur Avg Loss: 0.21480098, Log Avg loss: 0.16072910, Global Avg Loss: 0.78441178, Time: 0.0208 Steps: 77320, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000686, Sample Num: 10976, Cur Loss: 0.11028157, Cur Avg Loss: 0.21412258, Log Avg loss: 0.16826289, Global Avg Loss: 0.78433210, Time: 0.0208 Steps: 77330, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000696, Sample Num: 11136, Cur Loss: 0.26351273, Cur Avg Loss: 0.21295531, Log Avg loss: 0.13288090, Global Avg Loss: 0.78424787, Time: 0.0208 Steps: 77340, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000706, Sample Num: 11296, Cur Loss: 0.25549072, Cur Avg Loss: 0.21294393, Log Avg loss: 0.21215158, Global Avg Loss: 0.78417391, Time: 0.0208 Steps: 77350, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000716, Sample Num: 11456, Cur Loss: 0.18127592, Cur Avg Loss: 0.21326451, Log Avg loss: 0.23589716, Global Avg Loss: 0.78410304, Time: 0.0208 Steps: 77360, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000726, Sample Num: 11616, Cur Loss: 0.18322809, Cur Avg Loss: 0.21309618, Log Avg loss: 0.20104399, Global Avg Loss: 0.78402768, Time: 0.0208 Steps: 77370, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000736, Sample Num: 11776, Cur Loss: 0.49662399, Cur Avg Loss: 0.21389152, Log Avg loss: 0.27163325, Global Avg Loss: 0.78396146, Time: 0.0208 Steps: 77380, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000746, Sample Num: 11936, Cur Loss: 0.08743581, Cur Avg Loss: 0.21472149, Log Avg loss: 0.27580743, Global Avg Loss: 0.78389580, Time: 0.0208 Steps: 77390, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000756, Sample Num: 12096, Cur Loss: 0.08808307, Cur Avg Loss: 0.21443530, Log Avg loss: 0.19308501, Global Avg Loss: 0.78381947, Time: 0.0209 Steps: 77400, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000766, Sample Num: 12256, Cur Loss: 0.31398243, Cur Avg Loss: 0.21511491, Log Avg loss: 0.26649363, Global Avg Loss: 0.78375264, Time: 0.0208 Steps: 77410, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000776, Sample Num: 12416, Cur Loss: 0.28839064, Cur Avg Loss: 0.21548586, Log Avg loss: 0.24390093, Global Avg Loss: 0.78368291, Time: 0.0208 Steps: 77420, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000786, Sample Num: 12576, Cur Loss: 0.42432493, Cur Avg Loss: 0.21565196, Log Avg loss: 0.22854082, Global Avg Loss: 0.78361121, Time: 0.0207 Steps: 77430, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000796, Sample Num: 12736, Cur Loss: 0.10750017, Cur Avg Loss: 0.21592197, Log Avg loss: 0.23714537, Global Avg Loss: 0.78354064, Time: 0.0208 Steps: 77440, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000806, Sample Num: 12896, Cur Loss: 0.13349769, Cur Avg Loss: 0.21599752, Log Avg loss: 0.22201114, Global Avg Loss: 0.78346814, Time: 0.0208 Steps: 77450, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000816, Sample Num: 13056, Cur Loss: 0.09186179, Cur Avg Loss: 0.21569957, Log Avg loss: 0.19168452, Global Avg Loss: 0.78339174, Time: 0.0208 Steps: 77460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000826, Sample Num: 13216, Cur Loss: 0.17028335, Cur Avg Loss: 0.21529628, Log Avg loss: 0.18238777, Global Avg Loss: 0.78331416, Time: 0.0207 Steps: 77470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000836, Sample Num: 13376, Cur Loss: 0.14651603, Cur Avg Loss: 0.21479555, Log Avg loss: 0.17343585, Global Avg Loss: 0.78323545, Time: 0.0208 Steps: 77480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000846, Sample Num: 13536, Cur Loss: 0.25619146, Cur Avg Loss: 0.21449742, Log Avg loss: 0.18957340, Global Avg Loss: 0.78315884, Time: 0.0208 Steps: 77490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000856, Sample Num: 13696, Cur Loss: 0.20053324, Cur Avg Loss: 0.21525906, Log Avg loss: 0.27969337, Global Avg Loss: 0.78309388, Time: 0.0208 Steps: 77500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000866, Sample Num: 13856, Cur Loss: 0.24153480, Cur Avg Loss: 0.21467237, Log Avg loss: 0.16445242, Global Avg Loss: 0.78301406, Time: 0.0208 Steps: 77510, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000876, Sample Num: 14016, Cur Loss: 0.10878912, Cur Avg Loss: 0.21441177, Log Avg loss: 0.19184342, Global Avg Loss: 0.78293780, Time: 0.0207 Steps: 77520, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000886, Sample Num: 14176, Cur Loss: 0.19888085, Cur Avg Loss: 0.21452749, Log Avg loss: 0.22466470, Global Avg Loss: 0.78286579, Time: 0.0208 Steps: 77530, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000896, Sample Num: 14336, Cur Loss: 0.52428216, Cur Avg Loss: 0.21413977, Log Avg loss: 0.17978804, Global Avg Loss: 0.78278802, Time: 0.0208 Steps: 77540, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000906, Sample Num: 14496, Cur Loss: 0.41280827, Cur Avg Loss: 0.21477249, Log Avg loss: 0.27146353, Global Avg Loss: 0.78272208, Time: 0.0207 Steps: 77550, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000916, Sample Num: 14656, Cur Loss: 0.05146495, Cur Avg Loss: 0.21498863, Log Avg loss: 0.23457142, Global Avg Loss: 0.78265141, Time: 0.0208 Steps: 77560, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000926, Sample Num: 14816, Cur Loss: 0.23221420, Cur Avg Loss: 0.21542635, Log Avg loss: 0.25552100, Global Avg Loss: 0.78258345, Time: 0.0207 Steps: 77570, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000936, Sample Num: 14976, Cur Loss: 0.20485345, Cur Avg Loss: 0.21546149, Log Avg loss: 0.21871597, Global Avg Loss: 0.78251077, Time: 0.0207 Steps: 77580, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000946, Sample Num: 15136, Cur Loss: 0.27879995, Cur Avg Loss: 0.21540177, Log Avg loss: 0.20981208, Global Avg Loss: 0.78243696, Time: 0.0208 Steps: 77590, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000956, Sample Num: 15296, Cur Loss: 0.22756302, Cur Avg Loss: 0.21518080, Log Avg loss: 0.19427687, Global Avg Loss: 0.78236117, Time: 0.0210 Steps: 77600, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000966, Sample Num: 15456, Cur Loss: 0.18406747, Cur Avg Loss: 0.21577900, Log Avg loss: 0.27296710, Global Avg Loss: 0.78229553, Time: 0.0207 Steps: 77610, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000976, Sample Num: 15616, Cur Loss: 0.25501814, Cur Avg Loss: 0.21585485, Log Avg loss: 0.22318179, Global Avg Loss: 0.78222350, Time: 0.0209 Steps: 77620, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000986, Sample Num: 15776, Cur Loss: 0.49631292, Cur Avg Loss: 0.21672746, Log Avg loss: 0.30189444, Global Avg Loss: 0.78216162, Time: 0.0208 Steps: 77630, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000996, Sample Num: 15936, Cur Loss: 0.15253851, Cur Avg Loss: 0.21649573, Log Avg loss: 0.19364679, Global Avg Loss: 0.78208582, Time: 0.0208 Steps: 77640, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001006, Sample Num: 16096, Cur Loss: 0.12267271, Cur Avg Loss: 0.21640047, Log Avg loss: 0.20691306, Global Avg Loss: 0.78201175, Time: 0.0208 Steps: 77650, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001016, Sample Num: 16256, Cur Loss: 0.44932103, Cur Avg Loss: 0.21617723, Log Avg loss: 0.19371879, Global Avg Loss: 0.78193600, Time: 0.0208 Steps: 77660, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001026, Sample Num: 16416, Cur Loss: 0.11820836, Cur Avg Loss: 0.21543247, Log Avg loss: 0.13976529, Global Avg Loss: 0.78185332, Time: 0.0246 Steps: 77670, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001036, Sample Num: 16576, Cur Loss: 0.03764408, Cur Avg Loss: 0.21505186, Log Avg loss: 0.17600089, Global Avg Loss: 0.78177533, Time: 0.0208 Steps: 77680, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001046, Sample Num: 16736, Cur Loss: 0.48469350, Cur Avg Loss: 0.21591307, Log Avg loss: 0.30513388, Global Avg Loss: 0.78171397, Time: 0.0208 Steps: 77690, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001056, Sample Num: 16896, Cur Loss: 0.11964068, Cur Avg Loss: 0.21582954, Log Avg loss: 0.20709278, Global Avg Loss: 0.78164002, Time: 0.0208 Steps: 77700, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001066, Sample Num: 17056, Cur Loss: 0.09211171, Cur Avg Loss: 0.21592821, Log Avg loss: 0.22634826, Global Avg Loss: 0.78156856, Time: 0.0208 Steps: 77710, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001076, Sample Num: 17216, Cur Loss: 0.22431855, Cur Avg Loss: 0.21621801, Log Avg loss: 0.24711035, Global Avg Loss: 0.78149980, Time: 0.0207 Steps: 77720, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001086, Sample Num: 17376, Cur Loss: 0.13753241, Cur Avg Loss: 0.21626645, Log Avg loss: 0.22147866, Global Avg Loss: 0.78142775, Time: 0.0208 Steps: 77730, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001096, Sample Num: 17536, Cur Loss: 0.33030897, Cur Avg Loss: 0.21657362, Log Avg loss: 0.24993195, Global Avg Loss: 0.78135938, Time: 0.0208 Steps: 77740, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001106, Sample Num: 17696, Cur Loss: 0.04029660, Cur Avg Loss: 0.21594760, Log Avg loss: 0.14733634, Global Avg Loss: 0.78127783, Time: 0.0208 Steps: 77750, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001116, Sample Num: 17856, Cur Loss: 0.11357278, Cur Avg Loss: 0.21563647, Log Avg loss: 0.18122497, Global Avg Loss: 0.78120067, Time: 0.0207 Steps: 77760, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001126, Sample Num: 18016, Cur Loss: 0.15920168, Cur Avg Loss: 0.21547274, Log Avg loss: 0.19720050, Global Avg Loss: 0.78112557, Time: 0.0208 Steps: 77770, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001136, Sample Num: 18176, Cur Loss: 0.05274276, Cur Avg Loss: 0.21512311, Log Avg loss: 0.17575469, Global Avg Loss: 0.78104774, Time: 0.0208 Steps: 77780, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001146, Sample Num: 18336, Cur Loss: 0.15389879, Cur Avg Loss: 0.21470585, Log Avg loss: 0.16730527, Global Avg Loss: 0.78096885, Time: 0.0208 Steps: 77790, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001156, Sample Num: 18496, Cur Loss: 0.68852031, Cur Avg Loss: 0.21524242, Log Avg loss: 0.27673357, Global Avg Loss: 0.78090403, Time: 0.0208 Steps: 77800, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001166, Sample Num: 18656, Cur Loss: 0.05589971, Cur Avg Loss: 0.21587404, Log Avg loss: 0.28888864, Global Avg Loss: 0.78084080, Time: 0.0208 Steps: 77810, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001176, Sample Num: 18816, Cur Loss: 0.04921386, Cur Avg Loss: 0.21595267, Log Avg loss: 0.22512197, Global Avg Loss: 0.78076939, Time: 0.0208 Steps: 77820, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001186, Sample Num: 18976, Cur Loss: 0.19760665, Cur Avg Loss: 0.21606374, Log Avg loss: 0.22912492, Global Avg Loss: 0.78069851, Time: 0.0207 Steps: 77830, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001196, Sample Num: 19136, Cur Loss: 0.59586364, Cur Avg Loss: 0.21582058, Log Avg loss: 0.18698179, Global Avg Loss: 0.78062224, Time: 0.0207 Steps: 77840, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001206, Sample Num: 19296, Cur Loss: 0.12349936, Cur Avg Loss: 0.21598163, Log Avg loss: 0.23524300, Global Avg Loss: 0.78055218, Time: 0.0207 Steps: 77850, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001216, Sample Num: 19456, Cur Loss: 0.36378944, Cur Avg Loss: 0.21628296, Log Avg loss: 0.25262308, Global Avg Loss: 0.78048438, Time: 0.0207 Steps: 77860, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001226, Sample Num: 19616, Cur Loss: 0.05495919, Cur Avg Loss: 0.21664441, Log Avg loss: 0.26059712, Global Avg Loss: 0.78041761, Time: 0.0207 Steps: 77870, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001236, Sample Num: 19776, Cur Loss: 0.21992160, Cur Avg Loss: 0.21639349, Log Avg loss: 0.18563126, Global Avg Loss: 0.78034124, Time: 0.0207 Steps: 77880, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001246, Sample Num: 19936, Cur Loss: 0.16694732, Cur Avg Loss: 0.21600372, Log Avg loss: 0.16782788, Global Avg Loss: 0.78026260, Time: 0.0209 Steps: 77890, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001256, Sample Num: 20096, Cur Loss: 0.08339030, Cur Avg Loss: 0.21563748, Log Avg loss: 0.17000402, Global Avg Loss: 0.78018427, Time: 0.0209 Steps: 77900, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001266, Sample Num: 20256, Cur Loss: 0.23269022, Cur Avg Loss: 0.21567713, Log Avg loss: 0.22065756, Global Avg Loss: 0.78011245, Time: 0.0209 Steps: 77910, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001276, Sample Num: 20416, Cur Loss: 0.27769983, Cur Avg Loss: 0.21564879, Log Avg loss: 0.21206007, Global Avg Loss: 0.78003955, Time: 0.0209 Steps: 77920, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001286, Sample Num: 20576, Cur Loss: 0.05771234, Cur Avg Loss: 0.21535736, Log Avg loss: 0.17817098, Global Avg Loss: 0.77996231, Time: 0.0243 Steps: 77930, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001296, Sample Num: 20736, Cur Loss: 0.26711959, Cur Avg Loss: 0.21525520, Log Avg loss: 0.20211712, Global Avg Loss: 0.77988817, Time: 0.0209 Steps: 77940, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001306, Sample Num: 20896, Cur Loss: 0.21850583, Cur Avg Loss: 0.21563145, Log Avg loss: 0.26439385, Global Avg Loss: 0.77982204, Time: 0.0209 Steps: 77950, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001316, Sample Num: 21056, Cur Loss: 0.15465027, Cur Avg Loss: 0.21566468, Log Avg loss: 0.22000423, Global Avg Loss: 0.77975023, Time: 0.0209 Steps: 77960, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001326, Sample Num: 21216, Cur Loss: 0.09360730, Cur Avg Loss: 0.21501578, Log Avg loss: 0.12962165, Global Avg Loss: 0.77966685, Time: 0.0209 Steps: 77970, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001336, Sample Num: 21376, Cur Loss: 0.08205581, Cur Avg Loss: 0.21434631, Log Avg loss: 0.12557424, Global Avg Loss: 0.77958297, Time: 0.0209 Steps: 77980, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001346, Sample Num: 21536, Cur Loss: 0.06144910, Cur Avg Loss: 0.21392242, Log Avg loss: 0.15729060, Global Avg Loss: 0.77950318, Time: 0.0209 Steps: 77990, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001356, Sample Num: 21696, Cur Loss: 0.21452908, Cur Avg Loss: 0.21379336, Log Avg loss: 0.19642189, Global Avg Loss: 0.77942843, Time: 0.0209 Steps: 78000, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001366, Sample Num: 21856, Cur Loss: 0.21322720, Cur Avg Loss: 0.21369240, Log Avg loss: 0.20000143, Global Avg Loss: 0.77935415, Time: 0.0209 Steps: 78010, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001376, Sample Num: 22016, Cur Loss: 0.18486698, Cur Avg Loss: 0.21336243, Log Avg loss: 0.16828890, Global Avg Loss: 0.77927583, Time: 0.0209 Steps: 78020, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001386, Sample Num: 22176, Cur Loss: 0.20852485, Cur Avg Loss: 0.21324893, Log Avg loss: 0.19763133, Global Avg Loss: 0.77920129, Time: 0.0209 Steps: 78030, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001396, Sample Num: 22336, Cur Loss: 0.08428951, Cur Avg Loss: 0.21339695, Log Avg loss: 0.23391334, Global Avg Loss: 0.77913142, Time: 0.0209 Steps: 78040, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001406, Sample Num: 22496, Cur Loss: 0.05739968, Cur Avg Loss: 0.21360136, Log Avg loss: 0.24213699, Global Avg Loss: 0.77906261, Time: 0.0208 Steps: 78050, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001416, Sample Num: 22656, Cur Loss: 0.17253637, Cur Avg Loss: 0.21399175, Log Avg loss: 0.26887984, Global Avg Loss: 0.77899726, Time: 0.0209 Steps: 78060, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001426, Sample Num: 22816, Cur Loss: 0.11355677, Cur Avg Loss: 0.21368845, Log Avg loss: 0.17074176, Global Avg Loss: 0.77891935, Time: 0.0209 Steps: 78070, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001436, Sample Num: 22976, Cur Loss: 0.24740066, Cur Avg Loss: 0.21362460, Log Avg loss: 0.20451950, Global Avg Loss: 0.77884578, Time: 0.0209 Steps: 78080, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001446, Sample Num: 23136, Cur Loss: 0.41909939, Cur Avg Loss: 0.21446110, Log Avg loss: 0.33458282, Global Avg Loss: 0.77878889, Time: 0.0209 Steps: 78090, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001456, Sample Num: 23296, Cur Loss: 0.06286895, Cur Avg Loss: 0.21431191, Log Avg loss: 0.19273771, Global Avg Loss: 0.77871385, Time: 0.0209 Steps: 78100, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001466, Sample Num: 23456, Cur Loss: 0.28235054, Cur Avg Loss: 0.21414264, Log Avg loss: 0.18949698, Global Avg Loss: 0.77863842, Time: 0.0209 Steps: 78110, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001476, Sample Num: 23616, Cur Loss: 0.24612965, Cur Avg Loss: 0.21402843, Log Avg loss: 0.19728657, Global Avg Loss: 0.77856400, Time: 0.0210 Steps: 78120, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001486, Sample Num: 23776, Cur Loss: 0.33929694, Cur Avg Loss: 0.21409829, Log Avg loss: 0.22440841, Global Avg Loss: 0.77849307, Time: 0.0209 Steps: 78130, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001496, Sample Num: 23936, Cur Loss: 0.14764042, Cur Avg Loss: 0.21449470, Log Avg loss: 0.27340146, Global Avg Loss: 0.77842843, Time: 0.0208 Steps: 78140, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001506, Sample Num: 24096, Cur Loss: 0.10410170, Cur Avg Loss: 0.21465567, Log Avg loss: 0.23873755, Global Avg Loss: 0.77835937, Time: 0.0209 Steps: 78150, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001516, Sample Num: 24256, Cur Loss: 0.05654430, Cur Avg Loss: 0.21459727, Log Avg loss: 0.20580114, Global Avg Loss: 0.77828612, Time: 0.0209 Steps: 78160, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001526, Sample Num: 24416, Cur Loss: 0.26712036, Cur Avg Loss: 0.21506812, Log Avg loss: 0.28644914, Global Avg Loss: 0.77822320, Time: 0.0209 Steps: 78170, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001536, Sample Num: 24576, Cur Loss: 0.07518369, Cur Avg Loss: 0.21503600, Log Avg loss: 0.21013458, Global Avg Loss: 0.77815054, Time: 0.0254 Steps: 78180, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001546, Sample Num: 24736, Cur Loss: 0.31318003, Cur Avg Loss: 0.21537648, Log Avg loss: 0.26767398, Global Avg Loss: 0.77808525, Time: 0.0209 Steps: 78190, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001556, Sample Num: 24896, Cur Loss: 0.14621024, Cur Avg Loss: 0.21529188, Log Avg loss: 0.20221373, Global Avg Loss: 0.77801161, Time: 0.0209 Steps: 78200, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001566, Sample Num: 25056, Cur Loss: 0.25682372, Cur Avg Loss: 0.21503120, Log Avg loss: 0.17446886, Global Avg Loss: 0.77793444, Time: 0.0209 Steps: 78210, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001576, Sample Num: 25216, Cur Loss: 0.08287983, Cur Avg Loss: 0.21462038, Log Avg loss: 0.15028606, Global Avg Loss: 0.77785420, Time: 0.0209 Steps: 78220, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001586, Sample Num: 25376, Cur Loss: 0.09300128, Cur Avg Loss: 0.21479518, Log Avg loss: 0.24234342, Global Avg Loss: 0.77778574, Time: 0.0209 Steps: 78230, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001596, Sample Num: 25536, Cur Loss: 0.15669677, Cur Avg Loss: 0.21451111, Log Avg loss: 0.16945736, Global Avg Loss: 0.77770799, Time: 0.0209 Steps: 78240, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001606, Sample Num: 25696, Cur Loss: 0.13068873, Cur Avg Loss: 0.21425906, Log Avg loss: 0.17403206, Global Avg Loss: 0.77763084, Time: 0.0209 Steps: 78250, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001616, Sample Num: 25856, Cur Loss: 0.21960644, Cur Avg Loss: 0.21412007, Log Avg loss: 0.19179885, Global Avg Loss: 0.77755599, Time: 0.0209 Steps: 78260, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001626, Sample Num: 26016, Cur Loss: 0.32205230, Cur Avg Loss: 0.21440536, Log Avg loss: 0.26050852, Global Avg Loss: 0.77748993, Time: 0.0210 Steps: 78270, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001636, Sample Num: 26176, Cur Loss: 0.14721395, Cur Avg Loss: 0.21453027, Log Avg loss: 0.23483958, Global Avg Loss: 0.77742061, Time: 0.0209 Steps: 78280, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001646, Sample Num: 26336, Cur Loss: 0.07811551, Cur Avg Loss: 0.21435774, Log Avg loss: 0.18613291, Global Avg Loss: 0.77734508, Time: 0.0209 Steps: 78290, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001656, Sample Num: 26496, Cur Loss: 0.18431050, Cur Avg Loss: 0.21383340, Log Avg loss: 0.12752667, Global Avg Loss: 0.77726209, Time: 0.0209 Steps: 78300, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001666, Sample Num: 26656, Cur Loss: 0.25136554, Cur Avg Loss: 0.21374482, Log Avg loss: 0.19907506, Global Avg Loss: 0.77718826, Time: 0.0209 Steps: 78310, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001676, Sample Num: 26816, Cur Loss: 0.39540279, Cur Avg Loss: 0.21387192, Log Avg loss: 0.23504827, Global Avg Loss: 0.77711904, Time: 0.0209 Steps: 78320, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001686, Sample Num: 26976, Cur Loss: 0.21555763, Cur Avg Loss: 0.21368730, Log Avg loss: 0.18274402, Global Avg Loss: 0.77704316, Time: 0.0209 Steps: 78330, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001696, Sample Num: 27136, Cur Loss: 0.15140557, Cur Avg Loss: 0.21369753, Log Avg loss: 0.21542279, Global Avg Loss: 0.77697147, Time: 0.0209 Steps: 78340, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001706, Sample Num: 27296, Cur Loss: 0.20338604, Cur Avg Loss: 0.21414327, Log Avg loss: 0.28974015, Global Avg Loss: 0.77690928, Time: 0.0209 Steps: 78350, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001716, Sample Num: 27456, Cur Loss: 0.10025845, Cur Avg Loss: 0.21438987, Log Avg loss: 0.25646054, Global Avg Loss: 0.77684286, Time: 0.0209 Steps: 78360, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001726, Sample Num: 27616, Cur Loss: 0.18964201, Cur Avg Loss: 0.21438357, Log Avg loss: 0.21330246, Global Avg Loss: 0.77677095, Time: 0.0209 Steps: 78370, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001736, Sample Num: 27776, Cur Loss: 0.17930768, Cur Avg Loss: 0.21412187, Log Avg loss: 0.16895188, Global Avg Loss: 0.77669341, Time: 0.0210 Steps: 78380, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001746, Sample Num: 27936, Cur Loss: 0.23078416, Cur Avg Loss: 0.21401482, Log Avg loss: 0.19543032, Global Avg Loss: 0.77661926, Time: 0.0209 Steps: 78390, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001756, Sample Num: 28096, Cur Loss: 0.18115342, Cur Avg Loss: 0.21410119, Log Avg loss: 0.22918296, Global Avg Loss: 0.77654943, Time: 0.0209 Steps: 78400, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001766, Sample Num: 28256, Cur Loss: 0.07323888, Cur Avg Loss: 0.21415858, Log Avg loss: 0.22423463, Global Avg Loss: 0.77647899, Time: 0.0209 Steps: 78410, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001776, Sample Num: 28416, Cur Loss: 0.21211369, Cur Avg Loss: 0.21441574, Log Avg loss: 0.25983043, Global Avg Loss: 0.77641311, Time: 0.0209 Steps: 78420, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001786, Sample Num: 28576, Cur Loss: 0.21686353, Cur Avg Loss: 0.21423731, Log Avg loss: 0.18254915, Global Avg Loss: 0.77633739, Time: 0.0209 Steps: 78430, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001796, Sample Num: 28736, Cur Loss: 0.13394780, Cur Avg Loss: 0.21438194, Log Avg loss: 0.24021242, Global Avg Loss: 0.77626904, Time: 0.0256 Steps: 78440, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001806, Sample Num: 28896, Cur Loss: 0.33830225, Cur Avg Loss: 0.21441783, Log Avg loss: 0.22086294, Global Avg Loss: 0.77619824, Time: 0.0211 Steps: 78450, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001816, Sample Num: 29056, Cur Loss: 0.25402594, Cur Avg Loss: 0.21481790, Log Avg loss: 0.28707147, Global Avg Loss: 0.77613590, Time: 0.0210 Steps: 78460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001826, Sample Num: 29216, Cur Loss: 0.33321354, Cur Avg Loss: 0.21476580, Log Avg loss: 0.20530506, Global Avg Loss: 0.77606316, Time: 0.0210 Steps: 78470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001836, Sample Num: 29376, Cur Loss: 0.19065654, Cur Avg Loss: 0.21457451, Log Avg loss: 0.17964505, Global Avg Loss: 0.77598716, Time: 0.0219 Steps: 78480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001846, Sample Num: 29536, Cur Loss: 0.27236381, Cur Avg Loss: 0.21430439, Log Avg loss: 0.16471023, Global Avg Loss: 0.77590928, Time: 0.0220 Steps: 78490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001856, Sample Num: 29696, Cur Loss: 0.21136490, Cur Avg Loss: 0.21458775, Log Avg loss: 0.26689453, Global Avg Loss: 0.77584444, Time: 0.0209 Steps: 78500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001866, Sample Num: 29856, Cur Loss: 0.08771463, Cur Avg Loss: 0.21448225, Log Avg loss: 0.19490178, Global Avg Loss: 0.77577044, Time: 0.0209 Steps: 78510, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001876, Sample Num: 30016, Cur Loss: 0.46103644, Cur Avg Loss: 0.21484368, Log Avg loss: 0.28228705, Global Avg Loss: 0.77570759, Time: 0.0208 Steps: 78520, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001886, Sample Num: 30176, Cur Loss: 0.10640255, Cur Avg Loss: 0.21521132, Log Avg loss: 0.28418069, Global Avg Loss: 0.77564500, Time: 0.0219 Steps: 78530, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001896, Sample Num: 30336, Cur Loss: 0.34778887, Cur Avg Loss: 0.21557776, Log Avg loss: 0.28468757, Global Avg Loss: 0.77558249, Time: 0.0209 Steps: 78540, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001906, Sample Num: 30496, Cur Loss: 0.11403438, Cur Avg Loss: 0.21555022, Log Avg loss: 0.21032846, Global Avg Loss: 0.77551053, Time: 0.0219 Steps: 78550, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001916, Sample Num: 30656, Cur Loss: 0.30589509, Cur Avg Loss: 0.21555137, Log Avg loss: 0.21577147, Global Avg Loss: 0.77543928, Time: 0.0210 Steps: 78560, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001926, Sample Num: 30816, Cur Loss: 0.13385025, Cur Avg Loss: 0.21534305, Log Avg loss: 0.17542882, Global Avg Loss: 0.77536292, Time: 0.0209 Steps: 78570, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001936, Sample Num: 30976, Cur Loss: 0.51603901, Cur Avg Loss: 0.21562087, Log Avg loss: 0.26912989, Global Avg Loss: 0.77529849, Time: 0.0209 Steps: 78580, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001946, Sample Num: 31136, Cur Loss: 0.22754516, Cur Avg Loss: 0.21563380, Log Avg loss: 0.21813640, Global Avg Loss: 0.77522760, Time: 0.0219 Steps: 78590, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001956, Sample Num: 31296, Cur Loss: 0.19401038, Cur Avg Loss: 0.21548543, Log Avg loss: 0.18661223, Global Avg Loss: 0.77515271, Time: 0.0211 Steps: 78600, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001966, Sample Num: 31456, Cur Loss: 0.38737738, Cur Avg Loss: 0.21541982, Log Avg loss: 0.20258752, Global Avg Loss: 0.77507987, Time: 0.0208 Steps: 78610, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001976, Sample Num: 31616, Cur Loss: 0.18336840, Cur Avg Loss: 0.21535106, Log Avg loss: 0.20183197, Global Avg Loss: 0.77500696, Time: 0.0219 Steps: 78620, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001986, Sample Num: 31776, Cur Loss: 0.27502295, Cur Avg Loss: 0.21510990, Log Avg loss: 0.16745762, Global Avg Loss: 0.77492969, Time: 0.0219 Steps: 78630, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001996, Sample Num: 31936, Cur Loss: 0.07606025, Cur Avg Loss: 0.21491919, Log Avg loss: 0.17704417, Global Avg Loss: 0.77485367, Time: 0.0209 Steps: 78640, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002006, Sample Num: 32096, Cur Loss: 0.49121612, Cur Avg Loss: 0.21515617, Log Avg loss: 0.26245734, Global Avg Loss: 0.77478852, Time: 0.0219 Steps: 78650, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002016, Sample Num: 32256, Cur Loss: 0.07808102, Cur Avg Loss: 0.21493575, Log Avg loss: 0.17071797, Global Avg Loss: 0.77471172, Time: 0.0209 Steps: 78660, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002026, Sample Num: 32416, Cur Loss: 0.64149928, Cur Avg Loss: 0.21521634, Log Avg loss: 0.27178440, Global Avg Loss: 0.77464779, Time: 0.0209 Steps: 78670, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002036, Sample Num: 32576, Cur Loss: 0.26444256, Cur Avg Loss: 0.21526261, Log Avg loss: 0.22463621, Global Avg Loss: 0.77457789, Time: 0.0219 Steps: 78680, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002046, Sample Num: 32736, Cur Loss: 0.48784766, Cur Avg Loss: 0.21521509, Log Avg loss: 0.20554050, Global Avg Loss: 0.77450557, Time: 0.0211 Steps: 78690, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002056, Sample Num: 32896, Cur Loss: 0.27397114, Cur Avg Loss: 0.21525038, Log Avg loss: 0.22246977, Global Avg Loss: 0.77443543, Time: 0.0209 Steps: 78700, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002066, Sample Num: 33056, Cur Loss: 0.21157452, Cur Avg Loss: 0.21522088, Log Avg loss: 0.20915553, Global Avg Loss: 0.77436361, Time: 0.0210 Steps: 78710, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002076, Sample Num: 33216, Cur Loss: 0.12679583, Cur Avg Loss: 0.21486773, Log Avg loss: 0.14190712, Global Avg Loss: 0.77428327, Time: 0.0209 Steps: 78720, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002086, Sample Num: 33376, Cur Loss: 0.19872406, Cur Avg Loss: 0.21480597, Log Avg loss: 0.20198526, Global Avg Loss: 0.77421058, Time: 0.0209 Steps: 78730, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002096, Sample Num: 33536, Cur Loss: 0.21243474, Cur Avg Loss: 0.21452019, Log Avg loss: 0.15490637, Global Avg Loss: 0.77413193, Time: 0.0209 Steps: 78740, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002106, Sample Num: 33696, Cur Loss: 0.29776889, Cur Avg Loss: 0.21446143, Log Avg loss: 0.20214495, Global Avg Loss: 0.77405929, Time: 0.0209 Steps: 78750, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002116, Sample Num: 33856, Cur Loss: 0.20019048, Cur Avg Loss: 0.21471976, Log Avg loss: 0.26912489, Global Avg Loss: 0.77399518, Time: 0.0209 Steps: 78760, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002126, Sample Num: 34016, Cur Loss: 0.04899165, Cur Avg Loss: 0.21455953, Log Avg loss: 0.18065536, Global Avg Loss: 0.77391986, Time: 0.0209 Steps: 78770, Updated lr: 0.000026 ***** Running evaluation checkpoint-78773 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-78773 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.682048, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.324407, "eval_total_loss": 228.058163, "eval_mae": 0.456951, "eval_mse": 0.324415, "eval_r2": 0.793781, "eval_sp_statistic": 0.895725, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.922331, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.285945, "test_total_loss": 143.544631, "test_mae": 0.348234, "test_mse": 0.286046, "test_r2": 0.815383, "test_sp_statistic": 0.876184, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.924408, "test_ps_pvalue": 0.0, "lr": 2.624656235182551e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7738935262598835, "train_cur_epoch_loss": 456.40117104537785, "train_cur_epoch_avg_loss": 0.21437349508942125, "train_cur_epoch_time": 44.682048082351685, "train_cur_epoch_avg_time": 0.02098734057414358, "epoch": 37, "step": 78773} ################################################## Training, Epoch: 0038, Batch: 000007, Sample Num: 112, Cur Loss: 0.30320936, Cur Avg Loss: 0.15421044, Log Avg loss: 0.13270747, Global Avg Loss: 0.77383846, Time: 0.0224 Steps: 78780, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000017, Sample Num: 272, Cur Loss: 0.35768887, Cur Avg Loss: 0.22658837, Log Avg loss: 0.27725291, Global Avg Loss: 0.77377544, Time: 0.0211 Steps: 78790, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000027, Sample Num: 432, Cur Loss: 0.24921615, Cur Avg Loss: 0.22515940, Log Avg loss: 0.22273016, Global Avg Loss: 0.77370551, Time: 0.0211 Steps: 78800, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000037, Sample Num: 592, Cur Loss: 0.17593405, Cur Avg Loss: 0.21374771, Log Avg loss: 0.18293614, Global Avg Loss: 0.77363055, Time: 0.0211 Steps: 78810, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000047, Sample Num: 752, Cur Loss: 0.32556081, Cur Avg Loss: 0.19924151, Log Avg loss: 0.14556856, Global Avg Loss: 0.77355086, Time: 0.0210 Steps: 78820, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000057, Sample Num: 912, Cur Loss: 0.38347712, Cur Avg Loss: 0.21126651, Log Avg loss: 0.26778404, Global Avg Loss: 0.77348670, Time: 0.0211 Steps: 78830, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000067, Sample Num: 1072, Cur Loss: 0.17910790, Cur Avg Loss: 0.21029033, Log Avg loss: 0.20472607, Global Avg Loss: 0.77341456, Time: 0.0210 Steps: 78840, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000077, Sample Num: 1232, Cur Loss: 0.14798662, Cur Avg Loss: 0.21264053, Log Avg loss: 0.22838691, Global Avg Loss: 0.77334544, Time: 0.0210 Steps: 78850, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000087, Sample Num: 1392, Cur Loss: 0.07572341, Cur Avg Loss: 0.20736371, Log Avg loss: 0.16673219, Global Avg Loss: 0.77326852, Time: 0.0210 Steps: 78860, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000097, Sample Num: 1552, Cur Loss: 0.36436638, Cur Avg Loss: 0.20699913, Log Avg loss: 0.20382726, Global Avg Loss: 0.77319632, Time: 0.0210 Steps: 78870, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000107, Sample Num: 1712, Cur Loss: 0.17691916, Cur Avg Loss: 0.20076902, Log Avg loss: 0.14033700, Global Avg Loss: 0.77311609, Time: 0.0210 Steps: 78880, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000117, Sample Num: 1872, Cur Loss: 0.27373838, Cur Avg Loss: 0.20813206, Log Avg loss: 0.28691650, Global Avg Loss: 0.77305446, Time: 0.0211 Steps: 78890, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000127, Sample Num: 2032, Cur Loss: 0.05420440, Cur Avg Loss: 0.20762444, Log Avg loss: 0.20168534, Global Avg Loss: 0.77298204, Time: 0.0211 Steps: 78900, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000137, Sample Num: 2192, Cur Loss: 0.40242010, Cur Avg Loss: 0.20713746, Log Avg loss: 0.20095276, Global Avg Loss: 0.77290955, Time: 0.0210 Steps: 78910, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000147, Sample Num: 2352, Cur Loss: 0.13536392, Cur Avg Loss: 0.20877462, Log Avg loss: 0.23120376, Global Avg Loss: 0.77284091, Time: 0.0210 Steps: 78920, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000157, Sample Num: 2512, Cur Loss: 0.17318283, Cur Avg Loss: 0.20788730, Log Avg loss: 0.19484365, Global Avg Loss: 0.77276768, Time: 0.0210 Steps: 78930, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000167, Sample Num: 2672, Cur Loss: 0.14428934, Cur Avg Loss: 0.20723326, Log Avg loss: 0.19696483, Global Avg Loss: 0.77269474, Time: 0.0210 Steps: 78940, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000177, Sample Num: 2832, Cur Loss: 0.10058814, Cur Avg Loss: 0.20395439, Log Avg loss: 0.14919722, Global Avg Loss: 0.77261577, Time: 0.0210 Steps: 78950, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000187, Sample Num: 2992, Cur Loss: 0.30482489, Cur Avg Loss: 0.20153661, Log Avg loss: 0.15874193, Global Avg Loss: 0.77253802, Time: 0.0210 Steps: 78960, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000197, Sample Num: 3152, Cur Loss: 0.17734045, Cur Avg Loss: 0.19920824, Log Avg loss: 0.15566784, Global Avg Loss: 0.77245991, Time: 0.0210 Steps: 78970, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000207, Sample Num: 3312, Cur Loss: 0.14292048, Cur Avg Loss: 0.20088610, Log Avg loss: 0.23393987, Global Avg Loss: 0.77239172, Time: 0.0210 Steps: 78980, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000217, Sample Num: 3472, Cur Loss: 0.17639276, Cur Avg Loss: 0.19986558, Log Avg loss: 0.17874072, Global Avg Loss: 0.77231657, Time: 0.0210 Steps: 78990, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000227, Sample Num: 3632, Cur Loss: 0.16854782, Cur Avg Loss: 0.19893259, Log Avg loss: 0.17868692, Global Avg Loss: 0.77224142, Time: 0.0210 Steps: 79000, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000237, Sample Num: 3792, Cur Loss: 0.18592390, Cur Avg Loss: 0.19931558, Log Avg loss: 0.20800925, Global Avg Loss: 0.77217001, Time: 0.0210 Steps: 79010, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000247, Sample Num: 3952, Cur Loss: 0.09713829, Cur Avg Loss: 0.19995528, Log Avg loss: 0.21511622, Global Avg Loss: 0.77209952, Time: 0.0210 Steps: 79020, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000257, Sample Num: 4112, Cur Loss: 0.14788993, Cur Avg Loss: 0.20315097, Log Avg loss: 0.28208460, Global Avg Loss: 0.77203751, Time: 0.0245 Steps: 79030, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000267, Sample Num: 4272, Cur Loss: 0.22079991, Cur Avg Loss: 0.20262735, Log Avg loss: 0.18917039, Global Avg Loss: 0.77196377, Time: 0.0209 Steps: 79040, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000277, Sample Num: 4432, Cur Loss: 0.09757048, Cur Avg Loss: 0.20170641, Log Avg loss: 0.17711710, Global Avg Loss: 0.77188852, Time: 0.0209 Steps: 79050, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000287, Sample Num: 4592, Cur Loss: 0.21297386, Cur Avg Loss: 0.20097698, Log Avg loss: 0.18077180, Global Avg Loss: 0.77181375, Time: 0.0209 Steps: 79060, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000297, Sample Num: 4752, Cur Loss: 0.03726292, Cur Avg Loss: 0.19933443, Log Avg loss: 0.15219334, Global Avg Loss: 0.77173539, Time: 0.0209 Steps: 79070, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000307, Sample Num: 4912, Cur Loss: 0.18538710, Cur Avg Loss: 0.19879744, Log Avg loss: 0.18284896, Global Avg Loss: 0.77166092, Time: 0.0209 Steps: 79080, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000317, Sample Num: 5072, Cur Loss: 0.27102387, Cur Avg Loss: 0.20155038, Log Avg loss: 0.28606543, Global Avg Loss: 0.77159952, Time: 0.0209 Steps: 79090, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000327, Sample Num: 5232, Cur Loss: 0.07468912, Cur Avg Loss: 0.20020862, Log Avg loss: 0.15767490, Global Avg Loss: 0.77152191, Time: 0.0209 Steps: 79100, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000337, Sample Num: 5392, Cur Loss: 0.13479067, Cur Avg Loss: 0.19810061, Log Avg loss: 0.12916883, Global Avg Loss: 0.77144071, Time: 0.0209 Steps: 79110, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000347, Sample Num: 5552, Cur Loss: 0.17810962, Cur Avg Loss: 0.19861769, Log Avg loss: 0.21604316, Global Avg Loss: 0.77137051, Time: 0.0209 Steps: 79120, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000357, Sample Num: 5712, Cur Loss: 0.26043156, Cur Avg Loss: 0.19834618, Log Avg loss: 0.18892491, Global Avg Loss: 0.77129691, Time: 0.0209 Steps: 79130, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000367, Sample Num: 5872, Cur Loss: 0.33557594, Cur Avg Loss: 0.19903411, Log Avg loss: 0.22359298, Global Avg Loss: 0.77122770, Time: 0.0209 Steps: 79140, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000377, Sample Num: 6032, Cur Loss: 0.12529886, Cur Avg Loss: 0.19947486, Log Avg loss: 0.21565032, Global Avg Loss: 0.77115751, Time: 0.0209 Steps: 79150, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000387, Sample Num: 6192, Cur Loss: 0.24514428, Cur Avg Loss: 0.19993153, Log Avg loss: 0.21714808, Global Avg Loss: 0.77108752, Time: 0.0209 Steps: 79160, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000397, Sample Num: 6352, Cur Loss: 0.33081603, Cur Avg Loss: 0.20078371, Log Avg loss: 0.23376318, Global Avg Loss: 0.77101965, Time: 0.0209 Steps: 79170, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000407, Sample Num: 6512, Cur Loss: 0.13547410, Cur Avg Loss: 0.19944787, Log Avg loss: 0.14641510, Global Avg Loss: 0.77094077, Time: 0.0209 Steps: 79180, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000417, Sample Num: 6672, Cur Loss: 0.23731402, Cur Avg Loss: 0.20041478, Log Avg loss: 0.23976786, Global Avg Loss: 0.77087369, Time: 0.0209 Steps: 79190, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000427, Sample Num: 6832, Cur Loss: 0.44474587, Cur Avg Loss: 0.20316924, Log Avg loss: 0.31803038, Global Avg Loss: 0.77081652, Time: 0.0209 Steps: 79200, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000437, Sample Num: 6992, Cur Loss: 0.12837920, Cur Avg Loss: 0.20291356, Log Avg loss: 0.19199601, Global Avg Loss: 0.77074344, Time: 0.0209 Steps: 79210, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000447, Sample Num: 7152, Cur Loss: 0.12295405, Cur Avg Loss: 0.20296124, Log Avg loss: 0.20504484, Global Avg Loss: 0.77067203, Time: 0.0209 Steps: 79220, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000457, Sample Num: 7312, Cur Loss: 0.20574760, Cur Avg Loss: 0.20353128, Log Avg loss: 0.22901202, Global Avg Loss: 0.77060367, Time: 0.0209 Steps: 79230, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000467, Sample Num: 7472, Cur Loss: 0.21523130, Cur Avg Loss: 0.20342548, Log Avg loss: 0.19859033, Global Avg Loss: 0.77053148, Time: 0.0209 Steps: 79240, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000477, Sample Num: 7632, Cur Loss: 0.20856699, Cur Avg Loss: 0.20372806, Log Avg loss: 0.21785868, Global Avg Loss: 0.77046174, Time: 0.0209 Steps: 79250, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000487, Sample Num: 7792, Cur Loss: 0.16133668, Cur Avg Loss: 0.20291836, Log Avg loss: 0.16429569, Global Avg Loss: 0.77038526, Time: 0.0209 Steps: 79260, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000497, Sample Num: 7952, Cur Loss: 0.09891774, Cur Avg Loss: 0.20249185, Log Avg loss: 0.18172051, Global Avg Loss: 0.77031100, Time: 0.0209 Steps: 79270, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000507, Sample Num: 8112, Cur Loss: 0.43587449, Cur Avg Loss: 0.20465375, Log Avg loss: 0.31210047, Global Avg Loss: 0.77025321, Time: 0.0209 Steps: 79280, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000517, Sample Num: 8272, Cur Loss: 0.11198185, Cur Avg Loss: 0.20381894, Log Avg loss: 0.16149396, Global Avg Loss: 0.77017643, Time: 0.0247 Steps: 79290, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000527, Sample Num: 8432, Cur Loss: 0.17868076, Cur Avg Loss: 0.20412233, Log Avg loss: 0.21980736, Global Avg Loss: 0.77010703, Time: 0.0209 Steps: 79300, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000537, Sample Num: 8592, Cur Loss: 0.28941321, Cur Avg Loss: 0.20399086, Log Avg loss: 0.19706243, Global Avg Loss: 0.77003477, Time: 0.0209 Steps: 79310, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000547, Sample Num: 8752, Cur Loss: 0.51346612, Cur Avg Loss: 0.20372714, Log Avg loss: 0.18956580, Global Avg Loss: 0.76996159, Time: 0.0211 Steps: 79320, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000557, Sample Num: 8912, Cur Loss: 0.11312614, Cur Avg Loss: 0.20280862, Log Avg loss: 0.15256521, Global Avg Loss: 0.76988377, Time: 0.0209 Steps: 79330, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000567, Sample Num: 9072, Cur Loss: 0.06264071, Cur Avg Loss: 0.20221121, Log Avg loss: 0.16893588, Global Avg Loss: 0.76980802, Time: 0.0209 Steps: 79340, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000577, Sample Num: 9232, Cur Loss: 0.11118901, Cur Avg Loss: 0.20130337, Log Avg loss: 0.14982860, Global Avg Loss: 0.76972989, Time: 0.0209 Steps: 79350, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000587, Sample Num: 9392, Cur Loss: 0.22036186, Cur Avg Loss: 0.20120624, Log Avg loss: 0.19560158, Global Avg Loss: 0.76965755, Time: 0.0209 Steps: 79360, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000597, Sample Num: 9552, Cur Loss: 0.52803481, Cur Avg Loss: 0.20188450, Log Avg loss: 0.24169873, Global Avg Loss: 0.76959103, Time: 0.0209 Steps: 79370, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000607, Sample Num: 9712, Cur Loss: 1.09318876, Cur Avg Loss: 0.20373469, Log Avg loss: 0.31419112, Global Avg Loss: 0.76953366, Time: 0.0209 Steps: 79380, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000617, Sample Num: 9872, Cur Loss: 0.23310527, Cur Avg Loss: 0.20283071, Log Avg loss: 0.14795899, Global Avg Loss: 0.76945536, Time: 0.0209 Steps: 79390, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000627, Sample Num: 10032, Cur Loss: 0.26444849, Cur Avg Loss: 0.20299095, Log Avg loss: 0.21287740, Global Avg Loss: 0.76938527, Time: 0.0209 Steps: 79400, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000637, Sample Num: 10192, Cur Loss: 0.05772719, Cur Avg Loss: 0.20253981, Log Avg loss: 0.17425332, Global Avg Loss: 0.76931032, Time: 0.0209 Steps: 79410, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000647, Sample Num: 10352, Cur Loss: 0.12958068, Cur Avg Loss: 0.20215373, Log Avg loss: 0.17756045, Global Avg Loss: 0.76923581, Time: 0.0209 Steps: 79420, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000657, Sample Num: 10512, Cur Loss: 0.31369564, Cur Avg Loss: 0.20211258, Log Avg loss: 0.19945046, Global Avg Loss: 0.76916408, Time: 0.0209 Steps: 79430, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000667, Sample Num: 10672, Cur Loss: 0.09381705, Cur Avg Loss: 0.20242651, Log Avg loss: 0.22305173, Global Avg Loss: 0.76909533, Time: 0.0209 Steps: 79440, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000677, Sample Num: 10832, Cur Loss: 0.11377068, Cur Avg Loss: 0.20274824, Log Avg loss: 0.22420773, Global Avg Loss: 0.76902675, Time: 0.0210 Steps: 79450, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000687, Sample Num: 10992, Cur Loss: 0.15972073, Cur Avg Loss: 0.20162184, Log Avg loss: 0.12536417, Global Avg Loss: 0.76894575, Time: 0.0209 Steps: 79460, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000697, Sample Num: 11152, Cur Loss: 0.42075199, Cur Avg Loss: 0.20186996, Log Avg loss: 0.21891622, Global Avg Loss: 0.76887653, Time: 0.0209 Steps: 79470, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000707, Sample Num: 11312, Cur Loss: 0.23121071, Cur Avg Loss: 0.20082742, Log Avg loss: 0.12816229, Global Avg Loss: 0.76879592, Time: 0.0210 Steps: 79480, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000717, Sample Num: 11472, Cur Loss: 0.24681701, Cur Avg Loss: 0.20122688, Log Avg loss: 0.22946899, Global Avg Loss: 0.76872807, Time: 0.0209 Steps: 79490, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000727, Sample Num: 11632, Cur Loss: 0.36695629, Cur Avg Loss: 0.20133628, Log Avg loss: 0.20918017, Global Avg Loss: 0.76865769, Time: 0.0209 Steps: 79500, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000737, Sample Num: 11792, Cur Loss: 0.15134010, Cur Avg Loss: 0.20119082, Log Avg loss: 0.19061538, Global Avg Loss: 0.76858499, Time: 0.0210 Steps: 79510, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000747, Sample Num: 11952, Cur Loss: 0.18103918, Cur Avg Loss: 0.20152179, Log Avg loss: 0.22591461, Global Avg Loss: 0.76851674, Time: 0.0209 Steps: 79520, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000757, Sample Num: 12112, Cur Loss: 0.09014383, Cur Avg Loss: 0.20053643, Log Avg loss: 0.12693002, Global Avg Loss: 0.76843607, Time: 0.0210 Steps: 79530, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000767, Sample Num: 12272, Cur Loss: 0.13094437, Cur Avg Loss: 0.20082344, Log Avg loss: 0.22255028, Global Avg Loss: 0.76836744, Time: 0.0210 Steps: 79540, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000777, Sample Num: 12432, Cur Loss: 0.17399523, Cur Avg Loss: 0.20128999, Log Avg loss: 0.23707448, Global Avg Loss: 0.76830065, Time: 0.0210 Steps: 79550, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000787, Sample Num: 12592, Cur Loss: 0.03855479, Cur Avg Loss: 0.20141441, Log Avg loss: 0.21108137, Global Avg Loss: 0.76823062, Time: 0.0209 Steps: 79560, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000797, Sample Num: 12752, Cur Loss: 0.59416318, Cur Avg Loss: 0.20270847, Log Avg loss: 0.30455095, Global Avg Loss: 0.76817234, Time: 0.0209 Steps: 79570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000807, Sample Num: 12912, Cur Loss: 0.38197351, Cur Avg Loss: 0.20291405, Log Avg loss: 0.21929886, Global Avg Loss: 0.76810337, Time: 0.0209 Steps: 79580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000817, Sample Num: 13072, Cur Loss: 0.22099167, Cur Avg Loss: 0.20272560, Log Avg loss: 0.18751755, Global Avg Loss: 0.76803043, Time: 0.0209 Steps: 79590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000827, Sample Num: 13232, Cur Loss: 0.12523805, Cur Avg Loss: 0.20321696, Log Avg loss: 0.24336122, Global Avg Loss: 0.76796451, Time: 0.0209 Steps: 79600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000837, Sample Num: 13392, Cur Loss: 0.13892038, Cur Avg Loss: 0.20274092, Log Avg loss: 0.16337230, Global Avg Loss: 0.76788857, Time: 0.0209 Steps: 79610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000847, Sample Num: 13552, Cur Loss: 0.05978841, Cur Avg Loss: 0.20269523, Log Avg loss: 0.19887151, Global Avg Loss: 0.76781710, Time: 0.0209 Steps: 79620, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000857, Sample Num: 13712, Cur Loss: 0.19800650, Cur Avg Loss: 0.20321348, Log Avg loss: 0.24710873, Global Avg Loss: 0.76775171, Time: 0.0209 Steps: 79630, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000867, Sample Num: 13872, Cur Loss: 0.29759338, Cur Avg Loss: 0.20342202, Log Avg loss: 0.22129372, Global Avg Loss: 0.76768309, Time: 0.0209 Steps: 79640, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000877, Sample Num: 14032, Cur Loss: 0.31012890, Cur Avg Loss: 0.20376372, Log Avg loss: 0.23338955, Global Avg Loss: 0.76761601, Time: 0.0209 Steps: 79650, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000887, Sample Num: 14192, Cur Loss: 0.28416747, Cur Avg Loss: 0.20357032, Log Avg loss: 0.18660874, Global Avg Loss: 0.76754308, Time: 0.0209 Steps: 79660, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000897, Sample Num: 14352, Cur Loss: 0.12459725, Cur Avg Loss: 0.20280630, Log Avg loss: 0.13503791, Global Avg Loss: 0.76746369, Time: 0.0209 Steps: 79670, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000907, Sample Num: 14512, Cur Loss: 0.34416085, Cur Avg Loss: 0.20370902, Log Avg loss: 0.28468310, Global Avg Loss: 0.76740310, Time: 0.0209 Steps: 79680, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000917, Sample Num: 14672, Cur Loss: 0.17341404, Cur Avg Loss: 0.20389268, Log Avg loss: 0.22055074, Global Avg Loss: 0.76733448, Time: 0.0209 Steps: 79690, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000927, Sample Num: 14832, Cur Loss: 0.17670156, Cur Avg Loss: 0.20420893, Log Avg loss: 0.23320924, Global Avg Loss: 0.76726746, Time: 0.0209 Steps: 79700, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000937, Sample Num: 14992, Cur Loss: 0.29070652, Cur Avg Loss: 0.20520386, Log Avg loss: 0.29743386, Global Avg Loss: 0.76720852, Time: 0.0210 Steps: 79710, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000947, Sample Num: 15152, Cur Loss: 0.13819905, Cur Avg Loss: 0.20553608, Log Avg loss: 0.23666445, Global Avg Loss: 0.76714196, Time: 0.0209 Steps: 79720, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000957, Sample Num: 15312, Cur Loss: 0.15154006, Cur Avg Loss: 0.20524936, Log Avg loss: 0.17809739, Global Avg Loss: 0.76706808, Time: 0.0209 Steps: 79730, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000967, Sample Num: 15472, Cur Loss: 0.23142931, Cur Avg Loss: 0.20572899, Log Avg loss: 0.25162923, Global Avg Loss: 0.76700344, Time: 0.0210 Steps: 79740, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000977, Sample Num: 15632, Cur Loss: 0.03915410, Cur Avg Loss: 0.20558364, Log Avg loss: 0.19152824, Global Avg Loss: 0.76693128, Time: 0.0209 Steps: 79750, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000987, Sample Num: 15792, Cur Loss: 0.12606069, Cur Avg Loss: 0.20523159, Log Avg loss: 0.17083645, Global Avg Loss: 0.76685655, Time: 0.0209 Steps: 79760, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000997, Sample Num: 15952, Cur Loss: 0.32441086, Cur Avg Loss: 0.20570695, Log Avg loss: 0.25262543, Global Avg Loss: 0.76679208, Time: 0.0209 Steps: 79770, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001007, Sample Num: 16112, Cur Loss: 0.80021322, Cur Avg Loss: 0.20577627, Log Avg loss: 0.21268744, Global Avg Loss: 0.76672263, Time: 0.0209 Steps: 79780, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001017, Sample Num: 16272, Cur Loss: 0.24904886, Cur Avg Loss: 0.20592113, Log Avg loss: 0.22050822, Global Avg Loss: 0.76665417, Time: 0.0210 Steps: 79790, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001027, Sample Num: 16432, Cur Loss: 0.10770991, Cur Avg Loss: 0.20558070, Log Avg loss: 0.17095934, Global Avg Loss: 0.76657953, Time: 0.0246 Steps: 79800, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001037, Sample Num: 16592, Cur Loss: 0.34906951, Cur Avg Loss: 0.20549805, Log Avg loss: 0.19700979, Global Avg Loss: 0.76650816, Time: 0.0209 Steps: 79810, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001047, Sample Num: 16752, Cur Loss: 0.09850836, Cur Avg Loss: 0.20545385, Log Avg loss: 0.20086974, Global Avg Loss: 0.76643730, Time: 0.0209 Steps: 79820, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001057, Sample Num: 16912, Cur Loss: 0.31951398, Cur Avg Loss: 0.20521160, Log Avg loss: 0.17984856, Global Avg Loss: 0.76636382, Time: 0.0209 Steps: 79830, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001067, Sample Num: 17072, Cur Loss: 0.13261229, Cur Avg Loss: 0.20559160, Log Avg loss: 0.24575713, Global Avg Loss: 0.76629861, Time: 0.0209 Steps: 79840, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001077, Sample Num: 17232, Cur Loss: 0.11788443, Cur Avg Loss: 0.20570011, Log Avg loss: 0.21727895, Global Avg Loss: 0.76622985, Time: 0.0209 Steps: 79850, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001087, Sample Num: 17392, Cur Loss: 0.20194633, Cur Avg Loss: 0.20651977, Log Avg loss: 0.29479714, Global Avg Loss: 0.76617082, Time: 0.0209 Steps: 79860, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001097, Sample Num: 17552, Cur Loss: 0.13449571, Cur Avg Loss: 0.20659224, Log Avg loss: 0.21446928, Global Avg Loss: 0.76610175, Time: 0.0209 Steps: 79870, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001107, Sample Num: 17712, Cur Loss: 0.14921892, Cur Avg Loss: 0.20592567, Log Avg loss: 0.13280244, Global Avg Loss: 0.76602246, Time: 0.0209 Steps: 79880, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001117, Sample Num: 17872, Cur Loss: 0.15092781, Cur Avg Loss: 0.20578065, Log Avg loss: 0.18972701, Global Avg Loss: 0.76595033, Time: 0.0209 Steps: 79890, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001127, Sample Num: 18032, Cur Loss: 0.08958904, Cur Avg Loss: 0.20592552, Log Avg loss: 0.22210737, Global Avg Loss: 0.76588226, Time: 0.0209 Steps: 79900, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001137, Sample Num: 18192, Cur Loss: 0.14012571, Cur Avg Loss: 0.20539882, Log Avg loss: 0.14604021, Global Avg Loss: 0.76580470, Time: 0.0209 Steps: 79910, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001147, Sample Num: 18352, Cur Loss: 0.07943467, Cur Avg Loss: 0.20536935, Log Avg loss: 0.20201829, Global Avg Loss: 0.76573415, Time: 0.0209 Steps: 79920, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001157, Sample Num: 18512, Cur Loss: 0.15369920, Cur Avg Loss: 0.20535572, Log Avg loss: 0.20379327, Global Avg Loss: 0.76566385, Time: 0.0209 Steps: 79930, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001167, Sample Num: 18672, Cur Loss: 0.18798371, Cur Avg Loss: 0.20520560, Log Avg loss: 0.18783574, Global Avg Loss: 0.76559156, Time: 0.0209 Steps: 79940, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001177, Sample Num: 18832, Cur Loss: 0.29545701, Cur Avg Loss: 0.20533814, Log Avg loss: 0.22080613, Global Avg Loss: 0.76552342, Time: 0.0209 Steps: 79950, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001187, Sample Num: 18992, Cur Loss: 0.25466862, Cur Avg Loss: 0.20534482, Log Avg loss: 0.20613039, Global Avg Loss: 0.76545346, Time: 0.0209 Steps: 79960, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001197, Sample Num: 19152, Cur Loss: 0.09475842, Cur Avg Loss: 0.20525702, Log Avg loss: 0.19483519, Global Avg Loss: 0.76538211, Time: 0.0209 Steps: 79970, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001207, Sample Num: 19312, Cur Loss: 0.30337793, Cur Avg Loss: 0.20524781, Log Avg loss: 0.20414614, Global Avg Loss: 0.76531194, Time: 0.0209 Steps: 79980, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001217, Sample Num: 19472, Cur Loss: 0.08605646, Cur Avg Loss: 0.20552612, Log Avg loss: 0.23911772, Global Avg Loss: 0.76524616, Time: 0.0209 Steps: 79990, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001227, Sample Num: 19632, Cur Loss: 0.20417809, Cur Avg Loss: 0.20568504, Log Avg loss: 0.22502542, Global Avg Loss: 0.76517863, Time: 0.0209 Steps: 80000, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001237, Sample Num: 19792, Cur Loss: 0.55175012, Cur Avg Loss: 0.20553911, Log Avg loss: 0.18763340, Global Avg Loss: 0.76510644, Time: 0.0209 Steps: 80010, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001247, Sample Num: 19952, Cur Loss: 0.17911384, Cur Avg Loss: 0.20547976, Log Avg loss: 0.19813844, Global Avg Loss: 0.76503559, Time: 0.0209 Steps: 80020, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001257, Sample Num: 20112, Cur Loss: 0.72905672, Cur Avg Loss: 0.20607033, Log Avg loss: 0.27971454, Global Avg Loss: 0.76497495, Time: 0.0209 Steps: 80030, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001267, Sample Num: 20272, Cur Loss: 0.25659364, Cur Avg Loss: 0.20583191, Log Avg loss: 0.17586308, Global Avg Loss: 0.76490135, Time: 0.0209 Steps: 80040, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001277, Sample Num: 20432, Cur Loss: 0.11799389, Cur Avg Loss: 0.20561148, Log Avg loss: 0.17768281, Global Avg Loss: 0.76482799, Time: 0.0209 Steps: 80050, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001287, Sample Num: 20592, Cur Loss: 0.09351259, Cur Avg Loss: 0.20570890, Log Avg loss: 0.21814848, Global Avg Loss: 0.76475971, Time: 0.0209 Steps: 80060, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001297, Sample Num: 20752, Cur Loss: 0.25294510, Cur Avg Loss: 0.20584846, Log Avg loss: 0.22381089, Global Avg Loss: 0.76469215, Time: 0.0209 Steps: 80070, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001307, Sample Num: 20912, Cur Loss: 0.22114152, Cur Avg Loss: 0.20597098, Log Avg loss: 0.22186198, Global Avg Loss: 0.76462436, Time: 0.0209 Steps: 80080, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001317, Sample Num: 21072, Cur Loss: 0.09349606, Cur Avg Loss: 0.20562049, Log Avg loss: 0.15981072, Global Avg Loss: 0.76454884, Time: 0.0209 Steps: 80090, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001327, Sample Num: 21232, Cur Loss: 0.22328453, Cur Avg Loss: 0.20564939, Log Avg loss: 0.20945612, Global Avg Loss: 0.76447954, Time: 0.0209 Steps: 80100, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001337, Sample Num: 21392, Cur Loss: 0.26637936, Cur Avg Loss: 0.20594150, Log Avg loss: 0.24470396, Global Avg Loss: 0.76441466, Time: 0.0208 Steps: 80110, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001347, Sample Num: 21552, Cur Loss: 0.11695769, Cur Avg Loss: 0.20646984, Log Avg loss: 0.27710845, Global Avg Loss: 0.76435384, Time: 0.0209 Steps: 80120, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001357, Sample Num: 21712, Cur Loss: 0.21603197, Cur Avg Loss: 0.20616334, Log Avg loss: 0.16487799, Global Avg Loss: 0.76427903, Time: 0.0209 Steps: 80130, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001367, Sample Num: 21872, Cur Loss: 0.13200037, Cur Avg Loss: 0.20596797, Log Avg loss: 0.17945701, Global Avg Loss: 0.76420605, Time: 0.0208 Steps: 80140, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001377, Sample Num: 22032, Cur Loss: 0.17881009, Cur Avg Loss: 0.20614539, Log Avg loss: 0.23039866, Global Avg Loss: 0.76413945, Time: 0.0208 Steps: 80150, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001387, Sample Num: 22192, Cur Loss: 0.15406430, Cur Avg Loss: 0.20573088, Log Avg loss: 0.14865218, Global Avg Loss: 0.76406267, Time: 0.0209 Steps: 80160, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001397, Sample Num: 22352, Cur Loss: 0.10442297, Cur Avg Loss: 0.20532379, Log Avg loss: 0.14886036, Global Avg Loss: 0.76398593, Time: 0.0208 Steps: 80170, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001407, Sample Num: 22512, Cur Loss: 0.45840725, Cur Avg Loss: 0.20499622, Log Avg loss: 0.15923457, Global Avg Loss: 0.76391051, Time: 0.0208 Steps: 80180, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001417, Sample Num: 22672, Cur Loss: 0.08590122, Cur Avg Loss: 0.20456305, Log Avg loss: 0.14361720, Global Avg Loss: 0.76383315, Time: 0.0208 Steps: 80190, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001427, Sample Num: 22832, Cur Loss: 0.21969223, Cur Avg Loss: 0.20459946, Log Avg loss: 0.20975754, Global Avg Loss: 0.76376407, Time: 0.0209 Steps: 80200, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001437, Sample Num: 22992, Cur Loss: 0.52419657, Cur Avg Loss: 0.20517807, Log Avg loss: 0.28774669, Global Avg Loss: 0.76370472, Time: 0.0209 Steps: 80210, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001447, Sample Num: 23152, Cur Loss: 0.18346250, Cur Avg Loss: 0.20567444, Log Avg loss: 0.27700244, Global Avg Loss: 0.76364405, Time: 0.0208 Steps: 80220, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001457, Sample Num: 23312, Cur Loss: 0.19173966, Cur Avg Loss: 0.20593619, Log Avg loss: 0.24381163, Global Avg Loss: 0.76357926, Time: 0.0208 Steps: 80230, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001467, Sample Num: 23472, Cur Loss: 0.30114624, Cur Avg Loss: 0.20556925, Log Avg loss: 0.15210505, Global Avg Loss: 0.76350305, Time: 0.0208 Steps: 80240, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001477, Sample Num: 23632, Cur Loss: 0.08514035, Cur Avg Loss: 0.20556673, Log Avg loss: 0.20519706, Global Avg Loss: 0.76343348, Time: 0.0209 Steps: 80250, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001487, Sample Num: 23792, Cur Loss: 0.04866365, Cur Avg Loss: 0.20633266, Log Avg loss: 0.31946075, Global Avg Loss: 0.76337816, Time: 0.0209 Steps: 80260, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001497, Sample Num: 23952, Cur Loss: 0.04664690, Cur Avg Loss: 0.20622040, Log Avg loss: 0.18952794, Global Avg Loss: 0.76330667, Time: 0.0209 Steps: 80270, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001507, Sample Num: 24112, Cur Loss: 0.36104783, Cur Avg Loss: 0.20674056, Log Avg loss: 0.28460816, Global Avg Loss: 0.76324704, Time: 0.0208 Steps: 80280, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001517, Sample Num: 24272, Cur Loss: 0.12856874, Cur Avg Loss: 0.20707071, Log Avg loss: 0.25682428, Global Avg Loss: 0.76318397, Time: 0.0208 Steps: 80290, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001527, Sample Num: 24432, Cur Loss: 0.29302126, Cur Avg Loss: 0.20690126, Log Avg loss: 0.18119561, Global Avg Loss: 0.76311149, Time: 0.0208 Steps: 80300, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001537, Sample Num: 24592, Cur Loss: 0.13143450, Cur Avg Loss: 0.20644104, Log Avg loss: 0.13616636, Global Avg Loss: 0.76303343, Time: 0.0245 Steps: 80310, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001547, Sample Num: 24752, Cur Loss: 0.09460977, Cur Avg Loss: 0.20689134, Log Avg loss: 0.27610154, Global Avg Loss: 0.76297280, Time: 0.0208 Steps: 80320, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001557, Sample Num: 24912, Cur Loss: 0.25179225, Cur Avg Loss: 0.20698325, Log Avg loss: 0.22120224, Global Avg Loss: 0.76290536, Time: 0.0208 Steps: 80330, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001567, Sample Num: 25072, Cur Loss: 0.16835365, Cur Avg Loss: 0.20704033, Log Avg loss: 0.21592774, Global Avg Loss: 0.76283728, Time: 0.0208 Steps: 80340, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001577, Sample Num: 25232, Cur Loss: 0.15749210, Cur Avg Loss: 0.20679648, Log Avg loss: 0.16858549, Global Avg Loss: 0.76276332, Time: 0.0208 Steps: 80350, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001587, Sample Num: 25392, Cur Loss: 0.08292721, Cur Avg Loss: 0.20669595, Log Avg loss: 0.19084125, Global Avg Loss: 0.76269215, Time: 0.0208 Steps: 80360, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001597, Sample Num: 25552, Cur Loss: 0.23763204, Cur Avg Loss: 0.20647303, Log Avg loss: 0.17109610, Global Avg Loss: 0.76261854, Time: 0.0208 Steps: 80370, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001607, Sample Num: 25712, Cur Loss: 0.14210381, Cur Avg Loss: 0.20630251, Log Avg loss: 0.17907066, Global Avg Loss: 0.76254594, Time: 0.0208 Steps: 80380, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001617, Sample Num: 25872, Cur Loss: 0.40136492, Cur Avg Loss: 0.20630400, Log Avg loss: 0.20654399, Global Avg Loss: 0.76247678, Time: 0.0209 Steps: 80390, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001627, Sample Num: 26032, Cur Loss: 0.30345333, Cur Avg Loss: 0.20612325, Log Avg loss: 0.17689550, Global Avg Loss: 0.76240395, Time: 0.0208 Steps: 80400, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001637, Sample Num: 26192, Cur Loss: 0.25427362, Cur Avg Loss: 0.20637546, Log Avg loss: 0.24740956, Global Avg Loss: 0.76233990, Time: 0.0208 Steps: 80410, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001647, Sample Num: 26352, Cur Loss: 0.36886063, Cur Avg Loss: 0.20619643, Log Avg loss: 0.17688877, Global Avg Loss: 0.76226710, Time: 0.0208 Steps: 80420, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001657, Sample Num: 26512, Cur Loss: 0.36143845, Cur Avg Loss: 0.20614863, Log Avg loss: 0.19827654, Global Avg Loss: 0.76219698, Time: 0.0209 Steps: 80430, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001667, Sample Num: 26672, Cur Loss: 0.10163294, Cur Avg Loss: 0.20606709, Log Avg loss: 0.19255558, Global Avg Loss: 0.76212616, Time: 0.0208 Steps: 80440, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001677, Sample Num: 26832, Cur Loss: 0.49775237, Cur Avg Loss: 0.20616336, Log Avg loss: 0.22221184, Global Avg Loss: 0.76205905, Time: 0.0208 Steps: 80450, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001687, Sample Num: 26992, Cur Loss: 0.62406969, Cur Avg Loss: 0.20650523, Log Avg loss: 0.26383678, Global Avg Loss: 0.76199713, Time: 0.0208 Steps: 80460, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001697, Sample Num: 27152, Cur Loss: 0.09568629, Cur Avg Loss: 0.20628070, Log Avg loss: 0.16840340, Global Avg Loss: 0.76192336, Time: 0.0209 Steps: 80470, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001707, Sample Num: 27312, Cur Loss: 0.07909541, Cur Avg Loss: 0.20614631, Log Avg loss: 0.18333939, Global Avg Loss: 0.76185147, Time: 0.0208 Steps: 80480, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001717, Sample Num: 27472, Cur Loss: 0.14234796, Cur Avg Loss: 0.20608239, Log Avg loss: 0.19517189, Global Avg Loss: 0.76178107, Time: 0.0208 Steps: 80490, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001727, Sample Num: 27632, Cur Loss: 0.28697988, Cur Avg Loss: 0.20609292, Log Avg loss: 0.20790050, Global Avg Loss: 0.76171226, Time: 0.0208 Steps: 80500, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001737, Sample Num: 27792, Cur Loss: 0.40571821, Cur Avg Loss: 0.20618378, Log Avg loss: 0.22187558, Global Avg Loss: 0.76164521, Time: 0.0209 Steps: 80510, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001747, Sample Num: 27952, Cur Loss: 0.08848119, Cur Avg Loss: 0.20648827, Log Avg loss: 0.25937813, Global Avg Loss: 0.76158283, Time: 0.0209 Steps: 80520, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001757, Sample Num: 28112, Cur Loss: 0.03406236, Cur Avg Loss: 0.20615334, Log Avg loss: 0.14764166, Global Avg Loss: 0.76150660, Time: 0.0208 Steps: 80530, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001767, Sample Num: 28272, Cur Loss: 0.07444005, Cur Avg Loss: 0.20642195, Log Avg loss: 0.25361605, Global Avg Loss: 0.76144354, Time: 0.0208 Steps: 80540, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001777, Sample Num: 28432, Cur Loss: 0.12629066, Cur Avg Loss: 0.20638052, Log Avg loss: 0.19906021, Global Avg Loss: 0.76137372, Time: 0.0208 Steps: 80550, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001787, Sample Num: 28592, Cur Loss: 0.19320060, Cur Avg Loss: 0.20666933, Log Avg loss: 0.25799019, Global Avg Loss: 0.76131123, Time: 0.0208 Steps: 80560, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001797, Sample Num: 28752, Cur Loss: 0.30299717, Cur Avg Loss: 0.20685516, Log Avg loss: 0.24006355, Global Avg Loss: 0.76124654, Time: 0.0246 Steps: 80570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001807, Sample Num: 28912, Cur Loss: 0.05833385, Cur Avg Loss: 0.20678911, Log Avg loss: 0.19492015, Global Avg Loss: 0.76117626, Time: 0.0208 Steps: 80580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001817, Sample Num: 29072, Cur Loss: 0.17234452, Cur Avg Loss: 0.20701591, Log Avg loss: 0.24799816, Global Avg Loss: 0.76111258, Time: 0.0208 Steps: 80590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001827, Sample Num: 29232, Cur Loss: 0.12342734, Cur Avg Loss: 0.20670411, Log Avg loss: 0.15004939, Global Avg Loss: 0.76103676, Time: 0.0208 Steps: 80600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001837, Sample Num: 29392, Cur Loss: 0.09615269, Cur Avg Loss: 0.20656965, Log Avg loss: 0.18200371, Global Avg Loss: 0.76096493, Time: 0.0208 Steps: 80610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001847, Sample Num: 29552, Cur Loss: 0.16364837, Cur Avg Loss: 0.20674973, Log Avg loss: 0.23983089, Global Avg Loss: 0.76090029, Time: 0.0209 Steps: 80620, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001857, Sample Num: 29712, Cur Loss: 0.06881480, Cur Avg Loss: 0.20657064, Log Avg loss: 0.17349271, Global Avg Loss: 0.76082744, Time: 0.0209 Steps: 80630, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001867, Sample Num: 29872, Cur Loss: 0.20236135, Cur Avg Loss: 0.20676616, Log Avg loss: 0.24307439, Global Avg Loss: 0.76076323, Time: 0.0209 Steps: 80640, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001877, Sample Num: 30032, Cur Loss: 0.10198206, Cur Avg Loss: 0.20680516, Log Avg loss: 0.21408744, Global Avg Loss: 0.76069545, Time: 0.0209 Steps: 80650, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001887, Sample Num: 30192, Cur Loss: 0.32537723, Cur Avg Loss: 0.20715913, Log Avg loss: 0.27359869, Global Avg Loss: 0.76063506, Time: 0.0208 Steps: 80660, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001897, Sample Num: 30352, Cur Loss: 0.35654080, Cur Avg Loss: 0.20732550, Log Avg loss: 0.23871873, Global Avg Loss: 0.76057036, Time: 0.0208 Steps: 80670, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001907, Sample Num: 30512, Cur Loss: 0.18649957, Cur Avg Loss: 0.20747655, Log Avg loss: 0.23613061, Global Avg Loss: 0.76050536, Time: 0.0209 Steps: 80680, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001917, Sample Num: 30672, Cur Loss: 0.07684960, Cur Avg Loss: 0.20750910, Log Avg loss: 0.21371710, Global Avg Loss: 0.76043760, Time: 0.0209 Steps: 80690, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001927, Sample Num: 30832, Cur Loss: 0.10985461, Cur Avg Loss: 0.20721022, Log Avg loss: 0.14991538, Global Avg Loss: 0.76036194, Time: 0.0208 Steps: 80700, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001937, Sample Num: 30992, Cur Loss: 0.23999050, Cur Avg Loss: 0.20698262, Log Avg loss: 0.16312355, Global Avg Loss: 0.76028795, Time: 0.0208 Steps: 80710, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001947, Sample Num: 31152, Cur Loss: 0.21403742, Cur Avg Loss: 0.20712995, Log Avg loss: 0.23566861, Global Avg Loss: 0.76022295, Time: 0.0208 Steps: 80720, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001957, Sample Num: 31312, Cur Loss: 0.09942505, Cur Avg Loss: 0.20685361, Log Avg loss: 0.15304880, Global Avg Loss: 0.76014774, Time: 0.0208 Steps: 80730, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001967, Sample Num: 31472, Cur Loss: 0.21646793, Cur Avg Loss: 0.20715477, Log Avg loss: 0.26609327, Global Avg Loss: 0.76008655, Time: 0.0208 Steps: 80740, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001977, Sample Num: 31632, Cur Loss: 0.33715042, Cur Avg Loss: 0.20706529, Log Avg loss: 0.18946303, Global Avg Loss: 0.76001589, Time: 0.0209 Steps: 80750, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001987, Sample Num: 31792, Cur Loss: 0.15326118, Cur Avg Loss: 0.20720284, Log Avg loss: 0.23439793, Global Avg Loss: 0.75995080, Time: 0.0208 Steps: 80760, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001997, Sample Num: 31952, Cur Loss: 0.04293303, Cur Avg Loss: 0.20750841, Log Avg loss: 0.26822498, Global Avg Loss: 0.75988992, Time: 0.0208 Steps: 80770, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002007, Sample Num: 32112, Cur Loss: 0.17206934, Cur Avg Loss: 0.20747591, Log Avg loss: 0.20098518, Global Avg Loss: 0.75982073, Time: 0.0208 Steps: 80780, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002017, Sample Num: 32272, Cur Loss: 0.16663972, Cur Avg Loss: 0.20734291, Log Avg loss: 0.18064910, Global Avg Loss: 0.75974905, Time: 0.0208 Steps: 80790, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002027, Sample Num: 32432, Cur Loss: 0.29561210, Cur Avg Loss: 0.20733497, Log Avg loss: 0.20573505, Global Avg Loss: 0.75968048, Time: 0.0208 Steps: 80800, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002037, Sample Num: 32592, Cur Loss: 0.14114651, Cur Avg Loss: 0.20728472, Log Avg loss: 0.19709911, Global Avg Loss: 0.75961086, Time: 0.0208 Steps: 80810, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002047, Sample Num: 32752, Cur Loss: 0.09375934, Cur Avg Loss: 0.20741201, Log Avg loss: 0.23333952, Global Avg Loss: 0.75954575, Time: 0.0209 Steps: 80820, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002057, Sample Num: 32912, Cur Loss: 0.19774579, Cur Avg Loss: 0.20767928, Log Avg loss: 0.26239082, Global Avg Loss: 0.75948424, Time: 0.0209 Steps: 80830, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002067, Sample Num: 33072, Cur Loss: 0.55324519, Cur Avg Loss: 0.20786804, Log Avg loss: 0.24669447, Global Avg Loss: 0.75942081, Time: 0.0208 Steps: 80840, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002077, Sample Num: 33232, Cur Loss: 0.09801154, Cur Avg Loss: 0.20757856, Log Avg loss: 0.14774409, Global Avg Loss: 0.75934515, Time: 0.0208 Steps: 80850, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002087, Sample Num: 33392, Cur Loss: 0.18442033, Cur Avg Loss: 0.20728997, Log Avg loss: 0.14734871, Global Avg Loss: 0.75926946, Time: 0.0208 Steps: 80860, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002097, Sample Num: 33552, Cur Loss: 0.38180691, Cur Avg Loss: 0.20710541, Log Avg loss: 0.16858845, Global Avg Loss: 0.75919642, Time: 0.0208 Steps: 80870, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002107, Sample Num: 33712, Cur Loss: 0.10056261, Cur Avg Loss: 0.20684334, Log Avg loss: 0.15188672, Global Avg Loss: 0.75912134, Time: 0.0208 Steps: 80880, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002117, Sample Num: 33872, Cur Loss: 0.23401862, Cur Avg Loss: 0.20665044, Log Avg loss: 0.16600748, Global Avg Loss: 0.75904801, Time: 0.0208 Steps: 80890, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002127, Sample Num: 34032, Cur Loss: 0.14955159, Cur Avg Loss: 0.20690548, Log Avg loss: 0.26089684, Global Avg Loss: 0.75898644, Time: 0.0208 Steps: 80900, Updated lr: 0.000024 ***** Running evaluation checkpoint-80902 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-80902 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.706666, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.40578, "eval_total_loss": 285.263025, "eval_mae": 0.544443, "eval_mse": 0.40575, "eval_r2": 0.742079, "eval_sp_statistic": 0.895358, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.919691, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.335383, "test_total_loss": 168.362164, "test_mae": 0.445702, "test_mse": 0.335481, "test_r2": 0.783478, "test_sp_statistic": 0.868002, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.913469, "test_ps_pvalue": 0.0, "lr": 2.4227596017069703e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7589744045450817, "train_cur_epoch_loss": 440.63253243640065, "train_cur_epoch_avg_loss": 0.20696690109741694, "train_cur_epoch_time": 44.706666231155396, "train_cur_epoch_avg_time": 0.020998903819236917, "epoch": 38, "step": 80902} ################################################## Training, Epoch: 0039, Batch: 000008, Sample Num: 128, Cur Loss: 0.29198292, Cur Avg Loss: 0.22754444, Log Avg loss: 0.23649321, Global Avg Loss: 0.75892186, Time: 0.0208 Steps: 80910, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000018, Sample Num: 288, Cur Loss: 0.05607315, Cur Avg Loss: 0.18204677, Log Avg loss: 0.14564863, Global Avg Loss: 0.75884607, Time: 0.0209 Steps: 80920, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000028, Sample Num: 448, Cur Loss: 0.24820888, Cur Avg Loss: 0.16940323, Log Avg loss: 0.14664486, Global Avg Loss: 0.75877043, Time: 0.0209 Steps: 80930, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000038, Sample Num: 608, Cur Loss: 0.16775200, Cur Avg Loss: 0.16283565, Log Avg loss: 0.14444643, Global Avg Loss: 0.75869453, Time: 0.0209 Steps: 80940, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000048, Sample Num: 768, Cur Loss: 0.07647038, Cur Avg Loss: 0.17230429, Log Avg loss: 0.20828511, Global Avg Loss: 0.75862653, Time: 0.0209 Steps: 80950, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000058, Sample Num: 928, Cur Loss: 0.25124192, Cur Avg Loss: 0.17488644, Log Avg loss: 0.18728077, Global Avg Loss: 0.75855596, Time: 0.0209 Steps: 80960, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000068, Sample Num: 1088, Cur Loss: 0.08338219, Cur Avg Loss: 0.16833370, Log Avg loss: 0.13032784, Global Avg Loss: 0.75847837, Time: 0.0208 Steps: 80970, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000078, Sample Num: 1248, Cur Loss: 0.25309974, Cur Avg Loss: 0.16526472, Log Avg loss: 0.14439566, Global Avg Loss: 0.75840254, Time: 0.0209 Steps: 80980, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000088, Sample Num: 1408, Cur Loss: 0.53593886, Cur Avg Loss: 0.16890855, Log Avg loss: 0.19733038, Global Avg Loss: 0.75833327, Time: 0.0209 Steps: 80990, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000098, Sample Num: 1568, Cur Loss: 0.11485539, Cur Avg Loss: 0.17292880, Log Avg loss: 0.20830706, Global Avg Loss: 0.75826536, Time: 0.0209 Steps: 81000, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000108, Sample Num: 1728, Cur Loss: 0.27383143, Cur Avg Loss: 0.17277336, Log Avg loss: 0.17125002, Global Avg Loss: 0.75819290, Time: 0.0209 Steps: 81010, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000118, Sample Num: 1888, Cur Loss: 0.06167085, Cur Avg Loss: 0.17266937, Log Avg loss: 0.17154632, Global Avg Loss: 0.75812049, Time: 0.0209 Steps: 81020, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000128, Sample Num: 2048, Cur Loss: 0.37300396, Cur Avg Loss: 0.17550834, Log Avg loss: 0.20900816, Global Avg Loss: 0.75805273, Time: 0.0209 Steps: 81030, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000138, Sample Num: 2208, Cur Loss: 0.06060426, Cur Avg Loss: 0.17796498, Log Avg loss: 0.20941001, Global Avg Loss: 0.75798503, Time: 0.0209 Steps: 81040, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000148, Sample Num: 2368, Cur Loss: 0.09463517, Cur Avg Loss: 0.18123750, Log Avg loss: 0.22639818, Global Avg Loss: 0.75791944, Time: 0.0209 Steps: 81050, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000158, Sample Num: 2528, Cur Loss: 0.30396122, Cur Avg Loss: 0.18450495, Log Avg loss: 0.23286331, Global Avg Loss: 0.75785466, Time: 0.0208 Steps: 81060, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000168, Sample Num: 2688, Cur Loss: 0.20235616, Cur Avg Loss: 0.18684337, Log Avg loss: 0.22379030, Global Avg Loss: 0.75778879, Time: 0.0208 Steps: 81070, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000178, Sample Num: 2848, Cur Loss: 0.43645188, Cur Avg Loss: 0.19332742, Log Avg loss: 0.30225945, Global Avg Loss: 0.75773260, Time: 0.0210 Steps: 81080, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000188, Sample Num: 3008, Cur Loss: 0.10138030, Cur Avg Loss: 0.19382709, Log Avg loss: 0.20272132, Global Avg Loss: 0.75766416, Time: 0.0209 Steps: 81090, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000198, Sample Num: 3168, Cur Loss: 0.09226517, Cur Avg Loss: 0.19137334, Log Avg loss: 0.14524277, Global Avg Loss: 0.75758865, Time: 0.0209 Steps: 81100, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000208, Sample Num: 3328, Cur Loss: 0.17087889, Cur Avg Loss: 0.19265233, Log Avg loss: 0.21797645, Global Avg Loss: 0.75752212, Time: 0.0209 Steps: 81110, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000218, Sample Num: 3488, Cur Loss: 0.20762591, Cur Avg Loss: 0.19473971, Log Avg loss: 0.23815704, Global Avg Loss: 0.75745809, Time: 0.0208 Steps: 81120, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000228, Sample Num: 3648, Cur Loss: 0.55155885, Cur Avg Loss: 0.19828920, Log Avg loss: 0.27566812, Global Avg Loss: 0.75739871, Time: 0.0208 Steps: 81130, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000238, Sample Num: 3808, Cur Loss: 0.14702810, Cur Avg Loss: 0.19945420, Log Avg loss: 0.22601617, Global Avg Loss: 0.75733322, Time: 0.0209 Steps: 81140, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000248, Sample Num: 3968, Cur Loss: 0.06011115, Cur Avg Loss: 0.19902453, Log Avg loss: 0.18879849, Global Avg Loss: 0.75726316, Time: 0.0209 Steps: 81150, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000258, Sample Num: 4128, Cur Loss: 0.56606644, Cur Avg Loss: 0.19766957, Log Avg loss: 0.16406655, Global Avg Loss: 0.75719007, Time: 0.0249 Steps: 81160, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000268, Sample Num: 4288, Cur Loss: 0.08386239, Cur Avg Loss: 0.19725008, Log Avg loss: 0.18642707, Global Avg Loss: 0.75711975, Time: 0.0211 Steps: 81170, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000278, Sample Num: 4448, Cur Loss: 0.14078806, Cur Avg Loss: 0.19750649, Log Avg loss: 0.20437836, Global Avg Loss: 0.75705166, Time: 0.0208 Steps: 81180, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000288, Sample Num: 4608, Cur Loss: 0.11723226, Cur Avg Loss: 0.19813500, Log Avg loss: 0.21560749, Global Avg Loss: 0.75698498, Time: 0.0210 Steps: 81190, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000298, Sample Num: 4768, Cur Loss: 0.09379187, Cur Avg Loss: 0.19593440, Log Avg loss: 0.13255741, Global Avg Loss: 0.75690808, Time: 0.0209 Steps: 81200, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000308, Sample Num: 4928, Cur Loss: 0.10539335, Cur Avg Loss: 0.19553585, Log Avg loss: 0.18365884, Global Avg Loss: 0.75683749, Time: 0.0210 Steps: 81210, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000318, Sample Num: 5088, Cur Loss: 0.13650446, Cur Avg Loss: 0.19559798, Log Avg loss: 0.19751171, Global Avg Loss: 0.75676862, Time: 0.0211 Steps: 81220, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000328, Sample Num: 5248, Cur Loss: 0.16560771, Cur Avg Loss: 0.19504318, Log Avg loss: 0.17740064, Global Avg Loss: 0.75669730, Time: 0.0209 Steps: 81230, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000338, Sample Num: 5408, Cur Loss: 0.09213449, Cur Avg Loss: 0.19481807, Log Avg loss: 0.18743422, Global Avg Loss: 0.75662723, Time: 0.0209 Steps: 81240, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000348, Sample Num: 5568, Cur Loss: 0.25691229, Cur Avg Loss: 0.19662365, Log Avg loss: 0.25765238, Global Avg Loss: 0.75656581, Time: 0.0211 Steps: 81250, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000358, Sample Num: 5728, Cur Loss: 0.10970346, Cur Avg Loss: 0.19720965, Log Avg loss: 0.21760250, Global Avg Loss: 0.75649949, Time: 0.0209 Steps: 81260, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000368, Sample Num: 5888, Cur Loss: 0.14548023, Cur Avg Loss: 0.19811448, Log Avg loss: 0.23050728, Global Avg Loss: 0.75643477, Time: 0.0210 Steps: 81270, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000378, Sample Num: 6048, Cur Loss: 0.16378170, Cur Avg Loss: 0.19771445, Log Avg loss: 0.18299322, Global Avg Loss: 0.75636421, Time: 0.0209 Steps: 81280, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000388, Sample Num: 6208, Cur Loss: 0.15995152, Cur Avg Loss: 0.19743349, Log Avg loss: 0.18681320, Global Avg Loss: 0.75629415, Time: 0.0211 Steps: 81290, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000398, Sample Num: 6368, Cur Loss: 0.20773894, Cur Avg Loss: 0.19650535, Log Avg loss: 0.16049368, Global Avg Loss: 0.75622087, Time: 0.0211 Steps: 81300, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000408, Sample Num: 6528, Cur Loss: 0.17713152, Cur Avg Loss: 0.19676642, Log Avg loss: 0.20715703, Global Avg Loss: 0.75615334, Time: 0.0209 Steps: 81310, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000418, Sample Num: 6688, Cur Loss: 0.15953302, Cur Avg Loss: 0.19694387, Log Avg loss: 0.20418396, Global Avg Loss: 0.75608546, Time: 0.0210 Steps: 81320, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000428, Sample Num: 6848, Cur Loss: 0.29727691, Cur Avg Loss: 0.19690650, Log Avg loss: 0.19534425, Global Avg Loss: 0.75601652, Time: 0.0209 Steps: 81330, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000438, Sample Num: 7008, Cur Loss: 0.06653041, Cur Avg Loss: 0.19669773, Log Avg loss: 0.18776239, Global Avg Loss: 0.75594665, Time: 0.0209 Steps: 81340, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000448, Sample Num: 7168, Cur Loss: 0.22933102, Cur Avg Loss: 0.19493488, Log Avg loss: 0.11772210, Global Avg Loss: 0.75586820, Time: 0.0211 Steps: 81350, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000458, Sample Num: 7328, Cur Loss: 0.05204773, Cur Avg Loss: 0.19378823, Log Avg loss: 0.14241831, Global Avg Loss: 0.75579280, Time: 0.0209 Steps: 81360, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000468, Sample Num: 7488, Cur Loss: 0.20904627, Cur Avg Loss: 0.19274729, Log Avg loss: 0.14507209, Global Avg Loss: 0.75571775, Time: 0.0208 Steps: 81370, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000478, Sample Num: 7648, Cur Loss: 0.25026426, Cur Avg Loss: 0.19244400, Log Avg loss: 0.17825015, Global Avg Loss: 0.75564679, Time: 0.0211 Steps: 81380, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000488, Sample Num: 7808, Cur Loss: 0.30716074, Cur Avg Loss: 0.19242130, Log Avg loss: 0.19133631, Global Avg Loss: 0.75557745, Time: 0.0209 Steps: 81390, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000498, Sample Num: 7968, Cur Loss: 0.13075033, Cur Avg Loss: 0.19320460, Log Avg loss: 0.23142959, Global Avg Loss: 0.75551306, Time: 0.0209 Steps: 81400, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000508, Sample Num: 8128, Cur Loss: 0.58511889, Cur Avg Loss: 0.19559359, Log Avg loss: 0.31456536, Global Avg Loss: 0.75545890, Time: 0.0211 Steps: 81410, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000518, Sample Num: 8288, Cur Loss: 0.12019198, Cur Avg Loss: 0.19620088, Log Avg loss: 0.22705085, Global Avg Loss: 0.75539400, Time: 0.0247 Steps: 81420, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000528, Sample Num: 8448, Cur Loss: 0.16927531, Cur Avg Loss: 0.19633181, Log Avg loss: 0.20311445, Global Avg Loss: 0.75532618, Time: 0.0209 Steps: 81430, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000538, Sample Num: 8608, Cur Loss: 0.14416863, Cur Avg Loss: 0.19512415, Log Avg loss: 0.13135955, Global Avg Loss: 0.75524956, Time: 0.0210 Steps: 81440, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000548, Sample Num: 8768, Cur Loss: 0.22360176, Cur Avg Loss: 0.19545118, Log Avg loss: 0.21304545, Global Avg Loss: 0.75518299, Time: 0.0210 Steps: 81450, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000558, Sample Num: 8928, Cur Loss: 0.11394855, Cur Avg Loss: 0.19539558, Log Avg loss: 0.19234877, Global Avg Loss: 0.75511390, Time: 0.0210 Steps: 81460, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000568, Sample Num: 9088, Cur Loss: 0.08183399, Cur Avg Loss: 0.19419742, Log Avg loss: 0.12734005, Global Avg Loss: 0.75503684, Time: 0.0210 Steps: 81470, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000578, Sample Num: 9248, Cur Loss: 0.04677992, Cur Avg Loss: 0.19555290, Log Avg loss: 0.27254416, Global Avg Loss: 0.75497762, Time: 0.0210 Steps: 81480, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000588, Sample Num: 9408, Cur Loss: 0.46408859, Cur Avg Loss: 0.19592948, Log Avg loss: 0.21769540, Global Avg Loss: 0.75491169, Time: 0.0210 Steps: 81490, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000598, Sample Num: 9568, Cur Loss: 0.18091601, Cur Avg Loss: 0.19680239, Log Avg loss: 0.24812994, Global Avg Loss: 0.75484951, Time: 0.0210 Steps: 81500, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000608, Sample Num: 9728, Cur Loss: 0.23931642, Cur Avg Loss: 0.19634390, Log Avg loss: 0.16892575, Global Avg Loss: 0.75477763, Time: 0.0210 Steps: 81510, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000618, Sample Num: 9888, Cur Loss: 0.31627524, Cur Avg Loss: 0.19620362, Log Avg loss: 0.18767484, Global Avg Loss: 0.75470806, Time: 0.0210 Steps: 81520, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000628, Sample Num: 10048, Cur Loss: 0.24204600, Cur Avg Loss: 0.19631477, Log Avg loss: 0.20318368, Global Avg Loss: 0.75464041, Time: 0.0209 Steps: 81530, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000638, Sample Num: 10208, Cur Loss: 0.05861191, Cur Avg Loss: 0.19752443, Log Avg loss: 0.27349126, Global Avg Loss: 0.75458141, Time: 0.0210 Steps: 81540, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000648, Sample Num: 10368, Cur Loss: 0.06052279, Cur Avg Loss: 0.19642020, Log Avg loss: 0.12597040, Global Avg Loss: 0.75450432, Time: 0.0210 Steps: 81550, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000658, Sample Num: 10528, Cur Loss: 0.19595738, Cur Avg Loss: 0.19580111, Log Avg loss: 0.15568390, Global Avg Loss: 0.75443090, Time: 0.0210 Steps: 81560, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000668, Sample Num: 10688, Cur Loss: 0.11237128, Cur Avg Loss: 0.19500413, Log Avg loss: 0.14256262, Global Avg Loss: 0.75435589, Time: 0.0209 Steps: 81570, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000678, Sample Num: 10848, Cur Loss: 0.16525079, Cur Avg Loss: 0.19479504, Log Avg loss: 0.18082802, Global Avg Loss: 0.75428559, Time: 0.0209 Steps: 81580, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000688, Sample Num: 11008, Cur Loss: 0.08820394, Cur Avg Loss: 0.19478480, Log Avg loss: 0.19409079, Global Avg Loss: 0.75421693, Time: 0.0209 Steps: 81590, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000698, Sample Num: 11168, Cur Loss: 0.27679390, Cur Avg Loss: 0.19450558, Log Avg loss: 0.17529517, Global Avg Loss: 0.75414598, Time: 0.0210 Steps: 81600, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000708, Sample Num: 11328, Cur Loss: 0.53063738, Cur Avg Loss: 0.19798796, Log Avg loss: 0.44105784, Global Avg Loss: 0.75410762, Time: 0.0209 Steps: 81610, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000718, Sample Num: 11488, Cur Loss: 0.11088143, Cur Avg Loss: 0.19667854, Log Avg loss: 0.10397201, Global Avg Loss: 0.75402796, Time: 0.0210 Steps: 81620, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000728, Sample Num: 11648, Cur Loss: 0.28350136, Cur Avg Loss: 0.19825720, Log Avg loss: 0.31160489, Global Avg Loss: 0.75397377, Time: 0.0209 Steps: 81630, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000738, Sample Num: 11808, Cur Loss: 0.34048617, Cur Avg Loss: 0.19882494, Log Avg loss: 0.24015632, Global Avg Loss: 0.75391083, Time: 0.0210 Steps: 81640, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000748, Sample Num: 11968, Cur Loss: 0.23411666, Cur Avg Loss: 0.19850148, Log Avg loss: 0.17463010, Global Avg Loss: 0.75383988, Time: 0.0209 Steps: 81650, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000758, Sample Num: 12128, Cur Loss: 0.14605802, Cur Avg Loss: 0.20071315, Log Avg loss: 0.36614596, Global Avg Loss: 0.75379241, Time: 0.0210 Steps: 81660, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000768, Sample Num: 12288, Cur Loss: 0.07249399, Cur Avg Loss: 0.19995008, Log Avg loss: 0.14210942, Global Avg Loss: 0.75371751, Time: 0.0255 Steps: 81670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000778, Sample Num: 12448, Cur Loss: 0.27857682, Cur Avg Loss: 0.20079755, Log Avg loss: 0.26588357, Global Avg Loss: 0.75365778, Time: 0.0210 Steps: 81680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000788, Sample Num: 12608, Cur Loss: 0.14843221, Cur Avg Loss: 0.20032587, Log Avg loss: 0.16362922, Global Avg Loss: 0.75358556, Time: 0.0208 Steps: 81690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000798, Sample Num: 12768, Cur Loss: 0.21012241, Cur Avg Loss: 0.20043912, Log Avg loss: 0.20936326, Global Avg Loss: 0.75351894, Time: 0.0209 Steps: 81700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000808, Sample Num: 12928, Cur Loss: 0.06803131, Cur Avg Loss: 0.20115510, Log Avg loss: 0.25829006, Global Avg Loss: 0.75345834, Time: 0.0209 Steps: 81710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000818, Sample Num: 13088, Cur Loss: 0.11113198, Cur Avg Loss: 0.20136954, Log Avg loss: 0.21869587, Global Avg Loss: 0.75339290, Time: 0.0208 Steps: 81720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000828, Sample Num: 13248, Cur Loss: 0.13262370, Cur Avg Loss: 0.20170103, Log Avg loss: 0.22881713, Global Avg Loss: 0.75332871, Time: 0.0209 Steps: 81730, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000838, Sample Num: 13408, Cur Loss: 0.21782172, Cur Avg Loss: 0.20104958, Log Avg loss: 0.14710980, Global Avg Loss: 0.75325455, Time: 0.0208 Steps: 81740, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000848, Sample Num: 13568, Cur Loss: 0.07576222, Cur Avg Loss: 0.20079055, Log Avg loss: 0.17908389, Global Avg Loss: 0.75318431, Time: 0.0210 Steps: 81750, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000858, Sample Num: 13728, Cur Loss: 0.10601211, Cur Avg Loss: 0.20082445, Log Avg loss: 0.20369928, Global Avg Loss: 0.75311711, Time: 0.0210 Steps: 81760, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000868, Sample Num: 13888, Cur Loss: 0.25418198, Cur Avg Loss: 0.20116211, Log Avg loss: 0.23013286, Global Avg Loss: 0.75305315, Time: 0.0209 Steps: 81770, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000878, Sample Num: 14048, Cur Loss: 0.41857189, Cur Avg Loss: 0.20148325, Log Avg loss: 0.22935794, Global Avg Loss: 0.75298911, Time: 0.0210 Steps: 81780, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000888, Sample Num: 14208, Cur Loss: 0.15519996, Cur Avg Loss: 0.20094314, Log Avg loss: 0.15352177, Global Avg Loss: 0.75291582, Time: 0.0208 Steps: 81790, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000898, Sample Num: 14368, Cur Loss: 0.18729918, Cur Avg Loss: 0.20128397, Log Avg loss: 0.23154985, Global Avg Loss: 0.75285208, Time: 0.0210 Steps: 81800, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000908, Sample Num: 14528, Cur Loss: 0.09438881, Cur Avg Loss: 0.20093014, Log Avg loss: 0.16915634, Global Avg Loss: 0.75278073, Time: 0.0208 Steps: 81810, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000918, Sample Num: 14688, Cur Loss: 0.27365246, Cur Avg Loss: 0.20075181, Log Avg loss: 0.18455894, Global Avg Loss: 0.75271129, Time: 0.0208 Steps: 81820, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000928, Sample Num: 14848, Cur Loss: 0.09070735, Cur Avg Loss: 0.20148330, Log Avg loss: 0.26863406, Global Avg Loss: 0.75265213, Time: 0.0210 Steps: 81830, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000938, Sample Num: 15008, Cur Loss: 0.14709786, Cur Avg Loss: 0.20119573, Log Avg loss: 0.17450982, Global Avg Loss: 0.75258149, Time: 0.0210 Steps: 81840, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000948, Sample Num: 15168, Cur Loss: 0.09246239, Cur Avg Loss: 0.20151980, Log Avg loss: 0.23191713, Global Avg Loss: 0.75251787, Time: 0.0208 Steps: 81850, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000958, Sample Num: 15328, Cur Loss: 0.30598193, Cur Avg Loss: 0.20179056, Log Avg loss: 0.22745876, Global Avg Loss: 0.75245373, Time: 0.0210 Steps: 81860, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000968, Sample Num: 15488, Cur Loss: 0.16261441, Cur Avg Loss: 0.20148949, Log Avg loss: 0.17264673, Global Avg Loss: 0.75238291, Time: 0.0210 Steps: 81870, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000978, Sample Num: 15648, Cur Loss: 0.07394168, Cur Avg Loss: 0.20111698, Log Avg loss: 0.16505829, Global Avg Loss: 0.75231118, Time: 0.0208 Steps: 81880, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000988, Sample Num: 15808, Cur Loss: 0.42816839, Cur Avg Loss: 0.20113784, Log Avg loss: 0.20317813, Global Avg Loss: 0.75224413, Time: 0.0211 Steps: 81890, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000998, Sample Num: 15968, Cur Loss: 0.12936032, Cur Avg Loss: 0.20083482, Log Avg loss: 0.17089579, Global Avg Loss: 0.75217314, Time: 0.0210 Steps: 81900, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001008, Sample Num: 16128, Cur Loss: 0.38347232, Cur Avg Loss: 0.20127358, Log Avg loss: 0.24506250, Global Avg Loss: 0.75211123, Time: 0.0210 Steps: 81910, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001018, Sample Num: 16288, Cur Loss: 0.56748885, Cur Avg Loss: 0.20158565, Log Avg loss: 0.23304230, Global Avg Loss: 0.75204787, Time: 0.0210 Steps: 81920, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001028, Sample Num: 16448, Cur Loss: 0.20041497, Cur Avg Loss: 0.20229038, Log Avg loss: 0.27403207, Global Avg Loss: 0.75198953, Time: 0.0246 Steps: 81930, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001038, Sample Num: 16608, Cur Loss: 0.41729403, Cur Avg Loss: 0.20235186, Log Avg loss: 0.20867159, Global Avg Loss: 0.75192322, Time: 0.0209 Steps: 81940, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001048, Sample Num: 16768, Cur Loss: 0.12987590, Cur Avg Loss: 0.20252968, Log Avg loss: 0.22098770, Global Avg Loss: 0.75185843, Time: 0.0208 Steps: 81950, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001058, Sample Num: 16928, Cur Loss: 0.30362031, Cur Avg Loss: 0.20260276, Log Avg loss: 0.21026154, Global Avg Loss: 0.75179235, Time: 0.0209 Steps: 81960, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001068, Sample Num: 17088, Cur Loss: 0.14007206, Cur Avg Loss: 0.20205050, Log Avg loss: 0.14362097, Global Avg Loss: 0.75171816, Time: 0.0209 Steps: 81970, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001078, Sample Num: 17248, Cur Loss: 0.25199822, Cur Avg Loss: 0.20186399, Log Avg loss: 0.18194468, Global Avg Loss: 0.75164865, Time: 0.0210 Steps: 81980, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001088, Sample Num: 17408, Cur Loss: 0.11249954, Cur Avg Loss: 0.20167349, Log Avg loss: 0.18113829, Global Avg Loss: 0.75157907, Time: 0.0210 Steps: 81990, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001098, Sample Num: 17568, Cur Loss: 0.19675252, Cur Avg Loss: 0.20201241, Log Avg loss: 0.23888674, Global Avg Loss: 0.75151655, Time: 0.0211 Steps: 82000, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001108, Sample Num: 17728, Cur Loss: 0.09121947, Cur Avg Loss: 0.20225613, Log Avg loss: 0.22901575, Global Avg Loss: 0.75145284, Time: 0.0210 Steps: 82010, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001118, Sample Num: 17888, Cur Loss: 0.03976788, Cur Avg Loss: 0.20200382, Log Avg loss: 0.17404808, Global Avg Loss: 0.75138244, Time: 0.0210 Steps: 82020, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001128, Sample Num: 18048, Cur Loss: 0.09110567, Cur Avg Loss: 0.20179042, Log Avg loss: 0.17793286, Global Avg Loss: 0.75131253, Time: 0.0210 Steps: 82030, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001138, Sample Num: 18208, Cur Loss: 0.15753160, Cur Avg Loss: 0.20146939, Log Avg loss: 0.16525700, Global Avg Loss: 0.75124110, Time: 0.0210 Steps: 82040, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001148, Sample Num: 18368, Cur Loss: 0.08455762, Cur Avg Loss: 0.20159537, Log Avg loss: 0.21593240, Global Avg Loss: 0.75117585, Time: 0.0210 Steps: 82050, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001158, Sample Num: 18528, Cur Loss: 0.07862297, Cur Avg Loss: 0.20139993, Log Avg loss: 0.17896245, Global Avg Loss: 0.75110612, Time: 0.0211 Steps: 82060, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001168, Sample Num: 18688, Cur Loss: 0.15310679, Cur Avg Loss: 0.20129878, Log Avg loss: 0.18958602, Global Avg Loss: 0.75103770, Time: 0.0211 Steps: 82070, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001178, Sample Num: 18848, Cur Loss: 0.14050813, Cur Avg Loss: 0.20096700, Log Avg loss: 0.16221502, Global Avg Loss: 0.75096596, Time: 0.0211 Steps: 82080, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001188, Sample Num: 19008, Cur Loss: 0.36731830, Cur Avg Loss: 0.20149047, Log Avg loss: 0.26315486, Global Avg Loss: 0.75090654, Time: 0.0210 Steps: 82090, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001198, Sample Num: 19168, Cur Loss: 0.23041904, Cur Avg Loss: 0.20133737, Log Avg loss: 0.18314976, Global Avg Loss: 0.75083739, Time: 0.0211 Steps: 82100, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001208, Sample Num: 19328, Cur Loss: 0.13639042, Cur Avg Loss: 0.20099932, Log Avg loss: 0.16050038, Global Avg Loss: 0.75076549, Time: 0.0210 Steps: 82110, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001218, Sample Num: 19488, Cur Loss: 0.16090915, Cur Avg Loss: 0.20077599, Log Avg loss: 0.17379800, Global Avg Loss: 0.75069523, Time: 0.0210 Steps: 82120, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001228, Sample Num: 19648, Cur Loss: 0.24693832, Cur Avg Loss: 0.20143982, Log Avg loss: 0.28229477, Global Avg Loss: 0.75063820, Time: 0.0210 Steps: 82130, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001238, Sample Num: 19808, Cur Loss: 0.15645754, Cur Avg Loss: 0.20108324, Log Avg loss: 0.15729525, Global Avg Loss: 0.75056596, Time: 0.0210 Steps: 82140, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001248, Sample Num: 19968, Cur Loss: 0.16241290, Cur Avg Loss: 0.20106570, Log Avg loss: 0.19889428, Global Avg Loss: 0.75049881, Time: 0.0210 Steps: 82150, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001258, Sample Num: 20128, Cur Loss: 0.33057863, Cur Avg Loss: 0.20216195, Log Avg loss: 0.33897396, Global Avg Loss: 0.75044872, Time: 0.0210 Steps: 82160, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001268, Sample Num: 20288, Cur Loss: 0.09855795, Cur Avg Loss: 0.20191492, Log Avg loss: 0.17083820, Global Avg Loss: 0.75037818, Time: 0.0210 Steps: 82170, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001278, Sample Num: 20448, Cur Loss: 0.11813742, Cur Avg Loss: 0.20209250, Log Avg loss: 0.22460915, Global Avg Loss: 0.75031421, Time: 0.0210 Steps: 82180, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001288, Sample Num: 20608, Cur Loss: 0.23751962, Cur Avg Loss: 0.20256222, Log Avg loss: 0.26259281, Global Avg Loss: 0.75025487, Time: 0.0211 Steps: 82190, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001298, Sample Num: 20768, Cur Loss: 0.17592593, Cur Avg Loss: 0.20307532, Log Avg loss: 0.26916317, Global Avg Loss: 0.75019634, Time: 0.0210 Steps: 82200, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001308, Sample Num: 20928, Cur Loss: 0.11090282, Cur Avg Loss: 0.20278730, Log Avg loss: 0.16540174, Global Avg Loss: 0.75012520, Time: 0.0210 Steps: 82210, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001318, Sample Num: 21088, Cur Loss: 0.14006734, Cur Avg Loss: 0.20241136, Log Avg loss: 0.15323882, Global Avg Loss: 0.75005261, Time: 0.0210 Steps: 82220, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001328, Sample Num: 21248, Cur Loss: 0.14467102, Cur Avg Loss: 0.20224139, Log Avg loss: 0.17983887, Global Avg Loss: 0.74998326, Time: 0.0210 Steps: 82230, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001338, Sample Num: 21408, Cur Loss: 0.11778486, Cur Avg Loss: 0.20258690, Log Avg loss: 0.24847047, Global Avg Loss: 0.74992228, Time: 0.0210 Steps: 82240, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001348, Sample Num: 21568, Cur Loss: 0.51947713, Cur Avg Loss: 0.20279550, Log Avg loss: 0.23070690, Global Avg Loss: 0.74985916, Time: 0.0210 Steps: 82250, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001358, Sample Num: 21728, Cur Loss: 0.16372614, Cur Avg Loss: 0.20237597, Log Avg loss: 0.14582322, Global Avg Loss: 0.74978573, Time: 0.0210 Steps: 82260, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001368, Sample Num: 21888, Cur Loss: 0.22052500, Cur Avg Loss: 0.20272842, Log Avg loss: 0.25059053, Global Avg Loss: 0.74972505, Time: 0.0210 Steps: 82270, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001378, Sample Num: 22048, Cur Loss: 0.31554922, Cur Avg Loss: 0.20279021, Log Avg loss: 0.21124365, Global Avg Loss: 0.74965960, Time: 0.0210 Steps: 82280, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001388, Sample Num: 22208, Cur Loss: 0.07513367, Cur Avg Loss: 0.20248982, Log Avg loss: 0.16109591, Global Avg Loss: 0.74958808, Time: 0.0210 Steps: 82290, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001398, Sample Num: 22368, Cur Loss: 0.22521161, Cur Avg Loss: 0.20301542, Log Avg loss: 0.27596795, Global Avg Loss: 0.74953053, Time: 0.0210 Steps: 82300, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001408, Sample Num: 22528, Cur Loss: 0.11707292, Cur Avg Loss: 0.20297030, Log Avg loss: 0.19666342, Global Avg Loss: 0.74946336, Time: 0.0210 Steps: 82310, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001418, Sample Num: 22688, Cur Loss: 0.18547334, Cur Avg Loss: 0.20372800, Log Avg loss: 0.31041217, Global Avg Loss: 0.74941003, Time: 0.0210 Steps: 82320, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001428, Sample Num: 22848, Cur Loss: 0.39230829, Cur Avg Loss: 0.20381621, Log Avg loss: 0.21632394, Global Avg Loss: 0.74934528, Time: 0.0210 Steps: 82330, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001438, Sample Num: 23008, Cur Loss: 0.09883735, Cur Avg Loss: 0.20402500, Log Avg loss: 0.23383985, Global Avg Loss: 0.74928267, Time: 0.0210 Steps: 82340, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001448, Sample Num: 23168, Cur Loss: 0.25880507, Cur Avg Loss: 0.20386713, Log Avg loss: 0.18116591, Global Avg Loss: 0.74921368, Time: 0.0210 Steps: 82350, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001458, Sample Num: 23328, Cur Loss: 0.03128388, Cur Avg Loss: 0.20331703, Log Avg loss: 0.12366268, Global Avg Loss: 0.74913773, Time: 0.0210 Steps: 82360, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001468, Sample Num: 23488, Cur Loss: 0.51962811, Cur Avg Loss: 0.20424891, Log Avg loss: 0.34011720, Global Avg Loss: 0.74908807, Time: 0.0210 Steps: 82370, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001478, Sample Num: 23648, Cur Loss: 0.35046089, Cur Avg Loss: 0.20489075, Log Avg loss: 0.29911223, Global Avg Loss: 0.74903345, Time: 0.0210 Steps: 82380, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001488, Sample Num: 23808, Cur Loss: 0.18283539, Cur Avg Loss: 0.20502943, Log Avg loss: 0.22552587, Global Avg Loss: 0.74896991, Time: 0.0210 Steps: 82390, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001498, Sample Num: 23968, Cur Loss: 0.19728993, Cur Avg Loss: 0.20509434, Log Avg loss: 0.21475387, Global Avg Loss: 0.74890508, Time: 0.0210 Steps: 82400, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001508, Sample Num: 24128, Cur Loss: 0.10550971, Cur Avg Loss: 0.20470619, Log Avg loss: 0.14656159, Global Avg Loss: 0.74883199, Time: 0.0210 Steps: 82410, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001518, Sample Num: 24288, Cur Loss: 0.10464697, Cur Avg Loss: 0.20439615, Log Avg loss: 0.15764238, Global Avg Loss: 0.74876026, Time: 0.0210 Steps: 82420, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001528, Sample Num: 24448, Cur Loss: 0.66426390, Cur Avg Loss: 0.20455965, Log Avg loss: 0.22937774, Global Avg Loss: 0.74869725, Time: 0.0210 Steps: 82430, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001538, Sample Num: 24608, Cur Loss: 0.25846523, Cur Avg Loss: 0.20503708, Log Avg loss: 0.27798953, Global Avg Loss: 0.74864015, Time: 0.0248 Steps: 82440, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001548, Sample Num: 24768, Cur Loss: 0.35693103, Cur Avg Loss: 0.20512217, Log Avg loss: 0.21820760, Global Avg Loss: 0.74857582, Time: 0.0210 Steps: 82450, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001558, Sample Num: 24928, Cur Loss: 0.11578045, Cur Avg Loss: 0.20516158, Log Avg loss: 0.21126349, Global Avg Loss: 0.74851066, Time: 0.0210 Steps: 82460, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001568, Sample Num: 25088, Cur Loss: 0.18019255, Cur Avg Loss: 0.20529139, Log Avg loss: 0.22551557, Global Avg Loss: 0.74844724, Time: 0.0210 Steps: 82470, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001578, Sample Num: 25248, Cur Loss: 0.10185600, Cur Avg Loss: 0.20506710, Log Avg loss: 0.16989784, Global Avg Loss: 0.74837710, Time: 0.0210 Steps: 82480, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001588, Sample Num: 25408, Cur Loss: 0.22668926, Cur Avg Loss: 0.20493120, Log Avg loss: 0.18348600, Global Avg Loss: 0.74830862, Time: 0.0210 Steps: 82490, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001598, Sample Num: 25568, Cur Loss: 0.58596182, Cur Avg Loss: 0.20578817, Log Avg loss: 0.34187471, Global Avg Loss: 0.74825935, Time: 0.0210 Steps: 82500, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001608, Sample Num: 25728, Cur Loss: 0.10465147, Cur Avg Loss: 0.20580349, Log Avg loss: 0.20825262, Global Avg Loss: 0.74819391, Time: 0.0210 Steps: 82510, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001618, Sample Num: 25888, Cur Loss: 0.09602377, Cur Avg Loss: 0.20588010, Log Avg loss: 0.21819902, Global Avg Loss: 0.74812968, Time: 0.0210 Steps: 82520, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001628, Sample Num: 26048, Cur Loss: 0.15257901, Cur Avg Loss: 0.20543808, Log Avg loss: 0.13391844, Global Avg Loss: 0.74805526, Time: 0.0210 Steps: 82530, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001638, Sample Num: 26208, Cur Loss: 0.25625169, Cur Avg Loss: 0.20521361, Log Avg loss: 0.16866982, Global Avg Loss: 0.74798506, Time: 0.0211 Steps: 82540, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001648, Sample Num: 26368, Cur Loss: 0.14083797, Cur Avg Loss: 0.20518213, Log Avg loss: 0.20002645, Global Avg Loss: 0.74791868, Time: 0.0211 Steps: 82550, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001658, Sample Num: 26528, Cur Loss: 0.40994054, Cur Avg Loss: 0.20516136, Log Avg loss: 0.20173782, Global Avg Loss: 0.74785253, Time: 0.0208 Steps: 82560, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001668, Sample Num: 26688, Cur Loss: 0.41662779, Cur Avg Loss: 0.20537697, Log Avg loss: 0.24112531, Global Avg Loss: 0.74779116, Time: 0.0208 Steps: 82570, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001678, Sample Num: 26848, Cur Loss: 0.06957483, Cur Avg Loss: 0.20485257, Log Avg loss: 0.11738304, Global Avg Loss: 0.74771482, Time: 0.0208 Steps: 82580, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001688, Sample Num: 27008, Cur Loss: 0.09439795, Cur Avg Loss: 0.20487649, Log Avg loss: 0.20888936, Global Avg Loss: 0.74764958, Time: 0.0208 Steps: 82590, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001698, Sample Num: 27168, Cur Loss: 0.08983817, Cur Avg Loss: 0.20488573, Log Avg loss: 0.20644656, Global Avg Loss: 0.74758406, Time: 0.0208 Steps: 82600, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001708, Sample Num: 27328, Cur Loss: 0.25597596, Cur Avg Loss: 0.20467317, Log Avg loss: 0.16857992, Global Avg Loss: 0.74751397, Time: 0.0208 Steps: 82610, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001718, Sample Num: 27488, Cur Loss: 0.22351061, Cur Avg Loss: 0.20438576, Log Avg loss: 0.15529653, Global Avg Loss: 0.74744229, Time: 0.0208 Steps: 82620, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001728, Sample Num: 27648, Cur Loss: 0.12655842, Cur Avg Loss: 0.20449676, Log Avg loss: 0.22356573, Global Avg Loss: 0.74737889, Time: 0.0208 Steps: 82630, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001738, Sample Num: 27808, Cur Loss: 0.27797371, Cur Avg Loss: 0.20456393, Log Avg loss: 0.21617241, Global Avg Loss: 0.74731461, Time: 0.0208 Steps: 82640, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001748, Sample Num: 27968, Cur Loss: 0.13234417, Cur Avg Loss: 0.20444758, Log Avg loss: 0.18422490, Global Avg Loss: 0.74724648, Time: 0.0208 Steps: 82650, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001758, Sample Num: 28128, Cur Loss: 0.06562031, Cur Avg Loss: 0.20407570, Log Avg loss: 0.13907058, Global Avg Loss: 0.74717291, Time: 0.0208 Steps: 82660, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001768, Sample Num: 28288, Cur Loss: 0.15294217, Cur Avg Loss: 0.20419411, Log Avg loss: 0.22501108, Global Avg Loss: 0.74710974, Time: 0.0208 Steps: 82670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001778, Sample Num: 28448, Cur Loss: 0.10179424, Cur Avg Loss: 0.20436894, Log Avg loss: 0.23527983, Global Avg Loss: 0.74704784, Time: 0.0208 Steps: 82680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001788, Sample Num: 28608, Cur Loss: 0.12246299, Cur Avg Loss: 0.20404601, Log Avg loss: 0.14662846, Global Avg Loss: 0.74697523, Time: 0.0208 Steps: 82690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001798, Sample Num: 28768, Cur Loss: 0.16666263, Cur Avg Loss: 0.20401587, Log Avg loss: 0.19862631, Global Avg Loss: 0.74690892, Time: 0.0211 Steps: 82700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001808, Sample Num: 28928, Cur Loss: 0.10426439, Cur Avg Loss: 0.20383940, Log Avg loss: 0.17211053, Global Avg Loss: 0.74683943, Time: 0.0210 Steps: 82710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001818, Sample Num: 29088, Cur Loss: 0.27438009, Cur Avg Loss: 0.20413909, Log Avg loss: 0.25832351, Global Avg Loss: 0.74678037, Time: 0.0210 Steps: 82720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001828, Sample Num: 29248, Cur Loss: 0.49598694, Cur Avg Loss: 0.20425819, Log Avg loss: 0.22591007, Global Avg Loss: 0.74671741, Time: 0.0210 Steps: 82730, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001838, Sample Num: 29408, Cur Loss: 0.17007814, Cur Avg Loss: 0.20427036, Log Avg loss: 0.20649556, Global Avg Loss: 0.74665212, Time: 0.0209 Steps: 82740, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001848, Sample Num: 29568, Cur Loss: 0.12883268, Cur Avg Loss: 0.20427860, Log Avg loss: 0.20579330, Global Avg Loss: 0.74658676, Time: 0.0210 Steps: 82750, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001858, Sample Num: 29728, Cur Loss: 0.03484084, Cur Avg Loss: 0.20408013, Log Avg loss: 0.16740140, Global Avg Loss: 0.74651677, Time: 0.0209 Steps: 82760, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001868, Sample Num: 29888, Cur Loss: 0.23011145, Cur Avg Loss: 0.20404805, Log Avg loss: 0.19808884, Global Avg Loss: 0.74645051, Time: 0.0209 Steps: 82770, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001878, Sample Num: 30048, Cur Loss: 0.06066887, Cur Avg Loss: 0.20386821, Log Avg loss: 0.17027448, Global Avg Loss: 0.74638091, Time: 0.0210 Steps: 82780, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001888, Sample Num: 30208, Cur Loss: 0.08884159, Cur Avg Loss: 0.20360802, Log Avg loss: 0.15474399, Global Avg Loss: 0.74630945, Time: 0.0210 Steps: 82790, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001898, Sample Num: 30368, Cur Loss: 0.16733769, Cur Avg Loss: 0.20354375, Log Avg loss: 0.19140959, Global Avg Loss: 0.74624243, Time: 0.0210 Steps: 82800, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001908, Sample Num: 30528, Cur Loss: 0.27246794, Cur Avg Loss: 0.20392835, Log Avg loss: 0.27692550, Global Avg Loss: 0.74618576, Time: 0.0209 Steps: 82810, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001918, Sample Num: 30688, Cur Loss: 0.24846436, Cur Avg Loss: 0.20402496, Log Avg loss: 0.22245819, Global Avg Loss: 0.74612252, Time: 0.0209 Steps: 82820, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001928, Sample Num: 30848, Cur Loss: 0.08827855, Cur Avg Loss: 0.20374600, Log Avg loss: 0.15024109, Global Avg Loss: 0.74605058, Time: 0.0209 Steps: 82830, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001938, Sample Num: 31008, Cur Loss: 0.11410765, Cur Avg Loss: 0.20396788, Log Avg loss: 0.24674560, Global Avg Loss: 0.74599031, Time: 0.0209 Steps: 82840, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001948, Sample Num: 31168, Cur Loss: 0.21125764, Cur Avg Loss: 0.20423097, Log Avg loss: 0.25521764, Global Avg Loss: 0.74593107, Time: 0.0209 Steps: 82850, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001958, Sample Num: 31328, Cur Loss: 0.05496578, Cur Avg Loss: 0.20431501, Log Avg loss: 0.22068625, Global Avg Loss: 0.74586768, Time: 0.0209 Steps: 82860, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001968, Sample Num: 31488, Cur Loss: 0.12646072, Cur Avg Loss: 0.20407222, Log Avg loss: 0.15653357, Global Avg Loss: 0.74579657, Time: 0.0210 Steps: 82870, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001978, Sample Num: 31648, Cur Loss: 0.32137537, Cur Avg Loss: 0.20431031, Log Avg loss: 0.25116749, Global Avg Loss: 0.74573689, Time: 0.0209 Steps: 82880, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001988, Sample Num: 31808, Cur Loss: 0.19869858, Cur Avg Loss: 0.20402171, Log Avg loss: 0.14693698, Global Avg Loss: 0.74566465, Time: 0.0209 Steps: 82890, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001998, Sample Num: 31968, Cur Loss: 0.05486810, Cur Avg Loss: 0.20424836, Log Avg loss: 0.24930637, Global Avg Loss: 0.74560477, Time: 0.0210 Steps: 82900, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002008, Sample Num: 32128, Cur Loss: 0.20097509, Cur Avg Loss: 0.20447068, Log Avg loss: 0.24888903, Global Avg Loss: 0.74554486, Time: 0.0209 Steps: 82910, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002018, Sample Num: 32288, Cur Loss: 0.09317883, Cur Avg Loss: 0.20410080, Log Avg loss: 0.12983030, Global Avg Loss: 0.74547061, Time: 0.0209 Steps: 82920, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002028, Sample Num: 32448, Cur Loss: 0.23229027, Cur Avg Loss: 0.20431405, Log Avg loss: 0.24734635, Global Avg Loss: 0.74541054, Time: 0.0210 Steps: 82930, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002038, Sample Num: 32608, Cur Loss: 0.14712484, Cur Avg Loss: 0.20458543, Log Avg loss: 0.25962212, Global Avg Loss: 0.74535197, Time: 0.0209 Steps: 82940, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002048, Sample Num: 32768, Cur Loss: 0.26753691, Cur Avg Loss: 0.20481555, Log Avg loss: 0.25171443, Global Avg Loss: 0.74529246, Time: 0.0255 Steps: 82950, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002058, Sample Num: 32928, Cur Loss: 0.10331272, Cur Avg Loss: 0.20475882, Log Avg loss: 0.19314009, Global Avg Loss: 0.74522590, Time: 0.0209 Steps: 82960, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002068, Sample Num: 33088, Cur Loss: 0.45841637, Cur Avg Loss: 0.20466508, Log Avg loss: 0.18537354, Global Avg Loss: 0.74515843, Time: 0.0209 Steps: 82970, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002078, Sample Num: 33248, Cur Loss: 0.10756657, Cur Avg Loss: 0.20457341, Log Avg loss: 0.18561671, Global Avg Loss: 0.74509100, Time: 0.0208 Steps: 82980, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002088, Sample Num: 33408, Cur Loss: 0.17564830, Cur Avg Loss: 0.20457798, Log Avg loss: 0.20552737, Global Avg Loss: 0.74502598, Time: 0.0209 Steps: 82990, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002098, Sample Num: 33568, Cur Loss: 0.28617150, Cur Avg Loss: 0.20460877, Log Avg loss: 0.21103789, Global Avg Loss: 0.74496164, Time: 0.0208 Steps: 83000, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002108, Sample Num: 33728, Cur Loss: 0.11829669, Cur Avg Loss: 0.20435907, Log Avg loss: 0.15197067, Global Avg Loss: 0.74489021, Time: 0.0209 Steps: 83010, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002118, Sample Num: 33888, Cur Loss: 0.17835927, Cur Avg Loss: 0.20470113, Log Avg loss: 0.27680726, Global Avg Loss: 0.74483383, Time: 0.0209 Steps: 83020, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002128, Sample Num: 34048, Cur Loss: 0.16096795, Cur Avg Loss: 0.20420376, Log Avg loss: 0.09886150, Global Avg Loss: 0.74475603, Time: 0.0209 Steps: 83030, Updated lr: 0.000022 ***** Running evaluation checkpoint-83031 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-83031 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.769092, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.34956, "eval_total_loss": 245.74077, "eval_mae": 0.487164, "eval_mse": 0.34955, "eval_r2": 0.777803, "eval_sp_statistic": 0.89683, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.923884, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.303072, "test_total_loss": 152.142276, "test_mae": 0.35245, "test_mse": 0.30318, "test_r2": 0.804325, "test_sp_statistic": 0.87453, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921475, "test_ps_pvalue": 0.0, "lr": 2.2208629682313894e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7447487656606638, "train_cur_epoch_loss": 434.68748506437987, "train_cur_epoch_avg_loss": 0.2041744880527853, "train_cur_epoch_time": 44.769091844558716, "train_cur_epoch_avg_time": 0.021028225384950076, "epoch": 39, "step": 83031} ################################################## Training, Epoch: 0040, Batch: 000009, Sample Num: 144, Cur Loss: 0.14972967, Cur Avg Loss: 0.21816257, Log Avg loss: 0.21053463, Global Avg Loss: 0.74469169, Time: 0.0210 Steps: 83040, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000019, Sample Num: 304, Cur Loss: 0.31829548, Cur Avg Loss: 0.24039279, Log Avg loss: 0.26039998, Global Avg Loss: 0.74463338, Time: 0.0210 Steps: 83050, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000029, Sample Num: 464, Cur Loss: 0.08331538, Cur Avg Loss: 0.21061036, Log Avg loss: 0.15402374, Global Avg Loss: 0.74456227, Time: 0.0209 Steps: 83060, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000039, Sample Num: 624, Cur Loss: 0.28839648, Cur Avg Loss: 0.19211707, Log Avg loss: 0.13848652, Global Avg Loss: 0.74448931, Time: 0.0209 Steps: 83070, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000049, Sample Num: 784, Cur Loss: 0.13395318, Cur Avg Loss: 0.17841762, Log Avg loss: 0.12498978, Global Avg Loss: 0.74441475, Time: 0.0209 Steps: 83080, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000059, Sample Num: 944, Cur Loss: 0.14790176, Cur Avg Loss: 0.19328358, Log Avg loss: 0.26612680, Global Avg Loss: 0.74435718, Time: 0.0209 Steps: 83090, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000069, Sample Num: 1104, Cur Loss: 0.17954461, Cur Avg Loss: 0.19140425, Log Avg loss: 0.18031621, Global Avg Loss: 0.74428931, Time: 0.0209 Steps: 83100, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000079, Sample Num: 1264, Cur Loss: 0.02892677, Cur Avg Loss: 0.19057214, Log Avg loss: 0.18483052, Global Avg Loss: 0.74422199, Time: 0.0210 Steps: 83110, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000089, Sample Num: 1424, Cur Loss: 0.23966077, Cur Avg Loss: 0.19218313, Log Avg loss: 0.20491001, Global Avg Loss: 0.74415711, Time: 0.0210 Steps: 83120, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000099, Sample Num: 1584, Cur Loss: 0.11134622, Cur Avg Loss: 0.19068271, Log Avg loss: 0.17732897, Global Avg Loss: 0.74408893, Time: 0.0209 Steps: 83130, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000109, Sample Num: 1744, Cur Loss: 0.45226347, Cur Avg Loss: 0.19788289, Log Avg loss: 0.26916462, Global Avg Loss: 0.74403180, Time: 0.0210 Steps: 83140, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000119, Sample Num: 1904, Cur Loss: 0.12102234, Cur Avg Loss: 0.20813973, Log Avg loss: 0.31993928, Global Avg Loss: 0.74398080, Time: 0.0208 Steps: 83150, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000129, Sample Num: 2064, Cur Loss: 0.14117077, Cur Avg Loss: 0.21140079, Log Avg loss: 0.25020745, Global Avg Loss: 0.74392142, Time: 0.0210 Steps: 83160, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000139, Sample Num: 2224, Cur Loss: 0.13896193, Cur Avg Loss: 0.21005304, Log Avg loss: 0.19266712, Global Avg Loss: 0.74385514, Time: 0.0209 Steps: 83170, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000149, Sample Num: 2384, Cur Loss: 0.11338928, Cur Avg Loss: 0.20646680, Log Avg loss: 0.15661806, Global Avg Loss: 0.74378454, Time: 0.0210 Steps: 83180, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000159, Sample Num: 2544, Cur Loss: 0.12174489, Cur Avg Loss: 0.20362578, Log Avg loss: 0.16129454, Global Avg Loss: 0.74371452, Time: 0.0209 Steps: 83190, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000169, Sample Num: 2704, Cur Loss: 0.08875404, Cur Avg Loss: 0.19720080, Log Avg loss: 0.09504356, Global Avg Loss: 0.74363656, Time: 0.0209 Steps: 83200, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000179, Sample Num: 2864, Cur Loss: 0.21625818, Cur Avg Loss: 0.19864473, Log Avg loss: 0.22304723, Global Avg Loss: 0.74357400, Time: 0.0209 Steps: 83210, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000189, Sample Num: 3024, Cur Loss: 0.16891599, Cur Avg Loss: 0.19609353, Log Avg loss: 0.15042692, Global Avg Loss: 0.74350272, Time: 0.0209 Steps: 83220, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000199, Sample Num: 3184, Cur Loss: 0.15887889, Cur Avg Loss: 0.19713983, Log Avg loss: 0.21691502, Global Avg Loss: 0.74343945, Time: 0.0209 Steps: 83230, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000209, Sample Num: 3344, Cur Loss: 0.08468647, Cur Avg Loss: 0.19665697, Log Avg loss: 0.18704804, Global Avg Loss: 0.74337261, Time: 0.0209 Steps: 83240, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000219, Sample Num: 3504, Cur Loss: 0.29302713, Cur Avg Loss: 0.19777209, Log Avg loss: 0.22107812, Global Avg Loss: 0.74330987, Time: 0.0209 Steps: 83250, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000229, Sample Num: 3664, Cur Loss: 0.12327543, Cur Avg Loss: 0.19894522, Log Avg loss: 0.22463672, Global Avg Loss: 0.74324758, Time: 0.0209 Steps: 83260, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000239, Sample Num: 3824, Cur Loss: 0.12418254, Cur Avg Loss: 0.19765183, Log Avg loss: 0.16803328, Global Avg Loss: 0.74317850, Time: 0.0210 Steps: 83270, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000249, Sample Num: 3984, Cur Loss: 0.11163902, Cur Avg Loss: 0.19628641, Log Avg loss: 0.16365278, Global Avg Loss: 0.74310891, Time: 0.0210 Steps: 83280, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000259, Sample Num: 4144, Cur Loss: 0.04047327, Cur Avg Loss: 0.19425276, Log Avg loss: 0.14361489, Global Avg Loss: 0.74303693, Time: 0.0245 Steps: 83290, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000269, Sample Num: 4304, Cur Loss: 0.11228730, Cur Avg Loss: 0.19454282, Log Avg loss: 0.20205551, Global Avg Loss: 0.74297199, Time: 0.0210 Steps: 83300, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000279, Sample Num: 4464, Cur Loss: 0.08469316, Cur Avg Loss: 0.19346699, Log Avg loss: 0.16452702, Global Avg Loss: 0.74290256, Time: 0.0209 Steps: 83310, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000289, Sample Num: 4624, Cur Loss: 0.09340123, Cur Avg Loss: 0.19360202, Log Avg loss: 0.19736931, Global Avg Loss: 0.74283708, Time: 0.0208 Steps: 83320, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000299, Sample Num: 4784, Cur Loss: 0.20508778, Cur Avg Loss: 0.19528319, Log Avg loss: 0.24386906, Global Avg Loss: 0.74277720, Time: 0.0208 Steps: 83330, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000309, Sample Num: 4944, Cur Loss: 0.20949891, Cur Avg Loss: 0.19476043, Log Avg loss: 0.17912990, Global Avg Loss: 0.74270957, Time: 0.0208 Steps: 83340, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000319, Sample Num: 5104, Cur Loss: 0.03477218, Cur Avg Loss: 0.19574892, Log Avg loss: 0.22629317, Global Avg Loss: 0.74264761, Time: 0.0208 Steps: 83350, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000329, Sample Num: 5264, Cur Loss: 0.09237684, Cur Avg Loss: 0.19564620, Log Avg loss: 0.19236970, Global Avg Loss: 0.74258160, Time: 0.0209 Steps: 83360, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000339, Sample Num: 5424, Cur Loss: 0.30045021, Cur Avg Loss: 0.19740759, Log Avg loss: 0.25535724, Global Avg Loss: 0.74252316, Time: 0.0208 Steps: 83370, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000349, Sample Num: 5584, Cur Loss: 0.31213325, Cur Avg Loss: 0.19658782, Log Avg loss: 0.16879749, Global Avg Loss: 0.74245435, Time: 0.0208 Steps: 83380, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000359, Sample Num: 5744, Cur Loss: 0.13523382, Cur Avg Loss: 0.19542491, Log Avg loss: 0.15483955, Global Avg Loss: 0.74238389, Time: 0.0208 Steps: 83390, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000369, Sample Num: 5904, Cur Loss: 0.15831405, Cur Avg Loss: 0.19594460, Log Avg loss: 0.21460130, Global Avg Loss: 0.74232060, Time: 0.0209 Steps: 83400, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000379, Sample Num: 6064, Cur Loss: 0.30934477, Cur Avg Loss: 0.19707517, Log Avg loss: 0.23879307, Global Avg Loss: 0.74226024, Time: 0.0208 Steps: 83410, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000389, Sample Num: 6224, Cur Loss: 0.05155141, Cur Avg Loss: 0.19651107, Log Avg loss: 0.17513197, Global Avg Loss: 0.74219225, Time: 0.0208 Steps: 83420, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000399, Sample Num: 6384, Cur Loss: 0.23100272, Cur Avg Loss: 0.19656457, Log Avg loss: 0.19864541, Global Avg Loss: 0.74212710, Time: 0.0209 Steps: 83430, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000409, Sample Num: 6544, Cur Loss: 0.19551678, Cur Avg Loss: 0.19575313, Log Avg loss: 0.16337702, Global Avg Loss: 0.74205774, Time: 0.0209 Steps: 83440, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000419, Sample Num: 6704, Cur Loss: 0.14387897, Cur Avg Loss: 0.19449863, Log Avg loss: 0.14318953, Global Avg Loss: 0.74198598, Time: 0.0209 Steps: 83450, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000429, Sample Num: 6864, Cur Loss: 0.14267437, Cur Avg Loss: 0.19362641, Log Avg loss: 0.15708030, Global Avg Loss: 0.74191589, Time: 0.0208 Steps: 83460, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000439, Sample Num: 7024, Cur Loss: 0.40630853, Cur Avg Loss: 0.19289624, Log Avg loss: 0.16157205, Global Avg Loss: 0.74184637, Time: 0.0208 Steps: 83470, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000449, Sample Num: 7184, Cur Loss: 0.25551403, Cur Avg Loss: 0.19226763, Log Avg loss: 0.16467153, Global Avg Loss: 0.74177723, Time: 0.0208 Steps: 83480, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000459, Sample Num: 7344, Cur Loss: 0.69154435, Cur Avg Loss: 0.19319436, Log Avg loss: 0.23480469, Global Avg Loss: 0.74171650, Time: 0.0208 Steps: 83490, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000469, Sample Num: 7504, Cur Loss: 0.08153857, Cur Avg Loss: 0.19341807, Log Avg loss: 0.20368606, Global Avg Loss: 0.74165207, Time: 0.0207 Steps: 83500, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000479, Sample Num: 7664, Cur Loss: 0.42052904, Cur Avg Loss: 0.19464256, Log Avg loss: 0.25207123, Global Avg Loss: 0.74159344, Time: 0.0207 Steps: 83510, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000489, Sample Num: 7824, Cur Loss: 0.18278161, Cur Avg Loss: 0.19534027, Log Avg loss: 0.22876065, Global Avg Loss: 0.74153204, Time: 0.0208 Steps: 83520, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000499, Sample Num: 7984, Cur Loss: 0.40660778, Cur Avg Loss: 0.19621611, Log Avg loss: 0.23904456, Global Avg Loss: 0.74147189, Time: 0.0210 Steps: 83530, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000509, Sample Num: 8144, Cur Loss: 0.07139997, Cur Avg Loss: 0.19578415, Log Avg loss: 0.17422951, Global Avg Loss: 0.74140398, Time: 0.0208 Steps: 83540, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000519, Sample Num: 8304, Cur Loss: 0.14992136, Cur Avg Loss: 0.19600404, Log Avg loss: 0.20719634, Global Avg Loss: 0.74134005, Time: 0.0222 Steps: 83550, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000529, Sample Num: 8464, Cur Loss: 0.06139502, Cur Avg Loss: 0.19675090, Log Avg loss: 0.23551279, Global Avg Loss: 0.74127951, Time: 0.0208 Steps: 83560, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000539, Sample Num: 8624, Cur Loss: 0.04586773, Cur Avg Loss: 0.19679623, Log Avg loss: 0.19919428, Global Avg Loss: 0.74121465, Time: 0.0208 Steps: 83570, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000549, Sample Num: 8784, Cur Loss: 0.05883204, Cur Avg Loss: 0.19810898, Log Avg loss: 0.26886622, Global Avg Loss: 0.74115813, Time: 0.0208 Steps: 83580, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000559, Sample Num: 8944, Cur Loss: 0.09850365, Cur Avg Loss: 0.19742854, Log Avg loss: 0.16007219, Global Avg Loss: 0.74108861, Time: 0.0208 Steps: 83590, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000569, Sample Num: 9104, Cur Loss: 0.19079861, Cur Avg Loss: 0.19735243, Log Avg loss: 0.19309828, Global Avg Loss: 0.74102307, Time: 0.0208 Steps: 83600, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000579, Sample Num: 9264, Cur Loss: 0.18520534, Cur Avg Loss: 0.19711005, Log Avg loss: 0.18331877, Global Avg Loss: 0.74095636, Time: 0.0208 Steps: 83610, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000589, Sample Num: 9424, Cur Loss: 0.09429629, Cur Avg Loss: 0.19676962, Log Avg loss: 0.17705820, Global Avg Loss: 0.74088893, Time: 0.0207 Steps: 83620, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000599, Sample Num: 9584, Cur Loss: 0.47064158, Cur Avg Loss: 0.19727840, Log Avg loss: 0.22724580, Global Avg Loss: 0.74082751, Time: 0.0207 Steps: 83630, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000609, Sample Num: 9744, Cur Loss: 0.13592269, Cur Avg Loss: 0.19621659, Log Avg loss: 0.13261420, Global Avg Loss: 0.74075479, Time: 0.0208 Steps: 83640, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000619, Sample Num: 9904, Cur Loss: 0.28886145, Cur Avg Loss: 0.19549207, Log Avg loss: 0.15136851, Global Avg Loss: 0.74068433, Time: 0.0208 Steps: 83650, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000629, Sample Num: 10064, Cur Loss: 0.19449948, Cur Avg Loss: 0.19608337, Log Avg loss: 0.23268532, Global Avg Loss: 0.74062361, Time: 0.0208 Steps: 83660, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000639, Sample Num: 10224, Cur Loss: 0.03211410, Cur Avg Loss: 0.19555366, Log Avg loss: 0.16223494, Global Avg Loss: 0.74055448, Time: 0.0208 Steps: 83670, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000649, Sample Num: 10384, Cur Loss: 0.20954403, Cur Avg Loss: 0.19561135, Log Avg loss: 0.19929748, Global Avg Loss: 0.74048980, Time: 0.0207 Steps: 83680, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000659, Sample Num: 10544, Cur Loss: 0.07587887, Cur Avg Loss: 0.19590455, Log Avg loss: 0.21493305, Global Avg Loss: 0.74042700, Time: 0.0207 Steps: 83690, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000669, Sample Num: 10704, Cur Loss: 0.13915023, Cur Avg Loss: 0.19713175, Log Avg loss: 0.27800433, Global Avg Loss: 0.74037176, Time: 0.0208 Steps: 83700, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000679, Sample Num: 10864, Cur Loss: 0.18080294, Cur Avg Loss: 0.19628513, Log Avg loss: 0.13964609, Global Avg Loss: 0.74029999, Time: 0.0208 Steps: 83710, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000689, Sample Num: 11024, Cur Loss: 0.07786314, Cur Avg Loss: 0.19596045, Log Avg loss: 0.17391476, Global Avg Loss: 0.74023234, Time: 0.0208 Steps: 83720, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000699, Sample Num: 11184, Cur Loss: 0.11243373, Cur Avg Loss: 0.19539730, Log Avg loss: 0.15659661, Global Avg Loss: 0.74016264, Time: 0.0207 Steps: 83730, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000709, Sample Num: 11344, Cur Loss: 0.09522841, Cur Avg Loss: 0.19485500, Log Avg loss: 0.15694769, Global Avg Loss: 0.74009299, Time: 0.0208 Steps: 83740, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000719, Sample Num: 11504, Cur Loss: 0.48474270, Cur Avg Loss: 0.19676049, Log Avg loss: 0.33185994, Global Avg Loss: 0.74004425, Time: 0.0207 Steps: 83750, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000729, Sample Num: 11664, Cur Loss: 0.21713422, Cur Avg Loss: 0.19673939, Log Avg loss: 0.19522233, Global Avg Loss: 0.73997920, Time: 0.0208 Steps: 83760, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000739, Sample Num: 11824, Cur Loss: 0.17578927, Cur Avg Loss: 0.19678931, Log Avg loss: 0.20042858, Global Avg Loss: 0.73991479, Time: 0.0208 Steps: 83770, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000749, Sample Num: 11984, Cur Loss: 0.10107179, Cur Avg Loss: 0.19630418, Log Avg loss: 0.16045312, Global Avg Loss: 0.73984563, Time: 0.0208 Steps: 83780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000759, Sample Num: 12144, Cur Loss: 0.27224448, Cur Avg Loss: 0.19643221, Log Avg loss: 0.20602162, Global Avg Loss: 0.73978192, Time: 0.0208 Steps: 83790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000769, Sample Num: 12304, Cur Loss: 0.20481741, Cur Avg Loss: 0.19686587, Log Avg loss: 0.22978031, Global Avg Loss: 0.73972106, Time: 0.0245 Steps: 83800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000779, Sample Num: 12464, Cur Loss: 0.63981068, Cur Avg Loss: 0.19733196, Log Avg loss: 0.23317442, Global Avg Loss: 0.73966062, Time: 0.0208 Steps: 83810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000789, Sample Num: 12624, Cur Loss: 0.27798921, Cur Avg Loss: 0.19741793, Log Avg loss: 0.20411502, Global Avg Loss: 0.73959673, Time: 0.0208 Steps: 83820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000799, Sample Num: 12784, Cur Loss: 0.38955233, Cur Avg Loss: 0.19816299, Log Avg loss: 0.25694830, Global Avg Loss: 0.73953915, Time: 0.0208 Steps: 83830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000809, Sample Num: 12944, Cur Loss: 0.10925429, Cur Avg Loss: 0.19811283, Log Avg loss: 0.19410481, Global Avg Loss: 0.73947409, Time: 0.0208 Steps: 83840, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000819, Sample Num: 13104, Cur Loss: 0.44535023, Cur Avg Loss: 0.19859310, Log Avg loss: 0.23744733, Global Avg Loss: 0.73941422, Time: 0.0208 Steps: 83850, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000829, Sample Num: 13264, Cur Loss: 0.10466906, Cur Avg Loss: 0.19848006, Log Avg loss: 0.18922218, Global Avg Loss: 0.73934861, Time: 0.0208 Steps: 83860, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000839, Sample Num: 13424, Cur Loss: 0.22746910, Cur Avg Loss: 0.19925261, Log Avg loss: 0.26329686, Global Avg Loss: 0.73929185, Time: 0.0208 Steps: 83870, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000849, Sample Num: 13584, Cur Loss: 0.09461148, Cur Avg Loss: 0.19944943, Log Avg loss: 0.21596294, Global Avg Loss: 0.73922946, Time: 0.0208 Steps: 83880, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000859, Sample Num: 13744, Cur Loss: 0.16567309, Cur Avg Loss: 0.19931157, Log Avg loss: 0.18760664, Global Avg Loss: 0.73916371, Time: 0.0208 Steps: 83890, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000869, Sample Num: 13904, Cur Loss: 0.21629740, Cur Avg Loss: 0.19893280, Log Avg loss: 0.16639695, Global Avg Loss: 0.73909544, Time: 0.0208 Steps: 83900, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000879, Sample Num: 14064, Cur Loss: 0.17180502, Cur Avg Loss: 0.19920382, Log Avg loss: 0.22275490, Global Avg Loss: 0.73903390, Time: 0.0208 Steps: 83910, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000889, Sample Num: 14224, Cur Loss: 0.22854090, Cur Avg Loss: 0.19890231, Log Avg loss: 0.17240017, Global Avg Loss: 0.73896638, Time: 0.0209 Steps: 83920, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000899, Sample Num: 14384, Cur Loss: 0.31107169, Cur Avg Loss: 0.19869703, Log Avg loss: 0.18044709, Global Avg Loss: 0.73889984, Time: 0.0208 Steps: 83930, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000909, Sample Num: 14544, Cur Loss: 0.16670257, Cur Avg Loss: 0.19839974, Log Avg loss: 0.17167405, Global Avg Loss: 0.73883226, Time: 0.0208 Steps: 83940, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000919, Sample Num: 14704, Cur Loss: 0.25540391, Cur Avg Loss: 0.19942572, Log Avg loss: 0.29268682, Global Avg Loss: 0.73877912, Time: 0.0208 Steps: 83950, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000929, Sample Num: 14864, Cur Loss: 0.44962817, Cur Avg Loss: 0.19978834, Log Avg loss: 0.23311347, Global Avg Loss: 0.73871889, Time: 0.0208 Steps: 83960, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000939, Sample Num: 15024, Cur Loss: 0.08196744, Cur Avg Loss: 0.19947946, Log Avg loss: 0.17078383, Global Avg Loss: 0.73865126, Time: 0.0208 Steps: 83970, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000949, Sample Num: 15184, Cur Loss: 0.16191651, Cur Avg Loss: 0.19917683, Log Avg loss: 0.17076049, Global Avg Loss: 0.73858363, Time: 0.0208 Steps: 83980, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000959, Sample Num: 15344, Cur Loss: 0.03648711, Cur Avg Loss: 0.19888283, Log Avg loss: 0.17098164, Global Avg Loss: 0.73851605, Time: 0.0208 Steps: 83990, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000969, Sample Num: 15504, Cur Loss: 0.18233821, Cur Avg Loss: 0.19858625, Log Avg loss: 0.17014422, Global Avg Loss: 0.73844839, Time: 0.0208 Steps: 84000, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000979, Sample Num: 15664, Cur Loss: 0.16073152, Cur Avg Loss: 0.19870287, Log Avg loss: 0.21000379, Global Avg Loss: 0.73838549, Time: 0.0208 Steps: 84010, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000989, Sample Num: 15824, Cur Loss: 0.52323407, Cur Avg Loss: 0.19893029, Log Avg loss: 0.22119426, Global Avg Loss: 0.73832393, Time: 0.0208 Steps: 84020, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000999, Sample Num: 15984, Cur Loss: 0.10932575, Cur Avg Loss: 0.19879394, Log Avg loss: 0.18530985, Global Avg Loss: 0.73825812, Time: 0.0208 Steps: 84030, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001009, Sample Num: 16144, Cur Loss: 0.16411318, Cur Avg Loss: 0.19876051, Log Avg loss: 0.19542086, Global Avg Loss: 0.73819353, Time: 0.0208 Steps: 84040, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001019, Sample Num: 16304, Cur Loss: 0.32727295, Cur Avg Loss: 0.19957273, Log Avg loss: 0.28152473, Global Avg Loss: 0.73813920, Time: 0.0208 Steps: 84050, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001029, Sample Num: 16464, Cur Loss: 0.31011325, Cur Avg Loss: 0.20060320, Log Avg loss: 0.30560895, Global Avg Loss: 0.73808774, Time: 0.0245 Steps: 84060, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001039, Sample Num: 16624, Cur Loss: 0.29322195, Cur Avg Loss: 0.20093406, Log Avg loss: 0.23497891, Global Avg Loss: 0.73802790, Time: 0.0208 Steps: 84070, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001049, Sample Num: 16784, Cur Loss: 0.12499628, Cur Avg Loss: 0.20068718, Log Avg loss: 0.17503677, Global Avg Loss: 0.73796094, Time: 0.0208 Steps: 84080, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001059, Sample Num: 16944, Cur Loss: 0.05308143, Cur Avg Loss: 0.20070679, Log Avg loss: 0.20276379, Global Avg Loss: 0.73789729, Time: 0.0208 Steps: 84090, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001069, Sample Num: 17104, Cur Loss: 0.28280166, Cur Avg Loss: 0.20001037, Log Avg loss: 0.12625897, Global Avg Loss: 0.73782456, Time: 0.0208 Steps: 84100, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001079, Sample Num: 17264, Cur Loss: 0.13322076, Cur Avg Loss: 0.20067348, Log Avg loss: 0.27155993, Global Avg Loss: 0.73776913, Time: 0.0208 Steps: 84110, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001089, Sample Num: 17424, Cur Loss: 0.14897078, Cur Avg Loss: 0.20135964, Log Avg loss: 0.27539666, Global Avg Loss: 0.73771416, Time: 0.0208 Steps: 84120, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001099, Sample Num: 17584, Cur Loss: 0.06322621, Cur Avg Loss: 0.20181589, Log Avg loss: 0.25150166, Global Avg Loss: 0.73765637, Time: 0.0208 Steps: 84130, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001109, Sample Num: 17744, Cur Loss: 0.14280038, Cur Avg Loss: 0.20163513, Log Avg loss: 0.18177010, Global Avg Loss: 0.73759030, Time: 0.0208 Steps: 84140, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001119, Sample Num: 17904, Cur Loss: 0.16771415, Cur Avg Loss: 0.20164582, Log Avg loss: 0.20283129, Global Avg Loss: 0.73752676, Time: 0.0208 Steps: 84150, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001129, Sample Num: 18064, Cur Loss: 0.28202534, Cur Avg Loss: 0.20202542, Log Avg loss: 0.24450239, Global Avg Loss: 0.73746817, Time: 0.0207 Steps: 84160, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001139, Sample Num: 18224, Cur Loss: 0.17102668, Cur Avg Loss: 0.20205678, Log Avg loss: 0.20559734, Global Avg Loss: 0.73740498, Time: 0.0208 Steps: 84170, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001149, Sample Num: 18384, Cur Loss: 0.03708727, Cur Avg Loss: 0.20207934, Log Avg loss: 0.20464842, Global Avg Loss: 0.73734170, Time: 0.0208 Steps: 84180, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001159, Sample Num: 18544, Cur Loss: 0.25716960, Cur Avg Loss: 0.20180095, Log Avg loss: 0.16981401, Global Avg Loss: 0.73727429, Time: 0.0208 Steps: 84190, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001169, Sample Num: 18704, Cur Loss: 0.35600874, Cur Avg Loss: 0.20289617, Log Avg loss: 0.32983239, Global Avg Loss: 0.73722590, Time: 0.0208 Steps: 84200, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001179, Sample Num: 18864, Cur Loss: 0.28607288, Cur Avg Loss: 0.20280350, Log Avg loss: 0.19197019, Global Avg Loss: 0.73716115, Time: 0.0208 Steps: 84210, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001189, Sample Num: 19024, Cur Loss: 0.22637120, Cur Avg Loss: 0.20387170, Log Avg loss: 0.32981290, Global Avg Loss: 0.73711278, Time: 0.0208 Steps: 84220, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001199, Sample Num: 19184, Cur Loss: 0.11827980, Cur Avg Loss: 0.20350364, Log Avg loss: 0.15974108, Global Avg Loss: 0.73704423, Time: 0.0208 Steps: 84230, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001209, Sample Num: 19344, Cur Loss: 0.03056636, Cur Avg Loss: 0.20305175, Log Avg loss: 0.14887050, Global Avg Loss: 0.73697441, Time: 0.0208 Steps: 84240, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001219, Sample Num: 19504, Cur Loss: 0.42748672, Cur Avg Loss: 0.20317322, Log Avg loss: 0.21785815, Global Avg Loss: 0.73691279, Time: 0.0208 Steps: 84250, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001229, Sample Num: 19664, Cur Loss: 0.19129448, Cur Avg Loss: 0.20344116, Log Avg loss: 0.23610293, Global Avg Loss: 0.73685336, Time: 0.0209 Steps: 84260, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001239, Sample Num: 19824, Cur Loss: 0.20036674, Cur Avg Loss: 0.20347683, Log Avg loss: 0.20786143, Global Avg Loss: 0.73679058, Time: 0.0208 Steps: 84270, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001249, Sample Num: 19984, Cur Loss: 0.04038729, Cur Avg Loss: 0.20359388, Log Avg loss: 0.21809634, Global Avg Loss: 0.73672904, Time: 0.0208 Steps: 84280, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001259, Sample Num: 20144, Cur Loss: 0.13384664, Cur Avg Loss: 0.20347891, Log Avg loss: 0.18911899, Global Avg Loss: 0.73666407, Time: 0.0208 Steps: 84290, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001269, Sample Num: 20304, Cur Loss: 0.12601967, Cur Avg Loss: 0.20315908, Log Avg loss: 0.16289238, Global Avg Loss: 0.73659601, Time: 0.0208 Steps: 84300, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001279, Sample Num: 20464, Cur Loss: 0.06446508, Cur Avg Loss: 0.20312291, Log Avg loss: 0.19853280, Global Avg Loss: 0.73653219, Time: 0.0208 Steps: 84310, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001289, Sample Num: 20624, Cur Loss: 0.07606183, Cur Avg Loss: 0.20314615, Log Avg loss: 0.20611883, Global Avg Loss: 0.73646929, Time: 0.0208 Steps: 84320, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001299, Sample Num: 20784, Cur Loss: 0.10287795, Cur Avg Loss: 0.20318381, Log Avg loss: 0.20803796, Global Avg Loss: 0.73640662, Time: 0.0209 Steps: 84330, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001309, Sample Num: 20944, Cur Loss: 0.23108914, Cur Avg Loss: 0.20297308, Log Avg loss: 0.17559938, Global Avg Loss: 0.73634013, Time: 0.0209 Steps: 84340, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001319, Sample Num: 21104, Cur Loss: 0.17382607, Cur Avg Loss: 0.20272163, Log Avg loss: 0.16980739, Global Avg Loss: 0.73627296, Time: 0.0209 Steps: 84350, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001329, Sample Num: 21264, Cur Loss: 0.57575083, Cur Avg Loss: 0.20258902, Log Avg loss: 0.18509707, Global Avg Loss: 0.73620763, Time: 0.0208 Steps: 84360, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001339, Sample Num: 21424, Cur Loss: 0.11686850, Cur Avg Loss: 0.20202677, Log Avg loss: 0.12730328, Global Avg Loss: 0.73613546, Time: 0.0208 Steps: 84370, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001349, Sample Num: 21584, Cur Loss: 0.16623545, Cur Avg Loss: 0.20184536, Log Avg loss: 0.17755485, Global Avg Loss: 0.73606926, Time: 0.0209 Steps: 84380, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001359, Sample Num: 21744, Cur Loss: 0.16257909, Cur Avg Loss: 0.20173365, Log Avg loss: 0.18666461, Global Avg Loss: 0.73600416, Time: 0.0209 Steps: 84390, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001369, Sample Num: 21904, Cur Loss: 0.29183060, Cur Avg Loss: 0.20164962, Log Avg loss: 0.19022942, Global Avg Loss: 0.73593949, Time: 0.0208 Steps: 84400, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001379, Sample Num: 22064, Cur Loss: 0.09974815, Cur Avg Loss: 0.20165944, Log Avg loss: 0.20300440, Global Avg Loss: 0.73587636, Time: 0.0209 Steps: 84410, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001389, Sample Num: 22224, Cur Loss: 0.46083593, Cur Avg Loss: 0.20201675, Log Avg loss: 0.25128953, Global Avg Loss: 0.73581895, Time: 0.0209 Steps: 84420, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001399, Sample Num: 22384, Cur Loss: 0.17459257, Cur Avg Loss: 0.20295769, Log Avg loss: 0.33365389, Global Avg Loss: 0.73577132, Time: 0.0209 Steps: 84430, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001409, Sample Num: 22544, Cur Loss: 0.05978035, Cur Avg Loss: 0.20250163, Log Avg loss: 0.13869863, Global Avg Loss: 0.73570061, Time: 0.0208 Steps: 84440, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001419, Sample Num: 22704, Cur Loss: 0.24623421, Cur Avg Loss: 0.20287289, Log Avg loss: 0.25518367, Global Avg Loss: 0.73564371, Time: 0.0208 Steps: 84450, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001429, Sample Num: 22864, Cur Loss: 0.20069425, Cur Avg Loss: 0.20269994, Log Avg loss: 0.17815912, Global Avg Loss: 0.73557771, Time: 0.0209 Steps: 84460, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001439, Sample Num: 23024, Cur Loss: 0.19848633, Cur Avg Loss: 0.20286506, Log Avg loss: 0.22646054, Global Avg Loss: 0.73551743, Time: 0.0208 Steps: 84470, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001449, Sample Num: 23184, Cur Loss: 0.08592981, Cur Avg Loss: 0.20247070, Log Avg loss: 0.14572223, Global Avg Loss: 0.73544762, Time: 0.0209 Steps: 84480, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001459, Sample Num: 23344, Cur Loss: 0.08686247, Cur Avg Loss: 0.20243854, Log Avg loss: 0.19777833, Global Avg Loss: 0.73538398, Time: 0.0208 Steps: 84490, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001469, Sample Num: 23504, Cur Loss: 0.47516918, Cur Avg Loss: 0.20240653, Log Avg loss: 0.19773686, Global Avg Loss: 0.73532035, Time: 0.0208 Steps: 84500, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001479, Sample Num: 23664, Cur Loss: 0.13766579, Cur Avg Loss: 0.20253161, Log Avg loss: 0.22090486, Global Avg Loss: 0.73525948, Time: 0.0208 Steps: 84510, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001489, Sample Num: 23824, Cur Loss: 0.21386418, Cur Avg Loss: 0.20266472, Log Avg loss: 0.22235249, Global Avg Loss: 0.73519880, Time: 0.0209 Steps: 84520, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001499, Sample Num: 23984, Cur Loss: 0.45211178, Cur Avg Loss: 0.20316084, Log Avg loss: 0.27703221, Global Avg Loss: 0.73514460, Time: 0.0208 Steps: 84530, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001509, Sample Num: 24144, Cur Loss: 0.07127356, Cur Avg Loss: 0.20304366, Log Avg loss: 0.18547819, Global Avg Loss: 0.73507958, Time: 0.0210 Steps: 84540, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001519, Sample Num: 24304, Cur Loss: 0.28497267, Cur Avg Loss: 0.20275498, Log Avg loss: 0.15919368, Global Avg Loss: 0.73501147, Time: 0.0208 Steps: 84550, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001529, Sample Num: 24464, Cur Loss: 0.21537429, Cur Avg Loss: 0.20274627, Log Avg loss: 0.20142349, Global Avg Loss: 0.73494837, Time: 0.0209 Steps: 84560, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001539, Sample Num: 24624, Cur Loss: 0.12873653, Cur Avg Loss: 0.20238768, Log Avg loss: 0.14755912, Global Avg Loss: 0.73487891, Time: 0.0246 Steps: 84570, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001549, Sample Num: 24784, Cur Loss: 0.10666262, Cur Avg Loss: 0.20221551, Log Avg loss: 0.17571797, Global Avg Loss: 0.73481280, Time: 0.0209 Steps: 84580, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001559, Sample Num: 24944, Cur Loss: 0.11184396, Cur Avg Loss: 0.20194691, Log Avg loss: 0.16034187, Global Avg Loss: 0.73474489, Time: 0.0208 Steps: 84590, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001569, Sample Num: 25104, Cur Loss: 0.23262408, Cur Avg Loss: 0.20189214, Log Avg loss: 0.19335359, Global Avg Loss: 0.73468089, Time: 0.0208 Steps: 84600, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001579, Sample Num: 25264, Cur Loss: 0.26455826, Cur Avg Loss: 0.20232290, Log Avg loss: 0.26990896, Global Avg Loss: 0.73462596, Time: 0.0209 Steps: 84610, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001589, Sample Num: 25424, Cur Loss: 0.13878334, Cur Avg Loss: 0.20213725, Log Avg loss: 0.17282251, Global Avg Loss: 0.73455957, Time: 0.0209 Steps: 84620, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001599, Sample Num: 25584, Cur Loss: 0.20643738, Cur Avg Loss: 0.20175702, Log Avg loss: 0.14133827, Global Avg Loss: 0.73448947, Time: 0.0209 Steps: 84630, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001609, Sample Num: 25744, Cur Loss: 0.43415451, Cur Avg Loss: 0.20194669, Log Avg loss: 0.23227461, Global Avg Loss: 0.73443014, Time: 0.0209 Steps: 84640, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001619, Sample Num: 25904, Cur Loss: 0.20625857, Cur Avg Loss: 0.20203694, Log Avg loss: 0.21655904, Global Avg Loss: 0.73436896, Time: 0.0209 Steps: 84650, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001629, Sample Num: 26064, Cur Loss: 0.40355575, Cur Avg Loss: 0.20190046, Log Avg loss: 0.17980475, Global Avg Loss: 0.73430346, Time: 0.0208 Steps: 84660, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001639, Sample Num: 26224, Cur Loss: 0.50787765, Cur Avg Loss: 0.20208913, Log Avg loss: 0.23282299, Global Avg Loss: 0.73424423, Time: 0.0209 Steps: 84670, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001649, Sample Num: 26384, Cur Loss: 0.16838583, Cur Avg Loss: 0.20169529, Log Avg loss: 0.13714469, Global Avg Loss: 0.73417372, Time: 0.0210 Steps: 84680, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001659, Sample Num: 26544, Cur Loss: 0.12022823, Cur Avg Loss: 0.20166616, Log Avg loss: 0.19686313, Global Avg Loss: 0.73411027, Time: 0.0209 Steps: 84690, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001669, Sample Num: 26704, Cur Loss: 0.37067986, Cur Avg Loss: 0.20129543, Log Avg loss: 0.13979073, Global Avg Loss: 0.73404010, Time: 0.0209 Steps: 84700, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001679, Sample Num: 26864, Cur Loss: 0.31460494, Cur Avg Loss: 0.20143218, Log Avg loss: 0.22425620, Global Avg Loss: 0.73397992, Time: 0.0209 Steps: 84710, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001689, Sample Num: 27024, Cur Loss: 0.26213327, Cur Avg Loss: 0.20152734, Log Avg loss: 0.21750500, Global Avg Loss: 0.73391896, Time: 0.0209 Steps: 84720, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001699, Sample Num: 27184, Cur Loss: 0.25040090, Cur Avg Loss: 0.20179385, Log Avg loss: 0.24680679, Global Avg Loss: 0.73386147, Time: 0.0209 Steps: 84730, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001709, Sample Num: 27344, Cur Loss: 0.44225842, Cur Avg Loss: 0.20156221, Log Avg loss: 0.16220614, Global Avg Loss: 0.73379401, Time: 0.0209 Steps: 84740, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001719, Sample Num: 27504, Cur Loss: 0.13991897, Cur Avg Loss: 0.20138313, Log Avg loss: 0.17077826, Global Avg Loss: 0.73372758, Time: 0.0209 Steps: 84750, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001729, Sample Num: 27664, Cur Loss: 0.07699835, Cur Avg Loss: 0.20132389, Log Avg loss: 0.19114141, Global Avg Loss: 0.73366357, Time: 0.0209 Steps: 84760, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001739, Sample Num: 27824, Cur Loss: 0.15166022, Cur Avg Loss: 0.20131535, Log Avg loss: 0.19983890, Global Avg Loss: 0.73360059, Time: 0.0209 Steps: 84770, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001749, Sample Num: 27984, Cur Loss: 0.16491485, Cur Avg Loss: 0.20136800, Log Avg loss: 0.21052373, Global Avg Loss: 0.73353889, Time: 0.0209 Steps: 84780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001759, Sample Num: 28144, Cur Loss: 0.16110981, Cur Avg Loss: 0.20111263, Log Avg loss: 0.15644866, Global Avg Loss: 0.73347083, Time: 0.0209 Steps: 84790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001769, Sample Num: 28304, Cur Loss: 0.18127424, Cur Avg Loss: 0.20121815, Log Avg loss: 0.21977808, Global Avg Loss: 0.73341026, Time: 0.0209 Steps: 84800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001779, Sample Num: 28464, Cur Loss: 0.13371828, Cur Avg Loss: 0.20127793, Log Avg loss: 0.21185271, Global Avg Loss: 0.73334876, Time: 0.0209 Steps: 84810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001789, Sample Num: 28624, Cur Loss: 0.25316298, Cur Avg Loss: 0.20108904, Log Avg loss: 0.16748625, Global Avg Loss: 0.73328204, Time: 0.0209 Steps: 84820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001799, Sample Num: 28784, Cur Loss: 0.28193194, Cur Avg Loss: 0.20149030, Log Avg loss: 0.27327501, Global Avg Loss: 0.73322782, Time: 0.0242 Steps: 84830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001809, Sample Num: 28944, Cur Loss: 0.26736894, Cur Avg Loss: 0.20176218, Log Avg loss: 0.25067365, Global Avg Loss: 0.73317094, Time: 0.0210 Steps: 84840, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001819, Sample Num: 29104, Cur Loss: 0.13538979, Cur Avg Loss: 0.20159911, Log Avg loss: 0.17210055, Global Avg Loss: 0.73310481, Time: 0.0210 Steps: 84850, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001829, Sample Num: 29264, Cur Loss: 0.52483058, Cur Avg Loss: 0.20239474, Log Avg loss: 0.34711867, Global Avg Loss: 0.73305933, Time: 0.0211 Steps: 84860, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001839, Sample Num: 29424, Cur Loss: 0.17557152, Cur Avg Loss: 0.20211362, Log Avg loss: 0.15069733, Global Avg Loss: 0.73299071, Time: 0.0210 Steps: 84870, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001849, Sample Num: 29584, Cur Loss: 0.17284460, Cur Avg Loss: 0.20192782, Log Avg loss: 0.16775970, Global Avg Loss: 0.73292412, Time: 0.0210 Steps: 84880, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001859, Sample Num: 29744, Cur Loss: 0.05162994, Cur Avg Loss: 0.20199028, Log Avg loss: 0.21353820, Global Avg Loss: 0.73286294, Time: 0.0211 Steps: 84890, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001869, Sample Num: 29904, Cur Loss: 0.22597331, Cur Avg Loss: 0.20206059, Log Avg loss: 0.21513241, Global Avg Loss: 0.73280196, Time: 0.0210 Steps: 84900, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001879, Sample Num: 30064, Cur Loss: 0.39667970, Cur Avg Loss: 0.20254730, Log Avg loss: 0.29351231, Global Avg Loss: 0.73275022, Time: 0.0211 Steps: 84910, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001889, Sample Num: 30224, Cur Loss: 0.15680760, Cur Avg Loss: 0.20276402, Log Avg loss: 0.24348563, Global Avg Loss: 0.73269260, Time: 0.0211 Steps: 84920, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001899, Sample Num: 30384, Cur Loss: 0.06933048, Cur Avg Loss: 0.20262477, Log Avg loss: 0.17632130, Global Avg Loss: 0.73262710, Time: 0.0210 Steps: 84930, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001909, Sample Num: 30544, Cur Loss: 0.37800646, Cur Avg Loss: 0.20261831, Log Avg loss: 0.20139098, Global Avg Loss: 0.73256455, Time: 0.0211 Steps: 84940, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001919, Sample Num: 30704, Cur Loss: 0.04765665, Cur Avg Loss: 0.20243449, Log Avg loss: 0.16734319, Global Avg Loss: 0.73249802, Time: 0.0211 Steps: 84950, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001929, Sample Num: 30864, Cur Loss: 0.21405578, Cur Avg Loss: 0.20239370, Log Avg loss: 0.19456599, Global Avg Loss: 0.73243470, Time: 0.0211 Steps: 84960, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001939, Sample Num: 31024, Cur Loss: 0.24208644, Cur Avg Loss: 0.20240779, Log Avg loss: 0.20512674, Global Avg Loss: 0.73237264, Time: 0.0210 Steps: 84970, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001949, Sample Num: 31184, Cur Loss: 0.38411486, Cur Avg Loss: 0.20239493, Log Avg loss: 0.19990036, Global Avg Loss: 0.73230998, Time: 0.0211 Steps: 84980, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001959, Sample Num: 31344, Cur Loss: 0.06365000, Cur Avg Loss: 0.20251021, Log Avg loss: 0.22497912, Global Avg Loss: 0.73225029, Time: 0.0211 Steps: 84990, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001969, Sample Num: 31504, Cur Loss: 0.08118993, Cur Avg Loss: 0.20249082, Log Avg loss: 0.19869179, Global Avg Loss: 0.73218752, Time: 0.0211 Steps: 85000, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001979, Sample Num: 31664, Cur Loss: 0.06610664, Cur Avg Loss: 0.20231886, Log Avg loss: 0.16846106, Global Avg Loss: 0.73212121, Time: 0.0210 Steps: 85010, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001989, Sample Num: 31824, Cur Loss: 0.07041247, Cur Avg Loss: 0.20212169, Log Avg loss: 0.16310173, Global Avg Loss: 0.73205428, Time: 0.0210 Steps: 85020, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001999, Sample Num: 31984, Cur Loss: 0.30069670, Cur Avg Loss: 0.20197319, Log Avg loss: 0.17243608, Global Avg Loss: 0.73198846, Time: 0.0211 Steps: 85030, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002009, Sample Num: 32144, Cur Loss: 0.20402625, Cur Avg Loss: 0.20206905, Log Avg loss: 0.22123111, Global Avg Loss: 0.73192840, Time: 0.0210 Steps: 85040, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002019, Sample Num: 32304, Cur Loss: 0.27763721, Cur Avg Loss: 0.20196769, Log Avg loss: 0.18160489, Global Avg Loss: 0.73186370, Time: 0.0211 Steps: 85050, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002029, Sample Num: 32464, Cur Loss: 0.16880618, Cur Avg Loss: 0.20211035, Log Avg loss: 0.23091227, Global Avg Loss: 0.73180480, Time: 0.0211 Steps: 85060, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002039, Sample Num: 32624, Cur Loss: 0.03696335, Cur Avg Loss: 0.20235735, Log Avg loss: 0.25247421, Global Avg Loss: 0.73174846, Time: 0.0212 Steps: 85070, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002049, Sample Num: 32784, Cur Loss: 0.04643828, Cur Avg Loss: 0.20290325, Log Avg loss: 0.31421260, Global Avg Loss: 0.73169938, Time: 0.0246 Steps: 85080, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002059, Sample Num: 32944, Cur Loss: 0.09734765, Cur Avg Loss: 0.20277244, Log Avg loss: 0.17596897, Global Avg Loss: 0.73163407, Time: 0.0210 Steps: 85090, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002069, Sample Num: 33104, Cur Loss: 0.07933927, Cur Avg Loss: 0.20237906, Log Avg loss: 0.12138337, Global Avg Loss: 0.73156236, Time: 0.0210 Steps: 85100, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002079, Sample Num: 33264, Cur Loss: 0.10362820, Cur Avg Loss: 0.20237092, Log Avg loss: 0.20068591, Global Avg Loss: 0.73149999, Time: 0.0210 Steps: 85110, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002089, Sample Num: 33424, Cur Loss: 0.11298041, Cur Avg Loss: 0.20184559, Log Avg loss: 0.09263021, Global Avg Loss: 0.73142493, Time: 0.0210 Steps: 85120, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002099, Sample Num: 33584, Cur Loss: 0.13718081, Cur Avg Loss: 0.20176956, Log Avg loss: 0.18588675, Global Avg Loss: 0.73136085, Time: 0.0209 Steps: 85130, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002109, Sample Num: 33744, Cur Loss: 0.31134206, Cur Avg Loss: 0.20209559, Log Avg loss: 0.27052867, Global Avg Loss: 0.73130672, Time: 0.0210 Steps: 85140, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002119, Sample Num: 33904, Cur Loss: 0.10402514, Cur Avg Loss: 0.20186565, Log Avg loss: 0.15337169, Global Avg Loss: 0.73123885, Time: 0.0210 Steps: 85150, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002129, Sample Num: 34055, Cur Loss: 0.24515375, Cur Avg Loss: 0.20245289, Log Avg loss: 0.32688774, Global Avg Loss: 0.73119137, Time: 0.0102 Steps: 85160, Updated lr: 0.000020 ***** Running evaluation checkpoint-85160 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-85160 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.664154, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.263487, "eval_total_loss": 185.23103, "eval_mae": 0.395216, "eval_mse": 0.2635, "eval_r2": 0.832502, "eval_sp_statistic": 0.90221, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924196, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.326869, "test_total_loss": 164.0881, "test_mae": 0.350639, "test_mse": 0.326986, "test_r2": 0.788961, "test_sp_statistic": 0.875294, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921042, "test_ps_pvalue": 0.0, "lr": 2.0189663347558086e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7311913686688113, "train_cur_epoch_loss": 431.02219426538795, "train_cur_epoch_avg_loss": 0.2024528859865608, "train_cur_epoch_time": 44.66415357589722, "train_cur_epoch_avg_time": 0.02097893545133735, "epoch": 40, "step": 85160} ################################################## Training, Epoch: 0041, Batch: 000010, Sample Num: 160, Cur Loss: 0.11320323, Cur Avg Loss: 0.23833039, Log Avg loss: 0.23833039, Global Avg Loss: 0.73113350, Time: 0.0208 Steps: 85170, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000020, Sample Num: 320, Cur Loss: 0.38855016, Cur Avg Loss: 0.20624913, Log Avg loss: 0.17416787, Global Avg Loss: 0.73106811, Time: 0.0208 Steps: 85180, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000030, Sample Num: 480, Cur Loss: 0.15896311, Cur Avg Loss: 0.21129068, Log Avg loss: 0.22137378, Global Avg Loss: 0.73100828, Time: 0.0208 Steps: 85190, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000040, Sample Num: 640, Cur Loss: 0.09696075, Cur Avg Loss: 0.21437979, Log Avg loss: 0.22364713, Global Avg Loss: 0.73094873, Time: 0.0208 Steps: 85200, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000050, Sample Num: 800, Cur Loss: 0.09086145, Cur Avg Loss: 0.20855383, Log Avg loss: 0.18524997, Global Avg Loss: 0.73088469, Time: 0.0208 Steps: 85210, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000060, Sample Num: 960, Cur Loss: 0.50048482, Cur Avg Loss: 0.22170023, Log Avg loss: 0.28743226, Global Avg Loss: 0.73083266, Time: 0.0208 Steps: 85220, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000070, Sample Num: 1120, Cur Loss: 0.11459874, Cur Avg Loss: 0.21533188, Log Avg loss: 0.17712175, Global Avg Loss: 0.73076769, Time: 0.0208 Steps: 85230, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000080, Sample Num: 1280, Cur Loss: 0.09693519, Cur Avg Loss: 0.21277392, Log Avg loss: 0.19486819, Global Avg Loss: 0.73070482, Time: 0.0209 Steps: 85240, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000090, Sample Num: 1440, Cur Loss: 0.25593987, Cur Avg Loss: 0.21263003, Log Avg loss: 0.21147892, Global Avg Loss: 0.73064391, Time: 0.0208 Steps: 85250, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000100, Sample Num: 1600, Cur Loss: 0.08714022, Cur Avg Loss: 0.20905429, Log Avg loss: 0.17687265, Global Avg Loss: 0.73057896, Time: 0.0208 Steps: 85260, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000110, Sample Num: 1760, Cur Loss: 0.06280436, Cur Avg Loss: 0.20379870, Log Avg loss: 0.15124274, Global Avg Loss: 0.73051102, Time: 0.0209 Steps: 85270, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000120, Sample Num: 1920, Cur Loss: 0.25607729, Cur Avg Loss: 0.20243987, Log Avg loss: 0.18749274, Global Avg Loss: 0.73044735, Time: 0.0208 Steps: 85280, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000130, Sample Num: 2080, Cur Loss: 0.09910131, Cur Avg Loss: 0.19827304, Log Avg loss: 0.14827119, Global Avg Loss: 0.73037909, Time: 0.0208 Steps: 85290, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000140, Sample Num: 2240, Cur Loss: 0.25090730, Cur Avg Loss: 0.19643391, Log Avg loss: 0.17252516, Global Avg Loss: 0.73031369, Time: 0.0209 Steps: 85300, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000150, Sample Num: 2400, Cur Loss: 0.19835426, Cur Avg Loss: 0.19447069, Log Avg loss: 0.16698564, Global Avg Loss: 0.73024766, Time: 0.0209 Steps: 85310, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000160, Sample Num: 2560, Cur Loss: 0.10539458, Cur Avg Loss: 0.19579677, Log Avg loss: 0.21568798, Global Avg Loss: 0.73018735, Time: 0.0208 Steps: 85320, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000170, Sample Num: 2720, Cur Loss: 0.40492240, Cur Avg Loss: 0.19865651, Log Avg loss: 0.24441237, Global Avg Loss: 0.73013042, Time: 0.0208 Steps: 85330, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000180, Sample Num: 2880, Cur Loss: 0.24565765, Cur Avg Loss: 0.19662815, Log Avg loss: 0.16214596, Global Avg Loss: 0.73006386, Time: 0.0208 Steps: 85340, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000190, Sample Num: 3040, Cur Loss: 0.07165150, Cur Avg Loss: 0.19479049, Log Avg loss: 0.16171258, Global Avg Loss: 0.72999727, Time: 0.0208 Steps: 85350, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000200, Sample Num: 3200, Cur Loss: 0.10107405, Cur Avg Loss: 0.19187477, Log Avg loss: 0.13647609, Global Avg Loss: 0.72992774, Time: 0.0207 Steps: 85360, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000210, Sample Num: 3360, Cur Loss: 0.04065987, Cur Avg Loss: 0.19361371, Log Avg loss: 0.22839248, Global Avg Loss: 0.72986899, Time: 0.0208 Steps: 85370, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000220, Sample Num: 3520, Cur Loss: 0.40338483, Cur Avg Loss: 0.19226823, Log Avg loss: 0.16401329, Global Avg Loss: 0.72980272, Time: 0.0208 Steps: 85380, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000230, Sample Num: 3680, Cur Loss: 0.19558364, Cur Avg Loss: 0.19204273, Log Avg loss: 0.18708156, Global Avg Loss: 0.72973916, Time: 0.0208 Steps: 85390, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000240, Sample Num: 3840, Cur Loss: 0.15026708, Cur Avg Loss: 0.19313389, Log Avg loss: 0.21823074, Global Avg Loss: 0.72967926, Time: 0.0208 Steps: 85400, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000250, Sample Num: 4000, Cur Loss: 0.12643266, Cur Avg Loss: 0.19260450, Log Avg loss: 0.17989901, Global Avg Loss: 0.72961489, Time: 0.0208 Steps: 85410, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000260, Sample Num: 4160, Cur Loss: 0.18579566, Cur Avg Loss: 0.19226648, Log Avg loss: 0.18381614, Global Avg Loss: 0.72955100, Time: 0.0245 Steps: 85420, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000270, Sample Num: 4320, Cur Loss: 0.22185762, Cur Avg Loss: 0.19370113, Log Avg loss: 0.23100200, Global Avg Loss: 0.72949264, Time: 0.0208 Steps: 85430, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000280, Sample Num: 4480, Cur Loss: 0.19025835, Cur Avg Loss: 0.19274186, Log Avg loss: 0.16684160, Global Avg Loss: 0.72942679, Time: 0.0208 Steps: 85440, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000290, Sample Num: 4640, Cur Loss: 0.33410069, Cur Avg Loss: 0.19481884, Log Avg loss: 0.25297405, Global Avg Loss: 0.72937103, Time: 0.0209 Steps: 85450, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000300, Sample Num: 4800, Cur Loss: 0.20267969, Cur Avg Loss: 0.19609962, Log Avg loss: 0.23324231, Global Avg Loss: 0.72931297, Time: 0.0207 Steps: 85460, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000310, Sample Num: 4960, Cur Loss: 0.07039239, Cur Avg Loss: 0.19566873, Log Avg loss: 0.18274203, Global Avg Loss: 0.72924903, Time: 0.0209 Steps: 85470, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000320, Sample Num: 5120, Cur Loss: 0.16378367, Cur Avg Loss: 0.19553201, Log Avg loss: 0.19129362, Global Avg Loss: 0.72918609, Time: 0.0208 Steps: 85480, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000330, Sample Num: 5280, Cur Loss: 0.24812254, Cur Avg Loss: 0.19409552, Log Avg loss: 0.14812809, Global Avg Loss: 0.72911812, Time: 0.0209 Steps: 85490, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000340, Sample Num: 5440, Cur Loss: 0.04492119, Cur Avg Loss: 0.19344916, Log Avg loss: 0.17211929, Global Avg Loss: 0.72905298, Time: 0.0209 Steps: 85500, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000350, Sample Num: 5600, Cur Loss: 0.08263265, Cur Avg Loss: 0.19281782, Log Avg loss: 0.17135229, Global Avg Loss: 0.72898776, Time: 0.0207 Steps: 85510, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000360, Sample Num: 5760, Cur Loss: 0.17695287, Cur Avg Loss: 0.19328021, Log Avg loss: 0.20946374, Global Avg Loss: 0.72892701, Time: 0.0208 Steps: 85520, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000370, Sample Num: 5920, Cur Loss: 0.04613278, Cur Avg Loss: 0.19330641, Log Avg loss: 0.19424945, Global Avg Loss: 0.72886450, Time: 0.0207 Steps: 85530, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000380, Sample Num: 6080, Cur Loss: 0.17220706, Cur Avg Loss: 0.19177654, Log Avg loss: 0.13517142, Global Avg Loss: 0.72879509, Time: 0.0209 Steps: 85540, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000390, Sample Num: 6240, Cur Loss: 0.13214567, Cur Avg Loss: 0.19258573, Log Avg loss: 0.22333517, Global Avg Loss: 0.72873601, Time: 0.0208 Steps: 85550, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000400, Sample Num: 6400, Cur Loss: 0.07314482, Cur Avg Loss: 0.19239022, Log Avg loss: 0.18476498, Global Avg Loss: 0.72867243, Time: 0.0207 Steps: 85560, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000410, Sample Num: 6560, Cur Loss: 0.32837456, Cur Avg Loss: 0.19245080, Log Avg loss: 0.19487408, Global Avg Loss: 0.72861005, Time: 0.0207 Steps: 85570, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000420, Sample Num: 6720, Cur Loss: 0.21791370, Cur Avg Loss: 0.19084895, Log Avg loss: 0.12517328, Global Avg Loss: 0.72853954, Time: 0.0208 Steps: 85580, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000430, Sample Num: 6880, Cur Loss: 0.36911556, Cur Avg Loss: 0.19001427, Log Avg loss: 0.15495757, Global Avg Loss: 0.72847252, Time: 0.0207 Steps: 85590, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000440, Sample Num: 7040, Cur Loss: 0.07508698, Cur Avg Loss: 0.19037391, Log Avg loss: 0.20583846, Global Avg Loss: 0.72841147, Time: 0.0208 Steps: 85600, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000450, Sample Num: 7200, Cur Loss: 0.20065866, Cur Avg Loss: 0.19119145, Log Avg loss: 0.22716314, Global Avg Loss: 0.72835292, Time: 0.0209 Steps: 85610, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000460, Sample Num: 7360, Cur Loss: 0.27023479, Cur Avg Loss: 0.19175884, Log Avg loss: 0.21729131, Global Avg Loss: 0.72829323, Time: 0.0208 Steps: 85620, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000470, Sample Num: 7520, Cur Loss: 0.42571470, Cur Avg Loss: 0.19210061, Log Avg loss: 0.20782214, Global Avg Loss: 0.72823244, Time: 0.0208 Steps: 85630, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000480, Sample Num: 7680, Cur Loss: 0.09538420, Cur Avg Loss: 0.19257194, Log Avg loss: 0.21472467, Global Avg Loss: 0.72817248, Time: 0.0208 Steps: 85640, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000490, Sample Num: 7840, Cur Loss: 0.20307344, Cur Avg Loss: 0.19210194, Log Avg loss: 0.16954156, Global Avg Loss: 0.72810726, Time: 0.0208 Steps: 85650, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000500, Sample Num: 8000, Cur Loss: 0.29605639, Cur Avg Loss: 0.19231296, Log Avg loss: 0.20265323, Global Avg Loss: 0.72804592, Time: 0.0207 Steps: 85660, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000510, Sample Num: 8160, Cur Loss: 0.07183997, Cur Avg Loss: 0.19124588, Log Avg loss: 0.13789204, Global Avg Loss: 0.72797703, Time: 0.0208 Steps: 85670, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000520, Sample Num: 8320, Cur Loss: 0.04537990, Cur Avg Loss: 0.19037048, Log Avg loss: 0.14572478, Global Avg Loss: 0.72790908, Time: 0.0209 Steps: 85680, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000530, Sample Num: 8480, Cur Loss: 0.39119530, Cur Avg Loss: 0.19012051, Log Avg loss: 0.17712196, Global Avg Loss: 0.72784480, Time: 0.0208 Steps: 85690, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000540, Sample Num: 8640, Cur Loss: 0.07504241, Cur Avg Loss: 0.18987744, Log Avg loss: 0.17699487, Global Avg Loss: 0.72778052, Time: 0.0208 Steps: 85700, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000550, Sample Num: 8800, Cur Loss: 0.66674399, Cur Avg Loss: 0.19069883, Log Avg loss: 0.23505369, Global Avg Loss: 0.72772303, Time: 0.0208 Steps: 85710, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000560, Sample Num: 8960, Cur Loss: 0.21323469, Cur Avg Loss: 0.19050117, Log Avg loss: 0.17963011, Global Avg Loss: 0.72765909, Time: 0.0208 Steps: 85720, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000570, Sample Num: 9120, Cur Loss: 0.13479832, Cur Avg Loss: 0.19049153, Log Avg loss: 0.18995148, Global Avg Loss: 0.72759637, Time: 0.0208 Steps: 85730, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000580, Sample Num: 9280, Cur Loss: 0.37228489, Cur Avg Loss: 0.19059380, Log Avg loss: 0.19642340, Global Avg Loss: 0.72753442, Time: 0.0208 Steps: 85740, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000590, Sample Num: 9440, Cur Loss: 0.12335173, Cur Avg Loss: 0.19059548, Log Avg loss: 0.19069276, Global Avg Loss: 0.72747182, Time: 0.0208 Steps: 85750, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000600, Sample Num: 9600, Cur Loss: 0.08672942, Cur Avg Loss: 0.18978421, Log Avg loss: 0.14191962, Global Avg Loss: 0.72740354, Time: 0.0208 Steps: 85760, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000610, Sample Num: 9760, Cur Loss: 0.12270525, Cur Avg Loss: 0.18885308, Log Avg loss: 0.13298501, Global Avg Loss: 0.72733423, Time: 0.0208 Steps: 85770, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000620, Sample Num: 9920, Cur Loss: 0.08564778, Cur Avg Loss: 0.18915906, Log Avg loss: 0.20782402, Global Avg Loss: 0.72727367, Time: 0.0208 Steps: 85780, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000630, Sample Num: 10080, Cur Loss: 0.12011427, Cur Avg Loss: 0.19020297, Log Avg loss: 0.25492503, Global Avg Loss: 0.72721861, Time: 0.0208 Steps: 85790, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000640, Sample Num: 10240, Cur Loss: 0.05268793, Cur Avg Loss: 0.19115023, Log Avg loss: 0.25082771, Global Avg Loss: 0.72716309, Time: 0.0208 Steps: 85800, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000650, Sample Num: 10400, Cur Loss: 0.17676380, Cur Avg Loss: 0.19111662, Log Avg loss: 0.18896574, Global Avg Loss: 0.72710037, Time: 0.0208 Steps: 85810, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000660, Sample Num: 10560, Cur Loss: 0.17014000, Cur Avg Loss: 0.19036192, Log Avg loss: 0.14130619, Global Avg Loss: 0.72703211, Time: 0.0208 Steps: 85820, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000670, Sample Num: 10720, Cur Loss: 0.28425100, Cur Avg Loss: 0.18980788, Log Avg loss: 0.15324160, Global Avg Loss: 0.72696526, Time: 0.0208 Steps: 85830, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000680, Sample Num: 10880, Cur Loss: 0.29326054, Cur Avg Loss: 0.18972950, Log Avg loss: 0.18447819, Global Avg Loss: 0.72690206, Time: 0.0208 Steps: 85840, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000690, Sample Num: 11040, Cur Loss: 0.19871821, Cur Avg Loss: 0.18980404, Log Avg loss: 0.19487270, Global Avg Loss: 0.72684009, Time: 0.0208 Steps: 85850, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000700, Sample Num: 11200, Cur Loss: 0.47363570, Cur Avg Loss: 0.18985161, Log Avg loss: 0.19313397, Global Avg Loss: 0.72677793, Time: 0.0208 Steps: 85860, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000710, Sample Num: 11360, Cur Loss: 0.21163450, Cur Avg Loss: 0.18950083, Log Avg loss: 0.16494577, Global Avg Loss: 0.72671250, Time: 0.0208 Steps: 85870, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000720, Sample Num: 11520, Cur Loss: 0.35549703, Cur Avg Loss: 0.18961794, Log Avg loss: 0.19793265, Global Avg Loss: 0.72665093, Time: 0.0208 Steps: 85880, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000730, Sample Num: 11680, Cur Loss: 0.20111445, Cur Avg Loss: 0.18941023, Log Avg loss: 0.17445516, Global Avg Loss: 0.72658664, Time: 0.0209 Steps: 85890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000740, Sample Num: 11840, Cur Loss: 0.15069944, Cur Avg Loss: 0.18923778, Log Avg loss: 0.17664950, Global Avg Loss: 0.72652262, Time: 0.0208 Steps: 85900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000750, Sample Num: 12000, Cur Loss: 0.13154997, Cur Avg Loss: 0.18981474, Log Avg loss: 0.23250915, Global Avg Loss: 0.72646511, Time: 0.0208 Steps: 85910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000760, Sample Num: 12160, Cur Loss: 0.38472474, Cur Avg Loss: 0.18998745, Log Avg loss: 0.20294095, Global Avg Loss: 0.72640418, Time: 0.0208 Steps: 85920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000770, Sample Num: 12320, Cur Loss: 0.08574049, Cur Avg Loss: 0.19064901, Log Avg loss: 0.24092800, Global Avg Loss: 0.72634769, Time: 0.0246 Steps: 85930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000780, Sample Num: 12480, Cur Loss: 0.15597548, Cur Avg Loss: 0.19059129, Log Avg loss: 0.18614674, Global Avg Loss: 0.72628483, Time: 0.0208 Steps: 85940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000790, Sample Num: 12640, Cur Loss: 0.30860126, Cur Avg Loss: 0.19044124, Log Avg loss: 0.17873731, Global Avg Loss: 0.72622112, Time: 0.0209 Steps: 85950, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000800, Sample Num: 12800, Cur Loss: 0.08998010, Cur Avg Loss: 0.19030695, Log Avg loss: 0.17969811, Global Avg Loss: 0.72615754, Time: 0.0209 Steps: 85960, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000810, Sample Num: 12960, Cur Loss: 0.27292252, Cur Avg Loss: 0.18974405, Log Avg loss: 0.14471169, Global Avg Loss: 0.72608991, Time: 0.0208 Steps: 85970, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000820, Sample Num: 13120, Cur Loss: 0.25796664, Cur Avg Loss: 0.18949408, Log Avg loss: 0.16924691, Global Avg Loss: 0.72602515, Time: 0.0208 Steps: 85980, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000830, Sample Num: 13280, Cur Loss: 0.08105428, Cur Avg Loss: 0.18912336, Log Avg loss: 0.15872441, Global Avg Loss: 0.72595917, Time: 0.0208 Steps: 85990, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000840, Sample Num: 13440, Cur Loss: 0.24824098, Cur Avg Loss: 0.18895770, Log Avg loss: 0.17520714, Global Avg Loss: 0.72589513, Time: 0.0209 Steps: 86000, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000850, Sample Num: 13600, Cur Loss: 0.12642017, Cur Avg Loss: 0.18954744, Log Avg loss: 0.23908622, Global Avg Loss: 0.72583853, Time: 0.0208 Steps: 86010, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000860, Sample Num: 13760, Cur Loss: 0.23507354, Cur Avg Loss: 0.18932408, Log Avg loss: 0.17033865, Global Avg Loss: 0.72577396, Time: 0.0208 Steps: 86020, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000870, Sample Num: 13920, Cur Loss: 0.16054836, Cur Avg Loss: 0.18851423, Log Avg loss: 0.11886685, Global Avg Loss: 0.72570341, Time: 0.0208 Steps: 86030, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000880, Sample Num: 14080, Cur Loss: 0.10835197, Cur Avg Loss: 0.18803769, Log Avg loss: 0.14657847, Global Avg Loss: 0.72563610, Time: 0.0208 Steps: 86040, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000890, Sample Num: 14240, Cur Loss: 0.09104280, Cur Avg Loss: 0.18843735, Log Avg loss: 0.22360779, Global Avg Loss: 0.72557776, Time: 0.0208 Steps: 86050, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000900, Sample Num: 14400, Cur Loss: 0.18853413, Cur Avg Loss: 0.18831082, Log Avg loss: 0.17704915, Global Avg Loss: 0.72551402, Time: 0.0208 Steps: 86060, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000910, Sample Num: 14560, Cur Loss: 0.15853600, Cur Avg Loss: 0.18831950, Log Avg loss: 0.18910078, Global Avg Loss: 0.72545170, Time: 0.0208 Steps: 86070, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000920, Sample Num: 14720, Cur Loss: 0.13159133, Cur Avg Loss: 0.18874905, Log Avg loss: 0.22783836, Global Avg Loss: 0.72539389, Time: 0.0208 Steps: 86080, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000930, Sample Num: 14880, Cur Loss: 0.12241647, Cur Avg Loss: 0.18906631, Log Avg loss: 0.21825448, Global Avg Loss: 0.72533498, Time: 0.0208 Steps: 86090, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000940, Sample Num: 15040, Cur Loss: 0.13316321, Cur Avg Loss: 0.18966535, Log Avg loss: 0.24537538, Global Avg Loss: 0.72527924, Time: 0.0208 Steps: 86100, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000950, Sample Num: 15200, Cur Loss: 0.32854500, Cur Avg Loss: 0.18974367, Log Avg loss: 0.19710585, Global Avg Loss: 0.72521790, Time: 0.0208 Steps: 86110, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000960, Sample Num: 15360, Cur Loss: 0.16793963, Cur Avg Loss: 0.18998122, Log Avg loss: 0.21254830, Global Avg Loss: 0.72515837, Time: 0.0208 Steps: 86120, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000970, Sample Num: 15520, Cur Loss: 0.13725156, Cur Avg Loss: 0.19003674, Log Avg loss: 0.19536709, Global Avg Loss: 0.72509686, Time: 0.0208 Steps: 86130, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000980, Sample Num: 15680, Cur Loss: 0.24748790, Cur Avg Loss: 0.19052634, Log Avg loss: 0.23801733, Global Avg Loss: 0.72504032, Time: 0.0208 Steps: 86140, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000990, Sample Num: 15840, Cur Loss: 0.30837843, Cur Avg Loss: 0.19176804, Log Avg loss: 0.31345451, Global Avg Loss: 0.72499254, Time: 0.0208 Steps: 86150, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001000, Sample Num: 16000, Cur Loss: 0.10036187, Cur Avg Loss: 0.19188784, Log Avg loss: 0.20374883, Global Avg Loss: 0.72493204, Time: 0.0208 Steps: 86160, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001010, Sample Num: 16160, Cur Loss: 0.33993465, Cur Avg Loss: 0.19194415, Log Avg loss: 0.19757472, Global Avg Loss: 0.72487084, Time: 0.0208 Steps: 86170, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001020, Sample Num: 16320, Cur Loss: 0.35692191, Cur Avg Loss: 0.19186408, Log Avg loss: 0.18377694, Global Avg Loss: 0.72480806, Time: 0.0208 Steps: 86180, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001030, Sample Num: 16480, Cur Loss: 0.21928035, Cur Avg Loss: 0.19236733, Log Avg loss: 0.24369924, Global Avg Loss: 0.72475224, Time: 0.0227 Steps: 86190, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001040, Sample Num: 16640, Cur Loss: 0.16763607, Cur Avg Loss: 0.19233626, Log Avg loss: 0.18913576, Global Avg Loss: 0.72469010, Time: 0.0209 Steps: 86200, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001050, Sample Num: 16800, Cur Loss: 0.15564883, Cur Avg Loss: 0.19195459, Log Avg loss: 0.15226093, Global Avg Loss: 0.72462370, Time: 0.0209 Steps: 86210, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001060, Sample Num: 16960, Cur Loss: 0.26742747, Cur Avg Loss: 0.19148117, Log Avg loss: 0.14177220, Global Avg Loss: 0.72455610, Time: 0.0209 Steps: 86220, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001070, Sample Num: 17120, Cur Loss: 0.25740904, Cur Avg Loss: 0.19176973, Log Avg loss: 0.22235646, Global Avg Loss: 0.72449786, Time: 0.0209 Steps: 86230, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001080, Sample Num: 17280, Cur Loss: 0.07187439, Cur Avg Loss: 0.19170756, Log Avg loss: 0.18505596, Global Avg Loss: 0.72443531, Time: 0.0209 Steps: 86240, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001090, Sample Num: 17440, Cur Loss: 0.44264987, Cur Avg Loss: 0.19238364, Log Avg loss: 0.26539988, Global Avg Loss: 0.72438209, Time: 0.0209 Steps: 86250, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001100, Sample Num: 17600, Cur Loss: 0.19993666, Cur Avg Loss: 0.19337260, Log Avg loss: 0.30116974, Global Avg Loss: 0.72433303, Time: 0.0209 Steps: 86260, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001110, Sample Num: 17760, Cur Loss: 0.23171069, Cur Avg Loss: 0.19317988, Log Avg loss: 0.17198086, Global Avg Loss: 0.72426900, Time: 0.0209 Steps: 86270, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001120, Sample Num: 17920, Cur Loss: 0.09697065, Cur Avg Loss: 0.19271324, Log Avg loss: 0.14091564, Global Avg Loss: 0.72420139, Time: 0.0209 Steps: 86280, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001130, Sample Num: 18080, Cur Loss: 0.18192104, Cur Avg Loss: 0.19275959, Log Avg loss: 0.19795064, Global Avg Loss: 0.72414040, Time: 0.0209 Steps: 86290, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001140, Sample Num: 18240, Cur Loss: 0.25580725, Cur Avg Loss: 0.19299514, Log Avg loss: 0.21961270, Global Avg Loss: 0.72408194, Time: 0.0210 Steps: 86300, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001150, Sample Num: 18400, Cur Loss: 0.16621469, Cur Avg Loss: 0.19399178, Log Avg loss: 0.30760817, Global Avg Loss: 0.72403369, Time: 0.0209 Steps: 86310, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001160, Sample Num: 18560, Cur Loss: 0.29842299, Cur Avg Loss: 0.19453535, Log Avg loss: 0.25704662, Global Avg Loss: 0.72397959, Time: 0.0209 Steps: 86320, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001170, Sample Num: 18720, Cur Loss: 0.18298987, Cur Avg Loss: 0.19440906, Log Avg loss: 0.17975858, Global Avg Loss: 0.72391655, Time: 0.0209 Steps: 86330, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001180, Sample Num: 18880, Cur Loss: 0.44231132, Cur Avg Loss: 0.19427258, Log Avg loss: 0.17830499, Global Avg Loss: 0.72385335, Time: 0.0209 Steps: 86340, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001190, Sample Num: 19040, Cur Loss: 0.27352366, Cur Avg Loss: 0.19416651, Log Avg loss: 0.18165065, Global Avg Loss: 0.72379056, Time: 0.0209 Steps: 86350, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001200, Sample Num: 19200, Cur Loss: 0.12745667, Cur Avg Loss: 0.19406402, Log Avg loss: 0.18186719, Global Avg Loss: 0.72372781, Time: 0.0209 Steps: 86360, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001210, Sample Num: 19360, Cur Loss: 0.59953445, Cur Avg Loss: 0.19421637, Log Avg loss: 0.21249788, Global Avg Loss: 0.72366862, Time: 0.0209 Steps: 86370, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001220, Sample Num: 19520, Cur Loss: 0.13561124, Cur Avg Loss: 0.19487431, Log Avg loss: 0.27448503, Global Avg Loss: 0.72361662, Time: 0.0209 Steps: 86380, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001230, Sample Num: 19680, Cur Loss: 0.14456618, Cur Avg Loss: 0.19469000, Log Avg loss: 0.17220485, Global Avg Loss: 0.72355279, Time: 0.0209 Steps: 86390, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001240, Sample Num: 19840, Cur Loss: 0.14347488, Cur Avg Loss: 0.19475478, Log Avg loss: 0.20272252, Global Avg Loss: 0.72349251, Time: 0.0210 Steps: 86400, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001250, Sample Num: 20000, Cur Loss: 0.20136735, Cur Avg Loss: 0.19448976, Log Avg loss: 0.16162695, Global Avg Loss: 0.72342749, Time: 0.0209 Steps: 86410, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001260, Sample Num: 20160, Cur Loss: 0.06438126, Cur Avg Loss: 0.19405492, Log Avg loss: 0.13969983, Global Avg Loss: 0.72335994, Time: 0.0210 Steps: 86420, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001270, Sample Num: 20320, Cur Loss: 0.10560350, Cur Avg Loss: 0.19398675, Log Avg loss: 0.18539780, Global Avg Loss: 0.72329770, Time: 0.0209 Steps: 86430, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001280, Sample Num: 20480, Cur Loss: 0.36594504, Cur Avg Loss: 0.19407769, Log Avg loss: 0.20562755, Global Avg Loss: 0.72323781, Time: 0.0255 Steps: 86440, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001290, Sample Num: 20640, Cur Loss: 0.11488214, Cur Avg Loss: 0.19400167, Log Avg loss: 0.18427108, Global Avg Loss: 0.72317547, Time: 0.0209 Steps: 86450, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001300, Sample Num: 20800, Cur Loss: 0.41450882, Cur Avg Loss: 0.19425524, Log Avg loss: 0.22696514, Global Avg Loss: 0.72311808, Time: 0.0208 Steps: 86460, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001310, Sample Num: 20960, Cur Loss: 0.29314727, Cur Avg Loss: 0.19464449, Log Avg loss: 0.24524752, Global Avg Loss: 0.72306281, Time: 0.0208 Steps: 86470, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001320, Sample Num: 21120, Cur Loss: 0.23002267, Cur Avg Loss: 0.19465453, Log Avg loss: 0.19596969, Global Avg Loss: 0.72300186, Time: 0.0208 Steps: 86480, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001330, Sample Num: 21280, Cur Loss: 0.16970463, Cur Avg Loss: 0.19520515, Log Avg loss: 0.26788618, Global Avg Loss: 0.72294924, Time: 0.0208 Steps: 86490, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001340, Sample Num: 21440, Cur Loss: 0.13780625, Cur Avg Loss: 0.19497981, Log Avg loss: 0.16501051, Global Avg Loss: 0.72288474, Time: 0.0208 Steps: 86500, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001350, Sample Num: 21600, Cur Loss: 0.26669815, Cur Avg Loss: 0.19485070, Log Avg loss: 0.17754970, Global Avg Loss: 0.72282170, Time: 0.0208 Steps: 86510, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001360, Sample Num: 21760, Cur Loss: 0.14999142, Cur Avg Loss: 0.19472325, Log Avg loss: 0.17751807, Global Avg Loss: 0.72275868, Time: 0.0208 Steps: 86520, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001370, Sample Num: 21920, Cur Loss: 0.06911355, Cur Avg Loss: 0.19436907, Log Avg loss: 0.14620033, Global Avg Loss: 0.72269204, Time: 0.0208 Steps: 86530, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001380, Sample Num: 22080, Cur Loss: 0.06570472, Cur Avg Loss: 0.19429414, Log Avg loss: 0.18402789, Global Avg Loss: 0.72262980, Time: 0.0208 Steps: 86540, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001390, Sample Num: 22240, Cur Loss: 0.45160007, Cur Avg Loss: 0.19410789, Log Avg loss: 0.16840516, Global Avg Loss: 0.72256576, Time: 0.0208 Steps: 86550, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001400, Sample Num: 22400, Cur Loss: 0.25679678, Cur Avg Loss: 0.19427818, Log Avg loss: 0.21794962, Global Avg Loss: 0.72250747, Time: 0.0208 Steps: 86560, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001410, Sample Num: 22560, Cur Loss: 0.11694513, Cur Avg Loss: 0.19424484, Log Avg loss: 0.18957683, Global Avg Loss: 0.72244591, Time: 0.0208 Steps: 86570, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001420, Sample Num: 22720, Cur Loss: 0.16770405, Cur Avg Loss: 0.19407170, Log Avg loss: 0.16965891, Global Avg Loss: 0.72238206, Time: 0.0208 Steps: 86580, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001430, Sample Num: 22880, Cur Loss: 0.16022032, Cur Avg Loss: 0.19383202, Log Avg loss: 0.15979698, Global Avg Loss: 0.72231709, Time: 0.0208 Steps: 86590, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001440, Sample Num: 23040, Cur Loss: 0.40339845, Cur Avg Loss: 0.19383774, Log Avg loss: 0.19465666, Global Avg Loss: 0.72225616, Time: 0.0209 Steps: 86600, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001450, Sample Num: 23200, Cur Loss: 0.21882394, Cur Avg Loss: 0.19343293, Log Avg loss: 0.13514040, Global Avg Loss: 0.72218837, Time: 0.0211 Steps: 86610, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001460, Sample Num: 23360, Cur Loss: 0.11511697, Cur Avg Loss: 0.19335450, Log Avg loss: 0.18198214, Global Avg Loss: 0.72212600, Time: 0.0211 Steps: 86620, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001470, Sample Num: 23520, Cur Loss: 0.42682099, Cur Avg Loss: 0.19336731, Log Avg loss: 0.19523656, Global Avg Loss: 0.72206518, Time: 0.0210 Steps: 86630, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001480, Sample Num: 23680, Cur Loss: 0.22808607, Cur Avg Loss: 0.19346833, Log Avg loss: 0.20831867, Global Avg Loss: 0.72200589, Time: 0.0211 Steps: 86640, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001490, Sample Num: 23840, Cur Loss: 0.15402168, Cur Avg Loss: 0.19368922, Log Avg loss: 0.22638066, Global Avg Loss: 0.72194869, Time: 0.0211 Steps: 86650, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001500, Sample Num: 24000, Cur Loss: 0.73991358, Cur Avg Loss: 0.19413172, Log Avg loss: 0.26006399, Global Avg Loss: 0.72189539, Time: 0.0210 Steps: 86660, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001510, Sample Num: 24160, Cur Loss: 0.10000513, Cur Avg Loss: 0.19428028, Log Avg loss: 0.21656454, Global Avg Loss: 0.72183709, Time: 0.0211 Steps: 86670, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001520, Sample Num: 24320, Cur Loss: 0.04282332, Cur Avg Loss: 0.19413602, Log Avg loss: 0.17235276, Global Avg Loss: 0.72177369, Time: 0.0210 Steps: 86680, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001530, Sample Num: 24480, Cur Loss: 0.10998771, Cur Avg Loss: 0.19438418, Log Avg loss: 0.23210456, Global Avg Loss: 0.72171721, Time: 0.0211 Steps: 86690, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001540, Sample Num: 24640, Cur Loss: 0.07436259, Cur Avg Loss: 0.19444644, Log Avg loss: 0.20397191, Global Avg Loss: 0.72165749, Time: 0.0243 Steps: 86700, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001550, Sample Num: 24800, Cur Loss: 0.15573236, Cur Avg Loss: 0.19462400, Log Avg loss: 0.22196903, Global Avg Loss: 0.72159986, Time: 0.0211 Steps: 86710, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001560, Sample Num: 24960, Cur Loss: 0.40686831, Cur Avg Loss: 0.19494839, Log Avg loss: 0.24522845, Global Avg Loss: 0.72154493, Time: 0.0211 Steps: 86720, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001570, Sample Num: 25120, Cur Loss: 0.13435225, Cur Avg Loss: 0.19473335, Log Avg loss: 0.16118793, Global Avg Loss: 0.72148032, Time: 0.0210 Steps: 86730, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001580, Sample Num: 25280, Cur Loss: 0.18857533, Cur Avg Loss: 0.19468849, Log Avg loss: 0.18764444, Global Avg Loss: 0.72141878, Time: 0.0210 Steps: 86740, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001590, Sample Num: 25440, Cur Loss: 0.09162530, Cur Avg Loss: 0.19477889, Log Avg loss: 0.20906270, Global Avg Loss: 0.72135972, Time: 0.0211 Steps: 86750, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001600, Sample Num: 25600, Cur Loss: 0.22818609, Cur Avg Loss: 0.19448874, Log Avg loss: 0.14835517, Global Avg Loss: 0.72129367, Time: 0.0211 Steps: 86760, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001610, Sample Num: 25760, Cur Loss: 0.34198436, Cur Avg Loss: 0.19461010, Log Avg loss: 0.21402650, Global Avg Loss: 0.72123521, Time: 0.0211 Steps: 86770, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001620, Sample Num: 25920, Cur Loss: 0.13501969, Cur Avg Loss: 0.19470413, Log Avg loss: 0.20984407, Global Avg Loss: 0.72117628, Time: 0.0211 Steps: 86780, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001630, Sample Num: 26080, Cur Loss: 0.28979224, Cur Avg Loss: 0.19478420, Log Avg loss: 0.20775514, Global Avg Loss: 0.72111712, Time: 0.0211 Steps: 86790, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001640, Sample Num: 26240, Cur Loss: 0.24682239, Cur Avg Loss: 0.19522763, Log Avg loss: 0.26750675, Global Avg Loss: 0.72106486, Time: 0.0211 Steps: 86800, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001650, Sample Num: 26400, Cur Loss: 0.50520092, Cur Avg Loss: 0.19528639, Log Avg loss: 0.20492315, Global Avg Loss: 0.72100541, Time: 0.0212 Steps: 86810, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001660, Sample Num: 26560, Cur Loss: 0.22113845, Cur Avg Loss: 0.19482948, Log Avg loss: 0.11943924, Global Avg Loss: 0.72093612, Time: 0.0211 Steps: 86820, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001670, Sample Num: 26720, Cur Loss: 0.16137908, Cur Avg Loss: 0.19490718, Log Avg loss: 0.20780554, Global Avg Loss: 0.72087702, Time: 0.0212 Steps: 86830, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001680, Sample Num: 26880, Cur Loss: 0.29133049, Cur Avg Loss: 0.19490162, Log Avg loss: 0.19397235, Global Avg Loss: 0.72081635, Time: 0.0211 Steps: 86840, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001690, Sample Num: 27040, Cur Loss: 0.11072522, Cur Avg Loss: 0.19502335, Log Avg loss: 0.21547376, Global Avg Loss: 0.72075816, Time: 0.0210 Steps: 86850, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001700, Sample Num: 27200, Cur Loss: 0.13436934, Cur Avg Loss: 0.19522129, Log Avg loss: 0.22867329, Global Avg Loss: 0.72070151, Time: 0.0211 Steps: 86860, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001710, Sample Num: 27360, Cur Loss: 0.24635005, Cur Avg Loss: 0.19492495, Log Avg loss: 0.14454745, Global Avg Loss: 0.72063519, Time: 0.0211 Steps: 86870, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001720, Sample Num: 27520, Cur Loss: 0.13232873, Cur Avg Loss: 0.19508593, Log Avg loss: 0.22261322, Global Avg Loss: 0.72057786, Time: 0.0210 Steps: 86880, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001730, Sample Num: 27680, Cur Loss: 0.40000364, Cur Avg Loss: 0.19499593, Log Avg loss: 0.17951707, Global Avg Loss: 0.72051559, Time: 0.0211 Steps: 86890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001740, Sample Num: 27840, Cur Loss: 0.25978994, Cur Avg Loss: 0.19485020, Log Avg loss: 0.16963824, Global Avg Loss: 0.72045220, Time: 0.0211 Steps: 86900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001750, Sample Num: 28000, Cur Loss: 0.12320143, Cur Avg Loss: 0.19480556, Log Avg loss: 0.18703756, Global Avg Loss: 0.72039083, Time: 0.0210 Steps: 86910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001760, Sample Num: 28160, Cur Loss: 0.36795652, Cur Avg Loss: 0.19473303, Log Avg loss: 0.18204064, Global Avg Loss: 0.72032889, Time: 0.0210 Steps: 86920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001770, Sample Num: 28320, Cur Loss: 0.36107731, Cur Avg Loss: 0.19473796, Log Avg loss: 0.19560567, Global Avg Loss: 0.72026853, Time: 0.0210 Steps: 86930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001780, Sample Num: 28480, Cur Loss: 0.11151293, Cur Avg Loss: 0.19464200, Log Avg loss: 0.17765737, Global Avg Loss: 0.72020612, Time: 0.0212 Steps: 86940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001790, Sample Num: 28640, Cur Loss: 0.27677834, Cur Avg Loss: 0.19467662, Log Avg loss: 0.20083838, Global Avg Loss: 0.72014638, Time: 0.0211 Steps: 86950, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001800, Sample Num: 28800, Cur Loss: 0.23582238, Cur Avg Loss: 0.19474135, Log Avg loss: 0.20632924, Global Avg Loss: 0.72008730, Time: 0.0211 Steps: 86960, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001810, Sample Num: 28960, Cur Loss: 0.13020158, Cur Avg Loss: 0.19448852, Log Avg loss: 0.14897768, Global Avg Loss: 0.72002163, Time: 0.0210 Steps: 86970, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001820, Sample Num: 29120, Cur Loss: 0.18705100, Cur Avg Loss: 0.19440482, Log Avg loss: 0.17925592, Global Avg Loss: 0.71995946, Time: 0.0210 Steps: 86980, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001830, Sample Num: 29280, Cur Loss: 0.05060494, Cur Avg Loss: 0.19441259, Log Avg loss: 0.19582590, Global Avg Loss: 0.71989921, Time: 0.0211 Steps: 86990, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001840, Sample Num: 29440, Cur Loss: 0.22981346, Cur Avg Loss: 0.19431014, Log Avg loss: 0.17556301, Global Avg Loss: 0.71983664, Time: 0.0211 Steps: 87000, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001850, Sample Num: 29600, Cur Loss: 0.12973656, Cur Avg Loss: 0.19424850, Log Avg loss: 0.18290566, Global Avg Loss: 0.71977493, Time: 0.0211 Steps: 87010, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001860, Sample Num: 29760, Cur Loss: 0.12548366, Cur Avg Loss: 0.19424436, Log Avg loss: 0.19347939, Global Avg Loss: 0.71971445, Time: 0.0210 Steps: 87020, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001870, Sample Num: 29920, Cur Loss: 0.34334710, Cur Avg Loss: 0.19431145, Log Avg loss: 0.20678969, Global Avg Loss: 0.71965551, Time: 0.0210 Steps: 87030, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001880, Sample Num: 30080, Cur Loss: 0.06345335, Cur Avg Loss: 0.19440050, Log Avg loss: 0.21105407, Global Avg Loss: 0.71959708, Time: 0.0210 Steps: 87040, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001890, Sample Num: 30240, Cur Loss: 0.20466894, Cur Avg Loss: 0.19430180, Log Avg loss: 0.17574510, Global Avg Loss: 0.71953460, Time: 0.0211 Steps: 87050, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001900, Sample Num: 30400, Cur Loss: 0.64860851, Cur Avg Loss: 0.19490648, Log Avg loss: 0.30919125, Global Avg Loss: 0.71948747, Time: 0.0212 Steps: 87060, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001910, Sample Num: 30560, Cur Loss: 0.45432776, Cur Avg Loss: 0.19493859, Log Avg loss: 0.20103980, Global Avg Loss: 0.71942793, Time: 0.0210 Steps: 87070, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001920, Sample Num: 30720, Cur Loss: 0.15774485, Cur Avg Loss: 0.19525072, Log Avg loss: 0.25486631, Global Avg Loss: 0.71937458, Time: 0.0211 Steps: 87080, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001930, Sample Num: 30880, Cur Loss: 0.32357389, Cur Avg Loss: 0.19592907, Log Avg loss: 0.32617321, Global Avg Loss: 0.71932943, Time: 0.0212 Steps: 87090, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001940, Sample Num: 31040, Cur Loss: 0.13879429, Cur Avg Loss: 0.19613063, Log Avg loss: 0.23503229, Global Avg Loss: 0.71927383, Time: 0.0210 Steps: 87100, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001950, Sample Num: 31200, Cur Loss: 0.27951208, Cur Avg Loss: 0.19652408, Log Avg loss: 0.27285280, Global Avg Loss: 0.71922258, Time: 0.0211 Steps: 87110, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001960, Sample Num: 31360, Cur Loss: 0.51275516, Cur Avg Loss: 0.19696962, Log Avg loss: 0.28384890, Global Avg Loss: 0.71917261, Time: 0.0210 Steps: 87120, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001970, Sample Num: 31520, Cur Loss: 0.58741343, Cur Avg Loss: 0.19698133, Log Avg loss: 0.19927735, Global Avg Loss: 0.71911294, Time: 0.0211 Steps: 87130, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001980, Sample Num: 31680, Cur Loss: 0.21470644, Cur Avg Loss: 0.19715942, Log Avg loss: 0.23224346, Global Avg Loss: 0.71905706, Time: 0.0210 Steps: 87140, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001990, Sample Num: 31840, Cur Loss: 0.15910187, Cur Avg Loss: 0.19705915, Log Avg loss: 0.17720519, Global Avg Loss: 0.71899489, Time: 0.0210 Steps: 87150, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002000, Sample Num: 32000, Cur Loss: 0.41383857, Cur Avg Loss: 0.19711330, Log Avg loss: 0.20788931, Global Avg Loss: 0.71893625, Time: 0.0211 Steps: 87160, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002010, Sample Num: 32160, Cur Loss: 0.11466823, Cur Avg Loss: 0.19686595, Log Avg loss: 0.14739570, Global Avg Loss: 0.71887068, Time: 0.0211 Steps: 87170, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002020, Sample Num: 32320, Cur Loss: 0.23712096, Cur Avg Loss: 0.19693584, Log Avg loss: 0.21098478, Global Avg Loss: 0.71881243, Time: 0.0211 Steps: 87180, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002030, Sample Num: 32480, Cur Loss: 0.16136190, Cur Avg Loss: 0.19702195, Log Avg loss: 0.21441569, Global Avg Loss: 0.71875458, Time: 0.0210 Steps: 87190, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002040, Sample Num: 32640, Cur Loss: 0.11761344, Cur Avg Loss: 0.19701850, Log Avg loss: 0.19631781, Global Avg Loss: 0.71869466, Time: 0.0210 Steps: 87200, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002050, Sample Num: 32800, Cur Loss: 0.01054783, Cur Avg Loss: 0.19690429, Log Avg loss: 0.17360613, Global Avg Loss: 0.71863216, Time: 0.0249 Steps: 87210, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002060, Sample Num: 32960, Cur Loss: 0.05276916, Cur Avg Loss: 0.19703985, Log Avg loss: 0.22482950, Global Avg Loss: 0.71857555, Time: 0.0210 Steps: 87220, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002070, Sample Num: 33120, Cur Loss: 0.16913682, Cur Avg Loss: 0.19713826, Log Avg loss: 0.21740968, Global Avg Loss: 0.71851809, Time: 0.0211 Steps: 87230, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002080, Sample Num: 33280, Cur Loss: 0.30043098, Cur Avg Loss: 0.19694574, Log Avg loss: 0.15709419, Global Avg Loss: 0.71845374, Time: 0.0210 Steps: 87240, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002090, Sample Num: 33440, Cur Loss: 0.16420826, Cur Avg Loss: 0.19662823, Log Avg loss: 0.13058590, Global Avg Loss: 0.71838636, Time: 0.0210 Steps: 87250, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002100, Sample Num: 33600, Cur Loss: 0.09896864, Cur Avg Loss: 0.19619873, Log Avg loss: 0.10643492, Global Avg Loss: 0.71831623, Time: 0.0211 Steps: 87260, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002110, Sample Num: 33760, Cur Loss: 0.76412475, Cur Avg Loss: 0.19649563, Log Avg loss: 0.25884280, Global Avg Loss: 0.71826358, Time: 0.0210 Steps: 87270, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002120, Sample Num: 33920, Cur Loss: 0.17719619, Cur Avg Loss: 0.19671739, Log Avg loss: 0.24350949, Global Avg Loss: 0.71820919, Time: 0.0210 Steps: 87280, Updated lr: 0.000018 ***** Running evaluation checkpoint-87289 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-87289 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.705732, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.235713, "eval_total_loss": 165.70606, "eval_mae": 0.344447, "eval_mse": 0.235757, "eval_r2": 0.850137, "eval_sp_statistic": 0.901912, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924169, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.342279, "test_total_loss": 171.823995, "test_mae": 0.35392, "test_mse": 0.342402, "test_r2": 0.779011, "test_sp_statistic": 0.874223, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921328, "test_ps_pvalue": 0.0, "lr": 1.8170697012802277e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7181480068385747, "train_cur_epoch_loss": 418.1644130963832, "train_cur_epoch_avg_loss": 0.19641353362911376, "train_cur_epoch_time": 44.70573163032532, "train_cur_epoch_avg_time": 0.02099846483340785, "epoch": 41, "step": 87289} ################################################## Training, Epoch: 0042, Batch: 000001, Sample Num: 16, Cur Loss: 0.11668065, Cur Avg Loss: 0.11668065, Log Avg loss: 0.12402294, Global Avg Loss: 0.71814112, Time: 0.0250 Steps: 87290, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000011, Sample Num: 176, Cur Loss: 0.06999655, Cur Avg Loss: 0.17973899, Log Avg loss: 0.18604482, Global Avg Loss: 0.71808017, Time: 0.0211 Steps: 87300, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000021, Sample Num: 336, Cur Loss: 0.14055729, Cur Avg Loss: 0.15655631, Log Avg loss: 0.13105536, Global Avg Loss: 0.71801293, Time: 0.0212 Steps: 87310, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000031, Sample Num: 496, Cur Loss: 0.27050734, Cur Avg Loss: 0.16614409, Log Avg loss: 0.18627842, Global Avg Loss: 0.71795204, Time: 0.0211 Steps: 87320, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000041, Sample Num: 656, Cur Loss: 0.17715807, Cur Avg Loss: 0.16665867, Log Avg loss: 0.16825389, Global Avg Loss: 0.71788909, Time: 0.0212 Steps: 87330, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000051, Sample Num: 816, Cur Loss: 0.14851063, Cur Avg Loss: 0.17333711, Log Avg loss: 0.20071870, Global Avg Loss: 0.71782988, Time: 0.0212 Steps: 87340, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000061, Sample Num: 976, Cur Loss: 0.26045245, Cur Avg Loss: 0.17831791, Log Avg loss: 0.20371996, Global Avg Loss: 0.71777102, Time: 0.0210 Steps: 87350, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000071, Sample Num: 1136, Cur Loss: 0.26557201, Cur Avg Loss: 0.18336798, Log Avg loss: 0.21417342, Global Avg Loss: 0.71771338, Time: 0.0210 Steps: 87360, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000081, Sample Num: 1296, Cur Loss: 0.12803030, Cur Avg Loss: 0.18205374, Log Avg loss: 0.17272266, Global Avg Loss: 0.71765100, Time: 0.0210 Steps: 87370, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000091, Sample Num: 1456, Cur Loss: 0.19563898, Cur Avg Loss: 0.18892637, Log Avg loss: 0.24459470, Global Avg Loss: 0.71759686, Time: 0.0211 Steps: 87380, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000101, Sample Num: 1616, Cur Loss: 0.09084336, Cur Avg Loss: 0.18523144, Log Avg loss: 0.15160757, Global Avg Loss: 0.71753209, Time: 0.0212 Steps: 87390, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000111, Sample Num: 1776, Cur Loss: 0.41012123, Cur Avg Loss: 0.18330072, Log Avg loss: 0.16380042, Global Avg Loss: 0.71746874, Time: 0.0210 Steps: 87400, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000121, Sample Num: 1936, Cur Loss: 0.22282225, Cur Avg Loss: 0.18413642, Log Avg loss: 0.19341269, Global Avg Loss: 0.71740878, Time: 0.0211 Steps: 87410, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000131, Sample Num: 2096, Cur Loss: 0.11174873, Cur Avg Loss: 0.18994527, Log Avg loss: 0.26023232, Global Avg Loss: 0.71735649, Time: 0.0211 Steps: 87420, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000141, Sample Num: 2256, Cur Loss: 0.09843792, Cur Avg Loss: 0.18678450, Log Avg loss: 0.14537845, Global Avg Loss: 0.71729107, Time: 0.0211 Steps: 87430, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000151, Sample Num: 2416, Cur Loss: 0.11288267, Cur Avg Loss: 0.18520411, Log Avg loss: 0.16292059, Global Avg Loss: 0.71722767, Time: 0.0210 Steps: 87440, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000161, Sample Num: 2576, Cur Loss: 0.10987377, Cur Avg Loss: 0.18538132, Log Avg loss: 0.18805722, Global Avg Loss: 0.71716716, Time: 0.0212 Steps: 87450, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000171, Sample Num: 2736, Cur Loss: 0.11555305, Cur Avg Loss: 0.18362776, Log Avg loss: 0.15539548, Global Avg Loss: 0.71710292, Time: 0.0211 Steps: 87460, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000181, Sample Num: 2896, Cur Loss: 0.20508465, Cur Avg Loss: 0.18188268, Log Avg loss: 0.15204180, Global Avg Loss: 0.71703832, Time: 0.0211 Steps: 87470, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000191, Sample Num: 3056, Cur Loss: 0.27603489, Cur Avg Loss: 0.18024870, Log Avg loss: 0.15067356, Global Avg Loss: 0.71697358, Time: 0.0211 Steps: 87480, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000201, Sample Num: 3216, Cur Loss: 0.28153861, Cur Avg Loss: 0.18403252, Log Avg loss: 0.25630352, Global Avg Loss: 0.71692093, Time: 0.0211 Steps: 87490, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000211, Sample Num: 3376, Cur Loss: 0.09144908, Cur Avg Loss: 0.19016493, Log Avg loss: 0.31342634, Global Avg Loss: 0.71687481, Time: 0.0211 Steps: 87500, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000221, Sample Num: 3536, Cur Loss: 0.14167081, Cur Avg Loss: 0.18879333, Log Avg loss: 0.15985255, Global Avg Loss: 0.71681116, Time: 0.0212 Steps: 87510, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000231, Sample Num: 3696, Cur Loss: 0.07874312, Cur Avg Loss: 0.19053027, Log Avg loss: 0.22891664, Global Avg Loss: 0.71675541, Time: 0.0212 Steps: 87520, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000241, Sample Num: 3856, Cur Loss: 0.44943410, Cur Avg Loss: 0.19083931, Log Avg loss: 0.19797830, Global Avg Loss: 0.71669615, Time: 0.0211 Steps: 87530, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000251, Sample Num: 4016, Cur Loss: 0.10565542, Cur Avg Loss: 0.18748689, Log Avg loss: 0.10669350, Global Avg Loss: 0.71662646, Time: 0.0212 Steps: 87540, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000261, Sample Num: 4176, Cur Loss: 0.05560119, Cur Avg Loss: 0.18585525, Log Avg loss: 0.14490098, Global Avg Loss: 0.71656116, Time: 0.0248 Steps: 87550, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000271, Sample Num: 4336, Cur Loss: 0.40453273, Cur Avg Loss: 0.18680754, Log Avg loss: 0.21166229, Global Avg Loss: 0.71650350, Time: 0.0210 Steps: 87560, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000281, Sample Num: 4496, Cur Loss: 0.06397706, Cur Avg Loss: 0.18537037, Log Avg loss: 0.14642317, Global Avg Loss: 0.71643840, Time: 0.0210 Steps: 87570, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000291, Sample Num: 4656, Cur Loss: 0.20845650, Cur Avg Loss: 0.18542359, Log Avg loss: 0.18691905, Global Avg Loss: 0.71637794, Time: 0.0210 Steps: 87580, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000301, Sample Num: 4816, Cur Loss: 0.08747968, Cur Avg Loss: 0.18479748, Log Avg loss: 0.16657769, Global Avg Loss: 0.71631517, Time: 0.0210 Steps: 87590, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000311, Sample Num: 4976, Cur Loss: 0.08557917, Cur Avg Loss: 0.18334471, Log Avg loss: 0.13961621, Global Avg Loss: 0.71624933, Time: 0.0210 Steps: 87600, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000321, Sample Num: 5136, Cur Loss: 0.26331055, Cur Avg Loss: 0.18284733, Log Avg loss: 0.16737908, Global Avg Loss: 0.71618668, Time: 0.0210 Steps: 87610, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000331, Sample Num: 5296, Cur Loss: 0.09865599, Cur Avg Loss: 0.18228104, Log Avg loss: 0.16410296, Global Avg Loss: 0.71612367, Time: 0.0210 Steps: 87620, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000341, Sample Num: 5456, Cur Loss: 0.30995011, Cur Avg Loss: 0.18252996, Log Avg loss: 0.19076911, Global Avg Loss: 0.71606372, Time: 0.0210 Steps: 87630, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000351, Sample Num: 5616, Cur Loss: 0.09337996, Cur Avg Loss: 0.18238192, Log Avg loss: 0.17733395, Global Avg Loss: 0.71600225, Time: 0.0210 Steps: 87640, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000361, Sample Num: 5776, Cur Loss: 0.23932770, Cur Avg Loss: 0.18212057, Log Avg loss: 0.17294724, Global Avg Loss: 0.71594030, Time: 0.0211 Steps: 87650, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000371, Sample Num: 5936, Cur Loss: 0.20530587, Cur Avg Loss: 0.18430635, Log Avg loss: 0.26321287, Global Avg Loss: 0.71588865, Time: 0.0210 Steps: 87660, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000381, Sample Num: 6096, Cur Loss: 0.34868008, Cur Avg Loss: 0.18532046, Log Avg loss: 0.22294393, Global Avg Loss: 0.71583242, Time: 0.0211 Steps: 87670, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000391, Sample Num: 6256, Cur Loss: 0.40728140, Cur Avg Loss: 0.18603798, Log Avg loss: 0.21337543, Global Avg Loss: 0.71577512, Time: 0.0210 Steps: 87680, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000401, Sample Num: 6416, Cur Loss: 0.58657902, Cur Avg Loss: 0.18707784, Log Avg loss: 0.22773653, Global Avg Loss: 0.71571946, Time: 0.0211 Steps: 87690, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000411, Sample Num: 6576, Cur Loss: 0.06519061, Cur Avg Loss: 0.18685505, Log Avg loss: 0.17792106, Global Avg Loss: 0.71565814, Time: 0.0211 Steps: 87700, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000421, Sample Num: 6736, Cur Loss: 0.04013785, Cur Avg Loss: 0.18566579, Log Avg loss: 0.13678725, Global Avg Loss: 0.71559214, Time: 0.0211 Steps: 87710, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000431, Sample Num: 6896, Cur Loss: 0.15650636, Cur Avg Loss: 0.18534211, Log Avg loss: 0.17171504, Global Avg Loss: 0.71553014, Time: 0.0211 Steps: 87720, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000441, Sample Num: 7056, Cur Loss: 0.05570531, Cur Avg Loss: 0.18489521, Log Avg loss: 0.16563398, Global Avg Loss: 0.71546746, Time: 0.0210 Steps: 87730, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000451, Sample Num: 7216, Cur Loss: 0.07260431, Cur Avg Loss: 0.18568904, Log Avg loss: 0.22069708, Global Avg Loss: 0.71541107, Time: 0.0210 Steps: 87740, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000461, Sample Num: 7376, Cur Loss: 0.09094651, Cur Avg Loss: 0.18508259, Log Avg loss: 0.15773164, Global Avg Loss: 0.71534752, Time: 0.0211 Steps: 87750, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000471, Sample Num: 7536, Cur Loss: 0.07372575, Cur Avg Loss: 0.18514282, Log Avg loss: 0.18791932, Global Avg Loss: 0.71528742, Time: 0.0211 Steps: 87760, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000481, Sample Num: 7696, Cur Loss: 0.27402127, Cur Avg Loss: 0.18700466, Log Avg loss: 0.27469733, Global Avg Loss: 0.71523722, Time: 0.0210 Steps: 87770, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000491, Sample Num: 7856, Cur Loss: 0.09931701, Cur Avg Loss: 0.18551856, Log Avg loss: 0.11403707, Global Avg Loss: 0.71516873, Time: 0.0211 Steps: 87780, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000501, Sample Num: 8016, Cur Loss: 0.23681171, Cur Avg Loss: 0.18561899, Log Avg loss: 0.19055035, Global Avg Loss: 0.71510897, Time: 0.0211 Steps: 87790, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000511, Sample Num: 8176, Cur Loss: 0.24078023, Cur Avg Loss: 0.18603579, Log Avg loss: 0.20691719, Global Avg Loss: 0.71505109, Time: 0.0211 Steps: 87800, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000521, Sample Num: 8336, Cur Loss: 0.07200744, Cur Avg Loss: 0.18633191, Log Avg loss: 0.20146398, Global Avg Loss: 0.71499260, Time: 0.0211 Steps: 87810, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000531, Sample Num: 8496, Cur Loss: 0.16714665, Cur Avg Loss: 0.18549869, Log Avg loss: 0.14208760, Global Avg Loss: 0.71492736, Time: 0.0210 Steps: 87820, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000541, Sample Num: 8656, Cur Loss: 0.30707002, Cur Avg Loss: 0.18522819, Log Avg loss: 0.17086455, Global Avg Loss: 0.71486542, Time: 0.0210 Steps: 87830, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000551, Sample Num: 8816, Cur Loss: 0.24562307, Cur Avg Loss: 0.18434600, Log Avg loss: 0.13661969, Global Avg Loss: 0.71479959, Time: 0.0211 Steps: 87840, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000561, Sample Num: 8976, Cur Loss: 0.07285292, Cur Avg Loss: 0.18419435, Log Avg loss: 0.17583822, Global Avg Loss: 0.71473824, Time: 0.0210 Steps: 87850, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000571, Sample Num: 9136, Cur Loss: 0.28363943, Cur Avg Loss: 0.18482192, Log Avg loss: 0.22002882, Global Avg Loss: 0.71468193, Time: 0.0211 Steps: 87860, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000581, Sample Num: 9296, Cur Loss: 0.33458444, Cur Avg Loss: 0.18497582, Log Avg loss: 0.19376343, Global Avg Loss: 0.71462265, Time: 0.0211 Steps: 87870, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000591, Sample Num: 9456, Cur Loss: 0.08022673, Cur Avg Loss: 0.18487641, Log Avg loss: 0.17910090, Global Avg Loss: 0.71456171, Time: 0.0210 Steps: 87880, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000601, Sample Num: 9616, Cur Loss: 0.27287352, Cur Avg Loss: 0.18526303, Log Avg loss: 0.20811225, Global Avg Loss: 0.71450409, Time: 0.0210 Steps: 87890, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000611, Sample Num: 9776, Cur Loss: 0.10828094, Cur Avg Loss: 0.18532444, Log Avg loss: 0.18901512, Global Avg Loss: 0.71444431, Time: 0.0211 Steps: 87900, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000621, Sample Num: 9936, Cur Loss: 0.09723882, Cur Avg Loss: 0.18545822, Log Avg loss: 0.19363230, Global Avg Loss: 0.71438506, Time: 0.0211 Steps: 87910, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000631, Sample Num: 10096, Cur Loss: 0.24019498, Cur Avg Loss: 0.18552244, Log Avg loss: 0.18951019, Global Avg Loss: 0.71432536, Time: 0.0211 Steps: 87920, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000641, Sample Num: 10256, Cur Loss: 0.27270612, Cur Avg Loss: 0.18570845, Log Avg loss: 0.19744576, Global Avg Loss: 0.71426658, Time: 0.0210 Steps: 87930, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000651, Sample Num: 10416, Cur Loss: 0.15487787, Cur Avg Loss: 0.18589285, Log Avg loss: 0.19771288, Global Avg Loss: 0.71420784, Time: 0.0210 Steps: 87940, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000661, Sample Num: 10576, Cur Loss: 0.25036904, Cur Avg Loss: 0.18591376, Log Avg loss: 0.18727522, Global Avg Loss: 0.71414793, Time: 0.0211 Steps: 87950, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000671, Sample Num: 10736, Cur Loss: 0.03802904, Cur Avg Loss: 0.18543401, Log Avg loss: 0.15372229, Global Avg Loss: 0.71408422, Time: 0.0211 Steps: 87960, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000681, Sample Num: 10896, Cur Loss: 0.09220012, Cur Avg Loss: 0.18664717, Log Avg loss: 0.26805023, Global Avg Loss: 0.71403351, Time: 0.0211 Steps: 87970, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000691, Sample Num: 11056, Cur Loss: 0.11864305, Cur Avg Loss: 0.18611934, Log Avg loss: 0.15017445, Global Avg Loss: 0.71396942, Time: 0.0211 Steps: 87980, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000701, Sample Num: 11216, Cur Loss: 0.17595714, Cur Avg Loss: 0.18596703, Log Avg loss: 0.17544236, Global Avg Loss: 0.71390822, Time: 0.0211 Steps: 87990, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000711, Sample Num: 11376, Cur Loss: 0.19387463, Cur Avg Loss: 0.18598543, Log Avg loss: 0.18727512, Global Avg Loss: 0.71384838, Time: 0.0211 Steps: 88000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000721, Sample Num: 11536, Cur Loss: 0.27194637, Cur Avg Loss: 0.18613874, Log Avg loss: 0.19703905, Global Avg Loss: 0.71378965, Time: 0.0211 Steps: 88010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000731, Sample Num: 11696, Cur Loss: 0.40738669, Cur Avg Loss: 0.18609291, Log Avg loss: 0.18278861, Global Avg Loss: 0.71372933, Time: 0.0211 Steps: 88020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000741, Sample Num: 11856, Cur Loss: 0.23706475, Cur Avg Loss: 0.18615720, Log Avg loss: 0.19085663, Global Avg Loss: 0.71366993, Time: 0.0211 Steps: 88030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000751, Sample Num: 12016, Cur Loss: 0.29594773, Cur Avg Loss: 0.18696620, Log Avg loss: 0.24691297, Global Avg Loss: 0.71361691, Time: 0.0211 Steps: 88040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000761, Sample Num: 12176, Cur Loss: 0.26634151, Cur Avg Loss: 0.18679827, Log Avg loss: 0.17418660, Global Avg Loss: 0.71355565, Time: 0.0211 Steps: 88050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000771, Sample Num: 12336, Cur Loss: 0.08132826, Cur Avg Loss: 0.18665373, Log Avg loss: 0.17565428, Global Avg Loss: 0.71349456, Time: 0.0248 Steps: 88060, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000781, Sample Num: 12496, Cur Loss: 0.14609198, Cur Avg Loss: 0.18641487, Log Avg loss: 0.16799867, Global Avg Loss: 0.71343263, Time: 0.0211 Steps: 88070, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000791, Sample Num: 12656, Cur Loss: 0.16606596, Cur Avg Loss: 0.18629902, Log Avg loss: 0.17725153, Global Avg Loss: 0.71337175, Time: 0.0211 Steps: 88080, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000801, Sample Num: 12816, Cur Loss: 0.15410627, Cur Avg Loss: 0.18633674, Log Avg loss: 0.18932074, Global Avg Loss: 0.71331226, Time: 0.0211 Steps: 88090, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000811, Sample Num: 12976, Cur Loss: 0.11620098, Cur Avg Loss: 0.18607471, Log Avg loss: 0.16508593, Global Avg Loss: 0.71325003, Time: 0.0210 Steps: 88100, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000821, Sample Num: 13136, Cur Loss: 0.13988751, Cur Avg Loss: 0.18523074, Log Avg loss: 0.11678457, Global Avg Loss: 0.71318234, Time: 0.0211 Steps: 88110, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000831, Sample Num: 13296, Cur Loss: 0.08570015, Cur Avg Loss: 0.18510323, Log Avg loss: 0.17463490, Global Avg Loss: 0.71312122, Time: 0.0211 Steps: 88120, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000841, Sample Num: 13456, Cur Loss: 0.10373728, Cur Avg Loss: 0.18526659, Log Avg loss: 0.19884131, Global Avg Loss: 0.71306287, Time: 0.0211 Steps: 88130, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000851, Sample Num: 13616, Cur Loss: 0.07935688, Cur Avg Loss: 0.18511394, Log Avg loss: 0.17227596, Global Avg Loss: 0.71300151, Time: 0.0211 Steps: 88140, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000861, Sample Num: 13776, Cur Loss: 0.04293370, Cur Avg Loss: 0.18449570, Log Avg loss: 0.13188366, Global Avg Loss: 0.71293559, Time: 0.0211 Steps: 88150, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000871, Sample Num: 13936, Cur Loss: 0.06770056, Cur Avg Loss: 0.18389419, Log Avg loss: 0.13210404, Global Avg Loss: 0.71286971, Time: 0.0211 Steps: 88160, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000881, Sample Num: 14096, Cur Loss: 0.21521804, Cur Avg Loss: 0.18371308, Log Avg loss: 0.16793842, Global Avg Loss: 0.71280790, Time: 0.0211 Steps: 88170, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000891, Sample Num: 14256, Cur Loss: 0.52206612, Cur Avg Loss: 0.18440678, Log Avg loss: 0.24552212, Global Avg Loss: 0.71275491, Time: 0.0211 Steps: 88180, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000901, Sample Num: 14416, Cur Loss: 0.09363510, Cur Avg Loss: 0.18398047, Log Avg loss: 0.14599583, Global Avg Loss: 0.71269064, Time: 0.0211 Steps: 88190, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000911, Sample Num: 14576, Cur Loss: 0.17090321, Cur Avg Loss: 0.18469407, Log Avg loss: 0.24898977, Global Avg Loss: 0.71263807, Time: 0.0211 Steps: 88200, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000921, Sample Num: 14736, Cur Loss: 0.06910437, Cur Avg Loss: 0.18437417, Log Avg loss: 0.15523090, Global Avg Loss: 0.71257488, Time: 0.0211 Steps: 88210, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000931, Sample Num: 14896, Cur Loss: 0.09174959, Cur Avg Loss: 0.18438223, Log Avg loss: 0.18512479, Global Avg Loss: 0.71251509, Time: 0.0211 Steps: 88220, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000941, Sample Num: 15056, Cur Loss: 0.17715880, Cur Avg Loss: 0.18433609, Log Avg loss: 0.18004068, Global Avg Loss: 0.71245474, Time: 0.0211 Steps: 88230, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000951, Sample Num: 15216, Cur Loss: 0.24781062, Cur Avg Loss: 0.18480924, Log Avg loss: 0.22933265, Global Avg Loss: 0.71239999, Time: 0.0211 Steps: 88240, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000961, Sample Num: 15376, Cur Loss: 0.27988794, Cur Avg Loss: 0.18526729, Log Avg loss: 0.22882787, Global Avg Loss: 0.71234519, Time: 0.0211 Steps: 88250, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000971, Sample Num: 15536, Cur Loss: 0.32541648, Cur Avg Loss: 0.18518966, Log Avg loss: 0.17772896, Global Avg Loss: 0.71228462, Time: 0.0211 Steps: 88260, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000981, Sample Num: 15696, Cur Loss: 0.43247178, Cur Avg Loss: 0.18634574, Log Avg loss: 0.29860139, Global Avg Loss: 0.71223775, Time: 0.0211 Steps: 88270, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000991, Sample Num: 15856, Cur Loss: 0.15037383, Cur Avg Loss: 0.18637896, Log Avg loss: 0.18963747, Global Avg Loss: 0.71217856, Time: 0.0211 Steps: 88280, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001001, Sample Num: 16016, Cur Loss: 0.06539698, Cur Avg Loss: 0.18629524, Log Avg loss: 0.17799903, Global Avg Loss: 0.71211805, Time: 0.0211 Steps: 88290, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001011, Sample Num: 16176, Cur Loss: 0.11598079, Cur Avg Loss: 0.18655821, Log Avg loss: 0.21288116, Global Avg Loss: 0.71206151, Time: 0.0211 Steps: 88300, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001021, Sample Num: 16336, Cur Loss: 0.12755913, Cur Avg Loss: 0.18653879, Log Avg loss: 0.18457604, Global Avg Loss: 0.71200178, Time: 0.0211 Steps: 88310, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001031, Sample Num: 16496, Cur Loss: 0.16529478, Cur Avg Loss: 0.18686819, Log Avg loss: 0.22049948, Global Avg Loss: 0.71194613, Time: 0.0212 Steps: 88320, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001041, Sample Num: 16656, Cur Loss: 0.44016930, Cur Avg Loss: 0.18751483, Log Avg loss: 0.25418380, Global Avg Loss: 0.71189431, Time: 0.0211 Steps: 88330, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001051, Sample Num: 16816, Cur Loss: 0.28420326, Cur Avg Loss: 0.18790009, Log Avg loss: 0.22800525, Global Avg Loss: 0.71183953, Time: 0.0211 Steps: 88340, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001061, Sample Num: 16976, Cur Loss: 0.37599981, Cur Avg Loss: 0.18752775, Log Avg loss: 0.14839536, Global Avg Loss: 0.71177576, Time: 0.0211 Steps: 88350, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001071, Sample Num: 17136, Cur Loss: 0.21514110, Cur Avg Loss: 0.18691271, Log Avg loss: 0.12165639, Global Avg Loss: 0.71170897, Time: 0.0210 Steps: 88360, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001081, Sample Num: 17296, Cur Loss: 0.05569657, Cur Avg Loss: 0.18621567, Log Avg loss: 0.11156263, Global Avg Loss: 0.71164106, Time: 0.0211 Steps: 88370, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001091, Sample Num: 17456, Cur Loss: 0.17461737, Cur Avg Loss: 0.18609137, Log Avg loss: 0.17265507, Global Avg Loss: 0.71158008, Time: 0.0212 Steps: 88380, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001101, Sample Num: 17616, Cur Loss: 0.16561908, Cur Avg Loss: 0.18604062, Log Avg loss: 0.18050378, Global Avg Loss: 0.71151999, Time: 0.0211 Steps: 88390, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001111, Sample Num: 17776, Cur Loss: 0.21217728, Cur Avg Loss: 0.18611642, Log Avg loss: 0.19446238, Global Avg Loss: 0.71146150, Time: 0.0211 Steps: 88400, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001121, Sample Num: 17936, Cur Loss: 0.09764360, Cur Avg Loss: 0.18631125, Log Avg loss: 0.20795685, Global Avg Loss: 0.71140455, Time: 0.0211 Steps: 88410, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001131, Sample Num: 18096, Cur Loss: 0.14108790, Cur Avg Loss: 0.18602118, Log Avg loss: 0.15350345, Global Avg Loss: 0.71134145, Time: 0.0211 Steps: 88420, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001141, Sample Num: 18256, Cur Loss: 0.19019228, Cur Avg Loss: 0.18600892, Log Avg loss: 0.18462304, Global Avg Loss: 0.71128189, Time: 0.0211 Steps: 88430, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001151, Sample Num: 18416, Cur Loss: 0.08179058, Cur Avg Loss: 0.18608754, Log Avg loss: 0.19505719, Global Avg Loss: 0.71122352, Time: 0.0211 Steps: 88440, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001161, Sample Num: 18576, Cur Loss: 0.09221542, Cur Avg Loss: 0.18591784, Log Avg loss: 0.16638590, Global Avg Loss: 0.71116192, Time: 0.0211 Steps: 88450, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001171, Sample Num: 18736, Cur Loss: 0.27500904, Cur Avg Loss: 0.18583483, Log Avg loss: 0.17619697, Global Avg Loss: 0.71110145, Time: 0.0211 Steps: 88460, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001181, Sample Num: 18896, Cur Loss: 0.17447408, Cur Avg Loss: 0.18614446, Log Avg loss: 0.22240303, Global Avg Loss: 0.71104621, Time: 0.0210 Steps: 88470, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001191, Sample Num: 19056, Cur Loss: 0.20111099, Cur Avg Loss: 0.18591013, Log Avg loss: 0.15823564, Global Avg Loss: 0.71098373, Time: 0.0210 Steps: 88480, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001201, Sample Num: 19216, Cur Loss: 0.18763819, Cur Avg Loss: 0.18754584, Log Avg loss: 0.38235884, Global Avg Loss: 0.71094659, Time: 0.0209 Steps: 88490, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001211, Sample Num: 19376, Cur Loss: 0.12441436, Cur Avg Loss: 0.18806214, Log Avg loss: 0.25006936, Global Avg Loss: 0.71089452, Time: 0.0210 Steps: 88500, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001221, Sample Num: 19536, Cur Loss: 0.29181868, Cur Avg Loss: 0.18802553, Log Avg loss: 0.18359256, Global Avg Loss: 0.71083494, Time: 0.0210 Steps: 88510, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001231, Sample Num: 19696, Cur Loss: 0.14921373, Cur Avg Loss: 0.18853454, Log Avg loss: 0.25068422, Global Avg Loss: 0.71078296, Time: 0.0210 Steps: 88520, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001241, Sample Num: 19856, Cur Loss: 0.16937286, Cur Avg Loss: 0.18866160, Log Avg loss: 0.20430285, Global Avg Loss: 0.71072575, Time: 0.0210 Steps: 88530, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001251, Sample Num: 20016, Cur Loss: 0.13742678, Cur Avg Loss: 0.18833483, Log Avg loss: 0.14778271, Global Avg Loss: 0.71066217, Time: 0.0210 Steps: 88540, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001261, Sample Num: 20176, Cur Loss: 0.10252422, Cur Avg Loss: 0.18797266, Log Avg loss: 0.14266476, Global Avg Loss: 0.71059802, Time: 0.0210 Steps: 88550, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001271, Sample Num: 20336, Cur Loss: 0.15928894, Cur Avg Loss: 0.18828574, Log Avg loss: 0.22776582, Global Avg Loss: 0.71054350, Time: 0.0210 Steps: 88560, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001281, Sample Num: 20496, Cur Loss: 0.07289791, Cur Avg Loss: 0.18796860, Log Avg loss: 0.14765997, Global Avg Loss: 0.71047995, Time: 0.0248 Steps: 88570, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001291, Sample Num: 20656, Cur Loss: 0.03273842, Cur Avg Loss: 0.18795414, Log Avg loss: 0.18610157, Global Avg Loss: 0.71042075, Time: 0.0209 Steps: 88580, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001301, Sample Num: 20816, Cur Loss: 0.13945997, Cur Avg Loss: 0.18779657, Log Avg loss: 0.16745451, Global Avg Loss: 0.71035946, Time: 0.0209 Steps: 88590, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001311, Sample Num: 20976, Cur Loss: 0.25217271, Cur Avg Loss: 0.18795442, Log Avg loss: 0.20849032, Global Avg Loss: 0.71030282, Time: 0.0209 Steps: 88600, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001321, Sample Num: 21136, Cur Loss: 0.05748013, Cur Avg Loss: 0.18816272, Log Avg loss: 0.21547113, Global Avg Loss: 0.71024697, Time: 0.0209 Steps: 88610, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001331, Sample Num: 21296, Cur Loss: 0.22641130, Cur Avg Loss: 0.18815514, Log Avg loss: 0.18715381, Global Avg Loss: 0.71018795, Time: 0.0209 Steps: 88620, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001341, Sample Num: 21456, Cur Loss: 0.29902157, Cur Avg Loss: 0.18793726, Log Avg loss: 0.15893684, Global Avg Loss: 0.71012575, Time: 0.0209 Steps: 88630, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001351, Sample Num: 21616, Cur Loss: 0.14982565, Cur Avg Loss: 0.18763578, Log Avg loss: 0.14720821, Global Avg Loss: 0.71006224, Time: 0.0209 Steps: 88640, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001361, Sample Num: 21776, Cur Loss: 0.07450240, Cur Avg Loss: 0.18759578, Log Avg loss: 0.18219135, Global Avg Loss: 0.71000270, Time: 0.0209 Steps: 88650, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001371, Sample Num: 21936, Cur Loss: 0.15058246, Cur Avg Loss: 0.18786550, Log Avg loss: 0.22457368, Global Avg Loss: 0.70994795, Time: 0.0209 Steps: 88660, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001381, Sample Num: 22096, Cur Loss: 0.43327913, Cur Avg Loss: 0.18787406, Log Avg loss: 0.18904805, Global Avg Loss: 0.70988920, Time: 0.0209 Steps: 88670, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001391, Sample Num: 22256, Cur Loss: 0.05943840, Cur Avg Loss: 0.18777625, Log Avg loss: 0.17426929, Global Avg Loss: 0.70982880, Time: 0.0211 Steps: 88680, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001401, Sample Num: 22416, Cur Loss: 0.08849324, Cur Avg Loss: 0.18744457, Log Avg loss: 0.14130814, Global Avg Loss: 0.70976470, Time: 0.0213 Steps: 88690, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001411, Sample Num: 22576, Cur Loss: 0.06875560, Cur Avg Loss: 0.18731528, Log Avg loss: 0.16920115, Global Avg Loss: 0.70970376, Time: 0.0212 Steps: 88700, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001421, Sample Num: 22736, Cur Loss: 0.04677509, Cur Avg Loss: 0.18738918, Log Avg loss: 0.19781633, Global Avg Loss: 0.70964605, Time: 0.0211 Steps: 88710, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001431, Sample Num: 22896, Cur Loss: 0.09934122, Cur Avg Loss: 0.18692512, Log Avg loss: 0.12098261, Global Avg Loss: 0.70957970, Time: 0.0212 Steps: 88720, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001441, Sample Num: 23056, Cur Loss: 0.11512378, Cur Avg Loss: 0.18700369, Log Avg loss: 0.19824657, Global Avg Loss: 0.70952207, Time: 0.0211 Steps: 88730, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001451, Sample Num: 23216, Cur Loss: 0.23508835, Cur Avg Loss: 0.18719137, Log Avg loss: 0.21423614, Global Avg Loss: 0.70946626, Time: 0.0211 Steps: 88740, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001461, Sample Num: 23376, Cur Loss: 0.07644119, Cur Avg Loss: 0.18710463, Log Avg loss: 0.17451871, Global Avg Loss: 0.70940599, Time: 0.0211 Steps: 88750, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001471, Sample Num: 23536, Cur Loss: 0.12939680, Cur Avg Loss: 0.18747874, Log Avg loss: 0.24213674, Global Avg Loss: 0.70935334, Time: 0.0211 Steps: 88760, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001481, Sample Num: 23696, Cur Loss: 0.11315048, Cur Avg Loss: 0.18742374, Log Avg loss: 0.17933285, Global Avg Loss: 0.70929363, Time: 0.0212 Steps: 88770, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001491, Sample Num: 23856, Cur Loss: 0.09857062, Cur Avg Loss: 0.18762709, Log Avg loss: 0.21774362, Global Avg Loss: 0.70923827, Time: 0.0211 Steps: 88780, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001501, Sample Num: 24016, Cur Loss: 0.25619781, Cur Avg Loss: 0.18776484, Log Avg loss: 0.20830283, Global Avg Loss: 0.70918185, Time: 0.0211 Steps: 88790, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001511, Sample Num: 24176, Cur Loss: 0.36309189, Cur Avg Loss: 0.18807293, Log Avg loss: 0.23431717, Global Avg Loss: 0.70912837, Time: 0.0211 Steps: 88800, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001521, Sample Num: 24336, Cur Loss: 0.33796421, Cur Avg Loss: 0.18859009, Log Avg loss: 0.26673262, Global Avg Loss: 0.70907856, Time: 0.0211 Steps: 88810, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001531, Sample Num: 24496, Cur Loss: 0.16968940, Cur Avg Loss: 0.18883313, Log Avg loss: 0.22579959, Global Avg Loss: 0.70902415, Time: 0.0211 Steps: 88820, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001541, Sample Num: 24656, Cur Loss: 0.37298113, Cur Avg Loss: 0.18905573, Log Avg loss: 0.22313686, Global Avg Loss: 0.70896945, Time: 0.0215 Steps: 88830, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001551, Sample Num: 24816, Cur Loss: 0.05155162, Cur Avg Loss: 0.18905621, Log Avg loss: 0.18912914, Global Avg Loss: 0.70891094, Time: 0.0211 Steps: 88840, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001561, Sample Num: 24976, Cur Loss: 0.13375615, Cur Avg Loss: 0.18933736, Log Avg loss: 0.23294371, Global Avg Loss: 0.70885737, Time: 0.0211 Steps: 88850, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001571, Sample Num: 25136, Cur Loss: 0.31227353, Cur Avg Loss: 0.18929089, Log Avg loss: 0.18203670, Global Avg Loss: 0.70879808, Time: 0.0211 Steps: 88860, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001581, Sample Num: 25296, Cur Loss: 0.09563334, Cur Avg Loss: 0.18919113, Log Avg loss: 0.17352017, Global Avg Loss: 0.70873785, Time: 0.0211 Steps: 88870, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001591, Sample Num: 25456, Cur Loss: 0.05996921, Cur Avg Loss: 0.18891217, Log Avg loss: 0.14480742, Global Avg Loss: 0.70867440, Time: 0.0211 Steps: 88880, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001601, Sample Num: 25616, Cur Loss: 0.21904784, Cur Avg Loss: 0.18878465, Log Avg loss: 0.16849720, Global Avg Loss: 0.70861363, Time: 0.0211 Steps: 88890, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001611, Sample Num: 25776, Cur Loss: 0.06217516, Cur Avg Loss: 0.18881933, Log Avg loss: 0.19437181, Global Avg Loss: 0.70855579, Time: 0.0210 Steps: 88900, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001621, Sample Num: 25936, Cur Loss: 0.09339292, Cur Avg Loss: 0.18878568, Log Avg loss: 0.18336377, Global Avg Loss: 0.70849672, Time: 0.0211 Steps: 88910, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001631, Sample Num: 26096, Cur Loss: 0.08525125, Cur Avg Loss: 0.18905411, Log Avg loss: 0.23256704, Global Avg Loss: 0.70844319, Time: 0.0211 Steps: 88920, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001641, Sample Num: 26256, Cur Loss: 0.19542454, Cur Avg Loss: 0.18883641, Log Avg loss: 0.15332999, Global Avg Loss: 0.70838077, Time: 0.0211 Steps: 88930, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001651, Sample Num: 26416, Cur Loss: 0.27714533, Cur Avg Loss: 0.18898207, Log Avg loss: 0.21288360, Global Avg Loss: 0.70832506, Time: 0.0211 Steps: 88940, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001661, Sample Num: 26576, Cur Loss: 0.27865380, Cur Avg Loss: 0.18956448, Log Avg loss: 0.28572103, Global Avg Loss: 0.70827755, Time: 0.0211 Steps: 88950, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001671, Sample Num: 26736, Cur Loss: 0.15190576, Cur Avg Loss: 0.18989141, Log Avg loss: 0.24419422, Global Avg Loss: 0.70822538, Time: 0.0211 Steps: 88960, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001681, Sample Num: 26896, Cur Loss: 0.23775008, Cur Avg Loss: 0.18994873, Log Avg loss: 0.19952634, Global Avg Loss: 0.70816820, Time: 0.0211 Steps: 88970, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001691, Sample Num: 27056, Cur Loss: 0.26833260, Cur Avg Loss: 0.18999386, Log Avg loss: 0.19758137, Global Avg Loss: 0.70811082, Time: 0.0211 Steps: 88980, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001701, Sample Num: 27216, Cur Loss: 0.13609178, Cur Avg Loss: 0.19004427, Log Avg loss: 0.19856756, Global Avg Loss: 0.70805356, Time: 0.0211 Steps: 88990, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001711, Sample Num: 27376, Cur Loss: 0.17616074, Cur Avg Loss: 0.18979739, Log Avg loss: 0.14780410, Global Avg Loss: 0.70799061, Time: 0.0211 Steps: 89000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001721, Sample Num: 27536, Cur Loss: 0.28761613, Cur Avg Loss: 0.18968621, Log Avg loss: 0.17066327, Global Avg Loss: 0.70793025, Time: 0.0211 Steps: 89010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001731, Sample Num: 27696, Cur Loss: 0.27474624, Cur Avg Loss: 0.18949003, Log Avg loss: 0.15572757, Global Avg Loss: 0.70786822, Time: 0.0211 Steps: 89020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001741, Sample Num: 27856, Cur Loss: 0.27651501, Cur Avg Loss: 0.18959258, Log Avg loss: 0.20734347, Global Avg Loss: 0.70781200, Time: 0.0211 Steps: 89030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001751, Sample Num: 28016, Cur Loss: 0.08059026, Cur Avg Loss: 0.18916152, Log Avg loss: 0.11411461, Global Avg Loss: 0.70774532, Time: 0.0211 Steps: 89040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001761, Sample Num: 28176, Cur Loss: 0.12018745, Cur Avg Loss: 0.18942529, Log Avg loss: 0.23561135, Global Avg Loss: 0.70769230, Time: 0.0211 Steps: 89050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001771, Sample Num: 28336, Cur Loss: 0.40792310, Cur Avg Loss: 0.18934978, Log Avg loss: 0.17605203, Global Avg Loss: 0.70763261, Time: 0.0211 Steps: 89060, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001781, Sample Num: 28496, Cur Loss: 0.26533371, Cur Avg Loss: 0.18949874, Log Avg loss: 0.21587978, Global Avg Loss: 0.70757740, Time: 0.0211 Steps: 89070, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001791, Sample Num: 28656, Cur Loss: 0.23347898, Cur Avg Loss: 0.18938707, Log Avg loss: 0.16949805, Global Avg Loss: 0.70751699, Time: 0.0211 Steps: 89080, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001801, Sample Num: 28816, Cur Loss: 0.25897127, Cur Avg Loss: 0.18932326, Log Avg loss: 0.17789568, Global Avg Loss: 0.70745754, Time: 0.0211 Steps: 89090, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001811, Sample Num: 28976, Cur Loss: 0.11439066, Cur Avg Loss: 0.18924625, Log Avg loss: 0.17537693, Global Avg Loss: 0.70739783, Time: 0.0211 Steps: 89100, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001821, Sample Num: 29136, Cur Loss: 0.28292215, Cur Avg Loss: 0.18951593, Log Avg loss: 0.23835409, Global Avg Loss: 0.70734519, Time: 0.0210 Steps: 89110, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001831, Sample Num: 29296, Cur Loss: 0.08622541, Cur Avg Loss: 0.18960553, Log Avg loss: 0.20592163, Global Avg Loss: 0.70728893, Time: 0.0210 Steps: 89120, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001841, Sample Num: 29456, Cur Loss: 0.29214251, Cur Avg Loss: 0.18922590, Log Avg loss: 0.11971603, Global Avg Loss: 0.70722300, Time: 0.0210 Steps: 89130, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001851, Sample Num: 29616, Cur Loss: 0.50779569, Cur Avg Loss: 0.18942955, Log Avg loss: 0.22692157, Global Avg Loss: 0.70716912, Time: 0.0211 Steps: 89140, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001861, Sample Num: 29776, Cur Loss: 0.20984507, Cur Avg Loss: 0.18981065, Log Avg loss: 0.26035200, Global Avg Loss: 0.70711900, Time: 0.0210 Steps: 89150, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001871, Sample Num: 29936, Cur Loss: 0.16654743, Cur Avg Loss: 0.18987341, Log Avg loss: 0.20155302, Global Avg Loss: 0.70706230, Time: 0.0210 Steps: 89160, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001881, Sample Num: 30096, Cur Loss: 0.13708594, Cur Avg Loss: 0.19008079, Log Avg loss: 0.22888139, Global Avg Loss: 0.70700867, Time: 0.0210 Steps: 89170, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001891, Sample Num: 30256, Cur Loss: 0.11377576, Cur Avg Loss: 0.19028669, Log Avg loss: 0.22901615, Global Avg Loss: 0.70695507, Time: 0.0210 Steps: 89180, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001901, Sample Num: 30416, Cur Loss: 0.14543766, Cur Avg Loss: 0.19027185, Log Avg loss: 0.18746694, Global Avg Loss: 0.70689683, Time: 0.0210 Steps: 89190, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001911, Sample Num: 30576, Cur Loss: 0.25827944, Cur Avg Loss: 0.19014103, Log Avg loss: 0.16527183, Global Avg Loss: 0.70683611, Time: 0.0210 Steps: 89200, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001921, Sample Num: 30736, Cur Loss: 0.23398049, Cur Avg Loss: 0.19007489, Log Avg loss: 0.17743591, Global Avg Loss: 0.70677677, Time: 0.0211 Steps: 89210, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001931, Sample Num: 30896, Cur Loss: 0.22721411, Cur Avg Loss: 0.19028676, Log Avg loss: 0.23098608, Global Avg Loss: 0.70672344, Time: 0.0211 Steps: 89220, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001941, Sample Num: 31056, Cur Loss: 0.16809119, Cur Avg Loss: 0.19062478, Log Avg loss: 0.25589599, Global Avg Loss: 0.70667291, Time: 0.0210 Steps: 89230, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001951, Sample Num: 31216, Cur Loss: 0.11966237, Cur Avg Loss: 0.19073164, Log Avg loss: 0.21147448, Global Avg Loss: 0.70661742, Time: 0.0210 Steps: 89240, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001961, Sample Num: 31376, Cur Loss: 0.11563241, Cur Avg Loss: 0.19056945, Log Avg loss: 0.15892625, Global Avg Loss: 0.70655606, Time: 0.0211 Steps: 89250, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001971, Sample Num: 31536, Cur Loss: 0.19176902, Cur Avg Loss: 0.19062124, Log Avg loss: 0.20077570, Global Avg Loss: 0.70649939, Time: 0.0211 Steps: 89260, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001981, Sample Num: 31696, Cur Loss: 0.10826534, Cur Avg Loss: 0.19108765, Log Avg loss: 0.28301718, Global Avg Loss: 0.70645195, Time: 0.0210 Steps: 89270, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001991, Sample Num: 31856, Cur Loss: 0.57385087, Cur Avg Loss: 0.19137246, Log Avg loss: 0.24779453, Global Avg Loss: 0.70640058, Time: 0.0211 Steps: 89280, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002001, Sample Num: 32016, Cur Loss: 0.11752449, Cur Avg Loss: 0.19160885, Log Avg loss: 0.23867331, Global Avg Loss: 0.70634820, Time: 0.0210 Steps: 89290, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002011, Sample Num: 32176, Cur Loss: 0.15483555, Cur Avg Loss: 0.19179816, Log Avg loss: 0.22967975, Global Avg Loss: 0.70629482, Time: 0.0210 Steps: 89300, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002021, Sample Num: 32336, Cur Loss: 0.16788317, Cur Avg Loss: 0.19212356, Log Avg loss: 0.25756111, Global Avg Loss: 0.70624458, Time: 0.0210 Steps: 89310, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002031, Sample Num: 32496, Cur Loss: 0.15668516, Cur Avg Loss: 0.19203315, Log Avg loss: 0.17376208, Global Avg Loss: 0.70618496, Time: 0.0211 Steps: 89320, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002041, Sample Num: 32656, Cur Loss: 0.23020691, Cur Avg Loss: 0.19216952, Log Avg loss: 0.21986443, Global Avg Loss: 0.70613052, Time: 0.0210 Steps: 89330, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002051, Sample Num: 32816, Cur Loss: 0.21138862, Cur Avg Loss: 0.19214712, Log Avg loss: 0.18757710, Global Avg Loss: 0.70607248, Time: 0.0247 Steps: 89340, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002061, Sample Num: 32976, Cur Loss: 0.15810201, Cur Avg Loss: 0.19253340, Log Avg loss: 0.27175944, Global Avg Loss: 0.70602387, Time: 0.0211 Steps: 89350, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002071, Sample Num: 33136, Cur Loss: 0.19926308, Cur Avg Loss: 0.19221961, Log Avg loss: 0.12754740, Global Avg Loss: 0.70595913, Time: 0.0210 Steps: 89360, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002081, Sample Num: 33296, Cur Loss: 0.13767177, Cur Avg Loss: 0.19188043, Log Avg loss: 0.12163617, Global Avg Loss: 0.70589375, Time: 0.0210 Steps: 89370, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002091, Sample Num: 33456, Cur Loss: 0.16376758, Cur Avg Loss: 0.19177035, Log Avg loss: 0.16886116, Global Avg Loss: 0.70583367, Time: 0.0210 Steps: 89380, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002101, Sample Num: 33616, Cur Loss: 0.23676747, Cur Avg Loss: 0.19184104, Log Avg loss: 0.20662228, Global Avg Loss: 0.70577782, Time: 0.0210 Steps: 89390, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002111, Sample Num: 33776, Cur Loss: 0.14713907, Cur Avg Loss: 0.19166787, Log Avg loss: 0.15528507, Global Avg Loss: 0.70571624, Time: 0.0210 Steps: 89400, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002121, Sample Num: 33936, Cur Loss: 0.14052540, Cur Avg Loss: 0.19177620, Log Avg loss: 0.21464506, Global Avg Loss: 0.70566132, Time: 0.0210 Steps: 89410, Updated lr: 0.000016 ***** Running evaluation checkpoint-89418 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-89418 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.080867, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.390175, "eval_total_loss": 274.293055, "eval_mae": 0.524071, "eval_mse": 0.39016, "eval_r2": 0.751989, "eval_sp_statistic": 0.899127, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925048, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.280922, "test_total_loss": 141.022935, "test_mae": 0.379788, "test_mse": 0.281012, "test_r2": 0.818632, "test_sp_statistic": 0.876424, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.924381, "test_ps_pvalue": 0.0, "lr": 1.6151730678046468e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7056130995246868, "train_cur_epoch_loss": 408.0907643660903, "train_cur_epoch_avg_loss": 0.19168189965527963, "train_cur_epoch_time": 45.08086705207825, "train_cur_epoch_avg_time": 0.021174667473968175, "epoch": 42, "step": 89418} ################################################## Training, Epoch: 0043, Batch: 000002, Sample Num: 32, Cur Loss: 0.13346207, Cur Avg Loss: 0.08613173, Log Avg loss: 0.15057105, Global Avg Loss: 0.70559924, Time: 0.0249 Steps: 89420, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000012, Sample Num: 192, Cur Loss: 0.58190686, Cur Avg Loss: 0.21354492, Log Avg loss: 0.23902755, Global Avg Loss: 0.70554707, Time: 0.0211 Steps: 89430, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000022, Sample Num: 352, Cur Loss: 0.15777437, Cur Avg Loss: 0.19682653, Log Avg loss: 0.17676446, Global Avg Loss: 0.70548795, Time: 0.0211 Steps: 89440, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000032, Sample Num: 512, Cur Loss: 0.41335058, Cur Avg Loss: 0.19583980, Log Avg loss: 0.19366900, Global Avg Loss: 0.70543073, Time: 0.0211 Steps: 89450, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000042, Sample Num: 672, Cur Loss: 0.12407660, Cur Avg Loss: 0.18470946, Log Avg loss: 0.14909238, Global Avg Loss: 0.70536854, Time: 0.0211 Steps: 89460, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000052, Sample Num: 832, Cur Loss: 0.07972997, Cur Avg Loss: 0.17276197, Log Avg loss: 0.12258249, Global Avg Loss: 0.70530341, Time: 0.0211 Steps: 89470, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000062, Sample Num: 992, Cur Loss: 0.19915886, Cur Avg Loss: 0.17667535, Log Avg loss: 0.19702494, Global Avg Loss: 0.70524660, Time: 0.0211 Steps: 89480, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000072, Sample Num: 1152, Cur Loss: 0.15055257, Cur Avg Loss: 0.18381833, Log Avg loss: 0.22810478, Global Avg Loss: 0.70519328, Time: 0.0211 Steps: 89490, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000082, Sample Num: 1312, Cur Loss: 0.28577805, Cur Avg Loss: 0.19085036, Log Avg loss: 0.24148097, Global Avg Loss: 0.70514147, Time: 0.0211 Steps: 89500, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000092, Sample Num: 1472, Cur Loss: 0.40600431, Cur Avg Loss: 0.18992173, Log Avg loss: 0.18230702, Global Avg Loss: 0.70508306, Time: 0.0213 Steps: 89510, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000102, Sample Num: 1632, Cur Loss: 0.18261212, Cur Avg Loss: 0.19249572, Log Avg loss: 0.21617635, Global Avg Loss: 0.70502845, Time: 0.0211 Steps: 89520, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000112, Sample Num: 1792, Cur Loss: 0.21570526, Cur Avg Loss: 0.19387738, Log Avg loss: 0.20797038, Global Avg Loss: 0.70497293, Time: 0.0212 Steps: 89530, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000122, Sample Num: 1952, Cur Loss: 0.13496256, Cur Avg Loss: 0.19890390, Log Avg loss: 0.25520096, Global Avg Loss: 0.70492270, Time: 0.0210 Steps: 89540, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000132, Sample Num: 2112, Cur Loss: 0.06596483, Cur Avg Loss: 0.19616110, Log Avg loss: 0.16269894, Global Avg Loss: 0.70486215, Time: 0.0213 Steps: 89550, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000142, Sample Num: 2272, Cur Loss: 0.02657148, Cur Avg Loss: 0.19776695, Log Avg loss: 0.21896410, Global Avg Loss: 0.70480789, Time: 0.0213 Steps: 89560, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000152, Sample Num: 2432, Cur Loss: 0.17404243, Cur Avg Loss: 0.19668842, Log Avg loss: 0.18137327, Global Avg Loss: 0.70474946, Time: 0.0211 Steps: 89570, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000162, Sample Num: 2592, Cur Loss: 0.27937946, Cur Avg Loss: 0.19794764, Log Avg loss: 0.21708778, Global Avg Loss: 0.70469502, Time: 0.0211 Steps: 89580, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000172, Sample Num: 2752, Cur Loss: 0.28026399, Cur Avg Loss: 0.19681116, Log Avg loss: 0.17840019, Global Avg Loss: 0.70463627, Time: 0.0212 Steps: 89590, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000182, Sample Num: 2912, Cur Loss: 0.08643457, Cur Avg Loss: 0.19578049, Log Avg loss: 0.17805301, Global Avg Loss: 0.70457750, Time: 0.0213 Steps: 89600, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000192, Sample Num: 3072, Cur Loss: 0.30496025, Cur Avg Loss: 0.20093244, Log Avg loss: 0.29469789, Global Avg Loss: 0.70453176, Time: 0.0211 Steps: 89610, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000202, Sample Num: 3232, Cur Loss: 0.23219858, Cur Avg Loss: 0.19806335, Log Avg loss: 0.14297686, Global Avg Loss: 0.70446910, Time: 0.0213 Steps: 89620, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000212, Sample Num: 3392, Cur Loss: 0.09204298, Cur Avg Loss: 0.19855970, Log Avg loss: 0.20858598, Global Avg Loss: 0.70441378, Time: 0.0211 Steps: 89630, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000222, Sample Num: 3552, Cur Loss: 0.07926631, Cur Avg Loss: 0.19640370, Log Avg loss: 0.15069649, Global Avg Loss: 0.70435201, Time: 0.0213 Steps: 89640, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000232, Sample Num: 3712, Cur Loss: 0.09916458, Cur Avg Loss: 0.19627374, Log Avg loss: 0.19338872, Global Avg Loss: 0.70429501, Time: 0.0211 Steps: 89650, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000242, Sample Num: 3872, Cur Loss: 0.12015865, Cur Avg Loss: 0.19448947, Log Avg loss: 0.15309438, Global Avg Loss: 0.70423353, Time: 0.0213 Steps: 89660, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000252, Sample Num: 4032, Cur Loss: 0.13778149, Cur Avg Loss: 0.19210268, Log Avg loss: 0.13434227, Global Avg Loss: 0.70416998, Time: 0.0213 Steps: 89670, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000262, Sample Num: 4192, Cur Loss: 0.18509153, Cur Avg Loss: 0.19320164, Log Avg loss: 0.22089555, Global Avg Loss: 0.70411609, Time: 0.0214 Steps: 89680, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000272, Sample Num: 4352, Cur Loss: 0.29945293, Cur Avg Loss: 0.19521036, Log Avg loss: 0.24783876, Global Avg Loss: 0.70406522, Time: 0.0213 Steps: 89690, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000282, Sample Num: 4512, Cur Loss: 0.06767584, Cur Avg Loss: 0.19280764, Log Avg loss: 0.12745362, Global Avg Loss: 0.70400094, Time: 0.0212 Steps: 89700, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000292, Sample Num: 4672, Cur Loss: 0.10717860, Cur Avg Loss: 0.19179583, Log Avg loss: 0.16326268, Global Avg Loss: 0.70394066, Time: 0.0212 Steps: 89710, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000302, Sample Num: 4832, Cur Loss: 0.32746178, Cur Avg Loss: 0.19372266, Log Avg loss: 0.24998613, Global Avg Loss: 0.70389006, Time: 0.0212 Steps: 89720, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000312, Sample Num: 4992, Cur Loss: 0.16172513, Cur Avg Loss: 0.19414155, Log Avg loss: 0.20679221, Global Avg Loss: 0.70383466, Time: 0.0212 Steps: 89730, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000322, Sample Num: 5152, Cur Loss: 0.26296619, Cur Avg Loss: 0.19437295, Log Avg loss: 0.20159260, Global Avg Loss: 0.70377870, Time: 0.0212 Steps: 89740, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000332, Sample Num: 5312, Cur Loss: 0.12732458, Cur Avg Loss: 0.19375056, Log Avg loss: 0.17370941, Global Avg Loss: 0.70371964, Time: 0.0212 Steps: 89750, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000342, Sample Num: 5472, Cur Loss: 0.37593001, Cur Avg Loss: 0.19323795, Log Avg loss: 0.17621933, Global Avg Loss: 0.70366087, Time: 0.0212 Steps: 89760, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000352, Sample Num: 5632, Cur Loss: 0.08769273, Cur Avg Loss: 0.19217881, Log Avg loss: 0.15595634, Global Avg Loss: 0.70359986, Time: 0.0212 Steps: 89770, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000362, Sample Num: 5792, Cur Loss: 0.10081564, Cur Avg Loss: 0.19177276, Log Avg loss: 0.17747969, Global Avg Loss: 0.70354125, Time: 0.0213 Steps: 89780, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000372, Sample Num: 5952, Cur Loss: 0.16309944, Cur Avg Loss: 0.19202315, Log Avg loss: 0.20108736, Global Avg Loss: 0.70348530, Time: 0.0212 Steps: 89790, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000382, Sample Num: 6112, Cur Loss: 0.13952129, Cur Avg Loss: 0.19134235, Log Avg loss: 0.16601666, Global Avg Loss: 0.70342544, Time: 0.0210 Steps: 89800, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000392, Sample Num: 6272, Cur Loss: 0.04392893, Cur Avg Loss: 0.19048994, Log Avg loss: 0.15792785, Global Avg Loss: 0.70336471, Time: 0.0210 Steps: 89810, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000402, Sample Num: 6432, Cur Loss: 0.15863429, Cur Avg Loss: 0.18955036, Log Avg loss: 0.15271861, Global Avg Loss: 0.70330340, Time: 0.0210 Steps: 89820, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000412, Sample Num: 6592, Cur Loss: 0.12391618, Cur Avg Loss: 0.18859272, Log Avg loss: 0.15009578, Global Avg Loss: 0.70324182, Time: 0.0210 Steps: 89830, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000422, Sample Num: 6752, Cur Loss: 0.03642510, Cur Avg Loss: 0.18874255, Log Avg loss: 0.19491551, Global Avg Loss: 0.70318523, Time: 0.0210 Steps: 89840, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000432, Sample Num: 6912, Cur Loss: 0.19033508, Cur Avg Loss: 0.18796936, Log Avg loss: 0.15534095, Global Avg Loss: 0.70312426, Time: 0.0211 Steps: 89850, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000442, Sample Num: 7072, Cur Loss: 0.26081845, Cur Avg Loss: 0.18978973, Log Avg loss: 0.26842939, Global Avg Loss: 0.70307589, Time: 0.0211 Steps: 89860, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000452, Sample Num: 7232, Cur Loss: 0.21878891, Cur Avg Loss: 0.18869384, Log Avg loss: 0.14025562, Global Avg Loss: 0.70301326, Time: 0.0211 Steps: 89870, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000462, Sample Num: 7392, Cur Loss: 0.12434144, Cur Avg Loss: 0.19044771, Log Avg loss: 0.26972252, Global Avg Loss: 0.70296505, Time: 0.0211 Steps: 89880, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000472, Sample Num: 7552, Cur Loss: 0.10273831, Cur Avg Loss: 0.19150619, Log Avg loss: 0.24040793, Global Avg Loss: 0.70291359, Time: 0.0210 Steps: 89890, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000482, Sample Num: 7712, Cur Loss: 0.20980257, Cur Avg Loss: 0.19104924, Log Avg loss: 0.16948132, Global Avg Loss: 0.70285426, Time: 0.0210 Steps: 89900, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000492, Sample Num: 7872, Cur Loss: 0.22648391, Cur Avg Loss: 0.19118248, Log Avg loss: 0.19760463, Global Avg Loss: 0.70279806, Time: 0.0211 Steps: 89910, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000502, Sample Num: 8032, Cur Loss: 0.60127032, Cur Avg Loss: 0.19136379, Log Avg loss: 0.20028420, Global Avg Loss: 0.70274218, Time: 0.0210 Steps: 89920, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000512, Sample Num: 8192, Cur Loss: 0.07820892, Cur Avg Loss: 0.19074705, Log Avg loss: 0.15978656, Global Avg Loss: 0.70268180, Time: 0.0256 Steps: 89930, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000522, Sample Num: 8352, Cur Loss: 0.14743857, Cur Avg Loss: 0.19017033, Log Avg loss: 0.16064254, Global Avg Loss: 0.70262154, Time: 0.0210 Steps: 89940, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000532, Sample Num: 8512, Cur Loss: 0.31070152, Cur Avg Loss: 0.19062984, Log Avg loss: 0.21461610, Global Avg Loss: 0.70256728, Time: 0.0211 Steps: 89950, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000542, Sample Num: 8672, Cur Loss: 0.24169950, Cur Avg Loss: 0.19019030, Log Avg loss: 0.16680665, Global Avg Loss: 0.70250773, Time: 0.0211 Steps: 89960, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000552, Sample Num: 8832, Cur Loss: 0.40073675, Cur Avg Loss: 0.19219625, Log Avg loss: 0.30091915, Global Avg Loss: 0.70246309, Time: 0.0211 Steps: 89970, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000562, Sample Num: 8992, Cur Loss: 0.38224572, Cur Avg Loss: 0.19252991, Log Avg loss: 0.21094785, Global Avg Loss: 0.70240847, Time: 0.0211 Steps: 89980, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000572, Sample Num: 9152, Cur Loss: 0.03792381, Cur Avg Loss: 0.19192843, Log Avg loss: 0.15812510, Global Avg Loss: 0.70234799, Time: 0.0211 Steps: 89990, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000582, Sample Num: 9312, Cur Loss: 0.20535569, Cur Avg Loss: 0.19219973, Log Avg loss: 0.20771807, Global Avg Loss: 0.70229303, Time: 0.0211 Steps: 90000, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000592, Sample Num: 9472, Cur Loss: 0.32038620, Cur Avg Loss: 0.19256782, Log Avg loss: 0.21399061, Global Avg Loss: 0.70223878, Time: 0.0211 Steps: 90010, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000602, Sample Num: 9632, Cur Loss: 0.13551298, Cur Avg Loss: 0.19301551, Log Avg loss: 0.21951909, Global Avg Loss: 0.70218515, Time: 0.0211 Steps: 90020, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000612, Sample Num: 9792, Cur Loss: 0.04828448, Cur Avg Loss: 0.19307963, Log Avg loss: 0.19693966, Global Avg Loss: 0.70212903, Time: 0.0211 Steps: 90030, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000622, Sample Num: 9952, Cur Loss: 0.04118737, Cur Avg Loss: 0.19172558, Log Avg loss: 0.10885724, Global Avg Loss: 0.70206314, Time: 0.0211 Steps: 90040, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000632, Sample Num: 10112, Cur Loss: 0.05392804, Cur Avg Loss: 0.19098512, Log Avg loss: 0.14492872, Global Avg Loss: 0.70200127, Time: 0.0211 Steps: 90050, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000642, Sample Num: 10272, Cur Loss: 0.18875851, Cur Avg Loss: 0.18972721, Log Avg loss: 0.11022738, Global Avg Loss: 0.70193557, Time: 0.0211 Steps: 90060, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000652, Sample Num: 10432, Cur Loss: 0.08095339, Cur Avg Loss: 0.18928719, Log Avg loss: 0.16103807, Global Avg Loss: 0.70187551, Time: 0.0210 Steps: 90070, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000662, Sample Num: 10592, Cur Loss: 0.05105429, Cur Avg Loss: 0.18886197, Log Avg loss: 0.16113752, Global Avg Loss: 0.70181548, Time: 0.0211 Steps: 90080, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000672, Sample Num: 10752, Cur Loss: 0.22018711, Cur Avg Loss: 0.18876405, Log Avg loss: 0.18228180, Global Avg Loss: 0.70175782, Time: 0.0211 Steps: 90090, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000682, Sample Num: 10912, Cur Loss: 0.03815303, Cur Avg Loss: 0.18888353, Log Avg loss: 0.19691253, Global Avg Loss: 0.70170178, Time: 0.0211 Steps: 90100, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000692, Sample Num: 11072, Cur Loss: 0.24853992, Cur Avg Loss: 0.18830620, Log Avg loss: 0.14893252, Global Avg Loss: 0.70164044, Time: 0.0211 Steps: 90110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000702, Sample Num: 11232, Cur Loss: 0.08793798, Cur Avg Loss: 0.18851621, Log Avg loss: 0.20304831, Global Avg Loss: 0.70158511, Time: 0.0211 Steps: 90120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000712, Sample Num: 11392, Cur Loss: 0.08254075, Cur Avg Loss: 0.18798762, Log Avg loss: 0.15088062, Global Avg Loss: 0.70152401, Time: 0.0210 Steps: 90130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000722, Sample Num: 11552, Cur Loss: 0.14508677, Cur Avg Loss: 0.18770551, Log Avg loss: 0.16761965, Global Avg Loss: 0.70146478, Time: 0.0210 Steps: 90140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000732, Sample Num: 11712, Cur Loss: 0.06064055, Cur Avg Loss: 0.18847936, Log Avg loss: 0.24435099, Global Avg Loss: 0.70141408, Time: 0.0211 Steps: 90150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000742, Sample Num: 11872, Cur Loss: 0.05478945, Cur Avg Loss: 0.18846011, Log Avg loss: 0.18705095, Global Avg Loss: 0.70135703, Time: 0.0211 Steps: 90160, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000752, Sample Num: 12032, Cur Loss: 0.52838039, Cur Avg Loss: 0.18839389, Log Avg loss: 0.18348090, Global Avg Loss: 0.70129959, Time: 0.0211 Steps: 90170, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000762, Sample Num: 12192, Cur Loss: 0.28842884, Cur Avg Loss: 0.18838530, Log Avg loss: 0.18773923, Global Avg Loss: 0.70124265, Time: 0.0210 Steps: 90180, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000772, Sample Num: 12352, Cur Loss: 0.10603298, Cur Avg Loss: 0.18761858, Log Avg loss: 0.12919460, Global Avg Loss: 0.70117922, Time: 0.0248 Steps: 90190, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000782, Sample Num: 12512, Cur Loss: 0.20905875, Cur Avg Loss: 0.18712899, Log Avg loss: 0.14933270, Global Avg Loss: 0.70111804, Time: 0.0211 Steps: 90200, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000792, Sample Num: 12672, Cur Loss: 0.12540343, Cur Avg Loss: 0.18725732, Log Avg loss: 0.19729209, Global Avg Loss: 0.70106219, Time: 0.0211 Steps: 90210, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000802, Sample Num: 12832, Cur Loss: 0.31074089, Cur Avg Loss: 0.18744715, Log Avg loss: 0.20248187, Global Avg Loss: 0.70100692, Time: 0.0210 Steps: 90220, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000812, Sample Num: 12992, Cur Loss: 0.03793302, Cur Avg Loss: 0.18706868, Log Avg loss: 0.15671598, Global Avg Loss: 0.70094660, Time: 0.0211 Steps: 90230, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000822, Sample Num: 13152, Cur Loss: 0.72701442, Cur Avg Loss: 0.18730571, Log Avg loss: 0.20655253, Global Avg Loss: 0.70089182, Time: 0.0211 Steps: 90240, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000832, Sample Num: 13312, Cur Loss: 0.21990018, Cur Avg Loss: 0.18808358, Log Avg loss: 0.25202395, Global Avg Loss: 0.70084208, Time: 0.0211 Steps: 90250, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000842, Sample Num: 13472, Cur Loss: 0.14175005, Cur Avg Loss: 0.18791882, Log Avg loss: 0.17421088, Global Avg Loss: 0.70078373, Time: 0.0211 Steps: 90260, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000852, Sample Num: 13632, Cur Loss: 0.31086200, Cur Avg Loss: 0.18848017, Log Avg loss: 0.23574566, Global Avg Loss: 0.70073222, Time: 0.0211 Steps: 90270, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000862, Sample Num: 13792, Cur Loss: 0.12259318, Cur Avg Loss: 0.18841179, Log Avg loss: 0.18258604, Global Avg Loss: 0.70067482, Time: 0.0211 Steps: 90280, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000872, Sample Num: 13952, Cur Loss: 0.24751852, Cur Avg Loss: 0.18811343, Log Avg loss: 0.16239471, Global Avg Loss: 0.70061521, Time: 0.0211 Steps: 90290, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000882, Sample Num: 14112, Cur Loss: 0.27332020, Cur Avg Loss: 0.18818734, Log Avg loss: 0.19463275, Global Avg Loss: 0.70055917, Time: 0.0211 Steps: 90300, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000892, Sample Num: 14272, Cur Loss: 0.23256262, Cur Avg Loss: 0.18783185, Log Avg loss: 0.15647759, Global Avg Loss: 0.70049893, Time: 0.0211 Steps: 90310, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000902, Sample Num: 14432, Cur Loss: 0.16895717, Cur Avg Loss: 0.18786029, Log Avg loss: 0.19039674, Global Avg Loss: 0.70044245, Time: 0.0211 Steps: 90320, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000912, Sample Num: 14592, Cur Loss: 0.08037928, Cur Avg Loss: 0.18762673, Log Avg loss: 0.16655995, Global Avg Loss: 0.70038335, Time: 0.0211 Steps: 90330, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000922, Sample Num: 14752, Cur Loss: 0.32262176, Cur Avg Loss: 0.18878593, Log Avg loss: 0.29450447, Global Avg Loss: 0.70033842, Time: 0.0211 Steps: 90340, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000932, Sample Num: 14912, Cur Loss: 0.24292932, Cur Avg Loss: 0.18872358, Log Avg loss: 0.18297548, Global Avg Loss: 0.70028116, Time: 0.0212 Steps: 90350, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000942, Sample Num: 15072, Cur Loss: 0.21327984, Cur Avg Loss: 0.18925427, Log Avg loss: 0.23871449, Global Avg Loss: 0.70023008, Time: 0.0212 Steps: 90360, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000952, Sample Num: 15232, Cur Loss: 0.15373623, Cur Avg Loss: 0.18980452, Log Avg loss: 0.24163744, Global Avg Loss: 0.70017933, Time: 0.0209 Steps: 90370, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000962, Sample Num: 15392, Cur Loss: 0.16668768, Cur Avg Loss: 0.18986446, Log Avg loss: 0.19557073, Global Avg Loss: 0.70012350, Time: 0.0211 Steps: 90380, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000972, Sample Num: 15552, Cur Loss: 0.13928798, Cur Avg Loss: 0.18935302, Log Avg loss: 0.14015250, Global Avg Loss: 0.70006155, Time: 0.0211 Steps: 90390, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000982, Sample Num: 15712, Cur Loss: 0.17381339, Cur Avg Loss: 0.18934155, Log Avg loss: 0.18822739, Global Avg Loss: 0.70000493, Time: 0.0210 Steps: 90400, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000992, Sample Num: 15872, Cur Loss: 0.10095386, Cur Avg Loss: 0.18885587, Log Avg loss: 0.14116173, Global Avg Loss: 0.69994312, Time: 0.0212 Steps: 90410, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001002, Sample Num: 16032, Cur Loss: 0.17227449, Cur Avg Loss: 0.18909482, Log Avg loss: 0.21279863, Global Avg Loss: 0.69988924, Time: 0.0211 Steps: 90420, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001012, Sample Num: 16192, Cur Loss: 0.07442245, Cur Avg Loss: 0.18896958, Log Avg loss: 0.17642059, Global Avg Loss: 0.69983135, Time: 0.0211 Steps: 90430, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001022, Sample Num: 16352, Cur Loss: 0.12409457, Cur Avg Loss: 0.18890665, Log Avg loss: 0.18253864, Global Avg Loss: 0.69977416, Time: 0.0210 Steps: 90440, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001032, Sample Num: 16512, Cur Loss: 0.14572546, Cur Avg Loss: 0.18847439, Log Avg loss: 0.14429693, Global Avg Loss: 0.69971274, Time: 0.0211 Steps: 90450, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001042, Sample Num: 16672, Cur Loss: 0.12821867, Cur Avg Loss: 0.18825446, Log Avg loss: 0.16555808, Global Avg Loss: 0.69965370, Time: 0.0211 Steps: 90460, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001052, Sample Num: 16832, Cur Loss: 0.22295506, Cur Avg Loss: 0.18856358, Log Avg loss: 0.22077347, Global Avg Loss: 0.69960076, Time: 0.0210 Steps: 90470, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001062, Sample Num: 16992, Cur Loss: 0.28086615, Cur Avg Loss: 0.18987923, Log Avg loss: 0.32828579, Global Avg Loss: 0.69955972, Time: 0.0211 Steps: 90480, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001072, Sample Num: 17152, Cur Loss: 0.14236876, Cur Avg Loss: 0.18966122, Log Avg loss: 0.16650794, Global Avg Loss: 0.69950082, Time: 0.0210 Steps: 90490, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001082, Sample Num: 17312, Cur Loss: 0.28549585, Cur Avg Loss: 0.18960578, Log Avg loss: 0.18366346, Global Avg Loss: 0.69944382, Time: 0.0210 Steps: 90500, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001092, Sample Num: 17472, Cur Loss: 0.11302248, Cur Avg Loss: 0.18906281, Log Avg loss: 0.13031262, Global Avg Loss: 0.69938094, Time: 0.0211 Steps: 90510, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001102, Sample Num: 17632, Cur Loss: 0.05956708, Cur Avg Loss: 0.18893064, Log Avg loss: 0.17449818, Global Avg Loss: 0.69932295, Time: 0.0209 Steps: 90520, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001112, Sample Num: 17792, Cur Loss: 0.08412342, Cur Avg Loss: 0.18864418, Log Avg loss: 0.15707675, Global Avg Loss: 0.69926306, Time: 0.0208 Steps: 90530, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001122, Sample Num: 17952, Cur Loss: 0.19264100, Cur Avg Loss: 0.18813398, Log Avg loss: 0.13139893, Global Avg Loss: 0.69920034, Time: 0.0208 Steps: 90540, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001132, Sample Num: 18112, Cur Loss: 0.18011120, Cur Avg Loss: 0.18813658, Log Avg loss: 0.18842871, Global Avg Loss: 0.69914393, Time: 0.0209 Steps: 90550, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001142, Sample Num: 18272, Cur Loss: 0.12151743, Cur Avg Loss: 0.18793001, Log Avg loss: 0.16454561, Global Avg Loss: 0.69908490, Time: 0.0208 Steps: 90560, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001152, Sample Num: 18432, Cur Loss: 0.08388168, Cur Avg Loss: 0.18838752, Log Avg loss: 0.24063515, Global Avg Loss: 0.69903428, Time: 0.0209 Steps: 90570, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001162, Sample Num: 18592, Cur Loss: 0.15208355, Cur Avg Loss: 0.18923167, Log Avg loss: 0.28647797, Global Avg Loss: 0.69898873, Time: 0.0208 Steps: 90580, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001172, Sample Num: 18752, Cur Loss: 0.06327432, Cur Avg Loss: 0.18907037, Log Avg loss: 0.17032773, Global Avg Loss: 0.69893037, Time: 0.0208 Steps: 90590, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001182, Sample Num: 18912, Cur Loss: 0.05891799, Cur Avg Loss: 0.18839696, Log Avg loss: 0.10947292, Global Avg Loss: 0.69886531, Time: 0.0209 Steps: 90600, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001192, Sample Num: 19072, Cur Loss: 0.04275608, Cur Avg Loss: 0.18788063, Log Avg loss: 0.12685131, Global Avg Loss: 0.69880218, Time: 0.0208 Steps: 90610, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001202, Sample Num: 19232, Cur Loss: 0.47946990, Cur Avg Loss: 0.18833589, Log Avg loss: 0.24260188, Global Avg Loss: 0.69875184, Time: 0.0209 Steps: 90620, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001212, Sample Num: 19392, Cur Loss: 0.06848752, Cur Avg Loss: 0.18835040, Log Avg loss: 0.19009495, Global Avg Loss: 0.69869572, Time: 0.0209 Steps: 90630, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001222, Sample Num: 19552, Cur Loss: 0.17997248, Cur Avg Loss: 0.18814635, Log Avg loss: 0.16341562, Global Avg Loss: 0.69863666, Time: 0.0208 Steps: 90640, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001232, Sample Num: 19712, Cur Loss: 0.35186917, Cur Avg Loss: 0.18862987, Log Avg loss: 0.24771600, Global Avg Loss: 0.69858692, Time: 0.0209 Steps: 90650, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001242, Sample Num: 19872, Cur Loss: 0.07699366, Cur Avg Loss: 0.18846980, Log Avg loss: 0.16874847, Global Avg Loss: 0.69852848, Time: 0.0209 Steps: 90660, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001252, Sample Num: 20032, Cur Loss: 0.30450875, Cur Avg Loss: 0.18875292, Log Avg loss: 0.22391726, Global Avg Loss: 0.69847613, Time: 0.0208 Steps: 90670, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001262, Sample Num: 20192, Cur Loss: 0.31808448, Cur Avg Loss: 0.18863515, Log Avg loss: 0.17389061, Global Avg Loss: 0.69841828, Time: 0.0209 Steps: 90680, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001272, Sample Num: 20352, Cur Loss: 0.14964277, Cur Avg Loss: 0.18888582, Log Avg loss: 0.22051930, Global Avg Loss: 0.69836558, Time: 0.0209 Steps: 90690, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001282, Sample Num: 20512, Cur Loss: 0.09387076, Cur Avg Loss: 0.18863519, Log Avg loss: 0.15675508, Global Avg Loss: 0.69830587, Time: 0.0247 Steps: 90700, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001292, Sample Num: 20672, Cur Loss: 0.19400156, Cur Avg Loss: 0.18893329, Log Avg loss: 0.22714966, Global Avg Loss: 0.69825393, Time: 0.0209 Steps: 90710, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001302, Sample Num: 20832, Cur Loss: 0.01299925, Cur Avg Loss: 0.18907760, Log Avg loss: 0.20772250, Global Avg Loss: 0.69819986, Time: 0.0208 Steps: 90720, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001312, Sample Num: 20992, Cur Loss: 0.37900579, Cur Avg Loss: 0.18914054, Log Avg loss: 0.19733535, Global Avg Loss: 0.69814465, Time: 0.0209 Steps: 90730, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001322, Sample Num: 21152, Cur Loss: 0.40666866, Cur Avg Loss: 0.18919311, Log Avg loss: 0.19609025, Global Avg Loss: 0.69808933, Time: 0.0210 Steps: 90740, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001332, Sample Num: 21312, Cur Loss: 0.36813468, Cur Avg Loss: 0.18895411, Log Avg loss: 0.15735906, Global Avg Loss: 0.69802974, Time: 0.0209 Steps: 90750, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001342, Sample Num: 21472, Cur Loss: 0.11157230, Cur Avg Loss: 0.18909905, Log Avg loss: 0.20840470, Global Avg Loss: 0.69797579, Time: 0.0210 Steps: 90760, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001352, Sample Num: 21632, Cur Loss: 0.12535332, Cur Avg Loss: 0.18909470, Log Avg loss: 0.18851058, Global Avg Loss: 0.69791967, Time: 0.0208 Steps: 90770, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001362, Sample Num: 21792, Cur Loss: 0.09833807, Cur Avg Loss: 0.18971269, Log Avg loss: 0.27326516, Global Avg Loss: 0.69787289, Time: 0.0208 Steps: 90780, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001372, Sample Num: 21952, Cur Loss: 0.22109984, Cur Avg Loss: 0.18992008, Log Avg loss: 0.21816644, Global Avg Loss: 0.69782005, Time: 0.0208 Steps: 90790, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001382, Sample Num: 22112, Cur Loss: 0.31775782, Cur Avg Loss: 0.19019969, Log Avg loss: 0.22856200, Global Avg Loss: 0.69776837, Time: 0.0208 Steps: 90800, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001392, Sample Num: 22272, Cur Loss: 0.32328850, Cur Avg Loss: 0.19056875, Log Avg loss: 0.24157299, Global Avg Loss: 0.69771813, Time: 0.0210 Steps: 90810, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001402, Sample Num: 22432, Cur Loss: 0.23129220, Cur Avg Loss: 0.19087947, Log Avg loss: 0.23413175, Global Avg Loss: 0.69766709, Time: 0.0209 Steps: 90820, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001412, Sample Num: 22592, Cur Loss: 0.11636747, Cur Avg Loss: 0.19054258, Log Avg loss: 0.14331068, Global Avg Loss: 0.69760606, Time: 0.0211 Steps: 90830, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001422, Sample Num: 22752, Cur Loss: 0.08913308, Cur Avg Loss: 0.19054239, Log Avg loss: 0.19051560, Global Avg Loss: 0.69755024, Time: 0.0208 Steps: 90840, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001432, Sample Num: 22912, Cur Loss: 0.05064818, Cur Avg Loss: 0.18999219, Log Avg loss: 0.11175358, Global Avg Loss: 0.69748576, Time: 0.0209 Steps: 90850, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001442, Sample Num: 23072, Cur Loss: 0.11283686, Cur Avg Loss: 0.18976036, Log Avg loss: 0.15656305, Global Avg Loss: 0.69742622, Time: 0.0210 Steps: 90860, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001452, Sample Num: 23232, Cur Loss: 0.49167114, Cur Avg Loss: 0.19011174, Log Avg loss: 0.24078017, Global Avg Loss: 0.69737597, Time: 0.0210 Steps: 90870, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001462, Sample Num: 23392, Cur Loss: 0.08977440, Cur Avg Loss: 0.19037761, Log Avg loss: 0.22898225, Global Avg Loss: 0.69732443, Time: 0.0209 Steps: 90880, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001472, Sample Num: 23552, Cur Loss: 0.12140437, Cur Avg Loss: 0.18992996, Log Avg loss: 0.12448386, Global Avg Loss: 0.69726140, Time: 0.0209 Steps: 90890, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001482, Sample Num: 23712, Cur Loss: 0.28520149, Cur Avg Loss: 0.19009921, Log Avg loss: 0.21501300, Global Avg Loss: 0.69720835, Time: 0.0209 Steps: 90900, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001492, Sample Num: 23872, Cur Loss: 0.05981452, Cur Avg Loss: 0.19035959, Log Avg loss: 0.22894690, Global Avg Loss: 0.69715684, Time: 0.0208 Steps: 90910, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001502, Sample Num: 24032, Cur Loss: 0.21725604, Cur Avg Loss: 0.19023911, Log Avg loss: 0.17226428, Global Avg Loss: 0.69709911, Time: 0.0208 Steps: 90920, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001512, Sample Num: 24192, Cur Loss: 0.15403223, Cur Avg Loss: 0.19080816, Log Avg loss: 0.27627929, Global Avg Loss: 0.69705283, Time: 0.0208 Steps: 90930, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001522, Sample Num: 24352, Cur Loss: 0.10419759, Cur Avg Loss: 0.19061080, Log Avg loss: 0.16076964, Global Avg Loss: 0.69699386, Time: 0.0210 Steps: 90940, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001532, Sample Num: 24512, Cur Loss: 0.13351864, Cur Avg Loss: 0.19088874, Log Avg loss: 0.23319152, Global Avg Loss: 0.69694287, Time: 0.0209 Steps: 90950, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001542, Sample Num: 24672, Cur Loss: 0.29659799, Cur Avg Loss: 0.19071078, Log Avg loss: 0.16344663, Global Avg Loss: 0.69688421, Time: 0.0246 Steps: 90960, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001552, Sample Num: 24832, Cur Loss: 0.04679667, Cur Avg Loss: 0.19092137, Log Avg loss: 0.22339522, Global Avg Loss: 0.69683217, Time: 0.0209 Steps: 90970, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001562, Sample Num: 24992, Cur Loss: 0.11468829, Cur Avg Loss: 0.19046852, Log Avg loss: 0.12018650, Global Avg Loss: 0.69676878, Time: 0.0209 Steps: 90980, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001572, Sample Num: 25152, Cur Loss: 0.66810334, Cur Avg Loss: 0.19051113, Log Avg loss: 0.19716580, Global Avg Loss: 0.69671388, Time: 0.0209 Steps: 90990, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001582, Sample Num: 25312, Cur Loss: 0.16423552, Cur Avg Loss: 0.19008006, Log Avg loss: 0.12231580, Global Avg Loss: 0.69665076, Time: 0.0209 Steps: 91000, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001592, Sample Num: 25472, Cur Loss: 0.24888197, Cur Avg Loss: 0.19026660, Log Avg loss: 0.21977820, Global Avg Loss: 0.69659836, Time: 0.0209 Steps: 91010, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001602, Sample Num: 25632, Cur Loss: 0.14273301, Cur Avg Loss: 0.19035822, Log Avg loss: 0.20494367, Global Avg Loss: 0.69654434, Time: 0.0209 Steps: 91020, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001612, Sample Num: 25792, Cur Loss: 0.03806324, Cur Avg Loss: 0.18983878, Log Avg loss: 0.10662416, Global Avg Loss: 0.69647954, Time: 0.0209 Steps: 91030, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001622, Sample Num: 25952, Cur Loss: 0.28719974, Cur Avg Loss: 0.18963931, Log Avg loss: 0.15748541, Global Avg Loss: 0.69642033, Time: 0.0209 Steps: 91040, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001632, Sample Num: 26112, Cur Loss: 0.05617315, Cur Avg Loss: 0.18963003, Log Avg loss: 0.18812421, Global Avg Loss: 0.69636451, Time: 0.0209 Steps: 91050, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001642, Sample Num: 26272, Cur Loss: 0.23797719, Cur Avg Loss: 0.18943864, Log Avg loss: 0.15820463, Global Avg Loss: 0.69630541, Time: 0.0209 Steps: 91060, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001652, Sample Num: 26432, Cur Loss: 0.21219233, Cur Avg Loss: 0.18936591, Log Avg loss: 0.17742262, Global Avg Loss: 0.69624843, Time: 0.0209 Steps: 91070, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001662, Sample Num: 26592, Cur Loss: 0.22429922, Cur Avg Loss: 0.18929991, Log Avg loss: 0.17839785, Global Avg Loss: 0.69619157, Time: 0.0209 Steps: 91080, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001672, Sample Num: 26752, Cur Loss: 0.28706512, Cur Avg Loss: 0.18942483, Log Avg loss: 0.21018505, Global Avg Loss: 0.69613822, Time: 0.0210 Steps: 91090, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001682, Sample Num: 26912, Cur Loss: 0.55365735, Cur Avg Loss: 0.18936586, Log Avg loss: 0.17950710, Global Avg Loss: 0.69608151, Time: 0.0209 Steps: 91100, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001692, Sample Num: 27072, Cur Loss: 0.26529562, Cur Avg Loss: 0.18918634, Log Avg loss: 0.15899073, Global Avg Loss: 0.69602256, Time: 0.0209 Steps: 91110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001702, Sample Num: 27232, Cur Loss: 0.18371165, Cur Avg Loss: 0.18899482, Log Avg loss: 0.15658892, Global Avg Loss: 0.69596336, Time: 0.0210 Steps: 91120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001712, Sample Num: 27392, Cur Loss: 0.45830679, Cur Avg Loss: 0.18970492, Log Avg loss: 0.31056438, Global Avg Loss: 0.69592107, Time: 0.0209 Steps: 91130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001722, Sample Num: 27552, Cur Loss: 0.07000235, Cur Avg Loss: 0.18970490, Log Avg loss: 0.18970244, Global Avg Loss: 0.69586553, Time: 0.0209 Steps: 91140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001732, Sample Num: 27712, Cur Loss: 0.17620438, Cur Avg Loss: 0.18965068, Log Avg loss: 0.18031296, Global Avg Loss: 0.69580896, Time: 0.0209 Steps: 91150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001742, Sample Num: 27872, Cur Loss: 0.14172252, Cur Avg Loss: 0.18939665, Log Avg loss: 0.14539836, Global Avg Loss: 0.69574859, Time: 0.0209 Steps: 91160, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001752, Sample Num: 28032, Cur Loss: 0.18089397, Cur Avg Loss: 0.18937627, Log Avg loss: 0.18582757, Global Avg Loss: 0.69569266, Time: 0.0209 Steps: 91170, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001762, Sample Num: 28192, Cur Loss: 0.24263296, Cur Avg Loss: 0.18972277, Log Avg loss: 0.25042798, Global Avg Loss: 0.69564382, Time: 0.0209 Steps: 91180, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001772, Sample Num: 28352, Cur Loss: 0.10116964, Cur Avg Loss: 0.18967075, Log Avg loss: 0.18050487, Global Avg Loss: 0.69558733, Time: 0.0209 Steps: 91190, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001782, Sample Num: 28512, Cur Loss: 0.05357816, Cur Avg Loss: 0.18947644, Log Avg loss: 0.15504461, Global Avg Loss: 0.69552806, Time: 0.0209 Steps: 91200, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001792, Sample Num: 28672, Cur Loss: 0.11171143, Cur Avg Loss: 0.18958998, Log Avg loss: 0.20982388, Global Avg Loss: 0.69547481, Time: 0.0255 Steps: 91210, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001802, Sample Num: 28832, Cur Loss: 0.07854250, Cur Avg Loss: 0.18958025, Log Avg loss: 0.18783639, Global Avg Loss: 0.69541916, Time: 0.0209 Steps: 91220, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001812, Sample Num: 28992, Cur Loss: 0.23652266, Cur Avg Loss: 0.18972249, Log Avg loss: 0.21535359, Global Avg Loss: 0.69536654, Time: 0.0209 Steps: 91230, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001822, Sample Num: 29152, Cur Loss: 0.08181834, Cur Avg Loss: 0.18965732, Log Avg loss: 0.17784889, Global Avg Loss: 0.69530982, Time: 0.0209 Steps: 91240, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001832, Sample Num: 29312, Cur Loss: 0.17842467, Cur Avg Loss: 0.18973353, Log Avg loss: 0.20361879, Global Avg Loss: 0.69525593, Time: 0.0209 Steps: 91250, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001842, Sample Num: 29472, Cur Loss: 0.48195642, Cur Avg Loss: 0.19036572, Log Avg loss: 0.30618328, Global Avg Loss: 0.69521330, Time: 0.0209 Steps: 91260, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001852, Sample Num: 29632, Cur Loss: 0.34001613, Cur Avg Loss: 0.19020864, Log Avg loss: 0.16127488, Global Avg Loss: 0.69515480, Time: 0.0209 Steps: 91270, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001862, Sample Num: 29792, Cur Loss: 0.10738152, Cur Avg Loss: 0.19052593, Log Avg loss: 0.24928745, Global Avg Loss: 0.69510595, Time: 0.0209 Steps: 91280, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001872, Sample Num: 29952, Cur Loss: 0.25121272, Cur Avg Loss: 0.19041968, Log Avg loss: 0.17063628, Global Avg Loss: 0.69504850, Time: 0.0209 Steps: 91290, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001882, Sample Num: 30112, Cur Loss: 0.05082416, Cur Avg Loss: 0.19058822, Log Avg loss: 0.22213845, Global Avg Loss: 0.69499670, Time: 0.0209 Steps: 91300, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001892, Sample Num: 30272, Cur Loss: 0.16445704, Cur Avg Loss: 0.19078091, Log Avg loss: 0.22704476, Global Avg Loss: 0.69494546, Time: 0.0209 Steps: 91310, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001902, Sample Num: 30432, Cur Loss: 0.21466054, Cur Avg Loss: 0.19108300, Log Avg loss: 0.24823958, Global Avg Loss: 0.69489654, Time: 0.0209 Steps: 91320, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001912, Sample Num: 30592, Cur Loss: 0.10818981, Cur Avg Loss: 0.19107247, Log Avg loss: 0.18906947, Global Avg Loss: 0.69484116, Time: 0.0210 Steps: 91330, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001922, Sample Num: 30752, Cur Loss: 0.51661944, Cur Avg Loss: 0.19172396, Log Avg loss: 0.31628950, Global Avg Loss: 0.69479971, Time: 0.0210 Steps: 91340, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001932, Sample Num: 30912, Cur Loss: 0.18062244, Cur Avg Loss: 0.19185911, Log Avg loss: 0.21783460, Global Avg Loss: 0.69474750, Time: 0.0210 Steps: 91350, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001942, Sample Num: 31072, Cur Loss: 0.06997439, Cur Avg Loss: 0.19162666, Log Avg loss: 0.14671601, Global Avg Loss: 0.69468751, Time: 0.0209 Steps: 91360, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001952, Sample Num: 31232, Cur Loss: 0.33318639, Cur Avg Loss: 0.19159614, Log Avg loss: 0.18567014, Global Avg Loss: 0.69463180, Time: 0.0209 Steps: 91370, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001962, Sample Num: 31392, Cur Loss: 0.15437308, Cur Avg Loss: 0.19157745, Log Avg loss: 0.18792976, Global Avg Loss: 0.69457635, Time: 0.0209 Steps: 91380, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001972, Sample Num: 31552, Cur Loss: 0.13133135, Cur Avg Loss: 0.19138451, Log Avg loss: 0.15352894, Global Avg Loss: 0.69451715, Time: 0.0209 Steps: 91390, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001982, Sample Num: 31712, Cur Loss: 0.38182610, Cur Avg Loss: 0.19120948, Log Avg loss: 0.15669361, Global Avg Loss: 0.69445831, Time: 0.0212 Steps: 91400, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001992, Sample Num: 31872, Cur Loss: 0.16953105, Cur Avg Loss: 0.19111679, Log Avg loss: 0.17274619, Global Avg Loss: 0.69440123, Time: 0.0209 Steps: 91410, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002002, Sample Num: 32032, Cur Loss: 0.14370328, Cur Avg Loss: 0.19110955, Log Avg loss: 0.18966716, Global Avg Loss: 0.69434602, Time: 0.0209 Steps: 91420, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002012, Sample Num: 32192, Cur Loss: 0.31840849, Cur Avg Loss: 0.19135243, Log Avg loss: 0.23997593, Global Avg Loss: 0.69429633, Time: 0.0209 Steps: 91430, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002022, Sample Num: 32352, Cur Loss: 0.39624316, Cur Avg Loss: 0.19142283, Log Avg loss: 0.20558805, Global Avg Loss: 0.69424288, Time: 0.0209 Steps: 91440, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002032, Sample Num: 32512, Cur Loss: 0.36878169, Cur Avg Loss: 0.19156764, Log Avg loss: 0.22084862, Global Avg Loss: 0.69419112, Time: 0.0209 Steps: 91450, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002042, Sample Num: 32672, Cur Loss: 0.08210808, Cur Avg Loss: 0.19164872, Log Avg loss: 0.20812337, Global Avg Loss: 0.69413797, Time: 0.0209 Steps: 91460, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002052, Sample Num: 32832, Cur Loss: 0.14442565, Cur Avg Loss: 0.19137141, Log Avg loss: 0.13474598, Global Avg Loss: 0.69407682, Time: 0.0245 Steps: 91470, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002062, Sample Num: 32992, Cur Loss: 0.12391631, Cur Avg Loss: 0.19145447, Log Avg loss: 0.20849806, Global Avg Loss: 0.69402373, Time: 0.0208 Steps: 91480, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002072, Sample Num: 33152, Cur Loss: 0.13384692, Cur Avg Loss: 0.19152979, Log Avg loss: 0.20705957, Global Avg Loss: 0.69397051, Time: 0.0207 Steps: 91490, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002082, Sample Num: 33312, Cur Loss: 0.14967927, Cur Avg Loss: 0.19152021, Log Avg loss: 0.18953620, Global Avg Loss: 0.69391538, Time: 0.0207 Steps: 91500, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002092, Sample Num: 33472, Cur Loss: 0.09804888, Cur Avg Loss: 0.19131361, Log Avg loss: 0.14829979, Global Avg Loss: 0.69385576, Time: 0.0208 Steps: 91510, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002102, Sample Num: 33632, Cur Loss: 0.09974774, Cur Avg Loss: 0.19117720, Log Avg loss: 0.16263992, Global Avg Loss: 0.69379771, Time: 0.0208 Steps: 91520, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002112, Sample Num: 33792, Cur Loss: 0.05904110, Cur Avg Loss: 0.19089645, Log Avg loss: 0.13188322, Global Avg Loss: 0.69373632, Time: 0.0209 Steps: 91530, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002122, Sample Num: 33952, Cur Loss: 0.11065197, Cur Avg Loss: 0.19083260, Log Avg loss: 0.17734772, Global Avg Loss: 0.69367991, Time: 0.0208 Steps: 91540, Updated lr: 0.000014 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.899237, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.237157, "eval_total_loss": 166.721445, "eval_mae": 0.332667, "eval_mse": 0.237219, "eval_r2": 0.849208, "eval_sp_statistic": 0.905039, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924539, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.396912, "test_total_loss": 199.249838, "test_mae": 0.384466, "test_mse": 0.397053, "test_r2": 0.743738, "test_sp_statistic": 0.87417, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.919932, "test_ps_pvalue": 0.0, "lr": 1.413276434329066e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6936404578076866, "train_cur_epoch_loss": 406.19085762184113, "train_cur_epoch_avg_loss": 0.1907895056936783, "train_cur_epoch_time": 44.89923691749573, "train_cur_epoch_avg_time": 0.021089355057536743, "epoch": 43, "step": 91547} ################################################## Training, Epoch: 0044, Batch: 000003, Sample Num: 48, Cur Loss: 0.09344587, Cur Avg Loss: 0.19468172, Log Avg loss: 0.18281151, Global Avg Loss: 0.69362411, Time: 0.0246 Steps: 91550, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000013, Sample Num: 208, Cur Loss: 0.13891830, Cur Avg Loss: 0.14012063, Log Avg loss: 0.12375231, Global Avg Loss: 0.69356187, Time: 0.0207 Steps: 91560, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000023, Sample Num: 368, Cur Loss: 0.15589660, Cur Avg Loss: 0.15198982, Log Avg loss: 0.16741976, Global Avg Loss: 0.69350441, Time: 0.0209 Steps: 91570, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000033, Sample Num: 528, Cur Loss: 0.07844169, Cur Avg Loss: 0.16826519, Log Avg loss: 0.20569855, Global Avg Loss: 0.69345114, Time: 0.0209 Steps: 91580, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000043, Sample Num: 688, Cur Loss: 0.15695556, Cur Avg Loss: 0.17175349, Log Avg loss: 0.18326488, Global Avg Loss: 0.69339544, Time: 0.0209 Steps: 91590, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000053, Sample Num: 848, Cur Loss: 0.14359030, Cur Avg Loss: 0.17198811, Log Avg loss: 0.17299697, Global Avg Loss: 0.69333863, Time: 0.0208 Steps: 91600, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000063, Sample Num: 1008, Cur Loss: 0.03957685, Cur Avg Loss: 0.16628200, Log Avg loss: 0.13603963, Global Avg Loss: 0.69327779, Time: 0.0209 Steps: 91610, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000073, Sample Num: 1168, Cur Loss: 0.20157740, Cur Avg Loss: 0.16656494, Log Avg loss: 0.16834747, Global Avg Loss: 0.69322050, Time: 0.0209 Steps: 91620, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000083, Sample Num: 1328, Cur Loss: 0.08943716, Cur Avg Loss: 0.17115234, Log Avg loss: 0.20464038, Global Avg Loss: 0.69316718, Time: 0.0209 Steps: 91630, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000093, Sample Num: 1488, Cur Loss: 0.08998799, Cur Avg Loss: 0.17486213, Log Avg loss: 0.20565338, Global Avg Loss: 0.69311398, Time: 0.0209 Steps: 91640, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000103, Sample Num: 1648, Cur Loss: 0.10249805, Cur Avg Loss: 0.17054394, Log Avg loss: 0.13038477, Global Avg Loss: 0.69305258, Time: 0.0208 Steps: 91650, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000113, Sample Num: 1808, Cur Loss: 0.07346319, Cur Avg Loss: 0.17357567, Log Avg loss: 0.20480244, Global Avg Loss: 0.69299931, Time: 0.0209 Steps: 91660, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000123, Sample Num: 1968, Cur Loss: 0.14888656, Cur Avg Loss: 0.17350713, Log Avg loss: 0.17273263, Global Avg Loss: 0.69294256, Time: 0.0208 Steps: 91670, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000133, Sample Num: 2128, Cur Loss: 0.33607191, Cur Avg Loss: 0.17229709, Log Avg loss: 0.15741356, Global Avg Loss: 0.69288415, Time: 0.0208 Steps: 91680, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000143, Sample Num: 2288, Cur Loss: 0.08050960, Cur Avg Loss: 0.17027539, Log Avg loss: 0.14338690, Global Avg Loss: 0.69282422, Time: 0.0209 Steps: 91690, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000153, Sample Num: 2448, Cur Loss: 0.14250250, Cur Avg Loss: 0.17136382, Log Avg loss: 0.18692831, Global Avg Loss: 0.69276905, Time: 0.0209 Steps: 91700, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000163, Sample Num: 2608, Cur Loss: 0.16581894, Cur Avg Loss: 0.16994809, Log Avg loss: 0.14828744, Global Avg Loss: 0.69270968, Time: 0.0209 Steps: 91710, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000173, Sample Num: 2768, Cur Loss: 0.09504105, Cur Avg Loss: 0.17206007, Log Avg loss: 0.20648523, Global Avg Loss: 0.69265667, Time: 0.0209 Steps: 91720, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000183, Sample Num: 2928, Cur Loss: 0.04134097, Cur Avg Loss: 0.17342880, Log Avg loss: 0.19710797, Global Avg Loss: 0.69260264, Time: 0.0209 Steps: 91730, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000193, Sample Num: 3088, Cur Loss: 0.19925027, Cur Avg Loss: 0.17266344, Log Avg loss: 0.15865730, Global Avg Loss: 0.69254444, Time: 0.0209 Steps: 91740, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000203, Sample Num: 3248, Cur Loss: 0.11253250, Cur Avg Loss: 0.17174042, Log Avg loss: 0.15392623, Global Avg Loss: 0.69248574, Time: 0.0209 Steps: 91750, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000213, Sample Num: 3408, Cur Loss: 0.14576010, Cur Avg Loss: 0.17321495, Log Avg loss: 0.20314788, Global Avg Loss: 0.69243241, Time: 0.0209 Steps: 91760, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000223, Sample Num: 3568, Cur Loss: 0.05099033, Cur Avg Loss: 0.17781748, Log Avg loss: 0.27585120, Global Avg Loss: 0.69238701, Time: 0.0209 Steps: 91770, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000233, Sample Num: 3728, Cur Loss: 0.05211323, Cur Avg Loss: 0.17889171, Log Avg loss: 0.20284707, Global Avg Loss: 0.69233368, Time: 0.0209 Steps: 91780, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000243, Sample Num: 3888, Cur Loss: 0.27042601, Cur Avg Loss: 0.17737623, Log Avg loss: 0.14206563, Global Avg Loss: 0.69227373, Time: 0.0209 Steps: 91790, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000253, Sample Num: 4048, Cur Loss: 0.49344075, Cur Avg Loss: 0.18145357, Log Avg loss: 0.28053301, Global Avg Loss: 0.69222888, Time: 0.0208 Steps: 91800, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000263, Sample Num: 4208, Cur Loss: 0.21464482, Cur Avg Loss: 0.17937217, Log Avg loss: 0.12671270, Global Avg Loss: 0.69216728, Time: 0.0212 Steps: 91810, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000273, Sample Num: 4368, Cur Loss: 0.08047456, Cur Avg Loss: 0.17806933, Log Avg loss: 0.14380468, Global Avg Loss: 0.69210756, Time: 0.0209 Steps: 91820, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000283, Sample Num: 4528, Cur Loss: 0.11909714, Cur Avg Loss: 0.17802000, Log Avg loss: 0.17667321, Global Avg Loss: 0.69205143, Time: 0.0208 Steps: 91830, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000293, Sample Num: 4688, Cur Loss: 0.02056271, Cur Avg Loss: 0.17735665, Log Avg loss: 0.15858382, Global Avg Loss: 0.69199334, Time: 0.0208 Steps: 91840, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000303, Sample Num: 4848, Cur Loss: 0.17564163, Cur Avg Loss: 0.17762273, Log Avg loss: 0.18541894, Global Avg Loss: 0.69193819, Time: 0.0208 Steps: 91850, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000313, Sample Num: 5008, Cur Loss: 0.12035793, Cur Avg Loss: 0.17783525, Log Avg loss: 0.18427451, Global Avg Loss: 0.69188292, Time: 0.0208 Steps: 91860, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000323, Sample Num: 5168, Cur Loss: 0.31757522, Cur Avg Loss: 0.17638967, Log Avg loss: 0.13114307, Global Avg Loss: 0.69182189, Time: 0.0209 Steps: 91870, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000333, Sample Num: 5328, Cur Loss: 0.16314484, Cur Avg Loss: 0.17596162, Log Avg loss: 0.16213557, Global Avg Loss: 0.69176424, Time: 0.0208 Steps: 91880, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000343, Sample Num: 5488, Cur Loss: 0.34133703, Cur Avg Loss: 0.17728226, Log Avg loss: 0.22125964, Global Avg Loss: 0.69171304, Time: 0.0209 Steps: 91890, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000353, Sample Num: 5648, Cur Loss: 0.45563915, Cur Avg Loss: 0.18024156, Log Avg loss: 0.28174544, Global Avg Loss: 0.69166843, Time: 0.0209 Steps: 91900, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000363, Sample Num: 5808, Cur Loss: 0.13309827, Cur Avg Loss: 0.18078215, Log Avg loss: 0.19986506, Global Avg Loss: 0.69161492, Time: 0.0209 Steps: 91910, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000373, Sample Num: 5968, Cur Loss: 0.19484228, Cur Avg Loss: 0.18220418, Log Avg loss: 0.23382401, Global Avg Loss: 0.69156511, Time: 0.0209 Steps: 91920, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000383, Sample Num: 6128, Cur Loss: 0.12212038, Cur Avg Loss: 0.18249741, Log Avg loss: 0.19343478, Global Avg Loss: 0.69151093, Time: 0.0209 Steps: 91930, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000393, Sample Num: 6288, Cur Loss: 0.34299660, Cur Avg Loss: 0.18217917, Log Avg loss: 0.16999050, Global Avg Loss: 0.69145420, Time: 0.0209 Steps: 91940, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000403, Sample Num: 6448, Cur Loss: 0.28173697, Cur Avg Loss: 0.18419265, Log Avg loss: 0.26332251, Global Avg Loss: 0.69140764, Time: 0.0209 Steps: 91950, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000413, Sample Num: 6608, Cur Loss: 0.36982054, Cur Avg Loss: 0.18398681, Log Avg loss: 0.17569119, Global Avg Loss: 0.69135156, Time: 0.0209 Steps: 91960, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000423, Sample Num: 6768, Cur Loss: 0.05112670, Cur Avg Loss: 0.18391628, Log Avg loss: 0.18100373, Global Avg Loss: 0.69129607, Time: 0.0208 Steps: 91970, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000433, Sample Num: 6928, Cur Loss: 0.24219711, Cur Avg Loss: 0.18417118, Log Avg loss: 0.19495316, Global Avg Loss: 0.69124211, Time: 0.0208 Steps: 91980, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000443, Sample Num: 7088, Cur Loss: 0.12841764, Cur Avg Loss: 0.18299707, Log Avg loss: 0.13215845, Global Avg Loss: 0.69118133, Time: 0.0208 Steps: 91990, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000453, Sample Num: 7248, Cur Loss: 0.05710160, Cur Avg Loss: 0.18220923, Log Avg loss: 0.14730786, Global Avg Loss: 0.69112221, Time: 0.0209 Steps: 92000, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000463, Sample Num: 7408, Cur Loss: 0.19865747, Cur Avg Loss: 0.18231425, Log Avg loss: 0.18707153, Global Avg Loss: 0.69106743, Time: 0.0209 Steps: 92010, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000473, Sample Num: 7568, Cur Loss: 0.12663530, Cur Avg Loss: 0.18274241, Log Avg loss: 0.20256608, Global Avg Loss: 0.69101435, Time: 0.0208 Steps: 92020, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000483, Sample Num: 7728, Cur Loss: 0.08249173, Cur Avg Loss: 0.18195705, Log Avg loss: 0.14480945, Global Avg Loss: 0.69095500, Time: 0.0208 Steps: 92030, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000493, Sample Num: 7888, Cur Loss: 0.34944567, Cur Avg Loss: 0.18342678, Log Avg loss: 0.25441510, Global Avg Loss: 0.69090757, Time: 0.0209 Steps: 92040, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000503, Sample Num: 8048, Cur Loss: 0.11746731, Cur Avg Loss: 0.18257980, Log Avg loss: 0.14082338, Global Avg Loss: 0.69084781, Time: 0.0208 Steps: 92050, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000513, Sample Num: 8208, Cur Loss: 0.21389206, Cur Avg Loss: 0.18253509, Log Avg loss: 0.18028617, Global Avg Loss: 0.69079235, Time: 0.0245 Steps: 92060, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000523, Sample Num: 8368, Cur Loss: 0.04003694, Cur Avg Loss: 0.18154189, Log Avg loss: 0.13059064, Global Avg Loss: 0.69073150, Time: 0.0209 Steps: 92070, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000533, Sample Num: 8528, Cur Loss: 0.15856256, Cur Avg Loss: 0.18168186, Log Avg loss: 0.18900233, Global Avg Loss: 0.69067701, Time: 0.0209 Steps: 92080, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000543, Sample Num: 8688, Cur Loss: 0.45050520, Cur Avg Loss: 0.18187715, Log Avg loss: 0.19228648, Global Avg Loss: 0.69062289, Time: 0.0209 Steps: 92090, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000553, Sample Num: 8848, Cur Loss: 0.05466227, Cur Avg Loss: 0.18127457, Log Avg loss: 0.14855405, Global Avg Loss: 0.69056404, Time: 0.0208 Steps: 92100, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000563, Sample Num: 9008, Cur Loss: 0.14153779, Cur Avg Loss: 0.18033557, Log Avg loss: 0.12840903, Global Avg Loss: 0.69050301, Time: 0.0209 Steps: 92110, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000573, Sample Num: 9168, Cur Loss: 0.08560080, Cur Avg Loss: 0.17983954, Log Avg loss: 0.15191301, Global Avg Loss: 0.69044454, Time: 0.0209 Steps: 92120, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000583, Sample Num: 9328, Cur Loss: 0.10329023, Cur Avg Loss: 0.17957474, Log Avg loss: 0.16440161, Global Avg Loss: 0.69038744, Time: 0.0208 Steps: 92130, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000593, Sample Num: 9488, Cur Loss: 0.17346808, Cur Avg Loss: 0.17934852, Log Avg loss: 0.16616009, Global Avg Loss: 0.69033055, Time: 0.0208 Steps: 92140, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000603, Sample Num: 9648, Cur Loss: 0.18431015, Cur Avg Loss: 0.17869817, Log Avg loss: 0.14013268, Global Avg Loss: 0.69027084, Time: 0.0209 Steps: 92150, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000613, Sample Num: 9808, Cur Loss: 0.09552132, Cur Avg Loss: 0.17846570, Log Avg loss: 0.16444721, Global Avg Loss: 0.69021379, Time: 0.0209 Steps: 92160, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000623, Sample Num: 9968, Cur Loss: 0.28462267, Cur Avg Loss: 0.17850984, Log Avg loss: 0.18121618, Global Avg Loss: 0.69015856, Time: 0.0209 Steps: 92170, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000633, Sample Num: 10128, Cur Loss: 0.25947845, Cur Avg Loss: 0.17778649, Log Avg loss: 0.13272139, Global Avg Loss: 0.69009809, Time: 0.0209 Steps: 92180, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000643, Sample Num: 10288, Cur Loss: 0.08167584, Cur Avg Loss: 0.17993488, Log Avg loss: 0.31592806, Global Avg Loss: 0.69005750, Time: 0.0209 Steps: 92190, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000653, Sample Num: 10448, Cur Loss: 0.22429338, Cur Avg Loss: 0.18082235, Log Avg loss: 0.23788699, Global Avg Loss: 0.69000846, Time: 0.0209 Steps: 92200, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000663, Sample Num: 10608, Cur Loss: 0.31769365, Cur Avg Loss: 0.18060789, Log Avg loss: 0.16660310, Global Avg Loss: 0.68995170, Time: 0.0209 Steps: 92210, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000673, Sample Num: 10768, Cur Loss: 0.06188487, Cur Avg Loss: 0.18052393, Log Avg loss: 0.17495770, Global Avg Loss: 0.68989585, Time: 0.0208 Steps: 92220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000683, Sample Num: 10928, Cur Loss: 0.04462095, Cur Avg Loss: 0.18028788, Log Avg loss: 0.16440151, Global Avg Loss: 0.68983888, Time: 0.0209 Steps: 92230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000693, Sample Num: 11088, Cur Loss: 0.10845648, Cur Avg Loss: 0.18009433, Log Avg loss: 0.16687471, Global Avg Loss: 0.68978218, Time: 0.0208 Steps: 92240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000703, Sample Num: 11248, Cur Loss: 0.10522375, Cur Avg Loss: 0.17968669, Log Avg loss: 0.15143719, Global Avg Loss: 0.68972382, Time: 0.0208 Steps: 92250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000713, Sample Num: 11408, Cur Loss: 0.19977197, Cur Avg Loss: 0.17973729, Log Avg loss: 0.18329473, Global Avg Loss: 0.68966893, Time: 0.0209 Steps: 92260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000723, Sample Num: 11568, Cur Loss: 0.15918958, Cur Avg Loss: 0.17975330, Log Avg loss: 0.18089465, Global Avg Loss: 0.68961379, Time: 0.0209 Steps: 92270, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000733, Sample Num: 11728, Cur Loss: 0.05978942, Cur Avg Loss: 0.17875330, Log Avg loss: 0.10645372, Global Avg Loss: 0.68955060, Time: 0.0208 Steps: 92280, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000743, Sample Num: 11888, Cur Loss: 0.16623425, Cur Avg Loss: 0.17931101, Log Avg loss: 0.22019129, Global Avg Loss: 0.68949974, Time: 0.0208 Steps: 92290, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000753, Sample Num: 12048, Cur Loss: 0.21763553, Cur Avg Loss: 0.18009656, Log Avg loss: 0.23846279, Global Avg Loss: 0.68945087, Time: 0.0208 Steps: 92300, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000763, Sample Num: 12208, Cur Loss: 0.09721479, Cur Avg Loss: 0.17948636, Log Avg loss: 0.13353804, Global Avg Loss: 0.68939065, Time: 0.0209 Steps: 92310, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000773, Sample Num: 12368, Cur Loss: 0.06879765, Cur Avg Loss: 0.18017686, Log Avg loss: 0.23286201, Global Avg Loss: 0.68934120, Time: 0.0245 Steps: 92320, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000783, Sample Num: 12528, Cur Loss: 0.20920660, Cur Avg Loss: 0.17966737, Log Avg loss: 0.14028392, Global Avg Loss: 0.68928173, Time: 0.0209 Steps: 92330, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000793, Sample Num: 12688, Cur Loss: 0.24214511, Cur Avg Loss: 0.18003382, Log Avg loss: 0.20872668, Global Avg Loss: 0.68922969, Time: 0.0208 Steps: 92340, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000803, Sample Num: 12848, Cur Loss: 0.11908723, Cur Avg Loss: 0.18028843, Log Avg loss: 0.20047916, Global Avg Loss: 0.68917677, Time: 0.0209 Steps: 92350, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000813, Sample Num: 13008, Cur Loss: 0.34462616, Cur Avg Loss: 0.18071922, Log Avg loss: 0.21531138, Global Avg Loss: 0.68912546, Time: 0.0209 Steps: 92360, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000823, Sample Num: 13168, Cur Loss: 0.22072439, Cur Avg Loss: 0.18084474, Log Avg loss: 0.19104963, Global Avg Loss: 0.68907154, Time: 0.0209 Steps: 92370, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000833, Sample Num: 13328, Cur Loss: 0.02040214, Cur Avg Loss: 0.18069174, Log Avg loss: 0.16809986, Global Avg Loss: 0.68901515, Time: 0.0209 Steps: 92380, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000843, Sample Num: 13488, Cur Loss: 0.17317942, Cur Avg Loss: 0.18143302, Log Avg loss: 0.24318216, Global Avg Loss: 0.68896689, Time: 0.0209 Steps: 92390, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000853, Sample Num: 13648, Cur Loss: 0.51638305, Cur Avg Loss: 0.18221347, Log Avg loss: 0.24800532, Global Avg Loss: 0.68891917, Time: 0.0209 Steps: 92400, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000863, Sample Num: 13808, Cur Loss: 0.02358999, Cur Avg Loss: 0.18154113, Log Avg loss: 0.12419004, Global Avg Loss: 0.68885806, Time: 0.0209 Steps: 92410, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000873, Sample Num: 13968, Cur Loss: 0.11650677, Cur Avg Loss: 0.18159544, Log Avg loss: 0.18628283, Global Avg Loss: 0.68880368, Time: 0.0209 Steps: 92420, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000883, Sample Num: 14128, Cur Loss: 0.19025379, Cur Avg Loss: 0.18126215, Log Avg loss: 0.15216583, Global Avg Loss: 0.68874562, Time: 0.0210 Steps: 92430, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000893, Sample Num: 14288, Cur Loss: 0.11937788, Cur Avg Loss: 0.18160728, Log Avg loss: 0.21208195, Global Avg Loss: 0.68869405, Time: 0.0209 Steps: 92440, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000903, Sample Num: 14448, Cur Loss: 0.08592257, Cur Avg Loss: 0.18120691, Log Avg loss: 0.14545394, Global Avg Loss: 0.68863529, Time: 0.0209 Steps: 92450, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000913, Sample Num: 14608, Cur Loss: 0.16063896, Cur Avg Loss: 0.18221269, Log Avg loss: 0.27303509, Global Avg Loss: 0.68859034, Time: 0.0208 Steps: 92460, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000923, Sample Num: 14768, Cur Loss: 0.15394440, Cur Avg Loss: 0.18225982, Log Avg loss: 0.18656265, Global Avg Loss: 0.68853605, Time: 0.0210 Steps: 92470, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000933, Sample Num: 14928, Cur Loss: 0.05010770, Cur Avg Loss: 0.18159872, Log Avg loss: 0.12057856, Global Avg Loss: 0.68847464, Time: 0.0208 Steps: 92480, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000943, Sample Num: 15088, Cur Loss: 0.07836342, Cur Avg Loss: 0.18158296, Log Avg loss: 0.18011324, Global Avg Loss: 0.68841967, Time: 0.0208 Steps: 92490, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000953, Sample Num: 15248, Cur Loss: 0.18395813, Cur Avg Loss: 0.18129018, Log Avg loss: 0.15368048, Global Avg Loss: 0.68836187, Time: 0.0209 Steps: 92500, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000963, Sample Num: 15408, Cur Loss: 0.18225557, Cur Avg Loss: 0.18082556, Log Avg loss: 0.13654763, Global Avg Loss: 0.68830222, Time: 0.0209 Steps: 92510, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000973, Sample Num: 15568, Cur Loss: 0.49307370, Cur Avg Loss: 0.18100372, Log Avg loss: 0.19816074, Global Avg Loss: 0.68824924, Time: 0.0208 Steps: 92520, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000983, Sample Num: 15728, Cur Loss: 0.13609980, Cur Avg Loss: 0.18029149, Log Avg loss: 0.11099071, Global Avg Loss: 0.68818685, Time: 0.0209 Steps: 92530, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000993, Sample Num: 15888, Cur Loss: 0.15528938, Cur Avg Loss: 0.18023340, Log Avg loss: 0.17452371, Global Avg Loss: 0.68813135, Time: 0.0209 Steps: 92540, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001003, Sample Num: 16048, Cur Loss: 0.27183220, Cur Avg Loss: 0.18029331, Log Avg loss: 0.18624269, Global Avg Loss: 0.68807712, Time: 0.0209 Steps: 92550, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001013, Sample Num: 16208, Cur Loss: 0.01996230, Cur Avg Loss: 0.18062745, Log Avg loss: 0.21414145, Global Avg Loss: 0.68802591, Time: 0.0209 Steps: 92560, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001023, Sample Num: 16368, Cur Loss: 0.05955083, Cur Avg Loss: 0.18122055, Log Avg loss: 0.24130123, Global Avg Loss: 0.68797766, Time: 0.0209 Steps: 92570, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001033, Sample Num: 16528, Cur Loss: 0.23090553, Cur Avg Loss: 0.18127459, Log Avg loss: 0.18680317, Global Avg Loss: 0.68792352, Time: 0.0208 Steps: 92580, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001043, Sample Num: 16688, Cur Loss: 0.07654478, Cur Avg Loss: 0.18078974, Log Avg loss: 0.13070430, Global Avg Loss: 0.68786334, Time: 0.0209 Steps: 92590, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001053, Sample Num: 16848, Cur Loss: 0.23397520, Cur Avg Loss: 0.18078468, Log Avg loss: 0.18025712, Global Avg Loss: 0.68780852, Time: 0.0208 Steps: 92600, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001063, Sample Num: 17008, Cur Loss: 0.20356736, Cur Avg Loss: 0.18076587, Log Avg loss: 0.17878488, Global Avg Loss: 0.68775356, Time: 0.0208 Steps: 92610, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001073, Sample Num: 17168, Cur Loss: 0.23862323, Cur Avg Loss: 0.18073087, Log Avg loss: 0.17701076, Global Avg Loss: 0.68769842, Time: 0.0209 Steps: 92620, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001083, Sample Num: 17328, Cur Loss: 0.06846395, Cur Avg Loss: 0.18079722, Log Avg loss: 0.18791688, Global Avg Loss: 0.68764446, Time: 0.0209 Steps: 92630, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001093, Sample Num: 17488, Cur Loss: 0.17973283, Cur Avg Loss: 0.18073583, Log Avg loss: 0.17408663, Global Avg Loss: 0.68758902, Time: 0.0209 Steps: 92640, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001103, Sample Num: 17648, Cur Loss: 0.15681103, Cur Avg Loss: 0.18027746, Log Avg loss: 0.13017852, Global Avg Loss: 0.68752886, Time: 0.0208 Steps: 92650, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001113, Sample Num: 17808, Cur Loss: 0.25948292, Cur Avg Loss: 0.18032220, Log Avg loss: 0.18525645, Global Avg Loss: 0.68747466, Time: 0.0209 Steps: 92660, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001123, Sample Num: 17968, Cur Loss: 0.13996926, Cur Avg Loss: 0.18082289, Log Avg loss: 0.23654920, Global Avg Loss: 0.68742600, Time: 0.0208 Steps: 92670, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001133, Sample Num: 18128, Cur Loss: 0.44365966, Cur Avg Loss: 0.18082389, Log Avg loss: 0.18093631, Global Avg Loss: 0.68737135, Time: 0.0208 Steps: 92680, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001143, Sample Num: 18288, Cur Loss: 0.17935893, Cur Avg Loss: 0.18074275, Log Avg loss: 0.17155042, Global Avg Loss: 0.68731570, Time: 0.0209 Steps: 92690, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001153, Sample Num: 18448, Cur Loss: 0.67282653, Cur Avg Loss: 0.18127338, Log Avg loss: 0.24192417, Global Avg Loss: 0.68726765, Time: 0.0208 Steps: 92700, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001163, Sample Num: 18608, Cur Loss: 0.12246601, Cur Avg Loss: 0.18168853, Log Avg loss: 0.22955472, Global Avg Loss: 0.68721828, Time: 0.0208 Steps: 92710, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001173, Sample Num: 18768, Cur Loss: 0.15269062, Cur Avg Loss: 0.18221008, Log Avg loss: 0.24286670, Global Avg Loss: 0.68717036, Time: 0.0209 Steps: 92720, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001183, Sample Num: 18928, Cur Loss: 0.08415745, Cur Avg Loss: 0.18240816, Log Avg loss: 0.20564273, Global Avg Loss: 0.68711843, Time: 0.0209 Steps: 92730, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001193, Sample Num: 19088, Cur Loss: 0.18434681, Cur Avg Loss: 0.18241667, Log Avg loss: 0.18342395, Global Avg Loss: 0.68706412, Time: 0.0208 Steps: 92740, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001203, Sample Num: 19248, Cur Loss: 0.17326847, Cur Avg Loss: 0.18246141, Log Avg loss: 0.18779870, Global Avg Loss: 0.68701029, Time: 0.0208 Steps: 92750, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001213, Sample Num: 19408, Cur Loss: 0.33938307, Cur Avg Loss: 0.18238814, Log Avg loss: 0.17357336, Global Avg Loss: 0.68695494, Time: 0.0208 Steps: 92760, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001223, Sample Num: 19568, Cur Loss: 0.09756000, Cur Avg Loss: 0.18220488, Log Avg loss: 0.15997601, Global Avg Loss: 0.68689813, Time: 0.0208 Steps: 92770, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001233, Sample Num: 19728, Cur Loss: 0.06762873, Cur Avg Loss: 0.18233334, Log Avg loss: 0.19804368, Global Avg Loss: 0.68684544, Time: 0.0208 Steps: 92780, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001243, Sample Num: 19888, Cur Loss: 0.38292173, Cur Avg Loss: 0.18247324, Log Avg loss: 0.19972358, Global Avg Loss: 0.68679294, Time: 0.0208 Steps: 92790, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001253, Sample Num: 20048, Cur Loss: 0.22410467, Cur Avg Loss: 0.18297100, Log Avg loss: 0.24484222, Global Avg Loss: 0.68674532, Time: 0.0208 Steps: 92800, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001263, Sample Num: 20208, Cur Loss: 0.05882486, Cur Avg Loss: 0.18227484, Log Avg loss: 0.09504543, Global Avg Loss: 0.68668157, Time: 0.0209 Steps: 92810, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001273, Sample Num: 20368, Cur Loss: 0.10171089, Cur Avg Loss: 0.18251964, Log Avg loss: 0.21343848, Global Avg Loss: 0.68663058, Time: 0.0208 Steps: 92820, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001283, Sample Num: 20528, Cur Loss: 0.15956916, Cur Avg Loss: 0.18215274, Log Avg loss: 0.13544612, Global Avg Loss: 0.68657120, Time: 0.0245 Steps: 92830, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001293, Sample Num: 20688, Cur Loss: 0.07445344, Cur Avg Loss: 0.18161965, Log Avg loss: 0.11322444, Global Avg Loss: 0.68650945, Time: 0.0208 Steps: 92840, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001303, Sample Num: 20848, Cur Loss: 0.29614264, Cur Avg Loss: 0.18134873, Log Avg loss: 0.14631815, Global Avg Loss: 0.68645127, Time: 0.0208 Steps: 92850, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001313, Sample Num: 21008, Cur Loss: 0.21206160, Cur Avg Loss: 0.18141216, Log Avg loss: 0.18967696, Global Avg Loss: 0.68639777, Time: 0.0208 Steps: 92860, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001323, Sample Num: 21168, Cur Loss: 0.23709261, Cur Avg Loss: 0.18159832, Log Avg loss: 0.20604184, Global Avg Loss: 0.68634605, Time: 0.0208 Steps: 92870, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001333, Sample Num: 21328, Cur Loss: 0.16317233, Cur Avg Loss: 0.18161308, Log Avg loss: 0.18356613, Global Avg Loss: 0.68629192, Time: 0.0208 Steps: 92880, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001343, Sample Num: 21488, Cur Loss: 0.17316401, Cur Avg Loss: 0.18250770, Log Avg loss: 0.30176056, Global Avg Loss: 0.68625052, Time: 0.0208 Steps: 92890, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001353, Sample Num: 21648, Cur Loss: 0.19996437, Cur Avg Loss: 0.18251634, Log Avg loss: 0.18367604, Global Avg Loss: 0.68619642, Time: 0.0209 Steps: 92900, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001363, Sample Num: 21808, Cur Loss: 0.09865571, Cur Avg Loss: 0.18255536, Log Avg loss: 0.18783510, Global Avg Loss: 0.68614278, Time: 0.0208 Steps: 92910, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001373, Sample Num: 21968, Cur Loss: 0.27245954, Cur Avg Loss: 0.18245439, Log Avg loss: 0.16869131, Global Avg Loss: 0.68608709, Time: 0.0208 Steps: 92920, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001383, Sample Num: 22128, Cur Loss: 0.15668149, Cur Avg Loss: 0.18308823, Log Avg loss: 0.27011473, Global Avg Loss: 0.68604233, Time: 0.0208 Steps: 92930, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001393, Sample Num: 22288, Cur Loss: 0.08883134, Cur Avg Loss: 0.18288660, Log Avg loss: 0.15500204, Global Avg Loss: 0.68598520, Time: 0.0208 Steps: 92940, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001403, Sample Num: 22448, Cur Loss: 0.16077779, Cur Avg Loss: 0.18268022, Log Avg loss: 0.15393116, Global Avg Loss: 0.68592795, Time: 0.0209 Steps: 92950, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001413, Sample Num: 22608, Cur Loss: 0.25844815, Cur Avg Loss: 0.18244967, Log Avg loss: 0.15010340, Global Avg Loss: 0.68587031, Time: 0.0208 Steps: 92960, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001423, Sample Num: 22768, Cur Loss: 0.06261669, Cur Avg Loss: 0.18244638, Log Avg loss: 0.18198075, Global Avg Loss: 0.68581611, Time: 0.0209 Steps: 92970, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001433, Sample Num: 22928, Cur Loss: 0.14947692, Cur Avg Loss: 0.18282210, Log Avg loss: 0.23628756, Global Avg Loss: 0.68576777, Time: 0.0208 Steps: 92980, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001443, Sample Num: 23088, Cur Loss: 0.16024056, Cur Avg Loss: 0.18261967, Log Avg loss: 0.15361119, Global Avg Loss: 0.68571054, Time: 0.0208 Steps: 92990, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001453, Sample Num: 23248, Cur Loss: 0.48128456, Cur Avg Loss: 0.18292496, Log Avg loss: 0.22697859, Global Avg Loss: 0.68566121, Time: 0.0208 Steps: 93000, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001463, Sample Num: 23408, Cur Loss: 0.22495045, Cur Avg Loss: 0.18329627, Log Avg loss: 0.23724817, Global Avg Loss: 0.68561300, Time: 0.0208 Steps: 93010, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001473, Sample Num: 23568, Cur Loss: 0.14707397, Cur Avg Loss: 0.18351118, Log Avg loss: 0.21495274, Global Avg Loss: 0.68556241, Time: 0.0208 Steps: 93020, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001483, Sample Num: 23728, Cur Loss: 0.06460790, Cur Avg Loss: 0.18324193, Log Avg loss: 0.14358007, Global Avg Loss: 0.68550415, Time: 0.0209 Steps: 93030, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001493, Sample Num: 23888, Cur Loss: 0.07285360, Cur Avg Loss: 0.18335017, Log Avg loss: 0.19940263, Global Avg Loss: 0.68545190, Time: 0.0209 Steps: 93040, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001503, Sample Num: 24048, Cur Loss: 0.18367642, Cur Avg Loss: 0.18327099, Log Avg loss: 0.17144990, Global Avg Loss: 0.68539666, Time: 0.0208 Steps: 93050, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001513, Sample Num: 24208, Cur Loss: 0.05915105, Cur Avg Loss: 0.18349352, Log Avg loss: 0.21693973, Global Avg Loss: 0.68534632, Time: 0.0209 Steps: 93060, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001523, Sample Num: 24368, Cur Loss: 0.08201852, Cur Avg Loss: 0.18326152, Log Avg loss: 0.14815939, Global Avg Loss: 0.68528860, Time: 0.0208 Steps: 93070, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001533, Sample Num: 24528, Cur Loss: 0.16052149, Cur Avg Loss: 0.18341842, Log Avg loss: 0.20731406, Global Avg Loss: 0.68523725, Time: 0.0209 Steps: 93080, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001543, Sample Num: 24688, Cur Loss: 0.05546845, Cur Avg Loss: 0.18341185, Log Avg loss: 0.18240564, Global Avg Loss: 0.68518324, Time: 0.0209 Steps: 93090, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001553, Sample Num: 24848, Cur Loss: 0.22230387, Cur Avg Loss: 0.18346475, Log Avg loss: 0.19162647, Global Avg Loss: 0.68513022, Time: 0.0209 Steps: 93100, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001563, Sample Num: 25008, Cur Loss: 0.39933202, Cur Avg Loss: 0.18425834, Log Avg loss: 0.30750272, Global Avg Loss: 0.68508967, Time: 0.0207 Steps: 93110, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001573, Sample Num: 25168, Cur Loss: 0.14362684, Cur Avg Loss: 0.18442549, Log Avg loss: 0.21055094, Global Avg Loss: 0.68503871, Time: 0.0212 Steps: 93120, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001583, Sample Num: 25328, Cur Loss: 0.20727426, Cur Avg Loss: 0.18433784, Log Avg loss: 0.17055173, Global Avg Loss: 0.68498346, Time: 0.0208 Steps: 93130, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001593, Sample Num: 25488, Cur Loss: 0.08589116, Cur Avg Loss: 0.18400771, Log Avg loss: 0.13174777, Global Avg Loss: 0.68492406, Time: 0.0207 Steps: 93140, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001603, Sample Num: 25648, Cur Loss: 0.28370097, Cur Avg Loss: 0.18411019, Log Avg loss: 0.20043422, Global Avg Loss: 0.68487205, Time: 0.0208 Steps: 93150, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001613, Sample Num: 25808, Cur Loss: 0.19030401, Cur Avg Loss: 0.18402884, Log Avg loss: 0.17098842, Global Avg Loss: 0.68481689, Time: 0.0212 Steps: 93160, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001623, Sample Num: 25968, Cur Loss: 0.22759104, Cur Avg Loss: 0.18407931, Log Avg loss: 0.19222054, Global Avg Loss: 0.68476402, Time: 0.0208 Steps: 93170, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001633, Sample Num: 26128, Cur Loss: 0.10942493, Cur Avg Loss: 0.18390509, Log Avg loss: 0.15562881, Global Avg Loss: 0.68470723, Time: 0.0208 Steps: 93180, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001643, Sample Num: 26288, Cur Loss: 0.25304449, Cur Avg Loss: 0.18390916, Log Avg loss: 0.18457434, Global Avg Loss: 0.68465357, Time: 0.0212 Steps: 93190, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001653, Sample Num: 26448, Cur Loss: 0.14973187, Cur Avg Loss: 0.18391552, Log Avg loss: 0.18496025, Global Avg Loss: 0.68459995, Time: 0.0209 Steps: 93200, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001663, Sample Num: 26608, Cur Loss: 0.30490214, Cur Avg Loss: 0.18379613, Log Avg loss: 0.16406172, Global Avg Loss: 0.68454410, Time: 0.0208 Steps: 93210, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001673, Sample Num: 26768, Cur Loss: 0.17437434, Cur Avg Loss: 0.18357916, Log Avg loss: 0.14749618, Global Avg Loss: 0.68448649, Time: 0.0209 Steps: 93220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001683, Sample Num: 26928, Cur Loss: 0.10258044, Cur Avg Loss: 0.18366414, Log Avg loss: 0.19788230, Global Avg Loss: 0.68443430, Time: 0.0209 Steps: 93230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001693, Sample Num: 27088, Cur Loss: 0.13352992, Cur Avg Loss: 0.18357700, Log Avg loss: 0.16891090, Global Avg Loss: 0.68437901, Time: 0.0212 Steps: 93240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001703, Sample Num: 27248, Cur Loss: 0.03473215, Cur Avg Loss: 0.18344155, Log Avg loss: 0.16051046, Global Avg Loss: 0.68432283, Time: 0.0212 Steps: 93250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001713, Sample Num: 27408, Cur Loss: 0.08700465, Cur Avg Loss: 0.18340372, Log Avg loss: 0.17696121, Global Avg Loss: 0.68426843, Time: 0.0215 Steps: 93260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001723, Sample Num: 27568, Cur Loss: 0.13132071, Cur Avg Loss: 0.18351020, Log Avg loss: 0.20175033, Global Avg Loss: 0.68421669, Time: 0.0209 Steps: 93270, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001733, Sample Num: 27728, Cur Loss: 0.19913800, Cur Avg Loss: 0.18358307, Log Avg loss: 0.19613721, Global Avg Loss: 0.68416437, Time: 0.0210 Steps: 93280, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001743, Sample Num: 27888, Cur Loss: 0.11310495, Cur Avg Loss: 0.18325510, Log Avg loss: 0.12641776, Global Avg Loss: 0.68410458, Time: 0.0213 Steps: 93290, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001753, Sample Num: 28048, Cur Loss: 0.41121066, Cur Avg Loss: 0.18316722, Log Avg loss: 0.16785087, Global Avg Loss: 0.68404925, Time: 0.0213 Steps: 93300, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001763, Sample Num: 28208, Cur Loss: 0.05318415, Cur Avg Loss: 0.18315999, Log Avg loss: 0.18189211, Global Avg Loss: 0.68399544, Time: 0.0209 Steps: 93310, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001773, Sample Num: 28368, Cur Loss: 0.35393029, Cur Avg Loss: 0.18326646, Log Avg loss: 0.20203762, Global Avg Loss: 0.68394379, Time: 0.0210 Steps: 93320, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001783, Sample Num: 28528, Cur Loss: 0.31759346, Cur Avg Loss: 0.18306233, Log Avg loss: 0.14686905, Global Avg Loss: 0.68388624, Time: 0.0212 Steps: 93330, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001793, Sample Num: 28688, Cur Loss: 0.38388422, Cur Avg Loss: 0.18335661, Log Avg loss: 0.23582813, Global Avg Loss: 0.68383824, Time: 0.0247 Steps: 93340, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001803, Sample Num: 28848, Cur Loss: 0.07119556, Cur Avg Loss: 0.18336880, Log Avg loss: 0.18555454, Global Avg Loss: 0.68378486, Time: 0.0211 Steps: 93350, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001813, Sample Num: 29008, Cur Loss: 0.15870120, Cur Avg Loss: 0.18339176, Log Avg loss: 0.18753100, Global Avg Loss: 0.68373171, Time: 0.0211 Steps: 93360, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001823, Sample Num: 29168, Cur Loss: 0.40361771, Cur Avg Loss: 0.18365927, Log Avg loss: 0.23215894, Global Avg Loss: 0.68368334, Time: 0.0211 Steps: 93370, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001833, Sample Num: 29328, Cur Loss: 0.36593932, Cur Avg Loss: 0.18382897, Log Avg loss: 0.21476511, Global Avg Loss: 0.68363313, Time: 0.0211 Steps: 93380, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001843, Sample Num: 29488, Cur Loss: 0.31544235, Cur Avg Loss: 0.18367932, Log Avg loss: 0.15624830, Global Avg Loss: 0.68357666, Time: 0.0211 Steps: 93390, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001853, Sample Num: 29648, Cur Loss: 0.06254500, Cur Avg Loss: 0.18389226, Log Avg loss: 0.22313773, Global Avg Loss: 0.68352736, Time: 0.0211 Steps: 93400, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001863, Sample Num: 29808, Cur Loss: 0.28449363, Cur Avg Loss: 0.18404159, Log Avg loss: 0.21171125, Global Avg Loss: 0.68347685, Time: 0.0211 Steps: 93410, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001873, Sample Num: 29968, Cur Loss: 0.43399948, Cur Avg Loss: 0.18422278, Log Avg loss: 0.21797879, Global Avg Loss: 0.68342702, Time: 0.0211 Steps: 93420, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001883, Sample Num: 30128, Cur Loss: 0.16067415, Cur Avg Loss: 0.18406763, Log Avg loss: 0.15500863, Global Avg Loss: 0.68337046, Time: 0.0212 Steps: 93430, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001893, Sample Num: 30288, Cur Loss: 0.09040255, Cur Avg Loss: 0.18390326, Log Avg loss: 0.15295193, Global Avg Loss: 0.68331370, Time: 0.0211 Steps: 93440, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001903, Sample Num: 30448, Cur Loss: 0.26856646, Cur Avg Loss: 0.18406034, Log Avg loss: 0.21379471, Global Avg Loss: 0.68326345, Time: 0.0211 Steps: 93450, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001913, Sample Num: 30608, Cur Loss: 0.19763497, Cur Avg Loss: 0.18406903, Log Avg loss: 0.18572355, Global Avg Loss: 0.68321022, Time: 0.0211 Steps: 93460, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001923, Sample Num: 30768, Cur Loss: 0.04456705, Cur Avg Loss: 0.18384119, Log Avg loss: 0.14025493, Global Avg Loss: 0.68315213, Time: 0.0211 Steps: 93470, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001933, Sample Num: 30928, Cur Loss: 0.53661382, Cur Avg Loss: 0.18388281, Log Avg loss: 0.19188744, Global Avg Loss: 0.68309958, Time: 0.0211 Steps: 93480, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001943, Sample Num: 31088, Cur Loss: 0.07672672, Cur Avg Loss: 0.18400779, Log Avg loss: 0.20816563, Global Avg Loss: 0.68304878, Time: 0.0211 Steps: 93490, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001953, Sample Num: 31248, Cur Loss: 0.21026309, Cur Avg Loss: 0.18377186, Log Avg loss: 0.13793033, Global Avg Loss: 0.68299048, Time: 0.0211 Steps: 93500, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001963, Sample Num: 31408, Cur Loss: 0.16920716, Cur Avg Loss: 0.18378941, Log Avg loss: 0.18721742, Global Avg Loss: 0.68293746, Time: 0.0211 Steps: 93510, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001973, Sample Num: 31568, Cur Loss: 0.12896445, Cur Avg Loss: 0.18418960, Log Avg loss: 0.26274731, Global Avg Loss: 0.68289253, Time: 0.0211 Steps: 93520, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001983, Sample Num: 31728, Cur Loss: 0.10326011, Cur Avg Loss: 0.18429055, Log Avg loss: 0.20420781, Global Avg Loss: 0.68284135, Time: 0.0211 Steps: 93530, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001993, Sample Num: 31888, Cur Loss: 0.15665737, Cur Avg Loss: 0.18402635, Log Avg loss: 0.13163606, Global Avg Loss: 0.68278242, Time: 0.0211 Steps: 93540, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002003, Sample Num: 32048, Cur Loss: 0.16150074, Cur Avg Loss: 0.18397104, Log Avg loss: 0.17294711, Global Avg Loss: 0.68272792, Time: 0.0211 Steps: 93550, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002013, Sample Num: 32208, Cur Loss: 0.36268467, Cur Avg Loss: 0.18419811, Log Avg loss: 0.22967930, Global Avg Loss: 0.68267950, Time: 0.0210 Steps: 93560, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002023, Sample Num: 32368, Cur Loss: 0.13147144, Cur Avg Loss: 0.18439324, Log Avg loss: 0.22367408, Global Avg Loss: 0.68263044, Time: 0.0211 Steps: 93570, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002033, Sample Num: 32528, Cur Loss: 0.16872610, Cur Avg Loss: 0.18439672, Log Avg loss: 0.18509937, Global Avg Loss: 0.68257728, Time: 0.0211 Steps: 93580, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002043, Sample Num: 32688, Cur Loss: 0.21540585, Cur Avg Loss: 0.18452198, Log Avg loss: 0.20998832, Global Avg Loss: 0.68252678, Time: 0.0211 Steps: 93590, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002053, Sample Num: 32848, Cur Loss: 0.12930582, Cur Avg Loss: 0.18442882, Log Avg loss: 0.16539609, Global Avg Loss: 0.68247153, Time: 0.0212 Steps: 93600, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002063, Sample Num: 33008, Cur Loss: 0.43194294, Cur Avg Loss: 0.18442812, Log Avg loss: 0.18428392, Global Avg Loss: 0.68241831, Time: 0.0211 Steps: 93610, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002073, Sample Num: 33168, Cur Loss: 0.16739964, Cur Avg Loss: 0.18434620, Log Avg loss: 0.16744637, Global Avg Loss: 0.68236331, Time: 0.0211 Steps: 93620, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002083, Sample Num: 33328, Cur Loss: 0.31635693, Cur Avg Loss: 0.18449127, Log Avg loss: 0.21456425, Global Avg Loss: 0.68231334, Time: 0.0211 Steps: 93630, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002093, Sample Num: 33488, Cur Loss: 0.35172665, Cur Avg Loss: 0.18441692, Log Avg loss: 0.16893073, Global Avg Loss: 0.68225852, Time: 0.0211 Steps: 93640, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002103, Sample Num: 33648, Cur Loss: 0.23196623, Cur Avg Loss: 0.18424855, Log Avg loss: 0.14900855, Global Avg Loss: 0.68220158, Time: 0.0211 Steps: 93650, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002113, Sample Num: 33808, Cur Loss: 0.04956558, Cur Avg Loss: 0.18409240, Log Avg loss: 0.15125390, Global Avg Loss: 0.68214489, Time: 0.0211 Steps: 93660, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002123, Sample Num: 33968, Cur Loss: 0.20900458, Cur Avg Loss: 0.18411439, Log Avg loss: 0.18876169, Global Avg Loss: 0.68209222, Time: 0.0211 Steps: 93670, Updated lr: 0.000012 ***** Running evaluation checkpoint-93676 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-93676 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.736298, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.310288, "eval_total_loss": 218.132415, "eval_mae": 0.451566, "eval_mse": 0.310279, "eval_r2": 0.802766, "eval_sp_statistic": 0.900067, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924898, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.305957, "test_total_loss": 153.590598, "test_mae": 0.359286, "test_mse": 0.306065, "test_r2": 0.802463, "test_sp_statistic": 0.873099, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.920459, "test_ps_pvalue": 0.0, "lr": 1.2113798008534852e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6820597256713214, "train_cur_epoch_loss": 391.92387106642127, "train_cur_epoch_avg_loss": 0.1840882438076192, "train_cur_epoch_time": 44.736297607421875, "train_cur_epoch_avg_time": 0.021012821797755696, "epoch": 44, "step": 93676} ################################################## Training, Epoch: 0045, Batch: 000004, Sample Num: 64, Cur Loss: 0.25001210, Cur Avg Loss: 0.15986463, Log Avg loss: 0.16884697, Global Avg Loss: 0.68203743, Time: 0.0232 Steps: 93680, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000014, Sample Num: 224, Cur Loss: 0.15489854, Cur Avg Loss: 0.21776726, Log Avg loss: 0.24092831, Global Avg Loss: 0.68199035, Time: 0.0211 Steps: 93690, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000024, Sample Num: 384, Cur Loss: 0.15520999, Cur Avg Loss: 0.19454500, Log Avg loss: 0.16203383, Global Avg Loss: 0.68193486, Time: 0.0211 Steps: 93700, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000034, Sample Num: 544, Cur Loss: 0.06356783, Cur Avg Loss: 0.18227905, Log Avg loss: 0.15284079, Global Avg Loss: 0.68187839, Time: 0.0212 Steps: 93710, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000044, Sample Num: 704, Cur Loss: 0.14620891, Cur Avg Loss: 0.20376608, Log Avg loss: 0.27682197, Global Avg Loss: 0.68183517, Time: 0.0212 Steps: 93720, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000054, Sample Num: 864, Cur Loss: 0.27359092, Cur Avg Loss: 0.19374102, Log Avg loss: 0.14963075, Global Avg Loss: 0.68177839, Time: 0.0211 Steps: 93730, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000064, Sample Num: 1024, Cur Loss: 0.11111113, Cur Avg Loss: 0.19189100, Log Avg loss: 0.18190089, Global Avg Loss: 0.68172507, Time: 0.0211 Steps: 93740, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000074, Sample Num: 1184, Cur Loss: 0.04528575, Cur Avg Loss: 0.19490459, Log Avg loss: 0.21419161, Global Avg Loss: 0.68167520, Time: 0.0212 Steps: 93750, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000084, Sample Num: 1344, Cur Loss: 0.04504748, Cur Avg Loss: 0.19696336, Log Avg loss: 0.21219818, Global Avg Loss: 0.68162513, Time: 0.0211 Steps: 93760, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000094, Sample Num: 1504, Cur Loss: 0.21775270, Cur Avg Loss: 0.19206933, Log Avg loss: 0.15095950, Global Avg Loss: 0.68156853, Time: 0.0211 Steps: 93770, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000104, Sample Num: 1664, Cur Loss: 0.14649697, Cur Avg Loss: 0.19120018, Log Avg loss: 0.18303014, Global Avg Loss: 0.68151537, Time: 0.0212 Steps: 93780, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000114, Sample Num: 1824, Cur Loss: 0.18451419, Cur Avg Loss: 0.19375087, Log Avg loss: 0.22027810, Global Avg Loss: 0.68146620, Time: 0.0211 Steps: 93790, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000124, Sample Num: 1984, Cur Loss: 0.08136242, Cur Avg Loss: 0.18958589, Log Avg loss: 0.14210515, Global Avg Loss: 0.68140869, Time: 0.0212 Steps: 93800, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000134, Sample Num: 2144, Cur Loss: 0.15062961, Cur Avg Loss: 0.19368838, Log Avg loss: 0.24455917, Global Avg Loss: 0.68136213, Time: 0.0211 Steps: 93810, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000144, Sample Num: 2304, Cur Loss: 0.27966052, Cur Avg Loss: 0.19280827, Log Avg loss: 0.18101482, Global Avg Loss: 0.68130880, Time: 0.0212 Steps: 93820, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000154, Sample Num: 2464, Cur Loss: 0.14753141, Cur Avg Loss: 0.19501312, Log Avg loss: 0.22676296, Global Avg Loss: 0.68126035, Time: 0.0211 Steps: 93830, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000164, Sample Num: 2624, Cur Loss: 0.12700626, Cur Avg Loss: 0.19590179, Log Avg loss: 0.20958725, Global Avg Loss: 0.68121009, Time: 0.0211 Steps: 93840, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000174, Sample Num: 2784, Cur Loss: 0.12992895, Cur Avg Loss: 0.19213988, Log Avg loss: 0.13044458, Global Avg Loss: 0.68115140, Time: 0.0211 Steps: 93850, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000184, Sample Num: 2944, Cur Loss: 0.04453743, Cur Avg Loss: 0.18693810, Log Avg loss: 0.09642713, Global Avg Loss: 0.68108911, Time: 0.0211 Steps: 93860, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000194, Sample Num: 3104, Cur Loss: 0.23960224, Cur Avg Loss: 0.18559108, Log Avg loss: 0.16080592, Global Avg Loss: 0.68103368, Time: 0.0211 Steps: 93870, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000204, Sample Num: 3264, Cur Loss: 0.50252759, Cur Avg Loss: 0.18573736, Log Avg loss: 0.18857527, Global Avg Loss: 0.68098122, Time: 0.0211 Steps: 93880, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000214, Sample Num: 3424, Cur Loss: 0.19844316, Cur Avg Loss: 0.18428933, Log Avg loss: 0.15474943, Global Avg Loss: 0.68092518, Time: 0.0212 Steps: 93890, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000224, Sample Num: 3584, Cur Loss: 0.10394695, Cur Avg Loss: 0.18296351, Log Avg loss: 0.15459095, Global Avg Loss: 0.68086912, Time: 0.0211 Steps: 93900, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000234, Sample Num: 3744, Cur Loss: 0.06809649, Cur Avg Loss: 0.17933033, Log Avg loss: 0.09794728, Global Avg Loss: 0.68080705, Time: 0.0212 Steps: 93910, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000244, Sample Num: 3904, Cur Loss: 0.07062667, Cur Avg Loss: 0.17924312, Log Avg loss: 0.17720233, Global Avg Loss: 0.68075343, Time: 0.0211 Steps: 93920, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000254, Sample Num: 4064, Cur Loss: 0.11448410, Cur Avg Loss: 0.18137288, Log Avg loss: 0.23333895, Global Avg Loss: 0.68070580, Time: 0.0212 Steps: 93930, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000264, Sample Num: 4224, Cur Loss: 0.21010123, Cur Avg Loss: 0.18345412, Log Avg loss: 0.23631766, Global Avg Loss: 0.68065849, Time: 0.0212 Steps: 93940, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000274, Sample Num: 4384, Cur Loss: 0.12191930, Cur Avg Loss: 0.18275921, Log Avg loss: 0.16441351, Global Avg Loss: 0.68060354, Time: 0.0211 Steps: 93950, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000284, Sample Num: 4544, Cur Loss: 0.10957405, Cur Avg Loss: 0.18374887, Log Avg loss: 0.21086553, Global Avg Loss: 0.68055355, Time: 0.0212 Steps: 93960, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000294, Sample Num: 4704, Cur Loss: 0.12014100, Cur Avg Loss: 0.18352063, Log Avg loss: 0.17703877, Global Avg Loss: 0.68049997, Time: 0.0211 Steps: 93970, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000304, Sample Num: 4864, Cur Loss: 0.18023816, Cur Avg Loss: 0.18156580, Log Avg loss: 0.12409384, Global Avg Loss: 0.68044076, Time: 0.0212 Steps: 93980, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000314, Sample Num: 5024, Cur Loss: 0.35432601, Cur Avg Loss: 0.18372299, Log Avg loss: 0.24930136, Global Avg Loss: 0.68039489, Time: 0.0210 Steps: 93990, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000324, Sample Num: 5184, Cur Loss: 0.08072668, Cur Avg Loss: 0.18387691, Log Avg loss: 0.18871007, Global Avg Loss: 0.68034258, Time: 0.0212 Steps: 94000, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000334, Sample Num: 5344, Cur Loss: 0.15361941, Cur Avg Loss: 0.18592347, Log Avg loss: 0.25223198, Global Avg Loss: 0.68029705, Time: 0.0211 Steps: 94010, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000344, Sample Num: 5504, Cur Loss: 0.06100696, Cur Avg Loss: 0.18516802, Log Avg loss: 0.15993595, Global Avg Loss: 0.68024170, Time: 0.0211 Steps: 94020, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000354, Sample Num: 5664, Cur Loss: 0.10858719, Cur Avg Loss: 0.18437918, Log Avg loss: 0.15724307, Global Avg Loss: 0.68018608, Time: 0.0212 Steps: 94030, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000364, Sample Num: 5824, Cur Loss: 0.09912275, Cur Avg Loss: 0.18220260, Log Avg loss: 0.10515180, Global Avg Loss: 0.68012493, Time: 0.0212 Steps: 94040, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000374, Sample Num: 5984, Cur Loss: 0.24158090, Cur Avg Loss: 0.18267903, Log Avg loss: 0.20002093, Global Avg Loss: 0.68007388, Time: 0.0211 Steps: 94050, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000384, Sample Num: 6144, Cur Loss: 0.27958024, Cur Avg Loss: 0.18239069, Log Avg loss: 0.17160674, Global Avg Loss: 0.68001983, Time: 0.0212 Steps: 94060, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000394, Sample Num: 6304, Cur Loss: 0.25660884, Cur Avg Loss: 0.18183784, Log Avg loss: 0.16060844, Global Avg Loss: 0.67996461, Time: 0.0212 Steps: 94070, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000404, Sample Num: 6464, Cur Loss: 0.07059388, Cur Avg Loss: 0.18126078, Log Avg loss: 0.15852471, Global Avg Loss: 0.67990919, Time: 0.0211 Steps: 94080, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000414, Sample Num: 6624, Cur Loss: 0.33633795, Cur Avg Loss: 0.18440931, Log Avg loss: 0.31160999, Global Avg Loss: 0.67987004, Time: 0.0211 Steps: 94090, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000424, Sample Num: 6784, Cur Loss: 0.06269162, Cur Avg Loss: 0.18298706, Log Avg loss: 0.12410603, Global Avg Loss: 0.67981098, Time: 0.0212 Steps: 94100, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000434, Sample Num: 6944, Cur Loss: 0.10578486, Cur Avg Loss: 0.18234882, Log Avg loss: 0.15528729, Global Avg Loss: 0.67975525, Time: 0.0212 Steps: 94110, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000444, Sample Num: 7104, Cur Loss: 0.23919570, Cur Avg Loss: 0.18166987, Log Avg loss: 0.15220342, Global Avg Loss: 0.67969920, Time: 0.0211 Steps: 94120, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000454, Sample Num: 7264, Cur Loss: 0.08239503, Cur Avg Loss: 0.18133066, Log Avg loss: 0.16626973, Global Avg Loss: 0.67964465, Time: 0.0211 Steps: 94130, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000464, Sample Num: 7424, Cur Loss: 0.23455130, Cur Avg Loss: 0.18192546, Log Avg loss: 0.20892933, Global Avg Loss: 0.67959465, Time: 0.0211 Steps: 94140, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000474, Sample Num: 7584, Cur Loss: 0.02326588, Cur Avg Loss: 0.18182922, Log Avg loss: 0.17736389, Global Avg Loss: 0.67954131, Time: 0.0211 Steps: 94150, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000484, Sample Num: 7744, Cur Loss: 0.16555819, Cur Avg Loss: 0.18197120, Log Avg loss: 0.18870078, Global Avg Loss: 0.67948918, Time: 0.0212 Steps: 94160, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000494, Sample Num: 7904, Cur Loss: 0.21794640, Cur Avg Loss: 0.18190752, Log Avg loss: 0.17882557, Global Avg Loss: 0.67943601, Time: 0.0211 Steps: 94170, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000504, Sample Num: 8064, Cur Loss: 0.17221569, Cur Avg Loss: 0.18146924, Log Avg loss: 0.15981811, Global Avg Loss: 0.67938084, Time: 0.0211 Steps: 94180, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000514, Sample Num: 8224, Cur Loss: 0.15832108, Cur Avg Loss: 0.18100307, Log Avg loss: 0.15750827, Global Avg Loss: 0.67932543, Time: 0.0250 Steps: 94190, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000524, Sample Num: 8384, Cur Loss: 0.10070767, Cur Avg Loss: 0.18148882, Log Avg loss: 0.20645640, Global Avg Loss: 0.67927523, Time: 0.0212 Steps: 94200, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000534, Sample Num: 8544, Cur Loss: 0.11693211, Cur Avg Loss: 0.17990094, Log Avg loss: 0.09669574, Global Avg Loss: 0.67921340, Time: 0.0212 Steps: 94210, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000544, Sample Num: 8704, Cur Loss: 0.31847030, Cur Avg Loss: 0.17927487, Log Avg loss: 0.14584308, Global Avg Loss: 0.67915679, Time: 0.0211 Steps: 94220, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000554, Sample Num: 8864, Cur Loss: 0.11093441, Cur Avg Loss: 0.17859564, Log Avg loss: 0.14164529, Global Avg Loss: 0.67909974, Time: 0.0211 Steps: 94230, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000564, Sample Num: 9024, Cur Loss: 0.25222069, Cur Avg Loss: 0.17859077, Log Avg loss: 0.17832080, Global Avg Loss: 0.67904660, Time: 0.0210 Steps: 94240, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000574, Sample Num: 9184, Cur Loss: 0.10819773, Cur Avg Loss: 0.17811680, Log Avg loss: 0.15138539, Global Avg Loss: 0.67899062, Time: 0.0211 Steps: 94250, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000584, Sample Num: 9344, Cur Loss: 0.06214267, Cur Avg Loss: 0.17817415, Log Avg loss: 0.18146604, Global Avg Loss: 0.67893784, Time: 0.0212 Steps: 94260, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000594, Sample Num: 9504, Cur Loss: 0.09066179, Cur Avg Loss: 0.17838810, Log Avg loss: 0.19088222, Global Avg Loss: 0.67888607, Time: 0.0211 Steps: 94270, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000604, Sample Num: 9664, Cur Loss: 0.19247785, Cur Avg Loss: 0.17835621, Log Avg loss: 0.17646225, Global Avg Loss: 0.67883277, Time: 0.0212 Steps: 94280, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000614, Sample Num: 9824, Cur Loss: 0.11447577, Cur Avg Loss: 0.17812487, Log Avg loss: 0.16415173, Global Avg Loss: 0.67877819, Time: 0.0211 Steps: 94290, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000624, Sample Num: 9984, Cur Loss: 0.13690764, Cur Avg Loss: 0.17778982, Log Avg loss: 0.15721782, Global Avg Loss: 0.67872288, Time: 0.0212 Steps: 94300, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000634, Sample Num: 10144, Cur Loss: 0.17692646, Cur Avg Loss: 0.17745935, Log Avg loss: 0.15683837, Global Avg Loss: 0.67866754, Time: 0.0213 Steps: 94310, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000644, Sample Num: 10304, Cur Loss: 0.15307437, Cur Avg Loss: 0.17782398, Log Avg loss: 0.20094111, Global Avg Loss: 0.67861689, Time: 0.0211 Steps: 94320, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000654, Sample Num: 10464, Cur Loss: 0.14328025, Cur Avg Loss: 0.17737216, Log Avg loss: 0.14827510, Global Avg Loss: 0.67856067, Time: 0.0212 Steps: 94330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000664, Sample Num: 10624, Cur Loss: 0.06607064, Cur Avg Loss: 0.17724658, Log Avg loss: 0.16903373, Global Avg Loss: 0.67850666, Time: 0.0211 Steps: 94340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000674, Sample Num: 10784, Cur Loss: 0.08051746, Cur Avg Loss: 0.17746743, Log Avg loss: 0.19213182, Global Avg Loss: 0.67845511, Time: 0.0212 Steps: 94350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000684, Sample Num: 10944, Cur Loss: 0.16162729, Cur Avg Loss: 0.17861896, Log Avg loss: 0.25623212, Global Avg Loss: 0.67841037, Time: 0.0211 Steps: 94360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000694, Sample Num: 11104, Cur Loss: 0.12683460, Cur Avg Loss: 0.17805828, Log Avg loss: 0.13970749, Global Avg Loss: 0.67835328, Time: 0.0212 Steps: 94370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000704, Sample Num: 11264, Cur Loss: 0.20868127, Cur Avg Loss: 0.17914438, Log Avg loss: 0.25452001, Global Avg Loss: 0.67830838, Time: 0.0211 Steps: 94380, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000714, Sample Num: 11424, Cur Loss: 0.14159229, Cur Avg Loss: 0.17949836, Log Avg loss: 0.20441819, Global Avg Loss: 0.67825817, Time: 0.0211 Steps: 94390, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000724, Sample Num: 11584, Cur Loss: 0.13919991, Cur Avg Loss: 0.17909958, Log Avg loss: 0.15062684, Global Avg Loss: 0.67820228, Time: 0.0211 Steps: 94400, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000734, Sample Num: 11744, Cur Loss: 0.30280495, Cur Avg Loss: 0.17917372, Log Avg loss: 0.18454158, Global Avg Loss: 0.67814999, Time: 0.0211 Steps: 94410, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000744, Sample Num: 11904, Cur Loss: 0.43793520, Cur Avg Loss: 0.17938735, Log Avg loss: 0.19506763, Global Avg Loss: 0.67809882, Time: 0.0211 Steps: 94420, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000754, Sample Num: 12064, Cur Loss: 0.13581781, Cur Avg Loss: 0.17885011, Log Avg loss: 0.13887969, Global Avg Loss: 0.67804172, Time: 0.0211 Steps: 94430, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000764, Sample Num: 12224, Cur Loss: 0.13784145, Cur Avg Loss: 0.17932394, Log Avg loss: 0.21505089, Global Avg Loss: 0.67799270, Time: 0.0211 Steps: 94440, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000774, Sample Num: 12384, Cur Loss: 0.11714543, Cur Avg Loss: 0.17907737, Log Avg loss: 0.16023909, Global Avg Loss: 0.67793788, Time: 0.0221 Steps: 94450, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000784, Sample Num: 12544, Cur Loss: 0.47546130, Cur Avg Loss: 0.17908522, Log Avg loss: 0.17969290, Global Avg Loss: 0.67788513, Time: 0.0220 Steps: 94460, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000794, Sample Num: 12704, Cur Loss: 0.38550833, Cur Avg Loss: 0.17995542, Log Avg loss: 0.24817917, Global Avg Loss: 0.67783965, Time: 0.0220 Steps: 94470, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000804, Sample Num: 12864, Cur Loss: 0.05348332, Cur Avg Loss: 0.17971277, Log Avg loss: 0.16044638, Global Avg Loss: 0.67778488, Time: 0.0220 Steps: 94480, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000814, Sample Num: 13024, Cur Loss: 0.24583943, Cur Avg Loss: 0.17949998, Log Avg loss: 0.16239127, Global Avg Loss: 0.67773034, Time: 0.0220 Steps: 94490, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000824, Sample Num: 13184, Cur Loss: 0.28662816, Cur Avg Loss: 0.17943858, Log Avg loss: 0.17444136, Global Avg Loss: 0.67767708, Time: 0.0220 Steps: 94500, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000834, Sample Num: 13344, Cur Loss: 0.11476793, Cur Avg Loss: 0.17953722, Log Avg loss: 0.18766481, Global Avg Loss: 0.67762523, Time: 0.0220 Steps: 94510, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000844, Sample Num: 13504, Cur Loss: 0.05261941, Cur Avg Loss: 0.17914173, Log Avg loss: 0.14615739, Global Avg Loss: 0.67756901, Time: 0.0220 Steps: 94520, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000854, Sample Num: 13664, Cur Loss: 0.10855524, Cur Avg Loss: 0.17941372, Log Avg loss: 0.20236992, Global Avg Loss: 0.67751874, Time: 0.0220 Steps: 94530, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000864, Sample Num: 13824, Cur Loss: 0.06752148, Cur Avg Loss: 0.17953529, Log Avg loss: 0.18991730, Global Avg Loss: 0.67746716, Time: 0.0219 Steps: 94540, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000874, Sample Num: 13984, Cur Loss: 0.12192423, Cur Avg Loss: 0.17915956, Log Avg loss: 0.14669689, Global Avg Loss: 0.67741102, Time: 0.0219 Steps: 94550, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000884, Sample Num: 14144, Cur Loss: 0.17363325, Cur Avg Loss: 0.17924645, Log Avg loss: 0.18684055, Global Avg Loss: 0.67735914, Time: 0.0220 Steps: 94560, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000894, Sample Num: 14304, Cur Loss: 0.62490952, Cur Avg Loss: 0.17950619, Log Avg loss: 0.20246723, Global Avg Loss: 0.67730893, Time: 0.0219 Steps: 94570, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000904, Sample Num: 14464, Cur Loss: 0.05038746, Cur Avg Loss: 0.17886807, Log Avg loss: 0.12182013, Global Avg Loss: 0.67725020, Time: 0.0219 Steps: 94580, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000914, Sample Num: 14624, Cur Loss: 0.17836913, Cur Avg Loss: 0.17889423, Log Avg loss: 0.18125938, Global Avg Loss: 0.67719776, Time: 0.0219 Steps: 94590, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000924, Sample Num: 14784, Cur Loss: 0.08604118, Cur Avg Loss: 0.17855197, Log Avg loss: 0.14726943, Global Avg Loss: 0.67714174, Time: 0.0220 Steps: 94600, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000934, Sample Num: 14944, Cur Loss: 0.13052383, Cur Avg Loss: 0.17844127, Log Avg loss: 0.16821201, Global Avg Loss: 0.67708795, Time: 0.0219 Steps: 94610, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000944, Sample Num: 15104, Cur Loss: 0.04333684, Cur Avg Loss: 0.17882146, Log Avg loss: 0.21433172, Global Avg Loss: 0.67703904, Time: 0.0219 Steps: 94620, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000954, Sample Num: 15264, Cur Loss: 0.04950318, Cur Avg Loss: 0.17870177, Log Avg loss: 0.16740251, Global Avg Loss: 0.67698519, Time: 0.0219 Steps: 94630, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000964, Sample Num: 15424, Cur Loss: 0.14219585, Cur Avg Loss: 0.17829475, Log Avg loss: 0.13946519, Global Avg Loss: 0.67692839, Time: 0.0219 Steps: 94640, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000974, Sample Num: 15584, Cur Loss: 0.24492185, Cur Avg Loss: 0.17897484, Log Avg loss: 0.24453533, Global Avg Loss: 0.67688271, Time: 0.0219 Steps: 94650, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000984, Sample Num: 15744, Cur Loss: 0.16359514, Cur Avg Loss: 0.17883674, Log Avg loss: 0.16538600, Global Avg Loss: 0.67682867, Time: 0.0220 Steps: 94660, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000994, Sample Num: 15904, Cur Loss: 0.06712766, Cur Avg Loss: 0.17896749, Log Avg loss: 0.19183356, Global Avg Loss: 0.67677744, Time: 0.0219 Steps: 94670, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001004, Sample Num: 16064, Cur Loss: 0.19192211, Cur Avg Loss: 0.17911295, Log Avg loss: 0.19357102, Global Avg Loss: 0.67672641, Time: 0.0219 Steps: 94680, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001014, Sample Num: 16224, Cur Loss: 0.11815191, Cur Avg Loss: 0.17898322, Log Avg loss: 0.16595918, Global Avg Loss: 0.67667247, Time: 0.0220 Steps: 94690, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001024, Sample Num: 16384, Cur Loss: 0.07697621, Cur Avg Loss: 0.17908618, Log Avg loss: 0.18952579, Global Avg Loss: 0.67662103, Time: 0.0262 Steps: 94700, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001034, Sample Num: 16544, Cur Loss: 0.28275436, Cur Avg Loss: 0.17968258, Log Avg loss: 0.24075439, Global Avg Loss: 0.67657500, Time: 0.0210 Steps: 94710, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001044, Sample Num: 16704, Cur Loss: 0.24545769, Cur Avg Loss: 0.18066038, Log Avg loss: 0.28176494, Global Avg Loss: 0.67653332, Time: 0.0210 Steps: 94720, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001054, Sample Num: 16864, Cur Loss: 0.27094030, Cur Avg Loss: 0.18036267, Log Avg loss: 0.14928187, Global Avg Loss: 0.67647766, Time: 0.0209 Steps: 94730, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001064, Sample Num: 17024, Cur Loss: 0.20516926, Cur Avg Loss: 0.18016421, Log Avg loss: 0.15924636, Global Avg Loss: 0.67642307, Time: 0.0210 Steps: 94740, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001074, Sample Num: 17184, Cur Loss: 0.10512600, Cur Avg Loss: 0.18041239, Log Avg loss: 0.20681844, Global Avg Loss: 0.67637351, Time: 0.0209 Steps: 94750, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001084, Sample Num: 17344, Cur Loss: 0.20281711, Cur Avg Loss: 0.18025043, Log Avg loss: 0.16285542, Global Avg Loss: 0.67631932, Time: 0.0209 Steps: 94760, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001094, Sample Num: 17504, Cur Loss: 0.39536372, Cur Avg Loss: 0.18021238, Log Avg loss: 0.17608830, Global Avg Loss: 0.67626653, Time: 0.0210 Steps: 94770, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001104, Sample Num: 17664, Cur Loss: 0.08537744, Cur Avg Loss: 0.18007902, Log Avg loss: 0.16548976, Global Avg Loss: 0.67621264, Time: 0.0209 Steps: 94780, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001114, Sample Num: 17824, Cur Loss: 0.19630539, Cur Avg Loss: 0.18012507, Log Avg loss: 0.18520897, Global Avg Loss: 0.67616084, Time: 0.0209 Steps: 94790, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001124, Sample Num: 17984, Cur Loss: 0.37387565, Cur Avg Loss: 0.18103105, Log Avg loss: 0.28195655, Global Avg Loss: 0.67611926, Time: 0.0210 Steps: 94800, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001134, Sample Num: 18144, Cur Loss: 0.11897475, Cur Avg Loss: 0.18066154, Log Avg loss: 0.13912852, Global Avg Loss: 0.67606262, Time: 0.0210 Steps: 94810, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001144, Sample Num: 18304, Cur Loss: 0.07050954, Cur Avg Loss: 0.18059485, Log Avg loss: 0.17303245, Global Avg Loss: 0.67600957, Time: 0.0209 Steps: 94820, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001154, Sample Num: 18464, Cur Loss: 0.22965731, Cur Avg Loss: 0.18088822, Log Avg loss: 0.21445008, Global Avg Loss: 0.67596090, Time: 0.0210 Steps: 94830, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001164, Sample Num: 18624, Cur Loss: 0.28945395, Cur Avg Loss: 0.18114542, Log Avg loss: 0.21082621, Global Avg Loss: 0.67591185, Time: 0.0209 Steps: 94840, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001174, Sample Num: 18784, Cur Loss: 0.09322859, Cur Avg Loss: 0.18059365, Log Avg loss: 0.11636733, Global Avg Loss: 0.67585286, Time: 0.0209 Steps: 94850, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001184, Sample Num: 18944, Cur Loss: 0.25499788, Cur Avg Loss: 0.18041056, Log Avg loss: 0.15891637, Global Avg Loss: 0.67579837, Time: 0.0210 Steps: 94860, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001194, Sample Num: 19104, Cur Loss: 0.10790509, Cur Avg Loss: 0.18031015, Log Avg loss: 0.16842112, Global Avg Loss: 0.67574488, Time: 0.0209 Steps: 94870, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001204, Sample Num: 19264, Cur Loss: 0.07525991, Cur Avg Loss: 0.18009862, Log Avg loss: 0.15484151, Global Avg Loss: 0.67568998, Time: 0.0210 Steps: 94880, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001214, Sample Num: 19424, Cur Loss: 0.25052452, Cur Avg Loss: 0.17994831, Log Avg loss: 0.16185132, Global Avg Loss: 0.67563583, Time: 0.0210 Steps: 94890, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001224, Sample Num: 19584, Cur Loss: 0.11375485, Cur Avg Loss: 0.18027666, Log Avg loss: 0.22013853, Global Avg Loss: 0.67558783, Time: 0.0210 Steps: 94900, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001234, Sample Num: 19744, Cur Loss: 0.12520671, Cur Avg Loss: 0.18044699, Log Avg loss: 0.20129545, Global Avg Loss: 0.67553786, Time: 0.0209 Steps: 94910, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001244, Sample Num: 19904, Cur Loss: 0.24216849, Cur Avg Loss: 0.18007353, Log Avg loss: 0.13398800, Global Avg Loss: 0.67548081, Time: 0.0210 Steps: 94920, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001254, Sample Num: 20064, Cur Loss: 0.35833654, Cur Avg Loss: 0.18044526, Log Avg loss: 0.22668869, Global Avg Loss: 0.67543353, Time: 0.0210 Steps: 94930, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001264, Sample Num: 20224, Cur Loss: 0.20597227, Cur Avg Loss: 0.18033512, Log Avg loss: 0.16652343, Global Avg Loss: 0.67537993, Time: 0.0209 Steps: 94940, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001274, Sample Num: 20384, Cur Loss: 0.25065881, Cur Avg Loss: 0.18026722, Log Avg loss: 0.17168551, Global Avg Loss: 0.67532688, Time: 0.0210 Steps: 94950, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001284, Sample Num: 20544, Cur Loss: 0.22698097, Cur Avg Loss: 0.18043210, Log Avg loss: 0.20143726, Global Avg Loss: 0.67527698, Time: 0.0247 Steps: 94960, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001294, Sample Num: 20704, Cur Loss: 0.10806105, Cur Avg Loss: 0.18073848, Log Avg loss: 0.22007836, Global Avg Loss: 0.67522905, Time: 0.0209 Steps: 94970, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001304, Sample Num: 20864, Cur Loss: 0.05077420, Cur Avg Loss: 0.18059140, Log Avg loss: 0.16155889, Global Avg Loss: 0.67517496, Time: 0.0210 Steps: 94980, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001314, Sample Num: 21024, Cur Loss: 0.10052655, Cur Avg Loss: 0.18038851, Log Avg loss: 0.15393179, Global Avg Loss: 0.67512009, Time: 0.0210 Steps: 94990, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001324, Sample Num: 21184, Cur Loss: 0.24984583, Cur Avg Loss: 0.18057247, Log Avg loss: 0.20474420, Global Avg Loss: 0.67507058, Time: 0.0209 Steps: 95000, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001334, Sample Num: 21344, Cur Loss: 0.06450479, Cur Avg Loss: 0.18107955, Log Avg loss: 0.24821755, Global Avg Loss: 0.67502565, Time: 0.0209 Steps: 95010, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001344, Sample Num: 21504, Cur Loss: 0.34509644, Cur Avg Loss: 0.18147991, Log Avg loss: 0.23488714, Global Avg Loss: 0.67497933, Time: 0.0210 Steps: 95020, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001354, Sample Num: 21664, Cur Loss: 0.19100454, Cur Avg Loss: 0.18131202, Log Avg loss: 0.15874751, Global Avg Loss: 0.67492501, Time: 0.0210 Steps: 95030, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001364, Sample Num: 21824, Cur Loss: 0.12738921, Cur Avg Loss: 0.18143062, Log Avg loss: 0.19748914, Global Avg Loss: 0.67487477, Time: 0.0210 Steps: 95040, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001374, Sample Num: 21984, Cur Loss: 0.05745686, Cur Avg Loss: 0.18166806, Log Avg loss: 0.21405507, Global Avg Loss: 0.67482629, Time: 0.0209 Steps: 95050, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001384, Sample Num: 22144, Cur Loss: 0.11134443, Cur Avg Loss: 0.18166985, Log Avg loss: 0.18191626, Global Avg Loss: 0.67477444, Time: 0.0210 Steps: 95060, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001394, Sample Num: 22304, Cur Loss: 0.24260420, Cur Avg Loss: 0.18155100, Log Avg loss: 0.16510131, Global Avg Loss: 0.67472083, Time: 0.0209 Steps: 95070, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001404, Sample Num: 22464, Cur Loss: 0.14775094, Cur Avg Loss: 0.18132355, Log Avg loss: 0.14961772, Global Avg Loss: 0.67466560, Time: 0.0210 Steps: 95080, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001414, Sample Num: 22624, Cur Loss: 0.15674295, Cur Avg Loss: 0.18149001, Log Avg loss: 0.20486093, Global Avg Loss: 0.67461619, Time: 0.0210 Steps: 95090, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001424, Sample Num: 22784, Cur Loss: 0.23458862, Cur Avg Loss: 0.18174138, Log Avg loss: 0.21728537, Global Avg Loss: 0.67456810, Time: 0.0210 Steps: 95100, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001434, Sample Num: 22944, Cur Loss: 0.02398595, Cur Avg Loss: 0.18133339, Log Avg loss: 0.12323510, Global Avg Loss: 0.67451014, Time: 0.0210 Steps: 95110, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001444, Sample Num: 23104, Cur Loss: 0.11904064, Cur Avg Loss: 0.18126455, Log Avg loss: 0.17139356, Global Avg Loss: 0.67445724, Time: 0.0210 Steps: 95120, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001454, Sample Num: 23264, Cur Loss: 0.38533652, Cur Avg Loss: 0.18173459, Log Avg loss: 0.24960771, Global Avg Loss: 0.67441258, Time: 0.0209 Steps: 95130, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001464, Sample Num: 23424, Cur Loss: 0.01552378, Cur Avg Loss: 0.18154970, Log Avg loss: 0.15466725, Global Avg Loss: 0.67435795, Time: 0.0209 Steps: 95140, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001474, Sample Num: 23584, Cur Loss: 0.10426709, Cur Avg Loss: 0.18174303, Log Avg loss: 0.21004609, Global Avg Loss: 0.67430915, Time: 0.0209 Steps: 95150, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001484, Sample Num: 23744, Cur Loss: 0.31471461, Cur Avg Loss: 0.18145477, Log Avg loss: 0.13896533, Global Avg Loss: 0.67425290, Time: 0.0209 Steps: 95160, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001494, Sample Num: 23904, Cur Loss: 0.21741039, Cur Avg Loss: 0.18171913, Log Avg loss: 0.22094958, Global Avg Loss: 0.67420527, Time: 0.0209 Steps: 95170, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001504, Sample Num: 24064, Cur Loss: 0.44227892, Cur Avg Loss: 0.18158558, Log Avg loss: 0.16163435, Global Avg Loss: 0.67415141, Time: 0.0210 Steps: 95180, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001514, Sample Num: 24224, Cur Loss: 0.08947638, Cur Avg Loss: 0.18198789, Log Avg loss: 0.24249475, Global Avg Loss: 0.67410607, Time: 0.0209 Steps: 95190, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001524, Sample Num: 24384, Cur Loss: 0.21298434, Cur Avg Loss: 0.18158092, Log Avg loss: 0.11996627, Global Avg Loss: 0.67404786, Time: 0.0210 Steps: 95200, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001534, Sample Num: 24544, Cur Loss: 0.27099895, Cur Avg Loss: 0.18197258, Log Avg loss: 0.24166093, Global Avg Loss: 0.67400245, Time: 0.0210 Steps: 95210, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001544, Sample Num: 24704, Cur Loss: 0.16690505, Cur Avg Loss: 0.18176800, Log Avg loss: 0.15038517, Global Avg Loss: 0.67394745, Time: 0.0209 Steps: 95220, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001554, Sample Num: 24864, Cur Loss: 0.14970914, Cur Avg Loss: 0.18171470, Log Avg loss: 0.17348537, Global Avg Loss: 0.67389490, Time: 0.0209 Steps: 95230, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001564, Sample Num: 25024, Cur Loss: 0.21639033, Cur Avg Loss: 0.18145025, Log Avg loss: 0.14035566, Global Avg Loss: 0.67383888, Time: 0.0209 Steps: 95240, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001574, Sample Num: 25184, Cur Loss: 0.19278687, Cur Avg Loss: 0.18143373, Log Avg loss: 0.17884926, Global Avg Loss: 0.67378691, Time: 0.0209 Steps: 95250, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001584, Sample Num: 25344, Cur Loss: 0.19978917, Cur Avg Loss: 0.18145757, Log Avg loss: 0.18520938, Global Avg Loss: 0.67373563, Time: 0.0209 Steps: 95260, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001594, Sample Num: 25504, Cur Loss: 0.35062885, Cur Avg Loss: 0.18164505, Log Avg loss: 0.21134283, Global Avg Loss: 0.67368709, Time: 0.0209 Steps: 95270, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001604, Sample Num: 25664, Cur Loss: 0.17742835, Cur Avg Loss: 0.18188202, Log Avg loss: 0.21965442, Global Avg Loss: 0.67363944, Time: 0.0209 Steps: 95280, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001614, Sample Num: 25824, Cur Loss: 0.30715793, Cur Avg Loss: 0.18191362, Log Avg loss: 0.18698314, Global Avg Loss: 0.67358837, Time: 0.0209 Steps: 95290, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001624, Sample Num: 25984, Cur Loss: 0.29541582, Cur Avg Loss: 0.18185635, Log Avg loss: 0.17261309, Global Avg Loss: 0.67353580, Time: 0.0209 Steps: 95300, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001634, Sample Num: 26144, Cur Loss: 0.26576275, Cur Avg Loss: 0.18213326, Log Avg loss: 0.22710323, Global Avg Loss: 0.67348896, Time: 0.0209 Steps: 95310, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001644, Sample Num: 26304, Cur Loss: 0.35723853, Cur Avg Loss: 0.18221114, Log Avg loss: 0.19493555, Global Avg Loss: 0.67343875, Time: 0.0209 Steps: 95320, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001654, Sample Num: 26464, Cur Loss: 0.02886013, Cur Avg Loss: 0.18221010, Log Avg loss: 0.18203971, Global Avg Loss: 0.67338721, Time: 0.0209 Steps: 95330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001664, Sample Num: 26624, Cur Loss: 0.20730953, Cur Avg Loss: 0.18225650, Log Avg loss: 0.18993039, Global Avg Loss: 0.67333650, Time: 0.0209 Steps: 95340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001674, Sample Num: 26784, Cur Loss: 0.15376164, Cur Avg Loss: 0.18225852, Log Avg loss: 0.18259527, Global Avg Loss: 0.67328503, Time: 0.0210 Steps: 95350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001684, Sample Num: 26944, Cur Loss: 0.06883566, Cur Avg Loss: 0.18195735, Log Avg loss: 0.13154088, Global Avg Loss: 0.67322822, Time: 0.0209 Steps: 95360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001694, Sample Num: 27104, Cur Loss: 0.14300449, Cur Avg Loss: 0.18201658, Log Avg loss: 0.19199188, Global Avg Loss: 0.67317776, Time: 0.0209 Steps: 95370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001704, Sample Num: 27264, Cur Loss: 0.03172047, Cur Avg Loss: 0.18179342, Log Avg loss: 0.14398940, Global Avg Loss: 0.67312228, Time: 0.0209 Steps: 95380, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001714, Sample Num: 27424, Cur Loss: 0.10470907, Cur Avg Loss: 0.18187475, Log Avg loss: 0.19573302, Global Avg Loss: 0.67307223, Time: 0.0209 Steps: 95390, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001724, Sample Num: 27584, Cur Loss: 0.14227678, Cur Avg Loss: 0.18186929, Log Avg loss: 0.18093346, Global Avg Loss: 0.67302064, Time: 0.0209 Steps: 95400, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001734, Sample Num: 27744, Cur Loss: 0.21839358, Cur Avg Loss: 0.18159554, Log Avg loss: 0.13440107, Global Avg Loss: 0.67296419, Time: 0.0210 Steps: 95410, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001744, Sample Num: 27904, Cur Loss: 0.18688621, Cur Avg Loss: 0.18151499, Log Avg loss: 0.16754779, Global Avg Loss: 0.67291122, Time: 0.0209 Steps: 95420, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001754, Sample Num: 28064, Cur Loss: 0.16586789, Cur Avg Loss: 0.18152365, Log Avg loss: 0.18303415, Global Avg Loss: 0.67285989, Time: 0.0209 Steps: 95430, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001764, Sample Num: 28224, Cur Loss: 0.24498568, Cur Avg Loss: 0.18188075, Log Avg loss: 0.24451727, Global Avg Loss: 0.67281501, Time: 0.0210 Steps: 95440, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001774, Sample Num: 28384, Cur Loss: 0.22537902, Cur Avg Loss: 0.18194795, Log Avg loss: 0.19380063, Global Avg Loss: 0.67276482, Time: 0.0209 Steps: 95450, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001784, Sample Num: 28544, Cur Loss: 0.12661149, Cur Avg Loss: 0.18190033, Log Avg loss: 0.17345361, Global Avg Loss: 0.67271252, Time: 0.0209 Steps: 95460, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001794, Sample Num: 28704, Cur Loss: 0.17905292, Cur Avg Loss: 0.18206433, Log Avg loss: 0.21132128, Global Avg Loss: 0.67266419, Time: 0.0248 Steps: 95470, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001804, Sample Num: 28864, Cur Loss: 0.09959704, Cur Avg Loss: 0.18170547, Log Avg loss: 0.11732543, Global Avg Loss: 0.67260603, Time: 0.0209 Steps: 95480, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001814, Sample Num: 29024, Cur Loss: 0.17932461, Cur Avg Loss: 0.18159333, Log Avg loss: 0.16136317, Global Avg Loss: 0.67255249, Time: 0.0210 Steps: 95490, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001824, Sample Num: 29184, Cur Loss: 0.08099671, Cur Avg Loss: 0.18125935, Log Avg loss: 0.12067612, Global Avg Loss: 0.67249470, Time: 0.0210 Steps: 95500, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001834, Sample Num: 29344, Cur Loss: 0.12756807, Cur Avg Loss: 0.18115203, Log Avg loss: 0.16157682, Global Avg Loss: 0.67244121, Time: 0.0209 Steps: 95510, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001844, Sample Num: 29504, Cur Loss: 0.17458713, Cur Avg Loss: 0.18169348, Log Avg loss: 0.28099637, Global Avg Loss: 0.67240023, Time: 0.0210 Steps: 95520, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001854, Sample Num: 29664, Cur Loss: 0.09885388, Cur Avg Loss: 0.18166264, Log Avg loss: 0.17597520, Global Avg Loss: 0.67234826, Time: 0.0209 Steps: 95530, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001864, Sample Num: 29824, Cur Loss: 0.11669830, Cur Avg Loss: 0.18177437, Log Avg loss: 0.20248861, Global Avg Loss: 0.67229908, Time: 0.0209 Steps: 95540, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001874, Sample Num: 29984, Cur Loss: 0.52235806, Cur Avg Loss: 0.18200204, Log Avg loss: 0.22444089, Global Avg Loss: 0.67225221, Time: 0.0210 Steps: 95550, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001884, Sample Num: 30144, Cur Loss: 0.20688990, Cur Avg Loss: 0.18184196, Log Avg loss: 0.15184137, Global Avg Loss: 0.67219775, Time: 0.0209 Steps: 95560, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001894, Sample Num: 30304, Cur Loss: 0.33041623, Cur Avg Loss: 0.18196968, Log Avg loss: 0.20603255, Global Avg Loss: 0.67214897, Time: 0.0209 Steps: 95570, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001904, Sample Num: 30464, Cur Loss: 0.36059269, Cur Avg Loss: 0.18191734, Log Avg loss: 0.17200509, Global Avg Loss: 0.67209665, Time: 0.0209 Steps: 95580, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001914, Sample Num: 30624, Cur Loss: 0.08374663, Cur Avg Loss: 0.18198387, Log Avg loss: 0.19465035, Global Avg Loss: 0.67204670, Time: 0.0210 Steps: 95590, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001924, Sample Num: 30784, Cur Loss: 0.33060083, Cur Avg Loss: 0.18232813, Log Avg loss: 0.24822032, Global Avg Loss: 0.67200237, Time: 0.0209 Steps: 95600, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001934, Sample Num: 30944, Cur Loss: 0.36853188, Cur Avg Loss: 0.18220085, Log Avg loss: 0.15771190, Global Avg Loss: 0.67194858, Time: 0.0210 Steps: 95610, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001944, Sample Num: 31104, Cur Loss: 0.38826737, Cur Avg Loss: 0.18236074, Log Avg loss: 0.21328262, Global Avg Loss: 0.67190061, Time: 0.0209 Steps: 95620, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001954, Sample Num: 31264, Cur Loss: 0.24190925, Cur Avg Loss: 0.18215774, Log Avg loss: 0.14269418, Global Avg Loss: 0.67184527, Time: 0.0209 Steps: 95630, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001964, Sample Num: 31424, Cur Loss: 0.20987082, Cur Avg Loss: 0.18209848, Log Avg loss: 0.17051953, Global Avg Loss: 0.67179285, Time: 0.0209 Steps: 95640, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001974, Sample Num: 31584, Cur Loss: 0.09070121, Cur Avg Loss: 0.18227239, Log Avg loss: 0.21642898, Global Avg Loss: 0.67174524, Time: 0.0209 Steps: 95650, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001984, Sample Num: 31744, Cur Loss: 0.28045857, Cur Avg Loss: 0.18232746, Log Avg loss: 0.19319715, Global Avg Loss: 0.67169522, Time: 0.0210 Steps: 95660, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001994, Sample Num: 31904, Cur Loss: 0.19420534, Cur Avg Loss: 0.18255047, Log Avg loss: 0.22679681, Global Avg Loss: 0.67164871, Time: 0.0210 Steps: 95670, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002004, Sample Num: 32064, Cur Loss: 0.04997227, Cur Avg Loss: 0.18243952, Log Avg loss: 0.16031538, Global Avg Loss: 0.67159527, Time: 0.0210 Steps: 95680, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002014, Sample Num: 32224, Cur Loss: 0.48163238, Cur Avg Loss: 0.18290981, Log Avg loss: 0.27715595, Global Avg Loss: 0.67155405, Time: 0.0209 Steps: 95690, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002024, Sample Num: 32384, Cur Loss: 0.08237928, Cur Avg Loss: 0.18265796, Log Avg loss: 0.13193602, Global Avg Loss: 0.67149767, Time: 0.0209 Steps: 95700, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002034, Sample Num: 32544, Cur Loss: 0.20259982, Cur Avg Loss: 0.18255967, Log Avg loss: 0.16266607, Global Avg Loss: 0.67144450, Time: 0.0210 Steps: 95710, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002044, Sample Num: 32704, Cur Loss: 0.14391153, Cur Avg Loss: 0.18309720, Log Avg loss: 0.29243056, Global Avg Loss: 0.67140491, Time: 0.0210 Steps: 95720, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002054, Sample Num: 32864, Cur Loss: 0.06235529, Cur Avg Loss: 0.18283053, Log Avg loss: 0.12832380, Global Avg Loss: 0.67134817, Time: 0.0247 Steps: 95730, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002064, Sample Num: 33024, Cur Loss: 0.27889347, Cur Avg Loss: 0.18284790, Log Avg loss: 0.18641485, Global Avg Loss: 0.67129752, Time: 0.0209 Steps: 95740, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002074, Sample Num: 33184, Cur Loss: 0.13768716, Cur Avg Loss: 0.18278837, Log Avg loss: 0.17050170, Global Avg Loss: 0.67124522, Time: 0.0209 Steps: 95750, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002084, Sample Num: 33344, Cur Loss: 0.15000588, Cur Avg Loss: 0.18263664, Log Avg loss: 0.15116686, Global Avg Loss: 0.67119091, Time: 0.0209 Steps: 95760, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002094, Sample Num: 33504, Cur Loss: 0.15007593, Cur Avg Loss: 0.18254610, Log Avg loss: 0.16367824, Global Avg Loss: 0.67113792, Time: 0.0209 Steps: 95770, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002104, Sample Num: 33664, Cur Loss: 0.15779480, Cur Avg Loss: 0.18267724, Log Avg loss: 0.21013856, Global Avg Loss: 0.67108979, Time: 0.0209 Steps: 95780, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002114, Sample Num: 33824, Cur Loss: 0.19656312, Cur Avg Loss: 0.18251650, Log Avg loss: 0.14869689, Global Avg Loss: 0.67103525, Time: 0.0209 Steps: 95790, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002124, Sample Num: 33984, Cur Loss: 0.19962513, Cur Avg Loss: 0.18234708, Log Avg loss: 0.14653039, Global Avg Loss: 0.67098050, Time: 0.0209 Steps: 95800, Updated lr: 0.000010 ***** Running evaluation checkpoint-95805 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-95805 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.044142, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.342344, "eval_total_loss": 240.667482, "eval_mae": 0.483353, "eval_mse": 0.342334, "eval_r2": 0.78239, "eval_sp_statistic": 0.901231, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925728, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.287991, "test_total_loss": 144.571327, "test_mae": 0.386008, "test_mse": 0.288082, "test_r2": 0.814069, "test_sp_statistic": 0.875988, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.922799, "test_ps_pvalue": 0.0, "lr": 1.0094831673779043e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6709538469627172, "train_cur_epoch_loss": 388.10644627641886, "train_cur_epoch_avg_loss": 0.18229518378413287, "train_cur_epoch_time": 45.04414176940918, "train_cur_epoch_avg_time": 0.02115741745862338, "epoch": 45, "step": 95805} ################################################## Training, Epoch: 0046, Batch: 000005, Sample Num: 80, Cur Loss: 0.40721461, Cur Avg Loss: 0.24525813, Log Avg loss: 0.20275436, Global Avg Loss: 0.67093163, Time: 0.0221 Steps: 95810, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000015, Sample Num: 240, Cur Loss: 0.37107545, Cur Avg Loss: 0.20223750, Log Avg loss: 0.18072719, Global Avg Loss: 0.67088047, Time: 0.0209 Steps: 95820, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000025, Sample Num: 400, Cur Loss: 0.39390546, Cur Avg Loss: 0.18172051, Log Avg loss: 0.15094501, Global Avg Loss: 0.67082622, Time: 0.0209 Steps: 95830, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000035, Sample Num: 560, Cur Loss: 0.09084553, Cur Avg Loss: 0.17741809, Log Avg loss: 0.16666207, Global Avg Loss: 0.67077361, Time: 0.0209 Steps: 95840, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000045, Sample Num: 720, Cur Loss: 0.24778727, Cur Avg Loss: 0.18936781, Log Avg loss: 0.23119179, Global Avg Loss: 0.67072775, Time: 0.0209 Steps: 95850, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000055, Sample Num: 880, Cur Loss: 0.48826420, Cur Avg Loss: 0.18128901, Log Avg loss: 0.14493445, Global Avg Loss: 0.67067290, Time: 0.0209 Steps: 95860, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000065, Sample Num: 1040, Cur Loss: 0.06540497, Cur Avg Loss: 0.17972172, Log Avg loss: 0.17110164, Global Avg Loss: 0.67062079, Time: 0.0209 Steps: 95870, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000075, Sample Num: 1200, Cur Loss: 0.08944447, Cur Avg Loss: 0.17435070, Log Avg loss: 0.13943903, Global Avg Loss: 0.67056539, Time: 0.0209 Steps: 95880, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000085, Sample Num: 1360, Cur Loss: 0.26494715, Cur Avg Loss: 0.19026193, Log Avg loss: 0.30959613, Global Avg Loss: 0.67052775, Time: 0.0209 Steps: 95890, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000095, Sample Num: 1520, Cur Loss: 0.13411190, Cur Avg Loss: 0.19065553, Log Avg loss: 0.19400112, Global Avg Loss: 0.67047806, Time: 0.0210 Steps: 95900, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000105, Sample Num: 1680, Cur Loss: 0.13922599, Cur Avg Loss: 0.18797293, Log Avg loss: 0.16248825, Global Avg Loss: 0.67042509, Time: 0.0209 Steps: 95910, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000115, Sample Num: 1840, Cur Loss: 0.09777978, Cur Avg Loss: 0.18007891, Log Avg loss: 0.09719176, Global Avg Loss: 0.67036533, Time: 0.0209 Steps: 95920, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000125, Sample Num: 2000, Cur Loss: 0.12067275, Cur Avg Loss: 0.17999523, Log Avg loss: 0.17903283, Global Avg Loss: 0.67031411, Time: 0.0209 Steps: 95930, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000135, Sample Num: 2160, Cur Loss: 0.20834947, Cur Avg Loss: 0.17925886, Log Avg loss: 0.17005434, Global Avg Loss: 0.67026197, Time: 0.0209 Steps: 95940, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000145, Sample Num: 2320, Cur Loss: 0.21955165, Cur Avg Loss: 0.17941964, Log Avg loss: 0.18159012, Global Avg Loss: 0.67021104, Time: 0.0209 Steps: 95950, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000155, Sample Num: 2480, Cur Loss: 0.06625178, Cur Avg Loss: 0.17983424, Log Avg loss: 0.18584590, Global Avg Loss: 0.67016056, Time: 0.0209 Steps: 95960, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000165, Sample Num: 2640, Cur Loss: 0.09027550, Cur Avg Loss: 0.17691534, Log Avg loss: 0.13167248, Global Avg Loss: 0.67010445, Time: 0.0209 Steps: 95970, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000175, Sample Num: 2800, Cur Loss: 0.19594753, Cur Avg Loss: 0.17646377, Log Avg loss: 0.16901275, Global Avg Loss: 0.67005224, Time: 0.0209 Steps: 95980, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000185, Sample Num: 2960, Cur Loss: 0.11104457, Cur Avg Loss: 0.17736891, Log Avg loss: 0.19320894, Global Avg Loss: 0.67000257, Time: 0.0209 Steps: 95990, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000195, Sample Num: 3120, Cur Loss: 0.05706017, Cur Avg Loss: 0.17433326, Log Avg loss: 0.11817373, Global Avg Loss: 0.66994509, Time: 0.0209 Steps: 96000, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000205, Sample Num: 3280, Cur Loss: 0.11584888, Cur Avg Loss: 0.17288551, Log Avg loss: 0.14465433, Global Avg Loss: 0.66989037, Time: 0.0209 Steps: 96010, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000215, Sample Num: 3440, Cur Loss: 0.40046763, Cur Avg Loss: 0.17190041, Log Avg loss: 0.15170582, Global Avg Loss: 0.66983641, Time: 0.0209 Steps: 96020, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000225, Sample Num: 3600, Cur Loss: 0.13090298, Cur Avg Loss: 0.17661345, Log Avg loss: 0.27794378, Global Avg Loss: 0.66979560, Time: 0.0209 Steps: 96030, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000235, Sample Num: 3760, Cur Loss: 0.06954124, Cur Avg Loss: 0.17851738, Log Avg loss: 0.22135596, Global Avg Loss: 0.66974891, Time: 0.0209 Steps: 96040, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000245, Sample Num: 3920, Cur Loss: 0.21225156, Cur Avg Loss: 0.17788560, Log Avg loss: 0.16303882, Global Avg Loss: 0.66969615, Time: 0.0209 Steps: 96050, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000255, Sample Num: 4080, Cur Loss: 0.24150972, Cur Avg Loss: 0.17717049, Log Avg loss: 0.15965007, Global Avg Loss: 0.66964305, Time: 0.0209 Steps: 96060, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000265, Sample Num: 4240, Cur Loss: 0.18099338, Cur Avg Loss: 0.17532954, Log Avg loss: 0.12838533, Global Avg Loss: 0.66958671, Time: 0.0210 Steps: 96070, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000275, Sample Num: 4400, Cur Loss: 0.06172555, Cur Avg Loss: 0.17614458, Log Avg loss: 0.19774332, Global Avg Loss: 0.66953760, Time: 0.0209 Steps: 96080, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000285, Sample Num: 4560, Cur Loss: 0.18885306, Cur Avg Loss: 0.17619321, Log Avg loss: 0.17753049, Global Avg Loss: 0.66948640, Time: 0.0209 Steps: 96090, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000295, Sample Num: 4720, Cur Loss: 0.05716853, Cur Avg Loss: 0.17676093, Log Avg loss: 0.19294096, Global Avg Loss: 0.66943681, Time: 0.0209 Steps: 96100, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000305, Sample Num: 4880, Cur Loss: 0.23360187, Cur Avg Loss: 0.17683075, Log Avg loss: 0.17889048, Global Avg Loss: 0.66938577, Time: 0.0209 Steps: 96110, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000315, Sample Num: 5040, Cur Loss: 0.11525284, Cur Avg Loss: 0.17564209, Log Avg loss: 0.13938788, Global Avg Loss: 0.66933063, Time: 0.0209 Steps: 96120, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000325, Sample Num: 5200, Cur Loss: 0.21564098, Cur Avg Loss: 0.17506821, Log Avg loss: 0.15699088, Global Avg Loss: 0.66927734, Time: 0.0210 Steps: 96130, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000335, Sample Num: 5360, Cur Loss: 0.32548320, Cur Avg Loss: 0.17581425, Log Avg loss: 0.20006065, Global Avg Loss: 0.66922853, Time: 0.0209 Steps: 96140, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000345, Sample Num: 5520, Cur Loss: 0.15149164, Cur Avg Loss: 0.17560675, Log Avg loss: 0.16865558, Global Avg Loss: 0.66917647, Time: 0.0209 Steps: 96150, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000355, Sample Num: 5680, Cur Loss: 0.65298027, Cur Avg Loss: 0.17610454, Log Avg loss: 0.19327817, Global Avg Loss: 0.66912698, Time: 0.0209 Steps: 96160, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000365, Sample Num: 5840, Cur Loss: 0.06010373, Cur Avg Loss: 0.17689084, Log Avg loss: 0.20480457, Global Avg Loss: 0.66907870, Time: 0.0209 Steps: 96170, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000375, Sample Num: 6000, Cur Loss: 0.24506567, Cur Avg Loss: 0.17562428, Log Avg loss: 0.12939476, Global Avg Loss: 0.66902259, Time: 0.0209 Steps: 96180, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000385, Sample Num: 6160, Cur Loss: 0.16767392, Cur Avg Loss: 0.17596468, Log Avg loss: 0.18872970, Global Avg Loss: 0.66897266, Time: 0.0209 Steps: 96190, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000395, Sample Num: 6320, Cur Loss: 0.04247465, Cur Avg Loss: 0.17761790, Log Avg loss: 0.24126704, Global Avg Loss: 0.66892820, Time: 0.0210 Steps: 96200, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000405, Sample Num: 6480, Cur Loss: 0.15406038, Cur Avg Loss: 0.17897215, Log Avg loss: 0.23246499, Global Avg Loss: 0.66888283, Time: 0.0209 Steps: 96210, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000415, Sample Num: 6640, Cur Loss: 0.07676953, Cur Avg Loss: 0.17900054, Log Avg loss: 0.18015006, Global Avg Loss: 0.66883204, Time: 0.0209 Steps: 96220, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000425, Sample Num: 6800, Cur Loss: 0.14419827, Cur Avg Loss: 0.17802456, Log Avg loss: 0.13752174, Global Avg Loss: 0.66877682, Time: 0.0210 Steps: 96230, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000435, Sample Num: 6960, Cur Loss: 0.11776287, Cur Avg Loss: 0.17798470, Log Avg loss: 0.17629041, Global Avg Loss: 0.66872565, Time: 0.0209 Steps: 96240, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000445, Sample Num: 7120, Cur Loss: 0.26525131, Cur Avg Loss: 0.17763117, Log Avg loss: 0.16225283, Global Avg Loss: 0.66867303, Time: 0.0209 Steps: 96250, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000455, Sample Num: 7280, Cur Loss: 0.13339972, Cur Avg Loss: 0.17814723, Log Avg loss: 0.20111184, Global Avg Loss: 0.66862446, Time: 0.0209 Steps: 96260, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000465, Sample Num: 7440, Cur Loss: 0.15206370, Cur Avg Loss: 0.17833826, Log Avg loss: 0.18703024, Global Avg Loss: 0.66857443, Time: 0.0209 Steps: 96270, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000475, Sample Num: 7600, Cur Loss: 0.19012794, Cur Avg Loss: 0.17708332, Log Avg loss: 0.11872838, Global Avg Loss: 0.66851732, Time: 0.0209 Steps: 96280, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000485, Sample Num: 7760, Cur Loss: 0.41180786, Cur Avg Loss: 0.17842903, Log Avg loss: 0.24235007, Global Avg Loss: 0.66847306, Time: 0.0209 Steps: 96290, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000495, Sample Num: 7920, Cur Loss: 0.16141510, Cur Avg Loss: 0.17825012, Log Avg loss: 0.16957334, Global Avg Loss: 0.66842126, Time: 0.0209 Steps: 96300, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000505, Sample Num: 8080, Cur Loss: 0.16870248, Cur Avg Loss: 0.17858661, Log Avg loss: 0.19524258, Global Avg Loss: 0.66837213, Time: 0.0209 Steps: 96310, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000515, Sample Num: 8240, Cur Loss: 0.14688849, Cur Avg Loss: 0.17852518, Log Avg loss: 0.17542330, Global Avg Loss: 0.66832095, Time: 0.0245 Steps: 96320, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000525, Sample Num: 8400, Cur Loss: 0.47913802, Cur Avg Loss: 0.17881103, Log Avg loss: 0.19353220, Global Avg Loss: 0.66827166, Time: 0.0209 Steps: 96330, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000535, Sample Num: 8560, Cur Loss: 0.31428143, Cur Avg Loss: 0.17833544, Log Avg loss: 0.15336682, Global Avg Loss: 0.66821821, Time: 0.0209 Steps: 96340, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000545, Sample Num: 8720, Cur Loss: 0.15647748, Cur Avg Loss: 0.17854412, Log Avg loss: 0.18970828, Global Avg Loss: 0.66816855, Time: 0.0208 Steps: 96350, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000555, Sample Num: 8880, Cur Loss: 0.09203160, Cur Avg Loss: 0.17795369, Log Avg loss: 0.14577558, Global Avg Loss: 0.66811434, Time: 0.0208 Steps: 96360, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000565, Sample Num: 9040, Cur Loss: 0.08605930, Cur Avg Loss: 0.17645046, Log Avg loss: 0.09302106, Global Avg Loss: 0.66805466, Time: 0.0209 Steps: 96370, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000575, Sample Num: 9200, Cur Loss: 0.13893692, Cur Avg Loss: 0.17726389, Log Avg loss: 0.22322259, Global Avg Loss: 0.66800851, Time: 0.0209 Steps: 96380, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000585, Sample Num: 9360, Cur Loss: 0.28401184, Cur Avg Loss: 0.17744461, Log Avg loss: 0.18783636, Global Avg Loss: 0.66795869, Time: 0.0209 Steps: 96390, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000595, Sample Num: 9520, Cur Loss: 0.15846950, Cur Avg Loss: 0.17717519, Log Avg loss: 0.16141371, Global Avg Loss: 0.66790615, Time: 0.0209 Steps: 96400, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000605, Sample Num: 9680, Cur Loss: 0.11988188, Cur Avg Loss: 0.17645058, Log Avg loss: 0.13333668, Global Avg Loss: 0.66785070, Time: 0.0209 Steps: 96410, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000615, Sample Num: 9840, Cur Loss: 0.08020981, Cur Avg Loss: 0.17537020, Log Avg loss: 0.11000726, Global Avg Loss: 0.66779284, Time: 0.0208 Steps: 96420, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000625, Sample Num: 10000, Cur Loss: 0.58503222, Cur Avg Loss: 0.17707941, Log Avg loss: 0.28219586, Global Avg Loss: 0.66775286, Time: 0.0209 Steps: 96430, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000635, Sample Num: 10160, Cur Loss: 0.26466486, Cur Avg Loss: 0.17692926, Log Avg loss: 0.16754491, Global Avg Loss: 0.66770099, Time: 0.0208 Steps: 96440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000645, Sample Num: 10320, Cur Loss: 0.12823761, Cur Avg Loss: 0.17649847, Log Avg loss: 0.14914284, Global Avg Loss: 0.66764722, Time: 0.0208 Steps: 96450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000655, Sample Num: 10480, Cur Loss: 0.03862400, Cur Avg Loss: 0.17572466, Log Avg loss: 0.12581392, Global Avg Loss: 0.66759105, Time: 0.0209 Steps: 96460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000665, Sample Num: 10640, Cur Loss: 0.36496323, Cur Avg Loss: 0.17646548, Log Avg loss: 0.22498965, Global Avg Loss: 0.66754517, Time: 0.0208 Steps: 96470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000675, Sample Num: 10800, Cur Loss: 0.57566845, Cur Avg Loss: 0.17699462, Log Avg loss: 0.21218239, Global Avg Loss: 0.66749798, Time: 0.0208 Steps: 96480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000685, Sample Num: 10960, Cur Loss: 0.21489727, Cur Avg Loss: 0.17705676, Log Avg loss: 0.18125076, Global Avg Loss: 0.66744758, Time: 0.0208 Steps: 96490, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000695, Sample Num: 11120, Cur Loss: 0.06856216, Cur Avg Loss: 0.17645619, Log Avg loss: 0.13531760, Global Avg Loss: 0.66739244, Time: 0.0208 Steps: 96500, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000705, Sample Num: 11280, Cur Loss: 0.28241587, Cur Avg Loss: 0.17639213, Log Avg loss: 0.17193986, Global Avg Loss: 0.66734110, Time: 0.0208 Steps: 96510, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000715, Sample Num: 11440, Cur Loss: 0.10612050, Cur Avg Loss: 0.17622412, Log Avg loss: 0.16437955, Global Avg Loss: 0.66728899, Time: 0.0208 Steps: 96520, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000725, Sample Num: 11600, Cur Loss: 0.10947142, Cur Avg Loss: 0.17541513, Log Avg loss: 0.11757212, Global Avg Loss: 0.66723204, Time: 0.0208 Steps: 96530, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000735, Sample Num: 11760, Cur Loss: 0.09912387, Cur Avg Loss: 0.17553788, Log Avg loss: 0.18443690, Global Avg Loss: 0.66718203, Time: 0.0208 Steps: 96540, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000745, Sample Num: 11920, Cur Loss: 0.02919075, Cur Avg Loss: 0.17645969, Log Avg loss: 0.24421300, Global Avg Loss: 0.66713823, Time: 0.0208 Steps: 96550, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000755, Sample Num: 12080, Cur Loss: 0.26288220, Cur Avg Loss: 0.17684160, Log Avg loss: 0.20529405, Global Avg Loss: 0.66709040, Time: 0.0208 Steps: 96560, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000765, Sample Num: 12240, Cur Loss: 0.14210400, Cur Avg Loss: 0.17672509, Log Avg loss: 0.16792834, Global Avg Loss: 0.66703871, Time: 0.0208 Steps: 96570, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000775, Sample Num: 12400, Cur Loss: 0.19066337, Cur Avg Loss: 0.17738843, Log Avg loss: 0.22813394, Global Avg Loss: 0.66699326, Time: 0.0210 Steps: 96580, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000785, Sample Num: 12560, Cur Loss: 0.28814805, Cur Avg Loss: 0.17751425, Log Avg loss: 0.18726557, Global Avg Loss: 0.66694360, Time: 0.0209 Steps: 96590, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000795, Sample Num: 12720, Cur Loss: 0.06471428, Cur Avg Loss: 0.17682393, Log Avg loss: 0.12263326, Global Avg Loss: 0.66688725, Time: 0.0209 Steps: 96600, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000805, Sample Num: 12880, Cur Loss: 0.22246256, Cur Avg Loss: 0.17708113, Log Avg loss: 0.19752892, Global Avg Loss: 0.66683867, Time: 0.0209 Steps: 96610, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000815, Sample Num: 13040, Cur Loss: 0.08268587, Cur Avg Loss: 0.17678841, Log Avg loss: 0.15322444, Global Avg Loss: 0.66678551, Time: 0.0209 Steps: 96620, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000825, Sample Num: 13200, Cur Loss: 0.15007074, Cur Avg Loss: 0.17662527, Log Avg loss: 0.16332940, Global Avg Loss: 0.66673341, Time: 0.0209 Steps: 96630, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000835, Sample Num: 13360, Cur Loss: 0.04281908, Cur Avg Loss: 0.17657972, Log Avg loss: 0.17282154, Global Avg Loss: 0.66668230, Time: 0.0209 Steps: 96640, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000845, Sample Num: 13520, Cur Loss: 0.17531818, Cur Avg Loss: 0.17656736, Log Avg loss: 0.17553525, Global Avg Loss: 0.66663148, Time: 0.0209 Steps: 96650, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000855, Sample Num: 13680, Cur Loss: 0.03996496, Cur Avg Loss: 0.17599844, Log Avg loss: 0.12792468, Global Avg Loss: 0.66657575, Time: 0.0209 Steps: 96660, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000865, Sample Num: 13840, Cur Loss: 0.08148038, Cur Avg Loss: 0.17622738, Log Avg loss: 0.19580202, Global Avg Loss: 0.66652705, Time: 0.0209 Steps: 96670, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000875, Sample Num: 14000, Cur Loss: 0.17348593, Cur Avg Loss: 0.17642991, Log Avg loss: 0.19394855, Global Avg Loss: 0.66647817, Time: 0.0208 Steps: 96680, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000885, Sample Num: 14160, Cur Loss: 0.27612689, Cur Avg Loss: 0.17706602, Log Avg loss: 0.23272562, Global Avg Loss: 0.66643331, Time: 0.0209 Steps: 96690, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000895, Sample Num: 14320, Cur Loss: 0.07343943, Cur Avg Loss: 0.17666599, Log Avg loss: 0.14126364, Global Avg Loss: 0.66637900, Time: 0.0209 Steps: 96700, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000905, Sample Num: 14480, Cur Loss: 0.45664594, Cur Avg Loss: 0.17685895, Log Avg loss: 0.19412873, Global Avg Loss: 0.66633017, Time: 0.0209 Steps: 96710, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000915, Sample Num: 14640, Cur Loss: 0.12446954, Cur Avg Loss: 0.17671974, Log Avg loss: 0.16412127, Global Avg Loss: 0.66627825, Time: 0.0209 Steps: 96720, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000925, Sample Num: 14800, Cur Loss: 0.30743006, Cur Avg Loss: 0.17705894, Log Avg loss: 0.20809617, Global Avg Loss: 0.66623088, Time: 0.0209 Steps: 96730, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000935, Sample Num: 14960, Cur Loss: 0.26991683, Cur Avg Loss: 0.17701013, Log Avg loss: 0.17249476, Global Avg Loss: 0.66617984, Time: 0.0209 Steps: 96740, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000945, Sample Num: 15120, Cur Loss: 0.10782391, Cur Avg Loss: 0.17651652, Log Avg loss: 0.13036426, Global Avg Loss: 0.66612446, Time: 0.0209 Steps: 96750, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000955, Sample Num: 15280, Cur Loss: 0.24298690, Cur Avg Loss: 0.17666055, Log Avg loss: 0.19027098, Global Avg Loss: 0.66607528, Time: 0.0209 Steps: 96760, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000965, Sample Num: 15440, Cur Loss: 0.06999701, Cur Avg Loss: 0.17692415, Log Avg loss: 0.20209778, Global Avg Loss: 0.66602733, Time: 0.0209 Steps: 96770, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000975, Sample Num: 15600, Cur Loss: 0.07329607, Cur Avg Loss: 0.17659445, Log Avg loss: 0.14477925, Global Avg Loss: 0.66597347, Time: 0.0209 Steps: 96780, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000985, Sample Num: 15760, Cur Loss: 0.21047339, Cur Avg Loss: 0.17614221, Log Avg loss: 0.13204841, Global Avg Loss: 0.66591831, Time: 0.0209 Steps: 96790, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000995, Sample Num: 15920, Cur Loss: 0.11621662, Cur Avg Loss: 0.17647018, Log Avg loss: 0.20877566, Global Avg Loss: 0.66587109, Time: 0.0209 Steps: 96800, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001005, Sample Num: 16080, Cur Loss: 0.10282616, Cur Avg Loss: 0.17660706, Log Avg loss: 0.19022620, Global Avg Loss: 0.66582195, Time: 0.0209 Steps: 96810, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001015, Sample Num: 16240, Cur Loss: 0.26285562, Cur Avg Loss: 0.17710981, Log Avg loss: 0.22763564, Global Avg Loss: 0.66577670, Time: 0.0209 Steps: 96820, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001025, Sample Num: 16400, Cur Loss: 0.87092638, Cur Avg Loss: 0.17790621, Log Avg loss: 0.25874094, Global Avg Loss: 0.66573466, Time: 0.0246 Steps: 96830, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001035, Sample Num: 16560, Cur Loss: 0.09592222, Cur Avg Loss: 0.17791552, Log Avg loss: 0.17886993, Global Avg Loss: 0.66568439, Time: 0.0209 Steps: 96840, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001045, Sample Num: 16720, Cur Loss: 0.18435073, Cur Avg Loss: 0.17796935, Log Avg loss: 0.18354047, Global Avg Loss: 0.66563460, Time: 0.0209 Steps: 96850, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001055, Sample Num: 16880, Cur Loss: 0.63595855, Cur Avg Loss: 0.17833873, Log Avg loss: 0.21693989, Global Avg Loss: 0.66558828, Time: 0.0209 Steps: 96860, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001065, Sample Num: 17040, Cur Loss: 0.06286801, Cur Avg Loss: 0.17815274, Log Avg loss: 0.15853003, Global Avg Loss: 0.66553593, Time: 0.0209 Steps: 96870, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001075, Sample Num: 17200, Cur Loss: 0.25257945, Cur Avg Loss: 0.17822587, Log Avg loss: 0.18601408, Global Avg Loss: 0.66548644, Time: 0.0208 Steps: 96880, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001085, Sample Num: 17360, Cur Loss: 0.08181997, Cur Avg Loss: 0.17804873, Log Avg loss: 0.15900618, Global Avg Loss: 0.66543416, Time: 0.0208 Steps: 96890, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001095, Sample Num: 17520, Cur Loss: 0.12983735, Cur Avg Loss: 0.17851554, Log Avg loss: 0.22916542, Global Avg Loss: 0.66538914, Time: 0.0209 Steps: 96900, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001105, Sample Num: 17680, Cur Loss: 0.18813798, Cur Avg Loss: 0.17830754, Log Avg loss: 0.15553126, Global Avg Loss: 0.66533653, Time: 0.0209 Steps: 96910, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001115, Sample Num: 17840, Cur Loss: 0.06521718, Cur Avg Loss: 0.17770349, Log Avg loss: 0.11095623, Global Avg Loss: 0.66527933, Time: 0.0209 Steps: 96920, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001125, Sample Num: 18000, Cur Loss: 0.18216820, Cur Avg Loss: 0.17756181, Log Avg loss: 0.16176374, Global Avg Loss: 0.66522738, Time: 0.0208 Steps: 96930, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001135, Sample Num: 18160, Cur Loss: 0.06385406, Cur Avg Loss: 0.17745581, Log Avg loss: 0.16553159, Global Avg Loss: 0.66517584, Time: 0.0209 Steps: 96940, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001145, Sample Num: 18320, Cur Loss: 0.26769337, Cur Avg Loss: 0.17771303, Log Avg loss: 0.20690717, Global Avg Loss: 0.66512857, Time: 0.0209 Steps: 96950, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001155, Sample Num: 18480, Cur Loss: 0.15036379, Cur Avg Loss: 0.17764084, Log Avg loss: 0.16937463, Global Avg Loss: 0.66507744, Time: 0.0208 Steps: 96960, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001165, Sample Num: 18640, Cur Loss: 0.18640271, Cur Avg Loss: 0.17779966, Log Avg loss: 0.19614307, Global Avg Loss: 0.66502908, Time: 0.0209 Steps: 96970, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001175, Sample Num: 18800, Cur Loss: 0.13313951, Cur Avg Loss: 0.17764595, Log Avg loss: 0.15973885, Global Avg Loss: 0.66497698, Time: 0.0208 Steps: 96980, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001185, Sample Num: 18960, Cur Loss: 0.11832854, Cur Avg Loss: 0.17723348, Log Avg loss: 0.12876889, Global Avg Loss: 0.66492169, Time: 0.0208 Steps: 96990, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001195, Sample Num: 19120, Cur Loss: 0.07326016, Cur Avg Loss: 0.17733599, Log Avg loss: 0.18948267, Global Avg Loss: 0.66487268, Time: 0.0208 Steps: 97000, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001205, Sample Num: 19280, Cur Loss: 0.13688546, Cur Avg Loss: 0.17705225, Log Avg loss: 0.14314608, Global Avg Loss: 0.66481890, Time: 0.0209 Steps: 97010, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001215, Sample Num: 19440, Cur Loss: 0.07659620, Cur Avg Loss: 0.17691679, Log Avg loss: 0.16059319, Global Avg Loss: 0.66476693, Time: 0.0209 Steps: 97020, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001225, Sample Num: 19600, Cur Loss: 0.25169927, Cur Avg Loss: 0.17741806, Log Avg loss: 0.23832312, Global Avg Loss: 0.66472298, Time: 0.0209 Steps: 97030, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001235, Sample Num: 19760, Cur Loss: 0.26765221, Cur Avg Loss: 0.17729538, Log Avg loss: 0.16226707, Global Avg Loss: 0.66467120, Time: 0.0209 Steps: 97040, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001245, Sample Num: 19920, Cur Loss: 0.19451997, Cur Avg Loss: 0.17733669, Log Avg loss: 0.18243799, Global Avg Loss: 0.66462151, Time: 0.0209 Steps: 97050, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001255, Sample Num: 20080, Cur Loss: 0.09390877, Cur Avg Loss: 0.17744591, Log Avg loss: 0.19104353, Global Avg Loss: 0.66457272, Time: 0.0209 Steps: 97060, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001265, Sample Num: 20240, Cur Loss: 0.13721986, Cur Avg Loss: 0.17782852, Log Avg loss: 0.22584632, Global Avg Loss: 0.66452752, Time: 0.0208 Steps: 97070, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001275, Sample Num: 20400, Cur Loss: 0.05950871, Cur Avg Loss: 0.17796706, Log Avg loss: 0.19549237, Global Avg Loss: 0.66447921, Time: 0.0209 Steps: 97080, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001285, Sample Num: 20560, Cur Loss: 0.05752382, Cur Avg Loss: 0.17832697, Log Avg loss: 0.22421634, Global Avg Loss: 0.66443386, Time: 0.0209 Steps: 97090, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001295, Sample Num: 20720, Cur Loss: 0.10110391, Cur Avg Loss: 0.17809622, Log Avg loss: 0.14844462, Global Avg Loss: 0.66438072, Time: 0.0209 Steps: 97100, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001305, Sample Num: 20880, Cur Loss: 0.03244293, Cur Avg Loss: 0.17823934, Log Avg loss: 0.19677303, Global Avg Loss: 0.66433257, Time: 0.0210 Steps: 97110, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001315, Sample Num: 21040, Cur Loss: 0.03797191, Cur Avg Loss: 0.17769185, Log Avg loss: 0.10624372, Global Avg Loss: 0.66427510, Time: 0.0209 Steps: 97120, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001325, Sample Num: 21200, Cur Loss: 0.25176743, Cur Avg Loss: 0.17734767, Log Avg loss: 0.13208901, Global Avg Loss: 0.66422031, Time: 0.0208 Steps: 97130, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001335, Sample Num: 21360, Cur Loss: 0.32933909, Cur Avg Loss: 0.17714191, Log Avg loss: 0.14987786, Global Avg Loss: 0.66416736, Time: 0.0209 Steps: 97140, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001345, Sample Num: 21520, Cur Loss: 0.17124701, Cur Avg Loss: 0.17715463, Log Avg loss: 0.17885377, Global Avg Loss: 0.66411741, Time: 0.0209 Steps: 97150, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001355, Sample Num: 21680, Cur Loss: 0.08365899, Cur Avg Loss: 0.17702937, Log Avg loss: 0.16018149, Global Avg Loss: 0.66406554, Time: 0.0209 Steps: 97160, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001365, Sample Num: 21840, Cur Loss: 0.10314031, Cur Avg Loss: 0.17737595, Log Avg loss: 0.22433727, Global Avg Loss: 0.66402029, Time: 0.0209 Steps: 97170, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001375, Sample Num: 22000, Cur Loss: 0.16171557, Cur Avg Loss: 0.17705169, Log Avg loss: 0.13278999, Global Avg Loss: 0.66396562, Time: 0.0209 Steps: 97180, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001385, Sample Num: 22160, Cur Loss: 0.06781032, Cur Avg Loss: 0.17702237, Log Avg loss: 0.17299056, Global Avg Loss: 0.66391511, Time: 0.0209 Steps: 97190, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001395, Sample Num: 22320, Cur Loss: 0.14123216, Cur Avg Loss: 0.17745058, Log Avg loss: 0.23675821, Global Avg Loss: 0.66387116, Time: 0.0209 Steps: 97200, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001405, Sample Num: 22480, Cur Loss: 0.32041684, Cur Avg Loss: 0.17764226, Log Avg loss: 0.20438111, Global Avg Loss: 0.66382389, Time: 0.0209 Steps: 97210, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001415, Sample Num: 22640, Cur Loss: 0.69093078, Cur Avg Loss: 0.17808438, Log Avg loss: 0.24020308, Global Avg Loss: 0.66378032, Time: 0.0208 Steps: 97220, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001425, Sample Num: 22800, Cur Loss: 0.07558193, Cur Avg Loss: 0.17762799, Log Avg loss: 0.11304865, Global Avg Loss: 0.66372368, Time: 0.0209 Steps: 97230, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001435, Sample Num: 22960, Cur Loss: 0.43071833, Cur Avg Loss: 0.17780400, Log Avg loss: 0.20288584, Global Avg Loss: 0.66367629, Time: 0.0209 Steps: 97240, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001445, Sample Num: 23120, Cur Loss: 0.09520650, Cur Avg Loss: 0.17756023, Log Avg loss: 0.14257898, Global Avg Loss: 0.66362270, Time: 0.0209 Steps: 97250, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001455, Sample Num: 23280, Cur Loss: 0.52105379, Cur Avg Loss: 0.17763848, Log Avg loss: 0.18894568, Global Avg Loss: 0.66357390, Time: 0.0209 Steps: 97260, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001465, Sample Num: 23440, Cur Loss: 0.12701628, Cur Avg Loss: 0.17782120, Log Avg loss: 0.20440715, Global Avg Loss: 0.66352669, Time: 0.0209 Steps: 97270, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001475, Sample Num: 23600, Cur Loss: 0.33349398, Cur Avg Loss: 0.17822303, Log Avg loss: 0.23708993, Global Avg Loss: 0.66348286, Time: 0.0208 Steps: 97280, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001485, Sample Num: 23760, Cur Loss: 0.16213167, Cur Avg Loss: 0.17792357, Log Avg loss: 0.13375323, Global Avg Loss: 0.66342841, Time: 0.0209 Steps: 97290, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001495, Sample Num: 23920, Cur Loss: 0.17012151, Cur Avg Loss: 0.17762113, Log Avg loss: 0.13270943, Global Avg Loss: 0.66337386, Time: 0.0209 Steps: 97300, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001505, Sample Num: 24080, Cur Loss: 0.16696867, Cur Avg Loss: 0.17760379, Log Avg loss: 0.17501099, Global Avg Loss: 0.66332368, Time: 0.0209 Steps: 97310, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001515, Sample Num: 24240, Cur Loss: 0.13878563, Cur Avg Loss: 0.17762028, Log Avg loss: 0.18010284, Global Avg Loss: 0.66327402, Time: 0.0209 Steps: 97320, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001525, Sample Num: 24400, Cur Loss: 0.19495960, Cur Avg Loss: 0.17776470, Log Avg loss: 0.19964408, Global Avg Loss: 0.66322639, Time: 0.0209 Steps: 97330, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001535, Sample Num: 24560, Cur Loss: 0.08035646, Cur Avg Loss: 0.17821532, Log Avg loss: 0.24693428, Global Avg Loss: 0.66318362, Time: 0.0209 Steps: 97340, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001545, Sample Num: 24720, Cur Loss: 0.14246622, Cur Avg Loss: 0.17831328, Log Avg loss: 0.19335108, Global Avg Loss: 0.66313536, Time: 0.0208 Steps: 97350, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001555, Sample Num: 24880, Cur Loss: 0.15854061, Cur Avg Loss: 0.17813123, Log Avg loss: 0.15000447, Global Avg Loss: 0.66308266, Time: 0.0208 Steps: 97360, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001565, Sample Num: 25040, Cur Loss: 0.10913467, Cur Avg Loss: 0.17791433, Log Avg loss: 0.14418616, Global Avg Loss: 0.66302936, Time: 0.0208 Steps: 97370, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001575, Sample Num: 25200, Cur Loss: 0.36260304, Cur Avg Loss: 0.17792005, Log Avg loss: 0.17881454, Global Avg Loss: 0.66297964, Time: 0.0209 Steps: 97380, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001585, Sample Num: 25360, Cur Loss: 0.16842416, Cur Avg Loss: 0.17791319, Log Avg loss: 0.17683338, Global Avg Loss: 0.66292972, Time: 0.0209 Steps: 97390, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001595, Sample Num: 25520, Cur Loss: 0.19544137, Cur Avg Loss: 0.17785415, Log Avg loss: 0.16849606, Global Avg Loss: 0.66287896, Time: 0.0209 Steps: 97400, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001605, Sample Num: 25680, Cur Loss: 0.24598536, Cur Avg Loss: 0.17797143, Log Avg loss: 0.19667782, Global Avg Loss: 0.66283110, Time: 0.0209 Steps: 97410, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001615, Sample Num: 25840, Cur Loss: 0.17614186, Cur Avg Loss: 0.17798501, Log Avg loss: 0.18016415, Global Avg Loss: 0.66278156, Time: 0.0209 Steps: 97420, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001625, Sample Num: 26000, Cur Loss: 0.26002330, Cur Avg Loss: 0.17816189, Log Avg loss: 0.20672819, Global Avg Loss: 0.66273475, Time: 0.0209 Steps: 97430, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001635, Sample Num: 26160, Cur Loss: 0.25549868, Cur Avg Loss: 0.17824451, Log Avg loss: 0.19167024, Global Avg Loss: 0.66268640, Time: 0.0209 Steps: 97440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001645, Sample Num: 26320, Cur Loss: 0.21365848, Cur Avg Loss: 0.17810790, Log Avg loss: 0.15577282, Global Avg Loss: 0.66263438, Time: 0.0209 Steps: 97450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001655, Sample Num: 26480, Cur Loss: 0.13412639, Cur Avg Loss: 0.17779651, Log Avg loss: 0.12657286, Global Avg Loss: 0.66257938, Time: 0.0209 Steps: 97460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001665, Sample Num: 26640, Cur Loss: 0.06306966, Cur Avg Loss: 0.17781365, Log Avg loss: 0.18065044, Global Avg Loss: 0.66252994, Time: 0.0208 Steps: 97470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001675, Sample Num: 26800, Cur Loss: 0.50163007, Cur Avg Loss: 0.17825921, Log Avg loss: 0.25244443, Global Avg Loss: 0.66248787, Time: 0.0209 Steps: 97480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001685, Sample Num: 26960, Cur Loss: 0.03856150, Cur Avg Loss: 0.17841889, Log Avg loss: 0.20516582, Global Avg Loss: 0.66244096, Time: 0.0208 Steps: 97490, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001695, Sample Num: 27120, Cur Loss: 0.19751063, Cur Avg Loss: 0.17817762, Log Avg loss: 0.13752306, Global Avg Loss: 0.66238712, Time: 0.0209 Steps: 97500, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001705, Sample Num: 27280, Cur Loss: 0.07039557, Cur Avg Loss: 0.17785192, Log Avg loss: 0.12264495, Global Avg Loss: 0.66233177, Time: 0.0209 Steps: 97510, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001715, Sample Num: 27440, Cur Loss: 0.17461158, Cur Avg Loss: 0.17778152, Log Avg loss: 0.16577968, Global Avg Loss: 0.66228085, Time: 0.0209 Steps: 97520, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001725, Sample Num: 27600, Cur Loss: 0.24026832, Cur Avg Loss: 0.17794385, Log Avg loss: 0.20578316, Global Avg Loss: 0.66223405, Time: 0.0209 Steps: 97530, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001735, Sample Num: 27760, Cur Loss: 0.23407917, Cur Avg Loss: 0.17819088, Log Avg loss: 0.22080246, Global Avg Loss: 0.66218879, Time: 0.0209 Steps: 97540, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001745, Sample Num: 27920, Cur Loss: 0.05670610, Cur Avg Loss: 0.17811763, Log Avg loss: 0.16540963, Global Avg Loss: 0.66213786, Time: 0.0209 Steps: 97550, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001755, Sample Num: 28080, Cur Loss: 0.18210374, Cur Avg Loss: 0.17805590, Log Avg loss: 0.16728468, Global Avg Loss: 0.66208714, Time: 0.0209 Steps: 97560, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001765, Sample Num: 28240, Cur Loss: 0.11497215, Cur Avg Loss: 0.17798876, Log Avg loss: 0.16620451, Global Avg Loss: 0.66203632, Time: 0.0209 Steps: 97570, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001775, Sample Num: 28400, Cur Loss: 0.17315698, Cur Avg Loss: 0.17815412, Log Avg loss: 0.20734069, Global Avg Loss: 0.66198972, Time: 0.0208 Steps: 97580, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001785, Sample Num: 28560, Cur Loss: 0.19139889, Cur Avg Loss: 0.17811396, Log Avg loss: 0.17098611, Global Avg Loss: 0.66193941, Time: 0.0209 Steps: 97590, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001795, Sample Num: 28720, Cur Loss: 0.14705700, Cur Avg Loss: 0.17786252, Log Avg loss: 0.13298036, Global Avg Loss: 0.66188521, Time: 0.0241 Steps: 97600, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001805, Sample Num: 28880, Cur Loss: 0.04680318, Cur Avg Loss: 0.17822220, Log Avg loss: 0.24278416, Global Avg Loss: 0.66184227, Time: 0.0208 Steps: 97610, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001815, Sample Num: 29040, Cur Loss: 0.16908744, Cur Avg Loss: 0.17814761, Log Avg loss: 0.16468488, Global Avg Loss: 0.66179135, Time: 0.0208 Steps: 97620, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001825, Sample Num: 29200, Cur Loss: 0.11779250, Cur Avg Loss: 0.17794962, Log Avg loss: 0.14201451, Global Avg Loss: 0.66173811, Time: 0.0209 Steps: 97630, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001835, Sample Num: 29360, Cur Loss: 0.06348390, Cur Avg Loss: 0.17808314, Log Avg loss: 0.20245066, Global Avg Loss: 0.66169107, Time: 0.0208 Steps: 97640, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001845, Sample Num: 29520, Cur Loss: 0.09993690, Cur Avg Loss: 0.17795360, Log Avg loss: 0.15418175, Global Avg Loss: 0.66163910, Time: 0.0209 Steps: 97650, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001855, Sample Num: 29680, Cur Loss: 0.10397351, Cur Avg Loss: 0.17789024, Log Avg loss: 0.16620116, Global Avg Loss: 0.66158836, Time: 0.0208 Steps: 97660, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001865, Sample Num: 29840, Cur Loss: 0.29185635, Cur Avg Loss: 0.17816364, Log Avg loss: 0.22887837, Global Avg Loss: 0.66154406, Time: 0.0209 Steps: 97670, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001875, Sample Num: 30000, Cur Loss: 0.12190560, Cur Avg Loss: 0.17830504, Log Avg loss: 0.20467764, Global Avg Loss: 0.66149729, Time: 0.0219 Steps: 97680, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001885, Sample Num: 30160, Cur Loss: 0.09821520, Cur Avg Loss: 0.17813286, Log Avg loss: 0.14584771, Global Avg Loss: 0.66144451, Time: 0.0219 Steps: 97690, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001895, Sample Num: 30320, Cur Loss: 0.36569718, Cur Avg Loss: 0.17826995, Log Avg loss: 0.20411218, Global Avg Loss: 0.66139770, Time: 0.0219 Steps: 97700, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001905, Sample Num: 30480, Cur Loss: 0.17265949, Cur Avg Loss: 0.17815779, Log Avg loss: 0.15690284, Global Avg Loss: 0.66134606, Time: 0.0219 Steps: 97710, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001915, Sample Num: 30640, Cur Loss: 0.30644378, Cur Avg Loss: 0.17821106, Log Avg loss: 0.18835903, Global Avg Loss: 0.66129766, Time: 0.0219 Steps: 97720, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001925, Sample Num: 30800, Cur Loss: 0.06156192, Cur Avg Loss: 0.17819576, Log Avg loss: 0.17526689, Global Avg Loss: 0.66124793, Time: 0.0219 Steps: 97730, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001935, Sample Num: 30960, Cur Loss: 0.11551011, Cur Avg Loss: 0.17845471, Log Avg loss: 0.22830102, Global Avg Loss: 0.66120363, Time: 0.0219 Steps: 97740, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001945, Sample Num: 31120, Cur Loss: 0.22271448, Cur Avg Loss: 0.17851638, Log Avg loss: 0.19045004, Global Avg Loss: 0.66115547, Time: 0.0219 Steps: 97750, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001955, Sample Num: 31280, Cur Loss: 0.12083329, Cur Avg Loss: 0.17912459, Log Avg loss: 0.29742167, Global Avg Loss: 0.66111827, Time: 0.0219 Steps: 97760, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001965, Sample Num: 31440, Cur Loss: 0.10554536, Cur Avg Loss: 0.17883983, Log Avg loss: 0.12316975, Global Avg Loss: 0.66106325, Time: 0.0219 Steps: 97770, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001975, Sample Num: 31600, Cur Loss: 0.08439076, Cur Avg Loss: 0.17866387, Log Avg loss: 0.14408757, Global Avg Loss: 0.66101037, Time: 0.0219 Steps: 97780, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001985, Sample Num: 31760, Cur Loss: 0.12924343, Cur Avg Loss: 0.17875295, Log Avg loss: 0.19634507, Global Avg Loss: 0.66096286, Time: 0.0219 Steps: 97790, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001995, Sample Num: 31920, Cur Loss: 0.35040200, Cur Avg Loss: 0.17900432, Log Avg loss: 0.22890287, Global Avg Loss: 0.66091868, Time: 0.0219 Steps: 97800, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002005, Sample Num: 32080, Cur Loss: 0.08433153, Cur Avg Loss: 0.17922748, Log Avg loss: 0.22374724, Global Avg Loss: 0.66087398, Time: 0.0219 Steps: 97810, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002015, Sample Num: 32240, Cur Loss: 0.45849130, Cur Avg Loss: 0.17977632, Log Avg loss: 0.28981878, Global Avg Loss: 0.66083605, Time: 0.0219 Steps: 97820, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002025, Sample Num: 32400, Cur Loss: 0.07965831, Cur Avg Loss: 0.18002199, Log Avg loss: 0.22952491, Global Avg Loss: 0.66079196, Time: 0.0219 Steps: 97830, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002035, Sample Num: 32560, Cur Loss: 0.14629216, Cur Avg Loss: 0.18018882, Log Avg loss: 0.21397184, Global Avg Loss: 0.66074630, Time: 0.0219 Steps: 97840, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002045, Sample Num: 32720, Cur Loss: 0.11427022, Cur Avg Loss: 0.18005794, Log Avg loss: 0.15342240, Global Avg Loss: 0.66069445, Time: 0.0219 Steps: 97850, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002055, Sample Num: 32880, Cur Loss: 0.05723392, Cur Avg Loss: 0.17989843, Log Avg loss: 0.14727907, Global Avg Loss: 0.66064198, Time: 0.0208 Steps: 97860, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002065, Sample Num: 33040, Cur Loss: 0.04107144, Cur Avg Loss: 0.17958779, Log Avg loss: 0.11575266, Global Avg Loss: 0.66058631, Time: 0.0208 Steps: 97870, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002075, Sample Num: 33200, Cur Loss: 0.03333278, Cur Avg Loss: 0.17947069, Log Avg loss: 0.15528868, Global Avg Loss: 0.66053469, Time: 0.0207 Steps: 97880, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002085, Sample Num: 33360, Cur Loss: 0.21089715, Cur Avg Loss: 0.17941197, Log Avg loss: 0.16722739, Global Avg Loss: 0.66048429, Time: 0.0208 Steps: 97890, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002095, Sample Num: 33520, Cur Loss: 0.57919300, Cur Avg Loss: 0.17949966, Log Avg loss: 0.19778390, Global Avg Loss: 0.66043703, Time: 0.0208 Steps: 97900, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002105, Sample Num: 33680, Cur Loss: 0.10977197, Cur Avg Loss: 0.17973613, Log Avg loss: 0.22927525, Global Avg Loss: 0.66039299, Time: 0.0207 Steps: 97910, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002115, Sample Num: 33840, Cur Loss: 0.47106487, Cur Avg Loss: 0.17978911, Log Avg loss: 0.19094107, Global Avg Loss: 0.66034505, Time: 0.0208 Steps: 97920, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002125, Sample Num: 34000, Cur Loss: 0.17197400, Cur Avg Loss: 0.17961252, Log Avg loss: 0.14226414, Global Avg Loss: 0.66029215, Time: 0.0208 Steps: 97930, Updated lr: 0.000008 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.753278, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.284213, "eval_total_loss": 199.801645, "eval_mae": 0.421401, "eval_mse": 0.284224, "eval_r2": 0.819329, "eval_sp_statistic": 0.901983, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925963, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.2951, "test_total_loss": 148.140303, "test_mae": 0.348512, "test_mse": 0.295205, "test_r2": 0.809472, "test_sp_statistic": 0.875603, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921748, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6602710237223735, "train_cur_epoch_loss": 382.2491289637983, "train_cur_epoch_avg_loss": 0.17954397790690385, "train_cur_epoch_time": 44.75327777862549, "train_cur_epoch_avg_time": 0.021020797453558237, "epoch": 46, "step": 97934} ################################################## Training, Epoch: 0047, Batch: 000006, Sample Num: 96, Cur Loss: 0.39590222, Cur Avg Loss: 0.18039502, Log Avg loss: 0.16549002, Global Avg Loss: 0.66024163, Time: 0.0220 Steps: 97940, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000016, Sample Num: 256, Cur Loss: 0.11700973, Cur Avg Loss: 0.14657004, Log Avg loss: 0.12627505, Global Avg Loss: 0.66018711, Time: 0.0208 Steps: 97950, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000026, Sample Num: 416, Cur Loss: 0.12798728, Cur Avg Loss: 0.13198333, Log Avg loss: 0.10864460, Global Avg Loss: 0.66013081, Time: 0.0208 Steps: 97960, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000036, Sample Num: 576, Cur Loss: 0.11487942, Cur Avg Loss: 0.14535190, Log Avg loss: 0.18011019, Global Avg Loss: 0.66008181, Time: 0.0208 Steps: 97970, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000046, Sample Num: 736, Cur Loss: 0.16464598, Cur Avg Loss: 0.16597294, Log Avg loss: 0.24020868, Global Avg Loss: 0.66003896, Time: 0.0209 Steps: 97980, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000056, Sample Num: 896, Cur Loss: 0.15898210, Cur Avg Loss: 0.16510100, Log Avg loss: 0.16109008, Global Avg Loss: 0.65998804, Time: 0.0208 Steps: 97990, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000066, Sample Num: 1056, Cur Loss: 0.31571019, Cur Avg Loss: 0.17346486, Log Avg loss: 0.22030245, Global Avg Loss: 0.65994317, Time: 0.0208 Steps: 98000, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000076, Sample Num: 1216, Cur Loss: 0.18412855, Cur Avg Loss: 0.17310408, Log Avg loss: 0.17072297, Global Avg Loss: 0.65989326, Time: 0.0210 Steps: 98010, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000086, Sample Num: 1376, Cur Loss: 0.28175378, Cur Avg Loss: 0.16942494, Log Avg loss: 0.14146345, Global Avg Loss: 0.65984037, Time: 0.0209 Steps: 98020, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000096, Sample Num: 1536, Cur Loss: 0.08950048, Cur Avg Loss: 0.16772117, Log Avg loss: 0.15306877, Global Avg Loss: 0.65978867, Time: 0.0209 Steps: 98030, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000106, Sample Num: 1696, Cur Loss: 0.31262171, Cur Avg Loss: 0.16977699, Log Avg loss: 0.18951288, Global Avg Loss: 0.65974071, Time: 0.0208 Steps: 98040, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000116, Sample Num: 1856, Cur Loss: 0.33044255, Cur Avg Loss: 0.17029419, Log Avg loss: 0.17577645, Global Avg Loss: 0.65969135, Time: 0.0208 Steps: 98050, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000126, Sample Num: 2016, Cur Loss: 0.08756195, Cur Avg Loss: 0.16848762, Log Avg loss: 0.14753148, Global Avg Loss: 0.65963912, Time: 0.0210 Steps: 98060, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000136, Sample Num: 2176, Cur Loss: 0.16836156, Cur Avg Loss: 0.16851509, Log Avg loss: 0.16886118, Global Avg Loss: 0.65958907, Time: 0.0208 Steps: 98070, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000146, Sample Num: 2336, Cur Loss: 0.08928677, Cur Avg Loss: 0.16595641, Log Avg loss: 0.13115829, Global Avg Loss: 0.65953520, Time: 0.0209 Steps: 98080, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000156, Sample Num: 2496, Cur Loss: 0.27487916, Cur Avg Loss: 0.16664241, Log Avg loss: 0.17665801, Global Avg Loss: 0.65948597, Time: 0.0208 Steps: 98090, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000166, Sample Num: 2656, Cur Loss: 0.08625741, Cur Avg Loss: 0.16808319, Log Avg loss: 0.19055948, Global Avg Loss: 0.65943817, Time: 0.0208 Steps: 98100, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000176, Sample Num: 2816, Cur Loss: 0.26599297, Cur Avg Loss: 0.17187148, Log Avg loss: 0.23475704, Global Avg Loss: 0.65939488, Time: 0.0208 Steps: 98110, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000186, Sample Num: 2976, Cur Loss: 0.04845041, Cur Avg Loss: 0.16931162, Log Avg loss: 0.12425803, Global Avg Loss: 0.65934034, Time: 0.0208 Steps: 98120, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000196, Sample Num: 3136, Cur Loss: 0.07839654, Cur Avg Loss: 0.16748241, Log Avg loss: 0.13345908, Global Avg Loss: 0.65928675, Time: 0.0208 Steps: 98130, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000206, Sample Num: 3296, Cur Loss: 0.29733765, Cur Avg Loss: 0.16921615, Log Avg loss: 0.20319760, Global Avg Loss: 0.65924028, Time: 0.0208 Steps: 98140, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000216, Sample Num: 3456, Cur Loss: 0.08934462, Cur Avg Loss: 0.16827423, Log Avg loss: 0.14887056, Global Avg Loss: 0.65918828, Time: 0.0208 Steps: 98150, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000226, Sample Num: 3616, Cur Loss: 0.34866354, Cur Avg Loss: 0.16928743, Log Avg loss: 0.19117260, Global Avg Loss: 0.65914060, Time: 0.0208 Steps: 98160, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000236, Sample Num: 3776, Cur Loss: 0.21916103, Cur Avg Loss: 0.17028892, Log Avg loss: 0.19292270, Global Avg Loss: 0.65909311, Time: 0.0208 Steps: 98170, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000246, Sample Num: 3936, Cur Loss: 0.09712473, Cur Avg Loss: 0.17359232, Log Avg loss: 0.25155235, Global Avg Loss: 0.65905160, Time: 0.0208 Steps: 98180, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000256, Sample Num: 4096, Cur Loss: 0.20345117, Cur Avg Loss: 0.17520953, Log Avg loss: 0.21499305, Global Avg Loss: 0.65900638, Time: 0.0253 Steps: 98190, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000266, Sample Num: 4256, Cur Loss: 0.09538003, Cur Avg Loss: 0.17260887, Log Avg loss: 0.10603202, Global Avg Loss: 0.65895007, Time: 0.0209 Steps: 98200, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000276, Sample Num: 4416, Cur Loss: 0.21314099, Cur Avg Loss: 0.17219253, Log Avg loss: 0.16111777, Global Avg Loss: 0.65889937, Time: 0.0209 Steps: 98210, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000286, Sample Num: 4576, Cur Loss: 0.10010594, Cur Avg Loss: 0.17424783, Log Avg loss: 0.23097420, Global Avg Loss: 0.65885581, Time: 0.0209 Steps: 98220, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000296, Sample Num: 4736, Cur Loss: 0.06442683, Cur Avg Loss: 0.17338670, Log Avg loss: 0.14875836, Global Avg Loss: 0.65880388, Time: 0.0209 Steps: 98230, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000306, Sample Num: 4896, Cur Loss: 0.16682905, Cur Avg Loss: 0.17437168, Log Avg loss: 0.20352703, Global Avg Loss: 0.65875753, Time: 0.0209 Steps: 98240, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000316, Sample Num: 5056, Cur Loss: 0.29525563, Cur Avg Loss: 0.17406319, Log Avg loss: 0.16462343, Global Avg Loss: 0.65870724, Time: 0.0209 Steps: 98250, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000326, Sample Num: 5216, Cur Loss: 0.13517219, Cur Avg Loss: 0.17505232, Log Avg loss: 0.20630882, Global Avg Loss: 0.65866120, Time: 0.0209 Steps: 98260, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000336, Sample Num: 5376, Cur Loss: 0.16893531, Cur Avg Loss: 0.17559669, Log Avg loss: 0.19334309, Global Avg Loss: 0.65861385, Time: 0.0209 Steps: 98270, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000346, Sample Num: 5536, Cur Loss: 0.29795837, Cur Avg Loss: 0.17735307, Log Avg loss: 0.23636734, Global Avg Loss: 0.65857089, Time: 0.0209 Steps: 98280, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000356, Sample Num: 5696, Cur Loss: 0.05152847, Cur Avg Loss: 0.17662414, Log Avg loss: 0.15140336, Global Avg Loss: 0.65851929, Time: 0.0209 Steps: 98290, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000366, Sample Num: 5856, Cur Loss: 0.15425427, Cur Avg Loss: 0.17729841, Log Avg loss: 0.20130254, Global Avg Loss: 0.65847277, Time: 0.0208 Steps: 98300, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000376, Sample Num: 6016, Cur Loss: 0.03995512, Cur Avg Loss: 0.17731434, Log Avg loss: 0.17789728, Global Avg Loss: 0.65842389, Time: 0.0209 Steps: 98310, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000386, Sample Num: 6176, Cur Loss: 0.12498109, Cur Avg Loss: 0.17633585, Log Avg loss: 0.13954438, Global Avg Loss: 0.65837112, Time: 0.0209 Steps: 98320, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000396, Sample Num: 6336, Cur Loss: 0.04221879, Cur Avg Loss: 0.17447347, Log Avg loss: 0.10258565, Global Avg Loss: 0.65831459, Time: 0.0209 Steps: 98330, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000406, Sample Num: 6496, Cur Loss: 0.02692172, Cur Avg Loss: 0.17552821, Log Avg loss: 0.21729594, Global Avg Loss: 0.65826975, Time: 0.0209 Steps: 98340, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000416, Sample Num: 6656, Cur Loss: 0.16879094, Cur Avg Loss: 0.17567263, Log Avg loss: 0.18153629, Global Avg Loss: 0.65822127, Time: 0.0209 Steps: 98350, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000426, Sample Num: 6816, Cur Loss: 0.09908248, Cur Avg Loss: 0.17637169, Log Avg loss: 0.20545267, Global Avg Loss: 0.65817524, Time: 0.0209 Steps: 98360, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000436, Sample Num: 6976, Cur Loss: 0.25552598, Cur Avg Loss: 0.17670291, Log Avg loss: 0.19081289, Global Avg Loss: 0.65812773, Time: 0.0209 Steps: 98370, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000446, Sample Num: 7136, Cur Loss: 0.19392893, Cur Avg Loss: 0.17647985, Log Avg loss: 0.16675431, Global Avg Loss: 0.65807778, Time: 0.0209 Steps: 98380, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000456, Sample Num: 7296, Cur Loss: 0.08823927, Cur Avg Loss: 0.17589891, Log Avg loss: 0.14998874, Global Avg Loss: 0.65802614, Time: 0.0208 Steps: 98390, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000466, Sample Num: 7456, Cur Loss: 0.12706733, Cur Avg Loss: 0.17649751, Log Avg loss: 0.20379391, Global Avg Loss: 0.65797998, Time: 0.0209 Steps: 98400, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000476, Sample Num: 7616, Cur Loss: 0.32777190, Cur Avg Loss: 0.17619647, Log Avg loss: 0.16216794, Global Avg Loss: 0.65792960, Time: 0.0209 Steps: 98410, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000486, Sample Num: 7776, Cur Loss: 0.11618660, Cur Avg Loss: 0.17571440, Log Avg loss: 0.15276783, Global Avg Loss: 0.65787827, Time: 0.0209 Steps: 98420, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000496, Sample Num: 7936, Cur Loss: 0.13890323, Cur Avg Loss: 0.17533139, Log Avg loss: 0.15671703, Global Avg Loss: 0.65782736, Time: 0.0209 Steps: 98430, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000506, Sample Num: 8096, Cur Loss: 0.10161263, Cur Avg Loss: 0.17536351, Log Avg loss: 0.17695681, Global Avg Loss: 0.65777851, Time: 0.0209 Steps: 98440, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000516, Sample Num: 8256, Cur Loss: 0.15885076, Cur Avg Loss: 0.17500449, Log Avg loss: 0.15683795, Global Avg Loss: 0.65772763, Time: 0.0245 Steps: 98450, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000526, Sample Num: 8416, Cur Loss: 0.07446325, Cur Avg Loss: 0.17527846, Log Avg loss: 0.18941537, Global Avg Loss: 0.65768006, Time: 0.0208 Steps: 98460, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000536, Sample Num: 8576, Cur Loss: 0.09849620, Cur Avg Loss: 0.17447717, Log Avg loss: 0.13232941, Global Avg Loss: 0.65762671, Time: 0.0207 Steps: 98470, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000546, Sample Num: 8736, Cur Loss: 0.06588908, Cur Avg Loss: 0.17337886, Log Avg loss: 0.11450956, Global Avg Loss: 0.65757156, Time: 0.0208 Steps: 98480, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000556, Sample Num: 8896, Cur Loss: 0.04155141, Cur Avg Loss: 0.17268964, Log Avg loss: 0.13505821, Global Avg Loss: 0.65751851, Time: 0.0207 Steps: 98490, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000566, Sample Num: 9056, Cur Loss: 0.24261160, Cur Avg Loss: 0.17384475, Log Avg loss: 0.23806844, Global Avg Loss: 0.65747592, Time: 0.0208 Steps: 98500, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000576, Sample Num: 9216, Cur Loss: 0.17644879, Cur Avg Loss: 0.17341589, Log Avg loss: 0.14914253, Global Avg Loss: 0.65742432, Time: 0.0208 Steps: 98510, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000586, Sample Num: 9376, Cur Loss: 0.18937908, Cur Avg Loss: 0.17341950, Log Avg loss: 0.17362746, Global Avg Loss: 0.65737522, Time: 0.0208 Steps: 98520, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000596, Sample Num: 9536, Cur Loss: 0.16718611, Cur Avg Loss: 0.17298369, Log Avg loss: 0.14744539, Global Avg Loss: 0.65732346, Time: 0.0208 Steps: 98530, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000606, Sample Num: 9696, Cur Loss: 0.14329928, Cur Avg Loss: 0.17216358, Log Avg loss: 0.12328493, Global Avg Loss: 0.65726927, Time: 0.0208 Steps: 98540, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000616, Sample Num: 9856, Cur Loss: 0.15167145, Cur Avg Loss: 0.17145366, Log Avg loss: 0.12843247, Global Avg Loss: 0.65721561, Time: 0.0207 Steps: 98550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000626, Sample Num: 10016, Cur Loss: 0.39347306, Cur Avg Loss: 0.17338572, Log Avg loss: 0.29240074, Global Avg Loss: 0.65717859, Time: 0.0208 Steps: 98560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000636, Sample Num: 10176, Cur Loss: 0.22931157, Cur Avg Loss: 0.17327688, Log Avg loss: 0.16646313, Global Avg Loss: 0.65712881, Time: 0.0208 Steps: 98570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000646, Sample Num: 10336, Cur Loss: 0.07195396, Cur Avg Loss: 0.17232267, Log Avg loss: 0.11163553, Global Avg Loss: 0.65707347, Time: 0.0208 Steps: 98580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000656, Sample Num: 10496, Cur Loss: 0.24906829, Cur Avg Loss: 0.17226745, Log Avg loss: 0.16869982, Global Avg Loss: 0.65702394, Time: 0.0209 Steps: 98590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000666, Sample Num: 10656, Cur Loss: 0.10087107, Cur Avg Loss: 0.17259609, Log Avg loss: 0.19415497, Global Avg Loss: 0.65697699, Time: 0.0207 Steps: 98600, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000676, Sample Num: 10816, Cur Loss: 0.46991530, Cur Avg Loss: 0.17296583, Log Avg loss: 0.19759061, Global Avg Loss: 0.65693041, Time: 0.0207 Steps: 98610, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000686, Sample Num: 10976, Cur Loss: 0.06112295, Cur Avg Loss: 0.17288787, Log Avg loss: 0.16761790, Global Avg Loss: 0.65688079, Time: 0.0208 Steps: 98620, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000696, Sample Num: 11136, Cur Loss: 0.07162133, Cur Avg Loss: 0.17279325, Log Avg loss: 0.16630176, Global Avg Loss: 0.65683105, Time: 0.0209 Steps: 98630, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000706, Sample Num: 11296, Cur Loss: 0.23320380, Cur Avg Loss: 0.17244658, Log Avg loss: 0.14831868, Global Avg Loss: 0.65677950, Time: 0.0208 Steps: 98640, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000716, Sample Num: 11456, Cur Loss: 0.06441353, Cur Avg Loss: 0.17210660, Log Avg loss: 0.14810435, Global Avg Loss: 0.65672793, Time: 0.0208 Steps: 98650, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000726, Sample Num: 11616, Cur Loss: 0.06993365, Cur Avg Loss: 0.17119134, Log Avg loss: 0.10565858, Global Avg Loss: 0.65667208, Time: 0.0208 Steps: 98660, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000736, Sample Num: 11776, Cur Loss: 0.21404105, Cur Avg Loss: 0.17131291, Log Avg loss: 0.18013851, Global Avg Loss: 0.65662378, Time: 0.0208 Steps: 98670, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000746, Sample Num: 11936, Cur Loss: 0.11451989, Cur Avg Loss: 0.17066921, Log Avg loss: 0.12329280, Global Avg Loss: 0.65656974, Time: 0.0208 Steps: 98680, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000756, Sample Num: 12096, Cur Loss: 0.05759705, Cur Avg Loss: 0.17124426, Log Avg loss: 0.21414304, Global Avg Loss: 0.65652491, Time: 0.0208 Steps: 98690, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000766, Sample Num: 12256, Cur Loss: 0.21117823, Cur Avg Loss: 0.17148735, Log Avg loss: 0.18986547, Global Avg Loss: 0.65647763, Time: 0.0208 Steps: 98700, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000776, Sample Num: 12416, Cur Loss: 0.27374154, Cur Avg Loss: 0.17181739, Log Avg loss: 0.19709788, Global Avg Loss: 0.65643109, Time: 0.0208 Steps: 98710, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000786, Sample Num: 12576, Cur Loss: 0.18697587, Cur Avg Loss: 0.17211557, Log Avg loss: 0.19525475, Global Avg Loss: 0.65638437, Time: 0.0208 Steps: 98720, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000796, Sample Num: 12736, Cur Loss: 0.11942131, Cur Avg Loss: 0.17212105, Log Avg loss: 0.17255139, Global Avg Loss: 0.65633537, Time: 0.0207 Steps: 98730, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000806, Sample Num: 12896, Cur Loss: 0.36636478, Cur Avg Loss: 0.17277095, Log Avg loss: 0.22450286, Global Avg Loss: 0.65629163, Time: 0.0208 Steps: 98740, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000816, Sample Num: 13056, Cur Loss: 0.19730502, Cur Avg Loss: 0.17217395, Log Avg loss: 0.12405648, Global Avg Loss: 0.65623774, Time: 0.0208 Steps: 98750, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000826, Sample Num: 13216, Cur Loss: 0.22562730, Cur Avg Loss: 0.17264865, Log Avg loss: 0.21138412, Global Avg Loss: 0.65619269, Time: 0.0208 Steps: 98760, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000836, Sample Num: 13376, Cur Loss: 0.28924423, Cur Avg Loss: 0.17239247, Log Avg loss: 0.15123156, Global Avg Loss: 0.65614157, Time: 0.0208 Steps: 98770, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000846, Sample Num: 13536, Cur Loss: 0.10902712, Cur Avg Loss: 0.17280791, Log Avg loss: 0.20753925, Global Avg Loss: 0.65609615, Time: 0.0207 Steps: 98780, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000856, Sample Num: 13696, Cur Loss: 0.22922193, Cur Avg Loss: 0.17321160, Log Avg loss: 0.20736321, Global Avg Loss: 0.65605073, Time: 0.0208 Steps: 98790, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000866, Sample Num: 13856, Cur Loss: 0.09677799, Cur Avg Loss: 0.17366092, Log Avg loss: 0.21212267, Global Avg Loss: 0.65600580, Time: 0.0208 Steps: 98800, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000876, Sample Num: 14016, Cur Loss: 0.21998855, Cur Avg Loss: 0.17438943, Log Avg loss: 0.23747830, Global Avg Loss: 0.65596344, Time: 0.0208 Steps: 98810, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000886, Sample Num: 14176, Cur Loss: 0.20239195, Cur Avg Loss: 0.17473682, Log Avg loss: 0.20516814, Global Avg Loss: 0.65591782, Time: 0.0208 Steps: 98820, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000896, Sample Num: 14336, Cur Loss: 0.09778537, Cur Avg Loss: 0.17519349, Log Avg loss: 0.21565460, Global Avg Loss: 0.65587328, Time: 0.0208 Steps: 98830, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000906, Sample Num: 14496, Cur Loss: 0.12828040, Cur Avg Loss: 0.17493304, Log Avg loss: 0.15159673, Global Avg Loss: 0.65582226, Time: 0.0208 Steps: 98840, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000916, Sample Num: 14656, Cur Loss: 0.07251621, Cur Avg Loss: 0.17410960, Log Avg loss: 0.09950647, Global Avg Loss: 0.65576598, Time: 0.0207 Steps: 98850, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000926, Sample Num: 14816, Cur Loss: 0.05549571, Cur Avg Loss: 0.17424063, Log Avg loss: 0.18624236, Global Avg Loss: 0.65571848, Time: 0.0207 Steps: 98860, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000936, Sample Num: 14976, Cur Loss: 0.07648103, Cur Avg Loss: 0.17441160, Log Avg loss: 0.19024390, Global Avg Loss: 0.65567140, Time: 0.0208 Steps: 98870, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000946, Sample Num: 15136, Cur Loss: 0.20322332, Cur Avg Loss: 0.17374131, Log Avg loss: 0.11100212, Global Avg Loss: 0.65561632, Time: 0.0208 Steps: 98880, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000956, Sample Num: 15296, Cur Loss: 0.03375554, Cur Avg Loss: 0.17419566, Log Avg loss: 0.21717699, Global Avg Loss: 0.65557198, Time: 0.0208 Steps: 98890, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000966, Sample Num: 15456, Cur Loss: 0.15440187, Cur Avg Loss: 0.17375607, Log Avg loss: 0.13173157, Global Avg Loss: 0.65551902, Time: 0.0208 Steps: 98900, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000976, Sample Num: 15616, Cur Loss: 0.10503167, Cur Avg Loss: 0.17343780, Log Avg loss: 0.14269236, Global Avg Loss: 0.65546717, Time: 0.0208 Steps: 98910, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000986, Sample Num: 15776, Cur Loss: 0.22271851, Cur Avg Loss: 0.17292261, Log Avg loss: 0.12264008, Global Avg Loss: 0.65541330, Time: 0.0208 Steps: 98920, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000996, Sample Num: 15936, Cur Loss: 0.06445063, Cur Avg Loss: 0.17268339, Log Avg loss: 0.14909673, Global Avg Loss: 0.65536213, Time: 0.0208 Steps: 98930, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001006, Sample Num: 16096, Cur Loss: 0.26302814, Cur Avg Loss: 0.17232205, Log Avg loss: 0.13633251, Global Avg Loss: 0.65530967, Time: 0.0208 Steps: 98940, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001016, Sample Num: 16256, Cur Loss: 0.07224558, Cur Avg Loss: 0.17219730, Log Avg loss: 0.15964747, Global Avg Loss: 0.65525957, Time: 0.0208 Steps: 98950, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001026, Sample Num: 16416, Cur Loss: 0.19749501, Cur Avg Loss: 0.17205103, Log Avg loss: 0.15718988, Global Avg Loss: 0.65520924, Time: 0.0246 Steps: 98960, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001036, Sample Num: 16576, Cur Loss: 0.17600384, Cur Avg Loss: 0.17198787, Log Avg loss: 0.16550721, Global Avg Loss: 0.65515976, Time: 0.0208 Steps: 98970, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001046, Sample Num: 16736, Cur Loss: 0.12735297, Cur Avg Loss: 0.17207986, Log Avg loss: 0.18161052, Global Avg Loss: 0.65511192, Time: 0.0208 Steps: 98980, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001056, Sample Num: 16896, Cur Loss: 0.09931841, Cur Avg Loss: 0.17264786, Log Avg loss: 0.23206095, Global Avg Loss: 0.65506918, Time: 0.0208 Steps: 98990, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001066, Sample Num: 17056, Cur Loss: 0.08570188, Cur Avg Loss: 0.17305087, Log Avg loss: 0.21560860, Global Avg Loss: 0.65502479, Time: 0.0208 Steps: 99000, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001076, Sample Num: 17216, Cur Loss: 0.13018455, Cur Avg Loss: 0.17281012, Log Avg loss: 0.14714625, Global Avg Loss: 0.65497350, Time: 0.0208 Steps: 99010, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001086, Sample Num: 17376, Cur Loss: 0.16032061, Cur Avg Loss: 0.17259475, Log Avg loss: 0.14942056, Global Avg Loss: 0.65492244, Time: 0.0207 Steps: 99020, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001096, Sample Num: 17536, Cur Loss: 0.18249030, Cur Avg Loss: 0.17281182, Log Avg loss: 0.19638524, Global Avg Loss: 0.65487614, Time: 0.0208 Steps: 99030, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001106, Sample Num: 17696, Cur Loss: 0.07373305, Cur Avg Loss: 0.17278137, Log Avg loss: 0.16944487, Global Avg Loss: 0.65482713, Time: 0.0208 Steps: 99040, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001116, Sample Num: 17856, Cur Loss: 0.18025768, Cur Avg Loss: 0.17265987, Log Avg loss: 0.15922152, Global Avg Loss: 0.65477709, Time: 0.0208 Steps: 99050, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001126, Sample Num: 18016, Cur Loss: 0.12369490, Cur Avg Loss: 0.17237032, Log Avg loss: 0.14005671, Global Avg Loss: 0.65472513, Time: 0.0208 Steps: 99060, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001136, Sample Num: 18176, Cur Loss: 0.07654953, Cur Avg Loss: 0.17223248, Log Avg loss: 0.15671203, Global Avg Loss: 0.65467486, Time: 0.0208 Steps: 99070, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001146, Sample Num: 18336, Cur Loss: 0.12547356, Cur Avg Loss: 0.17200313, Log Avg loss: 0.14594868, Global Avg Loss: 0.65462352, Time: 0.0208 Steps: 99080, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001156, Sample Num: 18496, Cur Loss: 0.05564507, Cur Avg Loss: 0.17209058, Log Avg loss: 0.18211170, Global Avg Loss: 0.65457583, Time: 0.0208 Steps: 99090, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001166, Sample Num: 18656, Cur Loss: 0.16010870, Cur Avg Loss: 0.17241623, Log Avg loss: 0.21006181, Global Avg Loss: 0.65453098, Time: 0.0208 Steps: 99100, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001176, Sample Num: 18816, Cur Loss: 0.35849565, Cur Avg Loss: 0.17259849, Log Avg loss: 0.19384992, Global Avg Loss: 0.65448449, Time: 0.0207 Steps: 99110, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001186, Sample Num: 18976, Cur Loss: 0.29454330, Cur Avg Loss: 0.17261776, Log Avg loss: 0.17488379, Global Avg Loss: 0.65443611, Time: 0.0208 Steps: 99120, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001196, Sample Num: 19136, Cur Loss: 0.09047017, Cur Avg Loss: 0.17264780, Log Avg loss: 0.17621102, Global Avg Loss: 0.65438787, Time: 0.0208 Steps: 99130, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001206, Sample Num: 19296, Cur Loss: 0.13122468, Cur Avg Loss: 0.17269832, Log Avg loss: 0.17874083, Global Avg Loss: 0.65433989, Time: 0.0208 Steps: 99140, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001216, Sample Num: 19456, Cur Loss: 0.22895828, Cur Avg Loss: 0.17270182, Log Avg loss: 0.17312345, Global Avg Loss: 0.65429136, Time: 0.0208 Steps: 99150, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001226, Sample Num: 19616, Cur Loss: 0.08271061, Cur Avg Loss: 0.17343187, Log Avg loss: 0.26220555, Global Avg Loss: 0.65425181, Time: 0.0209 Steps: 99160, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001236, Sample Num: 19776, Cur Loss: 0.16416940, Cur Avg Loss: 0.17362374, Log Avg loss: 0.19714784, Global Avg Loss: 0.65420572, Time: 0.0209 Steps: 99170, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001246, Sample Num: 19936, Cur Loss: 0.07403918, Cur Avg Loss: 0.17334218, Log Avg loss: 0.13854144, Global Avg Loss: 0.65415373, Time: 0.0210 Steps: 99180, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001256, Sample Num: 20096, Cur Loss: 0.11178985, Cur Avg Loss: 0.17338678, Log Avg loss: 0.17894289, Global Avg Loss: 0.65410582, Time: 0.0208 Steps: 99190, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001266, Sample Num: 20256, Cur Loss: 0.04475537, Cur Avg Loss: 0.17327495, Log Avg loss: 0.15922913, Global Avg Loss: 0.65405593, Time: 0.0209 Steps: 99200, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001276, Sample Num: 20416, Cur Loss: 0.12471749, Cur Avg Loss: 0.17320052, Log Avg loss: 0.16377839, Global Avg Loss: 0.65400651, Time: 0.0209 Steps: 99210, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001286, Sample Num: 20576, Cur Loss: 0.19419312, Cur Avg Loss: 0.17386434, Log Avg loss: 0.25856745, Global Avg Loss: 0.65396666, Time: 0.0246 Steps: 99220, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001296, Sample Num: 20736, Cur Loss: 0.16697486, Cur Avg Loss: 0.17420148, Log Avg loss: 0.21755812, Global Avg Loss: 0.65392268, Time: 0.0209 Steps: 99230, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001306, Sample Num: 20896, Cur Loss: 0.26617277, Cur Avg Loss: 0.17393636, Log Avg loss: 0.13957659, Global Avg Loss: 0.65387085, Time: 0.0209 Steps: 99240, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001316, Sample Num: 21056, Cur Loss: 0.05890502, Cur Avg Loss: 0.17393334, Log Avg loss: 0.17353819, Global Avg Loss: 0.65382246, Time: 0.0209 Steps: 99250, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001326, Sample Num: 21216, Cur Loss: 0.14114997, Cur Avg Loss: 0.17445969, Log Avg loss: 0.24372765, Global Avg Loss: 0.65378114, Time: 0.0209 Steps: 99260, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001336, Sample Num: 21376, Cur Loss: 0.27749273, Cur Avg Loss: 0.17474821, Log Avg loss: 0.21300678, Global Avg Loss: 0.65373674, Time: 0.0209 Steps: 99270, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001346, Sample Num: 21536, Cur Loss: 0.28768066, Cur Avg Loss: 0.17435871, Log Avg loss: 0.12232046, Global Avg Loss: 0.65368321, Time: 0.0209 Steps: 99280, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001356, Sample Num: 21696, Cur Loss: 0.06824969, Cur Avg Loss: 0.17429756, Log Avg loss: 0.16606741, Global Avg Loss: 0.65363410, Time: 0.0209 Steps: 99290, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001366, Sample Num: 21856, Cur Loss: 0.19460821, Cur Avg Loss: 0.17429800, Log Avg loss: 0.17435806, Global Avg Loss: 0.65358584, Time: 0.0209 Steps: 99300, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001376, Sample Num: 22016, Cur Loss: 0.10969853, Cur Avg Loss: 0.17467857, Log Avg loss: 0.22666439, Global Avg Loss: 0.65354285, Time: 0.0208 Steps: 99310, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001386, Sample Num: 22176, Cur Loss: 0.06763370, Cur Avg Loss: 0.17445082, Log Avg loss: 0.14311184, Global Avg Loss: 0.65349145, Time: 0.0209 Steps: 99320, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001396, Sample Num: 22336, Cur Loss: 0.03288148, Cur Avg Loss: 0.17423306, Log Avg loss: 0.14405210, Global Avg Loss: 0.65344017, Time: 0.0209 Steps: 99330, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001406, Sample Num: 22496, Cur Loss: 0.19122417, Cur Avg Loss: 0.17432553, Log Avg loss: 0.18723374, Global Avg Loss: 0.65339324, Time: 0.0209 Steps: 99340, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001416, Sample Num: 22656, Cur Loss: 0.07962299, Cur Avg Loss: 0.17418051, Log Avg loss: 0.15379024, Global Avg Loss: 0.65334295, Time: 0.0209 Steps: 99350, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001426, Sample Num: 22816, Cur Loss: 0.04112465, Cur Avg Loss: 0.17370957, Log Avg loss: 0.10702497, Global Avg Loss: 0.65328797, Time: 0.0209 Steps: 99360, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001436, Sample Num: 22976, Cur Loss: 0.10411283, Cur Avg Loss: 0.17357110, Log Avg loss: 0.15382492, Global Avg Loss: 0.65323770, Time: 0.0209 Steps: 99370, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001446, Sample Num: 23136, Cur Loss: 0.40673605, Cur Avg Loss: 0.17353475, Log Avg loss: 0.16831495, Global Avg Loss: 0.65318891, Time: 0.0209 Steps: 99380, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001456, Sample Num: 23296, Cur Loss: 0.10239971, Cur Avg Loss: 0.17353885, Log Avg loss: 0.17413167, Global Avg Loss: 0.65314071, Time: 0.0209 Steps: 99390, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001466, Sample Num: 23456, Cur Loss: 0.36063671, Cur Avg Loss: 0.17363320, Log Avg loss: 0.18737099, Global Avg Loss: 0.65309385, Time: 0.0209 Steps: 99400, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001476, Sample Num: 23616, Cur Loss: 0.16790228, Cur Avg Loss: 0.17366506, Log Avg loss: 0.17833518, Global Avg Loss: 0.65304609, Time: 0.0209 Steps: 99410, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001486, Sample Num: 23776, Cur Loss: 0.18458717, Cur Avg Loss: 0.17351967, Log Avg loss: 0.15206033, Global Avg Loss: 0.65299570, Time: 0.0209 Steps: 99420, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001496, Sample Num: 23936, Cur Loss: 0.04981036, Cur Avg Loss: 0.17343944, Log Avg loss: 0.16151699, Global Avg Loss: 0.65294627, Time: 0.0209 Steps: 99430, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001506, Sample Num: 24096, Cur Loss: 0.27376184, Cur Avg Loss: 0.17352862, Log Avg loss: 0.18686996, Global Avg Loss: 0.65289940, Time: 0.0208 Steps: 99440, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001516, Sample Num: 24256, Cur Loss: 0.05969796, Cur Avg Loss: 0.17349450, Log Avg loss: 0.16835662, Global Avg Loss: 0.65285068, Time: 0.0209 Steps: 99450, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001526, Sample Num: 24416, Cur Loss: 0.24189037, Cur Avg Loss: 0.17379676, Log Avg loss: 0.21962000, Global Avg Loss: 0.65280712, Time: 0.0209 Steps: 99460, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001536, Sample Num: 24576, Cur Loss: 0.14116770, Cur Avg Loss: 0.17381432, Log Avg loss: 0.17649306, Global Avg Loss: 0.65275924, Time: 0.0254 Steps: 99470, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001546, Sample Num: 24736, Cur Loss: 0.09688851, Cur Avg Loss: 0.17364846, Log Avg loss: 0.14817267, Global Avg Loss: 0.65270851, Time: 0.0209 Steps: 99480, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001556, Sample Num: 24896, Cur Loss: 0.06868080, Cur Avg Loss: 0.17355050, Log Avg loss: 0.15840658, Global Avg Loss: 0.65265883, Time: 0.0209 Steps: 99490, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001566, Sample Num: 25056, Cur Loss: 0.36032319, Cur Avg Loss: 0.17338043, Log Avg loss: 0.14691614, Global Avg Loss: 0.65260800, Time: 0.0209 Steps: 99500, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001576, Sample Num: 25216, Cur Loss: 0.15711090, Cur Avg Loss: 0.17328248, Log Avg loss: 0.15794433, Global Avg Loss: 0.65255829, Time: 0.0209 Steps: 99510, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001586, Sample Num: 25376, Cur Loss: 0.21602447, Cur Avg Loss: 0.17339742, Log Avg loss: 0.19151127, Global Avg Loss: 0.65251196, Time: 0.0209 Steps: 99520, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001596, Sample Num: 25536, Cur Loss: 0.08363376, Cur Avg Loss: 0.17332193, Log Avg loss: 0.16135043, Global Avg Loss: 0.65246262, Time: 0.0209 Steps: 99530, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001606, Sample Num: 25696, Cur Loss: 0.03644201, Cur Avg Loss: 0.17329285, Log Avg loss: 0.16865089, Global Avg Loss: 0.65241401, Time: 0.0209 Steps: 99540, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001616, Sample Num: 25856, Cur Loss: 0.08212306, Cur Avg Loss: 0.17343486, Log Avg loss: 0.19624209, Global Avg Loss: 0.65236819, Time: 0.0209 Steps: 99550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001626, Sample Num: 26016, Cur Loss: 0.04329760, Cur Avg Loss: 0.17338716, Log Avg loss: 0.16567847, Global Avg Loss: 0.65231930, Time: 0.0209 Steps: 99560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001636, Sample Num: 26176, Cur Loss: 0.21664760, Cur Avg Loss: 0.17342511, Log Avg loss: 0.17959646, Global Avg Loss: 0.65227183, Time: 0.0209 Steps: 99570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001646, Sample Num: 26336, Cur Loss: 0.35873559, Cur Avg Loss: 0.17366826, Log Avg loss: 0.21344702, Global Avg Loss: 0.65222776, Time: 0.0209 Steps: 99580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001656, Sample Num: 26496, Cur Loss: 0.12954298, Cur Avg Loss: 0.17394503, Log Avg loss: 0.21950213, Global Avg Loss: 0.65218431, Time: 0.0209 Steps: 99590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001666, Sample Num: 26656, Cur Loss: 0.02654584, Cur Avg Loss: 0.17365141, Log Avg loss: 0.12502802, Global Avg Loss: 0.65213138, Time: 0.0209 Steps: 99600, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001676, Sample Num: 26816, Cur Loss: 0.30117500, Cur Avg Loss: 0.17342604, Log Avg loss: 0.13587876, Global Avg Loss: 0.65207956, Time: 0.0209 Steps: 99610, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001686, Sample Num: 26976, Cur Loss: 0.19925371, Cur Avg Loss: 0.17370278, Log Avg loss: 0.22008459, Global Avg Loss: 0.65203619, Time: 0.0209 Steps: 99620, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001696, Sample Num: 27136, Cur Loss: 0.12219258, Cur Avg Loss: 0.17351026, Log Avg loss: 0.14105169, Global Avg Loss: 0.65198490, Time: 0.0210 Steps: 99630, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001706, Sample Num: 27296, Cur Loss: 0.08257893, Cur Avg Loss: 0.17350219, Log Avg loss: 0.17213201, Global Avg Loss: 0.65193674, Time: 0.0210 Steps: 99640, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001716, Sample Num: 27456, Cur Loss: 0.09943142, Cur Avg Loss: 0.17339171, Log Avg loss: 0.15454463, Global Avg Loss: 0.65188683, Time: 0.0209 Steps: 99650, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001726, Sample Num: 27616, Cur Loss: 0.36160725, Cur Avg Loss: 0.17338849, Log Avg loss: 0.17283524, Global Avg Loss: 0.65183876, Time: 0.0209 Steps: 99660, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001736, Sample Num: 27776, Cur Loss: 0.24541993, Cur Avg Loss: 0.17308484, Log Avg loss: 0.12067633, Global Avg Loss: 0.65178547, Time: 0.0209 Steps: 99670, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001746, Sample Num: 27936, Cur Loss: 0.48040190, Cur Avg Loss: 0.17317831, Log Avg loss: 0.18940379, Global Avg Loss: 0.65173908, Time: 0.0209 Steps: 99680, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001756, Sample Num: 28096, Cur Loss: 0.29559100, Cur Avg Loss: 0.17352139, Log Avg loss: 0.23342352, Global Avg Loss: 0.65169712, Time: 0.0210 Steps: 99690, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001766, Sample Num: 28256, Cur Loss: 0.29673147, Cur Avg Loss: 0.17345416, Log Avg loss: 0.16164834, Global Avg Loss: 0.65164797, Time: 0.0209 Steps: 99700, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001776, Sample Num: 28416, Cur Loss: 0.06922456, Cur Avg Loss: 0.17343147, Log Avg loss: 0.16942519, Global Avg Loss: 0.65159961, Time: 0.0209 Steps: 99710, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001786, Sample Num: 28576, Cur Loss: 0.30401734, Cur Avg Loss: 0.17376671, Log Avg loss: 0.23330397, Global Avg Loss: 0.65155766, Time: 0.0209 Steps: 99720, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001796, Sample Num: 28736, Cur Loss: 0.41206521, Cur Avg Loss: 0.17379166, Log Avg loss: 0.17824752, Global Avg Loss: 0.65151020, Time: 0.0210 Steps: 99730, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001806, Sample Num: 28896, Cur Loss: 0.08698275, Cur Avg Loss: 0.17374622, Log Avg loss: 0.16558617, Global Avg Loss: 0.65146148, Time: 0.0209 Steps: 99740, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001816, Sample Num: 29056, Cur Loss: 0.09287174, Cur Avg Loss: 0.17367512, Log Avg loss: 0.16083368, Global Avg Loss: 0.65141230, Time: 0.0209 Steps: 99750, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001826, Sample Num: 29216, Cur Loss: 0.17189747, Cur Avg Loss: 0.17334818, Log Avg loss: 0.11397729, Global Avg Loss: 0.65135842, Time: 0.0209 Steps: 99760, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001836, Sample Num: 29376, Cur Loss: 0.21648246, Cur Avg Loss: 0.17385172, Log Avg loss: 0.26579802, Global Avg Loss: 0.65131978, Time: 0.0210 Steps: 99770, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001846, Sample Num: 29536, Cur Loss: 0.59802938, Cur Avg Loss: 0.17421855, Log Avg loss: 0.24156799, Global Avg Loss: 0.65127871, Time: 0.0209 Steps: 99780, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001856, Sample Num: 29696, Cur Loss: 0.48540124, Cur Avg Loss: 0.17432265, Log Avg loss: 0.19353991, Global Avg Loss: 0.65123284, Time: 0.0209 Steps: 99790, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001866, Sample Num: 29856, Cur Loss: 0.25935060, Cur Avg Loss: 0.17467525, Log Avg loss: 0.24011672, Global Avg Loss: 0.65119165, Time: 0.0209 Steps: 99800, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001876, Sample Num: 30016, Cur Loss: 0.20052649, Cur Avg Loss: 0.17484924, Log Avg loss: 0.20731704, Global Avg Loss: 0.65114718, Time: 0.0209 Steps: 99810, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001886, Sample Num: 30176, Cur Loss: 0.05143716, Cur Avg Loss: 0.17469895, Log Avg loss: 0.14650410, Global Avg Loss: 0.65109662, Time: 0.0209 Steps: 99820, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001896, Sample Num: 30336, Cur Loss: 0.23178703, Cur Avg Loss: 0.17462404, Log Avg loss: 0.16049606, Global Avg Loss: 0.65104748, Time: 0.0209 Steps: 99830, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001906, Sample Num: 30496, Cur Loss: 0.08683366, Cur Avg Loss: 0.17483709, Log Avg loss: 0.21523064, Global Avg Loss: 0.65100383, Time: 0.0209 Steps: 99840, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001916, Sample Num: 30656, Cur Loss: 0.24407375, Cur Avg Loss: 0.17500136, Log Avg loss: 0.20631148, Global Avg Loss: 0.65095929, Time: 0.0209 Steps: 99850, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001926, Sample Num: 30816, Cur Loss: 0.39371550, Cur Avg Loss: 0.17530380, Log Avg loss: 0.23325084, Global Avg Loss: 0.65091746, Time: 0.0209 Steps: 99860, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001936, Sample Num: 30976, Cur Loss: 0.17246182, Cur Avg Loss: 0.17512917, Log Avg loss: 0.14149653, Global Avg Loss: 0.65086645, Time: 0.0209 Steps: 99870, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001946, Sample Num: 31136, Cur Loss: 0.14089769, Cur Avg Loss: 0.17510098, Log Avg loss: 0.16964360, Global Avg Loss: 0.65081827, Time: 0.0209 Steps: 99880, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001956, Sample Num: 31296, Cur Loss: 0.04564967, Cur Avg Loss: 0.17517003, Log Avg loss: 0.18860580, Global Avg Loss: 0.65077200, Time: 0.0210 Steps: 99890, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001966, Sample Num: 31456, Cur Loss: 0.08522125, Cur Avg Loss: 0.17519529, Log Avg loss: 0.18013592, Global Avg Loss: 0.65072489, Time: 0.0209 Steps: 99900, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001976, Sample Num: 31616, Cur Loss: 0.27123463, Cur Avg Loss: 0.17540661, Log Avg loss: 0.21695295, Global Avg Loss: 0.65068147, Time: 0.0209 Steps: 99910, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001986, Sample Num: 31776, Cur Loss: 0.67152274, Cur Avg Loss: 0.17564571, Log Avg loss: 0.22289182, Global Avg Loss: 0.65063866, Time: 0.0209 Steps: 99920, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001996, Sample Num: 31936, Cur Loss: 0.33042067, Cur Avg Loss: 0.17565011, Log Avg loss: 0.17652472, Global Avg Loss: 0.65059121, Time: 0.0209 Steps: 99930, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002006, Sample Num: 32096, Cur Loss: 0.11604233, Cur Avg Loss: 0.17557307, Log Avg loss: 0.16019608, Global Avg Loss: 0.65054215, Time: 0.0209 Steps: 99940, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002016, Sample Num: 32256, Cur Loss: 0.18305889, Cur Avg Loss: 0.17574999, Log Avg loss: 0.21123928, Global Avg Loss: 0.65049819, Time: 0.0209 Steps: 99950, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002026, Sample Num: 32416, Cur Loss: 0.12677360, Cur Avg Loss: 0.17575185, Log Avg loss: 0.17612721, Global Avg Loss: 0.65045074, Time: 0.0209 Steps: 99960, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002036, Sample Num: 32576, Cur Loss: 0.24586050, Cur Avg Loss: 0.17581868, Log Avg loss: 0.18935804, Global Avg Loss: 0.65040461, Time: 0.0209 Steps: 99970, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002046, Sample Num: 32736, Cur Loss: 0.27765942, Cur Avg Loss: 0.17566560, Log Avg loss: 0.14449811, Global Avg Loss: 0.65035401, Time: 0.0209 Steps: 99980, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002056, Sample Num: 32896, Cur Loss: 0.05912194, Cur Avg Loss: 0.17548308, Log Avg loss: 0.13814089, Global Avg Loss: 0.65030279, Time: 0.0210 Steps: 99990, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002066, Sample Num: 33056, Cur Loss: 0.25481662, Cur Avg Loss: 0.17539518, Log Avg loss: 0.15732302, Global Avg Loss: 0.65025349, Time: 0.0209 Steps: 100000, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002076, Sample Num: 33216, Cur Loss: 0.13724138, Cur Avg Loss: 0.17570443, Log Avg loss: 0.23959446, Global Avg Loss: 0.65021243, Time: 0.0210 Steps: 100010, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002086, Sample Num: 33376, Cur Loss: 0.18873361, Cur Avg Loss: 0.17559202, Log Avg loss: 0.15225648, Global Avg Loss: 0.65016264, Time: 0.0209 Steps: 100020, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002096, Sample Num: 33536, Cur Loss: 0.05739339, Cur Avg Loss: 0.17529691, Log Avg loss: 0.11373659, Global Avg Loss: 0.65010901, Time: 0.0209 Steps: 100030, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002106, Sample Num: 33696, Cur Loss: 0.11670204, Cur Avg Loss: 0.17505514, Log Avg loss: 0.12438038, Global Avg Loss: 0.65005646, Time: 0.0209 Steps: 100040, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002116, Sample Num: 33856, Cur Loss: 0.37878647, Cur Avg Loss: 0.17527624, Log Avg loss: 0.22183908, Global Avg Loss: 0.65001366, Time: 0.0209 Steps: 100050, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002126, Sample Num: 34016, Cur Loss: 0.18000183, Cur Avg Loss: 0.17521333, Log Avg loss: 0.16190201, Global Avg Loss: 0.64996488, Time: 0.0209 Steps: 100060, Updated lr: 0.000006 ***** Running evaluation checkpoint-100063 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-100063 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.588404, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.230834, "eval_total_loss": 162.276097, "eval_mae": 0.343956, "eval_mse": 0.230873, "eval_r2": 0.853242, "eval_sp_statistic": 0.905, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.926224, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.368659, "test_total_loss": 185.066742, "test_mae": 0.374837, "test_mse": 0.368789, "test_r2": 0.76198, "test_sp_statistic": 0.876798, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921952, "test_ps_pvalue": 0.0, "lr": 6.056899004267426e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6499489447810967, "train_cur_epoch_loss": 372.85882440395653, "train_cur_epoch_avg_loss": 0.1751333134823657, "train_cur_epoch_time": 44.588404417037964, "train_cur_epoch_avg_time": 0.02094335576187786, "epoch": 47, "step": 100063} ################################################## Training, Epoch: 0048, Batch: 000007, Sample Num: 112, Cur Loss: 0.10621533, Cur Avg Loss: 0.22936616, Log Avg loss: 0.19608455, Global Avg Loss: 0.64991952, Time: 0.0210 Steps: 100070, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000017, Sample Num: 272, Cur Loss: 0.06236745, Cur Avg Loss: 0.18356320, Log Avg loss: 0.15150113, Global Avg Loss: 0.64986972, Time: 0.0210 Steps: 100080, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000027, Sample Num: 432, Cur Loss: 0.19112821, Cur Avg Loss: 0.17029009, Log Avg loss: 0.14772581, Global Avg Loss: 0.64981955, Time: 0.0211 Steps: 100090, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000037, Sample Num: 592, Cur Loss: 0.24446519, Cur Avg Loss: 0.16435428, Log Avg loss: 0.14832760, Global Avg Loss: 0.64976945, Time: 0.0209 Steps: 100100, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000047, Sample Num: 752, Cur Loss: 0.12287775, Cur Avg Loss: 0.15699493, Log Avg loss: 0.12976534, Global Avg Loss: 0.64971751, Time: 0.0210 Steps: 100110, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000057, Sample Num: 912, Cur Loss: 0.09952218, Cur Avg Loss: 0.16048163, Log Avg loss: 0.17686909, Global Avg Loss: 0.64967028, Time: 0.0211 Steps: 100120, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000067, Sample Num: 1072, Cur Loss: 0.07948737, Cur Avg Loss: 0.15730941, Log Avg loss: 0.13922773, Global Avg Loss: 0.64961930, Time: 0.0211 Steps: 100130, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000077, Sample Num: 1232, Cur Loss: 0.10459381, Cur Avg Loss: 0.16518070, Log Avg loss: 0.21791836, Global Avg Loss: 0.64957620, Time: 0.0209 Steps: 100140, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000087, Sample Num: 1392, Cur Loss: 0.06543361, Cur Avg Loss: 0.16206048, Log Avg loss: 0.13803477, Global Avg Loss: 0.64952512, Time: 0.0211 Steps: 100150, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000097, Sample Num: 1552, Cur Loss: 0.71673840, Cur Avg Loss: 0.17393437, Log Avg loss: 0.27723728, Global Avg Loss: 0.64948795, Time: 0.0210 Steps: 100160, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000107, Sample Num: 1712, Cur Loss: 0.09285855, Cur Avg Loss: 0.17605248, Log Avg loss: 0.19659811, Global Avg Loss: 0.64944274, Time: 0.0210 Steps: 100170, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000117, Sample Num: 1872, Cur Loss: 0.32097840, Cur Avg Loss: 0.17487668, Log Avg loss: 0.16229560, Global Avg Loss: 0.64939411, Time: 0.0210 Steps: 100180, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000127, Sample Num: 2032, Cur Loss: 0.11595572, Cur Avg Loss: 0.17341932, Log Avg loss: 0.15636819, Global Avg Loss: 0.64934490, Time: 0.0209 Steps: 100190, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000137, Sample Num: 2192, Cur Loss: 0.38889039, Cur Avg Loss: 0.17411260, Log Avg loss: 0.18291733, Global Avg Loss: 0.64929835, Time: 0.0210 Steps: 100200, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000147, Sample Num: 2352, Cur Loss: 0.14184621, Cur Avg Loss: 0.17579741, Log Avg loss: 0.19887934, Global Avg Loss: 0.64925340, Time: 0.0211 Steps: 100210, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000157, Sample Num: 2512, Cur Loss: 0.17978804, Cur Avg Loss: 0.17741467, Log Avg loss: 0.20118831, Global Avg Loss: 0.64920869, Time: 0.0211 Steps: 100220, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000167, Sample Num: 2672, Cur Loss: 0.39188486, Cur Avg Loss: 0.17583705, Log Avg loss: 0.15106838, Global Avg Loss: 0.64915899, Time: 0.0209 Steps: 100230, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000177, Sample Num: 2832, Cur Loss: 0.34895623, Cur Avg Loss: 0.17536498, Log Avg loss: 0.16748139, Global Avg Loss: 0.64911094, Time: 0.0210 Steps: 100240, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000187, Sample Num: 2992, Cur Loss: 0.05893822, Cur Avg Loss: 0.17117236, Log Avg loss: 0.09696307, Global Avg Loss: 0.64905587, Time: 0.0211 Steps: 100250, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000197, Sample Num: 3152, Cur Loss: 0.31335825, Cur Avg Loss: 0.17266165, Log Avg loss: 0.20051133, Global Avg Loss: 0.64901113, Time: 0.0210 Steps: 100260, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000207, Sample Num: 3312, Cur Loss: 0.12485820, Cur Avg Loss: 0.16965057, Log Avg loss: 0.11033242, Global Avg Loss: 0.64895740, Time: 0.0211 Steps: 100270, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000217, Sample Num: 3472, Cur Loss: 0.25116432, Cur Avg Loss: 0.17325856, Log Avg loss: 0.24794386, Global Avg Loss: 0.64891741, Time: 0.0210 Steps: 100280, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000227, Sample Num: 3632, Cur Loss: 0.15371893, Cur Avg Loss: 0.17107641, Log Avg loss: 0.12372379, Global Avg Loss: 0.64886505, Time: 0.0209 Steps: 100290, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000237, Sample Num: 3792, Cur Loss: 0.22333400, Cur Avg Loss: 0.16977304, Log Avg loss: 0.14018657, Global Avg Loss: 0.64881433, Time: 0.0210 Steps: 100300, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000247, Sample Num: 3952, Cur Loss: 0.18968877, Cur Avg Loss: 0.16918447, Log Avg loss: 0.15523524, Global Avg Loss: 0.64876513, Time: 0.0211 Steps: 100310, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000257, Sample Num: 4112, Cur Loss: 0.12705421, Cur Avg Loss: 0.17072397, Log Avg loss: 0.20874969, Global Avg Loss: 0.64872127, Time: 0.0246 Steps: 100320, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000267, Sample Num: 4272, Cur Loss: 0.06900901, Cur Avg Loss: 0.16940093, Log Avg loss: 0.13539875, Global Avg Loss: 0.64867010, Time: 0.0210 Steps: 100330, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000277, Sample Num: 4432, Cur Loss: 0.16166818, Cur Avg Loss: 0.16937354, Log Avg loss: 0.16864222, Global Avg Loss: 0.64862226, Time: 0.0208 Steps: 100340, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000287, Sample Num: 4592, Cur Loss: 0.12336624, Cur Avg Loss: 0.16881250, Log Avg loss: 0.15327171, Global Avg Loss: 0.64857290, Time: 0.0209 Steps: 100350, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000297, Sample Num: 4752, Cur Loss: 0.85285151, Cur Avg Loss: 0.16989891, Log Avg loss: 0.20107897, Global Avg Loss: 0.64852831, Time: 0.0210 Steps: 100360, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000307, Sample Num: 4912, Cur Loss: 0.13909607, Cur Avg Loss: 0.17019949, Log Avg loss: 0.17912649, Global Avg Loss: 0.64848154, Time: 0.0210 Steps: 100370, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000317, Sample Num: 5072, Cur Loss: 0.11382630, Cur Avg Loss: 0.17178695, Log Avg loss: 0.22052197, Global Avg Loss: 0.64843891, Time: 0.0209 Steps: 100380, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000327, Sample Num: 5232, Cur Loss: 0.15904322, Cur Avg Loss: 0.17175176, Log Avg loss: 0.17063651, Global Avg Loss: 0.64839131, Time: 0.0209 Steps: 100390, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000337, Sample Num: 5392, Cur Loss: 0.25127503, Cur Avg Loss: 0.17135710, Log Avg loss: 0.15845170, Global Avg Loss: 0.64834252, Time: 0.0208 Steps: 100400, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000347, Sample Num: 5552, Cur Loss: 0.14158061, Cur Avg Loss: 0.17128945, Log Avg loss: 0.16900941, Global Avg Loss: 0.64829478, Time: 0.0210 Steps: 100410, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000357, Sample Num: 5712, Cur Loss: 0.11953890, Cur Avg Loss: 0.17116514, Log Avg loss: 0.16685188, Global Avg Loss: 0.64824684, Time: 0.0210 Steps: 100420, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000367, Sample Num: 5872, Cur Loss: 0.05184448, Cur Avg Loss: 0.17075391, Log Avg loss: 0.15607266, Global Avg Loss: 0.64819783, Time: 0.0209 Steps: 100430, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000377, Sample Num: 6032, Cur Loss: 0.09928322, Cur Avg Loss: 0.17029421, Log Avg loss: 0.15342340, Global Avg Loss: 0.64814857, Time: 0.0209 Steps: 100440, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000387, Sample Num: 6192, Cur Loss: 0.19870862, Cur Avg Loss: 0.16945314, Log Avg loss: 0.13774491, Global Avg Loss: 0.64809776, Time: 0.0210 Steps: 100450, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000397, Sample Num: 6352, Cur Loss: 0.39300740, Cur Avg Loss: 0.16965372, Log Avg loss: 0.17741622, Global Avg Loss: 0.64805090, Time: 0.0209 Steps: 100460, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000407, Sample Num: 6512, Cur Loss: 0.06137953, Cur Avg Loss: 0.16914385, Log Avg loss: 0.14890173, Global Avg Loss: 0.64800122, Time: 0.0210 Steps: 100470, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000417, Sample Num: 6672, Cur Loss: 0.19566239, Cur Avg Loss: 0.16848572, Log Avg loss: 0.14169974, Global Avg Loss: 0.64795083, Time: 0.0209 Steps: 100480, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000427, Sample Num: 6832, Cur Loss: 0.05819479, Cur Avg Loss: 0.16994259, Log Avg loss: 0.23069408, Global Avg Loss: 0.64790931, Time: 0.0208 Steps: 100490, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000437, Sample Num: 6992, Cur Loss: 0.08138342, Cur Avg Loss: 0.17002012, Log Avg loss: 0.17333076, Global Avg Loss: 0.64786209, Time: 0.0209 Steps: 100500, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000447, Sample Num: 7152, Cur Loss: 0.11288007, Cur Avg Loss: 0.17050678, Log Avg loss: 0.19177409, Global Avg Loss: 0.64781671, Time: 0.0208 Steps: 100510, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000457, Sample Num: 7312, Cur Loss: 0.06990921, Cur Avg Loss: 0.17015819, Log Avg loss: 0.15457603, Global Avg Loss: 0.64776764, Time: 0.0210 Steps: 100520, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000467, Sample Num: 7472, Cur Loss: 0.11870250, Cur Avg Loss: 0.17025609, Log Avg loss: 0.17473004, Global Avg Loss: 0.64772059, Time: 0.0208 Steps: 100530, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000477, Sample Num: 7632, Cur Loss: 0.14622498, Cur Avg Loss: 0.17046966, Log Avg loss: 0.18044351, Global Avg Loss: 0.64767411, Time: 0.0210 Steps: 100540, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000487, Sample Num: 7792, Cur Loss: 0.05518697, Cur Avg Loss: 0.16962467, Log Avg loss: 0.12931841, Global Avg Loss: 0.64762256, Time: 0.0208 Steps: 100550, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000497, Sample Num: 7952, Cur Loss: 0.35376418, Cur Avg Loss: 0.16951399, Log Avg loss: 0.16412393, Global Avg Loss: 0.64757448, Time: 0.0210 Steps: 100560, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000507, Sample Num: 8112, Cur Loss: 0.07047272, Cur Avg Loss: 0.16973152, Log Avg loss: 0.18054310, Global Avg Loss: 0.64752804, Time: 0.0209 Steps: 100570, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000517, Sample Num: 8272, Cur Loss: 0.06445514, Cur Avg Loss: 0.17003057, Log Avg loss: 0.18519232, Global Avg Loss: 0.64748207, Time: 0.0245 Steps: 100580, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000527, Sample Num: 8432, Cur Loss: 0.12617098, Cur Avg Loss: 0.17114634, Log Avg loss: 0.22883153, Global Avg Loss: 0.64744046, Time: 0.0208 Steps: 100590, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000537, Sample Num: 8592, Cur Loss: 0.15148214, Cur Avg Loss: 0.17108739, Log Avg loss: 0.16798057, Global Avg Loss: 0.64739280, Time: 0.0208 Steps: 100600, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000547, Sample Num: 8752, Cur Loss: 0.25365475, Cur Avg Loss: 0.17110993, Log Avg loss: 0.17232046, Global Avg Loss: 0.64734558, Time: 0.0208 Steps: 100610, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000557, Sample Num: 8912, Cur Loss: 0.15583792, Cur Avg Loss: 0.16995071, Log Avg loss: 0.10654145, Global Avg Loss: 0.64729183, Time: 0.0208 Steps: 100620, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000567, Sample Num: 9072, Cur Loss: 0.26533511, Cur Avg Loss: 0.17021139, Log Avg loss: 0.18473128, Global Avg Loss: 0.64724586, Time: 0.0208 Steps: 100630, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000577, Sample Num: 9232, Cur Loss: 0.13395855, Cur Avg Loss: 0.16945363, Log Avg loss: 0.12648881, Global Avg Loss: 0.64719412, Time: 0.0208 Steps: 100640, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000587, Sample Num: 9392, Cur Loss: 0.20197517, Cur Avg Loss: 0.16887380, Log Avg loss: 0.13541716, Global Avg Loss: 0.64714327, Time: 0.0208 Steps: 100650, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000597, Sample Num: 9552, Cur Loss: 0.35310048, Cur Avg Loss: 0.16891351, Log Avg loss: 0.17124448, Global Avg Loss: 0.64709599, Time: 0.0208 Steps: 100660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000607, Sample Num: 9712, Cur Loss: 0.22809413, Cur Avg Loss: 0.16808578, Log Avg loss: 0.11867025, Global Avg Loss: 0.64704350, Time: 0.0208 Steps: 100670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000617, Sample Num: 9872, Cur Loss: 0.13629207, Cur Avg Loss: 0.16784921, Log Avg loss: 0.15348938, Global Avg Loss: 0.64699448, Time: 0.0208 Steps: 100680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000627, Sample Num: 10032, Cur Loss: 0.18223305, Cur Avg Loss: 0.16822023, Log Avg loss: 0.19111248, Global Avg Loss: 0.64694920, Time: 0.0208 Steps: 100690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000637, Sample Num: 10192, Cur Loss: 0.27362812, Cur Avg Loss: 0.16862911, Log Avg loss: 0.19426597, Global Avg Loss: 0.64690425, Time: 0.0208 Steps: 100700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000647, Sample Num: 10352, Cur Loss: 0.10874616, Cur Avg Loss: 0.16849336, Log Avg loss: 0.15984594, Global Avg Loss: 0.64685589, Time: 0.0208 Steps: 100710, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000657, Sample Num: 10512, Cur Loss: 0.28817755, Cur Avg Loss: 0.16856650, Log Avg loss: 0.17329876, Global Avg Loss: 0.64680887, Time: 0.0208 Steps: 100720, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000667, Sample Num: 10672, Cur Loss: 0.05241778, Cur Avg Loss: 0.16849623, Log Avg loss: 0.16387944, Global Avg Loss: 0.64676093, Time: 0.0208 Steps: 100730, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000677, Sample Num: 10832, Cur Loss: 0.25854892, Cur Avg Loss: 0.16839742, Log Avg loss: 0.16180659, Global Avg Loss: 0.64671279, Time: 0.0208 Steps: 100740, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000687, Sample Num: 10992, Cur Loss: 0.04603361, Cur Avg Loss: 0.16780532, Log Avg loss: 0.12772041, Global Avg Loss: 0.64666128, Time: 0.0208 Steps: 100750, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000697, Sample Num: 11152, Cur Loss: 0.08109035, Cur Avg Loss: 0.16902654, Log Avg loss: 0.25292436, Global Avg Loss: 0.64662220, Time: 0.0209 Steps: 100760, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000707, Sample Num: 11312, Cur Loss: 0.22683534, Cur Avg Loss: 0.16972349, Log Avg loss: 0.21830109, Global Avg Loss: 0.64657969, Time: 0.0208 Steps: 100770, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000717, Sample Num: 11472, Cur Loss: 0.23377477, Cur Avg Loss: 0.16927041, Log Avg loss: 0.13723714, Global Avg Loss: 0.64652915, Time: 0.0208 Steps: 100780, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000727, Sample Num: 11632, Cur Loss: 0.09333953, Cur Avg Loss: 0.16858997, Log Avg loss: 0.11980244, Global Avg Loss: 0.64647689, Time: 0.0208 Steps: 100790, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000737, Sample Num: 11792, Cur Loss: 0.19549155, Cur Avg Loss: 0.16917540, Log Avg loss: 0.21173637, Global Avg Loss: 0.64643377, Time: 0.0208 Steps: 100800, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000747, Sample Num: 11952, Cur Loss: 0.20705074, Cur Avg Loss: 0.16938872, Log Avg loss: 0.18511006, Global Avg Loss: 0.64638800, Time: 0.0208 Steps: 100810, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000757, Sample Num: 12112, Cur Loss: 0.16377166, Cur Avg Loss: 0.16961397, Log Avg loss: 0.18644070, Global Avg Loss: 0.64634238, Time: 0.0208 Steps: 100820, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000767, Sample Num: 12272, Cur Loss: 0.08739202, Cur Avg Loss: 0.16946584, Log Avg loss: 0.15825232, Global Avg Loss: 0.64629398, Time: 0.0209 Steps: 100830, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000777, Sample Num: 12432, Cur Loss: 0.15801542, Cur Avg Loss: 0.17042095, Log Avg loss: 0.24367768, Global Avg Loss: 0.64625405, Time: 0.0210 Steps: 100840, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000787, Sample Num: 12592, Cur Loss: 0.17154624, Cur Avg Loss: 0.17052750, Log Avg loss: 0.17880660, Global Avg Loss: 0.64620770, Time: 0.0209 Steps: 100850, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000797, Sample Num: 12752, Cur Loss: 0.30153295, Cur Avg Loss: 0.17056819, Log Avg loss: 0.17377060, Global Avg Loss: 0.64616086, Time: 0.0209 Steps: 100860, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000807, Sample Num: 12912, Cur Loss: 0.14892244, Cur Avg Loss: 0.17051032, Log Avg loss: 0.16589766, Global Avg Loss: 0.64611325, Time: 0.0209 Steps: 100870, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000817, Sample Num: 13072, Cur Loss: 0.26824242, Cur Avg Loss: 0.17051375, Log Avg loss: 0.17079110, Global Avg Loss: 0.64606613, Time: 0.0209 Steps: 100880, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000827, Sample Num: 13232, Cur Loss: 0.18089178, Cur Avg Loss: 0.16982885, Log Avg loss: 0.11387200, Global Avg Loss: 0.64601338, Time: 0.0209 Steps: 100890, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000837, Sample Num: 13392, Cur Loss: 0.14092404, Cur Avg Loss: 0.16966956, Log Avg loss: 0.15649655, Global Avg Loss: 0.64596486, Time: 0.0209 Steps: 100900, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000847, Sample Num: 13552, Cur Loss: 0.21371758, Cur Avg Loss: 0.16986773, Log Avg loss: 0.18645458, Global Avg Loss: 0.64591933, Time: 0.0209 Steps: 100910, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000857, Sample Num: 13712, Cur Loss: 0.13623400, Cur Avg Loss: 0.16951891, Log Avg loss: 0.13997366, Global Avg Loss: 0.64586919, Time: 0.0210 Steps: 100920, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000867, Sample Num: 13872, Cur Loss: 0.33513409, Cur Avg Loss: 0.16976806, Log Avg loss: 0.19111995, Global Avg Loss: 0.64582414, Time: 0.0209 Steps: 100930, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000877, Sample Num: 14032, Cur Loss: 0.22954348, Cur Avg Loss: 0.16967220, Log Avg loss: 0.16136112, Global Avg Loss: 0.64577614, Time: 0.0209 Steps: 100940, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000887, Sample Num: 14192, Cur Loss: 0.39377806, Cur Avg Loss: 0.16933620, Log Avg loss: 0.13986942, Global Avg Loss: 0.64572603, Time: 0.0209 Steps: 100950, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000897, Sample Num: 14352, Cur Loss: 0.21791480, Cur Avg Loss: 0.16979279, Log Avg loss: 0.21029187, Global Avg Loss: 0.64568290, Time: 0.0209 Steps: 100960, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000907, Sample Num: 14512, Cur Loss: 0.23097537, Cur Avg Loss: 0.16953768, Log Avg loss: 0.14665447, Global Avg Loss: 0.64563347, Time: 0.0209 Steps: 100970, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000917, Sample Num: 14672, Cur Loss: 0.17632869, Cur Avg Loss: 0.16954534, Log Avg loss: 0.17024016, Global Avg Loss: 0.64558640, Time: 0.0209 Steps: 100980, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000927, Sample Num: 14832, Cur Loss: 0.10310400, Cur Avg Loss: 0.16966535, Log Avg loss: 0.18067010, Global Avg Loss: 0.64554036, Time: 0.0209 Steps: 100990, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000937, Sample Num: 14992, Cur Loss: 0.19043280, Cur Avg Loss: 0.16959293, Log Avg loss: 0.16288042, Global Avg Loss: 0.64549257, Time: 0.0209 Steps: 101000, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000947, Sample Num: 15152, Cur Loss: 0.02522864, Cur Avg Loss: 0.16956177, Log Avg loss: 0.16664198, Global Avg Loss: 0.64544517, Time: 0.0209 Steps: 101010, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000957, Sample Num: 15312, Cur Loss: 0.15786129, Cur Avg Loss: 0.16963969, Log Avg loss: 0.17701801, Global Avg Loss: 0.64539880, Time: 0.0209 Steps: 101020, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000967, Sample Num: 15472, Cur Loss: 0.02773725, Cur Avg Loss: 0.16982430, Log Avg loss: 0.18749176, Global Avg Loss: 0.64535347, Time: 0.0209 Steps: 101030, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000977, Sample Num: 15632, Cur Loss: 0.13144100, Cur Avg Loss: 0.17046312, Log Avg loss: 0.23223727, Global Avg Loss: 0.64531259, Time: 0.0209 Steps: 101040, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000987, Sample Num: 15792, Cur Loss: 0.24299434, Cur Avg Loss: 0.17096284, Log Avg loss: 0.21978512, Global Avg Loss: 0.64527048, Time: 0.0209 Steps: 101050, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000997, Sample Num: 15952, Cur Loss: 0.05238113, Cur Avg Loss: 0.17117199, Log Avg loss: 0.19181510, Global Avg Loss: 0.64522561, Time: 0.0209 Steps: 101060, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001007, Sample Num: 16112, Cur Loss: 0.08320858, Cur Avg Loss: 0.17111559, Log Avg loss: 0.16549286, Global Avg Loss: 0.64517814, Time: 0.0209 Steps: 101070, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001017, Sample Num: 16272, Cur Loss: 0.17260192, Cur Avg Loss: 0.17095621, Log Avg loss: 0.15490652, Global Avg Loss: 0.64512964, Time: 0.0209 Steps: 101080, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001027, Sample Num: 16432, Cur Loss: 0.22235164, Cur Avg Loss: 0.17110246, Log Avg loss: 0.18597563, Global Avg Loss: 0.64508422, Time: 0.0252 Steps: 101090, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001037, Sample Num: 16592, Cur Loss: 0.14237407, Cur Avg Loss: 0.17143527, Log Avg loss: 0.20561565, Global Avg Loss: 0.64504075, Time: 0.0209 Steps: 101100, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001047, Sample Num: 16752, Cur Loss: 0.10470422, Cur Avg Loss: 0.17162685, Log Avg loss: 0.19149336, Global Avg Loss: 0.64499589, Time: 0.0209 Steps: 101110, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001057, Sample Num: 16912, Cur Loss: 0.34573385, Cur Avg Loss: 0.17224586, Log Avg loss: 0.23705589, Global Avg Loss: 0.64495555, Time: 0.0209 Steps: 101120, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001067, Sample Num: 17072, Cur Loss: 0.05510621, Cur Avg Loss: 0.17319214, Log Avg loss: 0.27321397, Global Avg Loss: 0.64491879, Time: 0.0209 Steps: 101130, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001077, Sample Num: 17232, Cur Loss: 0.13027337, Cur Avg Loss: 0.17281092, Log Avg loss: 0.13213454, Global Avg Loss: 0.64486809, Time: 0.0209 Steps: 101140, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001087, Sample Num: 17392, Cur Loss: 0.02838197, Cur Avg Loss: 0.17321794, Log Avg loss: 0.21705499, Global Avg Loss: 0.64482580, Time: 0.0209 Steps: 101150, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001097, Sample Num: 17552, Cur Loss: 0.16618076, Cur Avg Loss: 0.17333915, Log Avg loss: 0.18651413, Global Avg Loss: 0.64478049, Time: 0.0209 Steps: 101160, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001107, Sample Num: 17712, Cur Loss: 0.15785301, Cur Avg Loss: 0.17323483, Log Avg loss: 0.16179089, Global Avg Loss: 0.64473275, Time: 0.0209 Steps: 101170, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001117, Sample Num: 17872, Cur Loss: 0.11677718, Cur Avg Loss: 0.17354828, Log Avg loss: 0.20824717, Global Avg Loss: 0.64468961, Time: 0.0209 Steps: 101180, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001127, Sample Num: 18032, Cur Loss: 0.07004392, Cur Avg Loss: 0.17322416, Log Avg loss: 0.13702037, Global Avg Loss: 0.64463944, Time: 0.0209 Steps: 101190, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001137, Sample Num: 18192, Cur Loss: 0.21075624, Cur Avg Loss: 0.17312033, Log Avg loss: 0.16141816, Global Avg Loss: 0.64459169, Time: 0.0209 Steps: 101200, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001147, Sample Num: 18352, Cur Loss: 0.28033036, Cur Avg Loss: 0.17308329, Log Avg loss: 0.16887172, Global Avg Loss: 0.64454469, Time: 0.0209 Steps: 101210, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001157, Sample Num: 18512, Cur Loss: 0.18282109, Cur Avg Loss: 0.17244720, Log Avg loss: 0.09948829, Global Avg Loss: 0.64449084, Time: 0.0209 Steps: 101220, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001167, Sample Num: 18672, Cur Loss: 0.17005421, Cur Avg Loss: 0.17203213, Log Avg loss: 0.12400797, Global Avg Loss: 0.64443942, Time: 0.0210 Steps: 101230, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001177, Sample Num: 18832, Cur Loss: 0.02973022, Cur Avg Loss: 0.17226283, Log Avg loss: 0.19918595, Global Avg Loss: 0.64439544, Time: 0.0209 Steps: 101240, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001187, Sample Num: 18992, Cur Loss: 0.03468478, Cur Avg Loss: 0.17190981, Log Avg loss: 0.13035928, Global Avg Loss: 0.64434467, Time: 0.0209 Steps: 101250, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001197, Sample Num: 19152, Cur Loss: 0.10371363, Cur Avg Loss: 0.17210687, Log Avg loss: 0.19549765, Global Avg Loss: 0.64430035, Time: 0.0209 Steps: 101260, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001207, Sample Num: 19312, Cur Loss: 0.35567710, Cur Avg Loss: 0.17186750, Log Avg loss: 0.14321535, Global Avg Loss: 0.64425087, Time: 0.0209 Steps: 101270, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001217, Sample Num: 19472, Cur Loss: 0.11384170, Cur Avg Loss: 0.17177315, Log Avg loss: 0.16038490, Global Avg Loss: 0.64420309, Time: 0.0209 Steps: 101280, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001227, Sample Num: 19632, Cur Loss: 0.26862353, Cur Avg Loss: 0.17181897, Log Avg loss: 0.17739521, Global Avg Loss: 0.64415701, Time: 0.0209 Steps: 101290, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001237, Sample Num: 19792, Cur Loss: 0.12723166, Cur Avg Loss: 0.17181421, Log Avg loss: 0.17123009, Global Avg Loss: 0.64411032, Time: 0.0209 Steps: 101300, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001247, Sample Num: 19952, Cur Loss: 0.13739303, Cur Avg Loss: 0.17193898, Log Avg loss: 0.18737275, Global Avg Loss: 0.64406524, Time: 0.0209 Steps: 101310, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001257, Sample Num: 20112, Cur Loss: 0.07826325, Cur Avg Loss: 0.17197099, Log Avg loss: 0.17596301, Global Avg Loss: 0.64401904, Time: 0.0210 Steps: 101320, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001267, Sample Num: 20272, Cur Loss: 0.09363478, Cur Avg Loss: 0.17197839, Log Avg loss: 0.17290804, Global Avg Loss: 0.64397254, Time: 0.0209 Steps: 101330, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001277, Sample Num: 20432, Cur Loss: 0.25595111, Cur Avg Loss: 0.17156963, Log Avg loss: 0.11977981, Global Avg Loss: 0.64392082, Time: 0.0209 Steps: 101340, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001287, Sample Num: 20592, Cur Loss: 0.08014217, Cur Avg Loss: 0.17138462, Log Avg loss: 0.14775950, Global Avg Loss: 0.64387186, Time: 0.0216 Steps: 101350, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001297, Sample Num: 20752, Cur Loss: 0.16081491, Cur Avg Loss: 0.17136138, Log Avg loss: 0.16836960, Global Avg Loss: 0.64382495, Time: 0.0211 Steps: 101360, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001307, Sample Num: 20912, Cur Loss: 0.05192979, Cur Avg Loss: 0.17142515, Log Avg loss: 0.17969673, Global Avg Loss: 0.64377916, Time: 0.0210 Steps: 101370, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001317, Sample Num: 21072, Cur Loss: 0.21387735, Cur Avg Loss: 0.17180122, Log Avg loss: 0.22095301, Global Avg Loss: 0.64373746, Time: 0.0209 Steps: 101380, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001327, Sample Num: 21232, Cur Loss: 0.19804712, Cur Avg Loss: 0.17198241, Log Avg loss: 0.19584547, Global Avg Loss: 0.64369328, Time: 0.0210 Steps: 101390, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001337, Sample Num: 21392, Cur Loss: 0.06105358, Cur Avg Loss: 0.17162776, Log Avg loss: 0.12456525, Global Avg Loss: 0.64364209, Time: 0.0210 Steps: 101400, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001347, Sample Num: 21552, Cur Loss: 0.17146859, Cur Avg Loss: 0.17151779, Log Avg loss: 0.15681578, Global Avg Loss: 0.64359408, Time: 0.0209 Steps: 101410, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001357, Sample Num: 21712, Cur Loss: 0.10848241, Cur Avg Loss: 0.17120621, Log Avg loss: 0.12923653, Global Avg Loss: 0.64354337, Time: 0.0209 Steps: 101420, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001367, Sample Num: 21872, Cur Loss: 0.11793279, Cur Avg Loss: 0.17122087, Log Avg loss: 0.17321011, Global Avg Loss: 0.64349699, Time: 0.0209 Steps: 101430, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001377, Sample Num: 22032, Cur Loss: 0.23885643, Cur Avg Loss: 0.17131265, Log Avg loss: 0.18385857, Global Avg Loss: 0.64345168, Time: 0.0210 Steps: 101440, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001387, Sample Num: 22192, Cur Loss: 0.06478991, Cur Avg Loss: 0.17161791, Log Avg loss: 0.21365277, Global Avg Loss: 0.64340932, Time: 0.0210 Steps: 101450, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001397, Sample Num: 22352, Cur Loss: 0.16663662, Cur Avg Loss: 0.17167965, Log Avg loss: 0.18024295, Global Avg Loss: 0.64336367, Time: 0.0209 Steps: 101460, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001407, Sample Num: 22512, Cur Loss: 0.14937653, Cur Avg Loss: 0.17171260, Log Avg loss: 0.17631509, Global Avg Loss: 0.64331764, Time: 0.0209 Steps: 101470, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001417, Sample Num: 22672, Cur Loss: 0.29996037, Cur Avg Loss: 0.17188080, Log Avg loss: 0.19554713, Global Avg Loss: 0.64327352, Time: 0.0210 Steps: 101480, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001427, Sample Num: 22832, Cur Loss: 0.16125125, Cur Avg Loss: 0.17220935, Log Avg loss: 0.21876484, Global Avg Loss: 0.64323169, Time: 0.0210 Steps: 101490, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001437, Sample Num: 22992, Cur Loss: 0.38596064, Cur Avg Loss: 0.17203897, Log Avg loss: 0.14772464, Global Avg Loss: 0.64318287, Time: 0.0210 Steps: 101500, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001447, Sample Num: 23152, Cur Loss: 0.32696545, Cur Avg Loss: 0.17207047, Log Avg loss: 0.17659764, Global Avg Loss: 0.64313691, Time: 0.0209 Steps: 101510, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001457, Sample Num: 23312, Cur Loss: 0.20805156, Cur Avg Loss: 0.17232231, Log Avg loss: 0.20876306, Global Avg Loss: 0.64309412, Time: 0.0209 Steps: 101520, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001467, Sample Num: 23472, Cur Loss: 0.55799204, Cur Avg Loss: 0.17242565, Log Avg loss: 0.18748259, Global Avg Loss: 0.64304924, Time: 0.0210 Steps: 101530, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001477, Sample Num: 23632, Cur Loss: 0.09307215, Cur Avg Loss: 0.17207797, Log Avg loss: 0.12107324, Global Avg Loss: 0.64299784, Time: 0.0209 Steps: 101540, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001487, Sample Num: 23792, Cur Loss: 0.13807584, Cur Avg Loss: 0.17203287, Log Avg loss: 0.16537232, Global Avg Loss: 0.64295080, Time: 0.0209 Steps: 101550, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001497, Sample Num: 23952, Cur Loss: 0.05395007, Cur Avg Loss: 0.17213985, Log Avg loss: 0.18804714, Global Avg Loss: 0.64290601, Time: 0.0210 Steps: 101560, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001507, Sample Num: 24112, Cur Loss: 0.15171093, Cur Avg Loss: 0.17204937, Log Avg loss: 0.15850508, Global Avg Loss: 0.64285832, Time: 0.0210 Steps: 101570, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001517, Sample Num: 24272, Cur Loss: 0.10829818, Cur Avg Loss: 0.17197962, Log Avg loss: 0.16146771, Global Avg Loss: 0.64281093, Time: 0.0208 Steps: 101580, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001527, Sample Num: 24432, Cur Loss: 0.09511615, Cur Avg Loss: 0.17179636, Log Avg loss: 0.14399594, Global Avg Loss: 0.64276183, Time: 0.0208 Steps: 101590, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001537, Sample Num: 24592, Cur Loss: 0.12779108, Cur Avg Loss: 0.17215782, Log Avg loss: 0.22735253, Global Avg Loss: 0.64272094, Time: 0.0247 Steps: 101600, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001547, Sample Num: 24752, Cur Loss: 0.14766717, Cur Avg Loss: 0.17223413, Log Avg loss: 0.18396336, Global Avg Loss: 0.64267579, Time: 0.0208 Steps: 101610, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001557, Sample Num: 24912, Cur Loss: 0.06129473, Cur Avg Loss: 0.17223332, Log Avg loss: 0.17210785, Global Avg Loss: 0.64262949, Time: 0.0208 Steps: 101620, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001567, Sample Num: 25072, Cur Loss: 0.11321244, Cur Avg Loss: 0.17186960, Log Avg loss: 0.11523810, Global Avg Loss: 0.64257759, Time: 0.0209 Steps: 101630, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001577, Sample Num: 25232, Cur Loss: 0.12049337, Cur Avg Loss: 0.17203461, Log Avg loss: 0.19789171, Global Avg Loss: 0.64253384, Time: 0.0208 Steps: 101640, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001587, Sample Num: 25392, Cur Loss: 0.23114660, Cur Avg Loss: 0.17243358, Log Avg loss: 0.23535135, Global Avg Loss: 0.64249379, Time: 0.0208 Steps: 101650, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001597, Sample Num: 25552, Cur Loss: 0.26796231, Cur Avg Loss: 0.17247452, Log Avg loss: 0.17897246, Global Avg Loss: 0.64244819, Time: 0.0208 Steps: 101660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001607, Sample Num: 25712, Cur Loss: 0.21870725, Cur Avg Loss: 0.17266664, Log Avg loss: 0.20334739, Global Avg Loss: 0.64240500, Time: 0.0208 Steps: 101670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001617, Sample Num: 25872, Cur Loss: 0.32108882, Cur Avg Loss: 0.17246602, Log Avg loss: 0.14022708, Global Avg Loss: 0.64235561, Time: 0.0208 Steps: 101680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001627, Sample Num: 26032, Cur Loss: 0.08601832, Cur Avg Loss: 0.17249406, Log Avg loss: 0.17702699, Global Avg Loss: 0.64230985, Time: 0.0208 Steps: 101690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001637, Sample Num: 26192, Cur Loss: 0.21499601, Cur Avg Loss: 0.17230471, Log Avg loss: 0.14149825, Global Avg Loss: 0.64226061, Time: 0.0208 Steps: 101700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001647, Sample Num: 26352, Cur Loss: 0.23229000, Cur Avg Loss: 0.17222716, Log Avg loss: 0.15953176, Global Avg Loss: 0.64221315, Time: 0.0208 Steps: 101710, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001657, Sample Num: 26512, Cur Loss: 0.08358955, Cur Avg Loss: 0.17211326, Log Avg loss: 0.15335415, Global Avg Loss: 0.64216509, Time: 0.0210 Steps: 101720, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001667, Sample Num: 26672, Cur Loss: 0.18263088, Cur Avg Loss: 0.17227627, Log Avg loss: 0.19928750, Global Avg Loss: 0.64212156, Time: 0.0208 Steps: 101730, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001677, Sample Num: 26832, Cur Loss: 0.22147062, Cur Avg Loss: 0.17211150, Log Avg loss: 0.14464446, Global Avg Loss: 0.64207266, Time: 0.0208 Steps: 101740, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001687, Sample Num: 26992, Cur Loss: 0.10837662, Cur Avg Loss: 0.17191129, Log Avg loss: 0.13833513, Global Avg Loss: 0.64202315, Time: 0.0208 Steps: 101750, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001697, Sample Num: 27152, Cur Loss: 0.11303748, Cur Avg Loss: 0.17197565, Log Avg loss: 0.18283310, Global Avg Loss: 0.64197803, Time: 0.0208 Steps: 101760, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001707, Sample Num: 27312, Cur Loss: 0.19521098, Cur Avg Loss: 0.17189348, Log Avg loss: 0.15794991, Global Avg Loss: 0.64193047, Time: 0.0208 Steps: 101770, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001717, Sample Num: 27472, Cur Loss: 0.08323437, Cur Avg Loss: 0.17211408, Log Avg loss: 0.20977099, Global Avg Loss: 0.64188800, Time: 0.0208 Steps: 101780, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001727, Sample Num: 27632, Cur Loss: 0.38252723, Cur Avg Loss: 0.17202193, Log Avg loss: 0.15619850, Global Avg Loss: 0.64184029, Time: 0.0208 Steps: 101790, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001737, Sample Num: 27792, Cur Loss: 0.37529257, Cur Avg Loss: 0.17179905, Log Avg loss: 0.13330813, Global Avg Loss: 0.64179034, Time: 0.0208 Steps: 101800, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001747, Sample Num: 27952, Cur Loss: 0.04704475, Cur Avg Loss: 0.17173023, Log Avg loss: 0.15977688, Global Avg Loss: 0.64174299, Time: 0.0208 Steps: 101810, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001757, Sample Num: 28112, Cur Loss: 0.32117203, Cur Avg Loss: 0.17176361, Log Avg loss: 0.17759446, Global Avg Loss: 0.64169741, Time: 0.0208 Steps: 101820, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001767, Sample Num: 28272, Cur Loss: 0.08080432, Cur Avg Loss: 0.17149834, Log Avg loss: 0.12489035, Global Avg Loss: 0.64164665, Time: 0.0208 Steps: 101830, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001777, Sample Num: 28432, Cur Loss: 0.30029052, Cur Avg Loss: 0.17180423, Log Avg loss: 0.22585601, Global Avg Loss: 0.64160583, Time: 0.0208 Steps: 101840, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001787, Sample Num: 28592, Cur Loss: 0.25711307, Cur Avg Loss: 0.17179546, Log Avg loss: 0.17023563, Global Avg Loss: 0.64155955, Time: 0.0208 Steps: 101850, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001797, Sample Num: 28752, Cur Loss: 0.04310492, Cur Avg Loss: 0.17184808, Log Avg loss: 0.18125140, Global Avg Loss: 0.64151436, Time: 0.0208 Steps: 101860, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001807, Sample Num: 28912, Cur Loss: 0.35885292, Cur Avg Loss: 0.17200998, Log Avg loss: 0.20110444, Global Avg Loss: 0.64147112, Time: 0.0211 Steps: 101870, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001817, Sample Num: 29072, Cur Loss: 0.05592321, Cur Avg Loss: 0.17229985, Log Avg loss: 0.22467899, Global Avg Loss: 0.64143021, Time: 0.0211 Steps: 101880, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001827, Sample Num: 29232, Cur Loss: 0.36171389, Cur Avg Loss: 0.17253443, Log Avg loss: 0.21515757, Global Avg Loss: 0.64138838, Time: 0.0211 Steps: 101890, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001837, Sample Num: 29392, Cur Loss: 0.11204592, Cur Avg Loss: 0.17237536, Log Avg loss: 0.14331313, Global Avg Loss: 0.64133950, Time: 0.0211 Steps: 101900, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001847, Sample Num: 29552, Cur Loss: 0.46288088, Cur Avg Loss: 0.17243768, Log Avg loss: 0.18388551, Global Avg Loss: 0.64129461, Time: 0.0212 Steps: 101910, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001857, Sample Num: 29712, Cur Loss: 0.31322598, Cur Avg Loss: 0.17249975, Log Avg loss: 0.18396479, Global Avg Loss: 0.64124974, Time: 0.0211 Steps: 101920, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001867, Sample Num: 29872, Cur Loss: 0.26590139, Cur Avg Loss: 0.17257859, Log Avg loss: 0.18721915, Global Avg Loss: 0.64120519, Time: 0.0211 Steps: 101930, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001877, Sample Num: 30032, Cur Loss: 0.16721600, Cur Avg Loss: 0.17251509, Log Avg loss: 0.16065891, Global Avg Loss: 0.64115805, Time: 0.0211 Steps: 101940, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001887, Sample Num: 30192, Cur Loss: 0.07573673, Cur Avg Loss: 0.17249994, Log Avg loss: 0.16965595, Global Avg Loss: 0.64111181, Time: 0.0211 Steps: 101950, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001897, Sample Num: 30352, Cur Loss: 0.06832211, Cur Avg Loss: 0.17250581, Log Avg loss: 0.17361503, Global Avg Loss: 0.64106596, Time: 0.0211 Steps: 101960, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001907, Sample Num: 30512, Cur Loss: 0.08245754, Cur Avg Loss: 0.17264781, Log Avg loss: 0.19958370, Global Avg Loss: 0.64102266, Time: 0.0211 Steps: 101970, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001917, Sample Num: 30672, Cur Loss: 0.15166357, Cur Avg Loss: 0.17248049, Log Avg loss: 0.14057384, Global Avg Loss: 0.64097359, Time: 0.0211 Steps: 101980, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001927, Sample Num: 30832, Cur Loss: 0.10921684, Cur Avg Loss: 0.17226036, Log Avg loss: 0.13006068, Global Avg Loss: 0.64092349, Time: 0.0211 Steps: 101990, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001937, Sample Num: 30992, Cur Loss: 0.17816889, Cur Avg Loss: 0.17209044, Log Avg loss: 0.13934712, Global Avg Loss: 0.64087432, Time: 0.0211 Steps: 102000, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001947, Sample Num: 31152, Cur Loss: 0.51587266, Cur Avg Loss: 0.17286254, Log Avg loss: 0.32241813, Global Avg Loss: 0.64084310, Time: 0.0211 Steps: 102010, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001957, Sample Num: 31312, Cur Loss: 0.13964261, Cur Avg Loss: 0.17317584, Log Avg loss: 0.23417629, Global Avg Loss: 0.64080324, Time: 0.0211 Steps: 102020, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001967, Sample Num: 31472, Cur Loss: 0.03402982, Cur Avg Loss: 0.17303304, Log Avg loss: 0.14508568, Global Avg Loss: 0.64075465, Time: 0.0211 Steps: 102030, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001977, Sample Num: 31632, Cur Loss: 0.05817668, Cur Avg Loss: 0.17274914, Log Avg loss: 0.11690662, Global Avg Loss: 0.64070332, Time: 0.0211 Steps: 102040, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001987, Sample Num: 31792, Cur Loss: 0.34296578, Cur Avg Loss: 0.17333844, Log Avg loss: 0.28984266, Global Avg Loss: 0.64066893, Time: 0.0211 Steps: 102050, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001997, Sample Num: 31952, Cur Loss: 0.13197729, Cur Avg Loss: 0.17345555, Log Avg loss: 0.19672465, Global Avg Loss: 0.64062544, Time: 0.0211 Steps: 102060, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002007, Sample Num: 32112, Cur Loss: 0.20902166, Cur Avg Loss: 0.17359118, Log Avg loss: 0.20067756, Global Avg Loss: 0.64058233, Time: 0.0211 Steps: 102070, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002017, Sample Num: 32272, Cur Loss: 0.05719408, Cur Avg Loss: 0.17376006, Log Avg loss: 0.20765512, Global Avg Loss: 0.64053992, Time: 0.0211 Steps: 102080, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002027, Sample Num: 32432, Cur Loss: 0.27045950, Cur Avg Loss: 0.17366293, Log Avg loss: 0.15407098, Global Avg Loss: 0.64049227, Time: 0.0212 Steps: 102090, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002037, Sample Num: 32592, Cur Loss: 0.14918351, Cur Avg Loss: 0.17383151, Log Avg loss: 0.20800292, Global Avg Loss: 0.64044991, Time: 0.0211 Steps: 102100, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002047, Sample Num: 32752, Cur Loss: 0.22702509, Cur Avg Loss: 0.17363317, Log Avg loss: 0.13323028, Global Avg Loss: 0.64040024, Time: 0.0211 Steps: 102110, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002057, Sample Num: 32912, Cur Loss: 0.24030629, Cur Avg Loss: 0.17363030, Log Avg loss: 0.17304445, Global Avg Loss: 0.64035447, Time: 0.0209 Steps: 102120, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002067, Sample Num: 33072, Cur Loss: 0.44178319, Cur Avg Loss: 0.17360794, Log Avg loss: 0.16900824, Global Avg Loss: 0.64030832, Time: 0.0208 Steps: 102130, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002077, Sample Num: 33232, Cur Loss: 0.62803751, Cur Avg Loss: 0.17364422, Log Avg loss: 0.18114325, Global Avg Loss: 0.64026337, Time: 0.0208 Steps: 102140, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002087, Sample Num: 33392, Cur Loss: 0.06848916, Cur Avg Loss: 0.17365897, Log Avg loss: 0.17672278, Global Avg Loss: 0.64021799, Time: 0.0208 Steps: 102150, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002097, Sample Num: 33552, Cur Loss: 0.48811105, Cur Avg Loss: 0.17410855, Log Avg loss: 0.26793578, Global Avg Loss: 0.64018155, Time: 0.0208 Steps: 102160, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002107, Sample Num: 33712, Cur Loss: 0.21047364, Cur Avg Loss: 0.17382646, Log Avg loss: 0.11467180, Global Avg Loss: 0.64013011, Time: 0.0208 Steps: 102170, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002117, Sample Num: 33872, Cur Loss: 0.08575222, Cur Avg Loss: 0.17377492, Log Avg loss: 0.16291520, Global Avg Loss: 0.64008341, Time: 0.0208 Steps: 102180, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002127, Sample Num: 34032, Cur Loss: 0.08094832, Cur Avg Loss: 0.17354573, Log Avg loss: 0.12502694, Global Avg Loss: 0.64003301, Time: 0.0208 Steps: 102190, Updated lr: 0.000004 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.778621, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.263284, "eval_total_loss": 185.088935, "eval_mae": 0.397072, "eval_mse": 0.263295, "eval_r2": 0.832632, "eval_sp_statistic": 0.903792, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.926552, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.309404, "test_total_loss": 155.320841, "test_mae": 0.3428, "test_mse": 0.309515, "test_r2": 0.800237, "test_sp_statistic": 0.876325, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.922693, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.640022090145999, "train_cur_epoch_loss": 369.29617456905544, "train_cur_epoch_avg_loss": 0.17345992229640933, "train_cur_epoch_time": 44.77862071990967, "train_cur_epoch_avg_time": 0.02103270113664146, "epoch": 48, "step": 102192} ################################################## Training, Epoch: 0049, Batch: 000008, Sample Num: 128, Cur Loss: 0.20174840, Cur Avg Loss: 0.15645224, Log Avg loss: 0.14160158, Global Avg Loss: 0.63998424, Time: 0.0211 Steps: 102200, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000018, Sample Num: 288, Cur Loss: 0.14789908, Cur Avg Loss: 0.14508795, Log Avg loss: 0.13599652, Global Avg Loss: 0.63993493, Time: 0.0209 Steps: 102210, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000028, Sample Num: 448, Cur Loss: 0.08273633, Cur Avg Loss: 0.14328225, Log Avg loss: 0.14003198, Global Avg Loss: 0.63988602, Time: 0.0209 Steps: 102220, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000038, Sample Num: 608, Cur Loss: 0.02570507, Cur Avg Loss: 0.14278209, Log Avg loss: 0.14138166, Global Avg Loss: 0.63983726, Time: 0.0209 Steps: 102230, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000048, Sample Num: 768, Cur Loss: 0.04033799, Cur Avg Loss: 0.14420656, Log Avg loss: 0.14961955, Global Avg Loss: 0.63978931, Time: 0.0208 Steps: 102240, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000058, Sample Num: 928, Cur Loss: 0.05592633, Cur Avg Loss: 0.15405246, Log Avg loss: 0.20131275, Global Avg Loss: 0.63974643, Time: 0.0209 Steps: 102250, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000068, Sample Num: 1088, Cur Loss: 0.17979230, Cur Avg Loss: 0.15720556, Log Avg loss: 0.17549356, Global Avg Loss: 0.63970103, Time: 0.0209 Steps: 102260, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000078, Sample Num: 1248, Cur Loss: 0.25083727, Cur Avg Loss: 0.15624912, Log Avg loss: 0.14974532, Global Avg Loss: 0.63965312, Time: 0.0209 Steps: 102270, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000088, Sample Num: 1408, Cur Loss: 0.14551757, Cur Avg Loss: 0.15640882, Log Avg loss: 0.15765450, Global Avg Loss: 0.63960600, Time: 0.0208 Steps: 102280, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000098, Sample Num: 1568, Cur Loss: 0.08012043, Cur Avg Loss: 0.15980211, Log Avg loss: 0.18966304, Global Avg Loss: 0.63956201, Time: 0.0209 Steps: 102290, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000108, Sample Num: 1728, Cur Loss: 0.17787731, Cur Avg Loss: 0.16292071, Log Avg loss: 0.19348296, Global Avg Loss: 0.63951841, Time: 0.0209 Steps: 102300, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000118, Sample Num: 1888, Cur Loss: 0.08092378, Cur Avg Loss: 0.16267546, Log Avg loss: 0.16002684, Global Avg Loss: 0.63947154, Time: 0.0209 Steps: 102310, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000128, Sample Num: 2048, Cur Loss: 0.11115980, Cur Avg Loss: 0.16284087, Log Avg loss: 0.16479268, Global Avg Loss: 0.63942515, Time: 0.0209 Steps: 102320, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000138, Sample Num: 2208, Cur Loss: 0.09527586, Cur Avg Loss: 0.16141364, Log Avg loss: 0.14314503, Global Avg Loss: 0.63937665, Time: 0.0209 Steps: 102330, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000148, Sample Num: 2368, Cur Loss: 0.25477976, Cur Avg Loss: 0.16262335, Log Avg loss: 0.17931738, Global Avg Loss: 0.63933170, Time: 0.0209 Steps: 102340, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000158, Sample Num: 2528, Cur Loss: 0.04061692, Cur Avg Loss: 0.16137830, Log Avg loss: 0.14295154, Global Avg Loss: 0.63928320, Time: 0.0208 Steps: 102350, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000168, Sample Num: 2688, Cur Loss: 0.05295258, Cur Avg Loss: 0.16573130, Log Avg loss: 0.23450874, Global Avg Loss: 0.63924365, Time: 0.0209 Steps: 102360, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000178, Sample Num: 2848, Cur Loss: 0.28067464, Cur Avg Loss: 0.16595175, Log Avg loss: 0.16965527, Global Avg Loss: 0.63919778, Time: 0.0209 Steps: 102370, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000188, Sample Num: 3008, Cur Loss: 0.38441515, Cur Avg Loss: 0.16899308, Log Avg loss: 0.22312885, Global Avg Loss: 0.63915714, Time: 0.0209 Steps: 102380, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000198, Sample Num: 3168, Cur Loss: 0.10271974, Cur Avg Loss: 0.16868121, Log Avg loss: 0.16281795, Global Avg Loss: 0.63911062, Time: 0.0209 Steps: 102390, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000208, Sample Num: 3328, Cur Loss: 0.23589212, Cur Avg Loss: 0.16832064, Log Avg loss: 0.16118148, Global Avg Loss: 0.63906395, Time: 0.0209 Steps: 102400, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000218, Sample Num: 3488, Cur Loss: 0.28329095, Cur Avg Loss: 0.16681630, Log Avg loss: 0.13552595, Global Avg Loss: 0.63901478, Time: 0.0209 Steps: 102410, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000228, Sample Num: 3648, Cur Loss: 0.20461847, Cur Avg Loss: 0.16563120, Log Avg loss: 0.13979605, Global Avg Loss: 0.63896604, Time: 0.0209 Steps: 102420, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000238, Sample Num: 3808, Cur Loss: 0.04110353, Cur Avg Loss: 0.16603093, Log Avg loss: 0.17514472, Global Avg Loss: 0.63892075, Time: 0.0209 Steps: 102430, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000248, Sample Num: 3968, Cur Loss: 0.17951845, Cur Avg Loss: 0.16531536, Log Avg loss: 0.14828477, Global Avg Loss: 0.63887286, Time: 0.0208 Steps: 102440, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000258, Sample Num: 4128, Cur Loss: 0.18694004, Cur Avg Loss: 0.16444368, Log Avg loss: 0.14282611, Global Avg Loss: 0.63882444, Time: 0.0246 Steps: 102450, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000268, Sample Num: 4288, Cur Loss: 0.52466649, Cur Avg Loss: 0.16520432, Log Avg loss: 0.18482876, Global Avg Loss: 0.63878013, Time: 0.0209 Steps: 102460, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000278, Sample Num: 4448, Cur Loss: 0.18299097, Cur Avg Loss: 0.16503160, Log Avg loss: 0.16040262, Global Avg Loss: 0.63873345, Time: 0.0209 Steps: 102470, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000288, Sample Num: 4608, Cur Loss: 0.20747094, Cur Avg Loss: 0.16444585, Log Avg loss: 0.14816198, Global Avg Loss: 0.63868558, Time: 0.0209 Steps: 102480, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000298, Sample Num: 4768, Cur Loss: 0.22462742, Cur Avg Loss: 0.16367629, Log Avg loss: 0.14151301, Global Avg Loss: 0.63863707, Time: 0.0209 Steps: 102490, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000308, Sample Num: 4928, Cur Loss: 0.26221675, Cur Avg Loss: 0.16391504, Log Avg loss: 0.17102981, Global Avg Loss: 0.63859145, Time: 0.0209 Steps: 102500, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000318, Sample Num: 5088, Cur Loss: 0.12052423, Cur Avg Loss: 0.16282060, Log Avg loss: 0.12911185, Global Avg Loss: 0.63854175, Time: 0.0208 Steps: 102510, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000328, Sample Num: 5248, Cur Loss: 0.44694984, Cur Avg Loss: 0.16474028, Log Avg loss: 0.22578618, Global Avg Loss: 0.63850149, Time: 0.0208 Steps: 102520, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000338, Sample Num: 5408, Cur Loss: 0.11229125, Cur Avg Loss: 0.16456539, Log Avg loss: 0.15882910, Global Avg Loss: 0.63845470, Time: 0.0209 Steps: 102530, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000348, Sample Num: 5568, Cur Loss: 0.11782235, Cur Avg Loss: 0.16495595, Log Avg loss: 0.17815661, Global Avg Loss: 0.63840981, Time: 0.0209 Steps: 102540, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000358, Sample Num: 5728, Cur Loss: 0.14930737, Cur Avg Loss: 0.16495063, Log Avg loss: 0.16476550, Global Avg Loss: 0.63836363, Time: 0.0209 Steps: 102550, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000368, Sample Num: 5888, Cur Loss: 0.02013805, Cur Avg Loss: 0.16449261, Log Avg loss: 0.14809569, Global Avg Loss: 0.63831582, Time: 0.0209 Steps: 102560, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000378, Sample Num: 6048, Cur Loss: 0.08609931, Cur Avg Loss: 0.16346936, Log Avg loss: 0.12581360, Global Avg Loss: 0.63826586, Time: 0.0208 Steps: 102570, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000388, Sample Num: 6208, Cur Loss: 0.20276730, Cur Avg Loss: 0.16397790, Log Avg loss: 0.18320075, Global Avg Loss: 0.63822149, Time: 0.0208 Steps: 102580, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000398, Sample Num: 6368, Cur Loss: 0.18519048, Cur Avg Loss: 0.16336109, Log Avg loss: 0.13942905, Global Avg Loss: 0.63817287, Time: 0.0209 Steps: 102590, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000408, Sample Num: 6528, Cur Loss: 0.31181574, Cur Avg Loss: 0.16346126, Log Avg loss: 0.16744786, Global Avg Loss: 0.63812699, Time: 0.0208 Steps: 102600, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000418, Sample Num: 6688, Cur Loss: 0.06879838, Cur Avg Loss: 0.16345486, Log Avg loss: 0.16319361, Global Avg Loss: 0.63808071, Time: 0.0210 Steps: 102610, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000428, Sample Num: 6848, Cur Loss: 0.28389537, Cur Avg Loss: 0.16392869, Log Avg loss: 0.18373493, Global Avg Loss: 0.63803643, Time: 0.0208 Steps: 102620, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000438, Sample Num: 7008, Cur Loss: 0.09201393, Cur Avg Loss: 0.16509843, Log Avg loss: 0.21516336, Global Avg Loss: 0.63799523, Time: 0.0208 Steps: 102630, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000448, Sample Num: 7168, Cur Loss: 0.17655775, Cur Avg Loss: 0.16506116, Log Avg loss: 0.16342877, Global Avg Loss: 0.63794899, Time: 0.0210 Steps: 102640, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000458, Sample Num: 7328, Cur Loss: 0.11594482, Cur Avg Loss: 0.16378115, Log Avg loss: 0.10643670, Global Avg Loss: 0.63789722, Time: 0.0209 Steps: 102650, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000468, Sample Num: 7488, Cur Loss: 0.19704512, Cur Avg Loss: 0.16578031, Log Avg loss: 0.25734184, Global Avg Loss: 0.63786015, Time: 0.0209 Steps: 102660, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000478, Sample Num: 7648, Cur Loss: 0.20991474, Cur Avg Loss: 0.16643807, Log Avg loss: 0.19722107, Global Avg Loss: 0.63781723, Time: 0.0209 Steps: 102670, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000488, Sample Num: 7808, Cur Loss: 0.09735214, Cur Avg Loss: 0.16636498, Log Avg loss: 0.16287122, Global Avg Loss: 0.63777097, Time: 0.0209 Steps: 102680, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000498, Sample Num: 7968, Cur Loss: 0.21513528, Cur Avg Loss: 0.16586569, Log Avg loss: 0.14150058, Global Avg Loss: 0.63772265, Time: 0.0209 Steps: 102690, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000508, Sample Num: 8128, Cur Loss: 0.42516062, Cur Avg Loss: 0.16532916, Log Avg loss: 0.13861016, Global Avg Loss: 0.63767405, Time: 0.0210 Steps: 102700, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000518, Sample Num: 8288, Cur Loss: 0.08112238, Cur Avg Loss: 0.16640271, Log Avg loss: 0.22093876, Global Avg Loss: 0.63763347, Time: 0.0212 Steps: 102710, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000528, Sample Num: 8448, Cur Loss: 0.03191351, Cur Avg Loss: 0.16733339, Log Avg loss: 0.21554242, Global Avg Loss: 0.63759238, Time: 0.0209 Steps: 102720, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000538, Sample Num: 8608, Cur Loss: 0.16512290, Cur Avg Loss: 0.16769452, Log Avg loss: 0.18676234, Global Avg Loss: 0.63754850, Time: 0.0209 Steps: 102730, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000548, Sample Num: 8768, Cur Loss: 0.12999499, Cur Avg Loss: 0.16688372, Log Avg loss: 0.12326278, Global Avg Loss: 0.63749844, Time: 0.0209 Steps: 102740, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000558, Sample Num: 8928, Cur Loss: 0.02608052, Cur Avg Loss: 0.16588335, Log Avg loss: 0.11106280, Global Avg Loss: 0.63744721, Time: 0.0209 Steps: 102750, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000568, Sample Num: 9088, Cur Loss: 0.07102723, Cur Avg Loss: 0.16605379, Log Avg loss: 0.17556444, Global Avg Loss: 0.63740226, Time: 0.0209 Steps: 102760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000578, Sample Num: 9248, Cur Loss: 0.15973273, Cur Avg Loss: 0.16494639, Log Avg loss: 0.10204614, Global Avg Loss: 0.63735016, Time: 0.0209 Steps: 102770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000588, Sample Num: 9408, Cur Loss: 0.45422244, Cur Avg Loss: 0.16587145, Log Avg loss: 0.21933977, Global Avg Loss: 0.63730949, Time: 0.0209 Steps: 102780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000598, Sample Num: 9568, Cur Loss: 0.10550539, Cur Avg Loss: 0.16558977, Log Avg loss: 0.14902731, Global Avg Loss: 0.63726199, Time: 0.0209 Steps: 102790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000608, Sample Num: 9728, Cur Loss: 0.11341636, Cur Avg Loss: 0.16482172, Log Avg loss: 0.11889195, Global Avg Loss: 0.63721157, Time: 0.0209 Steps: 102800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000618, Sample Num: 9888, Cur Loss: 0.09856427, Cur Avg Loss: 0.16419024, Log Avg loss: 0.12579670, Global Avg Loss: 0.63716182, Time: 0.0209 Steps: 102810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000628, Sample Num: 10048, Cur Loss: 0.07414822, Cur Avg Loss: 0.16370603, Log Avg loss: 0.13378149, Global Avg Loss: 0.63711287, Time: 0.0209 Steps: 102820, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000638, Sample Num: 10208, Cur Loss: 0.47000325, Cur Avg Loss: 0.16416490, Log Avg loss: 0.19298236, Global Avg Loss: 0.63706967, Time: 0.0209 Steps: 102830, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000648, Sample Num: 10368, Cur Loss: 0.17063396, Cur Avg Loss: 0.16560840, Log Avg loss: 0.25770358, Global Avg Loss: 0.63703279, Time: 0.0210 Steps: 102840, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000658, Sample Num: 10528, Cur Loss: 0.17682992, Cur Avg Loss: 0.16471006, Log Avg loss: 0.10649751, Global Avg Loss: 0.63698120, Time: 0.0209 Steps: 102850, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000668, Sample Num: 10688, Cur Loss: 0.05909068, Cur Avg Loss: 0.16410321, Log Avg loss: 0.12417262, Global Avg Loss: 0.63693135, Time: 0.0209 Steps: 102860, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000678, Sample Num: 10848, Cur Loss: 0.06756286, Cur Avg Loss: 0.16385877, Log Avg loss: 0.14753019, Global Avg Loss: 0.63688377, Time: 0.0209 Steps: 102870, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000688, Sample Num: 11008, Cur Loss: 0.26084247, Cur Avg Loss: 0.16423462, Log Avg loss: 0.18971725, Global Avg Loss: 0.63684031, Time: 0.0209 Steps: 102880, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000698, Sample Num: 11168, Cur Loss: 0.53721952, Cur Avg Loss: 0.16442602, Log Avg loss: 0.17759402, Global Avg Loss: 0.63679567, Time: 0.0209 Steps: 102890, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000708, Sample Num: 11328, Cur Loss: 0.09426065, Cur Avg Loss: 0.16512309, Log Avg loss: 0.21377890, Global Avg Loss: 0.63675456, Time: 0.0209 Steps: 102900, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000718, Sample Num: 11488, Cur Loss: 0.77891886, Cur Avg Loss: 0.16598702, Log Avg loss: 0.22715316, Global Avg Loss: 0.63671476, Time: 0.0209 Steps: 102910, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000728, Sample Num: 11648, Cur Loss: 0.05040237, Cur Avg Loss: 0.16534940, Log Avg loss: 0.11956789, Global Avg Loss: 0.63666451, Time: 0.0209 Steps: 102920, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000738, Sample Num: 11808, Cur Loss: 0.05328035, Cur Avg Loss: 0.16489914, Log Avg loss: 0.13212037, Global Avg Loss: 0.63661550, Time: 0.0209 Steps: 102930, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000748, Sample Num: 11968, Cur Loss: 0.08363404, Cur Avg Loss: 0.16472230, Log Avg loss: 0.15167155, Global Avg Loss: 0.63656839, Time: 0.0209 Steps: 102940, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000758, Sample Num: 12128, Cur Loss: 0.16925953, Cur Avg Loss: 0.16473361, Log Avg loss: 0.16557957, Global Avg Loss: 0.63652264, Time: 0.0208 Steps: 102950, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000768, Sample Num: 12288, Cur Loss: 0.09058247, Cur Avg Loss: 0.16470517, Log Avg loss: 0.16254979, Global Avg Loss: 0.63647660, Time: 0.0254 Steps: 102960, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000778, Sample Num: 12448, Cur Loss: 0.29587239, Cur Avg Loss: 0.16567582, Log Avg loss: 0.24022190, Global Avg Loss: 0.63643812, Time: 0.0209 Steps: 102970, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000788, Sample Num: 12608, Cur Loss: 0.29705977, Cur Avg Loss: 0.16576321, Log Avg loss: 0.17256188, Global Avg Loss: 0.63639307, Time: 0.0209 Steps: 102980, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000798, Sample Num: 12768, Cur Loss: 0.07879565, Cur Avg Loss: 0.16579231, Log Avg loss: 0.16808529, Global Avg Loss: 0.63634760, Time: 0.0209 Steps: 102990, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000808, Sample Num: 12928, Cur Loss: 0.12197989, Cur Avg Loss: 0.16562481, Log Avg loss: 0.15225817, Global Avg Loss: 0.63630060, Time: 0.0209 Steps: 103000, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000818, Sample Num: 13088, Cur Loss: 0.12583327, Cur Avg Loss: 0.16567322, Log Avg loss: 0.16958491, Global Avg Loss: 0.63625530, Time: 0.0209 Steps: 103010, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000828, Sample Num: 13248, Cur Loss: 0.20522857, Cur Avg Loss: 0.16649506, Log Avg loss: 0.23372194, Global Avg Loss: 0.63621622, Time: 0.0209 Steps: 103020, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000838, Sample Num: 13408, Cur Loss: 0.20342015, Cur Avg Loss: 0.16686256, Log Avg loss: 0.19729116, Global Avg Loss: 0.63617362, Time: 0.0209 Steps: 103030, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000848, Sample Num: 13568, Cur Loss: 0.14172570, Cur Avg Loss: 0.16695141, Log Avg loss: 0.17439701, Global Avg Loss: 0.63612881, Time: 0.0209 Steps: 103040, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000858, Sample Num: 13728, Cur Loss: 0.27067670, Cur Avg Loss: 0.16717356, Log Avg loss: 0.18601167, Global Avg Loss: 0.63608513, Time: 0.0209 Steps: 103050, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000868, Sample Num: 13888, Cur Loss: 0.08296204, Cur Avg Loss: 0.16758783, Log Avg loss: 0.20313209, Global Avg Loss: 0.63604312, Time: 0.0209 Steps: 103060, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000878, Sample Num: 14048, Cur Loss: 0.02973625, Cur Avg Loss: 0.16732074, Log Avg loss: 0.14413758, Global Avg Loss: 0.63599539, Time: 0.0209 Steps: 103070, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000888, Sample Num: 14208, Cur Loss: 0.27175540, Cur Avg Loss: 0.16784977, Log Avg loss: 0.21429874, Global Avg Loss: 0.63595448, Time: 0.0209 Steps: 103080, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000898, Sample Num: 14368, Cur Loss: 0.53668177, Cur Avg Loss: 0.16827090, Log Avg loss: 0.20566716, Global Avg Loss: 0.63591274, Time: 0.0208 Steps: 103090, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000908, Sample Num: 14528, Cur Loss: 0.17770089, Cur Avg Loss: 0.16846454, Log Avg loss: 0.18585358, Global Avg Loss: 0.63586909, Time: 0.0209 Steps: 103100, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000918, Sample Num: 14688, Cur Loss: 0.04248293, Cur Avg Loss: 0.16807675, Log Avg loss: 0.13286567, Global Avg Loss: 0.63582031, Time: 0.0209 Steps: 103110, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000928, Sample Num: 14848, Cur Loss: 0.08329941, Cur Avg Loss: 0.16748042, Log Avg loss: 0.11273692, Global Avg Loss: 0.63576958, Time: 0.0209 Steps: 103120, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000938, Sample Num: 15008, Cur Loss: 0.09907445, Cur Avg Loss: 0.16730432, Log Avg loss: 0.15096277, Global Avg Loss: 0.63572257, Time: 0.0209 Steps: 103130, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000948, Sample Num: 15168, Cur Loss: 0.31612614, Cur Avg Loss: 0.16762781, Log Avg loss: 0.19797104, Global Avg Loss: 0.63568013, Time: 0.0209 Steps: 103140, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000958, Sample Num: 15328, Cur Loss: 0.14909136, Cur Avg Loss: 0.16746931, Log Avg loss: 0.15244350, Global Avg Loss: 0.63563328, Time: 0.0209 Steps: 103150, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000968, Sample Num: 15488, Cur Loss: 0.13005532, Cur Avg Loss: 0.16723375, Log Avg loss: 0.14466703, Global Avg Loss: 0.63558569, Time: 0.0209 Steps: 103160, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000978, Sample Num: 15648, Cur Loss: 0.27437875, Cur Avg Loss: 0.16756337, Log Avg loss: 0.19947009, Global Avg Loss: 0.63554342, Time: 0.0208 Steps: 103170, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000988, Sample Num: 15808, Cur Loss: 0.11896069, Cur Avg Loss: 0.16762708, Log Avg loss: 0.17385848, Global Avg Loss: 0.63549867, Time: 0.0209 Steps: 103180, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000998, Sample Num: 15968, Cur Loss: 0.13774207, Cur Avg Loss: 0.16748982, Log Avg loss: 0.15392792, Global Avg Loss: 0.63545200, Time: 0.0208 Steps: 103190, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001008, Sample Num: 16128, Cur Loss: 0.09397297, Cur Avg Loss: 0.16694866, Log Avg loss: 0.11294162, Global Avg Loss: 0.63540137, Time: 0.0209 Steps: 103200, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001018, Sample Num: 16288, Cur Loss: 0.06403475, Cur Avg Loss: 0.16674689, Log Avg loss: 0.14640814, Global Avg Loss: 0.63535399, Time: 0.0209 Steps: 103210, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001028, Sample Num: 16448, Cur Loss: 0.41946146, Cur Avg Loss: 0.16688760, Log Avg loss: 0.18121144, Global Avg Loss: 0.63531000, Time: 0.0247 Steps: 103220, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001038, Sample Num: 16608, Cur Loss: 0.11256542, Cur Avg Loss: 0.16656337, Log Avg loss: 0.13323315, Global Avg Loss: 0.63526136, Time: 0.0210 Steps: 103230, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001048, Sample Num: 16768, Cur Loss: 0.07615063, Cur Avg Loss: 0.16640781, Log Avg loss: 0.15026069, Global Avg Loss: 0.63521438, Time: 0.0209 Steps: 103240, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001058, Sample Num: 16928, Cur Loss: 0.22778000, Cur Avg Loss: 0.16627194, Log Avg loss: 0.15203213, Global Avg Loss: 0.63516758, Time: 0.0209 Steps: 103250, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001068, Sample Num: 17088, Cur Loss: 0.06569082, Cur Avg Loss: 0.16644974, Log Avg loss: 0.18526145, Global Avg Loss: 0.63512401, Time: 0.0209 Steps: 103260, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001078, Sample Num: 17248, Cur Loss: 0.31755728, Cur Avg Loss: 0.16617983, Log Avg loss: 0.13735284, Global Avg Loss: 0.63507581, Time: 0.0210 Steps: 103270, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001088, Sample Num: 17408, Cur Loss: 0.12265792, Cur Avg Loss: 0.16578700, Log Avg loss: 0.12343998, Global Avg Loss: 0.63502628, Time: 0.0209 Steps: 103280, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001098, Sample Num: 17568, Cur Loss: 0.06604965, Cur Avg Loss: 0.16702950, Log Avg loss: 0.30221338, Global Avg Loss: 0.63499405, Time: 0.0209 Steps: 103290, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001108, Sample Num: 17728, Cur Loss: 0.21643119, Cur Avg Loss: 0.16673303, Log Avg loss: 0.13418089, Global Avg Loss: 0.63494557, Time: 0.0209 Steps: 103300, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001118, Sample Num: 17888, Cur Loss: 0.09461715, Cur Avg Loss: 0.16665660, Log Avg loss: 0.15818893, Global Avg Loss: 0.63489942, Time: 0.0209 Steps: 103310, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001128, Sample Num: 18048, Cur Loss: 0.21853387, Cur Avg Loss: 0.16635931, Log Avg loss: 0.13312196, Global Avg Loss: 0.63485086, Time: 0.0209 Steps: 103320, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001138, Sample Num: 18208, Cur Loss: 0.16211349, Cur Avg Loss: 0.16615443, Log Avg loss: 0.14304383, Global Avg Loss: 0.63480326, Time: 0.0209 Steps: 103330, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001148, Sample Num: 18368, Cur Loss: 0.08591571, Cur Avg Loss: 0.16621819, Log Avg loss: 0.17347374, Global Avg Loss: 0.63475862, Time: 0.0209 Steps: 103340, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001158, Sample Num: 18528, Cur Loss: 0.11757797, Cur Avg Loss: 0.16620433, Log Avg loss: 0.16461342, Global Avg Loss: 0.63471313, Time: 0.0210 Steps: 103350, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001168, Sample Num: 18688, Cur Loss: 0.21030092, Cur Avg Loss: 0.16598751, Log Avg loss: 0.14088012, Global Avg Loss: 0.63466535, Time: 0.0209 Steps: 103360, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001178, Sample Num: 18848, Cur Loss: 0.11950927, Cur Avg Loss: 0.16560940, Log Avg loss: 0.12144575, Global Avg Loss: 0.63461570, Time: 0.0209 Steps: 103370, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001188, Sample Num: 19008, Cur Loss: 0.14521171, Cur Avg Loss: 0.16546103, Log Avg loss: 0.14798302, Global Avg Loss: 0.63456863, Time: 0.0209 Steps: 103380, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001198, Sample Num: 19168, Cur Loss: 0.18487526, Cur Avg Loss: 0.16522200, Log Avg loss: 0.13682557, Global Avg Loss: 0.63452049, Time: 0.0209 Steps: 103390, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001208, Sample Num: 19328, Cur Loss: 0.33066908, Cur Avg Loss: 0.16601951, Log Avg loss: 0.26156121, Global Avg Loss: 0.63448442, Time: 0.0209 Steps: 103400, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001218, Sample Num: 19488, Cur Loss: 0.13480785, Cur Avg Loss: 0.16587363, Log Avg loss: 0.14825158, Global Avg Loss: 0.63443740, Time: 0.0209 Steps: 103410, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001228, Sample Num: 19648, Cur Loss: 0.30650479, Cur Avg Loss: 0.16609097, Log Avg loss: 0.19256276, Global Avg Loss: 0.63439467, Time: 0.0209 Steps: 103420, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001238, Sample Num: 19808, Cur Loss: 0.13905616, Cur Avg Loss: 0.16666719, Log Avg loss: 0.23742698, Global Avg Loss: 0.63435629, Time: 0.0208 Steps: 103430, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001248, Sample Num: 19968, Cur Loss: 0.19356500, Cur Avg Loss: 0.16681315, Log Avg loss: 0.18488246, Global Avg Loss: 0.63431284, Time: 0.0209 Steps: 103440, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001258, Sample Num: 20128, Cur Loss: 0.21620914, Cur Avg Loss: 0.16745594, Log Avg loss: 0.24767629, Global Avg Loss: 0.63427547, Time: 0.0209 Steps: 103450, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001268, Sample Num: 20288, Cur Loss: 0.11198674, Cur Avg Loss: 0.16766971, Log Avg loss: 0.19456204, Global Avg Loss: 0.63423297, Time: 0.0209 Steps: 103460, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001278, Sample Num: 20448, Cur Loss: 0.93548298, Cur Avg Loss: 0.16798987, Log Avg loss: 0.20858623, Global Avg Loss: 0.63419183, Time: 0.0209 Steps: 103470, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001288, Sample Num: 20608, Cur Loss: 0.13810256, Cur Avg Loss: 0.16814856, Log Avg loss: 0.18842905, Global Avg Loss: 0.63414875, Time: 0.0219 Steps: 103480, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001298, Sample Num: 20768, Cur Loss: 0.09437854, Cur Avg Loss: 0.16819422, Log Avg loss: 0.17407587, Global Avg Loss: 0.63410430, Time: 0.0220 Steps: 103490, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001308, Sample Num: 20928, Cur Loss: 0.11379056, Cur Avg Loss: 0.16793865, Log Avg loss: 0.13476565, Global Avg Loss: 0.63405605, Time: 0.0219 Steps: 103500, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001318, Sample Num: 21088, Cur Loss: 0.20810802, Cur Avg Loss: 0.16740771, Log Avg loss: 0.09796052, Global Avg Loss: 0.63400426, Time: 0.0220 Steps: 103510, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001328, Sample Num: 21248, Cur Loss: 0.11011158, Cur Avg Loss: 0.16745574, Log Avg loss: 0.17378573, Global Avg Loss: 0.63395980, Time: 0.0219 Steps: 103520, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001338, Sample Num: 21408, Cur Loss: 0.10870995, Cur Avg Loss: 0.16742383, Log Avg loss: 0.16318582, Global Avg Loss: 0.63391433, Time: 0.0219 Steps: 103530, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001348, Sample Num: 21568, Cur Loss: 0.10676550, Cur Avg Loss: 0.16747819, Log Avg loss: 0.17475266, Global Avg Loss: 0.63386998, Time: 0.0219 Steps: 103540, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001358, Sample Num: 21728, Cur Loss: 0.08614838, Cur Avg Loss: 0.16722915, Log Avg loss: 0.13365722, Global Avg Loss: 0.63382168, Time: 0.0220 Steps: 103550, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001368, Sample Num: 21888, Cur Loss: 0.15220132, Cur Avg Loss: 0.16691653, Log Avg loss: 0.12446327, Global Avg Loss: 0.63377249, Time: 0.0219 Steps: 103560, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001378, Sample Num: 22048, Cur Loss: 0.10845807, Cur Avg Loss: 0.16744843, Log Avg loss: 0.24021195, Global Avg Loss: 0.63373449, Time: 0.0219 Steps: 103570, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001388, Sample Num: 22208, Cur Loss: 0.09268600, Cur Avg Loss: 0.16716387, Log Avg loss: 0.12795151, Global Avg Loss: 0.63368566, Time: 0.0219 Steps: 103580, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001398, Sample Num: 22368, Cur Loss: 0.25496012, Cur Avg Loss: 0.16729832, Log Avg loss: 0.18596013, Global Avg Loss: 0.63364244, Time: 0.0219 Steps: 103590, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001408, Sample Num: 22528, Cur Loss: 0.03441384, Cur Avg Loss: 0.16770986, Log Avg loss: 0.22524366, Global Avg Loss: 0.63360302, Time: 0.0219 Steps: 103600, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001418, Sample Num: 22688, Cur Loss: 0.27422327, Cur Avg Loss: 0.16800961, Log Avg loss: 0.21021370, Global Avg Loss: 0.63356216, Time: 0.0219 Steps: 103610, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001428, Sample Num: 22848, Cur Loss: 0.67981958, Cur Avg Loss: 0.16864981, Log Avg loss: 0.25943094, Global Avg Loss: 0.63352605, Time: 0.0221 Steps: 103620, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001438, Sample Num: 23008, Cur Loss: 0.14738995, Cur Avg Loss: 0.16893247, Log Avg loss: 0.20929562, Global Avg Loss: 0.63348511, Time: 0.0219 Steps: 103630, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001448, Sample Num: 23168, Cur Loss: 0.12587002, Cur Avg Loss: 0.16894360, Log Avg loss: 0.17054391, Global Avg Loss: 0.63344045, Time: 0.0220 Steps: 103640, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001458, Sample Num: 23328, Cur Loss: 0.04041971, Cur Avg Loss: 0.16848942, Log Avg loss: 0.10272496, Global Avg Loss: 0.63338924, Time: 0.0219 Steps: 103650, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001468, Sample Num: 23488, Cur Loss: 0.07047801, Cur Avg Loss: 0.16845867, Log Avg loss: 0.16397590, Global Avg Loss: 0.63334396, Time: 0.0219 Steps: 103660, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001478, Sample Num: 23648, Cur Loss: 0.06277907, Cur Avg Loss: 0.16900531, Log Avg loss: 0.24925067, Global Avg Loss: 0.63330691, Time: 0.0219 Steps: 103670, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001488, Sample Num: 23808, Cur Loss: 0.11799560, Cur Avg Loss: 0.16875555, Log Avg loss: 0.13184157, Global Avg Loss: 0.63325854, Time: 0.0219 Steps: 103680, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001498, Sample Num: 23968, Cur Loss: 0.52515286, Cur Avg Loss: 0.16906805, Log Avg loss: 0.21556806, Global Avg Loss: 0.63321826, Time: 0.0219 Steps: 103690, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001508, Sample Num: 24128, Cur Loss: 0.05193102, Cur Avg Loss: 0.16925249, Log Avg loss: 0.19688108, Global Avg Loss: 0.63317618, Time: 0.0219 Steps: 103700, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001518, Sample Num: 24288, Cur Loss: 0.20229910, Cur Avg Loss: 0.16937913, Log Avg loss: 0.18847740, Global Avg Loss: 0.63313330, Time: 0.0219 Steps: 103710, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001528, Sample Num: 24448, Cur Loss: 0.05521421, Cur Avg Loss: 0.16919984, Log Avg loss: 0.14198378, Global Avg Loss: 0.63308595, Time: 0.0219 Steps: 103720, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001538, Sample Num: 24608, Cur Loss: 0.16269912, Cur Avg Loss: 0.16917505, Log Avg loss: 0.16538686, Global Avg Loss: 0.63304086, Time: 0.0246 Steps: 103730, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001548, Sample Num: 24768, Cur Loss: 0.23505229, Cur Avg Loss: 0.16949157, Log Avg loss: 0.21817178, Global Avg Loss: 0.63300087, Time: 0.0209 Steps: 103740, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001558, Sample Num: 24928, Cur Loss: 0.07749839, Cur Avg Loss: 0.16918520, Log Avg loss: 0.12175905, Global Avg Loss: 0.63295159, Time: 0.0208 Steps: 103750, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001568, Sample Num: 25088, Cur Loss: 0.18520170, Cur Avg Loss: 0.16916501, Log Avg loss: 0.16601945, Global Avg Loss: 0.63290659, Time: 0.0208 Steps: 103760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001578, Sample Num: 25248, Cur Loss: 0.11849764, Cur Avg Loss: 0.16886905, Log Avg loss: 0.12246356, Global Avg Loss: 0.63285740, Time: 0.0209 Steps: 103770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001588, Sample Num: 25408, Cur Loss: 0.08444236, Cur Avg Loss: 0.16907354, Log Avg loss: 0.20134100, Global Avg Loss: 0.63281582, Time: 0.0209 Steps: 103780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001598, Sample Num: 25568, Cur Loss: 0.13068499, Cur Avg Loss: 0.16871488, Log Avg loss: 0.11176046, Global Avg Loss: 0.63276562, Time: 0.0209 Steps: 103790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001608, Sample Num: 25728, Cur Loss: 0.15714312, Cur Avg Loss: 0.16878165, Log Avg loss: 0.17945112, Global Avg Loss: 0.63272195, Time: 0.0209 Steps: 103800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001618, Sample Num: 25888, Cur Loss: 0.20979002, Cur Avg Loss: 0.16871392, Log Avg loss: 0.15782246, Global Avg Loss: 0.63267620, Time: 0.0209 Steps: 103810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001628, Sample Num: 26048, Cur Loss: 0.05189411, Cur Avg Loss: 0.16898824, Log Avg loss: 0.21337403, Global Avg Loss: 0.63263581, Time: 0.0208 Steps: 103820, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001638, Sample Num: 26208, Cur Loss: 0.32049015, Cur Avg Loss: 0.16899910, Log Avg loss: 0.17076616, Global Avg Loss: 0.63259133, Time: 0.0208 Steps: 103830, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001648, Sample Num: 26368, Cur Loss: 0.07793267, Cur Avg Loss: 0.16879202, Log Avg loss: 0.13487351, Global Avg Loss: 0.63254340, Time: 0.0208 Steps: 103840, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001658, Sample Num: 26528, Cur Loss: 0.23790830, Cur Avg Loss: 0.16927035, Log Avg loss: 0.24809851, Global Avg Loss: 0.63250638, Time: 0.0208 Steps: 103850, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001668, Sample Num: 26688, Cur Loss: 0.13254800, Cur Avg Loss: 0.16915206, Log Avg loss: 0.14953941, Global Avg Loss: 0.63245988, Time: 0.0209 Steps: 103860, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001678, Sample Num: 26848, Cur Loss: 0.15350255, Cur Avg Loss: 0.16911668, Log Avg loss: 0.16321579, Global Avg Loss: 0.63241470, Time: 0.0208 Steps: 103870, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001688, Sample Num: 27008, Cur Loss: 0.21670181, Cur Avg Loss: 0.16889057, Log Avg loss: 0.13094863, Global Avg Loss: 0.63236643, Time: 0.0208 Steps: 103880, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001698, Sample Num: 27168, Cur Loss: 0.05550472, Cur Avg Loss: 0.16877571, Log Avg loss: 0.14938720, Global Avg Loss: 0.63231994, Time: 0.0208 Steps: 103890, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001708, Sample Num: 27328, Cur Loss: 0.14538194, Cur Avg Loss: 0.16885759, Log Avg loss: 0.18276086, Global Avg Loss: 0.63227667, Time: 0.0208 Steps: 103900, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001718, Sample Num: 27488, Cur Loss: 0.23584929, Cur Avg Loss: 0.16885813, Log Avg loss: 0.16895169, Global Avg Loss: 0.63223208, Time: 0.0208 Steps: 103910, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001728, Sample Num: 27648, Cur Loss: 0.31318820, Cur Avg Loss: 0.16890299, Log Avg loss: 0.17660880, Global Avg Loss: 0.63218824, Time: 0.0209 Steps: 103920, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001738, Sample Num: 27808, Cur Loss: 0.21491478, Cur Avg Loss: 0.16909423, Log Avg loss: 0.20214118, Global Avg Loss: 0.63214686, Time: 0.0210 Steps: 103930, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001748, Sample Num: 27968, Cur Loss: 0.15149598, Cur Avg Loss: 0.16904659, Log Avg loss: 0.16076641, Global Avg Loss: 0.63210151, Time: 0.0208 Steps: 103940, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001758, Sample Num: 28128, Cur Loss: 0.14777011, Cur Avg Loss: 0.16896889, Log Avg loss: 0.15538619, Global Avg Loss: 0.63205565, Time: 0.0208 Steps: 103950, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001768, Sample Num: 28288, Cur Loss: 0.14866903, Cur Avg Loss: 0.16932188, Log Avg loss: 0.23137833, Global Avg Loss: 0.63201711, Time: 0.0208 Steps: 103960, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001778, Sample Num: 28448, Cur Loss: 0.04924549, Cur Avg Loss: 0.16916412, Log Avg loss: 0.14127150, Global Avg Loss: 0.63196991, Time: 0.0209 Steps: 103970, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001788, Sample Num: 28608, Cur Loss: 0.36628690, Cur Avg Loss: 0.16912492, Log Avg loss: 0.16215582, Global Avg Loss: 0.63192472, Time: 0.0209 Steps: 103980, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001798, Sample Num: 28768, Cur Loss: 0.07601608, Cur Avg Loss: 0.16930584, Log Avg loss: 0.20165449, Global Avg Loss: 0.63188335, Time: 0.0246 Steps: 103990, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001808, Sample Num: 28928, Cur Loss: 0.09254002, Cur Avg Loss: 0.16938846, Log Avg loss: 0.18424247, Global Avg Loss: 0.63184031, Time: 0.0209 Steps: 104000, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001818, Sample Num: 29088, Cur Loss: 0.10258697, Cur Avg Loss: 0.16942673, Log Avg loss: 0.17634595, Global Avg Loss: 0.63179651, Time: 0.0208 Steps: 104010, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001828, Sample Num: 29248, Cur Loss: 0.20811936, Cur Avg Loss: 0.16964970, Log Avg loss: 0.21018717, Global Avg Loss: 0.63175598, Time: 0.0208 Steps: 104020, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001838, Sample Num: 29408, Cur Loss: 0.34538999, Cur Avg Loss: 0.17009635, Log Avg loss: 0.25174342, Global Avg Loss: 0.63171945, Time: 0.0208 Steps: 104030, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001848, Sample Num: 29568, Cur Loss: 0.18926316, Cur Avg Loss: 0.16991699, Log Avg loss: 0.13695129, Global Avg Loss: 0.63167190, Time: 0.0209 Steps: 104040, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001858, Sample Num: 29728, Cur Loss: 0.13623178, Cur Avg Loss: 0.17000919, Log Avg loss: 0.18704637, Global Avg Loss: 0.63162916, Time: 0.0208 Steps: 104050, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001868, Sample Num: 29888, Cur Loss: 0.14204471, Cur Avg Loss: 0.17011426, Log Avg loss: 0.18963606, Global Avg Loss: 0.63158669, Time: 0.0209 Steps: 104060, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001878, Sample Num: 30048, Cur Loss: 0.33412707, Cur Avg Loss: 0.17023577, Log Avg loss: 0.19293428, Global Avg Loss: 0.63154454, Time: 0.0209 Steps: 104070, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001888, Sample Num: 30208, Cur Loss: 0.49831802, Cur Avg Loss: 0.17025245, Log Avg loss: 0.17338446, Global Avg Loss: 0.63150052, Time: 0.0210 Steps: 104080, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001898, Sample Num: 30368, Cur Loss: 0.12381129, Cur Avg Loss: 0.17023930, Log Avg loss: 0.16775794, Global Avg Loss: 0.63145597, Time: 0.0209 Steps: 104090, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001908, Sample Num: 30528, Cur Loss: 0.16363716, Cur Avg Loss: 0.17006219, Log Avg loss: 0.13644600, Global Avg Loss: 0.63140842, Time: 0.0209 Steps: 104100, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001918, Sample Num: 30688, Cur Loss: 0.21207064, Cur Avg Loss: 0.16998633, Log Avg loss: 0.15551216, Global Avg Loss: 0.63136270, Time: 0.0209 Steps: 104110, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001928, Sample Num: 30848, Cur Loss: 0.04112050, Cur Avg Loss: 0.16993322, Log Avg loss: 0.15974675, Global Avg Loss: 0.63131741, Time: 0.0209 Steps: 104120, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001938, Sample Num: 31008, Cur Loss: 0.06133735, Cur Avg Loss: 0.16990595, Log Avg loss: 0.16464812, Global Avg Loss: 0.63127259, Time: 0.0208 Steps: 104130, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001948, Sample Num: 31168, Cur Loss: 0.26878762, Cur Avg Loss: 0.16970159, Log Avg loss: 0.13009674, Global Avg Loss: 0.63122447, Time: 0.0208 Steps: 104140, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001958, Sample Num: 31328, Cur Loss: 0.33259687, Cur Avg Loss: 0.16999421, Log Avg loss: 0.22699728, Global Avg Loss: 0.63118566, Time: 0.0209 Steps: 104150, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001968, Sample Num: 31488, Cur Loss: 0.32375818, Cur Avg Loss: 0.17011125, Log Avg loss: 0.19302726, Global Avg Loss: 0.63114359, Time: 0.0208 Steps: 104160, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001978, Sample Num: 31648, Cur Loss: 0.16200250, Cur Avg Loss: 0.16997059, Log Avg loss: 0.14228862, Global Avg Loss: 0.63109666, Time: 0.0209 Steps: 104170, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001988, Sample Num: 31808, Cur Loss: 0.27516371, Cur Avg Loss: 0.17001140, Log Avg loss: 0.17808432, Global Avg Loss: 0.63105318, Time: 0.0209 Steps: 104180, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001998, Sample Num: 31968, Cur Loss: 0.26117718, Cur Avg Loss: 0.17002728, Log Avg loss: 0.17318427, Global Avg Loss: 0.63100923, Time: 0.0209 Steps: 104190, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002008, Sample Num: 32128, Cur Loss: 0.24046949, Cur Avg Loss: 0.17027176, Log Avg loss: 0.21911812, Global Avg Loss: 0.63096970, Time: 0.0209 Steps: 104200, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002018, Sample Num: 32288, Cur Loss: 0.20357245, Cur Avg Loss: 0.17062502, Log Avg loss: 0.24156061, Global Avg Loss: 0.63093234, Time: 0.0208 Steps: 104210, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002028, Sample Num: 32448, Cur Loss: 0.34617162, Cur Avg Loss: 0.17083105, Log Avg loss: 0.21240624, Global Avg Loss: 0.63089218, Time: 0.0209 Steps: 104220, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002038, Sample Num: 32608, Cur Loss: 0.16435406, Cur Avg Loss: 0.17084292, Log Avg loss: 0.17325059, Global Avg Loss: 0.63084827, Time: 0.0209 Steps: 104230, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002048, Sample Num: 32768, Cur Loss: 0.41929600, Cur Avg Loss: 0.17106086, Log Avg loss: 0.21547812, Global Avg Loss: 0.63080842, Time: 0.0254 Steps: 104240, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002058, Sample Num: 32928, Cur Loss: 0.22702980, Cur Avg Loss: 0.17114738, Log Avg loss: 0.18886685, Global Avg Loss: 0.63076603, Time: 0.0207 Steps: 104250, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002068, Sample Num: 33088, Cur Loss: 0.10263281, Cur Avg Loss: 0.17132937, Log Avg loss: 0.20878290, Global Avg Loss: 0.63072556, Time: 0.0208 Steps: 104260, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002078, Sample Num: 33248, Cur Loss: 0.06661143, Cur Avg Loss: 0.17118381, Log Avg loss: 0.14108151, Global Avg Loss: 0.63067860, Time: 0.0207 Steps: 104270, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002088, Sample Num: 33408, Cur Loss: 0.37421983, Cur Avg Loss: 0.17103237, Log Avg loss: 0.13956250, Global Avg Loss: 0.63063150, Time: 0.0208 Steps: 104280, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002098, Sample Num: 33568, Cur Loss: 0.10524856, Cur Avg Loss: 0.17081834, Log Avg loss: 0.12612809, Global Avg Loss: 0.63058313, Time: 0.0207 Steps: 104290, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002108, Sample Num: 33728, Cur Loss: 0.25069630, Cur Avg Loss: 0.17079407, Log Avg loss: 0.16570421, Global Avg Loss: 0.63053856, Time: 0.0207 Steps: 104300, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002118, Sample Num: 33888, Cur Loss: 0.34558949, Cur Avg Loss: 0.17072850, Log Avg loss: 0.15690602, Global Avg Loss: 0.63049315, Time: 0.0207 Steps: 104310, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002128, Sample Num: 34048, Cur Loss: 0.12744194, Cur Avg Loss: 0.17069553, Log Avg loss: 0.16371181, Global Avg Loss: 0.63044840, Time: 0.0208 Steps: 104320, Updated lr: 0.000002 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.735286, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.267679, "eval_total_loss": 188.178059, "eval_mae": 0.40269, "eval_mse": 0.267688, "eval_r2": 0.82984, "eval_sp_statistic": 0.904303, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.926837, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.300507, "test_total_loss": 150.854465, "test_mae": 0.340725, "test_mse": 0.300614, "test_r2": 0.805981, "test_sp_statistic": 0.877523, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.923565, "test_ps_pvalue": 0.0, "lr": 2.0189663347558085e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6304429759679259, "train_cur_epoch_loss": 363.30425975006074, "train_cur_epoch_avg_loss": 0.17064549542041368, "train_cur_epoch_time": 44.735286235809326, "train_cur_epoch_avg_time": 0.021012346752376385, "epoch": 49, "step": 104321} ################################################## Training, Epoch: 0050, Batch: 000009, Sample Num: 144, Cur Loss: 0.11963327, Cur Avg Loss: 0.13423126, Log Avg loss: 0.12722531, Global Avg Loss: 0.63040017, Time: 0.0208 Steps: 104330, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000019, Sample Num: 304, Cur Loss: 0.20669852, Cur Avg Loss: 0.15000721, Log Avg loss: 0.16420557, Global Avg Loss: 0.63035549, Time: 0.0208 Steps: 104340, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000029, Sample Num: 464, Cur Loss: 0.09744640, Cur Avg Loss: 0.14687991, Log Avg loss: 0.14093803, Global Avg Loss: 0.63030859, Time: 0.0208 Steps: 104350, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000039, Sample Num: 624, Cur Loss: 0.10761130, Cur Avg Loss: 0.14098160, Log Avg loss: 0.12387650, Global Avg Loss: 0.63026006, Time: 0.0208 Steps: 104360, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000049, Sample Num: 784, Cur Loss: 0.44855797, Cur Avg Loss: 0.15091097, Log Avg loss: 0.18963552, Global Avg Loss: 0.63021784, Time: 0.0208 Steps: 104370, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000059, Sample Num: 944, Cur Loss: 0.36253327, Cur Avg Loss: 0.15500121, Log Avg loss: 0.17504337, Global Avg Loss: 0.63017424, Time: 0.0208 Steps: 104380, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000069, Sample Num: 1104, Cur Loss: 0.05054274, Cur Avg Loss: 0.15630255, Log Avg loss: 0.16398045, Global Avg Loss: 0.63012958, Time: 0.0208 Steps: 104390, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000079, Sample Num: 1264, Cur Loss: 0.06710231, Cur Avg Loss: 0.15692460, Log Avg loss: 0.16121676, Global Avg Loss: 0.63008466, Time: 0.0208 Steps: 104400, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000089, Sample Num: 1424, Cur Loss: 0.16497220, Cur Avg Loss: 0.15759184, Log Avg loss: 0.16286304, Global Avg Loss: 0.63003991, Time: 0.0209 Steps: 104410, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000099, Sample Num: 1584, Cur Loss: 0.14395732, Cur Avg Loss: 0.15418646, Log Avg loss: 0.12387854, Global Avg Loss: 0.62999144, Time: 0.0208 Steps: 104420, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000109, Sample Num: 1744, Cur Loss: 0.01812913, Cur Avg Loss: 0.15819375, Log Avg loss: 0.19786595, Global Avg Loss: 0.62995006, Time: 0.0208 Steps: 104430, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000119, Sample Num: 1904, Cur Loss: 0.14231460, Cur Avg Loss: 0.16236138, Log Avg loss: 0.20778857, Global Avg Loss: 0.62990964, Time: 0.0208 Steps: 104440, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000129, Sample Num: 2064, Cur Loss: 0.10723095, Cur Avg Loss: 0.16441831, Log Avg loss: 0.18889571, Global Avg Loss: 0.62986742, Time: 0.0208 Steps: 104450, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000139, Sample Num: 2224, Cur Loss: 0.32147565, Cur Avg Loss: 0.16549798, Log Avg loss: 0.17942573, Global Avg Loss: 0.62982430, Time: 0.0208 Steps: 104460, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000149, Sample Num: 2384, Cur Loss: 0.15755813, Cur Avg Loss: 0.16631321, Log Avg loss: 0.17764494, Global Avg Loss: 0.62978101, Time: 0.0208 Steps: 104470, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000159, Sample Num: 2544, Cur Loss: 0.08331582, Cur Avg Loss: 0.16195641, Log Avg loss: 0.09704016, Global Avg Loss: 0.62973002, Time: 0.0208 Steps: 104480, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000169, Sample Num: 2704, Cur Loss: 0.22532427, Cur Avg Loss: 0.16318045, Log Avg loss: 0.18264266, Global Avg Loss: 0.62968724, Time: 0.0208 Steps: 104490, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000179, Sample Num: 2864, Cur Loss: 0.07753357, Cur Avg Loss: 0.16060886, Log Avg loss: 0.11714890, Global Avg Loss: 0.62963819, Time: 0.0208 Steps: 104500, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000189, Sample Num: 3024, Cur Loss: 0.15510990, Cur Avg Loss: 0.16230523, Log Avg loss: 0.19267034, Global Avg Loss: 0.62959638, Time: 0.0212 Steps: 104510, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000199, Sample Num: 3184, Cur Loss: 0.07100265, Cur Avg Loss: 0.16103287, Log Avg loss: 0.13698520, Global Avg Loss: 0.62954925, Time: 0.0211 Steps: 104520, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000209, Sample Num: 3344, Cur Loss: 0.15721920, Cur Avg Loss: 0.15946236, Log Avg loss: 0.12820919, Global Avg Loss: 0.62950129, Time: 0.0213 Steps: 104530, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000219, Sample Num: 3504, Cur Loss: 0.11707798, Cur Avg Loss: 0.16195848, Log Avg loss: 0.21412737, Global Avg Loss: 0.62946155, Time: 0.0211 Steps: 104540, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000229, Sample Num: 3664, Cur Loss: 0.07153564, Cur Avg Loss: 0.16208071, Log Avg loss: 0.16475758, Global Avg Loss: 0.62941710, Time: 0.0212 Steps: 104550, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000239, Sample Num: 3824, Cur Loss: 0.06292769, Cur Avg Loss: 0.16330586, Log Avg loss: 0.19136177, Global Avg Loss: 0.62937521, Time: 0.0211 Steps: 104560, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000249, Sample Num: 3984, Cur Loss: 0.05048726, Cur Avg Loss: 0.16549474, Log Avg loss: 0.21780903, Global Avg Loss: 0.62933585, Time: 0.0211 Steps: 104570, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000259, Sample Num: 4144, Cur Loss: 0.11238591, Cur Avg Loss: 0.16452603, Log Avg loss: 0.14040524, Global Avg Loss: 0.62928910, Time: 0.0248 Steps: 104580, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000269, Sample Num: 4304, Cur Loss: 0.09909829, Cur Avg Loss: 0.16402056, Log Avg loss: 0.15092890, Global Avg Loss: 0.62924336, Time: 0.0211 Steps: 104590, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000279, Sample Num: 4464, Cur Loss: 0.13472900, Cur Avg Loss: 0.16551275, Log Avg loss: 0.20565254, Global Avg Loss: 0.62920287, Time: 0.0211 Steps: 104600, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000289, Sample Num: 4624, Cur Loss: 0.21656999, Cur Avg Loss: 0.16385213, Log Avg loss: 0.11752096, Global Avg Loss: 0.62915395, Time: 0.0211 Steps: 104610, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000299, Sample Num: 4784, Cur Loss: 0.27178049, Cur Avg Loss: 0.16393267, Log Avg loss: 0.16626005, Global Avg Loss: 0.62910971, Time: 0.0211 Steps: 104620, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000309, Sample Num: 4944, Cur Loss: 0.11225986, Cur Avg Loss: 0.16253948, Log Avg loss: 0.12088311, Global Avg Loss: 0.62906113, Time: 0.0210 Steps: 104630, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000319, Sample Num: 5104, Cur Loss: 0.10977136, Cur Avg Loss: 0.16219839, Log Avg loss: 0.15165865, Global Avg Loss: 0.62901551, Time: 0.0211 Steps: 104640, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000329, Sample Num: 5264, Cur Loss: 0.10888436, Cur Avg Loss: 0.16171819, Log Avg loss: 0.14639992, Global Avg Loss: 0.62896939, Time: 0.0211 Steps: 104650, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000339, Sample Num: 5424, Cur Loss: 0.14345598, Cur Avg Loss: 0.16305755, Log Avg loss: 0.20712256, Global Avg Loss: 0.62892909, Time: 0.0211 Steps: 104660, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000349, Sample Num: 5584, Cur Loss: 0.32703307, Cur Avg Loss: 0.16749410, Log Avg loss: 0.31789298, Global Avg Loss: 0.62889937, Time: 0.0211 Steps: 104670, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000359, Sample Num: 5744, Cur Loss: 0.18727158, Cur Avg Loss: 0.16770631, Log Avg loss: 0.17511248, Global Avg Loss: 0.62885602, Time: 0.0211 Steps: 104680, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000369, Sample Num: 5904, Cur Loss: 0.03921727, Cur Avg Loss: 0.16728783, Log Avg loss: 0.15226435, Global Avg Loss: 0.62881050, Time: 0.0211 Steps: 104690, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000379, Sample Num: 6064, Cur Loss: 0.13944720, Cur Avg Loss: 0.16662134, Log Avg loss: 0.14202817, Global Avg Loss: 0.62876400, Time: 0.0212 Steps: 104700, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000389, Sample Num: 6224, Cur Loss: 0.25015339, Cur Avg Loss: 0.16632653, Log Avg loss: 0.15515304, Global Avg Loss: 0.62871877, Time: 0.0211 Steps: 104710, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000399, Sample Num: 6384, Cur Loss: 0.10629235, Cur Avg Loss: 0.16570232, Log Avg loss: 0.14142078, Global Avg Loss: 0.62867224, Time: 0.0212 Steps: 104720, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000409, Sample Num: 6544, Cur Loss: 0.17803106, Cur Avg Loss: 0.16663712, Log Avg loss: 0.20393536, Global Avg Loss: 0.62863168, Time: 0.0211 Steps: 104730, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000419, Sample Num: 6704, Cur Loss: 0.13626069, Cur Avg Loss: 0.16624348, Log Avg loss: 0.15014358, Global Avg Loss: 0.62858600, Time: 0.0210 Steps: 104740, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000429, Sample Num: 6864, Cur Loss: 0.31747282, Cur Avg Loss: 0.16697095, Log Avg loss: 0.19745213, Global Avg Loss: 0.62854484, Time: 0.0211 Steps: 104750, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000439, Sample Num: 7024, Cur Loss: 0.18580467, Cur Avg Loss: 0.16638249, Log Avg loss: 0.14113728, Global Avg Loss: 0.62849832, Time: 0.0211 Steps: 104760, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000449, Sample Num: 7184, Cur Loss: 0.13249969, Cur Avg Loss: 0.16691540, Log Avg loss: 0.19031017, Global Avg Loss: 0.62845649, Time: 0.0211 Steps: 104770, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000459, Sample Num: 7344, Cur Loss: 0.05632089, Cur Avg Loss: 0.16740969, Log Avg loss: 0.18960358, Global Avg Loss: 0.62841461, Time: 0.0211 Steps: 104780, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000469, Sample Num: 7504, Cur Loss: 0.12455349, Cur Avg Loss: 0.16650623, Log Avg loss: 0.12503738, Global Avg Loss: 0.62836657, Time: 0.0212 Steps: 104790, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000479, Sample Num: 7664, Cur Loss: 0.07763040, Cur Avg Loss: 0.16693219, Log Avg loss: 0.18690988, Global Avg Loss: 0.62832445, Time: 0.0212 Steps: 104800, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000489, Sample Num: 7824, Cur Loss: 0.15444383, Cur Avg Loss: 0.16574015, Log Avg loss: 0.10864126, Global Avg Loss: 0.62827487, Time: 0.0211 Steps: 104810, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000499, Sample Num: 7984, Cur Loss: 0.13608713, Cur Avg Loss: 0.16600406, Log Avg loss: 0.17890908, Global Avg Loss: 0.62823200, Time: 0.0212 Steps: 104820, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000509, Sample Num: 8144, Cur Loss: 0.33263937, Cur Avg Loss: 0.16533404, Log Avg loss: 0.13190004, Global Avg Loss: 0.62818465, Time: 0.0211 Steps: 104830, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000519, Sample Num: 8304, Cur Loss: 0.25705621, Cur Avg Loss: 0.16466738, Log Avg loss: 0.13073474, Global Avg Loss: 0.62813720, Time: 0.0212 Steps: 104840, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000529, Sample Num: 8464, Cur Loss: 0.11103162, Cur Avg Loss: 0.16451978, Log Avg loss: 0.15685929, Global Avg Loss: 0.62809225, Time: 0.0212 Steps: 104850, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000539, Sample Num: 8624, Cur Loss: 0.06625794, Cur Avg Loss: 0.16521573, Log Avg loss: 0.20203124, Global Avg Loss: 0.62805162, Time: 0.0212 Steps: 104860, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000549, Sample Num: 8784, Cur Loss: 0.06870039, Cur Avg Loss: 0.16523228, Log Avg loss: 0.16612425, Global Avg Loss: 0.62800757, Time: 0.0211 Steps: 104870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000559, Sample Num: 8944, Cur Loss: 0.78804624, Cur Avg Loss: 0.16566091, Log Avg loss: 0.18919273, Global Avg Loss: 0.62796573, Time: 0.0212 Steps: 104880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000569, Sample Num: 9104, Cur Loss: 0.25309548, Cur Avg Loss: 0.16515813, Log Avg loss: 0.13705277, Global Avg Loss: 0.62791893, Time: 0.0212 Steps: 104890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000579, Sample Num: 9264, Cur Loss: 0.24500461, Cur Avg Loss: 0.16531711, Log Avg loss: 0.17436288, Global Avg Loss: 0.62787569, Time: 0.0212 Steps: 104900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000589, Sample Num: 9424, Cur Loss: 0.12675509, Cur Avg Loss: 0.16559895, Log Avg loss: 0.18191803, Global Avg Loss: 0.62783319, Time: 0.0212 Steps: 104910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000599, Sample Num: 9584, Cur Loss: 0.40262252, Cur Avg Loss: 0.16675973, Log Avg loss: 0.23512918, Global Avg Loss: 0.62779576, Time: 0.0211 Steps: 104920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000609, Sample Num: 9744, Cur Loss: 0.04813788, Cur Avg Loss: 0.16707913, Log Avg loss: 0.18621159, Global Avg Loss: 0.62775367, Time: 0.0211 Steps: 104930, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000619, Sample Num: 9904, Cur Loss: 0.14342624, Cur Avg Loss: 0.16706462, Log Avg loss: 0.16618098, Global Avg Loss: 0.62770969, Time: 0.0211 Steps: 104940, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000629, Sample Num: 10064, Cur Loss: 0.21555604, Cur Avg Loss: 0.16716749, Log Avg loss: 0.17353497, Global Avg Loss: 0.62766641, Time: 0.0212 Steps: 104950, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000639, Sample Num: 10224, Cur Loss: 0.34407091, Cur Avg Loss: 0.16742600, Log Avg loss: 0.18368622, Global Avg Loss: 0.62762411, Time: 0.0211 Steps: 104960, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000649, Sample Num: 10384, Cur Loss: 0.03837307, Cur Avg Loss: 0.16790127, Log Avg loss: 0.19827086, Global Avg Loss: 0.62758321, Time: 0.0212 Steps: 104970, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000659, Sample Num: 10544, Cur Loss: 0.08925381, Cur Avg Loss: 0.16782331, Log Avg loss: 0.16276403, Global Avg Loss: 0.62753893, Time: 0.0212 Steps: 104980, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000669, Sample Num: 10704, Cur Loss: 0.12258097, Cur Avg Loss: 0.16759571, Log Avg loss: 0.15259680, Global Avg Loss: 0.62749370, Time: 0.0212 Steps: 104990, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000679, Sample Num: 10864, Cur Loss: 0.15595725, Cur Avg Loss: 0.16725388, Log Avg loss: 0.14438554, Global Avg Loss: 0.62744769, Time: 0.0211 Steps: 105000, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000689, Sample Num: 11024, Cur Loss: 0.19732678, Cur Avg Loss: 0.16755243, Log Avg loss: 0.18782358, Global Avg Loss: 0.62740582, Time: 0.0212 Steps: 105010, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000699, Sample Num: 11184, Cur Loss: 0.01549878, Cur Avg Loss: 0.16639947, Log Avg loss: 0.08696060, Global Avg Loss: 0.62735436, Time: 0.0212 Steps: 105020, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000709, Sample Num: 11344, Cur Loss: 0.07166936, Cur Avg Loss: 0.16587526, Log Avg loss: 0.12923335, Global Avg Loss: 0.62730693, Time: 0.0212 Steps: 105030, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000719, Sample Num: 11504, Cur Loss: 0.07671355, Cur Avg Loss: 0.16559213, Log Avg loss: 0.14551798, Global Avg Loss: 0.62726107, Time: 0.0211 Steps: 105040, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000729, Sample Num: 11664, Cur Loss: 0.17315559, Cur Avg Loss: 0.16600120, Log Avg loss: 0.19541357, Global Avg Loss: 0.62721996, Time: 0.0213 Steps: 105050, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000739, Sample Num: 11824, Cur Loss: 0.10684016, Cur Avg Loss: 0.16481818, Log Avg loss: 0.07857591, Global Avg Loss: 0.62716774, Time: 0.0211 Steps: 105060, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000749, Sample Num: 11984, Cur Loss: 0.24692149, Cur Avg Loss: 0.16448150, Log Avg loss: 0.13960105, Global Avg Loss: 0.62712133, Time: 0.0212 Steps: 105070, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000759, Sample Num: 12144, Cur Loss: 0.11309363, Cur Avg Loss: 0.16423059, Log Avg loss: 0.14543723, Global Avg Loss: 0.62707549, Time: 0.0211 Steps: 105080, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000769, Sample Num: 12304, Cur Loss: 0.06543321, Cur Avg Loss: 0.16364071, Log Avg loss: 0.11886896, Global Avg Loss: 0.62702713, Time: 0.0247 Steps: 105090, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000779, Sample Num: 12464, Cur Loss: 0.44701025, Cur Avg Loss: 0.16525029, Log Avg loss: 0.28902713, Global Avg Loss: 0.62699497, Time: 0.0211 Steps: 105100, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000789, Sample Num: 12624, Cur Loss: 0.24826698, Cur Avg Loss: 0.16544610, Log Avg loss: 0.18069898, Global Avg Loss: 0.62695251, Time: 0.0211 Steps: 105110, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000799, Sample Num: 12784, Cur Loss: 0.16397682, Cur Avg Loss: 0.16512841, Log Avg loss: 0.14006271, Global Avg Loss: 0.62690620, Time: 0.0211 Steps: 105120, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000809, Sample Num: 12944, Cur Loss: 0.11734632, Cur Avg Loss: 0.16510901, Log Avg loss: 0.16355909, Global Avg Loss: 0.62686212, Time: 0.0212 Steps: 105130, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000819, Sample Num: 13104, Cur Loss: 0.18604234, Cur Avg Loss: 0.16592331, Log Avg loss: 0.23180048, Global Avg Loss: 0.62682455, Time: 0.0212 Steps: 105140, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000829, Sample Num: 13264, Cur Loss: 0.09134348, Cur Avg Loss: 0.16616960, Log Avg loss: 0.18634046, Global Avg Loss: 0.62678266, Time: 0.0212 Steps: 105150, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000839, Sample Num: 13424, Cur Loss: 0.04025123, Cur Avg Loss: 0.16640660, Log Avg loss: 0.18605412, Global Avg Loss: 0.62674075, Time: 0.0211 Steps: 105160, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000849, Sample Num: 13584, Cur Loss: 0.16539697, Cur Avg Loss: 0.16620951, Log Avg loss: 0.14967394, Global Avg Loss: 0.62669538, Time: 0.0211 Steps: 105170, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000859, Sample Num: 13744, Cur Loss: 0.14228210, Cur Avg Loss: 0.16581408, Log Avg loss: 0.13224212, Global Avg Loss: 0.62664837, Time: 0.0211 Steps: 105180, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000869, Sample Num: 13904, Cur Loss: 0.13141775, Cur Avg Loss: 0.16560987, Log Avg loss: 0.14806798, Global Avg Loss: 0.62660288, Time: 0.0211 Steps: 105190, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000879, Sample Num: 14064, Cur Loss: 0.07373699, Cur Avg Loss: 0.16504978, Log Avg loss: 0.11637764, Global Avg Loss: 0.62655438, Time: 0.0211 Steps: 105200, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000889, Sample Num: 14224, Cur Loss: 0.12704113, Cur Avg Loss: 0.16564420, Log Avg loss: 0.21789395, Global Avg Loss: 0.62651553, Time: 0.0211 Steps: 105210, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000899, Sample Num: 14384, Cur Loss: 0.09679942, Cur Avg Loss: 0.16559473, Log Avg loss: 0.16119703, Global Avg Loss: 0.62647131, Time: 0.0210 Steps: 105220, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000909, Sample Num: 14544, Cur Loss: 0.38837624, Cur Avg Loss: 0.16611894, Log Avg loss: 0.21324559, Global Avg Loss: 0.62643204, Time: 0.0210 Steps: 105230, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000919, Sample Num: 14704, Cur Loss: 0.17979544, Cur Avg Loss: 0.16590698, Log Avg loss: 0.14663918, Global Avg Loss: 0.62638645, Time: 0.0211 Steps: 105240, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000929, Sample Num: 14864, Cur Loss: 0.16631982, Cur Avg Loss: 0.16586569, Log Avg loss: 0.16207127, Global Avg Loss: 0.62634234, Time: 0.0210 Steps: 105250, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000939, Sample Num: 15024, Cur Loss: 0.11673781, Cur Avg Loss: 0.16512150, Log Avg loss: 0.09598650, Global Avg Loss: 0.62629195, Time: 0.0211 Steps: 105260, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000949, Sample Num: 15184, Cur Loss: 0.11883178, Cur Avg Loss: 0.16482665, Log Avg loss: 0.13713989, Global Avg Loss: 0.62624548, Time: 0.0211 Steps: 105270, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000959, Sample Num: 15344, Cur Loss: 0.38291556, Cur Avg Loss: 0.16492708, Log Avg loss: 0.17445836, Global Avg Loss: 0.62620257, Time: 0.0212 Steps: 105280, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000969, Sample Num: 15504, Cur Loss: 0.22029781, Cur Avg Loss: 0.16467307, Log Avg loss: 0.14031308, Global Avg Loss: 0.62615642, Time: 0.0212 Steps: 105290, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000979, Sample Num: 15664, Cur Loss: 0.14018619, Cur Avg Loss: 0.16478356, Log Avg loss: 0.17549030, Global Avg Loss: 0.62611363, Time: 0.0210 Steps: 105300, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000989, Sample Num: 15824, Cur Loss: 0.29675370, Cur Avg Loss: 0.16500158, Log Avg loss: 0.18634556, Global Avg Loss: 0.62607187, Time: 0.0211 Steps: 105310, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000999, Sample Num: 15984, Cur Loss: 0.07823217, Cur Avg Loss: 0.16479901, Log Avg loss: 0.14476442, Global Avg Loss: 0.62602617, Time: 0.0211 Steps: 105320, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001009, Sample Num: 16144, Cur Loss: 0.13467471, Cur Avg Loss: 0.16436986, Log Avg loss: 0.12149868, Global Avg Loss: 0.62597827, Time: 0.0210 Steps: 105330, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001019, Sample Num: 16304, Cur Loss: 0.14034525, Cur Avg Loss: 0.16463052, Log Avg loss: 0.19093020, Global Avg Loss: 0.62593697, Time: 0.0212 Steps: 105340, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001029, Sample Num: 16464, Cur Loss: 0.06463397, Cur Avg Loss: 0.16452449, Log Avg loss: 0.15372010, Global Avg Loss: 0.62589214, Time: 0.0248 Steps: 105350, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001039, Sample Num: 16624, Cur Loss: 0.19372009, Cur Avg Loss: 0.16488204, Log Avg loss: 0.20167470, Global Avg Loss: 0.62585188, Time: 0.0211 Steps: 105360, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001049, Sample Num: 16784, Cur Loss: 0.11741640, Cur Avg Loss: 0.16463638, Log Avg loss: 0.13911167, Global Avg Loss: 0.62580569, Time: 0.0210 Steps: 105370, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001059, Sample Num: 16944, Cur Loss: 0.04945809, Cur Avg Loss: 0.16473826, Log Avg loss: 0.17542630, Global Avg Loss: 0.62576295, Time: 0.0210 Steps: 105380, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001069, Sample Num: 17104, Cur Loss: 0.17488596, Cur Avg Loss: 0.16481894, Log Avg loss: 0.17336213, Global Avg Loss: 0.62572002, Time: 0.0211 Steps: 105390, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001079, Sample Num: 17264, Cur Loss: 0.15506414, Cur Avg Loss: 0.16507555, Log Avg loss: 0.19250768, Global Avg Loss: 0.62567892, Time: 0.0211 Steps: 105400, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001089, Sample Num: 17424, Cur Loss: 0.03427033, Cur Avg Loss: 0.16450027, Log Avg loss: 0.10242698, Global Avg Loss: 0.62562928, Time: 0.0211 Steps: 105410, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001099, Sample Num: 17584, Cur Loss: 0.13264406, Cur Avg Loss: 0.16450607, Log Avg loss: 0.16513793, Global Avg Loss: 0.62558560, Time: 0.0210 Steps: 105420, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001109, Sample Num: 17744, Cur Loss: 0.09685122, Cur Avg Loss: 0.16440193, Log Avg loss: 0.15295701, Global Avg Loss: 0.62554077, Time: 0.0210 Steps: 105430, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001119, Sample Num: 17904, Cur Loss: 0.11976592, Cur Avg Loss: 0.16439264, Log Avg loss: 0.16336192, Global Avg Loss: 0.62549694, Time: 0.0210 Steps: 105440, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001129, Sample Num: 18064, Cur Loss: 0.11749054, Cur Avg Loss: 0.16449803, Log Avg loss: 0.17629144, Global Avg Loss: 0.62545434, Time: 0.0210 Steps: 105450, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001139, Sample Num: 18224, Cur Loss: 0.24259996, Cur Avg Loss: 0.16510046, Log Avg loss: 0.23311491, Global Avg Loss: 0.62541714, Time: 0.0210 Steps: 105460, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001149, Sample Num: 18384, Cur Loss: 0.04039286, Cur Avg Loss: 0.16472819, Log Avg loss: 0.12232711, Global Avg Loss: 0.62536944, Time: 0.0210 Steps: 105470, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001159, Sample Num: 18544, Cur Loss: 0.08653566, Cur Avg Loss: 0.16463726, Log Avg loss: 0.15418926, Global Avg Loss: 0.62532477, Time: 0.0210 Steps: 105480, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001169, Sample Num: 18704, Cur Loss: 0.29079002, Cur Avg Loss: 0.16437704, Log Avg loss: 0.13421782, Global Avg Loss: 0.62527821, Time: 0.0211 Steps: 105490, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001179, Sample Num: 18864, Cur Loss: 0.07876848, Cur Avg Loss: 0.16428443, Log Avg loss: 0.15345757, Global Avg Loss: 0.62523349, Time: 0.0210 Steps: 105500, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001189, Sample Num: 19024, Cur Loss: 0.21989048, Cur Avg Loss: 0.16420674, Log Avg loss: 0.15504714, Global Avg Loss: 0.62518893, Time: 0.0211 Steps: 105510, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001199, Sample Num: 19184, Cur Loss: 0.26804960, Cur Avg Loss: 0.16425882, Log Avg loss: 0.17045076, Global Avg Loss: 0.62514583, Time: 0.0211 Steps: 105520, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001209, Sample Num: 19344, Cur Loss: 0.23367117, Cur Avg Loss: 0.16509516, Log Avg loss: 0.26537306, Global Avg Loss: 0.62511174, Time: 0.0210 Steps: 105530, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001219, Sample Num: 19504, Cur Loss: 0.19525762, Cur Avg Loss: 0.16497316, Log Avg loss: 0.15022262, Global Avg Loss: 0.62506674, Time: 0.0211 Steps: 105540, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001229, Sample Num: 19664, Cur Loss: 0.16364360, Cur Avg Loss: 0.16522150, Log Avg loss: 0.19549488, Global Avg Loss: 0.62502604, Time: 0.0210 Steps: 105550, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001239, Sample Num: 19824, Cur Loss: 0.14641297, Cur Avg Loss: 0.16529224, Log Avg loss: 0.17398572, Global Avg Loss: 0.62498332, Time: 0.0211 Steps: 105560, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001249, Sample Num: 19984, Cur Loss: 0.21561620, Cur Avg Loss: 0.16515157, Log Avg loss: 0.14772243, Global Avg Loss: 0.62493811, Time: 0.0211 Steps: 105570, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001259, Sample Num: 20144, Cur Loss: 0.24213138, Cur Avg Loss: 0.16533418, Log Avg loss: 0.18814235, Global Avg Loss: 0.62489674, Time: 0.0210 Steps: 105580, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001269, Sample Num: 20304, Cur Loss: 0.15786693, Cur Avg Loss: 0.16586506, Log Avg loss: 0.23270349, Global Avg Loss: 0.62485959, Time: 0.0210 Steps: 105590, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001279, Sample Num: 20464, Cur Loss: 0.05381106, Cur Avg Loss: 0.16574047, Log Avg loss: 0.14992898, Global Avg Loss: 0.62481462, Time: 0.0211 Steps: 105600, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001289, Sample Num: 20624, Cur Loss: 0.29574805, Cur Avg Loss: 0.16546037, Log Avg loss: 0.12963641, Global Avg Loss: 0.62476773, Time: 0.0211 Steps: 105610, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001299, Sample Num: 20784, Cur Loss: 0.08351663, Cur Avg Loss: 0.16572645, Log Avg loss: 0.20002363, Global Avg Loss: 0.62472752, Time: 0.0211 Steps: 105620, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001309, Sample Num: 20944, Cur Loss: 0.07787296, Cur Avg Loss: 0.16540764, Log Avg loss: 0.12399438, Global Avg Loss: 0.62468011, Time: 0.0212 Steps: 105630, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001319, Sample Num: 21104, Cur Loss: 0.13220674, Cur Avg Loss: 0.16508016, Log Avg loss: 0.12221334, Global Avg Loss: 0.62463255, Time: 0.0211 Steps: 105640, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001329, Sample Num: 21264, Cur Loss: 0.07411878, Cur Avg Loss: 0.16500485, Log Avg loss: 0.15507095, Global Avg Loss: 0.62458810, Time: 0.0212 Steps: 105650, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001339, Sample Num: 21424, Cur Loss: 0.40491566, Cur Avg Loss: 0.16541544, Log Avg loss: 0.21998276, Global Avg Loss: 0.62454981, Time: 0.0212 Steps: 105660, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001349, Sample Num: 21584, Cur Loss: 0.05584941, Cur Avg Loss: 0.16564266, Log Avg loss: 0.19606750, Global Avg Loss: 0.62450926, Time: 0.0212 Steps: 105670, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001359, Sample Num: 21744, Cur Loss: 0.41796806, Cur Avg Loss: 0.16578485, Log Avg loss: 0.18496629, Global Avg Loss: 0.62446767, Time: 0.0211 Steps: 105680, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001369, Sample Num: 21904, Cur Loss: 0.27224380, Cur Avg Loss: 0.16539167, Log Avg loss: 0.11195833, Global Avg Loss: 0.62441918, Time: 0.0211 Steps: 105690, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001379, Sample Num: 22064, Cur Loss: 0.05315252, Cur Avg Loss: 0.16529505, Log Avg loss: 0.15206876, Global Avg Loss: 0.62437449, Time: 0.0212 Steps: 105700, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001389, Sample Num: 22224, Cur Loss: 0.12196446, Cur Avg Loss: 0.16525284, Log Avg loss: 0.15943107, Global Avg Loss: 0.62433051, Time: 0.0211 Steps: 105710, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001399, Sample Num: 22384, Cur Loss: 0.12027294, Cur Avg Loss: 0.16557834, Log Avg loss: 0.21079035, Global Avg Loss: 0.62429139, Time: 0.0212 Steps: 105720, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001409, Sample Num: 22544, Cur Loss: 0.50379086, Cur Avg Loss: 0.16570667, Log Avg loss: 0.18366087, Global Avg Loss: 0.62424972, Time: 0.0212 Steps: 105730, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001419, Sample Num: 22704, Cur Loss: 0.70557290, Cur Avg Loss: 0.16605391, Log Avg loss: 0.21497989, Global Avg Loss: 0.62421101, Time: 0.0210 Steps: 105740, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001429, Sample Num: 22864, Cur Loss: 0.12983550, Cur Avg Loss: 0.16582255, Log Avg loss: 0.13299170, Global Avg Loss: 0.62416456, Time: 0.0212 Steps: 105750, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001439, Sample Num: 23024, Cur Loss: 0.12484352, Cur Avg Loss: 0.16560289, Log Avg loss: 0.13421404, Global Avg Loss: 0.62411823, Time: 0.0210 Steps: 105760, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001449, Sample Num: 23184, Cur Loss: 0.15124086, Cur Avg Loss: 0.16563210, Log Avg loss: 0.16983516, Global Avg Loss: 0.62407528, Time: 0.0211 Steps: 105770, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001459, Sample Num: 23344, Cur Loss: 0.20547605, Cur Avg Loss: 0.16605349, Log Avg loss: 0.22711363, Global Avg Loss: 0.62403776, Time: 0.0210 Steps: 105780, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001469, Sample Num: 23504, Cur Loss: 0.05109750, Cur Avg Loss: 0.16605703, Log Avg loss: 0.16657366, Global Avg Loss: 0.62399451, Time: 0.0211 Steps: 105790, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001479, Sample Num: 23664, Cur Loss: 0.33824760, Cur Avg Loss: 0.16633612, Log Avg loss: 0.20733418, Global Avg Loss: 0.62395513, Time: 0.0212 Steps: 105800, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001489, Sample Num: 23824, Cur Loss: 0.15357199, Cur Avg Loss: 0.16629788, Log Avg loss: 0.16064152, Global Avg Loss: 0.62391134, Time: 0.0212 Steps: 105810, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001499, Sample Num: 23984, Cur Loss: 0.09625228, Cur Avg Loss: 0.16623368, Log Avg loss: 0.15667497, Global Avg Loss: 0.62386719, Time: 0.0211 Steps: 105820, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001509, Sample Num: 24144, Cur Loss: 0.03970693, Cur Avg Loss: 0.16651172, Log Avg loss: 0.20818942, Global Avg Loss: 0.62382791, Time: 0.0211 Steps: 105830, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001519, Sample Num: 24304, Cur Loss: 0.24414507, Cur Avg Loss: 0.16674011, Log Avg loss: 0.20120438, Global Avg Loss: 0.62378798, Time: 0.0210 Steps: 105840, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001529, Sample Num: 24464, Cur Loss: 0.04163129, Cur Avg Loss: 0.16698582, Log Avg loss: 0.20430927, Global Avg Loss: 0.62374835, Time: 0.0211 Steps: 105850, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001539, Sample Num: 24624, Cur Loss: 0.08908227, Cur Avg Loss: 0.16680862, Log Avg loss: 0.13971436, Global Avg Loss: 0.62370263, Time: 0.0249 Steps: 105860, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001549, Sample Num: 24784, Cur Loss: 0.13290621, Cur Avg Loss: 0.16684994, Log Avg loss: 0.17320918, Global Avg Loss: 0.62366008, Time: 0.0210 Steps: 105870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001559, Sample Num: 24944, Cur Loss: 0.18633127, Cur Avg Loss: 0.16680918, Log Avg loss: 0.16049623, Global Avg Loss: 0.62361633, Time: 0.0210 Steps: 105880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001569, Sample Num: 25104, Cur Loss: 0.19068927, Cur Avg Loss: 0.16666821, Log Avg loss: 0.14469029, Global Avg Loss: 0.62357110, Time: 0.0210 Steps: 105890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001579, Sample Num: 25264, Cur Loss: 0.16183561, Cur Avg Loss: 0.16665608, Log Avg loss: 0.16475311, Global Avg Loss: 0.62352778, Time: 0.0210 Steps: 105900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001589, Sample Num: 25424, Cur Loss: 0.12595487, Cur Avg Loss: 0.16663456, Log Avg loss: 0.16323713, Global Avg Loss: 0.62348432, Time: 0.0211 Steps: 105910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001599, Sample Num: 25584, Cur Loss: 0.10813883, Cur Avg Loss: 0.16698607, Log Avg loss: 0.22284015, Global Avg Loss: 0.62344649, Time: 0.0210 Steps: 105920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001609, Sample Num: 25744, Cur Loss: 0.06365950, Cur Avg Loss: 0.16676110, Log Avg loss: 0.13078851, Global Avg Loss: 0.62339998, Time: 0.0211 Steps: 105930, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001619, Sample Num: 25904, Cur Loss: 0.06129420, Cur Avg Loss: 0.16655142, Log Avg loss: 0.13281385, Global Avg Loss: 0.62335368, Time: 0.0210 Steps: 105940, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001629, Sample Num: 26064, Cur Loss: 0.14268866, Cur Avg Loss: 0.16643868, Log Avg loss: 0.14818650, Global Avg Loss: 0.62330883, Time: 0.0210 Steps: 105950, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001639, Sample Num: 26224, Cur Loss: 0.06606857, Cur Avg Loss: 0.16681311, Log Avg loss: 0.22780813, Global Avg Loss: 0.62327150, Time: 0.0210 Steps: 105960, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001649, Sample Num: 26384, Cur Loss: 0.16543557, Cur Avg Loss: 0.16674965, Log Avg loss: 0.15634745, Global Avg Loss: 0.62322744, Time: 0.0211 Steps: 105970, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001659, Sample Num: 26544, Cur Loss: 0.12466028, Cur Avg Loss: 0.16671016, Log Avg loss: 0.16019835, Global Avg Loss: 0.62318375, Time: 0.0210 Steps: 105980, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001669, Sample Num: 26704, Cur Loss: 0.25413573, Cur Avg Loss: 0.16654172, Log Avg loss: 0.13859852, Global Avg Loss: 0.62313803, Time: 0.0210 Steps: 105990, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001679, Sample Num: 26864, Cur Loss: 0.14503771, Cur Avg Loss: 0.16672143, Log Avg loss: 0.19671456, Global Avg Loss: 0.62309780, Time: 0.0210 Steps: 106000, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001689, Sample Num: 27024, Cur Loss: 0.20556688, Cur Avg Loss: 0.16652593, Log Avg loss: 0.13370164, Global Avg Loss: 0.62305164, Time: 0.0211 Steps: 106010, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001699, Sample Num: 27184, Cur Loss: 0.18435857, Cur Avg Loss: 0.16664773, Log Avg loss: 0.18721979, Global Avg Loss: 0.62301053, Time: 0.0211 Steps: 106020, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001709, Sample Num: 27344, Cur Loss: 0.05685407, Cur Avg Loss: 0.16626365, Log Avg loss: 0.10100728, Global Avg Loss: 0.62296130, Time: 0.0210 Steps: 106030, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001719, Sample Num: 27504, Cur Loss: 0.07680847, Cur Avg Loss: 0.16614555, Log Avg loss: 0.14596316, Global Avg Loss: 0.62291631, Time: 0.0211 Steps: 106040, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001729, Sample Num: 27664, Cur Loss: 0.07184588, Cur Avg Loss: 0.16612958, Log Avg loss: 0.16338346, Global Avg Loss: 0.62287298, Time: 0.0212 Steps: 106050, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001739, Sample Num: 27824, Cur Loss: 0.07266372, Cur Avg Loss: 0.16641777, Log Avg loss: 0.21624584, Global Avg Loss: 0.62283464, Time: 0.0210 Steps: 106060, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001749, Sample Num: 27984, Cur Loss: 0.07874046, Cur Avg Loss: 0.16600747, Log Avg loss: 0.09465783, Global Avg Loss: 0.62278485, Time: 0.0210 Steps: 106070, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001759, Sample Num: 28144, Cur Loss: 0.05140237, Cur Avg Loss: 0.16586198, Log Avg loss: 0.14041575, Global Avg Loss: 0.62273938, Time: 0.0210 Steps: 106080, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001769, Sample Num: 28304, Cur Loss: 0.30315697, Cur Avg Loss: 0.16625686, Log Avg loss: 0.23571553, Global Avg Loss: 0.62270289, Time: 0.0211 Steps: 106090, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001779, Sample Num: 28464, Cur Loss: 0.32405645, Cur Avg Loss: 0.16641426, Log Avg loss: 0.19425888, Global Avg Loss: 0.62266251, Time: 0.0211 Steps: 106100, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001789, Sample Num: 28624, Cur Loss: 0.20975077, Cur Avg Loss: 0.16650723, Log Avg loss: 0.18304593, Global Avg Loss: 0.62262108, Time: 0.0211 Steps: 106110, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001799, Sample Num: 28784, Cur Loss: 0.32331932, Cur Avg Loss: 0.16650118, Log Avg loss: 0.16541911, Global Avg Loss: 0.62257800, Time: 0.0212 Steps: 106120, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001809, Sample Num: 28944, Cur Loss: 0.10414954, Cur Avg Loss: 0.16632961, Log Avg loss: 0.13546400, Global Avg Loss: 0.62253210, Time: 0.0211 Steps: 106130, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001819, Sample Num: 29104, Cur Loss: 0.13173495, Cur Avg Loss: 0.16614765, Log Avg loss: 0.13323021, Global Avg Loss: 0.62248600, Time: 0.0211 Steps: 106140, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001829, Sample Num: 29264, Cur Loss: 0.03932278, Cur Avg Loss: 0.16586268, Log Avg loss: 0.11402776, Global Avg Loss: 0.62243810, Time: 0.0212 Steps: 106150, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001839, Sample Num: 29424, Cur Loss: 0.10001216, Cur Avg Loss: 0.16587446, Log Avg loss: 0.16802812, Global Avg Loss: 0.62239530, Time: 0.0211 Steps: 106160, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001849, Sample Num: 29584, Cur Loss: 0.36398619, Cur Avg Loss: 0.16576837, Log Avg loss: 0.14625873, Global Avg Loss: 0.62235045, Time: 0.0211 Steps: 106170, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001859, Sample Num: 29744, Cur Loss: 0.12237920, Cur Avg Loss: 0.16594447, Log Avg loss: 0.19850624, Global Avg Loss: 0.62231053, Time: 0.0211 Steps: 106180, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001869, Sample Num: 29904, Cur Loss: 0.04501563, Cur Avg Loss: 0.16592295, Log Avg loss: 0.16192210, Global Avg Loss: 0.62226718, Time: 0.0211 Steps: 106190, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001879, Sample Num: 30064, Cur Loss: 0.22196792, Cur Avg Loss: 0.16581774, Log Avg loss: 0.14615374, Global Avg Loss: 0.62222235, Time: 0.0211 Steps: 106200, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001889, Sample Num: 30224, Cur Loss: 0.07792460, Cur Avg Loss: 0.16647012, Log Avg loss: 0.28905303, Global Avg Loss: 0.62219098, Time: 0.0211 Steps: 106210, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001899, Sample Num: 30384, Cur Loss: 0.15000269, Cur Avg Loss: 0.16644890, Log Avg loss: 0.16243878, Global Avg Loss: 0.62214769, Time: 0.0210 Steps: 106220, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001909, Sample Num: 30544, Cur Loss: 0.08235091, Cur Avg Loss: 0.16639597, Log Avg loss: 0.15634552, Global Avg Loss: 0.62210385, Time: 0.0211 Steps: 106230, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001919, Sample Num: 30704, Cur Loss: 0.25567049, Cur Avg Loss: 0.16669203, Log Avg loss: 0.22321046, Global Avg Loss: 0.62206630, Time: 0.0210 Steps: 106240, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001929, Sample Num: 30864, Cur Loss: 0.23306127, Cur Avg Loss: 0.16681829, Log Avg loss: 0.19104624, Global Avg Loss: 0.62202573, Time: 0.0211 Steps: 106250, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001939, Sample Num: 31024, Cur Loss: 0.09819093, Cur Avg Loss: 0.16692579, Log Avg loss: 0.18766254, Global Avg Loss: 0.62198486, Time: 0.0211 Steps: 106260, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001949, Sample Num: 31184, Cur Loss: 0.20564935, Cur Avg Loss: 0.16687605, Log Avg loss: 0.15723127, Global Avg Loss: 0.62194112, Time: 0.0211 Steps: 106270, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001959, Sample Num: 31344, Cur Loss: 0.14661872, Cur Avg Loss: 0.16704388, Log Avg loss: 0.19975498, Global Avg Loss: 0.62190140, Time: 0.0211 Steps: 106280, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001969, Sample Num: 31504, Cur Loss: 0.16661373, Cur Avg Loss: 0.16703908, Log Avg loss: 0.16609894, Global Avg Loss: 0.62185852, Time: 0.0211 Steps: 106290, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001979, Sample Num: 31664, Cur Loss: 0.44138113, Cur Avg Loss: 0.16712883, Log Avg loss: 0.18480025, Global Avg Loss: 0.62181740, Time: 0.0211 Steps: 106300, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001989, Sample Num: 31824, Cur Loss: 0.11393604, Cur Avg Loss: 0.16699053, Log Avg loss: 0.13962140, Global Avg Loss: 0.62177204, Time: 0.0212 Steps: 106310, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001999, Sample Num: 31984, Cur Loss: 0.18854740, Cur Avg Loss: 0.16706981, Log Avg loss: 0.18283739, Global Avg Loss: 0.62173076, Time: 0.0211 Steps: 106320, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002009, Sample Num: 32144, Cur Loss: 0.15110055, Cur Avg Loss: 0.16718184, Log Avg loss: 0.18957824, Global Avg Loss: 0.62169012, Time: 0.0211 Steps: 106330, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002019, Sample Num: 32304, Cur Loss: 0.53304136, Cur Avg Loss: 0.16717685, Log Avg loss: 0.16617329, Global Avg Loss: 0.62164728, Time: 0.0211 Steps: 106340, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002029, Sample Num: 32464, Cur Loss: 0.38337573, Cur Avg Loss: 0.16761994, Log Avg loss: 0.25708049, Global Avg Loss: 0.62161300, Time: 0.0210 Steps: 106350, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002039, Sample Num: 32624, Cur Loss: 0.15545782, Cur Avg Loss: 0.16791077, Log Avg loss: 0.22691931, Global Avg Loss: 0.62157589, Time: 0.0211 Steps: 106360, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002049, Sample Num: 32784, Cur Loss: 0.11556783, Cur Avg Loss: 0.16816671, Log Avg loss: 0.22035335, Global Avg Loss: 0.62153817, Time: 0.0248 Steps: 106370, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002059, Sample Num: 32944, Cur Loss: 0.07375361, Cur Avg Loss: 0.16802199, Log Avg loss: 0.13836989, Global Avg Loss: 0.62149275, Time: 0.0211 Steps: 106380, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002069, Sample Num: 33104, Cur Loss: 0.05184508, Cur Avg Loss: 0.16778502, Log Avg loss: 0.11899205, Global Avg Loss: 0.62144552, Time: 0.0212 Steps: 106390, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002079, Sample Num: 33264, Cur Loss: 0.26804757, Cur Avg Loss: 0.16788050, Log Avg loss: 0.18763457, Global Avg Loss: 0.62140475, Time: 0.0212 Steps: 106400, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002089, Sample Num: 33424, Cur Loss: 0.03558430, Cur Avg Loss: 0.16761215, Log Avg loss: 0.11182340, Global Avg Loss: 0.62135686, Time: 0.0211 Steps: 106410, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002099, Sample Num: 33584, Cur Loss: 0.19742171, Cur Avg Loss: 0.16773734, Log Avg loss: 0.19388971, Global Avg Loss: 0.62131669, Time: 0.0212 Steps: 106420, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002109, Sample Num: 33744, Cur Loss: 0.09679288, Cur Avg Loss: 0.16743032, Log Avg loss: 0.10298538, Global Avg Loss: 0.62126799, Time: 0.0211 Steps: 106430, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002119, Sample Num: 33904, Cur Loss: 0.12208122, Cur Avg Loss: 0.16749415, Log Avg loss: 0.18095749, Global Avg Loss: 0.62122662, Time: 0.0211 Steps: 106440, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002129, Sample Num: 34055, Cur Loss: 0.18002906, Cur Avg Loss: 0.16774396, Log Avg loss: 0.22067767, Global Avg Loss: 0.62118900, Time: 0.0102 Steps: 106450, Updated lr: 0.000000 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.085704, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.2608, "eval_total_loss": 183.342461, "eval_mae": 0.393739, "eval_mse": 0.260811, "eval_r2": 0.834211, "eval_sp_statistic": 0.904849, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.927023, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.302963, "test_total_loss": 152.087412, "test_mae": 0.339286, "test_mse": 0.303071, "test_r2": 0.804395, "test_sp_statistic": 0.877041, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.923178, "test_ps_pvalue": 0.0, "lr": 0.0, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6211889956411308, "train_cur_epoch_loss": 357.12689004838467, "train_cur_epoch_avg_loss": 0.16774395962817504, "train_cur_epoch_time": 45.08570408821106, "train_cur_epoch_avg_time": 0.021176939449605945, "epoch": 50, "step": 106450} ################################################## #########################Best Metric######################### {"epoch": 43, "global_step": 91547, "eval_avg_loss": 0.237157, "eval_total_loss": 166.721445, "eval_mae": 0.332667, "eval_mse": 0.237219, "eval_r2": 0.849208, "eval_sp_statistic": 0.905039, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924539, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.396912, "test_total_loss": 199.249838, "test_mae": 0.384466, "test_mse": 0.397053, "test_r2": 0.743738, "test_sp_statistic": 0.87417, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.919932, "test_ps_pvalue": 0.0} ################################################## Total Time: 39251.043123, Avg time per epoch(50 epochs): 785.020000 ++++++++++++Validation+++++++++++++ best sp_statistic global step: 91547 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226035651/checkpoint-91547 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## {"evaluation_avg_loss_91547": 0.237157, "evaluation_total_loss_91547": 166.721445, "evaluation_mae_91547": 0.332667, "evaluation_mse_91547": 0.237219, "evaluation_r2_91547": 0.849208, "evaluation_sp_statistic_91547": 0.905039, "evaluation_sp_pvalue_91547": 0.0, "evaluation_ps_statistic_91547": 0.924539, "evaluation_ps_pvalue_91547": 0.0} ++++++++++++Testing+++++++++++++ best sp_statistic global step: 91547 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226035651/checkpoint-91547 ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## {"evaluation_avg_loss_91547": 0.396912, "evaluation_total_loss_91547": 199.249838, "evaluation_mae_91547": 0.384466, "evaluation_mse_91547": 0.397053, "evaluation_r2_91547": 0.743738, "evaluation_sp_statistic_91547": 0.87417, "evaluation_sp_pvalue_91547": 0.0, "evaluation_ps_statistic_91547": 0.919932, "evaluation_ps_pvalue_91547": 0.0}