{"adam_epsilon": 1e-08, "alphabet": "gene_prot", "append_eos": true, "asl_gamma_neg": 4.0, "asl_gamma_pos": 1.0, "best_metric_type": "sp_statistic", "beta1": 0.9, "beta2": 0.98, "buffer_size": 4096, "cache_dir": null, "classifier_activate_func": "gelu", "classifier_size": 128, "codes_file": null, "config_path": "../config/luca_base/luca_base_config.json", "cross_atten": false, "dataset_name": "DMS_Bind_Reps_Strain", "dataset_type": "protein", "delete_old": true, "dev_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/dev/", "device": "cuda", "do_eval": true, "do_lower_case": false, "do_metrics": true, "do_predict": true, "do_train": true, "dropout_prob": 0.1, "early_stop_epoch": -1, "emb_activate_func": "gelu", "embedding_complete": true, "embedding_complete_seg_overlap": true, "embedding_fixed_len_a_time": 3072, "embedding_input_size": 2560, "embedding_input_size_a": null, "embedding_input_size_b": null, "eval_all_checkpoints": false, "evaluate_during_training": true, "evaluate_steps": 2000, "evaluate_strategy": "epoch", "fc_activate_func": "gelu", "focal_loss_alpha": 0.7, "focal_loss_gamma": 2.0, "focal_loss_reduce": false, "fp16": false, "fp16_embedding": false, "fp16_opt_level": "O1", "fusion_type": "concat", "gradient_accumulation_steps": 1, "hidden_size": 1024, "ignore_index": -100, "input_mode": "single", "input_type": "matrix", "intermediate_size": 4096, "label_filepath": "../dataset/DMS_Bind_Reps_Strain/protein/regression/label.txt", "label_size": 1, "label_type": "DMS_Bind_Reps_Strain", "learning_rate": 0.0001, "llm_dir": "..", "llm_dirpath": null, "llm_step": "3B", "llm_task_level": "token_level,span_level,seq_level,structure_level", "llm_time_str": null, "llm_type": "esm", "llm_version": "esm2", "lmdb_path": null, "local_rank": -1, "log_dir": "../logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226150046", "logging_steps": 10, "loss_reduction": "mean", "loss_type": "l2", "lr_decay_rate": 0.9, "lr_update_strategy": "step", "matrix_add_special_token": false, "matrix_dirpath": "../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/esm2/esm//3B", "matrix_embedding_exists": false, "matrix_encoder": false, "matrix_encoder_act": false, "matrix_fc_size": "128", "matrix_max_length": 100000, "matrix_max_length_a": null, "matrix_max_length_b": null, "matrix_pooling_type": "value_attention", "max_grad_norm": 1.0, "max_sentence_length": null, "max_sentences": null, "max_steps": -1, "model_dirpath": null, "model_type": "luca_base", "n_gpu": 1, "no_cuda": false, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "non_ignore": false, "not_append_eos": false, "not_matrix_encoder_shared": false, "not_prepend_bos": false, "not_save_emb_to_disk": false, "not_seq_encoder_shared": false, "num_attention_heads": 4, "num_hidden_layers": 2, "num_train_epochs": 50, "output_dir": "../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226150046", "output_mode": "regression", "overwrite_cache": false, "overwrite_output_dir": true, "per_gpu_eval_batch_size": 16, "per_gpu_train_batch_size": 16, "pos_weight": 1.0, "position_embedding_type": "absolute", "prepend_bos": true, "save_all": false, "save_steps": -1, "seed": 1221, "self_atten": false, "seq_fc_size": "null", "seq_max_length": 100000, "seq_max_length_a": null, "seq_max_length_b": null, "seq_pooling_type": "value_attention", "seq_subword": false, "seq_vocab_path": "gene_prot", "sigmoid": false, "task_level_type": "seq_level", "task_type": "regression", "tb_log_dir": "../tb-logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226150046", "test_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/test/", "time_str": "20250226150050", "train_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/train/", "trunc_type": "right", "vector_dirpath": "../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/esm2/esm//3B", "vector_fc_size": "null", "vocab_size": 39, "warmup_steps": 1000, "weight": null, "weight_decay": 0.01, "worker_num": 0} ################################################## n_gpu: 1 ################################################## Inputs: Input Name List: protein,embedding_matrix ################################################## Encoder Config: {'llm_type': 'esm', 'llm_version': 'esm2', 'llm_step': '3B', 'llm_dirpath': None, 'input_type': 'matrix', 'trunc_type': 'right', 'seq_max_length': 100000, 'atom_seq_max_length': None, 'vector_dirpath': '../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/esm2/esm//3B', 'matrix_dirpath': '../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/esm2/esm//3B', 'local_rank': -1, 'max_sentence_length': None, 'max_sentences': None, 'matrix_add_special_token': False, 'embedding_complete': True, 'embedding_complete_seg_overlap': True, 'embedding_fixed_len_a_time': 3072, 'matrix_embedding_exists': False, 'save_emb_to_disk': True, 'fp16_embedding': False} ################################################## Model Config: LucaConfig { "alphabet": "gene_prot", "attention_probs_dropout_prob": 0.1, "classifier_activate_func": "gelu", "classifier_dropout_prob": 0.1, "classifier_size": 128, "cls_token_id": 2, "cross_atten": false, "directionality": "bidi", "emb_activate_func": "gelu", "embedding_input_size": 2560, "fc_activate_func": "gelu", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "LABEL_0" }, "ignore_index": -100, "initializer_range": 0.02, "intermediate_size": 4096, "kernel_size": 7, "label2id": { "LABEL_0": 0 }, "layer_norm_eps": 1e-12, "loss_reduction": "mean", "matrix_fc_size": [ 128 ], "matrix_max_length": 100000, "matrix_pooling_type": "value_attention", "max_position_embeddings": 100002, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "num_attention_heads": 4, "num_hidden_layers": 2, "pad_token_id": 0, "pos_weight": 1.0, "position_embedding_type": "absolute", "self_atten": false, "sep_token_id": 3, "seq_fc_size": null, "seq_max_length": 100000, "seq_pooling_type": "value_attention", "token_dropout": null, "transformers_version": "4.29.0", "type_vocab_size": 2, "use_luca_layer_norm_v2": true, "vector_fc_size": null, "vocab_size": 39 } ################################################## Mode Architecture: LucaBase( (matrix_pooler): GlobalMaskValueAttentionPooling1D (2560 -> 2560) (linear): ModuleList( (0): ModuleList( (0): Linear(in_features=2560, out_features=128, bias=True) (1): GELU(approximate='none') ) ) (dropout): Dropout(p=0.1, inplace=False) (hidden_layer): Linear(in_features=128, out_features=128, bias=True) (hidden_act): GELU(approximate='none') (classifier): Linear(in_features=128, out_features=1, bias=True) (loss_fct): MaskedMSELoss( (criterion): MSELoss() ) ) ################################################## Model parameters: 20005249 ################################################## {"total_num": "19.080000M", "total_size": "76.310000MB", "param_sum": "19.080000M", "param_size": "76.310000MB", "buffer_sum": "0.000000M", "buffer_size": "0.000000MB", "trainable_num": "19.078492M", "trainable_size": "76.313969MB"} ################################################## Train dataset len: 34055, batch size: 16, batch num: 2129 Train dataset t_total: 106450, max_steps: -1 ***** Running training ***** Train Dataset Num examples = 34055 Train Dataset Num Epochs = 50 Logging Steps = 10 Saving Steps = -1 Evaluating Strategy = epoch Train Dataset Instantaneous batch size per GPU = 16 Train Dataset Total train batch size (w. parallel, distributed & accumulation) = 16 Train Dataset Gradient Accumulation steps = 1 Train Dataset Total optimization steps = 106450 ################################################## Training, Epoch: 0001, Batch: 000010, Sample Num: 160, Cur Loss: 70.83182526, Cur Avg Loss: 71.88814850, Log Avg loss: 71.88814850, Global Avg Loss: 71.88814850, Time: 0.0210 Steps: 10, Updated lr: 0.000001 Training, Epoch: 0001, Batch: 000020, Sample Num: 320, Cur Loss: 73.29515076, Cur Avg Loss: 71.36273308, Log Avg loss: 70.83731766, Global Avg Loss: 71.36273308, Time: 0.0210 Steps: 20, Updated lr: 0.000002 Training, Epoch: 0001, Batch: 000030, Sample Num: 480, Cur Loss: 72.96173096, Cur Avg Loss: 72.67614136, Log Avg loss: 75.30295792, Global Avg Loss: 72.67614136, Time: 0.0210 Steps: 30, Updated lr: 0.000003 Training, Epoch: 0001, Batch: 000040, Sample Num: 640, Cur Loss: 67.66802216, Cur Avg Loss: 72.15621729, Log Avg loss: 70.59644508, Global Avg Loss: 72.15621729, Time: 0.0210 Steps: 40, Updated lr: 0.000004 Training, Epoch: 0001, Batch: 000050, Sample Num: 800, Cur Loss: 63.16626740, Cur Avg Loss: 71.96011192, Log Avg loss: 71.17569046, Global Avg Loss: 71.96011192, Time: 0.0210 Steps: 50, Updated lr: 0.000005 Training, Epoch: 0001, Batch: 000060, Sample Num: 960, Cur Loss: 73.95346832, Cur Avg Loss: 71.89290282, Log Avg loss: 71.55685730, Global Avg Loss: 71.89290282, Time: 0.0210 Steps: 60, Updated lr: 0.000006 Training, Epoch: 0001, Batch: 000070, Sample Num: 1120, Cur Loss: 63.14691162, Cur Avg Loss: 71.55490848, Log Avg loss: 69.52694244, Global Avg Loss: 71.55490848, Time: 0.0209 Steps: 70, Updated lr: 0.000007 Training, Epoch: 0001, Batch: 000080, Sample Num: 1280, Cur Loss: 74.75345612, Cur Avg Loss: 71.16604280, Log Avg loss: 68.44398308, Global Avg Loss: 71.16604280, Time: 0.0211 Steps: 80, Updated lr: 0.000008 Training, Epoch: 0001, Batch: 000090, Sample Num: 1440, Cur Loss: 73.46148682, Cur Avg Loss: 70.81411463, Log Avg loss: 67.99868927, Global Avg Loss: 70.81411463, Time: 0.0210 Steps: 90, Updated lr: 0.000009 Training, Epoch: 0001, Batch: 000100, Sample Num: 1600, Cur Loss: 76.83807373, Cur Avg Loss: 71.03263145, Log Avg loss: 72.99928284, Global Avg Loss: 71.03263145, Time: 0.0211 Steps: 100, Updated lr: 0.000010 Training, Epoch: 0001, Batch: 000110, Sample Num: 1760, Cur Loss: 67.88798523, Cur Avg Loss: 71.03853167, Log Avg loss: 71.09753380, Global Avg Loss: 71.03853167, Time: 0.0209 Steps: 110, Updated lr: 0.000011 Training, Epoch: 0001, Batch: 000120, Sample Num: 1920, Cur Loss: 71.27017212, Cur Avg Loss: 71.17041260, Log Avg loss: 72.62110291, Global Avg Loss: 71.17041260, Time: 0.0210 Steps: 120, Updated lr: 0.000012 Training, Epoch: 0001, Batch: 000130, Sample Num: 2080, Cur Loss: 77.23480225, Cur Avg Loss: 71.30109943, Log Avg loss: 72.86934128, Global Avg Loss: 71.30109943, Time: 0.0209 Steps: 130, Updated lr: 0.000013 Training, Epoch: 0001, Batch: 000140, Sample Num: 2240, Cur Loss: 77.06033325, Cur Avg Loss: 71.40593477, Log Avg loss: 72.76879425, Global Avg Loss: 71.40593477, Time: 0.0211 Steps: 140, Updated lr: 0.000014 Training, Epoch: 0001, Batch: 000150, Sample Num: 2400, Cur Loss: 72.82645416, Cur Avg Loss: 71.64572685, Log Avg loss: 75.00281601, Global Avg Loss: 71.64572685, Time: 0.0210 Steps: 150, Updated lr: 0.000015 Training, Epoch: 0001, Batch: 000160, Sample Num: 2560, Cur Loss: 64.93380737, Cur Avg Loss: 71.52272124, Log Avg loss: 69.67763710, Global Avg Loss: 71.52272124, Time: 0.0210 Steps: 160, Updated lr: 0.000016 Training, Epoch: 0001, Batch: 000170, Sample Num: 2720, Cur Loss: 70.47185516, Cur Avg Loss: 71.39727438, Log Avg loss: 69.39012451, Global Avg Loss: 71.39727438, Time: 0.0210 Steps: 170, Updated lr: 0.000017 Training, Epoch: 0001, Batch: 000180, Sample Num: 2880, Cur Loss: 62.07156754, Cur Avg Loss: 71.34873303, Log Avg loss: 70.52353020, Global Avg Loss: 71.34873303, Time: 0.0210 Steps: 180, Updated lr: 0.000018 Training, Epoch: 0001, Batch: 000190, Sample Num: 3040, Cur Loss: 77.12212372, Cur Avg Loss: 71.37462606, Log Avg loss: 71.84070053, Global Avg Loss: 71.37462606, Time: 0.0210 Steps: 190, Updated lr: 0.000019 Training, Epoch: 0001, Batch: 000200, Sample Num: 3200, Cur Loss: 64.15548706, Cur Avg Loss: 71.34742399, Log Avg loss: 70.83058472, Global Avg Loss: 71.34742399, Time: 0.0210 Steps: 200, Updated lr: 0.000020 Training, Epoch: 0001, Batch: 000210, Sample Num: 3360, Cur Loss: 67.59258270, Cur Avg Loss: 71.24992140, Log Avg loss: 69.29986954, Global Avg Loss: 71.24992140, Time: 0.0210 Steps: 210, Updated lr: 0.000021 Training, Epoch: 0001, Batch: 000220, Sample Num: 3520, Cur Loss: 62.28026581, Cur Avg Loss: 71.10127498, Log Avg loss: 67.97970009, Global Avg Loss: 71.10127498, Time: 0.0210 Steps: 220, Updated lr: 0.000022 Training, Epoch: 0001, Batch: 000230, Sample Num: 3680, Cur Loss: 75.54181671, Cur Avg Loss: 70.90950113, Log Avg loss: 66.69047661, Global Avg Loss: 70.90950113, Time: 0.0211 Steps: 230, Updated lr: 0.000023 Training, Epoch: 0001, Batch: 000240, Sample Num: 3840, Cur Loss: 68.42121124, Cur Avg Loss: 70.91749908, Log Avg loss: 71.10145187, Global Avg Loss: 70.91749908, Time: 0.0211 Steps: 240, Updated lr: 0.000024 Training, Epoch: 0001, Batch: 000250, Sample Num: 4000, Cur Loss: 69.23279572, Cur Avg Loss: 70.79500771, Log Avg loss: 67.85521469, Global Avg Loss: 70.79500771, Time: 0.0210 Steps: 250, Updated lr: 0.000025 Training, Epoch: 0001, Batch: 000260, Sample Num: 4160, Cur Loss: 69.60492706, Cur Avg Loss: 70.63525398, Log Avg loss: 66.64141083, Global Avg Loss: 70.63525398, Time: 0.0210 Steps: 260, Updated lr: 0.000026 Training, Epoch: 0001, Batch: 000270, Sample Num: 4320, Cur Loss: 55.68511200, Cur Avg Loss: 70.39611486, Log Avg loss: 64.17849770, Global Avg Loss: 70.39611486, Time: 0.0210 Steps: 270, Updated lr: 0.000027 Training, Epoch: 0001, Batch: 000280, Sample Num: 4480, Cur Loss: 60.91697693, Cur Avg Loss: 70.18828295, Log Avg loss: 64.57682152, Global Avg Loss: 70.18828295, Time: 0.0210 Steps: 280, Updated lr: 0.000028 Training, Epoch: 0001, Batch: 000290, Sample Num: 4640, Cur Loss: 60.05079651, Cur Avg Loss: 69.95229413, Log Avg loss: 63.34460716, Global Avg Loss: 69.95229413, Time: 0.0210 Steps: 290, Updated lr: 0.000029 Training, Epoch: 0001, Batch: 000300, Sample Num: 4800, Cur Loss: 62.08605957, Cur Avg Loss: 69.65074586, Log Avg loss: 60.90584602, Global Avg Loss: 69.65074586, Time: 0.0210 Steps: 300, Updated lr: 0.000030 Training, Epoch: 0001, Batch: 000310, Sample Num: 4960, Cur Loss: 52.25639343, Cur Avg Loss: 69.26315900, Log Avg loss: 57.63555298, Global Avg Loss: 69.26315900, Time: 0.0210 Steps: 310, Updated lr: 0.000031 Training, Epoch: 0001, Batch: 000320, Sample Num: 5120, Cur Loss: 59.42348480, Cur Avg Loss: 68.84558003, Log Avg loss: 55.90063210, Global Avg Loss: 68.84558003, Time: 0.0210 Steps: 320, Updated lr: 0.000032 Training, Epoch: 0001, Batch: 000330, Sample Num: 5280, Cur Loss: 54.19661331, Cur Avg Loss: 68.33296684, Log Avg loss: 51.92934494, Global Avg Loss: 68.33296684, Time: 0.0210 Steps: 330, Updated lr: 0.000033 Training, Epoch: 0001, Batch: 000340, Sample Num: 5440, Cur Loss: 48.19434357, Cur Avg Loss: 67.80202326, Log Avg loss: 50.28088493, Global Avg Loss: 67.80202326, Time: 0.0210 Steps: 340, Updated lr: 0.000034 Training, Epoch: 0001, Batch: 000350, Sample Num: 5600, Cur Loss: 56.87677383, Cur Avg Loss: 67.32166381, Log Avg loss: 50.98944244, Global Avg Loss: 67.32166381, Time: 0.0210 Steps: 350, Updated lr: 0.000035 Training, Epoch: 0001, Batch: 000360, Sample Num: 5760, Cur Loss: 42.95371246, Cur Avg Loss: 66.74583184, Log Avg loss: 46.59171295, Global Avg Loss: 66.74583184, Time: 0.0210 Steps: 360, Updated lr: 0.000036 Training, Epoch: 0001, Batch: 000370, Sample Num: 5920, Cur Loss: 39.00234985, Cur Avg Loss: 66.05900135, Log Avg loss: 41.33310356, Global Avg Loss: 66.05900135, Time: 0.0210 Steps: 370, Updated lr: 0.000037 Training, Epoch: 0001, Batch: 000380, Sample Num: 6080, Cur Loss: 37.30493164, Cur Avg Loss: 65.33931286, Log Avg loss: 38.71083908, Global Avg Loss: 65.33931286, Time: 0.0210 Steps: 380, Updated lr: 0.000038 Training, Epoch: 0001, Batch: 000390, Sample Num: 6240, Cur Loss: 27.52929115, Cur Avg Loss: 64.56718815, Log Avg loss: 35.22644882, Global Avg Loss: 64.56718815, Time: 0.0210 Steps: 390, Updated lr: 0.000039 Training, Epoch: 0001, Batch: 000400, Sample Num: 6400, Cur Loss: 30.84651184, Cur Avg Loss: 63.75458109, Log Avg loss: 32.06290607, Global Avg Loss: 63.75458109, Time: 0.0209 Steps: 400, Updated lr: 0.000040 Training, Epoch: 0001, Batch: 000410, Sample Num: 6560, Cur Loss: 29.41937637, Cur Avg Loss: 62.91305193, Log Avg loss: 29.25188541, Global Avg Loss: 62.91305193, Time: 0.0210 Steps: 410, Updated lr: 0.000041 Training, Epoch: 0001, Batch: 000420, Sample Num: 6720, Cur Loss: 23.47639847, Cur Avg Loss: 62.01414581, Log Avg loss: 25.15899467, Global Avg Loss: 62.01414581, Time: 0.0210 Steps: 420, Updated lr: 0.000042 Training, Epoch: 0001, Batch: 000430, Sample Num: 6880, Cur Loss: 16.76692581, Cur Avg Loss: 61.05057787, Log Avg loss: 20.58072472, Global Avg Loss: 61.05057787, Time: 0.0210 Steps: 430, Updated lr: 0.000043 Training, Epoch: 0001, Batch: 000440, Sample Num: 7040, Cur Loss: 18.13956833, Cur Avg Loss: 60.04551503, Log Avg loss: 16.82781277, Global Avg Loss: 60.04551503, Time: 0.0210 Steps: 440, Updated lr: 0.000044 Training, Epoch: 0001, Batch: 000450, Sample Num: 7200, Cur Loss: 10.00717926, Cur Avg Loss: 59.00427115, Log Avg loss: 13.18954048, Global Avg Loss: 59.00427115, Time: 0.0210 Steps: 450, Updated lr: 0.000045 Training, Epoch: 0001, Batch: 000460, Sample Num: 7360, Cur Loss: 8.74394035, Cur Avg Loss: 57.93444274, Log Avg loss: 9.79216418, Global Avg Loss: 57.93444274, Time: 0.0209 Steps: 460, Updated lr: 0.000046 Training, Epoch: 0001, Batch: 000470, Sample Num: 7520, Cur Loss: 7.26677942, Cur Avg Loss: 56.85008038, Log Avg loss: 6.96941175, Global Avg Loss: 56.85008038, Time: 0.0210 Steps: 470, Updated lr: 0.000047 Training, Epoch: 0001, Batch: 000480, Sample Num: 7680, Cur Loss: 2.27948022, Cur Avg Loss: 55.75175220, Log Avg loss: 4.13032770, Global Avg Loss: 55.75175220, Time: 0.0210 Steps: 480, Updated lr: 0.000048 Training, Epoch: 0001, Batch: 000490, Sample Num: 7840, Cur Loss: 1.54005861, Cur Avg Loss: 54.66314753, Log Avg loss: 2.41012336, Global Avg Loss: 54.66314753, Time: 0.0209 Steps: 490, Updated lr: 0.000049 Training, Epoch: 0001, Batch: 000500, Sample Num: 8000, Cur Loss: 1.05100238, Cur Avg Loss: 53.60530212, Log Avg loss: 1.77087733, Global Avg Loss: 53.60530212, Time: 0.0210 Steps: 500, Updated lr: 0.000050 Training, Epoch: 0001, Batch: 000510, Sample Num: 8160, Cur Loss: 1.85381317, Cur Avg Loss: 52.60066300, Log Avg loss: 2.36870682, Global Avg Loss: 52.60066300, Time: 0.0210 Steps: 510, Updated lr: 0.000051 Training, Epoch: 0001, Batch: 000520, Sample Num: 8320, Cur Loss: 2.24227118, Cur Avg Loss: 51.62137356, Log Avg loss: 1.67761227, Global Avg Loss: 51.62137356, Time: 0.0209 Steps: 520, Updated lr: 0.000052 Training, Epoch: 0001, Batch: 000530, Sample Num: 8480, Cur Loss: 1.10324764, Cur Avg Loss: 50.67881322, Log Avg loss: 1.66567558, Global Avg Loss: 50.67881322, Time: 0.0210 Steps: 530, Updated lr: 0.000053 Training, Epoch: 0001, Batch: 000540, Sample Num: 8640, Cur Loss: 1.27996504, Cur Avg Loss: 49.76972402, Log Avg loss: 1.58799641, Global Avg Loss: 49.76972402, Time: 0.0210 Steps: 540, Updated lr: 0.000054 Training, Epoch: 0001, Batch: 000550, Sample Num: 8800, Cur Loss: 1.74432111, Cur Avg Loss: 48.90317326, Log Avg loss: 2.10943197, Global Avg Loss: 48.90317326, Time: 0.0210 Steps: 550, Updated lr: 0.000055 Training, Epoch: 0001, Batch: 000560, Sample Num: 8960, Cur Loss: 1.02943647, Cur Avg Loss: 48.05591883, Log Avg loss: 1.45692522, Global Avg Loss: 48.05591883, Time: 0.0209 Steps: 560, Updated lr: 0.000056 Training, Epoch: 0001, Batch: 000570, Sample Num: 9120, Cur Loss: 1.22700882, Cur Avg Loss: 47.24628630, Log Avg loss: 1.90686459, Global Avg Loss: 47.24628630, Time: 0.0209 Steps: 570, Updated lr: 0.000057 Training, Epoch: 0001, Batch: 000580, Sample Num: 9280, Cur Loss: 1.10266507, Cur Avg Loss: 46.45870293, Log Avg loss: 1.56645068, Global Avg Loss: 46.45870293, Time: 0.0209 Steps: 580, Updated lr: 0.000058 Training, Epoch: 0001, Batch: 000590, Sample Num: 9440, Cur Loss: 0.68196356, Cur Avg Loss: 45.69827554, Log Avg loss: 1.59348744, Global Avg Loss: 45.69827554, Time: 0.0209 Steps: 590, Updated lr: 0.000059 Training, Epoch: 0001, Batch: 000600, Sample Num: 9600, Cur Loss: 2.81867695, Cur Avg Loss: 44.96829912, Log Avg loss: 1.89969033, Global Avg Loss: 44.96829912, Time: 0.0210 Steps: 600, Updated lr: 0.000060 Training, Epoch: 0001, Batch: 000610, Sample Num: 9760, Cur Loss: 1.21191061, Cur Avg Loss: 44.25486587, Log Avg loss: 1.44887059, Global Avg Loss: 44.25486587, Time: 0.0209 Steps: 610, Updated lr: 0.000061 Training, Epoch: 0001, Batch: 000620, Sample Num: 9920, Cur Loss: 0.77601445, Cur Avg Loss: 43.56745090, Log Avg loss: 1.63513798, Global Avg Loss: 43.56745090, Time: 0.0210 Steps: 620, Updated lr: 0.000062 Training, Epoch: 0001, Batch: 000630, Sample Num: 10080, Cur Loss: 0.90588331, Cur Avg Loss: 42.90011235, Log Avg loss: 1.52512219, Global Avg Loss: 42.90011235, Time: 0.0210 Steps: 630, Updated lr: 0.000063 Training, Epoch: 0001, Batch: 000640, Sample Num: 10240, Cur Loss: 1.87736857, Cur Avg Loss: 42.25466780, Log Avg loss: 1.59166117, Global Avg Loss: 42.25466780, Time: 0.0209 Steps: 640, Updated lr: 0.000064 Training, Epoch: 0001, Batch: 000650, Sample Num: 10400, Cur Loss: 1.74336076, Cur Avg Loss: 41.63389305, Log Avg loss: 1.90430911, Global Avg Loss: 41.63389305, Time: 0.0209 Steps: 650, Updated lr: 0.000065 Training, Epoch: 0001, Batch: 000660, Sample Num: 10560, Cur Loss: 1.27831054, Cur Avg Loss: 41.02943023, Log Avg loss: 1.73934659, Global Avg Loss: 41.02943023, Time: 0.0209 Steps: 660, Updated lr: 0.000066 Training, Epoch: 0001, Batch: 000670, Sample Num: 10720, Cur Loss: 1.08312559, Cur Avg Loss: 40.43678272, Log Avg loss: 1.32204686, Global Avg Loss: 40.43678272, Time: 0.0209 Steps: 670, Updated lr: 0.000067 Training, Epoch: 0001, Batch: 000680, Sample Num: 10880, Cur Loss: 1.40617716, Cur Avg Loss: 39.86661494, Log Avg loss: 1.66537387, Global Avg Loss: 39.86661494, Time: 0.0210 Steps: 680, Updated lr: 0.000068 Training, Epoch: 0001, Batch: 000690, Sample Num: 11040, Cur Loss: 0.70912164, Cur Avg Loss: 39.31159383, Log Avg loss: 1.57015818, Global Avg Loss: 39.31159383, Time: 0.0210 Steps: 690, Updated lr: 0.000069 Training, Epoch: 0001, Batch: 000700, Sample Num: 11200, Cur Loss: 1.68472290, Cur Avg Loss: 38.77532499, Log Avg loss: 1.77277530, Global Avg Loss: 38.77532499, Time: 0.0210 Steps: 700, Updated lr: 0.000070 Training, Epoch: 0001, Batch: 000710, Sample Num: 11360, Cur Loss: 1.81379843, Cur Avg Loss: 38.25587082, Log Avg loss: 1.89407921, Global Avg Loss: 38.25587082, Time: 0.0210 Steps: 710, Updated lr: 0.000071 Training, Epoch: 0001, Batch: 000720, Sample Num: 11520, Cur Loss: 1.40473485, Cur Avg Loss: 37.74373166, Log Avg loss: 1.38185070, Global Avg Loss: 37.74373166, Time: 0.0210 Steps: 720, Updated lr: 0.000072 Training, Epoch: 0001, Batch: 000730, Sample Num: 11680, Cur Loss: 2.56999969, Cur Avg Loss: 37.24872428, Log Avg loss: 1.60819322, Global Avg Loss: 37.24872428, Time: 0.0210 Steps: 730, Updated lr: 0.000073 Training, Epoch: 0001, Batch: 000740, Sample Num: 11840, Cur Loss: 1.55242896, Cur Avg Loss: 36.77163876, Log Avg loss: 1.94439559, Global Avg Loss: 36.77163876, Time: 0.0210 Steps: 740, Updated lr: 0.000074 Training, Epoch: 0001, Batch: 000750, Sample Num: 12000, Cur Loss: 2.92001748, Cur Avg Loss: 36.30582506, Log Avg loss: 1.83561129, Global Avg Loss: 36.30582506, Time: 0.0210 Steps: 750, Updated lr: 0.000075 Training, Epoch: 0001, Batch: 000760, Sample Num: 12160, Cur Loss: 1.13406181, Cur Avg Loss: 35.85266795, Log Avg loss: 1.86588480, Global Avg Loss: 35.85266795, Time: 0.0210 Steps: 760, Updated lr: 0.000076 Training, Epoch: 0001, Batch: 000770, Sample Num: 12320, Cur Loss: 1.61097693, Cur Avg Loss: 35.41222871, Log Avg loss: 1.93884658, Global Avg Loss: 35.41222871, Time: 0.0211 Steps: 770, Updated lr: 0.000077 Training, Epoch: 0001, Batch: 000780, Sample Num: 12480, Cur Loss: 1.44898009, Cur Avg Loss: 34.98119896, Log Avg loss: 1.79190819, Global Avg Loss: 34.98119896, Time: 0.0210 Steps: 780, Updated lr: 0.000078 Training, Epoch: 0001, Batch: 000790, Sample Num: 12640, Cur Loss: 1.02076387, Cur Avg Loss: 34.55815666, Log Avg loss: 1.56085745, Global Avg Loss: 34.55815666, Time: 0.0210 Steps: 790, Updated lr: 0.000079 Training, Epoch: 0001, Batch: 000800, Sample Num: 12800, Cur Loss: 1.84853089, Cur Avg Loss: 34.14747655, Log Avg loss: 1.70374739, Global Avg Loss: 34.14747655, Time: 0.0210 Steps: 800, Updated lr: 0.000080 Training, Epoch: 0001, Batch: 000810, Sample Num: 12960, Cur Loss: 2.33741999, Cur Avg Loss: 33.74662583, Log Avg loss: 1.67856839, Global Avg Loss: 33.74662583, Time: 0.0210 Steps: 810, Updated lr: 0.000081 Training, Epoch: 0001, Batch: 000820, Sample Num: 13120, Cur Loss: 1.26982379, Cur Avg Loss: 33.35738365, Log Avg loss: 1.82876759, Global Avg Loss: 33.35738365, Time: 0.0210 Steps: 820, Updated lr: 0.000082 Training, Epoch: 0001, Batch: 000830, Sample Num: 13280, Cur Loss: 1.03937697, Cur Avg Loss: 32.97634941, Log Avg loss: 1.73154164, Global Avg Loss: 32.97634941, Time: 0.0210 Steps: 830, Updated lr: 0.000083 Training, Epoch: 0001, Batch: 000840, Sample Num: 13440, Cur Loss: 0.79085100, Cur Avg Loss: 32.60238803, Log Avg loss: 1.56359292, Global Avg Loss: 32.60238803, Time: 0.0210 Steps: 840, Updated lr: 0.000084 Training, Epoch: 0001, Batch: 000850, Sample Num: 13600, Cur Loss: 2.71126676, Cur Avg Loss: 32.23848401, Log Avg loss: 1.67054701, Global Avg Loss: 32.23848401, Time: 0.0210 Steps: 850, Updated lr: 0.000085 Training, Epoch: 0001, Batch: 000860, Sample Num: 13760, Cur Loss: 1.57688951, Cur Avg Loss: 31.88034689, Log Avg loss: 1.43869116, Global Avg Loss: 31.88034689, Time: 0.0210 Steps: 860, Updated lr: 0.000086 Training, Epoch: 0001, Batch: 000870, Sample Num: 13920, Cur Loss: 1.07007527, Cur Avg Loss: 31.53182907, Log Avg loss: 1.55929679, Global Avg Loss: 31.53182907, Time: 0.0210 Steps: 870, Updated lr: 0.000087 Training, Epoch: 0001, Batch: 000880, Sample Num: 14080, Cur Loss: 0.77336657, Cur Avg Loss: 31.19155483, Log Avg loss: 1.58769605, Global Avg Loss: 31.19155483, Time: 0.0209 Steps: 880, Updated lr: 0.000088 Training, Epoch: 0001, Batch: 000890, Sample Num: 14240, Cur Loss: 1.16406405, Cur Avg Loss: 30.86166083, Log Avg loss: 1.83098902, Global Avg Loss: 30.86166083, Time: 0.0210 Steps: 890, Updated lr: 0.000089 Training, Epoch: 0001, Batch: 000900, Sample Num: 14400, Cur Loss: 1.28475833, Cur Avg Loss: 30.53600510, Log Avg loss: 1.55264472, Global Avg Loss: 30.53600510, Time: 0.0210 Steps: 900, Updated lr: 0.000090 Training, Epoch: 0001, Batch: 000910, Sample Num: 14560, Cur Loss: 2.22389603, Cur Avg Loss: 30.22044299, Log Avg loss: 1.81985302, Global Avg Loss: 30.22044299, Time: 0.0210 Steps: 910, Updated lr: 0.000091 Training, Epoch: 0001, Batch: 000920, Sample Num: 14720, Cur Loss: 1.87127542, Cur Avg Loss: 29.91093619, Log Avg loss: 1.74581760, Global Avg Loss: 29.91093619, Time: 0.0210 Steps: 920, Updated lr: 0.000092 Training, Epoch: 0001, Batch: 000930, Sample Num: 14880, Cur Loss: 2.76380396, Cur Avg Loss: 29.60747759, Log Avg loss: 1.68928593, Global Avg Loss: 29.60747759, Time: 0.0210 Steps: 930, Updated lr: 0.000093 Training, Epoch: 0001, Batch: 000940, Sample Num: 15040, Cur Loss: 2.87947178, Cur Avg Loss: 29.31222588, Log Avg loss: 1.85381733, Global Avg Loss: 29.31222588, Time: 0.0210 Steps: 940, Updated lr: 0.000094 Training, Epoch: 0001, Batch: 000950, Sample Num: 15200, Cur Loss: 2.29429102, Cur Avg Loss: 29.02215162, Log Avg loss: 1.75517147, Global Avg Loss: 29.02215162, Time: 0.0210 Steps: 950, Updated lr: 0.000095 Training, Epoch: 0001, Batch: 000960, Sample Num: 15360, Cur Loss: 2.85970783, Cur Avg Loss: 28.73922270, Log Avg loss: 1.86097527, Global Avg Loss: 28.73922270, Time: 0.0210 Steps: 960, Updated lr: 0.000096 Training, Epoch: 0001, Batch: 000970, Sample Num: 15520, Cur Loss: 1.36245596, Cur Avg Loss: 28.46249019, Log Avg loss: 1.89616929, Global Avg Loss: 28.46249019, Time: 0.0210 Steps: 970, Updated lr: 0.000097 Training, Epoch: 0001, Batch: 000980, Sample Num: 15680, Cur Loss: 1.46349847, Cur Avg Loss: 28.19059149, Log Avg loss: 1.81641709, Global Avg Loss: 28.19059149, Time: 0.0210 Steps: 980, Updated lr: 0.000098 Training, Epoch: 0001, Batch: 000990, Sample Num: 15840, Cur Loss: 2.30466247, Cur Avg Loss: 27.92411091, Log Avg loss: 1.80901396, Global Avg Loss: 27.92411091, Time: 0.0210 Steps: 990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001000, Sample Num: 16000, Cur Loss: 1.16339493, Cur Avg Loss: 27.66101389, Log Avg loss: 1.61440888, Global Avg Loss: 27.66101389, Time: 0.0210 Steps: 1000, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001010, Sample Num: 16160, Cur Loss: 1.06458127, Cur Avg Loss: 27.40290535, Log Avg loss: 1.59205183, Global Avg Loss: 27.40290535, Time: 0.0210 Steps: 1010, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001020, Sample Num: 16320, Cur Loss: 1.96046257, Cur Avg Loss: 27.14949636, Log Avg loss: 1.55518799, Global Avg Loss: 27.14949636, Time: 0.0210 Steps: 1020, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001030, Sample Num: 16480, Cur Loss: 1.76231837, Cur Avg Loss: 26.89785717, Log Avg loss: 1.23066021, Global Avg Loss: 26.89785717, Time: 0.0210 Steps: 1030, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001040, Sample Num: 16640, Cur Loss: 1.53354168, Cur Avg Loss: 26.65238225, Log Avg loss: 1.36846559, Global Avg Loss: 26.65238225, Time: 0.0210 Steps: 1040, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001050, Sample Num: 16800, Cur Loss: 1.10918570, Cur Avg Loss: 26.41549951, Log Avg loss: 1.77969449, Global Avg Loss: 26.41549951, Time: 0.0210 Steps: 1050, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001060, Sample Num: 16960, Cur Loss: 1.28345585, Cur Avg Loss: 26.18139552, Log Avg loss: 1.60047655, Global Avg Loss: 26.18139552, Time: 0.0210 Steps: 1060, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001070, Sample Num: 17120, Cur Loss: 3.80274868, Cur Avg Loss: 25.95769316, Log Avg loss: 2.24524218, Global Avg Loss: 25.95769316, Time: 0.0210 Steps: 1070, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001080, Sample Num: 17280, Cur Loss: 2.65251637, Cur Avg Loss: 25.73460587, Log Avg loss: 1.86426643, Global Avg Loss: 25.73460587, Time: 0.0210 Steps: 1080, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001090, Sample Num: 17440, Cur Loss: 2.45821452, Cur Avg Loss: 25.51458480, Log Avg loss: 1.75230923, Global Avg Loss: 25.51458480, Time: 0.0210 Steps: 1090, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001100, Sample Num: 17600, Cur Loss: 2.12222481, Cur Avg Loss: 25.29901327, Log Avg loss: 1.80171601, Global Avg Loss: 25.29901327, Time: 0.0210 Steps: 1100, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001110, Sample Num: 17760, Cur Loss: 1.17870975, Cur Avg Loss: 25.08457294, Log Avg loss: 1.49613675, Global Avg Loss: 25.08457294, Time: 0.0210 Steps: 1110, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001120, Sample Num: 17920, Cur Loss: 0.99946064, Cur Avg Loss: 24.87625698, Log Avg loss: 1.75318602, Global Avg Loss: 24.87625698, Time: 0.0210 Steps: 1120, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001130, Sample Num: 18080, Cur Loss: 1.14533377, Cur Avg Loss: 24.67029155, Log Avg loss: 1.60216269, Global Avg Loss: 24.67029155, Time: 0.0210 Steps: 1130, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001140, Sample Num: 18240, Cur Loss: 2.78067827, Cur Avg Loss: 24.46713535, Log Avg loss: 1.51048558, Global Avg Loss: 24.46713535, Time: 0.0210 Steps: 1140, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001150, Sample Num: 18400, Cur Loss: 1.50216722, Cur Avg Loss: 24.27131336, Log Avg loss: 1.94760658, Global Avg Loss: 24.27131336, Time: 0.0210 Steps: 1150, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001160, Sample Num: 18560, Cur Loss: 2.02894783, Cur Avg Loss: 24.07679454, Log Avg loss: 1.70712999, Global Avg Loss: 24.07679454, Time: 0.0210 Steps: 1160, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001170, Sample Num: 18720, Cur Loss: 1.59414935, Cur Avg Loss: 23.88277703, Log Avg loss: 1.37674586, Global Avg Loss: 23.88277703, Time: 0.0210 Steps: 1170, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001180, Sample Num: 18880, Cur Loss: 0.55858243, Cur Avg Loss: 23.69763631, Log Avg loss: 2.03617216, Global Avg Loss: 23.69763631, Time: 0.0211 Steps: 1180, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001190, Sample Num: 19040, Cur Loss: 1.14435887, Cur Avg Loss: 23.51005315, Log Avg loss: 1.37524032, Global Avg Loss: 23.51005315, Time: 0.0210 Steps: 1190, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001200, Sample Num: 19200, Cur Loss: 1.10517013, Cur Avg Loss: 23.32634552, Log Avg loss: 1.46513690, Global Avg Loss: 23.32634552, Time: 0.0210 Steps: 1200, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001210, Sample Num: 19360, Cur Loss: 0.99541759, Cur Avg Loss: 23.14636143, Log Avg loss: 1.54827110, Global Avg Loss: 23.14636143, Time: 0.0210 Steps: 1210, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001220, Sample Num: 19520, Cur Loss: 0.97125715, Cur Avg Loss: 22.96865657, Log Avg loss: 1.46636822, Global Avg Loss: 22.96865657, Time: 0.0211 Steps: 1220, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001230, Sample Num: 19680, Cur Loss: 1.90605235, Cur Avg Loss: 22.79539790, Log Avg loss: 1.65784079, Global Avg Loss: 22.79539790, Time: 0.0210 Steps: 1230, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001240, Sample Num: 19840, Cur Loss: 1.60271454, Cur Avg Loss: 22.62531720, Log Avg loss: 1.70539102, Global Avg Loss: 22.62531720, Time: 0.0210 Steps: 1240, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001250, Sample Num: 20000, Cur Loss: 1.67121255, Cur Avg Loss: 22.45454811, Log Avg loss: 1.27918101, Global Avg Loss: 22.45454811, Time: 0.0210 Steps: 1250, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001260, Sample Num: 20160, Cur Loss: 0.77008688, Cur Avg Loss: 22.28460353, Log Avg loss: 1.04153072, Global Avg Loss: 22.28460353, Time: 0.0210 Steps: 1260, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001270, Sample Num: 20320, Cur Loss: 1.74612641, Cur Avg Loss: 22.12048186, Log Avg loss: 1.44115068, Global Avg Loss: 22.12048186, Time: 0.0210 Steps: 1270, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001280, Sample Num: 20480, Cur Loss: 1.24409950, Cur Avg Loss: 21.96289609, Log Avg loss: 1.94950358, Global Avg Loss: 21.96289609, Time: 0.0217 Steps: 1280, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001290, Sample Num: 20640, Cur Loss: 0.84802151, Cur Avg Loss: 21.80319969, Log Avg loss: 1.36206089, Global Avg Loss: 21.80319969, Time: 0.0210 Steps: 1290, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001300, Sample Num: 20800, Cur Loss: 2.67645073, Cur Avg Loss: 21.64686775, Log Avg loss: 1.48004699, Global Avg Loss: 21.64686775, Time: 0.0210 Steps: 1300, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001310, Sample Num: 20960, Cur Loss: 1.22471058, Cur Avg Loss: 21.49081201, Log Avg loss: 1.20356569, Global Avg Loss: 21.49081201, Time: 0.0209 Steps: 1310, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001320, Sample Num: 21120, Cur Loss: 1.23990417, Cur Avg Loss: 21.34004298, Log Avg loss: 1.58930082, Global Avg Loss: 21.34004298, Time: 0.0210 Steps: 1320, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001330, Sample Num: 21280, Cur Loss: 0.44111609, Cur Avg Loss: 21.19188446, Log Avg loss: 1.63495909, Global Avg Loss: 21.19188446, Time: 0.0209 Steps: 1330, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001340, Sample Num: 21440, Cur Loss: 1.32649732, Cur Avg Loss: 21.04265779, Log Avg loss: 1.19551079, Global Avg Loss: 21.04265779, Time: 0.0210 Steps: 1340, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001350, Sample Num: 21600, Cur Loss: 1.32362056, Cur Avg Loss: 20.89775517, Log Avg loss: 1.48080462, Global Avg Loss: 20.89775517, Time: 0.0209 Steps: 1350, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001360, Sample Num: 21760, Cur Loss: 1.28990757, Cur Avg Loss: 20.75610018, Log Avg loss: 1.63267629, Global Avg Loss: 20.75610018, Time: 0.0209 Steps: 1360, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001370, Sample Num: 21920, Cur Loss: 1.18464565, Cur Avg Loss: 20.61646037, Log Avg loss: 1.62544627, Global Avg Loss: 20.61646037, Time: 0.0209 Steps: 1370, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001380, Sample Num: 22080, Cur Loss: 1.81816769, Cur Avg Loss: 20.47897974, Log Avg loss: 1.64413326, Global Avg Loss: 20.47897974, Time: 0.0209 Steps: 1380, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001390, Sample Num: 22240, Cur Loss: 1.34457135, Cur Avg Loss: 20.34437303, Log Avg loss: 1.76864663, Global Avg Loss: 20.34437303, Time: 0.0209 Steps: 1390, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001400, Sample Num: 22400, Cur Loss: 2.31962562, Cur Avg Loss: 20.21228019, Log Avg loss: 1.85137633, Global Avg Loss: 20.21228019, Time: 0.0208 Steps: 1400, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001410, Sample Num: 22560, Cur Loss: 1.52852035, Cur Avg Loss: 20.07910978, Log Avg loss: 1.43525177, Global Avg Loss: 20.07910978, Time: 0.0210 Steps: 1410, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001420, Sample Num: 22720, Cur Loss: 1.49224401, Cur Avg Loss: 19.94804028, Log Avg loss: 1.46724120, Global Avg Loss: 19.94804028, Time: 0.0210 Steps: 1420, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001430, Sample Num: 22880, Cur Loss: 0.97480667, Cur Avg Loss: 19.82023940, Log Avg loss: 1.67251452, Global Avg Loss: 19.82023940, Time: 0.0210 Steps: 1430, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001440, Sample Num: 23040, Cur Loss: 0.94097561, Cur Avg Loss: 19.69258326, Log Avg loss: 1.43775458, Global Avg Loss: 19.69258326, Time: 0.0210 Steps: 1440, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001450, Sample Num: 23200, Cur Loss: 0.62357527, Cur Avg Loss: 19.56834673, Log Avg loss: 1.67828619, Global Avg Loss: 19.56834673, Time: 0.0210 Steps: 1450, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001460, Sample Num: 23360, Cur Loss: 1.55025125, Cur Avg Loss: 19.44443399, Log Avg loss: 1.47708737, Global Avg Loss: 19.44443399, Time: 0.0210 Steps: 1460, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001470, Sample Num: 23520, Cur Loss: 2.88139343, Cur Avg Loss: 19.32168613, Log Avg loss: 1.40049785, Global Avg Loss: 19.32168613, Time: 0.0210 Steps: 1470, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001480, Sample Num: 23680, Cur Loss: 0.86333102, Cur Avg Loss: 19.20063956, Log Avg loss: 1.40679442, Global Avg Loss: 19.20063956, Time: 0.0209 Steps: 1480, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001490, Sample Num: 23840, Cur Loss: 1.37780464, Cur Avg Loss: 19.08132443, Log Avg loss: 1.42268551, Global Avg Loss: 19.08132443, Time: 0.0209 Steps: 1490, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001500, Sample Num: 24000, Cur Loss: 2.09842277, Cur Avg Loss: 18.96403472, Log Avg loss: 1.48786792, Global Avg Loss: 18.96403472, Time: 0.0209 Steps: 1500, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001510, Sample Num: 24160, Cur Loss: 1.22612953, Cur Avg Loss: 18.85017360, Log Avg loss: 1.77100589, Global Avg Loss: 18.85017360, Time: 0.0209 Steps: 1510, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001520, Sample Num: 24320, Cur Loss: 0.93881321, Cur Avg Loss: 18.73534878, Log Avg loss: 1.39680049, Global Avg Loss: 18.73534878, Time: 0.0209 Steps: 1520, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001530, Sample Num: 24480, Cur Loss: 0.99797571, Cur Avg Loss: 18.62151144, Log Avg loss: 1.31823642, Global Avg Loss: 18.62151144, Time: 0.0209 Steps: 1530, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001540, Sample Num: 24640, Cur Loss: 2.50846839, Cur Avg Loss: 18.51456891, Log Avg loss: 2.15236129, Global Avg Loss: 18.51456891, Time: 0.0209 Steps: 1540, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001550, Sample Num: 24800, Cur Loss: 2.07415104, Cur Avg Loss: 18.40453996, Log Avg loss: 1.46008162, Global Avg Loss: 18.40453996, Time: 0.0209 Steps: 1550, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001560, Sample Num: 24960, Cur Loss: 1.10792613, Cur Avg Loss: 18.29571110, Log Avg loss: 1.42723811, Global Avg Loss: 18.29571110, Time: 0.0209 Steps: 1560, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001570, Sample Num: 25120, Cur Loss: 0.93941784, Cur Avg Loss: 18.18796927, Log Avg loss: 1.38024370, Global Avg Loss: 18.18796927, Time: 0.0209 Steps: 1570, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001580, Sample Num: 25280, Cur Loss: 1.34427238, Cur Avg Loss: 18.08168298, Log Avg loss: 1.39473537, Global Avg Loss: 18.08168298, Time: 0.0209 Steps: 1580, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001590, Sample Num: 25440, Cur Loss: 1.21674824, Cur Avg Loss: 17.97737801, Log Avg loss: 1.49719250, Global Avg Loss: 17.97737801, Time: 0.0209 Steps: 1590, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001600, Sample Num: 25600, Cur Loss: 1.72757459, Cur Avg Loss: 17.87460590, Log Avg loss: 1.53384104, Global Avg Loss: 17.87460590, Time: 0.0210 Steps: 1600, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001610, Sample Num: 25760, Cur Loss: 1.90154338, Cur Avg Loss: 17.77319117, Log Avg loss: 1.54683321, Global Avg Loss: 17.77319117, Time: 0.0209 Steps: 1610, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001620, Sample Num: 25920, Cur Loss: 1.04052031, Cur Avg Loss: 17.67454020, Log Avg loss: 1.79173386, Global Avg Loss: 17.67454020, Time: 0.0209 Steps: 1620, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001630, Sample Num: 26080, Cur Loss: 0.95232862, Cur Avg Loss: 17.57451385, Log Avg loss: 1.37024595, Global Avg Loss: 17.57451385, Time: 0.0211 Steps: 1630, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001640, Sample Num: 26240, Cur Loss: 2.38093185, Cur Avg Loss: 17.47537464, Log Avg loss: 1.31568280, Global Avg Loss: 17.47537464, Time: 0.0209 Steps: 1640, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001650, Sample Num: 26400, Cur Loss: 0.99607557, Cur Avg Loss: 17.37585359, Log Avg loss: 1.05440189, Global Avg Loss: 17.37585359, Time: 0.0209 Steps: 1650, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001660, Sample Num: 26560, Cur Loss: 2.13616061, Cur Avg Loss: 17.28034834, Log Avg loss: 1.52198254, Global Avg Loss: 17.28034834, Time: 0.0209 Steps: 1660, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001670, Sample Num: 26720, Cur Loss: 1.32810688, Cur Avg Loss: 17.18582421, Log Avg loss: 1.49481779, Global Avg Loss: 17.18582421, Time: 0.0209 Steps: 1670, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001680, Sample Num: 26880, Cur Loss: 1.42875075, Cur Avg Loss: 17.09212754, Log Avg loss: 1.44478318, Global Avg Loss: 17.09212754, Time: 0.0209 Steps: 1680, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001690, Sample Num: 27040, Cur Loss: 1.39307332, Cur Avg Loss: 17.00200430, Log Avg loss: 1.86130154, Global Avg Loss: 17.00200430, Time: 0.0209 Steps: 1690, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001700, Sample Num: 27200, Cur Loss: 1.21455407, Cur Avg Loss: 16.91118972, Log Avg loss: 1.56352446, Global Avg Loss: 16.91118972, Time: 0.0209 Steps: 1700, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001710, Sample Num: 27360, Cur Loss: 1.95771205, Cur Avg Loss: 16.82246852, Log Avg loss: 1.73986521, Global Avg Loss: 16.82246852, Time: 0.0209 Steps: 1710, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001720, Sample Num: 27520, Cur Loss: 1.04602230, Cur Avg Loss: 16.73316130, Log Avg loss: 1.46162688, Global Avg Loss: 16.73316130, Time: 0.0209 Steps: 1720, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001730, Sample Num: 27680, Cur Loss: 2.02355790, Cur Avg Loss: 16.64623019, Log Avg loss: 1.69407838, Global Avg Loss: 16.64623019, Time: 0.0209 Steps: 1730, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001740, Sample Num: 27840, Cur Loss: 1.27058506, Cur Avg Loss: 16.55797472, Log Avg loss: 1.28977821, Global Avg Loss: 16.55797472, Time: 0.0209 Steps: 1740, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001750, Sample Num: 28000, Cur Loss: 0.92181623, Cur Avg Loss: 16.47127542, Log Avg loss: 1.38559781, Global Avg Loss: 16.47127542, Time: 0.0209 Steps: 1750, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001760, Sample Num: 28160, Cur Loss: 1.77097678, Cur Avg Loss: 16.38801699, Log Avg loss: 1.81779118, Global Avg Loss: 16.38801699, Time: 0.0209 Steps: 1760, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001770, Sample Num: 28320, Cur Loss: 2.48418021, Cur Avg Loss: 16.30451981, Log Avg loss: 1.60901598, Global Avg Loss: 16.30451981, Time: 0.0210 Steps: 1770, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001780, Sample Num: 28480, Cur Loss: 1.39545536, Cur Avg Loss: 16.22047718, Log Avg loss: 1.34493191, Global Avg Loss: 16.22047718, Time: 0.0209 Steps: 1780, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001790, Sample Num: 28640, Cur Loss: 0.41817343, Cur Avg Loss: 16.13814133, Log Avg loss: 1.48235975, Global Avg Loss: 16.13814133, Time: 0.0209 Steps: 1790, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001800, Sample Num: 28800, Cur Loss: 1.79722643, Cur Avg Loss: 16.05541875, Log Avg loss: 1.24807816, Global Avg Loss: 16.05541875, Time: 0.0209 Steps: 1800, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001810, Sample Num: 28960, Cur Loss: 2.36987305, Cur Avg Loss: 15.97583387, Log Avg loss: 1.65055498, Global Avg Loss: 15.97583387, Time: 0.0209 Steps: 1810, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001820, Sample Num: 29120, Cur Loss: 1.06598568, Cur Avg Loss: 15.89580164, Log Avg loss: 1.40996838, Global Avg Loss: 15.89580164, Time: 0.0209 Steps: 1820, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001830, Sample Num: 29280, Cur Loss: 1.41608667, Cur Avg Loss: 15.81611494, Log Avg loss: 1.31313477, Global Avg Loss: 15.81611494, Time: 0.0209 Steps: 1830, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001840, Sample Num: 29440, Cur Loss: 1.26721919, Cur Avg Loss: 15.73863577, Log Avg loss: 1.55994886, Global Avg Loss: 15.73863577, Time: 0.0209 Steps: 1840, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001850, Sample Num: 29600, Cur Loss: 2.08496189, Cur Avg Loss: 15.66135901, Log Avg loss: 1.44243511, Global Avg Loss: 15.66135901, Time: 0.0209 Steps: 1850, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001860, Sample Num: 29760, Cur Loss: 1.76648593, Cur Avg Loss: 15.58425283, Log Avg loss: 1.31960880, Global Avg Loss: 15.58425283, Time: 0.0210 Steps: 1860, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001870, Sample Num: 29920, Cur Loss: 1.57706547, Cur Avg Loss: 15.50986801, Log Avg loss: 1.67429162, Global Avg Loss: 15.50986801, Time: 0.0209 Steps: 1870, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001880, Sample Num: 30080, Cur Loss: 2.01677608, Cur Avg Loss: 15.43491006, Log Avg loss: 1.41777292, Global Avg Loss: 15.43491006, Time: 0.0209 Steps: 1880, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001890, Sample Num: 30240, Cur Loss: 0.53470176, Cur Avg Loss: 15.36100347, Log Avg loss: 1.46656596, Global Avg Loss: 15.36100347, Time: 0.0210 Steps: 1890, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001900, Sample Num: 30400, Cur Loss: 1.47498274, Cur Avg Loss: 15.28727492, Log Avg loss: 1.35257782, Global Avg Loss: 15.28727492, Time: 0.0210 Steps: 1900, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001910, Sample Num: 30560, Cur Loss: 1.26680493, Cur Avg Loss: 15.21600797, Log Avg loss: 1.67528841, Global Avg Loss: 15.21600797, Time: 0.0209 Steps: 1910, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001920, Sample Num: 30720, Cur Loss: 0.99493426, Cur Avg Loss: 15.14650508, Log Avg loss: 1.87145308, Global Avg Loss: 15.14650508, Time: 0.0209 Steps: 1920, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001930, Sample Num: 30880, Cur Loss: 0.69678938, Cur Avg Loss: 15.07657888, Log Avg loss: 1.65074846, Global Avg Loss: 15.07657888, Time: 0.0210 Steps: 1930, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001940, Sample Num: 31040, Cur Loss: 0.93312508, Cur Avg Loss: 15.00734230, Log Avg loss: 1.64468209, Global Avg Loss: 15.00734230, Time: 0.0209 Steps: 1940, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001950, Sample Num: 31200, Cur Loss: 1.47982860, Cur Avg Loss: 14.93884353, Log Avg loss: 1.65008235, Global Avg Loss: 14.93884353, Time: 0.0209 Steps: 1950, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001960, Sample Num: 31360, Cur Loss: 1.58749127, Cur Avg Loss: 14.87137235, Log Avg loss: 1.71449186, Global Avg Loss: 14.87137235, Time: 0.0210 Steps: 1960, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001970, Sample Num: 31520, Cur Loss: 1.23845482, Cur Avg Loss: 14.80337791, Log Avg loss: 1.47646670, Global Avg Loss: 14.80337791, Time: 0.0209 Steps: 1970, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001980, Sample Num: 31680, Cur Loss: 1.89015818, Cur Avg Loss: 14.73764000, Log Avg loss: 1.78727170, Global Avg Loss: 14.73764000, Time: 0.0209 Steps: 1980, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001990, Sample Num: 31840, Cur Loss: 0.89206839, Cur Avg Loss: 14.67110489, Log Avg loss: 1.49715471, Global Avg Loss: 14.67110489, Time: 0.0209 Steps: 1990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002000, Sample Num: 32000, Cur Loss: 1.10680616, Cur Avg Loss: 14.60428724, Log Avg loss: 1.30757495, Global Avg Loss: 14.60428724, Time: 0.0210 Steps: 2000, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002010, Sample Num: 32160, Cur Loss: 1.92237246, Cur Avg Loss: 14.53883321, Log Avg loss: 1.44802612, Global Avg Loss: 14.53883321, Time: 0.0209 Steps: 2010, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002020, Sample Num: 32320, Cur Loss: 1.70162725, Cur Avg Loss: 14.47466134, Log Avg loss: 1.57611526, Global Avg Loss: 14.47466134, Time: 0.0209 Steps: 2020, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002030, Sample Num: 32480, Cur Loss: 0.79923940, Cur Avg Loss: 14.40953239, Log Avg loss: 1.25348455, Global Avg Loss: 14.40953239, Time: 0.0210 Steps: 2030, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002040, Sample Num: 32640, Cur Loss: 1.87255454, Cur Avg Loss: 14.34701455, Log Avg loss: 1.65589275, Global Avg Loss: 14.34701455, Time: 0.0209 Steps: 2040, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002050, Sample Num: 32800, Cur Loss: 1.76490951, Cur Avg Loss: 14.28414531, Log Avg loss: 1.45882053, Global Avg Loss: 14.28414531, Time: 0.0210 Steps: 2050, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002060, Sample Num: 32960, Cur Loss: 1.30718756, Cur Avg Loss: 14.22207200, Log Avg loss: 1.49704346, Global Avg Loss: 14.22207200, Time: 0.0209 Steps: 2060, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002070, Sample Num: 33120, Cur Loss: 1.50593448, Cur Avg Loss: 14.16072997, Log Avg loss: 1.52427126, Global Avg Loss: 14.16072997, Time: 0.0210 Steps: 2070, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002080, Sample Num: 33280, Cur Loss: 1.38887739, Cur Avg Loss: 14.09919037, Log Avg loss: 1.36049373, Global Avg Loss: 14.09919037, Time: 0.0210 Steps: 2080, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002090, Sample Num: 33440, Cur Loss: 1.46083593, Cur Avg Loss: 14.03824245, Log Avg loss: 1.36107497, Global Avg Loss: 14.03824245, Time: 0.0209 Steps: 2090, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002100, Sample Num: 33600, Cur Loss: 1.23596489, Cur Avg Loss: 13.97918966, Log Avg loss: 1.63715761, Global Avg Loss: 13.97918966, Time: 0.0208 Steps: 2100, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002110, Sample Num: 33760, Cur Loss: 0.97524065, Cur Avg Loss: 13.91897986, Log Avg loss: 1.27492121, Global Avg Loss: 13.91897986, Time: 0.0208 Steps: 2110, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002120, Sample Num: 33920, Cur Loss: 1.77953374, Cur Avg Loss: 13.86146348, Log Avg loss: 1.72550669, Global Avg Loss: 13.86146348, Time: 0.0209 Steps: 2120, Updated lr: 0.000099 ***** Running evaluation checkpoint-2129 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-2129 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.624385, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.310242, "eval_total_loss": 921.099993, "eval_mae": 0.943505, "eval_mse": 1.31032, "eval_r2": 0.167075, "eval_sp_statistic": 0.458753, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.4369, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 2.235967, "test_total_loss": 1122.455259, "test_mae": 1.426932, "test_mse": 2.235457, "test_r2": -0.442784, "test_sp_statistic": 0.285715, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.305418, "test_ps_pvalue": 0.0, "lr": 9.892935040303462e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 13.809980796110244, "train_cur_epoch_loss": 29401.44911491871, "train_cur_epoch_avg_loss": 13.809980796110244, "train_cur_epoch_time": 44.6243851184845, "train_cur_epoch_avg_time": 0.020960256044379753, "epoch": 1, "step": 2129} ################################################## Training, Epoch: 0002, Batch: 000001, Sample Num: 16, Cur Loss: 0.92488724, Cur Avg Loss: 0.92488724, Log Avg loss: 1.60714320, Global Avg Loss: 13.80393146, Time: 0.0247 Steps: 2130, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000011, Sample Num: 176, Cur Loss: 1.49606895, Cur Avg Loss: 1.34418781, Log Avg loss: 1.38611787, Global Avg Loss: 13.74590429, Time: 0.0209 Steps: 2140, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000021, Sample Num: 336, Cur Loss: 2.94810843, Cur Avg Loss: 1.37874430, Log Avg loss: 1.41675644, Global Avg Loss: 13.68855942, Time: 0.0209 Steps: 2150, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000031, Sample Num: 496, Cur Loss: 2.96593189, Cur Avg Loss: 1.40754481, Log Avg loss: 1.46802589, Global Avg Loss: 13.63198287, Time: 0.0209 Steps: 2160, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000041, Sample Num: 656, Cur Loss: 1.62151027, Cur Avg Loss: 1.48064651, Log Avg loss: 1.70726178, Global Avg Loss: 13.57703024, Time: 0.0209 Steps: 2170, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000051, Sample Num: 816, Cur Loss: 0.54632807, Cur Avg Loss: 1.40680416, Log Avg loss: 1.10405052, Global Avg Loss: 13.51981474, Time: 0.0209 Steps: 2180, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000061, Sample Num: 976, Cur Loss: 0.92135686, Cur Avg Loss: 1.38295257, Log Avg loss: 1.26130946, Global Avg Loss: 13.46383983, Time: 0.0208 Steps: 2190, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000071, Sample Num: 1136, Cur Loss: 2.01518607, Cur Avg Loss: 1.41983156, Log Avg loss: 1.64479342, Global Avg Loss: 13.41011689, Time: 0.0209 Steps: 2200, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000081, Sample Num: 1296, Cur Loss: 1.55381989, Cur Avg Loss: 1.40972578, Log Avg loss: 1.33797472, Global Avg Loss: 13.35549181, Time: 0.0208 Steps: 2210, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000091, Sample Num: 1456, Cur Loss: 1.48702645, Cur Avg Loss: 1.43148824, Log Avg loss: 1.60776418, Global Avg Loss: 13.30257412, Time: 0.0208 Steps: 2220, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000101, Sample Num: 1616, Cur Loss: 2.11267662, Cur Avg Loss: 1.44440979, Log Avg loss: 1.56199594, Global Avg Loss: 13.24992579, Time: 0.0208 Steps: 2230, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000111, Sample Num: 1776, Cur Loss: 1.51075411, Cur Avg Loss: 1.46715711, Log Avg loss: 1.69690501, Global Avg Loss: 13.19834980, Time: 0.0208 Steps: 2240, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000121, Sample Num: 1936, Cur Loss: 0.75029409, Cur Avg Loss: 1.43010401, Log Avg loss: 1.01881455, Global Avg Loss: 13.14421853, Time: 0.0208 Steps: 2250, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000131, Sample Num: 2096, Cur Loss: 1.68339467, Cur Avg Loss: 1.44986605, Log Avg loss: 1.68898677, Global Avg Loss: 13.09353167, Time: 0.0208 Steps: 2260, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000141, Sample Num: 2256, Cur Loss: 1.60320175, Cur Avg Loss: 1.44879877, Log Avg loss: 1.43481737, Global Avg Loss: 13.04217169, Time: 0.0208 Steps: 2270, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000151, Sample Num: 2416, Cur Loss: 1.48918355, Cur Avg Loss: 1.45548384, Log Avg loss: 1.54974333, Global Avg Loss: 12.99176630, Time: 0.0208 Steps: 2280, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000161, Sample Num: 2576, Cur Loss: 1.50185215, Cur Avg Loss: 1.46708995, Log Avg loss: 1.64234229, Global Avg Loss: 12.94220550, Time: 0.0208 Steps: 2290, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000171, Sample Num: 2736, Cur Loss: 1.15548956, Cur Avg Loss: 1.45667776, Log Avg loss: 1.28904150, Global Avg Loss: 12.89153957, Time: 0.0208 Steps: 2300, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000181, Sample Num: 2896, Cur Loss: 1.18149614, Cur Avg Loss: 1.44949347, Log Avg loss: 1.32664213, Global Avg Loss: 12.84147508, Time: 0.0208 Steps: 2310, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000191, Sample Num: 3056, Cur Loss: 1.31992459, Cur Avg Loss: 1.44644904, Log Avg loss: 1.39134473, Global Avg Loss: 12.79212107, Time: 0.0208 Steps: 2320, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000201, Sample Num: 3216, Cur Loss: 1.95311022, Cur Avg Loss: 1.44847969, Log Avg loss: 1.48726518, Global Avg Loss: 12.74360237, Time: 0.0208 Steps: 2330, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000211, Sample Num: 3376, Cur Loss: 1.46751821, Cur Avg Loss: 1.43786247, Log Avg loss: 1.22445635, Global Avg Loss: 12.69437525, Time: 0.0208 Steps: 2340, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000221, Sample Num: 3536, Cur Loss: 1.52263045, Cur Avg Loss: 1.43081579, Log Avg loss: 1.28213077, Global Avg Loss: 12.64581251, Time: 0.0208 Steps: 2350, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000231, Sample Num: 3696, Cur Loss: 0.39833492, Cur Avg Loss: 1.42884888, Log Avg loss: 1.38538025, Global Avg Loss: 12.59809882, Time: 0.0208 Steps: 2360, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000241, Sample Num: 3856, Cur Loss: 0.96611959, Cur Avg Loss: 1.42887835, Log Avg loss: 1.42955917, Global Avg Loss: 12.55097418, Time: 0.0208 Steps: 2370, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000251, Sample Num: 4016, Cur Loss: 0.95823562, Cur Avg Loss: 1.42746197, Log Avg loss: 1.39332714, Global Avg Loss: 12.50409331, Time: 0.0208 Steps: 2380, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000261, Sample Num: 4176, Cur Loss: 1.58216882, Cur Avg Loss: 1.43634277, Log Avg loss: 1.65925091, Global Avg Loss: 12.45871740, Time: 0.0246 Steps: 2390, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000271, Sample Num: 4336, Cur Loss: 2.43610597, Cur Avg Loss: 1.44197375, Log Avg loss: 1.58894216, Global Avg Loss: 12.41342667, Time: 0.0209 Steps: 2400, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000281, Sample Num: 4496, Cur Loss: 1.23096633, Cur Avg Loss: 1.44233775, Log Avg loss: 1.45220220, Global Avg Loss: 12.36794441, Time: 0.0209 Steps: 2410, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000291, Sample Num: 4656, Cur Loss: 2.35874891, Cur Avg Loss: 1.45571687, Log Avg loss: 1.83167013, Global Avg Loss: 12.32440608, Time: 0.0209 Steps: 2420, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000301, Sample Num: 4816, Cur Loss: 1.21938586, Cur Avg Loss: 1.45122578, Log Avg loss: 1.32053505, Global Avg Loss: 12.27912266, Time: 0.0208 Steps: 2430, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000311, Sample Num: 4976, Cur Loss: 1.66397107, Cur Avg Loss: 1.45669528, Log Avg loss: 1.62132726, Global Avg Loss: 12.23544317, Time: 0.0209 Steps: 2440, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000321, Sample Num: 5136, Cur Loss: 1.55924273, Cur Avg Loss: 1.45873045, Log Avg loss: 1.52202415, Global Avg Loss: 12.19171493, Time: 0.0209 Steps: 2450, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000331, Sample Num: 5296, Cur Loss: 2.59442186, Cur Avg Loss: 1.47328704, Log Avg loss: 1.94055359, Global Avg Loss: 12.15004355, Time: 0.0209 Steps: 2460, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000341, Sample Num: 5456, Cur Loss: 1.72975445, Cur Avg Loss: 1.46860146, Log Avg loss: 1.31350892, Global Avg Loss: 12.10617094, Time: 0.0208 Steps: 2470, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000351, Sample Num: 5616, Cur Loss: 1.40758181, Cur Avg Loss: 1.47222324, Log Avg loss: 1.59572589, Global Avg Loss: 12.06379011, Time: 0.0209 Steps: 2480, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000361, Sample Num: 5776, Cur Loss: 1.71583152, Cur Avg Loss: 1.46356560, Log Avg loss: 1.15968248, Global Avg Loss: 12.01999851, Time: 0.0209 Steps: 2490, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000371, Sample Num: 5936, Cur Loss: 1.12702072, Cur Avg Loss: 1.46866374, Log Avg loss: 1.65270644, Global Avg Loss: 11.97852934, Time: 0.0208 Steps: 2500, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000381, Sample Num: 6096, Cur Loss: 0.85957503, Cur Avg Loss: 1.46309676, Log Avg loss: 1.25656204, Global Avg Loss: 11.93581234, Time: 0.0209 Steps: 2510, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000391, Sample Num: 6256, Cur Loss: 1.35413861, Cur Avg Loss: 1.46260503, Log Avg loss: 1.44386994, Global Avg Loss: 11.89417765, Time: 0.0209 Steps: 2520, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000401, Sample Num: 6416, Cur Loss: 0.82098246, Cur Avg Loss: 1.45919220, Log Avg loss: 1.32575057, Global Avg Loss: 11.85240521, Time: 0.0209 Steps: 2530, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000411, Sample Num: 6576, Cur Loss: 1.77551246, Cur Avg Loss: 1.45659285, Log Avg loss: 1.35235884, Global Avg Loss: 11.81106645, Time: 0.0209 Steps: 2540, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000421, Sample Num: 6736, Cur Loss: 0.83421552, Cur Avg Loss: 1.45703459, Log Avg loss: 1.47519014, Global Avg Loss: 11.77053360, Time: 0.0209 Steps: 2550, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000431, Sample Num: 6896, Cur Loss: 1.09960532, Cur Avg Loss: 1.46353049, Log Avg loss: 1.73700776, Global Avg Loss: 11.73134014, Time: 0.0209 Steps: 2560, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000441, Sample Num: 7056, Cur Loss: 1.35849738, Cur Avg Loss: 1.46324712, Log Avg loss: 1.45103409, Global Avg Loss: 11.69133895, Time: 0.0208 Steps: 2570, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000451, Sample Num: 7216, Cur Loss: 2.18553710, Cur Avg Loss: 1.45529202, Log Avg loss: 1.10447207, Global Avg Loss: 11.65030458, Time: 0.0209 Steps: 2580, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000461, Sample Num: 7376, Cur Loss: 0.45692670, Cur Avg Loss: 1.45433527, Log Avg loss: 1.41118575, Global Avg Loss: 11.61077130, Time: 0.0209 Steps: 2590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000471, Sample Num: 7536, Cur Loss: 1.62280035, Cur Avg Loss: 1.44639090, Log Avg loss: 1.08015531, Global Avg Loss: 11.57026893, Time: 0.0208 Steps: 2600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000481, Sample Num: 7696, Cur Loss: 1.79535818, Cur Avg Loss: 1.44024201, Log Avg loss: 1.15062934, Global Avg Loss: 11.53034694, Time: 0.0208 Steps: 2610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000491, Sample Num: 7856, Cur Loss: 0.88557696, Cur Avg Loss: 1.44258836, Log Avg loss: 1.55544780, Global Avg Loss: 11.49227481, Time: 0.0209 Steps: 2620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000501, Sample Num: 8016, Cur Loss: 1.26719582, Cur Avg Loss: 1.44191769, Log Avg loss: 1.40898782, Global Avg Loss: 11.45393531, Time: 0.0208 Steps: 2630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000511, Sample Num: 8176, Cur Loss: 0.75186944, Cur Avg Loss: 1.43390203, Log Avg loss: 1.03231740, Global Avg Loss: 11.41445949, Time: 0.0209 Steps: 2640, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000521, Sample Num: 8336, Cur Loss: 1.57543743, Cur Avg Loss: 1.43346832, Log Avg loss: 1.41130601, Global Avg Loss: 11.37671174, Time: 0.0209 Steps: 2650, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000531, Sample Num: 8496, Cur Loss: 2.04938364, Cur Avg Loss: 1.43230466, Log Avg loss: 1.37167765, Global Avg Loss: 11.33909883, Time: 0.0209 Steps: 2660, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000541, Sample Num: 8656, Cur Loss: 1.19384742, Cur Avg Loss: 1.43368083, Log Avg loss: 1.50675596, Global Avg Loss: 11.30227358, Time: 0.0208 Steps: 2670, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000551, Sample Num: 8816, Cur Loss: 0.87013036, Cur Avg Loss: 1.43062908, Log Avg loss: 1.26552929, Global Avg Loss: 11.26482304, Time: 0.0208 Steps: 2680, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000561, Sample Num: 8976, Cur Loss: 2.65354013, Cur Avg Loss: 1.42965728, Log Avg loss: 1.37611080, Global Avg Loss: 11.22806203, Time: 0.0209 Steps: 2690, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000571, Sample Num: 9136, Cur Loss: 1.01483691, Cur Avg Loss: 1.43121565, Log Avg loss: 1.51864052, Global Avg Loss: 11.19210120, Time: 0.0209 Steps: 2700, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000581, Sample Num: 9296, Cur Loss: 0.92046672, Cur Avg Loss: 1.42796906, Log Avg loss: 1.24258870, Global Avg Loss: 11.15538714, Time: 0.0208 Steps: 2710, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000591, Sample Num: 9456, Cur Loss: 1.73965466, Cur Avg Loss: 1.42302918, Log Avg loss: 1.13602218, Global Avg Loss: 11.11855124, Time: 0.0208 Steps: 2720, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000601, Sample Num: 9616, Cur Loss: 1.10805750, Cur Avg Loss: 1.41723182, Log Avg loss: 1.07460799, Global Avg Loss: 11.08176023, Time: 0.0209 Steps: 2730, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000611, Sample Num: 9776, Cur Loss: 1.13085842, Cur Avg Loss: 1.41662394, Log Avg loss: 1.38008993, Global Avg Loss: 11.04635268, Time: 0.0208 Steps: 2740, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000621, Sample Num: 9936, Cur Loss: 1.28176045, Cur Avg Loss: 1.41856213, Log Avg loss: 1.53698542, Global Avg Loss: 11.01177316, Time: 0.0209 Steps: 2750, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000631, Sample Num: 10096, Cur Loss: 2.02420545, Cur Avg Loss: 1.42196175, Log Avg loss: 1.63307824, Global Avg Loss: 10.97779238, Time: 0.0209 Steps: 2760, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000641, Sample Num: 10256, Cur Loss: 0.70462090, Cur Avg Loss: 1.41961378, Log Avg loss: 1.27145736, Global Avg Loss: 10.94275146, Time: 0.0209 Steps: 2770, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000651, Sample Num: 10416, Cur Loss: 1.91882730, Cur Avg Loss: 1.41767342, Log Avg loss: 1.29329607, Global Avg Loss: 10.90804119, Time: 0.0209 Steps: 2780, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000661, Sample Num: 10576, Cur Loss: 1.01372838, Cur Avg Loss: 1.41685211, Log Avg loss: 1.36338457, Global Avg Loss: 10.87383095, Time: 0.0209 Steps: 2790, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000671, Sample Num: 10736, Cur Loss: 1.41597128, Cur Avg Loss: 1.42236761, Log Avg loss: 1.78694237, Global Avg Loss: 10.84137778, Time: 0.0209 Steps: 2800, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000681, Sample Num: 10896, Cur Loss: 2.42532325, Cur Avg Loss: 1.42269791, Log Avg loss: 1.44486126, Global Avg Loss: 10.80793822, Time: 0.0209 Steps: 2810, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000691, Sample Num: 11056, Cur Loss: 0.82786173, Cur Avg Loss: 1.42231188, Log Avg loss: 1.39602323, Global Avg Loss: 10.77456263, Time: 0.0209 Steps: 2820, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000701, Sample Num: 11216, Cur Loss: 1.39607406, Cur Avg Loss: 1.42121125, Log Avg loss: 1.34515738, Global Avg Loss: 10.74124318, Time: 0.0209 Steps: 2830, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000711, Sample Num: 11376, Cur Loss: 1.19810390, Cur Avg Loss: 1.42161684, Log Avg loss: 1.45004894, Global Avg Loss: 10.70852771, Time: 0.0208 Steps: 2840, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000721, Sample Num: 11536, Cur Loss: 0.86975276, Cur Avg Loss: 1.41838914, Log Avg loss: 1.18889972, Global Avg Loss: 10.67512550, Time: 0.0209 Steps: 2850, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000731, Sample Num: 11696, Cur Loss: 1.63708866, Cur Avg Loss: 1.41507016, Log Avg loss: 1.17577154, Global Avg Loss: 10.64191098, Time: 0.0209 Steps: 2860, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000741, Sample Num: 11856, Cur Loss: 1.46046424, Cur Avg Loss: 1.41116035, Log Avg loss: 1.12535327, Global Avg Loss: 10.60875224, Time: 0.0209 Steps: 2870, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000751, Sample Num: 12016, Cur Loss: 0.79923433, Cur Avg Loss: 1.41113113, Log Avg loss: 1.40896586, Global Avg Loss: 10.57680854, Time: 0.0209 Steps: 2880, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000761, Sample Num: 12176, Cur Loss: 0.76923078, Cur Avg Loss: 1.40549971, Log Avg loss: 0.98258032, Global Avg Loss: 10.54361052, Time: 0.0208 Steps: 2890, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000771, Sample Num: 12336, Cur Loss: 0.95930707, Cur Avg Loss: 1.40042786, Log Avg loss: 1.01445954, Global Avg Loss: 10.51075138, Time: 0.0246 Steps: 2900, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000781, Sample Num: 12496, Cur Loss: 1.29350579, Cur Avg Loss: 1.40265777, Log Avg loss: 1.57458435, Global Avg Loss: 10.48004290, Time: 0.0209 Steps: 2910, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000791, Sample Num: 12656, Cur Loss: 1.78082561, Cur Avg Loss: 1.40394448, Log Avg loss: 1.50443616, Global Avg Loss: 10.44930452, Time: 0.0208 Steps: 2920, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000801, Sample Num: 12816, Cur Loss: 1.01008034, Cur Avg Loss: 1.40672995, Log Avg loss: 1.62706070, Global Avg Loss: 10.41919447, Time: 0.0209 Steps: 2930, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000811, Sample Num: 12976, Cur Loss: 1.44784796, Cur Avg Loss: 1.40467020, Log Avg loss: 1.23968461, Global Avg Loss: 10.38797165, Time: 0.0209 Steps: 2940, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000821, Sample Num: 13136, Cur Loss: 1.42599916, Cur Avg Loss: 1.40462141, Log Avg loss: 1.40066468, Global Avg Loss: 10.35750620, Time: 0.0209 Steps: 2950, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000831, Sample Num: 13296, Cur Loss: 1.76210761, Cur Avg Loss: 1.40113534, Log Avg loss: 1.11492870, Global Avg Loss: 10.32628128, Time: 0.0209 Steps: 2960, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000841, Sample Num: 13456, Cur Loss: 0.84771127, Cur Avg Loss: 1.39808763, Log Avg loss: 1.14482250, Global Avg Loss: 10.29536728, Time: 0.0208 Steps: 2970, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000851, Sample Num: 13616, Cur Loss: 1.94331110, Cur Avg Loss: 1.39860210, Log Avg loss: 1.44186898, Global Avg Loss: 10.26565755, Time: 0.0208 Steps: 2980, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000861, Sample Num: 13776, Cur Loss: 0.74702120, Cur Avg Loss: 1.39548403, Log Avg loss: 1.13013657, Global Avg Loss: 10.23510397, Time: 0.0209 Steps: 2990, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000871, Sample Num: 13936, Cur Loss: 0.93600690, Cur Avg Loss: 1.38982747, Log Avg loss: 0.90279732, Global Avg Loss: 10.20399628, Time: 0.0209 Steps: 3000, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000881, Sample Num: 14096, Cur Loss: 1.60346949, Cur Avg Loss: 1.39017642, Log Avg loss: 1.42057010, Global Avg Loss: 10.17481546, Time: 0.0209 Steps: 3010, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000891, Sample Num: 14256, Cur Loss: 1.48344088, Cur Avg Loss: 1.39001410, Log Avg loss: 1.37571434, Global Avg Loss: 10.14567936, Time: 0.0209 Steps: 3020, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000901, Sample Num: 14416, Cur Loss: 0.62241614, Cur Avg Loss: 1.38801724, Log Avg loss: 1.21009630, Global Avg Loss: 10.11618899, Time: 0.0209 Steps: 3030, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000911, Sample Num: 14576, Cur Loss: 1.59669209, Cur Avg Loss: 1.38859731, Log Avg loss: 1.44086186, Global Avg Loss: 10.08765173, Time: 0.0209 Steps: 3040, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000921, Sample Num: 14736, Cur Loss: 1.22623575, Cur Avg Loss: 1.38893983, Log Avg loss: 1.42014333, Global Avg Loss: 10.05923367, Time: 0.0208 Steps: 3050, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000931, Sample Num: 14896, Cur Loss: 2.62097454, Cur Avg Loss: 1.38851379, Log Avg loss: 1.34927530, Global Avg Loss: 10.03076975, Time: 0.0209 Steps: 3060, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000941, Sample Num: 15056, Cur Loss: 0.92132944, Cur Avg Loss: 1.38824661, Log Avg loss: 1.36337208, Global Avg Loss: 10.00253719, Time: 0.0209 Steps: 3070, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000951, Sample Num: 15216, Cur Loss: 1.15967929, Cur Avg Loss: 1.38894457, Log Avg loss: 1.45462342, Global Avg Loss: 9.97478422, Time: 0.0209 Steps: 3080, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000961, Sample Num: 15376, Cur Loss: 1.12009728, Cur Avg Loss: 1.38942907, Log Avg loss: 1.43550512, Global Avg Loss: 9.94714901, Time: 0.0209 Steps: 3090, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000971, Sample Num: 15536, Cur Loss: 2.09140038, Cur Avg Loss: 1.39031801, Log Avg loss: 1.47574440, Global Avg Loss: 9.91982190, Time: 0.0209 Steps: 3100, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000981, Sample Num: 15696, Cur Loss: 0.55061024, Cur Avg Loss: 1.38990574, Log Avg loss: 1.34987429, Global Avg Loss: 9.89226580, Time: 0.0209 Steps: 3110, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000991, Sample Num: 15856, Cur Loss: 1.36849523, Cur Avg Loss: 1.38644385, Log Avg loss: 1.04683326, Global Avg Loss: 9.86391506, Time: 0.0209 Steps: 3120, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001001, Sample Num: 16016, Cur Loss: 0.85978031, Cur Avg Loss: 1.38447556, Log Avg loss: 1.18941724, Global Avg Loss: 9.83620101, Time: 0.0209 Steps: 3130, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001011, Sample Num: 16176, Cur Loss: 1.04253888, Cur Avg Loss: 1.38519838, Log Avg loss: 1.45755343, Global Avg Loss: 9.80951741, Time: 0.0208 Steps: 3140, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001021, Sample Num: 16336, Cur Loss: 0.84881246, Cur Avg Loss: 1.38453360, Log Avg loss: 1.31732355, Global Avg Loss: 9.78255807, Time: 0.0209 Steps: 3150, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001031, Sample Num: 16496, Cur Loss: 0.43700057, Cur Avg Loss: 1.38558101, Log Avg loss: 1.49252146, Global Avg Loss: 9.75632378, Time: 0.0210 Steps: 3160, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001041, Sample Num: 16656, Cur Loss: 1.58734083, Cur Avg Loss: 1.38446027, Log Avg loss: 1.26891198, Global Avg Loss: 9.72954961, Time: 0.0209 Steps: 3170, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001051, Sample Num: 16816, Cur Loss: 1.14704394, Cur Avg Loss: 1.38546367, Log Avg loss: 1.48991818, Global Avg Loss: 9.70363882, Time: 0.0209 Steps: 3180, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001061, Sample Num: 16976, Cur Loss: 1.02123821, Cur Avg Loss: 1.38439420, Log Avg loss: 1.27199255, Global Avg Loss: 9.67720732, Time: 0.0209 Steps: 3190, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001071, Sample Num: 17136, Cur Loss: 1.67577457, Cur Avg Loss: 1.38222119, Log Avg loss: 1.15166516, Global Avg Loss: 9.65056500, Time: 0.0209 Steps: 3200, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001081, Sample Num: 17296, Cur Loss: 1.40959394, Cur Avg Loss: 1.38525564, Log Avg loss: 1.71024499, Global Avg Loss: 9.62582880, Time: 0.0208 Steps: 3210, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001091, Sample Num: 17456, Cur Loss: 0.97105324, Cur Avg Loss: 1.38513625, Log Avg loss: 1.37223047, Global Avg Loss: 9.60019651, Time: 0.0209 Steps: 3220, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001101, Sample Num: 17616, Cur Loss: 1.41639781, Cur Avg Loss: 1.38386247, Log Avg loss: 1.24489323, Global Avg Loss: 9.57432870, Time: 0.0209 Steps: 3230, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001111, Sample Num: 17776, Cur Loss: 0.59379661, Cur Avg Loss: 1.38236324, Log Avg loss: 1.21729758, Global Avg Loss: 9.54853539, Time: 0.0209 Steps: 3240, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001121, Sample Num: 17936, Cur Loss: 1.36076069, Cur Avg Loss: 1.38109464, Log Avg loss: 1.24015381, Global Avg Loss: 9.52297114, Time: 0.0208 Steps: 3250, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001131, Sample Num: 18096, Cur Loss: 0.80299884, Cur Avg Loss: 1.38004581, Log Avg loss: 1.26247118, Global Avg Loss: 9.49763219, Time: 0.0209 Steps: 3260, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001141, Sample Num: 18256, Cur Loss: 1.49935865, Cur Avg Loss: 1.37770290, Log Avg loss: 1.11272039, Global Avg Loss: 9.47199025, Time: 0.0214 Steps: 3270, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001151, Sample Num: 18416, Cur Loss: 1.28887820, Cur Avg Loss: 1.37806609, Log Avg loss: 1.41950580, Global Avg Loss: 9.44744000, Time: 0.0209 Steps: 3280, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001161, Sample Num: 18576, Cur Loss: 1.18857121, Cur Avg Loss: 1.37965035, Log Avg loss: 1.56199824, Global Avg Loss: 9.42347209, Time: 0.0208 Steps: 3290, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001171, Sample Num: 18736, Cur Loss: 0.83174109, Cur Avg Loss: 1.37848545, Log Avg loss: 1.24324141, Global Avg Loss: 9.39868351, Time: 0.0209 Steps: 3300, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001181, Sample Num: 18896, Cur Loss: 1.37678599, Cur Avg Loss: 1.38020834, Log Avg loss: 1.58195849, Global Avg Loss: 9.37506803, Time: 0.0208 Steps: 3310, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001191, Sample Num: 19056, Cur Loss: 1.71063244, Cur Avg Loss: 1.38017510, Log Avg loss: 1.37624889, Global Avg Loss: 9.35097520, Time: 0.0209 Steps: 3320, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001201, Sample Num: 19216, Cur Loss: 2.00212169, Cur Avg Loss: 1.37871677, Log Avg loss: 1.20503042, Global Avg Loss: 9.32651290, Time: 0.0208 Steps: 3330, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001211, Sample Num: 19376, Cur Loss: 2.32065773, Cur Avg Loss: 1.37603365, Log Avg loss: 1.05379103, Global Avg Loss: 9.30174427, Time: 0.0209 Steps: 3340, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001221, Sample Num: 19536, Cur Loss: 1.07406056, Cur Avg Loss: 1.37612785, Log Avg loss: 1.38753558, Global Avg Loss: 9.27811977, Time: 0.0209 Steps: 3350, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001231, Sample Num: 19696, Cur Loss: 1.42500401, Cur Avg Loss: 1.37679279, Log Avg loss: 1.45798129, Global Avg Loss: 9.25484555, Time: 0.0209 Steps: 3360, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001241, Sample Num: 19856, Cur Loss: 1.24261189, Cur Avg Loss: 1.37465123, Log Avg loss: 1.11102495, Global Avg Loss: 9.23067991, Time: 0.0209 Steps: 3370, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001251, Sample Num: 20016, Cur Loss: 1.21985936, Cur Avg Loss: 1.37383152, Log Avg loss: 1.27210560, Global Avg Loss: 9.20713383, Time: 0.0208 Steps: 3380, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001261, Sample Num: 20176, Cur Loss: 1.35933971, Cur Avg Loss: 1.37714536, Log Avg loss: 1.79170690, Global Avg Loss: 9.18525941, Time: 0.0209 Steps: 3390, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001271, Sample Num: 20336, Cur Loss: 1.86907864, Cur Avg Loss: 1.37700282, Log Avg loss: 1.35902829, Global Avg Loss: 9.16224109, Time: 0.0209 Steps: 3400, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001281, Sample Num: 20496, Cur Loss: 1.64150155, Cur Avg Loss: 1.37659394, Log Avg loss: 1.32462570, Global Avg Loss: 9.13925688, Time: 0.0245 Steps: 3410, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001291, Sample Num: 20656, Cur Loss: 2.15363860, Cur Avg Loss: 1.37628075, Log Avg loss: 1.33616045, Global Avg Loss: 9.11644081, Time: 0.0209 Steps: 3420, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001301, Sample Num: 20816, Cur Loss: 1.00824618, Cur Avg Loss: 1.37643180, Log Avg loss: 1.39593303, Global Avg Loss: 9.09393204, Time: 0.0209 Steps: 3430, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001311, Sample Num: 20976, Cur Loss: 2.03705812, Cur Avg Loss: 1.37685925, Log Avg loss: 1.43247076, Global Avg Loss: 9.07166035, Time: 0.0209 Steps: 3440, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001321, Sample Num: 21136, Cur Loss: 0.67320502, Cur Avg Loss: 1.37646487, Log Avg loss: 1.32476093, Global Avg Loss: 9.04920557, Time: 0.0209 Steps: 3450, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001331, Sample Num: 21296, Cur Loss: 1.13021588, Cur Avg Loss: 1.37497333, Log Avg loss: 1.17794082, Global Avg Loss: 9.02645625, Time: 0.0209 Steps: 3460, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001341, Sample Num: 21456, Cur Loss: 1.24299765, Cur Avg Loss: 1.37568720, Log Avg loss: 1.47070435, Global Avg Loss: 9.00468175, Time: 0.0209 Steps: 3470, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001351, Sample Num: 21616, Cur Loss: 1.35693252, Cur Avg Loss: 1.37502273, Log Avg loss: 1.28591734, Global Avg Loss: 8.98250139, Time: 0.0209 Steps: 3480, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001361, Sample Num: 21776, Cur Loss: 1.38748801, Cur Avg Loss: 1.37252514, Log Avg loss: 1.03509944, Global Avg Loss: 8.95972946, Time: 0.0208 Steps: 3490, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001371, Sample Num: 21936, Cur Loss: 2.27393532, Cur Avg Loss: 1.37445476, Log Avg loss: 1.63707724, Global Avg Loss: 8.93880760, Time: 0.0209 Steps: 3500, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001381, Sample Num: 22096, Cur Loss: 3.09634972, Cur Avg Loss: 1.37425284, Log Avg loss: 1.34656836, Global Avg Loss: 8.91717729, Time: 0.0209 Steps: 3510, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001391, Sample Num: 22256, Cur Loss: 1.24583030, Cur Avg Loss: 1.37298396, Log Avg loss: 1.19775187, Global Avg Loss: 8.89524710, Time: 0.0209 Steps: 3520, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001401, Sample Num: 22416, Cur Loss: 1.89745736, Cur Avg Loss: 1.37454940, Log Avg loss: 1.59230245, Global Avg Loss: 8.87455887, Time: 0.0208 Steps: 3530, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001411, Sample Num: 22576, Cur Loss: 1.62157989, Cur Avg Loss: 1.37360739, Log Avg loss: 1.24163213, Global Avg Loss: 8.85299693, Time: 0.0209 Steps: 3540, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001421, Sample Num: 22736, Cur Loss: 0.69894618, Cur Avg Loss: 1.37487567, Log Avg loss: 1.55382991, Global Avg Loss: 8.83243590, Time: 0.0209 Steps: 3550, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001431, Sample Num: 22896, Cur Loss: 0.81205755, Cur Avg Loss: 1.37334520, Log Avg loss: 1.15586497, Global Avg Loss: 8.81087250, Time: 0.0209 Steps: 3560, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001441, Sample Num: 23056, Cur Loss: 0.71458924, Cur Avg Loss: 1.37179706, Log Avg loss: 1.15025801, Global Avg Loss: 8.78941419, Time: 0.0209 Steps: 3570, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001451, Sample Num: 23216, Cur Loss: 1.94770491, Cur Avg Loss: 1.37083477, Log Avg loss: 1.23216914, Global Avg Loss: 8.76830457, Time: 0.0209 Steps: 3580, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001461, Sample Num: 23376, Cur Loss: 1.49121618, Cur Avg Loss: 1.37049450, Log Avg loss: 1.32112111, Global Avg Loss: 8.74756033, Time: 0.0209 Steps: 3590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001471, Sample Num: 23536, Cur Loss: 0.36725229, Cur Avg Loss: 1.36712097, Log Avg loss: 0.87424824, Global Avg Loss: 8.72569002, Time: 0.0209 Steps: 3600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001481, Sample Num: 23696, Cur Loss: 1.95359361, Cur Avg Loss: 1.36526326, Log Avg loss: 1.09199449, Global Avg Loss: 8.70454405, Time: 0.0209 Steps: 3610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001491, Sample Num: 23856, Cur Loss: 0.69049847, Cur Avg Loss: 1.36471951, Log Avg loss: 1.28419020, Global Avg Loss: 8.68404583, Time: 0.0208 Steps: 3620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001501, Sample Num: 24016, Cur Loss: 1.49184203, Cur Avg Loss: 1.36300578, Log Avg loss: 1.10748867, Global Avg Loss: 8.66317377, Time: 0.0208 Steps: 3630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001511, Sample Num: 24176, Cur Loss: 1.43803060, Cur Avg Loss: 1.36177552, Log Avg loss: 1.17711347, Global Avg Loss: 8.64260767, Time: 0.0209 Steps: 3640, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001521, Sample Num: 24336, Cur Loss: 0.54915226, Cur Avg Loss: 1.36232617, Log Avg loss: 1.44552894, Global Avg Loss: 8.62288965, Time: 0.0208 Steps: 3650, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001531, Sample Num: 24496, Cur Loss: 0.81238931, Cur Avg Loss: 1.36075515, Log Avg loss: 1.12180397, Global Avg Loss: 8.60239488, Time: 0.0209 Steps: 3660, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001541, Sample Num: 24656, Cur Loss: 0.58148831, Cur Avg Loss: 1.35861116, Log Avg loss: 1.03036620, Global Avg Loss: 8.58176265, Time: 0.0246 Steps: 3670, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001551, Sample Num: 24816, Cur Loss: 1.15408707, Cur Avg Loss: 1.35764243, Log Avg loss: 1.20836053, Global Avg Loss: 8.56172623, Time: 0.0209 Steps: 3680, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001561, Sample Num: 24976, Cur Loss: 0.67402214, Cur Avg Loss: 1.35568349, Log Avg loss: 1.05185246, Global Avg Loss: 8.54137427, Time: 0.0209 Steps: 3690, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001571, Sample Num: 25136, Cur Loss: 1.72652984, Cur Avg Loss: 1.35657435, Log Avg loss: 1.49563776, Global Avg Loss: 8.52233174, Time: 0.0209 Steps: 3700, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001581, Sample Num: 25296, Cur Loss: 0.70182997, Cur Avg Loss: 1.35533300, Log Avg loss: 1.16031565, Global Avg Loss: 8.50248803, Time: 0.0209 Steps: 3710, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001591, Sample Num: 25456, Cur Loss: 1.50763977, Cur Avg Loss: 1.35503787, Log Avg loss: 1.30837772, Global Avg Loss: 8.48314902, Time: 0.0209 Steps: 3720, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001601, Sample Num: 25616, Cur Loss: 0.76974243, Cur Avg Loss: 1.35372236, Log Avg loss: 1.14442577, Global Avg Loss: 8.46347416, Time: 0.0209 Steps: 3730, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001611, Sample Num: 25776, Cur Loss: 1.51462972, Cur Avg Loss: 1.35548617, Log Avg loss: 1.63787155, Global Avg Loss: 8.44522389, Time: 0.0209 Steps: 3740, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001621, Sample Num: 25936, Cur Loss: 1.29422092, Cur Avg Loss: 1.35379988, Log Avg loss: 1.08213919, Global Avg Loss: 8.42558899, Time: 0.0209 Steps: 3750, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001631, Sample Num: 26096, Cur Loss: 1.73126757, Cur Avg Loss: 1.35599247, Log Avg loss: 1.71141082, Global Avg Loss: 8.40773214, Time: 0.0209 Steps: 3760, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001641, Sample Num: 26256, Cur Loss: 0.91369259, Cur Avg Loss: 1.35337842, Log Avg loss: 0.92702685, Global Avg Loss: 8.38788942, Time: 0.0208 Steps: 3770, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001651, Sample Num: 26416, Cur Loss: 1.07967556, Cur Avg Loss: 1.35099539, Log Avg loss: 0.95994050, Global Avg Loss: 8.36823876, Time: 0.0208 Steps: 3780, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001661, Sample Num: 26576, Cur Loss: 0.56597823, Cur Avg Loss: 1.34849478, Log Avg loss: 0.93564386, Global Avg Loss: 8.34862769, Time: 0.0209 Steps: 3790, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001671, Sample Num: 26736, Cur Loss: 1.14765573, Cur Avg Loss: 1.34701200, Log Avg loss: 1.10072223, Global Avg Loss: 8.32955425, Time: 0.0209 Steps: 3800, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001681, Sample Num: 26896, Cur Loss: 0.53761917, Cur Avg Loss: 1.34550716, Log Avg loss: 1.09404779, Global Avg Loss: 8.31056342, Time: 0.0209 Steps: 3810, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001691, Sample Num: 27056, Cur Loss: 1.02029240, Cur Avg Loss: 1.34517100, Log Avg loss: 1.28866311, Global Avg Loss: 8.29218149, Time: 0.0208 Steps: 3820, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001701, Sample Num: 27216, Cur Loss: 1.07392001, Cur Avg Loss: 1.34320977, Log Avg loss: 1.01156664, Global Avg Loss: 8.27317205, Time: 0.0209 Steps: 3830, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001711, Sample Num: 27376, Cur Loss: 0.22963817, Cur Avg Loss: 1.34182697, Log Avg loss: 1.10661246, Global Avg Loss: 8.25450913, Time: 0.0209 Steps: 3840, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001721, Sample Num: 27536, Cur Loss: 1.32770956, Cur Avg Loss: 1.34192625, Log Avg loss: 1.35891235, Global Avg Loss: 8.23659849, Time: 0.0209 Steps: 3850, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001731, Sample Num: 27696, Cur Loss: 1.19180894, Cur Avg Loss: 1.34153555, Log Avg loss: 1.27429679, Global Avg Loss: 8.21856144, Time: 0.0209 Steps: 3860, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001741, Sample Num: 27856, Cur Loss: 1.24168491, Cur Avg Loss: 1.34199566, Log Avg loss: 1.42163963, Global Avg Loss: 8.20099833, Time: 0.0208 Steps: 3870, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001751, Sample Num: 28016, Cur Loss: 1.28401160, Cur Avg Loss: 1.34078865, Log Avg loss: 1.13064924, Global Avg Loss: 8.18277579, Time: 0.0209 Steps: 3880, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001761, Sample Num: 28176, Cur Loss: 0.93812352, Cur Avg Loss: 1.34221298, Log Avg loss: 1.59161183, Global Avg Loss: 8.16583192, Time: 0.0209 Steps: 3890, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001771, Sample Num: 28336, Cur Loss: 3.02476740, Cur Avg Loss: 1.34213296, Log Avg loss: 1.32804285, Global Avg Loss: 8.14829913, Time: 0.0209 Steps: 3900, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001781, Sample Num: 28496, Cur Loss: 1.93999410, Cur Avg Loss: 1.34173020, Log Avg loss: 1.27040125, Global Avg Loss: 8.13070859, Time: 0.0209 Steps: 3910, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001791, Sample Num: 28656, Cur Loss: 1.65911150, Cur Avg Loss: 1.34057157, Log Avg loss: 1.13421870, Global Avg Loss: 8.11286041, Time: 0.0209 Steps: 3920, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001801, Sample Num: 28816, Cur Loss: 0.65188670, Cur Avg Loss: 1.33835752, Log Avg loss: 0.94182110, Global Avg Loss: 8.09461349, Time: 0.0209 Steps: 3930, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001811, Sample Num: 28976, Cur Loss: 1.35624933, Cur Avg Loss: 1.33802990, Log Avg loss: 1.27902535, Global Avg Loss: 8.07731504, Time: 0.0209 Steps: 3940, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001821, Sample Num: 29136, Cur Loss: 1.65318787, Cur Avg Loss: 1.33825635, Log Avg loss: 1.37926669, Global Avg Loss: 8.06035796, Time: 0.0209 Steps: 3950, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001831, Sample Num: 29296, Cur Loss: 2.07500911, Cur Avg Loss: 1.34015646, Log Avg loss: 1.68616610, Global Avg Loss: 8.04426151, Time: 0.0209 Steps: 3960, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001841, Sample Num: 29456, Cur Loss: 1.22709465, Cur Avg Loss: 1.33908947, Log Avg loss: 1.14372409, Global Avg Loss: 8.02687981, Time: 0.0209 Steps: 3970, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001851, Sample Num: 29616, Cur Loss: 0.69871867, Cur Avg Loss: 1.33758410, Log Avg loss: 1.06044620, Global Avg Loss: 8.00937620, Time: 0.0209 Steps: 3980, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001861, Sample Num: 29776, Cur Loss: 1.01667273, Cur Avg Loss: 1.33705658, Log Avg loss: 1.23941280, Global Avg Loss: 7.99240888, Time: 0.0209 Steps: 3990, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001871, Sample Num: 29936, Cur Loss: 0.91989827, Cur Avg Loss: 1.33615717, Log Avg loss: 1.16877559, Global Avg Loss: 7.97534979, Time: 0.0209 Steps: 4000, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001881, Sample Num: 30096, Cur Loss: 1.97120690, Cur Avg Loss: 1.33708092, Log Avg loss: 1.50991479, Global Avg Loss: 7.95922651, Time: 0.0209 Steps: 4010, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001891, Sample Num: 30256, Cur Loss: 1.15979791, Cur Avg Loss: 1.33782749, Log Avg loss: 1.47825772, Global Avg Loss: 7.94310470, Time: 0.0209 Steps: 4020, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001901, Sample Num: 30416, Cur Loss: 1.15783739, Cur Avg Loss: 1.33916245, Log Avg loss: 1.59160263, Global Avg Loss: 7.92734415, Time: 0.0209 Steps: 4030, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001911, Sample Num: 30576, Cur Loss: 0.96106070, Cur Avg Loss: 1.33828851, Log Avg loss: 1.17215415, Global Avg Loss: 7.91062338, Time: 0.0209 Steps: 4040, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001921, Sample Num: 30736, Cur Loss: 0.70792615, Cur Avg Loss: 1.33707175, Log Avg loss: 1.10454801, Global Avg Loss: 7.89381826, Time: 0.0209 Steps: 4050, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001931, Sample Num: 30896, Cur Loss: 1.62259007, Cur Avg Loss: 1.33713763, Log Avg loss: 1.34979409, Global Avg Loss: 7.87769997, Time: 0.0209 Steps: 4060, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001941, Sample Num: 31056, Cur Loss: 0.89680505, Cur Avg Loss: 1.33613655, Log Avg loss: 1.14282781, Global Avg Loss: 7.86115237, Time: 0.0208 Steps: 4070, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001951, Sample Num: 31216, Cur Loss: 0.99675596, Cur Avg Loss: 1.33597324, Log Avg loss: 1.30427455, Global Avg Loss: 7.84508160, Time: 0.0209 Steps: 4080, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001961, Sample Num: 31376, Cur Loss: 1.19062543, Cur Avg Loss: 1.33517791, Log Avg loss: 1.18000935, Global Avg Loss: 7.82878558, Time: 0.0209 Steps: 4090, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001971, Sample Num: 31536, Cur Loss: 1.92835808, Cur Avg Loss: 1.33565501, Log Avg loss: 1.42921306, Global Avg Loss: 7.81317686, Time: 0.0209 Steps: 4100, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001981, Sample Num: 31696, Cur Loss: 1.84204555, Cur Avg Loss: 1.33492279, Log Avg loss: 1.19060300, Global Avg Loss: 7.79706354, Time: 0.0209 Steps: 4110, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001991, Sample Num: 31856, Cur Loss: 2.41589403, Cur Avg Loss: 1.33337181, Log Avg loss: 1.02612250, Global Avg Loss: 7.78062922, Time: 0.0209 Steps: 4120, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002001, Sample Num: 32016, Cur Loss: 2.07650113, Cur Avg Loss: 1.33368088, Log Avg loss: 1.39521707, Global Avg Loss: 7.76516817, Time: 0.0208 Steps: 4130, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002011, Sample Num: 32176, Cur Loss: 0.96613050, Cur Avg Loss: 1.33239069, Log Avg loss: 1.07422263, Global Avg Loss: 7.74900647, Time: 0.0209 Steps: 4140, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002021, Sample Num: 32336, Cur Loss: 1.11348176, Cur Avg Loss: 1.33082700, Log Avg loss: 1.01636966, Global Avg Loss: 7.73278325, Time: 0.0209 Steps: 4150, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002031, Sample Num: 32496, Cur Loss: 1.78332686, Cur Avg Loss: 1.32991619, Log Avg loss: 1.14584195, Global Avg Loss: 7.71694926, Time: 0.0209 Steps: 4160, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002041, Sample Num: 32656, Cur Loss: 1.17733812, Cur Avg Loss: 1.32968053, Log Avg loss: 1.28181827, Global Avg Loss: 7.70151729, Time: 0.0209 Steps: 4170, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002051, Sample Num: 32816, Cur Loss: 0.99655008, Cur Avg Loss: 1.32888430, Log Avg loss: 1.16637240, Global Avg Loss: 7.68588297, Time: 0.0246 Steps: 4180, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002061, Sample Num: 32976, Cur Loss: 0.68015283, Cur Avg Loss: 1.32698159, Log Avg loss: 0.93673685, Global Avg Loss: 7.66977522, Time: 0.0208 Steps: 4190, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002071, Sample Num: 33136, Cur Loss: 0.94822830, Cur Avg Loss: 1.32543016, Log Avg loss: 1.00567985, Global Avg Loss: 7.65390833, Time: 0.0208 Steps: 4200, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002081, Sample Num: 33296, Cur Loss: 0.61283946, Cur Avg Loss: 1.32490857, Log Avg loss: 1.21688826, Global Avg Loss: 7.63861849, Time: 0.0208 Steps: 4210, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002091, Sample Num: 33456, Cur Loss: 0.94291735, Cur Avg Loss: 1.32410857, Log Avg loss: 1.15762700, Global Avg Loss: 7.62326069, Time: 0.0209 Steps: 4220, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002101, Sample Num: 33616, Cur Loss: 1.52829266, Cur Avg Loss: 1.32296676, Log Avg loss: 1.08421565, Global Avg Loss: 7.60780196, Time: 0.0209 Steps: 4230, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002111, Sample Num: 33776, Cur Loss: 0.93393505, Cur Avg Loss: 1.32279499, Log Avg loss: 1.28670492, Global Avg Loss: 7.59289371, Time: 0.0209 Steps: 4240, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002121, Sample Num: 33936, Cur Loss: 0.97642839, Cur Avg Loss: 1.32227573, Log Avg loss: 1.21266093, Global Avg Loss: 7.57788140, Time: 0.0208 Steps: 4250, Updated lr: 0.000097 ***** Running evaluation checkpoint-4258 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-4258 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.471648, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.070351, "eval_total_loss": 752.456885, "eval_mae": 0.779006, "eval_mse": 1.070543, "eval_r2": 0.319493, "eval_sp_statistic": 0.600207, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.583677, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.513765, "test_total_loss": 759.91028, "test_mae": 1.12482, "test_mse": 1.513687, "test_r2": 0.023053, "test_sp_statistic": 0.425001, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.487351, "test_ps_pvalue": 0.0, "lr": 9.691038406827881e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 7.565667083312811, "train_cur_epoch_loss": 2813.161325827241, "train_cur_epoch_avg_loss": 1.3213533705153786, "train_cur_epoch_time": 44.47164821624756, "train_cur_epoch_avg_time": 0.020888514897251083, "epoch": 2, "step": 4258} ################################################## Training, Epoch: 0003, Batch: 000002, Sample Num: 32, Cur Loss: 0.95674694, Cur Avg Loss: 1.05500859, Log Avg loss: 1.07245147, Global Avg Loss: 7.56261044, Time: 0.0247 Steps: 4260, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000012, Sample Num: 192, Cur Loss: 1.22763872, Cur Avg Loss: 1.24171044, Log Avg loss: 1.27905081, Global Avg Loss: 7.54789484, Time: 0.0209 Steps: 4270, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000022, Sample Num: 352, Cur Loss: 1.48249722, Cur Avg Loss: 1.26398616, Log Avg loss: 1.29071702, Global Avg Loss: 7.53327527, Time: 0.0208 Steps: 4280, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000032, Sample Num: 512, Cur Loss: 1.50391316, Cur Avg Loss: 1.22022179, Log Avg loss: 1.12394019, Global Avg Loss: 7.51833509, Time: 0.0208 Steps: 4290, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000042, Sample Num: 672, Cur Loss: 0.93817943, Cur Avg Loss: 1.17798530, Log Avg loss: 1.04282854, Global Avg Loss: 7.50327577, Time: 0.0208 Steps: 4300, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000052, Sample Num: 832, Cur Loss: 1.84164596, Cur Avg Loss: 1.19014737, Log Avg loss: 1.24122804, Global Avg Loss: 7.48874666, Time: 0.0209 Steps: 4310, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000062, Sample Num: 992, Cur Loss: 1.00607252, Cur Avg Loss: 1.15988021, Log Avg loss: 1.00249099, Global Avg Loss: 7.47373218, Time: 0.0208 Steps: 4320, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000072, Sample Num: 1152, Cur Loss: 1.48372841, Cur Avg Loss: 1.15765311, Log Avg loss: 1.14384511, Global Avg Loss: 7.45911350, Time: 0.0208 Steps: 4330, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000082, Sample Num: 1312, Cur Loss: 1.67515349, Cur Avg Loss: 1.14479856, Log Avg loss: 1.05224580, Global Avg Loss: 7.44435113, Time: 0.0209 Steps: 4340, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000092, Sample Num: 1472, Cur Loss: 0.30743700, Cur Avg Loss: 1.16112318, Log Avg loss: 1.29498504, Global Avg Loss: 7.43021466, Time: 0.0209 Steps: 4350, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000102, Sample Num: 1632, Cur Loss: 1.70784330, Cur Avg Loss: 1.18073305, Log Avg loss: 1.36114382, Global Avg Loss: 7.41629477, Time: 0.0208 Steps: 4360, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000112, Sample Num: 1792, Cur Loss: 1.09563875, Cur Avg Loss: 1.19120482, Log Avg loss: 1.29801689, Global Avg Loss: 7.40229414, Time: 0.0208 Steps: 4370, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000122, Sample Num: 1952, Cur Loss: 0.90850925, Cur Avg Loss: 1.19226640, Log Avg loss: 1.20415616, Global Avg Loss: 7.38814314, Time: 0.0208 Steps: 4380, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000132, Sample Num: 2112, Cur Loss: 1.51389766, Cur Avg Loss: 1.20497360, Log Avg loss: 1.36000144, Global Avg Loss: 7.37441161, Time: 0.0208 Steps: 4390, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000142, Sample Num: 2272, Cur Loss: 0.77153432, Cur Avg Loss: 1.20143586, Log Avg loss: 1.15473761, Global Avg Loss: 7.36027598, Time: 0.0208 Steps: 4400, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000152, Sample Num: 2432, Cur Loss: 1.48129702, Cur Avg Loss: 1.20093628, Log Avg loss: 1.19384223, Global Avg Loss: 7.34629314, Time: 0.0208 Steps: 4410, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000162, Sample Num: 2592, Cur Loss: 2.40838051, Cur Avg Loss: 1.23382647, Log Avg loss: 1.73375745, Global Avg Loss: 7.33359510, Time: 0.0208 Steps: 4420, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000172, Sample Num: 2752, Cur Loss: 0.72201133, Cur Avg Loss: 1.21822193, Log Avg loss: 0.96542833, Global Avg Loss: 7.31922000, Time: 0.0208 Steps: 4430, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000182, Sample Num: 2912, Cur Loss: 0.81277192, Cur Avg Loss: 1.21805304, Log Avg loss: 1.21514823, Global Avg Loss: 7.30547209, Time: 0.0208 Steps: 4440, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000192, Sample Num: 3072, Cur Loss: 1.53049994, Cur Avg Loss: 1.20590570, Log Avg loss: 0.98482397, Global Avg Loss: 7.29126839, Time: 0.0208 Steps: 4450, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000202, Sample Num: 3232, Cur Loss: 1.44446027, Cur Avg Loss: 1.22037156, Log Avg loss: 1.49811606, Global Avg Loss: 7.27827926, Time: 0.0209 Steps: 4460, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000212, Sample Num: 3392, Cur Loss: 0.40459338, Cur Avg Loss: 1.20385772, Log Avg loss: 0.87027831, Global Avg Loss: 7.26394369, Time: 0.0209 Steps: 4470, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000222, Sample Num: 3552, Cur Loss: 0.86559027, Cur Avg Loss: 1.19980987, Log Avg loss: 1.11399541, Global Avg Loss: 7.25021612, Time: 0.0208 Steps: 4480, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000232, Sample Num: 3712, Cur Loss: 1.55921602, Cur Avg Loss: 1.19690486, Log Avg loss: 1.13241347, Global Avg Loss: 7.23659073, Time: 0.0208 Steps: 4490, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000242, Sample Num: 3872, Cur Loss: 0.73245227, Cur Avg Loss: 1.18286322, Log Avg loss: 0.85709723, Global Avg Loss: 7.22241408, Time: 0.0209 Steps: 4500, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000252, Sample Num: 4032, Cur Loss: 0.79504138, Cur Avg Loss: 1.17142412, Log Avg loss: 0.89459787, Global Avg Loss: 7.20838344, Time: 0.0208 Steps: 4510, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000262, Sample Num: 4192, Cur Loss: 0.61242318, Cur Avg Loss: 1.16330724, Log Avg loss: 0.95876199, Global Avg Loss: 7.19455684, Time: 0.0246 Steps: 4520, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000272, Sample Num: 4352, Cur Loss: 0.88561654, Cur Avg Loss: 1.16513035, Log Avg loss: 1.21289565, Global Avg Loss: 7.18135229, Time: 0.0208 Steps: 4530, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000282, Sample Num: 4512, Cur Loss: 0.58112109, Cur Avg Loss: 1.15920128, Log Avg loss: 0.99793059, Global Avg Loss: 7.16773242, Time: 0.0208 Steps: 4540, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000292, Sample Num: 4672, Cur Loss: 1.16204786, Cur Avg Loss: 1.15845851, Log Avg loss: 1.13751241, Global Avg Loss: 7.15447919, Time: 0.0208 Steps: 4550, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000302, Sample Num: 4832, Cur Loss: 1.54377913, Cur Avg Loss: 1.15161492, Log Avg loss: 0.95178231, Global Avg Loss: 7.14087679, Time: 0.0208 Steps: 4560, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000312, Sample Num: 4992, Cur Loss: 0.67099428, Cur Avg Loss: 1.15215816, Log Avg loss: 1.16856385, Global Avg Loss: 7.12780827, Time: 0.0208 Steps: 4570, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000322, Sample Num: 5152, Cur Loss: 1.04317927, Cur Avg Loss: 1.15262487, Log Avg loss: 1.16718621, Global Avg Loss: 7.11479381, Time: 0.0209 Steps: 4580, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000332, Sample Num: 5312, Cur Loss: 1.13590062, Cur Avg Loss: 1.15629768, Log Avg loss: 1.27456213, Global Avg Loss: 7.10206999, Time: 0.0209 Steps: 4590, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000342, Sample Num: 5472, Cur Loss: 0.47456574, Cur Avg Loss: 1.15953720, Log Avg loss: 1.26708945, Global Avg Loss: 7.08938525, Time: 0.0209 Steps: 4600, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000352, Sample Num: 5632, Cur Loss: 1.28537941, Cur Avg Loss: 1.15751709, Log Avg loss: 1.08842908, Global Avg Loss: 7.07636799, Time: 0.0208 Steps: 4610, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000362, Sample Num: 5792, Cur Loss: 1.08770466, Cur Avg Loss: 1.16011993, Log Avg loss: 1.25174011, Global Avg Loss: 7.06376057, Time: 0.0208 Steps: 4620, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000372, Sample Num: 5952, Cur Loss: 0.65484923, Cur Avg Loss: 1.16005435, Log Avg loss: 1.15768032, Global Avg Loss: 7.05100446, Time: 0.0209 Steps: 4630, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000382, Sample Num: 6112, Cur Loss: 1.86146045, Cur Avg Loss: 1.16676227, Log Avg loss: 1.41629697, Global Avg Loss: 7.03886070, Time: 0.0209 Steps: 4640, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000392, Sample Num: 6272, Cur Loss: 1.59548521, Cur Avg Loss: 1.16904630, Log Avg loss: 1.25629627, Global Avg Loss: 7.02642507, Time: 0.0209 Steps: 4650, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000402, Sample Num: 6432, Cur Loss: 1.31371534, Cur Avg Loss: 1.16747942, Log Avg loss: 1.10605747, Global Avg Loss: 7.01372042, Time: 0.0209 Steps: 4660, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000412, Sample Num: 6592, Cur Loss: 1.24262643, Cur Avg Loss: 1.17097222, Log Avg loss: 1.31138299, Global Avg Loss: 7.00150985, Time: 0.0208 Steps: 4670, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000422, Sample Num: 6752, Cur Loss: 0.69901502, Cur Avg Loss: 1.16424631, Log Avg loss: 0.88713866, Global Avg Loss: 6.98844495, Time: 0.0208 Steps: 4680, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000432, Sample Num: 6912, Cur Loss: 0.97121245, Cur Avg Loss: 1.16354003, Log Avg loss: 1.13373492, Global Avg Loss: 6.97596156, Time: 0.0208 Steps: 4690, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000442, Sample Num: 7072, Cur Loss: 0.78682631, Cur Avg Loss: 1.16633435, Log Avg loss: 1.28704901, Global Avg Loss: 6.96385749, Time: 0.0208 Steps: 4700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000452, Sample Num: 7232, Cur Loss: 0.57946742, Cur Avg Loss: 1.16865772, Log Avg loss: 1.27135098, Global Avg Loss: 6.95177149, Time: 0.0209 Steps: 4710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000462, Sample Num: 7392, Cur Loss: 1.02598143, Cur Avg Loss: 1.16980366, Log Avg loss: 1.22159987, Global Avg Loss: 6.93963130, Time: 0.0208 Steps: 4720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000472, Sample Num: 7552, Cur Loss: 0.58768481, Cur Avg Loss: 1.16773028, Log Avg loss: 1.07194036, Global Avg Loss: 6.92722603, Time: 0.0208 Steps: 4730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000482, Sample Num: 7712, Cur Loss: 1.91140950, Cur Avg Loss: 1.17160953, Log Avg loss: 1.35470999, Global Avg Loss: 6.91546967, Time: 0.0208 Steps: 4740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000492, Sample Num: 7872, Cur Loss: 1.11163807, Cur Avg Loss: 1.17238305, Log Avg loss: 1.20966648, Global Avg Loss: 6.90345745, Time: 0.0209 Steps: 4750, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000502, Sample Num: 8032, Cur Loss: 1.17378366, Cur Avg Loss: 1.16540534, Log Avg loss: 0.82210216, Global Avg Loss: 6.89068150, Time: 0.0208 Steps: 4760, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000512, Sample Num: 8192, Cur Loss: 0.33241019, Cur Avg Loss: 1.16219291, Log Avg loss: 1.00092903, Global Avg Loss: 6.87833401, Time: 0.0253 Steps: 4770, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000522, Sample Num: 8352, Cur Loss: 1.27412176, Cur Avg Loss: 1.17073329, Log Avg loss: 1.60800065, Global Avg Loss: 6.86730820, Time: 0.0208 Steps: 4780, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000532, Sample Num: 8512, Cur Loss: 1.01660633, Cur Avg Loss: 1.16603571, Log Avg loss: 0.92082236, Global Avg Loss: 6.85489383, Time: 0.0208 Steps: 4790, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000542, Sample Num: 8672, Cur Loss: 1.31913459, Cur Avg Loss: 1.17276670, Log Avg loss: 1.53085523, Global Avg Loss: 6.84380208, Time: 0.0209 Steps: 4800, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000552, Sample Num: 8832, Cur Loss: 1.01241684, Cur Avg Loss: 1.17473670, Log Avg loss: 1.28151054, Global Avg Loss: 6.83223807, Time: 0.0208 Steps: 4810, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000562, Sample Num: 8992, Cur Loss: 1.01686549, Cur Avg Loss: 1.17142018, Log Avg loss: 0.98834847, Global Avg Loss: 6.82011381, Time: 0.0208 Steps: 4820, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000572, Sample Num: 9152, Cur Loss: 0.51367044, Cur Avg Loss: 1.16628404, Log Avg loss: 0.87763271, Global Avg Loss: 6.80781054, Time: 0.0209 Steps: 4830, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000582, Sample Num: 9312, Cur Loss: 2.73199344, Cur Avg Loss: 1.16782045, Log Avg loss: 1.25570303, Global Avg Loss: 6.79633924, Time: 0.0208 Steps: 4840, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000592, Sample Num: 9472, Cur Loss: 2.44396687, Cur Avg Loss: 1.16330977, Log Avg loss: 0.90078857, Global Avg Loss: 6.78418347, Time: 0.0208 Steps: 4850, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000602, Sample Num: 9632, Cur Loss: 1.01365530, Cur Avg Loss: 1.16299223, Log Avg loss: 1.14419371, Global Avg Loss: 6.77257855, Time: 0.0208 Steps: 4860, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000612, Sample Num: 9792, Cur Loss: 1.00834060, Cur Avg Loss: 1.16255223, Log Avg loss: 1.13606405, Global Avg Loss: 6.76100460, Time: 0.0208 Steps: 4870, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000622, Sample Num: 9952, Cur Loss: 0.84003651, Cur Avg Loss: 1.16185168, Log Avg loss: 1.11897798, Global Avg Loss: 6.74944307, Time: 0.0208 Steps: 4880, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000632, Sample Num: 10112, Cur Loss: 1.58197796, Cur Avg Loss: 1.16099800, Log Avg loss: 1.10789903, Global Avg Loss: 6.73790617, Time: 0.0209 Steps: 4890, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000642, Sample Num: 10272, Cur Loss: 1.11590207, Cur Avg Loss: 1.15878452, Log Avg loss: 1.01889272, Global Avg Loss: 6.72623471, Time: 0.0208 Steps: 4900, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000652, Sample Num: 10432, Cur Loss: 0.99438894, Cur Avg Loss: 1.16030795, Log Avg loss: 1.25811243, Global Avg Loss: 6.71509801, Time: 0.0208 Steps: 4910, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000662, Sample Num: 10592, Cur Loss: 0.88330591, Cur Avg Loss: 1.15634587, Log Avg loss: 0.89801816, Global Avg Loss: 6.70327468, Time: 0.0209 Steps: 4920, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000672, Sample Num: 10752, Cur Loss: 1.37490249, Cur Avg Loss: 1.15868177, Log Avg loss: 1.31331843, Global Avg Loss: 6.69234170, Time: 0.0208 Steps: 4930, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000682, Sample Num: 10912, Cur Loss: 0.58210427, Cur Avg Loss: 1.15746795, Log Avg loss: 1.07589933, Global Avg Loss: 6.68097239, Time: 0.0208 Steps: 4940, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000692, Sample Num: 11072, Cur Loss: 0.37357634, Cur Avg Loss: 1.15317309, Log Avg loss: 0.86026344, Global Avg Loss: 6.66921338, Time: 0.0208 Steps: 4950, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000702, Sample Num: 11232, Cur Loss: 0.91897595, Cur Avg Loss: 1.14762105, Log Avg loss: 0.76341961, Global Avg Loss: 6.65730654, Time: 0.0209 Steps: 4960, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000712, Sample Num: 11392, Cur Loss: 1.26072824, Cur Avg Loss: 1.14799490, Log Avg loss: 1.17423923, Global Avg Loss: 6.64627421, Time: 0.0208 Steps: 4970, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000722, Sample Num: 11552, Cur Loss: 1.32802808, Cur Avg Loss: 1.14831969, Log Avg loss: 1.17144498, Global Avg Loss: 6.63528057, Time: 0.0208 Steps: 4980, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000732, Sample Num: 11712, Cur Loss: 2.03509331, Cur Avg Loss: 1.14858918, Log Avg loss: 1.16804660, Global Avg Loss: 6.62432419, Time: 0.0208 Steps: 4990, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000742, Sample Num: 11872, Cur Loss: 1.21672726, Cur Avg Loss: 1.14610237, Log Avg loss: 0.96406761, Global Avg Loss: 6.61300368, Time: 0.0209 Steps: 5000, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000752, Sample Num: 12032, Cur Loss: 1.82055426, Cur Avg Loss: 1.14610392, Log Avg loss: 1.14621916, Global Avg Loss: 6.60209193, Time: 0.0208 Steps: 5010, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000762, Sample Num: 12192, Cur Loss: 2.45887637, Cur Avg Loss: 1.14629706, Log Avg loss: 1.16082115, Global Avg Loss: 6.59125275, Time: 0.0208 Steps: 5020, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000772, Sample Num: 12352, Cur Loss: 1.47679377, Cur Avg Loss: 1.15374374, Log Avg loss: 1.72118083, Global Avg Loss: 6.58157070, Time: 0.0243 Steps: 5030, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000782, Sample Num: 12512, Cur Loss: 1.55239046, Cur Avg Loss: 1.15482837, Log Avg loss: 1.23856153, Global Avg Loss: 6.57096949, Time: 0.0209 Steps: 5040, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000792, Sample Num: 12672, Cur Loss: 1.09986830, Cur Avg Loss: 1.15429524, Log Avg loss: 1.11260408, Global Avg Loss: 6.56016084, Time: 0.0209 Steps: 5050, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000802, Sample Num: 12832, Cur Loss: 0.62877184, Cur Avg Loss: 1.15189975, Log Avg loss: 0.96217705, Global Avg Loss: 6.54909764, Time: 0.0208 Steps: 5060, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000812, Sample Num: 12992, Cur Loss: 1.61594343, Cur Avg Loss: 1.15222125, Log Avg loss: 1.17800564, Global Avg Loss: 6.53850377, Time: 0.0209 Steps: 5070, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000822, Sample Num: 13152, Cur Loss: 1.18585849, Cur Avg Loss: 1.15250197, Log Avg loss: 1.17529673, Global Avg Loss: 6.52794627, Time: 0.0209 Steps: 5080, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000832, Sample Num: 13312, Cur Loss: 0.26272112, Cur Avg Loss: 1.15089452, Log Avg loss: 1.01876188, Global Avg Loss: 6.51712273, Time: 0.0210 Steps: 5090, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000842, Sample Num: 13472, Cur Loss: 1.28408051, Cur Avg Loss: 1.15318895, Log Avg loss: 1.34408539, Global Avg Loss: 6.50697952, Time: 0.0209 Steps: 5100, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000852, Sample Num: 13632, Cur Loss: 0.97679341, Cur Avg Loss: 1.15291827, Log Avg loss: 1.13012688, Global Avg Loss: 6.49645730, Time: 0.0209 Steps: 5110, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000862, Sample Num: 13792, Cur Loss: 0.48014456, Cur Avg Loss: 1.15296016, Log Avg loss: 1.15652959, Global Avg Loss: 6.48602775, Time: 0.0208 Steps: 5120, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000872, Sample Num: 13952, Cur Loss: 0.98366058, Cur Avg Loss: 1.15354139, Log Avg loss: 1.20364307, Global Avg Loss: 6.47573071, Time: 0.0209 Steps: 5130, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000882, Sample Num: 14112, Cur Loss: 0.46777180, Cur Avg Loss: 1.15143083, Log Avg loss: 0.96739060, Global Avg Loss: 6.46501409, Time: 0.0208 Steps: 5140, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000892, Sample Num: 14272, Cur Loss: 0.98152339, Cur Avg Loss: 1.14836697, Log Avg loss: 0.87813441, Global Avg Loss: 6.45416578, Time: 0.0209 Steps: 5150, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000902, Sample Num: 14432, Cur Loss: 1.62869596, Cur Avg Loss: 1.15021856, Log Avg loss: 1.31538049, Global Avg Loss: 6.44420690, Time: 0.0209 Steps: 5160, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000912, Sample Num: 14592, Cur Loss: 0.54783249, Cur Avg Loss: 1.14929358, Log Avg loss: 1.06586036, Global Avg Loss: 6.43380390, Time: 0.0208 Steps: 5170, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000922, Sample Num: 14752, Cur Loss: 0.71054184, Cur Avg Loss: 1.14705668, Log Avg loss: 0.94305103, Global Avg Loss: 6.42320400, Time: 0.0208 Steps: 5180, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000932, Sample Num: 14912, Cur Loss: 0.90876228, Cur Avg Loss: 1.14631728, Log Avg loss: 1.07814462, Global Avg Loss: 6.41290523, Time: 0.0208 Steps: 5190, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000942, Sample Num: 15072, Cur Loss: 1.38282001, Cur Avg Loss: 1.14617482, Log Avg loss: 1.13289768, Global Avg Loss: 6.40275137, Time: 0.0209 Steps: 5200, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000952, Sample Num: 15232, Cur Loss: 1.41639721, Cur Avg Loss: 1.14481687, Log Avg loss: 1.01689815, Global Avg Loss: 6.39241384, Time: 0.0208 Steps: 5210, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000962, Sample Num: 15392, Cur Loss: 1.02970779, Cur Avg Loss: 1.14553570, Log Avg loss: 1.21396775, Global Avg Loss: 6.38249344, Time: 0.0208 Steps: 5220, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000972, Sample Num: 15552, Cur Loss: 0.73576617, Cur Avg Loss: 1.14292527, Log Avg loss: 0.89180225, Global Avg Loss: 6.37199499, Time: 0.0209 Steps: 5230, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000982, Sample Num: 15712, Cur Loss: 0.66671473, Cur Avg Loss: 1.14076286, Log Avg loss: 0.93057656, Global Avg Loss: 6.36161060, Time: 0.0208 Steps: 5240, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000992, Sample Num: 15872, Cur Loss: 1.20172858, Cur Avg Loss: 1.14232793, Log Avg loss: 1.29601795, Global Avg Loss: 6.35196186, Time: 0.0208 Steps: 5250, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001002, Sample Num: 16032, Cur Loss: 0.42762005, Cur Avg Loss: 1.14440445, Log Avg loss: 1.35039499, Global Avg Loss: 6.34245317, Time: 0.0208 Steps: 5260, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001012, Sample Num: 16192, Cur Loss: 1.06015003, Cur Avg Loss: 1.14417172, Log Avg loss: 1.12085195, Global Avg Loss: 6.33254501, Time: 0.0208 Steps: 5270, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001022, Sample Num: 16352, Cur Loss: 0.41842788, Cur Avg Loss: 1.14240065, Log Avg loss: 0.96316882, Global Avg Loss: 6.32237574, Time: 0.0208 Steps: 5280, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001032, Sample Num: 16512, Cur Loss: 1.14294851, Cur Avg Loss: 1.14323810, Log Avg loss: 1.22882529, Global Avg Loss: 6.31274710, Time: 0.0208 Steps: 5290, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001042, Sample Num: 16672, Cur Loss: 0.83838153, Cur Avg Loss: 1.13987000, Log Avg loss: 0.79228181, Global Avg Loss: 6.30233113, Time: 0.0208 Steps: 5300, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001052, Sample Num: 16832, Cur Loss: 0.76082128, Cur Avg Loss: 1.13996570, Log Avg loss: 1.14993769, Global Avg Loss: 6.29262794, Time: 0.0208 Steps: 5310, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001062, Sample Num: 16992, Cur Loss: 2.40288591, Cur Avg Loss: 1.14115419, Log Avg loss: 1.26618340, Global Avg Loss: 6.28317973, Time: 0.0208 Steps: 5320, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001072, Sample Num: 17152, Cur Loss: 0.76105708, Cur Avg Loss: 1.14055938, Log Avg loss: 1.07739058, Global Avg Loss: 6.27341278, Time: 0.0208 Steps: 5330, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001082, Sample Num: 17312, Cur Loss: 2.05313253, Cur Avg Loss: 1.14102585, Log Avg loss: 1.19103125, Global Avg Loss: 6.26389521, Time: 0.0208 Steps: 5340, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001092, Sample Num: 17472, Cur Loss: 1.12529206, Cur Avg Loss: 1.14033824, Log Avg loss: 1.06593950, Global Avg Loss: 6.25417940, Time: 0.0208 Steps: 5350, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001102, Sample Num: 17632, Cur Loss: 1.44601059, Cur Avg Loss: 1.13832734, Log Avg loss: 0.91873677, Global Avg Loss: 6.24422522, Time: 0.0207 Steps: 5360, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001112, Sample Num: 17792, Cur Loss: 0.72917086, Cur Avg Loss: 1.13580110, Log Avg loss: 0.85740966, Global Avg Loss: 6.23419390, Time: 0.0208 Steps: 5370, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001122, Sample Num: 17952, Cur Loss: 0.97575867, Cur Avg Loss: 1.13563481, Log Avg loss: 1.11714380, Global Avg Loss: 6.22468266, Time: 0.0208 Steps: 5380, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001132, Sample Num: 18112, Cur Loss: 2.15531111, Cur Avg Loss: 1.14078252, Log Avg loss: 1.71835456, Global Avg Loss: 6.21632212, Time: 0.0208 Steps: 5390, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001142, Sample Num: 18272, Cur Loss: 0.50116169, Cur Avg Loss: 1.14218078, Log Avg loss: 1.30046483, Global Avg Loss: 6.20721868, Time: 0.0209 Steps: 5400, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001152, Sample Num: 18432, Cur Loss: 1.03684688, Cur Avg Loss: 1.14063011, Log Avg loss: 0.96354332, Global Avg Loss: 6.19752612, Time: 0.0208 Steps: 5410, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001162, Sample Num: 18592, Cur Loss: 1.12817931, Cur Avg Loss: 1.14301778, Log Avg loss: 1.41807691, Global Avg Loss: 6.18870795, Time: 0.0208 Steps: 5420, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001172, Sample Num: 18752, Cur Loss: 1.04068720, Cur Avg Loss: 1.14018662, Log Avg loss: 0.81120567, Global Avg Loss: 6.17880463, Time: 0.0208 Steps: 5430, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001182, Sample Num: 18912, Cur Loss: 1.02861142, Cur Avg Loss: 1.14109245, Log Avg loss: 1.24725601, Global Avg Loss: 6.16973929, Time: 0.0208 Steps: 5440, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001192, Sample Num: 19072, Cur Loss: 1.08155358, Cur Avg Loss: 1.14070337, Log Avg loss: 1.09471475, Global Avg Loss: 6.16042731, Time: 0.0208 Steps: 5450, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001202, Sample Num: 19232, Cur Loss: 0.89543027, Cur Avg Loss: 1.14076642, Log Avg loss: 1.14828170, Global Avg Loss: 6.15124756, Time: 0.0209 Steps: 5460, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001212, Sample Num: 19392, Cur Loss: 1.58557415, Cur Avg Loss: 1.14084095, Log Avg loss: 1.14979901, Global Avg Loss: 6.14210414, Time: 0.0208 Steps: 5470, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001222, Sample Num: 19552, Cur Loss: 0.55252206, Cur Avg Loss: 1.13911391, Log Avg loss: 0.92979719, Global Avg Loss: 6.13259264, Time: 0.0210 Steps: 5480, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001232, Sample Num: 19712, Cur Loss: 0.76209819, Cur Avg Loss: 1.13980547, Log Avg loss: 1.22431412, Global Avg Loss: 6.12365224, Time: 0.0209 Steps: 5490, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001242, Sample Num: 19872, Cur Loss: 1.55755520, Cur Avg Loss: 1.13883406, Log Avg loss: 1.01915548, Global Avg Loss: 6.11437133, Time: 0.0209 Steps: 5500, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001252, Sample Num: 20032, Cur Loss: 0.90412116, Cur Avg Loss: 1.13708026, Log Avg loss: 0.91925834, Global Avg Loss: 6.10494282, Time: 0.0208 Steps: 5510, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001262, Sample Num: 20192, Cur Loss: 2.05555582, Cur Avg Loss: 1.13924053, Log Avg loss: 1.40970727, Global Avg Loss: 6.09643696, Time: 0.0209 Steps: 5520, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001272, Sample Num: 20352, Cur Loss: 1.72245979, Cur Avg Loss: 1.13852869, Log Avg loss: 1.04869395, Global Avg Loss: 6.08730903, Time: 0.0208 Steps: 5530, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001282, Sample Num: 20512, Cur Loss: 0.56130797, Cur Avg Loss: 1.13656800, Log Avg loss: 0.88716806, Global Avg Loss: 6.07792249, Time: 0.0247 Steps: 5540, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001292, Sample Num: 20672, Cur Loss: 0.73947096, Cur Avg Loss: 1.13469417, Log Avg loss: 0.89446982, Global Avg Loss: 6.06858294, Time: 0.0209 Steps: 5550, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001302, Sample Num: 20832, Cur Loss: 0.52078545, Cur Avg Loss: 1.13384782, Log Avg loss: 1.02449858, Global Avg Loss: 6.05951085, Time: 0.0209 Steps: 5560, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001312, Sample Num: 20992, Cur Loss: 0.67065567, Cur Avg Loss: 1.13508317, Log Avg loss: 1.29592649, Global Avg Loss: 6.05095863, Time: 0.0209 Steps: 5570, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001322, Sample Num: 21152, Cur Loss: 1.22150195, Cur Avg Loss: 1.13347024, Log Avg loss: 0.92185331, Global Avg Loss: 6.04176668, Time: 0.0209 Steps: 5580, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001332, Sample Num: 21312, Cur Loss: 1.87068510, Cur Avg Loss: 1.13346413, Log Avg loss: 1.13265631, Global Avg Loss: 6.03298473, Time: 0.0209 Steps: 5590, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001342, Sample Num: 21472, Cur Loss: 1.83560824, Cur Avg Loss: 1.13290747, Log Avg loss: 1.05876102, Global Avg Loss: 6.02410219, Time: 0.0209 Steps: 5600, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001352, Sample Num: 21632, Cur Loss: 1.16338396, Cur Avg Loss: 1.13318736, Log Avg loss: 1.17074861, Global Avg Loss: 6.01545094, Time: 0.0209 Steps: 5610, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001362, Sample Num: 21792, Cur Loss: 1.36581969, Cur Avg Loss: 1.13304663, Log Avg loss: 1.11402004, Global Avg Loss: 6.00672953, Time: 0.0209 Steps: 5620, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001372, Sample Num: 21952, Cur Loss: 0.63953996, Cur Avg Loss: 1.13316518, Log Avg loss: 1.14931062, Global Avg Loss: 5.99810179, Time: 0.0209 Steps: 5630, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001382, Sample Num: 22112, Cur Loss: 1.68971729, Cur Avg Loss: 1.13232921, Log Avg loss: 1.01763505, Global Avg Loss: 5.98927117, Time: 0.0209 Steps: 5640, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001392, Sample Num: 22272, Cur Loss: 1.82164288, Cur Avg Loss: 1.13283749, Log Avg loss: 1.20308176, Global Avg Loss: 5.98080004, Time: 0.0209 Steps: 5650, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001402, Sample Num: 22432, Cur Loss: 1.84366632, Cur Avg Loss: 1.13319929, Log Avg loss: 1.18356095, Global Avg Loss: 5.97232435, Time: 0.0209 Steps: 5660, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001412, Sample Num: 22592, Cur Loss: 0.91291082, Cur Avg Loss: 1.13406824, Log Avg loss: 1.25589485, Global Avg Loss: 5.96400614, Time: 0.0209 Steps: 5670, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001422, Sample Num: 22752, Cur Loss: 0.62180227, Cur Avg Loss: 1.13419470, Log Avg loss: 1.15205094, Global Avg Loss: 5.95553438, Time: 0.0209 Steps: 5680, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001432, Sample Num: 22912, Cur Loss: 0.22760007, Cur Avg Loss: 1.13274646, Log Avg loss: 0.92680755, Global Avg Loss: 5.94669655, Time: 0.0209 Steps: 5690, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001442, Sample Num: 23072, Cur Loss: 1.26051950, Cur Avg Loss: 1.13441777, Log Avg loss: 1.37374967, Global Avg Loss: 5.93867384, Time: 0.0210 Steps: 5700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001452, Sample Num: 23232, Cur Loss: 0.15122803, Cur Avg Loss: 1.13462887, Log Avg loss: 1.16506874, Global Avg Loss: 5.93031376, Time: 0.0209 Steps: 5710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001462, Sample Num: 23392, Cur Loss: 1.69537592, Cur Avg Loss: 1.13235033, Log Avg loss: 0.80150699, Global Avg Loss: 5.92134731, Time: 0.0215 Steps: 5720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001472, Sample Num: 23552, Cur Loss: 2.01990390, Cur Avg Loss: 1.13307829, Log Avg loss: 1.23950532, Global Avg Loss: 5.91317656, Time: 0.0209 Steps: 5730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001482, Sample Num: 23712, Cur Loss: 0.50714284, Cur Avg Loss: 1.12950850, Log Avg loss: 0.60403639, Global Avg Loss: 5.90392719, Time: 0.0209 Steps: 5740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001492, Sample Num: 23872, Cur Loss: 0.62777269, Cur Avg Loss: 1.12726464, Log Avg loss: 0.79472457, Global Avg Loss: 5.89504162, Time: 0.0209 Steps: 5750, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001502, Sample Num: 24032, Cur Loss: 0.52209246, Cur Avg Loss: 1.12726059, Log Avg loss: 1.12665537, Global Avg Loss: 5.88676317, Time: 0.0209 Steps: 5760, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001512, Sample Num: 24192, Cur Loss: 0.59685612, Cur Avg Loss: 1.12716322, Log Avg loss: 1.11253854, Global Avg Loss: 5.87848895, Time: 0.0209 Steps: 5770, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001522, Sample Num: 24352, Cur Loss: 0.84739155, Cur Avg Loss: 1.12703027, Log Avg loss: 1.10692832, Global Avg Loss: 5.87023365, Time: 0.0209 Steps: 5780, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001532, Sample Num: 24512, Cur Loss: 1.61914313, Cur Avg Loss: 1.12560260, Log Avg loss: 0.90831148, Global Avg Loss: 5.86166384, Time: 0.0209 Steps: 5790, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001542, Sample Num: 24672, Cur Loss: 0.85592198, Cur Avg Loss: 1.12526351, Log Avg loss: 1.07331437, Global Avg Loss: 5.85340806, Time: 0.0210 Steps: 5800, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001552, Sample Num: 24832, Cur Loss: 0.93187416, Cur Avg Loss: 1.12353585, Log Avg loss: 0.85713130, Global Avg Loss: 5.84480862, Time: 0.0209 Steps: 5810, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001562, Sample Num: 24992, Cur Loss: 1.09399939, Cur Avg Loss: 1.12488330, Log Avg loss: 1.33400654, Global Avg Loss: 5.83705810, Time: 0.0209 Steps: 5820, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001572, Sample Num: 25152, Cur Loss: 0.50767148, Cur Avg Loss: 1.12412317, Log Avg loss: 1.00539181, Global Avg Loss: 5.82877051, Time: 0.0209 Steps: 5830, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001582, Sample Num: 25312, Cur Loss: 0.67118907, Cur Avg Loss: 1.12357135, Log Avg loss: 1.03682436, Global Avg Loss: 5.82056512, Time: 0.0208 Steps: 5840, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001592, Sample Num: 25472, Cur Loss: 0.69067144, Cur Avg Loss: 1.12260468, Log Avg loss: 0.96967866, Global Avg Loss: 5.81227301, Time: 0.0208 Steps: 5850, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001602, Sample Num: 25632, Cur Loss: 0.85865623, Cur Avg Loss: 1.12266956, Log Avg loss: 1.13299792, Global Avg Loss: 5.80428790, Time: 0.0208 Steps: 5860, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001612, Sample Num: 25792, Cur Loss: 1.84790492, Cur Avg Loss: 1.12161825, Log Avg loss: 0.95319754, Global Avg Loss: 5.79602369, Time: 0.0209 Steps: 5870, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001622, Sample Num: 25952, Cur Loss: 0.89856786, Cur Avg Loss: 1.11991674, Log Avg loss: 0.84563422, Global Avg Loss: 5.78760466, Time: 0.0209 Steps: 5880, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001632, Sample Num: 26112, Cur Loss: 0.79343867, Cur Avg Loss: 1.11797448, Log Avg loss: 0.80294053, Global Avg Loss: 5.77914173, Time: 0.0209 Steps: 5890, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001642, Sample Num: 26272, Cur Loss: 1.28318512, Cur Avg Loss: 1.11740550, Log Avg loss: 1.02454695, Global Avg Loss: 5.77108310, Time: 0.0208 Steps: 5900, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001652, Sample Num: 26432, Cur Loss: 0.99296653, Cur Avg Loss: 1.11690519, Log Avg loss: 1.03475442, Global Avg Loss: 5.76306900, Time: 0.0209 Steps: 5910, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001662, Sample Num: 26592, Cur Loss: 1.26981187, Cur Avg Loss: 1.11657927, Log Avg loss: 1.06273783, Global Avg Loss: 5.75512926, Time: 0.0209 Steps: 5920, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001672, Sample Num: 26752, Cur Loss: 0.76967120, Cur Avg Loss: 1.11596138, Log Avg loss: 1.01326735, Global Avg Loss: 5.74713286, Time: 0.0209 Steps: 5930, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001682, Sample Num: 26912, Cur Loss: 0.52562195, Cur Avg Loss: 1.11534306, Log Avg loss: 1.01195939, Global Avg Loss: 5.73916119, Time: 0.0210 Steps: 5940, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001692, Sample Num: 27072, Cur Loss: 0.25159955, Cur Avg Loss: 1.11552318, Log Avg loss: 1.14582063, Global Avg Loss: 5.73144129, Time: 0.0209 Steps: 5950, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001702, Sample Num: 27232, Cur Loss: 0.64091456, Cur Avg Loss: 1.11519383, Log Avg loss: 1.05946817, Global Avg Loss: 5.72360241, Time: 0.0209 Steps: 5960, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001712, Sample Num: 27392, Cur Loss: 1.41436017, Cur Avg Loss: 1.11331109, Log Avg loss: 0.79286816, Global Avg Loss: 5.71534322, Time: 0.0209 Steps: 5970, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001722, Sample Num: 27552, Cur Loss: 0.51790786, Cur Avg Loss: 1.11263203, Log Avg loss: 0.99637607, Global Avg Loss: 5.70745197, Time: 0.0208 Steps: 5980, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001732, Sample Num: 27712, Cur Loss: 1.32143557, Cur Avg Loss: 1.11112205, Log Avg loss: 0.85110498, Global Avg Loss: 5.69934455, Time: 0.0209 Steps: 5990, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001742, Sample Num: 27872, Cur Loss: 1.52020228, Cur Avg Loss: 1.11201529, Log Avg loss: 1.26672277, Global Avg Loss: 5.69195684, Time: 0.0209 Steps: 6000, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001752, Sample Num: 28032, Cur Loss: 1.44721174, Cur Avg Loss: 1.11261186, Log Avg loss: 1.21653486, Global Avg Loss: 5.68451022, Time: 0.0209 Steps: 6010, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001762, Sample Num: 28192, Cur Loss: 0.70885003, Cur Avg Loss: 1.11259862, Log Avg loss: 1.11027848, Global Avg Loss: 5.67691183, Time: 0.0209 Steps: 6020, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001772, Sample Num: 28352, Cur Loss: 1.08889008, Cur Avg Loss: 1.11174464, Log Avg loss: 0.96127381, Global Avg Loss: 5.66909153, Time: 0.0209 Steps: 6030, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001782, Sample Num: 28512, Cur Loss: 0.40508497, Cur Avg Loss: 1.11067313, Log Avg loss: 0.92080120, Global Avg Loss: 5.66123012, Time: 0.0209 Steps: 6040, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001792, Sample Num: 28672, Cur Loss: 0.68580890, Cur Avg Loss: 1.10854089, Log Avg loss: 0.72857637, Global Avg Loss: 5.65307698, Time: 0.0254 Steps: 6050, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001802, Sample Num: 28832, Cur Loss: 1.40861988, Cur Avg Loss: 1.10858812, Log Avg loss: 1.11705126, Global Avg Loss: 5.64559179, Time: 0.0209 Steps: 6060, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001812, Sample Num: 28992, Cur Loss: 0.79132324, Cur Avg Loss: 1.10824490, Log Avg loss: 1.04639686, Global Avg Loss: 5.63801486, Time: 0.0209 Steps: 6070, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001822, Sample Num: 29152, Cur Loss: 1.26209354, Cur Avg Loss: 1.10854793, Log Avg loss: 1.16345656, Global Avg Loss: 5.63065539, Time: 0.0209 Steps: 6080, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001832, Sample Num: 29312, Cur Loss: 0.69936603, Cur Avg Loss: 1.10776046, Log Avg loss: 0.96428377, Global Avg Loss: 5.62299304, Time: 0.0209 Steps: 6090, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001842, Sample Num: 29472, Cur Loss: 1.52461100, Cur Avg Loss: 1.10639510, Log Avg loss: 0.85626230, Global Avg Loss: 5.61517872, Time: 0.0208 Steps: 6100, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001852, Sample Num: 29632, Cur Loss: 1.08327270, Cur Avg Loss: 1.10598517, Log Avg loss: 1.03047509, Global Avg Loss: 5.60767512, Time: 0.0209 Steps: 6110, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001862, Sample Num: 29792, Cur Loss: 1.25212049, Cur Avg Loss: 1.10403559, Log Avg loss: 0.74297276, Global Avg Loss: 5.59972626, Time: 0.0209 Steps: 6120, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001872, Sample Num: 29952, Cur Loss: 0.24923351, Cur Avg Loss: 1.10308771, Log Avg loss: 0.92659307, Global Avg Loss: 5.59210288, Time: 0.0208 Steps: 6130, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001882, Sample Num: 30112, Cur Loss: 1.51672900, Cur Avg Loss: 1.10276517, Log Avg loss: 1.04238684, Global Avg Loss: 5.58469292, Time: 0.0209 Steps: 6140, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001892, Sample Num: 30272, Cur Loss: 1.89888453, Cur Avg Loss: 1.10301462, Log Avg loss: 1.14996075, Global Avg Loss: 5.57748197, Time: 0.0209 Steps: 6150, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001902, Sample Num: 30432, Cur Loss: 0.33319825, Cur Avg Loss: 1.10254713, Log Avg loss: 1.01409730, Global Avg Loss: 5.57007388, Time: 0.0208 Steps: 6160, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001912, Sample Num: 30592, Cur Loss: 1.46816015, Cur Avg Loss: 1.10251386, Log Avg loss: 1.09618638, Global Avg Loss: 5.56282284, Time: 0.0208 Steps: 6170, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001922, Sample Num: 30752, Cur Loss: 1.35406828, Cur Avg Loss: 1.10164846, Log Avg loss: 0.93618340, Global Avg Loss: 5.55533637, Time: 0.0209 Steps: 6180, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001932, Sample Num: 30912, Cur Loss: 0.68990529, Cur Avg Loss: 1.10112479, Log Avg loss: 1.00047531, Global Avg Loss: 5.54797795, Time: 0.0209 Steps: 6190, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001942, Sample Num: 31072, Cur Loss: 0.87265712, Cur Avg Loss: 1.10050172, Log Avg loss: 0.98012557, Global Avg Loss: 5.54061045, Time: 0.0209 Steps: 6200, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001952, Sample Num: 31232, Cur Loss: 1.71448612, Cur Avg Loss: 1.10040614, Log Avg loss: 1.08184397, Global Avg Loss: 5.53343047, Time: 0.0209 Steps: 6210, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001962, Sample Num: 31392, Cur Loss: 0.89525938, Cur Avg Loss: 1.09884141, Log Avg loss: 0.79340554, Global Avg Loss: 5.52580985, Time: 0.0208 Steps: 6220, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001972, Sample Num: 31552, Cur Loss: 0.64359134, Cur Avg Loss: 1.09778175, Log Avg loss: 0.88987689, Global Avg Loss: 5.51836855, Time: 0.0209 Steps: 6230, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001982, Sample Num: 31712, Cur Loss: 0.54208553, Cur Avg Loss: 1.09806024, Log Avg loss: 1.15297842, Global Avg Loss: 5.51137273, Time: 0.0208 Steps: 6240, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001992, Sample Num: 31872, Cur Loss: 0.94450420, Cur Avg Loss: 1.10015416, Log Avg loss: 1.51516855, Global Avg Loss: 5.50497880, Time: 0.0210 Steps: 6250, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002002, Sample Num: 32032, Cur Loss: 1.21681762, Cur Avg Loss: 1.09985019, Log Avg loss: 1.03929938, Global Avg Loss: 5.49784513, Time: 0.0208 Steps: 6260, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002012, Sample Num: 32192, Cur Loss: 0.61840004, Cur Avg Loss: 1.10054913, Log Avg loss: 1.24047811, Global Avg Loss: 5.49105507, Time: 0.0209 Steps: 6270, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002022, Sample Num: 32352, Cur Loss: 1.02513599, Cur Avg Loss: 1.09926575, Log Avg loss: 0.84104899, Global Avg Loss: 5.48365060, Time: 0.0209 Steps: 6280, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002032, Sample Num: 32512, Cur Loss: 1.45391524, Cur Avg Loss: 1.09962534, Log Avg loss: 1.17233543, Global Avg Loss: 5.47679637, Time: 0.0208 Steps: 6290, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002042, Sample Num: 32672, Cur Loss: 1.01471364, Cur Avg Loss: 1.10000736, Log Avg loss: 1.17763366, Global Avg Loss: 5.46997230, Time: 0.0208 Steps: 6300, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002052, Sample Num: 32832, Cur Loss: 0.71235669, Cur Avg Loss: 1.09922876, Log Avg loss: 0.94023728, Global Avg Loss: 5.46279364, Time: 0.0246 Steps: 6310, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002062, Sample Num: 32992, Cur Loss: 0.82917202, Cur Avg Loss: 1.09782725, Log Avg loss: 0.81023841, Global Avg Loss: 5.45543200, Time: 0.0208 Steps: 6320, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002072, Sample Num: 33152, Cur Loss: 0.39136046, Cur Avg Loss: 1.09755040, Log Avg loss: 1.04046298, Global Avg Loss: 5.44845732, Time: 0.0208 Steps: 6330, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002082, Sample Num: 33312, Cur Loss: 0.48238307, Cur Avg Loss: 1.09647187, Log Avg loss: 0.87300216, Global Avg Loss: 5.44124052, Time: 0.0208 Steps: 6340, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002092, Sample Num: 33472, Cur Loss: 0.71072739, Cur Avg Loss: 1.09489224, Log Avg loss: 0.76601143, Global Avg Loss: 5.43387795, Time: 0.0209 Steps: 6350, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002102, Sample Num: 33632, Cur Loss: 0.23314998, Cur Avg Loss: 1.09266339, Log Avg loss: 0.62638915, Global Avg Loss: 5.42631901, Time: 0.0208 Steps: 6360, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002112, Sample Num: 33792, Cur Loss: 1.28956270, Cur Avg Loss: 1.09344446, Log Avg loss: 1.25762477, Global Avg Loss: 5.41977475, Time: 0.0208 Steps: 6370, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002122, Sample Num: 33952, Cur Loss: 1.17732584, Cur Avg Loss: 1.09324105, Log Avg loss: 1.05028108, Global Avg Loss: 5.41292601, Time: 0.0208 Steps: 6380, Updated lr: 0.000095 ***** Running evaluation checkpoint-6387 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-6387 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.518679, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.860858, "eval_total_loss": 605.182851, "eval_mae": 0.732194, "eval_mse": 0.860916, "eval_r2": 0.452745, "eval_sp_statistic": 0.686275, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.682767, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.107466, "test_total_loss": 555.947881, "test_mae": 0.904611, "test_mse": 1.107535, "test_r2": 0.285187, "test_sp_statistic": 0.581783, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.622675, "test_ps_pvalue": 0.0, "lr": 9.489141773352301e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 5.407806373415221, "train_cur_epoch_loss": 2325.0488662570715, "train_cur_epoch_avg_loss": 1.092084953620043, "train_cur_epoch_time": 44.51867938041687, "train_cur_epoch_avg_time": 0.02091060562725076, "epoch": 3, "step": 6387} ################################################## Training, Epoch: 0004, Batch: 000003, Sample Num: 48, Cur Loss: 0.60799301, Cur Avg Loss: 1.36949543, Log Avg loss: 0.92998450, Global Avg Loss: 5.40591045, Time: 0.0246 Steps: 6390, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000013, Sample Num: 208, Cur Loss: 0.40186185, Cur Avg Loss: 1.02012294, Log Avg loss: 0.91531119, Global Avg Loss: 5.39889389, Time: 0.0209 Steps: 6400, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000023, Sample Num: 368, Cur Loss: 0.44492018, Cur Avg Loss: 0.98061773, Log Avg loss: 0.92926096, Global Avg Loss: 5.39192099, Time: 0.0207 Steps: 6410, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000033, Sample Num: 528, Cur Loss: 0.83137119, Cur Avg Loss: 1.00184368, Log Avg loss: 1.05066336, Global Avg Loss: 5.38515890, Time: 0.0207 Steps: 6420, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000043, Sample Num: 688, Cur Loss: 1.24119997, Cur Avg Loss: 1.00042102, Log Avg loss: 0.99572622, Global Avg Loss: 5.37833241, Time: 0.0207 Steps: 6430, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000053, Sample Num: 848, Cur Loss: 1.35188282, Cur Avg Loss: 0.99988708, Log Avg loss: 0.99759115, Global Avg Loss: 5.37153002, Time: 0.0208 Steps: 6440, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000063, Sample Num: 1008, Cur Loss: 1.14735103, Cur Avg Loss: 0.98594403, Log Avg loss: 0.91204584, Global Avg Loss: 5.36461609, Time: 0.0207 Steps: 6450, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000073, Sample Num: 1168, Cur Loss: 1.39706981, Cur Avg Loss: 0.98555062, Log Avg loss: 0.98307213, Global Avg Loss: 5.35783351, Time: 0.0207 Steps: 6460, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000083, Sample Num: 1328, Cur Loss: 1.10350966, Cur Avg Loss: 0.97922384, Log Avg loss: 0.93303837, Global Avg Loss: 5.35099457, Time: 0.0207 Steps: 6470, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000093, Sample Num: 1488, Cur Loss: 0.33618075, Cur Avg Loss: 0.97783717, Log Avg loss: 0.96632778, Global Avg Loss: 5.34422811, Time: 0.0207 Steps: 6480, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000103, Sample Num: 1648, Cur Loss: 0.89012706, Cur Avg Loss: 0.97217086, Log Avg loss: 0.91947417, Global Avg Loss: 5.33741031, Time: 0.0207 Steps: 6490, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000113, Sample Num: 1808, Cur Loss: 0.72304535, Cur Avg Loss: 0.96177198, Log Avg loss: 0.85466354, Global Avg Loss: 5.33051378, Time: 0.0207 Steps: 6500, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000123, Sample Num: 1968, Cur Loss: 0.46004373, Cur Avg Loss: 0.95115076, Log Avg loss: 0.83113099, Global Avg Loss: 5.32360228, Time: 0.0207 Steps: 6510, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000133, Sample Num: 2128, Cur Loss: 0.90932703, Cur Avg Loss: 0.95621684, Log Avg loss: 1.01852961, Global Avg Loss: 5.31699941, Time: 0.0207 Steps: 6520, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000143, Sample Num: 2288, Cur Loss: 0.35785520, Cur Avg Loss: 0.97336820, Log Avg loss: 1.20148137, Global Avg Loss: 5.31069693, Time: 0.0208 Steps: 6530, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000153, Sample Num: 2448, Cur Loss: 0.71417505, Cur Avg Loss: 0.97826557, Log Avg loss: 1.04829784, Global Avg Loss: 5.30417950, Time: 0.0207 Steps: 6540, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000163, Sample Num: 2608, Cur Loss: 1.34700561, Cur Avg Loss: 0.97701295, Log Avg loss: 0.95784795, Global Avg Loss: 5.29754388, Time: 0.0208 Steps: 6550, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000173, Sample Num: 2768, Cur Loss: 1.34340692, Cur Avg Loss: 0.97072370, Log Avg loss: 0.86820885, Global Avg Loss: 5.29079185, Time: 0.0208 Steps: 6560, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000183, Sample Num: 2928, Cur Loss: 1.06766260, Cur Avg Loss: 0.98387706, Log Avg loss: 1.21143020, Global Avg Loss: 5.28458277, Time: 0.0209 Steps: 6570, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000193, Sample Num: 3088, Cur Loss: 0.81509894, Cur Avg Loss: 0.97796366, Log Avg loss: 0.86974857, Global Avg Loss: 5.27787330, Time: 0.0207 Steps: 6580, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000203, Sample Num: 3248, Cur Loss: 0.65234172, Cur Avg Loss: 0.96906221, Log Avg loss: 0.79726415, Global Avg Loss: 5.27107419, Time: 0.0207 Steps: 6590, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000213, Sample Num: 3408, Cur Loss: 0.40833202, Cur Avg Loss: 0.95944542, Log Avg loss: 0.76422462, Global Avg Loss: 5.26424563, Time: 0.0207 Steps: 6600, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000223, Sample Num: 3568, Cur Loss: 0.55553758, Cur Avg Loss: 0.95528606, Log Avg loss: 0.86669173, Global Avg Loss: 5.25759275, Time: 0.0207 Steps: 6610, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000233, Sample Num: 3728, Cur Loss: 0.98417938, Cur Avg Loss: 0.95523151, Log Avg loss: 0.95401490, Global Avg Loss: 5.25109188, Time: 0.0207 Steps: 6620, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000243, Sample Num: 3888, Cur Loss: 1.17913818, Cur Avg Loss: 0.94834044, Log Avg loss: 0.78777859, Global Avg Loss: 5.24435988, Time: 0.0208 Steps: 6630, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000253, Sample Num: 4048, Cur Loss: 0.86061108, Cur Avg Loss: 0.94914174, Log Avg loss: 0.96861340, Global Avg Loss: 5.23792051, Time: 0.0208 Steps: 6640, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000263, Sample Num: 4208, Cur Loss: 0.88209051, Cur Avg Loss: 0.94507230, Log Avg loss: 0.84211544, Global Avg Loss: 5.23131027, Time: 0.0209 Steps: 6650, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000273, Sample Num: 4368, Cur Loss: 1.38864124, Cur Avg Loss: 0.94774939, Log Avg loss: 1.01815692, Global Avg Loss: 5.22498422, Time: 0.0208 Steps: 6660, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000283, Sample Num: 4528, Cur Loss: 0.97172868, Cur Avg Loss: 0.95769850, Log Avg loss: 1.22930899, Global Avg Loss: 5.21899370, Time: 0.0208 Steps: 6670, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000293, Sample Num: 4688, Cur Loss: 1.25485289, Cur Avg Loss: 0.96035283, Log Avg loss: 1.03547050, Global Avg Loss: 5.21273094, Time: 0.0208 Steps: 6680, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000303, Sample Num: 4848, Cur Loss: 1.32426834, Cur Avg Loss: 0.95896407, Log Avg loss: 0.91827350, Global Avg Loss: 5.20631172, Time: 0.0208 Steps: 6690, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000313, Sample Num: 5008, Cur Loss: 0.86037552, Cur Avg Loss: 0.96549362, Log Avg loss: 1.16333895, Global Avg Loss: 5.20027743, Time: 0.0208 Steps: 6700, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000323, Sample Num: 5168, Cur Loss: 1.15722084, Cur Avg Loss: 0.96635079, Log Avg loss: 0.99318009, Global Avg Loss: 5.19400754, Time: 0.0208 Steps: 6710, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000333, Sample Num: 5328, Cur Loss: 1.00538254, Cur Avg Loss: 0.96427575, Log Avg loss: 0.89725197, Global Avg Loss: 5.18761356, Time: 0.0208 Steps: 6720, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000343, Sample Num: 5488, Cur Loss: 1.19687200, Cur Avg Loss: 0.96106548, Log Avg loss: 0.85416347, Global Avg Loss: 5.18117456, Time: 0.0208 Steps: 6730, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000353, Sample Num: 5648, Cur Loss: 0.52629989, Cur Avg Loss: 0.95961138, Log Avg loss: 0.90973577, Global Avg Loss: 5.17483711, Time: 0.0208 Steps: 6740, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000363, Sample Num: 5808, Cur Loss: 0.93565059, Cur Avg Loss: 0.95844031, Log Avg loss: 0.91710171, Global Avg Loss: 5.16852935, Time: 0.0208 Steps: 6750, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000373, Sample Num: 5968, Cur Loss: 0.83053410, Cur Avg Loss: 0.96188812, Log Avg loss: 1.08704328, Global Avg Loss: 5.16249165, Time: 0.0208 Steps: 6760, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000383, Sample Num: 6128, Cur Loss: 0.41352314, Cur Avg Loss: 0.96021557, Log Avg loss: 0.89782972, Global Avg Loss: 5.15619230, Time: 0.0208 Steps: 6770, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000393, Sample Num: 6288, Cur Loss: 0.66151583, Cur Avg Loss: 0.95607298, Log Avg loss: 0.79741177, Global Avg Loss: 5.14976342, Time: 0.0207 Steps: 6780, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000403, Sample Num: 6448, Cur Loss: 0.38816339, Cur Avg Loss: 0.95514267, Log Avg loss: 0.91858146, Global Avg Loss: 5.14353193, Time: 0.0208 Steps: 6790, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000413, Sample Num: 6608, Cur Loss: 0.79963541, Cur Avg Loss: 0.94988013, Log Avg loss: 0.73779980, Global Avg Loss: 5.13705291, Time: 0.0208 Steps: 6800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000423, Sample Num: 6768, Cur Loss: 1.60002530, Cur Avg Loss: 0.95558045, Log Avg loss: 1.19100358, Global Avg Loss: 5.13125842, Time: 0.0208 Steps: 6810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000433, Sample Num: 6928, Cur Loss: 1.45822597, Cur Avg Loss: 0.95502077, Log Avg loss: 0.93134649, Global Avg Loss: 5.12510019, Time: 0.0207 Steps: 6820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000443, Sample Num: 7088, Cur Loss: 1.20752239, Cur Avg Loss: 0.95892433, Log Avg loss: 1.12794847, Global Avg Loss: 5.11924785, Time: 0.0207 Steps: 6830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000453, Sample Num: 7248, Cur Loss: 0.70430017, Cur Avg Loss: 0.95429761, Log Avg loss: 0.74933383, Global Avg Loss: 5.11285908, Time: 0.0208 Steps: 6840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000463, Sample Num: 7408, Cur Loss: 0.67823201, Cur Avg Loss: 0.95288181, Log Avg loss: 0.88874583, Global Avg Loss: 5.10669249, Time: 0.0208 Steps: 6850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000473, Sample Num: 7568, Cur Loss: 1.09318411, Cur Avg Loss: 0.95431498, Log Avg loss: 1.02067074, Global Avg Loss: 5.10073619, Time: 0.0208 Steps: 6860, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000483, Sample Num: 7728, Cur Loss: 0.78000790, Cur Avg Loss: 0.95030554, Log Avg loss: 0.76065913, Global Avg Loss: 5.09441876, Time: 0.0207 Steps: 6870, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000493, Sample Num: 7888, Cur Loss: 0.72824985, Cur Avg Loss: 0.94987954, Log Avg loss: 0.92930369, Global Avg Loss: 5.08836481, Time: 0.0208 Steps: 6880, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000503, Sample Num: 8048, Cur Loss: 1.33120251, Cur Avg Loss: 0.95280494, Log Avg loss: 1.09702750, Global Avg Loss: 5.08257187, Time: 0.0207 Steps: 6890, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000513, Sample Num: 8208, Cur Loss: 1.47152174, Cur Avg Loss: 0.95547567, Log Avg loss: 1.08981329, Global Avg Loss: 5.07678526, Time: 0.0244 Steps: 6900, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000523, Sample Num: 8368, Cur Loss: 0.60931349, Cur Avg Loss: 0.95187810, Log Avg loss: 0.76732283, Global Avg Loss: 5.07054871, Time: 0.0208 Steps: 6910, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000533, Sample Num: 8528, Cur Loss: 0.81652874, Cur Avg Loss: 0.96291959, Log Avg loss: 1.54038943, Global Avg Loss: 5.06544732, Time: 0.0208 Steps: 6920, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000543, Sample Num: 8688, Cur Loss: 0.87898827, Cur Avg Loss: 0.96355643, Log Avg loss: 0.99749990, Global Avg Loss: 5.05957727, Time: 0.0207 Steps: 6930, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000553, Sample Num: 8848, Cur Loss: 1.14099002, Cur Avg Loss: 0.96380072, Log Avg loss: 0.97706559, Global Avg Loss: 5.05369468, Time: 0.0207 Steps: 6940, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000563, Sample Num: 9008, Cur Loss: 0.20975354, Cur Avg Loss: 0.96412457, Log Avg loss: 0.98203368, Global Avg Loss: 5.04783618, Time: 0.0207 Steps: 6950, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000573, Sample Num: 9168, Cur Loss: 0.87890816, Cur Avg Loss: 0.96356034, Log Avg loss: 0.93179397, Global Avg Loss: 5.04192232, Time: 0.0208 Steps: 6960, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000583, Sample Num: 9328, Cur Loss: 0.57920694, Cur Avg Loss: 0.96406623, Log Avg loss: 0.99305407, Global Avg Loss: 5.03611333, Time: 0.0208 Steps: 6970, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000593, Sample Num: 9488, Cur Loss: 1.44518244, Cur Avg Loss: 0.96634623, Log Avg loss: 1.09927022, Global Avg Loss: 5.03047316, Time: 0.0208 Steps: 6980, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000603, Sample Num: 9648, Cur Loss: 0.93965584, Cur Avg Loss: 0.96329575, Log Avg loss: 0.78240213, Global Avg Loss: 5.02439580, Time: 0.0208 Steps: 6990, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000613, Sample Num: 9808, Cur Loss: 0.57576740, Cur Avg Loss: 0.96330478, Log Avg loss: 0.96384910, Global Avg Loss: 5.01859502, Time: 0.0208 Steps: 7000, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000623, Sample Num: 9968, Cur Loss: 0.68201250, Cur Avg Loss: 0.96061422, Log Avg loss: 0.79568304, Global Avg Loss: 5.01257089, Time: 0.0208 Steps: 7010, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000633, Sample Num: 10128, Cur Loss: 1.06980062, Cur Avg Loss: 0.96447602, Log Avg loss: 1.20506582, Global Avg Loss: 5.00714710, Time: 0.0207 Steps: 7020, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000643, Sample Num: 10288, Cur Loss: 1.11073399, Cur Avg Loss: 0.96460985, Log Avg loss: 0.97308129, Global Avg Loss: 5.00140874, Time: 0.0208 Steps: 7030, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000653, Sample Num: 10448, Cur Loss: 0.66734737, Cur Avg Loss: 0.96281556, Log Avg loss: 0.84744302, Global Avg Loss: 4.99550822, Time: 0.0208 Steps: 7040, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000663, Sample Num: 10608, Cur Loss: 1.47118819, Cur Avg Loss: 0.96152514, Log Avg loss: 0.87726063, Global Avg Loss: 4.98966673, Time: 0.0207 Steps: 7050, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000673, Sample Num: 10768, Cur Loss: 0.55134863, Cur Avg Loss: 0.95477440, Log Avg loss: 0.50720037, Global Avg Loss: 4.98331763, Time: 0.0208 Steps: 7060, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000683, Sample Num: 10928, Cur Loss: 0.75863516, Cur Avg Loss: 0.95627144, Log Avg loss: 1.05702194, Global Avg Loss: 4.97776417, Time: 0.0208 Steps: 7070, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000693, Sample Num: 11088, Cur Loss: 0.61121881, Cur Avg Loss: 0.95687277, Log Avg loss: 0.99794365, Global Avg Loss: 4.97214296, Time: 0.0208 Steps: 7080, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000703, Sample Num: 11248, Cur Loss: 0.51183248, Cur Avg Loss: 0.95708523, Log Avg loss: 0.97180911, Global Avg Loss: 4.96650074, Time: 0.0207 Steps: 7090, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000713, Sample Num: 11408, Cur Loss: 0.76956463, Cur Avg Loss: 0.95534032, Log Avg loss: 0.83267270, Global Avg Loss: 4.96067844, Time: 0.0208 Steps: 7100, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000723, Sample Num: 11568, Cur Loss: 0.32230461, Cur Avg Loss: 0.95156715, Log Avg loss: 0.68254047, Global Avg Loss: 4.95466137, Time: 0.0208 Steps: 7110, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000733, Sample Num: 11728, Cur Loss: 0.60861510, Cur Avg Loss: 0.94805647, Log Avg loss: 0.69423460, Global Avg Loss: 4.94867763, Time: 0.0207 Steps: 7120, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000743, Sample Num: 11888, Cur Loss: 1.23069787, Cur Avg Loss: 0.94781654, Log Avg loss: 0.93022955, Global Avg Loss: 4.94304165, Time: 0.0208 Steps: 7130, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000753, Sample Num: 12048, Cur Loss: 1.26517868, Cur Avg Loss: 0.95041721, Log Avg loss: 1.14364693, Global Avg Loss: 4.93772037, Time: 0.0208 Steps: 7140, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000763, Sample Num: 12208, Cur Loss: 0.86081183, Cur Avg Loss: 0.95034543, Log Avg loss: 0.94494050, Global Avg Loss: 4.93213607, Time: 0.0208 Steps: 7150, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000773, Sample Num: 12368, Cur Loss: 0.80026937, Cur Avg Loss: 0.95368997, Log Avg loss: 1.20887787, Global Avg Loss: 4.92693598, Time: 0.0244 Steps: 7160, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000783, Sample Num: 12528, Cur Loss: 0.96257687, Cur Avg Loss: 0.95217863, Log Avg loss: 0.83535199, Global Avg Loss: 4.92122945, Time: 0.0208 Steps: 7170, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000793, Sample Num: 12688, Cur Loss: 0.70845187, Cur Avg Loss: 0.95192054, Log Avg loss: 0.93171265, Global Avg Loss: 4.91567302, Time: 0.0207 Steps: 7180, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000803, Sample Num: 12848, Cur Loss: 0.49355811, Cur Avg Loss: 0.95224807, Log Avg loss: 0.97822099, Global Avg Loss: 4.91019673, Time: 0.0208 Steps: 7190, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000813, Sample Num: 13008, Cur Loss: 0.88961339, Cur Avg Loss: 0.95143637, Log Avg loss: 0.88625689, Global Avg Loss: 4.90460793, Time: 0.0207 Steps: 7200, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000823, Sample Num: 13168, Cur Loss: 0.30222756, Cur Avg Loss: 0.94857030, Log Avg loss: 0.71555861, Global Avg Loss: 4.89879787, Time: 0.0207 Steps: 7210, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000833, Sample Num: 13328, Cur Loss: 1.84626055, Cur Avg Loss: 0.94687318, Log Avg loss: 0.80720002, Global Avg Loss: 4.89313084, Time: 0.0208 Steps: 7220, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000843, Sample Num: 13488, Cur Loss: 0.82194006, Cur Avg Loss: 0.94640230, Log Avg loss: 0.90717815, Global Avg Loss: 4.88761777, Time: 0.0207 Steps: 7230, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000853, Sample Num: 13648, Cur Loss: 1.36714470, Cur Avg Loss: 0.94948567, Log Avg loss: 1.20941375, Global Avg Loss: 4.88253737, Time: 0.0207 Steps: 7240, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000863, Sample Num: 13808, Cur Loss: 0.93566632, Cur Avg Loss: 0.95289952, Log Avg loss: 1.24410101, Global Avg Loss: 4.87751884, Time: 0.0207 Steps: 7250, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000873, Sample Num: 13968, Cur Loss: 0.56478584, Cur Avg Loss: 0.95276407, Log Avg loss: 0.94107459, Global Avg Loss: 4.87209674, Time: 0.0207 Steps: 7260, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000883, Sample Num: 14128, Cur Loss: 0.43222278, Cur Avg Loss: 0.95062200, Log Avg loss: 0.76361932, Global Avg Loss: 4.86644547, Time: 0.0207 Steps: 7270, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000893, Sample Num: 14288, Cur Loss: 0.92335218, Cur Avg Loss: 0.94921844, Log Avg loss: 0.82528430, Global Avg Loss: 4.86089442, Time: 0.0207 Steps: 7280, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000903, Sample Num: 14448, Cur Loss: 1.22710717, Cur Avg Loss: 0.95197177, Log Avg loss: 1.19784376, Global Avg Loss: 4.85586966, Time: 0.0207 Steps: 7290, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000913, Sample Num: 14608, Cur Loss: 0.71672547, Cur Avg Loss: 0.95027104, Log Avg loss: 0.79669523, Global Avg Loss: 4.85030915, Time: 0.0207 Steps: 7300, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000923, Sample Num: 14768, Cur Loss: 0.86208510, Cur Avg Loss: 0.95060932, Log Avg loss: 0.98149429, Global Avg Loss: 4.84501665, Time: 0.0207 Steps: 7310, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000933, Sample Num: 14928, Cur Loss: 0.80693114, Cur Avg Loss: 0.95030242, Log Avg loss: 0.92197573, Global Avg Loss: 4.83965730, Time: 0.0208 Steps: 7320, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000943, Sample Num: 15088, Cur Loss: 0.85150898, Cur Avg Loss: 0.95078043, Log Avg loss: 0.99537863, Global Avg Loss: 4.83441272, Time: 0.0208 Steps: 7330, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000953, Sample Num: 15248, Cur Loss: 0.82887214, Cur Avg Loss: 0.94976615, Log Avg loss: 0.85411943, Global Avg Loss: 4.82898998, Time: 0.0207 Steps: 7340, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000963, Sample Num: 15408, Cur Loss: 0.95863044, Cur Avg Loss: 0.95209802, Log Avg loss: 1.17432565, Global Avg Loss: 4.82401765, Time: 0.0208 Steps: 7350, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000973, Sample Num: 15568, Cur Loss: 1.22281337, Cur Avg Loss: 0.95145272, Log Avg loss: 0.88931025, Global Avg Loss: 4.81867158, Time: 0.0208 Steps: 7360, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000983, Sample Num: 15728, Cur Loss: 0.87526500, Cur Avg Loss: 0.95167951, Log Avg loss: 0.97374570, Global Avg Loss: 4.81345458, Time: 0.0208 Steps: 7370, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000993, Sample Num: 15888, Cur Loss: 0.59784013, Cur Avg Loss: 0.94930687, Log Avg loss: 0.71607640, Global Avg Loss: 4.80790258, Time: 0.0207 Steps: 7380, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001003, Sample Num: 16048, Cur Loss: 1.07030964, Cur Avg Loss: 0.94863241, Log Avg loss: 0.88165879, Global Avg Loss: 4.80258966, Time: 0.0207 Steps: 7390, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001013, Sample Num: 16208, Cur Loss: 0.43088460, Cur Avg Loss: 0.94820255, Log Avg loss: 0.90508743, Global Avg Loss: 4.79732277, Time: 0.0207 Steps: 7400, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001023, Sample Num: 16368, Cur Loss: 0.97258109, Cur Avg Loss: 0.94843422, Log Avg loss: 0.97190280, Global Avg Loss: 4.79216026, Time: 0.0207 Steps: 7410, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001033, Sample Num: 16528, Cur Loss: 0.99767053, Cur Avg Loss: 0.94837636, Log Avg loss: 0.94245737, Global Avg Loss: 4.78697198, Time: 0.0209 Steps: 7420, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001043, Sample Num: 16688, Cur Loss: 1.60391641, Cur Avg Loss: 0.94812756, Log Avg loss: 0.92242591, Global Avg Loss: 4.78177071, Time: 0.0208 Steps: 7430, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001053, Sample Num: 16848, Cur Loss: 0.42756987, Cur Avg Loss: 0.94653951, Log Avg loss: 0.78090610, Global Avg Loss: 4.77639320, Time: 0.0209 Steps: 7440, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001063, Sample Num: 17008, Cur Loss: 0.43084621, Cur Avg Loss: 0.94448470, Log Avg loss: 0.72811332, Global Avg Loss: 4.77095927, Time: 0.0209 Steps: 7450, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001073, Sample Num: 17168, Cur Loss: 1.03381491, Cur Avg Loss: 0.94687541, Log Avg loss: 1.20100849, Global Avg Loss: 4.76617381, Time: 0.0208 Steps: 7460, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001083, Sample Num: 17328, Cur Loss: 0.55339301, Cur Avg Loss: 0.94808007, Log Avg loss: 1.07733907, Global Avg Loss: 4.76123561, Time: 0.0208 Steps: 7470, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001093, Sample Num: 17488, Cur Loss: 1.75735700, Cur Avg Loss: 0.94805519, Log Avg loss: 0.94536153, Global Avg Loss: 4.75613418, Time: 0.0209 Steps: 7480, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001103, Sample Num: 17648, Cur Loss: 1.07806563, Cur Avg Loss: 0.94717359, Log Avg loss: 0.85081462, Global Avg Loss: 4.75092013, Time: 0.0209 Steps: 7490, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001113, Sample Num: 17808, Cur Loss: 0.49170348, Cur Avg Loss: 0.94751536, Log Avg loss: 0.98521259, Global Avg Loss: 4.74589919, Time: 0.0208 Steps: 7500, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001123, Sample Num: 17968, Cur Loss: 1.52125502, Cur Avg Loss: 0.94971776, Log Avg loss: 1.19484431, Global Avg Loss: 4.74117075, Time: 0.0208 Steps: 7510, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001133, Sample Num: 18128, Cur Loss: 1.43356586, Cur Avg Loss: 0.94916362, Log Avg loss: 0.88693417, Global Avg Loss: 4.73604544, Time: 0.0208 Steps: 7520, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001143, Sample Num: 18288, Cur Loss: 0.94206989, Cur Avg Loss: 0.94864323, Log Avg loss: 0.88968273, Global Avg Loss: 4.73093739, Time: 0.0209 Steps: 7530, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001153, Sample Num: 18448, Cur Loss: 0.99895465, Cur Avg Loss: 0.94635858, Log Avg loss: 0.68522272, Global Avg Loss: 4.72557172, Time: 0.0208 Steps: 7540, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001163, Sample Num: 18608, Cur Loss: 1.16491449, Cur Avg Loss: 0.94554680, Log Avg loss: 0.85194940, Global Avg Loss: 4.72044109, Time: 0.0209 Steps: 7550, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001173, Sample Num: 18768, Cur Loss: 1.00222075, Cur Avg Loss: 0.94570972, Log Avg loss: 0.96465703, Global Avg Loss: 4.71547312, Time: 0.0208 Steps: 7560, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001183, Sample Num: 18928, Cur Loss: 0.27909482, Cur Avg Loss: 0.94460568, Log Avg loss: 0.81510222, Global Avg Loss: 4.71032072, Time: 0.0208 Steps: 7570, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001193, Sample Num: 19088, Cur Loss: 0.85623389, Cur Avg Loss: 0.94707259, Log Avg loss: 1.23890733, Global Avg Loss: 4.70574102, Time: 0.0209 Steps: 7580, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001203, Sample Num: 19248, Cur Loss: 1.03599298, Cur Avg Loss: 0.94966935, Log Avg loss: 1.25946266, Global Avg Loss: 4.70120047, Time: 0.0209 Steps: 7590, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001213, Sample Num: 19408, Cur Loss: 0.62777883, Cur Avg Loss: 0.95187019, Log Avg loss: 1.21663121, Global Avg Loss: 4.69661551, Time: 0.0209 Steps: 7600, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001223, Sample Num: 19568, Cur Loss: 1.03208911, Cur Avg Loss: 0.95094764, Log Avg loss: 0.83904278, Global Avg Loss: 4.69154642, Time: 0.0209 Steps: 7610, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001233, Sample Num: 19728, Cur Loss: 1.50768733, Cur Avg Loss: 0.95142433, Log Avg loss: 1.00972296, Global Avg Loss: 4.68671463, Time: 0.0208 Steps: 7620, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001243, Sample Num: 19888, Cur Loss: 1.29238784, Cur Avg Loss: 0.95187333, Log Avg loss: 1.00723537, Global Avg Loss: 4.68189225, Time: 0.0209 Steps: 7630, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001253, Sample Num: 20048, Cur Loss: 0.60216075, Cur Avg Loss: 0.95027366, Log Avg loss: 0.75143466, Global Avg Loss: 4.67674767, Time: 0.0209 Steps: 7640, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001263, Sample Num: 20208, Cur Loss: 0.75408649, Cur Avg Loss: 0.95103867, Log Avg loss: 1.04689431, Global Avg Loss: 4.67200276, Time: 0.0209 Steps: 7650, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001273, Sample Num: 20368, Cur Loss: 1.49834919, Cur Avg Loss: 0.95049791, Log Avg loss: 0.88220011, Global Avg Loss: 4.66705524, Time: 0.0208 Steps: 7660, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001283, Sample Num: 20528, Cur Loss: 0.61374187, Cur Avg Loss: 0.95126807, Log Avg loss: 1.04930942, Global Avg Loss: 4.66233849, Time: 0.0244 Steps: 7670, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001293, Sample Num: 20688, Cur Loss: 1.12440968, Cur Avg Loss: 0.95526081, Log Avg loss: 1.46753002, Global Avg Loss: 4.65817859, Time: 0.0207 Steps: 7680, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001303, Sample Num: 20848, Cur Loss: 0.76020026, Cur Avg Loss: 0.95701407, Log Avg loss: 1.18371031, Global Avg Loss: 4.65366042, Time: 0.0208 Steps: 7690, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001313, Sample Num: 21008, Cur Loss: 1.02307308, Cur Avg Loss: 0.95728138, Log Avg loss: 0.99211231, Global Avg Loss: 4.64890516, Time: 0.0208 Steps: 7700, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001323, Sample Num: 21168, Cur Loss: 0.77216953, Cur Avg Loss: 0.95579545, Log Avg loss: 0.76069163, Global Avg Loss: 4.64386209, Time: 0.0207 Steps: 7710, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001333, Sample Num: 21328, Cur Loss: 0.71079928, Cur Avg Loss: 0.95439053, Log Avg loss: 0.76852081, Global Avg Loss: 4.63884221, Time: 0.0208 Steps: 7720, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001343, Sample Num: 21488, Cur Loss: 0.10272863, Cur Avg Loss: 0.95456943, Log Avg loss: 0.97841621, Global Avg Loss: 4.63410686, Time: 0.0208 Steps: 7730, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001353, Sample Num: 21648, Cur Loss: 0.73255837, Cur Avg Loss: 0.95547434, Log Avg loss: 1.07700436, Global Avg Loss: 4.62951112, Time: 0.0208 Steps: 7740, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001363, Sample Num: 21808, Cur Loss: 1.34257543, Cur Avg Loss: 0.95425599, Log Avg loss: 0.78941276, Global Avg Loss: 4.62455616, Time: 0.0207 Steps: 7750, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001373, Sample Num: 21968, Cur Loss: 0.86074466, Cur Avg Loss: 0.95304928, Log Avg loss: 0.78857478, Global Avg Loss: 4.61961288, Time: 0.0208 Steps: 7760, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001383, Sample Num: 22128, Cur Loss: 0.53116447, Cur Avg Loss: 0.95208659, Log Avg loss: 0.81990907, Global Avg Loss: 4.61472266, Time: 0.0208 Steps: 7770, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001393, Sample Num: 22288, Cur Loss: 0.42338955, Cur Avg Loss: 0.95076189, Log Avg loss: 0.76755608, Global Avg Loss: 4.60977771, Time: 0.0207 Steps: 7780, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001403, Sample Num: 22448, Cur Loss: 2.14615893, Cur Avg Loss: 0.95092960, Log Avg loss: 0.97429101, Global Avg Loss: 4.60511085, Time: 0.0207 Steps: 7790, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001413, Sample Num: 22608, Cur Loss: 1.11798441, Cur Avg Loss: 0.95042927, Log Avg loss: 0.88023394, Global Avg Loss: 4.60033537, Time: 0.0207 Steps: 7800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001423, Sample Num: 22768, Cur Loss: 1.06541955, Cur Avg Loss: 0.95038057, Log Avg loss: 0.94349855, Global Avg Loss: 4.59565312, Time: 0.0207 Steps: 7810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001433, Sample Num: 22928, Cur Loss: 1.63396335, Cur Avg Loss: 0.95127572, Log Avg loss: 1.07865596, Global Avg Loss: 4.59115568, Time: 0.0208 Steps: 7820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001443, Sample Num: 23088, Cur Loss: 0.62841666, Cur Avg Loss: 0.95071193, Log Avg loss: 0.86992032, Global Avg Loss: 4.58640314, Time: 0.0208 Steps: 7830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001453, Sample Num: 23248, Cur Loss: 1.17565370, Cur Avg Loss: 0.95192260, Log Avg loss: 1.12662274, Global Avg Loss: 4.58199016, Time: 0.0207 Steps: 7840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001463, Sample Num: 23408, Cur Loss: 0.50002068, Cur Avg Loss: 0.95092979, Log Avg loss: 0.80667354, Global Avg Loss: 4.57718084, Time: 0.0207 Steps: 7850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001473, Sample Num: 23568, Cur Loss: 1.13363481, Cur Avg Loss: 0.94844263, Log Avg loss: 0.58457151, Global Avg Loss: 4.57210118, Time: 0.0208 Steps: 7860, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001483, Sample Num: 23728, Cur Loss: 1.32813597, Cur Avg Loss: 0.94799167, Log Avg loss: 0.88156511, Global Avg Loss: 4.56741181, Time: 0.0207 Steps: 7870, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001493, Sample Num: 23888, Cur Loss: 0.60016036, Cur Avg Loss: 0.94743977, Log Avg loss: 0.86559397, Global Avg Loss: 4.56271407, Time: 0.0207 Steps: 7880, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001503, Sample Num: 24048, Cur Loss: 0.70268691, Cur Avg Loss: 0.94633961, Log Avg loss: 0.78208568, Global Avg Loss: 4.55792240, Time: 0.0207 Steps: 7890, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001513, Sample Num: 24208, Cur Loss: 0.56982493, Cur Avg Loss: 0.94393483, Log Avg loss: 0.58249545, Global Avg Loss: 4.55289022, Time: 0.0208 Steps: 7900, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001523, Sample Num: 24368, Cur Loss: 0.43403232, Cur Avg Loss: 0.94446476, Log Avg loss: 1.02464378, Global Avg Loss: 4.54842973, Time: 0.0208 Steps: 7910, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001533, Sample Num: 24528, Cur Loss: 1.06811333, Cur Avg Loss: 0.94510684, Log Avg loss: 1.04289601, Global Avg Loss: 4.54400355, Time: 0.0207 Steps: 7920, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001543, Sample Num: 24688, Cur Loss: 0.63593858, Cur Avg Loss: 0.94602344, Log Avg loss: 1.08653694, Global Avg Loss: 4.53964356, Time: 0.0208 Steps: 7930, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001553, Sample Num: 24848, Cur Loss: 0.42078358, Cur Avg Loss: 0.94508212, Log Avg loss: 0.79983755, Global Avg Loss: 4.53493348, Time: 0.0208 Steps: 7940, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001563, Sample Num: 25008, Cur Loss: 0.69033164, Cur Avg Loss: 0.94381065, Log Avg loss: 0.74635150, Global Avg Loss: 4.53016797, Time: 0.0208 Steps: 7950, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001573, Sample Num: 25168, Cur Loss: 1.35431552, Cur Avg Loss: 0.94450410, Log Avg loss: 1.05288994, Global Avg Loss: 4.52579953, Time: 0.0208 Steps: 7960, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001583, Sample Num: 25328, Cur Loss: 1.18484068, Cur Avg Loss: 0.94422542, Log Avg loss: 0.90038843, Global Avg Loss: 4.52125071, Time: 0.0207 Steps: 7970, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001593, Sample Num: 25488, Cur Loss: 1.26844573, Cur Avg Loss: 0.94271738, Log Avg loss: 0.70399580, Global Avg Loss: 4.51646718, Time: 0.0208 Steps: 7980, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001603, Sample Num: 25648, Cur Loss: 0.61574233, Cur Avg Loss: 0.94069550, Log Avg loss: 0.61860861, Global Avg Loss: 4.51158876, Time: 0.0209 Steps: 7990, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001613, Sample Num: 25808, Cur Loss: 0.77024168, Cur Avg Loss: 0.94107489, Log Avg loss: 1.00189220, Global Avg Loss: 4.50720164, Time: 0.0207 Steps: 8000, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001623, Sample Num: 25968, Cur Loss: 0.65789491, Cur Avg Loss: 0.93958765, Log Avg loss: 0.69969586, Global Avg Loss: 4.50244820, Time: 0.0208 Steps: 8010, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001633, Sample Num: 26128, Cur Loss: 0.44309899, Cur Avg Loss: 0.93888586, Log Avg loss: 0.82498469, Global Avg Loss: 4.49786283, Time: 0.0207 Steps: 8020, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001643, Sample Num: 26288, Cur Loss: 0.71099186, Cur Avg Loss: 0.93753705, Log Avg loss: 0.71727723, Global Avg Loss: 4.49315476, Time: 0.0208 Steps: 8030, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001653, Sample Num: 26448, Cur Loss: 0.99919343, Cur Avg Loss: 0.93701329, Log Avg loss: 0.85095817, Global Avg Loss: 4.48862466, Time: 0.0207 Steps: 8040, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001663, Sample Num: 26608, Cur Loss: 1.11871326, Cur Avg Loss: 0.93656613, Log Avg loss: 0.86265178, Global Avg Loss: 4.48412035, Time: 0.0207 Steps: 8050, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001673, Sample Num: 26768, Cur Loss: 0.93098104, Cur Avg Loss: 0.93750545, Log Avg loss: 1.09371382, Global Avg Loss: 4.47991389, Time: 0.0208 Steps: 8060, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001683, Sample Num: 26928, Cur Loss: 0.42285383, Cur Avg Loss: 0.93726377, Log Avg loss: 0.89683160, Global Avg Loss: 4.47547388, Time: 0.0208 Steps: 8070, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001693, Sample Num: 27088, Cur Loss: 1.62504077, Cur Avg Loss: 0.93794329, Log Avg loss: 1.05230528, Global Avg Loss: 4.47123729, Time: 0.0208 Steps: 8080, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001703, Sample Num: 27248, Cur Loss: 0.68846464, Cur Avg Loss: 0.93681149, Log Avg loss: 0.74519862, Global Avg Loss: 4.46663155, Time: 0.0208 Steps: 8090, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001713, Sample Num: 27408, Cur Loss: 0.55779266, Cur Avg Loss: 0.93486496, Log Avg loss: 0.60336990, Global Avg Loss: 4.46186210, Time: 0.0208 Steps: 8100, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001723, Sample Num: 27568, Cur Loss: 0.53712159, Cur Avg Loss: 0.93462894, Log Avg loss: 0.89419937, Global Avg Loss: 4.45746301, Time: 0.0208 Steps: 8110, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001733, Sample Num: 27728, Cur Loss: 0.89542729, Cur Avg Loss: 0.93401703, Log Avg loss: 0.82858481, Global Avg Loss: 4.45299394, Time: 0.0207 Steps: 8120, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001743, Sample Num: 27888, Cur Loss: 1.03866208, Cur Avg Loss: 0.93331008, Log Avg loss: 0.81079494, Global Avg Loss: 4.44851399, Time: 0.0208 Steps: 8130, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001753, Sample Num: 28048, Cur Loss: 1.05173635, Cur Avg Loss: 0.93296683, Log Avg loss: 0.87313971, Global Avg Loss: 4.44412164, Time: 0.0207 Steps: 8140, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001763, Sample Num: 28208, Cur Loss: 1.05604017, Cur Avg Loss: 0.93106599, Log Avg loss: 0.59784734, Global Avg Loss: 4.43940229, Time: 0.0207 Steps: 8150, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001773, Sample Num: 28368, Cur Loss: 0.95686311, Cur Avg Loss: 0.93131277, Log Avg loss: 0.97482132, Global Avg Loss: 4.43515648, Time: 0.0207 Steps: 8160, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001783, Sample Num: 28528, Cur Loss: 0.80831981, Cur Avg Loss: 0.92943791, Log Avg loss: 0.59702498, Global Avg Loss: 4.43045864, Time: 0.0208 Steps: 8170, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001793, Sample Num: 28688, Cur Loss: 0.83740032, Cur Avg Loss: 0.92883262, Log Avg loss: 0.82090979, Global Avg Loss: 4.42604599, Time: 0.0248 Steps: 8180, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001803, Sample Num: 28848, Cur Loss: 0.73063910, Cur Avg Loss: 0.92870275, Log Avg loss: 0.90541648, Global Avg Loss: 4.42174730, Time: 0.0211 Steps: 8190, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001813, Sample Num: 29008, Cur Loss: 0.57657647, Cur Avg Loss: 0.92805622, Log Avg loss: 0.81148758, Global Avg Loss: 4.41734454, Time: 0.0207 Steps: 8200, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001823, Sample Num: 29168, Cur Loss: 1.30224609, Cur Avg Loss: 0.92739499, Log Avg loss: 0.80751404, Global Avg Loss: 4.41294767, Time: 0.0208 Steps: 8210, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001833, Sample Num: 29328, Cur Loss: 1.24062812, Cur Avg Loss: 0.92781404, Log Avg loss: 1.00420675, Global Avg Loss: 4.40880078, Time: 0.0207 Steps: 8220, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001843, Sample Num: 29488, Cur Loss: 0.67042488, Cur Avg Loss: 0.92653741, Log Avg loss: 0.69252976, Global Avg Loss: 4.40428527, Time: 0.0207 Steps: 8230, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001853, Sample Num: 29648, Cur Loss: 0.32268533, Cur Avg Loss: 0.92584328, Log Avg loss: 0.79791560, Global Avg Loss: 4.39990860, Time: 0.0208 Steps: 8240, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001863, Sample Num: 29808, Cur Loss: 0.98472410, Cur Avg Loss: 0.92497889, Log Avg loss: 0.76480703, Global Avg Loss: 4.39550242, Time: 0.0207 Steps: 8250, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001873, Sample Num: 29968, Cur Loss: 1.02399659, Cur Avg Loss: 0.92602389, Log Avg loss: 1.12070866, Global Avg Loss: 4.39153778, Time: 0.0210 Steps: 8260, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001883, Sample Num: 30128, Cur Loss: 1.02458668, Cur Avg Loss: 0.92547087, Log Avg loss: 0.82189006, Global Avg Loss: 4.38722140, Time: 0.0207 Steps: 8270, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001893, Sample Num: 30288, Cur Loss: 0.47177672, Cur Avg Loss: 0.92556069, Log Avg loss: 0.94247282, Global Avg Loss: 4.38306107, Time: 0.0208 Steps: 8280, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001903, Sample Num: 30448, Cur Loss: 1.11950576, Cur Avg Loss: 0.92543990, Log Avg loss: 0.90257417, Global Avg Loss: 4.37886266, Time: 0.0208 Steps: 8290, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001913, Sample Num: 30608, Cur Loss: 1.33326530, Cur Avg Loss: 0.92619542, Log Avg loss: 1.06997238, Global Avg Loss: 4.37487604, Time: 0.0208 Steps: 8300, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001923, Sample Num: 30768, Cur Loss: 0.40767384, Cur Avg Loss: 0.92502183, Log Avg loss: 0.70051367, Global Avg Loss: 4.37045443, Time: 0.0208 Steps: 8310, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001933, Sample Num: 30928, Cur Loss: 0.95855743, Cur Avg Loss: 0.92367453, Log Avg loss: 0.66458881, Global Avg Loss: 4.36600026, Time: 0.0207 Steps: 8320, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001943, Sample Num: 31088, Cur Loss: 0.71988499, Cur Avg Loss: 0.92295315, Log Avg loss: 0.78350945, Global Avg Loss: 4.36169955, Time: 0.0207 Steps: 8330, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001953, Sample Num: 31248, Cur Loss: 0.45315757, Cur Avg Loss: 0.92335763, Log Avg loss: 1.00194878, Global Avg Loss: 4.35767107, Time: 0.0213 Steps: 8340, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001963, Sample Num: 31408, Cur Loss: 0.68959737, Cur Avg Loss: 0.92237675, Log Avg loss: 0.73081068, Global Avg Loss: 4.35332753, Time: 0.0207 Steps: 8350, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001973, Sample Num: 31568, Cur Loss: 0.90086126, Cur Avg Loss: 0.92252318, Log Avg loss: 0.95126730, Global Avg Loss: 4.34925808, Time: 0.0207 Steps: 8360, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001983, Sample Num: 31728, Cur Loss: 2.27601862, Cur Avg Loss: 0.92646103, Log Avg loss: 1.70339889, Global Avg Loss: 4.34609696, Time: 0.0207 Steps: 8370, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001993, Sample Num: 31888, Cur Loss: 0.79390705, Cur Avg Loss: 0.92804148, Log Avg loss: 1.24144547, Global Avg Loss: 4.34239212, Time: 0.0207 Steps: 8380, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002003, Sample Num: 32048, Cur Loss: 1.25368810, Cur Avg Loss: 0.92885170, Log Avg loss: 1.09032784, Global Avg Loss: 4.33851600, Time: 0.0207 Steps: 8390, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002013, Sample Num: 32208, Cur Loss: 0.74440861, Cur Avg Loss: 0.92760221, Log Avg loss: 0.67733021, Global Avg Loss: 4.33415745, Time: 0.0208 Steps: 8400, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002023, Sample Num: 32368, Cur Loss: 0.90260124, Cur Avg Loss: 0.92727212, Log Avg loss: 0.86082394, Global Avg Loss: 4.33002744, Time: 0.0207 Steps: 8410, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002033, Sample Num: 32528, Cur Loss: 0.82704657, Cur Avg Loss: 0.92688951, Log Avg loss: 0.84948779, Global Avg Loss: 4.32589379, Time: 0.0207 Steps: 8420, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002043, Sample Num: 32688, Cur Loss: 0.63086975, Cur Avg Loss: 0.92857451, Log Avg loss: 1.27113500, Global Avg Loss: 4.32227011, Time: 0.0208 Steps: 8430, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002053, Sample Num: 32848, Cur Loss: 0.57451034, Cur Avg Loss: 0.92856365, Log Avg loss: 0.92634573, Global Avg Loss: 4.31824650, Time: 0.0246 Steps: 8440, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002063, Sample Num: 33008, Cur Loss: 0.32363927, Cur Avg Loss: 0.92844968, Log Avg loss: 0.90505168, Global Avg Loss: 4.31420722, Time: 0.0208 Steps: 8450, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002073, Sample Num: 33168, Cur Loss: 0.66342670, Cur Avg Loss: 0.92757781, Log Avg loss: 0.74771097, Global Avg Loss: 4.30999150, Time: 0.0208 Steps: 8460, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002083, Sample Num: 33328, Cur Loss: 0.38588065, Cur Avg Loss: 0.92631915, Log Avg loss: 0.66539785, Global Avg Loss: 4.30568856, Time: 0.0208 Steps: 8470, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002093, Sample Num: 33488, Cur Loss: 0.52567124, Cur Avg Loss: 0.92688855, Log Avg loss: 1.04549546, Global Avg Loss: 4.30184399, Time: 0.0208 Steps: 8480, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002103, Sample Num: 33648, Cur Loss: 0.89899743, Cur Avg Loss: 0.92792302, Log Avg loss: 1.14443715, Global Avg Loss: 4.29812502, Time: 0.0208 Steps: 8490, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002113, Sample Num: 33808, Cur Loss: 0.99243486, Cur Avg Loss: 0.92687096, Log Avg loss: 0.70562233, Global Avg Loss: 4.29389855, Time: 0.0210 Steps: 8500, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002123, Sample Num: 33968, Cur Loss: 1.06123984, Cur Avg Loss: 0.92647099, Log Avg loss: 0.84195815, Global Avg Loss: 4.28984221, Time: 0.0210 Steps: 8510, Updated lr: 0.000093 ***** Running evaluation checkpoint-8516 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-8516 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.414138, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.11596, "eval_total_loss": 784.520072, "eval_mae": 0.923376, "eval_mse": 1.115776, "eval_r2": 0.290739, "eval_sp_statistic": 0.697284, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.719363, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.121106, "test_total_loss": 562.795426, "test_mae": 0.929013, "test_mse": 1.121298, "test_r2": 0.276304, "test_sp_statistic": 0.580982, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.653776, "test_ps_pvalue": 0.0, "lr": 9.287245139876719e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 4.287535187988221, "train_cur_epoch_loss": 1972.9903539046645, "train_cur_epoch_avg_loss": 0.9267216317072168, "train_cur_epoch_time": 44.414137840270996, "train_cur_epoch_avg_time": 0.020861502038643023, "epoch": 4, "step": 8516} ################################################## Training, Epoch: 0005, Batch: 000004, Sample Num: 64, Cur Loss: 0.94044173, Cur Avg Loss: 1.11292116, Log Avg loss: 1.05441206, Global Avg Loss: 4.28604476, Time: 0.0246 Steps: 8520, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000014, Sample Num: 224, Cur Loss: 0.60826284, Cur Avg Loss: 0.94032225, Log Avg loss: 0.87128268, Global Avg Loss: 4.28204152, Time: 0.0208 Steps: 8530, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000024, Sample Num: 384, Cur Loss: 0.31448674, Cur Avg Loss: 0.86998764, Log Avg loss: 0.77151919, Global Avg Loss: 4.27793084, Time: 0.0208 Steps: 8540, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000034, Sample Num: 544, Cur Loss: 1.08012557, Cur Avg Loss: 0.86929842, Log Avg loss: 0.86764430, Global Avg Loss: 4.27394220, Time: 0.0208 Steps: 8550, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000044, Sample Num: 704, Cur Loss: 0.47755694, Cur Avg Loss: 0.82460719, Log Avg loss: 0.67265699, Global Avg Loss: 4.26973509, Time: 0.0208 Steps: 8560, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000054, Sample Num: 864, Cur Loss: 1.29307497, Cur Avg Loss: 0.81492880, Log Avg loss: 0.77234391, Global Avg Loss: 4.26565412, Time: 0.0208 Steps: 8570, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000064, Sample Num: 1024, Cur Loss: 1.38723993, Cur Avg Loss: 0.83516110, Log Avg loss: 0.94441552, Global Avg Loss: 4.26178321, Time: 0.0208 Steps: 8580, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000074, Sample Num: 1184, Cur Loss: 0.49987566, Cur Avg Loss: 0.81940138, Log Avg loss: 0.71853915, Global Avg Loss: 4.25765837, Time: 0.0208 Steps: 8590, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000084, Sample Num: 1344, Cur Loss: 0.82305259, Cur Avg Loss: 0.80443864, Log Avg loss: 0.69371435, Global Avg Loss: 4.25351424, Time: 0.0209 Steps: 8600, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000094, Sample Num: 1504, Cur Loss: 1.28040993, Cur Avg Loss: 0.80846369, Log Avg loss: 0.84227417, Global Avg Loss: 4.24955229, Time: 0.0208 Steps: 8610, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000104, Sample Num: 1664, Cur Loss: 0.43088749, Cur Avg Loss: 0.78315000, Log Avg loss: 0.54520126, Global Avg Loss: 4.24525490, Time: 0.0208 Steps: 8620, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000114, Sample Num: 1824, Cur Loss: 0.35674116, Cur Avg Loss: 0.77275719, Log Avg loss: 0.66467201, Global Avg Loss: 4.24110591, Time: 0.0209 Steps: 8630, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000124, Sample Num: 1984, Cur Loss: 0.61351132, Cur Avg Loss: 0.77714464, Log Avg loss: 0.82716157, Global Avg Loss: 4.23715458, Time: 0.0208 Steps: 8640, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000134, Sample Num: 2144, Cur Loss: 1.10746491, Cur Avg Loss: 0.76205935, Log Avg loss: 0.57500166, Global Avg Loss: 4.23292088, Time: 0.0207 Steps: 8650, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000144, Sample Num: 2304, Cur Loss: 0.40145171, Cur Avg Loss: 0.76483766, Log Avg loss: 0.80206704, Global Avg Loss: 4.22895916, Time: 0.0207 Steps: 8660, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000154, Sample Num: 2464, Cur Loss: 0.95852816, Cur Avg Loss: 0.75975215, Log Avg loss: 0.68652092, Global Avg Loss: 4.22487330, Time: 0.0209 Steps: 8670, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000164, Sample Num: 2624, Cur Loss: 0.32251796, Cur Avg Loss: 0.75857552, Log Avg loss: 0.74045531, Global Avg Loss: 4.22085899, Time: 0.0208 Steps: 8680, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000174, Sample Num: 2784, Cur Loss: 0.34714401, Cur Avg Loss: 0.75907276, Log Avg loss: 0.76722744, Global Avg Loss: 4.21688473, Time: 0.0208 Steps: 8690, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000184, Sample Num: 2944, Cur Loss: 0.84954190, Cur Avg Loss: 0.75893302, Log Avg loss: 0.75650168, Global Avg Loss: 4.21290728, Time: 0.0208 Steps: 8700, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000194, Sample Num: 3104, Cur Loss: 0.70330095, Cur Avg Loss: 0.76175368, Log Avg loss: 0.81365379, Global Avg Loss: 4.20900458, Time: 0.0209 Steps: 8710, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000204, Sample Num: 3264, Cur Loss: 0.60412222, Cur Avg Loss: 0.76740607, Log Avg loss: 0.87706249, Global Avg Loss: 4.20518354, Time: 0.0209 Steps: 8720, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000214, Sample Num: 3424, Cur Loss: 0.50616467, Cur Avg Loss: 0.77464689, Log Avg loss: 0.92235964, Global Avg Loss: 4.20142315, Time: 0.0209 Steps: 8730, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000224, Sample Num: 3584, Cur Loss: 0.40114588, Cur Avg Loss: 0.77877538, Log Avg loss: 0.86712502, Global Avg Loss: 4.19760816, Time: 0.0207 Steps: 8740, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000234, Sample Num: 3744, Cur Loss: 0.89414799, Cur Avg Loss: 0.78252797, Log Avg loss: 0.86658594, Global Avg Loss: 4.19380128, Time: 0.0209 Steps: 8750, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000244, Sample Num: 3904, Cur Loss: 0.71985209, Cur Avg Loss: 0.78758051, Log Avg loss: 0.90581005, Global Avg Loss: 4.19004787, Time: 0.0207 Steps: 8760, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000254, Sample Num: 4064, Cur Loss: 0.65398008, Cur Avg Loss: 0.79575695, Log Avg loss: 0.99526196, Global Avg Loss: 4.18640501, Time: 0.0209 Steps: 8770, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000264, Sample Num: 4224, Cur Loss: 0.54041785, Cur Avg Loss: 0.79881536, Log Avg loss: 0.87649900, Global Avg Loss: 4.18263518, Time: 0.0209 Steps: 8780, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000274, Sample Num: 4384, Cur Loss: 0.46313265, Cur Avg Loss: 0.79921298, Log Avg loss: 0.80971008, Global Avg Loss: 4.17879795, Time: 0.0208 Steps: 8790, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000284, Sample Num: 4544, Cur Loss: 0.93033248, Cur Avg Loss: 0.80437631, Log Avg loss: 0.94585153, Global Avg Loss: 4.17512415, Time: 0.0208 Steps: 8800, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000294, Sample Num: 4704, Cur Loss: 0.24975571, Cur Avg Loss: 0.82140409, Log Avg loss: 1.30499313, Global Avg Loss: 4.17186634, Time: 0.0208 Steps: 8810, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000304, Sample Num: 4864, Cur Loss: 0.70085788, Cur Avg Loss: 0.82764980, Log Avg loss: 1.01127374, Global Avg Loss: 4.16828290, Time: 0.0208 Steps: 8820, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000314, Sample Num: 5024, Cur Loss: 0.89579284, Cur Avg Loss: 0.83863351, Log Avg loss: 1.17253817, Global Avg Loss: 4.16489021, Time: 0.0208 Steps: 8830, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000324, Sample Num: 5184, Cur Loss: 0.65734428, Cur Avg Loss: 0.82944709, Log Avg loss: 0.54099357, Global Avg Loss: 4.16079078, Time: 0.0208 Steps: 8840, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000334, Sample Num: 5344, Cur Loss: 0.34162825, Cur Avg Loss: 0.83206228, Log Avg loss: 0.91679446, Global Avg Loss: 4.15712525, Time: 0.0207 Steps: 8850, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000344, Sample Num: 5504, Cur Loss: 0.67025602, Cur Avg Loss: 0.82699969, Log Avg loss: 0.65790927, Global Avg Loss: 4.15317580, Time: 0.0208 Steps: 8860, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000354, Sample Num: 5664, Cur Loss: 0.99885023, Cur Avg Loss: 0.82523042, Log Avg loss: 0.76436739, Global Avg Loss: 4.14935527, Time: 0.0208 Steps: 8870, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000364, Sample Num: 5824, Cur Loss: 0.34825844, Cur Avg Loss: 0.82408773, Log Avg loss: 0.78363664, Global Avg Loss: 4.14556504, Time: 0.0207 Steps: 8880, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000374, Sample Num: 5984, Cur Loss: 0.69728374, Cur Avg Loss: 0.82110309, Log Avg loss: 0.71246211, Global Avg Loss: 4.14170329, Time: 0.0207 Steps: 8890, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000384, Sample Num: 6144, Cur Loss: 0.45316643, Cur Avg Loss: 0.81550974, Log Avg loss: 0.60631828, Global Avg Loss: 4.13773094, Time: 0.0208 Steps: 8900, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000394, Sample Num: 6304, Cur Loss: 0.66802657, Cur Avg Loss: 0.81473205, Log Avg loss: 0.78486889, Global Avg Loss: 4.13396791, Time: 0.0207 Steps: 8910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000404, Sample Num: 6464, Cur Loss: 0.93242776, Cur Avg Loss: 0.81345316, Log Avg loss: 0.76306471, Global Avg Loss: 4.13018887, Time: 0.0208 Steps: 8920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000414, Sample Num: 6624, Cur Loss: 0.79533654, Cur Avg Loss: 0.81085337, Log Avg loss: 0.70582189, Global Avg Loss: 4.12635419, Time: 0.0208 Steps: 8930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000424, Sample Num: 6784, Cur Loss: 0.97832751, Cur Avg Loss: 0.81033411, Log Avg loss: 0.78883672, Global Avg Loss: 4.12262095, Time: 0.0208 Steps: 8940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000434, Sample Num: 6944, Cur Loss: 0.66778159, Cur Avg Loss: 0.80867061, Log Avg loss: 0.73813835, Global Avg Loss: 4.11883941, Time: 0.0208 Steps: 8950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000444, Sample Num: 7104, Cur Loss: 0.48435849, Cur Avg Loss: 0.81124043, Log Avg loss: 0.92277053, Global Avg Loss: 4.11527237, Time: 0.0208 Steps: 8960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000454, Sample Num: 7264, Cur Loss: 0.43187207, Cur Avg Loss: 0.81145495, Log Avg loss: 0.82097965, Global Avg Loss: 4.11159980, Time: 0.0208 Steps: 8970, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000464, Sample Num: 7424, Cur Loss: 0.51602554, Cur Avg Loss: 0.80868035, Log Avg loss: 0.68271341, Global Avg Loss: 4.10778144, Time: 0.0209 Steps: 8980, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000474, Sample Num: 7584, Cur Loss: 1.02272248, Cur Avg Loss: 0.80588902, Log Avg loss: 0.67637136, Global Avg Loss: 4.10396452, Time: 0.0208 Steps: 8990, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000484, Sample Num: 7744, Cur Loss: 0.62698257, Cur Avg Loss: 0.80298328, Log Avg loss: 0.66525120, Global Avg Loss: 4.10014373, Time: 0.0208 Steps: 9000, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000494, Sample Num: 7904, Cur Loss: 0.84057921, Cur Avg Loss: 0.80445809, Log Avg loss: 0.87583922, Global Avg Loss: 4.09656515, Time: 0.0208 Steps: 9010, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000504, Sample Num: 8064, Cur Loss: 1.42271662, Cur Avg Loss: 0.80486368, Log Avg loss: 0.82489949, Global Avg Loss: 4.09293802, Time: 0.0212 Steps: 9020, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000514, Sample Num: 8224, Cur Loss: 1.43045259, Cur Avg Loss: 0.80192055, Log Avg loss: 0.65358697, Global Avg Loss: 4.08912922, Time: 0.0245 Steps: 9030, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000524, Sample Num: 8384, Cur Loss: 1.63880444, Cur Avg Loss: 0.80633290, Log Avg loss: 1.03312789, Global Avg Loss: 4.08574868, Time: 0.0209 Steps: 9040, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000534, Sample Num: 8544, Cur Loss: 0.71033162, Cur Avg Loss: 0.80601627, Log Avg loss: 0.78942452, Global Avg Loss: 4.08210634, Time: 0.0208 Steps: 9050, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000544, Sample Num: 8704, Cur Loss: 0.37939700, Cur Avg Loss: 0.80501161, Log Avg loss: 0.75136290, Global Avg Loss: 4.07843002, Time: 0.0208 Steps: 9060, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000554, Sample Num: 8864, Cur Loss: 0.57187545, Cur Avg Loss: 0.80575720, Log Avg loss: 0.84631746, Global Avg Loss: 4.07486650, Time: 0.0208 Steps: 9070, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000564, Sample Num: 9024, Cur Loss: 0.59522694, Cur Avg Loss: 0.80057403, Log Avg loss: 0.51342607, Global Avg Loss: 4.07094421, Time: 0.0208 Steps: 9080, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000574, Sample Num: 9184, Cur Loss: 1.13100648, Cur Avg Loss: 0.79987647, Log Avg loss: 0.76053410, Global Avg Loss: 4.06730239, Time: 0.0208 Steps: 9090, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000584, Sample Num: 9344, Cur Loss: 0.39787161, Cur Avg Loss: 0.79678668, Log Avg loss: 0.61943293, Global Avg Loss: 4.06351353, Time: 0.0208 Steps: 9100, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000594, Sample Num: 9504, Cur Loss: 0.49806401, Cur Avg Loss: 0.79661542, Log Avg loss: 0.78661378, Global Avg Loss: 4.05991649, Time: 0.0207 Steps: 9110, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000604, Sample Num: 9664, Cur Loss: 1.49765706, Cur Avg Loss: 0.79708179, Log Avg loss: 0.82478443, Global Avg Loss: 4.05636920, Time: 0.0208 Steps: 9120, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000614, Sample Num: 9824, Cur Loss: 0.67015493, Cur Avg Loss: 0.79482333, Log Avg loss: 0.65841223, Global Avg Loss: 4.05264745, Time: 0.0208 Steps: 9130, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000624, Sample Num: 9984, Cur Loss: 1.01796079, Cur Avg Loss: 0.79550133, Log Avg loss: 0.83713061, Global Avg Loss: 4.04912938, Time: 0.0208 Steps: 9140, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000634, Sample Num: 10144, Cur Loss: 1.05211818, Cur Avg Loss: 0.79843992, Log Avg loss: 0.98180755, Global Avg Loss: 4.04577711, Time: 0.0207 Steps: 9150, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000644, Sample Num: 10304, Cur Loss: 0.24651298, Cur Avg Loss: 0.79571749, Log Avg loss: 0.62311560, Global Avg Loss: 4.04204058, Time: 0.0208 Steps: 9160, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000654, Sample Num: 10464, Cur Loss: 0.76069242, Cur Avg Loss: 0.79524171, Log Avg loss: 0.76460144, Global Avg Loss: 4.03846649, Time: 0.0208 Steps: 9170, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000664, Sample Num: 10624, Cur Loss: 0.32181993, Cur Avg Loss: 0.79137435, Log Avg loss: 0.53844909, Global Avg Loss: 4.03465384, Time: 0.0208 Steps: 9180, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000674, Sample Num: 10784, Cur Loss: 0.27385476, Cur Avg Loss: 0.78886037, Log Avg loss: 0.62193188, Global Avg Loss: 4.03094032, Time: 0.0208 Steps: 9190, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000684, Sample Num: 10944, Cur Loss: 0.43103045, Cur Avg Loss: 0.78372369, Log Avg loss: 0.43751153, Global Avg Loss: 4.02703442, Time: 0.0208 Steps: 9200, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000694, Sample Num: 11104, Cur Loss: 0.78858507, Cur Avg Loss: 0.78255764, Log Avg loss: 0.70279971, Global Avg Loss: 4.02342504, Time: 0.0208 Steps: 9210, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000704, Sample Num: 11264, Cur Loss: 0.55922568, Cur Avg Loss: 0.78198918, Log Avg loss: 0.74253839, Global Avg Loss: 4.01986660, Time: 0.0208 Steps: 9220, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000714, Sample Num: 11424, Cur Loss: 0.48339164, Cur Avg Loss: 0.77941032, Log Avg loss: 0.59785856, Global Avg Loss: 4.01615911, Time: 0.0208 Steps: 9230, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000724, Sample Num: 11584, Cur Loss: 0.33901960, Cur Avg Loss: 0.77656999, Log Avg loss: 0.57377059, Global Avg Loss: 4.01243359, Time: 0.0208 Steps: 9240, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000734, Sample Num: 11744, Cur Loss: 0.31664866, Cur Avg Loss: 0.77681834, Log Avg loss: 0.79479870, Global Avg Loss: 4.00895506, Time: 0.0209 Steps: 9250, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000744, Sample Num: 11904, Cur Loss: 0.51738799, Cur Avg Loss: 0.77932430, Log Avg loss: 0.96326177, Global Avg Loss: 4.00566598, Time: 0.0208 Steps: 9260, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000754, Sample Num: 12064, Cur Loss: 1.16607428, Cur Avg Loss: 0.78033993, Log Avg loss: 0.85590238, Global Avg Loss: 4.00226817, Time: 0.0208 Steps: 9270, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000764, Sample Num: 12224, Cur Loss: 1.33271277, Cur Avg Loss: 0.77952402, Log Avg loss: 0.71800472, Global Avg Loss: 3.99872910, Time: 0.0208 Steps: 9280, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000774, Sample Num: 12384, Cur Loss: 0.51147747, Cur Avg Loss: 0.77750584, Log Avg loss: 0.62331661, Global Avg Loss: 3.99509571, Time: 0.0209 Steps: 9290, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000784, Sample Num: 12544, Cur Loss: 0.71334147, Cur Avg Loss: 0.77544270, Log Avg loss: 0.61575632, Global Avg Loss: 3.99146202, Time: 0.0208 Steps: 9300, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000794, Sample Num: 12704, Cur Loss: 1.17450666, Cur Avg Loss: 0.77567471, Log Avg loss: 0.79386397, Global Avg Loss: 3.98802743, Time: 0.0207 Steps: 9310, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000804, Sample Num: 12864, Cur Loss: 0.90035379, Cur Avg Loss: 0.77462812, Log Avg loss: 0.69152879, Global Avg Loss: 3.98449042, Time: 0.0207 Steps: 9320, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000814, Sample Num: 13024, Cur Loss: 0.62116522, Cur Avg Loss: 0.77408681, Log Avg loss: 0.73056591, Global Avg Loss: 3.98100282, Time: 0.0207 Steps: 9330, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000824, Sample Num: 13184, Cur Loss: 0.94526571, Cur Avg Loss: 0.77271757, Log Avg loss: 0.66126117, Global Avg Loss: 3.97744849, Time: 0.0208 Steps: 9340, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000834, Sample Num: 13344, Cur Loss: 1.22018778, Cur Avg Loss: 0.77279968, Log Avg loss: 0.77956574, Global Avg Loss: 3.97402830, Time: 0.0207 Steps: 9350, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000844, Sample Num: 13504, Cur Loss: 0.95666748, Cur Avg Loss: 0.77300144, Log Avg loss: 0.78982759, Global Avg Loss: 3.97062638, Time: 0.0208 Steps: 9360, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000854, Sample Num: 13664, Cur Loss: 1.32031584, Cur Avg Loss: 0.77704902, Log Avg loss: 1.11866523, Global Avg Loss: 3.96758266, Time: 0.0208 Steps: 9370, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000864, Sample Num: 13824, Cur Loss: 0.69295514, Cur Avg Loss: 0.77919564, Log Avg loss: 0.96251655, Global Avg Loss: 3.96437896, Time: 0.0208 Steps: 9380, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000874, Sample Num: 13984, Cur Loss: 1.25839067, Cur Avg Loss: 0.77894514, Log Avg loss: 0.75730242, Global Avg Loss: 3.96096355, Time: 0.0207 Steps: 9390, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000884, Sample Num: 14144, Cur Loss: 1.45279825, Cur Avg Loss: 0.78173426, Log Avg loss: 1.02550359, Global Avg Loss: 3.95784072, Time: 0.0207 Steps: 9400, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000894, Sample Num: 14304, Cur Loss: 1.53659391, Cur Avg Loss: 0.78238152, Log Avg loss: 0.83959924, Global Avg Loss: 3.95452697, Time: 0.0207 Steps: 9410, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000904, Sample Num: 14464, Cur Loss: 0.92485040, Cur Avg Loss: 0.78152025, Log Avg loss: 0.70452214, Global Avg Loss: 3.95107685, Time: 0.0208 Steps: 9420, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000914, Sample Num: 14624, Cur Loss: 1.26635027, Cur Avg Loss: 0.78093579, Log Avg loss: 0.72810128, Global Avg Loss: 3.94765906, Time: 0.0207 Steps: 9430, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000924, Sample Num: 14784, Cur Loss: 0.91692102, Cur Avg Loss: 0.78031290, Log Avg loss: 0.72338019, Global Avg Loss: 3.94424351, Time: 0.0208 Steps: 9440, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000934, Sample Num: 14944, Cur Loss: 1.45300210, Cur Avg Loss: 0.78147371, Log Avg loss: 0.88873230, Global Avg Loss: 3.94101017, Time: 0.0208 Steps: 9450, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000944, Sample Num: 15104, Cur Loss: 0.65122151, Cur Avg Loss: 0.78095617, Log Avg loss: 0.73261850, Global Avg Loss: 3.93761864, Time: 0.0207 Steps: 9460, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000954, Sample Num: 15264, Cur Loss: 1.09623528, Cur Avg Loss: 0.78228457, Log Avg loss: 0.90768555, Global Avg Loss: 3.93441913, Time: 0.0208 Steps: 9470, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000964, Sample Num: 15424, Cur Loss: 0.35267851, Cur Avg Loss: 0.78107968, Log Avg loss: 0.66613273, Global Avg Loss: 3.93097157, Time: 0.0207 Steps: 9480, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000974, Sample Num: 15584, Cur Loss: 0.80379611, Cur Avg Loss: 0.78066003, Log Avg loss: 0.74020607, Global Avg Loss: 3.92760933, Time: 0.0207 Steps: 9490, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000984, Sample Num: 15744, Cur Loss: 1.61649323, Cur Avg Loss: 0.78580926, Log Avg loss: 1.28734438, Global Avg Loss: 3.92483010, Time: 0.0207 Steps: 9500, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000994, Sample Num: 15904, Cur Loss: 1.03553605, Cur Avg Loss: 0.78502442, Log Avg loss: 0.70779612, Global Avg Loss: 3.92144731, Time: 0.0207 Steps: 9510, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001004, Sample Num: 16064, Cur Loss: 0.55893105, Cur Avg Loss: 0.78655830, Log Avg loss: 0.93902594, Global Avg Loss: 3.91831452, Time: 0.0208 Steps: 9520, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001014, Sample Num: 16224, Cur Loss: 0.56802678, Cur Avg Loss: 0.78582848, Log Avg loss: 0.71255400, Global Avg Loss: 3.91495065, Time: 0.0207 Steps: 9530, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001024, Sample Num: 16384, Cur Loss: 1.08183718, Cur Avg Loss: 0.78644533, Log Avg loss: 0.84899466, Global Avg Loss: 3.91173686, Time: 0.0253 Steps: 9540, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001034, Sample Num: 16544, Cur Loss: 0.58423543, Cur Avg Loss: 0.78481716, Log Avg loss: 0.61809232, Global Avg Loss: 3.90828802, Time: 0.0208 Steps: 9550, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001044, Sample Num: 16704, Cur Loss: 1.05699730, Cur Avg Loss: 0.78586261, Log Avg loss: 0.89396223, Global Avg Loss: 3.90513496, Time: 0.0208 Steps: 9560, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001054, Sample Num: 16864, Cur Loss: 1.72299790, Cur Avg Loss: 0.78604327, Log Avg loss: 0.80490382, Global Avg Loss: 3.90189543, Time: 0.0207 Steps: 9570, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001064, Sample Num: 17024, Cur Loss: 0.73961425, Cur Avg Loss: 0.78525432, Log Avg loss: 0.70209970, Global Avg Loss: 3.89855535, Time: 0.0208 Steps: 9580, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001074, Sample Num: 17184, Cur Loss: 1.63851368, Cur Avg Loss: 0.78546463, Log Avg loss: 0.80784079, Global Avg Loss: 3.89533250, Time: 0.0207 Steps: 9590, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001084, Sample Num: 17344, Cur Loss: 0.53426337, Cur Avg Loss: 0.78428319, Log Avg loss: 0.65739673, Global Avg Loss: 3.89195965, Time: 0.0207 Steps: 9600, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001094, Sample Num: 17504, Cur Loss: 0.86062455, Cur Avg Loss: 0.78317511, Log Avg loss: 0.66305923, Global Avg Loss: 3.88859971, Time: 0.0208 Steps: 9610, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001104, Sample Num: 17664, Cur Loss: 0.85565466, Cur Avg Loss: 0.78365404, Log Avg loss: 0.83604913, Global Avg Loss: 3.88542658, Time: 0.0208 Steps: 9620, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001114, Sample Num: 17824, Cur Loss: 0.55011678, Cur Avg Loss: 0.78354540, Log Avg loss: 0.77155113, Global Avg Loss: 3.88219307, Time: 0.0208 Steps: 9630, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001124, Sample Num: 17984, Cur Loss: 0.95544732, Cur Avg Loss: 0.78327172, Log Avg loss: 0.75278422, Global Avg Loss: 3.87894679, Time: 0.0207 Steps: 9640, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001134, Sample Num: 18144, Cur Loss: 1.05178106, Cur Avg Loss: 0.78364106, Log Avg loss: 0.82515436, Global Avg Loss: 3.87578224, Time: 0.0208 Steps: 9650, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001144, Sample Num: 18304, Cur Loss: 0.32839906, Cur Avg Loss: 0.78226065, Log Avg loss: 0.62572241, Global Avg Loss: 3.87241779, Time: 0.0208 Steps: 9660, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001154, Sample Num: 18464, Cur Loss: 0.59208757, Cur Avg Loss: 0.78458308, Log Avg loss: 1.05026903, Global Avg Loss: 3.86949933, Time: 0.0209 Steps: 9670, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001164, Sample Num: 18624, Cur Loss: 0.40101004, Cur Avg Loss: 0.78214119, Log Avg loss: 0.50034717, Global Avg Loss: 3.86601880, Time: 0.0208 Steps: 9680, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001174, Sample Num: 18784, Cur Loss: 1.35260868, Cur Avg Loss: 0.78220328, Log Avg loss: 0.78943082, Global Avg Loss: 3.86284379, Time: 0.0208 Steps: 9690, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001184, Sample Num: 18944, Cur Loss: 0.92784083, Cur Avg Loss: 0.78255222, Log Avg loss: 0.82351770, Global Avg Loss: 3.85971046, Time: 0.0208 Steps: 9700, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001194, Sample Num: 19104, Cur Loss: 0.66878259, Cur Avg Loss: 0.78125191, Log Avg loss: 0.62729569, Global Avg Loss: 3.85638151, Time: 0.0207 Steps: 9710, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001204, Sample Num: 19264, Cur Loss: 0.62216210, Cur Avg Loss: 0.77891396, Log Avg loss: 0.49976234, Global Avg Loss: 3.85292820, Time: 0.0208 Steps: 9720, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001214, Sample Num: 19424, Cur Loss: 0.30092472, Cur Avg Loss: 0.77860928, Log Avg loss: 0.74192535, Global Avg Loss: 3.84973087, Time: 0.0207 Steps: 9730, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001224, Sample Num: 19584, Cur Loss: 0.35999450, Cur Avg Loss: 0.77699726, Log Avg loss: 0.58129881, Global Avg Loss: 3.84637519, Time: 0.0208 Steps: 9740, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001234, Sample Num: 19744, Cur Loss: 0.41542542, Cur Avg Loss: 0.77467050, Log Avg loss: 0.48987414, Global Avg Loss: 3.84293262, Time: 0.0207 Steps: 9750, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001244, Sample Num: 19904, Cur Loss: 0.63024139, Cur Avg Loss: 0.77581973, Log Avg loss: 0.91763551, Global Avg Loss: 3.83993539, Time: 0.0208 Steps: 9760, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001254, Sample Num: 20064, Cur Loss: 0.34057248, Cur Avg Loss: 0.77409916, Log Avg loss: 0.56005988, Global Avg Loss: 3.83657830, Time: 0.0208 Steps: 9770, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001264, Sample Num: 20224, Cur Loss: 0.53800386, Cur Avg Loss: 0.77370690, Log Avg loss: 0.72451733, Global Avg Loss: 3.83339624, Time: 0.0208 Steps: 9780, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001274, Sample Num: 20384, Cur Loss: 0.53258598, Cur Avg Loss: 0.77326622, Log Avg loss: 0.71756488, Global Avg Loss: 3.83021357, Time: 0.0208 Steps: 9790, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001284, Sample Num: 20544, Cur Loss: 0.41202939, Cur Avg Loss: 0.77142323, Log Avg loss: 0.53662606, Global Avg Loss: 3.82685276, Time: 0.0208 Steps: 9800, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001294, Sample Num: 20704, Cur Loss: 0.74199474, Cur Avg Loss: 0.77047653, Log Avg loss: 0.64892025, Global Avg Loss: 3.82361328, Time: 0.0207 Steps: 9810, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001304, Sample Num: 20864, Cur Loss: 0.61554968, Cur Avg Loss: 0.77076733, Log Avg loss: 0.80839645, Global Avg Loss: 3.82054280, Time: 0.0207 Steps: 9820, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001314, Sample Num: 21024, Cur Loss: 0.30125439, Cur Avg Loss: 0.77034446, Log Avg loss: 0.71520287, Global Avg Loss: 3.81738375, Time: 0.0207 Steps: 9830, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001324, Sample Num: 21184, Cur Loss: 0.26633915, Cur Avg Loss: 0.77034499, Log Avg loss: 0.77041404, Global Avg Loss: 3.81428724, Time: 0.0207 Steps: 9840, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001334, Sample Num: 21344, Cur Loss: 0.88598371, Cur Avg Loss: 0.77161338, Log Avg loss: 0.93954818, Global Avg Loss: 3.81136872, Time: 0.0207 Steps: 9850, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001344, Sample Num: 21504, Cur Loss: 0.65996945, Cur Avg Loss: 0.77255775, Log Avg loss: 0.89853625, Global Avg Loss: 3.80841453, Time: 0.0207 Steps: 9860, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001354, Sample Num: 21664, Cur Loss: 0.74792707, Cur Avg Loss: 0.77253912, Log Avg loss: 0.77003539, Global Avg Loss: 3.80533613, Time: 0.0207 Steps: 9870, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001364, Sample Num: 21824, Cur Loss: 0.44508174, Cur Avg Loss: 0.77244868, Log Avg loss: 0.76020403, Global Avg Loss: 3.80225401, Time: 0.0208 Steps: 9880, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001374, Sample Num: 21984, Cur Loss: 0.56377065, Cur Avg Loss: 0.77335821, Log Avg loss: 0.89741811, Global Avg Loss: 3.79931687, Time: 0.0207 Steps: 9890, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001384, Sample Num: 22144, Cur Loss: 0.81889498, Cur Avg Loss: 0.77246243, Log Avg loss: 0.64938246, Global Avg Loss: 3.79613512, Time: 0.0207 Steps: 9900, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001394, Sample Num: 22304, Cur Loss: 0.54587984, Cur Avg Loss: 0.77320874, Log Avg loss: 0.87649804, Global Avg Loss: 3.79318897, Time: 0.0208 Steps: 9910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001404, Sample Num: 22464, Cur Loss: 0.60745513, Cur Avg Loss: 0.77327523, Log Avg loss: 0.78254303, Global Avg Loss: 3.79015404, Time: 0.0207 Steps: 9920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001414, Sample Num: 22624, Cur Loss: 0.29220229, Cur Avg Loss: 0.77128012, Log Avg loss: 0.49116742, Global Avg Loss: 3.78683180, Time: 0.0207 Steps: 9930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001424, Sample Num: 22784, Cur Loss: 1.11318254, Cur Avg Loss: 0.77140903, Log Avg loss: 0.78963578, Global Avg Loss: 3.78381651, Time: 0.0207 Steps: 9940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001434, Sample Num: 22944, Cur Loss: 0.43692479, Cur Avg Loss: 0.77063125, Log Avg loss: 0.65987579, Global Avg Loss: 3.78067687, Time: 0.0208 Steps: 9950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001444, Sample Num: 23104, Cur Loss: 0.65953708, Cur Avg Loss: 0.76987436, Log Avg loss: 0.66133659, Global Avg Loss: 3.77754500, Time: 0.0207 Steps: 9960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001454, Sample Num: 23264, Cur Loss: 0.64742732, Cur Avg Loss: 0.77096014, Log Avg loss: 0.92774654, Global Avg Loss: 3.77468663, Time: 0.0207 Steps: 9970, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001464, Sample Num: 23424, Cur Loss: 0.68761635, Cur Avg Loss: 0.77233460, Log Avg loss: 0.97218183, Global Avg Loss: 3.77187851, Time: 0.0208 Steps: 9980, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001474, Sample Num: 23584, Cur Loss: 0.68525296, Cur Avg Loss: 0.77401164, Log Avg loss: 1.01953049, Global Avg Loss: 3.76912341, Time: 0.0209 Steps: 9990, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001484, Sample Num: 23744, Cur Loss: 0.94956923, Cur Avg Loss: 0.77649651, Log Avg loss: 1.14276492, Global Avg Loss: 3.76649705, Time: 0.0208 Steps: 10000, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001494, Sample Num: 23904, Cur Loss: 0.36068642, Cur Avg Loss: 0.77508326, Log Avg loss: 0.56535698, Global Avg Loss: 3.76329911, Time: 0.0207 Steps: 10010, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001504, Sample Num: 24064, Cur Loss: 0.79004693, Cur Avg Loss: 0.77490104, Log Avg loss: 0.74767737, Global Avg Loss: 3.76028950, Time: 0.0208 Steps: 10020, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001514, Sample Num: 24224, Cur Loss: 0.99104452, Cur Avg Loss: 0.77542790, Log Avg loss: 0.85466795, Global Avg Loss: 3.75739257, Time: 0.0207 Steps: 10030, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001524, Sample Num: 24384, Cur Loss: 0.72463125, Cur Avg Loss: 0.77559653, Log Avg loss: 0.80112785, Global Avg Loss: 3.75444809, Time: 0.0207 Steps: 10040, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001534, Sample Num: 24544, Cur Loss: 0.68884230, Cur Avg Loss: 0.77453283, Log Avg loss: 0.61242510, Global Avg Loss: 3.75132169, Time: 0.0207 Steps: 10050, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001544, Sample Num: 24704, Cur Loss: 0.36756742, Cur Avg Loss: 0.77324134, Log Avg loss: 0.57512669, Global Avg Loss: 3.74816444, Time: 0.0209 Steps: 10060, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001554, Sample Num: 24864, Cur Loss: 1.27747500, Cur Avg Loss: 0.77392172, Log Avg loss: 0.87897200, Global Avg Loss: 3.74531520, Time: 0.0208 Steps: 10070, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001564, Sample Num: 25024, Cur Loss: 0.47101262, Cur Avg Loss: 0.77422172, Log Avg loss: 0.82084118, Global Avg Loss: 3.74241393, Time: 0.0210 Steps: 10080, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001574, Sample Num: 25184, Cur Loss: 0.22693187, Cur Avg Loss: 0.77228624, Log Avg loss: 0.46957843, Global Avg Loss: 3.73917029, Time: 0.0210 Steps: 10090, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001584, Sample Num: 25344, Cur Loss: 0.53798789, Cur Avg Loss: 0.77072687, Log Avg loss: 0.52528087, Global Avg Loss: 3.73598822, Time: 0.0209 Steps: 10100, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001594, Sample Num: 25504, Cur Loss: 0.83749998, Cur Avg Loss: 0.77035886, Log Avg loss: 0.71206644, Global Avg Loss: 3.73299720, Time: 0.0208 Steps: 10110, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001604, Sample Num: 25664, Cur Loss: 0.72527003, Cur Avg Loss: 0.77032345, Log Avg loss: 0.76467916, Global Avg Loss: 3.73006408, Time: 0.0209 Steps: 10120, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001614, Sample Num: 25824, Cur Loss: 0.32969654, Cur Avg Loss: 0.77050137, Log Avg loss: 0.79904051, Global Avg Loss: 3.72717067, Time: 0.0209 Steps: 10130, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001624, Sample Num: 25984, Cur Loss: 0.33710858, Cur Avg Loss: 0.76937562, Log Avg loss: 0.58767891, Global Avg Loss: 3.72407452, Time: 0.0209 Steps: 10140, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001634, Sample Num: 26144, Cur Loss: 0.80702764, Cur Avg Loss: 0.76942717, Log Avg loss: 0.77779906, Global Avg Loss: 3.72117179, Time: 0.0209 Steps: 10150, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001644, Sample Num: 26304, Cur Loss: 0.64977503, Cur Avg Loss: 0.76798481, Log Avg loss: 0.53230257, Global Avg Loss: 3.71803314, Time: 0.0209 Steps: 10160, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001654, Sample Num: 26464, Cur Loss: 1.33294201, Cur Avg Loss: 0.76791930, Log Avg loss: 0.75715063, Global Avg Loss: 3.71512175, Time: 0.0208 Steps: 10170, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001664, Sample Num: 26624, Cur Loss: 0.42754105, Cur Avg Loss: 0.76830453, Log Avg loss: 0.83202039, Global Avg Loss: 3.71228963, Time: 0.0210 Steps: 10180, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001674, Sample Num: 26784, Cur Loss: 1.18443322, Cur Avg Loss: 0.76761061, Log Avg loss: 0.65214312, Global Avg Loss: 3.70928654, Time: 0.0208 Steps: 10190, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001684, Sample Num: 26944, Cur Loss: 1.58108616, Cur Avg Loss: 0.76774721, Log Avg loss: 0.79061340, Global Avg Loss: 3.70642509, Time: 0.0209 Steps: 10200, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001694, Sample Num: 27104, Cur Loss: 0.35035050, Cur Avg Loss: 0.76844702, Log Avg loss: 0.88629531, Global Avg Loss: 3.70366297, Time: 0.0209 Steps: 10210, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001704, Sample Num: 27264, Cur Loss: 0.82173282, Cur Avg Loss: 0.76815567, Log Avg loss: 0.71880054, Global Avg Loss: 3.70074236, Time: 0.0208 Steps: 10220, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001714, Sample Num: 27424, Cur Loss: 0.79442650, Cur Avg Loss: 0.76838648, Log Avg loss: 0.80771612, Global Avg Loss: 3.69791438, Time: 0.0209 Steps: 10230, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001724, Sample Num: 27584, Cur Loss: 0.91889971, Cur Avg Loss: 0.76712814, Log Avg loss: 0.55144903, Global Avg Loss: 3.69484166, Time: 0.0209 Steps: 10240, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001734, Sample Num: 27744, Cur Loss: 1.58451533, Cur Avg Loss: 0.76782917, Log Avg loss: 0.88868682, Global Avg Loss: 3.69210395, Time: 0.0209 Steps: 10250, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001744, Sample Num: 27904, Cur Loss: 0.28161824, Cur Avg Loss: 0.76714904, Log Avg loss: 0.64921529, Global Avg Loss: 3.68913817, Time: 0.0208 Steps: 10260, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001754, Sample Num: 28064, Cur Loss: 0.62333059, Cur Avg Loss: 0.76677818, Log Avg loss: 0.70209929, Global Avg Loss: 3.68622966, Time: 0.0208 Steps: 10270, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001764, Sample Num: 28224, Cur Loss: 0.37396622, Cur Avg Loss: 0.76501007, Log Avg loss: 0.45488380, Global Avg Loss: 3.68308633, Time: 0.0208 Steps: 10280, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001774, Sample Num: 28384, Cur Loss: 0.48666126, Cur Avg Loss: 0.76445279, Log Avg loss: 0.66614893, Global Avg Loss: 3.68015441, Time: 0.0209 Steps: 10290, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001784, Sample Num: 28544, Cur Loss: 1.35844409, Cur Avg Loss: 0.76481401, Log Avg loss: 0.82889393, Global Avg Loss: 3.67738620, Time: 0.0209 Steps: 10300, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001794, Sample Num: 28704, Cur Loss: 0.57200885, Cur Avg Loss: 0.76394937, Log Avg loss: 0.60969790, Global Avg Loss: 3.67441075, Time: 0.0248 Steps: 10310, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001804, Sample Num: 28864, Cur Loss: 1.23189330, Cur Avg Loss: 0.76351487, Log Avg loss: 0.68556602, Global Avg Loss: 3.67151458, Time: 0.0210 Steps: 10320, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001814, Sample Num: 29024, Cur Loss: 0.42786378, Cur Avg Loss: 0.76377956, Log Avg loss: 0.81152897, Global Avg Loss: 3.66874596, Time: 0.0209 Steps: 10330, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001824, Sample Num: 29184, Cur Loss: 1.24691939, Cur Avg Loss: 0.76372559, Log Avg loss: 0.75393571, Global Avg Loss: 3.66592700, Time: 0.0210 Steps: 10340, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001834, Sample Num: 29344, Cur Loss: 0.78570145, Cur Avg Loss: 0.76378908, Log Avg loss: 0.77536925, Global Avg Loss: 3.66313419, Time: 0.0210 Steps: 10350, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001844, Sample Num: 29504, Cur Loss: 0.91914594, Cur Avg Loss: 0.76372437, Log Avg loss: 0.75185722, Global Avg Loss: 3.66032407, Time: 0.0210 Steps: 10360, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001854, Sample Num: 29664, Cur Loss: 1.56521475, Cur Avg Loss: 0.76361603, Log Avg loss: 0.74363788, Global Avg Loss: 3.65751145, Time: 0.0210 Steps: 10370, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001864, Sample Num: 29824, Cur Loss: 1.61439848, Cur Avg Loss: 0.76598590, Log Avg loss: 1.20535974, Global Avg Loss: 3.65514907, Time: 0.0209 Steps: 10380, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001874, Sample Num: 29984, Cur Loss: 2.34847116, Cur Avg Loss: 0.76982472, Log Avg loss: 1.48538090, Global Avg Loss: 3.65306075, Time: 0.0209 Steps: 10390, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001884, Sample Num: 30144, Cur Loss: 0.42418718, Cur Avg Loss: 0.77132757, Log Avg loss: 1.05296131, Global Avg Loss: 3.65056065, Time: 0.0210 Steps: 10400, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001894, Sample Num: 30304, Cur Loss: 1.65506136, Cur Avg Loss: 0.77140076, Log Avg loss: 0.78519048, Global Avg Loss: 3.64780814, Time: 0.0209 Steps: 10410, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001904, Sample Num: 30464, Cur Loss: 1.54809189, Cur Avg Loss: 0.77123248, Log Avg loss: 0.73936045, Global Avg Loss: 3.64501692, Time: 0.0210 Steps: 10420, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001914, Sample Num: 30624, Cur Loss: 0.38939875, Cur Avg Loss: 0.77144289, Log Avg loss: 0.81150461, Global Avg Loss: 3.64230023, Time: 0.0210 Steps: 10430, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001924, Sample Num: 30784, Cur Loss: 0.40506092, Cur Avg Loss: 0.77071160, Log Avg loss: 0.63074296, Global Avg Loss: 3.63941559, Time: 0.0209 Steps: 10440, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001934, Sample Num: 30944, Cur Loss: 0.58196259, Cur Avg Loss: 0.76934930, Log Avg loss: 0.50724138, Global Avg Loss: 3.63641830, Time: 0.0209 Steps: 10450, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001944, Sample Num: 31104, Cur Loss: 0.91339034, Cur Avg Loss: 0.76985110, Log Avg loss: 0.86690061, Global Avg Loss: 3.63377057, Time: 0.0209 Steps: 10460, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001954, Sample Num: 31264, Cur Loss: 0.47982788, Cur Avg Loss: 0.76928295, Log Avg loss: 0.65883414, Global Avg Loss: 3.63092918, Time: 0.0210 Steps: 10470, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001964, Sample Num: 31424, Cur Loss: 0.62040526, Cur Avg Loss: 0.76905638, Log Avg loss: 0.72478413, Global Avg Loss: 3.62815614, Time: 0.0210 Steps: 10480, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001974, Sample Num: 31584, Cur Loss: 0.51732320, Cur Avg Loss: 0.76893770, Log Avg loss: 0.74562863, Global Avg Loss: 3.62540826, Time: 0.0209 Steps: 10490, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001984, Sample Num: 31744, Cur Loss: 0.49991444, Cur Avg Loss: 0.76911752, Log Avg loss: 0.80461468, Global Avg Loss: 3.62272179, Time: 0.0209 Steps: 10500, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001994, Sample Num: 31904, Cur Loss: 0.80306852, Cur Avg Loss: 0.76824063, Log Avg loss: 0.59426533, Global Avg Loss: 3.61984029, Time: 0.0209 Steps: 10510, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002004, Sample Num: 32064, Cur Loss: 0.16093098, Cur Avg Loss: 0.76751070, Log Avg loss: 0.62196279, Global Avg Loss: 3.61699060, Time: 0.0210 Steps: 10520, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002014, Sample Num: 32224, Cur Loss: 0.69987297, Cur Avg Loss: 0.76836250, Log Avg loss: 0.93906277, Global Avg Loss: 3.61444746, Time: 0.0210 Steps: 10530, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002024, Sample Num: 32384, Cur Loss: 0.80474627, Cur Avg Loss: 0.76731685, Log Avg loss: 0.55672311, Global Avg Loss: 3.61154639, Time: 0.0209 Steps: 10540, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002034, Sample Num: 32544, Cur Loss: 1.33866334, Cur Avg Loss: 0.76721390, Log Avg loss: 0.74637793, Global Avg Loss: 3.60883059, Time: 0.0209 Steps: 10550, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002044, Sample Num: 32704, Cur Loss: 0.69235647, Cur Avg Loss: 0.76760734, Log Avg loss: 0.84763194, Global Avg Loss: 3.60621582, Time: 0.0209 Steps: 10560, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002054, Sample Num: 32864, Cur Loss: 0.82041270, Cur Avg Loss: 0.76722662, Log Avg loss: 0.68940801, Global Avg Loss: 3.60345630, Time: 0.0210 Steps: 10570, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002064, Sample Num: 33024, Cur Loss: 1.10361862, Cur Avg Loss: 0.76653421, Log Avg loss: 0.62431218, Global Avg Loss: 3.60064048, Time: 0.0209 Steps: 10580, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002074, Sample Num: 33184, Cur Loss: 0.44879115, Cur Avg Loss: 0.76564664, Log Avg loss: 0.58245273, Global Avg Loss: 3.59779044, Time: 0.0209 Steps: 10590, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002084, Sample Num: 33344, Cur Loss: 1.01265633, Cur Avg Loss: 0.76459736, Log Avg loss: 0.54697760, Global Avg Loss: 3.59491232, Time: 0.0209 Steps: 10600, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002094, Sample Num: 33504, Cur Loss: 0.43067735, Cur Avg Loss: 0.76308301, Log Avg loss: 0.44749245, Global Avg Loss: 3.59194585, Time: 0.0209 Steps: 10610, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002104, Sample Num: 33664, Cur Loss: 0.82871133, Cur Avg Loss: 0.76238988, Log Avg loss: 0.61724847, Global Avg Loss: 3.58914482, Time: 0.0209 Steps: 10620, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002114, Sample Num: 33824, Cur Loss: 0.42869848, Cur Avg Loss: 0.76296378, Log Avg loss: 0.88371121, Global Avg Loss: 3.58659973, Time: 0.0209 Steps: 10630, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002124, Sample Num: 33984, Cur Loss: 0.58228385, Cur Avg Loss: 0.76214038, Log Avg loss: 0.58807378, Global Avg Loss: 3.58378156, Time: 0.0208 Steps: 10640, Updated lr: 0.000091 ***** Running evaluation checkpoint-10645 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-10645 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.490959, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.616968, "eval_total_loss": 433.728643, "eval_mae": 0.58162, "eval_mse": 0.61708, "eval_r2": 0.607743, "eval_sp_statistic": 0.75879, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.780746, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.749878, "test_total_loss": 376.438987, "test_mae": 0.598253, "test_mse": 0.750133, "test_r2": 0.515858, "test_sp_statistic": 0.688696, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.741531, "test_ps_pvalue": 0.0, "lr": 9.085348506401138e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.582506292145708, "train_cur_epoch_loss": 1623.129818983376, "train_cur_epoch_avg_loss": 0.7623907087756581, "train_cur_epoch_time": 44.49095940589905, "train_cur_epoch_avg_time": 0.020897585441944127, "epoch": 5, "step": 10645} ################################################## Training, Epoch: 0006, Batch: 000005, Sample Num: 80, Cur Loss: 0.61475480, Cur Avg Loss: 0.56488031, Log Avg loss: 0.71680581, Global Avg Loss: 3.58108957, Time: 0.0245 Steps: 10650, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000015, Sample Num: 240, Cur Loss: 0.53305662, Cur Avg Loss: 0.56834315, Log Avg loss: 0.57007457, Global Avg Loss: 3.57826497, Time: 0.0208 Steps: 10660, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000025, Sample Num: 400, Cur Loss: 1.21051693, Cur Avg Loss: 0.63347712, Log Avg loss: 0.73117808, Global Avg Loss: 3.57559666, Time: 0.0208 Steps: 10670, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000035, Sample Num: 560, Cur Loss: 0.72020507, Cur Avg Loss: 0.70837192, Log Avg loss: 0.89560892, Global Avg Loss: 3.57308731, Time: 0.0208 Steps: 10680, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000045, Sample Num: 720, Cur Loss: 0.63768280, Cur Avg Loss: 0.74723810, Log Avg loss: 0.88326972, Global Avg Loss: 3.57057111, Time: 0.0208 Steps: 10690, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000055, Sample Num: 880, Cur Loss: 0.55070436, Cur Avg Loss: 0.70054671, Log Avg loss: 0.49043544, Global Avg Loss: 3.56769248, Time: 0.0208 Steps: 10700, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000065, Sample Num: 1040, Cur Loss: 0.61537981, Cur Avg Loss: 0.68503904, Log Avg loss: 0.59974690, Global Avg Loss: 3.56492129, Time: 0.0208 Steps: 10710, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000075, Sample Num: 1200, Cur Loss: 0.34368742, Cur Avg Loss: 0.67301326, Log Avg loss: 0.59484564, Global Avg Loss: 3.56215070, Time: 0.0208 Steps: 10720, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000085, Sample Num: 1360, Cur Loss: 1.15175200, Cur Avg Loss: 0.70471998, Log Avg loss: 0.94252042, Global Avg Loss: 3.55970929, Time: 0.0208 Steps: 10730, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000095, Sample Num: 1520, Cur Loss: 0.74150836, Cur Avg Loss: 0.69778436, Log Avg loss: 0.63883162, Global Avg Loss: 3.55698966, Time: 0.0208 Steps: 10740, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000105, Sample Num: 1680, Cur Loss: 1.30147398, Cur Avg Loss: 0.69481894, Log Avg loss: 0.66664740, Global Avg Loss: 3.55430097, Time: 0.0209 Steps: 10750, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000115, Sample Num: 1840, Cur Loss: 1.01316547, Cur Avg Loss: 0.70528043, Log Avg loss: 0.81512614, Global Avg Loss: 3.55175527, Time: 0.0208 Steps: 10760, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000125, Sample Num: 2000, Cur Loss: 0.78235781, Cur Avg Loss: 0.71096664, Log Avg loss: 0.77635803, Global Avg Loss: 3.54917830, Time: 0.0208 Steps: 10770, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000135, Sample Num: 2160, Cur Loss: 0.57312727, Cur Avg Loss: 0.70829937, Log Avg loss: 0.67495850, Global Avg Loss: 3.54651205, Time: 0.0208 Steps: 10780, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000145, Sample Num: 2320, Cur Loss: 0.24824381, Cur Avg Loss: 0.70658356, Log Avg loss: 0.68342014, Global Avg Loss: 3.54385858, Time: 0.0208 Steps: 10790, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000155, Sample Num: 2480, Cur Loss: 0.99918187, Cur Avg Loss: 0.69878329, Log Avg loss: 0.58567938, Global Avg Loss: 3.54111953, Time: 0.0208 Steps: 10800, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000165, Sample Num: 2640, Cur Loss: 0.48343879, Cur Avg Loss: 0.69785920, Log Avg loss: 0.68353583, Global Avg Loss: 3.53847606, Time: 0.0208 Steps: 10810, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000175, Sample Num: 2800, Cur Loss: 0.51864493, Cur Avg Loss: 0.69261812, Log Avg loss: 0.60614017, Global Avg Loss: 3.53576596, Time: 0.0208 Steps: 10820, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000185, Sample Num: 2960, Cur Loss: 0.73947442, Cur Avg Loss: 0.69495695, Log Avg loss: 0.73588652, Global Avg Loss: 3.53318066, Time: 0.0208 Steps: 10830, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000195, Sample Num: 3120, Cur Loss: 0.37792787, Cur Avg Loss: 0.68256125, Log Avg loss: 0.45324081, Global Avg Loss: 3.53033938, Time: 0.0209 Steps: 10840, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000205, Sample Num: 3280, Cur Loss: 0.65044618, Cur Avg Loss: 0.68003303, Log Avg loss: 0.63073272, Global Avg Loss: 3.52766694, Time: 0.0208 Steps: 10850, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000215, Sample Num: 3440, Cur Loss: 0.54112315, Cur Avg Loss: 0.67944091, Log Avg loss: 0.66730251, Global Avg Loss: 3.52503308, Time: 0.0208 Steps: 10860, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000225, Sample Num: 3600, Cur Loss: 1.00566626, Cur Avg Loss: 0.68439995, Log Avg loss: 0.79101927, Global Avg Loss: 3.52251789, Time: 0.0208 Steps: 10870, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000235, Sample Num: 3760, Cur Loss: 0.80473799, Cur Avg Loss: 0.68466172, Log Avg loss: 0.69055150, Global Avg Loss: 3.51991498, Time: 0.0208 Steps: 10880, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000245, Sample Num: 3920, Cur Loss: 0.69139379, Cur Avg Loss: 0.68388772, Log Avg loss: 0.66569865, Global Avg Loss: 3.51729403, Time: 0.0207 Steps: 10890, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000255, Sample Num: 4080, Cur Loss: 0.70504391, Cur Avg Loss: 0.68582258, Log Avg loss: 0.73322681, Global Avg Loss: 3.51473984, Time: 0.0208 Steps: 10900, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000265, Sample Num: 4240, Cur Loss: 0.46748650, Cur Avg Loss: 0.68633211, Log Avg loss: 0.69932511, Global Avg Loss: 3.51215926, Time: 0.0208 Steps: 10910, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000275, Sample Num: 4400, Cur Loss: 0.30844581, Cur Avg Loss: 0.67954304, Log Avg loss: 0.49963272, Global Avg Loss: 3.50940053, Time: 0.0207 Steps: 10920, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000285, Sample Num: 4560, Cur Loss: 0.76840258, Cur Avg Loss: 0.68828410, Log Avg loss: 0.92866325, Global Avg Loss: 3.50703938, Time: 0.0208 Steps: 10930, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000295, Sample Num: 4720, Cur Loss: 0.99834853, Cur Avg Loss: 0.68801783, Log Avg loss: 0.68042895, Global Avg Loss: 3.50445564, Time: 0.0207 Steps: 10940, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000305, Sample Num: 4880, Cur Loss: 0.45015639, Cur Avg Loss: 0.68584268, Log Avg loss: 0.62167597, Global Avg Loss: 3.50182297, Time: 0.0207 Steps: 10950, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000315, Sample Num: 5040, Cur Loss: 0.80979526, Cur Avg Loss: 0.68317657, Log Avg loss: 0.60186007, Global Avg Loss: 3.49917702, Time: 0.0207 Steps: 10960, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000325, Sample Num: 5200, Cur Loss: 0.56654203, Cur Avg Loss: 0.69352012, Log Avg loss: 1.01934184, Global Avg Loss: 3.49691646, Time: 0.0208 Steps: 10970, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000335, Sample Num: 5360, Cur Loss: 0.58368176, Cur Avg Loss: 0.69261458, Log Avg loss: 0.66318474, Global Avg Loss: 3.49433564, Time: 0.0208 Steps: 10980, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000345, Sample Num: 5520, Cur Loss: 0.66217500, Cur Avg Loss: 0.69220897, Log Avg loss: 0.67862093, Global Avg Loss: 3.49177357, Time: 0.0207 Steps: 10990, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000355, Sample Num: 5680, Cur Loss: 0.60153657, Cur Avg Loss: 0.69244168, Log Avg loss: 0.70047026, Global Avg Loss: 3.48923603, Time: 0.0208 Steps: 11000, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000365, Sample Num: 5840, Cur Loss: 0.31309482, Cur Avg Loss: 0.69505937, Log Avg loss: 0.78798746, Global Avg Loss: 3.48678258, Time: 0.0208 Steps: 11010, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000375, Sample Num: 6000, Cur Loss: 0.58854896, Cur Avg Loss: 0.69493841, Log Avg loss: 0.69052325, Global Avg Loss: 3.48424513, Time: 0.0208 Steps: 11020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000385, Sample Num: 6160, Cur Loss: 0.57501608, Cur Avg Loss: 0.69196761, Log Avg loss: 0.58056246, Global Avg Loss: 3.48161260, Time: 0.0208 Steps: 11030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000395, Sample Num: 6320, Cur Loss: 0.98241556, Cur Avg Loss: 0.68776306, Log Avg loss: 0.52588797, Global Avg Loss: 3.47893532, Time: 0.0208 Steps: 11040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000405, Sample Num: 6480, Cur Loss: 0.84873700, Cur Avg Loss: 0.68345698, Log Avg loss: 0.51336679, Global Avg Loss: 3.47625154, Time: 0.0207 Steps: 11050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000415, Sample Num: 6640, Cur Loss: 0.69392705, Cur Avg Loss: 0.68182328, Log Avg loss: 0.61565861, Global Avg Loss: 3.47366511, Time: 0.0208 Steps: 11060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000425, Sample Num: 6800, Cur Loss: 0.38934225, Cur Avg Loss: 0.68441157, Log Avg loss: 0.79182534, Global Avg Loss: 3.47124249, Time: 0.0208 Steps: 11070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000435, Sample Num: 6960, Cur Loss: 1.11799276, Cur Avg Loss: 0.68825805, Log Avg loss: 0.85173363, Global Avg Loss: 3.46887832, Time: 0.0207 Steps: 11080, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000445, Sample Num: 7120, Cur Loss: 0.77295899, Cur Avg Loss: 0.69289123, Log Avg loss: 0.89443476, Global Avg Loss: 3.46655691, Time: 0.0208 Steps: 11090, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000455, Sample Num: 7280, Cur Loss: 1.43840218, Cur Avg Loss: 0.69832405, Log Avg loss: 0.94008440, Global Avg Loss: 3.46428080, Time: 0.0207 Steps: 11100, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000465, Sample Num: 7440, Cur Loss: 0.84452677, Cur Avg Loss: 0.69973284, Log Avg loss: 0.76383291, Global Avg Loss: 3.46185016, Time: 0.0208 Steps: 11110, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000475, Sample Num: 7600, Cur Loss: 0.85359949, Cur Avg Loss: 0.69898040, Log Avg loss: 0.66399178, Global Avg Loss: 3.45933410, Time: 0.0207 Steps: 11120, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000485, Sample Num: 7760, Cur Loss: 0.35100052, Cur Avg Loss: 0.69485929, Log Avg loss: 0.49910657, Global Avg Loss: 3.45667441, Time: 0.0208 Steps: 11130, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000495, Sample Num: 7920, Cur Loss: 0.60678643, Cur Avg Loss: 0.69775354, Log Avg loss: 0.83812442, Global Avg Loss: 3.45432383, Time: 0.0208 Steps: 11140, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000505, Sample Num: 8080, Cur Loss: 1.05112541, Cur Avg Loss: 0.69608333, Log Avg loss: 0.61340791, Global Avg Loss: 3.45177592, Time: 0.0208 Steps: 11150, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000515, Sample Num: 8240, Cur Loss: 1.03872955, Cur Avg Loss: 0.69481361, Log Avg loss: 0.63069301, Global Avg Loss: 3.44924807, Time: 0.0245 Steps: 11160, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000525, Sample Num: 8400, Cur Loss: 1.28272426, Cur Avg Loss: 0.69798903, Log Avg loss: 0.86152305, Global Avg Loss: 3.44693140, Time: 0.0208 Steps: 11170, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000535, Sample Num: 8560, Cur Loss: 0.49574128, Cur Avg Loss: 0.69608213, Log Avg loss: 0.59597012, Global Avg Loss: 3.44438134, Time: 0.0207 Steps: 11180, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000545, Sample Num: 8720, Cur Loss: 0.50691628, Cur Avg Loss: 0.69449469, Log Avg loss: 0.60956665, Global Avg Loss: 3.44184800, Time: 0.0208 Steps: 11190, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000555, Sample Num: 8880, Cur Loss: 0.21556111, Cur Avg Loss: 0.69025795, Log Avg loss: 0.45935521, Global Avg Loss: 3.43918506, Time: 0.0207 Steps: 11200, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000565, Sample Num: 9040, Cur Loss: 0.35262877, Cur Avg Loss: 0.68743393, Log Avg loss: 0.53070083, Global Avg Loss: 3.43659051, Time: 0.0207 Steps: 11210, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000575, Sample Num: 9200, Cur Loss: 0.19481005, Cur Avg Loss: 0.68332528, Log Avg loss: 0.45118678, Global Avg Loss: 3.43392973, Time: 0.0209 Steps: 11220, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000585, Sample Num: 9360, Cur Loss: 1.07416499, Cur Avg Loss: 0.68630884, Log Avg loss: 0.85786352, Global Avg Loss: 3.43163581, Time: 0.0207 Steps: 11230, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000595, Sample Num: 9520, Cur Loss: 0.99532014, Cur Avg Loss: 0.68468083, Log Avg loss: 0.58944207, Global Avg Loss: 3.42910717, Time: 0.0208 Steps: 11240, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000605, Sample Num: 9680, Cur Loss: 0.17088844, Cur Avg Loss: 0.68326311, Log Avg loss: 0.59890912, Global Avg Loss: 3.42659144, Time: 0.0207 Steps: 11250, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000615, Sample Num: 9840, Cur Loss: 0.57797348, Cur Avg Loss: 0.68086780, Log Avg loss: 0.53595148, Global Avg Loss: 3.42402426, Time: 0.0208 Steps: 11260, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000625, Sample Num: 10000, Cur Loss: 0.63462448, Cur Avg Loss: 0.68518008, Log Avg loss: 0.95038530, Global Avg Loss: 3.42182937, Time: 0.0208 Steps: 11270, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000635, Sample Num: 10160, Cur Loss: 0.50153482, Cur Avg Loss: 0.68486387, Log Avg loss: 0.66510050, Global Avg Loss: 3.41938546, Time: 0.0208 Steps: 11280, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000645, Sample Num: 10320, Cur Loss: 1.94196439, Cur Avg Loss: 0.68803559, Log Avg loss: 0.88944004, Global Avg Loss: 3.41714459, Time: 0.0207 Steps: 11290, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000655, Sample Num: 10480, Cur Loss: 0.58221900, Cur Avg Loss: 0.68669206, Log Avg loss: 0.60003432, Global Avg Loss: 3.41465157, Time: 0.0207 Steps: 11300, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000665, Sample Num: 10640, Cur Loss: 0.63504058, Cur Avg Loss: 0.68639073, Log Avg loss: 0.66665336, Global Avg Loss: 3.41222187, Time: 0.0207 Steps: 11310, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000675, Sample Num: 10800, Cur Loss: 1.08859205, Cur Avg Loss: 0.68689637, Log Avg loss: 0.72052140, Global Avg Loss: 3.40984404, Time: 0.0207 Steps: 11320, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000685, Sample Num: 10960, Cur Loss: 1.19587815, Cur Avg Loss: 0.69086197, Log Avg loss: 0.95854031, Global Avg Loss: 3.40768049, Time: 0.0208 Steps: 11330, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000695, Sample Num: 11120, Cur Loss: 0.44764853, Cur Avg Loss: 0.69294834, Log Avg loss: 0.83586479, Global Avg Loss: 3.40541257, Time: 0.0207 Steps: 11340, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000705, Sample Num: 11280, Cur Loss: 0.41316119, Cur Avg Loss: 0.69320653, Log Avg loss: 0.71115080, Global Avg Loss: 3.40303877, Time: 0.0208 Steps: 11350, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000715, Sample Num: 11440, Cur Loss: 1.33158910, Cur Avg Loss: 0.69439781, Log Avg loss: 0.77838248, Global Avg Loss: 3.40072834, Time: 0.0208 Steps: 11360, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000725, Sample Num: 11600, Cur Loss: 0.58059841, Cur Avg Loss: 0.69325005, Log Avg loss: 0.61118537, Global Avg Loss: 3.39827491, Time: 0.0208 Steps: 11370, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000735, Sample Num: 11760, Cur Loss: 0.83427763, Cur Avg Loss: 0.69176608, Log Avg loss: 0.58417849, Global Avg Loss: 3.39580207, Time: 0.0207 Steps: 11380, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000745, Sample Num: 11920, Cur Loss: 0.80898005, Cur Avg Loss: 0.69037656, Log Avg loss: 0.58824645, Global Avg Loss: 3.39333714, Time: 0.0207 Steps: 11390, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000755, Sample Num: 12080, Cur Loss: 0.18318638, Cur Avg Loss: 0.68778006, Log Avg loss: 0.49434114, Global Avg Loss: 3.39079416, Time: 0.0208 Steps: 11400, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000765, Sample Num: 12240, Cur Loss: 0.41862375, Cur Avg Loss: 0.68612302, Log Avg loss: 0.56101629, Global Avg Loss: 3.38831407, Time: 0.0207 Steps: 11410, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000775, Sample Num: 12400, Cur Loss: 0.40181226, Cur Avg Loss: 0.68389864, Log Avg loss: 0.51373383, Global Avg Loss: 3.38579693, Time: 0.0207 Steps: 11420, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000785, Sample Num: 12560, Cur Loss: 0.24932992, Cur Avg Loss: 0.68250578, Log Avg loss: 0.57455912, Global Avg Loss: 3.38333740, Time: 0.0207 Steps: 11430, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000795, Sample Num: 12720, Cur Loss: 0.51178223, Cur Avg Loss: 0.68122757, Log Avg loss: 0.58088783, Global Avg Loss: 3.38088771, Time: 0.0207 Steps: 11440, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000805, Sample Num: 12880, Cur Loss: 0.61457455, Cur Avg Loss: 0.68140981, Log Avg loss: 0.69589818, Global Avg Loss: 3.37854274, Time: 0.0207 Steps: 11450, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000815, Sample Num: 13040, Cur Loss: 0.60552108, Cur Avg Loss: 0.68278156, Log Avg loss: 0.79320696, Global Avg Loss: 3.37628678, Time: 0.0207 Steps: 11460, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000825, Sample Num: 13200, Cur Loss: 1.03381157, Cur Avg Loss: 0.68418710, Log Avg loss: 0.79873870, Global Avg Loss: 3.37403957, Time: 0.0208 Steps: 11470, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000835, Sample Num: 13360, Cur Loss: 1.46273458, Cur Avg Loss: 0.68433097, Log Avg loss: 0.69620026, Global Avg Loss: 3.37170695, Time: 0.0208 Steps: 11480, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000845, Sample Num: 13520, Cur Loss: 0.37137389, Cur Avg Loss: 0.68372122, Log Avg loss: 0.63280779, Global Avg Loss: 3.36932323, Time: 0.0208 Steps: 11490, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000855, Sample Num: 13680, Cur Loss: 0.48585227, Cur Avg Loss: 0.68371074, Log Avg loss: 0.68282465, Global Avg Loss: 3.36698714, Time: 0.0208 Steps: 11500, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000865, Sample Num: 13840, Cur Loss: 0.65092421, Cur Avg Loss: 0.68320428, Log Avg loss: 0.63990228, Global Avg Loss: 3.36461783, Time: 0.0208 Steps: 11510, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000875, Sample Num: 14000, Cur Loss: 0.78429013, Cur Avg Loss: 0.68203498, Log Avg loss: 0.58089013, Global Avg Loss: 3.36220140, Time: 0.0208 Steps: 11520, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000885, Sample Num: 14160, Cur Loss: 0.28626645, Cur Avg Loss: 0.68060098, Log Avg loss: 0.55512639, Global Avg Loss: 3.35976681, Time: 0.0208 Steps: 11530, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000895, Sample Num: 14320, Cur Loss: 0.64547491, Cur Avg Loss: 0.67923077, Log Avg loss: 0.55796660, Global Avg Loss: 3.35733891, Time: 0.0208 Steps: 11540, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000905, Sample Num: 14480, Cur Loss: 0.39653173, Cur Avg Loss: 0.67844579, Log Avg loss: 0.60819007, Global Avg Loss: 3.35495869, Time: 0.0208 Steps: 11550, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000915, Sample Num: 14640, Cur Loss: 1.37000632, Cur Avg Loss: 0.67878742, Log Avg loss: 0.70970492, Global Avg Loss: 3.35267041, Time: 0.0208 Steps: 11560, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000925, Sample Num: 14800, Cur Loss: 1.22863913, Cur Avg Loss: 0.67943710, Log Avg loss: 0.73888314, Global Avg Loss: 3.35041130, Time: 0.0208 Steps: 11570, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000935, Sample Num: 14960, Cur Loss: 1.06729245, Cur Avg Loss: 0.67798041, Log Avg loss: 0.54323710, Global Avg Loss: 3.34798715, Time: 0.0208 Steps: 11580, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000945, Sample Num: 15120, Cur Loss: 0.34227833, Cur Avg Loss: 0.67773714, Log Avg loss: 0.65499054, Global Avg Loss: 3.34566360, Time: 0.0208 Steps: 11590, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000955, Sample Num: 15280, Cur Loss: 1.56482351, Cur Avg Loss: 0.67832374, Log Avg loss: 0.73375805, Global Avg Loss: 3.34341195, Time: 0.0207 Steps: 11600, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000965, Sample Num: 15440, Cur Loss: 0.57902443, Cur Avg Loss: 0.67846302, Log Avg loss: 0.69176371, Global Avg Loss: 3.34112802, Time: 0.0208 Steps: 11610, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000975, Sample Num: 15600, Cur Loss: 0.76474410, Cur Avg Loss: 0.68142423, Log Avg loss: 0.96718168, Global Avg Loss: 3.33908504, Time: 0.0208 Steps: 11620, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000985, Sample Num: 15760, Cur Loss: 1.45885038, Cur Avg Loss: 0.68767752, Log Avg loss: 1.29737326, Global Avg Loss: 3.33732948, Time: 0.0208 Steps: 11630, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000995, Sample Num: 15920, Cur Loss: 0.59461516, Cur Avg Loss: 0.68817719, Log Avg loss: 0.73739428, Global Avg Loss: 3.33509586, Time: 0.0208 Steps: 11640, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001005, Sample Num: 16080, Cur Loss: 0.68741918, Cur Avg Loss: 0.68833744, Log Avg loss: 0.70428220, Global Avg Loss: 3.33283765, Time: 0.0208 Steps: 11650, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001015, Sample Num: 16240, Cur Loss: 0.64764357, Cur Avg Loss: 0.68916350, Log Avg loss: 0.77218312, Global Avg Loss: 3.33064155, Time: 0.0208 Steps: 11660, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001025, Sample Num: 16400, Cur Loss: 0.54946536, Cur Avg Loss: 0.68860454, Log Avg loss: 0.63187011, Global Avg Loss: 3.32832897, Time: 0.0244 Steps: 11670, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001035, Sample Num: 16560, Cur Loss: 0.59013879, Cur Avg Loss: 0.68817319, Log Avg loss: 0.64395981, Global Avg Loss: 3.32603071, Time: 0.0208 Steps: 11680, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001045, Sample Num: 16720, Cur Loss: 0.75716460, Cur Avg Loss: 0.69184254, Log Avg loss: 1.07161952, Global Avg Loss: 3.32410222, Time: 0.0208 Steps: 11690, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001055, Sample Num: 16880, Cur Loss: 0.86802530, Cur Avg Loss: 0.69042275, Log Avg loss: 0.54205514, Global Avg Loss: 3.32172440, Time: 0.0208 Steps: 11700, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001065, Sample Num: 17040, Cur Loss: 0.68728715, Cur Avg Loss: 0.68986150, Log Avg loss: 0.63064955, Global Avg Loss: 3.31942630, Time: 0.0208 Steps: 11710, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001075, Sample Num: 17200, Cur Loss: 1.00419188, Cur Avg Loss: 0.69192122, Log Avg loss: 0.91128162, Global Avg Loss: 3.31737157, Time: 0.0209 Steps: 11720, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001085, Sample Num: 17360, Cur Loss: 1.10946441, Cur Avg Loss: 0.69141958, Log Avg loss: 0.63749344, Global Avg Loss: 3.31508693, Time: 0.0208 Steps: 11730, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001095, Sample Num: 17520, Cur Loss: 0.78761846, Cur Avg Loss: 0.69073123, Log Avg loss: 0.61604484, Global Avg Loss: 3.31278792, Time: 0.0208 Steps: 11740, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001105, Sample Num: 17680, Cur Loss: 0.64070570, Cur Avg Loss: 0.68850121, Log Avg loss: 0.44431408, Global Avg Loss: 3.31034667, Time: 0.0209 Steps: 11750, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001115, Sample Num: 17840, Cur Loss: 0.31123424, Cur Avg Loss: 0.68776932, Log Avg loss: 0.60689498, Global Avg Loss: 3.30804781, Time: 0.0209 Steps: 11760, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001125, Sample Num: 18000, Cur Loss: 1.10153270, Cur Avg Loss: 0.68751670, Log Avg loss: 0.65934949, Global Avg Loss: 3.30579743, Time: 0.0209 Steps: 11770, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001135, Sample Num: 18160, Cur Loss: 0.23516621, Cur Avg Loss: 0.68641124, Log Avg loss: 0.56204762, Global Avg Loss: 3.30346827, Time: 0.0209 Steps: 11780, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001145, Sample Num: 18320, Cur Loss: 0.41711742, Cur Avg Loss: 0.68734578, Log Avg loss: 0.79341585, Global Avg Loss: 3.30133930, Time: 0.0209 Steps: 11790, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001155, Sample Num: 18480, Cur Loss: 1.05610991, Cur Avg Loss: 0.68733275, Log Avg loss: 0.68584141, Global Avg Loss: 3.29912278, Time: 0.0209 Steps: 11800, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001165, Sample Num: 18640, Cur Loss: 1.05272722, Cur Avg Loss: 0.68786761, Log Avg loss: 0.74964381, Global Avg Loss: 3.29696403, Time: 0.0209 Steps: 11810, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001175, Sample Num: 18800, Cur Loss: 0.35817009, Cur Avg Loss: 0.68727291, Log Avg loss: 0.61799052, Global Avg Loss: 3.29469756, Time: 0.0209 Steps: 11820, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001185, Sample Num: 18960, Cur Loss: 0.93678188, Cur Avg Loss: 0.68774975, Log Avg loss: 0.74377753, Global Avg Loss: 3.29254125, Time: 0.0208 Steps: 11830, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001195, Sample Num: 19120, Cur Loss: 0.81137633, Cur Avg Loss: 0.68922711, Log Avg loss: 0.86429454, Global Avg Loss: 3.29049036, Time: 0.0208 Steps: 11840, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001205, Sample Num: 19280, Cur Loss: 0.78951418, Cur Avg Loss: 0.69073815, Log Avg loss: 0.87130786, Global Avg Loss: 3.28844886, Time: 0.0209 Steps: 11850, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001215, Sample Num: 19440, Cur Loss: 0.48992354, Cur Avg Loss: 0.69007728, Log Avg loss: 0.61044164, Global Avg Loss: 3.28619084, Time: 0.0208 Steps: 11860, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001225, Sample Num: 19600, Cur Loss: 0.14994572, Cur Avg Loss: 0.68951282, Log Avg loss: 0.62093183, Global Avg Loss: 3.28394547, Time: 0.0209 Steps: 11870, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001235, Sample Num: 19760, Cur Loss: 0.35098994, Cur Avg Loss: 0.68837691, Log Avg loss: 0.54922697, Global Avg Loss: 3.28164351, Time: 0.0208 Steps: 11880, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001245, Sample Num: 19920, Cur Loss: 0.47627723, Cur Avg Loss: 0.69015315, Log Avg loss: 0.90951882, Global Avg Loss: 3.27964846, Time: 0.0208 Steps: 11890, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001255, Sample Num: 20080, Cur Loss: 0.80304182, Cur Avg Loss: 0.69037271, Log Avg loss: 0.71770792, Global Avg Loss: 3.27749557, Time: 0.0208 Steps: 11900, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001265, Sample Num: 20240, Cur Loss: 0.59338176, Cur Avg Loss: 0.68931414, Log Avg loss: 0.55646434, Global Avg Loss: 3.27521090, Time: 0.0208 Steps: 11910, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001275, Sample Num: 20400, Cur Loss: 1.13089466, Cur Avg Loss: 0.69108044, Log Avg loss: 0.91451678, Global Avg Loss: 3.27323046, Time: 0.0209 Steps: 11920, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001285, Sample Num: 20560, Cur Loss: 0.63203764, Cur Avg Loss: 0.69159392, Log Avg loss: 0.75706329, Global Avg Loss: 3.27112135, Time: 0.0238 Steps: 11930, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001295, Sample Num: 20720, Cur Loss: 0.85638744, Cur Avg Loss: 0.69142402, Log Avg loss: 0.66959180, Global Avg Loss: 3.26894251, Time: 0.0209 Steps: 11940, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001305, Sample Num: 20880, Cur Loss: 0.87871796, Cur Avg Loss: 0.69300904, Log Avg loss: 0.89826854, Global Avg Loss: 3.26695868, Time: 0.0209 Steps: 11950, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001315, Sample Num: 21040, Cur Loss: 0.64991510, Cur Avg Loss: 0.69263260, Log Avg loss: 0.64350750, Global Avg Loss: 3.26476516, Time: 0.0209 Steps: 11960, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001325, Sample Num: 21200, Cur Loss: 0.34481555, Cur Avg Loss: 0.69183381, Log Avg loss: 0.58679277, Global Avg Loss: 3.26252793, Time: 0.0209 Steps: 11970, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001335, Sample Num: 21360, Cur Loss: 0.27468884, Cur Avg Loss: 0.69206059, Log Avg loss: 0.72210859, Global Avg Loss: 3.26040738, Time: 0.0208 Steps: 11980, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001345, Sample Num: 21520, Cur Loss: 1.58046949, Cur Avg Loss: 0.69591808, Log Avg loss: 1.21089391, Global Avg Loss: 3.25869802, Time: 0.0208 Steps: 11990, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001355, Sample Num: 21680, Cur Loss: 1.00351882, Cur Avg Loss: 0.69723080, Log Avg loss: 0.87379121, Global Avg Loss: 3.25671060, Time: 0.0209 Steps: 12000, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001365, Sample Num: 21840, Cur Loss: 0.65523988, Cur Avg Loss: 0.69838519, Log Avg loss: 0.85480559, Global Avg Loss: 3.25471068, Time: 0.0209 Steps: 12010, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001375, Sample Num: 22000, Cur Loss: 0.30071676, Cur Avg Loss: 0.69773486, Log Avg loss: 0.60896448, Global Avg Loss: 3.25250956, Time: 0.0209 Steps: 12020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001385, Sample Num: 22160, Cur Loss: 1.00326025, Cur Avg Loss: 0.69768545, Log Avg loss: 0.69089107, Global Avg Loss: 3.25038020, Time: 0.0208 Steps: 12030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001395, Sample Num: 22320, Cur Loss: 0.27087754, Cur Avg Loss: 0.69864270, Log Avg loss: 0.83122206, Global Avg Loss: 3.24837093, Time: 0.0208 Steps: 12040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001405, Sample Num: 22480, Cur Loss: 0.42801273, Cur Avg Loss: 0.69679310, Log Avg loss: 0.43877374, Global Avg Loss: 3.24603932, Time: 0.0208 Steps: 12050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001415, Sample Num: 22640, Cur Loss: 0.48816156, Cur Avg Loss: 0.69617263, Log Avg loss: 0.60899739, Global Avg Loss: 3.24385272, Time: 0.0209 Steps: 12060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001425, Sample Num: 22800, Cur Loss: 0.37385735, Cur Avg Loss: 0.69589249, Log Avg loss: 0.65625156, Global Avg Loss: 3.24170889, Time: 0.0209 Steps: 12070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001435, Sample Num: 22960, Cur Loss: 0.28526586, Cur Avg Loss: 0.69595025, Log Avg loss: 0.70418129, Global Avg Loss: 3.23960829, Time: 0.0209 Steps: 12080, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001445, Sample Num: 23120, Cur Loss: 0.65876997, Cur Avg Loss: 0.69539540, Log Avg loss: 0.61577480, Global Avg Loss: 3.23743803, Time: 0.0208 Steps: 12090, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001455, Sample Num: 23280, Cur Loss: 1.93377805, Cur Avg Loss: 0.69734781, Log Avg loss: 0.97947119, Global Avg Loss: 3.23557195, Time: 0.0209 Steps: 12100, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001465, Sample Num: 23440, Cur Loss: 0.74762678, Cur Avg Loss: 0.69852600, Log Avg loss: 0.86995309, Global Avg Loss: 3.23361850, Time: 0.0208 Steps: 12110, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001475, Sample Num: 23600, Cur Loss: 0.19478028, Cur Avg Loss: 0.69802241, Log Avg loss: 0.62424524, Global Avg Loss: 3.23146556, Time: 0.0209 Steps: 12120, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001485, Sample Num: 23760, Cur Loss: 1.38478351, Cur Avg Loss: 0.69884504, Log Avg loss: 0.82018331, Global Avg Loss: 3.22947769, Time: 0.0209 Steps: 12130, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001495, Sample Num: 23920, Cur Loss: 0.72766858, Cur Avg Loss: 0.69948228, Log Avg loss: 0.79411243, Global Avg Loss: 3.22747162, Time: 0.0209 Steps: 12140, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001505, Sample Num: 24080, Cur Loss: 0.60062814, Cur Avg Loss: 0.70273058, Log Avg loss: 1.18835197, Global Avg Loss: 3.22579333, Time: 0.0208 Steps: 12150, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001515, Sample Num: 24240, Cur Loss: 0.45697239, Cur Avg Loss: 0.70374155, Log Avg loss: 0.85589193, Global Avg Loss: 3.22384440, Time: 0.0209 Steps: 12160, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001525, Sample Num: 24400, Cur Loss: 0.43428254, Cur Avg Loss: 0.70213995, Log Avg loss: 0.45949793, Global Avg Loss: 3.22157296, Time: 0.0209 Steps: 12170, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001535, Sample Num: 24560, Cur Loss: 0.62882733, Cur Avg Loss: 0.70068365, Log Avg loss: 0.47859827, Global Avg Loss: 3.21932093, Time: 0.0209 Steps: 12180, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001545, Sample Num: 24720, Cur Loss: 1.06247640, Cur Avg Loss: 0.70107776, Log Avg loss: 0.76157357, Global Avg Loss: 3.21730473, Time: 0.0219 Steps: 12190, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001555, Sample Num: 24880, Cur Loss: 0.50717258, Cur Avg Loss: 0.70142811, Log Avg loss: 0.75555750, Global Avg Loss: 3.21528690, Time: 0.0219 Steps: 12200, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001565, Sample Num: 25040, Cur Loss: 0.27968472, Cur Avg Loss: 0.70087157, Log Avg loss: 0.61432967, Global Avg Loss: 3.21315672, Time: 0.0219 Steps: 12210, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001575, Sample Num: 25200, Cur Loss: 0.71610194, Cur Avg Loss: 0.70095497, Log Avg loss: 0.71400709, Global Avg Loss: 3.21111158, Time: 0.0219 Steps: 12220, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001585, Sample Num: 25360, Cur Loss: 0.33216265, Cur Avg Loss: 0.70011975, Log Avg loss: 0.56857135, Global Avg Loss: 3.20895088, Time: 0.0219 Steps: 12230, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001595, Sample Num: 25520, Cur Loss: 0.43326306, Cur Avg Loss: 0.69944092, Log Avg loss: 0.59184618, Global Avg Loss: 3.20681272, Time: 0.0219 Steps: 12240, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001605, Sample Num: 25680, Cur Loss: 0.57647938, Cur Avg Loss: 0.69907413, Log Avg loss: 0.64057247, Global Avg Loss: 3.20471783, Time: 0.0219 Steps: 12250, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001615, Sample Num: 25840, Cur Loss: 1.33471251, Cur Avg Loss: 0.69872459, Log Avg loss: 0.64262207, Global Avg Loss: 3.20262803, Time: 0.0219 Steps: 12260, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001625, Sample Num: 26000, Cur Loss: 0.45287186, Cur Avg Loss: 0.69821572, Log Avg loss: 0.61603396, Global Avg Loss: 3.20051997, Time: 0.0220 Steps: 12270, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001635, Sample Num: 26160, Cur Loss: 1.46497273, Cur Avg Loss: 0.69844914, Log Avg loss: 0.73638019, Global Avg Loss: 3.19851334, Time: 0.0219 Steps: 12280, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001645, Sample Num: 26320, Cur Loss: 0.68022871, Cur Avg Loss: 0.69733897, Log Avg loss: 0.51582606, Global Avg Loss: 3.19633052, Time: 0.0219 Steps: 12290, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001655, Sample Num: 26480, Cur Loss: 0.94460779, Cur Avg Loss: 0.69709074, Log Avg loss: 0.65625680, Global Avg Loss: 3.19426542, Time: 0.0219 Steps: 12300, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001665, Sample Num: 26640, Cur Loss: 0.57219619, Cur Avg Loss: 0.69833952, Log Avg loss: 0.90501328, Global Avg Loss: 3.19240575, Time: 0.0220 Steps: 12310, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001675, Sample Num: 26800, Cur Loss: 1.21220398, Cur Avg Loss: 0.69843105, Log Avg loss: 0.71366937, Global Avg Loss: 3.19039379, Time: 0.0219 Steps: 12320, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001685, Sample Num: 26960, Cur Loss: 0.40169677, Cur Avg Loss: 0.69878690, Log Avg loss: 0.75839227, Global Avg Loss: 3.18842136, Time: 0.0219 Steps: 12330, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001695, Sample Num: 27120, Cur Loss: 0.32879680, Cur Avg Loss: 0.69811185, Log Avg loss: 0.58436643, Global Avg Loss: 3.18631111, Time: 0.0219 Steps: 12340, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001705, Sample Num: 27280, Cur Loss: 0.41570801, Cur Avg Loss: 0.69713290, Log Avg loss: 0.53120013, Global Avg Loss: 3.18416122, Time: 0.0219 Steps: 12350, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001715, Sample Num: 27440, Cur Loss: 0.44246060, Cur Avg Loss: 0.69682066, Log Avg loss: 0.64358352, Global Avg Loss: 3.18210574, Time: 0.0220 Steps: 12360, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001725, Sample Num: 27600, Cur Loss: 0.66349721, Cur Avg Loss: 0.69672019, Log Avg loss: 0.67949090, Global Avg Loss: 3.18008260, Time: 0.0219 Steps: 12370, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001735, Sample Num: 27760, Cur Loss: 1.02880144, Cur Avg Loss: 0.69677776, Log Avg loss: 0.70670864, Global Avg Loss: 3.17808473, Time: 0.0220 Steps: 12380, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001745, Sample Num: 27920, Cur Loss: 0.78553468, Cur Avg Loss: 0.69717971, Log Avg loss: 0.76691751, Global Avg Loss: 3.17613867, Time: 0.0219 Steps: 12390, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001755, Sample Num: 28080, Cur Loss: 0.28093395, Cur Avg Loss: 0.69662155, Log Avg loss: 0.59922180, Global Avg Loss: 3.17406051, Time: 0.0220 Steps: 12400, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001765, Sample Num: 28240, Cur Loss: 0.40848133, Cur Avg Loss: 0.69671755, Log Avg loss: 0.71356580, Global Avg Loss: 3.17207784, Time: 0.0219 Steps: 12410, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001775, Sample Num: 28400, Cur Loss: 0.41105902, Cur Avg Loss: 0.69562735, Log Avg loss: 0.50320657, Global Avg Loss: 3.16992899, Time: 0.0220 Steps: 12420, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001785, Sample Num: 28560, Cur Loss: 0.62291515, Cur Avg Loss: 0.69446760, Log Avg loss: 0.48861222, Global Avg Loss: 3.16777185, Time: 0.0219 Steps: 12430, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001795, Sample Num: 28720, Cur Loss: 0.78464949, Cur Avg Loss: 0.69386526, Log Avg loss: 0.58634830, Global Avg Loss: 3.16569675, Time: 0.0246 Steps: 12440, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001805, Sample Num: 28880, Cur Loss: 0.71913815, Cur Avg Loss: 0.69374960, Log Avg loss: 0.67298908, Global Avg Loss: 3.16369458, Time: 0.0209 Steps: 12450, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001815, Sample Num: 29040, Cur Loss: 0.52810609, Cur Avg Loss: 0.69322570, Log Avg loss: 0.59866117, Global Avg Loss: 3.16163597, Time: 0.0209 Steps: 12460, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001825, Sample Num: 29200, Cur Loss: 0.70162416, Cur Avg Loss: 0.69340238, Log Avg loss: 0.72546924, Global Avg Loss: 3.15968234, Time: 0.0209 Steps: 12470, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001835, Sample Num: 29360, Cur Loss: 0.40113592, Cur Avg Loss: 0.69419265, Log Avg loss: 0.83841836, Global Avg Loss: 3.15782236, Time: 0.0208 Steps: 12480, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001845, Sample Num: 29520, Cur Loss: 0.73802012, Cur Avg Loss: 0.69471561, Log Avg loss: 0.79067767, Global Avg Loss: 3.15592712, Time: 0.0209 Steps: 12490, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001855, Sample Num: 29680, Cur Loss: 0.99630558, Cur Avg Loss: 0.69449805, Log Avg loss: 0.65435911, Global Avg Loss: 3.15392587, Time: 0.0209 Steps: 12500, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001865, Sample Num: 29840, Cur Loss: 1.75578785, Cur Avg Loss: 0.69514683, Log Avg loss: 0.81549526, Global Avg Loss: 3.15205662, Time: 0.0209 Steps: 12510, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001875, Sample Num: 30000, Cur Loss: 0.64261055, Cur Avg Loss: 0.69544429, Log Avg loss: 0.75091962, Global Avg Loss: 3.15013878, Time: 0.0209 Steps: 12520, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001885, Sample Num: 30160, Cur Loss: 0.57686293, Cur Avg Loss: 0.69608927, Log Avg loss: 0.81702375, Global Avg Loss: 3.14827676, Time: 0.0209 Steps: 12530, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001895, Sample Num: 30320, Cur Loss: 1.34916735, Cur Avg Loss: 0.69663722, Log Avg loss: 0.79992485, Global Avg Loss: 3.14640407, Time: 0.0208 Steps: 12540, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001905, Sample Num: 30480, Cur Loss: 0.33931431, Cur Avg Loss: 0.69561517, Log Avg loss: 0.50193775, Global Avg Loss: 3.14429692, Time: 0.0209 Steps: 12550, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001915, Sample Num: 30640, Cur Loss: 0.34309173, Cur Avg Loss: 0.69559817, Log Avg loss: 0.69236019, Global Avg Loss: 3.14234474, Time: 0.0209 Steps: 12560, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001925, Sample Num: 30800, Cur Loss: 0.55858165, Cur Avg Loss: 0.69479967, Log Avg loss: 0.54188559, Global Avg Loss: 3.14027596, Time: 0.0209 Steps: 12570, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001935, Sample Num: 30960, Cur Loss: 0.79480743, Cur Avg Loss: 0.69504146, Log Avg loss: 0.74158692, Global Avg Loss: 3.13836921, Time: 0.0209 Steps: 12580, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001945, Sample Num: 31120, Cur Loss: 0.84935391, Cur Avg Loss: 0.69510780, Log Avg loss: 0.70794419, Global Avg Loss: 3.13643877, Time: 0.0209 Steps: 12590, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001955, Sample Num: 31280, Cur Loss: 0.92679048, Cur Avg Loss: 0.69430153, Log Avg loss: 0.53748116, Global Avg Loss: 3.13437611, Time: 0.0208 Steps: 12600, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001965, Sample Num: 31440, Cur Loss: 0.65594435, Cur Avg Loss: 0.69416984, Log Avg loss: 0.66842588, Global Avg Loss: 3.13242056, Time: 0.0209 Steps: 12610, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001975, Sample Num: 31600, Cur Loss: 0.53043777, Cur Avg Loss: 0.69474863, Log Avg loss: 0.80848096, Global Avg Loss: 3.13057908, Time: 0.0209 Steps: 12620, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001985, Sample Num: 31760, Cur Loss: 0.38977975, Cur Avg Loss: 0.69441371, Log Avg loss: 0.62826644, Global Avg Loss: 3.12859784, Time: 0.0209 Steps: 12630, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001995, Sample Num: 31920, Cur Loss: 0.52938861, Cur Avg Loss: 0.69456608, Log Avg loss: 0.72481129, Global Avg Loss: 3.12669611, Time: 0.0209 Steps: 12640, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002005, Sample Num: 32080, Cur Loss: 1.10648465, Cur Avg Loss: 0.69620428, Log Avg loss: 1.02302514, Global Avg Loss: 3.12503313, Time: 0.0209 Steps: 12650, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002015, Sample Num: 32240, Cur Loss: 0.34205228, Cur Avg Loss: 0.69564181, Log Avg loss: 0.58286708, Global Avg Loss: 3.12302510, Time: 0.0209 Steps: 12660, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002025, Sample Num: 32400, Cur Loss: 0.58993763, Cur Avg Loss: 0.69508456, Log Avg loss: 0.58279783, Global Avg Loss: 3.12102018, Time: 0.0208 Steps: 12670, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002035, Sample Num: 32560, Cur Loss: 0.42980936, Cur Avg Loss: 0.69478464, Log Avg loss: 0.63405229, Global Avg Loss: 3.11905885, Time: 0.0209 Steps: 12680, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002045, Sample Num: 32720, Cur Loss: 0.14122814, Cur Avg Loss: 0.69479699, Log Avg loss: 0.69730934, Global Avg Loss: 3.11715046, Time: 0.0209 Steps: 12690, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002055, Sample Num: 32880, Cur Loss: 0.45227462, Cur Avg Loss: 0.69381042, Log Avg loss: 0.49205735, Global Avg Loss: 3.11508346, Time: 0.0208 Steps: 12700, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002065, Sample Num: 33040, Cur Loss: 0.30540115, Cur Avg Loss: 0.69243636, Log Avg loss: 0.41006754, Global Avg Loss: 3.11295520, Time: 0.0207 Steps: 12710, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002075, Sample Num: 33200, Cur Loss: 0.98369527, Cur Avg Loss: 0.69283941, Log Avg loss: 0.77606759, Global Avg Loss: 3.11111802, Time: 0.0207 Steps: 12720, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002085, Sample Num: 33360, Cur Loss: 0.39143518, Cur Avg Loss: 0.69226173, Log Avg loss: 0.57239311, Global Avg Loss: 3.10912374, Time: 0.0207 Steps: 12730, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002095, Sample Num: 33520, Cur Loss: 0.69949234, Cur Avg Loss: 0.69186010, Log Avg loss: 0.60812049, Global Avg Loss: 3.10716063, Time: 0.0207 Steps: 12740, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002105, Sample Num: 33680, Cur Loss: 0.66576892, Cur Avg Loss: 0.69165622, Log Avg loss: 0.64894342, Global Avg Loss: 3.10523261, Time: 0.0207 Steps: 12750, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002115, Sample Num: 33840, Cur Loss: 1.12323523, Cur Avg Loss: 0.69134104, Log Avg loss: 0.62499645, Global Avg Loss: 3.10328885, Time: 0.0208 Steps: 12760, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002125, Sample Num: 34000, Cur Loss: 0.93605196, Cur Avg Loss: 0.69298966, Log Avg loss: 1.04167151, Global Avg Loss: 3.10167443, Time: 0.0208 Steps: 12770, Updated lr: 0.000089 ***** Running evaluation checkpoint-12774 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-12774 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.761395, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.645751, "eval_total_loss": 453.963099, "eval_mae": 0.543364, "eval_mse": 0.64601, "eval_r2": 0.589354, "eval_sp_statistic": 0.786644, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.806147, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.922642, "test_total_loss": 463.166199, "test_mae": 0.586679, "test_mse": 0.922977, "test_r2": 0.404303, "test_sp_statistic": 0.71116, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.764485, "test_ps_pvalue": 0.0, "lr": 8.883451872925558e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.100920158236722, "train_cur_epoch_loss": 1475.374621424824, "train_cur_epoch_avg_loss": 0.6929894886917914, "train_cur_epoch_time": 44.76139497756958, "train_cur_epoch_avg_time": 0.02102461013507261, "epoch": 6, "step": 12774} ################################################## Training, Epoch: 0007, Batch: 000006, Sample Num: 96, Cur Loss: 0.52083594, Cur Avg Loss: 0.67219706, Log Avg loss: 0.68047848, Global Avg Loss: 3.09977991, Time: 0.0216 Steps: 12780, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000016, Sample Num: 256, Cur Loss: 0.72730702, Cur Avg Loss: 0.63184377, Log Avg loss: 0.60763179, Global Avg Loss: 3.09783140, Time: 0.0208 Steps: 12790, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000026, Sample Num: 416, Cur Loss: 0.34573340, Cur Avg Loss: 0.71803607, Log Avg loss: 0.85594375, Global Avg Loss: 3.09607992, Time: 0.0208 Steps: 12800, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000036, Sample Num: 576, Cur Loss: 0.32511389, Cur Avg Loss: 0.63680926, Log Avg loss: 0.42561956, Global Avg Loss: 3.09399526, Time: 0.0207 Steps: 12810, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000046, Sample Num: 736, Cur Loss: 0.46024609, Cur Avg Loss: 0.61029270, Log Avg loss: 0.51483308, Global Avg Loss: 3.09198343, Time: 0.0208 Steps: 12820, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000056, Sample Num: 896, Cur Loss: 0.92117608, Cur Avg Loss: 0.62353731, Log Avg loss: 0.68446252, Global Avg Loss: 3.09010695, Time: 0.0207 Steps: 12830, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000066, Sample Num: 1056, Cur Loss: 1.12297082, Cur Avg Loss: 0.62026796, Log Avg loss: 0.60195962, Global Avg Loss: 3.08816914, Time: 0.0207 Steps: 12840, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000076, Sample Num: 1216, Cur Loss: 0.69633657, Cur Avg Loss: 0.64024436, Log Avg loss: 0.77208858, Global Avg Loss: 3.08636674, Time: 0.0208 Steps: 12850, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000086, Sample Num: 1376, Cur Loss: 0.71658897, Cur Avg Loss: 0.63275401, Log Avg loss: 0.57582737, Global Avg Loss: 3.08441454, Time: 0.0207 Steps: 12860, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000096, Sample Num: 1536, Cur Loss: 0.64514673, Cur Avg Loss: 0.61509716, Log Avg loss: 0.46324821, Global Avg Loss: 3.08237789, Time: 0.0207 Steps: 12870, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000106, Sample Num: 1696, Cur Loss: 0.61681426, Cur Avg Loss: 0.60482804, Log Avg loss: 0.50624449, Global Avg Loss: 3.08037779, Time: 0.0208 Steps: 12880, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000116, Sample Num: 1856, Cur Loss: 0.72654462, Cur Avg Loss: 0.63142426, Log Avg loss: 0.91334417, Global Avg Loss: 3.07869661, Time: 0.0207 Steps: 12890, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000126, Sample Num: 2016, Cur Loss: 0.83570009, Cur Avg Loss: 0.62456904, Log Avg loss: 0.54504856, Global Avg Loss: 3.07673254, Time: 0.0207 Steps: 12900, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000136, Sample Num: 2176, Cur Loss: 0.38711062, Cur Avg Loss: 0.63170021, Log Avg loss: 0.72155289, Global Avg Loss: 3.07490824, Time: 0.0207 Steps: 12910, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000146, Sample Num: 2336, Cur Loss: 1.23706222, Cur Avg Loss: 0.63949040, Log Avg loss: 0.74543703, Global Avg Loss: 3.07310524, Time: 0.0208 Steps: 12920, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000156, Sample Num: 2496, Cur Loss: 0.45547372, Cur Avg Loss: 0.62777720, Log Avg loss: 0.45676444, Global Avg Loss: 3.07108177, Time: 0.0207 Steps: 12930, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000166, Sample Num: 2656, Cur Loss: 0.37926710, Cur Avg Loss: 0.62821729, Log Avg loss: 0.63508280, Global Avg Loss: 3.06919924, Time: 0.0208 Steps: 12940, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000176, Sample Num: 2816, Cur Loss: 0.78928000, Cur Avg Loss: 0.62519910, Log Avg loss: 0.57509706, Global Avg Loss: 3.06727329, Time: 0.0207 Steps: 12950, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000186, Sample Num: 2976, Cur Loss: 0.31755739, Cur Avg Loss: 0.61556484, Log Avg loss: 0.44600194, Global Avg Loss: 3.06525071, Time: 0.0208 Steps: 12960, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000196, Sample Num: 3136, Cur Loss: 0.45001525, Cur Avg Loss: 0.61477504, Log Avg loss: 0.60008473, Global Avg Loss: 3.06335004, Time: 0.0208 Steps: 12970, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000206, Sample Num: 3296, Cur Loss: 0.81244302, Cur Avg Loss: 0.61106564, Log Avg loss: 0.53836145, Global Avg Loss: 3.06140475, Time: 0.0207 Steps: 12980, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000216, Sample Num: 3456, Cur Loss: 0.77399117, Cur Avg Loss: 0.60834979, Log Avg loss: 0.55240326, Global Avg Loss: 3.05947326, Time: 0.0208 Steps: 12990, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000226, Sample Num: 3616, Cur Loss: 0.42988425, Cur Avg Loss: 0.60507528, Log Avg loss: 0.53434575, Global Avg Loss: 3.05753085, Time: 0.0207 Steps: 13000, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000236, Sample Num: 3776, Cur Loss: 0.49325877, Cur Avg Loss: 0.60336742, Log Avg loss: 0.56476993, Global Avg Loss: 3.05561482, Time: 0.0207 Steps: 13010, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000246, Sample Num: 3936, Cur Loss: 1.71348596, Cur Avg Loss: 0.60251837, Log Avg loss: 0.58248080, Global Avg Loss: 3.05371533, Time: 0.0207 Steps: 13020, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000256, Sample Num: 4096, Cur Loss: 1.17275238, Cur Avg Loss: 0.60397794, Log Avg loss: 0.63988337, Global Avg Loss: 3.05186281, Time: 0.0253 Steps: 13030, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000266, Sample Num: 4256, Cur Loss: 0.19605571, Cur Avg Loss: 0.60771977, Log Avg loss: 0.70351050, Global Avg Loss: 3.05006193, Time: 0.0209 Steps: 13040, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000276, Sample Num: 4416, Cur Loss: 0.43357706, Cur Avg Loss: 0.60237478, Log Avg loss: 0.46019797, Global Avg Loss: 3.04807736, Time: 0.0208 Steps: 13050, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000286, Sample Num: 4576, Cur Loss: 0.79318762, Cur Avg Loss: 0.60347474, Log Avg loss: 0.63383377, Global Avg Loss: 3.04622878, Time: 0.0212 Steps: 13060, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000296, Sample Num: 4736, Cur Loss: 0.41200054, Cur Avg Loss: 0.60878144, Log Avg loss: 0.76055299, Global Avg Loss: 3.04447999, Time: 0.0209 Steps: 13070, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000306, Sample Num: 4896, Cur Loss: 0.28683409, Cur Avg Loss: 0.60590730, Log Avg loss: 0.52083285, Global Avg Loss: 3.04255059, Time: 0.0209 Steps: 13080, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000316, Sample Num: 5056, Cur Loss: 0.23896036, Cur Avg Loss: 0.60258470, Log Avg loss: 0.50091299, Global Avg Loss: 3.04060893, Time: 0.0209 Steps: 13090, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000326, Sample Num: 5216, Cur Loss: 0.30523106, Cur Avg Loss: 0.60121083, Log Avg loss: 0.55779662, Global Avg Loss: 3.03871365, Time: 0.0209 Steps: 13100, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000336, Sample Num: 5376, Cur Loss: 0.76973605, Cur Avg Loss: 0.60069230, Log Avg loss: 0.58378833, Global Avg Loss: 3.03684109, Time: 0.0209 Steps: 13110, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000346, Sample Num: 5536, Cur Loss: 0.69664079, Cur Avg Loss: 0.59984562, Log Avg loss: 0.57139691, Global Avg Loss: 3.03496194, Time: 0.0209 Steps: 13120, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000356, Sample Num: 5696, Cur Loss: 0.56152844, Cur Avg Loss: 0.59522061, Log Avg loss: 0.43519525, Global Avg Loss: 3.03298192, Time: 0.0209 Steps: 13130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000366, Sample Num: 5856, Cur Loss: 0.32317010, Cur Avg Loss: 0.59999431, Log Avg loss: 0.76993825, Global Avg Loss: 3.03125967, Time: 0.0209 Steps: 13140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000376, Sample Num: 6016, Cur Loss: 0.66890472, Cur Avg Loss: 0.59844212, Log Avg loss: 0.54163181, Global Avg Loss: 3.02936641, Time: 0.0209 Steps: 13150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000386, Sample Num: 6176, Cur Loss: 0.52555692, Cur Avg Loss: 0.59265122, Log Avg loss: 0.37491340, Global Avg Loss: 3.02734935, Time: 0.0209 Steps: 13160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000396, Sample Num: 6336, Cur Loss: 0.51285976, Cur Avg Loss: 0.59759059, Log Avg loss: 0.78825047, Global Avg Loss: 3.02564920, Time: 0.0209 Steps: 13170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000406, Sample Num: 6496, Cur Loss: 0.27278635, Cur Avg Loss: 0.59286605, Log Avg loss: 0.40577405, Global Avg Loss: 3.02366144, Time: 0.0209 Steps: 13180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000416, Sample Num: 6656, Cur Loss: 0.16569939, Cur Avg Loss: 0.59189314, Log Avg loss: 0.55239294, Global Avg Loss: 3.02178784, Time: 0.0209 Steps: 13190, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000426, Sample Num: 6816, Cur Loss: 0.67110723, Cur Avg Loss: 0.59693787, Log Avg loss: 0.80679885, Global Avg Loss: 3.02010982, Time: 0.0208 Steps: 13200, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000436, Sample Num: 6976, Cur Loss: 1.06983662, Cur Avg Loss: 0.60027753, Log Avg loss: 0.74254709, Global Avg Loss: 3.01838570, Time: 0.0209 Steps: 13210, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000446, Sample Num: 7136, Cur Loss: 1.18586326, Cur Avg Loss: 0.60136171, Log Avg loss: 0.64863177, Global Avg Loss: 3.01659315, Time: 0.0209 Steps: 13220, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000456, Sample Num: 7296, Cur Loss: 0.37283927, Cur Avg Loss: 0.60169732, Log Avg loss: 0.61666542, Global Avg Loss: 3.01477914, Time: 0.0209 Steps: 13230, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000466, Sample Num: 7456, Cur Loss: 0.63543308, Cur Avg Loss: 0.60614132, Log Avg loss: 0.80878809, Global Avg Loss: 3.01311299, Time: 0.0209 Steps: 13240, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000476, Sample Num: 7616, Cur Loss: 0.60747570, Cur Avg Loss: 0.60433996, Log Avg loss: 0.52039618, Global Avg Loss: 3.01123169, Time: 0.0209 Steps: 13250, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000486, Sample Num: 7776, Cur Loss: 0.75672227, Cur Avg Loss: 0.60507152, Log Avg loss: 0.63989414, Global Avg Loss: 3.00944335, Time: 0.0210 Steps: 13260, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000496, Sample Num: 7936, Cur Loss: 0.58805454, Cur Avg Loss: 0.61026751, Log Avg loss: 0.86279227, Global Avg Loss: 3.00782568, Time: 0.0209 Steps: 13270, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000506, Sample Num: 8096, Cur Loss: 0.75700390, Cur Avg Loss: 0.61552620, Log Avg loss: 0.87635765, Global Avg Loss: 3.00622066, Time: 0.0209 Steps: 13280, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000516, Sample Num: 8256, Cur Loss: 0.30190527, Cur Avg Loss: 0.61429861, Log Avg loss: 0.55218242, Global Avg Loss: 3.00437413, Time: 0.0218 Steps: 13290, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000526, Sample Num: 8416, Cur Loss: 0.26211143, Cur Avg Loss: 0.61095438, Log Avg loss: 0.43839214, Global Avg Loss: 3.00244482, Time: 0.0208 Steps: 13300, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000536, Sample Num: 8576, Cur Loss: 0.19566546, Cur Avg Loss: 0.61068346, Log Avg loss: 0.59643274, Global Avg Loss: 3.00063715, Time: 0.0208 Steps: 13310, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000546, Sample Num: 8736, Cur Loss: 1.01096761, Cur Avg Loss: 0.61053642, Log Avg loss: 0.60265520, Global Avg Loss: 2.99883686, Time: 0.0208 Steps: 13320, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000556, Sample Num: 8896, Cur Loss: 0.44249386, Cur Avg Loss: 0.60870143, Log Avg loss: 0.50851115, Global Avg Loss: 2.99696865, Time: 0.0208 Steps: 13330, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000566, Sample Num: 9056, Cur Loss: 1.20808303, Cur Avg Loss: 0.61052004, Log Avg loss: 0.71163461, Global Avg Loss: 2.99525551, Time: 0.0208 Steps: 13340, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000576, Sample Num: 9216, Cur Loss: 0.37870663, Cur Avg Loss: 0.61053560, Log Avg loss: 0.61141620, Global Avg Loss: 2.99346986, Time: 0.0208 Steps: 13350, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000586, Sample Num: 9376, Cur Loss: 0.59614396, Cur Avg Loss: 0.61245898, Log Avg loss: 0.72324593, Global Avg Loss: 2.99177059, Time: 0.0207 Steps: 13360, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000596, Sample Num: 9536, Cur Loss: 0.74875093, Cur Avg Loss: 0.61171026, Log Avg loss: 0.56783527, Global Avg Loss: 2.98995762, Time: 0.0208 Steps: 13370, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000606, Sample Num: 9696, Cur Loss: 0.34226483, Cur Avg Loss: 0.61186758, Log Avg loss: 0.62124356, Global Avg Loss: 2.98818728, Time: 0.0208 Steps: 13380, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000616, Sample Num: 9856, Cur Loss: 0.45714444, Cur Avg Loss: 0.61177154, Log Avg loss: 0.60595173, Global Avg Loss: 2.98640817, Time: 0.0209 Steps: 13390, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000626, Sample Num: 10016, Cur Loss: 0.60950363, Cur Avg Loss: 0.60961035, Log Avg loss: 0.47648127, Global Avg Loss: 2.98453509, Time: 0.0209 Steps: 13400, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000636, Sample Num: 10176, Cur Loss: 0.67728108, Cur Avg Loss: 0.60882226, Log Avg loss: 0.55948745, Global Avg Loss: 2.98272670, Time: 0.0209 Steps: 13410, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000646, Sample Num: 10336, Cur Loss: 0.48719794, Cur Avg Loss: 0.60998148, Log Avg loss: 0.68370805, Global Avg Loss: 2.98101357, Time: 0.0209 Steps: 13420, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000656, Sample Num: 10496, Cur Loss: 0.87399513, Cur Avg Loss: 0.60955999, Log Avg loss: 0.58233202, Global Avg Loss: 2.97922751, Time: 0.0209 Steps: 13430, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000666, Sample Num: 10656, Cur Loss: 0.59728765, Cur Avg Loss: 0.60606193, Log Avg loss: 0.37658856, Global Avg Loss: 2.97729102, Time: 0.0210 Steps: 13440, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000676, Sample Num: 10816, Cur Loss: 0.27244467, Cur Avg Loss: 0.60627526, Log Avg loss: 0.62048352, Global Avg Loss: 2.97553875, Time: 0.0209 Steps: 13450, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000686, Sample Num: 10976, Cur Loss: 0.77599263, Cur Avg Loss: 0.60791612, Log Avg loss: 0.71883830, Global Avg Loss: 2.97386215, Time: 0.0209 Steps: 13460, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000696, Sample Num: 11136, Cur Loss: 0.41497982, Cur Avg Loss: 0.60640613, Log Avg loss: 0.50282068, Global Avg Loss: 2.97202767, Time: 0.0210 Steps: 13470, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000706, Sample Num: 11296, Cur Loss: 0.30218256, Cur Avg Loss: 0.60528089, Log Avg loss: 0.52696399, Global Avg Loss: 2.97021383, Time: 0.0209 Steps: 13480, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000716, Sample Num: 11456, Cur Loss: 0.49852261, Cur Avg Loss: 0.60514220, Log Avg loss: 0.59535116, Global Avg Loss: 2.96845337, Time: 0.0209 Steps: 13490, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000726, Sample Num: 11616, Cur Loss: 0.97571599, Cur Avg Loss: 0.60472486, Log Avg loss: 0.57484292, Global Avg Loss: 2.96668032, Time: 0.0209 Steps: 13500, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000736, Sample Num: 11776, Cur Loss: 1.19230330, Cur Avg Loss: 0.60483339, Log Avg loss: 0.61271282, Global Avg Loss: 2.96493793, Time: 0.0209 Steps: 13510, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000746, Sample Num: 11936, Cur Loss: 1.05739391, Cur Avg Loss: 0.60399907, Log Avg loss: 0.54259331, Global Avg Loss: 2.96314626, Time: 0.0209 Steps: 13520, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000756, Sample Num: 12096, Cur Loss: 0.86295950, Cur Avg Loss: 0.60424943, Log Avg loss: 0.62292584, Global Avg Loss: 2.96141661, Time: 0.0209 Steps: 13530, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000766, Sample Num: 12256, Cur Loss: 0.45072746, Cur Avg Loss: 0.60362720, Log Avg loss: 0.55658658, Global Avg Loss: 2.95964051, Time: 0.0209 Steps: 13540, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000776, Sample Num: 12416, Cur Loss: 0.36180246, Cur Avg Loss: 0.60127121, Log Avg loss: 0.42080233, Global Avg Loss: 2.95776683, Time: 0.0209 Steps: 13550, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000786, Sample Num: 12576, Cur Loss: 0.30191258, Cur Avg Loss: 0.60122827, Log Avg loss: 0.59789629, Global Avg Loss: 2.95602651, Time: 0.0209 Steps: 13560, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000796, Sample Num: 12736, Cur Loss: 0.60224730, Cur Avg Loss: 0.60092554, Log Avg loss: 0.57713107, Global Avg Loss: 2.95427346, Time: 0.0209 Steps: 13570, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000806, Sample Num: 12896, Cur Loss: 0.45905769, Cur Avg Loss: 0.59869904, Log Avg loss: 0.42146995, Global Avg Loss: 2.95240836, Time: 0.0209 Steps: 13580, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000816, Sample Num: 13056, Cur Loss: 0.81273973, Cur Avg Loss: 0.59805989, Log Avg loss: 0.54654424, Global Avg Loss: 2.95063804, Time: 0.0209 Steps: 13590, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000826, Sample Num: 13216, Cur Loss: 0.55343622, Cur Avg Loss: 0.59945254, Log Avg loss: 0.71309266, Global Avg Loss: 2.94899279, Time: 0.0209 Steps: 13600, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000836, Sample Num: 13376, Cur Loss: 0.96476060, Cur Avg Loss: 0.60035053, Log Avg loss: 0.67452437, Global Avg Loss: 2.94732161, Time: 0.0209 Steps: 13610, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000846, Sample Num: 13536, Cur Loss: 0.45164150, Cur Avg Loss: 0.60559472, Log Avg loss: 1.04400939, Global Avg Loss: 2.94592417, Time: 0.0209 Steps: 13620, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000856, Sample Num: 13696, Cur Loss: 0.37850404, Cur Avg Loss: 0.60449593, Log Avg loss: 0.51153795, Global Avg Loss: 2.94413812, Time: 0.0209 Steps: 13630, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000866, Sample Num: 13856, Cur Loss: 0.33175594, Cur Avg Loss: 0.60326236, Log Avg loss: 0.49766848, Global Avg Loss: 2.94234452, Time: 0.0208 Steps: 13640, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000876, Sample Num: 14016, Cur Loss: 0.25176519, Cur Avg Loss: 0.60469373, Log Avg loss: 0.72865046, Global Avg Loss: 2.94072277, Time: 0.0209 Steps: 13650, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000886, Sample Num: 14176, Cur Loss: 0.48964459, Cur Avg Loss: 0.60551700, Log Avg loss: 0.67763581, Global Avg Loss: 2.93906604, Time: 0.0208 Steps: 13660, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000896, Sample Num: 14336, Cur Loss: 0.88508660, Cur Avg Loss: 0.60813117, Log Avg loss: 0.83974614, Global Avg Loss: 2.93753033, Time: 0.0209 Steps: 13670, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000906, Sample Num: 14496, Cur Loss: 0.50213635, Cur Avg Loss: 0.60874398, Log Avg loss: 0.66365214, Global Avg Loss: 2.93586814, Time: 0.0209 Steps: 13680, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000916, Sample Num: 14656, Cur Loss: 0.15894040, Cur Avg Loss: 0.60850408, Log Avg loss: 0.58676873, Global Avg Loss: 2.93415222, Time: 0.0209 Steps: 13690, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000926, Sample Num: 14816, Cur Loss: 0.41093349, Cur Avg Loss: 0.60705853, Log Avg loss: 0.47464627, Global Avg Loss: 2.93235696, Time: 0.0209 Steps: 13700, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000936, Sample Num: 14976, Cur Loss: 0.36772102, Cur Avg Loss: 0.60896392, Log Avg loss: 0.78540303, Global Avg Loss: 2.93079098, Time: 0.0209 Steps: 13710, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000946, Sample Num: 15136, Cur Loss: 0.53833127, Cur Avg Loss: 0.61078778, Log Avg loss: 0.78150119, Global Avg Loss: 2.92922444, Time: 0.0209 Steps: 13720, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000956, Sample Num: 15296, Cur Loss: 0.62429845, Cur Avg Loss: 0.61195605, Log Avg loss: 0.72247441, Global Avg Loss: 2.92761719, Time: 0.0209 Steps: 13730, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000966, Sample Num: 15456, Cur Loss: 0.42596373, Cur Avg Loss: 0.61293050, Log Avg loss: 0.70608838, Global Avg Loss: 2.92600036, Time: 0.0209 Steps: 13740, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000976, Sample Num: 15616, Cur Loss: 0.66410923, Cur Avg Loss: 0.61264085, Log Avg loss: 0.58466035, Global Avg Loss: 2.92429757, Time: 0.0209 Steps: 13750, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000986, Sample Num: 15776, Cur Loss: 0.35809052, Cur Avg Loss: 0.61509150, Log Avg loss: 0.85427526, Global Avg Loss: 2.92279319, Time: 0.0209 Steps: 13760, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000996, Sample Num: 15936, Cur Loss: 1.00086808, Cur Avg Loss: 0.61627294, Log Avg loss: 0.73276269, Global Avg Loss: 2.92120276, Time: 0.0209 Steps: 13770, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001006, Sample Num: 16096, Cur Loss: 0.95722473, Cur Avg Loss: 0.61786554, Log Avg loss: 0.77648798, Global Avg Loss: 2.91964636, Time: 0.0209 Steps: 13780, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001016, Sample Num: 16256, Cur Loss: 0.53264058, Cur Avg Loss: 0.61823804, Log Avg loss: 0.65571196, Global Avg Loss: 2.91800464, Time: 0.0209 Steps: 13790, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001026, Sample Num: 16416, Cur Loss: 0.83277720, Cur Avg Loss: 0.61853049, Log Avg loss: 0.64824381, Global Avg Loss: 2.91635988, Time: 0.0246 Steps: 13800, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001036, Sample Num: 16576, Cur Loss: 0.26001456, Cur Avg Loss: 0.62057605, Log Avg loss: 0.83045033, Global Avg Loss: 2.91484945, Time: 0.0208 Steps: 13810, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001046, Sample Num: 16736, Cur Loss: 0.30823272, Cur Avg Loss: 0.61956264, Log Avg loss: 0.51457305, Global Avg Loss: 2.91311264, Time: 0.0208 Steps: 13820, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001056, Sample Num: 16896, Cur Loss: 0.65143847, Cur Avg Loss: 0.62073059, Log Avg loss: 0.74289790, Global Avg Loss: 2.91154343, Time: 0.0208 Steps: 13830, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001066, Sample Num: 17056, Cur Loss: 0.26378405, Cur Avg Loss: 0.62048800, Log Avg loss: 0.59487063, Global Avg Loss: 2.90986953, Time: 0.0207 Steps: 13840, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001076, Sample Num: 17216, Cur Loss: 1.09637642, Cur Avg Loss: 0.61934254, Log Avg loss: 0.49723665, Global Avg Loss: 2.90812756, Time: 0.0208 Steps: 13850, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001086, Sample Num: 17376, Cur Loss: 0.24031296, Cur Avg Loss: 0.62014452, Log Avg loss: 0.70643755, Global Avg Loss: 2.90653904, Time: 0.0208 Steps: 13860, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001096, Sample Num: 17536, Cur Loss: 0.60917020, Cur Avg Loss: 0.61911834, Log Avg loss: 0.50767568, Global Avg Loss: 2.90480950, Time: 0.0208 Steps: 13870, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001106, Sample Num: 17696, Cur Loss: 0.33318368, Cur Avg Loss: 0.61783917, Log Avg loss: 0.47764221, Global Avg Loss: 2.90306082, Time: 0.0207 Steps: 13880, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001116, Sample Num: 17856, Cur Loss: 1.26786232, Cur Avg Loss: 0.61736808, Log Avg loss: 0.56526484, Global Avg Loss: 2.90137774, Time: 0.0208 Steps: 13890, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001126, Sample Num: 18016, Cur Loss: 0.69304430, Cur Avg Loss: 0.61768688, Log Avg loss: 0.65326523, Global Avg Loss: 2.89976040, Time: 0.0207 Steps: 13900, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001136, Sample Num: 18176, Cur Loss: 1.03389335, Cur Avg Loss: 0.61799159, Log Avg loss: 0.65230152, Global Avg Loss: 2.89814468, Time: 0.0208 Steps: 13910, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001146, Sample Num: 18336, Cur Loss: 0.26938915, Cur Avg Loss: 0.61612966, Log Avg loss: 0.40461443, Global Avg Loss: 2.89635335, Time: 0.0207 Steps: 13920, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001156, Sample Num: 18496, Cur Loss: 0.27551436, Cur Avg Loss: 0.61763215, Log Avg loss: 0.78981762, Global Avg Loss: 2.89484112, Time: 0.0208 Steps: 13930, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001166, Sample Num: 18656, Cur Loss: 0.42525938, Cur Avg Loss: 0.61690444, Log Avg loss: 0.53278172, Global Avg Loss: 2.89314668, Time: 0.0207 Steps: 13940, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001176, Sample Num: 18816, Cur Loss: 0.52562976, Cur Avg Loss: 0.61726982, Log Avg loss: 0.65987224, Global Avg Loss: 2.89154576, Time: 0.0208 Steps: 13950, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001186, Sample Num: 18976, Cur Loss: 0.52854234, Cur Avg Loss: 0.61705805, Log Avg loss: 0.59215409, Global Avg Loss: 2.88989864, Time: 0.0208 Steps: 13960, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001196, Sample Num: 19136, Cur Loss: 0.76100051, Cur Avg Loss: 0.61577636, Log Avg loss: 0.46376835, Global Avg Loss: 2.88816196, Time: 0.0208 Steps: 13970, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001206, Sample Num: 19296, Cur Loss: 0.91873884, Cur Avg Loss: 0.61622100, Log Avg loss: 0.66940025, Global Avg Loss: 2.88657487, Time: 0.0207 Steps: 13980, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001216, Sample Num: 19456, Cur Loss: 0.51268595, Cur Avg Loss: 0.61867350, Log Avg loss: 0.91444435, Global Avg Loss: 2.88516519, Time: 0.0208 Steps: 13990, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001226, Sample Num: 19616, Cur Loss: 0.62950492, Cur Avg Loss: 0.62069220, Log Avg loss: 0.86616650, Global Avg Loss: 2.88372305, Time: 0.0208 Steps: 14000, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001236, Sample Num: 19776, Cur Loss: 0.29972231, Cur Avg Loss: 0.61906923, Log Avg loss: 0.42009311, Global Avg Loss: 2.88196457, Time: 0.0208 Steps: 14010, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001246, Sample Num: 19936, Cur Loss: 0.77148116, Cur Avg Loss: 0.62024096, Log Avg loss: 0.76506625, Global Avg Loss: 2.88045466, Time: 0.0207 Steps: 14020, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001256, Sample Num: 20096, Cur Loss: 0.35984939, Cur Avg Loss: 0.62150272, Log Avg loss: 0.77871872, Global Avg Loss: 2.87895663, Time: 0.0208 Steps: 14030, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001266, Sample Num: 20256, Cur Loss: 0.48686630, Cur Avg Loss: 0.62203497, Log Avg loss: 0.68888554, Global Avg Loss: 2.87739675, Time: 0.0208 Steps: 14040, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001276, Sample Num: 20416, Cur Loss: 0.40249395, Cur Avg Loss: 0.62081071, Log Avg loss: 0.46581891, Global Avg Loss: 2.87568032, Time: 0.0208 Steps: 14050, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001286, Sample Num: 20576, Cur Loss: 0.36155814, Cur Avg Loss: 0.62062534, Log Avg loss: 0.59697228, Global Avg Loss: 2.87405962, Time: 0.0209 Steps: 14060, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001296, Sample Num: 20736, Cur Loss: 0.55958235, Cur Avg Loss: 0.62003418, Log Avg loss: 0.54401148, Global Avg Loss: 2.87240358, Time: 0.0208 Steps: 14070, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001306, Sample Num: 20896, Cur Loss: 0.25603873, Cur Avg Loss: 0.62095748, Log Avg loss: 0.74061694, Global Avg Loss: 2.87088953, Time: 0.0208 Steps: 14080, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001316, Sample Num: 21056, Cur Loss: 0.19535029, Cur Avg Loss: 0.61975891, Log Avg loss: 0.46322499, Global Avg Loss: 2.86918075, Time: 0.0208 Steps: 14090, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001326, Sample Num: 21216, Cur Loss: 0.74315584, Cur Avg Loss: 0.62030191, Log Avg loss: 0.69176183, Global Avg Loss: 2.86763649, Time: 0.0208 Steps: 14100, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001336, Sample Num: 21376, Cur Loss: 0.26289791, Cur Avg Loss: 0.61926799, Log Avg loss: 0.48216993, Global Avg Loss: 2.86594586, Time: 0.0208 Steps: 14110, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001346, Sample Num: 21536, Cur Loss: 0.29243481, Cur Avg Loss: 0.61898587, Log Avg loss: 0.58129382, Global Avg Loss: 2.86432784, Time: 0.0208 Steps: 14120, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001356, Sample Num: 21696, Cur Loss: 1.06938338, Cur Avg Loss: 0.61858800, Log Avg loss: 0.56503508, Global Avg Loss: 2.86270060, Time: 0.0208 Steps: 14130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001366, Sample Num: 21856, Cur Loss: 0.76536226, Cur Avg Loss: 0.61869850, Log Avg loss: 0.63368205, Global Avg Loss: 2.86112420, Time: 0.0208 Steps: 14140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001376, Sample Num: 22016, Cur Loss: 0.24994092, Cur Avg Loss: 0.61880761, Log Avg loss: 0.63371190, Global Avg Loss: 2.85955006, Time: 0.0208 Steps: 14150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001386, Sample Num: 22176, Cur Loss: 0.36449233, Cur Avg Loss: 0.61833649, Log Avg loss: 0.55351057, Global Avg Loss: 2.85792150, Time: 0.0208 Steps: 14160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001396, Sample Num: 22336, Cur Loss: 0.74246025, Cur Avg Loss: 0.61841604, Log Avg loss: 0.62944148, Global Avg Loss: 2.85634883, Time: 0.0208 Steps: 14170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001406, Sample Num: 22496, Cur Loss: 0.36843041, Cur Avg Loss: 0.62001735, Log Avg loss: 0.84356112, Global Avg Loss: 2.85492937, Time: 0.0208 Steps: 14180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001416, Sample Num: 22656, Cur Loss: 0.29296300, Cur Avg Loss: 0.62079381, Log Avg loss: 0.72996374, Global Avg Loss: 2.85343186, Time: 0.0208 Steps: 14190, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001426, Sample Num: 22816, Cur Loss: 0.45656055, Cur Avg Loss: 0.62330850, Log Avg loss: 0.97938906, Global Avg Loss: 2.85211211, Time: 0.0208 Steps: 14200, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001436, Sample Num: 22976, Cur Loss: 0.12460482, Cur Avg Loss: 0.62462225, Log Avg loss: 0.81196279, Global Avg Loss: 2.85067640, Time: 0.0208 Steps: 14210, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001446, Sample Num: 23136, Cur Loss: 0.55680358, Cur Avg Loss: 0.62407423, Log Avg loss: 0.54537769, Global Avg Loss: 2.84905523, Time: 0.0209 Steps: 14220, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001456, Sample Num: 23296, Cur Loss: 0.58381796, Cur Avg Loss: 0.62407704, Log Avg loss: 0.62448373, Global Avg Loss: 2.84749194, Time: 0.0208 Steps: 14230, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001466, Sample Num: 23456, Cur Loss: 0.58694112, Cur Avg Loss: 0.62387790, Log Avg loss: 0.59488335, Global Avg Loss: 2.84591005, Time: 0.0208 Steps: 14240, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001476, Sample Num: 23616, Cur Loss: 0.78872824, Cur Avg Loss: 0.62384999, Log Avg loss: 0.61975842, Global Avg Loss: 2.84434784, Time: 0.0208 Steps: 14250, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001486, Sample Num: 23776, Cur Loss: 0.27048424, Cur Avg Loss: 0.62291563, Log Avg loss: 0.48500414, Global Avg Loss: 2.84269332, Time: 0.0208 Steps: 14260, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001496, Sample Num: 23936, Cur Loss: 0.76280463, Cur Avg Loss: 0.62288028, Log Avg loss: 0.61762654, Global Avg Loss: 2.84113406, Time: 0.0208 Steps: 14270, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001506, Sample Num: 24096, Cur Loss: 0.70986480, Cur Avg Loss: 0.62175238, Log Avg loss: 0.45301927, Global Avg Loss: 2.83946171, Time: 0.0208 Steps: 14280, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001516, Sample Num: 24256, Cur Loss: 0.55744076, Cur Avg Loss: 0.62142240, Log Avg loss: 0.57172767, Global Avg Loss: 2.83787477, Time: 0.0208 Steps: 14290, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001526, Sample Num: 24416, Cur Loss: 0.80315536, Cur Avg Loss: 0.62136138, Log Avg loss: 0.61210992, Global Avg Loss: 2.83631829, Time: 0.0208 Steps: 14300, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001536, Sample Num: 24576, Cur Loss: 0.60338008, Cur Avg Loss: 0.62128245, Log Avg loss: 0.60923762, Global Avg Loss: 2.83476198, Time: 0.0253 Steps: 14310, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001546, Sample Num: 24736, Cur Loss: 0.71979022, Cur Avg Loss: 0.62009870, Log Avg loss: 0.43827542, Global Avg Loss: 2.83308846, Time: 0.0209 Steps: 14320, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001556, Sample Num: 24896, Cur Loss: 0.85391265, Cur Avg Loss: 0.62117435, Log Avg loss: 0.78746959, Global Avg Loss: 2.83166095, Time: 0.0209 Steps: 14330, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001566, Sample Num: 25056, Cur Loss: 0.93952572, Cur Avg Loss: 0.62015802, Log Avg loss: 0.46201674, Global Avg Loss: 2.83000848, Time: 0.0209 Steps: 14340, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001576, Sample Num: 25216, Cur Loss: 0.78508180, Cur Avg Loss: 0.61986471, Log Avg loss: 0.57393291, Global Avg Loss: 2.82843630, Time: 0.0209 Steps: 14350, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001586, Sample Num: 25376, Cur Loss: 0.60111225, Cur Avg Loss: 0.61991496, Log Avg loss: 0.62783492, Global Avg Loss: 2.82690385, Time: 0.0209 Steps: 14360, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001596, Sample Num: 25536, Cur Loss: 0.51994801, Cur Avg Loss: 0.62093250, Log Avg loss: 0.78231339, Global Avg Loss: 2.82548103, Time: 0.0209 Steps: 14370, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001606, Sample Num: 25696, Cur Loss: 0.40194365, Cur Avg Loss: 0.62132062, Log Avg loss: 0.68326421, Global Avg Loss: 2.82399131, Time: 0.0208 Steps: 14380, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001616, Sample Num: 25856, Cur Loss: 0.23517907, Cur Avg Loss: 0.62131445, Log Avg loss: 0.62032462, Global Avg Loss: 2.82245992, Time: 0.0208 Steps: 14390, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001626, Sample Num: 26016, Cur Loss: 0.80654109, Cur Avg Loss: 0.62088793, Log Avg loss: 0.55196251, Global Avg Loss: 2.82088319, Time: 0.0208 Steps: 14400, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001636, Sample Num: 26176, Cur Loss: 0.57535392, Cur Avg Loss: 0.62002636, Log Avg loss: 0.47993467, Global Avg Loss: 2.81925866, Time: 0.0208 Steps: 14410, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001646, Sample Num: 26336, Cur Loss: 0.25374463, Cur Avg Loss: 0.61890797, Log Avg loss: 0.43593886, Global Avg Loss: 2.81760587, Time: 0.0209 Steps: 14420, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001656, Sample Num: 26496, Cur Loss: 0.29243088, Cur Avg Loss: 0.61834526, Log Avg loss: 0.52572283, Global Avg Loss: 2.81601759, Time: 0.0209 Steps: 14430, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001666, Sample Num: 26656, Cur Loss: 0.38330346, Cur Avg Loss: 0.61763724, Log Avg loss: 0.50038961, Global Avg Loss: 2.81441397, Time: 0.0209 Steps: 14440, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001676, Sample Num: 26816, Cur Loss: 0.24705043, Cur Avg Loss: 0.61732010, Log Avg loss: 0.56448548, Global Avg Loss: 2.81285693, Time: 0.0209 Steps: 14450, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001686, Sample Num: 26976, Cur Loss: 0.59461302, Cur Avg Loss: 0.61694875, Log Avg loss: 0.55470935, Global Avg Loss: 2.81129528, Time: 0.0209 Steps: 14460, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001696, Sample Num: 27136, Cur Loss: 1.10446489, Cur Avg Loss: 0.61675315, Log Avg loss: 0.58377610, Global Avg Loss: 2.80975587, Time: 0.0209 Steps: 14470, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001706, Sample Num: 27296, Cur Loss: 0.46685719, Cur Avg Loss: 0.61596450, Log Avg loss: 0.48220834, Global Avg Loss: 2.80814845, Time: 0.0208 Steps: 14480, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001716, Sample Num: 27456, Cur Loss: 0.81689256, Cur Avg Loss: 0.61556221, Log Avg loss: 0.54693204, Global Avg Loss: 2.80658791, Time: 0.0209 Steps: 14490, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001726, Sample Num: 27616, Cur Loss: 0.35747468, Cur Avg Loss: 0.61385696, Log Avg loss: 0.32123520, Global Avg Loss: 2.80487388, Time: 0.0209 Steps: 14500, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001736, Sample Num: 27776, Cur Loss: 0.90060055, Cur Avg Loss: 0.61421724, Log Avg loss: 0.67640265, Global Avg Loss: 2.80340698, Time: 0.0209 Steps: 14510, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001746, Sample Num: 27936, Cur Loss: 0.59688342, Cur Avg Loss: 0.61375471, Log Avg loss: 0.53345982, Global Avg Loss: 2.80184365, Time: 0.0208 Steps: 14520, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001756, Sample Num: 28096, Cur Loss: 0.51177186, Cur Avg Loss: 0.61425929, Log Avg loss: 0.70235905, Global Avg Loss: 2.80039872, Time: 0.0208 Steps: 14530, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001766, Sample Num: 28256, Cur Loss: 0.57289803, Cur Avg Loss: 0.61479357, Log Avg loss: 0.70861213, Global Avg Loss: 2.79896008, Time: 0.0208 Steps: 14540, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001776, Sample Num: 28416, Cur Loss: 0.23256981, Cur Avg Loss: 0.61398194, Log Avg loss: 0.47064792, Global Avg Loss: 2.79735986, Time: 0.0208 Steps: 14550, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001786, Sample Num: 28576, Cur Loss: 0.30655640, Cur Avg Loss: 0.61341422, Log Avg loss: 0.51258784, Global Avg Loss: 2.79579065, Time: 0.0208 Steps: 14560, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001796, Sample Num: 28736, Cur Loss: 0.44030023, Cur Avg Loss: 0.61285732, Log Avg loss: 0.51339538, Global Avg Loss: 2.79422415, Time: 0.0209 Steps: 14570, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001806, Sample Num: 28896, Cur Loss: 0.97638237, Cur Avg Loss: 0.61269673, Log Avg loss: 0.58385345, Global Avg Loss: 2.79270812, Time: 0.0208 Steps: 14580, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001816, Sample Num: 29056, Cur Loss: 1.07222927, Cur Avg Loss: 0.61360844, Log Avg loss: 0.77826444, Global Avg Loss: 2.79132742, Time: 0.0207 Steps: 14590, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001826, Sample Num: 29216, Cur Loss: 0.78880876, Cur Avg Loss: 0.61426603, Log Avg loss: 0.73368311, Global Avg Loss: 2.78991807, Time: 0.0207 Steps: 14600, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001836, Sample Num: 29376, Cur Loss: 0.51183951, Cur Avg Loss: 0.61425301, Log Avg loss: 0.61187550, Global Avg Loss: 2.78842728, Time: 0.0207 Steps: 14610, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001846, Sample Num: 29536, Cur Loss: 0.33388415, Cur Avg Loss: 0.61402178, Log Avg loss: 0.57156836, Global Avg Loss: 2.78691096, Time: 0.0207 Steps: 14620, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001856, Sample Num: 29696, Cur Loss: 0.27482331, Cur Avg Loss: 0.61430165, Log Avg loss: 0.66596664, Global Avg Loss: 2.78546124, Time: 0.0208 Steps: 14630, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001866, Sample Num: 29856, Cur Loss: 0.62373179, Cur Avg Loss: 0.61303587, Log Avg loss: 0.37810668, Global Avg Loss: 2.78381687, Time: 0.0208 Steps: 14640, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001876, Sample Num: 30016, Cur Loss: 0.30061215, Cur Avg Loss: 0.61226845, Log Avg loss: 0.46906814, Global Avg Loss: 2.78223684, Time: 0.0208 Steps: 14650, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001886, Sample Num: 30176, Cur Loss: 0.43144107, Cur Avg Loss: 0.61112893, Log Avg loss: 0.39735412, Global Avg Loss: 2.78061004, Time: 0.0207 Steps: 14660, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001896, Sample Num: 30336, Cur Loss: 0.35078609, Cur Avg Loss: 0.61036554, Log Avg loss: 0.46639104, Global Avg Loss: 2.77903253, Time: 0.0208 Steps: 14670, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001906, Sample Num: 30496, Cur Loss: 0.86954880, Cur Avg Loss: 0.61030935, Log Avg loss: 0.59965459, Global Avg Loss: 2.77754794, Time: 0.0208 Steps: 14680, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001916, Sample Num: 30656, Cur Loss: 0.89193636, Cur Avg Loss: 0.61004752, Log Avg loss: 0.56014291, Global Avg Loss: 2.77603847, Time: 0.0207 Steps: 14690, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001926, Sample Num: 30816, Cur Loss: 0.65244842, Cur Avg Loss: 0.60981426, Log Avg loss: 0.56512243, Global Avg Loss: 2.77453445, Time: 0.0207 Steps: 14700, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001936, Sample Num: 30976, Cur Loss: 0.49291822, Cur Avg Loss: 0.60953577, Log Avg loss: 0.55589873, Global Avg Loss: 2.77302620, Time: 0.0208 Steps: 14710, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001946, Sample Num: 31136, Cur Loss: 0.96214080, Cur Avg Loss: 0.61046956, Log Avg loss: 0.79125020, Global Avg Loss: 2.77167988, Time: 0.0207 Steps: 14720, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001956, Sample Num: 31296, Cur Loss: 0.40103161, Cur Avg Loss: 0.61011102, Log Avg loss: 0.54033981, Global Avg Loss: 2.77016505, Time: 0.0207 Steps: 14730, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001966, Sample Num: 31456, Cur Loss: 0.75964707, Cur Avg Loss: 0.61040998, Log Avg loss: 0.66888725, Global Avg Loss: 2.76873949, Time: 0.0208 Steps: 14740, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001976, Sample Num: 31616, Cur Loss: 0.83368862, Cur Avg Loss: 0.61049241, Log Avg loss: 0.62669748, Global Avg Loss: 2.76728726, Time: 0.0207 Steps: 14750, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001986, Sample Num: 31776, Cur Loss: 0.32688040, Cur Avg Loss: 0.60994698, Log Avg loss: 0.50217057, Global Avg Loss: 2.76575263, Time: 0.0207 Steps: 14760, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001996, Sample Num: 31936, Cur Loss: 0.27980715, Cur Avg Loss: 0.60979402, Log Avg loss: 0.57941655, Global Avg Loss: 2.76427237, Time: 0.0208 Steps: 14770, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002006, Sample Num: 32096, Cur Loss: 0.33193132, Cur Avg Loss: 0.60926709, Log Avg loss: 0.50409100, Global Avg Loss: 2.76274316, Time: 0.0208 Steps: 14780, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002016, Sample Num: 32256, Cur Loss: 0.93728167, Cur Avg Loss: 0.60944401, Log Avg loss: 0.64493374, Global Avg Loss: 2.76131124, Time: 0.0208 Steps: 14790, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002026, Sample Num: 32416, Cur Loss: 0.68802619, Cur Avg Loss: 0.60916505, Log Avg loss: 0.55292671, Global Avg Loss: 2.75981909, Time: 0.0208 Steps: 14800, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002036, Sample Num: 32576, Cur Loss: 0.91552019, Cur Avg Loss: 0.60932008, Log Avg loss: 0.64073019, Global Avg Loss: 2.75838824, Time: 0.0207 Steps: 14810, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002046, Sample Num: 32736, Cur Loss: 0.27049351, Cur Avg Loss: 0.60920166, Log Avg loss: 0.58509088, Global Avg Loss: 2.75692177, Time: 0.0207 Steps: 14820, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002056, Sample Num: 32896, Cur Loss: 0.81098652, Cur Avg Loss: 0.60820825, Log Avg loss: 0.40495619, Global Avg Loss: 2.75533582, Time: 0.0210 Steps: 14830, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002066, Sample Num: 33056, Cur Loss: 0.26681167, Cur Avg Loss: 0.60717627, Log Avg loss: 0.39500077, Global Avg Loss: 2.75374530, Time: 0.0209 Steps: 14840, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002076, Sample Num: 33216, Cur Loss: 0.34658802, Cur Avg Loss: 0.60720632, Log Avg loss: 0.61341418, Global Avg Loss: 2.75230400, Time: 0.0209 Steps: 14850, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002086, Sample Num: 33376, Cur Loss: 0.85901070, Cur Avg Loss: 0.60631470, Log Avg loss: 0.42121522, Global Avg Loss: 2.75073530, Time: 0.0209 Steps: 14860, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002096, Sample Num: 33536, Cur Loss: 0.24538459, Cur Avg Loss: 0.60629635, Log Avg loss: 0.60246891, Global Avg Loss: 2.74929060, Time: 0.0209 Steps: 14870, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002106, Sample Num: 33696, Cur Loss: 0.39623907, Cur Avg Loss: 0.60636063, Log Avg loss: 0.61983422, Global Avg Loss: 2.74785952, Time: 0.0208 Steps: 14880, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002116, Sample Num: 33856, Cur Loss: 0.47092646, Cur Avg Loss: 0.60592090, Log Avg loss: 0.51331224, Global Avg Loss: 2.74635881, Time: 0.0208 Steps: 14890, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002126, Sample Num: 34016, Cur Loss: 0.49585861, Cur Avg Loss: 0.60510217, Log Avg loss: 0.43186077, Global Avg Loss: 2.74480546, Time: 0.0208 Steps: 14900, Updated lr: 0.000087 ***** Running evaluation checkpoint-14903 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-14903 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.553417, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.569675, "eval_total_loss": 400.481364, "eval_mae": 0.576522, "eval_mse": 0.569795, "eval_r2": 0.637801, "eval_sp_statistic": 0.781514, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.804459, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.72748, "test_total_loss": 365.195086, "test_mae": 0.572287, "test_mse": 0.727739, "test_r2": 0.530311, "test_sp_statistic": 0.711694, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.768246, "test_ps_pvalue": 0.0, "lr": 8.681555239449977e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.7443813508566333, "train_cur_epoch_loss": 1288.361170500517, "train_cur_epoch_avg_loss": 0.6051485065761, "train_cur_epoch_time": 44.553417444229126, "train_cur_epoch_avg_time": 0.020926922237777888, "epoch": 7, "step": 14903} ################################################## Training, Epoch: 0008, Batch: 000007, Sample Num: 112, Cur Loss: 0.94602519, Cur Avg Loss: 0.69384503, Log Avg loss: 0.67708620, Global Avg Loss: 2.74341866, Time: 0.0209 Steps: 14910, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000017, Sample Num: 272, Cur Loss: 0.27779391, Cur Avg Loss: 0.49654932, Log Avg loss: 0.35844233, Global Avg Loss: 2.74182015, Time: 0.0208 Steps: 14920, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000027, Sample Num: 432, Cur Loss: 0.64493591, Cur Avg Loss: 0.50276428, Log Avg loss: 0.51332969, Global Avg Loss: 2.74032752, Time: 0.0208 Steps: 14930, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000037, Sample Num: 592, Cur Loss: 0.64095604, Cur Avg Loss: 0.51433226, Log Avg loss: 0.54556582, Global Avg Loss: 2.73885847, Time: 0.0208 Steps: 14940, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000047, Sample Num: 752, Cur Loss: 2.00041175, Cur Avg Loss: 0.56545113, Log Avg loss: 0.75459093, Global Avg Loss: 2.73753120, Time: 0.0208 Steps: 14950, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000057, Sample Num: 912, Cur Loss: 1.67212713, Cur Avg Loss: 0.60878144, Log Avg loss: 0.81243391, Global Avg Loss: 2.73624437, Time: 0.0208 Steps: 14960, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000067, Sample Num: 1072, Cur Loss: 1.04286480, Cur Avg Loss: 0.64503911, Log Avg loss: 0.85170782, Global Avg Loss: 2.73498550, Time: 0.0208 Steps: 14970, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000077, Sample Num: 1232, Cur Loss: 0.39667052, Cur Avg Loss: 0.64505749, Log Avg loss: 0.64518069, Global Avg Loss: 2.73359043, Time: 0.0208 Steps: 14980, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000087, Sample Num: 1392, Cur Loss: 0.20182128, Cur Avg Loss: 0.65663835, Log Avg loss: 0.74581090, Global Avg Loss: 2.73226436, Time: 0.0208 Steps: 14990, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000097, Sample Num: 1552, Cur Loss: 0.36223286, Cur Avg Loss: 0.63849654, Log Avg loss: 0.48066284, Global Avg Loss: 2.73076330, Time: 0.0209 Steps: 15000, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000107, Sample Num: 1712, Cur Loss: 0.62505710, Cur Avg Loss: 0.62932376, Log Avg loss: 0.54034773, Global Avg Loss: 2.72930399, Time: 0.0209 Steps: 15010, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000117, Sample Num: 1872, Cur Loss: 0.61858231, Cur Avg Loss: 0.62575760, Log Avg loss: 0.58759971, Global Avg Loss: 2.72787809, Time: 0.0208 Steps: 15020, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000127, Sample Num: 2032, Cur Loss: 0.46304604, Cur Avg Loss: 0.63800691, Log Avg loss: 0.78132389, Global Avg Loss: 2.72658298, Time: 0.0208 Steps: 15030, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000137, Sample Num: 2192, Cur Loss: 0.63590842, Cur Avg Loss: 0.63228771, Log Avg loss: 0.55965390, Global Avg Loss: 2.72514220, Time: 0.0207 Steps: 15040, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000147, Sample Num: 2352, Cur Loss: 0.24510252, Cur Avg Loss: 0.62630598, Log Avg loss: 0.54435626, Global Avg Loss: 2.72369317, Time: 0.0208 Steps: 15050, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000157, Sample Num: 2512, Cur Loss: 1.10909808, Cur Avg Loss: 0.62586776, Log Avg loss: 0.61942586, Global Avg Loss: 2.72229592, Time: 0.0207 Steps: 15060, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000167, Sample Num: 2672, Cur Loss: 0.52556205, Cur Avg Loss: 0.61105924, Log Avg loss: 0.37856556, Global Avg Loss: 2.72074069, Time: 0.0207 Steps: 15070, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000177, Sample Num: 2832, Cur Loss: 0.23199609, Cur Avg Loss: 0.61828926, Log Avg loss: 0.73903049, Global Avg Loss: 2.71942656, Time: 0.0209 Steps: 15080, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000187, Sample Num: 2992, Cur Loss: 0.29229239, Cur Avg Loss: 0.60355548, Log Avg loss: 0.34276768, Global Avg Loss: 2.71785157, Time: 0.0207 Steps: 15090, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000197, Sample Num: 3152, Cur Loss: 0.39112231, Cur Avg Loss: 0.59779528, Log Avg loss: 0.49007956, Global Avg Loss: 2.71637622, Time: 0.0207 Steps: 15100, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000207, Sample Num: 3312, Cur Loss: 0.31418371, Cur Avg Loss: 0.59196713, Log Avg loss: 0.47715252, Global Avg Loss: 2.71489427, Time: 0.0209 Steps: 15110, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000217, Sample Num: 3472, Cur Loss: 0.31687549, Cur Avg Loss: 0.58454210, Log Avg loss: 0.43084401, Global Avg Loss: 2.71338366, Time: 0.0207 Steps: 15120, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000227, Sample Num: 3632, Cur Loss: 0.18677519, Cur Avg Loss: 0.59101563, Log Avg loss: 0.73149117, Global Avg Loss: 2.71207375, Time: 0.0209 Steps: 15130, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000237, Sample Num: 3792, Cur Loss: 1.08333933, Cur Avg Loss: 0.59248147, Log Avg loss: 0.62575603, Global Avg Loss: 2.71069573, Time: 0.0208 Steps: 15140, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000247, Sample Num: 3952, Cur Loss: 0.65039992, Cur Avg Loss: 0.58793911, Log Avg loss: 0.48028519, Global Avg Loss: 2.70922351, Time: 0.0208 Steps: 15150, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000257, Sample Num: 4112, Cur Loss: 0.22657073, Cur Avg Loss: 0.59006895, Log Avg loss: 0.64267590, Global Avg Loss: 2.70786036, Time: 0.0245 Steps: 15160, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000267, Sample Num: 4272, Cur Loss: 0.20686144, Cur Avg Loss: 0.58742811, Log Avg loss: 0.51955852, Global Avg Loss: 2.70641784, Time: 0.0209 Steps: 15170, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000277, Sample Num: 4432, Cur Loss: 0.35456848, Cur Avg Loss: 0.58402036, Log Avg loss: 0.49303360, Global Avg Loss: 2.70495974, Time: 0.0209 Steps: 15180, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000287, Sample Num: 4592, Cur Loss: 0.40065271, Cur Avg Loss: 0.58100363, Log Avg loss: 0.49744024, Global Avg Loss: 2.70350647, Time: 0.0208 Steps: 15190, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000297, Sample Num: 4752, Cur Loss: 0.39860409, Cur Avg Loss: 0.57649976, Log Avg loss: 0.44723853, Global Avg Loss: 2.70202209, Time: 0.0208 Steps: 15200, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000307, Sample Num: 4912, Cur Loss: 0.39496392, Cur Avg Loss: 0.57922645, Log Avg loss: 0.66020930, Global Avg Loss: 2.70067967, Time: 0.0208 Steps: 15210, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000317, Sample Num: 5072, Cur Loss: 0.17261325, Cur Avg Loss: 0.57876170, Log Avg loss: 0.56449363, Global Avg Loss: 2.69927613, Time: 0.0208 Steps: 15220, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000327, Sample Num: 5232, Cur Loss: 0.33581257, Cur Avg Loss: 0.57774999, Log Avg loss: 0.54567887, Global Avg Loss: 2.69786208, Time: 0.0208 Steps: 15230, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000337, Sample Num: 5392, Cur Loss: 0.24815965, Cur Avg Loss: 0.57246330, Log Avg loss: 0.39958873, Global Avg Loss: 2.69635403, Time: 0.0208 Steps: 15240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000347, Sample Num: 5552, Cur Loss: 0.40385234, Cur Avg Loss: 0.57013233, Log Avg loss: 0.49157842, Global Avg Loss: 2.69490827, Time: 0.0207 Steps: 15250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000357, Sample Num: 5712, Cur Loss: 0.53960574, Cur Avg Loss: 0.56496498, Log Avg loss: 0.38565806, Global Avg Loss: 2.69339500, Time: 0.0208 Steps: 15260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000367, Sample Num: 5872, Cur Loss: 1.01487291, Cur Avg Loss: 0.56392078, Log Avg loss: 0.52664284, Global Avg Loss: 2.69197604, Time: 0.0208 Steps: 15270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000377, Sample Num: 6032, Cur Loss: 0.16907144, Cur Avg Loss: 0.56216602, Log Avg loss: 0.49776614, Global Avg Loss: 2.69054004, Time: 0.0208 Steps: 15280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000387, Sample Num: 6192, Cur Loss: 1.00668716, Cur Avg Loss: 0.56710341, Log Avg loss: 0.75324316, Global Avg Loss: 2.68927301, Time: 0.0208 Steps: 15290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000397, Sample Num: 6352, Cur Loss: 0.46894297, Cur Avg Loss: 0.56664362, Log Avg loss: 0.54884975, Global Avg Loss: 2.68787404, Time: 0.0208 Steps: 15300, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000407, Sample Num: 6512, Cur Loss: 0.50649410, Cur Avg Loss: 0.56875294, Log Avg loss: 0.65249286, Global Avg Loss: 2.68654459, Time: 0.0208 Steps: 15310, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000417, Sample Num: 6672, Cur Loss: 0.59882104, Cur Avg Loss: 0.57249767, Log Avg loss: 0.72490829, Global Avg Loss: 2.68526415, Time: 0.0208 Steps: 15320, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000427, Sample Num: 6832, Cur Loss: 0.41870722, Cur Avg Loss: 0.57194009, Log Avg loss: 0.54868884, Global Avg Loss: 2.68387043, Time: 0.0208 Steps: 15330, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000437, Sample Num: 6992, Cur Loss: 0.64345121, Cur Avg Loss: 0.57021404, Log Avg loss: 0.49651195, Global Avg Loss: 2.68244451, Time: 0.0208 Steps: 15340, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000447, Sample Num: 7152, Cur Loss: 1.00011265, Cur Avg Loss: 0.56964660, Log Avg loss: 0.54484955, Global Avg Loss: 2.68105194, Time: 0.0208 Steps: 15350, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000457, Sample Num: 7312, Cur Loss: 0.92028910, Cur Avg Loss: 0.57260792, Log Avg loss: 0.70497895, Global Avg Loss: 2.67976544, Time: 0.0208 Steps: 15360, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000467, Sample Num: 7472, Cur Loss: 1.04009616, Cur Avg Loss: 0.57717761, Log Avg loss: 0.78601244, Global Avg Loss: 2.67853333, Time: 0.0208 Steps: 15370, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000477, Sample Num: 7632, Cur Loss: 2.58571696, Cur Avg Loss: 0.58424524, Log Avg loss: 0.91430324, Global Avg Loss: 2.67738623, Time: 0.0208 Steps: 15380, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000487, Sample Num: 7792, Cur Loss: 0.49710879, Cur Avg Loss: 0.58529056, Log Avg loss: 0.63515252, Global Avg Loss: 2.67605924, Time: 0.0208 Steps: 15390, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000497, Sample Num: 7952, Cur Loss: 1.04135704, Cur Avg Loss: 0.58522401, Log Avg loss: 0.58198299, Global Avg Loss: 2.67469945, Time: 0.0208 Steps: 15400, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000507, Sample Num: 8112, Cur Loss: 1.25119472, Cur Avg Loss: 0.58700662, Log Avg loss: 0.67560243, Global Avg Loss: 2.67340218, Time: 0.0208 Steps: 15410, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000517, Sample Num: 8272, Cur Loss: 0.27515161, Cur Avg Loss: 0.58582683, Log Avg loss: 0.52601144, Global Avg Loss: 2.67200958, Time: 0.0210 Steps: 15420, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000527, Sample Num: 8432, Cur Loss: 0.30509132, Cur Avg Loss: 0.58644439, Log Avg loss: 0.61837238, Global Avg Loss: 2.67067864, Time: 0.0209 Steps: 15430, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000537, Sample Num: 8592, Cur Loss: 0.24748304, Cur Avg Loss: 0.58472333, Log Avg loss: 0.49402333, Global Avg Loss: 2.66926889, Time: 0.0208 Steps: 15440, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000547, Sample Num: 8752, Cur Loss: 0.19222079, Cur Avg Loss: 0.58285408, Log Avg loss: 0.48247501, Global Avg Loss: 2.66785349, Time: 0.0208 Steps: 15450, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000557, Sample Num: 8912, Cur Loss: 0.31024891, Cur Avg Loss: 0.58207635, Log Avg loss: 0.53953495, Global Avg Loss: 2.66647683, Time: 0.0209 Steps: 15460, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000567, Sample Num: 9072, Cur Loss: 0.28258073, Cur Avg Loss: 0.57973307, Log Avg loss: 0.44921200, Global Avg Loss: 2.66504356, Time: 0.0208 Steps: 15470, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000577, Sample Num: 9232, Cur Loss: 0.45096904, Cur Avg Loss: 0.57756765, Log Avg loss: 0.45478859, Global Avg Loss: 2.66361575, Time: 0.0209 Steps: 15480, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000587, Sample Num: 9392, Cur Loss: 1.02421343, Cur Avg Loss: 0.57762390, Log Avg loss: 0.58086933, Global Avg Loss: 2.66227117, Time: 0.0208 Steps: 15490, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000597, Sample Num: 9552, Cur Loss: 0.56796694, Cur Avg Loss: 0.57658261, Log Avg loss: 0.51545931, Global Avg Loss: 2.66088614, Time: 0.0209 Steps: 15500, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000607, Sample Num: 9712, Cur Loss: 0.81323993, Cur Avg Loss: 0.57800141, Log Avg loss: 0.66270362, Global Avg Loss: 2.65959782, Time: 0.0208 Steps: 15510, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000617, Sample Num: 9872, Cur Loss: 0.42147684, Cur Avg Loss: 0.58114304, Log Avg loss: 0.77183971, Global Avg Loss: 2.65838148, Time: 0.0208 Steps: 15520, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000627, Sample Num: 10032, Cur Loss: 0.60968316, Cur Avg Loss: 0.58444106, Log Avg loss: 0.78792910, Global Avg Loss: 2.65717706, Time: 0.0208 Steps: 15530, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000637, Sample Num: 10192, Cur Loss: 0.25164223, Cur Avg Loss: 0.58308695, Log Avg loss: 0.49818432, Global Avg Loss: 2.65578775, Time: 0.0207 Steps: 15540, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000647, Sample Num: 10352, Cur Loss: 0.59641063, Cur Avg Loss: 0.58192481, Log Avg loss: 0.50789611, Global Avg Loss: 2.65440647, Time: 0.0208 Steps: 15550, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000657, Sample Num: 10512, Cur Loss: 0.70285398, Cur Avg Loss: 0.58289240, Log Avg loss: 0.64549565, Global Avg Loss: 2.65311540, Time: 0.0208 Steps: 15560, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000667, Sample Num: 10672, Cur Loss: 0.26034039, Cur Avg Loss: 0.58109861, Log Avg loss: 0.46324694, Global Avg Loss: 2.65170893, Time: 0.0207 Steps: 15570, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000677, Sample Num: 10832, Cur Loss: 0.60720235, Cur Avg Loss: 0.57943603, Log Avg loss: 0.46854157, Global Avg Loss: 2.65030767, Time: 0.0208 Steps: 15580, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000687, Sample Num: 10992, Cur Loss: 0.17459111, Cur Avg Loss: 0.57716713, Log Avg loss: 0.42356284, Global Avg Loss: 2.64887935, Time: 0.0208 Steps: 15590, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000697, Sample Num: 11152, Cur Loss: 0.24570322, Cur Avg Loss: 0.57481642, Log Avg loss: 0.41332222, Global Avg Loss: 2.64744630, Time: 0.0208 Steps: 15600, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000707, Sample Num: 11312, Cur Loss: 0.62472266, Cur Avg Loss: 0.57615908, Log Avg loss: 0.66974292, Global Avg Loss: 2.64617936, Time: 0.0208 Steps: 15610, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000717, Sample Num: 11472, Cur Loss: 0.60802531, Cur Avg Loss: 0.57584894, Log Avg loss: 0.55392192, Global Avg Loss: 2.64483988, Time: 0.0208 Steps: 15620, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000727, Sample Num: 11632, Cur Loss: 0.90189517, Cur Avg Loss: 0.57854274, Log Avg loss: 0.77168783, Global Avg Loss: 2.64364145, Time: 0.0208 Steps: 15630, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000737, Sample Num: 11792, Cur Loss: 0.49282265, Cur Avg Loss: 0.57701871, Log Avg loss: 0.46622239, Global Avg Loss: 2.64224924, Time: 0.0208 Steps: 15640, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000747, Sample Num: 11952, Cur Loss: 0.62339038, Cur Avg Loss: 0.57791136, Log Avg loss: 0.64369961, Global Avg Loss: 2.64097221, Time: 0.0208 Steps: 15650, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000757, Sample Num: 12112, Cur Loss: 0.93431824, Cur Avg Loss: 0.57833114, Log Avg loss: 0.60968854, Global Avg Loss: 2.63967509, Time: 0.0207 Steps: 15660, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000767, Sample Num: 12272, Cur Loss: 0.52297992, Cur Avg Loss: 0.57611584, Log Avg loss: 0.40841733, Global Avg Loss: 2.63825119, Time: 0.0207 Steps: 15670, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000777, Sample Num: 12432, Cur Loss: 0.49262652, Cur Avg Loss: 0.57426457, Log Avg loss: 0.43227238, Global Avg Loss: 2.63684431, Time: 0.0207 Steps: 15680, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000787, Sample Num: 12592, Cur Loss: 0.36382246, Cur Avg Loss: 0.57308206, Log Avg loss: 0.48120131, Global Avg Loss: 2.63547042, Time: 0.0207 Steps: 15690, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000797, Sample Num: 12752, Cur Loss: 0.43842992, Cur Avg Loss: 0.57070492, Log Avg loss: 0.38362346, Global Avg Loss: 2.63403612, Time: 0.0207 Steps: 15700, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000807, Sample Num: 12912, Cur Loss: 0.46912265, Cur Avg Loss: 0.57185823, Log Avg loss: 0.66377727, Global Avg Loss: 2.63278198, Time: 0.0207 Steps: 15710, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000817, Sample Num: 13072, Cur Loss: 1.20159626, Cur Avg Loss: 0.57241154, Log Avg loss: 0.61706403, Global Avg Loss: 2.63149971, Time: 0.0208 Steps: 15720, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000827, Sample Num: 13232, Cur Loss: 0.36140490, Cur Avg Loss: 0.57218423, Log Avg loss: 0.55361241, Global Avg Loss: 2.63017874, Time: 0.0208 Steps: 15730, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000837, Sample Num: 13392, Cur Loss: 0.45654905, Cur Avg Loss: 0.57197274, Log Avg loss: 0.55448242, Global Avg Loss: 2.62886000, Time: 0.0208 Steps: 15740, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000847, Sample Num: 13552, Cur Loss: 0.72238922, Cur Avg Loss: 0.57147362, Log Avg loss: 0.52969780, Global Avg Loss: 2.62752720, Time: 0.0207 Steps: 15750, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000857, Sample Num: 13712, Cur Loss: 0.24477415, Cur Avg Loss: 0.57105956, Log Avg loss: 0.53598833, Global Avg Loss: 2.62620008, Time: 0.0208 Steps: 15760, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000867, Sample Num: 13872, Cur Loss: 0.83742458, Cur Avg Loss: 0.56954811, Log Avg loss: 0.44001679, Global Avg Loss: 2.62481379, Time: 0.0208 Steps: 15770, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000877, Sample Num: 14032, Cur Loss: 0.49382937, Cur Avg Loss: 0.57075517, Log Avg loss: 0.67540748, Global Avg Loss: 2.62357843, Time: 0.0207 Steps: 15780, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000887, Sample Num: 14192, Cur Loss: 0.90408510, Cur Avg Loss: 0.57103523, Log Avg loss: 0.59559657, Global Avg Loss: 2.62229408, Time: 0.0208 Steps: 15790, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000897, Sample Num: 14352, Cur Loss: 1.75388432, Cur Avg Loss: 0.57530791, Log Avg loss: 0.95429437, Global Avg Loss: 2.62123838, Time: 0.0208 Steps: 15800, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000907, Sample Num: 14512, Cur Loss: 0.61713207, Cur Avg Loss: 0.57563524, Log Avg loss: 0.60499708, Global Avg Loss: 2.61996309, Time: 0.0207 Steps: 15810, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000917, Sample Num: 14672, Cur Loss: 0.34049380, Cur Avg Loss: 0.57649399, Log Avg loss: 0.65438243, Global Avg Loss: 2.61872062, Time: 0.0207 Steps: 15820, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000927, Sample Num: 14832, Cur Loss: 0.67714816, Cur Avg Loss: 0.57545228, Log Avg loss: 0.47992721, Global Avg Loss: 2.61736952, Time: 0.0207 Steps: 15830, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000937, Sample Num: 14992, Cur Loss: 0.93777949, Cur Avg Loss: 0.57804268, Log Avg loss: 0.81817269, Global Avg Loss: 2.61623367, Time: 0.0209 Steps: 15840, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000947, Sample Num: 15152, Cur Loss: 0.74919319, Cur Avg Loss: 0.57768603, Log Avg loss: 0.54426827, Global Avg Loss: 2.61492643, Time: 0.0208 Steps: 15850, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000957, Sample Num: 15312, Cur Loss: 0.94701934, Cur Avg Loss: 0.57620747, Log Avg loss: 0.43618777, Global Avg Loss: 2.61355270, Time: 0.0208 Steps: 15860, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000967, Sample Num: 15472, Cur Loss: 0.42050180, Cur Avg Loss: 0.57660564, Log Avg loss: 0.61471045, Global Avg Loss: 2.61229319, Time: 0.0208 Steps: 15870, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000977, Sample Num: 15632, Cur Loss: 0.61951864, Cur Avg Loss: 0.57664096, Log Avg loss: 0.58005704, Global Avg Loss: 2.61101344, Time: 0.0208 Steps: 15880, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000987, Sample Num: 15792, Cur Loss: 0.33427280, Cur Avg Loss: 0.57626850, Log Avg loss: 0.53987836, Global Avg Loss: 2.60971002, Time: 0.0207 Steps: 15890, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000997, Sample Num: 15952, Cur Loss: 0.42681921, Cur Avg Loss: 0.57573133, Log Avg loss: 0.52271299, Global Avg Loss: 2.60839745, Time: 0.0207 Steps: 15900, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001007, Sample Num: 16112, Cur Loss: 1.28734875, Cur Avg Loss: 0.57733594, Log Avg loss: 0.73731596, Global Avg Loss: 2.60722141, Time: 0.0208 Steps: 15910, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001017, Sample Num: 16272, Cur Loss: 0.36178881, Cur Avg Loss: 0.57691344, Log Avg loss: 0.53436692, Global Avg Loss: 2.60591936, Time: 0.0208 Steps: 15920, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001027, Sample Num: 16432, Cur Loss: 0.49790359, Cur Avg Loss: 0.57740457, Log Avg loss: 0.62735296, Global Avg Loss: 2.60467732, Time: 0.0248 Steps: 15930, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001037, Sample Num: 16592, Cur Loss: 0.98412991, Cur Avg Loss: 0.57685586, Log Avg loss: 0.52050271, Global Avg Loss: 2.60336981, Time: 0.0211 Steps: 15940, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001047, Sample Num: 16752, Cur Loss: 0.42556730, Cur Avg Loss: 0.57633520, Log Avg loss: 0.52234300, Global Avg Loss: 2.60206509, Time: 0.0211 Steps: 15950, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001057, Sample Num: 16912, Cur Loss: 0.44185403, Cur Avg Loss: 0.57836013, Log Avg loss: 0.79037023, Global Avg Loss: 2.60092995, Time: 0.0211 Steps: 15960, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001067, Sample Num: 17072, Cur Loss: 0.54272771, Cur Avg Loss: 0.57748477, Log Avg loss: 0.48495916, Global Avg Loss: 2.59960498, Time: 0.0211 Steps: 15970, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001077, Sample Num: 17232, Cur Loss: 0.58965278, Cur Avg Loss: 0.57728850, Log Avg loss: 0.55634715, Global Avg Loss: 2.59832634, Time: 0.0211 Steps: 15980, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001087, Sample Num: 17392, Cur Loss: 0.28327140, Cur Avg Loss: 0.57642570, Log Avg loss: 0.48350162, Global Avg Loss: 2.59700375, Time: 0.0211 Steps: 15990, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001097, Sample Num: 17552, Cur Loss: 0.86693233, Cur Avg Loss: 0.57663914, Log Avg loss: 0.59983988, Global Avg Loss: 2.59575553, Time: 0.0211 Steps: 16000, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001107, Sample Num: 17712, Cur Loss: 0.68917716, Cur Avg Loss: 0.57690272, Log Avg loss: 0.60581792, Global Avg Loss: 2.59451259, Time: 0.0211 Steps: 16010, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001117, Sample Num: 17872, Cur Loss: 0.50341189, Cur Avg Loss: 0.57752018, Log Avg loss: 0.64587322, Global Avg Loss: 2.59329621, Time: 0.0211 Steps: 16020, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001127, Sample Num: 18032, Cur Loss: 0.88498038, Cur Avg Loss: 0.57866660, Log Avg loss: 0.70672203, Global Avg Loss: 2.59211931, Time: 0.0211 Steps: 16030, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001137, Sample Num: 18192, Cur Loss: 0.31028575, Cur Avg Loss: 0.57857807, Log Avg loss: 0.56860065, Global Avg Loss: 2.59085776, Time: 0.0212 Steps: 16040, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001147, Sample Num: 18352, Cur Loss: 0.42862427, Cur Avg Loss: 0.57754030, Log Avg loss: 0.45954574, Global Avg Loss: 2.58952984, Time: 0.0211 Steps: 16050, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001157, Sample Num: 18512, Cur Loss: 0.31109399, Cur Avg Loss: 0.57719883, Log Avg loss: 0.53803155, Global Avg Loss: 2.58825245, Time: 0.0211 Steps: 16060, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001167, Sample Num: 18672, Cur Loss: 0.87689090, Cur Avg Loss: 0.57761577, Log Avg loss: 0.62585573, Global Avg Loss: 2.58703129, Time: 0.0211 Steps: 16070, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001177, Sample Num: 18832, Cur Loss: 0.69322312, Cur Avg Loss: 0.57824963, Log Avg loss: 0.65222163, Global Avg Loss: 2.58582805, Time: 0.0211 Steps: 16080, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001187, Sample Num: 18992, Cur Loss: 0.84150410, Cur Avg Loss: 0.58117107, Log Avg loss: 0.92502458, Global Avg Loss: 2.58479586, Time: 0.0211 Steps: 16090, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001197, Sample Num: 19152, Cur Loss: 0.57997960, Cur Avg Loss: 0.58131700, Log Avg loss: 0.59863841, Global Avg Loss: 2.58356222, Time: 0.0211 Steps: 16100, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001207, Sample Num: 19312, Cur Loss: 0.57473588, Cur Avg Loss: 0.58047268, Log Avg loss: 0.47940764, Global Avg Loss: 2.58225610, Time: 0.0211 Steps: 16110, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001217, Sample Num: 19472, Cur Loss: 0.40280092, Cur Avg Loss: 0.58027643, Log Avg loss: 0.55658895, Global Avg Loss: 2.58099948, Time: 0.0211 Steps: 16120, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001227, Sample Num: 19632, Cur Loss: 0.76119137, Cur Avg Loss: 0.58045965, Log Avg loss: 0.60275835, Global Avg Loss: 2.57977305, Time: 0.0211 Steps: 16130, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001237, Sample Num: 19792, Cur Loss: 0.43306983, Cur Avg Loss: 0.58003692, Log Avg loss: 0.52816728, Global Avg Loss: 2.57850192, Time: 0.0211 Steps: 16140, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001247, Sample Num: 19952, Cur Loss: 0.61884648, Cur Avg Loss: 0.57933515, Log Avg loss: 0.49252637, Global Avg Loss: 2.57721029, Time: 0.0211 Steps: 16150, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001257, Sample Num: 20112, Cur Loss: 0.93800408, Cur Avg Loss: 0.57970985, Log Avg loss: 0.62643464, Global Avg Loss: 2.57600313, Time: 0.0211 Steps: 16160, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001267, Sample Num: 20272, Cur Loss: 0.94713032, Cur Avg Loss: 0.58025584, Log Avg loss: 0.64888672, Global Avg Loss: 2.57481134, Time: 0.0211 Steps: 16170, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001277, Sample Num: 20432, Cur Loss: 0.35773277, Cur Avg Loss: 0.57924208, Log Avg loss: 0.45079853, Global Avg Loss: 2.57349860, Time: 0.0211 Steps: 16180, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001287, Sample Num: 20592, Cur Loss: 0.25224310, Cur Avg Loss: 0.57855402, Log Avg loss: 0.49068927, Global Avg Loss: 2.57221212, Time: 0.0209 Steps: 16190, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001297, Sample Num: 20752, Cur Loss: 0.72368073, Cur Avg Loss: 0.57958346, Log Avg loss: 0.71207272, Global Avg Loss: 2.57106389, Time: 0.0209 Steps: 16200, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001307, Sample Num: 20912, Cur Loss: 0.96600229, Cur Avg Loss: 0.58038229, Log Avg loss: 0.68399065, Global Avg Loss: 2.56989975, Time: 0.0209 Steps: 16210, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001317, Sample Num: 21072, Cur Loss: 0.40556425, Cur Avg Loss: 0.57953061, Log Avg loss: 0.46821496, Global Avg Loss: 2.56860401, Time: 0.0209 Steps: 16220, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001327, Sample Num: 21232, Cur Loss: 0.29948047, Cur Avg Loss: 0.57869302, Log Avg loss: 0.46838345, Global Avg Loss: 2.56730998, Time: 0.0208 Steps: 16230, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001337, Sample Num: 21392, Cur Loss: 0.81812847, Cur Avg Loss: 0.57726831, Log Avg loss: 0.38820910, Global Avg Loss: 2.56596817, Time: 0.0208 Steps: 16240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001347, Sample Num: 21552, Cur Loss: 0.80545092, Cur Avg Loss: 0.57809342, Log Avg loss: 0.68841043, Global Avg Loss: 2.56481275, Time: 0.0209 Steps: 16250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001357, Sample Num: 21712, Cur Loss: 0.51899838, Cur Avg Loss: 0.57769833, Log Avg loss: 0.52447963, Global Avg Loss: 2.56355793, Time: 0.0208 Steps: 16260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001367, Sample Num: 21872, Cur Loss: 0.93176699, Cur Avg Loss: 0.57750575, Log Avg loss: 0.55137218, Global Avg Loss: 2.56232118, Time: 0.0209 Steps: 16270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001377, Sample Num: 22032, Cur Loss: 0.57521564, Cur Avg Loss: 0.57721762, Log Avg loss: 0.53783104, Global Avg Loss: 2.56107764, Time: 0.0209 Steps: 16280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001387, Sample Num: 22192, Cur Loss: 0.39473671, Cur Avg Loss: 0.57997075, Log Avg loss: 0.95907705, Global Avg Loss: 2.56009421, Time: 0.0209 Steps: 16290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001397, Sample Num: 22352, Cur Loss: 0.38624355, Cur Avg Loss: 0.57969253, Log Avg loss: 0.54110352, Global Avg Loss: 2.55885557, Time: 0.0209 Steps: 16300, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001407, Sample Num: 22512, Cur Loss: 0.49751976, Cur Avg Loss: 0.57954275, Log Avg loss: 0.55861736, Global Avg Loss: 2.55762918, Time: 0.0209 Steps: 16310, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001417, Sample Num: 22672, Cur Loss: 0.39961696, Cur Avg Loss: 0.58003945, Log Avg loss: 0.64992606, Global Avg Loss: 2.55646024, Time: 0.0209 Steps: 16320, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001427, Sample Num: 22832, Cur Loss: 0.39686525, Cur Avg Loss: 0.57932458, Log Avg loss: 0.47802670, Global Avg Loss: 2.55518747, Time: 0.0209 Steps: 16330, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001437, Sample Num: 22992, Cur Loss: 0.43986696, Cur Avg Loss: 0.58075165, Log Avg loss: 0.78439524, Global Avg Loss: 2.55410376, Time: 0.0209 Steps: 16340, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001447, Sample Num: 23152, Cur Loss: 0.77774942, Cur Avg Loss: 0.58025519, Log Avg loss: 0.50891283, Global Avg Loss: 2.55285288, Time: 0.0209 Steps: 16350, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001457, Sample Num: 23312, Cur Loss: 0.78676081, Cur Avg Loss: 0.58029836, Log Avg loss: 0.58654568, Global Avg Loss: 2.55165098, Time: 0.0209 Steps: 16360, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001467, Sample Num: 23472, Cur Loss: 0.97094721, Cur Avg Loss: 0.58288144, Log Avg loss: 0.95923642, Global Avg Loss: 2.55067821, Time: 0.0209 Steps: 16370, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001477, Sample Num: 23632, Cur Loss: 1.52637720, Cur Avg Loss: 0.58349772, Log Avg loss: 0.67390562, Global Avg Loss: 2.54953244, Time: 0.0209 Steps: 16380, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001487, Sample Num: 23792, Cur Loss: 0.65721059, Cur Avg Loss: 0.58248857, Log Avg loss: 0.43343699, Global Avg Loss: 2.54824135, Time: 0.0209 Steps: 16390, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001497, Sample Num: 23952, Cur Loss: 0.38512242, Cur Avg Loss: 0.58226511, Log Avg loss: 0.54903735, Global Avg Loss: 2.54702233, Time: 0.0209 Steps: 16400, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001507, Sample Num: 24112, Cur Loss: 0.26371163, Cur Avg Loss: 0.58092395, Log Avg loss: 0.38015136, Global Avg Loss: 2.54570187, Time: 0.0209 Steps: 16410, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001517, Sample Num: 24272, Cur Loss: 0.55765629, Cur Avg Loss: 0.58043900, Log Avg loss: 0.50735756, Global Avg Loss: 2.54446049, Time: 0.0209 Steps: 16420, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001527, Sample Num: 24432, Cur Loss: 0.55814981, Cur Avg Loss: 0.57957460, Log Avg loss: 0.44844477, Global Avg Loss: 2.54318476, Time: 0.0208 Steps: 16430, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001537, Sample Num: 24592, Cur Loss: 0.25995591, Cur Avg Loss: 0.58005961, Log Avg loss: 0.65412125, Global Avg Loss: 2.54203570, Time: 0.0243 Steps: 16440, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001547, Sample Num: 24752, Cur Loss: 0.61211896, Cur Avg Loss: 0.58000424, Log Avg loss: 0.57149402, Global Avg Loss: 2.54083780, Time: 0.0207 Steps: 16450, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001557, Sample Num: 24912, Cur Loss: 0.48765737, Cur Avg Loss: 0.58038508, Log Avg loss: 0.63930117, Global Avg Loss: 2.53968255, Time: 0.0208 Steps: 16460, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001567, Sample Num: 25072, Cur Loss: 0.68105131, Cur Avg Loss: 0.57927728, Log Avg loss: 0.40679190, Global Avg Loss: 2.53838754, Time: 0.0208 Steps: 16470, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001577, Sample Num: 25232, Cur Loss: 0.31719649, Cur Avg Loss: 0.57861272, Log Avg loss: 0.47447626, Global Avg Loss: 2.53713517, Time: 0.0208 Steps: 16480, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001587, Sample Num: 25392, Cur Loss: 0.44557053, Cur Avg Loss: 0.57772076, Log Avg loss: 0.43705887, Global Avg Loss: 2.53586162, Time: 0.0207 Steps: 16490, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001597, Sample Num: 25552, Cur Loss: 0.39432159, Cur Avg Loss: 0.57673454, Log Avg loss: 0.42022185, Global Avg Loss: 2.53457941, Time: 0.0209 Steps: 16500, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001607, Sample Num: 25712, Cur Loss: 0.22278184, Cur Avg Loss: 0.57698526, Log Avg loss: 0.61702491, Global Avg Loss: 2.53341796, Time: 0.0209 Steps: 16510, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001617, Sample Num: 25872, Cur Loss: 0.64526021, Cur Avg Loss: 0.57579610, Log Avg loss: 0.38469748, Global Avg Loss: 2.53211729, Time: 0.0209 Steps: 16520, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001627, Sample Num: 26032, Cur Loss: 0.18786611, Cur Avg Loss: 0.57519947, Log Avg loss: 0.47872452, Global Avg Loss: 2.53087506, Time: 0.0209 Steps: 16530, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001637, Sample Num: 26192, Cur Loss: 0.56002116, Cur Avg Loss: 0.57480035, Log Avg loss: 0.50986430, Global Avg Loss: 2.52965317, Time: 0.0209 Steps: 16540, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001647, Sample Num: 26352, Cur Loss: 0.56590176, Cur Avg Loss: 0.57570289, Log Avg loss: 0.72344892, Global Avg Loss: 2.52856181, Time: 0.0209 Steps: 16550, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001657, Sample Num: 26512, Cur Loss: 0.73587734, Cur Avg Loss: 0.57534509, Log Avg loss: 0.51641430, Global Avg Loss: 2.52734674, Time: 0.0209 Steps: 16560, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001667, Sample Num: 26672, Cur Loss: 0.20000198, Cur Avg Loss: 0.57505300, Log Avg loss: 0.52665427, Global Avg Loss: 2.52613933, Time: 0.0209 Steps: 16570, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001677, Sample Num: 26832, Cur Loss: 0.78319412, Cur Avg Loss: 0.57499509, Log Avg loss: 0.56534195, Global Avg Loss: 2.52495670, Time: 0.0209 Steps: 16580, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001687, Sample Num: 26992, Cur Loss: 0.49385804, Cur Avg Loss: 0.57457259, Log Avg loss: 0.50371880, Global Avg Loss: 2.52373835, Time: 0.0209 Steps: 16590, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001697, Sample Num: 27152, Cur Loss: 0.59528172, Cur Avg Loss: 0.57337828, Log Avg loss: 0.37189852, Global Avg Loss: 2.52244206, Time: 0.0210 Steps: 16600, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001707, Sample Num: 27312, Cur Loss: 0.41844007, Cur Avg Loss: 0.57458873, Log Avg loss: 0.78000172, Global Avg Loss: 2.52139303, Time: 0.0209 Steps: 16610, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001717, Sample Num: 27472, Cur Loss: 0.38840598, Cur Avg Loss: 0.57412389, Log Avg loss: 0.49477649, Global Avg Loss: 2.52017365, Time: 0.0209 Steps: 16620, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001727, Sample Num: 27632, Cur Loss: 1.01287735, Cur Avg Loss: 0.57567752, Log Avg loss: 0.84243537, Global Avg Loss: 2.51916478, Time: 0.0209 Steps: 16630, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001737, Sample Num: 27792, Cur Loss: 0.87839746, Cur Avg Loss: 0.57684673, Log Avg loss: 0.77876827, Global Avg Loss: 2.51811887, Time: 0.0209 Steps: 16640, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001747, Sample Num: 27952, Cur Loss: 0.94716907, Cur Avg Loss: 0.57765375, Log Avg loss: 0.71783380, Global Avg Loss: 2.51703762, Time: 0.0209 Steps: 16650, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001757, Sample Num: 28112, Cur Loss: 0.48201531, Cur Avg Loss: 0.57739927, Log Avg loss: 0.53294253, Global Avg Loss: 2.51584669, Time: 0.0209 Steps: 16660, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001767, Sample Num: 28272, Cur Loss: 1.41219318, Cur Avg Loss: 0.57795839, Log Avg loss: 0.67619480, Global Avg Loss: 2.51474312, Time: 0.0209 Steps: 16670, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001777, Sample Num: 28432, Cur Loss: 0.24716830, Cur Avg Loss: 0.57823059, Log Avg loss: 0.62632864, Global Avg Loss: 2.51361097, Time: 0.0209 Steps: 16680, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001787, Sample Num: 28592, Cur Loss: 1.25811875, Cur Avg Loss: 0.57865433, Log Avg loss: 0.65395330, Global Avg Loss: 2.51249674, Time: 0.0209 Steps: 16690, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001797, Sample Num: 28752, Cur Loss: 2.18027401, Cur Avg Loss: 0.58227017, Log Avg loss: 1.22841999, Global Avg Loss: 2.51172783, Time: 0.0209 Steps: 16700, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001807, Sample Num: 28912, Cur Loss: 0.96610087, Cur Avg Loss: 0.58328431, Log Avg loss: 0.76552514, Global Avg Loss: 2.51068283, Time: 0.0209 Steps: 16710, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001817, Sample Num: 29072, Cur Loss: 0.19872713, Cur Avg Loss: 0.58347483, Log Avg loss: 0.61790265, Global Avg Loss: 2.50955078, Time: 0.0209 Steps: 16720, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001827, Sample Num: 29232, Cur Loss: 0.36179599, Cur Avg Loss: 0.58349756, Log Avg loss: 0.58762792, Global Avg Loss: 2.50840199, Time: 0.0210 Steps: 16730, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001837, Sample Num: 29392, Cur Loss: 0.36219984, Cur Avg Loss: 0.58315856, Log Avg loss: 0.52122312, Global Avg Loss: 2.50721491, Time: 0.0209 Steps: 16740, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001847, Sample Num: 29552, Cur Loss: 0.54098606, Cur Avg Loss: 0.58293388, Log Avg loss: 0.54166026, Global Avg Loss: 2.50604144, Time: 0.0209 Steps: 16750, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001857, Sample Num: 29712, Cur Loss: 0.91735005, Cur Avg Loss: 0.58337950, Log Avg loss: 0.66568415, Global Avg Loss: 2.50494338, Time: 0.0209 Steps: 16760, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001867, Sample Num: 29872, Cur Loss: 0.36103931, Cur Avg Loss: 0.58313104, Log Avg loss: 0.53699225, Global Avg Loss: 2.50376988, Time: 0.0209 Steps: 16770, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001877, Sample Num: 30032, Cur Loss: 0.48390108, Cur Avg Loss: 0.58223324, Log Avg loss: 0.41461366, Global Avg Loss: 2.50252485, Time: 0.0209 Steps: 16780, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001887, Sample Num: 30192, Cur Loss: 0.47348842, Cur Avg Loss: 0.58198268, Log Avg loss: 0.53495306, Global Avg Loss: 2.50135298, Time: 0.0208 Steps: 16790, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001897, Sample Num: 30352, Cur Loss: 0.91412902, Cur Avg Loss: 0.58133991, Log Avg loss: 0.46004989, Global Avg Loss: 2.50013792, Time: 0.0209 Steps: 16800, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001907, Sample Num: 30512, Cur Loss: 0.44913298, Cur Avg Loss: 0.58122107, Log Avg loss: 0.55867623, Global Avg Loss: 2.49898298, Time: 0.0211 Steps: 16810, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001917, Sample Num: 30672, Cur Loss: 0.14606845, Cur Avg Loss: 0.58089295, Log Avg loss: 0.51832076, Global Avg Loss: 2.49780541, Time: 0.0210 Steps: 16820, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001927, Sample Num: 30832, Cur Loss: 0.63021290, Cur Avg Loss: 0.58044095, Log Avg loss: 0.49379357, Global Avg Loss: 2.49661468, Time: 0.0210 Steps: 16830, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001937, Sample Num: 30992, Cur Loss: 0.68588471, Cur Avg Loss: 0.58043628, Log Avg loss: 0.57953536, Global Avg Loss: 2.49547627, Time: 0.0210 Steps: 16840, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001947, Sample Num: 31152, Cur Loss: 0.71989679, Cur Avg Loss: 0.58105852, Log Avg loss: 0.70158682, Global Avg Loss: 2.49441164, Time: 0.0209 Steps: 16850, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001957, Sample Num: 31312, Cur Loss: 0.51210320, Cur Avg Loss: 0.58074545, Log Avg loss: 0.51978958, Global Avg Loss: 2.49324046, Time: 0.0210 Steps: 16860, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001967, Sample Num: 31472, Cur Loss: 0.85574496, Cur Avg Loss: 0.58104165, Log Avg loss: 0.63900958, Global Avg Loss: 2.49214133, Time: 0.0210 Steps: 16870, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001977, Sample Num: 31632, Cur Loss: 0.54130590, Cur Avg Loss: 0.58048692, Log Avg loss: 0.47137179, Global Avg Loss: 2.49094419, Time: 0.0211 Steps: 16880, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001987, Sample Num: 31792, Cur Loss: 0.82567942, Cur Avg Loss: 0.58086401, Log Avg loss: 0.65541332, Global Avg Loss: 2.48985743, Time: 0.0210 Steps: 16890, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001997, Sample Num: 31952, Cur Loss: 0.63924307, Cur Avg Loss: 0.58142174, Log Avg loss: 0.69224377, Global Avg Loss: 2.48879376, Time: 0.0210 Steps: 16900, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002007, Sample Num: 32112, Cur Loss: 0.26327926, Cur Avg Loss: 0.58065539, Log Avg loss: 0.42761459, Global Avg Loss: 2.48757485, Time: 0.0210 Steps: 16910, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002017, Sample Num: 32272, Cur Loss: 0.34826440, Cur Avg Loss: 0.57982005, Log Avg loss: 0.41216746, Global Avg Loss: 2.48634825, Time: 0.0209 Steps: 16920, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002027, Sample Num: 32432, Cur Loss: 0.75798702, Cur Avg Loss: 0.57968025, Log Avg loss: 0.55148249, Global Avg Loss: 2.48520538, Time: 0.0210 Steps: 16930, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002037, Sample Num: 32592, Cur Loss: 0.49761492, Cur Avg Loss: 0.57907520, Log Avg loss: 0.45643186, Global Avg Loss: 2.48400776, Time: 0.0209 Steps: 16940, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002047, Sample Num: 32752, Cur Loss: 0.75511253, Cur Avg Loss: 0.57913002, Log Avg loss: 0.59029756, Global Avg Loss: 2.48289053, Time: 0.0209 Steps: 16950, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002057, Sample Num: 32912, Cur Loss: 0.32107842, Cur Avg Loss: 0.57864028, Log Avg loss: 0.47838856, Global Avg Loss: 2.48170863, Time: 0.0210 Steps: 16960, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002067, Sample Num: 33072, Cur Loss: 0.56612933, Cur Avg Loss: 0.57831242, Log Avg loss: 0.51087170, Global Avg Loss: 2.48054726, Time: 0.0211 Steps: 16970, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002077, Sample Num: 33232, Cur Loss: 0.25257251, Cur Avg Loss: 0.57720401, Log Avg loss: 0.34809681, Global Avg Loss: 2.47929140, Time: 0.0210 Steps: 16980, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002087, Sample Num: 33392, Cur Loss: 0.11050048, Cur Avg Loss: 0.57636265, Log Avg loss: 0.40161208, Global Avg Loss: 2.47806852, Time: 0.0211 Steps: 16990, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002097, Sample Num: 33552, Cur Loss: 0.37437537, Cur Avg Loss: 0.57624559, Log Avg loss: 0.55181575, Global Avg Loss: 2.47693543, Time: 0.0210 Steps: 17000, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002107, Sample Num: 33712, Cur Loss: 1.07049131, Cur Avg Loss: 0.57667831, Log Avg loss: 0.66741958, Global Avg Loss: 2.47587163, Time: 0.0210 Steps: 17010, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002117, Sample Num: 33872, Cur Loss: 0.65143275, Cur Avg Loss: 0.57666651, Log Avg loss: 0.57417906, Global Avg Loss: 2.47475430, Time: 0.0210 Steps: 17020, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002127, Sample Num: 34032, Cur Loss: 0.27971736, Cur Avg Loss: 0.57604192, Log Avg loss: 0.44381697, Global Avg Loss: 2.47356174, Time: 0.0211 Steps: 17030, Updated lr: 0.000085 ***** Running evaluation checkpoint-17032 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-17032 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.520652, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.510922, "eval_total_loss": 359.178195, "eval_mae": 0.500104, "eval_mse": 0.511089, "eval_r2": 0.675118, "eval_sp_statistic": 0.809728, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.835233, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.821434, "test_total_loss": 412.36002, "test_mae": 0.555588, "test_mse": 0.821727, "test_r2": 0.46965, "test_sp_statistic": 0.741635, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.791046, "test_ps_pvalue": 0.0, "lr": 8.479658605974395e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.4733086615136375, "train_cur_epoch_loss": 1225.8778510838747, "train_cur_epoch_avg_loss": 0.5757998361126702, "train_cur_epoch_time": 44.520652294158936, "train_cur_epoch_avg_time": 0.020911532312897576, "epoch": 8, "step": 17032} ################################################## Training, Epoch: 0009, Batch: 000008, Sample Num: 128, Cur Loss: 0.32413831, Cur Avg Loss: 0.46306157, Log Avg loss: 0.43411776, Global Avg Loss: 2.47236488, Time: 0.0209 Steps: 17040, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000018, Sample Num: 288, Cur Loss: 0.30516556, Cur Avg Loss: 0.46802547, Log Avg loss: 0.47199658, Global Avg Loss: 2.47119165, Time: 0.0209 Steps: 17050, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000028, Sample Num: 448, Cur Loss: 0.45424154, Cur Avg Loss: 0.48265032, Log Avg loss: 0.50897507, Global Avg Loss: 2.47004146, Time: 0.0209 Steps: 17060, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000038, Sample Num: 608, Cur Loss: 0.38176563, Cur Avg Loss: 0.46336809, Log Avg loss: 0.40937785, Global Avg Loss: 2.46883428, Time: 0.0208 Steps: 17070, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000048, Sample Num: 768, Cur Loss: 1.13505816, Cur Avg Loss: 0.48660881, Log Avg loss: 0.57492352, Global Avg Loss: 2.46772543, Time: 0.0209 Steps: 17080, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000058, Sample Num: 928, Cur Loss: 0.15666395, Cur Avg Loss: 0.47730929, Log Avg loss: 0.43267161, Global Avg Loss: 2.46653464, Time: 0.0209 Steps: 17090, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000068, Sample Num: 1088, Cur Loss: 0.48100466, Cur Avg Loss: 0.46782270, Log Avg loss: 0.41280046, Global Avg Loss: 2.46533363, Time: 0.0209 Steps: 17100, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000078, Sample Num: 1248, Cur Loss: 0.26712245, Cur Avg Loss: 0.47537439, Log Avg loss: 0.52672590, Global Avg Loss: 2.46420060, Time: 0.0209 Steps: 17110, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000088, Sample Num: 1408, Cur Loss: 0.86920524, Cur Avg Loss: 0.49330568, Log Avg loss: 0.63316978, Global Avg Loss: 2.46313108, Time: 0.0208 Steps: 17120, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000098, Sample Num: 1568, Cur Loss: 0.37038839, Cur Avg Loss: 0.49081498, Log Avg loss: 0.46889682, Global Avg Loss: 2.46196690, Time: 0.0209 Steps: 17130, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000108, Sample Num: 1728, Cur Loss: 0.52642202, Cur Avg Loss: 0.48568171, Log Avg loss: 0.43537565, Global Avg Loss: 2.46078452, Time: 0.0209 Steps: 17140, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000118, Sample Num: 1888, Cur Loss: 0.65129000, Cur Avg Loss: 0.48138559, Log Avg loss: 0.43498747, Global Avg Loss: 2.45960330, Time: 0.0209 Steps: 17150, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000128, Sample Num: 2048, Cur Loss: 0.68854457, Cur Avg Loss: 0.48068382, Log Avg loss: 0.47240296, Global Avg Loss: 2.45844526, Time: 0.0209 Steps: 17160, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000138, Sample Num: 2208, Cur Loss: 0.57122278, Cur Avg Loss: 0.47767580, Log Avg loss: 0.43917311, Global Avg Loss: 2.45726921, Time: 0.0208 Steps: 17170, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000148, Sample Num: 2368, Cur Loss: 0.38683686, Cur Avg Loss: 0.47354123, Log Avg loss: 0.41648414, Global Avg Loss: 2.45608133, Time: 0.0208 Steps: 17180, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000158, Sample Num: 2528, Cur Loss: 0.27522373, Cur Avg Loss: 0.47439828, Log Avg loss: 0.48708257, Global Avg Loss: 2.45493590, Time: 0.0208 Steps: 17190, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000168, Sample Num: 2688, Cur Loss: 0.27476844, Cur Avg Loss: 0.48708083, Log Avg loss: 0.68746526, Global Avg Loss: 2.45390830, Time: 0.0208 Steps: 17200, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000178, Sample Num: 2848, Cur Loss: 1.32972527, Cur Avg Loss: 0.49488417, Log Avg loss: 0.62598026, Global Avg Loss: 2.45284617, Time: 0.0209 Steps: 17210, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000188, Sample Num: 3008, Cur Loss: 0.35526854, Cur Avg Loss: 0.49591616, Log Avg loss: 0.51428547, Global Avg Loss: 2.45172040, Time: 0.0209 Steps: 17220, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000198, Sample Num: 3168, Cur Loss: 0.91953921, Cur Avg Loss: 0.49688736, Log Avg loss: 0.51514597, Global Avg Loss: 2.45059645, Time: 0.0208 Steps: 17230, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000208, Sample Num: 3328, Cur Loss: 0.74942291, Cur Avg Loss: 0.50515362, Log Avg loss: 0.66882550, Global Avg Loss: 2.44956294, Time: 0.0209 Steps: 17240, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000218, Sample Num: 3488, Cur Loss: 0.61902523, Cur Avg Loss: 0.51447204, Log Avg loss: 0.70829530, Global Avg Loss: 2.44855351, Time: 0.0209 Steps: 17250, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000228, Sample Num: 3648, Cur Loss: 0.71091127, Cur Avg Loss: 0.51157179, Log Avg loss: 0.44834641, Global Avg Loss: 2.44739464, Time: 0.0209 Steps: 17260, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000238, Sample Num: 3808, Cur Loss: 0.31696618, Cur Avg Loss: 0.51563753, Log Avg loss: 0.60833623, Global Avg Loss: 2.44632975, Time: 0.0209 Steps: 17270, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000248, Sample Num: 3968, Cur Loss: 0.85406131, Cur Avg Loss: 0.51887897, Log Avg loss: 0.59602542, Global Avg Loss: 2.44525898, Time: 0.0209 Steps: 17280, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000258, Sample Num: 4128, Cur Loss: 0.64908600, Cur Avg Loss: 0.51969722, Log Avg loss: 0.53998967, Global Avg Loss: 2.44415703, Time: 0.0248 Steps: 17290, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000268, Sample Num: 4288, Cur Loss: 0.55875093, Cur Avg Loss: 0.52443799, Log Avg loss: 0.64675001, Global Avg Loss: 2.44311806, Time: 0.0210 Steps: 17300, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000278, Sample Num: 4448, Cur Loss: 0.54331726, Cur Avg Loss: 0.52809520, Log Avg loss: 0.62610820, Global Avg Loss: 2.44206838, Time: 0.0210 Steps: 17310, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000288, Sample Num: 4608, Cur Loss: 0.44309682, Cur Avg Loss: 0.52888708, Log Avg loss: 0.55090150, Global Avg Loss: 2.44097648, Time: 0.0210 Steps: 17320, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000298, Sample Num: 4768, Cur Loss: 0.25442192, Cur Avg Loss: 0.53264735, Log Avg loss: 0.64094312, Global Avg Loss: 2.43993780, Time: 0.0210 Steps: 17330, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000308, Sample Num: 4928, Cur Loss: 0.13073206, Cur Avg Loss: 0.53223662, Log Avg loss: 0.51999673, Global Avg Loss: 2.43883057, Time: 0.0210 Steps: 17340, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000318, Sample Num: 5088, Cur Loss: 0.57788849, Cur Avg Loss: 0.53035892, Log Avg loss: 0.47252585, Global Avg Loss: 2.43769725, Time: 0.0210 Steps: 17350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000328, Sample Num: 5248, Cur Loss: 0.21251509, Cur Avg Loss: 0.53111235, Log Avg loss: 0.55507147, Global Avg Loss: 2.43661279, Time: 0.0210 Steps: 17360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000338, Sample Num: 5408, Cur Loss: 1.43540609, Cur Avg Loss: 0.54143825, Log Avg loss: 0.88012776, Global Avg Loss: 2.43571671, Time: 0.0210 Steps: 17370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000348, Sample Num: 5568, Cur Loss: 0.36857364, Cur Avg Loss: 0.54184069, Log Avg loss: 0.55544317, Global Avg Loss: 2.43463485, Time: 0.0210 Steps: 17380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000358, Sample Num: 5728, Cur Loss: 0.32366392, Cur Avg Loss: 0.53766279, Log Avg loss: 0.39227174, Global Avg Loss: 2.43346040, Time: 0.0210 Steps: 17390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000368, Sample Num: 5888, Cur Loss: 0.35188898, Cur Avg Loss: 0.53486996, Log Avg loss: 0.43488677, Global Avg Loss: 2.43231180, Time: 0.0210 Steps: 17400, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000378, Sample Num: 6048, Cur Loss: 0.23296246, Cur Avg Loss: 0.53226979, Log Avg loss: 0.43658347, Global Avg Loss: 2.43116549, Time: 0.0210 Steps: 17410, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000388, Sample Num: 6208, Cur Loss: 0.64221704, Cur Avg Loss: 0.53118054, Log Avg loss: 0.49000680, Global Avg Loss: 2.43005116, Time: 0.0210 Steps: 17420, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000398, Sample Num: 6368, Cur Loss: 0.12157273, Cur Avg Loss: 0.53354710, Log Avg loss: 0.62536969, Global Avg Loss: 2.42901577, Time: 0.0210 Steps: 17430, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000408, Sample Num: 6528, Cur Loss: 0.35469517, Cur Avg Loss: 0.53510814, Log Avg loss: 0.59723755, Global Avg Loss: 2.42796544, Time: 0.0210 Steps: 17440, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000418, Sample Num: 6688, Cur Loss: 0.52823704, Cur Avg Loss: 0.53451737, Log Avg loss: 0.51041399, Global Avg Loss: 2.42686655, Time: 0.0210 Steps: 17450, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000428, Sample Num: 6848, Cur Loss: 0.25561514, Cur Avg Loss: 0.53448782, Log Avg loss: 0.53325248, Global Avg Loss: 2.42578201, Time: 0.0210 Steps: 17460, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000438, Sample Num: 7008, Cur Loss: 0.23817767, Cur Avg Loss: 0.53305829, Log Avg loss: 0.47187455, Global Avg Loss: 2.42466357, Time: 0.0210 Steps: 17470, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000448, Sample Num: 7168, Cur Loss: 0.11464978, Cur Avg Loss: 0.52946756, Log Avg loss: 0.37219371, Global Avg Loss: 2.42348939, Time: 0.0210 Steps: 17480, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000458, Sample Num: 7328, Cur Loss: 0.64691347, Cur Avg Loss: 0.53013451, Log Avg loss: 0.56001393, Global Avg Loss: 2.42242394, Time: 0.0210 Steps: 17490, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000468, Sample Num: 7488, Cur Loss: 0.35047233, Cur Avg Loss: 0.52791290, Log Avg loss: 0.42616312, Global Avg Loss: 2.42128322, Time: 0.0210 Steps: 17500, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000478, Sample Num: 7648, Cur Loss: 0.74301910, Cur Avg Loss: 0.53278784, Log Avg loss: 0.76093503, Global Avg Loss: 2.42033499, Time: 0.0209 Steps: 17510, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000488, Sample Num: 7808, Cur Loss: 0.39688715, Cur Avg Loss: 0.53210354, Log Avg loss: 0.49939395, Global Avg Loss: 2.41923856, Time: 0.0210 Steps: 17520, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000498, Sample Num: 7968, Cur Loss: 0.46097705, Cur Avg Loss: 0.53288053, Log Avg loss: 0.57079760, Global Avg Loss: 2.41818412, Time: 0.0209 Steps: 17530, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000508, Sample Num: 8128, Cur Loss: 1.46583951, Cur Avg Loss: 0.53732665, Log Avg loss: 0.75874345, Global Avg Loss: 2.41723803, Time: 0.0210 Steps: 17540, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000518, Sample Num: 8288, Cur Loss: 0.25597820, Cur Avg Loss: 0.54178749, Log Avg loss: 0.76839816, Global Avg Loss: 2.41629852, Time: 0.0248 Steps: 17550, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000528, Sample Num: 8448, Cur Loss: 0.45732015, Cur Avg Loss: 0.53991453, Log Avg loss: 0.44289502, Global Avg Loss: 2.41517471, Time: 0.0211 Steps: 17560, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000538, Sample Num: 8608, Cur Loss: 0.14541027, Cur Avg Loss: 0.53644186, Log Avg loss: 0.35308500, Global Avg Loss: 2.41400107, Time: 0.0211 Steps: 17570, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000548, Sample Num: 8768, Cur Loss: 1.09821916, Cur Avg Loss: 0.54138515, Log Avg loss: 0.80733411, Global Avg Loss: 2.41308715, Time: 0.0210 Steps: 17580, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000558, Sample Num: 8928, Cur Loss: 0.58186924, Cur Avg Loss: 0.54338159, Log Avg loss: 0.65278662, Global Avg Loss: 2.41208642, Time: 0.0211 Steps: 17590, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000568, Sample Num: 9088, Cur Loss: 0.39482033, Cur Avg Loss: 0.53984607, Log Avg loss: 0.34256405, Global Avg Loss: 2.41091055, Time: 0.0211 Steps: 17600, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000578, Sample Num: 9248, Cur Loss: 0.29903942, Cur Avg Loss: 0.53738780, Log Avg loss: 0.39775795, Global Avg Loss: 2.40976736, Time: 0.0211 Steps: 17610, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000588, Sample Num: 9408, Cur Loss: 0.39635670, Cur Avg Loss: 0.53406416, Log Avg loss: 0.34195771, Global Avg Loss: 2.40859381, Time: 0.0211 Steps: 17620, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000598, Sample Num: 9568, Cur Loss: 1.19182658, Cur Avg Loss: 0.53414758, Log Avg loss: 0.53905276, Global Avg Loss: 2.40753337, Time: 0.0211 Steps: 17630, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000608, Sample Num: 9728, Cur Loss: 0.27208522, Cur Avg Loss: 0.53348188, Log Avg loss: 0.49367323, Global Avg Loss: 2.40644842, Time: 0.0211 Steps: 17640, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000618, Sample Num: 9888, Cur Loss: 0.58784270, Cur Avg Loss: 0.53276067, Log Avg loss: 0.48891077, Global Avg Loss: 2.40536200, Time: 0.0211 Steps: 17650, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000628, Sample Num: 10048, Cur Loss: 0.50860107, Cur Avg Loss: 0.53380474, Log Avg loss: 0.59832851, Global Avg Loss: 2.40433876, Time: 0.0211 Steps: 17660, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000638, Sample Num: 10208, Cur Loss: 0.56387836, Cur Avg Loss: 0.53341855, Log Avg loss: 0.50916586, Global Avg Loss: 2.40326622, Time: 0.0211 Steps: 17670, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000648, Sample Num: 10368, Cur Loss: 0.18637586, Cur Avg Loss: 0.53357309, Log Avg loss: 0.54343264, Global Avg Loss: 2.40221428, Time: 0.0211 Steps: 17680, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000658, Sample Num: 10528, Cur Loss: 0.33572632, Cur Avg Loss: 0.53203203, Log Avg loss: 0.43217120, Global Avg Loss: 2.40110063, Time: 0.0211 Steps: 17690, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000668, Sample Num: 10688, Cur Loss: 0.93119514, Cur Avg Loss: 0.53230311, Log Avg loss: 0.55014039, Global Avg Loss: 2.40005489, Time: 0.0211 Steps: 17700, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000678, Sample Num: 10848, Cur Loss: 0.15984815, Cur Avg Loss: 0.53367888, Log Avg loss: 0.62558026, Global Avg Loss: 2.39905293, Time: 0.0211 Steps: 17710, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000688, Sample Num: 11008, Cur Loss: 0.37700182, Cur Avg Loss: 0.53212452, Log Avg loss: 0.42673862, Global Avg Loss: 2.39793989, Time: 0.0211 Steps: 17720, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000698, Sample Num: 11168, Cur Loss: 0.77142727, Cur Avg Loss: 0.53148881, Log Avg loss: 0.48775211, Global Avg Loss: 2.39686251, Time: 0.0211 Steps: 17730, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000708, Sample Num: 11328, Cur Loss: 0.18734424, Cur Avg Loss: 0.53052118, Log Avg loss: 0.46298051, Global Avg Loss: 2.39577239, Time: 0.0211 Steps: 17740, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000718, Sample Num: 11488, Cur Loss: 0.18352801, Cur Avg Loss: 0.52981712, Log Avg loss: 0.47996974, Global Avg Loss: 2.39469306, Time: 0.0211 Steps: 17750, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000728, Sample Num: 11648, Cur Loss: 0.60655504, Cur Avg Loss: 0.52951496, Log Avg loss: 0.50781990, Global Avg Loss: 2.39363063, Time: 0.0211 Steps: 17760, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000738, Sample Num: 11808, Cur Loss: 0.14850858, Cur Avg Loss: 0.52800793, Log Avg loss: 0.41829647, Global Avg Loss: 2.39251902, Time: 0.0211 Steps: 17770, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000748, Sample Num: 11968, Cur Loss: 0.40652418, Cur Avg Loss: 0.53056731, Log Avg loss: 0.71944945, Global Avg Loss: 2.39157804, Time: 0.0210 Steps: 17780, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000758, Sample Num: 12128, Cur Loss: 0.52334076, Cur Avg Loss: 0.53126213, Log Avg loss: 0.58323458, Global Avg Loss: 2.39056154, Time: 0.0211 Steps: 17790, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000768, Sample Num: 12288, Cur Loss: 0.27787718, Cur Avg Loss: 0.52979578, Log Avg loss: 0.41864626, Global Avg Loss: 2.38945372, Time: 0.0256 Steps: 17800, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000778, Sample Num: 12448, Cur Loss: 0.18662524, Cur Avg Loss: 0.52771292, Log Avg loss: 0.36774964, Global Avg Loss: 2.38831857, Time: 0.0210 Steps: 17810, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000788, Sample Num: 12608, Cur Loss: 0.52981865, Cur Avg Loss: 0.52604552, Log Avg loss: 0.39632128, Global Avg Loss: 2.38720073, Time: 0.0210 Steps: 17820, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000798, Sample Num: 12768, Cur Loss: 0.51229942, Cur Avg Loss: 0.52502169, Log Avg loss: 0.44434460, Global Avg Loss: 2.38611107, Time: 0.0210 Steps: 17830, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000808, Sample Num: 12928, Cur Loss: 0.46359369, Cur Avg Loss: 0.52653685, Log Avg loss: 0.64744655, Global Avg Loss: 2.38513649, Time: 0.0210 Steps: 17840, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000818, Sample Num: 13088, Cur Loss: 0.28006917, Cur Avg Loss: 0.52528023, Log Avg loss: 0.42374464, Global Avg Loss: 2.38403767, Time: 0.0210 Steps: 17850, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000828, Sample Num: 13248, Cur Loss: 0.81499803, Cur Avg Loss: 0.52371712, Log Avg loss: 0.39585484, Global Avg Loss: 2.38292446, Time: 0.0210 Steps: 17860, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000838, Sample Num: 13408, Cur Loss: 0.55527216, Cur Avg Loss: 0.52451873, Log Avg loss: 0.59089190, Global Avg Loss: 2.38192165, Time: 0.0210 Steps: 17870, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000848, Sample Num: 13568, Cur Loss: 0.24415356, Cur Avg Loss: 0.52356738, Log Avg loss: 0.44384473, Global Avg Loss: 2.38083771, Time: 0.0210 Steps: 17880, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000858, Sample Num: 13728, Cur Loss: 0.29706842, Cur Avg Loss: 0.52252546, Log Avg loss: 0.43417019, Global Avg Loss: 2.37974958, Time: 0.0210 Steps: 17890, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000868, Sample Num: 13888, Cur Loss: 0.46662888, Cur Avg Loss: 0.52217160, Log Avg loss: 0.49181081, Global Avg Loss: 2.37869486, Time: 0.0210 Steps: 17900, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000878, Sample Num: 14048, Cur Loss: 0.51598370, Cur Avg Loss: 0.52261736, Log Avg loss: 0.56130896, Global Avg Loss: 2.37768013, Time: 0.0209 Steps: 17910, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000888, Sample Num: 14208, Cur Loss: 0.55570352, Cur Avg Loss: 0.52326030, Log Avg loss: 0.57971120, Global Avg Loss: 2.37667680, Time: 0.0210 Steps: 17920, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000898, Sample Num: 14368, Cur Loss: 0.34278217, Cur Avg Loss: 0.52314002, Log Avg loss: 0.51245888, Global Avg Loss: 2.37563708, Time: 0.0210 Steps: 17930, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000908, Sample Num: 14528, Cur Loss: 0.65037674, Cur Avg Loss: 0.52279482, Log Avg loss: 0.49179538, Global Avg Loss: 2.37458700, Time: 0.0210 Steps: 17940, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000918, Sample Num: 14688, Cur Loss: 0.45933896, Cur Avg Loss: 0.52132676, Log Avg loss: 0.38802770, Global Avg Loss: 2.37348028, Time: 0.0210 Steps: 17950, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000928, Sample Num: 14848, Cur Loss: 0.63431776, Cur Avg Loss: 0.52203045, Log Avg loss: 0.58662888, Global Avg Loss: 2.37248538, Time: 0.0210 Steps: 17960, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000938, Sample Num: 15008, Cur Loss: 0.86698729, Cur Avg Loss: 0.52340963, Log Avg loss: 0.65139781, Global Avg Loss: 2.37152762, Time: 0.0210 Steps: 17970, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000948, Sample Num: 15168, Cur Loss: 0.34438041, Cur Avg Loss: 0.52356830, Log Avg loss: 0.53845126, Global Avg Loss: 2.37050811, Time: 0.0210 Steps: 17980, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000958, Sample Num: 15328, Cur Loss: 0.69670337, Cur Avg Loss: 0.52466899, Log Avg loss: 0.62901433, Global Avg Loss: 2.36954008, Time: 0.0209 Steps: 17990, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000968, Sample Num: 15488, Cur Loss: 0.34984529, Cur Avg Loss: 0.52433468, Log Avg loss: 0.49230740, Global Avg Loss: 2.36849717, Time: 0.0210 Steps: 18000, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000978, Sample Num: 15648, Cur Loss: 0.48051256, Cur Avg Loss: 0.52346886, Log Avg loss: 0.43965789, Global Avg Loss: 2.36742619, Time: 0.0210 Steps: 18010, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000988, Sample Num: 15808, Cur Loss: 0.45407915, Cur Avg Loss: 0.52384218, Log Avg loss: 0.56035281, Global Avg Loss: 2.36642337, Time: 0.0210 Steps: 18020, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000998, Sample Num: 15968, Cur Loss: 0.80620766, Cur Avg Loss: 0.52388974, Log Avg loss: 0.52858889, Global Avg Loss: 2.36540405, Time: 0.0209 Steps: 18030, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001008, Sample Num: 16128, Cur Loss: 0.59162509, Cur Avg Loss: 0.52308086, Log Avg loss: 0.44235438, Global Avg Loss: 2.36433806, Time: 0.0209 Steps: 18040, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001018, Sample Num: 16288, Cur Loss: 0.09702696, Cur Avg Loss: 0.52182961, Log Avg loss: 0.39570314, Global Avg Loss: 2.36324741, Time: 0.0209 Steps: 18050, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001028, Sample Num: 16448, Cur Loss: 0.41520667, Cur Avg Loss: 0.52227578, Log Avg loss: 0.56769645, Global Avg Loss: 2.36225319, Time: 0.0246 Steps: 18060, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001038, Sample Num: 16608, Cur Loss: 1.32834840, Cur Avg Loss: 0.52239368, Log Avg loss: 0.53451368, Global Avg Loss: 2.36124171, Time: 0.0209 Steps: 18070, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001048, Sample Num: 16768, Cur Loss: 0.13731599, Cur Avg Loss: 0.52258582, Log Avg loss: 0.54253046, Global Avg Loss: 2.36023579, Time: 0.0209 Steps: 18080, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001058, Sample Num: 16928, Cur Loss: 0.37923706, Cur Avg Loss: 0.52136475, Log Avg loss: 0.39339594, Global Avg Loss: 2.35914854, Time: 0.0210 Steps: 18090, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001068, Sample Num: 17088, Cur Loss: 0.47313982, Cur Avg Loss: 0.52280045, Log Avg loss: 0.67469802, Global Avg Loss: 2.35821790, Time: 0.0209 Steps: 18100, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001078, Sample Num: 17248, Cur Loss: 0.38708335, Cur Avg Loss: 0.52361629, Log Avg loss: 0.61074771, Global Avg Loss: 2.35725298, Time: 0.0210 Steps: 18110, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001088, Sample Num: 17408, Cur Loss: 0.37472051, Cur Avg Loss: 0.52249500, Log Avg loss: 0.40161995, Global Avg Loss: 2.35617371, Time: 0.0209 Steps: 18120, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001098, Sample Num: 17568, Cur Loss: 0.39395621, Cur Avg Loss: 0.52161405, Log Avg loss: 0.42576653, Global Avg Loss: 2.35510895, Time: 0.0210 Steps: 18130, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001108, Sample Num: 17728, Cur Loss: 0.39861125, Cur Avg Loss: 0.52157766, Log Avg loss: 0.51758228, Global Avg Loss: 2.35409599, Time: 0.0210 Steps: 18140, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001118, Sample Num: 17888, Cur Loss: 0.41792417, Cur Avg Loss: 0.52023257, Log Avg loss: 0.37119632, Global Avg Loss: 2.35300348, Time: 0.0211 Steps: 18150, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001128, Sample Num: 18048, Cur Loss: 0.58215141, Cur Avg Loss: 0.51931813, Log Avg loss: 0.41708447, Global Avg Loss: 2.35193744, Time: 0.0209 Steps: 18160, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001138, Sample Num: 18208, Cur Loss: 0.71412063, Cur Avg Loss: 0.52009135, Log Avg loss: 0.60731018, Global Avg Loss: 2.35097727, Time: 0.0210 Steps: 18170, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001148, Sample Num: 18368, Cur Loss: 0.52532804, Cur Avg Loss: 0.52018127, Log Avg loss: 0.53041373, Global Avg Loss: 2.34997586, Time: 0.0209 Steps: 18180, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001158, Sample Num: 18528, Cur Loss: 0.32296717, Cur Avg Loss: 0.51979045, Log Avg loss: 0.47492450, Global Avg Loss: 2.34894505, Time: 0.0210 Steps: 18190, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001168, Sample Num: 18688, Cur Loss: 0.24337080, Cur Avg Loss: 0.52019945, Log Avg loss: 0.56756209, Global Avg Loss: 2.34796627, Time: 0.0209 Steps: 18200, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001178, Sample Num: 18848, Cur Loss: 0.63604164, Cur Avg Loss: 0.51934305, Log Avg loss: 0.41931560, Global Avg Loss: 2.34690715, Time: 0.0209 Steps: 18210, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001188, Sample Num: 19008, Cur Loss: 0.25167218, Cur Avg Loss: 0.51939346, Log Avg loss: 0.52533120, Global Avg Loss: 2.34590738, Time: 0.0211 Steps: 18220, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001198, Sample Num: 19168, Cur Loss: 0.26921052, Cur Avg Loss: 0.51978889, Log Avg loss: 0.56676556, Global Avg Loss: 2.34493144, Time: 0.0209 Steps: 18230, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001208, Sample Num: 19328, Cur Loss: 0.39523673, Cur Avg Loss: 0.52012837, Log Avg loss: 0.56079829, Global Avg Loss: 2.34395330, Time: 0.0208 Steps: 18240, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001218, Sample Num: 19488, Cur Loss: 0.62776059, Cur Avg Loss: 0.52035247, Log Avg loss: 0.54742420, Global Avg Loss: 2.34296890, Time: 0.0209 Steps: 18250, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001228, Sample Num: 19648, Cur Loss: 0.55590826, Cur Avg Loss: 0.52096681, Log Avg loss: 0.59579372, Global Avg Loss: 2.34201207, Time: 0.0209 Steps: 18260, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001238, Sample Num: 19808, Cur Loss: 0.65253437, Cur Avg Loss: 0.52083889, Log Avg loss: 0.50512933, Global Avg Loss: 2.34100666, Time: 0.0210 Steps: 18270, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001248, Sample Num: 19968, Cur Loss: 0.92656338, Cur Avg Loss: 0.52221201, Log Avg loss: 0.69220453, Global Avg Loss: 2.34010469, Time: 0.0211 Steps: 18280, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001258, Sample Num: 20128, Cur Loss: 1.02378058, Cur Avg Loss: 0.52398516, Log Avg loss: 0.74527422, Global Avg Loss: 2.33923272, Time: 0.0210 Steps: 18290, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001268, Sample Num: 20288, Cur Loss: 0.51157504, Cur Avg Loss: 0.52330487, Log Avg loss: 0.43772433, Global Avg Loss: 2.33819364, Time: 0.0210 Steps: 18300, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001278, Sample Num: 20448, Cur Loss: 0.23485990, Cur Avg Loss: 0.52240126, Log Avg loss: 0.40782450, Global Avg Loss: 2.33713937, Time: 0.0210 Steps: 18310, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001288, Sample Num: 20608, Cur Loss: 0.52701676, Cur Avg Loss: 0.52228999, Log Avg loss: 0.50806883, Global Avg Loss: 2.33614097, Time: 0.0211 Steps: 18320, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001298, Sample Num: 20768, Cur Loss: 0.81303942, Cur Avg Loss: 0.52164197, Log Avg loss: 0.43817773, Global Avg Loss: 2.33510553, Time: 0.0210 Steps: 18330, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001308, Sample Num: 20928, Cur Loss: 0.64025366, Cur Avg Loss: 0.52132775, Log Avg loss: 0.48054110, Global Avg Loss: 2.33409432, Time: 0.0210 Steps: 18340, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001318, Sample Num: 21088, Cur Loss: 0.17504327, Cur Avg Loss: 0.52147092, Log Avg loss: 0.54019826, Global Avg Loss: 2.33311672, Time: 0.0210 Steps: 18350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001328, Sample Num: 21248, Cur Loss: 0.61362469, Cur Avg Loss: 0.52158210, Log Avg loss: 0.53623490, Global Avg Loss: 2.33213803, Time: 0.0210 Steps: 18360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001338, Sample Num: 21408, Cur Loss: 0.38241285, Cur Avg Loss: 0.52203587, Log Avg loss: 0.58229665, Global Avg Loss: 2.33118547, Time: 0.0209 Steps: 18370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001348, Sample Num: 21568, Cur Loss: 0.37351671, Cur Avg Loss: 0.52185954, Log Avg loss: 0.49826711, Global Avg Loss: 2.33018824, Time: 0.0209 Steps: 18380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001358, Sample Num: 21728, Cur Loss: 0.37574112, Cur Avg Loss: 0.52124650, Log Avg loss: 0.43860833, Global Avg Loss: 2.32915964, Time: 0.0210 Steps: 18390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001368, Sample Num: 21888, Cur Loss: 0.22787052, Cur Avg Loss: 0.52056066, Log Avg loss: 0.42742312, Global Avg Loss: 2.32812609, Time: 0.0210 Steps: 18400, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001378, Sample Num: 22048, Cur Loss: 0.71692765, Cur Avg Loss: 0.52100474, Log Avg loss: 0.58175576, Global Avg Loss: 2.32717749, Time: 0.0211 Steps: 18410, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001388, Sample Num: 22208, Cur Loss: 0.97677296, Cur Avg Loss: 0.52130859, Log Avg loss: 0.56317839, Global Avg Loss: 2.32621984, Time: 0.0210 Steps: 18420, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001398, Sample Num: 22368, Cur Loss: 0.25358331, Cur Avg Loss: 0.52149596, Log Avg loss: 0.54750340, Global Avg Loss: 2.32525472, Time: 0.0210 Steps: 18430, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001408, Sample Num: 22528, Cur Loss: 0.61929446, Cur Avg Loss: 0.52197742, Log Avg loss: 0.58928610, Global Avg Loss: 2.32431330, Time: 0.0210 Steps: 18440, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001418, Sample Num: 22688, Cur Loss: 0.15126367, Cur Avg Loss: 0.52207828, Log Avg loss: 0.53627869, Global Avg Loss: 2.32334418, Time: 0.0210 Steps: 18450, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001428, Sample Num: 22848, Cur Loss: 0.40568501, Cur Avg Loss: 0.52151491, Log Avg loss: 0.44162962, Global Avg Loss: 2.32232483, Time: 0.0210 Steps: 18460, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001438, Sample Num: 23008, Cur Loss: 0.92203504, Cur Avg Loss: 0.52095076, Log Avg loss: 0.44038951, Global Avg Loss: 2.32130592, Time: 0.0209 Steps: 18470, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001448, Sample Num: 23168, Cur Loss: 0.57798892, Cur Avg Loss: 0.52178241, Log Avg loss: 0.64137406, Global Avg Loss: 2.32039686, Time: 0.0210 Steps: 18480, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001458, Sample Num: 23328, Cur Loss: 0.92706990, Cur Avg Loss: 0.52250263, Log Avg loss: 0.62679051, Global Avg Loss: 2.31948091, Time: 0.0210 Steps: 18490, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001468, Sample Num: 23488, Cur Loss: 0.58047938, Cur Avg Loss: 0.52231891, Log Avg loss: 0.49553237, Global Avg Loss: 2.31849499, Time: 0.0210 Steps: 18500, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001478, Sample Num: 23648, Cur Loss: 0.43396381, Cur Avg Loss: 0.52214724, Log Avg loss: 0.49694635, Global Avg Loss: 2.31751090, Time: 0.0211 Steps: 18510, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001488, Sample Num: 23808, Cur Loss: 0.33368549, Cur Avg Loss: 0.52165151, Log Avg loss: 0.44838276, Global Avg Loss: 2.31650165, Time: 0.0210 Steps: 18520, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001498, Sample Num: 23968, Cur Loss: 0.38207558, Cur Avg Loss: 0.52265068, Log Avg loss: 0.67132735, Global Avg Loss: 2.31561381, Time: 0.0210 Steps: 18530, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001508, Sample Num: 24128, Cur Loss: 0.54846936, Cur Avg Loss: 0.52157324, Log Avg loss: 0.36017180, Global Avg Loss: 2.31455909, Time: 0.0210 Steps: 18540, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001518, Sample Num: 24288, Cur Loss: 0.20189682, Cur Avg Loss: 0.52184513, Log Avg loss: 0.56284633, Global Avg Loss: 2.31361477, Time: 0.0210 Steps: 18550, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001528, Sample Num: 24448, Cur Loss: 0.34155989, Cur Avg Loss: 0.52124673, Log Avg loss: 0.43040997, Global Avg Loss: 2.31260011, Time: 0.0210 Steps: 18560, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001538, Sample Num: 24608, Cur Loss: 0.30531362, Cur Avg Loss: 0.52082275, Log Avg loss: 0.45603845, Global Avg Loss: 2.31160035, Time: 0.0248 Steps: 18570, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001548, Sample Num: 24768, Cur Loss: 0.38606682, Cur Avg Loss: 0.52024740, Log Avg loss: 0.43175894, Global Avg Loss: 2.31058860, Time: 0.0210 Steps: 18580, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001558, Sample Num: 24928, Cur Loss: 0.77785301, Cur Avg Loss: 0.52058474, Log Avg loss: 0.57280383, Global Avg Loss: 2.30965380, Time: 0.0210 Steps: 18590, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001568, Sample Num: 25088, Cur Loss: 0.41952986, Cur Avg Loss: 0.51979296, Log Avg loss: 0.39643388, Global Avg Loss: 2.30862519, Time: 0.0210 Steps: 18600, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001578, Sample Num: 25248, Cur Loss: 1.04822123, Cur Avg Loss: 0.51944239, Log Avg loss: 0.46447421, Global Avg Loss: 2.30763424, Time: 0.0210 Steps: 18610, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001588, Sample Num: 25408, Cur Loss: 0.26506215, Cur Avg Loss: 0.51933380, Log Avg loss: 0.50219831, Global Avg Loss: 2.30666462, Time: 0.0210 Steps: 18620, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001598, Sample Num: 25568, Cur Loss: 0.39849547, Cur Avg Loss: 0.51847092, Log Avg loss: 0.38144508, Global Avg Loss: 2.30563122, Time: 0.0210 Steps: 18630, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001608, Sample Num: 25728, Cur Loss: 0.24386977, Cur Avg Loss: 0.51835264, Log Avg loss: 0.49945201, Global Avg Loss: 2.30466224, Time: 0.0210 Steps: 18640, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001618, Sample Num: 25888, Cur Loss: 1.56767428, Cur Avg Loss: 0.51892866, Log Avg loss: 0.61155262, Global Avg Loss: 2.30375441, Time: 0.0210 Steps: 18650, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001628, Sample Num: 26048, Cur Loss: 0.28076762, Cur Avg Loss: 0.51918268, Log Avg loss: 0.56028250, Global Avg Loss: 2.30282007, Time: 0.0210 Steps: 18660, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001638, Sample Num: 26208, Cur Loss: 0.36078638, Cur Avg Loss: 0.51887819, Log Avg loss: 0.46930741, Global Avg Loss: 2.30183801, Time: 0.0210 Steps: 18670, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001648, Sample Num: 26368, Cur Loss: 0.50802755, Cur Avg Loss: 0.51889189, Log Avg loss: 0.52113613, Global Avg Loss: 2.30088474, Time: 0.0210 Steps: 18680, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001658, Sample Num: 26528, Cur Loss: 0.17708489, Cur Avg Loss: 0.51817126, Log Avg loss: 0.39941157, Global Avg Loss: 2.29986737, Time: 0.0211 Steps: 18690, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001668, Sample Num: 26688, Cur Loss: 0.82318997, Cur Avg Loss: 0.51721626, Log Avg loss: 0.35887626, Global Avg Loss: 2.29882940, Time: 0.0210 Steps: 18700, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001678, Sample Num: 26848, Cur Loss: 0.16704012, Cur Avg Loss: 0.51665354, Log Avg loss: 0.42279269, Global Avg Loss: 2.29782671, Time: 0.0210 Steps: 18710, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001688, Sample Num: 27008, Cur Loss: 0.40662879, Cur Avg Loss: 0.51610777, Log Avg loss: 0.42452687, Global Avg Loss: 2.29682602, Time: 0.0210 Steps: 18720, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001698, Sample Num: 27168, Cur Loss: 0.15225074, Cur Avg Loss: 0.51556191, Log Avg loss: 0.42342101, Global Avg Loss: 2.29582580, Time: 0.0210 Steps: 18730, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001708, Sample Num: 27328, Cur Loss: 0.06785829, Cur Avg Loss: 0.51466775, Log Avg loss: 0.36283946, Global Avg Loss: 2.29479432, Time: 0.0210 Steps: 18740, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001718, Sample Num: 27488, Cur Loss: 0.60256606, Cur Avg Loss: 0.51454264, Log Avg loss: 0.49317309, Global Avg Loss: 2.29383346, Time: 0.0211 Steps: 18750, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001728, Sample Num: 27648, Cur Loss: 0.52064341, Cur Avg Loss: 0.51500300, Log Avg loss: 0.59409395, Global Avg Loss: 2.29292742, Time: 0.0210 Steps: 18760, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001738, Sample Num: 27808, Cur Loss: 0.59762037, Cur Avg Loss: 0.51498197, Log Avg loss: 0.51134762, Global Avg Loss: 2.29197825, Time: 0.0210 Steps: 18770, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001748, Sample Num: 27968, Cur Loss: 0.48820537, Cur Avg Loss: 0.51651911, Log Avg loss: 0.78367399, Global Avg Loss: 2.29117511, Time: 0.0210 Steps: 18780, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001758, Sample Num: 28128, Cur Loss: 0.54073042, Cur Avg Loss: 0.51608993, Log Avg loss: 0.44106891, Global Avg Loss: 2.29019049, Time: 0.0210 Steps: 18790, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001768, Sample Num: 28288, Cur Loss: 0.55990267, Cur Avg Loss: 0.51620384, Log Avg loss: 0.53623002, Global Avg Loss: 2.28925753, Time: 0.0210 Steps: 18800, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001778, Sample Num: 28448, Cur Loss: 0.25042599, Cur Avg Loss: 0.51622717, Log Avg loss: 0.52035107, Global Avg Loss: 2.28831712, Time: 0.0211 Steps: 18810, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001788, Sample Num: 28608, Cur Loss: 0.66810912, Cur Avg Loss: 0.51633288, Log Avg loss: 0.53512787, Global Avg Loss: 2.28738556, Time: 0.0210 Steps: 18820, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001798, Sample Num: 28768, Cur Loss: 0.89949715, Cur Avg Loss: 0.51662534, Log Avg loss: 0.56891749, Global Avg Loss: 2.28647294, Time: 0.0247 Steps: 18830, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001808, Sample Num: 28928, Cur Loss: 0.31839865, Cur Avg Loss: 0.51638670, Log Avg loss: 0.47347869, Global Avg Loss: 2.28551063, Time: 0.0209 Steps: 18840, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001818, Sample Num: 29088, Cur Loss: 0.89000767, Cur Avg Loss: 0.51700344, Log Avg loss: 0.62851085, Global Avg Loss: 2.28463158, Time: 0.0209 Steps: 18850, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001828, Sample Num: 29248, Cur Loss: 0.32720265, Cur Avg Loss: 0.51653949, Log Avg loss: 0.43219349, Global Avg Loss: 2.28364938, Time: 0.0210 Steps: 18860, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001838, Sample Num: 29408, Cur Loss: 0.70140868, Cur Avg Loss: 0.51553825, Log Avg loss: 0.33251090, Global Avg Loss: 2.28261539, Time: 0.0210 Steps: 18870, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001848, Sample Num: 29568, Cur Loss: 0.51765108, Cur Avg Loss: 0.51563083, Log Avg loss: 0.53264775, Global Avg Loss: 2.28168850, Time: 0.0209 Steps: 18880, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001858, Sample Num: 29728, Cur Loss: 0.25003767, Cur Avg Loss: 0.51536729, Log Avg loss: 0.46666561, Global Avg Loss: 2.28072766, Time: 0.0210 Steps: 18890, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001868, Sample Num: 29888, Cur Loss: 0.26464033, Cur Avg Loss: 0.51492523, Log Avg loss: 0.43278996, Global Avg Loss: 2.27974992, Time: 0.0210 Steps: 18900, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001878, Sample Num: 30048, Cur Loss: 0.13411632, Cur Avg Loss: 0.51486921, Log Avg loss: 0.50440463, Global Avg Loss: 2.27881108, Time: 0.0210 Steps: 18910, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001888, Sample Num: 30208, Cur Loss: 0.53865027, Cur Avg Loss: 0.51474098, Log Avg loss: 0.49065974, Global Avg Loss: 2.27786597, Time: 0.0210 Steps: 18920, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001898, Sample Num: 30368, Cur Loss: 0.24366403, Cur Avg Loss: 0.51454770, Log Avg loss: 0.47805627, Global Avg Loss: 2.27691520, Time: 0.0209 Steps: 18930, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001908, Sample Num: 30528, Cur Loss: 0.28687692, Cur Avg Loss: 0.51405861, Log Avg loss: 0.42122882, Global Avg Loss: 2.27593542, Time: 0.0210 Steps: 18940, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001918, Sample Num: 30688, Cur Loss: 0.91572195, Cur Avg Loss: 0.51462696, Log Avg loss: 0.62306839, Global Avg Loss: 2.27506320, Time: 0.0210 Steps: 18950, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001928, Sample Num: 30848, Cur Loss: 0.89662349, Cur Avg Loss: 0.51470129, Log Avg loss: 0.52895707, Global Avg Loss: 2.27414226, Time: 0.0210 Steps: 18960, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001938, Sample Num: 31008, Cur Loss: 0.63322246, Cur Avg Loss: 0.51528346, Log Avg loss: 0.62752654, Global Avg Loss: 2.27327425, Time: 0.0210 Steps: 18970, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001948, Sample Num: 31168, Cur Loss: 0.83208686, Cur Avg Loss: 0.51572184, Log Avg loss: 0.60068032, Global Avg Loss: 2.27239301, Time: 0.0209 Steps: 18980, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001958, Sample Num: 31328, Cur Loss: 0.22246459, Cur Avg Loss: 0.51532732, Log Avg loss: 0.43847409, Global Avg Loss: 2.27142728, Time: 0.0210 Steps: 18990, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001968, Sample Num: 31488, Cur Loss: 0.95180821, Cur Avg Loss: 0.51475483, Log Avg loss: 0.40266134, Global Avg Loss: 2.27044372, Time: 0.0210 Steps: 19000, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001978, Sample Num: 31648, Cur Loss: 0.43717483, Cur Avg Loss: 0.51553845, Log Avg loss: 0.66975461, Global Avg Loss: 2.26960169, Time: 0.0210 Steps: 19010, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001988, Sample Num: 31808, Cur Loss: 0.24833854, Cur Avg Loss: 0.51645592, Log Avg loss: 0.69793267, Global Avg Loss: 2.26877537, Time: 0.0210 Steps: 19020, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001998, Sample Num: 31968, Cur Loss: 0.30332869, Cur Avg Loss: 0.51618523, Log Avg loss: 0.46237089, Global Avg Loss: 2.26782613, Time: 0.0210 Steps: 19030, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002008, Sample Num: 32128, Cur Loss: 0.40540612, Cur Avg Loss: 0.51599398, Log Avg loss: 0.47778259, Global Avg Loss: 2.26688598, Time: 0.0210 Steps: 19040, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002018, Sample Num: 32288, Cur Loss: 0.77938628, Cur Avg Loss: 0.51534134, Log Avg loss: 0.38429101, Global Avg Loss: 2.26589774, Time: 0.0210 Steps: 19050, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002028, Sample Num: 32448, Cur Loss: 0.24904384, Cur Avg Loss: 0.51482112, Log Avg loss: 0.40984012, Global Avg Loss: 2.26492394, Time: 0.0209 Steps: 19060, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002038, Sample Num: 32608, Cur Loss: 0.40966922, Cur Avg Loss: 0.51467701, Log Avg loss: 0.48545299, Global Avg Loss: 2.26399082, Time: 0.0210 Steps: 19070, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002048, Sample Num: 32768, Cur Loss: 0.55638963, Cur Avg Loss: 0.51455112, Log Avg loss: 0.48889325, Global Avg Loss: 2.26306047, Time: 0.0255 Steps: 19080, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002058, Sample Num: 32928, Cur Loss: 1.00933707, Cur Avg Loss: 0.51424610, Log Avg loss: 0.45177843, Global Avg Loss: 2.26211166, Time: 0.0209 Steps: 19090, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002068, Sample Num: 33088, Cur Loss: 0.55584592, Cur Avg Loss: 0.51374038, Log Avg loss: 0.40966271, Global Avg Loss: 2.26114179, Time: 0.0209 Steps: 19100, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002078, Sample Num: 33248, Cur Loss: 0.37489074, Cur Avg Loss: 0.51382684, Log Avg loss: 0.53170768, Global Avg Loss: 2.26023680, Time: 0.0209 Steps: 19110, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002088, Sample Num: 33408, Cur Loss: 0.29707766, Cur Avg Loss: 0.51290222, Log Avg loss: 0.32076704, Global Avg Loss: 2.25922244, Time: 0.0209 Steps: 19120, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002098, Sample Num: 33568, Cur Loss: 0.27959251, Cur Avg Loss: 0.51258950, Log Avg loss: 0.44729256, Global Avg Loss: 2.25827527, Time: 0.0209 Steps: 19130, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002108, Sample Num: 33728, Cur Loss: 0.26025286, Cur Avg Loss: 0.51212091, Log Avg loss: 0.41381029, Global Avg Loss: 2.25731160, Time: 0.0209 Steps: 19140, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002118, Sample Num: 33888, Cur Loss: 0.50186354, Cur Avg Loss: 0.51153374, Log Avg loss: 0.38775981, Global Avg Loss: 2.25633533, Time: 0.0209 Steps: 19150, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002128, Sample Num: 34048, Cur Loss: 0.39755034, Cur Avg Loss: 0.51143877, Log Avg loss: 0.49132289, Global Avg Loss: 2.25541413, Time: 0.0209 Steps: 19160, Updated lr: 0.000083 ***** Running evaluation checkpoint-19161 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-19161 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.832618, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.456304, "eval_total_loss": 320.781975, "eval_mae": 0.468755, "eval_mse": 0.45648, "eval_r2": 0.709832, "eval_sp_statistic": 0.823185, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.849031, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.938875, "test_total_loss": 471.31538, "test_mae": 0.651855, "test_mse": 0.939161, "test_r2": 0.393857, "test_sp_statistic": 0.747118, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.7973, "test_ps_pvalue": 0.0, "lr": 8.277761972498815e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.255327107407062, "train_cur_epoch_loss": 1088.9295821264386, "train_cur_epoch_avg_loss": 0.5114746745544568, "train_cur_epoch_time": 44.83261847496033, "train_cur_epoch_avg_time": 0.02105806410284656, "epoch": 9, "step": 19161} ################################################## Training, Epoch: 0010, Batch: 000009, Sample Num: 144, Cur Loss: 0.04184922, Cur Avg Loss: 0.47977086, Log Avg loss: 0.49058217, Global Avg Loss: 2.25449351, Time: 0.0211 Steps: 19170, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000019, Sample Num: 304, Cur Loss: 0.71304107, Cur Avg Loss: 0.60755044, Log Avg loss: 0.72255206, Global Avg Loss: 2.25369479, Time: 0.0209 Steps: 19180, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000029, Sample Num: 464, Cur Loss: 0.50961202, Cur Avg Loss: 0.60432474, Log Avg loss: 0.59819589, Global Avg Loss: 2.25283211, Time: 0.0209 Steps: 19190, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000039, Sample Num: 624, Cur Loss: 0.12301583, Cur Avg Loss: 0.55948357, Log Avg loss: 0.42944421, Global Avg Loss: 2.25188243, Time: 0.0209 Steps: 19200, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000049, Sample Num: 784, Cur Loss: 0.19369411, Cur Avg Loss: 0.51365077, Log Avg loss: 0.33490283, Global Avg Loss: 2.25088452, Time: 0.0209 Steps: 19210, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000059, Sample Num: 944, Cur Loss: 0.53465581, Cur Avg Loss: 0.51567612, Log Avg loss: 0.52560036, Global Avg Loss: 2.24998687, Time: 0.0209 Steps: 19220, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000069, Sample Num: 1104, Cur Loss: 0.78265077, Cur Avg Loss: 0.51939178, Log Avg loss: 0.54131415, Global Avg Loss: 2.24909832, Time: 0.0209 Steps: 19230, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000079, Sample Num: 1264, Cur Loss: 0.25530130, Cur Avg Loss: 0.51760177, Log Avg loss: 0.50525070, Global Avg Loss: 2.24819196, Time: 0.0208 Steps: 19240, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000089, Sample Num: 1424, Cur Loss: 0.25690222, Cur Avg Loss: 0.49932442, Log Avg loss: 0.35493337, Global Avg Loss: 2.24720845, Time: 0.0209 Steps: 19250, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000099, Sample Num: 1584, Cur Loss: 0.52523625, Cur Avg Loss: 0.50964754, Log Avg loss: 0.60152332, Global Avg Loss: 2.24635399, Time: 0.0210 Steps: 19260, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000109, Sample Num: 1744, Cur Loss: 0.52009702, Cur Avg Loss: 0.49854325, Log Avg loss: 0.38861080, Global Avg Loss: 2.24538993, Time: 0.0209 Steps: 19270, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000119, Sample Num: 1904, Cur Loss: 0.32692686, Cur Avg Loss: 0.48359324, Log Avg loss: 0.32063811, Global Avg Loss: 2.24439161, Time: 0.0209 Steps: 19280, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000129, Sample Num: 2064, Cur Loss: 0.56686586, Cur Avg Loss: 0.47990700, Log Avg loss: 0.43604077, Global Avg Loss: 2.24345416, Time: 0.0209 Steps: 19290, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000139, Sample Num: 2224, Cur Loss: 0.28546432, Cur Avg Loss: 0.48161344, Log Avg loss: 0.50362641, Global Avg Loss: 2.24255269, Time: 0.0209 Steps: 19300, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000149, Sample Num: 2384, Cur Loss: 0.24214190, Cur Avg Loss: 0.47984001, Log Avg loss: 0.45518944, Global Avg Loss: 2.24162708, Time: 0.0209 Steps: 19310, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000159, Sample Num: 2544, Cur Loss: 0.83488894, Cur Avg Loss: 0.48500426, Log Avg loss: 0.56195156, Global Avg Loss: 2.24075768, Time: 0.0208 Steps: 19320, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000169, Sample Num: 2704, Cur Loss: 0.31351000, Cur Avg Loss: 0.48727412, Log Avg loss: 0.52336484, Global Avg Loss: 2.23986922, Time: 0.0209 Steps: 19330, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000179, Sample Num: 2864, Cur Loss: 0.30096620, Cur Avg Loss: 0.48529307, Log Avg loss: 0.45181344, Global Avg Loss: 2.23894468, Time: 0.0208 Steps: 19340, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000189, Sample Num: 3024, Cur Loss: 0.36320299, Cur Avg Loss: 0.48604267, Log Avg loss: 0.49946038, Global Avg Loss: 2.23804572, Time: 0.0209 Steps: 19350, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000199, Sample Num: 3184, Cur Loss: 0.26373237, Cur Avg Loss: 0.48751862, Log Avg loss: 0.51541416, Global Avg Loss: 2.23715594, Time: 0.0209 Steps: 19360, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000209, Sample Num: 3344, Cur Loss: 0.15532334, Cur Avg Loss: 0.48350819, Log Avg loss: 0.40370069, Global Avg Loss: 2.23620939, Time: 0.0209 Steps: 19370, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000219, Sample Num: 3504, Cur Loss: 0.85399652, Cur Avg Loss: 0.48603136, Log Avg loss: 0.53876554, Global Avg Loss: 2.23533352, Time: 0.0209 Steps: 19380, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000229, Sample Num: 3664, Cur Loss: 0.20760612, Cur Avg Loss: 0.48297606, Log Avg loss: 0.41606501, Global Avg Loss: 2.23439527, Time: 0.0208 Steps: 19390, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000239, Sample Num: 3824, Cur Loss: 0.37229440, Cur Avg Loss: 0.47813105, Log Avg loss: 0.36718035, Global Avg Loss: 2.23343278, Time: 0.0209 Steps: 19400, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000249, Sample Num: 3984, Cur Loss: 0.61235309, Cur Avg Loss: 0.48455049, Log Avg loss: 0.63797497, Global Avg Loss: 2.23261081, Time: 0.0208 Steps: 19410, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000259, Sample Num: 4144, Cur Loss: 0.19670837, Cur Avg Loss: 0.48490828, Log Avg loss: 0.49381729, Global Avg Loss: 2.23171545, Time: 0.0248 Steps: 19420, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000269, Sample Num: 4304, Cur Loss: 0.16760898, Cur Avg Loss: 0.48765014, Log Avg loss: 0.55866440, Global Avg Loss: 2.23085438, Time: 0.0210 Steps: 19430, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000279, Sample Num: 4464, Cur Loss: 0.22542045, Cur Avg Loss: 0.48616987, Log Avg loss: 0.44635054, Global Avg Loss: 2.22993642, Time: 0.0210 Steps: 19440, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000289, Sample Num: 4624, Cur Loss: 0.52943844, Cur Avg Loss: 0.48379465, Log Avg loss: 0.41752597, Global Avg Loss: 2.22900459, Time: 0.0211 Steps: 19450, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000299, Sample Num: 4784, Cur Loss: 0.17775841, Cur Avg Loss: 0.48124076, Log Avg loss: 0.40743331, Global Avg Loss: 2.22806854, Time: 0.0211 Steps: 19460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000309, Sample Num: 4944, Cur Loss: 0.26475707, Cur Avg Loss: 0.48369135, Log Avg loss: 0.55696397, Global Avg Loss: 2.22721024, Time: 0.0210 Steps: 19470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000319, Sample Num: 5104, Cur Loss: 0.18975183, Cur Avg Loss: 0.48395832, Log Avg loss: 0.49220774, Global Avg Loss: 2.22631958, Time: 0.0210 Steps: 19480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000329, Sample Num: 5264, Cur Loss: 0.21191572, Cur Avg Loss: 0.48575928, Log Avg loss: 0.54320980, Global Avg Loss: 2.22545600, Time: 0.0211 Steps: 19490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000339, Sample Num: 5424, Cur Loss: 0.20587651, Cur Avg Loss: 0.48742320, Log Avg loss: 0.54216618, Global Avg Loss: 2.22459278, Time: 0.0211 Steps: 19500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000349, Sample Num: 5584, Cur Loss: 0.40219879, Cur Avg Loss: 0.49013972, Log Avg loss: 0.58222987, Global Avg Loss: 2.22375097, Time: 0.0210 Steps: 19510, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000359, Sample Num: 5744, Cur Loss: 0.63131070, Cur Avg Loss: 0.48688864, Log Avg loss: 0.37342608, Global Avg Loss: 2.22280306, Time: 0.0210 Steps: 19520, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000369, Sample Num: 5904, Cur Loss: 0.54463559, Cur Avg Loss: 0.48813751, Log Avg loss: 0.53297175, Global Avg Loss: 2.22193781, Time: 0.0211 Steps: 19530, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000379, Sample Num: 6064, Cur Loss: 0.36210793, Cur Avg Loss: 0.49172798, Log Avg loss: 0.62421633, Global Avg Loss: 2.22112014, Time: 0.0210 Steps: 19540, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000389, Sample Num: 6224, Cur Loss: 0.40251893, Cur Avg Loss: 0.49436222, Log Avg loss: 0.59419980, Global Avg Loss: 2.22028796, Time: 0.0210 Steps: 19550, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000399, Sample Num: 6384, Cur Loss: 0.55377507, Cur Avg Loss: 0.50069169, Log Avg loss: 0.74690833, Global Avg Loss: 2.21953470, Time: 0.0211 Steps: 19560, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000409, Sample Num: 6544, Cur Loss: 0.55572754, Cur Avg Loss: 0.50285751, Log Avg loss: 0.58927383, Global Avg Loss: 2.21870166, Time: 0.0210 Steps: 19570, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000419, Sample Num: 6704, Cur Loss: 0.36063182, Cur Avg Loss: 0.50027958, Log Avg loss: 0.39484212, Global Avg Loss: 2.21777017, Time: 0.0211 Steps: 19580, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000429, Sample Num: 6864, Cur Loss: 0.11310996, Cur Avg Loss: 0.50028591, Log Avg loss: 0.50055116, Global Avg Loss: 2.21689359, Time: 0.0210 Steps: 19590, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000439, Sample Num: 7024, Cur Loss: 0.62907732, Cur Avg Loss: 0.50278699, Log Avg loss: 0.61008334, Global Avg Loss: 2.21607379, Time: 0.0210 Steps: 19600, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000449, Sample Num: 7184, Cur Loss: 0.64215612, Cur Avg Loss: 0.50167215, Log Avg loss: 0.45273055, Global Avg Loss: 2.21517458, Time: 0.0210 Steps: 19610, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000459, Sample Num: 7344, Cur Loss: 0.17190157, Cur Avg Loss: 0.49902998, Log Avg loss: 0.38039675, Global Avg Loss: 2.21423942, Time: 0.0210 Steps: 19620, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000469, Sample Num: 7504, Cur Loss: 0.60963762, Cur Avg Loss: 0.49728982, Log Avg loss: 0.41741648, Global Avg Loss: 2.21332408, Time: 0.0211 Steps: 19630, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000479, Sample Num: 7664, Cur Loss: 0.29309222, Cur Avg Loss: 0.49743285, Log Avg loss: 0.50414076, Global Avg Loss: 2.21245382, Time: 0.0210 Steps: 19640, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000489, Sample Num: 7824, Cur Loss: 0.36292249, Cur Avg Loss: 0.49376633, Log Avg loss: 0.31814025, Global Avg Loss: 2.21148979, Time: 0.0210 Steps: 19650, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000499, Sample Num: 7984, Cur Loss: 0.21824622, Cur Avg Loss: 0.49148739, Log Avg loss: 0.38004714, Global Avg Loss: 2.21055824, Time: 0.0210 Steps: 19660, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000509, Sample Num: 8144, Cur Loss: 1.12146854, Cur Avg Loss: 0.49087874, Log Avg loss: 0.46050677, Global Avg Loss: 2.20966853, Time: 0.0210 Steps: 19670, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000519, Sample Num: 8304, Cur Loss: 0.29219961, Cur Avg Loss: 0.48857125, Log Avg loss: 0.37112045, Global Avg Loss: 2.20873431, Time: 0.0211 Steps: 19680, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000529, Sample Num: 8464, Cur Loss: 0.94834101, Cur Avg Loss: 0.48879506, Log Avg loss: 0.50041044, Global Avg Loss: 2.20786670, Time: 0.0210 Steps: 19690, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000539, Sample Num: 8624, Cur Loss: 0.31374705, Cur Avg Loss: 0.48859961, Log Avg loss: 0.47826025, Global Avg Loss: 2.20698873, Time: 0.0209 Steps: 19700, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000549, Sample Num: 8784, Cur Loss: 0.30506465, Cur Avg Loss: 0.48883601, Log Avg loss: 0.50157840, Global Avg Loss: 2.20612347, Time: 0.0209 Steps: 19710, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000559, Sample Num: 8944, Cur Loss: 0.42183495, Cur Avg Loss: 0.49086252, Log Avg loss: 0.60211761, Global Avg Loss: 2.20531008, Time: 0.0210 Steps: 19720, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000569, Sample Num: 9104, Cur Loss: 1.13257718, Cur Avg Loss: 0.49106775, Log Avg loss: 0.50253989, Global Avg Loss: 2.20444705, Time: 0.0209 Steps: 19730, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000579, Sample Num: 9264, Cur Loss: 1.04641998, Cur Avg Loss: 0.49207270, Log Avg loss: 0.54925466, Global Avg Loss: 2.20360855, Time: 0.0209 Steps: 19740, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000589, Sample Num: 9424, Cur Loss: 0.78190029, Cur Avg Loss: 0.49370942, Log Avg loss: 0.58847550, Global Avg Loss: 2.20279076, Time: 0.0210 Steps: 19750, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000599, Sample Num: 9584, Cur Loss: 0.54237747, Cur Avg Loss: 0.49192885, Log Avg loss: 0.38705338, Global Avg Loss: 2.20187187, Time: 0.0210 Steps: 19760, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000609, Sample Num: 9744, Cur Loss: 0.73601437, Cur Avg Loss: 0.48956279, Log Avg loss: 0.34783569, Global Avg Loss: 2.20093406, Time: 0.0210 Steps: 19770, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000619, Sample Num: 9904, Cur Loss: 0.63615692, Cur Avg Loss: 0.48861921, Log Avg loss: 0.43115523, Global Avg Loss: 2.20003933, Time: 0.0210 Steps: 19780, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000629, Sample Num: 10064, Cur Loss: 0.42299777, Cur Avg Loss: 0.48969410, Log Avg loss: 0.55622989, Global Avg Loss: 2.19920871, Time: 0.0210 Steps: 19790, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000639, Sample Num: 10224, Cur Loss: 0.38241568, Cur Avg Loss: 0.48941194, Log Avg loss: 0.47166363, Global Avg Loss: 2.19833621, Time: 0.0210 Steps: 19800, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000649, Sample Num: 10384, Cur Loss: 0.64073622, Cur Avg Loss: 0.49097555, Log Avg loss: 0.59089080, Global Avg Loss: 2.19752478, Time: 0.0209 Steps: 19810, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000659, Sample Num: 10544, Cur Loss: 0.30281204, Cur Avg Loss: 0.49026596, Log Avg loss: 0.44421307, Global Avg Loss: 2.19664016, Time: 0.0209 Steps: 19820, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000669, Sample Num: 10704, Cur Loss: 0.99766827, Cur Avg Loss: 0.49079056, Log Avg loss: 0.52536201, Global Avg Loss: 2.19579736, Time: 0.0209 Steps: 19830, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000679, Sample Num: 10864, Cur Loss: 0.23949316, Cur Avg Loss: 0.49102908, Log Avg loss: 0.50698589, Global Avg Loss: 2.19494614, Time: 0.0210 Steps: 19840, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000689, Sample Num: 11024, Cur Loss: 0.54584050, Cur Avg Loss: 0.48968222, Log Avg loss: 0.39823060, Global Avg Loss: 2.19404100, Time: 0.0209 Steps: 19850, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000699, Sample Num: 11184, Cur Loss: 0.40350956, Cur Avg Loss: 0.48757928, Log Avg loss: 0.34268650, Global Avg Loss: 2.19310879, Time: 0.0209 Steps: 19860, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000709, Sample Num: 11344, Cur Loss: 0.17685217, Cur Avg Loss: 0.48578449, Log Avg loss: 0.36032883, Global Avg Loss: 2.19218641, Time: 0.0209 Steps: 19870, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000719, Sample Num: 11504, Cur Loss: 0.43657026, Cur Avg Loss: 0.48542020, Log Avg loss: 0.45959235, Global Avg Loss: 2.19131488, Time: 0.0209 Steps: 19880, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000729, Sample Num: 11664, Cur Loss: 0.23932789, Cur Avg Loss: 0.48539409, Log Avg loss: 0.48351677, Global Avg Loss: 2.19045626, Time: 0.0210 Steps: 19890, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000739, Sample Num: 11824, Cur Loss: 0.75123274, Cur Avg Loss: 0.48559515, Log Avg loss: 0.50025243, Global Avg Loss: 2.18960691, Time: 0.0209 Steps: 19900, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000749, Sample Num: 11984, Cur Loss: 0.42498827, Cur Avg Loss: 0.48411252, Log Avg loss: 0.37454603, Global Avg Loss: 2.18869528, Time: 0.0210 Steps: 19910, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000759, Sample Num: 12144, Cur Loss: 0.24123384, Cur Avg Loss: 0.48195498, Log Avg loss: 0.32035532, Global Avg Loss: 2.18775736, Time: 0.0209 Steps: 19920, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000769, Sample Num: 12304, Cur Loss: 0.47195065, Cur Avg Loss: 0.48189275, Log Avg loss: 0.47716925, Global Avg Loss: 2.18689906, Time: 0.0248 Steps: 19930, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000779, Sample Num: 12464, Cur Loss: 0.52521867, Cur Avg Loss: 0.48261119, Log Avg loss: 0.53785915, Global Avg Loss: 2.18607206, Time: 0.0211 Steps: 19940, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000789, Sample Num: 12624, Cur Loss: 0.49212289, Cur Avg Loss: 0.48404125, Log Avg loss: 0.59544271, Global Avg Loss: 2.18527475, Time: 0.0211 Steps: 19950, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000799, Sample Num: 12784, Cur Loss: 0.21088648, Cur Avg Loss: 0.48268237, Log Avg loss: 0.37546691, Global Avg Loss: 2.18436803, Time: 0.0210 Steps: 19960, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000809, Sample Num: 12944, Cur Loss: 0.27060831, Cur Avg Loss: 0.47985425, Log Avg loss: 0.25388774, Global Avg Loss: 2.18340134, Time: 0.0210 Steps: 19970, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000819, Sample Num: 13104, Cur Loss: 0.25908062, Cur Avg Loss: 0.47985297, Log Avg loss: 0.47974938, Global Avg Loss: 2.18254866, Time: 0.0210 Steps: 19980, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000829, Sample Num: 13264, Cur Loss: 0.51976037, Cur Avg Loss: 0.48020550, Log Avg loss: 0.50907764, Global Avg Loss: 2.18171151, Time: 0.0210 Steps: 19990, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000839, Sample Num: 13424, Cur Loss: 0.68558592, Cur Avg Loss: 0.48136538, Log Avg loss: 0.57751945, Global Avg Loss: 2.18090941, Time: 0.0210 Steps: 20000, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000849, Sample Num: 13584, Cur Loss: 0.72607672, Cur Avg Loss: 0.48359533, Log Avg loss: 0.67068796, Global Avg Loss: 2.18015468, Time: 0.0211 Steps: 20010, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000859, Sample Num: 13744, Cur Loss: 0.38664532, Cur Avg Loss: 0.48274810, Log Avg loss: 0.41081874, Global Avg Loss: 2.17927090, Time: 0.0211 Steps: 20020, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000869, Sample Num: 13904, Cur Loss: 0.52765667, Cur Avg Loss: 0.48315636, Log Avg loss: 0.51822582, Global Avg Loss: 2.17844162, Time: 0.0210 Steps: 20030, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000879, Sample Num: 14064, Cur Loss: 0.34924972, Cur Avg Loss: 0.48218274, Log Avg loss: 0.39757497, Global Avg Loss: 2.17755296, Time: 0.0211 Steps: 20040, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000889, Sample Num: 14224, Cur Loss: 0.22041485, Cur Avg Loss: 0.48234077, Log Avg loss: 0.49623181, Global Avg Loss: 2.17671440, Time: 0.0211 Steps: 20050, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000899, Sample Num: 14384, Cur Loss: 0.54447228, Cur Avg Loss: 0.48300073, Log Avg loss: 0.54167110, Global Avg Loss: 2.17589932, Time: 0.0211 Steps: 20060, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000909, Sample Num: 14544, Cur Loss: 0.45818618, Cur Avg Loss: 0.48414438, Log Avg loss: 0.58695815, Global Avg Loss: 2.17510762, Time: 0.0211 Steps: 20070, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000919, Sample Num: 14704, Cur Loss: 0.75623339, Cur Avg Loss: 0.48666299, Log Avg loss: 0.71560516, Global Avg Loss: 2.17438078, Time: 0.0211 Steps: 20080, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000929, Sample Num: 14864, Cur Loss: 0.82705623, Cur Avg Loss: 0.48749246, Log Avg loss: 0.56372074, Global Avg Loss: 2.17357905, Time: 0.0210 Steps: 20090, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000939, Sample Num: 15024, Cur Loss: 0.40551537, Cur Avg Loss: 0.48632607, Log Avg loss: 0.37796781, Global Avg Loss: 2.17268572, Time: 0.0210 Steps: 20100, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000949, Sample Num: 15184, Cur Loss: 0.24862151, Cur Avg Loss: 0.48586357, Log Avg loss: 0.44243504, Global Avg Loss: 2.17182532, Time: 0.0210 Steps: 20110, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000959, Sample Num: 15344, Cur Loss: 0.38738826, Cur Avg Loss: 0.48583327, Log Avg loss: 0.48295819, Global Avg Loss: 2.17098593, Time: 0.0210 Steps: 20120, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000969, Sample Num: 15504, Cur Loss: 0.14722702, Cur Avg Loss: 0.48442381, Log Avg loss: 0.34925599, Global Avg Loss: 2.17008094, Time: 0.0210 Steps: 20130, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000979, Sample Num: 15664, Cur Loss: 0.93030238, Cur Avg Loss: 0.48529116, Log Avg loss: 0.56933716, Global Avg Loss: 2.16928613, Time: 0.0211 Steps: 20140, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000989, Sample Num: 15824, Cur Loss: 0.86826724, Cur Avg Loss: 0.48569022, Log Avg loss: 0.52475904, Global Avg Loss: 2.16846999, Time: 0.0210 Steps: 20150, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000999, Sample Num: 15984, Cur Loss: 0.39114499, Cur Avg Loss: 0.48820279, Log Avg loss: 0.73669589, Global Avg Loss: 2.16775979, Time: 0.0210 Steps: 20160, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001009, Sample Num: 16144, Cur Loss: 0.61049902, Cur Avg Loss: 0.48808948, Log Avg loss: 0.47676909, Global Avg Loss: 2.16692142, Time: 0.0211 Steps: 20170, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001019, Sample Num: 16304, Cur Loss: 0.26407605, Cur Avg Loss: 0.48774780, Log Avg loss: 0.45327262, Global Avg Loss: 2.16607224, Time: 0.0210 Steps: 20180, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001029, Sample Num: 16464, Cur Loss: 0.82130903, Cur Avg Loss: 0.48778499, Log Avg loss: 0.49157499, Global Avg Loss: 2.16524287, Time: 0.0211 Steps: 20190, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001039, Sample Num: 16624, Cur Loss: 1.27840281, Cur Avg Loss: 0.48787704, Log Avg loss: 0.49734893, Global Avg Loss: 2.16441718, Time: 0.0210 Steps: 20200, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001049, Sample Num: 16784, Cur Loss: 0.37849289, Cur Avg Loss: 0.48802189, Log Avg loss: 0.50307120, Global Avg Loss: 2.16359513, Time: 0.0210 Steps: 20210, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001059, Sample Num: 16944, Cur Loss: 0.78040445, Cur Avg Loss: 0.48825967, Log Avg loss: 0.51320270, Global Avg Loss: 2.16277892, Time: 0.0210 Steps: 20220, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001069, Sample Num: 17104, Cur Loss: 0.17163879, Cur Avg Loss: 0.48952705, Log Avg loss: 0.62374338, Global Avg Loss: 2.16201815, Time: 0.0210 Steps: 20230, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001079, Sample Num: 17264, Cur Loss: 0.73248017, Cur Avg Loss: 0.48857509, Log Avg loss: 0.38681046, Global Avg Loss: 2.16114107, Time: 0.0210 Steps: 20240, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001089, Sample Num: 17424, Cur Loss: 0.53396785, Cur Avg Loss: 0.48907426, Log Avg loss: 0.54293475, Global Avg Loss: 2.16034195, Time: 0.0210 Steps: 20250, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001099, Sample Num: 17584, Cur Loss: 0.21645647, Cur Avg Loss: 0.48803143, Log Avg loss: 0.37446737, Global Avg Loss: 2.15946048, Time: 0.0210 Steps: 20260, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001109, Sample Num: 17744, Cur Loss: 0.28922677, Cur Avg Loss: 0.48813618, Log Avg loss: 0.49964800, Global Avg Loss: 2.15864162, Time: 0.0210 Steps: 20270, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001119, Sample Num: 17904, Cur Loss: 0.49290410, Cur Avg Loss: 0.48805293, Log Avg loss: 0.47881999, Global Avg Loss: 2.15781331, Time: 0.0210 Steps: 20280, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001129, Sample Num: 18064, Cur Loss: 0.58295798, Cur Avg Loss: 0.48807812, Log Avg loss: 0.49089684, Global Avg Loss: 2.15699176, Time: 0.0210 Steps: 20290, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001139, Sample Num: 18224, Cur Loss: 0.34147313, Cur Avg Loss: 0.48899065, Log Avg loss: 0.59201577, Global Avg Loss: 2.15622084, Time: 0.0210 Steps: 20300, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001149, Sample Num: 18384, Cur Loss: 0.32653397, Cur Avg Loss: 0.48838000, Log Avg loss: 0.41882690, Global Avg Loss: 2.15536540, Time: 0.0210 Steps: 20310, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001159, Sample Num: 18544, Cur Loss: 0.49287295, Cur Avg Loss: 0.48797793, Log Avg loss: 0.44178004, Global Avg Loss: 2.15452210, Time: 0.0210 Steps: 20320, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001169, Sample Num: 18704, Cur Loss: 0.75701499, Cur Avg Loss: 0.48772506, Log Avg loss: 0.45841784, Global Avg Loss: 2.15368782, Time: 0.0210 Steps: 20330, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001179, Sample Num: 18864, Cur Loss: 0.40967584, Cur Avg Loss: 0.48757902, Log Avg loss: 0.47050640, Global Avg Loss: 2.15286029, Time: 0.0210 Steps: 20340, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001189, Sample Num: 19024, Cur Loss: 0.53324813, Cur Avg Loss: 0.48867499, Log Avg loss: 0.61789003, Global Avg Loss: 2.15210601, Time: 0.0210 Steps: 20350, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001199, Sample Num: 19184, Cur Loss: 0.34953132, Cur Avg Loss: 0.48813026, Log Avg loss: 0.42336155, Global Avg Loss: 2.15125692, Time: 0.0210 Steps: 20360, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001209, Sample Num: 19344, Cur Loss: 0.48027331, Cur Avg Loss: 0.48742064, Log Avg loss: 0.40233778, Global Avg Loss: 2.15039834, Time: 0.0210 Steps: 20370, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001219, Sample Num: 19504, Cur Loss: 1.01705122, Cur Avg Loss: 0.48830272, Log Avg loss: 0.59494528, Global Avg Loss: 2.14963512, Time: 0.0210 Steps: 20380, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001229, Sample Num: 19664, Cur Loss: 0.39403117, Cur Avg Loss: 0.48804107, Log Avg loss: 0.45614625, Global Avg Loss: 2.14880457, Time: 0.0210 Steps: 20390, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001239, Sample Num: 19824, Cur Loss: 0.35788390, Cur Avg Loss: 0.48800857, Log Avg loss: 0.48401470, Global Avg Loss: 2.14798850, Time: 0.0210 Steps: 20400, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001249, Sample Num: 19984, Cur Loss: 0.61422330, Cur Avg Loss: 0.48895782, Log Avg loss: 0.60657004, Global Avg Loss: 2.14723327, Time: 0.0210 Steps: 20410, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001259, Sample Num: 20144, Cur Loss: 0.77055651, Cur Avg Loss: 0.49034533, Log Avg loss: 0.66364516, Global Avg Loss: 2.14650673, Time: 0.0210 Steps: 20420, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001269, Sample Num: 20304, Cur Loss: 0.15360403, Cur Avg Loss: 0.49110347, Log Avg loss: 0.58655385, Global Avg Loss: 2.14574317, Time: 0.0210 Steps: 20430, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001279, Sample Num: 20464, Cur Loss: 0.71609700, Cur Avg Loss: 0.49199542, Log Avg loss: 0.60518287, Global Avg Loss: 2.14498947, Time: 0.0210 Steps: 20440, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001289, Sample Num: 20624, Cur Loss: 0.58858597, Cur Avg Loss: 0.49136219, Log Avg loss: 0.41037258, Global Avg Loss: 2.14414125, Time: 0.0211 Steps: 20450, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001299, Sample Num: 20784, Cur Loss: 0.25049442, Cur Avg Loss: 0.49238617, Log Avg loss: 0.62437711, Global Avg Loss: 2.14339845, Time: 0.0210 Steps: 20460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001309, Sample Num: 20944, Cur Loss: 0.51626736, Cur Avg Loss: 0.49247235, Log Avg loss: 0.50366653, Global Avg Loss: 2.14259741, Time: 0.0210 Steps: 20470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001319, Sample Num: 21104, Cur Loss: 0.32874531, Cur Avg Loss: 0.49297724, Log Avg loss: 0.55906773, Global Avg Loss: 2.14182420, Time: 0.0210 Steps: 20480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001329, Sample Num: 21264, Cur Loss: 0.24759334, Cur Avg Loss: 0.49349074, Log Avg loss: 0.56122197, Global Avg Loss: 2.14105280, Time: 0.0211 Steps: 20490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001339, Sample Num: 21424, Cur Loss: 0.94959307, Cur Avg Loss: 0.49305601, Log Avg loss: 0.43527960, Global Avg Loss: 2.14022072, Time: 0.0210 Steps: 20500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001349, Sample Num: 21584, Cur Loss: 0.40422225, Cur Avg Loss: 0.49419017, Log Avg loss: 0.64605460, Global Avg Loss: 2.13949221, Time: 0.0209 Steps: 20510, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001359, Sample Num: 21744, Cur Loss: 0.61961794, Cur Avg Loss: 0.49448375, Log Avg loss: 0.53408829, Global Avg Loss: 2.13870985, Time: 0.0210 Steps: 20520, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001369, Sample Num: 21904, Cur Loss: 0.80078787, Cur Avg Loss: 0.49570780, Log Avg loss: 0.66205577, Global Avg Loss: 2.13799058, Time: 0.0210 Steps: 20530, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001379, Sample Num: 22064, Cur Loss: 0.59618145, Cur Avg Loss: 0.49657494, Log Avg loss: 0.61528687, Global Avg Loss: 2.13724925, Time: 0.0210 Steps: 20540, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001389, Sample Num: 22224, Cur Loss: 0.33088082, Cur Avg Loss: 0.49681822, Log Avg loss: 0.53036530, Global Avg Loss: 2.13646731, Time: 0.0210 Steps: 20550, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001399, Sample Num: 22384, Cur Loss: 0.60415065, Cur Avg Loss: 0.49600634, Log Avg loss: 0.38323682, Global Avg Loss: 2.13561457, Time: 0.0210 Steps: 20560, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001409, Sample Num: 22544, Cur Loss: 0.30005449, Cur Avg Loss: 0.49553622, Log Avg loss: 0.42976586, Global Avg Loss: 2.13478528, Time: 0.0211 Steps: 20570, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001419, Sample Num: 22704, Cur Loss: 0.48592603, Cur Avg Loss: 0.49499644, Log Avg loss: 0.41894262, Global Avg Loss: 2.13395154, Time: 0.0210 Steps: 20580, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001429, Sample Num: 22864, Cur Loss: 1.29019082, Cur Avg Loss: 0.49632515, Log Avg loss: 0.68486839, Global Avg Loss: 2.13324776, Time: 0.0210 Steps: 20590, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001439, Sample Num: 23024, Cur Loss: 0.41013080, Cur Avg Loss: 0.49560988, Log Avg loss: 0.39339845, Global Avg Loss: 2.13240317, Time: 0.0210 Steps: 20600, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001449, Sample Num: 23184, Cur Loss: 0.33158037, Cur Avg Loss: 0.49507217, Log Avg loss: 0.41769565, Global Avg Loss: 2.13157119, Time: 0.0210 Steps: 20610, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001459, Sample Num: 23344, Cur Loss: 0.13546644, Cur Avg Loss: 0.49466852, Log Avg loss: 0.43617929, Global Avg Loss: 2.13074899, Time: 0.0210 Steps: 20620, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001469, Sample Num: 23504, Cur Loss: 0.33986509, Cur Avg Loss: 0.49329634, Log Avg loss: 0.29309519, Global Avg Loss: 2.12985822, Time: 0.0210 Steps: 20630, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001479, Sample Num: 23664, Cur Loss: 0.53965735, Cur Avg Loss: 0.49308977, Log Avg loss: 0.46274399, Global Avg Loss: 2.12905051, Time: 0.0210 Steps: 20640, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001489, Sample Num: 23824, Cur Loss: 0.24414372, Cur Avg Loss: 0.49376327, Log Avg loss: 0.59337417, Global Avg Loss: 2.12830684, Time: 0.0210 Steps: 20650, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001499, Sample Num: 23984, Cur Loss: 0.51429093, Cur Avg Loss: 0.49284308, Log Avg loss: 0.35582669, Global Avg Loss: 2.12744891, Time: 0.0210 Steps: 20660, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001509, Sample Num: 24144, Cur Loss: 0.31603241, Cur Avg Loss: 0.49299937, Log Avg loss: 0.51642695, Global Avg Loss: 2.12666951, Time: 0.0210 Steps: 20670, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001519, Sample Num: 24304, Cur Loss: 0.23868147, Cur Avg Loss: 0.49294301, Log Avg loss: 0.48443859, Global Avg Loss: 2.12587539, Time: 0.0210 Steps: 20680, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001529, Sample Num: 24464, Cur Loss: 1.20079291, Cur Avg Loss: 0.49354874, Log Avg loss: 0.58555893, Global Avg Loss: 2.12513092, Time: 0.0210 Steps: 20690, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001539, Sample Num: 24624, Cur Loss: 0.62860876, Cur Avg Loss: 0.49382841, Log Avg loss: 0.53659010, Global Avg Loss: 2.12436351, Time: 0.0247 Steps: 20700, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001549, Sample Num: 24784, Cur Loss: 0.43531334, Cur Avg Loss: 0.49392595, Log Avg loss: 0.50893842, Global Avg Loss: 2.12358349, Time: 0.0211 Steps: 20710, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001559, Sample Num: 24944, Cur Loss: 0.38302243, Cur Avg Loss: 0.49345004, Log Avg loss: 0.41973061, Global Avg Loss: 2.12276116, Time: 0.0210 Steps: 20720, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001569, Sample Num: 25104, Cur Loss: 0.66807330, Cur Avg Loss: 0.49351550, Log Avg loss: 0.50372096, Global Avg Loss: 2.12198015, Time: 0.0211 Steps: 20730, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001579, Sample Num: 25264, Cur Loss: 0.41523999, Cur Avg Loss: 0.49303971, Log Avg loss: 0.41838844, Global Avg Loss: 2.12115875, Time: 0.0210 Steps: 20740, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001589, Sample Num: 25424, Cur Loss: 0.82437277, Cur Avg Loss: 0.49247063, Log Avg loss: 0.40261342, Global Avg Loss: 2.12033053, Time: 0.0210 Steps: 20750, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001599, Sample Num: 25584, Cur Loss: 0.67485088, Cur Avg Loss: 0.49162349, Log Avg loss: 0.35701186, Global Avg Loss: 2.11948115, Time: 0.0210 Steps: 20760, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001609, Sample Num: 25744, Cur Loss: 0.20641850, Cur Avg Loss: 0.49098140, Log Avg loss: 0.38831134, Global Avg Loss: 2.11864765, Time: 0.0211 Steps: 20770, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001619, Sample Num: 25904, Cur Loss: 0.34064287, Cur Avg Loss: 0.49047140, Log Avg loss: 0.40841252, Global Avg Loss: 2.11782463, Time: 0.0210 Steps: 20780, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001629, Sample Num: 26064, Cur Loss: 0.94832575, Cur Avg Loss: 0.49066955, Log Avg loss: 0.52275070, Global Avg Loss: 2.11705740, Time: 0.0210 Steps: 20790, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001639, Sample Num: 26224, Cur Loss: 0.58014226, Cur Avg Loss: 0.49133955, Log Avg loss: 0.60048250, Global Avg Loss: 2.11632828, Time: 0.0210 Steps: 20800, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001649, Sample Num: 26384, Cur Loss: 0.79355812, Cur Avg Loss: 0.49134375, Log Avg loss: 0.49203208, Global Avg Loss: 2.11554774, Time: 0.0210 Steps: 20810, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001659, Sample Num: 26544, Cur Loss: 0.51793498, Cur Avg Loss: 0.49105566, Log Avg loss: 0.44354974, Global Avg Loss: 2.11474467, Time: 0.0211 Steps: 20820, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001669, Sample Num: 26704, Cur Loss: 0.25868464, Cur Avg Loss: 0.48960163, Log Avg loss: 0.24837728, Global Avg Loss: 2.11384867, Time: 0.0210 Steps: 20830, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001679, Sample Num: 26864, Cur Loss: 0.17200468, Cur Avg Loss: 0.48912163, Log Avg loss: 0.40900968, Global Avg Loss: 2.11303061, Time: 0.0210 Steps: 20840, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001689, Sample Num: 27024, Cur Loss: 0.25287226, Cur Avg Loss: 0.48857576, Log Avg loss: 0.39692469, Global Avg Loss: 2.11220754, Time: 0.0210 Steps: 20850, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001699, Sample Num: 27184, Cur Loss: 0.16140720, Cur Avg Loss: 0.48798914, Log Avg loss: 0.38890914, Global Avg Loss: 2.11138141, Time: 0.0210 Steps: 20860, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001709, Sample Num: 27344, Cur Loss: 0.13862237, Cur Avg Loss: 0.48756078, Log Avg loss: 0.41478250, Global Avg Loss: 2.11056848, Time: 0.0210 Steps: 20870, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001719, Sample Num: 27504, Cur Loss: 0.21275447, Cur Avg Loss: 0.48653344, Log Avg loss: 0.31096052, Global Avg Loss: 2.10970659, Time: 0.0210 Steps: 20880, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001729, Sample Num: 27664, Cur Loss: 0.34602618, Cur Avg Loss: 0.48665094, Log Avg loss: 0.50684989, Global Avg Loss: 2.10893931, Time: 0.0210 Steps: 20890, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001739, Sample Num: 27824, Cur Loss: 0.40861362, Cur Avg Loss: 0.48712177, Log Avg loss: 0.56852831, Global Avg Loss: 2.10820227, Time: 0.0210 Steps: 20900, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001749, Sample Num: 27984, Cur Loss: 0.47834858, Cur Avg Loss: 0.48835021, Log Avg loss: 0.70197608, Global Avg Loss: 2.10752976, Time: 0.0210 Steps: 20910, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001759, Sample Num: 28144, Cur Loss: 0.20196551, Cur Avg Loss: 0.48903216, Log Avg loss: 0.60830470, Global Avg Loss: 2.10681311, Time: 0.0210 Steps: 20920, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001769, Sample Num: 28304, Cur Loss: 0.39018393, Cur Avg Loss: 0.48863325, Log Avg loss: 0.41846437, Global Avg Loss: 2.10600645, Time: 0.0210 Steps: 20930, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001779, Sample Num: 28464, Cur Loss: 0.40939754, Cur Avg Loss: 0.48777422, Log Avg loss: 0.33581297, Global Avg Loss: 2.10516108, Time: 0.0210 Steps: 20940, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001789, Sample Num: 28624, Cur Loss: 0.34888417, Cur Avg Loss: 0.48739432, Log Avg loss: 0.41980925, Global Avg Loss: 2.10435662, Time: 0.0211 Steps: 20950, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001799, Sample Num: 28784, Cur Loss: 0.38188055, Cur Avg Loss: 0.48774611, Log Avg loss: 0.55068083, Global Avg Loss: 2.10361536, Time: 0.0241 Steps: 20960, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001809, Sample Num: 28944, Cur Loss: 0.35164696, Cur Avg Loss: 0.48774324, Log Avg loss: 0.48722684, Global Avg Loss: 2.10284455, Time: 0.0210 Steps: 20970, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001819, Sample Num: 29104, Cur Loss: 0.57694310, Cur Avg Loss: 0.48883439, Log Avg loss: 0.68622378, Global Avg Loss: 2.10216933, Time: 0.0210 Steps: 20980, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001829, Sample Num: 29264, Cur Loss: 0.29131696, Cur Avg Loss: 0.48868806, Log Avg loss: 0.46207057, Global Avg Loss: 2.10138795, Time: 0.0210 Steps: 20990, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001839, Sample Num: 29424, Cur Loss: 0.57219160, Cur Avg Loss: 0.48873431, Log Avg loss: 0.49719460, Global Avg Loss: 2.10062405, Time: 0.0210 Steps: 21000, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001849, Sample Num: 29584, Cur Loss: 0.49146789, Cur Avg Loss: 0.48858352, Log Avg loss: 0.46085222, Global Avg Loss: 2.09984358, Time: 0.0210 Steps: 21010, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001859, Sample Num: 29744, Cur Loss: 0.76283687, Cur Avg Loss: 0.48828445, Log Avg loss: 0.43298684, Global Avg Loss: 2.09905059, Time: 0.0210 Steps: 21020, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001869, Sample Num: 29904, Cur Loss: 0.17947443, Cur Avg Loss: 0.48826285, Log Avg loss: 0.48424628, Global Avg Loss: 2.09828274, Time: 0.0209 Steps: 21030, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001879, Sample Num: 30064, Cur Loss: 0.13236575, Cur Avg Loss: 0.48846148, Log Avg loss: 0.52558611, Global Avg Loss: 2.09753526, Time: 0.0208 Steps: 21040, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001889, Sample Num: 30224, Cur Loss: 0.42138553, Cur Avg Loss: 0.48851086, Log Avg loss: 0.49778933, Global Avg Loss: 2.09677528, Time: 0.0209 Steps: 21050, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001899, Sample Num: 30384, Cur Loss: 0.35690144, Cur Avg Loss: 0.48834968, Log Avg loss: 0.45790380, Global Avg Loss: 2.09599709, Time: 0.0209 Steps: 21060, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001909, Sample Num: 30544, Cur Loss: 0.35382283, Cur Avg Loss: 0.48764383, Log Avg loss: 0.35360237, Global Avg Loss: 2.09517014, Time: 0.0210 Steps: 21070, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001919, Sample Num: 30704, Cur Loss: 0.38535574, Cur Avg Loss: 0.48773245, Log Avg loss: 0.50464907, Global Avg Loss: 2.09441562, Time: 0.0211 Steps: 21080, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001929, Sample Num: 30864, Cur Loss: 0.21190475, Cur Avg Loss: 0.48742751, Log Avg loss: 0.42891041, Global Avg Loss: 2.09362591, Time: 0.0210 Steps: 21090, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001939, Sample Num: 31024, Cur Loss: 0.32654801, Cur Avg Loss: 0.48745270, Log Avg loss: 0.49231243, Global Avg Loss: 2.09286699, Time: 0.0210 Steps: 21100, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001949, Sample Num: 31184, Cur Loss: 0.24930562, Cur Avg Loss: 0.48692269, Log Avg loss: 0.38415378, Global Avg Loss: 2.09205756, Time: 0.0209 Steps: 21110, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001959, Sample Num: 31344, Cur Loss: 0.42216644, Cur Avg Loss: 0.48652808, Log Avg loss: 0.40961789, Global Avg Loss: 2.09126095, Time: 0.0210 Steps: 21120, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001969, Sample Num: 31504, Cur Loss: 0.90037209, Cur Avg Loss: 0.48659966, Log Avg loss: 0.50062219, Global Avg Loss: 2.09050816, Time: 0.0210 Steps: 21130, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001979, Sample Num: 31664, Cur Loss: 0.23158760, Cur Avg Loss: 0.48651670, Log Avg loss: 0.47018147, Global Avg Loss: 2.08974169, Time: 0.0209 Steps: 21140, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001989, Sample Num: 31824, Cur Loss: 0.53022748, Cur Avg Loss: 0.48692722, Log Avg loss: 0.56816999, Global Avg Loss: 2.08902227, Time: 0.0210 Steps: 21150, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001999, Sample Num: 31984, Cur Loss: 0.80757630, Cur Avg Loss: 0.48759886, Log Avg loss: 0.62118791, Global Avg Loss: 2.08832858, Time: 0.0210 Steps: 21160, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002009, Sample Num: 32144, Cur Loss: 0.76735359, Cur Avg Loss: 0.48796901, Log Avg loss: 0.56196206, Global Avg Loss: 2.08760758, Time: 0.0210 Steps: 21170, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002019, Sample Num: 32304, Cur Loss: 0.59790206, Cur Avg Loss: 0.48767793, Log Avg loss: 0.42919907, Global Avg Loss: 2.08682457, Time: 0.0210 Steps: 21180, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002029, Sample Num: 32464, Cur Loss: 0.43604261, Cur Avg Loss: 0.48726769, Log Avg loss: 0.40444077, Global Avg Loss: 2.08603062, Time: 0.0209 Steps: 21190, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002039, Sample Num: 32624, Cur Loss: 0.70319194, Cur Avg Loss: 0.48688312, Log Avg loss: 0.40885393, Global Avg Loss: 2.08523950, Time: 0.0209 Steps: 21200, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002049, Sample Num: 32784, Cur Loss: 0.45026767, Cur Avg Loss: 0.48692605, Log Avg loss: 0.49568017, Global Avg Loss: 2.08449006, Time: 0.0246 Steps: 21210, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002059, Sample Num: 32944, Cur Loss: 0.12648597, Cur Avg Loss: 0.48635127, Log Avg loss: 0.36857880, Global Avg Loss: 2.08368143, Time: 0.0210 Steps: 21220, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002069, Sample Num: 33104, Cur Loss: 0.24422789, Cur Avg Loss: 0.48623351, Log Avg loss: 0.46198564, Global Avg Loss: 2.08291756, Time: 0.0209 Steps: 21230, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002079, Sample Num: 33264, Cur Loss: 0.71230268, Cur Avg Loss: 0.48642768, Log Avg loss: 0.52660240, Global Avg Loss: 2.08218483, Time: 0.0209 Steps: 21240, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002089, Sample Num: 33424, Cur Loss: 0.26532489, Cur Avg Loss: 0.48633436, Log Avg loss: 0.46693287, Global Avg Loss: 2.08142471, Time: 0.0209 Steps: 21250, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002099, Sample Num: 33584, Cur Loss: 0.58864325, Cur Avg Loss: 0.48585705, Log Avg loss: 0.38614601, Global Avg Loss: 2.08062731, Time: 0.0210 Steps: 21260, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002109, Sample Num: 33744, Cur Loss: 0.25989974, Cur Avg Loss: 0.48539353, Log Avg loss: 0.38810154, Global Avg Loss: 2.07983158, Time: 0.0210 Steps: 21270, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002119, Sample Num: 33904, Cur Loss: 0.67570895, Cur Avg Loss: 0.48506814, Log Avg loss: 0.41644403, Global Avg Loss: 2.07904991, Time: 0.0210 Steps: 21280, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002129, Sample Num: 34055, Cur Loss: 0.18433081, Cur Avg Loss: 0.48535936, Log Avg loss: 0.54706707, Global Avg Loss: 2.07833033, Time: 0.0101 Steps: 21290, Updated lr: 0.000081 ***** Running evaluation checkpoint-21290 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-21290 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.864339, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.449219, "eval_total_loss": 315.800848, "eval_mae": 0.461551, "eval_mse": 0.449397, "eval_r2": 0.714334, "eval_sp_statistic": 0.825303, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.857171, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.089678, "test_total_loss": 547.018589, "test_mae": 0.754268, "test_mse": 1.089932, "test_r2": 0.296548, "test_sp_statistic": 0.751833, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.798374, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.078330332265634, "train_cur_epoch_loss": 1033.3300689086318, "train_cur_epoch_avg_loss": 0.4853593559927815, "train_cur_epoch_time": 44.864338874816895, "train_cur_epoch_avg_time": 0.021072963304282242, "epoch": 10, "step": 21290} ################################################## Training, Epoch: 0011, Batch: 000010, Sample Num: 160, Cur Loss: 0.75324488, Cur Avg Loss: 0.47446836, Log Avg loss: 0.47446836, Global Avg Loss: 2.07757735, Time: 0.0209 Steps: 21300, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000020, Sample Num: 320, Cur Loss: 0.36164147, Cur Avg Loss: 0.45293763, Log Avg loss: 0.43140689, Global Avg Loss: 2.07680486, Time: 0.0209 Steps: 21310, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000030, Sample Num: 480, Cur Loss: 0.24975243, Cur Avg Loss: 0.54482950, Log Avg loss: 0.72861324, Global Avg Loss: 2.07617250, Time: 0.0208 Steps: 21320, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000040, Sample Num: 640, Cur Loss: 0.53275436, Cur Avg Loss: 0.56194836, Log Avg loss: 0.61330494, Global Avg Loss: 2.07548667, Time: 0.0209 Steps: 21330, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000050, Sample Num: 800, Cur Loss: 0.21915889, Cur Avg Loss: 0.53360474, Log Avg loss: 0.42023025, Global Avg Loss: 2.07471101, Time: 0.0209 Steps: 21340, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000060, Sample Num: 960, Cur Loss: 0.95190263, Cur Avg Loss: 0.52611356, Log Avg loss: 0.48865765, Global Avg Loss: 2.07396813, Time: 0.0209 Steps: 21350, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000070, Sample Num: 1120, Cur Loss: 0.91443658, Cur Avg Loss: 0.54696102, Log Avg loss: 0.67204580, Global Avg Loss: 2.07331180, Time: 0.0208 Steps: 21360, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000080, Sample Num: 1280, Cur Loss: 0.33859894, Cur Avg Loss: 0.52147326, Log Avg loss: 0.34305893, Global Avg Loss: 2.07250214, Time: 0.0210 Steps: 21370, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000090, Sample Num: 1440, Cur Loss: 0.55281550, Cur Avg Loss: 0.54140343, Log Avg loss: 0.70084483, Global Avg Loss: 2.07186057, Time: 0.0210 Steps: 21380, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000100, Sample Num: 1600, Cur Loss: 0.59727508, Cur Avg Loss: 0.53570583, Log Avg loss: 0.48442737, Global Avg Loss: 2.07111844, Time: 0.0210 Steps: 21390, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000110, Sample Num: 1760, Cur Loss: 0.27715334, Cur Avg Loss: 0.51862549, Log Avg loss: 0.34782213, Global Avg Loss: 2.07031316, Time: 0.0210 Steps: 21400, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000120, Sample Num: 1920, Cur Loss: 0.59568304, Cur Avg Loss: 0.51873527, Log Avg loss: 0.51994284, Global Avg Loss: 2.06958902, Time: 0.0210 Steps: 21410, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000130, Sample Num: 2080, Cur Loss: 0.30100071, Cur Avg Loss: 0.51429518, Log Avg loss: 0.46101412, Global Avg Loss: 2.06883806, Time: 0.0209 Steps: 21420, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000140, Sample Num: 2240, Cur Loss: 0.33008528, Cur Avg Loss: 0.50478520, Log Avg loss: 0.38115550, Global Avg Loss: 2.06805052, Time: 0.0210 Steps: 21430, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000150, Sample Num: 2400, Cur Loss: 0.34041232, Cur Avg Loss: 0.49570915, Log Avg loss: 0.36864445, Global Avg Loss: 2.06725789, Time: 0.0210 Steps: 21440, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000160, Sample Num: 2560, Cur Loss: 0.50810301, Cur Avg Loss: 0.49289295, Log Avg loss: 0.45064984, Global Avg Loss: 2.06650423, Time: 0.0210 Steps: 21450, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000170, Sample Num: 2720, Cur Loss: 0.33530071, Cur Avg Loss: 0.48840630, Log Avg loss: 0.41662000, Global Avg Loss: 2.06573541, Time: 0.0210 Steps: 21460, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000180, Sample Num: 2880, Cur Loss: 0.31868720, Cur Avg Loss: 0.48879520, Log Avg loss: 0.49540648, Global Avg Loss: 2.06500400, Time: 0.0209 Steps: 21470, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000190, Sample Num: 3040, Cur Loss: 0.47207302, Cur Avg Loss: 0.50098165, Log Avg loss: 0.72033779, Global Avg Loss: 2.06437799, Time: 0.0209 Steps: 21480, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000200, Sample Num: 3200, Cur Loss: 0.16779764, Cur Avg Loss: 0.50318841, Log Avg loss: 0.54511687, Global Avg Loss: 2.06367103, Time: 0.0209 Steps: 21490, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000210, Sample Num: 3360, Cur Loss: 0.57777727, Cur Avg Loss: 0.49724122, Log Avg loss: 0.37829735, Global Avg Loss: 2.06288714, Time: 0.0210 Steps: 21500, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000220, Sample Num: 3520, Cur Loss: 0.28537789, Cur Avg Loss: 0.49684263, Log Avg loss: 0.48847231, Global Avg Loss: 2.06215519, Time: 0.0209 Steps: 21510, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000230, Sample Num: 3680, Cur Loss: 0.30350164, Cur Avg Loss: 0.48768044, Log Avg loss: 0.28611219, Global Avg Loss: 2.06132989, Time: 0.0210 Steps: 21520, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000240, Sample Num: 3840, Cur Loss: 0.96110779, Cur Avg Loss: 0.48784266, Log Avg loss: 0.49157374, Global Avg Loss: 2.06060079, Time: 0.0210 Steps: 21530, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000250, Sample Num: 4000, Cur Loss: 0.31989020, Cur Avg Loss: 0.48721601, Log Avg loss: 0.47217643, Global Avg Loss: 2.05986336, Time: 0.0210 Steps: 21540, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000260, Sample Num: 4160, Cur Loss: 0.45480913, Cur Avg Loss: 0.48328053, Log Avg loss: 0.38489343, Global Avg Loss: 2.05908611, Time: 0.0248 Steps: 21550, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000270, Sample Num: 4320, Cur Loss: 0.42178845, Cur Avg Loss: 0.47623332, Log Avg loss: 0.29300580, Global Avg Loss: 2.05826697, Time: 0.0210 Steps: 21560, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000280, Sample Num: 4480, Cur Loss: 0.52977622, Cur Avg Loss: 0.47944365, Log Avg loss: 0.56612262, Global Avg Loss: 2.05757520, Time: 0.0210 Steps: 21570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000290, Sample Num: 4640, Cur Loss: 0.83979559, Cur Avg Loss: 0.48209812, Log Avg loss: 0.55642335, Global Avg Loss: 2.05687958, Time: 0.0210 Steps: 21580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000300, Sample Num: 4800, Cur Loss: 0.26619673, Cur Avg Loss: 0.48208733, Log Avg loss: 0.48177439, Global Avg Loss: 2.05615002, Time: 0.0210 Steps: 21590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000310, Sample Num: 4960, Cur Loss: 0.56428063, Cur Avg Loss: 0.49072568, Log Avg loss: 0.74987611, Global Avg Loss: 2.05554527, Time: 0.0210 Steps: 21600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000320, Sample Num: 5120, Cur Loss: 0.48511076, Cur Avg Loss: 0.49235327, Log Avg loss: 0.54280858, Global Avg Loss: 2.05484525, Time: 0.0210 Steps: 21610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000330, Sample Num: 5280, Cur Loss: 0.56315351, Cur Avg Loss: 0.48952592, Log Avg loss: 0.39905074, Global Avg Loss: 2.05407939, Time: 0.0210 Steps: 21620, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000340, Sample Num: 5440, Cur Loss: 0.56850898, Cur Avg Loss: 0.49248904, Log Avg loss: 0.59027206, Global Avg Loss: 2.05340264, Time: 0.0210 Steps: 21630, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000350, Sample Num: 5600, Cur Loss: 0.23780343, Cur Avg Loss: 0.49296804, Log Avg loss: 0.50925394, Global Avg Loss: 2.05268908, Time: 0.0210 Steps: 21640, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000360, Sample Num: 5760, Cur Loss: 0.36492333, Cur Avg Loss: 0.48802972, Log Avg loss: 0.31518876, Global Avg Loss: 2.05188653, Time: 0.0210 Steps: 21650, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000370, Sample Num: 5920, Cur Loss: 0.54786140, Cur Avg Loss: 0.48919266, Log Avg loss: 0.53105835, Global Avg Loss: 2.05118440, Time: 0.0210 Steps: 21660, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000380, Sample Num: 6080, Cur Loss: 0.51563227, Cur Avg Loss: 0.48816347, Log Avg loss: 0.45008350, Global Avg Loss: 2.05044554, Time: 0.0210 Steps: 21670, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000390, Sample Num: 6240, Cur Loss: 1.33896613, Cur Avg Loss: 0.49000702, Log Avg loss: 0.56006189, Global Avg Loss: 2.04975810, Time: 0.0210 Steps: 21680, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000400, Sample Num: 6400, Cur Loss: 0.49790686, Cur Avg Loss: 0.48796510, Log Avg loss: 0.40833030, Global Avg Loss: 2.04900133, Time: 0.0210 Steps: 21690, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000410, Sample Num: 6560, Cur Loss: 0.29744011, Cur Avg Loss: 0.48813333, Log Avg loss: 0.49486253, Global Avg Loss: 2.04828514, Time: 0.0210 Steps: 21700, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000420, Sample Num: 6720, Cur Loss: 0.31612581, Cur Avg Loss: 0.48700147, Log Avg loss: 0.44059519, Global Avg Loss: 2.04754461, Time: 0.0209 Steps: 21710, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000430, Sample Num: 6880, Cur Loss: 0.67394561, Cur Avg Loss: 0.48761574, Log Avg loss: 0.51341507, Global Avg Loss: 2.04683828, Time: 0.0209 Steps: 21720, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000440, Sample Num: 7040, Cur Loss: 0.38466781, Cur Avg Loss: 0.48996770, Log Avg loss: 0.59110186, Global Avg Loss: 2.04616836, Time: 0.0210 Steps: 21730, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000450, Sample Num: 7200, Cur Loss: 1.01724803, Cur Avg Loss: 0.49555927, Log Avg loss: 0.74158845, Global Avg Loss: 2.04556828, Time: 0.0210 Steps: 21740, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000460, Sample Num: 7360, Cur Loss: 0.47360834, Cur Avg Loss: 0.49849450, Log Avg loss: 0.63057959, Global Avg Loss: 2.04491771, Time: 0.0210 Steps: 21750, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000470, Sample Num: 7520, Cur Loss: 0.28484905, Cur Avg Loss: 0.49943763, Log Avg loss: 0.54282178, Global Avg Loss: 2.04422741, Time: 0.0210 Steps: 21760, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000480, Sample Num: 7680, Cur Loss: 0.56813157, Cur Avg Loss: 0.49923126, Log Avg loss: 0.48953182, Global Avg Loss: 2.04351326, Time: 0.0210 Steps: 21770, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000490, Sample Num: 7840, Cur Loss: 0.33551860, Cur Avg Loss: 0.49848672, Log Avg loss: 0.46274882, Global Avg Loss: 2.04278748, Time: 0.0210 Steps: 21780, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000500, Sample Num: 8000, Cur Loss: 0.50778651, Cur Avg Loss: 0.50118372, Log Avg loss: 0.63333683, Global Avg Loss: 2.04214064, Time: 0.0210 Steps: 21790, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000510, Sample Num: 8160, Cur Loss: 0.29137051, Cur Avg Loss: 0.50083593, Log Avg loss: 0.48344614, Global Avg Loss: 2.04142565, Time: 0.0209 Steps: 21800, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000520, Sample Num: 8320, Cur Loss: 0.38956243, Cur Avg Loss: 0.50080117, Log Avg loss: 0.49902841, Global Avg Loss: 2.04071845, Time: 0.0210 Steps: 21810, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000530, Sample Num: 8480, Cur Loss: 0.48009354, Cur Avg Loss: 0.50104787, Log Avg loss: 0.51387654, Global Avg Loss: 2.04001871, Time: 0.0209 Steps: 21820, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000540, Sample Num: 8640, Cur Loss: 0.40557045, Cur Avg Loss: 0.49751510, Log Avg loss: 0.31027836, Global Avg Loss: 2.03922634, Time: 0.0210 Steps: 21830, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000550, Sample Num: 8800, Cur Loss: 0.32970291, Cur Avg Loss: 0.49354981, Log Avg loss: 0.27942402, Global Avg Loss: 2.03842057, Time: 0.0209 Steps: 21840, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000560, Sample Num: 8960, Cur Loss: 0.73401690, Cur Avg Loss: 0.49252561, Log Avg loss: 0.43619466, Global Avg Loss: 2.03768728, Time: 0.0209 Steps: 21850, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000570, Sample Num: 9120, Cur Loss: 0.61379635, Cur Avg Loss: 0.49540104, Log Avg loss: 0.65642525, Global Avg Loss: 2.03705541, Time: 0.0209 Steps: 21860, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000580, Sample Num: 9280, Cur Loss: 0.43460095, Cur Avg Loss: 0.49231658, Log Avg loss: 0.31650196, Global Avg Loss: 2.03626870, Time: 0.0209 Steps: 21870, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000590, Sample Num: 9440, Cur Loss: 0.20770624, Cur Avg Loss: 0.49236766, Log Avg loss: 0.49533076, Global Avg Loss: 2.03556443, Time: 0.0209 Steps: 21880, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000600, Sample Num: 9600, Cur Loss: 0.40129745, Cur Avg Loss: 0.49075835, Log Avg loss: 0.39580858, Global Avg Loss: 2.03481534, Time: 0.0209 Steps: 21890, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000610, Sample Num: 9760, Cur Loss: 0.18284565, Cur Avg Loss: 0.48937718, Log Avg loss: 0.40650725, Global Avg Loss: 2.03407182, Time: 0.0210 Steps: 21900, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000620, Sample Num: 9920, Cur Loss: 0.58414489, Cur Avg Loss: 0.48863577, Log Avg loss: 0.44340944, Global Avg Loss: 2.03334582, Time: 0.0210 Steps: 21910, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000630, Sample Num: 10080, Cur Loss: 0.44131365, Cur Avg Loss: 0.48653473, Log Avg loss: 0.35627041, Global Avg Loss: 2.03258073, Time: 0.0209 Steps: 21920, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000640, Sample Num: 10240, Cur Loss: 0.45502067, Cur Avg Loss: 0.48432399, Log Avg loss: 0.34504760, Global Avg Loss: 2.03181122, Time: 0.0209 Steps: 21930, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000650, Sample Num: 10400, Cur Loss: 0.20551217, Cur Avg Loss: 0.48157346, Log Avg loss: 0.30553965, Global Avg Loss: 2.03102441, Time: 0.0209 Steps: 21940, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000660, Sample Num: 10560, Cur Loss: 0.15055683, Cur Avg Loss: 0.47868166, Log Avg loss: 0.29071429, Global Avg Loss: 2.03023156, Time: 0.0209 Steps: 21950, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000670, Sample Num: 10720, Cur Loss: 1.31533384, Cur Avg Loss: 0.47858327, Log Avg loss: 0.47208944, Global Avg Loss: 2.02952202, Time: 0.0209 Steps: 21960, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000680, Sample Num: 10880, Cur Loss: 0.34417182, Cur Avg Loss: 0.47779134, Log Avg loss: 0.42473262, Global Avg Loss: 2.02879157, Time: 0.0209 Steps: 21970, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000690, Sample Num: 11040, Cur Loss: 0.36126018, Cur Avg Loss: 0.47811883, Log Avg loss: 0.50038787, Global Avg Loss: 2.02809621, Time: 0.0209 Steps: 21980, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000700, Sample Num: 11200, Cur Loss: 0.12802066, Cur Avg Loss: 0.47947926, Log Avg loss: 0.57334919, Global Avg Loss: 2.02743466, Time: 0.0209 Steps: 21990, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000710, Sample Num: 11360, Cur Loss: 0.28209320, Cur Avg Loss: 0.47645610, Log Avg loss: 0.26483422, Global Avg Loss: 2.02663348, Time: 0.0209 Steps: 22000, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000720, Sample Num: 11520, Cur Loss: 0.66468489, Cur Avg Loss: 0.47534672, Log Avg loss: 0.39658119, Global Avg Loss: 2.02589289, Time: 0.0209 Steps: 22010, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000730, Sample Num: 11680, Cur Loss: 0.87872696, Cur Avg Loss: 0.47600796, Log Avg loss: 0.52361694, Global Avg Loss: 2.02521065, Time: 0.0209 Steps: 22020, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000740, Sample Num: 11840, Cur Loss: 0.36099672, Cur Avg Loss: 0.47491891, Log Avg loss: 0.39541838, Global Avg Loss: 2.02447085, Time: 0.0209 Steps: 22030, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000750, Sample Num: 12000, Cur Loss: 1.10668480, Cur Avg Loss: 0.47373798, Log Avg loss: 0.38634917, Global Avg Loss: 2.02372760, Time: 0.0209 Steps: 22040, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000760, Sample Num: 12160, Cur Loss: 0.38503444, Cur Avg Loss: 0.47201829, Log Avg loss: 0.34304141, Global Avg Loss: 2.02296538, Time: 0.0209 Steps: 22050, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000770, Sample Num: 12320, Cur Loss: 0.34374553, Cur Avg Loss: 0.47317072, Log Avg loss: 0.56075567, Global Avg Loss: 2.02230255, Time: 0.0247 Steps: 22060, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000780, Sample Num: 12480, Cur Loss: 0.26639104, Cur Avg Loss: 0.47103526, Log Avg loss: 0.30660509, Global Avg Loss: 2.02152516, Time: 0.0209 Steps: 22070, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000790, Sample Num: 12640, Cur Loss: 0.47597352, Cur Avg Loss: 0.47149698, Log Avg loss: 0.50751101, Global Avg Loss: 2.02083947, Time: 0.0209 Steps: 22080, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000800, Sample Num: 12800, Cur Loss: 0.10824753, Cur Avg Loss: 0.47116195, Log Avg loss: 0.44469407, Global Avg Loss: 2.02012595, Time: 0.0209 Steps: 22090, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000810, Sample Num: 12960, Cur Loss: 0.70100474, Cur Avg Loss: 0.47037783, Log Avg loss: 0.40764867, Global Avg Loss: 2.01939633, Time: 0.0209 Steps: 22100, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000820, Sample Num: 13120, Cur Loss: 0.22188258, Cur Avg Loss: 0.47084709, Log Avg loss: 0.50885710, Global Avg Loss: 2.01871313, Time: 0.0209 Steps: 22110, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000830, Sample Num: 13280, Cur Loss: 0.83962166, Cur Avg Loss: 0.47200948, Log Avg loss: 0.56732577, Global Avg Loss: 2.01805699, Time: 0.0209 Steps: 22120, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000840, Sample Num: 13440, Cur Loss: 0.24412449, Cur Avg Loss: 0.47156386, Log Avg loss: 0.43457712, Global Avg Loss: 2.01734146, Time: 0.0209 Steps: 22130, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000850, Sample Num: 13600, Cur Loss: 0.53068817, Cur Avg Loss: 0.47181997, Log Avg loss: 0.49333306, Global Avg Loss: 2.01665311, Time: 0.0209 Steps: 22140, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000860, Sample Num: 13760, Cur Loss: 0.15997241, Cur Avg Loss: 0.47153163, Log Avg loss: 0.44702250, Global Avg Loss: 2.01594447, Time: 0.0209 Steps: 22150, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000870, Sample Num: 13920, Cur Loss: 0.48496056, Cur Avg Loss: 0.47115783, Log Avg loss: 0.43901156, Global Avg Loss: 2.01523286, Time: 0.0209 Steps: 22160, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000880, Sample Num: 14080, Cur Loss: 0.31694728, Cur Avg Loss: 0.47031886, Log Avg loss: 0.39732845, Global Avg Loss: 2.01450308, Time: 0.0209 Steps: 22170, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000890, Sample Num: 14240, Cur Loss: 0.21372320, Cur Avg Loss: 0.47131149, Log Avg loss: 0.55866307, Global Avg Loss: 2.01384671, Time: 0.0209 Steps: 22180, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000900, Sample Num: 14400, Cur Loss: 0.22667190, Cur Avg Loss: 0.47035021, Log Avg loss: 0.38479554, Global Avg Loss: 2.01311257, Time: 0.0209 Steps: 22190, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000910, Sample Num: 14560, Cur Loss: 0.46786296, Cur Avg Loss: 0.47011065, Log Avg loss: 0.44855039, Global Avg Loss: 2.01240781, Time: 0.0209 Steps: 22200, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000920, Sample Num: 14720, Cur Loss: 1.76450753, Cur Avg Loss: 0.47116243, Log Avg loss: 0.56687506, Global Avg Loss: 2.01175697, Time: 0.0209 Steps: 22210, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000930, Sample Num: 14880, Cur Loss: 0.62500525, Cur Avg Loss: 0.47406128, Log Avg loss: 0.74075542, Global Avg Loss: 2.01118496, Time: 0.0209 Steps: 22220, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000940, Sample Num: 15040, Cur Loss: 0.22290757, Cur Avg Loss: 0.47478951, Log Avg loss: 0.54251465, Global Avg Loss: 2.01052429, Time: 0.0209 Steps: 22230, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000950, Sample Num: 15200, Cur Loss: 0.57669747, Cur Avg Loss: 0.47451965, Log Avg loss: 0.44915253, Global Avg Loss: 2.00982223, Time: 0.0209 Steps: 22240, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000960, Sample Num: 15360, Cur Loss: 0.67539346, Cur Avg Loss: 0.47399150, Log Avg loss: 0.42381721, Global Avg Loss: 2.00910942, Time: 0.0209 Steps: 22250, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000970, Sample Num: 15520, Cur Loss: 0.56184894, Cur Avg Loss: 0.47268752, Log Avg loss: 0.34750523, Global Avg Loss: 2.00836297, Time: 0.0209 Steps: 22260, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000980, Sample Num: 15680, Cur Loss: 0.60063297, Cur Avg Loss: 0.47340447, Log Avg loss: 0.54294882, Global Avg Loss: 2.00770495, Time: 0.0209 Steps: 22270, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000990, Sample Num: 15840, Cur Loss: 0.55926114, Cur Avg Loss: 0.47409027, Log Avg loss: 0.54129932, Global Avg Loss: 2.00704677, Time: 0.0209 Steps: 22280, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001000, Sample Num: 16000, Cur Loss: 0.92191112, Cur Avg Loss: 0.47459632, Log Avg loss: 0.52469513, Global Avg Loss: 2.00638175, Time: 0.0210 Steps: 22290, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001010, Sample Num: 16160, Cur Loss: 0.52652788, Cur Avg Loss: 0.47569679, Log Avg loss: 0.58574308, Global Avg Loss: 2.00574469, Time: 0.0209 Steps: 22300, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001020, Sample Num: 16320, Cur Loss: 0.43058571, Cur Avg Loss: 0.47637538, Log Avg loss: 0.54491377, Global Avg Loss: 2.00508990, Time: 0.0209 Steps: 22310, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001030, Sample Num: 16480, Cur Loss: 1.21374381, Cur Avg Loss: 0.47655851, Log Avg loss: 0.49523747, Global Avg Loss: 2.00441344, Time: 0.0247 Steps: 22320, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001040, Sample Num: 16640, Cur Loss: 0.72294843, Cur Avg Loss: 0.47465697, Log Avg loss: 0.27879785, Global Avg Loss: 2.00364066, Time: 0.0210 Steps: 22330, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001050, Sample Num: 16800, Cur Loss: 0.71309310, Cur Avg Loss: 0.47375166, Log Avg loss: 0.37959999, Global Avg Loss: 2.00291370, Time: 0.0209 Steps: 22340, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001060, Sample Num: 16960, Cur Loss: 0.50984067, Cur Avg Loss: 0.47317840, Log Avg loss: 0.41298648, Global Avg Loss: 2.00220232, Time: 0.0209 Steps: 22350, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001070, Sample Num: 17120, Cur Loss: 0.22382322, Cur Avg Loss: 0.47483310, Log Avg loss: 0.65023094, Global Avg Loss: 2.00159768, Time: 0.0209 Steps: 22360, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001080, Sample Num: 17280, Cur Loss: 0.72479045, Cur Avg Loss: 0.47574126, Log Avg loss: 0.57291449, Global Avg Loss: 2.00095902, Time: 0.0209 Steps: 22370, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001090, Sample Num: 17440, Cur Loss: 0.51838064, Cur Avg Loss: 0.47696210, Log Avg loss: 0.60881233, Global Avg Loss: 2.00033697, Time: 0.0209 Steps: 22380, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001100, Sample Num: 17600, Cur Loss: 0.13300392, Cur Avg Loss: 0.47773431, Log Avg loss: 0.56190512, Global Avg Loss: 1.99969453, Time: 0.0209 Steps: 22390, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001110, Sample Num: 17760, Cur Loss: 0.18937525, Cur Avg Loss: 0.47762000, Log Avg loss: 0.46504570, Global Avg Loss: 1.99900942, Time: 0.0209 Steps: 22400, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001120, Sample Num: 17920, Cur Loss: 0.43216234, Cur Avg Loss: 0.47854144, Log Avg loss: 0.58082162, Global Avg Loss: 1.99837658, Time: 0.0209 Steps: 22410, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001130, Sample Num: 18080, Cur Loss: 0.22448868, Cur Avg Loss: 0.47809162, Log Avg loss: 0.42771198, Global Avg Loss: 1.99767602, Time: 0.0210 Steps: 22420, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001140, Sample Num: 18240, Cur Loss: 0.37161547, Cur Avg Loss: 0.47752844, Log Avg loss: 0.41388886, Global Avg Loss: 1.99696991, Time: 0.0209 Steps: 22430, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001150, Sample Num: 18400, Cur Loss: 0.67885244, Cur Avg Loss: 0.47775424, Log Avg loss: 0.50349587, Global Avg Loss: 1.99630437, Time: 0.0209 Steps: 22440, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001160, Sample Num: 18560, Cur Loss: 0.43294576, Cur Avg Loss: 0.47670803, Log Avg loss: 0.35639423, Global Avg Loss: 1.99557390, Time: 0.0209 Steps: 22450, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001170, Sample Num: 18720, Cur Loss: 0.26528764, Cur Avg Loss: 0.47581909, Log Avg loss: 0.37270140, Global Avg Loss: 1.99485134, Time: 0.0209 Steps: 22460, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001180, Sample Num: 18880, Cur Loss: 0.43812811, Cur Avg Loss: 0.47641661, Log Avg loss: 0.54632650, Global Avg Loss: 1.99420669, Time: 0.0209 Steps: 22470, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001190, Sample Num: 19040, Cur Loss: 0.51357013, Cur Avg Loss: 0.47680807, Log Avg loss: 0.52300004, Global Avg Loss: 1.99355224, Time: 0.0209 Steps: 22480, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001200, Sample Num: 19200, Cur Loss: 0.53573340, Cur Avg Loss: 0.47685209, Log Avg loss: 0.48209036, Global Avg Loss: 1.99288018, Time: 0.0209 Steps: 22490, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001210, Sample Num: 19360, Cur Loss: 0.22558464, Cur Avg Loss: 0.47737246, Log Avg loss: 0.53981679, Global Avg Loss: 1.99223438, Time: 0.0209 Steps: 22500, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001220, Sample Num: 19520, Cur Loss: 0.13573596, Cur Avg Loss: 0.47670469, Log Avg loss: 0.39590520, Global Avg Loss: 1.99152521, Time: 0.0209 Steps: 22510, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001230, Sample Num: 19680, Cur Loss: 0.24074697, Cur Avg Loss: 0.47695083, Log Avg loss: 0.50697963, Global Avg Loss: 1.99086600, Time: 0.0209 Steps: 22520, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001240, Sample Num: 19840, Cur Loss: 0.13466138, Cur Avg Loss: 0.47681529, Log Avg loss: 0.46014405, Global Avg Loss: 1.99018658, Time: 0.0209 Steps: 22530, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001250, Sample Num: 20000, Cur Loss: 0.21736556, Cur Avg Loss: 0.47804759, Log Avg loss: 0.63085232, Global Avg Loss: 1.98958351, Time: 0.0209 Steps: 22540, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001260, Sample Num: 20160, Cur Loss: 0.29021281, Cur Avg Loss: 0.47849807, Log Avg loss: 0.53480827, Global Avg Loss: 1.98893837, Time: 0.0209 Steps: 22550, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001270, Sample Num: 20320, Cur Loss: 0.52271992, Cur Avg Loss: 0.47867986, Log Avg loss: 0.50158583, Global Avg Loss: 1.98827909, Time: 0.0209 Steps: 22560, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001280, Sample Num: 20480, Cur Loss: 0.86889040, Cur Avg Loss: 0.47793924, Log Avg loss: 0.38388012, Global Avg Loss: 1.98756823, Time: 0.0254 Steps: 22570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001290, Sample Num: 20640, Cur Loss: 0.74376583, Cur Avg Loss: 0.47791857, Log Avg loss: 0.47527240, Global Avg Loss: 1.98689848, Time: 0.0211 Steps: 22580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001300, Sample Num: 20800, Cur Loss: 1.03697026, Cur Avg Loss: 0.47888611, Log Avg loss: 0.60369998, Global Avg Loss: 1.98628618, Time: 0.0210 Steps: 22590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001310, Sample Num: 20960, Cur Loss: 0.34777367, Cur Avg Loss: 0.47837572, Log Avg loss: 0.41202385, Global Avg Loss: 1.98558960, Time: 0.0211 Steps: 22600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001320, Sample Num: 21120, Cur Loss: 0.78971726, Cur Avg Loss: 0.47820800, Log Avg loss: 0.45623782, Global Avg Loss: 1.98491320, Time: 0.0210 Steps: 22610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001330, Sample Num: 21280, Cur Loss: 0.72125924, Cur Avg Loss: 0.47810700, Log Avg loss: 0.46477435, Global Avg Loss: 1.98424116, Time: 0.0210 Steps: 22620, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001340, Sample Num: 21440, Cur Loss: 0.66175878, Cur Avg Loss: 0.47900950, Log Avg loss: 0.59904233, Global Avg Loss: 1.98362905, Time: 0.0210 Steps: 22630, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001350, Sample Num: 21600, Cur Loss: 0.30403054, Cur Avg Loss: 0.47839291, Log Avg loss: 0.39576960, Global Avg Loss: 1.98292770, Time: 0.0210 Steps: 22640, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001360, Sample Num: 21760, Cur Loss: 0.31776464, Cur Avg Loss: 0.47830310, Log Avg loss: 0.46617855, Global Avg Loss: 1.98225806, Time: 0.0211 Steps: 22650, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001370, Sample Num: 21920, Cur Loss: 0.14629035, Cur Avg Loss: 0.47739271, Log Avg loss: 0.35357952, Global Avg Loss: 1.98153931, Time: 0.0211 Steps: 22660, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001380, Sample Num: 22080, Cur Loss: 0.13465512, Cur Avg Loss: 0.47771321, Log Avg loss: 0.52162195, Global Avg Loss: 1.98089532, Time: 0.0210 Steps: 22670, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001390, Sample Num: 22240, Cur Loss: 0.13189274, Cur Avg Loss: 0.47663712, Log Avg loss: 0.32813659, Global Avg Loss: 1.98016659, Time: 0.0210 Steps: 22680, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001400, Sample Num: 22400, Cur Loss: 0.41685605, Cur Avg Loss: 0.47675509, Log Avg loss: 0.49315273, Global Avg Loss: 1.97951123, Time: 0.0210 Steps: 22690, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001410, Sample Num: 22560, Cur Loss: 0.21637122, Cur Avg Loss: 0.47722753, Log Avg loss: 0.54336920, Global Avg Loss: 1.97887857, Time: 0.0210 Steps: 22700, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001420, Sample Num: 22720, Cur Loss: 0.86494005, Cur Avg Loss: 0.47685921, Log Avg loss: 0.42492590, Global Avg Loss: 1.97819431, Time: 0.0211 Steps: 22710, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001430, Sample Num: 22880, Cur Loss: 0.25698006, Cur Avg Loss: 0.47673819, Log Avg loss: 0.45955386, Global Avg Loss: 1.97752590, Time: 0.0210 Steps: 22720, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001440, Sample Num: 23040, Cur Loss: 0.19433427, Cur Avg Loss: 0.47754669, Log Avg loss: 0.59316215, Global Avg Loss: 1.97691685, Time: 0.0210 Steps: 22730, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001450, Sample Num: 23200, Cur Loss: 0.13724139, Cur Avg Loss: 0.47693457, Log Avg loss: 0.38878875, Global Avg Loss: 1.97621847, Time: 0.0210 Steps: 22740, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001460, Sample Num: 23360, Cur Loss: 0.52960014, Cur Avg Loss: 0.47724314, Log Avg loss: 0.52198666, Global Avg Loss: 1.97557924, Time: 0.0210 Steps: 22750, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001470, Sample Num: 23520, Cur Loss: 0.59071958, Cur Avg Loss: 0.47735497, Log Avg loss: 0.49368115, Global Avg Loss: 1.97492814, Time: 0.0211 Steps: 22760, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001480, Sample Num: 23680, Cur Loss: 0.59044361, Cur Avg Loss: 0.47665649, Log Avg loss: 0.37398107, Global Avg Loss: 1.97422505, Time: 0.0211 Steps: 22770, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001490, Sample Num: 23840, Cur Loss: 0.28480095, Cur Avg Loss: 0.47594104, Log Avg loss: 0.37005434, Global Avg Loss: 1.97352085, Time: 0.0210 Steps: 22780, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001500, Sample Num: 24000, Cur Loss: 0.26202208, Cur Avg Loss: 0.47492222, Log Avg loss: 0.32311701, Global Avg Loss: 1.97279667, Time: 0.0211 Steps: 22790, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001510, Sample Num: 24160, Cur Loss: 0.50832772, Cur Avg Loss: 0.47512700, Log Avg loss: 0.50584511, Global Avg Loss: 1.97215327, Time: 0.0211 Steps: 22800, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001520, Sample Num: 24320, Cur Loss: 0.51020157, Cur Avg Loss: 0.47520262, Log Avg loss: 0.48662069, Global Avg Loss: 1.97150201, Time: 0.0210 Steps: 22810, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001530, Sample Num: 24480, Cur Loss: 0.33279499, Cur Avg Loss: 0.47592385, Log Avg loss: 0.58555035, Global Avg Loss: 1.97089467, Time: 0.0210 Steps: 22820, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001540, Sample Num: 24640, Cur Loss: 0.25655094, Cur Avg Loss: 0.47559913, Log Avg loss: 0.42591813, Global Avg Loss: 1.97021793, Time: 0.0246 Steps: 22830, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001550, Sample Num: 24800, Cur Loss: 0.35958165, Cur Avg Loss: 0.47559499, Log Avg loss: 0.47495700, Global Avg Loss: 1.96956327, Time: 0.0209 Steps: 22840, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001560, Sample Num: 24960, Cur Loss: 0.52286416, Cur Avg Loss: 0.47527615, Log Avg loss: 0.42585559, Global Avg Loss: 1.96888768, Time: 0.0209 Steps: 22850, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001570, Sample Num: 25120, Cur Loss: 0.61305773, Cur Avg Loss: 0.47585711, Log Avg loss: 0.56648758, Global Avg Loss: 1.96827421, Time: 0.0209 Steps: 22860, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001580, Sample Num: 25280, Cur Loss: 0.51879692, Cur Avg Loss: 0.47637483, Log Avg loss: 0.55765654, Global Avg Loss: 1.96765741, Time: 0.0209 Steps: 22870, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001590, Sample Num: 25440, Cur Loss: 0.60355389, Cur Avg Loss: 0.47673110, Log Avg loss: 0.53302096, Global Avg Loss: 1.96703039, Time: 0.0209 Steps: 22880, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001600, Sample Num: 25600, Cur Loss: 0.41445634, Cur Avg Loss: 0.47677945, Log Avg loss: 0.48446701, Global Avg Loss: 1.96638269, Time: 0.0208 Steps: 22890, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001610, Sample Num: 25760, Cur Loss: 0.29600805, Cur Avg Loss: 0.47677233, Log Avg loss: 0.47563325, Global Avg Loss: 1.96573171, Time: 0.0209 Steps: 22900, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001620, Sample Num: 25920, Cur Loss: 0.34519529, Cur Avg Loss: 0.47677807, Log Avg loss: 0.47770309, Global Avg Loss: 1.96508220, Time: 0.0209 Steps: 22910, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001630, Sample Num: 26080, Cur Loss: 0.72416490, Cur Avg Loss: 0.47790724, Log Avg loss: 0.66083207, Global Avg Loss: 1.96451316, Time: 0.0209 Steps: 22920, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001640, Sample Num: 26240, Cur Loss: 0.30028298, Cur Avg Loss: 0.47745309, Log Avg loss: 0.40342710, Global Avg Loss: 1.96383235, Time: 0.0209 Steps: 22930, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001650, Sample Num: 26400, Cur Loss: 0.31852239, Cur Avg Loss: 0.47740074, Log Avg loss: 0.46881588, Global Avg Loss: 1.96318065, Time: 0.0209 Steps: 22940, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001660, Sample Num: 26560, Cur Loss: 0.22758478, Cur Avg Loss: 0.47704606, Log Avg loss: 0.41852370, Global Avg Loss: 1.96250759, Time: 0.0209 Steps: 22950, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001670, Sample Num: 26720, Cur Loss: 1.39887261, Cur Avg Loss: 0.47745519, Log Avg loss: 0.54537044, Global Avg Loss: 1.96189037, Time: 0.0209 Steps: 22960, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001680, Sample Num: 26880, Cur Loss: 0.67680538, Cur Avg Loss: 0.47851137, Log Avg loss: 0.65489282, Global Avg Loss: 1.96132137, Time: 0.0209 Steps: 22970, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001690, Sample Num: 27040, Cur Loss: 0.11212522, Cur Avg Loss: 0.47912064, Log Avg loss: 0.58147790, Global Avg Loss: 1.96072092, Time: 0.0209 Steps: 22980, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001700, Sample Num: 27200, Cur Loss: 0.34374076, Cur Avg Loss: 0.47927295, Log Avg loss: 0.50501438, Global Avg Loss: 1.96008772, Time: 0.0209 Steps: 22990, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001710, Sample Num: 27360, Cur Loss: 0.13098228, Cur Avg Loss: 0.47900225, Log Avg loss: 0.43298271, Global Avg Loss: 1.95942377, Time: 0.0209 Steps: 23000, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001720, Sample Num: 27520, Cur Loss: 0.40494132, Cur Avg Loss: 0.47861414, Log Avg loss: 0.41224805, Global Avg Loss: 1.95875137, Time: 0.0209 Steps: 23010, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001730, Sample Num: 27680, Cur Loss: 0.76020741, Cur Avg Loss: 0.47792348, Log Avg loss: 0.35913010, Global Avg Loss: 1.95805649, Time: 0.0209 Steps: 23020, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001740, Sample Num: 27840, Cur Loss: 0.34512746, Cur Avg Loss: 0.47732385, Log Avg loss: 0.37358691, Global Avg Loss: 1.95736849, Time: 0.0209 Steps: 23030, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001750, Sample Num: 28000, Cur Loss: 0.95466626, Cur Avg Loss: 0.47717713, Log Avg loss: 0.45164870, Global Avg Loss: 1.95671496, Time: 0.0209 Steps: 23040, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001760, Sample Num: 28160, Cur Loss: 0.31075847, Cur Avg Loss: 0.47679651, Log Avg loss: 0.41018659, Global Avg Loss: 1.95604402, Time: 0.0209 Steps: 23050, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001770, Sample Num: 28320, Cur Loss: 0.54995406, Cur Avg Loss: 0.47673321, Log Avg loss: 0.46559319, Global Avg Loss: 1.95539768, Time: 0.0209 Steps: 23060, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001780, Sample Num: 28480, Cur Loss: 0.41635337, Cur Avg Loss: 0.47617332, Log Avg loss: 0.37707282, Global Avg Loss: 1.95471354, Time: 0.0209 Steps: 23070, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001790, Sample Num: 28640, Cur Loss: 0.35702544, Cur Avg Loss: 0.47536062, Log Avg loss: 0.33069985, Global Avg Loss: 1.95400989, Time: 0.0209 Steps: 23080, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001800, Sample Num: 28800, Cur Loss: 0.85925305, Cur Avg Loss: 0.47535962, Log Avg loss: 0.47518067, Global Avg Loss: 1.95336943, Time: 0.0210 Steps: 23090, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001810, Sample Num: 28960, Cur Loss: 0.35862309, Cur Avg Loss: 0.47487302, Log Avg loss: 0.38728597, Global Avg Loss: 1.95269147, Time: 0.0209 Steps: 23100, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001820, Sample Num: 29120, Cur Loss: 0.68730319, Cur Avg Loss: 0.47438747, Log Avg loss: 0.38650230, Global Avg Loss: 1.95201376, Time: 0.0209 Steps: 23110, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001830, Sample Num: 29280, Cur Loss: 0.40484473, Cur Avg Loss: 0.47445606, Log Avg loss: 0.48693883, Global Avg Loss: 1.95138008, Time: 0.0209 Steps: 23120, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001840, Sample Num: 29440, Cur Loss: 1.01624823, Cur Avg Loss: 0.47487729, Log Avg loss: 0.55196267, Global Avg Loss: 1.95077505, Time: 0.0210 Steps: 23130, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001850, Sample Num: 29600, Cur Loss: 0.63267595, Cur Avg Loss: 0.47455782, Log Avg loss: 0.41577541, Global Avg Loss: 1.95011170, Time: 0.0209 Steps: 23140, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001860, Sample Num: 29760, Cur Loss: 1.08352542, Cur Avg Loss: 0.47642393, Log Avg loss: 0.82165505, Global Avg Loss: 1.94962425, Time: 0.0209 Steps: 23150, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001870, Sample Num: 29920, Cur Loss: 0.25557524, Cur Avg Loss: 0.47737200, Log Avg loss: 0.65371304, Global Avg Loss: 1.94906470, Time: 0.0209 Steps: 23160, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001880, Sample Num: 30080, Cur Loss: 0.32549727, Cur Avg Loss: 0.47712338, Log Avg loss: 0.43063082, Global Avg Loss: 1.94840935, Time: 0.0209 Steps: 23170, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001890, Sample Num: 30240, Cur Loss: 0.42597592, Cur Avg Loss: 0.47673057, Log Avg loss: 0.40288301, Global Avg Loss: 1.94774260, Time: 0.0209 Steps: 23180, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001900, Sample Num: 30400, Cur Loss: 0.19699121, Cur Avg Loss: 0.47616980, Log Avg loss: 0.37018428, Global Avg Loss: 1.94706233, Time: 0.0209 Steps: 23190, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001910, Sample Num: 30560, Cur Loss: 0.92668426, Cur Avg Loss: 0.47574493, Log Avg loss: 0.39501968, Global Avg Loss: 1.94639334, Time: 0.0209 Steps: 23200, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001920, Sample Num: 30720, Cur Loss: 0.45167050, Cur Avg Loss: 0.47702611, Log Avg loss: 0.72173110, Global Avg Loss: 1.94586570, Time: 0.0209 Steps: 23210, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001930, Sample Num: 30880, Cur Loss: 0.42072546, Cur Avg Loss: 0.47742566, Log Avg loss: 0.55413923, Global Avg Loss: 1.94526634, Time: 0.0209 Steps: 23220, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001940, Sample Num: 31040, Cur Loss: 0.23404613, Cur Avg Loss: 0.47724057, Log Avg loss: 0.44151740, Global Avg Loss: 1.94461900, Time: 0.0209 Steps: 23230, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001950, Sample Num: 31200, Cur Loss: 0.51124334, Cur Avg Loss: 0.47662295, Log Avg loss: 0.35680511, Global Avg Loss: 1.94393578, Time: 0.0209 Steps: 23240, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001960, Sample Num: 31360, Cur Loss: 0.81810260, Cur Avg Loss: 0.47706849, Log Avg loss: 0.56394888, Global Avg Loss: 1.94334224, Time: 0.0209 Steps: 23250, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001970, Sample Num: 31520, Cur Loss: 0.39691162, Cur Avg Loss: 0.47692853, Log Avg loss: 0.44949679, Global Avg Loss: 1.94270000, Time: 0.0209 Steps: 23260, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001980, Sample Num: 31680, Cur Loss: 0.39976764, Cur Avg Loss: 0.47708939, Log Avg loss: 0.50877723, Global Avg Loss: 1.94208379, Time: 0.0209 Steps: 23270, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001990, Sample Num: 31840, Cur Loss: 0.45941269, Cur Avg Loss: 0.47612494, Log Avg loss: 0.28516572, Global Avg Loss: 1.94137205, Time: 0.0209 Steps: 23280, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002000, Sample Num: 32000, Cur Loss: 0.15274692, Cur Avg Loss: 0.47526389, Log Avg loss: 0.30391330, Global Avg Loss: 1.94066898, Time: 0.0209 Steps: 23290, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002010, Sample Num: 32160, Cur Loss: 0.61466444, Cur Avg Loss: 0.47537620, Log Avg loss: 0.49783971, Global Avg Loss: 1.94004974, Time: 0.0209 Steps: 23300, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002020, Sample Num: 32320, Cur Loss: 0.69938576, Cur Avg Loss: 0.47607116, Log Avg loss: 0.61575800, Global Avg Loss: 1.93948162, Time: 0.0209 Steps: 23310, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002030, Sample Num: 32480, Cur Loss: 0.97115099, Cur Avg Loss: 0.47626196, Log Avg loss: 0.51480332, Global Avg Loss: 1.93887069, Time: 0.0209 Steps: 23320, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002040, Sample Num: 32640, Cur Loss: 0.38721770, Cur Avg Loss: 0.47612884, Log Avg loss: 0.44910502, Global Avg Loss: 1.93823213, Time: 0.0209 Steps: 23330, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002050, Sample Num: 32800, Cur Loss: 0.27135250, Cur Avg Loss: 0.47683769, Log Avg loss: 0.62144389, Global Avg Loss: 1.93766795, Time: 0.0247 Steps: 23340, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002060, Sample Num: 32960, Cur Loss: 0.19217855, Cur Avg Loss: 0.47744427, Log Avg loss: 0.60179157, Global Avg Loss: 1.93709584, Time: 0.0210 Steps: 23350, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002070, Sample Num: 33120, Cur Loss: 0.26847351, Cur Avg Loss: 0.47716331, Log Avg loss: 0.41928713, Global Avg Loss: 1.93644610, Time: 0.0210 Steps: 23360, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002080, Sample Num: 33280, Cur Loss: 0.94847673, Cur Avg Loss: 0.47750625, Log Avg loss: 0.54849357, Global Avg Loss: 1.93585219, Time: 0.0210 Steps: 23370, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002090, Sample Num: 33440, Cur Loss: 1.40132046, Cur Avg Loss: 0.47790724, Log Avg loss: 0.56131359, Global Avg Loss: 1.93526428, Time: 0.0210 Steps: 23380, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002100, Sample Num: 33600, Cur Loss: 0.39879414, Cur Avg Loss: 0.47807529, Log Avg loss: 0.51319810, Global Avg Loss: 1.93465630, Time: 0.0210 Steps: 23390, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002110, Sample Num: 33760, Cur Loss: 0.19933225, Cur Avg Loss: 0.47716401, Log Avg loss: 0.28579475, Global Avg Loss: 1.93395166, Time: 0.0210 Steps: 23400, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002120, Sample Num: 33920, Cur Loss: 0.57344127, Cur Avg Loss: 0.47694635, Log Avg loss: 0.43102078, Global Avg Loss: 1.93330966, Time: 0.0209 Steps: 23410, Updated lr: 0.000079 ***** Running evaluation checkpoint-23419 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-23419 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.783848, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.404621, "eval_total_loss": 284.448721, "eval_mae": 0.470116, "eval_mse": 0.404727, "eval_r2": 0.742729, "eval_sp_statistic": 0.836753, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.865482, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.777661, "test_total_loss": 390.385723, "test_mae": 0.57684, "test_mse": 0.777903, "test_r2": 0.497935, "test_sp_statistic": 0.762092, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.8061, "test_ps_pvalue": 0.0, "lr": 7.873968705547652e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.932726476169689, "train_cur_epoch_loss": 1014.8685714825988, "train_cur_epoch_avg_loss": 0.47668791521023895, "train_cur_epoch_time": 44.78384828567505, "train_cur_epoch_avg_time": 0.02103515654564352, "epoch": 11, "step": 23419} ################################################## Training, Epoch: 0012, Batch: 000001, Sample Num: 16, Cur Loss: 0.28806421, Cur Avg Loss: 0.28806421, Log Avg loss: 0.40303683, Global Avg Loss: 1.93265625, Time: 0.0245 Steps: 23420, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000011, Sample Num: 176, Cur Loss: 0.33941635, Cur Avg Loss: 0.44217588, Log Avg loss: 0.45758705, Global Avg Loss: 1.93202669, Time: 0.0209 Steps: 23430, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000021, Sample Num: 336, Cur Loss: 0.22273652, Cur Avg Loss: 0.45034172, Log Avg loss: 0.45932414, Global Avg Loss: 1.93139840, Time: 0.0209 Steps: 23440, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000031, Sample Num: 496, Cur Loss: 1.08237743, Cur Avg Loss: 0.46023294, Log Avg loss: 0.48100450, Global Avg Loss: 1.93077990, Time: 0.0209 Steps: 23450, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000041, Sample Num: 656, Cur Loss: 0.55043387, Cur Avg Loss: 0.45376176, Log Avg loss: 0.43370109, Global Avg Loss: 1.93014176, Time: 0.0209 Steps: 23460, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000051, Sample Num: 816, Cur Loss: 0.97239190, Cur Avg Loss: 0.43796314, Log Avg loss: 0.37318879, Global Avg Loss: 1.92947838, Time: 0.0209 Steps: 23470, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000061, Sample Num: 976, Cur Loss: 0.22618032, Cur Avg Loss: 0.43002479, Log Avg loss: 0.38953921, Global Avg Loss: 1.92882252, Time: 0.0209 Steps: 23480, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000071, Sample Num: 1136, Cur Loss: 0.17630833, Cur Avg Loss: 0.42242625, Log Avg loss: 0.37607517, Global Avg Loss: 1.92816150, Time: 0.0209 Steps: 23490, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000081, Sample Num: 1296, Cur Loss: 0.65425050, Cur Avg Loss: 0.42329445, Log Avg loss: 0.42945864, Global Avg Loss: 1.92752375, Time: 0.0209 Steps: 23500, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000091, Sample Num: 1456, Cur Loss: 0.71368062, Cur Avg Loss: 0.43251735, Log Avg loss: 0.50722290, Global Avg Loss: 1.92691963, Time: 0.0209 Steps: 23510, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000101, Sample Num: 1616, Cur Loss: 0.28935942, Cur Avg Loss: 0.44580594, Log Avg loss: 0.56673205, Global Avg Loss: 1.92634132, Time: 0.0209 Steps: 23520, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000111, Sample Num: 1776, Cur Loss: 0.20837760, Cur Avg Loss: 0.45221597, Log Avg loss: 0.51695730, Global Avg Loss: 1.92574234, Time: 0.0210 Steps: 23530, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000121, Sample Num: 1936, Cur Loss: 0.65010977, Cur Avg Loss: 0.47486681, Log Avg loss: 0.72629118, Global Avg Loss: 1.92523281, Time: 0.0210 Steps: 23540, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000131, Sample Num: 2096, Cur Loss: 0.26775247, Cur Avg Loss: 0.48817562, Log Avg loss: 0.64921220, Global Avg Loss: 1.92469097, Time: 0.0209 Steps: 23550, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000141, Sample Num: 2256, Cur Loss: 0.46366024, Cur Avg Loss: 0.49070529, Log Avg loss: 0.52384393, Global Avg Loss: 1.92409638, Time: 0.0208 Steps: 23560, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000151, Sample Num: 2416, Cur Loss: 0.33633548, Cur Avg Loss: 0.48833872, Log Avg loss: 0.45497008, Global Avg Loss: 1.92347308, Time: 0.0208 Steps: 23570, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000161, Sample Num: 2576, Cur Loss: 0.33796704, Cur Avg Loss: 0.48457697, Log Avg loss: 0.42777454, Global Avg Loss: 1.92283877, Time: 0.0209 Steps: 23580, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000171, Sample Num: 2736, Cur Loss: 0.18194106, Cur Avg Loss: 0.48372885, Log Avg loss: 0.47007422, Global Avg Loss: 1.92222293, Time: 0.0208 Steps: 23590, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000181, Sample Num: 2896, Cur Loss: 0.43208760, Cur Avg Loss: 0.48198743, Log Avg loss: 0.45220905, Global Avg Loss: 1.92160005, Time: 0.0208 Steps: 23600, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000191, Sample Num: 3056, Cur Loss: 0.17543590, Cur Avg Loss: 0.47637152, Log Avg loss: 0.37472352, Global Avg Loss: 1.92094487, Time: 0.0208 Steps: 23610, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000201, Sample Num: 3216, Cur Loss: 0.17349926, Cur Avg Loss: 0.46949631, Log Avg loss: 0.33817982, Global Avg Loss: 1.92027477, Time: 0.0208 Steps: 23620, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000211, Sample Num: 3376, Cur Loss: 0.48812696, Cur Avg Loss: 0.46817011, Log Avg loss: 0.44151345, Global Avg Loss: 1.91964897, Time: 0.0209 Steps: 23630, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000221, Sample Num: 3536, Cur Loss: 0.88090658, Cur Avg Loss: 0.47716763, Log Avg loss: 0.66701539, Global Avg Loss: 1.91911909, Time: 0.0208 Steps: 23640, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000231, Sample Num: 3696, Cur Loss: 0.24145344, Cur Avg Loss: 0.47822585, Log Avg loss: 0.50161244, Global Avg Loss: 1.91851973, Time: 0.0208 Steps: 23650, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000241, Sample Num: 3856, Cur Loss: 0.32584053, Cur Avg Loss: 0.47543496, Log Avg loss: 0.41096543, Global Avg Loss: 1.91788255, Time: 0.0209 Steps: 23660, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000251, Sample Num: 4016, Cur Loss: 0.20024106, Cur Avg Loss: 0.47068930, Log Avg loss: 0.35631882, Global Avg Loss: 1.91722283, Time: 0.0209 Steps: 23670, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000261, Sample Num: 4176, Cur Loss: 0.25653774, Cur Avg Loss: 0.46538418, Log Avg loss: 0.33222577, Global Avg Loss: 1.91655349, Time: 0.0247 Steps: 23680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000271, Sample Num: 4336, Cur Loss: 0.06567895, Cur Avg Loss: 0.45788926, Log Avg loss: 0.26227197, Global Avg Loss: 1.91585519, Time: 0.0210 Steps: 23690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000281, Sample Num: 4496, Cur Loss: 0.68217540, Cur Avg Loss: 0.45352141, Log Avg loss: 0.33515267, Global Avg Loss: 1.91518822, Time: 0.0210 Steps: 23700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000291, Sample Num: 4656, Cur Loss: 0.35792232, Cur Avg Loss: 0.44613836, Log Avg loss: 0.23867463, Global Avg Loss: 1.91448113, Time: 0.0209 Steps: 23710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000301, Sample Num: 4816, Cur Loss: 0.40019763, Cur Avg Loss: 0.44222366, Log Avg loss: 0.32830568, Global Avg Loss: 1.91381242, Time: 0.0210 Steps: 23720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000311, Sample Num: 4976, Cur Loss: 0.21845663, Cur Avg Loss: 0.44149357, Log Avg loss: 0.41951813, Global Avg Loss: 1.91318272, Time: 0.0209 Steps: 23730, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000321, Sample Num: 5136, Cur Loss: 0.17481279, Cur Avg Loss: 0.44046283, Log Avg loss: 0.40840664, Global Avg Loss: 1.91254886, Time: 0.0210 Steps: 23740, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000331, Sample Num: 5296, Cur Loss: 0.86660838, Cur Avg Loss: 0.44361275, Log Avg loss: 0.54472523, Global Avg Loss: 1.91197293, Time: 0.0210 Steps: 23750, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000341, Sample Num: 5456, Cur Loss: 0.19842139, Cur Avg Loss: 0.44389291, Log Avg loss: 0.45316627, Global Avg Loss: 1.91135896, Time: 0.0209 Steps: 23760, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000351, Sample Num: 5616, Cur Loss: 0.56694424, Cur Avg Loss: 0.44256973, Log Avg loss: 0.39744933, Global Avg Loss: 1.91072206, Time: 0.0210 Steps: 23770, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000361, Sample Num: 5776, Cur Loss: 0.62987489, Cur Avg Loss: 0.44603836, Log Avg loss: 0.56778702, Global Avg Loss: 1.91015733, Time: 0.0210 Steps: 23780, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000371, Sample Num: 5936, Cur Loss: 0.57734495, Cur Avg Loss: 0.44814816, Log Avg loss: 0.52431196, Global Avg Loss: 1.90957479, Time: 0.0210 Steps: 23790, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000381, Sample Num: 6096, Cur Loss: 0.51637185, Cur Avg Loss: 0.44945849, Log Avg loss: 0.49807169, Global Avg Loss: 1.90898172, Time: 0.0209 Steps: 23800, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000391, Sample Num: 6256, Cur Loss: 0.71349132, Cur Avg Loss: 0.44934315, Log Avg loss: 0.44494887, Global Avg Loss: 1.90836684, Time: 0.0210 Steps: 23810, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000401, Sample Num: 6416, Cur Loss: 0.12286317, Cur Avg Loss: 0.44781549, Log Avg loss: 0.38808411, Global Avg Loss: 1.90772860, Time: 0.0210 Steps: 23820, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000411, Sample Num: 6576, Cur Loss: 0.26896468, Cur Avg Loss: 0.44826398, Log Avg loss: 0.46624821, Global Avg Loss: 1.90712370, Time: 0.0210 Steps: 23830, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000421, Sample Num: 6736, Cur Loss: 0.51392806, Cur Avg Loss: 0.44670872, Log Avg loss: 0.38278771, Global Avg Loss: 1.90648430, Time: 0.0210 Steps: 23840, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000431, Sample Num: 6896, Cur Loss: 0.41549519, Cur Avg Loss: 0.44807418, Log Avg loss: 0.50555988, Global Avg Loss: 1.90589691, Time: 0.0210 Steps: 23850, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000441, Sample Num: 7056, Cur Loss: 0.26324654, Cur Avg Loss: 0.44644244, Log Avg loss: 0.37611436, Global Avg Loss: 1.90525576, Time: 0.0210 Steps: 23860, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000451, Sample Num: 7216, Cur Loss: 0.63084352, Cur Avg Loss: 0.44798577, Log Avg loss: 0.51604662, Global Avg Loss: 1.90467377, Time: 0.0209 Steps: 23870, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000461, Sample Num: 7376, Cur Loss: 0.32446748, Cur Avg Loss: 0.44799954, Log Avg loss: 0.44862069, Global Avg Loss: 1.90406403, Time: 0.0210 Steps: 23880, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000471, Sample Num: 7536, Cur Loss: 0.16100562, Cur Avg Loss: 0.44699253, Log Avg loss: 0.40056958, Global Avg Loss: 1.90343469, Time: 0.0209 Steps: 23890, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000481, Sample Num: 7696, Cur Loss: 0.29250944, Cur Avg Loss: 0.44722228, Log Avg loss: 0.45804323, Global Avg Loss: 1.90282993, Time: 0.0210 Steps: 23900, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000491, Sample Num: 7856, Cur Loss: 0.22614460, Cur Avg Loss: 0.44471234, Log Avg loss: 0.32398419, Global Avg Loss: 1.90216960, Time: 0.0210 Steps: 23910, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000501, Sample Num: 8016, Cur Loss: 0.26304436, Cur Avg Loss: 0.44146449, Log Avg loss: 0.28199507, Global Avg Loss: 1.90149227, Time: 0.0209 Steps: 23920, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000511, Sample Num: 8176, Cur Loss: 0.22699472, Cur Avg Loss: 0.44069152, Log Avg loss: 0.40196568, Global Avg Loss: 1.90086564, Time: 0.0210 Steps: 23930, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000521, Sample Num: 8336, Cur Loss: 0.30046329, Cur Avg Loss: 0.43925373, Log Avg loss: 0.36578298, Global Avg Loss: 1.90022442, Time: 0.0210 Steps: 23940, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000531, Sample Num: 8496, Cur Loss: 0.47387961, Cur Avg Loss: 0.43833556, Log Avg loss: 0.39049846, Global Avg Loss: 1.89959405, Time: 0.0211 Steps: 23950, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000541, Sample Num: 8656, Cur Loss: 0.40998209, Cur Avg Loss: 0.43904752, Log Avg loss: 0.47685290, Global Avg Loss: 1.89900025, Time: 0.0210 Steps: 23960, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000551, Sample Num: 8816, Cur Loss: 0.30410141, Cur Avg Loss: 0.43853396, Log Avg loss: 0.41075047, Global Avg Loss: 1.89837937, Time: 0.0210 Steps: 23970, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000561, Sample Num: 8976, Cur Loss: 0.43863946, Cur Avg Loss: 0.43524889, Log Avg loss: 0.25424118, Global Avg Loss: 1.89769374, Time: 0.0210 Steps: 23980, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000571, Sample Num: 9136, Cur Loss: 0.12632933, Cur Avg Loss: 0.43504487, Log Avg loss: 0.42359970, Global Avg Loss: 1.89707928, Time: 0.0210 Steps: 23990, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000581, Sample Num: 9296, Cur Loss: 0.61107898, Cur Avg Loss: 0.43696641, Log Avg loss: 0.54668628, Global Avg Loss: 1.89651662, Time: 0.0210 Steps: 24000, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000591, Sample Num: 9456, Cur Loss: 0.16435760, Cur Avg Loss: 0.43455911, Log Avg loss: 0.29469484, Global Avg Loss: 1.89584947, Time: 0.0210 Steps: 24010, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000601, Sample Num: 9616, Cur Loss: 0.57645786, Cur Avg Loss: 0.43311309, Log Avg loss: 0.34765320, Global Avg Loss: 1.89520493, Time: 0.0210 Steps: 24020, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000611, Sample Num: 9776, Cur Loss: 0.28545702, Cur Avg Loss: 0.43191002, Log Avg loss: 0.35960588, Global Avg Loss: 1.89456589, Time: 0.0209 Steps: 24030, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000621, Sample Num: 9936, Cur Loss: 0.45748031, Cur Avg Loss: 0.43244197, Log Avg loss: 0.46494414, Global Avg Loss: 1.89397121, Time: 0.0210 Steps: 24040, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000631, Sample Num: 10096, Cur Loss: 0.23898721, Cur Avg Loss: 0.43199977, Log Avg loss: 0.40453916, Global Avg Loss: 1.89335190, Time: 0.0210 Steps: 24050, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000641, Sample Num: 10256, Cur Loss: 0.59796917, Cur Avg Loss: 0.43149005, Log Avg loss: 0.39932648, Global Avg Loss: 1.89273094, Time: 0.0211 Steps: 24060, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000651, Sample Num: 10416, Cur Loss: 0.31192264, Cur Avg Loss: 0.43026226, Log Avg loss: 0.35156060, Global Avg Loss: 1.89209066, Time: 0.0210 Steps: 24070, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000661, Sample Num: 10576, Cur Loss: 0.20284525, Cur Avg Loss: 0.43080311, Log Avg loss: 0.46601284, Global Avg Loss: 1.89149843, Time: 0.0210 Steps: 24080, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000671, Sample Num: 10736, Cur Loss: 0.29005724, Cur Avg Loss: 0.42928550, Log Avg loss: 0.32897157, Global Avg Loss: 1.89084981, Time: 0.0210 Steps: 24090, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000681, Sample Num: 10896, Cur Loss: 0.20503291, Cur Avg Loss: 0.42799533, Log Avg loss: 0.34142502, Global Avg Loss: 1.89020689, Time: 0.0210 Steps: 24100, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000691, Sample Num: 11056, Cur Loss: 0.32697222, Cur Avg Loss: 0.42971875, Log Avg loss: 0.54708365, Global Avg Loss: 1.88964981, Time: 0.0211 Steps: 24110, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000701, Sample Num: 11216, Cur Loss: 0.13456038, Cur Avg Loss: 0.42907214, Log Avg loss: 0.38439101, Global Avg Loss: 1.88902574, Time: 0.0210 Steps: 24120, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000711, Sample Num: 11376, Cur Loss: 0.40858892, Cur Avg Loss: 0.42747127, Log Avg loss: 0.31525052, Global Avg Loss: 1.88837354, Time: 0.0210 Steps: 24130, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000721, Sample Num: 11536, Cur Loss: 0.17602414, Cur Avg Loss: 0.42653065, Log Avg loss: 0.35965268, Global Avg Loss: 1.88774026, Time: 0.0210 Steps: 24140, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000731, Sample Num: 11696, Cur Loss: 0.48314992, Cur Avg Loss: 0.42604249, Log Avg loss: 0.39084561, Global Avg Loss: 1.88712043, Time: 0.0210 Steps: 24150, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000741, Sample Num: 11856, Cur Loss: 0.24628660, Cur Avg Loss: 0.42477625, Log Avg loss: 0.33221455, Global Avg Loss: 1.88647684, Time: 0.0209 Steps: 24160, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000751, Sample Num: 12016, Cur Loss: 0.43248963, Cur Avg Loss: 0.42628583, Log Avg loss: 0.53814594, Global Avg Loss: 1.88591899, Time: 0.0210 Steps: 24170, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000761, Sample Num: 12176, Cur Loss: 0.30767477, Cur Avg Loss: 0.42588792, Log Avg loss: 0.39600471, Global Avg Loss: 1.88530281, Time: 0.0211 Steps: 24180, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000771, Sample Num: 12336, Cur Loss: 0.24442278, Cur Avg Loss: 0.42464992, Log Avg loss: 0.33043801, Global Avg Loss: 1.88466004, Time: 0.0247 Steps: 24190, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000781, Sample Num: 12496, Cur Loss: 0.93256581, Cur Avg Loss: 0.42452474, Log Avg loss: 0.41487294, Global Avg Loss: 1.88405269, Time: 0.0210 Steps: 24200, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000791, Sample Num: 12656, Cur Loss: 0.32137975, Cur Avg Loss: 0.42483661, Log Avg loss: 0.44919416, Global Avg Loss: 1.88346002, Time: 0.0210 Steps: 24210, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000801, Sample Num: 12816, Cur Loss: 0.37362790, Cur Avg Loss: 0.42578727, Log Avg loss: 0.50098419, Global Avg Loss: 1.88288922, Time: 0.0209 Steps: 24220, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000811, Sample Num: 12976, Cur Loss: 0.67321289, Cur Avg Loss: 0.42707423, Log Avg loss: 0.53016008, Global Avg Loss: 1.88233093, Time: 0.0210 Steps: 24230, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000821, Sample Num: 13136, Cur Loss: 0.64471698, Cur Avg Loss: 0.42655232, Log Avg loss: 0.38422535, Global Avg Loss: 1.88171290, Time: 0.0210 Steps: 24240, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000831, Sample Num: 13296, Cur Loss: 0.31075442, Cur Avg Loss: 0.42598155, Log Avg loss: 0.37912127, Global Avg Loss: 1.88109328, Time: 0.0209 Steps: 24250, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000841, Sample Num: 13456, Cur Loss: 0.76082683, Cur Avg Loss: 0.42704049, Log Avg loss: 0.51503824, Global Avg Loss: 1.88053019, Time: 0.0210 Steps: 24260, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000851, Sample Num: 13616, Cur Loss: 0.57681537, Cur Avg Loss: 0.42541275, Log Avg loss: 0.28851992, Global Avg Loss: 1.87987423, Time: 0.0210 Steps: 24270, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000861, Sample Num: 13776, Cur Loss: 0.10718968, Cur Avg Loss: 0.42480347, Log Avg loss: 0.37295388, Global Avg Loss: 1.87925359, Time: 0.0210 Steps: 24280, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000871, Sample Num: 13936, Cur Loss: 0.59287995, Cur Avg Loss: 0.42461858, Log Avg loss: 0.40869972, Global Avg Loss: 1.87864817, Time: 0.0209 Steps: 24290, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000881, Sample Num: 14096, Cur Loss: 0.55762833, Cur Avg Loss: 0.42467997, Log Avg loss: 0.43002666, Global Avg Loss: 1.87805203, Time: 0.0210 Steps: 24300, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000891, Sample Num: 14256, Cur Loss: 0.46293917, Cur Avg Loss: 0.42551674, Log Avg loss: 0.49923586, Global Avg Loss: 1.87748485, Time: 0.0210 Steps: 24310, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000901, Sample Num: 14416, Cur Loss: 0.62638825, Cur Avg Loss: 0.42531135, Log Avg loss: 0.40701184, Global Avg Loss: 1.87688022, Time: 0.0210 Steps: 24320, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000911, Sample Num: 14576, Cur Loss: 0.33624557, Cur Avg Loss: 0.42530419, Log Avg loss: 0.42465866, Global Avg Loss: 1.87628333, Time: 0.0210 Steps: 24330, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000921, Sample Num: 14736, Cur Loss: 0.12176378, Cur Avg Loss: 0.42501750, Log Avg loss: 0.39890044, Global Avg Loss: 1.87567635, Time: 0.0209 Steps: 24340, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000931, Sample Num: 14896, Cur Loss: 0.64334369, Cur Avg Loss: 0.42670821, Log Avg loss: 0.58242183, Global Avg Loss: 1.87514524, Time: 0.0210 Steps: 24350, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000941, Sample Num: 15056, Cur Loss: 0.50917202, Cur Avg Loss: 0.42720597, Log Avg loss: 0.47354819, Global Avg Loss: 1.87456988, Time: 0.0210 Steps: 24360, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000951, Sample Num: 15216, Cur Loss: 0.13570961, Cur Avg Loss: 0.42639818, Log Avg loss: 0.35038442, Global Avg Loss: 1.87394444, Time: 0.0210 Steps: 24370, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000961, Sample Num: 15376, Cur Loss: 0.41369128, Cur Avg Loss: 0.42792018, Log Avg loss: 0.57266304, Global Avg Loss: 1.87341069, Time: 0.0209 Steps: 24380, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000971, Sample Num: 15536, Cur Loss: 0.45713219, Cur Avg Loss: 0.42996695, Log Avg loss: 0.62666104, Global Avg Loss: 1.87289952, Time: 0.0210 Steps: 24390, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000981, Sample Num: 15696, Cur Loss: 0.62071645, Cur Avg Loss: 0.43060154, Log Avg loss: 0.49222027, Global Avg Loss: 1.87233367, Time: 0.0209 Steps: 24400, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000991, Sample Num: 15856, Cur Loss: 0.18860543, Cur Avg Loss: 0.43122514, Log Avg loss: 0.49240089, Global Avg Loss: 1.87176835, Time: 0.0209 Steps: 24410, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001001, Sample Num: 16016, Cur Loss: 0.25478929, Cur Avg Loss: 0.43018927, Log Avg loss: 0.32753365, Global Avg Loss: 1.87113599, Time: 0.0210 Steps: 24420, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001011, Sample Num: 16176, Cur Loss: 0.11124249, Cur Avg Loss: 0.42873508, Log Avg loss: 0.28317149, Global Avg Loss: 1.87048598, Time: 0.0209 Steps: 24430, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001021, Sample Num: 16336, Cur Loss: 0.44518727, Cur Avg Loss: 0.42852949, Log Avg loss: 0.40774356, Global Avg Loss: 1.86988748, Time: 0.0210 Steps: 24440, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001031, Sample Num: 16496, Cur Loss: 0.46923178, Cur Avg Loss: 0.42773816, Log Avg loss: 0.34694346, Global Avg Loss: 1.86926460, Time: 0.0209 Steps: 24450, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001041, Sample Num: 16656, Cur Loss: 0.46252012, Cur Avg Loss: 0.42741852, Log Avg loss: 0.39446404, Global Avg Loss: 1.86866165, Time: 0.0208 Steps: 24460, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001051, Sample Num: 16816, Cur Loss: 0.32705986, Cur Avg Loss: 0.42633282, Log Avg loss: 0.31331145, Global Avg Loss: 1.86802604, Time: 0.0208 Steps: 24470, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001061, Sample Num: 16976, Cur Loss: 0.76629412, Cur Avg Loss: 0.42751072, Log Avg loss: 0.55130843, Global Avg Loss: 1.86748816, Time: 0.0209 Steps: 24480, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001071, Sample Num: 17136, Cur Loss: 0.29223916, Cur Avg Loss: 0.42824408, Log Avg loss: 0.50605357, Global Avg Loss: 1.86693225, Time: 0.0209 Steps: 24490, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001081, Sample Num: 17296, Cur Loss: 0.36735344, Cur Avg Loss: 0.42724545, Log Avg loss: 0.32029157, Global Avg Loss: 1.86630097, Time: 0.0208 Steps: 24500, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001091, Sample Num: 17456, Cur Loss: 0.44947654, Cur Avg Loss: 0.42653058, Log Avg loss: 0.34925305, Global Avg Loss: 1.86568202, Time: 0.0208 Steps: 24510, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001101, Sample Num: 17616, Cur Loss: 0.32764032, Cur Avg Loss: 0.42706774, Log Avg loss: 0.48567239, Global Avg Loss: 1.86511921, Time: 0.0208 Steps: 24520, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001111, Sample Num: 17776, Cur Loss: 1.13832879, Cur Avg Loss: 0.42779247, Log Avg loss: 0.50758532, Global Avg Loss: 1.86456579, Time: 0.0209 Steps: 24530, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001121, Sample Num: 17936, Cur Loss: 0.43440884, Cur Avg Loss: 0.42754644, Log Avg loss: 0.40021213, Global Avg Loss: 1.86396907, Time: 0.0208 Steps: 24540, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001131, Sample Num: 18096, Cur Loss: 0.45523161, Cur Avg Loss: 0.42712123, Log Avg loss: 0.37945541, Global Avg Loss: 1.86336438, Time: 0.0209 Steps: 24550, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001141, Sample Num: 18256, Cur Loss: 0.60020041, Cur Avg Loss: 0.42657737, Log Avg loss: 0.36506627, Global Avg Loss: 1.86275432, Time: 0.0209 Steps: 24560, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001151, Sample Num: 18416, Cur Loss: 0.20048711, Cur Avg Loss: 0.42665082, Log Avg loss: 0.43503129, Global Avg Loss: 1.86217324, Time: 0.0208 Steps: 24570, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001161, Sample Num: 18576, Cur Loss: 0.14341795, Cur Avg Loss: 0.42564247, Log Avg loss: 0.30958222, Global Avg Loss: 1.86154159, Time: 0.0208 Steps: 24580, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001171, Sample Num: 18736, Cur Loss: 0.73380280, Cur Avg Loss: 0.42614569, Log Avg loss: 0.48456893, Global Avg Loss: 1.86098162, Time: 0.0208 Steps: 24590, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001181, Sample Num: 18896, Cur Loss: 0.26583570, Cur Avg Loss: 0.42522139, Log Avg loss: 0.31698630, Global Avg Loss: 1.86035398, Time: 0.0208 Steps: 24600, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001191, Sample Num: 19056, Cur Loss: 0.81050396, Cur Avg Loss: 0.42855750, Log Avg loss: 0.82255144, Global Avg Loss: 1.85993228, Time: 0.0209 Steps: 24610, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001201, Sample Num: 19216, Cur Loss: 1.25666475, Cur Avg Loss: 0.42992317, Log Avg loss: 0.59257550, Global Avg Loss: 1.85941751, Time: 0.0209 Steps: 24620, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001211, Sample Num: 19376, Cur Loss: 0.16558206, Cur Avg Loss: 0.43057651, Log Avg loss: 0.50904254, Global Avg Loss: 1.85886924, Time: 0.0208 Steps: 24630, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001221, Sample Num: 19536, Cur Loss: 0.31529006, Cur Avg Loss: 0.43154430, Log Avg loss: 0.54874363, Global Avg Loss: 1.85833754, Time: 0.0208 Steps: 24640, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001231, Sample Num: 19696, Cur Loss: 0.09785023, Cur Avg Loss: 0.43282270, Log Avg loss: 0.58891526, Global Avg Loss: 1.85782256, Time: 0.0208 Steps: 24650, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001241, Sample Num: 19856, Cur Loss: 0.56903422, Cur Avg Loss: 0.43220425, Log Avg loss: 0.35607332, Global Avg Loss: 1.85721358, Time: 0.0208 Steps: 24660, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001251, Sample Num: 20016, Cur Loss: 0.44777563, Cur Avg Loss: 0.43253093, Log Avg loss: 0.47307152, Global Avg Loss: 1.85665251, Time: 0.0208 Steps: 24670, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001261, Sample Num: 20176, Cur Loss: 0.48760635, Cur Avg Loss: 0.43215048, Log Avg loss: 0.38455613, Global Avg Loss: 1.85605604, Time: 0.0208 Steps: 24680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001271, Sample Num: 20336, Cur Loss: 0.36542866, Cur Avg Loss: 0.43186869, Log Avg loss: 0.39633527, Global Avg Loss: 1.85546482, Time: 0.0208 Steps: 24690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001281, Sample Num: 20496, Cur Loss: 0.15403518, Cur Avg Loss: 0.43228832, Log Avg loss: 0.48562318, Global Avg Loss: 1.85491023, Time: 0.0246 Steps: 24700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001291, Sample Num: 20656, Cur Loss: 0.59435600, Cur Avg Loss: 0.43260266, Log Avg loss: 0.47286967, Global Avg Loss: 1.85435093, Time: 0.0209 Steps: 24710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001301, Sample Num: 20816, Cur Loss: 0.22348306, Cur Avg Loss: 0.43149999, Log Avg loss: 0.28914463, Global Avg Loss: 1.85371775, Time: 0.0209 Steps: 24720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001311, Sample Num: 20976, Cur Loss: 0.21151644, Cur Avg Loss: 0.43139809, Log Avg loss: 0.41814176, Global Avg Loss: 1.85313725, Time: 0.0209 Steps: 24730, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001321, Sample Num: 21136, Cur Loss: 0.32973757, Cur Avg Loss: 0.43096207, Log Avg loss: 0.37379985, Global Avg Loss: 1.85253930, Time: 0.0210 Steps: 24740, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001331, Sample Num: 21296, Cur Loss: 0.24590836, Cur Avg Loss: 0.43158241, Log Avg loss: 0.51352855, Global Avg Loss: 1.85199828, Time: 0.0209 Steps: 24750, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001341, Sample Num: 21456, Cur Loss: 1.02852452, Cur Avg Loss: 0.43192519, Log Avg loss: 0.47754965, Global Avg Loss: 1.85144318, Time: 0.0208 Steps: 24760, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001351, Sample Num: 21616, Cur Loss: 0.48089746, Cur Avg Loss: 0.43190309, Log Avg loss: 0.42893957, Global Avg Loss: 1.85086889, Time: 0.0209 Steps: 24770, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001361, Sample Num: 21776, Cur Loss: 0.76175225, Cur Avg Loss: 0.43350057, Log Avg loss: 0.64932012, Global Avg Loss: 1.85038400, Time: 0.0210 Steps: 24780, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001371, Sample Num: 21936, Cur Loss: 0.67520887, Cur Avg Loss: 0.43516451, Log Avg loss: 0.66162646, Global Avg Loss: 1.84990447, Time: 0.0210 Steps: 24790, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001381, Sample Num: 22096, Cur Loss: 1.08180833, Cur Avg Loss: 0.43569382, Log Avg loss: 0.50826277, Global Avg Loss: 1.84936349, Time: 0.0209 Steps: 24800, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001391, Sample Num: 22256, Cur Loss: 0.34830594, Cur Avg Loss: 0.43552733, Log Avg loss: 0.41253442, Global Avg Loss: 1.84878436, Time: 0.0209 Steps: 24810, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001401, Sample Num: 22416, Cur Loss: 0.62926847, Cur Avg Loss: 0.43546210, Log Avg loss: 0.42638927, Global Avg Loss: 1.84821127, Time: 0.0209 Steps: 24820, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001411, Sample Num: 22576, Cur Loss: 0.40026829, Cur Avg Loss: 0.43548193, Log Avg loss: 0.43825942, Global Avg Loss: 1.84764343, Time: 0.0209 Steps: 24830, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001421, Sample Num: 22736, Cur Loss: 0.39693016, Cur Avg Loss: 0.43550189, Log Avg loss: 0.43831857, Global Avg Loss: 1.84707607, Time: 0.0209 Steps: 24840, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001431, Sample Num: 22896, Cur Loss: 0.43956882, Cur Avg Loss: 0.43452839, Log Avg loss: 0.29619316, Global Avg Loss: 1.84645197, Time: 0.0209 Steps: 24850, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001441, Sample Num: 23056, Cur Loss: 0.23313257, Cur Avg Loss: 0.43396120, Log Avg loss: 0.35279752, Global Avg Loss: 1.84585114, Time: 0.0209 Steps: 24860, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001451, Sample Num: 23216, Cur Loss: 0.76554763, Cur Avg Loss: 0.43464403, Log Avg loss: 0.53303969, Global Avg Loss: 1.84532327, Time: 0.0209 Steps: 24870, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001461, Sample Num: 23376, Cur Loss: 0.29191500, Cur Avg Loss: 0.43506812, Log Avg loss: 0.49660333, Global Avg Loss: 1.84478118, Time: 0.0209 Steps: 24880, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001471, Sample Num: 23536, Cur Loss: 0.23659615, Cur Avg Loss: 0.43581513, Log Avg loss: 0.54495378, Global Avg Loss: 1.84425896, Time: 0.0209 Steps: 24890, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001481, Sample Num: 23696, Cur Loss: 0.26044279, Cur Avg Loss: 0.43655570, Log Avg loss: 0.54549287, Global Avg Loss: 1.84373736, Time: 0.0209 Steps: 24900, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001491, Sample Num: 23856, Cur Loss: 0.41042572, Cur Avg Loss: 0.43757092, Log Avg loss: 0.58792528, Global Avg Loss: 1.84323322, Time: 0.0209 Steps: 24910, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001501, Sample Num: 24016, Cur Loss: 0.98995835, Cur Avg Loss: 0.43802895, Log Avg loss: 0.50632075, Global Avg Loss: 1.84269674, Time: 0.0209 Steps: 24920, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001511, Sample Num: 24176, Cur Loss: 0.75616670, Cur Avg Loss: 0.43858781, Log Avg loss: 0.52247222, Global Avg Loss: 1.84216717, Time: 0.0209 Steps: 24930, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001521, Sample Num: 24336, Cur Loss: 0.42535806, Cur Avg Loss: 0.43837720, Log Avg loss: 0.40655452, Global Avg Loss: 1.84159154, Time: 0.0209 Steps: 24940, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001531, Sample Num: 24496, Cur Loss: 0.15593769, Cur Avg Loss: 0.43780117, Log Avg loss: 0.35018648, Global Avg Loss: 1.84099378, Time: 0.0209 Steps: 24950, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001541, Sample Num: 24656, Cur Loss: 0.98492098, Cur Avg Loss: 0.43771511, Log Avg loss: 0.42454043, Global Avg Loss: 1.84042630, Time: 0.0210 Steps: 24960, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001551, Sample Num: 24816, Cur Loss: 0.17451678, Cur Avg Loss: 0.43789630, Log Avg loss: 0.46581772, Global Avg Loss: 1.83987579, Time: 0.0212 Steps: 24970, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001561, Sample Num: 24976, Cur Loss: 0.31290871, Cur Avg Loss: 0.43774214, Log Avg loss: 0.41383126, Global Avg Loss: 1.83930492, Time: 0.0210 Steps: 24980, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001571, Sample Num: 25136, Cur Loss: 0.57912636, Cur Avg Loss: 0.43741880, Log Avg loss: 0.38694586, Global Avg Loss: 1.83872374, Time: 0.0209 Steps: 24990, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001581, Sample Num: 25296, Cur Loss: 0.05725610, Cur Avg Loss: 0.43760024, Log Avg loss: 0.46610464, Global Avg Loss: 1.83817469, Time: 0.0209 Steps: 25000, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001591, Sample Num: 25456, Cur Loss: 0.75493902, Cur Avg Loss: 0.43756044, Log Avg loss: 0.43126756, Global Avg Loss: 1.83761216, Time: 0.0210 Steps: 25010, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001601, Sample Num: 25616, Cur Loss: 0.14725280, Cur Avg Loss: 0.43695564, Log Avg loss: 0.34073271, Global Avg Loss: 1.83701388, Time: 0.0210 Steps: 25020, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001611, Sample Num: 25776, Cur Loss: 0.29541111, Cur Avg Loss: 0.43669716, Log Avg loss: 0.39531433, Global Avg Loss: 1.83643789, Time: 0.0209 Steps: 25030, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001621, Sample Num: 25936, Cur Loss: 0.58863032, Cur Avg Loss: 0.43606425, Log Avg loss: 0.33410165, Global Avg Loss: 1.83583792, Time: 0.0210 Steps: 25040, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001631, Sample Num: 26096, Cur Loss: 0.67902130, Cur Avg Loss: 0.43650478, Log Avg loss: 0.50791442, Global Avg Loss: 1.83530781, Time: 0.0209 Steps: 25050, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001641, Sample Num: 26256, Cur Loss: 0.27196121, Cur Avg Loss: 0.43553482, Log Avg loss: 0.27733442, Global Avg Loss: 1.83468611, Time: 0.0210 Steps: 25060, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001651, Sample Num: 26416, Cur Loss: 0.51633358, Cur Avg Loss: 0.43603095, Log Avg loss: 0.51744639, Global Avg Loss: 1.83416069, Time: 0.0209 Steps: 25070, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001661, Sample Num: 26576, Cur Loss: 0.36157161, Cur Avg Loss: 0.43626610, Log Avg loss: 0.47508962, Global Avg Loss: 1.83361879, Time: 0.0209 Steps: 25080, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001671, Sample Num: 26736, Cur Loss: 0.31369433, Cur Avg Loss: 0.43648989, Log Avg loss: 0.47366165, Global Avg Loss: 1.83307676, Time: 0.0209 Steps: 25090, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001681, Sample Num: 26896, Cur Loss: 0.27397102, Cur Avg Loss: 0.43681408, Log Avg loss: 0.49098655, Global Avg Loss: 1.83254206, Time: 0.0208 Steps: 25100, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001691, Sample Num: 27056, Cur Loss: 0.41107401, Cur Avg Loss: 0.43643086, Log Avg loss: 0.37201093, Global Avg Loss: 1.83196041, Time: 0.0209 Steps: 25110, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001701, Sample Num: 27216, Cur Loss: 0.21739262, Cur Avg Loss: 0.43660589, Log Avg loss: 0.46620352, Global Avg Loss: 1.83141672, Time: 0.0209 Steps: 25120, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001711, Sample Num: 27376, Cur Loss: 0.26087812, Cur Avg Loss: 0.43578564, Log Avg loss: 0.29626075, Global Avg Loss: 1.83080583, Time: 0.0208 Steps: 25130, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001721, Sample Num: 27536, Cur Loss: 0.22842687, Cur Avg Loss: 0.43505799, Log Avg loss: 0.31055713, Global Avg Loss: 1.83020112, Time: 0.0208 Steps: 25140, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001731, Sample Num: 27696, Cur Loss: 0.61137748, Cur Avg Loss: 0.43466121, Log Avg loss: 0.36637596, Global Avg Loss: 1.82961908, Time: 0.0209 Steps: 25150, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001741, Sample Num: 27856, Cur Loss: 0.38969126, Cur Avg Loss: 0.43419670, Log Avg loss: 0.35378897, Global Avg Loss: 1.82903250, Time: 0.0209 Steps: 25160, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001751, Sample Num: 28016, Cur Loss: 0.30228281, Cur Avg Loss: 0.43369207, Log Avg loss: 0.34583649, Global Avg Loss: 1.82844323, Time: 0.0209 Steps: 25170, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001761, Sample Num: 28176, Cur Loss: 0.29127151, Cur Avg Loss: 0.43318144, Log Avg loss: 0.34377035, Global Avg Loss: 1.82785361, Time: 0.0210 Steps: 25180, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001771, Sample Num: 28336, Cur Loss: 0.40235105, Cur Avg Loss: 0.43238047, Log Avg loss: 0.29132930, Global Avg Loss: 1.82724363, Time: 0.0209 Steps: 25190, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001781, Sample Num: 28496, Cur Loss: 0.26070353, Cur Avg Loss: 0.43166428, Log Avg loss: 0.30482716, Global Avg Loss: 1.82663950, Time: 0.0208 Steps: 25200, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001791, Sample Num: 28656, Cur Loss: 0.45252359, Cur Avg Loss: 0.43134118, Log Avg loss: 0.37379676, Global Avg Loss: 1.82606320, Time: 0.0208 Steps: 25210, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001801, Sample Num: 28816, Cur Loss: 0.34188193, Cur Avg Loss: 0.43209164, Log Avg loss: 0.56649906, Global Avg Loss: 1.82556377, Time: 0.0211 Steps: 25220, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001811, Sample Num: 28976, Cur Loss: 0.78126633, Cur Avg Loss: 0.43196618, Log Avg loss: 0.40937075, Global Avg Loss: 1.82500246, Time: 0.0210 Steps: 25230, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001821, Sample Num: 29136, Cur Loss: 0.38014477, Cur Avg Loss: 0.43171840, Log Avg loss: 0.38684531, Global Avg Loss: 1.82443267, Time: 0.0210 Steps: 25240, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001831, Sample Num: 29296, Cur Loss: 0.83123052, Cur Avg Loss: 0.43267005, Log Avg loss: 0.60596553, Global Avg Loss: 1.82395011, Time: 0.0210 Steps: 25250, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001841, Sample Num: 29456, Cur Loss: 0.49107635, Cur Avg Loss: 0.43298885, Log Avg loss: 0.49136254, Global Avg Loss: 1.82342256, Time: 0.0210 Steps: 25260, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001851, Sample Num: 29616, Cur Loss: 0.18201967, Cur Avg Loss: 0.43241439, Log Avg loss: 0.32665553, Global Avg Loss: 1.82283025, Time: 0.0210 Steps: 25270, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001861, Sample Num: 29776, Cur Loss: 0.28582251, Cur Avg Loss: 0.43377661, Log Avg loss: 0.68592268, Global Avg Loss: 1.82238052, Time: 0.0210 Steps: 25280, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001871, Sample Num: 29936, Cur Loss: 0.54560518, Cur Avg Loss: 0.43417040, Log Avg loss: 0.50745575, Global Avg Loss: 1.82186058, Time: 0.0210 Steps: 25290, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001881, Sample Num: 30096, Cur Loss: 0.20271170, Cur Avg Loss: 0.43405454, Log Avg loss: 0.41237603, Global Avg Loss: 1.82130348, Time: 0.0210 Steps: 25300, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001891, Sample Num: 30256, Cur Loss: 0.25445789, Cur Avg Loss: 0.43411079, Log Avg loss: 0.44469304, Global Avg Loss: 1.82075958, Time: 0.0210 Steps: 25310, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001901, Sample Num: 30416, Cur Loss: 0.28231055, Cur Avg Loss: 0.43396027, Log Avg loss: 0.40549522, Global Avg Loss: 1.82020062, Time: 0.0209 Steps: 25320, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001911, Sample Num: 30576, Cur Loss: 0.69531935, Cur Avg Loss: 0.43411916, Log Avg loss: 0.46432417, Global Avg Loss: 1.81966534, Time: 0.0209 Steps: 25330, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001921, Sample Num: 30736, Cur Loss: 0.32473889, Cur Avg Loss: 0.43334684, Log Avg loss: 0.28575763, Global Avg Loss: 1.81906001, Time: 0.0209 Steps: 25340, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001931, Sample Num: 30896, Cur Loss: 0.47572690, Cur Avg Loss: 0.43269879, Log Avg loss: 0.30820857, Global Avg Loss: 1.81846401, Time: 0.0209 Steps: 25350, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001941, Sample Num: 31056, Cur Loss: 0.74801052, Cur Avg Loss: 0.43314342, Log Avg loss: 0.51900197, Global Avg Loss: 1.81795161, Time: 0.0210 Steps: 25360, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001951, Sample Num: 31216, Cur Loss: 0.43704951, Cur Avg Loss: 0.43285112, Log Avg loss: 0.37611456, Global Avg Loss: 1.81738328, Time: 0.0210 Steps: 25370, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001961, Sample Num: 31376, Cur Loss: 0.25468686, Cur Avg Loss: 0.43265701, Log Avg loss: 0.39478549, Global Avg Loss: 1.81682276, Time: 0.0209 Steps: 25380, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001971, Sample Num: 31536, Cur Loss: 0.29491502, Cur Avg Loss: 0.43333603, Log Avg loss: 0.56649303, Global Avg Loss: 1.81633031, Time: 0.0210 Steps: 25390, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001981, Sample Num: 31696, Cur Loss: 0.31111878, Cur Avg Loss: 0.43339030, Log Avg loss: 0.44408686, Global Avg Loss: 1.81579006, Time: 0.0210 Steps: 25400, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001991, Sample Num: 31856, Cur Loss: 0.35363829, Cur Avg Loss: 0.43306202, Log Avg loss: 0.36802891, Global Avg Loss: 1.81522030, Time: 0.0210 Steps: 25410, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002001, Sample Num: 32016, Cur Loss: 0.33514771, Cur Avg Loss: 0.43309675, Log Avg loss: 0.44001269, Global Avg Loss: 1.81467931, Time: 0.0210 Steps: 25420, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002011, Sample Num: 32176, Cur Loss: 0.56900918, Cur Avg Loss: 0.43324415, Log Avg loss: 0.46273740, Global Avg Loss: 1.81414767, Time: 0.0209 Steps: 25430, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002021, Sample Num: 32336, Cur Loss: 0.21962836, Cur Avg Loss: 0.43303763, Log Avg loss: 0.39150690, Global Avg Loss: 1.81358846, Time: 0.0210 Steps: 25440, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002031, Sample Num: 32496, Cur Loss: 0.21256867, Cur Avg Loss: 0.43338772, Log Avg loss: 0.50414091, Global Avg Loss: 1.81307394, Time: 0.0210 Steps: 25450, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002041, Sample Num: 32656, Cur Loss: 0.58198565, Cur Avg Loss: 0.43448152, Log Avg loss: 0.65663201, Global Avg Loss: 1.81261972, Time: 0.0210 Steps: 25460, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002051, Sample Num: 32816, Cur Loss: 0.27969903, Cur Avg Loss: 0.43408251, Log Avg loss: 0.35264543, Global Avg Loss: 1.81204651, Time: 0.0247 Steps: 25470, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002061, Sample Num: 32976, Cur Loss: 0.21277225, Cur Avg Loss: 0.43331427, Log Avg loss: 0.27574726, Global Avg Loss: 1.81144357, Time: 0.0210 Steps: 25480, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002071, Sample Num: 33136, Cur Loss: 0.43299133, Cur Avg Loss: 0.43295176, Log Avg loss: 0.35823991, Global Avg Loss: 1.81087346, Time: 0.0209 Steps: 25490, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002081, Sample Num: 33296, Cur Loss: 0.57182527, Cur Avg Loss: 0.43292128, Log Avg loss: 0.42660773, Global Avg Loss: 1.81033061, Time: 0.0210 Steps: 25500, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002091, Sample Num: 33456, Cur Loss: 0.34900710, Cur Avg Loss: 0.43315258, Log Avg loss: 0.48128630, Global Avg Loss: 1.80980962, Time: 0.0210 Steps: 25510, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002101, Sample Num: 33616, Cur Loss: 0.40014780, Cur Avg Loss: 0.43319728, Log Avg loss: 0.44254525, Global Avg Loss: 1.80927386, Time: 0.0210 Steps: 25520, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002111, Sample Num: 33776, Cur Loss: 0.18255401, Cur Avg Loss: 0.43346690, Log Avg loss: 0.49011294, Global Avg Loss: 1.80875715, Time: 0.0210 Steps: 25530, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002121, Sample Num: 33936, Cur Loss: 0.30580497, Cur Avg Loss: 0.43300918, Log Avg loss: 0.33638529, Global Avg Loss: 1.80818065, Time: 0.0210 Steps: 25540, Updated lr: 0.000077 ***** Running evaluation checkpoint-25548 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-25548 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.759490, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.536385, "eval_total_loss": 377.078365, "eval_mae": 0.509718, "eval_mse": 0.536603, "eval_r2": 0.6589, "eval_sp_statistic": 0.843868, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.867116, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.096985, "test_total_loss": 550.686482, "test_mae": 0.777701, "test_mse": 1.097269, "test_r2": 0.291813, "test_sp_statistic": 0.76224, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.813814, "test_ps_pvalue": 0.0, "lr": 7.672072072072073e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.8077971145059522, "train_cur_epoch_loss": 923.0793359801173, "train_cur_epoch_avg_loss": 0.4335741362048461, "train_cur_epoch_time": 44.75948977470398, "train_cur_epoch_avg_time": 0.021023715253501166, "epoch": 12, "step": 25548} ################################################## Training, Epoch: 0013, Batch: 000002, Sample Num: 32, Cur Loss: 0.21078423, Cur Avg Loss: 0.37051497, Log Avg loss: 0.54078897, Global Avg Loss: 1.80768461, Time: 0.0258 Steps: 25550, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000012, Sample Num: 192, Cur Loss: 0.55174524, Cur Avg Loss: 0.66955195, Log Avg loss: 0.72935934, Global Avg Loss: 1.80726273, Time: 0.0210 Steps: 25560, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000022, Sample Num: 352, Cur Loss: 0.85737765, Cur Avg Loss: 0.60742747, Log Avg loss: 0.53287809, Global Avg Loss: 1.80676434, Time: 0.0210 Steps: 25570, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000032, Sample Num: 512, Cur Loss: 0.18406564, Cur Avg Loss: 0.52671260, Log Avg loss: 0.34913989, Global Avg Loss: 1.80619451, Time: 0.0209 Steps: 25580, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000042, Sample Num: 672, Cur Loss: 0.11475278, Cur Avg Loss: 0.51131840, Log Avg loss: 0.46205697, Global Avg Loss: 1.80566925, Time: 0.0212 Steps: 25590, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000052, Sample Num: 832, Cur Loss: 0.13981289, Cur Avg Loss: 0.49513667, Log Avg loss: 0.42717341, Global Avg Loss: 1.80513077, Time: 0.0211 Steps: 25600, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000062, Sample Num: 992, Cur Loss: 0.31482086, Cur Avg Loss: 0.46073824, Log Avg loss: 0.28186642, Global Avg Loss: 1.80453598, Time: 0.0211 Steps: 25610, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000072, Sample Num: 1152, Cur Loss: 0.26191401, Cur Avg Loss: 0.44607475, Log Avg loss: 0.35516107, Global Avg Loss: 1.80397026, Time: 0.0212 Steps: 25620, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000082, Sample Num: 1312, Cur Loss: 0.60995716, Cur Avg Loss: 0.43563597, Log Avg loss: 0.36047680, Global Avg Loss: 1.80340706, Time: 0.0211 Steps: 25630, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000092, Sample Num: 1472, Cur Loss: 0.15244772, Cur Avg Loss: 0.42726240, Log Avg loss: 0.35859908, Global Avg Loss: 1.80284356, Time: 0.0211 Steps: 25640, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000102, Sample Num: 1632, Cur Loss: 0.10905795, Cur Avg Loss: 0.42766065, Log Avg loss: 0.43132451, Global Avg Loss: 1.80230885, Time: 0.0211 Steps: 25650, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000112, Sample Num: 1792, Cur Loss: 0.23395425, Cur Avg Loss: 0.41800286, Log Avg loss: 0.31949346, Global Avg Loss: 1.80173098, Time: 0.0210 Steps: 25660, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000122, Sample Num: 1952, Cur Loss: 1.01146901, Cur Avg Loss: 0.41083638, Log Avg loss: 0.33057185, Global Avg Loss: 1.80115788, Time: 0.0211 Steps: 25670, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000132, Sample Num: 2112, Cur Loss: 0.34109032, Cur Avg Loss: 0.40747609, Log Avg loss: 0.36648045, Global Avg Loss: 1.80059920, Time: 0.0210 Steps: 25680, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000142, Sample Num: 2272, Cur Loss: 0.32407230, Cur Avg Loss: 0.39986531, Log Avg loss: 0.29940301, Global Avg Loss: 1.80001485, Time: 0.0212 Steps: 25690, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000152, Sample Num: 2432, Cur Loss: 1.02570188, Cur Avg Loss: 0.40777088, Log Avg loss: 0.52003010, Global Avg Loss: 1.79951680, Time: 0.0211 Steps: 25700, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000162, Sample Num: 2592, Cur Loss: 0.24376123, Cur Avg Loss: 0.40064185, Log Avg loss: 0.29228056, Global Avg Loss: 1.79893056, Time: 0.0211 Steps: 25710, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000172, Sample Num: 2752, Cur Loss: 0.41188177, Cur Avg Loss: 0.39912939, Log Avg loss: 0.37462750, Global Avg Loss: 1.79837679, Time: 0.0211 Steps: 25720, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000182, Sample Num: 2912, Cur Loss: 0.34950542, Cur Avg Loss: 0.39979859, Log Avg loss: 0.41130887, Global Avg Loss: 1.79783770, Time: 0.0210 Steps: 25730, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000192, Sample Num: 3072, Cur Loss: 0.15135103, Cur Avg Loss: 0.39727880, Log Avg loss: 0.35141861, Global Avg Loss: 1.79727577, Time: 0.0211 Steps: 25740, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000202, Sample Num: 3232, Cur Loss: 0.17206834, Cur Avg Loss: 0.39596846, Log Avg loss: 0.37080997, Global Avg Loss: 1.79672180, Time: 0.0212 Steps: 25750, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000212, Sample Num: 3392, Cur Loss: 0.33957165, Cur Avg Loss: 0.39879328, Log Avg loss: 0.45585466, Global Avg Loss: 1.79620128, Time: 0.0211 Steps: 25760, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000222, Sample Num: 3552, Cur Loss: 0.33159354, Cur Avg Loss: 0.40456924, Log Avg loss: 0.52701945, Global Avg Loss: 1.79570877, Time: 0.0212 Steps: 25770, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000232, Sample Num: 3712, Cur Loss: 0.17372122, Cur Avg Loss: 0.40142910, Log Avg loss: 0.33171805, Global Avg Loss: 1.79514089, Time: 0.0214 Steps: 25780, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000242, Sample Num: 3872, Cur Loss: 0.31817704, Cur Avg Loss: 0.40399833, Log Avg loss: 0.46360445, Global Avg Loss: 1.79462459, Time: 0.0212 Steps: 25790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000252, Sample Num: 4032, Cur Loss: 0.36263213, Cur Avg Loss: 0.40342416, Log Avg loss: 0.38952925, Global Avg Loss: 1.79407998, Time: 0.0212 Steps: 25800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000262, Sample Num: 4192, Cur Loss: 0.62236887, Cur Avg Loss: 0.40522298, Log Avg loss: 0.45055338, Global Avg Loss: 1.79355944, Time: 0.0249 Steps: 25810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000272, Sample Num: 4352, Cur Loss: 0.88732469, Cur Avg Loss: 0.41086709, Log Avg loss: 0.55874278, Global Avg Loss: 1.79308120, Time: 0.0212 Steps: 25820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000282, Sample Num: 4512, Cur Loss: 0.17221905, Cur Avg Loss: 0.41661554, Log Avg loss: 0.57297313, Global Avg Loss: 1.79260884, Time: 0.0212 Steps: 25830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000292, Sample Num: 4672, Cur Loss: 0.68791729, Cur Avg Loss: 0.41682167, Log Avg loss: 0.42263452, Global Avg Loss: 1.79207866, Time: 0.0212 Steps: 25840, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000302, Sample Num: 4832, Cur Loss: 0.98308659, Cur Avg Loss: 0.41948983, Log Avg loss: 0.49740033, Global Avg Loss: 1.79157782, Time: 0.0212 Steps: 25850, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000312, Sample Num: 4992, Cur Loss: 0.26708171, Cur Avg Loss: 0.41873350, Log Avg loss: 0.39589238, Global Avg Loss: 1.79103811, Time: 0.0212 Steps: 25860, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000322, Sample Num: 5152, Cur Loss: 0.38484770, Cur Avg Loss: 0.41580577, Log Avg loss: 0.32446053, Global Avg Loss: 1.79047121, Time: 0.0212 Steps: 25870, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000332, Sample Num: 5312, Cur Loss: 0.20353377, Cur Avg Loss: 0.41501173, Log Avg loss: 0.38944369, Global Avg Loss: 1.78992985, Time: 0.0212 Steps: 25880, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000342, Sample Num: 5472, Cur Loss: 0.23339602, Cur Avg Loss: 0.41560734, Log Avg loss: 0.43538131, Global Avg Loss: 1.78940666, Time: 0.0212 Steps: 25890, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000352, Sample Num: 5632, Cur Loss: 0.67471528, Cur Avg Loss: 0.41639002, Log Avg loss: 0.44315794, Global Avg Loss: 1.78888687, Time: 0.0211 Steps: 25900, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000362, Sample Num: 5792, Cur Loss: 0.21755026, Cur Avg Loss: 0.41879776, Log Avg loss: 0.50355010, Global Avg Loss: 1.78839079, Time: 0.0211 Steps: 25910, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000372, Sample Num: 5952, Cur Loss: 0.25746736, Cur Avg Loss: 0.41618138, Log Avg loss: 0.32146848, Global Avg Loss: 1.78782485, Time: 0.0211 Steps: 25920, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000382, Sample Num: 6112, Cur Loss: 0.30129960, Cur Avg Loss: 0.41633587, Log Avg loss: 0.42208269, Global Avg Loss: 1.78729815, Time: 0.0212 Steps: 25930, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000392, Sample Num: 6272, Cur Loss: 0.47958136, Cur Avg Loss: 0.41896844, Log Avg loss: 0.51953279, Global Avg Loss: 1.78680942, Time: 0.0211 Steps: 25940, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000402, Sample Num: 6432, Cur Loss: 0.17145574, Cur Avg Loss: 0.41578671, Log Avg loss: 0.29106278, Global Avg Loss: 1.78623302, Time: 0.0212 Steps: 25950, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000412, Sample Num: 6592, Cur Loss: 0.64324319, Cur Avg Loss: 0.41439243, Log Avg loss: 0.35834252, Global Avg Loss: 1.78568299, Time: 0.0212 Steps: 25960, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000422, Sample Num: 6752, Cur Loss: 0.33146441, Cur Avg Loss: 0.40990331, Log Avg loss: 0.22495171, Global Avg Loss: 1.78508201, Time: 0.0212 Steps: 25970, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000432, Sample Num: 6912, Cur Loss: 0.19948685, Cur Avg Loss: 0.40812079, Log Avg loss: 0.33289818, Global Avg Loss: 1.78452305, Time: 0.0211 Steps: 25980, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000442, Sample Num: 7072, Cur Loss: 1.08521450, Cur Avg Loss: 0.40822690, Log Avg loss: 0.41281103, Global Avg Loss: 1.78399527, Time: 0.0212 Steps: 25990, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000452, Sample Num: 7232, Cur Loss: 0.73590255, Cur Avg Loss: 0.40728919, Log Avg loss: 0.36584235, Global Avg Loss: 1.78344982, Time: 0.0211 Steps: 26000, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000462, Sample Num: 7392, Cur Loss: 0.43966690, Cur Avg Loss: 0.40775691, Log Avg loss: 0.42889798, Global Avg Loss: 1.78292904, Time: 0.0211 Steps: 26010, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000472, Sample Num: 7552, Cur Loss: 0.82222891, Cur Avg Loss: 0.41059294, Log Avg loss: 0.54161728, Global Avg Loss: 1.78245198, Time: 0.0211 Steps: 26020, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000482, Sample Num: 7712, Cur Loss: 0.14499384, Cur Avg Loss: 0.40873913, Log Avg loss: 0.32123942, Global Avg Loss: 1.78189062, Time: 0.0212 Steps: 26030, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000492, Sample Num: 7872, Cur Loss: 0.36043030, Cur Avg Loss: 0.40865896, Log Avg loss: 0.40479468, Global Avg Loss: 1.78136179, Time: 0.0211 Steps: 26040, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000502, Sample Num: 8032, Cur Loss: 0.78889734, Cur Avg Loss: 0.41087253, Log Avg loss: 0.51977997, Global Avg Loss: 1.78087749, Time: 0.0212 Steps: 26050, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000512, Sample Num: 8192, Cur Loss: 0.45018688, Cur Avg Loss: 0.41142627, Log Avg loss: 0.43922404, Global Avg Loss: 1.78036266, Time: 0.0255 Steps: 26060, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000522, Sample Num: 8352, Cur Loss: 0.32222006, Cur Avg Loss: 0.40965499, Log Avg loss: 0.31896571, Global Avg Loss: 1.77980209, Time: 0.0211 Steps: 26070, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000532, Sample Num: 8512, Cur Loss: 0.20106822, Cur Avg Loss: 0.40769850, Log Avg loss: 0.30556974, Global Avg Loss: 1.77923682, Time: 0.0211 Steps: 26080, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000542, Sample Num: 8672, Cur Loss: 0.36133066, Cur Avg Loss: 0.40633874, Log Avg loss: 0.33399946, Global Avg Loss: 1.77868288, Time: 0.0211 Steps: 26090, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000552, Sample Num: 8832, Cur Loss: 0.25995946, Cur Avg Loss: 0.40392447, Log Avg loss: 0.27307092, Global Avg Loss: 1.77810601, Time: 0.0212 Steps: 26100, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000562, Sample Num: 8992, Cur Loss: 0.40165615, Cur Avg Loss: 0.40300514, Log Avg loss: 0.35225850, Global Avg Loss: 1.77755992, Time: 0.0211 Steps: 26110, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000572, Sample Num: 9152, Cur Loss: 0.18689060, Cur Avg Loss: 0.40121305, Log Avg loss: 0.30049706, Global Avg Loss: 1.77699443, Time: 0.0210 Steps: 26120, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000582, Sample Num: 9312, Cur Loss: 0.70406413, Cur Avg Loss: 0.40070209, Log Avg loss: 0.37147530, Global Avg Loss: 1.77645654, Time: 0.0211 Steps: 26130, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000592, Sample Num: 9472, Cur Loss: 0.27406189, Cur Avg Loss: 0.40072285, Log Avg loss: 0.40193121, Global Avg Loss: 1.77593070, Time: 0.0210 Steps: 26140, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000602, Sample Num: 9632, Cur Loss: 0.81569159, Cur Avg Loss: 0.40018823, Log Avg loss: 0.36853900, Global Avg Loss: 1.77539250, Time: 0.0211 Steps: 26150, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000612, Sample Num: 9792, Cur Loss: 0.57416207, Cur Avg Loss: 0.40185182, Log Avg loss: 0.50199973, Global Avg Loss: 1.77490573, Time: 0.0211 Steps: 26160, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000622, Sample Num: 9952, Cur Loss: 0.25283211, Cur Avg Loss: 0.40105776, Log Avg loss: 0.35246107, Global Avg Loss: 1.77436219, Time: 0.0211 Steps: 26170, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000632, Sample Num: 10112, Cur Loss: 0.38422161, Cur Avg Loss: 0.40168441, Log Avg loss: 0.44066195, Global Avg Loss: 1.77385276, Time: 0.0211 Steps: 26180, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000642, Sample Num: 10272, Cur Loss: 0.38044870, Cur Avg Loss: 0.40240072, Log Avg loss: 0.44767187, Global Avg Loss: 1.77334639, Time: 0.0211 Steps: 26190, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000652, Sample Num: 10432, Cur Loss: 0.52994287, Cur Avg Loss: 0.40685683, Log Avg loss: 0.69293876, Global Avg Loss: 1.77293402, Time: 0.0211 Steps: 26200, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000662, Sample Num: 10592, Cur Loss: 0.37069216, Cur Avg Loss: 0.40788946, Log Avg loss: 0.47521688, Global Avg Loss: 1.77243890, Time: 0.0210 Steps: 26210, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000672, Sample Num: 10752, Cur Loss: 0.27017710, Cur Avg Loss: 0.40663407, Log Avg loss: 0.32352756, Global Avg Loss: 1.77188630, Time: 0.0211 Steps: 26220, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000682, Sample Num: 10912, Cur Loss: 1.11204422, Cur Avg Loss: 0.40828484, Log Avg loss: 0.51921681, Global Avg Loss: 1.77140873, Time: 0.0211 Steps: 26230, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000692, Sample Num: 11072, Cur Loss: 0.37611771, Cur Avg Loss: 0.40704622, Log Avg loss: 0.32257229, Global Avg Loss: 1.77085658, Time: 0.0211 Steps: 26240, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000702, Sample Num: 11232, Cur Loss: 0.32765692, Cur Avg Loss: 0.40690182, Log Avg loss: 0.39690936, Global Avg Loss: 1.77033317, Time: 0.0211 Steps: 26250, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000712, Sample Num: 11392, Cur Loss: 0.56280327, Cur Avg Loss: 0.40634695, Log Avg loss: 0.36739508, Global Avg Loss: 1.76979892, Time: 0.0211 Steps: 26260, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000722, Sample Num: 11552, Cur Loss: 0.29615098, Cur Avg Loss: 0.40545052, Log Avg loss: 0.34162415, Global Avg Loss: 1.76925527, Time: 0.0210 Steps: 26270, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000732, Sample Num: 11712, Cur Loss: 0.33909282, Cur Avg Loss: 0.40729351, Log Avg loss: 0.54035778, Global Avg Loss: 1.76878765, Time: 0.0211 Steps: 26280, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000742, Sample Num: 11872, Cur Loss: 0.20701581, Cur Avg Loss: 0.40612438, Log Avg loss: 0.32054392, Global Avg Loss: 1.76823678, Time: 0.0211 Steps: 26290, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000752, Sample Num: 12032, Cur Loss: 0.26062983, Cur Avg Loss: 0.40496715, Log Avg loss: 0.31910085, Global Avg Loss: 1.76768578, Time: 0.0211 Steps: 26300, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000762, Sample Num: 12192, Cur Loss: 0.42670783, Cur Avg Loss: 0.40453945, Log Avg loss: 0.37237667, Global Avg Loss: 1.76715544, Time: 0.0210 Steps: 26310, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000772, Sample Num: 12352, Cur Loss: 0.54204249, Cur Avg Loss: 0.40456185, Log Avg loss: 0.40626859, Global Avg Loss: 1.76663839, Time: 0.0212 Steps: 26320, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000782, Sample Num: 12512, Cur Loss: 0.35798061, Cur Avg Loss: 0.40556990, Log Avg loss: 0.48339099, Global Avg Loss: 1.76615102, Time: 0.0211 Steps: 26330, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000792, Sample Num: 12672, Cur Loss: 0.38005728, Cur Avg Loss: 0.40490415, Log Avg loss: 0.35284296, Global Avg Loss: 1.76561446, Time: 0.0211 Steps: 26340, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000802, Sample Num: 12832, Cur Loss: 0.27474055, Cur Avg Loss: 0.40426098, Log Avg loss: 0.35332195, Global Avg Loss: 1.76507848, Time: 0.0210 Steps: 26350, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000812, Sample Num: 12992, Cur Loss: 0.27860266, Cur Avg Loss: 0.40288724, Log Avg loss: 0.29271327, Global Avg Loss: 1.76451992, Time: 0.0212 Steps: 26360, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000822, Sample Num: 13152, Cur Loss: 0.43964180, Cur Avg Loss: 0.40266840, Log Avg loss: 0.38489808, Global Avg Loss: 1.76399674, Time: 0.0210 Steps: 26370, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000832, Sample Num: 13312, Cur Loss: 0.48561069, Cur Avg Loss: 0.40443727, Log Avg loss: 0.54983875, Global Avg Loss: 1.76353649, Time: 0.0211 Steps: 26380, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000842, Sample Num: 13472, Cur Loss: 0.30334413, Cur Avg Loss: 0.40387157, Log Avg loss: 0.35680551, Global Avg Loss: 1.76300343, Time: 0.0211 Steps: 26390, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000852, Sample Num: 13632, Cur Loss: 0.72708571, Cur Avg Loss: 0.40400371, Log Avg loss: 0.41512986, Global Avg Loss: 1.76249287, Time: 0.0211 Steps: 26400, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000862, Sample Num: 13792, Cur Loss: 0.57415205, Cur Avg Loss: 0.40349319, Log Avg loss: 0.35999689, Global Avg Loss: 1.76196183, Time: 0.0211 Steps: 26410, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000872, Sample Num: 13952, Cur Loss: 0.45653689, Cur Avg Loss: 0.40336921, Log Avg loss: 0.39268143, Global Avg Loss: 1.76144355, Time: 0.0213 Steps: 26420, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000882, Sample Num: 14112, Cur Loss: 0.58798528, Cur Avg Loss: 0.40395827, Log Avg loss: 0.45532457, Global Avg Loss: 1.76094937, Time: 0.0210 Steps: 26430, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000892, Sample Num: 14272, Cur Loss: 0.18795471, Cur Avg Loss: 0.40397569, Log Avg loss: 0.40551198, Global Avg Loss: 1.76043672, Time: 0.0211 Steps: 26440, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000902, Sample Num: 14432, Cur Loss: 0.27802971, Cur Avg Loss: 0.40279070, Log Avg loss: 0.29708962, Global Avg Loss: 1.75988347, Time: 0.0211 Steps: 26450, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000912, Sample Num: 14592, Cur Loss: 0.61095744, Cur Avg Loss: 0.40317851, Log Avg loss: 0.43815905, Global Avg Loss: 1.75938396, Time: 0.0211 Steps: 26460, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000922, Sample Num: 14752, Cur Loss: 1.03768694, Cur Avg Loss: 0.40474096, Log Avg loss: 0.54723653, Global Avg Loss: 1.75892602, Time: 0.0210 Steps: 26470, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000932, Sample Num: 14912, Cur Loss: 0.24187207, Cur Avg Loss: 0.40446409, Log Avg loss: 0.37893641, Global Avg Loss: 1.75840488, Time: 0.0210 Steps: 26480, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000942, Sample Num: 15072, Cur Loss: 0.32341439, Cur Avg Loss: 0.40452276, Log Avg loss: 0.40999140, Global Avg Loss: 1.75789585, Time: 0.0211 Steps: 26490, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000952, Sample Num: 15232, Cur Loss: 0.42107418, Cur Avg Loss: 0.40369628, Log Avg loss: 0.32584150, Global Avg Loss: 1.75735545, Time: 0.0211 Steps: 26500, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000962, Sample Num: 15392, Cur Loss: 0.17944071, Cur Avg Loss: 0.40480924, Log Avg loss: 0.51076343, Global Avg Loss: 1.75688522, Time: 0.0212 Steps: 26510, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000972, Sample Num: 15552, Cur Loss: 0.41512042, Cur Avg Loss: 0.40434518, Log Avg loss: 0.35970239, Global Avg Loss: 1.75635838, Time: 0.0210 Steps: 26520, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000982, Sample Num: 15712, Cur Loss: 0.32947600, Cur Avg Loss: 0.40542639, Log Avg loss: 0.51051970, Global Avg Loss: 1.75588878, Time: 0.0211 Steps: 26530, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000992, Sample Num: 15872, Cur Loss: 0.34370267, Cur Avg Loss: 0.40484664, Log Avg loss: 0.34791494, Global Avg Loss: 1.75535827, Time: 0.0211 Steps: 26540, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001002, Sample Num: 16032, Cur Loss: 1.39456260, Cur Avg Loss: 0.40515855, Log Avg loss: 0.43610065, Global Avg Loss: 1.75486138, Time: 0.0211 Steps: 26550, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001012, Sample Num: 16192, Cur Loss: 0.59907842, Cur Avg Loss: 0.40660561, Log Avg loss: 0.55160103, Global Avg Loss: 1.75440834, Time: 0.0211 Steps: 26560, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001022, Sample Num: 16352, Cur Loss: 0.16898713, Cur Avg Loss: 0.40631037, Log Avg loss: 0.37643211, Global Avg Loss: 1.75388972, Time: 0.0211 Steps: 26570, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001032, Sample Num: 16512, Cur Loss: 0.59220970, Cur Avg Loss: 0.40654337, Log Avg loss: 0.43035549, Global Avg Loss: 1.75339178, Time: 0.0210 Steps: 26580, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001042, Sample Num: 16672, Cur Loss: 0.51006901, Cur Avg Loss: 0.40593300, Log Avg loss: 0.34294268, Global Avg Loss: 1.75286133, Time: 0.0210 Steps: 26590, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001052, Sample Num: 16832, Cur Loss: 0.26795387, Cur Avg Loss: 0.40607427, Log Avg loss: 0.42079472, Global Avg Loss: 1.75236056, Time: 0.0210 Steps: 26600, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001062, Sample Num: 16992, Cur Loss: 0.24709764, Cur Avg Loss: 0.40629319, Log Avg loss: 0.42932381, Global Avg Loss: 1.75186336, Time: 0.0211 Steps: 26610, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001072, Sample Num: 17152, Cur Loss: 0.58586442, Cur Avg Loss: 0.40590240, Log Avg loss: 0.36440043, Global Avg Loss: 1.75134215, Time: 0.0210 Steps: 26620, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001082, Sample Num: 17312, Cur Loss: 0.72128999, Cur Avg Loss: 0.40681334, Log Avg loss: 0.50446655, Global Avg Loss: 1.75087393, Time: 0.0210 Steps: 26630, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001092, Sample Num: 17472, Cur Loss: 0.18812793, Cur Avg Loss: 0.40774383, Log Avg loss: 0.50842234, Global Avg Loss: 1.75040754, Time: 0.0210 Steps: 26640, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001102, Sample Num: 17632, Cur Loss: 0.68981218, Cur Avg Loss: 0.40781700, Log Avg loss: 0.41580772, Global Avg Loss: 1.74990675, Time: 0.0210 Steps: 26650, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001112, Sample Num: 17792, Cur Loss: 0.14504102, Cur Avg Loss: 0.40724148, Log Avg loss: 0.34381830, Global Avg Loss: 1.74937934, Time: 0.0210 Steps: 26660, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001122, Sample Num: 17952, Cur Loss: 0.58959478, Cur Avg Loss: 0.40895376, Log Avg loss: 0.59936016, Global Avg Loss: 1.74894814, Time: 0.0211 Steps: 26670, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001132, Sample Num: 18112, Cur Loss: 0.18132262, Cur Avg Loss: 0.40869934, Log Avg loss: 0.38015329, Global Avg Loss: 1.74843510, Time: 0.0210 Steps: 26680, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001142, Sample Num: 18272, Cur Loss: 0.48131093, Cur Avg Loss: 0.40897642, Log Avg loss: 0.44034131, Global Avg Loss: 1.74794499, Time: 0.0210 Steps: 26690, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001152, Sample Num: 18432, Cur Loss: 0.21252254, Cur Avg Loss: 0.40901389, Log Avg loss: 0.41329280, Global Avg Loss: 1.74744512, Time: 0.0211 Steps: 26700, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001162, Sample Num: 18592, Cur Loss: 0.88467586, Cur Avg Loss: 0.40942794, Log Avg loss: 0.45712703, Global Avg Loss: 1.74696203, Time: 0.0210 Steps: 26710, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001172, Sample Num: 18752, Cur Loss: 0.54003334, Cur Avg Loss: 0.40927817, Log Avg loss: 0.39187431, Global Avg Loss: 1.74645489, Time: 0.0211 Steps: 26720, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001182, Sample Num: 18912, Cur Loss: 0.37872458, Cur Avg Loss: 0.40927414, Log Avg loss: 0.40880214, Global Avg Loss: 1.74595446, Time: 0.0210 Steps: 26730, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001192, Sample Num: 19072, Cur Loss: 0.12933159, Cur Avg Loss: 0.40847256, Log Avg loss: 0.31372632, Global Avg Loss: 1.74541885, Time: 0.0210 Steps: 26740, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001202, Sample Num: 19232, Cur Loss: 0.42085618, Cur Avg Loss: 0.40963059, Log Avg loss: 0.54766805, Global Avg Loss: 1.74497109, Time: 0.0210 Steps: 26750, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001212, Sample Num: 19392, Cur Loss: 0.09162432, Cur Avg Loss: 0.40955921, Log Avg loss: 0.40097860, Global Avg Loss: 1.74446885, Time: 0.0211 Steps: 26760, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001222, Sample Num: 19552, Cur Loss: 0.38521481, Cur Avg Loss: 0.40858543, Log Avg loss: 0.29056340, Global Avg Loss: 1.74392574, Time: 0.0210 Steps: 26770, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001232, Sample Num: 19712, Cur Loss: 0.24008429, Cur Avg Loss: 0.40735453, Log Avg loss: 0.25693891, Global Avg Loss: 1.74337048, Time: 0.0210 Steps: 26780, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001242, Sample Num: 19872, Cur Loss: 0.54393029, Cur Avg Loss: 0.40767427, Log Avg loss: 0.44706613, Global Avg Loss: 1.74288660, Time: 0.0211 Steps: 26790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001252, Sample Num: 20032, Cur Loss: 0.36252946, Cur Avg Loss: 0.40786423, Log Avg loss: 0.43145714, Global Avg Loss: 1.74239726, Time: 0.0210 Steps: 26800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001262, Sample Num: 20192, Cur Loss: 0.83424449, Cur Avg Loss: 0.40788452, Log Avg loss: 0.41042535, Global Avg Loss: 1.74190045, Time: 0.0210 Steps: 26810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001272, Sample Num: 20352, Cur Loss: 0.35349426, Cur Avg Loss: 0.40758377, Log Avg loss: 0.36962792, Global Avg Loss: 1.74138879, Time: 0.0210 Steps: 26820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001282, Sample Num: 20512, Cur Loss: 0.18761531, Cur Avg Loss: 0.40676277, Log Avg loss: 0.30233176, Global Avg Loss: 1.74085242, Time: 0.0250 Steps: 26830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001292, Sample Num: 20672, Cur Loss: 0.27950364, Cur Avg Loss: 0.40590607, Log Avg loss: 0.29607815, Global Avg Loss: 1.74031413, Time: 0.0212 Steps: 26840, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001302, Sample Num: 20832, Cur Loss: 0.27819505, Cur Avg Loss: 0.40580426, Log Avg loss: 0.39265019, Global Avg Loss: 1.73981221, Time: 0.0212 Steps: 26850, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001312, Sample Num: 20992, Cur Loss: 0.32997483, Cur Avg Loss: 0.40565953, Log Avg loss: 0.38681478, Global Avg Loss: 1.73930849, Time: 0.0211 Steps: 26860, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001322, Sample Num: 21152, Cur Loss: 0.55480015, Cur Avg Loss: 0.40577347, Log Avg loss: 0.42072322, Global Avg Loss: 1.73881776, Time: 0.0211 Steps: 26870, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001332, Sample Num: 21312, Cur Loss: 0.28352025, Cur Avg Loss: 0.40551796, Log Avg loss: 0.37173906, Global Avg Loss: 1.73830917, Time: 0.0211 Steps: 26880, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001342, Sample Num: 21472, Cur Loss: 0.40261891, Cur Avg Loss: 0.40514660, Log Avg loss: 0.35568150, Global Avg Loss: 1.73779500, Time: 0.0211 Steps: 26890, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001352, Sample Num: 21632, Cur Loss: 0.83330691, Cur Avg Loss: 0.40594584, Log Avg loss: 0.51320431, Global Avg Loss: 1.73733976, Time: 0.0211 Steps: 26900, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001362, Sample Num: 21792, Cur Loss: 0.28918833, Cur Avg Loss: 0.40547660, Log Avg loss: 0.34203539, Global Avg Loss: 1.73682125, Time: 0.0211 Steps: 26910, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001372, Sample Num: 21952, Cur Loss: 0.58275825, Cur Avg Loss: 0.40627439, Log Avg loss: 0.51493262, Global Avg Loss: 1.73636735, Time: 0.0211 Steps: 26920, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001382, Sample Num: 22112, Cur Loss: 0.76644421, Cur Avg Loss: 0.40574234, Log Avg loss: 0.33274609, Global Avg Loss: 1.73584614, Time: 0.0211 Steps: 26930, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001392, Sample Num: 22272, Cur Loss: 0.57254303, Cur Avg Loss: 0.40578944, Log Avg loss: 0.41229797, Global Avg Loss: 1.73535485, Time: 0.0212 Steps: 26940, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001402, Sample Num: 22432, Cur Loss: 0.13620932, Cur Avg Loss: 0.40497467, Log Avg loss: 0.29155916, Global Avg Loss: 1.73481912, Time: 0.0212 Steps: 26950, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001412, Sample Num: 22592, Cur Loss: 0.10974707, Cur Avg Loss: 0.40444805, Log Avg loss: 0.33061591, Global Avg Loss: 1.73429827, Time: 0.0211 Steps: 26960, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001422, Sample Num: 22752, Cur Loss: 0.18287577, Cur Avg Loss: 0.40359486, Log Avg loss: 0.28312393, Global Avg Loss: 1.73376020, Time: 0.0211 Steps: 26970, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001432, Sample Num: 22912, Cur Loss: 0.27062368, Cur Avg Loss: 0.40337985, Log Avg loss: 0.37280578, Global Avg Loss: 1.73325577, Time: 0.0211 Steps: 26980, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001442, Sample Num: 23072, Cur Loss: 0.19796127, Cur Avg Loss: 0.40296357, Log Avg loss: 0.34335159, Global Avg Loss: 1.73274080, Time: 0.0211 Steps: 26990, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001452, Sample Num: 23232, Cur Loss: 0.66140628, Cur Avg Loss: 0.40227588, Log Avg loss: 0.30311137, Global Avg Loss: 1.73221131, Time: 0.0211 Steps: 27000, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001462, Sample Num: 23392, Cur Loss: 0.17144574, Cur Avg Loss: 0.40292147, Log Avg loss: 0.49666108, Global Avg Loss: 1.73175386, Time: 0.0211 Steps: 27010, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001472, Sample Num: 23552, Cur Loss: 0.14763170, Cur Avg Loss: 0.40218481, Log Avg loss: 0.29448571, Global Avg Loss: 1.73122194, Time: 0.0211 Steps: 27020, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001482, Sample Num: 23712, Cur Loss: 0.57753837, Cur Avg Loss: 0.40251280, Log Avg loss: 0.45079187, Global Avg Loss: 1.73074823, Time: 0.0211 Steps: 27030, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001492, Sample Num: 23872, Cur Loss: 0.97850430, Cur Avg Loss: 0.40268100, Log Avg loss: 0.42760828, Global Avg Loss: 1.73026630, Time: 0.0211 Steps: 27040, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001502, Sample Num: 24032, Cur Loss: 0.45634091, Cur Avg Loss: 0.40355014, Log Avg loss: 0.53322602, Global Avg Loss: 1.72982377, Time: 0.0211 Steps: 27050, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001512, Sample Num: 24192, Cur Loss: 0.42184153, Cur Avg Loss: 0.40352308, Log Avg loss: 0.39945962, Global Avg Loss: 1.72933214, Time: 0.0211 Steps: 27060, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001522, Sample Num: 24352, Cur Loss: 0.56540024, Cur Avg Loss: 0.40460524, Log Avg loss: 0.56822787, Global Avg Loss: 1.72890321, Time: 0.0211 Steps: 27070, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001532, Sample Num: 24512, Cur Loss: 0.60559547, Cur Avg Loss: 0.40504102, Log Avg loss: 0.47136572, Global Avg Loss: 1.72843883, Time: 0.0211 Steps: 27080, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001542, Sample Num: 24672, Cur Loss: 0.90257084, Cur Avg Loss: 0.40554632, Log Avg loss: 0.48295918, Global Avg Loss: 1.72797907, Time: 0.0212 Steps: 27090, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001552, Sample Num: 24832, Cur Loss: 0.27689517, Cur Avg Loss: 0.40669758, Log Avg loss: 0.58422097, Global Avg Loss: 1.72755702, Time: 0.0211 Steps: 27100, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001562, Sample Num: 24992, Cur Loss: 0.45682380, Cur Avg Loss: 0.40652083, Log Avg loss: 0.37908995, Global Avg Loss: 1.72705962, Time: 0.0212 Steps: 27110, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001572, Sample Num: 25152, Cur Loss: 0.39242935, Cur Avg Loss: 0.40647984, Log Avg loss: 0.40007634, Global Avg Loss: 1.72657032, Time: 0.0211 Steps: 27120, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001582, Sample Num: 25312, Cur Loss: 0.21723115, Cur Avg Loss: 0.40570236, Log Avg loss: 0.28348256, Global Avg Loss: 1.72603840, Time: 0.0211 Steps: 27130, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001592, Sample Num: 25472, Cur Loss: 0.24878579, Cur Avg Loss: 0.40550185, Log Avg loss: 0.37378223, Global Avg Loss: 1.72554015, Time: 0.0211 Steps: 27140, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001602, Sample Num: 25632, Cur Loss: 0.29004809, Cur Avg Loss: 0.40557713, Log Avg loss: 0.41756036, Global Avg Loss: 1.72505839, Time: 0.0211 Steps: 27150, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001612, Sample Num: 25792, Cur Loss: 0.47345036, Cur Avg Loss: 0.40581459, Log Avg loss: 0.44385668, Global Avg Loss: 1.72458666, Time: 0.0211 Steps: 27160, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001622, Sample Num: 25952, Cur Loss: 0.29656333, Cur Avg Loss: 0.40533585, Log Avg loss: 0.32816253, Global Avg Loss: 1.72407271, Time: 0.0211 Steps: 27170, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001632, Sample Num: 26112, Cur Loss: 0.45488262, Cur Avg Loss: 0.40479755, Log Avg loss: 0.31748576, Global Avg Loss: 1.72355520, Time: 0.0212 Steps: 27180, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001642, Sample Num: 26272, Cur Loss: 0.45899537, Cur Avg Loss: 0.40455956, Log Avg loss: 0.36571978, Global Avg Loss: 1.72305581, Time: 0.0211 Steps: 27190, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001652, Sample Num: 26432, Cur Loss: 0.39883810, Cur Avg Loss: 0.40447744, Log Avg loss: 0.39099367, Global Avg Loss: 1.72256608, Time: 0.0211 Steps: 27200, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001662, Sample Num: 26592, Cur Loss: 0.24796957, Cur Avg Loss: 0.40431165, Log Avg loss: 0.37692283, Global Avg Loss: 1.72207154, Time: 0.0211 Steps: 27210, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001672, Sample Num: 26752, Cur Loss: 0.67358434, Cur Avg Loss: 0.40455014, Log Avg loss: 0.44418680, Global Avg Loss: 1.72160208, Time: 0.0211 Steps: 27220, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001682, Sample Num: 26912, Cur Loss: 0.14567326, Cur Avg Loss: 0.40355129, Log Avg loss: 0.23654348, Global Avg Loss: 1.72105670, Time: 0.0211 Steps: 27230, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001692, Sample Num: 27072, Cur Loss: 0.22547084, Cur Avg Loss: 0.40406918, Log Avg loss: 0.49117877, Global Avg Loss: 1.72060520, Time: 0.0211 Steps: 27240, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001702, Sample Num: 27232, Cur Loss: 0.42253304, Cur Avg Loss: 0.40365483, Log Avg loss: 0.33354555, Global Avg Loss: 1.72009619, Time: 0.0211 Steps: 27250, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001712, Sample Num: 27392, Cur Loss: 0.23795086, Cur Avg Loss: 0.40294033, Log Avg loss: 0.28133309, Global Avg Loss: 1.71956840, Time: 0.0211 Steps: 27260, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001722, Sample Num: 27552, Cur Loss: 0.16033405, Cur Avg Loss: 0.40207020, Log Avg loss: 0.25310404, Global Avg Loss: 1.71903064, Time: 0.0212 Steps: 27270, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001732, Sample Num: 27712, Cur Loss: 0.35987866, Cur Avg Loss: 0.40141707, Log Avg loss: 0.28894787, Global Avg Loss: 1.71850642, Time: 0.0211 Steps: 27280, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001742, Sample Num: 27872, Cur Loss: 0.45834100, Cur Avg Loss: 0.40156147, Log Avg loss: 0.42657151, Global Avg Loss: 1.71803301, Time: 0.0211 Steps: 27290, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001752, Sample Num: 28032, Cur Loss: 0.40189856, Cur Avg Loss: 0.40144888, Log Avg loss: 0.38183674, Global Avg Loss: 1.71754356, Time: 0.0211 Steps: 27300, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001762, Sample Num: 28192, Cur Loss: 0.12051459, Cur Avg Loss: 0.40107556, Log Avg loss: 0.33566888, Global Avg Loss: 1.71703756, Time: 0.0211 Steps: 27310, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001772, Sample Num: 28352, Cur Loss: 0.18271998, Cur Avg Loss: 0.40036785, Log Avg loss: 0.27567037, Global Avg Loss: 1.71650998, Time: 0.0211 Steps: 27320, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001782, Sample Num: 28512, Cur Loss: 0.23229930, Cur Avg Loss: 0.40044829, Log Avg loss: 0.41470153, Global Avg Loss: 1.71603365, Time: 0.0211 Steps: 27330, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001792, Sample Num: 28672, Cur Loss: 0.21699338, Cur Avg Loss: 0.40048894, Log Avg loss: 0.40773269, Global Avg Loss: 1.71555512, Time: 0.0256 Steps: 27340, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001802, Sample Num: 28832, Cur Loss: 0.39446664, Cur Avg Loss: 0.40040218, Log Avg loss: 0.38485504, Global Avg Loss: 1.71506857, Time: 0.0211 Steps: 27350, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001812, Sample Num: 28992, Cur Loss: 0.09418765, Cur Avg Loss: 0.40037031, Log Avg loss: 0.39462645, Global Avg Loss: 1.71458595, Time: 0.0211 Steps: 27360, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001822, Sample Num: 29152, Cur Loss: 0.41426510, Cur Avg Loss: 0.40004841, Log Avg loss: 0.34172026, Global Avg Loss: 1.71408436, Time: 0.0211 Steps: 27370, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001832, Sample Num: 29312, Cur Loss: 0.11015110, Cur Avg Loss: 0.40041323, Log Avg loss: 0.46688456, Global Avg Loss: 1.71362884, Time: 0.0211 Steps: 27380, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001842, Sample Num: 29472, Cur Loss: 1.15442824, Cur Avg Loss: 0.40115149, Log Avg loss: 0.53639945, Global Avg Loss: 1.71319904, Time: 0.0211 Steps: 27390, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001852, Sample Num: 29632, Cur Loss: 0.12581721, Cur Avg Loss: 0.40119750, Log Avg loss: 0.40967301, Global Avg Loss: 1.71272330, Time: 0.0211 Steps: 27400, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001862, Sample Num: 29792, Cur Loss: 0.30716306, Cur Avg Loss: 0.40094091, Log Avg loss: 0.35342099, Global Avg Loss: 1.71222739, Time: 0.0211 Steps: 27410, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001872, Sample Num: 29952, Cur Loss: 0.43415082, Cur Avg Loss: 0.40119186, Log Avg loss: 0.44791933, Global Avg Loss: 1.71176630, Time: 0.0211 Steps: 27420, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001882, Sample Num: 30112, Cur Loss: 0.62072384, Cur Avg Loss: 0.40230767, Log Avg loss: 0.61118633, Global Avg Loss: 1.71136506, Time: 0.0211 Steps: 27430, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001892, Sample Num: 30272, Cur Loss: 0.42567980, Cur Avg Loss: 0.40354204, Log Avg loss: 0.63584962, Global Avg Loss: 1.71097311, Time: 0.0211 Steps: 27440, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001902, Sample Num: 30432, Cur Loss: 0.46339145, Cur Avg Loss: 0.40355295, Log Avg loss: 0.40561887, Global Avg Loss: 1.71049757, Time: 0.0212 Steps: 27450, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001912, Sample Num: 30592, Cur Loss: 0.25039580, Cur Avg Loss: 0.40378322, Log Avg loss: 0.44758056, Global Avg Loss: 1.71003766, Time: 0.0211 Steps: 27460, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001922, Sample Num: 30752, Cur Loss: 0.33160287, Cur Avg Loss: 0.40340408, Log Avg loss: 0.33091128, Global Avg Loss: 1.70953561, Time: 0.0211 Steps: 27470, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001932, Sample Num: 30912, Cur Loss: 0.18107301, Cur Avg Loss: 0.40304045, Log Avg loss: 0.33315094, Global Avg Loss: 1.70903475, Time: 0.0211 Steps: 27480, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001942, Sample Num: 31072, Cur Loss: 0.23543149, Cur Avg Loss: 0.40276663, Log Avg loss: 0.34986522, Global Avg Loss: 1.70854032, Time: 0.0211 Steps: 27490, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001952, Sample Num: 31232, Cur Loss: 0.11945014, Cur Avg Loss: 0.40276608, Log Avg loss: 0.40265883, Global Avg Loss: 1.70806546, Time: 0.0211 Steps: 27500, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001962, Sample Num: 31392, Cur Loss: 0.89097714, Cur Avg Loss: 0.40269247, Log Avg loss: 0.38832481, Global Avg Loss: 1.70758573, Time: 0.0212 Steps: 27510, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001972, Sample Num: 31552, Cur Loss: 0.23395425, Cur Avg Loss: 0.40220412, Log Avg loss: 0.30638917, Global Avg Loss: 1.70707657, Time: 0.0211 Steps: 27520, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001982, Sample Num: 31712, Cur Loss: 0.23608279, Cur Avg Loss: 0.40190533, Log Avg loss: 0.34298441, Global Avg Loss: 1.70658108, Time: 0.0211 Steps: 27530, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001992, Sample Num: 31872, Cur Loss: 0.22919717, Cur Avg Loss: 0.40135572, Log Avg loss: 0.29242270, Global Avg Loss: 1.70606758, Time: 0.0211 Steps: 27540, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002002, Sample Num: 32032, Cur Loss: 0.81273228, Cur Avg Loss: 0.40121810, Log Avg loss: 0.37380276, Global Avg Loss: 1.70558400, Time: 0.0211 Steps: 27550, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002012, Sample Num: 32192, Cur Loss: 0.61273092, Cur Avg Loss: 0.40123639, Log Avg loss: 0.40489974, Global Avg Loss: 1.70511206, Time: 0.0211 Steps: 27560, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002022, Sample Num: 32352, Cur Loss: 0.55174327, Cur Avg Loss: 0.40092328, Log Avg loss: 0.33792431, Global Avg Loss: 1.70461616, Time: 0.0211 Steps: 27570, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002032, Sample Num: 32512, Cur Loss: 0.60968655, Cur Avg Loss: 0.40101429, Log Avg loss: 0.41941690, Global Avg Loss: 1.70415017, Time: 0.0211 Steps: 27580, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002042, Sample Num: 32672, Cur Loss: 0.22920662, Cur Avg Loss: 0.40110060, Log Avg loss: 0.41863949, Global Avg Loss: 1.70368424, Time: 0.0211 Steps: 27590, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002052, Sample Num: 32832, Cur Loss: 0.40326560, Cur Avg Loss: 0.40094924, Log Avg loss: 0.37004172, Global Avg Loss: 1.70320103, Time: 0.0250 Steps: 27600, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002062, Sample Num: 32992, Cur Loss: 0.30329433, Cur Avg Loss: 0.40014542, Log Avg loss: 0.23520054, Global Avg Loss: 1.70266934, Time: 0.0212 Steps: 27610, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002072, Sample Num: 33152, Cur Loss: 0.67081821, Cur Avg Loss: 0.40005214, Log Avg loss: 0.38081709, Global Avg Loss: 1.70219076, Time: 0.0213 Steps: 27620, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002082, Sample Num: 33312, Cur Loss: 0.19437864, Cur Avg Loss: 0.40022978, Log Avg loss: 0.43703822, Global Avg Loss: 1.70173287, Time: 0.0212 Steps: 27630, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002092, Sample Num: 33472, Cur Loss: 0.22358285, Cur Avg Loss: 0.40033927, Log Avg loss: 0.42313430, Global Avg Loss: 1.70127028, Time: 0.0212 Steps: 27640, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002102, Sample Num: 33632, Cur Loss: 0.24650544, Cur Avg Loss: 0.40015470, Log Avg loss: 0.36154310, Global Avg Loss: 1.70078575, Time: 0.0212 Steps: 27650, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002112, Sample Num: 33792, Cur Loss: 0.39528868, Cur Avg Loss: 0.39989837, Log Avg loss: 0.34601747, Global Avg Loss: 1.70029595, Time: 0.0212 Steps: 27660, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002122, Sample Num: 33952, Cur Loss: 0.21587294, Cur Avg Loss: 0.40007762, Log Avg loss: 0.43793493, Global Avg Loss: 1.69983973, Time: 0.0211 Steps: 27670, Updated lr: 0.000075 ***** Running evaluation checkpoint-27677 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-27677 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.108568, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.367483, "eval_total_loss": 258.340265, "eval_mae": 0.431974, "eval_mse": 0.367615, "eval_r2": 0.76632, "eval_sp_statistic": 0.8484, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.876078, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.700643, "test_total_loss": 351.722554, "test_mae": 0.544232, "test_mse": 0.70086, "test_r2": 0.547659, "test_sp_statistic": 0.767031, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.814837, "test_ps_pvalue": 0.0, "lr": 7.470175438596491e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.699505316038446, "train_cur_epoch_loss": 851.6079505980015, "train_cur_epoch_avg_loss": 0.40000373442837084, "train_cur_epoch_time": 45.1085684299469, "train_cur_epoch_avg_time": 0.0211876789243527, "epoch": 13, "step": 27677} ################################################## Training, Epoch: 0014, Batch: 000003, Sample Num: 48, Cur Loss: 0.19669737, Cur Avg Loss: 0.28979255, Log Avg loss: 0.35126238, Global Avg Loss: 1.69935253, Time: 0.0247 Steps: 27680, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000013, Sample Num: 208, Cur Loss: 0.39399126, Cur Avg Loss: 0.35510850, Log Avg loss: 0.37470328, Global Avg Loss: 1.69887414, Time: 0.0213 Steps: 27690, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000023, Sample Num: 368, Cur Loss: 0.25189078, Cur Avg Loss: 0.32889122, Log Avg loss: 0.29480876, Global Avg Loss: 1.69836726, Time: 0.0214 Steps: 27700, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000033, Sample Num: 528, Cur Loss: 0.40506467, Cur Avg Loss: 0.35240913, Log Avg loss: 0.40650031, Global Avg Loss: 1.69790105, Time: 0.0213 Steps: 27710, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000043, Sample Num: 688, Cur Loss: 0.32874435, Cur Avg Loss: 0.34482571, Log Avg loss: 0.31980043, Global Avg Loss: 1.69740390, Time: 0.0214 Steps: 27720, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000053, Sample Num: 848, Cur Loss: 1.14755797, Cur Avg Loss: 0.41442743, Log Avg loss: 0.71371480, Global Avg Loss: 1.69704916, Time: 0.0213 Steps: 27730, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000063, Sample Num: 1008, Cur Loss: 0.40060505, Cur Avg Loss: 0.43871695, Log Avg loss: 0.56745145, Global Avg Loss: 1.69664195, Time: 0.0213 Steps: 27740, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000073, Sample Num: 1168, Cur Loss: 0.11687092, Cur Avg Loss: 0.42888208, Log Avg loss: 0.36692239, Global Avg Loss: 1.69616278, Time: 0.0213 Steps: 27750, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000083, Sample Num: 1328, Cur Loss: 0.34830725, Cur Avg Loss: 0.42751457, Log Avg loss: 0.41753179, Global Avg Loss: 1.69570217, Time: 0.0211 Steps: 27760, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000093, Sample Num: 1488, Cur Loss: 0.74729085, Cur Avg Loss: 0.43866838, Log Avg loss: 0.53124500, Global Avg Loss: 1.69528285, Time: 0.0210 Steps: 27770, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000103, Sample Num: 1648, Cur Loss: 0.39721596, Cur Avg Loss: 0.46389179, Log Avg loss: 0.69846951, Global Avg Loss: 1.69492403, Time: 0.0211 Steps: 27780, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000113, Sample Num: 1808, Cur Loss: 0.30795452, Cur Avg Loss: 0.46206650, Log Avg loss: 0.44326594, Global Avg Loss: 1.69447363, Time: 0.0211 Steps: 27790, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000123, Sample Num: 1968, Cur Loss: 0.13310361, Cur Avg Loss: 0.45406905, Log Avg loss: 0.36369784, Global Avg Loss: 1.69399493, Time: 0.0211 Steps: 27800, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000133, Sample Num: 2128, Cur Loss: 0.74631584, Cur Avg Loss: 0.45008210, Log Avg loss: 0.40104262, Global Avg Loss: 1.69353001, Time: 0.0211 Steps: 27810, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000143, Sample Num: 2288, Cur Loss: 0.12515950, Cur Avg Loss: 0.44288353, Log Avg loss: 0.34714254, Global Avg Loss: 1.69304605, Time: 0.0211 Steps: 27820, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000153, Sample Num: 2448, Cur Loss: 0.30890143, Cur Avg Loss: 0.43215665, Log Avg loss: 0.27876227, Global Avg Loss: 1.69253786, Time: 0.0211 Steps: 27830, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000163, Sample Num: 2608, Cur Loss: 0.37591776, Cur Avg Loss: 0.42745313, Log Avg loss: 0.35548930, Global Avg Loss: 1.69205760, Time: 0.0211 Steps: 27840, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000173, Sample Num: 2768, Cur Loss: 0.51192582, Cur Avg Loss: 0.42413031, Log Avg loss: 0.36996845, Global Avg Loss: 1.69158288, Time: 0.0211 Steps: 27850, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000183, Sample Num: 2928, Cur Loss: 0.16939734, Cur Avg Loss: 0.42652552, Log Avg loss: 0.46796263, Global Avg Loss: 1.69114368, Time: 0.0211 Steps: 27860, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000193, Sample Num: 3088, Cur Loss: 0.09483980, Cur Avg Loss: 0.42199351, Log Avg loss: 0.33905766, Global Avg Loss: 1.69065854, Time: 0.0210 Steps: 27870, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000203, Sample Num: 3248, Cur Loss: 0.34246987, Cur Avg Loss: 0.41968994, Log Avg loss: 0.37523100, Global Avg Loss: 1.69018672, Time: 0.0210 Steps: 27880, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000213, Sample Num: 3408, Cur Loss: 0.34549356, Cur Avg Loss: 0.41557708, Log Avg loss: 0.33208602, Global Avg Loss: 1.68969977, Time: 0.0211 Steps: 27890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000223, Sample Num: 3568, Cur Loss: 0.63052344, Cur Avg Loss: 0.41762650, Log Avg loss: 0.46127911, Global Avg Loss: 1.68925947, Time: 0.0211 Steps: 27900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000233, Sample Num: 3728, Cur Loss: 0.28983566, Cur Avg Loss: 0.41631245, Log Avg loss: 0.38700921, Global Avg Loss: 1.68879289, Time: 0.0210 Steps: 27910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000243, Sample Num: 3888, Cur Loss: 0.18244404, Cur Avg Loss: 0.41527682, Log Avg loss: 0.39114656, Global Avg Loss: 1.68832811, Time: 0.0211 Steps: 27920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000253, Sample Num: 4048, Cur Loss: 0.27854475, Cur Avg Loss: 0.41064855, Log Avg loss: 0.29818172, Global Avg Loss: 1.68783039, Time: 0.0211 Steps: 27930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000263, Sample Num: 4208, Cur Loss: 0.33531842, Cur Avg Loss: 0.41109166, Log Avg loss: 0.42230233, Global Avg Loss: 1.68737744, Time: 0.0213 Steps: 27940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000273, Sample Num: 4368, Cur Loss: 0.23525766, Cur Avg Loss: 0.40838781, Log Avg loss: 0.33727655, Global Avg Loss: 1.68689440, Time: 0.0211 Steps: 27950, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000283, Sample Num: 4528, Cur Loss: 0.13750973, Cur Avg Loss: 0.40377002, Log Avg loss: 0.27770443, Global Avg Loss: 1.68639040, Time: 0.0212 Steps: 27960, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000293, Sample Num: 4688, Cur Loss: 0.89951617, Cur Avg Loss: 0.40450302, Log Avg loss: 0.42524677, Global Avg Loss: 1.68593951, Time: 0.0212 Steps: 27970, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000303, Sample Num: 4848, Cur Loss: 0.27046973, Cur Avg Loss: 0.40341892, Log Avg loss: 0.37165485, Global Avg Loss: 1.68546978, Time: 0.0211 Steps: 27980, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000313, Sample Num: 5008, Cur Loss: 0.47815689, Cur Avg Loss: 0.39939409, Log Avg loss: 0.27744167, Global Avg Loss: 1.68496674, Time: 0.0211 Steps: 27990, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000323, Sample Num: 5168, Cur Loss: 0.44664344, Cur Avg Loss: 0.39857577, Log Avg loss: 0.37296252, Global Avg Loss: 1.68449816, Time: 0.0211 Steps: 28000, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000333, Sample Num: 5328, Cur Loss: 0.05336235, Cur Avg Loss: 0.39438392, Log Avg loss: 0.25898701, Global Avg Loss: 1.68398924, Time: 0.0212 Steps: 28010, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000343, Sample Num: 5488, Cur Loss: 1.39686120, Cur Avg Loss: 0.39988767, Log Avg loss: 0.58316258, Global Avg Loss: 1.68359636, Time: 0.0211 Steps: 28020, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000353, Sample Num: 5648, Cur Loss: 0.34581396, Cur Avg Loss: 0.39994463, Log Avg loss: 0.40189847, Global Avg Loss: 1.68313910, Time: 0.0212 Steps: 28030, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000363, Sample Num: 5808, Cur Loss: 0.26557595, Cur Avg Loss: 0.39842567, Log Avg loss: 0.34480630, Global Avg Loss: 1.68266181, Time: 0.0211 Steps: 28040, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000373, Sample Num: 5968, Cur Loss: 0.39965516, Cur Avg Loss: 0.39549976, Log Avg loss: 0.28928920, Global Avg Loss: 1.68216506, Time: 0.0212 Steps: 28050, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000383, Sample Num: 6128, Cur Loss: 0.97072709, Cur Avg Loss: 0.39617141, Log Avg loss: 0.42122391, Global Avg Loss: 1.68171569, Time: 0.0211 Steps: 28060, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000393, Sample Num: 6288, Cur Loss: 0.41293848, Cur Avg Loss: 0.39556581, Log Avg loss: 0.37237141, Global Avg Loss: 1.68124923, Time: 0.0211 Steps: 28070, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000403, Sample Num: 6448, Cur Loss: 0.28691605, Cur Avg Loss: 0.39592762, Log Avg loss: 0.41014668, Global Avg Loss: 1.68079656, Time: 0.0212 Steps: 28080, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000413, Sample Num: 6608, Cur Loss: 0.47745860, Cur Avg Loss: 0.39441694, Log Avg loss: 0.33353679, Global Avg Loss: 1.68031694, Time: 0.0211 Steps: 28090, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000423, Sample Num: 6768, Cur Loss: 0.17721030, Cur Avg Loss: 0.39350893, Log Avg loss: 0.35600781, Global Avg Loss: 1.67984566, Time: 0.0212 Steps: 28100, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000433, Sample Num: 6928, Cur Loss: 0.42076522, Cur Avg Loss: 0.39344312, Log Avg loss: 0.39065941, Global Avg Loss: 1.67938703, Time: 0.0212 Steps: 28110, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000443, Sample Num: 7088, Cur Loss: 0.88870877, Cur Avg Loss: 0.39156210, Log Avg loss: 0.31011400, Global Avg Loss: 1.67890009, Time: 0.0211 Steps: 28120, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000453, Sample Num: 7248, Cur Loss: 0.36388135, Cur Avg Loss: 0.39049847, Log Avg loss: 0.34337982, Global Avg Loss: 1.67842533, Time: 0.0211 Steps: 28130, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000463, Sample Num: 7408, Cur Loss: 0.53046477, Cur Avg Loss: 0.39303142, Log Avg loss: 0.50777370, Global Avg Loss: 1.67800932, Time: 0.0211 Steps: 28140, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000473, Sample Num: 7568, Cur Loss: 0.68050724, Cur Avg Loss: 0.39600510, Log Avg loss: 0.53368692, Global Avg Loss: 1.67760281, Time: 0.0211 Steps: 28150, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000483, Sample Num: 7728, Cur Loss: 1.01064324, Cur Avg Loss: 0.39976250, Log Avg loss: 0.57748716, Global Avg Loss: 1.67721214, Time: 0.0211 Steps: 28160, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000493, Sample Num: 7888, Cur Loss: 0.14227983, Cur Avg Loss: 0.39939311, Log Avg loss: 0.38155161, Global Avg Loss: 1.67675220, Time: 0.0212 Steps: 28170, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000503, Sample Num: 8048, Cur Loss: 0.62005180, Cur Avg Loss: 0.39906897, Log Avg loss: 0.38308920, Global Avg Loss: 1.67629313, Time: 0.0212 Steps: 28180, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000513, Sample Num: 8208, Cur Loss: 0.20073891, Cur Avg Loss: 0.39769876, Log Avg loss: 0.32877688, Global Avg Loss: 1.67581512, Time: 0.0247 Steps: 28190, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000523, Sample Num: 8368, Cur Loss: 0.17216842, Cur Avg Loss: 0.39821465, Log Avg loss: 0.42467990, Global Avg Loss: 1.67537145, Time: 0.0211 Steps: 28200, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000533, Sample Num: 8528, Cur Loss: 0.18872452, Cur Avg Loss: 0.40045467, Log Avg loss: 0.51760765, Global Avg Loss: 1.67496104, Time: 0.0211 Steps: 28210, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000543, Sample Num: 8688, Cur Loss: 0.32177722, Cur Avg Loss: 0.39902437, Log Avg loss: 0.32278953, Global Avg Loss: 1.67448189, Time: 0.0211 Steps: 28220, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000553, Sample Num: 8848, Cur Loss: 0.38625726, Cur Avg Loss: 0.39753237, Log Avg loss: 0.31651641, Global Avg Loss: 1.67400085, Time: 0.0211 Steps: 28230, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000563, Sample Num: 9008, Cur Loss: 0.39797819, Cur Avg Loss: 0.39788563, Log Avg loss: 0.41742105, Global Avg Loss: 1.67355589, Time: 0.0211 Steps: 28240, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000573, Sample Num: 9168, Cur Loss: 0.14639352, Cur Avg Loss: 0.39743443, Log Avg loss: 0.37203209, Global Avg Loss: 1.67309517, Time: 0.0211 Steps: 28250, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000583, Sample Num: 9328, Cur Loss: 0.44279540, Cur Avg Loss: 0.39573233, Log Avg loss: 0.29820169, Global Avg Loss: 1.67260865, Time: 0.0211 Steps: 28260, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000593, Sample Num: 9488, Cur Loss: 0.11182705, Cur Avg Loss: 0.39442540, Log Avg loss: 0.31823136, Global Avg Loss: 1.67212957, Time: 0.0212 Steps: 28270, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000603, Sample Num: 9648, Cur Loss: 0.08031969, Cur Avg Loss: 0.39146044, Log Avg loss: 0.21563874, Global Avg Loss: 1.67161454, Time: 0.0211 Steps: 28280, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000613, Sample Num: 9808, Cur Loss: 0.37942091, Cur Avg Loss: 0.39088500, Log Avg loss: 0.35618602, Global Avg Loss: 1.67114956, Time: 0.0211 Steps: 28290, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000623, Sample Num: 9968, Cur Loss: 0.08641589, Cur Avg Loss: 0.39068287, Log Avg loss: 0.37829228, Global Avg Loss: 1.67069272, Time: 0.0211 Steps: 28300, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000633, Sample Num: 10128, Cur Loss: 0.91203403, Cur Avg Loss: 0.39065315, Log Avg loss: 0.38880111, Global Avg Loss: 1.67023992, Time: 0.0211 Steps: 28310, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000643, Sample Num: 10288, Cur Loss: 0.87623554, Cur Avg Loss: 0.39050569, Log Avg loss: 0.38117146, Global Avg Loss: 1.66978474, Time: 0.0211 Steps: 28320, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000653, Sample Num: 10448, Cur Loss: 0.19097391, Cur Avg Loss: 0.39094073, Log Avg loss: 0.41891436, Global Avg Loss: 1.66934320, Time: 0.0211 Steps: 28330, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000663, Sample Num: 10608, Cur Loss: 0.29060134, Cur Avg Loss: 0.39100096, Log Avg loss: 0.39493356, Global Avg Loss: 1.66889352, Time: 0.0211 Steps: 28340, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000673, Sample Num: 10768, Cur Loss: 0.21330771, Cur Avg Loss: 0.39108284, Log Avg loss: 0.39651190, Global Avg Loss: 1.66844470, Time: 0.0211 Steps: 28350, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000683, Sample Num: 10928, Cur Loss: 0.49499410, Cur Avg Loss: 0.38974086, Log Avg loss: 0.29942564, Global Avg Loss: 1.66796198, Time: 0.0211 Steps: 28360, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000693, Sample Num: 11088, Cur Loss: 0.43490756, Cur Avg Loss: 0.39013814, Log Avg loss: 0.41727227, Global Avg Loss: 1.66752113, Time: 0.0211 Steps: 28370, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000703, Sample Num: 11248, Cur Loss: 0.23869736, Cur Avg Loss: 0.38828944, Log Avg loss: 0.26017445, Global Avg Loss: 1.66702523, Time: 0.0211 Steps: 28380, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000713, Sample Num: 11408, Cur Loss: 0.98381913, Cur Avg Loss: 0.38780424, Log Avg loss: 0.35369436, Global Avg Loss: 1.66656263, Time: 0.0211 Steps: 28390, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000723, Sample Num: 11568, Cur Loss: 0.24095318, Cur Avg Loss: 0.38558574, Log Avg loss: 0.22740709, Global Avg Loss: 1.66605588, Time: 0.0211 Steps: 28400, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000733, Sample Num: 11728, Cur Loss: 0.29275447, Cur Avg Loss: 0.38474046, Log Avg loss: 0.32362626, Global Avg Loss: 1.66558336, Time: 0.0211 Steps: 28410, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000743, Sample Num: 11888, Cur Loss: 0.43676743, Cur Avg Loss: 0.38494881, Log Avg loss: 0.40022139, Global Avg Loss: 1.66513813, Time: 0.0211 Steps: 28420, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000753, Sample Num: 12048, Cur Loss: 0.18250647, Cur Avg Loss: 0.38552768, Log Avg loss: 0.42853732, Global Avg Loss: 1.66470316, Time: 0.0211 Steps: 28430, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000763, Sample Num: 12208, Cur Loss: 0.24902946, Cur Avg Loss: 0.38435777, Log Avg loss: 0.29626343, Global Avg Loss: 1.66422200, Time: 0.0211 Steps: 28440, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000773, Sample Num: 12368, Cur Loss: 0.64496362, Cur Avg Loss: 0.38379244, Log Avg loss: 0.34065821, Global Avg Loss: 1.66375677, Time: 0.0249 Steps: 28450, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000783, Sample Num: 12528, Cur Loss: 0.17286165, Cur Avg Loss: 0.38496676, Log Avg loss: 0.47574168, Global Avg Loss: 1.66333934, Time: 0.0212 Steps: 28460, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000793, Sample Num: 12688, Cur Loss: 0.40031844, Cur Avg Loss: 0.38525485, Log Avg loss: 0.40781228, Global Avg Loss: 1.66289834, Time: 0.0212 Steps: 28470, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000803, Sample Num: 12848, Cur Loss: 0.53536201, Cur Avg Loss: 0.38470338, Log Avg loss: 0.34097150, Global Avg Loss: 1.66243418, Time: 0.0212 Steps: 28480, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000813, Sample Num: 13008, Cur Loss: 0.28357822, Cur Avg Loss: 0.38369914, Log Avg loss: 0.30305880, Global Avg Loss: 1.66195704, Time: 0.0212 Steps: 28490, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000823, Sample Num: 13168, Cur Loss: 0.64729548, Cur Avg Loss: 0.38440436, Log Avg loss: 0.44173841, Global Avg Loss: 1.66152889, Time: 0.0211 Steps: 28500, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000833, Sample Num: 13328, Cur Loss: 0.25721234, Cur Avg Loss: 0.38394732, Log Avg loss: 0.34633302, Global Avg Loss: 1.66106758, Time: 0.0211 Steps: 28510, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000843, Sample Num: 13488, Cur Loss: 0.39081025, Cur Avg Loss: 0.38328072, Log Avg loss: 0.32775285, Global Avg Loss: 1.66060008, Time: 0.0212 Steps: 28520, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000853, Sample Num: 13648, Cur Loss: 0.40561163, Cur Avg Loss: 0.38568466, Log Avg loss: 0.58833726, Global Avg Loss: 1.66022424, Time: 0.0212 Steps: 28530, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000863, Sample Num: 13808, Cur Loss: 0.42673165, Cur Avg Loss: 0.38578498, Log Avg loss: 0.39434206, Global Avg Loss: 1.65978070, Time: 0.0211 Steps: 28540, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000873, Sample Num: 13968, Cur Loss: 0.27436191, Cur Avg Loss: 0.38597609, Log Avg loss: 0.40246876, Global Avg Loss: 1.65934031, Time: 0.0212 Steps: 28550, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000883, Sample Num: 14128, Cur Loss: 0.64739376, Cur Avg Loss: 0.38711331, Log Avg loss: 0.48639259, Global Avg Loss: 1.65892961, Time: 0.0212 Steps: 28560, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000893, Sample Num: 14288, Cur Loss: 0.40989655, Cur Avg Loss: 0.38635429, Log Avg loss: 0.31933281, Global Avg Loss: 1.65846073, Time: 0.0211 Steps: 28570, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000903, Sample Num: 14448, Cur Loss: 0.10313606, Cur Avg Loss: 0.38597050, Log Avg loss: 0.35169807, Global Avg Loss: 1.65800350, Time: 0.0211 Steps: 28580, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000913, Sample Num: 14608, Cur Loss: 0.15249793, Cur Avg Loss: 0.38576478, Log Avg loss: 0.36718812, Global Avg Loss: 1.65755201, Time: 0.0211 Steps: 28590, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000923, Sample Num: 14768, Cur Loss: 0.45851555, Cur Avg Loss: 0.38520302, Log Avg loss: 0.33391430, Global Avg Loss: 1.65708920, Time: 0.0211 Steps: 28600, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000933, Sample Num: 14928, Cur Loss: 0.21669373, Cur Avg Loss: 0.38431453, Log Avg loss: 0.30230696, Global Avg Loss: 1.65661566, Time: 0.0212 Steps: 28610, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000943, Sample Num: 15088, Cur Loss: 0.25596219, Cur Avg Loss: 0.38511362, Log Avg loss: 0.45966862, Global Avg Loss: 1.65619744, Time: 0.0212 Steps: 28620, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000953, Sample Num: 15248, Cur Loss: 0.22177923, Cur Avg Loss: 0.38544329, Log Avg loss: 0.41653125, Global Avg Loss: 1.65576445, Time: 0.0212 Steps: 28630, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000963, Sample Num: 15408, Cur Loss: 0.55709440, Cur Avg Loss: 0.38465590, Log Avg loss: 0.30961803, Global Avg Loss: 1.65529442, Time: 0.0212 Steps: 28640, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000973, Sample Num: 15568, Cur Loss: 0.51069170, Cur Avg Loss: 0.38570300, Log Avg loss: 0.48653832, Global Avg Loss: 1.65488648, Time: 0.0212 Steps: 28650, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000983, Sample Num: 15728, Cur Loss: 0.19138989, Cur Avg Loss: 0.38475031, Log Avg loss: 0.29205399, Global Avg Loss: 1.65441096, Time: 0.0212 Steps: 28660, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000993, Sample Num: 15888, Cur Loss: 0.37882549, Cur Avg Loss: 0.38460853, Log Avg loss: 0.37067106, Global Avg Loss: 1.65396320, Time: 0.0212 Steps: 28670, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001003, Sample Num: 16048, Cur Loss: 0.33240977, Cur Avg Loss: 0.38330747, Log Avg loss: 0.25411260, Global Avg Loss: 1.65347511, Time: 0.0212 Steps: 28680, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001013, Sample Num: 16208, Cur Loss: 0.47418633, Cur Avg Loss: 0.38262271, Log Avg loss: 0.31394175, Global Avg Loss: 1.65300821, Time: 0.0212 Steps: 28690, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001023, Sample Num: 16368, Cur Loss: 0.64450347, Cur Avg Loss: 0.38425296, Log Avg loss: 0.54939713, Global Avg Loss: 1.65262367, Time: 0.0212 Steps: 28700, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001033, Sample Num: 16528, Cur Loss: 0.99228710, Cur Avg Loss: 0.38473350, Log Avg loss: 0.43389290, Global Avg Loss: 1.65219918, Time: 0.0212 Steps: 28710, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001043, Sample Num: 16688, Cur Loss: 0.16565123, Cur Avg Loss: 0.38496056, Log Avg loss: 0.40841551, Global Avg Loss: 1.65176610, Time: 0.0211 Steps: 28720, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001053, Sample Num: 16848, Cur Loss: 0.34946477, Cur Avg Loss: 0.38500693, Log Avg loss: 0.38984362, Global Avg Loss: 1.65132687, Time: 0.0212 Steps: 28730, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001063, Sample Num: 17008, Cur Loss: 0.40609881, Cur Avg Loss: 0.38480749, Log Avg loss: 0.36380566, Global Avg Loss: 1.65087888, Time: 0.0212 Steps: 28740, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001073, Sample Num: 17168, Cur Loss: 0.54300463, Cur Avg Loss: 0.38470974, Log Avg loss: 0.37431954, Global Avg Loss: 1.65043486, Time: 0.0211 Steps: 28750, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001083, Sample Num: 17328, Cur Loss: 0.18434526, Cur Avg Loss: 0.38470432, Log Avg loss: 0.38412226, Global Avg Loss: 1.64999456, Time: 0.0211 Steps: 28760, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001093, Sample Num: 17488, Cur Loss: 0.44917509, Cur Avg Loss: 0.38533430, Log Avg loss: 0.45356118, Global Avg Loss: 1.64957869, Time: 0.0211 Steps: 28770, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001103, Sample Num: 17648, Cur Loss: 0.39245600, Cur Avg Loss: 0.38553004, Log Avg loss: 0.40692511, Global Avg Loss: 1.64914692, Time: 0.0211 Steps: 28780, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001113, Sample Num: 17808, Cur Loss: 0.15719537, Cur Avg Loss: 0.38544768, Log Avg loss: 0.37636311, Global Avg Loss: 1.64870482, Time: 0.0212 Steps: 28790, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001123, Sample Num: 17968, Cur Loss: 0.11000958, Cur Avg Loss: 0.38526814, Log Avg loss: 0.36528476, Global Avg Loss: 1.64825919, Time: 0.0211 Steps: 28800, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001133, Sample Num: 18128, Cur Loss: 0.34876090, Cur Avg Loss: 0.38508488, Log Avg loss: 0.36450571, Global Avg Loss: 1.64781360, Time: 0.0211 Steps: 28810, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001143, Sample Num: 18288, Cur Loss: 0.49627119, Cur Avg Loss: 0.38519424, Log Avg loss: 0.39758409, Global Avg Loss: 1.64737979, Time: 0.0211 Steps: 28820, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001153, Sample Num: 18448, Cur Loss: 0.37030578, Cur Avg Loss: 0.38463105, Log Avg loss: 0.32025830, Global Avg Loss: 1.64691947, Time: 0.0211 Steps: 28830, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001163, Sample Num: 18608, Cur Loss: 0.31463531, Cur Avg Loss: 0.38440510, Log Avg loss: 0.35835312, Global Avg Loss: 1.64647267, Time: 0.0211 Steps: 28840, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001173, Sample Num: 18768, Cur Loss: 0.16557984, Cur Avg Loss: 0.38355251, Log Avg loss: 0.28439606, Global Avg Loss: 1.64600054, Time: 0.0211 Steps: 28850, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001183, Sample Num: 18928, Cur Loss: 0.31795901, Cur Avg Loss: 0.38238399, Log Avg loss: 0.24531740, Global Avg Loss: 1.64551521, Time: 0.0211 Steps: 28860, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001193, Sample Num: 19088, Cur Loss: 0.28171521, Cur Avg Loss: 0.38301823, Log Avg loss: 0.45804800, Global Avg Loss: 1.64510389, Time: 0.0211 Steps: 28870, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001203, Sample Num: 19248, Cur Loss: 0.41943461, Cur Avg Loss: 0.38303373, Log Avg loss: 0.38488327, Global Avg Loss: 1.64466753, Time: 0.0211 Steps: 28880, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001213, Sample Num: 19408, Cur Loss: 0.29433736, Cur Avg Loss: 0.38321710, Log Avg loss: 0.40527604, Global Avg Loss: 1.64423852, Time: 0.0211 Steps: 28890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001223, Sample Num: 19568, Cur Loss: 0.33962876, Cur Avg Loss: 0.38303098, Log Avg loss: 0.36045547, Global Avg Loss: 1.64379431, Time: 0.0211 Steps: 28900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001233, Sample Num: 19728, Cur Loss: 0.17561051, Cur Avg Loss: 0.38244611, Log Avg loss: 0.31091668, Global Avg Loss: 1.64333326, Time: 0.0211 Steps: 28910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001243, Sample Num: 19888, Cur Loss: 0.25661528, Cur Avg Loss: 0.38291725, Log Avg loss: 0.44100839, Global Avg Loss: 1.64291752, Time: 0.0211 Steps: 28920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001253, Sample Num: 20048, Cur Loss: 0.10590076, Cur Avg Loss: 0.38289933, Log Avg loss: 0.38067214, Global Avg Loss: 1.64248121, Time: 0.0211 Steps: 28930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001263, Sample Num: 20208, Cur Loss: 0.67952764, Cur Avg Loss: 0.38302815, Log Avg loss: 0.39916888, Global Avg Loss: 1.64205160, Time: 0.0211 Steps: 28940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001273, Sample Num: 20368, Cur Loss: 0.74327481, Cur Avg Loss: 0.38231913, Log Avg loss: 0.29276974, Global Avg Loss: 1.64158552, Time: 0.0211 Steps: 28950, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001283, Sample Num: 20528, Cur Loss: 0.21100442, Cur Avg Loss: 0.38196862, Log Avg loss: 0.33734881, Global Avg Loss: 1.64113516, Time: 0.0248 Steps: 28960, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001293, Sample Num: 20688, Cur Loss: 0.36021340, Cur Avg Loss: 0.38259778, Log Avg loss: 0.46331868, Global Avg Loss: 1.64072860, Time: 0.0211 Steps: 28970, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001303, Sample Num: 20848, Cur Loss: 0.34992141, Cur Avg Loss: 0.38236434, Log Avg loss: 0.35218145, Global Avg Loss: 1.64028397, Time: 0.0211 Steps: 28980, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001313, Sample Num: 21008, Cur Loss: 0.48053053, Cur Avg Loss: 0.38335380, Log Avg loss: 0.51228054, Global Avg Loss: 1.63989487, Time: 0.0211 Steps: 28990, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001323, Sample Num: 21168, Cur Loss: 0.14301637, Cur Avg Loss: 0.38317854, Log Avg loss: 0.36016598, Global Avg Loss: 1.63945358, Time: 0.0211 Steps: 29000, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001333, Sample Num: 21328, Cur Loss: 0.50824916, Cur Avg Loss: 0.38242584, Log Avg loss: 0.28284413, Global Avg Loss: 1.63898595, Time: 0.0211 Steps: 29010, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001343, Sample Num: 21488, Cur Loss: 0.34601820, Cur Avg Loss: 0.38209142, Log Avg loss: 0.33751304, Global Avg Loss: 1.63853747, Time: 0.0211 Steps: 29020, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001353, Sample Num: 21648, Cur Loss: 0.67767239, Cur Avg Loss: 0.38316537, Log Avg loss: 0.52739691, Global Avg Loss: 1.63815472, Time: 0.0211 Steps: 29030, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001363, Sample Num: 21808, Cur Loss: 0.43425292, Cur Avg Loss: 0.38401069, Log Avg loss: 0.49838205, Global Avg Loss: 1.63776223, Time: 0.0211 Steps: 29040, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001373, Sample Num: 21968, Cur Loss: 0.63312882, Cur Avg Loss: 0.38405755, Log Avg loss: 0.39044500, Global Avg Loss: 1.63733286, Time: 0.0211 Steps: 29050, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001383, Sample Num: 22128, Cur Loss: 0.19590181, Cur Avg Loss: 0.38473115, Log Avg loss: 0.47721627, Global Avg Loss: 1.63693365, Time: 0.0211 Steps: 29060, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001393, Sample Num: 22288, Cur Loss: 0.17839833, Cur Avg Loss: 0.38441163, Log Avg loss: 0.34022242, Global Avg Loss: 1.63648758, Time: 0.0211 Steps: 29070, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001403, Sample Num: 22448, Cur Loss: 0.10598168, Cur Avg Loss: 0.38439384, Log Avg loss: 0.38191593, Global Avg Loss: 1.63605616, Time: 0.0211 Steps: 29080, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001413, Sample Num: 22608, Cur Loss: 0.33502951, Cur Avg Loss: 0.38390666, Log Avg loss: 0.31555503, Global Avg Loss: 1.63560223, Time: 0.0211 Steps: 29090, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001423, Sample Num: 22768, Cur Loss: 0.30087009, Cur Avg Loss: 0.38421163, Log Avg loss: 0.42730432, Global Avg Loss: 1.63518700, Time: 0.0211 Steps: 29100, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001433, Sample Num: 22928, Cur Loss: 0.10863337, Cur Avg Loss: 0.38378168, Log Avg loss: 0.32259915, Global Avg Loss: 1.63473610, Time: 0.0211 Steps: 29110, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001443, Sample Num: 23088, Cur Loss: 0.78217971, Cur Avg Loss: 0.38424055, Log Avg loss: 0.44999607, Global Avg Loss: 1.63432925, Time: 0.0211 Steps: 29120, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001453, Sample Num: 23248, Cur Loss: 0.22452927, Cur Avg Loss: 0.38507572, Log Avg loss: 0.50559180, Global Avg Loss: 1.63394177, Time: 0.0211 Steps: 29130, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001463, Sample Num: 23408, Cur Loss: 0.21705592, Cur Avg Loss: 0.38508385, Log Avg loss: 0.38626438, Global Avg Loss: 1.63351360, Time: 0.0211 Steps: 29140, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001473, Sample Num: 23568, Cur Loss: 0.76002038, Cur Avg Loss: 0.38471057, Log Avg loss: 0.33009978, Global Avg Loss: 1.63306646, Time: 0.0211 Steps: 29150, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001483, Sample Num: 23728, Cur Loss: 0.14870881, Cur Avg Loss: 0.38426236, Log Avg loss: 0.31824057, Global Avg Loss: 1.63261556, Time: 0.0211 Steps: 29160, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001493, Sample Num: 23888, Cur Loss: 0.50498462, Cur Avg Loss: 0.38388879, Log Avg loss: 0.32848847, Global Avg Loss: 1.63216848, Time: 0.0211 Steps: 29170, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001503, Sample Num: 24048, Cur Loss: 0.30399349, Cur Avg Loss: 0.38421032, Log Avg loss: 0.43221526, Global Avg Loss: 1.63175726, Time: 0.0211 Steps: 29180, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001513, Sample Num: 24208, Cur Loss: 0.71064317, Cur Avg Loss: 0.38451352, Log Avg loss: 0.43008495, Global Avg Loss: 1.63134558, Time: 0.0211 Steps: 29190, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001523, Sample Num: 24368, Cur Loss: 0.29250252, Cur Avg Loss: 0.38430084, Log Avg loss: 0.35212133, Global Avg Loss: 1.63090749, Time: 0.0211 Steps: 29200, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001533, Sample Num: 24528, Cur Loss: 0.30505788, Cur Avg Loss: 0.38403908, Log Avg loss: 0.34417324, Global Avg Loss: 1.63046698, Time: 0.0211 Steps: 29210, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001543, Sample Num: 24688, Cur Loss: 0.29129910, Cur Avg Loss: 0.38343061, Log Avg loss: 0.29015236, Global Avg Loss: 1.63000828, Time: 0.0213 Steps: 29220, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001553, Sample Num: 24848, Cur Loss: 0.11167158, Cur Avg Loss: 0.38281918, Log Avg loss: 0.28847585, Global Avg Loss: 1.62954933, Time: 0.0210 Steps: 29230, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001563, Sample Num: 25008, Cur Loss: 1.16639888, Cur Avg Loss: 0.38313324, Log Avg loss: 0.43190731, Global Avg Loss: 1.62913974, Time: 0.0210 Steps: 29240, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001573, Sample Num: 25168, Cur Loss: 0.19535814, Cur Avg Loss: 0.38327511, Log Avg loss: 0.40544805, Global Avg Loss: 1.62872138, Time: 0.0211 Steps: 29250, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001583, Sample Num: 25328, Cur Loss: 0.46854827, Cur Avg Loss: 0.38292435, Log Avg loss: 0.32775030, Global Avg Loss: 1.62827676, Time: 0.0210 Steps: 29260, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001593, Sample Num: 25488, Cur Loss: 0.14032385, Cur Avg Loss: 0.38269139, Log Avg loss: 0.34581342, Global Avg Loss: 1.62783861, Time: 0.0210 Steps: 29270, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001603, Sample Num: 25648, Cur Loss: 0.10792337, Cur Avg Loss: 0.38223469, Log Avg loss: 0.30948242, Global Avg Loss: 1.62738835, Time: 0.0210 Steps: 29280, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001613, Sample Num: 25808, Cur Loss: 0.45596370, Cur Avg Loss: 0.38168887, Log Avg loss: 0.29419403, Global Avg Loss: 1.62693318, Time: 0.0210 Steps: 29290, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001623, Sample Num: 25968, Cur Loss: 0.38507676, Cur Avg Loss: 0.38105141, Log Avg loss: 0.27822976, Global Avg Loss: 1.62647287, Time: 0.0210 Steps: 29300, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001633, Sample Num: 26128, Cur Loss: 0.28265178, Cur Avg Loss: 0.38056708, Log Avg loss: 0.30196018, Global Avg Loss: 1.62602097, Time: 0.0211 Steps: 29310, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001643, Sample Num: 26288, Cur Loss: 0.31362203, Cur Avg Loss: 0.38065880, Log Avg loss: 0.39563617, Global Avg Loss: 1.62560133, Time: 0.0210 Steps: 29320, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001653, Sample Num: 26448, Cur Loss: 0.23811260, Cur Avg Loss: 0.38036116, Log Avg loss: 0.33145893, Global Avg Loss: 1.62516010, Time: 0.0210 Steps: 29330, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001663, Sample Num: 26608, Cur Loss: 0.23865248, Cur Avg Loss: 0.38018783, Log Avg loss: 0.35153629, Global Avg Loss: 1.62472601, Time: 0.0210 Steps: 29340, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001673, Sample Num: 26768, Cur Loss: 0.26961410, Cur Avg Loss: 0.38019179, Log Avg loss: 0.38085003, Global Avg Loss: 1.62430220, Time: 0.0210 Steps: 29350, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001683, Sample Num: 26928, Cur Loss: 0.28292099, Cur Avg Loss: 0.37965070, Log Avg loss: 0.28912769, Global Avg Loss: 1.62384744, Time: 0.0210 Steps: 29360, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001693, Sample Num: 27088, Cur Loss: 0.51735723, Cur Avg Loss: 0.37962240, Log Avg loss: 0.37485841, Global Avg Loss: 1.62342218, Time: 0.0212 Steps: 29370, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001703, Sample Num: 27248, Cur Loss: 0.51005805, Cur Avg Loss: 0.37940755, Log Avg loss: 0.34303471, Global Avg Loss: 1.62298637, Time: 0.0210 Steps: 29380, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001713, Sample Num: 27408, Cur Loss: 0.47615260, Cur Avg Loss: 0.37940383, Log Avg loss: 0.37877015, Global Avg Loss: 1.62256303, Time: 0.0210 Steps: 29390, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001723, Sample Num: 27568, Cur Loss: 0.57771856, Cur Avg Loss: 0.37950202, Log Avg loss: 0.39632175, Global Avg Loss: 1.62214594, Time: 0.0211 Steps: 29400, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001733, Sample Num: 27728, Cur Loss: 0.36837888, Cur Avg Loss: 0.37927321, Log Avg loss: 0.33984827, Global Avg Loss: 1.62170993, Time: 0.0210 Steps: 29410, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001743, Sample Num: 27888, Cur Loss: 0.20486096, Cur Avg Loss: 0.37880185, Log Avg loss: 0.29711667, Global Avg Loss: 1.62125970, Time: 0.0211 Steps: 29420, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001753, Sample Num: 28048, Cur Loss: 0.28699583, Cur Avg Loss: 0.37822585, Log Avg loss: 0.27782871, Global Avg Loss: 1.62080321, Time: 0.0210 Steps: 29430, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001763, Sample Num: 28208, Cur Loss: 0.20745409, Cur Avg Loss: 0.37761065, Log Avg loss: 0.26976530, Global Avg Loss: 1.62034430, Time: 0.0210 Steps: 29440, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001773, Sample Num: 28368, Cur Loss: 0.51360488, Cur Avg Loss: 0.37762690, Log Avg loss: 0.38049262, Global Avg Loss: 1.61992330, Time: 0.0210 Steps: 29450, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001783, Sample Num: 28528, Cur Loss: 0.61348909, Cur Avg Loss: 0.37765186, Log Avg loss: 0.38207745, Global Avg Loss: 1.61950312, Time: 0.0210 Steps: 29460, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001793, Sample Num: 28688, Cur Loss: 0.32999772, Cur Avg Loss: 0.37771274, Log Avg loss: 0.38856662, Global Avg Loss: 1.61908543, Time: 0.0248 Steps: 29470, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001803, Sample Num: 28848, Cur Loss: 0.70029074, Cur Avg Loss: 0.37832241, Log Avg loss: 0.48763637, Global Avg Loss: 1.61870163, Time: 0.0211 Steps: 29480, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001813, Sample Num: 29008, Cur Loss: 0.10972011, Cur Avg Loss: 0.37766878, Log Avg loss: 0.25982030, Global Avg Loss: 1.61824083, Time: 0.0211 Steps: 29490, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001823, Sample Num: 29168, Cur Loss: 0.53395903, Cur Avg Loss: 0.37800991, Log Avg loss: 0.43985508, Global Avg Loss: 1.61784138, Time: 0.0210 Steps: 29500, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001833, Sample Num: 29328, Cur Loss: 0.13910219, Cur Avg Loss: 0.37744366, Log Avg loss: 0.27421785, Global Avg Loss: 1.61738607, Time: 0.0210 Steps: 29510, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001843, Sample Num: 29488, Cur Loss: 0.41690886, Cur Avg Loss: 0.37759056, Log Avg loss: 0.40451685, Global Avg Loss: 1.61697520, Time: 0.0210 Steps: 29520, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001853, Sample Num: 29648, Cur Loss: 0.23578592, Cur Avg Loss: 0.37769684, Log Avg loss: 0.39728388, Global Avg Loss: 1.61656217, Time: 0.0211 Steps: 29530, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001863, Sample Num: 29808, Cur Loss: 0.80424309, Cur Avg Loss: 0.37815270, Log Avg loss: 0.46262374, Global Avg Loss: 1.61617153, Time: 0.0210 Steps: 29540, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001873, Sample Num: 29968, Cur Loss: 0.55980301, Cur Avg Loss: 0.37805000, Log Avg loss: 0.35891641, Global Avg Loss: 1.61574607, Time: 0.0210 Steps: 29550, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001883, Sample Num: 30128, Cur Loss: 0.10562932, Cur Avg Loss: 0.37776252, Log Avg loss: 0.32391735, Global Avg Loss: 1.61530905, Time: 0.0210 Steps: 29560, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001893, Sample Num: 30288, Cur Loss: 0.61979413, Cur Avg Loss: 0.37749169, Log Avg loss: 0.32649414, Global Avg Loss: 1.61487320, Time: 0.0210 Steps: 29570, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001903, Sample Num: 30448, Cur Loss: 0.45087156, Cur Avg Loss: 0.37723282, Log Avg loss: 0.32823013, Global Avg Loss: 1.61443822, Time: 0.0210 Steps: 29580, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001913, Sample Num: 30608, Cur Loss: 0.40908265, Cur Avg Loss: 0.37680663, Log Avg loss: 0.29570184, Global Avg Loss: 1.61399256, Time: 0.0210 Steps: 29590, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001923, Sample Num: 30768, Cur Loss: 0.31298000, Cur Avg Loss: 0.37680791, Log Avg loss: 0.37705226, Global Avg Loss: 1.61357467, Time: 0.0210 Steps: 29600, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001933, Sample Num: 30928, Cur Loss: 0.49951747, Cur Avg Loss: 0.37668681, Log Avg loss: 0.35340025, Global Avg Loss: 1.61314908, Time: 0.0210 Steps: 29610, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001943, Sample Num: 31088, Cur Loss: 0.27995372, Cur Avg Loss: 0.37691620, Log Avg loss: 0.42125805, Global Avg Loss: 1.61274669, Time: 0.0210 Steps: 29620, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001953, Sample Num: 31248, Cur Loss: 0.12933107, Cur Avg Loss: 0.37671935, Log Avg loss: 0.33847119, Global Avg Loss: 1.61231662, Time: 0.0210 Steps: 29630, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001963, Sample Num: 31408, Cur Loss: 0.57897139, Cur Avg Loss: 0.37650584, Log Avg loss: 0.33480577, Global Avg Loss: 1.61188561, Time: 0.0210 Steps: 29640, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001973, Sample Num: 31568, Cur Loss: 0.44701469, Cur Avg Loss: 0.37658159, Log Avg loss: 0.39145276, Global Avg Loss: 1.61147400, Time: 0.0210 Steps: 29650, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001983, Sample Num: 31728, Cur Loss: 0.31105334, Cur Avg Loss: 0.37761408, Log Avg loss: 0.58132308, Global Avg Loss: 1.61112668, Time: 0.0210 Steps: 29660, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001993, Sample Num: 31888, Cur Loss: 1.79311037, Cur Avg Loss: 0.37864546, Log Avg loss: 0.58316859, Global Avg Loss: 1.61078022, Time: 0.0210 Steps: 29670, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002003, Sample Num: 32048, Cur Loss: 0.38442758, Cur Avg Loss: 0.37905930, Log Avg loss: 0.46153868, Global Avg Loss: 1.61039301, Time: 0.0208 Steps: 29680, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002013, Sample Num: 32208, Cur Loss: 0.21139351, Cur Avg Loss: 0.37878481, Log Avg loss: 0.32380330, Global Avg Loss: 1.60995966, Time: 0.0208 Steps: 29690, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002023, Sample Num: 32368, Cur Loss: 0.17026545, Cur Avg Loss: 0.37878946, Log Avg loss: 0.37972643, Global Avg Loss: 1.60954544, Time: 0.0207 Steps: 29700, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002033, Sample Num: 32528, Cur Loss: 0.25438923, Cur Avg Loss: 0.37925703, Log Avg loss: 0.47384628, Global Avg Loss: 1.60916318, Time: 0.0207 Steps: 29710, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002043, Sample Num: 32688, Cur Loss: 0.49243766, Cur Avg Loss: 0.37965085, Log Avg loss: 0.45971502, Global Avg Loss: 1.60877642, Time: 0.0207 Steps: 29720, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002053, Sample Num: 32848, Cur Loss: 0.79493380, Cur Avg Loss: 0.37972861, Log Avg loss: 0.39561347, Global Avg Loss: 1.60836836, Time: 0.0246 Steps: 29730, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002063, Sample Num: 33008, Cur Loss: 0.32378289, Cur Avg Loss: 0.38001231, Log Avg loss: 0.43825589, Global Avg Loss: 1.60797492, Time: 0.0209 Steps: 29740, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002073, Sample Num: 33168, Cur Loss: 0.41922170, Cur Avg Loss: 0.37978356, Log Avg loss: 0.33259311, Global Avg Loss: 1.60754622, Time: 0.0209 Steps: 29750, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002083, Sample Num: 33328, Cur Loss: 0.25181082, Cur Avg Loss: 0.37919475, Log Avg loss: 0.25713394, Global Avg Loss: 1.60709245, Time: 0.0209 Steps: 29760, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002093, Sample Num: 33488, Cur Loss: 0.16482797, Cur Avg Loss: 0.37882169, Log Avg loss: 0.30111272, Global Avg Loss: 1.60665376, Time: 0.0209 Steps: 29770, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002103, Sample Num: 33648, Cur Loss: 0.47850710, Cur Avg Loss: 0.37878071, Log Avg loss: 0.37020367, Global Avg Loss: 1.60623856, Time: 0.0209 Steps: 29780, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002113, Sample Num: 33808, Cur Loss: 0.37647882, Cur Avg Loss: 0.37861909, Log Avg loss: 0.34463145, Global Avg Loss: 1.60581506, Time: 0.0209 Steps: 29790, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002123, Sample Num: 33968, Cur Loss: 0.22220030, Cur Avg Loss: 0.37814981, Log Avg loss: 0.27899027, Global Avg Loss: 1.60536982, Time: 0.0209 Steps: 29800, Updated lr: 0.000073 ***** Running evaluation checkpoint-29806 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-29806 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.044045, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.357514, "eval_total_loss": 251.332091, "eval_mae": 0.415159, "eval_mse": 0.357647, "eval_r2": 0.772656, "eval_sp_statistic": 0.84931, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.882882, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.801869, "test_total_loss": 402.538284, "test_mae": 0.62148, "test_mse": 0.802072, "test_r2": 0.482335, "test_sp_statistic": 0.777572, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.829656, "test_ps_pvalue": 0.0, "lr": 7.26827880512091e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.605135107254329, "train_cur_epoch_loss": 805.4483748264611, "train_cur_epoch_avg_loss": 0.3783223930608084, "train_cur_epoch_time": 45.044045209884644, "train_cur_epoch_avg_time": 0.021157372104220123, "epoch": 14, "step": 29806} ################################################## Training, Epoch: 0015, Batch: 000004, Sample Num: 64, Cur Loss: 0.14570902, Cur Avg Loss: 0.26787194, Log Avg loss: 0.37078213, Global Avg Loss: 1.60495567, Time: 0.0245 Steps: 29810, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000014, Sample Num: 224, Cur Loss: 0.27596620, Cur Avg Loss: 0.26866565, Log Avg loss: 0.26898313, Global Avg Loss: 1.60450766, Time: 0.0208 Steps: 29820, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000024, Sample Num: 384, Cur Loss: 0.05210892, Cur Avg Loss: 0.23558088, Log Avg loss: 0.18926220, Global Avg Loss: 1.60403322, Time: 0.0208 Steps: 29830, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000034, Sample Num: 544, Cur Loss: 0.31543702, Cur Avg Loss: 0.27583408, Log Avg loss: 0.37244179, Global Avg Loss: 1.60362049, Time: 0.0208 Steps: 29840, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000044, Sample Num: 704, Cur Loss: 0.30804068, Cur Avg Loss: 0.28654377, Log Avg loss: 0.32295670, Global Avg Loss: 1.60319146, Time: 0.0207 Steps: 29850, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000054, Sample Num: 864, Cur Loss: 0.22114360, Cur Avg Loss: 0.29052104, Log Avg loss: 0.30802104, Global Avg Loss: 1.60275771, Time: 0.0208 Steps: 29860, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000064, Sample Num: 1024, Cur Loss: 0.40235025, Cur Avg Loss: 0.30003071, Log Avg loss: 0.35138289, Global Avg Loss: 1.60233877, Time: 0.0208 Steps: 29870, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000074, Sample Num: 1184, Cur Loss: 0.23678499, Cur Avg Loss: 0.30732612, Log Avg loss: 0.35401674, Global Avg Loss: 1.60192099, Time: 0.0208 Steps: 29880, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000084, Sample Num: 1344, Cur Loss: 0.21837212, Cur Avg Loss: 0.30707067, Log Avg loss: 0.30518041, Global Avg Loss: 1.60148715, Time: 0.0208 Steps: 29890, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000094, Sample Num: 1504, Cur Loss: 0.24910660, Cur Avg Loss: 0.32552594, Log Avg loss: 0.48055016, Global Avg Loss: 1.60111226, Time: 0.0208 Steps: 29900, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000104, Sample Num: 1664, Cur Loss: 0.39607885, Cur Avg Loss: 0.33311185, Log Avg loss: 0.40441938, Global Avg Loss: 1.60071216, Time: 0.0208 Steps: 29910, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000114, Sample Num: 1824, Cur Loss: 0.81000525, Cur Avg Loss: 0.33918032, Log Avg loss: 0.40229242, Global Avg Loss: 1.60031162, Time: 0.0208 Steps: 29920, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000124, Sample Num: 1984, Cur Loss: 0.16111130, Cur Avg Loss: 0.34049587, Log Avg loss: 0.35549319, Global Avg Loss: 1.59989571, Time: 0.0208 Steps: 29930, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000134, Sample Num: 2144, Cur Loss: 0.33680379, Cur Avg Loss: 0.34779309, Log Avg loss: 0.43827858, Global Avg Loss: 1.59950772, Time: 0.0209 Steps: 29940, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000144, Sample Num: 2304, Cur Loss: 0.09713358, Cur Avg Loss: 0.35802258, Log Avg loss: 0.49509782, Global Avg Loss: 1.59913897, Time: 0.0208 Steps: 29950, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000154, Sample Num: 2464, Cur Loss: 0.42288303, Cur Avg Loss: 0.36067329, Log Avg loss: 0.39884340, Global Avg Loss: 1.59873834, Time: 0.0208 Steps: 29960, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000164, Sample Num: 2624, Cur Loss: 0.44133055, Cur Avg Loss: 0.35903745, Log Avg loss: 0.33384548, Global Avg Loss: 1.59831629, Time: 0.0208 Steps: 29970, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000174, Sample Num: 2784, Cur Loss: 0.10827471, Cur Avg Loss: 0.35342760, Log Avg loss: 0.26142623, Global Avg Loss: 1.59787036, Time: 0.0208 Steps: 29980, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000184, Sample Num: 2944, Cur Loss: 0.27092460, Cur Avg Loss: 0.35280705, Log Avg loss: 0.34200938, Global Avg Loss: 1.59745160, Time: 0.0208 Steps: 29990, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000194, Sample Num: 3104, Cur Loss: 0.40273035, Cur Avg Loss: 0.35480779, Log Avg loss: 0.39162142, Global Avg Loss: 1.59704966, Time: 0.0208 Steps: 30000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000204, Sample Num: 3264, Cur Loss: 0.25304529, Cur Avg Loss: 0.34976262, Log Avg loss: 0.25188637, Global Avg Loss: 1.59660142, Time: 0.0208 Steps: 30010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000214, Sample Num: 3424, Cur Loss: 0.29124773, Cur Avg Loss: 0.34899833, Log Avg loss: 0.33340672, Global Avg Loss: 1.59618063, Time: 0.0208 Steps: 30020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000224, Sample Num: 3584, Cur Loss: 0.58463073, Cur Avg Loss: 0.34420902, Log Avg loss: 0.24171780, Global Avg Loss: 1.59572960, Time: 0.0209 Steps: 30030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000234, Sample Num: 3744, Cur Loss: 0.43207717, Cur Avg Loss: 0.34127356, Log Avg loss: 0.27551941, Global Avg Loss: 1.59529011, Time: 0.0207 Steps: 30040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000244, Sample Num: 3904, Cur Loss: 0.09729114, Cur Avg Loss: 0.33675566, Log Avg loss: 0.23103667, Global Avg Loss: 1.59483612, Time: 0.0208 Steps: 30050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000254, Sample Num: 4064, Cur Loss: 0.46141779, Cur Avg Loss: 0.33771688, Log Avg loss: 0.36117063, Global Avg Loss: 1.59442572, Time: 0.0208 Steps: 30060, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000264, Sample Num: 4224, Cur Loss: 0.37378573, Cur Avg Loss: 0.34299119, Log Avg loss: 0.47695882, Global Avg Loss: 1.59405410, Time: 0.0208 Steps: 30070, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000274, Sample Num: 4384, Cur Loss: 0.35388815, Cur Avg Loss: 0.34006486, Log Avg loss: 0.26280963, Global Avg Loss: 1.59361153, Time: 0.0209 Steps: 30080, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000284, Sample Num: 4544, Cur Loss: 0.92076886, Cur Avg Loss: 0.34160986, Log Avg loss: 0.38394285, Global Avg Loss: 1.59320951, Time: 0.0209 Steps: 30090, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000294, Sample Num: 4704, Cur Loss: 0.41041511, Cur Avg Loss: 0.34759453, Log Avg loss: 0.51755914, Global Avg Loss: 1.59285215, Time: 0.0209 Steps: 30100, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000304, Sample Num: 4864, Cur Loss: 0.38012764, Cur Avg Loss: 0.34828915, Log Avg loss: 0.36871108, Global Avg Loss: 1.59244560, Time: 0.0209 Steps: 30110, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000314, Sample Num: 5024, Cur Loss: 0.29997590, Cur Avg Loss: 0.34832062, Log Avg loss: 0.34927730, Global Avg Loss: 1.59203286, Time: 0.0209 Steps: 30120, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000324, Sample Num: 5184, Cur Loss: 0.42290094, Cur Avg Loss: 0.34853079, Log Avg loss: 0.35512992, Global Avg Loss: 1.59162234, Time: 0.0209 Steps: 30130, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000334, Sample Num: 5344, Cur Loss: 0.20500627, Cur Avg Loss: 0.34819797, Log Avg loss: 0.33741478, Global Avg Loss: 1.59120621, Time: 0.0209 Steps: 30140, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000344, Sample Num: 5504, Cur Loss: 0.09393679, Cur Avg Loss: 0.34804570, Log Avg loss: 0.34295976, Global Avg Loss: 1.59079220, Time: 0.0209 Steps: 30150, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000354, Sample Num: 5664, Cur Loss: 0.30111879, Cur Avg Loss: 0.34726335, Log Avg loss: 0.32035073, Global Avg Loss: 1.59037096, Time: 0.0209 Steps: 30160, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000364, Sample Num: 5824, Cur Loss: 0.87132561, Cur Avg Loss: 0.34835304, Log Avg loss: 0.38692807, Global Avg Loss: 1.58997208, Time: 0.0209 Steps: 30170, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000374, Sample Num: 5984, Cur Loss: 0.12682925, Cur Avg Loss: 0.34902793, Log Avg loss: 0.37359384, Global Avg Loss: 1.58956903, Time: 0.0209 Steps: 30180, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000384, Sample Num: 6144, Cur Loss: 0.71457386, Cur Avg Loss: 0.35278457, Log Avg loss: 0.49328295, Global Avg Loss: 1.58920591, Time: 0.0210 Steps: 30190, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000394, Sample Num: 6304, Cur Loss: 0.27026787, Cur Avg Loss: 0.35232701, Log Avg loss: 0.33475658, Global Avg Loss: 1.58879052, Time: 0.0209 Steps: 30200, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000404, Sample Num: 6464, Cur Loss: 0.49661922, Cur Avg Loss: 0.35544979, Log Avg loss: 0.47848718, Global Avg Loss: 1.58842300, Time: 0.0209 Steps: 30210, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000414, Sample Num: 6624, Cur Loss: 0.13328952, Cur Avg Loss: 0.35562856, Log Avg loss: 0.36285115, Global Avg Loss: 1.58801745, Time: 0.0209 Steps: 30220, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000424, Sample Num: 6784, Cur Loss: 0.29362854, Cur Avg Loss: 0.35382768, Log Avg loss: 0.27927106, Global Avg Loss: 1.58758452, Time: 0.0209 Steps: 30230, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000434, Sample Num: 6944, Cur Loss: 0.11848155, Cur Avg Loss: 0.35483070, Log Avg loss: 0.39735887, Global Avg Loss: 1.58719092, Time: 0.0209 Steps: 30240, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000444, Sample Num: 7104, Cur Loss: 0.65682179, Cur Avg Loss: 0.35575100, Log Avg loss: 0.39569210, Global Avg Loss: 1.58679704, Time: 0.0209 Steps: 30250, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000454, Sample Num: 7264, Cur Loss: 0.51221228, Cur Avg Loss: 0.35700954, Log Avg loss: 0.41288844, Global Avg Loss: 1.58640910, Time: 0.0209 Steps: 30260, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000464, Sample Num: 7424, Cur Loss: 0.17873926, Cur Avg Loss: 0.35579100, Log Avg loss: 0.30046964, Global Avg Loss: 1.58598428, Time: 0.0209 Steps: 30270, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000474, Sample Num: 7584, Cur Loss: 0.51850259, Cur Avg Loss: 0.35412650, Log Avg loss: 0.27689327, Global Avg Loss: 1.58555195, Time: 0.0209 Steps: 30280, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000484, Sample Num: 7744, Cur Loss: 0.12297178, Cur Avg Loss: 0.35142608, Log Avg loss: 0.22342661, Global Avg Loss: 1.58510225, Time: 0.0209 Steps: 30290, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000494, Sample Num: 7904, Cur Loss: 0.53334701, Cur Avg Loss: 0.35534990, Log Avg loss: 0.54526248, Global Avg Loss: 1.58475907, Time: 0.0209 Steps: 30300, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000504, Sample Num: 8064, Cur Loss: 0.45909747, Cur Avg Loss: 0.35790656, Log Avg loss: 0.48420552, Global Avg Loss: 1.58439597, Time: 0.0210 Steps: 30310, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000514, Sample Num: 8224, Cur Loss: 0.24831712, Cur Avg Loss: 0.35905455, Log Avg loss: 0.41691341, Global Avg Loss: 1.58401092, Time: 0.0246 Steps: 30320, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000524, Sample Num: 8384, Cur Loss: 0.34452873, Cur Avg Loss: 0.35944544, Log Avg loss: 0.37953720, Global Avg Loss: 1.58361380, Time: 0.0209 Steps: 30330, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000534, Sample Num: 8544, Cur Loss: 0.36252886, Cur Avg Loss: 0.35897721, Log Avg loss: 0.33444202, Global Avg Loss: 1.58320207, Time: 0.0208 Steps: 30340, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000544, Sample Num: 8704, Cur Loss: 0.44587028, Cur Avg Loss: 0.35690600, Log Avg loss: 0.24630351, Global Avg Loss: 1.58276158, Time: 0.0208 Steps: 30350, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000554, Sample Num: 8864, Cur Loss: 0.20068155, Cur Avg Loss: 0.35598473, Log Avg loss: 0.30586720, Global Avg Loss: 1.58234099, Time: 0.0208 Steps: 30360, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000564, Sample Num: 9024, Cur Loss: 0.59851563, Cur Avg Loss: 0.35507781, Log Avg loss: 0.30483482, Global Avg Loss: 1.58192035, Time: 0.0209 Steps: 30370, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000574, Sample Num: 9184, Cur Loss: 0.36793989, Cur Avg Loss: 0.36030440, Log Avg loss: 0.65508400, Global Avg Loss: 1.58161526, Time: 0.0209 Steps: 30380, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000584, Sample Num: 9344, Cur Loss: 0.35486382, Cur Avg Loss: 0.36022243, Log Avg loss: 0.35551700, Global Avg Loss: 1.58121181, Time: 0.0209 Steps: 30390, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000594, Sample Num: 9504, Cur Loss: 0.06340045, Cur Avg Loss: 0.35834716, Log Avg loss: 0.24883170, Global Avg Loss: 1.58077353, Time: 0.0208 Steps: 30400, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000604, Sample Num: 9664, Cur Loss: 0.19175722, Cur Avg Loss: 0.35752442, Log Avg loss: 0.30865360, Global Avg Loss: 1.58035520, Time: 0.0208 Steps: 30410, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000614, Sample Num: 9824, Cur Loss: 0.16172723, Cur Avg Loss: 0.35729743, Log Avg loss: 0.34358729, Global Avg Loss: 1.57994864, Time: 0.0208 Steps: 30420, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000624, Sample Num: 9984, Cur Loss: 0.19777796, Cur Avg Loss: 0.35570398, Log Avg loss: 0.25786597, Global Avg Loss: 1.57951417, Time: 0.0211 Steps: 30430, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000634, Sample Num: 10144, Cur Loss: 0.30969888, Cur Avg Loss: 0.35578387, Log Avg loss: 0.36076886, Global Avg Loss: 1.57911380, Time: 0.0210 Steps: 30440, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000644, Sample Num: 10304, Cur Loss: 0.23696353, Cur Avg Loss: 0.35447476, Log Avg loss: 0.27147746, Global Avg Loss: 1.57868436, Time: 0.0210 Steps: 30450, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000654, Sample Num: 10464, Cur Loss: 0.27229491, Cur Avg Loss: 0.35388939, Log Avg loss: 0.31619161, Global Avg Loss: 1.57826988, Time: 0.0210 Steps: 30460, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000664, Sample Num: 10624, Cur Loss: 0.22110584, Cur Avg Loss: 0.35423185, Log Avg loss: 0.37662884, Global Avg Loss: 1.57787552, Time: 0.0210 Steps: 30470, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000674, Sample Num: 10784, Cur Loss: 0.27781922, Cur Avg Loss: 0.35418090, Log Avg loss: 0.35079782, Global Avg Loss: 1.57747293, Time: 0.0210 Steps: 30480, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000684, Sample Num: 10944, Cur Loss: 0.27391246, Cur Avg Loss: 0.35421169, Log Avg loss: 0.35628684, Global Avg Loss: 1.57707241, Time: 0.0210 Steps: 30490, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000694, Sample Num: 11104, Cur Loss: 0.26615423, Cur Avg Loss: 0.35429569, Log Avg loss: 0.36004122, Global Avg Loss: 1.57667338, Time: 0.0209 Steps: 30500, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000704, Sample Num: 11264, Cur Loss: 0.23867995, Cur Avg Loss: 0.35404794, Log Avg loss: 0.33685424, Global Avg Loss: 1.57626702, Time: 0.0210 Steps: 30510, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000714, Sample Num: 11424, Cur Loss: 0.59205306, Cur Avg Loss: 0.35395698, Log Avg loss: 0.34755355, Global Avg Loss: 1.57586443, Time: 0.0210 Steps: 30520, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000724, Sample Num: 11584, Cur Loss: 0.26695341, Cur Avg Loss: 0.35369837, Log Avg loss: 0.33523306, Global Avg Loss: 1.57545806, Time: 0.0210 Steps: 30530, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000734, Sample Num: 11744, Cur Loss: 0.40914944, Cur Avg Loss: 0.35459152, Log Avg loss: 0.41925613, Global Avg Loss: 1.57507948, Time: 0.0210 Steps: 30540, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000744, Sample Num: 11904, Cur Loss: 0.27649170, Cur Avg Loss: 0.35516726, Log Avg loss: 0.39742635, Global Avg Loss: 1.57469399, Time: 0.0210 Steps: 30550, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000754, Sample Num: 12064, Cur Loss: 0.10541476, Cur Avg Loss: 0.35345318, Log Avg loss: 0.22592577, Global Avg Loss: 1.57425264, Time: 0.0210 Steps: 30560, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000764, Sample Num: 12224, Cur Loss: 0.22161418, Cur Avg Loss: 0.35360029, Log Avg loss: 0.36469201, Global Avg Loss: 1.57385697, Time: 0.0210 Steps: 30570, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000774, Sample Num: 12384, Cur Loss: 0.37535524, Cur Avg Loss: 0.35318750, Log Avg loss: 0.32165054, Global Avg Loss: 1.57344749, Time: 0.0210 Steps: 30580, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000784, Sample Num: 12544, Cur Loss: 0.49394858, Cur Avg Loss: 0.35728296, Log Avg loss: 0.67427173, Global Avg Loss: 1.57315354, Time: 0.0210 Steps: 30590, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000794, Sample Num: 12704, Cur Loss: 0.51107520, Cur Avg Loss: 0.35882271, Log Avg loss: 0.47953893, Global Avg Loss: 1.57279615, Time: 0.0211 Steps: 30600, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000804, Sample Num: 12864, Cur Loss: 0.28556412, Cur Avg Loss: 0.35832354, Log Avg loss: 0.31868912, Global Avg Loss: 1.57238645, Time: 0.0213 Steps: 30610, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000814, Sample Num: 13024, Cur Loss: 0.22450729, Cur Avg Loss: 0.35797532, Log Avg loss: 0.32997873, Global Avg Loss: 1.57198070, Time: 0.0210 Steps: 30620, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000824, Sample Num: 13184, Cur Loss: 0.08303855, Cur Avg Loss: 0.35754193, Log Avg loss: 0.32226427, Global Avg Loss: 1.57157269, Time: 0.0209 Steps: 30630, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000834, Sample Num: 13344, Cur Loss: 0.30938163, Cur Avg Loss: 0.35688158, Log Avg loss: 0.30246862, Global Avg Loss: 1.57115849, Time: 0.0209 Steps: 30640, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000844, Sample Num: 13504, Cur Loss: 0.18343517, Cur Avg Loss: 0.35718628, Log Avg loss: 0.38259787, Global Avg Loss: 1.57077071, Time: 0.0209 Steps: 30650, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000854, Sample Num: 13664, Cur Loss: 0.50066262, Cur Avg Loss: 0.35751281, Log Avg loss: 0.38507255, Global Avg Loss: 1.57038398, Time: 0.0210 Steps: 30660, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000864, Sample Num: 13824, Cur Loss: 0.27254066, Cur Avg Loss: 0.35767913, Log Avg loss: 0.37188260, Global Avg Loss: 1.56999321, Time: 0.0210 Steps: 30670, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000874, Sample Num: 13984, Cur Loss: 0.32504067, Cur Avg Loss: 0.35793301, Log Avg loss: 0.37986824, Global Avg Loss: 1.56960530, Time: 0.0210 Steps: 30680, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000884, Sample Num: 14144, Cur Loss: 0.30080605, Cur Avg Loss: 0.35739216, Log Avg loss: 0.31012211, Global Avg Loss: 1.56919491, Time: 0.0210 Steps: 30690, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000894, Sample Num: 14304, Cur Loss: 0.18957588, Cur Avg Loss: 0.35737477, Log Avg loss: 0.35583728, Global Avg Loss: 1.56879968, Time: 0.0209 Steps: 30700, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000904, Sample Num: 14464, Cur Loss: 0.24140482, Cur Avg Loss: 0.35684220, Log Avg loss: 0.30923064, Global Avg Loss: 1.56838953, Time: 0.0210 Steps: 30710, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000914, Sample Num: 14624, Cur Loss: 0.55786443, Cur Avg Loss: 0.35744919, Log Avg loss: 0.41232078, Global Avg Loss: 1.56801320, Time: 0.0210 Steps: 30720, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000924, Sample Num: 14784, Cur Loss: 0.10183267, Cur Avg Loss: 0.35643369, Log Avg loss: 0.26361739, Global Avg Loss: 1.56758873, Time: 0.0210 Steps: 30730, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000934, Sample Num: 14944, Cur Loss: 0.40639901, Cur Avg Loss: 0.35646052, Log Avg loss: 0.35893937, Global Avg Loss: 1.56719555, Time: 0.0210 Steps: 30740, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000944, Sample Num: 15104, Cur Loss: 0.43210548, Cur Avg Loss: 0.35600216, Log Avg loss: 0.31319120, Global Avg Loss: 1.56678774, Time: 0.0211 Steps: 30750, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000954, Sample Num: 15264, Cur Loss: 0.13735604, Cur Avg Loss: 0.35477751, Log Avg loss: 0.23917078, Global Avg Loss: 1.56635614, Time: 0.0209 Steps: 30760, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000964, Sample Num: 15424, Cur Loss: 0.53366435, Cur Avg Loss: 0.35445052, Log Avg loss: 0.32325512, Global Avg Loss: 1.56595214, Time: 0.0210 Steps: 30770, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000974, Sample Num: 15584, Cur Loss: 0.22469118, Cur Avg Loss: 0.35405611, Log Avg loss: 0.31603516, Global Avg Loss: 1.56554606, Time: 0.0211 Steps: 30780, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000984, Sample Num: 15744, Cur Loss: 0.34581131, Cur Avg Loss: 0.35381982, Log Avg loss: 0.33080563, Global Avg Loss: 1.56514504, Time: 0.0210 Steps: 30790, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000994, Sample Num: 15904, Cur Loss: 0.42666835, Cur Avg Loss: 0.35428432, Log Avg loss: 0.39999096, Global Avg Loss: 1.56476674, Time: 0.0210 Steps: 30800, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001004, Sample Num: 16064, Cur Loss: 0.34252620, Cur Avg Loss: 0.35443196, Log Avg loss: 0.36910731, Global Avg Loss: 1.56437867, Time: 0.0210 Steps: 30810, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001014, Sample Num: 16224, Cur Loss: 0.32993868, Cur Avg Loss: 0.35423065, Log Avg loss: 0.33401942, Global Avg Loss: 1.56397946, Time: 0.0210 Steps: 30820, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001024, Sample Num: 16384, Cur Loss: 0.46225828, Cur Avg Loss: 0.35463779, Log Avg loss: 0.39592155, Global Avg Loss: 1.56360059, Time: 0.0255 Steps: 30830, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001034, Sample Num: 16544, Cur Loss: 0.31155026, Cur Avg Loss: 0.35454528, Log Avg loss: 0.34507157, Global Avg Loss: 1.56320547, Time: 0.0209 Steps: 30840, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001044, Sample Num: 16704, Cur Loss: 0.56231761, Cur Avg Loss: 0.35568021, Log Avg loss: 0.47303259, Global Avg Loss: 1.56285210, Time: 0.0209 Steps: 30850, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001054, Sample Num: 16864, Cur Loss: 0.29786497, Cur Avg Loss: 0.35485404, Log Avg loss: 0.26860190, Global Avg Loss: 1.56243270, Time: 0.0208 Steps: 30860, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001064, Sample Num: 17024, Cur Loss: 0.33559021, Cur Avg Loss: 0.35497540, Log Avg loss: 0.36776714, Global Avg Loss: 1.56204570, Time: 0.0208 Steps: 30870, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001074, Sample Num: 17184, Cur Loss: 0.43763423, Cur Avg Loss: 0.35476610, Log Avg loss: 0.33249597, Global Avg Loss: 1.56164753, Time: 0.0209 Steps: 30880, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001084, Sample Num: 17344, Cur Loss: 0.43753809, Cur Avg Loss: 0.35518858, Log Avg loss: 0.40056329, Global Avg Loss: 1.56127166, Time: 0.0208 Steps: 30890, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001094, Sample Num: 17504, Cur Loss: 0.44457230, Cur Avg Loss: 0.35478124, Log Avg loss: 0.31062531, Global Avg Loss: 1.56086692, Time: 0.0208 Steps: 30900, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001104, Sample Num: 17664, Cur Loss: 0.22711611, Cur Avg Loss: 0.35582399, Log Avg loss: 0.46990121, Global Avg Loss: 1.56051397, Time: 0.0208 Steps: 30910, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001114, Sample Num: 17824, Cur Loss: 0.20542079, Cur Avg Loss: 0.35551562, Log Avg loss: 0.32147130, Global Avg Loss: 1.56011324, Time: 0.0209 Steps: 30920, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001124, Sample Num: 17984, Cur Loss: 0.37353468, Cur Avg Loss: 0.35486356, Log Avg loss: 0.28222419, Global Avg Loss: 1.55970009, Time: 0.0208 Steps: 30930, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001134, Sample Num: 18144, Cur Loss: 0.48311383, Cur Avg Loss: 0.35336792, Log Avg loss: 0.18525736, Global Avg Loss: 1.55925586, Time: 0.0208 Steps: 30940, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001144, Sample Num: 18304, Cur Loss: 0.34459865, Cur Avg Loss: 0.35354301, Log Avg loss: 0.37339826, Global Avg Loss: 1.55887270, Time: 0.0208 Steps: 30950, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001154, Sample Num: 18464, Cur Loss: 0.33754265, Cur Avg Loss: 0.35361688, Log Avg loss: 0.36206742, Global Avg Loss: 1.55848614, Time: 0.0210 Steps: 30960, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001164, Sample Num: 18624, Cur Loss: 0.49973696, Cur Avg Loss: 0.35372990, Log Avg loss: 0.36677248, Global Avg Loss: 1.55810134, Time: 0.0208 Steps: 30970, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001174, Sample Num: 18784, Cur Loss: 0.56872123, Cur Avg Loss: 0.35331049, Log Avg loss: 0.30449125, Global Avg Loss: 1.55769669, Time: 0.0208 Steps: 30980, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001184, Sample Num: 18944, Cur Loss: 0.47674677, Cur Avg Loss: 0.35324070, Log Avg loss: 0.34504775, Global Avg Loss: 1.55730539, Time: 0.0208 Steps: 30990, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001194, Sample Num: 19104, Cur Loss: 0.21132575, Cur Avg Loss: 0.35360790, Log Avg loss: 0.39708403, Global Avg Loss: 1.55693112, Time: 0.0208 Steps: 31000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001204, Sample Num: 19264, Cur Loss: 0.96189129, Cur Avg Loss: 0.35452284, Log Avg loss: 0.46376715, Global Avg Loss: 1.55657860, Time: 0.0208 Steps: 31010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001214, Sample Num: 19424, Cur Loss: 0.20987734, Cur Avg Loss: 0.35649376, Log Avg loss: 0.59379300, Global Avg Loss: 1.55626823, Time: 0.0208 Steps: 31020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001224, Sample Num: 19584, Cur Loss: 0.41618297, Cur Avg Loss: 0.35692493, Log Avg loss: 0.40926833, Global Avg Loss: 1.55589859, Time: 0.0208 Steps: 31030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001234, Sample Num: 19744, Cur Loss: 0.19405274, Cur Avg Loss: 0.35678601, Log Avg loss: 0.33978252, Global Avg Loss: 1.55550680, Time: 0.0208 Steps: 31040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001244, Sample Num: 19904, Cur Loss: 0.66732013, Cur Avg Loss: 0.35762870, Log Avg loss: 0.46161639, Global Avg Loss: 1.55515450, Time: 0.0208 Steps: 31050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001254, Sample Num: 20064, Cur Loss: 0.29655185, Cur Avg Loss: 0.35756436, Log Avg loss: 0.34956035, Global Avg Loss: 1.55476635, Time: 0.0208 Steps: 31060, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001264, Sample Num: 20224, Cur Loss: 0.14089224, Cur Avg Loss: 0.35771852, Log Avg loss: 0.37705064, Global Avg Loss: 1.55438729, Time: 0.0209 Steps: 31070, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001274, Sample Num: 20384, Cur Loss: 0.37054569, Cur Avg Loss: 0.35775817, Log Avg loss: 0.36277012, Global Avg Loss: 1.55400389, Time: 0.0208 Steps: 31080, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001284, Sample Num: 20544, Cur Loss: 1.05866349, Cur Avg Loss: 0.35929029, Log Avg loss: 0.55448157, Global Avg Loss: 1.55368240, Time: 0.0213 Steps: 31090, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001294, Sample Num: 20704, Cur Loss: 0.35308105, Cur Avg Loss: 0.35986068, Log Avg loss: 0.43309878, Global Avg Loss: 1.55332208, Time: 0.0209 Steps: 31100, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001304, Sample Num: 20864, Cur Loss: 0.07511649, Cur Avg Loss: 0.35939792, Log Avg loss: 0.29951697, Global Avg Loss: 1.55291906, Time: 0.0208 Steps: 31110, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001314, Sample Num: 21024, Cur Loss: 0.70442712, Cur Avg Loss: 0.35912872, Log Avg loss: 0.32402495, Global Avg Loss: 1.55252417, Time: 0.0208 Steps: 31120, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001324, Sample Num: 21184, Cur Loss: 0.16130285, Cur Avg Loss: 0.35855128, Log Avg loss: 0.28267614, Global Avg Loss: 1.55211625, Time: 0.0208 Steps: 31130, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001334, Sample Num: 21344, Cur Loss: 0.24609873, Cur Avg Loss: 0.35885904, Log Avg loss: 0.39960580, Global Avg Loss: 1.55174615, Time: 0.0208 Steps: 31140, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001344, Sample Num: 21504, Cur Loss: 0.21659091, Cur Avg Loss: 0.35894960, Log Avg loss: 0.37103085, Global Avg Loss: 1.55136710, Time: 0.0209 Steps: 31150, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001354, Sample Num: 21664, Cur Loss: 0.20773989, Cur Avg Loss: 0.35790347, Log Avg loss: 0.21730283, Global Avg Loss: 1.55093897, Time: 0.0209 Steps: 31160, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001364, Sample Num: 21824, Cur Loss: 0.71340847, Cur Avg Loss: 0.35932814, Log Avg loss: 0.55222864, Global Avg Loss: 1.55061856, Time: 0.0208 Steps: 31170, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001374, Sample Num: 21984, Cur Loss: 0.44647986, Cur Avg Loss: 0.35931059, Log Avg loss: 0.35691667, Global Avg Loss: 1.55023572, Time: 0.0208 Steps: 31180, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001384, Sample Num: 22144, Cur Loss: 0.31982908, Cur Avg Loss: 0.35997884, Log Avg loss: 0.45179619, Global Avg Loss: 1.54988354, Time: 0.0208 Steps: 31190, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001394, Sample Num: 22304, Cur Loss: 0.71995574, Cur Avg Loss: 0.35994814, Log Avg loss: 0.35570000, Global Avg Loss: 1.54950079, Time: 0.0208 Steps: 31200, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001404, Sample Num: 22464, Cur Loss: 0.20804903, Cur Avg Loss: 0.35965622, Log Avg loss: 0.31896312, Global Avg Loss: 1.54910652, Time: 0.0208 Steps: 31210, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001414, Sample Num: 22624, Cur Loss: 0.22469945, Cur Avg Loss: 0.35889058, Log Avg loss: 0.25139364, Global Avg Loss: 1.54869085, Time: 0.0208 Steps: 31220, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001424, Sample Num: 22784, Cur Loss: 0.19628599, Cur Avg Loss: 0.35908380, Log Avg loss: 0.38640584, Global Avg Loss: 1.54831868, Time: 0.0207 Steps: 31230, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001434, Sample Num: 22944, Cur Loss: 0.24501011, Cur Avg Loss: 0.35874349, Log Avg loss: 0.31028261, Global Avg Loss: 1.54792238, Time: 0.0209 Steps: 31240, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001444, Sample Num: 23104, Cur Loss: 0.46129835, Cur Avg Loss: 0.35863276, Log Avg loss: 0.34275405, Global Avg Loss: 1.54753673, Time: 0.0208 Steps: 31250, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001454, Sample Num: 23264, Cur Loss: 0.62816054, Cur Avg Loss: 0.35834536, Log Avg loss: 0.31684528, Global Avg Loss: 1.54714303, Time: 0.0208 Steps: 31260, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001464, Sample Num: 23424, Cur Loss: 0.34324378, Cur Avg Loss: 0.35800828, Log Avg loss: 0.30899730, Global Avg Loss: 1.54674708, Time: 0.0208 Steps: 31270, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001474, Sample Num: 23584, Cur Loss: 0.41189349, Cur Avg Loss: 0.35732669, Log Avg loss: 0.25754124, Global Avg Loss: 1.54633493, Time: 0.0208 Steps: 31280, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001484, Sample Num: 23744, Cur Loss: 0.24934195, Cur Avg Loss: 0.35746529, Log Avg loss: 0.37789482, Global Avg Loss: 1.54596151, Time: 0.0208 Steps: 31290, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001494, Sample Num: 23904, Cur Loss: 0.38387856, Cur Avg Loss: 0.35738855, Log Avg loss: 0.34600080, Global Avg Loss: 1.54557813, Time: 0.0208 Steps: 31300, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001504, Sample Num: 24064, Cur Loss: 0.21709263, Cur Avg Loss: 0.35771643, Log Avg loss: 0.40670197, Global Avg Loss: 1.54521439, Time: 0.0208 Steps: 31310, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001514, Sample Num: 24224, Cur Loss: 0.58226776, Cur Avg Loss: 0.35754189, Log Avg loss: 0.33129088, Global Avg Loss: 1.54482680, Time: 0.0209 Steps: 31320, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001524, Sample Num: 24384, Cur Loss: 0.17366242, Cur Avg Loss: 0.35636841, Log Avg loss: 0.17870363, Global Avg Loss: 1.54439076, Time: 0.0209 Steps: 31330, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001534, Sample Num: 24544, Cur Loss: 0.18988518, Cur Avg Loss: 0.35653167, Log Avg loss: 0.38141226, Global Avg Loss: 1.54401967, Time: 0.0208 Steps: 31340, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001544, Sample Num: 24704, Cur Loss: 0.36606687, Cur Avg Loss: 0.35639929, Log Avg loss: 0.33609206, Global Avg Loss: 1.54363437, Time: 0.0210 Steps: 31350, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001554, Sample Num: 24864, Cur Loss: 0.11088631, Cur Avg Loss: 0.35622753, Log Avg loss: 0.32970835, Global Avg Loss: 1.54324728, Time: 0.0209 Steps: 31360, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001564, Sample Num: 25024, Cur Loss: 0.33271298, Cur Avg Loss: 0.35599044, Log Avg loss: 0.31914574, Global Avg Loss: 1.54285706, Time: 0.0208 Steps: 31370, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001574, Sample Num: 25184, Cur Loss: 0.20069128, Cur Avg Loss: 0.35640706, Log Avg loss: 0.42156707, Global Avg Loss: 1.54249974, Time: 0.0208 Steps: 31380, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001584, Sample Num: 25344, Cur Loss: 0.12220303, Cur Avg Loss: 0.35559943, Log Avg loss: 0.22847747, Global Avg Loss: 1.54208112, Time: 0.0208 Steps: 31390, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001594, Sample Num: 25504, Cur Loss: 0.32577556, Cur Avg Loss: 0.35537490, Log Avg loss: 0.31981083, Global Avg Loss: 1.54169187, Time: 0.0209 Steps: 31400, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001604, Sample Num: 25664, Cur Loss: 0.18422359, Cur Avg Loss: 0.35539327, Log Avg loss: 0.35832124, Global Avg Loss: 1.54131512, Time: 0.0209 Steps: 31410, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001614, Sample Num: 25824, Cur Loss: 0.20351642, Cur Avg Loss: 0.35478028, Log Avg loss: 0.25645677, Global Avg Loss: 1.54090619, Time: 0.0209 Steps: 31420, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001624, Sample Num: 25984, Cur Loss: 0.32256344, Cur Avg Loss: 0.35490962, Log Avg loss: 0.37578393, Global Avg Loss: 1.54053548, Time: 0.0208 Steps: 31430, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001634, Sample Num: 26144, Cur Loss: 0.18656449, Cur Avg Loss: 0.35507922, Log Avg loss: 0.38262249, Global Avg Loss: 1.54016719, Time: 0.0209 Steps: 31440, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001644, Sample Num: 26304, Cur Loss: 0.32709616, Cur Avg Loss: 0.35566495, Log Avg loss: 0.45137362, Global Avg Loss: 1.53982099, Time: 0.0208 Steps: 31450, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001654, Sample Num: 26464, Cur Loss: 0.37792957, Cur Avg Loss: 0.35648461, Log Avg loss: 0.49123710, Global Avg Loss: 1.53948768, Time: 0.0209 Steps: 31460, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001664, Sample Num: 26624, Cur Loss: 0.14105800, Cur Avg Loss: 0.35625970, Log Avg loss: 0.31905876, Global Avg Loss: 1.53909988, Time: 0.0209 Steps: 31470, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001674, Sample Num: 26784, Cur Loss: 0.82251942, Cur Avg Loss: 0.35647440, Log Avg loss: 0.39220063, Global Avg Loss: 1.53873555, Time: 0.0209 Steps: 31480, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001684, Sample Num: 26944, Cur Loss: 0.31259745, Cur Avg Loss: 0.35620623, Log Avg loss: 0.31131409, Global Avg Loss: 1.53834577, Time: 0.0208 Steps: 31490, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001694, Sample Num: 27104, Cur Loss: 0.55899799, Cur Avg Loss: 0.35609161, Log Avg loss: 0.33678997, Global Avg Loss: 1.53796432, Time: 0.0209 Steps: 31500, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001704, Sample Num: 27264, Cur Loss: 0.17085332, Cur Avg Loss: 0.35582784, Log Avg loss: 0.31114574, Global Avg Loss: 1.53757498, Time: 0.0209 Steps: 31510, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001714, Sample Num: 27424, Cur Loss: 0.61670232, Cur Avg Loss: 0.35562548, Log Avg loss: 0.32114243, Global Avg Loss: 1.53718906, Time: 0.0208 Steps: 31520, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001724, Sample Num: 27584, Cur Loss: 0.18772307, Cur Avg Loss: 0.35563249, Log Avg loss: 0.35683556, Global Avg Loss: 1.53681470, Time: 0.0209 Steps: 31530, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001734, Sample Num: 27744, Cur Loss: 0.25877494, Cur Avg Loss: 0.35517056, Log Avg loss: 0.27553269, Global Avg Loss: 1.53641480, Time: 0.0209 Steps: 31540, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001744, Sample Num: 27904, Cur Loss: 0.53935242, Cur Avg Loss: 0.35520276, Log Avg loss: 0.36078680, Global Avg Loss: 1.53604218, Time: 0.0209 Steps: 31550, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001754, Sample Num: 28064, Cur Loss: 0.24966261, Cur Avg Loss: 0.35481719, Log Avg loss: 0.28757295, Global Avg Loss: 1.53564659, Time: 0.0209 Steps: 31560, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001764, Sample Num: 28224, Cur Loss: 0.39254960, Cur Avg Loss: 0.35427148, Log Avg loss: 0.25855391, Global Avg Loss: 1.53524206, Time: 0.0209 Steps: 31570, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001774, Sample Num: 28384, Cur Loss: 0.34953257, Cur Avg Loss: 0.35410093, Log Avg loss: 0.32401570, Global Avg Loss: 1.53485852, Time: 0.0209 Steps: 31580, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001784, Sample Num: 28544, Cur Loss: 0.20797396, Cur Avg Loss: 0.35352556, Log Avg loss: 0.25145652, Global Avg Loss: 1.53445225, Time: 0.0208 Steps: 31590, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001794, Sample Num: 28704, Cur Loss: 0.66724354, Cur Avg Loss: 0.35348585, Log Avg loss: 0.34640107, Global Avg Loss: 1.53407629, Time: 0.0246 Steps: 31600, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001804, Sample Num: 28864, Cur Loss: 0.19471298, Cur Avg Loss: 0.35382102, Log Avg loss: 0.41394941, Global Avg Loss: 1.53372193, Time: 0.0208 Steps: 31610, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001814, Sample Num: 29024, Cur Loss: 0.38912678, Cur Avg Loss: 0.35364067, Log Avg loss: 0.32110701, Global Avg Loss: 1.53333843, Time: 0.0208 Steps: 31620, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001824, Sample Num: 29184, Cur Loss: 0.87472647, Cur Avg Loss: 0.35439739, Log Avg loss: 0.49166584, Global Avg Loss: 1.53300910, Time: 0.0209 Steps: 31630, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001834, Sample Num: 29344, Cur Loss: 0.25660995, Cur Avg Loss: 0.35419306, Log Avg loss: 0.31692378, Global Avg Loss: 1.53262475, Time: 0.0208 Steps: 31640, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001844, Sample Num: 29504, Cur Loss: 0.26698688, Cur Avg Loss: 0.35404133, Log Avg loss: 0.32621331, Global Avg Loss: 1.53224358, Time: 0.0209 Steps: 31650, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001854, Sample Num: 29664, Cur Loss: 0.55584586, Cur Avg Loss: 0.35395232, Log Avg loss: 0.33753838, Global Avg Loss: 1.53186622, Time: 0.0209 Steps: 31660, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001864, Sample Num: 29824, Cur Loss: 0.18773393, Cur Avg Loss: 0.35347199, Log Avg loss: 0.26442023, Global Avg Loss: 1.53146602, Time: 0.0209 Steps: 31670, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001874, Sample Num: 29984, Cur Loss: 0.44784182, Cur Avg Loss: 0.35364061, Log Avg loss: 0.38507137, Global Avg Loss: 1.53110415, Time: 0.0208 Steps: 31680, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001884, Sample Num: 30144, Cur Loss: 0.59506774, Cur Avg Loss: 0.35395106, Log Avg loss: 0.41212957, Global Avg Loss: 1.53075105, Time: 0.0208 Steps: 31690, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001894, Sample Num: 30304, Cur Loss: 0.35734355, Cur Avg Loss: 0.35366645, Log Avg loss: 0.30004526, Global Avg Loss: 1.53036282, Time: 0.0209 Steps: 31700, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001904, Sample Num: 30464, Cur Loss: 0.39297834, Cur Avg Loss: 0.35313177, Log Avg loss: 0.25186380, Global Avg Loss: 1.52995963, Time: 0.0210 Steps: 31710, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001914, Sample Num: 30624, Cur Loss: 0.15355074, Cur Avg Loss: 0.35265707, Log Avg loss: 0.26227367, Global Avg Loss: 1.52955998, Time: 0.0208 Steps: 31720, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001924, Sample Num: 30784, Cur Loss: 0.31567878, Cur Avg Loss: 0.35272538, Log Avg loss: 0.36579954, Global Avg Loss: 1.52919321, Time: 0.0208 Steps: 31730, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001934, Sample Num: 30944, Cur Loss: 0.28343904, Cur Avg Loss: 0.35271141, Log Avg loss: 0.35002290, Global Avg Loss: 1.52882170, Time: 0.0208 Steps: 31740, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001944, Sample Num: 31104, Cur Loss: 0.49013412, Cur Avg Loss: 0.35319892, Log Avg loss: 0.44748498, Global Avg Loss: 1.52848112, Time: 0.0208 Steps: 31750, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001954, Sample Num: 31264, Cur Loss: 0.38729483, Cur Avg Loss: 0.35302978, Log Avg loss: 0.32014871, Global Avg Loss: 1.52810067, Time: 0.0208 Steps: 31760, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001964, Sample Num: 31424, Cur Loss: 0.30937958, Cur Avg Loss: 0.35342543, Log Avg loss: 0.43073560, Global Avg Loss: 1.52775526, Time: 0.0208 Steps: 31770, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001974, Sample Num: 31584, Cur Loss: 0.04704121, Cur Avg Loss: 0.35347400, Log Avg loss: 0.36301336, Global Avg Loss: 1.52738876, Time: 0.0208 Steps: 31780, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001984, Sample Num: 31744, Cur Loss: 0.35628211, Cur Avg Loss: 0.35388392, Log Avg loss: 0.43480172, Global Avg Loss: 1.52704507, Time: 0.0208 Steps: 31790, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001994, Sample Num: 31904, Cur Loss: 0.09808185, Cur Avg Loss: 0.35364022, Log Avg loss: 0.30528995, Global Avg Loss: 1.52666087, Time: 0.0209 Steps: 31800, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002004, Sample Num: 32064, Cur Loss: 0.12965414, Cur Avg Loss: 0.35315409, Log Avg loss: 0.25622031, Global Avg Loss: 1.52626148, Time: 0.0209 Steps: 31810, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002014, Sample Num: 32224, Cur Loss: 0.07538077, Cur Avg Loss: 0.35291287, Log Avg loss: 0.30457236, Global Avg Loss: 1.52587755, Time: 0.0209 Steps: 31820, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002024, Sample Num: 32384, Cur Loss: 0.06099940, Cur Avg Loss: 0.35291383, Log Avg loss: 0.35310735, Global Avg Loss: 1.52550910, Time: 0.0208 Steps: 31830, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002034, Sample Num: 32544, Cur Loss: 0.26126158, Cur Avg Loss: 0.35355034, Log Avg loss: 0.48237885, Global Avg Loss: 1.52518148, Time: 0.0209 Steps: 31840, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002044, Sample Num: 32704, Cur Loss: 0.44483501, Cur Avg Loss: 0.35309268, Log Avg loss: 0.26000461, Global Avg Loss: 1.52478425, Time: 0.0209 Steps: 31850, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002054, Sample Num: 32864, Cur Loss: 0.25766760, Cur Avg Loss: 0.35335525, Log Avg loss: 0.40702408, Global Avg Loss: 1.52443342, Time: 0.0247 Steps: 31860, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002064, Sample Num: 33024, Cur Loss: 0.14343834, Cur Avg Loss: 0.35281621, Log Avg loss: 0.24209718, Global Avg Loss: 1.52403105, Time: 0.0210 Steps: 31870, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002074, Sample Num: 33184, Cur Loss: 0.19227661, Cur Avg Loss: 0.35223778, Log Avg loss: 0.23285049, Global Avg Loss: 1.52362604, Time: 0.0209 Steps: 31880, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002084, Sample Num: 33344, Cur Loss: 0.12882970, Cur Avg Loss: 0.35179825, Log Avg loss: 0.26064065, Global Avg Loss: 1.52323000, Time: 0.0211 Steps: 31890, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002094, Sample Num: 33504, Cur Loss: 0.20863211, Cur Avg Loss: 0.35153458, Log Avg loss: 0.29658591, Global Avg Loss: 1.52284547, Time: 0.0210 Steps: 31900, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002104, Sample Num: 33664, Cur Loss: 0.11967716, Cur Avg Loss: 0.35097891, Log Avg loss: 0.23461978, Global Avg Loss: 1.52244176, Time: 0.0210 Steps: 31910, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002114, Sample Num: 33824, Cur Loss: 0.33016062, Cur Avg Loss: 0.35108586, Log Avg loss: 0.37358826, Global Avg Loss: 1.52208185, Time: 0.0209 Steps: 31920, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002124, Sample Num: 33984, Cur Loss: 0.17005417, Cur Avg Loss: 0.35130981, Log Avg loss: 0.39865413, Global Avg Loss: 1.52173000, Time: 0.0209 Steps: 31930, Updated lr: 0.000071 ***** Running evaluation checkpoint-31935 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-31935 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.807739, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.413333, "eval_total_loss": 290.572771, "eval_mae": 0.435604, "eval_mse": 0.413497, "eval_r2": 0.737154, "eval_sp_statistic": 0.848663, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.882137, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.003239, "test_total_loss": 503.626087, "test_mae": 0.724096, "test_mse": 1.003471, "test_r2": 0.352351, "test_sp_statistic": 0.780676, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.831706, "test_ps_pvalue": 0.0, "lr": 7.066382171645329e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5215448621839531, "train_cur_epoch_loss": 747.8781670220196, "train_cur_epoch_avg_loss": 0.35128143119869404, "train_cur_epoch_time": 44.80773949623108, "train_cur_epoch_avg_time": 0.021046378344871337, "epoch": 15, "step": 31935} ################################################## Training, Epoch: 0016, Batch: 000005, Sample Num: 80, Cur Loss: 0.38825786, Cur Avg Loss: 0.42397486, Log Avg loss: 0.38160006, Global Avg Loss: 1.52137304, Time: 0.0246 Steps: 31940, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000015, Sample Num: 240, Cur Loss: 0.60791492, Cur Avg Loss: 0.42199535, Log Avg loss: 0.42100559, Global Avg Loss: 1.52102864, Time: 0.0208 Steps: 31950, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000025, Sample Num: 400, Cur Loss: 0.71709681, Cur Avg Loss: 0.36796887, Log Avg loss: 0.28692914, Global Avg Loss: 1.52064250, Time: 0.0209 Steps: 31960, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000035, Sample Num: 560, Cur Loss: 0.26068914, Cur Avg Loss: 0.36214442, Log Avg loss: 0.34758329, Global Avg Loss: 1.52027558, Time: 0.0208 Steps: 31970, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000045, Sample Num: 720, Cur Loss: 0.25160402, Cur Avg Loss: 0.34850555, Log Avg loss: 0.30076952, Global Avg Loss: 1.51989424, Time: 0.0208 Steps: 31980, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000055, Sample Num: 880, Cur Loss: 0.35896564, Cur Avg Loss: 0.34220183, Log Avg loss: 0.31383510, Global Avg Loss: 1.51951723, Time: 0.0208 Steps: 31990, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000065, Sample Num: 1040, Cur Loss: 0.21329176, Cur Avg Loss: 0.34706622, Log Avg loss: 0.37382035, Global Avg Loss: 1.51915920, Time: 0.0208 Steps: 32000, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000075, Sample Num: 1200, Cur Loss: 0.36351717, Cur Avg Loss: 0.34449547, Log Avg loss: 0.32778561, Global Avg Loss: 1.51878701, Time: 0.0208 Steps: 32010, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000085, Sample Num: 1360, Cur Loss: 0.23262195, Cur Avg Loss: 0.33473892, Log Avg loss: 0.26156479, Global Avg Loss: 1.51839438, Time: 0.0209 Steps: 32020, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000095, Sample Num: 1520, Cur Loss: 0.25781497, Cur Avg Loss: 0.32779134, Log Avg loss: 0.26873686, Global Avg Loss: 1.51800423, Time: 0.0208 Steps: 32030, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000105, Sample Num: 1680, Cur Loss: 0.11079390, Cur Avg Loss: 0.32511579, Log Avg loss: 0.29969816, Global Avg Loss: 1.51762398, Time: 0.0208 Steps: 32040, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000115, Sample Num: 1840, Cur Loss: 0.54952252, Cur Avg Loss: 0.32485730, Log Avg loss: 0.32214310, Global Avg Loss: 1.51725098, Time: 0.0208 Steps: 32050, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000125, Sample Num: 2000, Cur Loss: 0.31946999, Cur Avg Loss: 0.32600972, Log Avg loss: 0.33926261, Global Avg Loss: 1.51688354, Time: 0.0208 Steps: 32060, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000135, Sample Num: 2160, Cur Loss: 0.16254482, Cur Avg Loss: 0.33978713, Log Avg loss: 0.51200470, Global Avg Loss: 1.51657020, Time: 0.0208 Steps: 32070, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000145, Sample Num: 2320, Cur Loss: 0.86930567, Cur Avg Loss: 0.33949186, Log Avg loss: 0.33550566, Global Avg Loss: 1.51620204, Time: 0.0208 Steps: 32080, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000155, Sample Num: 2480, Cur Loss: 0.23371638, Cur Avg Loss: 0.33672134, Log Avg loss: 0.29654880, Global Avg Loss: 1.51582197, Time: 0.0208 Steps: 32090, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000165, Sample Num: 2640, Cur Loss: 0.39911637, Cur Avg Loss: 0.33273930, Log Avg loss: 0.27101772, Global Avg Loss: 1.51543418, Time: 0.0208 Steps: 32100, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000175, Sample Num: 2800, Cur Loss: 0.16769764, Cur Avg Loss: 0.33064747, Log Avg loss: 0.29613227, Global Avg Loss: 1.51505445, Time: 0.0208 Steps: 32110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000185, Sample Num: 2960, Cur Loss: 0.34823313, Cur Avg Loss: 0.33819223, Log Avg loss: 0.47022548, Global Avg Loss: 1.51472916, Time: 0.0208 Steps: 32120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000195, Sample Num: 3120, Cur Loss: 0.21937177, Cur Avg Loss: 0.33270362, Log Avg loss: 0.23116435, Global Avg Loss: 1.51432967, Time: 0.0208 Steps: 32130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000205, Sample Num: 3280, Cur Loss: 0.23686530, Cur Avg Loss: 0.32945336, Log Avg loss: 0.26607334, Global Avg Loss: 1.51394129, Time: 0.0208 Steps: 32140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000215, Sample Num: 3440, Cur Loss: 0.12128977, Cur Avg Loss: 0.32826469, Log Avg loss: 0.30389697, Global Avg Loss: 1.51356492, Time: 0.0208 Steps: 32150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000225, Sample Num: 3600, Cur Loss: 0.13182263, Cur Avg Loss: 0.33002543, Log Avg loss: 0.36788125, Global Avg Loss: 1.51320867, Time: 0.0208 Steps: 32160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000235, Sample Num: 3760, Cur Loss: 0.82356322, Cur Avg Loss: 0.33676940, Log Avg loss: 0.48850879, Global Avg Loss: 1.51289015, Time: 0.0208 Steps: 32170, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000245, Sample Num: 3920, Cur Loss: 0.12313166, Cur Avg Loss: 0.33699782, Log Avg loss: 0.34236563, Global Avg Loss: 1.51252640, Time: 0.0208 Steps: 32180, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000255, Sample Num: 4080, Cur Loss: 0.13137214, Cur Avg Loss: 0.33351250, Log Avg loss: 0.24812218, Global Avg Loss: 1.51213361, Time: 0.0208 Steps: 32190, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000265, Sample Num: 4240, Cur Loss: 0.39657253, Cur Avg Loss: 0.33190866, Log Avg loss: 0.29101071, Global Avg Loss: 1.51175438, Time: 0.0211 Steps: 32200, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000275, Sample Num: 4400, Cur Loss: 0.43956912, Cur Avg Loss: 0.33202908, Log Avg loss: 0.33522036, Global Avg Loss: 1.51138911, Time: 0.0210 Steps: 32210, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000285, Sample Num: 4560, Cur Loss: 0.40979367, Cur Avg Loss: 0.33033008, Log Avg loss: 0.28360763, Global Avg Loss: 1.51100805, Time: 0.0209 Steps: 32220, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000295, Sample Num: 4720, Cur Loss: 0.30340502, Cur Avg Loss: 0.33314768, Log Avg loss: 0.41344928, Global Avg Loss: 1.51066751, Time: 0.0210 Steps: 32230, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000305, Sample Num: 4880, Cur Loss: 0.23299666, Cur Avg Loss: 0.33153257, Log Avg loss: 0.28388675, Global Avg Loss: 1.51028699, Time: 0.0209 Steps: 32240, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000315, Sample Num: 5040, Cur Loss: 0.30954331, Cur Avg Loss: 0.33326969, Log Avg loss: 0.38625194, Global Avg Loss: 1.50993845, Time: 0.0210 Steps: 32250, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000325, Sample Num: 5200, Cur Loss: 0.48198402, Cur Avg Loss: 0.33728438, Log Avg loss: 0.46374703, Global Avg Loss: 1.50961415, Time: 0.0209 Steps: 32260, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000335, Sample Num: 5360, Cur Loss: 0.10649488, Cur Avg Loss: 0.33618448, Log Avg loss: 0.30043761, Global Avg Loss: 1.50923945, Time: 0.0210 Steps: 32270, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000345, Sample Num: 5520, Cur Loss: 0.44492239, Cur Avg Loss: 0.33822147, Log Avg loss: 0.40646060, Global Avg Loss: 1.50889782, Time: 0.0210 Steps: 32280, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000355, Sample Num: 5680, Cur Loss: 0.13185352, Cur Avg Loss: 0.33710198, Log Avg loss: 0.29847974, Global Avg Loss: 1.50852296, Time: 0.0210 Steps: 32290, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000365, Sample Num: 5840, Cur Loss: 0.23731190, Cur Avg Loss: 0.33717710, Log Avg loss: 0.33984386, Global Avg Loss: 1.50816114, Time: 0.0209 Steps: 32300, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000375, Sample Num: 6000, Cur Loss: 0.33014774, Cur Avg Loss: 0.33442344, Log Avg loss: 0.23391488, Global Avg Loss: 1.50776676, Time: 0.0210 Steps: 32310, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000385, Sample Num: 6160, Cur Loss: 0.40673503, Cur Avg Loss: 0.34164855, Log Avg loss: 0.61259001, Global Avg Loss: 1.50748979, Time: 0.0210 Steps: 32320, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000395, Sample Num: 6320, Cur Loss: 0.28257543, Cur Avg Loss: 0.34052806, Log Avg loss: 0.29738932, Global Avg Loss: 1.50711549, Time: 0.0209 Steps: 32330, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000405, Sample Num: 6480, Cur Loss: 0.38888207, Cur Avg Loss: 0.33999633, Log Avg loss: 0.31899307, Global Avg Loss: 1.50674810, Time: 0.0210 Steps: 32340, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000415, Sample Num: 6640, Cur Loss: 0.36051396, Cur Avg Loss: 0.33907359, Log Avg loss: 0.30170257, Global Avg Loss: 1.50637560, Time: 0.0209 Steps: 32350, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000425, Sample Num: 6800, Cur Loss: 0.76637590, Cur Avg Loss: 0.33995299, Log Avg loss: 0.37644804, Global Avg Loss: 1.50602643, Time: 0.0209 Steps: 32360, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000435, Sample Num: 6960, Cur Loss: 0.26609978, Cur Avg Loss: 0.34555012, Log Avg loss: 0.58342826, Global Avg Loss: 1.50574141, Time: 0.0210 Steps: 32370, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000445, Sample Num: 7120, Cur Loss: 0.21274596, Cur Avg Loss: 0.34491292, Log Avg loss: 0.31719460, Global Avg Loss: 1.50537435, Time: 0.0210 Steps: 32380, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000455, Sample Num: 7280, Cur Loss: 0.86384463, Cur Avg Loss: 0.34658739, Log Avg loss: 0.42110144, Global Avg Loss: 1.50503959, Time: 0.0210 Steps: 32390, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000465, Sample Num: 7440, Cur Loss: 0.15221763, Cur Avg Loss: 0.34781475, Log Avg loss: 0.40365933, Global Avg Loss: 1.50469966, Time: 0.0209 Steps: 32400, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000475, Sample Num: 7600, Cur Loss: 0.24054024, Cur Avg Loss: 0.34687351, Log Avg loss: 0.30310606, Global Avg Loss: 1.50432891, Time: 0.0210 Steps: 32410, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000485, Sample Num: 7760, Cur Loss: 0.33909258, Cur Avg Loss: 0.34741970, Log Avg loss: 0.37336359, Global Avg Loss: 1.50398007, Time: 0.0209 Steps: 32420, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000495, Sample Num: 7920, Cur Loss: 0.79185569, Cur Avg Loss: 0.34758362, Log Avg loss: 0.35553392, Global Avg Loss: 1.50362593, Time: 0.0210 Steps: 32430, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000505, Sample Num: 8080, Cur Loss: 0.40557885, Cur Avg Loss: 0.34809756, Log Avg loss: 0.37353771, Global Avg Loss: 1.50327757, Time: 0.0209 Steps: 32440, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000515, Sample Num: 8240, Cur Loss: 0.30954188, Cur Avg Loss: 0.34652570, Log Avg loss: 0.26714669, Global Avg Loss: 1.50289664, Time: 0.0246 Steps: 32450, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000525, Sample Num: 8400, Cur Loss: 0.18221444, Cur Avg Loss: 0.34473388, Log Avg loss: 0.25245477, Global Avg Loss: 1.50251141, Time: 0.0209 Steps: 32460, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000535, Sample Num: 8560, Cur Loss: 0.22524562, Cur Avg Loss: 0.34537125, Log Avg loss: 0.37883317, Global Avg Loss: 1.50216535, Time: 0.0209 Steps: 32470, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000545, Sample Num: 8720, Cur Loss: 0.21781281, Cur Avg Loss: 0.34315529, Log Avg loss: 0.22460193, Global Avg Loss: 1.50177201, Time: 0.0210 Steps: 32480, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000555, Sample Num: 8880, Cur Loss: 0.33520299, Cur Avg Loss: 0.34307597, Log Avg loss: 0.33875274, Global Avg Loss: 1.50141405, Time: 0.0209 Steps: 32490, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000565, Sample Num: 9040, Cur Loss: 0.19678387, Cur Avg Loss: 0.34260663, Log Avg loss: 0.31655829, Global Avg Loss: 1.50104947, Time: 0.0209 Steps: 32500, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000575, Sample Num: 9200, Cur Loss: 0.20664963, Cur Avg Loss: 0.34301267, Log Avg loss: 0.36595384, Global Avg Loss: 1.50070032, Time: 0.0209 Steps: 32510, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000585, Sample Num: 9360, Cur Loss: 0.17233253, Cur Avg Loss: 0.34227567, Log Avg loss: 0.29989842, Global Avg Loss: 1.50033107, Time: 0.0209 Steps: 32520, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000595, Sample Num: 9520, Cur Loss: 0.16819832, Cur Avg Loss: 0.34364490, Log Avg loss: 0.42374485, Global Avg Loss: 1.50000012, Time: 0.0209 Steps: 32530, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000605, Sample Num: 9680, Cur Loss: 0.21887363, Cur Avg Loss: 0.34586252, Log Avg loss: 0.47781102, Global Avg Loss: 1.49968599, Time: 0.0210 Steps: 32540, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000615, Sample Num: 9840, Cur Loss: 0.32028693, Cur Avg Loss: 0.34750186, Log Avg loss: 0.44668187, Global Avg Loss: 1.49936248, Time: 0.0209 Steps: 32550, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000625, Sample Num: 10000, Cur Loss: 0.30481640, Cur Avg Loss: 0.34697606, Log Avg loss: 0.31463901, Global Avg Loss: 1.49899862, Time: 0.0210 Steps: 32560, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000635, Sample Num: 10160, Cur Loss: 0.43934053, Cur Avg Loss: 0.34728307, Log Avg loss: 0.36647165, Global Avg Loss: 1.49865090, Time: 0.0209 Steps: 32570, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000645, Sample Num: 10320, Cur Loss: 0.16756412, Cur Avg Loss: 0.34735661, Log Avg loss: 0.35202597, Global Avg Loss: 1.49829896, Time: 0.0210 Steps: 32580, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000655, Sample Num: 10480, Cur Loss: 0.09689753, Cur Avg Loss: 0.34500176, Log Avg loss: 0.19311390, Global Avg Loss: 1.49789848, Time: 0.0209 Steps: 32590, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000665, Sample Num: 10640, Cur Loss: 0.21658868, Cur Avg Loss: 0.34320576, Log Avg loss: 0.22556784, Global Avg Loss: 1.49750819, Time: 0.0209 Steps: 32600, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000675, Sample Num: 10800, Cur Loss: 0.11821170, Cur Avg Loss: 0.34305455, Log Avg loss: 0.33299903, Global Avg Loss: 1.49715109, Time: 0.0209 Steps: 32610, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000685, Sample Num: 10960, Cur Loss: 0.40470955, Cur Avg Loss: 0.34460492, Log Avg loss: 0.44925493, Global Avg Loss: 1.49682984, Time: 0.0209 Steps: 32620, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000695, Sample Num: 11120, Cur Loss: 0.40844595, Cur Avg Loss: 0.34404066, Log Avg loss: 0.30538877, Global Avg Loss: 1.49646471, Time: 0.0210 Steps: 32630, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000705, Sample Num: 11280, Cur Loss: 0.11768861, Cur Avg Loss: 0.34275611, Log Avg loss: 0.25348006, Global Avg Loss: 1.49608389, Time: 0.0210 Steps: 32640, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000715, Sample Num: 11440, Cur Loss: 0.25384399, Cur Avg Loss: 0.34190037, Log Avg loss: 0.28157059, Global Avg Loss: 1.49571191, Time: 0.0210 Steps: 32650, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000725, Sample Num: 11600, Cur Loss: 0.25427631, Cur Avg Loss: 0.34195159, Log Avg loss: 0.34561385, Global Avg Loss: 1.49535977, Time: 0.0209 Steps: 32660, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000735, Sample Num: 11760, Cur Loss: 0.25663453, Cur Avg Loss: 0.34261488, Log Avg loss: 0.39070362, Global Avg Loss: 1.49502164, Time: 0.0210 Steps: 32670, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000745, Sample Num: 11920, Cur Loss: 0.17285883, Cur Avg Loss: 0.34338993, Log Avg loss: 0.40035578, Global Avg Loss: 1.49468668, Time: 0.0209 Steps: 32680, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000755, Sample Num: 12080, Cur Loss: 0.48022404, Cur Avg Loss: 0.34355606, Log Avg loss: 0.35593313, Global Avg Loss: 1.49433833, Time: 0.0209 Steps: 32690, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000765, Sample Num: 12240, Cur Loss: 0.25152797, Cur Avg Loss: 0.34296715, Log Avg loss: 0.29850426, Global Avg Loss: 1.49397263, Time: 0.0209 Steps: 32700, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000775, Sample Num: 12400, Cur Loss: 0.44188017, Cur Avg Loss: 0.34245540, Log Avg loss: 0.30330635, Global Avg Loss: 1.49360862, Time: 0.0210 Steps: 32710, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000785, Sample Num: 12560, Cur Loss: 0.38164052, Cur Avg Loss: 0.34231673, Log Avg loss: 0.33157018, Global Avg Loss: 1.49325348, Time: 0.0210 Steps: 32720, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000795, Sample Num: 12720, Cur Loss: 0.13406023, Cur Avg Loss: 0.34229634, Log Avg loss: 0.34069546, Global Avg Loss: 1.49290134, Time: 0.0210 Steps: 32730, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000805, Sample Num: 12880, Cur Loss: 0.07288039, Cur Avg Loss: 0.34129114, Log Avg loss: 0.26137810, Global Avg Loss: 1.49252518, Time: 0.0211 Steps: 32740, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000815, Sample Num: 13040, Cur Loss: 0.15570062, Cur Avg Loss: 0.34058041, Log Avg loss: 0.28336611, Global Avg Loss: 1.49215598, Time: 0.0210 Steps: 32750, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000825, Sample Num: 13200, Cur Loss: 0.49295211, Cur Avg Loss: 0.34112072, Log Avg loss: 0.38515619, Global Avg Loss: 1.49181806, Time: 0.0209 Steps: 32760, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000835, Sample Num: 13360, Cur Loss: 0.25022319, Cur Avg Loss: 0.34126810, Log Avg loss: 0.35342739, Global Avg Loss: 1.49147068, Time: 0.0209 Steps: 32770, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000845, Sample Num: 13520, Cur Loss: 0.29335883, Cur Avg Loss: 0.34011211, Log Avg loss: 0.24358657, Global Avg Loss: 1.49108999, Time: 0.0213 Steps: 32780, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000855, Sample Num: 13680, Cur Loss: 0.20484656, Cur Avg Loss: 0.34000121, Log Avg loss: 0.33063008, Global Avg Loss: 1.49073608, Time: 0.0209 Steps: 32790, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000865, Sample Num: 13840, Cur Loss: 0.24640043, Cur Avg Loss: 0.34052972, Log Avg loss: 0.38571780, Global Avg Loss: 1.49039919, Time: 0.0210 Steps: 32800, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000875, Sample Num: 14000, Cur Loss: 0.20058610, Cur Avg Loss: 0.34004201, Log Avg loss: 0.29785510, Global Avg Loss: 1.49003572, Time: 0.0210 Steps: 32810, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000885, Sample Num: 14160, Cur Loss: 0.20480376, Cur Avg Loss: 0.33936304, Log Avg loss: 0.27995242, Global Avg Loss: 1.48966702, Time: 0.0211 Steps: 32820, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000895, Sample Num: 14320, Cur Loss: 0.17662635, Cur Avg Loss: 0.33956803, Log Avg loss: 0.35771039, Global Avg Loss: 1.48932222, Time: 0.0211 Steps: 32830, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000905, Sample Num: 14480, Cur Loss: 0.38726002, Cur Avg Loss: 0.33993225, Log Avg loss: 0.37252916, Global Avg Loss: 1.48898215, Time: 0.0209 Steps: 32840, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000915, Sample Num: 14640, Cur Loss: 0.24346609, Cur Avg Loss: 0.34007591, Log Avg loss: 0.35307791, Global Avg Loss: 1.48863637, Time: 0.0209 Steps: 32850, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000925, Sample Num: 14800, Cur Loss: 0.82142174, Cur Avg Loss: 0.34110064, Log Avg loss: 0.43486336, Global Avg Loss: 1.48831568, Time: 0.0209 Steps: 32860, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000935, Sample Num: 14960, Cur Loss: 0.36444467, Cur Avg Loss: 0.34055375, Log Avg loss: 0.28996624, Global Avg Loss: 1.48795111, Time: 0.0209 Steps: 32870, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000945, Sample Num: 15120, Cur Loss: 0.12965640, Cur Avg Loss: 0.34044173, Log Avg loss: 0.32996748, Global Avg Loss: 1.48759892, Time: 0.0210 Steps: 32880, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000955, Sample Num: 15280, Cur Loss: 0.20457959, Cur Avg Loss: 0.34086747, Log Avg loss: 0.38110029, Global Avg Loss: 1.48726250, Time: 0.0209 Steps: 32890, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000965, Sample Num: 15440, Cur Loss: 0.45384568, Cur Avg Loss: 0.34031303, Log Avg loss: 0.28736348, Global Avg Loss: 1.48689779, Time: 0.0210 Steps: 32900, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000975, Sample Num: 15600, Cur Loss: 0.16064084, Cur Avg Loss: 0.33949390, Log Avg loss: 0.26044802, Global Avg Loss: 1.48652512, Time: 0.0209 Steps: 32910, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000985, Sample Num: 15760, Cur Loss: 0.24582201, Cur Avg Loss: 0.33960800, Log Avg loss: 0.35073282, Global Avg Loss: 1.48618010, Time: 0.0210 Steps: 32920, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000995, Sample Num: 15920, Cur Loss: 0.14642739, Cur Avg Loss: 0.33961255, Log Avg loss: 0.34006076, Global Avg Loss: 1.48583206, Time: 0.0210 Steps: 32930, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001005, Sample Num: 16080, Cur Loss: 0.14786343, Cur Avg Loss: 0.33934773, Log Avg loss: 0.31299828, Global Avg Loss: 1.48547601, Time: 0.0209 Steps: 32940, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001015, Sample Num: 16240, Cur Loss: 0.15880719, Cur Avg Loss: 0.33880378, Log Avg loss: 0.28413727, Global Avg Loss: 1.48511141, Time: 0.0211 Steps: 32950, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001025, Sample Num: 16400, Cur Loss: 0.32049000, Cur Avg Loss: 0.33840908, Log Avg loss: 0.29834611, Global Avg Loss: 1.48475135, Time: 0.0247 Steps: 32960, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001035, Sample Num: 16560, Cur Loss: 0.40600204, Cur Avg Loss: 0.33750465, Log Avg loss: 0.24480099, Global Avg Loss: 1.48437526, Time: 0.0209 Steps: 32970, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001045, Sample Num: 16720, Cur Loss: 0.86401606, Cur Avg Loss: 0.33755560, Log Avg loss: 0.34282869, Global Avg Loss: 1.48402913, Time: 0.0209 Steps: 32980, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001055, Sample Num: 16880, Cur Loss: 0.30479291, Cur Avg Loss: 0.33828609, Log Avg loss: 0.41462289, Global Avg Loss: 1.48370497, Time: 0.0209 Steps: 32990, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001065, Sample Num: 17040, Cur Loss: 0.22466382, Cur Avg Loss: 0.33745445, Log Avg loss: 0.24971625, Global Avg Loss: 1.48333104, Time: 0.0209 Steps: 33000, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001075, Sample Num: 17200, Cur Loss: 0.16868670, Cur Avg Loss: 0.33703679, Log Avg loss: 0.29255546, Global Avg Loss: 1.48297030, Time: 0.0209 Steps: 33010, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001085, Sample Num: 17360, Cur Loss: 0.81921804, Cur Avg Loss: 0.33639386, Log Avg loss: 0.26727938, Global Avg Loss: 1.48260214, Time: 0.0209 Steps: 33020, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001095, Sample Num: 17520, Cur Loss: 0.36795801, Cur Avg Loss: 0.33673798, Log Avg loss: 0.37407467, Global Avg Loss: 1.48226652, Time: 0.0209 Steps: 33030, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001105, Sample Num: 17680, Cur Loss: 0.21503250, Cur Avg Loss: 0.33703403, Log Avg loss: 0.36945122, Global Avg Loss: 1.48192971, Time: 0.0209 Steps: 33040, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001115, Sample Num: 17840, Cur Loss: 0.39676777, Cur Avg Loss: 0.33641573, Log Avg loss: 0.26809391, Global Avg Loss: 1.48156244, Time: 0.0209 Steps: 33050, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001125, Sample Num: 18000, Cur Loss: 0.27521262, Cur Avg Loss: 0.33700177, Log Avg loss: 0.40234593, Global Avg Loss: 1.48123600, Time: 0.0209 Steps: 33060, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001135, Sample Num: 18160, Cur Loss: 0.27213377, Cur Avg Loss: 0.33900272, Log Avg loss: 0.56410908, Global Avg Loss: 1.48095867, Time: 0.0209 Steps: 33070, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001145, Sample Num: 18320, Cur Loss: 0.51205039, Cur Avg Loss: 0.33844777, Log Avg loss: 0.27546064, Global Avg Loss: 1.48059425, Time: 0.0209 Steps: 33080, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001155, Sample Num: 18480, Cur Loss: 0.35419768, Cur Avg Loss: 0.33869091, Log Avg loss: 0.36653022, Global Avg Loss: 1.48025758, Time: 0.0208 Steps: 33090, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001165, Sample Num: 18640, Cur Loss: 0.45461524, Cur Avg Loss: 0.33910507, Log Avg loss: 0.38694134, Global Avg Loss: 1.47992727, Time: 0.0209 Steps: 33100, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001175, Sample Num: 18800, Cur Loss: 0.19499667, Cur Avg Loss: 0.33936687, Log Avg loss: 0.36986603, Global Avg Loss: 1.47959200, Time: 0.0209 Steps: 33110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001185, Sample Num: 18960, Cur Loss: 0.26466125, Cur Avg Loss: 0.34128181, Log Avg loss: 0.56628761, Global Avg Loss: 1.47931625, Time: 0.0209 Steps: 33120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001195, Sample Num: 19120, Cur Loss: 0.23777717, Cur Avg Loss: 0.34345625, Log Avg loss: 0.60112743, Global Avg Loss: 1.47905117, Time: 0.0209 Steps: 33130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001205, Sample Num: 19280, Cur Loss: 0.57226956, Cur Avg Loss: 0.34479166, Log Avg loss: 0.50437280, Global Avg Loss: 1.47875706, Time: 0.0209 Steps: 33140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001215, Sample Num: 19440, Cur Loss: 0.34438798, Cur Avg Loss: 0.34663007, Log Avg loss: 0.56815889, Global Avg Loss: 1.47848237, Time: 0.0209 Steps: 33150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001225, Sample Num: 19600, Cur Loss: 0.57944185, Cur Avg Loss: 0.34708287, Log Avg loss: 0.40209728, Global Avg Loss: 1.47815777, Time: 0.0209 Steps: 33160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001235, Sample Num: 19760, Cur Loss: 0.21779603, Cur Avg Loss: 0.34672778, Log Avg loss: 0.30322942, Global Avg Loss: 1.47780356, Time: 0.0209 Steps: 33170, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001245, Sample Num: 19920, Cur Loss: 0.11356854, Cur Avg Loss: 0.34604321, Log Avg loss: 0.26149943, Global Avg Loss: 1.47743698, Time: 0.0209 Steps: 33180, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001255, Sample Num: 20080, Cur Loss: 0.21707915, Cur Avg Loss: 0.34649724, Log Avg loss: 0.40302337, Global Avg Loss: 1.47711326, Time: 0.0209 Steps: 33190, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001265, Sample Num: 20240, Cur Loss: 0.37765002, Cur Avg Loss: 0.34886643, Log Avg loss: 0.64619988, Global Avg Loss: 1.47686299, Time: 0.0208 Steps: 33200, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001275, Sample Num: 20400, Cur Loss: 0.66596997, Cur Avg Loss: 0.34919293, Log Avg loss: 0.39049574, Global Avg Loss: 1.47653587, Time: 0.0209 Steps: 33210, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001285, Sample Num: 20560, Cur Loss: 0.15392286, Cur Avg Loss: 0.34886077, Log Avg loss: 0.30651022, Global Avg Loss: 1.47618366, Time: 0.0216 Steps: 33220, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001295, Sample Num: 20720, Cur Loss: 0.40211287, Cur Avg Loss: 0.34951819, Log Avg loss: 0.43399622, Global Avg Loss: 1.47587003, Time: 0.0209 Steps: 33230, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001305, Sample Num: 20880, Cur Loss: 0.13650219, Cur Avg Loss: 0.34859555, Log Avg loss: 0.22911400, Global Avg Loss: 1.47549496, Time: 0.0209 Steps: 33240, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001315, Sample Num: 21040, Cur Loss: 0.15664890, Cur Avg Loss: 0.34864497, Log Avg loss: 0.35509375, Global Avg Loss: 1.47515799, Time: 0.0209 Steps: 33250, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001325, Sample Num: 21200, Cur Loss: 0.31735441, Cur Avg Loss: 0.34834649, Log Avg loss: 0.30909715, Global Avg Loss: 1.47480740, Time: 0.0210 Steps: 33260, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001335, Sample Num: 21360, Cur Loss: 0.43004620, Cur Avg Loss: 0.34811365, Log Avg loss: 0.31726271, Global Avg Loss: 1.47445948, Time: 0.0209 Steps: 33270, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001345, Sample Num: 21520, Cur Loss: 0.15322337, Cur Avg Loss: 0.34791285, Log Avg loss: 0.32110594, Global Avg Loss: 1.47411292, Time: 0.0210 Steps: 33280, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001355, Sample Num: 21680, Cur Loss: 0.33830351, Cur Avg Loss: 0.34786283, Log Avg loss: 0.34113415, Global Avg Loss: 1.47377258, Time: 0.0209 Steps: 33290, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001365, Sample Num: 21840, Cur Loss: 0.20544010, Cur Avg Loss: 0.34746343, Log Avg loss: 0.29334548, Global Avg Loss: 1.47341810, Time: 0.0210 Steps: 33300, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001375, Sample Num: 22000, Cur Loss: 0.13066635, Cur Avg Loss: 0.34777490, Log Avg loss: 0.39029007, Global Avg Loss: 1.47309293, Time: 0.0209 Steps: 33310, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001385, Sample Num: 22160, Cur Loss: 0.54035109, Cur Avg Loss: 0.34893294, Log Avg loss: 0.50816318, Global Avg Loss: 1.47280334, Time: 0.0210 Steps: 33320, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001395, Sample Num: 22320, Cur Loss: 0.40142494, Cur Avg Loss: 0.34881287, Log Avg loss: 0.33218368, Global Avg Loss: 1.47246112, Time: 0.0209 Steps: 33330, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001405, Sample Num: 22480, Cur Loss: 0.40393844, Cur Avg Loss: 0.34901775, Log Avg loss: 0.37759921, Global Avg Loss: 1.47213273, Time: 0.0209 Steps: 33340, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001415, Sample Num: 22640, Cur Loss: 0.46109891, Cur Avg Loss: 0.34840709, Log Avg loss: 0.26260861, Global Avg Loss: 1.47177005, Time: 0.0209 Steps: 33350, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001425, Sample Num: 22800, Cur Loss: 0.91842389, Cur Avg Loss: 0.34918369, Log Avg loss: 0.45907276, Global Avg Loss: 1.47146648, Time: 0.0209 Steps: 33360, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001435, Sample Num: 22960, Cur Loss: 0.11647063, Cur Avg Loss: 0.34882299, Log Avg loss: 0.29742329, Global Avg Loss: 1.47111466, Time: 0.0210 Steps: 33370, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001445, Sample Num: 23120, Cur Loss: 0.30242994, Cur Avg Loss: 0.34844376, Log Avg loss: 0.29402459, Global Avg Loss: 1.47076203, Time: 0.0210 Steps: 33380, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001455, Sample Num: 23280, Cur Loss: 0.15462084, Cur Avg Loss: 0.34820502, Log Avg loss: 0.31370705, Global Avg Loss: 1.47041550, Time: 0.0210 Steps: 33390, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001465, Sample Num: 23440, Cur Loss: 0.33437145, Cur Avg Loss: 0.34782573, Log Avg loss: 0.29263837, Global Avg Loss: 1.47006287, Time: 0.0210 Steps: 33400, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001475, Sample Num: 23600, Cur Loss: 0.53235745, Cur Avg Loss: 0.34827351, Log Avg loss: 0.41387402, Global Avg Loss: 1.46974674, Time: 0.0210 Steps: 33410, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001485, Sample Num: 23760, Cur Loss: 0.42959362, Cur Avg Loss: 0.34813208, Log Avg loss: 0.32727112, Global Avg Loss: 1.46940489, Time: 0.0210 Steps: 33420, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001495, Sample Num: 23920, Cur Loss: 0.26333719, Cur Avg Loss: 0.34849301, Log Avg loss: 0.40209125, Global Avg Loss: 1.46908562, Time: 0.0210 Steps: 33430, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001505, Sample Num: 24080, Cur Loss: 0.50269151, Cur Avg Loss: 0.34785397, Log Avg loss: 0.25231662, Global Avg Loss: 1.46872175, Time: 0.0210 Steps: 33440, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001515, Sample Num: 24240, Cur Loss: 0.39777812, Cur Avg Loss: 0.34790099, Log Avg loss: 0.35497826, Global Avg Loss: 1.46838879, Time: 0.0209 Steps: 33450, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001525, Sample Num: 24400, Cur Loss: 0.42703214, Cur Avg Loss: 0.34767513, Log Avg loss: 0.31345703, Global Avg Loss: 1.46804363, Time: 0.0210 Steps: 33460, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001535, Sample Num: 24560, Cur Loss: 0.52598244, Cur Avg Loss: 0.34764717, Log Avg loss: 0.34338274, Global Avg Loss: 1.46770761, Time: 0.0210 Steps: 33470, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001545, Sample Num: 24720, Cur Loss: 0.09780005, Cur Avg Loss: 0.34786282, Log Avg loss: 0.38096530, Global Avg Loss: 1.46738301, Time: 0.0209 Steps: 33480, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001555, Sample Num: 24880, Cur Loss: 0.16400790, Cur Avg Loss: 0.34864850, Log Avg loss: 0.47003688, Global Avg Loss: 1.46708521, Time: 0.0209 Steps: 33490, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001565, Sample Num: 25040, Cur Loss: 0.50543225, Cur Avg Loss: 0.34897382, Log Avg loss: 0.39956007, Global Avg Loss: 1.46676654, Time: 0.0209 Steps: 33500, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001575, Sample Num: 25200, Cur Loss: 0.29475608, Cur Avg Loss: 0.34917253, Log Avg loss: 0.38027141, Global Avg Loss: 1.46644231, Time: 0.0209 Steps: 33510, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001585, Sample Num: 25360, Cur Loss: 0.13841105, Cur Avg Loss: 0.34868819, Log Avg loss: 0.27240360, Global Avg Loss: 1.46608610, Time: 0.0209 Steps: 33520, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001595, Sample Num: 25520, Cur Loss: 0.35514492, Cur Avg Loss: 0.34844812, Log Avg loss: 0.31039834, Global Avg Loss: 1.46574142, Time: 0.0209 Steps: 33530, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001605, Sample Num: 25680, Cur Loss: 0.25176662, Cur Avg Loss: 0.34884232, Log Avg loss: 0.41171643, Global Avg Loss: 1.46542716, Time: 0.0209 Steps: 33540, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001615, Sample Num: 25840, Cur Loss: 0.15791442, Cur Avg Loss: 0.34838430, Log Avg loss: 0.27487262, Global Avg Loss: 1.46507230, Time: 0.0209 Steps: 33550, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001625, Sample Num: 26000, Cur Loss: 0.77247369, Cur Avg Loss: 0.34914411, Log Avg loss: 0.47185233, Global Avg Loss: 1.46477635, Time: 0.0209 Steps: 33560, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001635, Sample Num: 26160, Cur Loss: 0.22567171, Cur Avg Loss: 0.34884468, Log Avg loss: 0.30018794, Global Avg Loss: 1.46442944, Time: 0.0209 Steps: 33570, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001645, Sample Num: 26320, Cur Loss: 0.97893226, Cur Avg Loss: 0.34883573, Log Avg loss: 0.34737215, Global Avg Loss: 1.46409678, Time: 0.0209 Steps: 33580, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001655, Sample Num: 26480, Cur Loss: 0.18365321, Cur Avg Loss: 0.34832279, Log Avg loss: 0.26394454, Global Avg Loss: 1.46373949, Time: 0.0209 Steps: 33590, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001665, Sample Num: 26640, Cur Loss: 0.74213028, Cur Avg Loss: 0.34838926, Log Avg loss: 0.35938925, Global Avg Loss: 1.46341081, Time: 0.0209 Steps: 33600, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001675, Sample Num: 26800, Cur Loss: 0.27268982, Cur Avg Loss: 0.34838506, Log Avg loss: 0.34768611, Global Avg Loss: 1.46307885, Time: 0.0209 Steps: 33610, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001685, Sample Num: 26960, Cur Loss: 0.16020387, Cur Avg Loss: 0.34762069, Log Avg loss: 0.21958845, Global Avg Loss: 1.46270898, Time: 0.0209 Steps: 33620, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001695, Sample Num: 27120, Cur Loss: 0.32028574, Cur Avg Loss: 0.34705632, Log Avg loss: 0.25196067, Global Avg Loss: 1.46234896, Time: 0.0209 Steps: 33630, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001705, Sample Num: 27280, Cur Loss: 0.23610622, Cur Avg Loss: 0.34682529, Log Avg loss: 0.30766530, Global Avg Loss: 1.46200572, Time: 0.0209 Steps: 33640, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001715, Sample Num: 27440, Cur Loss: 0.22188070, Cur Avg Loss: 0.34643163, Log Avg loss: 0.27931248, Global Avg Loss: 1.46165425, Time: 0.0209 Steps: 33650, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001725, Sample Num: 27600, Cur Loss: 0.13001701, Cur Avg Loss: 0.34582035, Log Avg loss: 0.24098567, Global Avg Loss: 1.46129160, Time: 0.0209 Steps: 33660, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001735, Sample Num: 27760, Cur Loss: 0.20720965, Cur Avg Loss: 0.34520620, Log Avg loss: 0.23926511, Global Avg Loss: 1.46092866, Time: 0.0209 Steps: 33670, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001745, Sample Num: 27920, Cur Loss: 0.29912588, Cur Avg Loss: 0.34517910, Log Avg loss: 0.34047738, Global Avg Loss: 1.46059598, Time: 0.0209 Steps: 33680, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001755, Sample Num: 28080, Cur Loss: 0.28635412, Cur Avg Loss: 0.34466188, Log Avg loss: 0.25440724, Global Avg Loss: 1.46023796, Time: 0.0209 Steps: 33690, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001765, Sample Num: 28240, Cur Loss: 0.34616387, Cur Avg Loss: 0.34413665, Log Avg loss: 0.25195895, Global Avg Loss: 1.45987942, Time: 0.0209 Steps: 33700, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001775, Sample Num: 28400, Cur Loss: 0.11291294, Cur Avg Loss: 0.34376538, Log Avg loss: 0.27823562, Global Avg Loss: 1.45952889, Time: 0.0209 Steps: 33710, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001785, Sample Num: 28560, Cur Loss: 0.59888577, Cur Avg Loss: 0.34444298, Log Avg loss: 0.46471756, Global Avg Loss: 1.45923386, Time: 0.0209 Steps: 33720, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001795, Sample Num: 28720, Cur Loss: 0.47162786, Cur Avg Loss: 0.34440519, Log Avg loss: 0.33765919, Global Avg Loss: 1.45890135, Time: 0.0246 Steps: 33730, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001805, Sample Num: 28880, Cur Loss: 0.22020946, Cur Avg Loss: 0.34425764, Log Avg loss: 0.31777385, Global Avg Loss: 1.45856314, Time: 0.0209 Steps: 33740, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001815, Sample Num: 29040, Cur Loss: 0.11035232, Cur Avg Loss: 0.34449782, Log Avg loss: 0.38785010, Global Avg Loss: 1.45824589, Time: 0.0209 Steps: 33750, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001825, Sample Num: 29200, Cur Loss: 0.26756227, Cur Avg Loss: 0.34417319, Log Avg loss: 0.28525296, Global Avg Loss: 1.45789844, Time: 0.0210 Steps: 33760, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001835, Sample Num: 29360, Cur Loss: 0.34868211, Cur Avg Loss: 0.34433361, Log Avg loss: 0.37360972, Global Avg Loss: 1.45757736, Time: 0.0209 Steps: 33770, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001845, Sample Num: 29520, Cur Loss: 0.29557246, Cur Avg Loss: 0.34441119, Log Avg loss: 0.35864621, Global Avg Loss: 1.45725204, Time: 0.0209 Steps: 33780, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001855, Sample Num: 29680, Cur Loss: 0.29552794, Cur Avg Loss: 0.34405489, Log Avg loss: 0.27831808, Global Avg Loss: 1.45690314, Time: 0.0209 Steps: 33790, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001865, Sample Num: 29840, Cur Loss: 0.19154958, Cur Avg Loss: 0.34409814, Log Avg loss: 0.35212061, Global Avg Loss: 1.45657628, Time: 0.0209 Steps: 33800, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001875, Sample Num: 30000, Cur Loss: 0.48739275, Cur Avg Loss: 0.34425306, Log Avg loss: 0.37314588, Global Avg Loss: 1.45625583, Time: 0.0209 Steps: 33810, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001885, Sample Num: 30160, Cur Loss: 0.22622976, Cur Avg Loss: 0.34427896, Log Avg loss: 0.34913652, Global Avg Loss: 1.45592847, Time: 0.0210 Steps: 33820, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001895, Sample Num: 30320, Cur Loss: 0.14642987, Cur Avg Loss: 0.34437066, Log Avg loss: 0.36165449, Global Avg Loss: 1.45560501, Time: 0.0209 Steps: 33830, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001905, Sample Num: 30480, Cur Loss: 0.26637840, Cur Avg Loss: 0.34438841, Log Avg loss: 0.34775295, Global Avg Loss: 1.45527763, Time: 0.0209 Steps: 33840, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001915, Sample Num: 30640, Cur Loss: 0.54404819, Cur Avg Loss: 0.34426628, Log Avg loss: 0.32100009, Global Avg Loss: 1.45494254, Time: 0.0209 Steps: 33850, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001925, Sample Num: 30800, Cur Loss: 0.57296145, Cur Avg Loss: 0.34465220, Log Avg loss: 0.41855676, Global Avg Loss: 1.45463646, Time: 0.0210 Steps: 33860, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001935, Sample Num: 30960, Cur Loss: 0.28406700, Cur Avg Loss: 0.34496039, Log Avg loss: 0.40428543, Global Avg Loss: 1.45432635, Time: 0.0209 Steps: 33870, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001945, Sample Num: 31120, Cur Loss: 0.20415561, Cur Avg Loss: 0.34458784, Log Avg loss: 0.27250089, Global Avg Loss: 1.45397752, Time: 0.0209 Steps: 33880, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001955, Sample Num: 31280, Cur Loss: 0.26786357, Cur Avg Loss: 0.34429658, Log Avg loss: 0.28764535, Global Avg Loss: 1.45363337, Time: 0.0209 Steps: 33890, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001965, Sample Num: 31440, Cur Loss: 0.42766547, Cur Avg Loss: 0.34480209, Log Avg loss: 0.44362920, Global Avg Loss: 1.45333544, Time: 0.0209 Steps: 33900, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001975, Sample Num: 31600, Cur Loss: 0.39706752, Cur Avg Loss: 0.34443040, Log Avg loss: 0.27139378, Global Avg Loss: 1.45298688, Time: 0.0210 Steps: 33910, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001985, Sample Num: 31760, Cur Loss: 0.51027983, Cur Avg Loss: 0.34420467, Log Avg loss: 0.29962312, Global Avg Loss: 1.45264686, Time: 0.0209 Steps: 33920, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001995, Sample Num: 31920, Cur Loss: 0.47804534, Cur Avg Loss: 0.34408380, Log Avg loss: 0.32009140, Global Avg Loss: 1.45231307, Time: 0.0210 Steps: 33930, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002005, Sample Num: 32080, Cur Loss: 0.54099131, Cur Avg Loss: 0.34393329, Log Avg loss: 0.31390579, Global Avg Loss: 1.45197765, Time: 0.0209 Steps: 33940, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002015, Sample Num: 32240, Cur Loss: 0.47029278, Cur Avg Loss: 0.34383279, Log Avg loss: 0.32368305, Global Avg Loss: 1.45164531, Time: 0.0209 Steps: 33950, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002025, Sample Num: 32400, Cur Loss: 0.21107596, Cur Avg Loss: 0.34374318, Log Avg loss: 0.32568650, Global Avg Loss: 1.45131375, Time: 0.0209 Steps: 33960, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002035, Sample Num: 32560, Cur Loss: 0.25403747, Cur Avg Loss: 0.34424168, Log Avg loss: 0.44518845, Global Avg Loss: 1.45101757, Time: 0.0209 Steps: 33970, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002045, Sample Num: 32720, Cur Loss: 0.34274232, Cur Avg Loss: 0.34392341, Log Avg loss: 0.27915518, Global Avg Loss: 1.45067271, Time: 0.0209 Steps: 33980, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002055, Sample Num: 32880, Cur Loss: 0.27209434, Cur Avg Loss: 0.34405125, Log Avg loss: 0.37019514, Global Avg Loss: 1.45035482, Time: 0.0210 Steps: 33990, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002065, Sample Num: 33040, Cur Loss: 0.17530821, Cur Avg Loss: 0.34347964, Log Avg loss: 0.22601324, Global Avg Loss: 1.44999472, Time: 0.0209 Steps: 34000, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002075, Sample Num: 33200, Cur Loss: 0.12906037, Cur Avg Loss: 0.34330300, Log Avg loss: 0.30682678, Global Avg Loss: 1.44965860, Time: 0.0209 Steps: 34010, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002085, Sample Num: 33360, Cur Loss: 0.20008051, Cur Avg Loss: 0.34288801, Log Avg loss: 0.25677770, Global Avg Loss: 1.44930796, Time: 0.0209 Steps: 34020, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002095, Sample Num: 33520, Cur Loss: 0.36599594, Cur Avg Loss: 0.34243525, Log Avg loss: 0.24803400, Global Avg Loss: 1.44895495, Time: 0.0209 Steps: 34030, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002105, Sample Num: 33680, Cur Loss: 0.15697540, Cur Avg Loss: 0.34209636, Log Avg loss: 0.27109965, Global Avg Loss: 1.44860893, Time: 0.0209 Steps: 34040, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002115, Sample Num: 33840, Cur Loss: 0.43532771, Cur Avg Loss: 0.34265289, Log Avg loss: 0.45980229, Global Avg Loss: 1.44831853, Time: 0.0209 Steps: 34050, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002125, Sample Num: 34000, Cur Loss: 0.82308567, Cur Avg Loss: 0.34219755, Log Avg loss: 0.24589238, Global Avg Loss: 1.44796550, Time: 0.0209 Steps: 34060, Updated lr: 0.000069 ***** Running evaluation checkpoint-34064 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-34064 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.756870, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.417764, "eval_total_loss": 293.687789, "eval_mae": 0.522772, "eval_mse": 0.417825, "eval_r2": 0.734403, "eval_sp_statistic": 0.853167, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.888653, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.525339, "test_total_loss": 263.720201, "test_mae": 0.46794, "test_mse": 0.525511, "test_r2": 0.660831, "test_sp_statistic": 0.792778, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.840723, "test_ps_pvalue": 0.0, "lr": 6.86448553816975e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4478366181549445, "train_cur_epoch_loss": 728.5713869854808, "train_cur_epoch_avg_loss": 0.34221295771981247, "train_cur_epoch_time": 44.75687026977539, "train_cur_epoch_avg_time": 0.021022484861331795, "epoch": 16, "step": 34064} ################################################## Training, Epoch: 0017, Batch: 000006, Sample Num: 96, Cur Loss: 0.31155980, Cur Avg Loss: 0.38156429, Log Avg loss: 0.36909887, Global Avg Loss: 1.44764884, Time: 0.0242 Steps: 34070, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000016, Sample Num: 256, Cur Loss: 0.17025334, Cur Avg Loss: 0.28074964, Log Avg loss: 0.22026085, Global Avg Loss: 1.44728869, Time: 0.0210 Steps: 34080, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000026, Sample Num: 416, Cur Loss: 0.16546050, Cur Avg Loss: 0.30236450, Log Avg loss: 0.33694827, Global Avg Loss: 1.44696298, Time: 0.0209 Steps: 34090, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000036, Sample Num: 576, Cur Loss: 0.30240619, Cur Avg Loss: 0.29002869, Log Avg loss: 0.25795558, Global Avg Loss: 1.44661430, Time: 0.0209 Steps: 34100, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000046, Sample Num: 736, Cur Loss: 0.49733502, Cur Avg Loss: 0.28327610, Log Avg loss: 0.25896676, Global Avg Loss: 1.44626612, Time: 0.0209 Steps: 34110, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000056, Sample Num: 896, Cur Loss: 0.29108557, Cur Avg Loss: 0.31261363, Log Avg loss: 0.44756629, Global Avg Loss: 1.44597342, Time: 0.0209 Steps: 34120, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000066, Sample Num: 1056, Cur Loss: 0.18804000, Cur Avg Loss: 0.31655925, Log Avg loss: 0.33865474, Global Avg Loss: 1.44564897, Time: 0.0209 Steps: 34130, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000076, Sample Num: 1216, Cur Loss: 0.58622509, Cur Avg Loss: 0.31021840, Log Avg loss: 0.26836880, Global Avg Loss: 1.44530413, Time: 0.0209 Steps: 34140, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000086, Sample Num: 1376, Cur Loss: 0.44011402, Cur Avg Loss: 0.31422742, Log Avg loss: 0.34469594, Global Avg Loss: 1.44498185, Time: 0.0209 Steps: 34150, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000096, Sample Num: 1536, Cur Loss: 0.17605200, Cur Avg Loss: 0.31737334, Log Avg loss: 0.34442827, Global Avg Loss: 1.44465967, Time: 0.0209 Steps: 34160, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000106, Sample Num: 1696, Cur Loss: 0.12296870, Cur Avg Loss: 0.30805998, Log Avg loss: 0.21865171, Global Avg Loss: 1.44430088, Time: 0.0209 Steps: 34170, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000116, Sample Num: 1856, Cur Loss: 0.37555015, Cur Avg Loss: 0.30582979, Log Avg loss: 0.28218977, Global Avg Loss: 1.44396088, Time: 0.0209 Steps: 34180, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000126, Sample Num: 2016, Cur Loss: 0.40568697, Cur Avg Loss: 0.31257713, Log Avg loss: 0.39084633, Global Avg Loss: 1.44365286, Time: 0.0209 Steps: 34190, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000136, Sample Num: 2176, Cur Loss: 0.80422533, Cur Avg Loss: 0.32814834, Log Avg loss: 0.52434548, Global Avg Loss: 1.44338406, Time: 0.0209 Steps: 34200, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000146, Sample Num: 2336, Cur Loss: 0.22159821, Cur Avg Loss: 0.32984743, Log Avg loss: 0.35295513, Global Avg Loss: 1.44306531, Time: 0.0209 Steps: 34210, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000156, Sample Num: 2496, Cur Loss: 0.17999253, Cur Avg Loss: 0.32455850, Log Avg loss: 0.24734013, Global Avg Loss: 1.44271589, Time: 0.0209 Steps: 34220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000166, Sample Num: 2656, Cur Loss: 0.23272610, Cur Avg Loss: 0.32006174, Log Avg loss: 0.24991228, Global Avg Loss: 1.44236742, Time: 0.0209 Steps: 34230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000176, Sample Num: 2816, Cur Loss: 0.43669340, Cur Avg Loss: 0.31969169, Log Avg loss: 0.31354886, Global Avg Loss: 1.44203774, Time: 0.0209 Steps: 34240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000186, Sample Num: 2976, Cur Loss: 1.43008518, Cur Avg Loss: 0.32449630, Log Avg loss: 0.40905742, Global Avg Loss: 1.44173614, Time: 0.0209 Steps: 34250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000196, Sample Num: 3136, Cur Loss: 0.40358847, Cur Avg Loss: 0.33161994, Log Avg loss: 0.46411971, Global Avg Loss: 1.44145079, Time: 0.0209 Steps: 34260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000206, Sample Num: 3296, Cur Loss: 0.13154420, Cur Avg Loss: 0.32885056, Log Avg loss: 0.27457060, Global Avg Loss: 1.44111029, Time: 0.0209 Steps: 34270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000216, Sample Num: 3456, Cur Loss: 0.22677509, Cur Avg Loss: 0.32926559, Log Avg loss: 0.33781523, Global Avg Loss: 1.44078845, Time: 0.0209 Steps: 34280, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000226, Sample Num: 3616, Cur Loss: 0.24664317, Cur Avg Loss: 0.32898242, Log Avg loss: 0.32286601, Global Avg Loss: 1.44046243, Time: 0.0209 Steps: 34290, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000236, Sample Num: 3776, Cur Loss: 0.34824255, Cur Avg Loss: 0.32855725, Log Avg loss: 0.31894829, Global Avg Loss: 1.44013545, Time: 0.0209 Steps: 34300, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000246, Sample Num: 3936, Cur Loss: 0.15889725, Cur Avg Loss: 0.33065133, Log Avg loss: 0.38007175, Global Avg Loss: 1.43982649, Time: 0.0209 Steps: 34310, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000256, Sample Num: 4096, Cur Loss: 0.53354108, Cur Avg Loss: 0.32922744, Log Avg loss: 0.29419958, Global Avg Loss: 1.43949268, Time: 0.0254 Steps: 34320, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000266, Sample Num: 4256, Cur Loss: 0.07453481, Cur Avg Loss: 0.32781142, Log Avg loss: 0.29156130, Global Avg Loss: 1.43915830, Time: 0.0211 Steps: 34330, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000276, Sample Num: 4416, Cur Loss: 0.16491258, Cur Avg Loss: 0.32647574, Log Avg loss: 0.29094664, Global Avg Loss: 1.43882393, Time: 0.0210 Steps: 34340, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000286, Sample Num: 4576, Cur Loss: 0.11511897, Cur Avg Loss: 0.32683907, Log Avg loss: 0.33686701, Global Avg Loss: 1.43850313, Time: 0.0210 Steps: 34350, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000296, Sample Num: 4736, Cur Loss: 0.21519811, Cur Avg Loss: 0.32268274, Log Avg loss: 0.20381171, Global Avg Loss: 1.43814379, Time: 0.0211 Steps: 34360, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000306, Sample Num: 4896, Cur Loss: 0.97202849, Cur Avg Loss: 0.32409719, Log Avg loss: 0.36596495, Global Avg Loss: 1.43783184, Time: 0.0209 Steps: 34370, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000316, Sample Num: 5056, Cur Loss: 0.16221288, Cur Avg Loss: 0.32340997, Log Avg loss: 0.30238090, Global Avg Loss: 1.43750157, Time: 0.0210 Steps: 34380, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000326, Sample Num: 5216, Cur Loss: 0.26779845, Cur Avg Loss: 0.32141278, Log Avg loss: 0.25830180, Global Avg Loss: 1.43715868, Time: 0.0210 Steps: 34390, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000336, Sample Num: 5376, Cur Loss: 0.09344558, Cur Avg Loss: 0.32167142, Log Avg loss: 0.33010300, Global Avg Loss: 1.43683687, Time: 0.0209 Steps: 34400, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000346, Sample Num: 5536, Cur Loss: 0.58875138, Cur Avg Loss: 0.32213050, Log Avg loss: 0.33755571, Global Avg Loss: 1.43651740, Time: 0.0210 Steps: 34410, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000356, Sample Num: 5696, Cur Loss: 0.27832428, Cur Avg Loss: 0.32380455, Log Avg loss: 0.38172659, Global Avg Loss: 1.43621095, Time: 0.0210 Steps: 34420, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000366, Sample Num: 5856, Cur Loss: 0.32221848, Cur Avg Loss: 0.32259433, Log Avg loss: 0.27951055, Global Avg Loss: 1.43587500, Time: 0.0210 Steps: 34430, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000376, Sample Num: 6016, Cur Loss: 0.16391784, Cur Avg Loss: 0.32078836, Log Avg loss: 0.25468996, Global Avg Loss: 1.43553203, Time: 0.0210 Steps: 34440, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000386, Sample Num: 6176, Cur Loss: 0.10471166, Cur Avg Loss: 0.31826350, Log Avg loss: 0.22332851, Global Avg Loss: 1.43518015, Time: 0.0209 Steps: 34450, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000396, Sample Num: 6336, Cur Loss: 0.19706146, Cur Avg Loss: 0.31633874, Log Avg loss: 0.24204303, Global Avg Loss: 1.43483391, Time: 0.0210 Steps: 34460, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000406, Sample Num: 6496, Cur Loss: 0.05003832, Cur Avg Loss: 0.31704819, Log Avg loss: 0.34514231, Global Avg Loss: 1.43451779, Time: 0.0210 Steps: 34470, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000416, Sample Num: 6656, Cur Loss: 0.46448821, Cur Avg Loss: 0.31752176, Log Avg loss: 0.33674866, Global Avg Loss: 1.43419941, Time: 0.0210 Steps: 34480, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000426, Sample Num: 6816, Cur Loss: 0.52159792, Cur Avg Loss: 0.31945607, Log Avg loss: 0.39992374, Global Avg Loss: 1.43389953, Time: 0.0209 Steps: 34490, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000436, Sample Num: 6976, Cur Loss: 0.35597298, Cur Avg Loss: 0.31766543, Log Avg loss: 0.24138410, Global Avg Loss: 1.43355388, Time: 0.0209 Steps: 34500, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000446, Sample Num: 7136, Cur Loss: 0.61495394, Cur Avg Loss: 0.32259625, Log Avg loss: 0.53757979, Global Avg Loss: 1.43329425, Time: 0.0210 Steps: 34510, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000456, Sample Num: 7296, Cur Loss: 1.14653480, Cur Avg Loss: 0.33236554, Log Avg loss: 0.76807620, Global Avg Loss: 1.43310154, Time: 0.0210 Steps: 34520, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000466, Sample Num: 7456, Cur Loss: 0.12139641, Cur Avg Loss: 0.34373526, Log Avg loss: 0.86219432, Global Avg Loss: 1.43293621, Time: 0.0209 Steps: 34530, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000476, Sample Num: 7616, Cur Loss: 0.39814448, Cur Avg Loss: 0.34498486, Log Avg loss: 0.40321632, Global Avg Loss: 1.43263808, Time: 0.0210 Steps: 34540, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000486, Sample Num: 7776, Cur Loss: 0.31349072, Cur Avg Loss: 0.34391779, Log Avg loss: 0.29312492, Global Avg Loss: 1.43230827, Time: 0.0209 Steps: 34550, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000496, Sample Num: 7936, Cur Loss: 0.23750356, Cur Avg Loss: 0.34258699, Log Avg loss: 0.27791049, Global Avg Loss: 1.43197424, Time: 0.0210 Steps: 34560, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000506, Sample Num: 8096, Cur Loss: 0.16216345, Cur Avg Loss: 0.34193202, Log Avg loss: 0.30944519, Global Avg Loss: 1.43164953, Time: 0.0210 Steps: 34570, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000516, Sample Num: 8256, Cur Loss: 0.21678634, Cur Avg Loss: 0.34075149, Log Avg loss: 0.28101673, Global Avg Loss: 1.43131678, Time: 0.0211 Steps: 34580, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000526, Sample Num: 8416, Cur Loss: 0.21559906, Cur Avg Loss: 0.34230474, Log Avg loss: 0.42245257, Global Avg Loss: 1.43102512, Time: 0.0209 Steps: 34590, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000536, Sample Num: 8576, Cur Loss: 0.24376310, Cur Avg Loss: 0.34228216, Log Avg loss: 0.34109427, Global Avg Loss: 1.43071011, Time: 0.0210 Steps: 34600, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000546, Sample Num: 8736, Cur Loss: 0.23719065, Cur Avg Loss: 0.34140347, Log Avg loss: 0.29430602, Global Avg Loss: 1.43038176, Time: 0.0209 Steps: 34610, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000556, Sample Num: 8896, Cur Loss: 0.47862858, Cur Avg Loss: 0.34110269, Log Avg loss: 0.32467984, Global Avg Loss: 1.43006238, Time: 0.0210 Steps: 34620, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000566, Sample Num: 9056, Cur Loss: 0.37042990, Cur Avg Loss: 0.33974839, Log Avg loss: 0.26444937, Global Avg Loss: 1.42972579, Time: 0.0210 Steps: 34630, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000576, Sample Num: 9216, Cur Loss: 0.24649166, Cur Avg Loss: 0.33841308, Log Avg loss: 0.26283447, Global Avg Loss: 1.42938893, Time: 0.0209 Steps: 34640, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000586, Sample Num: 9376, Cur Loss: 0.20023704, Cur Avg Loss: 0.33673936, Log Avg loss: 0.24033293, Global Avg Loss: 1.42904577, Time: 0.0209 Steps: 34650, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000596, Sample Num: 9536, Cur Loss: 0.16289049, Cur Avg Loss: 0.33655705, Log Avg loss: 0.32587408, Global Avg Loss: 1.42872748, Time: 0.0209 Steps: 34660, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000606, Sample Num: 9696, Cur Loss: 0.21157806, Cur Avg Loss: 0.33677959, Log Avg loss: 0.35004291, Global Avg Loss: 1.42841635, Time: 0.0209 Steps: 34670, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000616, Sample Num: 9856, Cur Loss: 0.17925143, Cur Avg Loss: 0.33618236, Log Avg loss: 0.29999015, Global Avg Loss: 1.42809097, Time: 0.0209 Steps: 34680, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000626, Sample Num: 10016, Cur Loss: 0.16799110, Cur Avg Loss: 0.33675617, Log Avg loss: 0.37210296, Global Avg Loss: 1.42778656, Time: 0.0209 Steps: 34690, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000636, Sample Num: 10176, Cur Loss: 0.27367899, Cur Avg Loss: 0.33595546, Log Avg loss: 0.28583056, Global Avg Loss: 1.42745747, Time: 0.0209 Steps: 34700, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000646, Sample Num: 10336, Cur Loss: 0.23582320, Cur Avg Loss: 0.33544699, Log Avg loss: 0.30310886, Global Avg Loss: 1.42713354, Time: 0.0209 Steps: 34710, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000656, Sample Num: 10496, Cur Loss: 0.15453494, Cur Avg Loss: 0.33549956, Log Avg loss: 0.33889548, Global Avg Loss: 1.42682011, Time: 0.0209 Steps: 34720, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000666, Sample Num: 10656, Cur Loss: 0.28255668, Cur Avg Loss: 0.33359498, Log Avg loss: 0.20865446, Global Avg Loss: 1.42646936, Time: 0.0209 Steps: 34730, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000676, Sample Num: 10816, Cur Loss: 0.33349919, Cur Avg Loss: 0.33144073, Log Avg loss: 0.18796763, Global Avg Loss: 1.42611285, Time: 0.0209 Steps: 34740, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000686, Sample Num: 10976, Cur Loss: 0.57492030, Cur Avg Loss: 0.33149128, Log Avg loss: 0.33490830, Global Avg Loss: 1.42579884, Time: 0.0209 Steps: 34750, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000696, Sample Num: 11136, Cur Loss: 0.41621804, Cur Avg Loss: 0.33110769, Log Avg loss: 0.30479346, Global Avg Loss: 1.42547634, Time: 0.0209 Steps: 34760, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000706, Sample Num: 11296, Cur Loss: 0.26309082, Cur Avg Loss: 0.33135919, Log Avg loss: 0.34886360, Global Avg Loss: 1.42516670, Time: 0.0209 Steps: 34770, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000716, Sample Num: 11456, Cur Loss: 0.24061224, Cur Avg Loss: 0.33183144, Log Avg loss: 0.36517227, Global Avg Loss: 1.42486193, Time: 0.0209 Steps: 34780, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000726, Sample Num: 11616, Cur Loss: 0.19890146, Cur Avg Loss: 0.33145096, Log Avg loss: 0.30420891, Global Avg Loss: 1.42453981, Time: 0.0209 Steps: 34790, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000736, Sample Num: 11776, Cur Loss: 0.43613699, Cur Avg Loss: 0.33126990, Log Avg loss: 0.31812485, Global Avg Loss: 1.42422187, Time: 0.0210 Steps: 34800, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000746, Sample Num: 11936, Cur Loss: 0.22941695, Cur Avg Loss: 0.33028147, Log Avg loss: 0.25753291, Global Avg Loss: 1.42388671, Time: 0.0209 Steps: 34810, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000756, Sample Num: 12096, Cur Loss: 0.28000644, Cur Avg Loss: 0.33108944, Log Avg loss: 0.39136418, Global Avg Loss: 1.42359018, Time: 0.0209 Steps: 34820, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000766, Sample Num: 12256, Cur Loss: 0.22163808, Cur Avg Loss: 0.33063571, Log Avg loss: 0.29633349, Global Avg Loss: 1.42326654, Time: 0.0209 Steps: 34830, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000776, Sample Num: 12416, Cur Loss: 0.43691444, Cur Avg Loss: 0.33079523, Log Avg loss: 0.34301476, Global Avg Loss: 1.42295648, Time: 0.0209 Steps: 34840, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000786, Sample Num: 12576, Cur Loss: 0.39192799, Cur Avg Loss: 0.32996124, Log Avg loss: 0.26524313, Global Avg Loss: 1.42262428, Time: 0.0209 Steps: 34850, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000796, Sample Num: 12736, Cur Loss: 0.22849153, Cur Avg Loss: 0.32951181, Log Avg loss: 0.29418715, Global Avg Loss: 1.42230057, Time: 0.0209 Steps: 34860, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000806, Sample Num: 12896, Cur Loss: 0.61328030, Cur Avg Loss: 0.32998165, Log Avg loss: 0.36738075, Global Avg Loss: 1.42199804, Time: 0.0209 Steps: 34870, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000816, Sample Num: 13056, Cur Loss: 0.20162004, Cur Avg Loss: 0.32888304, Log Avg loss: 0.24033511, Global Avg Loss: 1.42165926, Time: 0.0209 Steps: 34880, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000826, Sample Num: 13216, Cur Loss: 0.17794269, Cur Avg Loss: 0.32948955, Log Avg loss: 0.37898055, Global Avg Loss: 1.42136042, Time: 0.0209 Steps: 34890, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000836, Sample Num: 13376, Cur Loss: 0.23484963, Cur Avg Loss: 0.32889485, Log Avg loss: 0.27977237, Global Avg Loss: 1.42103331, Time: 0.0209 Steps: 34900, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000846, Sample Num: 13536, Cur Loss: 0.23715676, Cur Avg Loss: 0.32899103, Log Avg loss: 0.33703201, Global Avg Loss: 1.42072280, Time: 0.0209 Steps: 34910, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000856, Sample Num: 13696, Cur Loss: 0.38076887, Cur Avg Loss: 0.32885167, Log Avg loss: 0.31706158, Global Avg Loss: 1.42040675, Time: 0.0209 Steps: 34920, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000866, Sample Num: 13856, Cur Loss: 0.35163230, Cur Avg Loss: 0.32894607, Log Avg loss: 0.33702656, Global Avg Loss: 1.42009659, Time: 0.0209 Steps: 34930, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000876, Sample Num: 14016, Cur Loss: 0.89364803, Cur Avg Loss: 0.33020502, Log Avg loss: 0.43923077, Global Avg Loss: 1.41981586, Time: 0.0209 Steps: 34940, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000886, Sample Num: 14176, Cur Loss: 0.18980312, Cur Avg Loss: 0.32961851, Log Avg loss: 0.27824023, Global Avg Loss: 1.41948923, Time: 0.0209 Steps: 34950, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000896, Sample Num: 14336, Cur Loss: 0.23190497, Cur Avg Loss: 0.33003444, Log Avg loss: 0.36688522, Global Avg Loss: 1.41918814, Time: 0.0209 Steps: 34960, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000906, Sample Num: 14496, Cur Loss: 0.14689654, Cur Avg Loss: 0.33024401, Log Avg loss: 0.34902203, Global Avg Loss: 1.41888212, Time: 0.0210 Steps: 34970, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000916, Sample Num: 14656, Cur Loss: 0.40979132, Cur Avg Loss: 0.32994581, Log Avg loss: 0.30292904, Global Avg Loss: 1.41856309, Time: 0.0209 Steps: 34980, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000926, Sample Num: 14816, Cur Loss: 0.66079664, Cur Avg Loss: 0.33006779, Log Avg loss: 0.34124031, Global Avg Loss: 1.41825520, Time: 0.0210 Steps: 34990, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000936, Sample Num: 14976, Cur Loss: 0.29324129, Cur Avg Loss: 0.32984409, Log Avg loss: 0.30912977, Global Avg Loss: 1.41793830, Time: 0.0209 Steps: 35000, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000946, Sample Num: 15136, Cur Loss: 0.19153810, Cur Avg Loss: 0.32944439, Log Avg loss: 0.29203279, Global Avg Loss: 1.41761671, Time: 0.0210 Steps: 35010, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000956, Sample Num: 15296, Cur Loss: 0.16470926, Cur Avg Loss: 0.32911969, Log Avg loss: 0.29840316, Global Avg Loss: 1.41729712, Time: 0.0210 Steps: 35020, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000966, Sample Num: 15456, Cur Loss: 0.15791471, Cur Avg Loss: 0.32938826, Log Avg loss: 0.35506345, Global Avg Loss: 1.41699388, Time: 0.0209 Steps: 35030, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000976, Sample Num: 15616, Cur Loss: 0.23993319, Cur Avg Loss: 0.32799331, Log Avg loss: 0.19324085, Global Avg Loss: 1.41664464, Time: 0.0210 Steps: 35040, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000986, Sample Num: 15776, Cur Loss: 0.29559678, Cur Avg Loss: 0.32821428, Log Avg loss: 0.34978147, Global Avg Loss: 1.41634025, Time: 0.0209 Steps: 35050, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000996, Sample Num: 15936, Cur Loss: 0.25015703, Cur Avg Loss: 0.32798610, Log Avg loss: 0.30548702, Global Avg Loss: 1.41602341, Time: 0.0210 Steps: 35060, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001006, Sample Num: 16096, Cur Loss: 0.17729363, Cur Avg Loss: 0.32749098, Log Avg loss: 0.27817732, Global Avg Loss: 1.41569896, Time: 0.0209 Steps: 35070, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001016, Sample Num: 16256, Cur Loss: 0.61570358, Cur Avg Loss: 0.32779396, Log Avg loss: 0.35827381, Global Avg Loss: 1.41539753, Time: 0.0209 Steps: 35080, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001026, Sample Num: 16416, Cur Loss: 0.32855180, Cur Avg Loss: 0.32817835, Log Avg loss: 0.36723208, Global Avg Loss: 1.41509882, Time: 0.0248 Steps: 35090, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001036, Sample Num: 16576, Cur Loss: 0.39463016, Cur Avg Loss: 0.32760602, Log Avg loss: 0.26888530, Global Avg Loss: 1.41477226, Time: 0.0209 Steps: 35100, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001046, Sample Num: 16736, Cur Loss: 0.42972994, Cur Avg Loss: 0.32700179, Log Avg loss: 0.26440351, Global Avg Loss: 1.41444462, Time: 0.0209 Steps: 35110, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001056, Sample Num: 16896, Cur Loss: 0.11305957, Cur Avg Loss: 0.32729890, Log Avg loss: 0.35837688, Global Avg Loss: 1.41414391, Time: 0.0209 Steps: 35120, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001066, Sample Num: 17056, Cur Loss: 0.55355513, Cur Avg Loss: 0.32730351, Log Avg loss: 0.32779015, Global Avg Loss: 1.41383467, Time: 0.0210 Steps: 35130, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001076, Sample Num: 17216, Cur Loss: 0.31622580, Cur Avg Loss: 0.32670656, Log Avg loss: 0.26307166, Global Avg Loss: 1.41350719, Time: 0.0209 Steps: 35140, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001086, Sample Num: 17376, Cur Loss: 0.35532820, Cur Avg Loss: 0.32644336, Log Avg loss: 0.29812285, Global Avg Loss: 1.41318987, Time: 0.0209 Steps: 35150, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001096, Sample Num: 17536, Cur Loss: 0.20031054, Cur Avg Loss: 0.32643837, Log Avg loss: 0.32589625, Global Avg Loss: 1.41288063, Time: 0.0210 Steps: 35160, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001106, Sample Num: 17696, Cur Loss: 0.18457344, Cur Avg Loss: 0.32687334, Log Avg loss: 0.37454657, Global Avg Loss: 1.41258540, Time: 0.0209 Steps: 35170, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001116, Sample Num: 17856, Cur Loss: 0.23185524, Cur Avg Loss: 0.32762948, Log Avg loss: 0.41125864, Global Avg Loss: 1.41230077, Time: 0.0209 Steps: 35180, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001126, Sample Num: 18016, Cur Loss: 0.24848622, Cur Avg Loss: 0.32763996, Log Avg loss: 0.32880922, Global Avg Loss: 1.41199287, Time: 0.0209 Steps: 35190, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001136, Sample Num: 18176, Cur Loss: 0.38673633, Cur Avg Loss: 0.32811838, Log Avg loss: 0.38198870, Global Avg Loss: 1.41170026, Time: 0.0209 Steps: 35200, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001146, Sample Num: 18336, Cur Loss: 0.33473054, Cur Avg Loss: 0.32754380, Log Avg loss: 0.26227156, Global Avg Loss: 1.41137381, Time: 0.0209 Steps: 35210, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001156, Sample Num: 18496, Cur Loss: 0.39523786, Cur Avg Loss: 0.32784293, Log Avg loss: 0.36212331, Global Avg Loss: 1.41107589, Time: 0.0209 Steps: 35220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001166, Sample Num: 18656, Cur Loss: 0.34998864, Cur Avg Loss: 0.32761616, Log Avg loss: 0.30140083, Global Avg Loss: 1.41076091, Time: 0.0209 Steps: 35230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001176, Sample Num: 18816, Cur Loss: 0.22030479, Cur Avg Loss: 0.32757655, Log Avg loss: 0.32295804, Global Avg Loss: 1.41045223, Time: 0.0209 Steps: 35240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001186, Sample Num: 18976, Cur Loss: 0.37005788, Cur Avg Loss: 0.32724382, Log Avg loss: 0.28811532, Global Avg Loss: 1.41013384, Time: 0.0209 Steps: 35250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001196, Sample Num: 19136, Cur Loss: 0.13797513, Cur Avg Loss: 0.32661946, Log Avg loss: 0.25256960, Global Avg Loss: 1.40980554, Time: 0.0209 Steps: 35260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001206, Sample Num: 19296, Cur Loss: 0.53637892, Cur Avg Loss: 0.32705894, Log Avg loss: 0.37962059, Global Avg Loss: 1.40951346, Time: 0.0209 Steps: 35270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001216, Sample Num: 19456, Cur Loss: 0.33295727, Cur Avg Loss: 0.32706511, Log Avg loss: 0.32781006, Global Avg Loss: 1.40920685, Time: 0.0209 Steps: 35280, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001226, Sample Num: 19616, Cur Loss: 0.14287980, Cur Avg Loss: 0.32632360, Log Avg loss: 0.23615624, Global Avg Loss: 1.40887445, Time: 0.0209 Steps: 35290, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001236, Sample Num: 19776, Cur Loss: 0.29059303, Cur Avg Loss: 0.32539660, Log Avg loss: 0.21174532, Global Avg Loss: 1.40853532, Time: 0.0209 Steps: 35300, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001246, Sample Num: 19936, Cur Loss: 0.67468226, Cur Avg Loss: 0.32624395, Log Avg loss: 0.43097684, Global Avg Loss: 1.40825847, Time: 0.0209 Steps: 35310, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001256, Sample Num: 20096, Cur Loss: 0.26896548, Cur Avg Loss: 0.32613928, Log Avg loss: 0.31309715, Global Avg Loss: 1.40794840, Time: 0.0209 Steps: 35320, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001266, Sample Num: 20256, Cur Loss: 0.13991803, Cur Avg Loss: 0.32633855, Log Avg loss: 0.35136667, Global Avg Loss: 1.40764934, Time: 0.0209 Steps: 35330, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001276, Sample Num: 20416, Cur Loss: 0.44512740, Cur Avg Loss: 0.32660102, Log Avg loss: 0.35983022, Global Avg Loss: 1.40735284, Time: 0.0209 Steps: 35340, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001286, Sample Num: 20576, Cur Loss: 0.12844060, Cur Avg Loss: 0.32585255, Log Avg loss: 0.23034819, Global Avg Loss: 1.40701989, Time: 0.0247 Steps: 35350, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001296, Sample Num: 20736, Cur Loss: 0.30688438, Cur Avg Loss: 0.32551446, Log Avg loss: 0.28203571, Global Avg Loss: 1.40670173, Time: 0.0210 Steps: 35360, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001306, Sample Num: 20896, Cur Loss: 0.20665957, Cur Avg Loss: 0.32552056, Log Avg loss: 0.32631153, Global Avg Loss: 1.40639628, Time: 0.0210 Steps: 35370, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001316, Sample Num: 21056, Cur Loss: 0.37780631, Cur Avg Loss: 0.32597587, Log Avg loss: 0.38543842, Global Avg Loss: 1.40610771, Time: 0.0210 Steps: 35380, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001326, Sample Num: 21216, Cur Loss: 0.62219167, Cur Avg Loss: 0.32575525, Log Avg loss: 0.29672218, Global Avg Loss: 1.40579424, Time: 0.0209 Steps: 35390, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001336, Sample Num: 21376, Cur Loss: 0.19490035, Cur Avg Loss: 0.32532010, Log Avg loss: 0.26761872, Global Avg Loss: 1.40547272, Time: 0.0210 Steps: 35400, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001346, Sample Num: 21536, Cur Loss: 0.10817927, Cur Avg Loss: 0.32486689, Log Avg loss: 0.26431893, Global Avg Loss: 1.40515045, Time: 0.0209 Steps: 35410, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001356, Sample Num: 21696, Cur Loss: 0.12359538, Cur Avg Loss: 0.32512769, Log Avg loss: 0.36023139, Global Avg Loss: 1.40485544, Time: 0.0211 Steps: 35420, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001366, Sample Num: 21856, Cur Loss: 0.40279630, Cur Avg Loss: 0.32480069, Log Avg loss: 0.28045881, Global Avg Loss: 1.40453808, Time: 0.0209 Steps: 35430, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001376, Sample Num: 22016, Cur Loss: 0.25971973, Cur Avg Loss: 0.32470250, Log Avg loss: 0.31129063, Global Avg Loss: 1.40422961, Time: 0.0209 Steps: 35440, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001386, Sample Num: 22176, Cur Loss: 0.19146934, Cur Avg Loss: 0.32442742, Log Avg loss: 0.28657555, Global Avg Loss: 1.40391433, Time: 0.0210 Steps: 35450, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001396, Sample Num: 22336, Cur Loss: 0.38042974, Cur Avg Loss: 0.32427971, Log Avg loss: 0.30380802, Global Avg Loss: 1.40360409, Time: 0.0210 Steps: 35460, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001406, Sample Num: 22496, Cur Loss: 0.62547523, Cur Avg Loss: 0.32430437, Log Avg loss: 0.32774619, Global Avg Loss: 1.40330078, Time: 0.0210 Steps: 35470, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001416, Sample Num: 22656, Cur Loss: 1.01050925, Cur Avg Loss: 0.32545440, Log Avg loss: 0.48714895, Global Avg Loss: 1.40304256, Time: 0.0210 Steps: 35480, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001426, Sample Num: 22816, Cur Loss: 0.61779541, Cur Avg Loss: 0.32663501, Log Avg loss: 0.49380837, Global Avg Loss: 1.40278636, Time: 0.0209 Steps: 35490, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001436, Sample Num: 22976, Cur Loss: 0.86323863, Cur Avg Loss: 0.32882762, Log Avg loss: 0.64149495, Global Avg Loss: 1.40257192, Time: 0.0209 Steps: 35500, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001446, Sample Num: 23136, Cur Loss: 0.78268045, Cur Avg Loss: 0.33026725, Log Avg loss: 0.53699818, Global Avg Loss: 1.40232816, Time: 0.0209 Steps: 35510, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001456, Sample Num: 23296, Cur Loss: 0.21860901, Cur Avg Loss: 0.33100714, Log Avg loss: 0.43799495, Global Avg Loss: 1.40205667, Time: 0.0210 Steps: 35520, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001466, Sample Num: 23456, Cur Loss: 0.17818782, Cur Avg Loss: 0.33128759, Log Avg loss: 0.37212154, Global Avg Loss: 1.40176679, Time: 0.0209 Steps: 35530, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001476, Sample Num: 23616, Cur Loss: 0.40584821, Cur Avg Loss: 0.33166941, Log Avg loss: 0.38764396, Global Avg Loss: 1.40148145, Time: 0.0210 Steps: 35540, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001486, Sample Num: 23776, Cur Loss: 0.20202014, Cur Avg Loss: 0.33209446, Log Avg loss: 0.39483103, Global Avg Loss: 1.40119828, Time: 0.0210 Steps: 35550, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001496, Sample Num: 23936, Cur Loss: 0.28523433, Cur Avg Loss: 0.33200957, Log Avg loss: 0.31939492, Global Avg Loss: 1.40089406, Time: 0.0209 Steps: 35560, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001506, Sample Num: 24096, Cur Loss: 0.48601082, Cur Avg Loss: 0.33224623, Log Avg loss: 0.36765089, Global Avg Loss: 1.40060358, Time: 0.0209 Steps: 35570, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001516, Sample Num: 24256, Cur Loss: 0.18985251, Cur Avg Loss: 0.33215350, Log Avg loss: 0.31818822, Global Avg Loss: 1.40029936, Time: 0.0209 Steps: 35580, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001526, Sample Num: 24416, Cur Loss: 0.18109171, Cur Avg Loss: 0.33209237, Log Avg loss: 0.32282569, Global Avg Loss: 1.39999661, Time: 0.0209 Steps: 35590, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001536, Sample Num: 24576, Cur Loss: 0.09385099, Cur Avg Loss: 0.33119666, Log Avg loss: 0.19451075, Global Avg Loss: 1.39965800, Time: 0.0255 Steps: 35600, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001546, Sample Num: 24736, Cur Loss: 0.49201924, Cur Avg Loss: 0.33129625, Log Avg loss: 0.34659367, Global Avg Loss: 1.39936227, Time: 0.0210 Steps: 35610, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001556, Sample Num: 24896, Cur Loss: 0.08084529, Cur Avg Loss: 0.33109567, Log Avg loss: 0.30008606, Global Avg Loss: 1.39905366, Time: 0.0209 Steps: 35620, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001566, Sample Num: 25056, Cur Loss: 0.23248783, Cur Avg Loss: 0.33113166, Log Avg loss: 0.33673115, Global Avg Loss: 1.39875551, Time: 0.0209 Steps: 35630, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001576, Sample Num: 25216, Cur Loss: 0.11030571, Cur Avg Loss: 0.33076355, Log Avg loss: 0.27311850, Global Avg Loss: 1.39843967, Time: 0.0209 Steps: 35640, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001586, Sample Num: 25376, Cur Loss: 0.32042333, Cur Avg Loss: 0.33088715, Log Avg loss: 0.35036527, Global Avg Loss: 1.39814568, Time: 0.0209 Steps: 35650, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001596, Sample Num: 25536, Cur Loss: 0.26963925, Cur Avg Loss: 0.33037245, Log Avg loss: 0.24874192, Global Avg Loss: 1.39782336, Time: 0.0209 Steps: 35660, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001606, Sample Num: 25696, Cur Loss: 0.26145878, Cur Avg Loss: 0.33021686, Log Avg loss: 0.30538449, Global Avg Loss: 1.39751710, Time: 0.0209 Steps: 35670, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001616, Sample Num: 25856, Cur Loss: 0.13756448, Cur Avg Loss: 0.32961662, Log Avg loss: 0.23321864, Global Avg Loss: 1.39719078, Time: 0.0209 Steps: 35680, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001626, Sample Num: 26016, Cur Loss: 0.27636218, Cur Avg Loss: 0.32900355, Log Avg loss: 0.22993042, Global Avg Loss: 1.39686372, Time: 0.0209 Steps: 35690, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001636, Sample Num: 26176, Cur Loss: 0.32661498, Cur Avg Loss: 0.32881657, Log Avg loss: 0.29841415, Global Avg Loss: 1.39655604, Time: 0.0209 Steps: 35700, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001646, Sample Num: 26336, Cur Loss: 0.39721546, Cur Avg Loss: 0.32852314, Log Avg loss: 0.28051701, Global Avg Loss: 1.39624351, Time: 0.0209 Steps: 35710, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001656, Sample Num: 26496, Cur Loss: 0.18414070, Cur Avg Loss: 0.32879208, Log Avg loss: 0.37305988, Global Avg Loss: 1.39595706, Time: 0.0209 Steps: 35720, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001666, Sample Num: 26656, Cur Loss: 0.25796252, Cur Avg Loss: 0.32946632, Log Avg loss: 0.44112124, Global Avg Loss: 1.39568983, Time: 0.0209 Steps: 35730, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001676, Sample Num: 26816, Cur Loss: 0.13776425, Cur Avg Loss: 0.32917510, Log Avg loss: 0.28065752, Global Avg Loss: 1.39537784, Time: 0.0209 Steps: 35740, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001686, Sample Num: 26976, Cur Loss: 0.16223407, Cur Avg Loss: 0.32860463, Log Avg loss: 0.23299344, Global Avg Loss: 1.39505270, Time: 0.0209 Steps: 35750, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001696, Sample Num: 27136, Cur Loss: 0.61211854, Cur Avg Loss: 0.32884711, Log Avg loss: 0.36973051, Global Avg Loss: 1.39476598, Time: 0.0209 Steps: 35760, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001706, Sample Num: 27296, Cur Loss: 0.30958554, Cur Avg Loss: 0.32901961, Log Avg loss: 0.35827533, Global Avg Loss: 1.39447621, Time: 0.0209 Steps: 35770, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001716, Sample Num: 27456, Cur Loss: 0.35871056, Cur Avg Loss: 0.32925329, Log Avg loss: 0.36911830, Global Avg Loss: 1.39418964, Time: 0.0209 Steps: 35780, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001726, Sample Num: 27616, Cur Loss: 0.37699926, Cur Avg Loss: 0.32885510, Log Avg loss: 0.26052600, Global Avg Loss: 1.39387288, Time: 0.0209 Steps: 35790, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001736, Sample Num: 27776, Cur Loss: 0.17884031, Cur Avg Loss: 0.32857315, Log Avg loss: 0.27990832, Global Avg Loss: 1.39356172, Time: 0.0209 Steps: 35800, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001746, Sample Num: 27936, Cur Loss: 0.33452708, Cur Avg Loss: 0.32853929, Log Avg loss: 0.32266075, Global Avg Loss: 1.39326267, Time: 0.0209 Steps: 35810, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001756, Sample Num: 28096, Cur Loss: 0.37557250, Cur Avg Loss: 0.32794370, Log Avg loss: 0.22395368, Global Avg Loss: 1.39293623, Time: 0.0209 Steps: 35820, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001766, Sample Num: 28256, Cur Loss: 0.36700866, Cur Avg Loss: 0.32822315, Log Avg loss: 0.37729597, Global Avg Loss: 1.39265277, Time: 0.0209 Steps: 35830, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001776, Sample Num: 28416, Cur Loss: 0.20254612, Cur Avg Loss: 0.32836413, Log Avg loss: 0.35326122, Global Avg Loss: 1.39236276, Time: 0.0209 Steps: 35840, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001786, Sample Num: 28576, Cur Loss: 0.36316401, Cur Avg Loss: 0.32841572, Log Avg loss: 0.33757817, Global Avg Loss: 1.39206854, Time: 0.0209 Steps: 35850, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001796, Sample Num: 28736, Cur Loss: 0.09729324, Cur Avg Loss: 0.32801779, Log Avg loss: 0.25694751, Global Avg Loss: 1.39175199, Time: 0.0246 Steps: 35860, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001806, Sample Num: 28896, Cur Loss: 0.09895645, Cur Avg Loss: 0.32810522, Log Avg loss: 0.34380619, Global Avg Loss: 1.39145984, Time: 0.0209 Steps: 35870, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001816, Sample Num: 29056, Cur Loss: 0.55041277, Cur Avg Loss: 0.32783610, Log Avg loss: 0.27923453, Global Avg Loss: 1.39114986, Time: 0.0209 Steps: 35880, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001826, Sample Num: 29216, Cur Loss: 0.23858827, Cur Avg Loss: 0.32770472, Log Avg loss: 0.30384441, Global Avg Loss: 1.39084690, Time: 0.0209 Steps: 35890, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001836, Sample Num: 29376, Cur Loss: 0.51875114, Cur Avg Loss: 0.32822398, Log Avg loss: 0.42304135, Global Avg Loss: 1.39057732, Time: 0.0209 Steps: 35900, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001846, Sample Num: 29536, Cur Loss: 0.42780483, Cur Avg Loss: 0.32827639, Log Avg loss: 0.33789917, Global Avg Loss: 1.39028418, Time: 0.0209 Steps: 35910, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001856, Sample Num: 29696, Cur Loss: 0.56395864, Cur Avg Loss: 0.32901408, Log Avg loss: 0.46519157, Global Avg Loss: 1.39002663, Time: 0.0209 Steps: 35920, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001866, Sample Num: 29856, Cur Loss: 0.15104347, Cur Avg Loss: 0.32869176, Log Avg loss: 0.26887004, Global Avg Loss: 1.38971459, Time: 0.0209 Steps: 35930, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001876, Sample Num: 30016, Cur Loss: 0.37900388, Cur Avg Loss: 0.32903701, Log Avg loss: 0.39346077, Global Avg Loss: 1.38943740, Time: 0.0209 Steps: 35940, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001886, Sample Num: 30176, Cur Loss: 0.25897050, Cur Avg Loss: 0.32961320, Log Avg loss: 0.43770509, Global Avg Loss: 1.38917266, Time: 0.0209 Steps: 35950, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001896, Sample Num: 30336, Cur Loss: 0.05804130, Cur Avg Loss: 0.32914824, Log Avg loss: 0.24145724, Global Avg Loss: 1.38885349, Time: 0.0209 Steps: 35960, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001906, Sample Num: 30496, Cur Loss: 0.33614165, Cur Avg Loss: 0.32904134, Log Avg loss: 0.30877379, Global Avg Loss: 1.38855322, Time: 0.0209 Steps: 35970, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001916, Sample Num: 30656, Cur Loss: 0.09235086, Cur Avg Loss: 0.32897756, Log Avg loss: 0.31682024, Global Avg Loss: 1.38825535, Time: 0.0209 Steps: 35980, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001926, Sample Num: 30816, Cur Loss: 0.45428807, Cur Avg Loss: 0.32877676, Log Avg loss: 0.29030398, Global Avg Loss: 1.38795028, Time: 0.0209 Steps: 35990, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001936, Sample Num: 30976, Cur Loss: 0.47613469, Cur Avg Loss: 0.32879325, Log Avg loss: 0.33196879, Global Avg Loss: 1.38765695, Time: 0.0209 Steps: 36000, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001946, Sample Num: 31136, Cur Loss: 0.40457758, Cur Avg Loss: 0.32932853, Log Avg loss: 0.43295833, Global Avg Loss: 1.38739183, Time: 0.0210 Steps: 36010, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001956, Sample Num: 31296, Cur Loss: 0.18770331, Cur Avg Loss: 0.32934341, Log Avg loss: 0.33223875, Global Avg Loss: 1.38709890, Time: 0.0209 Steps: 36020, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001966, Sample Num: 31456, Cur Loss: 0.89814955, Cur Avg Loss: 0.32946070, Log Avg loss: 0.35240272, Global Avg Loss: 1.38681172, Time: 0.0209 Steps: 36030, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001976, Sample Num: 31616, Cur Loss: 0.22484724, Cur Avg Loss: 0.32935461, Log Avg loss: 0.30849801, Global Avg Loss: 1.38651252, Time: 0.0209 Steps: 36040, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001986, Sample Num: 31776, Cur Loss: 0.19968930, Cur Avg Loss: 0.32906312, Log Avg loss: 0.27146520, Global Avg Loss: 1.38620322, Time: 0.0209 Steps: 36050, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001996, Sample Num: 31936, Cur Loss: 0.31743720, Cur Avg Loss: 0.32868869, Log Avg loss: 0.25432646, Global Avg Loss: 1.38588933, Time: 0.0209 Steps: 36060, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002006, Sample Num: 32096, Cur Loss: 0.37023139, Cur Avg Loss: 0.32925746, Log Avg loss: 0.44278405, Global Avg Loss: 1.38562786, Time: 0.0209 Steps: 36070, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002016, Sample Num: 32256, Cur Loss: 0.63130927, Cur Avg Loss: 0.32937786, Log Avg loss: 0.35353023, Global Avg Loss: 1.38534181, Time: 0.0209 Steps: 36080, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002026, Sample Num: 32416, Cur Loss: 0.35072345, Cur Avg Loss: 0.32932603, Log Avg loss: 0.31887767, Global Avg Loss: 1.38504630, Time: 0.0209 Steps: 36090, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002036, Sample Num: 32576, Cur Loss: 0.16437648, Cur Avg Loss: 0.32952206, Log Avg loss: 0.36923717, Global Avg Loss: 1.38476492, Time: 0.0209 Steps: 36100, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002046, Sample Num: 32736, Cur Loss: 0.52424431, Cur Avg Loss: 0.32940232, Log Avg loss: 0.30502228, Global Avg Loss: 1.38446590, Time: 0.0209 Steps: 36110, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002056, Sample Num: 32896, Cur Loss: 0.19176027, Cur Avg Loss: 0.32897096, Log Avg loss: 0.24071511, Global Avg Loss: 1.38414925, Time: 0.0212 Steps: 36120, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002066, Sample Num: 33056, Cur Loss: 0.61384559, Cur Avg Loss: 0.32890805, Log Avg loss: 0.31597327, Global Avg Loss: 1.38385360, Time: 0.0211 Steps: 36130, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002076, Sample Num: 33216, Cur Loss: 0.24427406, Cur Avg Loss: 0.32860221, Log Avg loss: 0.26541749, Global Avg Loss: 1.38354413, Time: 0.0212 Steps: 36140, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002086, Sample Num: 33376, Cur Loss: 0.25334406, Cur Avg Loss: 0.32848767, Log Avg loss: 0.30470829, Global Avg Loss: 1.38324569, Time: 0.0211 Steps: 36150, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002096, Sample Num: 33536, Cur Loss: 0.28450465, Cur Avg Loss: 0.32830052, Log Avg loss: 0.28926090, Global Avg Loss: 1.38294315, Time: 0.0211 Steps: 36160, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002106, Sample Num: 33696, Cur Loss: 0.22987303, Cur Avg Loss: 0.32814462, Log Avg loss: 0.29546767, Global Avg Loss: 1.38264250, Time: 0.0211 Steps: 36170, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002116, Sample Num: 33856, Cur Loss: 0.17580192, Cur Avg Loss: 0.32770822, Log Avg loss: 0.23580342, Global Avg Loss: 1.38232552, Time: 0.0211 Steps: 36180, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002126, Sample Num: 34016, Cur Loss: 0.33398670, Cur Avg Loss: 0.32731218, Log Avg loss: 0.24350868, Global Avg Loss: 1.38201084, Time: 0.0211 Steps: 36190, Updated lr: 0.000067 ***** Running evaluation checkpoint-36193 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-36193 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.823971, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.467628, "eval_total_loss": 328.742211, "eval_mae": 0.561386, "eval_mse": 0.467714, "eval_r2": 0.70269, "eval_sp_statistic": 0.839609, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.880288, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.569348, "test_total_loss": 285.812807, "test_mae": 0.484473, "test_mse": 0.569525, "test_r2": 0.632424, "test_sp_statistic": 0.778849, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.822562, "test_ps_pvalue": 0.0, "lr": 6.662588904694168e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.381951507277682, "train_cur_epoch_loss": 697.864342071116, "train_cur_epoch_avg_loss": 0.32778973324148236, "train_cur_epoch_time": 44.823970556259155, "train_cur_epoch_avg_time": 0.021054002140093543, "epoch": 17, "step": 36193} ################################################## Training, Epoch: 0018, Batch: 000007, Sample Num: 112, Cur Loss: 0.28557566, Cur Avg Loss: 0.42587908, Log Avg loss: 0.49798078, Global Avg Loss: 1.38176663, Time: 0.0232 Steps: 36200, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000017, Sample Num: 272, Cur Loss: 0.37129402, Cur Avg Loss: 0.33154931, Log Avg loss: 0.26551846, Global Avg Loss: 1.38145836, Time: 0.0210 Steps: 36210, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000027, Sample Num: 432, Cur Loss: 1.21611750, Cur Avg Loss: 0.37036893, Log Avg loss: 0.43636229, Global Avg Loss: 1.38119743, Time: 0.0209 Steps: 36220, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000037, Sample Num: 592, Cur Loss: 0.12436387, Cur Avg Loss: 0.38350265, Log Avg loss: 0.41896369, Global Avg Loss: 1.38093184, Time: 0.0210 Steps: 36230, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000047, Sample Num: 752, Cur Loss: 1.15679348, Cur Avg Loss: 0.37857775, Log Avg loss: 0.36035561, Global Avg Loss: 1.38065022, Time: 0.0210 Steps: 36240, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000057, Sample Num: 912, Cur Loss: 0.20455754, Cur Avg Loss: 0.36538709, Log Avg loss: 0.30339098, Global Avg Loss: 1.38035305, Time: 0.0209 Steps: 36250, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000067, Sample Num: 1072, Cur Loss: 0.47379130, Cur Avg Loss: 0.35813253, Log Avg loss: 0.31678159, Global Avg Loss: 1.38005973, Time: 0.0210 Steps: 36260, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000077, Sample Num: 1232, Cur Loss: 0.09557282, Cur Avg Loss: 0.34783691, Log Avg loss: 0.27885619, Global Avg Loss: 1.37975612, Time: 0.0210 Steps: 36270, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000087, Sample Num: 1392, Cur Loss: 0.51447606, Cur Avg Loss: 0.33868810, Log Avg loss: 0.26824233, Global Avg Loss: 1.37944975, Time: 0.0209 Steps: 36280, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000097, Sample Num: 1552, Cur Loss: 0.55880588, Cur Avg Loss: 0.32889002, Log Avg loss: 0.24364672, Global Avg Loss: 1.37913677, Time: 0.0210 Steps: 36290, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000107, Sample Num: 1712, Cur Loss: 0.11994956, Cur Avg Loss: 0.31917742, Log Avg loss: 0.22496520, Global Avg Loss: 1.37881881, Time: 0.0210 Steps: 36300, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000117, Sample Num: 1872, Cur Loss: 0.64526618, Cur Avg Loss: 0.32160142, Log Avg loss: 0.34753825, Global Avg Loss: 1.37853479, Time: 0.0210 Steps: 36310, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000127, Sample Num: 2032, Cur Loss: 0.15140855, Cur Avg Loss: 0.31756450, Log Avg loss: 0.27033248, Global Avg Loss: 1.37822967, Time: 0.0208 Steps: 36320, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000137, Sample Num: 2192, Cur Loss: 0.40437180, Cur Avg Loss: 0.31545658, Log Avg loss: 0.28868606, Global Avg Loss: 1.37792977, Time: 0.0208 Steps: 36330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000147, Sample Num: 2352, Cur Loss: 0.26494497, Cur Avg Loss: 0.31958025, Log Avg loss: 0.37607439, Global Avg Loss: 1.37765408, Time: 0.0208 Steps: 36340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000157, Sample Num: 2512, Cur Loss: 0.80593133, Cur Avg Loss: 0.32747257, Log Avg loss: 0.44348982, Global Avg Loss: 1.37739709, Time: 0.0208 Steps: 36350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000167, Sample Num: 2672, Cur Loss: 0.08989394, Cur Avg Loss: 0.31640505, Log Avg loss: 0.14264491, Global Avg Loss: 1.37705750, Time: 0.0208 Steps: 36360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000177, Sample Num: 2832, Cur Loss: 0.17316759, Cur Avg Loss: 0.31969265, Log Avg loss: 0.37459552, Global Avg Loss: 1.37678187, Time: 0.0208 Steps: 36370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000187, Sample Num: 2992, Cur Loss: 0.95685720, Cur Avg Loss: 0.33133951, Log Avg loss: 0.53748902, Global Avg Loss: 1.37655117, Time: 0.0209 Steps: 36380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000197, Sample Num: 3152, Cur Loss: 0.17245592, Cur Avg Loss: 0.32968765, Log Avg loss: 0.29879775, Global Avg Loss: 1.37625500, Time: 0.0208 Steps: 36390, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000207, Sample Num: 3312, Cur Loss: 0.10232576, Cur Avg Loss: 0.32691885, Log Avg loss: 0.27237364, Global Avg Loss: 1.37595173, Time: 0.0209 Steps: 36400, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000217, Sample Num: 3472, Cur Loss: 0.17792687, Cur Avg Loss: 0.32557492, Log Avg loss: 0.29775561, Global Avg Loss: 1.37565561, Time: 0.0209 Steps: 36410, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000227, Sample Num: 3632, Cur Loss: 0.22646850, Cur Avg Loss: 0.32126019, Log Avg loss: 0.22763048, Global Avg Loss: 1.37534039, Time: 0.0209 Steps: 36420, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000237, Sample Num: 3792, Cur Loss: 0.21188304, Cur Avg Loss: 0.31718477, Log Avg loss: 0.22467271, Global Avg Loss: 1.37502453, Time: 0.0209 Steps: 36430, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000247, Sample Num: 3952, Cur Loss: 0.27877027, Cur Avg Loss: 0.31582266, Log Avg loss: 0.28354071, Global Avg Loss: 1.37472500, Time: 0.0208 Steps: 36440, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000257, Sample Num: 4112, Cur Loss: 0.61902368, Cur Avg Loss: 0.31886742, Log Avg loss: 0.39407290, Global Avg Loss: 1.37445596, Time: 0.0246 Steps: 36450, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000267, Sample Num: 4272, Cur Loss: 0.10672517, Cur Avg Loss: 0.32216796, Log Avg loss: 0.40699183, Global Avg Loss: 1.37419061, Time: 0.0210 Steps: 36460, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000277, Sample Num: 4432, Cur Loss: 0.16416559, Cur Avg Loss: 0.32345397, Log Avg loss: 0.35779041, Global Avg Loss: 1.37391192, Time: 0.0210 Steps: 36470, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000287, Sample Num: 4592, Cur Loss: 0.44283360, Cur Avg Loss: 0.32403553, Log Avg loss: 0.34014495, Global Avg Loss: 1.37362854, Time: 0.0209 Steps: 36480, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000297, Sample Num: 4752, Cur Loss: 0.47105423, Cur Avg Loss: 0.32530768, Log Avg loss: 0.36181819, Global Avg Loss: 1.37335125, Time: 0.0210 Steps: 36490, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000307, Sample Num: 4912, Cur Loss: 0.25124079, Cur Avg Loss: 0.32335826, Log Avg loss: 0.26546044, Global Avg Loss: 1.37304772, Time: 0.0209 Steps: 36500, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000317, Sample Num: 5072, Cur Loss: 0.11926403, Cur Avg Loss: 0.32079540, Log Avg loss: 0.24211582, Global Avg Loss: 1.37273796, Time: 0.0210 Steps: 36510, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000327, Sample Num: 5232, Cur Loss: 0.22008801, Cur Avg Loss: 0.32310848, Log Avg loss: 0.39643289, Global Avg Loss: 1.37247063, Time: 0.0210 Steps: 36520, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000337, Sample Num: 5392, Cur Loss: 0.19160666, Cur Avg Loss: 0.32241712, Log Avg loss: 0.29980986, Global Avg Loss: 1.37217699, Time: 0.0210 Steps: 36530, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000347, Sample Num: 5552, Cur Loss: 0.24946862, Cur Avg Loss: 0.32219488, Log Avg loss: 0.31470542, Global Avg Loss: 1.37188759, Time: 0.0210 Steps: 36540, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000357, Sample Num: 5712, Cur Loss: 0.23843808, Cur Avg Loss: 0.32001138, Log Avg loss: 0.24424371, Global Avg Loss: 1.37157907, Time: 0.0210 Steps: 36550, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000367, Sample Num: 5872, Cur Loss: 0.21771318, Cur Avg Loss: 0.32009182, Log Avg loss: 0.32296343, Global Avg Loss: 1.37129225, Time: 0.0210 Steps: 36560, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000377, Sample Num: 6032, Cur Loss: 0.47245100, Cur Avg Loss: 0.32062242, Log Avg loss: 0.34009564, Global Avg Loss: 1.37101027, Time: 0.0210 Steps: 36570, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000387, Sample Num: 6192, Cur Loss: 0.64986634, Cur Avg Loss: 0.32269178, Log Avg loss: 0.40070673, Global Avg Loss: 1.37074501, Time: 0.0210 Steps: 36580, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000397, Sample Num: 6352, Cur Loss: 0.49888399, Cur Avg Loss: 0.32315135, Log Avg loss: 0.34093667, Global Avg Loss: 1.37046357, Time: 0.0210 Steps: 36590, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000407, Sample Num: 6512, Cur Loss: 0.45194197, Cur Avg Loss: 0.32388556, Log Avg loss: 0.35303362, Global Avg Loss: 1.37018558, Time: 0.0209 Steps: 36600, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000417, Sample Num: 6672, Cur Loss: 0.39540362, Cur Avg Loss: 0.32374588, Log Avg loss: 0.31806086, Global Avg Loss: 1.36989820, Time: 0.0209 Steps: 36610, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000427, Sample Num: 6832, Cur Loss: 0.31901115, Cur Avg Loss: 0.32240557, Log Avg loss: 0.26651461, Global Avg Loss: 1.36959689, Time: 0.0209 Steps: 36620, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000437, Sample Num: 6992, Cur Loss: 0.46113861, Cur Avg Loss: 0.32125235, Log Avg loss: 0.27201004, Global Avg Loss: 1.36929725, Time: 0.0210 Steps: 36630, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000447, Sample Num: 7152, Cur Loss: 0.22086576, Cur Avg Loss: 0.32067067, Log Avg loss: 0.29525124, Global Avg Loss: 1.36900411, Time: 0.0210 Steps: 36640, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000457, Sample Num: 7312, Cur Loss: 0.06523627, Cur Avg Loss: 0.32097158, Log Avg loss: 0.33442208, Global Avg Loss: 1.36872183, Time: 0.0210 Steps: 36650, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000467, Sample Num: 7472, Cur Loss: 0.34069201, Cur Avg Loss: 0.31995710, Log Avg loss: 0.27359533, Global Avg Loss: 1.36842310, Time: 0.0209 Steps: 36660, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000477, Sample Num: 7632, Cur Loss: 0.21050687, Cur Avg Loss: 0.32151220, Log Avg loss: 0.39413566, Global Avg Loss: 1.36815741, Time: 0.0210 Steps: 36670, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000487, Sample Num: 7792, Cur Loss: 0.12265189, Cur Avg Loss: 0.31837287, Log Avg loss: 0.16862692, Global Avg Loss: 1.36783038, Time: 0.0210 Steps: 36680, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000497, Sample Num: 7952, Cur Loss: 0.30437541, Cur Avg Loss: 0.31882505, Log Avg loss: 0.34084586, Global Avg Loss: 1.36755048, Time: 0.0210 Steps: 36690, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000507, Sample Num: 8112, Cur Loss: 0.21688019, Cur Avg Loss: 0.31786362, Log Avg loss: 0.27008048, Global Avg Loss: 1.36725144, Time: 0.0209 Steps: 36700, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000517, Sample Num: 8272, Cur Loss: 0.19352680, Cur Avg Loss: 0.31823107, Log Avg loss: 0.33686099, Global Avg Loss: 1.36697075, Time: 0.0212 Steps: 36710, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000527, Sample Num: 8432, Cur Loss: 0.09285745, Cur Avg Loss: 0.31550050, Log Avg loss: 0.17432983, Global Avg Loss: 1.36664596, Time: 0.0210 Steps: 36720, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000537, Sample Num: 8592, Cur Loss: 0.26758653, Cur Avg Loss: 0.31425805, Log Avg loss: 0.24878085, Global Avg Loss: 1.36634161, Time: 0.0210 Steps: 36730, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000547, Sample Num: 8752, Cur Loss: 0.20420752, Cur Avg Loss: 0.31260244, Log Avg loss: 0.22369668, Global Avg Loss: 1.36603061, Time: 0.0210 Steps: 36740, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000557, Sample Num: 8912, Cur Loss: 0.32607964, Cur Avg Loss: 0.31108889, Log Avg loss: 0.22829739, Global Avg Loss: 1.36572102, Time: 0.0210 Steps: 36750, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000567, Sample Num: 9072, Cur Loss: 0.25527287, Cur Avg Loss: 0.30976150, Log Avg loss: 0.23582577, Global Avg Loss: 1.36541365, Time: 0.0210 Steps: 36760, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000577, Sample Num: 9232, Cur Loss: 0.07200707, Cur Avg Loss: 0.30836459, Log Avg loss: 0.22915980, Global Avg Loss: 1.36510463, Time: 0.0210 Steps: 36770, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000587, Sample Num: 9392, Cur Loss: 0.22747838, Cur Avg Loss: 0.30894730, Log Avg loss: 0.34256981, Global Avg Loss: 1.36482662, Time: 0.0210 Steps: 36780, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000597, Sample Num: 9552, Cur Loss: 0.61863637, Cur Avg Loss: 0.30787076, Log Avg loss: 0.24467801, Global Avg Loss: 1.36452215, Time: 0.0211 Steps: 36790, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000607, Sample Num: 9712, Cur Loss: 0.14112304, Cur Avg Loss: 0.30682050, Log Avg loss: 0.24411970, Global Avg Loss: 1.36421769, Time: 0.0210 Steps: 36800, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000617, Sample Num: 9872, Cur Loss: 0.37212047, Cur Avg Loss: 0.30832677, Log Avg loss: 0.39975781, Global Avg Loss: 1.36395568, Time: 0.0210 Steps: 36810, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000627, Sample Num: 10032, Cur Loss: 0.34112832, Cur Avg Loss: 0.30784904, Log Avg loss: 0.27837277, Global Avg Loss: 1.36366084, Time: 0.0210 Steps: 36820, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000637, Sample Num: 10192, Cur Loss: 0.22126254, Cur Avg Loss: 0.30895192, Log Avg loss: 0.37810263, Global Avg Loss: 1.36339325, Time: 0.0211 Steps: 36830, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000647, Sample Num: 10352, Cur Loss: 0.47585773, Cur Avg Loss: 0.30870506, Log Avg loss: 0.29297991, Global Avg Loss: 1.36310269, Time: 0.0210 Steps: 36840, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000657, Sample Num: 10512, Cur Loss: 0.29841307, Cur Avg Loss: 0.30956272, Log Avg loss: 0.36505345, Global Avg Loss: 1.36283185, Time: 0.0210 Steps: 36850, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000667, Sample Num: 10672, Cur Loss: 0.55153131, Cur Avg Loss: 0.31254235, Log Avg loss: 0.50830371, Global Avg Loss: 1.36260002, Time: 0.0210 Steps: 36860, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000677, Sample Num: 10832, Cur Loss: 0.13693607, Cur Avg Loss: 0.31184173, Log Avg loss: 0.26511086, Global Avg Loss: 1.36230235, Time: 0.0210 Steps: 36870, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000687, Sample Num: 10992, Cur Loss: 0.19527537, Cur Avg Loss: 0.31107410, Log Avg loss: 0.25910568, Global Avg Loss: 1.36200322, Time: 0.0210 Steps: 36880, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000697, Sample Num: 11152, Cur Loss: 0.34961319, Cur Avg Loss: 0.31136095, Log Avg loss: 0.33106754, Global Avg Loss: 1.36172376, Time: 0.0211 Steps: 36890, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000707, Sample Num: 11312, Cur Loss: 0.67159325, Cur Avg Loss: 0.31192531, Log Avg loss: 0.35126073, Global Avg Loss: 1.36144992, Time: 0.0210 Steps: 36900, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000717, Sample Num: 11472, Cur Loss: 0.18428436, Cur Avg Loss: 0.31158184, Log Avg loss: 0.28729883, Global Avg Loss: 1.36115890, Time: 0.0210 Steps: 36910, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000727, Sample Num: 11632, Cur Loss: 0.33599055, Cur Avg Loss: 0.31154854, Log Avg loss: 0.30916085, Global Avg Loss: 1.36087396, Time: 0.0210 Steps: 36920, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000737, Sample Num: 11792, Cur Loss: 0.40112936, Cur Avg Loss: 0.31161987, Log Avg loss: 0.31680567, Global Avg Loss: 1.36059125, Time: 0.0210 Steps: 36930, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000747, Sample Num: 11952, Cur Loss: 0.28703898, Cur Avg Loss: 0.31070829, Log Avg loss: 0.24352497, Global Avg Loss: 1.36028885, Time: 0.0210 Steps: 36940, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000757, Sample Num: 12112, Cur Loss: 0.28603044, Cur Avg Loss: 0.31029832, Log Avg loss: 0.27967314, Global Avg Loss: 1.35999639, Time: 0.0210 Steps: 36950, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000767, Sample Num: 12272, Cur Loss: 0.31536290, Cur Avg Loss: 0.31163048, Log Avg loss: 0.41247478, Global Avg Loss: 1.35974003, Time: 0.0210 Steps: 36960, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000777, Sample Num: 12432, Cur Loss: 0.18871951, Cur Avg Loss: 0.31084019, Log Avg loss: 0.25022518, Global Avg Loss: 1.35943992, Time: 0.0210 Steps: 36970, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000787, Sample Num: 12592, Cur Loss: 0.44335431, Cur Avg Loss: 0.31164302, Log Avg loss: 0.37402299, Global Avg Loss: 1.35917344, Time: 0.0210 Steps: 36980, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000797, Sample Num: 12752, Cur Loss: 0.39746827, Cur Avg Loss: 0.31098567, Log Avg loss: 0.25925247, Global Avg Loss: 1.35887609, Time: 0.0210 Steps: 36990, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000807, Sample Num: 12912, Cur Loss: 0.37635827, Cur Avg Loss: 0.31185465, Log Avg loss: 0.38111182, Global Avg Loss: 1.35861183, Time: 0.0210 Steps: 37000, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000817, Sample Num: 13072, Cur Loss: 0.33775061, Cur Avg Loss: 0.31152490, Log Avg loss: 0.28491473, Global Avg Loss: 1.35832172, Time: 0.0210 Steps: 37010, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000827, Sample Num: 13232, Cur Loss: 0.25097829, Cur Avg Loss: 0.31085037, Log Avg loss: 0.25574128, Global Avg Loss: 1.35802388, Time: 0.0210 Steps: 37020, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000837, Sample Num: 13392, Cur Loss: 0.48527184, Cur Avg Loss: 0.31117100, Log Avg loss: 0.33768686, Global Avg Loss: 1.35774834, Time: 0.0210 Steps: 37030, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000847, Sample Num: 13552, Cur Loss: 0.41672549, Cur Avg Loss: 0.31106208, Log Avg loss: 0.30194550, Global Avg Loss: 1.35746330, Time: 0.0210 Steps: 37040, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000857, Sample Num: 13712, Cur Loss: 0.60044992, Cur Avg Loss: 0.31197778, Log Avg loss: 0.38953771, Global Avg Loss: 1.35720205, Time: 0.0210 Steps: 37050, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000867, Sample Num: 13872, Cur Loss: 0.17834947, Cur Avg Loss: 0.31248110, Log Avg loss: 0.35561517, Global Avg Loss: 1.35693179, Time: 0.0210 Steps: 37060, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000877, Sample Num: 14032, Cur Loss: 0.35343564, Cur Avg Loss: 0.31311096, Log Avg loss: 0.36771991, Global Avg Loss: 1.35666494, Time: 0.0210 Steps: 37070, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000887, Sample Num: 14192, Cur Loss: 0.20302849, Cur Avg Loss: 0.31424847, Log Avg loss: 0.41400833, Global Avg Loss: 1.35641071, Time: 0.0210 Steps: 37080, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000897, Sample Num: 14352, Cur Loss: 0.59527528, Cur Avg Loss: 0.31432875, Log Avg loss: 0.32144948, Global Avg Loss: 1.35613167, Time: 0.0210 Steps: 37090, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000907, Sample Num: 14512, Cur Loss: 0.18055990, Cur Avg Loss: 0.31340983, Log Avg loss: 0.23098266, Global Avg Loss: 1.35582840, Time: 0.0210 Steps: 37100, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000917, Sample Num: 14672, Cur Loss: 0.34122974, Cur Avg Loss: 0.31236812, Log Avg loss: 0.21788478, Global Avg Loss: 1.35552176, Time: 0.0209 Steps: 37110, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000927, Sample Num: 14832, Cur Loss: 0.16791016, Cur Avg Loss: 0.31197604, Log Avg loss: 0.27602231, Global Avg Loss: 1.35523095, Time: 0.0210 Steps: 37120, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000937, Sample Num: 14992, Cur Loss: 0.19858414, Cur Avg Loss: 0.31104370, Log Avg loss: 0.22461597, Global Avg Loss: 1.35492644, Time: 0.0210 Steps: 37130, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000947, Sample Num: 15152, Cur Loss: 0.18739802, Cur Avg Loss: 0.31092798, Log Avg loss: 0.30008498, Global Avg Loss: 1.35464243, Time: 0.0209 Steps: 37140, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000957, Sample Num: 15312, Cur Loss: 0.17417185, Cur Avg Loss: 0.31146034, Log Avg loss: 0.36187446, Global Avg Loss: 1.35437519, Time: 0.0209 Steps: 37150, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000967, Sample Num: 15472, Cur Loss: 0.42991388, Cur Avg Loss: 0.31156751, Log Avg loss: 0.32182398, Global Avg Loss: 1.35409733, Time: 0.0210 Steps: 37160, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000977, Sample Num: 15632, Cur Loss: 0.21495774, Cur Avg Loss: 0.31132377, Log Avg loss: 0.28775445, Global Avg Loss: 1.35381044, Time: 0.0210 Steps: 37170, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000987, Sample Num: 15792, Cur Loss: 0.34813917, Cur Avg Loss: 0.31190350, Log Avg loss: 0.36854306, Global Avg Loss: 1.35354545, Time: 0.0210 Steps: 37180, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000997, Sample Num: 15952, Cur Loss: 0.11487538, Cur Avg Loss: 0.31147804, Log Avg loss: 0.26948466, Global Avg Loss: 1.35325395, Time: 0.0210 Steps: 37190, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001007, Sample Num: 16112, Cur Loss: 0.11764217, Cur Avg Loss: 0.31172027, Log Avg loss: 0.33587084, Global Avg Loss: 1.35298046, Time: 0.0210 Steps: 37200, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001017, Sample Num: 16272, Cur Loss: 0.22169495, Cur Avg Loss: 0.31074863, Log Avg loss: 0.21290489, Global Avg Loss: 1.35267407, Time: 0.0210 Steps: 37210, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001027, Sample Num: 16432, Cur Loss: 0.43849719, Cur Avg Loss: 0.31078511, Log Avg loss: 0.31449488, Global Avg Loss: 1.35239514, Time: 0.0247 Steps: 37220, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001037, Sample Num: 16592, Cur Loss: 0.36343965, Cur Avg Loss: 0.30997594, Log Avg loss: 0.22687419, Global Avg Loss: 1.35209283, Time: 0.0210 Steps: 37230, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001047, Sample Num: 16752, Cur Loss: 0.17536631, Cur Avg Loss: 0.30991167, Log Avg loss: 0.30324704, Global Avg Loss: 1.35181118, Time: 0.0210 Steps: 37240, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001057, Sample Num: 16912, Cur Loss: 0.36321563, Cur Avg Loss: 0.30985216, Log Avg loss: 0.30362096, Global Avg Loss: 1.35152979, Time: 0.0210 Steps: 37250, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001067, Sample Num: 17072, Cur Loss: 0.18500572, Cur Avg Loss: 0.30948949, Log Avg loss: 0.27115597, Global Avg Loss: 1.35123983, Time: 0.0210 Steps: 37260, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001077, Sample Num: 17232, Cur Loss: 0.13412142, Cur Avg Loss: 0.30895468, Log Avg loss: 0.25188997, Global Avg Loss: 1.35094486, Time: 0.0210 Steps: 37270, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001087, Sample Num: 17392, Cur Loss: 0.46975362, Cur Avg Loss: 0.31068864, Log Avg loss: 0.49743655, Global Avg Loss: 1.35071592, Time: 0.0210 Steps: 37280, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001097, Sample Num: 17552, Cur Loss: 0.11980294, Cur Avg Loss: 0.31047240, Log Avg loss: 0.28696640, Global Avg Loss: 1.35043065, Time: 0.0210 Steps: 37290, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001107, Sample Num: 17712, Cur Loss: 0.48878342, Cur Avg Loss: 0.31164558, Log Avg loss: 0.44034363, Global Avg Loss: 1.35018666, Time: 0.0210 Steps: 37300, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001117, Sample Num: 17872, Cur Loss: 0.32555050, Cur Avg Loss: 0.31106907, Log Avg loss: 0.24725001, Global Avg Loss: 1.34989105, Time: 0.0210 Steps: 37310, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001127, Sample Num: 18032, Cur Loss: 0.35226506, Cur Avg Loss: 0.31092370, Log Avg loss: 0.29468540, Global Avg Loss: 1.34960830, Time: 0.0210 Steps: 37320, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001137, Sample Num: 18192, Cur Loss: 0.25878376, Cur Avg Loss: 0.31097483, Log Avg loss: 0.31673665, Global Avg Loss: 1.34933162, Time: 0.0210 Steps: 37330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001147, Sample Num: 18352, Cur Loss: 0.12241686, Cur Avg Loss: 0.31037864, Log Avg loss: 0.24259222, Global Avg Loss: 1.34903522, Time: 0.0210 Steps: 37340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001157, Sample Num: 18512, Cur Loss: 0.53014427, Cur Avg Loss: 0.31050862, Log Avg loss: 0.32541731, Global Avg Loss: 1.34876116, Time: 0.0210 Steps: 37350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001167, Sample Num: 18672, Cur Loss: 0.83112079, Cur Avg Loss: 0.31068657, Log Avg loss: 0.33127497, Global Avg Loss: 1.34848881, Time: 0.0210 Steps: 37360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001177, Sample Num: 18832, Cur Loss: 0.74729884, Cur Avg Loss: 0.31033542, Log Avg loss: 0.26935653, Global Avg Loss: 1.34820005, Time: 0.0210 Steps: 37370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001187, Sample Num: 18992, Cur Loss: 0.27826226, Cur Avg Loss: 0.31182719, Log Avg loss: 0.48740923, Global Avg Loss: 1.34796976, Time: 0.0210 Steps: 37380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001197, Sample Num: 19152, Cur Loss: 0.31096244, Cur Avg Loss: 0.31236324, Log Avg loss: 0.37599244, Global Avg Loss: 1.34770981, Time: 0.0210 Steps: 37390, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001207, Sample Num: 19312, Cur Loss: 0.25527847, Cur Avg Loss: 0.31181015, Log Avg loss: 0.24560530, Global Avg Loss: 1.34741513, Time: 0.0210 Steps: 37400, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001217, Sample Num: 19472, Cur Loss: 0.13301334, Cur Avg Loss: 0.31110403, Log Avg loss: 0.22587461, Global Avg Loss: 1.34711533, Time: 0.0210 Steps: 37410, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001227, Sample Num: 19632, Cur Loss: 0.39144862, Cur Avg Loss: 0.31116982, Log Avg loss: 0.31917616, Global Avg Loss: 1.34684063, Time: 0.0210 Steps: 37420, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001237, Sample Num: 19792, Cur Loss: 0.16920885, Cur Avg Loss: 0.31191479, Log Avg loss: 0.40332289, Global Avg Loss: 1.34658855, Time: 0.0210 Steps: 37430, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001247, Sample Num: 19952, Cur Loss: 0.16473415, Cur Avg Loss: 0.31180693, Log Avg loss: 0.29846463, Global Avg Loss: 1.34630860, Time: 0.0210 Steps: 37440, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001257, Sample Num: 20112, Cur Loss: 0.38693222, Cur Avg Loss: 0.31180079, Log Avg loss: 0.31103481, Global Avg Loss: 1.34603216, Time: 0.0210 Steps: 37450, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001267, Sample Num: 20272, Cur Loss: 0.19753978, Cur Avg Loss: 0.31159835, Log Avg loss: 0.28615174, Global Avg Loss: 1.34574923, Time: 0.0210 Steps: 37460, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001277, Sample Num: 20432, Cur Loss: 0.47496748, Cur Avg Loss: 0.31158591, Log Avg loss: 0.31001051, Global Avg Loss: 1.34547281, Time: 0.0210 Steps: 37470, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001287, Sample Num: 20592, Cur Loss: 0.68613929, Cur Avg Loss: 0.31199488, Log Avg loss: 0.36421970, Global Avg Loss: 1.34521100, Time: 0.0210 Steps: 37480, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001297, Sample Num: 20752, Cur Loss: 0.29948562, Cur Avg Loss: 0.31143445, Log Avg loss: 0.23930705, Global Avg Loss: 1.34491601, Time: 0.0210 Steps: 37490, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001307, Sample Num: 20912, Cur Loss: 0.23867635, Cur Avg Loss: 0.31115650, Log Avg loss: 0.27510670, Global Avg Loss: 1.34463073, Time: 0.0210 Steps: 37500, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001317, Sample Num: 21072, Cur Loss: 0.30093950, Cur Avg Loss: 0.31185612, Log Avg loss: 0.40329669, Global Avg Loss: 1.34437978, Time: 0.0210 Steps: 37510, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001327, Sample Num: 21232, Cur Loss: 0.62722552, Cur Avg Loss: 0.31240411, Log Avg loss: 0.38457478, Global Avg Loss: 1.34412396, Time: 0.0210 Steps: 37520, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001337, Sample Num: 21392, Cur Loss: 0.21793748, Cur Avg Loss: 0.31206061, Log Avg loss: 0.26647754, Global Avg Loss: 1.34383682, Time: 0.0210 Steps: 37530, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001347, Sample Num: 21552, Cur Loss: 0.14921282, Cur Avg Loss: 0.31144534, Log Avg loss: 0.22918341, Global Avg Loss: 1.34353990, Time: 0.0210 Steps: 37540, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001357, Sample Num: 21712, Cur Loss: 0.30843604, Cur Avg Loss: 0.31133092, Log Avg loss: 0.29591884, Global Avg Loss: 1.34326090, Time: 0.0210 Steps: 37550, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001367, Sample Num: 21872, Cur Loss: 0.31064686, Cur Avg Loss: 0.31131183, Log Avg loss: 0.30872179, Global Avg Loss: 1.34298547, Time: 0.0210 Steps: 37560, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001377, Sample Num: 22032, Cur Loss: 0.14912751, Cur Avg Loss: 0.31076804, Log Avg loss: 0.23643119, Global Avg Loss: 1.34269094, Time: 0.0210 Steps: 37570, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001387, Sample Num: 22192, Cur Loss: 0.30257881, Cur Avg Loss: 0.31059430, Log Avg loss: 0.28667104, Global Avg Loss: 1.34240993, Time: 0.0211 Steps: 37580, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001397, Sample Num: 22352, Cur Loss: 0.17149404, Cur Avg Loss: 0.31022672, Log Avg loss: 0.25924299, Global Avg Loss: 1.34212178, Time: 0.0210 Steps: 37590, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001407, Sample Num: 22512, Cur Loss: 0.14897656, Cur Avg Loss: 0.31020544, Log Avg loss: 0.30723216, Global Avg Loss: 1.34184654, Time: 0.0210 Steps: 37600, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001417, Sample Num: 22672, Cur Loss: 0.29912159, Cur Avg Loss: 0.31009954, Log Avg loss: 0.29519999, Global Avg Loss: 1.34156825, Time: 0.0210 Steps: 37610, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001427, Sample Num: 22832, Cur Loss: 0.13925149, Cur Avg Loss: 0.31006801, Log Avg loss: 0.30559970, Global Avg Loss: 1.34129287, Time: 0.0209 Steps: 37620, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001437, Sample Num: 22992, Cur Loss: 0.44207141, Cur Avg Loss: 0.31071312, Log Avg loss: 0.40277124, Global Avg Loss: 1.34104347, Time: 0.0210 Steps: 37630, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001447, Sample Num: 23152, Cur Loss: 0.20253244, Cur Avg Loss: 0.31008478, Log Avg loss: 0.21979150, Global Avg Loss: 1.34074558, Time: 0.0210 Steps: 37640, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001457, Sample Num: 23312, Cur Loss: 0.24311893, Cur Avg Loss: 0.31059196, Log Avg loss: 0.38398091, Global Avg Loss: 1.34049146, Time: 0.0210 Steps: 37650, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001467, Sample Num: 23472, Cur Loss: 0.24955593, Cur Avg Loss: 0.31081155, Log Avg loss: 0.34280612, Global Avg Loss: 1.34022654, Time: 0.0209 Steps: 37660, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001477, Sample Num: 23632, Cur Loss: 0.60940790, Cur Avg Loss: 0.31266245, Log Avg loss: 0.58418898, Global Avg Loss: 1.34002584, Time: 0.0210 Steps: 37670, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001487, Sample Num: 23792, Cur Loss: 0.12390934, Cur Avg Loss: 0.31310357, Log Avg loss: 0.37825684, Global Avg Loss: 1.33977059, Time: 0.0210 Steps: 37680, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001497, Sample Num: 23952, Cur Loss: 0.23466878, Cur Avg Loss: 0.31284900, Log Avg loss: 0.27499515, Global Avg Loss: 1.33948808, Time: 0.0210 Steps: 37690, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001507, Sample Num: 24112, Cur Loss: 0.15070300, Cur Avg Loss: 0.31294189, Log Avg loss: 0.32684678, Global Avg Loss: 1.33921948, Time: 0.0210 Steps: 37700, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001517, Sample Num: 24272, Cur Loss: 0.06051809, Cur Avg Loss: 0.31250401, Log Avg loss: 0.24651590, Global Avg Loss: 1.33892971, Time: 0.0210 Steps: 37710, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001527, Sample Num: 24432, Cur Loss: 0.41680780, Cur Avg Loss: 0.31248075, Log Avg loss: 0.30895258, Global Avg Loss: 1.33865665, Time: 0.0210 Steps: 37720, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001537, Sample Num: 24592, Cur Loss: 0.30215934, Cur Avg Loss: 0.31257187, Log Avg loss: 0.32648564, Global Avg Loss: 1.33838839, Time: 0.0247 Steps: 37730, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001547, Sample Num: 24752, Cur Loss: 0.29089206, Cur Avg Loss: 0.31272974, Log Avg loss: 0.33699462, Global Avg Loss: 1.33812305, Time: 0.0210 Steps: 37740, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001557, Sample Num: 24912, Cur Loss: 0.31405604, Cur Avg Loss: 0.31359700, Log Avg loss: 0.44776276, Global Avg Loss: 1.33788719, Time: 0.0210 Steps: 37750, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001567, Sample Num: 25072, Cur Loss: 0.23450899, Cur Avg Loss: 0.31346635, Log Avg loss: 0.29312304, Global Avg Loss: 1.33761051, Time: 0.0210 Steps: 37760, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001577, Sample Num: 25232, Cur Loss: 0.41351455, Cur Avg Loss: 0.31379693, Log Avg loss: 0.36559928, Global Avg Loss: 1.33735315, Time: 0.0210 Steps: 37770, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001587, Sample Num: 25392, Cur Loss: 0.17200825, Cur Avg Loss: 0.31370402, Log Avg loss: 0.29905225, Global Avg Loss: 1.33707833, Time: 0.0209 Steps: 37780, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001597, Sample Num: 25552, Cur Loss: 0.31705099, Cur Avg Loss: 0.31302757, Log Avg loss: 0.20567499, Global Avg Loss: 1.33677893, Time: 0.0210 Steps: 37790, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001607, Sample Num: 25712, Cur Loss: 0.47761050, Cur Avg Loss: 0.31292773, Log Avg loss: 0.29698344, Global Avg Loss: 1.33650386, Time: 0.0210 Steps: 37800, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001617, Sample Num: 25872, Cur Loss: 0.07294662, Cur Avg Loss: 0.31244078, Log Avg loss: 0.23418757, Global Avg Loss: 1.33621232, Time: 0.0212 Steps: 37810, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001627, Sample Num: 26032, Cur Loss: 0.22896206, Cur Avg Loss: 0.31175133, Log Avg loss: 0.20026779, Global Avg Loss: 1.33591196, Time: 0.0209 Steps: 37820, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001637, Sample Num: 26192, Cur Loss: 0.32174352, Cur Avg Loss: 0.31197724, Log Avg loss: 0.34873178, Global Avg Loss: 1.33565101, Time: 0.0209 Steps: 37830, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001647, Sample Num: 26352, Cur Loss: 0.46185702, Cur Avg Loss: 0.31145653, Log Avg loss: 0.22621633, Global Avg Loss: 1.33535782, Time: 0.0210 Steps: 37840, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001657, Sample Num: 26512, Cur Loss: 0.07531068, Cur Avg Loss: 0.31096679, Log Avg loss: 0.23030713, Global Avg Loss: 1.33506586, Time: 0.0209 Steps: 37850, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001667, Sample Num: 26672, Cur Loss: 0.11946835, Cur Avg Loss: 0.31045713, Log Avg loss: 0.22600579, Global Avg Loss: 1.33477292, Time: 0.0209 Steps: 37860, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001677, Sample Num: 26832, Cur Loss: 0.06460225, Cur Avg Loss: 0.30984933, Log Avg loss: 0.20852893, Global Avg Loss: 1.33447553, Time: 0.0209 Steps: 37870, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001687, Sample Num: 26992, Cur Loss: 0.21208268, Cur Avg Loss: 0.30950353, Log Avg loss: 0.25151328, Global Avg Loss: 1.33418963, Time: 0.0209 Steps: 37880, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001697, Sample Num: 27152, Cur Loss: 0.32176220, Cur Avg Loss: 0.30917513, Log Avg loss: 0.25377520, Global Avg Loss: 1.33390449, Time: 0.0208 Steps: 37890, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001707, Sample Num: 27312, Cur Loss: 0.16000488, Cur Avg Loss: 0.30907260, Log Avg loss: 0.29167193, Global Avg Loss: 1.33362949, Time: 0.0214 Steps: 37900, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001717, Sample Num: 27472, Cur Loss: 0.09102480, Cur Avg Loss: 0.30858211, Log Avg loss: 0.22485671, Global Avg Loss: 1.33333702, Time: 0.0216 Steps: 37910, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001727, Sample Num: 27632, Cur Loss: 0.45049044, Cur Avg Loss: 0.30879639, Log Avg loss: 0.34558700, Global Avg Loss: 1.33307654, Time: 0.0213 Steps: 37920, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001737, Sample Num: 27792, Cur Loss: 0.25529057, Cur Avg Loss: 0.30886817, Log Avg loss: 0.32126466, Global Avg Loss: 1.33280978, Time: 0.0210 Steps: 37930, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001747, Sample Num: 27952, Cur Loss: 0.19364972, Cur Avg Loss: 0.30846395, Log Avg loss: 0.23825128, Global Avg Loss: 1.33252128, Time: 0.0209 Steps: 37940, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001757, Sample Num: 28112, Cur Loss: 0.20985033, Cur Avg Loss: 0.30794932, Log Avg loss: 0.21804340, Global Avg Loss: 1.33222761, Time: 0.0209 Steps: 37950, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001767, Sample Num: 28272, Cur Loss: 0.54940569, Cur Avg Loss: 0.30805019, Log Avg loss: 0.32577243, Global Avg Loss: 1.33196248, Time: 0.0209 Steps: 37960, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001777, Sample Num: 28432, Cur Loss: 0.16095591, Cur Avg Loss: 0.30792630, Log Avg loss: 0.28603484, Global Avg Loss: 1.33168701, Time: 0.0209 Steps: 37970, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001787, Sample Num: 28592, Cur Loss: 0.28783217, Cur Avg Loss: 0.30792799, Log Avg loss: 0.30822932, Global Avg Loss: 1.33141754, Time: 0.0209 Steps: 37980, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001797, Sample Num: 28752, Cur Loss: 0.40371904, Cur Avg Loss: 0.30840850, Log Avg loss: 0.39427537, Global Avg Loss: 1.33117086, Time: 0.0210 Steps: 37990, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001807, Sample Num: 28912, Cur Loss: 0.48902062, Cur Avg Loss: 0.30897764, Log Avg loss: 0.41125158, Global Avg Loss: 1.33092878, Time: 0.0209 Steps: 38000, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001817, Sample Num: 29072, Cur Loss: 0.40608138, Cur Avg Loss: 0.30939998, Log Avg loss: 0.38571835, Global Avg Loss: 1.33068010, Time: 0.0209 Steps: 38010, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001827, Sample Num: 29232, Cur Loss: 0.11366931, Cur Avg Loss: 0.30931477, Log Avg loss: 0.29383200, Global Avg Loss: 1.33040739, Time: 0.0209 Steps: 38020, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001837, Sample Num: 29392, Cur Loss: 0.38320696, Cur Avg Loss: 0.30968278, Log Avg loss: 0.37691695, Global Avg Loss: 1.33015667, Time: 0.0209 Steps: 38030, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001847, Sample Num: 29552, Cur Loss: 0.10886736, Cur Avg Loss: 0.30979893, Log Avg loss: 0.33113525, Global Avg Loss: 1.32989405, Time: 0.0209 Steps: 38040, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001857, Sample Num: 29712, Cur Loss: 0.57241774, Cur Avg Loss: 0.30974127, Log Avg loss: 0.29909185, Global Avg Loss: 1.32962314, Time: 0.0209 Steps: 38050, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001867, Sample Num: 29872, Cur Loss: 0.29911348, Cur Avg Loss: 0.31009092, Log Avg loss: 0.37502190, Global Avg Loss: 1.32937232, Time: 0.0209 Steps: 38060, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001877, Sample Num: 30032, Cur Loss: 0.33276606, Cur Avg Loss: 0.31067824, Log Avg loss: 0.42033061, Global Avg Loss: 1.32913354, Time: 0.0209 Steps: 38070, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001887, Sample Num: 30192, Cur Loss: 0.48364007, Cur Avg Loss: 0.31122898, Log Avg loss: 0.41460238, Global Avg Loss: 1.32889338, Time: 0.0209 Steps: 38080, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001897, Sample Num: 30352, Cur Loss: 0.25726300, Cur Avg Loss: 0.31144075, Log Avg loss: 0.35140264, Global Avg Loss: 1.32863676, Time: 0.0210 Steps: 38090, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001907, Sample Num: 30512, Cur Loss: 0.24995472, Cur Avg Loss: 0.31122006, Log Avg loss: 0.26935445, Global Avg Loss: 1.32835873, Time: 0.0209 Steps: 38100, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001917, Sample Num: 30672, Cur Loss: 0.20221400, Cur Avg Loss: 0.31084904, Log Avg loss: 0.24009556, Global Avg Loss: 1.32807317, Time: 0.0209 Steps: 38110, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001927, Sample Num: 30832, Cur Loss: 0.27521107, Cur Avg Loss: 0.31105529, Log Avg loss: 0.35059359, Global Avg Loss: 1.32781675, Time: 0.0209 Steps: 38120, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001937, Sample Num: 30992, Cur Loss: 0.37468684, Cur Avg Loss: 0.31134829, Log Avg loss: 0.36781019, Global Avg Loss: 1.32756498, Time: 0.0209 Steps: 38130, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001947, Sample Num: 31152, Cur Loss: 0.11651429, Cur Avg Loss: 0.31159443, Log Avg loss: 0.35927113, Global Avg Loss: 1.32731110, Time: 0.0209 Steps: 38140, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001957, Sample Num: 31312, Cur Loss: 0.34367263, Cur Avg Loss: 0.31188616, Log Avg loss: 0.36868526, Global Avg Loss: 1.32705982, Time: 0.0209 Steps: 38150, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001967, Sample Num: 31472, Cur Loss: 0.39322454, Cur Avg Loss: 0.31217638, Log Avg loss: 0.36897261, Global Avg Loss: 1.32680875, Time: 0.0209 Steps: 38160, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001977, Sample Num: 31632, Cur Loss: 0.28916880, Cur Avg Loss: 0.31192947, Log Avg loss: 0.26336190, Global Avg Loss: 1.32653014, Time: 0.0209 Steps: 38170, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001987, Sample Num: 31792, Cur Loss: 0.36296976, Cur Avg Loss: 0.31193899, Log Avg loss: 0.31382193, Global Avg Loss: 1.32626489, Time: 0.0209 Steps: 38180, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001997, Sample Num: 31952, Cur Loss: 0.38156664, Cur Avg Loss: 0.31195822, Log Avg loss: 0.31577889, Global Avg Loss: 1.32600030, Time: 0.0209 Steps: 38190, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002007, Sample Num: 32112, Cur Loss: 0.27688032, Cur Avg Loss: 0.31171204, Log Avg loss: 0.26254981, Global Avg Loss: 1.32572191, Time: 0.0208 Steps: 38200, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002017, Sample Num: 32272, Cur Loss: 0.21764886, Cur Avg Loss: 0.31152843, Log Avg loss: 0.27467783, Global Avg Loss: 1.32544684, Time: 0.0209 Steps: 38210, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002027, Sample Num: 32432, Cur Loss: 0.29820389, Cur Avg Loss: 0.31171688, Log Avg loss: 0.34972695, Global Avg Loss: 1.32519155, Time: 0.0209 Steps: 38220, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002037, Sample Num: 32592, Cur Loss: 0.36525196, Cur Avg Loss: 0.31173628, Log Avg loss: 0.31566861, Global Avg Loss: 1.32492748, Time: 0.0209 Steps: 38230, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002047, Sample Num: 32752, Cur Loss: 0.22607343, Cur Avg Loss: 0.31150038, Log Avg loss: 0.26344784, Global Avg Loss: 1.32464990, Time: 0.0209 Steps: 38240, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002057, Sample Num: 32912, Cur Loss: 0.14315207, Cur Avg Loss: 0.31098111, Log Avg loss: 0.20468785, Global Avg Loss: 1.32435710, Time: 0.0209 Steps: 38250, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002067, Sample Num: 33072, Cur Loss: 0.23736587, Cur Avg Loss: 0.31106068, Log Avg loss: 0.32742657, Global Avg Loss: 1.32409653, Time: 0.0209 Steps: 38260, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002077, Sample Num: 33232, Cur Loss: 0.34529877, Cur Avg Loss: 0.31060127, Log Avg loss: 0.21564229, Global Avg Loss: 1.32380689, Time: 0.0209 Steps: 38270, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002087, Sample Num: 33392, Cur Loss: 0.45116338, Cur Avg Loss: 0.31072634, Log Avg loss: 0.33670236, Global Avg Loss: 1.32354903, Time: 0.0209 Steps: 38280, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002097, Sample Num: 33552, Cur Loss: 0.12744623, Cur Avg Loss: 0.31083694, Log Avg loss: 0.33391985, Global Avg Loss: 1.32329057, Time: 0.0209 Steps: 38290, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002107, Sample Num: 33712, Cur Loss: 1.33449924, Cur Avg Loss: 0.31090704, Log Avg loss: 0.32560678, Global Avg Loss: 1.32303008, Time: 0.0209 Steps: 38300, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002117, Sample Num: 33872, Cur Loss: 0.43945116, Cur Avg Loss: 0.31126944, Log Avg loss: 0.38762685, Global Avg Loss: 1.32278591, Time: 0.0208 Steps: 38310, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002127, Sample Num: 34032, Cur Loss: 0.45533407, Cur Avg Loss: 0.31127976, Log Avg loss: 0.31346411, Global Avg Loss: 1.32252252, Time: 0.0209 Steps: 38320, Updated lr: 0.000065 ***** Running evaluation checkpoint-38322 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-38322 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.784706, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.557964, "eval_total_loss": 392.248442, "eval_mae": 0.637641, "eval_mse": 0.55801, "eval_r2": 0.645292, "eval_sp_statistic": 0.857998, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.894385, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.514373, "test_total_loss": 258.215388, "test_mae": 0.492989, "test_mse": 0.514541, "test_r2": 0.667911, "test_sp_statistic": 0.780794, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.828225, "test_ps_pvalue": 0.0, "lr": 6.460692271218587e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.322470179724894, "train_cur_epoch_loss": 662.7313245162368, "train_cur_epoch_avg_loss": 0.31128761132749494, "train_cur_epoch_time": 44.784706115722656, "train_cur_epoch_avg_time": 0.021035559471922337, "epoch": 18, "step": 38322} ################################################## Training, Epoch: 0019, Batch: 000008, Sample Num: 128, Cur Loss: 0.24590601, Cur Avg Loss: 0.22386815, Log Avg loss: 0.24302290, Global Avg Loss: 1.32224089, Time: 0.0210 Steps: 38330, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000018, Sample Num: 288, Cur Loss: 0.10942980, Cur Avg Loss: 0.31431016, Log Avg loss: 0.38666376, Global Avg Loss: 1.32199687, Time: 0.0208 Steps: 38340, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000028, Sample Num: 448, Cur Loss: 0.25303411, Cur Avg Loss: 0.34721559, Log Avg loss: 0.40644539, Global Avg Loss: 1.32175813, Time: 0.0209 Steps: 38350, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000038, Sample Num: 608, Cur Loss: 0.45303279, Cur Avg Loss: 0.33848488, Log Avg loss: 0.31403888, Global Avg Loss: 1.32149543, Time: 0.0208 Steps: 38360, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000048, Sample Num: 768, Cur Loss: 0.14850305, Cur Avg Loss: 0.33363193, Log Avg loss: 0.31519073, Global Avg Loss: 1.32123317, Time: 0.0208 Steps: 38370, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000058, Sample Num: 928, Cur Loss: 0.62368429, Cur Avg Loss: 0.32077324, Log Avg loss: 0.25905153, Global Avg Loss: 1.32095641, Time: 0.0209 Steps: 38380, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000068, Sample Num: 1088, Cur Loss: 0.70562375, Cur Avg Loss: 0.31294940, Log Avg loss: 0.26757111, Global Avg Loss: 1.32068202, Time: 0.0209 Steps: 38390, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000078, Sample Num: 1248, Cur Loss: 0.10448934, Cur Avg Loss: 0.30187135, Log Avg loss: 0.22654059, Global Avg Loss: 1.32039709, Time: 0.0208 Steps: 38400, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000088, Sample Num: 1408, Cur Loss: 0.28719091, Cur Avg Loss: 0.30069299, Log Avg loss: 0.29150182, Global Avg Loss: 1.32012922, Time: 0.0208 Steps: 38410, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000098, Sample Num: 1568, Cur Loss: 0.21169490, Cur Avg Loss: 0.30150586, Log Avg loss: 0.30865915, Global Avg Loss: 1.31986595, Time: 0.0208 Steps: 38420, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000108, Sample Num: 1728, Cur Loss: 0.30695403, Cur Avg Loss: 0.30238629, Log Avg loss: 0.31101448, Global Avg Loss: 1.31960343, Time: 0.0209 Steps: 38430, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000118, Sample Num: 1888, Cur Loss: 0.49834684, Cur Avg Loss: 0.30532541, Log Avg loss: 0.33706795, Global Avg Loss: 1.31934783, Time: 0.0208 Steps: 38440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000128, Sample Num: 2048, Cur Loss: 0.17890146, Cur Avg Loss: 0.30787053, Log Avg loss: 0.33790287, Global Avg Loss: 1.31909258, Time: 0.0208 Steps: 38450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000138, Sample Num: 2208, Cur Loss: 0.05010924, Cur Avg Loss: 0.30392194, Log Avg loss: 0.25338000, Global Avg Loss: 1.31881548, Time: 0.0208 Steps: 38460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000148, Sample Num: 2368, Cur Loss: 0.19929880, Cur Avg Loss: 0.30512304, Log Avg loss: 0.32169830, Global Avg Loss: 1.31855629, Time: 0.0209 Steps: 38470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000158, Sample Num: 2528, Cur Loss: 0.53278357, Cur Avg Loss: 0.30570199, Log Avg loss: 0.31427040, Global Avg Loss: 1.31829530, Time: 0.0208 Steps: 38480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000168, Sample Num: 2688, Cur Loss: 0.08020053, Cur Avg Loss: 0.30196324, Log Avg loss: 0.24289095, Global Avg Loss: 1.31801590, Time: 0.0209 Steps: 38490, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000178, Sample Num: 2848, Cur Loss: 0.07187561, Cur Avg Loss: 0.30075756, Log Avg loss: 0.28050219, Global Avg Loss: 1.31774642, Time: 0.0208 Steps: 38500, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000188, Sample Num: 3008, Cur Loss: 0.08540386, Cur Avg Loss: 0.29559575, Log Avg loss: 0.20371545, Global Avg Loss: 1.31745713, Time: 0.0209 Steps: 38510, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000198, Sample Num: 3168, Cur Loss: 0.12215937, Cur Avg Loss: 0.29144110, Log Avg loss: 0.21333376, Global Avg Loss: 1.31717050, Time: 0.0209 Steps: 38520, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000208, Sample Num: 3328, Cur Loss: 0.25394151, Cur Avg Loss: 0.28992973, Log Avg loss: 0.26000458, Global Avg Loss: 1.31689612, Time: 0.0208 Steps: 38530, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000218, Sample Num: 3488, Cur Loss: 0.33035976, Cur Avg Loss: 0.29049951, Log Avg loss: 0.30235084, Global Avg Loss: 1.31663288, Time: 0.0208 Steps: 38540, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000228, Sample Num: 3648, Cur Loss: 0.10596849, Cur Avg Loss: 0.28989115, Log Avg loss: 0.27662901, Global Avg Loss: 1.31636310, Time: 0.0209 Steps: 38550, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000238, Sample Num: 3808, Cur Loss: 0.31089142, Cur Avg Loss: 0.28919000, Log Avg loss: 0.27320383, Global Avg Loss: 1.31609257, Time: 0.0208 Steps: 38560, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000248, Sample Num: 3968, Cur Loss: 0.24958459, Cur Avg Loss: 0.28726864, Log Avg loss: 0.24154030, Global Avg Loss: 1.31581397, Time: 0.0208 Steps: 38570, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000258, Sample Num: 4128, Cur Loss: 0.24326065, Cur Avg Loss: 0.29075332, Log Avg loss: 0.37717332, Global Avg Loss: 1.31557067, Time: 0.0247 Steps: 38580, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000268, Sample Num: 4288, Cur Loss: 0.26769722, Cur Avg Loss: 0.29149309, Log Avg loss: 0.31057902, Global Avg Loss: 1.31531025, Time: 0.0209 Steps: 38590, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000278, Sample Num: 4448, Cur Loss: 0.23675996, Cur Avg Loss: 0.29156945, Log Avg loss: 0.29361608, Global Avg Loss: 1.31504556, Time: 0.0209 Steps: 38600, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000288, Sample Num: 4608, Cur Loss: 0.15902960, Cur Avg Loss: 0.28966960, Log Avg loss: 0.23685371, Global Avg Loss: 1.31476631, Time: 0.0209 Steps: 38610, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000298, Sample Num: 4768, Cur Loss: 0.06675130, Cur Avg Loss: 0.29102118, Log Avg loss: 0.32994656, Global Avg Loss: 1.31451130, Time: 0.0208 Steps: 38620, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000308, Sample Num: 4928, Cur Loss: 0.23296463, Cur Avg Loss: 0.28932138, Log Avg loss: 0.23866756, Global Avg Loss: 1.31423280, Time: 0.0208 Steps: 38630, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000318, Sample Num: 5088, Cur Loss: 0.09540895, Cur Avg Loss: 0.29129513, Log Avg loss: 0.35208646, Global Avg Loss: 1.31398380, Time: 0.0208 Steps: 38640, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000328, Sample Num: 5248, Cur Loss: 0.32166275, Cur Avg Loss: 0.28882430, Log Avg loss: 0.21025196, Global Avg Loss: 1.31369823, Time: 0.0209 Steps: 38650, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000338, Sample Num: 5408, Cur Loss: 0.13406326, Cur Avg Loss: 0.28647129, Log Avg loss: 0.20929267, Global Avg Loss: 1.31341256, Time: 0.0209 Steps: 38660, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000348, Sample Num: 5568, Cur Loss: 0.62058741, Cur Avg Loss: 0.28621730, Log Avg loss: 0.27763232, Global Avg Loss: 1.31314471, Time: 0.0209 Steps: 38670, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000358, Sample Num: 5728, Cur Loss: 0.47128659, Cur Avg Loss: 0.28902268, Log Avg loss: 0.38664995, Global Avg Loss: 1.31290518, Time: 0.0209 Steps: 38680, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000368, Sample Num: 5888, Cur Loss: 0.02676905, Cur Avg Loss: 0.28857025, Log Avg loss: 0.27237325, Global Avg Loss: 1.31263624, Time: 0.0209 Steps: 38690, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000378, Sample Num: 6048, Cur Loss: 0.13016711, Cur Avg Loss: 0.28789252, Log Avg loss: 0.26295196, Global Avg Loss: 1.31236500, Time: 0.0209 Steps: 38700, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000388, Sample Num: 6208, Cur Loss: 0.19176511, Cur Avg Loss: 0.28704611, Log Avg loss: 0.25505203, Global Avg Loss: 1.31209187, Time: 0.0208 Steps: 38710, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000398, Sample Num: 6368, Cur Loss: 0.27667218, Cur Avg Loss: 0.28675092, Log Avg loss: 0.27529749, Global Avg Loss: 1.31182410, Time: 0.0208 Steps: 38720, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000408, Sample Num: 6528, Cur Loss: 0.55032396, Cur Avg Loss: 0.28607989, Log Avg loss: 0.25937285, Global Avg Loss: 1.31155236, Time: 0.0209 Steps: 38730, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000418, Sample Num: 6688, Cur Loss: 0.23681739, Cur Avg Loss: 0.28583422, Log Avg loss: 0.27581070, Global Avg Loss: 1.31128500, Time: 0.0209 Steps: 38740, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000428, Sample Num: 6848, Cur Loss: 0.38095701, Cur Avg Loss: 0.28451093, Log Avg loss: 0.22919769, Global Avg Loss: 1.31100575, Time: 0.0209 Steps: 38750, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000438, Sample Num: 7008, Cur Loss: 0.08461526, Cur Avg Loss: 0.28581611, Log Avg loss: 0.34167759, Global Avg Loss: 1.31075567, Time: 0.0208 Steps: 38760, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000448, Sample Num: 7168, Cur Loss: 0.32003969, Cur Avg Loss: 0.28531443, Log Avg loss: 0.26334081, Global Avg Loss: 1.31048551, Time: 0.0208 Steps: 38770, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000458, Sample Num: 7328, Cur Loss: 0.22053590, Cur Avg Loss: 0.28538935, Log Avg loss: 0.28874599, Global Avg Loss: 1.31022204, Time: 0.0210 Steps: 38780, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000468, Sample Num: 7488, Cur Loss: 0.21417496, Cur Avg Loss: 0.28846113, Log Avg loss: 0.42914860, Global Avg Loss: 1.30999490, Time: 0.0209 Steps: 38790, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000478, Sample Num: 7648, Cur Loss: 0.09522630, Cur Avg Loss: 0.28827773, Log Avg loss: 0.27969448, Global Avg Loss: 1.30972936, Time: 0.0209 Steps: 38800, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000488, Sample Num: 7808, Cur Loss: 0.12689598, Cur Avg Loss: 0.28809630, Log Avg loss: 0.27942413, Global Avg Loss: 1.30946388, Time: 0.0208 Steps: 38810, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000498, Sample Num: 7968, Cur Loss: 0.46991694, Cur Avg Loss: 0.28954037, Log Avg loss: 0.36001086, Global Avg Loss: 1.30921930, Time: 0.0208 Steps: 38820, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000508, Sample Num: 8128, Cur Loss: 0.41456079, Cur Avg Loss: 0.28965129, Log Avg loss: 0.29517505, Global Avg Loss: 1.30895815, Time: 0.0209 Steps: 38830, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000518, Sample Num: 8288, Cur Loss: 0.26678154, Cur Avg Loss: 0.28924762, Log Avg loss: 0.26874142, Global Avg Loss: 1.30869033, Time: 0.0210 Steps: 38840, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000528, Sample Num: 8448, Cur Loss: 0.14783245, Cur Avg Loss: 0.28918358, Log Avg loss: 0.28586635, Global Avg Loss: 1.30842706, Time: 0.0209 Steps: 38850, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000538, Sample Num: 8608, Cur Loss: 0.23171230, Cur Avg Loss: 0.28831757, Log Avg loss: 0.24259220, Global Avg Loss: 1.30815278, Time: 0.0209 Steps: 38860, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000548, Sample Num: 8768, Cur Loss: 0.29921523, Cur Avg Loss: 0.28810712, Log Avg loss: 0.27678448, Global Avg Loss: 1.30788744, Time: 0.0209 Steps: 38870, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000558, Sample Num: 8928, Cur Loss: 0.17036235, Cur Avg Loss: 0.28745390, Log Avg loss: 0.25165775, Global Avg Loss: 1.30761578, Time: 0.0209 Steps: 38880, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000568, Sample Num: 9088, Cur Loss: 0.24782920, Cur Avg Loss: 0.28612991, Log Avg loss: 0.21225111, Global Avg Loss: 1.30733412, Time: 0.0209 Steps: 38890, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000578, Sample Num: 9248, Cur Loss: 0.24822605, Cur Avg Loss: 0.28566784, Log Avg loss: 0.25942223, Global Avg Loss: 1.30706474, Time: 0.0210 Steps: 38900, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000588, Sample Num: 9408, Cur Loss: 0.25987846, Cur Avg Loss: 0.28478569, Log Avg loss: 0.23379737, Global Avg Loss: 1.30678890, Time: 0.0209 Steps: 38910, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000598, Sample Num: 9568, Cur Loss: 0.17238033, Cur Avg Loss: 0.28361865, Log Avg loss: 0.21499661, Global Avg Loss: 1.30650838, Time: 0.0209 Steps: 38920, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000608, Sample Num: 9728, Cur Loss: 0.19003038, Cur Avg Loss: 0.28450748, Log Avg loss: 0.33765972, Global Avg Loss: 1.30625951, Time: 0.0209 Steps: 38930, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000618, Sample Num: 9888, Cur Loss: 0.04908440, Cur Avg Loss: 0.28365415, Log Avg loss: 0.23177162, Global Avg Loss: 1.30598358, Time: 0.0209 Steps: 38940, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000628, Sample Num: 10048, Cur Loss: 0.30902064, Cur Avg Loss: 0.28336641, Log Avg loss: 0.26558424, Global Avg Loss: 1.30571647, Time: 0.0209 Steps: 38950, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000638, Sample Num: 10208, Cur Loss: 0.21346901, Cur Avg Loss: 0.28408695, Log Avg loss: 0.32933697, Global Avg Loss: 1.30546585, Time: 0.0209 Steps: 38960, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000648, Sample Num: 10368, Cur Loss: 0.12509406, Cur Avg Loss: 0.28410622, Log Avg loss: 0.28533576, Global Avg Loss: 1.30520408, Time: 0.0209 Steps: 38970, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000658, Sample Num: 10528, Cur Loss: 0.38524976, Cur Avg Loss: 0.28369935, Log Avg loss: 0.25733384, Global Avg Loss: 1.30493526, Time: 0.0209 Steps: 38980, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000668, Sample Num: 10688, Cur Loss: 0.13049054, Cur Avg Loss: 0.28349113, Log Avg loss: 0.26979014, Global Avg Loss: 1.30466977, Time: 0.0209 Steps: 38990, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000678, Sample Num: 10848, Cur Loss: 0.29159367, Cur Avg Loss: 0.28238431, Log Avg loss: 0.20844868, Global Avg Loss: 1.30438869, Time: 0.0209 Steps: 39000, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000688, Sample Num: 11008, Cur Loss: 0.38980550, Cur Avg Loss: 0.28281364, Log Avg loss: 0.31192233, Global Avg Loss: 1.30413427, Time: 0.0209 Steps: 39010, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000698, Sample Num: 11168, Cur Loss: 0.06226075, Cur Avg Loss: 0.28401471, Log Avg loss: 0.36664865, Global Avg Loss: 1.30389402, Time: 0.0209 Steps: 39020, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000708, Sample Num: 11328, Cur Loss: 0.46016693, Cur Avg Loss: 0.28420779, Log Avg loss: 0.29768471, Global Avg Loss: 1.30363621, Time: 0.0209 Steps: 39030, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000718, Sample Num: 11488, Cur Loss: 0.12047105, Cur Avg Loss: 0.28369087, Log Avg loss: 0.24709255, Global Avg Loss: 1.30336558, Time: 0.0209 Steps: 39040, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000728, Sample Num: 11648, Cur Loss: 0.46887213, Cur Avg Loss: 0.28369027, Log Avg loss: 0.28364745, Global Avg Loss: 1.30310445, Time: 0.0209 Steps: 39050, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000738, Sample Num: 11808, Cur Loss: 0.56162065, Cur Avg Loss: 0.28407285, Log Avg loss: 0.31192495, Global Avg Loss: 1.30285069, Time: 0.0210 Steps: 39060, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000748, Sample Num: 11968, Cur Loss: 0.10743015, Cur Avg Loss: 0.28435222, Log Avg loss: 0.30496964, Global Avg Loss: 1.30259528, Time: 0.0209 Steps: 39070, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000758, Sample Num: 12128, Cur Loss: 0.20714790, Cur Avg Loss: 0.28682653, Log Avg loss: 0.47190504, Global Avg Loss: 1.30238272, Time: 0.0209 Steps: 39080, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000768, Sample Num: 12288, Cur Loss: 0.33170074, Cur Avg Loss: 0.28721499, Log Avg loss: 0.31666024, Global Avg Loss: 1.30213055, Time: 0.0254 Steps: 39090, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000778, Sample Num: 12448, Cur Loss: 0.20960802, Cur Avg Loss: 0.28610951, Log Avg loss: 0.20120831, Global Avg Loss: 1.30184899, Time: 0.0209 Steps: 39100, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000788, Sample Num: 12608, Cur Loss: 0.32731402, Cur Avg Loss: 0.28678328, Log Avg loss: 0.33920237, Global Avg Loss: 1.30160285, Time: 0.0209 Steps: 39110, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000798, Sample Num: 12768, Cur Loss: 0.22560960, Cur Avg Loss: 0.28617701, Log Avg loss: 0.23840309, Global Avg Loss: 1.30133107, Time: 0.0209 Steps: 39120, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000808, Sample Num: 12928, Cur Loss: 0.46153840, Cur Avg Loss: 0.28695951, Log Avg loss: 0.34940287, Global Avg Loss: 1.30108780, Time: 0.0208 Steps: 39130, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000818, Sample Num: 13088, Cur Loss: 0.16832939, Cur Avg Loss: 0.28660940, Log Avg loss: 0.25832077, Global Avg Loss: 1.30082138, Time: 0.0208 Steps: 39140, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000828, Sample Num: 13248, Cur Loss: 0.44999945, Cur Avg Loss: 0.28774036, Log Avg loss: 0.38025299, Global Avg Loss: 1.30058624, Time: 0.0208 Steps: 39150, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000838, Sample Num: 13408, Cur Loss: 0.09260732, Cur Avg Loss: 0.28784972, Log Avg loss: 0.29690478, Global Avg Loss: 1.30032994, Time: 0.0209 Steps: 39160, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000848, Sample Num: 13568, Cur Loss: 0.13788185, Cur Avg Loss: 0.28947876, Log Avg loss: 0.42599232, Global Avg Loss: 1.30010672, Time: 0.0209 Steps: 39170, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000858, Sample Num: 13728, Cur Loss: 0.47388858, Cur Avg Loss: 0.28926649, Log Avg loss: 0.27126580, Global Avg Loss: 1.29984413, Time: 0.0208 Steps: 39180, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000868, Sample Num: 13888, Cur Loss: 0.24730554, Cur Avg Loss: 0.28962311, Log Avg loss: 0.32022136, Global Avg Loss: 1.29959416, Time: 0.0209 Steps: 39190, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000878, Sample Num: 14048, Cur Loss: 0.23806088, Cur Avg Loss: 0.28937760, Log Avg loss: 0.26806710, Global Avg Loss: 1.29933101, Time: 0.0209 Steps: 39200, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000888, Sample Num: 14208, Cur Loss: 0.52323413, Cur Avg Loss: 0.28919912, Log Avg loss: 0.27352848, Global Avg Loss: 1.29906940, Time: 0.0209 Steps: 39210, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000898, Sample Num: 14368, Cur Loss: 0.10122560, Cur Avg Loss: 0.28922773, Log Avg loss: 0.29176864, Global Avg Loss: 1.29881256, Time: 0.0209 Steps: 39220, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000908, Sample Num: 14528, Cur Loss: 0.26282084, Cur Avg Loss: 0.28870489, Log Avg loss: 0.24175384, Global Avg Loss: 1.29854311, Time: 0.0208 Steps: 39230, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000918, Sample Num: 14688, Cur Loss: 0.26021776, Cur Avg Loss: 0.28919453, Log Avg loss: 0.33365333, Global Avg Loss: 1.29829722, Time: 0.0209 Steps: 39240, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000928, Sample Num: 14848, Cur Loss: 0.35728660, Cur Avg Loss: 0.28911869, Log Avg loss: 0.28215657, Global Avg Loss: 1.29803833, Time: 0.0209 Steps: 39250, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000938, Sample Num: 15008, Cur Loss: 0.17444256, Cur Avg Loss: 0.28940580, Log Avg loss: 0.31605005, Global Avg Loss: 1.29778820, Time: 0.0209 Steps: 39260, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000948, Sample Num: 15168, Cur Loss: 0.36744606, Cur Avg Loss: 0.28952914, Log Avg loss: 0.30109818, Global Avg Loss: 1.29753440, Time: 0.0209 Steps: 39270, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000958, Sample Num: 15328, Cur Loss: 0.16172782, Cur Avg Loss: 0.28961823, Log Avg loss: 0.29806398, Global Avg Loss: 1.29727995, Time: 0.0209 Steps: 39280, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000968, Sample Num: 15488, Cur Loss: 0.30992937, Cur Avg Loss: 0.28937677, Log Avg loss: 0.26624522, Global Avg Loss: 1.29701753, Time: 0.0209 Steps: 39290, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000978, Sample Num: 15648, Cur Loss: 0.12132890, Cur Avg Loss: 0.29020840, Log Avg loss: 0.37070965, Global Avg Loss: 1.29678183, Time: 0.0208 Steps: 39300, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000988, Sample Num: 15808, Cur Loss: 0.39052129, Cur Avg Loss: 0.28981618, Log Avg loss: 0.25145767, Global Avg Loss: 1.29651591, Time: 0.0208 Steps: 39310, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000998, Sample Num: 15968, Cur Loss: 0.44269267, Cur Avg Loss: 0.28979337, Log Avg loss: 0.28753932, Global Avg Loss: 1.29625931, Time: 0.0209 Steps: 39320, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001008, Sample Num: 16128, Cur Loss: 0.41500020, Cur Avg Loss: 0.28974819, Log Avg loss: 0.28523916, Global Avg Loss: 1.29600225, Time: 0.0208 Steps: 39330, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001018, Sample Num: 16288, Cur Loss: 0.22324315, Cur Avg Loss: 0.28940771, Log Avg loss: 0.25508765, Global Avg Loss: 1.29573765, Time: 0.0209 Steps: 39340, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001028, Sample Num: 16448, Cur Loss: 0.32993886, Cur Avg Loss: 0.28956403, Log Avg loss: 0.30547688, Global Avg Loss: 1.29548600, Time: 0.0247 Steps: 39350, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001038, Sample Num: 16608, Cur Loss: 0.33386907, Cur Avg Loss: 0.28938367, Log Avg loss: 0.27084274, Global Avg Loss: 1.29522567, Time: 0.0209 Steps: 39360, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001048, Sample Num: 16768, Cur Loss: 0.13832891, Cur Avg Loss: 0.28979887, Log Avg loss: 0.33289706, Global Avg Loss: 1.29498124, Time: 0.0210 Steps: 39370, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001058, Sample Num: 16928, Cur Loss: 0.36714762, Cur Avg Loss: 0.29046135, Log Avg loss: 0.35988890, Global Avg Loss: 1.29474379, Time: 0.0209 Steps: 39380, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001068, Sample Num: 17088, Cur Loss: 0.23738080, Cur Avg Loss: 0.29074791, Log Avg loss: 0.32106649, Global Avg Loss: 1.29449660, Time: 0.0209 Steps: 39390, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001078, Sample Num: 17248, Cur Loss: 0.17035973, Cur Avg Loss: 0.29074449, Log Avg loss: 0.29037941, Global Avg Loss: 1.29424175, Time: 0.0210 Steps: 39400, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001088, Sample Num: 17408, Cur Loss: 0.48644346, Cur Avg Loss: 0.28999228, Log Avg loss: 0.20890390, Global Avg Loss: 1.29396635, Time: 0.0209 Steps: 39410, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001098, Sample Num: 17568, Cur Loss: 0.49867100, Cur Avg Loss: 0.29053071, Log Avg loss: 0.34911131, Global Avg Loss: 1.29372666, Time: 0.0209 Steps: 39420, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001108, Sample Num: 17728, Cur Loss: 0.23999929, Cur Avg Loss: 0.29080409, Log Avg loss: 0.32082102, Global Avg Loss: 1.29347992, Time: 0.0209 Steps: 39430, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001118, Sample Num: 17888, Cur Loss: 0.18321940, Cur Avg Loss: 0.29096754, Log Avg loss: 0.30907812, Global Avg Loss: 1.29323032, Time: 0.0210 Steps: 39440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001128, Sample Num: 18048, Cur Loss: 0.18213642, Cur Avg Loss: 0.29098759, Log Avg loss: 0.29322887, Global Avg Loss: 1.29297684, Time: 0.0210 Steps: 39450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001138, Sample Num: 18208, Cur Loss: 0.34502026, Cur Avg Loss: 0.29061180, Log Avg loss: 0.24822317, Global Avg Loss: 1.29271207, Time: 0.0210 Steps: 39460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001148, Sample Num: 18368, Cur Loss: 0.29632193, Cur Avg Loss: 0.29081034, Log Avg loss: 0.31340439, Global Avg Loss: 1.29246396, Time: 0.0210 Steps: 39470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001158, Sample Num: 18528, Cur Loss: 0.19687805, Cur Avg Loss: 0.29115489, Log Avg loss: 0.33070904, Global Avg Loss: 1.29222035, Time: 0.0210 Steps: 39480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001168, Sample Num: 18688, Cur Loss: 0.08922374, Cur Avg Loss: 0.29167082, Log Avg loss: 0.35141602, Global Avg Loss: 1.29198212, Time: 0.0210 Steps: 39490, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001178, Sample Num: 18848, Cur Loss: 0.77185667, Cur Avg Loss: 0.29187053, Log Avg loss: 0.31519658, Global Avg Loss: 1.29173483, Time: 0.0209 Steps: 39500, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001188, Sample Num: 19008, Cur Loss: 0.23146698, Cur Avg Loss: 0.29269914, Log Avg loss: 0.39030882, Global Avg Loss: 1.29150668, Time: 0.0210 Steps: 39510, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001198, Sample Num: 19168, Cur Loss: 0.37298733, Cur Avg Loss: 0.29269061, Log Avg loss: 0.29167695, Global Avg Loss: 1.29125368, Time: 0.0210 Steps: 39520, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001208, Sample Num: 19328, Cur Loss: 0.43422553, Cur Avg Loss: 0.29282864, Log Avg loss: 0.30936532, Global Avg Loss: 1.29100529, Time: 0.0210 Steps: 39530, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001218, Sample Num: 19488, Cur Loss: 0.24158368, Cur Avg Loss: 0.29289297, Log Avg loss: 0.30066361, Global Avg Loss: 1.29075483, Time: 0.0209 Steps: 39540, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001228, Sample Num: 19648, Cur Loss: 0.55583459, Cur Avg Loss: 0.29282197, Log Avg loss: 0.28417393, Global Avg Loss: 1.29050032, Time: 0.0210 Steps: 39550, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001238, Sample Num: 19808, Cur Loss: 0.20453382, Cur Avg Loss: 0.29164914, Log Avg loss: 0.14762621, Global Avg Loss: 1.29021142, Time: 0.0209 Steps: 39560, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001248, Sample Num: 19968, Cur Loss: 0.18616495, Cur Avg Loss: 0.29141153, Log Avg loss: 0.26199512, Global Avg Loss: 1.28995157, Time: 0.0210 Steps: 39570, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001258, Sample Num: 20128, Cur Loss: 0.32019261, Cur Avg Loss: 0.29181125, Log Avg loss: 0.34169632, Global Avg Loss: 1.28971200, Time: 0.0210 Steps: 39580, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001268, Sample Num: 20288, Cur Loss: 0.24545063, Cur Avg Loss: 0.29120731, Log Avg loss: 0.21523138, Global Avg Loss: 1.28944059, Time: 0.0210 Steps: 39590, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001278, Sample Num: 20448, Cur Loss: 0.43349928, Cur Avg Loss: 0.29143375, Log Avg loss: 0.32014683, Global Avg Loss: 1.28919582, Time: 0.0209 Steps: 39600, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001288, Sample Num: 20608, Cur Loss: 0.23827749, Cur Avg Loss: 0.29187643, Log Avg loss: 0.34845133, Global Avg Loss: 1.28895832, Time: 0.0209 Steps: 39610, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001298, Sample Num: 20768, Cur Loss: 0.08022422, Cur Avg Loss: 0.29157138, Log Avg loss: 0.25228040, Global Avg Loss: 1.28869667, Time: 0.0209 Steps: 39620, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001308, Sample Num: 20928, Cur Loss: 0.61736572, Cur Avg Loss: 0.29150347, Log Avg loss: 0.28268816, Global Avg Loss: 1.28844282, Time: 0.0209 Steps: 39630, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001318, Sample Num: 21088, Cur Loss: 0.11573306, Cur Avg Loss: 0.29155504, Log Avg loss: 0.29830073, Global Avg Loss: 1.28819303, Time: 0.0210 Steps: 39640, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001328, Sample Num: 21248, Cur Loss: 0.21302611, Cur Avg Loss: 0.29130307, Log Avg loss: 0.25809355, Global Avg Loss: 1.28793323, Time: 0.0209 Steps: 39650, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001338, Sample Num: 21408, Cur Loss: 0.44144166, Cur Avg Loss: 0.29096348, Log Avg loss: 0.24586545, Global Avg Loss: 1.28767048, Time: 0.0209 Steps: 39660, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001348, Sample Num: 21568, Cur Loss: 0.37958819, Cur Avg Loss: 0.29117140, Log Avg loss: 0.31899227, Global Avg Loss: 1.28742630, Time: 0.0209 Steps: 39670, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001358, Sample Num: 21728, Cur Loss: 0.09316547, Cur Avg Loss: 0.29170763, Log Avg loss: 0.36399082, Global Avg Loss: 1.28719358, Time: 0.0209 Steps: 39680, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001368, Sample Num: 21888, Cur Loss: 0.20766465, Cur Avg Loss: 0.29183593, Log Avg loss: 0.30925872, Global Avg Loss: 1.28694719, Time: 0.0209 Steps: 39690, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001378, Sample Num: 22048, Cur Loss: 0.29468599, Cur Avg Loss: 0.29198578, Log Avg loss: 0.31248580, Global Avg Loss: 1.28670173, Time: 0.0209 Steps: 39700, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001388, Sample Num: 22208, Cur Loss: 0.41834408, Cur Avg Loss: 0.29153806, Log Avg loss: 0.22984145, Global Avg Loss: 1.28643558, Time: 0.0209 Steps: 39710, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001398, Sample Num: 22368, Cur Loss: 0.21719722, Cur Avg Loss: 0.29081207, Log Avg loss: 0.19004548, Global Avg Loss: 1.28615955, Time: 0.0209 Steps: 39720, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001408, Sample Num: 22528, Cur Loss: 0.41755518, Cur Avg Loss: 0.29101723, Log Avg loss: 0.31969808, Global Avg Loss: 1.28591630, Time: 0.0210 Steps: 39730, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001418, Sample Num: 22688, Cur Loss: 0.37988377, Cur Avg Loss: 0.29163225, Log Avg loss: 0.37822700, Global Avg Loss: 1.28568789, Time: 0.0209 Steps: 39740, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001428, Sample Num: 22848, Cur Loss: 0.24268027, Cur Avg Loss: 0.29188772, Log Avg loss: 0.32811359, Global Avg Loss: 1.28544699, Time: 0.0209 Steps: 39750, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001438, Sample Num: 23008, Cur Loss: 0.39260119, Cur Avg Loss: 0.29304637, Log Avg loss: 0.45850161, Global Avg Loss: 1.28523901, Time: 0.0209 Steps: 39760, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001448, Sample Num: 23168, Cur Loss: 0.19435574, Cur Avg Loss: 0.29369286, Log Avg loss: 0.38665860, Global Avg Loss: 1.28501306, Time: 0.0209 Steps: 39770, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001458, Sample Num: 23328, Cur Loss: 0.15229318, Cur Avg Loss: 0.29298937, Log Avg loss: 0.19112439, Global Avg Loss: 1.28473808, Time: 0.0209 Steps: 39780, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001468, Sample Num: 23488, Cur Loss: 0.52886075, Cur Avg Loss: 0.29371132, Log Avg loss: 0.39897126, Global Avg Loss: 1.28451547, Time: 0.0208 Steps: 39790, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001478, Sample Num: 23648, Cur Loss: 0.29727459, Cur Avg Loss: 0.29347881, Log Avg loss: 0.25934556, Global Avg Loss: 1.28425789, Time: 0.0209 Steps: 39800, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001488, Sample Num: 23808, Cur Loss: 0.07133797, Cur Avg Loss: 0.29340474, Log Avg loss: 0.28245746, Global Avg Loss: 1.28400624, Time: 0.0209 Steps: 39810, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001498, Sample Num: 23968, Cur Loss: 0.17258850, Cur Avg Loss: 0.29350497, Log Avg loss: 0.30841880, Global Avg Loss: 1.28376124, Time: 0.0209 Steps: 39820, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001508, Sample Num: 24128, Cur Loss: 0.43006220, Cur Avg Loss: 0.29379898, Log Avg loss: 0.33784253, Global Avg Loss: 1.28352375, Time: 0.0209 Steps: 39830, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001518, Sample Num: 24288, Cur Loss: 0.33120355, Cur Avg Loss: 0.29361769, Log Avg loss: 0.26627870, Global Avg Loss: 1.28326842, Time: 0.0209 Steps: 39840, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001528, Sample Num: 24448, Cur Loss: 0.18171018, Cur Avg Loss: 0.29333499, Log Avg loss: 0.25042106, Global Avg Loss: 1.28300924, Time: 0.0209 Steps: 39850, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001538, Sample Num: 24608, Cur Loss: 0.31836808, Cur Avg Loss: 0.29309043, Log Avg loss: 0.25572206, Global Avg Loss: 1.28275151, Time: 0.0247 Steps: 39860, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001548, Sample Num: 24768, Cur Loss: 0.33265120, Cur Avg Loss: 0.29305961, Log Avg loss: 0.28831938, Global Avg Loss: 1.28250209, Time: 0.0209 Steps: 39870, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001558, Sample Num: 24928, Cur Loss: 0.13134979, Cur Avg Loss: 0.29260207, Log Avg loss: 0.22177409, Global Avg Loss: 1.28223611, Time: 0.0210 Steps: 39880, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001568, Sample Num: 25088, Cur Loss: 0.10932305, Cur Avg Loss: 0.29232450, Log Avg loss: 0.24907905, Global Avg Loss: 1.28197711, Time: 0.0209 Steps: 39890, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001578, Sample Num: 25248, Cur Loss: 0.47885919, Cur Avg Loss: 0.29326615, Log Avg loss: 0.44091754, Global Avg Loss: 1.28176632, Time: 0.0209 Steps: 39900, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001588, Sample Num: 25408, Cur Loss: 0.20945093, Cur Avg Loss: 0.29380903, Log Avg loss: 0.37947563, Global Avg Loss: 1.28154024, Time: 0.0208 Steps: 39910, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001598, Sample Num: 25568, Cur Loss: 0.09335604, Cur Avg Loss: 0.29363214, Log Avg loss: 0.26554180, Global Avg Loss: 1.28128573, Time: 0.0209 Steps: 39920, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001608, Sample Num: 25728, Cur Loss: 0.10999978, Cur Avg Loss: 0.29281680, Log Avg loss: 0.16252628, Global Avg Loss: 1.28100555, Time: 0.0209 Steps: 39930, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001618, Sample Num: 25888, Cur Loss: 0.48460186, Cur Avg Loss: 0.29236890, Log Avg loss: 0.22034530, Global Avg Loss: 1.28073999, Time: 0.0209 Steps: 39940, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001628, Sample Num: 26048, Cur Loss: 0.54996049, Cur Avg Loss: 0.29310055, Log Avg loss: 0.41148223, Global Avg Loss: 1.28052240, Time: 0.0209 Steps: 39950, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001638, Sample Num: 26208, Cur Loss: 0.30212808, Cur Avg Loss: 0.29358801, Log Avg loss: 0.37294670, Global Avg Loss: 1.28029528, Time: 0.0208 Steps: 39960, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001648, Sample Num: 26368, Cur Loss: 0.68902206, Cur Avg Loss: 0.29369345, Log Avg loss: 0.31096448, Global Avg Loss: 1.28005277, Time: 0.0209 Steps: 39970, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001658, Sample Num: 26528, Cur Loss: 0.42108613, Cur Avg Loss: 0.29385973, Log Avg loss: 0.32126264, Global Avg Loss: 1.27981295, Time: 0.0209 Steps: 39980, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001668, Sample Num: 26688, Cur Loss: 0.14860642, Cur Avg Loss: 0.29350920, Log Avg loss: 0.23539180, Global Avg Loss: 1.27955178, Time: 0.0209 Steps: 39990, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001678, Sample Num: 26848, Cur Loss: 0.22599520, Cur Avg Loss: 0.29338951, Log Avg loss: 0.27342438, Global Avg Loss: 1.27930025, Time: 0.0208 Steps: 40000, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001688, Sample Num: 27008, Cur Loss: 0.26372364, Cur Avg Loss: 0.29345993, Log Avg loss: 0.30527712, Global Avg Loss: 1.27905680, Time: 0.0209 Steps: 40010, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001698, Sample Num: 27168, Cur Loss: 0.37929177, Cur Avg Loss: 0.29309487, Log Avg loss: 0.23147144, Global Avg Loss: 1.27879504, Time: 0.0209 Steps: 40020, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001708, Sample Num: 27328, Cur Loss: 0.43247253, Cur Avg Loss: 0.29310163, Log Avg loss: 0.29424963, Global Avg Loss: 1.27854908, Time: 0.0209 Steps: 40030, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001718, Sample Num: 27488, Cur Loss: 0.54634207, Cur Avg Loss: 0.29271164, Log Avg loss: 0.22610211, Global Avg Loss: 1.27828623, Time: 0.0209 Steps: 40040, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001728, Sample Num: 27648, Cur Loss: 0.57676125, Cur Avg Loss: 0.29218357, Log Avg loss: 0.20146042, Global Avg Loss: 1.27801736, Time: 0.0209 Steps: 40050, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001738, Sample Num: 27808, Cur Loss: 0.24062933, Cur Avg Loss: 0.29266335, Log Avg loss: 0.37556970, Global Avg Loss: 1.27779209, Time: 0.0208 Steps: 40060, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001748, Sample Num: 27968, Cur Loss: 0.24109513, Cur Avg Loss: 0.29293514, Log Avg loss: 0.34017255, Global Avg Loss: 1.27755809, Time: 0.0208 Steps: 40070, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001758, Sample Num: 28128, Cur Loss: 0.54585749, Cur Avg Loss: 0.29258926, Log Avg loss: 0.23212958, Global Avg Loss: 1.27729726, Time: 0.0209 Steps: 40080, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001768, Sample Num: 28288, Cur Loss: 0.45498917, Cur Avg Loss: 0.29302627, Log Avg loss: 0.36985237, Global Avg Loss: 1.27707091, Time: 0.0209 Steps: 40090, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001778, Sample Num: 28448, Cur Loss: 0.31614780, Cur Avg Loss: 0.29290287, Log Avg loss: 0.27108574, Global Avg Loss: 1.27682004, Time: 0.0209 Steps: 40100, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001788, Sample Num: 28608, Cur Loss: 0.23515886, Cur Avg Loss: 0.29311891, Log Avg loss: 0.33153135, Global Avg Loss: 1.27658436, Time: 0.0208 Steps: 40110, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001798, Sample Num: 28768, Cur Loss: 0.35382298, Cur Avg Loss: 0.29368045, Log Avg loss: 0.39408356, Global Avg Loss: 1.27636440, Time: 0.0209 Steps: 40120, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001808, Sample Num: 28928, Cur Loss: 0.43234110, Cur Avg Loss: 0.29428615, Log Avg loss: 0.40319116, Global Avg Loss: 1.27614681, Time: 0.0208 Steps: 40130, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001818, Sample Num: 29088, Cur Loss: 0.17186210, Cur Avg Loss: 0.29477841, Log Avg loss: 0.38377939, Global Avg Loss: 1.27592450, Time: 0.0208 Steps: 40140, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001828, Sample Num: 29248, Cur Loss: 0.19660601, Cur Avg Loss: 0.29507879, Log Avg loss: 0.34968793, Global Avg Loss: 1.27569380, Time: 0.0209 Steps: 40150, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001838, Sample Num: 29408, Cur Loss: 0.26756108, Cur Avg Loss: 0.29496113, Log Avg loss: 0.27345133, Global Avg Loss: 1.27544424, Time: 0.0208 Steps: 40160, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001848, Sample Num: 29568, Cur Loss: 0.22492225, Cur Avg Loss: 0.29499494, Log Avg loss: 0.30121007, Global Avg Loss: 1.27520171, Time: 0.0209 Steps: 40170, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001858, Sample Num: 29728, Cur Loss: 0.37945646, Cur Avg Loss: 0.29528597, Log Avg loss: 0.34906882, Global Avg Loss: 1.27497122, Time: 0.0209 Steps: 40180, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001868, Sample Num: 29888, Cur Loss: 0.38400090, Cur Avg Loss: 0.29519018, Log Avg loss: 0.27739200, Global Avg Loss: 1.27472300, Time: 0.0209 Steps: 40190, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001878, Sample Num: 30048, Cur Loss: 0.20052290, Cur Avg Loss: 0.29496880, Log Avg loss: 0.25361394, Global Avg Loss: 1.27446900, Time: 0.0209 Steps: 40200, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001888, Sample Num: 30208, Cur Loss: 0.23637158, Cur Avg Loss: 0.29486217, Log Avg loss: 0.27483825, Global Avg Loss: 1.27422039, Time: 0.0209 Steps: 40210, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001898, Sample Num: 30368, Cur Loss: 0.19926611, Cur Avg Loss: 0.29427678, Log Avg loss: 0.18375414, Global Avg Loss: 1.27394927, Time: 0.0208 Steps: 40220, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001908, Sample Num: 30528, Cur Loss: 0.16905032, Cur Avg Loss: 0.29481791, Log Avg loss: 0.39752593, Global Avg Loss: 1.27373141, Time: 0.0209 Steps: 40230, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001918, Sample Num: 30688, Cur Loss: 0.28499955, Cur Avg Loss: 0.29437915, Log Avg loss: 0.21066304, Global Avg Loss: 1.27346723, Time: 0.0210 Steps: 40240, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001928, Sample Num: 30848, Cur Loss: 0.62198728, Cur Avg Loss: 0.29466776, Log Avg loss: 0.35002387, Global Avg Loss: 1.27323781, Time: 0.0210 Steps: 40250, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001938, Sample Num: 31008, Cur Loss: 0.37552106, Cur Avg Loss: 0.29440794, Log Avg loss: 0.24431467, Global Avg Loss: 1.27298224, Time: 0.0209 Steps: 40260, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001948, Sample Num: 31168, Cur Loss: 0.22920406, Cur Avg Loss: 0.29457323, Log Avg loss: 0.32660611, Global Avg Loss: 1.27274723, Time: 0.0208 Steps: 40270, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001958, Sample Num: 31328, Cur Loss: 0.19224219, Cur Avg Loss: 0.29438409, Log Avg loss: 0.25753861, Global Avg Loss: 1.27249519, Time: 0.0209 Steps: 40280, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001968, Sample Num: 31488, Cur Loss: 0.20382003, Cur Avg Loss: 0.29429865, Log Avg loss: 0.27757022, Global Avg Loss: 1.27224825, Time: 0.0209 Steps: 40290, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001978, Sample Num: 31648, Cur Loss: 0.25922886, Cur Avg Loss: 0.29443826, Log Avg loss: 0.32191290, Global Avg Loss: 1.27201243, Time: 0.0210 Steps: 40300, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001988, Sample Num: 31808, Cur Loss: 0.10817556, Cur Avg Loss: 0.29397624, Log Avg loss: 0.20258982, Global Avg Loss: 1.27174713, Time: 0.0209 Steps: 40310, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001998, Sample Num: 31968, Cur Loss: 0.46699607, Cur Avg Loss: 0.29373452, Log Avg loss: 0.24568006, Global Avg Loss: 1.27149265, Time: 0.0209 Steps: 40320, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002008, Sample Num: 32128, Cur Loss: 0.18122175, Cur Avg Loss: 0.29360807, Log Avg loss: 0.26834205, Global Avg Loss: 1.27124392, Time: 0.0209 Steps: 40330, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002018, Sample Num: 32288, Cur Loss: 0.34994400, Cur Avg Loss: 0.29415796, Log Avg loss: 0.40457597, Global Avg Loss: 1.27102908, Time: 0.0208 Steps: 40340, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002028, Sample Num: 32448, Cur Loss: 0.35940230, Cur Avg Loss: 0.29462593, Log Avg loss: 0.38906409, Global Avg Loss: 1.27081050, Time: 0.0209 Steps: 40350, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002038, Sample Num: 32608, Cur Loss: 0.14869596, Cur Avg Loss: 0.29427006, Log Avg loss: 0.22209939, Global Avg Loss: 1.27055066, Time: 0.0209 Steps: 40360, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002048, Sample Num: 32768, Cur Loss: 0.35356858, Cur Avg Loss: 0.29431025, Log Avg loss: 0.30250003, Global Avg Loss: 1.27031086, Time: 0.0254 Steps: 40370, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002058, Sample Num: 32928, Cur Loss: 0.36126554, Cur Avg Loss: 0.29376063, Log Avg loss: 0.18119770, Global Avg Loss: 1.27004115, Time: 0.0209 Steps: 40380, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002068, Sample Num: 33088, Cur Loss: 0.31925946, Cur Avg Loss: 0.29366918, Log Avg loss: 0.27484904, Global Avg Loss: 1.26979475, Time: 0.0210 Steps: 40390, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002078, Sample Num: 33248, Cur Loss: 0.27480954, Cur Avg Loss: 0.29357334, Log Avg loss: 0.27375482, Global Avg Loss: 1.26954821, Time: 0.0209 Steps: 40400, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002088, Sample Num: 33408, Cur Loss: 0.03842799, Cur Avg Loss: 0.29325754, Log Avg loss: 0.22763408, Global Avg Loss: 1.26929037, Time: 0.0209 Steps: 40410, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002098, Sample Num: 33568, Cur Loss: 0.15296312, Cur Avg Loss: 0.29344342, Log Avg loss: 0.33225433, Global Avg Loss: 1.26905855, Time: 0.0209 Steps: 40420, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002108, Sample Num: 33728, Cur Loss: 0.33902976, Cur Avg Loss: 0.29361197, Log Avg loss: 0.32897349, Global Avg Loss: 1.26882603, Time: 0.0209 Steps: 40430, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002118, Sample Num: 33888, Cur Loss: 0.16248217, Cur Avg Loss: 0.29372592, Log Avg loss: 0.31774770, Global Avg Loss: 1.26859084, Time: 0.0209 Steps: 40440, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002128, Sample Num: 34048, Cur Loss: 0.24299291, Cur Avg Loss: 0.29350741, Log Avg loss: 0.24722779, Global Avg Loss: 1.26833834, Time: 0.0209 Steps: 40450, Updated lr: 0.000063 ***** Running evaluation checkpoint-40451 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-40451 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.600489, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.322392, "eval_total_loss": 226.641372, "eval_mae": 0.42699, "eval_mse": 0.322483, "eval_r2": 0.795008, "eval_sp_statistic": 0.86487, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.898961, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.654209, "test_total_loss": 328.412936, "test_mae": 0.545189, "test_mse": 0.654376, "test_r2": 0.57766, "test_sp_statistic": 0.789648, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.826336, "test_ps_pvalue": 0.0, "lr": 6.258795637743007e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2683114663340387, "train_cur_epoch_loss": 624.7648972608149, "train_cur_epoch_avg_loss": 0.29345462529864486, "train_cur_epoch_time": 44.600489139556885, "train_cur_epoch_avg_time": 0.020949032005428317, "epoch": 19, "step": 40451} ################################################## Training, Epoch: 0020, Batch: 000009, Sample Num: 144, Cur Loss: 0.19341670, Cur Avg Loss: 0.24323980, Log Avg loss: 0.23702765, Global Avg Loss: 1.26808345, Time: 0.0209 Steps: 40460, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000019, Sample Num: 304, Cur Loss: 0.10995933, Cur Avg Loss: 0.28073820, Log Avg loss: 0.31448676, Global Avg Loss: 1.26784782, Time: 0.0208 Steps: 40470, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000029, Sample Num: 464, Cur Loss: 0.30058989, Cur Avg Loss: 0.30943647, Log Avg loss: 0.36396319, Global Avg Loss: 1.26762453, Time: 0.0208 Steps: 40480, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000039, Sample Num: 624, Cur Loss: 0.08548898, Cur Avg Loss: 0.31369774, Log Avg loss: 0.32605541, Global Avg Loss: 1.26739198, Time: 0.0208 Steps: 40490, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000049, Sample Num: 784, Cur Loss: 0.33024502, Cur Avg Loss: 0.31032257, Log Avg loss: 0.29715940, Global Avg Loss: 1.26715242, Time: 0.0208 Steps: 40500, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000059, Sample Num: 944, Cur Loss: 0.29278490, Cur Avg Loss: 0.32459872, Log Avg loss: 0.39455186, Global Avg Loss: 1.26693701, Time: 0.0208 Steps: 40510, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000069, Sample Num: 1104, Cur Loss: 0.15885204, Cur Avg Loss: 0.31273414, Log Avg loss: 0.24273315, Global Avg Loss: 1.26668425, Time: 0.0208 Steps: 40520, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000079, Sample Num: 1264, Cur Loss: 0.29939866, Cur Avg Loss: 0.29806737, Log Avg loss: 0.19686664, Global Avg Loss: 1.26642029, Time: 0.0208 Steps: 40530, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000089, Sample Num: 1424, Cur Loss: 0.15799457, Cur Avg Loss: 0.28643827, Log Avg loss: 0.19456834, Global Avg Loss: 1.26615590, Time: 0.0210 Steps: 40540, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000099, Sample Num: 1584, Cur Loss: 0.73485816, Cur Avg Loss: 0.29671351, Log Avg loss: 0.38816322, Global Avg Loss: 1.26593938, Time: 0.0210 Steps: 40550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000109, Sample Num: 1744, Cur Loss: 0.38146231, Cur Avg Loss: 0.30035877, Log Avg loss: 0.33644683, Global Avg Loss: 1.26571021, Time: 0.0209 Steps: 40560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000119, Sample Num: 1904, Cur Loss: 0.12840119, Cur Avg Loss: 0.29901166, Log Avg loss: 0.28432814, Global Avg Loss: 1.26546831, Time: 0.0210 Steps: 40570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000129, Sample Num: 2064, Cur Loss: 0.33055288, Cur Avg Loss: 0.29566488, Log Avg loss: 0.25583822, Global Avg Loss: 1.26521951, Time: 0.0209 Steps: 40580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000139, Sample Num: 2224, Cur Loss: 0.19178511, Cur Avg Loss: 0.29746572, Log Avg loss: 0.32069659, Global Avg Loss: 1.26498682, Time: 0.0209 Steps: 40590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000149, Sample Num: 2384, Cur Loss: 0.38132954, Cur Avg Loss: 0.29629573, Log Avg loss: 0.28003275, Global Avg Loss: 1.26474422, Time: 0.0209 Steps: 40600, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000159, Sample Num: 2544, Cur Loss: 0.50921762, Cur Avg Loss: 0.29259004, Log Avg loss: 0.23737526, Global Avg Loss: 1.26449123, Time: 0.0210 Steps: 40610, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000169, Sample Num: 2704, Cur Loss: 0.15168606, Cur Avg Loss: 0.29165184, Log Avg loss: 0.27673455, Global Avg Loss: 1.26424806, Time: 0.0209 Steps: 40620, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000179, Sample Num: 2864, Cur Loss: 0.16690518, Cur Avg Loss: 0.28641979, Log Avg loss: 0.19799814, Global Avg Loss: 1.26398563, Time: 0.0210 Steps: 40630, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000189, Sample Num: 3024, Cur Loss: 0.07186365, Cur Avg Loss: 0.29045919, Log Avg loss: 0.36276449, Global Avg Loss: 1.26376388, Time: 0.0209 Steps: 40640, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000199, Sample Num: 3184, Cur Loss: 0.13569985, Cur Avg Loss: 0.29110093, Log Avg loss: 0.30322979, Global Avg Loss: 1.26352758, Time: 0.0209 Steps: 40650, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000209, Sample Num: 3344, Cur Loss: 0.30935204, Cur Avg Loss: 0.29089053, Log Avg loss: 0.28670353, Global Avg Loss: 1.26328734, Time: 0.0209 Steps: 40660, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000219, Sample Num: 3504, Cur Loss: 0.12483026, Cur Avg Loss: 0.28948983, Log Avg loss: 0.26021528, Global Avg Loss: 1.26304070, Time: 0.0210 Steps: 40670, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000229, Sample Num: 3664, Cur Loss: 0.34121701, Cur Avg Loss: 0.28547064, Log Avg loss: 0.19745019, Global Avg Loss: 1.26277876, Time: 0.0210 Steps: 40680, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000239, Sample Num: 3824, Cur Loss: 0.20290558, Cur Avg Loss: 0.28260762, Log Avg loss: 0.21704452, Global Avg Loss: 1.26252176, Time: 0.0209 Steps: 40690, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000249, Sample Num: 3984, Cur Loss: 0.28015766, Cur Avg Loss: 0.27993046, Log Avg loss: 0.21594632, Global Avg Loss: 1.26226461, Time: 0.0209 Steps: 40700, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000259, Sample Num: 4144, Cur Loss: 0.08545559, Cur Avg Loss: 0.27676752, Log Avg loss: 0.19801026, Global Avg Loss: 1.26200319, Time: 0.0246 Steps: 40710, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000269, Sample Num: 4304, Cur Loss: 0.32934701, Cur Avg Loss: 0.27702192, Log Avg loss: 0.28361092, Global Avg Loss: 1.26176292, Time: 0.0208 Steps: 40720, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000279, Sample Num: 4464, Cur Loss: 0.39365390, Cur Avg Loss: 0.27699469, Log Avg loss: 0.27626239, Global Avg Loss: 1.26152096, Time: 0.0209 Steps: 40730, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000289, Sample Num: 4624, Cur Loss: 0.38232306, Cur Avg Loss: 0.27811277, Log Avg loss: 0.30930696, Global Avg Loss: 1.26128723, Time: 0.0209 Steps: 40740, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000299, Sample Num: 4784, Cur Loss: 0.23834223, Cur Avg Loss: 0.27714214, Log Avg loss: 0.24909098, Global Avg Loss: 1.26103884, Time: 0.0209 Steps: 40750, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000309, Sample Num: 4944, Cur Loss: 0.06011274, Cur Avg Loss: 0.27676628, Log Avg loss: 0.26552816, Global Avg Loss: 1.26079460, Time: 0.0209 Steps: 40760, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000319, Sample Num: 5104, Cur Loss: 0.20557615, Cur Avg Loss: 0.27772301, Log Avg loss: 0.30728606, Global Avg Loss: 1.26056073, Time: 0.0209 Steps: 40770, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000329, Sample Num: 5264, Cur Loss: 0.25739336, Cur Avg Loss: 0.28037470, Log Avg loss: 0.36496346, Global Avg Loss: 1.26034111, Time: 0.0209 Steps: 40780, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000339, Sample Num: 5424, Cur Loss: 0.31983411, Cur Avg Loss: 0.28174357, Log Avg loss: 0.32677956, Global Avg Loss: 1.26011224, Time: 0.0208 Steps: 40790, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000349, Sample Num: 5584, Cur Loss: 0.64720577, Cur Avg Loss: 0.28225376, Log Avg loss: 0.29954894, Global Avg Loss: 1.25987681, Time: 0.0208 Steps: 40800, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000359, Sample Num: 5744, Cur Loss: 0.11949860, Cur Avg Loss: 0.27968355, Log Avg loss: 0.18998337, Global Avg Loss: 1.25961464, Time: 0.0208 Steps: 40810, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000369, Sample Num: 5904, Cur Loss: 0.24071787, Cur Avg Loss: 0.27835511, Log Avg loss: 0.23066427, Global Avg Loss: 1.25936257, Time: 0.0208 Steps: 40820, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000379, Sample Num: 6064, Cur Loss: 0.27540031, Cur Avg Loss: 0.27945860, Log Avg loss: 0.32017721, Global Avg Loss: 1.25913255, Time: 0.0208 Steps: 40830, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000389, Sample Num: 6224, Cur Loss: 0.72852314, Cur Avg Loss: 0.27865409, Log Avg loss: 0.24816318, Global Avg Loss: 1.25888500, Time: 0.0208 Steps: 40840, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000399, Sample Num: 6384, Cur Loss: 0.07858388, Cur Avg Loss: 0.27702549, Log Avg loss: 0.21367310, Global Avg Loss: 1.25862914, Time: 0.0208 Steps: 40850, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000409, Sample Num: 6544, Cur Loss: 0.51602620, Cur Avg Loss: 0.28190410, Log Avg loss: 0.47656064, Global Avg Loss: 1.25843774, Time: 0.0208 Steps: 40860, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000419, Sample Num: 6704, Cur Loss: 0.06247537, Cur Avg Loss: 0.28142224, Log Avg loss: 0.26171388, Global Avg Loss: 1.25819386, Time: 0.0208 Steps: 40870, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000429, Sample Num: 6864, Cur Loss: 0.56771767, Cur Avg Loss: 0.28261710, Log Avg loss: 0.33268186, Global Avg Loss: 1.25796746, Time: 0.0208 Steps: 40880, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000439, Sample Num: 7024, Cur Loss: 0.33404422, Cur Avg Loss: 0.28340789, Log Avg loss: 0.31733271, Global Avg Loss: 1.25773742, Time: 0.0208 Steps: 40890, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000449, Sample Num: 7184, Cur Loss: 0.09305704, Cur Avg Loss: 0.28178236, Log Avg loss: 0.21042169, Global Avg Loss: 1.25748135, Time: 0.0208 Steps: 40900, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000459, Sample Num: 7344, Cur Loss: 0.35333511, Cur Avg Loss: 0.28014350, Log Avg loss: 0.20655873, Global Avg Loss: 1.25722447, Time: 0.0208 Steps: 40910, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000469, Sample Num: 7504, Cur Loss: 0.06219416, Cur Avg Loss: 0.27918146, Log Avg loss: 0.23502382, Global Avg Loss: 1.25697466, Time: 0.0208 Steps: 40920, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000479, Sample Num: 7664, Cur Loss: 0.58359426, Cur Avg Loss: 0.27947444, Log Avg loss: 0.29321533, Global Avg Loss: 1.25673920, Time: 0.0208 Steps: 40930, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000489, Sample Num: 7824, Cur Loss: 0.23759194, Cur Avg Loss: 0.27848204, Log Avg loss: 0.23094569, Global Avg Loss: 1.25648864, Time: 0.0208 Steps: 40940, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000499, Sample Num: 7984, Cur Loss: 0.36799225, Cur Avg Loss: 0.28093728, Log Avg loss: 0.40099868, Global Avg Loss: 1.25627973, Time: 0.0208 Steps: 40950, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000509, Sample Num: 8144, Cur Loss: 0.69258761, Cur Avg Loss: 0.28216395, Log Avg loss: 0.34337456, Global Avg Loss: 1.25605685, Time: 0.0208 Steps: 40960, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000519, Sample Num: 8304, Cur Loss: 0.50170928, Cur Avg Loss: 0.28441227, Log Avg loss: 0.39885220, Global Avg Loss: 1.25584762, Time: 0.0210 Steps: 40970, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000529, Sample Num: 8464, Cur Loss: 0.05856183, Cur Avg Loss: 0.28362830, Log Avg loss: 0.24294018, Global Avg Loss: 1.25560045, Time: 0.0208 Steps: 40980, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000539, Sample Num: 8624, Cur Loss: 0.31167212, Cur Avg Loss: 0.28292537, Log Avg loss: 0.24574027, Global Avg Loss: 1.25535408, Time: 0.0208 Steps: 40990, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000549, Sample Num: 8784, Cur Loss: 0.09484798, Cur Avg Loss: 0.28178158, Log Avg loss: 0.22013121, Global Avg Loss: 1.25510159, Time: 0.0208 Steps: 41000, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000559, Sample Num: 8944, Cur Loss: 0.19223833, Cur Avg Loss: 0.28165603, Log Avg loss: 0.27476336, Global Avg Loss: 1.25486254, Time: 0.0208 Steps: 41010, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000569, Sample Num: 9104, Cur Loss: 0.16597331, Cur Avg Loss: 0.28189755, Log Avg loss: 0.29539877, Global Avg Loss: 1.25462864, Time: 0.0208 Steps: 41020, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000579, Sample Num: 9264, Cur Loss: 0.24710223, Cur Avg Loss: 0.28111274, Log Avg loss: 0.23645689, Global Avg Loss: 1.25438049, Time: 0.0208 Steps: 41030, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000589, Sample Num: 9424, Cur Loss: 0.64044601, Cur Avg Loss: 0.28076090, Log Avg loss: 0.26038944, Global Avg Loss: 1.25413829, Time: 0.0208 Steps: 41040, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000599, Sample Num: 9584, Cur Loss: 0.60285354, Cur Avg Loss: 0.28109903, Log Avg loss: 0.30101489, Global Avg Loss: 1.25390610, Time: 0.0208 Steps: 41050, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000609, Sample Num: 9744, Cur Loss: 0.14557846, Cur Avg Loss: 0.28055068, Log Avg loss: 0.24770461, Global Avg Loss: 1.25366104, Time: 0.0209 Steps: 41060, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000619, Sample Num: 9904, Cur Loss: 0.25957161, Cur Avg Loss: 0.28105254, Log Avg loss: 0.31161565, Global Avg Loss: 1.25343167, Time: 0.0208 Steps: 41070, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000629, Sample Num: 10064, Cur Loss: 0.23077750, Cur Avg Loss: 0.28075526, Log Avg loss: 0.26235342, Global Avg Loss: 1.25319041, Time: 0.0208 Steps: 41080, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000639, Sample Num: 10224, Cur Loss: 0.38063067, Cur Avg Loss: 0.28123435, Log Avg loss: 0.31136900, Global Avg Loss: 1.25296120, Time: 0.0209 Steps: 41090, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000649, Sample Num: 10384, Cur Loss: 0.10777926, Cur Avg Loss: 0.27996290, Log Avg loss: 0.19871717, Global Avg Loss: 1.25270470, Time: 0.0208 Steps: 41100, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000659, Sample Num: 10544, Cur Loss: 0.15431301, Cur Avg Loss: 0.28022671, Log Avg loss: 0.29734843, Global Avg Loss: 1.25247231, Time: 0.0208 Steps: 41110, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000669, Sample Num: 10704, Cur Loss: 0.31357938, Cur Avg Loss: 0.28087781, Log Avg loss: 0.32378488, Global Avg Loss: 1.25224646, Time: 0.0209 Steps: 41120, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000679, Sample Num: 10864, Cur Loss: 0.48324662, Cur Avg Loss: 0.28124327, Log Avg loss: 0.30569295, Global Avg Loss: 1.25201632, Time: 0.0208 Steps: 41130, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000689, Sample Num: 11024, Cur Loss: 0.47096413, Cur Avg Loss: 0.28187545, Log Avg loss: 0.32480057, Global Avg Loss: 1.25179094, Time: 0.0208 Steps: 41140, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000699, Sample Num: 11184, Cur Loss: 0.35506222, Cur Avg Loss: 0.28297002, Log Avg loss: 0.35838602, Global Avg Loss: 1.25157383, Time: 0.0209 Steps: 41150, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000709, Sample Num: 11344, Cur Loss: 0.18524694, Cur Avg Loss: 0.28326097, Log Avg loss: 0.30359826, Global Avg Loss: 1.25134352, Time: 0.0208 Steps: 41160, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000719, Sample Num: 11504, Cur Loss: 0.10567515, Cur Avg Loss: 0.28401266, Log Avg loss: 0.33730744, Global Avg Loss: 1.25112150, Time: 0.0208 Steps: 41170, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000729, Sample Num: 11664, Cur Loss: 0.19571547, Cur Avg Loss: 0.28320619, Log Avg loss: 0.22522096, Global Avg Loss: 1.25087238, Time: 0.0208 Steps: 41180, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000739, Sample Num: 11824, Cur Loss: 0.21970409, Cur Avg Loss: 0.28300863, Log Avg loss: 0.26860612, Global Avg Loss: 1.25063390, Time: 0.0208 Steps: 41190, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000749, Sample Num: 11984, Cur Loss: 0.11812178, Cur Avg Loss: 0.28288590, Log Avg loss: 0.27381646, Global Avg Loss: 1.25039681, Time: 0.0208 Steps: 41200, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000759, Sample Num: 12144, Cur Loss: 0.26187128, Cur Avg Loss: 0.28168622, Log Avg loss: 0.19183034, Global Avg Loss: 1.25013994, Time: 0.0209 Steps: 41210, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000769, Sample Num: 12304, Cur Loss: 0.15296781, Cur Avg Loss: 0.28051022, Log Avg loss: 0.19125192, Global Avg Loss: 1.24988305, Time: 0.0246 Steps: 41220, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000779, Sample Num: 12464, Cur Loss: 0.43978447, Cur Avg Loss: 0.27947864, Log Avg loss: 0.20014945, Global Avg Loss: 1.24962845, Time: 0.0209 Steps: 41230, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000789, Sample Num: 12624, Cur Loss: 0.23841918, Cur Avg Loss: 0.27937999, Log Avg loss: 0.27169558, Global Avg Loss: 1.24939132, Time: 0.0208 Steps: 41240, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000799, Sample Num: 12784, Cur Loss: 0.26176900, Cur Avg Loss: 0.27916228, Log Avg loss: 0.26198467, Global Avg Loss: 1.24915195, Time: 0.0208 Steps: 41250, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000809, Sample Num: 12944, Cur Loss: 0.49769124, Cur Avg Loss: 0.27909567, Log Avg loss: 0.27377385, Global Avg Loss: 1.24891555, Time: 0.0208 Steps: 41260, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000819, Sample Num: 13104, Cur Loss: 0.44005999, Cur Avg Loss: 0.28057051, Log Avg loss: 0.39988471, Global Avg Loss: 1.24870982, Time: 0.0208 Steps: 41270, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000829, Sample Num: 13264, Cur Loss: 0.31902820, Cur Avg Loss: 0.28222460, Log Avg loss: 0.41769476, Global Avg Loss: 1.24850851, Time: 0.0209 Steps: 41280, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000839, Sample Num: 13424, Cur Loss: 0.37117687, Cur Avg Loss: 0.28219222, Log Avg loss: 0.27950840, Global Avg Loss: 1.24827383, Time: 0.0210 Steps: 41290, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000849, Sample Num: 13584, Cur Loss: 0.10285178, Cur Avg Loss: 0.28190072, Log Avg loss: 0.25744325, Global Avg Loss: 1.24803392, Time: 0.0209 Steps: 41300, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000859, Sample Num: 13744, Cur Loss: 0.10629545, Cur Avg Loss: 0.28134680, Log Avg loss: 0.23431890, Global Avg Loss: 1.24778853, Time: 0.0208 Steps: 41310, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000869, Sample Num: 13904, Cur Loss: 0.17711380, Cur Avg Loss: 0.28139336, Log Avg loss: 0.28539338, Global Avg Loss: 1.24755561, Time: 0.0209 Steps: 41320, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000879, Sample Num: 14064, Cur Loss: 0.25698701, Cur Avg Loss: 0.28121396, Log Avg loss: 0.26562383, Global Avg Loss: 1.24731803, Time: 0.0208 Steps: 41330, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000889, Sample Num: 14224, Cur Loss: 0.06493054, Cur Avg Loss: 0.28105099, Log Avg loss: 0.26672593, Global Avg Loss: 1.24708083, Time: 0.0208 Steps: 41340, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000899, Sample Num: 14384, Cur Loss: 0.10374813, Cur Avg Loss: 0.28082535, Log Avg loss: 0.26076616, Global Avg Loss: 1.24684230, Time: 0.0208 Steps: 41350, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000909, Sample Num: 14544, Cur Loss: 0.35536033, Cur Avg Loss: 0.28173555, Log Avg loss: 0.36356215, Global Avg Loss: 1.24662874, Time: 0.0209 Steps: 41360, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000919, Sample Num: 14704, Cur Loss: 0.22613549, Cur Avg Loss: 0.28181592, Log Avg loss: 0.28912218, Global Avg Loss: 1.24639729, Time: 0.0209 Steps: 41370, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000929, Sample Num: 14864, Cur Loss: 0.43174246, Cur Avg Loss: 0.28218913, Log Avg loss: 0.31648641, Global Avg Loss: 1.24617257, Time: 0.0208 Steps: 41380, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000939, Sample Num: 15024, Cur Loss: 0.32082534, Cur Avg Loss: 0.28160591, Log Avg loss: 0.22742530, Global Avg Loss: 1.24592643, Time: 0.0208 Steps: 41390, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000949, Sample Num: 15184, Cur Loss: 0.15958185, Cur Avg Loss: 0.28120517, Log Avg loss: 0.24357569, Global Avg Loss: 1.24568432, Time: 0.0209 Steps: 41400, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000959, Sample Num: 15344, Cur Loss: 0.16867864, Cur Avg Loss: 0.28135753, Log Avg loss: 0.29581596, Global Avg Loss: 1.24545494, Time: 0.0209 Steps: 41410, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000969, Sample Num: 15504, Cur Loss: 0.25006956, Cur Avg Loss: 0.28073372, Log Avg loss: 0.22091086, Global Avg Loss: 1.24520758, Time: 0.0208 Steps: 41420, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000979, Sample Num: 15664, Cur Loss: 1.03168726, Cur Avg Loss: 0.28157588, Log Avg loss: 0.36318078, Global Avg Loss: 1.24499469, Time: 0.0209 Steps: 41430, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000989, Sample Num: 15824, Cur Loss: 0.71383369, Cur Avg Loss: 0.28193701, Log Avg loss: 0.31729219, Global Avg Loss: 1.24477082, Time: 0.0208 Steps: 41440, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000999, Sample Num: 15984, Cur Loss: 0.31869584, Cur Avg Loss: 0.28166123, Log Avg loss: 0.25438580, Global Avg Loss: 1.24453189, Time: 0.0208 Steps: 41450, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001009, Sample Num: 16144, Cur Loss: 0.26078415, Cur Avg Loss: 0.28160556, Log Avg loss: 0.27604421, Global Avg Loss: 1.24429829, Time: 0.0208 Steps: 41460, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001019, Sample Num: 16304, Cur Loss: 0.20170286, Cur Avg Loss: 0.28128990, Log Avg loss: 0.24944005, Global Avg Loss: 1.24405839, Time: 0.0208 Steps: 41470, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001029, Sample Num: 16464, Cur Loss: 0.91073728, Cur Avg Loss: 0.28264504, Log Avg loss: 0.42073409, Global Avg Loss: 1.24385991, Time: 0.0211 Steps: 41480, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001039, Sample Num: 16624, Cur Loss: 0.14571193, Cur Avg Loss: 0.28214046, Log Avg loss: 0.23021864, Global Avg Loss: 1.24361560, Time: 0.0210 Steps: 41490, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001049, Sample Num: 16784, Cur Loss: 0.22079667, Cur Avg Loss: 0.28241585, Log Avg loss: 0.31102969, Global Avg Loss: 1.24339088, Time: 0.0210 Steps: 41500, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001059, Sample Num: 16944, Cur Loss: 0.21194680, Cur Avg Loss: 0.28252632, Log Avg loss: 0.29411446, Global Avg Loss: 1.24316219, Time: 0.0210 Steps: 41510, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001069, Sample Num: 17104, Cur Loss: 0.16891587, Cur Avg Loss: 0.28348417, Log Avg loss: 0.38491988, Global Avg Loss: 1.24295548, Time: 0.0210 Steps: 41520, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001079, Sample Num: 17264, Cur Loss: 0.15444423, Cur Avg Loss: 0.28336082, Log Avg loss: 0.27017523, Global Avg Loss: 1.24272125, Time: 0.0210 Steps: 41530, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001089, Sample Num: 17424, Cur Loss: 0.34347358, Cur Avg Loss: 0.28384731, Log Avg loss: 0.33633924, Global Avg Loss: 1.24250305, Time: 0.0210 Steps: 41540, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001099, Sample Num: 17584, Cur Loss: 0.26454663, Cur Avg Loss: 0.28345748, Log Avg loss: 0.24100539, Global Avg Loss: 1.24226202, Time: 0.0210 Steps: 41550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001109, Sample Num: 17744, Cur Loss: 0.56695902, Cur Avg Loss: 0.28321349, Log Avg loss: 0.25639857, Global Avg Loss: 1.24202480, Time: 0.0210 Steps: 41560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001119, Sample Num: 17904, Cur Loss: 0.24544209, Cur Avg Loss: 0.28273986, Log Avg loss: 0.23021387, Global Avg Loss: 1.24178141, Time: 0.0210 Steps: 41570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001129, Sample Num: 18064, Cur Loss: 0.27277997, Cur Avg Loss: 0.28419613, Log Avg loss: 0.44715377, Global Avg Loss: 1.24159030, Time: 0.0210 Steps: 41580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001139, Sample Num: 18224, Cur Loss: 0.10440074, Cur Avg Loss: 0.28395368, Log Avg loss: 0.25658006, Global Avg Loss: 1.24135346, Time: 0.0210 Steps: 41590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001149, Sample Num: 18384, Cur Loss: 0.20916331, Cur Avg Loss: 0.28411941, Log Avg loss: 0.30299659, Global Avg Loss: 1.24112789, Time: 0.0210 Steps: 41600, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001159, Sample Num: 18544, Cur Loss: 0.29251090, Cur Avg Loss: 0.28401308, Log Avg loss: 0.27179583, Global Avg Loss: 1.24089494, Time: 0.0210 Steps: 41610, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001169, Sample Num: 18704, Cur Loss: 0.58988672, Cur Avg Loss: 0.28465772, Log Avg loss: 0.35937197, Global Avg Loss: 1.24068313, Time: 0.0211 Steps: 41620, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001179, Sample Num: 18864, Cur Loss: 0.11303991, Cur Avg Loss: 0.28480969, Log Avg loss: 0.30257389, Global Avg Loss: 1.24045779, Time: 0.0210 Steps: 41630, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001189, Sample Num: 19024, Cur Loss: 0.39235061, Cur Avg Loss: 0.28455739, Log Avg loss: 0.25481195, Global Avg Loss: 1.24022108, Time: 0.0210 Steps: 41640, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001199, Sample Num: 19184, Cur Loss: 0.34461248, Cur Avg Loss: 0.28427033, Log Avg loss: 0.25013872, Global Avg Loss: 1.23998337, Time: 0.0210 Steps: 41650, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001209, Sample Num: 19344, Cur Loss: 0.37036133, Cur Avg Loss: 0.28468437, Log Avg loss: 0.33432752, Global Avg Loss: 1.23976598, Time: 0.0210 Steps: 41660, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001219, Sample Num: 19504, Cur Loss: 0.40774831, Cur Avg Loss: 0.28469667, Log Avg loss: 0.28618451, Global Avg Loss: 1.23953713, Time: 0.0210 Steps: 41670, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001229, Sample Num: 19664, Cur Loss: 0.05256363, Cur Avg Loss: 0.28452582, Log Avg loss: 0.26369924, Global Avg Loss: 1.23930301, Time: 0.0210 Steps: 41680, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001239, Sample Num: 19824, Cur Loss: 0.28074181, Cur Avg Loss: 0.28515440, Log Avg loss: 0.36240680, Global Avg Loss: 1.23909267, Time: 0.0210 Steps: 41690, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001249, Sample Num: 19984, Cur Loss: 0.49164611, Cur Avg Loss: 0.28562039, Log Avg loss: 0.34335640, Global Avg Loss: 1.23887787, Time: 0.0210 Steps: 41700, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001259, Sample Num: 20144, Cur Loss: 0.18150221, Cur Avg Loss: 0.28503879, Log Avg loss: 0.21239666, Global Avg Loss: 1.23863177, Time: 0.0211 Steps: 41710, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001269, Sample Num: 20304, Cur Loss: 0.53381443, Cur Avg Loss: 0.28518038, Log Avg loss: 0.30300664, Global Avg Loss: 1.23840750, Time: 0.0210 Steps: 41720, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001279, Sample Num: 20464, Cur Loss: 0.15546922, Cur Avg Loss: 0.28486620, Log Avg loss: 0.24499647, Global Avg Loss: 1.23816945, Time: 0.0209 Steps: 41730, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001289, Sample Num: 20624, Cur Loss: 0.06372610, Cur Avg Loss: 0.28449170, Log Avg loss: 0.23659294, Global Avg Loss: 1.23792949, Time: 0.0210 Steps: 41740, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001299, Sample Num: 20784, Cur Loss: 0.05161486, Cur Avg Loss: 0.28415028, Log Avg loss: 0.24014126, Global Avg Loss: 1.23769050, Time: 0.0209 Steps: 41750, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001309, Sample Num: 20944, Cur Loss: 0.26086721, Cur Avg Loss: 0.28379423, Log Avg loss: 0.23754342, Global Avg Loss: 1.23745100, Time: 0.0209 Steps: 41760, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001319, Sample Num: 21104, Cur Loss: 0.59439540, Cur Avg Loss: 0.28437542, Log Avg loss: 0.36045302, Global Avg Loss: 1.23724104, Time: 0.0209 Steps: 41770, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001329, Sample Num: 21264, Cur Loss: 0.29695240, Cur Avg Loss: 0.28395720, Log Avg loss: 0.22879450, Global Avg Loss: 1.23699967, Time: 0.0209 Steps: 41780, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001339, Sample Num: 21424, Cur Loss: 0.21131235, Cur Avg Loss: 0.28367004, Log Avg loss: 0.24550610, Global Avg Loss: 1.23676241, Time: 0.0208 Steps: 41790, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001349, Sample Num: 21584, Cur Loss: 0.16017373, Cur Avg Loss: 0.28402447, Log Avg loss: 0.33148245, Global Avg Loss: 1.23654584, Time: 0.0209 Steps: 41800, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001359, Sample Num: 21744, Cur Loss: 0.29307055, Cur Avg Loss: 0.28514118, Log Avg loss: 0.43578634, Global Avg Loss: 1.23635432, Time: 0.0210 Steps: 41810, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001369, Sample Num: 21904, Cur Loss: 0.23083973, Cur Avg Loss: 0.28551575, Log Avg loss: 0.33641939, Global Avg Loss: 1.23613912, Time: 0.0209 Steps: 41820, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001379, Sample Num: 22064, Cur Loss: 0.29257652, Cur Avg Loss: 0.28505967, Log Avg loss: 0.22262159, Global Avg Loss: 1.23589683, Time: 0.0209 Steps: 41830, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001389, Sample Num: 22224, Cur Loss: 0.29082659, Cur Avg Loss: 0.28570277, Log Avg loss: 0.37438646, Global Avg Loss: 1.23569092, Time: 0.0208 Steps: 41840, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001399, Sample Num: 22384, Cur Loss: 0.16514176, Cur Avg Loss: 0.28625210, Log Avg loss: 0.36255461, Global Avg Loss: 1.23548229, Time: 0.0209 Steps: 41850, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001409, Sample Num: 22544, Cur Loss: 0.41373640, Cur Avg Loss: 0.28635454, Log Avg loss: 0.30068628, Global Avg Loss: 1.23525897, Time: 0.0209 Steps: 41860, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001419, Sample Num: 22704, Cur Loss: 0.83574736, Cur Avg Loss: 0.28665628, Log Avg loss: 0.32917119, Global Avg Loss: 1.23504257, Time: 0.0209 Steps: 41870, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001429, Sample Num: 22864, Cur Loss: 0.27254122, Cur Avg Loss: 0.28681469, Log Avg loss: 0.30929345, Global Avg Loss: 1.23482152, Time: 0.0209 Steps: 41880, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001439, Sample Num: 23024, Cur Loss: 0.31058103, Cur Avg Loss: 0.28732303, Log Avg loss: 0.35996351, Global Avg Loss: 1.23461268, Time: 0.0209 Steps: 41890, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001449, Sample Num: 23184, Cur Loss: 0.05969452, Cur Avg Loss: 0.28714945, Log Avg loss: 0.26217154, Global Avg Loss: 1.23438059, Time: 0.0208 Steps: 41900, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001459, Sample Num: 23344, Cur Loss: 0.35416934, Cur Avg Loss: 0.28695341, Log Avg loss: 0.25854703, Global Avg Loss: 1.23414775, Time: 0.0208 Steps: 41910, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001469, Sample Num: 23504, Cur Loss: 0.12556487, Cur Avg Loss: 0.28663816, Log Avg loss: 0.24064345, Global Avg Loss: 1.23391075, Time: 0.0209 Steps: 41920, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001479, Sample Num: 23664, Cur Loss: 0.37806109, Cur Avg Loss: 0.28634973, Log Avg loss: 0.24397953, Global Avg Loss: 1.23367466, Time: 0.0209 Steps: 41930, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001489, Sample Num: 23824, Cur Loss: 0.28278622, Cur Avg Loss: 0.28629470, Log Avg loss: 0.27815553, Global Avg Loss: 1.23344683, Time: 0.0209 Steps: 41940, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001499, Sample Num: 23984, Cur Loss: 0.12987122, Cur Avg Loss: 0.28620846, Log Avg loss: 0.27336836, Global Avg Loss: 1.23321796, Time: 0.0209 Steps: 41950, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001509, Sample Num: 24144, Cur Loss: 0.56262505, Cur Avg Loss: 0.28697653, Log Avg loss: 0.40210960, Global Avg Loss: 1.23301989, Time: 0.0208 Steps: 41960, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001519, Sample Num: 24304, Cur Loss: 0.50844479, Cur Avg Loss: 0.28693249, Log Avg loss: 0.28028673, Global Avg Loss: 1.23279289, Time: 0.0208 Steps: 41970, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001529, Sample Num: 24464, Cur Loss: 0.09902823, Cur Avg Loss: 0.28658889, Log Avg loss: 0.23439606, Global Avg Loss: 1.23255506, Time: 0.0209 Steps: 41980, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001539, Sample Num: 24624, Cur Loss: 0.23594001, Cur Avg Loss: 0.28647435, Log Avg loss: 0.26896204, Global Avg Loss: 1.23232558, Time: 0.0246 Steps: 41990, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001549, Sample Num: 24784, Cur Loss: 0.10729964, Cur Avg Loss: 0.28582510, Log Avg loss: 0.18590529, Global Avg Loss: 1.23207643, Time: 0.0208 Steps: 42000, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001559, Sample Num: 24944, Cur Loss: 0.33561262, Cur Avg Loss: 0.28565792, Log Avg loss: 0.25976176, Global Avg Loss: 1.23184499, Time: 0.0208 Steps: 42010, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001569, Sample Num: 25104, Cur Loss: 0.14697550, Cur Avg Loss: 0.28511899, Log Avg loss: 0.20109914, Global Avg Loss: 1.23159969, Time: 0.0209 Steps: 42020, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001579, Sample Num: 25264, Cur Loss: 0.30466127, Cur Avg Loss: 0.28524450, Log Avg loss: 0.30493769, Global Avg Loss: 1.23137921, Time: 0.0208 Steps: 42030, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001589, Sample Num: 25424, Cur Loss: 0.21421231, Cur Avg Loss: 0.28501033, Log Avg loss: 0.24803364, Global Avg Loss: 1.23114530, Time: 0.0208 Steps: 42040, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001599, Sample Num: 25584, Cur Loss: 0.28811538, Cur Avg Loss: 0.28441372, Log Avg loss: 0.18961332, Global Avg Loss: 1.23089761, Time: 0.0208 Steps: 42050, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001609, Sample Num: 25744, Cur Loss: 0.09771313, Cur Avg Loss: 0.28378754, Log Avg loss: 0.18366060, Global Avg Loss: 1.23064863, Time: 0.0208 Steps: 42060, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001619, Sample Num: 25904, Cur Loss: 0.22539257, Cur Avg Loss: 0.28360422, Log Avg loss: 0.25410815, Global Avg Loss: 1.23041650, Time: 0.0208 Steps: 42070, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001629, Sample Num: 26064, Cur Loss: 0.13524443, Cur Avg Loss: 0.28284566, Log Avg loss: 0.16003584, Global Avg Loss: 1.23016214, Time: 0.0208 Steps: 42080, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001639, Sample Num: 26224, Cur Loss: 0.18173100, Cur Avg Loss: 0.28244191, Log Avg loss: 0.21666969, Global Avg Loss: 1.22992134, Time: 0.0208 Steps: 42090, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001649, Sample Num: 26384, Cur Loss: 0.65136528, Cur Avg Loss: 0.28266095, Log Avg loss: 0.31856272, Global Avg Loss: 1.22970487, Time: 0.0208 Steps: 42100, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001659, Sample Num: 26544, Cur Loss: 0.27424383, Cur Avg Loss: 0.28262667, Log Avg loss: 0.27697409, Global Avg Loss: 1.22947862, Time: 0.0208 Steps: 42110, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001669, Sample Num: 26704, Cur Loss: 0.44424206, Cur Avg Loss: 0.28250326, Log Avg loss: 0.26202903, Global Avg Loss: 1.22924893, Time: 0.0208 Steps: 42120, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001679, Sample Num: 26864, Cur Loss: 0.17992514, Cur Avg Loss: 0.28262979, Log Avg loss: 0.30374727, Global Avg Loss: 1.22902926, Time: 0.0208 Steps: 42130, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001689, Sample Num: 27024, Cur Loss: 0.33560008, Cur Avg Loss: 0.28243619, Log Avg loss: 0.24993099, Global Avg Loss: 1.22879691, Time: 0.0208 Steps: 42140, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001699, Sample Num: 27184, Cur Loss: 0.56279874, Cur Avg Loss: 0.28263661, Log Avg loss: 0.31648699, Global Avg Loss: 1.22858047, Time: 0.0208 Steps: 42150, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001709, Sample Num: 27344, Cur Loss: 0.34923011, Cur Avg Loss: 0.28259374, Log Avg loss: 0.27531047, Global Avg Loss: 1.22835436, Time: 0.0208 Steps: 42160, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001719, Sample Num: 27504, Cur Loss: 0.16907088, Cur Avg Loss: 0.28242009, Log Avg loss: 0.25274286, Global Avg Loss: 1.22812301, Time: 0.0208 Steps: 42170, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001729, Sample Num: 27664, Cur Loss: 0.42314869, Cur Avg Loss: 0.28264858, Log Avg loss: 0.32192632, Global Avg Loss: 1.22790817, Time: 0.0208 Steps: 42180, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001739, Sample Num: 27824, Cur Loss: 0.13142890, Cur Avg Loss: 0.28230875, Log Avg loss: 0.22355253, Global Avg Loss: 1.22767011, Time: 0.0208 Steps: 42190, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001749, Sample Num: 27984, Cur Loss: 0.27316374, Cur Avg Loss: 0.28212665, Log Avg loss: 0.25046034, Global Avg Loss: 1.22743855, Time: 0.0208 Steps: 42200, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001759, Sample Num: 28144, Cur Loss: 0.18724857, Cur Avg Loss: 0.28167230, Log Avg loss: 0.20220534, Global Avg Loss: 1.22719566, Time: 0.0208 Steps: 42210, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001769, Sample Num: 28304, Cur Loss: 0.20276655, Cur Avg Loss: 0.28139724, Log Avg loss: 0.23301396, Global Avg Loss: 1.22696018, Time: 0.0208 Steps: 42220, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001779, Sample Num: 28464, Cur Loss: 0.28903222, Cur Avg Loss: 0.28171111, Log Avg loss: 0.33723552, Global Avg Loss: 1.22674950, Time: 0.0208 Steps: 42230, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001789, Sample Num: 28624, Cur Loss: 0.50250864, Cur Avg Loss: 0.28184124, Log Avg loss: 0.30499178, Global Avg Loss: 1.22653128, Time: 0.0208 Steps: 42240, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001799, Sample Num: 28784, Cur Loss: 0.68261611, Cur Avg Loss: 0.28308424, Log Avg loss: 0.50545639, Global Avg Loss: 1.22636061, Time: 0.0210 Steps: 42250, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001809, Sample Num: 28944, Cur Loss: 0.40132901, Cur Avg Loss: 0.28385139, Log Avg loss: 0.42186127, Global Avg Loss: 1.22617024, Time: 0.0208 Steps: 42260, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001819, Sample Num: 29104, Cur Loss: 0.14920531, Cur Avg Loss: 0.28388374, Log Avg loss: 0.28973661, Global Avg Loss: 1.22594870, Time: 0.0208 Steps: 42270, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001829, Sample Num: 29264, Cur Loss: 0.19221476, Cur Avg Loss: 0.28415349, Log Avg loss: 0.33321993, Global Avg Loss: 1.22573756, Time: 0.0209 Steps: 42280, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001839, Sample Num: 29424, Cur Loss: 0.43346620, Cur Avg Loss: 0.28399371, Log Avg loss: 0.25476981, Global Avg Loss: 1.22550796, Time: 0.0209 Steps: 42290, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001849, Sample Num: 29584, Cur Loss: 0.26014495, Cur Avg Loss: 0.28457232, Log Avg loss: 0.39097971, Global Avg Loss: 1.22531067, Time: 0.0209 Steps: 42300, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001859, Sample Num: 29744, Cur Loss: 1.42110622, Cur Avg Loss: 0.28511425, Log Avg loss: 0.38531625, Global Avg Loss: 1.22511214, Time: 0.0209 Steps: 42310, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001869, Sample Num: 29904, Cur Loss: 0.25035802, Cur Avg Loss: 0.28513482, Log Avg loss: 0.28895886, Global Avg Loss: 1.22489093, Time: 0.0209 Steps: 42320, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001879, Sample Num: 30064, Cur Loss: 0.35543543, Cur Avg Loss: 0.28541579, Log Avg loss: 0.33793016, Global Avg Loss: 1.22468139, Time: 0.0208 Steps: 42330, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001889, Sample Num: 30224, Cur Loss: 0.22411023, Cur Avg Loss: 0.28564255, Log Avg loss: 0.32824931, Global Avg Loss: 1.22446967, Time: 0.0208 Steps: 42340, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001899, Sample Num: 30384, Cur Loss: 0.25928980, Cur Avg Loss: 0.28561356, Log Avg loss: 0.28013873, Global Avg Loss: 1.22424669, Time: 0.0208 Steps: 42350, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001909, Sample Num: 30544, Cur Loss: 0.20441228, Cur Avg Loss: 0.28564339, Log Avg loss: 0.29130806, Global Avg Loss: 1.22402645, Time: 0.0209 Steps: 42360, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001919, Sample Num: 30704, Cur Loss: 0.41091168, Cur Avg Loss: 0.28535334, Log Avg loss: 0.22998277, Global Avg Loss: 1.22379184, Time: 0.0208 Steps: 42370, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001929, Sample Num: 30864, Cur Loss: 0.11506168, Cur Avg Loss: 0.28540052, Log Avg loss: 0.29445303, Global Avg Loss: 1.22357255, Time: 0.0208 Steps: 42380, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001939, Sample Num: 31024, Cur Loss: 0.35233837, Cur Avg Loss: 0.28513518, Log Avg loss: 0.23395257, Global Avg Loss: 1.22333910, Time: 0.0208 Steps: 42390, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001949, Sample Num: 31184, Cur Loss: 0.26923859, Cur Avg Loss: 0.28520491, Log Avg loss: 0.29872568, Global Avg Loss: 1.22312103, Time: 0.0208 Steps: 42400, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001959, Sample Num: 31344, Cur Loss: 0.39454785, Cur Avg Loss: 0.28513464, Log Avg loss: 0.27143867, Global Avg Loss: 1.22289663, Time: 0.0208 Steps: 42410, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001969, Sample Num: 31504, Cur Loss: 0.38055211, Cur Avg Loss: 0.28595851, Log Avg loss: 0.44735456, Global Avg Loss: 1.22271380, Time: 0.0208 Steps: 42420, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001979, Sample Num: 31664, Cur Loss: 0.07836274, Cur Avg Loss: 0.28630320, Log Avg loss: 0.35417180, Global Avg Loss: 1.22250910, Time: 0.0210 Steps: 42430, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001989, Sample Num: 31824, Cur Loss: 0.44716412, Cur Avg Loss: 0.28647950, Log Avg loss: 0.32136935, Global Avg Loss: 1.22229677, Time: 0.0208 Steps: 42440, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001999, Sample Num: 31984, Cur Loss: 0.21833208, Cur Avg Loss: 0.28658173, Log Avg loss: 0.30691588, Global Avg Loss: 1.22208113, Time: 0.0208 Steps: 42450, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002009, Sample Num: 32144, Cur Loss: 0.52644229, Cur Avg Loss: 0.28682948, Log Avg loss: 0.33635421, Global Avg Loss: 1.22187253, Time: 0.0208 Steps: 42460, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002019, Sample Num: 32304, Cur Loss: 0.23109362, Cur Avg Loss: 0.28699464, Log Avg loss: 0.32017538, Global Avg Loss: 1.22166021, Time: 0.0208 Steps: 42470, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002029, Sample Num: 32464, Cur Loss: 0.05612518, Cur Avg Loss: 0.28706800, Log Avg loss: 0.30188057, Global Avg Loss: 1.22144369, Time: 0.0208 Steps: 42480, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002039, Sample Num: 32624, Cur Loss: 0.35232860, Cur Avg Loss: 0.28720878, Log Avg loss: 0.31577211, Global Avg Loss: 1.22123054, Time: 0.0208 Steps: 42490, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002049, Sample Num: 32784, Cur Loss: 0.14257750, Cur Avg Loss: 0.28728419, Log Avg loss: 0.30266134, Global Avg Loss: 1.22101441, Time: 0.0246 Steps: 42500, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002059, Sample Num: 32944, Cur Loss: 0.49295858, Cur Avg Loss: 0.28724569, Log Avg loss: 0.27935628, Global Avg Loss: 1.22079290, Time: 0.0209 Steps: 42510, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002069, Sample Num: 33104, Cur Loss: 0.32440507, Cur Avg Loss: 0.28681234, Log Avg loss: 0.19758556, Global Avg Loss: 1.22055225, Time: 0.0208 Steps: 42520, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002079, Sample Num: 33264, Cur Loss: 0.15474293, Cur Avg Loss: 0.28663767, Log Avg loss: 0.25049815, Global Avg Loss: 1.22032417, Time: 0.0208 Steps: 42530, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002089, Sample Num: 33424, Cur Loss: 0.25063968, Cur Avg Loss: 0.28626652, Log Avg loss: 0.20910550, Global Avg Loss: 1.22008646, Time: 0.0209 Steps: 42540, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002099, Sample Num: 33584, Cur Loss: 0.15952905, Cur Avg Loss: 0.28586304, Log Avg loss: 0.20157581, Global Avg Loss: 1.21984709, Time: 0.0208 Steps: 42550, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002109, Sample Num: 33744, Cur Loss: 0.20093325, Cur Avg Loss: 0.28603501, Log Avg loss: 0.32213112, Global Avg Loss: 1.21963616, Time: 0.0208 Steps: 42560, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002119, Sample Num: 33904, Cur Loss: 0.68612051, Cur Avg Loss: 0.28586187, Log Avg loss: 0.24934661, Global Avg Loss: 1.21940823, Time: 0.0209 Steps: 42570, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002129, Sample Num: 34055, Cur Loss: 0.03863248, Cur Avg Loss: 0.28563928, Log Avg loss: 0.23847184, Global Avg Loss: 1.21917786, Time: 0.0101 Steps: 42580, Updated lr: 0.000061 ***** Running evaluation checkpoint-42580 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-42580 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.598354, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.300247, "eval_total_loss": 211.07391, "eval_mae": 0.381432, "eval_mse": 0.300364, "eval_r2": 0.809069, "eval_sp_statistic": 0.875044, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.899576, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.596813, "test_total_loss": 299.600039, "test_mae": 0.535765, "test_mse": 0.596969, "test_r2": 0.614711, "test_sp_statistic": 0.780797, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.813061, "test_ps_pvalue": 0.0, "lr": 6.056899004267425e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2191778569368092, "train_cur_epoch_loss": 608.1260236911476, "train_cur_epoch_avg_loss": 0.285639278389454, "train_cur_epoch_time": 44.59835433959961, "train_cur_epoch_avg_time": 0.020948029281164682, "epoch": 20, "step": 42580} ################################################## Training, Epoch: 0021, Batch: 000010, Sample Num: 160, Cur Loss: 0.17423150, Cur Avg Loss: 0.23060185, Log Avg loss: 0.23060185, Global Avg Loss: 1.21894574, Time: 0.0209 Steps: 42590, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000020, Sample Num: 320, Cur Loss: 0.15571889, Cur Avg Loss: 0.24881065, Log Avg loss: 0.26701945, Global Avg Loss: 1.21872229, Time: 0.0210 Steps: 42600, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000030, Sample Num: 480, Cur Loss: 0.16049199, Cur Avg Loss: 0.24558867, Log Avg loss: 0.23914472, Global Avg Loss: 1.21849239, Time: 0.0209 Steps: 42610, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000040, Sample Num: 640, Cur Loss: 0.08267635, Cur Avg Loss: 0.26125666, Log Avg loss: 0.30826061, Global Avg Loss: 1.21827882, Time: 0.0209 Steps: 42620, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000050, Sample Num: 800, Cur Loss: 0.32656497, Cur Avg Loss: 0.25667599, Log Avg loss: 0.23835330, Global Avg Loss: 1.21804895, Time: 0.0209 Steps: 42630, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000060, Sample Num: 960, Cur Loss: 0.18180923, Cur Avg Loss: 0.25005223, Log Avg loss: 0.21693347, Global Avg Loss: 1.21781417, Time: 0.0209 Steps: 42640, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000070, Sample Num: 1120, Cur Loss: 0.34915769, Cur Avg Loss: 0.24420345, Log Avg loss: 0.20911077, Global Avg Loss: 1.21757766, Time: 0.0209 Steps: 42650, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000080, Sample Num: 1280, Cur Loss: 0.22593877, Cur Avg Loss: 0.24777243, Log Avg loss: 0.27275527, Global Avg Loss: 1.21735619, Time: 0.0209 Steps: 42660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000090, Sample Num: 1440, Cur Loss: 0.12880599, Cur Avg Loss: 0.25084626, Log Avg loss: 0.27543693, Global Avg Loss: 1.21713544, Time: 0.0209 Steps: 42670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000100, Sample Num: 1600, Cur Loss: 0.27067637, Cur Avg Loss: 0.25407566, Log Avg loss: 0.28314023, Global Avg Loss: 1.21691661, Time: 0.0209 Steps: 42680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000110, Sample Num: 1760, Cur Loss: 0.19746828, Cur Avg Loss: 0.25483349, Log Avg loss: 0.26241183, Global Avg Loss: 1.21669302, Time: 0.0209 Steps: 42690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000120, Sample Num: 1920, Cur Loss: 0.38173735, Cur Avg Loss: 0.25674410, Log Avg loss: 0.27776072, Global Avg Loss: 1.21647313, Time: 0.0209 Steps: 42700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000130, Sample Num: 2080, Cur Loss: 0.11522372, Cur Avg Loss: 0.24957496, Log Avg loss: 0.16354534, Global Avg Loss: 1.21622660, Time: 0.0209 Steps: 42710, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000140, Sample Num: 2240, Cur Loss: 0.42829329, Cur Avg Loss: 0.25257626, Log Avg loss: 0.29159312, Global Avg Loss: 1.21601016, Time: 0.0209 Steps: 42720, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000150, Sample Num: 2400, Cur Loss: 0.04827049, Cur Avg Loss: 0.25271763, Log Avg loss: 0.25469680, Global Avg Loss: 1.21578518, Time: 0.0209 Steps: 42730, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000160, Sample Num: 2560, Cur Loss: 0.36348844, Cur Avg Loss: 0.25563492, Log Avg loss: 0.29939429, Global Avg Loss: 1.21557077, Time: 0.0209 Steps: 42740, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000170, Sample Num: 2720, Cur Loss: 0.15686333, Cur Avg Loss: 0.25256142, Log Avg loss: 0.20338536, Global Avg Loss: 1.21533400, Time: 0.0209 Steps: 42750, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000180, Sample Num: 2880, Cur Loss: 0.23786125, Cur Avg Loss: 0.24922050, Log Avg loss: 0.19242493, Global Avg Loss: 1.21509478, Time: 0.0209 Steps: 42760, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000190, Sample Num: 3040, Cur Loss: 0.38659596, Cur Avg Loss: 0.24698960, Log Avg loss: 0.20683347, Global Avg Loss: 1.21485904, Time: 0.0209 Steps: 42770, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000200, Sample Num: 3200, Cur Loss: 0.28953791, Cur Avg Loss: 0.25296506, Log Avg loss: 0.36649877, Global Avg Loss: 1.21466073, Time: 0.0209 Steps: 42780, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000210, Sample Num: 3360, Cur Loss: 1.02270138, Cur Avg Loss: 0.25790714, Log Avg loss: 0.35674878, Global Avg Loss: 1.21446024, Time: 0.0209 Steps: 42790, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000220, Sample Num: 3520, Cur Loss: 0.21119568, Cur Avg Loss: 0.25929910, Log Avg loss: 0.28853024, Global Avg Loss: 1.21424390, Time: 0.0209 Steps: 42800, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000230, Sample Num: 3680, Cur Loss: 0.25419185, Cur Avg Loss: 0.25900766, Log Avg loss: 0.25259591, Global Avg Loss: 1.21401927, Time: 0.0209 Steps: 42810, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000240, Sample Num: 3840, Cur Loss: 0.22751740, Cur Avg Loss: 0.26336032, Log Avg loss: 0.36347146, Global Avg Loss: 1.21382064, Time: 0.0209 Steps: 42820, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000250, Sample Num: 4000, Cur Loss: 0.41966265, Cur Avg Loss: 0.26771387, Log Avg loss: 0.37219919, Global Avg Loss: 1.21362413, Time: 0.0209 Steps: 42830, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000260, Sample Num: 4160, Cur Loss: 0.48926300, Cur Avg Loss: 0.26668845, Log Avg loss: 0.24105302, Global Avg Loss: 1.21339711, Time: 0.0247 Steps: 42840, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000270, Sample Num: 4320, Cur Loss: 0.10359100, Cur Avg Loss: 0.26588276, Log Avg loss: 0.24493465, Global Avg Loss: 1.21317110, Time: 0.0209 Steps: 42850, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000280, Sample Num: 4480, Cur Loss: 0.29625940, Cur Avg Loss: 0.26506874, Log Avg loss: 0.24309031, Global Avg Loss: 1.21294476, Time: 0.0209 Steps: 42860, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000290, Sample Num: 4640, Cur Loss: 0.10545078, Cur Avg Loss: 0.26477136, Log Avg loss: 0.25644458, Global Avg Loss: 1.21272164, Time: 0.0209 Steps: 42870, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000300, Sample Num: 4800, Cur Loss: 0.05008394, Cur Avg Loss: 0.26514999, Log Avg loss: 0.27613039, Global Avg Loss: 1.21250322, Time: 0.0210 Steps: 42880, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000310, Sample Num: 4960, Cur Loss: 0.41299006, Cur Avg Loss: 0.26441672, Log Avg loss: 0.24241872, Global Avg Loss: 1.21227704, Time: 0.0209 Steps: 42890, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000320, Sample Num: 5120, Cur Loss: 0.16194665, Cur Avg Loss: 0.26424220, Log Avg loss: 0.25883208, Global Avg Loss: 1.21205479, Time: 0.0210 Steps: 42900, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000330, Sample Num: 5280, Cur Loss: 0.88281721, Cur Avg Loss: 0.26617230, Log Avg loss: 0.32793526, Global Avg Loss: 1.21184875, Time: 0.0209 Steps: 42910, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000340, Sample Num: 5440, Cur Loss: 0.21886060, Cur Avg Loss: 0.26413392, Log Avg loss: 0.19686750, Global Avg Loss: 1.21161227, Time: 0.0209 Steps: 42920, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000350, Sample Num: 5600, Cur Loss: 0.62187272, Cur Avg Loss: 0.26473503, Log Avg loss: 0.28517261, Global Avg Loss: 1.21139647, Time: 0.0209 Steps: 42930, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000360, Sample Num: 5760, Cur Loss: 0.38328326, Cur Avg Loss: 0.26848016, Log Avg loss: 0.39955972, Global Avg Loss: 1.21120741, Time: 0.0209 Steps: 42940, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000370, Sample Num: 5920, Cur Loss: 0.16909534, Cur Avg Loss: 0.26731449, Log Avg loss: 0.22535051, Global Avg Loss: 1.21097787, Time: 0.0210 Steps: 42950, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000380, Sample Num: 6080, Cur Loss: 0.20539920, Cur Avg Loss: 0.26824137, Log Avg loss: 0.30253575, Global Avg Loss: 1.21076641, Time: 0.0209 Steps: 42960, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000390, Sample Num: 6240, Cur Loss: 0.38025033, Cur Avg Loss: 0.26904804, Log Avg loss: 0.29970153, Global Avg Loss: 1.21055438, Time: 0.0209 Steps: 42970, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000400, Sample Num: 6400, Cur Loss: 0.18550803, Cur Avg Loss: 0.26950676, Log Avg loss: 0.28739699, Global Avg Loss: 1.21033960, Time: 0.0209 Steps: 42980, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000410, Sample Num: 6560, Cur Loss: 0.59076136, Cur Avg Loss: 0.26957869, Log Avg loss: 0.27245574, Global Avg Loss: 1.21012143, Time: 0.0209 Steps: 42990, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000420, Sample Num: 6720, Cur Loss: 0.12531605, Cur Avg Loss: 0.26837112, Log Avg loss: 0.21886082, Global Avg Loss: 1.20989091, Time: 0.0209 Steps: 43000, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000430, Sample Num: 6880, Cur Loss: 0.32473761, Cur Avg Loss: 0.26718062, Log Avg loss: 0.21717984, Global Avg Loss: 1.20966010, Time: 0.0209 Steps: 43010, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000440, Sample Num: 7040, Cur Loss: 0.32225174, Cur Avg Loss: 0.26577048, Log Avg loss: 0.20513424, Global Avg Loss: 1.20942660, Time: 0.0209 Steps: 43020, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000450, Sample Num: 7200, Cur Loss: 0.08857703, Cur Avg Loss: 0.26392607, Log Avg loss: 0.18277210, Global Avg Loss: 1.20918801, Time: 0.0209 Steps: 43030, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000460, Sample Num: 7360, Cur Loss: 0.39649403, Cur Avg Loss: 0.26426771, Log Avg loss: 0.27964166, Global Avg Loss: 1.20897203, Time: 0.0209 Steps: 43040, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000470, Sample Num: 7520, Cur Loss: 0.21392989, Cur Avg Loss: 0.26318406, Log Avg loss: 0.21333606, Global Avg Loss: 1.20874076, Time: 0.0209 Steps: 43050, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000480, Sample Num: 7680, Cur Loss: 0.31268996, Cur Avg Loss: 0.26197223, Log Avg loss: 0.20501612, Global Avg Loss: 1.20850766, Time: 0.0210 Steps: 43060, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000490, Sample Num: 7840, Cur Loss: 0.56093454, Cur Avg Loss: 0.26307235, Log Avg loss: 0.31587794, Global Avg Loss: 1.20830041, Time: 0.0210 Steps: 43070, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000500, Sample Num: 8000, Cur Loss: 0.14867240, Cur Avg Loss: 0.26290702, Log Avg loss: 0.25480629, Global Avg Loss: 1.20807908, Time: 0.0210 Steps: 43080, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000510, Sample Num: 8160, Cur Loss: 0.14350396, Cur Avg Loss: 0.26440008, Log Avg loss: 0.33905308, Global Avg Loss: 1.20787740, Time: 0.0209 Steps: 43090, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000520, Sample Num: 8320, Cur Loss: 0.15448800, Cur Avg Loss: 0.26446175, Log Avg loss: 0.26760683, Global Avg Loss: 1.20765924, Time: 0.0209 Steps: 43100, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000530, Sample Num: 8480, Cur Loss: 0.37765682, Cur Avg Loss: 0.26416394, Log Avg loss: 0.24867793, Global Avg Loss: 1.20743679, Time: 0.0209 Steps: 43110, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000540, Sample Num: 8640, Cur Loss: 0.23577344, Cur Avg Loss: 0.26407291, Log Avg loss: 0.25924782, Global Avg Loss: 1.20721690, Time: 0.0209 Steps: 43120, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000550, Sample Num: 8800, Cur Loss: 0.12910685, Cur Avg Loss: 0.26317376, Log Avg loss: 0.21461976, Global Avg Loss: 1.20698675, Time: 0.0210 Steps: 43130, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000560, Sample Num: 8960, Cur Loss: 0.27898258, Cur Avg Loss: 0.26302344, Log Avg loss: 0.25475605, Global Avg Loss: 1.20676602, Time: 0.0209 Steps: 43140, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000570, Sample Num: 9120, Cur Loss: 0.21002358, Cur Avg Loss: 0.26331788, Log Avg loss: 0.27980642, Global Avg Loss: 1.20655120, Time: 0.0209 Steps: 43150, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000580, Sample Num: 9280, Cur Loss: 0.16720702, Cur Avg Loss: 0.26300171, Log Avg loss: 0.24497987, Global Avg Loss: 1.20632841, Time: 0.0209 Steps: 43160, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000590, Sample Num: 9440, Cur Loss: 0.52635521, Cur Avg Loss: 0.26388927, Log Avg loss: 0.31536789, Global Avg Loss: 1.20612202, Time: 0.0209 Steps: 43170, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000600, Sample Num: 9600, Cur Loss: 0.16665128, Cur Avg Loss: 0.26383609, Log Avg loss: 0.26069835, Global Avg Loss: 1.20590308, Time: 0.0209 Steps: 43180, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000610, Sample Num: 9760, Cur Loss: 0.17405018, Cur Avg Loss: 0.26266719, Log Avg loss: 0.19253366, Global Avg Loss: 1.20566844, Time: 0.0209 Steps: 43190, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000620, Sample Num: 9920, Cur Loss: 0.36085084, Cur Avg Loss: 0.26196120, Log Avg loss: 0.21889535, Global Avg Loss: 1.20544003, Time: 0.0209 Steps: 43200, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000630, Sample Num: 10080, Cur Loss: 0.22412688, Cur Avg Loss: 0.26117355, Log Avg loss: 0.21233937, Global Avg Loss: 1.20521019, Time: 0.0209 Steps: 43210, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000640, Sample Num: 10240, Cur Loss: 0.54311645, Cur Avg Loss: 0.26252671, Log Avg loss: 0.34777615, Global Avg Loss: 1.20501181, Time: 0.0209 Steps: 43220, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000650, Sample Num: 10400, Cur Loss: 0.46910203, Cur Avg Loss: 0.26328273, Log Avg loss: 0.31166785, Global Avg Loss: 1.20480516, Time: 0.0209 Steps: 43230, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000660, Sample Num: 10560, Cur Loss: 0.25900403, Cur Avg Loss: 0.26272175, Log Avg loss: 0.22625765, Global Avg Loss: 1.20457885, Time: 0.0209 Steps: 43240, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000670, Sample Num: 10720, Cur Loss: 0.16417196, Cur Avg Loss: 0.26226606, Log Avg loss: 0.23219057, Global Avg Loss: 1.20435402, Time: 0.0209 Steps: 43250, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000680, Sample Num: 10880, Cur Loss: 0.11246224, Cur Avg Loss: 0.26170930, Log Avg loss: 0.22440636, Global Avg Loss: 1.20412750, Time: 0.0209 Steps: 43260, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000690, Sample Num: 11040, Cur Loss: 0.33865952, Cur Avg Loss: 0.26335413, Log Avg loss: 0.37520312, Global Avg Loss: 1.20393593, Time: 0.0209 Steps: 43270, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000700, Sample Num: 11200, Cur Loss: 0.22277112, Cur Avg Loss: 0.26293097, Log Avg loss: 0.23373262, Global Avg Loss: 1.20371176, Time: 0.0209 Steps: 43280, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000710, Sample Num: 11360, Cur Loss: 0.13251585, Cur Avg Loss: 0.26228091, Log Avg loss: 0.21677645, Global Avg Loss: 1.20348377, Time: 0.0209 Steps: 43290, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000720, Sample Num: 11520, Cur Loss: 0.44171938, Cur Avg Loss: 0.26249730, Log Avg loss: 0.27786105, Global Avg Loss: 1.20327000, Time: 0.0209 Steps: 43300, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000730, Sample Num: 11680, Cur Loss: 0.55059206, Cur Avg Loss: 0.26266121, Log Avg loss: 0.27446288, Global Avg Loss: 1.20305555, Time: 0.0209 Steps: 43310, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000740, Sample Num: 11840, Cur Loss: 0.08549318, Cur Avg Loss: 0.26257435, Log Avg loss: 0.25623336, Global Avg Loss: 1.20283698, Time: 0.0209 Steps: 43320, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000750, Sample Num: 12000, Cur Loss: 0.14502664, Cur Avg Loss: 0.26202046, Log Avg loss: 0.22103289, Global Avg Loss: 1.20261040, Time: 0.0209 Steps: 43330, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000760, Sample Num: 12160, Cur Loss: 0.42982531, Cur Avg Loss: 0.26365462, Log Avg loss: 0.38621676, Global Avg Loss: 1.20242203, Time: 0.0209 Steps: 43340, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000770, Sample Num: 12320, Cur Loss: 0.21491623, Cur Avg Loss: 0.26479889, Log Avg loss: 0.35176317, Global Avg Loss: 1.20222580, Time: 0.0247 Steps: 43350, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000780, Sample Num: 12480, Cur Loss: 0.56953657, Cur Avg Loss: 0.26665597, Log Avg loss: 0.40965087, Global Avg Loss: 1.20204301, Time: 0.0209 Steps: 43360, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000790, Sample Num: 12640, Cur Loss: 0.26704502, Cur Avg Loss: 0.26775298, Log Avg loss: 0.35332005, Global Avg Loss: 1.20184731, Time: 0.0209 Steps: 43370, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000800, Sample Num: 12800, Cur Loss: 0.31073350, Cur Avg Loss: 0.26819408, Log Avg loss: 0.30304102, Global Avg Loss: 1.20164012, Time: 0.0211 Steps: 43380, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000810, Sample Num: 12960, Cur Loss: 0.34689850, Cur Avg Loss: 0.26804367, Log Avg loss: 0.25601072, Global Avg Loss: 1.20142218, Time: 0.0208 Steps: 43390, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000820, Sample Num: 13120, Cur Loss: 0.06540293, Cur Avg Loss: 0.26801087, Log Avg loss: 0.26535384, Global Avg Loss: 1.20120650, Time: 0.0208 Steps: 43400, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000830, Sample Num: 13280, Cur Loss: 0.34684855, Cur Avg Loss: 0.26738853, Log Avg loss: 0.21635681, Global Avg Loss: 1.20097963, Time: 0.0208 Steps: 43410, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000840, Sample Num: 13440, Cur Loss: 0.13417807, Cur Avg Loss: 0.26730033, Log Avg loss: 0.25997974, Global Avg Loss: 1.20076291, Time: 0.0208 Steps: 43420, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000850, Sample Num: 13600, Cur Loss: 0.50862533, Cur Avg Loss: 0.26743322, Log Avg loss: 0.27859575, Global Avg Loss: 1.20055057, Time: 0.0208 Steps: 43430, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000860, Sample Num: 13760, Cur Loss: 0.16122049, Cur Avg Loss: 0.26689201, Log Avg loss: 0.22088993, Global Avg Loss: 1.20032505, Time: 0.0208 Steps: 43440, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000870, Sample Num: 13920, Cur Loss: 0.15911372, Cur Avg Loss: 0.26621559, Log Avg loss: 0.20804296, Global Avg Loss: 1.20009668, Time: 0.0208 Steps: 43450, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000880, Sample Num: 14080, Cur Loss: 0.22638291, Cur Avg Loss: 0.26551592, Log Avg loss: 0.20464503, Global Avg Loss: 1.19986763, Time: 0.0208 Steps: 43460, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000890, Sample Num: 14240, Cur Loss: 0.31540233, Cur Avg Loss: 0.26578958, Log Avg loss: 0.28987122, Global Avg Loss: 1.19965829, Time: 0.0209 Steps: 43470, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000900, Sample Num: 14400, Cur Loss: 0.27097550, Cur Avg Loss: 0.26538194, Log Avg loss: 0.22910209, Global Avg Loss: 1.19943507, Time: 0.0208 Steps: 43480, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000910, Sample Num: 14560, Cur Loss: 0.14941597, Cur Avg Loss: 0.26494990, Log Avg loss: 0.22606640, Global Avg Loss: 1.19921126, Time: 0.0208 Steps: 43490, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000920, Sample Num: 14720, Cur Loss: 0.31370121, Cur Avg Loss: 0.26549888, Log Avg loss: 0.31545576, Global Avg Loss: 1.19900809, Time: 0.0208 Steps: 43500, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000930, Sample Num: 14880, Cur Loss: 0.19795395, Cur Avg Loss: 0.26504136, Log Avg loss: 0.22295004, Global Avg Loss: 1.19878377, Time: 0.0208 Steps: 43510, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000940, Sample Num: 15040, Cur Loss: 0.17262055, Cur Avg Loss: 0.26552955, Log Avg loss: 0.31093139, Global Avg Loss: 1.19857975, Time: 0.0208 Steps: 43520, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000950, Sample Num: 15200, Cur Loss: 0.21634416, Cur Avg Loss: 0.26565946, Log Avg loss: 0.27787051, Global Avg Loss: 1.19836824, Time: 0.0208 Steps: 43530, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000960, Sample Num: 15360, Cur Loss: 0.27352938, Cur Avg Loss: 0.26535116, Log Avg loss: 0.23606320, Global Avg Loss: 1.19814723, Time: 0.0208 Steps: 43540, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000970, Sample Num: 15520, Cur Loss: 0.32695234, Cur Avg Loss: 0.26519228, Log Avg loss: 0.24993904, Global Avg Loss: 1.19792950, Time: 0.0208 Steps: 43550, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000980, Sample Num: 15680, Cur Loss: 0.30769914, Cur Avg Loss: 0.26614276, Log Avg loss: 0.35833937, Global Avg Loss: 1.19773676, Time: 0.0209 Steps: 43560, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000990, Sample Num: 15840, Cur Loss: 0.17661545, Cur Avg Loss: 0.26592144, Log Avg loss: 0.24423192, Global Avg Loss: 1.19751791, Time: 0.0208 Steps: 43570, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001000, Sample Num: 16000, Cur Loss: 0.07459022, Cur Avg Loss: 0.26501333, Log Avg loss: 0.17511071, Global Avg Loss: 1.19728331, Time: 0.0208 Steps: 43580, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001010, Sample Num: 16160, Cur Loss: 0.55977929, Cur Avg Loss: 0.26592682, Log Avg loss: 0.35727563, Global Avg Loss: 1.19709060, Time: 0.0209 Steps: 43590, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001020, Sample Num: 16320, Cur Loss: 0.18155281, Cur Avg Loss: 0.26649267, Log Avg loss: 0.32364341, Global Avg Loss: 1.19689027, Time: 0.0208 Steps: 43600, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001030, Sample Num: 16480, Cur Loss: 0.27787486, Cur Avg Loss: 0.26695744, Log Avg loss: 0.31436449, Global Avg Loss: 1.19668790, Time: 0.0208 Steps: 43610, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001040, Sample Num: 16640, Cur Loss: 0.28289238, Cur Avg Loss: 0.26719291, Log Avg loss: 0.29144576, Global Avg Loss: 1.19648037, Time: 0.0208 Steps: 43620, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001050, Sample Num: 16800, Cur Loss: 0.19330600, Cur Avg Loss: 0.26658588, Log Avg loss: 0.20345574, Global Avg Loss: 1.19625277, Time: 0.0208 Steps: 43630, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001060, Sample Num: 16960, Cur Loss: 0.38405675, Cur Avg Loss: 0.26589157, Log Avg loss: 0.19298843, Global Avg Loss: 1.19602287, Time: 0.0209 Steps: 43640, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001070, Sample Num: 17120, Cur Loss: 0.46775430, Cur Avg Loss: 0.26653564, Log Avg loss: 0.33480704, Global Avg Loss: 1.19582557, Time: 0.0208 Steps: 43650, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001080, Sample Num: 17280, Cur Loss: 0.30862927, Cur Avg Loss: 0.26640649, Log Avg loss: 0.25258723, Global Avg Loss: 1.19560953, Time: 0.0208 Steps: 43660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001090, Sample Num: 17440, Cur Loss: 0.10207184, Cur Avg Loss: 0.26680004, Log Avg loss: 0.30930427, Global Avg Loss: 1.19540658, Time: 0.0208 Steps: 43670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001100, Sample Num: 17600, Cur Loss: 0.13108055, Cur Avg Loss: 0.26624734, Log Avg loss: 0.20600253, Global Avg Loss: 1.19518006, Time: 0.0208 Steps: 43680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001110, Sample Num: 17760, Cur Loss: 0.29091647, Cur Avg Loss: 0.26660096, Log Avg loss: 0.30549930, Global Avg Loss: 1.19497643, Time: 0.0208 Steps: 43690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001120, Sample Num: 17920, Cur Loss: 0.06968229, Cur Avg Loss: 0.26886134, Log Avg loss: 0.51976298, Global Avg Loss: 1.19482192, Time: 0.0208 Steps: 43700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001130, Sample Num: 18080, Cur Loss: 0.18824410, Cur Avg Loss: 0.26790382, Log Avg loss: 0.16066218, Global Avg Loss: 1.19458532, Time: 0.0208 Steps: 43710, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001140, Sample Num: 18240, Cur Loss: 0.51697719, Cur Avg Loss: 0.26897117, Log Avg loss: 0.38958202, Global Avg Loss: 1.19440120, Time: 0.0208 Steps: 43720, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001150, Sample Num: 18400, Cur Loss: 0.27231354, Cur Avg Loss: 0.26918512, Log Avg loss: 0.29357446, Global Avg Loss: 1.19419520, Time: 0.0208 Steps: 43730, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001160, Sample Num: 18560, Cur Loss: 0.58445346, Cur Avg Loss: 0.26953748, Log Avg loss: 0.31005980, Global Avg Loss: 1.19399306, Time: 0.0207 Steps: 43740, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001170, Sample Num: 18720, Cur Loss: 0.51087117, Cur Avg Loss: 0.26980361, Log Avg loss: 0.30067446, Global Avg Loss: 1.19378888, Time: 0.0208 Steps: 43750, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001180, Sample Num: 18880, Cur Loss: 0.47010458, Cur Avg Loss: 0.27081598, Log Avg loss: 0.38926294, Global Avg Loss: 1.19360503, Time: 0.0208 Steps: 43760, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001190, Sample Num: 19040, Cur Loss: 0.33743346, Cur Avg Loss: 0.27098301, Log Avg loss: 0.29069319, Global Avg Loss: 1.19339874, Time: 0.0208 Steps: 43770, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001200, Sample Num: 19200, Cur Loss: 0.36654574, Cur Avg Loss: 0.27192451, Log Avg loss: 0.38396192, Global Avg Loss: 1.19321385, Time: 0.0208 Steps: 43780, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001210, Sample Num: 19360, Cur Loss: 0.35704869, Cur Avg Loss: 0.27281003, Log Avg loss: 0.37907297, Global Avg Loss: 1.19302794, Time: 0.0208 Steps: 43790, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001220, Sample Num: 19520, Cur Loss: 0.62624043, Cur Avg Loss: 0.27363802, Log Avg loss: 0.37382539, Global Avg Loss: 1.19284090, Time: 0.0208 Steps: 43800, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001230, Sample Num: 19680, Cur Loss: 0.17869900, Cur Avg Loss: 0.27356077, Log Avg loss: 0.26413529, Global Avg Loss: 1.19262892, Time: 0.0208 Steps: 43810, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001240, Sample Num: 19840, Cur Loss: 0.23403689, Cur Avg Loss: 0.27355814, Log Avg loss: 0.27323545, Global Avg Loss: 1.19241911, Time: 0.0208 Steps: 43820, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001250, Sample Num: 20000, Cur Loss: 0.14538927, Cur Avg Loss: 0.27300012, Log Avg loss: 0.20380583, Global Avg Loss: 1.19219355, Time: 0.0208 Steps: 43830, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001260, Sample Num: 20160, Cur Loss: 0.42805341, Cur Avg Loss: 0.27286509, Log Avg loss: 0.25598533, Global Avg Loss: 1.19198000, Time: 0.0208 Steps: 43840, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001270, Sample Num: 20320, Cur Loss: 0.14559180, Cur Avg Loss: 0.27266577, Log Avg loss: 0.24755188, Global Avg Loss: 1.19176462, Time: 0.0208 Steps: 43850, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001280, Sample Num: 20480, Cur Loss: 0.41864744, Cur Avg Loss: 0.27248220, Log Avg loss: 0.24916869, Global Avg Loss: 1.19154971, Time: 0.0254 Steps: 43860, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001290, Sample Num: 20640, Cur Loss: 0.41026774, Cur Avg Loss: 0.27218593, Log Avg loss: 0.23426354, Global Avg Loss: 1.19133150, Time: 0.0209 Steps: 43870, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001300, Sample Num: 20800, Cur Loss: 0.49049658, Cur Avg Loss: 0.27257564, Log Avg loss: 0.32284775, Global Avg Loss: 1.19113358, Time: 0.0209 Steps: 43880, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001310, Sample Num: 20960, Cur Loss: 0.29551205, Cur Avg Loss: 0.27272336, Log Avg loss: 0.29192801, Global Avg Loss: 1.19092870, Time: 0.0209 Steps: 43890, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001320, Sample Num: 21120, Cur Loss: 0.55174255, Cur Avg Loss: 0.27264384, Log Avg loss: 0.26222655, Global Avg Loss: 1.19071715, Time: 0.0210 Steps: 43900, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001330, Sample Num: 21280, Cur Loss: 0.29805163, Cur Avg Loss: 0.27259053, Log Avg loss: 0.26555354, Global Avg Loss: 1.19050646, Time: 0.0209 Steps: 43910, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001340, Sample Num: 21440, Cur Loss: 0.09808278, Cur Avg Loss: 0.27227311, Log Avg loss: 0.23005648, Global Avg Loss: 1.19028778, Time: 0.0209 Steps: 43920, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001350, Sample Num: 21600, Cur Loss: 0.09633707, Cur Avg Loss: 0.27234154, Log Avg loss: 0.28151096, Global Avg Loss: 1.19008091, Time: 0.0209 Steps: 43930, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001360, Sample Num: 21760, Cur Loss: 0.69729507, Cur Avg Loss: 0.27263770, Log Avg loss: 0.31261951, Global Avg Loss: 1.18988121, Time: 0.0209 Steps: 43940, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001370, Sample Num: 21920, Cur Loss: 0.05942986, Cur Avg Loss: 0.27216169, Log Avg loss: 0.20742379, Global Avg Loss: 1.18965767, Time: 0.0209 Steps: 43950, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001380, Sample Num: 22080, Cur Loss: 0.30882248, Cur Avg Loss: 0.27190005, Log Avg loss: 0.23605462, Global Avg Loss: 1.18944075, Time: 0.0209 Steps: 43960, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001390, Sample Num: 22240, Cur Loss: 0.26512697, Cur Avg Loss: 0.27189087, Log Avg loss: 0.27062457, Global Avg Loss: 1.18923178, Time: 0.0210 Steps: 43970, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001400, Sample Num: 22400, Cur Loss: 0.07966127, Cur Avg Loss: 0.27141977, Log Avg loss: 0.20593697, Global Avg Loss: 1.18900820, Time: 0.0209 Steps: 43980, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001410, Sample Num: 22560, Cur Loss: 0.30596739, Cur Avg Loss: 0.27156044, Log Avg loss: 0.29125438, Global Avg Loss: 1.18880412, Time: 0.0209 Steps: 43990, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001420, Sample Num: 22720, Cur Loss: 0.59385860, Cur Avg Loss: 0.27241108, Log Avg loss: 0.39235143, Global Avg Loss: 1.18862311, Time: 0.0209 Steps: 44000, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001430, Sample Num: 22880, Cur Loss: 0.05458710, Cur Avg Loss: 0.27215725, Log Avg loss: 0.23611255, Global Avg Loss: 1.18840668, Time: 0.0209 Steps: 44010, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001440, Sample Num: 23040, Cur Loss: 0.40880445, Cur Avg Loss: 0.27251298, Log Avg loss: 0.32338368, Global Avg Loss: 1.18821017, Time: 0.0209 Steps: 44020, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001450, Sample Num: 23200, Cur Loss: 0.38964927, Cur Avg Loss: 0.27282974, Log Avg loss: 0.31844272, Global Avg Loss: 1.18801263, Time: 0.0209 Steps: 44030, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001460, Sample Num: 23360, Cur Loss: 0.46106243, Cur Avg Loss: 0.27305383, Log Avg loss: 0.30554709, Global Avg Loss: 1.18781226, Time: 0.0209 Steps: 44040, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001470, Sample Num: 23520, Cur Loss: 0.11495017, Cur Avg Loss: 0.27308374, Log Avg loss: 0.27745060, Global Avg Loss: 1.18760559, Time: 0.0209 Steps: 44050, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001480, Sample Num: 23680, Cur Loss: 0.24336511, Cur Avg Loss: 0.27333314, Log Avg loss: 0.30999388, Global Avg Loss: 1.18740640, Time: 0.0210 Steps: 44060, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001490, Sample Num: 23840, Cur Loss: 0.49834400, Cur Avg Loss: 0.27353667, Log Avg loss: 0.30365925, Global Avg Loss: 1.18720587, Time: 0.0209 Steps: 44070, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001500, Sample Num: 24000, Cur Loss: 0.15429227, Cur Avg Loss: 0.27300484, Log Avg loss: 0.19376271, Global Avg Loss: 1.18698050, Time: 0.0209 Steps: 44080, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001510, Sample Num: 24160, Cur Loss: 0.10046808, Cur Avg Loss: 0.27279676, Log Avg loss: 0.24158419, Global Avg Loss: 1.18676608, Time: 0.0209 Steps: 44090, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001520, Sample Num: 24320, Cur Loss: 0.42147914, Cur Avg Loss: 0.27242617, Log Avg loss: 0.21646737, Global Avg Loss: 1.18654605, Time: 0.0209 Steps: 44100, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001530, Sample Num: 24480, Cur Loss: 0.11018237, Cur Avg Loss: 0.27226711, Log Avg loss: 0.24809023, Global Avg Loss: 1.18633330, Time: 0.0209 Steps: 44110, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001540, Sample Num: 24640, Cur Loss: 0.21562937, Cur Avg Loss: 0.27251674, Log Avg loss: 0.31071001, Global Avg Loss: 1.18613484, Time: 0.0246 Steps: 44120, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001550, Sample Num: 24800, Cur Loss: 0.22393465, Cur Avg Loss: 0.27177804, Log Avg loss: 0.15801874, Global Avg Loss: 1.18590186, Time: 0.0209 Steps: 44130, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001560, Sample Num: 24960, Cur Loss: 0.42470962, Cur Avg Loss: 0.27217164, Log Avg loss: 0.33317945, Global Avg Loss: 1.18570867, Time: 0.0209 Steps: 44140, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001570, Sample Num: 25120, Cur Loss: 0.24158596, Cur Avg Loss: 0.27211973, Log Avg loss: 0.26402196, Global Avg Loss: 1.18549991, Time: 0.0209 Steps: 44150, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001580, Sample Num: 25280, Cur Loss: 0.20328456, Cur Avg Loss: 0.27193546, Log Avg loss: 0.24300498, Global Avg Loss: 1.18528649, Time: 0.0209 Steps: 44160, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001590, Sample Num: 25440, Cur Loss: 0.26383844, Cur Avg Loss: 0.27153842, Log Avg loss: 0.20880645, Global Avg Loss: 1.18506541, Time: 0.0210 Steps: 44170, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001600, Sample Num: 25600, Cur Loss: 0.37909669, Cur Avg Loss: 0.27152443, Log Avg loss: 0.26929921, Global Avg Loss: 1.18485813, Time: 0.0209 Steps: 44180, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001610, Sample Num: 25760, Cur Loss: 0.61276776, Cur Avg Loss: 0.27146707, Log Avg loss: 0.26228917, Global Avg Loss: 1.18464936, Time: 0.0209 Steps: 44190, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001620, Sample Num: 25920, Cur Loss: 0.33930480, Cur Avg Loss: 0.27108890, Log Avg loss: 0.21020425, Global Avg Loss: 1.18442890, Time: 0.0209 Steps: 44200, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001630, Sample Num: 26080, Cur Loss: 0.20143902, Cur Avg Loss: 0.27121066, Log Avg loss: 0.29093562, Global Avg Loss: 1.18422679, Time: 0.0209 Steps: 44210, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001640, Sample Num: 26240, Cur Loss: 0.12320647, Cur Avg Loss: 0.27096671, Log Avg loss: 0.23120273, Global Avg Loss: 1.18401127, Time: 0.0209 Steps: 44220, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001650, Sample Num: 26400, Cur Loss: 0.24285363, Cur Avg Loss: 0.27114280, Log Avg loss: 0.30002185, Global Avg Loss: 1.18381141, Time: 0.0209 Steps: 44230, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001660, Sample Num: 26560, Cur Loss: 0.20319991, Cur Avg Loss: 0.27087304, Log Avg loss: 0.22636177, Global Avg Loss: 1.18359499, Time: 0.0209 Steps: 44240, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001670, Sample Num: 26720, Cur Loss: 0.59078962, Cur Avg Loss: 0.27065784, Log Avg loss: 0.23493498, Global Avg Loss: 1.18338060, Time: 0.0211 Steps: 44250, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001680, Sample Num: 26880, Cur Loss: 0.28677589, Cur Avg Loss: 0.27054489, Log Avg loss: 0.25168329, Global Avg Loss: 1.18317010, Time: 0.0209 Steps: 44260, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001690, Sample Num: 27040, Cur Loss: 0.08043442, Cur Avg Loss: 0.27055804, Log Avg loss: 0.27276652, Global Avg Loss: 1.18296445, Time: 0.0209 Steps: 44270, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001700, Sample Num: 27200, Cur Loss: 0.83139521, Cur Avg Loss: 0.27081342, Log Avg loss: 0.31397285, Global Avg Loss: 1.18276820, Time: 0.0209 Steps: 44280, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001710, Sample Num: 27360, Cur Loss: 0.11220023, Cur Avg Loss: 0.27031580, Log Avg loss: 0.18571996, Global Avg Loss: 1.18254308, Time: 0.0209 Steps: 44290, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001720, Sample Num: 27520, Cur Loss: 0.18531199, Cur Avg Loss: 0.26996679, Log Avg loss: 0.21028667, Global Avg Loss: 1.18232361, Time: 0.0209 Steps: 44300, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001730, Sample Num: 27680, Cur Loss: 0.09636521, Cur Avg Loss: 0.26989310, Log Avg loss: 0.25721833, Global Avg Loss: 1.18211483, Time: 0.0209 Steps: 44310, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001740, Sample Num: 27840, Cur Loss: 0.47566473, Cur Avg Loss: 0.26991787, Log Avg loss: 0.27420298, Global Avg Loss: 1.18190998, Time: 0.0209 Steps: 44320, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001750, Sample Num: 28000, Cur Loss: 0.37653592, Cur Avg Loss: 0.27012900, Log Avg loss: 0.30686518, Global Avg Loss: 1.18171259, Time: 0.0209 Steps: 44330, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001760, Sample Num: 28160, Cur Loss: 0.40855342, Cur Avg Loss: 0.26983837, Log Avg loss: 0.21897806, Global Avg Loss: 1.18149546, Time: 0.0209 Steps: 44340, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001770, Sample Num: 28320, Cur Loss: 0.26514265, Cur Avg Loss: 0.26992269, Log Avg loss: 0.28476278, Global Avg Loss: 1.18129327, Time: 0.0210 Steps: 44350, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001780, Sample Num: 28480, Cur Loss: 0.56762069, Cur Avg Loss: 0.26994955, Log Avg loss: 0.27470467, Global Avg Loss: 1.18108889, Time: 0.0209 Steps: 44360, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001790, Sample Num: 28640, Cur Loss: 0.47497174, Cur Avg Loss: 0.27020921, Log Avg loss: 0.31642809, Global Avg Loss: 1.18089402, Time: 0.0209 Steps: 44370, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001800, Sample Num: 28800, Cur Loss: 0.51091516, Cur Avg Loss: 0.27095258, Log Avg loss: 0.40401556, Global Avg Loss: 1.18071897, Time: 0.0208 Steps: 44380, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001810, Sample Num: 28960, Cur Loss: 0.26472482, Cur Avg Loss: 0.27148013, Log Avg loss: 0.36643905, Global Avg Loss: 1.18053553, Time: 0.0208 Steps: 44390, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001820, Sample Num: 29120, Cur Loss: 0.13097438, Cur Avg Loss: 0.27162364, Log Avg loss: 0.29759910, Global Avg Loss: 1.18033667, Time: 0.0208 Steps: 44400, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001830, Sample Num: 29280, Cur Loss: 0.25736019, Cur Avg Loss: 0.27147321, Log Avg loss: 0.24409507, Global Avg Loss: 1.18012585, Time: 0.0209 Steps: 44410, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001840, Sample Num: 29440, Cur Loss: 0.13508010, Cur Avg Loss: 0.27150811, Log Avg loss: 0.27789430, Global Avg Loss: 1.17992274, Time: 0.0209 Steps: 44420, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001850, Sample Num: 29600, Cur Loss: 0.09553728, Cur Avg Loss: 0.27136229, Log Avg loss: 0.24453285, Global Avg Loss: 1.17971221, Time: 0.0208 Steps: 44430, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001860, Sample Num: 29760, Cur Loss: 0.07332721, Cur Avg Loss: 0.27114424, Log Avg loss: 0.23080511, Global Avg Loss: 1.17949868, Time: 0.0208 Steps: 44440, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001870, Sample Num: 29920, Cur Loss: 0.21144053, Cur Avg Loss: 0.27128169, Log Avg loss: 0.29684711, Global Avg Loss: 1.17930011, Time: 0.0211 Steps: 44450, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001880, Sample Num: 30080, Cur Loss: 0.44018993, Cur Avg Loss: 0.27123906, Log Avg loss: 0.26326743, Global Avg Loss: 1.17909408, Time: 0.0209 Steps: 44460, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001890, Sample Num: 30240, Cur Loss: 0.10068830, Cur Avg Loss: 0.27115021, Log Avg loss: 0.25444499, Global Avg Loss: 1.17888615, Time: 0.0209 Steps: 44470, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001900, Sample Num: 30400, Cur Loss: 0.27152622, Cur Avg Loss: 0.27145974, Log Avg loss: 0.32996161, Global Avg Loss: 1.17869529, Time: 0.0208 Steps: 44480, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001910, Sample Num: 30560, Cur Loss: 0.53308702, Cur Avg Loss: 0.27211626, Log Avg loss: 0.39685430, Global Avg Loss: 1.17851956, Time: 0.0211 Steps: 44490, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001920, Sample Num: 30720, Cur Loss: 0.26363042, Cur Avg Loss: 0.27226819, Log Avg loss: 0.30128846, Global Avg Loss: 1.17832243, Time: 0.0209 Steps: 44500, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001930, Sample Num: 30880, Cur Loss: 0.45179006, Cur Avg Loss: 0.27223390, Log Avg loss: 0.26564861, Global Avg Loss: 1.17811738, Time: 0.0208 Steps: 44510, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001940, Sample Num: 31040, Cur Loss: 0.17039149, Cur Avg Loss: 0.27231852, Log Avg loss: 0.28865020, Global Avg Loss: 1.17791759, Time: 0.0209 Steps: 44520, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001950, Sample Num: 31200, Cur Loss: 0.11946458, Cur Avg Loss: 0.27180297, Log Avg loss: 0.17178627, Global Avg Loss: 1.17769164, Time: 0.0209 Steps: 44530, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001960, Sample Num: 31360, Cur Loss: 0.29687670, Cur Avg Loss: 0.27162472, Log Avg loss: 0.23686633, Global Avg Loss: 1.17748041, Time: 0.0208 Steps: 44540, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001970, Sample Num: 31520, Cur Loss: 0.30505112, Cur Avg Loss: 0.27130965, Log Avg loss: 0.20955602, Global Avg Loss: 1.17726315, Time: 0.0208 Steps: 44550, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001980, Sample Num: 31680, Cur Loss: 0.11218201, Cur Avg Loss: 0.27106245, Log Avg loss: 0.22236470, Global Avg Loss: 1.17704885, Time: 0.0209 Steps: 44560, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001990, Sample Num: 31840, Cur Loss: 0.52236009, Cur Avg Loss: 0.27094999, Log Avg loss: 0.24868214, Global Avg Loss: 1.17684056, Time: 0.0208 Steps: 44570, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002000, Sample Num: 32000, Cur Loss: 1.12521338, Cur Avg Loss: 0.27170809, Log Avg loss: 0.42256974, Global Avg Loss: 1.17667136, Time: 0.0208 Steps: 44580, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002010, Sample Num: 32160, Cur Loss: 0.27364811, Cur Avg Loss: 0.27254255, Log Avg loss: 0.43943621, Global Avg Loss: 1.17650603, Time: 0.0209 Steps: 44590, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002020, Sample Num: 32320, Cur Loss: 0.20158073, Cur Avg Loss: 0.27284395, Log Avg loss: 0.33342523, Global Avg Loss: 1.17631699, Time: 0.0209 Steps: 44600, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002030, Sample Num: 32480, Cur Loss: 0.30870277, Cur Avg Loss: 0.27321196, Log Avg loss: 0.34754889, Global Avg Loss: 1.17613121, Time: 0.0209 Steps: 44610, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002040, Sample Num: 32640, Cur Loss: 0.38265389, Cur Avg Loss: 0.27329094, Log Avg loss: 0.28932327, Global Avg Loss: 1.17593247, Time: 0.0209 Steps: 44620, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002050, Sample Num: 32800, Cur Loss: 0.20194583, Cur Avg Loss: 0.27316502, Log Avg loss: 0.24747897, Global Avg Loss: 1.17572443, Time: 0.0247 Steps: 44630, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002060, Sample Num: 32960, Cur Loss: 0.49563757, Cur Avg Loss: 0.27313758, Log Avg loss: 0.26751168, Global Avg Loss: 1.17552098, Time: 0.0209 Steps: 44640, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002070, Sample Num: 33120, Cur Loss: 0.31341153, Cur Avg Loss: 0.27330206, Log Avg loss: 0.30718542, Global Avg Loss: 1.17532650, Time: 0.0209 Steps: 44650, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002080, Sample Num: 33280, Cur Loss: 0.54174727, Cur Avg Loss: 0.27337020, Log Avg loss: 0.28747397, Global Avg Loss: 1.17512770, Time: 0.0209 Steps: 44660, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002090, Sample Num: 33440, Cur Loss: 0.17000511, Cur Avg Loss: 0.27283625, Log Avg loss: 0.16177548, Global Avg Loss: 1.17490085, Time: 0.0209 Steps: 44670, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002100, Sample Num: 33600, Cur Loss: 0.26155591, Cur Avg Loss: 0.27287834, Log Avg loss: 0.28167514, Global Avg Loss: 1.17470093, Time: 0.0209 Steps: 44680, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002110, Sample Num: 33760, Cur Loss: 0.78084207, Cur Avg Loss: 0.27372893, Log Avg loss: 0.45235298, Global Avg Loss: 1.17453930, Time: 0.0210 Steps: 44690, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002120, Sample Num: 33920, Cur Loss: 0.16544276, Cur Avg Loss: 0.27359319, Log Avg loss: 0.24495164, Global Avg Loss: 1.17433134, Time: 0.0210 Steps: 44700, Updated lr: 0.000059 ***** Running evaluation checkpoint-44709 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-44709 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.612820, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.30412, "eval_total_loss": 213.796399, "eval_mae": 0.389747, "eval_mse": 0.304212, "eval_r2": 0.806623, "eval_sp_statistic": 0.859025, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.902288, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.851789, "test_total_loss": 427.59831, "test_mae": 0.685001, "test_mse": 0.851934, "test_r2": 0.450154, "test_sp_statistic": 0.794459, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.837052, "test_ps_pvalue": 0.0, "lr": 5.855002370791844e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.174124123598216, "train_cur_epoch_loss": 581.3222935833037, "train_cur_epoch_avg_loss": 0.2730494568263521, "train_cur_epoch_time": 44.61282014846802, "train_cur_epoch_avg_time": 0.020954823930703626, "epoch": 21, "step": 44709} ################################################## Training, Epoch: 0022, Batch: 000001, Sample Num: 16, Cur Loss: 0.16012043, Cur Avg Loss: 0.16012043, Log Avg loss: 0.14648518, Global Avg Loss: 1.17410144, Time: 0.0248 Steps: 44710, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000011, Sample Num: 176, Cur Loss: 0.17380145, Cur Avg Loss: 0.21293981, Log Avg loss: 0.21822175, Global Avg Loss: 1.17388770, Time: 0.0209 Steps: 44720, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000021, Sample Num: 336, Cur Loss: 0.38963288, Cur Avg Loss: 0.23574723, Log Avg loss: 0.26083538, Global Avg Loss: 1.17368357, Time: 0.0209 Steps: 44730, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000031, Sample Num: 496, Cur Loss: 0.12283326, Cur Avg Loss: 0.24000037, Log Avg loss: 0.24893198, Global Avg Loss: 1.17347688, Time: 0.0209 Steps: 44740, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000041, Sample Num: 656, Cur Loss: 0.22544293, Cur Avg Loss: 0.23840412, Log Avg loss: 0.23345573, Global Avg Loss: 1.17326682, Time: 0.0209 Steps: 44750, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000051, Sample Num: 816, Cur Loss: 0.41311157, Cur Avg Loss: 0.25361099, Log Avg loss: 0.31595917, Global Avg Loss: 1.17307528, Time: 0.0210 Steps: 44760, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000061, Sample Num: 976, Cur Loss: 0.08510110, Cur Avg Loss: 0.24545113, Log Avg loss: 0.20383585, Global Avg Loss: 1.17285879, Time: 0.0209 Steps: 44770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000071, Sample Num: 1136, Cur Loss: 0.15753999, Cur Avg Loss: 0.25081425, Log Avg loss: 0.28352923, Global Avg Loss: 1.17266019, Time: 0.0209 Steps: 44780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000081, Sample Num: 1296, Cur Loss: 0.15266405, Cur Avg Loss: 0.24515217, Log Avg loss: 0.20495140, Global Avg Loss: 1.17244413, Time: 0.0209 Steps: 44790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000091, Sample Num: 1456, Cur Loss: 0.31540310, Cur Avg Loss: 0.24286892, Log Avg loss: 0.22437460, Global Avg Loss: 1.17223251, Time: 0.0209 Steps: 44800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000101, Sample Num: 1616, Cur Loss: 0.41223457, Cur Avg Loss: 0.24112052, Log Avg loss: 0.22521010, Global Avg Loss: 1.17202117, Time: 0.0209 Steps: 44810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000111, Sample Num: 1776, Cur Loss: 0.24733737, Cur Avg Loss: 0.24034983, Log Avg loss: 0.23256590, Global Avg Loss: 1.17181156, Time: 0.0209 Steps: 44820, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000121, Sample Num: 1936, Cur Loss: 0.08472594, Cur Avg Loss: 0.23717050, Log Avg loss: 0.20187988, Global Avg Loss: 1.17159521, Time: 0.0209 Steps: 44830, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000131, Sample Num: 2096, Cur Loss: 0.10720868, Cur Avg Loss: 0.23559736, Log Avg loss: 0.21656245, Global Avg Loss: 1.17138222, Time: 0.0209 Steps: 44840, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000141, Sample Num: 2256, Cur Loss: 0.20000511, Cur Avg Loss: 0.23545548, Log Avg loss: 0.23359685, Global Avg Loss: 1.17117313, Time: 0.0209 Steps: 44850, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000151, Sample Num: 2416, Cur Loss: 0.20664935, Cur Avg Loss: 0.24071810, Log Avg loss: 0.31492090, Global Avg Loss: 1.17098225, Time: 0.0209 Steps: 44860, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000161, Sample Num: 2576, Cur Loss: 0.28699023, Cur Avg Loss: 0.24580954, Log Avg loss: 0.32269042, Global Avg Loss: 1.17079320, Time: 0.0209 Steps: 44870, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000171, Sample Num: 2736, Cur Loss: 0.32176971, Cur Avg Loss: 0.25503850, Log Avg loss: 0.40362462, Global Avg Loss: 1.17062226, Time: 0.0209 Steps: 44880, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000181, Sample Num: 2896, Cur Loss: 0.26867720, Cur Avg Loss: 0.26000765, Log Avg loss: 0.34498019, Global Avg Loss: 1.17043833, Time: 0.0209 Steps: 44890, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000191, Sample Num: 3056, Cur Loss: 0.26924253, Cur Avg Loss: 0.26042956, Log Avg loss: 0.26806621, Global Avg Loss: 1.17023736, Time: 0.0209 Steps: 44900, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000201, Sample Num: 3216, Cur Loss: 0.20976420, Cur Avg Loss: 0.25757941, Log Avg loss: 0.20314144, Global Avg Loss: 1.17002202, Time: 0.0209 Steps: 44910, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000211, Sample Num: 3376, Cur Loss: 0.35014904, Cur Avg Loss: 0.25585008, Log Avg loss: 0.22109051, Global Avg Loss: 1.16981077, Time: 0.0210 Steps: 44920, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000221, Sample Num: 3536, Cur Loss: 0.19009680, Cur Avg Loss: 0.25747730, Log Avg loss: 0.29181173, Global Avg Loss: 1.16961536, Time: 0.0209 Steps: 44930, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000231, Sample Num: 3696, Cur Loss: 0.32315361, Cur Avg Loss: 0.25816373, Log Avg loss: 0.27333378, Global Avg Loss: 1.16941592, Time: 0.0209 Steps: 44940, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000241, Sample Num: 3856, Cur Loss: 0.60310149, Cur Avg Loss: 0.25679256, Log Avg loss: 0.22511856, Global Avg Loss: 1.16920584, Time: 0.0209 Steps: 44950, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000251, Sample Num: 4016, Cur Loss: 0.12085528, Cur Avg Loss: 0.25556818, Log Avg loss: 0.22606066, Global Avg Loss: 1.16899606, Time: 0.0209 Steps: 44960, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000261, Sample Num: 4176, Cur Loss: 0.17146499, Cur Avg Loss: 0.25364899, Log Avg loss: 0.20547735, Global Avg Loss: 1.16878181, Time: 0.0246 Steps: 44970, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000271, Sample Num: 4336, Cur Loss: 0.09483911, Cur Avg Loss: 0.25323287, Log Avg loss: 0.24237215, Global Avg Loss: 1.16857585, Time: 0.0219 Steps: 44980, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000281, Sample Num: 4496, Cur Loss: 0.30852479, Cur Avg Loss: 0.25390009, Log Avg loss: 0.27198157, Global Avg Loss: 1.16837656, Time: 0.0219 Steps: 44990, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000291, Sample Num: 4656, Cur Loss: 0.08573047, Cur Avg Loss: 0.25120665, Log Avg loss: 0.17552103, Global Avg Loss: 1.16815592, Time: 0.0219 Steps: 45000, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000301, Sample Num: 4816, Cur Loss: 0.20563725, Cur Avg Loss: 0.25334529, Log Avg loss: 0.31557968, Global Avg Loss: 1.16796650, Time: 0.0220 Steps: 45010, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000311, Sample Num: 4976, Cur Loss: 0.07064140, Cur Avg Loss: 0.25349864, Log Avg loss: 0.25811473, Global Avg Loss: 1.16776441, Time: 0.0210 Steps: 45020, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000321, Sample Num: 5136, Cur Loss: 0.28244755, Cur Avg Loss: 0.25402736, Log Avg loss: 0.27047039, Global Avg Loss: 1.16756514, Time: 0.0209 Steps: 45030, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000331, Sample Num: 5296, Cur Loss: 0.28299776, Cur Avg Loss: 0.25233960, Log Avg loss: 0.19816259, Global Avg Loss: 1.16734991, Time: 0.0209 Steps: 45040, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000341, Sample Num: 5456, Cur Loss: 0.13741513, Cur Avg Loss: 0.25010000, Log Avg loss: 0.17596920, Global Avg Loss: 1.16712985, Time: 0.0209 Steps: 45050, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000351, Sample Num: 5616, Cur Loss: 0.24381644, Cur Avg Loss: 0.24926898, Log Avg loss: 0.22093123, Global Avg Loss: 1.16691986, Time: 0.0220 Steps: 45060, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000361, Sample Num: 5776, Cur Loss: 0.31447956, Cur Avg Loss: 0.24758408, Log Avg loss: 0.18844405, Global Avg Loss: 1.16670276, Time: 0.0219 Steps: 45070, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000371, Sample Num: 5936, Cur Loss: 0.23630777, Cur Avg Loss: 0.24745041, Log Avg loss: 0.24262478, Global Avg Loss: 1.16649777, Time: 0.0210 Steps: 45080, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000381, Sample Num: 6096, Cur Loss: 0.23460805, Cur Avg Loss: 0.25050587, Log Avg loss: 0.36386371, Global Avg Loss: 1.16631976, Time: 0.0209 Steps: 45090, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000391, Sample Num: 6256, Cur Loss: 0.23180214, Cur Avg Loss: 0.25395893, Log Avg loss: 0.38552056, Global Avg Loss: 1.16614664, Time: 0.0220 Steps: 45100, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000401, Sample Num: 6416, Cur Loss: 0.39401436, Cur Avg Loss: 0.25607355, Log Avg loss: 0.33875482, Global Avg Loss: 1.16596322, Time: 0.0209 Steps: 45110, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000411, Sample Num: 6576, Cur Loss: 0.27473646, Cur Avg Loss: 0.25791154, Log Avg loss: 0.33161518, Global Avg Loss: 1.16577830, Time: 0.0221 Steps: 45120, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000421, Sample Num: 6736, Cur Loss: 0.24803071, Cur Avg Loss: 0.25761468, Log Avg loss: 0.24541352, Global Avg Loss: 1.16557437, Time: 0.0219 Steps: 45130, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000431, Sample Num: 6896, Cur Loss: 0.77748120, Cur Avg Loss: 0.25989699, Log Avg loss: 0.35598259, Global Avg Loss: 1.16539502, Time: 0.0219 Steps: 45140, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000441, Sample Num: 7056, Cur Loss: 0.27669829, Cur Avg Loss: 0.25901191, Log Avg loss: 0.22086470, Global Avg Loss: 1.16518582, Time: 0.0220 Steps: 45150, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000451, Sample Num: 7216, Cur Loss: 0.02632632, Cur Avg Loss: 0.25742253, Log Avg loss: 0.18733101, Global Avg Loss: 1.16496929, Time: 0.0209 Steps: 45160, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000461, Sample Num: 7376, Cur Loss: 0.10437050, Cur Avg Loss: 0.25762875, Log Avg loss: 0.26692935, Global Avg Loss: 1.16477047, Time: 0.0209 Steps: 45170, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000471, Sample Num: 7536, Cur Loss: 0.18674713, Cur Avg Loss: 0.25617419, Log Avg loss: 0.18911886, Global Avg Loss: 1.16455453, Time: 0.0209 Steps: 45180, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000481, Sample Num: 7696, Cur Loss: 0.17532250, Cur Avg Loss: 0.25653012, Log Avg loss: 0.27329435, Global Avg Loss: 1.16435730, Time: 0.0209 Steps: 45190, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000491, Sample Num: 7856, Cur Loss: 0.10039225, Cur Avg Loss: 0.25749941, Log Avg loss: 0.30412232, Global Avg Loss: 1.16416698, Time: 0.0209 Steps: 45200, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000501, Sample Num: 8016, Cur Loss: 0.69125915, Cur Avg Loss: 0.25895065, Log Avg loss: 0.33020656, Global Avg Loss: 1.16398252, Time: 0.0209 Steps: 45210, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000511, Sample Num: 8176, Cur Loss: 0.18842857, Cur Avg Loss: 0.25771618, Log Avg loss: 0.19586903, Global Avg Loss: 1.16376843, Time: 0.0219 Steps: 45220, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000521, Sample Num: 8336, Cur Loss: 0.13321927, Cur Avg Loss: 0.25775979, Log Avg loss: 0.25998863, Global Avg Loss: 1.16356861, Time: 0.0219 Steps: 45230, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000531, Sample Num: 8496, Cur Loss: 0.24429521, Cur Avg Loss: 0.25812357, Log Avg loss: 0.27707652, Global Avg Loss: 1.16337266, Time: 0.0219 Steps: 45240, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000541, Sample Num: 8656, Cur Loss: 0.12169285, Cur Avg Loss: 0.25953912, Log Avg loss: 0.33470483, Global Avg Loss: 1.16318953, Time: 0.0219 Steps: 45250, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000551, Sample Num: 8816, Cur Loss: 0.17765456, Cur Avg Loss: 0.25893842, Log Avg loss: 0.22644046, Global Avg Loss: 1.16298256, Time: 0.0219 Steps: 45260, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000561, Sample Num: 8976, Cur Loss: 0.30374348, Cur Avg Loss: 0.25868918, Log Avg loss: 0.24495586, Global Avg Loss: 1.16277977, Time: 0.0219 Steps: 45270, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000571, Sample Num: 9136, Cur Loss: 0.47375613, Cur Avg Loss: 0.25874152, Log Avg loss: 0.26167795, Global Avg Loss: 1.16258076, Time: 0.0220 Steps: 45280, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000581, Sample Num: 9296, Cur Loss: 0.17098799, Cur Avg Loss: 0.26026644, Log Avg loss: 0.34733916, Global Avg Loss: 1.16240076, Time: 0.0219 Steps: 45290, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000591, Sample Num: 9456, Cur Loss: 0.20625633, Cur Avg Loss: 0.25924970, Log Avg loss: 0.20017723, Global Avg Loss: 1.16218834, Time: 0.0219 Steps: 45300, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000601, Sample Num: 9616, Cur Loss: 0.29386806, Cur Avg Loss: 0.26003346, Log Avg loss: 0.30635353, Global Avg Loss: 1.16199946, Time: 0.0220 Steps: 45310, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000611, Sample Num: 9776, Cur Loss: 0.45803475, Cur Avg Loss: 0.25991102, Log Avg loss: 0.25255275, Global Avg Loss: 1.16179879, Time: 0.0219 Steps: 45320, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000621, Sample Num: 9936, Cur Loss: 0.12140705, Cur Avg Loss: 0.25879278, Log Avg loss: 0.19046808, Global Avg Loss: 1.16158451, Time: 0.0219 Steps: 45330, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000631, Sample Num: 10096, Cur Loss: 0.39831907, Cur Avg Loss: 0.25805962, Log Avg loss: 0.21253008, Global Avg Loss: 1.16137519, Time: 0.0219 Steps: 45340, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000641, Sample Num: 10256, Cur Loss: 0.32544738, Cur Avg Loss: 0.25815432, Log Avg loss: 0.26413000, Global Avg Loss: 1.16117734, Time: 0.0219 Steps: 45350, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000651, Sample Num: 10416, Cur Loss: 0.10026712, Cur Avg Loss: 0.25733583, Log Avg loss: 0.20487107, Global Avg Loss: 1.16096651, Time: 0.0219 Steps: 45360, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000661, Sample Num: 10576, Cur Loss: 0.21592730, Cur Avg Loss: 0.25670946, Log Avg loss: 0.21593282, Global Avg Loss: 1.16075822, Time: 0.0219 Steps: 45370, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000671, Sample Num: 10736, Cur Loss: 0.05602796, Cur Avg Loss: 0.25577295, Log Avg loss: 0.19386943, Global Avg Loss: 1.16054515, Time: 0.0219 Steps: 45380, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000681, Sample Num: 10896, Cur Loss: 0.21041261, Cur Avg Loss: 0.25561965, Log Avg loss: 0.24533324, Global Avg Loss: 1.16034352, Time: 0.0219 Steps: 45390, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000691, Sample Num: 11056, Cur Loss: 0.10587461, Cur Avg Loss: 0.25578373, Log Avg loss: 0.26695741, Global Avg Loss: 1.16014674, Time: 0.0219 Steps: 45400, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000701, Sample Num: 11216, Cur Loss: 0.55375946, Cur Avg Loss: 0.25693236, Log Avg loss: 0.33630265, Global Avg Loss: 1.15996532, Time: 0.0219 Steps: 45410, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000711, Sample Num: 11376, Cur Loss: 0.18384561, Cur Avg Loss: 0.25789706, Log Avg loss: 0.32552267, Global Avg Loss: 1.15978160, Time: 0.0220 Steps: 45420, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000721, Sample Num: 11536, Cur Loss: 0.07665920, Cur Avg Loss: 0.25716318, Log Avg loss: 0.20498416, Global Avg Loss: 1.15957143, Time: 0.0219 Steps: 45430, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000731, Sample Num: 11696, Cur Loss: 0.13017499, Cur Avg Loss: 0.25697929, Log Avg loss: 0.24372092, Global Avg Loss: 1.15936988, Time: 0.0219 Steps: 45440, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000741, Sample Num: 11856, Cur Loss: 0.33785638, Cur Avg Loss: 0.25661607, Log Avg loss: 0.23006446, Global Avg Loss: 1.15916541, Time: 0.0219 Steps: 45450, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000751, Sample Num: 12016, Cur Loss: 0.22478576, Cur Avg Loss: 0.25707779, Log Avg loss: 0.29129139, Global Avg Loss: 1.15897450, Time: 0.0219 Steps: 45460, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000761, Sample Num: 12176, Cur Loss: 0.16401905, Cur Avg Loss: 0.25656470, Log Avg loss: 0.21803182, Global Avg Loss: 1.15876756, Time: 0.0219 Steps: 45470, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000771, Sample Num: 12336, Cur Loss: 0.34097192, Cur Avg Loss: 0.25813954, Log Avg loss: 0.37798465, Global Avg Loss: 1.15859589, Time: 0.0246 Steps: 45480, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000781, Sample Num: 12496, Cur Loss: 0.17022955, Cur Avg Loss: 0.25780739, Log Avg loss: 0.23219894, Global Avg Loss: 1.15839224, Time: 0.0209 Steps: 45490, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000791, Sample Num: 12656, Cur Loss: 0.18193050, Cur Avg Loss: 0.25749140, Log Avg loss: 0.23281268, Global Avg Loss: 1.15818882, Time: 0.0209 Steps: 45500, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000801, Sample Num: 12816, Cur Loss: 0.37894934, Cur Avg Loss: 0.25766881, Log Avg loss: 0.27170169, Global Avg Loss: 1.15799403, Time: 0.0209 Steps: 45510, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000811, Sample Num: 12976, Cur Loss: 0.59345615, Cur Avg Loss: 0.25758203, Log Avg loss: 0.25063107, Global Avg Loss: 1.15779469, Time: 0.0209 Steps: 45520, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000821, Sample Num: 13136, Cur Loss: 0.26767695, Cur Avg Loss: 0.25755435, Log Avg loss: 0.25530915, Global Avg Loss: 1.15759648, Time: 0.0209 Steps: 45530, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000831, Sample Num: 13296, Cur Loss: 0.22106248, Cur Avg Loss: 0.25769811, Log Avg loss: 0.26950093, Global Avg Loss: 1.15740146, Time: 0.0209 Steps: 45540, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000841, Sample Num: 13456, Cur Loss: 0.11399320, Cur Avg Loss: 0.25717752, Log Avg loss: 0.21391683, Global Avg Loss: 1.15719433, Time: 0.0209 Steps: 45550, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000851, Sample Num: 13616, Cur Loss: 0.05811291, Cur Avg Loss: 0.25622199, Log Avg loss: 0.17586161, Global Avg Loss: 1.15697894, Time: 0.0209 Steps: 45560, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000861, Sample Num: 13776, Cur Loss: 0.34875539, Cur Avg Loss: 0.25705332, Log Avg loss: 0.32779986, Global Avg Loss: 1.15679698, Time: 0.0209 Steps: 45570, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000871, Sample Num: 13936, Cur Loss: 0.35394004, Cur Avg Loss: 0.25692797, Log Avg loss: 0.24613465, Global Avg Loss: 1.15659719, Time: 0.0209 Steps: 45580, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000881, Sample Num: 14096, Cur Loss: 0.12989779, Cur Avg Loss: 0.25769735, Log Avg loss: 0.32471086, Global Avg Loss: 1.15641471, Time: 0.0209 Steps: 45590, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000891, Sample Num: 14256, Cur Loss: 0.37187606, Cur Avg Loss: 0.25743113, Log Avg loss: 0.23397671, Global Avg Loss: 1.15621242, Time: 0.0209 Steps: 45600, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000901, Sample Num: 14416, Cur Loss: 0.38677043, Cur Avg Loss: 0.25753865, Log Avg loss: 0.26711853, Global Avg Loss: 1.15601749, Time: 0.0209 Steps: 45610, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000911, Sample Num: 14576, Cur Loss: 0.23716030, Cur Avg Loss: 0.25674380, Log Avg loss: 0.18512820, Global Avg Loss: 1.15580467, Time: 0.0209 Steps: 45620, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000921, Sample Num: 14736, Cur Loss: 0.06700649, Cur Avg Loss: 0.25682393, Log Avg loss: 0.26412387, Global Avg Loss: 1.15560925, Time: 0.0211 Steps: 45630, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000931, Sample Num: 14896, Cur Loss: 0.23973463, Cur Avg Loss: 0.25662250, Log Avg loss: 0.23807061, Global Avg Loss: 1.15540822, Time: 0.0209 Steps: 45640, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000941, Sample Num: 15056, Cur Loss: 0.14502729, Cur Avg Loss: 0.25691235, Log Avg loss: 0.28389780, Global Avg Loss: 1.15521730, Time: 0.0209 Steps: 45650, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000951, Sample Num: 15216, Cur Loss: 0.08146895, Cur Avg Loss: 0.25713210, Log Avg loss: 0.27780999, Global Avg Loss: 1.15502514, Time: 0.0209 Steps: 45660, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000961, Sample Num: 15376, Cur Loss: 0.30845469, Cur Avg Loss: 0.25690170, Log Avg loss: 0.23499068, Global Avg Loss: 1.15482369, Time: 0.0209 Steps: 45670, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000971, Sample Num: 15536, Cur Loss: 0.11424104, Cur Avg Loss: 0.25592976, Log Avg loss: 0.16252614, Global Avg Loss: 1.15460646, Time: 0.0209 Steps: 45680, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000981, Sample Num: 15696, Cur Loss: 0.34542024, Cur Avg Loss: 0.25611974, Log Avg loss: 0.27456776, Global Avg Loss: 1.15441385, Time: 0.0209 Steps: 45690, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000991, Sample Num: 15856, Cur Loss: 0.20556837, Cur Avg Loss: 0.25607501, Log Avg loss: 0.25168618, Global Avg Loss: 1.15421632, Time: 0.0209 Steps: 45700, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001001, Sample Num: 16016, Cur Loss: 0.39153966, Cur Avg Loss: 0.25663564, Log Avg loss: 0.31219438, Global Avg Loss: 1.15403211, Time: 0.0209 Steps: 45710, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001011, Sample Num: 16176, Cur Loss: 0.23765929, Cur Avg Loss: 0.25853179, Log Avg loss: 0.44833678, Global Avg Loss: 1.15387776, Time: 0.0209 Steps: 45720, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001021, Sample Num: 16336, Cur Loss: 0.10643962, Cur Avg Loss: 0.25823284, Log Avg loss: 0.22800828, Global Avg Loss: 1.15367529, Time: 0.0209 Steps: 45730, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001031, Sample Num: 16496, Cur Loss: 0.19573855, Cur Avg Loss: 0.25801537, Log Avg loss: 0.23581208, Global Avg Loss: 1.15347462, Time: 0.0210 Steps: 45740, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001041, Sample Num: 16656, Cur Loss: 0.24325690, Cur Avg Loss: 0.25732592, Log Avg loss: 0.18624322, Global Avg Loss: 1.15326321, Time: 0.0209 Steps: 45750, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001051, Sample Num: 16816, Cur Loss: 0.27686703, Cur Avg Loss: 0.25706508, Log Avg loss: 0.22991232, Global Avg Loss: 1.15306143, Time: 0.0209 Steps: 45760, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001061, Sample Num: 16976, Cur Loss: 0.15939173, Cur Avg Loss: 0.25773182, Log Avg loss: 0.32780572, Global Avg Loss: 1.15288112, Time: 0.0209 Steps: 45770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001071, Sample Num: 17136, Cur Loss: 0.35396004, Cur Avg Loss: 0.25764885, Log Avg loss: 0.24884572, Global Avg Loss: 1.15268365, Time: 0.0208 Steps: 45780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001081, Sample Num: 17296, Cur Loss: 0.52740818, Cur Avg Loss: 0.25777419, Log Avg loss: 0.27119828, Global Avg Loss: 1.15249114, Time: 0.0209 Steps: 45790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001091, Sample Num: 17456, Cur Loss: 0.10944147, Cur Avg Loss: 0.25774309, Log Avg loss: 0.25438150, Global Avg Loss: 1.15229505, Time: 0.0208 Steps: 45800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001101, Sample Num: 17616, Cur Loss: 0.16998339, Cur Avg Loss: 0.25737603, Log Avg loss: 0.21732970, Global Avg Loss: 1.15209095, Time: 0.0208 Steps: 45810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001111, Sample Num: 17776, Cur Loss: 0.25854075, Cur Avg Loss: 0.25760962, Log Avg loss: 0.28332725, Global Avg Loss: 1.15190135, Time: 0.0208 Steps: 45820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001121, Sample Num: 17936, Cur Loss: 0.32881749, Cur Avg Loss: 0.25779284, Log Avg loss: 0.27814946, Global Avg Loss: 1.15171070, Time: 0.0209 Steps: 45830, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001131, Sample Num: 18096, Cur Loss: 0.39411855, Cur Avg Loss: 0.25811425, Log Avg loss: 0.29414436, Global Avg Loss: 1.15152362, Time: 0.0209 Steps: 45840, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001141, Sample Num: 18256, Cur Loss: 0.16248432, Cur Avg Loss: 0.25744331, Log Avg loss: 0.18155976, Global Avg Loss: 1.15131207, Time: 0.0209 Steps: 45850, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001151, Sample Num: 18416, Cur Loss: 0.12445810, Cur Avg Loss: 0.25743445, Log Avg loss: 0.25642277, Global Avg Loss: 1.15111693, Time: 0.0209 Steps: 45860, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001161, Sample Num: 18576, Cur Loss: 0.35744065, Cur Avg Loss: 0.25670340, Log Avg loss: 0.17256009, Global Avg Loss: 1.15090360, Time: 0.0209 Steps: 45870, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001171, Sample Num: 18736, Cur Loss: 0.66158664, Cur Avg Loss: 0.25757586, Log Avg loss: 0.35886850, Global Avg Loss: 1.15073097, Time: 0.0209 Steps: 45880, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001181, Sample Num: 18896, Cur Loss: 0.21774206, Cur Avg Loss: 0.25813405, Log Avg loss: 0.32349845, Global Avg Loss: 1.15055070, Time: 0.0209 Steps: 45890, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001191, Sample Num: 19056, Cur Loss: 0.15581638, Cur Avg Loss: 0.25763520, Log Avg loss: 0.19872078, Global Avg Loss: 1.15034333, Time: 0.0208 Steps: 45900, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001201, Sample Num: 19216, Cur Loss: 0.12621844, Cur Avg Loss: 0.25743161, Log Avg loss: 0.23318395, Global Avg Loss: 1.15014356, Time: 0.0209 Steps: 45910, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001211, Sample Num: 19376, Cur Loss: 0.53229403, Cur Avg Loss: 0.25742217, Log Avg loss: 0.25628798, Global Avg Loss: 1.14994890, Time: 0.0209 Steps: 45920, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001221, Sample Num: 19536, Cur Loss: 0.12696736, Cur Avg Loss: 0.25787125, Log Avg loss: 0.31225502, Global Avg Loss: 1.14976652, Time: 0.0209 Steps: 45930, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001231, Sample Num: 19696, Cur Loss: 0.19087741, Cur Avg Loss: 0.25797458, Log Avg loss: 0.27059137, Global Avg Loss: 1.14957514, Time: 0.0210 Steps: 45940, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001241, Sample Num: 19856, Cur Loss: 0.23274967, Cur Avg Loss: 0.25837529, Log Avg loss: 0.30770284, Global Avg Loss: 1.14939193, Time: 0.0209 Steps: 45950, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001251, Sample Num: 20016, Cur Loss: 0.44350868, Cur Avg Loss: 0.25888972, Log Avg loss: 0.32273032, Global Avg Loss: 1.14921206, Time: 0.0208 Steps: 45960, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001261, Sample Num: 20176, Cur Loss: 0.52681333, Cur Avg Loss: 0.25929080, Log Avg loss: 0.30946539, Global Avg Loss: 1.14902939, Time: 0.0208 Steps: 45970, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001271, Sample Num: 20336, Cur Loss: 1.11744428, Cur Avg Loss: 0.26060726, Log Avg loss: 0.42661385, Global Avg Loss: 1.14887228, Time: 0.0209 Steps: 45980, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001281, Sample Num: 20496, Cur Loss: 0.17359340, Cur Avg Loss: 0.26097174, Log Avg loss: 0.30729713, Global Avg Loss: 1.14868929, Time: 0.0246 Steps: 45990, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001291, Sample Num: 20656, Cur Loss: 0.16506797, Cur Avg Loss: 0.26115815, Log Avg loss: 0.28503713, Global Avg Loss: 1.14850154, Time: 0.0210 Steps: 46000, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001301, Sample Num: 20816, Cur Loss: 0.65672731, Cur Avg Loss: 0.26092075, Log Avg loss: 0.23027188, Global Avg Loss: 1.14830196, Time: 0.0210 Steps: 46010, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001311, Sample Num: 20976, Cur Loss: 0.12359136, Cur Avg Loss: 0.26060452, Log Avg loss: 0.21946342, Global Avg Loss: 1.14810013, Time: 0.0210 Steps: 46020, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001321, Sample Num: 21136, Cur Loss: 0.13757512, Cur Avg Loss: 0.26045106, Log Avg loss: 0.24033196, Global Avg Loss: 1.14790292, Time: 0.0209 Steps: 46030, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001331, Sample Num: 21296, Cur Loss: 0.23072754, Cur Avg Loss: 0.26050799, Log Avg loss: 0.26802912, Global Avg Loss: 1.14771181, Time: 0.0210 Steps: 46040, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001341, Sample Num: 21456, Cur Loss: 0.12214261, Cur Avg Loss: 0.26029244, Log Avg loss: 0.23160250, Global Avg Loss: 1.14751287, Time: 0.0209 Steps: 46050, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001351, Sample Num: 21616, Cur Loss: 0.27892372, Cur Avg Loss: 0.26011923, Log Avg loss: 0.23689126, Global Avg Loss: 1.14731517, Time: 0.0209 Steps: 46060, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001361, Sample Num: 21776, Cur Loss: 0.25459090, Cur Avg Loss: 0.26062169, Log Avg loss: 0.32850473, Global Avg Loss: 1.14713743, Time: 0.0209 Steps: 46070, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001371, Sample Num: 21936, Cur Loss: 0.29240799, Cur Avg Loss: 0.26146635, Log Avg loss: 0.37642357, Global Avg Loss: 1.14697018, Time: 0.0209 Steps: 46080, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001381, Sample Num: 22096, Cur Loss: 0.37312305, Cur Avg Loss: 0.26111699, Log Avg loss: 0.21322064, Global Avg Loss: 1.14676759, Time: 0.0209 Steps: 46090, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001391, Sample Num: 22256, Cur Loss: 0.61235869, Cur Avg Loss: 0.26158213, Log Avg loss: 0.32581816, Global Avg Loss: 1.14658951, Time: 0.0209 Steps: 46100, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001401, Sample Num: 22416, Cur Loss: 0.38345140, Cur Avg Loss: 0.26160668, Log Avg loss: 0.26502166, Global Avg Loss: 1.14639832, Time: 0.0209 Steps: 46110, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001411, Sample Num: 22576, Cur Loss: 0.45140260, Cur Avg Loss: 0.26168682, Log Avg loss: 0.27291413, Global Avg Loss: 1.14620892, Time: 0.0209 Steps: 46120, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001421, Sample Num: 22736, Cur Loss: 0.11034250, Cur Avg Loss: 0.26179732, Log Avg loss: 0.27738857, Global Avg Loss: 1.14602058, Time: 0.0209 Steps: 46130, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001431, Sample Num: 22896, Cur Loss: 0.46749529, Cur Avg Loss: 0.26234508, Log Avg loss: 0.34018127, Global Avg Loss: 1.14584593, Time: 0.0209 Steps: 46140, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001441, Sample Num: 23056, Cur Loss: 0.13619982, Cur Avg Loss: 0.26253721, Log Avg loss: 0.29003212, Global Avg Loss: 1.14566049, Time: 0.0209 Steps: 46150, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001451, Sample Num: 23216, Cur Loss: 0.33099374, Cur Avg Loss: 0.26264134, Log Avg loss: 0.27764587, Global Avg Loss: 1.14547244, Time: 0.0209 Steps: 46160, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001461, Sample Num: 23376, Cur Loss: 0.13237193, Cur Avg Loss: 0.26229414, Log Avg loss: 0.21191521, Global Avg Loss: 1.14527024, Time: 0.0210 Steps: 46170, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001471, Sample Num: 23536, Cur Loss: 0.13535403, Cur Avg Loss: 0.26164651, Log Avg loss: 0.16702766, Global Avg Loss: 1.14505841, Time: 0.0209 Steps: 46180, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001481, Sample Num: 23696, Cur Loss: 0.39529645, Cur Avg Loss: 0.26147489, Log Avg loss: 0.23622998, Global Avg Loss: 1.14486165, Time: 0.0209 Steps: 46190, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001491, Sample Num: 23856, Cur Loss: 0.21010762, Cur Avg Loss: 0.26106746, Log Avg loss: 0.20072703, Global Avg Loss: 1.14465729, Time: 0.0209 Steps: 46200, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001501, Sample Num: 24016, Cur Loss: 0.06536467, Cur Avg Loss: 0.26168481, Log Avg loss: 0.35373207, Global Avg Loss: 1.14448614, Time: 0.0209 Steps: 46210, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001511, Sample Num: 24176, Cur Loss: 0.11007638, Cur Avg Loss: 0.26109906, Log Avg loss: 0.17317778, Global Avg Loss: 1.14427599, Time: 0.0209 Steps: 46220, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001521, Sample Num: 24336, Cur Loss: 0.32825980, Cur Avg Loss: 0.26081745, Log Avg loss: 0.21826554, Global Avg Loss: 1.14407568, Time: 0.0209 Steps: 46230, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001531, Sample Num: 24496, Cur Loss: 0.40930513, Cur Avg Loss: 0.26130812, Log Avg loss: 0.33593978, Global Avg Loss: 1.14390091, Time: 0.0209 Steps: 46240, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001541, Sample Num: 24656, Cur Loss: 0.15808740, Cur Avg Loss: 0.26190168, Log Avg loss: 0.35277610, Global Avg Loss: 1.14372986, Time: 0.0248 Steps: 46250, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001551, Sample Num: 24816, Cur Loss: 0.53532553, Cur Avg Loss: 0.26233457, Log Avg loss: 0.32904202, Global Avg Loss: 1.14355375, Time: 0.0212 Steps: 46260, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001561, Sample Num: 24976, Cur Loss: 0.25393310, Cur Avg Loss: 0.26246880, Log Avg loss: 0.28328780, Global Avg Loss: 1.14336782, Time: 0.0211 Steps: 46270, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001571, Sample Num: 25136, Cur Loss: 0.34674025, Cur Avg Loss: 0.26326649, Log Avg loss: 0.38778642, Global Avg Loss: 1.14320456, Time: 0.0211 Steps: 46280, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001581, Sample Num: 25296, Cur Loss: 0.20353445, Cur Avg Loss: 0.26340006, Log Avg loss: 0.28438348, Global Avg Loss: 1.14301903, Time: 0.0211 Steps: 46290, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001591, Sample Num: 25456, Cur Loss: 0.27118295, Cur Avg Loss: 0.26331737, Log Avg loss: 0.25024452, Global Avg Loss: 1.14282621, Time: 0.0211 Steps: 46300, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001601, Sample Num: 25616, Cur Loss: 0.09058806, Cur Avg Loss: 0.26306035, Log Avg loss: 0.22216898, Global Avg Loss: 1.14262740, Time: 0.0211 Steps: 46310, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001611, Sample Num: 25776, Cur Loss: 0.08754631, Cur Avg Loss: 0.26279205, Log Avg loss: 0.21983692, Global Avg Loss: 1.14242818, Time: 0.0211 Steps: 46320, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001621, Sample Num: 25936, Cur Loss: 0.08842363, Cur Avg Loss: 0.26248127, Log Avg loss: 0.21241464, Global Avg Loss: 1.14222745, Time: 0.0211 Steps: 46330, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001631, Sample Num: 26096, Cur Loss: 0.13103451, Cur Avg Loss: 0.26216581, Log Avg loss: 0.21102869, Global Avg Loss: 1.14202650, Time: 0.0211 Steps: 46340, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001641, Sample Num: 26256, Cur Loss: 0.13365093, Cur Avg Loss: 0.26213619, Log Avg loss: 0.25730623, Global Avg Loss: 1.14183562, Time: 0.0211 Steps: 46350, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001651, Sample Num: 26416, Cur Loss: 0.28953663, Cur Avg Loss: 0.26198491, Log Avg loss: 0.23715929, Global Avg Loss: 1.14164048, Time: 0.0212 Steps: 46360, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001661, Sample Num: 26576, Cur Loss: 0.19570009, Cur Avg Loss: 0.26131761, Log Avg loss: 0.15114624, Global Avg Loss: 1.14142687, Time: 0.0212 Steps: 46370, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001671, Sample Num: 26736, Cur Loss: 0.13019167, Cur Avg Loss: 0.26069548, Log Avg loss: 0.15735995, Global Avg Loss: 1.14121470, Time: 0.0211 Steps: 46380, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001681, Sample Num: 26896, Cur Loss: 0.41908586, Cur Avg Loss: 0.26112977, Log Avg loss: 0.33369891, Global Avg Loss: 1.14104062, Time: 0.0212 Steps: 46390, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001691, Sample Num: 27056, Cur Loss: 0.43206334, Cur Avg Loss: 0.26145223, Log Avg loss: 0.31565835, Global Avg Loss: 1.14086274, Time: 0.0211 Steps: 46400, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001701, Sample Num: 27216, Cur Loss: 0.13195774, Cur Avg Loss: 0.26136433, Log Avg loss: 0.24650132, Global Avg Loss: 1.14067003, Time: 0.0211 Steps: 46410, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001711, Sample Num: 27376, Cur Loss: 0.07101545, Cur Avg Loss: 0.26103625, Log Avg loss: 0.20522926, Global Avg Loss: 1.14046852, Time: 0.0211 Steps: 46420, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001721, Sample Num: 27536, Cur Loss: 0.26481825, Cur Avg Loss: 0.26074646, Log Avg loss: 0.21116274, Global Avg Loss: 1.14026836, Time: 0.0211 Steps: 46430, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001731, Sample Num: 27696, Cur Loss: 0.24546492, Cur Avg Loss: 0.26066794, Log Avg loss: 0.24715578, Global Avg Loss: 1.14007605, Time: 0.0211 Steps: 46440, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001741, Sample Num: 27856, Cur Loss: 0.20496069, Cur Avg Loss: 0.26045313, Log Avg loss: 0.22326802, Global Avg Loss: 1.13987867, Time: 0.0211 Steps: 46450, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001751, Sample Num: 28016, Cur Loss: 0.15778759, Cur Avg Loss: 0.26031920, Log Avg loss: 0.23700263, Global Avg Loss: 1.13968434, Time: 0.0211 Steps: 46460, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001761, Sample Num: 28176, Cur Loss: 0.17675859, Cur Avg Loss: 0.26070248, Log Avg loss: 0.32781418, Global Avg Loss: 1.13950963, Time: 0.0212 Steps: 46470, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001771, Sample Num: 28336, Cur Loss: 0.11610575, Cur Avg Loss: 0.26028004, Log Avg loss: 0.18588875, Global Avg Loss: 1.13930446, Time: 0.0211 Steps: 46480, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001781, Sample Num: 28496, Cur Loss: 0.13295822, Cur Avg Loss: 0.26044988, Log Avg loss: 0.29052949, Global Avg Loss: 1.13912189, Time: 0.0211 Steps: 46490, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001791, Sample Num: 28656, Cur Loss: 0.15899941, Cur Avg Loss: 0.25983752, Log Avg loss: 0.15077600, Global Avg Loss: 1.13890934, Time: 0.0211 Steps: 46500, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001801, Sample Num: 28816, Cur Loss: 0.17295299, Cur Avg Loss: 0.26010695, Log Avg loss: 0.30836091, Global Avg Loss: 1.13873077, Time: 0.0209 Steps: 46510, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001811, Sample Num: 28976, Cur Loss: 0.10830183, Cur Avg Loss: 0.26023988, Log Avg loss: 0.28418174, Global Avg Loss: 1.13854707, Time: 0.0209 Steps: 46520, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001821, Sample Num: 29136, Cur Loss: 0.24351709, Cur Avg Loss: 0.26009403, Log Avg loss: 0.23367943, Global Avg Loss: 1.13835260, Time: 0.0209 Steps: 46530, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001831, Sample Num: 29296, Cur Loss: 0.21524113, Cur Avg Loss: 0.25981362, Log Avg loss: 0.20875136, Global Avg Loss: 1.13815286, Time: 0.0209 Steps: 46540, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001841, Sample Num: 29456, Cur Loss: 0.75894403, Cur Avg Loss: 0.25997669, Log Avg loss: 0.28983581, Global Avg Loss: 1.13797062, Time: 0.0209 Steps: 46550, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001851, Sample Num: 29616, Cur Loss: 0.38201332, Cur Avg Loss: 0.26031389, Log Avg loss: 0.32239225, Global Avg Loss: 1.13779546, Time: 0.0209 Steps: 46560, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001861, Sample Num: 29776, Cur Loss: 0.41410354, Cur Avg Loss: 0.26086757, Log Avg loss: 0.36335348, Global Avg Loss: 1.13762916, Time: 0.0209 Steps: 46570, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001871, Sample Num: 29936, Cur Loss: 0.08694719, Cur Avg Loss: 0.26076357, Log Avg loss: 0.24140851, Global Avg Loss: 1.13743676, Time: 0.0209 Steps: 46580, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001881, Sample Num: 30096, Cur Loss: 0.11807904, Cur Avg Loss: 0.26044280, Log Avg loss: 0.20042760, Global Avg Loss: 1.13723564, Time: 0.0209 Steps: 46590, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001891, Sample Num: 30256, Cur Loss: 0.22046371, Cur Avg Loss: 0.26026575, Log Avg loss: 0.22696182, Global Avg Loss: 1.13704030, Time: 0.0209 Steps: 46600, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001901, Sample Num: 30416, Cur Loss: 0.29241943, Cur Avg Loss: 0.26040416, Log Avg loss: 0.28657825, Global Avg Loss: 1.13685784, Time: 0.0209 Steps: 46610, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001911, Sample Num: 30576, Cur Loss: 0.16144720, Cur Avg Loss: 0.26062429, Log Avg loss: 0.30247018, Global Avg Loss: 1.13667886, Time: 0.0209 Steps: 46620, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001921, Sample Num: 30736, Cur Loss: 0.35379118, Cur Avg Loss: 0.26111898, Log Avg loss: 0.35565423, Global Avg Loss: 1.13651137, Time: 0.0210 Steps: 46630, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001931, Sample Num: 30896, Cur Loss: 0.28998619, Cur Avg Loss: 0.26109709, Log Avg loss: 0.25689279, Global Avg Loss: 1.13632277, Time: 0.0208 Steps: 46640, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001941, Sample Num: 31056, Cur Loss: 0.82592392, Cur Avg Loss: 0.26181711, Log Avg loss: 0.40085194, Global Avg Loss: 1.13616511, Time: 0.0210 Steps: 46650, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001951, Sample Num: 31216, Cur Loss: 0.22871946, Cur Avg Loss: 0.26246360, Log Avg loss: 0.38794825, Global Avg Loss: 1.13600476, Time: 0.0209 Steps: 46660, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001961, Sample Num: 31376, Cur Loss: 0.08694087, Cur Avg Loss: 0.26304402, Log Avg loss: 0.37628296, Global Avg Loss: 1.13584197, Time: 0.0209 Steps: 46670, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001971, Sample Num: 31536, Cur Loss: 0.38401002, Cur Avg Loss: 0.26303147, Log Avg loss: 0.26057099, Global Avg Loss: 1.13565447, Time: 0.0209 Steps: 46680, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001981, Sample Num: 31696, Cur Loss: 0.15750659, Cur Avg Loss: 0.26318771, Log Avg loss: 0.29398218, Global Avg Loss: 1.13547420, Time: 0.0209 Steps: 46690, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001991, Sample Num: 31856, Cur Loss: 0.23186147, Cur Avg Loss: 0.26348927, Log Avg loss: 0.32322790, Global Avg Loss: 1.13530027, Time: 0.0209 Steps: 46700, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002001, Sample Num: 32016, Cur Loss: 0.45259586, Cur Avg Loss: 0.26357592, Log Avg loss: 0.28082964, Global Avg Loss: 1.13511734, Time: 0.0209 Steps: 46710, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002011, Sample Num: 32176, Cur Loss: 0.34621096, Cur Avg Loss: 0.26367287, Log Avg loss: 0.28307195, Global Avg Loss: 1.13493497, Time: 0.0210 Steps: 46720, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002021, Sample Num: 32336, Cur Loss: 0.11877464, Cur Avg Loss: 0.26342486, Log Avg loss: 0.21355058, Global Avg Loss: 1.13473779, Time: 0.0209 Steps: 46730, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002031, Sample Num: 32496, Cur Loss: 0.20328996, Cur Avg Loss: 0.26340430, Log Avg loss: 0.25924910, Global Avg Loss: 1.13455048, Time: 0.0209 Steps: 46740, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002041, Sample Num: 32656, Cur Loss: 0.44068789, Cur Avg Loss: 0.26315773, Log Avg loss: 0.21307822, Global Avg Loss: 1.13435338, Time: 0.0209 Steps: 46750, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002051, Sample Num: 32816, Cur Loss: 0.66459280, Cur Avg Loss: 0.26326531, Log Avg loss: 0.28522370, Global Avg Loss: 1.13417178, Time: 0.0245 Steps: 46760, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002061, Sample Num: 32976, Cur Loss: 0.05815616, Cur Avg Loss: 0.26272066, Log Avg loss: 0.15101201, Global Avg Loss: 1.13396157, Time: 0.0209 Steps: 46770, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002071, Sample Num: 33136, Cur Loss: 0.52943099, Cur Avg Loss: 0.26272156, Log Avg loss: 0.26290748, Global Avg Loss: 1.13377537, Time: 0.0208 Steps: 46780, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002081, Sample Num: 33296, Cur Loss: 0.03640854, Cur Avg Loss: 0.26231051, Log Avg loss: 0.17718118, Global Avg Loss: 1.13357093, Time: 0.0208 Steps: 46790, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002091, Sample Num: 33456, Cur Loss: 0.07017145, Cur Avg Loss: 0.26179851, Log Avg loss: 0.15525137, Global Avg Loss: 1.13336188, Time: 0.0209 Steps: 46800, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002101, Sample Num: 33616, Cur Loss: 0.10237860, Cur Avg Loss: 0.26171262, Log Avg loss: 0.24375291, Global Avg Loss: 1.13317184, Time: 0.0209 Steps: 46810, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002111, Sample Num: 33776, Cur Loss: 0.22927624, Cur Avg Loss: 0.26191369, Log Avg loss: 0.30416002, Global Avg Loss: 1.13299477, Time: 0.0210 Steps: 46820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002121, Sample Num: 33936, Cur Loss: 0.09843128, Cur Avg Loss: 0.26212502, Log Avg loss: 0.30673483, Global Avg Loss: 1.13281833, Time: 0.0208 Steps: 46830, Updated lr: 0.000057 ***** Running evaluation checkpoint-46838 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-46838 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.932739, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.28821, "eval_total_loss": 202.611713, "eval_mae": 0.368563, "eval_mse": 0.288316, "eval_r2": 0.816727, "eval_sp_statistic": 0.87754, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.906096, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.691406, "test_total_loss": 347.085633, "test_mae": 0.593979, "test_mse": 0.691539, "test_r2": 0.553675, "test_sp_statistic": 0.792858, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.83685, "test_ps_pvalue": 0.0, "lr": 5.653105737316264e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.132652778393319, "train_cur_epoch_loss": 557.2753924336284, "train_cur_epoch_avg_loss": 0.26175452909047836, "train_cur_epoch_time": 44.93273854255676, "train_cur_epoch_avg_time": 0.021105090907729808, "epoch": 22, "step": 46838} ################################################## Training, Epoch: 0023, Batch: 000002, Sample Num: 32, Cur Loss: 0.38333699, Cur Avg Loss: 0.23929211, Log Avg loss: 0.17868192, Global Avg Loss: 1.13261463, Time: 0.0247 Steps: 46840, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000012, Sample Num: 192, Cur Loss: 0.08645239, Cur Avg Loss: 0.23226080, Log Avg loss: 0.23085454, Global Avg Loss: 1.13242216, Time: 0.0208 Steps: 46850, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000022, Sample Num: 352, Cur Loss: 0.21496540, Cur Avg Loss: 0.26970584, Log Avg loss: 0.31463989, Global Avg Loss: 1.13224764, Time: 0.0209 Steps: 46860, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000032, Sample Num: 512, Cur Loss: 0.24619865, Cur Avg Loss: 0.26369944, Log Avg loss: 0.25048534, Global Avg Loss: 1.13205951, Time: 0.0208 Steps: 46870, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000042, Sample Num: 672, Cur Loss: 0.09714201, Cur Avg Loss: 0.25352111, Log Avg loss: 0.22095045, Global Avg Loss: 1.13186516, Time: 0.0208 Steps: 46880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000052, Sample Num: 832, Cur Loss: 0.07580990, Cur Avg Loss: 0.25153057, Log Avg loss: 0.24317034, Global Avg Loss: 1.13167563, Time: 0.0208 Steps: 46890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000062, Sample Num: 992, Cur Loss: 0.64575315, Cur Avg Loss: 0.26950952, Log Avg loss: 0.36300002, Global Avg Loss: 1.13151174, Time: 0.0208 Steps: 46900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000072, Sample Num: 1152, Cur Loss: 0.14355624, Cur Avg Loss: 0.27286198, Log Avg loss: 0.29364723, Global Avg Loss: 1.13133313, Time: 0.0208 Steps: 46910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000082, Sample Num: 1312, Cur Loss: 0.16219446, Cur Avg Loss: 0.26450753, Log Avg loss: 0.20435551, Global Avg Loss: 1.13113556, Time: 0.0208 Steps: 46920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000092, Sample Num: 1472, Cur Loss: 0.21769640, Cur Avg Loss: 0.26545491, Log Avg loss: 0.27322340, Global Avg Loss: 1.13095275, Time: 0.0208 Steps: 46930, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000102, Sample Num: 1632, Cur Loss: 0.19348982, Cur Avg Loss: 0.25743854, Log Avg loss: 0.18368800, Global Avg Loss: 1.13075095, Time: 0.0208 Steps: 46940, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000112, Sample Num: 1792, Cur Loss: 0.27786252, Cur Avg Loss: 0.25374959, Log Avg loss: 0.21612229, Global Avg Loss: 1.13055614, Time: 0.0208 Steps: 46950, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000122, Sample Num: 1952, Cur Loss: 0.04392011, Cur Avg Loss: 0.25200092, Log Avg loss: 0.23241575, Global Avg Loss: 1.13036488, Time: 0.0208 Steps: 46960, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000132, Sample Num: 2112, Cur Loss: 0.31690055, Cur Avg Loss: 0.25070737, Log Avg loss: 0.23492607, Global Avg Loss: 1.13017424, Time: 0.0208 Steps: 46970, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000142, Sample Num: 2272, Cur Loss: 0.20912370, Cur Avg Loss: 0.24709844, Log Avg loss: 0.19946065, Global Avg Loss: 1.12997613, Time: 0.0208 Steps: 46980, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000152, Sample Num: 2432, Cur Loss: 0.11869896, Cur Avg Loss: 0.24591810, Log Avg loss: 0.22915721, Global Avg Loss: 1.12978443, Time: 0.0208 Steps: 46990, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000162, Sample Num: 2592, Cur Loss: 0.10961550, Cur Avg Loss: 0.24502492, Log Avg loss: 0.23144867, Global Avg Loss: 1.12959330, Time: 0.0208 Steps: 47000, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000172, Sample Num: 2752, Cur Loss: 0.17817307, Cur Avg Loss: 0.24473694, Log Avg loss: 0.24007165, Global Avg Loss: 1.12940408, Time: 0.0208 Steps: 47010, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000182, Sample Num: 2912, Cur Loss: 0.09908025, Cur Avg Loss: 0.24553920, Log Avg loss: 0.25933802, Global Avg Loss: 1.12921903, Time: 0.0208 Steps: 47020, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000192, Sample Num: 3072, Cur Loss: 0.67208648, Cur Avg Loss: 0.24773839, Log Avg loss: 0.28776357, Global Avg Loss: 1.12904011, Time: 0.0209 Steps: 47030, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000202, Sample Num: 3232, Cur Loss: 0.12547229, Cur Avg Loss: 0.24804588, Log Avg loss: 0.25394983, Global Avg Loss: 1.12885408, Time: 0.0209 Steps: 47040, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000212, Sample Num: 3392, Cur Loss: 0.27419329, Cur Avg Loss: 0.24850184, Log Avg loss: 0.25771213, Global Avg Loss: 1.12866893, Time: 0.0208 Steps: 47050, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000222, Sample Num: 3552, Cur Loss: 0.15660138, Cur Avg Loss: 0.24761315, Log Avg loss: 0.22877286, Global Avg Loss: 1.12847771, Time: 0.0208 Steps: 47060, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000232, Sample Num: 3712, Cur Loss: 0.34111679, Cur Avg Loss: 0.24728272, Log Avg loss: 0.23994717, Global Avg Loss: 1.12828894, Time: 0.0208 Steps: 47070, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000242, Sample Num: 3872, Cur Loss: 0.23047173, Cur Avg Loss: 0.24669282, Log Avg loss: 0.23300719, Global Avg Loss: 1.12809878, Time: 0.0208 Steps: 47080, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000252, Sample Num: 4032, Cur Loss: 0.24728429, Cur Avg Loss: 0.24724848, Log Avg loss: 0.26069556, Global Avg Loss: 1.12791458, Time: 0.0208 Steps: 47090, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000262, Sample Num: 4192, Cur Loss: 0.11625081, Cur Avg Loss: 0.24499264, Log Avg loss: 0.18814552, Global Avg Loss: 1.12771505, Time: 0.0210 Steps: 47100, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000272, Sample Num: 4352, Cur Loss: 0.07718334, Cur Avg Loss: 0.24446271, Log Avg loss: 0.23057847, Global Avg Loss: 1.12752462, Time: 0.0209 Steps: 47110, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000282, Sample Num: 4512, Cur Loss: 0.16475171, Cur Avg Loss: 0.24612303, Log Avg loss: 0.29128359, Global Avg Loss: 1.12734715, Time: 0.0209 Steps: 47120, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000292, Sample Num: 4672, Cur Loss: 0.22147842, Cur Avg Loss: 0.24764597, Log Avg loss: 0.29059304, Global Avg Loss: 1.12716960, Time: 0.0208 Steps: 47130, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000302, Sample Num: 4832, Cur Loss: 0.32817703, Cur Avg Loss: 0.24753288, Log Avg loss: 0.24423074, Global Avg Loss: 1.12698230, Time: 0.0208 Steps: 47140, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000312, Sample Num: 4992, Cur Loss: 0.12545887, Cur Avg Loss: 0.24870289, Log Avg loss: 0.28403714, Global Avg Loss: 1.12680352, Time: 0.0209 Steps: 47150, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000322, Sample Num: 5152, Cur Loss: 0.26729968, Cur Avg Loss: 0.25109145, Log Avg loss: 0.32561455, Global Avg Loss: 1.12663364, Time: 0.0208 Steps: 47160, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000332, Sample Num: 5312, Cur Loss: 0.42798144, Cur Avg Loss: 0.25295034, Log Avg loss: 0.31280633, Global Avg Loss: 1.12646111, Time: 0.0208 Steps: 47170, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000342, Sample Num: 5472, Cur Loss: 0.09603667, Cur Avg Loss: 0.25253731, Log Avg loss: 0.23882485, Global Avg Loss: 1.12627297, Time: 0.0208 Steps: 47180, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000352, Sample Num: 5632, Cur Loss: 0.41517568, Cur Avg Loss: 0.25185196, Log Avg loss: 0.22841314, Global Avg Loss: 1.12608270, Time: 0.0208 Steps: 47190, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000362, Sample Num: 5792, Cur Loss: 0.28906393, Cur Avg Loss: 0.25252414, Log Avg loss: 0.27618466, Global Avg Loss: 1.12590264, Time: 0.0209 Steps: 47200, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000372, Sample Num: 5952, Cur Loss: 0.11647156, Cur Avg Loss: 0.25047834, Log Avg loss: 0.17642051, Global Avg Loss: 1.12570152, Time: 0.0208 Steps: 47210, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000382, Sample Num: 6112, Cur Loss: 0.15305857, Cur Avg Loss: 0.25057000, Log Avg loss: 0.25397977, Global Avg Loss: 1.12551691, Time: 0.0208 Steps: 47220, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000392, Sample Num: 6272, Cur Loss: 0.32372928, Cur Avg Loss: 0.25126583, Log Avg loss: 0.27784647, Global Avg Loss: 1.12533743, Time: 0.0210 Steps: 47230, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000402, Sample Num: 6432, Cur Loss: 0.28848162, Cur Avg Loss: 0.25168501, Log Avg loss: 0.26811679, Global Avg Loss: 1.12515597, Time: 0.0209 Steps: 47240, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000412, Sample Num: 6592, Cur Loss: 0.20792863, Cur Avg Loss: 0.25004397, Log Avg loss: 0.18407429, Global Avg Loss: 1.12495680, Time: 0.0210 Steps: 47250, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000422, Sample Num: 6752, Cur Loss: 0.16652581, Cur Avg Loss: 0.24965910, Log Avg loss: 0.23380228, Global Avg Loss: 1.12476824, Time: 0.0208 Steps: 47260, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000432, Sample Num: 6912, Cur Loss: 0.28172910, Cur Avg Loss: 0.24773518, Log Avg loss: 0.16654588, Global Avg Loss: 1.12456553, Time: 0.0209 Steps: 47270, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000442, Sample Num: 7072, Cur Loss: 0.45113331, Cur Avg Loss: 0.24893630, Log Avg loss: 0.30082462, Global Avg Loss: 1.12439130, Time: 0.0208 Steps: 47280, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000452, Sample Num: 7232, Cur Loss: 0.16130966, Cur Avg Loss: 0.25064126, Log Avg loss: 0.32600054, Global Avg Loss: 1.12422247, Time: 0.0209 Steps: 47290, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000462, Sample Num: 7392, Cur Loss: 0.53006721, Cur Avg Loss: 0.25174208, Log Avg loss: 0.30149938, Global Avg Loss: 1.12404853, Time: 0.0208 Steps: 47300, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000472, Sample Num: 7552, Cur Loss: 0.06396773, Cur Avg Loss: 0.25073153, Log Avg loss: 0.20404373, Global Avg Loss: 1.12385407, Time: 0.0209 Steps: 47310, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000482, Sample Num: 7712, Cur Loss: 0.09749999, Cur Avg Loss: 0.24971568, Log Avg loss: 0.20176766, Global Avg Loss: 1.12365921, Time: 0.0208 Steps: 47320, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000492, Sample Num: 7872, Cur Loss: 0.46129462, Cur Avg Loss: 0.25003743, Log Avg loss: 0.26554572, Global Avg Loss: 1.12347791, Time: 0.0208 Steps: 47330, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000502, Sample Num: 8032, Cur Loss: 0.16959688, Cur Avg Loss: 0.24935759, Log Avg loss: 0.21590943, Global Avg Loss: 1.12328619, Time: 0.0209 Steps: 47340, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000512, Sample Num: 8192, Cur Loss: 0.09460540, Cur Avg Loss: 0.24813375, Log Avg loss: 0.18669699, Global Avg Loss: 1.12308839, Time: 0.0253 Steps: 47350, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000522, Sample Num: 8352, Cur Loss: 0.25581688, Cur Avg Loss: 0.24789028, Log Avg loss: 0.23542482, Global Avg Loss: 1.12290096, Time: 0.0209 Steps: 47360, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000532, Sample Num: 8512, Cur Loss: 0.12818886, Cur Avg Loss: 0.24910313, Log Avg loss: 0.31241395, Global Avg Loss: 1.12272986, Time: 0.0208 Steps: 47370, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000542, Sample Num: 8672, Cur Loss: 0.05523346, Cur Avg Loss: 0.24845259, Log Avg loss: 0.21384362, Global Avg Loss: 1.12253804, Time: 0.0208 Steps: 47380, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000552, Sample Num: 8832, Cur Loss: 0.21533656, Cur Avg Loss: 0.24764332, Log Avg loss: 0.20378105, Global Avg Loss: 1.12234416, Time: 0.0208 Steps: 47390, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000562, Sample Num: 8992, Cur Loss: 0.15186438, Cur Avg Loss: 0.24680033, Log Avg loss: 0.20026749, Global Avg Loss: 1.12214963, Time: 0.0208 Steps: 47400, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000572, Sample Num: 9152, Cur Loss: 0.23515218, Cur Avg Loss: 0.24769409, Log Avg loss: 0.29792290, Global Avg Loss: 1.12197578, Time: 0.0209 Steps: 47410, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000582, Sample Num: 9312, Cur Loss: 0.25916988, Cur Avg Loss: 0.24796376, Log Avg loss: 0.26338917, Global Avg Loss: 1.12179472, Time: 0.0208 Steps: 47420, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000592, Sample Num: 9472, Cur Loss: 0.18760805, Cur Avg Loss: 0.24982218, Log Avg loss: 0.35798199, Global Avg Loss: 1.12163368, Time: 0.0208 Steps: 47430, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000602, Sample Num: 9632, Cur Loss: 0.08227551, Cur Avg Loss: 0.24962300, Log Avg loss: 0.23783156, Global Avg Loss: 1.12144738, Time: 0.0212 Steps: 47440, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000612, Sample Num: 9792, Cur Loss: 0.10289247, Cur Avg Loss: 0.24915770, Log Avg loss: 0.22114700, Global Avg Loss: 1.12125765, Time: 0.0211 Steps: 47450, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000622, Sample Num: 9952, Cur Loss: 0.32085252, Cur Avg Loss: 0.24867070, Log Avg loss: 0.21886619, Global Avg Loss: 1.12106751, Time: 0.0209 Steps: 47460, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000632, Sample Num: 10112, Cur Loss: 0.14781815, Cur Avg Loss: 0.24880353, Log Avg loss: 0.25706570, Global Avg Loss: 1.12088550, Time: 0.0209 Steps: 47470, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000642, Sample Num: 10272, Cur Loss: 0.33782816, Cur Avg Loss: 0.24918187, Log Avg loss: 0.27309295, Global Avg Loss: 1.12070694, Time: 0.0208 Steps: 47480, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000652, Sample Num: 10432, Cur Loss: 0.05369360, Cur Avg Loss: 0.24861058, Log Avg loss: 0.21193372, Global Avg Loss: 1.12051558, Time: 0.0209 Steps: 47490, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000662, Sample Num: 10592, Cur Loss: 0.52847427, Cur Avg Loss: 0.24854580, Log Avg loss: 0.24432186, Global Avg Loss: 1.12033112, Time: 0.0208 Steps: 47500, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000672, Sample Num: 10752, Cur Loss: 0.10853411, Cur Avg Loss: 0.24816876, Log Avg loss: 0.22320873, Global Avg Loss: 1.12014229, Time: 0.0208 Steps: 47510, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000682, Sample Num: 10912, Cur Loss: 0.12171640, Cur Avg Loss: 0.24875335, Log Avg loss: 0.28803803, Global Avg Loss: 1.11996718, Time: 0.0211 Steps: 47520, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000692, Sample Num: 11072, Cur Loss: 0.03244809, Cur Avg Loss: 0.24860761, Log Avg loss: 0.23866829, Global Avg Loss: 1.11978177, Time: 0.0210 Steps: 47530, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000702, Sample Num: 11232, Cur Loss: 0.05538257, Cur Avg Loss: 0.24904698, Log Avg loss: 0.27945134, Global Avg Loss: 1.11960500, Time: 0.0211 Steps: 47540, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000712, Sample Num: 11392, Cur Loss: 0.12592442, Cur Avg Loss: 0.24844153, Log Avg loss: 0.20593891, Global Avg Loss: 1.11941285, Time: 0.0210 Steps: 47550, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000722, Sample Num: 11552, Cur Loss: 0.20414436, Cur Avg Loss: 0.24767775, Log Avg loss: 0.19329663, Global Avg Loss: 1.11921813, Time: 0.0208 Steps: 47560, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000732, Sample Num: 11712, Cur Loss: 0.16896579, Cur Avg Loss: 0.24897579, Log Avg loss: 0.34269393, Global Avg Loss: 1.11905489, Time: 0.0208 Steps: 47570, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000742, Sample Num: 11872, Cur Loss: 0.48069900, Cur Avg Loss: 0.24979937, Log Avg loss: 0.31008594, Global Avg Loss: 1.11888487, Time: 0.0208 Steps: 47580, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000752, Sample Num: 12032, Cur Loss: 0.06002587, Cur Avg Loss: 0.24879648, Log Avg loss: 0.17438195, Global Avg Loss: 1.11868640, Time: 0.0208 Steps: 47590, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000762, Sample Num: 12192, Cur Loss: 0.08130921, Cur Avg Loss: 0.24816359, Log Avg loss: 0.20057023, Global Avg Loss: 1.11849352, Time: 0.0208 Steps: 47600, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000772, Sample Num: 12352, Cur Loss: 0.21685669, Cur Avg Loss: 0.24800780, Log Avg loss: 0.23613657, Global Avg Loss: 1.11830819, Time: 0.0227 Steps: 47610, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000782, Sample Num: 12512, Cur Loss: 0.27977538, Cur Avg Loss: 0.24877883, Log Avg loss: 0.30830253, Global Avg Loss: 1.11813809, Time: 0.0208 Steps: 47620, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000792, Sample Num: 12672, Cur Loss: 0.34911621, Cur Avg Loss: 0.24911474, Log Avg loss: 0.27538269, Global Avg Loss: 1.11796115, Time: 0.0208 Steps: 47630, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000802, Sample Num: 12832, Cur Loss: 0.13592067, Cur Avg Loss: 0.24878581, Log Avg loss: 0.22273474, Global Avg Loss: 1.11777324, Time: 0.0208 Steps: 47640, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000812, Sample Num: 12992, Cur Loss: 0.38099891, Cur Avg Loss: 0.24896338, Log Avg loss: 0.26320424, Global Avg Loss: 1.11759390, Time: 0.0208 Steps: 47650, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000822, Sample Num: 13152, Cur Loss: 0.18108463, Cur Avg Loss: 0.24849653, Log Avg loss: 0.21058856, Global Avg Loss: 1.11740359, Time: 0.0208 Steps: 47660, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000832, Sample Num: 13312, Cur Loss: 0.25203377, Cur Avg Loss: 0.24857091, Log Avg loss: 0.25468491, Global Avg Loss: 1.11722261, Time: 0.0208 Steps: 47670, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000842, Sample Num: 13472, Cur Loss: 0.10987879, Cur Avg Loss: 0.24842277, Log Avg loss: 0.23609733, Global Avg Loss: 1.11703781, Time: 0.0208 Steps: 47680, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000852, Sample Num: 13632, Cur Loss: 0.14148137, Cur Avg Loss: 0.24798993, Log Avg loss: 0.21154505, Global Avg Loss: 1.11684794, Time: 0.0208 Steps: 47690, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000862, Sample Num: 13792, Cur Loss: 0.42003685, Cur Avg Loss: 0.24809354, Log Avg loss: 0.25692088, Global Avg Loss: 1.11666766, Time: 0.0208 Steps: 47700, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000872, Sample Num: 13952, Cur Loss: 0.35284263, Cur Avg Loss: 0.24768895, Log Avg loss: 0.21281341, Global Avg Loss: 1.11647821, Time: 0.0208 Steps: 47710, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000882, Sample Num: 14112, Cur Loss: 0.20997182, Cur Avg Loss: 0.24852809, Log Avg loss: 0.32170067, Global Avg Loss: 1.11631166, Time: 0.0208 Steps: 47720, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000892, Sample Num: 14272, Cur Loss: 0.46044689, Cur Avg Loss: 0.24873442, Log Avg loss: 0.26693283, Global Avg Loss: 1.11613371, Time: 0.0208 Steps: 47730, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000902, Sample Num: 14432, Cur Loss: 0.18523583, Cur Avg Loss: 0.24876600, Log Avg loss: 0.25158260, Global Avg Loss: 1.11595261, Time: 0.0208 Steps: 47740, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000912, Sample Num: 14592, Cur Loss: 0.19595575, Cur Avg Loss: 0.24834015, Log Avg loss: 0.20992900, Global Avg Loss: 1.11576287, Time: 0.0207 Steps: 47750, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000922, Sample Num: 14752, Cur Loss: 0.09093431, Cur Avg Loss: 0.24774733, Log Avg loss: 0.19368177, Global Avg Loss: 1.11556980, Time: 0.0208 Steps: 47760, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000932, Sample Num: 14912, Cur Loss: 0.85396075, Cur Avg Loss: 0.24872037, Log Avg loss: 0.33843499, Global Avg Loss: 1.11540712, Time: 0.0208 Steps: 47770, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000942, Sample Num: 15072, Cur Loss: 0.12240658, Cur Avg Loss: 0.24841428, Log Avg loss: 0.21988621, Global Avg Loss: 1.11521970, Time: 0.0208 Steps: 47780, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000952, Sample Num: 15232, Cur Loss: 0.10778707, Cur Avg Loss: 0.24806113, Log Avg loss: 0.21479451, Global Avg Loss: 1.11503128, Time: 0.0208 Steps: 47790, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000962, Sample Num: 15392, Cur Loss: 0.36014420, Cur Avg Loss: 0.24763840, Log Avg loss: 0.20739503, Global Avg Loss: 1.11484140, Time: 0.0208 Steps: 47800, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000972, Sample Num: 15552, Cur Loss: 0.21080390, Cur Avg Loss: 0.24712068, Log Avg loss: 0.19731545, Global Avg Loss: 1.11464949, Time: 0.0208 Steps: 47810, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000982, Sample Num: 15712, Cur Loss: 0.11270205, Cur Avg Loss: 0.24710065, Log Avg loss: 0.24515427, Global Avg Loss: 1.11446766, Time: 0.0208 Steps: 47820, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000992, Sample Num: 15872, Cur Loss: 0.18135557, Cur Avg Loss: 0.24715970, Log Avg loss: 0.25295814, Global Avg Loss: 1.11428754, Time: 0.0207 Steps: 47830, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001002, Sample Num: 16032, Cur Loss: 0.15300870, Cur Avg Loss: 0.24758188, Log Avg loss: 0.28946245, Global Avg Loss: 1.11411513, Time: 0.0208 Steps: 47840, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001012, Sample Num: 16192, Cur Loss: 0.20804861, Cur Avg Loss: 0.24706209, Log Avg loss: 0.19497903, Global Avg Loss: 1.11392304, Time: 0.0208 Steps: 47850, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001022, Sample Num: 16352, Cur Loss: 0.20799571, Cur Avg Loss: 0.24713166, Log Avg loss: 0.25417162, Global Avg Loss: 1.11374341, Time: 0.0208 Steps: 47860, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001032, Sample Num: 16512, Cur Loss: 0.32251471, Cur Avg Loss: 0.24732422, Log Avg loss: 0.26700461, Global Avg Loss: 1.11356652, Time: 0.0209 Steps: 47870, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001042, Sample Num: 16672, Cur Loss: 0.08808783, Cur Avg Loss: 0.24636999, Log Avg loss: 0.14789265, Global Avg Loss: 1.11336484, Time: 0.0208 Steps: 47880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001052, Sample Num: 16832, Cur Loss: 0.16833311, Cur Avg Loss: 0.24566552, Log Avg loss: 0.17226006, Global Avg Loss: 1.11316832, Time: 0.0208 Steps: 47890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001062, Sample Num: 16992, Cur Loss: 0.15365526, Cur Avg Loss: 0.24589225, Log Avg loss: 0.26974406, Global Avg Loss: 1.11299224, Time: 0.0208 Steps: 47900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001072, Sample Num: 17152, Cur Loss: 0.65259075, Cur Avg Loss: 0.24681345, Log Avg loss: 0.34464481, Global Avg Loss: 1.11283187, Time: 0.0208 Steps: 47910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001082, Sample Num: 17312, Cur Loss: 0.28733006, Cur Avg Loss: 0.24676505, Log Avg loss: 0.24157720, Global Avg Loss: 1.11265005, Time: 0.0208 Steps: 47920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001092, Sample Num: 17472, Cur Loss: 0.09043639, Cur Avg Loss: 0.24670155, Log Avg loss: 0.23983084, Global Avg Loss: 1.11246795, Time: 0.0208 Steps: 47930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001102, Sample Num: 17632, Cur Loss: 0.15788853, Cur Avg Loss: 0.24658879, Log Avg loss: 0.23427500, Global Avg Loss: 1.11228477, Time: 0.0208 Steps: 47940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001112, Sample Num: 17792, Cur Loss: 0.31143790, Cur Avg Loss: 0.24670969, Log Avg loss: 0.26003320, Global Avg Loss: 1.11210703, Time: 0.0207 Steps: 47950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001122, Sample Num: 17952, Cur Loss: 0.08058451, Cur Avg Loss: 0.24672335, Log Avg loss: 0.24824203, Global Avg Loss: 1.11192691, Time: 0.0208 Steps: 47960, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001132, Sample Num: 18112, Cur Loss: 0.18834421, Cur Avg Loss: 0.24603003, Log Avg loss: 0.16823967, Global Avg Loss: 1.11173018, Time: 0.0208 Steps: 47970, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001142, Sample Num: 18272, Cur Loss: 0.13742201, Cur Avg Loss: 0.24574203, Log Avg loss: 0.21314061, Global Avg Loss: 1.11154290, Time: 0.0208 Steps: 47980, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001152, Sample Num: 18432, Cur Loss: 0.31107521, Cur Avg Loss: 0.24612042, Log Avg loss: 0.28933287, Global Avg Loss: 1.11137157, Time: 0.0208 Steps: 47990, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001162, Sample Num: 18592, Cur Loss: 0.25755075, Cur Avg Loss: 0.24533738, Log Avg loss: 0.15513115, Global Avg Loss: 1.11117235, Time: 0.0207 Steps: 48000, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001172, Sample Num: 18752, Cur Loss: 0.26314923, Cur Avg Loss: 0.24493244, Log Avg loss: 0.19787844, Global Avg Loss: 1.11098212, Time: 0.0208 Steps: 48010, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001182, Sample Num: 18912, Cur Loss: 0.11156883, Cur Avg Loss: 0.24486866, Log Avg loss: 0.23739309, Global Avg Loss: 1.11080020, Time: 0.0207 Steps: 48020, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001192, Sample Num: 19072, Cur Loss: 0.16951981, Cur Avg Loss: 0.24472416, Log Avg loss: 0.22764465, Global Avg Loss: 1.11061632, Time: 0.0208 Steps: 48030, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001202, Sample Num: 19232, Cur Loss: 0.31092080, Cur Avg Loss: 0.24537239, Log Avg loss: 0.32264154, Global Avg Loss: 1.11045230, Time: 0.0208 Steps: 48040, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001212, Sample Num: 19392, Cur Loss: 0.54975647, Cur Avg Loss: 0.24598629, Log Avg loss: 0.31977644, Global Avg Loss: 1.11028775, Time: 0.0208 Steps: 48050, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001222, Sample Num: 19552, Cur Loss: 0.30801982, Cur Avg Loss: 0.24740312, Log Avg loss: 0.41912303, Global Avg Loss: 1.11014393, Time: 0.0208 Steps: 48060, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001232, Sample Num: 19712, Cur Loss: 0.46942139, Cur Avg Loss: 0.24769640, Log Avg loss: 0.28353493, Global Avg Loss: 1.10997197, Time: 0.0208 Steps: 48070, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001242, Sample Num: 19872, Cur Loss: 0.17370713, Cur Avg Loss: 0.24848798, Log Avg loss: 0.34601067, Global Avg Loss: 1.10981308, Time: 0.0208 Steps: 48080, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001252, Sample Num: 20032, Cur Loss: 0.26980367, Cur Avg Loss: 0.24899945, Log Avg loss: 0.31252441, Global Avg Loss: 1.10964729, Time: 0.0207 Steps: 48090, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001262, Sample Num: 20192, Cur Loss: 0.19802710, Cur Avg Loss: 0.24908592, Log Avg loss: 0.25991136, Global Avg Loss: 1.10947063, Time: 0.0208 Steps: 48100, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001272, Sample Num: 20352, Cur Loss: 0.24544173, Cur Avg Loss: 0.24873975, Log Avg loss: 0.20505346, Global Avg Loss: 1.10928264, Time: 0.0208 Steps: 48110, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001282, Sample Num: 20512, Cur Loss: 0.21524855, Cur Avg Loss: 0.24856011, Log Avg loss: 0.22571047, Global Avg Loss: 1.10909902, Time: 0.0247 Steps: 48120, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001292, Sample Num: 20672, Cur Loss: 0.29438412, Cur Avg Loss: 0.24868756, Log Avg loss: 0.26502571, Global Avg Loss: 1.10892365, Time: 0.0209 Steps: 48130, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001302, Sample Num: 20832, Cur Loss: 0.34138611, Cur Avg Loss: 0.24859887, Log Avg loss: 0.23714101, Global Avg Loss: 1.10874255, Time: 0.0208 Steps: 48140, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001312, Sample Num: 20992, Cur Loss: 0.36901969, Cur Avg Loss: 0.24955556, Log Avg loss: 0.37411593, Global Avg Loss: 1.10858998, Time: 0.0208 Steps: 48150, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001322, Sample Num: 21152, Cur Loss: 0.24577716, Cur Avg Loss: 0.24950986, Log Avg loss: 0.24351430, Global Avg Loss: 1.10841036, Time: 0.0208 Steps: 48160, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001332, Sample Num: 21312, Cur Loss: 0.28514528, Cur Avg Loss: 0.24941289, Log Avg loss: 0.23659395, Global Avg Loss: 1.10822937, Time: 0.0208 Steps: 48170, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001342, Sample Num: 21472, Cur Loss: 0.48327482, Cur Avg Loss: 0.24920506, Log Avg loss: 0.22152173, Global Avg Loss: 1.10804533, Time: 0.0208 Steps: 48180, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001352, Sample Num: 21632, Cur Loss: 0.10837186, Cur Avg Loss: 0.24852237, Log Avg loss: 0.15690533, Global Avg Loss: 1.10784796, Time: 0.0208 Steps: 48190, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001362, Sample Num: 21792, Cur Loss: 0.27501813, Cur Avg Loss: 0.24861107, Log Avg loss: 0.26060261, Global Avg Loss: 1.10767218, Time: 0.0208 Steps: 48200, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001372, Sample Num: 21952, Cur Loss: 0.25825998, Cur Avg Loss: 0.24809996, Log Avg loss: 0.17848727, Global Avg Loss: 1.10747944, Time: 0.0207 Steps: 48210, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001382, Sample Num: 22112, Cur Loss: 0.29300815, Cur Avg Loss: 0.24816361, Log Avg loss: 0.25689690, Global Avg Loss: 1.10730305, Time: 0.0208 Steps: 48220, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001392, Sample Num: 22272, Cur Loss: 0.19287854, Cur Avg Loss: 0.24814041, Log Avg loss: 0.24493414, Global Avg Loss: 1.10712424, Time: 0.0208 Steps: 48230, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001402, Sample Num: 22432, Cur Loss: 0.47915572, Cur Avg Loss: 0.24800737, Log Avg loss: 0.22948793, Global Avg Loss: 1.10694231, Time: 0.0208 Steps: 48240, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001412, Sample Num: 22592, Cur Loss: 0.10426685, Cur Avg Loss: 0.24772574, Log Avg loss: 0.20824148, Global Avg Loss: 1.10675605, Time: 0.0208 Steps: 48250, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001422, Sample Num: 22752, Cur Loss: 0.36239967, Cur Avg Loss: 0.24812751, Log Avg loss: 0.30485765, Global Avg Loss: 1.10658989, Time: 0.0208 Steps: 48260, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001432, Sample Num: 22912, Cur Loss: 0.07054086, Cur Avg Loss: 0.24795702, Log Avg loss: 0.22371321, Global Avg Loss: 1.10640699, Time: 0.0207 Steps: 48270, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001442, Sample Num: 23072, Cur Loss: 0.27265811, Cur Avg Loss: 0.24844249, Log Avg loss: 0.31796093, Global Avg Loss: 1.10624368, Time: 0.0208 Steps: 48280, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001452, Sample Num: 23232, Cur Loss: 0.28505784, Cur Avg Loss: 0.24858296, Log Avg loss: 0.26883892, Global Avg Loss: 1.10607027, Time: 0.0207 Steps: 48290, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001462, Sample Num: 23392, Cur Loss: 0.20386487, Cur Avg Loss: 0.24866379, Log Avg loss: 0.26040111, Global Avg Loss: 1.10589518, Time: 0.0207 Steps: 48300, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001472, Sample Num: 23552, Cur Loss: 0.10056928, Cur Avg Loss: 0.24912884, Log Avg loss: 0.31711887, Global Avg Loss: 1.10573191, Time: 0.0208 Steps: 48310, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001482, Sample Num: 23712, Cur Loss: 0.41764224, Cur Avg Loss: 0.24932905, Log Avg loss: 0.27879974, Global Avg Loss: 1.10556077, Time: 0.0208 Steps: 48320, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001492, Sample Num: 23872, Cur Loss: 0.25844163, Cur Avg Loss: 0.24904611, Log Avg loss: 0.20711414, Global Avg Loss: 1.10537487, Time: 0.0208 Steps: 48330, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001502, Sample Num: 24032, Cur Loss: 0.10276750, Cur Avg Loss: 0.24898796, Log Avg loss: 0.24031217, Global Avg Loss: 1.10519592, Time: 0.0208 Steps: 48340, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001512, Sample Num: 24192, Cur Loss: 0.14377686, Cur Avg Loss: 0.24892636, Log Avg loss: 0.23967451, Global Avg Loss: 1.10501691, Time: 0.0208 Steps: 48350, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001522, Sample Num: 24352, Cur Loss: 0.13626999, Cur Avg Loss: 0.24849819, Log Avg loss: 0.18375911, Global Avg Loss: 1.10482641, Time: 0.0208 Steps: 48360, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001532, Sample Num: 24512, Cur Loss: 0.30820587, Cur Avg Loss: 0.24826913, Log Avg loss: 0.21340521, Global Avg Loss: 1.10464212, Time: 0.0208 Steps: 48370, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001542, Sample Num: 24672, Cur Loss: 0.06641681, Cur Avg Loss: 0.24814626, Log Avg loss: 0.22932319, Global Avg Loss: 1.10446119, Time: 0.0237 Steps: 48380, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001552, Sample Num: 24832, Cur Loss: 0.24063879, Cur Avg Loss: 0.24816565, Log Avg loss: 0.25115604, Global Avg Loss: 1.10428485, Time: 0.0208 Steps: 48390, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001562, Sample Num: 24992, Cur Loss: 0.45178166, Cur Avg Loss: 0.24826037, Log Avg loss: 0.26296035, Global Avg Loss: 1.10411102, Time: 0.0208 Steps: 48400, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001572, Sample Num: 25152, Cur Loss: 0.13963880, Cur Avg Loss: 0.24846984, Log Avg loss: 0.28118883, Global Avg Loss: 1.10394103, Time: 0.0208 Steps: 48410, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001582, Sample Num: 25312, Cur Loss: 0.12840590, Cur Avg Loss: 0.24850333, Log Avg loss: 0.25376737, Global Avg Loss: 1.10376545, Time: 0.0208 Steps: 48420, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001592, Sample Num: 25472, Cur Loss: 0.11921370, Cur Avg Loss: 0.24812757, Log Avg loss: 0.18868284, Global Avg Loss: 1.10357650, Time: 0.0208 Steps: 48430, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001602, Sample Num: 25632, Cur Loss: 0.55826551, Cur Avg Loss: 0.24816809, Log Avg loss: 0.25461837, Global Avg Loss: 1.10340124, Time: 0.0208 Steps: 48440, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001612, Sample Num: 25792, Cur Loss: 0.27478033, Cur Avg Loss: 0.24776450, Log Avg loss: 0.18311003, Global Avg Loss: 1.10321129, Time: 0.0208 Steps: 48450, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001622, Sample Num: 25952, Cur Loss: 0.16288570, Cur Avg Loss: 0.24755530, Log Avg loss: 0.21383228, Global Avg Loss: 1.10302777, Time: 0.0208 Steps: 48460, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001632, Sample Num: 26112, Cur Loss: 0.20591907, Cur Avg Loss: 0.24815155, Log Avg loss: 0.34486324, Global Avg Loss: 1.10287135, Time: 0.0209 Steps: 48470, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001642, Sample Num: 26272, Cur Loss: 0.26002920, Cur Avg Loss: 0.24816038, Log Avg loss: 0.24960089, Global Avg Loss: 1.10269534, Time: 0.0208 Steps: 48480, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001652, Sample Num: 26432, Cur Loss: 0.51916945, Cur Avg Loss: 0.24853902, Log Avg loss: 0.31071202, Global Avg Loss: 1.10253201, Time: 0.0208 Steps: 48490, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001662, Sample Num: 26592, Cur Loss: 0.14179149, Cur Avg Loss: 0.24842286, Log Avg loss: 0.22923420, Global Avg Loss: 1.10235195, Time: 0.0208 Steps: 48500, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001672, Sample Num: 26752, Cur Loss: 0.28374714, Cur Avg Loss: 0.24779242, Log Avg loss: 0.14301280, Global Avg Loss: 1.10215419, Time: 0.0208 Steps: 48510, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001682, Sample Num: 26912, Cur Loss: 0.12731345, Cur Avg Loss: 0.24736194, Log Avg loss: 0.17538477, Global Avg Loss: 1.10196318, Time: 0.0208 Steps: 48520, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001692, Sample Num: 27072, Cur Loss: 0.28102222, Cur Avg Loss: 0.24744622, Log Avg loss: 0.26162303, Global Avg Loss: 1.10179002, Time: 0.0208 Steps: 48530, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001702, Sample Num: 27232, Cur Loss: 0.20507175, Cur Avg Loss: 0.24699159, Log Avg loss: 0.17006865, Global Avg Loss: 1.10159807, Time: 0.0208 Steps: 48540, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001712, Sample Num: 27392, Cur Loss: 0.17467441, Cur Avg Loss: 0.24683748, Log Avg loss: 0.22060806, Global Avg Loss: 1.10141661, Time: 0.0209 Steps: 48550, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001722, Sample Num: 27552, Cur Loss: 0.10191135, Cur Avg Loss: 0.24675655, Log Avg loss: 0.23290092, Global Avg Loss: 1.10123776, Time: 0.0208 Steps: 48560, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001732, Sample Num: 27712, Cur Loss: 0.30144852, Cur Avg Loss: 0.24649004, Log Avg loss: 0.20059687, Global Avg Loss: 1.10105233, Time: 0.0209 Steps: 48570, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001742, Sample Num: 27872, Cur Loss: 0.27003247, Cur Avg Loss: 0.24677232, Log Avg loss: 0.29566220, Global Avg Loss: 1.10088654, Time: 0.0208 Steps: 48580, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001752, Sample Num: 28032, Cur Loss: 0.05099675, Cur Avg Loss: 0.24665305, Log Avg loss: 0.22587636, Global Avg Loss: 1.10070646, Time: 0.0208 Steps: 48590, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001762, Sample Num: 28192, Cur Loss: 0.30487746, Cur Avg Loss: 0.24639462, Log Avg loss: 0.20111856, Global Avg Loss: 1.10052136, Time: 0.0208 Steps: 48600, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001772, Sample Num: 28352, Cur Loss: 0.47865549, Cur Avg Loss: 0.24694152, Log Avg loss: 0.34330569, Global Avg Loss: 1.10036559, Time: 0.0208 Steps: 48610, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001782, Sample Num: 28512, Cur Loss: 0.16000214, Cur Avg Loss: 0.24701791, Log Avg loss: 0.26055288, Global Avg Loss: 1.10019286, Time: 0.0208 Steps: 48620, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001792, Sample Num: 28672, Cur Loss: 0.16612747, Cur Avg Loss: 0.24685450, Log Avg loss: 0.21773626, Global Avg Loss: 1.10001139, Time: 0.0253 Steps: 48630, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001802, Sample Num: 28832, Cur Loss: 0.05268794, Cur Avg Loss: 0.24676273, Log Avg loss: 0.23031739, Global Avg Loss: 1.09983259, Time: 0.0209 Steps: 48640, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001812, Sample Num: 28992, Cur Loss: 0.23738775, Cur Avg Loss: 0.24712706, Log Avg loss: 0.31277916, Global Avg Loss: 1.09967081, Time: 0.0209 Steps: 48650, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001822, Sample Num: 29152, Cur Loss: 0.49391419, Cur Avg Loss: 0.24740639, Log Avg loss: 0.29802097, Global Avg Loss: 1.09950607, Time: 0.0209 Steps: 48660, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001832, Sample Num: 29312, Cur Loss: 0.09482181, Cur Avg Loss: 0.24727481, Log Avg loss: 0.22330018, Global Avg Loss: 1.09932604, Time: 0.0209 Steps: 48670, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001842, Sample Num: 29472, Cur Loss: 0.06696800, Cur Avg Loss: 0.24702504, Log Avg loss: 0.20126751, Global Avg Loss: 1.09914156, Time: 0.0209 Steps: 48680, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001852, Sample Num: 29632, Cur Loss: 0.25388691, Cur Avg Loss: 0.24694706, Log Avg loss: 0.23258385, Global Avg Loss: 1.09896358, Time: 0.0209 Steps: 48690, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001862, Sample Num: 29792, Cur Loss: 0.44757080, Cur Avg Loss: 0.24686553, Log Avg loss: 0.23176576, Global Avg Loss: 1.09878551, Time: 0.0209 Steps: 48700, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001872, Sample Num: 29952, Cur Loss: 0.21887064, Cur Avg Loss: 0.24684981, Log Avg loss: 0.24392305, Global Avg Loss: 1.09861001, Time: 0.0209 Steps: 48710, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001882, Sample Num: 30112, Cur Loss: 0.29008263, Cur Avg Loss: 0.24688179, Log Avg loss: 0.25286728, Global Avg Loss: 1.09843642, Time: 0.0209 Steps: 48720, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001892, Sample Num: 30272, Cur Loss: 0.26442984, Cur Avg Loss: 0.24684681, Log Avg loss: 0.24026413, Global Avg Loss: 1.09826031, Time: 0.0208 Steps: 48730, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001902, Sample Num: 30432, Cur Loss: 0.42054573, Cur Avg Loss: 0.24737328, Log Avg loss: 0.34698205, Global Avg Loss: 1.09810617, Time: 0.0209 Steps: 48740, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001912, Sample Num: 30592, Cur Loss: 0.60441840, Cur Avg Loss: 0.24724454, Log Avg loss: 0.22275837, Global Avg Loss: 1.09792661, Time: 0.0209 Steps: 48750, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001922, Sample Num: 30752, Cur Loss: 0.08619599, Cur Avg Loss: 0.24731873, Log Avg loss: 0.26150272, Global Avg Loss: 1.09775507, Time: 0.0209 Steps: 48760, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001932, Sample Num: 30912, Cur Loss: 0.18173580, Cur Avg Loss: 0.24710780, Log Avg loss: 0.20656685, Global Avg Loss: 1.09757234, Time: 0.0209 Steps: 48770, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001942, Sample Num: 31072, Cur Loss: 0.28727758, Cur Avg Loss: 0.24746733, Log Avg loss: 0.31692953, Global Avg Loss: 1.09741231, Time: 0.0208 Steps: 48780, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001952, Sample Num: 31232, Cur Loss: 0.26086223, Cur Avg Loss: 0.24780916, Log Avg loss: 0.31419288, Global Avg Loss: 1.09725178, Time: 0.0208 Steps: 48790, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001962, Sample Num: 31392, Cur Loss: 0.21776789, Cur Avg Loss: 0.24806150, Log Avg loss: 0.29731738, Global Avg Loss: 1.09708786, Time: 0.0209 Steps: 48800, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001972, Sample Num: 31552, Cur Loss: 0.09729278, Cur Avg Loss: 0.24850991, Log Avg loss: 0.33648756, Global Avg Loss: 1.09693203, Time: 0.0209 Steps: 48810, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001982, Sample Num: 31712, Cur Loss: 0.31180441, Cur Avg Loss: 0.24862543, Log Avg loss: 0.27140596, Global Avg Loss: 1.09676293, Time: 0.0208 Steps: 48820, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001992, Sample Num: 31872, Cur Loss: 0.94156587, Cur Avg Loss: 0.24929824, Log Avg loss: 0.38264915, Global Avg Loss: 1.09661669, Time: 0.0209 Steps: 48830, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002002, Sample Num: 32032, Cur Loss: 0.15226284, Cur Avg Loss: 0.24931670, Log Avg loss: 0.25299428, Global Avg Loss: 1.09644396, Time: 0.0209 Steps: 48840, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002012, Sample Num: 32192, Cur Loss: 0.07764748, Cur Avg Loss: 0.24897249, Log Avg loss: 0.18006272, Global Avg Loss: 1.09625637, Time: 0.0209 Steps: 48850, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002022, Sample Num: 32352, Cur Loss: 0.10689908, Cur Avg Loss: 0.24927213, Log Avg loss: 0.30955848, Global Avg Loss: 1.09609536, Time: 0.0209 Steps: 48860, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002032, Sample Num: 32512, Cur Loss: 0.14243691, Cur Avg Loss: 0.24987482, Log Avg loss: 0.37173918, Global Avg Loss: 1.09594713, Time: 0.0210 Steps: 48870, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002042, Sample Num: 32672, Cur Loss: 0.59635168, Cur Avg Loss: 0.24991202, Log Avg loss: 0.25747164, Global Avg Loss: 1.09577560, Time: 0.0208 Steps: 48880, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002052, Sample Num: 32832, Cur Loss: 0.21540792, Cur Avg Loss: 0.24958210, Log Avg loss: 0.18221181, Global Avg Loss: 1.09558874, Time: 0.0246 Steps: 48890, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002062, Sample Num: 32992, Cur Loss: 0.11535975, Cur Avg Loss: 0.24991596, Log Avg loss: 0.31842388, Global Avg Loss: 1.09542981, Time: 0.0208 Steps: 48900, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002072, Sample Num: 33152, Cur Loss: 0.07945565, Cur Avg Loss: 0.24939133, Log Avg loss: 0.14121220, Global Avg Loss: 1.09523471, Time: 0.0208 Steps: 48910, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002082, Sample Num: 33312, Cur Loss: 0.17843354, Cur Avg Loss: 0.24869914, Log Avg loss: 0.10527761, Global Avg Loss: 1.09503235, Time: 0.0208 Steps: 48920, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002092, Sample Num: 33472, Cur Loss: 0.54019737, Cur Avg Loss: 0.24874865, Log Avg loss: 0.25905670, Global Avg Loss: 1.09486150, Time: 0.0208 Steps: 48930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002102, Sample Num: 33632, Cur Loss: 0.06805521, Cur Avg Loss: 0.24884049, Log Avg loss: 0.26805432, Global Avg Loss: 1.09469255, Time: 0.0208 Steps: 48940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002112, Sample Num: 33792, Cur Loss: 0.44459617, Cur Avg Loss: 0.24896894, Log Avg loss: 0.27596807, Global Avg Loss: 1.09452530, Time: 0.0208 Steps: 48950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002122, Sample Num: 33952, Cur Loss: 0.09713708, Cur Avg Loss: 0.24870394, Log Avg loss: 0.19273554, Global Avg Loss: 1.09434111, Time: 0.0208 Steps: 48960, Updated lr: 0.000055 ***** Running evaluation checkpoint-48967 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-48967 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.579898, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.293404, "eval_total_loss": 206.262847, "eval_mae": 0.385973, "eval_mse": 0.293505, "eval_r2": 0.813429, "eval_sp_statistic": 0.862943, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.903717, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.744876, "test_total_loss": 373.927509, "test_mae": 0.63474, "test_mse": 0.744984, "test_r2": 0.519181, "test_sp_statistic": 0.799467, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.84307, "test_ps_pvalue": 0.0, "lr": 5.451209103840683e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0942096573935698, "train_cur_epoch_loss": 528.9734592046589, "train_cur_epoch_avg_loss": 0.24846099539908825, "train_cur_epoch_time": 44.57989764213562, "train_cur_epoch_avg_time": 0.02093936009494393, "epoch": 23, "step": 48967} ################################################## Training, Epoch: 0024, Batch: 000003, Sample Num: 48, Cur Loss: 0.17598143, Cur Avg Loss: 0.16795860, Log Avg loss: 0.17275823, Global Avg Loss: 1.09415291, Time: 0.0247 Steps: 48970, Updated lr: 0.000055 Training, Epoch: 0024, Batch: 000013, Sample Num: 208, Cur Loss: 0.11204708, Cur Avg Loss: 0.28809049, Log Avg loss: 0.32413006, Global Avg Loss: 1.09399570, Time: 0.0210 Steps: 48980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000023, Sample Num: 368, Cur Loss: 0.23834428, Cur Avg Loss: 0.24099673, Log Avg loss: 0.17977483, Global Avg Loss: 1.09380909, Time: 0.0209 Steps: 48990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000033, Sample Num: 528, Cur Loss: 0.07189242, Cur Avg Loss: 0.23657325, Log Avg loss: 0.22639925, Global Avg Loss: 1.09363207, Time: 0.0209 Steps: 49000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000043, Sample Num: 688, Cur Loss: 0.25044608, Cur Avg Loss: 0.24039666, Log Avg loss: 0.25301391, Global Avg Loss: 1.09346055, Time: 0.0210 Steps: 49010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000053, Sample Num: 848, Cur Loss: 0.13035296, Cur Avg Loss: 0.22969484, Log Avg loss: 0.18367704, Global Avg Loss: 1.09327495, Time: 0.0209 Steps: 49020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000063, Sample Num: 1008, Cur Loss: 0.20614430, Cur Avg Loss: 0.22469736, Log Avg loss: 0.19821068, Global Avg Loss: 1.09309240, Time: 0.0209 Steps: 49030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000073, Sample Num: 1168, Cur Loss: 0.10228000, Cur Avg Loss: 0.22757163, Log Avg loss: 0.24567951, Global Avg Loss: 1.09291960, Time: 0.0210 Steps: 49040, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000083, Sample Num: 1328, Cur Loss: 0.20292772, Cur Avg Loss: 0.23074058, Log Avg loss: 0.25387392, Global Avg Loss: 1.09274854, Time: 0.0209 Steps: 49050, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000093, Sample Num: 1488, Cur Loss: 0.31016517, Cur Avg Loss: 0.23564725, Log Avg loss: 0.27637263, Global Avg Loss: 1.09258213, Time: 0.0210 Steps: 49060, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000103, Sample Num: 1648, Cur Loss: 0.51370996, Cur Avg Loss: 0.23622946, Log Avg loss: 0.24164405, Global Avg Loss: 1.09240872, Time: 0.0209 Steps: 49070, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000113, Sample Num: 1808, Cur Loss: 0.19367027, Cur Avg Loss: 0.23030971, Log Avg loss: 0.16933631, Global Avg Loss: 1.09222065, Time: 0.0209 Steps: 49080, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000123, Sample Num: 1968, Cur Loss: 0.05356385, Cur Avg Loss: 0.22737317, Log Avg loss: 0.19419025, Global Avg Loss: 1.09203771, Time: 0.0209 Steps: 49090, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000133, Sample Num: 2128, Cur Loss: 0.16759044, Cur Avg Loss: 0.22525951, Log Avg loss: 0.19926151, Global Avg Loss: 1.09185588, Time: 0.0210 Steps: 49100, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000143, Sample Num: 2288, Cur Loss: 0.13169910, Cur Avg Loss: 0.22653474, Log Avg loss: 0.24349530, Global Avg Loss: 1.09168314, Time: 0.0210 Steps: 49110, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000153, Sample Num: 2448, Cur Loss: 0.12186238, Cur Avg Loss: 0.22877135, Log Avg loss: 0.26075478, Global Avg Loss: 1.09151397, Time: 0.0209 Steps: 49120, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000163, Sample Num: 2608, Cur Loss: 0.09825069, Cur Avg Loss: 0.22923263, Log Avg loss: 0.23629028, Global Avg Loss: 1.09133990, Time: 0.0209 Steps: 49130, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000173, Sample Num: 2768, Cur Loss: 0.20243870, Cur Avg Loss: 0.23063138, Log Avg loss: 0.25343094, Global Avg Loss: 1.09116938, Time: 0.0209 Steps: 49140, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000183, Sample Num: 2928, Cur Loss: 1.21793985, Cur Avg Loss: 0.23982561, Log Avg loss: 0.39888585, Global Avg Loss: 1.09102853, Time: 0.0210 Steps: 49150, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000193, Sample Num: 3088, Cur Loss: 0.18272039, Cur Avg Loss: 0.24283967, Log Avg loss: 0.29799697, Global Avg Loss: 1.09086722, Time: 0.0208 Steps: 49160, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000203, Sample Num: 3248, Cur Loss: 0.19810785, Cur Avg Loss: 0.23996996, Log Avg loss: 0.18458448, Global Avg Loss: 1.09068290, Time: 0.0208 Steps: 49170, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000213, Sample Num: 3408, Cur Loss: 0.55012250, Cur Avg Loss: 0.24102709, Log Avg loss: 0.26248686, Global Avg Loss: 1.09051450, Time: 0.0209 Steps: 49180, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000223, Sample Num: 3568, Cur Loss: 0.11961462, Cur Avg Loss: 0.24083339, Log Avg loss: 0.23670750, Global Avg Loss: 1.09034093, Time: 0.0208 Steps: 49190, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000233, Sample Num: 3728, Cur Loss: 0.22234300, Cur Avg Loss: 0.24106039, Log Avg loss: 0.24612252, Global Avg Loss: 1.09016934, Time: 0.0209 Steps: 49200, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000243, Sample Num: 3888, Cur Loss: 0.12790897, Cur Avg Loss: 0.24337983, Log Avg loss: 0.29742294, Global Avg Loss: 1.09000824, Time: 0.0208 Steps: 49210, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000253, Sample Num: 4048, Cur Loss: 0.67643094, Cur Avg Loss: 0.24953651, Log Avg loss: 0.39914360, Global Avg Loss: 1.08986788, Time: 0.0209 Steps: 49220, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000263, Sample Num: 4208, Cur Loss: 0.34309751, Cur Avg Loss: 0.25085890, Log Avg loss: 0.28431540, Global Avg Loss: 1.08970425, Time: 0.0209 Steps: 49230, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000273, Sample Num: 4368, Cur Loss: 0.20012707, Cur Avg Loss: 0.24989545, Log Avg loss: 0.22455672, Global Avg Loss: 1.08952855, Time: 0.0208 Steps: 49240, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000283, Sample Num: 4528, Cur Loss: 0.33158839, Cur Avg Loss: 0.24999586, Log Avg loss: 0.25273719, Global Avg Loss: 1.08935864, Time: 0.0208 Steps: 49250, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000293, Sample Num: 4688, Cur Loss: 0.14586060, Cur Avg Loss: 0.24811836, Log Avg loss: 0.19498510, Global Avg Loss: 1.08917708, Time: 0.0208 Steps: 49260, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000303, Sample Num: 4848, Cur Loss: 0.15003940, Cur Avg Loss: 0.24636181, Log Avg loss: 0.19489496, Global Avg Loss: 1.08899557, Time: 0.0210 Steps: 49270, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000313, Sample Num: 5008, Cur Loss: 0.48182714, Cur Avg Loss: 0.24718737, Log Avg loss: 0.27220169, Global Avg Loss: 1.08882983, Time: 0.0209 Steps: 49280, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000323, Sample Num: 5168, Cur Loss: 0.23205785, Cur Avg Loss: 0.24688222, Log Avg loss: 0.23733101, Global Avg Loss: 1.08865708, Time: 0.0209 Steps: 49290, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000333, Sample Num: 5328, Cur Loss: 0.19461922, Cur Avg Loss: 0.24606998, Log Avg loss: 0.21983460, Global Avg Loss: 1.08848084, Time: 0.0209 Steps: 49300, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000343, Sample Num: 5488, Cur Loss: 0.34314030, Cur Avg Loss: 0.24661118, Log Avg loss: 0.26463305, Global Avg Loss: 1.08831377, Time: 0.0209 Steps: 49310, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000353, Sample Num: 5648, Cur Loss: 0.10733686, Cur Avg Loss: 0.24815109, Log Avg loss: 0.30097026, Global Avg Loss: 1.08815413, Time: 0.0209 Steps: 49320, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000363, Sample Num: 5808, Cur Loss: 0.11512588, Cur Avg Loss: 0.24724822, Log Avg loss: 0.21537675, Global Avg Loss: 1.08797720, Time: 0.0208 Steps: 49330, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000373, Sample Num: 5968, Cur Loss: 0.27484596, Cur Avg Loss: 0.24575145, Log Avg loss: 0.19141875, Global Avg Loss: 1.08779549, Time: 0.0209 Steps: 49340, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000383, Sample Num: 6128, Cur Loss: 0.22840515, Cur Avg Loss: 0.24692855, Log Avg loss: 0.29083446, Global Avg Loss: 1.08763400, Time: 0.0209 Steps: 49350, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000393, Sample Num: 6288, Cur Loss: 0.09628381, Cur Avg Loss: 0.24664829, Log Avg loss: 0.23591442, Global Avg Loss: 1.08746145, Time: 0.0209 Steps: 49360, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000403, Sample Num: 6448, Cur Loss: 0.27125084, Cur Avg Loss: 0.24901942, Log Avg loss: 0.34220478, Global Avg Loss: 1.08731049, Time: 0.0209 Steps: 49370, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000413, Sample Num: 6608, Cur Loss: 0.14059861, Cur Avg Loss: 0.24741146, Log Avg loss: 0.18261057, Global Avg Loss: 1.08712728, Time: 0.0209 Steps: 49380, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000423, Sample Num: 6768, Cur Loss: 0.28980267, Cur Avg Loss: 0.24826113, Log Avg loss: 0.28335263, Global Avg Loss: 1.08696454, Time: 0.0209 Steps: 49390, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000433, Sample Num: 6928, Cur Loss: 0.14615092, Cur Avg Loss: 0.24782759, Log Avg loss: 0.22948871, Global Avg Loss: 1.08679096, Time: 0.0209 Steps: 49400, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000443, Sample Num: 7088, Cur Loss: 0.23145978, Cur Avg Loss: 0.24629991, Log Avg loss: 0.18015159, Global Avg Loss: 1.08660747, Time: 0.0209 Steps: 49410, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000453, Sample Num: 7248, Cur Loss: 0.09487206, Cur Avg Loss: 0.24619265, Log Avg loss: 0.24144080, Global Avg Loss: 1.08643645, Time: 0.0209 Steps: 49420, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000463, Sample Num: 7408, Cur Loss: 0.23807305, Cur Avg Loss: 0.24608659, Log Avg loss: 0.24128221, Global Avg Loss: 1.08626547, Time: 0.0209 Steps: 49430, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000473, Sample Num: 7568, Cur Loss: 0.09525770, Cur Avg Loss: 0.24583081, Log Avg loss: 0.23398829, Global Avg Loss: 1.08609309, Time: 0.0209 Steps: 49440, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000483, Sample Num: 7728, Cur Loss: 0.18993092, Cur Avg Loss: 0.24533524, Log Avg loss: 0.22189454, Global Avg Loss: 1.08591833, Time: 0.0209 Steps: 49450, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000493, Sample Num: 7888, Cur Loss: 0.16759099, Cur Avg Loss: 0.24551796, Log Avg loss: 0.25434326, Global Avg Loss: 1.08575020, Time: 0.0209 Steps: 49460, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000503, Sample Num: 8048, Cur Loss: 0.11081609, Cur Avg Loss: 0.24573060, Log Avg loss: 0.25621366, Global Avg Loss: 1.08558251, Time: 0.0209 Steps: 49470, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000513, Sample Num: 8208, Cur Loss: 0.12365028, Cur Avg Loss: 0.24540919, Log Avg loss: 0.22924225, Global Avg Loss: 1.08540944, Time: 0.0245 Steps: 49480, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000523, Sample Num: 8368, Cur Loss: 0.46806729, Cur Avg Loss: 0.24402629, Log Avg loss: 0.17308360, Global Avg Loss: 1.08522510, Time: 0.0208 Steps: 49490, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000533, Sample Num: 8528, Cur Loss: 0.33328873, Cur Avg Loss: 0.24479159, Log Avg loss: 0.28481685, Global Avg Loss: 1.08506340, Time: 0.0208 Steps: 49500, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000543, Sample Num: 8688, Cur Loss: 0.20234653, Cur Avg Loss: 0.24377606, Log Avg loss: 0.18964812, Global Avg Loss: 1.08488254, Time: 0.0208 Steps: 49510, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000553, Sample Num: 8848, Cur Loss: 0.16260415, Cur Avg Loss: 0.24424496, Log Avg loss: 0.26970650, Global Avg Loss: 1.08471793, Time: 0.0211 Steps: 49520, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000563, Sample Num: 9008, Cur Loss: 0.10676640, Cur Avg Loss: 0.24374616, Log Avg loss: 0.21616227, Global Avg Loss: 1.08454257, Time: 0.0208 Steps: 49530, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000573, Sample Num: 9168, Cur Loss: 0.16702634, Cur Avg Loss: 0.24373352, Log Avg loss: 0.24302236, Global Avg Loss: 1.08437270, Time: 0.0208 Steps: 49540, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000583, Sample Num: 9328, Cur Loss: 0.57722479, Cur Avg Loss: 0.24391051, Log Avg loss: 0.25405202, Global Avg Loss: 1.08420513, Time: 0.0208 Steps: 49550, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000593, Sample Num: 9488, Cur Loss: 0.18256067, Cur Avg Loss: 0.24396140, Log Avg loss: 0.24692782, Global Avg Loss: 1.08403619, Time: 0.0208 Steps: 49560, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000603, Sample Num: 9648, Cur Loss: 0.04966599, Cur Avg Loss: 0.24308568, Log Avg loss: 0.19115561, Global Avg Loss: 1.08385606, Time: 0.0208 Steps: 49570, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000613, Sample Num: 9808, Cur Loss: 0.47411025, Cur Avg Loss: 0.24298624, Log Avg loss: 0.23699030, Global Avg Loss: 1.08368525, Time: 0.0208 Steps: 49580, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000623, Sample Num: 9968, Cur Loss: 0.18691368, Cur Avg Loss: 0.24301974, Log Avg loss: 0.24507334, Global Avg Loss: 1.08351614, Time: 0.0208 Steps: 49590, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000633, Sample Num: 10128, Cur Loss: 0.13476904, Cur Avg Loss: 0.24411483, Log Avg loss: 0.31233853, Global Avg Loss: 1.08336066, Time: 0.0208 Steps: 49600, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000643, Sample Num: 10288, Cur Loss: 0.25514925, Cur Avg Loss: 0.24338332, Log Avg loss: 0.19707906, Global Avg Loss: 1.08318202, Time: 0.0209 Steps: 49610, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000653, Sample Num: 10448, Cur Loss: 0.37592921, Cur Avg Loss: 0.24264368, Log Avg loss: 0.19508475, Global Avg Loss: 1.08300304, Time: 0.0208 Steps: 49620, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000663, Sample Num: 10608, Cur Loss: 0.24483156, Cur Avg Loss: 0.24218970, Log Avg loss: 0.21254490, Global Avg Loss: 1.08282765, Time: 0.0208 Steps: 49630, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000673, Sample Num: 10768, Cur Loss: 0.18206534, Cur Avg Loss: 0.24303001, Log Avg loss: 0.29874252, Global Avg Loss: 1.08266969, Time: 0.0208 Steps: 49640, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000683, Sample Num: 10928, Cur Loss: 0.11984821, Cur Avg Loss: 0.24314079, Log Avg loss: 0.25059638, Global Avg Loss: 1.08250210, Time: 0.0208 Steps: 49650, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000693, Sample Num: 11088, Cur Loss: 0.30637902, Cur Avg Loss: 0.24335777, Log Avg loss: 0.25817762, Global Avg Loss: 1.08233611, Time: 0.0209 Steps: 49660, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000703, Sample Num: 11248, Cur Loss: 0.16047728, Cur Avg Loss: 0.24268569, Log Avg loss: 0.19611050, Global Avg Loss: 1.08215769, Time: 0.0208 Steps: 49670, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000713, Sample Num: 11408, Cur Loss: 0.09600638, Cur Avg Loss: 0.24214266, Log Avg loss: 0.20396710, Global Avg Loss: 1.08198092, Time: 0.0210 Steps: 49680, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000723, Sample Num: 11568, Cur Loss: 0.56096035, Cur Avg Loss: 0.24298694, Log Avg loss: 0.30318413, Global Avg Loss: 1.08182419, Time: 0.0208 Steps: 49690, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000733, Sample Num: 11728, Cur Loss: 0.09810974, Cur Avg Loss: 0.24190094, Log Avg loss: 0.16338306, Global Avg Loss: 1.08163939, Time: 0.0208 Steps: 49700, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000743, Sample Num: 11888, Cur Loss: 0.03863959, Cur Avg Loss: 0.24218932, Log Avg loss: 0.26332775, Global Avg Loss: 1.08147477, Time: 0.0208 Steps: 49710, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000753, Sample Num: 12048, Cur Loss: 0.72234255, Cur Avg Loss: 0.24329446, Log Avg loss: 0.32540690, Global Avg Loss: 1.08132271, Time: 0.0209 Steps: 49720, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000763, Sample Num: 12208, Cur Loss: 0.24898337, Cur Avg Loss: 0.24278935, Log Avg loss: 0.20475403, Global Avg Loss: 1.08114644, Time: 0.0208 Steps: 49730, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000773, Sample Num: 12368, Cur Loss: 0.30719447, Cur Avg Loss: 0.24251147, Log Avg loss: 0.22130937, Global Avg Loss: 1.08097358, Time: 0.0226 Steps: 49740, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000783, Sample Num: 12528, Cur Loss: 0.21784633, Cur Avg Loss: 0.24191234, Log Avg loss: 0.19559999, Global Avg Loss: 1.08079561, Time: 0.0208 Steps: 49750, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000793, Sample Num: 12688, Cur Loss: 0.22642967, Cur Avg Loss: 0.24166670, Log Avg loss: 0.22243252, Global Avg Loss: 1.08062311, Time: 0.0208 Steps: 49760, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000803, Sample Num: 12848, Cur Loss: 0.21640661, Cur Avg Loss: 0.24281720, Log Avg loss: 0.33405194, Global Avg Loss: 1.08047311, Time: 0.0208 Steps: 49770, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000813, Sample Num: 13008, Cur Loss: 0.24998009, Cur Avg Loss: 0.24263751, Log Avg loss: 0.22820820, Global Avg Loss: 1.08030190, Time: 0.0208 Steps: 49780, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000823, Sample Num: 13168, Cur Loss: 0.34825423, Cur Avg Loss: 0.24313361, Log Avg loss: 0.28346697, Global Avg Loss: 1.08014186, Time: 0.0208 Steps: 49790, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000833, Sample Num: 13328, Cur Loss: 0.35472661, Cur Avg Loss: 0.24321753, Log Avg loss: 0.25012419, Global Avg Loss: 1.07997519, Time: 0.0208 Steps: 49800, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000843, Sample Num: 13488, Cur Loss: 0.16762005, Cur Avg Loss: 0.24284757, Log Avg loss: 0.21202968, Global Avg Loss: 1.07980094, Time: 0.0208 Steps: 49810, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000853, Sample Num: 13648, Cur Loss: 0.22350419, Cur Avg Loss: 0.24338861, Log Avg loss: 0.28899849, Global Avg Loss: 1.07964221, Time: 0.0208 Steps: 49820, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000863, Sample Num: 13808, Cur Loss: 0.05375788, Cur Avg Loss: 0.24396536, Log Avg loss: 0.29316218, Global Avg Loss: 1.07948437, Time: 0.0208 Steps: 49830, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000873, Sample Num: 13968, Cur Loss: 0.11728309, Cur Avg Loss: 0.24357532, Log Avg loss: 0.20991468, Global Avg Loss: 1.07930990, Time: 0.0208 Steps: 49840, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000883, Sample Num: 14128, Cur Loss: 0.28769988, Cur Avg Loss: 0.24412648, Log Avg loss: 0.29224254, Global Avg Loss: 1.07915202, Time: 0.0208 Steps: 49850, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000893, Sample Num: 14288, Cur Loss: 0.21159664, Cur Avg Loss: 0.24587097, Log Avg loss: 0.39990934, Global Avg Loss: 1.07901579, Time: 0.0208 Steps: 49860, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000903, Sample Num: 14448, Cur Loss: 0.21599348, Cur Avg Loss: 0.24654344, Log Avg loss: 0.30659541, Global Avg Loss: 1.07886090, Time: 0.0208 Steps: 49870, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000913, Sample Num: 14608, Cur Loss: 0.28787452, Cur Avg Loss: 0.24675483, Log Avg loss: 0.26584346, Global Avg Loss: 1.07869790, Time: 0.0208 Steps: 49880, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000923, Sample Num: 14768, Cur Loss: 0.11516165, Cur Avg Loss: 0.24737785, Log Avg loss: 0.30425975, Global Avg Loss: 1.07854267, Time: 0.0208 Steps: 49890, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000933, Sample Num: 14928, Cur Loss: 0.12606831, Cur Avg Loss: 0.24646832, Log Avg loss: 0.16251868, Global Avg Loss: 1.07835910, Time: 0.0208 Steps: 49900, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000943, Sample Num: 15088, Cur Loss: 0.09462891, Cur Avg Loss: 0.24604125, Log Avg loss: 0.20619553, Global Avg Loss: 1.07818436, Time: 0.0209 Steps: 49910, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000953, Sample Num: 15248, Cur Loss: 0.29389238, Cur Avg Loss: 0.24577082, Log Avg loss: 0.22026916, Global Avg Loss: 1.07801250, Time: 0.0208 Steps: 49920, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000963, Sample Num: 15408, Cur Loss: 0.21956919, Cur Avg Loss: 0.24600261, Log Avg loss: 0.26809220, Global Avg Loss: 1.07785029, Time: 0.0208 Steps: 49930, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000973, Sample Num: 15568, Cur Loss: 0.29033631, Cur Avg Loss: 0.24701959, Log Avg loss: 0.34495496, Global Avg Loss: 1.07770353, Time: 0.0208 Steps: 49940, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000983, Sample Num: 15728, Cur Loss: 0.13987376, Cur Avg Loss: 0.24710188, Log Avg loss: 0.25510810, Global Avg Loss: 1.07753885, Time: 0.0208 Steps: 49950, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000993, Sample Num: 15888, Cur Loss: 0.16404907, Cur Avg Loss: 0.24832862, Log Avg loss: 0.36891737, Global Avg Loss: 1.07739701, Time: 0.0208 Steps: 49960, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001003, Sample Num: 16048, Cur Loss: 0.07219319, Cur Avg Loss: 0.24726980, Log Avg loss: 0.14212948, Global Avg Loss: 1.07720984, Time: 0.0208 Steps: 49970, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001013, Sample Num: 16208, Cur Loss: 0.11878780, Cur Avg Loss: 0.24662163, Log Avg loss: 0.18161018, Global Avg Loss: 1.07703065, Time: 0.0208 Steps: 49980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001023, Sample Num: 16368, Cur Loss: 0.06358685, Cur Avg Loss: 0.24577237, Log Avg loss: 0.15974156, Global Avg Loss: 1.07684716, Time: 0.0208 Steps: 49990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001033, Sample Num: 16528, Cur Loss: 0.39368662, Cur Avg Loss: 0.24607814, Log Avg loss: 0.27735862, Global Avg Loss: 1.07668726, Time: 0.0208 Steps: 50000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001043, Sample Num: 16688, Cur Loss: 0.45226669, Cur Avg Loss: 0.24633891, Log Avg loss: 0.27327691, Global Avg Loss: 1.07652661, Time: 0.0208 Steps: 50010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001053, Sample Num: 16848, Cur Loss: 0.46710852, Cur Avg Loss: 0.24688025, Log Avg loss: 0.30334129, Global Avg Loss: 1.07637204, Time: 0.0208 Steps: 50020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001063, Sample Num: 17008, Cur Loss: 0.19656783, Cur Avg Loss: 0.24725123, Log Avg loss: 0.28631605, Global Avg Loss: 1.07621412, Time: 0.0208 Steps: 50030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001073, Sample Num: 17168, Cur Loss: 0.52469587, Cur Avg Loss: 0.24758088, Log Avg loss: 0.28262220, Global Avg Loss: 1.07605553, Time: 0.0208 Steps: 50040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001083, Sample Num: 17328, Cur Loss: 0.35148820, Cur Avg Loss: 0.24735840, Log Avg loss: 0.22348685, Global Avg Loss: 1.07588518, Time: 0.0208 Steps: 50050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001093, Sample Num: 17488, Cur Loss: 0.32754207, Cur Avg Loss: 0.24711561, Log Avg loss: 0.22082089, Global Avg Loss: 1.07571438, Time: 0.0208 Steps: 50060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001103, Sample Num: 17648, Cur Loss: 0.25481310, Cur Avg Loss: 0.24657655, Log Avg loss: 0.18765765, Global Avg Loss: 1.07553701, Time: 0.0207 Steps: 50070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001113, Sample Num: 17808, Cur Loss: 0.74950373, Cur Avg Loss: 0.24662762, Log Avg loss: 0.25226010, Global Avg Loss: 1.07537262, Time: 0.0208 Steps: 50080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001123, Sample Num: 17968, Cur Loss: 0.21490890, Cur Avg Loss: 0.24707877, Log Avg loss: 0.29729226, Global Avg Loss: 1.07521728, Time: 0.0208 Steps: 50090, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001133, Sample Num: 18128, Cur Loss: 0.26926056, Cur Avg Loss: 0.24762219, Log Avg loss: 0.30864791, Global Avg Loss: 1.07506428, Time: 0.0209 Steps: 50100, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001143, Sample Num: 18288, Cur Loss: 0.34654638, Cur Avg Loss: 0.24847744, Log Avg loss: 0.34537799, Global Avg Loss: 1.07491866, Time: 0.0208 Steps: 50110, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001153, Sample Num: 18448, Cur Loss: 0.39155680, Cur Avg Loss: 0.24936859, Log Avg loss: 0.35122690, Global Avg Loss: 1.07477427, Time: 0.0208 Steps: 50120, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001163, Sample Num: 18608, Cur Loss: 0.41951114, Cur Avg Loss: 0.24955703, Log Avg loss: 0.27128334, Global Avg Loss: 1.07461399, Time: 0.0209 Steps: 50130, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001173, Sample Num: 18768, Cur Loss: 0.52013719, Cur Avg Loss: 0.24992296, Log Avg loss: 0.29248059, Global Avg Loss: 1.07445800, Time: 0.0208 Steps: 50140, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001183, Sample Num: 18928, Cur Loss: 0.07512675, Cur Avg Loss: 0.25028148, Log Avg loss: 0.29233652, Global Avg Loss: 1.07430204, Time: 0.0208 Steps: 50150, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001193, Sample Num: 19088, Cur Loss: 0.25081086, Cur Avg Loss: 0.25023220, Log Avg loss: 0.24440217, Global Avg Loss: 1.07413659, Time: 0.0208 Steps: 50160, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001203, Sample Num: 19248, Cur Loss: 0.22970220, Cur Avg Loss: 0.25035964, Log Avg loss: 0.26556348, Global Avg Loss: 1.07397542, Time: 0.0208 Steps: 50170, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001213, Sample Num: 19408, Cur Loss: 0.32759631, Cur Avg Loss: 0.25021743, Log Avg loss: 0.23310949, Global Avg Loss: 1.07380785, Time: 0.0208 Steps: 50180, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001223, Sample Num: 19568, Cur Loss: 0.26160204, Cur Avg Loss: 0.25039900, Log Avg loss: 0.27242384, Global Avg Loss: 1.07364818, Time: 0.0208 Steps: 50190, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001233, Sample Num: 19728, Cur Loss: 0.37023979, Cur Avg Loss: 0.25036894, Log Avg loss: 0.24669175, Global Avg Loss: 1.07348345, Time: 0.0208 Steps: 50200, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001243, Sample Num: 19888, Cur Loss: 0.27620652, Cur Avg Loss: 0.25004051, Log Avg loss: 0.20954553, Global Avg Loss: 1.07331139, Time: 0.0208 Steps: 50210, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001253, Sample Num: 20048, Cur Loss: 0.04342626, Cur Avg Loss: 0.24953188, Log Avg loss: 0.18630842, Global Avg Loss: 1.07313476, Time: 0.0208 Steps: 50220, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001263, Sample Num: 20208, Cur Loss: 0.18790658, Cur Avg Loss: 0.24915356, Log Avg loss: 0.20175085, Global Avg Loss: 1.07296128, Time: 0.0208 Steps: 50230, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001273, Sample Num: 20368, Cur Loss: 0.40170431, Cur Avg Loss: 0.24959635, Log Avg loss: 0.30552004, Global Avg Loss: 1.07280853, Time: 0.0208 Steps: 50240, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001283, Sample Num: 20528, Cur Loss: 0.07227149, Cur Avg Loss: 0.24911170, Log Avg loss: 0.18741600, Global Avg Loss: 1.07263233, Time: 0.0244 Steps: 50250, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001293, Sample Num: 20688, Cur Loss: 0.18333693, Cur Avg Loss: 0.24883653, Log Avg loss: 0.21353272, Global Avg Loss: 1.07246140, Time: 0.0208 Steps: 50260, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001303, Sample Num: 20848, Cur Loss: 0.10810843, Cur Avg Loss: 0.24918209, Log Avg loss: 0.29386314, Global Avg Loss: 1.07230652, Time: 0.0207 Steps: 50270, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001313, Sample Num: 21008, Cur Loss: 0.62022316, Cur Avg Loss: 0.24953743, Log Avg loss: 0.29583778, Global Avg Loss: 1.07215209, Time: 0.0208 Steps: 50280, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001323, Sample Num: 21168, Cur Loss: 0.36732408, Cur Avg Loss: 0.24961565, Log Avg loss: 0.25988595, Global Avg Loss: 1.07199057, Time: 0.0208 Steps: 50290, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001333, Sample Num: 21328, Cur Loss: 0.13492244, Cur Avg Loss: 0.24965190, Log Avg loss: 0.25444828, Global Avg Loss: 1.07182804, Time: 0.0207 Steps: 50300, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001343, Sample Num: 21488, Cur Loss: 0.58270091, Cur Avg Loss: 0.24983332, Log Avg loss: 0.27401579, Global Avg Loss: 1.07166946, Time: 0.0208 Steps: 50310, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001353, Sample Num: 21648, Cur Loss: 0.07191495, Cur Avg Loss: 0.24964585, Log Avg loss: 0.22446885, Global Avg Loss: 1.07150110, Time: 0.0208 Steps: 50320, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001363, Sample Num: 21808, Cur Loss: 0.17549454, Cur Avg Loss: 0.24916431, Log Avg loss: 0.18401144, Global Avg Loss: 1.07132476, Time: 0.0208 Steps: 50330, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001373, Sample Num: 21968, Cur Loss: 0.05310366, Cur Avg Loss: 0.24895621, Log Avg loss: 0.22059298, Global Avg Loss: 1.07115576, Time: 0.0208 Steps: 50340, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001383, Sample Num: 22128, Cur Loss: 0.17633036, Cur Avg Loss: 0.24910673, Log Avg loss: 0.26977296, Global Avg Loss: 1.07099660, Time: 0.0209 Steps: 50350, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001393, Sample Num: 22288, Cur Loss: 0.24352911, Cur Avg Loss: 0.24916357, Log Avg loss: 0.25702508, Global Avg Loss: 1.07083497, Time: 0.0208 Steps: 50360, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001403, Sample Num: 22448, Cur Loss: 0.47986978, Cur Avg Loss: 0.24940333, Log Avg loss: 0.28280193, Global Avg Loss: 1.07067852, Time: 0.0208 Steps: 50370, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001413, Sample Num: 22608, Cur Loss: 0.51229489, Cur Avg Loss: 0.24993117, Log Avg loss: 0.32398666, Global Avg Loss: 1.07053031, Time: 0.0208 Steps: 50380, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001423, Sample Num: 22768, Cur Loss: 0.13696948, Cur Avg Loss: 0.25067198, Log Avg loss: 0.35534856, Global Avg Loss: 1.07038838, Time: 0.0208 Steps: 50390, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001433, Sample Num: 22928, Cur Loss: 0.21414885, Cur Avg Loss: 0.25072672, Log Avg loss: 0.25851644, Global Avg Loss: 1.07022730, Time: 0.0209 Steps: 50400, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001443, Sample Num: 23088, Cur Loss: 0.31249493, Cur Avg Loss: 0.25043757, Log Avg loss: 0.20900132, Global Avg Loss: 1.07005645, Time: 0.0208 Steps: 50410, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001453, Sample Num: 23248, Cur Loss: 0.14720541, Cur Avg Loss: 0.25037817, Log Avg loss: 0.24180686, Global Avg Loss: 1.06989218, Time: 0.0208 Steps: 50420, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001463, Sample Num: 23408, Cur Loss: 0.18097448, Cur Avg Loss: 0.25052777, Log Avg loss: 0.27226492, Global Avg Loss: 1.06973402, Time: 0.0208 Steps: 50430, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001473, Sample Num: 23568, Cur Loss: 0.28161883, Cur Avg Loss: 0.25065106, Log Avg loss: 0.26868853, Global Avg Loss: 1.06957520, Time: 0.0208 Steps: 50440, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001483, Sample Num: 23728, Cur Loss: 0.36466575, Cur Avg Loss: 0.25040045, Log Avg loss: 0.21348518, Global Avg Loss: 1.06940551, Time: 0.0208 Steps: 50450, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001493, Sample Num: 23888, Cur Loss: 0.49301758, Cur Avg Loss: 0.24994523, Log Avg loss: 0.18243701, Global Avg Loss: 1.06922974, Time: 0.0208 Steps: 50460, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001503, Sample Num: 24048, Cur Loss: 0.32973117, Cur Avg Loss: 0.25042216, Log Avg loss: 0.32162699, Global Avg Loss: 1.06908161, Time: 0.0208 Steps: 50470, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001513, Sample Num: 24208, Cur Loss: 0.40560350, Cur Avg Loss: 0.25006531, Log Avg loss: 0.19643115, Global Avg Loss: 1.06890874, Time: 0.0208 Steps: 50480, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001523, Sample Num: 24368, Cur Loss: 0.27594560, Cur Avg Loss: 0.25021905, Log Avg loss: 0.27348052, Global Avg Loss: 1.06875120, Time: 0.0208 Steps: 50490, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001533, Sample Num: 24528, Cur Loss: 0.16408232, Cur Avg Loss: 0.25003038, Log Avg loss: 0.22129613, Global Avg Loss: 1.06858338, Time: 0.0208 Steps: 50500, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001543, Sample Num: 24688, Cur Loss: 0.17730048, Cur Avg Loss: 0.24947371, Log Avg loss: 0.16413508, Global Avg Loss: 1.06840432, Time: 0.0208 Steps: 50510, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001553, Sample Num: 24848, Cur Loss: 0.14829078, Cur Avg Loss: 0.24958424, Log Avg loss: 0.26664016, Global Avg Loss: 1.06824562, Time: 0.0208 Steps: 50520, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001563, Sample Num: 25008, Cur Loss: 0.24856395, Cur Avg Loss: 0.24921026, Log Avg loss: 0.19113109, Global Avg Loss: 1.06807204, Time: 0.0208 Steps: 50530, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001573, Sample Num: 25168, Cur Loss: 0.11676243, Cur Avg Loss: 0.24876321, Log Avg loss: 0.17888840, Global Avg Loss: 1.06789610, Time: 0.0208 Steps: 50540, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001583, Sample Num: 25328, Cur Loss: 0.14561290, Cur Avg Loss: 0.24888380, Log Avg loss: 0.26785208, Global Avg Loss: 1.06773783, Time: 0.0208 Steps: 50550, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001593, Sample Num: 25488, Cur Loss: 0.45165509, Cur Avg Loss: 0.24877544, Log Avg loss: 0.23162222, Global Avg Loss: 1.06757246, Time: 0.0208 Steps: 50560, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001603, Sample Num: 25648, Cur Loss: 0.28381342, Cur Avg Loss: 0.24899897, Log Avg loss: 0.28460826, Global Avg Loss: 1.06741763, Time: 0.0208 Steps: 50570, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001613, Sample Num: 25808, Cur Loss: 0.27817225, Cur Avg Loss: 0.24896642, Log Avg loss: 0.24374795, Global Avg Loss: 1.06725479, Time: 0.0208 Steps: 50580, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001623, Sample Num: 25968, Cur Loss: 0.11761382, Cur Avg Loss: 0.24857222, Log Avg loss: 0.18498883, Global Avg Loss: 1.06708039, Time: 0.0208 Steps: 50590, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001633, Sample Num: 26128, Cur Loss: 0.09670109, Cur Avg Loss: 0.24864027, Log Avg loss: 0.25968384, Global Avg Loss: 1.06692083, Time: 0.0208 Steps: 50600, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001643, Sample Num: 26288, Cur Loss: 0.21421546, Cur Avg Loss: 0.24860444, Log Avg loss: 0.24275376, Global Avg Loss: 1.06675798, Time: 0.0208 Steps: 50610, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001653, Sample Num: 26448, Cur Loss: 0.27075019, Cur Avg Loss: 0.24821173, Log Avg loss: 0.18369005, Global Avg Loss: 1.06658353, Time: 0.0208 Steps: 50620, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001663, Sample Num: 26608, Cur Loss: 0.12881504, Cur Avg Loss: 0.24790082, Log Avg loss: 0.19650739, Global Avg Loss: 1.06641168, Time: 0.0208 Steps: 50630, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001673, Sample Num: 26768, Cur Loss: 0.05340288, Cur Avg Loss: 0.24732134, Log Avg loss: 0.15095311, Global Avg Loss: 1.06623090, Time: 0.0208 Steps: 50640, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001683, Sample Num: 26928, Cur Loss: 0.17988539, Cur Avg Loss: 0.24728399, Log Avg loss: 0.24103452, Global Avg Loss: 1.06606798, Time: 0.0208 Steps: 50650, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001693, Sample Num: 27088, Cur Loss: 0.32397869, Cur Avg Loss: 0.24710275, Log Avg loss: 0.21660151, Global Avg Loss: 1.06590030, Time: 0.0208 Steps: 50660, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001703, Sample Num: 27248, Cur Loss: 0.21768025, Cur Avg Loss: 0.24693337, Log Avg loss: 0.21825622, Global Avg Loss: 1.06573301, Time: 0.0209 Steps: 50670, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001713, Sample Num: 27408, Cur Loss: 0.14798194, Cur Avg Loss: 0.24668112, Log Avg loss: 0.20372330, Global Avg Loss: 1.06556293, Time: 0.0209 Steps: 50680, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001723, Sample Num: 27568, Cur Loss: 0.19573574, Cur Avg Loss: 0.24708970, Log Avg loss: 0.31707884, Global Avg Loss: 1.06541527, Time: 0.0208 Steps: 50690, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001733, Sample Num: 27728, Cur Loss: 0.16547646, Cur Avg Loss: 0.24695804, Log Avg loss: 0.22427387, Global Avg Loss: 1.06524936, Time: 0.0208 Steps: 50700, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001743, Sample Num: 27888, Cur Loss: 0.21453084, Cur Avg Loss: 0.24674505, Log Avg loss: 0.20983360, Global Avg Loss: 1.06508067, Time: 0.0208 Steps: 50710, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001753, Sample Num: 28048, Cur Loss: 0.69489324, Cur Avg Loss: 0.24672821, Log Avg loss: 0.24379279, Global Avg Loss: 1.06491875, Time: 0.0208 Steps: 50720, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001763, Sample Num: 28208, Cur Loss: 0.28917271, Cur Avg Loss: 0.24648647, Log Avg loss: 0.20411013, Global Avg Loss: 1.06474906, Time: 0.0208 Steps: 50730, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001773, Sample Num: 28368, Cur Loss: 0.19640958, Cur Avg Loss: 0.24667740, Log Avg loss: 0.28033736, Global Avg Loss: 1.06459447, Time: 0.0209 Steps: 50740, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001783, Sample Num: 28528, Cur Loss: 0.34823591, Cur Avg Loss: 0.24655582, Log Avg loss: 0.22500028, Global Avg Loss: 1.06442903, Time: 0.0208 Steps: 50750, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001793, Sample Num: 28688, Cur Loss: 0.26483077, Cur Avg Loss: 0.24635125, Log Avg loss: 0.20987664, Global Avg Loss: 1.06426068, Time: 0.0245 Steps: 50760, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001803, Sample Num: 28848, Cur Loss: 0.31135753, Cur Avg Loss: 0.24693906, Log Avg loss: 0.35233342, Global Avg Loss: 1.06412045, Time: 0.0208 Steps: 50770, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001813, Sample Num: 29008, Cur Loss: 0.23035209, Cur Avg Loss: 0.24727990, Log Avg loss: 0.30873363, Global Avg Loss: 1.06397170, Time: 0.0208 Steps: 50780, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001823, Sample Num: 29168, Cur Loss: 0.25478092, Cur Avg Loss: 0.24751068, Log Avg loss: 0.28935082, Global Avg Loss: 1.06381918, Time: 0.0208 Steps: 50790, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001833, Sample Num: 29328, Cur Loss: 0.22953741, Cur Avg Loss: 0.24751091, Log Avg loss: 0.24755287, Global Avg Loss: 1.06365850, Time: 0.0208 Steps: 50800, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001843, Sample Num: 29488, Cur Loss: 0.53285873, Cur Avg Loss: 0.24796471, Log Avg loss: 0.33114518, Global Avg Loss: 1.06351433, Time: 0.0209 Steps: 50810, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001853, Sample Num: 29648, Cur Loss: 0.20053390, Cur Avg Loss: 0.24807620, Log Avg loss: 0.26862388, Global Avg Loss: 1.06335792, Time: 0.0208 Steps: 50820, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001863, Sample Num: 29808, Cur Loss: 0.12009275, Cur Avg Loss: 0.24794173, Log Avg loss: 0.22302600, Global Avg Loss: 1.06319260, Time: 0.0208 Steps: 50830, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001873, Sample Num: 29968, Cur Loss: 0.07198366, Cur Avg Loss: 0.24814244, Log Avg loss: 0.28553429, Global Avg Loss: 1.06303964, Time: 0.0208 Steps: 50840, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001883, Sample Num: 30128, Cur Loss: 0.19683373, Cur Avg Loss: 0.24789734, Log Avg loss: 0.20199026, Global Avg Loss: 1.06287030, Time: 0.0208 Steps: 50850, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001893, Sample Num: 30288, Cur Loss: 0.16598707, Cur Avg Loss: 0.24793323, Log Avg loss: 0.25469087, Global Avg Loss: 1.06271140, Time: 0.0208 Steps: 50860, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001903, Sample Num: 30448, Cur Loss: 0.21321762, Cur Avg Loss: 0.24798338, Log Avg loss: 0.25747702, Global Avg Loss: 1.06255311, Time: 0.0208 Steps: 50870, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001913, Sample Num: 30608, Cur Loss: 0.35911590, Cur Avg Loss: 0.24795022, Log Avg loss: 0.24164007, Global Avg Loss: 1.06239177, Time: 0.0208 Steps: 50880, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001923, Sample Num: 30768, Cur Loss: 0.29467297, Cur Avg Loss: 0.24809400, Log Avg loss: 0.27559815, Global Avg Loss: 1.06223716, Time: 0.0207 Steps: 50890, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001933, Sample Num: 30928, Cur Loss: 0.20824957, Cur Avg Loss: 0.24801112, Log Avg loss: 0.23207356, Global Avg Loss: 1.06207406, Time: 0.0208 Steps: 50900, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001943, Sample Num: 31088, Cur Loss: 0.08066665, Cur Avg Loss: 0.24756997, Log Avg loss: 0.16229579, Global Avg Loss: 1.06189732, Time: 0.0208 Steps: 50910, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001953, Sample Num: 31248, Cur Loss: 0.15833144, Cur Avg Loss: 0.24802317, Log Avg loss: 0.33607969, Global Avg Loss: 1.06175478, Time: 0.0208 Steps: 50920, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001963, Sample Num: 31408, Cur Loss: 0.62366372, Cur Avg Loss: 0.24865555, Log Avg loss: 0.37215961, Global Avg Loss: 1.06161938, Time: 0.0208 Steps: 50930, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001973, Sample Num: 31568, Cur Loss: 0.08548371, Cur Avg Loss: 0.24864395, Log Avg loss: 0.24636712, Global Avg Loss: 1.06145934, Time: 0.0208 Steps: 50940, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001983, Sample Num: 31728, Cur Loss: 0.55948114, Cur Avg Loss: 0.24871120, Log Avg loss: 0.26197879, Global Avg Loss: 1.06130243, Time: 0.0208 Steps: 50950, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001993, Sample Num: 31888, Cur Loss: 0.26690775, Cur Avg Loss: 0.24851356, Log Avg loss: 0.20932212, Global Avg Loss: 1.06113524, Time: 0.0208 Steps: 50960, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002003, Sample Num: 32048, Cur Loss: 0.23689380, Cur Avg Loss: 0.24843215, Log Avg loss: 0.23220710, Global Avg Loss: 1.06097261, Time: 0.0207 Steps: 50970, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002013, Sample Num: 32208, Cur Loss: 0.22576842, Cur Avg Loss: 0.24861767, Log Avg loss: 0.28577708, Global Avg Loss: 1.06082055, Time: 0.0208 Steps: 50980, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002023, Sample Num: 32368, Cur Loss: 0.07345943, Cur Avg Loss: 0.24812617, Log Avg loss: 0.14918745, Global Avg Loss: 1.06064176, Time: 0.0208 Steps: 50990, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002033, Sample Num: 32528, Cur Loss: 0.11628237, Cur Avg Loss: 0.24828192, Log Avg loss: 0.27979016, Global Avg Loss: 1.06048866, Time: 0.0208 Steps: 51000, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002043, Sample Num: 32688, Cur Loss: 0.13015127, Cur Avg Loss: 0.24818107, Log Avg loss: 0.22767820, Global Avg Loss: 1.06032539, Time: 0.0208 Steps: 51010, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002053, Sample Num: 32848, Cur Loss: 0.13218577, Cur Avg Loss: 0.24806666, Log Avg loss: 0.22469212, Global Avg Loss: 1.06016161, Time: 0.0244 Steps: 51020, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002063, Sample Num: 33008, Cur Loss: 0.14901924, Cur Avg Loss: 0.24785475, Log Avg loss: 0.20435059, Global Avg Loss: 1.05999390, Time: 0.0208 Steps: 51030, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002073, Sample Num: 33168, Cur Loss: 0.09650449, Cur Avg Loss: 0.24775948, Log Avg loss: 0.22810482, Global Avg Loss: 1.05983091, Time: 0.0207 Steps: 51040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002083, Sample Num: 33328, Cur Loss: 0.10454676, Cur Avg Loss: 0.24751111, Log Avg loss: 0.19602340, Global Avg Loss: 1.05966170, Time: 0.0207 Steps: 51050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002093, Sample Num: 33488, Cur Loss: 0.06986920, Cur Avg Loss: 0.24727180, Log Avg loss: 0.19742465, Global Avg Loss: 1.05949284, Time: 0.0208 Steps: 51060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002103, Sample Num: 33648, Cur Loss: 0.25692052, Cur Avg Loss: 0.24725408, Log Avg loss: 0.24354517, Global Avg Loss: 1.05933306, Time: 0.0208 Steps: 51070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002113, Sample Num: 33808, Cur Loss: 0.25070834, Cur Avg Loss: 0.24737691, Log Avg loss: 0.27320704, Global Avg Loss: 1.05917916, Time: 0.0207 Steps: 51080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002123, Sample Num: 33968, Cur Loss: 0.21984974, Cur Avg Loss: 0.24738294, Log Avg loss: 0.24865792, Global Avg Loss: 1.05902052, Time: 0.0208 Steps: 51090, Updated lr: 0.000052 ***** Running evaluation checkpoint-51096 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-51096 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.516616, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.346193, "eval_total_loss": 243.373612, "eval_mae": 0.406609, "eval_mse": 0.346325, "eval_r2": 0.779853, "eval_sp_statistic": 0.877585, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.90223, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.875313, "test_total_loss": 439.407037, "test_mae": 0.728918, "test_mse": 0.875423, "test_r2": 0.434995, "test_sp_statistic": 0.771678, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.791789, "test_ps_pvalue": 0.0, "lr": 5.249312470365102e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0589185048914185, "train_cur_epoch_loss": 526.3356323409826, "train_cur_epoch_avg_loss": 0.2472219973419364, "train_cur_epoch_time": 44.516615867614746, "train_cur_epoch_avg_time": 0.020909636386855213, "epoch": 24, "step": 51096} ################################################## Training, Epoch: 0025, Batch: 000004, Sample Num: 64, Cur Loss: 0.29299200, Cur Avg Loss: 0.27539127, Log Avg loss: 0.22432164, Global Avg Loss: 1.05885717, Time: 0.0245 Steps: 51100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000014, Sample Num: 224, Cur Loss: 0.15396881, Cur Avg Loss: 0.26210011, Log Avg loss: 0.25678364, Global Avg Loss: 1.05870024, Time: 0.0209 Steps: 51110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000024, Sample Num: 384, Cur Loss: 0.12485253, Cur Avg Loss: 0.27277642, Log Avg loss: 0.28772326, Global Avg Loss: 1.05854942, Time: 0.0209 Steps: 51120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000034, Sample Num: 544, Cur Loss: 0.12874055, Cur Avg Loss: 0.24501862, Log Avg loss: 0.17839990, Global Avg Loss: 1.05837728, Time: 0.0209 Steps: 51130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000044, Sample Num: 704, Cur Loss: 0.65042520, Cur Avg Loss: 0.25245972, Log Avg loss: 0.27775944, Global Avg Loss: 1.05822464, Time: 0.0209 Steps: 51140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000054, Sample Num: 864, Cur Loss: 0.38439959, Cur Avg Loss: 0.26261431, Log Avg loss: 0.30729452, Global Avg Loss: 1.05807783, Time: 0.0208 Steps: 51150, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000064, Sample Num: 1024, Cur Loss: 0.11735355, Cur Avg Loss: 0.26630630, Log Avg loss: 0.28624304, Global Avg Loss: 1.05792696, Time: 0.0209 Steps: 51160, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000074, Sample Num: 1184, Cur Loss: 0.32740730, Cur Avg Loss: 0.26497117, Log Avg loss: 0.25642633, Global Avg Loss: 1.05777033, Time: 0.0209 Steps: 51170, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000084, Sample Num: 1344, Cur Loss: 0.18022278, Cur Avg Loss: 0.25482098, Log Avg loss: 0.17970964, Global Avg Loss: 1.05759877, Time: 0.0209 Steps: 51180, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000094, Sample Num: 1504, Cur Loss: 0.31682155, Cur Avg Loss: 0.25217384, Log Avg loss: 0.22993781, Global Avg Loss: 1.05743708, Time: 0.0208 Steps: 51190, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000104, Sample Num: 1664, Cur Loss: 0.12401551, Cur Avg Loss: 0.24692563, Log Avg loss: 0.19759246, Global Avg Loss: 1.05726914, Time: 0.0209 Steps: 51200, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000114, Sample Num: 1824, Cur Loss: 0.56474423, Cur Avg Loss: 0.25367515, Log Avg loss: 0.32387014, Global Avg Loss: 1.05712593, Time: 0.0208 Steps: 51210, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000124, Sample Num: 1984, Cur Loss: 0.12095463, Cur Avg Loss: 0.25190046, Log Avg loss: 0.23166897, Global Avg Loss: 1.05696477, Time: 0.0208 Steps: 51220, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000134, Sample Num: 2144, Cur Loss: 0.18752557, Cur Avg Loss: 0.24715310, Log Avg loss: 0.18828589, Global Avg Loss: 1.05679521, Time: 0.0209 Steps: 51230, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000144, Sample Num: 2304, Cur Loss: 0.29108214, Cur Avg Loss: 0.25099481, Log Avg loss: 0.30247364, Global Avg Loss: 1.05664799, Time: 0.0208 Steps: 51240, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000154, Sample Num: 2464, Cur Loss: 0.45824996, Cur Avg Loss: 0.24882408, Log Avg loss: 0.21756556, Global Avg Loss: 1.05648427, Time: 0.0208 Steps: 51250, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000164, Sample Num: 2624, Cur Loss: 0.11956376, Cur Avg Loss: 0.24833469, Log Avg loss: 0.24079814, Global Avg Loss: 1.05632514, Time: 0.0209 Steps: 51260, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000174, Sample Num: 2784, Cur Loss: 0.11212189, Cur Avg Loss: 0.24624882, Log Avg loss: 0.21204062, Global Avg Loss: 1.05616047, Time: 0.0208 Steps: 51270, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000184, Sample Num: 2944, Cur Loss: 0.12319493, Cur Avg Loss: 0.24663020, Log Avg loss: 0.25326620, Global Avg Loss: 1.05600390, Time: 0.0208 Steps: 51280, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000194, Sample Num: 3104, Cur Loss: 0.18055013, Cur Avg Loss: 0.24151246, Log Avg loss: 0.14734608, Global Avg Loss: 1.05582674, Time: 0.0209 Steps: 51290, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000204, Sample Num: 3264, Cur Loss: 0.08257672, Cur Avg Loss: 0.24145113, Log Avg loss: 0.24026115, Global Avg Loss: 1.05566776, Time: 0.0208 Steps: 51300, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000214, Sample Num: 3424, Cur Loss: 0.22858097, Cur Avg Loss: 0.24474578, Log Avg loss: 0.31195678, Global Avg Loss: 1.05552281, Time: 0.0209 Steps: 51310, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000224, Sample Num: 3584, Cur Loss: 0.09810445, Cur Avg Loss: 0.24678072, Log Avg loss: 0.29032848, Global Avg Loss: 1.05537371, Time: 0.0208 Steps: 51320, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000234, Sample Num: 3744, Cur Loss: 0.18020028, Cur Avg Loss: 0.24498078, Log Avg loss: 0.20466195, Global Avg Loss: 1.05520798, Time: 0.0209 Steps: 51330, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000244, Sample Num: 3904, Cur Loss: 0.11184502, Cur Avg Loss: 0.24390250, Log Avg loss: 0.21867094, Global Avg Loss: 1.05504504, Time: 0.0209 Steps: 51340, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000254, Sample Num: 4064, Cur Loss: 0.15492804, Cur Avg Loss: 0.24635400, Log Avg loss: 0.30617042, Global Avg Loss: 1.05489920, Time: 0.0208 Steps: 51350, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000264, Sample Num: 4224, Cur Loss: 0.09040727, Cur Avg Loss: 0.24819789, Log Avg loss: 0.29503267, Global Avg Loss: 1.05475125, Time: 0.0208 Steps: 51360, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000274, Sample Num: 4384, Cur Loss: 0.10998274, Cur Avg Loss: 0.24799751, Log Avg loss: 0.24270770, Global Avg Loss: 1.05459317, Time: 0.0208 Steps: 51370, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000284, Sample Num: 4544, Cur Loss: 0.19033004, Cur Avg Loss: 0.24769027, Log Avg loss: 0.23927178, Global Avg Loss: 1.05443449, Time: 0.0209 Steps: 51380, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000294, Sample Num: 4704, Cur Loss: 0.11059675, Cur Avg Loss: 0.24599160, Log Avg loss: 0.19774944, Global Avg Loss: 1.05426778, Time: 0.0208 Steps: 51390, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000304, Sample Num: 4864, Cur Loss: 0.18031013, Cur Avg Loss: 0.24460357, Log Avg loss: 0.20379551, Global Avg Loss: 1.05410232, Time: 0.0208 Steps: 51400, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000314, Sample Num: 5024, Cur Loss: 0.53508973, Cur Avg Loss: 0.24543482, Log Avg loss: 0.27070459, Global Avg Loss: 1.05394994, Time: 0.0208 Steps: 51410, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000324, Sample Num: 5184, Cur Loss: 0.11544874, Cur Avg Loss: 0.24450505, Log Avg loss: 0.21531030, Global Avg Loss: 1.05378684, Time: 0.0208 Steps: 51420, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000334, Sample Num: 5344, Cur Loss: 0.28825876, Cur Avg Loss: 0.24454874, Log Avg loss: 0.24596426, Global Avg Loss: 1.05362977, Time: 0.0208 Steps: 51430, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000344, Sample Num: 5504, Cur Loss: 0.29565528, Cur Avg Loss: 0.24597920, Log Avg loss: 0.29375658, Global Avg Loss: 1.05348205, Time: 0.0208 Steps: 51440, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000354, Sample Num: 5664, Cur Loss: 0.43867031, Cur Avg Loss: 0.24735416, Log Avg loss: 0.29465294, Global Avg Loss: 1.05333456, Time: 0.0208 Steps: 51450, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000364, Sample Num: 5824, Cur Loss: 0.16365042, Cur Avg Loss: 0.24671626, Log Avg loss: 0.22413474, Global Avg Loss: 1.05317343, Time: 0.0208 Steps: 51460, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000374, Sample Num: 5984, Cur Loss: 0.42422211, Cur Avg Loss: 0.24739457, Log Avg loss: 0.27208479, Global Avg Loss: 1.05302167, Time: 0.0208 Steps: 51470, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000384, Sample Num: 6144, Cur Loss: 0.48951679, Cur Avg Loss: 0.24628589, Log Avg loss: 0.20482140, Global Avg Loss: 1.05285691, Time: 0.0208 Steps: 51480, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000394, Sample Num: 6304, Cur Loss: 0.14409590, Cur Avg Loss: 0.24668299, Log Avg loss: 0.26193170, Global Avg Loss: 1.05270330, Time: 0.0208 Steps: 51490, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000404, Sample Num: 6464, Cur Loss: 0.29402387, Cur Avg Loss: 0.24685316, Log Avg loss: 0.25355780, Global Avg Loss: 1.05254813, Time: 0.0208 Steps: 51500, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000414, Sample Num: 6624, Cur Loss: 0.31507385, Cur Avg Loss: 0.24581483, Log Avg loss: 0.20386627, Global Avg Loss: 1.05238337, Time: 0.0208 Steps: 51510, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000424, Sample Num: 6784, Cur Loss: 0.13148916, Cur Avg Loss: 0.24529800, Log Avg loss: 0.22390136, Global Avg Loss: 1.05222256, Time: 0.0208 Steps: 51520, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000434, Sample Num: 6944, Cur Loss: 0.32862177, Cur Avg Loss: 0.24501845, Log Avg loss: 0.23316555, Global Avg Loss: 1.05206361, Time: 0.0208 Steps: 51530, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000444, Sample Num: 7104, Cur Loss: 0.54905224, Cur Avg Loss: 0.24542363, Log Avg loss: 0.26300849, Global Avg Loss: 1.05191052, Time: 0.0208 Steps: 51540, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000454, Sample Num: 7264, Cur Loss: 0.24455881, Cur Avg Loss: 0.24521713, Log Avg loss: 0.23604848, Global Avg Loss: 1.05175225, Time: 0.0208 Steps: 51550, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000464, Sample Num: 7424, Cur Loss: 0.24991508, Cur Avg Loss: 0.24526178, Log Avg loss: 0.24728885, Global Avg Loss: 1.05159623, Time: 0.0208 Steps: 51560, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000474, Sample Num: 7584, Cur Loss: 0.04281024, Cur Avg Loss: 0.24344258, Log Avg loss: 0.15903172, Global Avg Loss: 1.05142315, Time: 0.0208 Steps: 51570, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000484, Sample Num: 7744, Cur Loss: 0.15642545, Cur Avg Loss: 0.24181763, Log Avg loss: 0.16479496, Global Avg Loss: 1.05125125, Time: 0.0208 Steps: 51580, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000494, Sample Num: 7904, Cur Loss: 0.15204717, Cur Avg Loss: 0.24114616, Log Avg loss: 0.20864681, Global Avg Loss: 1.05108793, Time: 0.0208 Steps: 51590, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000504, Sample Num: 8064, Cur Loss: 0.24354184, Cur Avg Loss: 0.24037740, Log Avg loss: 0.20240053, Global Avg Loss: 1.05092345, Time: 0.0208 Steps: 51600, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000514, Sample Num: 8224, Cur Loss: 0.26922810, Cur Avg Loss: 0.24113663, Log Avg loss: 0.27940208, Global Avg Loss: 1.05077396, Time: 0.0246 Steps: 51610, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000524, Sample Num: 8384, Cur Loss: 0.28087804, Cur Avg Loss: 0.24114779, Log Avg loss: 0.24172151, Global Avg Loss: 1.05061723, Time: 0.0208 Steps: 51620, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000534, Sample Num: 8544, Cur Loss: 0.70643514, Cur Avg Loss: 0.24133167, Log Avg loss: 0.25096710, Global Avg Loss: 1.05046235, Time: 0.0208 Steps: 51630, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000544, Sample Num: 8704, Cur Loss: 0.24261700, Cur Avg Loss: 0.24053343, Log Avg loss: 0.19790730, Global Avg Loss: 1.05029725, Time: 0.0208 Steps: 51640, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000554, Sample Num: 8864, Cur Loss: 0.09375101, Cur Avg Loss: 0.24007529, Log Avg loss: 0.21515220, Global Avg Loss: 1.05013556, Time: 0.0208 Steps: 51650, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000564, Sample Num: 9024, Cur Loss: 0.10806603, Cur Avg Loss: 0.23962143, Log Avg loss: 0.21447755, Global Avg Loss: 1.04997380, Time: 0.0208 Steps: 51660, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000574, Sample Num: 9184, Cur Loss: 0.09552719, Cur Avg Loss: 0.23852864, Log Avg loss: 0.17689536, Global Avg Loss: 1.04980483, Time: 0.0208 Steps: 51670, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000584, Sample Num: 9344, Cur Loss: 0.36086515, Cur Avg Loss: 0.23869564, Log Avg loss: 0.24828158, Global Avg Loss: 1.04964973, Time: 0.0208 Steps: 51680, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000594, Sample Num: 9504, Cur Loss: 0.39313310, Cur Avg Loss: 0.23939936, Log Avg loss: 0.28049645, Global Avg Loss: 1.04950093, Time: 0.0207 Steps: 51690, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000604, Sample Num: 9664, Cur Loss: 0.18581510, Cur Avg Loss: 0.23862150, Log Avg loss: 0.19241692, Global Avg Loss: 1.04933515, Time: 0.0209 Steps: 51700, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000614, Sample Num: 9824, Cur Loss: 0.16958669, Cur Avg Loss: 0.23956817, Log Avg loss: 0.29674662, Global Avg Loss: 1.04918961, Time: 0.0209 Steps: 51710, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000624, Sample Num: 9984, Cur Loss: 0.13898140, Cur Avg Loss: 0.24056982, Log Avg loss: 0.30207132, Global Avg Loss: 1.04904516, Time: 0.0208 Steps: 51720, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000634, Sample Num: 10144, Cur Loss: 0.16801764, Cur Avg Loss: 0.24043061, Log Avg loss: 0.23174384, Global Avg Loss: 1.04888716, Time: 0.0209 Steps: 51730, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000644, Sample Num: 10304, Cur Loss: 0.21069525, Cur Avg Loss: 0.23973719, Log Avg loss: 0.19577475, Global Avg Loss: 1.04872228, Time: 0.0208 Steps: 51740, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000654, Sample Num: 10464, Cur Loss: 0.27741995, Cur Avg Loss: 0.24002318, Log Avg loss: 0.25844082, Global Avg Loss: 1.04856957, Time: 0.0208 Steps: 51750, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000664, Sample Num: 10624, Cur Loss: 0.43140793, Cur Avg Loss: 0.24024394, Log Avg loss: 0.25468163, Global Avg Loss: 1.04841619, Time: 0.0208 Steps: 51760, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000674, Sample Num: 10784, Cur Loss: 0.15458186, Cur Avg Loss: 0.23938652, Log Avg loss: 0.18245360, Global Avg Loss: 1.04824892, Time: 0.0208 Steps: 51770, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000684, Sample Num: 10944, Cur Loss: 0.41663176, Cur Avg Loss: 0.23985845, Log Avg loss: 0.27166663, Global Avg Loss: 1.04809894, Time: 0.0207 Steps: 51780, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000694, Sample Num: 11104, Cur Loss: 0.19166842, Cur Avg Loss: 0.24089361, Log Avg loss: 0.31169851, Global Avg Loss: 1.04795675, Time: 0.0208 Steps: 51790, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000704, Sample Num: 11264, Cur Loss: 0.14140749, Cur Avg Loss: 0.24065167, Log Avg loss: 0.22386103, Global Avg Loss: 1.04779766, Time: 0.0208 Steps: 51800, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000714, Sample Num: 11424, Cur Loss: 0.08679849, Cur Avg Loss: 0.24014404, Log Avg loss: 0.20440722, Global Avg Loss: 1.04763487, Time: 0.0207 Steps: 51810, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000724, Sample Num: 11584, Cur Loss: 0.07261774, Cur Avg Loss: 0.23882160, Log Avg loss: 0.14439928, Global Avg Loss: 1.04746057, Time: 0.0207 Steps: 51820, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000734, Sample Num: 11744, Cur Loss: 0.37117827, Cur Avg Loss: 0.23968215, Log Avg loss: 0.30198571, Global Avg Loss: 1.04731674, Time: 0.0208 Steps: 51830, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000744, Sample Num: 11904, Cur Loss: 0.26122677, Cur Avg Loss: 0.23935737, Log Avg loss: 0.21551870, Global Avg Loss: 1.04715628, Time: 0.0208 Steps: 51840, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000754, Sample Num: 12064, Cur Loss: 0.10632318, Cur Avg Loss: 0.23864572, Log Avg loss: 0.18569873, Global Avg Loss: 1.04699014, Time: 0.0208 Steps: 51850, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000764, Sample Num: 12224, Cur Loss: 0.15769897, Cur Avg Loss: 0.23961468, Log Avg loss: 0.31267425, Global Avg Loss: 1.04684854, Time: 0.0207 Steps: 51860, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000774, Sample Num: 12384, Cur Loss: 0.15627109, Cur Avg Loss: 0.23924331, Log Avg loss: 0.21087054, Global Avg Loss: 1.04668738, Time: 0.0245 Steps: 51870, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000784, Sample Num: 12544, Cur Loss: 0.19435997, Cur Avg Loss: 0.23872540, Log Avg loss: 0.19863908, Global Avg Loss: 1.04652391, Time: 0.0208 Steps: 51880, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000794, Sample Num: 12704, Cur Loss: 0.39090449, Cur Avg Loss: 0.24146945, Log Avg loss: 0.45660316, Global Avg Loss: 1.04641023, Time: 0.0208 Steps: 51890, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000804, Sample Num: 12864, Cur Loss: 0.08201447, Cur Avg Loss: 0.24220352, Log Avg loss: 0.30048866, Global Avg Loss: 1.04626650, Time: 0.0208 Steps: 51900, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000814, Sample Num: 13024, Cur Loss: 0.42168751, Cur Avg Loss: 0.24204259, Log Avg loss: 0.22910399, Global Avg Loss: 1.04610908, Time: 0.0208 Steps: 51910, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000824, Sample Num: 13184, Cur Loss: 0.03532403, Cur Avg Loss: 0.24183317, Log Avg loss: 0.22478601, Global Avg Loss: 1.04595089, Time: 0.0208 Steps: 51920, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000834, Sample Num: 13344, Cur Loss: 0.07032144, Cur Avg Loss: 0.24144315, Log Avg loss: 0.20930584, Global Avg Loss: 1.04578978, Time: 0.0208 Steps: 51930, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000844, Sample Num: 13504, Cur Loss: 0.15255374, Cur Avg Loss: 0.24058223, Log Avg loss: 0.16878151, Global Avg Loss: 1.04562093, Time: 0.0208 Steps: 51940, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000854, Sample Num: 13664, Cur Loss: 0.13174590, Cur Avg Loss: 0.23978300, Log Avg loss: 0.17232817, Global Avg Loss: 1.04545283, Time: 0.0208 Steps: 51950, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000864, Sample Num: 13824, Cur Loss: 0.59872979, Cur Avg Loss: 0.24116931, Log Avg loss: 0.35955964, Global Avg Loss: 1.04532083, Time: 0.0208 Steps: 51960, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000874, Sample Num: 13984, Cur Loss: 0.67591661, Cur Avg Loss: 0.24313875, Log Avg loss: 0.41329872, Global Avg Loss: 1.04519921, Time: 0.0209 Steps: 51970, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000884, Sample Num: 14144, Cur Loss: 0.25173301, Cur Avg Loss: 0.24310701, Log Avg loss: 0.24033324, Global Avg Loss: 1.04504437, Time: 0.0208 Steps: 51980, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000894, Sample Num: 14304, Cur Loss: 0.20998818, Cur Avg Loss: 0.24323792, Log Avg loss: 0.25481024, Global Avg Loss: 1.04489238, Time: 0.0208 Steps: 51990, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000904, Sample Num: 14464, Cur Loss: 0.34775057, Cur Avg Loss: 0.24264976, Log Avg loss: 0.19006806, Global Avg Loss: 1.04472799, Time: 0.0208 Steps: 52000, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000914, Sample Num: 14624, Cur Loss: 0.24990577, Cur Avg Loss: 0.24296204, Log Avg loss: 0.27119264, Global Avg Loss: 1.04457926, Time: 0.0208 Steps: 52010, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000924, Sample Num: 14784, Cur Loss: 0.18335536, Cur Avg Loss: 0.24264899, Log Avg loss: 0.21403567, Global Avg Loss: 1.04441960, Time: 0.0208 Steps: 52020, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000934, Sample Num: 14944, Cur Loss: 0.18155414, Cur Avg Loss: 0.24234772, Log Avg loss: 0.21451043, Global Avg Loss: 1.04426009, Time: 0.0208 Steps: 52030, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000944, Sample Num: 15104, Cur Loss: 0.20703189, Cur Avg Loss: 0.24274927, Log Avg loss: 0.28025452, Global Avg Loss: 1.04411328, Time: 0.0208 Steps: 52040, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000954, Sample Num: 15264, Cur Loss: 0.11176787, Cur Avg Loss: 0.24244685, Log Avg loss: 0.21389798, Global Avg Loss: 1.04395378, Time: 0.0208 Steps: 52050, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000964, Sample Num: 15424, Cur Loss: 0.20462784, Cur Avg Loss: 0.24293362, Log Avg loss: 0.28937169, Global Avg Loss: 1.04380883, Time: 0.0208 Steps: 52060, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000974, Sample Num: 15584, Cur Loss: 0.21748632, Cur Avg Loss: 0.24248298, Log Avg loss: 0.19904070, Global Avg Loss: 1.04364660, Time: 0.0208 Steps: 52070, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000984, Sample Num: 15744, Cur Loss: 0.52338916, Cur Avg Loss: 0.24352786, Log Avg loss: 0.34529927, Global Avg Loss: 1.04351251, Time: 0.0208 Steps: 52080, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000994, Sample Num: 15904, Cur Loss: 0.17324877, Cur Avg Loss: 0.24395292, Log Avg loss: 0.28577890, Global Avg Loss: 1.04336704, Time: 0.0208 Steps: 52090, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001004, Sample Num: 16064, Cur Loss: 0.25099987, Cur Avg Loss: 0.24358578, Log Avg loss: 0.20709203, Global Avg Loss: 1.04320653, Time: 0.0208 Steps: 52100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001014, Sample Num: 16224, Cur Loss: 0.28760618, Cur Avg Loss: 0.24363284, Log Avg loss: 0.24835763, Global Avg Loss: 1.04305399, Time: 0.0208 Steps: 52110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001024, Sample Num: 16384, Cur Loss: 0.23424532, Cur Avg Loss: 0.24338726, Log Avg loss: 0.21848563, Global Avg Loss: 1.04289579, Time: 0.0253 Steps: 52120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001034, Sample Num: 16544, Cur Loss: 0.73134810, Cur Avg Loss: 0.24362287, Log Avg loss: 0.26774925, Global Avg Loss: 1.04274709, Time: 0.0208 Steps: 52130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001044, Sample Num: 16704, Cur Loss: 0.33079374, Cur Avg Loss: 0.24441877, Log Avg loss: 0.32671467, Global Avg Loss: 1.04260976, Time: 0.0208 Steps: 52140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001054, Sample Num: 16864, Cur Loss: 0.20476002, Cur Avg Loss: 0.24481933, Log Avg loss: 0.28663862, Global Avg Loss: 1.04246480, Time: 0.0208 Steps: 52150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001064, Sample Num: 17024, Cur Loss: 0.13834968, Cur Avg Loss: 0.24443478, Log Avg loss: 0.20390266, Global Avg Loss: 1.04230404, Time: 0.0208 Steps: 52160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001074, Sample Num: 17184, Cur Loss: 0.06584363, Cur Avg Loss: 0.24435272, Log Avg loss: 0.23562151, Global Avg Loss: 1.04214941, Time: 0.0208 Steps: 52170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001084, Sample Num: 17344, Cur Loss: 0.22950618, Cur Avg Loss: 0.24417554, Log Avg loss: 0.22514605, Global Avg Loss: 1.04199284, Time: 0.0208 Steps: 52180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001094, Sample Num: 17504, Cur Loss: 0.21333331, Cur Avg Loss: 0.24448079, Log Avg loss: 0.27757006, Global Avg Loss: 1.04184637, Time: 0.0208 Steps: 52190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001104, Sample Num: 17664, Cur Loss: 0.24143787, Cur Avg Loss: 0.24409780, Log Avg loss: 0.20219856, Global Avg Loss: 1.04168552, Time: 0.0208 Steps: 52200, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001114, Sample Num: 17824, Cur Loss: 0.26763314, Cur Avg Loss: 0.24389732, Log Avg loss: 0.22176484, Global Avg Loss: 1.04152847, Time: 0.0208 Steps: 52210, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001124, Sample Num: 17984, Cur Loss: 0.10029603, Cur Avg Loss: 0.24323541, Log Avg loss: 0.16949904, Global Avg Loss: 1.04136148, Time: 0.0208 Steps: 52220, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001134, Sample Num: 18144, Cur Loss: 0.47329021, Cur Avg Loss: 0.24311243, Log Avg loss: 0.22928927, Global Avg Loss: 1.04120600, Time: 0.0210 Steps: 52230, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001144, Sample Num: 18304, Cur Loss: 0.09497363, Cur Avg Loss: 0.24294867, Log Avg loss: 0.22437778, Global Avg Loss: 1.04104964, Time: 0.0207 Steps: 52240, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001154, Sample Num: 18464, Cur Loss: 0.94535291, Cur Avg Loss: 0.24306447, Log Avg loss: 0.25631254, Global Avg Loss: 1.04089945, Time: 0.0208 Steps: 52250, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001164, Sample Num: 18624, Cur Loss: 0.26544225, Cur Avg Loss: 0.24343160, Log Avg loss: 0.28579835, Global Avg Loss: 1.04075496, Time: 0.0208 Steps: 52260, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001174, Sample Num: 18784, Cur Loss: 0.19759431, Cur Avg Loss: 0.24341337, Log Avg loss: 0.24129088, Global Avg Loss: 1.04060201, Time: 0.0207 Steps: 52270, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001184, Sample Num: 18944, Cur Loss: 0.14776714, Cur Avg Loss: 0.24503644, Log Avg loss: 0.43558464, Global Avg Loss: 1.04048629, Time: 0.0208 Steps: 52280, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001194, Sample Num: 19104, Cur Loss: 0.15137947, Cur Avg Loss: 0.24525484, Log Avg loss: 0.27111361, Global Avg Loss: 1.04033915, Time: 0.0208 Steps: 52290, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001204, Sample Num: 19264, Cur Loss: 0.33740169, Cur Avg Loss: 0.24513938, Log Avg loss: 0.23135310, Global Avg Loss: 1.04018447, Time: 0.0207 Steps: 52300, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001214, Sample Num: 19424, Cur Loss: 0.32515413, Cur Avg Loss: 0.24473038, Log Avg loss: 0.19548749, Global Avg Loss: 1.04002299, Time: 0.0207 Steps: 52310, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001224, Sample Num: 19584, Cur Loss: 0.17162013, Cur Avg Loss: 0.24479025, Log Avg loss: 0.25205854, Global Avg Loss: 1.03987239, Time: 0.0208 Steps: 52320, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001234, Sample Num: 19744, Cur Loss: 0.07615028, Cur Avg Loss: 0.24465970, Log Avg loss: 0.22868060, Global Avg Loss: 1.03971737, Time: 0.0208 Steps: 52330, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001244, Sample Num: 19904, Cur Loss: 0.41379744, Cur Avg Loss: 0.24515718, Log Avg loss: 0.30654620, Global Avg Loss: 1.03957729, Time: 0.0208 Steps: 52340, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001254, Sample Num: 20064, Cur Loss: 0.29335922, Cur Avg Loss: 0.24474543, Log Avg loss: 0.19352391, Global Avg Loss: 1.03941568, Time: 0.0208 Steps: 52350, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001264, Sample Num: 20224, Cur Loss: 0.25488046, Cur Avg Loss: 0.24557126, Log Avg loss: 0.34912982, Global Avg Loss: 1.03928384, Time: 0.0208 Steps: 52360, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001274, Sample Num: 20384, Cur Loss: 0.65347183, Cur Avg Loss: 0.24594304, Log Avg loss: 0.29293631, Global Avg Loss: 1.03914133, Time: 0.0208 Steps: 52370, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001284, Sample Num: 20544, Cur Loss: 0.39483988, Cur Avg Loss: 0.24595955, Log Avg loss: 0.24806288, Global Avg Loss: 1.03899030, Time: 0.0245 Steps: 52380, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001294, Sample Num: 20704, Cur Loss: 0.29217428, Cur Avg Loss: 0.24582487, Log Avg loss: 0.22853150, Global Avg Loss: 1.03883560, Time: 0.0208 Steps: 52390, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001304, Sample Num: 20864, Cur Loss: 0.08924500, Cur Avg Loss: 0.24585102, Log Avg loss: 0.24923454, Global Avg Loss: 1.03868492, Time: 0.0208 Steps: 52400, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001314, Sample Num: 21024, Cur Loss: 0.44444734, Cur Avg Loss: 0.24652040, Log Avg loss: 0.33380824, Global Avg Loss: 1.03855042, Time: 0.0207 Steps: 52410, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001324, Sample Num: 21184, Cur Loss: 0.36383632, Cur Avg Loss: 0.24664269, Log Avg loss: 0.26271153, Global Avg Loss: 1.03840242, Time: 0.0208 Steps: 52420, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001334, Sample Num: 21344, Cur Loss: 0.15111819, Cur Avg Loss: 0.24751645, Log Avg loss: 0.36320210, Global Avg Loss: 1.03827364, Time: 0.0208 Steps: 52430, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001344, Sample Num: 21504, Cur Loss: 0.15801404, Cur Avg Loss: 0.24722022, Log Avg loss: 0.20770279, Global Avg Loss: 1.03811525, Time: 0.0208 Steps: 52440, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001354, Sample Num: 21664, Cur Loss: 0.20800687, Cur Avg Loss: 0.24746548, Log Avg loss: 0.28042860, Global Avg Loss: 1.03797079, Time: 0.0208 Steps: 52450, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001364, Sample Num: 21824, Cur Loss: 0.10083810, Cur Avg Loss: 0.24755036, Log Avg loss: 0.25904270, Global Avg Loss: 1.03782231, Time: 0.0208 Steps: 52460, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001374, Sample Num: 21984, Cur Loss: 0.04191480, Cur Avg Loss: 0.24777492, Log Avg loss: 0.27840596, Global Avg Loss: 1.03767758, Time: 0.0208 Steps: 52470, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001384, Sample Num: 22144, Cur Loss: 0.56957716, Cur Avg Loss: 0.24795671, Log Avg loss: 0.27293356, Global Avg Loss: 1.03753186, Time: 0.0207 Steps: 52480, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001394, Sample Num: 22304, Cur Loss: 0.31503272, Cur Avg Loss: 0.24784487, Log Avg loss: 0.23236613, Global Avg Loss: 1.03737847, Time: 0.0208 Steps: 52490, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001404, Sample Num: 22464, Cur Loss: 0.18115452, Cur Avg Loss: 0.24744346, Log Avg loss: 0.19148797, Global Avg Loss: 1.03721734, Time: 0.0207 Steps: 52500, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001414, Sample Num: 22624, Cur Loss: 0.57741100, Cur Avg Loss: 0.24825808, Log Avg loss: 0.36263044, Global Avg Loss: 1.03708888, Time: 0.0207 Steps: 52510, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001424, Sample Num: 22784, Cur Loss: 0.14603412, Cur Avg Loss: 0.24811315, Log Avg loss: 0.22761974, Global Avg Loss: 1.03693475, Time: 0.0207 Steps: 52520, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001434, Sample Num: 22944, Cur Loss: 0.06158511, Cur Avg Loss: 0.24764868, Log Avg loss: 0.18150860, Global Avg Loss: 1.03677190, Time: 0.0207 Steps: 52530, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001444, Sample Num: 23104, Cur Loss: 0.15963721, Cur Avg Loss: 0.24710679, Log Avg loss: 0.16939920, Global Avg Loss: 1.03660682, Time: 0.0208 Steps: 52540, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001454, Sample Num: 23264, Cur Loss: 0.47998565, Cur Avg Loss: 0.24710005, Log Avg loss: 0.24612759, Global Avg Loss: 1.03645639, Time: 0.0208 Steps: 52550, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001464, Sample Num: 23424, Cur Loss: 0.30463549, Cur Avg Loss: 0.24677867, Log Avg loss: 0.20004886, Global Avg Loss: 1.03629726, Time: 0.0207 Steps: 52560, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001474, Sample Num: 23584, Cur Loss: 0.40645346, Cur Avg Loss: 0.24676129, Log Avg loss: 0.24421727, Global Avg Loss: 1.03614659, Time: 0.0208 Steps: 52570, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001484, Sample Num: 23744, Cur Loss: 0.10189877, Cur Avg Loss: 0.24654824, Log Avg loss: 0.21514490, Global Avg Loss: 1.03599044, Time: 0.0208 Steps: 52580, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001494, Sample Num: 23904, Cur Loss: 0.15006202, Cur Avg Loss: 0.24612054, Log Avg loss: 0.18264983, Global Avg Loss: 1.03582818, Time: 0.0208 Steps: 52590, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001504, Sample Num: 24064, Cur Loss: 0.21777561, Cur Avg Loss: 0.24570704, Log Avg loss: 0.18392956, Global Avg Loss: 1.03566622, Time: 0.0207 Steps: 52600, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001514, Sample Num: 24224, Cur Loss: 0.07828467, Cur Avg Loss: 0.24521787, Log Avg loss: 0.17164800, Global Avg Loss: 1.03550199, Time: 0.0207 Steps: 52610, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001524, Sample Num: 24384, Cur Loss: 0.21696034, Cur Avg Loss: 0.24499814, Log Avg loss: 0.21173018, Global Avg Loss: 1.03534544, Time: 0.0208 Steps: 52620, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001534, Sample Num: 24544, Cur Loss: 0.06159317, Cur Avg Loss: 0.24459949, Log Avg loss: 0.18384510, Global Avg Loss: 1.03518365, Time: 0.0207 Steps: 52630, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001544, Sample Num: 24704, Cur Loss: 0.14848788, Cur Avg Loss: 0.24395488, Log Avg loss: 0.14507216, Global Avg Loss: 1.03501456, Time: 0.0209 Steps: 52640, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001554, Sample Num: 24864, Cur Loss: 0.07990774, Cur Avg Loss: 0.24355402, Log Avg loss: 0.18166070, Global Avg Loss: 1.03485248, Time: 0.0208 Steps: 52650, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001564, Sample Num: 25024, Cur Loss: 0.09927423, Cur Avg Loss: 0.24332722, Log Avg loss: 0.20808288, Global Avg Loss: 1.03469547, Time: 0.0208 Steps: 52660, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001574, Sample Num: 25184, Cur Loss: 0.12544176, Cur Avg Loss: 0.24291064, Log Avg loss: 0.17775739, Global Avg Loss: 1.03453278, Time: 0.0208 Steps: 52670, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001584, Sample Num: 25344, Cur Loss: 0.40183473, Cur Avg Loss: 0.24314425, Log Avg loss: 0.27991502, Global Avg Loss: 1.03438953, Time: 0.0208 Steps: 52680, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001594, Sample Num: 25504, Cur Loss: 0.12073927, Cur Avg Loss: 0.24338912, Log Avg loss: 0.28217679, Global Avg Loss: 1.03424677, Time: 0.0207 Steps: 52690, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001604, Sample Num: 25664, Cur Loss: 0.09466572, Cur Avg Loss: 0.24296909, Log Avg loss: 0.17601637, Global Avg Loss: 1.03408392, Time: 0.0208 Steps: 52700, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001614, Sample Num: 25824, Cur Loss: 0.18972340, Cur Avg Loss: 0.24278199, Log Avg loss: 0.21277065, Global Avg Loss: 1.03392810, Time: 0.0208 Steps: 52710, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001624, Sample Num: 25984, Cur Loss: 0.38634822, Cur Avg Loss: 0.24266494, Log Avg loss: 0.22377308, Global Avg Loss: 1.03377443, Time: 0.0208 Steps: 52720, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001634, Sample Num: 26144, Cur Loss: 0.13084623, Cur Avg Loss: 0.24229492, Log Avg loss: 0.18220318, Global Avg Loss: 1.03361293, Time: 0.0208 Steps: 52730, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001644, Sample Num: 26304, Cur Loss: 0.17304781, Cur Avg Loss: 0.24193373, Log Avg loss: 0.18291610, Global Avg Loss: 1.03345163, Time: 0.0208 Steps: 52740, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001654, Sample Num: 26464, Cur Loss: 0.20975742, Cur Avg Loss: 0.24171780, Log Avg loss: 0.20621805, Global Avg Loss: 1.03329481, Time: 0.0208 Steps: 52750, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001664, Sample Num: 26624, Cur Loss: 0.22898246, Cur Avg Loss: 0.24202718, Log Avg loss: 0.29319847, Global Avg Loss: 1.03315453, Time: 0.0208 Steps: 52760, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001674, Sample Num: 26784, Cur Loss: 0.27594280, Cur Avg Loss: 0.24212318, Log Avg loss: 0.25809757, Global Avg Loss: 1.03300766, Time: 0.0208 Steps: 52770, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001684, Sample Num: 26944, Cur Loss: 0.37756997, Cur Avg Loss: 0.24199231, Log Avg loss: 0.22008570, Global Avg Loss: 1.03285364, Time: 0.0208 Steps: 52780, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001694, Sample Num: 27104, Cur Loss: 0.19164464, Cur Avg Loss: 0.24147794, Log Avg loss: 0.15485684, Global Avg Loss: 1.03268732, Time: 0.0208 Steps: 52790, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001704, Sample Num: 27264, Cur Loss: 0.11056709, Cur Avg Loss: 0.24122378, Log Avg loss: 0.19816905, Global Avg Loss: 1.03252927, Time: 0.0208 Steps: 52800, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001714, Sample Num: 27424, Cur Loss: 0.54156405, Cur Avg Loss: 0.24125622, Log Avg loss: 0.24678542, Global Avg Loss: 1.03238048, Time: 0.0208 Steps: 52810, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001724, Sample Num: 27584, Cur Loss: 0.07347569, Cur Avg Loss: 0.24143912, Log Avg loss: 0.27278752, Global Avg Loss: 1.03223667, Time: 0.0208 Steps: 52820, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001734, Sample Num: 27744, Cur Loss: 0.16151927, Cur Avg Loss: 0.24123125, Log Avg loss: 0.20539452, Global Avg Loss: 1.03208016, Time: 0.0208 Steps: 52830, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001744, Sample Num: 27904, Cur Loss: 0.07867694, Cur Avg Loss: 0.24125148, Log Avg loss: 0.24475953, Global Avg Loss: 1.03193116, Time: 0.0208 Steps: 52840, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001754, Sample Num: 28064, Cur Loss: 0.09105867, Cur Avg Loss: 0.24121991, Log Avg loss: 0.23571330, Global Avg Loss: 1.03178050, Time: 0.0208 Steps: 52850, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001764, Sample Num: 28224, Cur Loss: 0.09519389, Cur Avg Loss: 0.24099602, Log Avg loss: 0.20172645, Global Avg Loss: 1.03162348, Time: 0.0208 Steps: 52860, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001774, Sample Num: 28384, Cur Loss: 0.12577984, Cur Avg Loss: 0.24087384, Log Avg loss: 0.21932042, Global Avg Loss: 1.03146983, Time: 0.0208 Steps: 52870, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001784, Sample Num: 28544, Cur Loss: 0.37216368, Cur Avg Loss: 0.24068229, Log Avg loss: 0.20670180, Global Avg Loss: 1.03131386, Time: 0.0208 Steps: 52880, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001794, Sample Num: 28704, Cur Loss: 0.52227509, Cur Avg Loss: 0.24073242, Log Avg loss: 0.24967569, Global Avg Loss: 1.03116608, Time: 0.0246 Steps: 52890, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001804, Sample Num: 28864, Cur Loss: 0.34485561, Cur Avg Loss: 0.24061409, Log Avg loss: 0.21938492, Global Avg Loss: 1.03101262, Time: 0.0208 Steps: 52900, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001814, Sample Num: 29024, Cur Loss: 0.26161313, Cur Avg Loss: 0.24047709, Log Avg loss: 0.21576340, Global Avg Loss: 1.03085854, Time: 0.0208 Steps: 52910, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001824, Sample Num: 29184, Cur Loss: 0.08161956, Cur Avg Loss: 0.24091126, Log Avg loss: 0.31966887, Global Avg Loss: 1.03072415, Time: 0.0208 Steps: 52920, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001834, Sample Num: 29344, Cur Loss: 0.28904143, Cur Avg Loss: 0.24083991, Log Avg loss: 0.22782639, Global Avg Loss: 1.03057246, Time: 0.0208 Steps: 52930, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001844, Sample Num: 29504, Cur Loss: 0.25894171, Cur Avg Loss: 0.24118349, Log Avg loss: 0.30419598, Global Avg Loss: 1.03043525, Time: 0.0208 Steps: 52940, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001854, Sample Num: 29664, Cur Loss: 0.33622366, Cur Avg Loss: 0.24132518, Log Avg loss: 0.26745201, Global Avg Loss: 1.03029116, Time: 0.0208 Steps: 52950, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001864, Sample Num: 29824, Cur Loss: 0.37817568, Cur Avg Loss: 0.24156444, Log Avg loss: 0.28592366, Global Avg Loss: 1.03015060, Time: 0.0208 Steps: 52960, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001874, Sample Num: 29984, Cur Loss: 0.16252950, Cur Avg Loss: 0.24194927, Log Avg loss: 0.31368129, Global Avg Loss: 1.03001535, Time: 0.0208 Steps: 52970, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001884, Sample Num: 30144, Cur Loss: 0.34120315, Cur Avg Loss: 0.24226118, Log Avg loss: 0.30071416, Global Avg Loss: 1.02987769, Time: 0.0208 Steps: 52980, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001894, Sample Num: 30304, Cur Loss: 0.29411465, Cur Avg Loss: 0.24218601, Log Avg loss: 0.22802422, Global Avg Loss: 1.02972637, Time: 0.0208 Steps: 52990, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001904, Sample Num: 30464, Cur Loss: 0.19149971, Cur Avg Loss: 0.24229053, Log Avg loss: 0.26208583, Global Avg Loss: 1.02958153, Time: 0.0208 Steps: 53000, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001914, Sample Num: 30624, Cur Loss: 0.09513327, Cur Avg Loss: 0.24208528, Log Avg loss: 0.20300538, Global Avg Loss: 1.02942560, Time: 0.0208 Steps: 53010, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001924, Sample Num: 30784, Cur Loss: 0.17319591, Cur Avg Loss: 0.24211168, Log Avg loss: 0.24716564, Global Avg Loss: 1.02927806, Time: 0.0208 Steps: 53020, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001934, Sample Num: 30944, Cur Loss: 0.23802201, Cur Avg Loss: 0.24219439, Log Avg loss: 0.25810664, Global Avg Loss: 1.02913264, Time: 0.0208 Steps: 53030, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001944, Sample Num: 31104, Cur Loss: 0.49492386, Cur Avg Loss: 0.24258901, Log Avg loss: 0.31890965, Global Avg Loss: 1.02899874, Time: 0.0208 Steps: 53040, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001954, Sample Num: 31264, Cur Loss: 0.50650293, Cur Avg Loss: 0.24223515, Log Avg loss: 0.17344433, Global Avg Loss: 1.02883746, Time: 0.0208 Steps: 53050, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001964, Sample Num: 31424, Cur Loss: 0.21993428, Cur Avg Loss: 0.24218110, Log Avg loss: 0.23162028, Global Avg Loss: 1.02868721, Time: 0.0209 Steps: 53060, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001974, Sample Num: 31584, Cur Loss: 0.07711557, Cur Avg Loss: 0.24216178, Log Avg loss: 0.23836724, Global Avg Loss: 1.02853829, Time: 0.0208 Steps: 53070, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001984, Sample Num: 31744, Cur Loss: 0.27386478, Cur Avg Loss: 0.24230376, Log Avg loss: 0.27032902, Global Avg Loss: 1.02839545, Time: 0.0208 Steps: 53080, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001994, Sample Num: 31904, Cur Loss: 0.07270572, Cur Avg Loss: 0.24210238, Log Avg loss: 0.20214984, Global Avg Loss: 1.02823982, Time: 0.0208 Steps: 53090, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002004, Sample Num: 32064, Cur Loss: 0.23891521, Cur Avg Loss: 0.24205140, Log Avg loss: 0.23188530, Global Avg Loss: 1.02808985, Time: 0.0208 Steps: 53100, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002014, Sample Num: 32224, Cur Loss: 0.23263812, Cur Avg Loss: 0.24191216, Log Avg loss: 0.21400818, Global Avg Loss: 1.02793657, Time: 0.0208 Steps: 53110, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002024, Sample Num: 32384, Cur Loss: 0.14114918, Cur Avg Loss: 0.24180763, Log Avg loss: 0.22075613, Global Avg Loss: 1.02778461, Time: 0.0208 Steps: 53120, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002034, Sample Num: 32544, Cur Loss: 0.22144890, Cur Avg Loss: 0.24181130, Log Avg loss: 0.24255474, Global Avg Loss: 1.02763682, Time: 0.0208 Steps: 53130, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002044, Sample Num: 32704, Cur Loss: 0.21936303, Cur Avg Loss: 0.24181122, Log Avg loss: 0.24179418, Global Avg Loss: 1.02748894, Time: 0.0208 Steps: 53140, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002054, Sample Num: 32864, Cur Loss: 0.04586384, Cur Avg Loss: 0.24162183, Log Avg loss: 0.20291111, Global Avg Loss: 1.02733379, Time: 0.0209 Steps: 53150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002064, Sample Num: 33024, Cur Loss: 0.04142242, Cur Avg Loss: 0.24126807, Log Avg loss: 0.16860560, Global Avg Loss: 1.02717226, Time: 0.0208 Steps: 53160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002074, Sample Num: 33184, Cur Loss: 0.22611655, Cur Avg Loss: 0.24094506, Log Avg loss: 0.17427636, Global Avg Loss: 1.02701185, Time: 0.0207 Steps: 53170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002084, Sample Num: 33344, Cur Loss: 0.17819938, Cur Avg Loss: 0.24084447, Log Avg loss: 0.21998165, Global Avg Loss: 1.02686009, Time: 0.0208 Steps: 53180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002094, Sample Num: 33504, Cur Loss: 0.08163388, Cur Avg Loss: 0.24088199, Log Avg loss: 0.24870077, Global Avg Loss: 1.02671380, Time: 0.0208 Steps: 53190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002104, Sample Num: 33664, Cur Loss: 0.05772636, Cur Avg Loss: 0.24057415, Log Avg loss: 0.17611226, Global Avg Loss: 1.02655391, Time: 0.0208 Steps: 53200, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002114, Sample Num: 33824, Cur Loss: 0.30698696, Cur Avg Loss: 0.24057317, Log Avg loss: 0.24036781, Global Avg Loss: 1.02640616, Time: 0.0208 Steps: 53210, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002124, Sample Num: 33984, Cur Loss: 0.10900705, Cur Avg Loss: 0.24027554, Log Avg loss: 0.17735664, Global Avg Loss: 1.02624662, Time: 0.0208 Steps: 53220, Updated lr: 0.000050 ***** Running evaluation checkpoint-53225 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-53225 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.422452, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.277864, "eval_total_loss": 195.338371, "eval_mae": 0.379869, "eval_mse": 0.277949, "eval_r2": 0.823317, "eval_sp_statistic": 0.880896, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.909966, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.610202, "test_total_loss": 306.321321, "test_mae": 0.545662, "test_mse": 0.61032, "test_r2": 0.606094, "test_sp_statistic": 0.800146, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.840875, "test_ps_pvalue": 0.0, "lr": 5.0474158368895205e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0261637331520157, "train_cur_epoch_loss": 511.0647710841149, "train_cur_epoch_avg_loss": 0.240049211406348, "train_cur_epoch_time": 44.422452449798584, "train_cur_epoch_avg_time": 0.020865407444715163, "epoch": 25, "step": 53225} ################################################## Training, Epoch: 0026, Batch: 000005, Sample Num: 80, Cur Loss: 0.34653473, Cur Avg Loss: 0.28202714, Log Avg loss: 0.21296509, Global Avg Loss: 1.02609383, Time: 0.0245 Steps: 53230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000015, Sample Num: 240, Cur Loss: 0.07659381, Cur Avg Loss: 0.20708321, Log Avg loss: 0.16961124, Global Avg Loss: 1.02593296, Time: 0.0210 Steps: 53240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000025, Sample Num: 400, Cur Loss: 0.09261616, Cur Avg Loss: 0.23083867, Log Avg loss: 0.26647187, Global Avg Loss: 1.02579034, Time: 0.0210 Steps: 53250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000035, Sample Num: 560, Cur Loss: 0.30938944, Cur Avg Loss: 0.23871361, Log Avg loss: 0.25840098, Global Avg Loss: 1.02564626, Time: 0.0210 Steps: 53260, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000045, Sample Num: 720, Cur Loss: 0.23826844, Cur Avg Loss: 0.23176040, Log Avg loss: 0.20742416, Global Avg Loss: 1.02549266, Time: 0.0210 Steps: 53270, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000055, Sample Num: 880, Cur Loss: 0.49874756, Cur Avg Loss: 0.25877573, Log Avg loss: 0.38034469, Global Avg Loss: 1.02537157, Time: 0.0209 Steps: 53280, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000065, Sample Num: 1040, Cur Loss: 0.21856079, Cur Avg Loss: 0.25679294, Log Avg loss: 0.24588757, Global Avg Loss: 1.02522530, Time: 0.0209 Steps: 53290, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000075, Sample Num: 1200, Cur Loss: 0.11702724, Cur Avg Loss: 0.24424349, Log Avg loss: 0.16267210, Global Avg Loss: 1.02506347, Time: 0.0210 Steps: 53300, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000085, Sample Num: 1360, Cur Loss: 0.24723890, Cur Avg Loss: 0.25163511, Log Avg loss: 0.30707224, Global Avg Loss: 1.02492879, Time: 0.0209 Steps: 53310, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000095, Sample Num: 1520, Cur Loss: 0.11794175, Cur Avg Loss: 0.25161647, Log Avg loss: 0.25145804, Global Avg Loss: 1.02478373, Time: 0.0209 Steps: 53320, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000105, Sample Num: 1680, Cur Loss: 0.08923381, Cur Avg Loss: 0.24348881, Log Avg loss: 0.16627603, Global Avg Loss: 1.02462275, Time: 0.0209 Steps: 53330, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000115, Sample Num: 1840, Cur Loss: 0.13323449, Cur Avg Loss: 0.23975528, Log Avg loss: 0.20055321, Global Avg Loss: 1.02446825, Time: 0.0209 Steps: 53340, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000125, Sample Num: 2000, Cur Loss: 0.16441652, Cur Avg Loss: 0.23371942, Log Avg loss: 0.16430706, Global Avg Loss: 1.02430702, Time: 0.0209 Steps: 53350, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000135, Sample Num: 2160, Cur Loss: 0.10362981, Cur Avg Loss: 0.23023681, Log Avg loss: 0.18670417, Global Avg Loss: 1.02415005, Time: 0.0209 Steps: 53360, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000145, Sample Num: 2320, Cur Loss: 0.25348973, Cur Avg Loss: 0.22816420, Log Avg loss: 0.20018404, Global Avg Loss: 1.02399566, Time: 0.0209 Steps: 53370, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000155, Sample Num: 2480, Cur Loss: 0.08118440, Cur Avg Loss: 0.22884159, Log Avg loss: 0.23866363, Global Avg Loss: 1.02384854, Time: 0.0210 Steps: 53380, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000165, Sample Num: 2640, Cur Loss: 0.23455513, Cur Avg Loss: 0.23513817, Log Avg loss: 0.33273515, Global Avg Loss: 1.02371910, Time: 0.0209 Steps: 53390, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000175, Sample Num: 2800, Cur Loss: 0.17623037, Cur Avg Loss: 0.23135566, Log Avg loss: 0.16894433, Global Avg Loss: 1.02355903, Time: 0.0209 Steps: 53400, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000185, Sample Num: 2960, Cur Loss: 0.12269004, Cur Avg Loss: 0.22984050, Log Avg loss: 0.20332510, Global Avg Loss: 1.02340545, Time: 0.0209 Steps: 53410, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000195, Sample Num: 3120, Cur Loss: 0.12744227, Cur Avg Loss: 0.22941344, Log Avg loss: 0.22151293, Global Avg Loss: 1.02325534, Time: 0.0209 Steps: 53420, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000205, Sample Num: 3280, Cur Loss: 0.12705812, Cur Avg Loss: 0.22609225, Log Avg loss: 0.16132908, Global Avg Loss: 1.02309402, Time: 0.0209 Steps: 53430, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000215, Sample Num: 3440, Cur Loss: 0.10920990, Cur Avg Loss: 0.22423234, Log Avg loss: 0.18610407, Global Avg Loss: 1.02293740, Time: 0.0209 Steps: 53440, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000225, Sample Num: 3600, Cur Loss: 0.62484318, Cur Avg Loss: 0.22534873, Log Avg loss: 0.24935118, Global Avg Loss: 1.02279267, Time: 0.0209 Steps: 53450, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000235, Sample Num: 3760, Cur Loss: 0.20646256, Cur Avg Loss: 0.22644100, Log Avg loss: 0.25101696, Global Avg Loss: 1.02264830, Time: 0.0209 Steps: 53460, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000245, Sample Num: 3920, Cur Loss: 0.26877618, Cur Avg Loss: 0.22631220, Log Avg loss: 0.22328550, Global Avg Loss: 1.02249881, Time: 0.0209 Steps: 53470, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000255, Sample Num: 4080, Cur Loss: 0.27706641, Cur Avg Loss: 0.22555119, Log Avg loss: 0.20690654, Global Avg Loss: 1.02234630, Time: 0.0209 Steps: 53480, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000265, Sample Num: 4240, Cur Loss: 0.16339424, Cur Avg Loss: 0.22534685, Log Avg loss: 0.22013617, Global Avg Loss: 1.02219633, Time: 0.0208 Steps: 53490, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000275, Sample Num: 4400, Cur Loss: 0.15089998, Cur Avg Loss: 0.22262390, Log Avg loss: 0.15046558, Global Avg Loss: 1.02203339, Time: 0.0207 Steps: 53500, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000285, Sample Num: 4560, Cur Loss: 0.38765070, Cur Avg Loss: 0.22390258, Log Avg loss: 0.25906622, Global Avg Loss: 1.02189080, Time: 0.0207 Steps: 53510, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000295, Sample Num: 4720, Cur Loss: 0.10611306, Cur Avg Loss: 0.22492451, Log Avg loss: 0.25404967, Global Avg Loss: 1.02174734, Time: 0.0208 Steps: 53520, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000305, Sample Num: 4880, Cur Loss: 0.08759817, Cur Avg Loss: 0.22675259, Log Avg loss: 0.28068078, Global Avg Loss: 1.02160890, Time: 0.0207 Steps: 53530, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000315, Sample Num: 5040, Cur Loss: 0.61365163, Cur Avg Loss: 0.23065292, Log Avg loss: 0.34961313, Global Avg Loss: 1.02148338, Time: 0.0208 Steps: 53540, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000325, Sample Num: 5200, Cur Loss: 0.31836161, Cur Avg Loss: 0.22967459, Log Avg loss: 0.19885706, Global Avg Loss: 1.02132977, Time: 0.0207 Steps: 53550, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000335, Sample Num: 5360, Cur Loss: 0.29183835, Cur Avg Loss: 0.22951974, Log Avg loss: 0.22448727, Global Avg Loss: 1.02118099, Time: 0.0208 Steps: 53560, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000345, Sample Num: 5520, Cur Loss: 0.22154787, Cur Avg Loss: 0.22958151, Log Avg loss: 0.23165063, Global Avg Loss: 1.02103361, Time: 0.0207 Steps: 53570, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000355, Sample Num: 5680, Cur Loss: 0.32739070, Cur Avg Loss: 0.22753480, Log Avg loss: 0.15692343, Global Avg Loss: 1.02087233, Time: 0.0207 Steps: 53580, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000365, Sample Num: 5840, Cur Loss: 0.16407810, Cur Avg Loss: 0.22693705, Log Avg loss: 0.20571679, Global Avg Loss: 1.02072022, Time: 0.0208 Steps: 53590, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000375, Sample Num: 6000, Cur Loss: 0.24299094, Cur Avg Loss: 0.22655600, Log Avg loss: 0.21264769, Global Avg Loss: 1.02056946, Time: 0.0208 Steps: 53600, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000385, Sample Num: 6160, Cur Loss: 0.11995463, Cur Avg Loss: 0.22588631, Log Avg loss: 0.20077326, Global Avg Loss: 1.02041654, Time: 0.0208 Steps: 53610, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000395, Sample Num: 6320, Cur Loss: 0.05593276, Cur Avg Loss: 0.22467407, Log Avg loss: 0.17800265, Global Avg Loss: 1.02025944, Time: 0.0208 Steps: 53620, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000405, Sample Num: 6480, Cur Loss: 0.30475795, Cur Avg Loss: 0.22424524, Log Avg loss: 0.20730654, Global Avg Loss: 1.02010785, Time: 0.0207 Steps: 53630, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000415, Sample Num: 6640, Cur Loss: 0.76619267, Cur Avg Loss: 0.22509852, Log Avg loss: 0.25965608, Global Avg Loss: 1.01996608, Time: 0.0207 Steps: 53640, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000425, Sample Num: 6800, Cur Loss: 0.25356022, Cur Avg Loss: 0.22820639, Log Avg loss: 0.35718313, Global Avg Loss: 1.01984254, Time: 0.0207 Steps: 53650, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000435, Sample Num: 6960, Cur Loss: 0.45682019, Cur Avg Loss: 0.22987259, Log Avg loss: 0.30068615, Global Avg Loss: 1.01970852, Time: 0.0208 Steps: 53660, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000445, Sample Num: 7120, Cur Loss: 0.16867237, Cur Avg Loss: 0.23122654, Log Avg loss: 0.29012319, Global Avg Loss: 1.01957258, Time: 0.0208 Steps: 53670, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000455, Sample Num: 7280, Cur Loss: 0.18334189, Cur Avg Loss: 0.23143032, Log Avg loss: 0.24049880, Global Avg Loss: 1.01942745, Time: 0.0208 Steps: 53680, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000465, Sample Num: 7440, Cur Loss: 0.11486153, Cur Avg Loss: 0.23052492, Log Avg loss: 0.18932902, Global Avg Loss: 1.01927284, Time: 0.0207 Steps: 53690, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000475, Sample Num: 7600, Cur Loss: 0.24355695, Cur Avg Loss: 0.22992504, Log Avg loss: 0.20203055, Global Avg Loss: 1.01912065, Time: 0.0208 Steps: 53700, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000485, Sample Num: 7760, Cur Loss: 0.11536162, Cur Avg Loss: 0.22891529, Log Avg loss: 0.18095241, Global Avg Loss: 1.01896460, Time: 0.0208 Steps: 53710, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000495, Sample Num: 7920, Cur Loss: 0.28805688, Cur Avg Loss: 0.22849984, Log Avg loss: 0.20835045, Global Avg Loss: 1.01881370, Time: 0.0208 Steps: 53720, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000505, Sample Num: 8080, Cur Loss: 0.22778782, Cur Avg Loss: 0.22731727, Log Avg loss: 0.16878023, Global Avg Loss: 1.01865550, Time: 0.0207 Steps: 53730, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000515, Sample Num: 8240, Cur Loss: 0.16982391, Cur Avg Loss: 0.22660935, Log Avg loss: 0.19085929, Global Avg Loss: 1.01850146, Time: 0.0244 Steps: 53740, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000525, Sample Num: 8400, Cur Loss: 0.53726476, Cur Avg Loss: 0.22714298, Log Avg loss: 0.25462477, Global Avg Loss: 1.01835934, Time: 0.0207 Steps: 53750, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000535, Sample Num: 8560, Cur Loss: 0.41748589, Cur Avg Loss: 0.22795313, Log Avg loss: 0.27048635, Global Avg Loss: 1.01822023, Time: 0.0208 Steps: 53760, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000545, Sample Num: 8720, Cur Loss: 0.07464181, Cur Avg Loss: 0.22733552, Log Avg loss: 0.19429327, Global Avg Loss: 1.01806700, Time: 0.0207 Steps: 53770, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000555, Sample Num: 8880, Cur Loss: 0.47726193, Cur Avg Loss: 0.22788278, Log Avg loss: 0.25770841, Global Avg Loss: 1.01792562, Time: 0.0207 Steps: 53780, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000565, Sample Num: 9040, Cur Loss: 0.15024358, Cur Avg Loss: 0.22704696, Log Avg loss: 0.18065909, Global Avg Loss: 1.01776996, Time: 0.0207 Steps: 53790, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000575, Sample Num: 9200, Cur Loss: 0.53204334, Cur Avg Loss: 0.22736471, Log Avg loss: 0.24531749, Global Avg Loss: 1.01762638, Time: 0.0207 Steps: 53800, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000585, Sample Num: 9360, Cur Loss: 0.10202082, Cur Avg Loss: 0.22618183, Log Avg loss: 0.15816619, Global Avg Loss: 1.01746666, Time: 0.0208 Steps: 53810, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000595, Sample Num: 9520, Cur Loss: 0.36749458, Cur Avg Loss: 0.22723358, Log Avg loss: 0.28876057, Global Avg Loss: 1.01733126, Time: 0.0207 Steps: 53820, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000605, Sample Num: 9680, Cur Loss: 0.15556848, Cur Avg Loss: 0.22712549, Log Avg loss: 0.22069458, Global Avg Loss: 1.01718327, Time: 0.0207 Steps: 53830, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000615, Sample Num: 9840, Cur Loss: 0.17740026, Cur Avg Loss: 0.22906532, Log Avg loss: 0.34642466, Global Avg Loss: 1.01705869, Time: 0.0207 Steps: 53840, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000625, Sample Num: 10000, Cur Loss: 0.33045334, Cur Avg Loss: 0.22955749, Log Avg loss: 0.25982630, Global Avg Loss: 1.01691807, Time: 0.0208 Steps: 53850, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000635, Sample Num: 10160, Cur Loss: 0.35913199, Cur Avg Loss: 0.22924513, Log Avg loss: 0.20972249, Global Avg Loss: 1.01676820, Time: 0.0207 Steps: 53860, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000645, Sample Num: 10320, Cur Loss: 0.08026038, Cur Avg Loss: 0.22918051, Log Avg loss: 0.22507743, Global Avg Loss: 1.01662124, Time: 0.0208 Steps: 53870, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000655, Sample Num: 10480, Cur Loss: 0.01710289, Cur Avg Loss: 0.22889474, Log Avg loss: 0.21046259, Global Avg Loss: 1.01647162, Time: 0.0208 Steps: 53880, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000665, Sample Num: 10640, Cur Loss: 0.25429565, Cur Avg Loss: 0.22952869, Log Avg loss: 0.27105200, Global Avg Loss: 1.01633330, Time: 0.0207 Steps: 53890, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000675, Sample Num: 10800, Cur Loss: 0.39237404, Cur Avg Loss: 0.22919117, Log Avg loss: 0.20674648, Global Avg Loss: 1.01618309, Time: 0.0208 Steps: 53900, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000685, Sample Num: 10960, Cur Loss: 0.17195131, Cur Avg Loss: 0.22843916, Log Avg loss: 0.17767802, Global Avg Loss: 1.01602756, Time: 0.0207 Steps: 53910, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000695, Sample Num: 11120, Cur Loss: 0.45192063, Cur Avg Loss: 0.22975093, Log Avg loss: 0.31960758, Global Avg Loss: 1.01589840, Time: 0.0207 Steps: 53920, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000705, Sample Num: 11280, Cur Loss: 0.21648407, Cur Avg Loss: 0.22989339, Log Avg loss: 0.23979411, Global Avg Loss: 1.01575449, Time: 0.0207 Steps: 53930, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000715, Sample Num: 11440, Cur Loss: 0.09951223, Cur Avg Loss: 0.22948586, Log Avg loss: 0.20075495, Global Avg Loss: 1.01560339, Time: 0.0207 Steps: 53940, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000725, Sample Num: 11600, Cur Loss: 0.14709023, Cur Avg Loss: 0.22876704, Log Avg loss: 0.17737180, Global Avg Loss: 1.01544802, Time: 0.0207 Steps: 53950, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000735, Sample Num: 11760, Cur Loss: 0.25500995, Cur Avg Loss: 0.22879616, Log Avg loss: 0.23090685, Global Avg Loss: 1.01530263, Time: 0.0207 Steps: 53960, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000745, Sample Num: 11920, Cur Loss: 0.22385091, Cur Avg Loss: 0.22806877, Log Avg loss: 0.17460544, Global Avg Loss: 1.01514686, Time: 0.0207 Steps: 53970, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000755, Sample Num: 12080, Cur Loss: 0.42772460, Cur Avg Loss: 0.22878771, Log Avg loss: 0.28234935, Global Avg Loss: 1.01501110, Time: 0.0207 Steps: 53980, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000765, Sample Num: 12240, Cur Loss: 0.10773692, Cur Avg Loss: 0.22851363, Log Avg loss: 0.20782070, Global Avg Loss: 1.01486160, Time: 0.0207 Steps: 53990, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000775, Sample Num: 12400, Cur Loss: 0.10413840, Cur Avg Loss: 0.22771679, Log Avg loss: 0.16675777, Global Avg Loss: 1.01470454, Time: 0.0207 Steps: 54000, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000785, Sample Num: 12560, Cur Loss: 0.13930750, Cur Avg Loss: 0.22756452, Log Avg loss: 0.21576395, Global Avg Loss: 1.01455662, Time: 0.0207 Steps: 54010, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000795, Sample Num: 12720, Cur Loss: 0.05862541, Cur Avg Loss: 0.22714169, Log Avg loss: 0.19394981, Global Avg Loss: 1.01440471, Time: 0.0208 Steps: 54020, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000805, Sample Num: 12880, Cur Loss: 0.08754084, Cur Avg Loss: 0.22627646, Log Avg loss: 0.15749054, Global Avg Loss: 1.01424611, Time: 0.0209 Steps: 54030, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000815, Sample Num: 13040, Cur Loss: 0.14339368, Cur Avg Loss: 0.22737469, Log Avg loss: 0.31578217, Global Avg Loss: 1.01411686, Time: 0.0207 Steps: 54040, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000825, Sample Num: 13200, Cur Loss: 0.06079822, Cur Avg Loss: 0.22782942, Log Avg loss: 0.26488964, Global Avg Loss: 1.01397824, Time: 0.0208 Steps: 54050, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000835, Sample Num: 13360, Cur Loss: 0.08744694, Cur Avg Loss: 0.22753743, Log Avg loss: 0.20344874, Global Avg Loss: 1.01382831, Time: 0.0208 Steps: 54060, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000845, Sample Num: 13520, Cur Loss: 0.15542817, Cur Avg Loss: 0.22745428, Log Avg loss: 0.22051117, Global Avg Loss: 1.01368159, Time: 0.0208 Steps: 54070, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000855, Sample Num: 13680, Cur Loss: 0.22168326, Cur Avg Loss: 0.22774577, Log Avg loss: 0.25237636, Global Avg Loss: 1.01354082, Time: 0.0208 Steps: 54080, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000865, Sample Num: 13840, Cur Loss: 0.08365658, Cur Avg Loss: 0.22903620, Log Avg loss: 0.33936790, Global Avg Loss: 1.01341618, Time: 0.0208 Steps: 54090, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000875, Sample Num: 14000, Cur Loss: 0.15224168, Cur Avg Loss: 0.22957189, Log Avg loss: 0.27590968, Global Avg Loss: 1.01327985, Time: 0.0208 Steps: 54100, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000885, Sample Num: 14160, Cur Loss: 0.24231204, Cur Avg Loss: 0.22912961, Log Avg loss: 0.19043005, Global Avg Loss: 1.01312778, Time: 0.0208 Steps: 54110, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000895, Sample Num: 14320, Cur Loss: 0.45319405, Cur Avg Loss: 0.22858711, Log Avg loss: 0.18057530, Global Avg Loss: 1.01297395, Time: 0.0208 Steps: 54120, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000905, Sample Num: 14480, Cur Loss: 0.18601468, Cur Avg Loss: 0.22883646, Log Avg loss: 0.25115312, Global Avg Loss: 1.01283321, Time: 0.0209 Steps: 54130, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000915, Sample Num: 14640, Cur Loss: 0.04715629, Cur Avg Loss: 0.22873376, Log Avg loss: 0.21944011, Global Avg Loss: 1.01268667, Time: 0.0208 Steps: 54140, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000925, Sample Num: 14800, Cur Loss: 0.21331003, Cur Avg Loss: 0.22959458, Log Avg loss: 0.30835902, Global Avg Loss: 1.01255660, Time: 0.0208 Steps: 54150, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000935, Sample Num: 14960, Cur Loss: 0.09615209, Cur Avg Loss: 0.22964024, Log Avg loss: 0.23386370, Global Avg Loss: 1.01241282, Time: 0.0208 Steps: 54160, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000945, Sample Num: 15120, Cur Loss: 0.23398095, Cur Avg Loss: 0.22972080, Log Avg loss: 0.23725310, Global Avg Loss: 1.01226972, Time: 0.0208 Steps: 54170, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000955, Sample Num: 15280, Cur Loss: 0.18556036, Cur Avg Loss: 0.22936747, Log Avg loss: 0.19597860, Global Avg Loss: 1.01211906, Time: 0.0208 Steps: 54180, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000965, Sample Num: 15440, Cur Loss: 0.34519362, Cur Avg Loss: 0.22883688, Log Avg loss: 0.17816468, Global Avg Loss: 1.01196516, Time: 0.0207 Steps: 54190, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000975, Sample Num: 15600, Cur Loss: 0.09433068, Cur Avg Loss: 0.22858448, Log Avg loss: 0.20422882, Global Avg Loss: 1.01181614, Time: 0.0208 Steps: 54200, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000985, Sample Num: 15760, Cur Loss: 0.40042222, Cur Avg Loss: 0.22873681, Log Avg loss: 0.24358876, Global Avg Loss: 1.01167442, Time: 0.0207 Steps: 54210, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000995, Sample Num: 15920, Cur Loss: 0.30114716, Cur Avg Loss: 0.22882907, Log Avg loss: 0.23791631, Global Avg Loss: 1.01153172, Time: 0.0208 Steps: 54220, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001005, Sample Num: 16080, Cur Loss: 0.16158459, Cur Avg Loss: 0.22926149, Log Avg loss: 0.27228698, Global Avg Loss: 1.01139540, Time: 0.0207 Steps: 54230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001015, Sample Num: 16240, Cur Loss: 0.25480360, Cur Avg Loss: 0.22922930, Log Avg loss: 0.22599480, Global Avg Loss: 1.01125060, Time: 0.0208 Steps: 54240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001025, Sample Num: 16400, Cur Loss: 0.30021149, Cur Avg Loss: 0.22954287, Log Avg loss: 0.26137017, Global Avg Loss: 1.01111237, Time: 0.0256 Steps: 54250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001035, Sample Num: 16560, Cur Loss: 0.35496235, Cur Avg Loss: 0.22999893, Log Avg loss: 0.27674445, Global Avg Loss: 1.01097703, Time: 0.0219 Steps: 54260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001045, Sample Num: 16720, Cur Loss: 0.46112406, Cur Avg Loss: 0.22987909, Log Avg loss: 0.21747662, Global Avg Loss: 1.01083082, Time: 0.0219 Steps: 54270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001055, Sample Num: 16880, Cur Loss: 0.26425061, Cur Avg Loss: 0.23097538, Log Avg loss: 0.34553749, Global Avg Loss: 1.01070825, Time: 0.0219 Steps: 54280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001065, Sample Num: 17040, Cur Loss: 0.21812326, Cur Avg Loss: 0.23074197, Log Avg loss: 0.20611671, Global Avg Loss: 1.01056005, Time: 0.0219 Steps: 54290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001075, Sample Num: 17200, Cur Loss: 0.04979789, Cur Avg Loss: 0.23010330, Log Avg loss: 0.16208545, Global Avg Loss: 1.01040379, Time: 0.0219 Steps: 54300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001085, Sample Num: 17360, Cur Loss: 0.12558205, Cur Avg Loss: 0.22994033, Log Avg loss: 0.21242082, Global Avg Loss: 1.01025686, Time: 0.0219 Steps: 54310, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001095, Sample Num: 17520, Cur Loss: 0.08047895, Cur Avg Loss: 0.22984607, Log Avg loss: 0.21961850, Global Avg Loss: 1.01011131, Time: 0.0219 Steps: 54320, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001105, Sample Num: 17680, Cur Loss: 0.57809556, Cur Avg Loss: 0.23029559, Log Avg loss: 0.27951826, Global Avg Loss: 1.00997683, Time: 0.0219 Steps: 54330, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001115, Sample Num: 17840, Cur Loss: 0.27685633, Cur Avg Loss: 0.23043480, Log Avg loss: 0.24581771, Global Avg Loss: 1.00983621, Time: 0.0219 Steps: 54340, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001125, Sample Num: 18000, Cur Loss: 0.29541868, Cur Avg Loss: 0.23051733, Log Avg loss: 0.23971888, Global Avg Loss: 1.00969451, Time: 0.0219 Steps: 54350, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001135, Sample Num: 18160, Cur Loss: 0.13952141, Cur Avg Loss: 0.23028844, Log Avg loss: 0.20453875, Global Avg Loss: 1.00954640, Time: 0.0219 Steps: 54360, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001145, Sample Num: 18320, Cur Loss: 0.09927802, Cur Avg Loss: 0.23055107, Log Avg loss: 0.26035967, Global Avg Loss: 1.00940860, Time: 0.0218 Steps: 54370, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001155, Sample Num: 18480, Cur Loss: 0.35279691, Cur Avg Loss: 0.23080406, Log Avg loss: 0.25977131, Global Avg Loss: 1.00927075, Time: 0.0219 Steps: 54380, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001165, Sample Num: 18640, Cur Loss: 0.11559293, Cur Avg Loss: 0.23107333, Log Avg loss: 0.26217415, Global Avg Loss: 1.00913339, Time: 0.0219 Steps: 54390, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001175, Sample Num: 18800, Cur Loss: 0.43647084, Cur Avg Loss: 0.23163231, Log Avg loss: 0.29675318, Global Avg Loss: 1.00900244, Time: 0.0219 Steps: 54400, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001185, Sample Num: 18960, Cur Loss: 0.05531814, Cur Avg Loss: 0.23145543, Log Avg loss: 0.21067226, Global Avg Loss: 1.00885571, Time: 0.0219 Steps: 54410, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001195, Sample Num: 19120, Cur Loss: 0.16034792, Cur Avg Loss: 0.23124656, Log Avg loss: 0.20649574, Global Avg Loss: 1.00870828, Time: 0.0219 Steps: 54420, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001205, Sample Num: 19280, Cur Loss: 0.08275314, Cur Avg Loss: 0.23105281, Log Avg loss: 0.20789986, Global Avg Loss: 1.00856115, Time: 0.0219 Steps: 54430, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001215, Sample Num: 19440, Cur Loss: 0.20271823, Cur Avg Loss: 0.23104522, Log Avg loss: 0.23012960, Global Avg Loss: 1.00841816, Time: 0.0219 Steps: 54440, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001225, Sample Num: 19600, Cur Loss: 0.22116704, Cur Avg Loss: 0.23101698, Log Avg loss: 0.22758633, Global Avg Loss: 1.00827476, Time: 0.0219 Steps: 54450, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001235, Sample Num: 19760, Cur Loss: 0.18759608, Cur Avg Loss: 0.23107110, Log Avg loss: 0.23770084, Global Avg Loss: 1.00813326, Time: 0.0219 Steps: 54460, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001245, Sample Num: 19920, Cur Loss: 0.65354317, Cur Avg Loss: 0.23137792, Log Avg loss: 0.26927008, Global Avg Loss: 1.00799762, Time: 0.0219 Steps: 54470, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001255, Sample Num: 20080, Cur Loss: 0.20824921, Cur Avg Loss: 0.23120877, Log Avg loss: 0.21014972, Global Avg Loss: 1.00785117, Time: 0.0219 Steps: 54480, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001265, Sample Num: 20240, Cur Loss: 0.18189947, Cur Avg Loss: 0.23130613, Log Avg loss: 0.24352473, Global Avg Loss: 1.00771090, Time: 0.0219 Steps: 54490, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001275, Sample Num: 20400, Cur Loss: 0.11701085, Cur Avg Loss: 0.23061125, Log Avg loss: 0.14270847, Global Avg Loss: 1.00755218, Time: 0.0219 Steps: 54500, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001285, Sample Num: 20560, Cur Loss: 0.14067031, Cur Avg Loss: 0.23115556, Log Avg loss: 0.30055590, Global Avg Loss: 1.00742248, Time: 0.0208 Steps: 54510, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001295, Sample Num: 20720, Cur Loss: 0.12634043, Cur Avg Loss: 0.23131263, Log Avg loss: 0.25149572, Global Avg Loss: 1.00728383, Time: 0.0208 Steps: 54520, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001305, Sample Num: 20880, Cur Loss: 0.03579369, Cur Avg Loss: 0.23107354, Log Avg loss: 0.20011168, Global Avg Loss: 1.00713581, Time: 0.0208 Steps: 54530, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001315, Sample Num: 21040, Cur Loss: 0.20857888, Cur Avg Loss: 0.23085931, Log Avg loss: 0.20290187, Global Avg Loss: 1.00698835, Time: 0.0208 Steps: 54540, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001325, Sample Num: 21200, Cur Loss: 0.30840969, Cur Avg Loss: 0.23164017, Log Avg loss: 0.33432393, Global Avg Loss: 1.00686504, Time: 0.0208 Steps: 54550, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001335, Sample Num: 21360, Cur Loss: 0.11234819, Cur Avg Loss: 0.23154953, Log Avg loss: 0.21953935, Global Avg Loss: 1.00672074, Time: 0.0208 Steps: 54560, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001345, Sample Num: 21520, Cur Loss: 0.18082123, Cur Avg Loss: 0.23112491, Log Avg loss: 0.17443819, Global Avg Loss: 1.00656822, Time: 0.0207 Steps: 54570, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001355, Sample Num: 21680, Cur Loss: 0.18610284, Cur Avg Loss: 0.23103753, Log Avg loss: 0.21928503, Global Avg Loss: 1.00642397, Time: 0.0208 Steps: 54580, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001365, Sample Num: 21840, Cur Loss: 0.31614256, Cur Avg Loss: 0.23121653, Log Avg loss: 0.25547099, Global Avg Loss: 1.00628641, Time: 0.0208 Steps: 54590, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001375, Sample Num: 22000, Cur Loss: 0.10194353, Cur Avg Loss: 0.23128791, Log Avg loss: 0.24103165, Global Avg Loss: 1.00614626, Time: 0.0208 Steps: 54600, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001385, Sample Num: 22160, Cur Loss: 0.47803599, Cur Avg Loss: 0.23155783, Log Avg loss: 0.26867130, Global Avg Loss: 1.00601121, Time: 0.0208 Steps: 54610, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001395, Sample Num: 22320, Cur Loss: 0.16659218, Cur Avg Loss: 0.23141401, Log Avg loss: 0.21149480, Global Avg Loss: 1.00586575, Time: 0.0208 Steps: 54620, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001405, Sample Num: 22480, Cur Loss: 0.20089976, Cur Avg Loss: 0.23105611, Log Avg loss: 0.18112954, Global Avg Loss: 1.00571478, Time: 0.0208 Steps: 54630, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001415, Sample Num: 22640, Cur Loss: 0.20481066, Cur Avg Loss: 0.23081258, Log Avg loss: 0.19659626, Global Avg Loss: 1.00556670, Time: 0.0207 Steps: 54640, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001425, Sample Num: 22800, Cur Loss: 0.23759234, Cur Avg Loss: 0.23082996, Log Avg loss: 0.23328917, Global Avg Loss: 1.00542539, Time: 0.0208 Steps: 54650, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001435, Sample Num: 22960, Cur Loss: 0.09080110, Cur Avg Loss: 0.23085488, Log Avg loss: 0.23440665, Global Avg Loss: 1.00528433, Time: 0.0207 Steps: 54660, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001445, Sample Num: 23120, Cur Loss: 0.19401398, Cur Avg Loss: 0.23041010, Log Avg loss: 0.16658284, Global Avg Loss: 1.00513092, Time: 0.0207 Steps: 54670, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001455, Sample Num: 23280, Cur Loss: 0.20167649, Cur Avg Loss: 0.23072474, Log Avg loss: 0.27619084, Global Avg Loss: 1.00499761, Time: 0.0207 Steps: 54680, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001465, Sample Num: 23440, Cur Loss: 0.23764488, Cur Avg Loss: 0.23040223, Log Avg loss: 0.18347754, Global Avg Loss: 1.00484739, Time: 0.0207 Steps: 54690, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001475, Sample Num: 23600, Cur Loss: 0.10121783, Cur Avg Loss: 0.23036747, Log Avg loss: 0.22527434, Global Avg Loss: 1.00470488, Time: 0.0207 Steps: 54700, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001485, Sample Num: 23760, Cur Loss: 0.14822006, Cur Avg Loss: 0.23062790, Log Avg loss: 0.26904155, Global Avg Loss: 1.00457041, Time: 0.0207 Steps: 54710, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001495, Sample Num: 23920, Cur Loss: 0.03167189, Cur Avg Loss: 0.23140050, Log Avg loss: 0.34613227, Global Avg Loss: 1.00445008, Time: 0.0207 Steps: 54720, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001505, Sample Num: 24080, Cur Loss: 0.31563252, Cur Avg Loss: 0.23154451, Log Avg loss: 0.25307297, Global Avg Loss: 1.00431279, Time: 0.0207 Steps: 54730, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001515, Sample Num: 24240, Cur Loss: 0.44817507, Cur Avg Loss: 0.23188390, Log Avg loss: 0.28296285, Global Avg Loss: 1.00418102, Time: 0.0207 Steps: 54740, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001525, Sample Num: 24400, Cur Loss: 0.11896212, Cur Avg Loss: 0.23189624, Log Avg loss: 0.23376599, Global Avg Loss: 1.00404030, Time: 0.0208 Steps: 54750, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001535, Sample Num: 24560, Cur Loss: 0.03954379, Cur Avg Loss: 0.23204681, Log Avg loss: 0.25500818, Global Avg Loss: 1.00390352, Time: 0.0208 Steps: 54760, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001545, Sample Num: 24720, Cur Loss: 0.28334522, Cur Avg Loss: 0.23214203, Log Avg loss: 0.24675802, Global Avg Loss: 1.00376528, Time: 0.0208 Steps: 54770, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001555, Sample Num: 24880, Cur Loss: 0.12176353, Cur Avg Loss: 0.23197839, Log Avg loss: 0.20669622, Global Avg Loss: 1.00361977, Time: 0.0208 Steps: 54780, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001565, Sample Num: 25040, Cur Loss: 0.17528677, Cur Avg Loss: 0.23153969, Log Avg loss: 0.16332235, Global Avg Loss: 1.00346640, Time: 0.0208 Steps: 54790, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001575, Sample Num: 25200, Cur Loss: 0.14643718, Cur Avg Loss: 0.23115155, Log Avg loss: 0.17040739, Global Avg Loss: 1.00331439, Time: 0.0208 Steps: 54800, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001585, Sample Num: 25360, Cur Loss: 0.16559283, Cur Avg Loss: 0.23086637, Log Avg loss: 0.18595036, Global Avg Loss: 1.00316526, Time: 0.0208 Steps: 54810, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001595, Sample Num: 25520, Cur Loss: 0.32724410, Cur Avg Loss: 0.23119655, Log Avg loss: 0.28352966, Global Avg Loss: 1.00303399, Time: 0.0209 Steps: 54820, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001605, Sample Num: 25680, Cur Loss: 0.13486078, Cur Avg Loss: 0.23088623, Log Avg loss: 0.18139025, Global Avg Loss: 1.00288413, Time: 0.0208 Steps: 54830, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001615, Sample Num: 25840, Cur Loss: 0.18292281, Cur Avg Loss: 0.23092277, Log Avg loss: 0.23678779, Global Avg Loss: 1.00274444, Time: 0.0208 Steps: 54840, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001625, Sample Num: 26000, Cur Loss: 0.49838504, Cur Avg Loss: 0.23111461, Log Avg loss: 0.26209716, Global Avg Loss: 1.00260941, Time: 0.0209 Steps: 54850, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001635, Sample Num: 26160, Cur Loss: 0.45941591, Cur Avg Loss: 0.23105241, Log Avg loss: 0.22094394, Global Avg Loss: 1.00246692, Time: 0.0208 Steps: 54860, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001645, Sample Num: 26320, Cur Loss: 0.15904450, Cur Avg Loss: 0.23096050, Log Avg loss: 0.21593327, Global Avg Loss: 1.00232358, Time: 0.0208 Steps: 54870, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001655, Sample Num: 26480, Cur Loss: 0.24655744, Cur Avg Loss: 0.23107841, Log Avg loss: 0.25047495, Global Avg Loss: 1.00218658, Time: 0.0208 Steps: 54880, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001665, Sample Num: 26640, Cur Loss: 0.31893024, Cur Avg Loss: 0.23085260, Log Avg loss: 0.19348128, Global Avg Loss: 1.00203925, Time: 0.0208 Steps: 54890, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001675, Sample Num: 26800, Cur Loss: 0.15063809, Cur Avg Loss: 0.23086096, Log Avg loss: 0.23225289, Global Avg Loss: 1.00189903, Time: 0.0208 Steps: 54900, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001685, Sample Num: 26960, Cur Loss: 0.21970555, Cur Avg Loss: 0.23079482, Log Avg loss: 0.21971667, Global Avg Loss: 1.00175658, Time: 0.0208 Steps: 54910, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001695, Sample Num: 27120, Cur Loss: 0.56456101, Cur Avg Loss: 0.23109046, Log Avg loss: 0.28090597, Global Avg Loss: 1.00162533, Time: 0.0208 Steps: 54920, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001705, Sample Num: 27280, Cur Loss: 0.24135213, Cur Avg Loss: 0.23087910, Log Avg loss: 0.19505246, Global Avg Loss: 1.00147849, Time: 0.0208 Steps: 54930, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001715, Sample Num: 27440, Cur Loss: 0.23445329, Cur Avg Loss: 0.23077196, Log Avg loss: 0.21250476, Global Avg Loss: 1.00133489, Time: 0.0208 Steps: 54940, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001725, Sample Num: 27600, Cur Loss: 0.07082698, Cur Avg Loss: 0.23034972, Log Avg loss: 0.15793675, Global Avg Loss: 1.00118140, Time: 0.0208 Steps: 54950, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001735, Sample Num: 27760, Cur Loss: 0.33787984, Cur Avg Loss: 0.23042534, Log Avg loss: 0.24346982, Global Avg Loss: 1.00104353, Time: 0.0208 Steps: 54960, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001745, Sample Num: 27920, Cur Loss: 0.26089388, Cur Avg Loss: 0.23044081, Log Avg loss: 0.23312431, Global Avg Loss: 1.00090384, Time: 0.0208 Steps: 54970, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001755, Sample Num: 28080, Cur Loss: 0.49122706, Cur Avg Loss: 0.23020032, Log Avg loss: 0.18823528, Global Avg Loss: 1.00075603, Time: 0.0210 Steps: 54980, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001765, Sample Num: 28240, Cur Loss: 0.12789349, Cur Avg Loss: 0.23036157, Log Avg loss: 0.25866055, Global Avg Loss: 1.00062107, Time: 0.0208 Steps: 54990, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001775, Sample Num: 28400, Cur Loss: 0.30779445, Cur Avg Loss: 0.23039711, Log Avg loss: 0.23666948, Global Avg Loss: 1.00048217, Time: 0.0208 Steps: 55000, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001785, Sample Num: 28560, Cur Loss: 0.12173698, Cur Avg Loss: 0.23027474, Log Avg loss: 0.20855382, Global Avg Loss: 1.00033821, Time: 0.0208 Steps: 55010, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001795, Sample Num: 28720, Cur Loss: 0.17015909, Cur Avg Loss: 0.22994316, Log Avg loss: 0.17075598, Global Avg Loss: 1.00018743, Time: 0.0245 Steps: 55020, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001805, Sample Num: 28880, Cur Loss: 0.09166834, Cur Avg Loss: 0.22951640, Log Avg loss: 0.15291326, Global Avg Loss: 1.00003347, Time: 0.0208 Steps: 55030, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001815, Sample Num: 29040, Cur Loss: 0.19870940, Cur Avg Loss: 0.22969663, Log Avg loss: 0.26222760, Global Avg Loss: 0.99989942, Time: 0.0209 Steps: 55040, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001825, Sample Num: 29200, Cur Loss: 0.58884180, Cur Avg Loss: 0.23058910, Log Avg loss: 0.39257239, Global Avg Loss: 0.99978910, Time: 0.0208 Steps: 55050, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001835, Sample Num: 29360, Cur Loss: 0.09884907, Cur Avg Loss: 0.23042410, Log Avg loss: 0.20031156, Global Avg Loss: 0.99964390, Time: 0.0208 Steps: 55060, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001845, Sample Num: 29520, Cur Loss: 0.15211511, Cur Avg Loss: 0.23017402, Log Avg loss: 0.18428437, Global Avg Loss: 0.99949584, Time: 0.0209 Steps: 55070, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001855, Sample Num: 29680, Cur Loss: 0.17253435, Cur Avg Loss: 0.22996012, Log Avg loss: 0.19049698, Global Avg Loss: 0.99934896, Time: 0.0209 Steps: 55080, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001865, Sample Num: 29840, Cur Loss: 0.20602116, Cur Avg Loss: 0.22998023, Log Avg loss: 0.23371093, Global Avg Loss: 0.99920998, Time: 0.0208 Steps: 55090, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001875, Sample Num: 30000, Cur Loss: 0.40280697, Cur Avg Loss: 0.23035090, Log Avg loss: 0.29948031, Global Avg Loss: 0.99908299, Time: 0.0208 Steps: 55100, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001885, Sample Num: 30160, Cur Loss: 0.26525557, Cur Avg Loss: 0.23063295, Log Avg loss: 0.28351694, Global Avg Loss: 0.99895314, Time: 0.0208 Steps: 55110, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001895, Sample Num: 30320, Cur Loss: 0.31176180, Cur Avg Loss: 0.23066180, Log Avg loss: 0.23609970, Global Avg Loss: 0.99881475, Time: 0.0208 Steps: 55120, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001905, Sample Num: 30480, Cur Loss: 0.06562912, Cur Avg Loss: 0.23095260, Log Avg loss: 0.28605896, Global Avg Loss: 0.99868546, Time: 0.0208 Steps: 55130, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001915, Sample Num: 30640, Cur Loss: 0.17678948, Cur Avg Loss: 0.23104293, Log Avg loss: 0.24825124, Global Avg Loss: 0.99854936, Time: 0.0208 Steps: 55140, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001925, Sample Num: 30800, Cur Loss: 0.26641220, Cur Avg Loss: 0.23102132, Log Avg loss: 0.22688267, Global Avg Loss: 0.99840944, Time: 0.0208 Steps: 55150, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001935, Sample Num: 30960, Cur Loss: 0.26710412, Cur Avg Loss: 0.23131831, Log Avg loss: 0.28848905, Global Avg Loss: 0.99828074, Time: 0.0209 Steps: 55160, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001945, Sample Num: 31120, Cur Loss: 0.16234961, Cur Avg Loss: 0.23138028, Log Avg loss: 0.24337088, Global Avg Loss: 0.99814391, Time: 0.0209 Steps: 55170, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001955, Sample Num: 31280, Cur Loss: 0.14041251, Cur Avg Loss: 0.23179236, Log Avg loss: 0.31194344, Global Avg Loss: 0.99801955, Time: 0.0208 Steps: 55180, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001965, Sample Num: 31440, Cur Loss: 0.15504956, Cur Avg Loss: 0.23193054, Log Avg loss: 0.25894498, Global Avg Loss: 0.99788564, Time: 0.0209 Steps: 55190, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001975, Sample Num: 31600, Cur Loss: 0.37892824, Cur Avg Loss: 0.23187939, Log Avg loss: 0.22182752, Global Avg Loss: 0.99774505, Time: 0.0208 Steps: 55200, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001985, Sample Num: 31760, Cur Loss: 0.10207145, Cur Avg Loss: 0.23183799, Log Avg loss: 0.22366169, Global Avg Loss: 0.99760484, Time: 0.0209 Steps: 55210, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001995, Sample Num: 31920, Cur Loss: 0.14984493, Cur Avg Loss: 0.23165248, Log Avg loss: 0.19482826, Global Avg Loss: 0.99745946, Time: 0.0208 Steps: 55220, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002005, Sample Num: 32080, Cur Loss: 0.25698936, Cur Avg Loss: 0.23177257, Log Avg loss: 0.25573002, Global Avg Loss: 0.99732516, Time: 0.0210 Steps: 55230, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002015, Sample Num: 32240, Cur Loss: 0.20593718, Cur Avg Loss: 0.23200891, Log Avg loss: 0.27939647, Global Avg Loss: 0.99719520, Time: 0.0208 Steps: 55240, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002025, Sample Num: 32400, Cur Loss: 0.18459262, Cur Avg Loss: 0.23206738, Log Avg loss: 0.24384818, Global Avg Loss: 0.99705884, Time: 0.0208 Steps: 55250, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002035, Sample Num: 32560, Cur Loss: 0.18830918, Cur Avg Loss: 0.23196110, Log Avg loss: 0.21044012, Global Avg Loss: 0.99691650, Time: 0.0208 Steps: 55260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002045, Sample Num: 32720, Cur Loss: 0.11698666, Cur Avg Loss: 0.23229101, Log Avg loss: 0.29942732, Global Avg Loss: 0.99679030, Time: 0.0208 Steps: 55270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002055, Sample Num: 32880, Cur Loss: 0.18289641, Cur Avg Loss: 0.23268036, Log Avg loss: 0.31230279, Global Avg Loss: 0.99666648, Time: 0.0209 Steps: 55280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002065, Sample Num: 33040, Cur Loss: 0.17553906, Cur Avg Loss: 0.23237685, Log Avg loss: 0.17000544, Global Avg Loss: 0.99651696, Time: 0.0209 Steps: 55290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002075, Sample Num: 33200, Cur Loss: 0.13026252, Cur Avg Loss: 0.23204431, Log Avg loss: 0.16337500, Global Avg Loss: 0.99636630, Time: 0.0209 Steps: 55300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002085, Sample Num: 33360, Cur Loss: 0.11026330, Cur Avg Loss: 0.23191866, Log Avg loss: 0.20584523, Global Avg Loss: 0.99622338, Time: 0.0209 Steps: 55310, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002095, Sample Num: 33520, Cur Loss: 0.10675108, Cur Avg Loss: 0.23163319, Log Avg loss: 0.17211238, Global Avg Loss: 0.99607441, Time: 0.0209 Steps: 55320, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002105, Sample Num: 33680, Cur Loss: 0.08882033, Cur Avg Loss: 0.23144604, Log Avg loss: 0.19224001, Global Avg Loss: 0.99592913, Time: 0.0208 Steps: 55330, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002115, Sample Num: 33840, Cur Loss: 0.14888996, Cur Avg Loss: 0.23163360, Log Avg loss: 0.27111477, Global Avg Loss: 0.99579815, Time: 0.0208 Steps: 55340, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002125, Sample Num: 34000, Cur Loss: 0.44372287, Cur Avg Loss: 0.23162038, Log Avg loss: 0.22882383, Global Avg Loss: 0.99565958, Time: 0.0208 Steps: 55350, Updated lr: 0.000048 ***** Running evaluation checkpoint-55354 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-55354 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.605114, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.287145, "eval_total_loss": 201.863056, "eval_mae": 0.374869, "eval_mse": 0.287247, "eval_r2": 0.817407, "eval_sp_statistic": 0.866655, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.905258, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.867643, "test_total_loss": 435.556706, "test_mae": 0.726609, "test_mse": 0.867734, "test_r2": 0.439957, "test_sp_statistic": 0.803349, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.848197, "test_ps_pvalue": 0.0, "lr": 4.845519203413941e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9955962621370498, "train_cur_epoch_loss": 492.67079731822014, "train_cur_epoch_avg_loss": 0.23140948676290285, "train_cur_epoch_time": 44.60511350631714, "train_cur_epoch_avg_time": 0.020951204089392737, "epoch": 26, "step": 55354} ################################################## Training, Epoch: 0027, Batch: 000006, Sample Num: 96, Cur Loss: 0.30931807, Cur Avg Loss: 0.16797940, Log Avg loss: 0.14853642, Global Avg Loss: 0.99550656, Time: 0.0211 Steps: 55360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000016, Sample Num: 256, Cur Loss: 0.20650446, Cur Avg Loss: 0.21835893, Log Avg loss: 0.24858665, Global Avg Loss: 0.99537167, Time: 0.0209 Steps: 55370, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000026, Sample Num: 416, Cur Loss: 0.26501101, Cur Avg Loss: 0.22362776, Log Avg loss: 0.23205789, Global Avg Loss: 0.99523384, Time: 0.0209 Steps: 55380, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000036, Sample Num: 576, Cur Loss: 0.13313197, Cur Avg Loss: 0.21236268, Log Avg loss: 0.18307347, Global Avg Loss: 0.99508721, Time: 0.0208 Steps: 55390, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000046, Sample Num: 736, Cur Loss: 0.28823486, Cur Avg Loss: 0.22673662, Log Avg loss: 0.27848282, Global Avg Loss: 0.99495786, Time: 0.0209 Steps: 55400, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000056, Sample Num: 896, Cur Loss: 0.23552510, Cur Avg Loss: 0.22063872, Log Avg loss: 0.19258837, Global Avg Loss: 0.99481305, Time: 0.0208 Steps: 55410, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000066, Sample Num: 1056, Cur Loss: 0.51131052, Cur Avg Loss: 0.23004889, Log Avg loss: 0.28274583, Global Avg Loss: 0.99468457, Time: 0.0207 Steps: 55420, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000076, Sample Num: 1216, Cur Loss: 0.17846744, Cur Avg Loss: 0.22283773, Log Avg loss: 0.17524407, Global Avg Loss: 0.99453673, Time: 0.0207 Steps: 55430, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000086, Sample Num: 1376, Cur Loss: 0.54424888, Cur Avg Loss: 0.22811080, Log Avg loss: 0.26818617, Global Avg Loss: 0.99440572, Time: 0.0207 Steps: 55440, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000096, Sample Num: 1536, Cur Loss: 0.59697026, Cur Avg Loss: 0.23510456, Log Avg loss: 0.29525085, Global Avg Loss: 0.99427963, Time: 0.0208 Steps: 55450, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000106, Sample Num: 1696, Cur Loss: 0.20802097, Cur Avg Loss: 0.23213834, Log Avg loss: 0.20366263, Global Avg Loss: 0.99413707, Time: 0.0209 Steps: 55460, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000116, Sample Num: 1856, Cur Loss: 0.07337825, Cur Avg Loss: 0.23158140, Log Avg loss: 0.22567781, Global Avg Loss: 0.99399854, Time: 0.0208 Steps: 55470, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000126, Sample Num: 2016, Cur Loss: 0.17065221, Cur Avg Loss: 0.22293303, Log Avg loss: 0.12261195, Global Avg Loss: 0.99384148, Time: 0.0207 Steps: 55480, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000136, Sample Num: 2176, Cur Loss: 0.13483760, Cur Avg Loss: 0.22201707, Log Avg loss: 0.21047595, Global Avg Loss: 0.99370030, Time: 0.0208 Steps: 55490, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000146, Sample Num: 2336, Cur Loss: 0.56527257, Cur Avg Loss: 0.22413199, Log Avg loss: 0.25289498, Global Avg Loss: 0.99356682, Time: 0.0209 Steps: 55500, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000156, Sample Num: 2496, Cur Loss: 0.11753935, Cur Avg Loss: 0.22567726, Log Avg loss: 0.24823823, Global Avg Loss: 0.99343256, Time: 0.0207 Steps: 55510, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000166, Sample Num: 2656, Cur Loss: 0.20671690, Cur Avg Loss: 0.22503265, Log Avg loss: 0.21497673, Global Avg Loss: 0.99329234, Time: 0.0209 Steps: 55520, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000176, Sample Num: 2816, Cur Loss: 0.07816879, Cur Avg Loss: 0.22229331, Log Avg loss: 0.17682025, Global Avg Loss: 0.99314531, Time: 0.0209 Steps: 55530, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000186, Sample Num: 2976, Cur Loss: 0.14616498, Cur Avg Loss: 0.21949550, Log Avg loss: 0.17025408, Global Avg Loss: 0.99299715, Time: 0.0208 Steps: 55540, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000196, Sample Num: 3136, Cur Loss: 0.13965230, Cur Avg Loss: 0.21921234, Log Avg loss: 0.21394550, Global Avg Loss: 0.99285691, Time: 0.0207 Steps: 55550, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000206, Sample Num: 3296, Cur Loss: 0.16049509, Cur Avg Loss: 0.22002126, Log Avg loss: 0.23587615, Global Avg Loss: 0.99272066, Time: 0.0210 Steps: 55560, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000216, Sample Num: 3456, Cur Loss: 0.29370522, Cur Avg Loss: 0.22441751, Log Avg loss: 0.31498019, Global Avg Loss: 0.99259870, Time: 0.0207 Steps: 55570, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000226, Sample Num: 3616, Cur Loss: 0.14067493, Cur Avg Loss: 0.22366738, Log Avg loss: 0.20746456, Global Avg Loss: 0.99245744, Time: 0.0208 Steps: 55580, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000236, Sample Num: 3776, Cur Loss: 0.11340210, Cur Avg Loss: 0.22271404, Log Avg loss: 0.20116868, Global Avg Loss: 0.99231509, Time: 0.0209 Steps: 55590, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000246, Sample Num: 3936, Cur Loss: 0.13696247, Cur Avg Loss: 0.22378774, Log Avg loss: 0.24912691, Global Avg Loss: 0.99218143, Time: 0.0209 Steps: 55600, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000256, Sample Num: 4096, Cur Loss: 0.14141184, Cur Avg Loss: 0.22450485, Log Avg loss: 0.24214581, Global Avg Loss: 0.99204655, Time: 0.0252 Steps: 55610, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000266, Sample Num: 4256, Cur Loss: 0.10463884, Cur Avg Loss: 0.22465674, Log Avg loss: 0.22854520, Global Avg Loss: 0.99190928, Time: 0.0209 Steps: 55620, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000276, Sample Num: 4416, Cur Loss: 0.15400678, Cur Avg Loss: 0.22268850, Log Avg loss: 0.17033316, Global Avg Loss: 0.99176159, Time: 0.0209 Steps: 55630, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000286, Sample Num: 4576, Cur Loss: 0.20283517, Cur Avg Loss: 0.22176517, Log Avg loss: 0.19628138, Global Avg Loss: 0.99161863, Time: 0.0208 Steps: 55640, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000296, Sample Num: 4736, Cur Loss: 0.16166684, Cur Avg Loss: 0.22148159, Log Avg loss: 0.21337103, Global Avg Loss: 0.99147878, Time: 0.0208 Steps: 55650, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000306, Sample Num: 4896, Cur Loss: 0.28746337, Cur Avg Loss: 0.22429619, Log Avg loss: 0.30760860, Global Avg Loss: 0.99135591, Time: 0.0209 Steps: 55660, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000316, Sample Num: 5056, Cur Loss: 0.26915938, Cur Avg Loss: 0.22305787, Log Avg loss: 0.18516500, Global Avg Loss: 0.99121110, Time: 0.0209 Steps: 55670, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000326, Sample Num: 5216, Cur Loss: 0.13725013, Cur Avg Loss: 0.22066229, Log Avg loss: 0.14496220, Global Avg Loss: 0.99105911, Time: 0.0208 Steps: 55680, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000336, Sample Num: 5376, Cur Loss: 0.20398721, Cur Avg Loss: 0.22106559, Log Avg loss: 0.23421322, Global Avg Loss: 0.99092321, Time: 0.0208 Steps: 55690, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000346, Sample Num: 5536, Cur Loss: 0.50196844, Cur Avg Loss: 0.22184258, Log Avg loss: 0.24794947, Global Avg Loss: 0.99078982, Time: 0.0208 Steps: 55700, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000356, Sample Num: 5696, Cur Loss: 0.41359234, Cur Avg Loss: 0.22157133, Log Avg loss: 0.21218593, Global Avg Loss: 0.99065006, Time: 0.0208 Steps: 55710, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000366, Sample Num: 5856, Cur Loss: 0.28407061, Cur Avg Loss: 0.22297056, Log Avg loss: 0.27278315, Global Avg Loss: 0.99052123, Time: 0.0208 Steps: 55720, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000376, Sample Num: 6016, Cur Loss: 0.28917843, Cur Avg Loss: 0.22260401, Log Avg loss: 0.20918820, Global Avg Loss: 0.99038103, Time: 0.0208 Steps: 55730, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000386, Sample Num: 6176, Cur Loss: 0.24921659, Cur Avg Loss: 0.22051221, Log Avg loss: 0.14186059, Global Avg Loss: 0.99022880, Time: 0.0208 Steps: 55740, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000396, Sample Num: 6336, Cur Loss: 0.21317932, Cur Avg Loss: 0.22071486, Log Avg loss: 0.22853702, Global Avg Loss: 0.99009217, Time: 0.0208 Steps: 55750, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000406, Sample Num: 6496, Cur Loss: 0.18538375, Cur Avg Loss: 0.21961169, Log Avg loss: 0.17592620, Global Avg Loss: 0.98994616, Time: 0.0208 Steps: 55760, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000416, Sample Num: 6656, Cur Loss: 0.40954307, Cur Avg Loss: 0.21831824, Log Avg loss: 0.16580430, Global Avg Loss: 0.98979838, Time: 0.0208 Steps: 55770, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000426, Sample Num: 6816, Cur Loss: 0.09057407, Cur Avg Loss: 0.21838209, Log Avg loss: 0.22103839, Global Avg Loss: 0.98966056, Time: 0.0208 Steps: 55780, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000436, Sample Num: 6976, Cur Loss: 0.02945474, Cur Avg Loss: 0.21846983, Log Avg loss: 0.22220731, Global Avg Loss: 0.98952300, Time: 0.0208 Steps: 55790, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000446, Sample Num: 7136, Cur Loss: 0.18930769, Cur Avg Loss: 0.21837414, Log Avg loss: 0.21420217, Global Avg Loss: 0.98938406, Time: 0.0208 Steps: 55800, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000456, Sample Num: 7296, Cur Loss: 0.17442793, Cur Avg Loss: 0.21817261, Log Avg loss: 0.20918417, Global Avg Loss: 0.98924426, Time: 0.0209 Steps: 55810, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000466, Sample Num: 7456, Cur Loss: 0.05739285, Cur Avg Loss: 0.21758176, Log Avg loss: 0.19063908, Global Avg Loss: 0.98910119, Time: 0.0209 Steps: 55820, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000476, Sample Num: 7616, Cur Loss: 0.05371586, Cur Avg Loss: 0.21598024, Log Avg loss: 0.14134944, Global Avg Loss: 0.98894935, Time: 0.0208 Steps: 55830, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000486, Sample Num: 7776, Cur Loss: 0.17436527, Cur Avg Loss: 0.21552880, Log Avg loss: 0.19404014, Global Avg Loss: 0.98880699, Time: 0.0208 Steps: 55840, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000496, Sample Num: 7936, Cur Loss: 0.15561306, Cur Avg Loss: 0.21572719, Log Avg loss: 0.22536904, Global Avg Loss: 0.98867030, Time: 0.0208 Steps: 55850, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000506, Sample Num: 8096, Cur Loss: 0.06599270, Cur Avg Loss: 0.21599333, Log Avg loss: 0.22919402, Global Avg Loss: 0.98853434, Time: 0.0208 Steps: 55860, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000516, Sample Num: 8256, Cur Loss: 0.08130171, Cur Avg Loss: 0.21587653, Log Avg loss: 0.20996657, Global Avg Loss: 0.98839498, Time: 0.0244 Steps: 55870, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000526, Sample Num: 8416, Cur Loss: 0.45080483, Cur Avg Loss: 0.21696426, Log Avg loss: 0.27309073, Global Avg Loss: 0.98826698, Time: 0.0207 Steps: 55880, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000536, Sample Num: 8576, Cur Loss: 0.22070333, Cur Avg Loss: 0.21743285, Log Avg loss: 0.24208066, Global Avg Loss: 0.98813347, Time: 0.0207 Steps: 55890, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000546, Sample Num: 8736, Cur Loss: 0.10597051, Cur Avg Loss: 0.21697934, Log Avg loss: 0.19267153, Global Avg Loss: 0.98799117, Time: 0.0207 Steps: 55900, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000556, Sample Num: 8896, Cur Loss: 0.08022450, Cur Avg Loss: 0.21625750, Log Avg loss: 0.17684482, Global Avg Loss: 0.98784609, Time: 0.0207 Steps: 55910, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000566, Sample Num: 9056, Cur Loss: 0.14601795, Cur Avg Loss: 0.21659107, Log Avg loss: 0.23513763, Global Avg Loss: 0.98771148, Time: 0.0207 Steps: 55920, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000576, Sample Num: 9216, Cur Loss: 0.13758925, Cur Avg Loss: 0.21628296, Log Avg loss: 0.19884414, Global Avg Loss: 0.98757044, Time: 0.0207 Steps: 55930, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000586, Sample Num: 9376, Cur Loss: 0.18114656, Cur Avg Loss: 0.21590797, Log Avg loss: 0.19430866, Global Avg Loss: 0.98742863, Time: 0.0207 Steps: 55940, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000596, Sample Num: 9536, Cur Loss: 0.19892576, Cur Avg Loss: 0.21590425, Log Avg loss: 0.21568594, Global Avg Loss: 0.98729070, Time: 0.0207 Steps: 55950, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000606, Sample Num: 9696, Cur Loss: 0.04337208, Cur Avg Loss: 0.21502098, Log Avg loss: 0.16237793, Global Avg Loss: 0.98714328, Time: 0.0207 Steps: 55960, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000616, Sample Num: 9856, Cur Loss: 0.09955842, Cur Avg Loss: 0.21437312, Log Avg loss: 0.17511325, Global Avg Loss: 0.98699820, Time: 0.0207 Steps: 55970, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000626, Sample Num: 10016, Cur Loss: 0.30095747, Cur Avg Loss: 0.21505775, Log Avg loss: 0.25723096, Global Avg Loss: 0.98686784, Time: 0.0207 Steps: 55980, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000636, Sample Num: 10176, Cur Loss: 0.15999797, Cur Avg Loss: 0.21547757, Log Avg loss: 0.24175778, Global Avg Loss: 0.98673476, Time: 0.0207 Steps: 55990, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000646, Sample Num: 10336, Cur Loss: 0.32794523, Cur Avg Loss: 0.21611624, Log Avg loss: 0.25673597, Global Avg Loss: 0.98660440, Time: 0.0207 Steps: 56000, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000656, Sample Num: 10496, Cur Loss: 0.33955616, Cur Avg Loss: 0.21696704, Log Avg loss: 0.27192889, Global Avg Loss: 0.98647681, Time: 0.0207 Steps: 56010, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000666, Sample Num: 10656, Cur Loss: 0.22172584, Cur Avg Loss: 0.21595566, Log Avg loss: 0.14960876, Global Avg Loss: 0.98632742, Time: 0.0207 Steps: 56020, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000676, Sample Num: 10816, Cur Loss: 0.21256572, Cur Avg Loss: 0.21565049, Log Avg loss: 0.19532622, Global Avg Loss: 0.98618624, Time: 0.0207 Steps: 56030, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000686, Sample Num: 10976, Cur Loss: 0.10800716, Cur Avg Loss: 0.21564703, Log Avg loss: 0.21541352, Global Avg Loss: 0.98604870, Time: 0.0208 Steps: 56040, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000696, Sample Num: 11136, Cur Loss: 0.32413894, Cur Avg Loss: 0.21586730, Log Avg loss: 0.23097749, Global Avg Loss: 0.98591399, Time: 0.0207 Steps: 56050, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000706, Sample Num: 11296, Cur Loss: 0.61420631, Cur Avg Loss: 0.21590757, Log Avg loss: 0.21871012, Global Avg Loss: 0.98577714, Time: 0.0207 Steps: 56060, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000716, Sample Num: 11456, Cur Loss: 0.18143678, Cur Avg Loss: 0.21589277, Log Avg loss: 0.21484844, Global Avg Loss: 0.98563964, Time: 0.0207 Steps: 56070, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000726, Sample Num: 11616, Cur Loss: 0.14037666, Cur Avg Loss: 0.21587046, Log Avg loss: 0.21427275, Global Avg Loss: 0.98550209, Time: 0.0207 Steps: 56080, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000736, Sample Num: 11776, Cur Loss: 0.20251769, Cur Avg Loss: 0.21638585, Log Avg loss: 0.25380335, Global Avg Loss: 0.98537164, Time: 0.0207 Steps: 56090, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000746, Sample Num: 11936, Cur Loss: 0.31023258, Cur Avg Loss: 0.21606959, Log Avg loss: 0.19279289, Global Avg Loss: 0.98523036, Time: 0.0207 Steps: 56100, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000756, Sample Num: 12096, Cur Loss: 0.29522809, Cur Avg Loss: 0.21518902, Log Avg loss: 0.14949870, Global Avg Loss: 0.98508142, Time: 0.0207 Steps: 56110, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000766, Sample Num: 12256, Cur Loss: 0.15465693, Cur Avg Loss: 0.21558569, Log Avg loss: 0.24557358, Global Avg Loss: 0.98494965, Time: 0.0207 Steps: 56120, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000776, Sample Num: 12416, Cur Loss: 0.21967775, Cur Avg Loss: 0.21605137, Log Avg loss: 0.25172265, Global Avg Loss: 0.98481902, Time: 0.0208 Steps: 56130, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000786, Sample Num: 12576, Cur Loss: 0.16962206, Cur Avg Loss: 0.21639759, Log Avg loss: 0.24326416, Global Avg Loss: 0.98468693, Time: 0.0208 Steps: 56140, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000796, Sample Num: 12736, Cur Loss: 0.14323705, Cur Avg Loss: 0.21884091, Log Avg loss: 0.41088591, Global Avg Loss: 0.98458473, Time: 0.0208 Steps: 56150, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000806, Sample Num: 12896, Cur Loss: 0.15167820, Cur Avg Loss: 0.21852445, Log Avg loss: 0.19333414, Global Avg Loss: 0.98444384, Time: 0.0208 Steps: 56160, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000816, Sample Num: 13056, Cur Loss: 0.21094213, Cur Avg Loss: 0.21852873, Log Avg loss: 0.21887335, Global Avg Loss: 0.98430755, Time: 0.0207 Steps: 56170, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000826, Sample Num: 13216, Cur Loss: 0.14862828, Cur Avg Loss: 0.21959522, Log Avg loss: 0.30662120, Global Avg Loss: 0.98418692, Time: 0.0207 Steps: 56180, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000836, Sample Num: 13376, Cur Loss: 0.25821948, Cur Avg Loss: 0.21999001, Log Avg loss: 0.25259971, Global Avg Loss: 0.98405672, Time: 0.0207 Steps: 56190, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000846, Sample Num: 13536, Cur Loss: 0.18837088, Cur Avg Loss: 0.22058992, Log Avg loss: 0.27074264, Global Avg Loss: 0.98392980, Time: 0.0208 Steps: 56200, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000856, Sample Num: 13696, Cur Loss: 0.15036713, Cur Avg Loss: 0.22089592, Log Avg loss: 0.24678287, Global Avg Loss: 0.98379866, Time: 0.0207 Steps: 56210, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000866, Sample Num: 13856, Cur Loss: 0.19069776, Cur Avg Loss: 0.22072129, Log Avg loss: 0.20577316, Global Avg Loss: 0.98366027, Time: 0.0207 Steps: 56220, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000876, Sample Num: 14016, Cur Loss: 0.29580897, Cur Avg Loss: 0.21977939, Log Avg loss: 0.13821075, Global Avg Loss: 0.98350991, Time: 0.0207 Steps: 56230, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000886, Sample Num: 14176, Cur Loss: 0.23817609, Cur Avg Loss: 0.21953682, Log Avg loss: 0.19828800, Global Avg Loss: 0.98337029, Time: 0.0208 Steps: 56240, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000896, Sample Num: 14336, Cur Loss: 0.14264840, Cur Avg Loss: 0.21976015, Log Avg loss: 0.23954677, Global Avg Loss: 0.98323805, Time: 0.0207 Steps: 56250, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000906, Sample Num: 14496, Cur Loss: 0.15535687, Cur Avg Loss: 0.21982634, Log Avg loss: 0.22575748, Global Avg Loss: 0.98310342, Time: 0.0207 Steps: 56260, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000916, Sample Num: 14656, Cur Loss: 0.41912994, Cur Avg Loss: 0.22005339, Log Avg loss: 0.24062345, Global Avg Loss: 0.98297147, Time: 0.0207 Steps: 56270, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000926, Sample Num: 14816, Cur Loss: 0.49214363, Cur Avg Loss: 0.22034363, Log Avg loss: 0.24692970, Global Avg Loss: 0.98284068, Time: 0.0207 Steps: 56280, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000936, Sample Num: 14976, Cur Loss: 0.06734747, Cur Avg Loss: 0.21991724, Log Avg loss: 0.18043348, Global Avg Loss: 0.98269814, Time: 0.0207 Steps: 56290, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000946, Sample Num: 15136, Cur Loss: 0.18929932, Cur Avg Loss: 0.22008329, Log Avg loss: 0.23562557, Global Avg Loss: 0.98256544, Time: 0.0207 Steps: 56300, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000956, Sample Num: 15296, Cur Loss: 0.31976894, Cur Avg Loss: 0.21961530, Log Avg loss: 0.17534367, Global Avg Loss: 0.98242209, Time: 0.0207 Steps: 56310, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000966, Sample Num: 15456, Cur Loss: 0.19432577, Cur Avg Loss: 0.21950710, Log Avg loss: 0.20916343, Global Avg Loss: 0.98228479, Time: 0.0207 Steps: 56320, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000976, Sample Num: 15616, Cur Loss: 0.29144982, Cur Avg Loss: 0.21934857, Log Avg loss: 0.20403472, Global Avg Loss: 0.98214663, Time: 0.0207 Steps: 56330, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000986, Sample Num: 15776, Cur Loss: 0.27769625, Cur Avg Loss: 0.21962563, Log Avg loss: 0.24666658, Global Avg Loss: 0.98201609, Time: 0.0207 Steps: 56340, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000996, Sample Num: 15936, Cur Loss: 0.19112730, Cur Avg Loss: 0.21905002, Log Avg loss: 0.16229461, Global Avg Loss: 0.98187062, Time: 0.0207 Steps: 56350, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001006, Sample Num: 16096, Cur Loss: 0.21800454, Cur Avg Loss: 0.21947129, Log Avg loss: 0.26143010, Global Avg Loss: 0.98174279, Time: 0.0207 Steps: 56360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001016, Sample Num: 16256, Cur Loss: 0.14452147, Cur Avg Loss: 0.21909403, Log Avg loss: 0.18114162, Global Avg Loss: 0.98160076, Time: 0.0209 Steps: 56370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001026, Sample Num: 16416, Cur Loss: 0.08019947, Cur Avg Loss: 0.21843319, Log Avg loss: 0.15129212, Global Avg Loss: 0.98145349, Time: 0.0246 Steps: 56380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001036, Sample Num: 16576, Cur Loss: 0.27098438, Cur Avg Loss: 0.21861596, Log Avg loss: 0.23736733, Global Avg Loss: 0.98132154, Time: 0.0208 Steps: 56390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001046, Sample Num: 16736, Cur Loss: 0.14372349, Cur Avg Loss: 0.21918143, Log Avg loss: 0.27776429, Global Avg Loss: 0.98119680, Time: 0.0207 Steps: 56400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001056, Sample Num: 16896, Cur Loss: 0.12356655, Cur Avg Loss: 0.21944829, Log Avg loss: 0.24736242, Global Avg Loss: 0.98106671, Time: 0.0207 Steps: 56410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001066, Sample Num: 17056, Cur Loss: 0.11051886, Cur Avg Loss: 0.21919839, Log Avg loss: 0.19280844, Global Avg Loss: 0.98092699, Time: 0.0207 Steps: 56420, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001076, Sample Num: 17216, Cur Loss: 0.15144044, Cur Avg Loss: 0.21882766, Log Avg loss: 0.17930784, Global Avg Loss: 0.98078494, Time: 0.0207 Steps: 56430, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001086, Sample Num: 17376, Cur Loss: 0.71705669, Cur Avg Loss: 0.21918271, Log Avg loss: 0.25738618, Global Avg Loss: 0.98065677, Time: 0.0207 Steps: 56440, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001096, Sample Num: 17536, Cur Loss: 0.14010739, Cur Avg Loss: 0.21918552, Log Avg loss: 0.21949028, Global Avg Loss: 0.98052193, Time: 0.0207 Steps: 56450, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001106, Sample Num: 17696, Cur Loss: 0.20965692, Cur Avg Loss: 0.21957699, Log Avg loss: 0.26248279, Global Avg Loss: 0.98039475, Time: 0.0207 Steps: 56460, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001116, Sample Num: 17856, Cur Loss: 0.16420195, Cur Avg Loss: 0.21974813, Log Avg loss: 0.23867632, Global Avg Loss: 0.98026340, Time: 0.0207 Steps: 56470, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001126, Sample Num: 18016, Cur Loss: 0.24272856, Cur Avg Loss: 0.21947737, Log Avg loss: 0.18925996, Global Avg Loss: 0.98012335, Time: 0.0208 Steps: 56480, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001136, Sample Num: 18176, Cur Loss: 0.11732829, Cur Avg Loss: 0.21881585, Log Avg loss: 0.14432885, Global Avg Loss: 0.97997540, Time: 0.0207 Steps: 56490, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001146, Sample Num: 18336, Cur Loss: 0.19867834, Cur Avg Loss: 0.21906573, Log Avg loss: 0.24745285, Global Avg Loss: 0.97984575, Time: 0.0207 Steps: 56500, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001156, Sample Num: 18496, Cur Loss: 0.26952773, Cur Avg Loss: 0.21916723, Log Avg loss: 0.23079820, Global Avg Loss: 0.97971320, Time: 0.0208 Steps: 56510, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001166, Sample Num: 18656, Cur Loss: 0.16035275, Cur Avg Loss: 0.21929841, Log Avg loss: 0.23446282, Global Avg Loss: 0.97958134, Time: 0.0207 Steps: 56520, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001176, Sample Num: 18816, Cur Loss: 0.19150218, Cur Avg Loss: 0.21974560, Log Avg loss: 0.27188852, Global Avg Loss: 0.97945615, Time: 0.0207 Steps: 56530, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001186, Sample Num: 18976, Cur Loss: 0.14953747, Cur Avg Loss: 0.22054270, Log Avg loss: 0.31428098, Global Avg Loss: 0.97933851, Time: 0.0207 Steps: 56540, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001196, Sample Num: 19136, Cur Loss: 0.05602731, Cur Avg Loss: 0.22088302, Log Avg loss: 0.26124573, Global Avg Loss: 0.97921152, Time: 0.0207 Steps: 56550, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001206, Sample Num: 19296, Cur Loss: 0.17227146, Cur Avg Loss: 0.22088617, Log Avg loss: 0.22126255, Global Avg Loss: 0.97907751, Time: 0.0208 Steps: 56560, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001216, Sample Num: 19456, Cur Loss: 0.07281131, Cur Avg Loss: 0.22093031, Log Avg loss: 0.22625379, Global Avg Loss: 0.97894444, Time: 0.0207 Steps: 56570, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001226, Sample Num: 19616, Cur Loss: 0.23624071, Cur Avg Loss: 0.22093366, Log Avg loss: 0.22134070, Global Avg Loss: 0.97881054, Time: 0.0207 Steps: 56580, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001236, Sample Num: 19776, Cur Loss: 0.14565149, Cur Avg Loss: 0.22051894, Log Avg loss: 0.16967437, Global Avg Loss: 0.97866755, Time: 0.0207 Steps: 56590, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001246, Sample Num: 19936, Cur Loss: 0.08466373, Cur Avg Loss: 0.22007151, Log Avg loss: 0.16476872, Global Avg Loss: 0.97852376, Time: 0.0207 Steps: 56600, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001256, Sample Num: 20096, Cur Loss: 0.31016776, Cur Avg Loss: 0.22044228, Log Avg loss: 0.26664015, Global Avg Loss: 0.97839800, Time: 0.0207 Steps: 56610, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001266, Sample Num: 20256, Cur Loss: 0.28876352, Cur Avg Loss: 0.22070242, Log Avg loss: 0.25337684, Global Avg Loss: 0.97826995, Time: 0.0207 Steps: 56620, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001276, Sample Num: 20416, Cur Loss: 0.11918691, Cur Avg Loss: 0.22050056, Log Avg loss: 0.19494483, Global Avg Loss: 0.97813163, Time: 0.0207 Steps: 56630, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001286, Sample Num: 20576, Cur Loss: 0.37073600, Cur Avg Loss: 0.22072353, Log Avg loss: 0.24917438, Global Avg Loss: 0.97800293, Time: 0.0210 Steps: 56640, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001296, Sample Num: 20736, Cur Loss: 0.09958474, Cur Avg Loss: 0.22069597, Log Avg loss: 0.21715231, Global Avg Loss: 0.97786862, Time: 0.0209 Steps: 56650, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001306, Sample Num: 20896, Cur Loss: 0.24859868, Cur Avg Loss: 0.22044346, Log Avg loss: 0.18771730, Global Avg Loss: 0.97772917, Time: 0.0208 Steps: 56660, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001316, Sample Num: 21056, Cur Loss: 0.30220222, Cur Avg Loss: 0.22084606, Log Avg loss: 0.27342555, Global Avg Loss: 0.97760489, Time: 0.0208 Steps: 56670, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001326, Sample Num: 21216, Cur Loss: 0.24877495, Cur Avg Loss: 0.22079411, Log Avg loss: 0.21395818, Global Avg Loss: 0.97747016, Time: 0.0209 Steps: 56680, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001336, Sample Num: 21376, Cur Loss: 0.73560578, Cur Avg Loss: 0.22091251, Log Avg loss: 0.23661255, Global Avg Loss: 0.97733947, Time: 0.0208 Steps: 56690, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001346, Sample Num: 21536, Cur Loss: 0.17650884, Cur Avg Loss: 0.22035043, Log Avg loss: 0.14525622, Global Avg Loss: 0.97719272, Time: 0.0209 Steps: 56700, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001356, Sample Num: 21696, Cur Loss: 0.08778728, Cur Avg Loss: 0.22024471, Log Avg loss: 0.20601511, Global Avg Loss: 0.97705673, Time: 0.0210 Steps: 56710, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001366, Sample Num: 21856, Cur Loss: 0.16081737, Cur Avg Loss: 0.22056907, Log Avg loss: 0.26455129, Global Avg Loss: 0.97693111, Time: 0.0210 Steps: 56720, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001376, Sample Num: 22016, Cur Loss: 0.21456395, Cur Avg Loss: 0.22094566, Log Avg loss: 0.27238884, Global Avg Loss: 0.97680692, Time: 0.0209 Steps: 56730, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001386, Sample Num: 22176, Cur Loss: 0.27080256, Cur Avg Loss: 0.22088275, Log Avg loss: 0.21222545, Global Avg Loss: 0.97667217, Time: 0.0210 Steps: 56740, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001396, Sample Num: 22336, Cur Loss: 0.15699580, Cur Avg Loss: 0.22084138, Log Avg loss: 0.21510735, Global Avg Loss: 0.97653797, Time: 0.0210 Steps: 56750, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001406, Sample Num: 22496, Cur Loss: 0.23162076, Cur Avg Loss: 0.22068533, Log Avg loss: 0.19890108, Global Avg Loss: 0.97640097, Time: 0.0210 Steps: 56760, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001416, Sample Num: 22656, Cur Loss: 0.25856543, Cur Avg Loss: 0.22091655, Log Avg loss: 0.25342697, Global Avg Loss: 0.97627362, Time: 0.0210 Steps: 56770, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001426, Sample Num: 22816, Cur Loss: 0.14904062, Cur Avg Loss: 0.22114989, Log Avg loss: 0.25419033, Global Avg Loss: 0.97614645, Time: 0.0209 Steps: 56780, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001436, Sample Num: 22976, Cur Loss: 0.06016905, Cur Avg Loss: 0.22090346, Log Avg loss: 0.18576285, Global Avg Loss: 0.97600727, Time: 0.0210 Steps: 56790, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001446, Sample Num: 23136, Cur Loss: 0.44063258, Cur Avg Loss: 0.22213956, Log Avg loss: 0.39964339, Global Avg Loss: 0.97590580, Time: 0.0210 Steps: 56800, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001456, Sample Num: 23296, Cur Loss: 0.11332000, Cur Avg Loss: 0.22288483, Log Avg loss: 0.33065116, Global Avg Loss: 0.97579222, Time: 0.0210 Steps: 56810, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001466, Sample Num: 23456, Cur Loss: 0.10593387, Cur Avg Loss: 0.22267787, Log Avg loss: 0.19254406, Global Avg Loss: 0.97565437, Time: 0.0210 Steps: 56820, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001476, Sample Num: 23616, Cur Loss: 0.13394384, Cur Avg Loss: 0.22252506, Log Avg loss: 0.20012307, Global Avg Loss: 0.97551790, Time: 0.0210 Steps: 56830, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001486, Sample Num: 23776, Cur Loss: 0.12701569, Cur Avg Loss: 0.22228739, Log Avg loss: 0.18720801, Global Avg Loss: 0.97537921, Time: 0.0210 Steps: 56840, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001496, Sample Num: 23936, Cur Loss: 0.42505139, Cur Avg Loss: 0.22254814, Log Avg loss: 0.26129497, Global Avg Loss: 0.97525361, Time: 0.0209 Steps: 56850, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001506, Sample Num: 24096, Cur Loss: 0.18276973, Cur Avg Loss: 0.22237952, Log Avg loss: 0.19715442, Global Avg Loss: 0.97511676, Time: 0.0210 Steps: 56860, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001516, Sample Num: 24256, Cur Loss: 0.30920747, Cur Avg Loss: 0.22262370, Log Avg loss: 0.25939698, Global Avg Loss: 0.97499091, Time: 0.0210 Steps: 56870, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001526, Sample Num: 24416, Cur Loss: 0.29715627, Cur Avg Loss: 0.22271530, Log Avg loss: 0.23660165, Global Avg Loss: 0.97486109, Time: 0.0210 Steps: 56880, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001536, Sample Num: 24576, Cur Loss: 0.27992126, Cur Avg Loss: 0.22342919, Log Avg loss: 0.33236877, Global Avg Loss: 0.97474816, Time: 0.0255 Steps: 56890, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001546, Sample Num: 24736, Cur Loss: 0.20220286, Cur Avg Loss: 0.22328872, Log Avg loss: 0.20171271, Global Avg Loss: 0.97461230, Time: 0.0209 Steps: 56900, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001556, Sample Num: 24896, Cur Loss: 0.09408735, Cur Avg Loss: 0.22300988, Log Avg loss: 0.17990167, Global Avg Loss: 0.97447266, Time: 0.0210 Steps: 56910, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001566, Sample Num: 25056, Cur Loss: 0.06537078, Cur Avg Loss: 0.22265808, Log Avg loss: 0.16791794, Global Avg Loss: 0.97433096, Time: 0.0210 Steps: 56920, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001576, Sample Num: 25216, Cur Loss: 0.06381412, Cur Avg Loss: 0.22224037, Log Avg loss: 0.15682627, Global Avg Loss: 0.97418736, Time: 0.0210 Steps: 56930, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001586, Sample Num: 25376, Cur Loss: 0.35345742, Cur Avg Loss: 0.22207071, Log Avg loss: 0.19533264, Global Avg Loss: 0.97405057, Time: 0.0209 Steps: 56940, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001596, Sample Num: 25536, Cur Loss: 0.16294813, Cur Avg Loss: 0.22217930, Log Avg loss: 0.23940149, Global Avg Loss: 0.97392157, Time: 0.0209 Steps: 56950, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001606, Sample Num: 25696, Cur Loss: 0.44756663, Cur Avg Loss: 0.22256116, Log Avg loss: 0.28350658, Global Avg Loss: 0.97380036, Time: 0.0209 Steps: 56960, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001616, Sample Num: 25856, Cur Loss: 0.28607449, Cur Avg Loss: 0.22251203, Log Avg loss: 0.21462159, Global Avg Loss: 0.97366710, Time: 0.0209 Steps: 56970, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001626, Sample Num: 26016, Cur Loss: 0.10194977, Cur Avg Loss: 0.22230747, Log Avg loss: 0.18925064, Global Avg Loss: 0.97352944, Time: 0.0209 Steps: 56980, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001636, Sample Num: 26176, Cur Loss: 0.26186618, Cur Avg Loss: 0.22226513, Log Avg loss: 0.21537940, Global Avg Loss: 0.97339641, Time: 0.0209 Steps: 56990, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001646, Sample Num: 26336, Cur Loss: 0.09133275, Cur Avg Loss: 0.22194734, Log Avg loss: 0.16995706, Global Avg Loss: 0.97325545, Time: 0.0209 Steps: 57000, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001656, Sample Num: 26496, Cur Loss: 0.07530415, Cur Avg Loss: 0.22185373, Log Avg loss: 0.20644585, Global Avg Loss: 0.97312095, Time: 0.0210 Steps: 57010, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001666, Sample Num: 26656, Cur Loss: 0.18777452, Cur Avg Loss: 0.22183178, Log Avg loss: 0.21819684, Global Avg Loss: 0.97298855, Time: 0.0209 Steps: 57020, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001676, Sample Num: 26816, Cur Loss: 0.31070453, Cur Avg Loss: 0.22164092, Log Avg loss: 0.18984313, Global Avg Loss: 0.97285123, Time: 0.0209 Steps: 57030, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001686, Sample Num: 26976, Cur Loss: 0.11863859, Cur Avg Loss: 0.22158159, Log Avg loss: 0.21163813, Global Avg Loss: 0.97271778, Time: 0.0209 Steps: 57040, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001696, Sample Num: 27136, Cur Loss: 0.17829645, Cur Avg Loss: 0.22156770, Log Avg loss: 0.21922676, Global Avg Loss: 0.97258570, Time: 0.0209 Steps: 57050, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001706, Sample Num: 27296, Cur Loss: 0.42259020, Cur Avg Loss: 0.22215176, Log Avg loss: 0.32120754, Global Avg Loss: 0.97247155, Time: 0.0209 Steps: 57060, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001716, Sample Num: 27456, Cur Loss: 0.30602989, Cur Avg Loss: 0.22245054, Log Avg loss: 0.27342272, Global Avg Loss: 0.97234906, Time: 0.0209 Steps: 57070, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001726, Sample Num: 27616, Cur Loss: 0.12090182, Cur Avg Loss: 0.22241138, Log Avg loss: 0.21569144, Global Avg Loss: 0.97221650, Time: 0.0210 Steps: 57080, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001736, Sample Num: 27776, Cur Loss: 0.23381490, Cur Avg Loss: 0.22234509, Log Avg loss: 0.21090279, Global Avg Loss: 0.97208314, Time: 0.0209 Steps: 57090, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001746, Sample Num: 27936, Cur Loss: 0.09039317, Cur Avg Loss: 0.22309800, Log Avg loss: 0.35380434, Global Avg Loss: 0.97197486, Time: 0.0209 Steps: 57100, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001756, Sample Num: 28096, Cur Loss: 0.18332171, Cur Avg Loss: 0.22315791, Log Avg loss: 0.23361755, Global Avg Loss: 0.97184557, Time: 0.0209 Steps: 57110, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001766, Sample Num: 28256, Cur Loss: 0.29053703, Cur Avg Loss: 0.22320587, Log Avg loss: 0.23162735, Global Avg Loss: 0.97171598, Time: 0.0209 Steps: 57120, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001776, Sample Num: 28416, Cur Loss: 0.13642919, Cur Avg Loss: 0.22328644, Log Avg loss: 0.23751614, Global Avg Loss: 0.97158747, Time: 0.0209 Steps: 57130, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001786, Sample Num: 28576, Cur Loss: 0.15276298, Cur Avg Loss: 0.22311391, Log Avg loss: 0.19247273, Global Avg Loss: 0.97145112, Time: 0.0209 Steps: 57140, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001796, Sample Num: 28736, Cur Loss: 0.15931126, Cur Avg Loss: 0.22310560, Log Avg loss: 0.22162160, Global Avg Loss: 0.97131992, Time: 0.0247 Steps: 57150, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001806, Sample Num: 28896, Cur Loss: 0.28167397, Cur Avg Loss: 0.22298647, Log Avg loss: 0.20159060, Global Avg Loss: 0.97118525, Time: 0.0219 Steps: 57160, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001816, Sample Num: 29056, Cur Loss: 0.11401977, Cur Avg Loss: 0.22313413, Log Avg loss: 0.24980120, Global Avg Loss: 0.97105907, Time: 0.0219 Steps: 57170, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001826, Sample Num: 29216, Cur Loss: 0.70366198, Cur Avg Loss: 0.22431355, Log Avg loss: 0.43849564, Global Avg Loss: 0.97096593, Time: 0.0219 Steps: 57180, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001836, Sample Num: 29376, Cur Loss: 0.15297544, Cur Avg Loss: 0.22475806, Log Avg loss: 0.30592493, Global Avg Loss: 0.97084965, Time: 0.0219 Steps: 57190, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001846, Sample Num: 29536, Cur Loss: 0.22399926, Cur Avg Loss: 0.22463740, Log Avg loss: 0.20248520, Global Avg Loss: 0.97071532, Time: 0.0219 Steps: 57200, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001856, Sample Num: 29696, Cur Loss: 0.27332354, Cur Avg Loss: 0.22455796, Log Avg loss: 0.20989244, Global Avg Loss: 0.97058233, Time: 0.0220 Steps: 57210, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001866, Sample Num: 29856, Cur Loss: 0.07903452, Cur Avg Loss: 0.22467869, Log Avg loss: 0.24708704, Global Avg Loss: 0.97045589, Time: 0.0219 Steps: 57220, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001876, Sample Num: 30016, Cur Loss: 0.29270199, Cur Avg Loss: 0.22489516, Log Avg loss: 0.26528919, Global Avg Loss: 0.97033267, Time: 0.0219 Steps: 57230, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001886, Sample Num: 30176, Cur Loss: 0.23852015, Cur Avg Loss: 0.22461167, Log Avg loss: 0.17142829, Global Avg Loss: 0.97019310, Time: 0.0220 Steps: 57240, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001896, Sample Num: 30336, Cur Loss: 0.31415725, Cur Avg Loss: 0.22517232, Log Avg loss: 0.33091151, Global Avg Loss: 0.97008144, Time: 0.0219 Steps: 57250, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001906, Sample Num: 30496, Cur Loss: 0.55711490, Cur Avg Loss: 0.22557630, Log Avg loss: 0.30216923, Global Avg Loss: 0.96996479, Time: 0.0219 Steps: 57260, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001916, Sample Num: 30656, Cur Loss: 0.20882842, Cur Avg Loss: 0.22567892, Log Avg loss: 0.24523976, Global Avg Loss: 0.96983825, Time: 0.0219 Steps: 57270, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001926, Sample Num: 30816, Cur Loss: 0.26229274, Cur Avg Loss: 0.22588149, Log Avg loss: 0.26469396, Global Avg Loss: 0.96971514, Time: 0.0219 Steps: 57280, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001936, Sample Num: 30976, Cur Loss: 0.26161429, Cur Avg Loss: 0.22574646, Log Avg loss: 0.19973815, Global Avg Loss: 0.96958074, Time: 0.0219 Steps: 57290, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001946, Sample Num: 31136, Cur Loss: 0.43383807, Cur Avg Loss: 0.22603292, Log Avg loss: 0.28149278, Global Avg Loss: 0.96946066, Time: 0.0219 Steps: 57300, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001956, Sample Num: 31296, Cur Loss: 0.15288064, Cur Avg Loss: 0.22595303, Log Avg loss: 0.21040555, Global Avg Loss: 0.96932821, Time: 0.0219 Steps: 57310, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001966, Sample Num: 31456, Cur Loss: 0.17216168, Cur Avg Loss: 0.22575171, Log Avg loss: 0.18637418, Global Avg Loss: 0.96919161, Time: 0.0219 Steps: 57320, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001976, Sample Num: 31616, Cur Loss: 0.20118728, Cur Avg Loss: 0.22591839, Log Avg loss: 0.25868754, Global Avg Loss: 0.96906768, Time: 0.0219 Steps: 57330, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001986, Sample Num: 31776, Cur Loss: 0.11478481, Cur Avg Loss: 0.22584254, Log Avg loss: 0.21085425, Global Avg Loss: 0.96893545, Time: 0.0219 Steps: 57340, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001996, Sample Num: 31936, Cur Loss: 0.16221684, Cur Avg Loss: 0.22534895, Log Avg loss: 0.12732332, Global Avg Loss: 0.96878870, Time: 0.0219 Steps: 57350, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002006, Sample Num: 32096, Cur Loss: 0.05111510, Cur Avg Loss: 0.22515236, Log Avg loss: 0.18591211, Global Avg Loss: 0.96865222, Time: 0.0219 Steps: 57360, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002016, Sample Num: 32256, Cur Loss: 0.06609488, Cur Avg Loss: 0.22497152, Log Avg loss: 0.18869499, Global Avg Loss: 0.96851626, Time: 0.0219 Steps: 57370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002026, Sample Num: 32416, Cur Loss: 0.11773080, Cur Avg Loss: 0.22502555, Log Avg loss: 0.23591729, Global Avg Loss: 0.96838859, Time: 0.0219 Steps: 57380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002036, Sample Num: 32576, Cur Loss: 0.23602621, Cur Avg Loss: 0.22524503, Log Avg loss: 0.26971299, Global Avg Loss: 0.96826685, Time: 0.0219 Steps: 57390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002046, Sample Num: 32736, Cur Loss: 0.22997601, Cur Avg Loss: 0.22534837, Log Avg loss: 0.24638692, Global Avg Loss: 0.96814108, Time: 0.0220 Steps: 57400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002056, Sample Num: 32896, Cur Loss: 0.24581158, Cur Avg Loss: 0.22521497, Log Avg loss: 0.19792283, Global Avg Loss: 0.96800692, Time: 0.0214 Steps: 57410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002066, Sample Num: 33056, Cur Loss: 0.15647560, Cur Avg Loss: 0.22498712, Log Avg loss: 0.17814073, Global Avg Loss: 0.96786936, Time: 0.0210 Steps: 57420, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002076, Sample Num: 33216, Cur Loss: 0.28945541, Cur Avg Loss: 0.22540672, Log Avg loss: 0.31209632, Global Avg Loss: 0.96775518, Time: 0.0210 Steps: 57430, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002086, Sample Num: 33376, Cur Loss: 0.07105506, Cur Avg Loss: 0.22520990, Log Avg loss: 0.18434974, Global Avg Loss: 0.96761879, Time: 0.0210 Steps: 57440, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002096, Sample Num: 33536, Cur Loss: 0.04780217, Cur Avg Loss: 0.22494093, Log Avg loss: 0.16883415, Global Avg Loss: 0.96747975, Time: 0.0211 Steps: 57450, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002106, Sample Num: 33696, Cur Loss: 0.28197652, Cur Avg Loss: 0.22497327, Log Avg loss: 0.23175091, Global Avg Loss: 0.96735171, Time: 0.0211 Steps: 57460, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002116, Sample Num: 33856, Cur Loss: 0.20233876, Cur Avg Loss: 0.22503212, Log Avg loss: 0.23742549, Global Avg Loss: 0.96722470, Time: 0.0212 Steps: 57470, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002126, Sample Num: 34016, Cur Loss: 0.09923598, Cur Avg Loss: 0.22490304, Log Avg loss: 0.19758999, Global Avg Loss: 0.96709080, Time: 0.0211 Steps: 57480, Updated lr: 0.000046 ***** Running evaluation checkpoint-57483 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-57483 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.714762, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.284862, "eval_total_loss": 200.257982, "eval_mae": 0.35897, "eval_mse": 0.284969, "eval_r2": 0.818855, "eval_sp_statistic": 0.884002, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.910813, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.790675, "test_total_loss": 396.919078, "test_mae": 0.670882, "test_mse": 0.790789, "test_r2": 0.489618, "test_sp_statistic": 0.795557, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.833764, "test_ps_pvalue": 0.0, "lr": 4.6436225699383594e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9670455986891018, "train_cur_epoch_loss": 478.4466551113874, "train_cur_epoch_avg_loss": 0.2247283490424553, "train_cur_epoch_time": 44.71476221084595, "train_cur_epoch_avg_time": 0.02100270653398119, "epoch": 27, "step": 57483} ################################################## Training, Epoch: 0028, Batch: 000007, Sample Num: 112, Cur Loss: 0.16013229, Cur Avg Loss: 0.20759150, Log Avg loss: 0.17559374, Global Avg Loss: 0.96695313, Time: 0.0210 Steps: 57490, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000017, Sample Num: 272, Cur Loss: 0.08898506, Cur Avg Loss: 0.18187131, Log Avg loss: 0.16386719, Global Avg Loss: 0.96681346, Time: 0.0209 Steps: 57500, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000027, Sample Num: 432, Cur Loss: 0.13681142, Cur Avg Loss: 0.18838847, Log Avg loss: 0.19946762, Global Avg Loss: 0.96668003, Time: 0.0210 Steps: 57510, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000037, Sample Num: 592, Cur Loss: 0.09174938, Cur Avg Loss: 0.18125302, Log Avg loss: 0.16198733, Global Avg Loss: 0.96654013, Time: 0.0210 Steps: 57520, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000047, Sample Num: 752, Cur Loss: 0.20668924, Cur Avg Loss: 0.19294070, Log Avg loss: 0.23618510, Global Avg Loss: 0.96641318, Time: 0.0210 Steps: 57530, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000057, Sample Num: 912, Cur Loss: 0.17139618, Cur Avg Loss: 0.18861333, Log Avg loss: 0.16827469, Global Avg Loss: 0.96627447, Time: 0.0210 Steps: 57540, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000067, Sample Num: 1072, Cur Loss: 0.18143719, Cur Avg Loss: 0.20957592, Log Avg loss: 0.32906270, Global Avg Loss: 0.96616375, Time: 0.0209 Steps: 57550, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000077, Sample Num: 1232, Cur Loss: 0.22054501, Cur Avg Loss: 0.20490161, Log Avg loss: 0.17358370, Global Avg Loss: 0.96602605, Time: 0.0211 Steps: 57560, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000087, Sample Num: 1392, Cur Loss: 0.24293575, Cur Avg Loss: 0.20854013, Log Avg loss: 0.23655674, Global Avg Loss: 0.96589934, Time: 0.0209 Steps: 57570, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000097, Sample Num: 1552, Cur Loss: 0.16637051, Cur Avg Loss: 0.20886079, Log Avg loss: 0.21165054, Global Avg Loss: 0.96576835, Time: 0.0208 Steps: 57580, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000107, Sample Num: 1712, Cur Loss: 0.17886259, Cur Avg Loss: 0.21021058, Log Avg loss: 0.22330355, Global Avg Loss: 0.96563943, Time: 0.0210 Steps: 57590, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000117, Sample Num: 1872, Cur Loss: 0.11769459, Cur Avg Loss: 0.20326075, Log Avg loss: 0.12889760, Global Avg Loss: 0.96549416, Time: 0.0209 Steps: 57600, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000127, Sample Num: 2032, Cur Loss: 0.34304351, Cur Avg Loss: 0.21256472, Log Avg loss: 0.32142118, Global Avg Loss: 0.96538236, Time: 0.0210 Steps: 57610, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000137, Sample Num: 2192, Cur Loss: 0.15189418, Cur Avg Loss: 0.21599461, Log Avg loss: 0.25955420, Global Avg Loss: 0.96525986, Time: 0.0211 Steps: 57620, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000147, Sample Num: 2352, Cur Loss: 0.47709179, Cur Avg Loss: 0.21476326, Log Avg loss: 0.19789377, Global Avg Loss: 0.96512671, Time: 0.0210 Steps: 57630, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000157, Sample Num: 2512, Cur Loss: 0.17487574, Cur Avg Loss: 0.20963067, Log Avg loss: 0.13418152, Global Avg Loss: 0.96498255, Time: 0.0210 Steps: 57640, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000167, Sample Num: 2672, Cur Loss: 0.25615287, Cur Avg Loss: 0.20657844, Log Avg loss: 0.15865851, Global Avg Loss: 0.96484268, Time: 0.0210 Steps: 57650, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000177, Sample Num: 2832, Cur Loss: 0.08466795, Cur Avg Loss: 0.20726506, Log Avg loss: 0.21873152, Global Avg Loss: 0.96471329, Time: 0.0211 Steps: 57660, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000187, Sample Num: 2992, Cur Loss: 0.18757583, Cur Avg Loss: 0.20885377, Log Avg loss: 0.23697402, Global Avg Loss: 0.96458710, Time: 0.0209 Steps: 57670, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000197, Sample Num: 3152, Cur Loss: 0.08749980, Cur Avg Loss: 0.20936182, Log Avg loss: 0.21886229, Global Avg Loss: 0.96445781, Time: 0.0209 Steps: 57680, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000207, Sample Num: 3312, Cur Loss: 0.63678938, Cur Avg Loss: 0.21124980, Log Avg loss: 0.24844295, Global Avg Loss: 0.96433369, Time: 0.0210 Steps: 57690, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000217, Sample Num: 3472, Cur Loss: 0.20684533, Cur Avg Loss: 0.21110414, Log Avg loss: 0.20808907, Global Avg Loss: 0.96420263, Time: 0.0211 Steps: 57700, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000227, Sample Num: 3632, Cur Loss: 0.14901091, Cur Avg Loss: 0.21137665, Log Avg loss: 0.21729011, Global Avg Loss: 0.96407320, Time: 0.0209 Steps: 57710, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000237, Sample Num: 3792, Cur Loss: 0.09429760, Cur Avg Loss: 0.21223291, Log Avg loss: 0.23166995, Global Avg Loss: 0.96394632, Time: 0.0209 Steps: 57720, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000247, Sample Num: 3952, Cur Loss: 0.28387752, Cur Avg Loss: 0.20918013, Log Avg loss: 0.13682941, Global Avg Loss: 0.96380304, Time: 0.0210 Steps: 57730, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000257, Sample Num: 4112, Cur Loss: 0.29346946, Cur Avg Loss: 0.20785682, Log Avg loss: 0.17517103, Global Avg Loss: 0.96366646, Time: 0.0247 Steps: 57740, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000267, Sample Num: 4272, Cur Loss: 0.18302512, Cur Avg Loss: 0.21151685, Log Avg loss: 0.30557947, Global Avg Loss: 0.96355250, Time: 0.0210 Steps: 57750, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000277, Sample Num: 4432, Cur Loss: 0.32279202, Cur Avg Loss: 0.21285698, Log Avg loss: 0.24863840, Global Avg Loss: 0.96342873, Time: 0.0210 Steps: 57760, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000287, Sample Num: 4592, Cur Loss: 0.22981712, Cur Avg Loss: 0.21185029, Log Avg loss: 0.18396524, Global Avg Loss: 0.96329381, Time: 0.0210 Steps: 57770, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000297, Sample Num: 4752, Cur Loss: 0.20500109, Cur Avg Loss: 0.21132287, Log Avg loss: 0.19618574, Global Avg Loss: 0.96316104, Time: 0.0210 Steps: 57780, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000307, Sample Num: 4912, Cur Loss: 0.55043799, Cur Avg Loss: 0.21308342, Log Avg loss: 0.26537173, Global Avg Loss: 0.96304030, Time: 0.0210 Steps: 57790, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000317, Sample Num: 5072, Cur Loss: 0.11030348, Cur Avg Loss: 0.21363838, Log Avg loss: 0.23067559, Global Avg Loss: 0.96291359, Time: 0.0210 Steps: 57800, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000327, Sample Num: 5232, Cur Loss: 0.22267991, Cur Avg Loss: 0.21527276, Log Avg loss: 0.26708261, Global Avg Loss: 0.96279323, Time: 0.0210 Steps: 57810, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000337, Sample Num: 5392, Cur Loss: 0.28408733, Cur Avg Loss: 0.21594641, Log Avg loss: 0.23797476, Global Avg Loss: 0.96266787, Time: 0.0210 Steps: 57820, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000347, Sample Num: 5552, Cur Loss: 0.24671093, Cur Avg Loss: 0.21737336, Log Avg loss: 0.26546180, Global Avg Loss: 0.96254731, Time: 0.0210 Steps: 57830, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000357, Sample Num: 5712, Cur Loss: 0.29412860, Cur Avg Loss: 0.21948253, Log Avg loss: 0.29267066, Global Avg Loss: 0.96243149, Time: 0.0210 Steps: 57840, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000367, Sample Num: 5872, Cur Loss: 0.27531797, Cur Avg Loss: 0.21901632, Log Avg loss: 0.20237243, Global Avg Loss: 0.96230011, Time: 0.0210 Steps: 57850, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000377, Sample Num: 6032, Cur Loss: 0.26775026, Cur Avg Loss: 0.21877376, Log Avg loss: 0.20987195, Global Avg Loss: 0.96217006, Time: 0.0210 Steps: 57860, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000387, Sample Num: 6192, Cur Loss: 0.21118669, Cur Avg Loss: 0.21834267, Log Avg loss: 0.20209062, Global Avg Loss: 0.96203872, Time: 0.0210 Steps: 57870, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000397, Sample Num: 6352, Cur Loss: 0.03653312, Cur Avg Loss: 0.21713829, Log Avg loss: 0.17052885, Global Avg Loss: 0.96190197, Time: 0.0210 Steps: 57880, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000407, Sample Num: 6512, Cur Loss: 0.04181889, Cur Avg Loss: 0.21667614, Log Avg loss: 0.19832881, Global Avg Loss: 0.96177007, Time: 0.0210 Steps: 57890, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000417, Sample Num: 6672, Cur Loss: 0.34389347, Cur Avg Loss: 0.21590734, Log Avg loss: 0.18461704, Global Avg Loss: 0.96163585, Time: 0.0210 Steps: 57900, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000427, Sample Num: 6832, Cur Loss: 0.28335786, Cur Avg Loss: 0.21441936, Log Avg loss: 0.15237049, Global Avg Loss: 0.96149610, Time: 0.0210 Steps: 57910, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000437, Sample Num: 6992, Cur Loss: 0.17974719, Cur Avg Loss: 0.21488360, Log Avg loss: 0.23470678, Global Avg Loss: 0.96137062, Time: 0.0210 Steps: 57920, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000447, Sample Num: 7152, Cur Loss: 0.24521111, Cur Avg Loss: 0.21526166, Log Avg loss: 0.23178307, Global Avg Loss: 0.96124468, Time: 0.0210 Steps: 57930, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000457, Sample Num: 7312, Cur Loss: 0.15776485, Cur Avg Loss: 0.21650148, Log Avg loss: 0.27192105, Global Avg Loss: 0.96112570, Time: 0.0210 Steps: 57940, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000467, Sample Num: 7472, Cur Loss: 0.15070269, Cur Avg Loss: 0.21691399, Log Avg loss: 0.23576582, Global Avg Loss: 0.96100053, Time: 0.0210 Steps: 57950, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000477, Sample Num: 7632, Cur Loss: 0.18112442, Cur Avg Loss: 0.21668440, Log Avg loss: 0.20596264, Global Avg Loss: 0.96087027, Time: 0.0210 Steps: 57960, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000487, Sample Num: 7792, Cur Loss: 0.11458067, Cur Avg Loss: 0.21547359, Log Avg loss: 0.15771803, Global Avg Loss: 0.96073172, Time: 0.0210 Steps: 57970, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000497, Sample Num: 7952, Cur Loss: 0.07025532, Cur Avg Loss: 0.21458071, Log Avg loss: 0.17109739, Global Avg Loss: 0.96059553, Time: 0.0210 Steps: 57980, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000507, Sample Num: 8112, Cur Loss: 0.20924598, Cur Avg Loss: 0.21570205, Log Avg loss: 0.27143245, Global Avg Loss: 0.96047669, Time: 0.0210 Steps: 57990, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000517, Sample Num: 8272, Cur Loss: 0.12189861, Cur Avg Loss: 0.21568765, Log Avg loss: 0.21495778, Global Avg Loss: 0.96034815, Time: 0.0246 Steps: 58000, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000527, Sample Num: 8432, Cur Loss: 0.17467350, Cur Avg Loss: 0.21581419, Log Avg loss: 0.22235626, Global Avg Loss: 0.96022093, Time: 0.0209 Steps: 58010, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000537, Sample Num: 8592, Cur Loss: 0.34705266, Cur Avg Loss: 0.21699357, Log Avg loss: 0.27914703, Global Avg Loss: 0.96010355, Time: 0.0209 Steps: 58020, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000547, Sample Num: 8752, Cur Loss: 0.21338794, Cur Avg Loss: 0.21685861, Log Avg loss: 0.20961124, Global Avg Loss: 0.95997422, Time: 0.0209 Steps: 58030, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000557, Sample Num: 8912, Cur Loss: 0.23195264, Cur Avg Loss: 0.21652864, Log Avg loss: 0.19847938, Global Avg Loss: 0.95984302, Time: 0.0209 Steps: 58040, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000567, Sample Num: 9072, Cur Loss: 0.06056473, Cur Avg Loss: 0.21685624, Log Avg loss: 0.23510337, Global Avg Loss: 0.95971817, Time: 0.0209 Steps: 58050, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000577, Sample Num: 9232, Cur Loss: 0.13623816, Cur Avg Loss: 0.21833815, Log Avg loss: 0.30236233, Global Avg Loss: 0.95960495, Time: 0.0209 Steps: 58060, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000587, Sample Num: 9392, Cur Loss: 0.16442274, Cur Avg Loss: 0.21922201, Log Avg loss: 0.27022088, Global Avg Loss: 0.95948623, Time: 0.0209 Steps: 58070, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000597, Sample Num: 9552, Cur Loss: 0.15689328, Cur Avg Loss: 0.21840900, Log Avg loss: 0.17068540, Global Avg Loss: 0.95935042, Time: 0.0210 Steps: 58080, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000607, Sample Num: 9712, Cur Loss: 0.20911001, Cur Avg Loss: 0.21750064, Log Avg loss: 0.16327146, Global Avg Loss: 0.95921338, Time: 0.0209 Steps: 58090, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000617, Sample Num: 9872, Cur Loss: 0.11635566, Cur Avg Loss: 0.21793464, Log Avg loss: 0.24427830, Global Avg Loss: 0.95909032, Time: 0.0209 Steps: 58100, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000627, Sample Num: 10032, Cur Loss: 0.14488539, Cur Avg Loss: 0.21851182, Log Avg loss: 0.25412409, Global Avg Loss: 0.95896901, Time: 0.0209 Steps: 58110, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000637, Sample Num: 10192, Cur Loss: 0.07536556, Cur Avg Loss: 0.21719928, Log Avg loss: 0.13490288, Global Avg Loss: 0.95882722, Time: 0.0209 Steps: 58120, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000647, Sample Num: 10352, Cur Loss: 0.31368947, Cur Avg Loss: 0.21764362, Log Avg loss: 0.24594812, Global Avg Loss: 0.95870459, Time: 0.0209 Steps: 58130, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000657, Sample Num: 10512, Cur Loss: 0.13048872, Cur Avg Loss: 0.21798764, Log Avg loss: 0.24024594, Global Avg Loss: 0.95858101, Time: 0.0209 Steps: 58140, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000667, Sample Num: 10672, Cur Loss: 0.10551389, Cur Avg Loss: 0.21817401, Log Avg loss: 0.23041852, Global Avg Loss: 0.95845579, Time: 0.0210 Steps: 58150, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000677, Sample Num: 10832, Cur Loss: 0.25674298, Cur Avg Loss: 0.21762587, Log Avg loss: 0.18106498, Global Avg Loss: 0.95832213, Time: 0.0210 Steps: 58160, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000687, Sample Num: 10992, Cur Loss: 0.13156976, Cur Avg Loss: 0.21673546, Log Avg loss: 0.15645427, Global Avg Loss: 0.95818428, Time: 0.0209 Steps: 58170, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000697, Sample Num: 11152, Cur Loss: 0.27368402, Cur Avg Loss: 0.21625427, Log Avg loss: 0.18319679, Global Avg Loss: 0.95805107, Time: 0.0209 Steps: 58180, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000707, Sample Num: 11312, Cur Loss: 0.12694438, Cur Avg Loss: 0.21621037, Log Avg loss: 0.21315016, Global Avg Loss: 0.95792306, Time: 0.0209 Steps: 58190, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000717, Sample Num: 11472, Cur Loss: 0.18299179, Cur Avg Loss: 0.21533680, Log Avg loss: 0.15357552, Global Avg Loss: 0.95778486, Time: 0.0209 Steps: 58200, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000727, Sample Num: 11632, Cur Loss: 0.17560041, Cur Avg Loss: 0.21606140, Log Avg loss: 0.26801540, Global Avg Loss: 0.95766636, Time: 0.0209 Steps: 58210, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000737, Sample Num: 11792, Cur Loss: 0.10376676, Cur Avg Loss: 0.21606186, Log Avg loss: 0.21609493, Global Avg Loss: 0.95753899, Time: 0.0210 Steps: 58220, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000747, Sample Num: 11952, Cur Loss: 0.08354199, Cur Avg Loss: 0.21541766, Log Avg loss: 0.16794049, Global Avg Loss: 0.95740339, Time: 0.0209 Steps: 58230, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000757, Sample Num: 12112, Cur Loss: 0.29765788, Cur Avg Loss: 0.21492068, Log Avg loss: 0.17779619, Global Avg Loss: 0.95726952, Time: 0.0209 Steps: 58240, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000767, Sample Num: 12272, Cur Loss: 0.47391117, Cur Avg Loss: 0.21495794, Log Avg loss: 0.21777874, Global Avg Loss: 0.95714257, Time: 0.0209 Steps: 58250, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000777, Sample Num: 12432, Cur Loss: 0.16988675, Cur Avg Loss: 0.21567094, Log Avg loss: 0.27035767, Global Avg Loss: 0.95702469, Time: 0.0208 Steps: 58260, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000787, Sample Num: 12592, Cur Loss: 0.15018657, Cur Avg Loss: 0.21581207, Log Avg loss: 0.22677832, Global Avg Loss: 0.95689937, Time: 0.0211 Steps: 58270, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000797, Sample Num: 12752, Cur Loss: 0.21796606, Cur Avg Loss: 0.21581372, Log Avg loss: 0.21594323, Global Avg Loss: 0.95677223, Time: 0.0210 Steps: 58280, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000807, Sample Num: 12912, Cur Loss: 0.21209693, Cur Avg Loss: 0.21519522, Log Avg loss: 0.16590112, Global Avg Loss: 0.95663655, Time: 0.0211 Steps: 58290, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000817, Sample Num: 13072, Cur Loss: 0.26090837, Cur Avg Loss: 0.21515012, Log Avg loss: 0.21151064, Global Avg Loss: 0.95650874, Time: 0.0210 Steps: 58300, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000827, Sample Num: 13232, Cur Loss: 0.09177917, Cur Avg Loss: 0.21502857, Log Avg loss: 0.20509780, Global Avg Loss: 0.95637988, Time: 0.0211 Steps: 58310, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000837, Sample Num: 13392, Cur Loss: 0.11194316, Cur Avg Loss: 0.21528889, Log Avg loss: 0.23681736, Global Avg Loss: 0.95625650, Time: 0.0210 Steps: 58320, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000847, Sample Num: 13552, Cur Loss: 0.22265324, Cur Avg Loss: 0.21511233, Log Avg loss: 0.20033372, Global Avg Loss: 0.95612690, Time: 0.0211 Steps: 58330, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000857, Sample Num: 13712, Cur Loss: 0.28834546, Cur Avg Loss: 0.21649761, Log Avg loss: 0.33383143, Global Avg Loss: 0.95602024, Time: 0.0210 Steps: 58340, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000867, Sample Num: 13872, Cur Loss: 0.11908384, Cur Avg Loss: 0.21644520, Log Avg loss: 0.21195365, Global Avg Loss: 0.95589272, Time: 0.0210 Steps: 58350, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000877, Sample Num: 14032, Cur Loss: 0.43576607, Cur Avg Loss: 0.21638784, Log Avg loss: 0.21141434, Global Avg Loss: 0.95576515, Time: 0.0211 Steps: 58360, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000887, Sample Num: 14192, Cur Loss: 0.08765854, Cur Avg Loss: 0.21616458, Log Avg loss: 0.19658505, Global Avg Loss: 0.95563509, Time: 0.0211 Steps: 58370, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000897, Sample Num: 14352, Cur Loss: 0.13402505, Cur Avg Loss: 0.21613089, Log Avg loss: 0.21314248, Global Avg Loss: 0.95550791, Time: 0.0211 Steps: 58380, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000907, Sample Num: 14512, Cur Loss: 0.13863006, Cur Avg Loss: 0.21512252, Log Avg loss: 0.12467192, Global Avg Loss: 0.95536562, Time: 0.0211 Steps: 58390, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000917, Sample Num: 14672, Cur Loss: 0.03639238, Cur Avg Loss: 0.21523717, Log Avg loss: 0.22563595, Global Avg Loss: 0.95524066, Time: 0.0211 Steps: 58400, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000927, Sample Num: 14832, Cur Loss: 0.16900674, Cur Avg Loss: 0.21578952, Log Avg loss: 0.26643997, Global Avg Loss: 0.95512274, Time: 0.0211 Steps: 58410, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000937, Sample Num: 14992, Cur Loss: 0.13457760, Cur Avg Loss: 0.21533422, Log Avg loss: 0.17312764, Global Avg Loss: 0.95498888, Time: 0.0211 Steps: 58420, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000947, Sample Num: 15152, Cur Loss: 0.32032222, Cur Avg Loss: 0.21579965, Log Avg loss: 0.25940991, Global Avg Loss: 0.95486983, Time: 0.0211 Steps: 58430, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000957, Sample Num: 15312, Cur Loss: 0.23687465, Cur Avg Loss: 0.21570867, Log Avg loss: 0.20709330, Global Avg Loss: 0.95474188, Time: 0.0211 Steps: 58440, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000967, Sample Num: 15472, Cur Loss: 0.30298898, Cur Avg Loss: 0.21569491, Log Avg loss: 0.21437799, Global Avg Loss: 0.95461521, Time: 0.0211 Steps: 58450, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000977, Sample Num: 15632, Cur Loss: 0.11464799, Cur Avg Loss: 0.21565303, Log Avg loss: 0.21160296, Global Avg Loss: 0.95448811, Time: 0.0210 Steps: 58460, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000987, Sample Num: 15792, Cur Loss: 0.19340713, Cur Avg Loss: 0.21553421, Log Avg loss: 0.20392575, Global Avg Loss: 0.95435975, Time: 0.0210 Steps: 58470, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000997, Sample Num: 15952, Cur Loss: 0.10494880, Cur Avg Loss: 0.21537321, Log Avg loss: 0.19948291, Global Avg Loss: 0.95423066, Time: 0.0211 Steps: 58480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001007, Sample Num: 16112, Cur Loss: 0.14729241, Cur Avg Loss: 0.21640422, Log Avg loss: 0.31919519, Global Avg Loss: 0.95412209, Time: 0.0211 Steps: 58490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001017, Sample Num: 16272, Cur Loss: 0.16599810, Cur Avg Loss: 0.21677335, Log Avg loss: 0.25394560, Global Avg Loss: 0.95400240, Time: 0.0210 Steps: 58500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001027, Sample Num: 16432, Cur Loss: 0.14061815, Cur Avg Loss: 0.21639598, Log Avg loss: 0.17801697, Global Avg Loss: 0.95386978, Time: 0.0249 Steps: 58510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001037, Sample Num: 16592, Cur Loss: 0.15778212, Cur Avg Loss: 0.21588813, Log Avg loss: 0.16373209, Global Avg Loss: 0.95373476, Time: 0.0212 Steps: 58520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001047, Sample Num: 16752, Cur Loss: 0.05828160, Cur Avg Loss: 0.21640490, Log Avg loss: 0.26999357, Global Avg Loss: 0.95361794, Time: 0.0211 Steps: 58530, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001057, Sample Num: 16912, Cur Loss: 0.11648335, Cur Avg Loss: 0.21615373, Log Avg loss: 0.18985615, Global Avg Loss: 0.95348747, Time: 0.0211 Steps: 58540, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001067, Sample Num: 17072, Cur Loss: 0.18339302, Cur Avg Loss: 0.21648194, Log Avg loss: 0.25117439, Global Avg Loss: 0.95336752, Time: 0.0211 Steps: 58550, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001077, Sample Num: 17232, Cur Loss: 0.46633109, Cur Avg Loss: 0.21617332, Log Avg loss: 0.18324360, Global Avg Loss: 0.95323601, Time: 0.0211 Steps: 58560, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001087, Sample Num: 17392, Cur Loss: 0.09113007, Cur Avg Loss: 0.21587892, Log Avg loss: 0.18417146, Global Avg Loss: 0.95310470, Time: 0.0211 Steps: 58570, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001097, Sample Num: 17552, Cur Loss: 0.17647876, Cur Avg Loss: 0.21652829, Log Avg loss: 0.28711458, Global Avg Loss: 0.95299102, Time: 0.0211 Steps: 58580, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001107, Sample Num: 17712, Cur Loss: 0.17298844, Cur Avg Loss: 0.21704738, Log Avg loss: 0.27399210, Global Avg Loss: 0.95287513, Time: 0.0211 Steps: 58590, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001117, Sample Num: 17872, Cur Loss: 0.15336646, Cur Avg Loss: 0.21688369, Log Avg loss: 0.19876317, Global Avg Loss: 0.95274644, Time: 0.0211 Steps: 58600, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001127, Sample Num: 18032, Cur Loss: 0.11389405, Cur Avg Loss: 0.21734502, Log Avg loss: 0.26887604, Global Avg Loss: 0.95262976, Time: 0.0211 Steps: 58610, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001137, Sample Num: 18192, Cur Loss: 0.17771068, Cur Avg Loss: 0.21719944, Log Avg loss: 0.20079194, Global Avg Loss: 0.95250150, Time: 0.0211 Steps: 58620, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001147, Sample Num: 18352, Cur Loss: 0.38206786, Cur Avg Loss: 0.21796198, Log Avg loss: 0.30466304, Global Avg Loss: 0.95239100, Time: 0.0211 Steps: 58630, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001157, Sample Num: 18512, Cur Loss: 1.00304163, Cur Avg Loss: 0.21893913, Log Avg loss: 0.33101791, Global Avg Loss: 0.95228504, Time: 0.0212 Steps: 58640, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001167, Sample Num: 18672, Cur Loss: 0.49784717, Cur Avg Loss: 0.21953078, Log Avg loss: 0.28798496, Global Avg Loss: 0.95217177, Time: 0.0211 Steps: 58650, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001177, Sample Num: 18832, Cur Loss: 0.11168948, Cur Avg Loss: 0.21909845, Log Avg loss: 0.16864512, Global Avg Loss: 0.95203820, Time: 0.0211 Steps: 58660, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001187, Sample Num: 18992, Cur Loss: 0.08136888, Cur Avg Loss: 0.21885825, Log Avg loss: 0.19058660, Global Avg Loss: 0.95190842, Time: 0.0211 Steps: 58670, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001197, Sample Num: 19152, Cur Loss: 0.19275720, Cur Avg Loss: 0.21849900, Log Avg loss: 0.17585596, Global Avg Loss: 0.95177617, Time: 0.0211 Steps: 58680, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001207, Sample Num: 19312, Cur Loss: 0.16508853, Cur Avg Loss: 0.21836631, Log Avg loss: 0.20248384, Global Avg Loss: 0.95164850, Time: 0.0211 Steps: 58690, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001217, Sample Num: 19472, Cur Loss: 0.35176468, Cur Avg Loss: 0.21808064, Log Avg loss: 0.18360047, Global Avg Loss: 0.95151765, Time: 0.0211 Steps: 58700, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001227, Sample Num: 19632, Cur Loss: 0.14138557, Cur Avg Loss: 0.21847133, Log Avg loss: 0.26601853, Global Avg Loss: 0.95140089, Time: 0.0211 Steps: 58710, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001237, Sample Num: 19792, Cur Loss: 0.46138293, Cur Avg Loss: 0.21882834, Log Avg loss: 0.26263337, Global Avg Loss: 0.95128360, Time: 0.0210 Steps: 58720, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001247, Sample Num: 19952, Cur Loss: 0.06740733, Cur Avg Loss: 0.21910543, Log Avg loss: 0.25338060, Global Avg Loss: 0.95116476, Time: 0.0211 Steps: 58730, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001257, Sample Num: 20112, Cur Loss: 0.29045063, Cur Avg Loss: 0.21948467, Log Avg loss: 0.26677649, Global Avg Loss: 0.95104825, Time: 0.0211 Steps: 58740, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001267, Sample Num: 20272, Cur Loss: 0.27251875, Cur Avg Loss: 0.21936299, Log Avg loss: 0.20406741, Global Avg Loss: 0.95092111, Time: 0.0211 Steps: 58750, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001277, Sample Num: 20432, Cur Loss: 0.08328509, Cur Avg Loss: 0.21947824, Log Avg loss: 0.23408137, Global Avg Loss: 0.95079911, Time: 0.0211 Steps: 58760, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001287, Sample Num: 20592, Cur Loss: 0.24831614, Cur Avg Loss: 0.21907755, Log Avg loss: 0.16790912, Global Avg Loss: 0.95066590, Time: 0.0213 Steps: 58770, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001297, Sample Num: 20752, Cur Loss: 0.09921087, Cur Avg Loss: 0.21876714, Log Avg loss: 0.17881764, Global Avg Loss: 0.95053459, Time: 0.0212 Steps: 58780, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001307, Sample Num: 20912, Cur Loss: 0.03244396, Cur Avg Loss: 0.21857782, Log Avg loss: 0.19402247, Global Avg Loss: 0.95040591, Time: 0.0212 Steps: 58790, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001317, Sample Num: 21072, Cur Loss: 0.10052419, Cur Avg Loss: 0.21859632, Log Avg loss: 0.22101462, Global Avg Loss: 0.95028186, Time: 0.0212 Steps: 58800, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001327, Sample Num: 21232, Cur Loss: 0.24255718, Cur Avg Loss: 0.21840091, Log Avg loss: 0.19266544, Global Avg Loss: 0.95015304, Time: 0.0212 Steps: 58810, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001337, Sample Num: 21392, Cur Loss: 0.18347019, Cur Avg Loss: 0.21802978, Log Avg loss: 0.16878084, Global Avg Loss: 0.95002020, Time: 0.0213 Steps: 58820, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001347, Sample Num: 21552, Cur Loss: 0.34501392, Cur Avg Loss: 0.21795638, Log Avg loss: 0.20814285, Global Avg Loss: 0.94989409, Time: 0.0212 Steps: 58830, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001357, Sample Num: 21712, Cur Loss: 0.22709250, Cur Avg Loss: 0.21773334, Log Avg loss: 0.18768998, Global Avg Loss: 0.94976455, Time: 0.0212 Steps: 58840, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001367, Sample Num: 21872, Cur Loss: 0.21715534, Cur Avg Loss: 0.21742899, Log Avg loss: 0.17612787, Global Avg Loss: 0.94963309, Time: 0.0212 Steps: 58850, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001377, Sample Num: 22032, Cur Loss: 0.21639860, Cur Avg Loss: 0.21748683, Log Avg loss: 0.22539445, Global Avg Loss: 0.94951005, Time: 0.0212 Steps: 58860, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001387, Sample Num: 22192, Cur Loss: 0.38148141, Cur Avg Loss: 0.21754990, Log Avg loss: 0.22623449, Global Avg Loss: 0.94938719, Time: 0.0212 Steps: 58870, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001397, Sample Num: 22352, Cur Loss: 0.42232740, Cur Avg Loss: 0.21785697, Log Avg loss: 0.26044655, Global Avg Loss: 0.94927018, Time: 0.0212 Steps: 58880, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001407, Sample Num: 22512, Cur Loss: 0.21585320, Cur Avg Loss: 0.21829833, Log Avg loss: 0.27995739, Global Avg Loss: 0.94915653, Time: 0.0212 Steps: 58890, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001417, Sample Num: 22672, Cur Loss: 0.13768756, Cur Avg Loss: 0.21877514, Log Avg loss: 0.28586144, Global Avg Loss: 0.94904391, Time: 0.0213 Steps: 58900, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001427, Sample Num: 22832, Cur Loss: 0.43317232, Cur Avg Loss: 0.21974173, Log Avg loss: 0.35670817, Global Avg Loss: 0.94894336, Time: 0.0212 Steps: 58910, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001437, Sample Num: 22992, Cur Loss: 0.29472893, Cur Avg Loss: 0.21938526, Log Avg loss: 0.16851678, Global Avg Loss: 0.94881091, Time: 0.0212 Steps: 58920, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001447, Sample Num: 23152, Cur Loss: 0.13390920, Cur Avg Loss: 0.21937618, Log Avg loss: 0.21807124, Global Avg Loss: 0.94868691, Time: 0.0214 Steps: 58930, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001457, Sample Num: 23312, Cur Loss: 0.24721164, Cur Avg Loss: 0.21964805, Log Avg loss: 0.25898800, Global Avg Loss: 0.94856989, Time: 0.0212 Steps: 58940, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001467, Sample Num: 23472, Cur Loss: 0.42510474, Cur Avg Loss: 0.21985259, Log Avg loss: 0.24965413, Global Avg Loss: 0.94845133, Time: 0.0212 Steps: 58950, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001477, Sample Num: 23632, Cur Loss: 0.64278972, Cur Avg Loss: 0.22018086, Log Avg loss: 0.26833837, Global Avg Loss: 0.94833598, Time: 0.0212 Steps: 58960, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001487, Sample Num: 23792, Cur Loss: 0.04958908, Cur Avg Loss: 0.21967107, Log Avg loss: 0.14437481, Global Avg Loss: 0.94819964, Time: 0.0212 Steps: 58970, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001497, Sample Num: 23952, Cur Loss: 0.26156655, Cur Avg Loss: 0.21932611, Log Avg loss: 0.16803045, Global Avg Loss: 0.94806737, Time: 0.0212 Steps: 58980, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001507, Sample Num: 24112, Cur Loss: 0.28128111, Cur Avg Loss: 0.21921536, Log Avg loss: 0.20263643, Global Avg Loss: 0.94794100, Time: 0.0212 Steps: 58990, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001517, Sample Num: 24272, Cur Loss: 0.40806028, Cur Avg Loss: 0.22023242, Log Avg loss: 0.37350215, Global Avg Loss: 0.94784364, Time: 0.0212 Steps: 59000, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001527, Sample Num: 24432, Cur Loss: 0.13455540, Cur Avg Loss: 0.22037897, Log Avg loss: 0.24261102, Global Avg Loss: 0.94772413, Time: 0.0212 Steps: 59010, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001537, Sample Num: 24592, Cur Loss: 0.16275245, Cur Avg Loss: 0.22018053, Log Avg loss: 0.18987924, Global Avg Loss: 0.94759572, Time: 0.0247 Steps: 59020, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001547, Sample Num: 24752, Cur Loss: 0.25440088, Cur Avg Loss: 0.22001148, Log Avg loss: 0.19402812, Global Avg Loss: 0.94746807, Time: 0.0210 Steps: 59030, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001557, Sample Num: 24912, Cur Loss: 0.23894839, Cur Avg Loss: 0.22027410, Log Avg loss: 0.26090124, Global Avg Loss: 0.94735178, Time: 0.0210 Steps: 59040, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001567, Sample Num: 25072, Cur Loss: 0.09384245, Cur Avg Loss: 0.22055091, Log Avg loss: 0.26364984, Global Avg Loss: 0.94723599, Time: 0.0210 Steps: 59050, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001577, Sample Num: 25232, Cur Loss: 0.11328312, Cur Avg Loss: 0.22039511, Log Avg loss: 0.19598202, Global Avg Loss: 0.94710879, Time: 0.0210 Steps: 59060, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001587, Sample Num: 25392, Cur Loss: 0.09245721, Cur Avg Loss: 0.21989701, Log Avg loss: 0.14134638, Global Avg Loss: 0.94697238, Time: 0.0210 Steps: 59070, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001597, Sample Num: 25552, Cur Loss: 0.20523590, Cur Avg Loss: 0.21950295, Log Avg loss: 0.15696511, Global Avg Loss: 0.94683867, Time: 0.0210 Steps: 59080, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001607, Sample Num: 25712, Cur Loss: 0.25221235, Cur Avg Loss: 0.21944485, Log Avg loss: 0.21016753, Global Avg Loss: 0.94671400, Time: 0.0209 Steps: 59090, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001617, Sample Num: 25872, Cur Loss: 0.70765829, Cur Avg Loss: 0.21928360, Log Avg loss: 0.19336995, Global Avg Loss: 0.94658653, Time: 0.0210 Steps: 59100, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001627, Sample Num: 26032, Cur Loss: 0.58309817, Cur Avg Loss: 0.21953534, Log Avg loss: 0.26024189, Global Avg Loss: 0.94647041, Time: 0.0210 Steps: 59110, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001637, Sample Num: 26192, Cur Loss: 0.09596179, Cur Avg Loss: 0.21923049, Log Avg loss: 0.16963111, Global Avg Loss: 0.94633901, Time: 0.0210 Steps: 59120, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001647, Sample Num: 26352, Cur Loss: 0.25725207, Cur Avg Loss: 0.21920278, Log Avg loss: 0.21466762, Global Avg Loss: 0.94621527, Time: 0.0210 Steps: 59130, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001657, Sample Num: 26512, Cur Loss: 0.12989423, Cur Avg Loss: 0.21965746, Log Avg loss: 0.29454245, Global Avg Loss: 0.94610508, Time: 0.0210 Steps: 59140, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001667, Sample Num: 26672, Cur Loss: 0.34445941, Cur Avg Loss: 0.21984258, Log Avg loss: 0.25051673, Global Avg Loss: 0.94598748, Time: 0.0209 Steps: 59150, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001677, Sample Num: 26832, Cur Loss: 0.31545994, Cur Avg Loss: 0.22000762, Log Avg loss: 0.24752091, Global Avg Loss: 0.94586942, Time: 0.0210 Steps: 59160, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001687, Sample Num: 26992, Cur Loss: 0.12184905, Cur Avg Loss: 0.22034969, Log Avg loss: 0.27771421, Global Avg Loss: 0.94575650, Time: 0.0210 Steps: 59170, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001697, Sample Num: 27152, Cur Loss: 0.17603269, Cur Avg Loss: 0.22045506, Log Avg loss: 0.23823005, Global Avg Loss: 0.94563694, Time: 0.0210 Steps: 59180, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001707, Sample Num: 27312, Cur Loss: 0.16920930, Cur Avg Loss: 0.22040825, Log Avg loss: 0.21246492, Global Avg Loss: 0.94551308, Time: 0.0210 Steps: 59190, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001717, Sample Num: 27472, Cur Loss: 0.10831973, Cur Avg Loss: 0.22011107, Log Avg loss: 0.16938282, Global Avg Loss: 0.94538197, Time: 0.0210 Steps: 59200, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001727, Sample Num: 27632, Cur Loss: 0.07908248, Cur Avg Loss: 0.21980527, Log Avg loss: 0.16730011, Global Avg Loss: 0.94525056, Time: 0.0210 Steps: 59210, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001737, Sample Num: 27792, Cur Loss: 0.20003910, Cur Avg Loss: 0.21965112, Log Avg loss: 0.19302864, Global Avg Loss: 0.94512354, Time: 0.0210 Steps: 59220, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001747, Sample Num: 27952, Cur Loss: 0.23976192, Cur Avg Loss: 0.21939876, Log Avg loss: 0.17556435, Global Avg Loss: 0.94499361, Time: 0.0209 Steps: 59230, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001757, Sample Num: 28112, Cur Loss: 0.17345014, Cur Avg Loss: 0.21918363, Log Avg loss: 0.18159979, Global Avg Loss: 0.94486475, Time: 0.0210 Steps: 59240, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001767, Sample Num: 28272, Cur Loss: 0.10958545, Cur Avg Loss: 0.21914727, Log Avg loss: 0.21275832, Global Avg Loss: 0.94474119, Time: 0.0210 Steps: 59250, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001777, Sample Num: 28432, Cur Loss: 0.31292182, Cur Avg Loss: 0.21878218, Log Avg loss: 0.15427190, Global Avg Loss: 0.94460780, Time: 0.0210 Steps: 59260, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001787, Sample Num: 28592, Cur Loss: 0.14799643, Cur Avg Loss: 0.21869285, Log Avg loss: 0.20281790, Global Avg Loss: 0.94448264, Time: 0.0210 Steps: 59270, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001797, Sample Num: 28752, Cur Loss: 0.15308315, Cur Avg Loss: 0.21923626, Log Avg loss: 0.31634462, Global Avg Loss: 0.94437668, Time: 0.0247 Steps: 59280, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001807, Sample Num: 28912, Cur Loss: 0.13786945, Cur Avg Loss: 0.21913131, Log Avg loss: 0.20027069, Global Avg Loss: 0.94425118, Time: 0.0210 Steps: 59290, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001817, Sample Num: 29072, Cur Loss: 0.26959872, Cur Avg Loss: 0.21955578, Log Avg loss: 0.29625842, Global Avg Loss: 0.94414191, Time: 0.0209 Steps: 59300, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001827, Sample Num: 29232, Cur Loss: 0.32186615, Cur Avg Loss: 0.21953097, Log Avg loss: 0.21502207, Global Avg Loss: 0.94401897, Time: 0.0210 Steps: 59310, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001837, Sample Num: 29392, Cur Loss: 0.17046219, Cur Avg Loss: 0.21965139, Log Avg loss: 0.24165289, Global Avg Loss: 0.94390057, Time: 0.0210 Steps: 59320, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001847, Sample Num: 29552, Cur Loss: 0.18503599, Cur Avg Loss: 0.21966248, Log Avg loss: 0.22170056, Global Avg Loss: 0.94377884, Time: 0.0209 Steps: 59330, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001857, Sample Num: 29712, Cur Loss: 0.14987835, Cur Avg Loss: 0.21972705, Log Avg loss: 0.23165179, Global Avg Loss: 0.94365884, Time: 0.0209 Steps: 59340, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001867, Sample Num: 29872, Cur Loss: 0.06664722, Cur Avg Loss: 0.21964305, Log Avg loss: 0.20404518, Global Avg Loss: 0.94353422, Time: 0.0210 Steps: 59350, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001877, Sample Num: 30032, Cur Loss: 0.38340628, Cur Avg Loss: 0.21954171, Log Avg loss: 0.20062040, Global Avg Loss: 0.94340906, Time: 0.0210 Steps: 59360, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001887, Sample Num: 30192, Cur Loss: 0.21975364, Cur Avg Loss: 0.21948303, Log Avg loss: 0.20846915, Global Avg Loss: 0.94328527, Time: 0.0210 Steps: 59370, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001897, Sample Num: 30352, Cur Loss: 0.21987548, Cur Avg Loss: 0.21913337, Log Avg loss: 0.15315209, Global Avg Loss: 0.94315221, Time: 0.0209 Steps: 59380, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001907, Sample Num: 30512, Cur Loss: 0.19757691, Cur Avg Loss: 0.21935823, Log Avg loss: 0.26201422, Global Avg Loss: 0.94303752, Time: 0.0209 Steps: 59390, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001917, Sample Num: 30672, Cur Loss: 0.29303950, Cur Avg Loss: 0.21966653, Log Avg loss: 0.27846059, Global Avg Loss: 0.94292564, Time: 0.0210 Steps: 59400, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001927, Sample Num: 30832, Cur Loss: 0.25331092, Cur Avg Loss: 0.21950799, Log Avg loss: 0.18911522, Global Avg Loss: 0.94279876, Time: 0.0209 Steps: 59410, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001937, Sample Num: 30992, Cur Loss: 0.24255177, Cur Avg Loss: 0.21992652, Log Avg loss: 0.30057803, Global Avg Loss: 0.94269067, Time: 0.0209 Steps: 59420, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001947, Sample Num: 31152, Cur Loss: 0.25904036, Cur Avg Loss: 0.22007478, Log Avg loss: 0.24879142, Global Avg Loss: 0.94257391, Time: 0.0210 Steps: 59430, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001957, Sample Num: 31312, Cur Loss: 0.17548221, Cur Avg Loss: 0.21996965, Log Avg loss: 0.19950064, Global Avg Loss: 0.94244890, Time: 0.0210 Steps: 59440, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001967, Sample Num: 31472, Cur Loss: 0.10033325, Cur Avg Loss: 0.21963426, Log Avg loss: 0.15399977, Global Avg Loss: 0.94231628, Time: 0.0210 Steps: 59450, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001977, Sample Num: 31632, Cur Loss: 0.22241428, Cur Avg Loss: 0.21941887, Log Avg loss: 0.17705028, Global Avg Loss: 0.94218758, Time: 0.0210 Steps: 59460, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001987, Sample Num: 31792, Cur Loss: 0.29964370, Cur Avg Loss: 0.21951150, Log Avg loss: 0.23782630, Global Avg Loss: 0.94206914, Time: 0.0210 Steps: 59470, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001997, Sample Num: 31952, Cur Loss: 0.52739632, Cur Avg Loss: 0.21986710, Log Avg loss: 0.29052360, Global Avg Loss: 0.94195960, Time: 0.0209 Steps: 59480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002007, Sample Num: 32112, Cur Loss: 0.94038868, Cur Avg Loss: 0.22041959, Log Avg loss: 0.33075197, Global Avg Loss: 0.94185685, Time: 0.0210 Steps: 59490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002017, Sample Num: 32272, Cur Loss: 0.36028388, Cur Avg Loss: 0.22070115, Log Avg loss: 0.27721006, Global Avg Loss: 0.94174515, Time: 0.0210 Steps: 59500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002027, Sample Num: 32432, Cur Loss: 0.19148184, Cur Avg Loss: 0.22093899, Log Avg loss: 0.26891182, Global Avg Loss: 0.94163209, Time: 0.0210 Steps: 59510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002037, Sample Num: 32592, Cur Loss: 0.26771098, Cur Avg Loss: 0.22093267, Log Avg loss: 0.21965132, Global Avg Loss: 0.94151079, Time: 0.0209 Steps: 59520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002047, Sample Num: 32752, Cur Loss: 0.19766848, Cur Avg Loss: 0.22104225, Log Avg loss: 0.24336457, Global Avg Loss: 0.94139351, Time: 0.0210 Steps: 59530, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002057, Sample Num: 32912, Cur Loss: 0.28598887, Cur Avg Loss: 0.22088656, Log Avg loss: 0.18901502, Global Avg Loss: 0.94126714, Time: 0.0211 Steps: 59540, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002067, Sample Num: 33072, Cur Loss: 0.32382086, Cur Avg Loss: 0.22091568, Log Avg loss: 0.22690645, Global Avg Loss: 0.94114718, Time: 0.0210 Steps: 59550, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002077, Sample Num: 33232, Cur Loss: 0.13600415, Cur Avg Loss: 0.22085993, Log Avg loss: 0.20933695, Global Avg Loss: 0.94102432, Time: 0.0210 Steps: 59560, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002087, Sample Num: 33392, Cur Loss: 0.57302231, Cur Avg Loss: 0.22070207, Log Avg loss: 0.18791381, Global Avg Loss: 0.94089789, Time: 0.0210 Steps: 59570, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002097, Sample Num: 33552, Cur Loss: 0.18392590, Cur Avg Loss: 0.22044204, Log Avg loss: 0.16617472, Global Avg Loss: 0.94076786, Time: 0.0210 Steps: 59580, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002107, Sample Num: 33712, Cur Loss: 0.35539579, Cur Avg Loss: 0.22057644, Log Avg loss: 0.24875968, Global Avg Loss: 0.94065173, Time: 0.0210 Steps: 59590, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002117, Sample Num: 33872, Cur Loss: 0.34341595, Cur Avg Loss: 0.22049693, Log Avg loss: 0.20374502, Global Avg Loss: 0.94052809, Time: 0.0210 Steps: 59600, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002127, Sample Num: 34032, Cur Loss: 0.26568109, Cur Avg Loss: 0.22026500, Log Avg loss: 0.17116465, Global Avg Loss: 0.94039902, Time: 0.0210 Steps: 59610, Updated lr: 0.000044 ***** Running evaluation checkpoint-59612 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-59612 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.901667, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.283651, "eval_total_loss": 199.406527, "eval_mae": 0.358394, "eval_mse": 0.283762, "eval_r2": 0.819622, "eval_sp_statistic": 0.884926, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.913188, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.877287, "test_total_loss": 440.398229, "test_mae": 0.733362, "test_mse": 0.877398, "test_r2": 0.43372, "test_sp_statistic": 0.795905, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.82818, "test_ps_pvalue": 0.0, "lr": 4.441725936462779e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9403729323608451, "train_cur_epoch_loss": 468.8290944490582, "train_cur_epoch_avg_loss": 0.22021094149791365, "train_cur_epoch_time": 44.90166687965393, "train_cur_epoch_avg_time": 0.02109049642069231, "epoch": 28, "step": 59612} ################################################## Training, Epoch: 0029, Batch: 000008, Sample Num: 128, Cur Loss: 0.14997673, Cur Avg Loss: 0.17985736, Log Avg loss: 0.17642965, Global Avg Loss: 0.94027088, Time: 0.0212 Steps: 59620, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000018, Sample Num: 288, Cur Loss: 0.18782066, Cur Avg Loss: 0.25847153, Log Avg loss: 0.32136287, Global Avg Loss: 0.94016709, Time: 0.0208 Steps: 59630, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000028, Sample Num: 448, Cur Loss: 0.14316714, Cur Avg Loss: 0.23671232, Log Avg loss: 0.19754574, Global Avg Loss: 0.94004258, Time: 0.0208 Steps: 59640, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000038, Sample Num: 608, Cur Loss: 0.17983815, Cur Avg Loss: 0.23155570, Log Avg loss: 0.21711717, Global Avg Loss: 0.93992138, Time: 0.0208 Steps: 59650, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000048, Sample Num: 768, Cur Loss: 0.11588836, Cur Avg Loss: 0.22537882, Log Avg loss: 0.20190665, Global Avg Loss: 0.93979768, Time: 0.0208 Steps: 59660, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000058, Sample Num: 928, Cur Loss: 0.52158225, Cur Avg Loss: 0.23049915, Log Avg loss: 0.25507677, Global Avg Loss: 0.93968293, Time: 0.0209 Steps: 59670, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000068, Sample Num: 1088, Cur Loss: 0.05095506, Cur Avg Loss: 0.21888972, Log Avg loss: 0.15155503, Global Avg Loss: 0.93955087, Time: 0.0209 Steps: 59680, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000078, Sample Num: 1248, Cur Loss: 0.19406876, Cur Avg Loss: 0.21536001, Log Avg loss: 0.19135794, Global Avg Loss: 0.93942552, Time: 0.0208 Steps: 59690, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000088, Sample Num: 1408, Cur Loss: 0.10107773, Cur Avg Loss: 0.21333183, Log Avg loss: 0.19751204, Global Avg Loss: 0.93930125, Time: 0.0209 Steps: 59700, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000098, Sample Num: 1568, Cur Loss: 0.36482194, Cur Avg Loss: 0.21822088, Log Avg loss: 0.26124450, Global Avg Loss: 0.93918769, Time: 0.0209 Steps: 59710, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000108, Sample Num: 1728, Cur Loss: 0.08689663, Cur Avg Loss: 0.21677419, Log Avg loss: 0.20259660, Global Avg Loss: 0.93906435, Time: 0.0208 Steps: 59720, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000118, Sample Num: 1888, Cur Loss: 0.25540680, Cur Avg Loss: 0.21398672, Log Avg loss: 0.18388213, Global Avg Loss: 0.93893792, Time: 0.0208 Steps: 59730, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000128, Sample Num: 2048, Cur Loss: 0.16754651, Cur Avg Loss: 0.21423816, Log Avg loss: 0.21720513, Global Avg Loss: 0.93881710, Time: 0.0208 Steps: 59740, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000138, Sample Num: 2208, Cur Loss: 0.20051762, Cur Avg Loss: 0.20967102, Log Avg loss: 0.15121156, Global Avg Loss: 0.93868529, Time: 0.0208 Steps: 59750, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000148, Sample Num: 2368, Cur Loss: 0.44090897, Cur Avg Loss: 0.21721916, Log Avg loss: 0.32138357, Global Avg Loss: 0.93858199, Time: 0.0208 Steps: 59760, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000158, Sample Num: 2528, Cur Loss: 0.10521418, Cur Avg Loss: 0.21764695, Log Avg loss: 0.22397820, Global Avg Loss: 0.93846243, Time: 0.0209 Steps: 59770, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000168, Sample Num: 2688, Cur Loss: 0.18061894, Cur Avg Loss: 0.21920908, Log Avg loss: 0.24389067, Global Avg Loss: 0.93834624, Time: 0.0208 Steps: 59780, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000178, Sample Num: 2848, Cur Loss: 0.29569069, Cur Avg Loss: 0.21896568, Log Avg loss: 0.21487670, Global Avg Loss: 0.93822524, Time: 0.0208 Steps: 59790, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000188, Sample Num: 3008, Cur Loss: 0.19185808, Cur Avg Loss: 0.21981443, Log Avg loss: 0.23492211, Global Avg Loss: 0.93810763, Time: 0.0208 Steps: 59800, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000198, Sample Num: 3168, Cur Loss: 0.41912228, Cur Avg Loss: 0.22148847, Log Avg loss: 0.25296042, Global Avg Loss: 0.93799308, Time: 0.0208 Steps: 59810, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000208, Sample Num: 3328, Cur Loss: 0.26758870, Cur Avg Loss: 0.22097353, Log Avg loss: 0.21077781, Global Avg Loss: 0.93787151, Time: 0.0208 Steps: 59820, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000218, Sample Num: 3488, Cur Loss: 0.13663062, Cur Avg Loss: 0.22183102, Log Avg loss: 0.23966680, Global Avg Loss: 0.93775481, Time: 0.0208 Steps: 59830, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000228, Sample Num: 3648, Cur Loss: 0.22148320, Cur Avg Loss: 0.21923269, Log Avg loss: 0.16258892, Global Avg Loss: 0.93762527, Time: 0.0208 Steps: 59840, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000238, Sample Num: 3808, Cur Loss: 0.06652470, Cur Avg Loss: 0.21558567, Log Avg loss: 0.13243362, Global Avg Loss: 0.93749074, Time: 0.0208 Steps: 59850, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000248, Sample Num: 3968, Cur Loss: 0.10072291, Cur Avg Loss: 0.21358402, Log Avg loss: 0.16594474, Global Avg Loss: 0.93736185, Time: 0.0208 Steps: 59860, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000258, Sample Num: 4128, Cur Loss: 0.15793341, Cur Avg Loss: 0.21140569, Log Avg loss: 0.15738321, Global Avg Loss: 0.93723157, Time: 0.0248 Steps: 59870, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000268, Sample Num: 4288, Cur Loss: 0.24514413, Cur Avg Loss: 0.21146864, Log Avg loss: 0.21309265, Global Avg Loss: 0.93711064, Time: 0.0210 Steps: 59880, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000278, Sample Num: 4448, Cur Loss: 0.33601457, Cur Avg Loss: 0.21059985, Log Avg loss: 0.18731629, Global Avg Loss: 0.93698544, Time: 0.0210 Steps: 59890, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000288, Sample Num: 4608, Cur Loss: 0.14204724, Cur Avg Loss: 0.21263587, Log Avg loss: 0.26923739, Global Avg Loss: 0.93687396, Time: 0.0209 Steps: 59900, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000298, Sample Num: 4768, Cur Loss: 0.03688179, Cur Avg Loss: 0.21244462, Log Avg loss: 0.20693668, Global Avg Loss: 0.93675212, Time: 0.0210 Steps: 59910, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000308, Sample Num: 4928, Cur Loss: 0.03043730, Cur Avg Loss: 0.21049364, Log Avg loss: 0.15235435, Global Avg Loss: 0.93662122, Time: 0.0210 Steps: 59920, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000318, Sample Num: 5088, Cur Loss: 0.29841688, Cur Avg Loss: 0.20965790, Log Avg loss: 0.18391717, Global Avg Loss: 0.93649562, Time: 0.0210 Steps: 59930, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000328, Sample Num: 5248, Cur Loss: 0.05179799, Cur Avg Loss: 0.20891520, Log Avg loss: 0.18529708, Global Avg Loss: 0.93637029, Time: 0.0209 Steps: 59940, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000338, Sample Num: 5408, Cur Loss: 0.18120989, Cur Avg Loss: 0.20851343, Log Avg loss: 0.19533547, Global Avg Loss: 0.93624669, Time: 0.0210 Steps: 59950, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000348, Sample Num: 5568, Cur Loss: 0.43622428, Cur Avg Loss: 0.20885570, Log Avg loss: 0.22042465, Global Avg Loss: 0.93612730, Time: 0.0211 Steps: 59960, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000358, Sample Num: 5728, Cur Loss: 0.53746176, Cur Avg Loss: 0.21030520, Log Avg loss: 0.26074749, Global Avg Loss: 0.93601468, Time: 0.0210 Steps: 59970, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000368, Sample Num: 5888, Cur Loss: 0.45868614, Cur Avg Loss: 0.21081538, Log Avg loss: 0.22907994, Global Avg Loss: 0.93589682, Time: 0.0210 Steps: 59980, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000378, Sample Num: 6048, Cur Loss: 0.11904915, Cur Avg Loss: 0.21009903, Log Avg loss: 0.18373741, Global Avg Loss: 0.93577144, Time: 0.0209 Steps: 59990, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000388, Sample Num: 6208, Cur Loss: 0.23918605, Cur Avg Loss: 0.21047015, Log Avg loss: 0.22449850, Global Avg Loss: 0.93565289, Time: 0.0210 Steps: 60000, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000398, Sample Num: 6368, Cur Loss: 0.30883810, Cur Avg Loss: 0.21004889, Log Avg loss: 0.19370390, Global Avg Loss: 0.93552926, Time: 0.0209 Steps: 60010, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000408, Sample Num: 6528, Cur Loss: 0.34005103, Cur Avg Loss: 0.20928700, Log Avg loss: 0.17896381, Global Avg Loss: 0.93540320, Time: 0.0209 Steps: 60020, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000418, Sample Num: 6688, Cur Loss: 0.19977160, Cur Avg Loss: 0.20864431, Log Avg loss: 0.18242260, Global Avg Loss: 0.93527777, Time: 0.0209 Steps: 60030, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000428, Sample Num: 6848, Cur Loss: 0.29510424, Cur Avg Loss: 0.20833439, Log Avg loss: 0.19537973, Global Avg Loss: 0.93515454, Time: 0.0210 Steps: 60040, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000438, Sample Num: 7008, Cur Loss: 0.14877400, Cur Avg Loss: 0.20882947, Log Avg loss: 0.23001888, Global Avg Loss: 0.93503711, Time: 0.0210 Steps: 60050, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000448, Sample Num: 7168, Cur Loss: 0.15988825, Cur Avg Loss: 0.20779945, Log Avg loss: 0.16268454, Global Avg Loss: 0.93490851, Time: 0.0210 Steps: 60060, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000458, Sample Num: 7328, Cur Loss: 0.25160986, Cur Avg Loss: 0.20773512, Log Avg loss: 0.20485329, Global Avg Loss: 0.93478698, Time: 0.0209 Steps: 60070, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000468, Sample Num: 7488, Cur Loss: 0.37745306, Cur Avg Loss: 0.20900379, Log Avg loss: 0.26710861, Global Avg Loss: 0.93467585, Time: 0.0210 Steps: 60080, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000478, Sample Num: 7648, Cur Loss: 0.21093448, Cur Avg Loss: 0.20792909, Log Avg loss: 0.15763313, Global Avg Loss: 0.93454654, Time: 0.0211 Steps: 60090, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000488, Sample Num: 7808, Cur Loss: 0.12957296, Cur Avg Loss: 0.20780959, Log Avg loss: 0.20209758, Global Avg Loss: 0.93442466, Time: 0.0209 Steps: 60100, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000498, Sample Num: 7968, Cur Loss: 0.16036709, Cur Avg Loss: 0.20678245, Log Avg loss: 0.15665828, Global Avg Loss: 0.93429527, Time: 0.0210 Steps: 60110, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000508, Sample Num: 8128, Cur Loss: 0.13849336, Cur Avg Loss: 0.20651700, Log Avg loss: 0.19329714, Global Avg Loss: 0.93417202, Time: 0.0210 Steps: 60120, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000518, Sample Num: 8288, Cur Loss: 0.35148418, Cur Avg Loss: 0.20649428, Log Avg loss: 0.20534029, Global Avg Loss: 0.93405081, Time: 0.0228 Steps: 60130, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000528, Sample Num: 8448, Cur Loss: 0.20160738, Cur Avg Loss: 0.20599416, Log Avg loss: 0.18008815, Global Avg Loss: 0.93392544, Time: 0.0209 Steps: 60140, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000538, Sample Num: 8608, Cur Loss: 0.27931350, Cur Avg Loss: 0.20634615, Log Avg loss: 0.22493109, Global Avg Loss: 0.93380757, Time: 0.0209 Steps: 60150, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000548, Sample Num: 8768, Cur Loss: 0.42863375, Cur Avg Loss: 0.20730483, Log Avg loss: 0.25888185, Global Avg Loss: 0.93369538, Time: 0.0209 Steps: 60160, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000558, Sample Num: 8928, Cur Loss: 0.14558224, Cur Avg Loss: 0.20669526, Log Avg loss: 0.17329053, Global Avg Loss: 0.93356901, Time: 0.0209 Steps: 60170, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000568, Sample Num: 9088, Cur Loss: 0.19509184, Cur Avg Loss: 0.20563604, Log Avg loss: 0.14653199, Global Avg Loss: 0.93343823, Time: 0.0209 Steps: 60180, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000578, Sample Num: 9248, Cur Loss: 0.15484910, Cur Avg Loss: 0.20603296, Log Avg loss: 0.22857766, Global Avg Loss: 0.93332112, Time: 0.0209 Steps: 60190, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000588, Sample Num: 9408, Cur Loss: 0.10421710, Cur Avg Loss: 0.20552177, Log Avg loss: 0.17597536, Global Avg Loss: 0.93319532, Time: 0.0209 Steps: 60200, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000598, Sample Num: 9568, Cur Loss: 0.16261342, Cur Avg Loss: 0.20541000, Log Avg loss: 0.19883754, Global Avg Loss: 0.93307335, Time: 0.0209 Steps: 60210, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000608, Sample Num: 9728, Cur Loss: 0.11610726, Cur Avg Loss: 0.20459938, Log Avg loss: 0.15612424, Global Avg Loss: 0.93294433, Time: 0.0209 Steps: 60220, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000618, Sample Num: 9888, Cur Loss: 0.21198951, Cur Avg Loss: 0.20563025, Log Avg loss: 0.26830739, Global Avg Loss: 0.93283398, Time: 0.0209 Steps: 60230, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000628, Sample Num: 10048, Cur Loss: 0.22429439, Cur Avg Loss: 0.20518355, Log Avg loss: 0.17757764, Global Avg Loss: 0.93270861, Time: 0.0209 Steps: 60240, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000638, Sample Num: 10208, Cur Loss: 0.27120644, Cur Avg Loss: 0.20535694, Log Avg loss: 0.21624596, Global Avg Loss: 0.93258969, Time: 0.0209 Steps: 60250, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000648, Sample Num: 10368, Cur Loss: 0.09645347, Cur Avg Loss: 0.20477555, Log Avg loss: 0.16768287, Global Avg Loss: 0.93246276, Time: 0.0209 Steps: 60260, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000658, Sample Num: 10528, Cur Loss: 0.29569241, Cur Avg Loss: 0.20503099, Log Avg loss: 0.22158323, Global Avg Loss: 0.93234481, Time: 0.0209 Steps: 60270, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000668, Sample Num: 10688, Cur Loss: 0.26179180, Cur Avg Loss: 0.20536720, Log Avg loss: 0.22748994, Global Avg Loss: 0.93222788, Time: 0.0209 Steps: 60280, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000678, Sample Num: 10848, Cur Loss: 0.12833169, Cur Avg Loss: 0.20607856, Log Avg loss: 0.25359747, Global Avg Loss: 0.93211532, Time: 0.0209 Steps: 60290, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000688, Sample Num: 11008, Cur Loss: 0.36207944, Cur Avg Loss: 0.20794094, Log Avg loss: 0.33420983, Global Avg Loss: 0.93201616, Time: 0.0209 Steps: 60300, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000698, Sample Num: 11168, Cur Loss: 0.05538578, Cur Avg Loss: 0.20727704, Log Avg loss: 0.16160095, Global Avg Loss: 0.93188842, Time: 0.0209 Steps: 60310, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000708, Sample Num: 11328, Cur Loss: 0.04318553, Cur Avg Loss: 0.20686280, Log Avg loss: 0.17794875, Global Avg Loss: 0.93176343, Time: 0.0209 Steps: 60320, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000718, Sample Num: 11488, Cur Loss: 0.11325283, Cur Avg Loss: 0.20765574, Log Avg loss: 0.26379578, Global Avg Loss: 0.93165271, Time: 0.0209 Steps: 60330, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000728, Sample Num: 11648, Cur Loss: 0.15362617, Cur Avg Loss: 0.20743171, Log Avg loss: 0.19134631, Global Avg Loss: 0.93153002, Time: 0.0209 Steps: 60340, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000738, Sample Num: 11808, Cur Loss: 0.12149088, Cur Avg Loss: 0.20729841, Log Avg loss: 0.19759413, Global Avg Loss: 0.93140841, Time: 0.0209 Steps: 60350, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000748, Sample Num: 11968, Cur Loss: 0.11439048, Cur Avg Loss: 0.20696053, Log Avg loss: 0.18202525, Global Avg Loss: 0.93128426, Time: 0.0209 Steps: 60360, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000758, Sample Num: 12128, Cur Loss: 0.11086141, Cur Avg Loss: 0.20677465, Log Avg loss: 0.19287075, Global Avg Loss: 0.93116194, Time: 0.0209 Steps: 60370, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000768, Sample Num: 12288, Cur Loss: 0.29172578, Cur Avg Loss: 0.20672264, Log Avg loss: 0.20278055, Global Avg Loss: 0.93104131, Time: 0.0254 Steps: 60380, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000778, Sample Num: 12448, Cur Loss: 0.08608332, Cur Avg Loss: 0.20606661, Log Avg loss: 0.15568330, Global Avg Loss: 0.93091292, Time: 0.0210 Steps: 60390, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000788, Sample Num: 12608, Cur Loss: 0.24192044, Cur Avg Loss: 0.20610575, Log Avg loss: 0.20915130, Global Avg Loss: 0.93079342, Time: 0.0210 Steps: 60400, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000798, Sample Num: 12768, Cur Loss: 0.23844168, Cur Avg Loss: 0.20636142, Log Avg loss: 0.22650758, Global Avg Loss: 0.93067684, Time: 0.0209 Steps: 60410, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000808, Sample Num: 12928, Cur Loss: 0.82110012, Cur Avg Loss: 0.20683594, Log Avg loss: 0.24470255, Global Avg Loss: 0.93056330, Time: 0.0209 Steps: 60420, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000818, Sample Num: 13088, Cur Loss: 0.10408752, Cur Avg Loss: 0.20672401, Log Avg loss: 0.19768040, Global Avg Loss: 0.93044202, Time: 0.0210 Steps: 60430, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000828, Sample Num: 13248, Cur Loss: 0.50618124, Cur Avg Loss: 0.20700336, Log Avg loss: 0.22985434, Global Avg Loss: 0.93032611, Time: 0.0209 Steps: 60440, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000838, Sample Num: 13408, Cur Loss: 0.42657232, Cur Avg Loss: 0.20718982, Log Avg loss: 0.22262848, Global Avg Loss: 0.93020904, Time: 0.0209 Steps: 60450, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000848, Sample Num: 13568, Cur Loss: 0.26230514, Cur Avg Loss: 0.20710417, Log Avg loss: 0.19992703, Global Avg Loss: 0.93008825, Time: 0.0210 Steps: 60460, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000858, Sample Num: 13728, Cur Loss: 0.17288575, Cur Avg Loss: 0.20697213, Log Avg loss: 0.19577512, Global Avg Loss: 0.92996682, Time: 0.0209 Steps: 60470, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000868, Sample Num: 13888, Cur Loss: 0.08867600, Cur Avg Loss: 0.20663841, Log Avg loss: 0.17800469, Global Avg Loss: 0.92984248, Time: 0.0211 Steps: 60480, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000878, Sample Num: 14048, Cur Loss: 0.31727973, Cur Avg Loss: 0.20683086, Log Avg loss: 0.22353591, Global Avg Loss: 0.92972572, Time: 0.0210 Steps: 60490, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000888, Sample Num: 14208, Cur Loss: 0.24014992, Cur Avg Loss: 0.20766866, Log Avg loss: 0.28122789, Global Avg Loss: 0.92961853, Time: 0.0210 Steps: 60500, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000898, Sample Num: 14368, Cur Loss: 0.18292958, Cur Avg Loss: 0.20742944, Log Avg loss: 0.18618601, Global Avg Loss: 0.92949567, Time: 0.0209 Steps: 60510, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000908, Sample Num: 14528, Cur Loss: 0.10237330, Cur Avg Loss: 0.20783417, Log Avg loss: 0.24417908, Global Avg Loss: 0.92938243, Time: 0.0210 Steps: 60520, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000918, Sample Num: 14688, Cur Loss: 0.17536709, Cur Avg Loss: 0.20766109, Log Avg loss: 0.19194521, Global Avg Loss: 0.92926060, Time: 0.0210 Steps: 60530, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000928, Sample Num: 14848, Cur Loss: 0.25181893, Cur Avg Loss: 0.20863166, Log Avg loss: 0.29772992, Global Avg Loss: 0.92915628, Time: 0.0209 Steps: 60540, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000938, Sample Num: 15008, Cur Loss: 0.52383614, Cur Avg Loss: 0.20919000, Log Avg loss: 0.26100447, Global Avg Loss: 0.92904594, Time: 0.0209 Steps: 60550, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000948, Sample Num: 15168, Cur Loss: 0.14666171, Cur Avg Loss: 0.20881334, Log Avg loss: 0.17348264, Global Avg Loss: 0.92892117, Time: 0.0209 Steps: 60560, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000958, Sample Num: 15328, Cur Loss: 0.23928851, Cur Avg Loss: 0.20844575, Log Avg loss: 0.17359819, Global Avg Loss: 0.92879647, Time: 0.0209 Steps: 60570, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000968, Sample Num: 15488, Cur Loss: 0.43650904, Cur Avg Loss: 0.20852701, Log Avg loss: 0.21631124, Global Avg Loss: 0.92867886, Time: 0.0209 Steps: 60580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000978, Sample Num: 15648, Cur Loss: 0.12663186, Cur Avg Loss: 0.20819041, Log Avg loss: 0.17560768, Global Avg Loss: 0.92855457, Time: 0.0209 Steps: 60590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000988, Sample Num: 15808, Cur Loss: 0.30378965, Cur Avg Loss: 0.20838403, Log Avg loss: 0.22731994, Global Avg Loss: 0.92843886, Time: 0.0209 Steps: 60600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000998, Sample Num: 15968, Cur Loss: 0.09956988, Cur Avg Loss: 0.20783647, Log Avg loss: 0.15373781, Global Avg Loss: 0.92831104, Time: 0.0210 Steps: 60610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001008, Sample Num: 16128, Cur Loss: 0.11091100, Cur Avg Loss: 0.20751577, Log Avg loss: 0.17551043, Global Avg Loss: 0.92818685, Time: 0.0210 Steps: 60620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001018, Sample Num: 16288, Cur Loss: 0.26289481, Cur Avg Loss: 0.20769057, Log Avg loss: 0.22531027, Global Avg Loss: 0.92807093, Time: 0.0209 Steps: 60630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001028, Sample Num: 16448, Cur Loss: 0.43810460, Cur Avg Loss: 0.20748914, Log Avg loss: 0.18698293, Global Avg Loss: 0.92794871, Time: 0.0210 Steps: 60640, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001038, Sample Num: 16608, Cur Loss: 0.16477433, Cur Avg Loss: 0.20711548, Log Avg loss: 0.16870337, Global Avg Loss: 0.92782353, Time: 0.0210 Steps: 60650, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001048, Sample Num: 16768, Cur Loss: 0.14063713, Cur Avg Loss: 0.20790074, Log Avg loss: 0.28941118, Global Avg Loss: 0.92771829, Time: 0.0211 Steps: 60660, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001058, Sample Num: 16928, Cur Loss: 0.09499251, Cur Avg Loss: 0.20788874, Log Avg loss: 0.20663076, Global Avg Loss: 0.92759943, Time: 0.0211 Steps: 60670, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001068, Sample Num: 17088, Cur Loss: 0.37907630, Cur Avg Loss: 0.20791696, Log Avg loss: 0.21090222, Global Avg Loss: 0.92748132, Time: 0.0211 Steps: 60680, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001078, Sample Num: 17248, Cur Loss: 0.06794203, Cur Avg Loss: 0.20795603, Log Avg loss: 0.21212952, Global Avg Loss: 0.92736345, Time: 0.0211 Steps: 60690, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001088, Sample Num: 17408, Cur Loss: 0.12544262, Cur Avg Loss: 0.20815759, Log Avg loss: 0.22988527, Global Avg Loss: 0.92724855, Time: 0.0211 Steps: 60700, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001098, Sample Num: 17568, Cur Loss: 0.18853281, Cur Avg Loss: 0.20795633, Log Avg loss: 0.18605918, Global Avg Loss: 0.92712646, Time: 0.0213 Steps: 60710, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001108, Sample Num: 17728, Cur Loss: 0.08642650, Cur Avg Loss: 0.20752011, Log Avg loss: 0.15962295, Global Avg Loss: 0.92700006, Time: 0.0211 Steps: 60720, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001118, Sample Num: 17888, Cur Loss: 0.17598842, Cur Avg Loss: 0.20675401, Log Avg loss: 0.12187096, Global Avg Loss: 0.92686748, Time: 0.0211 Steps: 60730, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001128, Sample Num: 18048, Cur Loss: 0.20153981, Cur Avg Loss: 0.20620474, Log Avg loss: 0.14479627, Global Avg Loss: 0.92673873, Time: 0.0211 Steps: 60740, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001138, Sample Num: 18208, Cur Loss: 0.23545457, Cur Avg Loss: 0.20629156, Log Avg loss: 0.21608411, Global Avg Loss: 0.92662175, Time: 0.0211 Steps: 60750, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001148, Sample Num: 18368, Cur Loss: 0.42402560, Cur Avg Loss: 0.20651583, Log Avg loss: 0.23203794, Global Avg Loss: 0.92650743, Time: 0.0211 Steps: 60760, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001158, Sample Num: 18528, Cur Loss: 0.15259437, Cur Avg Loss: 0.20673974, Log Avg loss: 0.23244437, Global Avg Loss: 0.92639322, Time: 0.0211 Steps: 60770, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001168, Sample Num: 18688, Cur Loss: 0.22182196, Cur Avg Loss: 0.20777143, Log Avg loss: 0.32724210, Global Avg Loss: 0.92629464, Time: 0.0210 Steps: 60780, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001178, Sample Num: 18848, Cur Loss: 0.28640777, Cur Avg Loss: 0.20800505, Log Avg loss: 0.23529185, Global Avg Loss: 0.92618097, Time: 0.0210 Steps: 60790, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001188, Sample Num: 19008, Cur Loss: 0.70292109, Cur Avg Loss: 0.20904743, Log Avg loss: 0.33183919, Global Avg Loss: 0.92608322, Time: 0.0211 Steps: 60800, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001198, Sample Num: 19168, Cur Loss: 0.08960283, Cur Avg Loss: 0.20888499, Log Avg loss: 0.18958687, Global Avg Loss: 0.92596210, Time: 0.0210 Steps: 60810, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001208, Sample Num: 19328, Cur Loss: 0.38543999, Cur Avg Loss: 0.20878126, Log Avg loss: 0.19635495, Global Avg Loss: 0.92584214, Time: 0.0210 Steps: 60820, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001218, Sample Num: 19488, Cur Loss: 0.33542305, Cur Avg Loss: 0.20962269, Log Avg loss: 0.31126695, Global Avg Loss: 0.92574111, Time: 0.0210 Steps: 60830, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001228, Sample Num: 19648, Cur Loss: 0.44020557, Cur Avg Loss: 0.20989620, Log Avg loss: 0.24321051, Global Avg Loss: 0.92562892, Time: 0.0210 Steps: 60840, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001238, Sample Num: 19808, Cur Loss: 0.16650389, Cur Avg Loss: 0.21025069, Log Avg loss: 0.25378141, Global Avg Loss: 0.92551851, Time: 0.0210 Steps: 60850, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001248, Sample Num: 19968, Cur Loss: 0.15464383, Cur Avg Loss: 0.21052254, Log Avg loss: 0.24417806, Global Avg Loss: 0.92540656, Time: 0.0210 Steps: 60860, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001258, Sample Num: 20128, Cur Loss: 0.34016752, Cur Avg Loss: 0.21087705, Log Avg loss: 0.25511935, Global Avg Loss: 0.92529644, Time: 0.0210 Steps: 60870, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001268, Sample Num: 20288, Cur Loss: 0.24571976, Cur Avg Loss: 0.21093621, Log Avg loss: 0.21837848, Global Avg Loss: 0.92518033, Time: 0.0210 Steps: 60880, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001278, Sample Num: 20448, Cur Loss: 0.21286039, Cur Avg Loss: 0.21077039, Log Avg loss: 0.18974531, Global Avg Loss: 0.92505955, Time: 0.0210 Steps: 60890, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001288, Sample Num: 20608, Cur Loss: 0.28034419, Cur Avg Loss: 0.21099950, Log Avg loss: 0.24027869, Global Avg Loss: 0.92494710, Time: 0.0211 Steps: 60900, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001298, Sample Num: 20768, Cur Loss: 0.06668435, Cur Avg Loss: 0.21096985, Log Avg loss: 0.20715142, Global Avg Loss: 0.92482926, Time: 0.0210 Steps: 60910, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001308, Sample Num: 20928, Cur Loss: 0.16175823, Cur Avg Loss: 0.21128531, Log Avg loss: 0.25223229, Global Avg Loss: 0.92471885, Time: 0.0210 Steps: 60920, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001318, Sample Num: 21088, Cur Loss: 0.39106402, Cur Avg Loss: 0.21167702, Log Avg loss: 0.26291188, Global Avg Loss: 0.92461023, Time: 0.0210 Steps: 60930, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001328, Sample Num: 21248, Cur Loss: 0.12379950, Cur Avg Loss: 0.21140896, Log Avg loss: 0.17607869, Global Avg Loss: 0.92448740, Time: 0.0210 Steps: 60940, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001338, Sample Num: 21408, Cur Loss: 0.23765427, Cur Avg Loss: 0.21145221, Log Avg loss: 0.21719666, Global Avg Loss: 0.92437136, Time: 0.0210 Steps: 60950, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001348, Sample Num: 21568, Cur Loss: 0.25824815, Cur Avg Loss: 0.21175171, Log Avg loss: 0.25182412, Global Avg Loss: 0.92426103, Time: 0.0210 Steps: 60960, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001358, Sample Num: 21728, Cur Loss: 0.25623369, Cur Avg Loss: 0.21128414, Log Avg loss: 0.14825545, Global Avg Loss: 0.92413376, Time: 0.0210 Steps: 60970, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001368, Sample Num: 21888, Cur Loss: 0.17748788, Cur Avg Loss: 0.21107279, Log Avg loss: 0.18237218, Global Avg Loss: 0.92401212, Time: 0.0210 Steps: 60980, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001378, Sample Num: 22048, Cur Loss: 0.22559205, Cur Avg Loss: 0.21102103, Log Avg loss: 0.20393997, Global Avg Loss: 0.92389405, Time: 0.0210 Steps: 60990, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001388, Sample Num: 22208, Cur Loss: 0.13098887, Cur Avg Loss: 0.21102979, Log Avg loss: 0.21223681, Global Avg Loss: 0.92377739, Time: 0.0212 Steps: 61000, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001398, Sample Num: 22368, Cur Loss: 0.13861248, Cur Avg Loss: 0.21105523, Log Avg loss: 0.21458654, Global Avg Loss: 0.92366114, Time: 0.0210 Steps: 61010, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001408, Sample Num: 22528, Cur Loss: 0.14893495, Cur Avg Loss: 0.21168995, Log Avg loss: 0.30042447, Global Avg Loss: 0.92355901, Time: 0.0211 Steps: 61020, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001418, Sample Num: 22688, Cur Loss: 0.21841338, Cur Avg Loss: 0.21253180, Log Avg loss: 0.33106415, Global Avg Loss: 0.92346193, Time: 0.0211 Steps: 61030, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001428, Sample Num: 22848, Cur Loss: 0.12602647, Cur Avg Loss: 0.21318377, Log Avg loss: 0.30563255, Global Avg Loss: 0.92336071, Time: 0.0210 Steps: 61040, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001438, Sample Num: 23008, Cur Loss: 0.25727183, Cur Avg Loss: 0.21381565, Log Avg loss: 0.30404823, Global Avg Loss: 0.92325927, Time: 0.0210 Steps: 61050, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001448, Sample Num: 23168, Cur Loss: 0.18421157, Cur Avg Loss: 0.21384745, Log Avg loss: 0.21842072, Global Avg Loss: 0.92314383, Time: 0.0210 Steps: 61060, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001458, Sample Num: 23328, Cur Loss: 0.19889066, Cur Avg Loss: 0.21424463, Log Avg loss: 0.27175549, Global Avg Loss: 0.92303717, Time: 0.0211 Steps: 61070, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001468, Sample Num: 23488, Cur Loss: 0.24581800, Cur Avg Loss: 0.21434786, Log Avg loss: 0.22939964, Global Avg Loss: 0.92292361, Time: 0.0210 Steps: 61080, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001478, Sample Num: 23648, Cur Loss: 0.15878932, Cur Avg Loss: 0.21382680, Log Avg loss: 0.13733435, Global Avg Loss: 0.92279501, Time: 0.0210 Steps: 61090, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001488, Sample Num: 23808, Cur Loss: 0.25088328, Cur Avg Loss: 0.21398143, Log Avg loss: 0.23683602, Global Avg Loss: 0.92268274, Time: 0.0210 Steps: 61100, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001498, Sample Num: 23968, Cur Loss: 0.09481948, Cur Avg Loss: 0.21369840, Log Avg loss: 0.17158346, Global Avg Loss: 0.92255983, Time: 0.0210 Steps: 61110, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001508, Sample Num: 24128, Cur Loss: 0.19939882, Cur Avg Loss: 0.21347638, Log Avg loss: 0.18021828, Global Avg Loss: 0.92243838, Time: 0.0210 Steps: 61120, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001518, Sample Num: 24288, Cur Loss: 0.41674599, Cur Avg Loss: 0.21349072, Log Avg loss: 0.21565264, Global Avg Loss: 0.92232276, Time: 0.0210 Steps: 61130, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001528, Sample Num: 24448, Cur Loss: 0.35135311, Cur Avg Loss: 0.21374290, Log Avg loss: 0.25202447, Global Avg Loss: 0.92221312, Time: 0.0210 Steps: 61140, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001538, Sample Num: 24608, Cur Loss: 0.18038693, Cur Avg Loss: 0.21316146, Log Avg loss: 0.12431722, Global Avg Loss: 0.92208264, Time: 0.0251 Steps: 61150, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001548, Sample Num: 24768, Cur Loss: 0.17358977, Cur Avg Loss: 0.21280603, Log Avg loss: 0.15814104, Global Avg Loss: 0.92195773, Time: 0.0211 Steps: 61160, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001558, Sample Num: 24928, Cur Loss: 0.36416018, Cur Avg Loss: 0.21289448, Log Avg loss: 0.22658691, Global Avg Loss: 0.92184406, Time: 0.0210 Steps: 61170, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001568, Sample Num: 25088, Cur Loss: 0.05524900, Cur Avg Loss: 0.21268478, Log Avg loss: 0.18001297, Global Avg Loss: 0.92172280, Time: 0.0211 Steps: 61180, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001578, Sample Num: 25248, Cur Loss: 0.17700490, Cur Avg Loss: 0.21315280, Log Avg loss: 0.28653806, Global Avg Loss: 0.92161900, Time: 0.0211 Steps: 61190, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001588, Sample Num: 25408, Cur Loss: 0.36200455, Cur Avg Loss: 0.21338092, Log Avg loss: 0.24937849, Global Avg Loss: 0.92150915, Time: 0.0211 Steps: 61200, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001598, Sample Num: 25568, Cur Loss: 0.17890444, Cur Avg Loss: 0.21318253, Log Avg loss: 0.18167764, Global Avg Loss: 0.92138828, Time: 0.0211 Steps: 61210, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001608, Sample Num: 25728, Cur Loss: 0.29888043, Cur Avg Loss: 0.21337081, Log Avg loss: 0.24345844, Global Avg Loss: 0.92127755, Time: 0.0211 Steps: 61220, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001618, Sample Num: 25888, Cur Loss: 0.09600689, Cur Avg Loss: 0.21347374, Log Avg loss: 0.23002506, Global Avg Loss: 0.92116465, Time: 0.0211 Steps: 61230, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001628, Sample Num: 26048, Cur Loss: 0.15869761, Cur Avg Loss: 0.21320579, Log Avg loss: 0.16985049, Global Avg Loss: 0.92104197, Time: 0.0211 Steps: 61240, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001638, Sample Num: 26208, Cur Loss: 0.18644525, Cur Avg Loss: 0.21294510, Log Avg loss: 0.17050506, Global Avg Loss: 0.92091943, Time: 0.0211 Steps: 61250, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001648, Sample Num: 26368, Cur Loss: 0.24795616, Cur Avg Loss: 0.21271915, Log Avg loss: 0.17570861, Global Avg Loss: 0.92079779, Time: 0.0211 Steps: 61260, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001658, Sample Num: 26528, Cur Loss: 0.03852135, Cur Avg Loss: 0.21251399, Log Avg loss: 0.17870426, Global Avg Loss: 0.92067667, Time: 0.0211 Steps: 61270, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001668, Sample Num: 26688, Cur Loss: 0.32309806, Cur Avg Loss: 0.21274011, Log Avg loss: 0.25023107, Global Avg Loss: 0.92056726, Time: 0.0211 Steps: 61280, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001678, Sample Num: 26848, Cur Loss: 0.11933613, Cur Avg Loss: 0.21292380, Log Avg loss: 0.24356230, Global Avg Loss: 0.92045680, Time: 0.0210 Steps: 61290, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001688, Sample Num: 27008, Cur Loss: 0.17393878, Cur Avg Loss: 0.21302625, Log Avg loss: 0.23021857, Global Avg Loss: 0.92034420, Time: 0.0211 Steps: 61300, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001698, Sample Num: 27168, Cur Loss: 0.28427228, Cur Avg Loss: 0.21347305, Log Avg loss: 0.28889169, Global Avg Loss: 0.92024121, Time: 0.0210 Steps: 61310, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001708, Sample Num: 27328, Cur Loss: 0.12072205, Cur Avg Loss: 0.21340226, Log Avg loss: 0.20138334, Global Avg Loss: 0.92012398, Time: 0.0211 Steps: 61320, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001718, Sample Num: 27488, Cur Loss: 0.08728082, Cur Avg Loss: 0.21325366, Log Avg loss: 0.18787195, Global Avg Loss: 0.92000458, Time: 0.0210 Steps: 61330, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001728, Sample Num: 27648, Cur Loss: 0.31292212, Cur Avg Loss: 0.21302990, Log Avg loss: 0.17458726, Global Avg Loss: 0.91988306, Time: 0.0211 Steps: 61340, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001738, Sample Num: 27808, Cur Loss: 0.04749676, Cur Avg Loss: 0.21242496, Log Avg loss: 0.10789166, Global Avg Loss: 0.91975071, Time: 0.0211 Steps: 61350, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001748, Sample Num: 27968, Cur Loss: 0.32585755, Cur Avg Loss: 0.21238982, Log Avg loss: 0.20628330, Global Avg Loss: 0.91963443, Time: 0.0211 Steps: 61360, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001758, Sample Num: 28128, Cur Loss: 0.15699634, Cur Avg Loss: 0.21224716, Log Avg loss: 0.18730913, Global Avg Loss: 0.91951510, Time: 0.0211 Steps: 61370, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001768, Sample Num: 28288, Cur Loss: 0.24731521, Cur Avg Loss: 0.21239747, Log Avg loss: 0.23882236, Global Avg Loss: 0.91940420, Time: 0.0210 Steps: 61380, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001778, Sample Num: 28448, Cur Loss: 0.19517189, Cur Avg Loss: 0.21248402, Log Avg loss: 0.22778608, Global Avg Loss: 0.91929154, Time: 0.0211 Steps: 61390, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001788, Sample Num: 28608, Cur Loss: 0.21564111, Cur Avg Loss: 0.21256753, Log Avg loss: 0.22741631, Global Avg Loss: 0.91917886, Time: 0.0211 Steps: 61400, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001798, Sample Num: 28768, Cur Loss: 0.09068005, Cur Avg Loss: 0.21238232, Log Avg loss: 0.17926636, Global Avg Loss: 0.91905837, Time: 0.0212 Steps: 61410, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001808, Sample Num: 28928, Cur Loss: 0.09432827, Cur Avg Loss: 0.21227027, Log Avg loss: 0.19212383, Global Avg Loss: 0.91894002, Time: 0.0211 Steps: 61420, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001818, Sample Num: 29088, Cur Loss: 0.33033603, Cur Avg Loss: 0.21218700, Log Avg loss: 0.19713094, Global Avg Loss: 0.91882252, Time: 0.0210 Steps: 61430, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001828, Sample Num: 29248, Cur Loss: 0.31706405, Cur Avg Loss: 0.21257615, Log Avg loss: 0.28332510, Global Avg Loss: 0.91871908, Time: 0.0211 Steps: 61440, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001838, Sample Num: 29408, Cur Loss: 0.19719179, Cur Avg Loss: 0.21305339, Log Avg loss: 0.30029134, Global Avg Loss: 0.91861844, Time: 0.0210 Steps: 61450, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001848, Sample Num: 29568, Cur Loss: 0.15246272, Cur Avg Loss: 0.21298724, Log Avg loss: 0.20083011, Global Avg Loss: 0.91850165, Time: 0.0210 Steps: 61460, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001858, Sample Num: 29728, Cur Loss: 0.08318333, Cur Avg Loss: 0.21282663, Log Avg loss: 0.18314492, Global Avg Loss: 0.91838203, Time: 0.0210 Steps: 61470, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001868, Sample Num: 29888, Cur Loss: 0.20810433, Cur Avg Loss: 0.21294761, Log Avg loss: 0.23542525, Global Avg Loss: 0.91827094, Time: 0.0211 Steps: 61480, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001878, Sample Num: 30048, Cur Loss: 0.14819810, Cur Avg Loss: 0.21264927, Log Avg loss: 0.15692054, Global Avg Loss: 0.91814712, Time: 0.0212 Steps: 61490, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001888, Sample Num: 30208, Cur Loss: 0.11218837, Cur Avg Loss: 0.21317248, Log Avg loss: 0.31143029, Global Avg Loss: 0.91804847, Time: 0.0211 Steps: 61500, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001898, Sample Num: 30368, Cur Loss: 0.60064876, Cur Avg Loss: 0.21327390, Log Avg loss: 0.23242278, Global Avg Loss: 0.91793700, Time: 0.0210 Steps: 61510, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001908, Sample Num: 30528, Cur Loss: 0.20276946, Cur Avg Loss: 0.21323142, Log Avg loss: 0.20516793, Global Avg Loss: 0.91782114, Time: 0.0211 Steps: 61520, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001918, Sample Num: 30688, Cur Loss: 0.14986056, Cur Avg Loss: 0.21316699, Log Avg loss: 0.20087377, Global Avg Loss: 0.91770462, Time: 0.0211 Steps: 61530, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001928, Sample Num: 30848, Cur Loss: 0.20451993, Cur Avg Loss: 0.21304883, Log Avg loss: 0.19038673, Global Avg Loss: 0.91758644, Time: 0.0210 Steps: 61540, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001938, Sample Num: 31008, Cur Loss: 0.16554725, Cur Avg Loss: 0.21316378, Log Avg loss: 0.23532532, Global Avg Loss: 0.91747559, Time: 0.0211 Steps: 61550, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001948, Sample Num: 31168, Cur Loss: 0.49048492, Cur Avg Loss: 0.21339327, Log Avg loss: 0.25786880, Global Avg Loss: 0.91736844, Time: 0.0210 Steps: 61560, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001958, Sample Num: 31328, Cur Loss: 0.12676059, Cur Avg Loss: 0.21335658, Log Avg loss: 0.20620884, Global Avg Loss: 0.91725294, Time: 0.0210 Steps: 61570, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001968, Sample Num: 31488, Cur Loss: 0.17855683, Cur Avg Loss: 0.21316934, Log Avg loss: 0.17650830, Global Avg Loss: 0.91713265, Time: 0.0210 Steps: 61580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001978, Sample Num: 31648, Cur Loss: 0.29127240, Cur Avg Loss: 0.21372031, Log Avg loss: 0.32215220, Global Avg Loss: 0.91703605, Time: 0.0211 Steps: 61590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001988, Sample Num: 31808, Cur Loss: 0.19284250, Cur Avg Loss: 0.21410539, Log Avg loss: 0.29027264, Global Avg Loss: 0.91693430, Time: 0.0210 Steps: 61600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001998, Sample Num: 31968, Cur Loss: 0.17342404, Cur Avg Loss: 0.21403372, Log Avg loss: 0.19978690, Global Avg Loss: 0.91681790, Time: 0.0210 Steps: 61610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002008, Sample Num: 32128, Cur Loss: 0.32952422, Cur Avg Loss: 0.21426176, Log Avg loss: 0.25982405, Global Avg Loss: 0.91671128, Time: 0.0210 Steps: 61620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002018, Sample Num: 32288, Cur Loss: 0.50841570, Cur Avg Loss: 0.21444833, Log Avg loss: 0.25191106, Global Avg Loss: 0.91660341, Time: 0.0210 Steps: 61630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002028, Sample Num: 32448, Cur Loss: 0.52977204, Cur Avg Loss: 0.21498992, Log Avg loss: 0.32428289, Global Avg Loss: 0.91650731, Time: 0.0211 Steps: 61640, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002038, Sample Num: 32608, Cur Loss: 0.26540297, Cur Avg Loss: 0.21567363, Log Avg loss: 0.35433082, Global Avg Loss: 0.91641613, Time: 0.0210 Steps: 61650, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002048, Sample Num: 32768, Cur Loss: 0.10556644, Cur Avg Loss: 0.21581073, Log Avg loss: 0.24375154, Global Avg Loss: 0.91630703, Time: 0.0255 Steps: 61660, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002058, Sample Num: 32928, Cur Loss: 0.12756209, Cur Avg Loss: 0.21586301, Log Avg loss: 0.22657019, Global Avg Loss: 0.91619519, Time: 0.0210 Steps: 61670, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002068, Sample Num: 33088, Cur Loss: 0.45999098, Cur Avg Loss: 0.21608562, Log Avg loss: 0.26189857, Global Avg Loss: 0.91608911, Time: 0.0211 Steps: 61680, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002078, Sample Num: 33248, Cur Loss: 0.33585128, Cur Avg Loss: 0.21603443, Log Avg loss: 0.20544851, Global Avg Loss: 0.91597391, Time: 0.0210 Steps: 61690, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002088, Sample Num: 33408, Cur Loss: 0.04256772, Cur Avg Loss: 0.21575937, Log Avg loss: 0.15860034, Global Avg Loss: 0.91585116, Time: 0.0210 Steps: 61700, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002098, Sample Num: 33568, Cur Loss: 0.13342969, Cur Avg Loss: 0.21555271, Log Avg loss: 0.17240369, Global Avg Loss: 0.91573069, Time: 0.0211 Steps: 61710, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002108, Sample Num: 33728, Cur Loss: 0.04175010, Cur Avg Loss: 0.21577847, Log Avg loss: 0.26314166, Global Avg Loss: 0.91562496, Time: 0.0210 Steps: 61720, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002118, Sample Num: 33888, Cur Loss: 0.12810358, Cur Avg Loss: 0.21537664, Log Avg loss: 0.13067082, Global Avg Loss: 0.91549780, Time: 0.0210 Steps: 61730, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002128, Sample Num: 34048, Cur Loss: 0.06859812, Cur Avg Loss: 0.21520256, Log Avg loss: 0.17833394, Global Avg Loss: 0.91537840, Time: 0.0211 Steps: 61740, Updated lr: 0.000042 ***** Running evaluation checkpoint-61741 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-61741 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.863527, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.268011, "eval_total_loss": 188.411836, "eval_mae": 0.356992, "eval_mse": 0.268107, "eval_r2": 0.829574, "eval_sp_statistic": 0.883473, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.913563, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.776827, "test_total_loss": 389.967272, "test_mae": 0.668924, "test_mse": 0.776928, "test_r2": 0.498564, "test_sp_statistic": 0.801716, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.837222, "test_ps_pvalue": 0.0, "lr": 4.239829302987198e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9153665498066974, "train_cur_epoch_loss": 458.1349077206105, "train_cur_epoch_avg_loss": 0.21518783829056387, "train_cur_epoch_time": 44.863527059555054, "train_cur_epoch_avg_time": 0.021072581991336332, "epoch": 29, "step": 61741} ################################################## Training, Epoch: 0030, Batch: 000009, Sample Num: 144, Cur Loss: 0.10242859, Cur Avg Loss: 0.17099546, Log Avg loss: 0.17228089, Global Avg Loss: 0.91525806, Time: 0.0211 Steps: 61750, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000019, Sample Num: 304, Cur Loss: 0.06835853, Cur Avg Loss: 0.16434826, Log Avg loss: 0.15836579, Global Avg Loss: 0.91513550, Time: 0.0210 Steps: 61760, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000029, Sample Num: 464, Cur Loss: 0.55167621, Cur Avg Loss: 0.17741541, Log Avg loss: 0.20224299, Global Avg Loss: 0.91502009, Time: 0.0210 Steps: 61770, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000039, Sample Num: 624, Cur Loss: 0.39712161, Cur Avg Loss: 0.19680051, Log Avg loss: 0.25301729, Global Avg Loss: 0.91491294, Time: 0.0211 Steps: 61780, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000049, Sample Num: 784, Cur Loss: 0.16009186, Cur Avg Loss: 0.19482378, Log Avg loss: 0.18711452, Global Avg Loss: 0.91479515, Time: 0.0211 Steps: 61790, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000059, Sample Num: 944, Cur Loss: 0.35929951, Cur Avg Loss: 0.20575229, Log Avg loss: 0.25930202, Global Avg Loss: 0.91468909, Time: 0.0210 Steps: 61800, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000069, Sample Num: 1104, Cur Loss: 0.22989476, Cur Avg Loss: 0.20667032, Log Avg loss: 0.21208666, Global Avg Loss: 0.91457542, Time: 0.0211 Steps: 61810, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000079, Sample Num: 1264, Cur Loss: 0.16495343, Cur Avg Loss: 0.21092943, Log Avg loss: 0.24031734, Global Avg Loss: 0.91446635, Time: 0.0211 Steps: 61820, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000089, Sample Num: 1424, Cur Loss: 0.33871269, Cur Avg Loss: 0.21546399, Log Avg loss: 0.25128703, Global Avg Loss: 0.91435909, Time: 0.0211 Steps: 61830, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000099, Sample Num: 1584, Cur Loss: 0.19334394, Cur Avg Loss: 0.21280312, Log Avg loss: 0.18912138, Global Avg Loss: 0.91424181, Time: 0.0211 Steps: 61840, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000109, Sample Num: 1744, Cur Loss: 0.17380475, Cur Avg Loss: 0.21087039, Log Avg loss: 0.19173629, Global Avg Loss: 0.91412500, Time: 0.0211 Steps: 61850, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000119, Sample Num: 1904, Cur Loss: 0.14197871, Cur Avg Loss: 0.20597523, Log Avg loss: 0.15261797, Global Avg Loss: 0.91400189, Time: 0.0210 Steps: 61860, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000129, Sample Num: 2064, Cur Loss: 0.15543315, Cur Avg Loss: 0.20707926, Log Avg loss: 0.22021731, Global Avg Loss: 0.91388976, Time: 0.0211 Steps: 61870, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000139, Sample Num: 2224, Cur Loss: 0.16314256, Cur Avg Loss: 0.20736023, Log Avg loss: 0.21098475, Global Avg Loss: 0.91377617, Time: 0.0211 Steps: 61880, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000149, Sample Num: 2384, Cur Loss: 0.14595112, Cur Avg Loss: 0.20829834, Log Avg loss: 0.22133795, Global Avg Loss: 0.91366429, Time: 0.0211 Steps: 61890, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000159, Sample Num: 2544, Cur Loss: 0.06521188, Cur Avg Loss: 0.20644865, Log Avg loss: 0.17888835, Global Avg Loss: 0.91354558, Time: 0.0211 Steps: 61900, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000169, Sample Num: 2704, Cur Loss: 0.12931222, Cur Avg Loss: 0.21058785, Log Avg loss: 0.27640115, Global Avg Loss: 0.91344267, Time: 0.0211 Steps: 61910, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000179, Sample Num: 2864, Cur Loss: 0.18390252, Cur Avg Loss: 0.21303448, Log Avg loss: 0.25438241, Global Avg Loss: 0.91333623, Time: 0.0211 Steps: 61920, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000189, Sample Num: 3024, Cur Loss: 0.46242592, Cur Avg Loss: 0.21264937, Log Avg loss: 0.20575596, Global Avg Loss: 0.91322197, Time: 0.0211 Steps: 61930, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000199, Sample Num: 3184, Cur Loss: 0.19471644, Cur Avg Loss: 0.21100442, Log Avg loss: 0.17991496, Global Avg Loss: 0.91310358, Time: 0.0210 Steps: 61940, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000209, Sample Num: 3344, Cur Loss: 0.20812111, Cur Avg Loss: 0.20864056, Log Avg loss: 0.16159969, Global Avg Loss: 0.91298228, Time: 0.0211 Steps: 61950, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000219, Sample Num: 3504, Cur Loss: 0.05409000, Cur Avg Loss: 0.20606654, Log Avg loss: 0.15226956, Global Avg Loss: 0.91285950, Time: 0.0211 Steps: 61960, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000229, Sample Num: 3664, Cur Loss: 0.04719844, Cur Avg Loss: 0.20722327, Log Avg loss: 0.23255564, Global Avg Loss: 0.91274972, Time: 0.0210 Steps: 61970, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000239, Sample Num: 3824, Cur Loss: 0.37353978, Cur Avg Loss: 0.20827534, Log Avg loss: 0.23236769, Global Avg Loss: 0.91263995, Time: 0.0210 Steps: 61980, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000249, Sample Num: 3984, Cur Loss: 0.18661390, Cur Avg Loss: 0.20572690, Log Avg loss: 0.14481910, Global Avg Loss: 0.91251609, Time: 0.0210 Steps: 61990, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000259, Sample Num: 4144, Cur Loss: 0.16806713, Cur Avg Loss: 0.20314903, Log Avg loss: 0.13896006, Global Avg Loss: 0.91239132, Time: 0.0249 Steps: 62000, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000269, Sample Num: 4304, Cur Loss: 0.23795961, Cur Avg Loss: 0.20292339, Log Avg loss: 0.19707930, Global Avg Loss: 0.91227596, Time: 0.0211 Steps: 62010, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000279, Sample Num: 4464, Cur Loss: 0.13429977, Cur Avg Loss: 0.20181555, Log Avg loss: 0.17201467, Global Avg Loss: 0.91215661, Time: 0.0211 Steps: 62020, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000289, Sample Num: 4624, Cur Loss: 0.15771064, Cur Avg Loss: 0.20128126, Log Avg loss: 0.18637461, Global Avg Loss: 0.91203960, Time: 0.0212 Steps: 62030, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000299, Sample Num: 4784, Cur Loss: 0.29752660, Cur Avg Loss: 0.20094587, Log Avg loss: 0.19125309, Global Avg Loss: 0.91192342, Time: 0.0211 Steps: 62040, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000309, Sample Num: 4944, Cur Loss: 0.28532630, Cur Avg Loss: 0.20169620, Log Avg loss: 0.22413115, Global Avg Loss: 0.91181257, Time: 0.0211 Steps: 62050, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000319, Sample Num: 5104, Cur Loss: 0.16963580, Cur Avg Loss: 0.20263697, Log Avg loss: 0.23170679, Global Avg Loss: 0.91170299, Time: 0.0212 Steps: 62060, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000329, Sample Num: 5264, Cur Loss: 0.09160414, Cur Avg Loss: 0.20495055, Log Avg loss: 0.27875365, Global Avg Loss: 0.91160101, Time: 0.0211 Steps: 62070, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000339, Sample Num: 5424, Cur Loss: 0.07631020, Cur Avg Loss: 0.20588979, Log Avg loss: 0.23679101, Global Avg Loss: 0.91149231, Time: 0.0211 Steps: 62080, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000349, Sample Num: 5584, Cur Loss: 0.22241247, Cur Avg Loss: 0.20694475, Log Avg loss: 0.24270787, Global Avg Loss: 0.91138460, Time: 0.0211 Steps: 62090, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000359, Sample Num: 5744, Cur Loss: 0.04239001, Cur Avg Loss: 0.20639300, Log Avg loss: 0.18713683, Global Avg Loss: 0.91126797, Time: 0.0211 Steps: 62100, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000369, Sample Num: 5904, Cur Loss: 0.02119848, Cur Avg Loss: 0.20439470, Log Avg loss: 0.13265567, Global Avg Loss: 0.91114261, Time: 0.0211 Steps: 62110, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000379, Sample Num: 6064, Cur Loss: 0.21422414, Cur Avg Loss: 0.20366113, Log Avg loss: 0.17659256, Global Avg Loss: 0.91102437, Time: 0.0211 Steps: 62120, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000389, Sample Num: 6224, Cur Loss: 0.10647162, Cur Avg Loss: 0.20286846, Log Avg loss: 0.17282610, Global Avg Loss: 0.91090555, Time: 0.0211 Steps: 62130, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000399, Sample Num: 6384, Cur Loss: 0.18189912, Cur Avg Loss: 0.20138311, Log Avg loss: 0.14360297, Global Avg Loss: 0.91078207, Time: 0.0211 Steps: 62140, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000409, Sample Num: 6544, Cur Loss: 0.20381120, Cur Avg Loss: 0.20212932, Log Avg loss: 0.23190330, Global Avg Loss: 0.91067284, Time: 0.0210 Steps: 62150, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000419, Sample Num: 6704, Cur Loss: 0.11403379, Cur Avg Loss: 0.20319128, Log Avg loss: 0.24662534, Global Avg Loss: 0.91056601, Time: 0.0210 Steps: 62160, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000429, Sample Num: 6864, Cur Loss: 0.03220825, Cur Avg Loss: 0.20330188, Log Avg loss: 0.20793613, Global Avg Loss: 0.91045299, Time: 0.0211 Steps: 62170, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000439, Sample Num: 7024, Cur Loss: 0.21330711, Cur Avg Loss: 0.20345110, Log Avg loss: 0.20985254, Global Avg Loss: 0.91034032, Time: 0.0211 Steps: 62180, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000449, Sample Num: 7184, Cur Loss: 0.15862738, Cur Avg Loss: 0.20238434, Log Avg loss: 0.15555363, Global Avg Loss: 0.91021895, Time: 0.0211 Steps: 62190, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000459, Sample Num: 7344, Cur Loss: 0.05812206, Cur Avg Loss: 0.20084940, Log Avg loss: 0.13193023, Global Avg Loss: 0.91009383, Time: 0.0211 Steps: 62200, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000469, Sample Num: 7504, Cur Loss: 0.51157004, Cur Avg Loss: 0.20185752, Log Avg loss: 0.24813055, Global Avg Loss: 0.90998742, Time: 0.0211 Steps: 62210, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000479, Sample Num: 7664, Cur Loss: 0.28666484, Cur Avg Loss: 0.20188363, Log Avg loss: 0.20310822, Global Avg Loss: 0.90987381, Time: 0.0212 Steps: 62220, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000489, Sample Num: 7824, Cur Loss: 0.15478522, Cur Avg Loss: 0.20133793, Log Avg loss: 0.17519879, Global Avg Loss: 0.90975575, Time: 0.0211 Steps: 62230, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000499, Sample Num: 7984, Cur Loss: 0.40517175, Cur Avg Loss: 0.20108962, Log Avg loss: 0.18894744, Global Avg Loss: 0.90963994, Time: 0.0210 Steps: 62240, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000509, Sample Num: 8144, Cur Loss: 0.26709390, Cur Avg Loss: 0.20192835, Log Avg loss: 0.24378068, Global Avg Loss: 0.90953297, Time: 0.0211 Steps: 62250, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000519, Sample Num: 8304, Cur Loss: 0.06155708, Cur Avg Loss: 0.20037275, Log Avg loss: 0.12119301, Global Avg Loss: 0.90940635, Time: 0.0220 Steps: 62260, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000529, Sample Num: 8464, Cur Loss: 0.23264821, Cur Avg Loss: 0.19996855, Log Avg loss: 0.17899058, Global Avg Loss: 0.90928906, Time: 0.0219 Steps: 62270, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000539, Sample Num: 8624, Cur Loss: 0.20935908, Cur Avg Loss: 0.20055709, Log Avg loss: 0.23169070, Global Avg Loss: 0.90918026, Time: 0.0219 Steps: 62280, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000549, Sample Num: 8784, Cur Loss: 0.17989746, Cur Avg Loss: 0.20007565, Log Avg loss: 0.17412624, Global Avg Loss: 0.90906225, Time: 0.0219 Steps: 62290, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000559, Sample Num: 8944, Cur Loss: 0.13485871, Cur Avg Loss: 0.19972145, Log Avg loss: 0.18027534, Global Avg Loss: 0.90894527, Time: 0.0219 Steps: 62300, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000569, Sample Num: 9104, Cur Loss: 0.40214738, Cur Avg Loss: 0.20171832, Log Avg loss: 0.31334357, Global Avg Loss: 0.90884969, Time: 0.0220 Steps: 62310, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000579, Sample Num: 9264, Cur Loss: 0.21496405, Cur Avg Loss: 0.20185200, Log Avg loss: 0.20945866, Global Avg Loss: 0.90873746, Time: 0.0219 Steps: 62320, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000589, Sample Num: 9424, Cur Loss: 0.17212173, Cur Avg Loss: 0.20175931, Log Avg loss: 0.19639255, Global Avg Loss: 0.90862317, Time: 0.0219 Steps: 62330, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000599, Sample Num: 9584, Cur Loss: 0.19429253, Cur Avg Loss: 0.20363397, Log Avg loss: 0.31405103, Global Avg Loss: 0.90852780, Time: 0.0219 Steps: 62340, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000609, Sample Num: 9744, Cur Loss: 0.14465928, Cur Avg Loss: 0.20393203, Log Avg loss: 0.22178574, Global Avg Loss: 0.90841765, Time: 0.0219 Steps: 62350, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000619, Sample Num: 9904, Cur Loss: 0.41758132, Cur Avg Loss: 0.20312361, Log Avg loss: 0.15389084, Global Avg Loss: 0.90829666, Time: 0.0220 Steps: 62360, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000629, Sample Num: 10064, Cur Loss: 0.36066800, Cur Avg Loss: 0.20353128, Log Avg loss: 0.22876664, Global Avg Loss: 0.90818771, Time: 0.0219 Steps: 62370, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000639, Sample Num: 10224, Cur Loss: 0.17832586, Cur Avg Loss: 0.20346000, Log Avg loss: 0.19897602, Global Avg Loss: 0.90807402, Time: 0.0219 Steps: 62380, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000649, Sample Num: 10384, Cur Loss: 0.12661745, Cur Avg Loss: 0.20470127, Log Avg loss: 0.28401857, Global Avg Loss: 0.90797399, Time: 0.0219 Steps: 62390, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000659, Sample Num: 10544, Cur Loss: 0.38079613, Cur Avg Loss: 0.20478685, Log Avg loss: 0.21034115, Global Avg Loss: 0.90786219, Time: 0.0219 Steps: 62400, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000669, Sample Num: 10704, Cur Loss: 0.04202581, Cur Avg Loss: 0.20397812, Log Avg loss: 0.15068269, Global Avg Loss: 0.90774087, Time: 0.0219 Steps: 62410, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000679, Sample Num: 10864, Cur Loss: 0.07124149, Cur Avg Loss: 0.20397985, Log Avg loss: 0.20409533, Global Avg Loss: 0.90762814, Time: 0.0219 Steps: 62420, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000689, Sample Num: 11024, Cur Loss: 0.03364391, Cur Avg Loss: 0.20409324, Log Avg loss: 0.21179237, Global Avg Loss: 0.90751668, Time: 0.0219 Steps: 62430, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000699, Sample Num: 11184, Cur Loss: 0.21461631, Cur Avg Loss: 0.20390752, Log Avg loss: 0.19111156, Global Avg Loss: 0.90740195, Time: 0.0219 Steps: 62440, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000709, Sample Num: 11344, Cur Loss: 0.41332954, Cur Avg Loss: 0.20368571, Log Avg loss: 0.18818115, Global Avg Loss: 0.90728678, Time: 0.0219 Steps: 62450, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000719, Sample Num: 11504, Cur Loss: 0.09218366, Cur Avg Loss: 0.20337246, Log Avg loss: 0.18116321, Global Avg Loss: 0.90717052, Time: 0.0219 Steps: 62460, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000729, Sample Num: 11664, Cur Loss: 0.11700937, Cur Avg Loss: 0.20290940, Log Avg loss: 0.16961542, Global Avg Loss: 0.90705246, Time: 0.0219 Steps: 62470, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000739, Sample Num: 11824, Cur Loss: 0.35459477, Cur Avg Loss: 0.20240787, Log Avg loss: 0.16584631, Global Avg Loss: 0.90693383, Time: 0.0219 Steps: 62480, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000749, Sample Num: 11984, Cur Loss: 0.13189477, Cur Avg Loss: 0.20275842, Log Avg loss: 0.22866434, Global Avg Loss: 0.90682529, Time: 0.0219 Steps: 62490, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000759, Sample Num: 12144, Cur Loss: 0.45928943, Cur Avg Loss: 0.20286992, Log Avg loss: 0.21122136, Global Avg Loss: 0.90671399, Time: 0.0219 Steps: 62500, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000769, Sample Num: 12304, Cur Loss: 0.07906767, Cur Avg Loss: 0.20240327, Log Avg loss: 0.16698437, Global Avg Loss: 0.90659565, Time: 0.0247 Steps: 62510, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000779, Sample Num: 12464, Cur Loss: 0.15811162, Cur Avg Loss: 0.20228539, Log Avg loss: 0.19321987, Global Avg Loss: 0.90648155, Time: 0.0211 Steps: 62520, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000789, Sample Num: 12624, Cur Loss: 0.10272802, Cur Avg Loss: 0.20159198, Log Avg loss: 0.14757555, Global Avg Loss: 0.90636018, Time: 0.0210 Steps: 62530, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000799, Sample Num: 12784, Cur Loss: 0.09778741, Cur Avg Loss: 0.20068669, Log Avg loss: 0.12925919, Global Avg Loss: 0.90623593, Time: 0.0211 Steps: 62540, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000809, Sample Num: 12944, Cur Loss: 0.14027895, Cur Avg Loss: 0.20054436, Log Avg loss: 0.18917236, Global Avg Loss: 0.90612129, Time: 0.0210 Steps: 62550, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000819, Sample Num: 13104, Cur Loss: 0.21277919, Cur Avg Loss: 0.20036745, Log Avg loss: 0.18605570, Global Avg Loss: 0.90600619, Time: 0.0210 Steps: 62560, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000829, Sample Num: 13264, Cur Loss: 0.26702252, Cur Avg Loss: 0.20103059, Log Avg loss: 0.25534186, Global Avg Loss: 0.90590220, Time: 0.0210 Steps: 62570, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000839, Sample Num: 13424, Cur Loss: 0.34325179, Cur Avg Loss: 0.20141590, Log Avg loss: 0.23335793, Global Avg Loss: 0.90579473, Time: 0.0210 Steps: 62580, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000849, Sample Num: 13584, Cur Loss: 0.37390822, Cur Avg Loss: 0.20173149, Log Avg loss: 0.22820936, Global Avg Loss: 0.90568647, Time: 0.0210 Steps: 62590, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000859, Sample Num: 13744, Cur Loss: 0.79784197, Cur Avg Loss: 0.20252142, Log Avg loss: 0.26958668, Global Avg Loss: 0.90558486, Time: 0.0210 Steps: 62600, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000869, Sample Num: 13904, Cur Loss: 0.11889705, Cur Avg Loss: 0.20272889, Log Avg loss: 0.22055063, Global Avg Loss: 0.90547544, Time: 0.0210 Steps: 62610, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000879, Sample Num: 14064, Cur Loss: 0.15254444, Cur Avg Loss: 0.20319651, Log Avg loss: 0.24383269, Global Avg Loss: 0.90536978, Time: 0.0210 Steps: 62620, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000889, Sample Num: 14224, Cur Loss: 0.21805091, Cur Avg Loss: 0.20429038, Log Avg loss: 0.30044127, Global Avg Loss: 0.90527320, Time: 0.0210 Steps: 62630, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000899, Sample Num: 14384, Cur Loss: 0.06421205, Cur Avg Loss: 0.20411962, Log Avg loss: 0.18893911, Global Avg Loss: 0.90515884, Time: 0.0210 Steps: 62640, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000909, Sample Num: 14544, Cur Loss: 0.22530851, Cur Avg Loss: 0.20397801, Log Avg loss: 0.19124763, Global Avg Loss: 0.90504489, Time: 0.0210 Steps: 62650, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000919, Sample Num: 14704, Cur Loss: 0.25581187, Cur Avg Loss: 0.20454029, Log Avg loss: 0.25565128, Global Avg Loss: 0.90494125, Time: 0.0210 Steps: 62660, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000929, Sample Num: 14864, Cur Loss: 0.26190782, Cur Avg Loss: 0.20422132, Log Avg loss: 0.17490767, Global Avg Loss: 0.90482476, Time: 0.0210 Steps: 62670, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000939, Sample Num: 15024, Cur Loss: 0.05201859, Cur Avg Loss: 0.20337817, Log Avg loss: 0.12504990, Global Avg Loss: 0.90470036, Time: 0.0210 Steps: 62680, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000949, Sample Num: 15184, Cur Loss: 0.07100834, Cur Avg Loss: 0.20280857, Log Avg loss: 0.14932288, Global Avg Loss: 0.90457986, Time: 0.0210 Steps: 62690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000959, Sample Num: 15344, Cur Loss: 0.16103990, Cur Avg Loss: 0.20263768, Log Avg loss: 0.18642047, Global Avg Loss: 0.90446532, Time: 0.0210 Steps: 62700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000969, Sample Num: 15504, Cur Loss: 0.11148244, Cur Avg Loss: 0.20266794, Log Avg loss: 0.20556959, Global Avg Loss: 0.90435387, Time: 0.0210 Steps: 62710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000979, Sample Num: 15664, Cur Loss: 0.06281462, Cur Avg Loss: 0.20278896, Log Avg loss: 0.21451631, Global Avg Loss: 0.90424389, Time: 0.0210 Steps: 62720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000989, Sample Num: 15824, Cur Loss: 0.30822504, Cur Avg Loss: 0.20349244, Log Avg loss: 0.27236235, Global Avg Loss: 0.90414316, Time: 0.0210 Steps: 62730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000999, Sample Num: 15984, Cur Loss: 0.19806075, Cur Avg Loss: 0.20362283, Log Avg loss: 0.21651888, Global Avg Loss: 0.90403356, Time: 0.0210 Steps: 62740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001009, Sample Num: 16144, Cur Loss: 0.19775462, Cur Avg Loss: 0.20337620, Log Avg loss: 0.17873735, Global Avg Loss: 0.90391797, Time: 0.0210 Steps: 62750, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001019, Sample Num: 16304, Cur Loss: 0.64783835, Cur Avg Loss: 0.20445229, Log Avg loss: 0.31303011, Global Avg Loss: 0.90382382, Time: 0.0210 Steps: 62760, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001029, Sample Num: 16464, Cur Loss: 0.30998021, Cur Avg Loss: 0.20453930, Log Avg loss: 0.21340562, Global Avg Loss: 0.90371383, Time: 0.0242 Steps: 62770, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001039, Sample Num: 16624, Cur Loss: 0.03911612, Cur Avg Loss: 0.20466170, Log Avg loss: 0.21725657, Global Avg Loss: 0.90360449, Time: 0.0211 Steps: 62780, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001049, Sample Num: 16784, Cur Loss: 0.10701160, Cur Avg Loss: 0.20500088, Log Avg loss: 0.24024168, Global Avg Loss: 0.90349884, Time: 0.0211 Steps: 62790, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001059, Sample Num: 16944, Cur Loss: 0.18949962, Cur Avg Loss: 0.20632712, Log Avg loss: 0.34544946, Global Avg Loss: 0.90340998, Time: 0.0211 Steps: 62800, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001069, Sample Num: 17104, Cur Loss: 0.13711980, Cur Avg Loss: 0.20583858, Log Avg loss: 0.15410300, Global Avg Loss: 0.90329068, Time: 0.0210 Steps: 62810, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001079, Sample Num: 17264, Cur Loss: 0.09248353, Cur Avg Loss: 0.20599117, Log Avg loss: 0.22230296, Global Avg Loss: 0.90318228, Time: 0.0210 Steps: 62820, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001089, Sample Num: 17424, Cur Loss: 0.76360261, Cur Avg Loss: 0.20640642, Log Avg loss: 0.25121108, Global Avg Loss: 0.90307851, Time: 0.0210 Steps: 62830, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001099, Sample Num: 17584, Cur Loss: 0.24152020, Cur Avg Loss: 0.20672839, Log Avg loss: 0.24179162, Global Avg Loss: 0.90297328, Time: 0.0210 Steps: 62840, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001109, Sample Num: 17744, Cur Loss: 0.41182074, Cur Avg Loss: 0.20697952, Log Avg loss: 0.23457801, Global Avg Loss: 0.90286693, Time: 0.0210 Steps: 62850, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001119, Sample Num: 17904, Cur Loss: 0.12511641, Cur Avg Loss: 0.20703435, Log Avg loss: 0.21311576, Global Avg Loss: 0.90275720, Time: 0.0210 Steps: 62860, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001129, Sample Num: 18064, Cur Loss: 0.13913575, Cur Avg Loss: 0.20764809, Log Avg loss: 0.27632535, Global Avg Loss: 0.90265756, Time: 0.0210 Steps: 62870, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001139, Sample Num: 18224, Cur Loss: 0.24006532, Cur Avg Loss: 0.20736447, Log Avg loss: 0.17534315, Global Avg Loss: 0.90254189, Time: 0.0210 Steps: 62880, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001149, Sample Num: 18384, Cur Loss: 0.35560259, Cur Avg Loss: 0.20745131, Log Avg loss: 0.21734301, Global Avg Loss: 0.90243294, Time: 0.0210 Steps: 62890, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001159, Sample Num: 18544, Cur Loss: 0.12707615, Cur Avg Loss: 0.20747690, Log Avg loss: 0.21041695, Global Avg Loss: 0.90232292, Time: 0.0210 Steps: 62900, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001169, Sample Num: 18704, Cur Loss: 0.10238987, Cur Avg Loss: 0.20751754, Log Avg loss: 0.21222842, Global Avg Loss: 0.90221323, Time: 0.0210 Steps: 62910, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001179, Sample Num: 18864, Cur Loss: 0.66881859, Cur Avg Loss: 0.20771032, Log Avg loss: 0.23024582, Global Avg Loss: 0.90210643, Time: 0.0210 Steps: 62920, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001189, Sample Num: 19024, Cur Loss: 0.21983752, Cur Avg Loss: 0.20814509, Log Avg loss: 0.25940395, Global Avg Loss: 0.90200430, Time: 0.0211 Steps: 62930, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001199, Sample Num: 19184, Cur Loss: 0.25552124, Cur Avg Loss: 0.20841071, Log Avg loss: 0.23999348, Global Avg Loss: 0.90189912, Time: 0.0210 Steps: 62940, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001209, Sample Num: 19344, Cur Loss: 0.46902078, Cur Avg Loss: 0.20833372, Log Avg loss: 0.19910316, Global Avg Loss: 0.90178748, Time: 0.0210 Steps: 62950, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001219, Sample Num: 19504, Cur Loss: 0.13592735, Cur Avg Loss: 0.20789424, Log Avg loss: 0.15476095, Global Avg Loss: 0.90166883, Time: 0.0210 Steps: 62960, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001229, Sample Num: 19664, Cur Loss: 0.09872946, Cur Avg Loss: 0.20770969, Log Avg loss: 0.18521237, Global Avg Loss: 0.90155505, Time: 0.0211 Steps: 62970, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001239, Sample Num: 19824, Cur Loss: 0.21512157, Cur Avg Loss: 0.20803681, Log Avg loss: 0.24823956, Global Avg Loss: 0.90145131, Time: 0.0210 Steps: 62980, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001249, Sample Num: 19984, Cur Loss: 0.40010262, Cur Avg Loss: 0.20756207, Log Avg loss: 0.14874175, Global Avg Loss: 0.90133182, Time: 0.0210 Steps: 62990, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001259, Sample Num: 20144, Cur Loss: 0.16314067, Cur Avg Loss: 0.20748827, Log Avg loss: 0.19827142, Global Avg Loss: 0.90122022, Time: 0.0210 Steps: 63000, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001269, Sample Num: 20304, Cur Loss: 0.19084999, Cur Avg Loss: 0.20722128, Log Avg loss: 0.17360736, Global Avg Loss: 0.90110474, Time: 0.0210 Steps: 63010, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001279, Sample Num: 20464, Cur Loss: 0.04796289, Cur Avg Loss: 0.20755627, Log Avg loss: 0.25006666, Global Avg Loss: 0.90100144, Time: 0.0210 Steps: 63020, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001289, Sample Num: 20624, Cur Loss: 0.17377515, Cur Avg Loss: 0.20723287, Log Avg loss: 0.16586918, Global Avg Loss: 0.90088481, Time: 0.0211 Steps: 63030, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001299, Sample Num: 20784, Cur Loss: 0.06051436, Cur Avg Loss: 0.20707835, Log Avg loss: 0.18716079, Global Avg Loss: 0.90077159, Time: 0.0211 Steps: 63040, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001309, Sample Num: 20944, Cur Loss: 0.65066105, Cur Avg Loss: 0.20743346, Log Avg loss: 0.25356283, Global Avg Loss: 0.90066894, Time: 0.0211 Steps: 63050, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001319, Sample Num: 21104, Cur Loss: 0.08795758, Cur Avg Loss: 0.20714479, Log Avg loss: 0.16935797, Global Avg Loss: 0.90055297, Time: 0.0210 Steps: 63060, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001329, Sample Num: 21264, Cur Loss: 0.07498255, Cur Avg Loss: 0.20669913, Log Avg loss: 0.14791678, Global Avg Loss: 0.90043363, Time: 0.0211 Steps: 63070, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001339, Sample Num: 21424, Cur Loss: 0.21813275, Cur Avg Loss: 0.20670405, Log Avg loss: 0.20735792, Global Avg Loss: 0.90032376, Time: 0.0210 Steps: 63080, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001349, Sample Num: 21584, Cur Loss: 0.09435073, Cur Avg Loss: 0.20678381, Log Avg loss: 0.21746289, Global Avg Loss: 0.90021553, Time: 0.0210 Steps: 63090, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001359, Sample Num: 21744, Cur Loss: 0.06558248, Cur Avg Loss: 0.20670729, Log Avg loss: 0.19638540, Global Avg Loss: 0.90010398, Time: 0.0210 Steps: 63100, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001369, Sample Num: 21904, Cur Loss: 0.20392892, Cur Avg Loss: 0.20666574, Log Avg loss: 0.20101896, Global Avg Loss: 0.89999321, Time: 0.0210 Steps: 63110, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001379, Sample Num: 22064, Cur Loss: 0.18537110, Cur Avg Loss: 0.20681434, Log Avg loss: 0.22715733, Global Avg Loss: 0.89988661, Time: 0.0210 Steps: 63120, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001389, Sample Num: 22224, Cur Loss: 0.20764878, Cur Avg Loss: 0.20731243, Log Avg loss: 0.27599953, Global Avg Loss: 0.89978779, Time: 0.0210 Steps: 63130, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001399, Sample Num: 22384, Cur Loss: 0.31774166, Cur Avg Loss: 0.20766949, Log Avg loss: 0.25726404, Global Avg Loss: 0.89968603, Time: 0.0210 Steps: 63140, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001409, Sample Num: 22544, Cur Loss: 0.09314115, Cur Avg Loss: 0.20730978, Log Avg loss: 0.15698655, Global Avg Loss: 0.89956842, Time: 0.0210 Steps: 63150, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001419, Sample Num: 22704, Cur Loss: 0.41909099, Cur Avg Loss: 0.20707353, Log Avg loss: 0.17378697, Global Avg Loss: 0.89945351, Time: 0.0210 Steps: 63160, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001429, Sample Num: 22864, Cur Loss: 0.09226428, Cur Avg Loss: 0.20742045, Log Avg loss: 0.25664708, Global Avg Loss: 0.89935175, Time: 0.0210 Steps: 63170, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001439, Sample Num: 23024, Cur Loss: 0.11260236, Cur Avg Loss: 0.20731662, Log Avg loss: 0.19248037, Global Avg Loss: 0.89923987, Time: 0.0210 Steps: 63180, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001449, Sample Num: 23184, Cur Loss: 0.30833390, Cur Avg Loss: 0.20748398, Log Avg loss: 0.23156609, Global Avg Loss: 0.89913421, Time: 0.0211 Steps: 63190, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001459, Sample Num: 23344, Cur Loss: 0.19441539, Cur Avg Loss: 0.20754598, Log Avg loss: 0.21653023, Global Avg Loss: 0.89902620, Time: 0.0211 Steps: 63200, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001469, Sample Num: 23504, Cur Loss: 0.20878386, Cur Avg Loss: 0.20788036, Log Avg loss: 0.25666605, Global Avg Loss: 0.89892458, Time: 0.0210 Steps: 63210, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001479, Sample Num: 23664, Cur Loss: 0.09034576, Cur Avg Loss: 0.20811848, Log Avg loss: 0.24309839, Global Avg Loss: 0.89882084, Time: 0.0210 Steps: 63220, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001489, Sample Num: 23824, Cur Loss: 0.11881645, Cur Avg Loss: 0.20797425, Log Avg loss: 0.18664353, Global Avg Loss: 0.89870821, Time: 0.0210 Steps: 63230, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001499, Sample Num: 23984, Cur Loss: 0.21151885, Cur Avg Loss: 0.20828790, Log Avg loss: 0.25499032, Global Avg Loss: 0.89860642, Time: 0.0209 Steps: 63240, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001509, Sample Num: 24144, Cur Loss: 0.16419926, Cur Avg Loss: 0.20817015, Log Avg loss: 0.19051832, Global Avg Loss: 0.89849446, Time: 0.0210 Steps: 63250, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001519, Sample Num: 24304, Cur Loss: 0.36806092, Cur Avg Loss: 0.20850240, Log Avg loss: 0.25864022, Global Avg Loss: 0.89839332, Time: 0.0210 Steps: 63260, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001529, Sample Num: 24464, Cur Loss: 0.09821035, Cur Avg Loss: 0.20834678, Log Avg loss: 0.18470697, Global Avg Loss: 0.89828052, Time: 0.0210 Steps: 63270, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001539, Sample Num: 24624, Cur Loss: 0.31416950, Cur Avg Loss: 0.20846843, Log Avg loss: 0.22706942, Global Avg Loss: 0.89817445, Time: 0.0247 Steps: 63280, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001549, Sample Num: 24784, Cur Loss: 0.55892408, Cur Avg Loss: 0.20903769, Log Avg loss: 0.29664622, Global Avg Loss: 0.89807940, Time: 0.0210 Steps: 63290, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001559, Sample Num: 24944, Cur Loss: 0.03669558, Cur Avg Loss: 0.20955685, Log Avg loss: 0.28997524, Global Avg Loss: 0.89798334, Time: 0.0210 Steps: 63300, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001569, Sample Num: 25104, Cur Loss: 0.09104159, Cur Avg Loss: 0.20920075, Log Avg loss: 0.15368461, Global Avg Loss: 0.89786577, Time: 0.0210 Steps: 63310, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001579, Sample Num: 25264, Cur Loss: 0.17862567, Cur Avg Loss: 0.20889841, Log Avg loss: 0.16146115, Global Avg Loss: 0.89774947, Time: 0.0210 Steps: 63320, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001589, Sample Num: 25424, Cur Loss: 0.13932674, Cur Avg Loss: 0.20876749, Log Avg loss: 0.18809520, Global Avg Loss: 0.89763742, Time: 0.0210 Steps: 63330, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001599, Sample Num: 25584, Cur Loss: 0.14330715, Cur Avg Loss: 0.20906055, Log Avg loss: 0.25562825, Global Avg Loss: 0.89753606, Time: 0.0210 Steps: 63340, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001609, Sample Num: 25744, Cur Loss: 0.23698771, Cur Avg Loss: 0.20884022, Log Avg loss: 0.17360848, Global Avg Loss: 0.89742178, Time: 0.0210 Steps: 63350, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001619, Sample Num: 25904, Cur Loss: 0.50725114, Cur Avg Loss: 0.20859402, Log Avg loss: 0.16898173, Global Avg Loss: 0.89730682, Time: 0.0210 Steps: 63360, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001629, Sample Num: 26064, Cur Loss: 0.16844226, Cur Avg Loss: 0.20818987, Log Avg loss: 0.14275790, Global Avg Loss: 0.89718775, Time: 0.0210 Steps: 63370, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001639, Sample Num: 26224, Cur Loss: 0.27328485, Cur Avg Loss: 0.20811544, Log Avg loss: 0.19599012, Global Avg Loss: 0.89707711, Time: 0.0210 Steps: 63380, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001649, Sample Num: 26384, Cur Loss: 0.31546283, Cur Avg Loss: 0.20785581, Log Avg loss: 0.16530284, Global Avg Loss: 0.89696167, Time: 0.0210 Steps: 63390, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001659, Sample Num: 26544, Cur Loss: 0.05265062, Cur Avg Loss: 0.20772644, Log Avg loss: 0.18639319, Global Avg Loss: 0.89684959, Time: 0.0210 Steps: 63400, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001669, Sample Num: 26704, Cur Loss: 0.22905985, Cur Avg Loss: 0.20767179, Log Avg loss: 0.19860460, Global Avg Loss: 0.89673948, Time: 0.0210 Steps: 63410, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001679, Sample Num: 26864, Cur Loss: 0.41427389, Cur Avg Loss: 0.20791946, Log Avg loss: 0.24925697, Global Avg Loss: 0.89663738, Time: 0.0211 Steps: 63420, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001689, Sample Num: 27024, Cur Loss: 0.47066757, Cur Avg Loss: 0.20825193, Log Avg loss: 0.26407310, Global Avg Loss: 0.89653766, Time: 0.0210 Steps: 63430, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001699, Sample Num: 27184, Cur Loss: 0.37765563, Cur Avg Loss: 0.20827432, Log Avg loss: 0.21205583, Global Avg Loss: 0.89642976, Time: 0.0210 Steps: 63440, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001709, Sample Num: 27344, Cur Loss: 0.11598646, Cur Avg Loss: 0.20795531, Log Avg loss: 0.15375628, Global Avg Loss: 0.89631272, Time: 0.0210 Steps: 63450, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001719, Sample Num: 27504, Cur Loss: 0.11497270, Cur Avg Loss: 0.20762907, Log Avg loss: 0.15187379, Global Avg Loss: 0.89619541, Time: 0.0211 Steps: 63460, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001729, Sample Num: 27664, Cur Loss: 0.28690010, Cur Avg Loss: 0.20741462, Log Avg loss: 0.17055153, Global Avg Loss: 0.89608108, Time: 0.0210 Steps: 63470, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001739, Sample Num: 27824, Cur Loss: 0.26995081, Cur Avg Loss: 0.20750377, Log Avg loss: 0.22291665, Global Avg Loss: 0.89597503, Time: 0.0210 Steps: 63480, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001749, Sample Num: 27984, Cur Loss: 0.16426602, Cur Avg Loss: 0.20751333, Log Avg loss: 0.20917583, Global Avg Loss: 0.89586686, Time: 0.0210 Steps: 63490, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001759, Sample Num: 28144, Cur Loss: 0.22375295, Cur Avg Loss: 0.20726299, Log Avg loss: 0.16347872, Global Avg Loss: 0.89575152, Time: 0.0210 Steps: 63500, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001769, Sample Num: 28304, Cur Loss: 0.09240761, Cur Avg Loss: 0.20721373, Log Avg loss: 0.19854951, Global Avg Loss: 0.89564175, Time: 0.0210 Steps: 63510, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001779, Sample Num: 28464, Cur Loss: 0.48644283, Cur Avg Loss: 0.20797954, Log Avg loss: 0.34345095, Global Avg Loss: 0.89555481, Time: 0.0210 Steps: 63520, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001789, Sample Num: 28624, Cur Loss: 0.12138319, Cur Avg Loss: 0.20798099, Log Avg loss: 0.20823870, Global Avg Loss: 0.89544663, Time: 0.0210 Steps: 63530, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001799, Sample Num: 28784, Cur Loss: 0.36089677, Cur Avg Loss: 0.20774753, Log Avg loss: 0.16598132, Global Avg Loss: 0.89533182, Time: 0.0233 Steps: 63540, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001809, Sample Num: 28944, Cur Loss: 0.25922471, Cur Avg Loss: 0.20793067, Log Avg loss: 0.24087782, Global Avg Loss: 0.89522884, Time: 0.0211 Steps: 63550, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001819, Sample Num: 29104, Cur Loss: 0.06342506, Cur Avg Loss: 0.20786757, Log Avg loss: 0.19645312, Global Avg Loss: 0.89511890, Time: 0.0210 Steps: 63560, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001829, Sample Num: 29264, Cur Loss: 0.18990147, Cur Avg Loss: 0.20790228, Log Avg loss: 0.21421626, Global Avg Loss: 0.89501179, Time: 0.0210 Steps: 63570, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001839, Sample Num: 29424, Cur Loss: 0.08541606, Cur Avg Loss: 0.20784839, Log Avg loss: 0.19799159, Global Avg Loss: 0.89490216, Time: 0.0210 Steps: 63580, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001849, Sample Num: 29584, Cur Loss: 0.23641646, Cur Avg Loss: 0.20806235, Log Avg loss: 0.24741004, Global Avg Loss: 0.89480034, Time: 0.0211 Steps: 63590, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001859, Sample Num: 29744, Cur Loss: 0.18802921, Cur Avg Loss: 0.20803495, Log Avg loss: 0.20296808, Global Avg Loss: 0.89469156, Time: 0.0211 Steps: 63600, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001869, Sample Num: 29904, Cur Loss: 0.58672452, Cur Avg Loss: 0.20788000, Log Avg loss: 0.17907553, Global Avg Loss: 0.89457906, Time: 0.0211 Steps: 63610, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001879, Sample Num: 30064, Cur Loss: 0.25838009, Cur Avg Loss: 0.20779903, Log Avg loss: 0.19266468, Global Avg Loss: 0.89446873, Time: 0.0212 Steps: 63620, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001889, Sample Num: 30224, Cur Loss: 0.22516863, Cur Avg Loss: 0.20824837, Log Avg loss: 0.29268062, Global Avg Loss: 0.89437415, Time: 0.0210 Steps: 63630, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001899, Sample Num: 30384, Cur Loss: 0.12727521, Cur Avg Loss: 0.20834667, Log Avg loss: 0.22691484, Global Avg Loss: 0.89426927, Time: 0.0210 Steps: 63640, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001909, Sample Num: 30544, Cur Loss: 0.11190251, Cur Avg Loss: 0.20825613, Log Avg loss: 0.19106211, Global Avg Loss: 0.89415879, Time: 0.0210 Steps: 63650, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001919, Sample Num: 30704, Cur Loss: 0.13056071, Cur Avg Loss: 0.20806796, Log Avg loss: 0.17214716, Global Avg Loss: 0.89404537, Time: 0.0210 Steps: 63660, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001929, Sample Num: 30864, Cur Loss: 0.38346291, Cur Avg Loss: 0.20830207, Log Avg loss: 0.25322714, Global Avg Loss: 0.89394473, Time: 0.0211 Steps: 63670, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001939, Sample Num: 31024, Cur Loss: 0.10226482, Cur Avg Loss: 0.20813007, Log Avg loss: 0.17495122, Global Avg Loss: 0.89383182, Time: 0.0210 Steps: 63680, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001949, Sample Num: 31184, Cur Loss: 0.24739817, Cur Avg Loss: 0.20824263, Log Avg loss: 0.23006865, Global Avg Loss: 0.89372760, Time: 0.0210 Steps: 63690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001959, Sample Num: 31344, Cur Loss: 0.30439109, Cur Avg Loss: 0.20842816, Log Avg loss: 0.24458716, Global Avg Loss: 0.89362570, Time: 0.0210 Steps: 63700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001969, Sample Num: 31504, Cur Loss: 0.42800674, Cur Avg Loss: 0.20892140, Log Avg loss: 0.30554772, Global Avg Loss: 0.89353339, Time: 0.0210 Steps: 63710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001979, Sample Num: 31664, Cur Loss: 0.37627771, Cur Avg Loss: 0.20892989, Log Avg loss: 0.21060065, Global Avg Loss: 0.89342621, Time: 0.0211 Steps: 63720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001989, Sample Num: 31824, Cur Loss: 0.13537905, Cur Avg Loss: 0.20880093, Log Avg loss: 0.18327972, Global Avg Loss: 0.89331478, Time: 0.0210 Steps: 63730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001999, Sample Num: 31984, Cur Loss: 0.07196692, Cur Avg Loss: 0.20877183, Log Avg loss: 0.20298395, Global Avg Loss: 0.89320648, Time: 0.0210 Steps: 63740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 002009, Sample Num: 32144, Cur Loss: 0.15525818, Cur Avg Loss: 0.20897443, Log Avg loss: 0.24947526, Global Avg Loss: 0.89310550, Time: 0.0210 Steps: 63750, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002019, Sample Num: 32304, Cur Loss: 0.14495732, Cur Avg Loss: 0.20918013, Log Avg loss: 0.25050458, Global Avg Loss: 0.89300472, Time: 0.0210 Steps: 63760, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002029, Sample Num: 32464, Cur Loss: 0.09006836, Cur Avg Loss: 0.20934077, Log Avg loss: 0.24177460, Global Avg Loss: 0.89290260, Time: 0.0210 Steps: 63770, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002039, Sample Num: 32624, Cur Loss: 0.23266853, Cur Avg Loss: 0.20952610, Log Avg loss: 0.24712974, Global Avg Loss: 0.89280135, Time: 0.0211 Steps: 63780, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002049, Sample Num: 32784, Cur Loss: 0.20273978, Cur Avg Loss: 0.20985466, Log Avg loss: 0.27684651, Global Avg Loss: 0.89270479, Time: 0.0247 Steps: 63790, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002059, Sample Num: 32944, Cur Loss: 0.34907404, Cur Avg Loss: 0.20980354, Log Avg loss: 0.19932906, Global Avg Loss: 0.89259611, Time: 0.0210 Steps: 63800, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002069, Sample Num: 33104, Cur Loss: 0.21984857, Cur Avg Loss: 0.20960554, Log Avg loss: 0.16883729, Global Avg Loss: 0.89248268, Time: 0.0210 Steps: 63810, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002079, Sample Num: 33264, Cur Loss: 0.25808933, Cur Avg Loss: 0.20951418, Log Avg loss: 0.19061328, Global Avg Loss: 0.89237271, Time: 0.0210 Steps: 63820, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002089, Sample Num: 33424, Cur Loss: 0.13294294, Cur Avg Loss: 0.20951218, Log Avg loss: 0.20909535, Global Avg Loss: 0.89226566, Time: 0.0210 Steps: 63830, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002099, Sample Num: 33584, Cur Loss: 0.12275004, Cur Avg Loss: 0.20959980, Log Avg loss: 0.22790396, Global Avg Loss: 0.89216159, Time: 0.0211 Steps: 63840, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002109, Sample Num: 33744, Cur Loss: 0.05822518, Cur Avg Loss: 0.20943853, Log Avg loss: 0.17558859, Global Avg Loss: 0.89204937, Time: 0.0210 Steps: 63850, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002119, Sample Num: 33904, Cur Loss: 0.20489019, Cur Avg Loss: 0.20920802, Log Avg loss: 0.16059247, Global Avg Loss: 0.89193483, Time: 0.0210 Steps: 63860, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002129, Sample Num: 34055, Cur Loss: 0.48688477, Cur Avg Loss: 0.20951000, Log Avg loss: 0.27350058, Global Avg Loss: 0.89183800, Time: 0.0106 Steps: 63870, Updated lr: 0.000040 ***** Running evaluation checkpoint-63870 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-63870 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.151670, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.352982, "eval_total_loss": 248.146482, "eval_mae": 0.478928, "eval_mse": 0.353052, "eval_r2": 0.775577, "eval_sp_statistic": 0.893131, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.916246, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.46123, "test_total_loss": 231.537683, "test_mae": 0.454127, "test_mse": 0.461368, "test_r2": 0.702229, "test_sp_statistic": 0.800754, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.841268, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8918379982317186, "train_cur_epoch_loss": 446.04679544456303, "train_cur_epoch_avg_loss": 0.2095100025573335, "train_cur_epoch_time": 45.15166974067688, "train_cur_epoch_avg_time": 0.02120792378613287, "epoch": 30, "step": 63870} ################################################## Training, Epoch: 0031, Batch: 000010, Sample Num: 160, Cur Loss: 0.27991974, Cur Avg Loss: 0.36564513, Log Avg loss: 0.36564513, Global Avg Loss: 0.89175563, Time: 0.0211 Steps: 63880, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000020, Sample Num: 320, Cur Loss: 0.12528086, Cur Avg Loss: 0.31056217, Log Avg loss: 0.25547922, Global Avg Loss: 0.89165604, Time: 0.0211 Steps: 63890, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000030, Sample Num: 480, Cur Loss: 0.22543311, Cur Avg Loss: 0.30019681, Log Avg loss: 0.27946610, Global Avg Loss: 0.89156023, Time: 0.0211 Steps: 63900, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000040, Sample Num: 640, Cur Loss: 0.29140243, Cur Avg Loss: 0.26319128, Log Avg loss: 0.15217466, Global Avg Loss: 0.89144454, Time: 0.0211 Steps: 63910, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000050, Sample Num: 800, Cur Loss: 0.20802085, Cur Avg Loss: 0.23960978, Log Avg loss: 0.14528378, Global Avg Loss: 0.89132781, Time: 0.0211 Steps: 63920, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000060, Sample Num: 960, Cur Loss: 0.05466945, Cur Avg Loss: 0.22235057, Log Avg loss: 0.13605452, Global Avg Loss: 0.89120967, Time: 0.0211 Steps: 63930, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000070, Sample Num: 1120, Cur Loss: 0.27438346, Cur Avg Loss: 0.22303828, Log Avg loss: 0.22716454, Global Avg Loss: 0.89110581, Time: 0.0211 Steps: 63940, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000080, Sample Num: 1280, Cur Loss: 0.11003385, Cur Avg Loss: 0.21632163, Log Avg loss: 0.16930510, Global Avg Loss: 0.89099294, Time: 0.0211 Steps: 63950, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000090, Sample Num: 1440, Cur Loss: 0.20003429, Cur Avg Loss: 0.20763960, Log Avg loss: 0.13818338, Global Avg Loss: 0.89087524, Time: 0.0212 Steps: 63960, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000100, Sample Num: 1600, Cur Loss: 0.12275945, Cur Avg Loss: 0.20559169, Log Avg loss: 0.18716052, Global Avg Loss: 0.89076524, Time: 0.0211 Steps: 63970, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000110, Sample Num: 1760, Cur Loss: 0.09421179, Cur Avg Loss: 0.20192705, Log Avg loss: 0.16528063, Global Avg Loss: 0.89065184, Time: 0.0211 Steps: 63980, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000120, Sample Num: 1920, Cur Loss: 0.30600166, Cur Avg Loss: 0.20427785, Log Avg loss: 0.23013667, Global Avg Loss: 0.89054862, Time: 0.0210 Steps: 63990, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000130, Sample Num: 2080, Cur Loss: 0.29207706, Cur Avg Loss: 0.20315590, Log Avg loss: 0.18969240, Global Avg Loss: 0.89043911, Time: 0.0211 Steps: 64000, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000140, Sample Num: 2240, Cur Loss: 0.18737696, Cur Avg Loss: 0.19993512, Log Avg loss: 0.15806505, Global Avg Loss: 0.89032470, Time: 0.0211 Steps: 64010, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000150, Sample Num: 2400, Cur Loss: 0.52180356, Cur Avg Loss: 0.20106201, Log Avg loss: 0.21683846, Global Avg Loss: 0.89021950, Time: 0.0211 Steps: 64020, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000160, Sample Num: 2560, Cur Loss: 0.17691547, Cur Avg Loss: 0.19769004, Log Avg loss: 0.14711053, Global Avg Loss: 0.89010344, Time: 0.0211 Steps: 64030, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000170, Sample Num: 2720, Cur Loss: 0.15321130, Cur Avg Loss: 0.19594638, Log Avg loss: 0.16804785, Global Avg Loss: 0.88999069, Time: 0.0211 Steps: 64040, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000180, Sample Num: 2880, Cur Loss: 0.16913091, Cur Avg Loss: 0.19491373, Log Avg loss: 0.17735865, Global Avg Loss: 0.88987943, Time: 0.0211 Steps: 64050, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000190, Sample Num: 3040, Cur Loss: 0.29857257, Cur Avg Loss: 0.19529200, Log Avg loss: 0.20210074, Global Avg Loss: 0.88977206, Time: 0.0211 Steps: 64060, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000200, Sample Num: 3200, Cur Loss: 0.11028743, Cur Avg Loss: 0.19871723, Log Avg loss: 0.26379658, Global Avg Loss: 0.88967436, Time: 0.0211 Steps: 64070, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000210, Sample Num: 3360, Cur Loss: 0.09766697, Cur Avg Loss: 0.19793986, Log Avg loss: 0.18239251, Global Avg Loss: 0.88956399, Time: 0.0211 Steps: 64080, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000220, Sample Num: 3520, Cur Loss: 0.17923629, Cur Avg Loss: 0.19649961, Log Avg loss: 0.16625445, Global Avg Loss: 0.88945113, Time: 0.0209 Steps: 64090, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000230, Sample Num: 3680, Cur Loss: 0.15258507, Cur Avg Loss: 0.19426914, Log Avg loss: 0.14519879, Global Avg Loss: 0.88933502, Time: 0.0209 Steps: 64100, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000240, Sample Num: 3840, Cur Loss: 0.20964375, Cur Avg Loss: 0.19341270, Log Avg loss: 0.17371457, Global Avg Loss: 0.88922340, Time: 0.0209 Steps: 64110, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000250, Sample Num: 4000, Cur Loss: 0.03470869, Cur Avg Loss: 0.19393792, Log Avg loss: 0.20654313, Global Avg Loss: 0.88911693, Time: 0.0209 Steps: 64120, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000260, Sample Num: 4160, Cur Loss: 0.10675939, Cur Avg Loss: 0.19162708, Log Avg loss: 0.13385600, Global Avg Loss: 0.88899916, Time: 0.0247 Steps: 64130, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000270, Sample Num: 4320, Cur Loss: 0.20662200, Cur Avg Loss: 0.19222093, Log Avg loss: 0.20766126, Global Avg Loss: 0.88889293, Time: 0.0209 Steps: 64140, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000280, Sample Num: 4480, Cur Loss: 0.17837512, Cur Avg Loss: 0.19126006, Log Avg loss: 0.16531634, Global Avg Loss: 0.88878014, Time: 0.0210 Steps: 64150, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000290, Sample Num: 4640, Cur Loss: 0.13007513, Cur Avg Loss: 0.19023003, Log Avg loss: 0.16138928, Global Avg Loss: 0.88866677, Time: 0.0209 Steps: 64160, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000300, Sample Num: 4800, Cur Loss: 0.20781511, Cur Avg Loss: 0.18960154, Log Avg loss: 0.17137527, Global Avg Loss: 0.88855499, Time: 0.0209 Steps: 64170, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000310, Sample Num: 4960, Cur Loss: 0.15712512, Cur Avg Loss: 0.18717979, Log Avg loss: 0.11452731, Global Avg Loss: 0.88843438, Time: 0.0209 Steps: 64180, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000320, Sample Num: 5120, Cur Loss: 0.38479620, Cur Avg Loss: 0.18912377, Log Avg loss: 0.24938733, Global Avg Loss: 0.88833483, Time: 0.0210 Steps: 64190, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000330, Sample Num: 5280, Cur Loss: 0.34176233, Cur Avg Loss: 0.19050234, Log Avg loss: 0.23461650, Global Avg Loss: 0.88823300, Time: 0.0208 Steps: 64200, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000340, Sample Num: 5440, Cur Loss: 0.11588092, Cur Avg Loss: 0.19172645, Log Avg loss: 0.23212218, Global Avg Loss: 0.88813082, Time: 0.0209 Steps: 64210, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000350, Sample Num: 5600, Cur Loss: 0.15836024, Cur Avg Loss: 0.19138037, Log Avg loss: 0.17961349, Global Avg Loss: 0.88802049, Time: 0.0210 Steps: 64220, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000360, Sample Num: 5760, Cur Loss: 0.03155181, Cur Avg Loss: 0.18986082, Log Avg loss: 0.13667656, Global Avg Loss: 0.88790352, Time: 0.0209 Steps: 64230, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000370, Sample Num: 5920, Cur Loss: 0.11294185, Cur Avg Loss: 0.18987533, Log Avg loss: 0.19039792, Global Avg Loss: 0.88779494, Time: 0.0208 Steps: 64240, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000380, Sample Num: 6080, Cur Loss: 0.06131978, Cur Avg Loss: 0.18843813, Log Avg loss: 0.13526148, Global Avg Loss: 0.88767781, Time: 0.0208 Steps: 64250, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000390, Sample Num: 6240, Cur Loss: 0.26721680, Cur Avg Loss: 0.18864828, Log Avg loss: 0.19663407, Global Avg Loss: 0.88757027, Time: 0.0208 Steps: 64260, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000400, Sample Num: 6400, Cur Loss: 0.48329622, Cur Avg Loss: 0.19078041, Log Avg loss: 0.27393358, Global Avg Loss: 0.88747480, Time: 0.0208 Steps: 64270, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000410, Sample Num: 6560, Cur Loss: 0.32873145, Cur Avg Loss: 0.19305407, Log Avg loss: 0.28400051, Global Avg Loss: 0.88738091, Time: 0.0209 Steps: 64280, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000420, Sample Num: 6720, Cur Loss: 0.06293555, Cur Avg Loss: 0.19354076, Log Avg loss: 0.21349480, Global Avg Loss: 0.88727609, Time: 0.0209 Steps: 64290, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000430, Sample Num: 6880, Cur Loss: 0.32556149, Cur Avg Loss: 0.19543167, Log Avg loss: 0.27484989, Global Avg Loss: 0.88718085, Time: 0.0208 Steps: 64300, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000440, Sample Num: 7040, Cur Loss: 0.27364048, Cur Avg Loss: 0.19602486, Log Avg loss: 0.22153226, Global Avg Loss: 0.88707734, Time: 0.0208 Steps: 64310, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000450, Sample Num: 7200, Cur Loss: 0.34517050, Cur Avg Loss: 0.19661610, Log Avg loss: 0.22263057, Global Avg Loss: 0.88697404, Time: 0.0209 Steps: 64320, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000460, Sample Num: 7360, Cur Loss: 0.19369163, Cur Avg Loss: 0.19629011, Log Avg loss: 0.18162065, Global Avg Loss: 0.88686439, Time: 0.0208 Steps: 64330, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000470, Sample Num: 7520, Cur Loss: 0.42878994, Cur Avg Loss: 0.19573271, Log Avg loss: 0.17009231, Global Avg Loss: 0.88675299, Time: 0.0207 Steps: 64340, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000480, Sample Num: 7680, Cur Loss: 0.05289143, Cur Avg Loss: 0.19432217, Log Avg loss: 0.12802685, Global Avg Loss: 0.88663508, Time: 0.0208 Steps: 64350, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000490, Sample Num: 7840, Cur Loss: 0.25350514, Cur Avg Loss: 0.19455492, Log Avg loss: 0.20572674, Global Avg Loss: 0.88652929, Time: 0.0209 Steps: 64360, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000500, Sample Num: 8000, Cur Loss: 0.24547204, Cur Avg Loss: 0.19509927, Log Avg loss: 0.22177237, Global Avg Loss: 0.88642601, Time: 0.0209 Steps: 64370, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000510, Sample Num: 8160, Cur Loss: 0.12296927, Cur Avg Loss: 0.19463337, Log Avg loss: 0.17133828, Global Avg Loss: 0.88631494, Time: 0.0210 Steps: 64380, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000520, Sample Num: 8320, Cur Loss: 0.41680884, Cur Avg Loss: 0.19584414, Log Avg loss: 0.25759346, Global Avg Loss: 0.88621730, Time: 0.0208 Steps: 64390, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000530, Sample Num: 8480, Cur Loss: 0.16553967, Cur Avg Loss: 0.19510081, Log Avg loss: 0.15644778, Global Avg Loss: 0.88610398, Time: 0.0207 Steps: 64400, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000540, Sample Num: 8640, Cur Loss: 0.22124779, Cur Avg Loss: 0.19507192, Log Avg loss: 0.19354053, Global Avg Loss: 0.88599646, Time: 0.0208 Steps: 64410, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000550, Sample Num: 8800, Cur Loss: 0.05767220, Cur Avg Loss: 0.19450606, Log Avg loss: 0.16394950, Global Avg Loss: 0.88588437, Time: 0.0208 Steps: 64420, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000560, Sample Num: 8960, Cur Loss: 0.22603731, Cur Avg Loss: 0.19431308, Log Avg loss: 0.18369960, Global Avg Loss: 0.88577539, Time: 0.0208 Steps: 64430, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000570, Sample Num: 9120, Cur Loss: 0.12425685, Cur Avg Loss: 0.19370622, Log Avg loss: 0.15972166, Global Avg Loss: 0.88566272, Time: 0.0207 Steps: 64440, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000580, Sample Num: 9280, Cur Loss: 0.36721525, Cur Avg Loss: 0.19379718, Log Avg loss: 0.19898200, Global Avg Loss: 0.88555617, Time: 0.0207 Steps: 64450, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000590, Sample Num: 9440, Cur Loss: 0.28617021, Cur Avg Loss: 0.19399957, Log Avg loss: 0.20573821, Global Avg Loss: 0.88545071, Time: 0.0207 Steps: 64460, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000600, Sample Num: 9600, Cur Loss: 0.16012040, Cur Avg Loss: 0.19363174, Log Avg loss: 0.17193011, Global Avg Loss: 0.88534003, Time: 0.0207 Steps: 64470, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000610, Sample Num: 9760, Cur Loss: 0.31971976, Cur Avg Loss: 0.19286701, Log Avg loss: 0.14698324, Global Avg Loss: 0.88522552, Time: 0.0207 Steps: 64480, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000620, Sample Num: 9920, Cur Loss: 0.12909447, Cur Avg Loss: 0.19196896, Log Avg loss: 0.13718781, Global Avg Loss: 0.88510953, Time: 0.0208 Steps: 64490, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000630, Sample Num: 10080, Cur Loss: 0.10227862, Cur Avg Loss: 0.19163095, Log Avg loss: 0.17067439, Global Avg Loss: 0.88499877, Time: 0.0208 Steps: 64500, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000640, Sample Num: 10240, Cur Loss: 0.04808811, Cur Avg Loss: 0.19154611, Log Avg loss: 0.18620097, Global Avg Loss: 0.88489044, Time: 0.0208 Steps: 64510, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000650, Sample Num: 10400, Cur Loss: 0.09261777, Cur Avg Loss: 0.19115948, Log Avg loss: 0.16641495, Global Avg Loss: 0.88477909, Time: 0.0208 Steps: 64520, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000660, Sample Num: 10560, Cur Loss: 0.04352777, Cur Avg Loss: 0.19121986, Log Avg loss: 0.19514492, Global Avg Loss: 0.88467222, Time: 0.0209 Steps: 64530, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000670, Sample Num: 10720, Cur Loss: 0.11672834, Cur Avg Loss: 0.19127679, Log Avg loss: 0.19503397, Global Avg Loss: 0.88456536, Time: 0.0208 Steps: 64540, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000680, Sample Num: 10880, Cur Loss: 0.11393715, Cur Avg Loss: 0.19196282, Log Avg loss: 0.23792674, Global Avg Loss: 0.88446518, Time: 0.0208 Steps: 64550, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000690, Sample Num: 11040, Cur Loss: 0.14194545, Cur Avg Loss: 0.19203219, Log Avg loss: 0.19674947, Global Avg Loss: 0.88435866, Time: 0.0208 Steps: 64560, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000700, Sample Num: 11200, Cur Loss: 0.19072990, Cur Avg Loss: 0.19245579, Log Avg loss: 0.22168458, Global Avg Loss: 0.88425603, Time: 0.0208 Steps: 64570, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000710, Sample Num: 11360, Cur Loss: 0.41640940, Cur Avg Loss: 0.19356718, Log Avg loss: 0.27136434, Global Avg Loss: 0.88416113, Time: 0.0208 Steps: 64580, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000720, Sample Num: 11520, Cur Loss: 0.03523096, Cur Avg Loss: 0.19308435, Log Avg loss: 0.15880328, Global Avg Loss: 0.88404883, Time: 0.0208 Steps: 64590, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000730, Sample Num: 11680, Cur Loss: 0.10546996, Cur Avg Loss: 0.19330370, Log Avg loss: 0.20909694, Global Avg Loss: 0.88394434, Time: 0.0208 Steps: 64600, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000740, Sample Num: 11840, Cur Loss: 0.46800658, Cur Avg Loss: 0.19383416, Log Avg loss: 0.23255756, Global Avg Loss: 0.88384353, Time: 0.0208 Steps: 64610, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000750, Sample Num: 12000, Cur Loss: 0.25610429, Cur Avg Loss: 0.19450859, Log Avg loss: 0.24441629, Global Avg Loss: 0.88374457, Time: 0.0208 Steps: 64620, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000760, Sample Num: 12160, Cur Loss: 0.09451018, Cur Avg Loss: 0.19635703, Log Avg loss: 0.33499056, Global Avg Loss: 0.88365967, Time: 0.0209 Steps: 64630, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000770, Sample Num: 12320, Cur Loss: 0.23374391, Cur Avg Loss: 0.19621446, Log Avg loss: 0.18537877, Global Avg Loss: 0.88355164, Time: 0.0246 Steps: 64640, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000780, Sample Num: 12480, Cur Loss: 0.18286347, Cur Avg Loss: 0.19606653, Log Avg loss: 0.18467589, Global Avg Loss: 0.88344354, Time: 0.0209 Steps: 64650, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000790, Sample Num: 12640, Cur Loss: 0.07065856, Cur Avg Loss: 0.19512238, Log Avg loss: 0.12147906, Global Avg Loss: 0.88332570, Time: 0.0209 Steps: 64660, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000800, Sample Num: 12800, Cur Loss: 0.12999134, Cur Avg Loss: 0.19516099, Log Avg loss: 0.19821066, Global Avg Loss: 0.88321976, Time: 0.0209 Steps: 64670, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000810, Sample Num: 12960, Cur Loss: 0.20019825, Cur Avg Loss: 0.19548766, Log Avg loss: 0.22162175, Global Avg Loss: 0.88311747, Time: 0.0209 Steps: 64680, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000820, Sample Num: 13120, Cur Loss: 0.08695196, Cur Avg Loss: 0.19514602, Log Avg loss: 0.16747312, Global Avg Loss: 0.88300684, Time: 0.0209 Steps: 64690, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000830, Sample Num: 13280, Cur Loss: 0.17816143, Cur Avg Loss: 0.19470433, Log Avg loss: 0.15848556, Global Avg Loss: 0.88289486, Time: 0.0209 Steps: 64700, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000840, Sample Num: 13440, Cur Loss: 0.16519399, Cur Avg Loss: 0.19475375, Log Avg loss: 0.19885585, Global Avg Loss: 0.88278915, Time: 0.0208 Steps: 64710, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000850, Sample Num: 13600, Cur Loss: 0.19515443, Cur Avg Loss: 0.19435932, Log Avg loss: 0.16122723, Global Avg Loss: 0.88267766, Time: 0.0209 Steps: 64720, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000860, Sample Num: 13760, Cur Loss: 0.25595909, Cur Avg Loss: 0.19419868, Log Avg loss: 0.18054368, Global Avg Loss: 0.88256919, Time: 0.0209 Steps: 64730, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000870, Sample Num: 13920, Cur Loss: 0.17324542, Cur Avg Loss: 0.19480827, Log Avg loss: 0.24723305, Global Avg Loss: 0.88247106, Time: 0.0209 Steps: 64740, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000880, Sample Num: 14080, Cur Loss: 0.30224332, Cur Avg Loss: 0.19528565, Log Avg loss: 0.23681762, Global Avg Loss: 0.88237134, Time: 0.0209 Steps: 64750, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000890, Sample Num: 14240, Cur Loss: 0.07877490, Cur Avg Loss: 0.19557258, Log Avg loss: 0.22082295, Global Avg Loss: 0.88226919, Time: 0.0208 Steps: 64760, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000900, Sample Num: 14400, Cur Loss: 0.11075406, Cur Avg Loss: 0.19598542, Log Avg loss: 0.23272808, Global Avg Loss: 0.88216890, Time: 0.0209 Steps: 64770, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000910, Sample Num: 14560, Cur Loss: 0.14663529, Cur Avg Loss: 0.19596719, Log Avg loss: 0.19432600, Global Avg Loss: 0.88206272, Time: 0.0208 Steps: 64780, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000920, Sample Num: 14720, Cur Loss: 0.30551359, Cur Avg Loss: 0.19560309, Log Avg loss: 0.16246998, Global Avg Loss: 0.88195166, Time: 0.0208 Steps: 64790, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000930, Sample Num: 14880, Cur Loss: 0.16906755, Cur Avg Loss: 0.19520387, Log Avg loss: 0.15847601, Global Avg Loss: 0.88184001, Time: 0.0210 Steps: 64800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000940, Sample Num: 15040, Cur Loss: 0.15614620, Cur Avg Loss: 0.19478920, Log Avg loss: 0.15622497, Global Avg Loss: 0.88172805, Time: 0.0208 Steps: 64810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000950, Sample Num: 15200, Cur Loss: 0.56568027, Cur Avg Loss: 0.19569452, Log Avg loss: 0.28079479, Global Avg Loss: 0.88163534, Time: 0.0208 Steps: 64820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000960, Sample Num: 15360, Cur Loss: 0.12037569, Cur Avg Loss: 0.19551663, Log Avg loss: 0.17861649, Global Avg Loss: 0.88152690, Time: 0.0208 Steps: 64830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000970, Sample Num: 15520, Cur Loss: 0.16040678, Cur Avg Loss: 0.19564230, Log Avg loss: 0.20770734, Global Avg Loss: 0.88142298, Time: 0.0209 Steps: 64840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000980, Sample Num: 15680, Cur Loss: 0.08465700, Cur Avg Loss: 0.19577537, Log Avg loss: 0.20868320, Global Avg Loss: 0.88131924, Time: 0.0208 Steps: 64850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000990, Sample Num: 15840, Cur Loss: 0.08453301, Cur Avg Loss: 0.19608015, Log Avg loss: 0.22594784, Global Avg Loss: 0.88121820, Time: 0.0208 Steps: 64860, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001000, Sample Num: 16000, Cur Loss: 0.27889627, Cur Avg Loss: 0.19625729, Log Avg loss: 0.21379484, Global Avg Loss: 0.88111531, Time: 0.0209 Steps: 64870, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001010, Sample Num: 16160, Cur Loss: 0.33887121, Cur Avg Loss: 0.19579316, Log Avg loss: 0.14937985, Global Avg Loss: 0.88100253, Time: 0.0209 Steps: 64880, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001020, Sample Num: 16320, Cur Loss: 0.23765731, Cur Avg Loss: 0.19648971, Log Avg loss: 0.26684165, Global Avg Loss: 0.88090788, Time: 0.0208 Steps: 64890, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001030, Sample Num: 16480, Cur Loss: 0.79774964, Cur Avg Loss: 0.19671498, Log Avg loss: 0.21969172, Global Avg Loss: 0.88080600, Time: 0.0208 Steps: 64900, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001040, Sample Num: 16640, Cur Loss: 0.35122865, Cur Avg Loss: 0.19699009, Log Avg loss: 0.22532637, Global Avg Loss: 0.88070502, Time: 0.0208 Steps: 64910, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001050, Sample Num: 16800, Cur Loss: 0.52258265, Cur Avg Loss: 0.19728251, Log Avg loss: 0.22769446, Global Avg Loss: 0.88060443, Time: 0.0208 Steps: 64920, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001060, Sample Num: 16960, Cur Loss: 0.09761690, Cur Avg Loss: 0.19794939, Log Avg loss: 0.26797152, Global Avg Loss: 0.88051008, Time: 0.0208 Steps: 64930, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001070, Sample Num: 17120, Cur Loss: 0.36127380, Cur Avg Loss: 0.19865666, Log Avg loss: 0.27362768, Global Avg Loss: 0.88041662, Time: 0.0208 Steps: 64940, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001080, Sample Num: 17280, Cur Loss: 0.12856811, Cur Avg Loss: 0.19855867, Log Avg loss: 0.18807351, Global Avg Loss: 0.88031003, Time: 0.0208 Steps: 64950, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001090, Sample Num: 17440, Cur Loss: 0.29140124, Cur Avg Loss: 0.19814942, Log Avg loss: 0.15395111, Global Avg Loss: 0.88019821, Time: 0.0208 Steps: 64960, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001100, Sample Num: 17600, Cur Loss: 1.23552799, Cur Avg Loss: 0.19948905, Log Avg loss: 0.34550882, Global Avg Loss: 0.88011591, Time: 0.0208 Steps: 64970, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001110, Sample Num: 17760, Cur Loss: 0.41870633, Cur Avg Loss: 0.19972008, Log Avg loss: 0.22513321, Global Avg Loss: 0.88001512, Time: 0.0208 Steps: 64980, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001120, Sample Num: 17920, Cur Loss: 0.18787341, Cur Avg Loss: 0.20014109, Log Avg loss: 0.24687330, Global Avg Loss: 0.87991769, Time: 0.0208 Steps: 64990, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001130, Sample Num: 18080, Cur Loss: 0.05338361, Cur Avg Loss: 0.20003218, Log Avg loss: 0.18783348, Global Avg Loss: 0.87981122, Time: 0.0208 Steps: 65000, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001140, Sample Num: 18240, Cur Loss: 0.21665509, Cur Avg Loss: 0.20027503, Log Avg loss: 0.22771692, Global Avg Loss: 0.87971091, Time: 0.0208 Steps: 65010, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001150, Sample Num: 18400, Cur Loss: 0.11353715, Cur Avg Loss: 0.20008037, Log Avg loss: 0.17788951, Global Avg Loss: 0.87960297, Time: 0.0208 Steps: 65020, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001160, Sample Num: 18560, Cur Loss: 0.18905726, Cur Avg Loss: 0.20049424, Log Avg loss: 0.24808949, Global Avg Loss: 0.87950586, Time: 0.0208 Steps: 65030, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001170, Sample Num: 18720, Cur Loss: 0.21555056, Cur Avg Loss: 0.20021118, Log Avg loss: 0.16737565, Global Avg Loss: 0.87939637, Time: 0.0208 Steps: 65040, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001180, Sample Num: 18880, Cur Loss: 0.21770309, Cur Avg Loss: 0.20022917, Log Avg loss: 0.20233441, Global Avg Loss: 0.87929229, Time: 0.0208 Steps: 65050, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001190, Sample Num: 19040, Cur Loss: 0.10292500, Cur Avg Loss: 0.20018329, Log Avg loss: 0.19477005, Global Avg Loss: 0.87918707, Time: 0.0208 Steps: 65060, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001200, Sample Num: 19200, Cur Loss: 0.26429549, Cur Avg Loss: 0.20074828, Log Avg loss: 0.26798167, Global Avg Loss: 0.87909314, Time: 0.0208 Steps: 65070, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001210, Sample Num: 19360, Cur Loss: 0.09615289, Cur Avg Loss: 0.20051199, Log Avg loss: 0.17215690, Global Avg Loss: 0.87898452, Time: 0.0208 Steps: 65080, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001220, Sample Num: 19520, Cur Loss: 0.13299786, Cur Avg Loss: 0.20045910, Log Avg loss: 0.19405947, Global Avg Loss: 0.87887929, Time: 0.0209 Steps: 65090, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001230, Sample Num: 19680, Cur Loss: 0.24288066, Cur Avg Loss: 0.20022161, Log Avg loss: 0.17124818, Global Avg Loss: 0.87877059, Time: 0.0208 Steps: 65100, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001240, Sample Num: 19840, Cur Loss: 0.41373524, Cur Avg Loss: 0.20038505, Log Avg loss: 0.22048813, Global Avg Loss: 0.87866949, Time: 0.0208 Steps: 65110, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001250, Sample Num: 20000, Cur Loss: 0.37196010, Cur Avg Loss: 0.20013627, Log Avg loss: 0.16928710, Global Avg Loss: 0.87856055, Time: 0.0208 Steps: 65120, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001260, Sample Num: 20160, Cur Loss: 0.26472029, Cur Avg Loss: 0.20068586, Log Avg loss: 0.26938454, Global Avg Loss: 0.87846702, Time: 0.0208 Steps: 65130, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001270, Sample Num: 20320, Cur Loss: 0.59957576, Cur Avg Loss: 0.20127137, Log Avg loss: 0.27504540, Global Avg Loss: 0.87837439, Time: 0.0208 Steps: 65140, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001280, Sample Num: 20480, Cur Loss: 0.56832266, Cur Avg Loss: 0.20156731, Log Avg loss: 0.23915176, Global Avg Loss: 0.87827627, Time: 0.0253 Steps: 65150, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001290, Sample Num: 20640, Cur Loss: 0.19784550, Cur Avg Loss: 0.20221312, Log Avg loss: 0.28487759, Global Avg Loss: 0.87818520, Time: 0.0208 Steps: 65160, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001300, Sample Num: 20800, Cur Loss: 0.24738246, Cur Avg Loss: 0.20305037, Log Avg loss: 0.31105494, Global Avg Loss: 0.87809818, Time: 0.0209 Steps: 65170, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001310, Sample Num: 20960, Cur Loss: 0.17227052, Cur Avg Loss: 0.20298251, Log Avg loss: 0.19416071, Global Avg Loss: 0.87799325, Time: 0.0208 Steps: 65180, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001320, Sample Num: 21120, Cur Loss: 0.39910236, Cur Avg Loss: 0.20347993, Log Avg loss: 0.26864177, Global Avg Loss: 0.87789978, Time: 0.0209 Steps: 65190, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001330, Sample Num: 21280, Cur Loss: 0.19586389, Cur Avg Loss: 0.20303860, Log Avg loss: 0.14478369, Global Avg Loss: 0.87778734, Time: 0.0209 Steps: 65200, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001340, Sample Num: 21440, Cur Loss: 0.35343900, Cur Avg Loss: 0.20273488, Log Avg loss: 0.16234067, Global Avg Loss: 0.87767762, Time: 0.0208 Steps: 65210, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001350, Sample Num: 21600, Cur Loss: 0.29518044, Cur Avg Loss: 0.20289962, Log Avg loss: 0.22497351, Global Avg Loss: 0.87757754, Time: 0.0208 Steps: 65220, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001360, Sample Num: 21760, Cur Loss: 0.19323532, Cur Avg Loss: 0.20248129, Log Avg loss: 0.14600748, Global Avg Loss: 0.87746539, Time: 0.0208 Steps: 65230, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001370, Sample Num: 21920, Cur Loss: 0.32307240, Cur Avg Loss: 0.20246539, Log Avg loss: 0.20030332, Global Avg Loss: 0.87736160, Time: 0.0209 Steps: 65240, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001380, Sample Num: 22080, Cur Loss: 0.60233879, Cur Avg Loss: 0.20293707, Log Avg loss: 0.26755669, Global Avg Loss: 0.87726814, Time: 0.0209 Steps: 65250, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001390, Sample Num: 22240, Cur Loss: 0.11248608, Cur Avg Loss: 0.20278546, Log Avg loss: 0.18186322, Global Avg Loss: 0.87716158, Time: 0.0209 Steps: 65260, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001400, Sample Num: 22400, Cur Loss: 0.45149148, Cur Avg Loss: 0.20280906, Log Avg loss: 0.20609006, Global Avg Loss: 0.87705877, Time: 0.0209 Steps: 65270, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001410, Sample Num: 22560, Cur Loss: 0.13133463, Cur Avg Loss: 0.20299443, Log Avg loss: 0.22894562, Global Avg Loss: 0.87695948, Time: 0.0209 Steps: 65280, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001420, Sample Num: 22720, Cur Loss: 0.07944597, Cur Avg Loss: 0.20271595, Log Avg loss: 0.16345109, Global Avg Loss: 0.87685020, Time: 0.0209 Steps: 65290, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001430, Sample Num: 22880, Cur Loss: 0.17545046, Cur Avg Loss: 0.20230874, Log Avg loss: 0.14448367, Global Avg Loss: 0.87673805, Time: 0.0208 Steps: 65300, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001440, Sample Num: 23040, Cur Loss: 0.25593829, Cur Avg Loss: 0.20252467, Log Avg loss: 0.23340361, Global Avg Loss: 0.87663954, Time: 0.0209 Steps: 65310, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001450, Sample Num: 23200, Cur Loss: 0.10510036, Cur Avg Loss: 0.20258849, Log Avg loss: 0.21177861, Global Avg Loss: 0.87653776, Time: 0.0208 Steps: 65320, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001460, Sample Num: 23360, Cur Loss: 0.33170936, Cur Avg Loss: 0.20294605, Log Avg loss: 0.25479205, Global Avg Loss: 0.87644259, Time: 0.0209 Steps: 65330, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001470, Sample Num: 23520, Cur Loss: 0.33086056, Cur Avg Loss: 0.20292162, Log Avg loss: 0.19935401, Global Avg Loss: 0.87633896, Time: 0.0209 Steps: 65340, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001480, Sample Num: 23680, Cur Loss: 0.17945085, Cur Avg Loss: 0.20284000, Log Avg loss: 0.19084183, Global Avg Loss: 0.87623406, Time: 0.0208 Steps: 65350, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001490, Sample Num: 23840, Cur Loss: 0.10411046, Cur Avg Loss: 0.20273969, Log Avg loss: 0.18789445, Global Avg Loss: 0.87612875, Time: 0.0208 Steps: 65360, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001500, Sample Num: 24000, Cur Loss: 0.07021006, Cur Avg Loss: 0.20222325, Log Avg loss: 0.12527406, Global Avg Loss: 0.87601389, Time: 0.0209 Steps: 65370, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001510, Sample Num: 24160, Cur Loss: 0.33412579, Cur Avg Loss: 0.20238607, Log Avg loss: 0.22680820, Global Avg Loss: 0.87591459, Time: 0.0208 Steps: 65380, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001520, Sample Num: 24320, Cur Loss: 0.93719900, Cur Avg Loss: 0.20263726, Log Avg loss: 0.24056716, Global Avg Loss: 0.87581743, Time: 0.0208 Steps: 65390, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001530, Sample Num: 24480, Cur Loss: 0.21290861, Cur Avg Loss: 0.20272812, Log Avg loss: 0.21653885, Global Avg Loss: 0.87571662, Time: 0.0208 Steps: 65400, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001540, Sample Num: 24640, Cur Loss: 0.32408655, Cur Avg Loss: 0.20268352, Log Avg loss: 0.19585992, Global Avg Loss: 0.87561268, Time: 0.0246 Steps: 65410, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001550, Sample Num: 24800, Cur Loss: 0.14140706, Cur Avg Loss: 0.20251260, Log Avg loss: 0.17619117, Global Avg Loss: 0.87550577, Time: 0.0209 Steps: 65420, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001560, Sample Num: 24960, Cur Loss: 0.24681786, Cur Avg Loss: 0.20259377, Log Avg loss: 0.21517431, Global Avg Loss: 0.87540485, Time: 0.0209 Steps: 65430, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001570, Sample Num: 25120, Cur Loss: 0.07854676, Cur Avg Loss: 0.20225728, Log Avg loss: 0.14976534, Global Avg Loss: 0.87529396, Time: 0.0209 Steps: 65440, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001580, Sample Num: 25280, Cur Loss: 0.08808728, Cur Avg Loss: 0.20205106, Log Avg loss: 0.16967458, Global Avg Loss: 0.87518615, Time: 0.0209 Steps: 65450, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001590, Sample Num: 25440, Cur Loss: 0.28133559, Cur Avg Loss: 0.20164593, Log Avg loss: 0.13763538, Global Avg Loss: 0.87507348, Time: 0.0210 Steps: 65460, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001600, Sample Num: 25600, Cur Loss: 0.30203524, Cur Avg Loss: 0.20142427, Log Avg loss: 0.16618082, Global Avg Loss: 0.87496520, Time: 0.0209 Steps: 65470, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001610, Sample Num: 25760, Cur Loss: 0.13168837, Cur Avg Loss: 0.20124632, Log Avg loss: 0.17277470, Global Avg Loss: 0.87485796, Time: 0.0209 Steps: 65480, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001620, Sample Num: 25920, Cur Loss: 0.15684246, Cur Avg Loss: 0.20113873, Log Avg loss: 0.18381674, Global Avg Loss: 0.87475245, Time: 0.0209 Steps: 65490, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001630, Sample Num: 26080, Cur Loss: 0.07205612, Cur Avg Loss: 0.20104031, Log Avg loss: 0.18509617, Global Avg Loss: 0.87464716, Time: 0.0209 Steps: 65500, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001640, Sample Num: 26240, Cur Loss: 0.19012937, Cur Avg Loss: 0.20107968, Log Avg loss: 0.20749623, Global Avg Loss: 0.87454532, Time: 0.0209 Steps: 65510, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001650, Sample Num: 26400, Cur Loss: 0.18377817, Cur Avg Loss: 0.20095628, Log Avg loss: 0.18071829, Global Avg Loss: 0.87443942, Time: 0.0209 Steps: 65520, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001660, Sample Num: 26560, Cur Loss: 0.43638635, Cur Avg Loss: 0.20130512, Log Avg loss: 0.25886375, Global Avg Loss: 0.87434548, Time: 0.0209 Steps: 65530, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001670, Sample Num: 26720, Cur Loss: 0.37198588, Cur Avg Loss: 0.20137514, Log Avg loss: 0.21299898, Global Avg Loss: 0.87424457, Time: 0.0209 Steps: 65540, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001680, Sample Num: 26880, Cur Loss: 0.13162045, Cur Avg Loss: 0.20109987, Log Avg loss: 0.15513026, Global Avg Loss: 0.87413487, Time: 0.0208 Steps: 65550, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001690, Sample Num: 27040, Cur Loss: 0.09840916, Cur Avg Loss: 0.20118165, Log Avg loss: 0.21492071, Global Avg Loss: 0.87403432, Time: 0.0209 Steps: 65560, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001700, Sample Num: 27200, Cur Loss: 0.14821886, Cur Avg Loss: 0.20114717, Log Avg loss: 0.19531907, Global Avg Loss: 0.87393081, Time: 0.0209 Steps: 65570, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001710, Sample Num: 27360, Cur Loss: 0.14526635, Cur Avg Loss: 0.20108650, Log Avg loss: 0.19077379, Global Avg Loss: 0.87382664, Time: 0.0208 Steps: 65580, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001720, Sample Num: 27520, Cur Loss: 0.11767878, Cur Avg Loss: 0.20097133, Log Avg loss: 0.18127641, Global Avg Loss: 0.87372105, Time: 0.0209 Steps: 65590, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001730, Sample Num: 27680, Cur Loss: 0.12809581, Cur Avg Loss: 0.20106920, Log Avg loss: 0.21790315, Global Avg Loss: 0.87362108, Time: 0.0209 Steps: 65600, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001740, Sample Num: 27840, Cur Loss: 0.12614360, Cur Avg Loss: 0.20086815, Log Avg loss: 0.16608580, Global Avg Loss: 0.87351324, Time: 0.0209 Steps: 65610, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001750, Sample Num: 28000, Cur Loss: 0.05107590, Cur Avg Loss: 0.20063355, Log Avg loss: 0.15981381, Global Avg Loss: 0.87340448, Time: 0.0209 Steps: 65620, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001760, Sample Num: 28160, Cur Loss: 0.20085624, Cur Avg Loss: 0.20117205, Log Avg loss: 0.29540947, Global Avg Loss: 0.87331641, Time: 0.0209 Steps: 65630, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001770, Sample Num: 28320, Cur Loss: 0.14836055, Cur Avg Loss: 0.20133788, Log Avg loss: 0.23052432, Global Avg Loss: 0.87321848, Time: 0.0209 Steps: 65640, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001780, Sample Num: 28480, Cur Loss: 0.22579499, Cur Avg Loss: 0.20121673, Log Avg loss: 0.17977238, Global Avg Loss: 0.87311285, Time: 0.0209 Steps: 65650, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001790, Sample Num: 28640, Cur Loss: 0.22466491, Cur Avg Loss: 0.20090870, Log Avg loss: 0.14607899, Global Avg Loss: 0.87300212, Time: 0.0209 Steps: 65660, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001800, Sample Num: 28800, Cur Loss: 0.07952724, Cur Avg Loss: 0.20108701, Log Avg loss: 0.23300535, Global Avg Loss: 0.87290467, Time: 0.0208 Steps: 65670, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001810, Sample Num: 28960, Cur Loss: 0.31049132, Cur Avg Loss: 0.20100729, Log Avg loss: 0.18665694, Global Avg Loss: 0.87280018, Time: 0.0208 Steps: 65680, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001820, Sample Num: 29120, Cur Loss: 0.26761842, Cur Avg Loss: 0.20112758, Log Avg loss: 0.22290054, Global Avg Loss: 0.87270125, Time: 0.0209 Steps: 65690, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001830, Sample Num: 29280, Cur Loss: 0.15529896, Cur Avg Loss: 0.20123800, Log Avg loss: 0.22133529, Global Avg Loss: 0.87260211, Time: 0.0208 Steps: 65700, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001840, Sample Num: 29440, Cur Loss: 0.22189535, Cur Avg Loss: 0.20138292, Log Avg loss: 0.22790301, Global Avg Loss: 0.87250400, Time: 0.0208 Steps: 65710, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001850, Sample Num: 29600, Cur Loss: 0.10930190, Cur Avg Loss: 0.20142203, Log Avg loss: 0.20861759, Global Avg Loss: 0.87240298, Time: 0.0209 Steps: 65720, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001860, Sample Num: 29760, Cur Loss: 0.11814915, Cur Avg Loss: 0.20143739, Log Avg loss: 0.20427946, Global Avg Loss: 0.87230133, Time: 0.0209 Steps: 65730, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001870, Sample Num: 29920, Cur Loss: 0.19309814, Cur Avg Loss: 0.20170069, Log Avg loss: 0.25067367, Global Avg Loss: 0.87220677, Time: 0.0209 Steps: 65740, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001880, Sample Num: 30080, Cur Loss: 0.25307712, Cur Avg Loss: 0.20158863, Log Avg loss: 0.18063477, Global Avg Loss: 0.87210159, Time: 0.0208 Steps: 65750, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001890, Sample Num: 30240, Cur Loss: 0.16528890, Cur Avg Loss: 0.20175532, Log Avg loss: 0.23309170, Global Avg Loss: 0.87200442, Time: 0.0208 Steps: 65760, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001900, Sample Num: 30400, Cur Loss: 0.04706440, Cur Avg Loss: 0.20204847, Log Avg loss: 0.25745480, Global Avg Loss: 0.87191098, Time: 0.0208 Steps: 65770, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001910, Sample Num: 30560, Cur Loss: 0.06270412, Cur Avg Loss: 0.20236536, Log Avg loss: 0.26257461, Global Avg Loss: 0.87181835, Time: 0.0209 Steps: 65780, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001920, Sample Num: 30720, Cur Loss: 0.17630266, Cur Avg Loss: 0.20234574, Log Avg loss: 0.19859694, Global Avg Loss: 0.87171602, Time: 0.0207 Steps: 65790, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001930, Sample Num: 30880, Cur Loss: 0.39265367, Cur Avg Loss: 0.20224833, Log Avg loss: 0.18354590, Global Avg Loss: 0.87161143, Time: 0.0208 Steps: 65800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001940, Sample Num: 31040, Cur Loss: 0.18534103, Cur Avg Loss: 0.20239529, Log Avg loss: 0.23075965, Global Avg Loss: 0.87151405, Time: 0.0208 Steps: 65810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001950, Sample Num: 31200, Cur Loss: 0.30568370, Cur Avg Loss: 0.20247897, Log Avg loss: 0.21871329, Global Avg Loss: 0.87141487, Time: 0.0208 Steps: 65820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001960, Sample Num: 31360, Cur Loss: 0.18802287, Cur Avg Loss: 0.20251493, Log Avg loss: 0.20952612, Global Avg Loss: 0.87131433, Time: 0.0209 Steps: 65830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001970, Sample Num: 31520, Cur Loss: 0.44989821, Cur Avg Loss: 0.20286055, Log Avg loss: 0.27060187, Global Avg Loss: 0.87122309, Time: 0.0208 Steps: 65840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001980, Sample Num: 31680, Cur Loss: 0.18504132, Cur Avg Loss: 0.20298371, Log Avg loss: 0.22724560, Global Avg Loss: 0.87112530, Time: 0.0209 Steps: 65850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001990, Sample Num: 31840, Cur Loss: 0.29249609, Cur Avg Loss: 0.20342110, Log Avg loss: 0.29002579, Global Avg Loss: 0.87103706, Time: 0.0208 Steps: 65860, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002000, Sample Num: 32000, Cur Loss: 0.57305312, Cur Avg Loss: 0.20349504, Log Avg loss: 0.21820758, Global Avg Loss: 0.87093795, Time: 0.0208 Steps: 65870, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002010, Sample Num: 32160, Cur Loss: 0.19004130, Cur Avg Loss: 0.20342210, Log Avg loss: 0.18883485, Global Avg Loss: 0.87083442, Time: 0.0208 Steps: 65880, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002020, Sample Num: 32320, Cur Loss: 0.38851517, Cur Avg Loss: 0.20345097, Log Avg loss: 0.20925364, Global Avg Loss: 0.87073401, Time: 0.0209 Steps: 65890, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002030, Sample Num: 32480, Cur Loss: 0.55231380, Cur Avg Loss: 0.20388504, Log Avg loss: 0.29156663, Global Avg Loss: 0.87064612, Time: 0.0207 Steps: 65900, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002040, Sample Num: 32640, Cur Loss: 0.05246378, Cur Avg Loss: 0.20420505, Log Avg loss: 0.26916820, Global Avg Loss: 0.87055487, Time: 0.0209 Steps: 65910, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002050, Sample Num: 32800, Cur Loss: 0.09451531, Cur Avg Loss: 0.20395753, Log Avg loss: 0.15346345, Global Avg Loss: 0.87044608, Time: 0.0246 Steps: 65920, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002060, Sample Num: 32960, Cur Loss: 0.19396576, Cur Avg Loss: 0.20388810, Log Avg loss: 0.18965412, Global Avg Loss: 0.87034282, Time: 0.0208 Steps: 65930, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002070, Sample Num: 33120, Cur Loss: 0.20748690, Cur Avg Loss: 0.20426136, Log Avg loss: 0.28115345, Global Avg Loss: 0.87025347, Time: 0.0209 Steps: 65940, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002080, Sample Num: 33280, Cur Loss: 0.13574597, Cur Avg Loss: 0.20404059, Log Avg loss: 0.15834191, Global Avg Loss: 0.87014553, Time: 0.0208 Steps: 65950, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002090, Sample Num: 33440, Cur Loss: 0.18054786, Cur Avg Loss: 0.20417469, Log Avg loss: 0.23206608, Global Avg Loss: 0.87004879, Time: 0.0209 Steps: 65960, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002100, Sample Num: 33600, Cur Loss: 0.09769076, Cur Avg Loss: 0.20393524, Log Avg loss: 0.15389168, Global Avg Loss: 0.86994023, Time: 0.0209 Steps: 65970, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002110, Sample Num: 33760, Cur Loss: 0.12320793, Cur Avg Loss: 0.20388728, Log Avg loss: 0.19381513, Global Avg Loss: 0.86983776, Time: 0.0209 Steps: 65980, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002120, Sample Num: 33920, Cur Loss: 0.08545823, Cur Avg Loss: 0.20363646, Log Avg loss: 0.15071302, Global Avg Loss: 0.86972878, Time: 0.0207 Steps: 65990, Updated lr: 0.000038 ***** Running evaluation checkpoint-65999 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-65999 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.610753, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.278615, "eval_total_loss": 195.866047, "eval_mae": 0.356794, "eval_mse": 0.278723, "eval_r2": 0.822825, "eval_sp_statistic": 0.880502, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.913203, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.794134, "test_total_loss": 398.655302, "test_mae": 0.684007, "test_mse": 0.794198, "test_r2": 0.487418, "test_sp_statistic": 0.800061, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.840338, "test_ps_pvalue": 0.0, "lr": 3.8360360360360366e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8696288601427098, "train_cur_epoch_loss": 432.94219349883497, "train_cur_epoch_avg_loss": 0.2033547174724448, "train_cur_epoch_time": 44.610753297805786, "train_cur_epoch_avg_time": 0.020953853122501545, "epoch": 31, "step": 65999} ################################################## Training, Epoch: 0032, Batch: 000001, Sample Num: 16, Cur Loss: 0.41883600, Cur Avg Loss: 0.41883600, Log Avg loss: 0.16517373, Global Avg Loss: 0.86962203, Time: 0.0248 Steps: 66000, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000011, Sample Num: 176, Cur Loss: 0.12635000, Cur Avg Loss: 0.15029503, Log Avg loss: 0.12344093, Global Avg Loss: 0.86950899, Time: 0.0209 Steps: 66010, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000021, Sample Num: 336, Cur Loss: 0.11611092, Cur Avg Loss: 0.14702723, Log Avg loss: 0.14343265, Global Avg Loss: 0.86939901, Time: 0.0208 Steps: 66020, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000031, Sample Num: 496, Cur Loss: 0.22431669, Cur Avg Loss: 0.15060748, Log Avg loss: 0.15812600, Global Avg Loss: 0.86929129, Time: 0.0208 Steps: 66030, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000041, Sample Num: 656, Cur Loss: 0.04613628, Cur Avg Loss: 0.15093659, Log Avg loss: 0.15195684, Global Avg Loss: 0.86918267, Time: 0.0208 Steps: 66040, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000051, Sample Num: 816, Cur Loss: 0.04060324, Cur Avg Loss: 0.14019908, Log Avg loss: 0.09617527, Global Avg Loss: 0.86906564, Time: 0.0208 Steps: 66050, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000061, Sample Num: 976, Cur Loss: 0.25429547, Cur Avg Loss: 0.15510422, Log Avg loss: 0.23112043, Global Avg Loss: 0.86896907, Time: 0.0208 Steps: 66060, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000071, Sample Num: 1136, Cur Loss: 0.16286555, Cur Avg Loss: 0.16350362, Log Avg loss: 0.21474000, Global Avg Loss: 0.86887005, Time: 0.0208 Steps: 66070, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000081, Sample Num: 1296, Cur Loss: 0.14214158, Cur Avg Loss: 0.16626781, Log Avg loss: 0.18589353, Global Avg Loss: 0.86876669, Time: 0.0208 Steps: 66080, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000091, Sample Num: 1456, Cur Loss: 0.07459342, Cur Avg Loss: 0.16990576, Log Avg loss: 0.19937313, Global Avg Loss: 0.86866540, Time: 0.0208 Steps: 66090, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000101, Sample Num: 1616, Cur Loss: 0.11998944, Cur Avg Loss: 0.16947506, Log Avg loss: 0.16555574, Global Avg Loss: 0.86855903, Time: 0.0209 Steps: 66100, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000111, Sample Num: 1776, Cur Loss: 0.26857346, Cur Avg Loss: 0.18277460, Log Avg loss: 0.31709989, Global Avg Loss: 0.86847562, Time: 0.0209 Steps: 66110, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000121, Sample Num: 1936, Cur Loss: 0.07281625, Cur Avg Loss: 0.18431394, Log Avg loss: 0.20140061, Global Avg Loss: 0.86837473, Time: 0.0208 Steps: 66120, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000131, Sample Num: 2096, Cur Loss: 0.05956383, Cur Avg Loss: 0.18287947, Log Avg loss: 0.16552247, Global Avg Loss: 0.86826845, Time: 0.0208 Steps: 66130, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000141, Sample Num: 2256, Cur Loss: 0.17644753, Cur Avg Loss: 0.18043452, Log Avg loss: 0.14840563, Global Avg Loss: 0.86815961, Time: 0.0208 Steps: 66140, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000151, Sample Num: 2416, Cur Loss: 0.09993611, Cur Avg Loss: 0.17975301, Log Avg loss: 0.17014375, Global Avg Loss: 0.86805409, Time: 0.0208 Steps: 66150, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000161, Sample Num: 2576, Cur Loss: 0.11802945, Cur Avg Loss: 0.17568067, Log Avg loss: 0.11418835, Global Avg Loss: 0.86794014, Time: 0.0208 Steps: 66160, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000171, Sample Num: 2736, Cur Loss: 0.14357847, Cur Avg Loss: 0.17789421, Log Avg loss: 0.21353214, Global Avg Loss: 0.86784124, Time: 0.0208 Steps: 66170, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000181, Sample Num: 2896, Cur Loss: 0.22326890, Cur Avg Loss: 0.17662959, Log Avg loss: 0.15500452, Global Avg Loss: 0.86773353, Time: 0.0208 Steps: 66180, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000191, Sample Num: 3056, Cur Loss: 0.16497472, Cur Avg Loss: 0.18445243, Log Avg loss: 0.32604585, Global Avg Loss: 0.86765169, Time: 0.0208 Steps: 66190, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000201, Sample Num: 3216, Cur Loss: 0.34601095, Cur Avg Loss: 0.18419911, Log Avg loss: 0.17936074, Global Avg Loss: 0.86754772, Time: 0.0209 Steps: 66200, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000211, Sample Num: 3376, Cur Loss: 0.12505446, Cur Avg Loss: 0.18513264, Log Avg loss: 0.20389655, Global Avg Loss: 0.86744749, Time: 0.0208 Steps: 66210, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000221, Sample Num: 3536, Cur Loss: 0.20270932, Cur Avg Loss: 0.18793275, Log Avg loss: 0.24701506, Global Avg Loss: 0.86735379, Time: 0.0208 Steps: 66220, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000231, Sample Num: 3696, Cur Loss: 0.49241507, Cur Avg Loss: 0.19018541, Log Avg loss: 0.23996916, Global Avg Loss: 0.86725907, Time: 0.0208 Steps: 66230, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000241, Sample Num: 3856, Cur Loss: 0.09829313, Cur Avg Loss: 0.18829933, Log Avg loss: 0.14473090, Global Avg Loss: 0.86714999, Time: 0.0208 Steps: 66240, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000251, Sample Num: 4016, Cur Loss: 0.17831913, Cur Avg Loss: 0.18761014, Log Avg loss: 0.17100073, Global Avg Loss: 0.86704491, Time: 0.0208 Steps: 66250, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000261, Sample Num: 4176, Cur Loss: 0.17602533, Cur Avg Loss: 0.18624814, Log Avg loss: 0.15206209, Global Avg Loss: 0.86693700, Time: 0.0209 Steps: 66260, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000271, Sample Num: 4336, Cur Loss: 0.14067495, Cur Avg Loss: 0.18599010, Log Avg loss: 0.17925514, Global Avg Loss: 0.86683323, Time: 0.0208 Steps: 66270, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000281, Sample Num: 4496, Cur Loss: 0.26686150, Cur Avg Loss: 0.18593652, Log Avg loss: 0.18448451, Global Avg Loss: 0.86673029, Time: 0.0208 Steps: 66280, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000291, Sample Num: 4656, Cur Loss: 0.11659466, Cur Avg Loss: 0.18315642, Log Avg loss: 0.10503551, Global Avg Loss: 0.86661538, Time: 0.0208 Steps: 66290, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000301, Sample Num: 4816, Cur Loss: 0.16702124, Cur Avg Loss: 0.18381899, Log Avg loss: 0.20309971, Global Avg Loss: 0.86651530, Time: 0.0208 Steps: 66300, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000311, Sample Num: 4976, Cur Loss: 0.17240521, Cur Avg Loss: 0.18302593, Log Avg loss: 0.15915488, Global Avg Loss: 0.86640863, Time: 0.0209 Steps: 66310, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000321, Sample Num: 5136, Cur Loss: 0.16867922, Cur Avg Loss: 0.18380330, Log Avg loss: 0.20797975, Global Avg Loss: 0.86630935, Time: 0.0208 Steps: 66320, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000331, Sample Num: 5296, Cur Loss: 0.10091485, Cur Avg Loss: 0.18240072, Log Avg loss: 0.13737789, Global Avg Loss: 0.86619945, Time: 0.0208 Steps: 66330, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000341, Sample Num: 5456, Cur Loss: 0.06587087, Cur Avg Loss: 0.18302206, Log Avg loss: 0.20358815, Global Avg Loss: 0.86609957, Time: 0.0208 Steps: 66340, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000351, Sample Num: 5616, Cur Loss: 0.05816071, Cur Avg Loss: 0.18351892, Log Avg loss: 0.20046188, Global Avg Loss: 0.86599925, Time: 0.0208 Steps: 66350, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000361, Sample Num: 5776, Cur Loss: 0.18052763, Cur Avg Loss: 0.18353858, Log Avg loss: 0.18422890, Global Avg Loss: 0.86589651, Time: 0.0208 Steps: 66360, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000371, Sample Num: 5936, Cur Loss: 0.13731101, Cur Avg Loss: 0.18396109, Log Avg loss: 0.19921345, Global Avg Loss: 0.86579606, Time: 0.0208 Steps: 66370, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000381, Sample Num: 6096, Cur Loss: 0.14269431, Cur Avg Loss: 0.18349480, Log Avg loss: 0.16619554, Global Avg Loss: 0.86569067, Time: 0.0207 Steps: 66380, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000391, Sample Num: 6256, Cur Loss: 0.30728337, Cur Avg Loss: 0.18528304, Log Avg loss: 0.25341511, Global Avg Loss: 0.86559845, Time: 0.0209 Steps: 66390, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000401, Sample Num: 6416, Cur Loss: 0.10377488, Cur Avg Loss: 0.18592063, Log Avg loss: 0.21085035, Global Avg Loss: 0.86549984, Time: 0.0208 Steps: 66400, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000411, Sample Num: 6576, Cur Loss: 0.21235268, Cur Avg Loss: 0.18525328, Log Avg loss: 0.15849249, Global Avg Loss: 0.86539338, Time: 0.0209 Steps: 66410, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000421, Sample Num: 6736, Cur Loss: 0.05761922, Cur Avg Loss: 0.18630956, Log Avg loss: 0.22972260, Global Avg Loss: 0.86529767, Time: 0.0208 Steps: 66420, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000431, Sample Num: 6896, Cur Loss: 0.07361171, Cur Avg Loss: 0.18707894, Log Avg loss: 0.21946966, Global Avg Loss: 0.86520045, Time: 0.0208 Steps: 66430, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000441, Sample Num: 7056, Cur Loss: 0.06162140, Cur Avg Loss: 0.18722267, Log Avg loss: 0.19341775, Global Avg Loss: 0.86509934, Time: 0.0208 Steps: 66440, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000451, Sample Num: 7216, Cur Loss: 0.17671302, Cur Avg Loss: 0.18668729, Log Avg loss: 0.16307707, Global Avg Loss: 0.86499370, Time: 0.0208 Steps: 66450, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000461, Sample Num: 7376, Cur Loss: 0.40519249, Cur Avg Loss: 0.18833277, Log Avg loss: 0.26254379, Global Avg Loss: 0.86490305, Time: 0.0208 Steps: 66460, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000471, Sample Num: 7536, Cur Loss: 0.17719063, Cur Avg Loss: 0.18794590, Log Avg loss: 0.17011123, Global Avg Loss: 0.86479852, Time: 0.0208 Steps: 66470, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000481, Sample Num: 7696, Cur Loss: 0.42480761, Cur Avg Loss: 0.18723640, Log Avg loss: 0.15381866, Global Avg Loss: 0.86469157, Time: 0.0208 Steps: 66480, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000491, Sample Num: 7856, Cur Loss: 0.12194002, Cur Avg Loss: 0.18735181, Log Avg loss: 0.19290332, Global Avg Loss: 0.86459054, Time: 0.0208 Steps: 66490, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000501, Sample Num: 8016, Cur Loss: 0.04350115, Cur Avg Loss: 0.18640108, Log Avg loss: 0.13972021, Global Avg Loss: 0.86448154, Time: 0.0208 Steps: 66500, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000511, Sample Num: 8176, Cur Loss: 0.27481055, Cur Avg Loss: 0.18635544, Log Avg loss: 0.18406902, Global Avg Loss: 0.86437923, Time: 0.0209 Steps: 66510, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000521, Sample Num: 8336, Cur Loss: 0.12681337, Cur Avg Loss: 0.18712604, Log Avg loss: 0.22650364, Global Avg Loss: 0.86428334, Time: 0.0209 Steps: 66520, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000531, Sample Num: 8496, Cur Loss: 0.26381707, Cur Avg Loss: 0.18785981, Log Avg loss: 0.22608921, Global Avg Loss: 0.86418741, Time: 0.0208 Steps: 66530, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000541, Sample Num: 8656, Cur Loss: 0.23075989, Cur Avg Loss: 0.18862582, Log Avg loss: 0.22930085, Global Avg Loss: 0.86409200, Time: 0.0208 Steps: 66540, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000551, Sample Num: 8816, Cur Loss: 0.19402447, Cur Avg Loss: 0.19118555, Log Avg loss: 0.32966712, Global Avg Loss: 0.86401170, Time: 0.0209 Steps: 66550, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000561, Sample Num: 8976, Cur Loss: 0.30593377, Cur Avg Loss: 0.19173284, Log Avg loss: 0.22188829, Global Avg Loss: 0.86391522, Time: 0.0208 Steps: 66560, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000571, Sample Num: 9136, Cur Loss: 0.07689319, Cur Avg Loss: 0.19078400, Log Avg loss: 0.13755424, Global Avg Loss: 0.86380611, Time: 0.0212 Steps: 66570, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000581, Sample Num: 9296, Cur Loss: 0.09638825, Cur Avg Loss: 0.19058415, Log Avg loss: 0.17917263, Global Avg Loss: 0.86370328, Time: 0.0208 Steps: 66580, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000591, Sample Num: 9456, Cur Loss: 0.09173086, Cur Avg Loss: 0.18946832, Log Avg loss: 0.12463883, Global Avg Loss: 0.86359229, Time: 0.0208 Steps: 66590, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000601, Sample Num: 9616, Cur Loss: 0.13224611, Cur Avg Loss: 0.18945034, Log Avg loss: 0.18838729, Global Avg Loss: 0.86349091, Time: 0.0208 Steps: 66600, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000611, Sample Num: 9776, Cur Loss: 0.47149491, Cur Avg Loss: 0.18975943, Log Avg loss: 0.20833573, Global Avg Loss: 0.86339256, Time: 0.0209 Steps: 66610, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000621, Sample Num: 9936, Cur Loss: 0.03113893, Cur Avg Loss: 0.18905554, Log Avg loss: 0.14604815, Global Avg Loss: 0.86328488, Time: 0.0207 Steps: 66620, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000631, Sample Num: 10096, Cur Loss: 0.04735731, Cur Avg Loss: 0.18816142, Log Avg loss: 0.13263625, Global Avg Loss: 0.86317522, Time: 0.0209 Steps: 66630, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000641, Sample Num: 10256, Cur Loss: 0.15432641, Cur Avg Loss: 0.18996618, Log Avg loss: 0.30384702, Global Avg Loss: 0.86309129, Time: 0.0208 Steps: 66640, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000651, Sample Num: 10416, Cur Loss: 0.17243549, Cur Avg Loss: 0.19000076, Log Avg loss: 0.19221681, Global Avg Loss: 0.86299063, Time: 0.0208 Steps: 66650, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000661, Sample Num: 10576, Cur Loss: 0.14509472, Cur Avg Loss: 0.18903144, Log Avg loss: 0.12592901, Global Avg Loss: 0.86288006, Time: 0.0208 Steps: 66660, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000671, Sample Num: 10736, Cur Loss: 0.06648572, Cur Avg Loss: 0.18958745, Log Avg loss: 0.22633959, Global Avg Loss: 0.86278459, Time: 0.0208 Steps: 66670, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000681, Sample Num: 10896, Cur Loss: 0.10160954, Cur Avg Loss: 0.18904431, Log Avg loss: 0.15259998, Global Avg Loss: 0.86267808, Time: 0.0208 Steps: 66680, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000691, Sample Num: 11056, Cur Loss: 0.19614598, Cur Avg Loss: 0.18903141, Log Avg loss: 0.18815259, Global Avg Loss: 0.86257694, Time: 0.0208 Steps: 66690, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000701, Sample Num: 11216, Cur Loss: 0.66845053, Cur Avg Loss: 0.19008176, Log Avg loss: 0.26266120, Global Avg Loss: 0.86248699, Time: 0.0208 Steps: 66700, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000711, Sample Num: 11376, Cur Loss: 0.04165253, Cur Avg Loss: 0.18977149, Log Avg loss: 0.16802157, Global Avg Loss: 0.86238289, Time: 0.0208 Steps: 66710, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000721, Sample Num: 11536, Cur Loss: 0.06812215, Cur Avg Loss: 0.18958725, Log Avg loss: 0.17648728, Global Avg Loss: 0.86228009, Time: 0.0208 Steps: 66720, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000731, Sample Num: 11696, Cur Loss: 0.13940376, Cur Avg Loss: 0.18958255, Log Avg loss: 0.18924413, Global Avg Loss: 0.86217923, Time: 0.0208 Steps: 66730, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000741, Sample Num: 11856, Cur Loss: 0.11995645, Cur Avg Loss: 0.18972194, Log Avg loss: 0.19991127, Global Avg Loss: 0.86208000, Time: 0.0209 Steps: 66740, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000751, Sample Num: 12016, Cur Loss: 0.39893502, Cur Avg Loss: 0.19001136, Log Avg loss: 0.21145713, Global Avg Loss: 0.86198253, Time: 0.0208 Steps: 66750, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000761, Sample Num: 12176, Cur Loss: 0.23242380, Cur Avg Loss: 0.18953683, Log Avg loss: 0.15389967, Global Avg Loss: 0.86187646, Time: 0.0209 Steps: 66760, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000771, Sample Num: 12336, Cur Loss: 0.12936217, Cur Avg Loss: 0.18967643, Log Avg loss: 0.20030034, Global Avg Loss: 0.86177738, Time: 0.0245 Steps: 66770, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000781, Sample Num: 12496, Cur Loss: 0.05518351, Cur Avg Loss: 0.19043615, Log Avg loss: 0.24901046, Global Avg Loss: 0.86168562, Time: 0.0208 Steps: 66780, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000791, Sample Num: 12656, Cur Loss: 0.37610433, Cur Avg Loss: 0.19034450, Log Avg loss: 0.18318673, Global Avg Loss: 0.86158403, Time: 0.0212 Steps: 66790, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000801, Sample Num: 12816, Cur Loss: 0.19729654, Cur Avg Loss: 0.19039092, Log Avg loss: 0.19406253, Global Avg Loss: 0.86148411, Time: 0.0212 Steps: 66800, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000811, Sample Num: 12976, Cur Loss: 0.06615801, Cur Avg Loss: 0.19052090, Log Avg loss: 0.20093223, Global Avg Loss: 0.86138524, Time: 0.0212 Steps: 66810, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000821, Sample Num: 13136, Cur Loss: 0.05313134, Cur Avg Loss: 0.19021118, Log Avg loss: 0.16509269, Global Avg Loss: 0.86128103, Time: 0.0212 Steps: 66820, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000831, Sample Num: 13296, Cur Loss: 0.24107313, Cur Avg Loss: 0.19022641, Log Avg loss: 0.19147699, Global Avg Loss: 0.86118081, Time: 0.0208 Steps: 66830, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000841, Sample Num: 13456, Cur Loss: 0.15107836, Cur Avg Loss: 0.19053603, Log Avg loss: 0.21626525, Global Avg Loss: 0.86108432, Time: 0.0208 Steps: 66840, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000851, Sample Num: 13616, Cur Loss: 0.10507575, Cur Avg Loss: 0.18988813, Log Avg loss: 0.13540030, Global Avg Loss: 0.86097577, Time: 0.0207 Steps: 66850, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000861, Sample Num: 13776, Cur Loss: 0.08807924, Cur Avg Loss: 0.19001985, Log Avg loss: 0.20122887, Global Avg Loss: 0.86087709, Time: 0.0208 Steps: 66860, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000871, Sample Num: 13936, Cur Loss: 0.14475420, Cur Avg Loss: 0.18944457, Log Avg loss: 0.13991319, Global Avg Loss: 0.86076927, Time: 0.0208 Steps: 66870, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000881, Sample Num: 14096, Cur Loss: 0.11472233, Cur Avg Loss: 0.18933031, Log Avg loss: 0.17937782, Global Avg Loss: 0.86066739, Time: 0.0208 Steps: 66880, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000891, Sample Num: 14256, Cur Loss: 0.19931263, Cur Avg Loss: 0.18923531, Log Avg loss: 0.18086591, Global Avg Loss: 0.86056576, Time: 0.0208 Steps: 66890, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000901, Sample Num: 14416, Cur Loss: 0.17754692, Cur Avg Loss: 0.18933070, Log Avg loss: 0.19783003, Global Avg Loss: 0.86046670, Time: 0.0212 Steps: 66900, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000911, Sample Num: 14576, Cur Loss: 0.13952157, Cur Avg Loss: 0.18949442, Log Avg loss: 0.20424526, Global Avg Loss: 0.86036862, Time: 0.0212 Steps: 66910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000921, Sample Num: 14736, Cur Loss: 0.09261202, Cur Avg Loss: 0.18952027, Log Avg loss: 0.19187559, Global Avg Loss: 0.86026873, Time: 0.0210 Steps: 66920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000931, Sample Num: 14896, Cur Loss: 0.28541818, Cur Avg Loss: 0.18906354, Log Avg loss: 0.14699919, Global Avg Loss: 0.86016216, Time: 0.0210 Steps: 66930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000941, Sample Num: 15056, Cur Loss: 0.32750177, Cur Avg Loss: 0.18916162, Log Avg loss: 0.19829276, Global Avg Loss: 0.86006328, Time: 0.0209 Steps: 66940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000951, Sample Num: 15216, Cur Loss: 0.16548258, Cur Avg Loss: 0.18917436, Log Avg loss: 0.19037274, Global Avg Loss: 0.85996326, Time: 0.0210 Steps: 66950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000961, Sample Num: 15376, Cur Loss: 0.13002294, Cur Avg Loss: 0.19048656, Log Avg loss: 0.31527654, Global Avg Loss: 0.85988191, Time: 0.0209 Steps: 66960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000971, Sample Num: 15536, Cur Loss: 0.45079073, Cur Avg Loss: 0.19126196, Log Avg loss: 0.26577792, Global Avg Loss: 0.85979320, Time: 0.0210 Steps: 66970, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000981, Sample Num: 15696, Cur Loss: 0.18764496, Cur Avg Loss: 0.19073039, Log Avg loss: 0.13911490, Global Avg Loss: 0.85968560, Time: 0.0209 Steps: 66980, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000991, Sample Num: 15856, Cur Loss: 0.08565076, Cur Avg Loss: 0.19017766, Log Avg loss: 0.13595477, Global Avg Loss: 0.85957757, Time: 0.0209 Steps: 66990, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001001, Sample Num: 16016, Cur Loss: 0.06791473, Cur Avg Loss: 0.19022114, Log Avg loss: 0.19453078, Global Avg Loss: 0.85947831, Time: 0.0210 Steps: 67000, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001011, Sample Num: 16176, Cur Loss: 0.15520863, Cur Avg Loss: 0.19014078, Log Avg loss: 0.18209634, Global Avg Loss: 0.85937722, Time: 0.0210 Steps: 67010, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001021, Sample Num: 16336, Cur Loss: 0.20371798, Cur Avg Loss: 0.19045984, Log Avg loss: 0.22271738, Global Avg Loss: 0.85928222, Time: 0.0210 Steps: 67020, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001031, Sample Num: 16496, Cur Loss: 0.19750057, Cur Avg Loss: 0.19040200, Log Avg loss: 0.18449586, Global Avg Loss: 0.85918155, Time: 0.0225 Steps: 67030, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001041, Sample Num: 16656, Cur Loss: 0.10622017, Cur Avg Loss: 0.19059551, Log Avg loss: 0.21054677, Global Avg Loss: 0.85908480, Time: 0.0211 Steps: 67040, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001051, Sample Num: 16816, Cur Loss: 0.06633686, Cur Avg Loss: 0.19056874, Log Avg loss: 0.18778133, Global Avg Loss: 0.85898468, Time: 0.0210 Steps: 67050, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001061, Sample Num: 16976, Cur Loss: 0.21729660, Cur Avg Loss: 0.19082863, Log Avg loss: 0.21814398, Global Avg Loss: 0.85888912, Time: 0.0209 Steps: 67060, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001071, Sample Num: 17136, Cur Loss: 0.23520185, Cur Avg Loss: 0.19102842, Log Avg loss: 0.21222588, Global Avg Loss: 0.85879270, Time: 0.0210 Steps: 67070, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001081, Sample Num: 17296, Cur Loss: 0.05206831, Cur Avg Loss: 0.19093693, Log Avg loss: 0.18113829, Global Avg Loss: 0.85869168, Time: 0.0210 Steps: 67080, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001091, Sample Num: 17456, Cur Loss: 0.08568229, Cur Avg Loss: 0.19051887, Log Avg loss: 0.14532610, Global Avg Loss: 0.85858535, Time: 0.0210 Steps: 67090, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001101, Sample Num: 17616, Cur Loss: 0.18351316, Cur Avg Loss: 0.19081955, Log Avg loss: 0.22362362, Global Avg Loss: 0.85849072, Time: 0.0210 Steps: 67100, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001111, Sample Num: 17776, Cur Loss: 0.28625298, Cur Avg Loss: 0.19146586, Log Avg loss: 0.26262553, Global Avg Loss: 0.85840193, Time: 0.0210 Steps: 67110, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001121, Sample Num: 17936, Cur Loss: 0.18743092, Cur Avg Loss: 0.19133016, Log Avg loss: 0.17625360, Global Avg Loss: 0.85830030, Time: 0.0210 Steps: 67120, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001131, Sample Num: 18096, Cur Loss: 0.19166276, Cur Avg Loss: 0.19114777, Log Avg loss: 0.17070167, Global Avg Loss: 0.85819787, Time: 0.0212 Steps: 67130, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001141, Sample Num: 18256, Cur Loss: 0.36801940, Cur Avg Loss: 0.19199742, Log Avg loss: 0.28809227, Global Avg Loss: 0.85811296, Time: 0.0210 Steps: 67140, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001151, Sample Num: 18416, Cur Loss: 0.32674971, Cur Avg Loss: 0.19169135, Log Avg loss: 0.15676984, Global Avg Loss: 0.85800852, Time: 0.0209 Steps: 67150, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001161, Sample Num: 18576, Cur Loss: 0.34218323, Cur Avg Loss: 0.19239505, Log Avg loss: 0.27339038, Global Avg Loss: 0.85792147, Time: 0.0210 Steps: 67160, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001171, Sample Num: 18736, Cur Loss: 0.24953836, Cur Avg Loss: 0.19320254, Log Avg loss: 0.28695240, Global Avg Loss: 0.85783646, Time: 0.0209 Steps: 67170, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001181, Sample Num: 18896, Cur Loss: 0.16624725, Cur Avg Loss: 0.19277181, Log Avg loss: 0.14233266, Global Avg Loss: 0.85772996, Time: 0.0210 Steps: 67180, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001191, Sample Num: 19056, Cur Loss: 0.31520456, Cur Avg Loss: 0.19265649, Log Avg loss: 0.17903767, Global Avg Loss: 0.85762895, Time: 0.0209 Steps: 67190, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001201, Sample Num: 19216, Cur Loss: 0.04382842, Cur Avg Loss: 0.19280188, Log Avg loss: 0.21011741, Global Avg Loss: 0.85753259, Time: 0.0209 Steps: 67200, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001211, Sample Num: 19376, Cur Loss: 0.05268141, Cur Avg Loss: 0.19256868, Log Avg loss: 0.16456183, Global Avg Loss: 0.85742949, Time: 0.0210 Steps: 67210, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001221, Sample Num: 19536, Cur Loss: 0.32092613, Cur Avg Loss: 0.19221215, Log Avg loss: 0.14903587, Global Avg Loss: 0.85732410, Time: 0.0209 Steps: 67220, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001231, Sample Num: 19696, Cur Loss: 0.22228713, Cur Avg Loss: 0.19237025, Log Avg loss: 0.21167436, Global Avg Loss: 0.85722807, Time: 0.0210 Steps: 67230, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001241, Sample Num: 19856, Cur Loss: 0.10568758, Cur Avg Loss: 0.19255207, Log Avg loss: 0.21493395, Global Avg Loss: 0.85713254, Time: 0.0210 Steps: 67240, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001251, Sample Num: 20016, Cur Loss: 0.27052653, Cur Avg Loss: 0.19246287, Log Avg loss: 0.18139420, Global Avg Loss: 0.85703206, Time: 0.0210 Steps: 67250, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001261, Sample Num: 20176, Cur Loss: 0.28786457, Cur Avg Loss: 0.19290772, Log Avg loss: 0.24855811, Global Avg Loss: 0.85694160, Time: 0.0209 Steps: 67260, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001271, Sample Num: 20336, Cur Loss: 0.02678519, Cur Avg Loss: 0.19265254, Log Avg loss: 0.16047380, Global Avg Loss: 0.85683806, Time: 0.0210 Steps: 67270, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001281, Sample Num: 20496, Cur Loss: 0.08898528, Cur Avg Loss: 0.19246688, Log Avg loss: 0.16886922, Global Avg Loss: 0.85673581, Time: 0.0247 Steps: 67280, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001291, Sample Num: 20656, Cur Loss: 0.20474723, Cur Avg Loss: 0.19217209, Log Avg loss: 0.15441053, Global Avg Loss: 0.85663144, Time: 0.0210 Steps: 67290, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001301, Sample Num: 20816, Cur Loss: 0.10049623, Cur Avg Loss: 0.19250776, Log Avg loss: 0.23584274, Global Avg Loss: 0.85653919, Time: 0.0209 Steps: 67300, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001311, Sample Num: 20976, Cur Loss: 0.25429818, Cur Avg Loss: 0.19229201, Log Avg loss: 0.16422305, Global Avg Loss: 0.85643634, Time: 0.0209 Steps: 67310, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001321, Sample Num: 21136, Cur Loss: 0.03835704, Cur Avg Loss: 0.19243803, Log Avg loss: 0.21158059, Global Avg Loss: 0.85634055, Time: 0.0210 Steps: 67320, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001331, Sample Num: 21296, Cur Loss: 0.07133859, Cur Avg Loss: 0.19219916, Log Avg loss: 0.16064509, Global Avg Loss: 0.85623722, Time: 0.0209 Steps: 67330, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001341, Sample Num: 21456, Cur Loss: 0.42088467, Cur Avg Loss: 0.19227851, Log Avg loss: 0.20283964, Global Avg Loss: 0.85614019, Time: 0.0209 Steps: 67340, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001351, Sample Num: 21616, Cur Loss: 0.60451895, Cur Avg Loss: 0.19271042, Log Avg loss: 0.25062876, Global Avg Loss: 0.85605029, Time: 0.0210 Steps: 67350, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001361, Sample Num: 21776, Cur Loss: 0.44142482, Cur Avg Loss: 0.19299012, Log Avg loss: 0.23077879, Global Avg Loss: 0.85595746, Time: 0.0209 Steps: 67360, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001371, Sample Num: 21936, Cur Loss: 0.18721613, Cur Avg Loss: 0.19303945, Log Avg loss: 0.19975290, Global Avg Loss: 0.85586006, Time: 0.0210 Steps: 67370, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001381, Sample Num: 22096, Cur Loss: 0.16883977, Cur Avg Loss: 0.19316295, Log Avg loss: 0.21009395, Global Avg Loss: 0.85576422, Time: 0.0209 Steps: 67380, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001391, Sample Num: 22256, Cur Loss: 0.19109043, Cur Avg Loss: 0.19338890, Log Avg loss: 0.22459264, Global Avg Loss: 0.85567056, Time: 0.0210 Steps: 67390, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001401, Sample Num: 22416, Cur Loss: 0.27937555, Cur Avg Loss: 0.19386093, Log Avg loss: 0.25952120, Global Avg Loss: 0.85558211, Time: 0.0209 Steps: 67400, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001411, Sample Num: 22576, Cur Loss: 0.73620009, Cur Avg Loss: 0.19485570, Log Avg loss: 0.33422248, Global Avg Loss: 0.85550477, Time: 0.0209 Steps: 67410, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001421, Sample Num: 22736, Cur Loss: 0.21006466, Cur Avg Loss: 0.19513430, Log Avg loss: 0.23444431, Global Avg Loss: 0.85541265, Time: 0.0209 Steps: 67420, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001431, Sample Num: 22896, Cur Loss: 0.49131393, Cur Avg Loss: 0.19533879, Log Avg loss: 0.22439799, Global Avg Loss: 0.85531907, Time: 0.0209 Steps: 67430, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001441, Sample Num: 23056, Cur Loss: 0.16816689, Cur Avg Loss: 0.19509694, Log Avg loss: 0.16048732, Global Avg Loss: 0.85521604, Time: 0.0209 Steps: 67440, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001451, Sample Num: 23216, Cur Loss: 0.07871538, Cur Avg Loss: 0.19485026, Log Avg loss: 0.15930337, Global Avg Loss: 0.85511287, Time: 0.0209 Steps: 67450, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001461, Sample Num: 23376, Cur Loss: 0.05581285, Cur Avg Loss: 0.19451679, Log Avg loss: 0.14613124, Global Avg Loss: 0.85500777, Time: 0.0209 Steps: 67460, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001471, Sample Num: 23536, Cur Loss: 0.22662845, Cur Avg Loss: 0.19474944, Log Avg loss: 0.22873999, Global Avg Loss: 0.85491495, Time: 0.0209 Steps: 67470, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001481, Sample Num: 23696, Cur Loss: 0.22735539, Cur Avg Loss: 0.19452674, Log Avg loss: 0.16176729, Global Avg Loss: 0.85481223, Time: 0.0209 Steps: 67480, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001491, Sample Num: 23856, Cur Loss: 0.53872883, Cur Avg Loss: 0.19488177, Log Avg loss: 0.24746091, Global Avg Loss: 0.85472224, Time: 0.0209 Steps: 67490, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001501, Sample Num: 24016, Cur Loss: 0.25493455, Cur Avg Loss: 0.19440930, Log Avg loss: 0.12396396, Global Avg Loss: 0.85461398, Time: 0.0209 Steps: 67500, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001511, Sample Num: 24176, Cur Loss: 0.01923915, Cur Avg Loss: 0.19442177, Log Avg loss: 0.19629425, Global Avg Loss: 0.85451646, Time: 0.0209 Steps: 67510, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001521, Sample Num: 24336, Cur Loss: 0.51469094, Cur Avg Loss: 0.19471074, Log Avg loss: 0.23837343, Global Avg Loss: 0.85442521, Time: 0.0209 Steps: 67520, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001531, Sample Num: 24496, Cur Loss: 0.29965571, Cur Avg Loss: 0.19487193, Log Avg loss: 0.21938955, Global Avg Loss: 0.85433117, Time: 0.0210 Steps: 67530, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001541, Sample Num: 24656, Cur Loss: 0.11168199, Cur Avg Loss: 0.19485690, Log Avg loss: 0.19255492, Global Avg Loss: 0.85423319, Time: 0.0247 Steps: 67540, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001551, Sample Num: 24816, Cur Loss: 0.23749459, Cur Avg Loss: 0.19468264, Log Avg loss: 0.16783048, Global Avg Loss: 0.85413158, Time: 0.0209 Steps: 67550, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001561, Sample Num: 24976, Cur Loss: 0.20146126, Cur Avg Loss: 0.19470248, Log Avg loss: 0.19777960, Global Avg Loss: 0.85403442, Time: 0.0209 Steps: 67560, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001571, Sample Num: 25136, Cur Loss: 0.10061513, Cur Avg Loss: 0.19421844, Log Avg loss: 0.11865919, Global Avg Loss: 0.85392559, Time: 0.0209 Steps: 67570, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001581, Sample Num: 25296, Cur Loss: 0.18630613, Cur Avg Loss: 0.19402948, Log Avg loss: 0.16434347, Global Avg Loss: 0.85382355, Time: 0.0209 Steps: 67580, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001591, Sample Num: 25456, Cur Loss: 0.06541638, Cur Avg Loss: 0.19395979, Log Avg loss: 0.18294279, Global Avg Loss: 0.85372430, Time: 0.0209 Steps: 67590, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001601, Sample Num: 25616, Cur Loss: 0.07373954, Cur Avg Loss: 0.19400756, Log Avg loss: 0.20160696, Global Avg Loss: 0.85362783, Time: 0.0209 Steps: 67600, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001611, Sample Num: 25776, Cur Loss: 0.05522097, Cur Avg Loss: 0.19345507, Log Avg loss: 0.10500212, Global Avg Loss: 0.85351710, Time: 0.0209 Steps: 67610, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001621, Sample Num: 25936, Cur Loss: 0.46856990, Cur Avg Loss: 0.19385716, Log Avg loss: 0.25863322, Global Avg Loss: 0.85342913, Time: 0.0209 Steps: 67620, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001631, Sample Num: 26096, Cur Loss: 0.14303856, Cur Avg Loss: 0.19402581, Log Avg loss: 0.22136490, Global Avg Loss: 0.85333567, Time: 0.0209 Steps: 67630, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001641, Sample Num: 26256, Cur Loss: 0.16136198, Cur Avg Loss: 0.19400586, Log Avg loss: 0.19075123, Global Avg Loss: 0.85323771, Time: 0.0210 Steps: 67640, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001651, Sample Num: 26416, Cur Loss: 0.14385122, Cur Avg Loss: 0.19408810, Log Avg loss: 0.20758447, Global Avg Loss: 0.85314227, Time: 0.0210 Steps: 67650, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001661, Sample Num: 26576, Cur Loss: 0.10388546, Cur Avg Loss: 0.19404549, Log Avg loss: 0.18701038, Global Avg Loss: 0.85304382, Time: 0.0209 Steps: 67660, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001671, Sample Num: 26736, Cur Loss: 0.14207155, Cur Avg Loss: 0.19449102, Log Avg loss: 0.26849284, Global Avg Loss: 0.85295744, Time: 0.0209 Steps: 67670, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001681, Sample Num: 26896, Cur Loss: 0.21275991, Cur Avg Loss: 0.19480035, Log Avg loss: 0.24648949, Global Avg Loss: 0.85286783, Time: 0.0209 Steps: 67680, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001691, Sample Num: 27056, Cur Loss: 0.24202877, Cur Avg Loss: 0.19462981, Log Avg loss: 0.16596290, Global Avg Loss: 0.85276635, Time: 0.0209 Steps: 67690, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001701, Sample Num: 27216, Cur Loss: 0.10865636, Cur Avg Loss: 0.19488047, Log Avg loss: 0.23726670, Global Avg Loss: 0.85267543, Time: 0.0209 Steps: 67700, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001711, Sample Num: 27376, Cur Loss: 0.34449911, Cur Avg Loss: 0.19481546, Log Avg loss: 0.18375705, Global Avg Loss: 0.85257664, Time: 0.0210 Steps: 67710, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001721, Sample Num: 27536, Cur Loss: 0.15370704, Cur Avg Loss: 0.19467363, Log Avg loss: 0.17040641, Global Avg Loss: 0.85247591, Time: 0.0209 Steps: 67720, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001731, Sample Num: 27696, Cur Loss: 0.12898958, Cur Avg Loss: 0.19450921, Log Avg loss: 0.16621199, Global Avg Loss: 0.85237458, Time: 0.0209 Steps: 67730, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001741, Sample Num: 27856, Cur Loss: 0.19431709, Cur Avg Loss: 0.19438969, Log Avg loss: 0.17370140, Global Avg Loss: 0.85227440, Time: 0.0209 Steps: 67740, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001751, Sample Num: 28016, Cur Loss: 0.32311007, Cur Avg Loss: 0.19467607, Log Avg loss: 0.24453412, Global Avg Loss: 0.85218469, Time: 0.0209 Steps: 67750, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001761, Sample Num: 28176, Cur Loss: 0.05731552, Cur Avg Loss: 0.19441106, Log Avg loss: 0.14800824, Global Avg Loss: 0.85208077, Time: 0.0209 Steps: 67760, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001771, Sample Num: 28336, Cur Loss: 0.14461347, Cur Avg Loss: 0.19433789, Log Avg loss: 0.18145255, Global Avg Loss: 0.85198181, Time: 0.0209 Steps: 67770, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001781, Sample Num: 28496, Cur Loss: 0.23666948, Cur Avg Loss: 0.19403504, Log Avg loss: 0.14040047, Global Avg Loss: 0.85187683, Time: 0.0210 Steps: 67780, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001791, Sample Num: 28656, Cur Loss: 0.18289283, Cur Avg Loss: 0.19373297, Log Avg loss: 0.13993512, Global Avg Loss: 0.85177181, Time: 0.0209 Steps: 67790, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001801, Sample Num: 28816, Cur Loss: 0.09771048, Cur Avg Loss: 0.19378294, Log Avg loss: 0.20273090, Global Avg Loss: 0.85167608, Time: 0.0209 Steps: 67800, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001811, Sample Num: 28976, Cur Loss: 0.18564326, Cur Avg Loss: 0.19383464, Log Avg loss: 0.20314675, Global Avg Loss: 0.85158044, Time: 0.0209 Steps: 67810, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001821, Sample Num: 29136, Cur Loss: 0.39454806, Cur Avg Loss: 0.19441409, Log Avg loss: 0.29935326, Global Avg Loss: 0.85149902, Time: 0.0209 Steps: 67820, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001831, Sample Num: 29296, Cur Loss: 0.21722816, Cur Avg Loss: 0.19464279, Log Avg loss: 0.23628827, Global Avg Loss: 0.85140832, Time: 0.0210 Steps: 67830, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001841, Sample Num: 29456, Cur Loss: 0.09773075, Cur Avg Loss: 0.19487289, Log Avg loss: 0.23700497, Global Avg Loss: 0.85131775, Time: 0.0209 Steps: 67840, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001851, Sample Num: 29616, Cur Loss: 0.31216845, Cur Avg Loss: 0.19483017, Log Avg loss: 0.18696385, Global Avg Loss: 0.85121983, Time: 0.0209 Steps: 67850, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001861, Sample Num: 29776, Cur Loss: 0.09847954, Cur Avg Loss: 0.19490639, Log Avg loss: 0.20901634, Global Avg Loss: 0.85112520, Time: 0.0209 Steps: 67860, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001871, Sample Num: 29936, Cur Loss: 0.56679696, Cur Avg Loss: 0.19503255, Log Avg loss: 0.21850928, Global Avg Loss: 0.85103199, Time: 0.0209 Steps: 67870, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001881, Sample Num: 30096, Cur Loss: 0.14841647, Cur Avg Loss: 0.19524407, Log Avg loss: 0.23481961, Global Avg Loss: 0.85094121, Time: 0.0209 Steps: 67880, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001891, Sample Num: 30256, Cur Loss: 0.14096692, Cur Avg Loss: 0.19546088, Log Avg loss: 0.23624361, Global Avg Loss: 0.85085067, Time: 0.0209 Steps: 67890, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001901, Sample Num: 30416, Cur Loss: 0.27044022, Cur Avg Loss: 0.19519381, Log Avg loss: 0.14469107, Global Avg Loss: 0.85074667, Time: 0.0209 Steps: 67900, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001911, Sample Num: 30576, Cur Loss: 0.05767511, Cur Avg Loss: 0.19498035, Log Avg loss: 0.15440136, Global Avg Loss: 0.85064413, Time: 0.0209 Steps: 67910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001921, Sample Num: 30736, Cur Loss: 0.14686850, Cur Avg Loss: 0.19497319, Log Avg loss: 0.19360472, Global Avg Loss: 0.85054739, Time: 0.0209 Steps: 67920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001931, Sample Num: 30896, Cur Loss: 0.05433646, Cur Avg Loss: 0.19491180, Log Avg loss: 0.18311854, Global Avg Loss: 0.85044914, Time: 0.0209 Steps: 67930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001941, Sample Num: 31056, Cur Loss: 0.27030796, Cur Avg Loss: 0.19519548, Log Avg loss: 0.24997530, Global Avg Loss: 0.85036075, Time: 0.0209 Steps: 67940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001951, Sample Num: 31216, Cur Loss: 0.14055413, Cur Avg Loss: 0.19527570, Log Avg loss: 0.21084647, Global Avg Loss: 0.85026664, Time: 0.0209 Steps: 67950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001961, Sample Num: 31376, Cur Loss: 0.10304634, Cur Avg Loss: 0.19578072, Log Avg loss: 0.29430955, Global Avg Loss: 0.85018483, Time: 0.0209 Steps: 67960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001971, Sample Num: 31536, Cur Loss: 0.28724012, Cur Avg Loss: 0.19573030, Log Avg loss: 0.18584293, Global Avg Loss: 0.85008709, Time: 0.0209 Steps: 67970, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001981, Sample Num: 31696, Cur Loss: 0.16406935, Cur Avg Loss: 0.19577880, Log Avg loss: 0.20533775, Global Avg Loss: 0.84999225, Time: 0.0209 Steps: 67980, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001991, Sample Num: 31856, Cur Loss: 0.28916731, Cur Avg Loss: 0.19594758, Log Avg loss: 0.22938262, Global Avg Loss: 0.84990097, Time: 0.0209 Steps: 67990, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002001, Sample Num: 32016, Cur Loss: 0.14409269, Cur Avg Loss: 0.19631668, Log Avg loss: 0.26980472, Global Avg Loss: 0.84981566, Time: 0.0210 Steps: 68000, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002011, Sample Num: 32176, Cur Loss: 0.12759423, Cur Avg Loss: 0.19629990, Log Avg loss: 0.19294268, Global Avg Loss: 0.84971907, Time: 0.0209 Steps: 68010, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002021, Sample Num: 32336, Cur Loss: 0.20600757, Cur Avg Loss: 0.19614906, Log Avg loss: 0.16581421, Global Avg Loss: 0.84961853, Time: 0.0209 Steps: 68020, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002031, Sample Num: 32496, Cur Loss: 0.20776330, Cur Avg Loss: 0.19601989, Log Avg loss: 0.16991550, Global Avg Loss: 0.84951862, Time: 0.0209 Steps: 68030, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002041, Sample Num: 32656, Cur Loss: 0.24291027, Cur Avg Loss: 0.19604767, Log Avg loss: 0.20168884, Global Avg Loss: 0.84942340, Time: 0.0209 Steps: 68040, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002051, Sample Num: 32816, Cur Loss: 0.16992041, Cur Avg Loss: 0.19605527, Log Avg loss: 0.19760646, Global Avg Loss: 0.84932762, Time: 0.0246 Steps: 68050, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002061, Sample Num: 32976, Cur Loss: 0.16099782, Cur Avg Loss: 0.19600412, Log Avg loss: 0.18551453, Global Avg Loss: 0.84923009, Time: 0.0209 Steps: 68060, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002071, Sample Num: 33136, Cur Loss: 0.17044142, Cur Avg Loss: 0.19583739, Log Avg loss: 0.16147467, Global Avg Loss: 0.84912905, Time: 0.0209 Steps: 68070, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002081, Sample Num: 33296, Cur Loss: 0.09680167, Cur Avg Loss: 0.19587277, Log Avg loss: 0.20319865, Global Avg Loss: 0.84903417, Time: 0.0209 Steps: 68080, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002091, Sample Num: 33456, Cur Loss: 0.12863874, Cur Avg Loss: 0.19591804, Log Avg loss: 0.20533856, Global Avg Loss: 0.84893964, Time: 0.0209 Steps: 68090, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002101, Sample Num: 33616, Cur Loss: 0.31084427, Cur Avg Loss: 0.19569691, Log Avg loss: 0.14945906, Global Avg Loss: 0.84883692, Time: 0.0209 Steps: 68100, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002111, Sample Num: 33776, Cur Loss: 0.14030091, Cur Avg Loss: 0.19565290, Log Avg loss: 0.18640541, Global Avg Loss: 0.84873966, Time: 0.0209 Steps: 68110, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002121, Sample Num: 33936, Cur Loss: 0.15082754, Cur Avg Loss: 0.19557336, Log Avg loss: 0.17878344, Global Avg Loss: 0.84864131, Time: 0.0209 Steps: 68120, Updated lr: 0.000036 ***** Running evaluation checkpoint-68128 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-68128 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.663928, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.324165, "eval_total_loss": 227.888023, "eval_mae": 0.386722, "eval_mse": 0.324293, "eval_r2": 0.793858, "eval_sp_statistic": 0.883765, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.913773, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.936441, "test_total_loss": 470.09328, "test_mae": 0.764253, "test_mse": 0.936526, "test_r2": 0.395558, "test_sp_statistic": 0.796306, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.83702, "test_ps_pvalue": 0.0, "lr": 3.634139402560455e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8485611675693693, "train_cur_epoch_loss": 416.1400836072862, "train_cur_epoch_avg_loss": 0.19546269779581316, "train_cur_epoch_time": 44.663928270339966, "train_cur_epoch_avg_time": 0.020978829624396414, "epoch": 32, "step": 68128} ################################################## Training, Epoch: 0033, Batch: 000002, Sample Num: 32, Cur Loss: 0.12053077, Cur Avg Loss: 0.08930419, Log Avg loss: 0.15075958, Global Avg Loss: 0.84853888, Time: 0.0249 Steps: 68130, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000012, Sample Num: 192, Cur Loss: 0.39859051, Cur Avg Loss: 0.22756468, Log Avg loss: 0.25521677, Global Avg Loss: 0.84845181, Time: 0.0210 Steps: 68140, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000022, Sample Num: 352, Cur Loss: 0.09190162, Cur Avg Loss: 0.23896782, Log Avg loss: 0.25265160, Global Avg Loss: 0.84836438, Time: 0.0211 Steps: 68150, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000032, Sample Num: 512, Cur Loss: 0.16885519, Cur Avg Loss: 0.23601341, Log Avg loss: 0.22951369, Global Avg Loss: 0.84827359, Time: 0.0212 Steps: 68160, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000042, Sample Num: 672, Cur Loss: 0.26156127, Cur Avg Loss: 0.22827073, Log Avg loss: 0.20349417, Global Avg Loss: 0.84817900, Time: 0.0210 Steps: 68170, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000052, Sample Num: 832, Cur Loss: 0.23625621, Cur Avg Loss: 0.22394827, Log Avg loss: 0.20579392, Global Avg Loss: 0.84808478, Time: 0.0211 Steps: 68180, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000062, Sample Num: 992, Cur Loss: 0.39072439, Cur Avg Loss: 0.23122246, Log Avg loss: 0.26904824, Global Avg Loss: 0.84799987, Time: 0.0210 Steps: 68190, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000072, Sample Num: 1152, Cur Loss: 0.11157211, Cur Avg Loss: 0.22134365, Log Avg loss: 0.16009508, Global Avg Loss: 0.84789900, Time: 0.0211 Steps: 68200, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000082, Sample Num: 1312, Cur Loss: 0.14305237, Cur Avg Loss: 0.22265281, Log Avg loss: 0.23207871, Global Avg Loss: 0.84780872, Time: 0.0211 Steps: 68210, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000092, Sample Num: 1472, Cur Loss: 0.14765306, Cur Avg Loss: 0.21829872, Log Avg loss: 0.18259518, Global Avg Loss: 0.84771121, Time: 0.0211 Steps: 68220, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000102, Sample Num: 1632, Cur Loss: 0.22643310, Cur Avg Loss: 0.20859361, Log Avg loss: 0.11930659, Global Avg Loss: 0.84760445, Time: 0.0211 Steps: 68230, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000112, Sample Num: 1792, Cur Loss: 0.30500984, Cur Avg Loss: 0.20666037, Log Avg loss: 0.18694135, Global Avg Loss: 0.84750764, Time: 0.0211 Steps: 68240, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000122, Sample Num: 1952, Cur Loss: 0.13307668, Cur Avg Loss: 0.20541105, Log Avg loss: 0.19141873, Global Avg Loss: 0.84741151, Time: 0.0212 Steps: 68250, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000132, Sample Num: 2112, Cur Loss: 0.12244608, Cur Avg Loss: 0.20389216, Log Avg loss: 0.18536166, Global Avg Loss: 0.84731452, Time: 0.0211 Steps: 68260, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000142, Sample Num: 2272, Cur Loss: 0.13668260, Cur Avg Loss: 0.19809356, Log Avg loss: 0.12155209, Global Avg Loss: 0.84720821, Time: 0.0211 Steps: 68270, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000152, Sample Num: 2432, Cur Loss: 0.20699425, Cur Avg Loss: 0.19776684, Log Avg loss: 0.19312740, Global Avg Loss: 0.84711242, Time: 0.0211 Steps: 68280, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000162, Sample Num: 2592, Cur Loss: 0.18148448, Cur Avg Loss: 0.19964046, Log Avg loss: 0.22811950, Global Avg Loss: 0.84702177, Time: 0.0212 Steps: 68290, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000172, Sample Num: 2752, Cur Loss: 0.09266251, Cur Avg Loss: 0.19860539, Log Avg loss: 0.18183712, Global Avg Loss: 0.84692438, Time: 0.0210 Steps: 68300, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000182, Sample Num: 2912, Cur Loss: 0.18607435, Cur Avg Loss: 0.19473893, Log Avg loss: 0.12823586, Global Avg Loss: 0.84681917, Time: 0.0211 Steps: 68310, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000192, Sample Num: 3072, Cur Loss: 0.68161023, Cur Avg Loss: 0.19601208, Log Avg loss: 0.21918337, Global Avg Loss: 0.84672731, Time: 0.0211 Steps: 68320, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000202, Sample Num: 3232, Cur Loss: 0.14624096, Cur Avg Loss: 0.19318640, Log Avg loss: 0.13893344, Global Avg Loss: 0.84662372, Time: 0.0211 Steps: 68330, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000212, Sample Num: 3392, Cur Loss: 0.08873439, Cur Avg Loss: 0.18995972, Log Avg loss: 0.12478083, Global Avg Loss: 0.84651810, Time: 0.0210 Steps: 68340, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000222, Sample Num: 3552, Cur Loss: 0.15747151, Cur Avg Loss: 0.18680598, Log Avg loss: 0.11994663, Global Avg Loss: 0.84641179, Time: 0.0211 Steps: 68350, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000232, Sample Num: 3712, Cur Loss: 0.14571905, Cur Avg Loss: 0.18568692, Log Avg loss: 0.16084392, Global Avg Loss: 0.84631151, Time: 0.0211 Steps: 68360, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000242, Sample Num: 3872, Cur Loss: 0.11149290, Cur Avg Loss: 0.18693803, Log Avg loss: 0.21596356, Global Avg Loss: 0.84621931, Time: 0.0211 Steps: 68370, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000252, Sample Num: 4032, Cur Loss: 0.13738593, Cur Avg Loss: 0.18970449, Log Avg loss: 0.25665305, Global Avg Loss: 0.84613309, Time: 0.0211 Steps: 68380, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000262, Sample Num: 4192, Cur Loss: 0.18162078, Cur Avg Loss: 0.19352572, Log Avg loss: 0.28982049, Global Avg Loss: 0.84605175, Time: 0.0212 Steps: 68390, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000272, Sample Num: 4352, Cur Loss: 0.18504629, Cur Avg Loss: 0.19257711, Log Avg loss: 0.16772358, Global Avg Loss: 0.84595258, Time: 0.0211 Steps: 68400, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000282, Sample Num: 4512, Cur Loss: 0.35440108, Cur Avg Loss: 0.19353748, Log Avg loss: 0.21965950, Global Avg Loss: 0.84586103, Time: 0.0210 Steps: 68410, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000292, Sample Num: 4672, Cur Loss: 0.20496669, Cur Avg Loss: 0.19470425, Log Avg loss: 0.22760742, Global Avg Loss: 0.84577066, Time: 0.0211 Steps: 68420, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000302, Sample Num: 4832, Cur Loss: 0.14208241, Cur Avg Loss: 0.19621440, Log Avg loss: 0.24031073, Global Avg Loss: 0.84568219, Time: 0.0210 Steps: 68430, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000312, Sample Num: 4992, Cur Loss: 0.26125515, Cur Avg Loss: 0.19436593, Log Avg loss: 0.13854203, Global Avg Loss: 0.84557886, Time: 0.0211 Steps: 68440, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000322, Sample Num: 5152, Cur Loss: 0.28151578, Cur Avg Loss: 0.19550867, Log Avg loss: 0.23116209, Global Avg Loss: 0.84548910, Time: 0.0211 Steps: 68450, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000332, Sample Num: 5312, Cur Loss: 0.49380773, Cur Avg Loss: 0.19890724, Log Avg loss: 0.30834112, Global Avg Loss: 0.84541064, Time: 0.0211 Steps: 68460, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000342, Sample Num: 5472, Cur Loss: 0.10732377, Cur Avg Loss: 0.19944815, Log Avg loss: 0.21740647, Global Avg Loss: 0.84531892, Time: 0.0211 Steps: 68470, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000352, Sample Num: 5632, Cur Loss: 0.35092473, Cur Avg Loss: 0.20134326, Log Avg loss: 0.26615591, Global Avg Loss: 0.84523435, Time: 0.0211 Steps: 68480, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000362, Sample Num: 5792, Cur Loss: 0.06495501, Cur Avg Loss: 0.20157134, Log Avg loss: 0.20959988, Global Avg Loss: 0.84514154, Time: 0.0210 Steps: 68490, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000372, Sample Num: 5952, Cur Loss: 0.18963125, Cur Avg Loss: 0.20092276, Log Avg loss: 0.17744410, Global Avg Loss: 0.84504407, Time: 0.0211 Steps: 68500, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000382, Sample Num: 6112, Cur Loss: 0.17410044, Cur Avg Loss: 0.20018763, Log Avg loss: 0.17284071, Global Avg Loss: 0.84494595, Time: 0.0210 Steps: 68510, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000392, Sample Num: 6272, Cur Loss: 0.34125778, Cur Avg Loss: 0.20087480, Log Avg loss: 0.22712485, Global Avg Loss: 0.84485578, Time: 0.0210 Steps: 68520, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000402, Sample Num: 6432, Cur Loss: 0.10932529, Cur Avg Loss: 0.19925645, Log Avg loss: 0.13581716, Global Avg Loss: 0.84475232, Time: 0.0210 Steps: 68530, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000412, Sample Num: 6592, Cur Loss: 0.07251161, Cur Avg Loss: 0.19940701, Log Avg loss: 0.20545949, Global Avg Loss: 0.84465904, Time: 0.0210 Steps: 68540, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000422, Sample Num: 6752, Cur Loss: 0.27629355, Cur Avg Loss: 0.19907300, Log Avg loss: 0.18531188, Global Avg Loss: 0.84456286, Time: 0.0211 Steps: 68550, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000432, Sample Num: 6912, Cur Loss: 0.37085825, Cur Avg Loss: 0.19818918, Log Avg loss: 0.16089188, Global Avg Loss: 0.84446314, Time: 0.0210 Steps: 68560, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000442, Sample Num: 7072, Cur Loss: 0.18162793, Cur Avg Loss: 0.19692521, Log Avg loss: 0.14232191, Global Avg Loss: 0.84436074, Time: 0.0210 Steps: 68570, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000452, Sample Num: 7232, Cur Loss: 0.17256220, Cur Avg Loss: 0.19617237, Log Avg loss: 0.16289646, Global Avg Loss: 0.84426138, Time: 0.0211 Steps: 68580, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000462, Sample Num: 7392, Cur Loss: 0.20813176, Cur Avg Loss: 0.19473055, Log Avg loss: 0.12956043, Global Avg Loss: 0.84415718, Time: 0.0210 Steps: 68590, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000472, Sample Num: 7552, Cur Loss: 0.15907297, Cur Avg Loss: 0.19370578, Log Avg loss: 0.14636156, Global Avg Loss: 0.84405546, Time: 0.0211 Steps: 68600, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000482, Sample Num: 7712, Cur Loss: 0.27733341, Cur Avg Loss: 0.19330817, Log Avg loss: 0.17454076, Global Avg Loss: 0.84395787, Time: 0.0210 Steps: 68610, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000492, Sample Num: 7872, Cur Loss: 0.12138107, Cur Avg Loss: 0.19224601, Log Avg loss: 0.14105006, Global Avg Loss: 0.84385544, Time: 0.0211 Steps: 68620, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000502, Sample Num: 8032, Cur Loss: 0.41431653, Cur Avg Loss: 0.19200979, Log Avg loss: 0.18038785, Global Avg Loss: 0.84375877, Time: 0.0211 Steps: 68630, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000512, Sample Num: 8192, Cur Loss: 0.06020869, Cur Avg Loss: 0.19100181, Log Avg loss: 0.14040114, Global Avg Loss: 0.84365630, Time: 0.0256 Steps: 68640, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000522, Sample Num: 8352, Cur Loss: 0.13654716, Cur Avg Loss: 0.19151829, Log Avg loss: 0.21796195, Global Avg Loss: 0.84356515, Time: 0.0211 Steps: 68650, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000532, Sample Num: 8512, Cur Loss: 0.34040108, Cur Avg Loss: 0.19447292, Log Avg loss: 0.34870451, Global Avg Loss: 0.84349308, Time: 0.0210 Steps: 68660, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000542, Sample Num: 8672, Cur Loss: 0.18641733, Cur Avg Loss: 0.19398736, Log Avg loss: 0.16815585, Global Avg Loss: 0.84339473, Time: 0.0211 Steps: 68670, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000552, Sample Num: 8832, Cur Loss: 0.07184835, Cur Avg Loss: 0.19438045, Log Avg loss: 0.21568579, Global Avg Loss: 0.84330334, Time: 0.0210 Steps: 68680, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000562, Sample Num: 8992, Cur Loss: 0.33925217, Cur Avg Loss: 0.19469879, Log Avg loss: 0.21227131, Global Avg Loss: 0.84321147, Time: 0.0210 Steps: 68690, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000572, Sample Num: 9152, Cur Loss: 0.07960874, Cur Avg Loss: 0.19397464, Log Avg loss: 0.15327736, Global Avg Loss: 0.84311104, Time: 0.0211 Steps: 68700, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000582, Sample Num: 9312, Cur Loss: 0.19117390, Cur Avg Loss: 0.19524979, Log Avg loss: 0.26818803, Global Avg Loss: 0.84302737, Time: 0.0211 Steps: 68710, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000592, Sample Num: 9472, Cur Loss: 0.13575101, Cur Avg Loss: 0.19502405, Log Avg loss: 0.18188630, Global Avg Loss: 0.84293116, Time: 0.0211 Steps: 68720, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000602, Sample Num: 9632, Cur Loss: 0.03694133, Cur Avg Loss: 0.19427744, Log Avg loss: 0.15007804, Global Avg Loss: 0.84283035, Time: 0.0211 Steps: 68730, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000612, Sample Num: 9792, Cur Loss: 0.12338519, Cur Avg Loss: 0.19366825, Log Avg loss: 0.15699515, Global Avg Loss: 0.84273058, Time: 0.0210 Steps: 68740, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000622, Sample Num: 9952, Cur Loss: 0.27730638, Cur Avg Loss: 0.19455184, Log Avg loss: 0.24862723, Global Avg Loss: 0.84264417, Time: 0.0210 Steps: 68750, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000632, Sample Num: 10112, Cur Loss: 0.14211082, Cur Avg Loss: 0.19383656, Log Avg loss: 0.14934639, Global Avg Loss: 0.84254334, Time: 0.0210 Steps: 68760, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000642, Sample Num: 10272, Cur Loss: 0.05976580, Cur Avg Loss: 0.19290035, Log Avg loss: 0.13373178, Global Avg Loss: 0.84244027, Time: 0.0211 Steps: 68770, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000652, Sample Num: 10432, Cur Loss: 0.29635388, Cur Avg Loss: 0.19341652, Log Avg loss: 0.22655450, Global Avg Loss: 0.84235072, Time: 0.0210 Steps: 68780, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000662, Sample Num: 10592, Cur Loss: 0.13207178, Cur Avg Loss: 0.19305890, Log Avg loss: 0.16974195, Global Avg Loss: 0.84225295, Time: 0.0210 Steps: 68790, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000672, Sample Num: 10752, Cur Loss: 0.32751179, Cur Avg Loss: 0.19342976, Log Avg loss: 0.21798095, Global Avg Loss: 0.84216221, Time: 0.0210 Steps: 68800, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000682, Sample Num: 10912, Cur Loss: 0.05127907, Cur Avg Loss: 0.19351369, Log Avg loss: 0.19915402, Global Avg Loss: 0.84206876, Time: 0.0210 Steps: 68810, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000692, Sample Num: 11072, Cur Loss: 0.28194347, Cur Avg Loss: 0.19345523, Log Avg loss: 0.18946829, Global Avg Loss: 0.84197394, Time: 0.0211 Steps: 68820, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000702, Sample Num: 11232, Cur Loss: 0.08768474, Cur Avg Loss: 0.19326797, Log Avg loss: 0.18030945, Global Avg Loss: 0.84187781, Time: 0.0211 Steps: 68830, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000712, Sample Num: 11392, Cur Loss: 0.19988526, Cur Avg Loss: 0.19293687, Log Avg loss: 0.16969369, Global Avg Loss: 0.84178016, Time: 0.0210 Steps: 68840, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000722, Sample Num: 11552, Cur Loss: 0.10550608, Cur Avg Loss: 0.19261639, Log Avg loss: 0.16979832, Global Avg Loss: 0.84168256, Time: 0.0210 Steps: 68850, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000732, Sample Num: 11712, Cur Loss: 0.09603578, Cur Avg Loss: 0.19236945, Log Avg loss: 0.17454024, Global Avg Loss: 0.84158568, Time: 0.0210 Steps: 68860, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000742, Sample Num: 11872, Cur Loss: 0.26851517, Cur Avg Loss: 0.19150665, Log Avg loss: 0.12834950, Global Avg Loss: 0.84148211, Time: 0.0210 Steps: 68870, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000752, Sample Num: 12032, Cur Loss: 0.10250095, Cur Avg Loss: 0.19049448, Log Avg loss: 0.11539170, Global Avg Loss: 0.84137670, Time: 0.0210 Steps: 68880, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000762, Sample Num: 12192, Cur Loss: 0.19517614, Cur Avg Loss: 0.19077063, Log Avg loss: 0.21153680, Global Avg Loss: 0.84128527, Time: 0.0211 Steps: 68890, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000772, Sample Num: 12352, Cur Loss: 0.49550578, Cur Avg Loss: 0.19180379, Log Avg loss: 0.27053101, Global Avg Loss: 0.84120243, Time: 0.0248 Steps: 68900, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000782, Sample Num: 12512, Cur Loss: 0.07503770, Cur Avg Loss: 0.19151233, Log Avg loss: 0.16901131, Global Avg Loss: 0.84110489, Time: 0.0210 Steps: 68910, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000792, Sample Num: 12672, Cur Loss: 0.49045366, Cur Avg Loss: 0.19166438, Log Avg loss: 0.20355443, Global Avg Loss: 0.84101238, Time: 0.0211 Steps: 68920, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000802, Sample Num: 12832, Cur Loss: 0.52720714, Cur Avg Loss: 0.19242668, Log Avg loss: 0.25280124, Global Avg Loss: 0.84092705, Time: 0.0211 Steps: 68930, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000812, Sample Num: 12992, Cur Loss: 0.08262240, Cur Avg Loss: 0.19220679, Log Avg loss: 0.17457176, Global Avg Loss: 0.84083039, Time: 0.0211 Steps: 68940, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000822, Sample Num: 13152, Cur Loss: 0.41853070, Cur Avg Loss: 0.19228510, Log Avg loss: 0.19864343, Global Avg Loss: 0.84073725, Time: 0.0211 Steps: 68950, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000832, Sample Num: 13312, Cur Loss: 0.26063132, Cur Avg Loss: 0.19293579, Log Avg loss: 0.24642245, Global Avg Loss: 0.84065107, Time: 0.0211 Steps: 68960, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000842, Sample Num: 13472, Cur Loss: 0.14651802, Cur Avg Loss: 0.19309098, Log Avg loss: 0.20600286, Global Avg Loss: 0.84055905, Time: 0.0210 Steps: 68970, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000852, Sample Num: 13632, Cur Loss: 0.06087439, Cur Avg Loss: 0.19312720, Log Avg loss: 0.19617682, Global Avg Loss: 0.84046564, Time: 0.0210 Steps: 68980, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000862, Sample Num: 13792, Cur Loss: 0.10391372, Cur Avg Loss: 0.19356018, Log Avg loss: 0.23045052, Global Avg Loss: 0.84037722, Time: 0.0211 Steps: 68990, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000872, Sample Num: 13952, Cur Loss: 0.10042349, Cur Avg Loss: 0.19381697, Log Avg loss: 0.21595238, Global Avg Loss: 0.84028672, Time: 0.0211 Steps: 69000, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000882, Sample Num: 14112, Cur Loss: 0.17246026, Cur Avg Loss: 0.19426638, Log Avg loss: 0.23345469, Global Avg Loss: 0.84019879, Time: 0.0210 Steps: 69010, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000892, Sample Num: 14272, Cur Loss: 0.30270827, Cur Avg Loss: 0.19503036, Log Avg loss: 0.26241321, Global Avg Loss: 0.84011507, Time: 0.0211 Steps: 69020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000902, Sample Num: 14432, Cur Loss: 0.28065073, Cur Avg Loss: 0.19512620, Log Avg loss: 0.20367511, Global Avg Loss: 0.84002287, Time: 0.0210 Steps: 69030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000912, Sample Num: 14592, Cur Loss: 0.09955452, Cur Avg Loss: 0.19591332, Log Avg loss: 0.26691155, Global Avg Loss: 0.83993986, Time: 0.0211 Steps: 69040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000922, Sample Num: 14752, Cur Loss: 0.10790907, Cur Avg Loss: 0.19637921, Log Avg loss: 0.23886885, Global Avg Loss: 0.83985281, Time: 0.0210 Steps: 69050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000932, Sample Num: 14912, Cur Loss: 0.21016639, Cur Avg Loss: 0.19587864, Log Avg loss: 0.14972542, Global Avg Loss: 0.83975288, Time: 0.0211 Steps: 69060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000942, Sample Num: 15072, Cur Loss: 0.30868042, Cur Avg Loss: 0.19562706, Log Avg loss: 0.17218037, Global Avg Loss: 0.83965623, Time: 0.0211 Steps: 69070, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000952, Sample Num: 15232, Cur Loss: 0.20340748, Cur Avg Loss: 0.19530813, Log Avg loss: 0.16526498, Global Avg Loss: 0.83955861, Time: 0.0211 Steps: 69080, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000962, Sample Num: 15392, Cur Loss: 0.37268656, Cur Avg Loss: 0.19563053, Log Avg loss: 0.22632295, Global Avg Loss: 0.83946985, Time: 0.0211 Steps: 69090, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000972, Sample Num: 15552, Cur Loss: 0.26242125, Cur Avg Loss: 0.19571220, Log Avg loss: 0.20356895, Global Avg Loss: 0.83937782, Time: 0.0211 Steps: 69100, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000982, Sample Num: 15712, Cur Loss: 0.25544652, Cur Avg Loss: 0.19537830, Log Avg loss: 0.16292256, Global Avg Loss: 0.83927994, Time: 0.0210 Steps: 69110, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000992, Sample Num: 15872, Cur Loss: 0.20319569, Cur Avg Loss: 0.19594612, Log Avg loss: 0.25170640, Global Avg Loss: 0.83919493, Time: 0.0211 Steps: 69120, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001002, Sample Num: 16032, Cur Loss: 0.09334756, Cur Avg Loss: 0.19570450, Log Avg loss: 0.17173624, Global Avg Loss: 0.83909838, Time: 0.0211 Steps: 69130, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001012, Sample Num: 16192, Cur Loss: 0.10272851, Cur Avg Loss: 0.19510475, Log Avg loss: 0.13500912, Global Avg Loss: 0.83899655, Time: 0.0211 Steps: 69140, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001022, Sample Num: 16352, Cur Loss: 0.07508923, Cur Avg Loss: 0.19453306, Log Avg loss: 0.13667836, Global Avg Loss: 0.83889498, Time: 0.0210 Steps: 69150, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001032, Sample Num: 16512, Cur Loss: 0.06058562, Cur Avg Loss: 0.19390803, Log Avg loss: 0.13002966, Global Avg Loss: 0.83879249, Time: 0.0211 Steps: 69160, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001042, Sample Num: 16672, Cur Loss: 0.32408106, Cur Avg Loss: 0.19474574, Log Avg loss: 0.28119775, Global Avg Loss: 0.83871187, Time: 0.0211 Steps: 69170, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001052, Sample Num: 16832, Cur Loss: 0.27491754, Cur Avg Loss: 0.19473421, Log Avg loss: 0.19353281, Global Avg Loss: 0.83861861, Time: 0.0210 Steps: 69180, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001062, Sample Num: 16992, Cur Loss: 0.10843381, Cur Avg Loss: 0.19505144, Log Avg loss: 0.22842418, Global Avg Loss: 0.83853042, Time: 0.0210 Steps: 69190, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001072, Sample Num: 17152, Cur Loss: 0.13561362, Cur Avg Loss: 0.19485018, Log Avg loss: 0.17347635, Global Avg Loss: 0.83843432, Time: 0.0210 Steps: 69200, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001082, Sample Num: 17312, Cur Loss: 0.25805369, Cur Avg Loss: 0.19505714, Log Avg loss: 0.21724262, Global Avg Loss: 0.83834456, Time: 0.0211 Steps: 69210, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001092, Sample Num: 17472, Cur Loss: 0.23410974, Cur Avg Loss: 0.19528439, Log Avg loss: 0.21987340, Global Avg Loss: 0.83825521, Time: 0.0211 Steps: 69220, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001102, Sample Num: 17632, Cur Loss: 0.12635298, Cur Avg Loss: 0.19471792, Log Avg loss: 0.13285912, Global Avg Loss: 0.83815332, Time: 0.0210 Steps: 69230, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001112, Sample Num: 17792, Cur Loss: 0.18244199, Cur Avg Loss: 0.19446972, Log Avg loss: 0.16711765, Global Avg Loss: 0.83805641, Time: 0.0210 Steps: 69240, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001122, Sample Num: 17952, Cur Loss: 0.07642466, Cur Avg Loss: 0.19408386, Log Avg loss: 0.15117714, Global Avg Loss: 0.83795722, Time: 0.0211 Steps: 69250, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001132, Sample Num: 18112, Cur Loss: 0.42425781, Cur Avg Loss: 0.19421647, Log Avg loss: 0.20909515, Global Avg Loss: 0.83786642, Time: 0.0210 Steps: 69260, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001142, Sample Num: 18272, Cur Loss: 0.07448892, Cur Avg Loss: 0.19345671, Log Avg loss: 0.10745202, Global Avg Loss: 0.83776098, Time: 0.0210 Steps: 69270, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001152, Sample Num: 18432, Cur Loss: 0.49664530, Cur Avg Loss: 0.19469680, Log Avg loss: 0.33631449, Global Avg Loss: 0.83768860, Time: 0.0211 Steps: 69280, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001162, Sample Num: 18592, Cur Loss: 0.22337779, Cur Avg Loss: 0.19554537, Log Avg loss: 0.29330051, Global Avg Loss: 0.83761003, Time: 0.0211 Steps: 69290, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001172, Sample Num: 18752, Cur Loss: 0.06934734, Cur Avg Loss: 0.19577666, Log Avg loss: 0.22265293, Global Avg Loss: 0.83752129, Time: 0.0210 Steps: 69300, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001182, Sample Num: 18912, Cur Loss: 0.13854021, Cur Avg Loss: 0.19549356, Log Avg loss: 0.16231438, Global Avg Loss: 0.83742387, Time: 0.0210 Steps: 69310, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001192, Sample Num: 19072, Cur Loss: 0.29337612, Cur Avg Loss: 0.19606803, Log Avg loss: 0.26397027, Global Avg Loss: 0.83734115, Time: 0.0210 Steps: 69320, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001202, Sample Num: 19232, Cur Loss: 0.06689848, Cur Avg Loss: 0.19692578, Log Avg loss: 0.29916978, Global Avg Loss: 0.83726352, Time: 0.0211 Steps: 69330, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001212, Sample Num: 19392, Cur Loss: 0.11661416, Cur Avg Loss: 0.19710411, Log Avg loss: 0.21853949, Global Avg Loss: 0.83717429, Time: 0.0211 Steps: 69340, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001222, Sample Num: 19552, Cur Loss: 0.49422386, Cur Avg Loss: 0.19705657, Log Avg loss: 0.19129471, Global Avg Loss: 0.83708116, Time: 0.0210 Steps: 69350, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001232, Sample Num: 19712, Cur Loss: 0.21590008, Cur Avg Loss: 0.19716811, Log Avg loss: 0.21079811, Global Avg Loss: 0.83699086, Time: 0.0210 Steps: 69360, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001242, Sample Num: 19872, Cur Loss: 0.02479343, Cur Avg Loss: 0.19746849, Log Avg loss: 0.23447507, Global Avg Loss: 0.83690401, Time: 0.0211 Steps: 69370, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001252, Sample Num: 20032, Cur Loss: 0.13909902, Cur Avg Loss: 0.19762220, Log Avg loss: 0.21671253, Global Avg Loss: 0.83681462, Time: 0.0210 Steps: 69380, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001262, Sample Num: 20192, Cur Loss: 0.32773045, Cur Avg Loss: 0.19752688, Log Avg loss: 0.18559349, Global Avg Loss: 0.83672077, Time: 0.0210 Steps: 69390, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001272, Sample Num: 20352, Cur Loss: 0.09003526, Cur Avg Loss: 0.19800442, Log Avg loss: 0.25826960, Global Avg Loss: 0.83663742, Time: 0.0211 Steps: 69400, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001282, Sample Num: 20512, Cur Loss: 0.10599536, Cur Avg Loss: 0.19796912, Log Avg loss: 0.19347977, Global Avg Loss: 0.83654476, Time: 0.0250 Steps: 69410, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001292, Sample Num: 20672, Cur Loss: 0.24127778, Cur Avg Loss: 0.19814096, Log Avg loss: 0.22017092, Global Avg Loss: 0.83645597, Time: 0.0211 Steps: 69420, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001302, Sample Num: 20832, Cur Loss: 0.23779094, Cur Avg Loss: 0.19801670, Log Avg loss: 0.18196203, Global Avg Loss: 0.83636170, Time: 0.0211 Steps: 69430, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001312, Sample Num: 20992, Cur Loss: 0.14419334, Cur Avg Loss: 0.19800668, Log Avg loss: 0.19670176, Global Avg Loss: 0.83626959, Time: 0.0211 Steps: 69440, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001322, Sample Num: 21152, Cur Loss: 0.37643820, Cur Avg Loss: 0.19809847, Log Avg loss: 0.21014160, Global Avg Loss: 0.83617943, Time: 0.0211 Steps: 69450, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001332, Sample Num: 21312, Cur Loss: 0.06997526, Cur Avg Loss: 0.19873326, Log Avg loss: 0.28265280, Global Avg Loss: 0.83609974, Time: 0.0211 Steps: 69460, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001342, Sample Num: 21472, Cur Loss: 0.16268808, Cur Avg Loss: 0.19834162, Log Avg loss: 0.14617458, Global Avg Loss: 0.83600043, Time: 0.0210 Steps: 69470, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001352, Sample Num: 21632, Cur Loss: 0.10598929, Cur Avg Loss: 0.19815889, Log Avg loss: 0.17363671, Global Avg Loss: 0.83590510, Time: 0.0210 Steps: 69480, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001362, Sample Num: 21792, Cur Loss: 0.10119449, Cur Avg Loss: 0.19838422, Log Avg loss: 0.22884838, Global Avg Loss: 0.83581774, Time: 0.0211 Steps: 69490, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001372, Sample Num: 21952, Cur Loss: 0.16854797, Cur Avg Loss: 0.19858827, Log Avg loss: 0.22638051, Global Avg Loss: 0.83573005, Time: 0.0210 Steps: 69500, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001382, Sample Num: 22112, Cur Loss: 0.10021052, Cur Avg Loss: 0.19832737, Log Avg loss: 0.16253134, Global Avg Loss: 0.83563320, Time: 0.0210 Steps: 69510, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001392, Sample Num: 22272, Cur Loss: 0.07471021, Cur Avg Loss: 0.19823820, Log Avg loss: 0.18591474, Global Avg Loss: 0.83553974, Time: 0.0211 Steps: 69520, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001402, Sample Num: 22432, Cur Loss: 0.17471951, Cur Avg Loss: 0.19830654, Log Avg loss: 0.20781947, Global Avg Loss: 0.83544946, Time: 0.0210 Steps: 69530, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001412, Sample Num: 22592, Cur Loss: 0.15626553, Cur Avg Loss: 0.19824346, Log Avg loss: 0.18940052, Global Avg Loss: 0.83535656, Time: 0.0210 Steps: 69540, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001422, Sample Num: 22752, Cur Loss: 0.07638425, Cur Avg Loss: 0.19827245, Log Avg loss: 0.20236538, Global Avg Loss: 0.83526554, Time: 0.0210 Steps: 69550, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001432, Sample Num: 22912, Cur Loss: 0.09928383, Cur Avg Loss: 0.19774868, Log Avg loss: 0.12326797, Global Avg Loss: 0.83516319, Time: 0.0210 Steps: 69560, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001442, Sample Num: 23072, Cur Loss: 0.28533399, Cur Avg Loss: 0.19814300, Log Avg loss: 0.25461070, Global Avg Loss: 0.83507974, Time: 0.0211 Steps: 69570, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001452, Sample Num: 23232, Cur Loss: 0.09163362, Cur Avg Loss: 0.19817414, Log Avg loss: 0.20266339, Global Avg Loss: 0.83498885, Time: 0.0211 Steps: 69580, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001462, Sample Num: 23392, Cur Loss: 0.25275505, Cur Avg Loss: 0.19832585, Log Avg loss: 0.22035455, Global Avg Loss: 0.83490053, Time: 0.0210 Steps: 69590, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001472, Sample Num: 23552, Cur Loss: 0.19821823, Cur Avg Loss: 0.19840435, Log Avg loss: 0.20988107, Global Avg Loss: 0.83481072, Time: 0.0210 Steps: 69600, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001482, Sample Num: 23712, Cur Loss: 0.10372928, Cur Avg Loss: 0.19847991, Log Avg loss: 0.20960194, Global Avg Loss: 0.83472091, Time: 0.0210 Steps: 69610, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001492, Sample Num: 23872, Cur Loss: 0.13318445, Cur Avg Loss: 0.19886871, Log Avg loss: 0.25648887, Global Avg Loss: 0.83463785, Time: 0.0210 Steps: 69620, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001502, Sample Num: 24032, Cur Loss: 0.09737996, Cur Avg Loss: 0.19914132, Log Avg loss: 0.23981525, Global Avg Loss: 0.83455243, Time: 0.0210 Steps: 69630, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001512, Sample Num: 24192, Cur Loss: 0.15303305, Cur Avg Loss: 0.19886172, Log Avg loss: 0.15686519, Global Avg Loss: 0.83445511, Time: 0.0210 Steps: 69640, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001522, Sample Num: 24352, Cur Loss: 0.05311206, Cur Avg Loss: 0.19845446, Log Avg loss: 0.13687772, Global Avg Loss: 0.83435496, Time: 0.0210 Steps: 69650, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001532, Sample Num: 24512, Cur Loss: 0.32428077, Cur Avg Loss: 0.19854859, Log Avg loss: 0.21287513, Global Avg Loss: 0.83426574, Time: 0.0211 Steps: 69660, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001542, Sample Num: 24672, Cur Loss: 0.08769275, Cur Avg Loss: 0.19824971, Log Avg loss: 0.15246164, Global Avg Loss: 0.83416788, Time: 0.0213 Steps: 69670, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001552, Sample Num: 24832, Cur Loss: 0.18518649, Cur Avg Loss: 0.19807970, Log Avg loss: 0.17186302, Global Avg Loss: 0.83407283, Time: 0.0212 Steps: 69680, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001562, Sample Num: 24992, Cur Loss: 0.34964597, Cur Avg Loss: 0.19815956, Log Avg loss: 0.21055380, Global Avg Loss: 0.83398336, Time: 0.0212 Steps: 69690, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001572, Sample Num: 25152, Cur Loss: 0.07754175, Cur Avg Loss: 0.19789550, Log Avg loss: 0.15664983, Global Avg Loss: 0.83388618, Time: 0.0212 Steps: 69700, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001582, Sample Num: 25312, Cur Loss: 0.10426907, Cur Avg Loss: 0.19804277, Log Avg loss: 0.22119342, Global Avg Loss: 0.83379829, Time: 0.0212 Steps: 69710, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001592, Sample Num: 25472, Cur Loss: 0.27772966, Cur Avg Loss: 0.19806258, Log Avg loss: 0.20119626, Global Avg Loss: 0.83370756, Time: 0.0212 Steps: 69720, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001602, Sample Num: 25632, Cur Loss: 0.11404777, Cur Avg Loss: 0.19807778, Log Avg loss: 0.20049818, Global Avg Loss: 0.83361675, Time: 0.0212 Steps: 69730, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001612, Sample Num: 25792, Cur Loss: 0.26365343, Cur Avg Loss: 0.19785077, Log Avg loss: 0.16148330, Global Avg Loss: 0.83352037, Time: 0.0212 Steps: 69740, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001622, Sample Num: 25952, Cur Loss: 0.16129681, Cur Avg Loss: 0.19793783, Log Avg loss: 0.21197273, Global Avg Loss: 0.83343126, Time: 0.0212 Steps: 69750, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001632, Sample Num: 26112, Cur Loss: 0.06915252, Cur Avg Loss: 0.19788316, Log Avg loss: 0.18901516, Global Avg Loss: 0.83333888, Time: 0.0212 Steps: 69760, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001642, Sample Num: 26272, Cur Loss: 0.12577276, Cur Avg Loss: 0.19767716, Log Avg loss: 0.16405833, Global Avg Loss: 0.83324296, Time: 0.0212 Steps: 69770, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001652, Sample Num: 26432, Cur Loss: 0.19311573, Cur Avg Loss: 0.19729930, Log Avg loss: 0.13525477, Global Avg Loss: 0.83314293, Time: 0.0212 Steps: 69780, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001662, Sample Num: 26592, Cur Loss: 0.19437373, Cur Avg Loss: 0.19730707, Log Avg loss: 0.19858977, Global Avg Loss: 0.83305201, Time: 0.0212 Steps: 69790, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001672, Sample Num: 26752, Cur Loss: 0.27322477, Cur Avg Loss: 0.19744891, Log Avg loss: 0.22102352, Global Avg Loss: 0.83296432, Time: 0.0212 Steps: 69800, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001682, Sample Num: 26912, Cur Loss: 0.20064126, Cur Avg Loss: 0.19733893, Log Avg loss: 0.17895065, Global Avg Loss: 0.83287064, Time: 0.0212 Steps: 69810, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001692, Sample Num: 27072, Cur Loss: 0.29013509, Cur Avg Loss: 0.19743851, Log Avg loss: 0.21418663, Global Avg Loss: 0.83278203, Time: 0.0212 Steps: 69820, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001702, Sample Num: 27232, Cur Loss: 0.08838488, Cur Avg Loss: 0.19808649, Log Avg loss: 0.30772566, Global Avg Loss: 0.83270684, Time: 0.0212 Steps: 69830, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001712, Sample Num: 27392, Cur Loss: 0.17328937, Cur Avg Loss: 0.19816797, Log Avg loss: 0.21203559, Global Avg Loss: 0.83261797, Time: 0.0212 Steps: 69840, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001722, Sample Num: 27552, Cur Loss: 0.30471787, Cur Avg Loss: 0.19833753, Log Avg loss: 0.22736672, Global Avg Loss: 0.83253132, Time: 0.0212 Steps: 69850, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001732, Sample Num: 27712, Cur Loss: 0.28753161, Cur Avg Loss: 0.19815955, Log Avg loss: 0.16751037, Global Avg Loss: 0.83243612, Time: 0.0212 Steps: 69860, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001742, Sample Num: 27872, Cur Loss: 0.05718708, Cur Avg Loss: 0.19785189, Log Avg loss: 0.14456490, Global Avg Loss: 0.83233767, Time: 0.0211 Steps: 69870, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001752, Sample Num: 28032, Cur Loss: 0.34853274, Cur Avg Loss: 0.19838525, Log Avg loss: 0.29129722, Global Avg Loss: 0.83226025, Time: 0.0212 Steps: 69880, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001762, Sample Num: 28192, Cur Loss: 0.12236176, Cur Avg Loss: 0.19816636, Log Avg loss: 0.15981725, Global Avg Loss: 0.83216403, Time: 0.0212 Steps: 69890, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001772, Sample Num: 28352, Cur Loss: 0.08774739, Cur Avg Loss: 0.19796057, Log Avg loss: 0.16169957, Global Avg Loss: 0.83206812, Time: 0.0212 Steps: 69900, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001782, Sample Num: 28512, Cur Loss: 0.13073355, Cur Avg Loss: 0.19816363, Log Avg loss: 0.23414615, Global Avg Loss: 0.83198259, Time: 0.0212 Steps: 69910, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001792, Sample Num: 28672, Cur Loss: 0.06494166, Cur Avg Loss: 0.19774885, Log Avg loss: 0.12383488, Global Avg Loss: 0.83188131, Time: 0.0257 Steps: 69920, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001802, Sample Num: 28832, Cur Loss: 0.22097163, Cur Avg Loss: 0.19755512, Log Avg loss: 0.16283882, Global Avg Loss: 0.83178564, Time: 0.0211 Steps: 69930, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001812, Sample Num: 28992, Cur Loss: 0.06640413, Cur Avg Loss: 0.19741623, Log Avg loss: 0.17238880, Global Avg Loss: 0.83169136, Time: 0.0210 Steps: 69940, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001822, Sample Num: 29152, Cur Loss: 0.12856182, Cur Avg Loss: 0.19752692, Log Avg loss: 0.21758417, Global Avg Loss: 0.83160356, Time: 0.0210 Steps: 69950, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001832, Sample Num: 29312, Cur Loss: 0.28607777, Cur Avg Loss: 0.19733155, Log Avg loss: 0.16173392, Global Avg Loss: 0.83150781, Time: 0.0211 Steps: 69960, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001842, Sample Num: 29472, Cur Loss: 0.15097308, Cur Avg Loss: 0.19739517, Log Avg loss: 0.20905074, Global Avg Loss: 0.83141885, Time: 0.0211 Steps: 69970, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001852, Sample Num: 29632, Cur Loss: 0.03112365, Cur Avg Loss: 0.19726587, Log Avg loss: 0.17344889, Global Avg Loss: 0.83132483, Time: 0.0211 Steps: 69980, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001862, Sample Num: 29792, Cur Loss: 0.09334338, Cur Avg Loss: 0.19717109, Log Avg loss: 0.17961799, Global Avg Loss: 0.83123172, Time: 0.0211 Steps: 69990, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001872, Sample Num: 29952, Cur Loss: 0.13170241, Cur Avg Loss: 0.19704185, Log Avg loss: 0.17297814, Global Avg Loss: 0.83113768, Time: 0.0211 Steps: 70000, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001882, Sample Num: 30112, Cur Loss: 0.21359889, Cur Avg Loss: 0.19705924, Log Avg loss: 0.20031352, Global Avg Loss: 0.83104757, Time: 0.0210 Steps: 70010, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001892, Sample Num: 30272, Cur Loss: 0.18586688, Cur Avg Loss: 0.19736640, Log Avg loss: 0.25517347, Global Avg Loss: 0.83096533, Time: 0.0210 Steps: 70020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001902, Sample Num: 30432, Cur Loss: 0.05261583, Cur Avg Loss: 0.19697911, Log Avg loss: 0.12370469, Global Avg Loss: 0.83086434, Time: 0.0211 Steps: 70030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001912, Sample Num: 30592, Cur Loss: 0.14114863, Cur Avg Loss: 0.19750204, Log Avg loss: 0.29696244, Global Avg Loss: 0.83078811, Time: 0.0212 Steps: 70040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001922, Sample Num: 30752, Cur Loss: 0.26417086, Cur Avg Loss: 0.19756206, Log Avg loss: 0.20903916, Global Avg Loss: 0.83069935, Time: 0.0211 Steps: 70050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001932, Sample Num: 30912, Cur Loss: 0.04936733, Cur Avg Loss: 0.19728331, Log Avg loss: 0.14370728, Global Avg Loss: 0.83060129, Time: 0.0210 Steps: 70060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001942, Sample Num: 31072, Cur Loss: 0.09604979, Cur Avg Loss: 0.19721005, Log Avg loss: 0.18305564, Global Avg Loss: 0.83050888, Time: 0.0211 Steps: 70070, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001952, Sample Num: 31232, Cur Loss: 0.18020624, Cur Avg Loss: 0.19718020, Log Avg loss: 0.19138289, Global Avg Loss: 0.83041768, Time: 0.0211 Steps: 70080, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001962, Sample Num: 31392, Cur Loss: 0.08718284, Cur Avg Loss: 0.19707937, Log Avg loss: 0.17739911, Global Avg Loss: 0.83032451, Time: 0.0211 Steps: 70090, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001972, Sample Num: 31552, Cur Loss: 0.19965383, Cur Avg Loss: 0.19741076, Log Avg loss: 0.26242823, Global Avg Loss: 0.83024350, Time: 0.0210 Steps: 70100, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001982, Sample Num: 31712, Cur Loss: 0.16496314, Cur Avg Loss: 0.19792425, Log Avg loss: 0.29918439, Global Avg Loss: 0.83016775, Time: 0.0210 Steps: 70110, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001992, Sample Num: 31872, Cur Loss: 0.23050199, Cur Avg Loss: 0.19827829, Log Avg loss: 0.26844923, Global Avg Loss: 0.83008764, Time: 0.0210 Steps: 70120, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002002, Sample Num: 32032, Cur Loss: 0.18676282, Cur Avg Loss: 0.19846832, Log Avg loss: 0.23632202, Global Avg Loss: 0.83000298, Time: 0.0210 Steps: 70130, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002012, Sample Num: 32192, Cur Loss: 0.37866896, Cur Avg Loss: 0.19856474, Log Avg loss: 0.21786934, Global Avg Loss: 0.82991570, Time: 0.0211 Steps: 70140, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002022, Sample Num: 32352, Cur Loss: 0.10507677, Cur Avg Loss: 0.19824819, Log Avg loss: 0.13455838, Global Avg Loss: 0.82981658, Time: 0.0211 Steps: 70150, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002032, Sample Num: 32512, Cur Loss: 0.13592747, Cur Avg Loss: 0.19794302, Log Avg loss: 0.13623766, Global Avg Loss: 0.82971772, Time: 0.0210 Steps: 70160, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002042, Sample Num: 32672, Cur Loss: 0.64222133, Cur Avg Loss: 0.19806299, Log Avg loss: 0.22244026, Global Avg Loss: 0.82963118, Time: 0.0210 Steps: 70170, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002052, Sample Num: 32832, Cur Loss: 0.17625359, Cur Avg Loss: 0.19845444, Log Avg loss: 0.27838816, Global Avg Loss: 0.82955263, Time: 0.0248 Steps: 70180, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002062, Sample Num: 32992, Cur Loss: 0.23106676, Cur Avg Loss: 0.19814806, Log Avg loss: 0.13527947, Global Avg Loss: 0.82945372, Time: 0.0211 Steps: 70190, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002072, Sample Num: 33152, Cur Loss: 0.15028292, Cur Avg Loss: 0.19813039, Log Avg loss: 0.19448631, Global Avg Loss: 0.82936327, Time: 0.0211 Steps: 70200, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002082, Sample Num: 33312, Cur Loss: 0.08284634, Cur Avg Loss: 0.19820454, Log Avg loss: 0.21356805, Global Avg Loss: 0.82927556, Time: 0.0211 Steps: 70210, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002092, Sample Num: 33472, Cur Loss: 0.14146209, Cur Avg Loss: 0.19804122, Log Avg loss: 0.16403789, Global Avg Loss: 0.82918082, Time: 0.0211 Steps: 70220, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002102, Sample Num: 33632, Cur Loss: 0.10375665, Cur Avg Loss: 0.19770616, Log Avg loss: 0.12761228, Global Avg Loss: 0.82908093, Time: 0.0210 Steps: 70230, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002112, Sample Num: 33792, Cur Loss: 0.15451553, Cur Avg Loss: 0.19746133, Log Avg loss: 0.14599779, Global Avg Loss: 0.82898368, Time: 0.0210 Steps: 70240, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002122, Sample Num: 33952, Cur Loss: 0.04678039, Cur Avg Loss: 0.19734502, Log Avg loss: 0.17278117, Global Avg Loss: 0.82889027, Time: 0.0210 Steps: 70250, Updated lr: 0.000034 ***** Running evaluation checkpoint-70257 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-70257 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.040138, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.293143, "eval_total_loss": 206.079628, "eval_mae": 0.363713, "eval_mse": 0.293258, "eval_r2": 0.813586, "eval_sp_statistic": 0.889324, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.915839, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.866482, "test_total_loss": 434.973808, "test_mae": 0.728821, "test_mse": 0.866573, "test_r2": 0.440706, "test_sp_statistic": 0.793372, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.831416, "test_ps_pvalue": 0.0, "lr": 3.432242769084875e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.828825301826749, "train_cur_epoch_loss": 420.00400627590716, "train_cur_epoch_avg_loss": 0.19727759806289674, "train_cur_epoch_time": 45.04013752937317, "train_cur_epoch_avg_time": 0.02115553665071544, "epoch": 33, "step": 70257} ################################################## Training, Epoch: 0034, Batch: 000003, Sample Num: 48, Cur Loss: 0.04728375, Cur Avg Loss: 0.08084646, Log Avg loss: 0.14804071, Global Avg Loss: 0.82879336, Time: 0.0248 Steps: 70260, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000013, Sample Num: 208, Cur Loss: 0.10917713, Cur Avg Loss: 0.14416322, Log Avg loss: 0.16315825, Global Avg Loss: 0.82869864, Time: 0.0213 Steps: 70270, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000023, Sample Num: 368, Cur Loss: 0.16173159, Cur Avg Loss: 0.15949974, Log Avg loss: 0.17943722, Global Avg Loss: 0.82860626, Time: 0.0212 Steps: 70280, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000033, Sample Num: 528, Cur Loss: 0.16469863, Cur Avg Loss: 0.16624801, Log Avg loss: 0.18176904, Global Avg Loss: 0.82851423, Time: 0.0212 Steps: 70290, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000043, Sample Num: 688, Cur Loss: 0.22381130, Cur Avg Loss: 0.16490132, Log Avg loss: 0.16045722, Global Avg Loss: 0.82841920, Time: 0.0212 Steps: 70300, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000053, Sample Num: 848, Cur Loss: 0.07867885, Cur Avg Loss: 0.17052960, Log Avg loss: 0.19473122, Global Avg Loss: 0.82832908, Time: 0.0212 Steps: 70310, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000063, Sample Num: 1008, Cur Loss: 0.22744934, Cur Avg Loss: 0.16857415, Log Avg loss: 0.15821023, Global Avg Loss: 0.82823378, Time: 0.0212 Steps: 70320, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000073, Sample Num: 1168, Cur Loss: 0.18149948, Cur Avg Loss: 0.16677323, Log Avg loss: 0.15542742, Global Avg Loss: 0.82813812, Time: 0.0212 Steps: 70330, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000083, Sample Num: 1328, Cur Loss: 0.08482064, Cur Avg Loss: 0.16202779, Log Avg loss: 0.12738615, Global Avg Loss: 0.82803849, Time: 0.0213 Steps: 70340, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000093, Sample Num: 1488, Cur Loss: 0.34284821, Cur Avg Loss: 0.16057684, Log Avg loss: 0.14853389, Global Avg Loss: 0.82794190, Time: 0.0212 Steps: 70350, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000103, Sample Num: 1648, Cur Loss: 0.19298926, Cur Avg Loss: 0.15936842, Log Avg loss: 0.14813017, Global Avg Loss: 0.82784528, Time: 0.0213 Steps: 70360, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000113, Sample Num: 1808, Cur Loss: 0.16140969, Cur Avg Loss: 0.16833169, Log Avg loss: 0.26065333, Global Avg Loss: 0.82776468, Time: 0.0212 Steps: 70370, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000123, Sample Num: 1968, Cur Loss: 0.56487948, Cur Avg Loss: 0.17586527, Log Avg loss: 0.26099472, Global Avg Loss: 0.82768415, Time: 0.0212 Steps: 70380, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000133, Sample Num: 2128, Cur Loss: 0.21226296, Cur Avg Loss: 0.17083766, Log Avg loss: 0.10899807, Global Avg Loss: 0.82758205, Time: 0.0212 Steps: 70390, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000143, Sample Num: 2288, Cur Loss: 0.45913026, Cur Avg Loss: 0.17173268, Log Avg loss: 0.18363641, Global Avg Loss: 0.82749058, Time: 0.0212 Steps: 70400, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000153, Sample Num: 2448, Cur Loss: 0.17266671, Cur Avg Loss: 0.17340648, Log Avg loss: 0.19734190, Global Avg Loss: 0.82740109, Time: 0.0212 Steps: 70410, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000163, Sample Num: 2608, Cur Loss: 0.48808137, Cur Avg Loss: 0.17838336, Log Avg loss: 0.25452963, Global Avg Loss: 0.82731973, Time: 0.0212 Steps: 70420, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000173, Sample Num: 2768, Cur Loss: 0.03340404, Cur Avg Loss: 0.17337274, Log Avg loss: 0.09169961, Global Avg Loss: 0.82721529, Time: 0.0213 Steps: 70430, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000183, Sample Num: 2928, Cur Loss: 0.09187943, Cur Avg Loss: 0.17454250, Log Avg loss: 0.19477927, Global Avg Loss: 0.82712550, Time: 0.0212 Steps: 70440, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000193, Sample Num: 3088, Cur Loss: 0.22925170, Cur Avg Loss: 0.17783399, Log Avg loss: 0.23806829, Global Avg Loss: 0.82704189, Time: 0.0212 Steps: 70450, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000203, Sample Num: 3248, Cur Loss: 0.13061850, Cur Avg Loss: 0.17893191, Log Avg loss: 0.20012175, Global Avg Loss: 0.82695292, Time: 0.0212 Steps: 70460, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000213, Sample Num: 3408, Cur Loss: 0.30001768, Cur Avg Loss: 0.18367412, Log Avg loss: 0.27994111, Global Avg Loss: 0.82687529, Time: 0.0212 Steps: 70470, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000223, Sample Num: 3568, Cur Loss: 0.34506959, Cur Avg Loss: 0.18408789, Log Avg loss: 0.19290101, Global Avg Loss: 0.82678534, Time: 0.0212 Steps: 70480, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000233, Sample Num: 3728, Cur Loss: 0.26133701, Cur Avg Loss: 0.18431613, Log Avg loss: 0.18940599, Global Avg Loss: 0.82669492, Time: 0.0212 Steps: 70490, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000243, Sample Num: 3888, Cur Loss: 0.10431598, Cur Avg Loss: 0.18286871, Log Avg loss: 0.14914390, Global Avg Loss: 0.82659881, Time: 0.0212 Steps: 70500, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000253, Sample Num: 4048, Cur Loss: 0.09050658, Cur Avg Loss: 0.18198607, Log Avg loss: 0.16053778, Global Avg Loss: 0.82650435, Time: 0.0212 Steps: 70510, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000263, Sample Num: 4208, Cur Loss: 0.44659793, Cur Avg Loss: 0.18383071, Log Avg loss: 0.23050006, Global Avg Loss: 0.82641983, Time: 0.0225 Steps: 70520, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000273, Sample Num: 4368, Cur Loss: 0.15169838, Cur Avg Loss: 0.18569611, Log Avg loss: 0.23475608, Global Avg Loss: 0.82633595, Time: 0.0210 Steps: 70530, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000283, Sample Num: 4528, Cur Loss: 0.13059938, Cur Avg Loss: 0.18483044, Log Avg loss: 0.16119779, Global Avg Loss: 0.82624165, Time: 0.0212 Steps: 70540, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000293, Sample Num: 4688, Cur Loss: 0.14033265, Cur Avg Loss: 0.18333131, Log Avg loss: 0.14090594, Global Avg Loss: 0.82614451, Time: 0.0211 Steps: 70550, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000303, Sample Num: 4848, Cur Loss: 0.60973501, Cur Avg Loss: 0.18602522, Log Avg loss: 0.26495685, Global Avg Loss: 0.82606498, Time: 0.0212 Steps: 70560, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000313, Sample Num: 5008, Cur Loss: 0.59463078, Cur Avg Loss: 0.18513681, Log Avg loss: 0.15821792, Global Avg Loss: 0.82597034, Time: 0.0211 Steps: 70570, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000323, Sample Num: 5168, Cur Loss: 0.09453903, Cur Avg Loss: 0.18470787, Log Avg loss: 0.17128213, Global Avg Loss: 0.82587758, Time: 0.0211 Steps: 70580, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000333, Sample Num: 5328, Cur Loss: 0.12678741, Cur Avg Loss: 0.18281304, Log Avg loss: 0.12160990, Global Avg Loss: 0.82577782, Time: 0.0211 Steps: 70590, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000343, Sample Num: 5488, Cur Loss: 0.38301882, Cur Avg Loss: 0.18292949, Log Avg loss: 0.18680712, Global Avg Loss: 0.82568731, Time: 0.0211 Steps: 70600, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000353, Sample Num: 5648, Cur Loss: 0.20661582, Cur Avg Loss: 0.18254525, Log Avg loss: 0.16936604, Global Avg Loss: 0.82559436, Time: 0.0211 Steps: 70610, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000363, Sample Num: 5808, Cur Loss: 0.17550677, Cur Avg Loss: 0.18158882, Log Avg loss: 0.14782693, Global Avg Loss: 0.82549839, Time: 0.0211 Steps: 70620, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000373, Sample Num: 5968, Cur Loss: 0.06187636, Cur Avg Loss: 0.17987967, Log Avg loss: 0.11783725, Global Avg Loss: 0.82539819, Time: 0.0212 Steps: 70630, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000383, Sample Num: 6128, Cur Loss: 0.12289935, Cur Avg Loss: 0.18020639, Log Avg loss: 0.19239327, Global Avg Loss: 0.82530858, Time: 0.0211 Steps: 70640, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000393, Sample Num: 6288, Cur Loss: 0.17908104, Cur Avg Loss: 0.17944281, Log Avg loss: 0.15019759, Global Avg Loss: 0.82521303, Time: 0.0211 Steps: 70650, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000403, Sample Num: 6448, Cur Loss: 0.17159510, Cur Avg Loss: 0.17752386, Log Avg loss: 0.10210921, Global Avg Loss: 0.82511069, Time: 0.0211 Steps: 70660, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000413, Sample Num: 6608, Cur Loss: 0.18343118, Cur Avg Loss: 0.17724365, Log Avg loss: 0.16595115, Global Avg Loss: 0.82501742, Time: 0.0211 Steps: 70670, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000423, Sample Num: 6768, Cur Loss: 0.14992750, Cur Avg Loss: 0.17706252, Log Avg loss: 0.16958175, Global Avg Loss: 0.82492468, Time: 0.0211 Steps: 70680, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000433, Sample Num: 6928, Cur Loss: 0.24316208, Cur Avg Loss: 0.17768669, Log Avg loss: 0.20408916, Global Avg Loss: 0.82483686, Time: 0.0211 Steps: 70690, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000443, Sample Num: 7088, Cur Loss: 0.12818563, Cur Avg Loss: 0.17775694, Log Avg loss: 0.18079889, Global Avg Loss: 0.82474576, Time: 0.0211 Steps: 70700, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000453, Sample Num: 7248, Cur Loss: 0.20386319, Cur Avg Loss: 0.17781913, Log Avg loss: 0.18057409, Global Avg Loss: 0.82465466, Time: 0.0212 Steps: 70710, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000463, Sample Num: 7408, Cur Loss: 0.06833591, Cur Avg Loss: 0.17833290, Log Avg loss: 0.20160667, Global Avg Loss: 0.82456656, Time: 0.0211 Steps: 70720, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000473, Sample Num: 7568, Cur Loss: 0.10150144, Cur Avg Loss: 0.17995332, Log Avg loss: 0.25497879, Global Avg Loss: 0.82448603, Time: 0.0211 Steps: 70730, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000483, Sample Num: 7728, Cur Loss: 0.04629161, Cur Avg Loss: 0.17926017, Log Avg loss: 0.14647403, Global Avg Loss: 0.82439019, Time: 0.0210 Steps: 70740, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000493, Sample Num: 7888, Cur Loss: 0.12904143, Cur Avg Loss: 0.17903607, Log Avg loss: 0.16821197, Global Avg Loss: 0.82429744, Time: 0.0211 Steps: 70750, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000503, Sample Num: 8048, Cur Loss: 0.25694570, Cur Avg Loss: 0.17980583, Log Avg loss: 0.21775486, Global Avg Loss: 0.82421172, Time: 0.0211 Steps: 70760, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000513, Sample Num: 8208, Cur Loss: 0.24691424, Cur Avg Loss: 0.18028489, Log Avg loss: 0.20438163, Global Avg Loss: 0.82412414, Time: 0.0246 Steps: 70770, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000523, Sample Num: 8368, Cur Loss: 0.30171698, Cur Avg Loss: 0.18017919, Log Avg loss: 0.17475693, Global Avg Loss: 0.82403240, Time: 0.0211 Steps: 70780, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000533, Sample Num: 8528, Cur Loss: 0.15788020, Cur Avg Loss: 0.17926081, Log Avg loss: 0.13122962, Global Avg Loss: 0.82393453, Time: 0.0211 Steps: 70790, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000543, Sample Num: 8688, Cur Loss: 0.12127402, Cur Avg Loss: 0.17860325, Log Avg loss: 0.14355538, Global Avg Loss: 0.82383843, Time: 0.0211 Steps: 70800, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000553, Sample Num: 8848, Cur Loss: 0.09512217, Cur Avg Loss: 0.17837308, Log Avg loss: 0.16587491, Global Avg Loss: 0.82374551, Time: 0.0211 Steps: 70810, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000563, Sample Num: 9008, Cur Loss: 0.42977971, Cur Avg Loss: 0.17904681, Log Avg loss: 0.21630408, Global Avg Loss: 0.82365974, Time: 0.0211 Steps: 70820, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000573, Sample Num: 9168, Cur Loss: 0.03334918, Cur Avg Loss: 0.17959106, Log Avg loss: 0.21023224, Global Avg Loss: 0.82357313, Time: 0.0211 Steps: 70830, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000583, Sample Num: 9328, Cur Loss: 0.15000986, Cur Avg Loss: 0.18055055, Log Avg loss: 0.23552916, Global Avg Loss: 0.82349012, Time: 0.0211 Steps: 70840, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000593, Sample Num: 9488, Cur Loss: 0.39374441, Cur Avg Loss: 0.18138993, Log Avg loss: 0.23032600, Global Avg Loss: 0.82340640, Time: 0.0211 Steps: 70850, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000603, Sample Num: 9648, Cur Loss: 0.32309294, Cur Avg Loss: 0.18071384, Log Avg loss: 0.14062187, Global Avg Loss: 0.82331004, Time: 0.0211 Steps: 70860, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000613, Sample Num: 9808, Cur Loss: 0.26688603, Cur Avg Loss: 0.18019816, Log Avg loss: 0.14910245, Global Avg Loss: 0.82321491, Time: 0.0211 Steps: 70870, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000623, Sample Num: 9968, Cur Loss: 0.27764329, Cur Avg Loss: 0.18015510, Log Avg loss: 0.17751539, Global Avg Loss: 0.82312381, Time: 0.0211 Steps: 70880, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000633, Sample Num: 10128, Cur Loss: 0.15157640, Cur Avg Loss: 0.18075119, Log Avg loss: 0.21788753, Global Avg Loss: 0.82303844, Time: 0.0211 Steps: 70890, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000643, Sample Num: 10288, Cur Loss: 0.06507012, Cur Avg Loss: 0.18033168, Log Avg loss: 0.15377690, Global Avg Loss: 0.82294404, Time: 0.0211 Steps: 70900, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000653, Sample Num: 10448, Cur Loss: 0.10462765, Cur Avg Loss: 0.17995467, Log Avg loss: 0.15571289, Global Avg Loss: 0.82284995, Time: 0.0211 Steps: 70910, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000663, Sample Num: 10608, Cur Loss: 0.21981087, Cur Avg Loss: 0.18034341, Log Avg loss: 0.20572806, Global Avg Loss: 0.82276293, Time: 0.0211 Steps: 70920, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000673, Sample Num: 10768, Cur Loss: 0.05993635, Cur Avg Loss: 0.17992696, Log Avg loss: 0.15231613, Global Avg Loss: 0.82266841, Time: 0.0211 Steps: 70930, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000683, Sample Num: 10928, Cur Loss: 0.11984789, Cur Avg Loss: 0.17915172, Log Avg loss: 0.12697854, Global Avg Loss: 0.82257034, Time: 0.0211 Steps: 70940, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000693, Sample Num: 11088, Cur Loss: 0.08474618, Cur Avg Loss: 0.17933243, Log Avg loss: 0.19167480, Global Avg Loss: 0.82248142, Time: 0.0211 Steps: 70950, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000703, Sample Num: 11248, Cur Loss: 0.16501293, Cur Avg Loss: 0.17953545, Log Avg loss: 0.19360427, Global Avg Loss: 0.82239279, Time: 0.0211 Steps: 70960, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000713, Sample Num: 11408, Cur Loss: 0.31433544, Cur Avg Loss: 0.18035089, Log Avg loss: 0.23767637, Global Avg Loss: 0.82231040, Time: 0.0211 Steps: 70970, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000723, Sample Num: 11568, Cur Loss: 0.13268319, Cur Avg Loss: 0.18152818, Log Avg loss: 0.26546921, Global Avg Loss: 0.82223195, Time: 0.0211 Steps: 70980, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000733, Sample Num: 11728, Cur Loss: 0.14993444, Cur Avg Loss: 0.18126610, Log Avg loss: 0.16231777, Global Avg Loss: 0.82213900, Time: 0.0211 Steps: 70990, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000743, Sample Num: 11888, Cur Loss: 0.10594189, Cur Avg Loss: 0.18182037, Log Avg loss: 0.22244846, Global Avg Loss: 0.82205453, Time: 0.0211 Steps: 71000, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000753, Sample Num: 12048, Cur Loss: 0.18021114, Cur Avg Loss: 0.18321162, Log Avg loss: 0.28658101, Global Avg Loss: 0.82197912, Time: 0.0211 Steps: 71010, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000763, Sample Num: 12208, Cur Loss: 0.26316005, Cur Avg Loss: 0.18313288, Log Avg loss: 0.17720424, Global Avg Loss: 0.82188834, Time: 0.0211 Steps: 71020, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000773, Sample Num: 12368, Cur Loss: 0.08409871, Cur Avg Loss: 0.18248243, Log Avg loss: 0.13285302, Global Avg Loss: 0.82179133, Time: 0.0215 Steps: 71030, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000783, Sample Num: 12528, Cur Loss: 0.14422452, Cur Avg Loss: 0.18195918, Log Avg loss: 0.14151187, Global Avg Loss: 0.82169557, Time: 0.0211 Steps: 71040, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000793, Sample Num: 12688, Cur Loss: 0.15178835, Cur Avg Loss: 0.18149904, Log Avg loss: 0.14547051, Global Avg Loss: 0.82160039, Time: 0.0210 Steps: 71050, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000803, Sample Num: 12848, Cur Loss: 0.08869985, Cur Avg Loss: 0.18117282, Log Avg loss: 0.15530338, Global Avg Loss: 0.82150663, Time: 0.0210 Steps: 71060, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000813, Sample Num: 13008, Cur Loss: 0.07565355, Cur Avg Loss: 0.18148452, Log Avg loss: 0.20651347, Global Avg Loss: 0.82142009, Time: 0.0210 Steps: 71070, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000823, Sample Num: 13168, Cur Loss: 0.07487267, Cur Avg Loss: 0.18238149, Log Avg loss: 0.25530538, Global Avg Loss: 0.82134045, Time: 0.0210 Steps: 71080, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000833, Sample Num: 13328, Cur Loss: 0.06297359, Cur Avg Loss: 0.18209446, Log Avg loss: 0.15847175, Global Avg Loss: 0.82124721, Time: 0.0210 Steps: 71090, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000843, Sample Num: 13488, Cur Loss: 0.05066637, Cur Avg Loss: 0.18139656, Log Avg loss: 0.12326180, Global Avg Loss: 0.82114904, Time: 0.0210 Steps: 71100, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000853, Sample Num: 13648, Cur Loss: 0.31683809, Cur Avg Loss: 0.18164741, Log Avg loss: 0.20279405, Global Avg Loss: 0.82106208, Time: 0.0211 Steps: 71110, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000863, Sample Num: 13808, Cur Loss: 0.15768304, Cur Avg Loss: 0.18180498, Log Avg loss: 0.19524588, Global Avg Loss: 0.82097409, Time: 0.0210 Steps: 71120, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000873, Sample Num: 13968, Cur Loss: 0.20041876, Cur Avg Loss: 0.18151273, Log Avg loss: 0.15629121, Global Avg Loss: 0.82088064, Time: 0.0210 Steps: 71130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000883, Sample Num: 14128, Cur Loss: 0.53705585, Cur Avg Loss: 0.18192368, Log Avg loss: 0.21779993, Global Avg Loss: 0.82079587, Time: 0.0210 Steps: 71140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000893, Sample Num: 14288, Cur Loss: 0.22251993, Cur Avg Loss: 0.18204556, Log Avg loss: 0.19280699, Global Avg Loss: 0.82070760, Time: 0.0210 Steps: 71150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000903, Sample Num: 14448, Cur Loss: 0.17711961, Cur Avg Loss: 0.18227928, Log Avg loss: 0.20315114, Global Avg Loss: 0.82062082, Time: 0.0210 Steps: 71160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000913, Sample Num: 14608, Cur Loss: 0.27267367, Cur Avg Loss: 0.18202522, Log Avg loss: 0.15908371, Global Avg Loss: 0.82052787, Time: 0.0210 Steps: 71170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000923, Sample Num: 14768, Cur Loss: 0.23534980, Cur Avg Loss: 0.18167548, Log Avg loss: 0.14974363, Global Avg Loss: 0.82043363, Time: 0.0210 Steps: 71180, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000933, Sample Num: 14928, Cur Loss: 0.22305548, Cur Avg Loss: 0.18268419, Log Avg loss: 0.27578836, Global Avg Loss: 0.82035712, Time: 0.0210 Steps: 71190, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000943, Sample Num: 15088, Cur Loss: 0.08849391, Cur Avg Loss: 0.18303584, Log Avg loss: 0.21584474, Global Avg Loss: 0.82027222, Time: 0.0210 Steps: 71200, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000953, Sample Num: 15248, Cur Loss: 0.10039859, Cur Avg Loss: 0.18293432, Log Avg loss: 0.17336060, Global Avg Loss: 0.82018137, Time: 0.0210 Steps: 71210, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000963, Sample Num: 15408, Cur Loss: 0.07441598, Cur Avg Loss: 0.18230317, Log Avg loss: 0.12215470, Global Avg Loss: 0.82008336, Time: 0.0210 Steps: 71220, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000973, Sample Num: 15568, Cur Loss: 0.13873462, Cur Avg Loss: 0.18221992, Log Avg loss: 0.17420290, Global Avg Loss: 0.81999269, Time: 0.0210 Steps: 71230, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000983, Sample Num: 15728, Cur Loss: 0.24542844, Cur Avg Loss: 0.18243641, Log Avg loss: 0.20350153, Global Avg Loss: 0.81990615, Time: 0.0210 Steps: 71240, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000993, Sample Num: 15888, Cur Loss: 0.46751300, Cur Avg Loss: 0.18251761, Log Avg loss: 0.19049889, Global Avg Loss: 0.81981781, Time: 0.0209 Steps: 71250, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001003, Sample Num: 16048, Cur Loss: 0.14516529, Cur Avg Loss: 0.18295312, Log Avg loss: 0.22619936, Global Avg Loss: 0.81973451, Time: 0.0210 Steps: 71260, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001013, Sample Num: 16208, Cur Loss: 0.15044841, Cur Avg Loss: 0.18256602, Log Avg loss: 0.14374007, Global Avg Loss: 0.81963966, Time: 0.0210 Steps: 71270, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001023, Sample Num: 16368, Cur Loss: 0.27677068, Cur Avg Loss: 0.18256350, Log Avg loss: 0.18230792, Global Avg Loss: 0.81955025, Time: 0.0210 Steps: 71280, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001033, Sample Num: 16528, Cur Loss: 0.10621598, Cur Avg Loss: 0.18276498, Log Avg loss: 0.20337713, Global Avg Loss: 0.81946382, Time: 0.0211 Steps: 71290, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001043, Sample Num: 16688, Cur Loss: 0.19502643, Cur Avg Loss: 0.18273379, Log Avg loss: 0.17951182, Global Avg Loss: 0.81937406, Time: 0.0211 Steps: 71300, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001053, Sample Num: 16848, Cur Loss: 0.14155367, Cur Avg Loss: 0.18249649, Log Avg loss: 0.15774623, Global Avg Loss: 0.81928128, Time: 0.0211 Steps: 71310, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001063, Sample Num: 17008, Cur Loss: 0.22723198, Cur Avg Loss: 0.18228402, Log Avg loss: 0.15991066, Global Avg Loss: 0.81918883, Time: 0.0211 Steps: 71320, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001073, Sample Num: 17168, Cur Loss: 0.41036037, Cur Avg Loss: 0.18235945, Log Avg loss: 0.19037794, Global Avg Loss: 0.81910067, Time: 0.0211 Steps: 71330, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001083, Sample Num: 17328, Cur Loss: 0.23980775, Cur Avg Loss: 0.18310846, Log Avg loss: 0.26347709, Global Avg Loss: 0.81902279, Time: 0.0211 Steps: 71340, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001093, Sample Num: 17488, Cur Loss: 0.15150066, Cur Avg Loss: 0.18389001, Log Avg loss: 0.26853190, Global Avg Loss: 0.81894563, Time: 0.0211 Steps: 71350, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001103, Sample Num: 17648, Cur Loss: 0.15300786, Cur Avg Loss: 0.18370319, Log Avg loss: 0.16328362, Global Avg Loss: 0.81885375, Time: 0.0211 Steps: 71360, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001113, Sample Num: 17808, Cur Loss: 0.32283753, Cur Avg Loss: 0.18368671, Log Avg loss: 0.18186907, Global Avg Loss: 0.81876450, Time: 0.0211 Steps: 71370, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001123, Sample Num: 17968, Cur Loss: 0.22770375, Cur Avg Loss: 0.18383163, Log Avg loss: 0.19996040, Global Avg Loss: 0.81867781, Time: 0.0211 Steps: 71380, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001133, Sample Num: 18128, Cur Loss: 0.14478761, Cur Avg Loss: 0.18375316, Log Avg loss: 0.17494151, Global Avg Loss: 0.81858764, Time: 0.0211 Steps: 71390, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001143, Sample Num: 18288, Cur Loss: 0.34401986, Cur Avg Loss: 0.18455624, Log Avg loss: 0.27554523, Global Avg Loss: 0.81851158, Time: 0.0211 Steps: 71400, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001153, Sample Num: 18448, Cur Loss: 0.22100443, Cur Avg Loss: 0.18457670, Log Avg loss: 0.18691509, Global Avg Loss: 0.81842314, Time: 0.0211 Steps: 71410, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001163, Sample Num: 18608, Cur Loss: 0.02599405, Cur Avg Loss: 0.18466376, Log Avg loss: 0.19470197, Global Avg Loss: 0.81833580, Time: 0.0211 Steps: 71420, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001173, Sample Num: 18768, Cur Loss: 0.34846133, Cur Avg Loss: 0.18532917, Log Avg loss: 0.26271653, Global Avg Loss: 0.81825802, Time: 0.0211 Steps: 71430, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001183, Sample Num: 18928, Cur Loss: 0.12164138, Cur Avg Loss: 0.18481670, Log Avg loss: 0.12470432, Global Avg Loss: 0.81816094, Time: 0.0211 Steps: 71440, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001193, Sample Num: 19088, Cur Loss: 0.16098540, Cur Avg Loss: 0.18465789, Log Avg loss: 0.16587009, Global Avg Loss: 0.81806964, Time: 0.0211 Steps: 71450, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001203, Sample Num: 19248, Cur Loss: 0.14699046, Cur Avg Loss: 0.18472025, Log Avg loss: 0.19215924, Global Avg Loss: 0.81798206, Time: 0.0210 Steps: 71460, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001213, Sample Num: 19408, Cur Loss: 0.08468589, Cur Avg Loss: 0.18434603, Log Avg loss: 0.13932735, Global Avg Loss: 0.81788710, Time: 0.0210 Steps: 71470, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001223, Sample Num: 19568, Cur Loss: 0.17662303, Cur Avg Loss: 0.18430695, Log Avg loss: 0.17956697, Global Avg Loss: 0.81779780, Time: 0.0211 Steps: 71480, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001233, Sample Num: 19728, Cur Loss: 0.08635943, Cur Avg Loss: 0.18439173, Log Avg loss: 0.19476029, Global Avg Loss: 0.81771065, Time: 0.0211 Steps: 71490, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001243, Sample Num: 19888, Cur Loss: 0.09380029, Cur Avg Loss: 0.18468829, Log Avg loss: 0.22125419, Global Avg Loss: 0.81762723, Time: 0.0210 Steps: 71500, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001253, Sample Num: 20048, Cur Loss: 0.22554193, Cur Avg Loss: 0.18476206, Log Avg loss: 0.19393197, Global Avg Loss: 0.81754001, Time: 0.0211 Steps: 71510, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001263, Sample Num: 20208, Cur Loss: 0.18986914, Cur Avg Loss: 0.18487582, Log Avg loss: 0.19912941, Global Avg Loss: 0.81745354, Time: 0.0211 Steps: 71520, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001273, Sample Num: 20368, Cur Loss: 0.12799317, Cur Avg Loss: 0.18460484, Log Avg loss: 0.15038042, Global Avg Loss: 0.81736029, Time: 0.0211 Steps: 71530, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001283, Sample Num: 20528, Cur Loss: 0.08084192, Cur Avg Loss: 0.18544003, Log Avg loss: 0.29175966, Global Avg Loss: 0.81728682, Time: 0.0248 Steps: 71540, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001293, Sample Num: 20688, Cur Loss: 0.07246511, Cur Avg Loss: 0.18536584, Log Avg loss: 0.17584770, Global Avg Loss: 0.81719717, Time: 0.0211 Steps: 71550, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001303, Sample Num: 20848, Cur Loss: 0.05137522, Cur Avg Loss: 0.18522600, Log Avg loss: 0.16714390, Global Avg Loss: 0.81710633, Time: 0.0211 Steps: 71560, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001313, Sample Num: 21008, Cur Loss: 0.04112096, Cur Avg Loss: 0.18539138, Log Avg loss: 0.20694054, Global Avg Loss: 0.81702107, Time: 0.0211 Steps: 71570, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001323, Sample Num: 21168, Cur Loss: 0.30384022, Cur Avg Loss: 0.18556402, Log Avg loss: 0.20823236, Global Avg Loss: 0.81693602, Time: 0.0211 Steps: 71580, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001333, Sample Num: 21328, Cur Loss: 0.06540094, Cur Avg Loss: 0.18561784, Log Avg loss: 0.19273820, Global Avg Loss: 0.81684883, Time: 0.0211 Steps: 71590, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001343, Sample Num: 21488, Cur Loss: 0.15902358, Cur Avg Loss: 0.18611181, Log Avg loss: 0.25195780, Global Avg Loss: 0.81676994, Time: 0.0211 Steps: 71600, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001353, Sample Num: 21648, Cur Loss: 0.13631415, Cur Avg Loss: 0.18617000, Log Avg loss: 0.19398480, Global Avg Loss: 0.81668297, Time: 0.0211 Steps: 71610, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001363, Sample Num: 21808, Cur Loss: 0.14044128, Cur Avg Loss: 0.18642093, Log Avg loss: 0.22037187, Global Avg Loss: 0.81659971, Time: 0.0211 Steps: 71620, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001373, Sample Num: 21968, Cur Loss: 0.22539733, Cur Avg Loss: 0.18653372, Log Avg loss: 0.20190722, Global Avg Loss: 0.81651389, Time: 0.0212 Steps: 71630, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001383, Sample Num: 22128, Cur Loss: 0.19704179, Cur Avg Loss: 0.18655205, Log Avg loss: 0.18906839, Global Avg Loss: 0.81642631, Time: 0.0211 Steps: 71640, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001393, Sample Num: 22288, Cur Loss: 0.27560127, Cur Avg Loss: 0.18675006, Log Avg loss: 0.21413519, Global Avg Loss: 0.81634225, Time: 0.0211 Steps: 71650, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001403, Sample Num: 22448, Cur Loss: 0.36687320, Cur Avg Loss: 0.18673010, Log Avg loss: 0.18394876, Global Avg Loss: 0.81625400, Time: 0.0210 Steps: 71660, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001413, Sample Num: 22608, Cur Loss: 0.12531564, Cur Avg Loss: 0.18690574, Log Avg loss: 0.21154798, Global Avg Loss: 0.81616963, Time: 0.0210 Steps: 71670, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001423, Sample Num: 22768, Cur Loss: 0.10156091, Cur Avg Loss: 0.18648520, Log Avg loss: 0.12706347, Global Avg Loss: 0.81607349, Time: 0.0211 Steps: 71680, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001433, Sample Num: 22928, Cur Loss: 0.23868757, Cur Avg Loss: 0.18666956, Log Avg loss: 0.21290471, Global Avg Loss: 0.81598935, Time: 0.0210 Steps: 71690, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001443, Sample Num: 23088, Cur Loss: 0.05568222, Cur Avg Loss: 0.18701356, Log Avg loss: 0.23630838, Global Avg Loss: 0.81590850, Time: 0.0210 Steps: 71700, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001453, Sample Num: 23248, Cur Loss: 0.08579537, Cur Avg Loss: 0.18728332, Log Avg loss: 0.22620909, Global Avg Loss: 0.81582627, Time: 0.0210 Steps: 71710, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001463, Sample Num: 23408, Cur Loss: 0.10008846, Cur Avg Loss: 0.18707602, Log Avg loss: 0.15695596, Global Avg Loss: 0.81573440, Time: 0.0210 Steps: 71720, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001473, Sample Num: 23568, Cur Loss: 0.08709494, Cur Avg Loss: 0.18703304, Log Avg loss: 0.18074438, Global Avg Loss: 0.81564588, Time: 0.0210 Steps: 71730, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001483, Sample Num: 23728, Cur Loss: 0.17564487, Cur Avg Loss: 0.18651938, Log Avg loss: 0.11085819, Global Avg Loss: 0.81554764, Time: 0.0213 Steps: 71740, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001493, Sample Num: 23888, Cur Loss: 0.11731137, Cur Avg Loss: 0.18672402, Log Avg loss: 0.21707223, Global Avg Loss: 0.81546423, Time: 0.0210 Steps: 71750, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001503, Sample Num: 24048, Cur Loss: 0.38521808, Cur Avg Loss: 0.18705819, Log Avg loss: 0.23694934, Global Avg Loss: 0.81538361, Time: 0.0210 Steps: 71760, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001513, Sample Num: 24208, Cur Loss: 0.08067439, Cur Avg Loss: 0.18701328, Log Avg loss: 0.18026245, Global Avg Loss: 0.81529511, Time: 0.0210 Steps: 71770, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001523, Sample Num: 24368, Cur Loss: 0.05816114, Cur Avg Loss: 0.18710893, Log Avg loss: 0.20158200, Global Avg Loss: 0.81520961, Time: 0.0210 Steps: 71780, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001533, Sample Num: 24528, Cur Loss: 0.39126685, Cur Avg Loss: 0.18716685, Log Avg loss: 0.19598675, Global Avg Loss: 0.81512336, Time: 0.0210 Steps: 71790, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001543, Sample Num: 24688, Cur Loss: 0.05392414, Cur Avg Loss: 0.18683825, Log Avg loss: 0.13646451, Global Avg Loss: 0.81502884, Time: 0.0245 Steps: 71800, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001553, Sample Num: 24848, Cur Loss: 0.16249529, Cur Avg Loss: 0.18668410, Log Avg loss: 0.16289806, Global Avg Loss: 0.81493803, Time: 0.0211 Steps: 71810, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001563, Sample Num: 25008, Cur Loss: 0.23374870, Cur Avg Loss: 0.18646184, Log Avg loss: 0.15194496, Global Avg Loss: 0.81484571, Time: 0.0211 Steps: 71820, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001573, Sample Num: 25168, Cur Loss: 0.09818757, Cur Avg Loss: 0.18675260, Log Avg loss: 0.23219945, Global Avg Loss: 0.81476460, Time: 0.0211 Steps: 71830, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001583, Sample Num: 25328, Cur Loss: 0.27502251, Cur Avg Loss: 0.18679815, Log Avg loss: 0.19396283, Global Avg Loss: 0.81467818, Time: 0.0211 Steps: 71840, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001593, Sample Num: 25488, Cur Loss: 0.13121656, Cur Avg Loss: 0.18658871, Log Avg loss: 0.15343503, Global Avg Loss: 0.81458615, Time: 0.0210 Steps: 71850, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001603, Sample Num: 25648, Cur Loss: 0.08103594, Cur Avg Loss: 0.18659516, Log Avg loss: 0.18762200, Global Avg Loss: 0.81449890, Time: 0.0210 Steps: 71860, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001613, Sample Num: 25808, Cur Loss: 0.03996589, Cur Avg Loss: 0.18620653, Log Avg loss: 0.12390959, Global Avg Loss: 0.81440282, Time: 0.0210 Steps: 71870, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001623, Sample Num: 25968, Cur Loss: 0.18241601, Cur Avg Loss: 0.18582174, Log Avg loss: 0.12375499, Global Avg Loss: 0.81430673, Time: 0.0210 Steps: 71880, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001633, Sample Num: 26128, Cur Loss: 0.33140278, Cur Avg Loss: 0.18604976, Log Avg loss: 0.22305617, Global Avg Loss: 0.81422449, Time: 0.0210 Steps: 71890, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001643, Sample Num: 26288, Cur Loss: 0.11732084, Cur Avg Loss: 0.18616430, Log Avg loss: 0.20486976, Global Avg Loss: 0.81413974, Time: 0.0210 Steps: 71900, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001653, Sample Num: 26448, Cur Loss: 0.23085131, Cur Avg Loss: 0.18607949, Log Avg loss: 0.17214457, Global Avg Loss: 0.81405046, Time: 0.0210 Steps: 71910, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001663, Sample Num: 26608, Cur Loss: 0.16263938, Cur Avg Loss: 0.18583816, Log Avg loss: 0.14594624, Global Avg Loss: 0.81395757, Time: 0.0211 Steps: 71920, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001673, Sample Num: 26768, Cur Loss: 0.15061232, Cur Avg Loss: 0.18597228, Log Avg loss: 0.20827696, Global Avg Loss: 0.81387336, Time: 0.0210 Steps: 71930, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001683, Sample Num: 26928, Cur Loss: 0.10399739, Cur Avg Loss: 0.18586228, Log Avg loss: 0.16745836, Global Avg Loss: 0.81378351, Time: 0.0210 Steps: 71940, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001693, Sample Num: 27088, Cur Loss: 0.25004295, Cur Avg Loss: 0.18556318, Log Avg loss: 0.13522510, Global Avg Loss: 0.81368920, Time: 0.0210 Steps: 71950, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001703, Sample Num: 27248, Cur Loss: 0.48896164, Cur Avg Loss: 0.18578431, Log Avg loss: 0.22322207, Global Avg Loss: 0.81360714, Time: 0.0210 Steps: 71960, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001713, Sample Num: 27408, Cur Loss: 0.09590433, Cur Avg Loss: 0.18570593, Log Avg loss: 0.17235831, Global Avg Loss: 0.81351804, Time: 0.0210 Steps: 71970, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001723, Sample Num: 27568, Cur Loss: 0.11184827, Cur Avg Loss: 0.18565748, Log Avg loss: 0.17735780, Global Avg Loss: 0.81342966, Time: 0.0210 Steps: 71980, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001733, Sample Num: 27728, Cur Loss: 0.52928883, Cur Avg Loss: 0.18598042, Log Avg loss: 0.24162197, Global Avg Loss: 0.81335023, Time: 0.0210 Steps: 71990, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001743, Sample Num: 27888, Cur Loss: 0.10777014, Cur Avg Loss: 0.18572689, Log Avg loss: 0.14179135, Global Avg Loss: 0.81325696, Time: 0.0210 Steps: 72000, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001753, Sample Num: 28048, Cur Loss: 0.35836834, Cur Avg Loss: 0.18554198, Log Avg loss: 0.15331207, Global Avg Loss: 0.81316531, Time: 0.0210 Steps: 72010, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001763, Sample Num: 28208, Cur Loss: 0.13269608, Cur Avg Loss: 0.18543548, Log Avg loss: 0.16676621, Global Avg Loss: 0.81307556, Time: 0.0210 Steps: 72020, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001773, Sample Num: 28368, Cur Loss: 0.05175067, Cur Avg Loss: 0.18544977, Log Avg loss: 0.18796924, Global Avg Loss: 0.81298878, Time: 0.0210 Steps: 72030, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001783, Sample Num: 28528, Cur Loss: 0.14389251, Cur Avg Loss: 0.18578232, Log Avg loss: 0.24474347, Global Avg Loss: 0.81290990, Time: 0.0210 Steps: 72040, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001793, Sample Num: 28688, Cur Loss: 0.15247205, Cur Avg Loss: 0.18629408, Log Avg loss: 0.27753959, Global Avg Loss: 0.81283559, Time: 0.0248 Steps: 72050, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001803, Sample Num: 28848, Cur Loss: 0.57763577, Cur Avg Loss: 0.18665217, Log Avg loss: 0.25085902, Global Avg Loss: 0.81275761, Time: 0.0211 Steps: 72060, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001813, Sample Num: 29008, Cur Loss: 0.38266551, Cur Avg Loss: 0.18739432, Log Avg loss: 0.32120305, Global Avg Loss: 0.81268940, Time: 0.0211 Steps: 72070, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001823, Sample Num: 29168, Cur Loss: 0.15942344, Cur Avg Loss: 0.18788967, Log Avg loss: 0.27769741, Global Avg Loss: 0.81261518, Time: 0.0211 Steps: 72080, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001833, Sample Num: 29328, Cur Loss: 0.19521412, Cur Avg Loss: 0.18813333, Log Avg loss: 0.23255114, Global Avg Loss: 0.81253472, Time: 0.0211 Steps: 72090, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001843, Sample Num: 29488, Cur Loss: 0.17933467, Cur Avg Loss: 0.18799343, Log Avg loss: 0.16234999, Global Avg Loss: 0.81244454, Time: 0.0211 Steps: 72100, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001853, Sample Num: 29648, Cur Loss: 0.21643883, Cur Avg Loss: 0.18808935, Log Avg loss: 0.20576724, Global Avg Loss: 0.81236040, Time: 0.0211 Steps: 72110, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001863, Sample Num: 29808, Cur Loss: 0.14213338, Cur Avg Loss: 0.18834859, Log Avg loss: 0.23638562, Global Avg Loss: 0.81228054, Time: 0.0211 Steps: 72120, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001873, Sample Num: 29968, Cur Loss: 0.03964256, Cur Avg Loss: 0.18834078, Log Avg loss: 0.18688697, Global Avg Loss: 0.81219384, Time: 0.0211 Steps: 72130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001883, Sample Num: 30128, Cur Loss: 0.09794718, Cur Avg Loss: 0.18848579, Log Avg loss: 0.21564655, Global Avg Loss: 0.81211114, Time: 0.0211 Steps: 72140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001893, Sample Num: 30288, Cur Loss: 0.23076144, Cur Avg Loss: 0.18864815, Log Avg loss: 0.21921952, Global Avg Loss: 0.81202897, Time: 0.0211 Steps: 72150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001903, Sample Num: 30448, Cur Loss: 0.22027820, Cur Avg Loss: 0.18837992, Log Avg loss: 0.13760474, Global Avg Loss: 0.81193551, Time: 0.0211 Steps: 72160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001913, Sample Num: 30608, Cur Loss: 0.08303040, Cur Avg Loss: 0.18855798, Log Avg loss: 0.22244190, Global Avg Loss: 0.81185383, Time: 0.0211 Steps: 72170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001923, Sample Num: 30768, Cur Loss: 0.13973793, Cur Avg Loss: 0.18870100, Log Avg loss: 0.21606025, Global Avg Loss: 0.81177128, Time: 0.0210 Steps: 72180, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001933, Sample Num: 30928, Cur Loss: 0.12135660, Cur Avg Loss: 0.18897093, Log Avg loss: 0.24088018, Global Avg Loss: 0.81169220, Time: 0.0210 Steps: 72190, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001943, Sample Num: 31088, Cur Loss: 0.04399031, Cur Avg Loss: 0.18895795, Log Avg loss: 0.18644843, Global Avg Loss: 0.81160560, Time: 0.0211 Steps: 72200, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001953, Sample Num: 31248, Cur Loss: 0.17181501, Cur Avg Loss: 0.18863858, Log Avg loss: 0.12658422, Global Avg Loss: 0.81151074, Time: 0.0211 Steps: 72210, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001963, Sample Num: 31408, Cur Loss: 0.12073743, Cur Avg Loss: 0.18835243, Log Avg loss: 0.13246803, Global Avg Loss: 0.81141671, Time: 0.0211 Steps: 72220, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001973, Sample Num: 31568, Cur Loss: 0.20969534, Cur Avg Loss: 0.18841006, Log Avg loss: 0.19972242, Global Avg Loss: 0.81133203, Time: 0.0211 Steps: 72230, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001983, Sample Num: 31728, Cur Loss: 0.09439611, Cur Avg Loss: 0.18866593, Log Avg loss: 0.23914979, Global Avg Loss: 0.81125282, Time: 0.0211 Steps: 72240, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001993, Sample Num: 31888, Cur Loss: 0.13700214, Cur Avg Loss: 0.18884214, Log Avg loss: 0.22378402, Global Avg Loss: 0.81117151, Time: 0.0211 Steps: 72250, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002003, Sample Num: 32048, Cur Loss: 0.55573291, Cur Avg Loss: 0.18893310, Log Avg loss: 0.20706086, Global Avg Loss: 0.81108791, Time: 0.0211 Steps: 72260, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002013, Sample Num: 32208, Cur Loss: 0.07258154, Cur Avg Loss: 0.18894716, Log Avg loss: 0.19176475, Global Avg Loss: 0.81100221, Time: 0.0210 Steps: 72270, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002023, Sample Num: 32368, Cur Loss: 0.30688193, Cur Avg Loss: 0.18887862, Log Avg loss: 0.17508074, Global Avg Loss: 0.81091423, Time: 0.0210 Steps: 72280, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002033, Sample Num: 32528, Cur Loss: 0.17093399, Cur Avg Loss: 0.18919500, Log Avg loss: 0.25319789, Global Avg Loss: 0.81083708, Time: 0.0211 Steps: 72290, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002043, Sample Num: 32688, Cur Loss: 0.46010333, Cur Avg Loss: 0.18946916, Log Avg loss: 0.24520588, Global Avg Loss: 0.81075885, Time: 0.0211 Steps: 72300, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002053, Sample Num: 32848, Cur Loss: 0.05791818, Cur Avg Loss: 0.18942480, Log Avg loss: 0.18036286, Global Avg Loss: 0.81067167, Time: 0.0240 Steps: 72310, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002063, Sample Num: 33008, Cur Loss: 0.11460593, Cur Avg Loss: 0.18924822, Log Avg loss: 0.15299526, Global Avg Loss: 0.81058073, Time: 0.0210 Steps: 72320, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002073, Sample Num: 33168, Cur Loss: 0.11395270, Cur Avg Loss: 0.18910704, Log Avg loss: 0.15998290, Global Avg Loss: 0.81049078, Time: 0.0210 Steps: 72330, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002083, Sample Num: 33328, Cur Loss: 0.16827142, Cur Avg Loss: 0.18916467, Log Avg loss: 0.20111057, Global Avg Loss: 0.81040654, Time: 0.0210 Steps: 72340, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002093, Sample Num: 33488, Cur Loss: 0.25655439, Cur Avg Loss: 0.18931820, Log Avg loss: 0.22129990, Global Avg Loss: 0.81032512, Time: 0.0211 Steps: 72350, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002103, Sample Num: 33648, Cur Loss: 0.12356232, Cur Avg Loss: 0.18901190, Log Avg loss: 0.12490212, Global Avg Loss: 0.81023039, Time: 0.0210 Steps: 72360, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002113, Sample Num: 33808, Cur Loss: 0.16959414, Cur Avg Loss: 0.18878320, Log Avg loss: 0.14068817, Global Avg Loss: 0.81013788, Time: 0.0210 Steps: 72370, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002123, Sample Num: 33968, Cur Loss: 0.13516816, Cur Avg Loss: 0.18874413, Log Avg loss: 0.18048743, Global Avg Loss: 0.81005088, Time: 0.0210 Steps: 72380, Updated lr: 0.000032 ***** Running evaluation checkpoint-72386 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-72386 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.061402, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.272245, "eval_total_loss": 191.388383, "eval_mae": 0.349723, "eval_mse": 0.27235, "eval_r2": 0.826877, "eval_sp_statistic": 0.886141, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.916552, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.795213, "test_total_loss": 399.196985, "test_mae": 0.689772, "test_mse": 0.795306, "test_r2": 0.486702, "test_sp_statistic": 0.79992, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.841047, "test_ps_pvalue": 0.0, "lr": 3.2303461356092936e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8099930192576587, "train_cur_epoch_loss": 401.3754615429789, "train_cur_epoch_avg_loss": 0.18852769447767914, "train_cur_epoch_time": 45.06140160560608, "train_cur_epoch_avg_time": 0.021165524474216102, "epoch": 34, "step": 72386} ################################################## Training, Epoch: 0035, Batch: 000004, Sample Num: 64, Cur Loss: 0.22885442, Cur Avg Loss: 0.34529120, Log Avg loss: 0.20528470, Global Avg Loss: 0.80996734, Time: 0.0248 Steps: 72390, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000014, Sample Num: 224, Cur Loss: 0.10476349, Cur Avg Loss: 0.17266499, Log Avg loss: 0.10361450, Global Avg Loss: 0.80986978, Time: 0.0211 Steps: 72400, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000024, Sample Num: 384, Cur Loss: 0.17430267, Cur Avg Loss: 0.15362261, Log Avg loss: 0.12696328, Global Avg Loss: 0.80977547, Time: 0.0211 Steps: 72410, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000034, Sample Num: 544, Cur Loss: 0.05147922, Cur Avg Loss: 0.16595833, Log Avg loss: 0.19556407, Global Avg Loss: 0.80969066, Time: 0.0211 Steps: 72420, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000044, Sample Num: 704, Cur Loss: 0.19710621, Cur Avg Loss: 0.17050758, Log Avg loss: 0.18597500, Global Avg Loss: 0.80960454, Time: 0.0211 Steps: 72430, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000054, Sample Num: 864, Cur Loss: 0.17503546, Cur Avg Loss: 0.16335491, Log Avg loss: 0.13188318, Global Avg Loss: 0.80951099, Time: 0.0210 Steps: 72440, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000064, Sample Num: 1024, Cur Loss: 0.12995264, Cur Avg Loss: 0.18078897, Log Avg loss: 0.27493291, Global Avg Loss: 0.80943720, Time: 0.0211 Steps: 72450, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000074, Sample Num: 1184, Cur Loss: 0.82793814, Cur Avg Loss: 0.19435683, Log Avg loss: 0.28119110, Global Avg Loss: 0.80936430, Time: 0.0211 Steps: 72460, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000084, Sample Num: 1344, Cur Loss: 0.43301177, Cur Avg Loss: 0.20123702, Log Avg loss: 0.25215043, Global Avg Loss: 0.80928741, Time: 0.0211 Steps: 72470, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000094, Sample Num: 1504, Cur Loss: 0.16696301, Cur Avg Loss: 0.19435108, Log Avg loss: 0.13650922, Global Avg Loss: 0.80919459, Time: 0.0210 Steps: 72480, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000104, Sample Num: 1664, Cur Loss: 0.20774138, Cur Avg Loss: 0.19203485, Log Avg loss: 0.17026230, Global Avg Loss: 0.80910645, Time: 0.0211 Steps: 72490, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000114, Sample Num: 1824, Cur Loss: 0.35518843, Cur Avg Loss: 0.18697772, Log Avg loss: 0.13438351, Global Avg Loss: 0.80901338, Time: 0.0211 Steps: 72500, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000124, Sample Num: 1984, Cur Loss: 0.15111452, Cur Avg Loss: 0.18589525, Log Avg loss: 0.17355509, Global Avg Loss: 0.80892574, Time: 0.0211 Steps: 72510, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000134, Sample Num: 2144, Cur Loss: 0.60303468, Cur Avg Loss: 0.18770560, Log Avg loss: 0.21015395, Global Avg Loss: 0.80884318, Time: 0.0211 Steps: 72520, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000144, Sample Num: 2304, Cur Loss: 0.28624979, Cur Avg Loss: 0.19034118, Log Avg loss: 0.22565794, Global Avg Loss: 0.80876277, Time: 0.0210 Steps: 72530, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000154, Sample Num: 2464, Cur Loss: 0.02973545, Cur Avg Loss: 0.19293633, Log Avg loss: 0.23030655, Global Avg Loss: 0.80868303, Time: 0.0211 Steps: 72540, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000164, Sample Num: 2624, Cur Loss: 0.12439814, Cur Avg Loss: 0.18787439, Log Avg loss: 0.10992043, Global Avg Loss: 0.80858671, Time: 0.0210 Steps: 72550, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000174, Sample Num: 2784, Cur Loss: 0.41500601, Cur Avg Loss: 0.18890389, Log Avg loss: 0.20578775, Global Avg Loss: 0.80850364, Time: 0.0211 Steps: 72560, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000184, Sample Num: 2944, Cur Loss: 0.72760177, Cur Avg Loss: 0.19035068, Log Avg loss: 0.21552486, Global Avg Loss: 0.80842193, Time: 0.0210 Steps: 72570, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000194, Sample Num: 3104, Cur Loss: 0.06334193, Cur Avg Loss: 0.18978989, Log Avg loss: 0.17947129, Global Avg Loss: 0.80833527, Time: 0.0211 Steps: 72580, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000204, Sample Num: 3264, Cur Loss: 0.02275735, Cur Avg Loss: 0.18862008, Log Avg loss: 0.16592581, Global Avg Loss: 0.80824677, Time: 0.0211 Steps: 72590, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000214, Sample Num: 3424, Cur Loss: 0.08303587, Cur Avg Loss: 0.18957113, Log Avg loss: 0.20897251, Global Avg Loss: 0.80816423, Time: 0.0211 Steps: 72600, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000224, Sample Num: 3584, Cur Loss: 0.28818873, Cur Avg Loss: 0.19139365, Log Avg loss: 0.23039560, Global Avg Loss: 0.80808466, Time: 0.0211 Steps: 72610, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000234, Sample Num: 3744, Cur Loss: 0.20922868, Cur Avg Loss: 0.18952507, Log Avg loss: 0.14766877, Global Avg Loss: 0.80799371, Time: 0.0211 Steps: 72620, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000244, Sample Num: 3904, Cur Loss: 0.20448980, Cur Avg Loss: 0.18731477, Log Avg loss: 0.13559387, Global Avg Loss: 0.80790114, Time: 0.0210 Steps: 72630, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000254, Sample Num: 4064, Cur Loss: 0.19782960, Cur Avg Loss: 0.18529411, Log Avg loss: 0.13599002, Global Avg Loss: 0.80780864, Time: 0.0210 Steps: 72640, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000264, Sample Num: 4224, Cur Loss: 0.10799877, Cur Avg Loss: 0.18457467, Log Avg loss: 0.16630087, Global Avg Loss: 0.80772034, Time: 0.0211 Steps: 72650, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000274, Sample Num: 4384, Cur Loss: 0.06745590, Cur Avg Loss: 0.18210481, Log Avg loss: 0.11690037, Global Avg Loss: 0.80762526, Time: 0.0212 Steps: 72660, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000284, Sample Num: 4544, Cur Loss: 0.13452266, Cur Avg Loss: 0.18288003, Log Avg loss: 0.20412115, Global Avg Loss: 0.80754221, Time: 0.0210 Steps: 72670, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000294, Sample Num: 4704, Cur Loss: 0.29851544, Cur Avg Loss: 0.18150799, Log Avg loss: 0.14254206, Global Avg Loss: 0.80745072, Time: 0.0211 Steps: 72680, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000304, Sample Num: 4864, Cur Loss: 0.22496024, Cur Avg Loss: 0.18174512, Log Avg loss: 0.18871671, Global Avg Loss: 0.80736560, Time: 0.0210 Steps: 72690, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000314, Sample Num: 5024, Cur Loss: 0.25528973, Cur Avg Loss: 0.18165895, Log Avg loss: 0.17903930, Global Avg Loss: 0.80727917, Time: 0.0210 Steps: 72700, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000324, Sample Num: 5184, Cur Loss: 0.16073905, Cur Avg Loss: 0.18305465, Log Avg loss: 0.22687984, Global Avg Loss: 0.80719935, Time: 0.0210 Steps: 72710, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000334, Sample Num: 5344, Cur Loss: 0.26499960, Cur Avg Loss: 0.18464687, Log Avg loss: 0.23623489, Global Avg Loss: 0.80712083, Time: 0.0210 Steps: 72720, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000344, Sample Num: 5504, Cur Loss: 0.08686496, Cur Avg Loss: 0.18298512, Log Avg loss: 0.12748267, Global Avg Loss: 0.80702738, Time: 0.0210 Steps: 72730, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000354, Sample Num: 5664, Cur Loss: 0.12620850, Cur Avg Loss: 0.18272131, Log Avg loss: 0.17364615, Global Avg Loss: 0.80694031, Time: 0.0210 Steps: 72740, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000364, Sample Num: 5824, Cur Loss: 0.09694639, Cur Avg Loss: 0.18303457, Log Avg loss: 0.19412377, Global Avg Loss: 0.80685607, Time: 0.0211 Steps: 72750, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000374, Sample Num: 5984, Cur Loss: 0.03986564, Cur Avg Loss: 0.18308335, Log Avg loss: 0.18485915, Global Avg Loss: 0.80677059, Time: 0.0210 Steps: 72760, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000384, Sample Num: 6144, Cur Loss: 0.10067654, Cur Avg Loss: 0.18211615, Log Avg loss: 0.14594278, Global Avg Loss: 0.80667978, Time: 0.0210 Steps: 72770, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000394, Sample Num: 6304, Cur Loss: 0.06196161, Cur Avg Loss: 0.18075087, Log Avg loss: 0.12832403, Global Avg Loss: 0.80658657, Time: 0.0210 Steps: 72780, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000404, Sample Num: 6464, Cur Loss: 0.28507733, Cur Avg Loss: 0.18110292, Log Avg loss: 0.19497391, Global Avg Loss: 0.80650255, Time: 0.0211 Steps: 72790, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000414, Sample Num: 6624, Cur Loss: 0.09782131, Cur Avg Loss: 0.17976140, Log Avg loss: 0.12556375, Global Avg Loss: 0.80640901, Time: 0.0210 Steps: 72800, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000424, Sample Num: 6784, Cur Loss: 0.22666842, Cur Avg Loss: 0.18247498, Log Avg loss: 0.29481725, Global Avg Loss: 0.80633875, Time: 0.0210 Steps: 72810, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000434, Sample Num: 6944, Cur Loss: 0.07470256, Cur Avg Loss: 0.18277370, Log Avg loss: 0.19543958, Global Avg Loss: 0.80625485, Time: 0.0210 Steps: 72820, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000444, Sample Num: 7104, Cur Loss: 0.12732878, Cur Avg Loss: 0.18228626, Log Avg loss: 0.16113114, Global Avg Loss: 0.80616627, Time: 0.0210 Steps: 72830, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000454, Sample Num: 7264, Cur Loss: 0.21543418, Cur Avg Loss: 0.18175822, Log Avg loss: 0.15831335, Global Avg Loss: 0.80607733, Time: 0.0210 Steps: 72840, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000464, Sample Num: 7424, Cur Loss: 0.21811017, Cur Avg Loss: 0.18176417, Log Avg loss: 0.18203427, Global Avg Loss: 0.80599167, Time: 0.0211 Steps: 72850, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000474, Sample Num: 7584, Cur Loss: 0.06477091, Cur Avg Loss: 0.18060745, Log Avg loss: 0.12693550, Global Avg Loss: 0.80589847, Time: 0.0210 Steps: 72860, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000484, Sample Num: 7744, Cur Loss: 0.05804100, Cur Avg Loss: 0.18040090, Log Avg loss: 0.17061078, Global Avg Loss: 0.80581129, Time: 0.0210 Steps: 72870, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000494, Sample Num: 7904, Cur Loss: 0.09500131, Cur Avg Loss: 0.17938083, Log Avg loss: 0.13000935, Global Avg Loss: 0.80571856, Time: 0.0210 Steps: 72880, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000504, Sample Num: 8064, Cur Loss: 0.08124714, Cur Avg Loss: 0.17889861, Log Avg loss: 0.15507667, Global Avg Loss: 0.80562930, Time: 0.0211 Steps: 72890, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000514, Sample Num: 8224, Cur Loss: 0.06526563, Cur Avg Loss: 0.17902472, Log Avg loss: 0.18538104, Global Avg Loss: 0.80554422, Time: 0.0249 Steps: 72900, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000524, Sample Num: 8384, Cur Loss: 0.20970502, Cur Avg Loss: 0.17854655, Log Avg loss: 0.15396862, Global Avg Loss: 0.80545485, Time: 0.0211 Steps: 72910, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000534, Sample Num: 8544, Cur Loss: 0.13130465, Cur Avg Loss: 0.17890306, Log Avg loss: 0.19758420, Global Avg Loss: 0.80537149, Time: 0.0212 Steps: 72920, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000544, Sample Num: 8704, Cur Loss: 0.11767189, Cur Avg Loss: 0.17977917, Log Avg loss: 0.22656308, Global Avg Loss: 0.80529212, Time: 0.0211 Steps: 72930, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000554, Sample Num: 8864, Cur Loss: 0.13284758, Cur Avg Loss: 0.17932736, Log Avg loss: 0.15474926, Global Avg Loss: 0.80520293, Time: 0.0211 Steps: 72940, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000564, Sample Num: 9024, Cur Loss: 0.19556683, Cur Avg Loss: 0.18003185, Log Avg loss: 0.21906069, Global Avg Loss: 0.80512259, Time: 0.0211 Steps: 72950, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000574, Sample Num: 9184, Cur Loss: 0.26382631, Cur Avg Loss: 0.17896655, Log Avg loss: 0.11888343, Global Avg Loss: 0.80502853, Time: 0.0211 Steps: 72960, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000584, Sample Num: 9344, Cur Loss: 0.11935447, Cur Avg Loss: 0.17940430, Log Avg loss: 0.20453123, Global Avg Loss: 0.80494624, Time: 0.0212 Steps: 72970, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000594, Sample Num: 9504, Cur Loss: 0.09421998, Cur Avg Loss: 0.17942150, Log Avg loss: 0.18042600, Global Avg Loss: 0.80486066, Time: 0.0211 Steps: 72980, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000604, Sample Num: 9664, Cur Loss: 0.03903448, Cur Avg Loss: 0.18017286, Log Avg loss: 0.22480319, Global Avg Loss: 0.80478119, Time: 0.0211 Steps: 72990, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000614, Sample Num: 9824, Cur Loss: 0.11505066, Cur Avg Loss: 0.18058455, Log Avg loss: 0.20545097, Global Avg Loss: 0.80469909, Time: 0.0212 Steps: 73000, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000624, Sample Num: 9984, Cur Loss: 0.22870488, Cur Avg Loss: 0.18003428, Log Avg loss: 0.14624778, Global Avg Loss: 0.80460890, Time: 0.0212 Steps: 73010, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000634, Sample Num: 10144, Cur Loss: 0.57040405, Cur Avg Loss: 0.18094467, Log Avg loss: 0.23775313, Global Avg Loss: 0.80453127, Time: 0.0211 Steps: 73020, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000644, Sample Num: 10304, Cur Loss: 0.32383627, Cur Avg Loss: 0.18094389, Log Avg loss: 0.18089451, Global Avg Loss: 0.80444588, Time: 0.0212 Steps: 73030, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000654, Sample Num: 10464, Cur Loss: 0.31541517, Cur Avg Loss: 0.18074939, Log Avg loss: 0.16822358, Global Avg Loss: 0.80435877, Time: 0.0211 Steps: 73040, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000664, Sample Num: 10624, Cur Loss: 0.08583254, Cur Avg Loss: 0.18049982, Log Avg loss: 0.16417776, Global Avg Loss: 0.80427114, Time: 0.0211 Steps: 73050, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000674, Sample Num: 10784, Cur Loss: 0.10349318, Cur Avg Loss: 0.17969244, Log Avg loss: 0.12608203, Global Avg Loss: 0.80417831, Time: 0.0211 Steps: 73060, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000684, Sample Num: 10944, Cur Loss: 0.18140689, Cur Avg Loss: 0.17940602, Log Avg loss: 0.16010152, Global Avg Loss: 0.80409017, Time: 0.0211 Steps: 73070, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000694, Sample Num: 11104, Cur Loss: 0.24825227, Cur Avg Loss: 0.17950788, Log Avg loss: 0.18647481, Global Avg Loss: 0.80400565, Time: 0.0211 Steps: 73080, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000704, Sample Num: 11264, Cur Loss: 0.29407862, Cur Avg Loss: 0.17906993, Log Avg loss: 0.14867666, Global Avg Loss: 0.80391599, Time: 0.0212 Steps: 73090, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000714, Sample Num: 11424, Cur Loss: 0.55697143, Cur Avg Loss: 0.17924137, Log Avg loss: 0.19131088, Global Avg Loss: 0.80383219, Time: 0.0211 Steps: 73100, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000724, Sample Num: 11584, Cur Loss: 0.16534527, Cur Avg Loss: 0.17925295, Log Avg loss: 0.18007972, Global Avg Loss: 0.80374687, Time: 0.0212 Steps: 73110, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000734, Sample Num: 11744, Cur Loss: 0.08504939, Cur Avg Loss: 0.17889480, Log Avg loss: 0.15296473, Global Avg Loss: 0.80365787, Time: 0.0211 Steps: 73120, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000744, Sample Num: 11904, Cur Loss: 0.22292581, Cur Avg Loss: 0.17883123, Log Avg loss: 0.17416476, Global Avg Loss: 0.80357179, Time: 0.0211 Steps: 73130, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000754, Sample Num: 12064, Cur Loss: 0.17720646, Cur Avg Loss: 0.17876521, Log Avg loss: 0.17385394, Global Avg Loss: 0.80348569, Time: 0.0212 Steps: 73140, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000764, Sample Num: 12224, Cur Loss: 0.13360001, Cur Avg Loss: 0.18033155, Log Avg loss: 0.29843328, Global Avg Loss: 0.80341665, Time: 0.0212 Steps: 73150, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000774, Sample Num: 12384, Cur Loss: 0.51802468, Cur Avg Loss: 0.18114050, Log Avg loss: 0.24294452, Global Avg Loss: 0.80334004, Time: 0.0212 Steps: 73160, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000784, Sample Num: 12544, Cur Loss: 0.31071696, Cur Avg Loss: 0.18176568, Log Avg loss: 0.23015407, Global Avg Loss: 0.80326171, Time: 0.0210 Steps: 73170, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000794, Sample Num: 12704, Cur Loss: 0.20731360, Cur Avg Loss: 0.18291562, Log Avg loss: 0.27307105, Global Avg Loss: 0.80318926, Time: 0.0210 Steps: 73180, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000804, Sample Num: 12864, Cur Loss: 0.14007431, Cur Avg Loss: 0.18274651, Log Avg loss: 0.16931968, Global Avg Loss: 0.80310265, Time: 0.0211 Steps: 73190, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000814, Sample Num: 13024, Cur Loss: 0.04370702, Cur Avg Loss: 0.18216139, Log Avg loss: 0.13511773, Global Avg Loss: 0.80301139, Time: 0.0210 Steps: 73200, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000824, Sample Num: 13184, Cur Loss: 0.09861961, Cur Avg Loss: 0.18148066, Log Avg loss: 0.12606854, Global Avg Loss: 0.80291893, Time: 0.0211 Steps: 73210, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000834, Sample Num: 13344, Cur Loss: 0.40622258, Cur Avg Loss: 0.18198768, Log Avg loss: 0.22376632, Global Avg Loss: 0.80283983, Time: 0.0211 Steps: 73220, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000844, Sample Num: 13504, Cur Loss: 0.14501989, Cur Avg Loss: 0.18246776, Log Avg loss: 0.22250633, Global Avg Loss: 0.80276058, Time: 0.0210 Steps: 73230, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000854, Sample Num: 13664, Cur Loss: 0.37352026, Cur Avg Loss: 0.18273685, Log Avg loss: 0.20544850, Global Avg Loss: 0.80267903, Time: 0.0211 Steps: 73240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000864, Sample Num: 13824, Cur Loss: 0.06919786, Cur Avg Loss: 0.18213060, Log Avg loss: 0.13035669, Global Avg Loss: 0.80258724, Time: 0.0211 Steps: 73250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000874, Sample Num: 13984, Cur Loss: 0.25504822, Cur Avg Loss: 0.18222393, Log Avg loss: 0.19028767, Global Avg Loss: 0.80250366, Time: 0.0210 Steps: 73260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000884, Sample Num: 14144, Cur Loss: 0.22801875, Cur Avg Loss: 0.18199160, Log Avg loss: 0.16168621, Global Avg Loss: 0.80241620, Time: 0.0211 Steps: 73270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000894, Sample Num: 14304, Cur Loss: 0.11470118, Cur Avg Loss: 0.18177589, Log Avg loss: 0.16270661, Global Avg Loss: 0.80232891, Time: 0.0211 Steps: 73280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000904, Sample Num: 14464, Cur Loss: 0.06333013, Cur Avg Loss: 0.18172922, Log Avg loss: 0.17755733, Global Avg Loss: 0.80224366, Time: 0.0210 Steps: 73290, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000914, Sample Num: 14624, Cur Loss: 0.10825910, Cur Avg Loss: 0.18153938, Log Avg loss: 0.16437761, Global Avg Loss: 0.80215664, Time: 0.0210 Steps: 73300, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000924, Sample Num: 14784, Cur Loss: 0.08842278, Cur Avg Loss: 0.18067248, Log Avg loss: 0.10143769, Global Avg Loss: 0.80206106, Time: 0.0210 Steps: 73310, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000934, Sample Num: 14944, Cur Loss: 0.50942820, Cur Avg Loss: 0.18077084, Log Avg loss: 0.18985934, Global Avg Loss: 0.80197756, Time: 0.0211 Steps: 73320, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000944, Sample Num: 15104, Cur Loss: 0.12739110, Cur Avg Loss: 0.18077676, Log Avg loss: 0.18133016, Global Avg Loss: 0.80189292, Time: 0.0210 Steps: 73330, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000954, Sample Num: 15264, Cur Loss: 0.14188157, Cur Avg Loss: 0.18078755, Log Avg loss: 0.18180614, Global Avg Loss: 0.80180837, Time: 0.0211 Steps: 73340, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000964, Sample Num: 15424, Cur Loss: 0.15885836, Cur Avg Loss: 0.18069216, Log Avg loss: 0.17159153, Global Avg Loss: 0.80172245, Time: 0.0211 Steps: 73350, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000974, Sample Num: 15584, Cur Loss: 0.11262685, Cur Avg Loss: 0.18033141, Log Avg loss: 0.14555473, Global Avg Loss: 0.80163301, Time: 0.0210 Steps: 73360, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000984, Sample Num: 15744, Cur Loss: 0.40823883, Cur Avg Loss: 0.18119347, Log Avg loss: 0.26515827, Global Avg Loss: 0.80155989, Time: 0.0210 Steps: 73370, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000994, Sample Num: 15904, Cur Loss: 0.28820056, Cur Avg Loss: 0.18150179, Log Avg loss: 0.21184083, Global Avg Loss: 0.80147952, Time: 0.0210 Steps: 73380, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001004, Sample Num: 16064, Cur Loss: 0.27859539, Cur Avg Loss: 0.18166150, Log Avg loss: 0.19753638, Global Avg Loss: 0.80139723, Time: 0.0210 Steps: 73390, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001014, Sample Num: 16224, Cur Loss: 0.18803503, Cur Avg Loss: 0.18160361, Log Avg loss: 0.17579182, Global Avg Loss: 0.80131200, Time: 0.0210 Steps: 73400, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001024, Sample Num: 16384, Cur Loss: 0.11588912, Cur Avg Loss: 0.18183113, Log Avg loss: 0.20490129, Global Avg Loss: 0.80123076, Time: 0.0255 Steps: 73410, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001034, Sample Num: 16544, Cur Loss: 0.05188354, Cur Avg Loss: 0.18137209, Log Avg loss: 0.13436630, Global Avg Loss: 0.80113993, Time: 0.0210 Steps: 73420, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001044, Sample Num: 16704, Cur Loss: 0.14986175, Cur Avg Loss: 0.18167011, Log Avg loss: 0.21248598, Global Avg Loss: 0.80105976, Time: 0.0210 Steps: 73430, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001054, Sample Num: 16864, Cur Loss: 0.09336463, Cur Avg Loss: 0.18130831, Log Avg loss: 0.14353584, Global Avg Loss: 0.80097023, Time: 0.0211 Steps: 73440, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001064, Sample Num: 17024, Cur Loss: 0.01894762, Cur Avg Loss: 0.18197313, Log Avg loss: 0.25204493, Global Avg Loss: 0.80089549, Time: 0.0211 Steps: 73450, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001074, Sample Num: 17184, Cur Loss: 0.10615692, Cur Avg Loss: 0.18171054, Log Avg loss: 0.15377124, Global Avg Loss: 0.80080740, Time: 0.0211 Steps: 73460, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001084, Sample Num: 17344, Cur Loss: 0.14365353, Cur Avg Loss: 0.18181545, Log Avg loss: 0.19308287, Global Avg Loss: 0.80072469, Time: 0.0211 Steps: 73470, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001094, Sample Num: 17504, Cur Loss: 0.37571257, Cur Avg Loss: 0.18163558, Log Avg loss: 0.16213764, Global Avg Loss: 0.80063778, Time: 0.0211 Steps: 73480, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001104, Sample Num: 17664, Cur Loss: 0.11026129, Cur Avg Loss: 0.18238752, Log Avg loss: 0.26464976, Global Avg Loss: 0.80056485, Time: 0.0210 Steps: 73490, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001114, Sample Num: 17824, Cur Loss: 0.11453950, Cur Avg Loss: 0.18268076, Log Avg loss: 0.21505437, Global Avg Loss: 0.80048518, Time: 0.0211 Steps: 73500, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001124, Sample Num: 17984, Cur Loss: 0.36709875, Cur Avg Loss: 0.18272427, Log Avg loss: 0.18757166, Global Avg Loss: 0.80040181, Time: 0.0210 Steps: 73510, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001134, Sample Num: 18144, Cur Loss: 0.12681344, Cur Avg Loss: 0.18321821, Log Avg loss: 0.23873674, Global Avg Loss: 0.80032541, Time: 0.0211 Steps: 73520, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001144, Sample Num: 18304, Cur Loss: 0.28885558, Cur Avg Loss: 0.18372281, Log Avg loss: 0.24094467, Global Avg Loss: 0.80024933, Time: 0.0211 Steps: 73530, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001154, Sample Num: 18464, Cur Loss: 0.05620130, Cur Avg Loss: 0.18357589, Log Avg loss: 0.16676775, Global Avg Loss: 0.80016319, Time: 0.0212 Steps: 73540, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001164, Sample Num: 18624, Cur Loss: 0.17070131, Cur Avg Loss: 0.18355377, Log Avg loss: 0.18100106, Global Avg Loss: 0.80007901, Time: 0.0210 Steps: 73550, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001174, Sample Num: 18784, Cur Loss: 0.28894588, Cur Avg Loss: 0.18362383, Log Avg loss: 0.19177935, Global Avg Loss: 0.79999632, Time: 0.0211 Steps: 73560, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001184, Sample Num: 18944, Cur Loss: 0.15932554, Cur Avg Loss: 0.18353168, Log Avg loss: 0.17271356, Global Avg Loss: 0.79991105, Time: 0.0211 Steps: 73570, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001194, Sample Num: 19104, Cur Loss: 0.12303247, Cur Avg Loss: 0.18344510, Log Avg loss: 0.17319416, Global Avg Loss: 0.79982588, Time: 0.0210 Steps: 73580, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001204, Sample Num: 19264, Cur Loss: 0.21401018, Cur Avg Loss: 0.18357481, Log Avg loss: 0.19906209, Global Avg Loss: 0.79974424, Time: 0.0211 Steps: 73590, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001214, Sample Num: 19424, Cur Loss: 0.20655909, Cur Avg Loss: 0.18354675, Log Avg loss: 0.18016842, Global Avg Loss: 0.79966006, Time: 0.0211 Steps: 73600, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001224, Sample Num: 19584, Cur Loss: 0.27093148, Cur Avg Loss: 0.18381449, Log Avg loss: 0.21631716, Global Avg Loss: 0.79958081, Time: 0.0211 Steps: 73610, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001234, Sample Num: 19744, Cur Loss: 0.04148027, Cur Avg Loss: 0.18354098, Log Avg loss: 0.15006407, Global Avg Loss: 0.79949259, Time: 0.0211 Steps: 73620, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001244, Sample Num: 19904, Cur Loss: 0.15155099, Cur Avg Loss: 0.18395161, Log Avg loss: 0.23462342, Global Avg Loss: 0.79941587, Time: 0.0210 Steps: 73630, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001254, Sample Num: 20064, Cur Loss: 0.17481963, Cur Avg Loss: 0.18389778, Log Avg loss: 0.17720082, Global Avg Loss: 0.79933138, Time: 0.0211 Steps: 73640, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001264, Sample Num: 20224, Cur Loss: 0.14659107, Cur Avg Loss: 0.18381730, Log Avg loss: 0.17372576, Global Avg Loss: 0.79924643, Time: 0.0210 Steps: 73650, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001274, Sample Num: 20384, Cur Loss: 0.13252838, Cur Avg Loss: 0.18387522, Log Avg loss: 0.19119623, Global Avg Loss: 0.79916388, Time: 0.0210 Steps: 73660, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001284, Sample Num: 20544, Cur Loss: 0.26199955, Cur Avg Loss: 0.18384739, Log Avg loss: 0.18030146, Global Avg Loss: 0.79907988, Time: 0.0212 Steps: 73670, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001294, Sample Num: 20704, Cur Loss: 0.43484941, Cur Avg Loss: 0.18387221, Log Avg loss: 0.18705954, Global Avg Loss: 0.79899682, Time: 0.0211 Steps: 73680, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001304, Sample Num: 20864, Cur Loss: 0.31224933, Cur Avg Loss: 0.18415832, Log Avg loss: 0.22118004, Global Avg Loss: 0.79891840, Time: 0.0211 Steps: 73690, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001314, Sample Num: 21024, Cur Loss: 0.26393002, Cur Avg Loss: 0.18381197, Log Avg loss: 0.13864906, Global Avg Loss: 0.79882881, Time: 0.0211 Steps: 73700, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001324, Sample Num: 21184, Cur Loss: 0.20322162, Cur Avg Loss: 0.18421646, Log Avg loss: 0.23736597, Global Avg Loss: 0.79875264, Time: 0.0214 Steps: 73710, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001334, Sample Num: 21344, Cur Loss: 0.05806578, Cur Avg Loss: 0.18425624, Log Avg loss: 0.18952250, Global Avg Loss: 0.79867000, Time: 0.0211 Steps: 73720, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001344, Sample Num: 21504, Cur Loss: 0.14733161, Cur Avg Loss: 0.18449373, Log Avg loss: 0.21617560, Global Avg Loss: 0.79859100, Time: 0.0211 Steps: 73730, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001354, Sample Num: 21664, Cur Loss: 0.45028150, Cur Avg Loss: 0.18505876, Log Avg loss: 0.26099816, Global Avg Loss: 0.79851809, Time: 0.0211 Steps: 73740, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001364, Sample Num: 21824, Cur Loss: 0.25861663, Cur Avg Loss: 0.18521649, Log Avg loss: 0.20657359, Global Avg Loss: 0.79843783, Time: 0.0210 Steps: 73750, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001374, Sample Num: 21984, Cur Loss: 0.09233502, Cur Avg Loss: 0.18482590, Log Avg loss: 0.13154961, Global Avg Loss: 0.79834742, Time: 0.0211 Steps: 73760, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001384, Sample Num: 22144, Cur Loss: 0.23638889, Cur Avg Loss: 0.18497893, Log Avg loss: 0.20600554, Global Avg Loss: 0.79826712, Time: 0.0211 Steps: 73770, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001394, Sample Num: 22304, Cur Loss: 0.12304744, Cur Avg Loss: 0.18498466, Log Avg loss: 0.18577739, Global Avg Loss: 0.79818411, Time: 0.0211 Steps: 73780, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001404, Sample Num: 22464, Cur Loss: 0.22882630, Cur Avg Loss: 0.18517235, Log Avg loss: 0.21133657, Global Avg Loss: 0.79810458, Time: 0.0211 Steps: 73790, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001414, Sample Num: 22624, Cur Loss: 0.51448208, Cur Avg Loss: 0.18538754, Log Avg loss: 0.21559998, Global Avg Loss: 0.79802565, Time: 0.0211 Steps: 73800, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001424, Sample Num: 22784, Cur Loss: 0.11693838, Cur Avg Loss: 0.18546260, Log Avg loss: 0.19607582, Global Avg Loss: 0.79794409, Time: 0.0211 Steps: 73810, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001434, Sample Num: 22944, Cur Loss: 0.27211589, Cur Avg Loss: 0.18522412, Log Avg loss: 0.15126456, Global Avg Loss: 0.79785649, Time: 0.0211 Steps: 73820, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001444, Sample Num: 23104, Cur Loss: 0.39302605, Cur Avg Loss: 0.18516178, Log Avg loss: 0.17622293, Global Avg Loss: 0.79777229, Time: 0.0211 Steps: 73830, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001454, Sample Num: 23264, Cur Loss: 0.08287193, Cur Avg Loss: 0.18479884, Log Avg loss: 0.13238906, Global Avg Loss: 0.79768218, Time: 0.0210 Steps: 73840, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001464, Sample Num: 23424, Cur Loss: 0.20505743, Cur Avg Loss: 0.18471436, Log Avg loss: 0.17243215, Global Avg Loss: 0.79759752, Time: 0.0211 Steps: 73850, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001474, Sample Num: 23584, Cur Loss: 0.17354932, Cur Avg Loss: 0.18440862, Log Avg loss: 0.13964733, Global Avg Loss: 0.79750843, Time: 0.0211 Steps: 73860, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001484, Sample Num: 23744, Cur Loss: 0.06165913, Cur Avg Loss: 0.18390012, Log Avg loss: 0.10894766, Global Avg Loss: 0.79741522, Time: 0.0211 Steps: 73870, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001494, Sample Num: 23904, Cur Loss: 0.17961213, Cur Avg Loss: 0.18420873, Log Avg loss: 0.23000588, Global Avg Loss: 0.79733842, Time: 0.0211 Steps: 73880, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001504, Sample Num: 24064, Cur Loss: 0.12327429, Cur Avg Loss: 0.18414062, Log Avg loss: 0.17396507, Global Avg Loss: 0.79725406, Time: 0.0211 Steps: 73890, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001514, Sample Num: 24224, Cur Loss: 0.10411527, Cur Avg Loss: 0.18381209, Log Avg loss: 0.13440106, Global Avg Loss: 0.79716436, Time: 0.0211 Steps: 73900, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001524, Sample Num: 24384, Cur Loss: 0.12881547, Cur Avg Loss: 0.18391549, Log Avg loss: 0.19957142, Global Avg Loss: 0.79708351, Time: 0.0211 Steps: 73910, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001534, Sample Num: 24544, Cur Loss: 0.36323905, Cur Avg Loss: 0.18429347, Log Avg loss: 0.24189623, Global Avg Loss: 0.79700840, Time: 0.0211 Steps: 73920, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001544, Sample Num: 24704, Cur Loss: 0.12568593, Cur Avg Loss: 0.18453862, Log Avg loss: 0.22214587, Global Avg Loss: 0.79693064, Time: 0.0211 Steps: 73930, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001554, Sample Num: 24864, Cur Loss: 0.11458485, Cur Avg Loss: 0.18439564, Log Avg loss: 0.16231901, Global Avg Loss: 0.79684481, Time: 0.0209 Steps: 73940, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001564, Sample Num: 25024, Cur Loss: 0.05762670, Cur Avg Loss: 0.18394563, Log Avg loss: 0.11401358, Global Avg Loss: 0.79675248, Time: 0.0208 Steps: 73950, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001574, Sample Num: 25184, Cur Loss: 0.05202914, Cur Avg Loss: 0.18421246, Log Avg loss: 0.22594478, Global Avg Loss: 0.79667530, Time: 0.0208 Steps: 73960, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001584, Sample Num: 25344, Cur Loss: 0.26480824, Cur Avg Loss: 0.18403520, Log Avg loss: 0.15613549, Global Avg Loss: 0.79658870, Time: 0.0209 Steps: 73970, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001594, Sample Num: 25504, Cur Loss: 0.07326708, Cur Avg Loss: 0.18422045, Log Avg loss: 0.21356304, Global Avg Loss: 0.79650990, Time: 0.0209 Steps: 73980, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001604, Sample Num: 25664, Cur Loss: 0.13786766, Cur Avg Loss: 0.18407441, Log Avg loss: 0.16079582, Global Avg Loss: 0.79642398, Time: 0.0209 Steps: 73990, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001614, Sample Num: 25824, Cur Loss: 0.23338082, Cur Avg Loss: 0.18368836, Log Avg loss: 0.12176619, Global Avg Loss: 0.79633281, Time: 0.0209 Steps: 74000, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001624, Sample Num: 25984, Cur Loss: 0.35164404, Cur Avg Loss: 0.18377371, Log Avg loss: 0.19754915, Global Avg Loss: 0.79625190, Time: 0.0209 Steps: 74010, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001634, Sample Num: 26144, Cur Loss: 0.50420117, Cur Avg Loss: 0.18371012, Log Avg loss: 0.17338290, Global Avg Loss: 0.79616775, Time: 0.0209 Steps: 74020, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001644, Sample Num: 26304, Cur Loss: 0.10147838, Cur Avg Loss: 0.18358733, Log Avg loss: 0.16352389, Global Avg Loss: 0.79608229, Time: 0.0209 Steps: 74030, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001654, Sample Num: 26464, Cur Loss: 0.15299669, Cur Avg Loss: 0.18367273, Log Avg loss: 0.19771195, Global Avg Loss: 0.79600148, Time: 0.0209 Steps: 74040, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001664, Sample Num: 26624, Cur Loss: 0.19788995, Cur Avg Loss: 0.18404633, Log Avg loss: 0.24583957, Global Avg Loss: 0.79592718, Time: 0.0208 Steps: 74050, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001674, Sample Num: 26784, Cur Loss: 0.15928508, Cur Avg Loss: 0.18425745, Log Avg loss: 0.21938844, Global Avg Loss: 0.79584933, Time: 0.0209 Steps: 74060, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001684, Sample Num: 26944, Cur Loss: 0.14346379, Cur Avg Loss: 0.18395135, Log Avg loss: 0.13271050, Global Avg Loss: 0.79575981, Time: 0.0209 Steps: 74070, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001694, Sample Num: 27104, Cur Loss: 0.12358180, Cur Avg Loss: 0.18371581, Log Avg loss: 0.14405085, Global Avg Loss: 0.79567183, Time: 0.0209 Steps: 74080, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001704, Sample Num: 27264, Cur Loss: 0.11569229, Cur Avg Loss: 0.18352899, Log Avg loss: 0.15188154, Global Avg Loss: 0.79558494, Time: 0.0209 Steps: 74090, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001714, Sample Num: 27424, Cur Loss: 0.48110488, Cur Avg Loss: 0.18419643, Log Avg loss: 0.29792861, Global Avg Loss: 0.79551778, Time: 0.0209 Steps: 74100, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001724, Sample Num: 27584, Cur Loss: 0.07600585, Cur Avg Loss: 0.18409601, Log Avg loss: 0.16688279, Global Avg Loss: 0.79543295, Time: 0.0209 Steps: 74110, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001734, Sample Num: 27744, Cur Loss: 0.36684582, Cur Avg Loss: 0.18412402, Log Avg loss: 0.18895272, Global Avg Loss: 0.79535113, Time: 0.0209 Steps: 74120, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001744, Sample Num: 27904, Cur Loss: 0.30848077, Cur Avg Loss: 0.18433570, Log Avg loss: 0.22104258, Global Avg Loss: 0.79527366, Time: 0.0208 Steps: 74130, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001754, Sample Num: 28064, Cur Loss: 0.13797443, Cur Avg Loss: 0.18424264, Log Avg loss: 0.16801196, Global Avg Loss: 0.79518905, Time: 0.0209 Steps: 74140, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001764, Sample Num: 28224, Cur Loss: 0.06790899, Cur Avg Loss: 0.18405982, Log Avg loss: 0.15199382, Global Avg Loss: 0.79510231, Time: 0.0209 Steps: 74150, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001774, Sample Num: 28384, Cur Loss: 0.10922785, Cur Avg Loss: 0.18393534, Log Avg loss: 0.16197710, Global Avg Loss: 0.79501694, Time: 0.0209 Steps: 74160, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001784, Sample Num: 28544, Cur Loss: 0.18648702, Cur Avg Loss: 0.18375678, Log Avg loss: 0.15207987, Global Avg Loss: 0.79493025, Time: 0.0209 Steps: 74170, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001794, Sample Num: 28704, Cur Loss: 0.36687768, Cur Avg Loss: 0.18396645, Log Avg loss: 0.22137069, Global Avg Loss: 0.79485293, Time: 0.0247 Steps: 74180, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001804, Sample Num: 28864, Cur Loss: 0.12973773, Cur Avg Loss: 0.18396621, Log Avg loss: 0.18392314, Global Avg Loss: 0.79477059, Time: 0.0208 Steps: 74190, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001814, Sample Num: 29024, Cur Loss: 0.45739752, Cur Avg Loss: 0.18435245, Log Avg loss: 0.25403110, Global Avg Loss: 0.79469771, Time: 0.0209 Steps: 74200, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001824, Sample Num: 29184, Cur Loss: 0.30679092, Cur Avg Loss: 0.18462900, Log Avg loss: 0.23479435, Global Avg Loss: 0.79462226, Time: 0.0209 Steps: 74210, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001834, Sample Num: 29344, Cur Loss: 0.38729778, Cur Avg Loss: 0.18465089, Log Avg loss: 0.18864419, Global Avg Loss: 0.79454061, Time: 0.0209 Steps: 74220, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001844, Sample Num: 29504, Cur Loss: 0.09121062, Cur Avg Loss: 0.18451975, Log Avg loss: 0.16046946, Global Avg Loss: 0.79445519, Time: 0.0210 Steps: 74230, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001854, Sample Num: 29664, Cur Loss: 0.41350076, Cur Avg Loss: 0.18456676, Log Avg loss: 0.19323486, Global Avg Loss: 0.79437421, Time: 0.0209 Steps: 74240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001864, Sample Num: 29824, Cur Loss: 0.23747376, Cur Avg Loss: 0.18491675, Log Avg loss: 0.24980531, Global Avg Loss: 0.79430087, Time: 0.0208 Steps: 74250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001874, Sample Num: 29984, Cur Loss: 0.11279880, Cur Avg Loss: 0.18511548, Log Avg loss: 0.22215750, Global Avg Loss: 0.79422382, Time: 0.0209 Steps: 74260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001884, Sample Num: 30144, Cur Loss: 0.07500062, Cur Avg Loss: 0.18530104, Log Avg loss: 0.22007660, Global Avg Loss: 0.79414652, Time: 0.0208 Steps: 74270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001894, Sample Num: 30304, Cur Loss: 0.04000239, Cur Avg Loss: 0.18520175, Log Avg loss: 0.16649455, Global Avg Loss: 0.79406202, Time: 0.0209 Steps: 74280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001904, Sample Num: 30464, Cur Loss: 0.22835916, Cur Avg Loss: 0.18563402, Log Avg loss: 0.26750589, Global Avg Loss: 0.79399114, Time: 0.0208 Steps: 74290, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001914, Sample Num: 30624, Cur Loss: 0.05359607, Cur Avg Loss: 0.18558651, Log Avg loss: 0.17654117, Global Avg Loss: 0.79390804, Time: 0.0209 Steps: 74300, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001924, Sample Num: 30784, Cur Loss: 0.06272581, Cur Avg Loss: 0.18540750, Log Avg loss: 0.15114435, Global Avg Loss: 0.79382154, Time: 0.0209 Steps: 74310, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001934, Sample Num: 30944, Cur Loss: 0.15817815, Cur Avg Loss: 0.18565850, Log Avg loss: 0.23395088, Global Avg Loss: 0.79374621, Time: 0.0209 Steps: 74320, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001944, Sample Num: 31104, Cur Loss: 0.19176111, Cur Avg Loss: 0.18593924, Log Avg loss: 0.24023498, Global Avg Loss: 0.79367174, Time: 0.0209 Steps: 74330, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001954, Sample Num: 31264, Cur Loss: 0.24161416, Cur Avg Loss: 0.18627166, Log Avg loss: 0.25089457, Global Avg Loss: 0.79359873, Time: 0.0209 Steps: 74340, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001964, Sample Num: 31424, Cur Loss: 0.15032786, Cur Avg Loss: 0.18645329, Log Avg loss: 0.22194365, Global Avg Loss: 0.79352184, Time: 0.0209 Steps: 74350, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001974, Sample Num: 31584, Cur Loss: 0.09019783, Cur Avg Loss: 0.18669969, Log Avg loss: 0.23509268, Global Avg Loss: 0.79344674, Time: 0.0209 Steps: 74360, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001984, Sample Num: 31744, Cur Loss: 0.19225070, Cur Avg Loss: 0.18724804, Log Avg loss: 0.29549262, Global Avg Loss: 0.79337979, Time: 0.0209 Steps: 74370, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001994, Sample Num: 31904, Cur Loss: 0.31953213, Cur Avg Loss: 0.18728238, Log Avg loss: 0.19409444, Global Avg Loss: 0.79329922, Time: 0.0209 Steps: 74380, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002004, Sample Num: 32064, Cur Loss: 0.06391956, Cur Avg Loss: 0.18702596, Log Avg loss: 0.13589567, Global Avg Loss: 0.79321084, Time: 0.0209 Steps: 74390, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002014, Sample Num: 32224, Cur Loss: 0.13887337, Cur Avg Loss: 0.18683540, Log Avg loss: 0.14864817, Global Avg Loss: 0.79312421, Time: 0.0209 Steps: 74400, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002024, Sample Num: 32384, Cur Loss: 0.19948436, Cur Avg Loss: 0.18664072, Log Avg loss: 0.14743118, Global Avg Loss: 0.79303743, Time: 0.0209 Steps: 74410, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002034, Sample Num: 32544, Cur Loss: 0.06553243, Cur Avg Loss: 0.18648755, Log Avg loss: 0.15548704, Global Avg Loss: 0.79295177, Time: 0.0209 Steps: 74420, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002044, Sample Num: 32704, Cur Loss: 0.33871469, Cur Avg Loss: 0.18671731, Log Avg loss: 0.23344989, Global Avg Loss: 0.79287659, Time: 0.0209 Steps: 74430, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002054, Sample Num: 32864, Cur Loss: 0.07110383, Cur Avg Loss: 0.18682215, Log Avg loss: 0.20825179, Global Avg Loss: 0.79279806, Time: 0.0234 Steps: 74440, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002064, Sample Num: 33024, Cur Loss: 0.20485561, Cur Avg Loss: 0.18675947, Log Avg loss: 0.17388427, Global Avg Loss: 0.79271493, Time: 0.0209 Steps: 74450, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002074, Sample Num: 33184, Cur Loss: 0.11106312, Cur Avg Loss: 0.18654657, Log Avg loss: 0.14260433, Global Avg Loss: 0.79262762, Time: 0.0212 Steps: 74460, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002084, Sample Num: 33344, Cur Loss: 0.13212663, Cur Avg Loss: 0.18628037, Log Avg loss: 0.13107085, Global Avg Loss: 0.79253878, Time: 0.0209 Steps: 74470, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002094, Sample Num: 33504, Cur Loss: 0.53177166, Cur Avg Loss: 0.18627616, Log Avg loss: 0.18539903, Global Avg Loss: 0.79245726, Time: 0.0209 Steps: 74480, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002104, Sample Num: 33664, Cur Loss: 0.59966099, Cur Avg Loss: 0.18647304, Log Avg loss: 0.22769978, Global Avg Loss: 0.79238145, Time: 0.0209 Steps: 74490, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002114, Sample Num: 33824, Cur Loss: 0.06276667, Cur Avg Loss: 0.18631363, Log Avg loss: 0.15277351, Global Avg Loss: 0.79229559, Time: 0.0209 Steps: 74500, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002124, Sample Num: 33984, Cur Loss: 0.09525186, Cur Avg Loss: 0.18612209, Log Avg loss: 0.14563044, Global Avg Loss: 0.79220880, Time: 0.0209 Steps: 74510, Updated lr: 0.000030 ***** Running evaluation checkpoint-74515 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-74515 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.933558, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.255822, "eval_total_loss": 179.842636, "eval_mae": 0.366146, "eval_mse": 0.255895, "eval_r2": 0.837336, "eval_sp_statistic": 0.888794, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.917949, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.63389, "test_total_loss": 318.212784, "test_mae": 0.582762, "test_mse": 0.633999, "test_r2": 0.590811, "test_sp_statistic": 0.797132, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.833099, "test_ps_pvalue": 0.0, "lr": 3.0284495021337127e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.792166996526895, "train_cur_epoch_loss": 396.16905421670526, "train_cur_epoch_avg_loss": 0.1860822236809325, "train_cur_epoch_time": 44.93355846405029, "train_cur_epoch_avg_time": 0.02110547602820587, "epoch": 35, "step": 74515} ################################################## Training, Epoch: 0036, Batch: 000005, Sample Num: 80, Cur Loss: 0.07826135, Cur Avg Loss: 0.16397206, Log Avg loss: 0.16655893, Global Avg Loss: 0.79212485, Time: 0.0247 Steps: 74520, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000015, Sample Num: 240, Cur Loss: 0.20132042, Cur Avg Loss: 0.17021860, Log Avg loss: 0.17334187, Global Avg Loss: 0.79204182, Time: 0.0209 Steps: 74530, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000025, Sample Num: 400, Cur Loss: 0.04144116, Cur Avg Loss: 0.16726246, Log Avg loss: 0.16282826, Global Avg Loss: 0.79195741, Time: 0.0209 Steps: 74540, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000035, Sample Num: 560, Cur Loss: 0.14374888, Cur Avg Loss: 0.16740219, Log Avg loss: 0.16775151, Global Avg Loss: 0.79187368, Time: 0.0209 Steps: 74550, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000045, Sample Num: 720, Cur Loss: 0.16388603, Cur Avg Loss: 0.17224882, Log Avg loss: 0.18921204, Global Avg Loss: 0.79179285, Time: 0.0209 Steps: 74560, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000055, Sample Num: 880, Cur Loss: 0.14432581, Cur Avg Loss: 0.19998480, Log Avg loss: 0.32479670, Global Avg Loss: 0.79173023, Time: 0.0209 Steps: 74570, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000065, Sample Num: 1040, Cur Loss: 0.06134919, Cur Avg Loss: 0.18844448, Log Avg loss: 0.12497272, Global Avg Loss: 0.79164082, Time: 0.0209 Steps: 74580, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000075, Sample Num: 1200, Cur Loss: 0.24612299, Cur Avg Loss: 0.19790906, Log Avg loss: 0.25942880, Global Avg Loss: 0.79156947, Time: 0.0208 Steps: 74590, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000085, Sample Num: 1360, Cur Loss: 0.23928304, Cur Avg Loss: 0.21298747, Log Avg loss: 0.32607561, Global Avg Loss: 0.79150707, Time: 0.0209 Steps: 74600, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000095, Sample Num: 1520, Cur Loss: 0.36549923, Cur Avg Loss: 0.21495477, Log Avg loss: 0.23167683, Global Avg Loss: 0.79143204, Time: 0.0209 Steps: 74610, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000105, Sample Num: 1680, Cur Loss: 0.22873092, Cur Avg Loss: 0.21638057, Log Avg loss: 0.22992563, Global Avg Loss: 0.79135679, Time: 0.0209 Steps: 74620, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000115, Sample Num: 1840, Cur Loss: 0.17148668, Cur Avg Loss: 0.21331039, Log Avg loss: 0.18107346, Global Avg Loss: 0.79127502, Time: 0.0209 Steps: 74630, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000125, Sample Num: 2000, Cur Loss: 0.11813837, Cur Avg Loss: 0.21317691, Log Avg loss: 0.21164197, Global Avg Loss: 0.79119736, Time: 0.0209 Steps: 74640, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000135, Sample Num: 2160, Cur Loss: 0.07019893, Cur Avg Loss: 0.21110924, Log Avg loss: 0.18526333, Global Avg Loss: 0.79111619, Time: 0.0209 Steps: 74650, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000145, Sample Num: 2320, Cur Loss: 0.06550467, Cur Avg Loss: 0.20965919, Log Avg loss: 0.19008344, Global Avg Loss: 0.79103569, Time: 0.0209 Steps: 74660, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000155, Sample Num: 2480, Cur Loss: 0.06406568, Cur Avg Loss: 0.20585798, Log Avg loss: 0.15074053, Global Avg Loss: 0.79094994, Time: 0.0209 Steps: 74670, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000165, Sample Num: 2640, Cur Loss: 0.06475849, Cur Avg Loss: 0.20248305, Log Avg loss: 0.15017154, Global Avg Loss: 0.79086413, Time: 0.0209 Steps: 74680, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000175, Sample Num: 2800, Cur Loss: 0.14700051, Cur Avg Loss: 0.19728475, Log Avg loss: 0.11151281, Global Avg Loss: 0.79077318, Time: 0.0209 Steps: 74690, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000185, Sample Num: 2960, Cur Loss: 0.11290367, Cur Avg Loss: 0.19752538, Log Avg loss: 0.20173641, Global Avg Loss: 0.79069432, Time: 0.0209 Steps: 74700, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000195, Sample Num: 3120, Cur Loss: 0.15710652, Cur Avg Loss: 0.19447185, Log Avg loss: 0.13798166, Global Avg Loss: 0.79060696, Time: 0.0209 Steps: 74710, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000205, Sample Num: 3280, Cur Loss: 0.09588383, Cur Avg Loss: 0.19117744, Log Avg loss: 0.12693636, Global Avg Loss: 0.79051814, Time: 0.0209 Steps: 74720, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000215, Sample Num: 3440, Cur Loss: 0.15840052, Cur Avg Loss: 0.18917034, Log Avg loss: 0.14802481, Global Avg Loss: 0.79043216, Time: 0.0209 Steps: 74730, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000225, Sample Num: 3600, Cur Loss: 0.11091600, Cur Avg Loss: 0.18957502, Log Avg loss: 0.19827571, Global Avg Loss: 0.79035293, Time: 0.0209 Steps: 74740, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000235, Sample Num: 3760, Cur Loss: 0.06300429, Cur Avg Loss: 0.19084991, Log Avg loss: 0.21953490, Global Avg Loss: 0.79027657, Time: 0.0209 Steps: 74750, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000245, Sample Num: 3920, Cur Loss: 0.37110144, Cur Avg Loss: 0.18881490, Log Avg loss: 0.14099209, Global Avg Loss: 0.79018972, Time: 0.0208 Steps: 74760, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000255, Sample Num: 4080, Cur Loss: 0.24536033, Cur Avg Loss: 0.18827076, Log Avg loss: 0.17493932, Global Avg Loss: 0.79010743, Time: 0.0208 Steps: 74770, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000265, Sample Num: 4240, Cur Loss: 0.21425058, Cur Avg Loss: 0.18749133, Log Avg loss: 0.16761595, Global Avg Loss: 0.79002419, Time: 0.0209 Steps: 74780, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000275, Sample Num: 4400, Cur Loss: 0.11499752, Cur Avg Loss: 0.18639810, Log Avg loss: 0.15742754, Global Avg Loss: 0.78993961, Time: 0.0209 Steps: 74790, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000285, Sample Num: 4560, Cur Loss: 0.04105901, Cur Avg Loss: 0.18444719, Log Avg loss: 0.13079705, Global Avg Loss: 0.78985149, Time: 0.0208 Steps: 74800, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000295, Sample Num: 4720, Cur Loss: 0.16718231, Cur Avg Loss: 0.18374824, Log Avg loss: 0.16382812, Global Avg Loss: 0.78976780, Time: 0.0208 Steps: 74810, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000305, Sample Num: 4880, Cur Loss: 0.09826379, Cur Avg Loss: 0.18198111, Log Avg loss: 0.12985094, Global Avg Loss: 0.78967960, Time: 0.0208 Steps: 74820, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000315, Sample Num: 5040, Cur Loss: 0.38867500, Cur Avg Loss: 0.18069823, Log Avg loss: 0.14157043, Global Avg Loss: 0.78959299, Time: 0.0209 Steps: 74830, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000325, Sample Num: 5200, Cur Loss: 0.10452911, Cur Avg Loss: 0.17950061, Log Avg loss: 0.14177532, Global Avg Loss: 0.78950643, Time: 0.0209 Steps: 74840, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000335, Sample Num: 5360, Cur Loss: 0.06217416, Cur Avg Loss: 0.17981481, Log Avg loss: 0.19002658, Global Avg Loss: 0.78942634, Time: 0.0208 Steps: 74850, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000345, Sample Num: 5520, Cur Loss: 0.08021780, Cur Avg Loss: 0.17980122, Log Avg loss: 0.17934592, Global Avg Loss: 0.78934485, Time: 0.0208 Steps: 74860, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000355, Sample Num: 5680, Cur Loss: 0.09198684, Cur Avg Loss: 0.17872870, Log Avg loss: 0.14172669, Global Avg Loss: 0.78925835, Time: 0.0209 Steps: 74870, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000365, Sample Num: 5840, Cur Loss: 0.14555247, Cur Avg Loss: 0.17767889, Log Avg loss: 0.14041064, Global Avg Loss: 0.78917170, Time: 0.0210 Steps: 74880, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000375, Sample Num: 6000, Cur Loss: 0.15349291, Cur Avg Loss: 0.17747138, Log Avg loss: 0.16989738, Global Avg Loss: 0.78908900, Time: 0.0208 Steps: 74890, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000385, Sample Num: 6160, Cur Loss: 0.10721219, Cur Avg Loss: 0.17728463, Log Avg loss: 0.17028121, Global Avg Loss: 0.78900639, Time: 0.0208 Steps: 74900, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000395, Sample Num: 6320, Cur Loss: 0.22460943, Cur Avg Loss: 0.17907073, Log Avg loss: 0.24783585, Global Avg Loss: 0.78893414, Time: 0.0209 Steps: 74910, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000405, Sample Num: 6480, Cur Loss: 0.12510851, Cur Avg Loss: 0.17812600, Log Avg loss: 0.14080892, Global Avg Loss: 0.78884763, Time: 0.0209 Steps: 74920, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000415, Sample Num: 6640, Cur Loss: 0.19125004, Cur Avg Loss: 0.17662113, Log Avg loss: 0.11567391, Global Avg Loss: 0.78875779, Time: 0.0208 Steps: 74930, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000425, Sample Num: 6800, Cur Loss: 0.17555402, Cur Avg Loss: 0.17613019, Log Avg loss: 0.15575617, Global Avg Loss: 0.78867333, Time: 0.0209 Steps: 74940, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000435, Sample Num: 6960, Cur Loss: 0.19672862, Cur Avg Loss: 0.17562377, Log Avg loss: 0.15410094, Global Avg Loss: 0.78858866, Time: 0.0209 Steps: 74950, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000445, Sample Num: 7120, Cur Loss: 0.16678260, Cur Avg Loss: 0.17505037, Log Avg loss: 0.15010754, Global Avg Loss: 0.78850348, Time: 0.0208 Steps: 74960, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000455, Sample Num: 7280, Cur Loss: 0.20649564, Cur Avg Loss: 0.17529925, Log Avg loss: 0.18637424, Global Avg Loss: 0.78842317, Time: 0.0209 Steps: 74970, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000465, Sample Num: 7440, Cur Loss: 0.03865931, Cur Avg Loss: 0.17493474, Log Avg loss: 0.15834973, Global Avg Loss: 0.78833914, Time: 0.0209 Steps: 74980, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000475, Sample Num: 7600, Cur Loss: 0.09748977, Cur Avg Loss: 0.17627395, Log Avg loss: 0.23854722, Global Avg Loss: 0.78826582, Time: 0.0209 Steps: 74990, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000485, Sample Num: 7760, Cur Loss: 0.15346184, Cur Avg Loss: 0.17628200, Log Avg loss: 0.17666419, Global Avg Loss: 0.78818427, Time: 0.0209 Steps: 75000, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000495, Sample Num: 7920, Cur Loss: 0.22163863, Cur Avg Loss: 0.17768282, Log Avg loss: 0.24562270, Global Avg Loss: 0.78811194, Time: 0.0208 Steps: 75010, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000505, Sample Num: 8080, Cur Loss: 0.06219715, Cur Avg Loss: 0.17704420, Log Avg loss: 0.14543275, Global Avg Loss: 0.78802627, Time: 0.0209 Steps: 75020, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000515, Sample Num: 8240, Cur Loss: 0.10105848, Cur Avg Loss: 0.17737146, Log Avg loss: 0.19389789, Global Avg Loss: 0.78794709, Time: 0.0245 Steps: 75030, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000525, Sample Num: 8400, Cur Loss: 0.18917996, Cur Avg Loss: 0.17663727, Log Avg loss: 0.13882676, Global Avg Loss: 0.78786059, Time: 0.0209 Steps: 75040, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000535, Sample Num: 8560, Cur Loss: 0.25418949, Cur Avg Loss: 0.17701458, Log Avg loss: 0.19682315, Global Avg Loss: 0.78778183, Time: 0.0209 Steps: 75050, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000545, Sample Num: 8720, Cur Loss: 0.17567819, Cur Avg Loss: 0.17722867, Log Avg loss: 0.18868273, Global Avg Loss: 0.78770202, Time: 0.0208 Steps: 75060, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000555, Sample Num: 8880, Cur Loss: 0.10167736, Cur Avg Loss: 0.17714690, Log Avg loss: 0.17268994, Global Avg Loss: 0.78762009, Time: 0.0208 Steps: 75070, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000565, Sample Num: 9040, Cur Loss: 0.21741718, Cur Avg Loss: 0.17724588, Log Avg loss: 0.18273956, Global Avg Loss: 0.78753953, Time: 0.0209 Steps: 75080, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000575, Sample Num: 9200, Cur Loss: 0.10592514, Cur Avg Loss: 0.17655658, Log Avg loss: 0.13761092, Global Avg Loss: 0.78745297, Time: 0.0209 Steps: 75090, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000585, Sample Num: 9360, Cur Loss: 0.04590014, Cur Avg Loss: 0.17680673, Log Avg loss: 0.19119045, Global Avg Loss: 0.78737358, Time: 0.0209 Steps: 75100, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000595, Sample Num: 9520, Cur Loss: 0.13073206, Cur Avg Loss: 0.17729427, Log Avg loss: 0.20581518, Global Avg Loss: 0.78729615, Time: 0.0209 Steps: 75110, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000605, Sample Num: 9680, Cur Loss: 0.08544536, Cur Avg Loss: 0.17733970, Log Avg loss: 0.18004301, Global Avg Loss: 0.78721531, Time: 0.0209 Steps: 75120, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000615, Sample Num: 9840, Cur Loss: 0.13477238, Cur Avg Loss: 0.17724665, Log Avg loss: 0.17161696, Global Avg Loss: 0.78713337, Time: 0.0209 Steps: 75130, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000625, Sample Num: 10000, Cur Loss: 0.10580514, Cur Avg Loss: 0.17770565, Log Avg loss: 0.20593421, Global Avg Loss: 0.78705603, Time: 0.0209 Steps: 75140, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000635, Sample Num: 10160, Cur Loss: 0.12354106, Cur Avg Loss: 0.17712978, Log Avg loss: 0.14113780, Global Avg Loss: 0.78697008, Time: 0.0209 Steps: 75150, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000645, Sample Num: 10320, Cur Loss: 0.18907098, Cur Avg Loss: 0.17753512, Log Avg loss: 0.20327409, Global Avg Loss: 0.78689241, Time: 0.0209 Steps: 75160, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000655, Sample Num: 10480, Cur Loss: 0.26489133, Cur Avg Loss: 0.17757478, Log Avg loss: 0.18013326, Global Avg Loss: 0.78681170, Time: 0.0209 Steps: 75170, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000665, Sample Num: 10640, Cur Loss: 0.12965254, Cur Avg Loss: 0.17734403, Log Avg loss: 0.16222951, Global Avg Loss: 0.78672862, Time: 0.0209 Steps: 75180, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000675, Sample Num: 10800, Cur Loss: 0.32854480, Cur Avg Loss: 0.17767087, Log Avg loss: 0.19940619, Global Avg Loss: 0.78665051, Time: 0.0209 Steps: 75190, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000685, Sample Num: 10960, Cur Loss: 0.35915110, Cur Avg Loss: 0.17888969, Log Avg loss: 0.26116017, Global Avg Loss: 0.78658063, Time: 0.0209 Steps: 75200, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000695, Sample Num: 11120, Cur Loss: 0.12672094, Cur Avg Loss: 0.17819297, Log Avg loss: 0.13046767, Global Avg Loss: 0.78649339, Time: 0.0209 Steps: 75210, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000705, Sample Num: 11280, Cur Loss: 0.02946879, Cur Avg Loss: 0.17811669, Log Avg loss: 0.17281527, Global Avg Loss: 0.78641181, Time: 0.0209 Steps: 75220, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000715, Sample Num: 11440, Cur Loss: 0.15095420, Cur Avg Loss: 0.17822373, Log Avg loss: 0.18576978, Global Avg Loss: 0.78633196, Time: 0.0209 Steps: 75230, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000725, Sample Num: 11600, Cur Loss: 0.31375751, Cur Avg Loss: 0.17838113, Log Avg loss: 0.18963504, Global Avg Loss: 0.78625266, Time: 0.0209 Steps: 75240, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000735, Sample Num: 11760, Cur Loss: 0.07828584, Cur Avg Loss: 0.17735320, Log Avg loss: 0.10282831, Global Avg Loss: 0.78616184, Time: 0.0209 Steps: 75250, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000745, Sample Num: 11920, Cur Loss: 0.05470936, Cur Avg Loss: 0.17764145, Log Avg loss: 0.19882825, Global Avg Loss: 0.78608380, Time: 0.0209 Steps: 75260, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000755, Sample Num: 12080, Cur Loss: 0.05521196, Cur Avg Loss: 0.17733646, Log Avg loss: 0.15461427, Global Avg Loss: 0.78599990, Time: 0.0209 Steps: 75270, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000765, Sample Num: 12240, Cur Loss: 0.08191024, Cur Avg Loss: 0.17676487, Log Avg loss: 0.13360972, Global Avg Loss: 0.78591324, Time: 0.0209 Steps: 75280, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000775, Sample Num: 12400, Cur Loss: 0.11552492, Cur Avg Loss: 0.17718291, Log Avg loss: 0.20916314, Global Avg Loss: 0.78583664, Time: 0.0209 Steps: 75290, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000785, Sample Num: 12560, Cur Loss: 0.33671075, Cur Avg Loss: 0.17681082, Log Avg loss: 0.14797362, Global Avg Loss: 0.78575193, Time: 0.0209 Steps: 75300, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000795, Sample Num: 12720, Cur Loss: 0.12532961, Cur Avg Loss: 0.17663502, Log Avg loss: 0.16283534, Global Avg Loss: 0.78566922, Time: 0.0209 Steps: 75310, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000805, Sample Num: 12880, Cur Loss: 0.09607315, Cur Avg Loss: 0.17682598, Log Avg loss: 0.19200662, Global Avg Loss: 0.78559040, Time: 0.0210 Steps: 75320, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000815, Sample Num: 13040, Cur Loss: 0.29989138, Cur Avg Loss: 0.17730631, Log Avg loss: 0.21597290, Global Avg Loss: 0.78551478, Time: 0.0209 Steps: 75330, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000825, Sample Num: 13200, Cur Loss: 0.24446803, Cur Avg Loss: 0.17721499, Log Avg loss: 0.16977256, Global Avg Loss: 0.78543305, Time: 0.0209 Steps: 75340, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000835, Sample Num: 13360, Cur Loss: 0.13639602, Cur Avg Loss: 0.17837243, Log Avg loss: 0.27386134, Global Avg Loss: 0.78536516, Time: 0.0209 Steps: 75350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000845, Sample Num: 13520, Cur Loss: 0.10524098, Cur Avg Loss: 0.17827060, Log Avg loss: 0.16976805, Global Avg Loss: 0.78528347, Time: 0.0209 Steps: 75360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000855, Sample Num: 13680, Cur Loss: 0.13613248, Cur Avg Loss: 0.17796398, Log Avg loss: 0.15205450, Global Avg Loss: 0.78519946, Time: 0.0209 Steps: 75370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000865, Sample Num: 13840, Cur Loss: 0.20486581, Cur Avg Loss: 0.17781875, Log Avg loss: 0.16540103, Global Avg Loss: 0.78511723, Time: 0.0209 Steps: 75380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000875, Sample Num: 14000, Cur Loss: 0.19187829, Cur Avg Loss: 0.17725669, Log Avg loss: 0.12863876, Global Avg Loss: 0.78503015, Time: 0.0209 Steps: 75390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000885, Sample Num: 14160, Cur Loss: 0.33400470, Cur Avg Loss: 0.17763177, Log Avg loss: 0.21045117, Global Avg Loss: 0.78495395, Time: 0.0209 Steps: 75400, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000895, Sample Num: 14320, Cur Loss: 0.12206876, Cur Avg Loss: 0.17861548, Log Avg loss: 0.26567406, Global Avg Loss: 0.78488509, Time: 0.0209 Steps: 75410, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000905, Sample Num: 14480, Cur Loss: 0.16548043, Cur Avg Loss: 0.17849559, Log Avg loss: 0.16776512, Global Avg Loss: 0.78480327, Time: 0.0209 Steps: 75420, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000915, Sample Num: 14640, Cur Loss: 0.13127959, Cur Avg Loss: 0.17827115, Log Avg loss: 0.15795945, Global Avg Loss: 0.78472016, Time: 0.0209 Steps: 75430, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000925, Sample Num: 14800, Cur Loss: 0.22009459, Cur Avg Loss: 0.17831479, Log Avg loss: 0.18230778, Global Avg Loss: 0.78464031, Time: 0.0209 Steps: 75440, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000935, Sample Num: 14960, Cur Loss: 0.27263016, Cur Avg Loss: 0.17873568, Log Avg loss: 0.21766875, Global Avg Loss: 0.78456516, Time: 0.0209 Steps: 75450, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000945, Sample Num: 15120, Cur Loss: 0.22459695, Cur Avg Loss: 0.17839221, Log Avg loss: 0.14627765, Global Avg Loss: 0.78448058, Time: 0.0209 Steps: 75460, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000955, Sample Num: 15280, Cur Loss: 0.25233912, Cur Avg Loss: 0.17821684, Log Avg loss: 0.16164386, Global Avg Loss: 0.78439805, Time: 0.0209 Steps: 75470, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000965, Sample Num: 15440, Cur Loss: 0.14651579, Cur Avg Loss: 0.17832838, Log Avg loss: 0.18898056, Global Avg Loss: 0.78431917, Time: 0.0209 Steps: 75480, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000975, Sample Num: 15600, Cur Loss: 0.50986958, Cur Avg Loss: 0.17845823, Log Avg loss: 0.19098916, Global Avg Loss: 0.78424057, Time: 0.0209 Steps: 75490, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000985, Sample Num: 15760, Cur Loss: 0.12507063, Cur Avg Loss: 0.17844853, Log Avg loss: 0.17750209, Global Avg Loss: 0.78416021, Time: 0.0209 Steps: 75500, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000995, Sample Num: 15920, Cur Loss: 0.24316265, Cur Avg Loss: 0.17848103, Log Avg loss: 0.18168255, Global Avg Loss: 0.78408042, Time: 0.0209 Steps: 75510, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001005, Sample Num: 16080, Cur Loss: 0.42015249, Cur Avg Loss: 0.17876107, Log Avg loss: 0.20662522, Global Avg Loss: 0.78400395, Time: 0.0209 Steps: 75520, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001015, Sample Num: 16240, Cur Loss: 0.14034808, Cur Avg Loss: 0.17863377, Log Avg loss: 0.16584001, Global Avg Loss: 0.78392211, Time: 0.0209 Steps: 75530, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001025, Sample Num: 16400, Cur Loss: 0.40837276, Cur Avg Loss: 0.17872209, Log Avg loss: 0.18768677, Global Avg Loss: 0.78384318, Time: 0.0245 Steps: 75540, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001035, Sample Num: 16560, Cur Loss: 0.16042154, Cur Avg Loss: 0.17882790, Log Avg loss: 0.18967287, Global Avg Loss: 0.78376453, Time: 0.0209 Steps: 75550, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001045, Sample Num: 16720, Cur Loss: 0.15685621, Cur Avg Loss: 0.17937579, Log Avg loss: 0.23608317, Global Avg Loss: 0.78369205, Time: 0.0208 Steps: 75560, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001055, Sample Num: 16880, Cur Loss: 0.05572172, Cur Avg Loss: 0.17866228, Log Avg loss: 0.10409983, Global Avg Loss: 0.78360212, Time: 0.0208 Steps: 75570, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001065, Sample Num: 17040, Cur Loss: 0.17874680, Cur Avg Loss: 0.17797805, Log Avg loss: 0.10579192, Global Avg Loss: 0.78351244, Time: 0.0208 Steps: 75580, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001075, Sample Num: 17200, Cur Loss: 0.03758254, Cur Avg Loss: 0.17791566, Log Avg loss: 0.17127110, Global Avg Loss: 0.78343145, Time: 0.0208 Steps: 75590, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001085, Sample Num: 17360, Cur Loss: 0.18673585, Cur Avg Loss: 0.17868470, Log Avg loss: 0.26135671, Global Avg Loss: 0.78336239, Time: 0.0209 Steps: 75600, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001095, Sample Num: 17520, Cur Loss: 0.15753256, Cur Avg Loss: 0.17882876, Log Avg loss: 0.19445902, Global Avg Loss: 0.78328450, Time: 0.0209 Steps: 75610, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001105, Sample Num: 17680, Cur Loss: 0.30184442, Cur Avg Loss: 0.17936969, Log Avg loss: 0.23860195, Global Avg Loss: 0.78321247, Time: 0.0208 Steps: 75620, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001115, Sample Num: 17840, Cur Loss: 0.17441624, Cur Avg Loss: 0.17945742, Log Avg loss: 0.18915081, Global Avg Loss: 0.78313393, Time: 0.0208 Steps: 75630, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001125, Sample Num: 18000, Cur Loss: 0.33291897, Cur Avg Loss: 0.17969618, Log Avg loss: 0.20631840, Global Avg Loss: 0.78305767, Time: 0.0209 Steps: 75640, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001135, Sample Num: 18160, Cur Loss: 0.07047610, Cur Avg Loss: 0.17982303, Log Avg loss: 0.19409365, Global Avg Loss: 0.78297981, Time: 0.0208 Steps: 75650, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001145, Sample Num: 18320, Cur Loss: 0.23255967, Cur Avg Loss: 0.18085863, Log Avg loss: 0.29839883, Global Avg Loss: 0.78291577, Time: 0.0209 Steps: 75660, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001155, Sample Num: 18480, Cur Loss: 0.11603394, Cur Avg Loss: 0.18059452, Log Avg loss: 0.15035454, Global Avg Loss: 0.78283217, Time: 0.0208 Steps: 75670, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001165, Sample Num: 18640, Cur Loss: 0.03866148, Cur Avg Loss: 0.18055494, Log Avg loss: 0.17598312, Global Avg Loss: 0.78275199, Time: 0.0208 Steps: 75680, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001175, Sample Num: 18800, Cur Loss: 0.20242190, Cur Avg Loss: 0.18036196, Log Avg loss: 0.15788003, Global Avg Loss: 0.78266943, Time: 0.0209 Steps: 75690, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001185, Sample Num: 18960, Cur Loss: 0.17318808, Cur Avg Loss: 0.18025796, Log Avg loss: 0.16803771, Global Avg Loss: 0.78258824, Time: 0.0208 Steps: 75700, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001195, Sample Num: 19120, Cur Loss: 0.12082713, Cur Avg Loss: 0.17989393, Log Avg loss: 0.13675715, Global Avg Loss: 0.78250293, Time: 0.0208 Steps: 75710, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001205, Sample Num: 19280, Cur Loss: 0.16783909, Cur Avg Loss: 0.17983356, Log Avg loss: 0.17261839, Global Avg Loss: 0.78242239, Time: 0.0208 Steps: 75720, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001215, Sample Num: 19440, Cur Loss: 0.05936588, Cur Avg Loss: 0.17952599, Log Avg loss: 0.14246415, Global Avg Loss: 0.78233788, Time: 0.0210 Steps: 75730, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001225, Sample Num: 19600, Cur Loss: 0.12128776, Cur Avg Loss: 0.17979976, Log Avg loss: 0.21306231, Global Avg Loss: 0.78226272, Time: 0.0209 Steps: 75740, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001235, Sample Num: 19760, Cur Loss: 0.26276034, Cur Avg Loss: 0.18005880, Log Avg loss: 0.21179171, Global Avg Loss: 0.78218741, Time: 0.0208 Steps: 75750, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001245, Sample Num: 19920, Cur Loss: 0.25031388, Cur Avg Loss: 0.18027208, Log Avg loss: 0.20661222, Global Avg Loss: 0.78211144, Time: 0.0208 Steps: 75760, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001255, Sample Num: 20080, Cur Loss: 0.05410555, Cur Avg Loss: 0.18044323, Log Avg loss: 0.20175115, Global Avg Loss: 0.78203484, Time: 0.0208 Steps: 75770, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001265, Sample Num: 20240, Cur Loss: 0.22271179, Cur Avg Loss: 0.18038688, Log Avg loss: 0.17331505, Global Avg Loss: 0.78195452, Time: 0.0208 Steps: 75780, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001275, Sample Num: 20400, Cur Loss: 0.37274003, Cur Avg Loss: 0.18084155, Log Avg loss: 0.23835685, Global Avg Loss: 0.78188279, Time: 0.0209 Steps: 75790, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001285, Sample Num: 20560, Cur Loss: 0.34240150, Cur Avg Loss: 0.18094172, Log Avg loss: 0.19371437, Global Avg Loss: 0.78180520, Time: 0.0210 Steps: 75800, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001295, Sample Num: 20720, Cur Loss: 0.12855153, Cur Avg Loss: 0.18073963, Log Avg loss: 0.15477125, Global Avg Loss: 0.78172248, Time: 0.0209 Steps: 75810, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001305, Sample Num: 20880, Cur Loss: 0.44042018, Cur Avg Loss: 0.18127262, Log Avg loss: 0.25029376, Global Avg Loss: 0.78165239, Time: 0.0209 Steps: 75820, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001315, Sample Num: 21040, Cur Loss: 0.31931856, Cur Avg Loss: 0.18187399, Log Avg loss: 0.26035386, Global Avg Loss: 0.78158365, Time: 0.0209 Steps: 75830, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001325, Sample Num: 21200, Cur Loss: 0.13850664, Cur Avg Loss: 0.18175054, Log Avg loss: 0.16551686, Global Avg Loss: 0.78150242, Time: 0.0209 Steps: 75840, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001335, Sample Num: 21360, Cur Loss: 0.50986624, Cur Avg Loss: 0.18180309, Log Avg loss: 0.18876509, Global Avg Loss: 0.78142427, Time: 0.0209 Steps: 75850, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001345, Sample Num: 21520, Cur Loss: 0.07449402, Cur Avg Loss: 0.18136492, Log Avg loss: 0.12286918, Global Avg Loss: 0.78133746, Time: 0.0209 Steps: 75860, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001355, Sample Num: 21680, Cur Loss: 0.20094107, Cur Avg Loss: 0.18163997, Log Avg loss: 0.21863456, Global Avg Loss: 0.78126329, Time: 0.0209 Steps: 75870, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001365, Sample Num: 21840, Cur Loss: 0.04954258, Cur Avg Loss: 0.18123333, Log Avg loss: 0.12613378, Global Avg Loss: 0.78117695, Time: 0.0209 Steps: 75880, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001375, Sample Num: 22000, Cur Loss: 0.05090938, Cur Avg Loss: 0.18127888, Log Avg loss: 0.18749573, Global Avg Loss: 0.78109872, Time: 0.0210 Steps: 75890, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001385, Sample Num: 22160, Cur Loss: 0.07547504, Cur Avg Loss: 0.18170994, Log Avg loss: 0.24098183, Global Avg Loss: 0.78102756, Time: 0.0209 Steps: 75900, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001395, Sample Num: 22320, Cur Loss: 0.22749873, Cur Avg Loss: 0.18142356, Log Avg loss: 0.14175931, Global Avg Loss: 0.78094335, Time: 0.0209 Steps: 75910, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001405, Sample Num: 22480, Cur Loss: 0.26045647, Cur Avg Loss: 0.18188613, Log Avg loss: 0.24641420, Global Avg Loss: 0.78087294, Time: 0.0209 Steps: 75920, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001415, Sample Num: 22640, Cur Loss: 0.17416693, Cur Avg Loss: 0.18188668, Log Avg loss: 0.18196369, Global Avg Loss: 0.78079407, Time: 0.0209 Steps: 75930, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001425, Sample Num: 22800, Cur Loss: 0.07778499, Cur Avg Loss: 0.18193931, Log Avg loss: 0.18938743, Global Avg Loss: 0.78071619, Time: 0.0209 Steps: 75940, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001435, Sample Num: 22960, Cur Loss: 0.22181921, Cur Avg Loss: 0.18229110, Log Avg loss: 0.23242073, Global Avg Loss: 0.78064400, Time: 0.0209 Steps: 75950, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001445, Sample Num: 23120, Cur Loss: 0.04537310, Cur Avg Loss: 0.18240072, Log Avg loss: 0.19813147, Global Avg Loss: 0.78056731, Time: 0.0209 Steps: 75960, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001455, Sample Num: 23280, Cur Loss: 0.16389176, Cur Avg Loss: 0.18226078, Log Avg loss: 0.16203892, Global Avg Loss: 0.78048589, Time: 0.0209 Steps: 75970, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001465, Sample Num: 23440, Cur Loss: 0.34554589, Cur Avg Loss: 0.18225884, Log Avg loss: 0.18197745, Global Avg Loss: 0.78040712, Time: 0.0209 Steps: 75980, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001475, Sample Num: 23600, Cur Loss: 0.05745235, Cur Avg Loss: 0.18214046, Log Avg loss: 0.16479691, Global Avg Loss: 0.78032611, Time: 0.0209 Steps: 75990, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001485, Sample Num: 23760, Cur Loss: 0.27156451, Cur Avg Loss: 0.18231288, Log Avg loss: 0.20774534, Global Avg Loss: 0.78025077, Time: 0.0210 Steps: 76000, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001495, Sample Num: 23920, Cur Loss: 0.23032954, Cur Avg Loss: 0.18213722, Log Avg loss: 0.15605157, Global Avg Loss: 0.78016865, Time: 0.0209 Steps: 76010, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001505, Sample Num: 24080, Cur Loss: 0.12885587, Cur Avg Loss: 0.18192973, Log Avg loss: 0.15090975, Global Avg Loss: 0.78008587, Time: 0.0209 Steps: 76020, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001515, Sample Num: 24240, Cur Loss: 0.11512426, Cur Avg Loss: 0.18175227, Log Avg loss: 0.15504506, Global Avg Loss: 0.78000366, Time: 0.0209 Steps: 76030, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001525, Sample Num: 24400, Cur Loss: 0.09157266, Cur Avg Loss: 0.18159353, Log Avg loss: 0.15754377, Global Avg Loss: 0.77992180, Time: 0.0209 Steps: 76040, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001535, Sample Num: 24560, Cur Loss: 0.23061582, Cur Avg Loss: 0.18141435, Log Avg loss: 0.15408988, Global Avg Loss: 0.77983951, Time: 0.0210 Steps: 76050, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001545, Sample Num: 24720, Cur Loss: 0.14709589, Cur Avg Loss: 0.18122142, Log Avg loss: 0.15160723, Global Avg Loss: 0.77975691, Time: 0.0209 Steps: 76060, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001555, Sample Num: 24880, Cur Loss: 0.14771940, Cur Avg Loss: 0.18114166, Log Avg loss: 0.16881862, Global Avg Loss: 0.77967660, Time: 0.0209 Steps: 76070, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001565, Sample Num: 25040, Cur Loss: 0.07777549, Cur Avg Loss: 0.18092918, Log Avg loss: 0.14788867, Global Avg Loss: 0.77959356, Time: 0.0210 Steps: 76080, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001575, Sample Num: 25200, Cur Loss: 0.08824725, Cur Avg Loss: 0.18061369, Log Avg loss: 0.13123899, Global Avg Loss: 0.77950835, Time: 0.0209 Steps: 76090, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001585, Sample Num: 25360, Cur Loss: 0.10097148, Cur Avg Loss: 0.18022371, Log Avg loss: 0.11880148, Global Avg Loss: 0.77942153, Time: 0.0209 Steps: 76100, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001595, Sample Num: 25520, Cur Loss: 0.13914099, Cur Avg Loss: 0.17996048, Log Avg loss: 0.13823948, Global Avg Loss: 0.77933728, Time: 0.0209 Steps: 76110, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001605, Sample Num: 25680, Cur Loss: 0.12019847, Cur Avg Loss: 0.18036133, Log Avg loss: 0.24429638, Global Avg Loss: 0.77926700, Time: 0.0209 Steps: 76120, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001615, Sample Num: 25840, Cur Loss: 0.11523204, Cur Avg Loss: 0.18006046, Log Avg loss: 0.13177044, Global Avg Loss: 0.77918194, Time: 0.0209 Steps: 76130, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001625, Sample Num: 26000, Cur Loss: 0.37555927, Cur Avg Loss: 0.18020649, Log Avg loss: 0.20379026, Global Avg Loss: 0.77910637, Time: 0.0209 Steps: 76140, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001635, Sample Num: 26160, Cur Loss: 0.11580293, Cur Avg Loss: 0.18005196, Log Avg loss: 0.15494049, Global Avg Loss: 0.77902441, Time: 0.0209 Steps: 76150, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001645, Sample Num: 26320, Cur Loss: 0.06185049, Cur Avg Loss: 0.18000854, Log Avg loss: 0.17291063, Global Avg Loss: 0.77894482, Time: 0.0209 Steps: 76160, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001655, Sample Num: 26480, Cur Loss: 0.27521914, Cur Avg Loss: 0.17999310, Log Avg loss: 0.17745244, Global Avg Loss: 0.77886586, Time: 0.0209 Steps: 76170, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001665, Sample Num: 26640, Cur Loss: 0.16855413, Cur Avg Loss: 0.18009585, Log Avg loss: 0.19710157, Global Avg Loss: 0.77878949, Time: 0.0209 Steps: 76180, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001675, Sample Num: 26800, Cur Loss: 0.09782477, Cur Avg Loss: 0.18049089, Log Avg loss: 0.24626530, Global Avg Loss: 0.77871960, Time: 0.0209 Steps: 76190, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001685, Sample Num: 26960, Cur Loss: 0.27053607, Cur Avg Loss: 0.18052108, Log Avg loss: 0.18557637, Global Avg Loss: 0.77864176, Time: 0.0209 Steps: 76200, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001695, Sample Num: 27120, Cur Loss: 0.31085584, Cur Avg Loss: 0.18079743, Log Avg loss: 0.22736282, Global Avg Loss: 0.77856942, Time: 0.0209 Steps: 76210, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001705, Sample Num: 27280, Cur Loss: 0.22148636, Cur Avg Loss: 0.18093349, Log Avg loss: 0.20399562, Global Avg Loss: 0.77849403, Time: 0.0209 Steps: 76220, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001715, Sample Num: 27440, Cur Loss: 0.22710340, Cur Avg Loss: 0.18127273, Log Avg loss: 0.23911387, Global Avg Loss: 0.77842328, Time: 0.0210 Steps: 76230, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001725, Sample Num: 27600, Cur Loss: 0.15843856, Cur Avg Loss: 0.18128657, Log Avg loss: 0.18366023, Global Avg Loss: 0.77834527, Time: 0.0209 Steps: 76240, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001735, Sample Num: 27760, Cur Loss: 0.06636173, Cur Avg Loss: 0.18107414, Log Avg loss: 0.14442887, Global Avg Loss: 0.77826213, Time: 0.0210 Steps: 76250, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001745, Sample Num: 27920, Cur Loss: 0.16488084, Cur Avg Loss: 0.18086346, Log Avg loss: 0.14431059, Global Avg Loss: 0.77817900, Time: 0.0209 Steps: 76260, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001755, Sample Num: 28080, Cur Loss: 0.18662761, Cur Avg Loss: 0.18078233, Log Avg loss: 0.16662562, Global Avg Loss: 0.77809882, Time: 0.0209 Steps: 76270, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001765, Sample Num: 28240, Cur Loss: 0.15878491, Cur Avg Loss: 0.18045978, Log Avg loss: 0.12385176, Global Avg Loss: 0.77801305, Time: 0.0209 Steps: 76280, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001775, Sample Num: 28400, Cur Loss: 0.07223094, Cur Avg Loss: 0.18063386, Log Avg loss: 0.21135975, Global Avg Loss: 0.77793877, Time: 0.0210 Steps: 76290, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001785, Sample Num: 28560, Cur Loss: 0.07207026, Cur Avg Loss: 0.18037558, Log Avg loss: 0.13452995, Global Avg Loss: 0.77785444, Time: 0.0209 Steps: 76300, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001795, Sample Num: 28720, Cur Loss: 0.11984376, Cur Avg Loss: 0.18020212, Log Avg loss: 0.14924031, Global Avg Loss: 0.77777207, Time: 0.0246 Steps: 76310, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001805, Sample Num: 28880, Cur Loss: 0.21314001, Cur Avg Loss: 0.18026858, Log Avg loss: 0.19219828, Global Avg Loss: 0.77769534, Time: 0.0209 Steps: 76320, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001815, Sample Num: 29040, Cur Loss: 0.28868726, Cur Avg Loss: 0.18038962, Log Avg loss: 0.20223783, Global Avg Loss: 0.77761995, Time: 0.0209 Steps: 76330, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001825, Sample Num: 29200, Cur Loss: 0.18634391, Cur Avg Loss: 0.18030039, Log Avg loss: 0.16410512, Global Avg Loss: 0.77753959, Time: 0.0209 Steps: 76340, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001835, Sample Num: 29360, Cur Loss: 0.11599719, Cur Avg Loss: 0.18040637, Log Avg loss: 0.19974658, Global Avg Loss: 0.77746391, Time: 0.0209 Steps: 76350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001845, Sample Num: 29520, Cur Loss: 0.36188051, Cur Avg Loss: 0.18096508, Log Avg loss: 0.28348942, Global Avg Loss: 0.77739922, Time: 0.0209 Steps: 76360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001855, Sample Num: 29680, Cur Loss: 0.07445163, Cur Avg Loss: 0.18110711, Log Avg loss: 0.20731110, Global Avg Loss: 0.77732457, Time: 0.0209 Steps: 76370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001865, Sample Num: 29840, Cur Loss: 0.19121416, Cur Avg Loss: 0.18148316, Log Avg loss: 0.25124109, Global Avg Loss: 0.77725569, Time: 0.0209 Steps: 76380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001875, Sample Num: 30000, Cur Loss: 0.08207105, Cur Avg Loss: 0.18141558, Log Avg loss: 0.16881127, Global Avg Loss: 0.77717604, Time: 0.0209 Steps: 76390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001885, Sample Num: 30160, Cur Loss: 0.21956700, Cur Avg Loss: 0.18140000, Log Avg loss: 0.17847942, Global Avg Loss: 0.77709768, Time: 0.0209 Steps: 76400, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001895, Sample Num: 30320, Cur Loss: 0.25333005, Cur Avg Loss: 0.18159014, Log Avg loss: 0.21743065, Global Avg Loss: 0.77702443, Time: 0.0209 Steps: 76410, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001905, Sample Num: 30480, Cur Loss: 0.23674239, Cur Avg Loss: 0.18168642, Log Avg loss: 0.19993118, Global Avg Loss: 0.77694892, Time: 0.0210 Steps: 76420, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001915, Sample Num: 30640, Cur Loss: 0.04765067, Cur Avg Loss: 0.18131483, Log Avg loss: 0.11052763, Global Avg Loss: 0.77686173, Time: 0.0209 Steps: 76430, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001925, Sample Num: 30800, Cur Loss: 0.23954961, Cur Avg Loss: 0.18121413, Log Avg loss: 0.16192980, Global Avg Loss: 0.77678128, Time: 0.0209 Steps: 76440, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001935, Sample Num: 30960, Cur Loss: 0.27155831, Cur Avg Loss: 0.18132145, Log Avg loss: 0.20198038, Global Avg Loss: 0.77670609, Time: 0.0209 Steps: 76450, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001945, Sample Num: 31120, Cur Loss: 0.06120040, Cur Avg Loss: 0.18128967, Log Avg loss: 0.17514105, Global Avg Loss: 0.77662742, Time: 0.0209 Steps: 76460, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001955, Sample Num: 31280, Cur Loss: 0.11062653, Cur Avg Loss: 0.18154867, Log Avg loss: 0.23192430, Global Avg Loss: 0.77655618, Time: 0.0209 Steps: 76470, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001965, Sample Num: 31440, Cur Loss: 0.33608982, Cur Avg Loss: 0.18171673, Log Avg loss: 0.21457147, Global Avg Loss: 0.77648270, Time: 0.0209 Steps: 76480, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001975, Sample Num: 31600, Cur Loss: 0.18744177, Cur Avg Loss: 0.18164729, Log Avg loss: 0.16800298, Global Avg Loss: 0.77640315, Time: 0.0209 Steps: 76490, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001985, Sample Num: 31760, Cur Loss: 0.12254385, Cur Avg Loss: 0.18180840, Log Avg loss: 0.21362767, Global Avg Loss: 0.77632959, Time: 0.0209 Steps: 76500, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001995, Sample Num: 31920, Cur Loss: 0.05480414, Cur Avg Loss: 0.18161064, Log Avg loss: 0.14235480, Global Avg Loss: 0.77624673, Time: 0.0209 Steps: 76510, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002005, Sample Num: 32080, Cur Loss: 0.14669412, Cur Avg Loss: 0.18186967, Log Avg loss: 0.23354712, Global Avg Loss: 0.77617580, Time: 0.0209 Steps: 76520, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002015, Sample Num: 32240, Cur Loss: 0.08149067, Cur Avg Loss: 0.18167370, Log Avg loss: 0.14238151, Global Avg Loss: 0.77609299, Time: 0.0209 Steps: 76530, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002025, Sample Num: 32400, Cur Loss: 0.20452107, Cur Avg Loss: 0.18180154, Log Avg loss: 0.20756017, Global Avg Loss: 0.77601871, Time: 0.0209 Steps: 76540, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002035, Sample Num: 32560, Cur Loss: 0.13523570, Cur Avg Loss: 0.18187469, Log Avg loss: 0.19668737, Global Avg Loss: 0.77594303, Time: 0.0209 Steps: 76550, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002045, Sample Num: 32720, Cur Loss: 0.12421092, Cur Avg Loss: 0.18177014, Log Avg loss: 0.16049432, Global Avg Loss: 0.77586264, Time: 0.0209 Steps: 76560, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002055, Sample Num: 32880, Cur Loss: 0.28662518, Cur Avg Loss: 0.18155615, Log Avg loss: 0.13779492, Global Avg Loss: 0.77577931, Time: 0.0209 Steps: 76570, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002065, Sample Num: 33040, Cur Loss: 0.20703393, Cur Avg Loss: 0.18136442, Log Avg loss: 0.14196502, Global Avg Loss: 0.77569654, Time: 0.0209 Steps: 76580, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002075, Sample Num: 33200, Cur Loss: 0.28409618, Cur Avg Loss: 0.18150484, Log Avg loss: 0.21050060, Global Avg Loss: 0.77562275, Time: 0.0210 Steps: 76590, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002085, Sample Num: 33360, Cur Loss: 0.05646168, Cur Avg Loss: 0.18149014, Log Avg loss: 0.17843960, Global Avg Loss: 0.77554479, Time: 0.0209 Steps: 76600, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002095, Sample Num: 33520, Cur Loss: 0.04574002, Cur Avg Loss: 0.18128297, Log Avg loss: 0.13808955, Global Avg Loss: 0.77546158, Time: 0.0209 Steps: 76610, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002105, Sample Num: 33680, Cur Loss: 0.12100738, Cur Avg Loss: 0.18147465, Log Avg loss: 0.22163204, Global Avg Loss: 0.77538930, Time: 0.0209 Steps: 76620, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002115, Sample Num: 33840, Cur Loss: 0.16021767, Cur Avg Loss: 0.18123452, Log Avg loss: 0.13068686, Global Avg Loss: 0.77530516, Time: 0.0209 Steps: 76630, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002125, Sample Num: 34000, Cur Loss: 0.05589654, Cur Avg Loss: 0.18090268, Log Avg loss: 0.11071681, Global Avg Loss: 0.77521845, Time: 0.0209 Steps: 76640, Updated lr: 0.000028 ***** Running evaluation checkpoint-76644 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-76644 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.675068, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.248578, "eval_total_loss": 174.750607, "eval_mae": 0.341666, "eval_mse": 0.248669, "eval_r2": 0.84193, "eval_sp_statistic": 0.891853, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.918239, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.632918, "test_total_loss": 317.724701, "test_mae": 0.584371, "test_mse": 0.633004, "test_r2": 0.591453, "test_sp_statistic": 0.802757, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.841933, "test_ps_pvalue": 0.0, "lr": 2.826552868658132e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7751863407298866, "train_cur_epoch_loss": 385.05815269984305, "train_cur_epoch_avg_loss": 0.18086338783459044, "train_cur_epoch_time": 44.67506766319275, "train_cur_epoch_avg_time": 0.020984061842739665, "epoch": 36, "step": 76644} ################################################## Training, Epoch: 0037, Batch: 000006, Sample Num: 96, Cur Loss: 0.10711565, Cur Avg Loss: 0.20624874, Log Avg loss: 0.18774606, Global Avg Loss: 0.77514181, Time: 0.0210 Steps: 76650, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000016, Sample Num: 256, Cur Loss: 0.08793615, Cur Avg Loss: 0.19489998, Log Avg loss: 0.18809072, Global Avg Loss: 0.77506523, Time: 0.0208 Steps: 76660, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000026, Sample Num: 416, Cur Loss: 0.16249105, Cur Avg Loss: 0.18509850, Log Avg loss: 0.16941613, Global Avg Loss: 0.77498623, Time: 0.0209 Steps: 76670, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000036, Sample Num: 576, Cur Loss: 0.13977174, Cur Avg Loss: 0.20121660, Log Avg loss: 0.24312367, Global Avg Loss: 0.77491687, Time: 0.0209 Steps: 76680, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000046, Sample Num: 736, Cur Loss: 0.17891425, Cur Avg Loss: 0.20843297, Log Avg loss: 0.23441189, Global Avg Loss: 0.77484639, Time: 0.0209 Steps: 76690, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000056, Sample Num: 896, Cur Loss: 0.14518858, Cur Avg Loss: 0.20732587, Log Avg loss: 0.20223322, Global Avg Loss: 0.77477174, Time: 0.0209 Steps: 76700, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000066, Sample Num: 1056, Cur Loss: 0.07540463, Cur Avg Loss: 0.19879708, Log Avg loss: 0.15103583, Global Avg Loss: 0.77469043, Time: 0.0209 Steps: 76710, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000076, Sample Num: 1216, Cur Loss: 0.14604385, Cur Avg Loss: 0.18554848, Log Avg loss: 0.09810773, Global Avg Loss: 0.77460224, Time: 0.0209 Steps: 76720, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000086, Sample Num: 1376, Cur Loss: 0.15117484, Cur Avg Loss: 0.18378814, Log Avg loss: 0.17040954, Global Avg Loss: 0.77452349, Time: 0.0209 Steps: 76730, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000096, Sample Num: 1536, Cur Loss: 0.11441377, Cur Avg Loss: 0.18363442, Log Avg loss: 0.18231244, Global Avg Loss: 0.77444632, Time: 0.0209 Steps: 76740, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000106, Sample Num: 1696, Cur Loss: 0.12198358, Cur Avg Loss: 0.18029851, Log Avg loss: 0.14827377, Global Avg Loss: 0.77436474, Time: 0.0209 Steps: 76750, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000116, Sample Num: 1856, Cur Loss: 0.14724261, Cur Avg Loss: 0.17912082, Log Avg loss: 0.16663734, Global Avg Loss: 0.77428556, Time: 0.0209 Steps: 76760, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000126, Sample Num: 2016, Cur Loss: 0.25583565, Cur Avg Loss: 0.18275979, Log Avg loss: 0.22497185, Global Avg Loss: 0.77421401, Time: 0.0209 Steps: 76770, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000136, Sample Num: 2176, Cur Loss: 0.24383320, Cur Avg Loss: 0.18191561, Log Avg loss: 0.17127896, Global Avg Loss: 0.77413548, Time: 0.0209 Steps: 76780, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000146, Sample Num: 2336, Cur Loss: 0.56698251, Cur Avg Loss: 0.18636192, Log Avg loss: 0.24683164, Global Avg Loss: 0.77406682, Time: 0.0209 Steps: 76790, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000156, Sample Num: 2496, Cur Loss: 0.16761149, Cur Avg Loss: 0.18302083, Log Avg loss: 0.13424097, Global Avg Loss: 0.77398350, Time: 0.0210 Steps: 76800, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000166, Sample Num: 2656, Cur Loss: 0.16238591, Cur Avg Loss: 0.18233781, Log Avg loss: 0.17168274, Global Avg Loss: 0.77390509, Time: 0.0209 Steps: 76810, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000176, Sample Num: 2816, Cur Loss: 0.21574602, Cur Avg Loss: 0.18079826, Log Avg loss: 0.15524166, Global Avg Loss: 0.77382456, Time: 0.0210 Steps: 76820, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000186, Sample Num: 2976, Cur Loss: 0.15045249, Cur Avg Loss: 0.18135516, Log Avg loss: 0.19115662, Global Avg Loss: 0.77374872, Time: 0.0209 Steps: 76830, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000196, Sample Num: 3136, Cur Loss: 0.10661817, Cur Avg Loss: 0.17971387, Log Avg loss: 0.14918586, Global Avg Loss: 0.77366744, Time: 0.0210 Steps: 76840, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000206, Sample Num: 3296, Cur Loss: 0.13898368, Cur Avg Loss: 0.17945917, Log Avg loss: 0.17446711, Global Avg Loss: 0.77358947, Time: 0.0209 Steps: 76850, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000216, Sample Num: 3456, Cur Loss: 0.14774828, Cur Avg Loss: 0.18064765, Log Avg loss: 0.20513038, Global Avg Loss: 0.77351551, Time: 0.0209 Steps: 76860, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000226, Sample Num: 3616, Cur Loss: 0.40032130, Cur Avg Loss: 0.18068934, Log Avg loss: 0.18158970, Global Avg Loss: 0.77343850, Time: 0.0209 Steps: 76870, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000236, Sample Num: 3776, Cur Loss: 0.13620847, Cur Avg Loss: 0.17846528, Log Avg loss: 0.12820166, Global Avg Loss: 0.77335457, Time: 0.0209 Steps: 76880, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000246, Sample Num: 3936, Cur Loss: 0.08823551, Cur Avg Loss: 0.17782768, Log Avg loss: 0.16278026, Global Avg Loss: 0.77327517, Time: 0.0209 Steps: 76890, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000256, Sample Num: 4096, Cur Loss: 0.10249647, Cur Avg Loss: 0.17636272, Log Avg loss: 0.14032458, Global Avg Loss: 0.77319286, Time: 0.0254 Steps: 76900, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000266, Sample Num: 4256, Cur Loss: 0.06162981, Cur Avg Loss: 0.17727058, Log Avg loss: 0.20051192, Global Avg Loss: 0.77311840, Time: 0.0209 Steps: 76910, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000276, Sample Num: 4416, Cur Loss: 0.14777060, Cur Avg Loss: 0.17771842, Log Avg loss: 0.18963090, Global Avg Loss: 0.77304254, Time: 0.0209 Steps: 76920, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000286, Sample Num: 4576, Cur Loss: 0.25187114, Cur Avg Loss: 0.17863981, Log Avg loss: 0.20407032, Global Avg Loss: 0.77296858, Time: 0.0209 Steps: 76930, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000296, Sample Num: 4736, Cur Loss: 0.10656795, Cur Avg Loss: 0.17736136, Log Avg loss: 0.14079756, Global Avg Loss: 0.77288642, Time: 0.0210 Steps: 76940, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000306, Sample Num: 4896, Cur Loss: 0.24584073, Cur Avg Loss: 0.17840179, Log Avg loss: 0.20919861, Global Avg Loss: 0.77281316, Time: 0.0209 Steps: 76950, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000316, Sample Num: 5056, Cur Loss: 0.11123850, Cur Avg Loss: 0.17691258, Log Avg loss: 0.13134264, Global Avg Loss: 0.77272981, Time: 0.0209 Steps: 76960, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000326, Sample Num: 5216, Cur Loss: 0.04599895, Cur Avg Loss: 0.17699170, Log Avg loss: 0.17949187, Global Avg Loss: 0.77265274, Time: 0.0209 Steps: 76970, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000336, Sample Num: 5376, Cur Loss: 0.19932455, Cur Avg Loss: 0.17659288, Log Avg loss: 0.16359152, Global Avg Loss: 0.77257362, Time: 0.0209 Steps: 76980, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000346, Sample Num: 5536, Cur Loss: 0.23821941, Cur Avg Loss: 0.17773032, Log Avg loss: 0.21594817, Global Avg Loss: 0.77250132, Time: 0.0209 Steps: 76990, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000356, Sample Num: 5696, Cur Loss: 0.11332464, Cur Avg Loss: 0.17833217, Log Avg loss: 0.19915621, Global Avg Loss: 0.77242686, Time: 0.0209 Steps: 77000, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000366, Sample Num: 5856, Cur Loss: 0.09042594, Cur Avg Loss: 0.17691791, Log Avg loss: 0.12657043, Global Avg Loss: 0.77234299, Time: 0.0209 Steps: 77010, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000376, Sample Num: 6016, Cur Loss: 0.11749305, Cur Avg Loss: 0.17582057, Log Avg loss: 0.13565771, Global Avg Loss: 0.77226033, Time: 0.0209 Steps: 77020, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000386, Sample Num: 6176, Cur Loss: 0.14503610, Cur Avg Loss: 0.17575135, Log Avg loss: 0.17314865, Global Avg Loss: 0.77218255, Time: 0.0209 Steps: 77030, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000396, Sample Num: 6336, Cur Loss: 0.13731970, Cur Avg Loss: 0.17462463, Log Avg loss: 0.13113312, Global Avg Loss: 0.77209934, Time: 0.0209 Steps: 77040, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000406, Sample Num: 6496, Cur Loss: 0.14967604, Cur Avg Loss: 0.17540503, Log Avg loss: 0.20630916, Global Avg Loss: 0.77202591, Time: 0.0208 Steps: 77050, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000416, Sample Num: 6656, Cur Loss: 0.27188540, Cur Avg Loss: 0.17518024, Log Avg loss: 0.16605380, Global Avg Loss: 0.77194727, Time: 0.0209 Steps: 77060, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000426, Sample Num: 6816, Cur Loss: 0.20817211, Cur Avg Loss: 0.17655571, Log Avg loss: 0.23377523, Global Avg Loss: 0.77187744, Time: 0.0209 Steps: 77070, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000436, Sample Num: 6976, Cur Loss: 0.18096919, Cur Avg Loss: 0.17637242, Log Avg loss: 0.16856427, Global Avg Loss: 0.77179917, Time: 0.0209 Steps: 77080, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000446, Sample Num: 7136, Cur Loss: 0.58332336, Cur Avg Loss: 0.17722108, Log Avg loss: 0.21422240, Global Avg Loss: 0.77172685, Time: 0.0209 Steps: 77090, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000456, Sample Num: 7296, Cur Loss: 0.06229492, Cur Avg Loss: 0.17592502, Log Avg loss: 0.11812077, Global Avg Loss: 0.77164207, Time: 0.0209 Steps: 77100, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000466, Sample Num: 7456, Cur Loss: 0.25381821, Cur Avg Loss: 0.17556691, Log Avg loss: 0.15923708, Global Avg Loss: 0.77156265, Time: 0.0209 Steps: 77110, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000476, Sample Num: 7616, Cur Loss: 0.15176107, Cur Avg Loss: 0.17515833, Log Avg loss: 0.15611850, Global Avg Loss: 0.77148285, Time: 0.0209 Steps: 77120, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000486, Sample Num: 7776, Cur Loss: 0.10449834, Cur Avg Loss: 0.17460755, Log Avg loss: 0.14839074, Global Avg Loss: 0.77140206, Time: 0.0209 Steps: 77130, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000496, Sample Num: 7936, Cur Loss: 0.06936228, Cur Avg Loss: 0.17452100, Log Avg loss: 0.17031434, Global Avg Loss: 0.77132414, Time: 0.0209 Steps: 77140, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000506, Sample Num: 8096, Cur Loss: 0.15582904, Cur Avg Loss: 0.17484506, Log Avg loss: 0.19091839, Global Avg Loss: 0.77124891, Time: 0.0209 Steps: 77150, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000516, Sample Num: 8256, Cur Loss: 0.13146222, Cur Avg Loss: 0.17483455, Log Avg loss: 0.17430290, Global Avg Loss: 0.77117155, Time: 0.0246 Steps: 77160, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000526, Sample Num: 8416, Cur Loss: 0.21638095, Cur Avg Loss: 0.17520694, Log Avg loss: 0.19442208, Global Avg Loss: 0.77109681, Time: 0.0208 Steps: 77170, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000536, Sample Num: 8576, Cur Loss: 0.07366691, Cur Avg Loss: 0.17466740, Log Avg loss: 0.14628759, Global Avg Loss: 0.77101585, Time: 0.0209 Steps: 77180, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000546, Sample Num: 8736, Cur Loss: 0.12128597, Cur Avg Loss: 0.17355021, Log Avg loss: 0.11366891, Global Avg Loss: 0.77093069, Time: 0.0212 Steps: 77190, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000556, Sample Num: 8896, Cur Loss: 0.10855471, Cur Avg Loss: 0.17264637, Log Avg loss: 0.12329671, Global Avg Loss: 0.77084680, Time: 0.0211 Steps: 77200, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000566, Sample Num: 9056, Cur Loss: 0.18345030, Cur Avg Loss: 0.17228823, Log Avg loss: 0.15237562, Global Avg Loss: 0.77076670, Time: 0.0210 Steps: 77210, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000576, Sample Num: 9216, Cur Loss: 0.01654316, Cur Avg Loss: 0.17244814, Log Avg loss: 0.18149917, Global Avg Loss: 0.77069039, Time: 0.0208 Steps: 77220, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000586, Sample Num: 9376, Cur Loss: 0.15260212, Cur Avg Loss: 0.17271128, Log Avg loss: 0.18786840, Global Avg Loss: 0.77061493, Time: 0.0207 Steps: 77230, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000596, Sample Num: 9536, Cur Loss: 0.37612927, Cur Avg Loss: 0.17277022, Log Avg loss: 0.17622405, Global Avg Loss: 0.77053797, Time: 0.0209 Steps: 77240, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000606, Sample Num: 9696, Cur Loss: 0.08986008, Cur Avg Loss: 0.17291032, Log Avg loss: 0.18126026, Global Avg Loss: 0.77046169, Time: 0.0208 Steps: 77250, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000616, Sample Num: 9856, Cur Loss: 0.20883316, Cur Avg Loss: 0.17304647, Log Avg loss: 0.18129728, Global Avg Loss: 0.77038543, Time: 0.0208 Steps: 77260, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000626, Sample Num: 10016, Cur Loss: 0.24715269, Cur Avg Loss: 0.17235711, Log Avg loss: 0.12989234, Global Avg Loss: 0.77030254, Time: 0.0209 Steps: 77270, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000636, Sample Num: 10176, Cur Loss: 0.40691990, Cur Avg Loss: 0.17352829, Log Avg loss: 0.24684421, Global Avg Loss: 0.77023481, Time: 0.0211 Steps: 77280, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000646, Sample Num: 10336, Cur Loss: 0.15447159, Cur Avg Loss: 0.17353094, Log Avg loss: 0.17369964, Global Avg Loss: 0.77015763, Time: 0.0210 Steps: 77290, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000656, Sample Num: 10496, Cur Loss: 0.26972347, Cur Avg Loss: 0.17429788, Log Avg loss: 0.22384212, Global Avg Loss: 0.77008695, Time: 0.0210 Steps: 77300, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000666, Sample Num: 10656, Cur Loss: 0.12852658, Cur Avg Loss: 0.17488450, Log Avg loss: 0.21336644, Global Avg Loss: 0.77001494, Time: 0.0210 Steps: 77310, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000676, Sample Num: 10816, Cur Loss: 0.07943609, Cur Avg Loss: 0.17448462, Log Avg loss: 0.14785287, Global Avg Loss: 0.76993447, Time: 0.0210 Steps: 77320, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000686, Sample Num: 10976, Cur Loss: 0.08750540, Cur Avg Loss: 0.17427090, Log Avg loss: 0.15982312, Global Avg Loss: 0.76985558, Time: 0.0208 Steps: 77330, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000696, Sample Num: 11136, Cur Loss: 0.20176095, Cur Avg Loss: 0.17366264, Log Avg loss: 0.13193653, Global Avg Loss: 0.76977309, Time: 0.0210 Steps: 77340, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000706, Sample Num: 11296, Cur Loss: 0.25407878, Cur Avg Loss: 0.17352659, Log Avg loss: 0.16405730, Global Avg Loss: 0.76969479, Time: 0.0208 Steps: 77350, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000716, Sample Num: 11456, Cur Loss: 0.12122405, Cur Avg Loss: 0.17298369, Log Avg loss: 0.13465499, Global Avg Loss: 0.76961270, Time: 0.0209 Steps: 77360, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000726, Sample Num: 11616, Cur Loss: 0.19260886, Cur Avg Loss: 0.17287763, Log Avg loss: 0.16528370, Global Avg Loss: 0.76953459, Time: 0.0209 Steps: 77370, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000736, Sample Num: 11776, Cur Loss: 0.46081069, Cur Avg Loss: 0.17456503, Log Avg loss: 0.29707004, Global Avg Loss: 0.76947353, Time: 0.0210 Steps: 77380, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000746, Sample Num: 11936, Cur Loss: 0.09441720, Cur Avg Loss: 0.17545408, Log Avg loss: 0.24088796, Global Avg Loss: 0.76940523, Time: 0.0208 Steps: 77390, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000756, Sample Num: 12096, Cur Loss: 0.05686413, Cur Avg Loss: 0.17524195, Log Avg loss: 0.15941742, Global Avg Loss: 0.76932642, Time: 0.0209 Steps: 77400, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000766, Sample Num: 12256, Cur Loss: 0.15878645, Cur Avg Loss: 0.17555602, Log Avg loss: 0.19929969, Global Avg Loss: 0.76925278, Time: 0.0210 Steps: 77410, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000776, Sample Num: 12416, Cur Loss: 0.17509425, Cur Avg Loss: 0.17579903, Log Avg loss: 0.19441344, Global Avg Loss: 0.76917853, Time: 0.0209 Steps: 77420, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000786, Sample Num: 12576, Cur Loss: 0.31758094, Cur Avg Loss: 0.17575610, Log Avg loss: 0.17242511, Global Avg Loss: 0.76910146, Time: 0.0209 Steps: 77430, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000796, Sample Num: 12736, Cur Loss: 0.07993095, Cur Avg Loss: 0.17560782, Log Avg loss: 0.16395300, Global Avg Loss: 0.76902332, Time: 0.0210 Steps: 77440, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000806, Sample Num: 12896, Cur Loss: 0.10051262, Cur Avg Loss: 0.17572350, Log Avg loss: 0.18493164, Global Avg Loss: 0.76894790, Time: 0.0210 Steps: 77450, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000816, Sample Num: 13056, Cur Loss: 0.11969766, Cur Avg Loss: 0.17553313, Log Avg loss: 0.16018925, Global Avg Loss: 0.76886931, Time: 0.0209 Steps: 77460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000826, Sample Num: 13216, Cur Loss: 0.13129225, Cur Avg Loss: 0.17571416, Log Avg loss: 0.19048628, Global Avg Loss: 0.76879465, Time: 0.0209 Steps: 77470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000836, Sample Num: 13376, Cur Loss: 0.10558723, Cur Avg Loss: 0.17521525, Log Avg loss: 0.13400463, Global Avg Loss: 0.76871272, Time: 0.0209 Steps: 77480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000846, Sample Num: 13536, Cur Loss: 0.19275847, Cur Avg Loss: 0.17473395, Log Avg loss: 0.13449789, Global Avg Loss: 0.76863088, Time: 0.0209 Steps: 77490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000856, Sample Num: 13696, Cur Loss: 0.03625972, Cur Avg Loss: 0.17509770, Log Avg loss: 0.20587089, Global Avg Loss: 0.76855826, Time: 0.0209 Steps: 77500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000866, Sample Num: 13856, Cur Loss: 0.26689959, Cur Avg Loss: 0.17537834, Log Avg loss: 0.19940071, Global Avg Loss: 0.76848483, Time: 0.0209 Steps: 77510, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000876, Sample Num: 14016, Cur Loss: 0.04944255, Cur Avg Loss: 0.17512679, Log Avg loss: 0.15334300, Global Avg Loss: 0.76840548, Time: 0.0210 Steps: 77520, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000886, Sample Num: 14176, Cur Loss: 0.23246700, Cur Avg Loss: 0.17563359, Log Avg loss: 0.22002910, Global Avg Loss: 0.76833475, Time: 0.0209 Steps: 77530, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000896, Sample Num: 14336, Cur Loss: 0.50828457, Cur Avg Loss: 0.17567725, Log Avg loss: 0.17954565, Global Avg Loss: 0.76825882, Time: 0.0209 Steps: 77540, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000906, Sample Num: 14496, Cur Loss: 0.42340547, Cur Avg Loss: 0.17615113, Log Avg loss: 0.21861025, Global Avg Loss: 0.76818794, Time: 0.0209 Steps: 77550, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000916, Sample Num: 14656, Cur Loss: 0.02902320, Cur Avg Loss: 0.17593718, Log Avg loss: 0.15655342, Global Avg Loss: 0.76810908, Time: 0.0209 Steps: 77560, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000926, Sample Num: 14816, Cur Loss: 0.15151584, Cur Avg Loss: 0.17615034, Log Avg loss: 0.19567593, Global Avg Loss: 0.76803529, Time: 0.0209 Steps: 77570, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000936, Sample Num: 14976, Cur Loss: 0.22316426, Cur Avg Loss: 0.17594718, Log Avg loss: 0.15713441, Global Avg Loss: 0.76795654, Time: 0.0209 Steps: 77580, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000946, Sample Num: 15136, Cur Loss: 0.30225402, Cur Avg Loss: 0.17571821, Log Avg loss: 0.15428685, Global Avg Loss: 0.76787745, Time: 0.0210 Steps: 77590, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000956, Sample Num: 15296, Cur Loss: 0.27518088, Cur Avg Loss: 0.17568538, Log Avg loss: 0.17258012, Global Avg Loss: 0.76780074, Time: 0.0209 Steps: 77600, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000966, Sample Num: 15456, Cur Loss: 0.10703769, Cur Avg Loss: 0.17598443, Log Avg loss: 0.20457340, Global Avg Loss: 0.76772816, Time: 0.0209 Steps: 77610, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000976, Sample Num: 15616, Cur Loss: 0.09904011, Cur Avg Loss: 0.17589094, Log Avg loss: 0.16685952, Global Avg Loss: 0.76765075, Time: 0.0209 Steps: 77620, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000986, Sample Num: 15776, Cur Loss: 0.33791065, Cur Avg Loss: 0.17589465, Log Avg loss: 0.17625727, Global Avg Loss: 0.76757457, Time: 0.0209 Steps: 77630, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000996, Sample Num: 15936, Cur Loss: 0.12089225, Cur Avg Loss: 0.17569877, Log Avg loss: 0.15638414, Global Avg Loss: 0.76749585, Time: 0.0209 Steps: 77640, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001006, Sample Num: 16096, Cur Loss: 0.15815158, Cur Avg Loss: 0.17587274, Log Avg loss: 0.19320058, Global Avg Loss: 0.76742189, Time: 0.0209 Steps: 77650, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001016, Sample Num: 16256, Cur Loss: 0.24825802, Cur Avg Loss: 0.17558932, Log Avg loss: 0.14707750, Global Avg Loss: 0.76734201, Time: 0.0210 Steps: 77660, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001026, Sample Num: 16416, Cur Loss: 0.03336526, Cur Avg Loss: 0.17507689, Log Avg loss: 0.12301419, Global Avg Loss: 0.76725905, Time: 0.0247 Steps: 77670, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001036, Sample Num: 16576, Cur Loss: 0.02619652, Cur Avg Loss: 0.17511809, Log Avg loss: 0.17934509, Global Avg Loss: 0.76718337, Time: 0.0209 Steps: 77680, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001046, Sample Num: 16736, Cur Loss: 0.39253592, Cur Avg Loss: 0.17591655, Log Avg loss: 0.25863663, Global Avg Loss: 0.76711791, Time: 0.0210 Steps: 77690, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001056, Sample Num: 16896, Cur Loss: 0.13338667, Cur Avg Loss: 0.17593320, Log Avg loss: 0.17767529, Global Avg Loss: 0.76704205, Time: 0.0209 Steps: 77700, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001066, Sample Num: 17056, Cur Loss: 0.04888498, Cur Avg Loss: 0.17621094, Log Avg loss: 0.20554027, Global Avg Loss: 0.76696979, Time: 0.0209 Steps: 77710, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001076, Sample Num: 17216, Cur Loss: 0.10584627, Cur Avg Loss: 0.17656866, Log Avg loss: 0.21470076, Global Avg Loss: 0.76689874, Time: 0.0209 Steps: 77720, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001086, Sample Num: 17376, Cur Loss: 0.04484185, Cur Avg Loss: 0.17661101, Log Avg loss: 0.18116823, Global Avg Loss: 0.76682338, Time: 0.0209 Steps: 77730, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001096, Sample Num: 17536, Cur Loss: 0.31142014, Cur Avg Loss: 0.17693272, Log Avg loss: 0.21187077, Global Avg Loss: 0.76675200, Time: 0.0209 Steps: 77740, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001106, Sample Num: 17696, Cur Loss: 0.07321480, Cur Avg Loss: 0.17694245, Log Avg loss: 0.17800867, Global Avg Loss: 0.76667627, Time: 0.0209 Steps: 77750, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001116, Sample Num: 17856, Cur Loss: 0.17652756, Cur Avg Loss: 0.17673016, Log Avg loss: 0.15325124, Global Avg Loss: 0.76659739, Time: 0.0209 Steps: 77760, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001126, Sample Num: 18016, Cur Loss: 0.14026989, Cur Avg Loss: 0.17689145, Log Avg loss: 0.19489119, Global Avg Loss: 0.76652387, Time: 0.0209 Steps: 77770, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001136, Sample Num: 18176, Cur Loss: 0.16729879, Cur Avg Loss: 0.17666511, Log Avg loss: 0.15117896, Global Avg Loss: 0.76644476, Time: 0.0209 Steps: 77780, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001146, Sample Num: 18336, Cur Loss: 0.11162339, Cur Avg Loss: 0.17629543, Log Avg loss: 0.13429992, Global Avg Loss: 0.76636350, Time: 0.0209 Steps: 77790, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001156, Sample Num: 18496, Cur Loss: 0.70283544, Cur Avg Loss: 0.17688460, Log Avg loss: 0.24440324, Global Avg Loss: 0.76629641, Time: 0.0209 Steps: 77800, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001166, Sample Num: 18656, Cur Loss: 0.08561552, Cur Avg Loss: 0.17739520, Log Avg loss: 0.23642041, Global Avg Loss: 0.76622831, Time: 0.0209 Steps: 77810, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001176, Sample Num: 18816, Cur Loss: 0.06061481, Cur Avg Loss: 0.17750569, Log Avg loss: 0.19038871, Global Avg Loss: 0.76615431, Time: 0.0209 Steps: 77820, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001186, Sample Num: 18976, Cur Loss: 0.17029600, Cur Avg Loss: 0.17743223, Log Avg loss: 0.16879356, Global Avg Loss: 0.76607756, Time: 0.0209 Steps: 77830, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001196, Sample Num: 19136, Cur Loss: 0.10953782, Cur Avg Loss: 0.17706429, Log Avg loss: 0.13342671, Global Avg Loss: 0.76599628, Time: 0.0210 Steps: 77840, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001206, Sample Num: 19296, Cur Loss: 0.22531673, Cur Avg Loss: 0.17711866, Log Avg loss: 0.18362144, Global Avg Loss: 0.76592148, Time: 0.0209 Steps: 77850, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001216, Sample Num: 19456, Cur Loss: 0.30176759, Cur Avg Loss: 0.17741266, Log Avg loss: 0.21286940, Global Avg Loss: 0.76585045, Time: 0.0209 Steps: 77860, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001226, Sample Num: 19616, Cur Loss: 0.08116558, Cur Avg Loss: 0.17740902, Log Avg loss: 0.17696573, Global Avg Loss: 0.76577482, Time: 0.0209 Steps: 77870, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001236, Sample Num: 19776, Cur Loss: 0.26484799, Cur Avg Loss: 0.17726921, Log Avg loss: 0.16012899, Global Avg Loss: 0.76569706, Time: 0.0209 Steps: 77880, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001246, Sample Num: 19936, Cur Loss: 0.09169725, Cur Avg Loss: 0.17703977, Log Avg loss: 0.14868034, Global Avg Loss: 0.76561784, Time: 0.0209 Steps: 77890, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001256, Sample Num: 20096, Cur Loss: 0.09352496, Cur Avg Loss: 0.17684671, Log Avg loss: 0.15279177, Global Avg Loss: 0.76553917, Time: 0.0209 Steps: 77900, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001266, Sample Num: 20256, Cur Loss: 0.25403380, Cur Avg Loss: 0.17717232, Log Avg loss: 0.21806962, Global Avg Loss: 0.76546890, Time: 0.0209 Steps: 77910, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001276, Sample Num: 20416, Cur Loss: 0.19412071, Cur Avg Loss: 0.17711012, Log Avg loss: 0.16923466, Global Avg Loss: 0.76539238, Time: 0.0209 Steps: 77920, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001286, Sample Num: 20576, Cur Loss: 0.04908455, Cur Avg Loss: 0.17711468, Log Avg loss: 0.17769741, Global Avg Loss: 0.76531697, Time: 0.0247 Steps: 77930, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001296, Sample Num: 20736, Cur Loss: 0.11976455, Cur Avg Loss: 0.17697541, Log Avg loss: 0.15906471, Global Avg Loss: 0.76523918, Time: 0.0211 Steps: 77940, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001306, Sample Num: 20896, Cur Loss: 0.07587221, Cur Avg Loss: 0.17689709, Log Avg loss: 0.16674695, Global Avg Loss: 0.76516241, Time: 0.0211 Steps: 77950, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001316, Sample Num: 21056, Cur Loss: 0.08184766, Cur Avg Loss: 0.17703189, Log Avg loss: 0.19463735, Global Avg Loss: 0.76508922, Time: 0.0210 Steps: 77960, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001326, Sample Num: 21216, Cur Loss: 0.07184612, Cur Avg Loss: 0.17679449, Log Avg loss: 0.14555161, Global Avg Loss: 0.76500977, Time: 0.0210 Steps: 77970, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001336, Sample Num: 21376, Cur Loss: 0.17347479, Cur Avg Loss: 0.17661197, Log Avg loss: 0.15241047, Global Avg Loss: 0.76493121, Time: 0.0210 Steps: 77980, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001346, Sample Num: 21536, Cur Loss: 0.15274303, Cur Avg Loss: 0.17637365, Log Avg loss: 0.14453455, Global Avg Loss: 0.76485166, Time: 0.0210 Steps: 77990, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001356, Sample Num: 21696, Cur Loss: 0.14392586, Cur Avg Loss: 0.17633757, Log Avg loss: 0.17148074, Global Avg Loss: 0.76477559, Time: 0.0209 Steps: 78000, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001366, Sample Num: 21856, Cur Loss: 0.19180191, Cur Avg Loss: 0.17636085, Log Avg loss: 0.17951785, Global Avg Loss: 0.76470056, Time: 0.0210 Steps: 78010, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001376, Sample Num: 22016, Cur Loss: 0.14166871, Cur Avg Loss: 0.17606937, Log Avg loss: 0.13625236, Global Avg Loss: 0.76462001, Time: 0.0210 Steps: 78020, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001386, Sample Num: 22176, Cur Loss: 0.17291541, Cur Avg Loss: 0.17614095, Log Avg loss: 0.18599086, Global Avg Loss: 0.76454586, Time: 0.0210 Steps: 78030, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001396, Sample Num: 22336, Cur Loss: 0.11241333, Cur Avg Loss: 0.17645521, Log Avg loss: 0.22001204, Global Avg Loss: 0.76447608, Time: 0.0210 Steps: 78040, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001406, Sample Num: 22496, Cur Loss: 0.08567004, Cur Avg Loss: 0.17651811, Log Avg loss: 0.18529835, Global Avg Loss: 0.76440188, Time: 0.0210 Steps: 78050, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001416, Sample Num: 22656, Cur Loss: 0.09716175, Cur Avg Loss: 0.17658523, Log Avg loss: 0.18602278, Global Avg Loss: 0.76432778, Time: 0.0210 Steps: 78060, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001426, Sample Num: 22816, Cur Loss: 0.13474146, Cur Avg Loss: 0.17654032, Log Avg loss: 0.17018032, Global Avg Loss: 0.76425168, Time: 0.0210 Steps: 78070, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001436, Sample Num: 22976, Cur Loss: 0.20090494, Cur Avg Loss: 0.17640760, Log Avg loss: 0.15748143, Global Avg Loss: 0.76417397, Time: 0.0210 Steps: 78080, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001446, Sample Num: 23136, Cur Loss: 0.22000688, Cur Avg Loss: 0.17688789, Log Avg loss: 0.24585884, Global Avg Loss: 0.76410759, Time: 0.0210 Steps: 78090, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001456, Sample Num: 23296, Cur Loss: 0.04685811, Cur Avg Loss: 0.17651306, Log Avg loss: 0.12231237, Global Avg Loss: 0.76402542, Time: 0.0210 Steps: 78100, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001466, Sample Num: 23456, Cur Loss: 0.26669639, Cur Avg Loss: 0.17641881, Log Avg loss: 0.16269520, Global Avg Loss: 0.76394843, Time: 0.0210 Steps: 78110, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001476, Sample Num: 23616, Cur Loss: 0.20812052, Cur Avg Loss: 0.17649258, Log Avg loss: 0.18730810, Global Avg Loss: 0.76387462, Time: 0.0210 Steps: 78120, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001486, Sample Num: 23776, Cur Loss: 0.22860011, Cur Avg Loss: 0.17656158, Log Avg loss: 0.18674553, Global Avg Loss: 0.76380075, Time: 0.0210 Steps: 78130, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001496, Sample Num: 23936, Cur Loss: 0.16064489, Cur Avg Loss: 0.17701332, Log Avg loss: 0.24414123, Global Avg Loss: 0.76373424, Time: 0.0210 Steps: 78140, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001506, Sample Num: 24096, Cur Loss: 0.07279266, Cur Avg Loss: 0.17738883, Log Avg loss: 0.23356582, Global Avg Loss: 0.76366640, Time: 0.0210 Steps: 78150, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001516, Sample Num: 24256, Cur Loss: 0.11962353, Cur Avg Loss: 0.17758959, Log Avg loss: 0.20782374, Global Avg Loss: 0.76359529, Time: 0.0211 Steps: 78160, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001526, Sample Num: 24416, Cur Loss: 0.26534516, Cur Avg Loss: 0.17767282, Log Avg loss: 0.19029061, Global Avg Loss: 0.76352195, Time: 0.0210 Steps: 78170, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001536, Sample Num: 24576, Cur Loss: 0.12580194, Cur Avg Loss: 0.17786898, Log Avg loss: 0.20780353, Global Avg Loss: 0.76345087, Time: 0.0255 Steps: 78180, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001546, Sample Num: 24736, Cur Loss: 0.23473470, Cur Avg Loss: 0.17800422, Log Avg loss: 0.19877611, Global Avg Loss: 0.76337865, Time: 0.0209 Steps: 78190, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001556, Sample Num: 24896, Cur Loss: 0.10664484, Cur Avg Loss: 0.17782651, Log Avg loss: 0.15035302, Global Avg Loss: 0.76330026, Time: 0.0209 Steps: 78200, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001566, Sample Num: 25056, Cur Loss: 0.11659595, Cur Avg Loss: 0.17763830, Log Avg loss: 0.14835259, Global Avg Loss: 0.76322163, Time: 0.0209 Steps: 78210, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001576, Sample Num: 25216, Cur Loss: 0.20846684, Cur Avg Loss: 0.17742338, Log Avg loss: 0.14376672, Global Avg Loss: 0.76314243, Time: 0.0208 Steps: 78220, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001586, Sample Num: 25376, Cur Loss: 0.04552702, Cur Avg Loss: 0.17758016, Log Avg loss: 0.20228990, Global Avg Loss: 0.76307074, Time: 0.0209 Steps: 78230, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001596, Sample Num: 25536, Cur Loss: 0.11217697, Cur Avg Loss: 0.17753865, Log Avg loss: 0.17095511, Global Avg Loss: 0.76299506, Time: 0.0209 Steps: 78240, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001606, Sample Num: 25696, Cur Loss: 0.14339663, Cur Avg Loss: 0.17738794, Log Avg loss: 0.15333395, Global Avg Loss: 0.76291715, Time: 0.0209 Steps: 78250, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001616, Sample Num: 25856, Cur Loss: 0.19982563, Cur Avg Loss: 0.17728572, Log Avg loss: 0.16086925, Global Avg Loss: 0.76284022, Time: 0.0209 Steps: 78260, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001626, Sample Num: 26016, Cur Loss: 0.16147390, Cur Avg Loss: 0.17750387, Log Avg loss: 0.21275769, Global Avg Loss: 0.76276994, Time: 0.0209 Steps: 78270, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001636, Sample Num: 26176, Cur Loss: 0.12252177, Cur Avg Loss: 0.17757349, Log Avg loss: 0.18889329, Global Avg Loss: 0.76269663, Time: 0.0209 Steps: 78280, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001646, Sample Num: 26336, Cur Loss: 0.09114040, Cur Avg Loss: 0.17737987, Log Avg loss: 0.14570380, Global Avg Loss: 0.76261782, Time: 0.0208 Steps: 78290, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001656, Sample Num: 26496, Cur Loss: 0.23393609, Cur Avg Loss: 0.17720141, Log Avg loss: 0.14782706, Global Avg Loss: 0.76253930, Time: 0.0209 Steps: 78300, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001666, Sample Num: 26656, Cur Loss: 0.20681408, Cur Avg Loss: 0.17714616, Log Avg loss: 0.16799664, Global Avg Loss: 0.76246338, Time: 0.0209 Steps: 78310, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001676, Sample Num: 26816, Cur Loss: 0.39612836, Cur Avg Loss: 0.17712158, Log Avg loss: 0.17302568, Global Avg Loss: 0.76238812, Time: 0.0209 Steps: 78320, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001686, Sample Num: 26976, Cur Loss: 0.20386605, Cur Avg Loss: 0.17687813, Log Avg loss: 0.13607701, Global Avg Loss: 0.76230816, Time: 0.0209 Steps: 78330, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001696, Sample Num: 27136, Cur Loss: 0.04689445, Cur Avg Loss: 0.17663572, Log Avg loss: 0.13576429, Global Avg Loss: 0.76222819, Time: 0.0209 Steps: 78340, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001706, Sample Num: 27296, Cur Loss: 0.09866055, Cur Avg Loss: 0.17682084, Log Avg loss: 0.20821724, Global Avg Loss: 0.76215748, Time: 0.0209 Steps: 78350, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001716, Sample Num: 27456, Cur Loss: 0.09214921, Cur Avg Loss: 0.17678313, Log Avg loss: 0.17035044, Global Avg Loss: 0.76208195, Time: 0.0209 Steps: 78360, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001726, Sample Num: 27616, Cur Loss: 0.13848951, Cur Avg Loss: 0.17679107, Log Avg loss: 0.17815267, Global Avg Loss: 0.76200744, Time: 0.0209 Steps: 78370, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001736, Sample Num: 27776, Cur Loss: 0.27501476, Cur Avg Loss: 0.17679512, Log Avg loss: 0.17749425, Global Avg Loss: 0.76193287, Time: 0.0209 Steps: 78380, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001746, Sample Num: 27936, Cur Loss: 0.28656313, Cur Avg Loss: 0.17668286, Log Avg loss: 0.15719450, Global Avg Loss: 0.76185572, Time: 0.0209 Steps: 78390, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001756, Sample Num: 28096, Cur Loss: 0.24514641, Cur Avg Loss: 0.17697092, Log Avg loss: 0.22726646, Global Avg Loss: 0.76178754, Time: 0.0209 Steps: 78400, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001766, Sample Num: 28256, Cur Loss: 0.06202077, Cur Avg Loss: 0.17721833, Log Avg loss: 0.22066378, Global Avg Loss: 0.76171852, Time: 0.0209 Steps: 78410, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001776, Sample Num: 28416, Cur Loss: 0.11486638, Cur Avg Loss: 0.17761372, Log Avg loss: 0.24744030, Global Avg Loss: 0.76165294, Time: 0.0208 Steps: 78420, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001786, Sample Num: 28576, Cur Loss: 0.13852158, Cur Avg Loss: 0.17753266, Log Avg loss: 0.16313524, Global Avg Loss: 0.76157663, Time: 0.0209 Steps: 78430, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001796, Sample Num: 28736, Cur Loss: 0.08041070, Cur Avg Loss: 0.17756997, Log Avg loss: 0.18423402, Global Avg Loss: 0.76150303, Time: 0.0246 Steps: 78440, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001806, Sample Num: 28896, Cur Loss: 0.46237457, Cur Avg Loss: 0.17793123, Log Avg loss: 0.24281332, Global Avg Loss: 0.76143691, Time: 0.0208 Steps: 78450, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001816, Sample Num: 29056, Cur Loss: 0.25471237, Cur Avg Loss: 0.17838496, Log Avg loss: 0.26032790, Global Avg Loss: 0.76137304, Time: 0.0209 Steps: 78460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001826, Sample Num: 29216, Cur Loss: 0.30579558, Cur Avg Loss: 0.17828482, Log Avg loss: 0.16009942, Global Avg Loss: 0.76129642, Time: 0.0208 Steps: 78470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001836, Sample Num: 29376, Cur Loss: 0.11816742, Cur Avg Loss: 0.17833114, Log Avg loss: 0.18679004, Global Avg Loss: 0.76122321, Time: 0.0209 Steps: 78480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001846, Sample Num: 29536, Cur Loss: 0.05853094, Cur Avg Loss: 0.17813979, Log Avg loss: 0.14300736, Global Avg Loss: 0.76114445, Time: 0.0209 Steps: 78490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001856, Sample Num: 29696, Cur Loss: 0.24036589, Cur Avg Loss: 0.17843616, Log Avg loss: 0.23314736, Global Avg Loss: 0.76107719, Time: 0.0209 Steps: 78500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001866, Sample Num: 29856, Cur Loss: 0.13095292, Cur Avg Loss: 0.17866236, Log Avg loss: 0.22064504, Global Avg Loss: 0.76100835, Time: 0.0209 Steps: 78510, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001876, Sample Num: 30016, Cur Loss: 0.16147846, Cur Avg Loss: 0.17877169, Log Avg loss: 0.19917135, Global Avg Loss: 0.76093680, Time: 0.0209 Steps: 78520, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001886, Sample Num: 30176, Cur Loss: 0.07443507, Cur Avg Loss: 0.17880900, Log Avg loss: 0.18580963, Global Avg Loss: 0.76086356, Time: 0.0209 Steps: 78530, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001896, Sample Num: 30336, Cur Loss: 0.28920877, Cur Avg Loss: 0.17916853, Log Avg loss: 0.24697600, Global Avg Loss: 0.76079813, Time: 0.0209 Steps: 78540, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001906, Sample Num: 30496, Cur Loss: 0.10706370, Cur Avg Loss: 0.17914083, Log Avg loss: 0.17388864, Global Avg Loss: 0.76072342, Time: 0.0209 Steps: 78550, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001916, Sample Num: 30656, Cur Loss: 0.31631792, Cur Avg Loss: 0.17908093, Log Avg loss: 0.16766386, Global Avg Loss: 0.76064792, Time: 0.0209 Steps: 78560, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001926, Sample Num: 30816, Cur Loss: 0.08806975, Cur Avg Loss: 0.17899623, Log Avg loss: 0.16276725, Global Avg Loss: 0.76057183, Time: 0.0209 Steps: 78570, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001936, Sample Num: 30976, Cur Loss: 0.41483027, Cur Avg Loss: 0.17929803, Log Avg loss: 0.23742464, Global Avg Loss: 0.76050525, Time: 0.0209 Steps: 78580, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001946, Sample Num: 31136, Cur Loss: 0.14214161, Cur Avg Loss: 0.17927361, Log Avg loss: 0.17454660, Global Avg Loss: 0.76043070, Time: 0.0209 Steps: 78590, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001956, Sample Num: 31296, Cur Loss: 0.23361759, Cur Avg Loss: 0.17921543, Log Avg loss: 0.16789206, Global Avg Loss: 0.76035531, Time: 0.0209 Steps: 78600, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001966, Sample Num: 31456, Cur Loss: 0.26566797, Cur Avg Loss: 0.17913745, Log Avg loss: 0.16388536, Global Avg Loss: 0.76027943, Time: 0.0209 Steps: 78610, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001976, Sample Num: 31616, Cur Loss: 0.12786615, Cur Avg Loss: 0.17900192, Log Avg loss: 0.15235623, Global Avg Loss: 0.76020211, Time: 0.0209 Steps: 78620, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001986, Sample Num: 31776, Cur Loss: 0.23692614, Cur Avg Loss: 0.17899302, Log Avg loss: 0.17723560, Global Avg Loss: 0.76012797, Time: 0.0209 Steps: 78630, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001996, Sample Num: 31936, Cur Loss: 0.04809017, Cur Avg Loss: 0.17867006, Log Avg loss: 0.11452882, Global Avg Loss: 0.76004587, Time: 0.0209 Steps: 78640, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002006, Sample Num: 32096, Cur Loss: 0.24181628, Cur Avg Loss: 0.17896204, Log Avg loss: 0.23724150, Global Avg Loss: 0.75997940, Time: 0.0208 Steps: 78650, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002016, Sample Num: 32256, Cur Loss: 0.12251712, Cur Avg Loss: 0.17890825, Log Avg loss: 0.16811932, Global Avg Loss: 0.75990416, Time: 0.0208 Steps: 78660, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002026, Sample Num: 32416, Cur Loss: 0.37078708, Cur Avg Loss: 0.17933901, Log Avg loss: 0.26618021, Global Avg Loss: 0.75984140, Time: 0.0208 Steps: 78670, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002036, Sample Num: 32576, Cur Loss: 0.17416817, Cur Avg Loss: 0.17942966, Log Avg loss: 0.19779497, Global Avg Loss: 0.75976996, Time: 0.0209 Steps: 78680, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002046, Sample Num: 32736, Cur Loss: 0.59918082, Cur Avg Loss: 0.17943894, Log Avg loss: 0.18132718, Global Avg Loss: 0.75969645, Time: 0.0209 Steps: 78690, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002056, Sample Num: 32896, Cur Loss: 0.30404514, Cur Avg Loss: 0.17955107, Log Avg loss: 0.20249400, Global Avg Loss: 0.75962565, Time: 0.0211 Steps: 78700, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002066, Sample Num: 33056, Cur Loss: 0.18006109, Cur Avg Loss: 0.17948834, Log Avg loss: 0.16659127, Global Avg Loss: 0.75955031, Time: 0.0209 Steps: 78710, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002076, Sample Num: 33216, Cur Loss: 0.11820348, Cur Avg Loss: 0.17913398, Log Avg loss: 0.10592228, Global Avg Loss: 0.75946728, Time: 0.0209 Steps: 78720, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002086, Sample Num: 33376, Cur Loss: 0.19286805, Cur Avg Loss: 0.17912544, Log Avg loss: 0.17735220, Global Avg Loss: 0.75939334, Time: 0.0209 Steps: 78730, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002096, Sample Num: 33536, Cur Loss: 0.35781643, Cur Avg Loss: 0.17903160, Log Avg loss: 0.15945769, Global Avg Loss: 0.75931715, Time: 0.0209 Steps: 78740, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002106, Sample Num: 33696, Cur Loss: 0.14453208, Cur Avg Loss: 0.17879609, Log Avg loss: 0.12943229, Global Avg Loss: 0.75923716, Time: 0.0209 Steps: 78750, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002116, Sample Num: 33856, Cur Loss: 0.25350899, Cur Avg Loss: 0.17881647, Log Avg loss: 0.18310880, Global Avg Loss: 0.75916401, Time: 0.0210 Steps: 78760, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002126, Sample Num: 34016, Cur Loss: 0.03255664, Cur Avg Loss: 0.17866414, Log Avg loss: 0.14643070, Global Avg Loss: 0.75908622, Time: 0.0209 Steps: 78770, Updated lr: 0.000026 ***** Running evaluation checkpoint-78773 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-78773 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.675963, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.281563, "eval_total_loss": 197.938876, "eval_mae": 0.360288, "eval_mse": 0.281671, "eval_r2": 0.820951, "eval_sp_statistic": 0.892805, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.91955, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.990149, "test_total_loss": 497.054827, "test_mae": 0.814208, "test_mse": 0.990194, "test_r2": 0.36092, "test_sp_statistic": 0.792464, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.827536, "test_ps_pvalue": 0.0, "lr": 2.624656235182551e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7590613677620105, "train_cur_epoch_loss": 380.15922381542623, "train_cur_epoch_avg_loss": 0.1785623409184717, "train_cur_epoch_time": 44.675962924957275, "train_cur_epoch_avg_time": 0.020984482350848883, "epoch": 37, "step": 78773} ################################################## Training, Epoch: 0038, Batch: 000007, Sample Num: 112, Cur Loss: 0.14653920, Cur Avg Loss: 0.18104096, Log Avg loss: 0.15865550, Global Avg Loss: 0.75901001, Time: 0.0210 Steps: 78780, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000017, Sample Num: 272, Cur Loss: 0.30198601, Cur Avg Loss: 0.21181537, Log Avg loss: 0.23335746, Global Avg Loss: 0.75894329, Time: 0.0209 Steps: 78790, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000027, Sample Num: 432, Cur Loss: 0.10387962, Cur Avg Loss: 0.18438092, Log Avg loss: 0.13774234, Global Avg Loss: 0.75886446, Time: 0.0209 Steps: 78800, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000037, Sample Num: 592, Cur Loss: 0.11296705, Cur Avg Loss: 0.17590436, Log Avg loss: 0.15301766, Global Avg Loss: 0.75878759, Time: 0.0209 Steps: 78810, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000047, Sample Num: 752, Cur Loss: 0.16816074, Cur Avg Loss: 0.16718178, Log Avg loss: 0.13490823, Global Avg Loss: 0.75870843, Time: 0.0209 Steps: 78820, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000057, Sample Num: 912, Cur Loss: 0.18049803, Cur Avg Loss: 0.16794894, Log Avg loss: 0.17155457, Global Avg Loss: 0.75863395, Time: 0.0209 Steps: 78830, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000067, Sample Num: 1072, Cur Loss: 0.08799168, Cur Avg Loss: 0.16963244, Log Avg loss: 0.17922841, Global Avg Loss: 0.75856046, Time: 0.0209 Steps: 78840, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000077, Sample Num: 1232, Cur Loss: 0.08821195, Cur Avg Loss: 0.17305956, Log Avg loss: 0.19602128, Global Avg Loss: 0.75848911, Time: 0.0209 Steps: 78850, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000087, Sample Num: 1392, Cur Loss: 0.07279557, Cur Avg Loss: 0.16505716, Log Avg loss: 0.10343865, Global Avg Loss: 0.75840605, Time: 0.0209 Steps: 78860, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000097, Sample Num: 1552, Cur Loss: 0.36215025, Cur Avg Loss: 0.16686560, Log Avg loss: 0.18259901, Global Avg Loss: 0.75833304, Time: 0.0209 Steps: 78870, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000107, Sample Num: 1712, Cur Loss: 0.12586014, Cur Avg Loss: 0.16191615, Log Avg loss: 0.11390649, Global Avg Loss: 0.75825135, Time: 0.0209 Steps: 78880, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000117, Sample Num: 1872, Cur Loss: 0.29573944, Cur Avg Loss: 0.16999253, Log Avg loss: 0.25640978, Global Avg Loss: 0.75818773, Time: 0.0209 Steps: 78890, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000127, Sample Num: 2032, Cur Loss: 0.13033891, Cur Avg Loss: 0.17186227, Log Avg loss: 0.19373832, Global Avg Loss: 0.75811619, Time: 0.0209 Steps: 78900, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000137, Sample Num: 2192, Cur Loss: 0.27545181, Cur Avg Loss: 0.16829139, Log Avg loss: 0.12294124, Global Avg Loss: 0.75803570, Time: 0.0209 Steps: 78910, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000147, Sample Num: 2352, Cur Loss: 0.11007293, Cur Avg Loss: 0.16943394, Log Avg loss: 0.18508685, Global Avg Loss: 0.75796310, Time: 0.0209 Steps: 78920, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000157, Sample Num: 2512, Cur Loss: 0.13271271, Cur Avg Loss: 0.17155938, Log Avg loss: 0.20280330, Global Avg Loss: 0.75789277, Time: 0.0209 Steps: 78930, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000167, Sample Num: 2672, Cur Loss: 0.18908688, Cur Avg Loss: 0.17683451, Log Avg loss: 0.25965405, Global Avg Loss: 0.75782965, Time: 0.0209 Steps: 78940, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000177, Sample Num: 2832, Cur Loss: 0.07619876, Cur Avg Loss: 0.17546297, Log Avg loss: 0.15255819, Global Avg Loss: 0.75775298, Time: 0.0209 Steps: 78950, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000187, Sample Num: 2992, Cur Loss: 0.14790115, Cur Avg Loss: 0.17518620, Log Avg loss: 0.17028735, Global Avg Loss: 0.75767858, Time: 0.0209 Steps: 78960, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000197, Sample Num: 3152, Cur Loss: 0.18865833, Cur Avg Loss: 0.17451412, Log Avg loss: 0.16194631, Global Avg Loss: 0.75760315, Time: 0.0209 Steps: 78970, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000207, Sample Num: 3312, Cur Loss: 0.19498762, Cur Avg Loss: 0.17669447, Log Avg loss: 0.21964743, Global Avg Loss: 0.75753503, Time: 0.0210 Steps: 78980, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000217, Sample Num: 3472, Cur Loss: 0.20125997, Cur Avg Loss: 0.17706386, Log Avg loss: 0.18471015, Global Avg Loss: 0.75746251, Time: 0.0208 Steps: 78990, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000227, Sample Num: 3632, Cur Loss: 0.09480134, Cur Avg Loss: 0.17464433, Log Avg loss: 0.12214056, Global Avg Loss: 0.75738209, Time: 0.0209 Steps: 79000, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000237, Sample Num: 3792, Cur Loss: 0.30256203, Cur Avg Loss: 0.17616296, Log Avg loss: 0.21063584, Global Avg Loss: 0.75731289, Time: 0.0209 Steps: 79010, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000247, Sample Num: 3952, Cur Loss: 0.08548611, Cur Avg Loss: 0.17538054, Log Avg loss: 0.15683709, Global Avg Loss: 0.75723690, Time: 0.0209 Steps: 79020, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000257, Sample Num: 4112, Cur Loss: 0.10369115, Cur Avg Loss: 0.17821353, Log Avg loss: 0.24818861, Global Avg Loss: 0.75717249, Time: 0.0246 Steps: 79030, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000267, Sample Num: 4272, Cur Loss: 0.27853513, Cur Avg Loss: 0.17712850, Log Avg loss: 0.14924325, Global Avg Loss: 0.75709558, Time: 0.0210 Steps: 79040, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000277, Sample Num: 4432, Cur Loss: 0.06852713, Cur Avg Loss: 0.17690688, Log Avg loss: 0.17098956, Global Avg Loss: 0.75702143, Time: 0.0210 Steps: 79050, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000287, Sample Num: 4592, Cur Loss: 0.07183058, Cur Avg Loss: 0.17468163, Log Avg loss: 0.11304220, Global Avg Loss: 0.75693998, Time: 0.0210 Steps: 79060, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000297, Sample Num: 4752, Cur Loss: 0.14221694, Cur Avg Loss: 0.17431041, Log Avg loss: 0.16365636, Global Avg Loss: 0.75686495, Time: 0.0210 Steps: 79070, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000307, Sample Num: 4912, Cur Loss: 0.09526163, Cur Avg Loss: 0.17441989, Log Avg loss: 0.17767137, Global Avg Loss: 0.75679170, Time: 0.0210 Steps: 79080, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000317, Sample Num: 5072, Cur Loss: 0.22235709, Cur Avg Loss: 0.17621415, Log Avg loss: 0.23129805, Global Avg Loss: 0.75672526, Time: 0.0210 Steps: 79090, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000327, Sample Num: 5232, Cur Loss: 0.04088480, Cur Avg Loss: 0.17468492, Log Avg loss: 0.12620834, Global Avg Loss: 0.75664555, Time: 0.0209 Steps: 79100, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000337, Sample Num: 5392, Cur Loss: 0.17319432, Cur Avg Loss: 0.17408675, Log Avg loss: 0.15452659, Global Avg Loss: 0.75656944, Time: 0.0210 Steps: 79110, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000347, Sample Num: 5552, Cur Loss: 0.27280650, Cur Avg Loss: 0.17426415, Log Avg loss: 0.18024241, Global Avg Loss: 0.75649660, Time: 0.0209 Steps: 79120, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000357, Sample Num: 5712, Cur Loss: 0.19657895, Cur Avg Loss: 0.17411169, Log Avg loss: 0.16882150, Global Avg Loss: 0.75642233, Time: 0.0210 Steps: 79130, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000367, Sample Num: 5872, Cur Loss: 0.09944168, Cur Avg Loss: 0.17450636, Log Avg loss: 0.18859607, Global Avg Loss: 0.75635058, Time: 0.0209 Steps: 79140, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000377, Sample Num: 6032, Cur Loss: 0.13056122, Cur Avg Loss: 0.17371520, Log Avg loss: 0.14467960, Global Avg Loss: 0.75627330, Time: 0.0210 Steps: 79150, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000387, Sample Num: 6192, Cur Loss: 0.14580958, Cur Avg Loss: 0.17197564, Log Avg loss: 0.10639412, Global Avg Loss: 0.75619120, Time: 0.0210 Steps: 79160, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000397, Sample Num: 6352, Cur Loss: 0.18664449, Cur Avg Loss: 0.17214189, Log Avg loss: 0.17857578, Global Avg Loss: 0.75611824, Time: 0.0209 Steps: 79170, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000407, Sample Num: 6512, Cur Loss: 0.12003004, Cur Avg Loss: 0.17103490, Log Avg loss: 0.12708747, Global Avg Loss: 0.75603880, Time: 0.0210 Steps: 79180, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000417, Sample Num: 6672, Cur Loss: 0.08076225, Cur Avg Loss: 0.17054876, Log Avg loss: 0.15076273, Global Avg Loss: 0.75596237, Time: 0.0210 Steps: 79190, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000427, Sample Num: 6832, Cur Loss: 0.29306942, Cur Avg Loss: 0.17270576, Log Avg loss: 0.26265282, Global Avg Loss: 0.75590008, Time: 0.0210 Steps: 79200, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000437, Sample Num: 6992, Cur Loss: 0.14155792, Cur Avg Loss: 0.17238140, Log Avg loss: 0.15853096, Global Avg Loss: 0.75582467, Time: 0.0210 Steps: 79210, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000447, Sample Num: 7152, Cur Loss: 0.16850404, Cur Avg Loss: 0.17263104, Log Avg loss: 0.18354043, Global Avg Loss: 0.75575243, Time: 0.0210 Steps: 79220, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000457, Sample Num: 7312, Cur Loss: 0.08170073, Cur Avg Loss: 0.17225062, Log Avg loss: 0.15524598, Global Avg Loss: 0.75567663, Time: 0.0209 Steps: 79230, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000467, Sample Num: 7472, Cur Loss: 0.14625105, Cur Avg Loss: 0.17234184, Log Avg loss: 0.17651072, Global Avg Loss: 0.75560354, Time: 0.0210 Steps: 79240, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000477, Sample Num: 7632, Cur Loss: 0.12745261, Cur Avg Loss: 0.17249967, Log Avg loss: 0.17987005, Global Avg Loss: 0.75553090, Time: 0.0210 Steps: 79250, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000487, Sample Num: 7792, Cur Loss: 0.15033188, Cur Avg Loss: 0.17207082, Log Avg loss: 0.15161472, Global Avg Loss: 0.75545470, Time: 0.0209 Steps: 79260, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000497, Sample Num: 7952, Cur Loss: 0.11382376, Cur Avg Loss: 0.17168370, Log Avg loss: 0.15283086, Global Avg Loss: 0.75537868, Time: 0.0210 Steps: 79270, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000507, Sample Num: 8112, Cur Loss: 0.28864232, Cur Avg Loss: 0.17096156, Log Avg loss: 0.13507155, Global Avg Loss: 0.75530044, Time: 0.0210 Steps: 79280, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000517, Sample Num: 8272, Cur Loss: 0.12345675, Cur Avg Loss: 0.17125540, Log Avg loss: 0.18615285, Global Avg Loss: 0.75522866, Time: 0.0209 Steps: 79290, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000527, Sample Num: 8432, Cur Loss: 0.09376594, Cur Avg Loss: 0.17174242, Log Avg loss: 0.19692112, Global Avg Loss: 0.75515825, Time: 0.0209 Steps: 79300, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000537, Sample Num: 8592, Cur Loss: 0.14220867, Cur Avg Loss: 0.17207903, Log Avg loss: 0.18981844, Global Avg Loss: 0.75508697, Time: 0.0209 Steps: 79310, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000547, Sample Num: 8752, Cur Loss: 0.34560847, Cur Avg Loss: 0.17211250, Log Avg loss: 0.17391002, Global Avg Loss: 0.75501370, Time: 0.0209 Steps: 79320, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000557, Sample Num: 8912, Cur Loss: 0.12682091, Cur Avg Loss: 0.17131911, Log Avg loss: 0.12792092, Global Avg Loss: 0.75493465, Time: 0.0208 Steps: 79330, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000567, Sample Num: 9072, Cur Loss: 0.03062009, Cur Avg Loss: 0.17094644, Log Avg loss: 0.15018832, Global Avg Loss: 0.75485843, Time: 0.0209 Steps: 79340, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000577, Sample Num: 9232, Cur Loss: 0.05054685, Cur Avg Loss: 0.17027546, Log Avg loss: 0.13223094, Global Avg Loss: 0.75477996, Time: 0.0209 Steps: 79350, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000587, Sample Num: 9392, Cur Loss: 0.20668271, Cur Avg Loss: 0.16971815, Log Avg loss: 0.13756169, Global Avg Loss: 0.75470219, Time: 0.0209 Steps: 79360, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000597, Sample Num: 9552, Cur Loss: 0.24155672, Cur Avg Loss: 0.16899137, Log Avg loss: 0.12632910, Global Avg Loss: 0.75462302, Time: 0.0209 Steps: 79370, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000607, Sample Num: 9712, Cur Loss: 0.76194507, Cur Avg Loss: 0.17053726, Log Avg loss: 0.26282666, Global Avg Loss: 0.75456106, Time: 0.0208 Steps: 79380, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000617, Sample Num: 9872, Cur Loss: 0.22995649, Cur Avg Loss: 0.16989762, Log Avg loss: 0.13107152, Global Avg Loss: 0.75448253, Time: 0.0208 Steps: 79390, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000627, Sample Num: 10032, Cur Loss: 0.20100109, Cur Avg Loss: 0.17014687, Log Avg loss: 0.18552583, Global Avg Loss: 0.75441087, Time: 0.0208 Steps: 79400, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000637, Sample Num: 10192, Cur Loss: 0.04401147, Cur Avg Loss: 0.16951035, Log Avg loss: 0.12960048, Global Avg Loss: 0.75433219, Time: 0.0208 Steps: 79410, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000647, Sample Num: 10352, Cur Loss: 0.03380259, Cur Avg Loss: 0.16875311, Log Avg loss: 0.12051714, Global Avg Loss: 0.75425238, Time: 0.0209 Steps: 79420, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000657, Sample Num: 10512, Cur Loss: 0.18679933, Cur Avg Loss: 0.16817567, Log Avg loss: 0.13081492, Global Avg Loss: 0.75417390, Time: 0.0209 Steps: 79430, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000667, Sample Num: 10672, Cur Loss: 0.06826490, Cur Avg Loss: 0.16903927, Log Avg loss: 0.22577806, Global Avg Loss: 0.75410738, Time: 0.0208 Steps: 79440, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000677, Sample Num: 10832, Cur Loss: 0.04540992, Cur Avg Loss: 0.16956978, Log Avg loss: 0.20495460, Global Avg Loss: 0.75403826, Time: 0.0209 Steps: 79450, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000687, Sample Num: 10992, Cur Loss: 0.08344609, Cur Avg Loss: 0.16871254, Log Avg loss: 0.11067746, Global Avg Loss: 0.75395729, Time: 0.0208 Steps: 79460, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000697, Sample Num: 11152, Cur Loss: 0.30250868, Cur Avg Loss: 0.16922653, Log Avg loss: 0.20453798, Global Avg Loss: 0.75388816, Time: 0.0209 Steps: 79470, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000707, Sample Num: 11312, Cur Loss: 0.20088322, Cur Avg Loss: 0.16897240, Log Avg loss: 0.15125930, Global Avg Loss: 0.75381234, Time: 0.0209 Steps: 79480, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000717, Sample Num: 11472, Cur Loss: 0.17681287, Cur Avg Loss: 0.16875096, Log Avg loss: 0.15309489, Global Avg Loss: 0.75373677, Time: 0.0209 Steps: 79490, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000727, Sample Num: 11632, Cur Loss: 0.18884736, Cur Avg Loss: 0.16882847, Log Avg loss: 0.17438623, Global Avg Loss: 0.75366389, Time: 0.0208 Steps: 79500, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000737, Sample Num: 11792, Cur Loss: 0.18272209, Cur Avg Loss: 0.16873462, Log Avg loss: 0.16191161, Global Avg Loss: 0.75358947, Time: 0.0209 Steps: 79510, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000747, Sample Num: 11952, Cur Loss: 0.17833385, Cur Avg Loss: 0.16913791, Log Avg loss: 0.19886060, Global Avg Loss: 0.75351971, Time: 0.0208 Steps: 79520, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000757, Sample Num: 12112, Cur Loss: 0.15343997, Cur Avg Loss: 0.16874541, Log Avg loss: 0.13942586, Global Avg Loss: 0.75344249, Time: 0.0209 Steps: 79530, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000767, Sample Num: 12272, Cur Loss: 0.13729921, Cur Avg Loss: 0.16941447, Log Avg loss: 0.22006224, Global Avg Loss: 0.75337543, Time: 0.0208 Steps: 79540, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000777, Sample Num: 12432, Cur Loss: 0.12874380, Cur Avg Loss: 0.16925512, Log Avg loss: 0.15703286, Global Avg Loss: 0.75330047, Time: 0.0211 Steps: 79550, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000787, Sample Num: 12592, Cur Loss: 0.02349108, Cur Avg Loss: 0.16927440, Log Avg loss: 0.17077203, Global Avg Loss: 0.75322725, Time: 0.0209 Steps: 79560, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000797, Sample Num: 12752, Cur Loss: 0.49575269, Cur Avg Loss: 0.17023462, Log Avg loss: 0.24580402, Global Avg Loss: 0.75316348, Time: 0.0209 Steps: 79570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000807, Sample Num: 12912, Cur Loss: 0.22693858, Cur Avg Loss: 0.17024725, Log Avg loss: 0.17125400, Global Avg Loss: 0.75309036, Time: 0.0209 Steps: 79580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000817, Sample Num: 13072, Cur Loss: 0.16386761, Cur Avg Loss: 0.16991852, Log Avg loss: 0.14338984, Global Avg Loss: 0.75301375, Time: 0.0209 Steps: 79590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000827, Sample Num: 13232, Cur Loss: 0.15904172, Cur Avg Loss: 0.17070703, Log Avg loss: 0.23512830, Global Avg Loss: 0.75294869, Time: 0.0208 Steps: 79600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000837, Sample Num: 13392, Cur Loss: 0.16693467, Cur Avg Loss: 0.17052552, Log Avg loss: 0.15551464, Global Avg Loss: 0.75287365, Time: 0.0209 Steps: 79610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000847, Sample Num: 13552, Cur Loss: 0.08103552, Cur Avg Loss: 0.17047601, Log Avg loss: 0.16633221, Global Avg Loss: 0.75279998, Time: 0.0209 Steps: 79620, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000857, Sample Num: 13712, Cur Loss: 0.15960824, Cur Avg Loss: 0.17068668, Log Avg loss: 0.18853049, Global Avg Loss: 0.75272912, Time: 0.0209 Steps: 79630, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000867, Sample Num: 13872, Cur Loss: 0.22009632, Cur Avg Loss: 0.17052818, Log Avg loss: 0.15694464, Global Avg Loss: 0.75265431, Time: 0.0209 Steps: 79640, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000877, Sample Num: 14032, Cur Loss: 0.20983551, Cur Avg Loss: 0.17064777, Log Avg loss: 0.18101650, Global Avg Loss: 0.75258254, Time: 0.0209 Steps: 79650, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000887, Sample Num: 14192, Cur Loss: 0.27976015, Cur Avg Loss: 0.17022773, Log Avg loss: 0.13339006, Global Avg Loss: 0.75250481, Time: 0.0209 Steps: 79660, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000897, Sample Num: 14352, Cur Loss: 0.08937410, Cur Avg Loss: 0.16953530, Log Avg loss: 0.10811668, Global Avg Loss: 0.75242393, Time: 0.0209 Steps: 79670, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000907, Sample Num: 14512, Cur Loss: 0.11139050, Cur Avg Loss: 0.16959796, Log Avg loss: 0.17521901, Global Avg Loss: 0.75235149, Time: 0.0210 Steps: 79680, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000917, Sample Num: 14672, Cur Loss: 0.11600064, Cur Avg Loss: 0.16961196, Log Avg loss: 0.17088163, Global Avg Loss: 0.75227852, Time: 0.0209 Steps: 79690, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000927, Sample Num: 14832, Cur Loss: 0.38446912, Cur Avg Loss: 0.17008691, Log Avg loss: 0.21363914, Global Avg Loss: 0.75221094, Time: 0.0209 Steps: 79700, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000937, Sample Num: 14992, Cur Loss: 0.17515633, Cur Avg Loss: 0.17049661, Log Avg loss: 0.20847647, Global Avg Loss: 0.75214272, Time: 0.0209 Steps: 79710, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000947, Sample Num: 15152, Cur Loss: 0.08140020, Cur Avg Loss: 0.17022537, Log Avg loss: 0.14481015, Global Avg Loss: 0.75206654, Time: 0.0210 Steps: 79720, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000957, Sample Num: 15312, Cur Loss: 0.08826436, Cur Avg Loss: 0.17050943, Log Avg loss: 0.19740987, Global Avg Loss: 0.75199697, Time: 0.0209 Steps: 79730, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000967, Sample Num: 15472, Cur Loss: 0.17290905, Cur Avg Loss: 0.17065400, Log Avg loss: 0.18448941, Global Avg Loss: 0.75192580, Time: 0.0208 Steps: 79740, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000977, Sample Num: 15632, Cur Loss: 0.10120335, Cur Avg Loss: 0.17043966, Log Avg loss: 0.14971260, Global Avg Loss: 0.75185029, Time: 0.0209 Steps: 79750, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000987, Sample Num: 15792, Cur Loss: 0.13540849, Cur Avg Loss: 0.17031074, Log Avg loss: 0.15771545, Global Avg Loss: 0.75177580, Time: 0.0208 Steps: 79760, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000997, Sample Num: 15952, Cur Loss: 0.28993720, Cur Avg Loss: 0.17052738, Log Avg loss: 0.19191008, Global Avg Loss: 0.75170562, Time: 0.0208 Steps: 79770, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001007, Sample Num: 16112, Cur Loss: 0.64983046, Cur Avg Loss: 0.17052602, Log Avg loss: 0.17039015, Global Avg Loss: 0.75163275, Time: 0.0208 Steps: 79780, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001017, Sample Num: 16272, Cur Loss: 0.22834247, Cur Avg Loss: 0.17064637, Log Avg loss: 0.18276561, Global Avg Loss: 0.75156145, Time: 0.0209 Steps: 79790, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001027, Sample Num: 16432, Cur Loss: 0.05967376, Cur Avg Loss: 0.17045750, Log Avg loss: 0.15124962, Global Avg Loss: 0.75148623, Time: 0.0245 Steps: 79800, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001037, Sample Num: 16592, Cur Loss: 0.30147541, Cur Avg Loss: 0.17024588, Log Avg loss: 0.14851260, Global Avg Loss: 0.75141068, Time: 0.0209 Steps: 79810, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001047, Sample Num: 16752, Cur Loss: 0.13488732, Cur Avg Loss: 0.17042877, Log Avg loss: 0.18939450, Global Avg Loss: 0.75134027, Time: 0.0208 Steps: 79820, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001057, Sample Num: 16912, Cur Loss: 0.26726741, Cur Avg Loss: 0.17034621, Log Avg loss: 0.16170134, Global Avg Loss: 0.75126640, Time: 0.0208 Steps: 79830, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001067, Sample Num: 17072, Cur Loss: 0.12113358, Cur Avg Loss: 0.17034456, Log Avg loss: 0.17017005, Global Avg Loss: 0.75119362, Time: 0.0208 Steps: 79840, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001077, Sample Num: 17232, Cur Loss: 0.11364454, Cur Avg Loss: 0.17089180, Log Avg loss: 0.22928287, Global Avg Loss: 0.75112826, Time: 0.0209 Steps: 79850, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001087, Sample Num: 17392, Cur Loss: 0.21447168, Cur Avg Loss: 0.17105404, Log Avg loss: 0.18852691, Global Avg Loss: 0.75105781, Time: 0.0208 Steps: 79860, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001097, Sample Num: 17552, Cur Loss: 0.22384590, Cur Avg Loss: 0.17103093, Log Avg loss: 0.16851938, Global Avg Loss: 0.75098488, Time: 0.0208 Steps: 79870, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001107, Sample Num: 17712, Cur Loss: 0.07081858, Cur Avg Loss: 0.17078309, Log Avg loss: 0.14359452, Global Avg Loss: 0.75090884, Time: 0.0209 Steps: 79880, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001117, Sample Num: 17872, Cur Loss: 0.05616906, Cur Avg Loss: 0.17084225, Log Avg loss: 0.17739197, Global Avg Loss: 0.75083705, Time: 0.0209 Steps: 79890, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001127, Sample Num: 18032, Cur Loss: 0.11271016, Cur Avg Loss: 0.17071060, Log Avg loss: 0.15600493, Global Avg Loss: 0.75076260, Time: 0.0208 Steps: 79900, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001137, Sample Num: 18192, Cur Loss: 0.08421077, Cur Avg Loss: 0.17043526, Log Avg loss: 0.13940479, Global Avg Loss: 0.75068610, Time: 0.0208 Steps: 79910, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001147, Sample Num: 18352, Cur Loss: 0.10401153, Cur Avg Loss: 0.17042628, Log Avg loss: 0.16940475, Global Avg Loss: 0.75061336, Time: 0.0209 Steps: 79920, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001157, Sample Num: 18512, Cur Loss: 0.14256601, Cur Avg Loss: 0.17015855, Log Avg loss: 0.13944993, Global Avg Loss: 0.75053690, Time: 0.0209 Steps: 79930, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001167, Sample Num: 18672, Cur Loss: 0.24156208, Cur Avg Loss: 0.17013097, Log Avg loss: 0.16693982, Global Avg Loss: 0.75046390, Time: 0.0208 Steps: 79940, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001177, Sample Num: 18832, Cur Loss: 0.38907176, Cur Avg Loss: 0.17069672, Log Avg loss: 0.23672017, Global Avg Loss: 0.75039964, Time: 0.0209 Steps: 79950, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001187, Sample Num: 18992, Cur Loss: 0.09776086, Cur Avg Loss: 0.17083724, Log Avg loss: 0.18737660, Global Avg Loss: 0.75032923, Time: 0.0208 Steps: 79960, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001197, Sample Num: 19152, Cur Loss: 0.17336924, Cur Avg Loss: 0.17085185, Log Avg loss: 0.17258532, Global Avg Loss: 0.75025698, Time: 0.0209 Steps: 79970, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001207, Sample Num: 19312, Cur Loss: 0.18749440, Cur Avg Loss: 0.17111480, Log Avg loss: 0.20259044, Global Avg Loss: 0.75018851, Time: 0.0209 Steps: 79980, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001217, Sample Num: 19472, Cur Loss: 0.13276272, Cur Avg Loss: 0.17152772, Log Avg loss: 0.22136720, Global Avg Loss: 0.75012239, Time: 0.0208 Steps: 79990, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001227, Sample Num: 19632, Cur Loss: 0.13038781, Cur Avg Loss: 0.17151798, Log Avg loss: 0.17033290, Global Avg Loss: 0.75004992, Time: 0.0208 Steps: 80000, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001237, Sample Num: 19792, Cur Loss: 0.30768660, Cur Avg Loss: 0.17141536, Log Avg loss: 0.15882308, Global Avg Loss: 0.74997603, Time: 0.0208 Steps: 80010, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001247, Sample Num: 19952, Cur Loss: 0.17262852, Cur Avg Loss: 0.17145885, Log Avg loss: 0.17683886, Global Avg Loss: 0.74990440, Time: 0.0208 Steps: 80020, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001257, Sample Num: 20112, Cur Loss: 0.90659082, Cur Avg Loss: 0.17218782, Log Avg loss: 0.26309062, Global Avg Loss: 0.74984357, Time: 0.0208 Steps: 80030, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001267, Sample Num: 20272, Cur Loss: 0.19327706, Cur Avg Loss: 0.17209688, Log Avg loss: 0.16066577, Global Avg Loss: 0.74976996, Time: 0.0209 Steps: 80040, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001277, Sample Num: 20432, Cur Loss: 0.06836223, Cur Avg Loss: 0.17219488, Log Avg loss: 0.18461155, Global Avg Loss: 0.74969936, Time: 0.0208 Steps: 80050, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001287, Sample Num: 20592, Cur Loss: 0.10661998, Cur Avg Loss: 0.17232602, Log Avg loss: 0.18907239, Global Avg Loss: 0.74962934, Time: 0.0209 Steps: 80060, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001297, Sample Num: 20752, Cur Loss: 0.30366531, Cur Avg Loss: 0.17268962, Log Avg loss: 0.21948506, Global Avg Loss: 0.74956313, Time: 0.0209 Steps: 80070, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001307, Sample Num: 20912, Cur Loss: 0.23499811, Cur Avg Loss: 0.17300129, Log Avg loss: 0.21342448, Global Avg Loss: 0.74949618, Time: 0.0209 Steps: 80080, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001317, Sample Num: 21072, Cur Loss: 0.19968455, Cur Avg Loss: 0.17305298, Log Avg loss: 0.17980948, Global Avg Loss: 0.74942505, Time: 0.0209 Steps: 80090, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001327, Sample Num: 21232, Cur Loss: 0.20810841, Cur Avg Loss: 0.17328028, Log Avg loss: 0.20321503, Global Avg Loss: 0.74935685, Time: 0.0209 Steps: 80100, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001337, Sample Num: 21392, Cur Loss: 0.21848318, Cur Avg Loss: 0.17330809, Log Avg loss: 0.17699813, Global Avg Loss: 0.74928541, Time: 0.0209 Steps: 80110, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001347, Sample Num: 21552, Cur Loss: 0.08623161, Cur Avg Loss: 0.17359683, Log Avg loss: 0.21220191, Global Avg Loss: 0.74921837, Time: 0.0209 Steps: 80120, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001357, Sample Num: 21712, Cur Loss: 0.41159552, Cur Avg Loss: 0.17355827, Log Avg loss: 0.16836391, Global Avg Loss: 0.74914588, Time: 0.0209 Steps: 80130, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001367, Sample Num: 21872, Cur Loss: 0.14666492, Cur Avg Loss: 0.17357902, Log Avg loss: 0.17639492, Global Avg Loss: 0.74907442, Time: 0.0209 Steps: 80140, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001377, Sample Num: 22032, Cur Loss: 0.13859631, Cur Avg Loss: 0.17358987, Log Avg loss: 0.17507296, Global Avg Loss: 0.74900280, Time: 0.0209 Steps: 80150, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001387, Sample Num: 22192, Cur Loss: 0.10082654, Cur Avg Loss: 0.17351875, Log Avg loss: 0.16372547, Global Avg Loss: 0.74892979, Time: 0.0209 Steps: 80160, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001397, Sample Num: 22352, Cur Loss: 0.08135502, Cur Avg Loss: 0.17313417, Log Avg loss: 0.11979281, Global Avg Loss: 0.74885131, Time: 0.0209 Steps: 80170, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001407, Sample Num: 22512, Cur Loss: 0.55182749, Cur Avg Loss: 0.17295262, Log Avg loss: 0.14759033, Global Avg Loss: 0.74877632, Time: 0.0210 Steps: 80180, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001417, Sample Num: 22672, Cur Loss: 0.26473957, Cur Avg Loss: 0.17286692, Log Avg loss: 0.16080992, Global Avg Loss: 0.74870300, Time: 0.0209 Steps: 80190, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001427, Sample Num: 22832, Cur Loss: 0.22032897, Cur Avg Loss: 0.17276595, Log Avg loss: 0.15845814, Global Avg Loss: 0.74862940, Time: 0.0210 Steps: 80200, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001437, Sample Num: 22992, Cur Loss: 0.37627321, Cur Avg Loss: 0.17278057, Log Avg loss: 0.17486631, Global Avg Loss: 0.74855787, Time: 0.0209 Steps: 80210, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001447, Sample Num: 23152, Cur Loss: 0.11489665, Cur Avg Loss: 0.17324589, Log Avg loss: 0.24011312, Global Avg Loss: 0.74849449, Time: 0.0210 Steps: 80220, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001457, Sample Num: 23312, Cur Loss: 0.14280286, Cur Avg Loss: 0.17332096, Log Avg loss: 0.18418363, Global Avg Loss: 0.74842415, Time: 0.0209 Steps: 80230, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001467, Sample Num: 23472, Cur Loss: 0.40359643, Cur Avg Loss: 0.17312001, Log Avg loss: 0.14384090, Global Avg Loss: 0.74834881, Time: 0.0209 Steps: 80240, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001477, Sample Num: 23632, Cur Loss: 0.10167028, Cur Avg Loss: 0.17323873, Log Avg loss: 0.19065463, Global Avg Loss: 0.74827931, Time: 0.0209 Steps: 80250, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001487, Sample Num: 23792, Cur Loss: 0.10380358, Cur Avg Loss: 0.17382861, Log Avg loss: 0.26095423, Global Avg Loss: 0.74821859, Time: 0.0209 Steps: 80260, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001497, Sample Num: 23952, Cur Loss: 0.04614805, Cur Avg Loss: 0.17361474, Log Avg loss: 0.14181186, Global Avg Loss: 0.74814305, Time: 0.0209 Steps: 80270, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001507, Sample Num: 24112, Cur Loss: 0.12208620, Cur Avg Loss: 0.17355900, Log Avg loss: 0.16521489, Global Avg Loss: 0.74807044, Time: 0.0209 Steps: 80280, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001517, Sample Num: 24272, Cur Loss: 0.12162802, Cur Avg Loss: 0.17320466, Log Avg loss: 0.11980673, Global Avg Loss: 0.74799219, Time: 0.0209 Steps: 80290, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001527, Sample Num: 24432, Cur Loss: 0.25443804, Cur Avg Loss: 0.17300637, Log Avg loss: 0.14292510, Global Avg Loss: 0.74791683, Time: 0.0209 Steps: 80300, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001537, Sample Num: 24592, Cur Loss: 0.12498532, Cur Avg Loss: 0.17274971, Log Avg loss: 0.13355731, Global Avg Loss: 0.74784034, Time: 0.0246 Steps: 80310, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001547, Sample Num: 24752, Cur Loss: 0.08135038, Cur Avg Loss: 0.17309213, Log Avg loss: 0.22572182, Global Avg Loss: 0.74777533, Time: 0.0209 Steps: 80320, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001557, Sample Num: 24912, Cur Loss: 0.12434597, Cur Avg Loss: 0.17327599, Log Avg loss: 0.20171969, Global Avg Loss: 0.74770736, Time: 0.0209 Steps: 80330, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001567, Sample Num: 25072, Cur Loss: 0.32315952, Cur Avg Loss: 0.17337136, Log Avg loss: 0.18822111, Global Avg Loss: 0.74763772, Time: 0.0210 Steps: 80340, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001577, Sample Num: 25232, Cur Loss: 0.15142801, Cur Avg Loss: 0.17311338, Log Avg loss: 0.13268739, Global Avg Loss: 0.74756118, Time: 0.0209 Steps: 80350, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001587, Sample Num: 25392, Cur Loss: 0.06104212, Cur Avg Loss: 0.17294627, Log Avg loss: 0.14659362, Global Avg Loss: 0.74748640, Time: 0.0209 Steps: 80360, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001597, Sample Num: 25552, Cur Loss: 0.20232916, Cur Avg Loss: 0.17299744, Log Avg loss: 0.18111774, Global Avg Loss: 0.74741593, Time: 0.0209 Steps: 80370, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001607, Sample Num: 25712, Cur Loss: 0.05261714, Cur Avg Loss: 0.17289208, Log Avg loss: 0.15606625, Global Avg Loss: 0.74734236, Time: 0.0209 Steps: 80380, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001617, Sample Num: 25872, Cur Loss: 0.30644989, Cur Avg Loss: 0.17287133, Log Avg loss: 0.16953609, Global Avg Loss: 0.74727048, Time: 0.0209 Steps: 80390, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001627, Sample Num: 26032, Cur Loss: 0.26425529, Cur Avg Loss: 0.17288893, Log Avg loss: 0.17573506, Global Avg Loss: 0.74719940, Time: 0.0209 Steps: 80400, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001637, Sample Num: 26192, Cur Loss: 0.15741378, Cur Avg Loss: 0.17336345, Log Avg loss: 0.25056843, Global Avg Loss: 0.74713763, Time: 0.0209 Steps: 80410, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001647, Sample Num: 26352, Cur Loss: 0.43870115, Cur Avg Loss: 0.17313309, Log Avg loss: 0.13542315, Global Avg Loss: 0.74706157, Time: 0.0209 Steps: 80420, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001657, Sample Num: 26512, Cur Loss: 0.16430287, Cur Avg Loss: 0.17315818, Log Avg loss: 0.17729074, Global Avg Loss: 0.74699073, Time: 0.0209 Steps: 80430, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001667, Sample Num: 26672, Cur Loss: 0.14217636, Cur Avg Loss: 0.17318618, Log Avg loss: 0.17782566, Global Avg Loss: 0.74691997, Time: 0.0209 Steps: 80440, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001677, Sample Num: 26832, Cur Loss: 0.36434728, Cur Avg Loss: 0.17313185, Log Avg loss: 0.16407367, Global Avg Loss: 0.74684752, Time: 0.0209 Steps: 80450, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001687, Sample Num: 26992, Cur Loss: 0.37513584, Cur Avg Loss: 0.17313080, Log Avg loss: 0.17295570, Global Avg Loss: 0.74677620, Time: 0.0209 Steps: 80460, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001697, Sample Num: 27152, Cur Loss: 0.16990450, Cur Avg Loss: 0.17297802, Log Avg loss: 0.14720367, Global Avg Loss: 0.74670169, Time: 0.0209 Steps: 80470, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001707, Sample Num: 27312, Cur Loss: 0.11338566, Cur Avg Loss: 0.17288125, Log Avg loss: 0.15646002, Global Avg Loss: 0.74662835, Time: 0.0209 Steps: 80480, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001717, Sample Num: 27472, Cur Loss: 0.06476405, Cur Avg Loss: 0.17256888, Log Avg loss: 0.11924608, Global Avg Loss: 0.74655040, Time: 0.0209 Steps: 80490, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001727, Sample Num: 27632, Cur Loss: 0.03988420, Cur Avg Loss: 0.17254841, Log Avg loss: 0.16903522, Global Avg Loss: 0.74647866, Time: 0.0209 Steps: 80500, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001737, Sample Num: 27792, Cur Loss: 0.18745343, Cur Avg Loss: 0.17241030, Log Avg loss: 0.14855729, Global Avg Loss: 0.74640439, Time: 0.0209 Steps: 80510, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001747, Sample Num: 27952, Cur Loss: 0.06038943, Cur Avg Loss: 0.17235671, Log Avg loss: 0.16304963, Global Avg Loss: 0.74633195, Time: 0.0209 Steps: 80520, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001757, Sample Num: 28112, Cur Loss: 0.03455274, Cur Avg Loss: 0.17219258, Log Avg loss: 0.14351767, Global Avg Loss: 0.74625709, Time: 0.0209 Steps: 80530, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001767, Sample Num: 28272, Cur Loss: 0.27668688, Cur Avg Loss: 0.17277894, Log Avg loss: 0.27580337, Global Avg Loss: 0.74619868, Time: 0.0209 Steps: 80540, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001777, Sample Num: 28432, Cur Loss: 0.10471197, Cur Avg Loss: 0.17273242, Log Avg loss: 0.16451173, Global Avg Loss: 0.74612646, Time: 0.0209 Steps: 80550, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001787, Sample Num: 28592, Cur Loss: 0.15810825, Cur Avg Loss: 0.17263274, Log Avg loss: 0.15491999, Global Avg Loss: 0.74605308, Time: 0.0209 Steps: 80560, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001797, Sample Num: 28752, Cur Loss: 0.42945206, Cur Avg Loss: 0.17288622, Log Avg loss: 0.21818346, Global Avg Loss: 0.74598756, Time: 0.0210 Steps: 80570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001807, Sample Num: 28912, Cur Loss: 0.07675346, Cur Avg Loss: 0.17285696, Log Avg loss: 0.16759845, Global Avg Loss: 0.74591578, Time: 0.0209 Steps: 80580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001817, Sample Num: 29072, Cur Loss: 0.11648633, Cur Avg Loss: 0.17327105, Log Avg loss: 0.24809685, Global Avg Loss: 0.74585401, Time: 0.0209 Steps: 80590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001827, Sample Num: 29232, Cur Loss: 0.15077049, Cur Avg Loss: 0.17334005, Log Avg loss: 0.18587748, Global Avg Loss: 0.74578453, Time: 0.0209 Steps: 80600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001837, Sample Num: 29392, Cur Loss: 0.19582802, Cur Avg Loss: 0.17332371, Log Avg loss: 0.17033876, Global Avg Loss: 0.74571315, Time: 0.0209 Steps: 80610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001847, Sample Num: 29552, Cur Loss: 0.22937082, Cur Avg Loss: 0.17342058, Log Avg loss: 0.19121515, Global Avg Loss: 0.74564437, Time: 0.0209 Steps: 80620, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001857, Sample Num: 29712, Cur Loss: 0.07861122, Cur Avg Loss: 0.17322845, Log Avg loss: 0.13774306, Global Avg Loss: 0.74556897, Time: 0.0209 Steps: 80630, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001867, Sample Num: 29872, Cur Loss: 0.15996003, Cur Avg Loss: 0.17347204, Log Avg loss: 0.21870520, Global Avg Loss: 0.74550364, Time: 0.0209 Steps: 80640, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001877, Sample Num: 30032, Cur Loss: 0.12565152, Cur Avg Loss: 0.17353497, Log Avg loss: 0.18528541, Global Avg Loss: 0.74543418, Time: 0.0209 Steps: 80650, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001887, Sample Num: 30192, Cur Loss: 0.35301560, Cur Avg Loss: 0.17367053, Log Avg loss: 0.19911440, Global Avg Loss: 0.74536644, Time: 0.0208 Steps: 80660, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001897, Sample Num: 30352, Cur Loss: 0.29873699, Cur Avg Loss: 0.17407626, Log Avg loss: 0.25063735, Global Avg Loss: 0.74530512, Time: 0.0209 Steps: 80670, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001907, Sample Num: 30512, Cur Loss: 0.13374479, Cur Avg Loss: 0.17406265, Log Avg loss: 0.17148056, Global Avg Loss: 0.74523399, Time: 0.0210 Steps: 80680, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001917, Sample Num: 30672, Cur Loss: 0.07042200, Cur Avg Loss: 0.17423279, Log Avg loss: 0.20667937, Global Avg Loss: 0.74516725, Time: 0.0209 Steps: 80690, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001927, Sample Num: 30832, Cur Loss: 0.12579696, Cur Avg Loss: 0.17391730, Log Avg loss: 0.11343760, Global Avg Loss: 0.74508897, Time: 0.0209 Steps: 80700, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001937, Sample Num: 30992, Cur Loss: 0.22659175, Cur Avg Loss: 0.17377193, Log Avg loss: 0.14575894, Global Avg Loss: 0.74501471, Time: 0.0209 Steps: 80710, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001947, Sample Num: 31152, Cur Loss: 0.20013942, Cur Avg Loss: 0.17373815, Log Avg loss: 0.16719562, Global Avg Loss: 0.74494313, Time: 0.0209 Steps: 80720, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001957, Sample Num: 31312, Cur Loss: 0.07128660, Cur Avg Loss: 0.17372906, Log Avg loss: 0.17195900, Global Avg Loss: 0.74487215, Time: 0.0208 Steps: 80730, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001967, Sample Num: 31472, Cur Loss: 0.09086858, Cur Avg Loss: 0.17387860, Log Avg loss: 0.20314368, Global Avg Loss: 0.74480506, Time: 0.0209 Steps: 80740, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001977, Sample Num: 31632, Cur Loss: 0.29530329, Cur Avg Loss: 0.17385904, Log Avg loss: 0.17001146, Global Avg Loss: 0.74473388, Time: 0.0209 Steps: 80750, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001987, Sample Num: 31792, Cur Loss: 0.08015485, Cur Avg Loss: 0.17395475, Log Avg loss: 0.19287568, Global Avg Loss: 0.74466554, Time: 0.0208 Steps: 80760, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001997, Sample Num: 31952, Cur Loss: 0.06599692, Cur Avg Loss: 0.17433569, Log Avg loss: 0.25002881, Global Avg Loss: 0.74460430, Time: 0.0209 Steps: 80770, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002007, Sample Num: 32112, Cur Loss: 0.16621372, Cur Avg Loss: 0.17436804, Log Avg loss: 0.18082854, Global Avg Loss: 0.74453451, Time: 0.0209 Steps: 80780, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002017, Sample Num: 32272, Cur Loss: 0.13837722, Cur Avg Loss: 0.17406185, Log Avg loss: 0.11261003, Global Avg Loss: 0.74445629, Time: 0.0209 Steps: 80790, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002027, Sample Num: 32432, Cur Loss: 0.40607718, Cur Avg Loss: 0.17419217, Log Avg loss: 0.20047707, Global Avg Loss: 0.74438897, Time: 0.0209 Steps: 80800, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002037, Sample Num: 32592, Cur Loss: 0.13374890, Cur Avg Loss: 0.17414031, Log Avg loss: 0.16362892, Global Avg Loss: 0.74431710, Time: 0.0209 Steps: 80810, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002047, Sample Num: 32752, Cur Loss: 0.18672641, Cur Avg Loss: 0.17426867, Log Avg loss: 0.20041502, Global Avg Loss: 0.74424980, Time: 0.0209 Steps: 80820, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002057, Sample Num: 32912, Cur Loss: 0.14589527, Cur Avg Loss: 0.17439986, Log Avg loss: 0.20125382, Global Avg Loss: 0.74418263, Time: 0.0209 Steps: 80830, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002067, Sample Num: 33072, Cur Loss: 0.31522259, Cur Avg Loss: 0.17439815, Log Avg loss: 0.17404753, Global Avg Loss: 0.74411210, Time: 0.0208 Steps: 80840, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002077, Sample Num: 33232, Cur Loss: 0.11482064, Cur Avg Loss: 0.17415563, Log Avg loss: 0.12402721, Global Avg Loss: 0.74403540, Time: 0.0209 Steps: 80850, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002087, Sample Num: 33392, Cur Loss: 0.28628305, Cur Avg Loss: 0.17391500, Log Avg loss: 0.12393455, Global Avg Loss: 0.74395872, Time: 0.0208 Steps: 80860, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002097, Sample Num: 33552, Cur Loss: 0.32304215, Cur Avg Loss: 0.17394101, Log Avg loss: 0.17937094, Global Avg Loss: 0.74388890, Time: 0.0208 Steps: 80870, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002107, Sample Num: 33712, Cur Loss: 0.06986971, Cur Avg Loss: 0.17371665, Log Avg loss: 0.12666823, Global Avg Loss: 0.74381259, Time: 0.0208 Steps: 80880, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002117, Sample Num: 33872, Cur Loss: 0.20465802, Cur Avg Loss: 0.17355242, Log Avg loss: 0.13894747, Global Avg Loss: 0.74373781, Time: 0.0209 Steps: 80890, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002127, Sample Num: 34032, Cur Loss: 0.12589976, Cur Avg Loss: 0.17364273, Log Avg loss: 0.19276116, Global Avg Loss: 0.74366971, Time: 0.0209 Steps: 80900, Updated lr: 0.000024 ***** Running evaluation checkpoint-80902 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-80902 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.695983, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.247005, "eval_total_loss": 173.644811, "eval_mae": 0.336486, "eval_mse": 0.247099, "eval_r2": 0.842928, "eval_sp_statistic": 0.894833, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.919549, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.683719, "test_total_loss": 343.22703, "test_mae": 0.621272, "test_mse": 0.683803, "test_r2": 0.558668, "test_sp_statistic": 0.794841, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.831326, "test_ps_pvalue": 0.0, "lr": 2.4227596017069703e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7436580346224936, "train_cur_epoch_loss": 369.88119431212544, "train_cur_epoch_avg_loss": 0.17373470846036893, "train_cur_epoch_time": 44.69598317146301, "train_cur_epoch_avg_time": 0.020993885942443877, "epoch": 38, "step": 80902} ################################################## Training, Epoch: 0039, Batch: 000008, Sample Num: 128, Cur Loss: 0.11922006, Cur Avg Loss: 0.19069506, Log Avg loss: 0.20686781, Global Avg Loss: 0.74360336, Time: 0.0226 Steps: 80910, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000018, Sample Num: 288, Cur Loss: 0.12216961, Cur Avg Loss: 0.15248660, Log Avg loss: 0.12191984, Global Avg Loss: 0.74352653, Time: 0.0210 Steps: 80920, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000028, Sample Num: 448, Cur Loss: 0.10292799, Cur Avg Loss: 0.15526315, Log Avg loss: 0.16026093, Global Avg Loss: 0.74345446, Time: 0.0208 Steps: 80930, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000038, Sample Num: 608, Cur Loss: 0.23199534, Cur Avg Loss: 0.14755497, Log Avg loss: 0.12597208, Global Avg Loss: 0.74337817, Time: 0.0209 Steps: 80940, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000048, Sample Num: 768, Cur Loss: 0.04450748, Cur Avg Loss: 0.15132235, Log Avg loss: 0.16563838, Global Avg Loss: 0.74330680, Time: 0.0209 Steps: 80950, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000058, Sample Num: 928, Cur Loss: 0.26002878, Cur Avg Loss: 0.14682751, Log Avg loss: 0.12525226, Global Avg Loss: 0.74323046, Time: 0.0210 Steps: 80960, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000068, Sample Num: 1088, Cur Loss: 0.09027448, Cur Avg Loss: 0.14869401, Log Avg loss: 0.15951976, Global Avg Loss: 0.74315837, Time: 0.0209 Steps: 80970, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000078, Sample Num: 1248, Cur Loss: 0.17064229, Cur Avg Loss: 0.14800705, Log Avg loss: 0.14333573, Global Avg Loss: 0.74308430, Time: 0.0209 Steps: 80980, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000088, Sample Num: 1408, Cur Loss: 0.51538289, Cur Avg Loss: 0.15098978, Log Avg loss: 0.17425500, Global Avg Loss: 0.74301407, Time: 0.0210 Steps: 80990, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000098, Sample Num: 1568, Cur Loss: 0.09861255, Cur Avg Loss: 0.15072284, Log Avg loss: 0.14837381, Global Avg Loss: 0.74294066, Time: 0.0209 Steps: 81000, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000108, Sample Num: 1728, Cur Loss: 0.33116636, Cur Avg Loss: 0.15060817, Log Avg loss: 0.14948439, Global Avg Loss: 0.74286740, Time: 0.0209 Steps: 81010, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000118, Sample Num: 1888, Cur Loss: 0.06066027, Cur Avg Loss: 0.15108355, Log Avg loss: 0.15621771, Global Avg Loss: 0.74279499, Time: 0.0209 Steps: 81020, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000128, Sample Num: 2048, Cur Loss: 0.26618427, Cur Avg Loss: 0.15164635, Log Avg loss: 0.15828734, Global Avg Loss: 0.74272286, Time: 0.0209 Steps: 81030, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000138, Sample Num: 2208, Cur Loss: 0.04866471, Cur Avg Loss: 0.15174149, Log Avg loss: 0.15295931, Global Avg Loss: 0.74265008, Time: 0.0210 Steps: 81040, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000148, Sample Num: 2368, Cur Loss: 0.05731904, Cur Avg Loss: 0.15538057, Log Avg loss: 0.20559988, Global Avg Loss: 0.74258382, Time: 0.0209 Steps: 81050, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000158, Sample Num: 2528, Cur Loss: 0.23365596, Cur Avg Loss: 0.15555724, Log Avg loss: 0.15817193, Global Avg Loss: 0.74251172, Time: 0.0208 Steps: 81060, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000168, Sample Num: 2688, Cur Loss: 0.08804801, Cur Avg Loss: 0.15598377, Log Avg loss: 0.16272297, Global Avg Loss: 0.74244021, Time: 0.0209 Steps: 81070, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000178, Sample Num: 2848, Cur Loss: 0.33423048, Cur Avg Loss: 0.16028858, Log Avg loss: 0.23260928, Global Avg Loss: 0.74237733, Time: 0.0209 Steps: 81080, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000188, Sample Num: 3008, Cur Loss: 0.10866098, Cur Avg Loss: 0.16309985, Log Avg loss: 0.21314059, Global Avg Loss: 0.74231206, Time: 0.0208 Steps: 81090, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000198, Sample Num: 3168, Cur Loss: 0.14960939, Cur Avg Loss: 0.16331223, Log Avg loss: 0.16730486, Global Avg Loss: 0.74224116, Time: 0.0208 Steps: 81100, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000208, Sample Num: 3328, Cur Loss: 0.09212569, Cur Avg Loss: 0.16608830, Log Avg loss: 0.22105455, Global Avg Loss: 0.74217690, Time: 0.0208 Steps: 81110, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000218, Sample Num: 3488, Cur Loss: 0.06267581, Cur Avg Loss: 0.16591412, Log Avg loss: 0.16229125, Global Avg Loss: 0.74210542, Time: 0.0209 Steps: 81120, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000228, Sample Num: 3648, Cur Loss: 0.54104680, Cur Avg Loss: 0.16881607, Log Avg loss: 0.23207853, Global Avg Loss: 0.74204255, Time: 0.0208 Steps: 81130, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000238, Sample Num: 3808, Cur Loss: 0.14613688, Cur Avg Loss: 0.16997820, Log Avg loss: 0.19647475, Global Avg Loss: 0.74197532, Time: 0.0209 Steps: 81140, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000248, Sample Num: 3968, Cur Loss: 0.07869495, Cur Avg Loss: 0.16943686, Log Avg loss: 0.15655300, Global Avg Loss: 0.74190318, Time: 0.0208 Steps: 81150, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000258, Sample Num: 4128, Cur Loss: 0.37769783, Cur Avg Loss: 0.16688452, Log Avg loss: 0.10358646, Global Avg Loss: 0.74182453, Time: 0.0247 Steps: 81160, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000268, Sample Num: 4288, Cur Loss: 0.10244931, Cur Avg Loss: 0.16726222, Log Avg loss: 0.17700681, Global Avg Loss: 0.74175494, Time: 0.0209 Steps: 81170, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000278, Sample Num: 4448, Cur Loss: 0.04384902, Cur Avg Loss: 0.16630288, Log Avg loss: 0.14059252, Global Avg Loss: 0.74168089, Time: 0.0208 Steps: 81180, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000288, Sample Num: 4608, Cur Loss: 0.33897150, Cur Avg Loss: 0.16681418, Log Avg loss: 0.18102840, Global Avg Loss: 0.74161183, Time: 0.0209 Steps: 81190, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000298, Sample Num: 4768, Cur Loss: 0.15005255, Cur Avg Loss: 0.16589574, Log Avg loss: 0.13944478, Global Avg Loss: 0.74153768, Time: 0.0210 Steps: 81200, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000308, Sample Num: 4928, Cur Loss: 0.13208820, Cur Avg Loss: 0.16610694, Log Avg loss: 0.17240049, Global Avg Loss: 0.74146759, Time: 0.0208 Steps: 81210, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000318, Sample Num: 5088, Cur Loss: 0.14561790, Cur Avg Loss: 0.16562648, Log Avg loss: 0.15082841, Global Avg Loss: 0.74139487, Time: 0.0209 Steps: 81220, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000328, Sample Num: 5248, Cur Loss: 0.12640968, Cur Avg Loss: 0.16530313, Log Avg loss: 0.15502052, Global Avg Loss: 0.74132269, Time: 0.0209 Steps: 81230, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000338, Sample Num: 5408, Cur Loss: 0.03843761, Cur Avg Loss: 0.16494062, Log Avg loss: 0.15305026, Global Avg Loss: 0.74125027, Time: 0.0209 Steps: 81240, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000348, Sample Num: 5568, Cur Loss: 0.10401864, Cur Avg Loss: 0.16437875, Log Avg loss: 0.14538782, Global Avg Loss: 0.74117694, Time: 0.0208 Steps: 81250, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000358, Sample Num: 5728, Cur Loss: 0.16947490, Cur Avg Loss: 0.16507973, Log Avg loss: 0.18947368, Global Avg Loss: 0.74110904, Time: 0.0209 Steps: 81260, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000368, Sample Num: 5888, Cur Loss: 0.11807515, Cur Avg Loss: 0.16559282, Log Avg loss: 0.18396145, Global Avg Loss: 0.74104049, Time: 0.0208 Steps: 81270, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000378, Sample Num: 6048, Cur Loss: 0.13813852, Cur Avg Loss: 0.16547156, Log Avg loss: 0.16100917, Global Avg Loss: 0.74096913, Time: 0.0208 Steps: 81280, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000388, Sample Num: 6208, Cur Loss: 0.11145976, Cur Avg Loss: 0.16502492, Log Avg loss: 0.14814204, Global Avg Loss: 0.74089620, Time: 0.0212 Steps: 81290, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000398, Sample Num: 6368, Cur Loss: 0.18653287, Cur Avg Loss: 0.16352646, Log Avg loss: 0.10538600, Global Avg Loss: 0.74081803, Time: 0.0209 Steps: 81300, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000408, Sample Num: 6528, Cur Loss: 0.08498305, Cur Avg Loss: 0.16324925, Log Avg loss: 0.15221643, Global Avg Loss: 0.74074564, Time: 0.0210 Steps: 81310, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000418, Sample Num: 6688, Cur Loss: 0.15955189, Cur Avg Loss: 0.16323777, Log Avg loss: 0.16276940, Global Avg Loss: 0.74067457, Time: 0.0209 Steps: 81320, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000428, Sample Num: 6848, Cur Loss: 0.28135553, Cur Avg Loss: 0.16334637, Log Avg loss: 0.16788587, Global Avg Loss: 0.74060414, Time: 0.0209 Steps: 81330, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000438, Sample Num: 7008, Cur Loss: 0.03145318, Cur Avg Loss: 0.16383349, Log Avg loss: 0.18468210, Global Avg Loss: 0.74053579, Time: 0.0209 Steps: 81340, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000448, Sample Num: 7168, Cur Loss: 0.32482672, Cur Avg Loss: 0.16338095, Log Avg loss: 0.14355960, Global Avg Loss: 0.74046241, Time: 0.0210 Steps: 81350, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000458, Sample Num: 7328, Cur Loss: 0.08216040, Cur Avg Loss: 0.16274105, Log Avg loss: 0.13407377, Global Avg Loss: 0.74038788, Time: 0.0209 Steps: 81360, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000468, Sample Num: 7488, Cur Loss: 0.27957699, Cur Avg Loss: 0.16242191, Log Avg loss: 0.14780510, Global Avg Loss: 0.74031505, Time: 0.0210 Steps: 81370, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000478, Sample Num: 7648, Cur Loss: 0.12948835, Cur Avg Loss: 0.16309284, Log Avg loss: 0.19449271, Global Avg Loss: 0.74024798, Time: 0.0209 Steps: 81380, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000488, Sample Num: 7808, Cur Loss: 0.35844952, Cur Avg Loss: 0.16286693, Log Avg loss: 0.15206820, Global Avg Loss: 0.74017571, Time: 0.0209 Steps: 81390, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000498, Sample Num: 7968, Cur Loss: 0.13212651, Cur Avg Loss: 0.16280568, Log Avg loss: 0.15981690, Global Avg Loss: 0.74010442, Time: 0.0209 Steps: 81400, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000508, Sample Num: 8128, Cur Loss: 0.49423450, Cur Avg Loss: 0.16378315, Log Avg loss: 0.21246103, Global Avg Loss: 0.74003960, Time: 0.0209 Steps: 81410, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000518, Sample Num: 8288, Cur Loss: 0.04966480, Cur Avg Loss: 0.16418902, Log Avg loss: 0.18480691, Global Avg Loss: 0.73997141, Time: 0.0210 Steps: 81420, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000528, Sample Num: 8448, Cur Loss: 0.14432088, Cur Avg Loss: 0.16442757, Log Avg loss: 0.17678466, Global Avg Loss: 0.73990225, Time: 0.0208 Steps: 81430, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000538, Sample Num: 8608, Cur Loss: 0.13133748, Cur Avg Loss: 0.16358755, Log Avg loss: 0.11923474, Global Avg Loss: 0.73982604, Time: 0.0208 Steps: 81440, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000548, Sample Num: 8768, Cur Loss: 0.10776813, Cur Avg Loss: 0.16350604, Log Avg loss: 0.15912085, Global Avg Loss: 0.73975474, Time: 0.0208 Steps: 81450, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000558, Sample Num: 8928, Cur Loss: 0.06057444, Cur Avg Loss: 0.16310488, Log Avg loss: 0.14112096, Global Avg Loss: 0.73968125, Time: 0.0208 Steps: 81460, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000568, Sample Num: 9088, Cur Loss: 0.07561354, Cur Avg Loss: 0.16207216, Log Avg loss: 0.10444668, Global Avg Loss: 0.73960328, Time: 0.0208 Steps: 81470, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000578, Sample Num: 9248, Cur Loss: 0.06053627, Cur Avg Loss: 0.16240792, Log Avg loss: 0.18147901, Global Avg Loss: 0.73953478, Time: 0.0208 Steps: 81480, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000588, Sample Num: 9408, Cur Loss: 0.31207812, Cur Avg Loss: 0.16233522, Log Avg loss: 0.15813298, Global Avg Loss: 0.73946344, Time: 0.0208 Steps: 81490, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000598, Sample Num: 9568, Cur Loss: 0.29544562, Cur Avg Loss: 0.16206914, Log Avg loss: 0.14642399, Global Avg Loss: 0.73939067, Time: 0.0208 Steps: 81500, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000608, Sample Num: 9728, Cur Loss: 0.32213938, Cur Avg Loss: 0.16241753, Log Avg loss: 0.18325080, Global Avg Loss: 0.73932244, Time: 0.0208 Steps: 81510, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000618, Sample Num: 9888, Cur Loss: 0.06358033, Cur Avg Loss: 0.16164008, Log Avg loss: 0.11437101, Global Avg Loss: 0.73924578, Time: 0.0209 Steps: 81520, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000628, Sample Num: 10048, Cur Loss: 0.19086233, Cur Avg Loss: 0.16178374, Log Avg loss: 0.17066195, Global Avg Loss: 0.73917604, Time: 0.0208 Steps: 81530, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000638, Sample Num: 10208, Cur Loss: 0.07948650, Cur Avg Loss: 0.16291252, Log Avg loss: 0.23380013, Global Avg Loss: 0.73911406, Time: 0.0208 Steps: 81540, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000648, Sample Num: 10368, Cur Loss: 0.07933518, Cur Avg Loss: 0.16239549, Log Avg loss: 0.12940911, Global Avg Loss: 0.73903930, Time: 0.0208 Steps: 81550, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000658, Sample Num: 10528, Cur Loss: 0.26385334, Cur Avg Loss: 0.16221773, Log Avg loss: 0.15069899, Global Avg Loss: 0.73896716, Time: 0.0208 Steps: 81560, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000668, Sample Num: 10688, Cur Loss: 0.09170496, Cur Avg Loss: 0.16176768, Log Avg loss: 0.13215430, Global Avg Loss: 0.73889277, Time: 0.0208 Steps: 81570, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000678, Sample Num: 10848, Cur Loss: 0.12598637, Cur Avg Loss: 0.16123177, Log Avg loss: 0.12543258, Global Avg Loss: 0.73881757, Time: 0.0209 Steps: 81580, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000688, Sample Num: 11008, Cur Loss: 0.08099379, Cur Avg Loss: 0.16141423, Log Avg loss: 0.17378517, Global Avg Loss: 0.73874832, Time: 0.0208 Steps: 81590, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000698, Sample Num: 11168, Cur Loss: 0.27500883, Cur Avg Loss: 0.16149085, Log Avg loss: 0.16676209, Global Avg Loss: 0.73867822, Time: 0.0209 Steps: 81600, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000708, Sample Num: 11328, Cur Loss: 0.27312624, Cur Avg Loss: 0.16266522, Log Avg loss: 0.24463683, Global Avg Loss: 0.73861769, Time: 0.0208 Steps: 81610, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000718, Sample Num: 11488, Cur Loss: 0.05738006, Cur Avg Loss: 0.16170895, Log Avg loss: 0.09400495, Global Avg Loss: 0.73853871, Time: 0.0209 Steps: 81620, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000728, Sample Num: 11648, Cur Loss: 0.29027945, Cur Avg Loss: 0.16304240, Log Avg loss: 0.25878360, Global Avg Loss: 0.73847994, Time: 0.0208 Steps: 81630, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000738, Sample Num: 11808, Cur Loss: 0.37140357, Cur Avg Loss: 0.16338040, Log Avg loss: 0.18798687, Global Avg Loss: 0.73841251, Time: 0.0209 Steps: 81640, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000748, Sample Num: 11968, Cur Loss: 0.11070976, Cur Avg Loss: 0.16305769, Log Avg loss: 0.13924159, Global Avg Loss: 0.73833912, Time: 0.0209 Steps: 81650, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000758, Sample Num: 12128, Cur Loss: 0.10209110, Cur Avg Loss: 0.16392927, Log Avg loss: 0.22912394, Global Avg Loss: 0.73827677, Time: 0.0208 Steps: 81660, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000768, Sample Num: 12288, Cur Loss: 0.03339926, Cur Avg Loss: 0.16309286, Log Avg loss: 0.09969281, Global Avg Loss: 0.73819858, Time: 0.0254 Steps: 81670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000778, Sample Num: 12448, Cur Loss: 0.11926239, Cur Avg Loss: 0.16398773, Log Avg loss: 0.23271393, Global Avg Loss: 0.73813669, Time: 0.0209 Steps: 81680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000788, Sample Num: 12608, Cur Loss: 0.09063444, Cur Avg Loss: 0.16352636, Log Avg loss: 0.12763137, Global Avg Loss: 0.73806195, Time: 0.0209 Steps: 81690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000798, Sample Num: 12768, Cur Loss: 0.01685191, Cur Avg Loss: 0.16278287, Log Avg loss: 0.10419600, Global Avg Loss: 0.73798437, Time: 0.0208 Steps: 81700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000808, Sample Num: 12928, Cur Loss: 0.09041516, Cur Avg Loss: 0.16366633, Log Avg loss: 0.23416665, Global Avg Loss: 0.73792271, Time: 0.0209 Steps: 81710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000818, Sample Num: 13088, Cur Loss: 0.05327893, Cur Avg Loss: 0.16380233, Log Avg loss: 0.17479080, Global Avg Loss: 0.73785380, Time: 0.0208 Steps: 81720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000828, Sample Num: 13248, Cur Loss: 0.09840652, Cur Avg Loss: 0.16390738, Log Avg loss: 0.17250049, Global Avg Loss: 0.73778463, Time: 0.0209 Steps: 81730, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000838, Sample Num: 13408, Cur Loss: 0.15275961, Cur Avg Loss: 0.16352196, Log Avg loss: 0.13160918, Global Avg Loss: 0.73771047, Time: 0.0208 Steps: 81740, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000848, Sample Num: 13568, Cur Loss: 0.13948587, Cur Avg Loss: 0.16345321, Log Avg loss: 0.15769186, Global Avg Loss: 0.73763952, Time: 0.0208 Steps: 81750, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000858, Sample Num: 13728, Cur Loss: 0.19619565, Cur Avg Loss: 0.16361464, Log Avg loss: 0.17730441, Global Avg Loss: 0.73757098, Time: 0.0208 Steps: 81760, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000868, Sample Num: 13888, Cur Loss: 0.19575214, Cur Avg Loss: 0.16415302, Log Avg loss: 0.21034623, Global Avg Loss: 0.73750651, Time: 0.0208 Steps: 81770, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000878, Sample Num: 14048, Cur Loss: 0.49544150, Cur Avg Loss: 0.16451158, Log Avg loss: 0.19563412, Global Avg Loss: 0.73744025, Time: 0.0209 Steps: 81780, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000888, Sample Num: 14208, Cur Loss: 0.20113023, Cur Avg Loss: 0.16443277, Log Avg loss: 0.15751349, Global Avg Loss: 0.73736934, Time: 0.0208 Steps: 81790, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000898, Sample Num: 14368, Cur Loss: 0.20769784, Cur Avg Loss: 0.16491680, Log Avg loss: 0.20789894, Global Avg Loss: 0.73730462, Time: 0.0208 Steps: 81800, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000908, Sample Num: 14528, Cur Loss: 0.04934736, Cur Avg Loss: 0.16468841, Log Avg loss: 0.14417817, Global Avg Loss: 0.73723212, Time: 0.0208 Steps: 81810, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000918, Sample Num: 14688, Cur Loss: 0.15119511, Cur Avg Loss: 0.16473867, Log Avg loss: 0.16930237, Global Avg Loss: 0.73716270, Time: 0.0208 Steps: 81820, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000928, Sample Num: 14848, Cur Loss: 0.08683305, Cur Avg Loss: 0.16484208, Log Avg loss: 0.17433501, Global Avg Loss: 0.73709392, Time: 0.0208 Steps: 81830, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000938, Sample Num: 15008, Cur Loss: 0.18263546, Cur Avg Loss: 0.16489629, Log Avg loss: 0.16992769, Global Avg Loss: 0.73702462, Time: 0.0208 Steps: 81840, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000948, Sample Num: 15168, Cur Loss: 0.10694528, Cur Avg Loss: 0.16509256, Log Avg loss: 0.18350232, Global Avg Loss: 0.73695700, Time: 0.0208 Steps: 81850, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000958, Sample Num: 15328, Cur Loss: 0.20970726, Cur Avg Loss: 0.16495269, Log Avg loss: 0.15169341, Global Avg Loss: 0.73688550, Time: 0.0208 Steps: 81860, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000968, Sample Num: 15488, Cur Loss: 0.12472562, Cur Avg Loss: 0.16466631, Log Avg loss: 0.13723056, Global Avg Loss: 0.73681225, Time: 0.0209 Steps: 81870, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000978, Sample Num: 15648, Cur Loss: 0.05196689, Cur Avg Loss: 0.16441343, Log Avg loss: 0.13993444, Global Avg Loss: 0.73673936, Time: 0.0208 Steps: 81880, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000988, Sample Num: 15808, Cur Loss: 0.22379999, Cur Avg Loss: 0.16422853, Log Avg loss: 0.14614544, Global Avg Loss: 0.73666724, Time: 0.0208 Steps: 81890, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000998, Sample Num: 15968, Cur Loss: 0.03743340, Cur Avg Loss: 0.16381184, Log Avg loss: 0.12264285, Global Avg Loss: 0.73659227, Time: 0.0208 Steps: 81900, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001008, Sample Num: 16128, Cur Loss: 0.31570482, Cur Avg Loss: 0.16401711, Log Avg loss: 0.18450299, Global Avg Loss: 0.73652486, Time: 0.0208 Steps: 81910, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001018, Sample Num: 16288, Cur Loss: 0.87510753, Cur Avg Loss: 0.16470875, Log Avg loss: 0.23442592, Global Avg Loss: 0.73646357, Time: 0.0208 Steps: 81920, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001028, Sample Num: 16448, Cur Loss: 0.21745569, Cur Avg Loss: 0.16518284, Log Avg loss: 0.21344572, Global Avg Loss: 0.73639973, Time: 0.0229 Steps: 81930, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001038, Sample Num: 16608, Cur Loss: 0.37410465, Cur Avg Loss: 0.16519451, Log Avg loss: 0.16639443, Global Avg Loss: 0.73633017, Time: 0.0209 Steps: 81940, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001048, Sample Num: 16768, Cur Loss: 0.11809006, Cur Avg Loss: 0.16503013, Log Avg loss: 0.14796710, Global Avg Loss: 0.73625838, Time: 0.0209 Steps: 81950, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001058, Sample Num: 16928, Cur Loss: 0.23330230, Cur Avg Loss: 0.16478523, Log Avg loss: 0.13912035, Global Avg Loss: 0.73618552, Time: 0.0209 Steps: 81960, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001068, Sample Num: 17088, Cur Loss: 0.14395484, Cur Avg Loss: 0.16459786, Log Avg loss: 0.14477370, Global Avg Loss: 0.73611337, Time: 0.0210 Steps: 81970, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001078, Sample Num: 17248, Cur Loss: 0.21095322, Cur Avg Loss: 0.16474185, Log Avg loss: 0.18011949, Global Avg Loss: 0.73604555, Time: 0.0208 Steps: 81980, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001088, Sample Num: 17408, Cur Loss: 0.08124663, Cur Avg Loss: 0.16473607, Log Avg loss: 0.16411382, Global Avg Loss: 0.73597579, Time: 0.0208 Steps: 81990, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001098, Sample Num: 17568, Cur Loss: 0.19503638, Cur Avg Loss: 0.16501952, Log Avg loss: 0.19585809, Global Avg Loss: 0.73590992, Time: 0.0209 Steps: 82000, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001108, Sample Num: 17728, Cur Loss: 0.07257846, Cur Avg Loss: 0.16494783, Log Avg loss: 0.15707638, Global Avg Loss: 0.73583934, Time: 0.0208 Steps: 82010, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001118, Sample Num: 17888, Cur Loss: 0.03463154, Cur Avg Loss: 0.16493662, Log Avg loss: 0.16369513, Global Avg Loss: 0.73576959, Time: 0.0208 Steps: 82020, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001128, Sample Num: 18048, Cur Loss: 0.16728625, Cur Avg Loss: 0.16496609, Log Avg loss: 0.16826052, Global Avg Loss: 0.73570040, Time: 0.0208 Steps: 82030, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001138, Sample Num: 18208, Cur Loss: 0.13389906, Cur Avg Loss: 0.16495258, Log Avg loss: 0.16342841, Global Avg Loss: 0.73563065, Time: 0.0207 Steps: 82040, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001148, Sample Num: 18368, Cur Loss: 0.17940265, Cur Avg Loss: 0.16481245, Log Avg loss: 0.14886576, Global Avg Loss: 0.73555913, Time: 0.0208 Steps: 82050, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001158, Sample Num: 18528, Cur Loss: 0.25767940, Cur Avg Loss: 0.16493135, Log Avg loss: 0.17858163, Global Avg Loss: 0.73549126, Time: 0.0209 Steps: 82060, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001168, Sample Num: 18688, Cur Loss: 0.13045764, Cur Avg Loss: 0.16530155, Log Avg loss: 0.20817048, Global Avg Loss: 0.73542701, Time: 0.0209 Steps: 82070, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001178, Sample Num: 18848, Cur Loss: 0.06139632, Cur Avg Loss: 0.16509296, Log Avg loss: 0.14072883, Global Avg Loss: 0.73535455, Time: 0.0208 Steps: 82080, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001188, Sample Num: 19008, Cur Loss: 0.15645733, Cur Avg Loss: 0.16609858, Log Avg loss: 0.28456096, Global Avg Loss: 0.73529964, Time: 0.0208 Steps: 82090, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001198, Sample Num: 19168, Cur Loss: 0.16546768, Cur Avg Loss: 0.16622782, Log Avg loss: 0.18158169, Global Avg Loss: 0.73523220, Time: 0.0208 Steps: 82100, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001208, Sample Num: 19328, Cur Loss: 0.09321687, Cur Avg Loss: 0.16581328, Log Avg loss: 0.11615124, Global Avg Loss: 0.73515680, Time: 0.0208 Steps: 82110, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001218, Sample Num: 19488, Cur Loss: 0.24389893, Cur Avg Loss: 0.16601565, Log Avg loss: 0.19046230, Global Avg Loss: 0.73509047, Time: 0.0210 Steps: 82120, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001228, Sample Num: 19648, Cur Loss: 0.29783806, Cur Avg Loss: 0.16692146, Log Avg loss: 0.27724957, Global Avg Loss: 0.73503472, Time: 0.0209 Steps: 82130, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001238, Sample Num: 19808, Cur Loss: 0.09410603, Cur Avg Loss: 0.16674594, Log Avg loss: 0.14519115, Global Avg Loss: 0.73496291, Time: 0.0210 Steps: 82140, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001248, Sample Num: 19968, Cur Loss: 0.12972388, Cur Avg Loss: 0.16718920, Log Avg loss: 0.22206514, Global Avg Loss: 0.73490048, Time: 0.0209 Steps: 82150, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001258, Sample Num: 20128, Cur Loss: 0.07180291, Cur Avg Loss: 0.16741519, Log Avg loss: 0.19561883, Global Avg Loss: 0.73483484, Time: 0.0209 Steps: 82160, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001268, Sample Num: 20288, Cur Loss: 0.10984471, Cur Avg Loss: 0.16728094, Log Avg loss: 0.15039172, Global Avg Loss: 0.73476372, Time: 0.0209 Steps: 82170, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001278, Sample Num: 20448, Cur Loss: 0.19737020, Cur Avg Loss: 0.16729531, Log Avg loss: 0.16911848, Global Avg Loss: 0.73469489, Time: 0.0208 Steps: 82180, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001288, Sample Num: 20608, Cur Loss: 0.15494582, Cur Avg Loss: 0.16733464, Log Avg loss: 0.17236092, Global Avg Loss: 0.73462647, Time: 0.0208 Steps: 82190, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001298, Sample Num: 20768, Cur Loss: 0.22865011, Cur Avg Loss: 0.16792280, Log Avg loss: 0.24367748, Global Avg Loss: 0.73456674, Time: 0.0208 Steps: 82200, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001308, Sample Num: 20928, Cur Loss: 0.08683793, Cur Avg Loss: 0.16756774, Log Avg loss: 0.12148022, Global Avg Loss: 0.73449217, Time: 0.0208 Steps: 82210, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001318, Sample Num: 21088, Cur Loss: 0.09862968, Cur Avg Loss: 0.16740927, Log Avg loss: 0.14668210, Global Avg Loss: 0.73442067, Time: 0.0208 Steps: 82220, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001328, Sample Num: 21248, Cur Loss: 0.06718421, Cur Avg Loss: 0.16702620, Log Avg loss: 0.11653755, Global Avg Loss: 0.73434553, Time: 0.0209 Steps: 82230, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001338, Sample Num: 21408, Cur Loss: 0.09948494, Cur Avg Loss: 0.16713914, Log Avg loss: 0.18213777, Global Avg Loss: 0.73427839, Time: 0.0208 Steps: 82240, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001348, Sample Num: 21568, Cur Loss: 0.64024901, Cur Avg Loss: 0.16754784, Log Avg loss: 0.22223101, Global Avg Loss: 0.73421613, Time: 0.0208 Steps: 82250, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001358, Sample Num: 21728, Cur Loss: 0.13746536, Cur Avg Loss: 0.16761298, Log Avg loss: 0.17639411, Global Avg Loss: 0.73414832, Time: 0.0208 Steps: 82260, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001368, Sample Num: 21888, Cur Loss: 0.14616013, Cur Avg Loss: 0.16738945, Log Avg loss: 0.13703448, Global Avg Loss: 0.73407574, Time: 0.0208 Steps: 82270, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001378, Sample Num: 22048, Cur Loss: 0.35693353, Cur Avg Loss: 0.16766812, Log Avg loss: 0.20579072, Global Avg Loss: 0.73401153, Time: 0.0208 Steps: 82280, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001388, Sample Num: 22208, Cur Loss: 0.05049382, Cur Avg Loss: 0.16743847, Log Avg loss: 0.13579242, Global Avg Loss: 0.73393884, Time: 0.0208 Steps: 82290, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001398, Sample Num: 22368, Cur Loss: 0.18050441, Cur Avg Loss: 0.16722750, Log Avg loss: 0.13794394, Global Avg Loss: 0.73386642, Time: 0.0208 Steps: 82300, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001408, Sample Num: 22528, Cur Loss: 0.08719377, Cur Avg Loss: 0.16711631, Log Avg loss: 0.15157186, Global Avg Loss: 0.73379568, Time: 0.0208 Steps: 82310, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001418, Sample Num: 22688, Cur Loss: 0.25269979, Cur Avg Loss: 0.16771945, Log Avg loss: 0.25264231, Global Avg Loss: 0.73373723, Time: 0.0208 Steps: 82320, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001428, Sample Num: 22848, Cur Loss: 0.33263794, Cur Avg Loss: 0.16779505, Log Avg loss: 0.17851462, Global Avg Loss: 0.73366979, Time: 0.0209 Steps: 82330, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001438, Sample Num: 23008, Cur Loss: 0.07080262, Cur Avg Loss: 0.16786357, Log Avg loss: 0.17764891, Global Avg Loss: 0.73360226, Time: 0.0209 Steps: 82340, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001448, Sample Num: 23168, Cur Loss: 0.24166524, Cur Avg Loss: 0.16781905, Log Avg loss: 0.16141698, Global Avg Loss: 0.73353278, Time: 0.0208 Steps: 82350, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001458, Sample Num: 23328, Cur Loss: 0.06002168, Cur Avg Loss: 0.16767391, Log Avg loss: 0.14665737, Global Avg Loss: 0.73346152, Time: 0.0208 Steps: 82360, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001468, Sample Num: 23488, Cur Loss: 0.32132053, Cur Avg Loss: 0.16842191, Log Avg loss: 0.27748066, Global Avg Loss: 0.73340616, Time: 0.0208 Steps: 82370, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001478, Sample Num: 23648, Cur Loss: 0.29767764, Cur Avg Loss: 0.16881130, Log Avg loss: 0.22597410, Global Avg Loss: 0.73334457, Time: 0.0208 Steps: 82380, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001488, Sample Num: 23808, Cur Loss: 0.06712487, Cur Avg Loss: 0.16874553, Log Avg loss: 0.15902376, Global Avg Loss: 0.73327486, Time: 0.0209 Steps: 82390, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001498, Sample Num: 23968, Cur Loss: 0.05655839, Cur Avg Loss: 0.16875514, Log Avg loss: 0.17018499, Global Avg Loss: 0.73320652, Time: 0.0209 Steps: 82400, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001508, Sample Num: 24128, Cur Loss: 0.09356977, Cur Avg Loss: 0.16853920, Log Avg loss: 0.13619171, Global Avg Loss: 0.73313408, Time: 0.0209 Steps: 82410, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001518, Sample Num: 24288, Cur Loss: 0.05597078, Cur Avg Loss: 0.16825722, Log Avg loss: 0.12573456, Global Avg Loss: 0.73306038, Time: 0.0208 Steps: 82420, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001528, Sample Num: 24448, Cur Loss: 0.17493194, Cur Avg Loss: 0.16811695, Log Avg loss: 0.14682441, Global Avg Loss: 0.73298926, Time: 0.0209 Steps: 82430, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001538, Sample Num: 24608, Cur Loss: 0.09323484, Cur Avg Loss: 0.16839164, Log Avg loss: 0.21036428, Global Avg Loss: 0.73292587, Time: 0.0247 Steps: 82440, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001548, Sample Num: 24768, Cur Loss: 0.31146497, Cur Avg Loss: 0.16848838, Log Avg loss: 0.18336676, Global Avg Loss: 0.73285922, Time: 0.0210 Steps: 82450, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001558, Sample Num: 24928, Cur Loss: 0.12955412, Cur Avg Loss: 0.16852663, Log Avg loss: 0.17444721, Global Avg Loss: 0.73279150, Time: 0.0209 Steps: 82460, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001568, Sample Num: 25088, Cur Loss: 0.10598148, Cur Avg Loss: 0.16863987, Log Avg loss: 0.18628327, Global Avg Loss: 0.73272523, Time: 0.0210 Steps: 82470, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001578, Sample Num: 25248, Cur Loss: 0.07826260, Cur Avg Loss: 0.16846504, Log Avg loss: 0.14105116, Global Avg Loss: 0.73265349, Time: 0.0210 Steps: 82480, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001588, Sample Num: 25408, Cur Loss: 0.10417251, Cur Avg Loss: 0.16838966, Log Avg loss: 0.15649519, Global Avg Loss: 0.73258365, Time: 0.0209 Steps: 82490, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001598, Sample Num: 25568, Cur Loss: 0.47943056, Cur Avg Loss: 0.16878145, Log Avg loss: 0.23099728, Global Avg Loss: 0.73252285, Time: 0.0209 Steps: 82500, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001608, Sample Num: 25728, Cur Loss: 0.08423482, Cur Avg Loss: 0.16865592, Log Avg loss: 0.14859708, Global Avg Loss: 0.73245208, Time: 0.0209 Steps: 82510, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001618, Sample Num: 25888, Cur Loss: 0.06810272, Cur Avg Loss: 0.16885517, Log Avg loss: 0.20089348, Global Avg Loss: 0.73238766, Time: 0.0210 Steps: 82520, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001628, Sample Num: 26048, Cur Loss: 0.10777757, Cur Avg Loss: 0.16859007, Log Avg loss: 0.12569683, Global Avg Loss: 0.73231415, Time: 0.0209 Steps: 82530, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001638, Sample Num: 26208, Cur Loss: 0.08088724, Cur Avg Loss: 0.16824491, Log Avg loss: 0.11205335, Global Avg Loss: 0.73223901, Time: 0.0210 Steps: 82540, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001648, Sample Num: 26368, Cur Loss: 0.12607554, Cur Avg Loss: 0.16819494, Log Avg loss: 0.16000996, Global Avg Loss: 0.73216969, Time: 0.0210 Steps: 82550, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001658, Sample Num: 26528, Cur Loss: 0.36701196, Cur Avg Loss: 0.16802169, Log Avg loss: 0.13947003, Global Avg Loss: 0.73209790, Time: 0.0210 Steps: 82560, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001668, Sample Num: 26688, Cur Loss: 0.60199279, Cur Avg Loss: 0.16839441, Log Avg loss: 0.23019144, Global Avg Loss: 0.73203711, Time: 0.0210 Steps: 82570, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001678, Sample Num: 26848, Cur Loss: 0.04228121, Cur Avg Loss: 0.16807721, Log Avg loss: 0.11516794, Global Avg Loss: 0.73196241, Time: 0.0210 Steps: 82580, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001688, Sample Num: 27008, Cur Loss: 0.07608746, Cur Avg Loss: 0.16803956, Log Avg loss: 0.16172229, Global Avg Loss: 0.73189337, Time: 0.0210 Steps: 82590, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001698, Sample Num: 27168, Cur Loss: 0.14373852, Cur Avg Loss: 0.16796010, Log Avg loss: 0.15454740, Global Avg Loss: 0.73182347, Time: 0.0210 Steps: 82600, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001708, Sample Num: 27328, Cur Loss: 0.17930964, Cur Avg Loss: 0.16788206, Log Avg loss: 0.15463031, Global Avg Loss: 0.73175360, Time: 0.0210 Steps: 82610, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001718, Sample Num: 27488, Cur Loss: 0.31274074, Cur Avg Loss: 0.16790799, Log Avg loss: 0.17233792, Global Avg Loss: 0.73168589, Time: 0.0210 Steps: 82620, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001728, Sample Num: 27648, Cur Loss: 0.13909087, Cur Avg Loss: 0.16775961, Log Avg loss: 0.14226655, Global Avg Loss: 0.73161456, Time: 0.0210 Steps: 82630, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001738, Sample Num: 27808, Cur Loss: 0.25619787, Cur Avg Loss: 0.16801386, Log Avg loss: 0.21194901, Global Avg Loss: 0.73155167, Time: 0.0210 Steps: 82640, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001748, Sample Num: 27968, Cur Loss: 0.10442897, Cur Avg Loss: 0.16786167, Log Avg loss: 0.14141057, Global Avg Loss: 0.73148027, Time: 0.0210 Steps: 82650, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001758, Sample Num: 28128, Cur Loss: 0.12506345, Cur Avg Loss: 0.16777719, Log Avg loss: 0.15301062, Global Avg Loss: 0.73141029, Time: 0.0210 Steps: 82660, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001768, Sample Num: 28288, Cur Loss: 0.11654866, Cur Avg Loss: 0.16796600, Log Avg loss: 0.20115844, Global Avg Loss: 0.73134615, Time: 0.0210 Steps: 82670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001778, Sample Num: 28448, Cur Loss: 0.12758514, Cur Avg Loss: 0.16827118, Log Avg loss: 0.22222711, Global Avg Loss: 0.73128457, Time: 0.0210 Steps: 82680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001788, Sample Num: 28608, Cur Loss: 0.13189988, Cur Avg Loss: 0.16802779, Log Avg loss: 0.12475268, Global Avg Loss: 0.73121122, Time: 0.0210 Steps: 82690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001798, Sample Num: 28768, Cur Loss: 0.16420123, Cur Avg Loss: 0.16777171, Log Avg loss: 0.12198566, Global Avg Loss: 0.73113756, Time: 0.0247 Steps: 82700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001808, Sample Num: 28928, Cur Loss: 0.06631353, Cur Avg Loss: 0.16757726, Log Avg loss: 0.13261432, Global Avg Loss: 0.73106519, Time: 0.0210 Steps: 82710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001818, Sample Num: 29088, Cur Loss: 0.11856204, Cur Avg Loss: 0.16761755, Log Avg loss: 0.17490303, Global Avg Loss: 0.73099796, Time: 0.0209 Steps: 82720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001828, Sample Num: 29248, Cur Loss: 0.39977670, Cur Avg Loss: 0.16779418, Log Avg loss: 0.19990431, Global Avg Loss: 0.73093376, Time: 0.0209 Steps: 82730, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001838, Sample Num: 29408, Cur Loss: 0.17009440, Cur Avg Loss: 0.16775055, Log Avg loss: 0.15977625, Global Avg Loss: 0.73086473, Time: 0.0209 Steps: 82740, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001848, Sample Num: 29568, Cur Loss: 0.10866116, Cur Avg Loss: 0.16795799, Log Avg loss: 0.20608517, Global Avg Loss: 0.73080131, Time: 0.0209 Steps: 82750, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001858, Sample Num: 29728, Cur Loss: 0.06981101, Cur Avg Loss: 0.16803290, Log Avg loss: 0.18187493, Global Avg Loss: 0.73073499, Time: 0.0210 Steps: 82760, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001868, Sample Num: 29888, Cur Loss: 0.28010082, Cur Avg Loss: 0.16799030, Log Avg loss: 0.16007583, Global Avg Loss: 0.73066604, Time: 0.0209 Steps: 82770, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001878, Sample Num: 30048, Cur Loss: 0.05454944, Cur Avg Loss: 0.16802883, Log Avg loss: 0.17522734, Global Avg Loss: 0.73059894, Time: 0.0209 Steps: 82780, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001888, Sample Num: 30208, Cur Loss: 0.04454152, Cur Avg Loss: 0.16793078, Log Avg loss: 0.14951551, Global Avg Loss: 0.73052875, Time: 0.0209 Steps: 82790, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001898, Sample Num: 30368, Cur Loss: 0.17219760, Cur Avg Loss: 0.16806138, Log Avg loss: 0.19271977, Global Avg Loss: 0.73046380, Time: 0.0210 Steps: 82800, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001908, Sample Num: 30528, Cur Loss: 0.32483464, Cur Avg Loss: 0.16847655, Log Avg loss: 0.24727508, Global Avg Loss: 0.73040545, Time: 0.0209 Steps: 82810, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001918, Sample Num: 30688, Cur Loss: 0.15435904, Cur Avg Loss: 0.16844524, Log Avg loss: 0.16247200, Global Avg Loss: 0.73033688, Time: 0.0209 Steps: 82820, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001928, Sample Num: 30848, Cur Loss: 0.09188819, Cur Avg Loss: 0.16838818, Log Avg loss: 0.15744272, Global Avg Loss: 0.73026771, Time: 0.0209 Steps: 82830, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001938, Sample Num: 31008, Cur Loss: 0.12399061, Cur Avg Loss: 0.16863989, Log Avg loss: 0.21716973, Global Avg Loss: 0.73020578, Time: 0.0210 Steps: 82840, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001948, Sample Num: 31168, Cur Loss: 0.09056486, Cur Avg Loss: 0.16882282, Log Avg loss: 0.20427582, Global Avg Loss: 0.73014230, Time: 0.0210 Steps: 82850, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001958, Sample Num: 31328, Cur Loss: 0.04712336, Cur Avg Loss: 0.16898087, Log Avg loss: 0.19976790, Global Avg Loss: 0.73007829, Time: 0.0209 Steps: 82860, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001968, Sample Num: 31488, Cur Loss: 0.12695304, Cur Avg Loss: 0.16894795, Log Avg loss: 0.16250323, Global Avg Loss: 0.73000980, Time: 0.0209 Steps: 82870, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001978, Sample Num: 31648, Cur Loss: 0.32537496, Cur Avg Loss: 0.16923265, Log Avg loss: 0.22526081, Global Avg Loss: 0.72994890, Time: 0.0210 Steps: 82880, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001988, Sample Num: 31808, Cur Loss: 0.28435007, Cur Avg Loss: 0.16901529, Log Avg loss: 0.12602294, Global Avg Loss: 0.72987604, Time: 0.0210 Steps: 82890, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001998, Sample Num: 31968, Cur Loss: 0.04451376, Cur Avg Loss: 0.16934139, Log Avg loss: 0.23417003, Global Avg Loss: 0.72981624, Time: 0.0210 Steps: 82900, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002008, Sample Num: 32128, Cur Loss: 0.14518759, Cur Avg Loss: 0.16968132, Log Avg loss: 0.23759772, Global Avg Loss: 0.72975687, Time: 0.0209 Steps: 82910, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002018, Sample Num: 32288, Cur Loss: 0.20445852, Cur Avg Loss: 0.16957244, Log Avg loss: 0.14770968, Global Avg Loss: 0.72968668, Time: 0.0209 Steps: 82920, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002028, Sample Num: 32448, Cur Loss: 0.15679510, Cur Avg Loss: 0.16967029, Log Avg loss: 0.18941743, Global Avg Loss: 0.72962153, Time: 0.0209 Steps: 82930, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002038, Sample Num: 32608, Cur Loss: 0.11509231, Cur Avg Loss: 0.16977975, Log Avg loss: 0.19197719, Global Avg Loss: 0.72955671, Time: 0.0209 Steps: 82940, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002048, Sample Num: 32768, Cur Loss: 0.27536166, Cur Avg Loss: 0.16997276, Log Avg loss: 0.20930823, Global Avg Loss: 0.72949399, Time: 0.0254 Steps: 82950, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002058, Sample Num: 32928, Cur Loss: 0.08536384, Cur Avg Loss: 0.16978096, Log Avg loss: 0.13050070, Global Avg Loss: 0.72942179, Time: 0.0209 Steps: 82960, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002068, Sample Num: 33088, Cur Loss: 0.40103057, Cur Avg Loss: 0.16996787, Log Avg loss: 0.20843406, Global Avg Loss: 0.72935900, Time: 0.0209 Steps: 82970, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002078, Sample Num: 33248, Cur Loss: 0.06273027, Cur Avg Loss: 0.16998712, Log Avg loss: 0.17396880, Global Avg Loss: 0.72929207, Time: 0.0209 Steps: 82980, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002088, Sample Num: 33408, Cur Loss: 0.17361671, Cur Avg Loss: 0.17003357, Log Avg loss: 0.17968447, Global Avg Loss: 0.72922584, Time: 0.0209 Steps: 82990, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002098, Sample Num: 33568, Cur Loss: 0.14669707, Cur Avg Loss: 0.17003682, Log Avg loss: 0.17071500, Global Avg Loss: 0.72915855, Time: 0.0210 Steps: 83000, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002108, Sample Num: 33728, Cur Loss: 0.17032787, Cur Avg Loss: 0.16987859, Log Avg loss: 0.13668243, Global Avg Loss: 0.72908717, Time: 0.0209 Steps: 83010, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002118, Sample Num: 33888, Cur Loss: 0.04131625, Cur Avg Loss: 0.16957492, Log Avg loss: 0.10556260, Global Avg Loss: 0.72901207, Time: 0.0210 Steps: 83020, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002128, Sample Num: 34048, Cur Loss: 0.06106424, Cur Avg Loss: 0.16918007, Log Avg loss: 0.08555043, Global Avg Loss: 0.72893457, Time: 0.0209 Steps: 83030, Updated lr: 0.000022 ***** Running evaluation checkpoint-83031 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-83031 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.689986, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.241436, "eval_total_loss": 169.729258, "eval_mae": 0.331994, "eval_mse": 0.241528, "eval_r2": 0.846469, "eval_sp_statistic": 0.896005, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.92129, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.719989, "test_total_loss": 361.434617, "test_mae": 0.652555, "test_mse": 0.720055, "test_r2": 0.53527, "test_sp_statistic": 0.801969, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.838669, "test_ps_pvalue": 0.0, "lr": 2.2208629682313894e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7289262487863322, "train_cur_epoch_loss": 360.05304594896734, "train_cur_epoch_avg_loss": 0.16911838701219697, "train_cur_epoch_time": 44.68998599052429, "train_cur_epoch_avg_time": 0.020991069042049927, "epoch": 39, "step": 83031} ################################################## Training, Epoch: 0040, Batch: 000009, Sample Num: 144, Cur Loss: 0.28027183, Cur Avg Loss: 0.17586299, Log Avg loss: 0.16206189, Global Avg Loss: 0.72886631, Time: 0.0221 Steps: 83040, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000019, Sample Num: 304, Cur Loss: 0.28721836, Cur Avg Loss: 0.19148928, Log Avg loss: 0.20555294, Global Avg Loss: 0.72880330, Time: 0.0219 Steps: 83050, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000029, Sample Num: 464, Cur Loss: 0.02972888, Cur Avg Loss: 0.17400843, Log Avg loss: 0.14079483, Global Avg Loss: 0.72873250, Time: 0.0219 Steps: 83060, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000039, Sample Num: 624, Cur Loss: 0.14657539, Cur Avg Loss: 0.15246410, Log Avg loss: 0.08998554, Global Avg Loss: 0.72865561, Time: 0.0218 Steps: 83070, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000049, Sample Num: 784, Cur Loss: 0.05408340, Cur Avg Loss: 0.13627502, Log Avg loss: 0.07313760, Global Avg Loss: 0.72857671, Time: 0.0219 Steps: 83080, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000059, Sample Num: 944, Cur Loss: 0.14973189, Cur Avg Loss: 0.15769129, Log Avg loss: 0.26263101, Global Avg Loss: 0.72852063, Time: 0.0219 Steps: 83090, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000069, Sample Num: 1104, Cur Loss: 0.20597136, Cur Avg Loss: 0.15951523, Log Avg loss: 0.17027647, Global Avg Loss: 0.72845345, Time: 0.0220 Steps: 83100, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000079, Sample Num: 1264, Cur Loss: 0.02846700, Cur Avg Loss: 0.16057119, Log Avg loss: 0.16785730, Global Avg Loss: 0.72838600, Time: 0.0219 Steps: 83110, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000089, Sample Num: 1424, Cur Loss: 0.16707730, Cur Avg Loss: 0.15863169, Log Avg loss: 0.14330969, Global Avg Loss: 0.72831561, Time: 0.0220 Steps: 83120, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000099, Sample Num: 1584, Cur Loss: 0.24964529, Cur Avg Loss: 0.16282949, Log Avg loss: 0.20018991, Global Avg Loss: 0.72825208, Time: 0.0209 Steps: 83130, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000109, Sample Num: 1744, Cur Loss: 0.18322337, Cur Avg Loss: 0.16469228, Log Avg loss: 0.18313392, Global Avg Loss: 0.72818651, Time: 0.0209 Steps: 83140, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000119, Sample Num: 1904, Cur Loss: 0.11623847, Cur Avg Loss: 0.17712014, Log Avg loss: 0.31258380, Global Avg Loss: 0.72813653, Time: 0.0209 Steps: 83150, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000129, Sample Num: 2064, Cur Loss: 0.22450624, Cur Avg Loss: 0.18126360, Log Avg loss: 0.23057071, Global Avg Loss: 0.72807670, Time: 0.0210 Steps: 83160, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000139, Sample Num: 2224, Cur Loss: 0.17688930, Cur Avg Loss: 0.18389882, Log Avg loss: 0.21789322, Global Avg Loss: 0.72801536, Time: 0.0209 Steps: 83170, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000149, Sample Num: 2384, Cur Loss: 0.21999271, Cur Avg Loss: 0.18125167, Log Avg loss: 0.14445627, Global Avg Loss: 0.72794520, Time: 0.0209 Steps: 83180, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000159, Sample Num: 2544, Cur Loss: 0.17594725, Cur Avg Loss: 0.17899888, Log Avg loss: 0.14543235, Global Avg Loss: 0.72787518, Time: 0.0209 Steps: 83190, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000169, Sample Num: 2704, Cur Loss: 0.10131860, Cur Avg Loss: 0.17483115, Log Avg loss: 0.10856426, Global Avg Loss: 0.72780074, Time: 0.0209 Steps: 83200, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000179, Sample Num: 2864, Cur Loss: 0.13321073, Cur Avg Loss: 0.17689081, Log Avg loss: 0.21169901, Global Avg Loss: 0.72773872, Time: 0.0209 Steps: 83210, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000189, Sample Num: 3024, Cur Loss: 0.31915510, Cur Avg Loss: 0.17497866, Log Avg loss: 0.14075116, Global Avg Loss: 0.72766818, Time: 0.0209 Steps: 83220, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000199, Sample Num: 3184, Cur Loss: 0.26043907, Cur Avg Loss: 0.17722192, Log Avg loss: 0.21961947, Global Avg Loss: 0.72760714, Time: 0.0208 Steps: 83230, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000209, Sample Num: 3344, Cur Loss: 0.07637893, Cur Avg Loss: 0.17559149, Log Avg loss: 0.14314609, Global Avg Loss: 0.72753693, Time: 0.0209 Steps: 83240, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000219, Sample Num: 3504, Cur Loss: 0.37047064, Cur Avg Loss: 0.17925212, Log Avg loss: 0.25575923, Global Avg Loss: 0.72748026, Time: 0.0208 Steps: 83250, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000229, Sample Num: 3664, Cur Loss: 0.07617015, Cur Avg Loss: 0.17919183, Log Avg loss: 0.17787140, Global Avg Loss: 0.72741425, Time: 0.0210 Steps: 83260, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000239, Sample Num: 3824, Cur Loss: 0.10820465, Cur Avg Loss: 0.17731187, Log Avg loss: 0.13426081, Global Avg Loss: 0.72734302, Time: 0.0209 Steps: 83270, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000249, Sample Num: 3984, Cur Loss: 0.09829009, Cur Avg Loss: 0.17604497, Log Avg loss: 0.14576599, Global Avg Loss: 0.72727318, Time: 0.0209 Steps: 83280, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000259, Sample Num: 4144, Cur Loss: 0.03729573, Cur Avg Loss: 0.17425484, Log Avg loss: 0.12968057, Global Avg Loss: 0.72720143, Time: 0.0257 Steps: 83290, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000269, Sample Num: 4304, Cur Loss: 0.08814123, Cur Avg Loss: 0.17357249, Log Avg loss: 0.15589986, Global Avg Loss: 0.72713285, Time: 0.0220 Steps: 83300, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000279, Sample Num: 4464, Cur Loss: 0.14633848, Cur Avg Loss: 0.17319494, Log Avg loss: 0.16303880, Global Avg Loss: 0.72706514, Time: 0.0212 Steps: 83310, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000289, Sample Num: 4624, Cur Loss: 0.13278769, Cur Avg Loss: 0.17307673, Log Avg loss: 0.16977865, Global Avg Loss: 0.72699825, Time: 0.0212 Steps: 83320, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000299, Sample Num: 4784, Cur Loss: 0.17256808, Cur Avg Loss: 0.17288970, Log Avg loss: 0.16748452, Global Avg Loss: 0.72693111, Time: 0.0220 Steps: 83330, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000309, Sample Num: 4944, Cur Loss: 0.22964747, Cur Avg Loss: 0.17249686, Log Avg loss: 0.16075106, Global Avg Loss: 0.72686317, Time: 0.0212 Steps: 83340, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000319, Sample Num: 5104, Cur Loss: 0.01562364, Cur Avg Loss: 0.17303353, Log Avg loss: 0.18961651, Global Avg Loss: 0.72679872, Time: 0.0212 Steps: 83350, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000329, Sample Num: 5264, Cur Loss: 0.04686992, Cur Avg Loss: 0.17248870, Log Avg loss: 0.15510863, Global Avg Loss: 0.72673014, Time: 0.0220 Steps: 83360, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000339, Sample Num: 5424, Cur Loss: 0.14675933, Cur Avg Loss: 0.17385113, Log Avg loss: 0.21867487, Global Avg Loss: 0.72666920, Time: 0.0212 Steps: 83370, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000349, Sample Num: 5584, Cur Loss: 0.19821538, Cur Avg Loss: 0.17299237, Log Avg loss: 0.14388059, Global Avg Loss: 0.72659930, Time: 0.0220 Steps: 83380, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000359, Sample Num: 5744, Cur Loss: 0.08590080, Cur Avg Loss: 0.17141191, Log Avg loss: 0.11625389, Global Avg Loss: 0.72652611, Time: 0.0209 Steps: 83390, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000369, Sample Num: 5904, Cur Loss: 0.19924617, Cur Avg Loss: 0.17086969, Log Avg loss: 0.15140390, Global Avg Loss: 0.72645715, Time: 0.0212 Steps: 83400, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000379, Sample Num: 6064, Cur Loss: 0.23529175, Cur Avg Loss: 0.17052015, Log Avg loss: 0.15762208, Global Avg Loss: 0.72638895, Time: 0.0220 Steps: 83410, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000389, Sample Num: 6224, Cur Loss: 0.05607740, Cur Avg Loss: 0.16960163, Log Avg loss: 0.13478980, Global Avg Loss: 0.72631803, Time: 0.0212 Steps: 83420, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000399, Sample Num: 6384, Cur Loss: 0.16922086, Cur Avg Loss: 0.16932458, Log Avg loss: 0.15854718, Global Avg Loss: 0.72624998, Time: 0.0219 Steps: 83430, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000409, Sample Num: 6544, Cur Loss: 0.17169991, Cur Avg Loss: 0.16903995, Log Avg loss: 0.15768355, Global Avg Loss: 0.72618184, Time: 0.0220 Steps: 83440, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000419, Sample Num: 6704, Cur Loss: 0.14791909, Cur Avg Loss: 0.16880057, Log Avg loss: 0.15900964, Global Avg Loss: 0.72611387, Time: 0.0212 Steps: 83450, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000429, Sample Num: 6864, Cur Loss: 0.20456494, Cur Avg Loss: 0.16853039, Log Avg loss: 0.15721007, Global Avg Loss: 0.72604571, Time: 0.0220 Steps: 83460, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000439, Sample Num: 7024, Cur Loss: 0.23505974, Cur Avg Loss: 0.16737371, Log Avg loss: 0.11775202, Global Avg Loss: 0.72597283, Time: 0.0212 Steps: 83470, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000449, Sample Num: 7184, Cur Loss: 0.18447207, Cur Avg Loss: 0.16667986, Log Avg loss: 0.13621979, Global Avg Loss: 0.72590219, Time: 0.0220 Steps: 83480, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000459, Sample Num: 7344, Cur Loss: 0.32827330, Cur Avg Loss: 0.16670631, Log Avg loss: 0.16789382, Global Avg Loss: 0.72583535, Time: 0.0211 Steps: 83490, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000469, Sample Num: 7504, Cur Loss: 0.04131662, Cur Avg Loss: 0.16657922, Log Avg loss: 0.16074582, Global Avg Loss: 0.72576768, Time: 0.0212 Steps: 83500, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000479, Sample Num: 7664, Cur Loss: 0.27964324, Cur Avg Loss: 0.16688961, Log Avg loss: 0.18144685, Global Avg Loss: 0.72570250, Time: 0.0212 Steps: 83510, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000489, Sample Num: 7824, Cur Loss: 0.15125576, Cur Avg Loss: 0.16738821, Log Avg loss: 0.19127148, Global Avg Loss: 0.72563851, Time: 0.0212 Steps: 83520, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000499, Sample Num: 7984, Cur Loss: 0.25038821, Cur Avg Loss: 0.16733445, Log Avg loss: 0.16470563, Global Avg Loss: 0.72557135, Time: 0.0219 Steps: 83530, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000509, Sample Num: 8144, Cur Loss: 0.11379194, Cur Avg Loss: 0.16679911, Log Avg loss: 0.14008549, Global Avg Loss: 0.72550127, Time: 0.0220 Steps: 83540, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000519, Sample Num: 8304, Cur Loss: 0.10962169, Cur Avg Loss: 0.16667979, Log Avg loss: 0.16060626, Global Avg Loss: 0.72543366, Time: 0.0212 Steps: 83550, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000529, Sample Num: 8464, Cur Loss: 0.08029591, Cur Avg Loss: 0.16695817, Log Avg loss: 0.18140641, Global Avg Loss: 0.72536855, Time: 0.0209 Steps: 83560, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000539, Sample Num: 8624, Cur Loss: 0.02693348, Cur Avg Loss: 0.16692888, Log Avg loss: 0.16537918, Global Avg Loss: 0.72530154, Time: 0.0220 Steps: 83570, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000549, Sample Num: 8784, Cur Loss: 0.06105579, Cur Avg Loss: 0.16715991, Log Avg loss: 0.17961227, Global Avg Loss: 0.72523625, Time: 0.0209 Steps: 83580, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000559, Sample Num: 8944, Cur Loss: 0.12772429, Cur Avg Loss: 0.16660891, Log Avg loss: 0.13635937, Global Avg Loss: 0.72516581, Time: 0.0213 Steps: 83590, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000569, Sample Num: 9104, Cur Loss: 0.12313794, Cur Avg Loss: 0.16678829, Log Avg loss: 0.17681536, Global Avg Loss: 0.72510021, Time: 0.0210 Steps: 83600, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000579, Sample Num: 9264, Cur Loss: 0.19366615, Cur Avg Loss: 0.16630559, Log Avg loss: 0.13883991, Global Avg Loss: 0.72503010, Time: 0.0213 Steps: 83610, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000589, Sample Num: 9424, Cur Loss: 0.15971708, Cur Avg Loss: 0.16641002, Log Avg loss: 0.17245677, Global Avg Loss: 0.72496401, Time: 0.0220 Steps: 83620, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000599, Sample Num: 9584, Cur Loss: 0.11948754, Cur Avg Loss: 0.16656844, Log Avg loss: 0.17589917, Global Avg Loss: 0.72489836, Time: 0.0210 Steps: 83630, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000609, Sample Num: 9744, Cur Loss: 0.17773247, Cur Avg Loss: 0.16606719, Log Avg loss: 0.13604229, Global Avg Loss: 0.72482796, Time: 0.0213 Steps: 83640, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000619, Sample Num: 9904, Cur Loss: 0.25229865, Cur Avg Loss: 0.16585009, Log Avg loss: 0.15262909, Global Avg Loss: 0.72475955, Time: 0.0220 Steps: 83650, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000629, Sample Num: 10064, Cur Loss: 0.03677230, Cur Avg Loss: 0.16596460, Log Avg loss: 0.17305255, Global Avg Loss: 0.72469361, Time: 0.0209 Steps: 83660, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000639, Sample Num: 10224, Cur Loss: 0.03471838, Cur Avg Loss: 0.16586184, Log Avg loss: 0.15939819, Global Avg Loss: 0.72462604, Time: 0.0220 Steps: 83670, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000649, Sample Num: 10384, Cur Loss: 0.14065954, Cur Avg Loss: 0.16591935, Log Avg loss: 0.16959435, Global Avg Loss: 0.72455972, Time: 0.0209 Steps: 83680, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000659, Sample Num: 10544, Cur Loss: 0.11419906, Cur Avg Loss: 0.16660973, Log Avg loss: 0.21141535, Global Avg Loss: 0.72449840, Time: 0.0213 Steps: 83690, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000669, Sample Num: 10704, Cur Loss: 0.08294585, Cur Avg Loss: 0.16716683, Log Avg loss: 0.20387987, Global Avg Loss: 0.72443620, Time: 0.0219 Steps: 83700, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000679, Sample Num: 10864, Cur Loss: 0.20964345, Cur Avg Loss: 0.16627819, Log Avg loss: 0.10682829, Global Avg Loss: 0.72436242, Time: 0.0209 Steps: 83710, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000689, Sample Num: 11024, Cur Loss: 0.10872389, Cur Avg Loss: 0.16594797, Log Avg loss: 0.14352599, Global Avg Loss: 0.72429304, Time: 0.0212 Steps: 83720, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000699, Sample Num: 11184, Cur Loss: 0.13261747, Cur Avg Loss: 0.16506298, Log Avg loss: 0.10408719, Global Avg Loss: 0.72421897, Time: 0.0220 Steps: 83730, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000709, Sample Num: 11344, Cur Loss: 0.08347514, Cur Avg Loss: 0.16474135, Log Avg loss: 0.14225898, Global Avg Loss: 0.72414947, Time: 0.0210 Steps: 83740, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000719, Sample Num: 11504, Cur Loss: 0.25573173, Cur Avg Loss: 0.16576507, Log Avg loss: 0.23834726, Global Avg Loss: 0.72409147, Time: 0.0213 Steps: 83750, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000729, Sample Num: 11664, Cur Loss: 0.15573663, Cur Avg Loss: 0.16562069, Log Avg loss: 0.15523918, Global Avg Loss: 0.72402355, Time: 0.0221 Steps: 83760, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000739, Sample Num: 11824, Cur Loss: 0.22573525, Cur Avg Loss: 0.16574382, Log Avg loss: 0.17472028, Global Avg Loss: 0.72395798, Time: 0.0209 Steps: 83770, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000749, Sample Num: 11984, Cur Loss: 0.12169985, Cur Avg Loss: 0.16550063, Log Avg loss: 0.14752904, Global Avg Loss: 0.72388918, Time: 0.0213 Steps: 83780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000759, Sample Num: 12144, Cur Loss: 0.13608801, Cur Avg Loss: 0.16584576, Log Avg loss: 0.19169604, Global Avg Loss: 0.72382566, Time: 0.0220 Steps: 83790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000769, Sample Num: 12304, Cur Loss: 0.11458425, Cur Avg Loss: 0.16619454, Log Avg loss: 0.19266676, Global Avg Loss: 0.72376228, Time: 0.0246 Steps: 83800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000779, Sample Num: 12464, Cur Loss: 0.22839309, Cur Avg Loss: 0.16632466, Log Avg loss: 0.17633115, Global Avg Loss: 0.72369696, Time: 0.0210 Steps: 83810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000789, Sample Num: 12624, Cur Loss: 0.20027338, Cur Avg Loss: 0.16660751, Log Avg loss: 0.18864093, Global Avg Loss: 0.72363313, Time: 0.0209 Steps: 83820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000799, Sample Num: 12784, Cur Loss: 0.28307253, Cur Avg Loss: 0.16724131, Log Avg loss: 0.21724876, Global Avg Loss: 0.72357272, Time: 0.0210 Steps: 83830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000809, Sample Num: 12944, Cur Loss: 0.07693779, Cur Avg Loss: 0.16727329, Log Avg loss: 0.16982787, Global Avg Loss: 0.72350667, Time: 0.0210 Steps: 83840, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000819, Sample Num: 13104, Cur Loss: 0.33396143, Cur Avg Loss: 0.16776026, Log Avg loss: 0.20715666, Global Avg Loss: 0.72344509, Time: 0.0208 Steps: 83850, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000829, Sample Num: 13264, Cur Loss: 0.04428699, Cur Avg Loss: 0.16732205, Log Avg loss: 0.13143240, Global Avg Loss: 0.72337450, Time: 0.0211 Steps: 83860, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000839, Sample Num: 13424, Cur Loss: 0.17237467, Cur Avg Loss: 0.16784424, Log Avg loss: 0.21113359, Global Avg Loss: 0.72331342, Time: 0.0210 Steps: 83870, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000849, Sample Num: 13584, Cur Loss: 0.03911971, Cur Avg Loss: 0.16765878, Log Avg loss: 0.15209919, Global Avg Loss: 0.72324532, Time: 0.0210 Steps: 83880, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000859, Sample Num: 13744, Cur Loss: 0.11334071, Cur Avg Loss: 0.16772122, Log Avg loss: 0.17302205, Global Avg Loss: 0.72317973, Time: 0.0210 Steps: 83890, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000869, Sample Num: 13904, Cur Loss: 0.13292035, Cur Avg Loss: 0.16754075, Log Avg loss: 0.15203817, Global Avg Loss: 0.72311166, Time: 0.0209 Steps: 83900, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000879, Sample Num: 14064, Cur Loss: 0.19133350, Cur Avg Loss: 0.16818099, Log Avg loss: 0.22381812, Global Avg Loss: 0.72305216, Time: 0.0210 Steps: 83910, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000889, Sample Num: 14224, Cur Loss: 0.12393863, Cur Avg Loss: 0.16785341, Log Avg loss: 0.13905905, Global Avg Loss: 0.72298257, Time: 0.0210 Steps: 83920, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000899, Sample Num: 14384, Cur Loss: 0.44070256, Cur Avg Loss: 0.16764466, Log Avg loss: 0.14908713, Global Avg Loss: 0.72291419, Time: 0.0209 Steps: 83930, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000909, Sample Num: 14544, Cur Loss: 0.17596236, Cur Avg Loss: 0.16701875, Log Avg loss: 0.11074936, Global Avg Loss: 0.72284126, Time: 0.0210 Steps: 83940, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000919, Sample Num: 14704, Cur Loss: 0.12515935, Cur Avg Loss: 0.16730565, Log Avg loss: 0.19338481, Global Avg Loss: 0.72277819, Time: 0.0209 Steps: 83950, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000929, Sample Num: 14864, Cur Loss: 0.20367974, Cur Avg Loss: 0.16739313, Log Avg loss: 0.17543226, Global Avg Loss: 0.72271300, Time: 0.0210 Steps: 83960, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000939, Sample Num: 15024, Cur Loss: 0.17520730, Cur Avg Loss: 0.16706973, Log Avg loss: 0.13702636, Global Avg Loss: 0.72264325, Time: 0.0210 Steps: 83970, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000949, Sample Num: 15184, Cur Loss: 0.11141083, Cur Avg Loss: 0.16700829, Log Avg loss: 0.16123909, Global Avg Loss: 0.72257640, Time: 0.0209 Steps: 83980, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000959, Sample Num: 15344, Cur Loss: 0.03658579, Cur Avg Loss: 0.16682528, Log Avg loss: 0.14945716, Global Avg Loss: 0.72250817, Time: 0.0210 Steps: 83990, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000969, Sample Num: 15504, Cur Loss: 0.09902886, Cur Avg Loss: 0.16625918, Log Avg loss: 0.11196988, Global Avg Loss: 0.72243548, Time: 0.0208 Steps: 84000, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000979, Sample Num: 15664, Cur Loss: 0.12308840, Cur Avg Loss: 0.16638525, Log Avg loss: 0.17860144, Global Avg Loss: 0.72237075, Time: 0.0210 Steps: 84010, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000989, Sample Num: 15824, Cur Loss: 0.54106528, Cur Avg Loss: 0.16657829, Log Avg loss: 0.18547733, Global Avg Loss: 0.72230685, Time: 0.0210 Steps: 84020, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000999, Sample Num: 15984, Cur Loss: 0.08524969, Cur Avg Loss: 0.16604980, Log Avg loss: 0.11378212, Global Avg Loss: 0.72223443, Time: 0.0210 Steps: 84030, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001009, Sample Num: 16144, Cur Loss: 0.11401863, Cur Avg Loss: 0.16566063, Log Avg loss: 0.12678254, Global Avg Loss: 0.72216358, Time: 0.0209 Steps: 84040, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001019, Sample Num: 16304, Cur Loss: 0.40737134, Cur Avg Loss: 0.16566303, Log Avg loss: 0.16590519, Global Avg Loss: 0.72209739, Time: 0.0210 Steps: 84050, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001029, Sample Num: 16464, Cur Loss: 0.11702384, Cur Avg Loss: 0.16550361, Log Avg loss: 0.14925852, Global Avg Loss: 0.72202925, Time: 0.0209 Steps: 84060, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001039, Sample Num: 16624, Cur Loss: 0.42695254, Cur Avg Loss: 0.16557526, Log Avg loss: 0.17294870, Global Avg Loss: 0.72196394, Time: 0.0209 Steps: 84070, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001049, Sample Num: 16784, Cur Loss: 0.04081598, Cur Avg Loss: 0.16516114, Log Avg loss: 0.12213389, Global Avg Loss: 0.72189260, Time: 0.0209 Steps: 84080, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001059, Sample Num: 16944, Cur Loss: 0.06162226, Cur Avg Loss: 0.16486927, Log Avg loss: 0.13425166, Global Avg Loss: 0.72182271, Time: 0.0209 Steps: 84090, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001069, Sample Num: 17104, Cur Loss: 0.30511862, Cur Avg Loss: 0.16450614, Log Avg loss: 0.12605061, Global Avg Loss: 0.72175187, Time: 0.0209 Steps: 84100, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001079, Sample Num: 17264, Cur Loss: 0.08982472, Cur Avg Loss: 0.16501886, Log Avg loss: 0.21982935, Global Avg Loss: 0.72169220, Time: 0.0209 Steps: 84110, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001089, Sample Num: 17424, Cur Loss: 0.17164657, Cur Avg Loss: 0.16542677, Log Avg loss: 0.20943947, Global Avg Loss: 0.72163130, Time: 0.0209 Steps: 84120, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001099, Sample Num: 17584, Cur Loss: 0.08987957, Cur Avg Loss: 0.16526016, Log Avg loss: 0.14711702, Global Avg Loss: 0.72156301, Time: 0.0209 Steps: 84130, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001109, Sample Num: 17744, Cur Loss: 0.21735395, Cur Avg Loss: 0.16544436, Log Avg loss: 0.18568790, Global Avg Loss: 0.72149932, Time: 0.0209 Steps: 84140, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001119, Sample Num: 17904, Cur Loss: 0.13072968, Cur Avg Loss: 0.16551372, Log Avg loss: 0.17320565, Global Avg Loss: 0.72143417, Time: 0.0209 Steps: 84150, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001129, Sample Num: 18064, Cur Loss: 0.13558742, Cur Avg Loss: 0.16588799, Log Avg loss: 0.20776808, Global Avg Loss: 0.72137313, Time: 0.0209 Steps: 84160, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001139, Sample Num: 18224, Cur Loss: 0.13798313, Cur Avg Loss: 0.16603392, Log Avg loss: 0.18250981, Global Avg Loss: 0.72130911, Time: 0.0209 Steps: 84170, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001149, Sample Num: 18384, Cur Loss: 0.03321092, Cur Avg Loss: 0.16582599, Log Avg loss: 0.14214232, Global Avg Loss: 0.72124031, Time: 0.0209 Steps: 84180, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001159, Sample Num: 18544, Cur Loss: 0.18518418, Cur Avg Loss: 0.16622597, Log Avg loss: 0.21218407, Global Avg Loss: 0.72117985, Time: 0.0208 Steps: 84190, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001169, Sample Num: 18704, Cur Loss: 0.16722056, Cur Avg Loss: 0.16639375, Log Avg loss: 0.18583922, Global Avg Loss: 0.72111627, Time: 0.0209 Steps: 84200, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001179, Sample Num: 18864, Cur Loss: 0.27289563, Cur Avg Loss: 0.16621232, Log Avg loss: 0.14500367, Global Avg Loss: 0.72104785, Time: 0.0209 Steps: 84210, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001189, Sample Num: 19024, Cur Loss: 0.24482049, Cur Avg Loss: 0.16669031, Log Avg loss: 0.22304509, Global Avg Loss: 0.72098872, Time: 0.0209 Steps: 84220, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001199, Sample Num: 19184, Cur Loss: 0.17475338, Cur Avg Loss: 0.16642640, Log Avg loss: 0.13504749, Global Avg Loss: 0.72091916, Time: 0.0209 Steps: 84230, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001209, Sample Num: 19344, Cur Loss: 0.04453237, Cur Avg Loss: 0.16635781, Log Avg loss: 0.15813437, Global Avg Loss: 0.72085235, Time: 0.0209 Steps: 84240, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001219, Sample Num: 19504, Cur Loss: 0.33670086, Cur Avg Loss: 0.16650397, Log Avg loss: 0.18417446, Global Avg Loss: 0.72078865, Time: 0.0209 Steps: 84250, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001229, Sample Num: 19664, Cur Loss: 0.20591381, Cur Avg Loss: 0.16687957, Log Avg loss: 0.21266442, Global Avg Loss: 0.72072834, Time: 0.0209 Steps: 84260, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001239, Sample Num: 19824, Cur Loss: 0.11271178, Cur Avg Loss: 0.16700468, Log Avg loss: 0.18238169, Global Avg Loss: 0.72066446, Time: 0.0209 Steps: 84270, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001249, Sample Num: 19984, Cur Loss: 0.03919045, Cur Avg Loss: 0.16654270, Log Avg loss: 0.10930283, Global Avg Loss: 0.72059192, Time: 0.0209 Steps: 84280, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001259, Sample Num: 20144, Cur Loss: 0.08529177, Cur Avg Loss: 0.16630484, Log Avg loss: 0.13659629, Global Avg Loss: 0.72052264, Time: 0.0209 Steps: 84290, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001269, Sample Num: 20304, Cur Loss: 0.13233244, Cur Avg Loss: 0.16635884, Log Avg loss: 0.17315715, Global Avg Loss: 0.72045771, Time: 0.0209 Steps: 84300, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001279, Sample Num: 20464, Cur Loss: 0.04084162, Cur Avg Loss: 0.16698004, Log Avg loss: 0.24581074, Global Avg Loss: 0.72040141, Time: 0.0209 Steps: 84310, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001289, Sample Num: 20624, Cur Loss: 0.09029189, Cur Avg Loss: 0.16681314, Log Avg loss: 0.14546669, Global Avg Loss: 0.72033322, Time: 0.0209 Steps: 84320, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001299, Sample Num: 20784, Cur Loss: 0.10316385, Cur Avg Loss: 0.16688909, Log Avg loss: 0.17667857, Global Avg Loss: 0.72026876, Time: 0.0209 Steps: 84330, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001309, Sample Num: 20944, Cur Loss: 0.13957801, Cur Avg Loss: 0.16690882, Log Avg loss: 0.16947214, Global Avg Loss: 0.72020345, Time: 0.0209 Steps: 84340, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001319, Sample Num: 21104, Cur Loss: 0.12976448, Cur Avg Loss: 0.16681344, Log Avg loss: 0.15432778, Global Avg Loss: 0.72013636, Time: 0.0209 Steps: 84350, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001329, Sample Num: 21264, Cur Loss: 0.04976430, Cur Avg Loss: 0.16656914, Log Avg loss: 0.13434599, Global Avg Loss: 0.72006692, Time: 0.0209 Steps: 84360, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001339, Sample Num: 21424, Cur Loss: 0.10761674, Cur Avg Loss: 0.16612167, Log Avg loss: 0.10665284, Global Avg Loss: 0.71999422, Time: 0.0209 Steps: 84370, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001349, Sample Num: 21584, Cur Loss: 0.07027823, Cur Avg Loss: 0.16579319, Log Avg loss: 0.12180964, Global Avg Loss: 0.71992333, Time: 0.0209 Steps: 84380, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001359, Sample Num: 21744, Cur Loss: 0.15443873, Cur Avg Loss: 0.16579608, Log Avg loss: 0.16618603, Global Avg Loss: 0.71985771, Time: 0.0209 Steps: 84390, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001369, Sample Num: 21904, Cur Loss: 0.21501836, Cur Avg Loss: 0.16566750, Log Avg loss: 0.14819412, Global Avg Loss: 0.71978998, Time: 0.0209 Steps: 84400, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001379, Sample Num: 22064, Cur Loss: 0.23477885, Cur Avg Loss: 0.16593860, Log Avg loss: 0.20305141, Global Avg Loss: 0.71972876, Time: 0.0209 Steps: 84410, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001389, Sample Num: 22224, Cur Loss: 0.43228674, Cur Avg Loss: 0.16636485, Log Avg loss: 0.22514501, Global Avg Loss: 0.71967017, Time: 0.0210 Steps: 84420, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001399, Sample Num: 22384, Cur Loss: 0.12829943, Cur Avg Loss: 0.16685260, Log Avg loss: 0.23460079, Global Avg Loss: 0.71961272, Time: 0.0209 Steps: 84430, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001409, Sample Num: 22544, Cur Loss: 0.05083252, Cur Avg Loss: 0.16648079, Log Avg loss: 0.11446507, Global Avg Loss: 0.71954106, Time: 0.0209 Steps: 84440, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001419, Sample Num: 22704, Cur Loss: 0.11368518, Cur Avg Loss: 0.16677562, Log Avg loss: 0.20831713, Global Avg Loss: 0.71948052, Time: 0.0209 Steps: 84450, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001429, Sample Num: 22864, Cur Loss: 0.05483303, Cur Avg Loss: 0.16681910, Log Avg loss: 0.17298956, Global Avg Loss: 0.71941582, Time: 0.0209 Steps: 84460, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001439, Sample Num: 23024, Cur Loss: 0.10324311, Cur Avg Loss: 0.16718631, Log Avg loss: 0.21965999, Global Avg Loss: 0.71935665, Time: 0.0209 Steps: 84470, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001449, Sample Num: 23184, Cur Loss: 0.04550073, Cur Avg Loss: 0.16681857, Log Avg loss: 0.11390121, Global Avg Loss: 0.71928498, Time: 0.0209 Steps: 84480, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001459, Sample Num: 23344, Cur Loss: 0.09308489, Cur Avg Loss: 0.16683757, Log Avg loss: 0.16959005, Global Avg Loss: 0.71921992, Time: 0.0209 Steps: 84490, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001469, Sample Num: 23504, Cur Loss: 0.31615338, Cur Avg Loss: 0.16691907, Log Avg loss: 0.17881003, Global Avg Loss: 0.71915597, Time: 0.0209 Steps: 84500, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001479, Sample Num: 23664, Cur Loss: 0.11625268, Cur Avg Loss: 0.16642535, Log Avg loss: 0.09389815, Global Avg Loss: 0.71908198, Time: 0.0209 Steps: 84510, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001489, Sample Num: 23824, Cur Loss: 0.27277565, Cur Avg Loss: 0.16662342, Log Avg loss: 0.19591774, Global Avg Loss: 0.71902009, Time: 0.0210 Steps: 84520, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001499, Sample Num: 23984, Cur Loss: 0.34110871, Cur Avg Loss: 0.16709605, Log Avg loss: 0.23747110, Global Avg Loss: 0.71896312, Time: 0.0209 Steps: 84530, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001509, Sample Num: 24144, Cur Loss: 0.10952392, Cur Avg Loss: 0.16707530, Log Avg loss: 0.16396396, Global Avg Loss: 0.71889747, Time: 0.0209 Steps: 84540, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001519, Sample Num: 24304, Cur Loss: 0.45773101, Cur Avg Loss: 0.16708299, Log Avg loss: 0.16824321, Global Avg Loss: 0.71883234, Time: 0.0209 Steps: 84550, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001529, Sample Num: 24464, Cur Loss: 0.13744180, Cur Avg Loss: 0.16713732, Log Avg loss: 0.17539096, Global Avg Loss: 0.71876807, Time: 0.0209 Steps: 84560, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001539, Sample Num: 24624, Cur Loss: 0.11170743, Cur Avg Loss: 0.16688449, Log Avg loss: 0.12822735, Global Avg Loss: 0.71869825, Time: 0.0246 Steps: 84570, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001549, Sample Num: 24784, Cur Loss: 0.14877468, Cur Avg Loss: 0.16682578, Log Avg loss: 0.15778962, Global Avg Loss: 0.71863193, Time: 0.0209 Steps: 84580, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001559, Sample Num: 24944, Cur Loss: 0.06430884, Cur Avg Loss: 0.16669025, Log Avg loss: 0.14569723, Global Avg Loss: 0.71856420, Time: 0.0209 Steps: 84590, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001569, Sample Num: 25104, Cur Loss: 0.13640575, Cur Avg Loss: 0.16653520, Log Avg loss: 0.14236210, Global Avg Loss: 0.71849609, Time: 0.0209 Steps: 84600, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001579, Sample Num: 25264, Cur Loss: 0.13930991, Cur Avg Loss: 0.16681630, Log Avg loss: 0.21092117, Global Avg Loss: 0.71843610, Time: 0.0209 Steps: 84610, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001589, Sample Num: 25424, Cur Loss: 0.13461179, Cur Avg Loss: 0.16682860, Log Avg loss: 0.16877018, Global Avg Loss: 0.71837114, Time: 0.0209 Steps: 84620, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001599, Sample Num: 25584, Cur Loss: 0.11698049, Cur Avg Loss: 0.16653274, Log Avg loss: 0.11952157, Global Avg Loss: 0.71830038, Time: 0.0209 Steps: 84630, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001609, Sample Num: 25744, Cur Loss: 0.35726672, Cur Avg Loss: 0.16657255, Log Avg loss: 0.17293793, Global Avg Loss: 0.71823595, Time: 0.0209 Steps: 84640, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001619, Sample Num: 25904, Cur Loss: 0.24294560, Cur Avg Loss: 0.16697931, Log Avg loss: 0.23242681, Global Avg Loss: 0.71817856, Time: 0.0209 Steps: 84650, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001629, Sample Num: 26064, Cur Loss: 0.24354076, Cur Avg Loss: 0.16698474, Log Avg loss: 0.16786304, Global Avg Loss: 0.71811355, Time: 0.0210 Steps: 84660, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001639, Sample Num: 26224, Cur Loss: 0.15766355, Cur Avg Loss: 0.16702180, Log Avg loss: 0.17305937, Global Avg Loss: 0.71804918, Time: 0.0209 Steps: 84670, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001649, Sample Num: 26384, Cur Loss: 0.13370623, Cur Avg Loss: 0.16684559, Log Avg loss: 0.13796495, Global Avg Loss: 0.71798068, Time: 0.0209 Steps: 84680, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001659, Sample Num: 26544, Cur Loss: 0.13286649, Cur Avg Loss: 0.16693178, Log Avg loss: 0.18114501, Global Avg Loss: 0.71791729, Time: 0.0209 Steps: 84690, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001669, Sample Num: 26704, Cur Loss: 0.25011775, Cur Avg Loss: 0.16650063, Log Avg loss: 0.09497262, Global Avg Loss: 0.71784374, Time: 0.0209 Steps: 84700, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001679, Sample Num: 26864, Cur Loss: 0.23965795, Cur Avg Loss: 0.16678771, Log Avg loss: 0.21470183, Global Avg Loss: 0.71778435, Time: 0.0209 Steps: 84710, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001689, Sample Num: 27024, Cur Loss: 0.29724979, Cur Avg Loss: 0.16690756, Log Avg loss: 0.18702975, Global Avg Loss: 0.71772170, Time: 0.0210 Steps: 84720, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001699, Sample Num: 27184, Cur Loss: 0.17514877, Cur Avg Loss: 0.16712324, Log Avg loss: 0.20355105, Global Avg Loss: 0.71766101, Time: 0.0209 Steps: 84730, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001709, Sample Num: 27344, Cur Loss: 0.36758327, Cur Avg Loss: 0.16684110, Log Avg loss: 0.11890656, Global Avg Loss: 0.71759036, Time: 0.0209 Steps: 84740, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001719, Sample Num: 27504, Cur Loss: 0.06937265, Cur Avg Loss: 0.16678816, Log Avg loss: 0.15773956, Global Avg Loss: 0.71752430, Time: 0.0209 Steps: 84750, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001729, Sample Num: 27664, Cur Loss: 0.09722541, Cur Avg Loss: 0.16671589, Log Avg loss: 0.15429367, Global Avg Loss: 0.71745785, Time: 0.0208 Steps: 84760, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001739, Sample Num: 27824, Cur Loss: 0.10257621, Cur Avg Loss: 0.16673334, Log Avg loss: 0.16975033, Global Avg Loss: 0.71739324, Time: 0.0208 Steps: 84770, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001749, Sample Num: 27984, Cur Loss: 0.10529318, Cur Avg Loss: 0.16680407, Log Avg loss: 0.17910406, Global Avg Loss: 0.71732974, Time: 0.0208 Steps: 84780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001759, Sample Num: 28144, Cur Loss: 0.14114995, Cur Avg Loss: 0.16667768, Log Avg loss: 0.14457090, Global Avg Loss: 0.71726219, Time: 0.0208 Steps: 84790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001769, Sample Num: 28304, Cur Loss: 0.23317620, Cur Avg Loss: 0.16667399, Log Avg loss: 0.16602525, Global Avg Loss: 0.71719719, Time: 0.0208 Steps: 84800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001779, Sample Num: 28464, Cur Loss: 0.16263504, Cur Avg Loss: 0.16665648, Log Avg loss: 0.16355914, Global Avg Loss: 0.71713191, Time: 0.0208 Steps: 84810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001789, Sample Num: 28624, Cur Loss: 0.29900536, Cur Avg Loss: 0.16647309, Log Avg loss: 0.13384906, Global Avg Loss: 0.71706314, Time: 0.0209 Steps: 84820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001799, Sample Num: 28784, Cur Loss: 0.17754644, Cur Avg Loss: 0.16660014, Log Avg loss: 0.18932791, Global Avg Loss: 0.71700093, Time: 0.0209 Steps: 84830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001809, Sample Num: 28944, Cur Loss: 0.34582293, Cur Avg Loss: 0.16706480, Log Avg loss: 0.25065779, Global Avg Loss: 0.71694596, Time: 0.0208 Steps: 84840, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001819, Sample Num: 29104, Cur Loss: 0.11222080, Cur Avg Loss: 0.16699559, Log Avg loss: 0.15447533, Global Avg Loss: 0.71687967, Time: 0.0208 Steps: 84850, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001829, Sample Num: 29264, Cur Loss: 0.20151868, Cur Avg Loss: 0.16713138, Log Avg loss: 0.19183144, Global Avg Loss: 0.71681780, Time: 0.0208 Steps: 84860, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001839, Sample Num: 29424, Cur Loss: 0.12011300, Cur Avg Loss: 0.16710993, Log Avg loss: 0.16318746, Global Avg Loss: 0.71675257, Time: 0.0208 Steps: 84870, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001849, Sample Num: 29584, Cur Loss: 0.18817131, Cur Avg Loss: 0.16712524, Log Avg loss: 0.16994135, Global Avg Loss: 0.71668815, Time: 0.0208 Steps: 84880, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001859, Sample Num: 29744, Cur Loss: 0.25311720, Cur Avg Loss: 0.16747873, Log Avg loss: 0.23283754, Global Avg Loss: 0.71663115, Time: 0.0208 Steps: 84890, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001869, Sample Num: 29904, Cur Loss: 0.12005278, Cur Avg Loss: 0.16763781, Log Avg loss: 0.19721187, Global Avg Loss: 0.71656997, Time: 0.0209 Steps: 84900, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001879, Sample Num: 30064, Cur Loss: 0.22041124, Cur Avg Loss: 0.16786198, Log Avg loss: 0.20975962, Global Avg Loss: 0.71651028, Time: 0.0208 Steps: 84910, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001889, Sample Num: 30224, Cur Loss: 0.13851282, Cur Avg Loss: 0.16792964, Log Avg loss: 0.18064223, Global Avg Loss: 0.71644718, Time: 0.0208 Steps: 84920, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001899, Sample Num: 30384, Cur Loss: 0.08855157, Cur Avg Loss: 0.16774745, Log Avg loss: 0.13333234, Global Avg Loss: 0.71637852, Time: 0.0208 Steps: 84930, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001909, Sample Num: 30544, Cur Loss: 0.35772479, Cur Avg Loss: 0.16805180, Log Avg loss: 0.22584745, Global Avg Loss: 0.71632077, Time: 0.0209 Steps: 84940, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001919, Sample Num: 30704, Cur Loss: 0.04397157, Cur Avg Loss: 0.16828608, Log Avg loss: 0.21301033, Global Avg Loss: 0.71626152, Time: 0.0208 Steps: 84950, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001929, Sample Num: 30864, Cur Loss: 0.11596903, Cur Avg Loss: 0.16803150, Log Avg loss: 0.11917628, Global Avg Loss: 0.71619124, Time: 0.0209 Steps: 84960, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001939, Sample Num: 31024, Cur Loss: 0.27644086, Cur Avg Loss: 0.16814022, Log Avg loss: 0.18911366, Global Avg Loss: 0.71612921, Time: 0.0208 Steps: 84970, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001949, Sample Num: 31184, Cur Loss: 0.30779666, Cur Avg Loss: 0.16815353, Log Avg loss: 0.17073326, Global Avg Loss: 0.71606503, Time: 0.0209 Steps: 84980, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001959, Sample Num: 31344, Cur Loss: 0.02587763, Cur Avg Loss: 0.16825938, Log Avg loss: 0.18889108, Global Avg Loss: 0.71600301, Time: 0.0208 Steps: 84990, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001969, Sample Num: 31504, Cur Loss: 0.06726093, Cur Avg Loss: 0.16824539, Log Avg loss: 0.16550305, Global Avg Loss: 0.71593824, Time: 0.0209 Steps: 85000, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001979, Sample Num: 31664, Cur Loss: 0.06303068, Cur Avg Loss: 0.16800602, Log Avg loss: 0.12087497, Global Avg Loss: 0.71586824, Time: 0.0208 Steps: 85010, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001989, Sample Num: 31824, Cur Loss: 0.07569965, Cur Avg Loss: 0.16788062, Log Avg loss: 0.14306353, Global Avg Loss: 0.71580087, Time: 0.0208 Steps: 85020, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001999, Sample Num: 31984, Cur Loss: 0.30786180, Cur Avg Loss: 0.16775799, Log Avg loss: 0.14336782, Global Avg Loss: 0.71573355, Time: 0.0210 Steps: 85030, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002009, Sample Num: 32144, Cur Loss: 0.11683958, Cur Avg Loss: 0.16775965, Log Avg loss: 0.16809080, Global Avg Loss: 0.71566915, Time: 0.0208 Steps: 85040, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002019, Sample Num: 32304, Cur Loss: 0.17709510, Cur Avg Loss: 0.16765744, Log Avg loss: 0.14712319, Global Avg Loss: 0.71560230, Time: 0.0208 Steps: 85050, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002029, Sample Num: 32464, Cur Loss: 0.25631705, Cur Avg Loss: 0.16799590, Log Avg loss: 0.23633192, Global Avg Loss: 0.71554596, Time: 0.0209 Steps: 85060, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002039, Sample Num: 32624, Cur Loss: 0.09597404, Cur Avg Loss: 0.16809516, Log Avg loss: 0.18823424, Global Avg Loss: 0.71548397, Time: 0.0208 Steps: 85070, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002049, Sample Num: 32784, Cur Loss: 0.01832194, Cur Avg Loss: 0.16846055, Log Avg loss: 0.24296358, Global Avg Loss: 0.71542843, Time: 0.0246 Steps: 85080, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002059, Sample Num: 32944, Cur Loss: 0.05286024, Cur Avg Loss: 0.16837852, Log Avg loss: 0.15156972, Global Avg Loss: 0.71536217, Time: 0.0210 Steps: 85090, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002069, Sample Num: 33104, Cur Loss: 0.10875802, Cur Avg Loss: 0.16806881, Log Avg loss: 0.10430019, Global Avg Loss: 0.71529036, Time: 0.0209 Steps: 85100, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002079, Sample Num: 33264, Cur Loss: 0.08457392, Cur Avg Loss: 0.16812446, Log Avg loss: 0.17963942, Global Avg Loss: 0.71522742, Time: 0.0210 Steps: 85110, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002089, Sample Num: 33424, Cur Loss: 0.12829158, Cur Avg Loss: 0.16772421, Log Avg loss: 0.08451210, Global Avg Loss: 0.71515333, Time: 0.0209 Steps: 85120, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002099, Sample Num: 33584, Cur Loss: 0.09145863, Cur Avg Loss: 0.16775596, Log Avg loss: 0.17438724, Global Avg Loss: 0.71508981, Time: 0.0209 Steps: 85130, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002109, Sample Num: 33744, Cur Loss: 0.29315621, Cur Avg Loss: 0.16786504, Log Avg loss: 0.19076141, Global Avg Loss: 0.71502822, Time: 0.0209 Steps: 85140, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002119, Sample Num: 33904, Cur Loss: 0.11913097, Cur Avg Loss: 0.16766869, Log Avg loss: 0.12625892, Global Avg Loss: 0.71495908, Time: 0.0210 Steps: 85150, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002129, Sample Num: 34055, Cur Loss: 0.34616688, Cur Avg Loss: 0.16786334, Log Avg loss: 0.20910866, Global Avg Loss: 0.71489968, Time: 0.0100 Steps: 85160, Updated lr: 0.000020 ***** Running evaluation checkpoint-85160 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-85160 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.114867, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.239206, "eval_total_loss": 168.161529, "eval_mae": 0.346392, "eval_mse": 0.239286, "eval_r2": 0.847894, "eval_sp_statistic": 0.89778, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.92218, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.590761, "test_total_loss": 296.561888, "test_mae": 0.556122, "test_mse": 0.590854, "test_r2": 0.618658, "test_sp_statistic": 0.797129, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.831975, "test_ps_pvalue": 0.0, "lr": 2.0189663347558086e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7148996759607523, "train_cur_epoch_loss": 357.3810418397188, "train_cur_epoch_avg_loss": 0.16786333576313706, "train_cur_epoch_time": 45.11486744880676, "train_cur_epoch_avg_time": 0.02119063759925165, "epoch": 40, "step": 85160} ################################################## Training, Epoch: 0041, Batch: 000010, Sample Num: 160, Cur Loss: 0.18167329, Cur Avg Loss: 0.24745135, Log Avg loss: 0.24745135, Global Avg Loss: 0.71484479, Time: 0.0209 Steps: 85170, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000020, Sample Num: 320, Cur Loss: 0.27016890, Cur Avg Loss: 0.18955261, Log Avg loss: 0.13165387, Global Avg Loss: 0.71477633, Time: 0.0209 Steps: 85180, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000030, Sample Num: 480, Cur Loss: 0.14424220, Cur Avg Loss: 0.18296507, Log Avg loss: 0.16978999, Global Avg Loss: 0.71471235, Time: 0.0209 Steps: 85190, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000040, Sample Num: 640, Cur Loss: 0.12041283, Cur Avg Loss: 0.18435457, Log Avg loss: 0.18852306, Global Avg Loss: 0.71465059, Time: 0.0208 Steps: 85200, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000050, Sample Num: 800, Cur Loss: 0.04795052, Cur Avg Loss: 0.18459670, Log Avg loss: 0.18556524, Global Avg Loss: 0.71458850, Time: 0.0209 Steps: 85210, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000060, Sample Num: 960, Cur Loss: 0.31952339, Cur Avg Loss: 0.18973247, Log Avg loss: 0.21541130, Global Avg Loss: 0.71452993, Time: 0.0208 Steps: 85220, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000070, Sample Num: 1120, Cur Loss: 0.13961461, Cur Avg Loss: 0.18245704, Log Avg loss: 0.13880444, Global Avg Loss: 0.71446238, Time: 0.0209 Steps: 85230, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000080, Sample Num: 1280, Cur Loss: 0.04606946, Cur Avg Loss: 0.17816960, Log Avg loss: 0.14815758, Global Avg Loss: 0.71439594, Time: 0.0209 Steps: 85240, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000090, Sample Num: 1440, Cur Loss: 0.20803706, Cur Avg Loss: 0.17680483, Log Avg loss: 0.16588662, Global Avg Loss: 0.71433160, Time: 0.0208 Steps: 85250, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000100, Sample Num: 1600, Cur Loss: 0.22671856, Cur Avg Loss: 0.17728013, Log Avg loss: 0.18155783, Global Avg Loss: 0.71426911, Time: 0.0209 Steps: 85260, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000110, Sample Num: 1760, Cur Loss: 0.03591781, Cur Avg Loss: 0.16933314, Log Avg loss: 0.08986320, Global Avg Loss: 0.71419588, Time: 0.0209 Steps: 85270, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000120, Sample Num: 1920, Cur Loss: 0.14013159, Cur Avg Loss: 0.16573927, Log Avg loss: 0.12620680, Global Avg Loss: 0.71412694, Time: 0.0208 Steps: 85280, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000130, Sample Num: 2080, Cur Loss: 0.04098110, Cur Avg Loss: 0.16124510, Log Avg loss: 0.10731495, Global Avg Loss: 0.71405579, Time: 0.0208 Steps: 85290, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000140, Sample Num: 2240, Cur Loss: 0.34468213, Cur Avg Loss: 0.16202122, Log Avg loss: 0.17211080, Global Avg Loss: 0.71399226, Time: 0.0209 Steps: 85300, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000150, Sample Num: 2400, Cur Loss: 0.09855081, Cur Avg Loss: 0.15824732, Log Avg loss: 0.10541271, Global Avg Loss: 0.71392092, Time: 0.0209 Steps: 85310, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000160, Sample Num: 2560, Cur Loss: 0.10320250, Cur Avg Loss: 0.15742287, Log Avg loss: 0.14505611, Global Avg Loss: 0.71385424, Time: 0.0208 Steps: 85320, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000170, Sample Num: 2720, Cur Loss: 0.28164259, Cur Avg Loss: 0.15652613, Log Avg loss: 0.14217828, Global Avg Loss: 0.71378725, Time: 0.0209 Steps: 85330, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000180, Sample Num: 2880, Cur Loss: 0.39830744, Cur Avg Loss: 0.15585461, Log Avg loss: 0.14443886, Global Avg Loss: 0.71372053, Time: 0.0209 Steps: 85340, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000190, Sample Num: 3040, Cur Loss: 0.04388428, Cur Avg Loss: 0.15394751, Log Avg loss: 0.11961964, Global Avg Loss: 0.71365092, Time: 0.0209 Steps: 85350, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000200, Sample Num: 3200, Cur Loss: 0.06093404, Cur Avg Loss: 0.15220876, Log Avg loss: 0.11917266, Global Avg Loss: 0.71358128, Time: 0.0209 Steps: 85360, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000210, Sample Num: 3360, Cur Loss: 0.03605800, Cur Avg Loss: 0.15575368, Log Avg loss: 0.22665196, Global Avg Loss: 0.71352424, Time: 0.0209 Steps: 85370, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000220, Sample Num: 3520, Cur Loss: 0.29306233, Cur Avg Loss: 0.15644248, Log Avg loss: 0.17090723, Global Avg Loss: 0.71346069, Time: 0.0209 Steps: 85380, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000230, Sample Num: 3680, Cur Loss: 0.20599931, Cur Avg Loss: 0.15614263, Log Avg loss: 0.14954605, Global Avg Loss: 0.71339465, Time: 0.0209 Steps: 85390, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000240, Sample Num: 3840, Cur Loss: 0.12431446, Cur Avg Loss: 0.15744089, Log Avg loss: 0.18730089, Global Avg Loss: 0.71333305, Time: 0.0209 Steps: 85400, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000250, Sample Num: 4000, Cur Loss: 0.05189323, Cur Avg Loss: 0.15700351, Log Avg loss: 0.14650643, Global Avg Loss: 0.71326668, Time: 0.0208 Steps: 85410, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000260, Sample Num: 4160, Cur Loss: 0.16277128, Cur Avg Loss: 0.15620926, Log Avg loss: 0.13635281, Global Avg Loss: 0.71319914, Time: 0.0246 Steps: 85420, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000270, Sample Num: 4320, Cur Loss: 0.17192873, Cur Avg Loss: 0.15707008, Log Avg loss: 0.17945158, Global Avg Loss: 0.71313667, Time: 0.0210 Steps: 85430, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000280, Sample Num: 4480, Cur Loss: 0.08241237, Cur Avg Loss: 0.15615382, Log Avg loss: 0.13141467, Global Avg Loss: 0.71306858, Time: 0.0208 Steps: 85440, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000290, Sample Num: 4640, Cur Loss: 0.29436365, Cur Avg Loss: 0.15682312, Log Avg loss: 0.17556370, Global Avg Loss: 0.71300568, Time: 0.0208 Steps: 85450, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000300, Sample Num: 4800, Cur Loss: 0.14705317, Cur Avg Loss: 0.15658160, Log Avg loss: 0.14957748, Global Avg Loss: 0.71293975, Time: 0.0210 Steps: 85460, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000310, Sample Num: 4960, Cur Loss: 0.04797095, Cur Avg Loss: 0.15685064, Log Avg loss: 0.16492172, Global Avg Loss: 0.71287563, Time: 0.0210 Steps: 85470, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000320, Sample Num: 5120, Cur Loss: 0.12347102, Cur Avg Loss: 0.15689754, Log Avg loss: 0.15835153, Global Avg Loss: 0.71281076, Time: 0.0209 Steps: 85480, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000330, Sample Num: 5280, Cur Loss: 0.16440298, Cur Avg Loss: 0.15537178, Log Avg loss: 0.10654756, Global Avg Loss: 0.71273984, Time: 0.0210 Steps: 85490, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000340, Sample Num: 5440, Cur Loss: 0.06478648, Cur Avg Loss: 0.15532362, Log Avg loss: 0.15373424, Global Avg Loss: 0.71267446, Time: 0.0209 Steps: 85500, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000350, Sample Num: 5600, Cur Loss: 0.15916446, Cur Avg Loss: 0.15581442, Log Avg loss: 0.17250150, Global Avg Loss: 0.71261129, Time: 0.0208 Steps: 85510, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000360, Sample Num: 5760, Cur Loss: 0.13541925, Cur Avg Loss: 0.15576380, Log Avg loss: 0.15399227, Global Avg Loss: 0.71254597, Time: 0.0208 Steps: 85520, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000370, Sample Num: 5920, Cur Loss: 0.05443512, Cur Avg Loss: 0.15656696, Log Avg loss: 0.18548065, Global Avg Loss: 0.71248435, Time: 0.0210 Steps: 85530, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000380, Sample Num: 6080, Cur Loss: 0.14783779, Cur Avg Loss: 0.15641104, Log Avg loss: 0.15064214, Global Avg Loss: 0.71241866, Time: 0.0208 Steps: 85540, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000390, Sample Num: 6240, Cur Loss: 0.09553942, Cur Avg Loss: 0.15543919, Log Avg loss: 0.11850876, Global Avg Loss: 0.71234924, Time: 0.0208 Steps: 85550, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000400, Sample Num: 6400, Cur Loss: 0.06569727, Cur Avg Loss: 0.15478433, Log Avg loss: 0.12924483, Global Avg Loss: 0.71228109, Time: 0.0210 Steps: 85560, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000410, Sample Num: 6560, Cur Loss: 0.38918370, Cur Avg Loss: 0.15536751, Log Avg loss: 0.17869481, Global Avg Loss: 0.71221873, Time: 0.0209 Steps: 85570, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000420, Sample Num: 6720, Cur Loss: 0.12971918, Cur Avg Loss: 0.15419971, Log Avg loss: 0.10631989, Global Avg Loss: 0.71214794, Time: 0.0210 Steps: 85580, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000430, Sample Num: 6880, Cur Loss: 0.14715838, Cur Avg Loss: 0.15416984, Log Avg loss: 0.15291527, Global Avg Loss: 0.71208260, Time: 0.0208 Steps: 85590, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000440, Sample Num: 7040, Cur Loss: 0.05067605, Cur Avg Loss: 0.15462306, Log Avg loss: 0.17411116, Global Avg Loss: 0.71201975, Time: 0.0209 Steps: 85600, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000450, Sample Num: 7200, Cur Loss: 0.26232162, Cur Avg Loss: 0.15596968, Log Avg loss: 0.21522118, Global Avg Loss: 0.71196172, Time: 0.0208 Steps: 85610, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000460, Sample Num: 7360, Cur Loss: 0.07905947, Cur Avg Loss: 0.15586434, Log Avg loss: 0.15112385, Global Avg Loss: 0.71189622, Time: 0.0210 Steps: 85620, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000470, Sample Num: 7520, Cur Loss: 0.46316856, Cur Avg Loss: 0.15682209, Log Avg loss: 0.20087862, Global Avg Loss: 0.71183654, Time: 0.0209 Steps: 85630, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000480, Sample Num: 7680, Cur Loss: 0.09333338, Cur Avg Loss: 0.15717335, Log Avg loss: 0.17368249, Global Avg Loss: 0.71177370, Time: 0.0209 Steps: 85640, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000490, Sample Num: 7840, Cur Loss: 0.07342964, Cur Avg Loss: 0.15651314, Log Avg loss: 0.12482348, Global Avg Loss: 0.71170517, Time: 0.0209 Steps: 85650, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000500, Sample Num: 8000, Cur Loss: 0.24927604, Cur Avg Loss: 0.15610792, Log Avg loss: 0.13625214, Global Avg Loss: 0.71163799, Time: 0.0208 Steps: 85660, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000510, Sample Num: 8160, Cur Loss: 0.13939932, Cur Avg Loss: 0.15546652, Log Avg loss: 0.12339617, Global Avg Loss: 0.71156933, Time: 0.0209 Steps: 85670, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000520, Sample Num: 8320, Cur Loss: 0.04411158, Cur Avg Loss: 0.15538433, Log Avg loss: 0.15119285, Global Avg Loss: 0.71150392, Time: 0.0209 Steps: 85680, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000530, Sample Num: 8480, Cur Loss: 0.38592905, Cur Avg Loss: 0.15563043, Log Avg loss: 0.16842768, Global Avg Loss: 0.71144055, Time: 0.0209 Steps: 85690, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000540, Sample Num: 8640, Cur Loss: 0.08656792, Cur Avg Loss: 0.15523499, Log Avg loss: 0.13427667, Global Avg Loss: 0.71137320, Time: 0.0209 Steps: 85700, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000550, Sample Num: 8800, Cur Loss: 0.32265636, Cur Avg Loss: 0.15548311, Log Avg loss: 0.16888166, Global Avg Loss: 0.71130991, Time: 0.0209 Steps: 85710, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000560, Sample Num: 8960, Cur Loss: 0.16923393, Cur Avg Loss: 0.15551736, Log Avg loss: 0.15740121, Global Avg Loss: 0.71124529, Time: 0.0209 Steps: 85720, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000570, Sample Num: 9120, Cur Loss: 0.10775350, Cur Avg Loss: 0.15617916, Log Avg loss: 0.19323982, Global Avg Loss: 0.71118487, Time: 0.0208 Steps: 85730, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000580, Sample Num: 9280, Cur Loss: 0.25465554, Cur Avg Loss: 0.15612971, Log Avg loss: 0.15331066, Global Avg Loss: 0.71111980, Time: 0.0209 Steps: 85740, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000590, Sample Num: 9440, Cur Loss: 0.11711424, Cur Avg Loss: 0.15670147, Log Avg loss: 0.18986409, Global Avg Loss: 0.71105901, Time: 0.0209 Steps: 85750, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000600, Sample Num: 9600, Cur Loss: 0.12590273, Cur Avg Loss: 0.15573297, Log Avg loss: 0.09859143, Global Avg Loss: 0.71098760, Time: 0.0208 Steps: 85760, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000610, Sample Num: 9760, Cur Loss: 0.27030084, Cur Avg Loss: 0.15523977, Log Avg loss: 0.12564727, Global Avg Loss: 0.71091935, Time: 0.0209 Steps: 85770, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000620, Sample Num: 9920, Cur Loss: 0.14207378, Cur Avg Loss: 0.15574944, Log Avg loss: 0.18683944, Global Avg Loss: 0.71085825, Time: 0.0209 Steps: 85780, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000630, Sample Num: 10080, Cur Loss: 0.10234877, Cur Avg Loss: 0.15768104, Log Avg loss: 0.27744025, Global Avg Loss: 0.71080773, Time: 0.0209 Steps: 85790, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000640, Sample Num: 10240, Cur Loss: 0.09373310, Cur Avg Loss: 0.15865774, Log Avg loss: 0.22019027, Global Avg Loss: 0.71075055, Time: 0.0209 Steps: 85800, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000650, Sample Num: 10400, Cur Loss: 0.09998884, Cur Avg Loss: 0.15821901, Log Avg loss: 0.13014001, Global Avg Loss: 0.71068289, Time: 0.0208 Steps: 85810, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000660, Sample Num: 10560, Cur Loss: 0.22606374, Cur Avg Loss: 0.15771993, Log Avg loss: 0.12527989, Global Avg Loss: 0.71061468, Time: 0.0209 Steps: 85820, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000670, Sample Num: 10720, Cur Loss: 0.46778178, Cur Avg Loss: 0.15782378, Log Avg loss: 0.16467743, Global Avg Loss: 0.71055107, Time: 0.0210 Steps: 85830, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000680, Sample Num: 10880, Cur Loss: 0.20688665, Cur Avg Loss: 0.15716994, Log Avg loss: 0.11336282, Global Avg Loss: 0.71048150, Time: 0.0211 Steps: 85840, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000690, Sample Num: 11040, Cur Loss: 0.15994734, Cur Avg Loss: 0.15706949, Log Avg loss: 0.15023904, Global Avg Loss: 0.71041624, Time: 0.0210 Steps: 85850, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000700, Sample Num: 11200, Cur Loss: 0.42915455, Cur Avg Loss: 0.15729075, Log Avg loss: 0.17255739, Global Avg Loss: 0.71035360, Time: 0.0211 Steps: 85860, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000710, Sample Num: 11360, Cur Loss: 0.18629256, Cur Avg Loss: 0.15718692, Log Avg loss: 0.14991941, Global Avg Loss: 0.71028833, Time: 0.0211 Steps: 85870, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000720, Sample Num: 11520, Cur Loss: 0.15069968, Cur Avg Loss: 0.15708782, Log Avg loss: 0.15005124, Global Avg Loss: 0.71022310, Time: 0.0211 Steps: 85880, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000730, Sample Num: 11680, Cur Loss: 0.21383911, Cur Avg Loss: 0.15703950, Log Avg loss: 0.15356051, Global Avg Loss: 0.71015829, Time: 0.0211 Steps: 85890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000740, Sample Num: 11840, Cur Loss: 0.05702440, Cur Avg Loss: 0.15685319, Log Avg loss: 0.14325292, Global Avg Loss: 0.71009229, Time: 0.0211 Steps: 85900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000750, Sample Num: 12000, Cur Loss: 0.07772236, Cur Avg Loss: 0.15759666, Log Avg loss: 0.21261297, Global Avg Loss: 0.71003438, Time: 0.0210 Steps: 85910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000760, Sample Num: 12160, Cur Loss: 0.12260194, Cur Avg Loss: 0.15737405, Log Avg loss: 0.14067841, Global Avg Loss: 0.70996812, Time: 0.0211 Steps: 85920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000770, Sample Num: 12320, Cur Loss: 0.05564271, Cur Avg Loss: 0.15757347, Log Avg loss: 0.17272916, Global Avg Loss: 0.70990560, Time: 0.0251 Steps: 85930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000780, Sample Num: 12480, Cur Loss: 0.13735011, Cur Avg Loss: 0.15768824, Log Avg loss: 0.16652617, Global Avg Loss: 0.70984237, Time: 0.0212 Steps: 85940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000790, Sample Num: 12640, Cur Loss: 0.17393816, Cur Avg Loss: 0.15728811, Log Avg loss: 0.12607782, Global Avg Loss: 0.70977445, Time: 0.0212 Steps: 85950, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000800, Sample Num: 12800, Cur Loss: 0.05855716, Cur Avg Loss: 0.15726312, Log Avg loss: 0.15528853, Global Avg Loss: 0.70970995, Time: 0.0212 Steps: 85960, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000810, Sample Num: 12960, Cur Loss: 0.14555407, Cur Avg Loss: 0.15682120, Log Avg loss: 0.12146765, Global Avg Loss: 0.70964152, Time: 0.0212 Steps: 85970, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000820, Sample Num: 13120, Cur Loss: 0.24561930, Cur Avg Loss: 0.15655226, Log Avg loss: 0.13476801, Global Avg Loss: 0.70957466, Time: 0.0212 Steps: 85980, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000830, Sample Num: 13280, Cur Loss: 0.07419825, Cur Avg Loss: 0.15649273, Log Avg loss: 0.15161163, Global Avg Loss: 0.70950977, Time: 0.0212 Steps: 85990, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000840, Sample Num: 13440, Cur Loss: 0.19042540, Cur Avg Loss: 0.15614716, Log Avg loss: 0.12746449, Global Avg Loss: 0.70944209, Time: 0.0212 Steps: 86000, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000850, Sample Num: 13600, Cur Loss: 0.06278011, Cur Avg Loss: 0.15646037, Log Avg loss: 0.18277036, Global Avg Loss: 0.70938086, Time: 0.0212 Steps: 86010, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000860, Sample Num: 13760, Cur Loss: 0.07526228, Cur Avg Loss: 0.15617258, Log Avg loss: 0.13170999, Global Avg Loss: 0.70931370, Time: 0.0213 Steps: 86020, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000870, Sample Num: 13920, Cur Loss: 0.05879601, Cur Avg Loss: 0.15551174, Log Avg loss: 0.09867975, Global Avg Loss: 0.70924272, Time: 0.0212 Steps: 86030, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000880, Sample Num: 14080, Cur Loss: 0.13112481, Cur Avg Loss: 0.15556929, Log Avg loss: 0.16057590, Global Avg Loss: 0.70917896, Time: 0.0212 Steps: 86040, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000890, Sample Num: 14240, Cur Loss: 0.05772465, Cur Avg Loss: 0.15572476, Log Avg loss: 0.16940649, Global Avg Loss: 0.70911623, Time: 0.0213 Steps: 86050, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000900, Sample Num: 14400, Cur Loss: 0.11278398, Cur Avg Loss: 0.15571639, Log Avg loss: 0.15497185, Global Avg Loss: 0.70905184, Time: 0.0212 Steps: 86060, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000910, Sample Num: 14560, Cur Loss: 0.10431486, Cur Avg Loss: 0.15565457, Log Avg loss: 0.15009045, Global Avg Loss: 0.70898690, Time: 0.0212 Steps: 86070, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000920, Sample Num: 14720, Cur Loss: 0.19890845, Cur Avg Loss: 0.15598895, Log Avg loss: 0.18641752, Global Avg Loss: 0.70892619, Time: 0.0212 Steps: 86080, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000930, Sample Num: 14880, Cur Loss: 0.08914507, Cur Avg Loss: 0.15637984, Log Avg loss: 0.19234212, Global Avg Loss: 0.70886618, Time: 0.0212 Steps: 86090, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000940, Sample Num: 15040, Cur Loss: 0.14681736, Cur Avg Loss: 0.15667128, Log Avg loss: 0.18377433, Global Avg Loss: 0.70880520, Time: 0.0213 Steps: 86100, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000950, Sample Num: 15200, Cur Loss: 0.26290107, Cur Avg Loss: 0.15658416, Log Avg loss: 0.14839575, Global Avg Loss: 0.70874012, Time: 0.0212 Steps: 86110, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000960, Sample Num: 15360, Cur Loss: 0.13839997, Cur Avg Loss: 0.15646367, Log Avg loss: 0.14501669, Global Avg Loss: 0.70867466, Time: 0.0213 Steps: 86120, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000970, Sample Num: 15520, Cur Loss: 0.08571618, Cur Avg Loss: 0.15652884, Log Avg loss: 0.16278560, Global Avg Loss: 0.70861128, Time: 0.0212 Steps: 86130, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000980, Sample Num: 15680, Cur Loss: 0.21069786, Cur Avg Loss: 0.15714247, Log Avg loss: 0.21666369, Global Avg Loss: 0.70855417, Time: 0.0212 Steps: 86140, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000990, Sample Num: 15840, Cur Loss: 0.15994336, Cur Avg Loss: 0.15740228, Log Avg loss: 0.18286405, Global Avg Loss: 0.70849315, Time: 0.0212 Steps: 86150, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001000, Sample Num: 16000, Cur Loss: 0.08097446, Cur Avg Loss: 0.15725678, Log Avg loss: 0.14285253, Global Avg Loss: 0.70842750, Time: 0.0212 Steps: 86160, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001010, Sample Num: 16160, Cur Loss: 0.11467366, Cur Avg Loss: 0.15723451, Log Avg loss: 0.15500680, Global Avg Loss: 0.70836327, Time: 0.0212 Steps: 86170, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001020, Sample Num: 16320, Cur Loss: 0.39531195, Cur Avg Loss: 0.15783732, Log Avg loss: 0.21872132, Global Avg Loss: 0.70830646, Time: 0.0212 Steps: 86180, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001030, Sample Num: 16480, Cur Loss: 0.13980803, Cur Avg Loss: 0.15872947, Log Avg loss: 0.24972884, Global Avg Loss: 0.70825325, Time: 0.0212 Steps: 86190, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001040, Sample Num: 16640, Cur Loss: 0.10559097, Cur Avg Loss: 0.15862477, Log Avg loss: 0.14784119, Global Avg Loss: 0.70818824, Time: 0.0212 Steps: 86200, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001050, Sample Num: 16800, Cur Loss: 0.10178050, Cur Avg Loss: 0.15852072, Log Avg loss: 0.14769901, Global Avg Loss: 0.70812322, Time: 0.0211 Steps: 86210, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001060, Sample Num: 16960, Cur Loss: 0.48281050, Cur Avg Loss: 0.15859646, Log Avg loss: 0.16654940, Global Avg Loss: 0.70806041, Time: 0.0211 Steps: 86220, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001070, Sample Num: 17120, Cur Loss: 0.30514741, Cur Avg Loss: 0.15887725, Log Avg loss: 0.18864071, Global Avg Loss: 0.70800017, Time: 0.0211 Steps: 86230, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001080, Sample Num: 17280, Cur Loss: 0.04378928, Cur Avg Loss: 0.15921741, Log Avg loss: 0.19561435, Global Avg Loss: 0.70794076, Time: 0.0210 Steps: 86240, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001090, Sample Num: 17440, Cur Loss: 0.55659842, Cur Avg Loss: 0.16028259, Log Avg loss: 0.27532207, Global Avg Loss: 0.70789060, Time: 0.0211 Steps: 86250, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001100, Sample Num: 17600, Cur Loss: 0.06346127, Cur Avg Loss: 0.16083774, Log Avg loss: 0.22134910, Global Avg Loss: 0.70783420, Time: 0.0211 Steps: 86260, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001110, Sample Num: 17760, Cur Loss: 0.22054125, Cur Avg Loss: 0.16054193, Log Avg loss: 0.12800283, Global Avg Loss: 0.70776699, Time: 0.0211 Steps: 86270, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001120, Sample Num: 17920, Cur Loss: 0.12236909, Cur Avg Loss: 0.16043969, Log Avg loss: 0.14909151, Global Avg Loss: 0.70770224, Time: 0.0211 Steps: 86280, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001130, Sample Num: 18080, Cur Loss: 0.18318473, Cur Avg Loss: 0.16094286, Log Avg loss: 0.21729795, Global Avg Loss: 0.70764540, Time: 0.0211 Steps: 86290, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001140, Sample Num: 18240, Cur Loss: 0.10232571, Cur Avg Loss: 0.16114855, Log Avg loss: 0.18439085, Global Avg Loss: 0.70758477, Time: 0.0211 Steps: 86300, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001150, Sample Num: 18400, Cur Loss: 0.20112821, Cur Avg Loss: 0.16107462, Log Avg loss: 0.15264731, Global Avg Loss: 0.70752048, Time: 0.0212 Steps: 86310, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001160, Sample Num: 18560, Cur Loss: 0.20663297, Cur Avg Loss: 0.16119188, Log Avg loss: 0.17467646, Global Avg Loss: 0.70745875, Time: 0.0211 Steps: 86320, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001170, Sample Num: 18720, Cur Loss: 0.14282905, Cur Avg Loss: 0.16109674, Log Avg loss: 0.15006029, Global Avg Loss: 0.70739418, Time: 0.0211 Steps: 86330, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001180, Sample Num: 18880, Cur Loss: 0.04828170, Cur Avg Loss: 0.16069147, Log Avg loss: 0.11327554, Global Avg Loss: 0.70732537, Time: 0.0211 Steps: 86340, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001190, Sample Num: 19040, Cur Loss: 0.42997164, Cur Avg Loss: 0.16074836, Log Avg loss: 0.16746056, Global Avg Loss: 0.70726285, Time: 0.0211 Steps: 86350, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001200, Sample Num: 19200, Cur Loss: 0.07625553, Cur Avg Loss: 0.16077277, Log Avg loss: 0.16367814, Global Avg Loss: 0.70719990, Time: 0.0211 Steps: 86360, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001210, Sample Num: 19360, Cur Loss: 0.17425212, Cur Avg Loss: 0.16064308, Log Avg loss: 0.14508073, Global Avg Loss: 0.70713482, Time: 0.0212 Steps: 86370, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001220, Sample Num: 19520, Cur Loss: 0.02386024, Cur Avg Loss: 0.16112104, Log Avg loss: 0.21895316, Global Avg Loss: 0.70707831, Time: 0.0211 Steps: 86380, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001230, Sample Num: 19680, Cur Loss: 0.18682554, Cur Avg Loss: 0.16095364, Log Avg loss: 0.14053156, Global Avg Loss: 0.70701273, Time: 0.0211 Steps: 86390, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001240, Sample Num: 19840, Cur Loss: 0.08394471, Cur Avg Loss: 0.16096516, Log Avg loss: 0.16238195, Global Avg Loss: 0.70694969, Time: 0.0211 Steps: 86400, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001250, Sample Num: 20000, Cur Loss: 0.21225575, Cur Avg Loss: 0.16086037, Log Avg loss: 0.14786648, Global Avg Loss: 0.70688499, Time: 0.0211 Steps: 86410, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001260, Sample Num: 20160, Cur Loss: 0.02889189, Cur Avg Loss: 0.16054875, Log Avg loss: 0.12159660, Global Avg Loss: 0.70681726, Time: 0.0211 Steps: 86420, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001270, Sample Num: 20320, Cur Loss: 0.07714596, Cur Avg Loss: 0.16054988, Log Avg loss: 0.16069207, Global Avg Loss: 0.70675408, Time: 0.0211 Steps: 86430, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001280, Sample Num: 20480, Cur Loss: 0.14899369, Cur Avg Loss: 0.16063849, Log Avg loss: 0.17189156, Global Avg Loss: 0.70669220, Time: 0.0256 Steps: 86440, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001290, Sample Num: 20640, Cur Loss: 0.04554913, Cur Avg Loss: 0.16058822, Log Avg loss: 0.15415366, Global Avg Loss: 0.70662828, Time: 0.0213 Steps: 86450, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001300, Sample Num: 20800, Cur Loss: 0.51484442, Cur Avg Loss: 0.16088994, Log Avg loss: 0.19981237, Global Avg Loss: 0.70656967, Time: 0.0211 Steps: 86460, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001310, Sample Num: 20960, Cur Loss: 0.16179010, Cur Avg Loss: 0.16113915, Log Avg loss: 0.19353616, Global Avg Loss: 0.70651034, Time: 0.0211 Steps: 86470, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001320, Sample Num: 21120, Cur Loss: 0.17795458, Cur Avg Loss: 0.16140420, Log Avg loss: 0.19612583, Global Avg Loss: 0.70645132, Time: 0.0211 Steps: 86480, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001330, Sample Num: 21280, Cur Loss: 0.17303523, Cur Avg Loss: 0.16175242, Log Avg loss: 0.20771745, Global Avg Loss: 0.70639365, Time: 0.0211 Steps: 86490, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001340, Sample Num: 21440, Cur Loss: 0.06410632, Cur Avg Loss: 0.16157255, Log Avg loss: 0.13765003, Global Avg Loss: 0.70632790, Time: 0.0211 Steps: 86500, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001350, Sample Num: 21600, Cur Loss: 0.28810173, Cur Avg Loss: 0.16154992, Log Avg loss: 0.15851654, Global Avg Loss: 0.70626458, Time: 0.0211 Steps: 86510, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001360, Sample Num: 21760, Cur Loss: 0.12706931, Cur Avg Loss: 0.16134672, Log Avg loss: 0.13391470, Global Avg Loss: 0.70619843, Time: 0.0212 Steps: 86520, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001370, Sample Num: 21920, Cur Loss: 0.05448384, Cur Avg Loss: 0.16108139, Log Avg loss: 0.12499737, Global Avg Loss: 0.70613126, Time: 0.0211 Steps: 86530, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001380, Sample Num: 22080, Cur Loss: 0.04376611, Cur Avg Loss: 0.16100814, Log Avg loss: 0.15097324, Global Avg Loss: 0.70606711, Time: 0.0211 Steps: 86540, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001390, Sample Num: 22240, Cur Loss: 0.33283687, Cur Avg Loss: 0.16099714, Log Avg loss: 0.15947868, Global Avg Loss: 0.70600396, Time: 0.0211 Steps: 86550, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001400, Sample Num: 22400, Cur Loss: 0.22436580, Cur Avg Loss: 0.16121841, Log Avg loss: 0.19197505, Global Avg Loss: 0.70594457, Time: 0.0211 Steps: 86560, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001410, Sample Num: 22560, Cur Loss: 0.11970289, Cur Avg Loss: 0.16117936, Log Avg loss: 0.15571228, Global Avg Loss: 0.70588101, Time: 0.0211 Steps: 86570, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001420, Sample Num: 22720, Cur Loss: 0.18949425, Cur Avg Loss: 0.16114732, Log Avg loss: 0.15662921, Global Avg Loss: 0.70581757, Time: 0.0211 Steps: 86580, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001430, Sample Num: 22880, Cur Loss: 0.11448640, Cur Avg Loss: 0.16132509, Log Avg loss: 0.18656874, Global Avg Loss: 0.70575761, Time: 0.0212 Steps: 86590, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001440, Sample Num: 23040, Cur Loss: 0.31182447, Cur Avg Loss: 0.16124415, Log Avg loss: 0.14967008, Global Avg Loss: 0.70569339, Time: 0.0212 Steps: 86600, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001450, Sample Num: 23200, Cur Loss: 0.26027498, Cur Avg Loss: 0.16111868, Log Avg loss: 0.14305051, Global Avg Loss: 0.70562843, Time: 0.0211 Steps: 86610, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001460, Sample Num: 23360, Cur Loss: 0.21368739, Cur Avg Loss: 0.16132988, Log Avg loss: 0.19195355, Global Avg Loss: 0.70556913, Time: 0.0211 Steps: 86620, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001470, Sample Num: 23520, Cur Loss: 0.24263366, Cur Avg Loss: 0.16103996, Log Avg loss: 0.11871272, Global Avg Loss: 0.70550139, Time: 0.0211 Steps: 86630, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001480, Sample Num: 23680, Cur Loss: 0.18957736, Cur Avg Loss: 0.16088693, Log Avg loss: 0.13839103, Global Avg Loss: 0.70543593, Time: 0.0211 Steps: 86640, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001490, Sample Num: 23840, Cur Loss: 0.11332643, Cur Avg Loss: 0.16135858, Log Avg loss: 0.23116294, Global Avg Loss: 0.70538120, Time: 0.0211 Steps: 86650, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001500, Sample Num: 24000, Cur Loss: 0.27328801, Cur Avg Loss: 0.16152783, Log Avg loss: 0.18674525, Global Avg Loss: 0.70532135, Time: 0.0213 Steps: 86660, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001510, Sample Num: 24160, Cur Loss: 0.07877672, Cur Avg Loss: 0.16161695, Log Avg loss: 0.17498572, Global Avg Loss: 0.70526016, Time: 0.0211 Steps: 86670, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001520, Sample Num: 24320, Cur Loss: 0.04376012, Cur Avg Loss: 0.16162403, Log Avg loss: 0.16269319, Global Avg Loss: 0.70519756, Time: 0.0211 Steps: 86680, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001530, Sample Num: 24480, Cur Loss: 0.08177546, Cur Avg Loss: 0.16153407, Log Avg loss: 0.14786042, Global Avg Loss: 0.70513327, Time: 0.0211 Steps: 86690, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001540, Sample Num: 24640, Cur Loss: 0.06777647, Cur Avg Loss: 0.16142131, Log Avg loss: 0.14416886, Global Avg Loss: 0.70506857, Time: 0.0249 Steps: 86700, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001550, Sample Num: 24800, Cur Loss: 0.10670133, Cur Avg Loss: 0.16126434, Log Avg loss: 0.13709094, Global Avg Loss: 0.70500307, Time: 0.0211 Steps: 86710, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001560, Sample Num: 24960, Cur Loss: 0.36940271, Cur Avg Loss: 0.16153612, Log Avg loss: 0.20366108, Global Avg Loss: 0.70494526, Time: 0.0211 Steps: 86720, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001570, Sample Num: 25120, Cur Loss: 0.23229131, Cur Avg Loss: 0.16126696, Log Avg loss: 0.11927880, Global Avg Loss: 0.70487773, Time: 0.0211 Steps: 86730, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001580, Sample Num: 25280, Cur Loss: 0.08222567, Cur Avg Loss: 0.16106765, Log Avg loss: 0.12977549, Global Avg Loss: 0.70481143, Time: 0.0211 Steps: 86740, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001590, Sample Num: 25440, Cur Loss: 0.07616007, Cur Avg Loss: 0.16111673, Log Avg loss: 0.16887096, Global Avg Loss: 0.70474965, Time: 0.0211 Steps: 86750, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001600, Sample Num: 25600, Cur Loss: 0.26634839, Cur Avg Loss: 0.16094257, Log Avg loss: 0.13325216, Global Avg Loss: 0.70468378, Time: 0.0211 Steps: 86760, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001610, Sample Num: 25760, Cur Loss: 0.28976476, Cur Avg Loss: 0.16101783, Log Avg loss: 0.17305938, Global Avg Loss: 0.70462251, Time: 0.0211 Steps: 86770, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001620, Sample Num: 25920, Cur Loss: 0.06292120, Cur Avg Loss: 0.16111102, Log Avg loss: 0.17611446, Global Avg Loss: 0.70456161, Time: 0.0211 Steps: 86780, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001630, Sample Num: 26080, Cur Loss: 0.20834678, Cur Avg Loss: 0.16115542, Log Avg loss: 0.16834852, Global Avg Loss: 0.70449982, Time: 0.0211 Steps: 86790, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001640, Sample Num: 26240, Cur Loss: 0.06305131, Cur Avg Loss: 0.16125976, Log Avg loss: 0.17826649, Global Avg Loss: 0.70443920, Time: 0.0211 Steps: 86800, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001650, Sample Num: 26400, Cur Loss: 0.22742066, Cur Avg Loss: 0.16144006, Log Avg loss: 0.19100983, Global Avg Loss: 0.70438005, Time: 0.0211 Steps: 86810, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001660, Sample Num: 26560, Cur Loss: 0.15213591, Cur Avg Loss: 0.16128105, Log Avg loss: 0.13504461, Global Avg Loss: 0.70431448, Time: 0.0211 Steps: 86820, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001670, Sample Num: 26720, Cur Loss: 0.10326425, Cur Avg Loss: 0.16135363, Log Avg loss: 0.17340192, Global Avg Loss: 0.70425333, Time: 0.0211 Steps: 86830, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001680, Sample Num: 26880, Cur Loss: 0.15441240, Cur Avg Loss: 0.16127148, Log Avg loss: 0.14755185, Global Avg Loss: 0.70418923, Time: 0.0211 Steps: 86840, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001690, Sample Num: 27040, Cur Loss: 0.13573244, Cur Avg Loss: 0.16167071, Log Avg loss: 0.22874212, Global Avg Loss: 0.70413448, Time: 0.0211 Steps: 86850, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001700, Sample Num: 27200, Cur Loss: 0.25199655, Cur Avg Loss: 0.16210633, Log Avg loss: 0.23572559, Global Avg Loss: 0.70408056, Time: 0.0211 Steps: 86860, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001710, Sample Num: 27360, Cur Loss: 0.28419033, Cur Avg Loss: 0.16189225, Log Avg loss: 0.12549823, Global Avg Loss: 0.70401395, Time: 0.0211 Steps: 86870, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001720, Sample Num: 27520, Cur Loss: 0.16183603, Cur Avg Loss: 0.16202137, Log Avg loss: 0.18410035, Global Avg Loss: 0.70395411, Time: 0.0211 Steps: 86880, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001730, Sample Num: 27680, Cur Loss: 0.12569101, Cur Avg Loss: 0.16190752, Log Avg loss: 0.14232618, Global Avg Loss: 0.70388947, Time: 0.0211 Steps: 86890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001740, Sample Num: 27840, Cur Loss: 0.23283176, Cur Avg Loss: 0.16192479, Log Avg loss: 0.16491231, Global Avg Loss: 0.70382745, Time: 0.0211 Steps: 86900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001750, Sample Num: 28000, Cur Loss: 0.07839641, Cur Avg Loss: 0.16190778, Log Avg loss: 0.15894787, Global Avg Loss: 0.70376476, Time: 0.0212 Steps: 86910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001760, Sample Num: 28160, Cur Loss: 0.23175642, Cur Avg Loss: 0.16171606, Log Avg loss: 0.12816587, Global Avg Loss: 0.70369854, Time: 0.0211 Steps: 86920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001770, Sample Num: 28320, Cur Loss: 0.34744114, Cur Avg Loss: 0.16174667, Log Avg loss: 0.16713389, Global Avg Loss: 0.70363681, Time: 0.0210 Steps: 86930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001780, Sample Num: 28480, Cur Loss: 0.19425415, Cur Avg Loss: 0.16176865, Log Avg loss: 0.16565846, Global Avg Loss: 0.70357493, Time: 0.0210 Steps: 86940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001790, Sample Num: 28640, Cur Loss: 0.15980439, Cur Avg Loss: 0.16177965, Log Avg loss: 0.16373734, Global Avg Loss: 0.70351285, Time: 0.0211 Steps: 86950, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001800, Sample Num: 28800, Cur Loss: 0.28049994, Cur Avg Loss: 0.16202824, Log Avg loss: 0.20652721, Global Avg Loss: 0.70345570, Time: 0.0211 Steps: 86960, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001810, Sample Num: 28960, Cur Loss: 0.10134277, Cur Avg Loss: 0.16183211, Log Avg loss: 0.12652803, Global Avg Loss: 0.70338936, Time: 0.0211 Steps: 86970, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001820, Sample Num: 29120, Cur Loss: 0.12762174, Cur Avg Loss: 0.16181739, Log Avg loss: 0.15915206, Global Avg Loss: 0.70332679, Time: 0.0210 Steps: 86980, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001830, Sample Num: 29280, Cur Loss: 0.03902593, Cur Avg Loss: 0.16189946, Log Avg loss: 0.17683715, Global Avg Loss: 0.70326627, Time: 0.0211 Steps: 86990, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001840, Sample Num: 29440, Cur Loss: 0.09806926, Cur Avg Loss: 0.16185789, Log Avg loss: 0.15425125, Global Avg Loss: 0.70320316, Time: 0.0212 Steps: 87000, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001850, Sample Num: 29600, Cur Loss: 0.12051310, Cur Avg Loss: 0.16167037, Log Avg loss: 0.12716523, Global Avg Loss: 0.70313696, Time: 0.0211 Steps: 87010, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001860, Sample Num: 29760, Cur Loss: 0.16139874, Cur Avg Loss: 0.16168931, Log Avg loss: 0.16519344, Global Avg Loss: 0.70307514, Time: 0.0211 Steps: 87020, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001870, Sample Num: 29920, Cur Loss: 0.32606795, Cur Avg Loss: 0.16182134, Log Avg loss: 0.18637858, Global Avg Loss: 0.70301577, Time: 0.0210 Steps: 87030, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001880, Sample Num: 30080, Cur Loss: 0.03392841, Cur Avg Loss: 0.16174251, Log Avg loss: 0.14700129, Global Avg Loss: 0.70295189, Time: 0.0210 Steps: 87040, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001890, Sample Num: 30240, Cur Loss: 0.18822163, Cur Avg Loss: 0.16181350, Log Avg loss: 0.17516073, Global Avg Loss: 0.70289126, Time: 0.0211 Steps: 87050, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001900, Sample Num: 30400, Cur Loss: 0.19335383, Cur Avg Loss: 0.16216648, Log Avg loss: 0.22887911, Global Avg Loss: 0.70283681, Time: 0.0211 Steps: 87060, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001910, Sample Num: 30560, Cur Loss: 0.41249412, Cur Avg Loss: 0.16221235, Log Avg loss: 0.17092696, Global Avg Loss: 0.70277572, Time: 0.0210 Steps: 87070, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001920, Sample Num: 30720, Cur Loss: 0.13134095, Cur Avg Loss: 0.16211344, Log Avg loss: 0.14322341, Global Avg Loss: 0.70271146, Time: 0.0211 Steps: 87080, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001930, Sample Num: 30880, Cur Loss: 0.26164883, Cur Avg Loss: 0.16244079, Log Avg loss: 0.22529088, Global Avg Loss: 0.70265664, Time: 0.0210 Steps: 87090, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001940, Sample Num: 31040, Cur Loss: 0.18037936, Cur Avg Loss: 0.16258185, Log Avg loss: 0.18980749, Global Avg Loss: 0.70259776, Time: 0.0210 Steps: 87100, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001950, Sample Num: 31200, Cur Loss: 0.14823249, Cur Avg Loss: 0.16273330, Log Avg loss: 0.19211325, Global Avg Loss: 0.70253916, Time: 0.0211 Steps: 87110, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001960, Sample Num: 31360, Cur Loss: 0.27753776, Cur Avg Loss: 0.16303885, Log Avg loss: 0.22262122, Global Avg Loss: 0.70248407, Time: 0.0210 Steps: 87120, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001970, Sample Num: 31520, Cur Loss: 0.29867807, Cur Avg Loss: 0.16290675, Log Avg loss: 0.13701544, Global Avg Loss: 0.70241917, Time: 0.0210 Steps: 87130, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001980, Sample Num: 31680, Cur Loss: 0.24983644, Cur Avg Loss: 0.16327335, Log Avg loss: 0.23549441, Global Avg Loss: 0.70236559, Time: 0.0210 Steps: 87140, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001990, Sample Num: 31840, Cur Loss: 0.06293669, Cur Avg Loss: 0.16328413, Log Avg loss: 0.16541707, Global Avg Loss: 0.70230398, Time: 0.0210 Steps: 87150, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002000, Sample Num: 32000, Cur Loss: 0.40734965, Cur Avg Loss: 0.16335309, Log Avg loss: 0.17707627, Global Avg Loss: 0.70224372, Time: 0.0210 Steps: 87160, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002010, Sample Num: 32160, Cur Loss: 0.12377924, Cur Avg Loss: 0.16321692, Log Avg loss: 0.13598354, Global Avg Loss: 0.70217876, Time: 0.0210 Steps: 87170, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002020, Sample Num: 32320, Cur Loss: 0.25600356, Cur Avg Loss: 0.16353457, Log Avg loss: 0.22738263, Global Avg Loss: 0.70212430, Time: 0.0210 Steps: 87180, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002030, Sample Num: 32480, Cur Loss: 0.08598153, Cur Avg Loss: 0.16369395, Log Avg loss: 0.19588795, Global Avg Loss: 0.70206624, Time: 0.0210 Steps: 87190, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002040, Sample Num: 32640, Cur Loss: 0.06278707, Cur Avg Loss: 0.16369454, Log Avg loss: 0.16381401, Global Avg Loss: 0.70200451, Time: 0.0211 Steps: 87200, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002050, Sample Num: 32800, Cur Loss: 0.01613835, Cur Avg Loss: 0.16369762, Log Avg loss: 0.16432735, Global Avg Loss: 0.70194286, Time: 0.0250 Steps: 87210, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002060, Sample Num: 32960, Cur Loss: 0.07275453, Cur Avg Loss: 0.16369077, Log Avg loss: 0.16228501, Global Avg Loss: 0.70188098, Time: 0.0212 Steps: 87220, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002070, Sample Num: 33120, Cur Loss: 0.08801601, Cur Avg Loss: 0.16360844, Log Avg loss: 0.14664861, Global Avg Loss: 0.70181733, Time: 0.0211 Steps: 87230, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002080, Sample Num: 33280, Cur Loss: 0.12111463, Cur Avg Loss: 0.16341838, Log Avg loss: 0.12407661, Global Avg Loss: 0.70175111, Time: 0.0211 Steps: 87240, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002090, Sample Num: 33440, Cur Loss: 0.18056989, Cur Avg Loss: 0.16313073, Log Avg loss: 0.10329893, Global Avg Loss: 0.70168252, Time: 0.0211 Steps: 87250, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002100, Sample Num: 33600, Cur Loss: 0.08687091, Cur Avg Loss: 0.16284957, Log Avg loss: 0.10408720, Global Avg Loss: 0.70161403, Time: 0.0212 Steps: 87260, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002110, Sample Num: 33760, Cur Loss: 0.09090980, Cur Avg Loss: 0.16289117, Log Avg loss: 0.17162679, Global Avg Loss: 0.70155330, Time: 0.0211 Steps: 87270, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002120, Sample Num: 33920, Cur Loss: 0.10455201, Cur Avg Loss: 0.16306653, Log Avg loss: 0.20006872, Global Avg Loss: 0.70149585, Time: 0.0211 Steps: 87280, Updated lr: 0.000018 ***** Running evaluation checkpoint-87289 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-87289 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.955443, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.25155, "eval_total_loss": 176.839673, "eval_mae": 0.335269, "eval_mse": 0.251646, "eval_r2": 0.840037, "eval_sp_statistic": 0.899989, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.922895, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.768993, "test_total_loss": 386.034641, "test_mae": 0.679032, "test_mse": 0.76905, "test_r2": 0.503648, "test_sp_statistic": 0.801373, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.837217, "test_ps_pvalue": 0.0, "lr": 1.8170697012802277e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7014343418795351, "train_cur_epoch_loss": 346.64586350508034, "train_cur_epoch_avg_loss": 0.16282097863085032, "train_cur_epoch_time": 44.955442905426025, "train_cur_epoch_avg_time": 0.02111575523974919, "epoch": 41, "step": 87289} ################################################## Training, Epoch: 0042, Batch: 000001, Sample Num: 16, Cur Loss: 0.16361772, Cur Avg Loss: 0.16361772, Log Avg loss: 0.11084344, Global Avg Loss: 0.70142818, Time: 0.0250 Steps: 87290, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000011, Sample Num: 176, Cur Loss: 0.08574856, Cur Avg Loss: 0.12872855, Log Avg loss: 0.12523964, Global Avg Loss: 0.70136218, Time: 0.0210 Steps: 87300, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000021, Sample Num: 336, Cur Loss: 0.26622933, Cur Avg Loss: 0.12712385, Log Avg loss: 0.12535868, Global Avg Loss: 0.70129621, Time: 0.0211 Steps: 87310, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000031, Sample Num: 496, Cur Loss: 0.20681947, Cur Avg Loss: 0.14030816, Log Avg loss: 0.16799521, Global Avg Loss: 0.70123513, Time: 0.0211 Steps: 87320, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000041, Sample Num: 656, Cur Loss: 0.13045479, Cur Avg Loss: 0.14328613, Log Avg loss: 0.15251785, Global Avg Loss: 0.70117230, Time: 0.0211 Steps: 87330, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000051, Sample Num: 816, Cur Loss: 0.08127102, Cur Avg Loss: 0.14759926, Log Avg loss: 0.16528308, Global Avg Loss: 0.70111094, Time: 0.0211 Steps: 87340, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000061, Sample Num: 976, Cur Loss: 0.16337626, Cur Avg Loss: 0.14839181, Log Avg loss: 0.15243380, Global Avg Loss: 0.70104813, Time: 0.0211 Steps: 87350, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000071, Sample Num: 1136, Cur Loss: 0.26267833, Cur Avg Loss: 0.14982562, Log Avg loss: 0.15857187, Global Avg Loss: 0.70098603, Time: 0.0211 Steps: 87360, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000081, Sample Num: 1296, Cur Loss: 0.16111802, Cur Avg Loss: 0.15176347, Log Avg loss: 0.16552223, Global Avg Loss: 0.70092475, Time: 0.0211 Steps: 87370, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000091, Sample Num: 1456, Cur Loss: 0.15158978, Cur Avg Loss: 0.15240869, Log Avg loss: 0.15763499, Global Avg Loss: 0.70086257, Time: 0.0211 Steps: 87380, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000101, Sample Num: 1616, Cur Loss: 0.06792046, Cur Avg Loss: 0.14982116, Log Avg loss: 0.12627457, Global Avg Loss: 0.70079682, Time: 0.0211 Steps: 87390, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000111, Sample Num: 1776, Cur Loss: 0.50444305, Cur Avg Loss: 0.15215282, Log Avg loss: 0.17570260, Global Avg Loss: 0.70073674, Time: 0.0211 Steps: 87400, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000121, Sample Num: 1936, Cur Loss: 0.08757969, Cur Avg Loss: 0.14842439, Log Avg loss: 0.10703881, Global Avg Loss: 0.70066882, Time: 0.0211 Steps: 87410, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000131, Sample Num: 2096, Cur Loss: 0.22535042, Cur Avg Loss: 0.15116398, Log Avg loss: 0.18431303, Global Avg Loss: 0.70060975, Time: 0.0211 Steps: 87420, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000141, Sample Num: 2256, Cur Loss: 0.12739171, Cur Avg Loss: 0.15043115, Log Avg loss: 0.14083107, Global Avg Loss: 0.70054573, Time: 0.0210 Steps: 87430, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000151, Sample Num: 2416, Cur Loss: 0.12587717, Cur Avg Loss: 0.15002829, Log Avg loss: 0.14434798, Global Avg Loss: 0.70048212, Time: 0.0212 Steps: 87440, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000161, Sample Num: 2576, Cur Loss: 0.05742342, Cur Avg Loss: 0.15153600, Log Avg loss: 0.17430242, Global Avg Loss: 0.70042195, Time: 0.0211 Steps: 87450, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000171, Sample Num: 2736, Cur Loss: 0.10109395, Cur Avg Loss: 0.14952901, Log Avg loss: 0.11721649, Global Avg Loss: 0.70035527, Time: 0.0211 Steps: 87460, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000181, Sample Num: 2896, Cur Loss: 0.21178770, Cur Avg Loss: 0.14759863, Log Avg loss: 0.11458915, Global Avg Loss: 0.70028830, Time: 0.0211 Steps: 87470, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000191, Sample Num: 3056, Cur Loss: 0.33299810, Cur Avg Loss: 0.14838901, Log Avg loss: 0.16269485, Global Avg Loss: 0.70022685, Time: 0.0211 Steps: 87480, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000201, Sample Num: 3216, Cur Loss: 0.34811798, Cur Avg Loss: 0.15234495, Log Avg loss: 0.22790352, Global Avg Loss: 0.70017286, Time: 0.0210 Steps: 87490, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000211, Sample Num: 3376, Cur Loss: 0.13432622, Cur Avg Loss: 0.15805657, Log Avg loss: 0.27286011, Global Avg Loss: 0.70012403, Time: 0.0211 Steps: 87500, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000221, Sample Num: 3536, Cur Loss: 0.06978914, Cur Avg Loss: 0.15771181, Log Avg loss: 0.15043722, Global Avg Loss: 0.70006121, Time: 0.0211 Steps: 87510, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000231, Sample Num: 3696, Cur Loss: 0.06352070, Cur Avg Loss: 0.15796678, Log Avg loss: 0.16360165, Global Avg Loss: 0.69999992, Time: 0.0211 Steps: 87520, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000241, Sample Num: 3856, Cur Loss: 0.56861979, Cur Avg Loss: 0.15870780, Log Avg loss: 0.17582550, Global Avg Loss: 0.69994003, Time: 0.0211 Steps: 87530, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000251, Sample Num: 4016, Cur Loss: 0.11558907, Cur Avg Loss: 0.15681015, Log Avg loss: 0.11107666, Global Avg Loss: 0.69987276, Time: 0.0211 Steps: 87540, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000261, Sample Num: 4176, Cur Loss: 0.07966982, Cur Avg Loss: 0.15627405, Log Avg loss: 0.14281795, Global Avg Loss: 0.69980914, Time: 0.0216 Steps: 87550, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000271, Sample Num: 4336, Cur Loss: 0.38632390, Cur Avg Loss: 0.15762059, Log Avg loss: 0.19276541, Global Avg Loss: 0.69975123, Time: 0.0212 Steps: 87560, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000281, Sample Num: 4496, Cur Loss: 0.06498732, Cur Avg Loss: 0.15691282, Log Avg loss: 0.13773213, Global Avg Loss: 0.69968705, Time: 0.0212 Steps: 87570, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000291, Sample Num: 4656, Cur Loss: 0.32366577, Cur Avg Loss: 0.15759807, Log Avg loss: 0.17685365, Global Avg Loss: 0.69962735, Time: 0.0211 Steps: 87580, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000301, Sample Num: 4816, Cur Loss: 0.03322937, Cur Avg Loss: 0.15699766, Log Avg loss: 0.13952581, Global Avg Loss: 0.69956340, Time: 0.0212 Steps: 87590, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000311, Sample Num: 4976, Cur Loss: 0.07732433, Cur Avg Loss: 0.15722084, Log Avg loss: 0.16393832, Global Avg Loss: 0.69950226, Time: 0.0211 Steps: 87600, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000321, Sample Num: 5136, Cur Loss: 0.21475092, Cur Avg Loss: 0.15704095, Log Avg loss: 0.15144651, Global Avg Loss: 0.69943970, Time: 0.0211 Steps: 87610, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000331, Sample Num: 5296, Cur Loss: 0.14678293, Cur Avg Loss: 0.15627538, Log Avg loss: 0.13170044, Global Avg Loss: 0.69937491, Time: 0.0211 Steps: 87620, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000341, Sample Num: 5456, Cur Loss: 0.26509458, Cur Avg Loss: 0.15555152, Log Avg loss: 0.13159181, Global Avg Loss: 0.69931011, Time: 0.0211 Steps: 87630, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000351, Sample Num: 5616, Cur Loss: 0.07031244, Cur Avg Loss: 0.15535345, Log Avg loss: 0.14859927, Global Avg Loss: 0.69924728, Time: 0.0211 Steps: 87640, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000361, Sample Num: 5776, Cur Loss: 0.25255936, Cur Avg Loss: 0.15461150, Log Avg loss: 0.12856914, Global Avg Loss: 0.69918217, Time: 0.0211 Steps: 87650, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000371, Sample Num: 5936, Cur Loss: 0.14072792, Cur Avg Loss: 0.15520897, Log Avg loss: 0.17677750, Global Avg Loss: 0.69912257, Time: 0.0211 Steps: 87660, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000381, Sample Num: 6096, Cur Loss: 0.23449221, Cur Avg Loss: 0.15506595, Log Avg loss: 0.14975997, Global Avg Loss: 0.69905991, Time: 0.0211 Steps: 87670, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000391, Sample Num: 6256, Cur Loss: 0.41230962, Cur Avg Loss: 0.15603401, Log Avg loss: 0.19291717, Global Avg Loss: 0.69900218, Time: 0.0212 Steps: 87680, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000401, Sample Num: 6416, Cur Loss: 0.30750349, Cur Avg Loss: 0.15650901, Log Avg loss: 0.17508156, Global Avg Loss: 0.69894244, Time: 0.0211 Steps: 87690, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000411, Sample Num: 6576, Cur Loss: 0.02239532, Cur Avg Loss: 0.15663132, Log Avg loss: 0.16153590, Global Avg Loss: 0.69888116, Time: 0.0212 Steps: 87700, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000421, Sample Num: 6736, Cur Loss: 0.03676256, Cur Avg Loss: 0.15528377, Log Avg loss: 0.09989956, Global Avg Loss: 0.69881287, Time: 0.0211 Steps: 87710, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000431, Sample Num: 6896, Cur Loss: 0.12678775, Cur Avg Loss: 0.15443603, Log Avg loss: 0.11874599, Global Avg Loss: 0.69874674, Time: 0.0211 Steps: 87720, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000441, Sample Num: 7056, Cur Loss: 0.02302012, Cur Avg Loss: 0.15374297, Log Avg loss: 0.12387226, Global Avg Loss: 0.69868121, Time: 0.0211 Steps: 87730, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000451, Sample Num: 7216, Cur Loss: 0.05454133, Cur Avg Loss: 0.15372744, Log Avg loss: 0.15304244, Global Avg Loss: 0.69861903, Time: 0.0210 Steps: 87740, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000461, Sample Num: 7376, Cur Loss: 0.07576309, Cur Avg Loss: 0.15252881, Log Avg loss: 0.09847048, Global Avg Loss: 0.69855063, Time: 0.0211 Steps: 87750, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000471, Sample Num: 7536, Cur Loss: 0.06822844, Cur Avg Loss: 0.15276070, Log Avg loss: 0.16345087, Global Avg Loss: 0.69848966, Time: 0.0211 Steps: 87760, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000481, Sample Num: 7696, Cur Loss: 0.25775829, Cur Avg Loss: 0.15487992, Log Avg loss: 0.25469511, Global Avg Loss: 0.69843910, Time: 0.0211 Steps: 87770, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000491, Sample Num: 7856, Cur Loss: 0.06290156, Cur Avg Loss: 0.15317305, Log Avg loss: 0.07107295, Global Avg Loss: 0.69836763, Time: 0.0212 Steps: 87780, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000501, Sample Num: 8016, Cur Loss: 0.23071072, Cur Avg Loss: 0.15282126, Log Avg loss: 0.13554828, Global Avg Loss: 0.69830352, Time: 0.0211 Steps: 87790, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000511, Sample Num: 8176, Cur Loss: 0.13694993, Cur Avg Loss: 0.15321027, Log Avg loss: 0.17269969, Global Avg Loss: 0.69824365, Time: 0.0211 Steps: 87800, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000521, Sample Num: 8336, Cur Loss: 0.08099084, Cur Avg Loss: 0.15335656, Log Avg loss: 0.16083166, Global Avg Loss: 0.69818245, Time: 0.0212 Steps: 87810, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000531, Sample Num: 8496, Cur Loss: 0.24669746, Cur Avg Loss: 0.15309082, Log Avg loss: 0.13924590, Global Avg Loss: 0.69811881, Time: 0.0211 Steps: 87820, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000541, Sample Num: 8656, Cur Loss: 0.22149634, Cur Avg Loss: 0.15282350, Log Avg loss: 0.13862878, Global Avg Loss: 0.69805510, Time: 0.0211 Steps: 87830, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000551, Sample Num: 8816, Cur Loss: 0.14696756, Cur Avg Loss: 0.15185527, Log Avg loss: 0.09947400, Global Avg Loss: 0.69798696, Time: 0.0211 Steps: 87840, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000561, Sample Num: 8976, Cur Loss: 0.09991959, Cur Avg Loss: 0.15212330, Log Avg loss: 0.16689177, Global Avg Loss: 0.69792650, Time: 0.0211 Steps: 87850, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000571, Sample Num: 9136, Cur Loss: 0.16796538, Cur Avg Loss: 0.15233986, Log Avg loss: 0.16448910, Global Avg Loss: 0.69786579, Time: 0.0212 Steps: 87860, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000581, Sample Num: 9296, Cur Loss: 0.27847177, Cur Avg Loss: 0.15209004, Log Avg loss: 0.13782534, Global Avg Loss: 0.69780206, Time: 0.0211 Steps: 87870, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000591, Sample Num: 9456, Cur Loss: 0.04628455, Cur Avg Loss: 0.15223090, Log Avg loss: 0.16041501, Global Avg Loss: 0.69774091, Time: 0.0211 Steps: 87880, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000601, Sample Num: 9616, Cur Loss: 0.28748208, Cur Avg Loss: 0.15239627, Log Avg loss: 0.16216942, Global Avg Loss: 0.69767997, Time: 0.0212 Steps: 87890, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000611, Sample Num: 9776, Cur Loss: 0.05902678, Cur Avg Loss: 0.15218839, Log Avg loss: 0.13969494, Global Avg Loss: 0.69761649, Time: 0.0212 Steps: 87900, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000621, Sample Num: 9936, Cur Loss: 0.05570237, Cur Avg Loss: 0.15214175, Log Avg loss: 0.14929166, Global Avg Loss: 0.69755412, Time: 0.0211 Steps: 87910, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000631, Sample Num: 10096, Cur Loss: 0.26788434, Cur Avg Loss: 0.15291197, Log Avg loss: 0.20074299, Global Avg Loss: 0.69749761, Time: 0.0211 Steps: 87920, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000641, Sample Num: 10256, Cur Loss: 0.19212815, Cur Avg Loss: 0.15287377, Log Avg loss: 0.15046337, Global Avg Loss: 0.69743540, Time: 0.0211 Steps: 87930, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000651, Sample Num: 10416, Cur Loss: 0.17765513, Cur Avg Loss: 0.15277856, Log Avg loss: 0.14667514, Global Avg Loss: 0.69737277, Time: 0.0211 Steps: 87940, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000661, Sample Num: 10576, Cur Loss: 0.13583572, Cur Avg Loss: 0.15261704, Log Avg loss: 0.14210262, Global Avg Loss: 0.69730963, Time: 0.0211 Steps: 87950, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000671, Sample Num: 10736, Cur Loss: 0.07231411, Cur Avg Loss: 0.15237333, Log Avg loss: 0.13626386, Global Avg Loss: 0.69724585, Time: 0.0212 Steps: 87960, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000681, Sample Num: 10896, Cur Loss: 0.07025412, Cur Avg Loss: 0.15342620, Log Avg loss: 0.22407401, Global Avg Loss: 0.69719206, Time: 0.0211 Steps: 87970, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000691, Sample Num: 11056, Cur Loss: 0.05961858, Cur Avg Loss: 0.15299279, Log Avg loss: 0.12347719, Global Avg Loss: 0.69712685, Time: 0.0211 Steps: 87980, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000701, Sample Num: 11216, Cur Loss: 0.08247013, Cur Avg Loss: 0.15266649, Log Avg loss: 0.13011902, Global Avg Loss: 0.69706241, Time: 0.0211 Steps: 87990, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000711, Sample Num: 11376, Cur Loss: 0.16854900, Cur Avg Loss: 0.15232280, Log Avg loss: 0.12823037, Global Avg Loss: 0.69699777, Time: 0.0211 Steps: 88000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000721, Sample Num: 11536, Cur Loss: 0.26104435, Cur Avg Loss: 0.15229543, Log Avg loss: 0.15034931, Global Avg Loss: 0.69693566, Time: 0.0211 Steps: 88010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000731, Sample Num: 11696, Cur Loss: 0.23077781, Cur Avg Loss: 0.15250333, Log Avg loss: 0.16749319, Global Avg Loss: 0.69687551, Time: 0.0212 Steps: 88020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000741, Sample Num: 11856, Cur Loss: 0.23948704, Cur Avg Loss: 0.15266887, Log Avg loss: 0.16476940, Global Avg Loss: 0.69681506, Time: 0.0212 Steps: 88030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000751, Sample Num: 12016, Cur Loss: 0.09131497, Cur Avg Loss: 0.15383472, Log Avg loss: 0.24022488, Global Avg Loss: 0.69676320, Time: 0.0211 Steps: 88040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000761, Sample Num: 12176, Cur Loss: 0.16923243, Cur Avg Loss: 0.15354581, Log Avg loss: 0.13184837, Global Avg Loss: 0.69669904, Time: 0.0211 Steps: 88050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000771, Sample Num: 12336, Cur Loss: 0.07409742, Cur Avg Loss: 0.15373443, Log Avg loss: 0.16808806, Global Avg Loss: 0.69663901, Time: 0.0247 Steps: 88060, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000781, Sample Num: 12496, Cur Loss: 0.09114886, Cur Avg Loss: 0.15320188, Log Avg loss: 0.11214259, Global Avg Loss: 0.69657265, Time: 0.0210 Steps: 88070, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000791, Sample Num: 12656, Cur Loss: 0.11601602, Cur Avg Loss: 0.15311904, Log Avg loss: 0.14664939, Global Avg Loss: 0.69651021, Time: 0.0211 Steps: 88080, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000801, Sample Num: 12816, Cur Loss: 0.11319517, Cur Avg Loss: 0.15296245, Log Avg loss: 0.14057563, Global Avg Loss: 0.69644710, Time: 0.0210 Steps: 88090, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000811, Sample Num: 12976, Cur Loss: 0.16394091, Cur Avg Loss: 0.15302875, Log Avg loss: 0.15833943, Global Avg Loss: 0.69638602, Time: 0.0210 Steps: 88100, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000821, Sample Num: 13136, Cur Loss: 0.07449552, Cur Avg Loss: 0.15248947, Log Avg loss: 0.10875417, Global Avg Loss: 0.69631933, Time: 0.0211 Steps: 88110, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000831, Sample Num: 13296, Cur Loss: 0.08583209, Cur Avg Loss: 0.15257071, Log Avg loss: 0.15924064, Global Avg Loss: 0.69625838, Time: 0.0210 Steps: 88120, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000841, Sample Num: 13456, Cur Loss: 0.10056283, Cur Avg Loss: 0.15302184, Log Avg loss: 0.19051043, Global Avg Loss: 0.69620099, Time: 0.0210 Steps: 88130, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000851, Sample Num: 13616, Cur Loss: 0.05567317, Cur Avg Loss: 0.15290125, Log Avg loss: 0.14276021, Global Avg Loss: 0.69613820, Time: 0.0210 Steps: 88140, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000861, Sample Num: 13776, Cur Loss: 0.05799935, Cur Avg Loss: 0.15266257, Log Avg loss: 0.13235043, Global Avg Loss: 0.69607425, Time: 0.0210 Steps: 88150, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000871, Sample Num: 13936, Cur Loss: 0.06082857, Cur Avg Loss: 0.15222173, Log Avg loss: 0.11426574, Global Avg Loss: 0.69600825, Time: 0.0210 Steps: 88160, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000881, Sample Num: 14096, Cur Loss: 0.13613312, Cur Avg Loss: 0.15189042, Log Avg loss: 0.12303310, Global Avg Loss: 0.69594327, Time: 0.0210 Steps: 88170, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000891, Sample Num: 14256, Cur Loss: 0.37149984, Cur Avg Loss: 0.15215262, Log Avg loss: 0.17525252, Global Avg Loss: 0.69588422, Time: 0.0210 Steps: 88180, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000901, Sample Num: 14416, Cur Loss: 0.12978660, Cur Avg Loss: 0.15195936, Log Avg loss: 0.13473976, Global Avg Loss: 0.69582059, Time: 0.0210 Steps: 88190, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000911, Sample Num: 14576, Cur Loss: 0.14368835, Cur Avg Loss: 0.15255658, Log Avg loss: 0.20636578, Global Avg Loss: 0.69576509, Time: 0.0210 Steps: 88200, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000921, Sample Num: 14736, Cur Loss: 0.11829028, Cur Avg Loss: 0.15234796, Log Avg loss: 0.13334294, Global Avg Loss: 0.69570133, Time: 0.0211 Steps: 88210, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000931, Sample Num: 14896, Cur Loss: 0.28215677, Cur Avg Loss: 0.15295804, Log Avg loss: 0.20914615, Global Avg Loss: 0.69564618, Time: 0.0211 Steps: 88220, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000941, Sample Num: 15056, Cur Loss: 0.07394120, Cur Avg Loss: 0.15286945, Log Avg loss: 0.14462229, Global Avg Loss: 0.69558373, Time: 0.0211 Steps: 88230, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000951, Sample Num: 15216, Cur Loss: 0.18115914, Cur Avg Loss: 0.15326268, Log Avg loss: 0.19026501, Global Avg Loss: 0.69552646, Time: 0.0211 Steps: 88240, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000961, Sample Num: 15376, Cur Loss: 0.13228261, Cur Avg Loss: 0.15358372, Log Avg loss: 0.18411513, Global Avg Loss: 0.69546851, Time: 0.0211 Steps: 88250, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000971, Sample Num: 15536, Cur Loss: 0.35532522, Cur Avg Loss: 0.15368689, Log Avg loss: 0.16360115, Global Avg Loss: 0.69540825, Time: 0.0211 Steps: 88260, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000981, Sample Num: 15696, Cur Loss: 0.26535046, Cur Avg Loss: 0.15442120, Log Avg loss: 0.22572308, Global Avg Loss: 0.69535504, Time: 0.0210 Steps: 88270, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000991, Sample Num: 15856, Cur Loss: 0.14872421, Cur Avg Loss: 0.15491724, Log Avg loss: 0.20357830, Global Avg Loss: 0.69529933, Time: 0.0211 Steps: 88280, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001001, Sample Num: 16016, Cur Loss: 0.08650301, Cur Avg Loss: 0.15497304, Log Avg loss: 0.16050301, Global Avg Loss: 0.69523876, Time: 0.0211 Steps: 88290, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001011, Sample Num: 16176, Cur Loss: 0.09255929, Cur Avg Loss: 0.15544446, Log Avg loss: 0.20263401, Global Avg Loss: 0.69518297, Time: 0.0211 Steps: 88300, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001021, Sample Num: 16336, Cur Loss: 0.11403340, Cur Avg Loss: 0.15506470, Log Avg loss: 0.11667050, Global Avg Loss: 0.69511746, Time: 0.0210 Steps: 88310, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001031, Sample Num: 16496, Cur Loss: 0.07856420, Cur Avg Loss: 0.15524169, Log Avg loss: 0.17331283, Global Avg Loss: 0.69505838, Time: 0.0211 Steps: 88320, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001041, Sample Num: 16656, Cur Loss: 0.25596395, Cur Avg Loss: 0.15572530, Log Avg loss: 0.20558537, Global Avg Loss: 0.69500297, Time: 0.0211 Steps: 88330, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001051, Sample Num: 16816, Cur Loss: 0.28026488, Cur Avg Loss: 0.15575350, Log Avg loss: 0.15868925, Global Avg Loss: 0.69494226, Time: 0.0210 Steps: 88340, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001061, Sample Num: 16976, Cur Loss: 0.12757586, Cur Avg Loss: 0.15520853, Log Avg loss: 0.09793222, Global Avg Loss: 0.69487469, Time: 0.0211 Steps: 88350, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001071, Sample Num: 17136, Cur Loss: 0.07613394, Cur Avg Loss: 0.15462012, Log Avg loss: 0.09218959, Global Avg Loss: 0.69480648, Time: 0.0211 Steps: 88360, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001081, Sample Num: 17296, Cur Loss: 0.03254148, Cur Avg Loss: 0.15417045, Log Avg loss: 0.10601106, Global Avg Loss: 0.69473985, Time: 0.0211 Steps: 88370, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001091, Sample Num: 17456, Cur Loss: 0.12711212, Cur Avg Loss: 0.15411021, Log Avg loss: 0.14759771, Global Avg Loss: 0.69467794, Time: 0.0211 Steps: 88380, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001101, Sample Num: 17616, Cur Loss: 0.13895197, Cur Avg Loss: 0.15410246, Log Avg loss: 0.15325745, Global Avg Loss: 0.69461669, Time: 0.0210 Steps: 88390, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001111, Sample Num: 17776, Cur Loss: 0.11628725, Cur Avg Loss: 0.15402092, Log Avg loss: 0.14504339, Global Avg Loss: 0.69455452, Time: 0.0211 Steps: 88400, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001121, Sample Num: 17936, Cur Loss: 0.06958545, Cur Avg Loss: 0.15422899, Log Avg loss: 0.17734513, Global Avg Loss: 0.69449602, Time: 0.0211 Steps: 88410, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001131, Sample Num: 18096, Cur Loss: 0.22861481, Cur Avg Loss: 0.15438137, Log Avg loss: 0.17146338, Global Avg Loss: 0.69443686, Time: 0.0211 Steps: 88420, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001141, Sample Num: 18256, Cur Loss: 0.17795420, Cur Avg Loss: 0.15449159, Log Avg loss: 0.16695715, Global Avg Loss: 0.69437722, Time: 0.0210 Steps: 88430, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001151, Sample Num: 18416, Cur Loss: 0.04611166, Cur Avg Loss: 0.15475033, Log Avg loss: 0.18427281, Global Avg Loss: 0.69431954, Time: 0.0210 Steps: 88440, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001161, Sample Num: 18576, Cur Loss: 0.16849159, Cur Avg Loss: 0.15495611, Log Avg loss: 0.17864108, Global Avg Loss: 0.69426124, Time: 0.0211 Steps: 88450, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001171, Sample Num: 18736, Cur Loss: 0.27966720, Cur Avg Loss: 0.15509473, Log Avg loss: 0.17118889, Global Avg Loss: 0.69420210, Time: 0.0211 Steps: 88460, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001181, Sample Num: 18896, Cur Loss: 0.05575158, Cur Avg Loss: 0.15502977, Log Avg loss: 0.14742342, Global Avg Loss: 0.69414030, Time: 0.0211 Steps: 88470, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001191, Sample Num: 19056, Cur Loss: 0.08361086, Cur Avg Loss: 0.15482840, Log Avg loss: 0.13104544, Global Avg Loss: 0.69407666, Time: 0.0211 Steps: 88480, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001201, Sample Num: 19216, Cur Loss: 0.09109697, Cur Avg Loss: 0.15508791, Log Avg loss: 0.18599660, Global Avg Loss: 0.69401924, Time: 0.0211 Steps: 88490, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001211, Sample Num: 19376, Cur Loss: 0.13782522, Cur Avg Loss: 0.15605581, Log Avg loss: 0.27230005, Global Avg Loss: 0.69397159, Time: 0.0210 Steps: 88500, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001221, Sample Num: 19536, Cur Loss: 0.10184467, Cur Avg Loss: 0.15615408, Log Avg loss: 0.16805501, Global Avg Loss: 0.69391217, Time: 0.0210 Steps: 88510, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001231, Sample Num: 19696, Cur Loss: 0.16202635, Cur Avg Loss: 0.15660600, Log Avg loss: 0.21178556, Global Avg Loss: 0.69385771, Time: 0.0210 Steps: 88520, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001241, Sample Num: 19856, Cur Loss: 0.11103449, Cur Avg Loss: 0.15709184, Log Avg loss: 0.21689798, Global Avg Loss: 0.69380383, Time: 0.0211 Steps: 88530, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001251, Sample Num: 20016, Cur Loss: 0.12983583, Cur Avg Loss: 0.15700025, Log Avg loss: 0.14563411, Global Avg Loss: 0.69374192, Time: 0.0211 Steps: 88540, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001261, Sample Num: 20176, Cur Loss: 0.09881735, Cur Avg Loss: 0.15706059, Log Avg loss: 0.16460888, Global Avg Loss: 0.69368216, Time: 0.0211 Steps: 88550, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001271, Sample Num: 20336, Cur Loss: 0.17652574, Cur Avg Loss: 0.15769304, Log Avg loss: 0.23744487, Global Avg Loss: 0.69363065, Time: 0.0211 Steps: 88560, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001281, Sample Num: 20496, Cur Loss: 0.07249926, Cur Avg Loss: 0.15763422, Log Avg loss: 0.15015899, Global Avg Loss: 0.69356929, Time: 0.0247 Steps: 88570, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001291, Sample Num: 20656, Cur Loss: 0.02614159, Cur Avg Loss: 0.15747362, Log Avg loss: 0.13690044, Global Avg Loss: 0.69350644, Time: 0.0211 Steps: 88580, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001301, Sample Num: 20816, Cur Loss: 0.11385054, Cur Avg Loss: 0.15754576, Log Avg loss: 0.16685901, Global Avg Loss: 0.69344700, Time: 0.0210 Steps: 88590, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001311, Sample Num: 20976, Cur Loss: 0.20132329, Cur Avg Loss: 0.15755728, Log Avg loss: 0.15905662, Global Avg Loss: 0.69338668, Time: 0.0211 Steps: 88600, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001321, Sample Num: 21136, Cur Loss: 0.08421370, Cur Avg Loss: 0.15757592, Log Avg loss: 0.16001868, Global Avg Loss: 0.69332649, Time: 0.0211 Steps: 88610, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001331, Sample Num: 21296, Cur Loss: 0.16641888, Cur Avg Loss: 0.15754105, Log Avg loss: 0.15293543, Global Avg Loss: 0.69326551, Time: 0.0210 Steps: 88620, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001341, Sample Num: 21456, Cur Loss: 0.38411385, Cur Avg Loss: 0.15757632, Log Avg loss: 0.16227088, Global Avg Loss: 0.69320560, Time: 0.0211 Steps: 88630, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001351, Sample Num: 21616, Cur Loss: 0.24812044, Cur Avg Loss: 0.15718235, Log Avg loss: 0.10434984, Global Avg Loss: 0.69313917, Time: 0.0210 Steps: 88640, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001361, Sample Num: 21776, Cur Loss: 0.06033525, Cur Avg Loss: 0.15742742, Log Avg loss: 0.19053725, Global Avg Loss: 0.69308247, Time: 0.0210 Steps: 88650, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001371, Sample Num: 21936, Cur Loss: 0.05690670, Cur Avg Loss: 0.15717487, Log Avg loss: 0.12280302, Global Avg Loss: 0.69301815, Time: 0.0211 Steps: 88660, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001381, Sample Num: 22096, Cur Loss: 0.32264018, Cur Avg Loss: 0.15716267, Log Avg loss: 0.15548958, Global Avg Loss: 0.69295753, Time: 0.0211 Steps: 88670, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001391, Sample Num: 22256, Cur Loss: 0.03273681, Cur Avg Loss: 0.15690606, Log Avg loss: 0.12146839, Global Avg Loss: 0.69289308, Time: 0.0210 Steps: 88680, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001401, Sample Num: 22416, Cur Loss: 0.04867229, Cur Avg Loss: 0.15668657, Log Avg loss: 0.12615504, Global Avg Loss: 0.69282918, Time: 0.0211 Steps: 88690, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001411, Sample Num: 22576, Cur Loss: 0.08292194, Cur Avg Loss: 0.15668093, Log Avg loss: 0.15589076, Global Avg Loss: 0.69276865, Time: 0.0211 Steps: 88700, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001421, Sample Num: 22736, Cur Loss: 0.09197766, Cur Avg Loss: 0.15708729, Log Avg loss: 0.21442473, Global Avg Loss: 0.69271473, Time: 0.0210 Steps: 88710, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001431, Sample Num: 22896, Cur Loss: 0.08313738, Cur Avg Loss: 0.15667261, Log Avg loss: 0.09774635, Global Avg Loss: 0.69264766, Time: 0.0211 Steps: 88720, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001441, Sample Num: 23056, Cur Loss: 0.07106440, Cur Avg Loss: 0.15647775, Log Avg loss: 0.12859368, Global Avg Loss: 0.69258409, Time: 0.0210 Steps: 88730, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001451, Sample Num: 23216, Cur Loss: 0.17598861, Cur Avg Loss: 0.15661413, Log Avg loss: 0.17626633, Global Avg Loss: 0.69252591, Time: 0.0210 Steps: 88740, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001461, Sample Num: 23376, Cur Loss: 0.07647070, Cur Avg Loss: 0.15669045, Log Avg loss: 0.16776536, Global Avg Loss: 0.69246678, Time: 0.0210 Steps: 88750, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001471, Sample Num: 23536, Cur Loss: 0.13576001, Cur Avg Loss: 0.15706309, Log Avg loss: 0.21150501, Global Avg Loss: 0.69241260, Time: 0.0211 Steps: 88760, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001481, Sample Num: 23696, Cur Loss: 0.15714926, Cur Avg Loss: 0.15715123, Log Avg loss: 0.17011611, Global Avg Loss: 0.69235376, Time: 0.0210 Steps: 88770, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001491, Sample Num: 23856, Cur Loss: 0.13652718, Cur Avg Loss: 0.15754536, Log Avg loss: 0.21591727, Global Avg Loss: 0.69230009, Time: 0.0211 Steps: 88780, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001501, Sample Num: 24016, Cur Loss: 0.16795361, Cur Avg Loss: 0.15786823, Log Avg loss: 0.20600810, Global Avg Loss: 0.69224533, Time: 0.0210 Steps: 88790, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001511, Sample Num: 24176, Cur Loss: 0.21958840, Cur Avg Loss: 0.15823847, Log Avg loss: 0.21381112, Global Avg Loss: 0.69219145, Time: 0.0211 Steps: 88800, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001521, Sample Num: 24336, Cur Loss: 0.13229564, Cur Avg Loss: 0.15835259, Log Avg loss: 0.17559663, Global Avg Loss: 0.69213328, Time: 0.0210 Steps: 88810, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001531, Sample Num: 24496, Cur Loss: 0.21389405, Cur Avg Loss: 0.15851792, Log Avg loss: 0.18366425, Global Avg Loss: 0.69207603, Time: 0.0210 Steps: 88820, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001541, Sample Num: 24656, Cur Loss: 0.23296183, Cur Avg Loss: 0.15872890, Log Avg loss: 0.19102972, Global Avg Loss: 0.69201963, Time: 0.0212 Steps: 88830, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001551, Sample Num: 24816, Cur Loss: 0.12284275, Cur Avg Loss: 0.15894708, Log Avg loss: 0.19256850, Global Avg Loss: 0.69196341, Time: 0.0211 Steps: 88840, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001561, Sample Num: 24976, Cur Loss: 0.12875620, Cur Avg Loss: 0.15903478, Log Avg loss: 0.17263649, Global Avg Loss: 0.69190496, Time: 0.0210 Steps: 88850, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001571, Sample Num: 25136, Cur Loss: 0.15918007, Cur Avg Loss: 0.15909699, Log Avg loss: 0.16880854, Global Avg Loss: 0.69184609, Time: 0.0211 Steps: 88860, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001581, Sample Num: 25296, Cur Loss: 0.03226173, Cur Avg Loss: 0.15894393, Log Avg loss: 0.13489842, Global Avg Loss: 0.69178342, Time: 0.0210 Steps: 88870, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001591, Sample Num: 25456, Cur Loss: 0.07749987, Cur Avg Loss: 0.15872917, Log Avg loss: 0.12477544, Global Avg Loss: 0.69171963, Time: 0.0211 Steps: 88880, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001601, Sample Num: 25616, Cur Loss: 0.20749035, Cur Avg Loss: 0.15854892, Log Avg loss: 0.12987061, Global Avg Loss: 0.69165642, Time: 0.0210 Steps: 88890, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001611, Sample Num: 25776, Cur Loss: 0.04800551, Cur Avg Loss: 0.15836302, Log Avg loss: 0.12860129, Global Avg Loss: 0.69159308, Time: 0.0210 Steps: 88900, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001621, Sample Num: 25936, Cur Loss: 0.11527894, Cur Avg Loss: 0.15835472, Log Avg loss: 0.15701715, Global Avg Loss: 0.69153296, Time: 0.0210 Steps: 88910, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001631, Sample Num: 26096, Cur Loss: 0.10850484, Cur Avg Loss: 0.15864156, Log Avg loss: 0.20513856, Global Avg Loss: 0.69147826, Time: 0.0210 Steps: 88920, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001641, Sample Num: 26256, Cur Loss: 0.15534189, Cur Avg Loss: 0.15837997, Log Avg loss: 0.11571454, Global Avg Loss: 0.69141351, Time: 0.0210 Steps: 88930, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001651, Sample Num: 26416, Cur Loss: 0.16046044, Cur Avg Loss: 0.15843238, Log Avg loss: 0.16703234, Global Avg Loss: 0.69135456, Time: 0.0211 Steps: 88940, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001661, Sample Num: 26576, Cur Loss: 0.21395278, Cur Avg Loss: 0.15851830, Log Avg loss: 0.17270345, Global Avg Loss: 0.69129625, Time: 0.0210 Steps: 88950, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001671, Sample Num: 26736, Cur Loss: 0.09386957, Cur Avg Loss: 0.15884053, Log Avg loss: 0.21236321, Global Avg Loss: 0.69124241, Time: 0.0211 Steps: 88960, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001681, Sample Num: 26896, Cur Loss: 0.09318627, Cur Avg Loss: 0.15862126, Log Avg loss: 0.12198159, Global Avg Loss: 0.69117843, Time: 0.0211 Steps: 88970, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001691, Sample Num: 27056, Cur Loss: 0.23062597, Cur Avg Loss: 0.15856900, Log Avg loss: 0.14978487, Global Avg Loss: 0.69111758, Time: 0.0211 Steps: 88980, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001701, Sample Num: 27216, Cur Loss: 0.18064268, Cur Avg Loss: 0.15869830, Log Avg loss: 0.18056163, Global Avg Loss: 0.69106021, Time: 0.0211 Steps: 88990, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001711, Sample Num: 27376, Cur Loss: 0.12838711, Cur Avg Loss: 0.15855025, Log Avg loss: 0.13336770, Global Avg Loss: 0.69099755, Time: 0.0210 Steps: 89000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001721, Sample Num: 27536, Cur Loss: 0.22729480, Cur Avg Loss: 0.15849036, Log Avg loss: 0.14824334, Global Avg Loss: 0.69093657, Time: 0.0210 Steps: 89010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001731, Sample Num: 27696, Cur Loss: 0.12122197, Cur Avg Loss: 0.15844102, Log Avg loss: 0.14994907, Global Avg Loss: 0.69087580, Time: 0.0212 Steps: 89020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001741, Sample Num: 27856, Cur Loss: 0.21029407, Cur Avg Loss: 0.15828555, Log Avg loss: 0.13137433, Global Avg Loss: 0.69081296, Time: 0.0211 Steps: 89030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001751, Sample Num: 28016, Cur Loss: 0.07353565, Cur Avg Loss: 0.15795257, Log Avg loss: 0.09998043, Global Avg Loss: 0.69074660, Time: 0.0210 Steps: 89040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001761, Sample Num: 28176, Cur Loss: 0.08800846, Cur Avg Loss: 0.15790194, Log Avg loss: 0.14903603, Global Avg Loss: 0.69068577, Time: 0.0210 Steps: 89050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001771, Sample Num: 28336, Cur Loss: 0.14920525, Cur Avg Loss: 0.15772451, Log Avg loss: 0.12647965, Global Avg Loss: 0.69062242, Time: 0.0211 Steps: 89060, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001781, Sample Num: 28496, Cur Loss: 0.27053845, Cur Avg Loss: 0.15807885, Log Avg loss: 0.22083310, Global Avg Loss: 0.69056967, Time: 0.0210 Steps: 89070, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001791, Sample Num: 28656, Cur Loss: 0.12578267, Cur Avg Loss: 0.15789954, Log Avg loss: 0.12596328, Global Avg Loss: 0.69050629, Time: 0.0210 Steps: 89080, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001801, Sample Num: 28816, Cur Loss: 0.07020901, Cur Avg Loss: 0.15790506, Log Avg loss: 0.15889362, Global Avg Loss: 0.69044662, Time: 0.0211 Steps: 89090, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001811, Sample Num: 28976, Cur Loss: 0.06792606, Cur Avg Loss: 0.15792647, Log Avg loss: 0.16178276, Global Avg Loss: 0.69038729, Time: 0.0211 Steps: 89100, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001821, Sample Num: 29136, Cur Loss: 0.24381164, Cur Avg Loss: 0.15819620, Log Avg loss: 0.20704490, Global Avg Loss: 0.69033304, Time: 0.0211 Steps: 89110, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001831, Sample Num: 29296, Cur Loss: 0.07988775, Cur Avg Loss: 0.15840633, Log Avg loss: 0.19666977, Global Avg Loss: 0.69027765, Time: 0.0211 Steps: 89120, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001841, Sample Num: 29456, Cur Loss: 0.25922304, Cur Avg Loss: 0.15826639, Log Avg loss: 0.13264504, Global Avg Loss: 0.69021509, Time: 0.0211 Steps: 89130, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001851, Sample Num: 29616, Cur Loss: 0.23944627, Cur Avg Loss: 0.15844283, Log Avg loss: 0.19092545, Global Avg Loss: 0.69015908, Time: 0.0210 Steps: 89140, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001861, Sample Num: 29776, Cur Loss: 0.24820811, Cur Avg Loss: 0.15848155, Log Avg loss: 0.16564688, Global Avg Loss: 0.69010024, Time: 0.0211 Steps: 89150, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001871, Sample Num: 29936, Cur Loss: 0.18442874, Cur Avg Loss: 0.15831761, Log Avg loss: 0.12780951, Global Avg Loss: 0.69003717, Time: 0.0211 Steps: 89160, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001881, Sample Num: 30096, Cur Loss: 0.08511268, Cur Avg Loss: 0.15820313, Log Avg loss: 0.13678406, Global Avg Loss: 0.68997513, Time: 0.0211 Steps: 89170, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001891, Sample Num: 30256, Cur Loss: 0.10668700, Cur Avg Loss: 0.15838334, Log Avg loss: 0.19228053, Global Avg Loss: 0.68991932, Time: 0.0211 Steps: 89180, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001901, Sample Num: 30416, Cur Loss: 0.13760331, Cur Avg Loss: 0.15860283, Log Avg loss: 0.20010760, Global Avg Loss: 0.68986440, Time: 0.0211 Steps: 89190, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001911, Sample Num: 30576, Cur Loss: 0.32609886, Cur Avg Loss: 0.15855135, Log Avg loss: 0.14876575, Global Avg Loss: 0.68980374, Time: 0.0211 Steps: 89200, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001921, Sample Num: 30736, Cur Loss: 0.26435852, Cur Avg Loss: 0.15854708, Log Avg loss: 0.15773196, Global Avg Loss: 0.68974410, Time: 0.0211 Steps: 89210, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001931, Sample Num: 30896, Cur Loss: 0.36901492, Cur Avg Loss: 0.15917721, Log Avg loss: 0.28022360, Global Avg Loss: 0.68969820, Time: 0.0211 Steps: 89220, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001941, Sample Num: 31056, Cur Loss: 0.13904890, Cur Avg Loss: 0.15945293, Log Avg loss: 0.21269508, Global Avg Loss: 0.68964474, Time: 0.0211 Steps: 89230, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001951, Sample Num: 31216, Cur Loss: 0.11163154, Cur Avg Loss: 0.15931872, Log Avg loss: 0.13326808, Global Avg Loss: 0.68958240, Time: 0.0211 Steps: 89240, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001961, Sample Num: 31376, Cur Loss: 0.05197583, Cur Avg Loss: 0.15920357, Log Avg loss: 0.13673757, Global Avg Loss: 0.68952045, Time: 0.0210 Steps: 89250, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001971, Sample Num: 31536, Cur Loss: 0.12884063, Cur Avg Loss: 0.15947612, Log Avg loss: 0.21292305, Global Avg Loss: 0.68946706, Time: 0.0211 Steps: 89260, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001981, Sample Num: 31696, Cur Loss: 0.11879820, Cur Avg Loss: 0.15985534, Log Avg loss: 0.23460080, Global Avg Loss: 0.68941611, Time: 0.0210 Steps: 89270, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001991, Sample Num: 31856, Cur Loss: 0.15151107, Cur Avg Loss: 0.16000795, Log Avg loss: 0.19023904, Global Avg Loss: 0.68936019, Time: 0.0211 Steps: 89280, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002001, Sample Num: 32016, Cur Loss: 0.16394900, Cur Avg Loss: 0.16026959, Log Avg loss: 0.21236364, Global Avg Loss: 0.68930677, Time: 0.0210 Steps: 89290, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002011, Sample Num: 32176, Cur Loss: 0.07342642, Cur Avg Loss: 0.16053758, Log Avg loss: 0.21416076, Global Avg Loss: 0.68925356, Time: 0.0211 Steps: 89300, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002021, Sample Num: 32336, Cur Loss: 0.11112983, Cur Avg Loss: 0.16086950, Log Avg loss: 0.22762007, Global Avg Loss: 0.68920188, Time: 0.0211 Steps: 89310, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002031, Sample Num: 32496, Cur Loss: 0.10140340, Cur Avg Loss: 0.16063766, Log Avg loss: 0.11378180, Global Avg Loss: 0.68913745, Time: 0.0211 Steps: 89320, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002041, Sample Num: 32656, Cur Loss: 0.11793942, Cur Avg Loss: 0.16084658, Log Avg loss: 0.20327791, Global Avg Loss: 0.68908306, Time: 0.0211 Steps: 89330, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002051, Sample Num: 32816, Cur Loss: 0.32455546, Cur Avg Loss: 0.16088476, Log Avg loss: 0.16867860, Global Avg Loss: 0.68902481, Time: 0.0248 Steps: 89340, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002061, Sample Num: 32976, Cur Loss: 0.05460346, Cur Avg Loss: 0.16096640, Log Avg loss: 0.17771032, Global Avg Loss: 0.68896759, Time: 0.0211 Steps: 89350, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002071, Sample Num: 33136, Cur Loss: 0.16540042, Cur Avg Loss: 0.16079946, Log Avg loss: 0.12639358, Global Avg Loss: 0.68890463, Time: 0.0211 Steps: 89360, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002081, Sample Num: 33296, Cur Loss: 0.19859685, Cur Avg Loss: 0.16057440, Log Avg loss: 0.11396381, Global Avg Loss: 0.68884030, Time: 0.0211 Steps: 89370, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002091, Sample Num: 33456, Cur Loss: 0.15539038, Cur Avg Loss: 0.16051844, Log Avg loss: 0.14887402, Global Avg Loss: 0.68877989, Time: 0.0211 Steps: 89380, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002101, Sample Num: 33616, Cur Loss: 0.26869985, Cur Avg Loss: 0.16051919, Log Avg loss: 0.16067591, Global Avg Loss: 0.68872081, Time: 0.0211 Steps: 89390, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002111, Sample Num: 33776, Cur Loss: 0.10047268, Cur Avg Loss: 0.16038636, Log Avg loss: 0.13247826, Global Avg Loss: 0.68865859, Time: 0.0211 Steps: 89400, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002121, Sample Num: 33936, Cur Loss: 0.05002875, Cur Avg Loss: 0.16025805, Log Avg loss: 0.13317182, Global Avg Loss: 0.68859646, Time: 0.0211 Steps: 89410, Updated lr: 0.000016 ***** Running evaluation checkpoint-89418 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-89418 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.049686, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.249995, "eval_total_loss": 175.746202, "eval_mae": 0.333361, "eval_mse": 0.25009, "eval_r2": 0.841026, "eval_sp_statistic": 0.897266, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.92135, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.739313, "test_total_loss": 371.135204, "test_mae": 0.660108, "test_mse": 0.739353, "test_r2": 0.522815, "test_sp_statistic": 0.797066, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.835837, "test_ps_pvalue": 0.0, "lr": 1.6151730678046468e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6885465087054998, "train_cur_epoch_loss": 340.9494471056387, "train_cur_epoch_avg_loss": 0.16014534857005105, "train_cur_epoch_time": 45.049686431884766, "train_cur_epoch_avg_time": 0.021160021809246015, "epoch": 42, "step": 89418} ################################################## Training, Epoch: 0043, Batch: 000002, Sample Num: 32, Cur Loss: 0.04470527, Cur Avg Loss: 0.03416223, Log Avg loss: 0.11104450, Global Avg Loss: 0.68853187, Time: 0.0249 Steps: 89420, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000012, Sample Num: 192, Cur Loss: 0.32317591, Cur Avg Loss: 0.16707533, Log Avg loss: 0.19365795, Global Avg Loss: 0.68847654, Time: 0.0211 Steps: 89430, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000022, Sample Num: 352, Cur Loss: 0.08095212, Cur Avg Loss: 0.14701629, Log Avg loss: 0.12294544, Global Avg Loss: 0.68841331, Time: 0.0211 Steps: 89440, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000032, Sample Num: 512, Cur Loss: 0.38841820, Cur Avg Loss: 0.14440850, Log Avg loss: 0.13867137, Global Avg Loss: 0.68835185, Time: 0.0210 Steps: 89450, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000042, Sample Num: 672, Cur Loss: 0.12189511, Cur Avg Loss: 0.14184986, Log Avg loss: 0.13366222, Global Avg Loss: 0.68828984, Time: 0.0210 Steps: 89460, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000052, Sample Num: 832, Cur Loss: 0.08369023, Cur Avg Loss: 0.13786260, Log Avg loss: 0.12111610, Global Avg Loss: 0.68822645, Time: 0.0211 Steps: 89470, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000062, Sample Num: 992, Cur Loss: 0.09054884, Cur Avg Loss: 0.13744185, Log Avg loss: 0.13525392, Global Avg Loss: 0.68816465, Time: 0.0211 Steps: 89480, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000072, Sample Num: 1152, Cur Loss: 0.07740918, Cur Avg Loss: 0.13876426, Log Avg loss: 0.14696321, Global Avg Loss: 0.68810418, Time: 0.0211 Steps: 89490, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000082, Sample Num: 1312, Cur Loss: 0.21970573, Cur Avg Loss: 0.14528708, Log Avg loss: 0.19225142, Global Avg Loss: 0.68804877, Time: 0.0211 Steps: 89500, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000092, Sample Num: 1472, Cur Loss: 0.29462287, Cur Avg Loss: 0.14374923, Log Avg loss: 0.13113881, Global Avg Loss: 0.68798656, Time: 0.0211 Steps: 89510, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000102, Sample Num: 1632, Cur Loss: 0.11529696, Cur Avg Loss: 0.14629267, Log Avg loss: 0.16969239, Global Avg Loss: 0.68792866, Time: 0.0211 Steps: 89520, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000112, Sample Num: 1792, Cur Loss: 0.20620799, Cur Avg Loss: 0.14504492, Log Avg loss: 0.13231781, Global Avg Loss: 0.68786660, Time: 0.0210 Steps: 89530, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000122, Sample Num: 1952, Cur Loss: 0.12877184, Cur Avg Loss: 0.14569799, Log Avg loss: 0.15301244, Global Avg Loss: 0.68780687, Time: 0.0210 Steps: 89540, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000132, Sample Num: 2112, Cur Loss: 0.09719698, Cur Avg Loss: 0.14451253, Log Avg loss: 0.13004986, Global Avg Loss: 0.68774458, Time: 0.0210 Steps: 89550, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000142, Sample Num: 2272, Cur Loss: 0.04743895, Cur Avg Loss: 0.14905029, Log Avg loss: 0.20894874, Global Avg Loss: 0.68769112, Time: 0.0210 Steps: 89560, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000152, Sample Num: 2432, Cur Loss: 0.18911114, Cur Avg Loss: 0.15086202, Log Avg loss: 0.17658858, Global Avg Loss: 0.68763406, Time: 0.0210 Steps: 89570, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000162, Sample Num: 2592, Cur Loss: 0.11199994, Cur Avg Loss: 0.15222860, Log Avg loss: 0.17300059, Global Avg Loss: 0.68757661, Time: 0.0210 Steps: 89580, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000172, Sample Num: 2752, Cur Loss: 0.41884017, Cur Avg Loss: 0.15255346, Log Avg loss: 0.15781624, Global Avg Loss: 0.68751748, Time: 0.0210 Steps: 89590, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000182, Sample Num: 2912, Cur Loss: 0.11739133, Cur Avg Loss: 0.15212156, Log Avg loss: 0.14469288, Global Avg Loss: 0.68745690, Time: 0.0210 Steps: 89600, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000192, Sample Num: 3072, Cur Loss: 0.08826313, Cur Avg Loss: 0.15558239, Log Avg loss: 0.21856939, Global Avg Loss: 0.68740457, Time: 0.0211 Steps: 89610, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000202, Sample Num: 3232, Cur Loss: 0.14882447, Cur Avg Loss: 0.15313800, Log Avg loss: 0.10620572, Global Avg Loss: 0.68733972, Time: 0.0210 Steps: 89620, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000212, Sample Num: 3392, Cur Loss: 0.09368783, Cur Avg Loss: 0.15397269, Log Avg loss: 0.17083340, Global Avg Loss: 0.68728209, Time: 0.0210 Steps: 89630, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000222, Sample Num: 3552, Cur Loss: 0.03138850, Cur Avg Loss: 0.15194381, Log Avg loss: 0.10893168, Global Avg Loss: 0.68721757, Time: 0.0211 Steps: 89640, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000232, Sample Num: 3712, Cur Loss: 0.09969634, Cur Avg Loss: 0.15264622, Log Avg loss: 0.16823973, Global Avg Loss: 0.68715968, Time: 0.0210 Steps: 89650, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000242, Sample Num: 3872, Cur Loss: 0.18446642, Cur Avg Loss: 0.15158094, Log Avg loss: 0.12686651, Global Avg Loss: 0.68709719, Time: 0.0210 Steps: 89660, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000252, Sample Num: 4032, Cur Loss: 0.14275807, Cur Avg Loss: 0.15256429, Log Avg loss: 0.17636135, Global Avg Loss: 0.68704024, Time: 0.0211 Steps: 89670, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000262, Sample Num: 4192, Cur Loss: 0.14416048, Cur Avg Loss: 0.15540243, Log Avg loss: 0.22692353, Global Avg Loss: 0.68698893, Time: 0.0247 Steps: 89680, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000272, Sample Num: 4352, Cur Loss: 0.18773447, Cur Avg Loss: 0.15683134, Log Avg loss: 0.19426864, Global Avg Loss: 0.68693399, Time: 0.0211 Steps: 89690, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000282, Sample Num: 4512, Cur Loss: 0.08673593, Cur Avg Loss: 0.15530591, Log Avg loss: 0.11381421, Global Avg Loss: 0.68687010, Time: 0.0211 Steps: 89700, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000292, Sample Num: 4672, Cur Loss: 0.06670539, Cur Avg Loss: 0.15459648, Log Avg loss: 0.13459077, Global Avg Loss: 0.68680854, Time: 0.0210 Steps: 89710, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000302, Sample Num: 4832, Cur Loss: 0.21332911, Cur Avg Loss: 0.15478157, Log Avg loss: 0.16018604, Global Avg Loss: 0.68674984, Time: 0.0210 Steps: 89720, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000312, Sample Num: 4992, Cur Loss: 0.05250531, Cur Avg Loss: 0.15494111, Log Avg loss: 0.15975927, Global Avg Loss: 0.68669111, Time: 0.0210 Steps: 89730, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000322, Sample Num: 5152, Cur Loss: 0.23573947, Cur Avg Loss: 0.15540971, Log Avg loss: 0.17002989, Global Avg Loss: 0.68663354, Time: 0.0210 Steps: 89740, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000332, Sample Num: 5312, Cur Loss: 0.16676740, Cur Avg Loss: 0.15459872, Log Avg loss: 0.12848499, Global Avg Loss: 0.68657135, Time: 0.0210 Steps: 89750, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000342, Sample Num: 5472, Cur Loss: 0.19457777, Cur Avg Loss: 0.15438404, Log Avg loss: 0.14725659, Global Avg Loss: 0.68651126, Time: 0.0210 Steps: 89760, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000352, Sample Num: 5632, Cur Loss: 0.05583160, Cur Avg Loss: 0.15383695, Log Avg loss: 0.13512663, Global Avg Loss: 0.68644984, Time: 0.0210 Steps: 89770, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000362, Sample Num: 5792, Cur Loss: 0.09566508, Cur Avg Loss: 0.15370216, Log Avg loss: 0.14895728, Global Avg Loss: 0.68638997, Time: 0.0210 Steps: 89780, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000372, Sample Num: 5952, Cur Loss: 0.14350979, Cur Avg Loss: 0.15383173, Log Avg loss: 0.15852241, Global Avg Loss: 0.68633119, Time: 0.0210 Steps: 89790, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000382, Sample Num: 6112, Cur Loss: 0.07575259, Cur Avg Loss: 0.15255889, Log Avg loss: 0.10520902, Global Avg Loss: 0.68626647, Time: 0.0211 Steps: 89800, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000392, Sample Num: 6272, Cur Loss: 0.07797382, Cur Avg Loss: 0.15164677, Log Avg loss: 0.11680398, Global Avg Loss: 0.68620306, Time: 0.0210 Steps: 89810, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000402, Sample Num: 6432, Cur Loss: 0.08872064, Cur Avg Loss: 0.15159930, Log Avg loss: 0.14973852, Global Avg Loss: 0.68614334, Time: 0.0210 Steps: 89820, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000412, Sample Num: 6592, Cur Loss: 0.06765819, Cur Avg Loss: 0.15073196, Log Avg loss: 0.11586490, Global Avg Loss: 0.68607985, Time: 0.0211 Steps: 89830, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000422, Sample Num: 6752, Cur Loss: 0.04205944, Cur Avg Loss: 0.14996543, Log Avg loss: 0.11838448, Global Avg Loss: 0.68601666, Time: 0.0211 Steps: 89840, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000432, Sample Num: 6912, Cur Loss: 0.15338597, Cur Avg Loss: 0.14940919, Log Avg loss: 0.12593566, Global Avg Loss: 0.68595433, Time: 0.0210 Steps: 89850, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000442, Sample Num: 7072, Cur Loss: 0.28446895, Cur Avg Loss: 0.15028399, Log Avg loss: 0.18807519, Global Avg Loss: 0.68589892, Time: 0.0210 Steps: 89860, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000452, Sample Num: 7232, Cur Loss: 0.05460874, Cur Avg Loss: 0.14927102, Log Avg loss: 0.10449779, Global Avg Loss: 0.68583423, Time: 0.0210 Steps: 89870, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000462, Sample Num: 7392, Cur Loss: 0.05666199, Cur Avg Loss: 0.15137752, Log Avg loss: 0.24659159, Global Avg Loss: 0.68578536, Time: 0.0210 Steps: 89880, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000472, Sample Num: 7552, Cur Loss: 0.10333771, Cur Avg Loss: 0.15199865, Log Avg loss: 0.18069452, Global Avg Loss: 0.68572917, Time: 0.0210 Steps: 89890, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000482, Sample Num: 7712, Cur Loss: 0.15913767, Cur Avg Loss: 0.15205520, Log Avg loss: 0.15472466, Global Avg Loss: 0.68567010, Time: 0.0211 Steps: 89900, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000492, Sample Num: 7872, Cur Loss: 0.07879615, Cur Avg Loss: 0.15186924, Log Avg loss: 0.14290564, Global Avg Loss: 0.68560974, Time: 0.0210 Steps: 89910, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000502, Sample Num: 8032, Cur Loss: 0.31657106, Cur Avg Loss: 0.15185125, Log Avg loss: 0.15096651, Global Avg Loss: 0.68555028, Time: 0.0210 Steps: 89920, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000512, Sample Num: 8192, Cur Loss: 0.07448892, Cur Avg Loss: 0.15179208, Log Avg loss: 0.14882170, Global Avg Loss: 0.68549060, Time: 0.0255 Steps: 89930, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000522, Sample Num: 8352, Cur Loss: 0.24273510, Cur Avg Loss: 0.15164321, Log Avg loss: 0.14402094, Global Avg Loss: 0.68543039, Time: 0.0210 Steps: 89940, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000532, Sample Num: 8512, Cur Loss: 0.21637705, Cur Avg Loss: 0.15162928, Log Avg loss: 0.15090225, Global Avg Loss: 0.68537097, Time: 0.0210 Steps: 89950, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000542, Sample Num: 8672, Cur Loss: 0.13693400, Cur Avg Loss: 0.15137988, Log Avg loss: 0.13811173, Global Avg Loss: 0.68531013, Time: 0.0210 Steps: 89960, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000552, Sample Num: 8832, Cur Loss: 0.37446809, Cur Avg Loss: 0.15330684, Log Avg loss: 0.25774826, Global Avg Loss: 0.68526261, Time: 0.0210 Steps: 89970, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000562, Sample Num: 8992, Cur Loss: 0.53358316, Cur Avg Loss: 0.15475298, Log Avg loss: 0.23457997, Global Avg Loss: 0.68521252, Time: 0.0210 Steps: 89980, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000572, Sample Num: 9152, Cur Loss: 0.02791222, Cur Avg Loss: 0.15477043, Log Avg loss: 0.15575100, Global Avg Loss: 0.68515369, Time: 0.0210 Steps: 89990, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000582, Sample Num: 9312, Cur Loss: 0.20836040, Cur Avg Loss: 0.15534591, Log Avg loss: 0.18826309, Global Avg Loss: 0.68509848, Time: 0.0210 Steps: 90000, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000592, Sample Num: 9472, Cur Loss: 0.29611263, Cur Avg Loss: 0.15598735, Log Avg loss: 0.19331945, Global Avg Loss: 0.68504384, Time: 0.0210 Steps: 90010, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000602, Sample Num: 9632, Cur Loss: 0.13907620, Cur Avg Loss: 0.15632576, Log Avg loss: 0.17635972, Global Avg Loss: 0.68498733, Time: 0.0210 Steps: 90020, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000612, Sample Num: 9792, Cur Loss: 0.04147907, Cur Avg Loss: 0.15718134, Log Avg loss: 0.20868693, Global Avg Loss: 0.68493443, Time: 0.0210 Steps: 90030, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000622, Sample Num: 9952, Cur Loss: 0.08567958, Cur Avg Loss: 0.15577843, Log Avg loss: 0.06992074, Global Avg Loss: 0.68486613, Time: 0.0210 Steps: 90040, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000632, Sample Num: 10112, Cur Loss: 0.04393726, Cur Avg Loss: 0.15511407, Log Avg loss: 0.11379074, Global Avg Loss: 0.68480271, Time: 0.0210 Steps: 90050, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000642, Sample Num: 10272, Cur Loss: 0.28745830, Cur Avg Loss: 0.15453436, Log Avg loss: 0.11789650, Global Avg Loss: 0.68473976, Time: 0.0210 Steps: 90060, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000652, Sample Num: 10432, Cur Loss: 0.07196919, Cur Avg Loss: 0.15462394, Log Avg loss: 0.16037506, Global Avg Loss: 0.68468154, Time: 0.0210 Steps: 90070, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000662, Sample Num: 10592, Cur Loss: 0.05603027, Cur Avg Loss: 0.15403027, Log Avg loss: 0.11532293, Global Avg Loss: 0.68461834, Time: 0.0210 Steps: 90080, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000672, Sample Num: 10752, Cur Loss: 0.21983606, Cur Avg Loss: 0.15363200, Log Avg loss: 0.12726669, Global Avg Loss: 0.68455647, Time: 0.0210 Steps: 90090, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000682, Sample Num: 10912, Cur Loss: 0.04971182, Cur Avg Loss: 0.15405325, Log Avg loss: 0.18236090, Global Avg Loss: 0.68450073, Time: 0.0210 Steps: 90100, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000692, Sample Num: 11072, Cur Loss: 0.19112045, Cur Avg Loss: 0.15394403, Log Avg loss: 0.14649516, Global Avg Loss: 0.68444103, Time: 0.0211 Steps: 90110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000702, Sample Num: 11232, Cur Loss: 0.06554043, Cur Avg Loss: 0.15369728, Log Avg loss: 0.13662231, Global Avg Loss: 0.68438024, Time: 0.0211 Steps: 90120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000712, Sample Num: 11392, Cur Loss: 0.02684018, Cur Avg Loss: 0.15332079, Log Avg loss: 0.12689101, Global Avg Loss: 0.68431839, Time: 0.0210 Steps: 90130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000722, Sample Num: 11552, Cur Loss: 0.05239901, Cur Avg Loss: 0.15299918, Log Avg loss: 0.13010084, Global Avg Loss: 0.68425690, Time: 0.0210 Steps: 90140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000732, Sample Num: 11712, Cur Loss: 0.03979412, Cur Avg Loss: 0.15326082, Log Avg loss: 0.17215153, Global Avg Loss: 0.68420010, Time: 0.0210 Steps: 90150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000742, Sample Num: 11872, Cur Loss: 0.02824136, Cur Avg Loss: 0.15294129, Log Avg loss: 0.12955134, Global Avg Loss: 0.68413858, Time: 0.0210 Steps: 90160, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000752, Sample Num: 12032, Cur Loss: 0.34771124, Cur Avg Loss: 0.15311752, Log Avg loss: 0.16619352, Global Avg Loss: 0.68408114, Time: 0.0210 Steps: 90170, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000762, Sample Num: 12192, Cur Loss: 0.12709767, Cur Avg Loss: 0.15317394, Log Avg loss: 0.15741673, Global Avg Loss: 0.68402274, Time: 0.0210 Steps: 90180, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000772, Sample Num: 12352, Cur Loss: 0.07730810, Cur Avg Loss: 0.15285460, Log Avg loss: 0.12852136, Global Avg Loss: 0.68396114, Time: 0.0247 Steps: 90190, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000782, Sample Num: 12512, Cur Loss: 0.10099116, Cur Avg Loss: 0.15254911, Log Avg loss: 0.12896506, Global Avg Loss: 0.68389961, Time: 0.0211 Steps: 90200, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000792, Sample Num: 12672, Cur Loss: 0.09131009, Cur Avg Loss: 0.15269928, Log Avg loss: 0.16444297, Global Avg Loss: 0.68384203, Time: 0.0209 Steps: 90210, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000802, Sample Num: 12832, Cur Loss: 0.14310564, Cur Avg Loss: 0.15298062, Log Avg loss: 0.17526209, Global Avg Loss: 0.68378566, Time: 0.0210 Steps: 90220, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000812, Sample Num: 12992, Cur Loss: 0.03364033, Cur Avg Loss: 0.15272759, Log Avg loss: 0.13243521, Global Avg Loss: 0.68372455, Time: 0.0211 Steps: 90230, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000822, Sample Num: 13152, Cur Loss: 0.43208119, Cur Avg Loss: 0.15279371, Log Avg loss: 0.15816246, Global Avg Loss: 0.68366631, Time: 0.0210 Steps: 90240, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000832, Sample Num: 13312, Cur Loss: 0.17680666, Cur Avg Loss: 0.15348730, Log Avg loss: 0.21050066, Global Avg Loss: 0.68361389, Time: 0.0210 Steps: 90250, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000842, Sample Num: 13472, Cur Loss: 0.16587184, Cur Avg Loss: 0.15356547, Log Avg loss: 0.16006877, Global Avg Loss: 0.68355588, Time: 0.0210 Steps: 90260, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000852, Sample Num: 13632, Cur Loss: 0.11940248, Cur Avg Loss: 0.15398034, Log Avg loss: 0.18891278, Global Avg Loss: 0.68350109, Time: 0.0210 Steps: 90270, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000862, Sample Num: 13792, Cur Loss: 0.04806650, Cur Avg Loss: 0.15392031, Log Avg loss: 0.14880521, Global Avg Loss: 0.68344186, Time: 0.0210 Steps: 90280, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000872, Sample Num: 13952, Cur Loss: 0.23461157, Cur Avg Loss: 0.15380539, Log Avg loss: 0.14389970, Global Avg Loss: 0.68338210, Time: 0.0210 Steps: 90290, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000882, Sample Num: 14112, Cur Loss: 0.28214550, Cur Avg Loss: 0.15438689, Log Avg loss: 0.20509325, Global Avg Loss: 0.68332914, Time: 0.0210 Steps: 90300, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000892, Sample Num: 14272, Cur Loss: 0.12736182, Cur Avg Loss: 0.15398004, Log Avg loss: 0.11809581, Global Avg Loss: 0.68326655, Time: 0.0210 Steps: 90310, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000902, Sample Num: 14432, Cur Loss: 0.17353788, Cur Avg Loss: 0.15380321, Log Avg loss: 0.13803068, Global Avg Loss: 0.68320618, Time: 0.0210 Steps: 90320, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000912, Sample Num: 14592, Cur Loss: 0.04792140, Cur Avg Loss: 0.15364221, Log Avg loss: 0.13911952, Global Avg Loss: 0.68314595, Time: 0.0210 Steps: 90330, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000922, Sample Num: 14752, Cur Loss: 0.26667646, Cur Avg Loss: 0.15454240, Log Avg loss: 0.23663950, Global Avg Loss: 0.68309652, Time: 0.0210 Steps: 90340, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000932, Sample Num: 14912, Cur Loss: 0.22146690, Cur Avg Loss: 0.15466788, Log Avg loss: 0.16623713, Global Avg Loss: 0.68303932, Time: 0.0210 Steps: 90350, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000942, Sample Num: 15072, Cur Loss: 0.05514958, Cur Avg Loss: 0.15435471, Log Avg loss: 0.12516737, Global Avg Loss: 0.68297758, Time: 0.0210 Steps: 90360, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000952, Sample Num: 15232, Cur Loss: 0.09985782, Cur Avg Loss: 0.15450639, Log Avg loss: 0.16879474, Global Avg Loss: 0.68292068, Time: 0.0210 Steps: 90370, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000962, Sample Num: 15392, Cur Loss: 0.09042270, Cur Avg Loss: 0.15443464, Log Avg loss: 0.14760403, Global Avg Loss: 0.68286145, Time: 0.0210 Steps: 90380, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000972, Sample Num: 15552, Cur Loss: 0.14780512, Cur Avg Loss: 0.15411081, Log Avg loss: 0.12295895, Global Avg Loss: 0.68279951, Time: 0.0210 Steps: 90390, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000982, Sample Num: 15712, Cur Loss: 0.28941467, Cur Avg Loss: 0.15448635, Log Avg loss: 0.19098827, Global Avg Loss: 0.68274510, Time: 0.0210 Steps: 90400, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000992, Sample Num: 15872, Cur Loss: 0.11566214, Cur Avg Loss: 0.15406144, Log Avg loss: 0.11233537, Global Avg Loss: 0.68268201, Time: 0.0210 Steps: 90410, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001002, Sample Num: 16032, Cur Loss: 0.10351701, Cur Avg Loss: 0.15409284, Log Avg loss: 0.15720822, Global Avg Loss: 0.68262390, Time: 0.0210 Steps: 90420, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001012, Sample Num: 16192, Cur Loss: 0.05842154, Cur Avg Loss: 0.15373178, Log Avg loss: 0.11755315, Global Avg Loss: 0.68256141, Time: 0.0210 Steps: 90430, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001022, Sample Num: 16352, Cur Loss: 0.12650226, Cur Avg Loss: 0.15397399, Log Avg loss: 0.17848570, Global Avg Loss: 0.68250567, Time: 0.0210 Steps: 90440, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001032, Sample Num: 16512, Cur Loss: 0.16395786, Cur Avg Loss: 0.15366327, Log Avg loss: 0.12190770, Global Avg Loss: 0.68244369, Time: 0.0211 Steps: 90450, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001042, Sample Num: 16672, Cur Loss: 0.16024220, Cur Avg Loss: 0.15375356, Log Avg loss: 0.16307174, Global Avg Loss: 0.68238628, Time: 0.0210 Steps: 90460, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001052, Sample Num: 16832, Cur Loss: 0.10931586, Cur Avg Loss: 0.15400551, Log Avg loss: 0.18025841, Global Avg Loss: 0.68233078, Time: 0.0210 Steps: 90470, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001062, Sample Num: 16992, Cur Loss: 0.31144676, Cur Avg Loss: 0.15399891, Log Avg loss: 0.15330401, Global Avg Loss: 0.68227231, Time: 0.0210 Steps: 90480, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001072, Sample Num: 17152, Cur Loss: 0.09173329, Cur Avg Loss: 0.15376513, Log Avg loss: 0.12893847, Global Avg Loss: 0.68221116, Time: 0.0210 Steps: 90490, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001082, Sample Num: 17312, Cur Loss: 0.28313500, Cur Avg Loss: 0.15390512, Log Avg loss: 0.16891164, Global Avg Loss: 0.68215444, Time: 0.0210 Steps: 90500, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001092, Sample Num: 17472, Cur Loss: 0.03358268, Cur Avg Loss: 0.15347495, Log Avg loss: 0.10693018, Global Avg Loss: 0.68209089, Time: 0.0210 Steps: 90510, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001102, Sample Num: 17632, Cur Loss: 0.02419826, Cur Avg Loss: 0.15324287, Log Avg loss: 0.12790000, Global Avg Loss: 0.68202967, Time: 0.0210 Steps: 90520, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001112, Sample Num: 17792, Cur Loss: 0.03384016, Cur Avg Loss: 0.15325820, Log Avg loss: 0.15494741, Global Avg Loss: 0.68197144, Time: 0.0210 Steps: 90530, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001122, Sample Num: 17952, Cur Loss: 0.08634865, Cur Avg Loss: 0.15277400, Log Avg loss: 0.09893089, Global Avg Loss: 0.68190705, Time: 0.0210 Steps: 90540, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001132, Sample Num: 18112, Cur Loss: 0.11423981, Cur Avg Loss: 0.15314045, Log Avg loss: 0.19425647, Global Avg Loss: 0.68185319, Time: 0.0210 Steps: 90550, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001142, Sample Num: 18272, Cur Loss: 0.10619646, Cur Avg Loss: 0.15312814, Log Avg loss: 0.15173534, Global Avg Loss: 0.68179466, Time: 0.0210 Steps: 90560, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001152, Sample Num: 18432, Cur Loss: 0.06868584, Cur Avg Loss: 0.15352952, Log Avg loss: 0.19936656, Global Avg Loss: 0.68174139, Time: 0.0210 Steps: 90570, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001162, Sample Num: 18592, Cur Loss: 0.17543380, Cur Avg Loss: 0.15403051, Log Avg loss: 0.21174441, Global Avg Loss: 0.68168950, Time: 0.0210 Steps: 90580, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001172, Sample Num: 18752, Cur Loss: 0.04845504, Cur Avg Loss: 0.15380580, Log Avg loss: 0.12769517, Global Avg Loss: 0.68162835, Time: 0.0210 Steps: 90590, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001182, Sample Num: 18912, Cur Loss: 0.10788967, Cur Avg Loss: 0.15344286, Log Avg loss: 0.11090574, Global Avg Loss: 0.68156536, Time: 0.0211 Steps: 90600, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001192, Sample Num: 19072, Cur Loss: 0.03422987, Cur Avg Loss: 0.15330954, Log Avg loss: 0.13755103, Global Avg Loss: 0.68150532, Time: 0.0210 Steps: 90610, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001202, Sample Num: 19232, Cur Loss: 0.52250063, Cur Avg Loss: 0.15385510, Log Avg loss: 0.21888639, Global Avg Loss: 0.68145427, Time: 0.0210 Steps: 90620, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001212, Sample Num: 19392, Cur Loss: 0.03767865, Cur Avg Loss: 0.15395650, Log Avg loss: 0.16614410, Global Avg Loss: 0.68139741, Time: 0.0210 Steps: 90630, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001222, Sample Num: 19552, Cur Loss: 0.16820598, Cur Avg Loss: 0.15369861, Log Avg loss: 0.12244245, Global Avg Loss: 0.68133574, Time: 0.0210 Steps: 90640, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001232, Sample Num: 19712, Cur Loss: 0.34913447, Cur Avg Loss: 0.15456299, Log Avg loss: 0.26018991, Global Avg Loss: 0.68128928, Time: 0.0211 Steps: 90650, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001242, Sample Num: 19872, Cur Loss: 0.06494032, Cur Avg Loss: 0.15456589, Log Avg loss: 0.15492335, Global Avg Loss: 0.68123122, Time: 0.0211 Steps: 90660, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001252, Sample Num: 20032, Cur Loss: 0.23044150, Cur Avg Loss: 0.15458128, Log Avg loss: 0.15649243, Global Avg Loss: 0.68117335, Time: 0.0211 Steps: 90670, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001262, Sample Num: 20192, Cur Loss: 0.17587903, Cur Avg Loss: 0.15462825, Log Avg loss: 0.16050993, Global Avg Loss: 0.68111593, Time: 0.0210 Steps: 90680, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001272, Sample Num: 20352, Cur Loss: 0.12951384, Cur Avg Loss: 0.15507796, Log Avg loss: 0.21183104, Global Avg Loss: 0.68106418, Time: 0.0210 Steps: 90690, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001282, Sample Num: 20512, Cur Loss: 0.07654954, Cur Avg Loss: 0.15484481, Log Avg loss: 0.12518860, Global Avg Loss: 0.68100290, Time: 0.0248 Steps: 90700, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001292, Sample Num: 20672, Cur Loss: 0.22579110, Cur Avg Loss: 0.15506135, Log Avg loss: 0.18282175, Global Avg Loss: 0.68094798, Time: 0.0210 Steps: 90710, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001302, Sample Num: 20832, Cur Loss: 0.02280468, Cur Avg Loss: 0.15491054, Log Avg loss: 0.13542523, Global Avg Loss: 0.68088784, Time: 0.0207 Steps: 90720, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001312, Sample Num: 20992, Cur Loss: 0.34710109, Cur Avg Loss: 0.15525202, Log Avg loss: 0.19971331, Global Avg Loss: 0.68083481, Time: 0.0207 Steps: 90730, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001322, Sample Num: 21152, Cur Loss: 0.16609672, Cur Avg Loss: 0.15522832, Log Avg loss: 0.15211866, Global Avg Loss: 0.68077654, Time: 0.0207 Steps: 90740, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001332, Sample Num: 21312, Cur Loss: 0.19672446, Cur Avg Loss: 0.15489277, Log Avg loss: 0.11053251, Global Avg Loss: 0.68071371, Time: 0.0207 Steps: 90750, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001342, Sample Num: 21472, Cur Loss: 0.25544119, Cur Avg Loss: 0.15477391, Log Avg loss: 0.13894214, Global Avg Loss: 0.68065401, Time: 0.0207 Steps: 90760, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001352, Sample Num: 21632, Cur Loss: 0.13589957, Cur Avg Loss: 0.15492587, Log Avg loss: 0.17531860, Global Avg Loss: 0.68059834, Time: 0.0207 Steps: 90770, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001362, Sample Num: 21792, Cur Loss: 0.14417617, Cur Avg Loss: 0.15501229, Log Avg loss: 0.16669637, Global Avg Loss: 0.68054173, Time: 0.0207 Steps: 90780, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001372, Sample Num: 21952, Cur Loss: 0.24498110, Cur Avg Loss: 0.15548515, Log Avg loss: 0.21988937, Global Avg Loss: 0.68049099, Time: 0.0207 Steps: 90790, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001382, Sample Num: 22112, Cur Loss: 0.19729389, Cur Avg Loss: 0.15563691, Log Avg loss: 0.17645795, Global Avg Loss: 0.68043548, Time: 0.0207 Steps: 90800, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001392, Sample Num: 22272, Cur Loss: 0.24588138, Cur Avg Loss: 0.15570923, Log Avg loss: 0.16570406, Global Avg Loss: 0.68037880, Time: 0.0207 Steps: 90810, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001402, Sample Num: 22432, Cur Loss: 0.23875713, Cur Avg Loss: 0.15606879, Log Avg loss: 0.20611957, Global Avg Loss: 0.68032658, Time: 0.0207 Steps: 90820, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001412, Sample Num: 22592, Cur Loss: 0.07618662, Cur Avg Loss: 0.15591917, Log Avg loss: 0.13494231, Global Avg Loss: 0.68026654, Time: 0.0207 Steps: 90830, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001422, Sample Num: 22752, Cur Loss: 0.08488932, Cur Avg Loss: 0.15588200, Log Avg loss: 0.15063322, Global Avg Loss: 0.68020823, Time: 0.0207 Steps: 90840, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001432, Sample Num: 22912, Cur Loss: 0.04919813, Cur Avg Loss: 0.15562738, Log Avg loss: 0.11942112, Global Avg Loss: 0.68014651, Time: 0.0207 Steps: 90850, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001442, Sample Num: 23072, Cur Loss: 0.21155941, Cur Avg Loss: 0.15546847, Log Avg loss: 0.13271172, Global Avg Loss: 0.68008626, Time: 0.0208 Steps: 90860, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001452, Sample Num: 23232, Cur Loss: 0.36442205, Cur Avg Loss: 0.15593273, Log Avg loss: 0.22287934, Global Avg Loss: 0.68003594, Time: 0.0207 Steps: 90870, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001462, Sample Num: 23392, Cur Loss: 0.10995090, Cur Avg Loss: 0.15606773, Log Avg loss: 0.17566968, Global Avg Loss: 0.67998044, Time: 0.0207 Steps: 90880, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001472, Sample Num: 23552, Cur Loss: 0.14426221, Cur Avg Loss: 0.15568850, Log Avg loss: 0.10024558, Global Avg Loss: 0.67991666, Time: 0.0208 Steps: 90890, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001482, Sample Num: 23712, Cur Loss: 0.17073098, Cur Avg Loss: 0.15571138, Log Avg loss: 0.15907877, Global Avg Loss: 0.67985936, Time: 0.0207 Steps: 90900, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001492, Sample Num: 23872, Cur Loss: 0.14951494, Cur Avg Loss: 0.15600679, Log Avg loss: 0.19978692, Global Avg Loss: 0.67980655, Time: 0.0207 Steps: 90910, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001502, Sample Num: 24032, Cur Loss: 0.19329238, Cur Avg Loss: 0.15596104, Log Avg loss: 0.14913512, Global Avg Loss: 0.67974819, Time: 0.0207 Steps: 90920, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001512, Sample Num: 24192, Cur Loss: 0.09254918, Cur Avg Loss: 0.15602416, Log Avg loss: 0.16550400, Global Avg Loss: 0.67969163, Time: 0.0207 Steps: 90930, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001522, Sample Num: 24352, Cur Loss: 0.19361293, Cur Avg Loss: 0.15598027, Log Avg loss: 0.14934389, Global Avg Loss: 0.67963332, Time: 0.0207 Steps: 90940, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001532, Sample Num: 24512, Cur Loss: 0.08662356, Cur Avg Loss: 0.15610155, Log Avg loss: 0.17456132, Global Avg Loss: 0.67957778, Time: 0.0207 Steps: 90950, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001542, Sample Num: 24672, Cur Loss: 0.16148630, Cur Avg Loss: 0.15608377, Log Avg loss: 0.15336024, Global Avg Loss: 0.67951993, Time: 0.0220 Steps: 90960, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001552, Sample Num: 24832, Cur Loss: 0.04871204, Cur Avg Loss: 0.15626948, Log Avg loss: 0.18490521, Global Avg Loss: 0.67946556, Time: 0.0219 Steps: 90970, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001562, Sample Num: 24992, Cur Loss: 0.06681371, Cur Avg Loss: 0.15595469, Log Avg loss: 0.10709878, Global Avg Loss: 0.67940265, Time: 0.0219 Steps: 90980, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001572, Sample Num: 25152, Cur Loss: 0.66197050, Cur Avg Loss: 0.15610107, Log Avg loss: 0.17896591, Global Avg Loss: 0.67934765, Time: 0.0218 Steps: 90990, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001582, Sample Num: 25312, Cur Loss: 0.17418702, Cur Avg Loss: 0.15581476, Log Avg loss: 0.11080760, Global Avg Loss: 0.67928517, Time: 0.0219 Steps: 91000, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001592, Sample Num: 25472, Cur Loss: 0.30422580, Cur Avg Loss: 0.15593527, Log Avg loss: 0.17500020, Global Avg Loss: 0.67922976, Time: 0.0218 Steps: 91010, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001602, Sample Num: 25632, Cur Loss: 0.10550579, Cur Avg Loss: 0.15618933, Log Avg loss: 0.19663559, Global Avg Loss: 0.67917674, Time: 0.0219 Steps: 91020, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001612, Sample Num: 25792, Cur Loss: 0.02398599, Cur Avg Loss: 0.15581154, Log Avg loss: 0.09528832, Global Avg Loss: 0.67911260, Time: 0.0219 Steps: 91030, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001622, Sample Num: 25952, Cur Loss: 0.32034644, Cur Avg Loss: 0.15564949, Log Avg loss: 0.12952728, Global Avg Loss: 0.67905223, Time: 0.0219 Steps: 91040, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001632, Sample Num: 26112, Cur Loss: 0.03929225, Cur Avg Loss: 0.15562593, Log Avg loss: 0.15180528, Global Avg Loss: 0.67899432, Time: 0.0218 Steps: 91050, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001642, Sample Num: 26272, Cur Loss: 0.15675342, Cur Avg Loss: 0.15537427, Log Avg loss: 0.11430217, Global Avg Loss: 0.67893231, Time: 0.0219 Steps: 91060, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001652, Sample Num: 26432, Cur Loss: 0.23399208, Cur Avg Loss: 0.15537822, Log Avg loss: 0.15602821, Global Avg Loss: 0.67887489, Time: 0.0219 Steps: 91070, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001662, Sample Num: 26592, Cur Loss: 0.22638130, Cur Avg Loss: 0.15543423, Log Avg loss: 0.16468574, Global Avg Loss: 0.67881844, Time: 0.0218 Steps: 91080, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001672, Sample Num: 26752, Cur Loss: 0.19125937, Cur Avg Loss: 0.15555727, Log Avg loss: 0.17600764, Global Avg Loss: 0.67876324, Time: 0.0218 Steps: 91090, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001682, Sample Num: 26912, Cur Loss: 0.23950878, Cur Avg Loss: 0.15537226, Log Avg loss: 0.12443837, Global Avg Loss: 0.67870239, Time: 0.0218 Steps: 91100, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001692, Sample Num: 27072, Cur Loss: 0.07643291, Cur Avg Loss: 0.15509853, Log Avg loss: 0.10905628, Global Avg Loss: 0.67863987, Time: 0.0218 Steps: 91110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001702, Sample Num: 27232, Cur Loss: 0.23468462, Cur Avg Loss: 0.15517734, Log Avg loss: 0.16851318, Global Avg Loss: 0.67858388, Time: 0.0218 Steps: 91120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001712, Sample Num: 27392, Cur Loss: 0.39048621, Cur Avg Loss: 0.15577330, Log Avg loss: 0.25720534, Global Avg Loss: 0.67853765, Time: 0.0219 Steps: 91130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001722, Sample Num: 27552, Cur Loss: 0.07885273, Cur Avg Loss: 0.15597738, Log Avg loss: 0.19091574, Global Avg Loss: 0.67848414, Time: 0.0219 Steps: 91140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001732, Sample Num: 27712, Cur Loss: 0.10677167, Cur Avg Loss: 0.15607079, Log Avg loss: 0.17215629, Global Avg Loss: 0.67842859, Time: 0.0219 Steps: 91150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001742, Sample Num: 27872, Cur Loss: 0.08058220, Cur Avg Loss: 0.15605929, Log Avg loss: 0.15406800, Global Avg Loss: 0.67837107, Time: 0.0218 Steps: 91160, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001752, Sample Num: 28032, Cur Loss: 0.15161416, Cur Avg Loss: 0.15596320, Log Avg loss: 0.13922343, Global Avg Loss: 0.67831194, Time: 0.0218 Steps: 91170, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001762, Sample Num: 28192, Cur Loss: 0.27413243, Cur Avg Loss: 0.15614405, Log Avg loss: 0.18782941, Global Avg Loss: 0.67825814, Time: 0.0219 Steps: 91180, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001772, Sample Num: 28352, Cur Loss: 0.05669037, Cur Avg Loss: 0.15605087, Log Avg loss: 0.13963138, Global Avg Loss: 0.67819908, Time: 0.0219 Steps: 91190, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001782, Sample Num: 28512, Cur Loss: 0.04585915, Cur Avg Loss: 0.15586574, Log Avg loss: 0.12306163, Global Avg Loss: 0.67813821, Time: 0.0219 Steps: 91200, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001792, Sample Num: 28672, Cur Loss: 0.09860360, Cur Avg Loss: 0.15591241, Log Avg loss: 0.16422922, Global Avg Loss: 0.67808186, Time: 0.0261 Steps: 91210, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001802, Sample Num: 28832, Cur Loss: 0.09489506, Cur Avg Loss: 0.15585356, Log Avg loss: 0.14530748, Global Avg Loss: 0.67802346, Time: 0.0208 Steps: 91220, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001812, Sample Num: 28992, Cur Loss: 0.16599694, Cur Avg Loss: 0.15579175, Log Avg loss: 0.14465385, Global Avg Loss: 0.67796499, Time: 0.0208 Steps: 91230, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001822, Sample Num: 29152, Cur Loss: 0.04450639, Cur Avg Loss: 0.15580570, Log Avg loss: 0.15833222, Global Avg Loss: 0.67790804, Time: 0.0208 Steps: 91240, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001832, Sample Num: 29312, Cur Loss: 0.17668243, Cur Avg Loss: 0.15605452, Log Avg loss: 0.20138974, Global Avg Loss: 0.67785582, Time: 0.0208 Steps: 91250, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001842, Sample Num: 29472, Cur Loss: 0.37086764, Cur Avg Loss: 0.15671224, Log Avg loss: 0.27720646, Global Avg Loss: 0.67781192, Time: 0.0208 Steps: 91260, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001852, Sample Num: 29632, Cur Loss: 0.32303363, Cur Avg Loss: 0.15647644, Log Avg loss: 0.11304231, Global Avg Loss: 0.67775004, Time: 0.0208 Steps: 91270, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001862, Sample Num: 29792, Cur Loss: 0.17070234, Cur Avg Loss: 0.15703276, Log Avg loss: 0.26006375, Global Avg Loss: 0.67770428, Time: 0.0208 Steps: 91280, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001872, Sample Num: 29952, Cur Loss: 0.29654503, Cur Avg Loss: 0.15696108, Log Avg loss: 0.14361510, Global Avg Loss: 0.67764578, Time: 0.0208 Steps: 91290, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001882, Sample Num: 30112, Cur Loss: 0.05561685, Cur Avg Loss: 0.15729289, Log Avg loss: 0.21940665, Global Avg Loss: 0.67759559, Time: 0.0208 Steps: 91300, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001892, Sample Num: 30272, Cur Loss: 0.16996215, Cur Avg Loss: 0.15744573, Log Avg loss: 0.18621047, Global Avg Loss: 0.67754177, Time: 0.0208 Steps: 91310, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001902, Sample Num: 30432, Cur Loss: 0.30507606, Cur Avg Loss: 0.15754178, Log Avg loss: 0.17571414, Global Avg Loss: 0.67748682, Time: 0.0208 Steps: 91320, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001912, Sample Num: 30592, Cur Loss: 0.11143261, Cur Avg Loss: 0.15765953, Log Avg loss: 0.18005510, Global Avg Loss: 0.67743235, Time: 0.0208 Steps: 91330, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001922, Sample Num: 30752, Cur Loss: 0.37134299, Cur Avg Loss: 0.15785716, Log Avg loss: 0.19564567, Global Avg Loss: 0.67737961, Time: 0.0208 Steps: 91340, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001932, Sample Num: 30912, Cur Loss: 0.24605945, Cur Avg Loss: 0.15790990, Log Avg loss: 0.16804505, Global Avg Loss: 0.67732385, Time: 0.0208 Steps: 91350, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001942, Sample Num: 31072, Cur Loss: 0.16547698, Cur Avg Loss: 0.15771450, Log Avg loss: 0.11996304, Global Avg Loss: 0.67726284, Time: 0.0208 Steps: 91360, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001952, Sample Num: 31232, Cur Loss: 0.33296680, Cur Avg Loss: 0.15765823, Log Avg loss: 0.14673225, Global Avg Loss: 0.67720478, Time: 0.0208 Steps: 91370, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001962, Sample Num: 31392, Cur Loss: 0.05093320, Cur Avg Loss: 0.15752307, Log Avg loss: 0.13113919, Global Avg Loss: 0.67714502, Time: 0.0208 Steps: 91380, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001972, Sample Num: 31552, Cur Loss: 0.14458369, Cur Avg Loss: 0.15731461, Log Avg loss: 0.11641411, Global Avg Loss: 0.67708366, Time: 0.0208 Steps: 91390, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001982, Sample Num: 31712, Cur Loss: 0.15676600, Cur Avg Loss: 0.15716788, Log Avg loss: 0.12823370, Global Avg Loss: 0.67702362, Time: 0.0208 Steps: 91400, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001992, Sample Num: 31872, Cur Loss: 0.05767192, Cur Avg Loss: 0.15713203, Log Avg loss: 0.15002525, Global Avg Loss: 0.67696596, Time: 0.0208 Steps: 91410, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002002, Sample Num: 32032, Cur Loss: 0.12861483, Cur Avg Loss: 0.15720587, Log Avg loss: 0.17191629, Global Avg Loss: 0.67691072, Time: 0.0208 Steps: 91420, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002012, Sample Num: 32192, Cur Loss: 0.31091231, Cur Avg Loss: 0.15769840, Log Avg loss: 0.25630188, Global Avg Loss: 0.67686472, Time: 0.0208 Steps: 91430, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002022, Sample Num: 32352, Cur Loss: 0.39890844, Cur Avg Loss: 0.15782170, Log Avg loss: 0.18263010, Global Avg Loss: 0.67681066, Time: 0.0208 Steps: 91440, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002032, Sample Num: 32512, Cur Loss: 0.36171520, Cur Avg Loss: 0.15797065, Log Avg loss: 0.18808756, Global Avg Loss: 0.67675722, Time: 0.0208 Steps: 91450, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002042, Sample Num: 32672, Cur Loss: 0.10745052, Cur Avg Loss: 0.15809793, Log Avg loss: 0.18396204, Global Avg Loss: 0.67670334, Time: 0.0208 Steps: 91460, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002052, Sample Num: 32832, Cur Loss: 0.09175254, Cur Avg Loss: 0.15794002, Log Avg loss: 0.12569451, Global Avg Loss: 0.67664310, Time: 0.0225 Steps: 91470, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002062, Sample Num: 32992, Cur Loss: 0.07196248, Cur Avg Loss: 0.15797114, Log Avg loss: 0.16435717, Global Avg Loss: 0.67658710, Time: 0.0207 Steps: 91480, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002072, Sample Num: 33152, Cur Loss: 0.14429742, Cur Avg Loss: 0.15801530, Log Avg loss: 0.16712073, Global Avg Loss: 0.67653142, Time: 0.0207 Steps: 91490, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002082, Sample Num: 33312, Cur Loss: 0.10351862, Cur Avg Loss: 0.15773954, Log Avg loss: 0.10060156, Global Avg Loss: 0.67646847, Time: 0.0208 Steps: 91500, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002092, Sample Num: 33472, Cur Loss: 0.10919959, Cur Avg Loss: 0.15761075, Log Avg loss: 0.13079808, Global Avg Loss: 0.67640885, Time: 0.0208 Steps: 91510, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002102, Sample Num: 33632, Cur Loss: 0.04228815, Cur Avg Loss: 0.15740853, Log Avg loss: 0.11510357, Global Avg Loss: 0.67634751, Time: 0.0207 Steps: 91520, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002112, Sample Num: 33792, Cur Loss: 0.06161632, Cur Avg Loss: 0.15709576, Log Avg loss: 0.09135058, Global Avg Loss: 0.67628360, Time: 0.0207 Steps: 91530, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002122, Sample Num: 33952, Cur Loss: 0.08354745, Cur Avg Loss: 0.15719417, Log Avg loss: 0.17797826, Global Avg Loss: 0.67622916, Time: 0.0208 Steps: 91540, Updated lr: 0.000014 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.886365, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.253377, "eval_total_loss": 178.124187, "eval_mae": 0.33685, "eval_mse": 0.253476, "eval_r2": 0.838874, "eval_sp_statistic": 0.899662, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.922721, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.70749, "test_total_loss": 355.159758, "test_mae": 0.643839, "test_mse": 0.707542, "test_r2": 0.543346, "test_sp_statistic": 0.800753, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.839473, "test_ps_pvalue": 0.0, "lr": 1.413276434329066e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6761884903056701, "train_cur_epoch_loss": 334.5760065848008, "train_cur_epoch_avg_loss": 0.15715171751282328, "train_cur_epoch_time": 44.88636493682861, "train_cur_epoch_avg_time": 0.02108330903561701, "epoch": 43, "step": 91547} ################################################## Training, Epoch: 0044, Batch: 000003, Sample Num: 48, Cur Loss: 0.08503874, Cur Avg Loss: 0.15083285, Log Avg loss: 0.14624850, Global Avg Loss: 0.67617127, Time: 0.0245 Steps: 91550, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000013, Sample Num: 208, Cur Loss: 0.14375132, Cur Avg Loss: 0.12162428, Log Avg loss: 0.11286170, Global Avg Loss: 0.67610975, Time: 0.0208 Steps: 91560, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000023, Sample Num: 368, Cur Loss: 0.11384305, Cur Avg Loss: 0.11402424, Log Avg loss: 0.10414420, Global Avg Loss: 0.67604729, Time: 0.0208 Steps: 91570, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000033, Sample Num: 528, Cur Loss: 0.12608379, Cur Avg Loss: 0.13817728, Log Avg loss: 0.19372927, Global Avg Loss: 0.67599462, Time: 0.0208 Steps: 91580, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000043, Sample Num: 688, Cur Loss: 0.11776055, Cur Avg Loss: 0.13935442, Log Avg loss: 0.14323899, Global Avg Loss: 0.67593646, Time: 0.0208 Steps: 91590, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000053, Sample Num: 848, Cur Loss: 0.11921972, Cur Avg Loss: 0.13923337, Log Avg loss: 0.13871283, Global Avg Loss: 0.67587781, Time: 0.0208 Steps: 91600, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000063, Sample Num: 1008, Cur Loss: 0.02969469, Cur Avg Loss: 0.13766772, Log Avg loss: 0.12936979, Global Avg Loss: 0.67581815, Time: 0.0208 Steps: 91610, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000073, Sample Num: 1168, Cur Loss: 0.06706771, Cur Avg Loss: 0.13455223, Log Avg loss: 0.11492465, Global Avg Loss: 0.67575693, Time: 0.0208 Steps: 91620, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000083, Sample Num: 1328, Cur Loss: 0.04364028, Cur Avg Loss: 0.13760712, Log Avg loss: 0.15990782, Global Avg Loss: 0.67570063, Time: 0.0208 Steps: 91630, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000093, Sample Num: 1488, Cur Loss: 0.04701474, Cur Avg Loss: 0.13583931, Log Avg loss: 0.12116644, Global Avg Loss: 0.67564012, Time: 0.0208 Steps: 91640, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000103, Sample Num: 1648, Cur Loss: 0.01858969, Cur Avg Loss: 0.13871961, Log Avg loss: 0.16550642, Global Avg Loss: 0.67558446, Time: 0.0207 Steps: 91650, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000113, Sample Num: 1808, Cur Loss: 0.11649363, Cur Avg Loss: 0.14134679, Log Avg loss: 0.16840670, Global Avg Loss: 0.67552913, Time: 0.0207 Steps: 91660, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000123, Sample Num: 1968, Cur Loss: 0.16422650, Cur Avg Loss: 0.14040918, Log Avg loss: 0.12981428, Global Avg Loss: 0.67546960, Time: 0.0207 Steps: 91670, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000133, Sample Num: 2128, Cur Loss: 0.19599992, Cur Avg Loss: 0.13904391, Log Avg loss: 0.12225107, Global Avg Loss: 0.67540926, Time: 0.0208 Steps: 91680, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000143, Sample Num: 2288, Cur Loss: 0.11678424, Cur Avg Loss: 0.13835991, Log Avg loss: 0.12926265, Global Avg Loss: 0.67534969, Time: 0.0207 Steps: 91690, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000153, Sample Num: 2448, Cur Loss: 0.10071416, Cur Avg Loss: 0.14047136, Log Avg loss: 0.17066507, Global Avg Loss: 0.67529465, Time: 0.0207 Steps: 91700, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000163, Sample Num: 2608, Cur Loss: 0.15458129, Cur Avg Loss: 0.13942299, Log Avg loss: 0.12338296, Global Avg Loss: 0.67523447, Time: 0.0207 Steps: 91710, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000173, Sample Num: 2768, Cur Loss: 0.06883484, Cur Avg Loss: 0.14117743, Log Avg loss: 0.16977488, Global Avg Loss: 0.67517937, Time: 0.0207 Steps: 91720, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000183, Sample Num: 2928, Cur Loss: 0.05711849, Cur Avg Loss: 0.14482226, Log Avg loss: 0.20787787, Global Avg Loss: 0.67512842, Time: 0.0207 Steps: 91730, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000193, Sample Num: 3088, Cur Loss: 0.10963282, Cur Avg Loss: 0.14496912, Log Avg loss: 0.14765650, Global Avg Loss: 0.67507093, Time: 0.0207 Steps: 91740, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000203, Sample Num: 3248, Cur Loss: 0.11679079, Cur Avg Loss: 0.14415855, Log Avg loss: 0.12851457, Global Avg Loss: 0.67501136, Time: 0.0207 Steps: 91750, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000213, Sample Num: 3408, Cur Loss: 0.12789644, Cur Avg Loss: 0.14567414, Log Avg loss: 0.17644061, Global Avg Loss: 0.67495702, Time: 0.0207 Steps: 91760, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000223, Sample Num: 3568, Cur Loss: 0.03735819, Cur Avg Loss: 0.15060484, Log Avg loss: 0.25562883, Global Avg Loss: 0.67491133, Time: 0.0207 Steps: 91770, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000233, Sample Num: 3728, Cur Loss: 0.05290917, Cur Avg Loss: 0.14982041, Log Avg loss: 0.13232758, Global Avg Loss: 0.67485221, Time: 0.0207 Steps: 91780, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000243, Sample Num: 3888, Cur Loss: 0.25703013, Cur Avg Loss: 0.14855815, Log Avg loss: 0.11914757, Global Avg Loss: 0.67479167, Time: 0.0207 Steps: 91790, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000253, Sample Num: 4048, Cur Loss: 0.26723659, Cur Avg Loss: 0.15112224, Log Avg loss: 0.21342966, Global Avg Loss: 0.67474141, Time: 0.0208 Steps: 91800, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000263, Sample Num: 4208, Cur Loss: 0.37568140, Cur Avg Loss: 0.15007255, Log Avg loss: 0.12351543, Global Avg Loss: 0.67468137, Time: 0.0209 Steps: 91810, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000273, Sample Num: 4368, Cur Loss: 0.14570951, Cur Avg Loss: 0.14930764, Log Avg loss: 0.12919049, Global Avg Loss: 0.67462196, Time: 0.0208 Steps: 91820, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000283, Sample Num: 4528, Cur Loss: 0.06882012, Cur Avg Loss: 0.14845610, Log Avg loss: 0.12520905, Global Avg Loss: 0.67456213, Time: 0.0209 Steps: 91830, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000293, Sample Num: 4688, Cur Loss: 0.04231765, Cur Avg Loss: 0.14874214, Log Avg loss: 0.15683690, Global Avg Loss: 0.67450576, Time: 0.0208 Steps: 91840, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000303, Sample Num: 4848, Cur Loss: 0.09987643, Cur Avg Loss: 0.14843177, Log Avg loss: 0.13933793, Global Avg Loss: 0.67444750, Time: 0.0208 Steps: 91850, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000313, Sample Num: 5008, Cur Loss: 0.04296017, Cur Avg Loss: 0.14952213, Log Avg loss: 0.18255998, Global Avg Loss: 0.67439395, Time: 0.0208 Steps: 91860, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000323, Sample Num: 5168, Cur Loss: 0.17315082, Cur Avg Loss: 0.14817683, Log Avg loss: 0.10606893, Global Avg Loss: 0.67433209, Time: 0.0209 Steps: 91870, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000333, Sample Num: 5328, Cur Loss: 0.10703602, Cur Avg Loss: 0.14729514, Log Avg loss: 0.11881668, Global Avg Loss: 0.67427163, Time: 0.0208 Steps: 91880, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000343, Sample Num: 5488, Cur Loss: 0.37990230, Cur Avg Loss: 0.14728880, Log Avg loss: 0.14707786, Global Avg Loss: 0.67421425, Time: 0.0208 Steps: 91890, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000353, Sample Num: 5648, Cur Loss: 0.22648008, Cur Avg Loss: 0.14703728, Log Avg loss: 0.13840993, Global Avg Loss: 0.67415595, Time: 0.0209 Steps: 91900, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000363, Sample Num: 5808, Cur Loss: 0.11299181, Cur Avg Loss: 0.14720511, Log Avg loss: 0.15312972, Global Avg Loss: 0.67409926, Time: 0.0208 Steps: 91910, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000373, Sample Num: 5968, Cur Loss: 0.14420822, Cur Avg Loss: 0.14826320, Log Avg loss: 0.18667158, Global Avg Loss: 0.67404623, Time: 0.0209 Steps: 91920, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000383, Sample Num: 6128, Cur Loss: 0.11700778, Cur Avg Loss: 0.14784278, Log Avg loss: 0.13216127, Global Avg Loss: 0.67398729, Time: 0.0208 Steps: 91930, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000393, Sample Num: 6288, Cur Loss: 0.23375149, Cur Avg Loss: 0.14732613, Log Avg loss: 0.12753830, Global Avg Loss: 0.67392785, Time: 0.0208 Steps: 91940, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000403, Sample Num: 6448, Cur Loss: 0.10555916, Cur Avg Loss: 0.14886943, Log Avg loss: 0.20952118, Global Avg Loss: 0.67387735, Time: 0.0208 Steps: 91950, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000413, Sample Num: 6608, Cur Loss: 0.43348169, Cur Avg Loss: 0.14969682, Log Avg loss: 0.18304074, Global Avg Loss: 0.67382397, Time: 0.0208 Steps: 91960, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000423, Sample Num: 6768, Cur Loss: 0.04216169, Cur Avg Loss: 0.14925796, Log Avg loss: 0.13113294, Global Avg Loss: 0.67376497, Time: 0.0209 Steps: 91970, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000433, Sample Num: 6928, Cur Loss: 0.12873188, Cur Avg Loss: 0.14900533, Log Avg loss: 0.13831899, Global Avg Loss: 0.67370675, Time: 0.0208 Steps: 91980, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000443, Sample Num: 7088, Cur Loss: 0.09799140, Cur Avg Loss: 0.14812091, Log Avg loss: 0.10982559, Global Avg Loss: 0.67364545, Time: 0.0208 Steps: 91990, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000453, Sample Num: 7248, Cur Loss: 0.07014973, Cur Avg Loss: 0.14779445, Log Avg loss: 0.13333239, Global Avg Loss: 0.67358672, Time: 0.0209 Steps: 92000, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000463, Sample Num: 7408, Cur Loss: 0.04069555, Cur Avg Loss: 0.14732226, Log Avg loss: 0.12593218, Global Avg Loss: 0.67352720, Time: 0.0208 Steps: 92010, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000473, Sample Num: 7568, Cur Loss: 0.05038162, Cur Avg Loss: 0.14719360, Log Avg loss: 0.14123648, Global Avg Loss: 0.67346936, Time: 0.0208 Steps: 92020, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000483, Sample Num: 7728, Cur Loss: 0.06916270, Cur Avg Loss: 0.14696470, Log Avg loss: 0.13613762, Global Avg Loss: 0.67341097, Time: 0.0208 Steps: 92030, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000493, Sample Num: 7888, Cur Loss: 0.27571124, Cur Avg Loss: 0.14868301, Log Avg loss: 0.23167752, Global Avg Loss: 0.67336298, Time: 0.0208 Steps: 92040, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000503, Sample Num: 8048, Cur Loss: 0.10910949, Cur Avg Loss: 0.14832410, Log Avg loss: 0.13062967, Global Avg Loss: 0.67330402, Time: 0.0209 Steps: 92050, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000513, Sample Num: 8208, Cur Loss: 0.28826198, Cur Avg Loss: 0.14835402, Log Avg loss: 0.14985934, Global Avg Loss: 0.67324716, Time: 0.0243 Steps: 92060, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000523, Sample Num: 8368, Cur Loss: 0.01855564, Cur Avg Loss: 0.14790780, Log Avg loss: 0.12501632, Global Avg Loss: 0.67318761, Time: 0.0209 Steps: 92070, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000533, Sample Num: 8528, Cur Loss: 0.17065492, Cur Avg Loss: 0.14752813, Log Avg loss: 0.12767147, Global Avg Loss: 0.67312837, Time: 0.0209 Steps: 92080, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000543, Sample Num: 8688, Cur Loss: 0.31926715, Cur Avg Loss: 0.14785761, Log Avg loss: 0.16541891, Global Avg Loss: 0.67307324, Time: 0.0209 Steps: 92090, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000553, Sample Num: 8848, Cur Loss: 0.04392187, Cur Avg Loss: 0.14725181, Log Avg loss: 0.11435679, Global Avg Loss: 0.67301257, Time: 0.0208 Steps: 92100, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000563, Sample Num: 9008, Cur Loss: 0.11993250, Cur Avg Loss: 0.14687863, Log Avg loss: 0.12624193, Global Avg Loss: 0.67295321, Time: 0.0208 Steps: 92110, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000573, Sample Num: 9168, Cur Loss: 0.09602001, Cur Avg Loss: 0.14628162, Log Avg loss: 0.11266992, Global Avg Loss: 0.67289239, Time: 0.0208 Steps: 92120, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000583, Sample Num: 9328, Cur Loss: 0.09483418, Cur Avg Loss: 0.14601566, Log Avg loss: 0.13077623, Global Avg Loss: 0.67283355, Time: 0.0208 Steps: 92130, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000593, Sample Num: 9488, Cur Loss: 0.16138715, Cur Avg Loss: 0.14595671, Log Avg loss: 0.14252006, Global Avg Loss: 0.67277599, Time: 0.0208 Steps: 92140, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000603, Sample Num: 9648, Cur Loss: 0.17479429, Cur Avg Loss: 0.14570492, Log Avg loss: 0.13077338, Global Avg Loss: 0.67271718, Time: 0.0209 Steps: 92150, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000613, Sample Num: 9808, Cur Loss: 0.08872240, Cur Avg Loss: 0.14584151, Log Avg loss: 0.15407792, Global Avg Loss: 0.67266090, Time: 0.0209 Steps: 92160, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000623, Sample Num: 9968, Cur Loss: 0.14882548, Cur Avg Loss: 0.14590476, Log Avg loss: 0.14978213, Global Avg Loss: 0.67260417, Time: 0.0209 Steps: 92170, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000633, Sample Num: 10128, Cur Loss: 0.23823105, Cur Avg Loss: 0.14549679, Log Avg loss: 0.12008010, Global Avg Loss: 0.67254423, Time: 0.0208 Steps: 92180, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000643, Sample Num: 10288, Cur Loss: 0.05354398, Cur Avg Loss: 0.14634141, Log Avg loss: 0.19980576, Global Avg Loss: 0.67249295, Time: 0.0209 Steps: 92190, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000653, Sample Num: 10448, Cur Loss: 0.11421418, Cur Avg Loss: 0.14733112, Log Avg loss: 0.21096991, Global Avg Loss: 0.67244290, Time: 0.0209 Steps: 92200, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000663, Sample Num: 10608, Cur Loss: 0.36915600, Cur Avg Loss: 0.14743838, Log Avg loss: 0.15444264, Global Avg Loss: 0.67238672, Time: 0.0209 Steps: 92210, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000673, Sample Num: 10768, Cur Loss: 0.07754180, Cur Avg Loss: 0.14763930, Log Avg loss: 0.16095971, Global Avg Loss: 0.67233126, Time: 0.0207 Steps: 92220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000683, Sample Num: 10928, Cur Loss: 0.02775536, Cur Avg Loss: 0.14804794, Log Avg loss: 0.17554987, Global Avg Loss: 0.67227740, Time: 0.0208 Steps: 92230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000693, Sample Num: 11088, Cur Loss: 0.05051874, Cur Avg Loss: 0.14820434, Log Avg loss: 0.15888650, Global Avg Loss: 0.67222174, Time: 0.0208 Steps: 92240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000703, Sample Num: 11248, Cur Loss: 0.06017115, Cur Avg Loss: 0.14791328, Log Avg loss: 0.12774260, Global Avg Loss: 0.67216272, Time: 0.0209 Steps: 92250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000713, Sample Num: 11408, Cur Loss: 0.27053207, Cur Avg Loss: 0.14819984, Log Avg loss: 0.16834488, Global Avg Loss: 0.67210811, Time: 0.0209 Steps: 92260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000723, Sample Num: 11568, Cur Loss: 0.08972158, Cur Avg Loss: 0.14785172, Log Avg loss: 0.12303066, Global Avg Loss: 0.67204860, Time: 0.0209 Steps: 92270, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000733, Sample Num: 11728, Cur Loss: 0.06680550, Cur Avg Loss: 0.14719921, Log Avg loss: 0.10002289, Global Avg Loss: 0.67198661, Time: 0.0207 Steps: 92280, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000743, Sample Num: 11888, Cur Loss: 0.21568832, Cur Avg Loss: 0.14772495, Log Avg loss: 0.18626153, Global Avg Loss: 0.67193398, Time: 0.0209 Steps: 92290, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000753, Sample Num: 12048, Cur Loss: 0.23179132, Cur Avg Loss: 0.14871401, Log Avg loss: 0.22220165, Global Avg Loss: 0.67188526, Time: 0.0207 Steps: 92300, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000763, Sample Num: 12208, Cur Loss: 0.12991397, Cur Avg Loss: 0.14818901, Log Avg loss: 0.10865643, Global Avg Loss: 0.67182424, Time: 0.0209 Steps: 92310, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000773, Sample Num: 12368, Cur Loss: 0.06055085, Cur Avg Loss: 0.14871108, Log Avg loss: 0.18854475, Global Avg Loss: 0.67177190, Time: 0.0209 Steps: 92320, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000783, Sample Num: 12528, Cur Loss: 0.14136805, Cur Avg Loss: 0.14833670, Log Avg loss: 0.11939758, Global Avg Loss: 0.67171207, Time: 0.0207 Steps: 92330, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000793, Sample Num: 12688, Cur Loss: 0.19850701, Cur Avg Loss: 0.14848094, Log Avg loss: 0.15977471, Global Avg Loss: 0.67165663, Time: 0.0208 Steps: 92340, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000803, Sample Num: 12848, Cur Loss: 0.09630750, Cur Avg Loss: 0.14846135, Log Avg loss: 0.14690806, Global Avg Loss: 0.67159981, Time: 0.0207 Steps: 92350, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000813, Sample Num: 13008, Cur Loss: 0.37558421, Cur Avg Loss: 0.14861251, Log Avg loss: 0.16075035, Global Avg Loss: 0.67154450, Time: 0.0207 Steps: 92360, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000823, Sample Num: 13168, Cur Loss: 0.17400113, Cur Avg Loss: 0.14905232, Log Avg loss: 0.18480848, Global Avg Loss: 0.67149180, Time: 0.0207 Steps: 92370, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000833, Sample Num: 13328, Cur Loss: 0.02337911, Cur Avg Loss: 0.14871616, Log Avg loss: 0.12105045, Global Avg Loss: 0.67143222, Time: 0.0207 Steps: 92380, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000843, Sample Num: 13488, Cur Loss: 0.15410611, Cur Avg Loss: 0.14927399, Log Avg loss: 0.19574134, Global Avg Loss: 0.67138073, Time: 0.0207 Steps: 92390, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000853, Sample Num: 13648, Cur Loss: 0.45512721, Cur Avg Loss: 0.14998009, Log Avg loss: 0.20950428, Global Avg Loss: 0.67133074, Time: 0.0207 Steps: 92400, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000863, Sample Num: 13808, Cur Loss: 0.04806709, Cur Avg Loss: 0.14945117, Log Avg loss: 0.10433420, Global Avg Loss: 0.67126939, Time: 0.0207 Steps: 92410, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000873, Sample Num: 13968, Cur Loss: 0.14867365, Cur Avg Loss: 0.14934956, Log Avg loss: 0.14058059, Global Avg Loss: 0.67121197, Time: 0.0207 Steps: 92420, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000883, Sample Num: 14128, Cur Loss: 0.11528414, Cur Avg Loss: 0.14900657, Log Avg loss: 0.11906333, Global Avg Loss: 0.67115223, Time: 0.0207 Steps: 92430, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000893, Sample Num: 14288, Cur Loss: 0.18381333, Cur Avg Loss: 0.14932895, Log Avg loss: 0.17779586, Global Avg Loss: 0.67109886, Time: 0.0207 Steps: 92440, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000903, Sample Num: 14448, Cur Loss: 0.06688973, Cur Avg Loss: 0.14916114, Log Avg loss: 0.13417536, Global Avg Loss: 0.67104078, Time: 0.0207 Steps: 92450, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000913, Sample Num: 14608, Cur Loss: 0.30916208, Cur Avg Loss: 0.15033617, Log Avg loss: 0.25644137, Global Avg Loss: 0.67099594, Time: 0.0207 Steps: 92460, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000923, Sample Num: 14768, Cur Loss: 0.11454871, Cur Avg Loss: 0.15044687, Log Avg loss: 0.16055356, Global Avg Loss: 0.67094074, Time: 0.0207 Steps: 92470, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000933, Sample Num: 14928, Cur Loss: 0.05028012, Cur Avg Loss: 0.14986391, Log Avg loss: 0.09605714, Global Avg Loss: 0.67087858, Time: 0.0207 Steps: 92480, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000943, Sample Num: 15088, Cur Loss: 0.06479097, Cur Avg Loss: 0.14983963, Log Avg loss: 0.14757374, Global Avg Loss: 0.67082200, Time: 0.0207 Steps: 92490, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000953, Sample Num: 15248, Cur Loss: 0.14692977, Cur Avg Loss: 0.14962108, Log Avg loss: 0.12901176, Global Avg Loss: 0.67076342, Time: 0.0207 Steps: 92500, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000963, Sample Num: 15408, Cur Loss: 0.09495839, Cur Avg Loss: 0.14938351, Log Avg loss: 0.12674368, Global Avg Loss: 0.67070462, Time: 0.0207 Steps: 92510, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000973, Sample Num: 15568, Cur Loss: 0.22923464, Cur Avg Loss: 0.14947388, Log Avg loss: 0.15817613, Global Avg Loss: 0.67064922, Time: 0.0207 Steps: 92520, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000983, Sample Num: 15728, Cur Loss: 0.13855276, Cur Avg Loss: 0.14927434, Log Avg loss: 0.12985891, Global Avg Loss: 0.67059077, Time: 0.0207 Steps: 92530, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000993, Sample Num: 15888, Cur Loss: 0.13382904, Cur Avg Loss: 0.14918922, Log Avg loss: 0.14082244, Global Avg Loss: 0.67053353, Time: 0.0207 Steps: 92540, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001003, Sample Num: 16048, Cur Loss: 0.27575326, Cur Avg Loss: 0.14919828, Log Avg loss: 0.15009756, Global Avg Loss: 0.67047729, Time: 0.0207 Steps: 92550, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001013, Sample Num: 16208, Cur Loss: 0.01497961, Cur Avg Loss: 0.14935723, Log Avg loss: 0.16530017, Global Avg Loss: 0.67042272, Time: 0.0207 Steps: 92560, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001023, Sample Num: 16368, Cur Loss: 0.04738491, Cur Avg Loss: 0.14989623, Log Avg loss: 0.20449686, Global Avg Loss: 0.67037238, Time: 0.0208 Steps: 92570, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001033, Sample Num: 16528, Cur Loss: 0.23360631, Cur Avg Loss: 0.14999259, Log Avg loss: 0.15985073, Global Avg Loss: 0.67031724, Time: 0.0208 Steps: 92580, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001043, Sample Num: 16688, Cur Loss: 0.06124366, Cur Avg Loss: 0.14969200, Log Avg loss: 0.11864089, Global Avg Loss: 0.67025766, Time: 0.0207 Steps: 92590, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001053, Sample Num: 16848, Cur Loss: 0.14888567, Cur Avg Loss: 0.14959083, Log Avg loss: 0.13903823, Global Avg Loss: 0.67020029, Time: 0.0207 Steps: 92600, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001063, Sample Num: 17008, Cur Loss: 0.09237222, Cur Avg Loss: 0.14962805, Log Avg loss: 0.15354714, Global Avg Loss: 0.67014450, Time: 0.0210 Steps: 92610, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001073, Sample Num: 17168, Cur Loss: 0.28785783, Cur Avg Loss: 0.14973883, Log Avg loss: 0.16151527, Global Avg Loss: 0.67008959, Time: 0.0210 Steps: 92620, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001083, Sample Num: 17328, Cur Loss: 0.23034512, Cur Avg Loss: 0.15012710, Log Avg loss: 0.19178856, Global Avg Loss: 0.67003795, Time: 0.0210 Steps: 92630, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001093, Sample Num: 17488, Cur Loss: 0.18521452, Cur Avg Loss: 0.15000413, Log Avg loss: 0.13668628, Global Avg Loss: 0.66998038, Time: 0.0210 Steps: 92640, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001103, Sample Num: 17648, Cur Loss: 0.14035428, Cur Avg Loss: 0.14966214, Log Avg loss: 0.11228259, Global Avg Loss: 0.66992018, Time: 0.0210 Steps: 92650, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001113, Sample Num: 17808, Cur Loss: 0.18375972, Cur Avg Loss: 0.14996768, Log Avg loss: 0.18366868, Global Avg Loss: 0.66986771, Time: 0.0210 Steps: 92660, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001123, Sample Num: 17968, Cur Loss: 0.14246452, Cur Avg Loss: 0.15024104, Log Avg loss: 0.18066599, Global Avg Loss: 0.66981492, Time: 0.0210 Steps: 92670, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001133, Sample Num: 18128, Cur Loss: 0.35007763, Cur Avg Loss: 0.15027475, Log Avg loss: 0.15406080, Global Avg Loss: 0.66975927, Time: 0.0210 Steps: 92680, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001143, Sample Num: 18288, Cur Loss: 0.22573404, Cur Avg Loss: 0.15045523, Log Avg loss: 0.17090317, Global Avg Loss: 0.66970545, Time: 0.0210 Steps: 92690, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001153, Sample Num: 18448, Cur Loss: 0.51602590, Cur Avg Loss: 0.15077751, Log Avg loss: 0.18761373, Global Avg Loss: 0.66965344, Time: 0.0210 Steps: 92700, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001163, Sample Num: 18608, Cur Loss: 0.06260911, Cur Avg Loss: 0.15062445, Log Avg loss: 0.13297756, Global Avg Loss: 0.66959556, Time: 0.0210 Steps: 92710, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001173, Sample Num: 18768, Cur Loss: 0.13055877, Cur Avg Loss: 0.15085459, Log Avg loss: 0.17761981, Global Avg Loss: 0.66954250, Time: 0.0210 Steps: 92720, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001183, Sample Num: 18928, Cur Loss: 0.02997366, Cur Avg Loss: 0.15114490, Log Avg loss: 0.18519762, Global Avg Loss: 0.66949026, Time: 0.0210 Steps: 92730, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001193, Sample Num: 19088, Cur Loss: 0.17306152, Cur Avg Loss: 0.15123139, Log Avg loss: 0.16146392, Global Avg Loss: 0.66943548, Time: 0.0210 Steps: 92740, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001203, Sample Num: 19248, Cur Loss: 0.13725418, Cur Avg Loss: 0.15139312, Log Avg loss: 0.17068727, Global Avg Loss: 0.66938171, Time: 0.0210 Steps: 92750, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001213, Sample Num: 19408, Cur Loss: 0.38028067, Cur Avg Loss: 0.15160183, Log Avg loss: 0.17670931, Global Avg Loss: 0.66932860, Time: 0.0210 Steps: 92760, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001223, Sample Num: 19568, Cur Loss: 0.15109977, Cur Avg Loss: 0.15134319, Log Avg loss: 0.11996973, Global Avg Loss: 0.66926938, Time: 0.0210 Steps: 92770, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001233, Sample Num: 19728, Cur Loss: 0.08373673, Cur Avg Loss: 0.15200595, Log Avg loss: 0.23306225, Global Avg Loss: 0.66922237, Time: 0.0210 Steps: 92780, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001243, Sample Num: 19888, Cur Loss: 0.18034221, Cur Avg Loss: 0.15201318, Log Avg loss: 0.15290484, Global Avg Loss: 0.66916672, Time: 0.0210 Steps: 92790, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001253, Sample Num: 20048, Cur Loss: 0.10446128, Cur Avg Loss: 0.15236755, Log Avg loss: 0.19641522, Global Avg Loss: 0.66911578, Time: 0.0210 Steps: 92800, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001263, Sample Num: 20208, Cur Loss: 0.05421206, Cur Avg Loss: 0.15193938, Log Avg loss: 0.09829029, Global Avg Loss: 0.66905427, Time: 0.0210 Steps: 92810, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001273, Sample Num: 20368, Cur Loss: 0.12826887, Cur Avg Loss: 0.15211218, Log Avg loss: 0.17393656, Global Avg Loss: 0.66900093, Time: 0.0210 Steps: 92820, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001283, Sample Num: 20528, Cur Loss: 0.07344311, Cur Avg Loss: 0.15187092, Log Avg loss: 0.12115893, Global Avg Loss: 0.66894192, Time: 0.0245 Steps: 92830, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001293, Sample Num: 20688, Cur Loss: 0.11304067, Cur Avg Loss: 0.15169756, Log Avg loss: 0.12945481, Global Avg Loss: 0.66888381, Time: 0.0209 Steps: 92840, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001303, Sample Num: 20848, Cur Loss: 0.33272591, Cur Avg Loss: 0.15153825, Log Avg loss: 0.13093919, Global Avg Loss: 0.66882587, Time: 0.0208 Steps: 92850, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001313, Sample Num: 21008, Cur Loss: 0.23014547, Cur Avg Loss: 0.15173664, Log Avg loss: 0.17758766, Global Avg Loss: 0.66877297, Time: 0.0208 Steps: 92860, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001323, Sample Num: 21168, Cur Loss: 0.23053928, Cur Avg Loss: 0.15197843, Log Avg loss: 0.18372449, Global Avg Loss: 0.66872074, Time: 0.0208 Steps: 92870, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001333, Sample Num: 21328, Cur Loss: 0.15694827, Cur Avg Loss: 0.15184282, Log Avg loss: 0.13390177, Global Avg Loss: 0.66866316, Time: 0.0208 Steps: 92880, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001343, Sample Num: 21488, Cur Loss: 0.12221821, Cur Avg Loss: 0.15186672, Log Avg loss: 0.15505304, Global Avg Loss: 0.66860787, Time: 0.0208 Steps: 92890, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001353, Sample Num: 21648, Cur Loss: 0.20527600, Cur Avg Loss: 0.15195481, Log Avg loss: 0.16378492, Global Avg Loss: 0.66855353, Time: 0.0208 Steps: 92900, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001363, Sample Num: 21808, Cur Loss: 0.10591551, Cur Avg Loss: 0.15194502, Log Avg loss: 0.15062096, Global Avg Loss: 0.66849778, Time: 0.0208 Steps: 92910, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001373, Sample Num: 21968, Cur Loss: 0.24211271, Cur Avg Loss: 0.15198549, Log Avg loss: 0.15750187, Global Avg Loss: 0.66844279, Time: 0.0208 Steps: 92920, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001383, Sample Num: 22128, Cur Loss: 0.13088353, Cur Avg Loss: 0.15230602, Log Avg loss: 0.19631374, Global Avg Loss: 0.66839198, Time: 0.0208 Steps: 92930, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001393, Sample Num: 22288, Cur Loss: 0.09951724, Cur Avg Loss: 0.15226141, Log Avg loss: 0.14609252, Global Avg Loss: 0.66833579, Time: 0.0208 Steps: 92940, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001403, Sample Num: 22448, Cur Loss: 0.19590282, Cur Avg Loss: 0.15203550, Log Avg loss: 0.12056601, Global Avg Loss: 0.66827685, Time: 0.0207 Steps: 92950, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001413, Sample Num: 22608, Cur Loss: 0.15924072, Cur Avg Loss: 0.15202247, Log Avg loss: 0.15019410, Global Avg Loss: 0.66822112, Time: 0.0208 Steps: 92960, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001423, Sample Num: 22768, Cur Loss: 0.12235937, Cur Avg Loss: 0.15214531, Log Avg loss: 0.16950338, Global Avg Loss: 0.66816748, Time: 0.0207 Steps: 92970, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001433, Sample Num: 22928, Cur Loss: 0.13723543, Cur Avg Loss: 0.15218645, Log Avg loss: 0.15803954, Global Avg Loss: 0.66811261, Time: 0.0207 Steps: 92980, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001443, Sample Num: 23088, Cur Loss: 0.12998955, Cur Avg Loss: 0.15204740, Log Avg loss: 0.13212286, Global Avg Loss: 0.66805498, Time: 0.0208 Steps: 92990, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001453, Sample Num: 23248, Cur Loss: 0.73985851, Cur Avg Loss: 0.15255937, Log Avg loss: 0.22643596, Global Avg Loss: 0.66800749, Time: 0.0208 Steps: 93000, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001463, Sample Num: 23408, Cur Loss: 0.12752876, Cur Avg Loss: 0.15272491, Log Avg loss: 0.17677751, Global Avg Loss: 0.66795467, Time: 0.0208 Steps: 93010, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001473, Sample Num: 23568, Cur Loss: 0.23361516, Cur Avg Loss: 0.15280984, Log Avg loss: 0.16523517, Global Avg Loss: 0.66790063, Time: 0.0208 Steps: 93020, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001483, Sample Num: 23728, Cur Loss: 0.12652814, Cur Avg Loss: 0.15252260, Log Avg loss: 0.11021296, Global Avg Loss: 0.66784068, Time: 0.0208 Steps: 93030, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001493, Sample Num: 23888, Cur Loss: 0.07589504, Cur Avg Loss: 0.15286401, Log Avg loss: 0.20349455, Global Avg Loss: 0.66779077, Time: 0.0207 Steps: 93040, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001503, Sample Num: 24048, Cur Loss: 0.19511159, Cur Avg Loss: 0.15278953, Log Avg loss: 0.14166902, Global Avg Loss: 0.66773423, Time: 0.0207 Steps: 93050, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001513, Sample Num: 24208, Cur Loss: 0.05606177, Cur Avg Loss: 0.15288961, Log Avg loss: 0.16793216, Global Avg Loss: 0.66768053, Time: 0.0208 Steps: 93060, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001523, Sample Num: 24368, Cur Loss: 0.11985340, Cur Avg Loss: 0.15279111, Log Avg loss: 0.13788773, Global Avg Loss: 0.66762360, Time: 0.0208 Steps: 93070, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001533, Sample Num: 24528, Cur Loss: 0.29946625, Cur Avg Loss: 0.15319742, Log Avg loss: 0.21507951, Global Avg Loss: 0.66757498, Time: 0.0207 Steps: 93080, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001543, Sample Num: 24688, Cur Loss: 0.08741664, Cur Avg Loss: 0.15318491, Log Avg loss: 0.15126681, Global Avg Loss: 0.66751952, Time: 0.0208 Steps: 93090, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001553, Sample Num: 24848, Cur Loss: 0.16239308, Cur Avg Loss: 0.15329450, Log Avg loss: 0.17020434, Global Avg Loss: 0.66746610, Time: 0.0207 Steps: 93100, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001563, Sample Num: 25008, Cur Loss: 0.34339762, Cur Avg Loss: 0.15399206, Log Avg loss: 0.26232294, Global Avg Loss: 0.66742259, Time: 0.0208 Steps: 93110, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001573, Sample Num: 25168, Cur Loss: 0.07560648, Cur Avg Loss: 0.15410382, Log Avg loss: 0.17157227, Global Avg Loss: 0.66736934, Time: 0.0207 Steps: 93120, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001583, Sample Num: 25328, Cur Loss: 0.09696572, Cur Avg Loss: 0.15389209, Log Avg loss: 0.12058676, Global Avg Loss: 0.66731063, Time: 0.0207 Steps: 93130, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001593, Sample Num: 25488, Cur Loss: 0.04594074, Cur Avg Loss: 0.15366699, Log Avg loss: 0.11803315, Global Avg Loss: 0.66725166, Time: 0.0207 Steps: 93140, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001603, Sample Num: 25648, Cur Loss: 0.19057947, Cur Avg Loss: 0.15385957, Log Avg loss: 0.18453731, Global Avg Loss: 0.66719983, Time: 0.0207 Steps: 93150, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001613, Sample Num: 25808, Cur Loss: 0.06265720, Cur Avg Loss: 0.15377031, Log Avg loss: 0.13946183, Global Avg Loss: 0.66714319, Time: 0.0208 Steps: 93160, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001623, Sample Num: 25968, Cur Loss: 0.15540235, Cur Avg Loss: 0.15381877, Log Avg loss: 0.16163592, Global Avg Loss: 0.66708893, Time: 0.0207 Steps: 93170, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001633, Sample Num: 26128, Cur Loss: 0.07142077, Cur Avg Loss: 0.15393222, Log Avg loss: 0.17234497, Global Avg Loss: 0.66703583, Time: 0.0207 Steps: 93180, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001643, Sample Num: 26288, Cur Loss: 0.04464010, Cur Avg Loss: 0.15394172, Log Avg loss: 0.15549245, Global Avg Loss: 0.66698094, Time: 0.0207 Steps: 93190, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001653, Sample Num: 26448, Cur Loss: 0.10662205, Cur Avg Loss: 0.15401147, Log Avg loss: 0.16547166, Global Avg Loss: 0.66692713, Time: 0.0208 Steps: 93200, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001663, Sample Num: 26608, Cur Loss: 0.26029515, Cur Avg Loss: 0.15384877, Log Avg loss: 0.12695464, Global Avg Loss: 0.66686920, Time: 0.0207 Steps: 93210, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001673, Sample Num: 26768, Cur Loss: 0.12682025, Cur Avg Loss: 0.15353317, Log Avg loss: 0.10104854, Global Avg Loss: 0.66680850, Time: 0.0208 Steps: 93220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001683, Sample Num: 26928, Cur Loss: 0.10189558, Cur Avg Loss: 0.15353926, Log Avg loss: 0.15455876, Global Avg Loss: 0.66675356, Time: 0.0208 Steps: 93230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001693, Sample Num: 27088, Cur Loss: 0.12931573, Cur Avg Loss: 0.15358738, Log Avg loss: 0.16168621, Global Avg Loss: 0.66669939, Time: 0.0208 Steps: 93240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001703, Sample Num: 27248, Cur Loss: 0.07685632, Cur Avg Loss: 0.15349490, Log Avg loss: 0.13783733, Global Avg Loss: 0.66664268, Time: 0.0208 Steps: 93250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001713, Sample Num: 27408, Cur Loss: 0.06852131, Cur Avg Loss: 0.15350369, Log Avg loss: 0.15500134, Global Avg Loss: 0.66658781, Time: 0.0208 Steps: 93260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001723, Sample Num: 27568, Cur Loss: 0.04801775, Cur Avg Loss: 0.15345992, Log Avg loss: 0.14596263, Global Avg Loss: 0.66653199, Time: 0.0208 Steps: 93270, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001733, Sample Num: 27728, Cur Loss: 0.22114010, Cur Avg Loss: 0.15351364, Log Avg loss: 0.16276912, Global Avg Loss: 0.66647799, Time: 0.0207 Steps: 93280, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001743, Sample Num: 27888, Cur Loss: 0.13157809, Cur Avg Loss: 0.15339208, Log Avg loss: 0.13232590, Global Avg Loss: 0.66642073, Time: 0.0207 Steps: 93290, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001753, Sample Num: 28048, Cur Loss: 0.24961579, Cur Avg Loss: 0.15326506, Log Avg loss: 0.13112535, Global Avg Loss: 0.66636336, Time: 0.0207 Steps: 93300, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001763, Sample Num: 28208, Cur Loss: 0.08091161, Cur Avg Loss: 0.15311458, Log Avg loss: 0.12673535, Global Avg Loss: 0.66630553, Time: 0.0207 Steps: 93310, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001773, Sample Num: 28368, Cur Loss: 0.24235061, Cur Avg Loss: 0.15313530, Log Avg loss: 0.15678746, Global Avg Loss: 0.66625093, Time: 0.0207 Steps: 93320, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001783, Sample Num: 28528, Cur Loss: 0.38363898, Cur Avg Loss: 0.15313501, Log Avg loss: 0.15308415, Global Avg Loss: 0.66619594, Time: 0.0208 Steps: 93330, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001793, Sample Num: 28688, Cur Loss: 0.17671233, Cur Avg Loss: 0.15320090, Log Avg loss: 0.16494991, Global Avg Loss: 0.66614224, Time: 0.0244 Steps: 93340, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001803, Sample Num: 28848, Cur Loss: 0.03441361, Cur Avg Loss: 0.15329053, Log Avg loss: 0.16935986, Global Avg Loss: 0.66608903, Time: 0.0207 Steps: 93350, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001813, Sample Num: 29008, Cur Loss: 0.13261826, Cur Avg Loss: 0.15346551, Log Avg loss: 0.18501423, Global Avg Loss: 0.66603750, Time: 0.0207 Steps: 93360, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001823, Sample Num: 29168, Cur Loss: 0.36829865, Cur Avg Loss: 0.15374199, Log Avg loss: 0.20386840, Global Avg Loss: 0.66598800, Time: 0.0207 Steps: 93370, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001833, Sample Num: 29328, Cur Loss: 0.08661086, Cur Avg Loss: 0.15374267, Log Avg loss: 0.15386646, Global Avg Loss: 0.66593316, Time: 0.0207 Steps: 93380, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001843, Sample Num: 29488, Cur Loss: 0.18048213, Cur Avg Loss: 0.15362699, Log Avg loss: 0.13242357, Global Avg Loss: 0.66587603, Time: 0.0207 Steps: 93390, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001853, Sample Num: 29648, Cur Loss: 0.04261241, Cur Avg Loss: 0.15374982, Log Avg loss: 0.17638676, Global Avg Loss: 0.66582362, Time: 0.0207 Steps: 93400, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001863, Sample Num: 29808, Cur Loss: 0.09341276, Cur Avg Loss: 0.15402081, Log Avg loss: 0.20423476, Global Avg Loss: 0.66577420, Time: 0.0207 Steps: 93410, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001873, Sample Num: 29968, Cur Loss: 0.44823587, Cur Avg Loss: 0.15425644, Log Avg loss: 0.19815535, Global Avg Loss: 0.66572415, Time: 0.0207 Steps: 93420, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001883, Sample Num: 30128, Cur Loss: 0.09677419, Cur Avg Loss: 0.15403046, Log Avg loss: 0.11170373, Global Avg Loss: 0.66566485, Time: 0.0207 Steps: 93430, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001893, Sample Num: 30288, Cur Loss: 0.10173742, Cur Avg Loss: 0.15397931, Log Avg loss: 0.14434739, Global Avg Loss: 0.66560906, Time: 0.0208 Steps: 93440, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001903, Sample Num: 30448, Cur Loss: 0.11863798, Cur Avg Loss: 0.15406445, Log Avg loss: 0.17018161, Global Avg Loss: 0.66555604, Time: 0.0207 Steps: 93450, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001913, Sample Num: 30608, Cur Loss: 0.27318469, Cur Avg Loss: 0.15414650, Log Avg loss: 0.16976033, Global Avg Loss: 0.66550300, Time: 0.0207 Steps: 93460, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001923, Sample Num: 30768, Cur Loss: 0.03158446, Cur Avg Loss: 0.15389949, Log Avg loss: 0.10664726, Global Avg Loss: 0.66544321, Time: 0.0207 Steps: 93470, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001933, Sample Num: 30928, Cur Loss: 0.51506144, Cur Avg Loss: 0.15382800, Log Avg loss: 0.14007965, Global Avg Loss: 0.66538701, Time: 0.0207 Steps: 93480, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001943, Sample Num: 31088, Cur Loss: 0.03609679, Cur Avg Loss: 0.15408981, Log Avg loss: 0.20469934, Global Avg Loss: 0.66533773, Time: 0.0207 Steps: 93490, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001953, Sample Num: 31248, Cur Loss: 0.28478330, Cur Avg Loss: 0.15408145, Log Avg loss: 0.15245686, Global Avg Loss: 0.66528287, Time: 0.0207 Steps: 93500, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001963, Sample Num: 31408, Cur Loss: 0.15223657, Cur Avg Loss: 0.15431023, Log Avg loss: 0.19898953, Global Avg Loss: 0.66523301, Time: 0.0207 Steps: 93510, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001973, Sample Num: 31568, Cur Loss: 0.10683189, Cur Avg Loss: 0.15443971, Log Avg loss: 0.17985772, Global Avg Loss: 0.66518111, Time: 0.0207 Steps: 93520, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001983, Sample Num: 31728, Cur Loss: 0.09716543, Cur Avg Loss: 0.15460462, Log Avg loss: 0.18714167, Global Avg Loss: 0.66513000, Time: 0.0207 Steps: 93530, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001993, Sample Num: 31888, Cur Loss: 0.14010428, Cur Avg Loss: 0.15452220, Log Avg loss: 0.13817756, Global Avg Loss: 0.66507366, Time: 0.0207 Steps: 93540, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002003, Sample Num: 32048, Cur Loss: 0.10637573, Cur Avg Loss: 0.15448790, Log Avg loss: 0.14765135, Global Avg Loss: 0.66501835, Time: 0.0207 Steps: 93550, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002013, Sample Num: 32208, Cur Loss: 0.20977855, Cur Avg Loss: 0.15468683, Log Avg loss: 0.19453406, Global Avg Loss: 0.66496807, Time: 0.0207 Steps: 93560, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002023, Sample Num: 32368, Cur Loss: 0.15196612, Cur Avg Loss: 0.15487863, Log Avg loss: 0.19348717, Global Avg Loss: 0.66491768, Time: 0.0207 Steps: 93570, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002033, Sample Num: 32528, Cur Loss: 0.07525533, Cur Avg Loss: 0.15493458, Log Avg loss: 0.16625266, Global Avg Loss: 0.66486439, Time: 0.0207 Steps: 93580, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002043, Sample Num: 32688, Cur Loss: 0.19144571, Cur Avg Loss: 0.15488258, Log Avg loss: 0.14431148, Global Avg Loss: 0.66480877, Time: 0.0207 Steps: 93590, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002053, Sample Num: 32848, Cur Loss: 0.05885976, Cur Avg Loss: 0.15473883, Log Avg loss: 0.12537101, Global Avg Loss: 0.66475114, Time: 0.0209 Steps: 93600, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002063, Sample Num: 33008, Cur Loss: 0.18122710, Cur Avg Loss: 0.15463024, Log Avg loss: 0.13233759, Global Avg Loss: 0.66469426, Time: 0.0209 Steps: 93610, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002073, Sample Num: 33168, Cur Loss: 0.10676385, Cur Avg Loss: 0.15458012, Log Avg loss: 0.14423864, Global Avg Loss: 0.66463867, Time: 0.0208 Steps: 93620, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002083, Sample Num: 33328, Cur Loss: 0.54806632, Cur Avg Loss: 0.15474518, Log Avg loss: 0.18896292, Global Avg Loss: 0.66458787, Time: 0.0209 Steps: 93630, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002093, Sample Num: 33488, Cur Loss: 0.18866804, Cur Avg Loss: 0.15461615, Log Avg loss: 0.12773817, Global Avg Loss: 0.66453054, Time: 0.0209 Steps: 93640, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002103, Sample Num: 33648, Cur Loss: 0.14568391, Cur Avg Loss: 0.15430826, Log Avg loss: 0.08986745, Global Avg Loss: 0.66446917, Time: 0.0208 Steps: 93650, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002113, Sample Num: 33808, Cur Loss: 0.04571968, Cur Avg Loss: 0.15422171, Log Avg loss: 0.13601986, Global Avg Loss: 0.66441275, Time: 0.0209 Steps: 93660, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002123, Sample Num: 33968, Cur Loss: 0.15826824, Cur Avg Loss: 0.15407690, Log Avg loss: 0.12347942, Global Avg Loss: 0.66435500, Time: 0.0208 Steps: 93670, Updated lr: 0.000012 ***** Running evaluation checkpoint-93676 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-93676 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.416047, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.230039, "eval_total_loss": 161.717511, "eval_mae": 0.324628, "eval_mse": 0.230124, "eval_r2": 0.853718, "eval_sp_statistic": 0.904954, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924244, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.602409, "test_total_loss": 302.409532, "test_mae": 0.567965, "test_mse": 0.602486, "test_r2": 0.61115, "test_sp_statistic": 0.801403, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.838495, "test_ps_pvalue": 0.0, "lr": 1.2113798008534852e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.664325027590356, "train_cur_epoch_loss": 328.283562541008, "train_cur_epoch_avg_loss": 0.1541961308318497, "train_cur_epoch_time": 44.41604685783386, "train_cur_epoch_avg_time": 0.02086239871199336, "epoch": 44, "step": 93676} ################################################## Training, Epoch: 0045, Batch: 000004, Sample Num: 64, Cur Loss: 0.20280871, Cur Avg Loss: 0.17384167, Log Avg loss: 0.18736693, Global Avg Loss: 0.66430408, Time: 0.0246 Steps: 93680, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000014, Sample Num: 224, Cur Loss: 0.19358528, Cur Avg Loss: 0.20090797, Log Avg loss: 0.21173448, Global Avg Loss: 0.66425578, Time: 0.0208 Steps: 93690, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000024, Sample Num: 384, Cur Loss: 0.12582880, Cur Avg Loss: 0.17274000, Log Avg loss: 0.13330485, Global Avg Loss: 0.66419911, Time: 0.0208 Steps: 93700, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000034, Sample Num: 544, Cur Loss: 0.08320896, Cur Avg Loss: 0.16298732, Log Avg loss: 0.13958089, Global Avg Loss: 0.66414313, Time: 0.0208 Steps: 93710, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000044, Sample Num: 704, Cur Loss: 0.15872402, Cur Avg Loss: 0.17787234, Log Avg loss: 0.22848140, Global Avg Loss: 0.66409665, Time: 0.0208 Steps: 93720, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000054, Sample Num: 864, Cur Loss: 0.18583983, Cur Avg Loss: 0.16638177, Log Avg loss: 0.11582326, Global Avg Loss: 0.66403815, Time: 0.0208 Steps: 93730, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000064, Sample Num: 1024, Cur Loss: 0.13653180, Cur Avg Loss: 0.16660643, Log Avg loss: 0.16781957, Global Avg Loss: 0.66398522, Time: 0.0208 Steps: 93740, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000074, Sample Num: 1184, Cur Loss: 0.05642349, Cur Avg Loss: 0.16544671, Log Avg loss: 0.15802454, Global Avg Loss: 0.66393125, Time: 0.0209 Steps: 93750, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000084, Sample Num: 1344, Cur Loss: 0.03880627, Cur Avg Loss: 0.16636126, Log Avg loss: 0.17312889, Global Avg Loss: 0.66387890, Time: 0.0208 Steps: 93760, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000094, Sample Num: 1504, Cur Loss: 0.18170556, Cur Avg Loss: 0.16235269, Log Avg loss: 0.12868077, Global Avg Loss: 0.66382182, Time: 0.0208 Steps: 93770, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000104, Sample Num: 1664, Cur Loss: 0.11250257, Cur Avg Loss: 0.16017215, Log Avg loss: 0.13967504, Global Avg Loss: 0.66376593, Time: 0.0209 Steps: 93780, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000114, Sample Num: 1824, Cur Loss: 0.10495939, Cur Avg Loss: 0.16214196, Log Avg loss: 0.18262795, Global Avg Loss: 0.66371463, Time: 0.0209 Steps: 93790, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000124, Sample Num: 1984, Cur Loss: 0.13401154, Cur Avg Loss: 0.15984414, Log Avg loss: 0.13364907, Global Avg Loss: 0.66365812, Time: 0.0209 Steps: 93800, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000134, Sample Num: 2144, Cur Loss: 0.18825692, Cur Avg Loss: 0.16300264, Log Avg loss: 0.20216795, Global Avg Loss: 0.66360893, Time: 0.0209 Steps: 93810, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000144, Sample Num: 2304, Cur Loss: 0.10530512, Cur Avg Loss: 0.16000887, Log Avg loss: 0.11989242, Global Avg Loss: 0.66355098, Time: 0.0209 Steps: 93820, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000154, Sample Num: 2464, Cur Loss: 0.15116820, Cur Avg Loss: 0.16068783, Log Avg loss: 0.17046482, Global Avg Loss: 0.66349842, Time: 0.0209 Steps: 93830, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000164, Sample Num: 2624, Cur Loss: 0.13526648, Cur Avg Loss: 0.16135220, Log Avg loss: 0.17158341, Global Avg Loss: 0.66344600, Time: 0.0209 Steps: 93840, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000174, Sample Num: 2784, Cur Loss: 0.14580522, Cur Avg Loss: 0.16097373, Log Avg loss: 0.15476692, Global Avg Loss: 0.66339180, Time: 0.0209 Steps: 93850, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000184, Sample Num: 2944, Cur Loss: 0.03790509, Cur Avg Loss: 0.15580376, Log Avg loss: 0.06584631, Global Avg Loss: 0.66332814, Time: 0.0209 Steps: 93860, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000194, Sample Num: 3104, Cur Loss: 0.18146715, Cur Avg Loss: 0.15561895, Log Avg loss: 0.15221839, Global Avg Loss: 0.66327369, Time: 0.0209 Steps: 93870, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000204, Sample Num: 3264, Cur Loss: 0.28228045, Cur Avg Loss: 0.15459276, Log Avg loss: 0.13468472, Global Avg Loss: 0.66321739, Time: 0.0209 Steps: 93880, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000214, Sample Num: 3424, Cur Loss: 0.17977802, Cur Avg Loss: 0.15495127, Log Avg loss: 0.16226475, Global Avg Loss: 0.66316403, Time: 0.0209 Steps: 93890, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000224, Sample Num: 3584, Cur Loss: 0.12191685, Cur Avg Loss: 0.15269556, Log Avg loss: 0.10442339, Global Avg Loss: 0.66310453, Time: 0.0209 Steps: 93900, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000234, Sample Num: 3744, Cur Loss: 0.07860634, Cur Avg Loss: 0.15065518, Log Avg loss: 0.10495072, Global Avg Loss: 0.66304509, Time: 0.0209 Steps: 93910, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000244, Sample Num: 3904, Cur Loss: 0.08212029, Cur Avg Loss: 0.15010147, Log Avg loss: 0.13714471, Global Avg Loss: 0.66298910, Time: 0.0209 Steps: 93920, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000254, Sample Num: 4064, Cur Loss: 0.12360622, Cur Avg Loss: 0.15133094, Log Avg loss: 0.18133004, Global Avg Loss: 0.66293782, Time: 0.0209 Steps: 93930, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000264, Sample Num: 4224, Cur Loss: 0.12593310, Cur Avg Loss: 0.15058847, Log Avg loss: 0.13172957, Global Avg Loss: 0.66288127, Time: 0.0209 Steps: 93940, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000274, Sample Num: 4384, Cur Loss: 0.06286816, Cur Avg Loss: 0.14934066, Log Avg loss: 0.11639851, Global Avg Loss: 0.66282310, Time: 0.0208 Steps: 93950, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000284, Sample Num: 4544, Cur Loss: 0.13891767, Cur Avg Loss: 0.14918268, Log Avg loss: 0.14485413, Global Avg Loss: 0.66276798, Time: 0.0209 Steps: 93960, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000294, Sample Num: 4704, Cur Loss: 0.07707408, Cur Avg Loss: 0.14783278, Log Avg loss: 0.10949569, Global Avg Loss: 0.66270910, Time: 0.0209 Steps: 93970, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000304, Sample Num: 4864, Cur Loss: 0.19623873, Cur Avg Loss: 0.14674735, Log Avg loss: 0.11483560, Global Avg Loss: 0.66265080, Time: 0.0209 Steps: 93980, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000314, Sample Num: 5024, Cur Loss: 0.23833369, Cur Avg Loss: 0.14767577, Log Avg loss: 0.17589966, Global Avg Loss: 0.66259902, Time: 0.0208 Steps: 93990, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000324, Sample Num: 5184, Cur Loss: 0.07621609, Cur Avg Loss: 0.14820824, Log Avg loss: 0.16492775, Global Avg Loss: 0.66254607, Time: 0.0208 Steps: 94000, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000334, Sample Num: 5344, Cur Loss: 0.08686426, Cur Avg Loss: 0.15017252, Log Avg loss: 0.21381524, Global Avg Loss: 0.66249834, Time: 0.0209 Steps: 94010, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000344, Sample Num: 5504, Cur Loss: 0.09359048, Cur Avg Loss: 0.15024416, Log Avg loss: 0.15263705, Global Avg Loss: 0.66244411, Time: 0.0209 Steps: 94020, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000354, Sample Num: 5664, Cur Loss: 0.03819907, Cur Avg Loss: 0.14936375, Log Avg loss: 0.11907743, Global Avg Loss: 0.66238632, Time: 0.0208 Steps: 94030, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000364, Sample Num: 5824, Cur Loss: 0.07292303, Cur Avg Loss: 0.14801250, Log Avg loss: 0.10017832, Global Avg Loss: 0.66232654, Time: 0.0209 Steps: 94040, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000374, Sample Num: 5984, Cur Loss: 0.18821128, Cur Avg Loss: 0.14746540, Log Avg loss: 0.12755109, Global Avg Loss: 0.66226968, Time: 0.0209 Steps: 94050, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000384, Sample Num: 6144, Cur Loss: 0.24459948, Cur Avg Loss: 0.14741160, Log Avg loss: 0.14539934, Global Avg Loss: 0.66221473, Time: 0.0209 Steps: 94060, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000394, Sample Num: 6304, Cur Loss: 0.19505811, Cur Avg Loss: 0.14713671, Log Avg loss: 0.13658085, Global Avg Loss: 0.66215885, Time: 0.0209 Steps: 94070, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000404, Sample Num: 6464, Cur Loss: 0.02759810, Cur Avg Loss: 0.14611852, Log Avg loss: 0.10600209, Global Avg Loss: 0.66209974, Time: 0.0209 Steps: 94080, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000414, Sample Num: 6624, Cur Loss: 0.16455019, Cur Avg Loss: 0.14757457, Log Avg loss: 0.20639903, Global Avg Loss: 0.66205130, Time: 0.0209 Steps: 94090, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000424, Sample Num: 6784, Cur Loss: 0.08881782, Cur Avg Loss: 0.14684405, Log Avg loss: 0.11660031, Global Avg Loss: 0.66199334, Time: 0.0209 Steps: 94100, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000434, Sample Num: 6944, Cur Loss: 0.08516411, Cur Avg Loss: 0.14642546, Log Avg loss: 0.12867751, Global Avg Loss: 0.66193667, Time: 0.0208 Steps: 94110, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000444, Sample Num: 7104, Cur Loss: 0.23412156, Cur Avg Loss: 0.14641178, Log Avg loss: 0.14581769, Global Avg Loss: 0.66188183, Time: 0.0209 Steps: 94120, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000454, Sample Num: 7264, Cur Loss: 0.07123387, Cur Avg Loss: 0.14657335, Log Avg loss: 0.15374749, Global Avg Loss: 0.66182785, Time: 0.0209 Steps: 94130, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000464, Sample Num: 7424, Cur Loss: 0.14137833, Cur Avg Loss: 0.14690690, Log Avg loss: 0.16204967, Global Avg Loss: 0.66177476, Time: 0.0209 Steps: 94140, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000474, Sample Num: 7584, Cur Loss: 0.02713697, Cur Avg Loss: 0.14753597, Log Avg loss: 0.17672485, Global Avg Loss: 0.66172324, Time: 0.0209 Steps: 94150, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000484, Sample Num: 7744, Cur Loss: 0.08510268, Cur Avg Loss: 0.14739221, Log Avg loss: 0.14057836, Global Avg Loss: 0.66166790, Time: 0.0209 Steps: 94160, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000494, Sample Num: 7904, Cur Loss: 0.22852167, Cur Avg Loss: 0.14714274, Log Avg loss: 0.13506798, Global Avg Loss: 0.66161198, Time: 0.0209 Steps: 94170, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000504, Sample Num: 8064, Cur Loss: 0.25000191, Cur Avg Loss: 0.14724938, Log Avg loss: 0.15251779, Global Avg Loss: 0.66155792, Time: 0.0209 Steps: 94180, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000514, Sample Num: 8224, Cur Loss: 0.09386043, Cur Avg Loss: 0.14663774, Log Avg loss: 0.11581072, Global Avg Loss: 0.66149998, Time: 0.0247 Steps: 94190, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000524, Sample Num: 8384, Cur Loss: 0.13101308, Cur Avg Loss: 0.14689706, Log Avg loss: 0.16022647, Global Avg Loss: 0.66144677, Time: 0.0208 Steps: 94200, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000534, Sample Num: 8544, Cur Loss: 0.04647938, Cur Avg Loss: 0.14555347, Log Avg loss: 0.07514909, Global Avg Loss: 0.66138453, Time: 0.0208 Steps: 94210, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000544, Sample Num: 8704, Cur Loss: 0.33184335, Cur Avg Loss: 0.14577427, Log Avg loss: 0.15756522, Global Avg Loss: 0.66133106, Time: 0.0208 Steps: 94220, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000554, Sample Num: 8864, Cur Loss: 0.15699354, Cur Avg Loss: 0.14531894, Log Avg loss: 0.12054869, Global Avg Loss: 0.66127367, Time: 0.0209 Steps: 94230, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000564, Sample Num: 9024, Cur Loss: 0.23548912, Cur Avg Loss: 0.14597290, Log Avg loss: 0.18220214, Global Avg Loss: 0.66122284, Time: 0.0208 Steps: 94240, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000574, Sample Num: 9184, Cur Loss: 0.10475774, Cur Avg Loss: 0.14562061, Log Avg loss: 0.12575191, Global Avg Loss: 0.66116602, Time: 0.0209 Steps: 94250, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000584, Sample Num: 9344, Cur Loss: 0.03268508, Cur Avg Loss: 0.14568128, Log Avg loss: 0.14916378, Global Avg Loss: 0.66111170, Time: 0.0208 Steps: 94260, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000594, Sample Num: 9504, Cur Loss: 0.03214624, Cur Avg Loss: 0.14572123, Log Avg loss: 0.14805407, Global Avg Loss: 0.66105728, Time: 0.0208 Steps: 94270, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000604, Sample Num: 9664, Cur Loss: 0.12439442, Cur Avg Loss: 0.14628107, Log Avg loss: 0.17953563, Global Avg Loss: 0.66100621, Time: 0.0208 Steps: 94280, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000614, Sample Num: 9824, Cur Loss: 0.02864086, Cur Avg Loss: 0.14581239, Log Avg loss: 0.11750428, Global Avg Loss: 0.66094856, Time: 0.0209 Steps: 94290, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000624, Sample Num: 9984, Cur Loss: 0.14932927, Cur Avg Loss: 0.14521284, Log Avg loss: 0.10840057, Global Avg Loss: 0.66088997, Time: 0.0209 Steps: 94300, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000634, Sample Num: 10144, Cur Loss: 0.22516996, Cur Avg Loss: 0.14438025, Log Avg loss: 0.09242658, Global Avg Loss: 0.66082969, Time: 0.0208 Steps: 94310, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000644, Sample Num: 10304, Cur Loss: 0.07615896, Cur Avg Loss: 0.14373462, Log Avg loss: 0.10280136, Global Avg Loss: 0.66077053, Time: 0.0209 Steps: 94320, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000654, Sample Num: 10464, Cur Loss: 0.08266123, Cur Avg Loss: 0.14333331, Log Avg loss: 0.11748875, Global Avg Loss: 0.66071294, Time: 0.0209 Steps: 94330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000664, Sample Num: 10624, Cur Loss: 0.09738033, Cur Avg Loss: 0.14372679, Log Avg loss: 0.16946058, Global Avg Loss: 0.66066086, Time: 0.0209 Steps: 94340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000674, Sample Num: 10784, Cur Loss: 0.07813394, Cur Avg Loss: 0.14456922, Log Avg loss: 0.20050632, Global Avg Loss: 0.66061209, Time: 0.0208 Steps: 94350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000684, Sample Num: 10944, Cur Loss: 0.29813585, Cur Avg Loss: 0.14509422, Log Avg loss: 0.18047953, Global Avg Loss: 0.66056121, Time: 0.0209 Steps: 94360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000694, Sample Num: 11104, Cur Loss: 0.10578051, Cur Avg Loss: 0.14515595, Log Avg loss: 0.14937822, Global Avg Loss: 0.66050704, Time: 0.0207 Steps: 94370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000704, Sample Num: 11264, Cur Loss: 0.07675309, Cur Avg Loss: 0.14576038, Log Avg loss: 0.18770785, Global Avg Loss: 0.66045695, Time: 0.0209 Steps: 94380, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000714, Sample Num: 11424, Cur Loss: 0.10447451, Cur Avg Loss: 0.14647836, Log Avg loss: 0.19702436, Global Avg Loss: 0.66040785, Time: 0.0208 Steps: 94390, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000724, Sample Num: 11584, Cur Loss: 0.10534424, Cur Avg Loss: 0.14617669, Log Avg loss: 0.12463764, Global Avg Loss: 0.66035109, Time: 0.0210 Steps: 94400, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000734, Sample Num: 11744, Cur Loss: 0.32688329, Cur Avg Loss: 0.14665321, Log Avg loss: 0.18115307, Global Avg Loss: 0.66030034, Time: 0.0209 Steps: 94410, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000744, Sample Num: 11904, Cur Loss: 0.40033111, Cur Avg Loss: 0.14712510, Log Avg loss: 0.18176139, Global Avg Loss: 0.66024965, Time: 0.0208 Steps: 94420, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000754, Sample Num: 12064, Cur Loss: 0.11733347, Cur Avg Loss: 0.14694310, Log Avg loss: 0.13340237, Global Avg Loss: 0.66019386, Time: 0.0209 Steps: 94430, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000764, Sample Num: 12224, Cur Loss: 0.15133440, Cur Avg Loss: 0.14696342, Log Avg loss: 0.14849590, Global Avg Loss: 0.66013968, Time: 0.0208 Steps: 94440, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000774, Sample Num: 12384, Cur Loss: 0.06967162, Cur Avg Loss: 0.14677258, Log Avg loss: 0.13219211, Global Avg Loss: 0.66008378, Time: 0.0210 Steps: 94450, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000784, Sample Num: 12544, Cur Loss: 0.37431276, Cur Avg Loss: 0.14665705, Log Avg loss: 0.13771520, Global Avg Loss: 0.66002848, Time: 0.0209 Steps: 94460, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000794, Sample Num: 12704, Cur Loss: 0.49894923, Cur Avg Loss: 0.14743005, Log Avg loss: 0.20803284, Global Avg Loss: 0.65998064, Time: 0.0209 Steps: 94470, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000804, Sample Num: 12864, Cur Loss: 0.04203701, Cur Avg Loss: 0.14780849, Log Avg loss: 0.17785680, Global Avg Loss: 0.65992961, Time: 0.0209 Steps: 94480, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000814, Sample Num: 13024, Cur Loss: 0.07094705, Cur Avg Loss: 0.14747491, Log Avg loss: 0.12065494, Global Avg Loss: 0.65987254, Time: 0.0209 Steps: 94490, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000824, Sample Num: 13184, Cur Loss: 0.13433407, Cur Avg Loss: 0.14728831, Log Avg loss: 0.13209913, Global Avg Loss: 0.65981669, Time: 0.0209 Steps: 94500, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000834, Sample Num: 13344, Cur Loss: 0.03640532, Cur Avg Loss: 0.14697840, Log Avg loss: 0.12144193, Global Avg Loss: 0.65975972, Time: 0.0209 Steps: 94510, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000844, Sample Num: 13504, Cur Loss: 0.04736087, Cur Avg Loss: 0.14704004, Log Avg loss: 0.15218087, Global Avg Loss: 0.65970602, Time: 0.0209 Steps: 94520, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000854, Sample Num: 13664, Cur Loss: 0.05437802, Cur Avg Loss: 0.14750603, Log Avg loss: 0.18683587, Global Avg Loss: 0.65965600, Time: 0.0209 Steps: 94530, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000864, Sample Num: 13824, Cur Loss: 0.05149392, Cur Avg Loss: 0.14769167, Log Avg loss: 0.16354550, Global Avg Loss: 0.65960352, Time: 0.0209 Steps: 94540, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000874, Sample Num: 13984, Cur Loss: 0.09656904, Cur Avg Loss: 0.14736704, Log Avg loss: 0.11931876, Global Avg Loss: 0.65954638, Time: 0.0209 Steps: 94550, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000884, Sample Num: 14144, Cur Loss: 0.10503254, Cur Avg Loss: 0.14762196, Log Avg loss: 0.16990166, Global Avg Loss: 0.65949460, Time: 0.0209 Steps: 94560, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000894, Sample Num: 14304, Cur Loss: 0.56471640, Cur Avg Loss: 0.14814352, Log Avg loss: 0.19424967, Global Avg Loss: 0.65944540, Time: 0.0209 Steps: 94570, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000904, Sample Num: 14464, Cur Loss: 0.05940325, Cur Avg Loss: 0.14786297, Log Avg loss: 0.12278192, Global Avg Loss: 0.65938866, Time: 0.0209 Steps: 94580, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000914, Sample Num: 14624, Cur Loss: 0.17718929, Cur Avg Loss: 0.14788631, Log Avg loss: 0.14999599, Global Avg Loss: 0.65933481, Time: 0.0208 Steps: 94590, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000924, Sample Num: 14784, Cur Loss: 0.02735037, Cur Avg Loss: 0.14749368, Log Avg loss: 0.11160732, Global Avg Loss: 0.65927691, Time: 0.0209 Steps: 94600, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000934, Sample Num: 14944, Cur Loss: 0.11177289, Cur Avg Loss: 0.14733634, Log Avg loss: 0.13279852, Global Avg Loss: 0.65922126, Time: 0.0210 Steps: 94610, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000944, Sample Num: 15104, Cur Loss: 0.01775007, Cur Avg Loss: 0.14742482, Log Avg loss: 0.15568880, Global Avg Loss: 0.65916804, Time: 0.0209 Steps: 94620, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000954, Sample Num: 15264, Cur Loss: 0.06695323, Cur Avg Loss: 0.14748913, Log Avg loss: 0.15355978, Global Avg Loss: 0.65911461, Time: 0.0209 Steps: 94630, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000964, Sample Num: 15424, Cur Loss: 0.17269547, Cur Avg Loss: 0.14724385, Log Avg loss: 0.12384447, Global Avg Loss: 0.65905806, Time: 0.0209 Steps: 94640, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000974, Sample Num: 15584, Cur Loss: 0.19138730, Cur Avg Loss: 0.14736178, Log Avg loss: 0.15872938, Global Avg Loss: 0.65900519, Time: 0.0209 Steps: 94650, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000984, Sample Num: 15744, Cur Loss: 0.16735058, Cur Avg Loss: 0.14755010, Log Avg loss: 0.16589246, Global Avg Loss: 0.65895310, Time: 0.0209 Steps: 94660, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000994, Sample Num: 15904, Cur Loss: 0.05778491, Cur Avg Loss: 0.14747140, Log Avg loss: 0.13972762, Global Avg Loss: 0.65889826, Time: 0.0209 Steps: 94670, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001004, Sample Num: 16064, Cur Loss: 0.14346701, Cur Avg Loss: 0.14764865, Log Avg loss: 0.16526751, Global Avg Loss: 0.65884612, Time: 0.0209 Steps: 94680, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001014, Sample Num: 16224, Cur Loss: 0.14319186, Cur Avg Loss: 0.14796456, Log Avg loss: 0.17968165, Global Avg Loss: 0.65879552, Time: 0.0209 Steps: 94690, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001024, Sample Num: 16384, Cur Loss: 0.07482996, Cur Avg Loss: 0.14767087, Log Avg loss: 0.11789073, Global Avg Loss: 0.65873840, Time: 0.0254 Steps: 94700, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001034, Sample Num: 16544, Cur Loss: 0.28532541, Cur Avg Loss: 0.14822958, Log Avg loss: 0.20544217, Global Avg Loss: 0.65869054, Time: 0.0210 Steps: 94710, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001044, Sample Num: 16704, Cur Loss: 0.26230139, Cur Avg Loss: 0.14851244, Log Avg loss: 0.17775949, Global Avg Loss: 0.65863976, Time: 0.0209 Steps: 94720, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001054, Sample Num: 16864, Cur Loss: 0.29455784, Cur Avg Loss: 0.14848547, Log Avg loss: 0.14567029, Global Avg Loss: 0.65858561, Time: 0.0209 Steps: 94730, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001064, Sample Num: 17024, Cur Loss: 0.11623890, Cur Avg Loss: 0.14845666, Log Avg loss: 0.14541965, Global Avg Loss: 0.65853145, Time: 0.0208 Steps: 94740, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001074, Sample Num: 17184, Cur Loss: 0.18845657, Cur Avg Loss: 0.14890696, Log Avg loss: 0.19681948, Global Avg Loss: 0.65848272, Time: 0.0209 Steps: 94750, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001084, Sample Num: 17344, Cur Loss: 0.20330527, Cur Avg Loss: 0.14900489, Log Avg loss: 0.15952205, Global Avg Loss: 0.65843006, Time: 0.0209 Steps: 94760, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001094, Sample Num: 17504, Cur Loss: 0.27505755, Cur Avg Loss: 0.14903246, Log Avg loss: 0.15202125, Global Avg Loss: 0.65837663, Time: 0.0209 Steps: 94770, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001104, Sample Num: 17664, Cur Loss: 0.10508925, Cur Avg Loss: 0.14903784, Log Avg loss: 0.14962587, Global Avg Loss: 0.65832295, Time: 0.0209 Steps: 94780, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001114, Sample Num: 17824, Cur Loss: 0.17728737, Cur Avg Loss: 0.14906779, Log Avg loss: 0.15237469, Global Avg Loss: 0.65826957, Time: 0.0209 Steps: 94790, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001124, Sample Num: 17984, Cur Loss: 0.07005994, Cur Avg Loss: 0.14957118, Log Avg loss: 0.20564893, Global Avg Loss: 0.65822183, Time: 0.0209 Steps: 94800, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001134, Sample Num: 18144, Cur Loss: 0.12353338, Cur Avg Loss: 0.14922421, Log Avg loss: 0.11022423, Global Avg Loss: 0.65816403, Time: 0.0209 Steps: 94810, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001144, Sample Num: 18304, Cur Loss: 0.07917823, Cur Avg Loss: 0.14911241, Log Avg loss: 0.13643455, Global Avg Loss: 0.65810901, Time: 0.0209 Steps: 94820, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001154, Sample Num: 18464, Cur Loss: 0.18417600, Cur Avg Loss: 0.14952095, Log Avg loss: 0.19625772, Global Avg Loss: 0.65806030, Time: 0.0209 Steps: 94830, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001164, Sample Num: 18624, Cur Loss: 0.17068464, Cur Avg Loss: 0.14969170, Log Avg loss: 0.16939617, Global Avg Loss: 0.65800878, Time: 0.0209 Steps: 94840, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001174, Sample Num: 18784, Cur Loss: 0.09434038, Cur Avg Loss: 0.14943640, Log Avg loss: 0.11971976, Global Avg Loss: 0.65795203, Time: 0.0209 Steps: 94850, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001184, Sample Num: 18944, Cur Loss: 0.20608076, Cur Avg Loss: 0.14982845, Log Avg loss: 0.19585561, Global Avg Loss: 0.65790331, Time: 0.0209 Steps: 94860, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001194, Sample Num: 19104, Cur Loss: 0.07166682, Cur Avg Loss: 0.14937548, Log Avg loss: 0.09574382, Global Avg Loss: 0.65784406, Time: 0.0209 Steps: 94870, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001204, Sample Num: 19264, Cur Loss: 0.09607209, Cur Avg Loss: 0.14899215, Log Avg loss: 0.10322224, Global Avg Loss: 0.65778560, Time: 0.0208 Steps: 94880, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001214, Sample Num: 19424, Cur Loss: 0.06246445, Cur Avg Loss: 0.14879057, Log Avg loss: 0.12452021, Global Avg Loss: 0.65772940, Time: 0.0209 Steps: 94890, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001224, Sample Num: 19584, Cur Loss: 0.14646219, Cur Avg Loss: 0.14917728, Log Avg loss: 0.19612347, Global Avg Loss: 0.65768076, Time: 0.0209 Steps: 94900, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001234, Sample Num: 19744, Cur Loss: 0.07991490, Cur Avg Loss: 0.14926463, Log Avg loss: 0.15995739, Global Avg Loss: 0.65762832, Time: 0.0209 Steps: 94910, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001244, Sample Num: 19904, Cur Loss: 0.28524429, Cur Avg Loss: 0.14928976, Log Avg loss: 0.15239049, Global Avg Loss: 0.65757509, Time: 0.0209 Steps: 94920, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001254, Sample Num: 20064, Cur Loss: 0.25509882, Cur Avg Loss: 0.14935251, Log Avg loss: 0.15715813, Global Avg Loss: 0.65752238, Time: 0.0209 Steps: 94930, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001264, Sample Num: 20224, Cur Loss: 0.17364113, Cur Avg Loss: 0.14940019, Log Avg loss: 0.15537928, Global Avg Loss: 0.65746949, Time: 0.0209 Steps: 94940, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001274, Sample Num: 20384, Cur Loss: 0.18566748, Cur Avg Loss: 0.14938720, Log Avg loss: 0.14774491, Global Avg Loss: 0.65741580, Time: 0.0208 Steps: 94950, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001284, Sample Num: 20544, Cur Loss: 0.18400200, Cur Avg Loss: 0.14958570, Log Avg loss: 0.17487479, Global Avg Loss: 0.65736499, Time: 0.0225 Steps: 94960, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001294, Sample Num: 20704, Cur Loss: 0.07140160, Cur Avg Loss: 0.14977213, Log Avg loss: 0.17370978, Global Avg Loss: 0.65731406, Time: 0.0209 Steps: 94970, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001304, Sample Num: 20864, Cur Loss: 0.05561203, Cur Avg Loss: 0.14984454, Log Avg loss: 0.15921502, Global Avg Loss: 0.65726162, Time: 0.0209 Steps: 94980, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001314, Sample Num: 21024, Cur Loss: 0.14690092, Cur Avg Loss: 0.14960236, Log Avg loss: 0.11802137, Global Avg Loss: 0.65720485, Time: 0.0209 Steps: 94990, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001324, Sample Num: 21184, Cur Loss: 0.17179477, Cur Avg Loss: 0.14939993, Log Avg loss: 0.12280144, Global Avg Loss: 0.65714860, Time: 0.0208 Steps: 95000, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001334, Sample Num: 21344, Cur Loss: 0.04171551, Cur Avg Loss: 0.14986286, Log Avg loss: 0.21115455, Global Avg Loss: 0.65710166, Time: 0.0209 Steps: 95010, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001344, Sample Num: 21504, Cur Loss: 0.24558905, Cur Avg Loss: 0.14975057, Log Avg loss: 0.13477151, Global Avg Loss: 0.65704669, Time: 0.0209 Steps: 95020, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001354, Sample Num: 21664, Cur Loss: 0.17775637, Cur Avg Loss: 0.14970809, Log Avg loss: 0.14399814, Global Avg Loss: 0.65699270, Time: 0.0208 Steps: 95030, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001364, Sample Num: 21824, Cur Loss: 0.09862040, Cur Avg Loss: 0.14978919, Log Avg loss: 0.16076987, Global Avg Loss: 0.65694049, Time: 0.0209 Steps: 95040, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001374, Sample Num: 21984, Cur Loss: 0.10303071, Cur Avg Loss: 0.15027441, Log Avg loss: 0.21645927, Global Avg Loss: 0.65689414, Time: 0.0209 Steps: 95050, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001384, Sample Num: 22144, Cur Loss: 0.06813063, Cur Avg Loss: 0.15026088, Log Avg loss: 0.14840118, Global Avg Loss: 0.65684065, Time: 0.0208 Steps: 95060, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001394, Sample Num: 22304, Cur Loss: 0.24191451, Cur Avg Loss: 0.15005862, Log Avg loss: 0.12206564, Global Avg Loss: 0.65678440, Time: 0.0209 Steps: 95070, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001404, Sample Num: 22464, Cur Loss: 0.21532279, Cur Avg Loss: 0.14993134, Log Avg loss: 0.13218866, Global Avg Loss: 0.65672923, Time: 0.0209 Steps: 95080, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001414, Sample Num: 22624, Cur Loss: 0.13778174, Cur Avg Loss: 0.15020330, Log Avg loss: 0.18838717, Global Avg Loss: 0.65667997, Time: 0.0208 Steps: 95090, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001424, Sample Num: 22784, Cur Loss: 0.18191212, Cur Avg Loss: 0.15040495, Log Avg loss: 0.17891711, Global Avg Loss: 0.65662974, Time: 0.0207 Steps: 95100, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001434, Sample Num: 22944, Cur Loss: 0.01712665, Cur Avg Loss: 0.15008297, Log Avg loss: 0.10423338, Global Avg Loss: 0.65657166, Time: 0.0208 Steps: 95110, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001444, Sample Num: 23104, Cur Loss: 0.12535691, Cur Avg Loss: 0.15008397, Log Avg loss: 0.15022798, Global Avg Loss: 0.65651842, Time: 0.0208 Steps: 95120, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001454, Sample Num: 23264, Cur Loss: 0.30824417, Cur Avg Loss: 0.15057479, Log Avg loss: 0.22144870, Global Avg Loss: 0.65647269, Time: 0.0207 Steps: 95130, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001464, Sample Num: 23424, Cur Loss: 0.03500016, Cur Avg Loss: 0.15036547, Log Avg loss: 0.11993122, Global Avg Loss: 0.65641630, Time: 0.0208 Steps: 95140, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001474, Sample Num: 23584, Cur Loss: 0.08885667, Cur Avg Loss: 0.15068964, Log Avg loss: 0.19814802, Global Avg Loss: 0.65636813, Time: 0.0207 Steps: 95150, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001484, Sample Num: 23744, Cur Loss: 0.19553459, Cur Avg Loss: 0.15036186, Log Avg loss: 0.10204709, Global Avg Loss: 0.65630988, Time: 0.0207 Steps: 95160, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001494, Sample Num: 23904, Cur Loss: 0.24646036, Cur Avg Loss: 0.15054480, Log Avg loss: 0.17769257, Global Avg Loss: 0.65625959, Time: 0.0207 Steps: 95170, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001504, Sample Num: 24064, Cur Loss: 0.28551042, Cur Avg Loss: 0.15036863, Log Avg loss: 0.12404919, Global Avg Loss: 0.65620367, Time: 0.0208 Steps: 95180, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001514, Sample Num: 24224, Cur Loss: 0.11220631, Cur Avg Loss: 0.15065770, Log Avg loss: 0.19413263, Global Avg Loss: 0.65615513, Time: 0.0208 Steps: 95190, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001524, Sample Num: 24384, Cur Loss: 0.14103386, Cur Avg Loss: 0.15037213, Log Avg loss: 0.10713793, Global Avg Loss: 0.65609746, Time: 0.0207 Steps: 95200, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001534, Sample Num: 24544, Cur Loss: 0.35065582, Cur Avg Loss: 0.15043599, Log Avg loss: 0.16016765, Global Avg Loss: 0.65604537, Time: 0.0208 Steps: 95210, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001544, Sample Num: 24704, Cur Loss: 0.25185758, Cur Avg Loss: 0.15023616, Log Avg loss: 0.11958242, Global Avg Loss: 0.65598904, Time: 0.0208 Steps: 95220, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001554, Sample Num: 24864, Cur Loss: 0.14279842, Cur Avg Loss: 0.15032687, Log Avg loss: 0.16433320, Global Avg Loss: 0.65593741, Time: 0.0208 Steps: 95230, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001564, Sample Num: 25024, Cur Loss: 0.24248907, Cur Avg Loss: 0.15038419, Log Avg loss: 0.15929056, Global Avg Loss: 0.65588526, Time: 0.0208 Steps: 95240, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001574, Sample Num: 25184, Cur Loss: 0.23792897, Cur Avg Loss: 0.15034728, Log Avg loss: 0.14457564, Global Avg Loss: 0.65583158, Time: 0.0208 Steps: 95250, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001584, Sample Num: 25344, Cur Loss: 0.17343883, Cur Avg Loss: 0.15026363, Log Avg loss: 0.13709631, Global Avg Loss: 0.65577712, Time: 0.0208 Steps: 95260, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001594, Sample Num: 25504, Cur Loss: 0.14342037, Cur Avg Loss: 0.15055705, Log Avg loss: 0.19703557, Global Avg Loss: 0.65572897, Time: 0.0208 Steps: 95270, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001604, Sample Num: 25664, Cur Loss: 0.18533635, Cur Avg Loss: 0.15088745, Log Avg loss: 0.20355198, Global Avg Loss: 0.65568151, Time: 0.0208 Steps: 95280, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001614, Sample Num: 25824, Cur Loss: 0.34183484, Cur Avg Loss: 0.15095081, Log Avg loss: 0.16111384, Global Avg Loss: 0.65562961, Time: 0.0208 Steps: 95290, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001624, Sample Num: 25984, Cur Loss: 0.16268960, Cur Avg Loss: 0.15086621, Log Avg loss: 0.13721170, Global Avg Loss: 0.65557522, Time: 0.0208 Steps: 95300, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001634, Sample Num: 26144, Cur Loss: 0.21544452, Cur Avg Loss: 0.15094141, Log Avg loss: 0.16315443, Global Avg Loss: 0.65552355, Time: 0.0209 Steps: 95310, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001644, Sample Num: 26304, Cur Loss: 0.21192065, Cur Avg Loss: 0.15114862, Log Avg loss: 0.18500733, Global Avg Loss: 0.65547419, Time: 0.0208 Steps: 95320, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001654, Sample Num: 26464, Cur Loss: 0.03036906, Cur Avg Loss: 0.15114416, Log Avg loss: 0.15041008, Global Avg Loss: 0.65542121, Time: 0.0208 Steps: 95330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001664, Sample Num: 26624, Cur Loss: 0.25424048, Cur Avg Loss: 0.15120810, Log Avg loss: 0.16178425, Global Avg Loss: 0.65536943, Time: 0.0208 Steps: 95340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001674, Sample Num: 26784, Cur Loss: 0.04668953, Cur Avg Loss: 0.15108849, Log Avg loss: 0.13118608, Global Avg Loss: 0.65531446, Time: 0.0208 Steps: 95350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001684, Sample Num: 26944, Cur Loss: 0.04376860, Cur Avg Loss: 0.15073425, Log Avg loss: 0.09143291, Global Avg Loss: 0.65525532, Time: 0.0208 Steps: 95360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001694, Sample Num: 27104, Cur Loss: 0.12069663, Cur Avg Loss: 0.15083300, Log Avg loss: 0.16746252, Global Avg Loss: 0.65520418, Time: 0.0208 Steps: 95370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001704, Sample Num: 27264, Cur Loss: 0.07420430, Cur Avg Loss: 0.15064100, Log Avg loss: 0.11811640, Global Avg Loss: 0.65514787, Time: 0.0208 Steps: 95380, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001714, Sample Num: 27424, Cur Loss: 0.08330399, Cur Avg Loss: 0.15081335, Log Avg loss: 0.18018198, Global Avg Loss: 0.65509807, Time: 0.0208 Steps: 95390, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001724, Sample Num: 27584, Cur Loss: 0.08866501, Cur Avg Loss: 0.15072870, Log Avg loss: 0.13622039, Global Avg Loss: 0.65504369, Time: 0.0208 Steps: 95400, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001734, Sample Num: 27744, Cur Loss: 0.18042648, Cur Avg Loss: 0.15049096, Log Avg loss: 0.10950446, Global Avg Loss: 0.65498651, Time: 0.0208 Steps: 95410, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001744, Sample Num: 27904, Cur Loss: 0.20315851, Cur Avg Loss: 0.15023396, Log Avg loss: 0.10567056, Global Avg Loss: 0.65492894, Time: 0.0208 Steps: 95420, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001754, Sample Num: 28064, Cur Loss: 0.21448204, Cur Avg Loss: 0.15022681, Log Avg loss: 0.14897872, Global Avg Loss: 0.65487592, Time: 0.0208 Steps: 95430, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001764, Sample Num: 28224, Cur Loss: 0.24540973, Cur Avg Loss: 0.15062141, Log Avg loss: 0.21983479, Global Avg Loss: 0.65483034, Time: 0.0208 Steps: 95440, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001774, Sample Num: 28384, Cur Loss: 0.16798043, Cur Avg Loss: 0.15057303, Log Avg loss: 0.14203967, Global Avg Loss: 0.65477661, Time: 0.0208 Steps: 95450, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001784, Sample Num: 28544, Cur Loss: 0.07920252, Cur Avg Loss: 0.15063523, Log Avg loss: 0.16166849, Global Avg Loss: 0.65472496, Time: 0.0208 Steps: 95460, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001794, Sample Num: 28704, Cur Loss: 0.18207422, Cur Avg Loss: 0.15093209, Log Avg loss: 0.20389170, Global Avg Loss: 0.65467774, Time: 0.0246 Steps: 95470, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001804, Sample Num: 28864, Cur Loss: 0.05819661, Cur Avg Loss: 0.15066939, Log Avg loss: 0.10354176, Global Avg Loss: 0.65462001, Time: 0.0208 Steps: 95480, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001814, Sample Num: 29024, Cur Loss: 0.31561720, Cur Avg Loss: 0.15076922, Log Avg loss: 0.16877811, Global Avg Loss: 0.65456913, Time: 0.0208 Steps: 95490, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001824, Sample Num: 29184, Cur Loss: 0.11380780, Cur Avg Loss: 0.15070320, Log Avg loss: 0.13872742, Global Avg Loss: 0.65451512, Time: 0.0208 Steps: 95500, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001834, Sample Num: 29344, Cur Loss: 0.11694203, Cur Avg Loss: 0.15081742, Log Avg loss: 0.17165071, Global Avg Loss: 0.65446456, Time: 0.0208 Steps: 95510, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001844, Sample Num: 29504, Cur Loss: 0.06717228, Cur Avg Loss: 0.15119435, Log Avg loss: 0.22032302, Global Avg Loss: 0.65441911, Time: 0.0208 Steps: 95520, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001854, Sample Num: 29664, Cur Loss: 0.05328804, Cur Avg Loss: 0.15118516, Log Avg loss: 0.14949160, Global Avg Loss: 0.65436626, Time: 0.0208 Steps: 95530, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001864, Sample Num: 29824, Cur Loss: 0.16902339, Cur Avg Loss: 0.15113501, Log Avg loss: 0.14183741, Global Avg Loss: 0.65431261, Time: 0.0208 Steps: 95540, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001874, Sample Num: 29984, Cur Loss: 0.53363848, Cur Avg Loss: 0.15124623, Log Avg loss: 0.17197633, Global Avg Loss: 0.65426213, Time: 0.0207 Steps: 95550, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001884, Sample Num: 30144, Cur Loss: 0.12526163, Cur Avg Loss: 0.15118317, Log Avg loss: 0.13936619, Global Avg Loss: 0.65420825, Time: 0.0208 Steps: 95560, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001894, Sample Num: 30304, Cur Loss: 0.29455489, Cur Avg Loss: 0.15137977, Log Avg loss: 0.18842021, Global Avg Loss: 0.65415951, Time: 0.0208 Steps: 95570, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001904, Sample Num: 30464, Cur Loss: 0.38234606, Cur Avg Loss: 0.15145155, Log Avg loss: 0.16504486, Global Avg Loss: 0.65410834, Time: 0.0208 Steps: 95580, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001914, Sample Num: 30624, Cur Loss: 0.11820889, Cur Avg Loss: 0.15157135, Log Avg loss: 0.17438250, Global Avg Loss: 0.65405815, Time: 0.0208 Steps: 95590, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001924, Sample Num: 30784, Cur Loss: 0.23734894, Cur Avg Loss: 0.15168294, Log Avg loss: 0.17304124, Global Avg Loss: 0.65400784, Time: 0.0208 Steps: 95600, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001934, Sample Num: 30944, Cur Loss: 0.44057694, Cur Avg Loss: 0.15192667, Log Avg loss: 0.19882084, Global Avg Loss: 0.65396023, Time: 0.0207 Steps: 95610, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001944, Sample Num: 31104, Cur Loss: 0.27901790, Cur Avg Loss: 0.15203427, Log Avg loss: 0.17284252, Global Avg Loss: 0.65390991, Time: 0.0208 Steps: 95620, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001954, Sample Num: 31264, Cur Loss: 0.24468513, Cur Avg Loss: 0.15194827, Log Avg loss: 0.13523118, Global Avg Loss: 0.65385568, Time: 0.0208 Steps: 95630, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001964, Sample Num: 31424, Cur Loss: 0.28543353, Cur Avg Loss: 0.15209407, Log Avg loss: 0.18058350, Global Avg Loss: 0.65380619, Time: 0.0208 Steps: 95640, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001974, Sample Num: 31584, Cur Loss: 0.06074961, Cur Avg Loss: 0.15224903, Log Avg loss: 0.18268213, Global Avg Loss: 0.65375694, Time: 0.0208 Steps: 95650, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001984, Sample Num: 31744, Cur Loss: 0.03249653, Cur Avg Loss: 0.15215745, Log Avg loss: 0.13407907, Global Avg Loss: 0.65370261, Time: 0.0207 Steps: 95660, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001994, Sample Num: 31904, Cur Loss: 0.14700906, Cur Avg Loss: 0.15225290, Log Avg loss: 0.17119108, Global Avg Loss: 0.65365217, Time: 0.0208 Steps: 95670, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002004, Sample Num: 32064, Cur Loss: 0.06283087, Cur Avg Loss: 0.15219620, Log Avg loss: 0.14089005, Global Avg Loss: 0.65359858, Time: 0.0208 Steps: 95680, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002014, Sample Num: 32224, Cur Loss: 0.29171145, Cur Avg Loss: 0.15236199, Log Avg loss: 0.18558563, Global Avg Loss: 0.65354967, Time: 0.0208 Steps: 95690, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002024, Sample Num: 32384, Cur Loss: 0.07429661, Cur Avg Loss: 0.15204852, Log Avg loss: 0.08891692, Global Avg Loss: 0.65349067, Time: 0.0208 Steps: 95700, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002034, Sample Num: 32544, Cur Loss: 0.15214950, Cur Avg Loss: 0.15197670, Log Avg loss: 0.13744076, Global Avg Loss: 0.65343676, Time: 0.0208 Steps: 95710, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002044, Sample Num: 32704, Cur Loss: 0.05502804, Cur Avg Loss: 0.15240721, Log Avg loss: 0.23997161, Global Avg Loss: 0.65339356, Time: 0.0208 Steps: 95720, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002054, Sample Num: 32864, Cur Loss: 0.03934069, Cur Avg Loss: 0.15227536, Log Avg loss: 0.12532626, Global Avg Loss: 0.65333840, Time: 0.0214 Steps: 95730, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002064, Sample Num: 33024, Cur Loss: 0.25250307, Cur Avg Loss: 0.15233108, Log Avg loss: 0.16377500, Global Avg Loss: 0.65328726, Time: 0.0208 Steps: 95740, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002074, Sample Num: 33184, Cur Loss: 0.07398470, Cur Avg Loss: 0.15229190, Log Avg loss: 0.14420549, Global Avg Loss: 0.65323410, Time: 0.0209 Steps: 95750, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002084, Sample Num: 33344, Cur Loss: 0.16351479, Cur Avg Loss: 0.15208194, Log Avg loss: 0.10853616, Global Avg Loss: 0.65317721, Time: 0.0208 Steps: 95760, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002094, Sample Num: 33504, Cur Loss: 0.10134964, Cur Avg Loss: 0.15205507, Log Avg loss: 0.14645532, Global Avg Loss: 0.65312430, Time: 0.0208 Steps: 95770, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002104, Sample Num: 33664, Cur Loss: 0.18803418, Cur Avg Loss: 0.15204729, Log Avg loss: 0.15041749, Global Avg Loss: 0.65307182, Time: 0.0209 Steps: 95780, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002114, Sample Num: 33824, Cur Loss: 0.12583387, Cur Avg Loss: 0.15185716, Log Avg loss: 0.11185386, Global Avg Loss: 0.65301532, Time: 0.0208 Steps: 95790, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002124, Sample Num: 33984, Cur Loss: 0.27230510, Cur Avg Loss: 0.15178521, Log Avg loss: 0.13657652, Global Avg Loss: 0.65296141, Time: 0.0208 Steps: 95800, Updated lr: 0.000010 ***** Running evaluation checkpoint-95805 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-95805 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.536480, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.234511, "eval_total_loss": 164.861249, "eval_mae": 0.323642, "eval_mse": 0.234599, "eval_r2": 0.850874, "eval_sp_statistic": 0.904111, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.923423, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.62632, "test_total_loss": 314.412412, "test_mae": 0.586303, "test_mse": 0.626393, "test_r2": 0.59572, "test_sp_statistic": 0.798991, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.839316, "test_ps_pvalue": 0.0, "lr": 1.0094831673779043e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6529337253041939, "train_cur_epoch_loss": 323.00426821410656, "train_cur_epoch_avg_loss": 0.15171642471306085, "train_cur_epoch_time": 44.53647994995117, "train_cur_epoch_avg_time": 0.020918966627501725, "epoch": 45, "step": 95805} ################################################## Training, Epoch: 0046, Batch: 000005, Sample Num: 80, Cur Loss: 0.41970032, Cur Avg Loss: 0.20253908, Log Avg loss: 0.16251702, Global Avg Loss: 0.65291022, Time: 0.0245 Steps: 95810, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000015, Sample Num: 240, Cur Loss: 0.33468193, Cur Avg Loss: 0.19488000, Log Avg loss: 0.19105046, Global Avg Loss: 0.65286202, Time: 0.0209 Steps: 95820, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000025, Sample Num: 400, Cur Loss: 0.25969779, Cur Avg Loss: 0.16787007, Log Avg loss: 0.12735518, Global Avg Loss: 0.65280718, Time: 0.0209 Steps: 95830, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000035, Sample Num: 560, Cur Loss: 0.17216089, Cur Avg Loss: 0.15120290, Log Avg loss: 0.10953499, Global Avg Loss: 0.65275050, Time: 0.0210 Steps: 95840, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000045, Sample Num: 720, Cur Loss: 0.14279330, Cur Avg Loss: 0.15296173, Log Avg loss: 0.15911763, Global Avg Loss: 0.65269900, Time: 0.0209 Steps: 95850, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000055, Sample Num: 880, Cur Loss: 0.45592850, Cur Avg Loss: 0.15100068, Log Avg loss: 0.14217595, Global Avg Loss: 0.65264574, Time: 0.0210 Steps: 95860, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000065, Sample Num: 1040, Cur Loss: 0.02868028, Cur Avg Loss: 0.14466034, Log Avg loss: 0.10978849, Global Avg Loss: 0.65258912, Time: 0.0209 Steps: 95870, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000075, Sample Num: 1200, Cur Loss: 0.10019223, Cur Avg Loss: 0.14106954, Log Avg loss: 0.11772931, Global Avg Loss: 0.65253333, Time: 0.0208 Steps: 95880, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000085, Sample Num: 1360, Cur Loss: 0.23518617, Cur Avg Loss: 0.14853000, Log Avg loss: 0.20448349, Global Avg Loss: 0.65248661, Time: 0.0208 Steps: 95890, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000095, Sample Num: 1520, Cur Loss: 0.09217566, Cur Avg Loss: 0.14795323, Log Avg loss: 0.14305061, Global Avg Loss: 0.65243348, Time: 0.0210 Steps: 95900, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000105, Sample Num: 1680, Cur Loss: 0.19281828, Cur Avg Loss: 0.14551643, Log Avg loss: 0.12236691, Global Avg Loss: 0.65237822, Time: 0.0208 Steps: 95910, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000115, Sample Num: 1840, Cur Loss: 0.09750611, Cur Avg Loss: 0.14117573, Log Avg loss: 0.09559839, Global Avg Loss: 0.65232017, Time: 0.0209 Steps: 95920, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000125, Sample Num: 2000, Cur Loss: 0.05608464, Cur Avg Loss: 0.14041726, Log Avg loss: 0.13169478, Global Avg Loss: 0.65226590, Time: 0.0209 Steps: 95930, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000135, Sample Num: 2160, Cur Loss: 0.13224252, Cur Avg Loss: 0.14057529, Log Avg loss: 0.14255068, Global Avg Loss: 0.65221277, Time: 0.0208 Steps: 95940, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000145, Sample Num: 2320, Cur Loss: 0.26437157, Cur Avg Loss: 0.14180916, Log Avg loss: 0.15846643, Global Avg Loss: 0.65216131, Time: 0.0209 Steps: 95950, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000155, Sample Num: 2480, Cur Loss: 0.03730598, Cur Avg Loss: 0.14183137, Log Avg loss: 0.14215337, Global Avg Loss: 0.65210816, Time: 0.0209 Steps: 95960, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000165, Sample Num: 2640, Cur Loss: 0.08758638, Cur Avg Loss: 0.13989773, Log Avg loss: 0.10992627, Global Avg Loss: 0.65205167, Time: 0.0208 Steps: 95970, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000175, Sample Num: 2800, Cur Loss: 0.09474781, Cur Avg Loss: 0.13898595, Log Avg loss: 0.12394167, Global Avg Loss: 0.65199665, Time: 0.0209 Steps: 95980, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000185, Sample Num: 2960, Cur Loss: 0.08784956, Cur Avg Loss: 0.13871362, Log Avg loss: 0.13394784, Global Avg Loss: 0.65194268, Time: 0.0209 Steps: 95990, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000195, Sample Num: 3120, Cur Loss: 0.05311338, Cur Avg Loss: 0.13654028, Log Avg loss: 0.09633348, Global Avg Loss: 0.65188480, Time: 0.0208 Steps: 96000, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000205, Sample Num: 3280, Cur Loss: 0.15399455, Cur Avg Loss: 0.13672300, Log Avg loss: 0.14028601, Global Avg Loss: 0.65183152, Time: 0.0210 Steps: 96010, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000215, Sample Num: 3440, Cur Loss: 0.39173821, Cur Avg Loss: 0.13741440, Log Avg loss: 0.15158808, Global Avg Loss: 0.65177942, Time: 0.0209 Steps: 96020, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000225, Sample Num: 3600, Cur Loss: 0.08084325, Cur Avg Loss: 0.14381361, Log Avg loss: 0.28139660, Global Avg Loss: 0.65174085, Time: 0.0209 Steps: 96030, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000235, Sample Num: 3760, Cur Loss: 0.06847648, Cur Avg Loss: 0.14680627, Log Avg loss: 0.21414126, Global Avg Loss: 0.65169528, Time: 0.0210 Steps: 96040, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000245, Sample Num: 3920, Cur Loss: 0.15247124, Cur Avg Loss: 0.14512759, Log Avg loss: 0.10567852, Global Avg Loss: 0.65163844, Time: 0.0209 Steps: 96050, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000255, Sample Num: 4080, Cur Loss: 0.10619076, Cur Avg Loss: 0.14320107, Log Avg loss: 0.09600123, Global Avg Loss: 0.65158059, Time: 0.0209 Steps: 96060, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000265, Sample Num: 4240, Cur Loss: 0.16072175, Cur Avg Loss: 0.14190424, Log Avg loss: 0.10883527, Global Avg Loss: 0.65152410, Time: 0.0208 Steps: 96070, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000275, Sample Num: 4400, Cur Loss: 0.05438018, Cur Avg Loss: 0.14115690, Log Avg loss: 0.12135223, Global Avg Loss: 0.65146892, Time: 0.0208 Steps: 96080, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000285, Sample Num: 4560, Cur Loss: 0.10156575, Cur Avg Loss: 0.14164025, Log Avg loss: 0.15493232, Global Avg Loss: 0.65141724, Time: 0.0207 Steps: 96090, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000295, Sample Num: 4720, Cur Loss: 0.06911322, Cur Avg Loss: 0.14257376, Log Avg loss: 0.16917906, Global Avg Loss: 0.65136706, Time: 0.0208 Steps: 96100, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000305, Sample Num: 4880, Cur Loss: 0.25209564, Cur Avg Loss: 0.14204271, Log Avg loss: 0.12637653, Global Avg Loss: 0.65131244, Time: 0.0207 Steps: 96110, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000315, Sample Num: 5040, Cur Loss: 0.05286676, Cur Avg Loss: 0.14131640, Log Avg loss: 0.11916394, Global Avg Loss: 0.65125708, Time: 0.0208 Steps: 96120, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000325, Sample Num: 5200, Cur Loss: 0.10440122, Cur Avg Loss: 0.14049434, Log Avg loss: 0.11459946, Global Avg Loss: 0.65120125, Time: 0.0207 Steps: 96130, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000335, Sample Num: 5360, Cur Loss: 0.05180293, Cur Avg Loss: 0.14007229, Log Avg loss: 0.12635569, Global Avg Loss: 0.65114666, Time: 0.0208 Steps: 96140, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000345, Sample Num: 5520, Cur Loss: 0.21497983, Cur Avg Loss: 0.13962842, Log Avg loss: 0.12475897, Global Avg Loss: 0.65109191, Time: 0.0208 Steps: 96150, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000355, Sample Num: 5680, Cur Loss: 0.39441252, Cur Avg Loss: 0.13957487, Log Avg loss: 0.13772717, Global Avg Loss: 0.65103853, Time: 0.0207 Steps: 96160, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000365, Sample Num: 5840, Cur Loss: 0.02942817, Cur Avg Loss: 0.14072804, Log Avg loss: 0.18166580, Global Avg Loss: 0.65098972, Time: 0.0207 Steps: 96170, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000375, Sample Num: 6000, Cur Loss: 0.16941519, Cur Avg Loss: 0.13986556, Log Avg loss: 0.10838476, Global Avg Loss: 0.65093330, Time: 0.0207 Steps: 96180, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000385, Sample Num: 6160, Cur Loss: 0.05102560, Cur Avg Loss: 0.13843654, Log Avg loss: 0.08484843, Global Avg Loss: 0.65087445, Time: 0.0207 Steps: 96190, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000395, Sample Num: 6320, Cur Loss: 0.03135920, Cur Avg Loss: 0.13973795, Log Avg loss: 0.18984225, Global Avg Loss: 0.65082653, Time: 0.0207 Steps: 96200, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000405, Sample Num: 6480, Cur Loss: 0.11679825, Cur Avg Loss: 0.14089412, Log Avg loss: 0.18656297, Global Avg Loss: 0.65077827, Time: 0.0208 Steps: 96210, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000415, Sample Num: 6640, Cur Loss: 0.16530423, Cur Avg Loss: 0.14118967, Log Avg loss: 0.15315949, Global Avg Loss: 0.65072656, Time: 0.0207 Steps: 96220, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000425, Sample Num: 6800, Cur Loss: 0.10448851, Cur Avg Loss: 0.14009429, Log Avg loss: 0.09463563, Global Avg Loss: 0.65066877, Time: 0.0208 Steps: 96230, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000435, Sample Num: 6960, Cur Loss: 0.03675305, Cur Avg Loss: 0.13973392, Log Avg loss: 0.12441835, Global Avg Loss: 0.65061409, Time: 0.0207 Steps: 96240, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000445, Sample Num: 7120, Cur Loss: 0.21218477, Cur Avg Loss: 0.13928458, Log Avg loss: 0.11973850, Global Avg Loss: 0.65055893, Time: 0.0207 Steps: 96250, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000455, Sample Num: 7280, Cur Loss: 0.11376099, Cur Avg Loss: 0.13921034, Log Avg loss: 0.13590639, Global Avg Loss: 0.65050547, Time: 0.0207 Steps: 96260, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000465, Sample Num: 7440, Cur Loss: 0.11756251, Cur Avg Loss: 0.13879606, Log Avg loss: 0.11994619, Global Avg Loss: 0.65045036, Time: 0.0207 Steps: 96270, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000475, Sample Num: 7600, Cur Loss: 0.10508299, Cur Avg Loss: 0.13822552, Log Avg loss: 0.11169568, Global Avg Loss: 0.65039440, Time: 0.0208 Steps: 96280, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000485, Sample Num: 7760, Cur Loss: 0.31569192, Cur Avg Loss: 0.13944939, Log Avg loss: 0.19758305, Global Avg Loss: 0.65034737, Time: 0.0208 Steps: 96290, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000495, Sample Num: 7920, Cur Loss: 0.21373263, Cur Avg Loss: 0.14023161, Log Avg loss: 0.17816949, Global Avg Loss: 0.65029834, Time: 0.0208 Steps: 96300, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000505, Sample Num: 8080, Cur Loss: 0.31601250, Cur Avg Loss: 0.14102829, Log Avg loss: 0.18046410, Global Avg Loss: 0.65024956, Time: 0.0207 Steps: 96310, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000515, Sample Num: 8240, Cur Loss: 0.17181723, Cur Avg Loss: 0.14155642, Log Avg loss: 0.16822658, Global Avg Loss: 0.65019951, Time: 0.0245 Steps: 96320, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000525, Sample Num: 8400, Cur Loss: 0.20392488, Cur Avg Loss: 0.14129079, Log Avg loss: 0.12761102, Global Avg Loss: 0.65014526, Time: 0.0207 Steps: 96330, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000535, Sample Num: 8560, Cur Loss: 0.32981223, Cur Avg Loss: 0.14135167, Log Avg loss: 0.14454810, Global Avg Loss: 0.65009278, Time: 0.0208 Steps: 96340, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000545, Sample Num: 8720, Cur Loss: 0.10864878, Cur Avg Loss: 0.14175498, Log Avg loss: 0.16333162, Global Avg Loss: 0.65004226, Time: 0.0208 Steps: 96350, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000555, Sample Num: 8880, Cur Loss: 0.08669907, Cur Avg Loss: 0.14147066, Log Avg loss: 0.12597538, Global Avg Loss: 0.64998788, Time: 0.0208 Steps: 96360, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000565, Sample Num: 9040, Cur Loss: 0.06271687, Cur Avg Loss: 0.14032710, Log Avg loss: 0.07685972, Global Avg Loss: 0.64992840, Time: 0.0207 Steps: 96370, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000575, Sample Num: 9200, Cur Loss: 0.06375897, Cur Avg Loss: 0.14042300, Log Avg loss: 0.14584141, Global Avg Loss: 0.64987610, Time: 0.0208 Steps: 96380, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000585, Sample Num: 9360, Cur Loss: 0.31416911, Cur Avg Loss: 0.14116751, Log Avg loss: 0.18397688, Global Avg Loss: 0.64982777, Time: 0.0208 Steps: 96390, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000595, Sample Num: 9520, Cur Loss: 0.11489639, Cur Avg Loss: 0.14137277, Log Avg loss: 0.15338029, Global Avg Loss: 0.64977627, Time: 0.0207 Steps: 96400, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000605, Sample Num: 9680, Cur Loss: 0.13312355, Cur Avg Loss: 0.14082817, Log Avg loss: 0.10842428, Global Avg Loss: 0.64972012, Time: 0.0207 Steps: 96410, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000615, Sample Num: 9840, Cur Loss: 0.04506614, Cur Avg Loss: 0.13978957, Log Avg loss: 0.07695414, Global Avg Loss: 0.64966071, Time: 0.0208 Steps: 96420, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000625, Sample Num: 10000, Cur Loss: 0.59870768, Cur Avg Loss: 0.14143426, Log Avg loss: 0.24258309, Global Avg Loss: 0.64961850, Time: 0.0207 Steps: 96430, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000635, Sample Num: 10160, Cur Loss: 0.29702193, Cur Avg Loss: 0.14146589, Log Avg loss: 0.14344269, Global Avg Loss: 0.64956601, Time: 0.0207 Steps: 96440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000645, Sample Num: 10320, Cur Loss: 0.15642285, Cur Avg Loss: 0.14138657, Log Avg loss: 0.13634936, Global Avg Loss: 0.64951280, Time: 0.0207 Steps: 96450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000655, Sample Num: 10480, Cur Loss: 0.03554057, Cur Avg Loss: 0.14103071, Log Avg loss: 0.11807776, Global Avg Loss: 0.64945771, Time: 0.0207 Steps: 96460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000665, Sample Num: 10640, Cur Loss: 0.26269615, Cur Avg Loss: 0.14147898, Log Avg loss: 0.17084075, Global Avg Loss: 0.64940810, Time: 0.0207 Steps: 96470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000675, Sample Num: 10800, Cur Loss: 0.42581868, Cur Avg Loss: 0.14174960, Log Avg loss: 0.15974573, Global Avg Loss: 0.64935734, Time: 0.0207 Steps: 96480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000685, Sample Num: 10960, Cur Loss: 0.14594057, Cur Avg Loss: 0.14164732, Log Avg loss: 0.13474344, Global Avg Loss: 0.64930401, Time: 0.0207 Steps: 96490, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000695, Sample Num: 11120, Cur Loss: 0.06591412, Cur Avg Loss: 0.14102565, Log Avg loss: 0.09844143, Global Avg Loss: 0.64924693, Time: 0.0207 Steps: 96500, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000705, Sample Num: 11280, Cur Loss: 0.35523084, Cur Avg Loss: 0.14160434, Log Avg loss: 0.18182324, Global Avg Loss: 0.64919849, Time: 0.0208 Steps: 96510, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000715, Sample Num: 11440, Cur Loss: 0.10646576, Cur Avg Loss: 0.14190417, Log Avg loss: 0.16304256, Global Avg Loss: 0.64914813, Time: 0.0207 Steps: 96520, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000725, Sample Num: 11600, Cur Loss: 0.05499063, Cur Avg Loss: 0.14117706, Log Avg loss: 0.08918818, Global Avg Loss: 0.64909012, Time: 0.0207 Steps: 96530, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000735, Sample Num: 11760, Cur Loss: 0.06343932, Cur Avg Loss: 0.14139713, Log Avg loss: 0.15735253, Global Avg Loss: 0.64903918, Time: 0.0207 Steps: 96540, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000745, Sample Num: 11920, Cur Loss: 0.06531417, Cur Avg Loss: 0.14199564, Log Avg loss: 0.18598635, Global Avg Loss: 0.64899122, Time: 0.0208 Steps: 96550, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000755, Sample Num: 12080, Cur Loss: 0.12651889, Cur Avg Loss: 0.14291935, Log Avg loss: 0.21173558, Global Avg Loss: 0.64894594, Time: 0.0207 Steps: 96560, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000765, Sample Num: 12240, Cur Loss: 0.09941050, Cur Avg Loss: 0.14287885, Log Avg loss: 0.13982117, Global Avg Loss: 0.64889322, Time: 0.0207 Steps: 96570, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000775, Sample Num: 12400, Cur Loss: 0.21712750, Cur Avg Loss: 0.14343082, Log Avg loss: 0.18565666, Global Avg Loss: 0.64884525, Time: 0.0214 Steps: 96580, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000785, Sample Num: 12560, Cur Loss: 0.23541763, Cur Avg Loss: 0.14369010, Log Avg loss: 0.16378378, Global Avg Loss: 0.64879503, Time: 0.0208 Steps: 96590, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000795, Sample Num: 12720, Cur Loss: 0.04185528, Cur Avg Loss: 0.14315267, Log Avg loss: 0.10096494, Global Avg Loss: 0.64873832, Time: 0.0208 Steps: 96600, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000805, Sample Num: 12880, Cur Loss: 0.07088870, Cur Avg Loss: 0.14312587, Log Avg loss: 0.14099486, Global Avg Loss: 0.64868577, Time: 0.0208 Steps: 96610, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000815, Sample Num: 13040, Cur Loss: 0.06855752, Cur Avg Loss: 0.14315279, Log Avg loss: 0.14532021, Global Avg Loss: 0.64863367, Time: 0.0207 Steps: 96620, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000825, Sample Num: 13200, Cur Loss: 0.04551068, Cur Avg Loss: 0.14320179, Log Avg loss: 0.14719507, Global Avg Loss: 0.64858178, Time: 0.0208 Steps: 96630, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000835, Sample Num: 13360, Cur Loss: 0.03048795, Cur Avg Loss: 0.14347692, Log Avg loss: 0.16617474, Global Avg Loss: 0.64853186, Time: 0.0208 Steps: 96640, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000845, Sample Num: 13520, Cur Loss: 0.15853557, Cur Avg Loss: 0.14360231, Log Avg loss: 0.15407312, Global Avg Loss: 0.64848070, Time: 0.0209 Steps: 96650, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000855, Sample Num: 13680, Cur Loss: 0.03508694, Cur Avg Loss: 0.14322236, Log Avg loss: 0.11111632, Global Avg Loss: 0.64842511, Time: 0.0208 Steps: 96660, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000865, Sample Num: 13840, Cur Loss: 0.08491651, Cur Avg Loss: 0.14363198, Log Avg loss: 0.17865410, Global Avg Loss: 0.64837651, Time: 0.0208 Steps: 96670, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000875, Sample Num: 14000, Cur Loss: 0.18378161, Cur Avg Loss: 0.14378408, Log Avg loss: 0.15694101, Global Avg Loss: 0.64832568, Time: 0.0208 Steps: 96680, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000885, Sample Num: 14160, Cur Loss: 0.32062840, Cur Avg Loss: 0.14432984, Log Avg loss: 0.19208355, Global Avg Loss: 0.64827849, Time: 0.0209 Steps: 96690, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000895, Sample Num: 14320, Cur Loss: 0.04175163, Cur Avg Loss: 0.14402551, Log Avg loss: 0.11709225, Global Avg Loss: 0.64822356, Time: 0.0208 Steps: 96700, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000905, Sample Num: 14480, Cur Loss: 0.25414243, Cur Avg Loss: 0.14410111, Log Avg loss: 0.15086772, Global Avg Loss: 0.64817213, Time: 0.0208 Steps: 96710, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000915, Sample Num: 14640, Cur Loss: 0.10486354, Cur Avg Loss: 0.14453761, Log Avg loss: 0.18404037, Global Avg Loss: 0.64812415, Time: 0.0208 Steps: 96720, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000925, Sample Num: 14800, Cur Loss: 0.30194238, Cur Avg Loss: 0.14519164, Log Avg loss: 0.20503595, Global Avg Loss: 0.64807834, Time: 0.0208 Steps: 96730, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000935, Sample Num: 14960, Cur Loss: 0.35733223, Cur Avg Loss: 0.14517612, Log Avg loss: 0.14374074, Global Avg Loss: 0.64802621, Time: 0.0208 Steps: 96740, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000945, Sample Num: 15120, Cur Loss: 0.09926711, Cur Avg Loss: 0.14500493, Log Avg loss: 0.12899862, Global Avg Loss: 0.64797256, Time: 0.0208 Steps: 96750, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000955, Sample Num: 15280, Cur Loss: 0.18767247, Cur Avg Loss: 0.14489034, Log Avg loss: 0.13406100, Global Avg Loss: 0.64791945, Time: 0.0208 Steps: 96760, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000965, Sample Num: 15440, Cur Loss: 0.09829758, Cur Avg Loss: 0.14514814, Log Avg loss: 0.16976869, Global Avg Loss: 0.64787004, Time: 0.0208 Steps: 96770, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000975, Sample Num: 15600, Cur Loss: 0.05778847, Cur Avg Loss: 0.14481222, Log Avg loss: 0.11239548, Global Avg Loss: 0.64781471, Time: 0.0208 Steps: 96780, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000985, Sample Num: 15760, Cur Loss: 0.11051320, Cur Avg Loss: 0.14445435, Log Avg loss: 0.10956223, Global Avg Loss: 0.64775910, Time: 0.0208 Steps: 96790, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000995, Sample Num: 15920, Cur Loss: 0.12755921, Cur Avg Loss: 0.14434611, Log Avg loss: 0.13368445, Global Avg Loss: 0.64770599, Time: 0.0208 Steps: 96800, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001005, Sample Num: 16080, Cur Loss: 0.06145192, Cur Avg Loss: 0.14422218, Log Avg loss: 0.13189066, Global Avg Loss: 0.64765271, Time: 0.0208 Steps: 96810, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001015, Sample Num: 16240, Cur Loss: 0.15900028, Cur Avg Loss: 0.14444417, Log Avg loss: 0.16675491, Global Avg Loss: 0.64760304, Time: 0.0208 Steps: 96820, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001025, Sample Num: 16400, Cur Loss: 0.13511208, Cur Avg Loss: 0.14444398, Log Avg loss: 0.14442457, Global Avg Loss: 0.64755108, Time: 0.0245 Steps: 96830, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001035, Sample Num: 16560, Cur Loss: 0.08367395, Cur Avg Loss: 0.14439554, Log Avg loss: 0.13943037, Global Avg Loss: 0.64749861, Time: 0.0208 Steps: 96840, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001045, Sample Num: 16720, Cur Loss: 0.26238972, Cur Avg Loss: 0.14496213, Log Avg loss: 0.20360371, Global Avg Loss: 0.64745277, Time: 0.0208 Steps: 96850, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001055, Sample Num: 16880, Cur Loss: 0.42812052, Cur Avg Loss: 0.14552978, Log Avg loss: 0.20484914, Global Avg Loss: 0.64740708, Time: 0.0207 Steps: 96860, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001065, Sample Num: 17040, Cur Loss: 0.04723735, Cur Avg Loss: 0.14508956, Log Avg loss: 0.09864667, Global Avg Loss: 0.64735043, Time: 0.0208 Steps: 96870, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001075, Sample Num: 17200, Cur Loss: 0.08922988, Cur Avg Loss: 0.14505614, Log Avg loss: 0.14149707, Global Avg Loss: 0.64729821, Time: 0.0207 Steps: 96880, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001085, Sample Num: 17360, Cur Loss: 0.04643254, Cur Avg Loss: 0.14482365, Log Avg loss: 0.11983144, Global Avg Loss: 0.64724377, Time: 0.0207 Steps: 96890, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001095, Sample Num: 17520, Cur Loss: 0.05287089, Cur Avg Loss: 0.14528838, Log Avg loss: 0.19571065, Global Avg Loss: 0.64719718, Time: 0.0208 Steps: 96900, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001105, Sample Num: 17680, Cur Loss: 0.13424823, Cur Avg Loss: 0.14509740, Log Avg loss: 0.12418516, Global Avg Loss: 0.64714321, Time: 0.0208 Steps: 96910, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001115, Sample Num: 17840, Cur Loss: 0.04827417, Cur Avg Loss: 0.14462153, Log Avg loss: 0.09203810, Global Avg Loss: 0.64708593, Time: 0.0208 Steps: 96920, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001125, Sample Num: 18000, Cur Loss: 0.12639081, Cur Avg Loss: 0.14470684, Log Avg loss: 0.15421898, Global Avg Loss: 0.64703508, Time: 0.0208 Steps: 96930, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001135, Sample Num: 18160, Cur Loss: 0.05087108, Cur Avg Loss: 0.14472774, Log Avg loss: 0.14707848, Global Avg Loss: 0.64698351, Time: 0.0208 Steps: 96940, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001145, Sample Num: 18320, Cur Loss: 0.30898452, Cur Avg Loss: 0.14503954, Log Avg loss: 0.18042946, Global Avg Loss: 0.64693539, Time: 0.0208 Steps: 96950, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001155, Sample Num: 18480, Cur Loss: 0.10709877, Cur Avg Loss: 0.14464967, Log Avg loss: 0.10000895, Global Avg Loss: 0.64687898, Time: 0.0208 Steps: 96960, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001165, Sample Num: 18640, Cur Loss: 0.15210655, Cur Avg Loss: 0.14506625, Log Avg loss: 0.19318149, Global Avg Loss: 0.64683219, Time: 0.0208 Steps: 96970, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001175, Sample Num: 18800, Cur Loss: 0.15251140, Cur Avg Loss: 0.14488676, Log Avg loss: 0.12397649, Global Avg Loss: 0.64677828, Time: 0.0208 Steps: 96980, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001185, Sample Num: 18960, Cur Loss: 0.11910441, Cur Avg Loss: 0.14458911, Log Avg loss: 0.10961525, Global Avg Loss: 0.64672290, Time: 0.0208 Steps: 96990, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001195, Sample Num: 19120, Cur Loss: 0.10064837, Cur Avg Loss: 0.14484279, Log Avg loss: 0.17490419, Global Avg Loss: 0.64667425, Time: 0.0208 Steps: 97000, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001205, Sample Num: 19280, Cur Loss: 0.19957177, Cur Avg Loss: 0.14457316, Log Avg loss: 0.11235187, Global Avg Loss: 0.64661918, Time: 0.0208 Steps: 97010, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001215, Sample Num: 19440, Cur Loss: 0.03633458, Cur Avg Loss: 0.14466846, Log Avg loss: 0.15615181, Global Avg Loss: 0.64656862, Time: 0.0207 Steps: 97020, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001225, Sample Num: 19600, Cur Loss: 0.38151544, Cur Avg Loss: 0.14543739, Log Avg loss: 0.23886341, Global Avg Loss: 0.64652660, Time: 0.0209 Steps: 97030, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001235, Sample Num: 19760, Cur Loss: 0.18497440, Cur Avg Loss: 0.14526923, Log Avg loss: 0.12466918, Global Avg Loss: 0.64647283, Time: 0.0208 Steps: 97040, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001245, Sample Num: 19920, Cur Loss: 0.22648296, Cur Avg Loss: 0.14522826, Log Avg loss: 0.14016776, Global Avg Loss: 0.64642066, Time: 0.0208 Steps: 97050, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001255, Sample Num: 20080, Cur Loss: 0.12596993, Cur Avg Loss: 0.14588358, Log Avg loss: 0.22747100, Global Avg Loss: 0.64637749, Time: 0.0208 Steps: 97060, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001265, Sample Num: 20240, Cur Loss: 0.09086771, Cur Avg Loss: 0.14612658, Log Avg loss: 0.17662357, Global Avg Loss: 0.64632910, Time: 0.0207 Steps: 97070, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001275, Sample Num: 20400, Cur Loss: 0.04415520, Cur Avg Loss: 0.14610652, Log Avg loss: 0.14356903, Global Avg Loss: 0.64627731, Time: 0.0208 Steps: 97080, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001285, Sample Num: 20560, Cur Loss: 0.07304370, Cur Avg Loss: 0.14642153, Log Avg loss: 0.18658538, Global Avg Loss: 0.64622996, Time: 0.0209 Steps: 97090, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001295, Sample Num: 20720, Cur Loss: 0.10008720, Cur Avg Loss: 0.14619117, Log Avg loss: 0.11658964, Global Avg Loss: 0.64617542, Time: 0.0208 Steps: 97100, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001305, Sample Num: 20880, Cur Loss: 0.02485086, Cur Avg Loss: 0.14626061, Log Avg loss: 0.15525357, Global Avg Loss: 0.64612487, Time: 0.0208 Steps: 97110, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001315, Sample Num: 21040, Cur Loss: 0.03428527, Cur Avg Loss: 0.14606489, Log Avg loss: 0.12052366, Global Avg Loss: 0.64607075, Time: 0.0208 Steps: 97120, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001325, Sample Num: 21200, Cur Loss: 0.28696752, Cur Avg Loss: 0.14601417, Log Avg loss: 0.13934403, Global Avg Loss: 0.64601858, Time: 0.0208 Steps: 97130, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001335, Sample Num: 21360, Cur Loss: 0.33110172, Cur Avg Loss: 0.14582192, Log Avg loss: 0.12034908, Global Avg Loss: 0.64596446, Time: 0.0208 Steps: 97140, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001345, Sample Num: 21520, Cur Loss: 0.09478490, Cur Avg Loss: 0.14611740, Log Avg loss: 0.18556290, Global Avg Loss: 0.64591707, Time: 0.0208 Steps: 97150, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001355, Sample Num: 21680, Cur Loss: 0.13405105, Cur Avg Loss: 0.14608624, Log Avg loss: 0.14189623, Global Avg Loss: 0.64586520, Time: 0.0208 Steps: 97160, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001365, Sample Num: 21840, Cur Loss: 0.05865225, Cur Avg Loss: 0.14627875, Log Avg loss: 0.17236401, Global Avg Loss: 0.64581647, Time: 0.0208 Steps: 97170, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001375, Sample Num: 22000, Cur Loss: 0.08599067, Cur Avg Loss: 0.14589144, Log Avg loss: 0.09302264, Global Avg Loss: 0.64575958, Time: 0.0208 Steps: 97180, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001385, Sample Num: 22160, Cur Loss: 0.06360036, Cur Avg Loss: 0.14593808, Log Avg loss: 0.15235099, Global Avg Loss: 0.64570882, Time: 0.0208 Steps: 97190, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001395, Sample Num: 22320, Cur Loss: 0.13163255, Cur Avg Loss: 0.14603123, Log Avg loss: 0.15893317, Global Avg Loss: 0.64565874, Time: 0.0208 Steps: 97200, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001405, Sample Num: 22480, Cur Loss: 0.43220419, Cur Avg Loss: 0.14628339, Log Avg loss: 0.18145985, Global Avg Loss: 0.64561098, Time: 0.0208 Steps: 97210, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001415, Sample Num: 22640, Cur Loss: 0.23392253, Cur Avg Loss: 0.14617659, Log Avg loss: 0.13117158, Global Avg Loss: 0.64555807, Time: 0.0208 Steps: 97220, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001425, Sample Num: 22800, Cur Loss: 0.09606852, Cur Avg Loss: 0.14592799, Log Avg loss: 0.11074986, Global Avg Loss: 0.64550306, Time: 0.0207 Steps: 97230, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001435, Sample Num: 22960, Cur Loss: 0.28913057, Cur Avg Loss: 0.14605768, Log Avg loss: 0.16453976, Global Avg Loss: 0.64545360, Time: 0.0208 Steps: 97240, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001445, Sample Num: 23120, Cur Loss: 0.05770814, Cur Avg Loss: 0.14599082, Log Avg loss: 0.13639535, Global Avg Loss: 0.64540126, Time: 0.0208 Steps: 97250, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001455, Sample Num: 23280, Cur Loss: 0.53742158, Cur Avg Loss: 0.14629341, Log Avg loss: 0.19001846, Global Avg Loss: 0.64535444, Time: 0.0208 Steps: 97260, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001465, Sample Num: 23440, Cur Loss: 0.16737461, Cur Avg Loss: 0.14665139, Log Avg loss: 0.19873730, Global Avg Loss: 0.64530852, Time: 0.0208 Steps: 97270, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001475, Sample Num: 23600, Cur Loss: 0.04652852, Cur Avg Loss: 0.14645802, Log Avg loss: 0.11812894, Global Avg Loss: 0.64525433, Time: 0.0208 Steps: 97280, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001485, Sample Num: 23760, Cur Loss: 0.18774754, Cur Avg Loss: 0.14618467, Log Avg loss: 0.10586571, Global Avg Loss: 0.64519889, Time: 0.0208 Steps: 97290, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001495, Sample Num: 23920, Cur Loss: 0.12616506, Cur Avg Loss: 0.14605278, Log Avg loss: 0.12646728, Global Avg Loss: 0.64514558, Time: 0.0207 Steps: 97300, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001505, Sample Num: 24080, Cur Loss: 0.09345357, Cur Avg Loss: 0.14611804, Log Avg loss: 0.15587487, Global Avg Loss: 0.64509530, Time: 0.0208 Steps: 97310, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001515, Sample Num: 24240, Cur Loss: 0.12485582, Cur Avg Loss: 0.14624674, Log Avg loss: 0.16561585, Global Avg Loss: 0.64504603, Time: 0.0207 Steps: 97320, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001525, Sample Num: 24400, Cur Loss: 0.14247671, Cur Avg Loss: 0.14663449, Log Avg loss: 0.20537823, Global Avg Loss: 0.64500085, Time: 0.0207 Steps: 97330, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001535, Sample Num: 24560, Cur Loss: 0.03703813, Cur Avg Loss: 0.14689713, Log Avg loss: 0.18695015, Global Avg Loss: 0.64495380, Time: 0.0207 Steps: 97340, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001545, Sample Num: 24720, Cur Loss: 0.18555054, Cur Avg Loss: 0.14698102, Log Avg loss: 0.15985808, Global Avg Loss: 0.64490397, Time: 0.0208 Steps: 97350, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001555, Sample Num: 24880, Cur Loss: 0.22310349, Cur Avg Loss: 0.14683546, Log Avg loss: 0.12434683, Global Avg Loss: 0.64485050, Time: 0.0208 Steps: 97360, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001565, Sample Num: 25040, Cur Loss: 0.08138978, Cur Avg Loss: 0.14688927, Log Avg loss: 0.15525589, Global Avg Loss: 0.64480022, Time: 0.0207 Steps: 97370, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001575, Sample Num: 25200, Cur Loss: 0.28698647, Cur Avg Loss: 0.14673083, Log Avg loss: 0.12193489, Global Avg Loss: 0.64474653, Time: 0.0207 Steps: 97380, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001585, Sample Num: 25360, Cur Loss: 0.17449723, Cur Avg Loss: 0.14680109, Log Avg loss: 0.15786685, Global Avg Loss: 0.64469653, Time: 0.0208 Steps: 97390, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001595, Sample Num: 25520, Cur Loss: 0.11898318, Cur Avg Loss: 0.14659745, Log Avg loss: 0.11432049, Global Avg Loss: 0.64464208, Time: 0.0208 Steps: 97400, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001605, Sample Num: 25680, Cur Loss: 0.20449433, Cur Avg Loss: 0.14679897, Log Avg loss: 0.17894157, Global Avg Loss: 0.64459427, Time: 0.0208 Steps: 97410, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001615, Sample Num: 25840, Cur Loss: 0.20482473, Cur Avg Loss: 0.14698700, Log Avg loss: 0.17716667, Global Avg Loss: 0.64454629, Time: 0.0208 Steps: 97420, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001625, Sample Num: 26000, Cur Loss: 0.26760229, Cur Avg Loss: 0.14723364, Log Avg loss: 0.18706553, Global Avg Loss: 0.64449934, Time: 0.0208 Steps: 97430, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001635, Sample Num: 26160, Cur Loss: 0.17853794, Cur Avg Loss: 0.14747783, Log Avg loss: 0.18715803, Global Avg Loss: 0.64445240, Time: 0.0208 Steps: 97440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001645, Sample Num: 26320, Cur Loss: 0.22100471, Cur Avg Loss: 0.14741206, Log Avg loss: 0.13665993, Global Avg Loss: 0.64440029, Time: 0.0208 Steps: 97450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001655, Sample Num: 26480, Cur Loss: 0.15655518, Cur Avg Loss: 0.14719867, Log Avg loss: 0.11209549, Global Avg Loss: 0.64434567, Time: 0.0208 Steps: 97460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001665, Sample Num: 26640, Cur Loss: 0.06525446, Cur Avg Loss: 0.14718534, Log Avg loss: 0.14497893, Global Avg Loss: 0.64429444, Time: 0.0208 Steps: 97470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001675, Sample Num: 26800, Cur Loss: 0.53018284, Cur Avg Loss: 0.14744762, Log Avg loss: 0.19111735, Global Avg Loss: 0.64424795, Time: 0.0208 Steps: 97480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001685, Sample Num: 26960, Cur Loss: 0.06266732, Cur Avg Loss: 0.14758174, Log Avg loss: 0.17004667, Global Avg Loss: 0.64419931, Time: 0.0208 Steps: 97490, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001695, Sample Num: 27120, Cur Loss: 0.21109600, Cur Avg Loss: 0.14742727, Log Avg loss: 0.12139946, Global Avg Loss: 0.64414569, Time: 0.0208 Steps: 97500, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001705, Sample Num: 27280, Cur Loss: 0.12268351, Cur Avg Loss: 0.14728387, Log Avg loss: 0.12297779, Global Avg Loss: 0.64409224, Time: 0.0208 Steps: 97510, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001715, Sample Num: 27440, Cur Loss: 0.08026053, Cur Avg Loss: 0.14724532, Log Avg loss: 0.14067271, Global Avg Loss: 0.64404062, Time: 0.0208 Steps: 97520, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001725, Sample Num: 27600, Cur Loss: 0.36621034, Cur Avg Loss: 0.14734950, Log Avg loss: 0.16521519, Global Avg Loss: 0.64399153, Time: 0.0208 Steps: 97530, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001735, Sample Num: 27760, Cur Loss: 0.29999214, Cur Avg Loss: 0.14776998, Log Avg loss: 0.22030302, Global Avg Loss: 0.64394809, Time: 0.0208 Steps: 97540, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001745, Sample Num: 27920, Cur Loss: 0.06393918, Cur Avg Loss: 0.14753433, Log Avg loss: 0.10665010, Global Avg Loss: 0.64389301, Time: 0.0208 Steps: 97550, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001755, Sample Num: 28080, Cur Loss: 0.24059141, Cur Avg Loss: 0.14736537, Log Avg loss: 0.11788170, Global Avg Loss: 0.64383909, Time: 0.0207 Steps: 97560, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001765, Sample Num: 28240, Cur Loss: 0.11620053, Cur Avg Loss: 0.14718802, Log Avg loss: 0.11606186, Global Avg Loss: 0.64378500, Time: 0.0208 Steps: 97570, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001775, Sample Num: 28400, Cur Loss: 0.19534560, Cur Avg Loss: 0.14739288, Log Avg loss: 0.18355193, Global Avg Loss: 0.64373783, Time: 0.0209 Steps: 97580, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001785, Sample Num: 28560, Cur Loss: 0.08981600, Cur Avg Loss: 0.14727903, Log Avg loss: 0.12706941, Global Avg Loss: 0.64368489, Time: 0.0208 Steps: 97590, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001795, Sample Num: 28720, Cur Loss: 0.12985343, Cur Avg Loss: 0.14722406, Log Avg loss: 0.13741332, Global Avg Loss: 0.64363302, Time: 0.0246 Steps: 97600, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001805, Sample Num: 28880, Cur Loss: 0.06587997, Cur Avg Loss: 0.14748362, Log Avg loss: 0.19407413, Global Avg Loss: 0.64358696, Time: 0.0208 Steps: 97610, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001815, Sample Num: 29040, Cur Loss: 0.31154922, Cur Avg Loss: 0.14758922, Log Avg loss: 0.16665042, Global Avg Loss: 0.64353811, Time: 0.0209 Steps: 97620, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001825, Sample Num: 29200, Cur Loss: 0.10898544, Cur Avg Loss: 0.14765638, Log Avg loss: 0.15984465, Global Avg Loss: 0.64348856, Time: 0.0208 Steps: 97630, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001835, Sample Num: 29360, Cur Loss: 0.08724245, Cur Avg Loss: 0.14777924, Log Avg loss: 0.17020094, Global Avg Loss: 0.64344009, Time: 0.0209 Steps: 97640, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001845, Sample Num: 29520, Cur Loss: 0.11694809, Cur Avg Loss: 0.14782756, Log Avg loss: 0.15669494, Global Avg Loss: 0.64339024, Time: 0.0209 Steps: 97650, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001855, Sample Num: 29680, Cur Loss: 0.08687131, Cur Avg Loss: 0.14801402, Log Avg loss: 0.18241660, Global Avg Loss: 0.64334304, Time: 0.0209 Steps: 97660, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001865, Sample Num: 29840, Cur Loss: 0.09974291, Cur Avg Loss: 0.14807598, Log Avg loss: 0.15956995, Global Avg Loss: 0.64329351, Time: 0.0209 Steps: 97670, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001875, Sample Num: 30000, Cur Loss: 0.13455997, Cur Avg Loss: 0.14808306, Log Avg loss: 0.14940313, Global Avg Loss: 0.64324295, Time: 0.0209 Steps: 97680, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001885, Sample Num: 30160, Cur Loss: 0.08712081, Cur Avg Loss: 0.14801250, Log Avg loss: 0.13478184, Global Avg Loss: 0.64319090, Time: 0.0209 Steps: 97690, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001895, Sample Num: 30320, Cur Loss: 0.34041283, Cur Avg Loss: 0.14823785, Log Avg loss: 0.19071548, Global Avg Loss: 0.64314459, Time: 0.0209 Steps: 97700, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001905, Sample Num: 30480, Cur Loss: 0.18448013, Cur Avg Loss: 0.14826185, Log Avg loss: 0.15281069, Global Avg Loss: 0.64309441, Time: 0.0209 Steps: 97710, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001915, Sample Num: 30640, Cur Loss: 0.29707512, Cur Avg Loss: 0.14838621, Log Avg loss: 0.17207666, Global Avg Loss: 0.64304620, Time: 0.0208 Steps: 97720, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001925, Sample Num: 30800, Cur Loss: 0.08376666, Cur Avg Loss: 0.14843523, Log Avg loss: 0.15782201, Global Avg Loss: 0.64299656, Time: 0.0209 Steps: 97730, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001935, Sample Num: 30960, Cur Loss: 0.07892466, Cur Avg Loss: 0.14859157, Log Avg loss: 0.17868754, Global Avg Loss: 0.64294905, Time: 0.0208 Steps: 97740, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001945, Sample Num: 31120, Cur Loss: 0.07218111, Cur Avg Loss: 0.14852889, Log Avg loss: 0.13640050, Global Avg Loss: 0.64289723, Time: 0.0208 Steps: 97750, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001955, Sample Num: 31280, Cur Loss: 0.11329105, Cur Avg Loss: 0.14871847, Log Avg loss: 0.18559110, Global Avg Loss: 0.64285045, Time: 0.0208 Steps: 97760, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001965, Sample Num: 31440, Cur Loss: 0.13340198, Cur Avg Loss: 0.14852205, Log Avg loss: 0.11012218, Global Avg Loss: 0.64279596, Time: 0.0208 Steps: 97770, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001975, Sample Num: 31600, Cur Loss: 0.19868134, Cur Avg Loss: 0.14851696, Log Avg loss: 0.14751676, Global Avg Loss: 0.64274531, Time: 0.0208 Steps: 97780, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001985, Sample Num: 31760, Cur Loss: 0.05183507, Cur Avg Loss: 0.14843141, Log Avg loss: 0.13153536, Global Avg Loss: 0.64269304, Time: 0.0208 Steps: 97790, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001995, Sample Num: 31920, Cur Loss: 0.28190529, Cur Avg Loss: 0.14861672, Log Avg loss: 0.18540098, Global Avg Loss: 0.64264628, Time: 0.0208 Steps: 97800, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002005, Sample Num: 32080, Cur Loss: 0.08328455, Cur Avg Loss: 0.14884138, Log Avg loss: 0.19366218, Global Avg Loss: 0.64260037, Time: 0.0208 Steps: 97810, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002015, Sample Num: 32240, Cur Loss: 0.29803464, Cur Avg Loss: 0.14922538, Log Avg loss: 0.22621584, Global Avg Loss: 0.64255781, Time: 0.0208 Steps: 97820, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002025, Sample Num: 32400, Cur Loss: 0.07288536, Cur Avg Loss: 0.14939280, Log Avg loss: 0.18312934, Global Avg Loss: 0.64251085, Time: 0.0209 Steps: 97830, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002035, Sample Num: 32560, Cur Loss: 0.19255319, Cur Avg Loss: 0.14947815, Log Avg loss: 0.16675984, Global Avg Loss: 0.64246222, Time: 0.0209 Steps: 97840, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002045, Sample Num: 32720, Cur Loss: 0.12477069, Cur Avg Loss: 0.14931519, Log Avg loss: 0.11615399, Global Avg Loss: 0.64240843, Time: 0.0208 Steps: 97850, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002055, Sample Num: 32880, Cur Loss: 0.02830186, Cur Avg Loss: 0.14904577, Log Avg loss: 0.09394915, Global Avg Loss: 0.64235239, Time: 0.0209 Steps: 97860, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002065, Sample Num: 33040, Cur Loss: 0.07266605, Cur Avg Loss: 0.14881158, Log Avg loss: 0.10068644, Global Avg Loss: 0.64229704, Time: 0.0208 Steps: 97870, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002075, Sample Num: 33200, Cur Loss: 0.05512320, Cur Avg Loss: 0.14867944, Log Avg loss: 0.12139254, Global Avg Loss: 0.64224382, Time: 0.0208 Steps: 97880, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002085, Sample Num: 33360, Cur Loss: 0.10661632, Cur Avg Loss: 0.14849701, Log Avg loss: 0.11064243, Global Avg Loss: 0.64218952, Time: 0.0208 Steps: 97890, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002095, Sample Num: 33520, Cur Loss: 0.57615089, Cur Avg Loss: 0.14875122, Log Avg loss: 0.20175427, Global Avg Loss: 0.64214453, Time: 0.0208 Steps: 97900, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002105, Sample Num: 33680, Cur Loss: 0.03104166, Cur Avg Loss: 0.14867839, Log Avg loss: 0.13341879, Global Avg Loss: 0.64209257, Time: 0.0207 Steps: 97910, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002115, Sample Num: 33840, Cur Loss: 0.24989727, Cur Avg Loss: 0.14875425, Log Avg loss: 0.16472380, Global Avg Loss: 0.64204382, Time: 0.0208 Steps: 97920, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002125, Sample Num: 34000, Cur Loss: 0.22511746, Cur Avg Loss: 0.14863200, Log Avg loss: 0.12277535, Global Avg Loss: 0.64199079, Time: 0.0208 Steps: 97930, Updated lr: 0.000008 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.458934, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.232943, "eval_total_loss": 163.758647, "eval_mae": 0.321684, "eval_mse": 0.233032, "eval_r2": 0.85187, "eval_sp_statistic": 0.905255, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.924942, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.704094, "test_total_loss": 353.454937, "test_mae": 0.648612, "test_mse": 0.704152, "test_r2": 0.545534, "test_sp_statistic": 0.798174, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.835459, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6419683867716718, "train_cur_epoch_loss": 316.21643732860684, "train_cur_epoch_avg_loss": 0.14852815280817608, "train_cur_epoch_time": 44.45893406867981, "train_cur_epoch_avg_time": 0.020882543010183094, "epoch": 46, "step": 97934} ################################################## Training, Epoch: 0047, Batch: 000006, Sample Num: 96, Cur Loss: 0.26212451, Cur Avg Loss: 0.14186069, Log Avg loss: 0.12246077, Global Avg Loss: 0.64193775, Time: 0.0210 Steps: 97940, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000016, Sample Num: 256, Cur Loss: 0.06706470, Cur Avg Loss: 0.12390210, Log Avg loss: 0.11312694, Global Avg Loss: 0.64188376, Time: 0.0208 Steps: 97950, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000026, Sample Num: 416, Cur Loss: 0.14518821, Cur Avg Loss: 0.11725195, Log Avg loss: 0.10661171, Global Avg Loss: 0.64182912, Time: 0.0208 Steps: 97960, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000036, Sample Num: 576, Cur Loss: 0.07726899, Cur Avg Loss: 0.11717767, Log Avg loss: 0.11698454, Global Avg Loss: 0.64177555, Time: 0.0209 Steps: 97970, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000046, Sample Num: 736, Cur Loss: 0.15880510, Cur Avg Loss: 0.13278554, Log Avg loss: 0.18897385, Global Avg Loss: 0.64172933, Time: 0.0208 Steps: 97980, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000056, Sample Num: 896, Cur Loss: 0.11391964, Cur Avg Loss: 0.13135576, Log Avg loss: 0.12477879, Global Avg Loss: 0.64167658, Time: 0.0208 Steps: 97990, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000066, Sample Num: 1056, Cur Loss: 0.32304749, Cur Avg Loss: 0.14031823, Log Avg loss: 0.19050809, Global Avg Loss: 0.64163054, Time: 0.0208 Steps: 98000, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000076, Sample Num: 1216, Cur Loss: 0.11159931, Cur Avg Loss: 0.13770248, Log Avg loss: 0.12043853, Global Avg Loss: 0.64157736, Time: 0.0209 Steps: 98010, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000086, Sample Num: 1376, Cur Loss: 0.20094937, Cur Avg Loss: 0.13515872, Log Avg loss: 0.11582611, Global Avg Loss: 0.64152373, Time: 0.0209 Steps: 98020, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000096, Sample Num: 1536, Cur Loss: 0.07719134, Cur Avg Loss: 0.13138200, Log Avg loss: 0.09890220, Global Avg Loss: 0.64146837, Time: 0.0208 Steps: 98030, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000106, Sample Num: 1696, Cur Loss: 0.15692395, Cur Avg Loss: 0.13390821, Log Avg loss: 0.15815984, Global Avg Loss: 0.64141908, Time: 0.0218 Steps: 98040, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000116, Sample Num: 1856, Cur Loss: 0.32374632, Cur Avg Loss: 0.13773856, Log Avg loss: 0.17834033, Global Avg Loss: 0.64137185, Time: 0.0219 Steps: 98050, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000126, Sample Num: 2016, Cur Loss: 0.07137631, Cur Avg Loss: 0.13409519, Log Avg loss: 0.09183198, Global Avg Loss: 0.64131581, Time: 0.0209 Steps: 98060, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000136, Sample Num: 2176, Cur Loss: 0.15833138, Cur Avg Loss: 0.13433286, Log Avg loss: 0.13732754, Global Avg Loss: 0.64126442, Time: 0.0209 Steps: 98070, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000146, Sample Num: 2336, Cur Loss: 0.08277002, Cur Avg Loss: 0.13279679, Log Avg loss: 0.11190621, Global Avg Loss: 0.64121044, Time: 0.0219 Steps: 98080, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000156, Sample Num: 2496, Cur Loss: 0.27006277, Cur Avg Loss: 0.13372056, Log Avg loss: 0.14720771, Global Avg Loss: 0.64116008, Time: 0.0208 Steps: 98090, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000166, Sample Num: 2656, Cur Loss: 0.08667824, Cur Avg Loss: 0.13710569, Log Avg loss: 0.18991370, Global Avg Loss: 0.64111408, Time: 0.0210 Steps: 98100, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000176, Sample Num: 2816, Cur Loss: 0.15305799, Cur Avg Loss: 0.14150450, Log Avg loss: 0.21452470, Global Avg Loss: 0.64107060, Time: 0.0209 Steps: 98110, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000186, Sample Num: 2976, Cur Loss: 0.03930599, Cur Avg Loss: 0.13952656, Log Avg loss: 0.10471489, Global Avg Loss: 0.64101594, Time: 0.0219 Steps: 98120, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000196, Sample Num: 3136, Cur Loss: 0.05466751, Cur Avg Loss: 0.13724668, Log Avg loss: 0.09484094, Global Avg Loss: 0.64096028, Time: 0.0209 Steps: 98130, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000206, Sample Num: 3296, Cur Loss: 0.34919864, Cur Avg Loss: 0.13910499, Log Avg loss: 0.17552770, Global Avg Loss: 0.64091286, Time: 0.0219 Steps: 98140, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000216, Sample Num: 3456, Cur Loss: 0.06309666, Cur Avg Loss: 0.13873232, Log Avg loss: 0.13105549, Global Avg Loss: 0.64086091, Time: 0.0209 Steps: 98150, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000226, Sample Num: 3616, Cur Loss: 0.18152490, Cur Avg Loss: 0.13814429, Log Avg loss: 0.12544271, Global Avg Loss: 0.64080840, Time: 0.0219 Steps: 98160, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000236, Sample Num: 3776, Cur Loss: 0.18487723, Cur Avg Loss: 0.13983646, Log Avg loss: 0.17807957, Global Avg Loss: 0.64076127, Time: 0.0218 Steps: 98170, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000246, Sample Num: 3936, Cur Loss: 0.09211457, Cur Avg Loss: 0.14355513, Log Avg loss: 0.23131560, Global Avg Loss: 0.64071956, Time: 0.0210 Steps: 98180, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000256, Sample Num: 4096, Cur Loss: 0.10878509, Cur Avg Loss: 0.14400240, Log Avg loss: 0.15500543, Global Avg Loss: 0.64067009, Time: 0.0261 Steps: 98190, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000266, Sample Num: 4256, Cur Loss: 0.11471324, Cur Avg Loss: 0.14196698, Log Avg loss: 0.08986018, Global Avg Loss: 0.64061400, Time: 0.0209 Steps: 98200, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000276, Sample Num: 4416, Cur Loss: 0.18930867, Cur Avg Loss: 0.14177594, Log Avg loss: 0.13669420, Global Avg Loss: 0.64056269, Time: 0.0209 Steps: 98210, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000286, Sample Num: 4576, Cur Loss: 0.11087157, Cur Avg Loss: 0.14209161, Log Avg loss: 0.15080418, Global Avg Loss: 0.64051283, Time: 0.0208 Steps: 98220, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000296, Sample Num: 4736, Cur Loss: 0.05888595, Cur Avg Loss: 0.14166319, Log Avg loss: 0.12941035, Global Avg Loss: 0.64046080, Time: 0.0208 Steps: 98230, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000306, Sample Num: 4896, Cur Loss: 0.10093351, Cur Avg Loss: 0.14183083, Log Avg loss: 0.14679283, Global Avg Loss: 0.64041055, Time: 0.0208 Steps: 98240, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000316, Sample Num: 5056, Cur Loss: 0.15445302, Cur Avg Loss: 0.14177011, Log Avg loss: 0.13991206, Global Avg Loss: 0.64035961, Time: 0.0208 Steps: 98250, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000326, Sample Num: 5216, Cur Loss: 0.05444727, Cur Avg Loss: 0.14252903, Log Avg loss: 0.16651115, Global Avg Loss: 0.64031138, Time: 0.0208 Steps: 98260, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000336, Sample Num: 5376, Cur Loss: 0.10901363, Cur Avg Loss: 0.14301214, Log Avg loss: 0.15876148, Global Avg Loss: 0.64026238, Time: 0.0209 Steps: 98270, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000346, Sample Num: 5536, Cur Loss: 0.17000943, Cur Avg Loss: 0.14295515, Log Avg loss: 0.14104011, Global Avg Loss: 0.64021158, Time: 0.0207 Steps: 98280, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000356, Sample Num: 5696, Cur Loss: 0.04747146, Cur Avg Loss: 0.14203180, Log Avg loss: 0.11008399, Global Avg Loss: 0.64015765, Time: 0.0208 Steps: 98290, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000366, Sample Num: 5856, Cur Loss: 0.08825189, Cur Avg Loss: 0.14319558, Log Avg loss: 0.18462615, Global Avg Loss: 0.64011131, Time: 0.0209 Steps: 98300, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000376, Sample Num: 6016, Cur Loss: 0.02458435, Cur Avg Loss: 0.14362959, Log Avg loss: 0.15951430, Global Avg Loss: 0.64006242, Time: 0.0208 Steps: 98310, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000386, Sample Num: 6176, Cur Loss: 0.04016656, Cur Avg Loss: 0.14255683, Log Avg loss: 0.10222116, Global Avg Loss: 0.64000772, Time: 0.0209 Steps: 98320, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000396, Sample Num: 6336, Cur Loss: 0.07189481, Cur Avg Loss: 0.14212653, Log Avg loss: 0.12551704, Global Avg Loss: 0.63995540, Time: 0.0209 Steps: 98330, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000406, Sample Num: 6496, Cur Loss: 0.06096048, Cur Avg Loss: 0.14311442, Log Avg loss: 0.18223457, Global Avg Loss: 0.63990885, Time: 0.0209 Steps: 98340, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000416, Sample Num: 6656, Cur Loss: 0.13996865, Cur Avg Loss: 0.14293623, Log Avg loss: 0.13570199, Global Avg Loss: 0.63985758, Time: 0.0208 Steps: 98350, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000426, Sample Num: 6816, Cur Loss: 0.14335464, Cur Avg Loss: 0.14266968, Log Avg loss: 0.13158110, Global Avg Loss: 0.63980591, Time: 0.0208 Steps: 98360, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000436, Sample Num: 6976, Cur Loss: 0.30606508, Cur Avg Loss: 0.14297673, Log Avg loss: 0.15605697, Global Avg Loss: 0.63975673, Time: 0.0208 Steps: 98370, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000446, Sample Num: 7136, Cur Loss: 0.12217586, Cur Avg Loss: 0.14227493, Log Avg loss: 0.11167643, Global Avg Loss: 0.63970306, Time: 0.0208 Steps: 98380, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000456, Sample Num: 7296, Cur Loss: 0.01870314, Cur Avg Loss: 0.14152507, Log Avg loss: 0.10808134, Global Avg Loss: 0.63964902, Time: 0.0208 Steps: 98390, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000466, Sample Num: 7456, Cur Loss: 0.11498520, Cur Avg Loss: 0.14210035, Log Avg loss: 0.16833320, Global Avg Loss: 0.63960113, Time: 0.0208 Steps: 98400, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000476, Sample Num: 7616, Cur Loss: 0.26243827, Cur Avg Loss: 0.14153229, Log Avg loss: 0.11506065, Global Avg Loss: 0.63954782, Time: 0.0208 Steps: 98410, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000486, Sample Num: 7776, Cur Loss: 0.12794238, Cur Avg Loss: 0.14082789, Log Avg loss: 0.10729832, Global Avg Loss: 0.63949374, Time: 0.0209 Steps: 98420, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000496, Sample Num: 7936, Cur Loss: 0.09173688, Cur Avg Loss: 0.14011848, Log Avg loss: 0.10564147, Global Avg Loss: 0.63943951, Time: 0.0208 Steps: 98430, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000506, Sample Num: 8096, Cur Loss: 0.07323070, Cur Avg Loss: 0.13978975, Log Avg loss: 0.12348444, Global Avg Loss: 0.63938709, Time: 0.0208 Steps: 98440, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000516, Sample Num: 8256, Cur Loss: 0.12892665, Cur Avg Loss: 0.14050163, Log Avg loss: 0.17652267, Global Avg Loss: 0.63934008, Time: 0.0245 Steps: 98450, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000526, Sample Num: 8416, Cur Loss: 0.04269348, Cur Avg Loss: 0.14117536, Log Avg loss: 0.17593993, Global Avg Loss: 0.63929301, Time: 0.0207 Steps: 98460, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000536, Sample Num: 8576, Cur Loss: 0.16004643, Cur Avg Loss: 0.14067756, Log Avg loss: 0.11449364, Global Avg Loss: 0.63923972, Time: 0.0208 Steps: 98470, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000546, Sample Num: 8736, Cur Loss: 0.05609511, Cur Avg Loss: 0.14010379, Log Avg loss: 0.10934938, Global Avg Loss: 0.63918591, Time: 0.0208 Steps: 98480, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000556, Sample Num: 8896, Cur Loss: 0.07562392, Cur Avg Loss: 0.14002365, Log Avg loss: 0.13564786, Global Avg Loss: 0.63913479, Time: 0.0207 Steps: 98490, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000566, Sample Num: 9056, Cur Loss: 0.18172702, Cur Avg Loss: 0.14150689, Log Avg loss: 0.22397520, Global Avg Loss: 0.63909264, Time: 0.0207 Steps: 98500, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000576, Sample Num: 9216, Cur Loss: 0.14221606, Cur Avg Loss: 0.14093831, Log Avg loss: 0.10875685, Global Avg Loss: 0.63903880, Time: 0.0207 Steps: 98510, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000586, Sample Num: 9376, Cur Loss: 0.09930240, Cur Avg Loss: 0.14091705, Log Avg loss: 0.13969225, Global Avg Loss: 0.63898812, Time: 0.0207 Steps: 98520, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000596, Sample Num: 9536, Cur Loss: 0.04525010, Cur Avg Loss: 0.14019028, Log Avg loss: 0.09760177, Global Avg Loss: 0.63893317, Time: 0.0208 Steps: 98530, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000606, Sample Num: 9696, Cur Loss: 0.07721078, Cur Avg Loss: 0.13925225, Log Avg loss: 0.08334569, Global Avg Loss: 0.63887679, Time: 0.0207 Steps: 98540, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000616, Sample Num: 9856, Cur Loss: 0.14395033, Cur Avg Loss: 0.13919875, Log Avg loss: 0.13595655, Global Avg Loss: 0.63882576, Time: 0.0207 Steps: 98550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000626, Sample Num: 10016, Cur Loss: 0.38147768, Cur Avg Loss: 0.14098376, Log Avg loss: 0.25094012, Global Avg Loss: 0.63878640, Time: 0.0207 Steps: 98560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000636, Sample Num: 10176, Cur Loss: 0.16651928, Cur Avg Loss: 0.14108419, Log Avg loss: 0.14737129, Global Avg Loss: 0.63873655, Time: 0.0207 Steps: 98570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000646, Sample Num: 10336, Cur Loss: 0.05658030, Cur Avg Loss: 0.14022883, Log Avg loss: 0.08582817, Global Avg Loss: 0.63868046, Time: 0.0207 Steps: 98580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000656, Sample Num: 10496, Cur Loss: 0.15054056, Cur Avg Loss: 0.14012985, Log Avg loss: 0.13373558, Global Avg Loss: 0.63862924, Time: 0.0208 Steps: 98590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000666, Sample Num: 10656, Cur Loss: 0.05710689, Cur Avg Loss: 0.14025107, Log Avg loss: 0.14820293, Global Avg Loss: 0.63857951, Time: 0.0207 Steps: 98600, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000676, Sample Num: 10816, Cur Loss: 0.33024192, Cur Avg Loss: 0.14051943, Log Avg loss: 0.15839243, Global Avg Loss: 0.63853081, Time: 0.0207 Steps: 98610, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000686, Sample Num: 10976, Cur Loss: 0.03004325, Cur Avg Loss: 0.14016434, Log Avg loss: 0.11615995, Global Avg Loss: 0.63847784, Time: 0.0207 Steps: 98620, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000696, Sample Num: 11136, Cur Loss: 0.06315205, Cur Avg Loss: 0.14045376, Log Avg loss: 0.16030798, Global Avg Loss: 0.63842936, Time: 0.0207 Steps: 98630, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000706, Sample Num: 11296, Cur Loss: 0.20651168, Cur Avg Loss: 0.14023589, Log Avg loss: 0.12507245, Global Avg Loss: 0.63837732, Time: 0.0207 Steps: 98640, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000716, Sample Num: 11456, Cur Loss: 0.06075491, Cur Avg Loss: 0.14000194, Log Avg loss: 0.12348527, Global Avg Loss: 0.63832512, Time: 0.0208 Steps: 98650, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000726, Sample Num: 11616, Cur Loss: 0.04585690, Cur Avg Loss: 0.13934326, Log Avg loss: 0.09218142, Global Avg Loss: 0.63826977, Time: 0.0207 Steps: 98660, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000736, Sample Num: 11776, Cur Loss: 0.24982849, Cur Avg Loss: 0.13943689, Log Avg loss: 0.14623460, Global Avg Loss: 0.63821990, Time: 0.0207 Steps: 98670, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000746, Sample Num: 11936, Cur Loss: 0.15950289, Cur Avg Loss: 0.13903721, Log Avg loss: 0.10962073, Global Avg Loss: 0.63816633, Time: 0.0207 Steps: 98680, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000756, Sample Num: 12096, Cur Loss: 0.04846045, Cur Avg Loss: 0.13940219, Log Avg loss: 0.16663001, Global Avg Loss: 0.63811855, Time: 0.0207 Steps: 98690, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000766, Sample Num: 12256, Cur Loss: 0.12224386, Cur Avg Loss: 0.13937482, Log Avg loss: 0.13730510, Global Avg Loss: 0.63806781, Time: 0.0207 Steps: 98700, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000776, Sample Num: 12416, Cur Loss: 0.19862482, Cur Avg Loss: 0.14016437, Log Avg loss: 0.20064417, Global Avg Loss: 0.63802350, Time: 0.0208 Steps: 98710, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000786, Sample Num: 12576, Cur Loss: 0.18760918, Cur Avg Loss: 0.14067332, Log Avg loss: 0.18016777, Global Avg Loss: 0.63797712, Time: 0.0208 Steps: 98720, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000796, Sample Num: 12736, Cur Loss: 0.12042740, Cur Avg Loss: 0.14102934, Log Avg loss: 0.16901281, Global Avg Loss: 0.63792962, Time: 0.0208 Steps: 98730, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000806, Sample Num: 12896, Cur Loss: 0.20950733, Cur Avg Loss: 0.14161043, Log Avg loss: 0.18786470, Global Avg Loss: 0.63788404, Time: 0.0209 Steps: 98740, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000816, Sample Num: 13056, Cur Loss: 0.15432748, Cur Avg Loss: 0.14125777, Log Avg loss: 0.11283384, Global Avg Loss: 0.63783087, Time: 0.0208 Steps: 98750, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000826, Sample Num: 13216, Cur Loss: 0.29597604, Cur Avg Loss: 0.14184859, Log Avg loss: 0.19005923, Global Avg Loss: 0.63778553, Time: 0.0209 Steps: 98760, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000836, Sample Num: 13376, Cur Loss: 0.32476825, Cur Avg Loss: 0.14182909, Log Avg loss: 0.14021834, Global Avg Loss: 0.63773515, Time: 0.0208 Steps: 98770, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000846, Sample Num: 13536, Cur Loss: 0.09322850, Cur Avg Loss: 0.14204586, Log Avg loss: 0.16016775, Global Avg Loss: 0.63768681, Time: 0.0208 Steps: 98780, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000856, Sample Num: 13696, Cur Loss: 0.16005065, Cur Avg Loss: 0.14228537, Log Avg loss: 0.16254780, Global Avg Loss: 0.63763871, Time: 0.0207 Steps: 98790, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000866, Sample Num: 13856, Cur Loss: 0.07279591, Cur Avg Loss: 0.14270678, Log Avg loss: 0.17877996, Global Avg Loss: 0.63759227, Time: 0.0208 Steps: 98800, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000876, Sample Num: 14016, Cur Loss: 0.17222288, Cur Avg Loss: 0.14295331, Log Avg loss: 0.16430264, Global Avg Loss: 0.63754437, Time: 0.0208 Steps: 98810, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000886, Sample Num: 14176, Cur Loss: 0.03707794, Cur Avg Loss: 0.14328571, Log Avg loss: 0.17240421, Global Avg Loss: 0.63749730, Time: 0.0207 Steps: 98820, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000896, Sample Num: 14336, Cur Loss: 0.15280691, Cur Avg Loss: 0.14374097, Log Avg loss: 0.18407634, Global Avg Loss: 0.63745142, Time: 0.0207 Steps: 98830, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000906, Sample Num: 14496, Cur Loss: 0.05377577, Cur Avg Loss: 0.14349564, Log Avg loss: 0.12151488, Global Avg Loss: 0.63739922, Time: 0.0208 Steps: 98840, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000916, Sample Num: 14656, Cur Loss: 0.04869174, Cur Avg Loss: 0.14296233, Log Avg loss: 0.09464390, Global Avg Loss: 0.63734431, Time: 0.0208 Steps: 98850, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000926, Sample Num: 14816, Cur Loss: 0.03935467, Cur Avg Loss: 0.14305785, Log Avg loss: 0.15180791, Global Avg Loss: 0.63729520, Time: 0.0207 Steps: 98860, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000936, Sample Num: 14976, Cur Loss: 0.05212148, Cur Avg Loss: 0.14315555, Log Avg loss: 0.15220201, Global Avg Loss: 0.63724614, Time: 0.0207 Steps: 98870, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000946, Sample Num: 15136, Cur Loss: 0.14413363, Cur Avg Loss: 0.14256849, Log Avg loss: 0.08762034, Global Avg Loss: 0.63719055, Time: 0.0208 Steps: 98880, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000956, Sample Num: 15296, Cur Loss: 0.04318415, Cur Avg Loss: 0.14334458, Log Avg loss: 0.21676259, Global Avg Loss: 0.63714804, Time: 0.0207 Steps: 98890, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000966, Sample Num: 15456, Cur Loss: 0.17278817, Cur Avg Loss: 0.14323717, Log Avg loss: 0.13296849, Global Avg Loss: 0.63709706, Time: 0.0208 Steps: 98900, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000976, Sample Num: 15616, Cur Loss: 0.07806708, Cur Avg Loss: 0.14299692, Log Avg loss: 0.11978855, Global Avg Loss: 0.63704476, Time: 0.0208 Steps: 98910, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000986, Sample Num: 15776, Cur Loss: 0.32020167, Cur Avg Loss: 0.14276468, Log Avg loss: 0.12009809, Global Avg Loss: 0.63699250, Time: 0.0208 Steps: 98920, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000996, Sample Num: 15936, Cur Loss: 0.10733490, Cur Avg Loss: 0.14265178, Log Avg loss: 0.13152005, Global Avg Loss: 0.63694140, Time: 0.0207 Steps: 98930, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001006, Sample Num: 16096, Cur Loss: 0.18776107, Cur Avg Loss: 0.14216853, Log Avg loss: 0.09403704, Global Avg Loss: 0.63688653, Time: 0.0207 Steps: 98940, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001016, Sample Num: 16256, Cur Loss: 0.04312239, Cur Avg Loss: 0.14230142, Log Avg loss: 0.15567018, Global Avg Loss: 0.63683790, Time: 0.0207 Steps: 98950, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001026, Sample Num: 16416, Cur Loss: 0.13259441, Cur Avg Loss: 0.14205315, Log Avg loss: 0.11682865, Global Avg Loss: 0.63678535, Time: 0.0246 Steps: 98960, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001036, Sample Num: 16576, Cur Loss: 0.09242486, Cur Avg Loss: 0.14186705, Log Avg loss: 0.12277348, Global Avg Loss: 0.63673342, Time: 0.0208 Steps: 98970, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001046, Sample Num: 16736, Cur Loss: 0.11927417, Cur Avg Loss: 0.14176557, Log Avg loss: 0.13125193, Global Avg Loss: 0.63668235, Time: 0.0208 Steps: 98980, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001056, Sample Num: 16896, Cur Loss: 0.08138239, Cur Avg Loss: 0.14227343, Log Avg loss: 0.19539566, Global Avg Loss: 0.63663777, Time: 0.0208 Steps: 98990, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001066, Sample Num: 17056, Cur Loss: 0.12870257, Cur Avg Loss: 0.14265716, Log Avg loss: 0.18317856, Global Avg Loss: 0.63659196, Time: 0.0208 Steps: 99000, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001076, Sample Num: 17216, Cur Loss: 0.21905580, Cur Avg Loss: 0.14267156, Log Avg loss: 0.14420734, Global Avg Loss: 0.63654223, Time: 0.0208 Steps: 99010, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001086, Sample Num: 17376, Cur Loss: 0.07808619, Cur Avg Loss: 0.14233585, Log Avg loss: 0.10621277, Global Avg Loss: 0.63648868, Time: 0.0208 Steps: 99020, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001096, Sample Num: 17536, Cur Loss: 0.12734254, Cur Avg Loss: 0.14256003, Log Avg loss: 0.16690669, Global Avg Loss: 0.63644126, Time: 0.0208 Steps: 99030, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001106, Sample Num: 17696, Cur Loss: 0.06138748, Cur Avg Loss: 0.14255475, Log Avg loss: 0.14197569, Global Avg Loss: 0.63639133, Time: 0.0208 Steps: 99040, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001116, Sample Num: 17856, Cur Loss: 0.11213376, Cur Avg Loss: 0.14254608, Log Avg loss: 0.14158663, Global Avg Loss: 0.63634138, Time: 0.0208 Steps: 99050, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001126, Sample Num: 18016, Cur Loss: 0.12853669, Cur Avg Loss: 0.14250362, Log Avg loss: 0.13776518, Global Avg Loss: 0.63629105, Time: 0.0208 Steps: 99060, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001136, Sample Num: 18176, Cur Loss: 0.08989155, Cur Avg Loss: 0.14274557, Log Avg loss: 0.16998932, Global Avg Loss: 0.63624398, Time: 0.0208 Steps: 99070, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001146, Sample Num: 18336, Cur Loss: 0.07829085, Cur Avg Loss: 0.14271502, Log Avg loss: 0.13924506, Global Avg Loss: 0.63619382, Time: 0.0208 Steps: 99080, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001156, Sample Num: 18496, Cur Loss: 0.05475260, Cur Avg Loss: 0.14276973, Log Avg loss: 0.14903968, Global Avg Loss: 0.63614465, Time: 0.0208 Steps: 99090, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001166, Sample Num: 18656, Cur Loss: 0.18276425, Cur Avg Loss: 0.14310819, Log Avg loss: 0.18223333, Global Avg Loss: 0.63609885, Time: 0.0208 Steps: 99100, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001176, Sample Num: 18816, Cur Loss: 0.44239557, Cur Avg Loss: 0.14352191, Log Avg loss: 0.19176181, Global Avg Loss: 0.63605402, Time: 0.0208 Steps: 99110, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001186, Sample Num: 18976, Cur Loss: 0.21034032, Cur Avg Loss: 0.14372904, Log Avg loss: 0.16808820, Global Avg Loss: 0.63600681, Time: 0.0208 Steps: 99120, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001196, Sample Num: 19136, Cur Loss: 0.14034107, Cur Avg Loss: 0.14374219, Log Avg loss: 0.14530180, Global Avg Loss: 0.63595731, Time: 0.0208 Steps: 99130, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001206, Sample Num: 19296, Cur Loss: 0.17609502, Cur Avg Loss: 0.14383046, Log Avg loss: 0.15438719, Global Avg Loss: 0.63590873, Time: 0.0208 Steps: 99140, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001216, Sample Num: 19456, Cur Loss: 0.28102934, Cur Avg Loss: 0.14384482, Log Avg loss: 0.14557705, Global Avg Loss: 0.63585928, Time: 0.0208 Steps: 99150, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001226, Sample Num: 19616, Cur Loss: 0.05321518, Cur Avg Loss: 0.14384046, Log Avg loss: 0.14330983, Global Avg Loss: 0.63580960, Time: 0.0208 Steps: 99160, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001236, Sample Num: 19776, Cur Loss: 0.11648502, Cur Avg Loss: 0.14398973, Log Avg loss: 0.16229034, Global Avg Loss: 0.63576186, Time: 0.0208 Steps: 99170, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001246, Sample Num: 19936, Cur Loss: 0.08331519, Cur Avg Loss: 0.14373234, Log Avg loss: 0.11191866, Global Avg Loss: 0.63570904, Time: 0.0208 Steps: 99180, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001256, Sample Num: 20096, Cur Loss: 0.09615124, Cur Avg Loss: 0.14384066, Log Avg loss: 0.15733698, Global Avg Loss: 0.63566081, Time: 0.0208 Steps: 99190, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001266, Sample Num: 20256, Cur Loss: 0.04769865, Cur Avg Loss: 0.14375148, Log Avg loss: 0.13255155, Global Avg Loss: 0.63561009, Time: 0.0208 Steps: 99200, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001276, Sample Num: 20416, Cur Loss: 0.08323502, Cur Avg Loss: 0.14360848, Log Avg loss: 0.12550377, Global Avg Loss: 0.63555868, Time: 0.0208 Steps: 99210, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001286, Sample Num: 20576, Cur Loss: 0.19941401, Cur Avg Loss: 0.14366743, Log Avg loss: 0.15118967, Global Avg Loss: 0.63550986, Time: 0.0248 Steps: 99220, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001296, Sample Num: 20736, Cur Loss: 0.10470203, Cur Avg Loss: 0.14380482, Log Avg loss: 0.16147322, Global Avg Loss: 0.63546209, Time: 0.0210 Steps: 99230, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001306, Sample Num: 20896, Cur Loss: 0.13641696, Cur Avg Loss: 0.14362644, Log Avg loss: 0.12050827, Global Avg Loss: 0.63541020, Time: 0.0210 Steps: 99240, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001316, Sample Num: 21056, Cur Loss: 0.05253786, Cur Avg Loss: 0.14357943, Log Avg loss: 0.13743979, Global Avg Loss: 0.63536003, Time: 0.0210 Steps: 99250, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001326, Sample Num: 21216, Cur Loss: 0.15759559, Cur Avg Loss: 0.14424472, Log Avg loss: 0.23179660, Global Avg Loss: 0.63531937, Time: 0.0210 Steps: 99260, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001336, Sample Num: 21376, Cur Loss: 0.22412616, Cur Avg Loss: 0.14451481, Log Avg loss: 0.18032918, Global Avg Loss: 0.63527353, Time: 0.0210 Steps: 99270, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001346, Sample Num: 21536, Cur Loss: 0.15598671, Cur Avg Loss: 0.14435129, Log Avg loss: 0.12250466, Global Avg Loss: 0.63522189, Time: 0.0210 Steps: 99280, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001356, Sample Num: 21696, Cur Loss: 0.13223459, Cur Avg Loss: 0.14456130, Log Avg loss: 0.17282905, Global Avg Loss: 0.63517532, Time: 0.0210 Steps: 99290, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001366, Sample Num: 21856, Cur Loss: 0.12853900, Cur Avg Loss: 0.14445562, Log Avg loss: 0.13012539, Global Avg Loss: 0.63512445, Time: 0.0210 Steps: 99300, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001376, Sample Num: 22016, Cur Loss: 0.13029963, Cur Avg Loss: 0.14467175, Log Avg loss: 0.17419507, Global Avg Loss: 0.63507804, Time: 0.0210 Steps: 99310, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001386, Sample Num: 22176, Cur Loss: 0.03930155, Cur Avg Loss: 0.14448935, Log Avg loss: 0.11939132, Global Avg Loss: 0.63502612, Time: 0.0209 Steps: 99320, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001396, Sample Num: 22336, Cur Loss: 0.05159002, Cur Avg Loss: 0.14423235, Log Avg loss: 0.10861233, Global Avg Loss: 0.63497312, Time: 0.0210 Steps: 99330, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001406, Sample Num: 22496, Cur Loss: 0.09887878, Cur Avg Loss: 0.14439462, Log Avg loss: 0.16704740, Global Avg Loss: 0.63492602, Time: 0.0209 Steps: 99340, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001416, Sample Num: 22656, Cur Loss: 0.07829873, Cur Avg Loss: 0.14427813, Log Avg loss: 0.12789994, Global Avg Loss: 0.63487499, Time: 0.0210 Steps: 99350, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001426, Sample Num: 22816, Cur Loss: 0.03525434, Cur Avg Loss: 0.14394299, Log Avg loss: 0.09648640, Global Avg Loss: 0.63482080, Time: 0.0210 Steps: 99360, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001436, Sample Num: 22976, Cur Loss: 0.16555282, Cur Avg Loss: 0.14392037, Log Avg loss: 0.14069497, Global Avg Loss: 0.63477107, Time: 0.0210 Steps: 99370, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001446, Sample Num: 23136, Cur Loss: 0.15313423, Cur Avg Loss: 0.14373697, Log Avg loss: 0.11740095, Global Avg Loss: 0.63471901, Time: 0.0210 Steps: 99380, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001456, Sample Num: 23296, Cur Loss: 0.09854030, Cur Avg Loss: 0.14351321, Log Avg loss: 0.11115747, Global Avg Loss: 0.63466634, Time: 0.0211 Steps: 99390, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001466, Sample Num: 23456, Cur Loss: 0.36970070, Cur Avg Loss: 0.14368013, Log Avg loss: 0.16798385, Global Avg Loss: 0.63461939, Time: 0.0210 Steps: 99400, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001476, Sample Num: 23616, Cur Loss: 0.20997055, Cur Avg Loss: 0.14369611, Log Avg loss: 0.14603845, Global Avg Loss: 0.63457024, Time: 0.0210 Steps: 99410, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001486, Sample Num: 23776, Cur Loss: 0.15144038, Cur Avg Loss: 0.14364803, Log Avg loss: 0.13655161, Global Avg Loss: 0.63452015, Time: 0.0210 Steps: 99420, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001496, Sample Num: 23936, Cur Loss: 0.10997311, Cur Avg Loss: 0.14385453, Log Avg loss: 0.17454105, Global Avg Loss: 0.63447388, Time: 0.0210 Steps: 99430, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001506, Sample Num: 24096, Cur Loss: 0.23470567, Cur Avg Loss: 0.14407810, Log Avg loss: 0.17752421, Global Avg Loss: 0.63442793, Time: 0.0210 Steps: 99440, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001516, Sample Num: 24256, Cur Loss: 0.08773968, Cur Avg Loss: 0.14401802, Log Avg loss: 0.13496954, Global Avg Loss: 0.63437771, Time: 0.0210 Steps: 99450, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001526, Sample Num: 24416, Cur Loss: 0.23338740, Cur Avg Loss: 0.14430007, Log Avg loss: 0.18705849, Global Avg Loss: 0.63433274, Time: 0.0209 Steps: 99460, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001536, Sample Num: 24576, Cur Loss: 0.04011941, Cur Avg Loss: 0.14408488, Log Avg loss: 0.11124625, Global Avg Loss: 0.63428015, Time: 0.0255 Steps: 99470, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001546, Sample Num: 24736, Cur Loss: 0.10267269, Cur Avg Loss: 0.14393210, Log Avg loss: 0.12046520, Global Avg Loss: 0.63422850, Time: 0.0208 Steps: 99480, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001556, Sample Num: 24896, Cur Loss: 0.03857128, Cur Avg Loss: 0.14380623, Log Avg loss: 0.12434755, Global Avg Loss: 0.63417725, Time: 0.0208 Steps: 99490, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001566, Sample Num: 25056, Cur Loss: 0.22235608, Cur Avg Loss: 0.14367816, Log Avg loss: 0.12374994, Global Avg Loss: 0.63412595, Time: 0.0207 Steps: 99500, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001576, Sample Num: 25216, Cur Loss: 0.09586641, Cur Avg Loss: 0.14366163, Log Avg loss: 0.14107272, Global Avg Loss: 0.63407640, Time: 0.0207 Steps: 99510, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001586, Sample Num: 25376, Cur Loss: 0.18170762, Cur Avg Loss: 0.14379350, Log Avg loss: 0.16457727, Global Avg Loss: 0.63402923, Time: 0.0208 Steps: 99520, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001596, Sample Num: 25536, Cur Loss: 0.06437033, Cur Avg Loss: 0.14379305, Log Avg loss: 0.14372139, Global Avg Loss: 0.63397996, Time: 0.0208 Steps: 99530, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001606, Sample Num: 25696, Cur Loss: 0.02159099, Cur Avg Loss: 0.14381895, Log Avg loss: 0.14795292, Global Avg Loss: 0.63393114, Time: 0.0208 Steps: 99540, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001616, Sample Num: 25856, Cur Loss: 0.07776517, Cur Avg Loss: 0.14393701, Log Avg loss: 0.16289709, Global Avg Loss: 0.63388382, Time: 0.0208 Steps: 99550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001626, Sample Num: 26016, Cur Loss: 0.03865244, Cur Avg Loss: 0.14393962, Log Avg loss: 0.14436076, Global Avg Loss: 0.63383465, Time: 0.0208 Steps: 99560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001636, Sample Num: 26176, Cur Loss: 0.27460676, Cur Avg Loss: 0.14387052, Log Avg loss: 0.13263471, Global Avg Loss: 0.63378431, Time: 0.0208 Steps: 99570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001646, Sample Num: 26336, Cur Loss: 0.30366787, Cur Avg Loss: 0.14406236, Log Avg loss: 0.17544762, Global Avg Loss: 0.63373829, Time: 0.0208 Steps: 99580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001656, Sample Num: 26496, Cur Loss: 0.12643170, Cur Avg Loss: 0.14411958, Log Avg loss: 0.15353925, Global Avg Loss: 0.63369007, Time: 0.0208 Steps: 99590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001666, Sample Num: 26656, Cur Loss: 0.06252110, Cur Avg Loss: 0.14377556, Log Avg loss: 0.08680502, Global Avg Loss: 0.63363516, Time: 0.0207 Steps: 99600, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001676, Sample Num: 26816, Cur Loss: 0.19332255, Cur Avg Loss: 0.14349393, Log Avg loss: 0.09657465, Global Avg Loss: 0.63358125, Time: 0.0208 Steps: 99610, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001686, Sample Num: 26976, Cur Loss: 0.16919377, Cur Avg Loss: 0.14375523, Log Avg loss: 0.18754868, Global Avg Loss: 0.63353647, Time: 0.0207 Steps: 99620, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001696, Sample Num: 27136, Cur Loss: 0.11104909, Cur Avg Loss: 0.14374627, Log Avg loss: 0.14223643, Global Avg Loss: 0.63348716, Time: 0.0207 Steps: 99630, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001706, Sample Num: 27296, Cur Loss: 0.06509735, Cur Avg Loss: 0.14367672, Log Avg loss: 0.13187983, Global Avg Loss: 0.63343682, Time: 0.0208 Steps: 99640, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001716, Sample Num: 27456, Cur Loss: 0.06859826, Cur Avg Loss: 0.14386064, Log Avg loss: 0.17523806, Global Avg Loss: 0.63339084, Time: 0.0207 Steps: 99650, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001726, Sample Num: 27616, Cur Loss: 0.15629351, Cur Avg Loss: 0.14370838, Log Avg loss: 0.11758046, Global Avg Loss: 0.63333908, Time: 0.0208 Steps: 99660, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001736, Sample Num: 27776, Cur Loss: 0.16670437, Cur Avg Loss: 0.14353664, Log Avg loss: 0.11389370, Global Avg Loss: 0.63328696, Time: 0.0208 Steps: 99670, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001746, Sample Num: 27936, Cur Loss: 0.35856789, Cur Avg Loss: 0.14360264, Log Avg loss: 0.15506140, Global Avg Loss: 0.63323899, Time: 0.0208 Steps: 99680, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001756, Sample Num: 28096, Cur Loss: 0.16601132, Cur Avg Loss: 0.14381600, Log Avg loss: 0.18106884, Global Avg Loss: 0.63319363, Time: 0.0208 Steps: 99690, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001766, Sample Num: 28256, Cur Loss: 0.22467767, Cur Avg Loss: 0.14388738, Log Avg loss: 0.15642147, Global Avg Loss: 0.63314581, Time: 0.0207 Steps: 99700, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001776, Sample Num: 28416, Cur Loss: 0.08232811, Cur Avg Loss: 0.14402288, Log Avg loss: 0.16795106, Global Avg Loss: 0.63309915, Time: 0.0208 Steps: 99710, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001786, Sample Num: 28576, Cur Loss: 0.27803895, Cur Avg Loss: 0.14450994, Log Avg loss: 0.23101157, Global Avg Loss: 0.63305883, Time: 0.0207 Steps: 99720, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001796, Sample Num: 28736, Cur Loss: 0.38829115, Cur Avg Loss: 0.14450253, Log Avg loss: 0.14317923, Global Avg Loss: 0.63300971, Time: 0.0220 Steps: 99730, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001806, Sample Num: 28896, Cur Loss: 0.07091872, Cur Avg Loss: 0.14436586, Log Avg loss: 0.11982081, Global Avg Loss: 0.63295826, Time: 0.0208 Steps: 99740, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001816, Sample Num: 29056, Cur Loss: 0.07605499, Cur Avg Loss: 0.14432896, Log Avg loss: 0.13766522, Global Avg Loss: 0.63290861, Time: 0.0208 Steps: 99750, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001826, Sample Num: 29216, Cur Loss: 0.06755270, Cur Avg Loss: 0.14417499, Log Avg loss: 0.11621287, Global Avg Loss: 0.63285681, Time: 0.0208 Steps: 99760, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001836, Sample Num: 29376, Cur Loss: 0.21496861, Cur Avg Loss: 0.14447448, Log Avg loss: 0.19916191, Global Avg Loss: 0.63281334, Time: 0.0208 Steps: 99770, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001846, Sample Num: 29536, Cur Loss: 0.37018213, Cur Avg Loss: 0.14485870, Log Avg loss: 0.21540173, Global Avg Loss: 0.63277151, Time: 0.0208 Steps: 99780, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001856, Sample Num: 29696, Cur Loss: 0.40399665, Cur Avg Loss: 0.14501890, Log Avg loss: 0.17459099, Global Avg Loss: 0.63272559, Time: 0.0208 Steps: 99790, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001866, Sample Num: 29856, Cur Loss: 0.09795402, Cur Avg Loss: 0.14543815, Log Avg loss: 0.22325146, Global Avg Loss: 0.63268456, Time: 0.0208 Steps: 99800, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001876, Sample Num: 30016, Cur Loss: 0.17641881, Cur Avg Loss: 0.14539400, Log Avg loss: 0.13715652, Global Avg Loss: 0.63263492, Time: 0.0208 Steps: 99810, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001886, Sample Num: 30176, Cur Loss: 0.06563845, Cur Avg Loss: 0.14530312, Log Avg loss: 0.12825354, Global Avg Loss: 0.63258439, Time: 0.0208 Steps: 99820, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001896, Sample Num: 30336, Cur Loss: 0.35969824, Cur Avg Loss: 0.14552668, Log Avg loss: 0.18768910, Global Avg Loss: 0.63253982, Time: 0.0208 Steps: 99830, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001906, Sample Num: 30496, Cur Loss: 0.09118859, Cur Avg Loss: 0.14572782, Log Avg loss: 0.18386456, Global Avg Loss: 0.63249488, Time: 0.0208 Steps: 99840, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001916, Sample Num: 30656, Cur Loss: 0.13192804, Cur Avg Loss: 0.14555909, Log Avg loss: 0.11339888, Global Avg Loss: 0.63244290, Time: 0.0209 Steps: 99850, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001926, Sample Num: 30816, Cur Loss: 0.31177127, Cur Avg Loss: 0.14590882, Log Avg loss: 0.21291711, Global Avg Loss: 0.63240088, Time: 0.0208 Steps: 99860, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001936, Sample Num: 30976, Cur Loss: 0.15870795, Cur Avg Loss: 0.14585205, Log Avg loss: 0.13491805, Global Avg Loss: 0.63235107, Time: 0.0209 Steps: 99870, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001946, Sample Num: 31136, Cur Loss: 0.08468354, Cur Avg Loss: 0.14601209, Log Avg loss: 0.17699531, Global Avg Loss: 0.63230548, Time: 0.0208 Steps: 99880, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001956, Sample Num: 31296, Cur Loss: 0.06100896, Cur Avg Loss: 0.14591279, Log Avg loss: 0.12658904, Global Avg Loss: 0.63225485, Time: 0.0209 Steps: 99890, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001966, Sample Num: 31456, Cur Loss: 0.13226810, Cur Avg Loss: 0.14607834, Log Avg loss: 0.17846101, Global Avg Loss: 0.63220943, Time: 0.0208 Steps: 99900, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001976, Sample Num: 31616, Cur Loss: 0.19321916, Cur Avg Loss: 0.14625927, Log Avg loss: 0.18182952, Global Avg Loss: 0.63216435, Time: 0.0208 Steps: 99910, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001986, Sample Num: 31776, Cur Loss: 0.43573740, Cur Avg Loss: 0.14642535, Log Avg loss: 0.17924221, Global Avg Loss: 0.63211902, Time: 0.0208 Steps: 99920, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001996, Sample Num: 31936, Cur Loss: 0.26291004, Cur Avg Loss: 0.14661219, Log Avg loss: 0.18371925, Global Avg Loss: 0.63207415, Time: 0.0208 Steps: 99930, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002006, Sample Num: 32096, Cur Loss: 0.12010714, Cur Avg Loss: 0.14653776, Log Avg loss: 0.13168119, Global Avg Loss: 0.63202408, Time: 0.0209 Steps: 99940, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002016, Sample Num: 32256, Cur Loss: 0.10833386, Cur Avg Loss: 0.14663316, Log Avg loss: 0.16577070, Global Avg Loss: 0.63197743, Time: 0.0208 Steps: 99950, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002026, Sample Num: 32416, Cur Loss: 0.19049904, Cur Avg Loss: 0.14667286, Log Avg loss: 0.15467588, Global Avg Loss: 0.63192968, Time: 0.0207 Steps: 99960, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002036, Sample Num: 32576, Cur Loss: 0.14386399, Cur Avg Loss: 0.14668359, Log Avg loss: 0.14885849, Global Avg Loss: 0.63188136, Time: 0.0208 Steps: 99970, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002046, Sample Num: 32736, Cur Loss: 0.22511923, Cur Avg Loss: 0.14658767, Log Avg loss: 0.12705826, Global Avg Loss: 0.63183087, Time: 0.0208 Steps: 99980, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002056, Sample Num: 32896, Cur Loss: 0.08268799, Cur Avg Loss: 0.14621795, Log Avg loss: 0.07057330, Global Avg Loss: 0.63177474, Time: 0.0209 Steps: 99990, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002066, Sample Num: 33056, Cur Loss: 0.41329056, Cur Avg Loss: 0.14624593, Log Avg loss: 0.15199904, Global Avg Loss: 0.63172676, Time: 0.0209 Steps: 100000, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002076, Sample Num: 33216, Cur Loss: 0.13264951, Cur Avg Loss: 0.14633662, Log Avg loss: 0.16507235, Global Avg Loss: 0.63168010, Time: 0.0208 Steps: 100010, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002086, Sample Num: 33376, Cur Loss: 0.16932644, Cur Avg Loss: 0.14632229, Log Avg loss: 0.14334852, Global Avg Loss: 0.63163128, Time: 0.0209 Steps: 100020, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002096, Sample Num: 33536, Cur Loss: 0.02685524, Cur Avg Loss: 0.14598791, Log Avg loss: 0.07623599, Global Avg Loss: 0.63157575, Time: 0.0209 Steps: 100030, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002106, Sample Num: 33696, Cur Loss: 0.11287477, Cur Avg Loss: 0.14583636, Log Avg loss: 0.11407067, Global Avg Loss: 0.63152402, Time: 0.0209 Steps: 100040, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002116, Sample Num: 33856, Cur Loss: 0.26444858, Cur Avg Loss: 0.14598508, Log Avg loss: 0.17730477, Global Avg Loss: 0.63147862, Time: 0.0209 Steps: 100050, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002126, Sample Num: 34016, Cur Loss: 0.14379053, Cur Avg Loss: 0.14590740, Log Avg loss: 0.12947087, Global Avg Loss: 0.63142845, Time: 0.0209 Steps: 100060, Updated lr: 0.000006 ***** Running evaluation checkpoint-100063 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-100063 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.625182, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.244972, "eval_total_loss": 172.215032, "eval_mae": 0.329262, "eval_mse": 0.245065, "eval_r2": 0.84422, "eval_sp_statistic": 0.905377, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925124, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.768797, "test_total_loss": 385.936258, "test_mae": 0.687584, "test_mse": 0.768839, "test_r2": 0.503785, "test_sp_statistic": 0.800052, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.838486, "test_ps_pvalue": 0.0, "lr": 6.056899004267426e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6314125290056296, "train_cur_epoch_loss": 310.49989979341626, "train_cur_epoch_avg_loss": 0.14584307176769198, "train_cur_epoch_time": 44.625181674957275, "train_cur_epoch_avg_time": 0.020960630190210088, "epoch": 47, "step": 100063} ################################################## Training, Epoch: 0048, Batch: 000007, Sample Num: 112, Cur Loss: 0.12916659, Cur Avg Loss: 0.19668362, Log Avg loss: 0.16775569, Global Avg Loss: 0.63138212, Time: 0.0210 Steps: 100070, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000017, Sample Num: 272, Cur Loss: 0.07217350, Cur Avg Loss: 0.16559137, Log Avg loss: 0.14382680, Global Avg Loss: 0.63133340, Time: 0.0208 Steps: 100080, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000027, Sample Num: 432, Cur Loss: 0.22058620, Cur Avg Loss: 0.14700237, Log Avg loss: 0.11540106, Global Avg Loss: 0.63128186, Time: 0.0208 Steps: 100090, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000037, Sample Num: 592, Cur Loss: 0.10388322, Cur Avg Loss: 0.14748092, Log Avg loss: 0.14877300, Global Avg Loss: 0.63123365, Time: 0.0208 Steps: 100100, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000047, Sample Num: 752, Cur Loss: 0.14226240, Cur Avg Loss: 0.13642446, Log Avg loss: 0.09551556, Global Avg Loss: 0.63118014, Time: 0.0208 Steps: 100110, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000057, Sample Num: 912, Cur Loss: 0.06470497, Cur Avg Loss: 0.13917411, Log Avg loss: 0.15209749, Global Avg Loss: 0.63113229, Time: 0.0208 Steps: 100120, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000067, Sample Num: 1072, Cur Loss: 0.07723943, Cur Avg Loss: 0.13829625, Log Avg loss: 0.13329246, Global Avg Loss: 0.63108257, Time: 0.0208 Steps: 100130, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000077, Sample Num: 1232, Cur Loss: 0.07275524, Cur Avg Loss: 0.13983750, Log Avg loss: 0.15016383, Global Avg Loss: 0.63103455, Time: 0.0208 Steps: 100140, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000087, Sample Num: 1392, Cur Loss: 0.05408454, Cur Avg Loss: 0.13517921, Log Avg loss: 0.09931041, Global Avg Loss: 0.63098145, Time: 0.0208 Steps: 100150, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000097, Sample Num: 1552, Cur Loss: 0.58990550, Cur Avg Loss: 0.14084991, Log Avg loss: 0.19018495, Global Avg Loss: 0.63093744, Time: 0.0208 Steps: 100160, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000107, Sample Num: 1712, Cur Loss: 0.10889260, Cur Avg Loss: 0.14316715, Log Avg loss: 0.16564445, Global Avg Loss: 0.63089099, Time: 0.0208 Steps: 100170, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000117, Sample Num: 1872, Cur Loss: 0.21628588, Cur Avg Loss: 0.14261220, Log Avg loss: 0.13667419, Global Avg Loss: 0.63084166, Time: 0.0208 Steps: 100180, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000127, Sample Num: 2032, Cur Loss: 0.10525782, Cur Avg Loss: 0.14183828, Log Avg loss: 0.13278337, Global Avg Loss: 0.63079195, Time: 0.0208 Steps: 100190, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000137, Sample Num: 2192, Cur Loss: 0.21385549, Cur Avg Loss: 0.14117135, Log Avg loss: 0.13270141, Global Avg Loss: 0.63074224, Time: 0.0208 Steps: 100200, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000147, Sample Num: 2352, Cur Loss: 0.14122812, Cur Avg Loss: 0.14322079, Log Avg loss: 0.17129810, Global Avg Loss: 0.63069639, Time: 0.0208 Steps: 100210, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000157, Sample Num: 2512, Cur Loss: 0.13942648, Cur Avg Loss: 0.14556123, Log Avg loss: 0.17996574, Global Avg Loss: 0.63065142, Time: 0.0208 Steps: 100220, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000167, Sample Num: 2672, Cur Loss: 0.31498012, Cur Avg Loss: 0.14521983, Log Avg loss: 0.13985973, Global Avg Loss: 0.63060245, Time: 0.0208 Steps: 100230, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000177, Sample Num: 2832, Cur Loss: 0.06954383, Cur Avg Loss: 0.14345017, Log Avg loss: 0.11389699, Global Avg Loss: 0.63055090, Time: 0.0208 Steps: 100240, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000187, Sample Num: 2992, Cur Loss: 0.05206279, Cur Avg Loss: 0.14020550, Log Avg loss: 0.08277484, Global Avg Loss: 0.63049626, Time: 0.0207 Steps: 100250, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000197, Sample Num: 3152, Cur Loss: 0.28050295, Cur Avg Loss: 0.14180192, Log Avg loss: 0.17165498, Global Avg Loss: 0.63045050, Time: 0.0207 Steps: 100260, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000207, Sample Num: 3312, Cur Loss: 0.05459352, Cur Avg Loss: 0.13881793, Log Avg loss: 0.08003318, Global Avg Loss: 0.63039560, Time: 0.0207 Steps: 100270, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000217, Sample Num: 3472, Cur Loss: 0.14324881, Cur Avg Loss: 0.14110130, Log Avg loss: 0.18836715, Global Avg Loss: 0.63035152, Time: 0.0207 Steps: 100280, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000227, Sample Num: 3632, Cur Loss: 0.06765050, Cur Avg Loss: 0.13888501, Log Avg loss: 0.09079153, Global Avg Loss: 0.63029772, Time: 0.0207 Steps: 100290, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000237, Sample Num: 3792, Cur Loss: 0.30976543, Cur Avg Loss: 0.13817386, Log Avg loss: 0.12203083, Global Avg Loss: 0.63024705, Time: 0.0207 Steps: 100300, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000247, Sample Num: 3952, Cur Loss: 0.18200904, Cur Avg Loss: 0.13811251, Log Avg loss: 0.13665848, Global Avg Loss: 0.63019784, Time: 0.0207 Steps: 100310, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000257, Sample Num: 4112, Cur Loss: 0.15296440, Cur Avg Loss: 0.13971052, Log Avg loss: 0.17918138, Global Avg Loss: 0.63015289, Time: 0.0246 Steps: 100320, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000267, Sample Num: 4272, Cur Loss: 0.01875539, Cur Avg Loss: 0.13872173, Log Avg loss: 0.11330961, Global Avg Loss: 0.63010137, Time: 0.0209 Steps: 100330, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000277, Sample Num: 4432, Cur Loss: 0.14016816, Cur Avg Loss: 0.13847094, Log Avg loss: 0.13177488, Global Avg Loss: 0.63005171, Time: 0.0208 Steps: 100340, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000287, Sample Num: 4592, Cur Loss: 0.06456894, Cur Avg Loss: 0.13906233, Log Avg loss: 0.15544406, Global Avg Loss: 0.63000441, Time: 0.0208 Steps: 100350, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000297, Sample Num: 4752, Cur Loss: 0.32841632, Cur Avg Loss: 0.13896608, Log Avg loss: 0.13620373, Global Avg Loss: 0.62995521, Time: 0.0208 Steps: 100360, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000307, Sample Num: 4912, Cur Loss: 0.10255609, Cur Avg Loss: 0.13954054, Log Avg loss: 0.15660193, Global Avg Loss: 0.62990805, Time: 0.0208 Steps: 100370, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000317, Sample Num: 5072, Cur Loss: 0.06386328, Cur Avg Loss: 0.13936442, Log Avg loss: 0.13395761, Global Avg Loss: 0.62985864, Time: 0.0208 Steps: 100380, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000327, Sample Num: 5232, Cur Loss: 0.06476563, Cur Avg Loss: 0.13899883, Log Avg loss: 0.12740955, Global Avg Loss: 0.62980859, Time: 0.0209 Steps: 100390, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000337, Sample Num: 5392, Cur Loss: 0.25214925, Cur Avg Loss: 0.13847694, Log Avg loss: 0.12141111, Global Avg Loss: 0.62975795, Time: 0.0208 Steps: 100400, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000347, Sample Num: 5552, Cur Loss: 0.07086273, Cur Avg Loss: 0.13792671, Log Avg loss: 0.11938404, Global Avg Loss: 0.62970713, Time: 0.0208 Steps: 100410, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000357, Sample Num: 5712, Cur Loss: 0.11046463, Cur Avg Loss: 0.13774076, Log Avg loss: 0.13128833, Global Avg Loss: 0.62965749, Time: 0.0208 Steps: 100420, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000367, Sample Num: 5872, Cur Loss: 0.13096815, Cur Avg Loss: 0.13733537, Log Avg loss: 0.12286296, Global Avg Loss: 0.62960703, Time: 0.0208 Steps: 100430, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000377, Sample Num: 6032, Cur Loss: 0.22300421, Cur Avg Loss: 0.13778589, Log Avg loss: 0.15431975, Global Avg Loss: 0.62955971, Time: 0.0208 Steps: 100440, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000387, Sample Num: 6192, Cur Loss: 0.12393353, Cur Avg Loss: 0.13754159, Log Avg loss: 0.12833170, Global Avg Loss: 0.62950981, Time: 0.0209 Steps: 100450, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000397, Sample Num: 6352, Cur Loss: 0.31171399, Cur Avg Loss: 0.13757084, Log Avg loss: 0.13870283, Global Avg Loss: 0.62946095, Time: 0.0207 Steps: 100460, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000407, Sample Num: 6512, Cur Loss: 0.03857277, Cur Avg Loss: 0.13753724, Log Avg loss: 0.13620320, Global Avg Loss: 0.62941186, Time: 0.0208 Steps: 100470, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000417, Sample Num: 6672, Cur Loss: 0.12575877, Cur Avg Loss: 0.13671802, Log Avg loss: 0.10337563, Global Avg Loss: 0.62935951, Time: 0.0208 Steps: 100480, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000427, Sample Num: 6832, Cur Loss: 0.10145518, Cur Avg Loss: 0.13789853, Log Avg loss: 0.18712586, Global Avg Loss: 0.62931550, Time: 0.0208 Steps: 100490, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000437, Sample Num: 6992, Cur Loss: 0.08368270, Cur Avg Loss: 0.13841121, Log Avg loss: 0.16030261, Global Avg Loss: 0.62926883, Time: 0.0208 Steps: 100500, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000447, Sample Num: 7152, Cur Loss: 0.08854870, Cur Avg Loss: 0.13789077, Log Avg loss: 0.11514759, Global Avg Loss: 0.62921768, Time: 0.0208 Steps: 100510, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000457, Sample Num: 7312, Cur Loss: 0.14545369, Cur Avg Loss: 0.13765736, Log Avg loss: 0.12722386, Global Avg Loss: 0.62916774, Time: 0.0208 Steps: 100520, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000467, Sample Num: 7472, Cur Loss: 0.05787022, Cur Avg Loss: 0.13755785, Log Avg loss: 0.13301020, Global Avg Loss: 0.62911839, Time: 0.0208 Steps: 100530, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000477, Sample Num: 7632, Cur Loss: 0.05799033, Cur Avg Loss: 0.13748357, Log Avg loss: 0.13401468, Global Avg Loss: 0.62906914, Time: 0.0208 Steps: 100540, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000487, Sample Num: 7792, Cur Loss: 0.04984225, Cur Avg Loss: 0.13668699, Log Avg loss: 0.09869019, Global Avg Loss: 0.62901639, Time: 0.0208 Steps: 100550, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000497, Sample Num: 7952, Cur Loss: 0.23060012, Cur Avg Loss: 0.13663753, Log Avg loss: 0.13422878, Global Avg Loss: 0.62896719, Time: 0.0208 Steps: 100560, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000507, Sample Num: 8112, Cur Loss: 0.04283636, Cur Avg Loss: 0.13655031, Log Avg loss: 0.13221553, Global Avg Loss: 0.62891780, Time: 0.0208 Steps: 100570, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000517, Sample Num: 8272, Cur Loss: 0.03987565, Cur Avg Loss: 0.13638636, Log Avg loss: 0.12807425, Global Avg Loss: 0.62886800, Time: 0.0208 Steps: 100580, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000527, Sample Num: 8432, Cur Loss: 0.13403316, Cur Avg Loss: 0.13698293, Log Avg loss: 0.16782542, Global Avg Loss: 0.62882217, Time: 0.0208 Steps: 100590, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000537, Sample Num: 8592, Cur Loss: 0.14136174, Cur Avg Loss: 0.13703849, Log Avg loss: 0.13996655, Global Avg Loss: 0.62877357, Time: 0.0208 Steps: 100600, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000547, Sample Num: 8752, Cur Loss: 0.22020134, Cur Avg Loss: 0.13715641, Log Avg loss: 0.14348889, Global Avg Loss: 0.62872534, Time: 0.0207 Steps: 100610, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000557, Sample Num: 8912, Cur Loss: 0.13105705, Cur Avg Loss: 0.13623081, Log Avg loss: 0.08560011, Global Avg Loss: 0.62867136, Time: 0.0208 Steps: 100620, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000567, Sample Num: 9072, Cur Loss: 0.26732868, Cur Avg Loss: 0.13635455, Log Avg loss: 0.14324712, Global Avg Loss: 0.62862312, Time: 0.0208 Steps: 100630, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000577, Sample Num: 9232, Cur Loss: 0.11502228, Cur Avg Loss: 0.13603876, Log Avg loss: 0.11813351, Global Avg Loss: 0.62857240, Time: 0.0207 Steps: 100640, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000587, Sample Num: 9392, Cur Loss: 0.24726945, Cur Avg Loss: 0.13558740, Log Avg loss: 0.10954408, Global Avg Loss: 0.62852083, Time: 0.0208 Steps: 100650, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000597, Sample Num: 9552, Cur Loss: 0.21904576, Cur Avg Loss: 0.13552906, Log Avg loss: 0.13210438, Global Avg Loss: 0.62847152, Time: 0.0207 Steps: 100660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000607, Sample Num: 9712, Cur Loss: 0.20658478, Cur Avg Loss: 0.13489993, Log Avg loss: 0.09734069, Global Avg Loss: 0.62841876, Time: 0.0207 Steps: 100670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000617, Sample Num: 9872, Cur Loss: 0.18613884, Cur Avg Loss: 0.13514057, Log Avg loss: 0.14974759, Global Avg Loss: 0.62837121, Time: 0.0208 Steps: 100680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000627, Sample Num: 10032, Cur Loss: 0.13251691, Cur Avg Loss: 0.13545105, Log Avg loss: 0.15460758, Global Avg Loss: 0.62832416, Time: 0.0208 Steps: 100690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000637, Sample Num: 10192, Cur Loss: 0.32018781, Cur Avg Loss: 0.13598167, Log Avg loss: 0.16925132, Global Avg Loss: 0.62827857, Time: 0.0207 Steps: 100700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000647, Sample Num: 10352, Cur Loss: 0.15043214, Cur Avg Loss: 0.13627724, Log Avg loss: 0.15510501, Global Avg Loss: 0.62823159, Time: 0.0207 Steps: 100710, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000657, Sample Num: 10512, Cur Loss: 0.17222513, Cur Avg Loss: 0.13592254, Log Avg loss: 0.11297351, Global Avg Loss: 0.62818043, Time: 0.0207 Steps: 100720, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000667, Sample Num: 10672, Cur Loss: 0.02685144, Cur Avg Loss: 0.13634607, Log Avg loss: 0.16417239, Global Avg Loss: 0.62813437, Time: 0.0208 Steps: 100730, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000677, Sample Num: 10832, Cur Loss: 0.23459360, Cur Avg Loss: 0.13619494, Log Avg loss: 0.12611429, Global Avg Loss: 0.62808453, Time: 0.0207 Steps: 100740, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000687, Sample Num: 10992, Cur Loss: 0.04032190, Cur Avg Loss: 0.13556760, Log Avg loss: 0.09309665, Global Avg Loss: 0.62803143, Time: 0.0207 Steps: 100750, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000697, Sample Num: 11152, Cur Loss: 0.04360555, Cur Avg Loss: 0.13596737, Log Avg loss: 0.16343154, Global Avg Loss: 0.62798532, Time: 0.0207 Steps: 100760, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000707, Sample Num: 11312, Cur Loss: 0.34581608, Cur Avg Loss: 0.13695876, Log Avg loss: 0.20605875, Global Avg Loss: 0.62794345, Time: 0.0207 Steps: 100770, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000717, Sample Num: 11472, Cur Loss: 0.38825971, Cur Avg Loss: 0.13695900, Log Avg loss: 0.13697603, Global Avg Loss: 0.62789474, Time: 0.0207 Steps: 100780, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000727, Sample Num: 11632, Cur Loss: 0.14356880, Cur Avg Loss: 0.13685034, Log Avg loss: 0.12905922, Global Avg Loss: 0.62784524, Time: 0.0208 Steps: 100790, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000737, Sample Num: 11792, Cur Loss: 0.09051713, Cur Avg Loss: 0.13775808, Log Avg loss: 0.20375064, Global Avg Loss: 0.62780317, Time: 0.0207 Steps: 100800, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000747, Sample Num: 11952, Cur Loss: 0.23012592, Cur Avg Loss: 0.13799162, Log Avg loss: 0.15520368, Global Avg Loss: 0.62775629, Time: 0.0208 Steps: 100810, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000757, Sample Num: 12112, Cur Loss: 0.14911546, Cur Avg Loss: 0.13826899, Log Avg loss: 0.15898833, Global Avg Loss: 0.62770979, Time: 0.0208 Steps: 100820, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000767, Sample Num: 12272, Cur Loss: 0.11428040, Cur Avg Loss: 0.13819673, Log Avg loss: 0.13272671, Global Avg Loss: 0.62766070, Time: 0.0207 Steps: 100830, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000777, Sample Num: 12432, Cur Loss: 0.12240874, Cur Avg Loss: 0.13896797, Log Avg loss: 0.19812262, Global Avg Loss: 0.62761811, Time: 0.0209 Steps: 100840, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000787, Sample Num: 12592, Cur Loss: 0.16796249, Cur Avg Loss: 0.13917517, Log Avg loss: 0.15527444, Global Avg Loss: 0.62757127, Time: 0.0209 Steps: 100850, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000797, Sample Num: 12752, Cur Loss: 0.15993479, Cur Avg Loss: 0.13924344, Log Avg loss: 0.14461627, Global Avg Loss: 0.62752339, Time: 0.0209 Steps: 100860, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000807, Sample Num: 12912, Cur Loss: 0.09381503, Cur Avg Loss: 0.13913770, Log Avg loss: 0.13071022, Global Avg Loss: 0.62747414, Time: 0.0209 Steps: 100870, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000817, Sample Num: 13072, Cur Loss: 0.22370312, Cur Avg Loss: 0.13950872, Log Avg loss: 0.16945003, Global Avg Loss: 0.62742873, Time: 0.0209 Steps: 100880, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000827, Sample Num: 13232, Cur Loss: 0.12755677, Cur Avg Loss: 0.13907166, Log Avg loss: 0.10336363, Global Avg Loss: 0.62737679, Time: 0.0209 Steps: 100890, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000837, Sample Num: 13392, Cur Loss: 0.11749534, Cur Avg Loss: 0.13908081, Log Avg loss: 0.13983805, Global Avg Loss: 0.62732847, Time: 0.0209 Steps: 100900, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000847, Sample Num: 13552, Cur Loss: 0.14645389, Cur Avg Loss: 0.13895389, Log Avg loss: 0.12833049, Global Avg Loss: 0.62727902, Time: 0.0209 Steps: 100910, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000857, Sample Num: 13712, Cur Loss: 0.12694266, Cur Avg Loss: 0.13881772, Log Avg loss: 0.12728368, Global Avg Loss: 0.62722948, Time: 0.0209 Steps: 100920, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000867, Sample Num: 13872, Cur Loss: 0.21349593, Cur Avg Loss: 0.13897195, Log Avg loss: 0.15218998, Global Avg Loss: 0.62718241, Time: 0.0209 Steps: 100930, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000877, Sample Num: 14032, Cur Loss: 0.16170406, Cur Avg Loss: 0.13882344, Log Avg loss: 0.12594758, Global Avg Loss: 0.62713275, Time: 0.0209 Steps: 100940, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000887, Sample Num: 14192, Cur Loss: 0.38195187, Cur Avg Loss: 0.13886924, Log Avg loss: 0.14288603, Global Avg Loss: 0.62708478, Time: 0.0208 Steps: 100950, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000897, Sample Num: 14352, Cur Loss: 0.24160704, Cur Avg Loss: 0.13924110, Log Avg loss: 0.17222463, Global Avg Loss: 0.62703973, Time: 0.0209 Steps: 100960, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000907, Sample Num: 14512, Cur Loss: 0.20623966, Cur Avg Loss: 0.13893527, Log Avg loss: 0.11150284, Global Avg Loss: 0.62698867, Time: 0.0209 Steps: 100970, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000917, Sample Num: 14672, Cur Loss: 0.03080602, Cur Avg Loss: 0.13870928, Log Avg loss: 0.11821172, Global Avg Loss: 0.62693829, Time: 0.0209 Steps: 100980, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000927, Sample Num: 14832, Cur Loss: 0.08774859, Cur Avg Loss: 0.13881721, Log Avg loss: 0.14871443, Global Avg Loss: 0.62689093, Time: 0.0209 Steps: 100990, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000937, Sample Num: 14992, Cur Loss: 0.17663318, Cur Avg Loss: 0.13871888, Log Avg loss: 0.12960385, Global Avg Loss: 0.62684170, Time: 0.0209 Steps: 101000, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000947, Sample Num: 15152, Cur Loss: 0.03144014, Cur Avg Loss: 0.13857690, Log Avg loss: 0.12527321, Global Avg Loss: 0.62679204, Time: 0.0209 Steps: 101010, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000957, Sample Num: 15312, Cur Loss: 0.14695545, Cur Avg Loss: 0.13891617, Log Avg loss: 0.17104455, Global Avg Loss: 0.62674693, Time: 0.0209 Steps: 101020, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000967, Sample Num: 15472, Cur Loss: 0.05044879, Cur Avg Loss: 0.13898749, Log Avg loss: 0.14581279, Global Avg Loss: 0.62669932, Time: 0.0209 Steps: 101030, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000977, Sample Num: 15632, Cur Loss: 0.07611568, Cur Avg Loss: 0.13919562, Log Avg loss: 0.15932212, Global Avg Loss: 0.62665307, Time: 0.0209 Steps: 101040, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000987, Sample Num: 15792, Cur Loss: 0.14694494, Cur Avg Loss: 0.13949473, Log Avg loss: 0.16871759, Global Avg Loss: 0.62660775, Time: 0.0209 Steps: 101050, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000997, Sample Num: 15952, Cur Loss: 0.03740328, Cur Avg Loss: 0.13941324, Log Avg loss: 0.13137003, Global Avg Loss: 0.62655875, Time: 0.0209 Steps: 101060, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001007, Sample Num: 16112, Cur Loss: 0.07906294, Cur Avg Loss: 0.13946730, Log Avg loss: 0.14485728, Global Avg Loss: 0.62651109, Time: 0.0209 Steps: 101070, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001017, Sample Num: 16272, Cur Loss: 0.13859487, Cur Avg Loss: 0.13962184, Log Avg loss: 0.15518408, Global Avg Loss: 0.62646446, Time: 0.0209 Steps: 101080, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001027, Sample Num: 16432, Cur Loss: 0.12014530, Cur Avg Loss: 0.13948331, Log Avg loss: 0.12539458, Global Avg Loss: 0.62641489, Time: 0.0246 Steps: 101090, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001037, Sample Num: 16592, Cur Loss: 0.11421229, Cur Avg Loss: 0.13972702, Log Avg loss: 0.16475636, Global Avg Loss: 0.62636923, Time: 0.0209 Steps: 101100, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001047, Sample Num: 16752, Cur Loss: 0.04867056, Cur Avg Loss: 0.13971393, Log Avg loss: 0.13835676, Global Avg Loss: 0.62632096, Time: 0.0209 Steps: 101110, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001057, Sample Num: 16912, Cur Loss: 0.07325494, Cur Avg Loss: 0.14046362, Log Avg loss: 0.21895610, Global Avg Loss: 0.62628068, Time: 0.0209 Steps: 101120, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001067, Sample Num: 17072, Cur Loss: 0.12482861, Cur Avg Loss: 0.14136267, Log Avg loss: 0.23639188, Global Avg Loss: 0.62624212, Time: 0.0209 Steps: 101130, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001077, Sample Num: 17232, Cur Loss: 0.13245526, Cur Avg Loss: 0.14127047, Log Avg loss: 0.13143273, Global Avg Loss: 0.62619320, Time: 0.0209 Steps: 101140, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001087, Sample Num: 17392, Cur Loss: 0.04755440, Cur Avg Loss: 0.14159185, Log Avg loss: 0.17620475, Global Avg Loss: 0.62614871, Time: 0.0209 Steps: 101150, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001097, Sample Num: 17552, Cur Loss: 0.35624969, Cur Avg Loss: 0.14174540, Log Avg loss: 0.15843607, Global Avg Loss: 0.62610248, Time: 0.0209 Steps: 101160, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001107, Sample Num: 17712, Cur Loss: 0.07810764, Cur Avg Loss: 0.14170130, Log Avg loss: 0.13686418, Global Avg Loss: 0.62605412, Time: 0.0209 Steps: 101170, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001117, Sample Num: 17872, Cur Loss: 0.08276193, Cur Avg Loss: 0.14220296, Log Avg loss: 0.19773675, Global Avg Loss: 0.62601179, Time: 0.0209 Steps: 101180, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001127, Sample Num: 18032, Cur Loss: 0.09761883, Cur Avg Loss: 0.14206546, Log Avg loss: 0.12670564, Global Avg Loss: 0.62596244, Time: 0.0209 Steps: 101190, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001137, Sample Num: 18192, Cur Loss: 0.17578411, Cur Avg Loss: 0.14217195, Log Avg loss: 0.15417373, Global Avg Loss: 0.62591582, Time: 0.0209 Steps: 101200, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001147, Sample Num: 18352, Cur Loss: 0.17410845, Cur Avg Loss: 0.14201447, Log Avg loss: 0.12410954, Global Avg Loss: 0.62586624, Time: 0.0209 Steps: 101210, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001157, Sample Num: 18512, Cur Loss: 0.14297828, Cur Avg Loss: 0.14167525, Log Avg loss: 0.10276609, Global Avg Loss: 0.62581456, Time: 0.0209 Steps: 101220, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001167, Sample Num: 18672, Cur Loss: 0.18835533, Cur Avg Loss: 0.14187485, Log Avg loss: 0.16496854, Global Avg Loss: 0.62576904, Time: 0.0209 Steps: 101230, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001177, Sample Num: 18832, Cur Loss: 0.03897062, Cur Avg Loss: 0.14200484, Log Avg loss: 0.15717487, Global Avg Loss: 0.62572275, Time: 0.0209 Steps: 101240, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001187, Sample Num: 18992, Cur Loss: 0.03114895, Cur Avg Loss: 0.14172370, Log Avg loss: 0.10863330, Global Avg Loss: 0.62567168, Time: 0.0209 Steps: 101250, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001197, Sample Num: 19152, Cur Loss: 0.06665674, Cur Avg Loss: 0.14187707, Log Avg loss: 0.16008196, Global Avg Loss: 0.62562570, Time: 0.0209 Steps: 101260, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001207, Sample Num: 19312, Cur Loss: 0.34300244, Cur Avg Loss: 0.14175842, Log Avg loss: 0.12755665, Global Avg Loss: 0.62557652, Time: 0.0209 Steps: 101270, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001217, Sample Num: 19472, Cur Loss: 0.12263246, Cur Avg Loss: 0.14180765, Log Avg loss: 0.14774922, Global Avg Loss: 0.62552934, Time: 0.0209 Steps: 101280, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001227, Sample Num: 19632, Cur Loss: 0.19758803, Cur Avg Loss: 0.14201186, Log Avg loss: 0.16686442, Global Avg Loss: 0.62548406, Time: 0.0209 Steps: 101290, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001237, Sample Num: 19792, Cur Loss: 0.06418324, Cur Avg Loss: 0.14194644, Log Avg loss: 0.13391915, Global Avg Loss: 0.62543553, Time: 0.0209 Steps: 101300, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001247, Sample Num: 19952, Cur Loss: 0.05526295, Cur Avg Loss: 0.14180987, Log Avg loss: 0.12491702, Global Avg Loss: 0.62538613, Time: 0.0209 Steps: 101310, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001257, Sample Num: 20112, Cur Loss: 0.10549983, Cur Avg Loss: 0.14186159, Log Avg loss: 0.14831017, Global Avg Loss: 0.62533904, Time: 0.0209 Steps: 101320, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001267, Sample Num: 20272, Cur Loss: 0.07662624, Cur Avg Loss: 0.14181065, Log Avg loss: 0.13540838, Global Avg Loss: 0.62529069, Time: 0.0209 Steps: 101330, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001277, Sample Num: 20432, Cur Loss: 0.34965384, Cur Avg Loss: 0.14151498, Log Avg loss: 0.10405360, Global Avg Loss: 0.62523926, Time: 0.0210 Steps: 101340, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001287, Sample Num: 20592, Cur Loss: 0.06543072, Cur Avg Loss: 0.14130418, Log Avg loss: 0.11438401, Global Avg Loss: 0.62518885, Time: 0.0209 Steps: 101350, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001297, Sample Num: 20752, Cur Loss: 0.14001457, Cur Avg Loss: 0.14141183, Log Avg loss: 0.15526646, Global Avg Loss: 0.62514249, Time: 0.0209 Steps: 101360, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001307, Sample Num: 20912, Cur Loss: 0.07009545, Cur Avg Loss: 0.14192184, Log Avg loss: 0.20807006, Global Avg Loss: 0.62510135, Time: 0.0209 Steps: 101370, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001317, Sample Num: 21072, Cur Loss: 0.25635108, Cur Avg Loss: 0.14209268, Log Avg loss: 0.16442188, Global Avg Loss: 0.62505591, Time: 0.0209 Steps: 101380, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001327, Sample Num: 21232, Cur Loss: 0.16155151, Cur Avg Loss: 0.14215777, Log Avg loss: 0.15073018, Global Avg Loss: 0.62500913, Time: 0.0209 Steps: 101390, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001337, Sample Num: 21392, Cur Loss: 0.07768840, Cur Avg Loss: 0.14186196, Log Avg loss: 0.10260861, Global Avg Loss: 0.62495761, Time: 0.0209 Steps: 101400, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001347, Sample Num: 21552, Cur Loss: 0.19632968, Cur Avg Loss: 0.14195221, Log Avg loss: 0.15401823, Global Avg Loss: 0.62491117, Time: 0.0209 Steps: 101410, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001357, Sample Num: 21712, Cur Loss: 0.04731952, Cur Avg Loss: 0.14151767, Log Avg loss: 0.08298501, Global Avg Loss: 0.62485773, Time: 0.0209 Steps: 101420, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001367, Sample Num: 21872, Cur Loss: 0.12408398, Cur Avg Loss: 0.14162731, Log Avg loss: 0.15650507, Global Avg Loss: 0.62481156, Time: 0.0209 Steps: 101430, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001377, Sample Num: 22032, Cur Loss: 0.14023441, Cur Avg Loss: 0.14181497, Log Avg loss: 0.16746859, Global Avg Loss: 0.62476647, Time: 0.0210 Steps: 101440, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001387, Sample Num: 22192, Cur Loss: 0.08790144, Cur Avg Loss: 0.14204792, Log Avg loss: 0.17412471, Global Avg Loss: 0.62472205, Time: 0.0210 Steps: 101450, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001397, Sample Num: 22352, Cur Loss: 0.09802860, Cur Avg Loss: 0.14195067, Log Avg loss: 0.12846290, Global Avg Loss: 0.62467314, Time: 0.0209 Steps: 101460, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001407, Sample Num: 22512, Cur Loss: 0.09578148, Cur Avg Loss: 0.14186700, Log Avg loss: 0.13017816, Global Avg Loss: 0.62462441, Time: 0.0209 Steps: 101470, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001417, Sample Num: 22672, Cur Loss: 0.36866909, Cur Avg Loss: 0.14198382, Log Avg loss: 0.15842048, Global Avg Loss: 0.62457847, Time: 0.0209 Steps: 101480, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001427, Sample Num: 22832, Cur Loss: 0.09178860, Cur Avg Loss: 0.14198763, Log Avg loss: 0.14252674, Global Avg Loss: 0.62453097, Time: 0.0211 Steps: 101490, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001437, Sample Num: 22992, Cur Loss: 0.34122351, Cur Avg Loss: 0.14194724, Log Avg loss: 0.13618375, Global Avg Loss: 0.62448286, Time: 0.0209 Steps: 101500, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001447, Sample Num: 23152, Cur Loss: 0.19933420, Cur Avg Loss: 0.14198637, Log Avg loss: 0.14760982, Global Avg Loss: 0.62443588, Time: 0.0209 Steps: 101510, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001457, Sample Num: 23312, Cur Loss: 0.13469653, Cur Avg Loss: 0.14222492, Log Avg loss: 0.17674326, Global Avg Loss: 0.62439178, Time: 0.0209 Steps: 101520, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001467, Sample Num: 23472, Cur Loss: 0.46081221, Cur Avg Loss: 0.14241741, Log Avg loss: 0.17046235, Global Avg Loss: 0.62434707, Time: 0.0209 Steps: 101530, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001477, Sample Num: 23632, Cur Loss: 0.11249278, Cur Avg Loss: 0.14230098, Log Avg loss: 0.12522149, Global Avg Loss: 0.62429792, Time: 0.0209 Steps: 101540, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001487, Sample Num: 23792, Cur Loss: 0.04896035, Cur Avg Loss: 0.14223875, Log Avg loss: 0.13304621, Global Avg Loss: 0.62424954, Time: 0.0209 Steps: 101550, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001497, Sample Num: 23952, Cur Loss: 0.13993846, Cur Avg Loss: 0.14211497, Log Avg loss: 0.12370972, Global Avg Loss: 0.62420026, Time: 0.0210 Steps: 101560, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001507, Sample Num: 24112, Cur Loss: 0.08579202, Cur Avg Loss: 0.14210980, Log Avg loss: 0.14133547, Global Avg Loss: 0.62415272, Time: 0.0209 Steps: 101570, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001517, Sample Num: 24272, Cur Loss: 0.12329190, Cur Avg Loss: 0.14202048, Log Avg loss: 0.12856025, Global Avg Loss: 0.62410393, Time: 0.0209 Steps: 101580, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001527, Sample Num: 24432, Cur Loss: 0.06186543, Cur Avg Loss: 0.14188821, Log Avg loss: 0.12182238, Global Avg Loss: 0.62405449, Time: 0.0210 Steps: 101590, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001537, Sample Num: 24592, Cur Loss: 0.12909000, Cur Avg Loss: 0.14244963, Log Avg loss: 0.22817858, Global Avg Loss: 0.62401552, Time: 0.0245 Steps: 101600, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001547, Sample Num: 24752, Cur Loss: 0.08972412, Cur Avg Loss: 0.14227273, Log Avg loss: 0.11508324, Global Avg Loss: 0.62396543, Time: 0.0208 Steps: 101610, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001557, Sample Num: 24912, Cur Loss: 0.14374205, Cur Avg Loss: 0.14229963, Log Avg loss: 0.14646123, Global Avg Loss: 0.62391845, Time: 0.0208 Steps: 101620, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001567, Sample Num: 25072, Cur Loss: 0.07895815, Cur Avg Loss: 0.14190224, Log Avg loss: 0.08002822, Global Avg Loss: 0.62386493, Time: 0.0209 Steps: 101630, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001577, Sample Num: 25232, Cur Loss: 0.13075788, Cur Avg Loss: 0.14207657, Log Avg loss: 0.16939448, Global Avg Loss: 0.62382021, Time: 0.0208 Steps: 101640, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001587, Sample Num: 25392, Cur Loss: 0.16124672, Cur Avg Loss: 0.14235149, Log Avg loss: 0.18570689, Global Avg Loss: 0.62377711, Time: 0.0208 Steps: 101650, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001597, Sample Num: 25552, Cur Loss: 0.22020531, Cur Avg Loss: 0.14237369, Log Avg loss: 0.14589638, Global Avg Loss: 0.62373011, Time: 0.0208 Steps: 101660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001607, Sample Num: 25712, Cur Loss: 0.23326811, Cur Avg Loss: 0.14257794, Log Avg loss: 0.17519704, Global Avg Loss: 0.62368599, Time: 0.0208 Steps: 101670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001617, Sample Num: 25872, Cur Loss: 0.14696653, Cur Avg Loss: 0.14242592, Log Avg loss: 0.11799576, Global Avg Loss: 0.62363626, Time: 0.0208 Steps: 101680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001627, Sample Num: 26032, Cur Loss: 0.09212666, Cur Avg Loss: 0.14268425, Log Avg loss: 0.18445587, Global Avg Loss: 0.62359307, Time: 0.0209 Steps: 101690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001637, Sample Num: 26192, Cur Loss: 0.15795428, Cur Avg Loss: 0.14267949, Log Avg loss: 0.14190620, Global Avg Loss: 0.62354571, Time: 0.0208 Steps: 101700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001647, Sample Num: 26352, Cur Loss: 0.07369730, Cur Avg Loss: 0.14245849, Log Avg loss: 0.10628010, Global Avg Loss: 0.62349485, Time: 0.0208 Steps: 101710, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001657, Sample Num: 26512, Cur Loss: 0.08807401, Cur Avg Loss: 0.14236221, Log Avg loss: 0.12650461, Global Avg Loss: 0.62344599, Time: 0.0209 Steps: 101720, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001667, Sample Num: 26672, Cur Loss: 0.12103502, Cur Avg Loss: 0.14243278, Log Avg loss: 0.15412711, Global Avg Loss: 0.62339986, Time: 0.0208 Steps: 101730, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001677, Sample Num: 26832, Cur Loss: 0.23160860, Cur Avg Loss: 0.14234663, Log Avg loss: 0.12798562, Global Avg Loss: 0.62335116, Time: 0.0208 Steps: 101740, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001687, Sample Num: 26992, Cur Loss: 0.06497686, Cur Avg Loss: 0.14198545, Log Avg loss: 0.08141563, Global Avg Loss: 0.62329790, Time: 0.0208 Steps: 101750, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001697, Sample Num: 27152, Cur Loss: 0.08563313, Cur Avg Loss: 0.14195821, Log Avg loss: 0.13736146, Global Avg Loss: 0.62325015, Time: 0.0208 Steps: 101760, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001707, Sample Num: 27312, Cur Loss: 0.20235872, Cur Avg Loss: 0.14189403, Log Avg loss: 0.13100369, Global Avg Loss: 0.62320178, Time: 0.0208 Steps: 101770, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001717, Sample Num: 27472, Cur Loss: 0.07159777, Cur Avg Loss: 0.14219347, Log Avg loss: 0.19330660, Global Avg Loss: 0.62315954, Time: 0.0208 Steps: 101780, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001727, Sample Num: 27632, Cur Loss: 0.51383173, Cur Avg Loss: 0.14232845, Log Avg loss: 0.16550612, Global Avg Loss: 0.62311458, Time: 0.0209 Steps: 101790, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001737, Sample Num: 27792, Cur Loss: 0.16497040, Cur Avg Loss: 0.14227474, Log Avg loss: 0.13299817, Global Avg Loss: 0.62306644, Time: 0.0208 Steps: 101800, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001747, Sample Num: 27952, Cur Loss: 0.06121366, Cur Avg Loss: 0.14210624, Log Avg loss: 0.11283774, Global Avg Loss: 0.62301632, Time: 0.0208 Steps: 101810, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001757, Sample Num: 28112, Cur Loss: 0.26057243, Cur Avg Loss: 0.14216282, Log Avg loss: 0.15204759, Global Avg Loss: 0.62297006, Time: 0.0208 Steps: 101820, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001767, Sample Num: 28272, Cur Loss: 0.08299267, Cur Avg Loss: 0.14189248, Log Avg loss: 0.09439369, Global Avg Loss: 0.62291816, Time: 0.0208 Steps: 101830, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001777, Sample Num: 28432, Cur Loss: 0.30112016, Cur Avg Loss: 0.14199900, Log Avg loss: 0.16082109, Global Avg Loss: 0.62287278, Time: 0.0208 Steps: 101840, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001787, Sample Num: 28592, Cur Loss: 0.21319288, Cur Avg Loss: 0.14211566, Log Avg loss: 0.16284601, Global Avg Loss: 0.62282761, Time: 0.0208 Steps: 101850, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001797, Sample Num: 28752, Cur Loss: 0.04177443, Cur Avg Loss: 0.14253676, Log Avg loss: 0.21778802, Global Avg Loss: 0.62278785, Time: 0.0235 Steps: 101860, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001807, Sample Num: 28912, Cur Loss: 0.28712869, Cur Avg Loss: 0.14262953, Log Avg loss: 0.15929955, Global Avg Loss: 0.62274235, Time: 0.0208 Steps: 101870, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001817, Sample Num: 29072, Cur Loss: 0.06465843, Cur Avg Loss: 0.14287979, Log Avg loss: 0.18810201, Global Avg Loss: 0.62269969, Time: 0.0208 Steps: 101880, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001827, Sample Num: 29232, Cur Loss: 0.27938923, Cur Avg Loss: 0.14307866, Log Avg loss: 0.17921355, Global Avg Loss: 0.62265616, Time: 0.0208 Steps: 101890, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001837, Sample Num: 29392, Cur Loss: 0.29166323, Cur Avg Loss: 0.14324354, Log Avg loss: 0.17336747, Global Avg Loss: 0.62261207, Time: 0.0208 Steps: 101900, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001847, Sample Num: 29552, Cur Loss: 0.48353243, Cur Avg Loss: 0.14327259, Log Avg loss: 0.14860873, Global Avg Loss: 0.62256556, Time: 0.0208 Steps: 101910, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001857, Sample Num: 29712, Cur Loss: 0.27255583, Cur Avg Loss: 0.14340286, Log Avg loss: 0.16746302, Global Avg Loss: 0.62252091, Time: 0.0208 Steps: 101920, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001867, Sample Num: 29872, Cur Loss: 0.17421407, Cur Avg Loss: 0.14339275, Log Avg loss: 0.14151617, Global Avg Loss: 0.62247372, Time: 0.0208 Steps: 101930, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001877, Sample Num: 30032, Cur Loss: 0.12970927, Cur Avg Loss: 0.14343384, Log Avg loss: 0.15110438, Global Avg Loss: 0.62242748, Time: 0.0208 Steps: 101940, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001887, Sample Num: 30192, Cur Loss: 0.20038903, Cur Avg Loss: 0.14366211, Log Avg loss: 0.18650876, Global Avg Loss: 0.62238472, Time: 0.0208 Steps: 101950, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001897, Sample Num: 30352, Cur Loss: 0.05035543, Cur Avg Loss: 0.14375240, Log Avg loss: 0.16079075, Global Avg Loss: 0.62233945, Time: 0.0209 Steps: 101960, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001907, Sample Num: 30512, Cur Loss: 0.08272222, Cur Avg Loss: 0.14379520, Log Avg loss: 0.15191310, Global Avg Loss: 0.62229331, Time: 0.0208 Steps: 101970, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001917, Sample Num: 30672, Cur Loss: 0.13434149, Cur Avg Loss: 0.14368112, Log Avg loss: 0.12192771, Global Avg Loss: 0.62224425, Time: 0.0208 Steps: 101980, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001927, Sample Num: 30832, Cur Loss: 0.07823162, Cur Avg Loss: 0.14363533, Log Avg loss: 0.13485581, Global Avg Loss: 0.62219646, Time: 0.0208 Steps: 101990, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001937, Sample Num: 30992, Cur Loss: 0.09252872, Cur Avg Loss: 0.14350493, Log Avg loss: 0.11837843, Global Avg Loss: 0.62214707, Time: 0.0208 Steps: 102000, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001947, Sample Num: 31152, Cur Loss: 0.49140036, Cur Avg Loss: 0.14382387, Log Avg loss: 0.20560103, Global Avg Loss: 0.62210623, Time: 0.0208 Steps: 102010, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001957, Sample Num: 31312, Cur Loss: 0.16089348, Cur Avg Loss: 0.14379438, Log Avg loss: 0.13805260, Global Avg Loss: 0.62205879, Time: 0.0208 Steps: 102020, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001967, Sample Num: 31472, Cur Loss: 0.04636857, Cur Avg Loss: 0.14368632, Log Avg loss: 0.12253989, Global Avg Loss: 0.62200983, Time: 0.0208 Steps: 102030, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001977, Sample Num: 31632, Cur Loss: 0.03230120, Cur Avg Loss: 0.14354682, Log Avg loss: 0.11610729, Global Avg Loss: 0.62196025, Time: 0.0209 Steps: 102040, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001987, Sample Num: 31792, Cur Loss: 0.17363080, Cur Avg Loss: 0.14369195, Log Avg loss: 0.17238432, Global Avg Loss: 0.62191620, Time: 0.0208 Steps: 102050, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001997, Sample Num: 31952, Cur Loss: 0.12026945, Cur Avg Loss: 0.14372688, Log Avg loss: 0.15066707, Global Avg Loss: 0.62187002, Time: 0.0208 Steps: 102060, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002007, Sample Num: 32112, Cur Loss: 0.28804544, Cur Avg Loss: 0.14398863, Log Avg loss: 0.19626010, Global Avg Loss: 0.62182832, Time: 0.0208 Steps: 102070, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002017, Sample Num: 32272, Cur Loss: 0.06156124, Cur Avg Loss: 0.14407261, Log Avg loss: 0.16092709, Global Avg Loss: 0.62178317, Time: 0.0208 Steps: 102080, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002027, Sample Num: 32432, Cur Loss: 0.17176139, Cur Avg Loss: 0.14399124, Log Avg loss: 0.12757987, Global Avg Loss: 0.62173476, Time: 0.0208 Steps: 102090, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002037, Sample Num: 32592, Cur Loss: 0.25498733, Cur Avg Loss: 0.14424274, Log Avg loss: 0.19522168, Global Avg Loss: 0.62169299, Time: 0.0209 Steps: 102100, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002047, Sample Num: 32752, Cur Loss: 0.09248361, Cur Avg Loss: 0.14414967, Log Avg loss: 0.12519083, Global Avg Loss: 0.62164437, Time: 0.0208 Steps: 102110, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002057, Sample Num: 32912, Cur Loss: 0.09503048, Cur Avg Loss: 0.14405390, Log Avg loss: 0.12444952, Global Avg Loss: 0.62159568, Time: 0.0209 Steps: 102120, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002067, Sample Num: 33072, Cur Loss: 0.33774033, Cur Avg Loss: 0.14405237, Log Avg loss: 0.14373696, Global Avg Loss: 0.62154889, Time: 0.0208 Steps: 102130, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002077, Sample Num: 33232, Cur Loss: 0.03285367, Cur Avg Loss: 0.14381043, Log Avg loss: 0.09380318, Global Avg Loss: 0.62149722, Time: 0.0209 Steps: 102140, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002087, Sample Num: 33392, Cur Loss: 0.04705343, Cur Avg Loss: 0.14371402, Log Avg loss: 0.12368841, Global Avg Loss: 0.62144849, Time: 0.0208 Steps: 102150, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002097, Sample Num: 33552, Cur Loss: 0.67492700, Cur Avg Loss: 0.14421337, Log Avg loss: 0.24842713, Global Avg Loss: 0.62141197, Time: 0.0209 Steps: 102160, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002107, Sample Num: 33712, Cur Loss: 0.13854612, Cur Avg Loss: 0.14403744, Log Avg loss: 0.10714598, Global Avg Loss: 0.62136164, Time: 0.0208 Steps: 102170, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002117, Sample Num: 33872, Cur Loss: 0.11034897, Cur Avg Loss: 0.14391707, Log Avg loss: 0.11855511, Global Avg Loss: 0.62131243, Time: 0.0208 Steps: 102180, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002127, Sample Num: 34032, Cur Loss: 0.06429136, Cur Avg Loss: 0.14373598, Log Avg loss: 0.10539982, Global Avg Loss: 0.62126195, Time: 0.0209 Steps: 102190, Updated lr: 0.000004 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.557715, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.238188, "eval_total_loss": 167.446083, "eval_mae": 0.323544, "eval_mse": 0.238279, "eval_r2": 0.848534, "eval_sp_statistic": 0.905492, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925398, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.730546, "test_total_loss": 366.733903, "test_mae": 0.661817, "test_mse": 0.730596, "test_r2": 0.528467, "test_sp_statistic": 0.800977, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.840501, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6212510265049528, "train_cur_epoch_loss": 305.85301070380956, "train_cur_epoch_avg_loss": 0.14366040897313742, "train_cur_epoch_time": 44.55771470069885, "train_cur_epoch_avg_time": 0.02092894067670214, "epoch": 48, "step": 102192} ################################################## Training, Epoch: 0049, Batch: 000008, Sample Num: 128, Cur Loss: 0.16148487, Cur Avg Loss: 0.14303475, Log Avg loss: 0.12708511, Global Avg Loss: 0.62121359, Time: 0.0212 Steps: 102200, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000018, Sample Num: 288, Cur Loss: 0.11712126, Cur Avg Loss: 0.13225466, Log Avg loss: 0.12363060, Global Avg Loss: 0.62116491, Time: 0.0209 Steps: 102210, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000028, Sample Num: 448, Cur Loss: 0.12431516, Cur Avg Loss: 0.13081342, Log Avg loss: 0.12821917, Global Avg Loss: 0.62111669, Time: 0.0209 Steps: 102220, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000038, Sample Num: 608, Cur Loss: 0.04717274, Cur Avg Loss: 0.11654150, Log Avg loss: 0.07658015, Global Avg Loss: 0.62106342, Time: 0.0209 Steps: 102230, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000048, Sample Num: 768, Cur Loss: 0.06871797, Cur Avg Loss: 0.11913651, Log Avg loss: 0.12899754, Global Avg Loss: 0.62101529, Time: 0.0209 Steps: 102240, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000058, Sample Num: 928, Cur Loss: 0.02125281, Cur Avg Loss: 0.12324787, Log Avg loss: 0.14298242, Global Avg Loss: 0.62096854, Time: 0.0209 Steps: 102250, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000068, Sample Num: 1088, Cur Loss: 0.16142875, Cur Avg Loss: 0.12627073, Log Avg loss: 0.14380326, Global Avg Loss: 0.62092188, Time: 0.0208 Steps: 102260, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000078, Sample Num: 1248, Cur Loss: 0.21818845, Cur Avg Loss: 0.12355063, Log Avg loss: 0.10505394, Global Avg Loss: 0.62087144, Time: 0.0209 Steps: 102270, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000088, Sample Num: 1408, Cur Loss: 0.10119258, Cur Avg Loss: 0.12198348, Log Avg loss: 0.10975979, Global Avg Loss: 0.62082147, Time: 0.0209 Steps: 102280, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000098, Sample Num: 1568, Cur Loss: 0.07207448, Cur Avg Loss: 0.12330603, Log Avg loss: 0.13494446, Global Avg Loss: 0.62077397, Time: 0.0209 Steps: 102290, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000108, Sample Num: 1728, Cur Loss: 0.20554397, Cur Avg Loss: 0.12565798, Log Avg loss: 0.14870706, Global Avg Loss: 0.62072782, Time: 0.0211 Steps: 102300, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000118, Sample Num: 1888, Cur Loss: 0.10785878, Cur Avg Loss: 0.12697830, Log Avg loss: 0.14123777, Global Avg Loss: 0.62068095, Time: 0.0209 Steps: 102310, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000128, Sample Num: 2048, Cur Loss: 0.06598017, Cur Avg Loss: 0.12642519, Log Avg loss: 0.11989847, Global Avg Loss: 0.62063201, Time: 0.0209 Steps: 102320, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000138, Sample Num: 2208, Cur Loss: 0.05781755, Cur Avg Loss: 0.12584418, Log Avg loss: 0.11840729, Global Avg Loss: 0.62058293, Time: 0.0209 Steps: 102330, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000148, Sample Num: 2368, Cur Loss: 0.37288415, Cur Avg Loss: 0.12903618, Log Avg loss: 0.17308580, Global Avg Loss: 0.62053921, Time: 0.0209 Steps: 102340, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000158, Sample Num: 2528, Cur Loss: 0.07704979, Cur Avg Loss: 0.12970091, Log Avg loss: 0.13953890, Global Avg Loss: 0.62049221, Time: 0.0209 Steps: 102350, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000168, Sample Num: 2688, Cur Loss: 0.05048837, Cur Avg Loss: 0.13339120, Log Avg loss: 0.19169776, Global Avg Loss: 0.62045032, Time: 0.0209 Steps: 102360, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000178, Sample Num: 2848, Cur Loss: 0.26654327, Cur Avg Loss: 0.13398051, Log Avg loss: 0.14388094, Global Avg Loss: 0.62040377, Time: 0.0209 Steps: 102370, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000188, Sample Num: 3008, Cur Loss: 0.37414211, Cur Avg Loss: 0.13692517, Log Avg loss: 0.18934008, Global Avg Loss: 0.62036166, Time: 0.0208 Steps: 102380, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000198, Sample Num: 3168, Cur Loss: 0.05857112, Cur Avg Loss: 0.13605501, Log Avg loss: 0.11969595, Global Avg Loss: 0.62031276, Time: 0.0208 Steps: 102390, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000208, Sample Num: 3328, Cur Loss: 0.20981042, Cur Avg Loss: 0.13554564, Log Avg loss: 0.12546008, Global Avg Loss: 0.62026444, Time: 0.0209 Steps: 102400, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000218, Sample Num: 3488, Cur Loss: 0.24173328, Cur Avg Loss: 0.13474055, Log Avg loss: 0.11799474, Global Avg Loss: 0.62021539, Time: 0.0209 Steps: 102410, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000228, Sample Num: 3648, Cur Loss: 0.29989773, Cur Avg Loss: 0.13545625, Log Avg loss: 0.15105854, Global Avg Loss: 0.62016959, Time: 0.0208 Steps: 102420, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000238, Sample Num: 3808, Cur Loss: 0.11167665, Cur Avg Loss: 0.13539921, Log Avg loss: 0.13409869, Global Avg Loss: 0.62012213, Time: 0.0209 Steps: 102430, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000248, Sample Num: 3968, Cur Loss: 0.16239716, Cur Avg Loss: 0.13583027, Log Avg loss: 0.14608959, Global Avg Loss: 0.62007586, Time: 0.0209 Steps: 102440, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000258, Sample Num: 4128, Cur Loss: 0.13643119, Cur Avg Loss: 0.13520091, Log Avg loss: 0.11959263, Global Avg Loss: 0.62002701, Time: 0.0248 Steps: 102450, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000268, Sample Num: 4288, Cur Loss: 0.20300436, Cur Avg Loss: 0.13450801, Log Avg loss: 0.11663117, Global Avg Loss: 0.61997787, Time: 0.0210 Steps: 102460, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000278, Sample Num: 4448, Cur Loss: 0.15859020, Cur Avg Loss: 0.13487895, Log Avg loss: 0.14482031, Global Avg Loss: 0.61993150, Time: 0.0210 Steps: 102470, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000288, Sample Num: 4608, Cur Loss: 0.18201134, Cur Avg Loss: 0.13401570, Log Avg loss: 0.11001737, Global Avg Loss: 0.61988175, Time: 0.0210 Steps: 102480, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000298, Sample Num: 4768, Cur Loss: 0.23830929, Cur Avg Loss: 0.13353750, Log Avg loss: 0.11976518, Global Avg Loss: 0.61983295, Time: 0.0209 Steps: 102490, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000308, Sample Num: 4928, Cur Loss: 0.16306700, Cur Avg Loss: 0.13385590, Log Avg loss: 0.14334423, Global Avg Loss: 0.61978646, Time: 0.0210 Steps: 102500, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000318, Sample Num: 5088, Cur Loss: 0.06947801, Cur Avg Loss: 0.13279140, Log Avg loss: 0.10000468, Global Avg Loss: 0.61973576, Time: 0.0210 Steps: 102510, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000328, Sample Num: 5248, Cur Loss: 0.09794478, Cur Avg Loss: 0.13361569, Log Avg loss: 0.15982835, Global Avg Loss: 0.61969090, Time: 0.0210 Steps: 102520, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000338, Sample Num: 5408, Cur Loss: 0.10660243, Cur Avg Loss: 0.13302967, Log Avg loss: 0.11380820, Global Avg Loss: 0.61964156, Time: 0.0210 Steps: 102530, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000348, Sample Num: 5568, Cur Loss: 0.07294039, Cur Avg Loss: 0.13371023, Log Avg loss: 0.15671292, Global Avg Loss: 0.61959641, Time: 0.0210 Steps: 102540, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000358, Sample Num: 5728, Cur Loss: 0.14766097, Cur Avg Loss: 0.13394510, Log Avg loss: 0.14211876, Global Avg Loss: 0.61954985, Time: 0.0210 Steps: 102550, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000368, Sample Num: 5888, Cur Loss: 0.01734907, Cur Avg Loss: 0.13386220, Log Avg loss: 0.13089420, Global Avg Loss: 0.61950221, Time: 0.0210 Steps: 102560, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000378, Sample Num: 6048, Cur Loss: 0.07542610, Cur Avg Loss: 0.13359384, Log Avg loss: 0.12371833, Global Avg Loss: 0.61945387, Time: 0.0210 Steps: 102570, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000388, Sample Num: 6208, Cur Loss: 0.15976650, Cur Avg Loss: 0.13390078, Log Avg loss: 0.14550324, Global Avg Loss: 0.61940767, Time: 0.0208 Steps: 102580, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000398, Sample Num: 6368, Cur Loss: 0.17243716, Cur Avg Loss: 0.13382818, Log Avg loss: 0.13101101, Global Avg Loss: 0.61936006, Time: 0.0210 Steps: 102590, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000408, Sample Num: 6528, Cur Loss: 0.26331860, Cur Avg Loss: 0.13513829, Log Avg loss: 0.18728074, Global Avg Loss: 0.61931795, Time: 0.0210 Steps: 102600, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000418, Sample Num: 6688, Cur Loss: 0.05410909, Cur Avg Loss: 0.13497847, Log Avg loss: 0.12845789, Global Avg Loss: 0.61927011, Time: 0.0210 Steps: 102610, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000428, Sample Num: 6848, Cur Loss: 0.24202749, Cur Avg Loss: 0.13510099, Log Avg loss: 0.14022235, Global Avg Loss: 0.61922343, Time: 0.0210 Steps: 102620, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000438, Sample Num: 7008, Cur Loss: 0.03308821, Cur Avg Loss: 0.13636534, Log Avg loss: 0.19047953, Global Avg Loss: 0.61918165, Time: 0.0210 Steps: 102630, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000448, Sample Num: 7168, Cur Loss: 0.11937536, Cur Avg Loss: 0.13623414, Log Avg loss: 0.13048764, Global Avg Loss: 0.61913404, Time: 0.0210 Steps: 102640, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000458, Sample Num: 7328, Cur Loss: 0.03663687, Cur Avg Loss: 0.13497616, Log Avg loss: 0.07861837, Global Avg Loss: 0.61908138, Time: 0.0210 Steps: 102650, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000468, Sample Num: 7488, Cur Loss: 0.14818850, Cur Avg Loss: 0.13552101, Log Avg loss: 0.16047539, Global Avg Loss: 0.61903671, Time: 0.0211 Steps: 102660, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000478, Sample Num: 7648, Cur Loss: 0.12569319, Cur Avg Loss: 0.13545785, Log Avg loss: 0.13250196, Global Avg Loss: 0.61898932, Time: 0.0209 Steps: 102670, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000488, Sample Num: 7808, Cur Loss: 0.10276999, Cur Avg Loss: 0.13524959, Log Avg loss: 0.12529494, Global Avg Loss: 0.61894124, Time: 0.0210 Steps: 102680, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000498, Sample Num: 7968, Cur Loss: 0.14205530, Cur Avg Loss: 0.13456026, Log Avg loss: 0.10092074, Global Avg Loss: 0.61889080, Time: 0.0210 Steps: 102690, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000508, Sample Num: 8128, Cur Loss: 0.35369971, Cur Avg Loss: 0.13415744, Log Avg loss: 0.11409693, Global Avg Loss: 0.61884164, Time: 0.0208 Steps: 102700, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000518, Sample Num: 8288, Cur Loss: 0.10586924, Cur Avg Loss: 0.13513431, Log Avg loss: 0.18475922, Global Avg Loss: 0.61879938, Time: 0.0245 Steps: 102710, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000528, Sample Num: 8448, Cur Loss: 0.03285279, Cur Avg Loss: 0.13599371, Log Avg loss: 0.18051099, Global Avg Loss: 0.61875671, Time: 0.0208 Steps: 102720, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000538, Sample Num: 8608, Cur Loss: 0.12222944, Cur Avg Loss: 0.13614015, Log Avg loss: 0.14387181, Global Avg Loss: 0.61871049, Time: 0.0207 Steps: 102730, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000548, Sample Num: 8768, Cur Loss: 0.14421736, Cur Avg Loss: 0.13560974, Log Avg loss: 0.10707394, Global Avg Loss: 0.61866069, Time: 0.0208 Steps: 102740, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000558, Sample Num: 8928, Cur Loss: 0.04739361, Cur Avg Loss: 0.13517412, Log Avg loss: 0.11130214, Global Avg Loss: 0.61861131, Time: 0.0208 Steps: 102750, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000568, Sample Num: 9088, Cur Loss: 0.06103203, Cur Avg Loss: 0.13561774, Log Avg loss: 0.16037178, Global Avg Loss: 0.61856672, Time: 0.0208 Steps: 102760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000578, Sample Num: 9248, Cur Loss: 0.09722915, Cur Avg Loss: 0.13466195, Log Avg loss: 0.08037301, Global Avg Loss: 0.61851435, Time: 0.0208 Steps: 102770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000588, Sample Num: 9408, Cur Loss: 0.37036517, Cur Avg Loss: 0.13520858, Log Avg loss: 0.16680358, Global Avg Loss: 0.61847040, Time: 0.0208 Steps: 102780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000598, Sample Num: 9568, Cur Loss: 0.04937573, Cur Avg Loss: 0.13479114, Log Avg loss: 0.11024610, Global Avg Loss: 0.61842096, Time: 0.0208 Steps: 102790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000608, Sample Num: 9728, Cur Loss: 0.10996095, Cur Avg Loss: 0.13486480, Log Avg loss: 0.13926941, Global Avg Loss: 0.61837435, Time: 0.0208 Steps: 102800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000618, Sample Num: 9888, Cur Loss: 0.05894168, Cur Avg Loss: 0.13442159, Log Avg loss: 0.10747432, Global Avg Loss: 0.61832465, Time: 0.0208 Steps: 102810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000628, Sample Num: 10048, Cur Loss: 0.06289744, Cur Avg Loss: 0.13414192, Log Avg loss: 0.11685843, Global Avg Loss: 0.61827588, Time: 0.0209 Steps: 102820, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000638, Sample Num: 10208, Cur Loss: 0.34031868, Cur Avg Loss: 0.13462843, Log Avg loss: 0.16518134, Global Avg Loss: 0.61823182, Time: 0.0209 Steps: 102830, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000648, Sample Num: 10368, Cur Loss: 0.07996739, Cur Avg Loss: 0.13578570, Log Avg loss: 0.20961953, Global Avg Loss: 0.61819209, Time: 0.0207 Steps: 102840, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000658, Sample Num: 10528, Cur Loss: 0.17388624, Cur Avg Loss: 0.13525122, Log Avg loss: 0.10061682, Global Avg Loss: 0.61814176, Time: 0.0208 Steps: 102850, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000668, Sample Num: 10688, Cur Loss: 0.03920452, Cur Avg Loss: 0.13462250, Log Avg loss: 0.09325297, Global Avg Loss: 0.61809073, Time: 0.0208 Steps: 102860, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000678, Sample Num: 10848, Cur Loss: 0.07230951, Cur Avg Loss: 0.13436974, Log Avg loss: 0.11748514, Global Avg Loss: 0.61804207, Time: 0.0208 Steps: 102870, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000688, Sample Num: 11008, Cur Loss: 0.26647508, Cur Avg Loss: 0.13503170, Log Avg loss: 0.17991228, Global Avg Loss: 0.61799948, Time: 0.0208 Steps: 102880, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000698, Sample Num: 11168, Cur Loss: 0.44246241, Cur Avg Loss: 0.13577246, Log Avg loss: 0.18673735, Global Avg Loss: 0.61795757, Time: 0.0208 Steps: 102890, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000708, Sample Num: 11328, Cur Loss: 0.03243337, Cur Avg Loss: 0.13592397, Log Avg loss: 0.14649900, Global Avg Loss: 0.61791175, Time: 0.0208 Steps: 102900, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000718, Sample Num: 11488, Cur Loss: 0.34175569, Cur Avg Loss: 0.13633956, Log Avg loss: 0.16576311, Global Avg Loss: 0.61786781, Time: 0.0209 Steps: 102910, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000728, Sample Num: 11648, Cur Loss: 0.06523483, Cur Avg Loss: 0.13591042, Log Avg loss: 0.10509838, Global Avg Loss: 0.61781799, Time: 0.0209 Steps: 102920, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000738, Sample Num: 11808, Cur Loss: 0.07924337, Cur Avg Loss: 0.13554896, Log Avg loss: 0.10923492, Global Avg Loss: 0.61776858, Time: 0.0208 Steps: 102930, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000748, Sample Num: 11968, Cur Loss: 0.09335868, Cur Avg Loss: 0.13543983, Log Avg loss: 0.12738595, Global Avg Loss: 0.61772094, Time: 0.0208 Steps: 102940, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000758, Sample Num: 12128, Cur Loss: 0.07142595, Cur Avg Loss: 0.13523676, Log Avg loss: 0.12004712, Global Avg Loss: 0.61767260, Time: 0.0208 Steps: 102950, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000768, Sample Num: 12288, Cur Loss: 0.15952209, Cur Avg Loss: 0.13530135, Log Avg loss: 0.14019755, Global Avg Loss: 0.61762623, Time: 0.0254 Steps: 102960, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000778, Sample Num: 12448, Cur Loss: 0.33466312, Cur Avg Loss: 0.13623643, Log Avg loss: 0.20805011, Global Avg Loss: 0.61758645, Time: 0.0208 Steps: 102970, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000788, Sample Num: 12608, Cur Loss: 0.21894769, Cur Avg Loss: 0.13645967, Log Avg loss: 0.15382803, Global Avg Loss: 0.61754142, Time: 0.0208 Steps: 102980, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000798, Sample Num: 12768, Cur Loss: 0.05317095, Cur Avg Loss: 0.13640971, Log Avg loss: 0.13247300, Global Avg Loss: 0.61749432, Time: 0.0208 Steps: 102990, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000808, Sample Num: 12928, Cur Loss: 0.08746138, Cur Avg Loss: 0.13633864, Log Avg loss: 0.13066687, Global Avg Loss: 0.61744705, Time: 0.0208 Steps: 103000, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000818, Sample Num: 13088, Cur Loss: 0.06979494, Cur Avg Loss: 0.13638376, Log Avg loss: 0.14002923, Global Avg Loss: 0.61740071, Time: 0.0208 Steps: 103010, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000828, Sample Num: 13248, Cur Loss: 0.08610629, Cur Avg Loss: 0.13676536, Log Avg loss: 0.16798016, Global Avg Loss: 0.61735708, Time: 0.0208 Steps: 103020, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000838, Sample Num: 13408, Cur Loss: 0.15122640, Cur Avg Loss: 0.13681315, Log Avg loss: 0.14077070, Global Avg Loss: 0.61731083, Time: 0.0208 Steps: 103030, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000848, Sample Num: 13568, Cur Loss: 0.23202643, Cur Avg Loss: 0.13700819, Log Avg loss: 0.15335197, Global Avg Loss: 0.61726580, Time: 0.0208 Steps: 103040, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000858, Sample Num: 13728, Cur Loss: 0.12755445, Cur Avg Loss: 0.13684452, Log Avg loss: 0.12296551, Global Avg Loss: 0.61721783, Time: 0.0208 Steps: 103050, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000868, Sample Num: 13888, Cur Loss: 0.05052300, Cur Avg Loss: 0.13705828, Log Avg loss: 0.15539891, Global Avg Loss: 0.61717302, Time: 0.0208 Steps: 103060, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000878, Sample Num: 14048, Cur Loss: 0.03056301, Cur Avg Loss: 0.13680293, Log Avg loss: 0.11463847, Global Avg Loss: 0.61712426, Time: 0.0208 Steps: 103070, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000888, Sample Num: 14208, Cur Loss: 0.13212490, Cur Avg Loss: 0.13742087, Log Avg loss: 0.19167605, Global Avg Loss: 0.61708299, Time: 0.0208 Steps: 103080, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000898, Sample Num: 14368, Cur Loss: 0.41542244, Cur Avg Loss: 0.13746451, Log Avg loss: 0.14134024, Global Avg Loss: 0.61703684, Time: 0.0208 Steps: 103090, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000908, Sample Num: 14528, Cur Loss: 0.10256280, Cur Avg Loss: 0.13743285, Log Avg loss: 0.13458964, Global Avg Loss: 0.61699005, Time: 0.0207 Steps: 103100, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000918, Sample Num: 14688, Cur Loss: 0.07373690, Cur Avg Loss: 0.13711764, Log Avg loss: 0.10849669, Global Avg Loss: 0.61694073, Time: 0.0208 Steps: 103110, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000928, Sample Num: 14848, Cur Loss: 0.15211037, Cur Avg Loss: 0.13695492, Log Avg loss: 0.12201715, Global Avg Loss: 0.61689274, Time: 0.0208 Steps: 103120, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000938, Sample Num: 15008, Cur Loss: 0.05527939, Cur Avg Loss: 0.13683125, Log Avg loss: 0.12535453, Global Avg Loss: 0.61684508, Time: 0.0208 Steps: 103130, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000948, Sample Num: 15168, Cur Loss: 0.31093749, Cur Avg Loss: 0.13730259, Log Avg loss: 0.18151390, Global Avg Loss: 0.61680287, Time: 0.0208 Steps: 103140, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000958, Sample Num: 15328, Cur Loss: 0.14123762, Cur Avg Loss: 0.13694325, Log Avg loss: 0.10287783, Global Avg Loss: 0.61675304, Time: 0.0208 Steps: 103150, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000968, Sample Num: 15488, Cur Loss: 0.03815490, Cur Avg Loss: 0.13655276, Log Avg loss: 0.09914388, Global Avg Loss: 0.61670287, Time: 0.0208 Steps: 103160, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000978, Sample Num: 15648, Cur Loss: 0.18360469, Cur Avg Loss: 0.13680081, Log Avg loss: 0.16081250, Global Avg Loss: 0.61665868, Time: 0.0208 Steps: 103170, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000988, Sample Num: 15808, Cur Loss: 0.06755605, Cur Avg Loss: 0.13709957, Log Avg loss: 0.16631821, Global Avg Loss: 0.61661503, Time: 0.0208 Steps: 103180, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000998, Sample Num: 15968, Cur Loss: 0.09541905, Cur Avg Loss: 0.13712820, Log Avg loss: 0.13995630, Global Avg Loss: 0.61656884, Time: 0.0208 Steps: 103190, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001008, Sample Num: 16128, Cur Loss: 0.03324820, Cur Avg Loss: 0.13663274, Log Avg loss: 0.08718602, Global Avg Loss: 0.61651755, Time: 0.0208 Steps: 103200, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001018, Sample Num: 16288, Cur Loss: 0.02437455, Cur Avg Loss: 0.13650307, Log Avg loss: 0.12343217, Global Avg Loss: 0.61646977, Time: 0.0208 Steps: 103210, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001028, Sample Num: 16448, Cur Loss: 0.37359715, Cur Avg Loss: 0.13684150, Log Avg loss: 0.17129362, Global Avg Loss: 0.61642664, Time: 0.0248 Steps: 103220, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001038, Sample Num: 16608, Cur Loss: 0.12474298, Cur Avg Loss: 0.13681853, Log Avg loss: 0.13445753, Global Avg Loss: 0.61637995, Time: 0.0211 Steps: 103230, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001048, Sample Num: 16768, Cur Loss: 0.08667314, Cur Avg Loss: 0.13663950, Log Avg loss: 0.11805663, Global Avg Loss: 0.61633168, Time: 0.0211 Steps: 103240, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001058, Sample Num: 16928, Cur Loss: 0.17306893, Cur Avg Loss: 0.13686232, Log Avg loss: 0.16021319, Global Avg Loss: 0.61628751, Time: 0.0210 Steps: 103250, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001068, Sample Num: 17088, Cur Loss: 0.05998306, Cur Avg Loss: 0.13715586, Log Avg loss: 0.16821297, Global Avg Loss: 0.61624412, Time: 0.0210 Steps: 103260, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001078, Sample Num: 17248, Cur Loss: 0.22536705, Cur Avg Loss: 0.13707588, Log Avg loss: 0.12853352, Global Avg Loss: 0.61619689, Time: 0.0210 Steps: 103270, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001088, Sample Num: 17408, Cur Loss: 0.13294365, Cur Avg Loss: 0.13680699, Log Avg loss: 0.10782060, Global Avg Loss: 0.61614767, Time: 0.0210 Steps: 103280, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001098, Sample Num: 17568, Cur Loss: 0.09329863, Cur Avg Loss: 0.13798224, Log Avg loss: 0.26584958, Global Avg Loss: 0.61611375, Time: 0.0210 Steps: 103290, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001108, Sample Num: 17728, Cur Loss: 0.08059335, Cur Avg Loss: 0.13770175, Log Avg loss: 0.10690462, Global Avg Loss: 0.61606446, Time: 0.0210 Steps: 103300, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001118, Sample Num: 17888, Cur Loss: 0.11208406, Cur Avg Loss: 0.13767339, Log Avg loss: 0.13453089, Global Avg Loss: 0.61601785, Time: 0.0210 Steps: 103310, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001128, Sample Num: 18048, Cur Loss: 0.15826391, Cur Avg Loss: 0.13749111, Log Avg loss: 0.11711246, Global Avg Loss: 0.61596956, Time: 0.0210 Steps: 103320, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001138, Sample Num: 18208, Cur Loss: 0.12399815, Cur Avg Loss: 0.13731432, Log Avg loss: 0.11737138, Global Avg Loss: 0.61592131, Time: 0.0210 Steps: 103330, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001148, Sample Num: 18368, Cur Loss: 0.21330099, Cur Avg Loss: 0.13757223, Log Avg loss: 0.16692327, Global Avg Loss: 0.61587786, Time: 0.0210 Steps: 103340, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001158, Sample Num: 18528, Cur Loss: 0.12082648, Cur Avg Loss: 0.13768043, Log Avg loss: 0.15010129, Global Avg Loss: 0.61583279, Time: 0.0210 Steps: 103350, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001168, Sample Num: 18688, Cur Loss: 0.13483293, Cur Avg Loss: 0.13732787, Log Avg loss: 0.09650166, Global Avg Loss: 0.61578255, Time: 0.0210 Steps: 103360, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001178, Sample Num: 18848, Cur Loss: 0.10293113, Cur Avg Loss: 0.13731950, Log Avg loss: 0.13634166, Global Avg Loss: 0.61573616, Time: 0.0210 Steps: 103370, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001188, Sample Num: 19008, Cur Loss: 0.06390994, Cur Avg Loss: 0.13711698, Log Avg loss: 0.11326034, Global Avg Loss: 0.61568756, Time: 0.0210 Steps: 103380, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001198, Sample Num: 19168, Cur Loss: 0.21194330, Cur Avg Loss: 0.13686095, Log Avg loss: 0.10644472, Global Avg Loss: 0.61563830, Time: 0.0210 Steps: 103390, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001208, Sample Num: 19328, Cur Loss: 0.23442963, Cur Avg Loss: 0.13711217, Log Avg loss: 0.16720832, Global Avg Loss: 0.61559494, Time: 0.0210 Steps: 103400, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001218, Sample Num: 19488, Cur Loss: 0.14273439, Cur Avg Loss: 0.13712392, Log Avg loss: 0.13854294, Global Avg Loss: 0.61554880, Time: 0.0210 Steps: 103410, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001228, Sample Num: 19648, Cur Loss: 0.28435105, Cur Avg Loss: 0.13771070, Log Avg loss: 0.20918046, Global Avg Loss: 0.61550951, Time: 0.0210 Steps: 103420, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001238, Sample Num: 19808, Cur Loss: 0.09724084, Cur Avg Loss: 0.13803882, Log Avg loss: 0.17833169, Global Avg Loss: 0.61546724, Time: 0.0210 Steps: 103430, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001248, Sample Num: 19968, Cur Loss: 0.15998329, Cur Avg Loss: 0.13791749, Log Avg loss: 0.12289682, Global Avg Loss: 0.61541962, Time: 0.0210 Steps: 103440, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001258, Sample Num: 20128, Cur Loss: 0.22830439, Cur Avg Loss: 0.13850305, Log Avg loss: 0.21158087, Global Avg Loss: 0.61538059, Time: 0.0210 Steps: 103450, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001268, Sample Num: 20288, Cur Loss: 0.03510386, Cur Avg Loss: 0.13840743, Log Avg loss: 0.12637886, Global Avg Loss: 0.61533332, Time: 0.0210 Steps: 103460, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001278, Sample Num: 20448, Cur Loss: 0.22416078, Cur Avg Loss: 0.13833615, Log Avg loss: 0.12929765, Global Avg Loss: 0.61528635, Time: 0.0210 Steps: 103470, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001288, Sample Num: 20608, Cur Loss: 0.12224597, Cur Avg Loss: 0.13839535, Log Avg loss: 0.14596128, Global Avg Loss: 0.61524099, Time: 0.0209 Steps: 103480, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001298, Sample Num: 20768, Cur Loss: 0.05076164, Cur Avg Loss: 0.13835681, Log Avg loss: 0.13339242, Global Avg Loss: 0.61519443, Time: 0.0208 Steps: 103490, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001308, Sample Num: 20928, Cur Loss: 0.29571533, Cur Avg Loss: 0.13831202, Log Avg loss: 0.13249932, Global Avg Loss: 0.61514780, Time: 0.0208 Steps: 103500, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001318, Sample Num: 21088, Cur Loss: 0.11589183, Cur Avg Loss: 0.13794760, Log Avg loss: 0.09028090, Global Avg Loss: 0.61509709, Time: 0.0209 Steps: 103510, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001328, Sample Num: 21248, Cur Loss: 0.11860271, Cur Avg Loss: 0.13781704, Log Avg loss: 0.12060926, Global Avg Loss: 0.61504932, Time: 0.0208 Steps: 103520, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001338, Sample Num: 21408, Cur Loss: 0.10418746, Cur Avg Loss: 0.13803719, Log Avg loss: 0.16727353, Global Avg Loss: 0.61500607, Time: 0.0208 Steps: 103530, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001348, Sample Num: 21568, Cur Loss: 0.07907784, Cur Avg Loss: 0.13835738, Log Avg loss: 0.18119869, Global Avg Loss: 0.61496417, Time: 0.0208 Steps: 103540, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001358, Sample Num: 21728, Cur Loss: 0.11010170, Cur Avg Loss: 0.13839797, Log Avg loss: 0.14386956, Global Avg Loss: 0.61491868, Time: 0.0208 Steps: 103550, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001368, Sample Num: 21888, Cur Loss: 0.07292858, Cur Avg Loss: 0.13813280, Log Avg loss: 0.10212299, Global Avg Loss: 0.61486916, Time: 0.0208 Steps: 103560, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001378, Sample Num: 22048, Cur Loss: 0.12736210, Cur Avg Loss: 0.13818803, Log Avg loss: 0.14574257, Global Avg Loss: 0.61482387, Time: 0.0208 Steps: 103570, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001388, Sample Num: 22208, Cur Loss: 0.08644858, Cur Avg Loss: 0.13793731, Log Avg loss: 0.10338806, Global Avg Loss: 0.61477449, Time: 0.0208 Steps: 103580, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001398, Sample Num: 22368, Cur Loss: 0.32959282, Cur Avg Loss: 0.13816116, Log Avg loss: 0.16923176, Global Avg Loss: 0.61473148, Time: 0.0208 Steps: 103590, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001408, Sample Num: 22528, Cur Loss: 0.05156162, Cur Avg Loss: 0.13839661, Log Avg loss: 0.17131278, Global Avg Loss: 0.61468868, Time: 0.0208 Steps: 103600, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001418, Sample Num: 22688, Cur Loss: 0.24303244, Cur Avg Loss: 0.13863307, Log Avg loss: 0.17192676, Global Avg Loss: 0.61464595, Time: 0.0208 Steps: 103610, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001428, Sample Num: 22848, Cur Loss: 0.17061901, Cur Avg Loss: 0.13876923, Log Avg loss: 0.15807682, Global Avg Loss: 0.61460189, Time: 0.0208 Steps: 103620, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001438, Sample Num: 23008, Cur Loss: 0.07888051, Cur Avg Loss: 0.13903057, Log Avg loss: 0.17634954, Global Avg Loss: 0.61455960, Time: 0.0209 Steps: 103630, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001448, Sample Num: 23168, Cur Loss: 0.06781442, Cur Avg Loss: 0.13908794, Log Avg loss: 0.14733818, Global Avg Loss: 0.61451451, Time: 0.0208 Steps: 103640, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001458, Sample Num: 23328, Cur Loss: 0.03878292, Cur Avg Loss: 0.13871908, Log Avg loss: 0.08530711, Global Avg Loss: 0.61446346, Time: 0.0208 Steps: 103650, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001468, Sample Num: 23488, Cur Loss: 0.08910233, Cur Avg Loss: 0.13878525, Log Avg loss: 0.14843356, Global Avg Loss: 0.61441850, Time: 0.0208 Steps: 103660, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001478, Sample Num: 23648, Cur Loss: 0.09536039, Cur Avg Loss: 0.13924174, Log Avg loss: 0.20625508, Global Avg Loss: 0.61437913, Time: 0.0208 Steps: 103670, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001488, Sample Num: 23808, Cur Loss: 0.07022125, Cur Avg Loss: 0.13909444, Log Avg loss: 0.11732229, Global Avg Loss: 0.61433119, Time: 0.0208 Steps: 103680, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001498, Sample Num: 23968, Cur Loss: 0.24480952, Cur Avg Loss: 0.13910650, Log Avg loss: 0.14090192, Global Avg Loss: 0.61428553, Time: 0.0208 Steps: 103690, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001508, Sample Num: 24128, Cur Loss: 0.03187026, Cur Avg Loss: 0.13938633, Log Avg loss: 0.18130393, Global Avg Loss: 0.61424378, Time: 0.0208 Steps: 103700, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001518, Sample Num: 24288, Cur Loss: 0.17765546, Cur Avg Loss: 0.13967838, Log Avg loss: 0.18371958, Global Avg Loss: 0.61420226, Time: 0.0209 Steps: 103710, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001528, Sample Num: 24448, Cur Loss: 0.02985276, Cur Avg Loss: 0.13950334, Log Avg loss: 0.11293256, Global Avg Loss: 0.61415393, Time: 0.0208 Steps: 103720, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001538, Sample Num: 24608, Cur Loss: 0.14186631, Cur Avg Loss: 0.13952515, Log Avg loss: 0.14285806, Global Avg Loss: 0.61410850, Time: 0.0246 Steps: 103730, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001548, Sample Num: 24768, Cur Loss: 0.15236436, Cur Avg Loss: 0.13957836, Log Avg loss: 0.14776134, Global Avg Loss: 0.61406355, Time: 0.0208 Steps: 103740, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001558, Sample Num: 24928, Cur Loss: 0.11214038, Cur Avg Loss: 0.13933394, Log Avg loss: 0.10149900, Global Avg Loss: 0.61401414, Time: 0.0208 Steps: 103750, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001568, Sample Num: 25088, Cur Loss: 0.16271396, Cur Avg Loss: 0.13918273, Log Avg loss: 0.11562310, Global Avg Loss: 0.61396611, Time: 0.0208 Steps: 103760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001578, Sample Num: 25248, Cur Loss: 0.12112805, Cur Avg Loss: 0.13890304, Log Avg loss: 0.09504781, Global Avg Loss: 0.61391610, Time: 0.0208 Steps: 103770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001588, Sample Num: 25408, Cur Loss: 0.05483210, Cur Avg Loss: 0.13918277, Log Avg loss: 0.18332485, Global Avg Loss: 0.61387461, Time: 0.0208 Steps: 103780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001598, Sample Num: 25568, Cur Loss: 0.11485304, Cur Avg Loss: 0.13888858, Log Avg loss: 0.09217006, Global Avg Loss: 0.61382435, Time: 0.0208 Steps: 103790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001608, Sample Num: 25728, Cur Loss: 0.08278765, Cur Avg Loss: 0.13904293, Log Avg loss: 0.16370847, Global Avg Loss: 0.61378098, Time: 0.0208 Steps: 103800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001618, Sample Num: 25888, Cur Loss: 0.08294941, Cur Avg Loss: 0.13890071, Log Avg loss: 0.11603220, Global Avg Loss: 0.61373303, Time: 0.0208 Steps: 103810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001628, Sample Num: 26048, Cur Loss: 0.04574976, Cur Avg Loss: 0.13937805, Log Avg loss: 0.21661143, Global Avg Loss: 0.61369478, Time: 0.0208 Steps: 103820, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001638, Sample Num: 26208, Cur Loss: 0.31797397, Cur Avg Loss: 0.13942640, Log Avg loss: 0.14729734, Global Avg Loss: 0.61364986, Time: 0.0208 Steps: 103830, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001648, Sample Num: 26368, Cur Loss: 0.06213591, Cur Avg Loss: 0.13928451, Log Avg loss: 0.11604389, Global Avg Loss: 0.61360194, Time: 0.0208 Steps: 103840, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001658, Sample Num: 26528, Cur Loss: 0.17275843, Cur Avg Loss: 0.13966509, Log Avg loss: 0.20238414, Global Avg Loss: 0.61356235, Time: 0.0208 Steps: 103850, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001668, Sample Num: 26688, Cur Loss: 0.05017065, Cur Avg Loss: 0.13953375, Log Avg loss: 0.11775772, Global Avg Loss: 0.61351461, Time: 0.0208 Steps: 103860, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001678, Sample Num: 26848, Cur Loss: 0.21911149, Cur Avg Loss: 0.13975939, Log Avg loss: 0.17739695, Global Avg Loss: 0.61347262, Time: 0.0208 Steps: 103870, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001688, Sample Num: 27008, Cur Loss: 0.18233716, Cur Avg Loss: 0.13955967, Log Avg loss: 0.10604589, Global Avg Loss: 0.61342377, Time: 0.0208 Steps: 103880, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001698, Sample Num: 27168, Cur Loss: 0.05171238, Cur Avg Loss: 0.13933212, Log Avg loss: 0.10092188, Global Avg Loss: 0.61337444, Time: 0.0208 Steps: 103890, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001708, Sample Num: 27328, Cur Loss: 0.15756494, Cur Avg Loss: 0.13955853, Log Avg loss: 0.17800242, Global Avg Loss: 0.61333254, Time: 0.0208 Steps: 103900, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001718, Sample Num: 27488, Cur Loss: 0.10340580, Cur Avg Loss: 0.13950493, Log Avg loss: 0.13034976, Global Avg Loss: 0.61328606, Time: 0.0208 Steps: 103910, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001728, Sample Num: 27648, Cur Loss: 0.26148993, Cur Avg Loss: 0.13967018, Log Avg loss: 0.16806081, Global Avg Loss: 0.61324322, Time: 0.0208 Steps: 103920, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001738, Sample Num: 27808, Cur Loss: 0.21980259, Cur Avg Loss: 0.13973973, Log Avg loss: 0.15175718, Global Avg Loss: 0.61319881, Time: 0.0208 Steps: 103930, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001748, Sample Num: 27968, Cur Loss: 0.18429819, Cur Avg Loss: 0.13964372, Log Avg loss: 0.12295808, Global Avg Loss: 0.61315165, Time: 0.0208 Steps: 103940, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001758, Sample Num: 28128, Cur Loss: 0.11191723, Cur Avg Loss: 0.13956575, Log Avg loss: 0.12593705, Global Avg Loss: 0.61310478, Time: 0.0208 Steps: 103950, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001768, Sample Num: 28288, Cur Loss: 0.06676821, Cur Avg Loss: 0.13997153, Log Avg loss: 0.21130694, Global Avg Loss: 0.61306613, Time: 0.0208 Steps: 103960, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001778, Sample Num: 28448, Cur Loss: 0.02580983, Cur Avg Loss: 0.13982270, Log Avg loss: 0.11351037, Global Avg Loss: 0.61301808, Time: 0.0207 Steps: 103970, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001788, Sample Num: 28608, Cur Loss: 0.25839242, Cur Avg Loss: 0.13961130, Log Avg loss: 0.10202369, Global Avg Loss: 0.61296894, Time: 0.0207 Steps: 103980, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001798, Sample Num: 28768, Cur Loss: 0.14231361, Cur Avg Loss: 0.13983418, Log Avg loss: 0.17968469, Global Avg Loss: 0.61292727, Time: 0.0208 Steps: 103990, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001808, Sample Num: 28928, Cur Loss: 0.07334545, Cur Avg Loss: 0.13984493, Log Avg loss: 0.14177848, Global Avg Loss: 0.61288197, Time: 0.0208 Steps: 104000, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001818, Sample Num: 29088, Cur Loss: 0.05696652, Cur Avg Loss: 0.13992768, Log Avg loss: 0.15488794, Global Avg Loss: 0.61283793, Time: 0.0208 Steps: 104010, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001828, Sample Num: 29248, Cur Loss: 0.25511399, Cur Avg Loss: 0.14015561, Log Avg loss: 0.18159442, Global Avg Loss: 0.61279648, Time: 0.0208 Steps: 104020, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001838, Sample Num: 29408, Cur Loss: 0.18019873, Cur Avg Loss: 0.14048514, Log Avg loss: 0.20072236, Global Avg Loss: 0.61275686, Time: 0.0209 Steps: 104030, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001848, Sample Num: 29568, Cur Loss: 0.09716036, Cur Avg Loss: 0.14038082, Log Avg loss: 0.12120757, Global Avg Loss: 0.61270962, Time: 0.0208 Steps: 104040, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001858, Sample Num: 29728, Cur Loss: 0.06477525, Cur Avg Loss: 0.14041721, Log Avg loss: 0.14714076, Global Avg Loss: 0.61266487, Time: 0.0208 Steps: 104050, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001868, Sample Num: 29888, Cur Loss: 0.16024938, Cur Avg Loss: 0.14051751, Log Avg loss: 0.15915436, Global Avg Loss: 0.61262129, Time: 0.0208 Steps: 104060, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001878, Sample Num: 30048, Cur Loss: 0.23619993, Cur Avg Loss: 0.14060141, Log Avg loss: 0.15627309, Global Avg Loss: 0.61257744, Time: 0.0208 Steps: 104070, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001888, Sample Num: 30208, Cur Loss: 0.26033896, Cur Avg Loss: 0.14048710, Log Avg loss: 0.11901965, Global Avg Loss: 0.61253002, Time: 0.0208 Steps: 104080, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001898, Sample Num: 30368, Cur Loss: 0.11520517, Cur Avg Loss: 0.14052658, Log Avg loss: 0.14798157, Global Avg Loss: 0.61248539, Time: 0.0208 Steps: 104090, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001908, Sample Num: 30528, Cur Loss: 0.15251458, Cur Avg Loss: 0.14033346, Log Avg loss: 0.10367933, Global Avg Loss: 0.61243651, Time: 0.0208 Steps: 104100, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001918, Sample Num: 30688, Cur Loss: 0.25800169, Cur Avg Loss: 0.14033655, Log Avg loss: 0.14092634, Global Avg Loss: 0.61239122, Time: 0.0208 Steps: 104110, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001928, Sample Num: 30848, Cur Loss: 0.01972174, Cur Avg Loss: 0.14041526, Log Avg loss: 0.15551086, Global Avg Loss: 0.61234734, Time: 0.0207 Steps: 104120, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001938, Sample Num: 31008, Cur Loss: 0.06499350, Cur Avg Loss: 0.14060041, Log Avg loss: 0.17629727, Global Avg Loss: 0.61230547, Time: 0.0207 Steps: 104130, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001948, Sample Num: 31168, Cur Loss: 0.18333259, Cur Avg Loss: 0.14040768, Log Avg loss: 0.10305688, Global Avg Loss: 0.61225657, Time: 0.0208 Steps: 104140, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001958, Sample Num: 31328, Cur Loss: 0.24032280, Cur Avg Loss: 0.14073445, Log Avg loss: 0.20439006, Global Avg Loss: 0.61221741, Time: 0.0207 Steps: 104150, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001968, Sample Num: 31488, Cur Loss: 0.33947307, Cur Avg Loss: 0.14098914, Log Avg loss: 0.19085658, Global Avg Loss: 0.61217695, Time: 0.0208 Steps: 104160, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001978, Sample Num: 31648, Cur Loss: 0.09657930, Cur Avg Loss: 0.14085519, Log Avg loss: 0.11449431, Global Avg Loss: 0.61212918, Time: 0.0208 Steps: 104170, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001988, Sample Num: 31808, Cur Loss: 0.26766166, Cur Avg Loss: 0.14084893, Log Avg loss: 0.13961026, Global Avg Loss: 0.61208382, Time: 0.0208 Steps: 104180, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001998, Sample Num: 31968, Cur Loss: 0.19835749, Cur Avg Loss: 0.14076493, Log Avg loss: 0.12406518, Global Avg Loss: 0.61203698, Time: 0.0208 Steps: 104190, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002008, Sample Num: 32128, Cur Loss: 0.17211333, Cur Avg Loss: 0.14118248, Log Avg loss: 0.22460868, Global Avg Loss: 0.61199980, Time: 0.0208 Steps: 104200, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002018, Sample Num: 32288, Cur Loss: 0.15967089, Cur Avg Loss: 0.14128818, Log Avg loss: 0.16251302, Global Avg Loss: 0.61195667, Time: 0.0208 Steps: 104210, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002028, Sample Num: 32448, Cur Loss: 0.37048495, Cur Avg Loss: 0.14135977, Log Avg loss: 0.15580782, Global Avg Loss: 0.61191290, Time: 0.0208 Steps: 104220, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002038, Sample Num: 32608, Cur Loss: 0.13544674, Cur Avg Loss: 0.14146935, Log Avg loss: 0.16369241, Global Avg Loss: 0.61186990, Time: 0.0208 Steps: 104230, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002048, Sample Num: 32768, Cur Loss: 0.30879062, Cur Avg Loss: 0.14167854, Log Avg loss: 0.18431020, Global Avg Loss: 0.61182888, Time: 0.0253 Steps: 104240, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002058, Sample Num: 32928, Cur Loss: 0.14446533, Cur Avg Loss: 0.14173161, Log Avg loss: 0.15260000, Global Avg Loss: 0.61178483, Time: 0.0209 Steps: 104250, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002068, Sample Num: 33088, Cur Loss: 0.15246351, Cur Avg Loss: 0.14158615, Log Avg loss: 0.11165053, Global Avg Loss: 0.61173686, Time: 0.0209 Steps: 104260, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002078, Sample Num: 33248, Cur Loss: 0.05024596, Cur Avg Loss: 0.14151184, Log Avg loss: 0.12614521, Global Avg Loss: 0.61169029, Time: 0.0209 Steps: 104270, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002088, Sample Num: 33408, Cur Loss: 0.40602982, Cur Avg Loss: 0.14152627, Log Avg loss: 0.14452508, Global Avg Loss: 0.61164549, Time: 0.0209 Steps: 104280, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002098, Sample Num: 33568, Cur Loss: 0.05174447, Cur Avg Loss: 0.14126170, Log Avg loss: 0.08601873, Global Avg Loss: 0.61159509, Time: 0.0208 Steps: 104290, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002108, Sample Num: 33728, Cur Loss: 0.18524691, Cur Avg Loss: 0.14120022, Log Avg loss: 0.12830278, Global Avg Loss: 0.61154875, Time: 0.0208 Steps: 104300, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002118, Sample Num: 33888, Cur Loss: 0.22382414, Cur Avg Loss: 0.14117938, Log Avg loss: 0.13678608, Global Avg Loss: 0.61150324, Time: 0.0208 Steps: 104310, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002128, Sample Num: 34048, Cur Loss: 0.09551321, Cur Avg Loss: 0.14120877, Log Avg loss: 0.14743245, Global Avg Loss: 0.61145875, Time: 0.0208 Steps: 104320, Updated lr: 0.000002 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.586447, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.241853, "eval_total_loss": 170.022556, "eval_mae": 0.326508, "eval_mse": 0.241945, "eval_r2": 0.846204, "eval_sp_statistic": 0.904963, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925388, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.765368, "test_total_loss": 384.214938, "test_mae": 0.683683, "test_mse": 0.765407, "test_r2": 0.506, "test_sp_statistic": 0.799003, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.838989, "test_ps_pvalue": 0.0, "lr": 2.0189663347558085e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6114536993416736, "train_cur_epoch_loss": 300.57646842859685, "train_cur_epoch_avg_loss": 0.14118199550427282, "train_cur_epoch_time": 44.58644700050354, "train_cur_epoch_avg_time": 0.020942436355332803, "epoch": 49, "step": 104321} ################################################## Training, Epoch: 0050, Batch: 000009, Sample Num: 144, Cur Loss: 0.06118213, Cur Avg Loss: 0.09435990, Log Avg loss: 0.09334526, Global Avg Loss: 0.61140909, Time: 0.0210 Steps: 104330, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000019, Sample Num: 304, Cur Loss: 0.20798640, Cur Avg Loss: 0.11753562, Log Avg loss: 0.13839378, Global Avg Loss: 0.61136376, Time: 0.0208 Steps: 104340, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000029, Sample Num: 464, Cur Loss: 0.09853043, Cur Avg Loss: 0.10570893, Log Avg loss: 0.08323822, Global Avg Loss: 0.61131315, Time: 0.0208 Steps: 104350, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000039, Sample Num: 624, Cur Loss: 0.10008149, Cur Avg Loss: 0.10295118, Log Avg loss: 0.09495368, Global Avg Loss: 0.61126367, Time: 0.0209 Steps: 104360, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000049, Sample Num: 784, Cur Loss: 0.28635433, Cur Avg Loss: 0.11403247, Log Avg loss: 0.15724951, Global Avg Loss: 0.61122017, Time: 0.0208 Steps: 104370, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000059, Sample Num: 944, Cur Loss: 0.43449366, Cur Avg Loss: 0.12100559, Log Avg loss: 0.15517388, Global Avg Loss: 0.61117648, Time: 0.0208 Steps: 104380, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000069, Sample Num: 1104, Cur Loss: 0.04421501, Cur Avg Loss: 0.12266624, Log Avg loss: 0.13246409, Global Avg Loss: 0.61113062, Time: 0.0208 Steps: 104390, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000079, Sample Num: 1264, Cur Loss: 0.07362732, Cur Avg Loss: 0.12257025, Log Avg loss: 0.12190789, Global Avg Loss: 0.61108376, Time: 0.0209 Steps: 104400, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000089, Sample Num: 1424, Cur Loss: 0.14526820, Cur Avg Loss: 0.12400715, Log Avg loss: 0.13535870, Global Avg Loss: 0.61103820, Time: 0.0208 Steps: 104410, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000099, Sample Num: 1584, Cur Loss: 0.10187817, Cur Avg Loss: 0.12079566, Log Avg loss: 0.09221335, Global Avg Loss: 0.61098851, Time: 0.0208 Steps: 104420, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000109, Sample Num: 1744, Cur Loss: 0.04015914, Cur Avg Loss: 0.12503127, Log Avg loss: 0.16696381, Global Avg Loss: 0.61094599, Time: 0.0208 Steps: 104430, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000119, Sample Num: 1904, Cur Loss: 0.05728551, Cur Avg Loss: 0.12951851, Log Avg loss: 0.17842945, Global Avg Loss: 0.61090458, Time: 0.0209 Steps: 104440, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000129, Sample Num: 2064, Cur Loss: 0.03911334, Cur Avg Loss: 0.13130785, Log Avg loss: 0.15260102, Global Avg Loss: 0.61086070, Time: 0.0208 Steps: 104450, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000139, Sample Num: 2224, Cur Loss: 0.26378167, Cur Avg Loss: 0.13348515, Log Avg loss: 0.16157230, Global Avg Loss: 0.61081769, Time: 0.0208 Steps: 104460, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000149, Sample Num: 2384, Cur Loss: 0.13894938, Cur Avg Loss: 0.13465089, Log Avg loss: 0.15085466, Global Avg Loss: 0.61077366, Time: 0.0208 Steps: 104470, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000159, Sample Num: 2544, Cur Loss: 0.11783096, Cur Avg Loss: 0.13037130, Log Avg loss: 0.06660546, Global Avg Loss: 0.61072158, Time: 0.0208 Steps: 104480, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000169, Sample Num: 2704, Cur Loss: 0.23383251, Cur Avg Loss: 0.13276064, Log Avg loss: 0.17075108, Global Avg Loss: 0.61067947, Time: 0.0208 Steps: 104490, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000179, Sample Num: 2864, Cur Loss: 0.04152005, Cur Avg Loss: 0.13061952, Log Avg loss: 0.09443468, Global Avg Loss: 0.61063007, Time: 0.0208 Steps: 104500, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000189, Sample Num: 3024, Cur Loss: 0.14364971, Cur Avg Loss: 0.13215314, Log Avg loss: 0.15960488, Global Avg Loss: 0.61058691, Time: 0.0208 Steps: 104510, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000199, Sample Num: 3184, Cur Loss: 0.07782403, Cur Avg Loss: 0.13232758, Log Avg loss: 0.13562450, Global Avg Loss: 0.61054147, Time: 0.0208 Steps: 104520, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000209, Sample Num: 3344, Cur Loss: 0.17524594, Cur Avg Loss: 0.13311088, Log Avg loss: 0.14869849, Global Avg Loss: 0.61049729, Time: 0.0209 Steps: 104530, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000219, Sample Num: 3504, Cur Loss: 0.02150850, Cur Avg Loss: 0.13442686, Log Avg loss: 0.16193084, Global Avg Loss: 0.61045438, Time: 0.0208 Steps: 104540, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000229, Sample Num: 3664, Cur Loss: 0.11051419, Cur Avg Loss: 0.13354487, Log Avg loss: 0.11422946, Global Avg Loss: 0.61040692, Time: 0.0208 Steps: 104550, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000239, Sample Num: 3824, Cur Loss: 0.05175270, Cur Avg Loss: 0.13419747, Log Avg loss: 0.14914181, Global Avg Loss: 0.61036280, Time: 0.0208 Steps: 104560, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000249, Sample Num: 3984, Cur Loss: 0.04397471, Cur Avg Loss: 0.13558138, Log Avg loss: 0.16865700, Global Avg Loss: 0.61032056, Time: 0.0208 Steps: 104570, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000259, Sample Num: 4144, Cur Loss: 0.05924588, Cur Avg Loss: 0.13522509, Log Avg loss: 0.12635347, Global Avg Loss: 0.61027428, Time: 0.0247 Steps: 104580, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000269, Sample Num: 4304, Cur Loss: 0.04865292, Cur Avg Loss: 0.13390794, Log Avg loss: 0.09979355, Global Avg Loss: 0.61022548, Time: 0.0209 Steps: 104590, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000279, Sample Num: 4464, Cur Loss: 0.04075562, Cur Avg Loss: 0.13398176, Log Avg loss: 0.13596764, Global Avg Loss: 0.61018014, Time: 0.0214 Steps: 104600, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000289, Sample Num: 4624, Cur Loss: 0.19033241, Cur Avg Loss: 0.13338867, Log Avg loss: 0.11684159, Global Avg Loss: 0.61013298, Time: 0.0213 Steps: 104610, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000299, Sample Num: 4784, Cur Loss: 0.16234292, Cur Avg Loss: 0.13337399, Log Avg loss: 0.13294973, Global Avg Loss: 0.61008737, Time: 0.0214 Steps: 104620, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000309, Sample Num: 4944, Cur Loss: 0.07112152, Cur Avg Loss: 0.13323207, Log Avg loss: 0.12898841, Global Avg Loss: 0.61004138, Time: 0.0210 Steps: 104630, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000319, Sample Num: 5104, Cur Loss: 0.07958977, Cur Avg Loss: 0.13305361, Log Avg loss: 0.12753926, Global Avg Loss: 0.60999527, Time: 0.0210 Steps: 104640, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000329, Sample Num: 5264, Cur Loss: 0.09405041, Cur Avg Loss: 0.13227799, Log Avg loss: 0.10753570, Global Avg Loss: 0.60994726, Time: 0.0210 Steps: 104650, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000339, Sample Num: 5424, Cur Loss: 0.11943337, Cur Avg Loss: 0.13296885, Log Avg loss: 0.15569821, Global Avg Loss: 0.60990386, Time: 0.0209 Steps: 104660, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000349, Sample Num: 5584, Cur Loss: 0.28038466, Cur Avg Loss: 0.13591518, Log Avg loss: 0.23579593, Global Avg Loss: 0.60986812, Time: 0.0209 Steps: 104670, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000359, Sample Num: 5744, Cur Loss: 0.18485364, Cur Avg Loss: 0.13641070, Log Avg loss: 0.15370409, Global Avg Loss: 0.60982454, Time: 0.0210 Steps: 104680, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000369, Sample Num: 5904, Cur Loss: 0.02482650, Cur Avg Loss: 0.13567357, Log Avg loss: 0.10921076, Global Avg Loss: 0.60977672, Time: 0.0209 Steps: 104690, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000379, Sample Num: 6064, Cur Loss: 0.21217075, Cur Avg Loss: 0.13521064, Log Avg loss: 0.11812857, Global Avg Loss: 0.60972976, Time: 0.0211 Steps: 104700, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000389, Sample Num: 6224, Cur Loss: 0.37368634, Cur Avg Loss: 0.13537318, Log Avg loss: 0.14153343, Global Avg Loss: 0.60968505, Time: 0.0209 Steps: 104710, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000399, Sample Num: 6384, Cur Loss: 0.10418870, Cur Avg Loss: 0.13532137, Log Avg loss: 0.13330602, Global Avg Loss: 0.60963956, Time: 0.0208 Steps: 104720, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000409, Sample Num: 6544, Cur Loss: 0.20975319, Cur Avg Loss: 0.13572778, Log Avg loss: 0.15194332, Global Avg Loss: 0.60959586, Time: 0.0210 Steps: 104730, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000419, Sample Num: 6704, Cur Loss: 0.10137395, Cur Avg Loss: 0.13613284, Log Avg loss: 0.15269982, Global Avg Loss: 0.60955223, Time: 0.0209 Steps: 104740, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000429, Sample Num: 6864, Cur Loss: 0.20759594, Cur Avg Loss: 0.13580128, Log Avg loss: 0.12190890, Global Avg Loss: 0.60950568, Time: 0.0209 Steps: 104750, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000439, Sample Num: 7024, Cur Loss: 0.18653391, Cur Avg Loss: 0.13564437, Log Avg loss: 0.12891316, Global Avg Loss: 0.60945981, Time: 0.0209 Steps: 104760, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000449, Sample Num: 7184, Cur Loss: 0.14123696, Cur Avg Loss: 0.13597185, Log Avg loss: 0.15034788, Global Avg Loss: 0.60941598, Time: 0.0209 Steps: 104770, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000459, Sample Num: 7344, Cur Loss: 0.10452706, Cur Avg Loss: 0.13653221, Log Avg loss: 0.16169257, Global Avg Loss: 0.60937325, Time: 0.0209 Steps: 104780, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000469, Sample Num: 7504, Cur Loss: 0.09600734, Cur Avg Loss: 0.13550315, Log Avg loss: 0.08826914, Global Avg Loss: 0.60932353, Time: 0.0209 Steps: 104790, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000479, Sample Num: 7664, Cur Loss: 0.05261348, Cur Avg Loss: 0.13569705, Log Avg loss: 0.14479098, Global Avg Loss: 0.60927920, Time: 0.0209 Steps: 104800, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000489, Sample Num: 7824, Cur Loss: 0.10576206, Cur Avg Loss: 0.13448907, Log Avg loss: 0.07662701, Global Avg Loss: 0.60922838, Time: 0.0209 Steps: 104810, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000499, Sample Num: 7984, Cur Loss: 0.09500148, Cur Avg Loss: 0.13430719, Log Avg loss: 0.12541336, Global Avg Loss: 0.60918222, Time: 0.0209 Steps: 104820, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000509, Sample Num: 8144, Cur Loss: 0.16906554, Cur Avg Loss: 0.13378846, Log Avg loss: 0.10790381, Global Avg Loss: 0.60913441, Time: 0.0210 Steps: 104830, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000519, Sample Num: 8304, Cur Loss: 0.15287519, Cur Avg Loss: 0.13316398, Log Avg loss: 0.10137792, Global Avg Loss: 0.60908597, Time: 0.0211 Steps: 104840, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000529, Sample Num: 8464, Cur Loss: 0.04561340, Cur Avg Loss: 0.13288050, Log Avg loss: 0.11816797, Global Avg Loss: 0.60903915, Time: 0.0211 Steps: 104850, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000539, Sample Num: 8624, Cur Loss: 0.04476159, Cur Avg Loss: 0.13304532, Log Avg loss: 0.14176404, Global Avg Loss: 0.60899459, Time: 0.0212 Steps: 104860, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000549, Sample Num: 8784, Cur Loss: 0.08758469, Cur Avg Loss: 0.13301702, Log Avg loss: 0.13149188, Global Avg Loss: 0.60894906, Time: 0.0211 Steps: 104870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000559, Sample Num: 8944, Cur Loss: 0.40200451, Cur Avg Loss: 0.13295003, Log Avg loss: 0.12927185, Global Avg Loss: 0.60890332, Time: 0.0219 Steps: 104880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000569, Sample Num: 9104, Cur Loss: 0.21152826, Cur Avg Loss: 0.13282510, Log Avg loss: 0.12584192, Global Avg Loss: 0.60885727, Time: 0.0211 Steps: 104890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000579, Sample Num: 9264, Cur Loss: 0.20126487, Cur Avg Loss: 0.13271552, Log Avg loss: 0.12648036, Global Avg Loss: 0.60881128, Time: 0.0219 Steps: 104900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000589, Sample Num: 9424, Cur Loss: 0.18104225, Cur Avg Loss: 0.13330574, Log Avg loss: 0.16747948, Global Avg Loss: 0.60876922, Time: 0.0211 Steps: 104910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000599, Sample Num: 9584, Cur Loss: 0.15068629, Cur Avg Loss: 0.13390606, Log Avg loss: 0.16926457, Global Avg Loss: 0.60872733, Time: 0.0219 Steps: 104920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000609, Sample Num: 9744, Cur Loss: 0.09325945, Cur Avg Loss: 0.13418200, Log Avg loss: 0.15071083, Global Avg Loss: 0.60868368, Time: 0.0211 Steps: 104930, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000619, Sample Num: 9904, Cur Loss: 0.10825507, Cur Avg Loss: 0.13444626, Log Avg loss: 0.15053969, Global Avg Loss: 0.60864002, Time: 0.0219 Steps: 104940, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000629, Sample Num: 10064, Cur Loss: 0.11426871, Cur Avg Loss: 0.13446802, Log Avg loss: 0.13581538, Global Avg Loss: 0.60859497, Time: 0.0211 Steps: 104950, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000639, Sample Num: 10224, Cur Loss: 0.30022797, Cur Avg Loss: 0.13500651, Log Avg loss: 0.16887724, Global Avg Loss: 0.60855307, Time: 0.0220 Steps: 104960, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000649, Sample Num: 10384, Cur Loss: 0.06487142, Cur Avg Loss: 0.13530902, Log Avg loss: 0.15463956, Global Avg Loss: 0.60850983, Time: 0.0211 Steps: 104970, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000659, Sample Num: 10544, Cur Loss: 0.12445225, Cur Avg Loss: 0.13551852, Log Avg loss: 0.14911472, Global Avg Loss: 0.60846607, Time: 0.0211 Steps: 104980, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000669, Sample Num: 10704, Cur Loss: 0.11560710, Cur Avg Loss: 0.13570178, Log Avg loss: 0.14777910, Global Avg Loss: 0.60842219, Time: 0.0211 Steps: 104990, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000679, Sample Num: 10864, Cur Loss: 0.11127136, Cur Avg Loss: 0.13559129, Log Avg loss: 0.12819898, Global Avg Loss: 0.60837646, Time: 0.0211 Steps: 105000, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000689, Sample Num: 11024, Cur Loss: 0.39938799, Cur Avg Loss: 0.13639054, Log Avg loss: 0.19065985, Global Avg Loss: 0.60833668, Time: 0.0212 Steps: 105010, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000699, Sample Num: 11184, Cur Loss: 0.01327520, Cur Avg Loss: 0.13540094, Log Avg loss: 0.06721751, Global Avg Loss: 0.60828515, Time: 0.0219 Steps: 105020, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000709, Sample Num: 11344, Cur Loss: 0.05699446, Cur Avg Loss: 0.13536472, Log Avg loss: 0.13283330, Global Avg Loss: 0.60823988, Time: 0.0211 Steps: 105030, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000719, Sample Num: 11504, Cur Loss: 0.08066835, Cur Avg Loss: 0.13520570, Log Avg loss: 0.12393119, Global Avg Loss: 0.60819378, Time: 0.0219 Steps: 105040, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000729, Sample Num: 11664, Cur Loss: 0.07012380, Cur Avg Loss: 0.13556177, Log Avg loss: 0.16116328, Global Avg Loss: 0.60815122, Time: 0.0211 Steps: 105050, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000739, Sample Num: 11824, Cur Loss: 0.10631059, Cur Avg Loss: 0.13488820, Log Avg loss: 0.08578480, Global Avg Loss: 0.60810150, Time: 0.0219 Steps: 105060, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000749, Sample Num: 11984, Cur Loss: 0.15726146, Cur Avg Loss: 0.13449252, Log Avg loss: 0.10525137, Global Avg Loss: 0.60805364, Time: 0.0211 Steps: 105070, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000759, Sample Num: 12144, Cur Loss: 0.08558165, Cur Avg Loss: 0.13415879, Log Avg loss: 0.10916243, Global Avg Loss: 0.60800617, Time: 0.0220 Steps: 105080, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000769, Sample Num: 12304, Cur Loss: 0.05314992, Cur Avg Loss: 0.13395515, Log Avg loss: 0.11849928, Global Avg Loss: 0.60795959, Time: 0.0247 Steps: 105090, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000779, Sample Num: 12464, Cur Loss: 0.30078766, Cur Avg Loss: 0.13532536, Log Avg loss: 0.24069436, Global Avg Loss: 0.60792464, Time: 0.0210 Steps: 105100, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000789, Sample Num: 12624, Cur Loss: 0.13337225, Cur Avg Loss: 0.13547101, Log Avg loss: 0.14681727, Global Avg Loss: 0.60788077, Time: 0.0210 Steps: 105110, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000799, Sample Num: 12784, Cur Loss: 0.09363341, Cur Avg Loss: 0.13512839, Log Avg loss: 0.10809523, Global Avg Loss: 0.60783323, Time: 0.0210 Steps: 105120, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000809, Sample Num: 12944, Cur Loss: 0.07318828, Cur Avg Loss: 0.13503386, Log Avg loss: 0.12748133, Global Avg Loss: 0.60778754, Time: 0.0210 Steps: 105130, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000819, Sample Num: 13104, Cur Loss: 0.06752174, Cur Avg Loss: 0.13560994, Log Avg loss: 0.18221502, Global Avg Loss: 0.60774706, Time: 0.0210 Steps: 105140, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000829, Sample Num: 13264, Cur Loss: 0.10270697, Cur Avg Loss: 0.13580254, Log Avg loss: 0.15157621, Global Avg Loss: 0.60770368, Time: 0.0210 Steps: 105150, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000839, Sample Num: 13424, Cur Loss: 0.04899987, Cur Avg Loss: 0.13609544, Log Avg loss: 0.16037672, Global Avg Loss: 0.60766114, Time: 0.0210 Steps: 105160, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000849, Sample Num: 13584, Cur Loss: 0.23713961, Cur Avg Loss: 0.13597271, Log Avg loss: 0.12567593, Global Avg Loss: 0.60761531, Time: 0.0210 Steps: 105170, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000859, Sample Num: 13744, Cur Loss: 0.09780146, Cur Avg Loss: 0.13580003, Log Avg loss: 0.12113919, Global Avg Loss: 0.60756906, Time: 0.0209 Steps: 105180, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000869, Sample Num: 13904, Cur Loss: 0.17015883, Cur Avg Loss: 0.13575939, Log Avg loss: 0.13226813, Global Avg Loss: 0.60752387, Time: 0.0210 Steps: 105190, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000879, Sample Num: 14064, Cur Loss: 0.04169209, Cur Avg Loss: 0.13555281, Log Avg loss: 0.11760105, Global Avg Loss: 0.60747730, Time: 0.0210 Steps: 105200, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000889, Sample Num: 14224, Cur Loss: 0.05445450, Cur Avg Loss: 0.13597891, Log Avg loss: 0.17343334, Global Avg Loss: 0.60743605, Time: 0.0210 Steps: 105210, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000899, Sample Num: 14384, Cur Loss: 0.08879285, Cur Avg Loss: 0.13582371, Log Avg loss: 0.12202643, Global Avg Loss: 0.60738992, Time: 0.0211 Steps: 105220, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000909, Sample Num: 14544, Cur Loss: 0.39938620, Cur Avg Loss: 0.13618825, Log Avg loss: 0.16896055, Global Avg Loss: 0.60734825, Time: 0.0210 Steps: 105230, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000919, Sample Num: 14704, Cur Loss: 0.11949722, Cur Avg Loss: 0.13622970, Log Avg loss: 0.13999751, Global Avg Loss: 0.60730384, Time: 0.0210 Steps: 105240, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000929, Sample Num: 14864, Cur Loss: 0.08402520, Cur Avg Loss: 0.13624459, Log Avg loss: 0.13761251, Global Avg Loss: 0.60725922, Time: 0.0209 Steps: 105250, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000939, Sample Num: 15024, Cur Loss: 0.14729843, Cur Avg Loss: 0.13580474, Log Avg loss: 0.09494324, Global Avg Loss: 0.60721055, Time: 0.0209 Steps: 105260, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000949, Sample Num: 15184, Cur Loss: 0.10930596, Cur Avg Loss: 0.13556892, Log Avg loss: 0.11342541, Global Avg Loss: 0.60716364, Time: 0.0210 Steps: 105270, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000959, Sample Num: 15344, Cur Loss: 0.26509380, Cur Avg Loss: 0.13565634, Log Avg loss: 0.14395190, Global Avg Loss: 0.60711964, Time: 0.0210 Steps: 105280, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000969, Sample Num: 15504, Cur Loss: 0.13443908, Cur Avg Loss: 0.13549130, Log Avg loss: 0.11966423, Global Avg Loss: 0.60707334, Time: 0.0210 Steps: 105290, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000979, Sample Num: 15664, Cur Loss: 0.13882074, Cur Avg Loss: 0.13598780, Log Avg loss: 0.18409857, Global Avg Loss: 0.60703318, Time: 0.0210 Steps: 105300, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000989, Sample Num: 15824, Cur Loss: 0.21311642, Cur Avg Loss: 0.13610681, Log Avg loss: 0.14775815, Global Avg Loss: 0.60698956, Time: 0.0210 Steps: 105310, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000999, Sample Num: 15984, Cur Loss: 0.03866763, Cur Avg Loss: 0.13554960, Log Avg loss: 0.08044198, Global Avg Loss: 0.60693957, Time: 0.0210 Steps: 105320, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001009, Sample Num: 16144, Cur Loss: 0.18850353, Cur Avg Loss: 0.13524584, Log Avg loss: 0.10489969, Global Avg Loss: 0.60689191, Time: 0.0210 Steps: 105330, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001019, Sample Num: 16304, Cur Loss: 0.06525554, Cur Avg Loss: 0.13515466, Log Avg loss: 0.12595480, Global Avg Loss: 0.60684625, Time: 0.0210 Steps: 105340, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001029, Sample Num: 16464, Cur Loss: 0.08220091, Cur Avg Loss: 0.13501198, Log Avg loss: 0.12047248, Global Avg Loss: 0.60680008, Time: 0.0246 Steps: 105350, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001039, Sample Num: 16624, Cur Loss: 0.14359979, Cur Avg Loss: 0.13517202, Log Avg loss: 0.15164075, Global Avg Loss: 0.60675688, Time: 0.0208 Steps: 105360, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001049, Sample Num: 16784, Cur Loss: 0.08183949, Cur Avg Loss: 0.13509834, Log Avg loss: 0.12744267, Global Avg Loss: 0.60671139, Time: 0.0208 Steps: 105370, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001059, Sample Num: 16944, Cur Loss: 0.03350603, Cur Avg Loss: 0.13533451, Log Avg loss: 0.16010899, Global Avg Loss: 0.60666901, Time: 0.0209 Steps: 105380, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001069, Sample Num: 17104, Cur Loss: 0.18128631, Cur Avg Loss: 0.13546487, Log Avg loss: 0.14926967, Global Avg Loss: 0.60662561, Time: 0.0208 Steps: 105390, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001079, Sample Num: 17264, Cur Loss: 0.06566179, Cur Avg Loss: 0.13570174, Log Avg loss: 0.16102371, Global Avg Loss: 0.60658334, Time: 0.0208 Steps: 105400, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001089, Sample Num: 17424, Cur Loss: 0.03676849, Cur Avg Loss: 0.13544479, Log Avg loss: 0.10771974, Global Avg Loss: 0.60653601, Time: 0.0208 Steps: 105410, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001099, Sample Num: 17584, Cur Loss: 0.11031330, Cur Avg Loss: 0.13506227, Log Avg loss: 0.09340599, Global Avg Loss: 0.60648733, Time: 0.0208 Steps: 105420, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001109, Sample Num: 17744, Cur Loss: 0.10387132, Cur Avg Loss: 0.13498622, Log Avg loss: 0.12662778, Global Avg Loss: 0.60644182, Time: 0.0208 Steps: 105430, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001119, Sample Num: 17904, Cur Loss: 0.15108587, Cur Avg Loss: 0.13481085, Log Avg loss: 0.11536227, Global Avg Loss: 0.60639525, Time: 0.0208 Steps: 105440, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001129, Sample Num: 18064, Cur Loss: 0.10236330, Cur Avg Loss: 0.13502989, Log Avg loss: 0.15954087, Global Avg Loss: 0.60635287, Time: 0.0208 Steps: 105450, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001139, Sample Num: 18224, Cur Loss: 0.16319847, Cur Avg Loss: 0.13543930, Log Avg loss: 0.18166092, Global Avg Loss: 0.60631260, Time: 0.0208 Steps: 105460, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001149, Sample Num: 18384, Cur Loss: 0.05514686, Cur Avg Loss: 0.13495118, Log Avg loss: 0.07935488, Global Avg Loss: 0.60626264, Time: 0.0208 Steps: 105470, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001159, Sample Num: 18544, Cur Loss: 0.05183540, Cur Avg Loss: 0.13498006, Log Avg loss: 0.13829786, Global Avg Loss: 0.60621827, Time: 0.0208 Steps: 105480, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001169, Sample Num: 18704, Cur Loss: 0.27660421, Cur Avg Loss: 0.13481312, Log Avg loss: 0.11546457, Global Avg Loss: 0.60617175, Time: 0.0208 Steps: 105490, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001179, Sample Num: 18864, Cur Loss: 0.11660720, Cur Avg Loss: 0.13468276, Log Avg loss: 0.11944443, Global Avg Loss: 0.60612561, Time: 0.0208 Steps: 105500, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001189, Sample Num: 19024, Cur Loss: 0.19433485, Cur Avg Loss: 0.13480895, Log Avg loss: 0.14968650, Global Avg Loss: 0.60608235, Time: 0.0208 Steps: 105510, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001199, Sample Num: 19184, Cur Loss: 0.22162397, Cur Avg Loss: 0.13513984, Log Avg loss: 0.17448293, Global Avg Loss: 0.60604145, Time: 0.0208 Steps: 105520, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001209, Sample Num: 19344, Cur Loss: 0.13914432, Cur Avg Loss: 0.13541111, Log Avg loss: 0.16793651, Global Avg Loss: 0.60599994, Time: 0.0208 Steps: 105530, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001219, Sample Num: 19504, Cur Loss: 0.24213964, Cur Avg Loss: 0.13551376, Log Avg loss: 0.14792372, Global Avg Loss: 0.60595653, Time: 0.0208 Steps: 105540, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001229, Sample Num: 19664, Cur Loss: 0.10269779, Cur Avg Loss: 0.13576405, Log Avg loss: 0.16627441, Global Avg Loss: 0.60591488, Time: 0.0208 Steps: 105550, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001239, Sample Num: 19824, Cur Loss: 0.19631840, Cur Avg Loss: 0.13575524, Log Avg loss: 0.13467287, Global Avg Loss: 0.60587024, Time: 0.0208 Steps: 105560, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001249, Sample Num: 19984, Cur Loss: 0.23761956, Cur Avg Loss: 0.13565313, Log Avg loss: 0.12300103, Global Avg Loss: 0.60582450, Time: 0.0208 Steps: 105570, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001259, Sample Num: 20144, Cur Loss: 0.24540359, Cur Avg Loss: 0.13581341, Log Avg loss: 0.15583214, Global Avg Loss: 0.60578188, Time: 0.0208 Steps: 105580, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001269, Sample Num: 20304, Cur Loss: 0.21280092, Cur Avg Loss: 0.13675495, Log Avg loss: 0.25529504, Global Avg Loss: 0.60574868, Time: 0.0208 Steps: 105590, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001279, Sample Num: 20464, Cur Loss: 0.05673563, Cur Avg Loss: 0.13689232, Log Avg loss: 0.15432502, Global Avg Loss: 0.60570593, Time: 0.0208 Steps: 105600, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001289, Sample Num: 20624, Cur Loss: 0.32382005, Cur Avg Loss: 0.13682983, Log Avg loss: 0.12883787, Global Avg Loss: 0.60566078, Time: 0.0210 Steps: 105610, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001299, Sample Num: 20784, Cur Loss: 0.06076341, Cur Avg Loss: 0.13696255, Log Avg loss: 0.15406986, Global Avg Loss: 0.60561802, Time: 0.0209 Steps: 105620, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001309, Sample Num: 20944, Cur Loss: 0.16227113, Cur Avg Loss: 0.13671202, Log Avg loss: 0.10416729, Global Avg Loss: 0.60557055, Time: 0.0209 Steps: 105630, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001319, Sample Num: 21104, Cur Loss: 0.06935641, Cur Avg Loss: 0.13649631, Log Avg loss: 0.10826073, Global Avg Loss: 0.60552348, Time: 0.0209 Steps: 105640, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001329, Sample Num: 21264, Cur Loss: 0.09606382, Cur Avg Loss: 0.13655665, Log Avg loss: 0.14451555, Global Avg Loss: 0.60547984, Time: 0.0209 Steps: 105650, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001339, Sample Num: 21424, Cur Loss: 0.27485245, Cur Avg Loss: 0.13675652, Log Avg loss: 0.16331892, Global Avg Loss: 0.60543799, Time: 0.0209 Steps: 105660, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001349, Sample Num: 21584, Cur Loss: 0.03081033, Cur Avg Loss: 0.13672932, Log Avg loss: 0.13308728, Global Avg Loss: 0.60539329, Time: 0.0209 Steps: 105670, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001359, Sample Num: 21744, Cur Loss: 0.34188175, Cur Avg Loss: 0.13688653, Log Avg loss: 0.15809450, Global Avg Loss: 0.60535097, Time: 0.0209 Steps: 105680, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001369, Sample Num: 21904, Cur Loss: 0.23255336, Cur Avg Loss: 0.13666007, Log Avg loss: 0.10588424, Global Avg Loss: 0.60530371, Time: 0.0210 Steps: 105690, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001379, Sample Num: 22064, Cur Loss: 0.05988256, Cur Avg Loss: 0.13663599, Log Avg loss: 0.13333907, Global Avg Loss: 0.60525906, Time: 0.0209 Steps: 105700, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001389, Sample Num: 22224, Cur Loss: 0.10056689, Cur Avg Loss: 0.13675660, Log Avg loss: 0.15338896, Global Avg Loss: 0.60521631, Time: 0.0209 Steps: 105710, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001399, Sample Num: 22384, Cur Loss: 0.01571901, Cur Avg Loss: 0.13673373, Log Avg loss: 0.13355616, Global Avg Loss: 0.60517170, Time: 0.0209 Steps: 105720, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001409, Sample Num: 22544, Cur Loss: 0.34613407, Cur Avg Loss: 0.13677682, Log Avg loss: 0.14280603, Global Avg Loss: 0.60512797, Time: 0.0209 Steps: 105730, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001419, Sample Num: 22704, Cur Loss: 0.19381085, Cur Avg Loss: 0.13684032, Log Avg loss: 0.14578736, Global Avg Loss: 0.60508453, Time: 0.0209 Steps: 105740, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001429, Sample Num: 22864, Cur Loss: 0.05976991, Cur Avg Loss: 0.13668128, Log Avg loss: 0.11411345, Global Avg Loss: 0.60503810, Time: 0.0209 Steps: 105750, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001439, Sample Num: 23024, Cur Loss: 0.09155321, Cur Avg Loss: 0.13650912, Log Avg loss: 0.11190772, Global Avg Loss: 0.60499147, Time: 0.0209 Steps: 105760, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001449, Sample Num: 23184, Cur Loss: 0.14127271, Cur Avg Loss: 0.13664331, Log Avg loss: 0.15595290, Global Avg Loss: 0.60494902, Time: 0.0209 Steps: 105770, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001459, Sample Num: 23344, Cur Loss: 0.13266604, Cur Avg Loss: 0.13696719, Log Avg loss: 0.18389809, Global Avg Loss: 0.60490921, Time: 0.0208 Steps: 105780, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001469, Sample Num: 23504, Cur Loss: 0.04220285, Cur Avg Loss: 0.13707955, Log Avg loss: 0.15347188, Global Avg Loss: 0.60486654, Time: 0.0209 Steps: 105790, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001479, Sample Num: 23664, Cur Loss: 0.32298625, Cur Avg Loss: 0.13751013, Log Avg loss: 0.20076312, Global Avg Loss: 0.60482834, Time: 0.0209 Steps: 105800, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001489, Sample Num: 23824, Cur Loss: 0.15517017, Cur Avg Loss: 0.13756207, Log Avg loss: 0.14524299, Global Avg Loss: 0.60478491, Time: 0.0209 Steps: 105810, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001499, Sample Num: 23984, Cur Loss: 0.14114816, Cur Avg Loss: 0.13748822, Log Avg loss: 0.12649240, Global Avg Loss: 0.60473971, Time: 0.0209 Steps: 105820, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001509, Sample Num: 24144, Cur Loss: 0.05983706, Cur Avg Loss: 0.13754158, Log Avg loss: 0.14553975, Global Avg Loss: 0.60469632, Time: 0.0208 Steps: 105830, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001519, Sample Num: 24304, Cur Loss: 0.18671495, Cur Avg Loss: 0.13775577, Log Avg loss: 0.17007728, Global Avg Loss: 0.60465526, Time: 0.0208 Steps: 105840, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001529, Sample Num: 24464, Cur Loss: 0.03803273, Cur Avg Loss: 0.13793382, Log Avg loss: 0.16497995, Global Avg Loss: 0.60461372, Time: 0.0208 Steps: 105850, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001539, Sample Num: 24624, Cur Loss: 0.06706753, Cur Avg Loss: 0.13782567, Log Avg loss: 0.12128926, Global Avg Loss: 0.60456806, Time: 0.0245 Steps: 105860, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001549, Sample Num: 24784, Cur Loss: 0.10285580, Cur Avg Loss: 0.13800272, Log Avg loss: 0.16525021, Global Avg Loss: 0.60452657, Time: 0.0208 Steps: 105870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001559, Sample Num: 24944, Cur Loss: 0.18895555, Cur Avg Loss: 0.13792723, Log Avg loss: 0.12623499, Global Avg Loss: 0.60448139, Time: 0.0208 Steps: 105880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001569, Sample Num: 25104, Cur Loss: 0.13620786, Cur Avg Loss: 0.13778374, Log Avg loss: 0.11541370, Global Avg Loss: 0.60443521, Time: 0.0210 Steps: 105890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001579, Sample Num: 25264, Cur Loss: 0.17791824, Cur Avg Loss: 0.13773289, Log Avg loss: 0.12975381, Global Avg Loss: 0.60439038, Time: 0.0208 Steps: 105900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001589, Sample Num: 25424, Cur Loss: 0.11765163, Cur Avg Loss: 0.13771913, Log Avg loss: 0.13554724, Global Avg Loss: 0.60434612, Time: 0.0208 Steps: 105910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001599, Sample Num: 25584, Cur Loss: 0.07076227, Cur Avg Loss: 0.13794587, Log Avg loss: 0.17397383, Global Avg Loss: 0.60430548, Time: 0.0208 Steps: 105920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001609, Sample Num: 25744, Cur Loss: 0.06686329, Cur Avg Loss: 0.13775865, Log Avg loss: 0.10782341, Global Avg Loss: 0.60425861, Time: 0.0208 Steps: 105930, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001619, Sample Num: 25904, Cur Loss: 0.03464298, Cur Avg Loss: 0.13769285, Log Avg loss: 0.12710424, Global Avg Loss: 0.60421357, Time: 0.0208 Steps: 105940, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001629, Sample Num: 26064, Cur Loss: 0.09932156, Cur Avg Loss: 0.13757091, Log Avg loss: 0.11782964, Global Avg Loss: 0.60416767, Time: 0.0208 Steps: 105950, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001639, Sample Num: 26224, Cur Loss: 0.06674962, Cur Avg Loss: 0.13781610, Log Avg loss: 0.17775773, Global Avg Loss: 0.60412743, Time: 0.0208 Steps: 105960, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001649, Sample Num: 26384, Cur Loss: 0.13892817, Cur Avg Loss: 0.13773011, Log Avg loss: 0.12363666, Global Avg Loss: 0.60408208, Time: 0.0210 Steps: 105970, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001659, Sample Num: 26544, Cur Loss: 0.09813079, Cur Avg Loss: 0.13758663, Log Avg loss: 0.11392579, Global Avg Loss: 0.60403583, Time: 0.0208 Steps: 105980, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001669, Sample Num: 26704, Cur Loss: 0.10859416, Cur Avg Loss: 0.13732836, Log Avg loss: 0.09448179, Global Avg Loss: 0.60398776, Time: 0.0208 Steps: 105990, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001679, Sample Num: 26864, Cur Loss: 0.10854535, Cur Avg Loss: 0.13750854, Log Avg loss: 0.16758002, Global Avg Loss: 0.60394659, Time: 0.0208 Steps: 106000, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001689, Sample Num: 27024, Cur Loss: 0.09998333, Cur Avg Loss: 0.13727668, Log Avg loss: 0.09834829, Global Avg Loss: 0.60389889, Time: 0.0208 Steps: 106010, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001699, Sample Num: 27184, Cur Loss: 0.13191378, Cur Avg Loss: 0.13734387, Log Avg loss: 0.14869200, Global Avg Loss: 0.60385596, Time: 0.0208 Steps: 106020, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001709, Sample Num: 27344, Cur Loss: 0.03485086, Cur Avg Loss: 0.13711691, Log Avg loss: 0.09855630, Global Avg Loss: 0.60380830, Time: 0.0208 Steps: 106030, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001719, Sample Num: 27504, Cur Loss: 0.05595053, Cur Avg Loss: 0.13714190, Log Avg loss: 0.14141301, Global Avg Loss: 0.60376470, Time: 0.0208 Steps: 106040, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001729, Sample Num: 27664, Cur Loss: 0.05568597, Cur Avg Loss: 0.13711108, Log Avg loss: 0.13181211, Global Avg Loss: 0.60372019, Time: 0.0208 Steps: 106050, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001739, Sample Num: 27824, Cur Loss: 0.04441964, Cur Avg Loss: 0.13751687, Log Avg loss: 0.20767915, Global Avg Loss: 0.60368285, Time: 0.0209 Steps: 106060, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001749, Sample Num: 27984, Cur Loss: 0.03867394, Cur Avg Loss: 0.13712044, Log Avg loss: 0.06818069, Global Avg Loss: 0.60363237, Time: 0.0208 Steps: 106070, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001759, Sample Num: 28144, Cur Loss: 0.02106574, Cur Avg Loss: 0.13707749, Log Avg loss: 0.12956555, Global Avg Loss: 0.60358768, Time: 0.0208 Steps: 106080, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001769, Sample Num: 28304, Cur Loss: 0.22382551, Cur Avg Loss: 0.13736150, Log Avg loss: 0.18731966, Global Avg Loss: 0.60354844, Time: 0.0209 Steps: 106090, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001779, Sample Num: 28464, Cur Loss: 0.32999411, Cur Avg Loss: 0.13754653, Log Avg loss: 0.17027816, Global Avg Loss: 0.60350760, Time: 0.0208 Steps: 106100, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001789, Sample Num: 28624, Cur Loss: 0.13931933, Cur Avg Loss: 0.13766772, Log Avg loss: 0.15922687, Global Avg Loss: 0.60346573, Time: 0.0208 Steps: 106110, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001799, Sample Num: 28784, Cur Loss: 0.14328399, Cur Avg Loss: 0.13764808, Log Avg loss: 0.13413516, Global Avg Loss: 0.60342151, Time: 0.0209 Steps: 106120, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001809, Sample Num: 28944, Cur Loss: 0.08716052, Cur Avg Loss: 0.13758360, Log Avg loss: 0.12598201, Global Avg Loss: 0.60337652, Time: 0.0208 Steps: 106130, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001819, Sample Num: 29104, Cur Loss: 0.08203252, Cur Avg Loss: 0.13747458, Log Avg loss: 0.11775330, Global Avg Loss: 0.60333077, Time: 0.0208 Steps: 106140, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001829, Sample Num: 29264, Cur Loss: 0.03650095, Cur Avg Loss: 0.13735833, Log Avg loss: 0.11621373, Global Avg Loss: 0.60328488, Time: 0.0208 Steps: 106150, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001839, Sample Num: 29424, Cur Loss: 0.14575161, Cur Avg Loss: 0.13738416, Log Avg loss: 0.14210808, Global Avg Loss: 0.60324144, Time: 0.0208 Steps: 106160, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001849, Sample Num: 29584, Cur Loss: 0.13679007, Cur Avg Loss: 0.13738660, Log Avg loss: 0.13783545, Global Avg Loss: 0.60319760, Time: 0.0208 Steps: 106170, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001859, Sample Num: 29744, Cur Loss: 0.08952283, Cur Avg Loss: 0.13735302, Log Avg loss: 0.13114295, Global Avg Loss: 0.60315314, Time: 0.0209 Steps: 106180, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001869, Sample Num: 29904, Cur Loss: 0.03451242, Cur Avg Loss: 0.13730635, Log Avg loss: 0.12863049, Global Avg Loss: 0.60310846, Time: 0.0208 Steps: 106190, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001879, Sample Num: 30064, Cur Loss: 0.16985841, Cur Avg Loss: 0.13744094, Log Avg loss: 0.16259705, Global Avg Loss: 0.60306698, Time: 0.0208 Steps: 106200, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001889, Sample Num: 30224, Cur Loss: 0.07829629, Cur Avg Loss: 0.13817746, Log Avg loss: 0.27656828, Global Avg Loss: 0.60303624, Time: 0.0208 Steps: 106210, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001899, Sample Num: 30384, Cur Loss: 0.18413913, Cur Avg Loss: 0.13820069, Log Avg loss: 0.14259005, Global Avg Loss: 0.60299289, Time: 0.0208 Steps: 106220, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001909, Sample Num: 30544, Cur Loss: 0.07257956, Cur Avg Loss: 0.13820868, Log Avg loss: 0.13972604, Global Avg Loss: 0.60294928, Time: 0.0209 Steps: 106230, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001919, Sample Num: 30704, Cur Loss: 0.15512338, Cur Avg Loss: 0.13830012, Log Avg loss: 0.15575440, Global Avg Loss: 0.60290718, Time: 0.0208 Steps: 106240, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001929, Sample Num: 30864, Cur Loss: 0.15368620, Cur Avg Loss: 0.13836012, Log Avg loss: 0.14987468, Global Avg Loss: 0.60286455, Time: 0.0209 Steps: 106250, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001939, Sample Num: 31024, Cur Loss: 0.06072737, Cur Avg Loss: 0.13840427, Log Avg loss: 0.14692130, Global Avg Loss: 0.60282164, Time: 0.0208 Steps: 106260, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001949, Sample Num: 31184, Cur Loss: 0.19255802, Cur Avg Loss: 0.13845744, Log Avg loss: 0.14876739, Global Avg Loss: 0.60277891, Time: 0.0208 Steps: 106270, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001959, Sample Num: 31344, Cur Loss: 0.07418051, Cur Avg Loss: 0.13849989, Log Avg loss: 0.14677321, Global Avg Loss: 0.60273601, Time: 0.0209 Steps: 106280, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001969, Sample Num: 31504, Cur Loss: 0.12639716, Cur Avg Loss: 0.13852509, Log Avg loss: 0.14346085, Global Avg Loss: 0.60269280, Time: 0.0208 Steps: 106290, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001979, Sample Num: 31664, Cur Loss: 0.38110566, Cur Avg Loss: 0.13851541, Log Avg loss: 0.13660970, Global Avg Loss: 0.60264895, Time: 0.0209 Steps: 106300, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001989, Sample Num: 31824, Cur Loss: 0.09888779, Cur Avg Loss: 0.13848277, Log Avg loss: 0.13202376, Global Avg Loss: 0.60260468, Time: 0.0208 Steps: 106310, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001999, Sample Num: 31984, Cur Loss: 0.14920589, Cur Avg Loss: 0.13858981, Log Avg loss: 0.15988077, Global Avg Loss: 0.60256304, Time: 0.0208 Steps: 106320, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002009, Sample Num: 32144, Cur Loss: 0.08278586, Cur Avg Loss: 0.13875551, Log Avg loss: 0.17187725, Global Avg Loss: 0.60252254, Time: 0.0209 Steps: 106330, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002019, Sample Num: 32304, Cur Loss: 0.41497964, Cur Avg Loss: 0.13877909, Log Avg loss: 0.14351790, Global Avg Loss: 0.60247937, Time: 0.0208 Steps: 106340, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002029, Sample Num: 32464, Cur Loss: 0.17316715, Cur Avg Loss: 0.13914047, Log Avg loss: 0.21210290, Global Avg Loss: 0.60244266, Time: 0.0208 Steps: 106350, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002039, Sample Num: 32624, Cur Loss: 0.12592669, Cur Avg Loss: 0.13942165, Log Avg loss: 0.19647342, Global Avg Loss: 0.60240450, Time: 0.0208 Steps: 106360, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002049, Sample Num: 32784, Cur Loss: 0.11272836, Cur Avg Loss: 0.13967800, Log Avg loss: 0.19194644, Global Avg Loss: 0.60236591, Time: 0.0245 Steps: 106370, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002059, Sample Num: 32944, Cur Loss: 0.05114897, Cur Avg Loss: 0.13959520, Log Avg loss: 0.12263015, Global Avg Loss: 0.60232081, Time: 0.0207 Steps: 106380, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002069, Sample Num: 33104, Cur Loss: 0.04578269, Cur Avg Loss: 0.13946664, Log Avg loss: 0.11299665, Global Avg Loss: 0.60227482, Time: 0.0207 Steps: 106390, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002079, Sample Num: 33264, Cur Loss: 0.12459339, Cur Avg Loss: 0.13950352, Log Avg loss: 0.14713238, Global Avg Loss: 0.60223204, Time: 0.0208 Steps: 106400, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002089, Sample Num: 33424, Cur Loss: 0.11036767, Cur Avg Loss: 0.13929902, Log Avg loss: 0.09678499, Global Avg Loss: 0.60218454, Time: 0.0208 Steps: 106410, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002099, Sample Num: 33584, Cur Loss: 0.08860835, Cur Avg Loss: 0.13955107, Log Avg loss: 0.19220418, Global Avg Loss: 0.60214602, Time: 0.0208 Steps: 106420, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002109, Sample Num: 33744, Cur Loss: 0.05207320, Cur Avg Loss: 0.13926418, Log Avg loss: 0.07904647, Global Avg Loss: 0.60209687, Time: 0.0207 Steps: 106430, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002119, Sample Num: 33904, Cur Loss: 0.07894526, Cur Avg Loss: 0.13918469, Log Avg loss: 0.12241996, Global Avg Loss: 0.60205180, Time: 0.0208 Steps: 106440, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002129, Sample Num: 34055, Cur Loss: 0.27694044, Cur Avg Loss: 0.13916171, Log Avg loss: 0.13429090, Global Avg Loss: 0.60200786, Time: 0.0101 Steps: 106450, Updated lr: 0.000000 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.771940, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.233872, "eval_total_loss": 164.411929, "eval_mae": 0.320608, "eval_mse": 0.233961, "eval_r2": 0.851279, "eval_sp_statistic": 0.90504, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925612, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.740266, "test_total_loss": 371.613462, "test_mae": 0.666628, "test_mse": 0.74031, "test_r2": 0.522198, "test_sp_statistic": 0.799697, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.839953, "test_ps_pvalue": 0.0, "lr": 0.0, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6020078594979299, "train_cur_epoch_loss": 296.2752745319158, "train_cur_epoch_avg_loss": 0.13916170715449308, "train_cur_epoch_time": 44.771939754486084, "train_cur_epoch_avg_time": 0.021029563059880733, "epoch": 50, "step": 106450} ################################################## #########################Best Metric######################### {"epoch": 48, "global_step": 102192, "eval_avg_loss": 0.238188, "eval_total_loss": 167.446083, "eval_mae": 0.323544, "eval_mse": 0.238279, "eval_r2": 0.848534, "eval_sp_statistic": 0.905492, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925398, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.730546, "test_total_loss": 366.733903, "test_mae": 0.661817, "test_mse": 0.730596, "test_r2": 0.528467, "test_sp_statistic": 0.800977, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.840501, "test_ps_pvalue": 0.0} ################################################## Total Time: 40364.981279, Avg time per epoch(50 epochs): 807.300000 ++++++++++++Validation+++++++++++++ best sp_statistic global step: 102192 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226150046/checkpoint-102192 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## {"evaluation_avg_loss_102192": 0.238188, "evaluation_total_loss_102192": 167.446083, "evaluation_mae_102192": 0.323544, "evaluation_mse_102192": 0.238279, "evaluation_r2_102192": 0.848534, "evaluation_sp_statistic_102192": 0.905492, "evaluation_sp_pvalue_102192": 0.0, "evaluation_ps_statistic_102192": 0.925398, "evaluation_ps_pvalue_102192": 0.0} ++++++++++++Testing+++++++++++++ best sp_statistic global step: 102192 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250226150046/checkpoint-102192 ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## {"evaluation_avg_loss_102192": 0.730546, "evaluation_total_loss_102192": 366.733903, "evaluation_mae_102192": 0.661817, "evaluation_mse_102192": 0.730596, "evaluation_r2_102192": 0.528467, "evaluation_sp_statistic_102192": 0.800977, "evaluation_sp_pvalue_102192": 0.0, "evaluation_ps_statistic_102192": 0.840501, "evaluation_ps_pvalue_102192": 0.0}