{"adam_epsilon": 1e-08, "alphabet": "gene_prot", "append_eos": true, "asl_gamma_neg": 4.0, "asl_gamma_pos": 1.0, "best_metric_type": "sp_statistic", "beta1": 0.9, "beta2": 0.98, "buffer_size": 4096, "cache_dir": null, "classifier_activate_func": "gelu", "classifier_size": 128, "codes_file": null, "config_path": "../config/luca_base/luca_base_config.json", "cross_atten": false, "dataset_name": "DMS_Bind_Reps_Strain", "dataset_type": "protein", "delete_old": true, "dev_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/dev/", "device": "cuda", "do_eval": true, "do_lower_case": false, "do_metrics": true, "do_predict": true, "do_train": true, "dropout_prob": 0.1, "early_stop_epoch": -1, "emb_activate_func": "gelu", "embedding_complete": true, "embedding_complete_seg_overlap": true, "embedding_fixed_len_a_time": 3072, "embedding_input_size": 2560, "embedding_input_size_a": null, "embedding_input_size_b": null, "eval_all_checkpoints": false, "evaluate_during_training": true, "evaluate_steps": 2000, "evaluate_strategy": "epoch", "fc_activate_func": "gelu", "focal_loss_alpha": 0.7, "focal_loss_gamma": 2.0, "focal_loss_reduce": false, "fp16": false, "fp16_embedding": false, "fp16_opt_level": "O1", "fusion_type": "concat", "gradient_accumulation_steps": 1, "hidden_size": 1024, "ignore_index": -100, "input_mode": "single", "input_type": "matrix", "intermediate_size": 4096, "label_filepath": "../dataset/DMS_Bind_Reps_Strain/protein/regression/label.txt", "label_size": 1, "label_type": "DMS_Bind_Reps_Strain", "learning_rate": 0.0001, "llm_dir": "..", "llm_dirpath": "../llm/models/lucaone_virus/v1.0/token_level,span_level,seq_level/lucaone_virus/20240815023346/checkpoint-step3800000", "llm_step": "3800000", "llm_task_level": "token_level,span_level,seq_level", "llm_time_str": "20240815023346", "llm_type": "lucaone_virus", "llm_version": "v1.0", "lmdb_path": null, "local_rank": -1, "log_dir": "../logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250225170110", "logging_steps": 10, "loss_reduction": "mean", "loss_type": "l2", "lr_decay_rate": 0.9, "lr_update_strategy": "step", "matrix_add_special_token": false, "matrix_dirpath": "../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000", "matrix_embedding_exists": false, "matrix_encoder": false, "matrix_encoder_act": false, "matrix_fc_size": "128", "matrix_max_length": 100000, "matrix_max_length_a": null, "matrix_max_length_b": null, "matrix_pooling_type": "value_attention", "max_grad_norm": 1.0, "max_sentence_length": null, "max_sentences": null, "max_steps": -1, "model_dirpath": null, "model_type": "luca_base", "n_gpu": 1, "no_cuda": false, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "non_ignore": false, "not_append_eos": false, "not_matrix_encoder_shared": false, "not_prepend_bos": false, "not_save_emb_to_disk": false, "not_seq_encoder_shared": false, "num_attention_heads": 4, "num_hidden_layers": 2, "num_train_epochs": 50, "output_dir": "../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250225170110", "output_mode": "regression", "overwrite_cache": false, "overwrite_output_dir": true, "per_gpu_eval_batch_size": 16, "per_gpu_train_batch_size": 16, "pos_weight": 1.0, "position_embedding_type": "absolute", "prepend_bos": true, "save_all": false, "save_steps": -1, "seed": 1221, "self_atten": false, "seq_fc_size": "null", "seq_max_length": 100000, "seq_max_length_a": null, "seq_max_length_b": null, "seq_pooling_type": "value_attention", "seq_subword": false, "seq_vocab_path": "gene_prot", "sigmoid": false, "task_level_type": "seq_level", "task_type": "regression", "tb_log_dir": "../tb-logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250225170110", "test_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/test/", "time_str": "20250225170115", "train_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/train/", "trunc_type": "right", "vector_dirpath": "../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000", "vector_fc_size": "null", "vocab_size": 39, "warmup_steps": 1000, "weight": null, "weight_decay": 0.01, "worker_num": 0} ################################################## n_gpu: 1 ################################################## Inputs: Input Name List: protein,embedding_matrix ################################################## Encoder Config: {'llm_type': 'lucaone_virus', 'llm_version': 'v1.0', 'llm_step': '3800000', 'llm_dirpath': '../llm/models/lucaone_virus/v1.0/token_level,span_level,seq_level/lucaone_virus/20240815023346/checkpoint-step3800000', 'input_type': 'matrix', 'trunc_type': 'right', 'seq_max_length': 100000, 'atom_seq_max_length': None, 'vector_dirpath': '../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000', 'matrix_dirpath': '../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/v1.0/lucaone_virus/20240815023346/3800000', 'local_rank': -1, 'max_sentence_length': None, 'max_sentences': None, 'matrix_add_special_token': False, 'embedding_complete': True, 'embedding_complete_seg_overlap': True, 'embedding_fixed_len_a_time': 3072, 'matrix_embedding_exists': False, 'save_emb_to_disk': True, 'fp16_embedding': False} ################################################## Model Config: LucaConfig { "alphabet": "gene_prot", "attention_probs_dropout_prob": 0.1, "classifier_activate_func": "gelu", "classifier_dropout_prob": 0.1, "classifier_size": 128, "cls_token_id": 2, "cross_atten": false, "directionality": "bidi", "emb_activate_func": "gelu", "embedding_input_size": 2560, "fc_activate_func": "gelu", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "LABEL_0" }, "ignore_index": -100, "initializer_range": 0.02, "intermediate_size": 4096, "kernel_size": 7, "label2id": { "LABEL_0": 0 }, "layer_norm_eps": 1e-12, "loss_reduction": "mean", "matrix_fc_size": [ 128 ], "matrix_max_length": 100000, "matrix_pooling_type": "value_attention", "max_position_embeddings": 100002, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "num_attention_heads": 4, "num_hidden_layers": 2, "pad_token_id": 0, "pos_weight": 1.0, "position_embedding_type": "absolute", "self_atten": false, "sep_token_id": 3, "seq_fc_size": null, "seq_max_length": 100000, "seq_pooling_type": "value_attention", "token_dropout": null, "transformers_version": "4.29.0", "type_vocab_size": 2, "use_luca_layer_norm_v2": true, "vector_fc_size": null, "vocab_size": 39 } ################################################## Mode Architecture: LucaBase( (matrix_pooler): GlobalMaskValueAttentionPooling1D (2560 -> 2560) (linear): ModuleList( (0): ModuleList( (0): Linear(in_features=2560, out_features=128, bias=True) (1): GELU(approximate='none') ) ) (dropout): Dropout(p=0.1, inplace=False) (hidden_layer): Linear(in_features=128, out_features=128, bias=True) (hidden_act): GELU(approximate='none') (classifier): Linear(in_features=128, out_features=1, bias=True) (loss_fct): MaskedMSELoss( (criterion): MSELoss() ) ) ################################################## Model parameters: 20005249 ################################################## {"total_num": "19.080000M", "total_size": "76.310000MB", "param_sum": "19.080000M", "param_size": "76.310000MB", "buffer_sum": "0.000000M", "buffer_size": "0.000000MB", "trainable_num": "19.078492M", "trainable_size": "76.313969MB"} ################################################## Train dataset len: 34055, batch size: 16, batch num: 2129 Train dataset t_total: 106450, max_steps: -1 ***** Running training ***** Train Dataset Num examples = 34055 Train Dataset Num Epochs = 50 Logging Steps = 10 Saving Steps = -1 Evaluating Strategy = epoch Train Dataset Instantaneous batch size per GPU = 16 Train Dataset Total train batch size (w. parallel, distributed & accumulation) = 16 Train Dataset Gradient Accumulation steps = 1 Train Dataset Total optimization steps = 106450 ################################################## Training, Epoch: 0001, Batch: 000010, Sample Num: 160, Cur Loss: 70.87425995, Cur Avg Loss: 71.93774948, Log Avg loss: 71.93774948, Global Avg Loss: 71.93774948, Time: 0.0210 Steps: 10, Updated lr: 0.000001 Training, Epoch: 0001, Batch: 000020, Sample Num: 320, Cur Loss: 73.35147858, Cur Avg Loss: 71.41352882, Log Avg loss: 70.88930817, Global Avg Loss: 71.41352882, Time: 0.0210 Steps: 20, Updated lr: 0.000002 Training, Epoch: 0001, Batch: 000030, Sample Num: 480, Cur Loss: 73.01605225, Cur Avg Loss: 72.72826029, Log Avg loss: 75.35772324, Global Avg Loss: 72.72826029, Time: 0.0209 Steps: 30, Updated lr: 0.000003 Training, Epoch: 0001, Batch: 000040, Sample Num: 640, Cur Loss: 67.71244812, Cur Avg Loss: 72.20841360, Log Avg loss: 70.64887352, Global Avg Loss: 72.20841360, Time: 0.0221 Steps: 40, Updated lr: 0.000004 Training, Epoch: 0001, Batch: 000050, Sample Num: 800, Cur Loss: 63.21402359, Cur Avg Loss: 72.01252045, Log Avg loss: 71.22894783, Global Avg Loss: 72.01252045, Time: 0.0220 Steps: 50, Updated lr: 0.000005 Training, Epoch: 0001, Batch: 000060, Sample Num: 960, Cur Loss: 74.01074219, Cur Avg Loss: 71.94546356, Log Avg loss: 71.61017914, Global Avg Loss: 71.94546356, Time: 0.0212 Steps: 60, Updated lr: 0.000006 Training, Epoch: 0001, Batch: 000070, Sample Num: 1120, Cur Loss: 63.19598770, Cur Avg Loss: 71.60817582, Log Avg loss: 69.58444939, Global Avg Loss: 71.60817582, Time: 0.0215 Steps: 70, Updated lr: 0.000007 Training, Epoch: 0001, Batch: 000080, Sample Num: 1280, Cur Loss: 74.81593323, Cur Avg Loss: 71.21966558, Log Avg loss: 68.50009384, Global Avg Loss: 71.21966558, Time: 0.0213 Steps: 80, Updated lr: 0.000008 Training, Epoch: 0001, Batch: 000090, Sample Num: 1440, Cur Loss: 73.52436829, Cur Avg Loss: 70.86844491, Log Avg loss: 68.05867958, Global Avg Loss: 70.86844491, Time: 0.0213 Steps: 90, Updated lr: 0.000009 Training, Epoch: 0001, Batch: 000100, Sample Num: 1600, Cur Loss: 76.90646362, Cur Avg Loss: 71.08774261, Log Avg loss: 73.06142197, Global Avg Loss: 71.08774261, Time: 0.0214 Steps: 100, Updated lr: 0.000010 Training, Epoch: 0001, Batch: 000110, Sample Num: 1760, Cur Loss: 67.94900513, Cur Avg Loss: 71.09439614, Log Avg loss: 71.16093140, Global Avg Loss: 71.09439614, Time: 0.0212 Steps: 110, Updated lr: 0.000011 Training, Epoch: 0001, Batch: 000120, Sample Num: 1920, Cur Loss: 71.33995056, Cur Avg Loss: 71.22711703, Log Avg loss: 72.68704681, Global Avg Loss: 71.22711703, Time: 0.0214 Steps: 120, Updated lr: 0.000012 Training, Epoch: 0001, Batch: 000130, Sample Num: 2080, Cur Loss: 77.29939270, Cur Avg Loss: 71.35836798, Log Avg loss: 72.93337936, Global Avg Loss: 71.35836798, Time: 0.0224 Steps: 130, Updated lr: 0.000013 Training, Epoch: 0001, Batch: 000140, Sample Num: 2240, Cur Loss: 77.13848114, Cur Avg Loss: 71.46396196, Log Avg loss: 72.83668365, Global Avg Loss: 71.46396196, Time: 0.0210 Steps: 140, Updated lr: 0.000014 Training, Epoch: 0001, Batch: 000150, Sample Num: 2400, Cur Loss: 72.87939453, Cur Avg Loss: 71.70387477, Log Avg loss: 75.06265411, Global Avg Loss: 71.70387477, Time: 0.0210 Steps: 150, Updated lr: 0.000015 Training, Epoch: 0001, Batch: 000160, Sample Num: 2560, Cur Loss: 64.97949219, Cur Avg Loss: 71.58092887, Log Avg loss: 69.73674049, Global Avg Loss: 71.58092887, Time: 0.0209 Steps: 160, Updated lr: 0.000016 Training, Epoch: 0001, Batch: 000170, Sample Num: 2720, Cur Loss: 70.51624298, Cur Avg Loss: 71.45491667, Log Avg loss: 69.43872147, Global Avg Loss: 71.45491667, Time: 0.0209 Steps: 170, Updated lr: 0.000017 Training, Epoch: 0001, Batch: 000180, Sample Num: 2880, Cur Loss: 62.10330963, Cur Avg Loss: 71.40545478, Log Avg loss: 70.56460266, Global Avg Loss: 71.40545478, Time: 0.0210 Steps: 180, Updated lr: 0.000018 Training, Epoch: 0001, Batch: 000190, Sample Num: 3040, Cur Loss: 77.14030457, Cur Avg Loss: 71.42956690, Log Avg loss: 71.86358490, Global Avg Loss: 71.42956690, Time: 0.0209 Steps: 190, Updated lr: 0.000019 Training, Epoch: 0001, Batch: 000200, Sample Num: 3200, Cur Loss: 64.14593506, Cur Avg Loss: 71.39916370, Log Avg loss: 70.82150307, Global Avg Loss: 71.39916370, Time: 0.0208 Steps: 200, Updated lr: 0.000020 Training, Epoch: 0001, Batch: 000210, Sample Num: 3360, Cur Loss: 67.53603363, Cur Avg Loss: 71.29702334, Log Avg loss: 69.25421600, Global Avg Loss: 71.29702334, Time: 0.0209 Steps: 210, Updated lr: 0.000021 Training, Epoch: 0001, Batch: 000220, Sample Num: 3520, Cur Loss: 62.23637390, Cur Avg Loss: 71.14247294, Log Avg loss: 67.89691467, Global Avg Loss: 71.14247294, Time: 0.0210 Steps: 220, Updated lr: 0.000022 Training, Epoch: 0001, Batch: 000230, Sample Num: 3680, Cur Loss: 75.49690247, Cur Avg Loss: 70.94554208, Log Avg loss: 66.61306305, Global Avg Loss: 70.94554208, Time: 0.0219 Steps: 230, Updated lr: 0.000023 Training, Epoch: 0001, Batch: 000240, Sample Num: 3840, Cur Loss: 68.23613739, Cur Avg Loss: 70.94740686, Log Avg loss: 70.99029694, Global Avg Loss: 70.94740686, Time: 0.0209 Steps: 240, Updated lr: 0.000024 Training, Epoch: 0001, Batch: 000250, Sample Num: 4000, Cur Loss: 69.24428558, Cur Avg Loss: 70.81909026, Log Avg loss: 67.73949165, Global Avg Loss: 70.81909026, Time: 0.0210 Steps: 250, Updated lr: 0.000025 Training, Epoch: 0001, Batch: 000260, Sample Num: 4160, Cur Loss: 69.49398041, Cur Avg Loss: 70.65366818, Log Avg loss: 66.51811638, Global Avg Loss: 70.65366818, Time: 0.0209 Steps: 260, Updated lr: 0.000026 Training, Epoch: 0001, Batch: 000270, Sample Num: 4320, Cur Loss: 55.76964569, Cur Avg Loss: 70.41302476, Log Avg loss: 64.15629578, Global Avg Loss: 70.41302476, Time: 0.0209 Steps: 270, Updated lr: 0.000027 Training, Epoch: 0001, Batch: 000280, Sample Num: 4480, Cur Loss: 60.94628906, Cur Avg Loss: 70.20443735, Log Avg loss: 64.57257729, Global Avg Loss: 70.20443735, Time: 0.0209 Steps: 280, Updated lr: 0.000028 Training, Epoch: 0001, Batch: 000290, Sample Num: 4640, Cur Loss: 60.20637894, Cur Avg Loss: 69.97233208, Log Avg loss: 63.47338448, Global Avg Loss: 69.97233208, Time: 0.0209 Steps: 290, Updated lr: 0.000029 Training, Epoch: 0001, Batch: 000300, Sample Num: 4800, Cur Loss: 62.39556122, Cur Avg Loss: 69.67658802, Log Avg loss: 61.10001030, Global Avg Loss: 69.67658802, Time: 0.0209 Steps: 300, Updated lr: 0.000030 Training, Epoch: 0001, Batch: 000310, Sample Num: 4960, Cur Loss: 52.51202393, Cur Avg Loss: 69.29139466, Log Avg loss: 57.73559380, Global Avg Loss: 69.29139466, Time: 0.0209 Steps: 310, Updated lr: 0.000031 Training, Epoch: 0001, Batch: 000320, Sample Num: 5120, Cur Loss: 59.38465118, Cur Avg Loss: 68.86817616, Log Avg loss: 55.74840279, Global Avg Loss: 68.86817616, Time: 0.0209 Steps: 320, Updated lr: 0.000032 Training, Epoch: 0001, Batch: 000330, Sample Num: 5280, Cur Loss: 53.34571457, Cur Avg Loss: 68.34122054, Log Avg loss: 51.47864075, Global Avg Loss: 68.34122054, Time: 0.0209 Steps: 330, Updated lr: 0.000033 Training, Epoch: 0001, Batch: 000340, Sample Num: 5440, Cur Loss: 46.67948151, Cur Avg Loss: 67.77709211, Log Avg loss: 49.16085396, Global Avg Loss: 67.77709211, Time: 0.0209 Steps: 340, Updated lr: 0.000034 Training, Epoch: 0001, Batch: 000350, Sample Num: 5600, Cur Loss: 53.69550323, Cur Avg Loss: 67.24152254, Log Avg loss: 49.03215714, Global Avg Loss: 67.24152254, Time: 0.0209 Steps: 350, Updated lr: 0.000035 Training, Epoch: 0001, Batch: 000360, Sample Num: 5760, Cur Loss: 40.16608429, Cur Avg Loss: 66.59240216, Log Avg loss: 43.87318878, Global Avg Loss: 66.59240216, Time: 0.0208 Steps: 360, Updated lr: 0.000036 Training, Epoch: 0001, Batch: 000370, Sample Num: 5920, Cur Loss: 35.11378860, Cur Avg Loss: 65.81213236, Log Avg loss: 37.72241936, Global Avg Loss: 65.81213236, Time: 0.0208 Steps: 370, Updated lr: 0.000037 Training, Epoch: 0001, Batch: 000380, Sample Num: 6080, Cur Loss: 31.81513596, Cur Avg Loss: 64.96973624, Log Avg loss: 33.80108013, Global Avg Loss: 64.96973624, Time: 0.0208 Steps: 380, Updated lr: 0.000038 Training, Epoch: 0001, Batch: 000390, Sample Num: 6240, Cur Loss: 22.37768555, Cur Avg Loss: 64.05701367, Log Avg loss: 29.37355576, Global Avg Loss: 64.05701367, Time: 0.0208 Steps: 390, Updated lr: 0.000039 Training, Epoch: 0001, Batch: 000400, Sample Num: 6400, Cur Loss: 23.55140305, Cur Avg Loss: 63.08288266, Log Avg loss: 25.09177322, Global Avg Loss: 63.08288266, Time: 0.0209 Steps: 400, Updated lr: 0.000040 Training, Epoch: 0001, Batch: 000410, Sample Num: 6560, Cur Loss: 20.47164154, Cur Avg Loss: 62.06044389, Log Avg loss: 21.16289330, Global Avg Loss: 62.06044389, Time: 0.0208 Steps: 410, Updated lr: 0.000041 Training, Epoch: 0001, Batch: 000420, Sample Num: 6720, Cur Loss: 14.76172924, Cur Avg Loss: 60.97946914, Log Avg loss: 16.65950451, Global Avg Loss: 60.97946914, Time: 0.0208 Steps: 420, Updated lr: 0.000042 Training, Epoch: 0001, Batch: 000430, Sample Num: 6880, Cur Loss: 8.78057861, Cur Avg Loss: 59.84052330, Log Avg loss: 12.00479774, Global Avg Loss: 59.84052330, Time: 0.0209 Steps: 430, Updated lr: 0.000043 Training, Epoch: 0001, Batch: 000440, Sample Num: 7040, Cur Loss: 8.46424103, Cur Avg Loss: 58.66674113, Log Avg loss: 8.19410772, Global Avg Loss: 58.66674113, Time: 0.0209 Steps: 440, Updated lr: 0.000044 Training, Epoch: 0001, Batch: 000450, Sample Num: 7200, Cur Loss: 3.20476437, Cur Avg Loss: 57.47588630, Log Avg loss: 5.07827413, Global Avg Loss: 57.47588630, Time: 0.0208 Steps: 450, Updated lr: 0.000045 Training, Epoch: 0001, Batch: 000460, Sample Num: 7360, Cur Loss: 2.98985910, Cur Avg Loss: 56.29564776, Log Avg loss: 3.18491319, Global Avg Loss: 56.29564776, Time: 0.0209 Steps: 460, Updated lr: 0.000046 Training, Epoch: 0001, Batch: 000470, Sample Num: 7520, Cur Loss: 1.23915029, Cur Avg Loss: 55.14362876, Log Avg loss: 2.15075462, Global Avg Loss: 55.14362876, Time: 0.0208 Steps: 470, Updated lr: 0.000047 Training, Epoch: 0001, Batch: 000480, Sample Num: 7680, Cur Loss: 2.90078497, Cur Avg Loss: 54.04458692, Log Avg loss: 2.38962075, Global Avg Loss: 54.04458692, Time: 0.0208 Steps: 480, Updated lr: 0.000048 Training, Epoch: 0001, Batch: 000490, Sample Num: 7840, Cur Loss: 1.38937938, Cur Avg Loss: 52.97733692, Log Avg loss: 1.74933676, Global Avg Loss: 52.97733692, Time: 0.0208 Steps: 490, Updated lr: 0.000049 Training, Epoch: 0001, Batch: 000500, Sample Num: 8000, Cur Loss: 1.35306799, Cur Avg Loss: 51.95847494, Log Avg loss: 2.03423799, Global Avg Loss: 51.95847494, Time: 0.0208 Steps: 500, Updated lr: 0.000050 Training, Epoch: 0001, Batch: 000510, Sample Num: 8160, Cur Loss: 1.73994744, Cur Avg Loss: 50.98209959, Log Avg loss: 2.16333207, Global Avg Loss: 50.98209959, Time: 0.0208 Steps: 510, Updated lr: 0.000051 Training, Epoch: 0001, Batch: 000520, Sample Num: 8320, Cur Loss: 1.88104701, Cur Avg Loss: 50.03403362, Log Avg loss: 1.68266929, Global Avg Loss: 50.03403362, Time: 0.0209 Steps: 520, Updated lr: 0.000052 Training, Epoch: 0001, Batch: 000530, Sample Num: 8480, Cur Loss: 1.13197398, Cur Avg Loss: 49.12265823, Log Avg loss: 1.73113810, Global Avg Loss: 49.12265823, Time: 0.0212 Steps: 530, Updated lr: 0.000053 Training, Epoch: 0001, Batch: 000540, Sample Num: 8640, Cur Loss: 1.33055043, Cur Avg Loss: 48.24332625, Log Avg loss: 1.63873126, Global Avg Loss: 48.24332625, Time: 0.0209 Steps: 540, Updated lr: 0.000054 Training, Epoch: 0001, Batch: 000550, Sample Num: 8800, Cur Loss: 1.97058046, Cur Avg Loss: 47.40179312, Log Avg loss: 1.95900366, Global Avg Loss: 47.40179312, Time: 0.0208 Steps: 550, Updated lr: 0.000055 Training, Epoch: 0001, Batch: 000560, Sample Num: 8960, Cur Loss: 1.47739530, Cur Avg Loss: 46.58136007, Log Avg loss: 1.45754269, Global Avg Loss: 46.58136007, Time: 0.0209 Steps: 560, Updated lr: 0.000056 Training, Epoch: 0001, Batch: 000570, Sample Num: 9120, Cur Loss: 0.93693078, Cur Avg Loss: 45.79675206, Log Avg loss: 1.85870319, Global Avg Loss: 45.79675206, Time: 0.0209 Steps: 570, Updated lr: 0.000057 Training, Epoch: 0001, Batch: 000580, Sample Num: 9280, Cur Loss: 1.08992457, Cur Avg Loss: 45.03283922, Log Avg loss: 1.48980755, Global Avg Loss: 45.03283922, Time: 0.0209 Steps: 580, Updated lr: 0.000058 Training, Epoch: 0001, Batch: 000590, Sample Num: 9440, Cur Loss: 0.64693940, Cur Avg Loss: 44.29808069, Log Avg loss: 1.68208620, Global Avg Loss: 44.29808069, Time: 0.0208 Steps: 590, Updated lr: 0.000059 Training, Epoch: 0001, Batch: 000600, Sample Num: 9600, Cur Loss: 2.77038765, Cur Avg Loss: 43.59156785, Log Avg loss: 1.90731022, Global Avg Loss: 43.59156785, Time: 0.0209 Steps: 600, Updated lr: 0.000060 Training, Epoch: 0001, Batch: 000610, Sample Num: 9760, Cur Loss: 1.21341872, Cur Avg Loss: 42.89925501, Log Avg loss: 1.36048424, Global Avg Loss: 42.89925501, Time: 0.0209 Steps: 610, Updated lr: 0.000061 Training, Epoch: 0001, Batch: 000620, Sample Num: 9920, Cur Loss: 1.12317681, Cur Avg Loss: 42.23410536, Log Avg loss: 1.65997666, Global Avg Loss: 42.23410536, Time: 0.0209 Steps: 620, Updated lr: 0.000062 Training, Epoch: 0001, Batch: 000630, Sample Num: 10080, Cur Loss: 1.16253495, Cur Avg Loss: 41.59093866, Log Avg loss: 1.71460369, Global Avg Loss: 41.59093866, Time: 0.0208 Steps: 630, Updated lr: 0.000063 Training, Epoch: 0001, Batch: 000640, Sample Num: 10240, Cur Loss: 1.23903573, Cur Avg Loss: 40.96827274, Log Avg loss: 1.74031944, Global Avg Loss: 40.96827274, Time: 0.0209 Steps: 640, Updated lr: 0.000064 Training, Epoch: 0001, Batch: 000650, Sample Num: 10400, Cur Loss: 1.13644254, Cur Avg Loss: 40.36720652, Log Avg loss: 1.89896846, Global Avg Loss: 40.36720652, Time: 0.0209 Steps: 650, Updated lr: 0.000065 Training, Epoch: 0001, Batch: 000660, Sample Num: 10560, Cur Loss: 1.72322631, Cur Avg Loss: 39.78352386, Log Avg loss: 1.84415097, Global Avg Loss: 39.78352386, Time: 0.0209 Steps: 660, Updated lr: 0.000066 Training, Epoch: 0001, Batch: 000670, Sample Num: 10720, Cur Loss: 0.99624091, Cur Avg Loss: 39.21089347, Log Avg loss: 1.41728793, Global Avg Loss: 39.21089347, Time: 0.0209 Steps: 670, Updated lr: 0.000067 Training, Epoch: 0001, Batch: 000680, Sample Num: 10880, Cur Loss: 1.92790782, Cur Avg Loss: 38.66074318, Log Avg loss: 1.80067366, Global Avg Loss: 38.66074318, Time: 0.0209 Steps: 680, Updated lr: 0.000068 Training, Epoch: 0001, Batch: 000690, Sample Num: 11040, Cur Loss: 0.85260940, Cur Avg Loss: 38.12518090, Log Avg loss: 1.70694556, Global Avg Loss: 38.12518090, Time: 0.0209 Steps: 690, Updated lr: 0.000069 Training, Epoch: 0001, Batch: 000700, Sample Num: 11200, Cur Loss: 1.60821760, Cur Avg Loss: 37.60696168, Log Avg loss: 1.84983580, Global Avg Loss: 37.60696168, Time: 0.0209 Steps: 700, Updated lr: 0.000070 Training, Epoch: 0001, Batch: 000710, Sample Num: 11360, Cur Loss: 1.90171313, Cur Avg Loss: 37.10320106, Log Avg loss: 1.83995740, Global Avg Loss: 37.10320106, Time: 0.0209 Steps: 710, Updated lr: 0.000071 Training, Epoch: 0001, Batch: 000720, Sample Num: 11520, Cur Loss: 1.71745849, Cur Avg Loss: 36.61018844, Log Avg loss: 1.60629300, Global Avg Loss: 36.61018844, Time: 0.0208 Steps: 720, Updated lr: 0.000072 Training, Epoch: 0001, Batch: 000730, Sample Num: 11680, Cur Loss: 3.06394243, Cur Avg Loss: 36.13248740, Log Avg loss: 1.73801214, Global Avg Loss: 36.13248740, Time: 0.0209 Steps: 730, Updated lr: 0.000073 Training, Epoch: 0001, Batch: 000740, Sample Num: 11840, Cur Loss: 1.51181448, Cur Avg Loss: 35.66860252, Log Avg loss: 1.80500640, Global Avg Loss: 35.66860252, Time: 0.0208 Steps: 740, Updated lr: 0.000074 Training, Epoch: 0001, Batch: 000750, Sample Num: 12000, Cur Loss: 2.59479904, Cur Avg Loss: 35.21789908, Log Avg loss: 1.86584462, Global Avg Loss: 35.21789908, Time: 0.0208 Steps: 750, Updated lr: 0.000075 Training, Epoch: 0001, Batch: 000760, Sample Num: 12160, Cur Loss: 1.02115989, Cur Avg Loss: 34.77958128, Log Avg loss: 1.90574627, Global Avg Loss: 34.77958128, Time: 0.0209 Steps: 760, Updated lr: 0.000076 Training, Epoch: 0001, Batch: 000770, Sample Num: 12320, Cur Loss: 1.89064050, Cur Avg Loss: 34.35498696, Log Avg loss: 2.08581827, Global Avg Loss: 34.35498696, Time: 0.0209 Steps: 770, Updated lr: 0.000077 Training, Epoch: 0001, Batch: 000780, Sample Num: 12480, Cur Loss: 1.75718725, Cur Avg Loss: 33.93743727, Log Avg loss: 1.78611140, Global Avg Loss: 33.93743727, Time: 0.0209 Steps: 780, Updated lr: 0.000078 Training, Epoch: 0001, Batch: 000790, Sample Num: 12640, Cur Loss: 1.21503198, Cur Avg Loss: 33.52741896, Log Avg loss: 1.54599053, Global Avg Loss: 33.52741896, Time: 0.0209 Steps: 790, Updated lr: 0.000079 Training, Epoch: 0001, Batch: 000800, Sample Num: 12800, Cur Loss: 1.30659914, Cur Avg Loss: 33.12985360, Log Avg loss: 1.72219011, Global Avg Loss: 33.12985360, Time: 0.0208 Steps: 800, Updated lr: 0.000080 Training, Epoch: 0001, Batch: 000810, Sample Num: 12960, Cur Loss: 1.77702355, Cur Avg Loss: 32.74102746, Log Avg loss: 1.63493633, Global Avg Loss: 32.74102746, Time: 0.0212 Steps: 810, Updated lr: 0.000081 Training, Epoch: 0001, Batch: 000820, Sample Num: 13120, Cur Loss: 1.54759753, Cur Avg Loss: 32.36523408, Log Avg loss: 1.92597030, Global Avg Loss: 32.36523408, Time: 0.0209 Steps: 820, Updated lr: 0.000082 Training, Epoch: 0001, Batch: 000830, Sample Num: 13280, Cur Loss: 1.11985254, Cur Avg Loss: 31.99675205, Log Avg loss: 1.78122578, Global Avg Loss: 31.99675205, Time: 0.0208 Steps: 830, Updated lr: 0.000083 Training, Epoch: 0001, Batch: 000840, Sample Num: 13440, Cur Loss: 1.04774284, Cur Avg Loss: 31.63724604, Log Avg loss: 1.79824677, Global Avg Loss: 31.63724604, Time: 0.0208 Steps: 840, Updated lr: 0.000084 Training, Epoch: 0001, Batch: 000850, Sample Num: 13600, Cur Loss: 3.19227958, Cur Avg Loss: 31.28482706, Log Avg loss: 1.68163333, Global Avg Loss: 31.28482706, Time: 0.0209 Steps: 850, Updated lr: 0.000085 Training, Epoch: 0001, Batch: 000860, Sample Num: 13760, Cur Loss: 1.38630390, Cur Avg Loss: 30.93979206, Log Avg loss: 1.61181694, Global Avg Loss: 30.93979206, Time: 0.0209 Steps: 860, Updated lr: 0.000086 Training, Epoch: 0001, Batch: 000870, Sample Num: 13920, Cur Loss: 1.05800390, Cur Avg Loss: 30.60477291, Log Avg loss: 1.79312624, Global Avg Loss: 30.60477291, Time: 0.0209 Steps: 870, Updated lr: 0.000087 Training, Epoch: 0001, Batch: 000880, Sample Num: 14080, Cur Loss: 1.05137634, Cur Avg Loss: 30.27663787, Log Avg loss: 1.72888883, Global Avg Loss: 30.27663787, Time: 0.0208 Steps: 880, Updated lr: 0.000088 Training, Epoch: 0001, Batch: 000890, Sample Num: 14240, Cur Loss: 1.28076029, Cur Avg Loss: 29.95646052, Log Avg loss: 1.78085389, Global Avg Loss: 29.95646052, Time: 0.0209 Steps: 890, Updated lr: 0.000089 Training, Epoch: 0001, Batch: 000900, Sample Num: 14400, Cur Loss: 1.67112708, Cur Avg Loss: 29.64040142, Log Avg loss: 1.51114165, Global Avg Loss: 29.64040142, Time: 0.0209 Steps: 900, Updated lr: 0.000090 Training, Epoch: 0001, Batch: 000910, Sample Num: 14560, Cur Loss: 1.87306547, Cur Avg Loss: 29.33554614, Log Avg loss: 1.89857110, Global Avg Loss: 29.33554614, Time: 0.0208 Steps: 910, Updated lr: 0.000091 Training, Epoch: 0001, Batch: 000920, Sample Num: 14720, Cur Loss: 1.90307641, Cur Avg Loss: 29.03701134, Log Avg loss: 1.87034451, Global Avg Loss: 29.03701134, Time: 0.0208 Steps: 920, Updated lr: 0.000092 Training, Epoch: 0001, Batch: 000930, Sample Num: 14880, Cur Loss: 2.64570498, Cur Avg Loss: 28.74349067, Log Avg loss: 1.73958869, Global Avg Loss: 28.74349067, Time: 0.0208 Steps: 930, Updated lr: 0.000093 Training, Epoch: 0001, Batch: 000940, Sample Num: 15040, Cur Loss: 2.30724478, Cur Avg Loss: 28.45697090, Log Avg loss: 1.81063225, Global Avg Loss: 28.45697090, Time: 0.0208 Steps: 940, Updated lr: 0.000094 Training, Epoch: 0001, Batch: 000950, Sample Num: 15200, Cur Loss: 2.22046280, Cur Avg Loss: 28.17448094, Log Avg loss: 1.62042487, Global Avg Loss: 28.17448094, Time: 0.0208 Steps: 950, Updated lr: 0.000095 Training, Epoch: 0001, Batch: 000960, Sample Num: 15360, Cur Loss: 2.39731455, Cur Avg Loss: 27.89949665, Log Avg loss: 1.77598941, Global Avg Loss: 27.89949665, Time: 0.0209 Steps: 960, Updated lr: 0.000096 Training, Epoch: 0001, Batch: 000970, Sample Num: 15520, Cur Loss: 1.32116473, Cur Avg Loss: 27.63200025, Log Avg loss: 1.95234550, Global Avg Loss: 27.63200025, Time: 0.0210 Steps: 970, Updated lr: 0.000097 Training, Epoch: 0001, Batch: 000980, Sample Num: 15680, Cur Loss: 1.21260262, Cur Avg Loss: 27.37003364, Log Avg loss: 1.95927230, Global Avg Loss: 27.37003364, Time: 0.0209 Steps: 980, Updated lr: 0.000098 Training, Epoch: 0001, Batch: 000990, Sample Num: 15840, Cur Loss: 1.82039189, Cur Avg Loss: 27.10983657, Log Avg loss: 1.61052436, Global Avg Loss: 27.10983657, Time: 0.0208 Steps: 990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001000, Sample Num: 16000, Cur Loss: 0.87098122, Cur Avg Loss: 26.85351492, Log Avg loss: 1.47767166, Global Avg Loss: 26.85351492, Time: 0.0209 Steps: 1000, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001010, Sample Num: 16160, Cur Loss: 1.33762443, Cur Avg Loss: 26.60325981, Log Avg loss: 1.57774805, Global Avg Loss: 26.60325981, Time: 0.0208 Steps: 1010, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001020, Sample Num: 16320, Cur Loss: 1.33239198, Cur Avg Loss: 26.35700401, Log Avg loss: 1.48516879, Global Avg Loss: 26.35700401, Time: 0.0208 Steps: 1020, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001030, Sample Num: 16480, Cur Loss: 1.26866794, Cur Avg Loss: 26.11361430, Log Avg loss: 1.28786345, Global Avg Loss: 26.11361430, Time: 0.0209 Steps: 1030, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001040, Sample Num: 16640, Cur Loss: 1.19493091, Cur Avg Loss: 25.87465430, Log Avg loss: 1.26177453, Global Avg Loss: 25.87465430, Time: 0.0209 Steps: 1040, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001050, Sample Num: 16800, Cur Loss: 1.68562698, Cur Avg Loss: 25.64568219, Log Avg loss: 1.83258224, Global Avg Loss: 25.64568219, Time: 0.0209 Steps: 1050, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001060, Sample Num: 16960, Cur Loss: 1.77563000, Cur Avg Loss: 25.41780136, Log Avg loss: 1.49031488, Global Avg Loss: 25.41780136, Time: 0.0210 Steps: 1060, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001070, Sample Num: 17120, Cur Loss: 2.94228888, Cur Avg Loss: 25.20070900, Log Avg loss: 2.18891819, Global Avg Loss: 25.20070900, Time: 0.0209 Steps: 1070, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001080, Sample Num: 17280, Cur Loss: 2.85633707, Cur Avg Loss: 24.98677957, Log Avg loss: 2.09633068, Global Avg Loss: 24.98677957, Time: 0.0209 Steps: 1080, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001090, Sample Num: 17440, Cur Loss: 3.04811311, Cur Avg Loss: 24.77542965, Log Avg loss: 1.94963813, Global Avg Loss: 24.77542965, Time: 0.0209 Steps: 1090, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001100, Sample Num: 17600, Cur Loss: 1.75024462, Cur Avg Loss: 24.56692391, Log Avg loss: 1.83979901, Global Avg Loss: 24.56692391, Time: 0.0209 Steps: 1100, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001110, Sample Num: 17760, Cur Loss: 0.78878832, Cur Avg Loss: 24.35884518, Log Avg loss: 1.47018459, Global Avg Loss: 24.35884518, Time: 0.0209 Steps: 1110, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001120, Sample Num: 17920, Cur Loss: 0.93414944, Cur Avg Loss: 24.15715999, Log Avg loss: 1.77010364, Global Avg Loss: 24.15715999, Time: 0.0209 Steps: 1120, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001130, Sample Num: 18080, Cur Loss: 1.79899287, Cur Avg Loss: 23.95877812, Log Avg loss: 1.74000862, Global Avg Loss: 23.95877812, Time: 0.0209 Steps: 1130, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001140, Sample Num: 18240, Cur Loss: 1.85352659, Cur Avg Loss: 23.76039132, Log Avg loss: 1.34268287, Global Avg Loss: 23.76039132, Time: 0.0210 Steps: 1140, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001150, Sample Num: 18400, Cur Loss: 1.68989873, Cur Avg Loss: 23.56744536, Log Avg loss: 1.57160609, Global Avg Loss: 23.56744536, Time: 0.0209 Steps: 1150, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001160, Sample Num: 18560, Cur Loss: 2.33740473, Cur Avg Loss: 23.37892076, Log Avg loss: 1.69859248, Global Avg Loss: 23.37892076, Time: 0.0209 Steps: 1160, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001170, Sample Num: 18720, Cur Loss: 1.73307228, Cur Avg Loss: 23.19323171, Log Avg loss: 1.65330079, Global Avg Loss: 23.19323171, Time: 0.0210 Steps: 1170, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001180, Sample Num: 18880, Cur Loss: 0.67344779, Cur Avg Loss: 23.01261918, Log Avg loss: 1.88095410, Global Avg Loss: 23.01261918, Time: 0.0210 Steps: 1180, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001190, Sample Num: 19040, Cur Loss: 1.20234561, Cur Avg Loss: 22.83020102, Log Avg loss: 1.30485795, Global Avg Loss: 22.83020102, Time: 0.0209 Steps: 1190, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001200, Sample Num: 19200, Cur Loss: 1.80044186, Cur Avg Loss: 22.65255729, Log Avg loss: 1.51295328, Global Avg Loss: 22.65255729, Time: 0.0210 Steps: 1200, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001210, Sample Num: 19360, Cur Loss: 0.57324445, Cur Avg Loss: 22.47581007, Log Avg loss: 1.26614371, Global Avg Loss: 22.47581007, Time: 0.0209 Steps: 1210, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001220, Sample Num: 19520, Cur Loss: 0.73288566, Cur Avg Loss: 22.30390863, Log Avg loss: 1.50383412, Global Avg Loss: 22.30390863, Time: 0.0209 Steps: 1220, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001230, Sample Num: 19680, Cur Loss: 1.90155101, Cur Avg Loss: 22.13556003, Log Avg loss: 1.59703108, Global Avg Loss: 22.13556003, Time: 0.0210 Steps: 1230, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001240, Sample Num: 19840, Cur Loss: 0.86142641, Cur Avg Loss: 21.97038392, Log Avg loss: 1.65372248, Global Avg Loss: 21.97038392, Time: 0.0210 Steps: 1240, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001250, Sample Num: 20000, Cur Loss: 1.88497496, Cur Avg Loss: 21.80591597, Log Avg loss: 1.41189029, Global Avg Loss: 21.80591597, Time: 0.0210 Steps: 1250, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001260, Sample Num: 20160, Cur Loss: 0.82585996, Cur Avg Loss: 21.64113790, Log Avg loss: 1.04387836, Global Avg Loss: 21.64113790, Time: 0.0210 Steps: 1260, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001270, Sample Num: 20320, Cur Loss: 1.90884614, Cur Avg Loss: 21.48343940, Log Avg loss: 1.61342955, Global Avg Loss: 21.48343940, Time: 0.0210 Steps: 1270, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001280, Sample Num: 20480, Cur Loss: 2.15078998, Cur Avg Loss: 21.33158914, Log Avg loss: 2.04660565, Global Avg Loss: 21.33158914, Time: 0.0244 Steps: 1280, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001290, Sample Num: 20640, Cur Loss: 0.92893410, Cur Avg Loss: 21.17680914, Log Avg loss: 1.36496948, Global Avg Loss: 21.17680914, Time: 0.0209 Steps: 1290, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001300, Sample Num: 20800, Cur Loss: 2.42071390, Cur Avg Loss: 21.02550336, Log Avg loss: 1.50705680, Global Avg Loss: 21.02550336, Time: 0.0210 Steps: 1300, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001310, Sample Num: 20960, Cur Loss: 0.79664153, Cur Avg Loss: 20.87337359, Log Avg loss: 1.09650335, Global Avg Loss: 20.87337359, Time: 0.0209 Steps: 1310, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001320, Sample Num: 21120, Cur Loss: 1.63852334, Cur Avg Loss: 20.72864973, Log Avg loss: 1.76982459, Global Avg Loss: 20.72864973, Time: 0.0209 Steps: 1320, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001330, Sample Num: 21280, Cur Loss: 0.30846244, Cur Avg Loss: 20.58601032, Log Avg loss: 1.75760892, Global Avg Loss: 20.58601032, Time: 0.0209 Steps: 1330, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001340, Sample Num: 21440, Cur Loss: 1.61310053, Cur Avg Loss: 20.44175980, Log Avg loss: 1.25644000, Global Avg Loss: 20.44175980, Time: 0.0210 Steps: 1340, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001350, Sample Num: 21600, Cur Loss: 1.15766215, Cur Avg Loss: 20.30076604, Log Avg loss: 1.40760182, Global Avg Loss: 20.30076604, Time: 0.0209 Steps: 1350, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001360, Sample Num: 21760, Cur Loss: 1.55254650, Cur Avg Loss: 20.16330966, Log Avg loss: 1.60669813, Global Avg Loss: 20.16330966, Time: 0.0210 Steps: 1360, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001370, Sample Num: 21920, Cur Loss: 1.27141643, Cur Avg Loss: 20.02923086, Log Avg loss: 1.79451470, Global Avg Loss: 20.02923086, Time: 0.0209 Steps: 1370, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001380, Sample Num: 22080, Cur Loss: 1.18922210, Cur Avg Loss: 19.89601170, Log Avg loss: 1.64498656, Global Avg Loss: 19.89601170, Time: 0.0210 Steps: 1380, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001390, Sample Num: 22240, Cur Loss: 2.04253721, Cur Avg Loss: 19.76421945, Log Avg loss: 1.57688883, Global Avg Loss: 19.76421945, Time: 0.0209 Steps: 1390, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001400, Sample Num: 22400, Cur Loss: 2.39399219, Cur Avg Loss: 19.63530132, Log Avg loss: 1.71568124, Global Avg Loss: 19.63530132, Time: 0.0210 Steps: 1400, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001410, Sample Num: 22560, Cur Loss: 1.00129557, Cur Avg Loss: 19.50568969, Log Avg loss: 1.36006177, Global Avg Loss: 19.50568969, Time: 0.0209 Steps: 1410, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001420, Sample Num: 22720, Cur Loss: 1.17889500, Cur Avg Loss: 19.37938246, Log Avg loss: 1.57006240, Global Avg Loss: 19.37938246, Time: 0.0209 Steps: 1420, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001430, Sample Num: 22880, Cur Loss: 0.93370795, Cur Avg Loss: 19.25614999, Log Avg loss: 1.75714052, Global Avg Loss: 19.25614999, Time: 0.0210 Steps: 1430, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001440, Sample Num: 23040, Cur Loss: 0.61237854, Cur Avg Loss: 19.13195735, Log Avg loss: 1.37240932, Global Avg Loss: 19.13195735, Time: 0.0209 Steps: 1440, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001450, Sample Num: 23200, Cur Loss: 0.52025330, Cur Avg Loss: 19.00942818, Log Avg loss: 1.36522694, Global Avg Loss: 19.00942818, Time: 0.0210 Steps: 1450, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001460, Sample Num: 23360, Cur Loss: 1.54934013, Cur Avg Loss: 18.89052923, Log Avg loss: 1.65018187, Global Avg Loss: 18.89052923, Time: 0.0209 Steps: 1460, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001470, Sample Num: 23520, Cur Loss: 2.02130604, Cur Avg Loss: 18.77073196, Log Avg loss: 1.28033045, Global Avg Loss: 18.77073196, Time: 0.0210 Steps: 1470, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001480, Sample Num: 23680, Cur Loss: 1.14841163, Cur Avg Loss: 18.65346432, Log Avg loss: 1.41512203, Global Avg Loss: 18.65346432, Time: 0.0210 Steps: 1480, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001490, Sample Num: 23840, Cur Loss: 1.31605101, Cur Avg Loss: 18.53940559, Log Avg loss: 1.65871264, Global Avg Loss: 18.53940559, Time: 0.0209 Steps: 1490, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001500, Sample Num: 24000, Cur Loss: 1.73921919, Cur Avg Loss: 18.42664795, Log Avg loss: 1.62575935, Global Avg Loss: 18.42664795, Time: 0.0210 Steps: 1500, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001510, Sample Num: 24160, Cur Loss: 1.45195770, Cur Avg Loss: 18.31695489, Log Avg loss: 1.86299673, Global Avg Loss: 18.31695489, Time: 0.0210 Steps: 1510, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001520, Sample Num: 24320, Cur Loss: 0.84219903, Cur Avg Loss: 18.20623988, Log Avg loss: 1.48827366, Global Avg Loss: 18.20623988, Time: 0.0209 Steps: 1520, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001530, Sample Num: 24480, Cur Loss: 0.58627284, Cur Avg Loss: 18.09545827, Log Avg loss: 1.25665357, Global Avg Loss: 18.09545827, Time: 0.0209 Steps: 1530, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001540, Sample Num: 24640, Cur Loss: 1.82368660, Cur Avg Loss: 17.98915970, Log Avg loss: 1.72547792, Global Avg Loss: 17.98915970, Time: 0.0247 Steps: 1540, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001550, Sample Num: 24800, Cur Loss: 1.83294451, Cur Avg Loss: 17.88209071, Log Avg loss: 1.39346617, Global Avg Loss: 17.88209071, Time: 0.0209 Steps: 1550, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001560, Sample Num: 24960, Cur Loss: 0.46922514, Cur Avg Loss: 17.77498143, Log Avg loss: 1.17304369, Global Avg Loss: 17.77498143, Time: 0.0209 Steps: 1560, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001570, Sample Num: 25120, Cur Loss: 1.69389737, Cur Avg Loss: 17.66956658, Log Avg loss: 1.22484998, Global Avg Loss: 17.66956658, Time: 0.0209 Steps: 1570, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001580, Sample Num: 25280, Cur Loss: 1.41092992, Cur Avg Loss: 17.56635033, Log Avg loss: 1.36139865, Global Avg Loss: 17.56635033, Time: 0.0209 Steps: 1580, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001590, Sample Num: 25440, Cur Loss: 1.10770321, Cur Avg Loss: 17.46433349, Log Avg loss: 1.34567212, Global Avg Loss: 17.46433349, Time: 0.0209 Steps: 1590, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001600, Sample Num: 25600, Cur Loss: 1.10877240, Cur Avg Loss: 17.36287888, Log Avg loss: 1.23159695, Global Avg Loss: 17.36287888, Time: 0.0209 Steps: 1600, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001610, Sample Num: 25760, Cur Loss: 1.52415967, Cur Avg Loss: 17.26288019, Log Avg loss: 1.26308990, Global Avg Loss: 17.26288019, Time: 0.0210 Steps: 1610, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001620, Sample Num: 25920, Cur Loss: 1.16995978, Cur Avg Loss: 17.16553745, Log Avg loss: 1.49335546, Global Avg Loss: 17.16553745, Time: 0.0209 Steps: 1620, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001630, Sample Num: 26080, Cur Loss: 0.96500754, Cur Avg Loss: 17.06628681, Log Avg loss: 0.98768294, Global Avg Loss: 17.06628681, Time: 0.0209 Steps: 1630, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001640, Sample Num: 26240, Cur Loss: 1.67235851, Cur Avg Loss: 16.96872746, Log Avg loss: 1.06655422, Global Avg Loss: 16.96872746, Time: 0.0209 Steps: 1640, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001650, Sample Num: 26400, Cur Loss: 0.72297734, Cur Avg Loss: 16.87270440, Log Avg loss: 1.12492294, Global Avg Loss: 16.87270440, Time: 0.0209 Steps: 1650, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001660, Sample Num: 26560, Cur Loss: 1.77777302, Cur Avg Loss: 16.78112780, Log Avg loss: 1.67098777, Global Avg Loss: 16.78112780, Time: 0.0209 Steps: 1660, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001670, Sample Num: 26720, Cur Loss: 0.80046797, Cur Avg Loss: 16.68878016, Log Avg loss: 1.35907184, Global Avg Loss: 16.68878016, Time: 0.0209 Steps: 1670, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001680, Sample Num: 26880, Cur Loss: 1.38485003, Cur Avg Loss: 16.59738447, Log Avg loss: 1.33430530, Global Avg Loss: 16.59738447, Time: 0.0209 Steps: 1680, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001690, Sample Num: 27040, Cur Loss: 1.09497571, Cur Avg Loss: 16.50780960, Log Avg loss: 1.45923162, Global Avg Loss: 16.50780960, Time: 0.0209 Steps: 1690, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001700, Sample Num: 27200, Cur Loss: 1.10158348, Cur Avg Loss: 16.41939699, Log Avg loss: 1.47766547, Global Avg Loss: 16.41939699, Time: 0.0209 Steps: 1700, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001710, Sample Num: 27360, Cur Loss: 1.02999020, Cur Avg Loss: 16.33153947, Log Avg loss: 1.39576045, Global Avg Loss: 16.33153947, Time: 0.0210 Steps: 1710, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001720, Sample Num: 27520, Cur Loss: 0.74775630, Cur Avg Loss: 16.24240203, Log Avg loss: 0.99989968, Global Avg Loss: 16.24240203, Time: 0.0209 Steps: 1720, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001730, Sample Num: 27680, Cur Loss: 1.78320932, Cur Avg Loss: 16.15648271, Log Avg loss: 1.37835990, Global Avg Loss: 16.15648271, Time: 0.0209 Steps: 1730, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001740, Sample Num: 27840, Cur Loss: 0.91311145, Cur Avg Loss: 16.06909716, Log Avg loss: 0.95139671, Global Avg Loss: 16.06909716, Time: 0.0209 Steps: 1740, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001750, Sample Num: 28000, Cur Loss: 1.00216043, Cur Avg Loss: 15.98408205, Log Avg loss: 1.19145398, Global Avg Loss: 15.98408205, Time: 0.0209 Steps: 1750, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001760, Sample Num: 28160, Cur Loss: 2.56273079, Cur Avg Loss: 15.90135369, Log Avg loss: 1.42389015, Global Avg Loss: 15.90135369, Time: 0.0209 Steps: 1760, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001770, Sample Num: 28320, Cur Loss: 1.74037313, Cur Avg Loss: 15.81891815, Log Avg loss: 1.31026376, Global Avg Loss: 15.81891815, Time: 0.0209 Steps: 1770, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001780, Sample Num: 28480, Cur Loss: 0.90718365, Cur Avg Loss: 15.73583686, Log Avg loss: 1.03044693, Global Avg Loss: 15.73583686, Time: 0.0209 Steps: 1780, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001790, Sample Num: 28640, Cur Loss: 0.99751282, Cur Avg Loss: 15.65456533, Log Avg loss: 1.18823358, Global Avg Loss: 15.65456533, Time: 0.0209 Steps: 1790, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001800, Sample Num: 28800, Cur Loss: 1.49271917, Cur Avg Loss: 15.57346512, Log Avg loss: 1.05652768, Global Avg Loss: 15.57346512, Time: 0.0211 Steps: 1800, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001810, Sample Num: 28960, Cur Loss: 1.40700555, Cur Avg Loss: 15.49444509, Log Avg loss: 1.27083977, Global Avg Loss: 15.49444509, Time: 0.0210 Steps: 1810, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001820, Sample Num: 29120, Cur Loss: 1.36078870, Cur Avg Loss: 15.41518679, Log Avg loss: 1.06943450, Global Avg Loss: 15.41518679, Time: 0.0210 Steps: 1820, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001830, Sample Num: 29280, Cur Loss: 1.39208341, Cur Avg Loss: 15.33720381, Log Avg loss: 1.14430069, Global Avg Loss: 15.33720381, Time: 0.0210 Steps: 1830, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001840, Sample Num: 29440, Cur Loss: 0.65213931, Cur Avg Loss: 15.26070101, Log Avg loss: 1.26068981, Global Avg Loss: 15.26070101, Time: 0.0210 Steps: 1840, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001850, Sample Num: 29600, Cur Loss: 2.77532959, Cur Avg Loss: 15.18427838, Log Avg loss: 1.12251346, Global Avg Loss: 15.18427838, Time: 0.0211 Steps: 1850, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001860, Sample Num: 29760, Cur Loss: 1.02036786, Cur Avg Loss: 15.10980208, Log Avg loss: 1.33168782, Global Avg Loss: 15.10980208, Time: 0.0211 Steps: 1860, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001870, Sample Num: 29920, Cur Loss: 1.28022981, Cur Avg Loss: 15.03575882, Log Avg loss: 1.26371177, Global Avg Loss: 15.03575882, Time: 0.0211 Steps: 1870, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001880, Sample Num: 30080, Cur Loss: 1.66257119, Cur Avg Loss: 14.96251618, Log Avg loss: 1.26614284, Global Avg Loss: 14.96251618, Time: 0.0211 Steps: 1880, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001890, Sample Num: 30240, Cur Loss: 0.40021765, Cur Avg Loss: 14.88867463, Log Avg loss: 1.00646328, Global Avg Loss: 14.88867463, Time: 0.0210 Steps: 1890, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001900, Sample Num: 30400, Cur Loss: 1.39353180, Cur Avg Loss: 14.81617156, Log Avg loss: 1.11309004, Global Avg Loss: 14.81617156, Time: 0.0210 Steps: 1900, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001910, Sample Num: 30560, Cur Loss: 1.00273895, Cur Avg Loss: 14.74533251, Log Avg loss: 1.28591334, Global Avg Loss: 14.74533251, Time: 0.0210 Steps: 1910, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001920, Sample Num: 30720, Cur Loss: 0.91101372, Cur Avg Loss: 14.67537343, Log Avg loss: 1.31318964, Global Avg Loss: 14.67537343, Time: 0.0210 Steps: 1920, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001930, Sample Num: 30880, Cur Loss: 1.10825014, Cur Avg Loss: 14.60582741, Log Avg loss: 1.25299149, Global Avg Loss: 14.60582741, Time: 0.0211 Steps: 1930, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001940, Sample Num: 31040, Cur Loss: 0.82734227, Cur Avg Loss: 14.53567326, Log Avg loss: 0.99592255, Global Avg Loss: 14.53567326, Time: 0.0210 Steps: 1940, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001950, Sample Num: 31200, Cur Loss: 1.04168141, Cur Avg Loss: 14.46702601, Log Avg loss: 1.14945894, Global Avg Loss: 14.46702601, Time: 0.0210 Steps: 1950, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001960, Sample Num: 31360, Cur Loss: 1.25508881, Cur Avg Loss: 14.39823514, Log Avg loss: 0.98401507, Global Avg Loss: 14.39823514, Time: 0.0210 Steps: 1960, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001970, Sample Num: 31520, Cur Loss: 0.90489459, Cur Avg Loss: 14.33072799, Log Avg loss: 1.09932757, Global Avg Loss: 14.33072799, Time: 0.0210 Steps: 1970, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001980, Sample Num: 31680, Cur Loss: 1.76177931, Cur Avg Loss: 14.26448560, Log Avg loss: 1.21473477, Global Avg Loss: 14.26448560, Time: 0.0210 Steps: 1980, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001990, Sample Num: 31840, Cur Loss: 1.18805254, Cur Avg Loss: 14.20034258, Log Avg loss: 1.50002412, Global Avg Loss: 14.20034258, Time: 0.0210 Steps: 1990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002000, Sample Num: 32000, Cur Loss: 1.12218547, Cur Avg Loss: 14.13493619, Log Avg loss: 1.11906567, Global Avg Loss: 14.13493619, Time: 0.0210 Steps: 2000, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002010, Sample Num: 32160, Cur Loss: 0.92184049, Cur Avg Loss: 14.06992678, Log Avg loss: 1.06804372, Global Avg Loss: 14.06992678, Time: 0.0211 Steps: 2010, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002020, Sample Num: 32320, Cur Loss: 1.12592888, Cur Avg Loss: 14.00507575, Log Avg loss: 0.97001920, Global Avg Loss: 14.00507575, Time: 0.0211 Steps: 2020, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002030, Sample Num: 32480, Cur Loss: 1.05874491, Cur Avg Loss: 13.94070039, Log Avg loss: 0.93687856, Global Avg Loss: 13.94070039, Time: 0.0211 Steps: 2030, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002040, Sample Num: 32640, Cur Loss: 1.46295953, Cur Avg Loss: 13.87807930, Log Avg loss: 1.16599787, Global Avg Loss: 13.87807930, Time: 0.0211 Steps: 2040, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002050, Sample Num: 32800, Cur Loss: 1.59363365, Cur Avg Loss: 13.81604715, Log Avg loss: 1.16148847, Global Avg Loss: 13.81604715, Time: 0.0247 Steps: 2050, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002060, Sample Num: 32960, Cur Loss: 1.19329858, Cur Avg Loss: 13.75418424, Log Avg loss: 1.07228634, Global Avg Loss: 13.75418424, Time: 0.0210 Steps: 2060, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002070, Sample Num: 33120, Cur Loss: 1.24186563, Cur Avg Loss: 13.69309853, Log Avg loss: 1.10944315, Global Avg Loss: 13.69309853, Time: 0.0210 Steps: 2070, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002080, Sample Num: 33280, Cur Loss: 0.52463353, Cur Avg Loss: 13.63112679, Log Avg loss: 0.80297717, Global Avg Loss: 13.63112679, Time: 0.0209 Steps: 2080, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002090, Sample Num: 33440, Cur Loss: 1.23560548, Cur Avg Loss: 13.56984380, Log Avg loss: 0.82298182, Global Avg Loss: 13.56984380, Time: 0.0210 Steps: 2090, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002100, Sample Num: 33600, Cur Loss: 0.87288457, Cur Avg Loss: 13.51157603, Log Avg loss: 1.33361113, Global Avg Loss: 13.51157603, Time: 0.0210 Steps: 2100, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002110, Sample Num: 33760, Cur Loss: 0.68351513, Cur Avg Loss: 13.45278029, Log Avg loss: 1.10567495, Global Avg Loss: 13.45278029, Time: 0.0210 Steps: 2110, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002120, Sample Num: 33920, Cur Loss: 0.75343025, Cur Avg Loss: 13.39365865, Log Avg loss: 0.91899286, Global Avg Loss: 13.39365865, Time: 0.0210 Steps: 2120, Updated lr: 0.000099 ***** Running evaluation checkpoint-2129 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-2129 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.687647, Avg time per batch (s): 0.020000 {"eval_avg_loss": 1.003723, "eval_total_loss": 705.617301, "eval_mae": 0.871187, "eval_mse": 1.003414, "eval_r2": 0.362164, "eval_sp_statistic": 0.65478, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.704274, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.038562, "test_total_loss": 521.358244, "test_mae": 0.910295, "test_mse": 1.038523, "test_r2": 0.329728, "test_sp_statistic": 0.653953, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.733747, "test_ps_pvalue": 0.0, "lr": 9.892935040303462e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 13.341987465408106, "train_cur_epoch_loss": 28405.09131385386, "train_cur_epoch_avg_loss": 13.341987465408106, "train_cur_epoch_time": 44.687647104263306, "train_cur_epoch_avg_time": 0.020989970457615456, "epoch": 1, "step": 2129} ################################################## Training, Epoch: 0002, Batch: 000001, Sample Num: 16, Cur Loss: 0.49740437, Cur Avg Loss: 0.49740437, Log Avg loss: 1.10323791, Global Avg Loss: 13.33595714, Time: 0.0247 Steps: 2130, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000011, Sample Num: 176, Cur Loss: 1.28667927, Cur Avg Loss: 1.22939497, Log Avg loss: 1.30259403, Global Avg Loss: 13.27972648, Time: 0.0209 Steps: 2140, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000021, Sample Num: 336, Cur Loss: 3.50471544, Cur Avg Loss: 1.19242938, Log Avg loss: 1.15176722, Global Avg Loss: 13.22331736, Time: 0.0209 Steps: 2150, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000031, Sample Num: 496, Cur Loss: 1.95617199, Cur Avg Loss: 1.19160928, Log Avg loss: 1.18988707, Global Avg Loss: 13.16760704, Time: 0.0209 Steps: 2160, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000041, Sample Num: 656, Cur Loss: 0.88696533, Cur Avg Loss: 1.17276436, Log Avg loss: 1.11434510, Global Avg Loss: 13.11206205, Time: 0.0209 Steps: 2170, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000051, Sample Num: 816, Cur Loss: 0.26247507, Cur Avg Loss: 1.08036911, Log Avg loss: 0.70154859, Global Avg Loss: 13.05513309, Time: 0.0209 Steps: 2180, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000061, Sample Num: 976, Cur Loss: 0.97718781, Cur Avg Loss: 1.03476847, Log Avg loss: 0.80220518, Global Avg Loss: 12.99918365, Time: 0.0209 Steps: 2190, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000071, Sample Num: 1136, Cur Loss: 0.78910196, Cur Avg Loss: 1.03741289, Log Avg loss: 1.05354385, Global Avg Loss: 12.94488529, Time: 0.0209 Steps: 2200, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000081, Sample Num: 1296, Cur Loss: 0.48714197, Cur Avg Loss: 1.03787056, Log Avg loss: 1.04112005, Global Avg Loss: 12.89102209, Time: 0.0208 Steps: 2210, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000091, Sample Num: 1456, Cur Loss: 1.14812291, Cur Avg Loss: 1.02899898, Log Avg loss: 0.95713920, Global Avg Loss: 12.83726587, Time: 0.0208 Steps: 2220, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000101, Sample Num: 1616, Cur Loss: 0.96809018, Cur Avg Loss: 1.02017984, Log Avg loss: 0.93992565, Global Avg Loss: 12.78391456, Time: 0.0208 Steps: 2230, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000111, Sample Num: 1776, Cur Loss: 0.43672889, Cur Avg Loss: 1.01886020, Log Avg loss: 1.00553180, Global Avg Loss: 12.73133250, Time: 0.0208 Steps: 2240, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000121, Sample Num: 1936, Cur Loss: 0.46761027, Cur Avg Loss: 0.99766426, Log Avg loss: 0.76238930, Global Avg Loss: 12.67813720, Time: 0.0208 Steps: 2250, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000131, Sample Num: 2096, Cur Loss: 0.60942858, Cur Avg Loss: 1.00103704, Log Avg loss: 1.04184777, Global Avg Loss: 12.62664919, Time: 0.0208 Steps: 2260, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000141, Sample Num: 2256, Cur Loss: 0.88737196, Cur Avg Loss: 0.99949494, Log Avg loss: 0.97929338, Global Avg Loss: 12.57533925, Time: 0.0208 Steps: 2270, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000151, Sample Num: 2416, Cur Loss: 1.33304298, Cur Avg Loss: 1.02002494, Log Avg loss: 1.30949790, Global Avg Loss: 12.52592767, Time: 0.0208 Steps: 2280, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000161, Sample Num: 2576, Cur Loss: 1.24411714, Cur Avg Loss: 1.03692638, Log Avg loss: 1.29213823, Global Avg Loss: 12.47687182, Time: 0.0208 Steps: 2290, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000171, Sample Num: 2736, Cur Loss: 1.12988806, Cur Avg Loss: 1.03854754, Log Avg loss: 1.06464819, Global Avg Loss: 12.42725345, Time: 0.0208 Steps: 2300, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000181, Sample Num: 2896, Cur Loss: 0.62439787, Cur Avg Loss: 1.03777707, Log Avg loss: 1.02460201, Global Avg Loss: 12.37789133, Time: 0.0208 Steps: 2310, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000191, Sample Num: 3056, Cur Loss: 0.66948253, Cur Avg Loss: 1.03979512, Log Avg loss: 1.07632181, Global Avg Loss: 12.32917766, Time: 0.0208 Steps: 2320, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000201, Sample Num: 3216, Cur Loss: 0.49947304, Cur Avg Loss: 1.03836439, Log Avg loss: 1.01103735, Global Avg Loss: 12.28060196, Time: 0.0207 Steps: 2330, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000211, Sample Num: 3376, Cur Loss: 0.67850697, Cur Avg Loss: 1.02394881, Log Avg loss: 0.73419574, Global Avg Loss: 12.23125834, Time: 0.0208 Steps: 2340, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000221, Sample Num: 3536, Cur Loss: 0.67210811, Cur Avg Loss: 1.02344489, Log Avg loss: 1.01281227, Global Avg Loss: 12.18352027, Time: 0.0208 Steps: 2350, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000231, Sample Num: 3696, Cur Loss: 0.35848910, Cur Avg Loss: 1.02726176, Log Avg loss: 1.11161451, Global Avg Loss: 12.13660542, Time: 0.0208 Steps: 2360, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000241, Sample Num: 3856, Cur Loss: 0.77346480, Cur Avg Loss: 1.03459965, Log Avg loss: 1.20410487, Global Avg Loss: 12.09047672, Time: 0.0208 Steps: 2370, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000251, Sample Num: 4016, Cur Loss: 0.45051298, Cur Avg Loss: 1.03296061, Log Avg loss: 0.99345969, Global Avg Loss: 12.04385060, Time: 0.0208 Steps: 2380, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000261, Sample Num: 4176, Cur Loss: 0.55659038, Cur Avg Loss: 1.03104640, Log Avg loss: 0.98299978, Global Avg Loss: 11.99757089, Time: 0.0244 Steps: 2390, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000271, Sample Num: 4336, Cur Loss: 1.42451215, Cur Avg Loss: 1.02546055, Log Avg loss: 0.87966980, Global Avg Loss: 11.95124630, Time: 0.0207 Steps: 2400, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000281, Sample Num: 4496, Cur Loss: 1.50048351, Cur Avg Loss: 1.03640841, Log Avg loss: 1.33309560, Global Avg Loss: 11.90718758, Time: 0.0208 Steps: 2410, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000291, Sample Num: 4656, Cur Loss: 2.80928016, Cur Avg Loss: 1.06381302, Log Avg loss: 1.83388256, Global Avg Loss: 11.86556236, Time: 0.0207 Steps: 2420, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000301, Sample Num: 4816, Cur Loss: 0.40312451, Cur Avg Loss: 1.05815932, Log Avg loss: 0.89363643, Global Avg Loss: 11.82041040, Time: 0.0208 Steps: 2430, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000311, Sample Num: 4976, Cur Loss: 1.29518390, Cur Avg Loss: 1.05515999, Log Avg loss: 0.96488039, Global Avg Loss: 11.77592052, Time: 0.0208 Steps: 2440, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000321, Sample Num: 5136, Cur Loss: 0.69336402, Cur Avg Loss: 1.05349269, Log Avg loss: 1.00163962, Global Avg Loss: 11.73194386, Time: 0.0208 Steps: 2450, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000331, Sample Num: 5296, Cur Loss: 0.98150194, Cur Avg Loss: 1.05486409, Log Avg loss: 1.09888584, Global Avg Loss: 11.68872005, Time: 0.0208 Steps: 2460, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000341, Sample Num: 5456, Cur Loss: 0.79361421, Cur Avg Loss: 1.04557239, Log Avg loss: 0.73801710, Global Avg Loss: 11.64438522, Time: 0.0208 Steps: 2470, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000351, Sample Num: 5616, Cur Loss: 1.28445089, Cur Avg Loss: 1.04170883, Log Avg loss: 0.90996160, Global Avg Loss: 11.60110126, Time: 0.0207 Steps: 2480, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000361, Sample Num: 5776, Cur Loss: 1.30613232, Cur Avg Loss: 1.05250984, Log Avg loss: 1.43162543, Global Avg Loss: 11.56025999, Time: 0.0207 Steps: 2490, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000371, Sample Num: 5936, Cur Loss: 0.68499660, Cur Avg Loss: 1.07080110, Log Avg loss: 1.73111556, Global Avg Loss: 11.52094341, Time: 0.0208 Steps: 2500, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000381, Sample Num: 6096, Cur Loss: 0.87249577, Cur Avg Loss: 1.06740343, Log Avg loss: 0.94134959, Global Avg Loss: 11.47879363, Time: 0.0207 Steps: 2510, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000391, Sample Num: 6256, Cur Loss: 0.73761529, Cur Avg Loss: 1.06452074, Log Avg loss: 0.95469025, Global Avg Loss: 11.43703132, Time: 0.0207 Steps: 2520, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000401, Sample Num: 6416, Cur Loss: 0.33144772, Cur Avg Loss: 1.05798387, Log Avg loss: 0.80239247, Global Avg Loss: 11.39499717, Time: 0.0208 Steps: 2530, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000411, Sample Num: 6576, Cur Loss: 1.59931409, Cur Avg Loss: 1.05443334, Log Avg loss: 0.91205710, Global Avg Loss: 11.35372575, Time: 0.0207 Steps: 2540, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000421, Sample Num: 6736, Cur Loss: 0.90463209, Cur Avg Loss: 1.05580800, Log Avg loss: 1.11230645, Global Avg Loss: 11.31356333, Time: 0.0207 Steps: 2550, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000431, Sample Num: 6896, Cur Loss: 1.02254033, Cur Avg Loss: 1.05419181, Log Avg loss: 0.98615033, Global Avg Loss: 11.27322187, Time: 0.0208 Steps: 2560, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000441, Sample Num: 7056, Cur Loss: 0.69023770, Cur Avg Loss: 1.05580450, Log Avg loss: 1.12531112, Global Avg Loss: 11.23373584, Time: 0.0207 Steps: 2570, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000451, Sample Num: 7216, Cur Loss: 0.92717344, Cur Avg Loss: 1.05005840, Log Avg loss: 0.79665564, Global Avg Loss: 11.19328204, Time: 0.0207 Steps: 2580, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000461, Sample Num: 7376, Cur Loss: 0.28159538, Cur Avg Loss: 1.04966371, Log Avg loss: 1.03186300, Global Avg Loss: 11.15404876, Time: 0.0208 Steps: 2590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000471, Sample Num: 7536, Cur Loss: 1.58855307, Cur Avg Loss: 1.04364331, Log Avg loss: 0.76610287, Global Avg Loss: 11.11409512, Time: 0.0208 Steps: 2600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000481, Sample Num: 7696, Cur Loss: 1.47204590, Cur Avg Loss: 1.04091544, Log Avg loss: 0.91243274, Global Avg Loss: 11.07500829, Time: 0.0207 Steps: 2610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000491, Sample Num: 7856, Cur Loss: 0.47253537, Cur Avg Loss: 1.03768418, Log Avg loss: 0.88226049, Global Avg Loss: 11.03610467, Time: 0.0208 Steps: 2620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000501, Sample Num: 8016, Cur Loss: 0.61524713, Cur Avg Loss: 1.03376681, Log Avg loss: 0.84142394, Global Avg Loss: 10.99734163, Time: 0.0208 Steps: 2630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000511, Sample Num: 8176, Cur Loss: 1.14451790, Cur Avg Loss: 1.02924082, Log Avg loss: 0.80248895, Global Avg Loss: 10.95872476, Time: 0.0207 Steps: 2640, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000521, Sample Num: 8336, Cur Loss: 1.25058913, Cur Avg Loss: 1.03102822, Log Avg loss: 1.12236429, Global Avg Loss: 10.92160642, Time: 0.0208 Steps: 2650, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000531, Sample Num: 8496, Cur Loss: 0.84235978, Cur Avg Loss: 1.02562974, Log Avg loss: 0.74436917, Global Avg Loss: 10.88334613, Time: 0.0207 Steps: 2660, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000541, Sample Num: 8656, Cur Loss: 0.68129373, Cur Avg Loss: 1.02397595, Log Avg loss: 0.93615928, Global Avg Loss: 10.84609075, Time: 0.0207 Steps: 2670, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000551, Sample Num: 8816, Cur Loss: 0.29664129, Cur Avg Loss: 1.01775419, Log Avg loss: 0.68115715, Global Avg Loss: 10.80816189, Time: 0.0206 Steps: 2680, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000561, Sample Num: 8976, Cur Loss: 1.31461358, Cur Avg Loss: 1.01608102, Log Avg loss: 0.92388956, Global Avg Loss: 10.77141739, Time: 0.0206 Steps: 2690, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000571, Sample Num: 9136, Cur Loss: 0.80950975, Cur Avg Loss: 1.01289675, Log Avg loss: 0.83425907, Global Avg Loss: 10.73461310, Time: 0.0206 Steps: 2700, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000581, Sample Num: 9296, Cur Loss: 0.32596129, Cur Avg Loss: 1.00994098, Log Avg loss: 0.84116664, Global Avg Loss: 10.69810591, Time: 0.0207 Steps: 2710, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000591, Sample Num: 9456, Cur Loss: 0.90263921, Cur Avg Loss: 1.00495255, Log Avg loss: 0.71512440, Global Avg Loss: 10.66140378, Time: 0.0206 Steps: 2720, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000601, Sample Num: 9616, Cur Loss: 0.68079746, Cur Avg Loss: 1.00155073, Log Avg loss: 0.80050354, Global Avg Loss: 10.62528326, Time: 0.0207 Steps: 2730, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000611, Sample Num: 9776, Cur Loss: 0.97021663, Cur Avg Loss: 1.00018846, Log Avg loss: 0.91831595, Global Avg Loss: 10.58985637, Time: 0.0207 Steps: 2740, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000621, Sample Num: 9936, Cur Loss: 0.61284012, Cur Avg Loss: 0.99919256, Log Avg loss: 0.93834302, Global Avg Loss: 10.55475996, Time: 0.0207 Steps: 2750, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000631, Sample Num: 10096, Cur Loss: 1.21895039, Cur Avg Loss: 0.99783935, Log Avg loss: 0.91380488, Global Avg Loss: 10.51982896, Time: 0.0207 Steps: 2760, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000641, Sample Num: 10256, Cur Loss: 0.35262579, Cur Avg Loss: 0.99776920, Log Avg loss: 0.99334316, Global Avg Loss: 10.48543732, Time: 0.0206 Steps: 2770, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000651, Sample Num: 10416, Cur Loss: 0.89219368, Cur Avg Loss: 0.99162456, Log Avg loss: 0.59775285, Global Avg Loss: 10.44987011, Time: 0.0206 Steps: 2780, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000661, Sample Num: 10576, Cur Loss: 0.82920808, Cur Avg Loss: 0.98974077, Log Avg loss: 0.86710629, Global Avg Loss: 10.41552329, Time: 0.0207 Steps: 2790, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000671, Sample Num: 10736, Cur Loss: 0.39753345, Cur Avg Loss: 0.99020972, Log Avg loss: 1.02120732, Global Avg Loss: 10.38197216, Time: 0.0207 Steps: 2800, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000681, Sample Num: 10896, Cur Loss: 1.38509703, Cur Avg Loss: 0.99296632, Log Avg loss: 1.17793391, Global Avg Loss: 10.34921757, Time: 0.0207 Steps: 2810, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000691, Sample Num: 11056, Cur Loss: 1.16047144, Cur Avg Loss: 0.99065740, Log Avg loss: 0.83342002, Global Avg Loss: 10.31547361, Time: 0.0206 Steps: 2820, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000701, Sample Num: 11216, Cur Loss: 0.80249399, Cur Avg Loss: 0.98929652, Log Avg loss: 0.89525984, Global Avg Loss: 10.28218663, Time: 0.0206 Steps: 2830, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000711, Sample Num: 11376, Cur Loss: 0.72273386, Cur Avg Loss: 0.98588698, Log Avg loss: 0.74687829, Global Avg Loss: 10.24861161, Time: 0.0207 Steps: 2840, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000721, Sample Num: 11536, Cur Loss: 1.01027679, Cur Avg Loss: 0.98317380, Log Avg loss: 0.79026666, Global Avg Loss: 10.21542443, Time: 0.0207 Steps: 2850, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000731, Sample Num: 11696, Cur Loss: 0.71363652, Cur Avg Loss: 0.97888392, Log Avg loss: 0.66958327, Global Avg Loss: 10.18204736, Time: 0.0207 Steps: 2860, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000741, Sample Num: 11856, Cur Loss: 0.77010798, Cur Avg Loss: 0.97484178, Log Avg loss: 0.67936107, Global Avg Loss: 10.14893696, Time: 0.0207 Steps: 2870, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000751, Sample Num: 12016, Cur Loss: 0.91889834, Cur Avg Loss: 0.97480235, Log Avg loss: 0.97188095, Global Avg Loss: 10.11707218, Time: 0.0206 Steps: 2880, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000761, Sample Num: 12176, Cur Loss: 0.43384880, Cur Avg Loss: 0.97047212, Log Avg loss: 0.64527211, Global Avg Loss: 10.08429779, Time: 0.0207 Steps: 2890, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000771, Sample Num: 12336, Cur Loss: 0.61945331, Cur Avg Loss: 0.96784085, Log Avg loss: 0.76760057, Global Avg Loss: 10.05217124, Time: 0.0244 Steps: 2900, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000781, Sample Num: 12496, Cur Loss: 1.44538248, Cur Avg Loss: 0.97175630, Log Avg loss: 1.27363760, Global Avg Loss: 10.02200446, Time: 0.0208 Steps: 2910, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000791, Sample Num: 12656, Cur Loss: 0.55622292, Cur Avg Loss: 0.97075558, Log Avg loss: 0.89259945, Global Avg Loss: 9.99073938, Time: 0.0209 Steps: 2920, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000801, Sample Num: 12816, Cur Loss: 0.71106255, Cur Avg Loss: 0.96983561, Log Avg loss: 0.89706590, Global Avg Loss: 9.95970295, Time: 0.0207 Steps: 2930, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000811, Sample Num: 12976, Cur Loss: 0.84937853, Cur Avg Loss: 0.96810918, Log Avg loss: 0.82982248, Global Avg Loss: 9.92864893, Time: 0.0208 Steps: 2940, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000821, Sample Num: 13136, Cur Loss: 0.69212073, Cur Avg Loss: 0.96810469, Log Avg loss: 0.96774044, Global Avg Loss: 9.89827297, Time: 0.0208 Steps: 2950, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000831, Sample Num: 13296, Cur Loss: 1.23987448, Cur Avg Loss: 0.96583383, Log Avg loss: 0.77939665, Global Avg Loss: 9.86746596, Time: 0.0207 Steps: 2960, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000841, Sample Num: 13456, Cur Loss: 0.52716154, Cur Avg Loss: 0.96307873, Log Avg loss: 0.73412980, Global Avg Loss: 9.83671398, Time: 0.0207 Steps: 2970, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000851, Sample Num: 13616, Cur Loss: 0.67142898, Cur Avg Loss: 0.95969246, Log Avg loss: 0.67490695, Global Avg Loss: 9.80596966, Time: 0.0207 Steps: 2980, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000861, Sample Num: 13776, Cur Loss: 0.50876343, Cur Avg Loss: 0.95691594, Log Avg loss: 0.72063398, Global Avg Loss: 9.77558393, Time: 0.0207 Steps: 2990, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000871, Sample Num: 13936, Cur Loss: 0.39126945, Cur Avg Loss: 0.95200628, Log Avg loss: 0.52928492, Global Avg Loss: 9.74476293, Time: 0.0208 Steps: 3000, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000881, Sample Num: 14096, Cur Loss: 1.42859209, Cur Avg Loss: 0.95232231, Log Avg loss: 0.97984813, Global Avg Loss: 9.71564361, Time: 0.0207 Steps: 3010, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000891, Sample Num: 14256, Cur Loss: 0.90699863, Cur Avg Loss: 0.95036398, Log Avg loss: 0.77783538, Global Avg Loss: 9.68604822, Time: 0.0208 Steps: 3020, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000901, Sample Num: 14416, Cur Loss: 0.41341633, Cur Avg Loss: 0.94944806, Log Avg loss: 0.86783968, Global Avg Loss: 9.65694522, Time: 0.0207 Steps: 3030, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000911, Sample Num: 14576, Cur Loss: 0.51244330, Cur Avg Loss: 0.94945875, Log Avg loss: 0.95042187, Global Avg Loss: 9.62830534, Time: 0.0207 Steps: 3040, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000921, Sample Num: 14736, Cur Loss: 0.88346040, Cur Avg Loss: 0.94847040, Log Avg loss: 0.85843140, Global Avg Loss: 9.59955166, Time: 0.0207 Steps: 3050, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000931, Sample Num: 14896, Cur Loss: 1.78682351, Cur Avg Loss: 0.94673755, Log Avg loss: 0.78714220, Global Avg Loss: 9.57075293, Time: 0.0207 Steps: 3060, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000941, Sample Num: 15056, Cur Loss: 0.64879519, Cur Avg Loss: 0.94695479, Log Avg loss: 0.96717965, Global Avg Loss: 9.54272826, Time: 0.0207 Steps: 3070, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000951, Sample Num: 15216, Cur Loss: 0.89681464, Cur Avg Loss: 0.95422341, Log Avg loss: 1.63820053, Global Avg Loss: 9.51706421, Time: 0.0207 Steps: 3080, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000961, Sample Num: 15376, Cur Loss: 0.78793049, Cur Avg Loss: 0.95415244, Log Avg loss: 0.94740364, Global Avg Loss: 9.48933068, Time: 0.0208 Steps: 3090, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000971, Sample Num: 15536, Cur Loss: 1.09683609, Cur Avg Loss: 0.95530962, Log Avg loss: 1.06651400, Global Avg Loss: 9.46216031, Time: 0.0207 Steps: 3100, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000981, Sample Num: 15696, Cur Loss: 0.38214228, Cur Avg Loss: 0.95441715, Log Avg loss: 0.86775819, Global Avg Loss: 9.43452557, Time: 0.0207 Steps: 3110, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000991, Sample Num: 15856, Cur Loss: 1.28926241, Cur Avg Loss: 0.95126572, Log Avg loss: 0.64211074, Global Avg Loss: 9.40634476, Time: 0.0208 Steps: 3120, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001001, Sample Num: 16016, Cur Loss: 0.58167982, Cur Avg Loss: 0.95060323, Log Avg loss: 0.88495008, Global Avg Loss: 9.37911985, Time: 0.0207 Steps: 3130, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001011, Sample Num: 16176, Cur Loss: 1.26332688, Cur Avg Loss: 0.94936141, Log Avg loss: 0.82505601, Global Avg Loss: 9.35187761, Time: 0.0207 Steps: 3140, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001021, Sample Num: 16336, Cur Loss: 0.54871655, Cur Avg Loss: 0.94739069, Log Avg loss: 0.74815088, Global Avg Loss: 9.32456419, Time: 0.0207 Steps: 3150, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001031, Sample Num: 16496, Cur Loss: 0.38694432, Cur Avg Loss: 0.94533839, Log Avg loss: 0.73579780, Global Avg Loss: 9.29738455, Time: 0.0210 Steps: 3160, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001041, Sample Num: 16656, Cur Loss: 0.93308353, Cur Avg Loss: 0.94323287, Log Avg loss: 0.72615414, Global Avg Loss: 9.27034597, Time: 0.0207 Steps: 3170, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001051, Sample Num: 16816, Cur Loss: 1.18084633, Cur Avg Loss: 0.94405963, Log Avg loss: 1.03012532, Global Avg Loss: 9.24443333, Time: 0.0208 Steps: 3180, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001061, Sample Num: 16976, Cur Loss: 0.73508644, Cur Avg Loss: 0.94351225, Log Avg loss: 0.88598323, Global Avg Loss: 9.21823129, Time: 0.0207 Steps: 3190, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001071, Sample Num: 17136, Cur Loss: 2.15038252, Cur Avg Loss: 0.94254527, Log Avg loss: 0.83994861, Global Avg Loss: 9.19204916, Time: 0.0207 Steps: 3200, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001081, Sample Num: 17296, Cur Loss: 0.63622969, Cur Avg Loss: 0.94055416, Log Avg loss: 0.72730603, Global Avg Loss: 9.16567924, Time: 0.0207 Steps: 3210, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001091, Sample Num: 17456, Cur Loss: 0.96424079, Cur Avg Loss: 0.94027727, Log Avg loss: 0.91034541, Global Avg Loss: 9.14004156, Time: 0.0207 Steps: 3220, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001101, Sample Num: 17616, Cur Loss: 0.90129668, Cur Avg Loss: 0.93796884, Log Avg loss: 0.68611910, Global Avg Loss: 9.11386842, Time: 0.0207 Steps: 3230, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001111, Sample Num: 17776, Cur Loss: 0.90504217, Cur Avg Loss: 0.93791496, Log Avg loss: 0.93198290, Global Avg Loss: 9.08861569, Time: 0.0207 Steps: 3240, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001121, Sample Num: 17936, Cur Loss: 0.51587927, Cur Avg Loss: 0.93997929, Log Avg loss: 1.16932563, Global Avg Loss: 9.06424864, Time: 0.0208 Steps: 3250, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001131, Sample Num: 18096, Cur Loss: 0.28895873, Cur Avg Loss: 0.94206176, Log Avg loss: 1.17550758, Global Avg Loss: 9.04005005, Time: 0.0207 Steps: 3260, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001141, Sample Num: 18256, Cur Loss: 0.43186274, Cur Avg Loss: 0.93862527, Log Avg loss: 0.54995835, Global Avg Loss: 9.01408647, Time: 0.0207 Steps: 3270, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001151, Sample Num: 18416, Cur Loss: 1.00867009, Cur Avg Loss: 0.93866166, Log Avg loss: 0.94281278, Global Avg Loss: 8.98947893, Time: 0.0207 Steps: 3280, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001161, Sample Num: 18576, Cur Loss: 0.42695671, Cur Avg Loss: 0.93796266, Log Avg loss: 0.85750778, Global Avg Loss: 8.96476169, Time: 0.0207 Steps: 3290, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001171, Sample Num: 18736, Cur Loss: 0.74625325, Cur Avg Loss: 0.93667626, Log Avg loss: 0.78732512, Global Avg Loss: 8.93998158, Time: 0.0207 Steps: 3300, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001181, Sample Num: 18896, Cur Loss: 1.85844517, Cur Avg Loss: 0.93782449, Log Avg loss: 1.07228313, Global Avg Loss: 8.91621210, Time: 0.0208 Steps: 3310, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001191, Sample Num: 19056, Cur Loss: 0.84837782, Cur Avg Loss: 0.94042748, Log Avg loss: 1.24784030, Global Avg Loss: 8.89311459, Time: 0.0207 Steps: 3320, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001201, Sample Num: 19216, Cur Loss: 0.99469310, Cur Avg Loss: 0.94016433, Log Avg loss: 0.90882265, Global Avg Loss: 8.86913774, Time: 0.0207 Steps: 3330, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001211, Sample Num: 19376, Cur Loss: 2.12845230, Cur Avg Loss: 0.93874138, Log Avg loss: 0.76784579, Global Avg Loss: 8.84488237, Time: 0.0207 Steps: 3340, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001221, Sample Num: 19536, Cur Loss: 0.40709025, Cur Avg Loss: 0.93780483, Log Avg loss: 0.82438853, Global Avg Loss: 8.82094060, Time: 0.0207 Steps: 3350, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001231, Sample Num: 19696, Cur Loss: 0.54786664, Cur Avg Loss: 0.93609718, Log Avg loss: 0.72759271, Global Avg Loss: 8.79685326, Time: 0.0207 Steps: 3360, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001241, Sample Num: 19856, Cur Loss: 0.62462497, Cur Avg Loss: 0.93262514, Log Avg loss: 0.50521782, Global Avg Loss: 8.77224900, Time: 0.0207 Steps: 3370, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001251, Sample Num: 20016, Cur Loss: 0.88827348, Cur Avg Loss: 0.93211914, Log Avg loss: 0.86932356, Global Avg Loss: 8.74886756, Time: 0.0207 Steps: 3380, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001261, Sample Num: 20176, Cur Loss: 0.86135501, Cur Avg Loss: 0.93471611, Log Avg loss: 1.25959719, Global Avg Loss: 8.72677532, Time: 0.0208 Steps: 3390, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001271, Sample Num: 20336, Cur Loss: 1.24578273, Cur Avg Loss: 0.93586383, Log Avg loss: 1.08059194, Global Avg Loss: 8.70428654, Time: 0.0207 Steps: 3400, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001281, Sample Num: 20496, Cur Loss: 0.56410581, Cur Avg Loss: 0.93427113, Log Avg loss: 0.73183924, Global Avg Loss: 8.68090693, Time: 0.0244 Steps: 3410, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001291, Sample Num: 20656, Cur Loss: 1.22529137, Cur Avg Loss: 0.93418369, Log Avg loss: 0.92298137, Global Avg Loss: 8.65822294, Time: 0.0208 Steps: 3420, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001301, Sample Num: 20816, Cur Loss: 0.69981396, Cur Avg Loss: 0.93455983, Log Avg loss: 0.98311998, Global Avg Loss: 8.63584655, Time: 0.0208 Steps: 3430, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001311, Sample Num: 20976, Cur Loss: 2.50901461, Cur Avg Loss: 0.93477095, Log Avg loss: 0.96223788, Global Avg Loss: 8.61353954, Time: 0.0207 Steps: 3440, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001321, Sample Num: 21136, Cur Loss: 0.30737400, Cur Avg Loss: 0.93243785, Log Avg loss: 0.62656791, Global Avg Loss: 8.59038890, Time: 0.0207 Steps: 3450, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001331, Sample Num: 21296, Cur Loss: 0.80832267, Cur Avg Loss: 0.92994146, Log Avg loss: 0.60016818, Global Avg Loss: 8.56729578, Time: 0.0207 Steps: 3460, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001341, Sample Num: 21456, Cur Loss: 0.83225024, Cur Avg Loss: 0.92874467, Log Avg loss: 0.76945278, Global Avg Loss: 8.54482361, Time: 0.0207 Steps: 3470, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001351, Sample Num: 21616, Cur Loss: 0.70818686, Cur Avg Loss: 0.92806377, Log Avg loss: 0.83675442, Global Avg Loss: 8.52267398, Time: 0.0207 Steps: 3480, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001361, Sample Num: 21776, Cur Loss: 0.85334671, Cur Avg Loss: 0.92585017, Log Avg loss: 0.62679267, Global Avg Loss: 8.50004968, Time: 0.0208 Steps: 3490, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001371, Sample Num: 21936, Cur Loss: 1.40274191, Cur Avg Loss: 0.92631072, Log Avg loss: 0.98899202, Global Avg Loss: 8.47858952, Time: 0.0207 Steps: 3500, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001381, Sample Num: 22096, Cur Loss: 3.69102287, Cur Avg Loss: 0.93039593, Log Avg loss: 1.49047836, Global Avg Loss: 8.45868037, Time: 0.0207 Steps: 3510, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001391, Sample Num: 22256, Cur Loss: 0.46871853, Cur Avg Loss: 0.92978540, Log Avg loss: 0.84547058, Global Avg Loss: 8.43705193, Time: 0.0208 Steps: 3520, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001401, Sample Num: 22416, Cur Loss: 1.31700826, Cur Avg Loss: 0.93014842, Log Avg loss: 0.98064447, Global Avg Loss: 8.41592896, Time: 0.0207 Steps: 3530, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001411, Sample Num: 22576, Cur Loss: 0.66428781, Cur Avg Loss: 0.92972009, Log Avg loss: 0.86971222, Global Avg Loss: 8.39461197, Time: 0.0207 Steps: 3540, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001421, Sample Num: 22736, Cur Loss: 0.87990719, Cur Avg Loss: 0.93308787, Log Avg loss: 1.40828135, Global Avg Loss: 8.37493216, Time: 0.0207 Steps: 3550, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001431, Sample Num: 22896, Cur Loss: 0.94830656, Cur Avg Loss: 0.93620521, Log Avg loss: 1.37917854, Global Avg Loss: 8.35528117, Time: 0.0207 Steps: 3560, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001441, Sample Num: 23056, Cur Loss: 0.26191074, Cur Avg Loss: 0.93518793, Log Avg loss: 0.78961597, Global Avg Loss: 8.33408883, Time: 0.0207 Steps: 3570, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001451, Sample Num: 23216, Cur Loss: 1.04965246, Cur Avg Loss: 0.93315895, Log Avg loss: 0.64078197, Global Avg Loss: 8.31259915, Time: 0.0207 Steps: 3580, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001461, Sample Num: 23376, Cur Loss: 1.01998186, Cur Avg Loss: 0.93309916, Log Avg loss: 0.92442372, Global Avg Loss: 8.29201927, Time: 0.0207 Steps: 3590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001471, Sample Num: 23536, Cur Loss: 1.97902250, Cur Avg Loss: 0.94630929, Log Avg loss: 2.87631021, Global Avg Loss: 8.27697563, Time: 0.0207 Steps: 3600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001481, Sample Num: 23696, Cur Loss: 1.17938757, Cur Avg Loss: 0.94738690, Log Avg loss: 1.10590333, Global Avg Loss: 8.25711117, Time: 0.0207 Steps: 3610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001491, Sample Num: 23856, Cur Loss: 0.47785541, Cur Avg Loss: 0.94846094, Log Avg loss: 1.10752569, Global Avg Loss: 8.23736093, Time: 0.0207 Steps: 3620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001501, Sample Num: 24016, Cur Loss: 1.21500826, Cur Avg Loss: 0.94917883, Log Avg loss: 1.05621675, Global Avg Loss: 8.21757817, Time: 0.0207 Steps: 3630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001511, Sample Num: 24176, Cur Loss: 0.42404246, Cur Avg Loss: 0.94722451, Log Avg loss: 0.65388091, Global Avg Loss: 8.19679878, Time: 0.0207 Steps: 3640, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001521, Sample Num: 24336, Cur Loss: 0.30945915, Cur Avg Loss: 0.94715843, Log Avg loss: 0.93717370, Global Avg Loss: 8.17690939, Time: 0.0207 Steps: 3650, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001531, Sample Num: 24496, Cur Loss: 0.74310094, Cur Avg Loss: 0.94736231, Log Avg loss: 0.97837166, Global Avg Loss: 8.15724126, Time: 0.0207 Steps: 3660, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001541, Sample Num: 24656, Cur Loss: 1.41685581, Cur Avg Loss: 0.95557788, Log Avg loss: 2.21338256, Global Avg Loss: 8.14104546, Time: 0.0209 Steps: 3670, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001551, Sample Num: 24816, Cur Loss: 0.66114235, Cur Avg Loss: 0.95559329, Log Avg loss: 0.95796825, Global Avg Loss: 8.12152623, Time: 0.0208 Steps: 3680, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001561, Sample Num: 24976, Cur Loss: 0.98532104, Cur Avg Loss: 0.95406018, Log Avg loss: 0.71627383, Global Avg Loss: 8.10145779, Time: 0.0207 Steps: 3690, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001571, Sample Num: 25136, Cur Loss: 0.81447250, Cur Avg Loss: 0.95414808, Log Avg loss: 0.96786943, Global Avg Loss: 8.08217782, Time: 0.0207 Steps: 3700, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001581, Sample Num: 25296, Cur Loss: 1.23169398, Cur Avg Loss: 0.95726964, Log Avg loss: 1.44766690, Global Avg Loss: 8.06429504, Time: 0.0208 Steps: 3710, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001591, Sample Num: 25456, Cur Loss: 1.32256222, Cur Avg Loss: 0.95707514, Log Avg loss: 0.92632482, Global Avg Loss: 8.04510695, Time: 0.0208 Steps: 3720, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001601, Sample Num: 25616, Cur Loss: 1.97979760, Cur Avg Loss: 0.95794148, Log Avg loss: 1.09577622, Global Avg Loss: 8.02647604, Time: 0.0208 Steps: 3730, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001611, Sample Num: 25776, Cur Loss: 0.61583728, Cur Avg Loss: 0.95791714, Log Avg loss: 0.95401997, Global Avg Loss: 8.00756573, Time: 0.0208 Steps: 3740, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001621, Sample Num: 25936, Cur Loss: 0.43184724, Cur Avg Loss: 0.95581462, Log Avg loss: 0.61709965, Global Avg Loss: 7.98785782, Time: 0.0207 Steps: 3750, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001631, Sample Num: 26096, Cur Loss: 0.78022295, Cur Avg Loss: 0.95559136, Log Avg loss: 0.91939974, Global Avg Loss: 7.96905873, Time: 0.0208 Steps: 3760, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001641, Sample Num: 26256, Cur Loss: 0.30248839, Cur Avg Loss: 0.95333617, Log Avg loss: 0.58551504, Global Avg Loss: 7.94947373, Time: 0.0208 Steps: 3770, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001651, Sample Num: 26416, Cur Loss: 0.41701168, Cur Avg Loss: 0.95120096, Log Avg loss: 0.60081354, Global Avg Loss: 7.93003283, Time: 0.0208 Steps: 3780, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001661, Sample Num: 26576, Cur Loss: 0.32082897, Cur Avg Loss: 0.94898160, Log Avg loss: 0.58256439, Global Avg Loss: 7.91064637, Time: 0.0208 Steps: 3790, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001671, Sample Num: 26736, Cur Loss: 0.77764046, Cur Avg Loss: 0.94730155, Log Avg loss: 0.66824575, Global Avg Loss: 7.89158742, Time: 0.0208 Steps: 3800, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001681, Sample Num: 26896, Cur Loss: 0.48513794, Cur Avg Loss: 0.94650842, Log Avg loss: 0.81397700, Global Avg Loss: 7.87301102, Time: 0.0208 Steps: 3810, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001691, Sample Num: 27056, Cur Loss: 0.54529881, Cur Avg Loss: 0.94711538, Log Avg loss: 1.04914433, Global Avg Loss: 7.85514749, Time: 0.0208 Steps: 3820, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001701, Sample Num: 27216, Cur Loss: 1.29897118, Cur Avg Loss: 0.94632432, Log Avg loss: 0.81255712, Global Avg Loss: 7.83675953, Time: 0.0208 Steps: 3830, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001711, Sample Num: 27376, Cur Loss: 0.23056895, Cur Avg Loss: 0.94649049, Log Avg loss: 0.97475593, Global Avg Loss: 7.81888973, Time: 0.0208 Steps: 3840, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001721, Sample Num: 27536, Cur Loss: 1.19556141, Cur Avg Loss: 0.94646827, Log Avg loss: 0.94266657, Global Avg Loss: 7.80102941, Time: 0.0208 Steps: 3850, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001731, Sample Num: 27696, Cur Loss: 0.46518892, Cur Avg Loss: 0.94443749, Log Avg loss: 0.59494019, Global Avg Loss: 7.78236078, Time: 0.0208 Steps: 3860, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001741, Sample Num: 27856, Cur Loss: 0.64444649, Cur Avg Loss: 0.94468217, Log Avg loss: 0.98703624, Global Avg Loss: 7.76480180, Time: 0.0208 Steps: 3870, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001751, Sample Num: 28016, Cur Loss: 0.43725449, Cur Avg Loss: 0.94317573, Log Avg loss: 0.68090299, Global Avg Loss: 7.74654433, Time: 0.0208 Steps: 3880, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001761, Sample Num: 28176, Cur Loss: 0.63762629, Cur Avg Loss: 0.94291636, Log Avg loss: 0.89750145, Global Avg Loss: 7.72893754, Time: 0.0208 Steps: 3890, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001771, Sample Num: 28336, Cur Loss: 1.48308265, Cur Avg Loss: 0.94386410, Log Avg loss: 1.11076036, Global Avg Loss: 7.71196785, Time: 0.0209 Steps: 3900, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001781, Sample Num: 28496, Cur Loss: 2.27934265, Cur Avg Loss: 0.94472152, Log Avg loss: 1.09657131, Global Avg Loss: 7.69504868, Time: 0.0208 Steps: 3910, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001791, Sample Num: 28656, Cur Loss: 1.13602531, Cur Avg Loss: 0.94362116, Log Avg loss: 0.74764806, Global Avg Loss: 7.67732572, Time: 0.0209 Steps: 3920, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001801, Sample Num: 28816, Cur Loss: 1.30751812, Cur Avg Loss: 0.94183289, Log Avg loss: 0.62155342, Global Avg Loss: 7.65937210, Time: 0.0209 Steps: 3930, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001811, Sample Num: 28976, Cur Loss: 1.09931481, Cur Avg Loss: 0.94117607, Log Avg loss: 0.82288223, Global Avg Loss: 7.64202060, Time: 0.0208 Steps: 3940, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001821, Sample Num: 29136, Cur Loss: 0.60758531, Cur Avg Loss: 0.93959016, Log Avg loss: 0.65238181, Global Avg Loss: 7.62432532, Time: 0.0208 Steps: 3950, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001831, Sample Num: 29296, Cur Loss: 1.14299202, Cur Avg Loss: 0.94015695, Log Avg loss: 1.04336941, Global Avg Loss: 7.60770674, Time: 0.0208 Steps: 3960, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001841, Sample Num: 29456, Cur Loss: 0.52218068, Cur Avg Loss: 0.93863014, Log Avg loss: 0.65907056, Global Avg Loss: 7.59020388, Time: 0.0209 Steps: 3970, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001851, Sample Num: 29616, Cur Loss: 0.30723760, Cur Avg Loss: 0.93768808, Log Avg loss: 0.76425613, Global Avg Loss: 7.57305326, Time: 0.0208 Steps: 3980, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001861, Sample Num: 29776, Cur Loss: 0.76625693, Cur Avg Loss: 0.93779781, Log Avg loss: 0.95810824, Global Avg Loss: 7.55647445, Time: 0.0209 Steps: 3990, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001871, Sample Num: 29936, Cur Loss: 0.70125753, Cur Avg Loss: 0.93777402, Log Avg loss: 0.93334759, Global Avg Loss: 7.53991663, Time: 0.0208 Steps: 4000, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001881, Sample Num: 30096, Cur Loss: 1.08214593, Cur Avg Loss: 0.93783197, Log Avg loss: 0.94867395, Global Avg Loss: 7.52347961, Time: 0.0209 Steps: 4010, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001891, Sample Num: 30256, Cur Loss: 1.60726810, Cur Avg Loss: 0.93740136, Log Avg loss: 0.85640298, Global Avg Loss: 7.50689485, Time: 0.0208 Steps: 4020, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001901, Sample Num: 30416, Cur Loss: 0.48662806, Cur Avg Loss: 0.93655695, Log Avg loss: 0.77687905, Global Avg Loss: 7.49019506, Time: 0.0208 Steps: 4030, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001911, Sample Num: 30576, Cur Loss: 2.32157207, Cur Avg Loss: 0.93719936, Log Avg loss: 1.05932236, Global Avg Loss: 7.47427705, Time: 0.0208 Steps: 4040, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001921, Sample Num: 30736, Cur Loss: 1.00232649, Cur Avg Loss: 0.94029512, Log Avg loss: 1.53189449, Global Avg Loss: 7.45960450, Time: 0.0208 Steps: 4050, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001931, Sample Num: 30896, Cur Loss: 0.36553761, Cur Avg Loss: 0.94041246, Log Avg loss: 0.96295393, Global Avg Loss: 7.44360290, Time: 0.0208 Steps: 4060, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001941, Sample Num: 31056, Cur Loss: 0.51329094, Cur Avg Loss: 0.93927843, Log Avg loss: 0.72029566, Global Avg Loss: 7.42708372, Time: 0.0208 Steps: 4070, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001951, Sample Num: 31216, Cur Loss: 1.29801714, Cur Avg Loss: 0.93847445, Log Avg loss: 0.78242354, Global Avg Loss: 7.41079779, Time: 0.0208 Steps: 4080, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001961, Sample Num: 31376, Cur Loss: 1.37805545, Cur Avg Loss: 0.93853606, Log Avg loss: 0.95055562, Global Avg Loss: 7.39500257, Time: 0.0208 Steps: 4090, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001971, Sample Num: 31536, Cur Loss: 0.84428871, Cur Avg Loss: 0.93757700, Log Avg loss: 0.74950601, Global Avg Loss: 7.37879405, Time: 0.0208 Steps: 4100, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001981, Sample Num: 31696, Cur Loss: 1.33256531, Cur Avg Loss: 0.93712803, Log Avg loss: 0.84863436, Global Avg Loss: 7.36290558, Time: 0.0208 Steps: 4110, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001991, Sample Num: 31856, Cur Loss: 1.38954186, Cur Avg Loss: 0.93558089, Log Avg loss: 0.62909276, Global Avg Loss: 7.34656137, Time: 0.0208 Steps: 4120, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002001, Sample Num: 32016, Cur Loss: 0.96457684, Cur Avg Loss: 0.93491099, Log Avg loss: 0.80153518, Global Avg Loss: 7.33071385, Time: 0.0208 Steps: 4130, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002011, Sample Num: 32176, Cur Loss: 0.44313174, Cur Avg Loss: 0.93316062, Log Avg loss: 0.58291134, Global Avg Loss: 7.31441481, Time: 0.0208 Steps: 4140, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002021, Sample Num: 32336, Cur Loss: 0.84027171, Cur Avg Loss: 0.93423315, Log Avg loss: 1.14991908, Global Avg Loss: 7.29956061, Time: 0.0208 Steps: 4150, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002031, Sample Num: 32496, Cur Loss: 1.91906762, Cur Avg Loss: 0.93465823, Log Avg loss: 1.02056613, Global Avg Loss: 7.28446687, Time: 0.0208 Steps: 4160, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002041, Sample Num: 32656, Cur Loss: 0.66151917, Cur Avg Loss: 0.93475493, Log Avg loss: 0.95439569, Global Avg Loss: 7.26928684, Time: 0.0208 Steps: 4170, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002051, Sample Num: 32816, Cur Loss: 0.52336341, Cur Avg Loss: 0.93394193, Log Avg loss: 0.76800742, Global Avg Loss: 7.25373354, Time: 0.0246 Steps: 4180, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002061, Sample Num: 32976, Cur Loss: 0.61018181, Cur Avg Loss: 0.93261886, Log Avg loss: 0.66125734, Global Avg Loss: 7.23799971, Time: 0.0209 Steps: 4190, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002071, Sample Num: 33136, Cur Loss: 0.68164539, Cur Avg Loss: 0.93434046, Log Avg loss: 1.28916216, Global Avg Loss: 7.22383581, Time: 0.0209 Steps: 4200, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002081, Sample Num: 33296, Cur Loss: 0.38811603, Cur Avg Loss: 0.93277828, Log Avg loss: 0.60925047, Global Avg Loss: 7.20812421, Time: 0.0208 Steps: 4210, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002091, Sample Num: 33456, Cur Loss: 1.42026997, Cur Avg Loss: 0.93316000, Log Avg loss: 1.01259731, Global Avg Loss: 7.19344286, Time: 0.0209 Steps: 4220, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002101, Sample Num: 33616, Cur Loss: 3.69524360, Cur Avg Loss: 0.93712308, Log Avg loss: 1.76580243, Global Avg Loss: 7.18061156, Time: 0.0209 Steps: 4230, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002111, Sample Num: 33776, Cur Loss: 1.63549829, Cur Avg Loss: 0.94006034, Log Avg loss: 1.55717909, Global Avg Loss: 7.16734875, Time: 0.0209 Steps: 4240, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002121, Sample Num: 33936, Cur Loss: 0.57951248, Cur Avg Loss: 0.94378476, Log Avg loss: 1.73000873, Global Avg Loss: 7.15455501, Time: 0.0209 Steps: 4250, Updated lr: 0.000097 ***** Running evaluation checkpoint-4258 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-4258 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.387188, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.626673, "eval_total_loss": 440.551084, "eval_mae": 0.549708, "eval_mse": 0.626844, "eval_r2": 0.601537, "eval_sp_statistic": 0.762036, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.813269, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.753829, "test_total_loss": 378.422201, "test_mae": 0.550101, "test_mse": 0.754106, "test_r2": 0.513293, "test_sp_statistic": 0.760984, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.845, "test_ps_pvalue": 0.0, "lr": 9.691038406827881e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 7.142569764753906, "train_cur_epoch_loss": 2007.9707444682717, "train_cur_epoch_avg_loss": 0.9431520640997049, "train_cur_epoch_time": 44.38718771934509, "train_cur_epoch_avg_time": 0.020848843456714464, "epoch": 2, "step": 4258} ################################################## Training, Epoch: 0003, Batch: 000002, Sample Num: 32, Cur Loss: 0.20796108, Cur Avg Loss: 0.37262657, Log Avg loss: 0.69485266, Global Avg Loss: 7.13939139, Time: 0.0246 Steps: 4260, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000012, Sample Num: 192, Cur Loss: 0.63455456, Cur Avg Loss: 0.61887106, Log Avg loss: 0.66811995, Global Avg Loss: 7.12423619, Time: 0.0209 Steps: 4270, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000022, Sample Num: 352, Cur Loss: 1.62183177, Cur Avg Loss: 0.81905967, Log Avg loss: 1.05928602, Global Avg Loss: 7.11006574, Time: 0.0209 Steps: 4280, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000032, Sample Num: 512, Cur Loss: 1.11300039, Cur Avg Loss: 1.00310443, Log Avg loss: 1.40800290, Global Avg Loss: 7.09677422, Time: 0.0209 Steps: 4290, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000042, Sample Num: 672, Cur Loss: 0.65989089, Cur Avg Loss: 0.99265757, Log Avg loss: 0.95922760, Global Avg Loss: 7.08250085, Time: 0.0209 Steps: 4300, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000052, Sample Num: 832, Cur Loss: 1.11965561, Cur Avg Loss: 0.98708745, Log Avg loss: 0.96369294, Global Avg Loss: 7.06830408, Time: 0.0209 Steps: 4310, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000062, Sample Num: 992, Cur Loss: 0.42614514, Cur Avg Loss: 0.98532931, Log Avg loss: 0.97618701, Global Avg Loss: 7.05420196, Time: 0.0209 Steps: 4320, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000072, Sample Num: 1152, Cur Loss: 0.78787112, Cur Avg Loss: 0.95336696, Log Avg loss: 0.75520038, Global Avg Loss: 7.03965461, Time: 0.0209 Steps: 4330, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000082, Sample Num: 1312, Cur Loss: 1.27449036, Cur Avg Loss: 0.92803564, Log Avg loss: 0.74565010, Global Avg Loss: 7.02515230, Time: 0.0209 Steps: 4340, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000092, Sample Num: 1472, Cur Loss: 0.69306153, Cur Avg Loss: 0.90433726, Log Avg loss: 0.71001053, Global Avg Loss: 7.01063473, Time: 0.0209 Steps: 4350, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000102, Sample Num: 1632, Cur Loss: 0.48484558, Cur Avg Loss: 0.89032871, Log Avg loss: 0.76145008, Global Avg Loss: 6.99630174, Time: 0.0209 Steps: 4360, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000112, Sample Num: 1792, Cur Loss: 0.98744029, Cur Avg Loss: 0.88676476, Log Avg loss: 0.85041243, Global Avg Loss: 6.98223792, Time: 0.0209 Steps: 4370, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000122, Sample Num: 1952, Cur Loss: 0.83751845, Cur Avg Loss: 0.89919611, Log Avg loss: 1.03842725, Global Avg Loss: 6.96866758, Time: 0.0210 Steps: 4380, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000132, Sample Num: 2112, Cur Loss: 0.62904024, Cur Avg Loss: 0.89140190, Log Avg loss: 0.79631257, Global Avg Loss: 6.95460754, Time: 0.0210 Steps: 4390, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000142, Sample Num: 2272, Cur Loss: 0.41874927, Cur Avg Loss: 0.88285126, Log Avg loss: 0.76998288, Global Avg Loss: 6.94055158, Time: 0.0210 Steps: 4400, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000152, Sample Num: 2432, Cur Loss: 1.14077508, Cur Avg Loss: 0.86237083, Log Avg loss: 0.57154867, Global Avg Loss: 6.92610939, Time: 0.0211 Steps: 4410, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000162, Sample Num: 2592, Cur Loss: 1.04203916, Cur Avg Loss: 0.89258694, Log Avg loss: 1.35187180, Global Avg Loss: 6.91349800, Time: 0.0209 Steps: 4420, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000172, Sample Num: 2752, Cur Loss: 1.60307109, Cur Avg Loss: 0.92901191, Log Avg loss: 1.51909651, Global Avg Loss: 6.90132102, Time: 0.0210 Steps: 4430, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000182, Sample Num: 2912, Cur Loss: 0.83100337, Cur Avg Loss: 0.94748242, Log Avg loss: 1.26517507, Global Avg Loss: 6.88862700, Time: 0.0209 Steps: 4440, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000192, Sample Num: 3072, Cur Loss: 0.85447890, Cur Avg Loss: 0.93135181, Log Avg loss: 0.63777484, Global Avg Loss: 6.87458014, Time: 0.0209 Steps: 4450, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000202, Sample Num: 3232, Cur Loss: 0.17572382, Cur Avg Loss: 0.92444628, Log Avg loss: 0.79185999, Global Avg Loss: 6.86094175, Time: 0.0209 Steps: 4460, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000212, Sample Num: 3392, Cur Loss: 0.58450377, Cur Avg Loss: 0.91429424, Log Avg loss: 0.70922298, Global Avg Loss: 6.84717952, Time: 0.0210 Steps: 4470, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000222, Sample Num: 3552, Cur Loss: 1.28993249, Cur Avg Loss: 0.91018811, Log Avg loss: 0.82313832, Global Avg Loss: 6.83373300, Time: 0.0209 Steps: 4480, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000232, Sample Num: 3712, Cur Loss: 1.52102566, Cur Avg Loss: 0.90206368, Log Avg loss: 0.72170133, Global Avg Loss: 6.82012045, Time: 0.0209 Steps: 4490, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000242, Sample Num: 3872, Cur Loss: 0.20896703, Cur Avg Loss: 0.89101719, Log Avg loss: 0.63473864, Global Avg Loss: 6.80637516, Time: 0.0209 Steps: 4500, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000252, Sample Num: 4032, Cur Loss: 0.69498420, Cur Avg Loss: 0.88340702, Log Avg loss: 0.69924071, Global Avg Loss: 6.79283384, Time: 0.0210 Steps: 4510, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000262, Sample Num: 4192, Cur Loss: 0.76605070, Cur Avg Loss: 0.88599741, Log Avg loss: 0.95127546, Global Avg Loss: 6.77991004, Time: 0.0247 Steps: 4520, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000272, Sample Num: 4352, Cur Loss: 1.77858198, Cur Avg Loss: 0.88755413, Log Avg loss: 0.92834019, Global Avg Loss: 6.76699267, Time: 0.0209 Steps: 4530, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000282, Sample Num: 4512, Cur Loss: 0.37872404, Cur Avg Loss: 0.89208732, Log Avg loss: 1.01538988, Global Avg Loss: 6.75432394, Time: 0.0209 Steps: 4540, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000292, Sample Num: 4672, Cur Loss: 0.37809163, Cur Avg Loss: 0.88009996, Log Avg loss: 0.54205653, Global Avg Loss: 6.74067060, Time: 0.0209 Steps: 4550, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000302, Sample Num: 4832, Cur Loss: 0.69832164, Cur Avg Loss: 0.87026370, Log Avg loss: 0.58304478, Global Avg Loss: 6.72716704, Time: 0.0209 Steps: 4560, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000312, Sample Num: 4992, Cur Loss: 0.52109283, Cur Avg Loss: 0.85976908, Log Avg loss: 0.54283158, Global Avg Loss: 6.71363458, Time: 0.0208 Steps: 4570, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000322, Sample Num: 5152, Cur Loss: 0.70342481, Cur Avg Loss: 0.84920933, Log Avg loss: 0.51974506, Global Avg Loss: 6.70011080, Time: 0.0209 Steps: 4580, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000332, Sample Num: 5312, Cur Loss: 1.05312681, Cur Avg Loss: 0.84849617, Log Avg loss: 0.82553251, Global Avg Loss: 6.68731215, Time: 0.0209 Steps: 4590, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000342, Sample Num: 5472, Cur Loss: 0.29080361, Cur Avg Loss: 0.84555034, Log Avg loss: 0.74774896, Global Avg Loss: 6.67440006, Time: 0.0209 Steps: 4600, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000352, Sample Num: 5632, Cur Loss: 1.11258733, Cur Avg Loss: 0.84049072, Log Avg loss: 0.66745160, Global Avg Loss: 6.66136980, Time: 0.0208 Steps: 4610, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000362, Sample Num: 5792, Cur Loss: 0.54607022, Cur Avg Loss: 0.84781474, Log Avg loss: 1.10562016, Global Avg Loss: 6.64934437, Time: 0.0209 Steps: 4620, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000372, Sample Num: 5952, Cur Loss: 0.49754125, Cur Avg Loss: 0.84526437, Log Avg loss: 0.75294100, Global Avg Loss: 6.63660916, Time: 0.0209 Steps: 4630, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000382, Sample Num: 6112, Cur Loss: 0.73145348, Cur Avg Loss: 0.84205615, Log Avg loss: 0.72271032, Global Avg Loss: 6.62386369, Time: 0.0209 Steps: 4640, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000392, Sample Num: 6272, Cur Loss: 0.62719703, Cur Avg Loss: 0.83825393, Log Avg loss: 0.69300927, Global Avg Loss: 6.61110916, Time: 0.0209 Steps: 4650, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000402, Sample Num: 6432, Cur Loss: 1.22672427, Cur Avg Loss: 0.83644303, Log Avg loss: 0.76545572, Global Avg Loss: 6.59856484, Time: 0.0209 Steps: 4660, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000412, Sample Num: 6592, Cur Loss: 0.66663325, Cur Avg Loss: 0.84040958, Log Avg loss: 0.99986507, Global Avg Loss: 6.58657619, Time: 0.0209 Steps: 4670, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000422, Sample Num: 6752, Cur Loss: 0.66591448, Cur Avg Loss: 0.84149968, Log Avg loss: 0.88641175, Global Avg Loss: 6.57439635, Time: 0.0208 Steps: 4680, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000432, Sample Num: 6912, Cur Loss: 0.57675385, Cur Avg Loss: 0.83698445, Log Avg loss: 0.64644174, Global Avg Loss: 6.56175679, Time: 0.0209 Steps: 4690, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000442, Sample Num: 7072, Cur Loss: 0.88060129, Cur Avg Loss: 0.83876492, Log Avg loss: 0.91568117, Global Avg Loss: 6.54974386, Time: 0.0209 Steps: 4700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000452, Sample Num: 7232, Cur Loss: 0.85960245, Cur Avg Loss: 0.83320779, Log Avg loss: 0.58758264, Global Avg Loss: 6.53708535, Time: 0.0209 Steps: 4710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000462, Sample Num: 7392, Cur Loss: 0.49308500, Cur Avg Loss: 0.83225358, Log Avg loss: 0.78912332, Global Avg Loss: 6.52490746, Time: 0.0209 Steps: 4720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000472, Sample Num: 7552, Cur Loss: 0.36228758, Cur Avg Loss: 0.83249941, Log Avg loss: 0.84385684, Global Avg Loss: 6.51289678, Time: 0.0209 Steps: 4730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000482, Sample Num: 7712, Cur Loss: 1.06733763, Cur Avg Loss: 0.83209324, Log Avg loss: 0.81292168, Global Avg Loss: 6.50087152, Time: 0.0209 Steps: 4740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000492, Sample Num: 7872, Cur Loss: 0.61331236, Cur Avg Loss: 0.83615861, Log Avg loss: 1.03210977, Global Avg Loss: 6.48935834, Time: 0.0209 Steps: 4750, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000502, Sample Num: 8032, Cur Loss: 0.76641828, Cur Avg Loss: 0.82973861, Log Avg loss: 0.51387448, Global Avg Loss: 6.47680480, Time: 0.0209 Steps: 4760, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000512, Sample Num: 8192, Cur Loss: 0.26966691, Cur Avg Loss: 0.82604677, Log Avg loss: 0.64071657, Global Avg Loss: 6.46456981, Time: 0.0254 Steps: 4770, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000522, Sample Num: 8352, Cur Loss: 1.32961380, Cur Avg Loss: 0.82791208, Log Avg loss: 0.92341575, Global Avg Loss: 6.45297744, Time: 0.0209 Steps: 4780, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000532, Sample Num: 8512, Cur Loss: 0.79240203, Cur Avg Loss: 0.82360640, Log Avg loss: 0.59884991, Global Avg Loss: 6.44075588, Time: 0.0209 Steps: 4790, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000542, Sample Num: 8672, Cur Loss: 1.13004470, Cur Avg Loss: 0.83149616, Log Avg loss: 1.25123131, Global Avg Loss: 6.42994437, Time: 0.0209 Steps: 4800, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000552, Sample Num: 8832, Cur Loss: 0.38197389, Cur Avg Loss: 0.83453793, Log Avg loss: 0.99940204, Global Avg Loss: 6.41865426, Time: 0.0209 Steps: 4810, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000562, Sample Num: 8992, Cur Loss: 0.72216976, Cur Avg Loss: 0.83750800, Log Avg loss: 1.00145551, Global Avg Loss: 6.40741526, Time: 0.0209 Steps: 4820, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000572, Sample Num: 9152, Cur Loss: 0.45978284, Cur Avg Loss: 0.83268800, Log Avg loss: 0.56180419, Global Avg Loss: 6.39531255, Time: 0.0209 Steps: 4830, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000582, Sample Num: 9312, Cur Loss: 0.51771796, Cur Avg Loss: 0.82727932, Log Avg loss: 0.51790255, Global Avg Loss: 6.38316914, Time: 0.0209 Steps: 4840, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000592, Sample Num: 9472, Cur Loss: 0.36513317, Cur Avg Loss: 0.82175793, Log Avg loss: 0.50041355, Global Avg Loss: 6.37103974, Time: 0.0209 Steps: 4850, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000602, Sample Num: 9632, Cur Loss: 0.33295843, Cur Avg Loss: 0.81683516, Log Avg loss: 0.52540720, Global Avg Loss: 6.35901169, Time: 0.0209 Steps: 4860, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000612, Sample Num: 9792, Cur Loss: 1.25169063, Cur Avg Loss: 0.82288281, Log Avg loss: 1.18695099, Global Avg Loss: 6.34839145, Time: 0.0209 Steps: 4870, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000622, Sample Num: 9952, Cur Loss: 0.20366806, Cur Avg Loss: 0.82592868, Log Avg loss: 1.01233586, Global Avg Loss: 6.33745690, Time: 0.0209 Steps: 4880, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000632, Sample Num: 10112, Cur Loss: 0.86891246, Cur Avg Loss: 0.82430227, Log Avg loss: 0.72313997, Global Avg Loss: 6.32597568, Time: 0.0209 Steps: 4890, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000642, Sample Num: 10272, Cur Loss: 0.90025139, Cur Avg Loss: 0.82149083, Log Avg loss: 0.64380768, Global Avg Loss: 6.31437942, Time: 0.0209 Steps: 4900, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000652, Sample Num: 10432, Cur Loss: 0.43558624, Cur Avg Loss: 0.81833042, Log Avg loss: 0.61543228, Global Avg Loss: 6.30277261, Time: 0.0209 Steps: 4910, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000662, Sample Num: 10592, Cur Loss: 0.95587301, Cur Avg Loss: 0.81722157, Log Avg loss: 0.74492399, Global Avg Loss: 6.29147617, Time: 0.0208 Steps: 4920, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000672, Sample Num: 10752, Cur Loss: 0.40435457, Cur Avg Loss: 0.81716172, Log Avg loss: 0.81319982, Global Avg Loss: 6.28036404, Time: 0.0209 Steps: 4930, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000682, Sample Num: 10912, Cur Loss: 0.30796376, Cur Avg Loss: 0.81697190, Log Avg loss: 0.80421635, Global Avg Loss: 6.26927872, Time: 0.0209 Steps: 4940, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000692, Sample Num: 11072, Cur Loss: 1.42555368, Cur Avg Loss: 0.82181801, Log Avg loss: 1.15232234, Global Avg Loss: 6.25894144, Time: 0.0209 Steps: 4950, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000702, Sample Num: 11232, Cur Loss: 0.40315998, Cur Avg Loss: 0.82064527, Log Avg loss: 0.73949196, Global Avg Loss: 6.24781352, Time: 0.0209 Steps: 4960, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000712, Sample Num: 11392, Cur Loss: 0.91706526, Cur Avg Loss: 0.82266735, Log Avg loss: 0.96461710, Global Avg Loss: 6.23718334, Time: 0.0209 Steps: 4970, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000722, Sample Num: 11552, Cur Loss: 1.37190461, Cur Avg Loss: 0.82663955, Log Avg loss: 1.10946028, Global Avg Loss: 6.22688671, Time: 0.0209 Steps: 4980, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000732, Sample Num: 11712, Cur Loss: 1.00155783, Cur Avg Loss: 0.82654344, Log Avg loss: 0.81960400, Global Avg Loss: 6.21605047, Time: 0.0209 Steps: 4990, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000742, Sample Num: 11872, Cur Loss: 0.46369728, Cur Avg Loss: 0.82291370, Log Avg loss: 0.55721705, Global Avg Loss: 6.20473280, Time: 0.0209 Steps: 5000, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000752, Sample Num: 12032, Cur Loss: 0.94850373, Cur Avg Loss: 0.82122362, Log Avg loss: 0.69582001, Global Avg Loss: 6.19373697, Time: 0.0208 Steps: 5010, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000762, Sample Num: 12192, Cur Loss: 1.56538343, Cur Avg Loss: 0.82169468, Log Avg loss: 0.85711841, Global Avg Loss: 6.18310626, Time: 0.0209 Steps: 5020, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000772, Sample Num: 12352, Cur Loss: 3.18158245, Cur Avg Loss: 0.83661617, Log Avg loss: 1.97363366, Global Avg Loss: 6.17473752, Time: 0.0246 Steps: 5030, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000782, Sample Num: 12512, Cur Loss: 1.29117262, Cur Avg Loss: 0.83950424, Log Avg loss: 1.06246307, Global Avg Loss: 6.16459412, Time: 0.0209 Steps: 5040, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000792, Sample Num: 12672, Cur Loss: 0.64384234, Cur Avg Loss: 0.83793621, Log Avg loss: 0.71531592, Global Avg Loss: 6.15380347, Time: 0.0209 Steps: 5050, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000802, Sample Num: 12832, Cur Loss: 0.24978071, Cur Avg Loss: 0.83353813, Log Avg loss: 0.48521078, Global Avg Loss: 6.14260072, Time: 0.0208 Steps: 5060, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000812, Sample Num: 12992, Cur Loss: 0.81455070, Cur Avg Loss: 0.83570671, Log Avg loss: 1.00962661, Global Avg Loss: 6.13247651, Time: 0.0209 Steps: 5070, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000822, Sample Num: 13152, Cur Loss: 0.66525501, Cur Avg Loss: 0.83228876, Log Avg loss: 0.55475139, Global Avg Loss: 6.12149674, Time: 0.0209 Steps: 5080, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000832, Sample Num: 13312, Cur Loss: 0.59927016, Cur Avg Loss: 0.83008370, Log Avg loss: 0.64882724, Global Avg Loss: 6.11074493, Time: 0.0209 Steps: 5090, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000842, Sample Num: 13472, Cur Loss: 1.63809204, Cur Avg Loss: 0.84650451, Log Avg loss: 2.21271571, Global Avg Loss: 6.10310174, Time: 0.0208 Steps: 5100, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000852, Sample Num: 13632, Cur Loss: 0.95744711, Cur Avg Loss: 0.85046345, Log Avg loss: 1.18380632, Global Avg Loss: 6.09347493, Time: 0.0209 Steps: 5110, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000862, Sample Num: 13792, Cur Loss: 0.27775091, Cur Avg Loss: 0.85071651, Log Avg loss: 0.87227754, Global Avg Loss: 6.08327728, Time: 0.0210 Steps: 5120, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000872, Sample Num: 13952, Cur Loss: 0.58417404, Cur Avg Loss: 0.84955905, Log Avg loss: 0.74978570, Global Avg Loss: 6.07288061, Time: 0.0211 Steps: 5130, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000882, Sample Num: 14112, Cur Loss: 0.95560759, Cur Avg Loss: 0.84863491, Log Avg loss: 0.76804987, Global Avg Loss: 6.06255993, Time: 0.0208 Steps: 5140, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000892, Sample Num: 14272, Cur Loss: 0.96088767, Cur Avg Loss: 0.84696110, Log Avg loss: 0.69933152, Global Avg Loss: 6.05214590, Time: 0.0208 Steps: 5150, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000902, Sample Num: 14432, Cur Loss: 0.59846789, Cur Avg Loss: 0.84356301, Log Avg loss: 0.54045321, Global Avg Loss: 6.04146432, Time: 0.0209 Steps: 5160, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000912, Sample Num: 14592, Cur Loss: 1.19667411, Cur Avg Loss: 0.84434613, Log Avg loss: 0.91498377, Global Avg Loss: 6.03154850, Time: 0.0209 Steps: 5170, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000922, Sample Num: 14752, Cur Loss: 0.36512625, Cur Avg Loss: 0.84190700, Log Avg loss: 0.61945778, Global Avg Loss: 6.02110045, Time: 0.0208 Steps: 5180, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000932, Sample Num: 14912, Cur Loss: 0.24892095, Cur Avg Loss: 0.83899232, Log Avg loss: 0.57025897, Global Avg Loss: 6.01059786, Time: 0.0208 Steps: 5190, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000942, Sample Num: 15072, Cur Loss: 0.97230244, Cur Avg Loss: 0.83774261, Log Avg loss: 0.72126950, Global Avg Loss: 6.00042608, Time: 0.0209 Steps: 5200, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000952, Sample Num: 15232, Cur Loss: 0.78960252, Cur Avg Loss: 0.83645117, Log Avg loss: 0.71479777, Global Avg Loss: 5.99028092, Time: 0.0208 Steps: 5210, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000962, Sample Num: 15392, Cur Loss: 0.57546091, Cur Avg Loss: 0.83596129, Log Avg loss: 0.78932481, Global Avg Loss: 5.98031740, Time: 0.0209 Steps: 5220, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000972, Sample Num: 15552, Cur Loss: 0.63219631, Cur Avg Loss: 0.83649470, Log Avg loss: 0.88780840, Global Avg Loss: 5.97058029, Time: 0.0209 Steps: 5230, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000982, Sample Num: 15712, Cur Loss: 0.54272503, Cur Avg Loss: 0.83625066, Log Avg loss: 0.81253016, Global Avg Loss: 5.96073668, Time: 0.0209 Steps: 5240, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000992, Sample Num: 15872, Cur Loss: 0.45718694, Cur Avg Loss: 0.83502517, Log Avg loss: 0.71468220, Global Avg Loss: 5.95074420, Time: 0.0209 Steps: 5250, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001002, Sample Num: 16032, Cur Loss: 0.40545681, Cur Avg Loss: 0.83340837, Log Avg loss: 0.67302170, Global Avg Loss: 5.94071050, Time: 0.0208 Steps: 5260, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001012, Sample Num: 16192, Cur Loss: 0.88155609, Cur Avg Loss: 0.83221330, Log Avg loss: 0.71246768, Global Avg Loss: 5.93078974, Time: 0.0208 Steps: 5270, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001022, Sample Num: 16352, Cur Loss: 0.38218188, Cur Avg Loss: 0.83099737, Log Avg loss: 0.70794501, Global Avg Loss: 5.92089799, Time: 0.0209 Steps: 5280, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001032, Sample Num: 16512, Cur Loss: 0.75074768, Cur Avg Loss: 0.83352462, Log Avg loss: 1.09180951, Global Avg Loss: 5.91176928, Time: 0.0209 Steps: 5290, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001042, Sample Num: 16672, Cur Loss: 0.89861125, Cur Avg Loss: 0.83327752, Log Avg loss: 0.80777707, Global Avg Loss: 5.90213910, Time: 0.0210 Steps: 5300, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001052, Sample Num: 16832, Cur Loss: 0.81187731, Cur Avg Loss: 0.83280455, Log Avg loss: 0.78352115, Global Avg Loss: 5.89249952, Time: 0.0209 Steps: 5310, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001062, Sample Num: 16992, Cur Loss: 1.19349480, Cur Avg Loss: 0.83177410, Log Avg loss: 0.72337006, Global Avg Loss: 5.88278311, Time: 0.0210 Steps: 5320, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001072, Sample Num: 17152, Cur Loss: 0.37972981, Cur Avg Loss: 0.82798565, Log Avg loss: 0.42565246, Global Avg Loss: 5.87254459, Time: 0.0209 Steps: 5330, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001082, Sample Num: 17312, Cur Loss: 0.55987895, Cur Avg Loss: 0.82519072, Log Avg loss: 0.52557481, Global Avg Loss: 5.86253154, Time: 0.0209 Steps: 5340, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001092, Sample Num: 17472, Cur Loss: 0.43829718, Cur Avg Loss: 0.82347915, Log Avg loss: 0.63828678, Global Avg Loss: 5.85276660, Time: 0.0209 Steps: 5350, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001102, Sample Num: 17632, Cur Loss: 1.00174570, Cur Avg Loss: 0.82116212, Log Avg loss: 0.56814227, Global Avg Loss: 5.84290722, Time: 0.0209 Steps: 5360, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001112, Sample Num: 17792, Cur Loss: 0.61519068, Cur Avg Loss: 0.81838608, Log Avg loss: 0.51246720, Global Avg Loss: 5.83298089, Time: 0.0210 Steps: 5370, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001122, Sample Num: 17952, Cur Loss: 0.34197646, Cur Avg Loss: 0.81728225, Log Avg loss: 0.69453597, Global Avg Loss: 5.82342988, Time: 0.0209 Steps: 5380, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001132, Sample Num: 18112, Cur Loss: 1.90758336, Cur Avg Loss: 0.82212746, Log Avg loss: 1.36575951, Global Avg Loss: 5.81515962, Time: 0.0209 Steps: 5390, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001142, Sample Num: 18272, Cur Loss: 0.22227861, Cur Avg Loss: 0.82371106, Log Avg loss: 1.00297544, Global Avg Loss: 5.80624817, Time: 0.0209 Steps: 5400, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001152, Sample Num: 18432, Cur Loss: 0.43217278, Cur Avg Loss: 0.82181507, Log Avg loss: 0.60529220, Global Avg Loss: 5.79663457, Time: 0.0210 Steps: 5410, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001162, Sample Num: 18592, Cur Loss: 0.66786784, Cur Avg Loss: 0.82119628, Log Avg loss: 0.74991185, Global Avg Loss: 5.78732327, Time: 0.0209 Steps: 5420, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001172, Sample Num: 18752, Cur Loss: 0.58348060, Cur Avg Loss: 0.81789843, Log Avg loss: 0.43468790, Global Avg Loss: 5.77746575, Time: 0.0209 Steps: 5430, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001182, Sample Num: 18912, Cur Loss: 0.16292322, Cur Avg Loss: 0.81605384, Log Avg loss: 0.59986864, Global Avg Loss: 5.76794811, Time: 0.0209 Steps: 5440, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001192, Sample Num: 19072, Cur Loss: 0.92947376, Cur Avg Loss: 0.81438307, Log Avg loss: 0.61689742, Global Avg Loss: 5.75849664, Time: 0.0209 Steps: 5450, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001202, Sample Num: 19232, Cur Loss: 0.99909484, Cur Avg Loss: 0.81638629, Log Avg loss: 1.05517092, Global Avg Loss: 5.74988249, Time: 0.0209 Steps: 5460, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001212, Sample Num: 19392, Cur Loss: 0.89585125, Cur Avg Loss: 0.81568735, Log Avg loss: 0.73167435, Global Avg Loss: 5.74070843, Time: 0.0209 Steps: 5470, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001222, Sample Num: 19552, Cur Loss: 0.84124261, Cur Avg Loss: 0.81451526, Log Avg loss: 0.67245785, Global Avg Loss: 5.73145980, Time: 0.0209 Steps: 5480, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001232, Sample Num: 19712, Cur Loss: 3.23128939, Cur Avg Loss: 0.82128201, Log Avg loss: 1.64817901, Global Avg Loss: 5.72402213, Time: 0.0210 Steps: 5490, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001242, Sample Num: 19872, Cur Loss: 1.11178374, Cur Avg Loss: 0.83102903, Log Avg loss: 2.03186234, Global Avg Loss: 5.71730911, Time: 0.0209 Steps: 5500, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001252, Sample Num: 20032, Cur Loss: 0.67236471, Cur Avg Loss: 0.83571044, Log Avg loss: 1.41714050, Global Avg Loss: 5.70950481, Time: 0.0210 Steps: 5510, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001262, Sample Num: 20192, Cur Loss: 0.94599313, Cur Avg Loss: 0.83925696, Log Avg loss: 1.28328197, Global Avg Loss: 5.70148629, Time: 0.0209 Steps: 5520, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001272, Sample Num: 20352, Cur Loss: 1.51989603, Cur Avg Loss: 0.83928690, Log Avg loss: 0.84306566, Global Avg Loss: 5.69270072, Time: 0.0209 Steps: 5530, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001282, Sample Num: 20512, Cur Loss: 0.75375801, Cur Avg Loss: 0.83762124, Log Avg loss: 0.62574892, Global Avg Loss: 5.68355460, Time: 0.0248 Steps: 5540, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001292, Sample Num: 20672, Cur Loss: 0.35947073, Cur Avg Loss: 0.83676130, Log Avg loss: 0.72651682, Global Avg Loss: 5.67462300, Time: 0.0209 Steps: 5550, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001302, Sample Num: 20832, Cur Loss: 0.36783171, Cur Avg Loss: 0.83965156, Log Avg loss: 1.21307318, Global Avg Loss: 5.66659863, Time: 0.0210 Steps: 5560, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001312, Sample Num: 20992, Cur Loss: 0.74093843, Cur Avg Loss: 0.84495996, Log Avg loss: 1.53611430, Global Avg Loss: 5.65918304, Time: 0.0209 Steps: 5570, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001322, Sample Num: 21152, Cur Loss: 0.60565519, Cur Avg Loss: 0.84788018, Log Avg loss: 1.23101245, Global Avg Loss: 5.65124725, Time: 0.0209 Steps: 5580, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001332, Sample Num: 21312, Cur Loss: 1.28093529, Cur Avg Loss: 0.84689045, Log Avg loss: 0.71604773, Global Avg Loss: 5.64241863, Time: 0.0209 Steps: 5590, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001342, Sample Num: 21472, Cur Loss: 1.33972096, Cur Avg Loss: 0.84530784, Log Avg loss: 0.63450516, Global Avg Loss: 5.63347593, Time: 0.0209 Steps: 5600, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001352, Sample Num: 21632, Cur Loss: 0.40625590, Cur Avg Loss: 0.84313761, Log Avg loss: 0.55189234, Global Avg Loss: 5.62441784, Time: 0.0210 Steps: 5610, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001362, Sample Num: 21792, Cur Loss: 0.66702300, Cur Avg Loss: 0.84140501, Log Avg loss: 0.60715789, Global Avg Loss: 5.61549034, Time: 0.0209 Steps: 5620, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001372, Sample Num: 21952, Cur Loss: 0.83390868, Cur Avg Loss: 0.84005543, Log Avg loss: 0.65624239, Global Avg Loss: 5.60668172, Time: 0.0209 Steps: 5630, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001382, Sample Num: 22112, Cur Loss: 0.65891409, Cur Avg Loss: 0.83772487, Log Avg loss: 0.51797200, Global Avg Loss: 5.59765919, Time: 0.0209 Steps: 5640, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001392, Sample Num: 22272, Cur Loss: 1.47038758, Cur Avg Loss: 0.83942193, Log Avg loss: 1.07395601, Global Avg Loss: 5.58965264, Time: 0.0209 Steps: 5650, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001402, Sample Num: 22432, Cur Loss: 1.05333948, Cur Avg Loss: 0.83925256, Log Avg loss: 0.81567571, Global Avg Loss: 5.58121805, Time: 0.0209 Steps: 5660, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001412, Sample Num: 22592, Cur Loss: 0.48445630, Cur Avg Loss: 0.84049531, Log Avg loss: 1.01472891, Global Avg Loss: 5.57316427, Time: 0.0209 Steps: 5670, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001422, Sample Num: 22752, Cur Loss: 1.77763224, Cur Avg Loss: 0.84157061, Log Avg loss: 0.99340349, Global Avg Loss: 5.56510132, Time: 0.0209 Steps: 5680, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001432, Sample Num: 22912, Cur Loss: 0.32038015, Cur Avg Loss: 0.84169607, Log Avg loss: 0.85953553, Global Avg Loss: 5.55683143, Time: 0.0209 Steps: 5690, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001442, Sample Num: 23072, Cur Loss: 0.44182077, Cur Avg Loss: 0.84061760, Log Avg loss: 0.68618090, Global Avg Loss: 5.54828643, Time: 0.0209 Steps: 5700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001452, Sample Num: 23232, Cur Loss: 0.23816638, Cur Avg Loss: 0.83970501, Log Avg loss: 0.70811032, Global Avg Loss: 5.53980976, Time: 0.0210 Steps: 5710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001462, Sample Num: 23392, Cur Loss: 0.98864007, Cur Avg Loss: 0.83776075, Log Avg loss: 0.55545362, Global Avg Loss: 5.53109585, Time: 0.0209 Steps: 5720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001472, Sample Num: 23552, Cur Loss: 2.66829681, Cur Avg Loss: 0.83995550, Log Avg loss: 1.16082730, Global Avg Loss: 5.52346886, Time: 0.0209 Steps: 5730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001482, Sample Num: 23712, Cur Loss: 0.26884872, Cur Avg Loss: 0.83727555, Log Avg loss: 0.44278726, Global Avg Loss: 5.51461749, Time: 0.0209 Steps: 5740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001492, Sample Num: 23872, Cur Loss: 0.59499729, Cur Avg Loss: 0.83573676, Log Avg loss: 0.60768878, Global Avg Loss: 5.50608371, Time: 0.0209 Steps: 5750, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001502, Sample Num: 24032, Cur Loss: 1.65764248, Cur Avg Loss: 0.83618112, Log Avg loss: 0.90247971, Global Avg Loss: 5.49809134, Time: 0.0209 Steps: 5760, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001512, Sample Num: 24192, Cur Loss: 0.47102398, Cur Avg Loss: 0.83505467, Log Avg loss: 0.66586164, Global Avg Loss: 5.48971659, Time: 0.0210 Steps: 5770, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001522, Sample Num: 24352, Cur Loss: 0.56782568, Cur Avg Loss: 0.83378247, Log Avg loss: 0.64142573, Global Avg Loss: 5.48132854, Time: 0.0209 Steps: 5780, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001532, Sample Num: 24512, Cur Loss: 1.47623551, Cur Avg Loss: 0.83202655, Log Avg loss: 0.56477558, Global Avg Loss: 5.47283709, Time: 0.0209 Steps: 5790, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001542, Sample Num: 24672, Cur Loss: 1.96737027, Cur Avg Loss: 0.83657656, Log Avg loss: 1.53363777, Global Avg Loss: 5.46604536, Time: 0.0213 Steps: 5800, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001552, Sample Num: 24832, Cur Loss: 0.55215275, Cur Avg Loss: 0.83632132, Log Avg loss: 0.79696344, Global Avg Loss: 5.45800908, Time: 0.0212 Steps: 5810, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001562, Sample Num: 24992, Cur Loss: 0.45240563, Cur Avg Loss: 0.83833306, Log Avg loss: 1.15055469, Global Avg Loss: 5.45060795, Time: 0.0212 Steps: 5820, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001572, Sample Num: 25152, Cur Loss: 0.49396384, Cur Avg Loss: 0.83726819, Log Avg loss: 0.67093647, Global Avg Loss: 5.44240955, Time: 0.0212 Steps: 5830, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001582, Sample Num: 25312, Cur Loss: 3.58099103, Cur Avg Loss: 0.84104094, Log Avg loss: 1.43411671, Global Avg Loss: 5.43554603, Time: 0.0213 Steps: 5840, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001592, Sample Num: 25472, Cur Loss: 3.51761341, Cur Avg Loss: 0.84659171, Log Avg loss: 1.72472375, Global Avg Loss: 5.42920275, Time: 0.0212 Steps: 5850, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001602, Sample Num: 25632, Cur Loss: 0.64294004, Cur Avg Loss: 0.84856461, Log Avg loss: 1.16264926, Global Avg Loss: 5.42192194, Time: 0.0212 Steps: 5860, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001612, Sample Num: 25792, Cur Loss: 1.18632817, Cur Avg Loss: 0.84926847, Log Avg loss: 0.96202740, Global Avg Loss: 5.41432416, Time: 0.0212 Steps: 5870, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001622, Sample Num: 25952, Cur Loss: 0.36441791, Cur Avg Loss: 0.84986992, Log Avg loss: 0.94682337, Global Avg Loss: 5.40672637, Time: 0.0212 Steps: 5880, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001632, Sample Num: 26112, Cur Loss: 0.42722398, Cur Avg Loss: 0.84759356, Log Avg loss: 0.47836878, Global Avg Loss: 5.39835904, Time: 0.0212 Steps: 5890, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001642, Sample Num: 26272, Cur Loss: 1.39549053, Cur Avg Loss: 0.84751030, Log Avg loss: 0.83392113, Global Avg Loss: 5.39062271, Time: 0.0212 Steps: 5900, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001652, Sample Num: 26432, Cur Loss: 0.90168214, Cur Avg Loss: 0.84699720, Log Avg loss: 0.76274635, Global Avg Loss: 5.38279212, Time: 0.0212 Steps: 5910, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001662, Sample Num: 26592, Cur Loss: 0.42169410, Cur Avg Loss: 0.84483899, Log Avg loss: 0.48830288, Global Avg Loss: 5.37452440, Time: 0.0212 Steps: 5920, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001672, Sample Num: 26752, Cur Loss: 1.08834171, Cur Avg Loss: 0.84454369, Log Avg loss: 0.79546452, Global Avg Loss: 5.36680255, Time: 0.0212 Steps: 5930, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001682, Sample Num: 26912, Cur Loss: 0.65621567, Cur Avg Loss: 0.84706541, Log Avg loss: 1.26869828, Global Avg Loss: 5.35990338, Time: 0.0212 Steps: 5940, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001692, Sample Num: 27072, Cur Loss: 0.91033322, Cur Avg Loss: 0.84608333, Log Avg loss: 0.68089723, Global Avg Loss: 5.35203951, Time: 0.0212 Steps: 5950, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001702, Sample Num: 27232, Cur Loss: 0.55666351, Cur Avg Loss: 0.84480118, Log Avg loss: 0.62786131, Global Avg Loss: 5.34411303, Time: 0.0212 Steps: 5960, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001712, Sample Num: 27392, Cur Loss: 1.29281569, Cur Avg Loss: 0.84313504, Log Avg loss: 0.55955782, Global Avg Loss: 5.33609870, Time: 0.0212 Steps: 5970, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001722, Sample Num: 27552, Cur Loss: 0.30055374, Cur Avg Loss: 0.84109752, Log Avg loss: 0.49227475, Global Avg Loss: 5.32799866, Time: 0.0212 Steps: 5980, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001732, Sample Num: 27712, Cur Loss: 0.86851311, Cur Avg Loss: 0.83956761, Log Avg loss: 0.57611722, Global Avg Loss: 5.32006564, Time: 0.0212 Steps: 5990, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001742, Sample Num: 27872, Cur Loss: 1.04910111, Cur Avg Loss: 0.83878790, Log Avg loss: 0.70374098, Global Avg Loss: 5.31237176, Time: 0.0212 Steps: 6000, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001752, Sample Num: 28032, Cur Loss: 0.39679927, Cur Avg Loss: 0.83690317, Log Avg loss: 0.50858321, Global Avg Loss: 5.30437877, Time: 0.0212 Steps: 6010, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001762, Sample Num: 28192, Cur Loss: 0.97108018, Cur Avg Loss: 0.83599886, Log Avg loss: 0.67756445, Global Avg Loss: 5.29669303, Time: 0.0212 Steps: 6020, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001772, Sample Num: 28352, Cur Loss: 0.36820439, Cur Avg Loss: 0.83455908, Log Avg loss: 0.58086974, Global Avg Loss: 5.28887243, Time: 0.0212 Steps: 6030, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001782, Sample Num: 28512, Cur Loss: 0.57889295, Cur Avg Loss: 0.83634058, Log Avg loss: 1.15202297, Global Avg Loss: 5.28202334, Time: 0.0212 Steps: 6040, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001792, Sample Num: 28672, Cur Loss: 0.32253376, Cur Avg Loss: 0.83517786, Log Avg loss: 0.62798089, Global Avg Loss: 5.27433071, Time: 0.0257 Steps: 6050, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001802, Sample Num: 28832, Cur Loss: 1.45789135, Cur Avg Loss: 0.83513209, Log Avg loss: 0.82692916, Global Avg Loss: 5.26699176, Time: 0.0209 Steps: 6060, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001812, Sample Num: 28992, Cur Loss: 0.63497102, Cur Avg Loss: 0.83477632, Log Avg loss: 0.77066690, Global Avg Loss: 5.25958431, Time: 0.0209 Steps: 6070, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001822, Sample Num: 29152, Cur Loss: 0.43401474, Cur Avg Loss: 0.83446520, Log Avg loss: 0.77809019, Global Avg Loss: 5.25221343, Time: 0.0209 Steps: 6080, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001832, Sample Num: 29312, Cur Loss: 0.26677987, Cur Avg Loss: 0.83279041, Log Avg loss: 0.52764297, Global Avg Loss: 5.24445551, Time: 0.0209 Steps: 6090, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001842, Sample Num: 29472, Cur Loss: 1.41790521, Cur Avg Loss: 0.83295753, Log Avg loss: 0.86357431, Global Avg Loss: 5.23727374, Time: 0.0209 Steps: 6100, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001852, Sample Num: 29632, Cur Loss: 0.79597795, Cur Avg Loss: 0.83750830, Log Avg loss: 1.67576053, Global Avg Loss: 5.23144475, Time: 0.0210 Steps: 6110, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001862, Sample Num: 29792, Cur Loss: 1.12674749, Cur Avg Loss: 0.83792223, Log Avg loss: 0.91458153, Global Avg Loss: 5.22439105, Time: 0.0210 Steps: 6120, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001872, Sample Num: 29952, Cur Loss: 0.16212887, Cur Avg Loss: 0.83632377, Log Avg loss: 0.53869087, Global Avg Loss: 5.21674717, Time: 0.0212 Steps: 6130, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001882, Sample Num: 30112, Cur Loss: 2.57002687, Cur Avg Loss: 0.83675093, Log Avg loss: 0.91671484, Global Avg Loss: 5.20974386, Time: 0.0209 Steps: 6140, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001892, Sample Num: 30272, Cur Loss: 0.67069590, Cur Avg Loss: 0.83665179, Log Avg loss: 0.81799334, Global Avg Loss: 5.20260280, Time: 0.0209 Steps: 6150, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001902, Sample Num: 30432, Cur Loss: 0.39154464, Cur Avg Loss: 0.83587693, Log Avg loss: 0.68927383, Global Avg Loss: 5.19527597, Time: 0.0210 Steps: 6160, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001912, Sample Num: 30592, Cur Loss: 1.14202416, Cur Avg Loss: 0.83568921, Log Avg loss: 0.79998482, Global Avg Loss: 5.18815232, Time: 0.0211 Steps: 6170, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001922, Sample Num: 30752, Cur Loss: 0.80962533, Cur Avg Loss: 0.83436787, Log Avg loss: 0.58172861, Global Avg Loss: 5.18069856, Time: 0.0210 Steps: 6180, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001932, Sample Num: 30912, Cur Loss: 0.39167130, Cur Avg Loss: 0.83310607, Log Avg loss: 0.59058810, Global Avg Loss: 5.17328320, Time: 0.0209 Steps: 6190, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001942, Sample Num: 31072, Cur Loss: 0.27252781, Cur Avg Loss: 0.83421242, Log Avg loss: 1.04795800, Global Avg Loss: 5.16662945, Time: 0.0210 Steps: 6200, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001952, Sample Num: 31232, Cur Loss: 1.11873269, Cur Avg Loss: 0.83576729, Log Avg loss: 1.13772481, Global Avg Loss: 5.16014168, Time: 0.0210 Steps: 6210, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001962, Sample Num: 31392, Cur Loss: 2.02140713, Cur Avg Loss: 0.83736573, Log Avg loss: 1.14937979, Global Avg Loss: 5.15369351, Time: 0.0209 Steps: 6220, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001972, Sample Num: 31552, Cur Loss: 0.45396391, Cur Avg Loss: 0.83620009, Log Avg loss: 0.60750192, Global Avg Loss: 5.14639625, Time: 0.0210 Steps: 6230, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001982, Sample Num: 31712, Cur Loss: 0.52408493, Cur Avg Loss: 0.83593861, Log Avg loss: 0.78437572, Global Avg Loss: 5.13940583, Time: 0.0210 Steps: 6240, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001992, Sample Num: 31872, Cur Loss: 0.67956066, Cur Avg Loss: 0.83521871, Log Avg loss: 0.69253401, Global Avg Loss: 5.13229084, Time: 0.0209 Steps: 6250, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002002, Sample Num: 32032, Cur Loss: 0.30592573, Cur Avg Loss: 0.83529403, Log Avg loss: 0.85029776, Global Avg Loss: 5.12545059, Time: 0.0209 Steps: 6260, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002012, Sample Num: 32192, Cur Loss: 0.55411816, Cur Avg Loss: 0.83563814, Log Avg loss: 0.90452855, Global Avg Loss: 5.11871866, Time: 0.0210 Steps: 6270, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002022, Sample Num: 32352, Cur Loss: 0.47999346, Cur Avg Loss: 0.83406576, Log Avg loss: 0.51770283, Global Avg Loss: 5.11139220, Time: 0.0210 Steps: 6280, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002032, Sample Num: 32512, Cur Loss: 0.60255426, Cur Avg Loss: 0.83336024, Log Avg loss: 0.69070495, Global Avg Loss: 5.10436408, Time: 0.0210 Steps: 6290, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002042, Sample Num: 32672, Cur Loss: 0.64294660, Cur Avg Loss: 0.83295687, Log Avg loss: 0.75099081, Global Avg Loss: 5.09745397, Time: 0.0209 Steps: 6300, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002052, Sample Num: 32832, Cur Loss: 0.41145578, Cur Avg Loss: 0.83329705, Log Avg loss: 0.90276312, Global Avg Loss: 5.09080628, Time: 0.0247 Steps: 6310, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002062, Sample Num: 32992, Cur Loss: 0.30758724, Cur Avg Loss: 0.83174392, Log Avg loss: 0.51304095, Global Avg Loss: 5.08356298, Time: 0.0210 Steps: 6320, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002072, Sample Num: 33152, Cur Loss: 0.69344914, Cur Avg Loss: 0.83114140, Log Avg loss: 0.70690248, Global Avg Loss: 5.07664882, Time: 0.0210 Steps: 6330, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002082, Sample Num: 33312, Cur Loss: 0.35787004, Cur Avg Loss: 0.83239284, Log Avg loss: 1.09168995, Global Avg Loss: 5.07036340, Time: 0.0210 Steps: 6340, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002092, Sample Num: 33472, Cur Loss: 0.71728581, Cur Avg Loss: 0.83124510, Log Avg loss: 0.59228568, Global Avg Loss: 5.06331131, Time: 0.0210 Steps: 6350, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002102, Sample Num: 33632, Cur Loss: 0.84136128, Cur Avg Loss: 0.82980375, Log Avg loss: 0.52827437, Global Avg Loss: 5.05618075, Time: 0.0211 Steps: 6360, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002112, Sample Num: 33792, Cur Loss: 1.28876090, Cur Avg Loss: 0.83193818, Log Avg loss: 1.28059481, Global Avg Loss: 5.05025361, Time: 0.0210 Steps: 6370, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002122, Sample Num: 33952, Cur Loss: 0.58334517, Cur Avg Loss: 0.83155582, Log Avg loss: 0.75080221, Global Avg Loss: 5.04351466, Time: 0.0209 Steps: 6380, Updated lr: 0.000095 ***** Running evaluation checkpoint-6387 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-6387 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.843865, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.715582, "eval_total_loss": 503.054106, "eval_mae": 0.715359, "eval_mse": 0.715426, "eval_r2": 0.545228, "eval_sp_statistic": 0.793375, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.836624, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.458787, "test_total_loss": 230.311118, "test_mae": 0.447726, "test_mse": 0.458935, "test_r2": 0.703799, "test_sp_statistic": 0.791295, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.869626, "test_ps_pvalue": 0.0, "lr": 9.489141773352301e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 5.038378395694269, "train_cur_epoch_loss": 1767.0607549771667, "train_cur_epoch_avg_loss": 0.829995657574996, "train_cur_epoch_time": 44.843865156173706, "train_cur_epoch_avg_time": 0.02106334671497121, "epoch": 3, "step": 6387} ################################################## Training, Epoch: 0004, Batch: 000003, Sample Num: 48, Cur Loss: 1.08808768, Cur Avg Loss: 0.93533260, Log Avg loss: 0.53052948, Global Avg Loss: 5.03645208, Time: 0.0247 Steps: 6390, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000013, Sample Num: 208, Cur Loss: 0.33157849, Cur Avg Loss: 0.79891394, Log Avg loss: 0.75798834, Global Avg Loss: 5.02976698, Time: 0.0210 Steps: 6400, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000023, Sample Num: 368, Cur Loss: 0.39064330, Cur Avg Loss: 0.70337153, Log Avg loss: 0.57916639, Global Avg Loss: 5.02282377, Time: 0.0210 Steps: 6410, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000033, Sample Num: 528, Cur Loss: 0.37872893, Cur Avg Loss: 0.63994747, Log Avg loss: 0.49407213, Global Avg Loss: 5.01576964, Time: 0.0209 Steps: 6420, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000043, Sample Num: 688, Cur Loss: 0.63298649, Cur Avg Loss: 0.65821216, Log Avg loss: 0.71848563, Global Avg Loss: 5.00908646, Time: 0.0210 Steps: 6430, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000053, Sample Num: 848, Cur Loss: 0.76733434, Cur Avg Loss: 0.64092873, Log Avg loss: 0.56660998, Global Avg Loss: 5.00218820, Time: 0.0210 Steps: 6440, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000063, Sample Num: 1008, Cur Loss: 0.38721240, Cur Avg Loss: 0.61194814, Log Avg loss: 0.45835100, Global Avg Loss: 4.99514350, Time: 0.0209 Steps: 6450, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000073, Sample Num: 1168, Cur Loss: 0.69899583, Cur Avg Loss: 0.63024973, Log Avg loss: 0.74554977, Global Avg Loss: 4.98856518, Time: 0.0211 Steps: 6460, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000083, Sample Num: 1328, Cur Loss: 0.46280628, Cur Avg Loss: 0.60972195, Log Avg loss: 0.45986918, Global Avg Loss: 4.98156565, Time: 0.0208 Steps: 6470, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000093, Sample Num: 1488, Cur Loss: 0.54209960, Cur Avg Loss: 0.61022644, Log Avg loss: 0.61441364, Global Avg Loss: 4.97482621, Time: 0.0209 Steps: 6480, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000103, Sample Num: 1648, Cur Loss: 0.63636523, Cur Avg Loss: 0.60761151, Log Avg loss: 0.58329276, Global Avg Loss: 4.96805960, Time: 0.0208 Steps: 6490, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000113, Sample Num: 1808, Cur Loss: 0.37333584, Cur Avg Loss: 0.60262308, Log Avg loss: 0.55124221, Global Avg Loss: 4.96126450, Time: 0.0209 Steps: 6500, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000123, Sample Num: 1968, Cur Loss: 1.24050355, Cur Avg Loss: 0.63056987, Log Avg loss: 0.94636865, Global Avg Loss: 4.95509722, Time: 0.0210 Steps: 6510, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000133, Sample Num: 2128, Cur Loss: 0.36046493, Cur Avg Loss: 0.64761412, Log Avg loss: 0.85725828, Global Avg Loss: 4.94881219, Time: 0.0209 Steps: 6520, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000143, Sample Num: 2288, Cur Loss: 0.63665515, Cur Avg Loss: 0.66105978, Log Avg loss: 0.83988707, Global Avg Loss: 4.94251981, Time: 0.0208 Steps: 6530, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000153, Sample Num: 2448, Cur Loss: 0.79713798, Cur Avg Loss: 0.65481479, Log Avg loss: 0.56551155, Global Avg Loss: 4.93582714, Time: 0.0208 Steps: 6540, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000163, Sample Num: 2608, Cur Loss: 0.62567872, Cur Avg Loss: 0.65854585, Log Avg loss: 0.71563107, Global Avg Loss: 4.92938409, Time: 0.0208 Steps: 6550, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000173, Sample Num: 2768, Cur Loss: 2.20567751, Cur Avg Loss: 0.66179397, Log Avg loss: 0.71473818, Global Avg Loss: 4.92295932, Time: 0.0209 Steps: 6560, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000183, Sample Num: 2928, Cur Loss: 0.42377466, Cur Avg Loss: 0.67131401, Log Avg loss: 0.83601083, Global Avg Loss: 4.91673870, Time: 0.0208 Steps: 6570, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000193, Sample Num: 3088, Cur Loss: 0.57604015, Cur Avg Loss: 0.65950376, Log Avg loss: 0.44337611, Global Avg Loss: 4.90994028, Time: 0.0209 Steps: 6580, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000203, Sample Num: 3248, Cur Loss: 0.46059206, Cur Avg Loss: 0.66648501, Log Avg loss: 0.80122322, Global Avg Loss: 4.90370550, Time: 0.0209 Steps: 6590, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000213, Sample Num: 3408, Cur Loss: 0.60474682, Cur Avg Loss: 0.65864776, Log Avg loss: 0.49955157, Global Avg Loss: 4.89703254, Time: 0.0208 Steps: 6600, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000223, Sample Num: 3568, Cur Loss: 0.64069337, Cur Avg Loss: 0.65724170, Log Avg loss: 0.62729262, Global Avg Loss: 4.89057303, Time: 0.0209 Steps: 6610, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000233, Sample Num: 3728, Cur Loss: 1.90910149, Cur Avg Loss: 0.65887980, Log Avg loss: 0.69540941, Global Avg Loss: 4.88423592, Time: 0.0208 Steps: 6620, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000243, Sample Num: 3888, Cur Loss: 0.56859052, Cur Avg Loss: 0.66175588, Log Avg loss: 0.72876842, Global Avg Loss: 4.87796825, Time: 0.0208 Steps: 6630, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000253, Sample Num: 4048, Cur Loss: 0.17894387, Cur Avg Loss: 0.67272570, Log Avg loss: 0.93929238, Global Avg Loss: 4.87203651, Time: 0.0208 Steps: 6640, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000263, Sample Num: 4208, Cur Loss: 0.84858799, Cur Avg Loss: 0.66889122, Log Avg loss: 0.57187884, Global Avg Loss: 4.86557011, Time: 0.0210 Steps: 6650, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000273, Sample Num: 4368, Cur Loss: 1.35729742, Cur Avg Loss: 0.68872826, Log Avg loss: 1.21044237, Global Avg Loss: 4.86008193, Time: 0.0208 Steps: 6660, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000283, Sample Num: 4528, Cur Loss: 1.05909836, Cur Avg Loss: 0.70405334, Log Avg loss: 1.12242806, Global Avg Loss: 4.85447825, Time: 0.0210 Steps: 6670, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000293, Sample Num: 4688, Cur Loss: 2.06241155, Cur Avg Loss: 0.73014508, Log Avg loss: 1.46854128, Global Avg Loss: 4.84940948, Time: 0.0210 Steps: 6680, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000303, Sample Num: 4848, Cur Loss: 0.98434222, Cur Avg Loss: 0.73439197, Log Avg loss: 0.85882604, Global Avg Loss: 4.84344448, Time: 0.0209 Steps: 6690, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000313, Sample Num: 5008, Cur Loss: 0.83156329, Cur Avg Loss: 0.73856990, Log Avg loss: 0.86516109, Global Avg Loss: 4.83750675, Time: 0.0210 Steps: 6700, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000323, Sample Num: 5168, Cur Loss: 0.42198747, Cur Avg Loss: 0.73804338, Log Avg loss: 0.72156317, Global Avg Loss: 4.83137270, Time: 0.0209 Steps: 6710, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000333, Sample Num: 5328, Cur Loss: 0.40993637, Cur Avg Loss: 0.73083858, Log Avg loss: 0.49812367, Global Avg Loss: 4.82492441, Time: 0.0209 Steps: 6720, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000343, Sample Num: 5488, Cur Loss: 0.85685301, Cur Avg Loss: 0.72541824, Log Avg loss: 0.54492092, Global Avg Loss: 4.81856482, Time: 0.0209 Steps: 6730, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000353, Sample Num: 5648, Cur Loss: 0.42577440, Cur Avg Loss: 0.72263580, Log Avg loss: 0.62719816, Global Avg Loss: 4.81234618, Time: 0.0209 Steps: 6740, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000363, Sample Num: 5808, Cur Loss: 0.50685203, Cur Avg Loss: 0.71656990, Log Avg loss: 0.50244372, Global Avg Loss: 4.80596114, Time: 0.0208 Steps: 6750, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000373, Sample Num: 5968, Cur Loss: 1.79998994, Cur Avg Loss: 0.72444209, Log Avg loss: 1.01020262, Global Avg Loss: 4.80034611, Time: 0.0209 Steps: 6760, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000383, Sample Num: 6128, Cur Loss: 0.35751131, Cur Avg Loss: 0.73447690, Log Avg loss: 1.10877502, Global Avg Loss: 4.79489327, Time: 0.0209 Steps: 6770, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000393, Sample Num: 6288, Cur Loss: 0.98572242, Cur Avg Loss: 0.72907339, Log Avg loss: 0.52211891, Global Avg Loss: 4.78859125, Time: 0.0209 Steps: 6780, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000403, Sample Num: 6448, Cur Loss: 0.55189860, Cur Avg Loss: 0.72291005, Log Avg loss: 0.48069116, Global Avg Loss: 4.78224677, Time: 0.0209 Steps: 6790, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000413, Sample Num: 6608, Cur Loss: 0.99034691, Cur Avg Loss: 0.72075878, Log Avg loss: 0.63406229, Global Avg Loss: 4.77614650, Time: 0.0209 Steps: 6800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000423, Sample Num: 6768, Cur Loss: 1.05373049, Cur Avg Loss: 0.72615036, Log Avg loss: 0.94882267, Global Avg Loss: 4.77052635, Time: 0.0209 Steps: 6810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000433, Sample Num: 6928, Cur Loss: 0.79970533, Cur Avg Loss: 0.72262087, Log Avg loss: 0.57332334, Global Avg Loss: 4.76437209, Time: 0.0209 Steps: 6820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000443, Sample Num: 7088, Cur Loss: 0.34789568, Cur Avg Loss: 0.71722998, Log Avg loss: 0.48380448, Global Avg Loss: 4.75810479, Time: 0.0209 Steps: 6830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000453, Sample Num: 7248, Cur Loss: 0.72164780, Cur Avg Loss: 0.71441180, Log Avg loss: 0.58956637, Global Avg Loss: 4.75201043, Time: 0.0208 Steps: 6840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000463, Sample Num: 7408, Cur Loss: 0.70613343, Cur Avg Loss: 0.71090425, Log Avg loss: 0.55201223, Global Avg Loss: 4.74587905, Time: 0.0209 Steps: 6850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000473, Sample Num: 7568, Cur Loss: 0.26997280, Cur Avg Loss: 0.71100630, Log Avg loss: 0.71573128, Global Avg Loss: 4.74000420, Time: 0.0209 Steps: 6860, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000483, Sample Num: 7728, Cur Loss: 0.34260809, Cur Avg Loss: 0.70673630, Log Avg loss: 0.50476561, Global Avg Loss: 4.73383937, Time: 0.0209 Steps: 6870, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000493, Sample Num: 7888, Cur Loss: 1.14969885, Cur Avg Loss: 0.70784866, Log Avg loss: 0.76157539, Global Avg Loss: 4.72806573, Time: 0.0209 Steps: 6880, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000503, Sample Num: 8048, Cur Loss: 1.85852790, Cur Avg Loss: 0.72046885, Log Avg loss: 1.34264430, Global Avg Loss: 4.72315220, Time: 0.0209 Steps: 6890, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000513, Sample Num: 8208, Cur Loss: 0.75344241, Cur Avg Loss: 0.72015828, Log Avg loss: 0.70453666, Global Avg Loss: 4.71732812, Time: 0.0245 Steps: 6900, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000523, Sample Num: 8368, Cur Loss: 0.38735235, Cur Avg Loss: 0.71615429, Log Avg loss: 0.51074978, Global Avg Loss: 4.71124045, Time: 0.0209 Steps: 6910, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000533, Sample Num: 8528, Cur Loss: 0.95199776, Cur Avg Loss: 0.72078827, Log Avg loss: 0.96314515, Global Avg Loss: 4.70582413, Time: 0.0209 Steps: 6920, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000543, Sample Num: 8688, Cur Loss: 0.23720153, Cur Avg Loss: 0.71865017, Log Avg loss: 0.60468959, Global Avg Loss: 4.69990618, Time: 0.0209 Steps: 6930, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000553, Sample Num: 8848, Cur Loss: 1.18827116, Cur Avg Loss: 0.72170879, Log Avg loss: 0.88779168, Global Avg Loss: 4.69441322, Time: 0.0209 Steps: 6940, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000563, Sample Num: 9008, Cur Loss: 0.82681549, Cur Avg Loss: 0.71941785, Log Avg loss: 0.59272914, Global Avg Loss: 4.68851152, Time: 0.0209 Steps: 6950, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000573, Sample Num: 9168, Cur Loss: 0.15905395, Cur Avg Loss: 0.72113148, Log Avg loss: 0.81760871, Global Avg Loss: 4.68294988, Time: 0.0209 Steps: 6960, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000583, Sample Num: 9328, Cur Loss: 1.20306563, Cur Avg Loss: 0.72793446, Log Avg loss: 1.11774496, Global Avg Loss: 4.67783481, Time: 0.0209 Steps: 6970, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000593, Sample Num: 9488, Cur Loss: 0.48281217, Cur Avg Loss: 0.72722943, Log Avg loss: 0.68612613, Global Avg Loss: 4.67211603, Time: 0.0209 Steps: 6980, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000603, Sample Num: 9648, Cur Loss: 0.32951117, Cur Avg Loss: 0.72995455, Log Avg loss: 0.89155466, Global Avg Loss: 4.66670750, Time: 0.0209 Steps: 6990, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000613, Sample Num: 9808, Cur Loss: 0.52467149, Cur Avg Loss: 0.73081731, Log Avg loss: 0.78284133, Global Avg Loss: 4.66115912, Time: 0.0209 Steps: 7000, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000623, Sample Num: 9968, Cur Loss: 0.19012173, Cur Avg Loss: 0.72734886, Log Avg loss: 0.51473280, Global Avg Loss: 4.65524410, Time: 0.0209 Steps: 7010, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000633, Sample Num: 10128, Cur Loss: 0.31659424, Cur Avg Loss: 0.72648608, Log Avg loss: 0.67273494, Global Avg Loss: 4.64957101, Time: 0.0210 Steps: 7020, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000643, Sample Num: 10288, Cur Loss: 1.37731731, Cur Avg Loss: 0.73021341, Log Avg loss: 0.96615363, Global Avg Loss: 4.64433144, Time: 0.0208 Steps: 7030, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000653, Sample Num: 10448, Cur Loss: 0.32870159, Cur Avg Loss: 0.72991457, Log Avg loss: 0.71069919, Global Avg Loss: 4.63874390, Time: 0.0209 Steps: 7040, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000663, Sample Num: 10608, Cur Loss: 1.54957950, Cur Avg Loss: 0.72857802, Log Avg loss: 0.64130106, Global Avg Loss: 4.63307376, Time: 0.0209 Steps: 7050, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000673, Sample Num: 10768, Cur Loss: 0.33145919, Cur Avg Loss: 0.72178711, Log Avg loss: 0.27154967, Global Avg Loss: 4.62689597, Time: 0.0208 Steps: 7060, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000683, Sample Num: 10928, Cur Loss: 0.73245078, Cur Avg Loss: 0.72058515, Log Avg loss: 0.63969362, Global Avg Loss: 4.62125636, Time: 0.0209 Steps: 7070, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000693, Sample Num: 11088, Cur Loss: 0.31731451, Cur Avg Loss: 0.71641094, Log Avg loss: 0.43131203, Global Avg Loss: 4.61533836, Time: 0.0208 Steps: 7080, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000703, Sample Num: 11248, Cur Loss: 0.67950433, Cur Avg Loss: 0.71371922, Log Avg loss: 0.52718319, Global Avg Loss: 4.60957227, Time: 0.0209 Steps: 7090, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000713, Sample Num: 11408, Cur Loss: 0.63944441, Cur Avg Loss: 0.71184297, Log Avg loss: 0.57994236, Global Avg Loss: 4.60389674, Time: 0.0208 Steps: 7100, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000723, Sample Num: 11568, Cur Loss: 0.17912677, Cur Avg Loss: 0.71246699, Log Avg loss: 0.75695987, Global Avg Loss: 4.59848614, Time: 0.0209 Steps: 7110, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000733, Sample Num: 11728, Cur Loss: 0.52284878, Cur Avg Loss: 0.71111674, Log Avg loss: 0.61349343, Global Avg Loss: 4.59288924, Time: 0.0208 Steps: 7120, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000743, Sample Num: 11888, Cur Loss: 0.69013166, Cur Avg Loss: 0.70864795, Log Avg loss: 0.52768573, Global Avg Loss: 4.58718769, Time: 0.0208 Steps: 7130, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000753, Sample Num: 12048, Cur Loss: 0.76382113, Cur Avg Loss: 0.70816849, Log Avg loss: 0.67254485, Global Avg Loss: 4.58170500, Time: 0.0209 Steps: 7140, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000763, Sample Num: 12208, Cur Loss: 0.46877050, Cur Avg Loss: 0.70640115, Log Avg loss: 0.57332068, Global Avg Loss: 4.57609887, Time: 0.0208 Steps: 7150, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000773, Sample Num: 12368, Cur Loss: 0.44418281, Cur Avg Loss: 0.70630872, Log Avg loss: 0.69925592, Global Avg Loss: 4.57068428, Time: 0.0212 Steps: 7160, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000783, Sample Num: 12528, Cur Loss: 0.40991732, Cur Avg Loss: 0.71798903, Log Avg loss: 1.62087734, Global Avg Loss: 4.56657018, Time: 0.0208 Steps: 7170, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000793, Sample Num: 12688, Cur Loss: 0.64692706, Cur Avg Loss: 0.72322511, Log Avg loss: 1.13321006, Global Avg Loss: 4.56178835, Time: 0.0209 Steps: 7180, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000803, Sample Num: 12848, Cur Loss: 0.31927416, Cur Avg Loss: 0.72747996, Log Avg loss: 1.06488977, Global Avg Loss: 4.55692479, Time: 0.0209 Steps: 7190, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000813, Sample Num: 13008, Cur Loss: 0.85159183, Cur Avg Loss: 0.72661590, Log Avg loss: 0.65723178, Global Avg Loss: 4.55150855, Time: 0.0208 Steps: 7200, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000823, Sample Num: 13168, Cur Loss: 0.20280658, Cur Avg Loss: 0.73420652, Log Avg loss: 1.35132352, Global Avg Loss: 4.54707001, Time: 0.0208 Steps: 7210, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000833, Sample Num: 13328, Cur Loss: 1.05355000, Cur Avg Loss: 0.73201482, Log Avg loss: 0.55163782, Global Avg Loss: 4.54153617, Time: 0.0208 Steps: 7220, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000843, Sample Num: 13488, Cur Loss: 0.25329542, Cur Avg Loss: 0.72962241, Log Avg loss: 0.53033533, Global Avg Loss: 4.53598818, Time: 0.0209 Steps: 7230, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000853, Sample Num: 13648, Cur Loss: 0.43363303, Cur Avg Loss: 0.72913082, Log Avg loss: 0.68768952, Global Avg Loss: 4.53067285, Time: 0.0210 Steps: 7240, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000863, Sample Num: 13808, Cur Loss: 0.66536486, Cur Avg Loss: 0.72669987, Log Avg loss: 0.51933959, Global Avg Loss: 4.52513997, Time: 0.0208 Steps: 7250, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000873, Sample Num: 13968, Cur Loss: 0.26449770, Cur Avg Loss: 0.72586599, Log Avg loss: 0.65390199, Global Avg Loss: 4.51980769, Time: 0.0208 Steps: 7260, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000883, Sample Num: 14128, Cur Loss: 0.40767598, Cur Avg Loss: 0.72407739, Log Avg loss: 0.56793261, Global Avg Loss: 4.51437182, Time: 0.0209 Steps: 7270, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000893, Sample Num: 14288, Cur Loss: 1.16311932, Cur Avg Loss: 0.72711983, Log Avg loss: 0.99576763, Global Avg Loss: 4.50953857, Time: 0.0209 Steps: 7280, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000903, Sample Num: 14448, Cur Loss: 0.65341461, Cur Avg Loss: 0.72946952, Log Avg loss: 0.93929717, Global Avg Loss: 4.50464112, Time: 0.0209 Steps: 7290, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000913, Sample Num: 14608, Cur Loss: 0.54751247, Cur Avg Loss: 0.73348426, Log Avg loss: 1.09601454, Global Avg Loss: 4.49997177, Time: 0.0209 Steps: 7300, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000923, Sample Num: 14768, Cur Loss: 1.44853091, Cur Avg Loss: 0.74002858, Log Avg loss: 1.33752522, Global Avg Loss: 4.49564558, Time: 0.0209 Steps: 7310, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000933, Sample Num: 14928, Cur Loss: 1.06997752, Cur Avg Loss: 0.74312577, Log Avg loss: 1.02899627, Global Avg Loss: 4.49090972, Time: 0.0209 Steps: 7320, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000943, Sample Num: 15088, Cur Loss: 0.56107879, Cur Avg Loss: 0.74109261, Log Avg loss: 0.55139873, Global Avg Loss: 4.48553522, Time: 0.0209 Steps: 7330, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000953, Sample Num: 15248, Cur Loss: 0.63200939, Cur Avg Loss: 0.74080156, Log Avg loss: 0.71335604, Global Avg Loss: 4.48039601, Time: 0.0208 Steps: 7340, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000963, Sample Num: 15408, Cur Loss: 0.41865084, Cur Avg Loss: 0.74389453, Log Avg loss: 1.03865429, Global Avg Loss: 4.47571337, Time: 0.0208 Steps: 7350, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000973, Sample Num: 15568, Cur Loss: 0.61736965, Cur Avg Loss: 0.74133049, Log Avg loss: 0.49441356, Global Avg Loss: 4.47030399, Time: 0.0209 Steps: 7360, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000983, Sample Num: 15728, Cur Loss: 0.67206270, Cur Avg Loss: 0.74018881, Log Avg loss: 0.62910304, Global Avg Loss: 4.46509205, Time: 0.0209 Steps: 7370, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000993, Sample Num: 15888, Cur Loss: 0.18279399, Cur Avg Loss: 0.73877252, Log Avg loss: 0.59955141, Global Avg Loss: 4.45985419, Time: 0.0210 Steps: 7380, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001003, Sample Num: 16048, Cur Loss: 0.64281738, Cur Avg Loss: 0.73774793, Log Avg loss: 0.63600608, Global Avg Loss: 4.45467984, Time: 0.0209 Steps: 7390, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001013, Sample Num: 16208, Cur Loss: 0.20205519, Cur Avg Loss: 0.73617932, Log Avg loss: 0.57884785, Global Avg Loss: 4.44944222, Time: 0.0210 Steps: 7400, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001023, Sample Num: 16368, Cur Loss: 0.48183441, Cur Avg Loss: 0.73510716, Log Avg loss: 0.62649766, Global Avg Loss: 4.44428306, Time: 0.0209 Steps: 7410, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001033, Sample Num: 16528, Cur Loss: 1.36418998, Cur Avg Loss: 0.73594526, Log Avg loss: 0.82168256, Global Avg Loss: 4.43940084, Time: 0.0210 Steps: 7420, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001043, Sample Num: 16688, Cur Loss: 1.96630096, Cur Avg Loss: 0.73929692, Log Avg loss: 1.08552327, Global Avg Loss: 4.43488688, Time: 0.0209 Steps: 7430, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001053, Sample Num: 16848, Cur Loss: 0.69922268, Cur Avg Loss: 0.74028655, Log Avg loss: 0.84350551, Global Avg Loss: 4.43005975, Time: 0.0209 Steps: 7440, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001063, Sample Num: 17008, Cur Loss: 0.23940659, Cur Avg Loss: 0.73690606, Log Avg loss: 0.38093967, Global Avg Loss: 4.42462469, Time: 0.0209 Steps: 7450, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001073, Sample Num: 17168, Cur Loss: 1.18398201, Cur Avg Loss: 0.73516896, Log Avg loss: 0.55051590, Global Avg Loss: 4.41943152, Time: 0.0209 Steps: 7460, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001083, Sample Num: 17328, Cur Loss: 0.63339400, Cur Avg Loss: 0.73391560, Log Avg loss: 0.59943003, Global Avg Loss: 4.41431773, Time: 0.0209 Steps: 7470, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001093, Sample Num: 17488, Cur Loss: 0.54270965, Cur Avg Loss: 0.73278572, Log Avg loss: 0.61041931, Global Avg Loss: 4.40923230, Time: 0.0209 Steps: 7480, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001103, Sample Num: 17648, Cur Loss: 0.71105468, Cur Avg Loss: 0.73015818, Log Avg loss: 0.44296815, Global Avg Loss: 4.40393689, Time: 0.0211 Steps: 7490, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001113, Sample Num: 17808, Cur Loss: 0.73980880, Cur Avg Loss: 0.73002961, Log Avg loss: 0.71584884, Global Avg Loss: 4.39901944, Time: 0.0209 Steps: 7500, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001123, Sample Num: 17968, Cur Loss: 0.41751564, Cur Avg Loss: 0.72920955, Log Avg loss: 0.63793706, Global Avg Loss: 4.39401134, Time: 0.0209 Steps: 7510, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001133, Sample Num: 18128, Cur Loss: 0.38803828, Cur Avg Loss: 0.72928016, Log Avg loss: 0.73720869, Global Avg Loss: 4.38914857, Time: 0.0209 Steps: 7520, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001143, Sample Num: 18288, Cur Loss: 0.41623041, Cur Avg Loss: 0.72720790, Log Avg loss: 0.49242095, Global Avg Loss: 4.38397363, Time: 0.0209 Steps: 7530, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001153, Sample Num: 18448, Cur Loss: 0.44998494, Cur Avg Loss: 0.72665485, Log Avg loss: 0.66344211, Global Avg Loss: 4.37903924, Time: 0.0209 Steps: 7540, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001163, Sample Num: 18608, Cur Loss: 1.27038026, Cur Avg Loss: 0.72735900, Log Avg loss: 0.80854710, Global Avg Loss: 4.37431011, Time: 0.0209 Steps: 7550, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001173, Sample Num: 18768, Cur Loss: 1.56231368, Cur Avg Loss: 0.73228646, Log Avg loss: 1.30535026, Global Avg Loss: 4.37025064, Time: 0.0209 Steps: 7560, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001183, Sample Num: 18928, Cur Loss: 0.59857452, Cur Avg Loss: 0.73380066, Log Avg loss: 0.91141635, Global Avg Loss: 4.36568151, Time: 0.0209 Steps: 7570, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001193, Sample Num: 19088, Cur Loss: 0.70057034, Cur Avg Loss: 0.73514503, Log Avg loss: 0.89418404, Global Avg Loss: 4.36110169, Time: 0.0209 Steps: 7580, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001203, Sample Num: 19248, Cur Loss: 0.59507984, Cur Avg Loss: 0.73516815, Log Avg loss: 0.73792597, Global Avg Loss: 4.35632808, Time: 0.0209 Steps: 7590, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001213, Sample Num: 19408, Cur Loss: 0.47378385, Cur Avg Loss: 0.73419055, Log Avg loss: 0.61658502, Global Avg Loss: 4.35140736, Time: 0.0208 Steps: 7600, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001223, Sample Num: 19568, Cur Loss: 0.62673783, Cur Avg Loss: 0.73413001, Log Avg loss: 0.72678703, Global Avg Loss: 4.34664439, Time: 0.0209 Steps: 7610, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001233, Sample Num: 19728, Cur Loss: 1.11166251, Cur Avg Loss: 0.73658881, Log Avg loss: 1.03729998, Global Avg Loss: 4.34230142, Time: 0.0209 Steps: 7620, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001243, Sample Num: 19888, Cur Loss: 1.24285388, Cur Avg Loss: 0.73902731, Log Avg loss: 1.03969397, Global Avg Loss: 4.33797297, Time: 0.0209 Steps: 7630, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001253, Sample Num: 20048, Cur Loss: 0.99020505, Cur Avg Loss: 0.73978215, Log Avg loss: 0.83360833, Global Avg Loss: 4.33338610, Time: 0.0209 Steps: 7640, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001263, Sample Num: 20208, Cur Loss: 0.49374723, Cur Avg Loss: 0.74084402, Log Avg loss: 0.87389745, Global Avg Loss: 4.32886390, Time: 0.0209 Steps: 7650, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001273, Sample Num: 20368, Cur Loss: 1.67068028, Cur Avg Loss: 0.74433770, Log Avg loss: 1.18558917, Global Avg Loss: 4.32476041, Time: 0.0209 Steps: 7660, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001283, Sample Num: 20528, Cur Loss: 0.52358055, Cur Avg Loss: 0.75264880, Log Avg loss: 1.81065201, Global Avg Loss: 4.32148256, Time: 0.0246 Steps: 7670, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001293, Sample Num: 20688, Cur Loss: 1.26762044, Cur Avg Loss: 0.75439064, Log Avg loss: 0.97786832, Global Avg Loss: 4.31712889, Time: 0.0209 Steps: 7680, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001303, Sample Num: 20848, Cur Loss: 0.30270097, Cur Avg Loss: 0.75442087, Log Avg loss: 0.75832906, Global Avg Loss: 4.31250107, Time: 0.0209 Steps: 7690, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001313, Sample Num: 21008, Cur Loss: 0.64162767, Cur Avg Loss: 0.75378651, Log Avg loss: 0.67113000, Global Avg Loss: 4.30777201, Time: 0.0209 Steps: 7700, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001323, Sample Num: 21168, Cur Loss: 0.84460902, Cur Avg Loss: 0.75269044, Log Avg loss: 0.60877624, Global Avg Loss: 4.30297435, Time: 0.0209 Steps: 7710, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001333, Sample Num: 21328, Cur Loss: 1.78096235, Cur Avg Loss: 0.75423855, Log Avg loss: 0.95905418, Global Avg Loss: 4.29864285, Time: 0.0209 Steps: 7720, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001343, Sample Num: 21488, Cur Loss: 0.23043926, Cur Avg Loss: 0.75448150, Log Avg loss: 0.78686595, Global Avg Loss: 4.29409980, Time: 0.0209 Steps: 7730, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001353, Sample Num: 21648, Cur Loss: 0.63376182, Cur Avg Loss: 0.75468516, Log Avg loss: 0.78203718, Global Avg Loss: 4.28956225, Time: 0.0209 Steps: 7740, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001363, Sample Num: 21808, Cur Loss: 0.81915212, Cur Avg Loss: 0.75604933, Log Avg loss: 0.94062167, Global Avg Loss: 4.28524104, Time: 0.0209 Steps: 7750, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001373, Sample Num: 21968, Cur Loss: 1.28753638, Cur Avg Loss: 0.76316905, Log Avg loss: 1.73358645, Global Avg Loss: 4.28195282, Time: 0.0209 Steps: 7760, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001383, Sample Num: 22128, Cur Loss: 0.37273484, Cur Avg Loss: 0.76685240, Log Avg loss: 1.27257662, Global Avg Loss: 4.27807975, Time: 0.0210 Steps: 7770, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001393, Sample Num: 22288, Cur Loss: 0.58478254, Cur Avg Loss: 0.76874619, Log Avg loss: 1.03065678, Global Avg Loss: 4.27390569, Time: 0.0209 Steps: 7780, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001403, Sample Num: 22448, Cur Loss: 1.09517264, Cur Avg Loss: 0.76856455, Log Avg loss: 0.74326274, Global Avg Loss: 4.26937341, Time: 0.0209 Steps: 7790, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001413, Sample Num: 22608, Cur Loss: 0.28661576, Cur Avg Loss: 0.76725772, Log Avg loss: 0.58390925, Global Avg Loss: 4.26464846, Time: 0.0209 Steps: 7800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001423, Sample Num: 22768, Cur Loss: 0.86480391, Cur Avg Loss: 0.76601860, Log Avg loss: 0.59093159, Global Avg Loss: 4.25994460, Time: 0.0209 Steps: 7810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001433, Sample Num: 22928, Cur Loss: 0.37869340, Cur Avg Loss: 0.76335217, Log Avg loss: 0.38391815, Global Avg Loss: 4.25498804, Time: 0.0209 Steps: 7820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001443, Sample Num: 23088, Cur Loss: 0.32524046, Cur Avg Loss: 0.76133182, Log Avg loss: 0.47181545, Global Avg Loss: 4.25015640, Time: 0.0208 Steps: 7830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001453, Sample Num: 23248, Cur Loss: 0.63281286, Cur Avg Loss: 0.76023759, Log Avg loss: 0.60234082, Global Avg Loss: 4.24550358, Time: 0.0209 Steps: 7840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001463, Sample Num: 23408, Cur Loss: 0.36397979, Cur Avg Loss: 0.75896332, Log Avg loss: 0.57381214, Global Avg Loss: 4.24082626, Time: 0.0209 Steps: 7850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001473, Sample Num: 23568, Cur Loss: 0.31965759, Cur Avg Loss: 0.75708503, Log Avg loss: 0.48229100, Global Avg Loss: 4.23604441, Time: 0.0209 Steps: 7860, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001483, Sample Num: 23728, Cur Loss: 1.13311386, Cur Avg Loss: 0.75566888, Log Avg loss: 0.54707032, Global Avg Loss: 4.23135702, Time: 0.0209 Steps: 7870, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001493, Sample Num: 23888, Cur Loss: 0.58617812, Cur Avg Loss: 0.75533144, Log Avg loss: 0.70528853, Global Avg Loss: 4.22688232, Time: 0.0209 Steps: 7880, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001503, Sample Num: 24048, Cur Loss: 0.70719302, Cur Avg Loss: 0.75493070, Log Avg loss: 0.69510014, Global Avg Loss: 4.22240604, Time: 0.0209 Steps: 7890, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001513, Sample Num: 24208, Cur Loss: 0.45162255, Cur Avg Loss: 0.75335541, Log Avg loss: 0.51659025, Global Avg Loss: 4.21771513, Time: 0.0209 Steps: 7900, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001523, Sample Num: 24368, Cur Loss: 0.38838118, Cur Avg Loss: 0.75246798, Log Avg loss: 0.61819834, Global Avg Loss: 4.21316454, Time: 0.0209 Steps: 7910, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001533, Sample Num: 24528, Cur Loss: 0.42814058, Cur Avg Loss: 0.75146865, Log Avg loss: 0.59927143, Global Avg Loss: 4.20860155, Time: 0.0209 Steps: 7920, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001543, Sample Num: 24688, Cur Loss: 0.35854119, Cur Avg Loss: 0.75004532, Log Avg loss: 0.53184811, Global Avg Loss: 4.20396504, Time: 0.0210 Steps: 7930, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001553, Sample Num: 24848, Cur Loss: 0.22452682, Cur Avg Loss: 0.74847120, Log Avg loss: 0.50558493, Global Avg Loss: 4.19930713, Time: 0.0209 Steps: 7940, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001563, Sample Num: 25008, Cur Loss: 0.52819914, Cur Avg Loss: 0.74674498, Log Avg loss: 0.47866349, Global Avg Loss: 4.19462707, Time: 0.0209 Steps: 7950, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001573, Sample Num: 25168, Cur Loss: 0.48216152, Cur Avg Loss: 0.74558276, Log Avg loss: 0.56392816, Global Avg Loss: 4.19006589, Time: 0.0209 Steps: 7960, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001583, Sample Num: 25328, Cur Loss: 0.78723115, Cur Avg Loss: 0.74434009, Log Avg loss: 0.54886697, Global Avg Loss: 4.18549726, Time: 0.0209 Steps: 7970, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001593, Sample Num: 25488, Cur Loss: 0.61961877, Cur Avg Loss: 0.74229943, Log Avg loss: 0.41926386, Global Avg Loss: 4.18077767, Time: 0.0209 Steps: 7980, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001603, Sample Num: 25648, Cur Loss: 0.63877153, Cur Avg Loss: 0.74143446, Log Avg loss: 0.60364393, Global Avg Loss: 4.17630066, Time: 0.0209 Steps: 7990, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001613, Sample Num: 25808, Cur Loss: 0.73770154, Cur Avg Loss: 0.74067535, Log Avg loss: 0.61899109, Global Avg Loss: 4.17185402, Time: 0.0209 Steps: 8000, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001623, Sample Num: 25968, Cur Loss: 0.62385690, Cur Avg Loss: 0.73900227, Log Avg loss: 0.46913341, Global Avg Loss: 4.16723140, Time: 0.0209 Steps: 8010, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001633, Sample Num: 26128, Cur Loss: 0.15568233, Cur Avg Loss: 0.73938121, Log Avg loss: 0.80088382, Global Avg Loss: 4.16303396, Time: 0.0209 Steps: 8020, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001643, Sample Num: 26288, Cur Loss: 1.13771653, Cur Avg Loss: 0.73855778, Log Avg loss: 0.60409151, Global Avg Loss: 4.15860190, Time: 0.0208 Steps: 8030, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001653, Sample Num: 26448, Cur Loss: 0.44599265, Cur Avg Loss: 0.73633850, Log Avg loss: 0.37171119, Global Avg Loss: 4.15389184, Time: 0.0209 Steps: 8040, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001663, Sample Num: 26608, Cur Loss: 0.61060983, Cur Avg Loss: 0.73495017, Log Avg loss: 0.50545918, Global Avg Loss: 4.14935962, Time: 0.0209 Steps: 8050, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001673, Sample Num: 26768, Cur Loss: 0.40473863, Cur Avg Loss: 0.73583964, Log Avg loss: 0.88375737, Global Avg Loss: 4.14530801, Time: 0.0209 Steps: 8060, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001683, Sample Num: 26928, Cur Loss: 0.31761685, Cur Avg Loss: 0.73513884, Log Avg loss: 0.61789527, Global Avg Loss: 4.14093699, Time: 0.0209 Steps: 8070, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001693, Sample Num: 27088, Cur Loss: 0.39918858, Cur Avg Loss: 0.73406621, Log Avg loss: 0.55354264, Global Avg Loss: 4.13649714, Time: 0.0209 Steps: 8080, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001703, Sample Num: 27248, Cur Loss: 0.67153192, Cur Avg Loss: 0.73390035, Log Avg loss: 0.70581979, Global Avg Loss: 4.13225650, Time: 0.0209 Steps: 8090, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001713, Sample Num: 27408, Cur Loss: 0.96816581, Cur Avg Loss: 0.73429883, Log Avg loss: 0.80216020, Global Avg Loss: 4.12814527, Time: 0.0209 Steps: 8100, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001723, Sample Num: 27568, Cur Loss: 0.68187475, Cur Avg Loss: 0.73519277, Log Avg loss: 0.88832453, Global Avg Loss: 4.12415043, Time: 0.0209 Steps: 8110, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001733, Sample Num: 27728, Cur Loss: 0.42002445, Cur Avg Loss: 0.73409117, Log Avg loss: 0.54428573, Global Avg Loss: 4.11974172, Time: 0.0209 Steps: 8120, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001743, Sample Num: 27888, Cur Loss: 0.50521827, Cur Avg Loss: 0.73332865, Log Avg loss: 0.60118457, Global Avg Loss: 4.11541386, Time: 0.0209 Steps: 8130, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001753, Sample Num: 28048, Cur Loss: 0.77249402, Cur Avg Loss: 0.73385407, Log Avg loss: 0.82543462, Global Avg Loss: 4.11137211, Time: 0.0209 Steps: 8140, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001763, Sample Num: 28208, Cur Loss: 0.41977093, Cur Avg Loss: 0.73377403, Log Avg loss: 0.71974329, Global Avg Loss: 4.10721060, Time: 0.0209 Steps: 8150, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001773, Sample Num: 28368, Cur Loss: 0.29602480, Cur Avg Loss: 0.73321625, Log Avg loss: 0.63487944, Global Avg Loss: 4.10295530, Time: 0.0209 Steps: 8160, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001783, Sample Num: 28528, Cur Loss: 0.18518922, Cur Avg Loss: 0.73110203, Log Avg loss: 0.35625048, Global Avg Loss: 4.09836937, Time: 0.0209 Steps: 8170, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001793, Sample Num: 28688, Cur Loss: 0.43130425, Cur Avg Loss: 0.73015310, Log Avg loss: 0.56095967, Global Avg Loss: 4.09404491, Time: 0.0245 Steps: 8180, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001803, Sample Num: 28848, Cur Loss: 0.44080275, Cur Avg Loss: 0.72928318, Log Avg loss: 0.57330543, Global Avg Loss: 4.08974608, Time: 0.0209 Steps: 8190, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001813, Sample Num: 29008, Cur Loss: 0.34132922, Cur Avg Loss: 0.72856368, Log Avg loss: 0.59883921, Global Avg Loss: 4.08548887, Time: 0.0209 Steps: 8200, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001823, Sample Num: 29168, Cur Loss: 0.85206956, Cur Avg Loss: 0.72817554, Log Avg loss: 0.65780574, Global Avg Loss: 4.08131386, Time: 0.0208 Steps: 8210, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001833, Sample Num: 29328, Cur Loss: 0.63084137, Cur Avg Loss: 0.72775176, Log Avg loss: 0.65049596, Global Avg Loss: 4.07714012, Time: 0.0209 Steps: 8220, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001843, Sample Num: 29488, Cur Loss: 0.59741580, Cur Avg Loss: 0.72654065, Log Avg loss: 0.50454447, Global Avg Loss: 4.07279918, Time: 0.0209 Steps: 8230, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001853, Sample Num: 29648, Cur Loss: 0.88627744, Cur Avg Loss: 0.72763065, Log Avg loss: 0.92851708, Global Avg Loss: 4.06898330, Time: 0.0210 Steps: 8240, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001863, Sample Num: 29808, Cur Loss: 0.54293489, Cur Avg Loss: 0.72802253, Log Avg loss: 0.80063909, Global Avg Loss: 4.06502167, Time: 0.0209 Steps: 8250, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001873, Sample Num: 29968, Cur Loss: 0.51084673, Cur Avg Loss: 0.72836933, Log Avg loss: 0.79297720, Global Avg Loss: 4.06106036, Time: 0.0209 Steps: 8260, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001883, Sample Num: 30128, Cur Loss: 0.38959783, Cur Avg Loss: 0.72699834, Log Avg loss: 0.47021188, Global Avg Loss: 4.05671834, Time: 0.0209 Steps: 8270, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001893, Sample Num: 30288, Cur Loss: 0.40701470, Cur Avg Loss: 0.72690916, Log Avg loss: 0.71011650, Global Avg Loss: 4.05267655, Time: 0.0209 Steps: 8280, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001903, Sample Num: 30448, Cur Loss: 0.80264962, Cur Avg Loss: 0.72841892, Log Avg loss: 1.01421708, Global Avg Loss: 4.04901134, Time: 0.0209 Steps: 8290, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001913, Sample Num: 30608, Cur Loss: 0.52552700, Cur Avg Loss: 0.72834503, Log Avg loss: 0.71428348, Global Avg Loss: 4.04499360, Time: 0.0209 Steps: 8300, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001923, Sample Num: 30768, Cur Loss: 0.22057363, Cur Avg Loss: 0.72724807, Log Avg loss: 0.51739999, Global Avg Loss: 4.04074860, Time: 0.0209 Steps: 8310, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001933, Sample Num: 30928, Cur Loss: 0.69067645, Cur Avg Loss: 0.72573492, Log Avg loss: 0.43475474, Global Avg Loss: 4.03641447, Time: 0.0209 Steps: 8320, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001943, Sample Num: 31088, Cur Loss: 0.45885611, Cur Avg Loss: 0.72559320, Log Avg loss: 0.69819981, Global Avg Loss: 4.03240701, Time: 0.0209 Steps: 8330, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001953, Sample Num: 31248, Cur Loss: 0.47475898, Cur Avg Loss: 0.72462278, Log Avg loss: 0.53606978, Global Avg Loss: 4.02821476, Time: 0.0209 Steps: 8340, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001963, Sample Num: 31408, Cur Loss: 0.32712233, Cur Avg Loss: 0.72327837, Log Avg loss: 0.46071454, Global Avg Loss: 4.02394230, Time: 0.0209 Steps: 8350, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001973, Sample Num: 31568, Cur Loss: 0.33807898, Cur Avg Loss: 0.72178829, Log Avg loss: 0.42928636, Global Avg Loss: 4.01964248, Time: 0.0209 Steps: 8360, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001983, Sample Num: 31728, Cur Loss: 1.11665463, Cur Avg Loss: 0.72099578, Log Avg loss: 0.56463284, Global Avg Loss: 4.01551463, Time: 0.0209 Steps: 8370, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001993, Sample Num: 31888, Cur Loss: 1.29551864, Cur Avg Loss: 0.72354796, Log Avg loss: 1.22964669, Global Avg Loss: 4.01219020, Time: 0.0209 Steps: 8380, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002003, Sample Num: 32048, Cur Loss: 0.55736208, Cur Avg Loss: 0.72387807, Log Avg loss: 0.78966891, Global Avg Loss: 4.00834930, Time: 0.0209 Steps: 8390, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002013, Sample Num: 32208, Cur Loss: 0.30506828, Cur Avg Loss: 0.72243558, Log Avg loss: 0.43350367, Global Avg Loss: 4.00409353, Time: 0.0209 Steps: 8400, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002023, Sample Num: 32368, Cur Loss: 0.57345861, Cur Avg Loss: 0.72181298, Log Avg loss: 0.59648428, Global Avg Loss: 4.00004167, Time: 0.0209 Steps: 8410, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002033, Sample Num: 32528, Cur Loss: 0.09984376, Cur Avg Loss: 0.72164441, Log Avg loss: 0.68754202, Global Avg Loss: 3.99610759, Time: 0.0209 Steps: 8420, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002043, Sample Num: 32688, Cur Loss: 0.89714283, Cur Avg Loss: 0.72236755, Log Avg loss: 0.86938334, Global Avg Loss: 3.99239854, Time: 0.0209 Steps: 8430, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002053, Sample Num: 32848, Cur Loss: 0.33244643, Cur Avg Loss: 0.72092625, Log Avg loss: 0.42646827, Global Avg Loss: 3.98817351, Time: 0.0248 Steps: 8440, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002063, Sample Num: 33008, Cur Loss: 0.95605588, Cur Avg Loss: 0.72118901, Log Avg loss: 0.77513355, Global Avg Loss: 3.98437109, Time: 0.0210 Steps: 8450, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002073, Sample Num: 33168, Cur Loss: 0.36939943, Cur Avg Loss: 0.71990774, Log Avg loss: 0.45558066, Global Avg Loss: 3.98019995, Time: 0.0210 Steps: 8460, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002083, Sample Num: 33328, Cur Loss: 0.67784512, Cur Avg Loss: 0.71897438, Log Avg loss: 0.52548896, Global Avg Loss: 3.97612119, Time: 0.0210 Steps: 8470, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002093, Sample Num: 33488, Cur Loss: 0.30849898, Cur Avg Loss: 0.71757943, Log Avg loss: 0.42701231, Global Avg Loss: 3.97193592, Time: 0.0210 Steps: 8480, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002103, Sample Num: 33648, Cur Loss: 0.31406492, Cur Avg Loss: 0.71669724, Log Avg loss: 0.53205512, Global Avg Loss: 3.96788423, Time: 0.0210 Steps: 8490, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002113, Sample Num: 33808, Cur Loss: 0.31788248, Cur Avg Loss: 0.71642372, Log Avg loss: 0.65890259, Global Avg Loss: 3.96399131, Time: 0.0210 Steps: 8500, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002123, Sample Num: 33968, Cur Loss: 0.89140832, Cur Avg Loss: 0.71649493, Log Avg loss: 0.73154022, Global Avg Loss: 3.96019290, Time: 0.0211 Steps: 8510, Updated lr: 0.000093 ***** Running evaluation checkpoint-8516 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-8516 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.686033, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.42829, "eval_total_loss": 301.087805, "eval_mae": 0.480787, "eval_mse": 0.428335, "eval_r2": 0.727722, "eval_sp_statistic": 0.823179, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.858997, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.640862, "test_total_loss": 321.712662, "test_mae": 0.495686, "test_mse": 0.641094, "test_r2": 0.586232, "test_sp_statistic": 0.818945, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.884286, "test_ps_pvalue": 0.0, "lr": 9.287245139876719e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.9577319878641255, "train_cur_epoch_loss": 1523.9227953515947, "train_cur_epoch_avg_loss": 0.7157927643736941, "train_cur_epoch_time": 44.68603324890137, "train_cur_epoch_avg_time": 0.020989212423157055, "epoch": 4, "step": 8516} ################################################## Training, Epoch: 0005, Batch: 000004, Sample Num: 64, Cur Loss: 0.54863793, Cur Avg Loss: 0.63518709, Log Avg loss: 0.53448133, Global Avg Loss: 3.95617211, Time: 0.0246 Steps: 8520, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000014, Sample Num: 224, Cur Loss: 0.49439126, Cur Avg Loss: 0.54644714, Log Avg loss: 0.51095116, Global Avg Loss: 3.95213316, Time: 0.0209 Steps: 8530, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000024, Sample Num: 384, Cur Loss: 0.59934705, Cur Avg Loss: 0.67124653, Log Avg loss: 0.84596569, Global Avg Loss: 3.94849596, Time: 0.0208 Steps: 8540, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000034, Sample Num: 544, Cur Loss: 0.26239187, Cur Avg Loss: 0.71972435, Log Avg loss: 0.83607111, Global Avg Loss: 3.94485570, Time: 0.0209 Steps: 8550, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000044, Sample Num: 704, Cur Loss: 0.30514762, Cur Avg Loss: 0.66016261, Log Avg loss: 0.45765271, Global Avg Loss: 3.94078186, Time: 0.0209 Steps: 8560, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000054, Sample Num: 864, Cur Loss: 0.58287889, Cur Avg Loss: 0.63260153, Log Avg loss: 0.51133275, Global Avg Loss: 3.93678017, Time: 0.0209 Steps: 8570, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000064, Sample Num: 1024, Cur Loss: 0.71724617, Cur Avg Loss: 0.62908814, Log Avg loss: 0.61011586, Global Avg Loss: 3.93290294, Time: 0.0209 Steps: 8580, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000074, Sample Num: 1184, Cur Loss: 0.50591761, Cur Avg Loss: 0.60613687, Log Avg loss: 0.45924870, Global Avg Loss: 3.92885911, Time: 0.0208 Steps: 8590, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000084, Sample Num: 1344, Cur Loss: 0.20333362, Cur Avg Loss: 0.57590535, Log Avg loss: 0.35219214, Global Avg Loss: 3.92470019, Time: 0.0210 Steps: 8600, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000094, Sample Num: 1504, Cur Loss: 0.48319510, Cur Avg Loss: 0.57342155, Log Avg loss: 0.55255764, Global Avg Loss: 3.92078365, Time: 0.0209 Steps: 8610, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000104, Sample Num: 1664, Cur Loss: 0.62521404, Cur Avg Loss: 0.55168769, Log Avg loss: 0.34738940, Global Avg Loss: 3.91663818, Time: 0.0209 Steps: 8620, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000114, Sample Num: 1824, Cur Loss: 0.27920347, Cur Avg Loss: 0.56021440, Log Avg loss: 0.64889214, Global Avg Loss: 3.91285169, Time: 0.0210 Steps: 8630, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000124, Sample Num: 1984, Cur Loss: 0.26436561, Cur Avg Loss: 0.59715645, Log Avg loss: 1.01829582, Global Avg Loss: 3.90950151, Time: 0.0210 Steps: 8640, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000134, Sample Num: 2144, Cur Loss: 0.70034111, Cur Avg Loss: 0.60226059, Log Avg loss: 0.66555198, Global Avg Loss: 3.90575127, Time: 0.0209 Steps: 8650, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000144, Sample Num: 2304, Cur Loss: 0.95149404, Cur Avg Loss: 0.67343785, Log Avg loss: 1.62721310, Global Avg Loss: 3.90312017, Time: 0.0210 Steps: 8660, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000154, Sample Num: 2464, Cur Loss: 1.21236777, Cur Avg Loss: 0.71630259, Log Avg loss: 1.33355492, Global Avg Loss: 3.90015643, Time: 0.0209 Steps: 8670, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000164, Sample Num: 2624, Cur Loss: 0.35527533, Cur Avg Loss: 0.71647008, Log Avg loss: 0.71904944, Global Avg Loss: 3.89649156, Time: 0.0210 Steps: 8680, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000174, Sample Num: 2784, Cur Loss: 0.45895374, Cur Avg Loss: 0.71133945, Log Avg loss: 0.62719699, Global Avg Loss: 3.89272942, Time: 0.0210 Steps: 8690, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000184, Sample Num: 2944, Cur Loss: 0.54943991, Cur Avg Loss: 0.70595216, Log Avg loss: 0.61221342, Global Avg Loss: 3.88895871, Time: 0.0210 Steps: 8700, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000194, Sample Num: 3104, Cur Loss: 0.38936025, Cur Avg Loss: 0.69466995, Log Avg loss: 0.48707719, Global Avg Loss: 3.88505299, Time: 0.0209 Steps: 8710, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000204, Sample Num: 3264, Cur Loss: 0.48704249, Cur Avg Loss: 0.68403250, Log Avg loss: 0.47766606, Global Avg Loss: 3.88114544, Time: 0.0209 Steps: 8720, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000214, Sample Num: 3424, Cur Loss: 0.59431213, Cur Avg Loss: 0.70131227, Log Avg loss: 1.05381958, Global Avg Loss: 3.87790681, Time: 0.0210 Steps: 8730, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000224, Sample Num: 3584, Cur Loss: 0.30480939, Cur Avg Loss: 0.69321063, Log Avg loss: 0.51983558, Global Avg Loss: 3.87406462, Time: 0.0210 Steps: 8740, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000234, Sample Num: 3744, Cur Loss: 0.45575672, Cur Avg Loss: 0.68549801, Log Avg loss: 0.51273521, Global Avg Loss: 3.87022310, Time: 0.0210 Steps: 8750, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000244, Sample Num: 3904, Cur Loss: 1.16474509, Cur Avg Loss: 0.68683675, Log Avg loss: 0.71816324, Global Avg Loss: 3.86662486, Time: 0.0209 Steps: 8760, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000254, Sample Num: 4064, Cur Loss: 0.42077142, Cur Avg Loss: 0.68248439, Log Avg loss: 0.57628695, Global Avg Loss: 3.86287305, Time: 0.0209 Steps: 8770, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000264, Sample Num: 4224, Cur Loss: 0.37911409, Cur Avg Loss: 0.68071326, Log Avg loss: 0.63572644, Global Avg Loss: 3.85919748, Time: 0.0210 Steps: 8780, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000274, Sample Num: 4384, Cur Loss: 0.18844926, Cur Avg Loss: 0.68018438, Log Avg loss: 0.66622202, Global Avg Loss: 3.85556497, Time: 0.0210 Steps: 8790, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000284, Sample Num: 4544, Cur Loss: 1.01395559, Cur Avg Loss: 0.67730097, Log Avg loss: 0.59829549, Global Avg Loss: 3.85186353, Time: 0.0209 Steps: 8800, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000294, Sample Num: 4704, Cur Loss: 0.11245463, Cur Avg Loss: 0.67746901, Log Avg loss: 0.68224143, Global Avg Loss: 3.84826578, Time: 0.0210 Steps: 8810, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000304, Sample Num: 4864, Cur Loss: 0.18909231, Cur Avg Loss: 0.66952918, Log Avg loss: 0.43609804, Global Avg Loss: 3.84439711, Time: 0.0209 Steps: 8820, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000314, Sample Num: 5024, Cur Loss: 0.69496608, Cur Avg Loss: 0.67100491, Log Avg loss: 0.71586705, Global Avg Loss: 3.84085404, Time: 0.0210 Steps: 8830, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000324, Sample Num: 5184, Cur Loss: 0.28183755, Cur Avg Loss: 0.66332537, Log Avg loss: 0.42218783, Global Avg Loss: 3.83698677, Time: 0.0209 Steps: 8840, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000334, Sample Num: 5344, Cur Loss: 0.31740904, Cur Avg Loss: 0.65708846, Log Avg loss: 0.45501265, Global Avg Loss: 3.83316533, Time: 0.0209 Steps: 8850, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000344, Sample Num: 5504, Cur Loss: 0.86465442, Cur Avg Loss: 0.65774205, Log Avg loss: 0.67957200, Global Avg Loss: 3.82960597, Time: 0.0209 Steps: 8860, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000354, Sample Num: 5664, Cur Loss: 0.78193998, Cur Avg Loss: 0.65354087, Log Avg loss: 0.50902021, Global Avg Loss: 3.82586235, Time: 0.0209 Steps: 8870, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000364, Sample Num: 5824, Cur Loss: 0.59920132, Cur Avg Loss: 0.65997653, Log Avg loss: 0.88779889, Global Avg Loss: 3.82255372, Time: 0.0209 Steps: 8880, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000374, Sample Num: 5984, Cur Loss: 1.30858946, Cur Avg Loss: 0.66914247, Log Avg loss: 1.00278265, Global Avg Loss: 3.81938188, Time: 0.0210 Steps: 8890, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000384, Sample Num: 6144, Cur Loss: 0.88877034, Cur Avg Loss: 0.66602427, Log Avg loss: 0.54940383, Global Avg Loss: 3.81570774, Time: 0.0209 Steps: 8900, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000394, Sample Num: 6304, Cur Loss: 0.62148631, Cur Avg Loss: 0.66107366, Log Avg loss: 0.47097022, Global Avg Loss: 3.81195383, Time: 0.0209 Steps: 8910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000404, Sample Num: 6464, Cur Loss: 0.39808795, Cur Avg Loss: 0.65530390, Log Avg loss: 0.42797500, Global Avg Loss: 3.80816013, Time: 0.0209 Steps: 8920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000414, Sample Num: 6624, Cur Loss: 0.47105944, Cur Avg Loss: 0.64943924, Log Avg loss: 0.41250712, Global Avg Loss: 3.80435761, Time: 0.0209 Steps: 8930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000424, Sample Num: 6784, Cur Loss: 1.29545534, Cur Avg Loss: 0.64824679, Log Avg loss: 0.59887932, Global Avg Loss: 3.80077206, Time: 0.0209 Steps: 8940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000434, Sample Num: 6944, Cur Loss: 0.54149234, Cur Avg Loss: 0.64995955, Log Avg loss: 0.72258067, Global Avg Loss: 3.79733274, Time: 0.0210 Steps: 8950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000444, Sample Num: 7104, Cur Loss: 0.68648863, Cur Avg Loss: 0.64846823, Log Avg loss: 0.58374475, Global Avg Loss: 3.79374615, Time: 0.0209 Steps: 8960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000454, Sample Num: 7264, Cur Loss: 0.26132900, Cur Avg Loss: 0.65350175, Log Avg loss: 0.87699002, Global Avg Loss: 3.79049447, Time: 0.0210 Steps: 8970, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000464, Sample Num: 7424, Cur Loss: 0.27927095, Cur Avg Loss: 0.66371094, Log Avg loss: 1.12720842, Global Avg Loss: 3.78752867, Time: 0.0209 Steps: 8980, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000474, Sample Num: 7584, Cur Loss: 0.63082963, Cur Avg Loss: 0.67141721, Log Avg loss: 1.02898788, Global Avg Loss: 3.78446022, Time: 0.0210 Steps: 8990, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000484, Sample Num: 7744, Cur Loss: 0.23859957, Cur Avg Loss: 0.68430805, Log Avg loss: 1.29533421, Global Avg Loss: 3.78169452, Time: 0.0209 Steps: 9000, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000494, Sample Num: 7904, Cur Loss: 0.54168320, Cur Avg Loss: 0.68332877, Log Avg loss: 0.63593152, Global Avg Loss: 3.77820311, Time: 0.0210 Steps: 9010, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000504, Sample Num: 8064, Cur Loss: 0.61877024, Cur Avg Loss: 0.67769212, Log Avg loss: 0.39924162, Global Avg Loss: 3.77445703, Time: 0.0209 Steps: 9020, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000514, Sample Num: 8224, Cur Loss: 0.35671234, Cur Avg Loss: 0.67337929, Log Avg loss: 0.45601263, Global Avg Loss: 3.77078212, Time: 0.0257 Steps: 9030, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000524, Sample Num: 8384, Cur Loss: 0.65280700, Cur Avg Loss: 0.67238609, Log Avg loss: 0.62133564, Global Avg Loss: 3.76729822, Time: 0.0219 Steps: 9040, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000534, Sample Num: 8544, Cur Loss: 0.44748712, Cur Avg Loss: 0.68007213, Log Avg loss: 1.08282054, Global Avg Loss: 3.76433195, Time: 0.0219 Steps: 9050, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000544, Sample Num: 8704, Cur Loss: 0.85732090, Cur Avg Loss: 0.67855905, Log Avg loss: 0.59776054, Global Avg Loss: 3.76083684, Time: 0.0219 Steps: 9060, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000554, Sample Num: 8864, Cur Loss: 0.44956639, Cur Avg Loss: 0.67755068, Log Avg loss: 0.62269556, Global Avg Loss: 3.75737692, Time: 0.0219 Steps: 9070, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000564, Sample Num: 9024, Cur Loss: 0.24609265, Cur Avg Loss: 0.67638183, Log Avg loss: 0.61162728, Global Avg Loss: 3.75391244, Time: 0.0219 Steps: 9080, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000574, Sample Num: 9184, Cur Loss: 0.48369834, Cur Avg Loss: 0.67269442, Log Avg loss: 0.46472436, Global Avg Loss: 3.75029397, Time: 0.0219 Steps: 9090, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000584, Sample Num: 9344, Cur Loss: 0.25827706, Cur Avg Loss: 0.66972773, Log Avg loss: 0.49944017, Global Avg Loss: 3.74672160, Time: 0.0219 Steps: 9100, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000594, Sample Num: 9504, Cur Loss: 0.77151000, Cur Avg Loss: 0.66729876, Log Avg loss: 0.52544690, Global Avg Loss: 3.74318563, Time: 0.0219 Steps: 9110, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000604, Sample Num: 9664, Cur Loss: 0.75714928, Cur Avg Loss: 0.66998031, Log Avg loss: 0.82926430, Global Avg Loss: 3.73999054, Time: 0.0219 Steps: 9120, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000614, Sample Num: 9824, Cur Loss: 0.49155885, Cur Avg Loss: 0.66607951, Log Avg loss: 0.43047079, Global Avg Loss: 3.73636565, Time: 0.0219 Steps: 9130, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000624, Sample Num: 9984, Cur Loss: 0.70509076, Cur Avg Loss: 0.66606063, Log Avg loss: 0.66490184, Global Avg Loss: 3.73300519, Time: 0.0219 Steps: 9140, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000634, Sample Num: 10144, Cur Loss: 0.69352150, Cur Avg Loss: 0.66549682, Log Avg loss: 0.63031514, Global Avg Loss: 3.72961427, Time: 0.0219 Steps: 9150, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000644, Sample Num: 10304, Cur Loss: 0.15804346, Cur Avg Loss: 0.66047138, Log Avg loss: 0.34185843, Global Avg Loss: 3.72591585, Time: 0.0219 Steps: 9160, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000654, Sample Num: 10464, Cur Loss: 0.22905031, Cur Avg Loss: 0.65730223, Log Avg loss: 0.45320910, Global Avg Loss: 3.72234692, Time: 0.0219 Steps: 9170, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000664, Sample Num: 10624, Cur Loss: 0.76835048, Cur Avg Loss: 0.65550935, Log Avg loss: 0.53825501, Global Avg Loss: 3.71887841, Time: 0.0219 Steps: 9180, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000674, Sample Num: 10784, Cur Loss: 1.19057512, Cur Avg Loss: 0.65949433, Log Avg loss: 0.92409636, Global Avg Loss: 3.71583730, Time: 0.0219 Steps: 9190, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000684, Sample Num: 10944, Cur Loss: 0.21846646, Cur Avg Loss: 0.65639581, Log Avg loss: 0.44755590, Global Avg Loss: 3.71228482, Time: 0.0219 Steps: 9200, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000694, Sample Num: 11104, Cur Loss: 0.48670542, Cur Avg Loss: 0.65349681, Log Avg loss: 0.45520487, Global Avg Loss: 3.70874836, Time: 0.0219 Steps: 9210, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000704, Sample Num: 11264, Cur Loss: 0.10915797, Cur Avg Loss: 0.65077915, Log Avg loss: 0.46217382, Global Avg Loss: 3.70522713, Time: 0.0219 Steps: 9220, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000714, Sample Num: 11424, Cur Loss: 0.38621682, Cur Avg Loss: 0.64904203, Log Avg loss: 0.52674858, Global Avg Loss: 3.70178349, Time: 0.0219 Steps: 9230, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000724, Sample Num: 11584, Cur Loss: 0.56258506, Cur Avg Loss: 0.64709809, Log Avg loss: 0.50830095, Global Avg Loss: 3.69832734, Time: 0.0219 Steps: 9240, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000734, Sample Num: 11744, Cur Loss: 0.18451752, Cur Avg Loss: 0.64471548, Log Avg loss: 0.47221430, Global Avg Loss: 3.69483965, Time: 0.0219 Steps: 9250, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000744, Sample Num: 11904, Cur Loss: 0.37292129, Cur Avg Loss: 0.64447041, Log Avg loss: 0.62648275, Global Avg Loss: 3.69152609, Time: 0.0219 Steps: 9260, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000754, Sample Num: 12064, Cur Loss: 0.28346843, Cur Avg Loss: 0.64195526, Log Avg loss: 0.45482775, Global Avg Loss: 3.68803451, Time: 0.0219 Steps: 9270, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000764, Sample Num: 12224, Cur Loss: 1.20458722, Cur Avg Loss: 0.63993026, Log Avg loss: 0.48724546, Global Avg Loss: 3.68458538, Time: 0.0219 Steps: 9280, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000774, Sample Num: 12384, Cur Loss: 0.35507217, Cur Avg Loss: 0.63706587, Log Avg loss: 0.41822665, Global Avg Loss: 3.68106939, Time: 0.0227 Steps: 9290, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000784, Sample Num: 12544, Cur Loss: 0.50350344, Cur Avg Loss: 0.63502034, Log Avg loss: 0.47669583, Global Avg Loss: 3.67762382, Time: 0.0209 Steps: 9300, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000794, Sample Num: 12704, Cur Loss: 0.67009854, Cur Avg Loss: 0.63385710, Log Avg loss: 0.54265897, Global Avg Loss: 3.67425651, Time: 0.0209 Steps: 9310, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000804, Sample Num: 12864, Cur Loss: 1.00299227, Cur Avg Loss: 0.63186139, Log Avg loss: 0.47340272, Global Avg Loss: 3.67082212, Time: 0.0209 Steps: 9320, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000814, Sample Num: 13024, Cur Loss: 0.23640339, Cur Avg Loss: 0.62977810, Log Avg loss: 0.46228162, Global Avg Loss: 3.66738317, Time: 0.0209 Steps: 9330, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000824, Sample Num: 13184, Cur Loss: 0.58072340, Cur Avg Loss: 0.62765376, Log Avg loss: 0.45473248, Global Avg Loss: 3.66394350, Time: 0.0209 Steps: 9340, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000834, Sample Num: 13344, Cur Loss: 0.77406418, Cur Avg Loss: 0.62535210, Log Avg loss: 0.43569534, Global Avg Loss: 3.66049083, Time: 0.0209 Steps: 9350, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000844, Sample Num: 13504, Cur Loss: 1.42712355, Cur Avg Loss: 0.62563769, Log Avg loss: 0.64945558, Global Avg Loss: 3.65727391, Time: 0.0209 Steps: 9360, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000854, Sample Num: 13664, Cur Loss: 1.14482176, Cur Avg Loss: 0.62956804, Log Avg loss: 0.96128911, Global Avg Loss: 3.65439666, Time: 0.0209 Steps: 9370, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000864, Sample Num: 13824, Cur Loss: 0.43042961, Cur Avg Loss: 0.62780742, Log Avg loss: 0.47745117, Global Avg Loss: 3.65100973, Time: 0.0209 Steps: 9380, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000874, Sample Num: 13984, Cur Loss: 1.96016502, Cur Avg Loss: 0.63884025, Log Avg loss: 1.59207630, Global Avg Loss: 3.64881704, Time: 0.0209 Steps: 9390, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000884, Sample Num: 14144, Cur Loss: 0.30930257, Cur Avg Loss: 0.63857832, Log Avg loss: 0.61568606, Global Avg Loss: 3.64559030, Time: 0.0211 Steps: 9400, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000894, Sample Num: 14304, Cur Loss: 0.83301067, Cur Avg Loss: 0.63733483, Log Avg loss: 0.52741039, Global Avg Loss: 3.64227662, Time: 0.0209 Steps: 9410, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000904, Sample Num: 14464, Cur Loss: 0.64128649, Cur Avg Loss: 0.63480590, Log Avg loss: 0.40871879, Global Avg Loss: 3.63884396, Time: 0.0209 Steps: 9420, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000914, Sample Num: 14624, Cur Loss: 0.59903127, Cur Avg Loss: 0.63524032, Log Avg loss: 0.67451259, Global Avg Loss: 3.63570045, Time: 0.0209 Steps: 9430, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000924, Sample Num: 14784, Cur Loss: 0.67894703, Cur Avg Loss: 0.63466103, Log Avg loss: 0.58171353, Global Avg Loss: 3.63246530, Time: 0.0209 Steps: 9440, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000934, Sample Num: 14944, Cur Loss: 0.73683745, Cur Avg Loss: 0.63418924, Log Avg loss: 0.59059625, Global Avg Loss: 3.62924639, Time: 0.0209 Steps: 9450, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000944, Sample Num: 15104, Cur Loss: 0.24736494, Cur Avg Loss: 0.63366652, Log Avg loss: 0.58484382, Global Avg Loss: 3.62602820, Time: 0.0209 Steps: 9460, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000954, Sample Num: 15264, Cur Loss: 0.27722776, Cur Avg Loss: 0.63166276, Log Avg loss: 0.44250862, Global Avg Loss: 3.62266651, Time: 0.0210 Steps: 9470, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000964, Sample Num: 15424, Cur Loss: 0.39285326, Cur Avg Loss: 0.62912527, Log Avg loss: 0.38704780, Global Avg Loss: 3.61925341, Time: 0.0209 Steps: 9480, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000974, Sample Num: 15584, Cur Loss: 0.69737828, Cur Avg Loss: 0.63011955, Log Avg loss: 0.72596906, Global Avg Loss: 3.61620464, Time: 0.0209 Steps: 9490, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000984, Sample Num: 15744, Cur Loss: 0.86731386, Cur Avg Loss: 0.63080366, Log Avg loss: 0.69743522, Global Avg Loss: 3.61313225, Time: 0.0209 Steps: 9500, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000994, Sample Num: 15904, Cur Loss: 1.65533876, Cur Avg Loss: 0.63213500, Log Avg loss: 0.76313873, Global Avg Loss: 3.61013541, Time: 0.0209 Steps: 9510, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001004, Sample Num: 16064, Cur Loss: 0.96506894, Cur Avg Loss: 0.63166434, Log Avg loss: 0.58488089, Global Avg Loss: 3.60695763, Time: 0.0209 Steps: 9520, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001014, Sample Num: 16224, Cur Loss: 0.21596503, Cur Avg Loss: 0.63047140, Log Avg loss: 0.51070083, Global Avg Loss: 3.60370867, Time: 0.0209 Steps: 9530, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001024, Sample Num: 16384, Cur Loss: 0.89495987, Cur Avg Loss: 0.62927045, Log Avg loss: 0.50749376, Global Avg Loss: 3.60046316, Time: 0.0255 Steps: 9540, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001034, Sample Num: 16544, Cur Loss: 1.78474438, Cur Avg Loss: 0.63023568, Log Avg loss: 0.72907518, Global Avg Loss: 3.59745647, Time: 0.0208 Steps: 9550, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001044, Sample Num: 16704, Cur Loss: 2.80825830, Cur Avg Loss: 0.63522171, Log Avg loss: 1.15077751, Global Avg Loss: 3.59489718, Time: 0.0209 Steps: 9560, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001054, Sample Num: 16864, Cur Loss: 1.19819713, Cur Avg Loss: 0.63646152, Log Avg loss: 0.76589752, Global Avg Loss: 3.59194107, Time: 0.0209 Steps: 9570, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001064, Sample Num: 17024, Cur Loss: 0.44941702, Cur Avg Loss: 0.63746242, Log Avg loss: 0.74295670, Global Avg Loss: 3.58896718, Time: 0.0209 Steps: 9580, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001074, Sample Num: 17184, Cur Loss: 1.01656592, Cur Avg Loss: 0.63741106, Log Avg loss: 0.63194702, Global Avg Loss: 3.58588374, Time: 0.0210 Steps: 9590, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001084, Sample Num: 17344, Cur Loss: 0.50960326, Cur Avg Loss: 0.63608332, Log Avg loss: 0.49348360, Global Avg Loss: 3.58266249, Time: 0.0210 Steps: 9600, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001094, Sample Num: 17504, Cur Loss: 0.47118402, Cur Avg Loss: 0.63593323, Log Avg loss: 0.61966391, Global Avg Loss: 3.57957925, Time: 0.0209 Steps: 9610, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001104, Sample Num: 17664, Cur Loss: 0.96966594, Cur Avg Loss: 0.63467241, Log Avg loss: 0.49673817, Global Avg Loss: 3.57637463, Time: 0.0209 Steps: 9620, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001114, Sample Num: 17824, Cur Loss: 0.44198954, Cur Avg Loss: 0.63463044, Log Avg loss: 0.62999720, Global Avg Loss: 3.57331505, Time: 0.0210 Steps: 9630, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001124, Sample Num: 17984, Cur Loss: 0.45471019, Cur Avg Loss: 0.63601744, Log Avg loss: 0.79052931, Global Avg Loss: 3.57042834, Time: 0.0209 Steps: 9640, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001134, Sample Num: 18144, Cur Loss: 0.60701644, Cur Avg Loss: 0.63624181, Log Avg loss: 0.66146118, Global Avg Loss: 3.56741387, Time: 0.0210 Steps: 9650, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001144, Sample Num: 18304, Cur Loss: 0.17975271, Cur Avg Loss: 0.63539627, Log Avg loss: 0.53951219, Global Avg Loss: 3.56427939, Time: 0.0209 Steps: 9660, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001154, Sample Num: 18464, Cur Loss: 0.80054235, Cur Avg Loss: 0.63573067, Log Avg loss: 0.67398537, Global Avg Loss: 3.56129047, Time: 0.0209 Steps: 9670, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001164, Sample Num: 18624, Cur Loss: 0.71128428, Cur Avg Loss: 0.63484404, Log Avg loss: 0.53252768, Global Avg Loss: 3.55816158, Time: 0.0209 Steps: 9680, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001174, Sample Num: 18784, Cur Loss: 1.02399552, Cur Avg Loss: 0.63423473, Log Avg loss: 0.56331065, Global Avg Loss: 3.55507092, Time: 0.0209 Steps: 9690, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001184, Sample Num: 18944, Cur Loss: 0.55001128, Cur Avg Loss: 0.63530558, Log Avg loss: 0.76102328, Global Avg Loss: 3.55219046, Time: 0.0209 Steps: 9700, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001194, Sample Num: 19104, Cur Loss: 0.46485460, Cur Avg Loss: 0.63318834, Log Avg loss: 0.38250774, Global Avg Loss: 3.54892611, Time: 0.0209 Steps: 9710, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001204, Sample Num: 19264, Cur Loss: 0.30627960, Cur Avg Loss: 0.63135744, Log Avg loss: 0.41274748, Global Avg Loss: 3.54569958, Time: 0.0209 Steps: 9720, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001214, Sample Num: 19424, Cur Loss: 0.13425957, Cur Avg Loss: 0.62941075, Log Avg loss: 0.39502910, Global Avg Loss: 3.54246149, Time: 0.0209 Steps: 9730, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001224, Sample Num: 19584, Cur Loss: 0.18355587, Cur Avg Loss: 0.62798145, Log Avg loss: 0.45446518, Global Avg Loss: 3.53929106, Time: 0.0209 Steps: 9740, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001234, Sample Num: 19744, Cur Loss: 0.40469423, Cur Avg Loss: 0.62575488, Log Avg loss: 0.35322215, Global Avg Loss: 3.53602330, Time: 0.0209 Steps: 9750, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001244, Sample Num: 19904, Cur Loss: 0.61505443, Cur Avg Loss: 0.62570702, Log Avg loss: 0.61980127, Global Avg Loss: 3.53303536, Time: 0.0209 Steps: 9760, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001254, Sample Num: 20064, Cur Loss: 0.33570766, Cur Avg Loss: 0.62402081, Log Avg loss: 0.41425610, Global Avg Loss: 3.52984316, Time: 0.0210 Steps: 9770, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001264, Sample Num: 20224, Cur Loss: 0.50672412, Cur Avg Loss: 0.62209287, Log Avg loss: 0.38032911, Global Avg Loss: 3.52662280, Time: 0.0209 Steps: 9780, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001274, Sample Num: 20384, Cur Loss: 0.74709153, Cur Avg Loss: 0.62210596, Log Avg loss: 0.62376032, Global Avg Loss: 3.52365767, Time: 0.0209 Steps: 9790, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001284, Sample Num: 20544, Cur Loss: 0.06367469, Cur Avg Loss: 0.62053418, Log Avg loss: 0.42029025, Global Avg Loss: 3.52049097, Time: 0.0248 Steps: 9800, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001294, Sample Num: 20704, Cur Loss: 0.48499727, Cur Avg Loss: 0.61905720, Log Avg loss: 0.42941296, Global Avg Loss: 3.51734002, Time: 0.0210 Steps: 9810, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001304, Sample Num: 20864, Cur Loss: 0.42963755, Cur Avg Loss: 0.61885719, Log Avg loss: 0.59297600, Global Avg Loss: 3.51436206, Time: 0.0210 Steps: 9820, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001314, Sample Num: 21024, Cur Loss: 0.19150087, Cur Avg Loss: 0.61754405, Log Avg loss: 0.44630964, Global Avg Loss: 3.51124094, Time: 0.0210 Steps: 9830, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001324, Sample Num: 21184, Cur Loss: 0.27357548, Cur Avg Loss: 0.61642981, Log Avg loss: 0.47001917, Global Avg Loss: 3.50815027, Time: 0.0210 Steps: 9840, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001334, Sample Num: 21344, Cur Loss: 0.23184043, Cur Avg Loss: 0.61649332, Log Avg loss: 0.62490250, Global Avg Loss: 3.50522312, Time: 0.0210 Steps: 9850, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001344, Sample Num: 21504, Cur Loss: 0.88418245, Cur Avg Loss: 0.61762748, Log Avg loss: 0.76892341, Global Avg Loss: 3.50244797, Time: 0.0210 Steps: 9860, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001354, Sample Num: 21664, Cur Loss: 0.47643858, Cur Avg Loss: 0.61693289, Log Avg loss: 0.52358063, Global Avg Loss: 3.49942986, Time: 0.0211 Steps: 9870, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001364, Sample Num: 21824, Cur Loss: 0.25019854, Cur Avg Loss: 0.61735436, Log Avg loss: 0.67442148, Global Avg Loss: 3.49657054, Time: 0.0210 Steps: 9880, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001374, Sample Num: 21984, Cur Loss: 0.69062543, Cur Avg Loss: 0.61800929, Log Avg loss: 0.70734189, Global Avg Loss: 3.49375029, Time: 0.0210 Steps: 9890, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001384, Sample Num: 22144, Cur Loss: 0.63271165, Cur Avg Loss: 0.61714700, Log Avg loss: 0.49866775, Global Avg Loss: 3.49072496, Time: 0.0210 Steps: 9900, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001394, Sample Num: 22304, Cur Loss: 0.44799542, Cur Avg Loss: 0.61933494, Log Avg loss: 0.92214614, Global Avg Loss: 3.48813305, Time: 0.0210 Steps: 9910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001404, Sample Num: 22464, Cur Loss: 0.32518917, Cur Avg Loss: 0.61851000, Log Avg loss: 0.50351377, Global Avg Loss: 3.48512436, Time: 0.0210 Steps: 9920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001414, Sample Num: 22624, Cur Loss: 0.69624579, Cur Avg Loss: 0.61749217, Log Avg loss: 0.47458776, Global Avg Loss: 3.48209260, Time: 0.0212 Steps: 9930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001424, Sample Num: 22784, Cur Loss: 0.44906881, Cur Avg Loss: 0.61583035, Log Avg loss: 0.38084992, Global Avg Loss: 3.47897264, Time: 0.0212 Steps: 9940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001434, Sample Num: 22944, Cur Loss: 0.31838560, Cur Avg Loss: 0.61408981, Log Avg loss: 0.36623604, Global Avg Loss: 3.47584426, Time: 0.0209 Steps: 9950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001444, Sample Num: 23104, Cur Loss: 0.30755150, Cur Avg Loss: 0.61338376, Log Avg loss: 0.51213650, Global Avg Loss: 3.47286865, Time: 0.0209 Steps: 9960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001454, Sample Num: 23264, Cur Loss: 0.40792763, Cur Avg Loss: 0.61397017, Log Avg loss: 0.69864793, Global Avg Loss: 3.47008608, Time: 0.0209 Steps: 9970, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001464, Sample Num: 23424, Cur Loss: 0.47229034, Cur Avg Loss: 0.61336102, Log Avg loss: 0.52479056, Global Avg Loss: 3.46713488, Time: 0.0209 Steps: 9980, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001474, Sample Num: 23584, Cur Loss: 0.61654782, Cur Avg Loss: 0.61323198, Log Avg loss: 0.59434091, Global Avg Loss: 3.46425921, Time: 0.0209 Steps: 9990, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001484, Sample Num: 23744, Cur Loss: 1.22786880, Cur Avg Loss: 0.61365347, Log Avg loss: 0.67578022, Global Avg Loss: 3.46147074, Time: 0.0209 Steps: 10000, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001494, Sample Num: 23904, Cur Loss: 0.27500868, Cur Avg Loss: 0.61241598, Log Avg loss: 0.42877351, Global Avg Loss: 3.45844107, Time: 0.0209 Steps: 10010, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001504, Sample Num: 24064, Cur Loss: 0.22682217, Cur Avg Loss: 0.61182853, Log Avg loss: 0.52406364, Global Avg Loss: 3.45551255, Time: 0.0208 Steps: 10020, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001514, Sample Num: 24224, Cur Loss: 0.76224136, Cur Avg Loss: 0.61197926, Log Avg loss: 0.63464886, Global Avg Loss: 3.45270012, Time: 0.0208 Steps: 10030, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001524, Sample Num: 24384, Cur Loss: 0.24696016, Cur Avg Loss: 0.61079336, Log Avg loss: 0.43124829, Global Avg Loss: 3.44969071, Time: 0.0208 Steps: 10040, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001534, Sample Num: 24544, Cur Loss: 0.72794294, Cur Avg Loss: 0.61075802, Log Avg loss: 0.60537184, Global Avg Loss: 3.44686054, Time: 0.0208 Steps: 10050, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001544, Sample Num: 24704, Cur Loss: 0.38137174, Cur Avg Loss: 0.61210000, Log Avg loss: 0.81795918, Global Avg Loss: 3.44424732, Time: 0.0207 Steps: 10060, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001554, Sample Num: 24864, Cur Loss: 0.31110203, Cur Avg Loss: 0.61335074, Log Avg loss: 0.80646535, Global Avg Loss: 3.44162787, Time: 0.0207 Steps: 10070, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001564, Sample Num: 25024, Cur Loss: 0.36082479, Cur Avg Loss: 0.61247496, Log Avg loss: 0.47637799, Global Avg Loss: 3.43868615, Time: 0.0208 Steps: 10080, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001574, Sample Num: 25184, Cur Loss: 0.24851790, Cur Avg Loss: 0.61268624, Log Avg loss: 0.64573166, Global Avg Loss: 3.43591811, Time: 0.0207 Steps: 10090, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001584, Sample Num: 25344, Cur Loss: 0.47489417, Cur Avg Loss: 0.61105820, Log Avg loss: 0.35480351, Global Avg Loss: 3.43286750, Time: 0.0208 Steps: 10100, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001594, Sample Num: 25504, Cur Loss: 0.48991477, Cur Avg Loss: 0.61038880, Log Avg loss: 0.50435596, Global Avg Loss: 3.42997086, Time: 0.0208 Steps: 10110, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001604, Sample Num: 25664, Cur Loss: 0.42781648, Cur Avg Loss: 0.60928787, Log Avg loss: 0.43380067, Global Avg Loss: 3.42701021, Time: 0.0207 Steps: 10120, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001614, Sample Num: 25824, Cur Loss: 0.30836692, Cur Avg Loss: 0.60872791, Log Avg loss: 0.51890943, Global Avg Loss: 3.42413943, Time: 0.0207 Steps: 10130, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001624, Sample Num: 25984, Cur Loss: 0.30706659, Cur Avg Loss: 0.60736815, Log Avg loss: 0.38790397, Global Avg Loss: 3.42114512, Time: 0.0208 Steps: 10140, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001634, Sample Num: 26144, Cur Loss: 0.83990294, Cur Avg Loss: 0.60798485, Log Avg loss: 0.70813616, Global Avg Loss: 3.41847220, Time: 0.0207 Steps: 10150, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001644, Sample Num: 26304, Cur Loss: 0.42061955, Cur Avg Loss: 0.60683596, Log Avg loss: 0.41910822, Global Avg Loss: 3.41552007, Time: 0.0208 Steps: 10160, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001654, Sample Num: 26464, Cur Loss: 0.94482887, Cur Avg Loss: 0.60648046, Log Avg loss: 0.54803528, Global Avg Loss: 3.41270052, Time: 0.0207 Steps: 10170, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001664, Sample Num: 26624, Cur Loss: 0.96750218, Cur Avg Loss: 0.60754825, Log Avg loss: 0.78416049, Global Avg Loss: 3.41011846, Time: 0.0208 Steps: 10180, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001674, Sample Num: 26784, Cur Loss: 2.90275645, Cur Avg Loss: 0.61117344, Log Avg loss: 1.21440474, Global Avg Loss: 3.40796368, Time: 0.0207 Steps: 10190, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001684, Sample Num: 26944, Cur Loss: 1.20096922, Cur Avg Loss: 0.61398746, Log Avg loss: 1.08505536, Global Avg Loss: 3.40568632, Time: 0.0207 Steps: 10200, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001694, Sample Num: 27104, Cur Loss: 0.29825610, Cur Avg Loss: 0.61485878, Log Avg loss: 0.76158920, Global Avg Loss: 3.40309661, Time: 0.0207 Steps: 10210, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001704, Sample Num: 27264, Cur Loss: 0.33828014, Cur Avg Loss: 0.61403877, Log Avg loss: 0.47512799, Global Avg Loss: 3.40023167, Time: 0.0208 Steps: 10220, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001714, Sample Num: 27424, Cur Loss: 0.24411699, Cur Avg Loss: 0.61276866, Log Avg loss: 0.39634318, Global Avg Loss: 3.39729532, Time: 0.0207 Steps: 10230, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001724, Sample Num: 27584, Cur Loss: 1.18559396, Cur Avg Loss: 0.61200877, Log Avg loss: 0.48176291, Global Avg Loss: 3.39444812, Time: 0.0207 Steps: 10240, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001734, Sample Num: 27744, Cur Loss: 0.95644051, Cur Avg Loss: 0.61233003, Log Avg loss: 0.66771630, Global Avg Loss: 3.39178789, Time: 0.0208 Steps: 10250, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001744, Sample Num: 27904, Cur Loss: 0.31551936, Cur Avg Loss: 0.61221987, Log Avg loss: 0.59311652, Global Avg Loss: 3.38906014, Time: 0.0207 Steps: 10260, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001754, Sample Num: 28064, Cur Loss: 0.50240940, Cur Avg Loss: 0.61129491, Log Avg loss: 0.44998288, Global Avg Loss: 3.38619833, Time: 0.0207 Steps: 10270, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001764, Sample Num: 28224, Cur Loss: 0.37645084, Cur Avg Loss: 0.60989526, Log Avg loss: 0.36439575, Global Avg Loss: 3.38325884, Time: 0.0207 Steps: 10280, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001774, Sample Num: 28384, Cur Loss: 0.16868506, Cur Avg Loss: 0.60926850, Log Avg loss: 0.49870821, Global Avg Loss: 3.38045558, Time: 0.0207 Steps: 10290, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001784, Sample Num: 28544, Cur Loss: 0.39538419, Cur Avg Loss: 0.60866290, Log Avg loss: 0.50123080, Global Avg Loss: 3.37766022, Time: 0.0207 Steps: 10300, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001794, Sample Num: 28704, Cur Loss: 0.42195505, Cur Avg Loss: 0.60776370, Log Avg loss: 0.44734477, Global Avg Loss: 3.37481801, Time: 0.0246 Steps: 10310, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001804, Sample Num: 28864, Cur Loss: 1.96289468, Cur Avg Loss: 0.60840814, Log Avg loss: 0.72402136, Global Avg Loss: 3.37224941, Time: 0.0207 Steps: 10320, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001814, Sample Num: 29024, Cur Loss: 0.40095645, Cur Avg Loss: 0.60818120, Log Avg loss: 0.56724091, Global Avg Loss: 3.36953401, Time: 0.0207 Steps: 10330, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001824, Sample Num: 29184, Cur Loss: 0.82227397, Cur Avg Loss: 0.60715000, Log Avg loss: 0.42008993, Global Avg Loss: 3.36668155, Time: 0.0207 Steps: 10340, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001834, Sample Num: 29344, Cur Loss: 0.86924183, Cur Avg Loss: 0.60673412, Log Avg loss: 0.53087901, Global Avg Loss: 3.36394164, Time: 0.0207 Steps: 10350, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001844, Sample Num: 29504, Cur Loss: 0.69728667, Cur Avg Loss: 0.60713095, Log Avg loss: 0.67990937, Global Avg Loss: 3.36135088, Time: 0.0207 Steps: 10360, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001854, Sample Num: 29664, Cur Loss: 0.87112045, Cur Avg Loss: 0.60650240, Log Avg loss: 0.49059800, Global Avg Loss: 3.35858255, Time: 0.0208 Steps: 10370, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001864, Sample Num: 29824, Cur Loss: 1.01014757, Cur Avg Loss: 0.60732033, Log Avg loss: 0.75896331, Global Avg Loss: 3.35607810, Time: 0.0207 Steps: 10380, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001874, Sample Num: 29984, Cur Loss: 0.52713692, Cur Avg Loss: 0.60694007, Log Avg loss: 0.53606103, Global Avg Loss: 3.35336394, Time: 0.0207 Steps: 10390, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001884, Sample Num: 30144, Cur Loss: 0.35174578, Cur Avg Loss: 0.60896972, Log Avg loss: 0.98932464, Global Avg Loss: 3.35109082, Time: 0.0207 Steps: 10400, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001894, Sample Num: 30304, Cur Loss: 0.96199298, Cur Avg Loss: 0.61064429, Log Avg loss: 0.92613435, Global Avg Loss: 3.34876137, Time: 0.0207 Steps: 10410, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001904, Sample Num: 30464, Cur Loss: 0.79557836, Cur Avg Loss: 0.61216004, Log Avg loss: 0.89924191, Global Avg Loss: 3.34641059, Time: 0.0207 Steps: 10420, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001914, Sample Num: 30624, Cur Loss: 0.88933688, Cur Avg Loss: 0.61270970, Log Avg loss: 0.71736605, Global Avg Loss: 3.34388993, Time: 0.0207 Steps: 10430, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001924, Sample Num: 30784, Cur Loss: 0.28240561, Cur Avg Loss: 0.61262629, Log Avg loss: 0.59666114, Global Avg Loss: 3.34125849, Time: 0.0207 Steps: 10440, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001934, Sample Num: 30944, Cur Loss: 0.85403275, Cur Avg Loss: 0.61246687, Log Avg loss: 0.58179415, Global Avg Loss: 3.33861785, Time: 0.0207 Steps: 10450, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001944, Sample Num: 31104, Cur Loss: 0.86700070, Cur Avg Loss: 0.61261450, Log Avg loss: 0.64116727, Global Avg Loss: 3.33603903, Time: 0.0207 Steps: 10460, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001954, Sample Num: 31264, Cur Loss: 0.31543991, Cur Avg Loss: 0.61180488, Log Avg loss: 0.45441454, Global Avg Loss: 3.33328676, Time: 0.0207 Steps: 10470, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001964, Sample Num: 31424, Cur Loss: 0.27730441, Cur Avg Loss: 0.61038985, Log Avg loss: 0.33389275, Global Avg Loss: 3.33042474, Time: 0.0207 Steps: 10480, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001974, Sample Num: 31584, Cur Loss: 0.56325591, Cur Avg Loss: 0.60938669, Log Avg loss: 0.41236618, Global Avg Loss: 3.32764299, Time: 0.0207 Steps: 10490, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001984, Sample Num: 31744, Cur Loss: 0.65933478, Cur Avg Loss: 0.60922019, Log Avg loss: 0.57635329, Global Avg Loss: 3.32502271, Time: 0.0207 Steps: 10500, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001994, Sample Num: 31904, Cur Loss: 0.60692221, Cur Avg Loss: 0.61108270, Log Avg loss: 0.98060394, Global Avg Loss: 3.32279206, Time: 0.0207 Steps: 10510, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002004, Sample Num: 32064, Cur Loss: 0.33391958, Cur Avg Loss: 0.61131633, Log Avg loss: 0.65790272, Global Avg Loss: 3.32025889, Time: 0.0207 Steps: 10520, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002014, Sample Num: 32224, Cur Loss: 0.35402995, Cur Avg Loss: 0.61181031, Log Avg loss: 0.71080397, Global Avg Loss: 3.31778078, Time: 0.0207 Steps: 10530, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002024, Sample Num: 32384, Cur Loss: 0.47997507, Cur Avg Loss: 0.61032848, Log Avg loss: 0.31188823, Global Avg Loss: 3.31492889, Time: 0.0207 Steps: 10540, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002034, Sample Num: 32544, Cur Loss: 0.63522929, Cur Avg Loss: 0.61017750, Log Avg loss: 0.57961877, Global Avg Loss: 3.31233618, Time: 0.0207 Steps: 10550, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002044, Sample Num: 32704, Cur Loss: 0.54975653, Cur Avg Loss: 0.60985915, Log Avg loss: 0.54510660, Global Avg Loss: 3.30971569, Time: 0.0207 Steps: 10560, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002054, Sample Num: 32864, Cur Loss: 0.62954414, Cur Avg Loss: 0.60909880, Log Avg loss: 0.45368384, Global Avg Loss: 3.30701368, Time: 0.0210 Steps: 10570, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002064, Sample Num: 33024, Cur Loss: 0.85861230, Cur Avg Loss: 0.60828805, Log Avg loss: 0.44175944, Global Avg Loss: 3.30430550, Time: 0.0208 Steps: 10580, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002074, Sample Num: 33184, Cur Loss: 0.33565438, Cur Avg Loss: 0.60819276, Log Avg loss: 0.58852576, Global Avg Loss: 3.30174102, Time: 0.0209 Steps: 10590, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002084, Sample Num: 33344, Cur Loss: 0.60175008, Cur Avg Loss: 0.60745226, Log Avg loss: 0.45387248, Global Avg Loss: 3.29905435, Time: 0.0208 Steps: 10600, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002094, Sample Num: 33504, Cur Loss: 0.89769065, Cur Avg Loss: 0.60667683, Log Avg loss: 0.44507522, Global Avg Loss: 3.29636446, Time: 0.0208 Steps: 10610, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002104, Sample Num: 33664, Cur Loss: 1.04617810, Cur Avg Loss: 0.60609996, Log Avg loss: 0.48530419, Global Avg Loss: 3.29371751, Time: 0.0208 Steps: 10620, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002114, Sample Num: 33824, Cur Loss: 0.52180165, Cur Avg Loss: 0.60624040, Log Avg loss: 0.63578882, Global Avg Loss: 3.29121710, Time: 0.0208 Steps: 10630, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002124, Sample Num: 33984, Cur Loss: 0.33284721, Cur Avg Loss: 0.60487171, Log Avg loss: 0.31553034, Global Avg Loss: 3.28842041, Time: 0.0209 Steps: 10640, Updated lr: 0.000091 ***** Running evaluation checkpoint-10645 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-10645 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.781603, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.472671, "eval_total_loss": 332.287401, "eval_mae": 0.554402, "eval_mse": 0.472632, "eval_r2": 0.699564, "eval_sp_statistic": 0.829756, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.871736, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.43018, "test_total_loss": 215.950361, "test_mae": 0.413813, "test_mse": 0.430336, "test_r2": 0.722257, "test_sp_statistic": 0.826198, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.899271, "test_ps_pvalue": 0.0, "lr": 9.085348506401138e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.286986468281263, "train_cur_epoch_loss": 1285.9253462031484, "train_cur_epoch_avg_loss": 0.6040043899498113, "train_cur_epoch_time": 44.78160333633423, "train_cur_epoch_avg_time": 0.021034102083764317, "epoch": 5, "step": 10645} ################################################## Training, Epoch: 0006, Batch: 000005, Sample Num: 80, Cur Loss: 0.19207782, Cur Avg Loss: 0.25132183, Log Avg loss: 0.24344502, Global Avg Loss: 3.28556127, Time: 0.0246 Steps: 10650, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000015, Sample Num: 240, Cur Loss: 0.38861406, Cur Avg Loss: 0.27780571, Log Avg loss: 0.29104765, Global Avg Loss: 3.28275216, Time: 0.0209 Steps: 10660, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000025, Sample Num: 400, Cur Loss: 1.06531346, Cur Avg Loss: 0.37243405, Log Avg loss: 0.51437657, Global Avg Loss: 3.28015762, Time: 0.0209 Steps: 10670, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000035, Sample Num: 560, Cur Loss: 0.43491018, Cur Avg Loss: 0.45693194, Log Avg loss: 0.66817668, Global Avg Loss: 3.27771195, Time: 0.0209 Steps: 10680, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000045, Sample Num: 720, Cur Loss: 0.39889610, Cur Avg Loss: 0.44031333, Log Avg loss: 0.38214819, Global Avg Loss: 3.27500328, Time: 0.0209 Steps: 10690, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000055, Sample Num: 880, Cur Loss: 0.33934379, Cur Avg Loss: 0.43327519, Log Avg loss: 0.40160352, Global Avg Loss: 3.27231786, Time: 0.0209 Steps: 10700, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000065, Sample Num: 1040, Cur Loss: 0.65371352, Cur Avg Loss: 0.43000921, Log Avg loss: 0.41204634, Global Avg Loss: 3.26964720, Time: 0.0210 Steps: 10710, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000075, Sample Num: 1200, Cur Loss: 0.33129442, Cur Avg Loss: 0.42724045, Log Avg loss: 0.40924350, Global Avg Loss: 3.26697892, Time: 0.0210 Steps: 10720, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000085, Sample Num: 1360, Cur Loss: 0.56038153, Cur Avg Loss: 0.43913891, Log Avg loss: 0.52837735, Global Avg Loss: 3.26442663, Time: 0.0209 Steps: 10730, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000095, Sample Num: 1520, Cur Loss: 0.17731659, Cur Avg Loss: 0.43258188, Log Avg loss: 0.37684715, Global Avg Loss: 3.26173801, Time: 0.0208 Steps: 10740, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000105, Sample Num: 1680, Cur Loss: 0.83357042, Cur Avg Loss: 0.44981447, Log Avg loss: 0.61352413, Global Avg Loss: 3.25927456, Time: 0.0209 Steps: 10750, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000115, Sample Num: 1840, Cur Loss: 0.70975202, Cur Avg Loss: 0.45416171, Log Avg loss: 0.49980775, Global Avg Loss: 3.25671000, Time: 0.0210 Steps: 10760, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000125, Sample Num: 2000, Cur Loss: 0.16335839, Cur Avg Loss: 0.46029910, Log Avg loss: 0.53087906, Global Avg Loss: 3.25417905, Time: 0.0209 Steps: 10770, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000135, Sample Num: 2160, Cur Loss: 0.69261432, Cur Avg Loss: 0.45785686, Log Avg loss: 0.42732879, Global Avg Loss: 3.25155674, Time: 0.0209 Steps: 10780, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000145, Sample Num: 2320, Cur Loss: 0.46761388, Cur Avg Loss: 0.47904708, Log Avg loss: 0.76511507, Global Avg Loss: 3.24925234, Time: 0.0207 Steps: 10790, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000155, Sample Num: 2480, Cur Loss: 0.76470959, Cur Avg Loss: 0.49102263, Log Avg loss: 0.66466819, Global Avg Loss: 3.24685921, Time: 0.0208 Steps: 10800, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000165, Sample Num: 2640, Cur Loss: 0.56391603, Cur Avg Loss: 0.50823340, Log Avg loss: 0.77500019, Global Avg Loss: 3.24457257, Time: 0.0209 Steps: 10810, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000175, Sample Num: 2800, Cur Loss: 0.34508672, Cur Avg Loss: 0.49919389, Log Avg loss: 0.35004206, Global Avg Loss: 3.24189740, Time: 0.0209 Steps: 10820, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000185, Sample Num: 2960, Cur Loss: 0.77775264, Cur Avg Loss: 0.49742346, Log Avg loss: 0.46644093, Global Avg Loss: 3.23933465, Time: 0.0208 Steps: 10830, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000195, Sample Num: 3120, Cur Loss: 0.10582708, Cur Avg Loss: 0.48717126, Log Avg loss: 0.29750561, Global Avg Loss: 3.23662079, Time: 0.0209 Steps: 10840, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000205, Sample Num: 3280, Cur Loss: 0.33726203, Cur Avg Loss: 0.48289906, Log Avg loss: 0.39959117, Global Avg Loss: 3.23400602, Time: 0.0209 Steps: 10850, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000215, Sample Num: 3440, Cur Loss: 2.34037447, Cur Avg Loss: 0.52902347, Log Avg loss: 1.47457383, Global Avg Loss: 3.23238591, Time: 0.0209 Steps: 10860, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000225, Sample Num: 3600, Cur Loss: 0.58883953, Cur Avg Loss: 0.56128621, Log Avg loss: 1.25493505, Global Avg Loss: 3.23056673, Time: 0.0208 Steps: 10870, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000235, Sample Num: 3760, Cur Loss: 0.23833060, Cur Avg Loss: 0.55594134, Log Avg loss: 0.43568175, Global Avg Loss: 3.22799790, Time: 0.0209 Steps: 10880, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000245, Sample Num: 3920, Cur Loss: 0.62382770, Cur Avg Loss: 0.55819829, Log Avg loss: 0.61123674, Global Avg Loss: 3.22559500, Time: 0.0208 Steps: 10890, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000255, Sample Num: 4080, Cur Loss: 0.65943277, Cur Avg Loss: 0.56623614, Log Avg loss: 0.76316330, Global Avg Loss: 3.22333589, Time: 0.0208 Steps: 10900, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000265, Sample Num: 4240, Cur Loss: 0.25271755, Cur Avg Loss: 0.56390459, Log Avg loss: 0.50445023, Global Avg Loss: 3.22084378, Time: 0.0208 Steps: 10910, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000275, Sample Num: 4400, Cur Loss: 0.19505250, Cur Avg Loss: 0.55904790, Log Avg loss: 0.43034556, Global Avg Loss: 3.21828838, Time: 0.0209 Steps: 10920, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000285, Sample Num: 4560, Cur Loss: 0.50884062, Cur Avg Loss: 0.55824015, Log Avg loss: 0.53602698, Global Avg Loss: 3.21583435, Time: 0.0208 Steps: 10930, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000295, Sample Num: 4720, Cur Loss: 1.31972611, Cur Avg Loss: 0.55841994, Log Avg loss: 0.56354390, Global Avg Loss: 3.21340995, Time: 0.0209 Steps: 10940, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000305, Sample Num: 4880, Cur Loss: 0.32788426, Cur Avg Loss: 0.56468824, Log Avg loss: 0.74960313, Global Avg Loss: 3.21115990, Time: 0.0208 Steps: 10950, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000315, Sample Num: 5040, Cur Loss: 0.34980837, Cur Avg Loss: 0.56157523, Log Avg loss: 0.46662841, Global Avg Loss: 3.20865576, Time: 0.0208 Steps: 10960, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000325, Sample Num: 5200, Cur Loss: 0.29759941, Cur Avg Loss: 0.56446683, Log Avg loss: 0.65555231, Global Avg Loss: 3.20632841, Time: 0.0207 Steps: 10970, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000335, Sample Num: 5360, Cur Loss: 0.27227908, Cur Avg Loss: 0.56592980, Log Avg loss: 0.61347633, Global Avg Loss: 3.20396698, Time: 0.0210 Steps: 10980, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000345, Sample Num: 5520, Cur Loss: 0.50907743, Cur Avg Loss: 0.56720862, Log Avg loss: 0.61004920, Global Avg Loss: 3.20160673, Time: 0.0209 Steps: 10990, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000355, Sample Num: 5680, Cur Loss: 0.52150655, Cur Avg Loss: 0.56749889, Log Avg loss: 0.57751292, Global Avg Loss: 3.19922119, Time: 0.0208 Steps: 11000, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000365, Sample Num: 5840, Cur Loss: 0.35960734, Cur Avg Loss: 0.56709864, Log Avg loss: 0.55288987, Global Avg Loss: 3.19681762, Time: 0.0208 Steps: 11010, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000375, Sample Num: 6000, Cur Loss: 0.51648474, Cur Avg Loss: 0.56474635, Log Avg loss: 0.47888790, Global Avg Loss: 3.19435126, Time: 0.0209 Steps: 11020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000385, Sample Num: 6160, Cur Loss: 0.19829661, Cur Avg Loss: 0.55972141, Log Avg loss: 0.37128609, Global Avg Loss: 3.19179181, Time: 0.0209 Steps: 11030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000395, Sample Num: 6320, Cur Loss: 0.28891689, Cur Avg Loss: 0.55634417, Log Avg loss: 0.42632041, Global Avg Loss: 3.18928686, Time: 0.0209 Steps: 11040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000405, Sample Num: 6480, Cur Loss: 0.55285776, Cur Avg Loss: 0.55069096, Log Avg loss: 0.32738916, Global Avg Loss: 3.18669690, Time: 0.0209 Steps: 11050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000415, Sample Num: 6640, Cur Loss: 0.48066175, Cur Avg Loss: 0.54724739, Log Avg loss: 0.40778297, Global Avg Loss: 3.18418432, Time: 0.0208 Steps: 11060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000425, Sample Num: 6800, Cur Loss: 0.57072884, Cur Avg Loss: 0.54503033, Log Avg loss: 0.45302219, Global Avg Loss: 3.18171715, Time: 0.0208 Steps: 11070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000435, Sample Num: 6960, Cur Loss: 0.32993159, Cur Avg Loss: 0.54533801, Log Avg loss: 0.55841427, Global Avg Loss: 3.17934955, Time: 0.0208 Steps: 11080, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000445, Sample Num: 7120, Cur Loss: 0.20291087, Cur Avg Loss: 0.54170820, Log Avg loss: 0.38381180, Global Avg Loss: 3.17682877, Time: 0.0207 Steps: 11090, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000455, Sample Num: 7280, Cur Loss: 0.63500088, Cur Avg Loss: 0.54140647, Log Avg loss: 0.52797950, Global Avg Loss: 3.17444242, Time: 0.0209 Steps: 11100, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000465, Sample Num: 7440, Cur Loss: 0.66749471, Cur Avg Loss: 0.53865288, Log Avg loss: 0.41336430, Global Avg Loss: 3.17195720, Time: 0.0209 Steps: 11110, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000475, Sample Num: 7600, Cur Loss: 0.50083739, Cur Avg Loss: 0.53749981, Log Avg loss: 0.48388222, Global Avg Loss: 3.16953987, Time: 0.0208 Steps: 11120, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000485, Sample Num: 7760, Cur Loss: 0.23178658, Cur Avg Loss: 0.53512190, Log Avg loss: 0.42217119, Global Avg Loss: 3.16707144, Time: 0.0209 Steps: 11130, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000495, Sample Num: 7920, Cur Loss: 0.11013177, Cur Avg Loss: 0.53435659, Log Avg loss: 0.49723875, Global Avg Loss: 3.16467482, Time: 0.0209 Steps: 11140, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000505, Sample Num: 8080, Cur Loss: 1.08099711, Cur Avg Loss: 0.53337905, Log Avg loss: 0.48499086, Global Avg Loss: 3.16227151, Time: 0.0209 Steps: 11150, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000515, Sample Num: 8240, Cur Loss: 1.19935715, Cur Avg Loss: 0.54535486, Log Avg loss: 1.15013353, Global Avg Loss: 3.16046852, Time: 0.0245 Steps: 11160, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000525, Sample Num: 8400, Cur Loss: 1.07282615, Cur Avg Loss: 0.55302579, Log Avg loss: 0.94807848, Global Avg Loss: 3.15848787, Time: 0.0208 Steps: 11170, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000535, Sample Num: 8560, Cur Loss: 0.55314744, Cur Avg Loss: 0.55994756, Log Avg loss: 0.92334047, Global Avg Loss: 3.15648863, Time: 0.0208 Steps: 11180, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000545, Sample Num: 8720, Cur Loss: 0.69996381, Cur Avg Loss: 0.56754252, Log Avg loss: 0.97387306, Global Avg Loss: 3.15453813, Time: 0.0207 Steps: 11190, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000555, Sample Num: 8880, Cur Loss: 0.15393735, Cur Avg Loss: 0.56970727, Log Avg loss: 0.68768607, Global Avg Loss: 3.15233558, Time: 0.0208 Steps: 11200, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000565, Sample Num: 9040, Cur Loss: 0.36513352, Cur Avg Loss: 0.56611996, Log Avg loss: 0.36702445, Global Avg Loss: 3.14985091, Time: 0.0208 Steps: 11210, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000575, Sample Num: 9200, Cur Loss: 0.16425008, Cur Avg Loss: 0.56218987, Log Avg loss: 0.34013950, Global Avg Loss: 3.14734671, Time: 0.0208 Steps: 11220, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000585, Sample Num: 9360, Cur Loss: 0.28661376, Cur Avg Loss: 0.56215554, Log Avg loss: 0.56018145, Global Avg Loss: 3.14504292, Time: 0.0208 Steps: 11230, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000595, Sample Num: 9520, Cur Loss: 0.60247612, Cur Avg Loss: 0.55948775, Log Avg loss: 0.40342190, Global Avg Loss: 3.14260375, Time: 0.0208 Steps: 11240, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000605, Sample Num: 9680, Cur Loss: 0.10749044, Cur Avg Loss: 0.55789879, Log Avg loss: 0.46335624, Global Avg Loss: 3.14022220, Time: 0.0208 Steps: 11250, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000615, Sample Num: 9840, Cur Loss: 0.41257313, Cur Avg Loss: 0.55881509, Log Avg loss: 0.61425094, Global Avg Loss: 3.13797888, Time: 0.0208 Steps: 11260, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000625, Sample Num: 10000, Cur Loss: 0.24725002, Cur Avg Loss: 0.55647927, Log Avg loss: 0.41282662, Global Avg Loss: 3.13556083, Time: 0.0207 Steps: 11270, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000635, Sample Num: 10160, Cur Loss: 0.35254508, Cur Avg Loss: 0.55441122, Log Avg loss: 0.42515774, Global Avg Loss: 3.13315799, Time: 0.0207 Steps: 11280, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000645, Sample Num: 10320, Cur Loss: 1.29145563, Cur Avg Loss: 0.55335240, Log Avg loss: 0.48611759, Global Avg Loss: 3.13081340, Time: 0.0207 Steps: 11290, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000655, Sample Num: 10480, Cur Loss: 0.35495174, Cur Avg Loss: 0.55452794, Log Avg loss: 0.63035038, Global Avg Loss: 3.12860060, Time: 0.0208 Steps: 11300, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000665, Sample Num: 10640, Cur Loss: 0.25967097, Cur Avg Loss: 0.55322278, Log Avg loss: 0.46773441, Global Avg Loss: 3.12624793, Time: 0.0207 Steps: 11310, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000675, Sample Num: 10800, Cur Loss: 0.77782583, Cur Avg Loss: 0.55302575, Log Avg loss: 0.53992353, Global Avg Loss: 3.12396319, Time: 0.0208 Steps: 11320, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000685, Sample Num: 10960, Cur Loss: 0.43076017, Cur Avg Loss: 0.55131672, Log Avg loss: 0.43595719, Global Avg Loss: 3.12159072, Time: 0.0208 Steps: 11330, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000695, Sample Num: 11120, Cur Loss: 0.31238112, Cur Avg Loss: 0.54955864, Log Avg loss: 0.42913010, Global Avg Loss: 3.11921642, Time: 0.0207 Steps: 11340, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000705, Sample Num: 11280, Cur Loss: 0.35112190, Cur Avg Loss: 0.54937402, Log Avg loss: 0.53654292, Global Avg Loss: 3.11694094, Time: 0.0207 Steps: 11350, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000715, Sample Num: 11440, Cur Loss: 2.08019805, Cur Avg Loss: 0.54849029, Log Avg loss: 0.48618740, Global Avg Loss: 3.11462513, Time: 0.0208 Steps: 11360, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000725, Sample Num: 11600, Cur Loss: 0.58101493, Cur Avg Loss: 0.55140919, Log Avg loss: 0.76011067, Global Avg Loss: 3.11255432, Time: 0.0208 Steps: 11370, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000735, Sample Num: 11760, Cur Loss: 0.66976458, Cur Avg Loss: 0.55486857, Log Avg loss: 0.80567329, Global Avg Loss: 3.11052718, Time: 0.0208 Steps: 11380, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000745, Sample Num: 11920, Cur Loss: 0.82253480, Cur Avg Loss: 0.56159490, Log Avg loss: 1.05598056, Global Avg Loss: 3.10872337, Time: 0.0207 Steps: 11390, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000755, Sample Num: 12080, Cur Loss: 0.14957854, Cur Avg Loss: 0.56575687, Log Avg loss: 0.87582340, Global Avg Loss: 3.10676468, Time: 0.0207 Steps: 11400, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000765, Sample Num: 12240, Cur Loss: 0.18333799, Cur Avg Loss: 0.56557495, Log Avg loss: 0.55184026, Global Avg Loss: 3.10452549, Time: 0.0208 Steps: 11410, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000775, Sample Num: 12400, Cur Loss: 0.61313391, Cur Avg Loss: 0.56406769, Log Avg loss: 0.44876222, Global Avg Loss: 3.10219995, Time: 0.0208 Steps: 11420, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000785, Sample Num: 12560, Cur Loss: 0.64348900, Cur Avg Loss: 0.56350426, Log Avg loss: 0.51983800, Global Avg Loss: 3.09994066, Time: 0.0207 Steps: 11430, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000795, Sample Num: 12720, Cur Loss: 0.49586260, Cur Avg Loss: 0.56500330, Log Avg loss: 0.68267786, Global Avg Loss: 3.09782767, Time: 0.0207 Steps: 11440, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000805, Sample Num: 12880, Cur Loss: 0.25026536, Cur Avg Loss: 0.56418071, Log Avg loss: 0.49878477, Global Avg Loss: 3.09555777, Time: 0.0207 Steps: 11450, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000815, Sample Num: 13040, Cur Loss: 0.31028077, Cur Avg Loss: 0.56433406, Log Avg loss: 0.57667926, Global Avg Loss: 3.09335979, Time: 0.0207 Steps: 11460, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000825, Sample Num: 13200, Cur Loss: 0.31943744, Cur Avg Loss: 0.56358062, Log Avg loss: 0.50217520, Global Avg Loss: 3.09110069, Time: 0.0207 Steps: 11470, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000835, Sample Num: 13360, Cur Loss: 0.39150184, Cur Avg Loss: 0.56157012, Log Avg loss: 0.39570356, Global Avg Loss: 3.08875279, Time: 0.0207 Steps: 11480, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000845, Sample Num: 13520, Cur Loss: 0.41297346, Cur Avg Loss: 0.56736348, Log Avg loss: 1.05110943, Global Avg Loss: 3.08697938, Time: 0.0207 Steps: 11490, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000855, Sample Num: 13680, Cur Loss: 3.04169512, Cur Avg Loss: 0.57490369, Log Avg loss: 1.21205073, Global Avg Loss: 3.08534901, Time: 0.0208 Steps: 11500, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000865, Sample Num: 13840, Cur Loss: 0.64341921, Cur Avg Loss: 0.57775327, Log Avg loss: 0.82139312, Global Avg Loss: 3.08338206, Time: 0.0208 Steps: 11510, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000875, Sample Num: 14000, Cur Loss: 0.28830761, Cur Avg Loss: 0.57719475, Log Avg loss: 0.52888259, Global Avg Loss: 3.08116461, Time: 0.0207 Steps: 11520, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000885, Sample Num: 14160, Cur Loss: 0.44719976, Cur Avg Loss: 0.57458342, Log Avg loss: 0.34609191, Global Avg Loss: 3.07879248, Time: 0.0207 Steps: 11530, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000895, Sample Num: 14320, Cur Loss: 0.66401696, Cur Avg Loss: 0.57220242, Log Avg loss: 0.36148400, Global Avg Loss: 3.07643779, Time: 0.0208 Steps: 11540, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000905, Sample Num: 14480, Cur Loss: 0.89303654, Cur Avg Loss: 0.57163942, Log Avg loss: 0.52125096, Global Avg Loss: 3.07422551, Time: 0.0208 Steps: 11550, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000915, Sample Num: 14640, Cur Loss: 0.76930243, Cur Avg Loss: 0.57068392, Log Avg loss: 0.48421094, Global Avg Loss: 3.07198501, Time: 0.0207 Steps: 11560, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000925, Sample Num: 14800, Cur Loss: 0.94949579, Cur Avg Loss: 0.56992442, Log Avg loss: 0.50043019, Global Avg Loss: 3.06976241, Time: 0.0207 Steps: 11570, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000935, Sample Num: 14960, Cur Loss: 0.69258362, Cur Avg Loss: 0.56776303, Log Avg loss: 0.36783411, Global Avg Loss: 3.06742914, Time: 0.0207 Steps: 11580, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000945, Sample Num: 15120, Cur Loss: 0.22803195, Cur Avg Loss: 0.56615738, Log Avg loss: 0.41602933, Global Avg Loss: 3.06514147, Time: 0.0207 Steps: 11590, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000955, Sample Num: 15280, Cur Loss: 1.42818344, Cur Avg Loss: 0.56662472, Log Avg loss: 0.61078843, Global Avg Loss: 3.06302565, Time: 0.0208 Steps: 11600, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000965, Sample Num: 15440, Cur Loss: 0.51545167, Cur Avg Loss: 0.56535881, Log Avg loss: 0.44446487, Global Avg Loss: 3.06077022, Time: 0.0207 Steps: 11610, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000975, Sample Num: 15600, Cur Loss: 0.37818053, Cur Avg Loss: 0.56676714, Log Avg loss: 0.70267083, Global Avg Loss: 3.05874087, Time: 0.0207 Steps: 11620, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000985, Sample Num: 15760, Cur Loss: 0.59769988, Cur Avg Loss: 0.56863114, Log Avg loss: 0.75037074, Global Avg Loss: 3.05675603, Time: 0.0207 Steps: 11630, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000995, Sample Num: 15920, Cur Loss: 0.40342832, Cur Avg Loss: 0.56797980, Log Avg loss: 0.50382345, Global Avg Loss: 3.05456279, Time: 0.0207 Steps: 11640, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001005, Sample Num: 16080, Cur Loss: 0.29224974, Cur Avg Loss: 0.56707487, Log Avg loss: 0.47703375, Global Avg Loss: 3.05235032, Time: 0.0207 Steps: 11650, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001015, Sample Num: 16240, Cur Loss: 0.26881862, Cur Avg Loss: 0.56600855, Log Avg loss: 0.45884322, Global Avg Loss: 3.05012604, Time: 0.0207 Steps: 11660, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001025, Sample Num: 16400, Cur Loss: 0.31044823, Cur Avg Loss: 0.56557321, Log Avg loss: 0.52138635, Global Avg Loss: 3.04795917, Time: 0.0245 Steps: 11670, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001035, Sample Num: 16560, Cur Loss: 0.56784564, Cur Avg Loss: 0.56437956, Log Avg loss: 0.44203079, Global Avg Loss: 3.04572807, Time: 0.0207 Steps: 11680, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001045, Sample Num: 16720, Cur Loss: 0.52458787, Cur Avg Loss: 0.56609805, Log Avg loss: 0.74396151, Global Avg Loss: 3.04375906, Time: 0.0207 Steps: 11690, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001055, Sample Num: 16880, Cur Loss: 0.43776619, Cur Avg Loss: 0.56498779, Log Avg loss: 0.44896537, Global Avg Loss: 3.04154129, Time: 0.0207 Steps: 11700, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001065, Sample Num: 17040, Cur Loss: 0.27397197, Cur Avg Loss: 0.56312456, Log Avg loss: 0.36655394, Global Avg Loss: 3.03925693, Time: 0.0207 Steps: 11710, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001075, Sample Num: 17200, Cur Loss: 0.62634563, Cur Avg Loss: 0.56247162, Log Avg loss: 0.49293317, Global Avg Loss: 3.03708430, Time: 0.0207 Steps: 11720, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001085, Sample Num: 17360, Cur Loss: 0.63655365, Cur Avg Loss: 0.56143327, Log Avg loss: 0.44981134, Global Avg Loss: 3.03487861, Time: 0.0207 Steps: 11730, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001095, Sample Num: 17520, Cur Loss: 0.75586808, Cur Avg Loss: 0.55997620, Log Avg loss: 0.40188348, Global Avg Loss: 3.03263585, Time: 0.0207 Steps: 11740, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001105, Sample Num: 17680, Cur Loss: 0.35483354, Cur Avg Loss: 0.55825626, Log Avg loss: 0.36992363, Global Avg Loss: 3.03036971, Time: 0.0207 Steps: 11750, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001115, Sample Num: 17840, Cur Loss: 1.10269439, Cur Avg Loss: 0.55877983, Log Avg loss: 0.61663366, Global Avg Loss: 3.02831722, Time: 0.0207 Steps: 11760, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001125, Sample Num: 18000, Cur Loss: 1.02917588, Cur Avg Loss: 0.56801390, Log Avg loss: 1.59761343, Global Avg Loss: 3.02710167, Time: 0.0208 Steps: 11770, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001135, Sample Num: 18160, Cur Loss: 0.51333404, Cur Avg Loss: 0.57157134, Log Avg loss: 0.97178254, Global Avg Loss: 3.02535691, Time: 0.0207 Steps: 11780, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001145, Sample Num: 18320, Cur Loss: 0.52887893, Cur Avg Loss: 0.57166488, Log Avg loss: 0.58228232, Global Avg Loss: 3.02328475, Time: 0.0207 Steps: 11790, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001155, Sample Num: 18480, Cur Loss: 1.17572796, Cur Avg Loss: 0.57196185, Log Avg loss: 0.60596449, Global Avg Loss: 3.02123618, Time: 0.0207 Steps: 11800, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001165, Sample Num: 18640, Cur Loss: 0.85853666, Cur Avg Loss: 0.57203245, Log Avg loss: 0.58018695, Global Avg Loss: 3.01916924, Time: 0.0207 Steps: 11810, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001175, Sample Num: 18800, Cur Loss: 0.28879535, Cur Avg Loss: 0.57392021, Log Avg loss: 0.79384359, Global Avg Loss: 3.01728656, Time: 0.0207 Steps: 11820, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001185, Sample Num: 18960, Cur Loss: 0.24487455, Cur Avg Loss: 0.57317128, Log Avg loss: 0.48517265, Global Avg Loss: 3.01514615, Time: 0.0207 Steps: 11830, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001195, Sample Num: 19120, Cur Loss: 0.67178953, Cur Avg Loss: 0.57302950, Log Avg loss: 0.55622859, Global Avg Loss: 3.01306936, Time: 0.0207 Steps: 11840, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001205, Sample Num: 19280, Cur Loss: 2.24971390, Cur Avg Loss: 0.57786550, Log Avg loss: 1.15576733, Global Avg Loss: 3.01150202, Time: 0.0207 Steps: 11850, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001215, Sample Num: 19440, Cur Loss: 0.58050102, Cur Avg Loss: 0.58236084, Log Avg loss: 1.12404967, Global Avg Loss: 3.00991057, Time: 0.0207 Steps: 11860, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001225, Sample Num: 19600, Cur Loss: 1.20169556, Cur Avg Loss: 0.58722796, Log Avg loss: 1.17858258, Global Avg Loss: 3.00836775, Time: 0.0207 Steps: 11870, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001235, Sample Num: 19760, Cur Loss: 0.26231456, Cur Avg Loss: 0.58881801, Log Avg loss: 0.78359944, Global Avg Loss: 3.00649505, Time: 0.0207 Steps: 11880, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001245, Sample Num: 19920, Cur Loss: 0.96885258, Cur Avg Loss: 0.59159137, Log Avg loss: 0.93410085, Global Avg Loss: 3.00475208, Time: 0.0207 Steps: 11890, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001255, Sample Num: 20080, Cur Loss: 1.00201178, Cur Avg Loss: 0.59323627, Log Avg loss: 0.79802606, Global Avg Loss: 3.00289769, Time: 0.0207 Steps: 11900, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001265, Sample Num: 20240, Cur Loss: 0.14990547, Cur Avg Loss: 0.59153776, Log Avg loss: 0.37837522, Global Avg Loss: 3.00069406, Time: 0.0207 Steps: 11910, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001275, Sample Num: 20400, Cur Loss: 0.17300171, Cur Avg Loss: 0.59025873, Log Avg loss: 0.42846174, Global Avg Loss: 2.99853614, Time: 0.0207 Steps: 11920, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001285, Sample Num: 20560, Cur Loss: 0.24846652, Cur Avg Loss: 0.58867179, Log Avg loss: 0.38633692, Global Avg Loss: 2.99634654, Time: 0.0245 Steps: 11930, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001295, Sample Num: 20720, Cur Loss: 2.00515366, Cur Avg Loss: 0.59069327, Log Avg loss: 0.85045337, Global Avg Loss: 2.99454931, Time: 0.0207 Steps: 11940, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001305, Sample Num: 20880, Cur Loss: 0.75423890, Cur Avg Loss: 0.59800587, Log Avg loss: 1.54498705, Global Avg Loss: 2.99333629, Time: 0.0207 Steps: 11950, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001315, Sample Num: 21040, Cur Loss: 0.56102842, Cur Avg Loss: 0.60006202, Log Avg loss: 0.86839010, Global Avg Loss: 2.99155957, Time: 0.0207 Steps: 11960, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001325, Sample Num: 21200, Cur Loss: 0.52564406, Cur Avg Loss: 0.59880201, Log Avg loss: 0.43311064, Global Avg Loss: 2.98942219, Time: 0.0208 Steps: 11970, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001335, Sample Num: 21360, Cur Loss: 0.12396124, Cur Avg Loss: 0.59789374, Log Avg loss: 0.47754826, Global Avg Loss: 2.98732547, Time: 0.0207 Steps: 11980, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001345, Sample Num: 21520, Cur Loss: 1.03275609, Cur Avg Loss: 0.59698555, Log Avg loss: 0.47574188, Global Avg Loss: 2.98523074, Time: 0.0207 Steps: 11990, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001355, Sample Num: 21680, Cur Loss: 0.52766043, Cur Avg Loss: 0.59650852, Log Avg loss: 0.53234847, Global Avg Loss: 2.98318667, Time: 0.0207 Steps: 12000, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001365, Sample Num: 21840, Cur Loss: 0.32729730, Cur Avg Loss: 0.59443974, Log Avg loss: 0.31412004, Global Avg Loss: 2.98096430, Time: 0.0208 Steps: 12010, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001375, Sample Num: 22000, Cur Loss: 0.23524889, Cur Avg Loss: 0.59301115, Log Avg loss: 0.39800815, Global Avg Loss: 2.97881541, Time: 0.0209 Steps: 12020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001385, Sample Num: 22160, Cur Loss: 0.42291632, Cur Avg Loss: 0.59285818, Log Avg loss: 0.57182482, Global Avg Loss: 2.97681459, Time: 0.0208 Steps: 12030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001395, Sample Num: 22320, Cur Loss: 0.16818610, Cur Avg Loss: 0.59228661, Log Avg loss: 0.51312376, Global Avg Loss: 2.97476834, Time: 0.0208 Steps: 12040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001405, Sample Num: 22480, Cur Loss: 0.31928730, Cur Avg Loss: 0.59116585, Log Avg loss: 0.43482076, Global Avg Loss: 2.97266050, Time: 0.0208 Steps: 12050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001415, Sample Num: 22640, Cur Loss: 0.13742498, Cur Avg Loss: 0.59026588, Log Avg loss: 0.46381893, Global Avg Loss: 2.97058020, Time: 0.0209 Steps: 12060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001425, Sample Num: 22800, Cur Loss: 0.22681010, Cur Avg Loss: 0.58930447, Log Avg loss: 0.45326586, Global Avg Loss: 2.96849460, Time: 0.0208 Steps: 12070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001435, Sample Num: 22960, Cur Loss: 0.24493672, Cur Avg Loss: 0.58879159, Log Avg loss: 0.51570582, Global Avg Loss: 2.96646415, Time: 0.0208 Steps: 12080, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001445, Sample Num: 23120, Cur Loss: 0.17169788, Cur Avg Loss: 0.58747208, Log Avg loss: 0.39812273, Global Avg Loss: 2.96433979, Time: 0.0208 Steps: 12090, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001455, Sample Num: 23280, Cur Loss: 1.54896164, Cur Avg Loss: 0.58766318, Log Avg loss: 0.61527718, Global Avg Loss: 2.96239842, Time: 0.0208 Steps: 12100, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001465, Sample Num: 23440, Cur Loss: 0.57322788, Cur Avg Loss: 0.58711560, Log Avg loss: 0.50744294, Global Avg Loss: 2.96037121, Time: 0.0208 Steps: 12110, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001475, Sample Num: 23600, Cur Loss: 0.18853426, Cur Avg Loss: 0.58620283, Log Avg loss: 0.45248162, Global Avg Loss: 2.95830199, Time: 0.0208 Steps: 12120, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001485, Sample Num: 23760, Cur Loss: 0.59184301, Cur Avg Loss: 0.58615019, Log Avg loss: 0.57838518, Global Avg Loss: 2.95633998, Time: 0.0208 Steps: 12130, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001495, Sample Num: 23920, Cur Loss: 0.70422393, Cur Avg Loss: 0.58586635, Log Avg loss: 0.54371730, Global Avg Loss: 2.95435265, Time: 0.0208 Steps: 12140, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001505, Sample Num: 24080, Cur Loss: 0.27887446, Cur Avg Loss: 0.58508612, Log Avg loss: 0.46844034, Global Avg Loss: 2.95230663, Time: 0.0209 Steps: 12150, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001515, Sample Num: 24240, Cur Loss: 0.16929147, Cur Avg Loss: 0.58445830, Log Avg loss: 0.48997177, Global Avg Loss: 2.95028168, Time: 0.0209 Steps: 12160, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001525, Sample Num: 24400, Cur Loss: 0.25689793, Cur Avg Loss: 0.58304392, Log Avg loss: 0.36876528, Global Avg Loss: 2.94816047, Time: 0.0209 Steps: 12170, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001535, Sample Num: 24560, Cur Loss: 0.51328403, Cur Avg Loss: 0.58181200, Log Avg loss: 0.39394387, Global Avg Loss: 2.94606341, Time: 0.0208 Steps: 12180, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001545, Sample Num: 24720, Cur Loss: 1.62558699, Cur Avg Loss: 0.58441401, Log Avg loss: 0.98382324, Global Avg Loss: 2.94445370, Time: 0.0209 Steps: 12190, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001555, Sample Num: 24880, Cur Loss: 0.60501969, Cur Avg Loss: 0.58511431, Log Avg loss: 0.69331102, Global Avg Loss: 2.94260850, Time: 0.0208 Steps: 12200, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001565, Sample Num: 25040, Cur Loss: 0.19376202, Cur Avg Loss: 0.58382908, Log Avg loss: 0.38397546, Global Avg Loss: 2.94051298, Time: 0.0209 Steps: 12210, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001575, Sample Num: 25200, Cur Loss: 0.92747414, Cur Avg Loss: 0.58381143, Log Avg loss: 0.58104966, Global Avg Loss: 2.93858216, Time: 0.0209 Steps: 12220, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001585, Sample Num: 25360, Cur Loss: 0.37413371, Cur Avg Loss: 0.58446991, Log Avg loss: 0.68817948, Global Avg Loss: 2.93674209, Time: 0.0208 Steps: 12230, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001595, Sample Num: 25520, Cur Loss: 0.27741906, Cur Avg Loss: 0.58341527, Log Avg loss: 0.41625518, Global Avg Loss: 2.93468287, Time: 0.0208 Steps: 12240, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001605, Sample Num: 25680, Cur Loss: 0.72836626, Cur Avg Loss: 0.58286935, Log Avg loss: 0.49579464, Global Avg Loss: 2.93269194, Time: 0.0208 Steps: 12250, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001615, Sample Num: 25840, Cur Loss: 2.10676289, Cur Avg Loss: 0.58363264, Log Avg loss: 0.70614141, Global Avg Loss: 2.93087583, Time: 0.0208 Steps: 12260, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001625, Sample Num: 26000, Cur Loss: 0.41668668, Cur Avg Loss: 0.58439903, Log Avg loss: 0.70817118, Global Avg Loss: 2.92906433, Time: 0.0209 Steps: 12270, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001635, Sample Num: 26160, Cur Loss: 0.88256282, Cur Avg Loss: 0.58363791, Log Avg loss: 0.45995651, Global Avg Loss: 2.92705366, Time: 0.0208 Steps: 12280, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001645, Sample Num: 26320, Cur Loss: 0.34050792, Cur Avg Loss: 0.58210047, Log Avg loss: 0.33072764, Global Avg Loss: 2.92494111, Time: 0.0209 Steps: 12290, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001655, Sample Num: 26480, Cur Loss: 0.64181960, Cur Avg Loss: 0.58122257, Log Avg loss: 0.43680933, Global Avg Loss: 2.92291824, Time: 0.0209 Steps: 12300, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001665, Sample Num: 26640, Cur Loss: 0.64883775, Cur Avg Loss: 0.58114179, Log Avg loss: 0.56777136, Global Avg Loss: 2.92100504, Time: 0.0208 Steps: 12310, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001675, Sample Num: 26800, Cur Loss: 0.30381089, Cur Avg Loss: 0.58068482, Log Avg loss: 0.50459939, Global Avg Loss: 2.91904367, Time: 0.0209 Steps: 12320, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001685, Sample Num: 26960, Cur Loss: 0.26780564, Cur Avg Loss: 0.57990133, Log Avg loss: 0.44866690, Global Avg Loss: 2.91704012, Time: 0.0208 Steps: 12330, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001695, Sample Num: 27120, Cur Loss: 0.29388922, Cur Avg Loss: 0.57982045, Log Avg loss: 0.56619322, Global Avg Loss: 2.91513506, Time: 0.0209 Steps: 12340, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001705, Sample Num: 27280, Cur Loss: 0.44620132, Cur Avg Loss: 0.57897910, Log Avg loss: 0.43636934, Global Avg Loss: 2.91312796, Time: 0.0209 Steps: 12350, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001715, Sample Num: 27440, Cur Loss: 0.82820857, Cur Avg Loss: 0.57872080, Log Avg loss: 0.53468045, Global Avg Loss: 2.91120365, Time: 0.0208 Steps: 12360, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001725, Sample Num: 27600, Cur Loss: 1.07253122, Cur Avg Loss: 0.58153370, Log Avg loss: 1.06394569, Global Avg Loss: 2.90971031, Time: 0.0209 Steps: 12370, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001735, Sample Num: 27760, Cur Loss: 1.14234638, Cur Avg Loss: 0.58317554, Log Avg loss: 0.86639452, Global Avg Loss: 2.90805982, Time: 0.0209 Steps: 12380, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001745, Sample Num: 27920, Cur Loss: 0.86847568, Cur Avg Loss: 0.58694837, Log Avg loss: 1.24153329, Global Avg Loss: 2.90671476, Time: 0.0209 Steps: 12390, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001755, Sample Num: 28080, Cur Loss: 0.13523862, Cur Avg Loss: 0.58729140, Log Avg loss: 0.64715114, Global Avg Loss: 2.90489253, Time: 0.0208 Steps: 12400, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001765, Sample Num: 28240, Cur Loss: 0.91886830, Cur Avg Loss: 0.59014834, Log Avg loss: 1.09154131, Global Avg Loss: 2.90343133, Time: 0.0209 Steps: 12410, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001775, Sample Num: 28400, Cur Loss: 0.27039808, Cur Avg Loss: 0.58922257, Log Avg loss: 0.42582389, Global Avg Loss: 2.90143648, Time: 0.0209 Steps: 12420, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001785, Sample Num: 28560, Cur Loss: 0.22992617, Cur Avg Loss: 0.58759878, Log Avg loss: 0.29937533, Global Avg Loss: 2.89934310, Time: 0.0208 Steps: 12430, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001795, Sample Num: 28720, Cur Loss: 0.27870905, Cur Avg Loss: 0.58703796, Log Avg loss: 0.48693163, Global Avg Loss: 2.89740387, Time: 0.0246 Steps: 12440, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001805, Sample Num: 28880, Cur Loss: 0.50026411, Cur Avg Loss: 0.58562832, Log Avg loss: 0.33259810, Global Avg Loss: 2.89534378, Time: 0.0209 Steps: 12450, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001815, Sample Num: 29040, Cur Loss: 0.37439477, Cur Avg Loss: 0.58569369, Log Avg loss: 0.59749394, Global Avg Loss: 2.89349960, Time: 0.0209 Steps: 12460, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001825, Sample Num: 29200, Cur Loss: 0.54420328, Cur Avg Loss: 0.58523798, Log Avg loss: 0.50252665, Global Avg Loss: 2.89158222, Time: 0.0209 Steps: 12470, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001835, Sample Num: 29360, Cur Loss: 0.25213024, Cur Avg Loss: 0.58531291, Log Avg loss: 0.59898616, Global Avg Loss: 2.88974520, Time: 0.0209 Steps: 12480, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001845, Sample Num: 29520, Cur Loss: 0.67922455, Cur Avg Loss: 0.58520199, Log Avg loss: 0.56484947, Global Avg Loss: 2.88788380, Time: 0.0209 Steps: 12490, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001855, Sample Num: 29680, Cur Loss: 0.55739474, Cur Avg Loss: 0.58475988, Log Avg loss: 0.50318910, Global Avg Loss: 2.88597604, Time: 0.0209 Steps: 12500, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001865, Sample Num: 29840, Cur Loss: 0.33863521, Cur Avg Loss: 0.58435001, Log Avg loss: 0.50832012, Global Avg Loss: 2.88407544, Time: 0.0209 Steps: 12510, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001875, Sample Num: 30000, Cur Loss: 0.21404123, Cur Avg Loss: 0.58296213, Log Avg loss: 0.32412295, Global Avg Loss: 2.88203075, Time: 0.0209 Steps: 12520, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001885, Sample Num: 30160, Cur Loss: 0.74763358, Cur Avg Loss: 0.58206689, Log Avg loss: 0.41420838, Global Avg Loss: 2.88006122, Time: 0.0209 Steps: 12530, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001895, Sample Num: 30320, Cur Loss: 0.67679238, Cur Avg Loss: 0.58102978, Log Avg loss: 0.38553434, Global Avg Loss: 2.87807196, Time: 0.0209 Steps: 12540, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001905, Sample Num: 30480, Cur Loss: 0.25791246, Cur Avg Loss: 0.58099682, Log Avg loss: 0.57475076, Global Avg Loss: 2.87623664, Time: 0.0209 Steps: 12550, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001915, Sample Num: 30640, Cur Loss: 0.24994190, Cur Avg Loss: 0.58187999, Log Avg loss: 0.75012428, Global Avg Loss: 2.87454388, Time: 0.0209 Steps: 12560, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001925, Sample Num: 30800, Cur Loss: 0.78741109, Cur Avg Loss: 0.58546856, Log Avg loss: 1.27268026, Global Avg Loss: 2.87326953, Time: 0.0209 Steps: 12570, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001935, Sample Num: 30960, Cur Loss: 0.36334047, Cur Avg Loss: 0.58526008, Log Avg loss: 0.54512780, Global Avg Loss: 2.87141886, Time: 0.0209 Steps: 12580, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001945, Sample Num: 31120, Cur Loss: 0.58443463, Cur Avg Loss: 0.58519243, Log Avg loss: 0.57210190, Global Avg Loss: 2.86959255, Time: 0.0209 Steps: 12590, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001955, Sample Num: 31280, Cur Loss: 0.46799499, Cur Avg Loss: 0.58503436, Log Avg loss: 0.55428913, Global Avg Loss: 2.86775501, Time: 0.0210 Steps: 12600, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001965, Sample Num: 31440, Cur Loss: 0.41360062, Cur Avg Loss: 0.58459656, Log Avg loss: 0.49900775, Global Avg Loss: 2.86587654, Time: 0.0209 Steps: 12610, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001975, Sample Num: 31600, Cur Loss: 0.63095403, Cur Avg Loss: 0.58466944, Log Avg loss: 0.59899061, Global Avg Loss: 2.86408028, Time: 0.0209 Steps: 12620, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001985, Sample Num: 31760, Cur Loss: 0.49016190, Cur Avg Loss: 0.58433481, Log Avg loss: 0.51824457, Global Avg Loss: 2.86222293, Time: 0.0209 Steps: 12630, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001995, Sample Num: 31920, Cur Loss: 0.92446697, Cur Avg Loss: 0.58717967, Log Avg loss: 1.15188510, Global Avg Loss: 2.86086981, Time: 0.0209 Steps: 12640, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002005, Sample Num: 32080, Cur Loss: 0.93572265, Cur Avg Loss: 0.58726241, Log Avg loss: 0.60376919, Global Avg Loss: 2.85908554, Time: 0.0209 Steps: 12650, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002015, Sample Num: 32240, Cur Loss: 0.45265776, Cur Avg Loss: 0.58624590, Log Avg loss: 0.38243417, Global Avg Loss: 2.85712926, Time: 0.0209 Steps: 12660, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002025, Sample Num: 32400, Cur Loss: 0.37451261, Cur Avg Loss: 0.58626925, Log Avg loss: 0.59097549, Global Avg Loss: 2.85534066, Time: 0.0209 Steps: 12670, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002035, Sample Num: 32560, Cur Loss: 0.45942318, Cur Avg Loss: 0.58595286, Log Avg loss: 0.52188402, Global Avg Loss: 2.85350040, Time: 0.0209 Steps: 12680, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002045, Sample Num: 32720, Cur Loss: 0.15515697, Cur Avg Loss: 0.58604848, Log Avg loss: 0.60550684, Global Avg Loss: 2.85172893, Time: 0.0209 Steps: 12690, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002055, Sample Num: 32880, Cur Loss: 0.37755534, Cur Avg Loss: 0.58693244, Log Avg loss: 0.76770095, Global Avg Loss: 2.85008796, Time: 0.0209 Steps: 12700, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002065, Sample Num: 33040, Cur Loss: 0.16363302, Cur Avg Loss: 0.58750422, Log Avg loss: 0.70500658, Global Avg Loss: 2.84840025, Time: 0.0208 Steps: 12710, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002075, Sample Num: 33200, Cur Loss: 0.14120896, Cur Avg Loss: 0.58646199, Log Avg loss: 0.37124091, Global Avg Loss: 2.84645280, Time: 0.0208 Steps: 12720, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002085, Sample Num: 33360, Cur Loss: 2.13848948, Cur Avg Loss: 0.58701947, Log Avg loss: 0.70269615, Global Avg Loss: 2.84476878, Time: 0.0207 Steps: 12730, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002095, Sample Num: 33520, Cur Loss: 0.26352367, Cur Avg Loss: 0.58673438, Log Avg loss: 0.52729430, Global Avg Loss: 2.84294972, Time: 0.0209 Steps: 12740, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002105, Sample Num: 33680, Cur Loss: 0.63384932, Cur Avg Loss: 0.58611481, Log Avg loss: 0.45631316, Global Avg Loss: 2.84107785, Time: 0.0209 Steps: 12750, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002115, Sample Num: 33840, Cur Loss: 0.30858833, Cur Avg Loss: 0.58575562, Log Avg loss: 0.51014635, Global Avg Loss: 2.83925110, Time: 0.0209 Steps: 12760, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002125, Sample Num: 34000, Cur Loss: 0.51115292, Cur Avg Loss: 0.58540044, Log Avg loss: 0.51028044, Global Avg Loss: 2.83742732, Time: 0.0211 Steps: 12770, Updated lr: 0.000089 ***** Running evaluation checkpoint-12774 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-12774 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.584524, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.429842, "eval_total_loss": 302.178727, "eval_mae": 0.527648, "eval_mse": 0.429835, "eval_r2": 0.726769, "eval_sp_statistic": 0.840759, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.878163, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.507512, "test_total_loss": 254.770956, "test_mae": 0.425282, "test_mse": 0.507698, "test_r2": 0.672327, "test_sp_statistic": 0.838055, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.900796, "test_ps_pvalue": 0.0, "lr": 8.883451872925558e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.8366793293575316, "train_cur_epoch_loss": 1245.7707983590662, "train_cur_epoch_avg_loss": 0.5851436347388757, "train_cur_epoch_time": 44.58452391624451, "train_cur_epoch_avg_time": 0.020941533074797795, "epoch": 6, "step": 12774} ################################################## Training, Epoch: 0007, Batch: 000006, Sample Num: 96, Cur Loss: 0.39484558, Cur Avg Loss: 0.32314770, Log Avg loss: 0.37337505, Global Avg Loss: 2.83549927, Time: 0.0248 Steps: 12780, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000016, Sample Num: 256, Cur Loss: 0.34104130, Cur Avg Loss: 0.32929771, Log Avg loss: 0.33298772, Global Avg Loss: 2.83354265, Time: 0.0211 Steps: 12790, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000026, Sample Num: 416, Cur Loss: 0.29197228, Cur Avg Loss: 0.64629858, Log Avg loss: 1.15349997, Global Avg Loss: 2.83223012, Time: 0.0211 Steps: 12800, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000036, Sample Num: 576, Cur Loss: 0.27590305, Cur Avg Loss: 0.54705786, Log Avg loss: 0.28903200, Global Avg Loss: 2.83024480, Time: 0.0210 Steps: 12810, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000046, Sample Num: 736, Cur Loss: 0.27121437, Cur Avg Loss: 0.57254952, Log Avg loss: 0.66431948, Global Avg Loss: 2.82855531, Time: 0.0210 Steps: 12820, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000056, Sample Num: 896, Cur Loss: 0.45577773, Cur Avg Loss: 0.53878979, Log Avg loss: 0.38349501, Global Avg Loss: 2.82664957, Time: 0.0210 Steps: 12830, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000066, Sample Num: 1056, Cur Loss: 0.33795556, Cur Avg Loss: 0.50790892, Log Avg loss: 0.33497609, Global Avg Loss: 2.82470901, Time: 0.0210 Steps: 12840, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000076, Sample Num: 1216, Cur Loss: 0.44482151, Cur Avg Loss: 0.49090342, Log Avg loss: 0.37866708, Global Avg Loss: 2.82280548, Time: 0.0210 Steps: 12850, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000086, Sample Num: 1376, Cur Loss: 0.23900798, Cur Avg Loss: 0.47791383, Log Avg loss: 0.37919298, Global Avg Loss: 2.82090531, Time: 0.0210 Steps: 12860, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000096, Sample Num: 1536, Cur Loss: 0.14272863, Cur Avg Loss: 0.47078476, Log Avg loss: 0.40947475, Global Avg Loss: 2.81903163, Time: 0.0210 Steps: 12870, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000106, Sample Num: 1696, Cur Loss: 0.26570728, Cur Avg Loss: 0.46623043, Log Avg loss: 0.42250887, Global Avg Loss: 2.81717098, Time: 0.0210 Steps: 12880, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000116, Sample Num: 1856, Cur Loss: 0.23371533, Cur Avg Loss: 0.47048649, Log Avg loss: 0.51560077, Global Avg Loss: 2.81538543, Time: 0.0211 Steps: 12890, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000126, Sample Num: 2016, Cur Loss: 0.69302702, Cur Avg Loss: 0.48930647, Log Avg loss: 0.70761815, Global Avg Loss: 2.81375150, Time: 0.0210 Steps: 12900, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000136, Sample Num: 2176, Cur Loss: 0.37267801, Cur Avg Loss: 0.49466153, Log Avg loss: 0.56213531, Global Avg Loss: 2.81200741, Time: 0.0210 Steps: 12910, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000146, Sample Num: 2336, Cur Loss: 0.69945139, Cur Avg Loss: 0.51912780, Log Avg loss: 0.85186911, Global Avg Loss: 2.81049028, Time: 0.0210 Steps: 12920, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000156, Sample Num: 2496, Cur Loss: 0.16740762, Cur Avg Loss: 0.50462423, Log Avg loss: 0.29287211, Global Avg Loss: 2.80854317, Time: 0.0210 Steps: 12930, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000166, Sample Num: 2656, Cur Loss: 0.19324468, Cur Avg Loss: 0.49547475, Log Avg loss: 0.35274290, Global Avg Loss: 2.80664533, Time: 0.0210 Steps: 12940, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000176, Sample Num: 2816, Cur Loss: 0.53043365, Cur Avg Loss: 0.49192915, Log Avg loss: 0.43307219, Global Avg Loss: 2.80481245, Time: 0.0210 Steps: 12950, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000186, Sample Num: 2976, Cur Loss: 0.77360392, Cur Avg Loss: 0.49719170, Log Avg loss: 0.58981253, Global Avg Loss: 2.80310335, Time: 0.0210 Steps: 12960, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000196, Sample Num: 3136, Cur Loss: 0.52000457, Cur Avg Loss: 0.50068759, Log Avg loss: 0.56571118, Global Avg Loss: 2.80137830, Time: 0.0210 Steps: 12970, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000206, Sample Num: 3296, Cur Loss: 0.25347805, Cur Avg Loss: 0.50163028, Log Avg loss: 0.52010692, Global Avg Loss: 2.79962077, Time: 0.0210 Steps: 12980, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000216, Sample Num: 3456, Cur Loss: 0.85929877, Cur Avg Loss: 0.49744958, Log Avg loss: 0.41132712, Global Avg Loss: 2.79778221, Time: 0.0210 Steps: 12990, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000226, Sample Num: 3616, Cur Loss: 0.25503147, Cur Avg Loss: 0.49773556, Log Avg loss: 0.50391292, Global Avg Loss: 2.79601769, Time: 0.0210 Steps: 13000, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000236, Sample Num: 3776, Cur Loss: 0.22029230, Cur Avg Loss: 0.49241551, Log Avg loss: 0.37218219, Global Avg Loss: 2.79415464, Time: 0.0210 Steps: 13010, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000246, Sample Num: 3936, Cur Loss: 0.79455328, Cur Avg Loss: 0.49668921, Log Avg loss: 0.59754866, Global Avg Loss: 2.79246753, Time: 0.0212 Steps: 13020, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000256, Sample Num: 4096, Cur Loss: 0.70338577, Cur Avg Loss: 0.49511473, Log Avg loss: 0.45638245, Global Avg Loss: 2.79067468, Time: 0.0254 Steps: 13030, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000266, Sample Num: 4256, Cur Loss: 0.42435294, Cur Avg Loss: 0.49313635, Log Avg loss: 0.44248978, Global Avg Loss: 2.78887393, Time: 0.0211 Steps: 13040, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000276, Sample Num: 4416, Cur Loss: 0.22464389, Cur Avg Loss: 0.48728459, Log Avg loss: 0.33162777, Global Avg Loss: 2.78699098, Time: 0.0211 Steps: 13050, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000286, Sample Num: 4576, Cur Loss: 0.93250048, Cur Avg Loss: 0.49179676, Log Avg loss: 0.61633278, Global Avg Loss: 2.78532891, Time: 0.0210 Steps: 13060, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000296, Sample Num: 4736, Cur Loss: 0.48630714, Cur Avg Loss: 0.48791339, Log Avg loss: 0.37684895, Global Avg Loss: 2.78348616, Time: 0.0211 Steps: 13070, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000306, Sample Num: 4896, Cur Loss: 0.41649634, Cur Avg Loss: 0.48356223, Log Avg loss: 0.35476793, Global Avg Loss: 2.78162934, Time: 0.0211 Steps: 13080, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000316, Sample Num: 5056, Cur Loss: 0.48072574, Cur Avg Loss: 0.48303602, Log Avg loss: 0.46693381, Global Avg Loss: 2.77986105, Time: 0.0210 Steps: 13090, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000326, Sample Num: 5216, Cur Loss: 0.31250650, Cur Avg Loss: 0.48497602, Log Avg loss: 0.54628000, Global Avg Loss: 2.77815603, Time: 0.0211 Steps: 13100, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000336, Sample Num: 5376, Cur Loss: 0.61311823, Cur Avg Loss: 0.48620049, Log Avg loss: 0.52611835, Global Avg Loss: 2.77643822, Time: 0.0211 Steps: 13110, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000346, Sample Num: 5536, Cur Loss: 0.64359051, Cur Avg Loss: 0.49217798, Log Avg loss: 0.69302159, Global Avg Loss: 2.77485025, Time: 0.0211 Steps: 13120, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000356, Sample Num: 5696, Cur Loss: 0.35087594, Cur Avg Loss: 0.49367032, Log Avg loss: 0.54530526, Global Avg Loss: 2.77315220, Time: 0.0211 Steps: 13130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000366, Sample Num: 5856, Cur Loss: 0.55549431, Cur Avg Loss: 0.49418850, Log Avg loss: 0.51263578, Global Avg Loss: 2.77143187, Time: 0.0211 Steps: 13140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000376, Sample Num: 6016, Cur Loss: 0.46606642, Cur Avg Loss: 0.49576100, Log Avg loss: 0.55331464, Global Avg Loss: 2.76974509, Time: 0.0211 Steps: 13150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000386, Sample Num: 6176, Cur Loss: 0.28427178, Cur Avg Loss: 0.49350832, Log Avg loss: 0.40880753, Global Avg Loss: 2.76795106, Time: 0.0210 Steps: 13160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000396, Sample Num: 6336, Cur Loss: 0.48007894, Cur Avg Loss: 0.49596428, Log Avg loss: 0.59076427, Global Avg Loss: 2.76629792, Time: 0.0211 Steps: 13170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000406, Sample Num: 6496, Cur Loss: 0.31303418, Cur Avg Loss: 0.49097002, Log Avg loss: 0.29319729, Global Avg Loss: 2.76442152, Time: 0.0209 Steps: 13180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000416, Sample Num: 6656, Cur Loss: 0.21919456, Cur Avg Loss: 0.48971696, Log Avg loss: 0.43884266, Global Avg Loss: 2.76265838, Time: 0.0209 Steps: 13190, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000426, Sample Num: 6816, Cur Loss: 0.63409257, Cur Avg Loss: 0.49389267, Log Avg loss: 0.66760243, Global Avg Loss: 2.76107121, Time: 0.0210 Steps: 13200, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000436, Sample Num: 6976, Cur Loss: 0.56694436, Cur Avg Loss: 0.49445624, Log Avg loss: 0.51846401, Global Avg Loss: 2.75937356, Time: 0.0210 Steps: 13210, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000446, Sample Num: 7136, Cur Loss: 0.55422407, Cur Avg Loss: 0.49313215, Log Avg loss: 0.43540195, Global Avg Loss: 2.75761563, Time: 0.0210 Steps: 13220, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000456, Sample Num: 7296, Cur Loss: 0.21784231, Cur Avg Loss: 0.49201001, Log Avg loss: 0.44196255, Global Avg Loss: 2.75586533, Time: 0.0210 Steps: 13230, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000466, Sample Num: 7456, Cur Loss: 0.21518698, Cur Avg Loss: 0.49419479, Log Avg loss: 0.59382076, Global Avg Loss: 2.75423237, Time: 0.0210 Steps: 13240, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000476, Sample Num: 7616, Cur Loss: 0.34141135, Cur Avg Loss: 0.49206216, Log Avg loss: 0.39268178, Global Avg Loss: 2.75245006, Time: 0.0209 Steps: 13250, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000486, Sample Num: 7776, Cur Loss: 1.24936700, Cur Avg Loss: 0.49518765, Log Avg loss: 0.64396109, Global Avg Loss: 2.75085995, Time: 0.0210 Steps: 13260, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000496, Sample Num: 7936, Cur Loss: 0.24908409, Cur Avg Loss: 0.49696841, Log Avg loss: 0.58351301, Global Avg Loss: 2.74922668, Time: 0.0209 Steps: 13270, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000506, Sample Num: 8096, Cur Loss: 0.27313921, Cur Avg Loss: 0.49218401, Log Avg loss: 0.25487786, Global Avg Loss: 2.74734841, Time: 0.0211 Steps: 13280, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000516, Sample Num: 8256, Cur Loss: 0.77496862, Cur Avg Loss: 0.49386743, Log Avg loss: 0.57904825, Global Avg Loss: 2.74571688, Time: 0.0247 Steps: 13290, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000526, Sample Num: 8416, Cur Loss: 0.72193289, Cur Avg Loss: 0.49542771, Log Avg loss: 0.57593817, Global Avg Loss: 2.74408547, Time: 0.0210 Steps: 13300, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000536, Sample Num: 8576, Cur Loss: 0.45434776, Cur Avg Loss: 0.49980508, Log Avg loss: 0.73005489, Global Avg Loss: 2.74257230, Time: 0.0210 Steps: 13310, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000546, Sample Num: 8736, Cur Loss: 1.40893328, Cur Avg Loss: 0.50116413, Log Avg loss: 0.57400954, Global Avg Loss: 2.74094425, Time: 0.0210 Steps: 13320, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000556, Sample Num: 8896, Cur Loss: 0.17871255, Cur Avg Loss: 0.49914395, Log Avg loss: 0.38884166, Global Avg Loss: 2.73917973, Time: 0.0210 Steps: 13330, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000566, Sample Num: 9056, Cur Loss: 0.38089716, Cur Avg Loss: 0.49612403, Log Avg loss: 0.32821663, Global Avg Loss: 2.73737241, Time: 0.0210 Steps: 13340, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000576, Sample Num: 9216, Cur Loss: 0.27495232, Cur Avg Loss: 0.49438366, Log Avg loss: 0.39587871, Global Avg Loss: 2.73561848, Time: 0.0210 Steps: 13350, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000586, Sample Num: 9376, Cur Loss: 0.92968071, Cur Avg Loss: 0.49989422, Log Avg loss: 0.81730276, Global Avg Loss: 2.73418262, Time: 0.0210 Steps: 13360, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000596, Sample Num: 9536, Cur Loss: 0.41769150, Cur Avg Loss: 0.49787940, Log Avg loss: 0.37981061, Global Avg Loss: 2.73242168, Time: 0.0210 Steps: 13370, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000606, Sample Num: 9696, Cur Loss: 0.22818677, Cur Avg Loss: 0.49536373, Log Avg loss: 0.34542963, Global Avg Loss: 2.73063768, Time: 0.0209 Steps: 13380, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000616, Sample Num: 9856, Cur Loss: 1.28235102, Cur Avg Loss: 0.49694635, Log Avg loss: 0.59285371, Global Avg Loss: 2.72904113, Time: 0.0210 Steps: 13390, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000626, Sample Num: 10016, Cur Loss: 0.21607584, Cur Avg Loss: 0.49644871, Log Avg loss: 0.46579356, Global Avg Loss: 2.72735214, Time: 0.0210 Steps: 13400, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000636, Sample Num: 10176, Cur Loss: 0.69483948, Cur Avg Loss: 0.49738111, Log Avg loss: 0.55574964, Global Avg Loss: 2.72573275, Time: 0.0209 Steps: 13410, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000646, Sample Num: 10336, Cur Loss: 0.19256178, Cur Avg Loss: 0.49734403, Log Avg loss: 0.49498539, Global Avg Loss: 2.72407049, Time: 0.0210 Steps: 13420, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000656, Sample Num: 10496, Cur Loss: 0.52314615, Cur Avg Loss: 0.49847295, Log Avg loss: 0.57140177, Global Avg Loss: 2.72246761, Time: 0.0210 Steps: 13430, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000666, Sample Num: 10656, Cur Loss: 0.67659819, Cur Avg Loss: 0.49770062, Log Avg loss: 0.44703579, Global Avg Loss: 2.72077458, Time: 0.0210 Steps: 13440, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000676, Sample Num: 10816, Cur Loss: 0.25564849, Cur Avg Loss: 0.49827735, Log Avg loss: 0.53668733, Global Avg Loss: 2.71915072, Time: 0.0210 Steps: 13450, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000686, Sample Num: 10976, Cur Loss: 0.52241850, Cur Avg Loss: 0.49629330, Log Avg loss: 0.36217130, Global Avg Loss: 2.71739963, Time: 0.0210 Steps: 13460, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000696, Sample Num: 11136, Cur Loss: 0.13690351, Cur Avg Loss: 0.49273891, Log Avg loss: 0.24890796, Global Avg Loss: 2.71556704, Time: 0.0210 Steps: 13470, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000706, Sample Num: 11296, Cur Loss: 0.52308732, Cur Avg Loss: 0.49555355, Log Avg loss: 0.69145211, Global Avg Loss: 2.71406547, Time: 0.0209 Steps: 13480, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000716, Sample Num: 11456, Cur Loss: 0.25218451, Cur Avg Loss: 0.49668017, Log Avg loss: 0.57622020, Global Avg Loss: 2.71248071, Time: 0.0210 Steps: 13490, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000726, Sample Num: 11616, Cur Loss: 0.90536690, Cur Avg Loss: 0.49771982, Log Avg loss: 0.57215827, Global Avg Loss: 2.71089528, Time: 0.0210 Steps: 13500, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000736, Sample Num: 11776, Cur Loss: 0.39695024, Cur Avg Loss: 0.49990071, Log Avg loss: 0.65823326, Global Avg Loss: 2.70937592, Time: 0.0210 Steps: 13510, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000746, Sample Num: 11936, Cur Loss: 1.12004519, Cur Avg Loss: 0.50865934, Log Avg loss: 1.15329441, Global Avg Loss: 2.70822497, Time: 0.0210 Steps: 13520, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000756, Sample Num: 12096, Cur Loss: 0.90622360, Cur Avg Loss: 0.52008467, Log Avg loss: 1.37241497, Global Avg Loss: 2.70723768, Time: 0.0211 Steps: 13530, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000766, Sample Num: 12256, Cur Loss: 0.52320516, Cur Avg Loss: 0.52402966, Log Avg loss: 0.82227043, Global Avg Loss: 2.70584553, Time: 0.0210 Steps: 13540, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000776, Sample Num: 12416, Cur Loss: 0.28532371, Cur Avg Loss: 0.52249679, Log Avg loss: 0.40507872, Global Avg Loss: 2.70414755, Time: 0.0211 Steps: 13550, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000786, Sample Num: 12576, Cur Loss: 0.27302107, Cur Avg Loss: 0.51987892, Log Avg loss: 0.31673215, Global Avg Loss: 2.70238692, Time: 0.0209 Steps: 13560, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000796, Sample Num: 12736, Cur Loss: 0.44667351, Cur Avg Loss: 0.51821692, Log Avg loss: 0.38758409, Global Avg Loss: 2.70068109, Time: 0.0209 Steps: 13570, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000806, Sample Num: 12896, Cur Loss: 0.96087360, Cur Avg Loss: 0.52177348, Log Avg loss: 0.80487570, Global Avg Loss: 2.69928506, Time: 0.0209 Steps: 13580, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000816, Sample Num: 13056, Cur Loss: 0.37728727, Cur Avg Loss: 0.52334278, Log Avg loss: 0.64982818, Global Avg Loss: 2.69777700, Time: 0.0209 Steps: 13590, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000826, Sample Num: 13216, Cur Loss: 0.36632904, Cur Avg Loss: 0.52354646, Log Avg loss: 0.54016683, Global Avg Loss: 2.69619052, Time: 0.0210 Steps: 13600, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000836, Sample Num: 13376, Cur Loss: 1.15364838, Cur Avg Loss: 0.52962665, Log Avg loss: 1.03185041, Global Avg Loss: 2.69496764, Time: 0.0209 Steps: 13610, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000846, Sample Num: 13536, Cur Loss: 0.38986251, Cur Avg Loss: 0.53079619, Log Avg loss: 0.62856992, Global Avg Loss: 2.69345046, Time: 0.0209 Steps: 13620, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000856, Sample Num: 13696, Cur Loss: 0.43695319, Cur Avg Loss: 0.52869809, Log Avg loss: 0.35119826, Global Avg Loss: 2.69173201, Time: 0.0210 Steps: 13630, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000866, Sample Num: 13856, Cur Loss: 0.20867828, Cur Avg Loss: 0.52727859, Log Avg loss: 0.40576987, Global Avg Loss: 2.69005609, Time: 0.0210 Steps: 13640, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000876, Sample Num: 14016, Cur Loss: 0.16709371, Cur Avg Loss: 0.52647131, Log Avg loss: 0.45656108, Global Avg Loss: 2.68841983, Time: 0.0209 Steps: 13650, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000886, Sample Num: 14176, Cur Loss: 0.46584004, Cur Avg Loss: 0.52411580, Log Avg loss: 0.31777290, Global Avg Loss: 2.68668436, Time: 0.0210 Steps: 13660, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000896, Sample Num: 14336, Cur Loss: 0.19119301, Cur Avg Loss: 0.52435896, Log Avg loss: 0.54590289, Global Avg Loss: 2.68511832, Time: 0.0210 Steps: 13670, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000906, Sample Num: 14496, Cur Loss: 0.21531315, Cur Avg Loss: 0.52475005, Log Avg loss: 0.55979205, Global Avg Loss: 2.68356472, Time: 0.0209 Steps: 13680, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000916, Sample Num: 14656, Cur Loss: 0.17324103, Cur Avg Loss: 0.52346804, Log Avg loss: 0.40731757, Global Avg Loss: 2.68190201, Time: 0.0209 Steps: 13690, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000926, Sample Num: 14816, Cur Loss: 0.40239879, Cur Avg Loss: 0.52205049, Log Avg loss: 0.39220272, Global Avg Loss: 2.68023069, Time: 0.0209 Steps: 13700, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000936, Sample Num: 14976, Cur Loss: 0.28475976, Cur Avg Loss: 0.52157300, Log Avg loss: 0.47735777, Global Avg Loss: 2.67862393, Time: 0.0209 Steps: 13710, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000946, Sample Num: 15136, Cur Loss: 0.51603287, Cur Avg Loss: 0.52179428, Log Avg loss: 0.54250631, Global Avg Loss: 2.67706699, Time: 0.0209 Steps: 13720, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000956, Sample Num: 15296, Cur Loss: 0.23960486, Cur Avg Loss: 0.52096920, Log Avg loss: 0.44291595, Global Avg Loss: 2.67543979, Time: 0.0210 Steps: 13730, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000966, Sample Num: 15456, Cur Loss: 0.23265219, Cur Avg Loss: 0.52131414, Log Avg loss: 0.55429024, Global Avg Loss: 2.67389601, Time: 0.0209 Steps: 13740, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000976, Sample Num: 15616, Cur Loss: 0.74860787, Cur Avg Loss: 0.52104167, Log Avg loss: 0.49472157, Global Avg Loss: 2.67231116, Time: 0.0209 Steps: 13750, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000986, Sample Num: 15776, Cur Loss: 0.64557576, Cur Avg Loss: 0.52018207, Log Avg loss: 0.43628516, Global Avg Loss: 2.67068614, Time: 0.0209 Steps: 13760, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000996, Sample Num: 15936, Cur Loss: 0.67695665, Cur Avg Loss: 0.51949606, Log Avg loss: 0.45185586, Global Avg Loss: 2.66907479, Time: 0.0209 Steps: 13770, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001006, Sample Num: 16096, Cur Loss: 0.47965229, Cur Avg Loss: 0.51815992, Log Avg loss: 0.38508027, Global Avg Loss: 2.66741732, Time: 0.0209 Steps: 13780, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001016, Sample Num: 16256, Cur Loss: 0.42770356, Cur Avg Loss: 0.51652485, Log Avg loss: 0.35203653, Global Avg Loss: 2.66573829, Time: 0.0209 Steps: 13790, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001026, Sample Num: 16416, Cur Loss: 0.99524999, Cur Avg Loss: 0.51659962, Log Avg loss: 0.52419668, Global Avg Loss: 2.66418645, Time: 0.0247 Steps: 13800, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001036, Sample Num: 16576, Cur Loss: 0.22941306, Cur Avg Loss: 0.51530100, Log Avg loss: 0.38206186, Global Avg Loss: 2.66253393, Time: 0.0210 Steps: 13810, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001046, Sample Num: 16736, Cur Loss: 0.14769369, Cur Avg Loss: 0.51315585, Log Avg loss: 0.29091843, Global Avg Loss: 2.66081786, Time: 0.0209 Steps: 13820, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001056, Sample Num: 16896, Cur Loss: 0.36233926, Cur Avg Loss: 0.51355987, Log Avg loss: 0.55582006, Global Avg Loss: 2.65929580, Time: 0.0210 Steps: 13830, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001066, Sample Num: 17056, Cur Loss: 0.29068449, Cur Avg Loss: 0.51152971, Log Avg loss: 0.29714546, Global Avg Loss: 2.65758905, Time: 0.0209 Steps: 13840, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001076, Sample Num: 17216, Cur Loss: 0.30635685, Cur Avg Loss: 0.51068666, Log Avg loss: 0.42081750, Global Avg Loss: 2.65597405, Time: 0.0209 Steps: 13850, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001086, Sample Num: 17376, Cur Loss: 0.24095505, Cur Avg Loss: 0.51012771, Log Avg loss: 0.44998405, Global Avg Loss: 2.65438243, Time: 0.0210 Steps: 13860, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001096, Sample Num: 17536, Cur Loss: 0.31906909, Cur Avg Loss: 0.50839723, Log Avg loss: 0.32046769, Global Avg Loss: 2.65269972, Time: 0.0210 Steps: 13870, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001106, Sample Num: 17696, Cur Loss: 0.24698301, Cur Avg Loss: 0.50615939, Log Avg loss: 0.26089218, Global Avg Loss: 2.65097652, Time: 0.0209 Steps: 13880, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001116, Sample Num: 17856, Cur Loss: 0.71373498, Cur Avg Loss: 0.50587274, Log Avg loss: 0.47416868, Global Avg Loss: 2.64940934, Time: 0.0210 Steps: 13890, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001126, Sample Num: 18016, Cur Loss: 0.68594587, Cur Avg Loss: 0.50748043, Log Avg loss: 0.68689851, Global Avg Loss: 2.64799746, Time: 0.0209 Steps: 13900, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001136, Sample Num: 18176, Cur Loss: 0.57563430, Cur Avg Loss: 0.50756447, Log Avg loss: 0.51702759, Global Avg Loss: 2.64646549, Time: 0.0209 Steps: 13910, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001146, Sample Num: 18336, Cur Loss: 0.33332646, Cur Avg Loss: 0.50966904, Log Avg loss: 0.74874838, Global Avg Loss: 2.64510219, Time: 0.0209 Steps: 13920, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001156, Sample Num: 18496, Cur Loss: 0.15882209, Cur Avg Loss: 0.50973193, Log Avg loss: 0.51693907, Global Avg Loss: 2.64357443, Time: 0.0209 Steps: 13930, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001166, Sample Num: 18656, Cur Loss: 0.37472895, Cur Avg Loss: 0.51007067, Log Avg loss: 0.54922917, Global Avg Loss: 2.64207203, Time: 0.0209 Steps: 13940, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001176, Sample Num: 18816, Cur Loss: 0.56710982, Cur Avg Loss: 0.50830515, Log Avg loss: 0.30244563, Global Avg Loss: 2.64039488, Time: 0.0210 Steps: 13950, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001186, Sample Num: 18976, Cur Loss: 0.29517788, Cur Avg Loss: 0.50697032, Log Avg loss: 0.34999403, Global Avg Loss: 2.63875419, Time: 0.0210 Steps: 13960, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001196, Sample Num: 19136, Cur Loss: 0.62071371, Cur Avg Loss: 0.50691343, Log Avg loss: 0.50016595, Global Avg Loss: 2.63722335, Time: 0.0210 Steps: 13970, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001206, Sample Num: 19296, Cur Loss: 0.30962005, Cur Avg Loss: 0.50664055, Log Avg loss: 0.47400504, Global Avg Loss: 2.63567598, Time: 0.0209 Steps: 13980, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001216, Sample Num: 19456, Cur Loss: 0.37367609, Cur Avg Loss: 0.50580489, Log Avg loss: 0.40502414, Global Avg Loss: 2.63408152, Time: 0.0209 Steps: 13990, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001226, Sample Num: 19616, Cur Loss: 0.65767300, Cur Avg Loss: 0.50728002, Log Avg loss: 0.68665572, Global Avg Loss: 2.63269050, Time: 0.0209 Steps: 14000, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001236, Sample Num: 19776, Cur Loss: 0.18644315, Cur Avg Loss: 0.50597178, Log Avg loss: 0.34558090, Global Avg Loss: 2.63105802, Time: 0.0210 Steps: 14010, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001246, Sample Num: 19936, Cur Loss: 0.86369896, Cur Avg Loss: 0.50653515, Log Avg loss: 0.57616823, Global Avg Loss: 2.62959234, Time: 0.0209 Steps: 14020, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001256, Sample Num: 20096, Cur Loss: 0.25119197, Cur Avg Loss: 0.50721523, Log Avg loss: 0.59195345, Global Avg Loss: 2.62813999, Time: 0.0210 Steps: 14030, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001266, Sample Num: 20256, Cur Loss: 0.30260906, Cur Avg Loss: 0.50711658, Log Avg loss: 0.49472610, Global Avg Loss: 2.62662047, Time: 0.0209 Steps: 14040, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001276, Sample Num: 20416, Cur Loss: 0.39770874, Cur Avg Loss: 0.50552737, Log Avg loss: 0.30433337, Global Avg Loss: 2.62496759, Time: 0.0209 Steps: 14050, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001286, Sample Num: 20576, Cur Loss: 0.25071219, Cur Avg Loss: 0.50471849, Log Avg loss: 0.40150476, Global Avg Loss: 2.62338618, Time: 0.0208 Steps: 14060, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001296, Sample Num: 20736, Cur Loss: 0.65185869, Cur Avg Loss: 0.50480838, Log Avg loss: 0.51636896, Global Avg Loss: 2.62188866, Time: 0.0210 Steps: 14070, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001306, Sample Num: 20896, Cur Loss: 0.24968296, Cur Avg Loss: 0.50442957, Log Avg loss: 0.45533574, Global Avg Loss: 2.62034991, Time: 0.0208 Steps: 14080, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001316, Sample Num: 21056, Cur Loss: 0.17789184, Cur Avg Loss: 0.50392463, Log Avg loss: 0.43797968, Global Avg Loss: 2.61880103, Time: 0.0209 Steps: 14090, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001326, Sample Num: 21216, Cur Loss: 0.41888529, Cur Avg Loss: 0.50431853, Log Avg loss: 0.55615517, Global Avg Loss: 2.61733816, Time: 0.0208 Steps: 14100, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001336, Sample Num: 21376, Cur Loss: 0.29424191, Cur Avg Loss: 0.50400749, Log Avg loss: 0.46276426, Global Avg Loss: 2.61581118, Time: 0.0208 Steps: 14110, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001346, Sample Num: 21536, Cur Loss: 0.09520228, Cur Avg Loss: 0.50314067, Log Avg loss: 0.38733318, Global Avg Loss: 2.61423294, Time: 0.0208 Steps: 14120, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001356, Sample Num: 21696, Cur Loss: 0.84204441, Cur Avg Loss: 0.50183852, Log Avg loss: 0.32656944, Global Avg Loss: 2.61261393, Time: 0.0208 Steps: 14130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001366, Sample Num: 21856, Cur Loss: 0.59805501, Cur Avg Loss: 0.50197405, Log Avg loss: 0.52035183, Global Avg Loss: 2.61113425, Time: 0.0208 Steps: 14140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001376, Sample Num: 22016, Cur Loss: 0.19329417, Cur Avg Loss: 0.50223153, Log Avg loss: 0.53740316, Global Avg Loss: 2.60966872, Time: 0.0208 Steps: 14150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001386, Sample Num: 22176, Cur Loss: 0.27534252, Cur Avg Loss: 0.50080256, Log Avg loss: 0.30417604, Global Avg Loss: 2.60804054, Time: 0.0208 Steps: 14160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001396, Sample Num: 22336, Cur Loss: 0.73557353, Cur Avg Loss: 0.50095441, Log Avg loss: 0.52200116, Global Avg Loss: 2.60656839, Time: 0.0208 Steps: 14170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001406, Sample Num: 22496, Cur Loss: 0.83367777, Cur Avg Loss: 0.50194302, Log Avg loss: 0.63995247, Global Avg Loss: 2.60518150, Time: 0.0208 Steps: 14180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001416, Sample Num: 22656, Cur Loss: 0.35453799, Cur Avg Loss: 0.50470234, Log Avg loss: 0.89266277, Global Avg Loss: 2.60397465, Time: 0.0208 Steps: 14190, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001426, Sample Num: 22816, Cur Loss: 0.32484758, Cur Avg Loss: 0.50907353, Log Avg loss: 1.12803375, Global Avg Loss: 2.60293525, Time: 0.0208 Steps: 14200, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001436, Sample Num: 22976, Cur Loss: 0.17276223, Cur Avg Loss: 0.51215760, Log Avg loss: 0.95194679, Global Avg Loss: 2.60177340, Time: 0.0208 Steps: 14210, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001446, Sample Num: 23136, Cur Loss: 0.41944313, Cur Avg Loss: 0.51219475, Log Avg loss: 0.51752835, Global Avg Loss: 2.60030769, Time: 0.0208 Steps: 14220, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001456, Sample Num: 23296, Cur Loss: 0.37305617, Cur Avg Loss: 0.51141749, Log Avg loss: 0.39902701, Global Avg Loss: 2.59876076, Time: 0.0208 Steps: 14230, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001466, Sample Num: 23456, Cur Loss: 1.02497339, Cur Avg Loss: 0.51168896, Log Avg loss: 0.55121495, Global Avg Loss: 2.59732288, Time: 0.0208 Steps: 14240, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001476, Sample Num: 23616, Cur Loss: 1.39959633, Cur Avg Loss: 0.51363250, Log Avg loss: 0.79855483, Global Avg Loss: 2.59606058, Time: 0.0208 Steps: 14250, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001486, Sample Num: 23776, Cur Loss: 0.25818244, Cur Avg Loss: 0.51296472, Log Avg loss: 0.41440048, Global Avg Loss: 2.59453067, Time: 0.0208 Steps: 14260, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001496, Sample Num: 23936, Cur Loss: 0.92907125, Cur Avg Loss: 0.51289703, Log Avg loss: 0.50283803, Global Avg Loss: 2.59306487, Time: 0.0208 Steps: 14270, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001506, Sample Num: 24096, Cur Loss: 0.22020951, Cur Avg Loss: 0.51245672, Log Avg loss: 0.44658602, Global Avg Loss: 2.59156173, Time: 0.0208 Steps: 14280, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001516, Sample Num: 24256, Cur Loss: 0.80731261, Cur Avg Loss: 0.51214301, Log Avg loss: 0.46489880, Global Avg Loss: 2.59007352, Time: 0.0208 Steps: 14290, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001526, Sample Num: 24416, Cur Loss: 0.48841125, Cur Avg Loss: 0.51224909, Log Avg loss: 0.52833133, Global Avg Loss: 2.58863174, Time: 0.0209 Steps: 14300, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001536, Sample Num: 24576, Cur Loss: 0.56339502, Cur Avg Loss: 0.51480674, Log Avg loss: 0.90510347, Global Avg Loss: 2.58745527, Time: 0.0254 Steps: 14310, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001546, Sample Num: 24736, Cur Loss: 0.45427006, Cur Avg Loss: 0.51363468, Log Avg loss: 0.33360615, Global Avg Loss: 2.58588135, Time: 0.0208 Steps: 14320, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001556, Sample Num: 24896, Cur Loss: 0.93283916, Cur Avg Loss: 0.51338499, Log Avg loss: 0.47478384, Global Avg Loss: 2.58440815, Time: 0.0207 Steps: 14330, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001566, Sample Num: 25056, Cur Loss: 0.42639127, Cur Avg Loss: 0.51200201, Log Avg loss: 0.29681023, Global Avg Loss: 2.58281289, Time: 0.0208 Steps: 14340, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001576, Sample Num: 25216, Cur Loss: 0.77654684, Cur Avg Loss: 0.51236360, Log Avg loss: 0.56898792, Global Avg Loss: 2.58140953, Time: 0.0209 Steps: 14350, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001586, Sample Num: 25376, Cur Loss: 0.63966489, Cur Avg Loss: 0.51224127, Log Avg loss: 0.49296256, Global Avg Loss: 2.57995518, Time: 0.0208 Steps: 14360, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001596, Sample Num: 25536, Cur Loss: 0.25954199, Cur Avg Loss: 0.51175058, Log Avg loss: 0.43392748, Global Avg Loss: 2.57846177, Time: 0.0212 Steps: 14370, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001606, Sample Num: 25696, Cur Loss: 0.67742443, Cur Avg Loss: 0.51118712, Log Avg loss: 0.42125757, Global Avg Loss: 2.57696163, Time: 0.0213 Steps: 14380, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001616, Sample Num: 25856, Cur Loss: 0.23327923, Cur Avg Loss: 0.51199338, Log Avg loss: 0.64147969, Global Avg Loss: 2.57561661, Time: 0.0210 Steps: 14390, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001626, Sample Num: 26016, Cur Loss: 0.79398715, Cur Avg Loss: 0.51127762, Log Avg loss: 0.39561037, Global Avg Loss: 2.57410272, Time: 0.0210 Steps: 14400, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001636, Sample Num: 26176, Cur Loss: 0.39312270, Cur Avg Loss: 0.51163984, Log Avg loss: 0.57053665, Global Avg Loss: 2.57271232, Time: 0.0210 Steps: 14410, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001646, Sample Num: 26336, Cur Loss: 0.08589222, Cur Avg Loss: 0.51031373, Log Avg loss: 0.29336291, Global Avg Loss: 2.57113163, Time: 0.0209 Steps: 14420, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001656, Sample Num: 26496, Cur Loss: 0.45136464, Cur Avg Loss: 0.50989236, Log Avg loss: 0.44053440, Global Avg Loss: 2.56965513, Time: 0.0210 Steps: 14430, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001666, Sample Num: 26656, Cur Loss: 0.35685015, Cur Avg Loss: 0.50904636, Log Avg loss: 0.36894891, Global Avg Loss: 2.56813109, Time: 0.0210 Steps: 14440, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001676, Sample Num: 26816, Cur Loss: 0.16476384, Cur Avg Loss: 0.50837585, Log Avg loss: 0.39666816, Global Avg Loss: 2.56662835, Time: 0.0212 Steps: 14450, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001686, Sample Num: 26976, Cur Loss: 0.71110201, Cur Avg Loss: 0.50777853, Log Avg loss: 0.40766810, Global Avg Loss: 2.56513529, Time: 0.0211 Steps: 14460, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001696, Sample Num: 27136, Cur Loss: 1.10365129, Cur Avg Loss: 0.50767938, Log Avg loss: 0.49096360, Global Avg Loss: 2.56370187, Time: 0.0210 Steps: 14470, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001706, Sample Num: 27296, Cur Loss: 0.30796957, Cur Avg Loss: 0.50751259, Log Avg loss: 0.47922496, Global Avg Loss: 2.56226231, Time: 0.0212 Steps: 14480, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001716, Sample Num: 27456, Cur Loss: 0.44272602, Cur Avg Loss: 0.50616883, Log Avg loss: 0.27692276, Global Avg Loss: 2.56068513, Time: 0.0210 Steps: 14490, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001726, Sample Num: 27616, Cur Loss: 0.41180861, Cur Avg Loss: 0.50507731, Log Avg loss: 0.31777211, Global Avg Loss: 2.55913829, Time: 0.0210 Steps: 14500, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001736, Sample Num: 27776, Cur Loss: 0.26557419, Cur Avg Loss: 0.50384722, Log Avg loss: 0.29153408, Global Avg Loss: 2.55757550, Time: 0.0210 Steps: 14510, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001746, Sample Num: 27936, Cur Loss: 0.48663676, Cur Avg Loss: 0.50282079, Log Avg loss: 0.32463267, Global Avg Loss: 2.55603766, Time: 0.0210 Steps: 14520, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001756, Sample Num: 28096, Cur Loss: 0.18121302, Cur Avg Loss: 0.50224190, Log Avg loss: 0.40116756, Global Avg Loss: 2.55455461, Time: 0.0212 Steps: 14530, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001766, Sample Num: 28256, Cur Loss: 0.48278797, Cur Avg Loss: 0.50222576, Log Avg loss: 0.49939187, Global Avg Loss: 2.55314116, Time: 0.0210 Steps: 14540, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001776, Sample Num: 28416, Cur Loss: 0.37239540, Cur Avg Loss: 0.50191717, Log Avg loss: 0.44741983, Global Avg Loss: 2.55169393, Time: 0.0212 Steps: 14550, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001786, Sample Num: 28576, Cur Loss: 0.13256189, Cur Avg Loss: 0.50091618, Log Avg loss: 0.32313938, Global Avg Loss: 2.55016333, Time: 0.0210 Steps: 14560, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001796, Sample Num: 28736, Cur Loss: 0.14213619, Cur Avg Loss: 0.50014727, Log Avg loss: 0.36282165, Global Avg Loss: 2.54866206, Time: 0.0248 Steps: 14570, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001806, Sample Num: 28896, Cur Loss: 0.26828480, Cur Avg Loss: 0.49895190, Log Avg loss: 0.28426237, Global Avg Loss: 2.54710898, Time: 0.0209 Steps: 14580, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001816, Sample Num: 29056, Cur Loss: 0.52922446, Cur Avg Loss: 0.49931827, Log Avg loss: 0.56548431, Global Avg Loss: 2.54575077, Time: 0.0210 Steps: 14590, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001826, Sample Num: 29216, Cur Loss: 0.43259174, Cur Avg Loss: 0.49861728, Log Avg loss: 0.37131857, Global Avg Loss: 2.54426143, Time: 0.0210 Steps: 14600, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001836, Sample Num: 29376, Cur Loss: 0.19904324, Cur Avg Loss: 0.49804112, Log Avg loss: 0.39283415, Global Avg Loss: 2.54278886, Time: 0.0209 Steps: 14610, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001846, Sample Num: 29536, Cur Loss: 0.23673643, Cur Avg Loss: 0.49746023, Log Avg loss: 0.39080918, Global Avg Loss: 2.54131692, Time: 0.0210 Steps: 14620, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001856, Sample Num: 29696, Cur Loss: 0.54455012, Cur Avg Loss: 0.49717777, Log Avg loss: 0.44503483, Global Avg Loss: 2.53988405, Time: 0.0211 Steps: 14630, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001866, Sample Num: 29856, Cur Loss: 0.77134311, Cur Avg Loss: 0.49682651, Log Avg loss: 0.43163269, Global Avg Loss: 2.53844399, Time: 0.0209 Steps: 14640, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001876, Sample Num: 30016, Cur Loss: 0.19170216, Cur Avg Loss: 0.49625863, Log Avg loss: 0.39029278, Global Avg Loss: 2.53697768, Time: 0.0209 Steps: 14650, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001886, Sample Num: 30176, Cur Loss: 0.25644135, Cur Avg Loss: 0.49549462, Log Avg loss: 0.35216679, Global Avg Loss: 2.53548735, Time: 0.0210 Steps: 14660, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001896, Sample Num: 30336, Cur Loss: 0.60635746, Cur Avg Loss: 0.49458468, Log Avg loss: 0.32296922, Global Avg Loss: 2.53397916, Time: 0.0209 Steps: 14670, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001906, Sample Num: 30496, Cur Loss: 0.43708569, Cur Avg Loss: 0.49435954, Log Avg loss: 0.45167280, Global Avg Loss: 2.53256070, Time: 0.0209 Steps: 14680, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001916, Sample Num: 30656, Cur Loss: 0.48439717, Cur Avg Loss: 0.49403704, Log Avg loss: 0.43256844, Global Avg Loss: 2.53113116, Time: 0.0209 Steps: 14690, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001926, Sample Num: 30816, Cur Loss: 0.20442230, Cur Avg Loss: 0.49314691, Log Avg loss: 0.32259909, Global Avg Loss: 2.52962876, Time: 0.0207 Steps: 14700, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001936, Sample Num: 30976, Cur Loss: 0.13110647, Cur Avg Loss: 0.49215730, Log Avg loss: 0.30155738, Global Avg Loss: 2.52811409, Time: 0.0208 Steps: 14710, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001946, Sample Num: 31136, Cur Loss: 0.56478310, Cur Avg Loss: 0.49255830, Log Avg loss: 0.57019210, Global Avg Loss: 2.52678398, Time: 0.0207 Steps: 14720, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001956, Sample Num: 31296, Cur Loss: 0.60452920, Cur Avg Loss: 0.49192235, Log Avg loss: 0.36816576, Global Avg Loss: 2.52531852, Time: 0.0208 Steps: 14730, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001966, Sample Num: 31456, Cur Loss: 0.51651073, Cur Avg Loss: 0.49195104, Log Avg loss: 0.49756443, Global Avg Loss: 2.52394284, Time: 0.0207 Steps: 14740, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001976, Sample Num: 31616, Cur Loss: 0.35034150, Cur Avg Loss: 0.49183094, Log Avg loss: 0.46821787, Global Avg Loss: 2.52254913, Time: 0.0208 Steps: 14750, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001986, Sample Num: 31776, Cur Loss: 0.44764036, Cur Avg Loss: 0.49193693, Log Avg loss: 0.51288015, Global Avg Loss: 2.52118757, Time: 0.0207 Steps: 14760, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001996, Sample Num: 31936, Cur Loss: 0.22142330, Cur Avg Loss: 0.49162699, Log Avg loss: 0.43007312, Global Avg Loss: 2.51977178, Time: 0.0207 Steps: 14770, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002006, Sample Num: 32096, Cur Loss: 0.28648746, Cur Avg Loss: 0.49128456, Log Avg loss: 0.42293622, Global Avg Loss: 2.51835308, Time: 0.0208 Steps: 14780, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002016, Sample Num: 32256, Cur Loss: 0.88016719, Cur Avg Loss: 0.49224936, Log Avg loss: 0.68578808, Global Avg Loss: 2.51711403, Time: 0.0207 Steps: 14790, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002026, Sample Num: 32416, Cur Loss: 0.30163023, Cur Avg Loss: 0.49251261, Log Avg loss: 0.54558396, Global Avg Loss: 2.51578191, Time: 0.0207 Steps: 14800, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002036, Sample Num: 32576, Cur Loss: 1.03005195, Cur Avg Loss: 0.49194058, Log Avg loss: 0.37604764, Global Avg Loss: 2.51433712, Time: 0.0208 Steps: 14810, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002046, Sample Num: 32736, Cur Loss: 0.12317858, Cur Avg Loss: 0.49099138, Log Avg loss: 0.29773466, Global Avg Loss: 2.51284144, Time: 0.0207 Steps: 14820, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002056, Sample Num: 32896, Cur Loss: 0.33785826, Cur Avg Loss: 0.49277775, Log Avg loss: 0.85826781, Global Avg Loss: 2.51172575, Time: 0.0208 Steps: 14830, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002066, Sample Num: 33056, Cur Loss: 0.56370902, Cur Avg Loss: 0.49276566, Log Avg loss: 0.49028039, Global Avg Loss: 2.51036359, Time: 0.0208 Steps: 14840, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002076, Sample Num: 33216, Cur Loss: 0.27598608, Cur Avg Loss: 0.49280477, Log Avg loss: 0.50088463, Global Avg Loss: 2.50901040, Time: 0.0210 Steps: 14850, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002086, Sample Num: 33376, Cur Loss: 0.27721789, Cur Avg Loss: 0.49204950, Log Avg loss: 0.33525509, Global Avg Loss: 2.50754758, Time: 0.0210 Steps: 14860, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002096, Sample Num: 33536, Cur Loss: 0.29127368, Cur Avg Loss: 0.49385689, Log Avg loss: 0.87087922, Global Avg Loss: 2.50644693, Time: 0.0209 Steps: 14870, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002106, Sample Num: 33696, Cur Loss: 0.87450886, Cur Avg Loss: 0.49499942, Log Avg loss: 0.73447295, Global Avg Loss: 2.50525608, Time: 0.0210 Steps: 14880, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002116, Sample Num: 33856, Cur Loss: 0.35827273, Cur Avg Loss: 0.49485779, Log Avg loss: 0.46503174, Global Avg Loss: 2.50388589, Time: 0.0209 Steps: 14890, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002126, Sample Num: 34016, Cur Loss: 0.26779348, Cur Avg Loss: 0.49466349, Log Avg loss: 0.45354919, Global Avg Loss: 2.50250982, Time: 0.0209 Steps: 14900, Updated lr: 0.000087 ***** Running evaluation checkpoint-14903 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-14903 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.684337, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.434459, "eval_total_loss": 305.424691, "eval_mae": 0.532609, "eval_mse": 0.434445, "eval_r2": 0.723838, "eval_sp_statistic": 0.851233, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.888711, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.349998, "test_total_loss": 175.699159, "test_mae": 0.369046, "test_mse": 0.350125, "test_r2": 0.774026, "test_sp_statistic": 0.847749, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.914359, "test_ps_pvalue": 0.0, "lr": 8.681555239449977e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.50205071309511, "train_cur_epoch_loss": 1052.3200240433216, "train_cur_epoch_avg_loss": 0.4942790155205832, "train_cur_epoch_time": 44.68433666229248, "train_cur_epoch_avg_time": 0.020988415529493886, "epoch": 7, "step": 14903} ################################################## Training, Epoch: 0008, Batch: 000007, Sample Num: 112, Cur Loss: 0.71965629, Cur Avg Loss: 0.42106918, Log Avg loss: 0.36129268, Global Avg Loss: 2.50107373, Time: 0.0210 Steps: 14910, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000017, Sample Num: 272, Cur Loss: 0.29580438, Cur Avg Loss: 0.32722996, Log Avg loss: 0.26154250, Global Avg Loss: 2.49957270, Time: 0.0209 Steps: 14920, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000027, Sample Num: 432, Cur Loss: 0.76000339, Cur Avg Loss: 0.37030228, Log Avg loss: 0.44352524, Global Avg Loss: 2.49819558, Time: 0.0207 Steps: 14930, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000037, Sample Num: 592, Cur Loss: 0.37969181, Cur Avg Loss: 0.38719562, Log Avg loss: 0.43280763, Global Avg Loss: 2.49681312, Time: 0.0208 Steps: 14940, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000047, Sample Num: 752, Cur Loss: 0.98677504, Cur Avg Loss: 0.42437218, Log Avg loss: 0.56192544, Global Avg Loss: 2.49551888, Time: 0.0208 Steps: 14950, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000057, Sample Num: 912, Cur Loss: 0.24056232, Cur Avg Loss: 0.42359656, Log Avg loss: 0.41995117, Global Avg Loss: 2.49413147, Time: 0.0208 Steps: 14960, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000067, Sample Num: 1072, Cur Loss: 0.25060052, Cur Avg Loss: 0.42676721, Log Avg loss: 0.44483989, Global Avg Loss: 2.49276254, Time: 0.0208 Steps: 14970, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000077, Sample Num: 1232, Cur Loss: 0.10944417, Cur Avg Loss: 0.50956438, Log Avg loss: 1.06430541, Global Avg Loss: 2.49180896, Time: 0.0208 Steps: 14980, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000087, Sample Num: 1392, Cur Loss: 0.81031632, Cur Avg Loss: 0.51774820, Log Avg loss: 0.58076363, Global Avg Loss: 2.49053408, Time: 0.0208 Steps: 14990, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000097, Sample Num: 1552, Cur Loss: 0.19016856, Cur Avg Loss: 0.53211585, Log Avg loss: 0.65711436, Global Avg Loss: 2.48931180, Time: 0.0208 Steps: 15000, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000107, Sample Num: 1712, Cur Loss: 0.53719848, Cur Avg Loss: 0.52412934, Log Avg loss: 0.44666018, Global Avg Loss: 2.48795094, Time: 0.0208 Steps: 15010, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000117, Sample Num: 1872, Cur Loss: 0.60193855, Cur Avg Loss: 0.52061854, Log Avg loss: 0.48305306, Global Avg Loss: 2.48661612, Time: 0.0207 Steps: 15020, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000127, Sample Num: 2032, Cur Loss: 0.67693383, Cur Avg Loss: 0.53382817, Log Avg loss: 0.68838076, Global Avg Loss: 2.48541969, Time: 0.0209 Steps: 15030, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000137, Sample Num: 2192, Cur Loss: 0.38056603, Cur Avg Loss: 0.51942198, Log Avg loss: 0.33646348, Global Avg Loss: 2.48399086, Time: 0.0208 Steps: 15040, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000147, Sample Num: 2352, Cur Loss: 0.38865837, Cur Avg Loss: 0.52066359, Log Avg loss: 0.53767359, Global Avg Loss: 2.48269763, Time: 0.0208 Steps: 15050, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000157, Sample Num: 2512, Cur Loss: 0.88728541, Cur Avg Loss: 0.53356314, Log Avg loss: 0.72318657, Global Avg Loss: 2.48152930, Time: 0.0208 Steps: 15060, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000167, Sample Num: 2672, Cur Loss: 0.22507523, Cur Avg Loss: 0.52425888, Log Avg loss: 0.37818190, Global Avg Loss: 2.48013358, Time: 0.0208 Steps: 15070, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000177, Sample Num: 2832, Cur Loss: 0.96177948, Cur Avg Loss: 0.52506054, Log Avg loss: 0.53844825, Global Avg Loss: 2.47884599, Time: 0.0208 Steps: 15080, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000187, Sample Num: 2992, Cur Loss: 0.36517066, Cur Avg Loss: 0.51394818, Log Avg loss: 0.31725953, Global Avg Loss: 2.47741352, Time: 0.0208 Steps: 15090, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000197, Sample Num: 3152, Cur Loss: 0.09280108, Cur Avg Loss: 0.50536701, Log Avg loss: 0.34489905, Global Avg Loss: 2.47600126, Time: 0.0208 Steps: 15100, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000207, Sample Num: 3312, Cur Loss: 0.31054354, Cur Avg Loss: 0.49789533, Log Avg loss: 0.35070320, Global Avg Loss: 2.47459471, Time: 0.0208 Steps: 15110, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000217, Sample Num: 3472, Cur Loss: 0.21630815, Cur Avg Loss: 0.48639927, Log Avg loss: 0.24843093, Global Avg Loss: 2.47312238, Time: 0.0208 Steps: 15120, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000227, Sample Num: 3632, Cur Loss: 0.55717158, Cur Avg Loss: 0.48275899, Log Avg loss: 0.40376483, Global Avg Loss: 2.47175466, Time: 0.0208 Steps: 15130, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000237, Sample Num: 3792, Cur Loss: 0.58649516, Cur Avg Loss: 0.48525982, Log Avg loss: 0.54202857, Global Avg Loss: 2.47048008, Time: 0.0208 Steps: 15140, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000247, Sample Num: 3952, Cur Loss: 0.63161075, Cur Avg Loss: 0.48665269, Log Avg loss: 0.51966387, Global Avg Loss: 2.46919241, Time: 0.0208 Steps: 15150, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000257, Sample Num: 4112, Cur Loss: 0.32906580, Cur Avg Loss: 0.48511774, Log Avg loss: 0.44720442, Global Avg Loss: 2.46785864, Time: 0.0243 Steps: 15160, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000267, Sample Num: 4272, Cur Loss: 0.18415928, Cur Avg Loss: 0.47979150, Log Avg loss: 0.34290707, Global Avg Loss: 2.46645788, Time: 0.0208 Steps: 15170, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000277, Sample Num: 4432, Cur Loss: 0.34633312, Cur Avg Loss: 0.47366921, Log Avg loss: 0.31020417, Global Avg Loss: 2.46503743, Time: 0.0208 Steps: 15180, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000287, Sample Num: 4592, Cur Loss: 0.18075877, Cur Avg Loss: 0.47030954, Log Avg loss: 0.37724654, Global Avg Loss: 2.46366298, Time: 0.0208 Steps: 15190, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000297, Sample Num: 4752, Cur Loss: 0.39576414, Cur Avg Loss: 0.46536620, Log Avg loss: 0.32349238, Global Avg Loss: 2.46225497, Time: 0.0208 Steps: 15200, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000307, Sample Num: 4912, Cur Loss: 0.31882519, Cur Avg Loss: 0.47185778, Log Avg loss: 0.66465781, Global Avg Loss: 2.46107312, Time: 0.0208 Steps: 15210, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000317, Sample Num: 5072, Cur Loss: 0.19456378, Cur Avg Loss: 0.47418321, Log Avg loss: 0.54557395, Global Avg Loss: 2.45981458, Time: 0.0208 Steps: 15220, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000327, Sample Num: 5232, Cur Loss: 0.27755862, Cur Avg Loss: 0.47897099, Log Avg loss: 0.63074339, Global Avg Loss: 2.45861361, Time: 0.0208 Steps: 15230, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000337, Sample Num: 5392, Cur Loss: 0.29629812, Cur Avg Loss: 0.47808574, Log Avg loss: 0.44913828, Global Avg Loss: 2.45729506, Time: 0.0208 Steps: 15240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000347, Sample Num: 5552, Cur Loss: 0.12956658, Cur Avg Loss: 0.47425327, Log Avg loss: 0.34509882, Global Avg Loss: 2.45591001, Time: 0.0208 Steps: 15250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000357, Sample Num: 5712, Cur Loss: 0.24003616, Cur Avg Loss: 0.47430401, Log Avg loss: 0.47606479, Global Avg Loss: 2.45461260, Time: 0.0208 Steps: 15260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000367, Sample Num: 5872, Cur Loss: 0.37738913, Cur Avg Loss: 0.46884153, Log Avg loss: 0.27383090, Global Avg Loss: 2.45318445, Time: 0.0208 Steps: 15270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000377, Sample Num: 6032, Cur Loss: 0.10113434, Cur Avg Loss: 0.46877992, Log Avg loss: 0.46651884, Global Avg Loss: 2.45188428, Time: 0.0208 Steps: 15280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000387, Sample Num: 6192, Cur Loss: 0.68739861, Cur Avg Loss: 0.46577728, Log Avg loss: 0.35257788, Global Avg Loss: 2.45051129, Time: 0.0208 Steps: 15290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000397, Sample Num: 6352, Cur Loss: 0.17869142, Cur Avg Loss: 0.46501979, Log Avg loss: 0.43570495, Global Avg Loss: 2.44919442, Time: 0.0208 Steps: 15300, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000407, Sample Num: 6512, Cur Loss: 0.45352286, Cur Avg Loss: 0.46477991, Log Avg loss: 0.45525649, Global Avg Loss: 2.44789204, Time: 0.0208 Steps: 15310, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000417, Sample Num: 6672, Cur Loss: 0.15290147, Cur Avg Loss: 0.46665380, Log Avg loss: 0.54292114, Global Avg Loss: 2.44664859, Time: 0.0208 Steps: 15320, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000427, Sample Num: 6832, Cur Loss: 0.28002208, Cur Avg Loss: 0.47277157, Log Avg loss: 0.72788260, Global Avg Loss: 2.44552741, Time: 0.0208 Steps: 15330, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000437, Sample Num: 6992, Cur Loss: 0.20161811, Cur Avg Loss: 0.48464201, Log Avg loss: 0.99151001, Global Avg Loss: 2.44457955, Time: 0.0208 Steps: 15340, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000447, Sample Num: 7152, Cur Loss: 0.64053434, Cur Avg Loss: 0.48857188, Log Avg loss: 0.66030723, Global Avg Loss: 2.44341716, Time: 0.0208 Steps: 15350, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000457, Sample Num: 7312, Cur Loss: 0.69499242, Cur Avg Loss: 0.49324839, Log Avg loss: 0.70228809, Global Avg Loss: 2.44228361, Time: 0.0208 Steps: 15360, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000467, Sample Num: 7472, Cur Loss: 0.44469857, Cur Avg Loss: 0.48965026, Log Avg loss: 0.32521578, Global Avg Loss: 2.44090621, Time: 0.0208 Steps: 15370, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000477, Sample Num: 7632, Cur Loss: 1.35979736, Cur Avg Loss: 0.49080725, Log Avg loss: 0.54483880, Global Avg Loss: 2.43967340, Time: 0.0208 Steps: 15380, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000487, Sample Num: 7792, Cur Loss: 0.10535763, Cur Avg Loss: 0.48817436, Log Avg loss: 0.36258568, Global Avg Loss: 2.43832376, Time: 0.0208 Steps: 15390, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000497, Sample Num: 7952, Cur Loss: 0.48353004, Cur Avg Loss: 0.48390596, Log Avg loss: 0.27603489, Global Avg Loss: 2.43691968, Time: 0.0208 Steps: 15400, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000507, Sample Num: 8112, Cur Loss: 0.32529581, Cur Avg Loss: 0.48108781, Log Avg loss: 0.34102557, Global Avg Loss: 2.43555959, Time: 0.0208 Steps: 15410, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000517, Sample Num: 8272, Cur Loss: 0.14710698, Cur Avg Loss: 0.47709250, Log Avg loss: 0.27453049, Global Avg Loss: 2.43415815, Time: 0.0245 Steps: 15420, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000527, Sample Num: 8432, Cur Loss: 0.30914545, Cur Avg Loss: 0.47520196, Log Avg loss: 0.37746106, Global Avg Loss: 2.43282522, Time: 0.0208 Steps: 15430, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000537, Sample Num: 8592, Cur Loss: 0.29270521, Cur Avg Loss: 0.47386980, Log Avg loss: 0.40366484, Global Avg Loss: 2.43151100, Time: 0.0208 Steps: 15440, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000547, Sample Num: 8752, Cur Loss: 0.16801870, Cur Avg Loss: 0.47068716, Log Avg loss: 0.29977942, Global Avg Loss: 2.43013124, Time: 0.0208 Steps: 15450, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000557, Sample Num: 8912, Cur Loss: 0.31116426, Cur Avg Loss: 0.46832753, Log Avg loss: 0.33925572, Global Avg Loss: 2.42877880, Time: 0.0208 Steps: 15460, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000567, Sample Num: 9072, Cur Loss: 0.11148176, Cur Avg Loss: 0.46698366, Log Avg loss: 0.39212979, Global Avg Loss: 2.42746228, Time: 0.0207 Steps: 15470, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000577, Sample Num: 9232, Cur Loss: 0.37407628, Cur Avg Loss: 0.46469935, Log Avg loss: 0.33517923, Global Avg Loss: 2.42611068, Time: 0.0208 Steps: 15480, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000587, Sample Num: 9392, Cur Loss: 0.90383470, Cur Avg Loss: 0.46455900, Log Avg loss: 0.45646058, Global Avg Loss: 2.42483912, Time: 0.0208 Steps: 15490, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000597, Sample Num: 9552, Cur Loss: 0.38232851, Cur Avg Loss: 0.46297712, Log Avg loss: 0.37012109, Global Avg Loss: 2.42351349, Time: 0.0208 Steps: 15500, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000607, Sample Num: 9712, Cur Loss: 0.54386580, Cur Avg Loss: 0.46242350, Log Avg loss: 0.42937222, Global Avg Loss: 2.42222778, Time: 0.0208 Steps: 15510, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000617, Sample Num: 9872, Cur Loss: 0.28915724, Cur Avg Loss: 0.46417625, Log Avg loss: 0.57056808, Global Avg Loss: 2.42103470, Time: 0.0208 Steps: 15520, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000627, Sample Num: 10032, Cur Loss: 0.11065388, Cur Avg Loss: 0.46657958, Log Avg loss: 0.61486520, Global Avg Loss: 2.41987168, Time: 0.0208 Steps: 15530, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000637, Sample Num: 10192, Cur Loss: 0.11133897, Cur Avg Loss: 0.46621270, Log Avg loss: 0.44320957, Global Avg Loss: 2.41859970, Time: 0.0208 Steps: 15540, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000647, Sample Num: 10352, Cur Loss: 0.74654567, Cur Avg Loss: 0.46563978, Log Avg loss: 0.42914480, Global Avg Loss: 2.41732030, Time: 0.0207 Steps: 15550, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000657, Sample Num: 10512, Cur Loss: 0.56017506, Cur Avg Loss: 0.46751719, Log Avg loss: 0.58898505, Global Avg Loss: 2.41614528, Time: 0.0208 Steps: 15560, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000667, Sample Num: 10672, Cur Loss: 0.24939045, Cur Avg Loss: 0.46632374, Log Avg loss: 0.38791402, Global Avg Loss: 2.41484263, Time: 0.0207 Steps: 15570, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000677, Sample Num: 10832, Cur Loss: 0.46673608, Cur Avg Loss: 0.46456947, Log Avg loss: 0.34755986, Global Avg Loss: 2.41351575, Time: 0.0207 Steps: 15580, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000687, Sample Num: 10992, Cur Loss: 0.30664700, Cur Avg Loss: 0.46603883, Log Avg loss: 0.56551486, Global Avg Loss: 2.41233037, Time: 0.0208 Steps: 15590, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000697, Sample Num: 11152, Cur Loss: 1.09176862, Cur Avg Loss: 0.46822328, Log Avg loss: 0.61829490, Global Avg Loss: 2.41118035, Time: 0.0207 Steps: 15600, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000707, Sample Num: 11312, Cur Loss: 0.49786955, Cur Avg Loss: 0.46868526, Log Avg loss: 0.50088506, Global Avg Loss: 2.40995658, Time: 0.0208 Steps: 15610, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000717, Sample Num: 11472, Cur Loss: 1.77680182, Cur Avg Loss: 0.46904269, Log Avg loss: 0.49431274, Global Avg Loss: 2.40873018, Time: 0.0207 Steps: 15620, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000727, Sample Num: 11632, Cur Loss: 0.57560396, Cur Avg Loss: 0.47159556, Log Avg loss: 0.65463692, Global Avg Loss: 2.40760792, Time: 0.0208 Steps: 15630, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000737, Sample Num: 11792, Cur Loss: 2.18245625, Cur Avg Loss: 0.47706684, Log Avg loss: 0.87482855, Global Avg Loss: 2.40662788, Time: 0.0208 Steps: 15640, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000747, Sample Num: 11952, Cur Loss: 0.51985431, Cur Avg Loss: 0.48054372, Log Avg loss: 0.73679006, Global Avg Loss: 2.40556089, Time: 0.0207 Steps: 15650, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000757, Sample Num: 12112, Cur Loss: 0.64061880, Cur Avg Loss: 0.48072297, Log Avg loss: 0.49411272, Global Avg Loss: 2.40434030, Time: 0.0207 Steps: 15660, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000767, Sample Num: 12272, Cur Loss: 0.97814310, Cur Avg Loss: 0.48107048, Log Avg loss: 0.50737661, Global Avg Loss: 2.40312973, Time: 0.0207 Steps: 15670, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000777, Sample Num: 12432, Cur Loss: 0.15560937, Cur Avg Loss: 0.48638120, Log Avg loss: 0.89371412, Global Avg Loss: 2.40216709, Time: 0.0209 Steps: 15680, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000787, Sample Num: 12592, Cur Loss: 0.49679592, Cur Avg Loss: 0.48534143, Log Avg loss: 0.40455116, Global Avg Loss: 2.40089391, Time: 0.0208 Steps: 15690, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000797, Sample Num: 12752, Cur Loss: 0.30241430, Cur Avg Loss: 0.48377097, Log Avg loss: 0.36017562, Global Avg Loss: 2.39959409, Time: 0.0208 Steps: 15700, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000807, Sample Num: 12912, Cur Loss: 0.37834466, Cur Avg Loss: 0.48567667, Log Avg loss: 0.63756061, Global Avg Loss: 2.39847249, Time: 0.0208 Steps: 15710, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000817, Sample Num: 13072, Cur Loss: 1.88167727, Cur Avg Loss: 0.48720146, Log Avg loss: 0.61025202, Global Avg Loss: 2.39733495, Time: 0.0208 Steps: 15720, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000827, Sample Num: 13232, Cur Loss: 0.76566648, Cur Avg Loss: 0.48937113, Log Avg loss: 0.66663385, Global Avg Loss: 2.39623469, Time: 0.0208 Steps: 15730, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000837, Sample Num: 13392, Cur Loss: 0.24130458, Cur Avg Loss: 0.48850614, Log Avg loss: 0.41697126, Global Avg Loss: 2.39497722, Time: 0.0208 Steps: 15740, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000847, Sample Num: 13552, Cur Loss: 0.78871191, Cur Avg Loss: 0.48913932, Log Avg loss: 0.54213658, Global Avg Loss: 2.39380081, Time: 0.0208 Steps: 15750, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000857, Sample Num: 13712, Cur Loss: 1.52910209, Cur Avg Loss: 0.49669536, Log Avg loss: 1.13669153, Global Avg Loss: 2.39300315, Time: 0.0208 Steps: 15760, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000867, Sample Num: 13872, Cur Loss: 1.77140498, Cur Avg Loss: 0.50062115, Log Avg loss: 0.83706194, Global Avg Loss: 2.39201651, Time: 0.0209 Steps: 15770, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000877, Sample Num: 14032, Cur Loss: 0.73573530, Cur Avg Loss: 0.50575216, Log Avg loss: 0.95060993, Global Avg Loss: 2.39110307, Time: 0.0208 Steps: 15780, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000887, Sample Num: 14192, Cur Loss: 0.24708666, Cur Avg Loss: 0.50498618, Log Avg loss: 0.43780990, Global Avg Loss: 2.38986602, Time: 0.0208 Steps: 15790, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000897, Sample Num: 14352, Cur Loss: 0.56767929, Cur Avg Loss: 0.50472919, Log Avg loss: 0.48193472, Global Avg Loss: 2.38865847, Time: 0.0208 Steps: 15800, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000907, Sample Num: 14512, Cur Loss: 1.11471868, Cur Avg Loss: 0.50624810, Log Avg loss: 0.64249385, Global Avg Loss: 2.38755400, Time: 0.0208 Steps: 15810, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000917, Sample Num: 14672, Cur Loss: 0.21659297, Cur Avg Loss: 0.50518090, Log Avg loss: 0.40838624, Global Avg Loss: 2.38630295, Time: 0.0208 Steps: 15820, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000927, Sample Num: 14832, Cur Loss: 0.29274642, Cur Avg Loss: 0.50250051, Log Avg loss: 0.25670892, Global Avg Loss: 2.38495766, Time: 0.0208 Steps: 15830, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000937, Sample Num: 14992, Cur Loss: 0.12202685, Cur Avg Loss: 0.50137894, Log Avg loss: 0.39740917, Global Avg Loss: 2.38370289, Time: 0.0208 Steps: 15840, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000947, Sample Num: 15152, Cur Loss: 0.71023500, Cur Avg Loss: 0.50043499, Log Avg loss: 0.41198654, Global Avg Loss: 2.38245891, Time: 0.0208 Steps: 15850, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000957, Sample Num: 15312, Cur Loss: 0.38801321, Cur Avg Loss: 0.49843145, Log Avg loss: 0.30869608, Global Avg Loss: 2.38115137, Time: 0.0208 Steps: 15860, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000967, Sample Num: 15472, Cur Loss: 0.23581813, Cur Avg Loss: 0.49838635, Log Avg loss: 0.49407031, Global Avg Loss: 2.37996228, Time: 0.0208 Steps: 15870, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000977, Sample Num: 15632, Cur Loss: 0.43116060, Cur Avg Loss: 0.49698940, Log Avg loss: 0.36190443, Global Avg Loss: 2.37869146, Time: 0.0208 Steps: 15880, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000987, Sample Num: 15792, Cur Loss: 0.36048776, Cur Avg Loss: 0.49562502, Log Avg loss: 0.36232490, Global Avg Loss: 2.37742251, Time: 0.0208 Steps: 15890, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000997, Sample Num: 15952, Cur Loss: 0.62561679, Cur Avg Loss: 0.49411487, Log Avg loss: 0.34506376, Global Avg Loss: 2.37614430, Time: 0.0208 Steps: 15900, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001007, Sample Num: 16112, Cur Loss: 0.20805249, Cur Avg Loss: 0.49240265, Log Avg loss: 0.32169449, Global Avg Loss: 2.37485300, Time: 0.0208 Steps: 15910, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001017, Sample Num: 16272, Cur Loss: 0.26476818, Cur Avg Loss: 0.49187642, Log Avg loss: 0.43888418, Global Avg Loss: 2.37363694, Time: 0.0208 Steps: 15920, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001027, Sample Num: 16432, Cur Loss: 0.31473953, Cur Avg Loss: 0.49186316, Log Avg loss: 0.49051514, Global Avg Loss: 2.37245482, Time: 0.0246 Steps: 15930, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001037, Sample Num: 16592, Cur Loss: 0.50427592, Cur Avg Loss: 0.49228483, Log Avg loss: 0.53559034, Global Avg Loss: 2.37130246, Time: 0.0208 Steps: 15940, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001047, Sample Num: 16752, Cur Loss: 0.48940122, Cur Avg Loss: 0.49187604, Log Avg loss: 0.44948416, Global Avg Loss: 2.37009755, Time: 0.0208 Steps: 15950, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001057, Sample Num: 16912, Cur Loss: 0.27573806, Cur Avg Loss: 0.49280168, Log Avg loss: 0.58971595, Global Avg Loss: 2.36898203, Time: 0.0206 Steps: 15960, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001067, Sample Num: 17072, Cur Loss: 0.35708505, Cur Avg Loss: 0.49351923, Log Avg loss: 0.56936435, Global Avg Loss: 2.36785515, Time: 0.0208 Steps: 15970, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001077, Sample Num: 17232, Cur Loss: 0.41953230, Cur Avg Loss: 0.49393732, Log Avg loss: 0.53854837, Global Avg Loss: 2.36671041, Time: 0.0208 Steps: 15980, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001087, Sample Num: 17392, Cur Loss: 0.26820886, Cur Avg Loss: 0.49386564, Log Avg loss: 0.48614555, Global Avg Loss: 2.36553432, Time: 0.0207 Steps: 15990, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001097, Sample Num: 17552, Cur Loss: 0.23710287, Cur Avg Loss: 0.49458072, Log Avg loss: 0.57230921, Global Avg Loss: 2.36441355, Time: 0.0208 Steps: 16000, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001107, Sample Num: 17712, Cur Loss: 0.50806910, Cur Avg Loss: 0.49485224, Log Avg loss: 0.52463800, Global Avg Loss: 2.36326441, Time: 0.0208 Steps: 16010, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001117, Sample Num: 17872, Cur Loss: 0.21544440, Cur Avg Loss: 0.49346096, Log Avg loss: 0.33944704, Global Avg Loss: 2.36200110, Time: 0.0208 Steps: 16020, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001127, Sample Num: 18032, Cur Loss: 0.41481000, Cur Avg Loss: 0.49202475, Log Avg loss: 0.33160012, Global Avg Loss: 2.36073448, Time: 0.0208 Steps: 16030, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001137, Sample Num: 18192, Cur Loss: 0.28373605, Cur Avg Loss: 0.49028015, Log Avg loss: 0.29366366, Global Avg Loss: 2.35944578, Time: 0.0207 Steps: 16040, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001147, Sample Num: 18352, Cur Loss: 0.22854775, Cur Avg Loss: 0.48873589, Log Avg loss: 0.31315356, Global Avg Loss: 2.35817083, Time: 0.0208 Steps: 16050, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001157, Sample Num: 18512, Cur Loss: 0.19490926, Cur Avg Loss: 0.48801009, Log Avg loss: 0.40476067, Global Avg Loss: 2.35695451, Time: 0.0208 Steps: 16060, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001167, Sample Num: 18672, Cur Loss: 0.48552489, Cur Avg Loss: 0.48689451, Log Avg loss: 0.35782202, Global Avg Loss: 2.35571050, Time: 0.0208 Steps: 16070, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001177, Sample Num: 18832, Cur Loss: 0.14958526, Cur Avg Loss: 0.48749736, Log Avg loss: 0.55784929, Global Avg Loss: 2.35459242, Time: 0.0208 Steps: 16080, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001187, Sample Num: 18992, Cur Loss: 0.44277319, Cur Avg Loss: 0.48917678, Log Avg loss: 0.68684515, Global Avg Loss: 2.35355591, Time: 0.0208 Steps: 16090, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001197, Sample Num: 19152, Cur Loss: 0.36987966, Cur Avg Loss: 0.48972657, Log Avg loss: 0.55498578, Global Avg Loss: 2.35243879, Time: 0.0208 Steps: 16100, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001207, Sample Num: 19312, Cur Loss: 0.70955324, Cur Avg Loss: 0.48838224, Log Avg loss: 0.32746625, Global Avg Loss: 2.35118182, Time: 0.0208 Steps: 16110, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001217, Sample Num: 19472, Cur Loss: 0.31974614, Cur Avg Loss: 0.48661019, Log Avg loss: 0.27272446, Global Avg Loss: 2.34989246, Time: 0.0208 Steps: 16120, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001227, Sample Num: 19632, Cur Loss: 0.30239975, Cur Avg Loss: 0.48670471, Log Avg loss: 0.49820678, Global Avg Loss: 2.34874448, Time: 0.0207 Steps: 16130, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001237, Sample Num: 19792, Cur Loss: 0.21443066, Cur Avg Loss: 0.48597633, Log Avg loss: 0.39660423, Global Avg Loss: 2.34753497, Time: 0.0208 Steps: 16140, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001247, Sample Num: 19952, Cur Loss: 0.40104797, Cur Avg Loss: 0.48525027, Log Avg loss: 0.39543748, Global Avg Loss: 2.34632625, Time: 0.0207 Steps: 16150, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001257, Sample Num: 20112, Cur Loss: 1.24947357, Cur Avg Loss: 0.48753114, Log Avg loss: 0.77195522, Global Avg Loss: 2.34535201, Time: 0.0207 Steps: 16160, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001267, Sample Num: 20272, Cur Loss: 0.77644157, Cur Avg Loss: 0.48861292, Log Avg loss: 0.62459317, Global Avg Loss: 2.34428784, Time: 0.0208 Steps: 16170, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001277, Sample Num: 20432, Cur Loss: 0.30655432, Cur Avg Loss: 0.48786057, Log Avg loss: 0.39253734, Global Avg Loss: 2.34308157, Time: 0.0207 Steps: 16180, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001287, Sample Num: 20592, Cur Loss: 0.31262052, Cur Avg Loss: 0.48711821, Log Avg loss: 0.39231914, Global Avg Loss: 2.34187665, Time: 0.0208 Steps: 16190, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001297, Sample Num: 20752, Cur Loss: 0.34692496, Cur Avg Loss: 0.48645575, Log Avg loss: 0.40119718, Global Avg Loss: 2.34067870, Time: 0.0208 Steps: 16200, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001307, Sample Num: 20912, Cur Loss: 0.30325830, Cur Avg Loss: 0.48518142, Log Avg loss: 0.31990108, Global Avg Loss: 2.33943207, Time: 0.0208 Steps: 16210, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001317, Sample Num: 21072, Cur Loss: 0.77383983, Cur Avg Loss: 0.48440401, Log Avg loss: 0.38279557, Global Avg Loss: 2.33822576, Time: 0.0207 Steps: 16220, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001327, Sample Num: 21232, Cur Loss: 0.60309017, Cur Avg Loss: 0.48415297, Log Avg loss: 0.45109147, Global Avg Loss: 2.33706302, Time: 0.0208 Steps: 16230, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001337, Sample Num: 21392, Cur Loss: 0.67318964, Cur Avg Loss: 0.48357612, Log Avg loss: 0.40702779, Global Avg Loss: 2.33587457, Time: 0.0207 Steps: 16240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001347, Sample Num: 21552, Cur Loss: 0.43740678, Cur Avg Loss: 0.48525694, Log Avg loss: 0.70998316, Global Avg Loss: 2.33487402, Time: 0.0208 Steps: 16250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001357, Sample Num: 21712, Cur Loss: 0.57858384, Cur Avg Loss: 0.48507875, Log Avg loss: 0.46107619, Global Avg Loss: 2.33372163, Time: 0.0208 Steps: 16260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001367, Sample Num: 21872, Cur Loss: 0.63593149, Cur Avg Loss: 0.48569775, Log Avg loss: 0.56969632, Global Avg Loss: 2.33263741, Time: 0.0207 Steps: 16270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001377, Sample Num: 22032, Cur Loss: 0.34535635, Cur Avg Loss: 0.48409559, Log Avg loss: 0.26508055, Global Avg Loss: 2.33136741, Time: 0.0207 Steps: 16280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001387, Sample Num: 22192, Cur Loss: 0.18669286, Cur Avg Loss: 0.48441387, Log Avg loss: 0.52823997, Global Avg Loss: 2.33026052, Time: 0.0207 Steps: 16290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001397, Sample Num: 22352, Cur Loss: 0.44580999, Cur Avg Loss: 0.48311340, Log Avg loss: 0.30273824, Global Avg Loss: 2.32901664, Time: 0.0207 Steps: 16300, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001407, Sample Num: 22512, Cur Loss: 0.27956122, Cur Avg Loss: 0.48417983, Log Avg loss: 0.63316060, Global Avg Loss: 2.32797687, Time: 0.0207 Steps: 16310, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001417, Sample Num: 22672, Cur Loss: 0.37289205, Cur Avg Loss: 0.48606372, Log Avg loss: 0.75112695, Global Avg Loss: 2.32701067, Time: 0.0207 Steps: 16320, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001427, Sample Num: 22832, Cur Loss: 0.45005631, Cur Avg Loss: 0.48709132, Log Avg loss: 0.63270172, Global Avg Loss: 2.32597312, Time: 0.0207 Steps: 16330, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001437, Sample Num: 22992, Cur Loss: 0.18953806, Cur Avg Loss: 0.48689517, Log Avg loss: 0.45890553, Global Avg Loss: 2.32483049, Time: 0.0207 Steps: 16340, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001447, Sample Num: 23152, Cur Loss: 0.56894916, Cur Avg Loss: 0.48642976, Log Avg loss: 0.41954933, Global Avg Loss: 2.32366518, Time: 0.0207 Steps: 16350, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001457, Sample Num: 23312, Cur Loss: 1.07390463, Cur Avg Loss: 0.48717203, Log Avg loss: 0.59457939, Global Avg Loss: 2.32260828, Time: 0.0207 Steps: 16360, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001467, Sample Num: 23472, Cur Loss: 0.65189338, Cur Avg Loss: 0.48882338, Log Avg loss: 0.72942423, Global Avg Loss: 2.32163504, Time: 0.0207 Steps: 16370, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001477, Sample Num: 23632, Cur Loss: 1.21835673, Cur Avg Loss: 0.48925267, Log Avg loss: 0.55223018, Global Avg Loss: 2.32055482, Time: 0.0208 Steps: 16380, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001487, Sample Num: 23792, Cur Loss: 0.43715116, Cur Avg Loss: 0.48795334, Log Avg loss: 0.29604176, Global Avg Loss: 2.31931961, Time: 0.0207 Steps: 16390, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001497, Sample Num: 23952, Cur Loss: 0.12353404, Cur Avg Loss: 0.48731211, Log Avg loss: 0.39196222, Global Avg Loss: 2.31814439, Time: 0.0207 Steps: 16400, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001507, Sample Num: 24112, Cur Loss: 0.55136573, Cur Avg Loss: 0.48603996, Log Avg loss: 0.29559813, Global Avg Loss: 2.31691188, Time: 0.0207 Steps: 16410, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001517, Sample Num: 24272, Cur Loss: 0.34321871, Cur Avg Loss: 0.48459111, Log Avg loss: 0.26624960, Global Avg Loss: 2.31566300, Time: 0.0208 Steps: 16420, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001527, Sample Num: 24432, Cur Loss: 0.68575561, Cur Avg Loss: 0.48612624, Log Avg loss: 0.71900646, Global Avg Loss: 2.31469121, Time: 0.0207 Steps: 16430, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001537, Sample Num: 24592, Cur Loss: 0.28335845, Cur Avg Loss: 0.48537950, Log Avg loss: 0.37135186, Global Avg Loss: 2.31350913, Time: 0.0245 Steps: 16440, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001547, Sample Num: 24752, Cur Loss: 0.42640108, Cur Avg Loss: 0.48521390, Log Avg loss: 0.45976046, Global Avg Loss: 2.31238223, Time: 0.0209 Steps: 16450, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001557, Sample Num: 24912, Cur Loss: 0.37853342, Cur Avg Loss: 0.48429588, Log Avg loss: 0.34227859, Global Avg Loss: 2.31118533, Time: 0.0209 Steps: 16460, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001567, Sample Num: 25072, Cur Loss: 0.68927127, Cur Avg Loss: 0.48352292, Log Avg loss: 0.36317232, Global Avg Loss: 2.31000256, Time: 0.0209 Steps: 16470, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001577, Sample Num: 25232, Cur Loss: 0.34650397, Cur Avg Loss: 0.48409947, Log Avg loss: 0.57444486, Global Avg Loss: 2.30894943, Time: 0.0208 Steps: 16480, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001587, Sample Num: 25392, Cur Loss: 0.57368612, Cur Avg Loss: 0.48340909, Log Avg loss: 0.37453646, Global Avg Loss: 2.30777635, Time: 0.0209 Steps: 16490, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001597, Sample Num: 25552, Cur Loss: 0.15644136, Cur Avg Loss: 0.48338995, Log Avg loss: 0.48035293, Global Avg Loss: 2.30666882, Time: 0.0210 Steps: 16500, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001607, Sample Num: 25712, Cur Loss: 0.12217389, Cur Avg Loss: 0.48265657, Log Avg loss: 0.36553558, Global Avg Loss: 2.30549309, Time: 0.0209 Steps: 16510, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001617, Sample Num: 25872, Cur Loss: 0.57408929, Cur Avg Loss: 0.48173469, Log Avg loss: 0.33358891, Global Avg Loss: 2.30429944, Time: 0.0209 Steps: 16520, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001627, Sample Num: 26032, Cur Loss: 0.18825617, Cur Avg Loss: 0.48040543, Log Avg loss: 0.26546334, Global Avg Loss: 2.30306603, Time: 0.0209 Steps: 16530, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001637, Sample Num: 26192, Cur Loss: 0.19295861, Cur Avg Loss: 0.47970133, Log Avg loss: 0.36514498, Global Avg Loss: 2.30189437, Time: 0.0209 Steps: 16540, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001647, Sample Num: 26352, Cur Loss: 0.31529421, Cur Avg Loss: 0.48475428, Log Avg loss: 1.31192133, Global Avg Loss: 2.30129620, Time: 0.0209 Steps: 16550, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001657, Sample Num: 26512, Cur Loss: 0.42524719, Cur Avg Loss: 0.48413260, Log Avg loss: 0.38174221, Global Avg Loss: 2.30013705, Time: 0.0209 Steps: 16560, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001667, Sample Num: 26672, Cur Loss: 0.35053688, Cur Avg Loss: 0.48471739, Log Avg loss: 0.58161702, Global Avg Loss: 2.29909992, Time: 0.0209 Steps: 16570, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001677, Sample Num: 26832, Cur Loss: 0.63920617, Cur Avg Loss: 0.48471992, Log Avg loss: 0.48514136, Global Avg Loss: 2.29800585, Time: 0.0209 Steps: 16580, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001687, Sample Num: 26992, Cur Loss: 0.50813389, Cur Avg Loss: 0.48478270, Log Avg loss: 0.49531251, Global Avg Loss: 2.29691924, Time: 0.0209 Steps: 16590, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001697, Sample Num: 27152, Cur Loss: 0.12217553, Cur Avg Loss: 0.48355568, Log Avg loss: 0.27655696, Global Avg Loss: 2.29570215, Time: 0.0209 Steps: 16600, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001707, Sample Num: 27312, Cur Loss: 0.12976986, Cur Avg Loss: 0.48308555, Log Avg loss: 0.40330334, Global Avg Loss: 2.29456284, Time: 0.0209 Steps: 16610, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001717, Sample Num: 27472, Cur Loss: 0.24904808, Cur Avg Loss: 0.48192302, Log Avg loss: 0.28347909, Global Avg Loss: 2.29335280, Time: 0.0209 Steps: 16620, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001727, Sample Num: 27632, Cur Loss: 0.63191992, Cur Avg Loss: 0.48194019, Log Avg loss: 0.48488995, Global Avg Loss: 2.29226533, Time: 0.0209 Steps: 16630, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001737, Sample Num: 27792, Cur Loss: 0.60075933, Cur Avg Loss: 0.48195934, Log Avg loss: 0.48526486, Global Avg Loss: 2.29117940, Time: 0.0209 Steps: 16640, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001747, Sample Num: 27952, Cur Loss: 0.13955674, Cur Avg Loss: 0.48143702, Log Avg loss: 0.39071135, Global Avg Loss: 2.29003797, Time: 0.0209 Steps: 16650, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001757, Sample Num: 28112, Cur Loss: 0.23405088, Cur Avg Loss: 0.48113966, Log Avg loss: 0.42919048, Global Avg Loss: 2.28892102, Time: 0.0210 Steps: 16660, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001767, Sample Num: 28272, Cur Loss: 0.81187588, Cur Avg Loss: 0.48110578, Log Avg loss: 0.47515374, Global Avg Loss: 2.28783298, Time: 0.0209 Steps: 16670, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001777, Sample Num: 28432, Cur Loss: 0.43984151, Cur Avg Loss: 0.48056740, Log Avg loss: 0.38543541, Global Avg Loss: 2.28669245, Time: 0.0209 Steps: 16680, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001787, Sample Num: 28592, Cur Loss: 0.58505338, Cur Avg Loss: 0.48025348, Log Avg loss: 0.42446896, Global Avg Loss: 2.28557668, Time: 0.0210 Steps: 16690, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001797, Sample Num: 28752, Cur Loss: 0.30763939, Cur Avg Loss: 0.48021063, Log Avg loss: 0.47255327, Global Avg Loss: 2.28449103, Time: 0.0246 Steps: 16700, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001807, Sample Num: 28912, Cur Loss: 0.36393732, Cur Avg Loss: 0.47990681, Log Avg loss: 0.42531084, Global Avg Loss: 2.28337842, Time: 0.0210 Steps: 16710, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001817, Sample Num: 29072, Cur Loss: 0.15038429, Cur Avg Loss: 0.47879361, Log Avg loss: 0.27763879, Global Avg Loss: 2.28217881, Time: 0.0207 Steps: 16720, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001827, Sample Num: 29232, Cur Loss: 0.37283209, Cur Avg Loss: 0.47845399, Log Avg loss: 0.41674411, Global Avg Loss: 2.28106379, Time: 0.0209 Steps: 16730, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001837, Sample Num: 29392, Cur Loss: 0.47761774, Cur Avg Loss: 0.47827078, Log Avg loss: 0.44479904, Global Avg Loss: 2.27996686, Time: 0.0210 Steps: 16740, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001847, Sample Num: 29552, Cur Loss: 0.51090437, Cur Avg Loss: 0.47807551, Log Avg loss: 0.44220493, Global Avg Loss: 2.27886969, Time: 0.0209 Steps: 16750, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001857, Sample Num: 29712, Cur Loss: 0.27007347, Cur Avg Loss: 0.47865472, Log Avg loss: 0.58563436, Global Avg Loss: 2.27785940, Time: 0.0208 Steps: 16760, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001867, Sample Num: 29872, Cur Loss: 0.57532358, Cur Avg Loss: 0.47837820, Log Avg loss: 0.42702727, Global Avg Loss: 2.27675575, Time: 0.0209 Steps: 16770, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001877, Sample Num: 30032, Cur Loss: 0.27151003, Cur Avg Loss: 0.47756141, Log Avg loss: 0.32506780, Global Avg Loss: 2.27559264, Time: 0.0214 Steps: 16780, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001887, Sample Num: 30192, Cur Loss: 0.55390567, Cur Avg Loss: 0.47738797, Log Avg loss: 0.44483319, Global Avg Loss: 2.27450226, Time: 0.0213 Steps: 16790, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001897, Sample Num: 30352, Cur Loss: 0.55834734, Cur Avg Loss: 0.47662436, Log Avg loss: 0.33253149, Global Avg Loss: 2.27334632, Time: 0.0213 Steps: 16800, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001907, Sample Num: 30512, Cur Loss: 0.45023644, Cur Avg Loss: 0.47655466, Log Avg loss: 0.46333136, Global Avg Loss: 2.27226957, Time: 0.0213 Steps: 16810, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001917, Sample Num: 30672, Cur Loss: 0.31264454, Cur Avg Loss: 0.47791360, Log Avg loss: 0.73706478, Global Avg Loss: 2.27135685, Time: 0.0210 Steps: 16820, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001927, Sample Num: 30832, Cur Loss: 0.23513973, Cur Avg Loss: 0.47736547, Log Avg loss: 0.37228851, Global Avg Loss: 2.27022846, Time: 0.0213 Steps: 16830, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001937, Sample Num: 30992, Cur Loss: 0.58310753, Cur Avg Loss: 0.47713324, Log Avg loss: 0.43238206, Global Avg Loss: 2.26913711, Time: 0.0213 Steps: 16840, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001947, Sample Num: 31152, Cur Loss: 0.25291812, Cur Avg Loss: 0.47649515, Log Avg loss: 0.35289832, Global Avg Loss: 2.26799987, Time: 0.0212 Steps: 16850, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001957, Sample Num: 31312, Cur Loss: 0.47359294, Cur Avg Loss: 0.47674894, Log Avg loss: 0.52616067, Global Avg Loss: 2.26696675, Time: 0.0212 Steps: 16860, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001967, Sample Num: 31472, Cur Loss: 1.36897087, Cur Avg Loss: 0.47936193, Log Avg loss: 0.99072372, Global Avg Loss: 2.26621024, Time: 0.0213 Steps: 16870, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001977, Sample Num: 31632, Cur Loss: 0.72974139, Cur Avg Loss: 0.48030689, Log Avg loss: 0.66618209, Global Avg Loss: 2.26526235, Time: 0.0210 Steps: 16880, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001987, Sample Num: 31792, Cur Loss: 1.42557120, Cur Avg Loss: 0.48168163, Log Avg loss: 0.75346724, Global Avg Loss: 2.26436727, Time: 0.0210 Steps: 16890, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001997, Sample Num: 31952, Cur Loss: 0.30240387, Cur Avg Loss: 0.48165304, Log Avg loss: 0.47597197, Global Avg Loss: 2.26330905, Time: 0.0212 Steps: 16900, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002007, Sample Num: 32112, Cur Loss: 0.28735891, Cur Avg Loss: 0.48081603, Log Avg loss: 0.31366453, Global Avg Loss: 2.26215609, Time: 0.0210 Steps: 16910, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002017, Sample Num: 32272, Cur Loss: 0.55011880, Cur Avg Loss: 0.48076431, Log Avg loss: 0.47038498, Global Avg Loss: 2.26109713, Time: 0.0212 Steps: 16920, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002027, Sample Num: 32432, Cur Loss: 0.67952061, Cur Avg Loss: 0.48105569, Log Avg loss: 0.53982621, Global Avg Loss: 2.26008043, Time: 0.0213 Steps: 16930, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002037, Sample Num: 32592, Cur Loss: 0.16071717, Cur Avg Loss: 0.47997802, Log Avg loss: 0.26153399, Global Avg Loss: 2.25890065, Time: 0.0212 Steps: 16940, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002047, Sample Num: 32752, Cur Loss: 0.88588876, Cur Avg Loss: 0.47984352, Log Avg loss: 0.45244646, Global Avg Loss: 2.25783489, Time: 0.0210 Steps: 16950, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002057, Sample Num: 32912, Cur Loss: 0.92717862, Cur Avg Loss: 0.47938640, Log Avg loss: 0.38581449, Global Avg Loss: 2.25673111, Time: 0.0211 Steps: 16960, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002067, Sample Num: 33072, Cur Loss: 0.83484381, Cur Avg Loss: 0.47936612, Log Avg loss: 0.47519430, Global Avg Loss: 2.25568129, Time: 0.0211 Steps: 16970, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002077, Sample Num: 33232, Cur Loss: 0.34873384, Cur Avg Loss: 0.47850131, Log Avg loss: 0.29974536, Global Avg Loss: 2.25452939, Time: 0.0211 Steps: 16980, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002087, Sample Num: 33392, Cur Loss: 0.25422448, Cur Avg Loss: 0.47765147, Log Avg loss: 0.30113918, Global Avg Loss: 2.25337966, Time: 0.0211 Steps: 16990, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002097, Sample Num: 33552, Cur Loss: 0.12762721, Cur Avg Loss: 0.47672368, Log Avg loss: 0.28309292, Global Avg Loss: 2.25222067, Time: 0.0209 Steps: 17000, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002107, Sample Num: 33712, Cur Loss: 0.57262719, Cur Avg Loss: 0.47648320, Log Avg loss: 0.42605487, Global Avg Loss: 2.25114708, Time: 0.0210 Steps: 17010, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002117, Sample Num: 33872, Cur Loss: 0.26689231, Cur Avg Loss: 0.47544723, Log Avg loss: 0.25716944, Global Avg Loss: 2.24997553, Time: 0.0209 Steps: 17020, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002127, Sample Num: 34032, Cur Loss: 0.20389222, Cur Avg Loss: 0.47474242, Log Avg loss: 0.32553439, Global Avg Loss: 2.24884550, Time: 0.0209 Steps: 17030, Updated lr: 0.000085 ***** Running evaluation checkpoint-17032 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-17032 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.574108, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.373622, "eval_total_loss": 262.656051, "eval_mae": 0.478143, "eval_mse": 0.373632, "eval_r2": 0.762495, "eval_sp_statistic": 0.859206, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.891331, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.465036, "test_total_loss": 233.447885, "test_mae": 0.405813, "test_mse": 0.465206, "test_r2": 0.699752, "test_sp_statistic": 0.855639, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.920402, "test_ps_pvalue": 0.0, "lr": 8.479658605974395e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.2486148470371417, "train_cur_epoch_loss": 1010.346297480166, "train_cur_epoch_avg_loss": 0.47456378463136023, "train_cur_epoch_time": 44.574108362197876, "train_cur_epoch_avg_time": 0.020936640846499707, "epoch": 8, "step": 17032} ################################################## Training, Epoch: 0009, Batch: 000008, Sample Num: 128, Cur Loss: 0.46296513, Cur Avg Loss: 0.43673434, Log Avg loss: 0.40630382, Global Avg Loss: 2.24776420, Time: 0.0210 Steps: 17040, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000018, Sample Num: 288, Cur Loss: 0.29193187, Cur Avg Loss: 0.36549510, Log Avg loss: 0.30850371, Global Avg Loss: 2.24662680, Time: 0.0210 Steps: 17050, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000028, Sample Num: 448, Cur Loss: 0.64559078, Cur Avg Loss: 0.34718563, Log Avg loss: 0.31422858, Global Avg Loss: 2.24549410, Time: 0.0210 Steps: 17060, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000038, Sample Num: 608, Cur Loss: 0.50282657, Cur Avg Loss: 0.33742436, Log Avg loss: 0.31009281, Global Avg Loss: 2.24436029, Time: 0.0210 Steps: 17070, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000048, Sample Num: 768, Cur Loss: 0.61304981, Cur Avg Loss: 0.38116381, Log Avg loss: 0.54737373, Global Avg Loss: 2.24336674, Time: 0.0210 Steps: 17080, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000058, Sample Num: 928, Cur Loss: 0.23193620, Cur Avg Loss: 0.38988859, Log Avg loss: 0.43176754, Global Avg Loss: 2.24230671, Time: 0.0210 Steps: 17090, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000068, Sample Num: 1088, Cur Loss: 0.21961093, Cur Avg Loss: 0.37617503, Log Avg loss: 0.29663633, Global Avg Loss: 2.24116889, Time: 0.0210 Steps: 17100, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000078, Sample Num: 1248, Cur Loss: 0.25621960, Cur Avg Loss: 0.37842064, Log Avg loss: 0.39369078, Global Avg Loss: 2.24008912, Time: 0.0210 Steps: 17110, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000088, Sample Num: 1408, Cur Loss: 1.08923352, Cur Avg Loss: 0.40463876, Log Avg loss: 0.60914010, Global Avg Loss: 2.23913647, Time: 0.0210 Steps: 17120, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000098, Sample Num: 1568, Cur Loss: 0.17738055, Cur Avg Loss: 0.45649661, Log Avg loss: 0.91284573, Global Avg Loss: 2.23836221, Time: 0.0210 Steps: 17130, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000108, Sample Num: 1728, Cur Loss: 0.33052915, Cur Avg Loss: 0.45180465, Log Avg loss: 0.40582338, Global Avg Loss: 2.23729306, Time: 0.0210 Steps: 17140, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000118, Sample Num: 1888, Cur Loss: 0.61290056, Cur Avg Loss: 0.45775955, Log Avg loss: 0.52207256, Global Avg Loss: 2.23629293, Time: 0.0210 Steps: 17150, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000128, Sample Num: 2048, Cur Loss: 0.35554028, Cur Avg Loss: 0.44413826, Log Avg loss: 0.28340704, Global Avg Loss: 2.23515488, Time: 0.0210 Steps: 17160, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000138, Sample Num: 2208, Cur Loss: 0.26770359, Cur Avg Loss: 0.43967774, Log Avg loss: 0.38258299, Global Avg Loss: 2.23407592, Time: 0.0210 Steps: 17170, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000148, Sample Num: 2368, Cur Loss: 0.22590998, Cur Avg Loss: 0.44135009, Log Avg loss: 0.46442856, Global Avg Loss: 2.23304586, Time: 0.0210 Steps: 17180, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000158, Sample Num: 2528, Cur Loss: 0.38882107, Cur Avg Loss: 0.43777851, Log Avg loss: 0.38491912, Global Avg Loss: 2.23197074, Time: 0.0210 Steps: 17190, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000168, Sample Num: 2688, Cur Loss: 0.33592030, Cur Avg Loss: 0.44129210, Log Avg loss: 0.49680692, Global Avg Loss: 2.23096193, Time: 0.0210 Steps: 17200, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000178, Sample Num: 2848, Cur Loss: 1.12037432, Cur Avg Loss: 0.44082089, Log Avg loss: 0.43290443, Global Avg Loss: 2.22991715, Time: 0.0210 Steps: 17210, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000188, Sample Num: 3008, Cur Loss: 0.32064635, Cur Avg Loss: 0.43504995, Log Avg loss: 0.33232734, Global Avg Loss: 2.22881518, Time: 0.0210 Steps: 17220, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000198, Sample Num: 3168, Cur Loss: 0.76103765, Cur Avg Loss: 0.44021739, Log Avg loss: 0.53736512, Global Avg Loss: 2.22783349, Time: 0.0210 Steps: 17230, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000208, Sample Num: 3328, Cur Loss: 1.42197978, Cur Avg Loss: 0.46828756, Log Avg loss: 1.02407696, Global Avg Loss: 2.22713526, Time: 0.0210 Steps: 17240, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000218, Sample Num: 3488, Cur Loss: 1.14672351, Cur Avg Loss: 0.47840977, Log Avg loss: 0.68895168, Global Avg Loss: 2.22624356, Time: 0.0210 Steps: 17250, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000228, Sample Num: 3648, Cur Loss: 0.43357062, Cur Avg Loss: 0.48993719, Log Avg loss: 0.74123497, Global Avg Loss: 2.22538318, Time: 0.0210 Steps: 17260, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000238, Sample Num: 3808, Cur Loss: 0.35466471, Cur Avg Loss: 0.48567211, Log Avg loss: 0.38842825, Global Avg Loss: 2.22431952, Time: 0.0210 Steps: 17270, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000248, Sample Num: 3968, Cur Loss: 0.34123495, Cur Avg Loss: 0.47904356, Log Avg loss: 0.32128412, Global Avg Loss: 2.22321822, Time: 0.0210 Steps: 17280, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000258, Sample Num: 4128, Cur Loss: 0.75676209, Cur Avg Loss: 0.47732744, Log Avg loss: 0.43476759, Global Avg Loss: 2.22218384, Time: 0.0257 Steps: 17290, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000268, Sample Num: 4288, Cur Loss: 0.24946439, Cur Avg Loss: 0.47160240, Log Avg loss: 0.32389645, Global Avg Loss: 2.22108656, Time: 0.0219 Steps: 17300, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000278, Sample Num: 4448, Cur Loss: 0.36972645, Cur Avg Loss: 0.46566942, Log Avg loss: 0.30666568, Global Avg Loss: 2.21998060, Time: 0.0220 Steps: 17310, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000288, Sample Num: 4608, Cur Loss: 0.09436107, Cur Avg Loss: 0.46054037, Log Avg loss: 0.31795283, Global Avg Loss: 2.21888243, Time: 0.0219 Steps: 17320, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000298, Sample Num: 4768, Cur Loss: 0.24446088, Cur Avg Loss: 0.45548513, Log Avg loss: 0.30989400, Global Avg Loss: 2.21778088, Time: 0.0219 Steps: 17330, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000308, Sample Num: 4928, Cur Loss: 0.05760680, Cur Avg Loss: 0.45005405, Log Avg loss: 0.28820785, Global Avg Loss: 2.21666809, Time: 0.0219 Steps: 17340, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000318, Sample Num: 5088, Cur Loss: 0.21158943, Cur Avg Loss: 0.44477567, Log Avg loss: 0.28220170, Global Avg Loss: 2.21555313, Time: 0.0220 Steps: 17350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000328, Sample Num: 5248, Cur Loss: 0.90919769, Cur Avg Loss: 0.44091857, Log Avg loss: 0.31826280, Global Avg Loss: 2.21446022, Time: 0.0219 Steps: 17360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000338, Sample Num: 5408, Cur Loss: 0.31398076, Cur Avg Loss: 0.43991556, Log Avg loss: 0.40701663, Global Avg Loss: 2.21341966, Time: 0.0208 Steps: 17370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000348, Sample Num: 5568, Cur Loss: 0.14930058, Cur Avg Loss: 0.44035516, Log Avg loss: 0.45521383, Global Avg Loss: 2.21240804, Time: 0.0208 Steps: 17380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000358, Sample Num: 5728, Cur Loss: 0.21177471, Cur Avg Loss: 0.43528505, Log Avg loss: 0.25884533, Global Avg Loss: 2.21128465, Time: 0.0208 Steps: 17390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000368, Sample Num: 5888, Cur Loss: 0.31094584, Cur Avg Loss: 0.43625030, Log Avg loss: 0.47080628, Global Avg Loss: 2.21028438, Time: 0.0208 Steps: 17400, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000378, Sample Num: 6048, Cur Loss: 0.25570953, Cur Avg Loss: 0.43239356, Log Avg loss: 0.29046551, Global Avg Loss: 2.20918167, Time: 0.0209 Steps: 17410, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000388, Sample Num: 6208, Cur Loss: 0.11950660, Cur Avg Loss: 0.43014139, Log Avg loss: 0.34500924, Global Avg Loss: 2.20811153, Time: 0.0208 Steps: 17420, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000398, Sample Num: 6368, Cur Loss: 0.27071154, Cur Avg Loss: 0.43064450, Log Avg loss: 0.45016526, Global Avg Loss: 2.20710296, Time: 0.0209 Steps: 17430, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000408, Sample Num: 6528, Cur Loss: 0.14791596, Cur Avg Loss: 0.42934357, Log Avg loss: 0.37756635, Global Avg Loss: 2.20605391, Time: 0.0208 Steps: 17440, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000418, Sample Num: 6688, Cur Loss: 0.30547786, Cur Avg Loss: 0.42647657, Log Avg loss: 0.30950287, Global Avg Loss: 2.20496706, Time: 0.0208 Steps: 17450, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000428, Sample Num: 6848, Cur Loss: 0.23963082, Cur Avg Loss: 0.42548645, Log Avg loss: 0.38409959, Global Avg Loss: 2.20392419, Time: 0.0208 Steps: 17460, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000438, Sample Num: 7008, Cur Loss: 0.16175140, Cur Avg Loss: 0.42259591, Log Avg loss: 0.29888069, Global Avg Loss: 2.20283372, Time: 0.0208 Steps: 17470, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000448, Sample Num: 7168, Cur Loss: 0.08683850, Cur Avg Loss: 0.41833494, Log Avg loss: 0.23170470, Global Avg Loss: 2.20170607, Time: 0.0209 Steps: 17480, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000458, Sample Num: 7328, Cur Loss: 0.40255159, Cur Avg Loss: 0.41857857, Log Avg loss: 0.42949294, Global Avg Loss: 2.20069280, Time: 0.0208 Steps: 17490, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000468, Sample Num: 7488, Cur Loss: 0.67902285, Cur Avg Loss: 0.41899835, Log Avg loss: 0.43822464, Global Avg Loss: 2.19968567, Time: 0.0209 Steps: 17500, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000478, Sample Num: 7648, Cur Loss: 0.34600875, Cur Avg Loss: 0.42944927, Log Avg loss: 0.91855234, Global Avg Loss: 2.19895402, Time: 0.0209 Steps: 17510, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000488, Sample Num: 7808, Cur Loss: 0.77915251, Cur Avg Loss: 0.42893925, Log Avg loss: 0.40455985, Global Avg Loss: 2.19792982, Time: 0.0209 Steps: 17520, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000498, Sample Num: 7968, Cur Loss: 0.24028918, Cur Avg Loss: 0.42941246, Log Avg loss: 0.45250509, Global Avg Loss: 2.19693414, Time: 0.0209 Steps: 17530, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000508, Sample Num: 8128, Cur Loss: 0.35078380, Cur Avg Loss: 0.42834311, Log Avg loss: 0.37508972, Global Avg Loss: 2.19589546, Time: 0.0208 Steps: 17540, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000518, Sample Num: 8288, Cur Loss: 0.26658392, Cur Avg Loss: 0.42766544, Log Avg loss: 0.39323978, Global Avg Loss: 2.19486831, Time: 0.0211 Steps: 17550, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000528, Sample Num: 8448, Cur Loss: 0.10806458, Cur Avg Loss: 0.43017571, Log Avg loss: 0.56020790, Global Avg Loss: 2.19393741, Time: 0.0210 Steps: 17560, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000538, Sample Num: 8608, Cur Loss: 0.25529128, Cur Avg Loss: 0.42918029, Log Avg loss: 0.37662182, Global Avg Loss: 2.19290308, Time: 0.0210 Steps: 17570, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000548, Sample Num: 8768, Cur Loss: 0.19974333, Cur Avg Loss: 0.42768312, Log Avg loss: 0.34713568, Global Avg Loss: 2.19185315, Time: 0.0209 Steps: 17580, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000558, Sample Num: 8928, Cur Loss: 0.92077684, Cur Avg Loss: 0.42635798, Log Avg loss: 0.35373993, Global Avg Loss: 2.19080818, Time: 0.0210 Steps: 17590, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000568, Sample Num: 9088, Cur Loss: 0.53477263, Cur Avg Loss: 0.42607947, Log Avg loss: 0.41053895, Global Avg Loss: 2.18979666, Time: 0.0210 Steps: 17600, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000578, Sample Num: 9248, Cur Loss: 0.38543606, Cur Avg Loss: 0.42644141, Log Avg loss: 0.44699925, Global Avg Loss: 2.18880700, Time: 0.0210 Steps: 17610, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000588, Sample Num: 9408, Cur Loss: 0.21248943, Cur Avg Loss: 0.42464331, Log Avg loss: 0.32071310, Global Avg Loss: 2.18774678, Time: 0.0209 Steps: 17620, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000598, Sample Num: 9568, Cur Loss: 0.79193258, Cur Avg Loss: 0.42516257, Log Avg loss: 0.45569515, Global Avg Loss: 2.18676434, Time: 0.0209 Steps: 17630, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000608, Sample Num: 9728, Cur Loss: 0.08839314, Cur Avg Loss: 0.42291434, Log Avg loss: 0.28847025, Global Avg Loss: 2.18568821, Time: 0.0209 Steps: 17640, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000618, Sample Num: 9888, Cur Loss: 0.30891791, Cur Avg Loss: 0.42248576, Log Avg loss: 0.39642831, Global Avg Loss: 2.18467446, Time: 0.0209 Steps: 17650, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000628, Sample Num: 10048, Cur Loss: 0.35704207, Cur Avg Loss: 0.42573375, Log Avg loss: 0.62645937, Global Avg Loss: 2.18379212, Time: 0.0210 Steps: 17660, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000638, Sample Num: 10208, Cur Loss: 0.24794005, Cur Avg Loss: 0.42502074, Log Avg loss: 0.38024369, Global Avg Loss: 2.18277144, Time: 0.0210 Steps: 17670, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000648, Sample Num: 10368, Cur Loss: 0.12259632, Cur Avg Loss: 0.42610417, Log Avg loss: 0.49522727, Global Avg Loss: 2.18181694, Time: 0.0209 Steps: 17680, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000658, Sample Num: 10528, Cur Loss: 0.73828775, Cur Avg Loss: 0.42863938, Log Avg loss: 0.59292094, Global Avg Loss: 2.18091876, Time: 0.0209 Steps: 17690, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000668, Sample Num: 10688, Cur Loss: 0.23069462, Cur Avg Loss: 0.42864618, Log Avg loss: 0.42909377, Global Avg Loss: 2.17992902, Time: 0.0210 Steps: 17700, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000678, Sample Num: 10848, Cur Loss: 0.18260238, Cur Avg Loss: 0.42764603, Log Avg loss: 0.36083566, Global Avg Loss: 2.17890187, Time: 0.0209 Steps: 17710, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000688, Sample Num: 11008, Cur Loss: 1.29022956, Cur Avg Loss: 0.43179002, Log Avg loss: 0.71275259, Global Avg Loss: 2.17807447, Time: 0.0209 Steps: 17720, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000698, Sample Num: 11168, Cur Loss: 0.20335273, Cur Avg Loss: 0.43332583, Log Avg loss: 0.53898974, Global Avg Loss: 2.17715000, Time: 0.0209 Steps: 17730, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000708, Sample Num: 11328, Cur Loss: 0.22823051, Cur Avg Loss: 0.43127627, Log Avg loss: 0.28821712, Global Avg Loss: 2.17608521, Time: 0.0209 Steps: 17740, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000718, Sample Num: 11488, Cur Loss: 0.08079273, Cur Avg Loss: 0.42936413, Log Avg loss: 0.29398413, Global Avg Loss: 2.17502487, Time: 0.0210 Steps: 17750, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000728, Sample Num: 11648, Cur Loss: 0.31861162, Cur Avg Loss: 0.42832430, Log Avg loss: 0.35366440, Global Avg Loss: 2.17399933, Time: 0.0209 Steps: 17760, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000738, Sample Num: 11808, Cur Loss: 0.17496175, Cur Avg Loss: 0.42609514, Log Avg loss: 0.26381233, Global Avg Loss: 2.17292438, Time: 0.0210 Steps: 17770, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000748, Sample Num: 11968, Cur Loss: 0.38877144, Cur Avg Loss: 0.42777973, Log Avg loss: 0.55210289, Global Avg Loss: 2.17201278, Time: 0.0209 Steps: 17780, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000758, Sample Num: 12128, Cur Loss: 0.47314662, Cur Avg Loss: 0.42700535, Log Avg loss: 0.36908183, Global Avg Loss: 2.17099933, Time: 0.0210 Steps: 17790, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000768, Sample Num: 12288, Cur Loss: 0.34784859, Cur Avg Loss: 0.42712859, Log Avg loss: 0.43646959, Global Avg Loss: 2.17002488, Time: 0.0255 Steps: 17800, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000778, Sample Num: 12448, Cur Loss: 0.35482943, Cur Avg Loss: 0.42636298, Log Avg loss: 0.36756451, Global Avg Loss: 2.16901283, Time: 0.0221 Steps: 17810, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000788, Sample Num: 12608, Cur Loss: 0.48340040, Cur Avg Loss: 0.42425576, Log Avg loss: 0.26031406, Global Avg Loss: 2.16794173, Time: 0.0220 Steps: 17820, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000798, Sample Num: 12768, Cur Loss: 0.50206894, Cur Avg Loss: 0.42266844, Log Avg loss: 0.29758756, Global Avg Loss: 2.16689274, Time: 0.0220 Steps: 17830, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000808, Sample Num: 12928, Cur Loss: 0.26919112, Cur Avg Loss: 0.42119349, Log Avg loss: 0.30349207, Global Avg Loss: 2.16584823, Time: 0.0220 Steps: 17840, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000818, Sample Num: 13088, Cur Loss: 0.32403651, Cur Avg Loss: 0.42097467, Log Avg loss: 0.40329399, Global Avg Loss: 2.16486080, Time: 0.0219 Steps: 17850, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000828, Sample Num: 13248, Cur Loss: 0.53018743, Cur Avg Loss: 0.41963262, Log Avg loss: 0.30985360, Global Avg Loss: 2.16382217, Time: 0.0219 Steps: 17860, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000838, Sample Num: 13408, Cur Loss: 0.37385696, Cur Avg Loss: 0.41865626, Log Avg loss: 0.33781361, Global Avg Loss: 2.16280034, Time: 0.0220 Steps: 17870, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000848, Sample Num: 13568, Cur Loss: 0.26721546, Cur Avg Loss: 0.41816366, Log Avg loss: 0.37688395, Global Avg Loss: 2.16180150, Time: 0.0219 Steps: 17880, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000858, Sample Num: 13728, Cur Loss: 0.12048778, Cur Avg Loss: 0.41738824, Log Avg loss: 0.35163182, Global Avg Loss: 2.16078967, Time: 0.0220 Steps: 17890, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000868, Sample Num: 13888, Cur Loss: 0.57659793, Cur Avg Loss: 0.41605506, Log Avg loss: 0.30166822, Global Avg Loss: 2.15975105, Time: 0.0220 Steps: 17900, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000878, Sample Num: 14048, Cur Loss: 0.24365489, Cur Avg Loss: 0.41564946, Log Avg loss: 0.38044372, Global Avg Loss: 2.15875758, Time: 0.0220 Steps: 17910, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000888, Sample Num: 14208, Cur Loss: 0.48892438, Cur Avg Loss: 0.41614416, Log Avg loss: 0.45957848, Global Avg Loss: 2.15780938, Time: 0.0219 Steps: 17920, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000898, Sample Num: 14368, Cur Loss: 0.59117395, Cur Avg Loss: 0.41710052, Log Avg loss: 0.50202548, Global Avg Loss: 2.15688591, Time: 0.0220 Steps: 17930, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000908, Sample Num: 14528, Cur Loss: 0.74055201, Cur Avg Loss: 0.41775552, Log Avg loss: 0.47657511, Global Avg Loss: 2.15594928, Time: 0.0220 Steps: 17940, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000918, Sample Num: 14688, Cur Loss: 0.34461600, Cur Avg Loss: 0.41677783, Log Avg loss: 0.32800321, Global Avg Loss: 2.15493093, Time: 0.0220 Steps: 17950, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000928, Sample Num: 14848, Cur Loss: 0.50880510, Cur Avg Loss: 0.41688277, Log Avg loss: 0.42651631, Global Avg Loss: 2.15396856, Time: 0.0219 Steps: 17960, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000938, Sample Num: 15008, Cur Loss: 1.11993337, Cur Avg Loss: 0.41828426, Log Avg loss: 0.54834277, Global Avg Loss: 2.15307505, Time: 0.0220 Steps: 17970, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000948, Sample Num: 15168, Cur Loss: 0.25001585, Cur Avg Loss: 0.41805325, Log Avg loss: 0.39638412, Global Avg Loss: 2.15209803, Time: 0.0220 Steps: 17980, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000958, Sample Num: 15328, Cur Loss: 0.34720573, Cur Avg Loss: 0.41667419, Log Avg loss: 0.28593896, Global Avg Loss: 2.15106070, Time: 0.0220 Steps: 17990, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000968, Sample Num: 15488, Cur Loss: 0.23670869, Cur Avg Loss: 0.41625316, Log Avg loss: 0.37591858, Global Avg Loss: 2.15007451, Time: 0.0220 Steps: 18000, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000978, Sample Num: 15648, Cur Loss: 0.19215597, Cur Avg Loss: 0.41497143, Log Avg loss: 0.29090024, Global Avg Loss: 2.14904221, Time: 0.0219 Steps: 18010, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000988, Sample Num: 15808, Cur Loss: 0.53578031, Cur Avg Loss: 0.41487393, Log Avg loss: 0.40533889, Global Avg Loss: 2.14807456, Time: 0.0219 Steps: 18020, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000998, Sample Num: 15968, Cur Loss: 0.45482272, Cur Avg Loss: 0.41444074, Log Avg loss: 0.37164155, Global Avg Loss: 2.14708929, Time: 0.0220 Steps: 18030, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001008, Sample Num: 16128, Cur Loss: 0.51967752, Cur Avg Loss: 0.41386749, Log Avg loss: 0.35665665, Global Avg Loss: 2.14609681, Time: 0.0219 Steps: 18040, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001018, Sample Num: 16288, Cur Loss: 0.38390493, Cur Avg Loss: 0.41352259, Log Avg loss: 0.37875697, Global Avg Loss: 2.14511768, Time: 0.0219 Steps: 18050, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001028, Sample Num: 16448, Cur Loss: 0.50229597, Cur Avg Loss: 0.41270367, Log Avg loss: 0.32933736, Global Avg Loss: 2.14411226, Time: 0.0211 Steps: 18060, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001038, Sample Num: 16608, Cur Loss: 0.84870410, Cur Avg Loss: 0.41459883, Log Avg loss: 0.60942131, Global Avg Loss: 2.14326296, Time: 0.0210 Steps: 18070, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001048, Sample Num: 16768, Cur Loss: 0.39752281, Cur Avg Loss: 0.41414032, Log Avg loss: 0.36654720, Global Avg Loss: 2.14228026, Time: 0.0210 Steps: 18080, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001058, Sample Num: 16928, Cur Loss: 0.40910304, Cur Avg Loss: 0.41510280, Log Avg loss: 0.51597040, Global Avg Loss: 2.14138125, Time: 0.0211 Steps: 18090, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001068, Sample Num: 17088, Cur Loss: 0.41941082, Cur Avg Loss: 0.41810806, Log Avg loss: 0.73606473, Global Avg Loss: 2.14060483, Time: 0.0211 Steps: 18100, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001078, Sample Num: 17248, Cur Loss: 0.24891251, Cur Avg Loss: 0.41784343, Log Avg loss: 0.38958039, Global Avg Loss: 2.13963795, Time: 0.0210 Steps: 18110, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001088, Sample Num: 17408, Cur Loss: 0.37763113, Cur Avg Loss: 0.41700297, Log Avg loss: 0.32640202, Global Avg Loss: 2.13863727, Time: 0.0211 Steps: 18120, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001098, Sample Num: 17568, Cur Loss: 0.25275490, Cur Avg Loss: 0.41638369, Log Avg loss: 0.34900544, Global Avg Loss: 2.13765016, Time: 0.0210 Steps: 18130, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001108, Sample Num: 17728, Cur Loss: 0.99591219, Cur Avg Loss: 0.41743766, Log Avg loss: 0.53316354, Global Avg Loss: 2.13676566, Time: 0.0211 Steps: 18140, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001118, Sample Num: 17888, Cur Loss: 0.33428031, Cur Avg Loss: 0.41681966, Log Avg loss: 0.34834533, Global Avg Loss: 2.13578030, Time: 0.0210 Steps: 18150, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001128, Sample Num: 18048, Cur Loss: 0.47855103, Cur Avg Loss: 0.41542407, Log Avg loss: 0.25939737, Global Avg Loss: 2.13474705, Time: 0.0211 Steps: 18160, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001138, Sample Num: 18208, Cur Loss: 0.35154945, Cur Avg Loss: 0.41570666, Log Avg loss: 0.44758241, Global Avg Loss: 2.13381851, Time: 0.0210 Steps: 18170, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001148, Sample Num: 18368, Cur Loss: 0.31890827, Cur Avg Loss: 0.41523568, Log Avg loss: 0.36163910, Global Avg Loss: 2.13284371, Time: 0.0211 Steps: 18180, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001158, Sample Num: 18528, Cur Loss: 0.25870374, Cur Avg Loss: 0.41467730, Log Avg loss: 0.35057516, Global Avg Loss: 2.13186390, Time: 0.0210 Steps: 18190, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001168, Sample Num: 18688, Cur Loss: 0.16667432, Cur Avg Loss: 0.41414709, Log Avg loss: 0.35274883, Global Avg Loss: 2.13088637, Time: 0.0210 Steps: 18200, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001178, Sample Num: 18848, Cur Loss: 0.38928825, Cur Avg Loss: 0.41351478, Log Avg loss: 0.33966072, Global Avg Loss: 2.12990272, Time: 0.0210 Steps: 18210, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001188, Sample Num: 19008, Cur Loss: 0.22332682, Cur Avg Loss: 0.41308317, Log Avg loss: 0.36223909, Global Avg Loss: 2.12893254, Time: 0.0211 Steps: 18220, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001198, Sample Num: 19168, Cur Loss: 0.30313742, Cur Avg Loss: 0.41239076, Log Avg loss: 0.33013255, Global Avg Loss: 2.12794581, Time: 0.0210 Steps: 18230, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001208, Sample Num: 19328, Cur Loss: 0.18474078, Cur Avg Loss: 0.41209870, Log Avg loss: 0.37710964, Global Avg Loss: 2.12698593, Time: 0.0211 Steps: 18240, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001218, Sample Num: 19488, Cur Loss: 0.26307672, Cur Avg Loss: 0.41137563, Log Avg loss: 0.32402874, Global Avg Loss: 2.12599800, Time: 0.0210 Steps: 18250, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001228, Sample Num: 19648, Cur Loss: 0.27179581, Cur Avg Loss: 0.41180162, Log Avg loss: 0.46368742, Global Avg Loss: 2.12508765, Time: 0.0211 Steps: 18260, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001238, Sample Num: 19808, Cur Loss: 0.80258405, Cur Avg Loss: 0.41302195, Log Avg loss: 0.56287857, Global Avg Loss: 2.12423258, Time: 0.0210 Steps: 18270, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001248, Sample Num: 19968, Cur Loss: 0.77887380, Cur Avg Loss: 0.41437855, Log Avg loss: 0.58232558, Global Avg Loss: 2.12338909, Time: 0.0208 Steps: 18280, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001258, Sample Num: 20128, Cur Loss: 1.27673256, Cur Avg Loss: 0.41613636, Log Avg loss: 0.63551087, Global Avg Loss: 2.12257559, Time: 0.0208 Steps: 18290, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001268, Sample Num: 20288, Cur Loss: 0.22297010, Cur Avg Loss: 0.41503268, Log Avg loss: 0.27618995, Global Avg Loss: 2.12156664, Time: 0.0208 Steps: 18300, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001278, Sample Num: 20448, Cur Loss: 0.21924710, Cur Avg Loss: 0.41356527, Log Avg loss: 0.22749819, Global Avg Loss: 2.12053220, Time: 0.0208 Steps: 18310, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001288, Sample Num: 20608, Cur Loss: 0.14367908, Cur Avg Loss: 0.41246601, Log Avg loss: 0.27198048, Global Avg Loss: 2.11952316, Time: 0.0210 Steps: 18320, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001298, Sample Num: 20768, Cur Loss: 0.50847352, Cur Avg Loss: 0.41268630, Log Avg loss: 0.44105970, Global Avg Loss: 2.11860747, Time: 0.0208 Steps: 18330, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001308, Sample Num: 20928, Cur Loss: 0.63216823, Cur Avg Loss: 0.41476961, Log Avg loss: 0.68518316, Global Avg Loss: 2.11782588, Time: 0.0210 Steps: 18340, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001318, Sample Num: 21088, Cur Loss: 0.10078040, Cur Avg Loss: 0.41506970, Log Avg loss: 0.45432097, Global Avg Loss: 2.11691934, Time: 0.0210 Steps: 18350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001328, Sample Num: 21248, Cur Loss: 0.48412284, Cur Avg Loss: 0.41539972, Log Avg loss: 0.45889707, Global Avg Loss: 2.11601628, Time: 0.0210 Steps: 18360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001338, Sample Num: 21408, Cur Loss: 0.27178916, Cur Avg Loss: 0.41556427, Log Avg loss: 0.43741651, Global Avg Loss: 2.11510251, Time: 0.0211 Steps: 18370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001348, Sample Num: 21568, Cur Loss: 0.37804106, Cur Avg Loss: 0.41656993, Log Avg loss: 0.55112694, Global Avg Loss: 2.11425160, Time: 0.0210 Steps: 18380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001358, Sample Num: 21728, Cur Loss: 0.38977915, Cur Avg Loss: 0.41874665, Log Avg loss: 0.71216832, Global Avg Loss: 2.11348918, Time: 0.0211 Steps: 18390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001368, Sample Num: 21888, Cur Loss: 0.36490288, Cur Avg Loss: 0.41893623, Log Avg loss: 0.44468081, Global Avg Loss: 2.11258222, Time: 0.0210 Steps: 18400, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001378, Sample Num: 22048, Cur Loss: 0.58668005, Cur Avg Loss: 0.41931057, Log Avg loss: 0.47052089, Global Avg Loss: 2.11169028, Time: 0.0210 Steps: 18410, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001388, Sample Num: 22208, Cur Loss: 0.69881719, Cur Avg Loss: 0.41917250, Log Avg loss: 0.40014705, Global Avg Loss: 2.11076110, Time: 0.0210 Steps: 18420, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001398, Sample Num: 22368, Cur Loss: 0.06306368, Cur Avg Loss: 0.41874496, Log Avg loss: 0.35940161, Global Avg Loss: 2.10981083, Time: 0.0211 Steps: 18430, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001408, Sample Num: 22528, Cur Loss: 0.34997600, Cur Avg Loss: 0.41774859, Log Avg loss: 0.27845687, Global Avg Loss: 2.10881768, Time: 0.0210 Steps: 18440, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001418, Sample Num: 22688, Cur Loss: 0.12003866, Cur Avg Loss: 0.41737870, Log Avg loss: 0.36529728, Global Avg Loss: 2.10787269, Time: 0.0210 Steps: 18450, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001428, Sample Num: 22848, Cur Loss: 0.24480450, Cur Avg Loss: 0.41665429, Log Avg loss: 0.31393328, Global Avg Loss: 2.10690089, Time: 0.0210 Steps: 18460, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001438, Sample Num: 23008, Cur Loss: 0.53875828, Cur Avg Loss: 0.41561938, Log Avg loss: 0.26783357, Global Avg Loss: 2.10590518, Time: 0.0211 Steps: 18470, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001448, Sample Num: 23168, Cur Loss: 0.69630945, Cur Avg Loss: 0.41591064, Log Avg loss: 0.45779382, Global Avg Loss: 2.10501335, Time: 0.0210 Steps: 18480, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001458, Sample Num: 23328, Cur Loss: 0.26437971, Cur Avg Loss: 0.41555822, Log Avg loss: 0.36452871, Global Avg Loss: 2.10407204, Time: 0.0211 Steps: 18490, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001468, Sample Num: 23488, Cur Loss: 0.50533110, Cur Avg Loss: 0.41536011, Log Avg loss: 0.38647576, Global Avg Loss: 2.10314361, Time: 0.0210 Steps: 18500, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001478, Sample Num: 23648, Cur Loss: 0.42248306, Cur Avg Loss: 0.41412805, Log Avg loss: 0.23326088, Global Avg Loss: 2.10213341, Time: 0.0211 Steps: 18510, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001488, Sample Num: 23808, Cur Loss: 0.16577157, Cur Avg Loss: 0.41388638, Log Avg loss: 0.37816828, Global Avg Loss: 2.10120254, Time: 0.0210 Steps: 18520, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001498, Sample Num: 23968, Cur Loss: 0.23548865, Cur Avg Loss: 0.41432615, Log Avg loss: 0.47976303, Global Avg Loss: 2.10032750, Time: 0.0211 Steps: 18530, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001508, Sample Num: 24128, Cur Loss: 0.39151227, Cur Avg Loss: 0.41410998, Log Avg loss: 0.38172807, Global Avg Loss: 2.09940054, Time: 0.0210 Steps: 18540, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001518, Sample Num: 24288, Cur Loss: 0.23974121, Cur Avg Loss: 0.41441977, Log Avg loss: 0.46113604, Global Avg Loss: 2.09851737, Time: 0.0211 Steps: 18550, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001528, Sample Num: 24448, Cur Loss: 0.29626858, Cur Avg Loss: 0.41403120, Log Avg loss: 0.35504695, Global Avg Loss: 2.09757800, Time: 0.0210 Steps: 18560, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001538, Sample Num: 24608, Cur Loss: 0.27694041, Cur Avg Loss: 0.41313925, Log Avg loss: 0.27684843, Global Avg Loss: 2.09659754, Time: 0.0250 Steps: 18570, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001548, Sample Num: 24768, Cur Loss: 0.28264165, Cur Avg Loss: 0.41265511, Log Avg loss: 0.33819537, Global Avg Loss: 2.09565114, Time: 0.0212 Steps: 18580, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001558, Sample Num: 24928, Cur Loss: 0.52281392, Cur Avg Loss: 0.41252907, Log Avg loss: 0.39301714, Global Avg Loss: 2.09473525, Time: 0.0210 Steps: 18590, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001568, Sample Num: 25088, Cur Loss: 0.31546548, Cur Avg Loss: 0.41182469, Log Avg loss: 0.30208225, Global Avg Loss: 2.09377146, Time: 0.0209 Steps: 18600, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001578, Sample Num: 25248, Cur Loss: 0.30180323, Cur Avg Loss: 0.41099146, Log Avg loss: 0.28034092, Global Avg Loss: 2.09279702, Time: 0.0210 Steps: 18610, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001588, Sample Num: 25408, Cur Loss: 0.27369857, Cur Avg Loss: 0.41047931, Log Avg loss: 0.32966218, Global Avg Loss: 2.09185012, Time: 0.0209 Steps: 18620, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001598, Sample Num: 25568, Cur Loss: 0.28547999, Cur Avg Loss: 0.40996961, Log Avg loss: 0.32902921, Global Avg Loss: 2.09090389, Time: 0.0210 Steps: 18630, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001608, Sample Num: 25728, Cur Loss: 0.13534859, Cur Avg Loss: 0.40930062, Log Avg loss: 0.30239694, Global Avg Loss: 2.08994439, Time: 0.0211 Steps: 18640, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001618, Sample Num: 25888, Cur Loss: 0.80288649, Cur Avg Loss: 0.40871736, Log Avg loss: 0.31492859, Global Avg Loss: 2.08899264, Time: 0.0209 Steps: 18650, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001628, Sample Num: 26048, Cur Loss: 0.32568747, Cur Avg Loss: 0.40862122, Log Avg loss: 0.39306636, Global Avg Loss: 2.08808378, Time: 0.0211 Steps: 18660, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001638, Sample Num: 26208, Cur Loss: 0.17609724, Cur Avg Loss: 0.40943437, Log Avg loss: 0.54181398, Global Avg Loss: 2.08725557, Time: 0.0211 Steps: 18670, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001648, Sample Num: 26368, Cur Loss: 0.17213100, Cur Avg Loss: 0.40953775, Log Avg loss: 0.42647197, Global Avg Loss: 2.08636650, Time: 0.0211 Steps: 18680, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001658, Sample Num: 26528, Cur Loss: 0.11269853, Cur Avg Loss: 0.40936164, Log Avg loss: 0.38033905, Global Avg Loss: 2.08545370, Time: 0.0209 Steps: 18690, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001668, Sample Num: 26688, Cur Loss: 0.85784835, Cur Avg Loss: 0.40982718, Log Avg loss: 0.48701305, Global Avg Loss: 2.08459892, Time: 0.0210 Steps: 18700, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001678, Sample Num: 26848, Cur Loss: 0.13486682, Cur Avg Loss: 0.41168310, Log Avg loss: 0.72125108, Global Avg Loss: 2.08387025, Time: 0.0211 Steps: 18710, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001688, Sample Num: 27008, Cur Loss: 0.45511636, Cur Avg Loss: 0.41129634, Log Avg loss: 0.34639737, Global Avg Loss: 2.08294211, Time: 0.0210 Steps: 18720, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001698, Sample Num: 27168, Cur Loss: 0.50949413, Cur Avg Loss: 0.41134231, Log Avg loss: 0.41910182, Global Avg Loss: 2.08205378, Time: 0.0208 Steps: 18730, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001708, Sample Num: 27328, Cur Loss: 0.08791885, Cur Avg Loss: 0.41100322, Log Avg loss: 0.35342596, Global Avg Loss: 2.08113135, Time: 0.0208 Steps: 18740, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001718, Sample Num: 27488, Cur Loss: 0.37370104, Cur Avg Loss: 0.41093005, Log Avg loss: 0.39843356, Global Avg Loss: 2.08023391, Time: 0.0209 Steps: 18750, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001728, Sample Num: 27648, Cur Loss: 0.49306738, Cur Avg Loss: 0.41115915, Log Avg loss: 0.45051772, Global Avg Loss: 2.07936520, Time: 0.0209 Steps: 18760, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001738, Sample Num: 27808, Cur Loss: 0.86187696, Cur Avg Loss: 0.41124749, Log Avg loss: 0.42651232, Global Avg Loss: 2.07848461, Time: 0.0209 Steps: 18770, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001748, Sample Num: 27968, Cur Loss: 0.26424772, Cur Avg Loss: 0.41099012, Log Avg loss: 0.36625902, Global Avg Loss: 2.07757289, Time: 0.0209 Steps: 18780, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001758, Sample Num: 28128, Cur Loss: 0.22483146, Cur Avg Loss: 0.41028441, Log Avg loss: 0.28692751, Global Avg Loss: 2.07661991, Time: 0.0210 Steps: 18790, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001768, Sample Num: 28288, Cur Loss: 0.44371378, Cur Avg Loss: 0.40989931, Log Avg loss: 0.34219876, Global Avg Loss: 2.07569734, Time: 0.0211 Steps: 18800, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001778, Sample Num: 28448, Cur Loss: 0.28447151, Cur Avg Loss: 0.40967434, Log Avg loss: 0.36989965, Global Avg Loss: 2.07479049, Time: 0.0210 Steps: 18810, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001788, Sample Num: 28608, Cur Loss: 0.25269791, Cur Avg Loss: 0.40966063, Log Avg loss: 0.40722286, Global Avg Loss: 2.07390443, Time: 0.0211 Steps: 18820, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001798, Sample Num: 28768, Cur Loss: 0.91741133, Cur Avg Loss: 0.40956963, Log Avg loss: 0.39329945, Global Avg Loss: 2.07301191, Time: 0.0211 Steps: 18830, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001808, Sample Num: 28928, Cur Loss: 0.35253531, Cur Avg Loss: 0.40933184, Log Avg loss: 0.36657652, Global Avg Loss: 2.07210616, Time: 0.0210 Steps: 18840, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001818, Sample Num: 29088, Cur Loss: 0.67782247, Cur Avg Loss: 0.41071703, Log Avg loss: 0.66115916, Global Avg Loss: 2.07135765, Time: 0.0209 Steps: 18850, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001828, Sample Num: 29248, Cur Loss: 0.52578950, Cur Avg Loss: 0.41038300, Log Avg loss: 0.34965642, Global Avg Loss: 2.07044476, Time: 0.0209 Steps: 18860, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001838, Sample Num: 29408, Cur Loss: 0.46294188, Cur Avg Loss: 0.40997609, Log Avg loss: 0.33559205, Global Avg Loss: 2.06952539, Time: 0.0209 Steps: 18870, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001848, Sample Num: 29568, Cur Loss: 0.17133448, Cur Avg Loss: 0.40974727, Log Avg loss: 0.36769095, Global Avg Loss: 2.06862400, Time: 0.0209 Steps: 18880, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001858, Sample Num: 29728, Cur Loss: 0.27815625, Cur Avg Loss: 0.40973995, Log Avg loss: 0.40838668, Global Avg Loss: 2.06774510, Time: 0.0209 Steps: 18890, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001868, Sample Num: 29888, Cur Loss: 0.29978073, Cur Avg Loss: 0.40952559, Log Avg loss: 0.36969772, Global Avg Loss: 2.06684666, Time: 0.0209 Steps: 18900, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001878, Sample Num: 30048, Cur Loss: 0.29789463, Cur Avg Loss: 0.40987038, Log Avg loss: 0.47427705, Global Avg Loss: 2.06600448, Time: 0.0209 Steps: 18910, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001888, Sample Num: 30208, Cur Loss: 0.54363120, Cur Avg Loss: 0.40994457, Log Avg loss: 0.42387715, Global Avg Loss: 2.06513654, Time: 0.0209 Steps: 18920, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001898, Sample Num: 30368, Cur Loss: 0.48033464, Cur Avg Loss: 0.40969234, Log Avg loss: 0.36207176, Global Avg Loss: 2.06423688, Time: 0.0209 Steps: 18930, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001908, Sample Num: 30528, Cur Loss: 0.22986871, Cur Avg Loss: 0.40869294, Log Avg loss: 0.21900653, Global Avg Loss: 2.06326263, Time: 0.0209 Steps: 18940, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001918, Sample Num: 30688, Cur Loss: 0.41381824, Cur Avg Loss: 0.40895570, Log Avg loss: 0.45909013, Global Avg Loss: 2.06241610, Time: 0.0209 Steps: 18950, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001928, Sample Num: 30848, Cur Loss: 1.22520733, Cur Avg Loss: 0.40962316, Log Avg loss: 0.53764367, Global Avg Loss: 2.06161190, Time: 0.0209 Steps: 18960, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001938, Sample Num: 31008, Cur Loss: 0.39792168, Cur Avg Loss: 0.40949449, Log Avg loss: 0.38468572, Global Avg Loss: 2.06072791, Time: 0.0209 Steps: 18970, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001948, Sample Num: 31168, Cur Loss: 0.36131030, Cur Avg Loss: 0.40869707, Log Avg loss: 0.25415680, Global Avg Loss: 2.05977608, Time: 0.0209 Steps: 18980, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001958, Sample Num: 31328, Cur Loss: 0.49894354, Cur Avg Loss: 0.40859087, Log Avg loss: 0.38790387, Global Avg Loss: 2.05889568, Time: 0.0209 Steps: 18990, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001968, Sample Num: 31488, Cur Loss: 0.25102255, Cur Avg Loss: 0.40774653, Log Avg loss: 0.24242362, Global Avg Loss: 2.05793964, Time: 0.0209 Steps: 19000, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001978, Sample Num: 31648, Cur Loss: 0.08648166, Cur Avg Loss: 0.40764064, Log Avg loss: 0.38680220, Global Avg Loss: 2.05706056, Time: 0.0209 Steps: 19010, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001988, Sample Num: 31808, Cur Loss: 0.89432156, Cur Avg Loss: 0.40905959, Log Avg loss: 0.68972900, Global Avg Loss: 2.05634167, Time: 0.0209 Steps: 19020, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001998, Sample Num: 31968, Cur Loss: 0.31716704, Cur Avg Loss: 0.40938647, Log Avg loss: 0.47436894, Global Avg Loss: 2.05551036, Time: 0.0209 Steps: 19030, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002008, Sample Num: 32128, Cur Loss: 0.38210708, Cur Avg Loss: 0.40900544, Log Avg loss: 0.33287555, Global Avg Loss: 2.05460562, Time: 0.0209 Steps: 19040, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002018, Sample Num: 32288, Cur Loss: 0.65984392, Cur Avg Loss: 0.40869003, Log Avg loss: 0.34535588, Global Avg Loss: 2.05370838, Time: 0.0209 Steps: 19050, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002028, Sample Num: 32448, Cur Loss: 0.21799013, Cur Avg Loss: 0.40851967, Log Avg loss: 0.37414068, Global Avg Loss: 2.05282718, Time: 0.0208 Steps: 19060, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002038, Sample Num: 32608, Cur Loss: 0.32861638, Cur Avg Loss: 0.40862601, Log Avg loss: 0.43019286, Global Avg Loss: 2.05197629, Time: 0.0209 Steps: 19070, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002048, Sample Num: 32768, Cur Loss: 0.41295213, Cur Avg Loss: 0.40873947, Log Avg loss: 0.43186174, Global Avg Loss: 2.05112718, Time: 0.0254 Steps: 19080, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002058, Sample Num: 32928, Cur Loss: 1.03886557, Cur Avg Loss: 0.40908672, Log Avg loss: 0.48020433, Global Avg Loss: 2.05030427, Time: 0.0209 Steps: 19090, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002068, Sample Num: 33088, Cur Loss: 0.16674533, Cur Avg Loss: 0.40867265, Log Avg loss: 0.32345743, Global Avg Loss: 2.04940016, Time: 0.0209 Steps: 19100, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002078, Sample Num: 33248, Cur Loss: 0.31657076, Cur Avg Loss: 0.40820594, Log Avg loss: 0.31168900, Global Avg Loss: 2.04849084, Time: 0.0210 Steps: 19110, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002088, Sample Num: 33408, Cur Loss: 0.54848075, Cur Avg Loss: 0.40871635, Log Avg loss: 0.51478092, Global Avg Loss: 2.04768869, Time: 0.0209 Steps: 19120, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002098, Sample Num: 33568, Cur Loss: 0.18598750, Cur Avg Loss: 0.40892226, Log Avg loss: 0.45191620, Global Avg Loss: 2.04685452, Time: 0.0209 Steps: 19130, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002108, Sample Num: 33728, Cur Loss: 0.21213645, Cur Avg Loss: 0.40896911, Log Avg loss: 0.41879798, Global Avg Loss: 2.04600392, Time: 0.0209 Steps: 19140, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002118, Sample Num: 33888, Cur Loss: 0.64437276, Cur Avg Loss: 0.40879841, Log Avg loss: 0.37281469, Global Avg Loss: 2.04513019, Time: 0.0209 Steps: 19150, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002128, Sample Num: 34048, Cur Loss: 0.25469461, Cur Avg Loss: 0.40886228, Log Avg loss: 0.42239070, Global Avg Loss: 2.04428325, Time: 0.0210 Steps: 19160, Updated lr: 0.000083 ***** Running evaluation checkpoint-19161 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-19161 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.021619, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.375489, "eval_total_loss": 263.968508, "eval_mae": 0.410329, "eval_mse": 0.375619, "eval_r2": 0.761232, "eval_sp_statistic": 0.870568, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.900663, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.688406, "test_total_loss": 345.579909, "test_mae": 0.585769, "test_mse": 0.688621, "test_r2": 0.555558, "test_sp_statistic": 0.867395, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.921197, "test_ps_pvalue": 0.0, "lr": 8.277761972498815e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.044204995087841, "train_cur_epoch_loss": 870.6038361415267, "train_cur_epoch_avg_loss": 0.4089261794934367, "train_cur_epoch_time": 45.021618604660034, "train_cur_epoch_avg_time": 0.021146838236101473, "epoch": 9, "step": 19161} ################################################## Training, Epoch: 0010, Batch: 000009, Sample Num: 144, Cur Loss: 0.36226600, Cur Avg Loss: 0.43455271, Log Avg loss: 0.44558682, Global Avg Loss: 2.04344929, Time: 0.0212 Steps: 19170, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000019, Sample Num: 304, Cur Loss: 0.27582628, Cur Avg Loss: 0.38282506, Log Avg loss: 0.33627017, Global Avg Loss: 2.04255921, Time: 0.0210 Steps: 19180, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000029, Sample Num: 464, Cur Loss: 0.27677923, Cur Avg Loss: 0.37074820, Log Avg loss: 0.34780217, Global Avg Loss: 2.04167606, Time: 0.0210 Steps: 19190, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000039, Sample Num: 624, Cur Loss: 0.26285356, Cur Avg Loss: 0.34907510, Log Avg loss: 0.28622313, Global Avg Loss: 2.04076176, Time: 0.0210 Steps: 19200, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000049, Sample Num: 784, Cur Loss: 0.24451968, Cur Avg Loss: 0.32577111, Log Avg loss: 0.23488555, Global Avg Loss: 2.03982169, Time: 0.0210 Steps: 19210, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000059, Sample Num: 944, Cur Loss: 0.44885075, Cur Avg Loss: 0.34100474, Log Avg loss: 0.41564949, Global Avg Loss: 2.03897665, Time: 0.0210 Steps: 19220, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000069, Sample Num: 1104, Cur Loss: 0.53381073, Cur Avg Loss: 0.36139612, Log Avg loss: 0.48170525, Global Avg Loss: 2.03816684, Time: 0.0210 Steps: 19230, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000079, Sample Num: 1264, Cur Loss: 0.27996764, Cur Avg Loss: 0.36890247, Log Avg loss: 0.42069628, Global Avg Loss: 2.03732615, Time: 0.0210 Steps: 19240, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000089, Sample Num: 1424, Cur Loss: 0.33139616, Cur Avg Loss: 0.36048457, Log Avg loss: 0.29398317, Global Avg Loss: 2.03642052, Time: 0.0210 Steps: 19250, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000099, Sample Num: 1584, Cur Loss: 0.79892242, Cur Avg Loss: 0.37186424, Log Avg loss: 0.47314336, Global Avg Loss: 2.03560885, Time: 0.0209 Steps: 19260, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000109, Sample Num: 1744, Cur Loss: 0.61376262, Cur Avg Loss: 0.37694350, Log Avg loss: 0.42722811, Global Avg Loss: 2.03477420, Time: 0.0209 Steps: 19270, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000119, Sample Num: 1904, Cur Loss: 0.47251213, Cur Avg Loss: 0.37057482, Log Avg loss: 0.30115620, Global Avg Loss: 2.03387502, Time: 0.0209 Steps: 19280, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000129, Sample Num: 2064, Cur Loss: 0.32702816, Cur Avg Loss: 0.36340806, Log Avg loss: 0.27812373, Global Avg Loss: 2.03296483, Time: 0.0208 Steps: 19290, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000139, Sample Num: 2224, Cur Loss: 0.05835278, Cur Avg Loss: 0.35705311, Log Avg loss: 0.27507423, Global Avg Loss: 2.03205400, Time: 0.0208 Steps: 19300, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000149, Sample Num: 2384, Cur Loss: 0.39395946, Cur Avg Loss: 0.35553188, Log Avg loss: 0.33438678, Global Avg Loss: 2.03117484, Time: 0.0208 Steps: 19310, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000159, Sample Num: 2544, Cur Loss: 0.55734807, Cur Avg Loss: 0.36095723, Log Avg loss: 0.44179489, Global Avg Loss: 2.03035218, Time: 0.0208 Steps: 19320, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000169, Sample Num: 2704, Cur Loss: 0.14458880, Cur Avg Loss: 0.36060490, Log Avg loss: 0.35500282, Global Avg Loss: 2.02948547, Time: 0.0208 Steps: 19330, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000179, Sample Num: 2864, Cur Loss: 0.07162297, Cur Avg Loss: 0.36066801, Log Avg loss: 0.36173457, Global Avg Loss: 2.02862314, Time: 0.0208 Steps: 19340, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000189, Sample Num: 3024, Cur Loss: 0.20631811, Cur Avg Loss: 0.36583640, Log Avg loss: 0.45835059, Global Avg Loss: 2.02781163, Time: 0.0208 Steps: 19350, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000199, Sample Num: 3184, Cur Loss: 0.23110387, Cur Avg Loss: 0.36374387, Log Avg loss: 0.32419517, Global Avg Loss: 2.02693166, Time: 0.0209 Steps: 19360, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000209, Sample Num: 3344, Cur Loss: 0.19953516, Cur Avg Loss: 0.35883898, Log Avg loss: 0.26123164, Global Avg Loss: 2.02602010, Time: 0.0209 Steps: 19370, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000219, Sample Num: 3504, Cur Loss: 0.26926017, Cur Avg Loss: 0.35439971, Log Avg loss: 0.26161883, Global Avg Loss: 2.02510967, Time: 0.0209 Steps: 19380, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000229, Sample Num: 3664, Cur Loss: 0.99331331, Cur Avg Loss: 0.35807726, Log Avg loss: 0.43861575, Global Avg Loss: 2.02429147, Time: 0.0209 Steps: 19390, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000239, Sample Num: 3824, Cur Loss: 0.66222215, Cur Avg Loss: 0.35715240, Log Avg loss: 0.33597303, Global Avg Loss: 2.02342120, Time: 0.0209 Steps: 19400, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000249, Sample Num: 3984, Cur Loss: 0.39820328, Cur Avg Loss: 0.35713319, Log Avg loss: 0.35667419, Global Avg Loss: 2.02256250, Time: 0.0208 Steps: 19410, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000259, Sample Num: 4144, Cur Loss: 0.14640254, Cur Avg Loss: 0.35411893, Log Avg loss: 0.27906374, Global Avg Loss: 2.02166471, Time: 0.0244 Steps: 19420, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000269, Sample Num: 4304, Cur Loss: 0.20946473, Cur Avg Loss: 0.35577984, Log Avg loss: 0.39879747, Global Avg Loss: 2.02082947, Time: 0.0209 Steps: 19430, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000279, Sample Num: 4464, Cur Loss: 0.09489179, Cur Avg Loss: 0.35204899, Log Avg loss: 0.25168920, Global Avg Loss: 2.01991942, Time: 0.0209 Steps: 19440, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000289, Sample Num: 4624, Cur Loss: 0.53237641, Cur Avg Loss: 0.35208765, Log Avg loss: 0.35316608, Global Avg Loss: 2.01906248, Time: 0.0209 Steps: 19450, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000299, Sample Num: 4784, Cur Loss: 0.16098386, Cur Avg Loss: 0.35108795, Log Avg loss: 0.32219677, Global Avg Loss: 2.01819050, Time: 0.0210 Steps: 19460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000309, Sample Num: 4944, Cur Loss: 0.22904240, Cur Avg Loss: 0.35487815, Log Avg loss: 0.46820496, Global Avg Loss: 2.01739441, Time: 0.0209 Steps: 19470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000319, Sample Num: 5104, Cur Loss: 0.11463778, Cur Avg Loss: 0.35665064, Log Avg loss: 0.41142066, Global Avg Loss: 2.01656999, Time: 0.0209 Steps: 19480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000329, Sample Num: 5264, Cur Loss: 0.15905412, Cur Avg Loss: 0.36121320, Log Avg loss: 0.50675877, Global Avg Loss: 2.01579533, Time: 0.0209 Steps: 19490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000339, Sample Num: 5424, Cur Loss: 0.63995582, Cur Avg Loss: 0.36830961, Log Avg loss: 0.60178149, Global Avg Loss: 2.01507020, Time: 0.0210 Steps: 19500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000349, Sample Num: 5584, Cur Loss: 0.72155035, Cur Avg Loss: 0.37737163, Log Avg loss: 0.68457428, Global Avg Loss: 2.01438824, Time: 0.0210 Steps: 19510, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000359, Sample Num: 5744, Cur Loss: 0.35129315, Cur Avg Loss: 0.37439762, Log Avg loss: 0.27060458, Global Avg Loss: 2.01349491, Time: 0.0210 Steps: 19520, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000369, Sample Num: 5904, Cur Loss: 0.26196611, Cur Avg Loss: 0.37313035, Log Avg loss: 0.32763522, Global Avg Loss: 2.01263170, Time: 0.0210 Steps: 19530, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000379, Sample Num: 6064, Cur Loss: 0.36960617, Cur Avg Loss: 0.37084586, Log Avg loss: 0.28654831, Global Avg Loss: 2.01174834, Time: 0.0210 Steps: 19540, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000389, Sample Num: 6224, Cur Loss: 0.34215477, Cur Avg Loss: 0.37076253, Log Avg loss: 0.36760454, Global Avg Loss: 2.01090734, Time: 0.0210 Steps: 19550, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000399, Sample Num: 6384, Cur Loss: 0.53355157, Cur Avg Loss: 0.36922852, Log Avg loss: 0.30955537, Global Avg Loss: 2.01003753, Time: 0.0210 Steps: 19560, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000409, Sample Num: 6544, Cur Loss: 0.50353569, Cur Avg Loss: 0.37068086, Log Avg loss: 0.42862940, Global Avg Loss: 2.00922945, Time: 0.0210 Steps: 19570, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000419, Sample Num: 6704, Cur Loss: 0.46511000, Cur Avg Loss: 0.37026083, Log Avg loss: 0.35308131, Global Avg Loss: 2.00838362, Time: 0.0209 Steps: 19580, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000429, Sample Num: 6864, Cur Loss: 0.20578620, Cur Avg Loss: 0.36949768, Log Avg loss: 0.33752200, Global Avg Loss: 2.00753070, Time: 0.0210 Steps: 19590, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000439, Sample Num: 7024, Cur Loss: 0.51719236, Cur Avg Loss: 0.37107777, Log Avg loss: 0.43886361, Global Avg Loss: 2.00673036, Time: 0.0209 Steps: 19600, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000449, Sample Num: 7184, Cur Loss: 0.26824421, Cur Avg Loss: 0.37007554, Log Avg loss: 0.32607753, Global Avg Loss: 2.00587332, Time: 0.0210 Steps: 19610, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000459, Sample Num: 7344, Cur Loss: 0.28466320, Cur Avg Loss: 0.36926094, Log Avg loss: 0.33268531, Global Avg Loss: 2.00502052, Time: 0.0209 Steps: 19620, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000469, Sample Num: 7504, Cur Loss: 0.27624393, Cur Avg Loss: 0.36627156, Log Avg loss: 0.22905916, Global Avg Loss: 2.00411581, Time: 0.0209 Steps: 19630, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000479, Sample Num: 7664, Cur Loss: 0.31404534, Cur Avg Loss: 0.36500289, Log Avg loss: 0.30550217, Global Avg Loss: 2.00325093, Time: 0.0209 Steps: 19640, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000489, Sample Num: 7824, Cur Loss: 0.53330797, Cur Avg Loss: 0.36404424, Log Avg loss: 0.31812483, Global Avg Loss: 2.00239336, Time: 0.0209 Steps: 19650, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000499, Sample Num: 7984, Cur Loss: 0.52986377, Cur Avg Loss: 0.36352819, Log Avg loss: 0.33829355, Global Avg Loss: 2.00154692, Time: 0.0209 Steps: 19660, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000509, Sample Num: 8144, Cur Loss: 0.75905591, Cur Avg Loss: 0.36414451, Log Avg loss: 0.39489879, Global Avg Loss: 2.00073012, Time: 0.0210 Steps: 19670, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000519, Sample Num: 8304, Cur Loss: 0.29015970, Cur Avg Loss: 0.36715935, Log Avg loss: 0.52061478, Global Avg Loss: 1.99997803, Time: 0.0234 Steps: 19680, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000529, Sample Num: 8464, Cur Loss: 0.46993923, Cur Avg Loss: 0.37044737, Log Avg loss: 0.54109526, Global Avg Loss: 1.99923710, Time: 0.0209 Steps: 19690, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000539, Sample Num: 8624, Cur Loss: 0.60930485, Cur Avg Loss: 0.37250256, Log Avg loss: 0.48122222, Global Avg Loss: 1.99846654, Time: 0.0210 Steps: 19700, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000549, Sample Num: 8784, Cur Loss: 0.18222421, Cur Avg Loss: 0.37122667, Log Avg loss: 0.30245616, Global Avg Loss: 1.99760606, Time: 0.0209 Steps: 19710, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000559, Sample Num: 8944, Cur Loss: 0.42654759, Cur Avg Loss: 0.36972830, Log Avg loss: 0.28746765, Global Avg Loss: 1.99673885, Time: 0.0210 Steps: 19720, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000569, Sample Num: 9104, Cur Loss: 0.72741151, Cur Avg Loss: 0.36868115, Log Avg loss: 0.31014549, Global Avg Loss: 1.99588401, Time: 0.0209 Steps: 19730, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000579, Sample Num: 9264, Cur Loss: 0.60756946, Cur Avg Loss: 0.36767880, Log Avg loss: 0.31064555, Global Avg Loss: 1.99503029, Time: 0.0209 Steps: 19740, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000589, Sample Num: 9424, Cur Loss: 0.54702514, Cur Avg Loss: 0.36667126, Log Avg loss: 0.30833452, Global Avg Loss: 1.99417627, Time: 0.0209 Steps: 19750, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000599, Sample Num: 9584, Cur Loss: 0.49555284, Cur Avg Loss: 0.36554223, Log Avg loss: 0.29904235, Global Avg Loss: 1.99331841, Time: 0.0209 Steps: 19760, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000609, Sample Num: 9744, Cur Loss: 0.12697604, Cur Avg Loss: 0.36551999, Log Avg loss: 0.36418799, Global Avg Loss: 1.99249436, Time: 0.0209 Steps: 19770, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000619, Sample Num: 9904, Cur Loss: 0.21405438, Cur Avg Loss: 0.36772084, Log Avg loss: 0.50175224, Global Avg Loss: 1.99174070, Time: 0.0209 Steps: 19780, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000629, Sample Num: 10064, Cur Loss: 0.19726183, Cur Avg Loss: 0.36756301, Log Avg loss: 0.35779346, Global Avg Loss: 1.99091506, Time: 0.0209 Steps: 19790, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000639, Sample Num: 10224, Cur Loss: 0.35688791, Cur Avg Loss: 0.36910826, Log Avg loss: 0.46630470, Global Avg Loss: 1.99014506, Time: 0.0209 Steps: 19800, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000649, Sample Num: 10384, Cur Loss: 0.39566264, Cur Avg Loss: 0.36893453, Log Avg loss: 0.35783298, Global Avg Loss: 1.98932107, Time: 0.0209 Steps: 19810, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000659, Sample Num: 10544, Cur Loss: 0.53389806, Cur Avg Loss: 0.36862668, Log Avg loss: 0.34864743, Global Avg Loss: 1.98849328, Time: 0.0209 Steps: 19820, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000669, Sample Num: 10704, Cur Loss: 0.77616215, Cur Avg Loss: 0.36824321, Log Avg loss: 0.34297256, Global Avg Loss: 1.98766347, Time: 0.0209 Steps: 19830, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000679, Sample Num: 10864, Cur Loss: 0.21501574, Cur Avg Loss: 0.36944613, Log Avg loss: 0.44992106, Global Avg Loss: 1.98688840, Time: 0.0209 Steps: 19840, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000689, Sample Num: 11024, Cur Loss: 0.50827461, Cur Avg Loss: 0.36917526, Log Avg loss: 0.35078306, Global Avg Loss: 1.98606416, Time: 0.0209 Steps: 19850, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000699, Sample Num: 11184, Cur Loss: 0.12868235, Cur Avg Loss: 0.36735443, Log Avg loss: 0.24189923, Global Avg Loss: 1.98518593, Time: 0.0209 Steps: 19860, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000709, Sample Num: 11344, Cur Loss: 0.20818496, Cur Avg Loss: 0.36653530, Log Avg loss: 0.30927875, Global Avg Loss: 1.98434250, Time: 0.0209 Steps: 19870, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000719, Sample Num: 11504, Cur Loss: 0.25266108, Cur Avg Loss: 0.36597780, Log Avg loss: 0.32645077, Global Avg Loss: 1.98350855, Time: 0.0209 Steps: 19880, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000729, Sample Num: 11664, Cur Loss: 0.11694968, Cur Avg Loss: 0.36575480, Log Avg loss: 0.34972097, Global Avg Loss: 1.98268714, Time: 0.0209 Steps: 19890, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000739, Sample Num: 11824, Cur Loss: 0.66295129, Cur Avg Loss: 0.36532254, Log Avg loss: 0.33381076, Global Avg Loss: 1.98185856, Time: 0.0208 Steps: 19900, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000749, Sample Num: 11984, Cur Loss: 0.19369639, Cur Avg Loss: 0.36421189, Log Avg loss: 0.28213489, Global Avg Loss: 1.98100485, Time: 0.0210 Steps: 19910, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000759, Sample Num: 12144, Cur Loss: 0.39798751, Cur Avg Loss: 0.36334961, Log Avg loss: 0.29876496, Global Avg Loss: 1.98016035, Time: 0.0209 Steps: 19920, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000769, Sample Num: 12304, Cur Loss: 0.68719375, Cur Avg Loss: 0.36321050, Log Avg loss: 0.35265171, Global Avg Loss: 1.97934374, Time: 0.0246 Steps: 19930, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000779, Sample Num: 12464, Cur Loss: 0.92632008, Cur Avg Loss: 0.36457368, Log Avg loss: 0.46940214, Global Avg Loss: 1.97858650, Time: 0.0211 Steps: 19940, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000789, Sample Num: 12624, Cur Loss: 0.86418027, Cur Avg Loss: 0.36815023, Log Avg loss: 0.64676363, Global Avg Loss: 1.97791892, Time: 0.0210 Steps: 19950, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000799, Sample Num: 12784, Cur Loss: 0.32568467, Cur Avg Loss: 0.36734445, Log Avg loss: 0.30376832, Global Avg Loss: 1.97708017, Time: 0.0210 Steps: 19960, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000809, Sample Num: 12944, Cur Loss: 0.25565329, Cur Avg Loss: 0.36566213, Log Avg loss: 0.23124534, Global Avg Loss: 1.97620594, Time: 0.0210 Steps: 19970, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000819, Sample Num: 13104, Cur Loss: 0.24813256, Cur Avg Loss: 0.36536798, Log Avg loss: 0.34157088, Global Avg Loss: 1.97538780, Time: 0.0210 Steps: 19980, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000829, Sample Num: 13264, Cur Loss: 0.46502694, Cur Avg Loss: 0.36599018, Log Avg loss: 0.41694887, Global Avg Loss: 1.97460819, Time: 0.0210 Steps: 19990, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000839, Sample Num: 13424, Cur Loss: 0.57490659, Cur Avg Loss: 0.36584586, Log Avg loss: 0.35388153, Global Avg Loss: 1.97379783, Time: 0.0210 Steps: 20000, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000849, Sample Num: 13584, Cur Loss: 0.33088261, Cur Avg Loss: 0.36716977, Log Avg loss: 0.47824520, Global Avg Loss: 1.97305043, Time: 0.0210 Steps: 20010, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000859, Sample Num: 13744, Cur Loss: 0.17433451, Cur Avg Loss: 0.36603599, Log Avg loss: 0.26977841, Global Avg Loss: 1.97219964, Time: 0.0210 Steps: 20020, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000869, Sample Num: 13904, Cur Loss: 0.42995742, Cur Avg Loss: 0.36770387, Log Avg loss: 0.51097496, Global Avg Loss: 1.97147012, Time: 0.0210 Steps: 20030, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000879, Sample Num: 14064, Cur Loss: 0.22533673, Cur Avg Loss: 0.36690271, Log Avg loss: 0.29728204, Global Avg Loss: 1.97063470, Time: 0.0210 Steps: 20040, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000889, Sample Num: 14224, Cur Loss: 0.21301140, Cur Avg Loss: 0.36699848, Log Avg loss: 0.37541667, Global Avg Loss: 1.96983908, Time: 0.0210 Steps: 20050, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000899, Sample Num: 14384, Cur Loss: 0.44025618, Cur Avg Loss: 0.36666752, Log Avg loss: 0.33724450, Global Avg Loss: 1.96902522, Time: 0.0210 Steps: 20060, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000909, Sample Num: 14544, Cur Loss: 0.36712921, Cur Avg Loss: 0.36800579, Log Avg loss: 0.48831688, Global Avg Loss: 1.96828745, Time: 0.0210 Steps: 20070, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000919, Sample Num: 14704, Cur Loss: 0.32812130, Cur Avg Loss: 0.36786567, Log Avg loss: 0.35512833, Global Avg Loss: 1.96748409, Time: 0.0210 Steps: 20080, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000929, Sample Num: 14864, Cur Loss: 0.39001912, Cur Avg Loss: 0.37018098, Log Avg loss: 0.58295809, Global Avg Loss: 1.96679492, Time: 0.0211 Steps: 20090, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000939, Sample Num: 15024, Cur Loss: 0.33425128, Cur Avg Loss: 0.36916864, Log Avg loss: 0.27512215, Global Avg Loss: 1.96595330, Time: 0.0210 Steps: 20100, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000949, Sample Num: 15184, Cur Loss: 0.11704919, Cur Avg Loss: 0.36835616, Log Avg loss: 0.29206414, Global Avg Loss: 1.96512093, Time: 0.0209 Steps: 20110, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000959, Sample Num: 15344, Cur Loss: 0.92236710, Cur Avg Loss: 0.37062146, Log Avg loss: 0.58559835, Global Avg Loss: 1.96443528, Time: 0.0210 Steps: 20120, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000969, Sample Num: 15504, Cur Loss: 0.18768249, Cur Avg Loss: 0.37190033, Log Avg loss: 0.49454422, Global Avg Loss: 1.96370508, Time: 0.0210 Steps: 20130, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000979, Sample Num: 15664, Cur Loss: 0.57210851, Cur Avg Loss: 0.37185500, Log Avg loss: 0.36746293, Global Avg Loss: 1.96291251, Time: 0.0210 Steps: 20140, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000989, Sample Num: 15824, Cur Loss: 0.48001552, Cur Avg Loss: 0.37150584, Log Avg loss: 0.33732243, Global Avg Loss: 1.96210577, Time: 0.0210 Steps: 20150, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000999, Sample Num: 15984, Cur Loss: 0.51478261, Cur Avg Loss: 0.37075248, Log Avg loss: 0.29624542, Global Avg Loss: 1.96127945, Time: 0.0210 Steps: 20160, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001009, Sample Num: 16144, Cur Loss: 0.41126245, Cur Avg Loss: 0.37055484, Log Avg loss: 0.35081031, Global Avg Loss: 1.96048100, Time: 0.0210 Steps: 20170, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001019, Sample Num: 16304, Cur Loss: 0.20796245, Cur Avg Loss: 0.36976729, Log Avg loss: 0.29030403, Global Avg Loss: 1.95965336, Time: 0.0210 Steps: 20180, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001029, Sample Num: 16464, Cur Loss: 0.35352367, Cur Avg Loss: 0.36937748, Log Avg loss: 0.32965616, Global Avg Loss: 1.95884603, Time: 0.0247 Steps: 20190, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001039, Sample Num: 16624, Cur Loss: 0.61375999, Cur Avg Loss: 0.36983924, Log Avg loss: 0.41735372, Global Avg Loss: 1.95808291, Time: 0.0210 Steps: 20200, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001049, Sample Num: 16784, Cur Loss: 0.19190028, Cur Avg Loss: 0.37057820, Log Avg loss: 0.44735651, Global Avg Loss: 1.95733540, Time: 0.0210 Steps: 20210, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001059, Sample Num: 16944, Cur Loss: 0.79972744, Cur Avg Loss: 0.37266990, Log Avg loss: 0.59208868, Global Avg Loss: 1.95666020, Time: 0.0210 Steps: 20220, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001069, Sample Num: 17104, Cur Loss: 0.13317691, Cur Avg Loss: 0.37188462, Log Avg loss: 0.28872436, Global Avg Loss: 1.95583572, Time: 0.0210 Steps: 20230, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001079, Sample Num: 17264, Cur Loss: 0.20600639, Cur Avg Loss: 0.37071432, Log Avg loss: 0.24560877, Global Avg Loss: 1.95499074, Time: 0.0210 Steps: 20240, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001089, Sample Num: 17424, Cur Loss: 0.28959009, Cur Avg Loss: 0.37013363, Log Avg loss: 0.30747686, Global Avg Loss: 1.95417716, Time: 0.0210 Steps: 20250, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001099, Sample Num: 17584, Cur Loss: 0.27635598, Cur Avg Loss: 0.36936265, Log Avg loss: 0.28540284, Global Avg Loss: 1.95335348, Time: 0.0211 Steps: 20260, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001109, Sample Num: 17744, Cur Loss: 0.28972495, Cur Avg Loss: 0.36856974, Log Avg loss: 0.28142965, Global Avg Loss: 1.95252865, Time: 0.0210 Steps: 20270, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001119, Sample Num: 17904, Cur Loss: 0.33370465, Cur Avg Loss: 0.36806331, Log Avg loss: 0.31189961, Global Avg Loss: 1.95171966, Time: 0.0210 Steps: 20280, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001129, Sample Num: 18064, Cur Loss: 0.69895631, Cur Avg Loss: 0.36935502, Log Avg loss: 0.51389753, Global Avg Loss: 1.95101103, Time: 0.0210 Steps: 20290, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001139, Sample Num: 18224, Cur Loss: 0.36800185, Cur Avg Loss: 0.37002336, Log Avg loss: 0.44547948, Global Avg Loss: 1.95026939, Time: 0.0210 Steps: 20300, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001149, Sample Num: 18384, Cur Loss: 0.58309710, Cur Avg Loss: 0.37072517, Log Avg loss: 0.45066071, Global Avg Loss: 1.94953103, Time: 0.0210 Steps: 20310, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001159, Sample Num: 18544, Cur Loss: 0.70818645, Cur Avg Loss: 0.37117227, Log Avg loss: 0.42254379, Global Avg Loss: 1.94877956, Time: 0.0209 Steps: 20320, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001169, Sample Num: 18704, Cur Loss: 0.15539436, Cur Avg Loss: 0.37091471, Log Avg loss: 0.34106408, Global Avg Loss: 1.94798875, Time: 0.0210 Steps: 20330, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001179, Sample Num: 18864, Cur Loss: 0.20432448, Cur Avg Loss: 0.37080321, Log Avg loss: 0.35776885, Global Avg Loss: 1.94720693, Time: 0.0210 Steps: 20340, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001189, Sample Num: 19024, Cur Loss: 0.34600016, Cur Avg Loss: 0.37029643, Log Avg loss: 0.31054758, Global Avg Loss: 1.94640267, Time: 0.0210 Steps: 20350, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001199, Sample Num: 19184, Cur Loss: 0.17993140, Cur Avg Loss: 0.36905409, Log Avg loss: 0.22133949, Global Avg Loss: 1.94555539, Time: 0.0210 Steps: 20360, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001209, Sample Num: 19344, Cur Loss: 0.73123431, Cur Avg Loss: 0.36882602, Log Avg loss: 0.34147973, Global Avg Loss: 1.94476792, Time: 0.0210 Steps: 20370, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001219, Sample Num: 19504, Cur Loss: 0.67163289, Cur Avg Loss: 0.36864282, Log Avg loss: 0.34649504, Global Avg Loss: 1.94398369, Time: 0.0209 Steps: 20380, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001229, Sample Num: 19664, Cur Loss: 0.26313302, Cur Avg Loss: 0.36908288, Log Avg loss: 0.42272525, Global Avg Loss: 1.94323761, Time: 0.0210 Steps: 20390, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001239, Sample Num: 19824, Cur Loss: 0.09491306, Cur Avg Loss: 0.36928908, Log Avg loss: 0.39463180, Global Avg Loss: 1.94247848, Time: 0.0210 Steps: 20400, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001249, Sample Num: 19984, Cur Loss: 0.29975906, Cur Avg Loss: 0.36880372, Log Avg loss: 0.30866674, Global Avg Loss: 1.94167799, Time: 0.0211 Steps: 20410, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001259, Sample Num: 20144, Cur Loss: 0.73493028, Cur Avg Loss: 0.37005268, Log Avg loss: 0.52604863, Global Avg Loss: 1.94098473, Time: 0.0209 Steps: 20420, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001269, Sample Num: 20304, Cur Loss: 0.25620228, Cur Avg Loss: 0.37251923, Log Avg loss: 0.68305702, Global Avg Loss: 1.94036901, Time: 0.0209 Steps: 20430, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001279, Sample Num: 20464, Cur Loss: 0.35796463, Cur Avg Loss: 0.37260205, Log Avg loss: 0.38311281, Global Avg Loss: 1.93960714, Time: 0.0209 Steps: 20440, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001289, Sample Num: 20624, Cur Loss: 0.38004270, Cur Avg Loss: 0.37227689, Log Avg loss: 0.33068896, Global Avg Loss: 1.93882038, Time: 0.0219 Steps: 20450, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001299, Sample Num: 20784, Cur Loss: 0.58649290, Cur Avg Loss: 0.37406072, Log Avg loss: 0.60399564, Global Avg Loss: 1.93816798, Time: 0.0219 Steps: 20460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001309, Sample Num: 20944, Cur Loss: 1.86391068, Cur Avg Loss: 0.37749696, Log Avg loss: 0.82386489, Global Avg Loss: 1.93762362, Time: 0.0219 Steps: 20470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001319, Sample Num: 21104, Cur Loss: 0.17462438, Cur Avg Loss: 0.37751360, Log Avg loss: 0.37969204, Global Avg Loss: 1.93686291, Time: 0.0219 Steps: 20480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001329, Sample Num: 21264, Cur Loss: 0.56627107, Cur Avg Loss: 0.37918452, Log Avg loss: 0.59957880, Global Avg Loss: 1.93621026, Time: 0.0218 Steps: 20490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001339, Sample Num: 21424, Cur Loss: 1.50131667, Cur Avg Loss: 0.38157358, Log Avg loss: 0.69907979, Global Avg Loss: 1.93560678, Time: 0.0219 Steps: 20500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001349, Sample Num: 21584, Cur Loss: 0.19697069, Cur Avg Loss: 0.38219555, Log Avg loss: 0.46547635, Global Avg Loss: 1.93488999, Time: 0.0219 Steps: 20510, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001359, Sample Num: 21744, Cur Loss: 0.53484726, Cur Avg Loss: 0.38279311, Log Avg loss: 0.46340450, Global Avg Loss: 1.93417289, Time: 0.0219 Steps: 20520, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001369, Sample Num: 21904, Cur Loss: 0.81230581, Cur Avg Loss: 0.38629950, Log Avg loss: 0.86281794, Global Avg Loss: 1.93365104, Time: 0.0219 Steps: 20530, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001379, Sample Num: 22064, Cur Loss: 0.40602964, Cur Avg Loss: 0.38593526, Log Avg loss: 0.33607103, Global Avg Loss: 1.93287325, Time: 0.0219 Steps: 20540, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001389, Sample Num: 22224, Cur Loss: 0.21328053, Cur Avg Loss: 0.38552528, Log Avg loss: 0.32898926, Global Avg Loss: 1.93209278, Time: 0.0219 Steps: 20550, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001399, Sample Num: 22384, Cur Loss: 0.63723147, Cur Avg Loss: 0.38513178, Log Avg loss: 0.33047446, Global Avg Loss: 1.93131378, Time: 0.0219 Steps: 20560, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001409, Sample Num: 22544, Cur Loss: 0.20927212, Cur Avg Loss: 0.38490474, Log Avg loss: 0.35314168, Global Avg Loss: 1.93054656, Time: 0.0219 Steps: 20570, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001419, Sample Num: 22704, Cur Loss: 0.26659086, Cur Avg Loss: 0.38394833, Log Avg loss: 0.24918971, Global Avg Loss: 1.92972957, Time: 0.0219 Steps: 20580, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001429, Sample Num: 22864, Cur Loss: 0.71928930, Cur Avg Loss: 0.38404686, Log Avg loss: 0.39802835, Global Avg Loss: 1.92898567, Time: 0.0219 Steps: 20590, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001439, Sample Num: 23024, Cur Loss: 0.24310699, Cur Avg Loss: 0.38280141, Log Avg loss: 0.20482673, Global Avg Loss: 1.92814870, Time: 0.0219 Steps: 20600, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001449, Sample Num: 23184, Cur Loss: 0.27389407, Cur Avg Loss: 0.38246318, Log Avg loss: 0.33379243, Global Avg Loss: 1.92737511, Time: 0.0219 Steps: 20610, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001459, Sample Num: 23344, Cur Loss: 0.17963025, Cur Avg Loss: 0.38234264, Log Avg loss: 0.36487609, Global Avg Loss: 1.92661735, Time: 0.0219 Steps: 20620, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001469, Sample Num: 23504, Cur Loss: 0.37832227, Cur Avg Loss: 0.38211520, Log Avg loss: 0.34893129, Global Avg Loss: 1.92585260, Time: 0.0219 Steps: 20630, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001479, Sample Num: 23664, Cur Loss: 0.42596394, Cur Avg Loss: 0.38174344, Log Avg loss: 0.32713266, Global Avg Loss: 1.92507803, Time: 0.0219 Steps: 20640, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001489, Sample Num: 23824, Cur Loss: 0.16407904, Cur Avg Loss: 0.38243198, Log Avg loss: 0.48426735, Global Avg Loss: 1.92438030, Time: 0.0219 Steps: 20650, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001499, Sample Num: 23984, Cur Loss: 0.59004360, Cur Avg Loss: 0.38181235, Log Avg loss: 0.28954818, Global Avg Loss: 1.92358899, Time: 0.0219 Steps: 20660, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001509, Sample Num: 24144, Cur Loss: 0.05526332, Cur Avg Loss: 0.38191525, Log Avg loss: 0.39733976, Global Avg Loss: 1.92285061, Time: 0.0219 Steps: 20670, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001519, Sample Num: 24304, Cur Loss: 0.12478647, Cur Avg Loss: 0.38223897, Log Avg loss: 0.43108871, Global Avg Loss: 1.92212925, Time: 0.0219 Steps: 20680, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001529, Sample Num: 24464, Cur Loss: 0.31311417, Cur Avg Loss: 0.38191641, Log Avg loss: 0.33292034, Global Avg Loss: 1.92136115, Time: 0.0219 Steps: 20690, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001539, Sample Num: 24624, Cur Loss: 0.32999837, Cur Avg Loss: 0.38193923, Log Avg loss: 0.38542836, Global Avg Loss: 1.92061915, Time: 0.0246 Steps: 20700, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001549, Sample Num: 24784, Cur Loss: 0.24337840, Cur Avg Loss: 0.38134820, Log Avg loss: 0.29038787, Global Avg Loss: 1.91983198, Time: 0.0209 Steps: 20710, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001559, Sample Num: 24944, Cur Loss: 0.22466463, Cur Avg Loss: 0.38169039, Log Avg loss: 0.43469607, Global Avg Loss: 1.91911521, Time: 0.0208 Steps: 20720, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001569, Sample Num: 25104, Cur Loss: 0.20088804, Cur Avg Loss: 0.38127476, Log Avg loss: 0.31647869, Global Avg Loss: 1.91834211, Time: 0.0209 Steps: 20730, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001579, Sample Num: 25264, Cur Loss: 0.29225746, Cur Avg Loss: 0.38088980, Log Avg loss: 0.32048847, Global Avg Loss: 1.91757169, Time: 0.0208 Steps: 20740, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001589, Sample Num: 25424, Cur Loss: 0.20823956, Cur Avg Loss: 0.38007882, Log Avg loss: 0.25202550, Global Avg Loss: 1.91676902, Time: 0.0209 Steps: 20750, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001599, Sample Num: 25584, Cur Loss: 1.26305497, Cur Avg Loss: 0.38175495, Log Avg loss: 0.64809266, Global Avg Loss: 1.91615790, Time: 0.0209 Steps: 20760, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001609, Sample Num: 25744, Cur Loss: 0.54129553, Cur Avg Loss: 0.38322050, Log Avg loss: 0.61756207, Global Avg Loss: 1.91553268, Time: 0.0209 Steps: 20770, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001619, Sample Num: 25904, Cur Loss: 0.12099416, Cur Avg Loss: 0.38380319, Log Avg loss: 0.47755741, Global Avg Loss: 1.91484068, Time: 0.0209 Steps: 20780, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001629, Sample Num: 26064, Cur Loss: 0.70327747, Cur Avg Loss: 0.38393993, Log Avg loss: 0.40607775, Global Avg Loss: 1.91411496, Time: 0.0209 Steps: 20790, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001639, Sample Num: 26224, Cur Loss: 0.33278012, Cur Avg Loss: 0.38371027, Log Avg loss: 0.34629871, Global Avg Loss: 1.91336120, Time: 0.0209 Steps: 20800, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001649, Sample Num: 26384, Cur Loss: 1.13451874, Cur Avg Loss: 0.38366123, Log Avg loss: 0.37562343, Global Avg Loss: 1.91262226, Time: 0.0209 Steps: 20810, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001659, Sample Num: 26544, Cur Loss: 0.35180712, Cur Avg Loss: 0.38349814, Log Avg loss: 0.35660499, Global Avg Loss: 1.91187490, Time: 0.0209 Steps: 20820, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001669, Sample Num: 26704, Cur Loss: 0.22843593, Cur Avg Loss: 0.38314315, Log Avg loss: 0.32425075, Global Avg Loss: 1.91111271, Time: 0.0209 Steps: 20830, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001679, Sample Num: 26864, Cur Loss: 0.07537645, Cur Avg Loss: 0.38233983, Log Avg loss: 0.24826505, Global Avg Loss: 1.91031480, Time: 0.0209 Steps: 20840, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001689, Sample Num: 27024, Cur Loss: 0.17022932, Cur Avg Loss: 0.38157927, Log Avg loss: 0.25388170, Global Avg Loss: 1.90952035, Time: 0.0208 Steps: 20850, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001699, Sample Num: 27184, Cur Loss: 0.07424725, Cur Avg Loss: 0.38136823, Log Avg loss: 0.34572376, Global Avg Loss: 1.90877069, Time: 0.0209 Steps: 20860, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001709, Sample Num: 27344, Cur Loss: 0.12042303, Cur Avg Loss: 0.38082027, Log Avg loss: 0.28772214, Global Avg Loss: 1.90799395, Time: 0.0209 Steps: 20870, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001719, Sample Num: 27504, Cur Loss: 0.21498308, Cur Avg Loss: 0.38017218, Log Avg loss: 0.26941286, Global Avg Loss: 1.90720919, Time: 0.0209 Steps: 20880, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001729, Sample Num: 27664, Cur Loss: 0.13548861, Cur Avg Loss: 0.38045307, Log Avg loss: 0.42873759, Global Avg Loss: 1.90650145, Time: 0.0209 Steps: 20890, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001739, Sample Num: 27824, Cur Loss: 0.50992095, Cur Avg Loss: 0.38080782, Log Avg loss: 0.44214472, Global Avg Loss: 1.90580080, Time: 0.0209 Steps: 20900, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001749, Sample Num: 27984, Cur Loss: 0.47662020, Cur Avg Loss: 0.38241833, Log Avg loss: 0.66248577, Global Avg Loss: 1.90520620, Time: 0.0209 Steps: 20910, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001759, Sample Num: 28144, Cur Loss: 0.15157820, Cur Avg Loss: 0.38200773, Log Avg loss: 0.31019355, Global Avg Loss: 1.90444376, Time: 0.0209 Steps: 20920, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001769, Sample Num: 28304, Cur Loss: 0.29811883, Cur Avg Loss: 0.38156774, Log Avg loss: 0.30417423, Global Avg Loss: 1.90367918, Time: 0.0208 Steps: 20930, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001779, Sample Num: 28464, Cur Loss: 0.20275834, Cur Avg Loss: 0.38150579, Log Avg loss: 0.37054727, Global Avg Loss: 1.90294703, Time: 0.0209 Steps: 20940, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001789, Sample Num: 28624, Cur Loss: 0.15510777, Cur Avg Loss: 0.38115381, Log Avg loss: 0.31853516, Global Avg Loss: 1.90219074, Time: 0.0209 Steps: 20950, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001799, Sample Num: 28784, Cur Loss: 0.24806072, Cur Avg Loss: 0.38110462, Log Avg loss: 0.37230475, Global Avg Loss: 1.90146084, Time: 0.0209 Steps: 20960, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001809, Sample Num: 28944, Cur Loss: 0.44130710, Cur Avg Loss: 0.38114614, Log Avg loss: 0.38861573, Global Avg Loss: 1.90073940, Time: 0.0209 Steps: 20970, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001819, Sample Num: 29104, Cur Loss: 0.59875667, Cur Avg Loss: 0.38115774, Log Avg loss: 0.38325660, Global Avg Loss: 1.90001610, Time: 0.0209 Steps: 20980, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001829, Sample Num: 29264, Cur Loss: 0.10870246, Cur Avg Loss: 0.38096160, Log Avg loss: 0.34528363, Global Avg Loss: 1.89927540, Time: 0.0209 Steps: 20990, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001839, Sample Num: 29424, Cur Loss: 0.32671040, Cur Avg Loss: 0.38085381, Log Avg loss: 0.36113902, Global Avg Loss: 1.89854296, Time: 0.0209 Steps: 21000, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001849, Sample Num: 29584, Cur Loss: 0.46685749, Cur Avg Loss: 0.38008576, Log Avg loss: 0.23884177, Global Avg Loss: 1.89775300, Time: 0.0209 Steps: 21010, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001859, Sample Num: 29744, Cur Loss: 2.15363455, Cur Avg Loss: 0.38253116, Log Avg loss: 0.83468496, Global Avg Loss: 1.89724726, Time: 0.0209 Steps: 21020, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001869, Sample Num: 29904, Cur Loss: 0.58588171, Cur Avg Loss: 0.38379633, Log Avg loss: 0.61899171, Global Avg Loss: 1.89663943, Time: 0.0209 Steps: 21030, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001879, Sample Num: 30064, Cur Loss: 0.13564309, Cur Avg Loss: 0.38629694, Log Avg loss: 0.85366109, Global Avg Loss: 1.89614372, Time: 0.0209 Steps: 21040, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001889, Sample Num: 30224, Cur Loss: 0.19985956, Cur Avg Loss: 0.38620866, Log Avg loss: 0.36962076, Global Avg Loss: 1.89541853, Time: 0.0209 Steps: 21050, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001899, Sample Num: 30384, Cur Loss: 0.63458145, Cur Avg Loss: 0.38630188, Log Avg loss: 0.40391005, Global Avg Loss: 1.89471031, Time: 0.0209 Steps: 21060, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001909, Sample Num: 30544, Cur Loss: 0.50002384, Cur Avg Loss: 0.38656672, Log Avg loss: 0.43686047, Global Avg Loss: 1.89401840, Time: 0.0209 Steps: 21070, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001919, Sample Num: 30704, Cur Loss: 0.37611908, Cur Avg Loss: 0.38714006, Log Avg loss: 0.49659089, Global Avg Loss: 1.89335549, Time: 0.0209 Steps: 21080, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001929, Sample Num: 30864, Cur Loss: 0.27883738, Cur Avg Loss: 0.38672524, Log Avg loss: 0.30712202, Global Avg Loss: 1.89260336, Time: 0.0209 Steps: 21090, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001939, Sample Num: 31024, Cur Loss: 0.21940532, Cur Avg Loss: 0.38585704, Log Avg loss: 0.21838129, Global Avg Loss: 1.89180989, Time: 0.0210 Steps: 21100, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001949, Sample Num: 31184, Cur Loss: 0.20129138, Cur Avg Loss: 0.38571690, Log Avg loss: 0.35854340, Global Avg Loss: 1.89108357, Time: 0.0209 Steps: 21110, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001959, Sample Num: 31344, Cur Loss: 0.47327265, Cur Avg Loss: 0.38576597, Log Avg loss: 0.39532861, Global Avg Loss: 1.89037535, Time: 0.0209 Steps: 21120, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001969, Sample Num: 31504, Cur Loss: 0.60402811, Cur Avg Loss: 0.38612908, Log Avg loss: 0.45726259, Global Avg Loss: 1.88969712, Time: 0.0209 Steps: 21130, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001979, Sample Num: 31664, Cur Loss: 0.46270555, Cur Avg Loss: 0.38642792, Log Avg loss: 0.44526974, Global Avg Loss: 1.88901385, Time: 0.0209 Steps: 21140, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001989, Sample Num: 31824, Cur Loss: 0.65424013, Cur Avg Loss: 0.38602537, Log Avg loss: 0.30636141, Global Avg Loss: 1.88826555, Time: 0.0209 Steps: 21150, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001999, Sample Num: 31984, Cur Loss: 0.41222826, Cur Avg Loss: 0.38595287, Log Avg loss: 0.37153250, Global Avg Loss: 1.88754876, Time: 0.0209 Steps: 21160, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002009, Sample Num: 32144, Cur Loss: 0.50967598, Cur Avg Loss: 0.38600503, Log Avg loss: 0.39643141, Global Avg Loss: 1.88684440, Time: 0.0209 Steps: 21170, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002019, Sample Num: 32304, Cur Loss: 0.43390021, Cur Avg Loss: 0.38577737, Log Avg loss: 0.34004015, Global Avg Loss: 1.88611409, Time: 0.0209 Steps: 21180, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002029, Sample Num: 32464, Cur Loss: 0.23657688, Cur Avg Loss: 0.38711569, Log Avg loss: 0.65732337, Global Avg Loss: 1.88553420, Time: 0.0209 Steps: 21190, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002039, Sample Num: 32624, Cur Loss: 0.46043617, Cur Avg Loss: 0.38740874, Log Avg loss: 0.44686753, Global Avg Loss: 1.88485558, Time: 0.0208 Steps: 21200, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002049, Sample Num: 32784, Cur Loss: 0.38074115, Cur Avg Loss: 0.38707416, Log Avg loss: 0.31885367, Global Avg Loss: 1.88411725, Time: 0.0246 Steps: 21210, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002059, Sample Num: 32944, Cur Loss: 0.21998551, Cur Avg Loss: 0.38656558, Log Avg loss: 0.28235690, Global Avg Loss: 1.88336241, Time: 0.0210 Steps: 21220, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002069, Sample Num: 33104, Cur Loss: 0.62738818, Cur Avg Loss: 0.38621918, Log Avg loss: 0.31489654, Global Avg Loss: 1.88262362, Time: 0.0209 Steps: 21230, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002079, Sample Num: 33264, Cur Loss: 0.43402934, Cur Avg Loss: 0.38590499, Log Avg loss: 0.32089908, Global Avg Loss: 1.88188834, Time: 0.0210 Steps: 21240, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002089, Sample Num: 33424, Cur Loss: 0.19795269, Cur Avg Loss: 0.38540247, Log Avg loss: 0.28092795, Global Avg Loss: 1.88113495, Time: 0.0210 Steps: 21250, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002099, Sample Num: 33584, Cur Loss: 0.36851716, Cur Avg Loss: 0.38504468, Log Avg loss: 0.31030192, Global Avg Loss: 1.88039608, Time: 0.0210 Steps: 21260, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002109, Sample Num: 33744, Cur Loss: 0.18184695, Cur Avg Loss: 0.38461978, Log Avg loss: 0.29543494, Global Avg Loss: 1.87965092, Time: 0.0210 Steps: 21270, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002119, Sample Num: 33904, Cur Loss: 0.57082200, Cur Avg Loss: 0.38441937, Log Avg loss: 0.34215135, Global Avg Loss: 1.87892841, Time: 0.0209 Steps: 21280, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002129, Sample Num: 34055, Cur Loss: 0.13365056, Cur Avg Loss: 0.38465648, Log Avg loss: 0.43490175, Global Avg Loss: 1.87825014, Time: 0.0100 Steps: 21290, Updated lr: 0.000081 ***** Running evaluation checkpoint-21290 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-21290 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.887171, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.329853, "eval_total_loss": 231.886805, "eval_mae": 0.423761, "eval_mse": 0.32991, "eval_r2": 0.790288, "eval_sp_statistic": 0.8771, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.902734, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.370896, "test_total_loss": 186.190033, "test_mae": 0.367527, "test_mse": 0.371033, "test_r2": 0.760532, "test_sp_statistic": 0.871281, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.923815, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.8782501440582742, "train_cur_epoch_loss": 818.9336561225355, "train_cur_epoch_avg_loss": 0.3846564847921726, "train_cur_epoch_time": 44.887171268463135, "train_cur_epoch_avg_time": 0.02108368777288076, "epoch": 10, "step": 21290} ################################################## Training, Epoch: 0011, Batch: 000010, Sample Num: 160, Cur Loss: 0.45932621, Cur Avg Loss: 0.59006971, Log Avg loss: 0.59006971, Global Avg Loss: 1.87764536, Time: 0.0212 Steps: 21300, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000020, Sample Num: 320, Cur Loss: 0.19252226, Cur Avg Loss: 0.60710694, Log Avg loss: 0.62414418, Global Avg Loss: 1.87705714, Time: 0.0210 Steps: 21310, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000030, Sample Num: 480, Cur Loss: 0.38482910, Cur Avg Loss: 0.56349447, Log Avg loss: 0.47626952, Global Avg Loss: 1.87640011, Time: 0.0219 Steps: 21320, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000040, Sample Num: 640, Cur Loss: 0.67016804, Cur Avg Loss: 0.56087702, Log Avg loss: 0.55302467, Global Avg Loss: 1.87577968, Time: 0.0211 Steps: 21330, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000050, Sample Num: 800, Cur Loss: 0.11197205, Cur Avg Loss: 0.50703476, Log Avg loss: 0.29166574, Global Avg Loss: 1.87503736, Time: 0.0211 Steps: 21340, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000060, Sample Num: 960, Cur Loss: 0.43821394, Cur Avg Loss: 0.46785854, Log Avg loss: 0.27197741, Global Avg Loss: 1.87428651, Time: 0.0219 Steps: 21350, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000070, Sample Num: 1120, Cur Loss: 0.29567137, Cur Avg Loss: 0.49885654, Log Avg loss: 0.68484455, Global Avg Loss: 1.87372966, Time: 0.0211 Steps: 21360, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000080, Sample Num: 1280, Cur Loss: 0.29468277, Cur Avg Loss: 0.49623901, Log Avg loss: 0.47791632, Global Avg Loss: 1.87307649, Time: 0.0219 Steps: 21370, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000090, Sample Num: 1440, Cur Loss: 0.13507879, Cur Avg Loss: 0.48513625, Log Avg loss: 0.39631415, Global Avg Loss: 1.87238577, Time: 0.0210 Steps: 21380, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000100, Sample Num: 1600, Cur Loss: 0.43859583, Cur Avg Loss: 0.46296248, Log Avg loss: 0.26339852, Global Avg Loss: 1.87163356, Time: 0.0219 Steps: 21390, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000110, Sample Num: 1760, Cur Loss: 0.18210210, Cur Avg Loss: 0.44546521, Log Avg loss: 0.27049249, Global Avg Loss: 1.87088536, Time: 0.0210 Steps: 21400, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000120, Sample Num: 1920, Cur Loss: 0.20763199, Cur Avg Loss: 0.45136401, Log Avg loss: 0.51625084, Global Avg Loss: 1.87025265, Time: 0.0211 Steps: 21410, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000130, Sample Num: 2080, Cur Loss: 0.16035572, Cur Avg Loss: 0.44032882, Log Avg loss: 0.30790651, Global Avg Loss: 1.86952326, Time: 0.0219 Steps: 21420, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000140, Sample Num: 2240, Cur Loss: 0.06491566, Cur Avg Loss: 0.42928043, Log Avg loss: 0.28565135, Global Avg Loss: 1.86878417, Time: 0.0212 Steps: 21430, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000150, Sample Num: 2400, Cur Loss: 0.26369140, Cur Avg Loss: 0.42044998, Log Avg loss: 0.29682373, Global Avg Loss: 1.86805098, Time: 0.0219 Steps: 21440, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000160, Sample Num: 2560, Cur Loss: 0.84616232, Cur Avg Loss: 0.43120600, Log Avg loss: 0.59254633, Global Avg Loss: 1.86745634, Time: 0.0210 Steps: 21450, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000170, Sample Num: 2720, Cur Loss: 0.30550775, Cur Avg Loss: 0.45505416, Log Avg loss: 0.83662473, Global Avg Loss: 1.86697599, Time: 0.0219 Steps: 21460, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000180, Sample Num: 2880, Cur Loss: 0.09224135, Cur Avg Loss: 0.46290357, Log Avg loss: 0.59634349, Global Avg Loss: 1.86638417, Time: 0.0210 Steps: 21470, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000190, Sample Num: 3040, Cur Loss: 0.59230459, Cur Avg Loss: 0.45610627, Log Avg loss: 0.33375482, Global Avg Loss: 1.86567066, Time: 0.0210 Steps: 21480, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000200, Sample Num: 3200, Cur Loss: 0.13190252, Cur Avg Loss: 0.44509355, Log Avg loss: 0.23585200, Global Avg Loss: 1.86491225, Time: 0.0210 Steps: 21490, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000210, Sample Num: 3360, Cur Loss: 0.42658558, Cur Avg Loss: 0.43772944, Log Avg loss: 0.29044713, Global Avg Loss: 1.86417994, Time: 0.0210 Steps: 21500, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000220, Sample Num: 3520, Cur Loss: 0.84059638, Cur Avg Loss: 0.44368347, Log Avg loss: 0.56871820, Global Avg Loss: 1.86357768, Time: 0.0210 Steps: 21510, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000230, Sample Num: 3680, Cur Loss: 1.10662889, Cur Avg Loss: 0.44945008, Log Avg loss: 0.57631555, Global Avg Loss: 1.86297951, Time: 0.0210 Steps: 21520, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000240, Sample Num: 3840, Cur Loss: 0.66239774, Cur Avg Loss: 0.46266577, Log Avg loss: 0.76662644, Global Avg Loss: 1.86247029, Time: 0.0212 Steps: 21530, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000250, Sample Num: 4000, Cur Loss: 0.25448057, Cur Avg Loss: 0.46325385, Log Avg loss: 0.47736792, Global Avg Loss: 1.86182725, Time: 0.0219 Steps: 21540, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000260, Sample Num: 4160, Cur Loss: 0.51358330, Cur Avg Loss: 0.45493340, Log Avg loss: 0.24692198, Global Avg Loss: 1.86107788, Time: 0.0247 Steps: 21550, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000270, Sample Num: 4320, Cur Loss: 0.09949385, Cur Avg Loss: 0.44431064, Log Avg loss: 0.16811903, Global Avg Loss: 1.86029265, Time: 0.0211 Steps: 21560, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000280, Sample Num: 4480, Cur Loss: 0.61418706, Cur Avg Loss: 0.44461244, Log Avg loss: 0.45276112, Global Avg Loss: 1.85964010, Time: 0.0211 Steps: 21570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000290, Sample Num: 4640, Cur Loss: 0.35109693, Cur Avg Loss: 0.44409087, Log Avg loss: 0.42948671, Global Avg Loss: 1.85897738, Time: 0.0211 Steps: 21580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000300, Sample Num: 4800, Cur Loss: 0.30074733, Cur Avg Loss: 0.44669388, Log Avg loss: 0.52218114, Global Avg Loss: 1.85835821, Time: 0.0211 Steps: 21590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000310, Sample Num: 4960, Cur Loss: 0.21821147, Cur Avg Loss: 0.44776544, Log Avg loss: 0.47991221, Global Avg Loss: 1.85772004, Time: 0.0211 Steps: 21600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000320, Sample Num: 5120, Cur Loss: 0.66615808, Cur Avg Loss: 0.44661057, Log Avg loss: 0.41080970, Global Avg Loss: 1.85705048, Time: 0.0211 Steps: 21610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000330, Sample Num: 5280, Cur Loss: 0.28153318, Cur Avg Loss: 0.44379855, Log Avg loss: 0.35381410, Global Avg Loss: 1.85635518, Time: 0.0211 Steps: 21620, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000340, Sample Num: 5440, Cur Loss: 0.48474979, Cur Avg Loss: 0.44212521, Log Avg loss: 0.38690484, Global Avg Loss: 1.85567583, Time: 0.0210 Steps: 21630, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000350, Sample Num: 5600, Cur Loss: 0.20734322, Cur Avg Loss: 0.43896663, Log Avg loss: 0.33157495, Global Avg Loss: 1.85497153, Time: 0.0211 Steps: 21640, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000360, Sample Num: 5760, Cur Loss: 0.32066453, Cur Avg Loss: 0.43402503, Log Avg loss: 0.26106891, Global Avg Loss: 1.85423532, Time: 0.0212 Steps: 21650, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000370, Sample Num: 5920, Cur Loss: 0.34334490, Cur Avg Loss: 0.43030617, Log Avg loss: 0.29642721, Global Avg Loss: 1.85351611, Time: 0.0211 Steps: 21660, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000380, Sample Num: 6080, Cur Loss: 0.33974791, Cur Avg Loss: 0.42689006, Log Avg loss: 0.30049424, Global Avg Loss: 1.85279944, Time: 0.0211 Steps: 21670, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000390, Sample Num: 6240, Cur Loss: 0.62607288, Cur Avg Loss: 0.42399864, Log Avg loss: 0.31412440, Global Avg Loss: 1.85208972, Time: 0.0211 Steps: 21680, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000400, Sample Num: 6400, Cur Loss: 0.16347827, Cur Avg Loss: 0.41887797, Log Avg loss: 0.21917214, Global Avg Loss: 1.85133687, Time: 0.0211 Steps: 21690, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000410, Sample Num: 6560, Cur Loss: 0.29822379, Cur Avg Loss: 0.41460122, Log Avg loss: 0.24353094, Global Avg Loss: 1.85059595, Time: 0.0211 Steps: 21700, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000420, Sample Num: 6720, Cur Loss: 0.30789924, Cur Avg Loss: 0.41126936, Log Avg loss: 0.27466310, Global Avg Loss: 1.84987005, Time: 0.0211 Steps: 21710, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000430, Sample Num: 6880, Cur Loss: 0.37612277, Cur Avg Loss: 0.41099216, Log Avg loss: 0.39934979, Global Avg Loss: 1.84920222, Time: 0.0211 Steps: 21720, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000440, Sample Num: 7040, Cur Loss: 0.08269224, Cur Avg Loss: 0.40927419, Log Avg loss: 0.33540163, Global Avg Loss: 1.84850558, Time: 0.0211 Steps: 21730, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000450, Sample Num: 7200, Cur Loss: 0.18624073, Cur Avg Loss: 0.40694741, Log Avg loss: 0.30456886, Global Avg Loss: 1.84779540, Time: 0.0211 Steps: 21740, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000460, Sample Num: 7360, Cur Loss: 0.49780253, Cur Avg Loss: 0.41146271, Log Avg loss: 0.61465129, Global Avg Loss: 1.84722843, Time: 0.0211 Steps: 21750, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000470, Sample Num: 7520, Cur Loss: 0.31776673, Cur Avg Loss: 0.41406881, Log Avg loss: 0.53394943, Global Avg Loss: 1.84662490, Time: 0.0211 Steps: 21760, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000480, Sample Num: 7680, Cur Loss: 0.34267390, Cur Avg Loss: 0.41210048, Log Avg loss: 0.31958926, Global Avg Loss: 1.84592346, Time: 0.0211 Steps: 21770, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000490, Sample Num: 7840, Cur Loss: 0.29717410, Cur Avg Loss: 0.40983602, Log Avg loss: 0.30114167, Global Avg Loss: 1.84521420, Time: 0.0210 Steps: 21780, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000500, Sample Num: 8000, Cur Loss: 0.15812799, Cur Avg Loss: 0.40934946, Log Avg loss: 0.38550802, Global Avg Loss: 1.84454430, Time: 0.0210 Steps: 21790, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000510, Sample Num: 8160, Cur Loss: 0.12222065, Cur Avg Loss: 0.40658243, Log Avg loss: 0.26823075, Global Avg Loss: 1.84382122, Time: 0.0211 Steps: 21800, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000520, Sample Num: 8320, Cur Loss: 0.19912641, Cur Avg Loss: 0.40578727, Log Avg loss: 0.36523425, Global Avg Loss: 1.84314328, Time: 0.0209 Steps: 21810, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000530, Sample Num: 8480, Cur Loss: 0.37099418, Cur Avg Loss: 0.40454881, Log Avg loss: 0.34014876, Global Avg Loss: 1.84245447, Time: 0.0209 Steps: 21820, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000540, Sample Num: 8640, Cur Loss: 0.26856703, Cur Avg Loss: 0.40288456, Log Avg loss: 0.31467938, Global Avg Loss: 1.84175461, Time: 0.0209 Steps: 21830, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000550, Sample Num: 8800, Cur Loss: 0.39785698, Cur Avg Loss: 0.40001893, Log Avg loss: 0.24527522, Global Avg Loss: 1.84102363, Time: 0.0209 Steps: 21840, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000560, Sample Num: 8960, Cur Loss: 0.34535980, Cur Avg Loss: 0.39729127, Log Avg loss: 0.24726991, Global Avg Loss: 1.84029422, Time: 0.0211 Steps: 21850, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000570, Sample Num: 9120, Cur Loss: 0.81354201, Cur Avg Loss: 0.39928137, Log Avg loss: 0.51072662, Global Avg Loss: 1.83968600, Time: 0.0210 Steps: 21860, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000580, Sample Num: 9280, Cur Loss: 0.88238329, Cur Avg Loss: 0.40298245, Log Avg loss: 0.61394444, Global Avg Loss: 1.83912553, Time: 0.0209 Steps: 21870, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000590, Sample Num: 9440, Cur Loss: 0.52235472, Cur Avg Loss: 0.40379433, Log Avg loss: 0.45088341, Global Avg Loss: 1.83849105, Time: 0.0211 Steps: 21880, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000600, Sample Num: 9600, Cur Loss: 0.39581919, Cur Avg Loss: 0.40257441, Log Avg loss: 0.33059881, Global Avg Loss: 1.83780220, Time: 0.0210 Steps: 21890, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000610, Sample Num: 9760, Cur Loss: 0.17667070, Cur Avg Loss: 0.39944960, Log Avg loss: 0.21196118, Global Avg Loss: 1.83705981, Time: 0.0209 Steps: 21900, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000620, Sample Num: 9920, Cur Loss: 0.27728438, Cur Avg Loss: 0.40215325, Log Avg loss: 0.56707554, Global Avg Loss: 1.83648017, Time: 0.0209 Steps: 21910, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000630, Sample Num: 10080, Cur Loss: 0.29046994, Cur Avg Loss: 0.40248666, Log Avg loss: 0.42315829, Global Avg Loss: 1.83583541, Time: 0.0209 Steps: 21920, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000640, Sample Num: 10240, Cur Loss: 0.21406540, Cur Avg Loss: 0.40085186, Log Avg loss: 0.29785925, Global Avg Loss: 1.83513410, Time: 0.0210 Steps: 21930, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000650, Sample Num: 10400, Cur Loss: 0.38669950, Cur Avg Loss: 0.39808016, Log Avg loss: 0.22069152, Global Avg Loss: 1.83439825, Time: 0.0210 Steps: 21940, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000660, Sample Num: 10560, Cur Loss: 0.10705587, Cur Avg Loss: 0.39642914, Log Avg loss: 0.28911273, Global Avg Loss: 1.83369425, Time: 0.0210 Steps: 21950, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000670, Sample Num: 10720, Cur Loss: 0.53059369, Cur Avg Loss: 0.39555252, Log Avg loss: 0.33769607, Global Avg Loss: 1.83301301, Time: 0.0210 Steps: 21960, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000680, Sample Num: 10880, Cur Loss: 0.10017308, Cur Avg Loss: 0.39344724, Log Avg loss: 0.25239349, Global Avg Loss: 1.83229357, Time: 0.0211 Steps: 21970, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000690, Sample Num: 11040, Cur Loss: 0.20532194, Cur Avg Loss: 0.39251585, Log Avg loss: 0.32918090, Global Avg Loss: 1.83160971, Time: 0.0209 Steps: 21980, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000700, Sample Num: 11200, Cur Loss: 0.29979351, Cur Avg Loss: 0.39319505, Log Avg loss: 0.44005993, Global Avg Loss: 1.83097690, Time: 0.0209 Steps: 21990, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000710, Sample Num: 11360, Cur Loss: 0.47561672, Cur Avg Loss: 0.39211478, Log Avg loss: 0.31649615, Global Avg Loss: 1.83028850, Time: 0.0209 Steps: 22000, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000720, Sample Num: 11520, Cur Loss: 0.11991984, Cur Avg Loss: 0.39106865, Log Avg loss: 0.31679313, Global Avg Loss: 1.82960086, Time: 0.0209 Steps: 22010, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000730, Sample Num: 11680, Cur Loss: 0.44552046, Cur Avg Loss: 0.39188553, Log Avg loss: 0.45070078, Global Avg Loss: 1.82897466, Time: 0.0209 Steps: 22020, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000740, Sample Num: 11840, Cur Loss: 0.36108452, Cur Avg Loss: 0.39090011, Log Avg loss: 0.31896436, Global Avg Loss: 1.82828923, Time: 0.0209 Steps: 22030, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000750, Sample Num: 12000, Cur Loss: 0.95126903, Cur Avg Loss: 0.39056453, Log Avg loss: 0.36573182, Global Avg Loss: 1.82762563, Time: 0.0208 Steps: 22040, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000760, Sample Num: 12160, Cur Loss: 0.24148639, Cur Avg Loss: 0.38958950, Log Avg loss: 0.31646256, Global Avg Loss: 1.82694030, Time: 0.0210 Steps: 22050, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000770, Sample Num: 12320, Cur Loss: 0.27754897, Cur Avg Loss: 0.39017783, Log Avg loss: 0.43489057, Global Avg Loss: 1.82630927, Time: 0.0247 Steps: 22060, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000780, Sample Num: 12480, Cur Loss: 0.22490111, Cur Avg Loss: 0.38894240, Log Avg loss: 0.29381448, Global Avg Loss: 1.82561489, Time: 0.0209 Steps: 22070, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000790, Sample Num: 12640, Cur Loss: 0.23655745, Cur Avg Loss: 0.38784152, Log Avg loss: 0.30197261, Global Avg Loss: 1.82492484, Time: 0.0209 Steps: 22080, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000800, Sample Num: 12800, Cur Loss: 0.23646650, Cur Avg Loss: 0.38563724, Log Avg loss: 0.21149924, Global Avg Loss: 1.82419445, Time: 0.0209 Steps: 22090, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000810, Sample Num: 12960, Cur Loss: 0.19870299, Cur Avg Loss: 0.38410168, Log Avg loss: 0.26125722, Global Avg Loss: 1.82348724, Time: 0.0210 Steps: 22100, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000820, Sample Num: 13120, Cur Loss: 0.09007441, Cur Avg Loss: 0.38294426, Log Avg loss: 0.28919329, Global Avg Loss: 1.82279330, Time: 0.0209 Steps: 22110, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000830, Sample Num: 13280, Cur Loss: 0.24294570, Cur Avg Loss: 0.38191169, Log Avg loss: 0.29724093, Global Avg Loss: 1.82210363, Time: 0.0209 Steps: 22120, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000840, Sample Num: 13440, Cur Loss: 0.23955132, Cur Avg Loss: 0.38210527, Log Avg loss: 0.39817218, Global Avg Loss: 1.82146019, Time: 0.0209 Steps: 22130, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000850, Sample Num: 13600, Cur Loss: 0.68470728, Cur Avg Loss: 0.38293577, Log Avg loss: 0.45269797, Global Avg Loss: 1.82084196, Time: 0.0209 Steps: 22140, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000860, Sample Num: 13760, Cur Loss: 0.17238808, Cur Avg Loss: 0.38280035, Log Avg loss: 0.37128974, Global Avg Loss: 1.82018753, Time: 0.0211 Steps: 22150, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000870, Sample Num: 13920, Cur Loss: 0.47517079, Cur Avg Loss: 0.38335901, Log Avg loss: 0.43140306, Global Avg Loss: 1.81956083, Time: 0.0208 Steps: 22160, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000880, Sample Num: 14080, Cur Loss: 0.50532734, Cur Avg Loss: 0.38281454, Log Avg loss: 0.33544573, Global Avg Loss: 1.81889140, Time: 0.0208 Steps: 22170, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000890, Sample Num: 14240, Cur Loss: 0.14448047, Cur Avg Loss: 0.38251865, Log Avg loss: 0.35648039, Global Avg Loss: 1.81823206, Time: 0.0208 Steps: 22180, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000900, Sample Num: 14400, Cur Loss: 0.11447955, Cur Avg Loss: 0.38076108, Log Avg loss: 0.22433719, Global Avg Loss: 1.81751377, Time: 0.0208 Steps: 22190, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000910, Sample Num: 14560, Cur Loss: 0.37184945, Cur Avg Loss: 0.38031746, Log Avg loss: 0.34039234, Global Avg Loss: 1.81684840, Time: 0.0208 Steps: 22200, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000920, Sample Num: 14720, Cur Loss: 1.34213948, Cur Avg Loss: 0.38053417, Log Avg loss: 0.40025442, Global Avg Loss: 1.81621058, Time: 0.0208 Steps: 22210, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000930, Sample Num: 14880, Cur Loss: 0.60602093, Cur Avg Loss: 0.38098031, Log Avg loss: 0.42202553, Global Avg Loss: 1.81558313, Time: 0.0208 Steps: 22220, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000940, Sample Num: 15040, Cur Loss: 0.13974145, Cur Avg Loss: 0.37902841, Log Avg loss: 0.19750175, Global Avg Loss: 1.81485525, Time: 0.0208 Steps: 22230, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000950, Sample Num: 15200, Cur Loss: 0.18692622, Cur Avg Loss: 0.37794285, Log Avg loss: 0.27589954, Global Avg Loss: 1.81416328, Time: 0.0209 Steps: 22240, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000960, Sample Num: 15360, Cur Loss: 0.55208111, Cur Avg Loss: 0.38030355, Log Avg loss: 0.60457064, Global Avg Loss: 1.81361964, Time: 0.0208 Steps: 22250, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000970, Sample Num: 15520, Cur Loss: 0.27156982, Cur Avg Loss: 0.38040288, Log Avg loss: 0.38993791, Global Avg Loss: 1.81298007, Time: 0.0208 Steps: 22260, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000980, Sample Num: 15680, Cur Loss: 0.35939264, Cur Avg Loss: 0.37994570, Log Avg loss: 0.33559933, Global Avg Loss: 1.81231667, Time: 0.0208 Steps: 22270, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000990, Sample Num: 15840, Cur Loss: 0.85107958, Cur Avg Loss: 0.38288504, Log Avg loss: 0.67094057, Global Avg Loss: 1.81180439, Time: 0.0208 Steps: 22280, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001000, Sample Num: 16000, Cur Loss: 0.71116698, Cur Avg Loss: 0.38302897, Log Avg loss: 0.39727778, Global Avg Loss: 1.81116979, Time: 0.0208 Steps: 22290, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001010, Sample Num: 16160, Cur Loss: 0.22984841, Cur Avg Loss: 0.38171516, Log Avg loss: 0.25033417, Global Avg Loss: 1.81046986, Time: 0.0208 Steps: 22300, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001020, Sample Num: 16320, Cur Loss: 0.35310990, Cur Avg Loss: 0.38099048, Log Avg loss: 0.30779791, Global Avg Loss: 1.80979632, Time: 0.0208 Steps: 22310, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001030, Sample Num: 16480, Cur Loss: 0.73030752, Cur Avg Loss: 0.38053927, Log Avg loss: 0.33451601, Global Avg Loss: 1.80913535, Time: 0.0245 Steps: 22320, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001040, Sample Num: 16640, Cur Loss: 0.31885719, Cur Avg Loss: 0.37905755, Log Avg loss: 0.22644006, Global Avg Loss: 1.80842657, Time: 0.0208 Steps: 22330, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001050, Sample Num: 16800, Cur Loss: 0.26856476, Cur Avg Loss: 0.37830997, Log Avg loss: 0.30056200, Global Avg Loss: 1.80775161, Time: 0.0208 Steps: 22340, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001060, Sample Num: 16960, Cur Loss: 0.22503753, Cur Avg Loss: 0.38061151, Log Avg loss: 0.62227365, Global Avg Loss: 1.80722120, Time: 0.0209 Steps: 22350, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001070, Sample Num: 17120, Cur Loss: 0.33107182, Cur Avg Loss: 0.38105674, Log Avg loss: 0.42825033, Global Avg Loss: 1.80660448, Time: 0.0208 Steps: 22360, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001080, Sample Num: 17280, Cur Loss: 0.81311738, Cur Avg Loss: 0.38165661, Log Avg loss: 0.44584339, Global Avg Loss: 1.80599619, Time: 0.0208 Steps: 22370, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001090, Sample Num: 17440, Cur Loss: 0.23157954, Cur Avg Loss: 0.38191821, Log Avg loss: 0.41017083, Global Avg Loss: 1.80537249, Time: 0.0208 Steps: 22380, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001100, Sample Num: 17600, Cur Loss: 0.29165578, Cur Avg Loss: 0.38162946, Log Avg loss: 0.35015517, Global Avg Loss: 1.80472255, Time: 0.0208 Steps: 22390, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001110, Sample Num: 17760, Cur Loss: 0.19043325, Cur Avg Loss: 0.38071373, Log Avg loss: 0.27998362, Global Avg Loss: 1.80404187, Time: 0.0208 Steps: 22400, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001120, Sample Num: 17920, Cur Loss: 0.21618108, Cur Avg Loss: 0.38020157, Log Avg loss: 0.32335246, Global Avg Loss: 1.80338114, Time: 0.0208 Steps: 22410, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001130, Sample Num: 18080, Cur Loss: 0.16149515, Cur Avg Loss: 0.37926262, Log Avg loss: 0.27409970, Global Avg Loss: 1.80269903, Time: 0.0208 Steps: 22420, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001140, Sample Num: 18240, Cur Loss: 0.08583222, Cur Avg Loss: 0.37812542, Log Avg loss: 0.24962171, Global Avg Loss: 1.80200662, Time: 0.0208 Steps: 22430, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001150, Sample Num: 18400, Cur Loss: 0.50166601, Cur Avg Loss: 0.37886760, Log Avg loss: 0.46347632, Global Avg Loss: 1.80141013, Time: 0.0208 Steps: 22440, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001160, Sample Num: 18560, Cur Loss: 0.36546323, Cur Avg Loss: 0.37823128, Log Avg loss: 0.30505479, Global Avg Loss: 1.80074360, Time: 0.0208 Steps: 22450, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001170, Sample Num: 18720, Cur Loss: 0.08999770, Cur Avg Loss: 0.37721147, Log Avg loss: 0.25891324, Global Avg Loss: 1.80005712, Time: 0.0208 Steps: 22460, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001180, Sample Num: 18880, Cur Loss: 0.26742384, Cur Avg Loss: 0.37728989, Log Avg loss: 0.38646450, Global Avg Loss: 1.79942802, Time: 0.0208 Steps: 22470, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001190, Sample Num: 19040, Cur Loss: 0.35820088, Cur Avg Loss: 0.37717730, Log Avg loss: 0.36389190, Global Avg Loss: 1.79878944, Time: 0.0208 Steps: 22480, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001200, Sample Num: 19200, Cur Loss: 0.29034933, Cur Avg Loss: 0.37732895, Log Avg loss: 0.39537516, Global Avg Loss: 1.79816542, Time: 0.0208 Steps: 22490, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001210, Sample Num: 19360, Cur Loss: 0.21946310, Cur Avg Loss: 0.37696302, Log Avg loss: 0.33305162, Global Avg Loss: 1.79751426, Time: 0.0208 Steps: 22500, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001220, Sample Num: 19520, Cur Loss: 0.09244198, Cur Avg Loss: 0.37594356, Log Avg loss: 0.25258898, Global Avg Loss: 1.79682793, Time: 0.0208 Steps: 22510, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001230, Sample Num: 19680, Cur Loss: 0.19947976, Cur Avg Loss: 0.37577149, Log Avg loss: 0.35477887, Global Avg Loss: 1.79618759, Time: 0.0208 Steps: 22520, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001240, Sample Num: 19840, Cur Loss: 0.21623293, Cur Avg Loss: 0.37581902, Log Avg loss: 0.38166533, Global Avg Loss: 1.79555975, Time: 0.0208 Steps: 22530, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001250, Sample Num: 20000, Cur Loss: 0.23053437, Cur Avg Loss: 0.37529312, Log Avg loss: 0.31008190, Global Avg Loss: 1.79490071, Time: 0.0208 Steps: 22540, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001260, Sample Num: 20160, Cur Loss: 0.29811329, Cur Avg Loss: 0.37552684, Log Avg loss: 0.40474196, Global Avg Loss: 1.79428423, Time: 0.0208 Steps: 22550, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001270, Sample Num: 20320, Cur Loss: 0.40571600, Cur Avg Loss: 0.37553096, Log Avg loss: 0.37604966, Global Avg Loss: 1.79365558, Time: 0.0208 Steps: 22560, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001280, Sample Num: 20480, Cur Loss: 0.79949117, Cur Avg Loss: 0.37470009, Log Avg loss: 0.26918007, Global Avg Loss: 1.79298014, Time: 0.0253 Steps: 22570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001290, Sample Num: 20640, Cur Loss: 0.55288255, Cur Avg Loss: 0.37401155, Log Avg loss: 0.28587837, Global Avg Loss: 1.79231269, Time: 0.0208 Steps: 22580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001300, Sample Num: 20800, Cur Loss: 0.42666304, Cur Avg Loss: 0.37478285, Log Avg loss: 0.47428027, Global Avg Loss: 1.79172923, Time: 0.0208 Steps: 22590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001310, Sample Num: 20960, Cur Loss: 0.29553416, Cur Avg Loss: 0.37459665, Log Avg loss: 0.35039013, Global Avg Loss: 1.79109147, Time: 0.0208 Steps: 22600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001320, Sample Num: 21120, Cur Loss: 0.48062003, Cur Avg Loss: 0.37405976, Log Avg loss: 0.30372756, Global Avg Loss: 1.79043363, Time: 0.0209 Steps: 22610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001330, Sample Num: 21280, Cur Loss: 0.91326189, Cur Avg Loss: 0.37445194, Log Avg loss: 0.42621945, Global Avg Loss: 1.78983053, Time: 0.0208 Steps: 22620, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001340, Sample Num: 21440, Cur Loss: 0.48496118, Cur Avg Loss: 0.37600081, Log Avg loss: 0.58200096, Global Avg Loss: 1.78929680, Time: 0.0209 Steps: 22630, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001350, Sample Num: 21600, Cur Loss: 0.25495914, Cur Avg Loss: 0.37616769, Log Avg loss: 0.39852899, Global Avg Loss: 1.78868251, Time: 0.0209 Steps: 22640, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001360, Sample Num: 21760, Cur Loss: 0.18688384, Cur Avg Loss: 0.37592618, Log Avg loss: 0.34332287, Global Avg Loss: 1.78804438, Time: 0.0208 Steps: 22650, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001370, Sample Num: 21920, Cur Loss: 0.43360808, Cur Avg Loss: 0.37482852, Log Avg loss: 0.22554720, Global Avg Loss: 1.78735484, Time: 0.0209 Steps: 22660, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001380, Sample Num: 22080, Cur Loss: 0.12684761, Cur Avg Loss: 0.37421443, Log Avg loss: 0.29008308, Global Avg Loss: 1.78669437, Time: 0.0208 Steps: 22670, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001390, Sample Num: 22240, Cur Loss: 0.13853222, Cur Avg Loss: 0.37345771, Log Avg loss: 0.26903112, Global Avg Loss: 1.78602521, Time: 0.0208 Steps: 22680, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001400, Sample Num: 22400, Cur Loss: 0.34437791, Cur Avg Loss: 0.37287769, Log Avg loss: 0.29225460, Global Avg Loss: 1.78536687, Time: 0.0209 Steps: 22690, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001410, Sample Num: 22560, Cur Loss: 0.17765574, Cur Avg Loss: 0.37255754, Log Avg loss: 0.32773690, Global Avg Loss: 1.78472474, Time: 0.0209 Steps: 22700, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001420, Sample Num: 22720, Cur Loss: 0.42944676, Cur Avg Loss: 0.37239142, Log Avg loss: 0.34896781, Global Avg Loss: 1.78409253, Time: 0.0208 Steps: 22710, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001430, Sample Num: 22880, Cur Loss: 0.12438353, Cur Avg Loss: 0.37202580, Log Avg loss: 0.32010820, Global Avg Loss: 1.78344817, Time: 0.0208 Steps: 22720, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001440, Sample Num: 23040, Cur Loss: 0.21243012, Cur Avg Loss: 0.37235782, Log Avg loss: 0.41983674, Global Avg Loss: 1.78284825, Time: 0.0208 Steps: 22730, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001450, Sample Num: 23200, Cur Loss: 0.21102946, Cur Avg Loss: 0.37198385, Log Avg loss: 0.31813166, Global Avg Loss: 1.78220414, Time: 0.0208 Steps: 22740, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001460, Sample Num: 23360, Cur Loss: 0.37488699, Cur Avg Loss: 0.37166390, Log Avg loss: 0.32527207, Global Avg Loss: 1.78156373, Time: 0.0209 Steps: 22750, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001470, Sample Num: 23520, Cur Loss: 0.44801927, Cur Avg Loss: 0.37283845, Log Avg loss: 0.54432172, Global Avg Loss: 1.78102013, Time: 0.0209 Steps: 22760, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001480, Sample Num: 23680, Cur Loss: 0.26749688, Cur Avg Loss: 0.37217394, Log Avg loss: 0.27449092, Global Avg Loss: 1.78035850, Time: 0.0208 Steps: 22770, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001490, Sample Num: 23840, Cur Loss: 0.18533766, Cur Avg Loss: 0.37283123, Log Avg loss: 0.47011134, Global Avg Loss: 1.77978332, Time: 0.0209 Steps: 22780, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001500, Sample Num: 24000, Cur Loss: 0.27971965, Cur Avg Loss: 0.37212098, Log Avg loss: 0.26629269, Global Avg Loss: 1.77911922, Time: 0.0208 Steps: 22790, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001510, Sample Num: 24160, Cur Loss: 0.67150795, Cur Avg Loss: 0.37232019, Log Avg loss: 0.40220137, Global Avg Loss: 1.77851531, Time: 0.0209 Steps: 22800, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001520, Sample Num: 24320, Cur Loss: 0.22750594, Cur Avg Loss: 0.37176296, Log Avg loss: 0.28762144, Global Avg Loss: 1.77786169, Time: 0.0208 Steps: 22810, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001530, Sample Num: 24480, Cur Loss: 0.43156448, Cur Avg Loss: 0.37196827, Log Avg loss: 0.40317590, Global Avg Loss: 1.77725929, Time: 0.0208 Steps: 22820, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001540, Sample Num: 24640, Cur Loss: 0.11794633, Cur Avg Loss: 0.37139318, Log Avg loss: 0.28340500, Global Avg Loss: 1.77660495, Time: 0.0246 Steps: 22830, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001550, Sample Num: 24800, Cur Loss: 0.23924413, Cur Avg Loss: 0.37080868, Log Avg loss: 0.28079440, Global Avg Loss: 1.77595004, Time: 0.0208 Steps: 22840, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001560, Sample Num: 24960, Cur Loss: 0.10162195, Cur Avg Loss: 0.37140073, Log Avg loss: 0.46316990, Global Avg Loss: 1.77537552, Time: 0.0208 Steps: 22850, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001570, Sample Num: 25120, Cur Loss: 0.17146751, Cur Avg Loss: 0.37084446, Log Avg loss: 0.28406590, Global Avg Loss: 1.77472316, Time: 0.0208 Steps: 22860, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001580, Sample Num: 25280, Cur Loss: 0.12908149, Cur Avg Loss: 0.37096540, Log Avg loss: 0.38995220, Global Avg Loss: 1.77411766, Time: 0.0209 Steps: 22870, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001590, Sample Num: 25440, Cur Loss: 0.20491526, Cur Avg Loss: 0.36997663, Log Avg loss: 0.21375213, Global Avg Loss: 1.77343568, Time: 0.0208 Steps: 22880, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001600, Sample Num: 25600, Cur Loss: 0.34023356, Cur Avg Loss: 0.36911176, Log Avg loss: 0.23159716, Global Avg Loss: 1.77276210, Time: 0.0209 Steps: 22890, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001610, Sample Num: 25760, Cur Loss: 0.70216238, Cur Avg Loss: 0.36914007, Log Avg loss: 0.37366900, Global Avg Loss: 1.77215114, Time: 0.0208 Steps: 22900, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001620, Sample Num: 25920, Cur Loss: 0.49242717, Cur Avg Loss: 0.36954583, Log Avg loss: 0.43487408, Global Avg Loss: 1.77156743, Time: 0.0208 Steps: 22910, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001630, Sample Num: 26080, Cur Loss: 0.26920235, Cur Avg Loss: 0.36934963, Log Avg loss: 0.33756477, Global Avg Loss: 1.77094177, Time: 0.0209 Steps: 22920, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001640, Sample Num: 26240, Cur Loss: 0.23388408, Cur Avg Loss: 0.36945360, Log Avg loss: 0.38640108, Global Avg Loss: 1.77033796, Time: 0.0208 Steps: 22930, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001650, Sample Num: 26400, Cur Loss: 0.28185955, Cur Avg Loss: 0.36919828, Log Avg loss: 0.32732545, Global Avg Loss: 1.76970892, Time: 0.0209 Steps: 22940, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001660, Sample Num: 26560, Cur Loss: 0.22186238, Cur Avg Loss: 0.36880313, Log Avg loss: 0.30360395, Global Avg Loss: 1.76907010, Time: 0.0209 Steps: 22950, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001670, Sample Num: 26720, Cur Loss: 1.25098872, Cur Avg Loss: 0.36901498, Log Avg loss: 0.40418109, Global Avg Loss: 1.76847564, Time: 0.0208 Steps: 22960, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001680, Sample Num: 26880, Cur Loss: 1.21222353, Cur Avg Loss: 0.37103534, Log Avg loss: 0.70843568, Global Avg Loss: 1.76801415, Time: 0.0208 Steps: 22970, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001690, Sample Num: 27040, Cur Loss: 0.36583742, Cur Avg Loss: 0.37045012, Log Avg loss: 0.27213392, Global Avg Loss: 1.76736320, Time: 0.0208 Steps: 22980, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001700, Sample Num: 27200, Cur Loss: 0.14633420, Cur Avg Loss: 0.37118081, Log Avg loss: 0.49466662, Global Avg Loss: 1.76680961, Time: 0.0208 Steps: 22990, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001710, Sample Num: 27360, Cur Loss: 0.33278501, Cur Avg Loss: 0.37161495, Log Avg loss: 0.44541874, Global Avg Loss: 1.76623509, Time: 0.0208 Steps: 23000, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001720, Sample Num: 27520, Cur Loss: 0.76103699, Cur Avg Loss: 0.37248270, Log Avg loss: 0.52086761, Global Avg Loss: 1.76569386, Time: 0.0208 Steps: 23010, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001730, Sample Num: 27680, Cur Loss: 0.56839836, Cur Avg Loss: 0.37316411, Log Avg loss: 0.49036758, Global Avg Loss: 1.76513986, Time: 0.0208 Steps: 23020, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001740, Sample Num: 27840, Cur Loss: 0.31772202, Cur Avg Loss: 0.37316527, Log Avg loss: 0.37336494, Global Avg Loss: 1.76453552, Time: 0.0209 Steps: 23030, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001750, Sample Num: 28000, Cur Loss: 0.36045486, Cur Avg Loss: 0.37266292, Log Avg loss: 0.28525451, Global Avg Loss: 1.76389348, Time: 0.0208 Steps: 23040, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001760, Sample Num: 28160, Cur Loss: 0.29923296, Cur Avg Loss: 0.37203243, Log Avg loss: 0.26169629, Global Avg Loss: 1.76324176, Time: 0.0208 Steps: 23050, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001770, Sample Num: 28320, Cur Loss: 0.29320398, Cur Avg Loss: 0.37142483, Log Avg loss: 0.26448793, Global Avg Loss: 1.76259183, Time: 0.0208 Steps: 23060, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001780, Sample Num: 28480, Cur Loss: 0.19334571, Cur Avg Loss: 0.37098706, Log Avg loss: 0.29350173, Global Avg Loss: 1.76195503, Time: 0.0210 Steps: 23070, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001790, Sample Num: 28640, Cur Loss: 0.36368945, Cur Avg Loss: 0.37044931, Log Avg loss: 0.27472910, Global Avg Loss: 1.76131065, Time: 0.0208 Steps: 23080, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001800, Sample Num: 28800, Cur Loss: 0.40721792, Cur Avg Loss: 0.37025068, Log Avg loss: 0.33469646, Global Avg Loss: 1.76069280, Time: 0.0209 Steps: 23090, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001810, Sample Num: 28960, Cur Loss: 0.45390654, Cur Avg Loss: 0.36966245, Log Avg loss: 0.26378045, Global Avg Loss: 1.76004479, Time: 0.0208 Steps: 23100, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001820, Sample Num: 29120, Cur Loss: 0.27340353, Cur Avg Loss: 0.36938539, Log Avg loss: 0.31923885, Global Avg Loss: 1.75942133, Time: 0.0208 Steps: 23110, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001830, Sample Num: 29280, Cur Loss: 0.40476206, Cur Avg Loss: 0.36924736, Log Avg loss: 0.34412448, Global Avg Loss: 1.75880918, Time: 0.0208 Steps: 23120, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001840, Sample Num: 29440, Cur Loss: 0.53198647, Cur Avg Loss: 0.36906434, Log Avg loss: 0.33557243, Global Avg Loss: 1.75819386, Time: 0.0209 Steps: 23130, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001850, Sample Num: 29600, Cur Loss: 0.33710679, Cur Avg Loss: 0.36963544, Log Avg loss: 0.47471818, Global Avg Loss: 1.75763920, Time: 0.0208 Steps: 23140, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001860, Sample Num: 29760, Cur Loss: 0.64855731, Cur Avg Loss: 0.36992387, Log Avg loss: 0.42328223, Global Avg Loss: 1.75706281, Time: 0.0208 Steps: 23150, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001870, Sample Num: 29920, Cur Loss: 0.25009242, Cur Avg Loss: 0.37009673, Log Avg loss: 0.40225013, Global Avg Loss: 1.75647783, Time: 0.0208 Steps: 23160, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001880, Sample Num: 30080, Cur Loss: 0.13172731, Cur Avg Loss: 0.37004038, Log Avg loss: 0.35950219, Global Avg Loss: 1.75587490, Time: 0.0208 Steps: 23170, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001890, Sample Num: 30240, Cur Loss: 0.19895548, Cur Avg Loss: 0.36947036, Log Avg loss: 0.26230659, Global Avg Loss: 1.75523057, Time: 0.0208 Steps: 23180, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001900, Sample Num: 30400, Cur Loss: 0.39370999, Cur Avg Loss: 0.36902349, Log Avg loss: 0.28456504, Global Avg Loss: 1.75459639, Time: 0.0208 Steps: 23190, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001910, Sample Num: 30560, Cur Loss: 0.36441499, Cur Avg Loss: 0.36906488, Log Avg loss: 0.37692924, Global Avg Loss: 1.75400256, Time: 0.0208 Steps: 23200, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001920, Sample Num: 30720, Cur Loss: 0.51576424, Cur Avg Loss: 0.36936313, Log Avg loss: 0.42632878, Global Avg Loss: 1.75343054, Time: 0.0208 Steps: 23210, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001930, Sample Num: 30880, Cur Loss: 0.12353420, Cur Avg Loss: 0.36892326, Log Avg loss: 0.28446734, Global Avg Loss: 1.75279791, Time: 0.0208 Steps: 23220, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001940, Sample Num: 31040, Cur Loss: 0.13543200, Cur Avg Loss: 0.36932342, Log Avg loss: 0.44655475, Global Avg Loss: 1.75223560, Time: 0.0208 Steps: 23230, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001950, Sample Num: 31200, Cur Loss: 1.30940890, Cur Avg Loss: 0.36977547, Log Avg loss: 0.45747415, Global Avg Loss: 1.75167847, Time: 0.0208 Steps: 23240, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001960, Sample Num: 31360, Cur Loss: 0.88473481, Cur Avg Loss: 0.36998079, Log Avg loss: 0.41001668, Global Avg Loss: 1.75110142, Time: 0.0208 Steps: 23250, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001970, Sample Num: 31520, Cur Loss: 0.34269446, Cur Avg Loss: 0.36948581, Log Avg loss: 0.27247030, Global Avg Loss: 1.75046572, Time: 0.0208 Steps: 23260, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001980, Sample Num: 31680, Cur Loss: 0.36966282, Cur Avg Loss: 0.36964967, Log Avg loss: 0.40193078, Global Avg Loss: 1.74988620, Time: 0.0208 Steps: 23270, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001990, Sample Num: 31840, Cur Loss: 0.23634219, Cur Avg Loss: 0.36875993, Log Avg loss: 0.19259184, Global Avg Loss: 1.74921726, Time: 0.0208 Steps: 23280, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002000, Sample Num: 32000, Cur Loss: 0.18203323, Cur Avg Loss: 0.36815800, Log Avg loss: 0.24837328, Global Avg Loss: 1.74857285, Time: 0.0208 Steps: 23290, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002010, Sample Num: 32160, Cur Loss: 0.35097805, Cur Avg Loss: 0.36807570, Log Avg loss: 0.35161533, Global Avg Loss: 1.74797329, Time: 0.0208 Steps: 23300, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002020, Sample Num: 32320, Cur Loss: 0.29653436, Cur Avg Loss: 0.36817705, Log Avg loss: 0.38854907, Global Avg Loss: 1.74739010, Time: 0.0208 Steps: 23310, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002030, Sample Num: 32480, Cur Loss: 0.74799782, Cur Avg Loss: 0.36832997, Log Avg loss: 0.39921843, Global Avg Loss: 1.74681198, Time: 0.0208 Steps: 23320, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002040, Sample Num: 32640, Cur Loss: 0.35732943, Cur Avg Loss: 0.36818493, Log Avg loss: 0.33874230, Global Avg Loss: 1.74620844, Time: 0.0208 Steps: 23330, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002050, Sample Num: 32800, Cur Loss: 0.11453301, Cur Avg Loss: 0.36810706, Log Avg loss: 0.35222226, Global Avg Loss: 1.74561118, Time: 0.0248 Steps: 23340, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002060, Sample Num: 32960, Cur Loss: 0.11159973, Cur Avg Loss: 0.36758723, Log Avg loss: 0.26102096, Global Avg Loss: 1.74497539, Time: 0.0210 Steps: 23350, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002070, Sample Num: 33120, Cur Loss: 0.12503593, Cur Avg Loss: 0.36765381, Log Avg loss: 0.38137100, Global Avg Loss: 1.74439165, Time: 0.0209 Steps: 23360, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002080, Sample Num: 33280, Cur Loss: 0.42903355, Cur Avg Loss: 0.36747191, Log Avg loss: 0.32981776, Global Avg Loss: 1.74378636, Time: 0.0209 Steps: 23370, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002090, Sample Num: 33440, Cur Loss: 0.52033091, Cur Avg Loss: 0.36695966, Log Avg loss: 0.26041066, Global Avg Loss: 1.74315189, Time: 0.0209 Steps: 23380, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002100, Sample Num: 33600, Cur Loss: 0.46196431, Cur Avg Loss: 0.36708998, Log Avg loss: 0.39432885, Global Avg Loss: 1.74257523, Time: 0.0210 Steps: 23390, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002110, Sample Num: 33760, Cur Loss: 0.20387848, Cur Avg Loss: 0.36640676, Log Avg loss: 0.22293018, Global Avg Loss: 1.74192581, Time: 0.0209 Steps: 23400, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002120, Sample Num: 33920, Cur Loss: 0.22577435, Cur Avg Loss: 0.36622660, Log Avg loss: 0.32821260, Global Avg Loss: 1.74132191, Time: 0.0209 Steps: 23410, Updated lr: 0.000079 ***** Running evaluation checkpoint-23419 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-23419 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.750522, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.334408, "eval_total_loss": 235.088733, "eval_mae": 0.391407, "eval_mse": 0.334511, "eval_r2": 0.787363, "eval_sp_statistic": 0.883663, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.904141, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.349906, "test_total_loss": 175.652621, "test_mae": 0.360991, "test_mse": 0.350031, "test_r2": 0.774087, "test_sp_statistic": 0.876552, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.930471, "test_ps_pvalue": 0.0, "lr": 7.873968705547652e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7407850817137311, "train_cur_epoch_loss": 779.5002616532147, "train_cur_epoch_avg_loss": 0.3661344582683019, "train_cur_epoch_time": 44.75052237510681, "train_cur_epoch_avg_time": 0.021019503229265765, "epoch": 11, "step": 23419} ################################################## Training, Epoch: 0012, Batch: 000001, Sample Num: 16, Cur Loss: 0.10943249, Cur Avg Loss: 0.10943249, Log Avg loss: 0.32092980, Global Avg Loss: 1.74071543, Time: 0.0248 Steps: 23420, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000011, Sample Num: 176, Cur Loss: 0.11327918, Cur Avg Loss: 0.23543464, Log Avg loss: 0.24803486, Global Avg Loss: 1.74007834, Time: 0.0209 Steps: 23430, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000021, Sample Num: 336, Cur Loss: 0.19900140, Cur Avg Loss: 0.28432589, Log Avg loss: 0.33810626, Global Avg Loss: 1.73948023, Time: 0.0209 Steps: 23440, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000031, Sample Num: 496, Cur Loss: 0.71428752, Cur Avg Loss: 0.27687606, Log Avg loss: 0.26123143, Global Avg Loss: 1.73884985, Time: 0.0209 Steps: 23450, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000041, Sample Num: 656, Cur Loss: 0.77723765, Cur Avg Loss: 0.30183814, Log Avg loss: 0.37922057, Global Avg Loss: 1.73827030, Time: 0.0210 Steps: 23460, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000051, Sample Num: 816, Cur Loss: 0.15009463, Cur Avg Loss: 0.28960893, Log Avg loss: 0.23946918, Global Avg Loss: 1.73763170, Time: 0.0210 Steps: 23470, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000061, Sample Num: 976, Cur Loss: 0.23676357, Cur Avg Loss: 0.27634459, Log Avg loss: 0.20869644, Global Avg Loss: 1.73698053, Time: 0.0209 Steps: 23480, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000071, Sample Num: 1136, Cur Loss: 0.15430222, Cur Avg Loss: 0.28320133, Log Avg loss: 0.32502747, Global Avg Loss: 1.73637944, Time: 0.0209 Steps: 23490, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000081, Sample Num: 1296, Cur Loss: 0.31260797, Cur Avg Loss: 0.28479588, Log Avg loss: 0.29611713, Global Avg Loss: 1.73576657, Time: 0.0209 Steps: 23500, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000091, Sample Num: 1456, Cur Loss: 0.19916520, Cur Avg Loss: 0.29611473, Log Avg loss: 0.38779745, Global Avg Loss: 1.73519321, Time: 0.0209 Steps: 23510, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000101, Sample Num: 1616, Cur Loss: 0.15407208, Cur Avg Loss: 0.30655417, Log Avg loss: 0.40155306, Global Avg Loss: 1.73462618, Time: 0.0210 Steps: 23520, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000111, Sample Num: 1776, Cur Loss: 0.75455946, Cur Avg Loss: 0.33578392, Log Avg loss: 0.63100437, Global Avg Loss: 1.73415715, Time: 0.0209 Steps: 23530, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000121, Sample Num: 1936, Cur Loss: 0.53402275, Cur Avg Loss: 0.33727897, Log Avg loss: 0.35387407, Global Avg Loss: 1.73357080, Time: 0.0210 Steps: 23540, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000131, Sample Num: 2096, Cur Loss: 0.14533946, Cur Avg Loss: 0.33118140, Log Avg loss: 0.25740084, Global Avg Loss: 1.73294397, Time: 0.0209 Steps: 23550, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000141, Sample Num: 2256, Cur Loss: 0.13269329, Cur Avg Loss: 0.33431181, Log Avg loss: 0.37532009, Global Avg Loss: 1.73236773, Time: 0.0209 Steps: 23560, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000151, Sample Num: 2416, Cur Loss: 0.20651159, Cur Avg Loss: 0.33210349, Log Avg loss: 0.30096616, Global Avg Loss: 1.73176044, Time: 0.0210 Steps: 23570, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000161, Sample Num: 2576, Cur Loss: 0.16342026, Cur Avg Loss: 0.33210321, Log Avg loss: 0.33209911, Global Avg Loss: 1.73116686, Time: 0.0209 Steps: 23580, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000171, Sample Num: 2736, Cur Loss: 0.05080511, Cur Avg Loss: 0.32936052, Log Avg loss: 0.28520308, Global Avg Loss: 1.73055390, Time: 0.0209 Steps: 23590, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000181, Sample Num: 2896, Cur Loss: 0.23145084, Cur Avg Loss: 0.32639213, Log Avg loss: 0.27563273, Global Avg Loss: 1.72993741, Time: 0.0209 Steps: 23600, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000191, Sample Num: 3056, Cur Loss: 0.14182121, Cur Avg Loss: 0.31985885, Log Avg loss: 0.20160642, Global Avg Loss: 1.72929008, Time: 0.0209 Steps: 23610, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000201, Sample Num: 3216, Cur Loss: 0.28644317, Cur Avg Loss: 0.31994701, Log Avg loss: 0.32163097, Global Avg Loss: 1.72869412, Time: 0.0210 Steps: 23620, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000211, Sample Num: 3376, Cur Loss: 0.21455647, Cur Avg Loss: 0.32008371, Log Avg loss: 0.32283142, Global Avg Loss: 1.72809917, Time: 0.0209 Steps: 23630, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000221, Sample Num: 3536, Cur Loss: 0.52807736, Cur Avg Loss: 0.32318185, Log Avg loss: 0.38855260, Global Avg Loss: 1.72753253, Time: 0.0211 Steps: 23640, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000231, Sample Num: 3696, Cur Loss: 0.51967138, Cur Avg Loss: 0.32363835, Log Avg loss: 0.33372703, Global Avg Loss: 1.72694318, Time: 0.0210 Steps: 23650, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000241, Sample Num: 3856, Cur Loss: 0.15827298, Cur Avg Loss: 0.32403859, Log Avg loss: 0.33328396, Global Avg Loss: 1.72635415, Time: 0.0209 Steps: 23660, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000251, Sample Num: 4016, Cur Loss: 0.19496110, Cur Avg Loss: 0.32298774, Log Avg loss: 0.29766230, Global Avg Loss: 1.72575056, Time: 0.0210 Steps: 23670, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000261, Sample Num: 4176, Cur Loss: 0.14547405, Cur Avg Loss: 0.32034506, Log Avg loss: 0.25401376, Global Avg Loss: 1.72512905, Time: 0.0209 Steps: 23680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000271, Sample Num: 4336, Cur Loss: 0.09165266, Cur Avg Loss: 0.31628696, Log Avg loss: 0.21037065, Global Avg Loss: 1.72448964, Time: 0.0209 Steps: 23690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000281, Sample Num: 4496, Cur Loss: 0.28576073, Cur Avg Loss: 0.31326978, Log Avg loss: 0.23150413, Global Avg Loss: 1.72385969, Time: 0.0209 Steps: 23700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000291, Sample Num: 4656, Cur Loss: 0.12181869, Cur Avg Loss: 0.30974616, Log Avg loss: 0.21073248, Global Avg Loss: 1.72322151, Time: 0.0208 Steps: 23710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000301, Sample Num: 4816, Cur Loss: 0.23697647, Cur Avg Loss: 0.30766924, Log Avg loss: 0.24723072, Global Avg Loss: 1.72259925, Time: 0.0209 Steps: 23720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000311, Sample Num: 4976, Cur Loss: 0.22125654, Cur Avg Loss: 0.30619518, Log Avg loss: 0.26182605, Global Avg Loss: 1.72198367, Time: 0.0209 Steps: 23730, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000321, Sample Num: 5136, Cur Loss: 0.13169473, Cur Avg Loss: 0.30518124, Log Avg loss: 0.27364768, Global Avg Loss: 1.72137359, Time: 0.0208 Steps: 23740, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000331, Sample Num: 5296, Cur Loss: 0.50961661, Cur Avg Loss: 0.30724563, Log Avg loss: 0.37351259, Global Avg Loss: 1.72080607, Time: 0.0208 Steps: 23750, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000341, Sample Num: 5456, Cur Loss: 0.53577864, Cur Avg Loss: 0.30593478, Log Avg loss: 0.26254552, Global Avg Loss: 1.72019232, Time: 0.0208 Steps: 23760, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000351, Sample Num: 5616, Cur Loss: 0.50850677, Cur Avg Loss: 0.31027483, Log Avg loss: 0.45827075, Global Avg Loss: 1.71966143, Time: 0.0208 Steps: 23770, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000361, Sample Num: 5776, Cur Loss: 0.55807883, Cur Avg Loss: 0.31113732, Log Avg loss: 0.34141061, Global Avg Loss: 1.71908185, Time: 0.0209 Steps: 23780, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000371, Sample Num: 5936, Cur Loss: 0.65562379, Cur Avg Loss: 0.31217351, Log Avg loss: 0.34957995, Global Avg Loss: 1.71850619, Time: 0.0209 Steps: 23790, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000381, Sample Num: 6096, Cur Loss: 1.07545424, Cur Avg Loss: 0.32228966, Log Avg loss: 0.69759907, Global Avg Loss: 1.71807723, Time: 0.0208 Steps: 23800, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000391, Sample Num: 6256, Cur Loss: 0.54203534, Cur Avg Loss: 0.33865168, Log Avg loss: 0.96204433, Global Avg Loss: 1.71775971, Time: 0.0209 Steps: 23810, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000401, Sample Num: 6416, Cur Loss: 0.34221768, Cur Avg Loss: 0.34310945, Log Avg loss: 0.51740856, Global Avg Loss: 1.71725578, Time: 0.0209 Steps: 23820, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000411, Sample Num: 6576, Cur Loss: 0.22716086, Cur Avg Loss: 0.34112035, Log Avg loss: 0.26135713, Global Avg Loss: 1.71664483, Time: 0.0208 Steps: 23830, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000421, Sample Num: 6736, Cur Loss: 0.44304603, Cur Avg Loss: 0.33962692, Log Avg loss: 0.27824704, Global Avg Loss: 1.71604147, Time: 0.0209 Steps: 23840, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000431, Sample Num: 6896, Cur Loss: 0.22974171, Cur Avg Loss: 0.34044908, Log Avg loss: 0.37506191, Global Avg Loss: 1.71547922, Time: 0.0208 Steps: 23850, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000441, Sample Num: 7056, Cur Loss: 0.59541035, Cur Avg Loss: 0.34185709, Log Avg loss: 0.40254265, Global Avg Loss: 1.71492895, Time: 0.0208 Steps: 23860, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000451, Sample Num: 7216, Cur Loss: 0.31476364, Cur Avg Loss: 0.34379254, Log Avg loss: 0.42914580, Global Avg Loss: 1.71439029, Time: 0.0209 Steps: 23870, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000461, Sample Num: 7376, Cur Loss: 0.10402659, Cur Avg Loss: 0.34390504, Log Avg loss: 0.34897871, Global Avg Loss: 1.71381851, Time: 0.0208 Steps: 23880, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000471, Sample Num: 7536, Cur Loss: 0.12741968, Cur Avg Loss: 0.34207713, Log Avg loss: 0.25781041, Global Avg Loss: 1.71320905, Time: 0.0208 Steps: 23890, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000481, Sample Num: 7696, Cur Loss: 0.33012688, Cur Avg Loss: 0.34099863, Log Avg loss: 0.29020130, Global Avg Loss: 1.71261365, Time: 0.0209 Steps: 23900, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000491, Sample Num: 7856, Cur Loss: 0.16824406, Cur Avg Loss: 0.33852899, Log Avg loss: 0.21973916, Global Avg Loss: 1.71198927, Time: 0.0209 Steps: 23910, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000501, Sample Num: 8016, Cur Loss: 0.22457543, Cur Avg Loss: 0.33529630, Log Avg loss: 0.17657152, Global Avg Loss: 1.71134738, Time: 0.0209 Steps: 23920, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000511, Sample Num: 8176, Cur Loss: 0.41744912, Cur Avg Loss: 0.33541208, Log Avg loss: 0.34121234, Global Avg Loss: 1.71077482, Time: 0.0209 Steps: 23930, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000521, Sample Num: 8336, Cur Loss: 0.24553673, Cur Avg Loss: 0.33618549, Log Avg loss: 0.37570723, Global Avg Loss: 1.71021715, Time: 0.0209 Steps: 23940, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000531, Sample Num: 8496, Cur Loss: 0.42261511, Cur Avg Loss: 0.33559320, Log Avg loss: 0.30473445, Global Avg Loss: 1.70963031, Time: 0.0209 Steps: 23950, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000541, Sample Num: 8656, Cur Loss: 0.24917872, Cur Avg Loss: 0.33485124, Log Avg loss: 0.29545315, Global Avg Loss: 1.70904008, Time: 0.0209 Steps: 23960, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000551, Sample Num: 8816, Cur Loss: 0.63132125, Cur Avg Loss: 0.33526392, Log Avg loss: 0.35759033, Global Avg Loss: 1.70847627, Time: 0.0208 Steps: 23970, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000561, Sample Num: 8976, Cur Loss: 0.16359413, Cur Avg Loss: 0.33299810, Log Avg loss: 0.20815128, Global Avg Loss: 1.70785062, Time: 0.0209 Steps: 23980, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000571, Sample Num: 9136, Cur Loss: 0.09978707, Cur Avg Loss: 0.33277119, Log Avg loss: 0.32004158, Global Avg Loss: 1.70727212, Time: 0.0209 Steps: 23990, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000581, Sample Num: 9296, Cur Loss: 0.60836220, Cur Avg Loss: 0.33488542, Log Avg loss: 0.45560806, Global Avg Loss: 1.70675059, Time: 0.0209 Steps: 24000, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000591, Sample Num: 9456, Cur Loss: 0.05367831, Cur Avg Loss: 0.33447582, Log Avg loss: 0.31067813, Global Avg Loss: 1.70616914, Time: 0.0209 Steps: 24010, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000601, Sample Num: 9616, Cur Loss: 0.41458926, Cur Avg Loss: 0.33336392, Log Avg loss: 0.26765044, Global Avg Loss: 1.70557026, Time: 0.0209 Steps: 24020, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000611, Sample Num: 9776, Cur Loss: 0.12311962, Cur Avg Loss: 0.33110735, Log Avg loss: 0.19548768, Global Avg Loss: 1.70494184, Time: 0.0209 Steps: 24030, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000621, Sample Num: 9936, Cur Loss: 0.10749888, Cur Avg Loss: 0.32931767, Log Avg loss: 0.21996767, Global Avg Loss: 1.70432413, Time: 0.0208 Steps: 24040, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000631, Sample Num: 10096, Cur Loss: 0.14328369, Cur Avg Loss: 0.32798010, Log Avg loss: 0.24491730, Global Avg Loss: 1.70371731, Time: 0.0209 Steps: 24050, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000641, Sample Num: 10256, Cur Loss: 0.43152934, Cur Avg Loss: 0.32719711, Log Avg loss: 0.27779048, Global Avg Loss: 1.70312465, Time: 0.0209 Steps: 24060, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000651, Sample Num: 10416, Cur Loss: 0.28485227, Cur Avg Loss: 0.32594635, Log Avg loss: 0.24577266, Global Avg Loss: 1.70251919, Time: 0.0209 Steps: 24070, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000661, Sample Num: 10576, Cur Loss: 0.08043543, Cur Avg Loss: 0.32504507, Log Avg loss: 0.26637195, Global Avg Loss: 1.70192278, Time: 0.0208 Steps: 24080, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000671, Sample Num: 10736, Cur Loss: 0.12687944, Cur Avg Loss: 0.32317069, Log Avg loss: 0.19927359, Global Avg Loss: 1.70129902, Time: 0.0209 Steps: 24090, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000681, Sample Num: 10896, Cur Loss: 0.47864434, Cur Avg Loss: 0.32539615, Log Avg loss: 0.47472506, Global Avg Loss: 1.70079007, Time: 0.0209 Steps: 24100, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000691, Sample Num: 11056, Cur Loss: 0.52283251, Cur Avg Loss: 0.32779864, Log Avg loss: 0.49140823, Global Avg Loss: 1.70028846, Time: 0.0209 Steps: 24110, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000701, Sample Num: 11216, Cur Loss: 0.18998924, Cur Avg Loss: 0.32767329, Log Avg loss: 0.31901113, Global Avg Loss: 1.69971579, Time: 0.0208 Steps: 24120, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000711, Sample Num: 11376, Cur Loss: 0.64605331, Cur Avg Loss: 0.32881329, Log Avg loss: 0.40872750, Global Avg Loss: 1.69918077, Time: 0.0209 Steps: 24130, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000721, Sample Num: 11536, Cur Loss: 0.16832882, Cur Avg Loss: 0.32833099, Log Avg loss: 0.29403948, Global Avg Loss: 1.69859869, Time: 0.0209 Steps: 24140, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000731, Sample Num: 11696, Cur Loss: 0.63167214, Cur Avg Loss: 0.32884729, Log Avg loss: 0.36607264, Global Avg Loss: 1.69804692, Time: 0.0209 Steps: 24150, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000741, Sample Num: 11856, Cur Loss: 0.16795018, Cur Avg Loss: 0.32884605, Log Avg loss: 0.32875557, Global Avg Loss: 1.69748016, Time: 0.0209 Steps: 24160, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000751, Sample Num: 12016, Cur Loss: 0.25388452, Cur Avg Loss: 0.33035605, Log Avg loss: 0.44224675, Global Avg Loss: 1.69696083, Time: 0.0209 Steps: 24170, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000761, Sample Num: 12176, Cur Loss: 0.88635230, Cur Avg Loss: 0.33029851, Log Avg loss: 0.32597750, Global Avg Loss: 1.69639384, Time: 0.0208 Steps: 24180, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000771, Sample Num: 12336, Cur Loss: 0.60166335, Cur Avg Loss: 0.33455832, Log Avg loss: 0.65872936, Global Avg Loss: 1.69596487, Time: 0.0246 Steps: 24190, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000781, Sample Num: 12496, Cur Loss: 0.46617788, Cur Avg Loss: 0.33594410, Log Avg loss: 0.44278767, Global Avg Loss: 1.69544703, Time: 0.0209 Steps: 24200, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000791, Sample Num: 12656, Cur Loss: 0.37269622, Cur Avg Loss: 0.33747435, Log Avg loss: 0.45698754, Global Avg Loss: 1.69493548, Time: 0.0209 Steps: 24210, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000801, Sample Num: 12816, Cur Loss: 0.12443545, Cur Avg Loss: 0.33839836, Log Avg loss: 0.41148706, Global Avg Loss: 1.69440557, Time: 0.0208 Steps: 24220, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000811, Sample Num: 12976, Cur Loss: 0.21588673, Cur Avg Loss: 0.33926730, Log Avg loss: 0.40886916, Global Avg Loss: 1.69387501, Time: 0.0209 Steps: 24230, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000821, Sample Num: 13136, Cur Loss: 0.29150400, Cur Avg Loss: 0.33919210, Log Avg loss: 0.33309353, Global Avg Loss: 1.69331364, Time: 0.0209 Steps: 24240, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000831, Sample Num: 13296, Cur Loss: 0.55587721, Cur Avg Loss: 0.34052739, Log Avg loss: 0.45015469, Global Avg Loss: 1.69280099, Time: 0.0208 Steps: 24250, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000841, Sample Num: 13456, Cur Loss: 0.91210634, Cur Avg Loss: 0.34237331, Log Avg loss: 0.49576956, Global Avg Loss: 1.69230758, Time: 0.0209 Steps: 24260, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000851, Sample Num: 13616, Cur Loss: 0.34231666, Cur Avg Loss: 0.34107951, Log Avg loss: 0.23227105, Global Avg Loss: 1.69170599, Time: 0.0209 Steps: 24270, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000861, Sample Num: 13776, Cur Loss: 0.15060064, Cur Avg Loss: 0.34016760, Log Avg loss: 0.26256406, Global Avg Loss: 1.69111739, Time: 0.0209 Steps: 24280, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000871, Sample Num: 13936, Cur Loss: 0.30438849, Cur Avg Loss: 0.33920962, Log Avg loss: 0.25672728, Global Avg Loss: 1.69052686, Time: 0.0209 Steps: 24290, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000881, Sample Num: 14096, Cur Loss: 0.14665398, Cur Avg Loss: 0.33843255, Log Avg loss: 0.27074996, Global Avg Loss: 1.68994259, Time: 0.0210 Steps: 24300, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000891, Sample Num: 14256, Cur Loss: 0.27942839, Cur Avg Loss: 0.33828589, Log Avg loss: 0.32536494, Global Avg Loss: 1.68938127, Time: 0.0208 Steps: 24310, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000901, Sample Num: 14416, Cur Loss: 0.41612217, Cur Avg Loss: 0.33830354, Log Avg loss: 0.33987630, Global Avg Loss: 1.68882637, Time: 0.0208 Steps: 24320, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000911, Sample Num: 14576, Cur Loss: 0.13827771, Cur Avg Loss: 0.33791279, Log Avg loss: 0.30270576, Global Avg Loss: 1.68825665, Time: 0.0208 Steps: 24330, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000921, Sample Num: 14736, Cur Loss: 0.39480644, Cur Avg Loss: 0.33785471, Log Avg loss: 0.33256387, Global Avg Loss: 1.68769967, Time: 0.0208 Steps: 24340, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000931, Sample Num: 14896, Cur Loss: 0.28648850, Cur Avg Loss: 0.33735495, Log Avg loss: 0.29132679, Global Avg Loss: 1.68712621, Time: 0.0208 Steps: 24350, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000941, Sample Num: 15056, Cur Loss: 0.30211228, Cur Avg Loss: 0.33677471, Log Avg loss: 0.28275478, Global Avg Loss: 1.68654971, Time: 0.0209 Steps: 24360, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000951, Sample Num: 15216, Cur Loss: 0.09684699, Cur Avg Loss: 0.33589218, Log Avg loss: 0.25284600, Global Avg Loss: 1.68596140, Time: 0.0209 Steps: 24370, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000961, Sample Num: 15376, Cur Loss: 0.19061285, Cur Avg Loss: 0.33558283, Log Avg loss: 0.30616328, Global Avg Loss: 1.68539544, Time: 0.0208 Steps: 24380, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000971, Sample Num: 15536, Cur Loss: 0.33275038, Cur Avg Loss: 0.33573962, Log Avg loss: 0.35080796, Global Avg Loss: 1.68484826, Time: 0.0208 Steps: 24390, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000981, Sample Num: 15696, Cur Loss: 0.32145658, Cur Avg Loss: 0.33631184, Log Avg loss: 0.39187368, Global Avg Loss: 1.68431835, Time: 0.0208 Steps: 24400, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000991, Sample Num: 15856, Cur Loss: 0.16148829, Cur Avg Loss: 0.33578008, Log Avg loss: 0.28361493, Global Avg Loss: 1.68374453, Time: 0.0208 Steps: 24410, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001001, Sample Num: 16016, Cur Loss: 0.29332638, Cur Avg Loss: 0.33580804, Log Avg loss: 0.33857892, Global Avg Loss: 1.68319368, Time: 0.0208 Steps: 24420, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001011, Sample Num: 16176, Cur Loss: 0.09340176, Cur Avg Loss: 0.33430333, Log Avg loss: 0.18368150, Global Avg Loss: 1.68257988, Time: 0.0208 Steps: 24430, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001021, Sample Num: 16336, Cur Loss: 0.16196598, Cur Avg Loss: 0.33336213, Log Avg loss: 0.23820721, Global Avg Loss: 1.68198889, Time: 0.0208 Steps: 24440, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001031, Sample Num: 16496, Cur Loss: 0.41843829, Cur Avg Loss: 0.33349085, Log Avg loss: 0.34663248, Global Avg Loss: 1.68144274, Time: 0.0210 Steps: 24450, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001041, Sample Num: 16656, Cur Loss: 0.15539078, Cur Avg Loss: 0.33270316, Log Avg loss: 0.25149300, Global Avg Loss: 1.68085813, Time: 0.0209 Steps: 24460, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001051, Sample Num: 16816, Cur Loss: 0.30721754, Cur Avg Loss: 0.33164633, Log Avg loss: 0.22162986, Global Avg Loss: 1.68026179, Time: 0.0209 Steps: 24470, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001061, Sample Num: 16976, Cur Loss: 0.35400975, Cur Avg Loss: 0.33199391, Log Avg loss: 0.36852477, Global Avg Loss: 1.67972595, Time: 0.0209 Steps: 24480, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001071, Sample Num: 17136, Cur Loss: 0.14473130, Cur Avg Loss: 0.33119689, Log Avg loss: 0.24663338, Global Avg Loss: 1.67914078, Time: 0.0209 Steps: 24490, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001081, Sample Num: 17296, Cur Loss: 0.14681637, Cur Avg Loss: 0.33020904, Log Avg loss: 0.22440949, Global Avg Loss: 1.67854701, Time: 0.0209 Steps: 24500, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001091, Sample Num: 17456, Cur Loss: 0.20877609, Cur Avg Loss: 0.32971334, Log Avg loss: 0.27612885, Global Avg Loss: 1.67797483, Time: 0.0208 Steps: 24510, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001101, Sample Num: 17616, Cur Loss: 0.14599660, Cur Avg Loss: 0.33045118, Log Avg loss: 0.41094948, Global Avg Loss: 1.67745810, Time: 0.0209 Steps: 24520, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001111, Sample Num: 17776, Cur Loss: 0.43707085, Cur Avg Loss: 0.33074934, Log Avg loss: 0.36357667, Global Avg Loss: 1.67692248, Time: 0.0209 Steps: 24530, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001121, Sample Num: 17936, Cur Loss: 0.30830997, Cur Avg Loss: 0.33296207, Log Avg loss: 0.57879659, Global Avg Loss: 1.67647499, Time: 0.0209 Steps: 24540, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001131, Sample Num: 18096, Cur Loss: 0.53884959, Cur Avg Loss: 0.33280800, Log Avg loss: 0.31553640, Global Avg Loss: 1.67592064, Time: 0.0209 Steps: 24550, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001141, Sample Num: 18256, Cur Loss: 0.84331036, Cur Avg Loss: 0.33348275, Log Avg loss: 0.40979690, Global Avg Loss: 1.67540512, Time: 0.0209 Steps: 24560, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001151, Sample Num: 18416, Cur Loss: 0.47918215, Cur Avg Loss: 0.33480687, Log Avg loss: 0.48588867, Global Avg Loss: 1.67492098, Time: 0.0209 Steps: 24570, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001161, Sample Num: 18576, Cur Loss: 0.17009674, Cur Avg Loss: 0.33645806, Log Avg loss: 0.52651009, Global Avg Loss: 1.67445377, Time: 0.0209 Steps: 24580, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001171, Sample Num: 18736, Cur Loss: 0.28530586, Cur Avg Loss: 0.33674203, Log Avg loss: 0.36971100, Global Avg Loss: 1.67392317, Time: 0.0209 Steps: 24590, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001181, Sample Num: 18896, Cur Loss: 0.08791083, Cur Avg Loss: 0.33553525, Log Avg loss: 0.19422217, Global Avg Loss: 1.67332167, Time: 0.0209 Steps: 24600, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001191, Sample Num: 19056, Cur Loss: 0.47065023, Cur Avg Loss: 0.33688252, Log Avg loss: 0.49599504, Global Avg Loss: 1.67284327, Time: 0.0209 Steps: 24610, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001201, Sample Num: 19216, Cur Loss: 0.39302945, Cur Avg Loss: 0.33661728, Log Avg loss: 0.30502732, Global Avg Loss: 1.67228770, Time: 0.0209 Steps: 24620, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001211, Sample Num: 19376, Cur Loss: 0.23754123, Cur Avg Loss: 0.33622718, Log Avg loss: 0.28937527, Global Avg Loss: 1.67172623, Time: 0.0209 Steps: 24630, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001221, Sample Num: 19536, Cur Loss: 1.13733518, Cur Avg Loss: 0.34059701, Log Avg loss: 0.86978366, Global Avg Loss: 1.67140076, Time: 0.0209 Steps: 24640, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001231, Sample Num: 19696, Cur Loss: 0.47643614, Cur Avg Loss: 0.34304294, Log Avg loss: 0.64169159, Global Avg Loss: 1.67098303, Time: 0.0209 Steps: 24650, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001241, Sample Num: 19856, Cur Loss: 0.18843570, Cur Avg Loss: 0.34238049, Log Avg loss: 0.26083237, Global Avg Loss: 1.67041119, Time: 0.0208 Steps: 24660, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001251, Sample Num: 20016, Cur Loss: 0.22323389, Cur Avg Loss: 0.34192125, Log Avg loss: 0.28492971, Global Avg Loss: 1.66984959, Time: 0.0209 Steps: 24670, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001261, Sample Num: 20176, Cur Loss: 0.21947007, Cur Avg Loss: 0.34174054, Log Avg loss: 0.31913359, Global Avg Loss: 1.66930230, Time: 0.0209 Steps: 24680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001271, Sample Num: 20336, Cur Loss: 0.19885397, Cur Avg Loss: 0.34098859, Log Avg loss: 0.24616709, Global Avg Loss: 1.66872589, Time: 0.0209 Steps: 24690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001281, Sample Num: 20496, Cur Loss: 0.22905903, Cur Avg Loss: 0.34206769, Log Avg loss: 0.47922227, Global Avg Loss: 1.66824431, Time: 0.0246 Steps: 24700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001291, Sample Num: 20656, Cur Loss: 0.29022613, Cur Avg Loss: 0.34303645, Log Avg loss: 0.46713370, Global Avg Loss: 1.66775823, Time: 0.0209 Steps: 24710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001301, Sample Num: 20816, Cur Loss: 0.19166823, Cur Avg Loss: 0.34244768, Log Avg loss: 0.26643770, Global Avg Loss: 1.66719135, Time: 0.0209 Steps: 24720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001311, Sample Num: 20976, Cur Loss: 0.25371766, Cur Avg Loss: 0.34236532, Log Avg loss: 0.33165030, Global Avg Loss: 1.66665130, Time: 0.0209 Steps: 24730, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001321, Sample Num: 21136, Cur Loss: 0.39140153, Cur Avg Loss: 0.34228772, Log Avg loss: 0.33211514, Global Avg Loss: 1.66611188, Time: 0.0208 Steps: 24740, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001331, Sample Num: 21296, Cur Loss: 0.21727361, Cur Avg Loss: 0.34253221, Log Avg loss: 0.37482841, Global Avg Loss: 1.66559015, Time: 0.0209 Steps: 24750, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001341, Sample Num: 21456, Cur Loss: 0.96079010, Cur Avg Loss: 0.34314237, Log Avg loss: 0.42435551, Global Avg Loss: 1.66508884, Time: 0.0208 Steps: 24760, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001351, Sample Num: 21616, Cur Loss: 0.12227046, Cur Avg Loss: 0.34220066, Log Avg loss: 0.21591683, Global Avg Loss: 1.66450379, Time: 0.0209 Steps: 24770, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001361, Sample Num: 21776, Cur Loss: 0.60349464, Cur Avg Loss: 0.34198574, Log Avg loss: 0.31294958, Global Avg Loss: 1.66395837, Time: 0.0209 Steps: 24780, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001371, Sample Num: 21936, Cur Loss: 0.40156770, Cur Avg Loss: 0.34195273, Log Avg loss: 0.33746021, Global Avg Loss: 1.66342328, Time: 0.0209 Steps: 24790, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001381, Sample Num: 22096, Cur Loss: 0.85075659, Cur Avg Loss: 0.34242560, Log Avg loss: 0.40725708, Global Avg Loss: 1.66291676, Time: 0.0209 Steps: 24800, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001391, Sample Num: 22256, Cur Loss: 0.81541163, Cur Avg Loss: 0.34303360, Log Avg loss: 0.42699804, Global Avg Loss: 1.66241860, Time: 0.0208 Steps: 24810, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001401, Sample Num: 22416, Cur Loss: 0.90897948, Cur Avg Loss: 0.34640283, Log Avg loss: 0.81506261, Global Avg Loss: 1.66207720, Time: 0.0208 Steps: 24820, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001411, Sample Num: 22576, Cur Loss: 0.80342400, Cur Avg Loss: 0.34732772, Log Avg loss: 0.47690524, Global Avg Loss: 1.66159989, Time: 0.0210 Steps: 24830, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001421, Sample Num: 22736, Cur Loss: 0.57664633, Cur Avg Loss: 0.34817527, Log Avg loss: 0.46776368, Global Avg Loss: 1.66111928, Time: 0.0209 Steps: 24840, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001431, Sample Num: 22896, Cur Loss: 0.48112237, Cur Avg Loss: 0.34920705, Log Avg loss: 0.49582393, Global Avg Loss: 1.66065035, Time: 0.0209 Steps: 24850, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001441, Sample Num: 23056, Cur Loss: 0.29299611, Cur Avg Loss: 0.34860533, Log Avg loss: 0.26249792, Global Avg Loss: 1.66008794, Time: 0.0209 Steps: 24860, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001451, Sample Num: 23216, Cur Loss: 0.29686433, Cur Avg Loss: 0.34794507, Log Avg loss: 0.25280223, Global Avg Loss: 1.65952208, Time: 0.0210 Steps: 24870, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001461, Sample Num: 23376, Cur Loss: 0.14805803, Cur Avg Loss: 0.34713290, Log Avg loss: 0.22928733, Global Avg Loss: 1.65894723, Time: 0.0209 Steps: 24880, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001471, Sample Num: 23536, Cur Loss: 0.18839774, Cur Avg Loss: 0.34660101, Log Avg loss: 0.26889113, Global Avg Loss: 1.65838875, Time: 0.0208 Steps: 24890, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001481, Sample Num: 23696, Cur Loss: 0.24444127, Cur Avg Loss: 0.34614627, Log Avg loss: 0.27925414, Global Avg Loss: 1.65783488, Time: 0.0210 Steps: 24900, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001491, Sample Num: 23856, Cur Loss: 0.28714937, Cur Avg Loss: 0.34627714, Log Avg loss: 0.36565930, Global Avg Loss: 1.65731614, Time: 0.0209 Steps: 24910, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001501, Sample Num: 24016, Cur Loss: 0.30171818, Cur Avg Loss: 0.34591272, Log Avg loss: 0.29157804, Global Avg Loss: 1.65676809, Time: 0.0210 Steps: 24920, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001511, Sample Num: 24176, Cur Loss: 0.49576426, Cur Avg Loss: 0.34573441, Log Avg loss: 0.31896951, Global Avg Loss: 1.65623147, Time: 0.0209 Steps: 24930, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001521, Sample Num: 24336, Cur Loss: 0.25792885, Cur Avg Loss: 0.34528306, Log Avg loss: 0.27708381, Global Avg Loss: 1.65567848, Time: 0.0208 Steps: 24940, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001531, Sample Num: 24496, Cur Loss: 0.29257905, Cur Avg Loss: 0.34469807, Log Avg loss: 0.25572195, Global Avg Loss: 1.65511738, Time: 0.0209 Steps: 24950, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001541, Sample Num: 24656, Cur Loss: 0.45984405, Cur Avg Loss: 0.34408978, Log Avg loss: 0.25095977, Global Avg Loss: 1.65455481, Time: 0.0225 Steps: 24960, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001551, Sample Num: 24816, Cur Loss: 0.20984255, Cur Avg Loss: 0.34409501, Log Avg loss: 0.34490216, Global Avg Loss: 1.65403032, Time: 0.0208 Steps: 24970, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001561, Sample Num: 24976, Cur Loss: 0.16361226, Cur Avg Loss: 0.34526868, Log Avg loss: 0.52730405, Global Avg Loss: 1.65357927, Time: 0.0211 Steps: 24980, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001571, Sample Num: 25136, Cur Loss: 0.65061867, Cur Avg Loss: 0.34712724, Log Avg loss: 0.63724852, Global Avg Loss: 1.65317258, Time: 0.0210 Steps: 24990, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001581, Sample Num: 25296, Cur Loss: 0.08999044, Cur Avg Loss: 0.34819051, Log Avg loss: 0.51522967, Global Avg Loss: 1.65271740, Time: 0.0210 Steps: 25000, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001591, Sample Num: 25456, Cur Loss: 0.40950090, Cur Avg Loss: 0.34848109, Log Avg loss: 0.39442207, Global Avg Loss: 1.65221428, Time: 0.0211 Steps: 25010, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001601, Sample Num: 25616, Cur Loss: 0.24572849, Cur Avg Loss: 0.34841551, Log Avg loss: 0.33798256, Global Avg Loss: 1.65168901, Time: 0.0210 Steps: 25020, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001611, Sample Num: 25776, Cur Loss: 0.12055337, Cur Avg Loss: 0.34780981, Log Avg loss: 0.25083639, Global Avg Loss: 1.65112934, Time: 0.0210 Steps: 25030, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001621, Sample Num: 25936, Cur Loss: 0.43295339, Cur Avg Loss: 0.34747886, Log Avg loss: 0.29416403, Global Avg Loss: 1.65058742, Time: 0.0210 Steps: 25040, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001631, Sample Num: 26096, Cur Loss: 0.46213174, Cur Avg Loss: 0.34734199, Log Avg loss: 0.32515497, Global Avg Loss: 1.65005831, Time: 0.0210 Steps: 25050, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001641, Sample Num: 26256, Cur Loss: 0.12892072, Cur Avg Loss: 0.34681590, Log Avg loss: 0.26101010, Global Avg Loss: 1.64950402, Time: 0.0210 Steps: 25060, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001651, Sample Num: 26416, Cur Loss: 0.45562044, Cur Avg Loss: 0.34736824, Log Avg loss: 0.43800801, Global Avg Loss: 1.64902077, Time: 0.0210 Steps: 25070, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001661, Sample Num: 26576, Cur Loss: 0.23014027, Cur Avg Loss: 0.34727924, Log Avg loss: 0.33258468, Global Avg Loss: 1.64849588, Time: 0.0210 Steps: 25080, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001671, Sample Num: 26736, Cur Loss: 0.14590241, Cur Avg Loss: 0.34684380, Log Avg loss: 0.27451653, Global Avg Loss: 1.64794826, Time: 0.0210 Steps: 25090, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001681, Sample Num: 26896, Cur Loss: 0.14709699, Cur Avg Loss: 0.34739964, Log Avg loss: 0.44028167, Global Avg Loss: 1.64746712, Time: 0.0210 Steps: 25100, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001691, Sample Num: 27056, Cur Loss: 0.48719504, Cur Avg Loss: 0.34825466, Log Avg loss: 0.49198378, Global Avg Loss: 1.64700695, Time: 0.0210 Steps: 25110, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001701, Sample Num: 27216, Cur Loss: 0.19270055, Cur Avg Loss: 0.34961918, Log Avg loss: 0.58035859, Global Avg Loss: 1.64658233, Time: 0.0210 Steps: 25120, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001711, Sample Num: 27376, Cur Loss: 0.36877382, Cur Avg Loss: 0.34929315, Log Avg loss: 0.29383631, Global Avg Loss: 1.64604403, Time: 0.0210 Steps: 25130, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001721, Sample Num: 27536, Cur Loss: 0.39130798, Cur Avg Loss: 0.34945390, Log Avg loss: 0.37695706, Global Avg Loss: 1.64553922, Time: 0.0210 Steps: 25140, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001731, Sample Num: 27696, Cur Loss: 0.11438126, Cur Avg Loss: 0.34866568, Log Avg loss: 0.21301296, Global Avg Loss: 1.64496963, Time: 0.0210 Steps: 25150, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001741, Sample Num: 27856, Cur Loss: 0.46623862, Cur Avg Loss: 0.34828849, Log Avg loss: 0.28299670, Global Avg Loss: 1.64442830, Time: 0.0210 Steps: 25160, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001751, Sample Num: 28016, Cur Loss: 0.21422449, Cur Avg Loss: 0.34821816, Log Avg loss: 0.33597413, Global Avg Loss: 1.64390846, Time: 0.0210 Steps: 25170, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001761, Sample Num: 28176, Cur Loss: 0.27529767, Cur Avg Loss: 0.34781141, Log Avg loss: 0.27659052, Global Avg Loss: 1.64336544, Time: 0.0210 Steps: 25180, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001771, Sample Num: 28336, Cur Loss: 0.43716836, Cur Avg Loss: 0.34740695, Log Avg loss: 0.27618025, Global Avg Loss: 1.64282269, Time: 0.0210 Steps: 25190, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001781, Sample Num: 28496, Cur Loss: 0.26130950, Cur Avg Loss: 0.34716356, Log Avg loss: 0.30406028, Global Avg Loss: 1.64229143, Time: 0.0210 Steps: 25200, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001791, Sample Num: 28656, Cur Loss: 0.20526454, Cur Avg Loss: 0.34659277, Log Avg loss: 0.24493447, Global Avg Loss: 1.64173715, Time: 0.0210 Steps: 25210, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001801, Sample Num: 28816, Cur Loss: 0.23697659, Cur Avg Loss: 0.34865067, Log Avg loss: 0.71721992, Global Avg Loss: 1.64137057, Time: 0.0210 Steps: 25220, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001811, Sample Num: 28976, Cur Loss: 0.52748334, Cur Avg Loss: 0.34992781, Log Avg loss: 0.57994150, Global Avg Loss: 1.64094986, Time: 0.0210 Steps: 25230, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001821, Sample Num: 29136, Cur Loss: 0.27701771, Cur Avg Loss: 0.35081178, Log Avg loss: 0.51089900, Global Avg Loss: 1.64050214, Time: 0.0210 Steps: 25240, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001831, Sample Num: 29296, Cur Loss: 0.47521698, Cur Avg Loss: 0.35053232, Log Avg loss: 0.29964190, Global Avg Loss: 1.63997111, Time: 0.0210 Steps: 25250, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001841, Sample Num: 29456, Cur Loss: 0.40929198, Cur Avg Loss: 0.35063164, Log Avg loss: 0.36881763, Global Avg Loss: 1.63946788, Time: 0.0210 Steps: 25260, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001851, Sample Num: 29616, Cur Loss: 0.12473699, Cur Avg Loss: 0.35002962, Log Avg loss: 0.23919749, Global Avg Loss: 1.63891376, Time: 0.0210 Steps: 25270, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001861, Sample Num: 29776, Cur Loss: 0.52045417, Cur Avg Loss: 0.35047079, Log Avg loss: 0.43213087, Global Avg Loss: 1.63843639, Time: 0.0210 Steps: 25280, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001871, Sample Num: 29936, Cur Loss: 0.73120880, Cur Avg Loss: 0.35063075, Log Avg loss: 0.38040006, Global Avg Loss: 1.63793895, Time: 0.0210 Steps: 25290, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001881, Sample Num: 30096, Cur Loss: 0.36529505, Cur Avg Loss: 0.35023591, Log Avg loss: 0.27636141, Global Avg Loss: 1.63740077, Time: 0.0210 Steps: 25300, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001891, Sample Num: 30256, Cur Loss: 0.17754422, Cur Avg Loss: 0.35097064, Log Avg loss: 0.48917371, Global Avg Loss: 1.63694711, Time: 0.0210 Steps: 25310, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001901, Sample Num: 30416, Cur Loss: 0.41460964, Cur Avg Loss: 0.35161144, Log Avg loss: 0.47278541, Global Avg Loss: 1.63648733, Time: 0.0210 Steps: 25320, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001911, Sample Num: 30576, Cur Loss: 1.01519310, Cur Avg Loss: 0.35197993, Log Avg loss: 0.42202976, Global Avg Loss: 1.63600787, Time: 0.0210 Steps: 25330, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001921, Sample Num: 30736, Cur Loss: 0.28462186, Cur Avg Loss: 0.35159357, Log Avg loss: 0.27776188, Global Avg Loss: 1.63547187, Time: 0.0210 Steps: 25340, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001931, Sample Num: 30896, Cur Loss: 0.45720136, Cur Avg Loss: 0.35188361, Log Avg loss: 0.40759899, Global Avg Loss: 1.63498750, Time: 0.0210 Steps: 25350, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001941, Sample Num: 31056, Cur Loss: 0.50204366, Cur Avg Loss: 0.35180126, Log Avg loss: 0.33589939, Global Avg Loss: 1.63447524, Time: 0.0210 Steps: 25360, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001951, Sample Num: 31216, Cur Loss: 0.33344531, Cur Avg Loss: 0.35203850, Log Avg loss: 0.39808728, Global Avg Loss: 1.63398790, Time: 0.0210 Steps: 25370, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001961, Sample Num: 31376, Cur Loss: 0.11405040, Cur Avg Loss: 0.35204219, Log Avg loss: 0.35276197, Global Avg Loss: 1.63348308, Time: 0.0210 Steps: 25380, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001971, Sample Num: 31536, Cur Loss: 0.10445708, Cur Avg Loss: 0.35258342, Log Avg loss: 0.45871795, Global Avg Loss: 1.63302039, Time: 0.0210 Steps: 25390, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001981, Sample Num: 31696, Cur Loss: 0.21657670, Cur Avg Loss: 0.35235319, Log Avg loss: 0.30697596, Global Avg Loss: 1.63249833, Time: 0.0210 Steps: 25400, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001991, Sample Num: 31856, Cur Loss: 0.15976344, Cur Avg Loss: 0.35218702, Log Avg loss: 0.31926812, Global Avg Loss: 1.63198151, Time: 0.0210 Steps: 25410, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002001, Sample Num: 32016, Cur Loss: 0.25380144, Cur Avg Loss: 0.35192039, Log Avg loss: 0.29883459, Global Avg Loss: 1.63145706, Time: 0.0210 Steps: 25420, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002011, Sample Num: 32176, Cur Loss: 0.19756095, Cur Avg Loss: 0.35184654, Log Avg loss: 0.33706923, Global Avg Loss: 1.63094806, Time: 0.0210 Steps: 25430, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002021, Sample Num: 32336, Cur Loss: 0.24670738, Cur Avg Loss: 0.35153207, Log Avg loss: 0.28829229, Global Avg Loss: 1.63042029, Time: 0.0210 Steps: 25440, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002031, Sample Num: 32496, Cur Loss: 0.07143361, Cur Avg Loss: 0.35157410, Log Avg loss: 0.36006905, Global Avg Loss: 1.62992113, Time: 0.0210 Steps: 25450, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002041, Sample Num: 32656, Cur Loss: 0.30036926, Cur Avg Loss: 0.35156205, Log Avg loss: 0.34911309, Global Avg Loss: 1.62941807, Time: 0.0210 Steps: 25460, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002051, Sample Num: 32816, Cur Loss: 0.30018297, Cur Avg Loss: 0.35141945, Log Avg loss: 0.32231596, Global Avg Loss: 1.62890487, Time: 0.0249 Steps: 25470, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002061, Sample Num: 32976, Cur Loss: 0.33295554, Cur Avg Loss: 0.35064333, Log Avg loss: 0.19146008, Global Avg Loss: 1.62834073, Time: 0.0211 Steps: 25480, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002071, Sample Num: 33136, Cur Loss: 0.22836310, Cur Avg Loss: 0.35020035, Log Avg loss: 0.25890250, Global Avg Loss: 1.62780348, Time: 0.0211 Steps: 25490, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002081, Sample Num: 33296, Cur Loss: 0.66860658, Cur Avg Loss: 0.35056229, Log Avg loss: 0.42552011, Global Avg Loss: 1.62733200, Time: 0.0211 Steps: 25500, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002091, Sample Num: 33456, Cur Loss: 0.17544910, Cur Avg Loss: 0.35017033, Log Avg loss: 0.26860484, Global Avg Loss: 1.62679937, Time: 0.0211 Steps: 25510, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002101, Sample Num: 33616, Cur Loss: 0.31100357, Cur Avg Loss: 0.34991600, Log Avg loss: 0.29673535, Global Avg Loss: 1.62627819, Time: 0.0212 Steps: 25520, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002111, Sample Num: 33776, Cur Loss: 0.37920812, Cur Avg Loss: 0.34996586, Log Avg loss: 0.36044011, Global Avg Loss: 1.62578236, Time: 0.0210 Steps: 25530, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002121, Sample Num: 33936, Cur Loss: 0.28079319, Cur Avg Loss: 0.34994350, Log Avg loss: 0.34522303, Global Avg Loss: 1.62528097, Time: 0.0211 Steps: 25540, Updated lr: 0.000077 ***** Running evaluation checkpoint-25548 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-25548 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.727993, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.35012, "eval_total_loss": 246.134273, "eval_mae": 0.393624, "eval_mse": 0.350247, "eval_r2": 0.77736, "eval_sp_statistic": 0.886572, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.902047, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.470634, "test_total_loss": 236.258447, "test_mae": 0.459507, "test_mse": 0.470788, "test_r2": 0.696149, "test_sp_statistic": 0.879973, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.920969, "test_ps_pvalue": 0.0, "lr": 7.672072072072073e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6249103137505296, "train_cur_epoch_loss": 745.7628670446575, "train_cur_epoch_avg_loss": 0.35028786615531116, "train_cur_epoch_time": 44.72799348831177, "train_cur_epoch_avg_time": 0.021008921319075512, "epoch": 12, "step": 25548} ################################################## Training, Epoch: 0013, Batch: 000002, Sample Num: 32, Cur Loss: 0.16656755, Cur Avg Loss: 0.29896671, Log Avg loss: 0.41306459, Global Avg Loss: 1.62480652, Time: 0.0249 Steps: 25550, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000012, Sample Num: 192, Cur Loss: 0.21118619, Cur Avg Loss: 0.58902478, Log Avg loss: 0.64703640, Global Avg Loss: 1.62442398, Time: 0.0211 Steps: 25560, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000022, Sample Num: 352, Cur Loss: 0.78925812, Cur Avg Loss: 0.68171180, Log Avg loss: 0.79293621, Global Avg Loss: 1.62409880, Time: 0.0211 Steps: 25570, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000032, Sample Num: 512, Cur Loss: 0.20854332, Cur Avg Loss: 0.61236900, Log Avg loss: 0.45981485, Global Avg Loss: 1.62364365, Time: 0.0211 Steps: 25580, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000042, Sample Num: 672, Cur Loss: 0.10739613, Cur Avg Loss: 0.55154427, Log Avg loss: 0.35690515, Global Avg Loss: 1.62314863, Time: 0.0211 Steps: 25590, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000052, Sample Num: 832, Cur Loss: 0.09585123, Cur Avg Loss: 0.48292159, Log Avg loss: 0.19470631, Global Avg Loss: 1.62259065, Time: 0.0211 Steps: 25600, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000062, Sample Num: 992, Cur Loss: 0.25956461, Cur Avg Loss: 0.43911001, Log Avg loss: 0.21128978, Global Avg Loss: 1.62203958, Time: 0.0211 Steps: 25610, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000072, Sample Num: 1152, Cur Loss: 0.19085351, Cur Avg Loss: 0.41965331, Log Avg loss: 0.29902179, Global Avg Loss: 1.62152317, Time: 0.0210 Steps: 25620, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000082, Sample Num: 1312, Cur Loss: 0.15394598, Cur Avg Loss: 0.40756575, Log Avg loss: 0.32053534, Global Avg Loss: 1.62101557, Time: 0.0211 Steps: 25630, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000092, Sample Num: 1472, Cur Loss: 0.07148466, Cur Avg Loss: 0.38476677, Log Avg loss: 0.19781506, Global Avg Loss: 1.62046050, Time: 0.0211 Steps: 25640, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000102, Sample Num: 1632, Cur Loss: 0.42767268, Cur Avg Loss: 0.38627085, Log Avg loss: 0.40010838, Global Avg Loss: 1.61998473, Time: 0.0211 Steps: 25650, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000112, Sample Num: 1792, Cur Loss: 0.29958546, Cur Avg Loss: 0.37944576, Log Avg loss: 0.30982989, Global Avg Loss: 1.61947415, Time: 0.0211 Steps: 25660, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000122, Sample Num: 1952, Cur Loss: 0.49554199, Cur Avg Loss: 0.36838865, Log Avg loss: 0.24454903, Global Avg Loss: 1.61893853, Time: 0.0211 Steps: 25670, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000132, Sample Num: 2112, Cur Loss: 0.42039150, Cur Avg Loss: 0.36641490, Log Avg loss: 0.34233512, Global Avg Loss: 1.61844141, Time: 0.0211 Steps: 25680, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000142, Sample Num: 2272, Cur Loss: 0.15539621, Cur Avg Loss: 0.35746521, Log Avg loss: 0.23932931, Global Avg Loss: 1.61790458, Time: 0.0211 Steps: 25690, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000152, Sample Num: 2432, Cur Loss: 0.82782370, Cur Avg Loss: 0.35687587, Log Avg loss: 0.34850722, Global Avg Loss: 1.61741065, Time: 0.0211 Steps: 25700, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000162, Sample Num: 2592, Cur Loss: 0.44884288, Cur Avg Loss: 0.35013103, Log Avg loss: 0.24760946, Global Avg Loss: 1.61687787, Time: 0.0211 Steps: 25710, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000172, Sample Num: 2752, Cur Loss: 0.28861520, Cur Avg Loss: 0.34482531, Log Avg loss: 0.25887262, Global Avg Loss: 1.61634987, Time: 0.0211 Steps: 25720, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000182, Sample Num: 2912, Cur Loss: 0.36138234, Cur Avg Loss: 0.34192194, Log Avg loss: 0.29198400, Global Avg Loss: 1.61583515, Time: 0.0210 Steps: 25730, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000192, Sample Num: 3072, Cur Loss: 0.24675563, Cur Avg Loss: 0.34295541, Log Avg loss: 0.36176468, Global Avg Loss: 1.61534795, Time: 0.0211 Steps: 25740, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000202, Sample Num: 3232, Cur Loss: 0.06859437, Cur Avg Loss: 0.34083158, Log Avg loss: 0.30005392, Global Avg Loss: 1.61483715, Time: 0.0210 Steps: 25750, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000212, Sample Num: 3392, Cur Loss: 0.37347591, Cur Avg Loss: 0.34190583, Log Avg loss: 0.36360576, Global Avg Loss: 1.61435143, Time: 0.0210 Steps: 25760, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000222, Sample Num: 3552, Cur Loss: 0.20122433, Cur Avg Loss: 0.33929370, Log Avg loss: 0.28391661, Global Avg Loss: 1.61383515, Time: 0.0210 Steps: 25770, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000232, Sample Num: 3712, Cur Loss: 0.21561475, Cur Avg Loss: 0.33276370, Log Avg loss: 0.18779749, Global Avg Loss: 1.61328200, Time: 0.0211 Steps: 25780, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000242, Sample Num: 3872, Cur Loss: 0.17186794, Cur Avg Loss: 0.33185627, Log Avg loss: 0.31080410, Global Avg Loss: 1.61277696, Time: 0.0210 Steps: 25790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000252, Sample Num: 4032, Cur Loss: 0.25408155, Cur Avg Loss: 0.34015159, Log Avg loss: 0.54089834, Global Avg Loss: 1.61236151, Time: 0.0211 Steps: 25800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000262, Sample Num: 4192, Cur Loss: 0.43635175, Cur Avg Loss: 0.33768271, Log Avg loss: 0.27546693, Global Avg Loss: 1.61184353, Time: 0.0211 Steps: 25810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000272, Sample Num: 4352, Cur Loss: 0.27720368, Cur Avg Loss: 0.33738079, Log Avg loss: 0.32947029, Global Avg Loss: 1.61134687, Time: 0.0211 Steps: 25820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000282, Sample Num: 4512, Cur Loss: 0.42653289, Cur Avg Loss: 0.33708074, Log Avg loss: 0.32891961, Global Avg Loss: 1.61085039, Time: 0.0211 Steps: 25830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000292, Sample Num: 4672, Cur Loss: 0.27773055, Cur Avg Loss: 0.33472276, Log Avg loss: 0.26822757, Global Avg Loss: 1.61033079, Time: 0.0211 Steps: 25840, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000302, Sample Num: 4832, Cur Loss: 0.83128345, Cur Avg Loss: 0.33791796, Log Avg loss: 0.43121790, Global Avg Loss: 1.60987466, Time: 0.0210 Steps: 25850, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000312, Sample Num: 4992, Cur Loss: 0.13017829, Cur Avg Loss: 0.33429030, Log Avg loss: 0.22473505, Global Avg Loss: 1.60933903, Time: 0.0211 Steps: 25860, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000322, Sample Num: 5152, Cur Loss: 0.25718001, Cur Avg Loss: 0.33141246, Log Avg loss: 0.24162376, Global Avg Loss: 1.60881034, Time: 0.0211 Steps: 25870, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000332, Sample Num: 5312, Cur Loss: 0.26000917, Cur Avg Loss: 0.33046869, Log Avg loss: 0.30007924, Global Avg Loss: 1.60830465, Time: 0.0211 Steps: 25880, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000342, Sample Num: 5472, Cur Loss: 0.32565078, Cur Avg Loss: 0.32970684, Log Avg loss: 0.30441346, Global Avg Loss: 1.60780102, Time: 0.0211 Steps: 25890, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000352, Sample Num: 5632, Cur Loss: 0.57941699, Cur Avg Loss: 0.32979777, Log Avg loss: 0.33290757, Global Avg Loss: 1.60730878, Time: 0.0211 Steps: 25900, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000362, Sample Num: 5792, Cur Loss: 0.16783756, Cur Avg Loss: 0.32878570, Log Avg loss: 0.29316076, Global Avg Loss: 1.60680159, Time: 0.0211 Steps: 25910, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000372, Sample Num: 5952, Cur Loss: 0.07743135, Cur Avg Loss: 0.32718057, Log Avg loss: 0.26907498, Global Avg Loss: 1.60628549, Time: 0.0211 Steps: 25920, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000382, Sample Num: 6112, Cur Loss: 0.28789669, Cur Avg Loss: 0.32556508, Log Avg loss: 0.26546871, Global Avg Loss: 1.60576840, Time: 0.0210 Steps: 25930, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000392, Sample Num: 6272, Cur Loss: 0.25394583, Cur Avg Loss: 0.32459131, Log Avg loss: 0.28739330, Global Avg Loss: 1.60526016, Time: 0.0211 Steps: 25940, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000402, Sample Num: 6432, Cur Loss: 0.14593002, Cur Avg Loss: 0.32062823, Log Avg loss: 0.16527575, Global Avg Loss: 1.60470525, Time: 0.0210 Steps: 25950, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000412, Sample Num: 6592, Cur Loss: 0.61897683, Cur Avg Loss: 0.32236013, Log Avg loss: 0.39198242, Global Avg Loss: 1.60423810, Time: 0.0211 Steps: 25960, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000422, Sample Num: 6752, Cur Loss: 0.33370918, Cur Avg Loss: 0.32192578, Log Avg loss: 0.30403043, Global Avg Loss: 1.60373744, Time: 0.0211 Steps: 25970, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000432, Sample Num: 6912, Cur Loss: 0.17505287, Cur Avg Loss: 0.32506957, Log Avg loss: 0.45773770, Global Avg Loss: 1.60329633, Time: 0.0210 Steps: 25980, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000442, Sample Num: 7072, Cur Loss: 0.32961777, Cur Avg Loss: 0.32423846, Log Avg loss: 0.28833428, Global Avg Loss: 1.60279038, Time: 0.0211 Steps: 25990, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000452, Sample Num: 7232, Cur Loss: 0.33477053, Cur Avg Loss: 0.32274641, Log Avg loss: 0.25679799, Global Avg Loss: 1.60227270, Time: 0.0211 Steps: 26000, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000462, Sample Num: 7392, Cur Loss: 0.23223978, Cur Avg Loss: 0.32127615, Log Avg loss: 0.25482035, Global Avg Loss: 1.60175464, Time: 0.0211 Steps: 26010, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000472, Sample Num: 7552, Cur Loss: 0.54211736, Cur Avg Loss: 0.31891998, Log Avg loss: 0.21006506, Global Avg Loss: 1.60121979, Time: 0.0210 Steps: 26020, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000482, Sample Num: 7712, Cur Loss: 0.17796758, Cur Avg Loss: 0.31896444, Log Avg loss: 0.32106283, Global Avg Loss: 1.60072799, Time: 0.0211 Steps: 26030, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000492, Sample Num: 7872, Cur Loss: 0.35407269, Cur Avg Loss: 0.32006295, Log Avg loss: 0.37301083, Global Avg Loss: 1.60025652, Time: 0.0210 Steps: 26040, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000502, Sample Num: 8032, Cur Loss: 0.24107368, Cur Avg Loss: 0.31979076, Log Avg loss: 0.30639944, Global Avg Loss: 1.59975983, Time: 0.0211 Steps: 26050, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000512, Sample Num: 8192, Cur Loss: 0.44457066, Cur Avg Loss: 0.31941847, Log Avg loss: 0.30072923, Global Avg Loss: 1.59926136, Time: 0.0256 Steps: 26060, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000522, Sample Num: 8352, Cur Loss: 0.11422075, Cur Avg Loss: 0.31889393, Log Avg loss: 0.29203739, Global Avg Loss: 1.59875993, Time: 0.0211 Steps: 26070, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000532, Sample Num: 8512, Cur Loss: 0.19320059, Cur Avg Loss: 0.31635217, Log Avg loss: 0.18367247, Global Avg Loss: 1.59821733, Time: 0.0211 Steps: 26080, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000542, Sample Num: 8672, Cur Loss: 0.41063324, Cur Avg Loss: 0.31602938, Log Avg loss: 0.29885716, Global Avg Loss: 1.59771930, Time: 0.0210 Steps: 26090, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000552, Sample Num: 8832, Cur Loss: 0.17769468, Cur Avg Loss: 0.31589470, Log Avg loss: 0.30859502, Global Avg Loss: 1.59722539, Time: 0.0210 Steps: 26100, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000562, Sample Num: 8992, Cur Loss: 0.52465481, Cur Avg Loss: 0.31660518, Log Avg loss: 0.35582348, Global Avg Loss: 1.59674994, Time: 0.0210 Steps: 26110, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000572, Sample Num: 9152, Cur Loss: 0.11470656, Cur Avg Loss: 0.31544739, Log Avg loss: 0.25037977, Global Avg Loss: 1.59623448, Time: 0.0210 Steps: 26120, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000582, Sample Num: 9312, Cur Loss: 0.06967200, Cur Avg Loss: 0.31350177, Log Avg loss: 0.20221205, Global Avg Loss: 1.59570098, Time: 0.0211 Steps: 26130, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000592, Sample Num: 9472, Cur Loss: 0.31664136, Cur Avg Loss: 0.31315931, Log Avg loss: 0.29322832, Global Avg Loss: 1.59520272, Time: 0.0210 Steps: 26140, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000602, Sample Num: 9632, Cur Loss: 0.40987194, Cur Avg Loss: 0.31364926, Log Avg loss: 0.34265410, Global Avg Loss: 1.59472373, Time: 0.0210 Steps: 26150, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000612, Sample Num: 9792, Cur Loss: 0.89840931, Cur Avg Loss: 0.31564579, Log Avg loss: 0.43583675, Global Avg Loss: 1.59428073, Time: 0.0210 Steps: 26160, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000622, Sample Num: 9952, Cur Loss: 0.13326010, Cur Avg Loss: 0.31512237, Log Avg loss: 0.28308937, Global Avg Loss: 1.59377970, Time: 0.0210 Steps: 26170, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000632, Sample Num: 10112, Cur Loss: 0.28209603, Cur Avg Loss: 0.31550142, Log Avg loss: 0.33907830, Global Avg Loss: 1.59330044, Time: 0.0211 Steps: 26180, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000642, Sample Num: 10272, Cur Loss: 0.11663616, Cur Avg Loss: 0.31399740, Log Avg loss: 0.21894327, Global Avg Loss: 1.59277568, Time: 0.0210 Steps: 26190, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000652, Sample Num: 10432, Cur Loss: 0.42704803, Cur Avg Loss: 0.31596775, Log Avg loss: 0.44246445, Global Avg Loss: 1.59233663, Time: 0.0211 Steps: 26200, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000662, Sample Num: 10592, Cur Loss: 0.12125865, Cur Avg Loss: 0.31659246, Log Avg loss: 0.35732333, Global Avg Loss: 1.59186543, Time: 0.0210 Steps: 26210, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000672, Sample Num: 10752, Cur Loss: 0.14512813, Cur Avg Loss: 0.31471046, Log Avg loss: 0.19012186, Global Avg Loss: 1.59133082, Time: 0.0210 Steps: 26220, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000682, Sample Num: 10912, Cur Loss: 0.42173472, Cur Avg Loss: 0.31378746, Log Avg loss: 0.25176221, Global Avg Loss: 1.59082012, Time: 0.0210 Steps: 26230, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000692, Sample Num: 11072, Cur Loss: 0.42656904, Cur Avg Loss: 0.31308025, Log Avg loss: 0.26484873, Global Avg Loss: 1.59031480, Time: 0.0212 Steps: 26240, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000702, Sample Num: 11232, Cur Loss: 0.33430922, Cur Avg Loss: 0.31176622, Log Avg loss: 0.22083471, Global Avg Loss: 1.58979309, Time: 0.0210 Steps: 26250, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000712, Sample Num: 11392, Cur Loss: 0.29272583, Cur Avg Loss: 0.31213185, Log Avg loss: 0.33779913, Global Avg Loss: 1.58931632, Time: 0.0210 Steps: 26260, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000722, Sample Num: 11552, Cur Loss: 0.23450884, Cur Avg Loss: 0.31182510, Log Avg loss: 0.28998508, Global Avg Loss: 1.58882171, Time: 0.0210 Steps: 26270, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000732, Sample Num: 11712, Cur Loss: 0.15138021, Cur Avg Loss: 0.31272313, Log Avg loss: 0.37756040, Global Avg Loss: 1.58836081, Time: 0.0210 Steps: 26280, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000742, Sample Num: 11872, Cur Loss: 0.34719732, Cur Avg Loss: 0.31157965, Log Avg loss: 0.22787687, Global Avg Loss: 1.58784332, Time: 0.0211 Steps: 26290, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000752, Sample Num: 12032, Cur Loss: 0.25061965, Cur Avg Loss: 0.31057707, Log Avg loss: 0.23618557, Global Avg Loss: 1.58732938, Time: 0.0210 Steps: 26300, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000762, Sample Num: 12192, Cur Loss: 0.17549205, Cur Avg Loss: 0.30952454, Log Avg loss: 0.23037420, Global Avg Loss: 1.58681362, Time: 0.0210 Steps: 26310, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000772, Sample Num: 12352, Cur Loss: 0.41745782, Cur Avg Loss: 0.30984026, Log Avg loss: 0.33389886, Global Avg Loss: 1.58633759, Time: 0.0248 Steps: 26320, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000782, Sample Num: 12512, Cur Loss: 0.47117096, Cur Avg Loss: 0.30962114, Log Avg loss: 0.29270435, Global Avg Loss: 1.58584628, Time: 0.0212 Steps: 26330, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000792, Sample Num: 12672, Cur Loss: 0.23632911, Cur Avg Loss: 0.30886599, Log Avg loss: 0.24981400, Global Avg Loss: 1.58533905, Time: 0.0211 Steps: 26340, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000802, Sample Num: 12832, Cur Loss: 0.23109779, Cur Avg Loss: 0.30886782, Log Avg loss: 0.30901204, Global Avg Loss: 1.58485467, Time: 0.0211 Steps: 26350, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000812, Sample Num: 12992, Cur Loss: 0.16041344, Cur Avg Loss: 0.30829266, Log Avg loss: 0.26216497, Global Avg Loss: 1.58435290, Time: 0.0211 Steps: 26360, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000822, Sample Num: 13152, Cur Loss: 0.23004462, Cur Avg Loss: 0.30839610, Log Avg loss: 0.31679568, Global Avg Loss: 1.58387221, Time: 0.0211 Steps: 26370, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000832, Sample Num: 13312, Cur Loss: 0.26219982, Cur Avg Loss: 0.30850716, Log Avg loss: 0.31763592, Global Avg Loss: 1.58339222, Time: 0.0211 Steps: 26380, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000842, Sample Num: 13472, Cur Loss: 0.37379065, Cur Avg Loss: 0.30841848, Log Avg loss: 0.30104035, Global Avg Loss: 1.58290629, Time: 0.0211 Steps: 26390, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000852, Sample Num: 13632, Cur Loss: 0.57021427, Cur Avg Loss: 0.30775729, Log Avg loss: 0.25208576, Global Avg Loss: 1.58240219, Time: 0.0211 Steps: 26400, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000862, Sample Num: 13792, Cur Loss: 0.65559483, Cur Avg Loss: 0.30886645, Log Avg loss: 0.40336621, Global Avg Loss: 1.58195576, Time: 0.0211 Steps: 26410, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000872, Sample Num: 13952, Cur Loss: 0.48534346, Cur Avg Loss: 0.31205123, Log Avg loss: 0.58657990, Global Avg Loss: 1.58157901, Time: 0.0212 Steps: 26420, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000882, Sample Num: 14112, Cur Loss: 0.44127926, Cur Avg Loss: 0.31297963, Log Avg loss: 0.39393581, Global Avg Loss: 1.58112965, Time: 0.0212 Steps: 26430, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000892, Sample Num: 14272, Cur Loss: 0.34880906, Cur Avg Loss: 0.31357774, Log Avg loss: 0.36633080, Global Avg Loss: 1.58067020, Time: 0.0211 Steps: 26440, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000902, Sample Num: 14432, Cur Loss: 0.25545111, Cur Avg Loss: 0.31305166, Log Avg loss: 0.26612503, Global Avg Loss: 1.58017321, Time: 0.0212 Steps: 26450, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000912, Sample Num: 14592, Cur Loss: 0.19276008, Cur Avg Loss: 0.31261507, Log Avg loss: 0.27323474, Global Avg Loss: 1.57967928, Time: 0.0211 Steps: 26460, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000922, Sample Num: 14752, Cur Loss: 0.86069268, Cur Avg Loss: 0.31245786, Log Avg loss: 0.29812050, Global Avg Loss: 1.57919512, Time: 0.0212 Steps: 26470, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000932, Sample Num: 14912, Cur Loss: 0.12099954, Cur Avg Loss: 0.31263323, Log Avg loss: 0.32880269, Global Avg Loss: 1.57872292, Time: 0.0212 Steps: 26480, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000942, Sample Num: 15072, Cur Loss: 0.50858635, Cur Avg Loss: 0.31278246, Log Avg loss: 0.32669094, Global Avg Loss: 1.57825027, Time: 0.0211 Steps: 26490, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000952, Sample Num: 15232, Cur Loss: 0.39115036, Cur Avg Loss: 0.31248227, Log Avg loss: 0.28420357, Global Avg Loss: 1.57776196, Time: 0.0212 Steps: 26500, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000962, Sample Num: 15392, Cur Loss: 0.13560055, Cur Avg Loss: 0.31394476, Log Avg loss: 0.45317394, Global Avg Loss: 1.57733774, Time: 0.0211 Steps: 26510, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000972, Sample Num: 15552, Cur Loss: 0.25789705, Cur Avg Loss: 0.31374085, Log Avg loss: 0.29412471, Global Avg Loss: 1.57685388, Time: 0.0211 Steps: 26520, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000982, Sample Num: 15712, Cur Loss: 0.25671202, Cur Avg Loss: 0.31629725, Log Avg loss: 0.56477968, Global Avg Loss: 1.57647239, Time: 0.0212 Steps: 26530, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000992, Sample Num: 15872, Cur Loss: 0.31902313, Cur Avg Loss: 0.31696987, Log Avg loss: 0.38302075, Global Avg Loss: 1.57602271, Time: 0.0211 Steps: 26540, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001002, Sample Num: 16032, Cur Loss: 0.41977090, Cur Avg Loss: 0.31653025, Log Avg loss: 0.27292010, Global Avg Loss: 1.57553190, Time: 0.0211 Steps: 26550, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001012, Sample Num: 16192, Cur Loss: 0.23159620, Cur Avg Loss: 0.31634213, Log Avg loss: 0.29749255, Global Avg Loss: 1.57505071, Time: 0.0212 Steps: 26560, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001022, Sample Num: 16352, Cur Loss: 0.70076907, Cur Avg Loss: 0.31836217, Log Avg loss: 0.52279070, Global Avg Loss: 1.57465468, Time: 0.0211 Steps: 26570, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001032, Sample Num: 16512, Cur Loss: 0.85779315, Cur Avg Loss: 0.32011221, Log Avg loss: 0.49896566, Global Avg Loss: 1.57424998, Time: 0.0213 Steps: 26580, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001042, Sample Num: 16672, Cur Loss: 0.38064197, Cur Avg Loss: 0.32092196, Log Avg loss: 0.40448824, Global Avg Loss: 1.57381006, Time: 0.0210 Steps: 26590, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001052, Sample Num: 16832, Cur Loss: 0.42865974, Cur Avg Loss: 0.32018304, Log Avg loss: 0.24318780, Global Avg Loss: 1.57330982, Time: 0.0211 Steps: 26600, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001062, Sample Num: 16992, Cur Loss: 0.18553452, Cur Avg Loss: 0.32034642, Log Avg loss: 0.33753358, Global Avg Loss: 1.57284542, Time: 0.0211 Steps: 26610, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001072, Sample Num: 17152, Cur Loss: 0.43128306, Cur Avg Loss: 0.31971023, Log Avg loss: 0.25214773, Global Avg Loss: 1.57234929, Time: 0.0211 Steps: 26620, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001082, Sample Num: 17312, Cur Loss: 0.33776072, Cur Avg Loss: 0.32028524, Log Avg loss: 0.38192630, Global Avg Loss: 1.57190227, Time: 0.0210 Steps: 26630, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001092, Sample Num: 17472, Cur Loss: 0.09149059, Cur Avg Loss: 0.32017164, Log Avg loss: 0.30787980, Global Avg Loss: 1.57142778, Time: 0.0211 Steps: 26640, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001102, Sample Num: 17632, Cur Loss: 0.54281473, Cur Avg Loss: 0.32057815, Log Avg loss: 0.36496907, Global Avg Loss: 1.57097508, Time: 0.0211 Steps: 26650, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001112, Sample Num: 17792, Cur Loss: 0.07734355, Cur Avg Loss: 0.31988405, Log Avg loss: 0.24339369, Global Avg Loss: 1.57047711, Time: 0.0211 Steps: 26660, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001122, Sample Num: 17952, Cur Loss: 0.34185371, Cur Avg Loss: 0.31982878, Log Avg loss: 0.31368346, Global Avg Loss: 1.57000587, Time: 0.0211 Steps: 26670, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001132, Sample Num: 18112, Cur Loss: 0.22897005, Cur Avg Loss: 0.31931232, Log Avg loss: 0.26136511, Global Avg Loss: 1.56951538, Time: 0.0210 Steps: 26680, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001142, Sample Num: 18272, Cur Loss: 0.20081246, Cur Avg Loss: 0.31870284, Log Avg loss: 0.24970979, Global Avg Loss: 1.56902088, Time: 0.0211 Steps: 26690, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001152, Sample Num: 18432, Cur Loss: 0.10409743, Cur Avg Loss: 0.31828745, Log Avg loss: 0.27084935, Global Avg Loss: 1.56853468, Time: 0.0210 Steps: 26700, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001162, Sample Num: 18592, Cur Loss: 0.28330338, Cur Avg Loss: 0.31875183, Log Avg loss: 0.37224847, Global Avg Loss: 1.56808680, Time: 0.0211 Steps: 26710, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001172, Sample Num: 18752, Cur Loss: 0.30744964, Cur Avg Loss: 0.31834619, Log Avg loss: 0.27121185, Global Avg Loss: 1.56760144, Time: 0.0210 Steps: 26720, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001182, Sample Num: 18912, Cur Loss: 0.43177688, Cur Avg Loss: 0.31759204, Log Avg loss: 0.22920537, Global Avg Loss: 1.56710073, Time: 0.0211 Steps: 26730, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001192, Sample Num: 19072, Cur Loss: 0.25516927, Cur Avg Loss: 0.31672098, Log Avg loss: 0.21376169, Global Avg Loss: 1.56659462, Time: 0.0211 Steps: 26740, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001202, Sample Num: 19232, Cur Loss: 0.34370536, Cur Avg Loss: 0.31652421, Log Avg loss: 0.29306942, Global Avg Loss: 1.56611853, Time: 0.0210 Steps: 26750, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001212, Sample Num: 19392, Cur Loss: 0.26225358, Cur Avg Loss: 0.31629983, Log Avg loss: 0.28932900, Global Avg Loss: 1.56564141, Time: 0.0211 Steps: 26760, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001222, Sample Num: 19552, Cur Loss: 0.18728246, Cur Avg Loss: 0.31540208, Log Avg loss: 0.20659500, Global Avg Loss: 1.56513373, Time: 0.0210 Steps: 26770, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001232, Sample Num: 19712, Cur Loss: 0.29011071, Cur Avg Loss: 0.31522810, Log Avg loss: 0.29396736, Global Avg Loss: 1.56465906, Time: 0.0211 Steps: 26780, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001242, Sample Num: 19872, Cur Loss: 0.53727180, Cur Avg Loss: 0.31577546, Log Avg loss: 0.38321065, Global Avg Loss: 1.56421806, Time: 0.0210 Steps: 26790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001252, Sample Num: 20032, Cur Loss: 0.34965548, Cur Avg Loss: 0.31568755, Log Avg loss: 0.30476825, Global Avg Loss: 1.56374812, Time: 0.0210 Steps: 26800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001262, Sample Num: 20192, Cur Loss: 0.37751138, Cur Avg Loss: 0.31561851, Log Avg loss: 0.30697574, Global Avg Loss: 1.56327935, Time: 0.0210 Steps: 26810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001272, Sample Num: 20352, Cur Loss: 0.18430923, Cur Avg Loss: 0.31587787, Log Avg loss: 0.34860880, Global Avg Loss: 1.56282645, Time: 0.0210 Steps: 26820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001282, Sample Num: 20512, Cur Loss: 0.39736038, Cur Avg Loss: 0.31561108, Log Avg loss: 0.28167568, Global Avg Loss: 1.56234894, Time: 0.0249 Steps: 26830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001292, Sample Num: 20672, Cur Loss: 0.67826569, Cur Avg Loss: 0.31704976, Log Avg loss: 0.50148824, Global Avg Loss: 1.56195369, Time: 0.0210 Steps: 26840, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001302, Sample Num: 20832, Cur Loss: 0.49196792, Cur Avg Loss: 0.31685273, Log Avg loss: 0.29139635, Global Avg Loss: 1.56148048, Time: 0.0210 Steps: 26850, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001312, Sample Num: 20992, Cur Loss: 0.40812182, Cur Avg Loss: 0.31743498, Log Avg loss: 0.39324364, Global Avg Loss: 1.56104555, Time: 0.0210 Steps: 26860, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001322, Sample Num: 21152, Cur Loss: 0.21888244, Cur Avg Loss: 0.31802093, Log Avg loss: 0.39489788, Global Avg Loss: 1.56061155, Time: 0.0210 Steps: 26870, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001332, Sample Num: 21312, Cur Loss: 0.21763852, Cur Avg Loss: 0.31826404, Log Avg loss: 0.35040312, Global Avg Loss: 1.56016132, Time: 0.0211 Steps: 26880, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001342, Sample Num: 21472, Cur Loss: 0.33024454, Cur Avg Loss: 0.31891261, Log Avg loss: 0.40530268, Global Avg Loss: 1.55973185, Time: 0.0210 Steps: 26890, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001352, Sample Num: 21632, Cur Loss: 0.55071688, Cur Avg Loss: 0.32009107, Log Avg loss: 0.47823967, Global Avg Loss: 1.55932981, Time: 0.0210 Steps: 26900, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001362, Sample Num: 21792, Cur Loss: 0.42621115, Cur Avg Loss: 0.32097459, Log Avg loss: 0.44042649, Global Avg Loss: 1.55891401, Time: 0.0210 Steps: 26910, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001372, Sample Num: 21952, Cur Loss: 0.10677998, Cur Avg Loss: 0.32078344, Log Avg loss: 0.29474894, Global Avg Loss: 1.55844441, Time: 0.0210 Steps: 26920, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001382, Sample Num: 22112, Cur Loss: 0.32415569, Cur Avg Loss: 0.32029049, Log Avg loss: 0.25265745, Global Avg Loss: 1.55795953, Time: 0.0210 Steps: 26930, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001392, Sample Num: 22272, Cur Loss: 0.78174114, Cur Avg Loss: 0.32042265, Log Avg loss: 0.33868744, Global Avg Loss: 1.55750694, Time: 0.0210 Steps: 26940, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001402, Sample Num: 22432, Cur Loss: 0.15127987, Cur Avg Loss: 0.32017522, Log Avg loss: 0.28573360, Global Avg Loss: 1.55703504, Time: 0.0210 Steps: 26950, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001412, Sample Num: 22592, Cur Loss: 0.34595549, Cur Avg Loss: 0.31960902, Log Avg loss: 0.24022697, Global Avg Loss: 1.55654661, Time: 0.0210 Steps: 26960, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001422, Sample Num: 22752, Cur Loss: 0.10785963, Cur Avg Loss: 0.31875643, Log Avg loss: 0.19837153, Global Avg Loss: 1.55604302, Time: 0.0210 Steps: 26970, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001432, Sample Num: 22912, Cur Loss: 0.18898246, Cur Avg Loss: 0.31807179, Log Avg loss: 0.22071575, Global Avg Loss: 1.55554809, Time: 0.0210 Steps: 26980, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001442, Sample Num: 23072, Cur Loss: 0.99476951, Cur Avg Loss: 0.31966565, Log Avg loss: 0.54790572, Global Avg Loss: 1.55517475, Time: 0.0210 Steps: 26990, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001452, Sample Num: 23232, Cur Loss: 0.39086527, Cur Avg Loss: 0.32132378, Log Avg loss: 0.56042680, Global Avg Loss: 1.55480633, Time: 0.0210 Steps: 27000, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001462, Sample Num: 23392, Cur Loss: 0.10130128, Cur Avg Loss: 0.32144942, Log Avg loss: 0.33969188, Global Avg Loss: 1.55435645, Time: 0.0210 Steps: 27010, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001472, Sample Num: 23552, Cur Loss: 0.18029276, Cur Avg Loss: 0.32085268, Log Avg loss: 0.23361000, Global Avg Loss: 1.55386765, Time: 0.0210 Steps: 27020, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001482, Sample Num: 23712, Cur Loss: 0.28041065, Cur Avg Loss: 0.32034413, Log Avg loss: 0.24548523, Global Avg Loss: 1.55338360, Time: 0.0210 Steps: 27030, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001492, Sample Num: 23872, Cur Loss: 0.21322948, Cur Avg Loss: 0.32016376, Log Avg loss: 0.29343331, Global Avg Loss: 1.55291764, Time: 0.0210 Steps: 27040, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001502, Sample Num: 24032, Cur Loss: 0.15552649, Cur Avg Loss: 0.32010201, Log Avg loss: 0.31088876, Global Avg Loss: 1.55245848, Time: 0.0210 Steps: 27050, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001512, Sample Num: 24192, Cur Loss: 0.57995677, Cur Avg Loss: 0.32028896, Log Avg loss: 0.34836796, Global Avg Loss: 1.55201351, Time: 0.0210 Steps: 27060, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001522, Sample Num: 24352, Cur Loss: 0.34539276, Cur Avg Loss: 0.32142804, Log Avg loss: 0.49365711, Global Avg Loss: 1.55162254, Time: 0.0210 Steps: 27070, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001532, Sample Num: 24512, Cur Loss: 0.40419540, Cur Avg Loss: 0.32166610, Log Avg loss: 0.35789967, Global Avg Loss: 1.55118173, Time: 0.0210 Steps: 27080, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001542, Sample Num: 24672, Cur Loss: 0.60451686, Cur Avg Loss: 0.32136658, Log Avg loss: 0.27547990, Global Avg Loss: 1.55071081, Time: 0.0211 Steps: 27090, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001552, Sample Num: 24832, Cur Loss: 0.17722407, Cur Avg Loss: 0.32175137, Log Avg loss: 0.38108503, Global Avg Loss: 1.55027922, Time: 0.0210 Steps: 27100, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001562, Sample Num: 24992, Cur Loss: 0.15369871, Cur Avg Loss: 0.32272941, Log Avg loss: 0.47452193, Global Avg Loss: 1.54988241, Time: 0.0210 Steps: 27110, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001572, Sample Num: 25152, Cur Loss: 0.48160338, Cur Avg Loss: 0.32354954, Log Avg loss: 0.45165410, Global Avg Loss: 1.54947745, Time: 0.0210 Steps: 27120, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001582, Sample Num: 25312, Cur Loss: 0.48340893, Cur Avg Loss: 0.32379114, Log Avg loss: 0.36177019, Global Avg Loss: 1.54903967, Time: 0.0211 Steps: 27130, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001592, Sample Num: 25472, Cur Loss: 0.15241379, Cur Avg Loss: 0.32459031, Log Avg loss: 0.45101956, Global Avg Loss: 1.54863509, Time: 0.0211 Steps: 27140, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001602, Sample Num: 25632, Cur Loss: 0.08242399, Cur Avg Loss: 0.32405769, Log Avg loss: 0.23926373, Global Avg Loss: 1.54815282, Time: 0.0211 Steps: 27150, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001612, Sample Num: 25792, Cur Loss: 0.08909986, Cur Avg Loss: 0.32326246, Log Avg loss: 0.19586694, Global Avg Loss: 1.54765493, Time: 0.0210 Steps: 27160, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001622, Sample Num: 25952, Cur Loss: 0.15589817, Cur Avg Loss: 0.32323027, Log Avg loss: 0.31804171, Global Avg Loss: 1.54720236, Time: 0.0211 Steps: 27170, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001632, Sample Num: 26112, Cur Loss: 0.65506649, Cur Avg Loss: 0.32288198, Log Avg loss: 0.26638902, Global Avg Loss: 1.54673113, Time: 0.0210 Steps: 27180, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001642, Sample Num: 26272, Cur Loss: 0.25763512, Cur Avg Loss: 0.32270606, Log Avg loss: 0.29399635, Global Avg Loss: 1.54627040, Time: 0.0210 Steps: 27190, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001652, Sample Num: 26432, Cur Loss: 0.28849062, Cur Avg Loss: 0.32244628, Log Avg loss: 0.27978973, Global Avg Loss: 1.54580478, Time: 0.0210 Steps: 27200, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001662, Sample Num: 26592, Cur Loss: 0.15848799, Cur Avg Loss: 0.32229135, Log Avg loss: 0.29669684, Global Avg Loss: 1.54534572, Time: 0.0210 Steps: 27210, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001672, Sample Num: 26752, Cur Loss: 0.26214936, Cur Avg Loss: 0.32200158, Log Avg loss: 0.27384148, Global Avg Loss: 1.54487859, Time: 0.0210 Steps: 27220, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001682, Sample Num: 26912, Cur Loss: 0.22912490, Cur Avg Loss: 0.32200708, Log Avg loss: 0.32292687, Global Avg Loss: 1.54442984, Time: 0.0210 Steps: 27230, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001692, Sample Num: 27072, Cur Loss: 0.15300873, Cur Avg Loss: 0.32249836, Log Avg loss: 0.40513203, Global Avg Loss: 1.54401160, Time: 0.0210 Steps: 27240, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001702, Sample Num: 27232, Cur Loss: 0.30595392, Cur Avg Loss: 0.32279157, Log Avg loss: 0.37240192, Global Avg Loss: 1.54358165, Time: 0.0210 Steps: 27250, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001712, Sample Num: 27392, Cur Loss: 0.28654653, Cur Avg Loss: 0.32294440, Log Avg loss: 0.34895698, Global Avg Loss: 1.54314342, Time: 0.0211 Steps: 27260, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001722, Sample Num: 27552, Cur Loss: 0.28040209, Cur Avg Loss: 0.32257712, Log Avg loss: 0.25969804, Global Avg Loss: 1.54267277, Time: 0.0211 Steps: 27270, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001732, Sample Num: 27712, Cur Loss: 0.45775843, Cur Avg Loss: 0.32307058, Log Avg loss: 0.40804552, Global Avg Loss: 1.54225685, Time: 0.0210 Steps: 27280, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001742, Sample Num: 27872, Cur Loss: 0.42629403, Cur Avg Loss: 0.32308492, Log Avg loss: 0.32556863, Global Avg Loss: 1.54181102, Time: 0.0210 Steps: 27290, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001752, Sample Num: 28032, Cur Loss: 0.23788851, Cur Avg Loss: 0.32298243, Log Avg loss: 0.30512793, Global Avg Loss: 1.54135802, Time: 0.0210 Steps: 27300, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001762, Sample Num: 28192, Cur Loss: 0.15428233, Cur Avg Loss: 0.32273254, Log Avg loss: 0.27895196, Global Avg Loss: 1.54089577, Time: 0.0210 Steps: 27310, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001772, Sample Num: 28352, Cur Loss: 0.58557612, Cur Avg Loss: 0.32294300, Log Avg loss: 0.36002564, Global Avg Loss: 1.54046353, Time: 0.0210 Steps: 27320, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001782, Sample Num: 28512, Cur Loss: 0.16714078, Cur Avg Loss: 0.32341323, Log Avg loss: 0.40673798, Global Avg Loss: 1.54004870, Time: 0.0210 Steps: 27330, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001792, Sample Num: 28672, Cur Loss: 0.22337726, Cur Avg Loss: 0.32251533, Log Avg loss: 0.16250935, Global Avg Loss: 1.53954485, Time: 0.0255 Steps: 27340, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001802, Sample Num: 28832, Cur Loss: 0.08454495, Cur Avg Loss: 0.32233725, Log Avg loss: 0.29042678, Global Avg Loss: 1.53908813, Time: 0.0211 Steps: 27350, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001812, Sample Num: 28992, Cur Loss: 0.04754201, Cur Avg Loss: 0.32229262, Log Avg loss: 0.31424887, Global Avg Loss: 1.53864046, Time: 0.0210 Steps: 27360, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001822, Sample Num: 29152, Cur Loss: 0.34118125, Cur Avg Loss: 0.32220531, Log Avg loss: 0.30638558, Global Avg Loss: 1.53819024, Time: 0.0210 Steps: 27370, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001832, Sample Num: 29312, Cur Loss: 0.16587457, Cur Avg Loss: 0.32221895, Log Avg loss: 0.32470445, Global Avg Loss: 1.53774703, Time: 0.0210 Steps: 27380, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001842, Sample Num: 29472, Cur Loss: 1.43136907, Cur Avg Loss: 0.32347277, Log Avg loss: 0.55317155, Global Avg Loss: 1.53738757, Time: 0.0210 Steps: 27390, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001852, Sample Num: 29632, Cur Loss: 0.05407540, Cur Avg Loss: 0.32356619, Log Avg loss: 0.34077393, Global Avg Loss: 1.53695085, Time: 0.0211 Steps: 27400, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001862, Sample Num: 29792, Cur Loss: 0.18108502, Cur Avg Loss: 0.32318879, Log Avg loss: 0.25329426, Global Avg Loss: 1.53648253, Time: 0.0210 Steps: 27410, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001872, Sample Num: 29952, Cur Loss: 0.49979001, Cur Avg Loss: 0.32263991, Log Avg loss: 0.22043967, Global Avg Loss: 1.53600258, Time: 0.0210 Steps: 27420, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001882, Sample Num: 30112, Cur Loss: 0.28501487, Cur Avg Loss: 0.32363197, Log Avg loss: 0.50934598, Global Avg Loss: 1.53562829, Time: 0.0210 Steps: 27430, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001892, Sample Num: 30272, Cur Loss: 0.42502272, Cur Avg Loss: 0.32458837, Log Avg loss: 0.50458198, Global Avg Loss: 1.53525255, Time: 0.0210 Steps: 27440, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001902, Sample Num: 30432, Cur Loss: 0.49137431, Cur Avg Loss: 0.32495976, Log Avg loss: 0.39522602, Global Avg Loss: 1.53483724, Time: 0.0210 Steps: 27450, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001912, Sample Num: 30592, Cur Loss: 0.36812839, Cur Avg Loss: 0.32515128, Log Avg loss: 0.36157992, Global Avg Loss: 1.53440998, Time: 0.0210 Steps: 27460, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001922, Sample Num: 30752, Cur Loss: 0.24815252, Cur Avg Loss: 0.32489278, Log Avg loss: 0.27546617, Global Avg Loss: 1.53395168, Time: 0.0210 Steps: 27470, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001932, Sample Num: 30912, Cur Loss: 0.14126554, Cur Avg Loss: 0.32417729, Log Avg loss: 0.18666028, Global Avg Loss: 1.53346140, Time: 0.0210 Steps: 27480, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001942, Sample Num: 31072, Cur Loss: 0.11145338, Cur Avg Loss: 0.32386623, Log Avg loss: 0.26377037, Global Avg Loss: 1.53299952, Time: 0.0210 Steps: 27490, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001952, Sample Num: 31232, Cur Loss: 0.10712703, Cur Avg Loss: 0.32362442, Log Avg loss: 0.27666395, Global Avg Loss: 1.53254267, Time: 0.0210 Steps: 27500, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001962, Sample Num: 31392, Cur Loss: 0.32266641, Cur Avg Loss: 0.32325166, Log Avg loss: 0.25048956, Global Avg Loss: 1.53207664, Time: 0.0210 Steps: 27510, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001972, Sample Num: 31552, Cur Loss: 0.21445233, Cur Avg Loss: 0.32295289, Log Avg loss: 0.26433403, Global Avg Loss: 1.53161598, Time: 0.0210 Steps: 27520, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001982, Sample Num: 31712, Cur Loss: 0.30897123, Cur Avg Loss: 0.32275748, Log Avg loss: 0.28422258, Global Avg Loss: 1.53116288, Time: 0.0210 Steps: 27530, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001992, Sample Num: 31872, Cur Loss: 0.09859669, Cur Avg Loss: 0.32238126, Log Avg loss: 0.24781467, Global Avg Loss: 1.53069688, Time: 0.0210 Steps: 27540, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002002, Sample Num: 32032, Cur Loss: 0.61007124, Cur Avg Loss: 0.32222705, Log Avg loss: 0.29150860, Global Avg Loss: 1.53024709, Time: 0.0210 Steps: 27550, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002012, Sample Num: 32192, Cur Loss: 0.48325107, Cur Avg Loss: 0.32247648, Log Avg loss: 0.37241277, Global Avg Loss: 1.52982697, Time: 0.0210 Steps: 27560, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002022, Sample Num: 32352, Cur Loss: 0.29175559, Cur Avg Loss: 0.32221501, Log Avg loss: 0.26960752, Global Avg Loss: 1.52936988, Time: 0.0210 Steps: 27570, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002032, Sample Num: 32512, Cur Loss: 0.42300197, Cur Avg Loss: 0.32267144, Log Avg loss: 0.41496154, Global Avg Loss: 1.52896581, Time: 0.0210 Steps: 27580, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002042, Sample Num: 32672, Cur Loss: 0.08968140, Cur Avg Loss: 0.32244284, Log Avg loss: 0.27599063, Global Avg Loss: 1.52851167, Time: 0.0210 Steps: 27590, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002052, Sample Num: 32832, Cur Loss: 0.52692747, Cur Avg Loss: 0.32244010, Log Avg loss: 0.32187954, Global Avg Loss: 1.52807448, Time: 0.0248 Steps: 27600, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002062, Sample Num: 32992, Cur Loss: 0.30872196, Cur Avg Loss: 0.32236946, Log Avg loss: 0.30787533, Global Avg Loss: 1.52763254, Time: 0.0211 Steps: 27610, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002072, Sample Num: 33152, Cur Loss: 0.54904318, Cur Avg Loss: 0.32251834, Log Avg loss: 0.35321774, Global Avg Loss: 1.52720734, Time: 0.0211 Steps: 27620, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002082, Sample Num: 33312, Cur Loss: 0.05829567, Cur Avg Loss: 0.32212276, Log Avg loss: 0.24015784, Global Avg Loss: 1.52674152, Time: 0.0211 Steps: 27630, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002092, Sample Num: 33472, Cur Loss: 0.23903422, Cur Avg Loss: 0.32201578, Log Avg loss: 0.29974318, Global Avg Loss: 1.52629760, Time: 0.0211 Steps: 27640, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002102, Sample Num: 33632, Cur Loss: 0.19545144, Cur Avg Loss: 0.32215591, Log Avg loss: 0.35147052, Global Avg Loss: 1.52587271, Time: 0.0210 Steps: 27650, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002112, Sample Num: 33792, Cur Loss: 0.45895317, Cur Avg Loss: 0.32220957, Log Avg loss: 0.33348793, Global Avg Loss: 1.52544162, Time: 0.0211 Steps: 27660, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002122, Sample Num: 33952, Cur Loss: 0.13192889, Cur Avg Loss: 0.32214099, Log Avg loss: 0.30765877, Global Avg Loss: 1.52500151, Time: 0.0211 Steps: 27670, Updated lr: 0.000075 ***** Running evaluation checkpoint-27677 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-27677 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.146567, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.275919, "eval_total_loss": 193.970839, "eval_mae": 0.371208, "eval_mse": 0.275994, "eval_r2": 0.82456, "eval_sp_statistic": 0.884996, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.909822, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.240274, "test_total_loss": 120.61778, "test_mae": 0.318339, "test_mse": 0.240358, "test_r2": 0.844871, "test_sp_statistic": 0.881991, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.932198, "test_ps_pvalue": 0.0, "lr": 7.470175438596491e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5246695640721724, "train_cur_epoch_loss": 685.0708291269839, "train_cur_epoch_avg_loss": 0.32178056793188536, "train_cur_epoch_time": 45.14656734466553, "train_cur_epoch_avg_time": 0.021205527169875775, "epoch": 13, "step": 27677} ################################################## Training, Epoch: 0014, Batch: 000003, Sample Num: 48, Cur Loss: 0.12994757, Cur Avg Loss: 0.17590577, Log Avg loss: 0.20153562, Global Avg Loss: 1.52452338, Time: 0.0247 Steps: 27680, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000013, Sample Num: 208, Cur Loss: 0.23790953, Cur Avg Loss: 0.28175852, Log Avg loss: 0.31351434, Global Avg Loss: 1.52408604, Time: 0.0212 Steps: 27690, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000023, Sample Num: 368, Cur Loss: 0.40299869, Cur Avg Loss: 0.25934365, Log Avg loss: 0.23020433, Global Avg Loss: 1.52361893, Time: 0.0211 Steps: 27700, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000033, Sample Num: 528, Cur Loss: 0.35283676, Cur Avg Loss: 0.28995429, Log Avg loss: 0.36035874, Global Avg Loss: 1.52319913, Time: 0.0211 Steps: 27710, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000043, Sample Num: 688, Cur Loss: 0.15453830, Cur Avg Loss: 0.26733488, Log Avg loss: 0.19269083, Global Avg Loss: 1.52271915, Time: 0.0211 Steps: 27720, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000053, Sample Num: 848, Cur Loss: 0.41100121, Cur Avg Loss: 0.30224547, Log Avg loss: 0.45236099, Global Avg Loss: 1.52233316, Time: 0.0211 Steps: 27730, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000063, Sample Num: 1008, Cur Loss: 0.20375526, Cur Avg Loss: 0.29628965, Log Avg loss: 0.26472385, Global Avg Loss: 1.52187980, Time: 0.0210 Steps: 27740, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000073, Sample Num: 1168, Cur Loss: 0.11418172, Cur Avg Loss: 0.30058015, Log Avg loss: 0.32761030, Global Avg Loss: 1.52144944, Time: 0.0211 Steps: 27750, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000083, Sample Num: 1328, Cur Loss: 0.16447586, Cur Avg Loss: 0.31701785, Log Avg loss: 0.43701303, Global Avg Loss: 1.52105879, Time: 0.0210 Steps: 27760, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000093, Sample Num: 1488, Cur Loss: 0.88866818, Cur Avg Loss: 0.32409811, Log Avg loss: 0.38286431, Global Avg Loss: 1.52064893, Time: 0.0211 Steps: 27770, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000103, Sample Num: 1648, Cur Loss: 0.20250043, Cur Avg Loss: 0.32213736, Log Avg loss: 0.30390230, Global Avg Loss: 1.52021093, Time: 0.0212 Steps: 27780, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000113, Sample Num: 1808, Cur Loss: 0.11378480, Cur Avg Loss: 0.30944323, Log Avg loss: 0.17869378, Global Avg Loss: 1.51972820, Time: 0.0210 Steps: 27790, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000123, Sample Num: 1968, Cur Loss: 0.13631497, Cur Avg Loss: 0.30362716, Log Avg loss: 0.23790549, Global Avg Loss: 1.51926711, Time: 0.0211 Steps: 27800, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000133, Sample Num: 2128, Cur Loss: 1.05849934, Cur Avg Loss: 0.31056260, Log Avg loss: 0.39586857, Global Avg Loss: 1.51886316, Time: 0.0212 Steps: 27810, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000143, Sample Num: 2288, Cur Loss: 0.06179727, Cur Avg Loss: 0.30943477, Log Avg loss: 0.29443456, Global Avg Loss: 1.51842303, Time: 0.0211 Steps: 27820, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000153, Sample Num: 2448, Cur Loss: 0.49509242, Cur Avg Loss: 0.30664347, Log Avg loss: 0.26672796, Global Avg Loss: 1.51797327, Time: 0.0211 Steps: 27830, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000163, Sample Num: 2608, Cur Loss: 0.19142881, Cur Avg Loss: 0.30846378, Log Avg loss: 0.33631442, Global Avg Loss: 1.51754882, Time: 0.0211 Steps: 27840, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000173, Sample Num: 2768, Cur Loss: 0.27315700, Cur Avg Loss: 0.30375959, Log Avg loss: 0.22708134, Global Avg Loss: 1.51708546, Time: 0.0211 Steps: 27850, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000183, Sample Num: 2928, Cur Loss: 0.08382043, Cur Avg Loss: 0.30201846, Log Avg loss: 0.27189691, Global Avg Loss: 1.51663851, Time: 0.0211 Steps: 27860, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000193, Sample Num: 3088, Cur Loss: 0.07087132, Cur Avg Loss: 0.29881192, Log Avg loss: 0.24013226, Global Avg Loss: 1.51618049, Time: 0.0211 Steps: 27870, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000203, Sample Num: 3248, Cur Loss: 0.07315002, Cur Avg Loss: 0.29424089, Log Avg loss: 0.20602010, Global Avg Loss: 1.51571056, Time: 0.0210 Steps: 27880, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000213, Sample Num: 3408, Cur Loss: 0.10371454, Cur Avg Loss: 0.29327359, Log Avg loss: 0.27363729, Global Avg Loss: 1.51526521, Time: 0.0211 Steps: 27890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000223, Sample Num: 3568, Cur Loss: 0.36874333, Cur Avg Loss: 0.29421684, Log Avg loss: 0.31430800, Global Avg Loss: 1.51483476, Time: 0.0211 Steps: 27900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000233, Sample Num: 3728, Cur Loss: 0.27925792, Cur Avg Loss: 0.29630793, Log Avg loss: 0.34293939, Global Avg Loss: 1.51441488, Time: 0.0212 Steps: 27910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000243, Sample Num: 3888, Cur Loss: 0.14504139, Cur Avg Loss: 0.29321260, Log Avg loss: 0.22109127, Global Avg Loss: 1.51395165, Time: 0.0211 Steps: 27920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000253, Sample Num: 4048, Cur Loss: 0.16300529, Cur Avg Loss: 0.29013407, Log Avg loss: 0.21532591, Global Avg Loss: 1.51348670, Time: 0.0211 Steps: 27930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000263, Sample Num: 4208, Cur Loss: 0.17321330, Cur Avg Loss: 0.28622900, Log Avg loss: 0.18743079, Global Avg Loss: 1.51301209, Time: 0.0211 Steps: 27940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000273, Sample Num: 4368, Cur Loss: 0.13527079, Cur Avg Loss: 0.28254102, Log Avg loss: 0.18554710, Global Avg Loss: 1.51253715, Time: 0.0210 Steps: 27950, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000283, Sample Num: 4528, Cur Loss: 0.13149761, Cur Avg Loss: 0.27937715, Log Avg loss: 0.19300345, Global Avg Loss: 1.51206521, Time: 0.0210 Steps: 27960, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000293, Sample Num: 4688, Cur Loss: 0.44752833, Cur Avg Loss: 0.27903167, Log Avg loss: 0.26925462, Global Avg Loss: 1.51162087, Time: 0.0210 Steps: 27970, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000303, Sample Num: 4848, Cur Loss: 0.19533695, Cur Avg Loss: 0.27818590, Log Avg loss: 0.25340492, Global Avg Loss: 1.51117119, Time: 0.0210 Steps: 27980, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000313, Sample Num: 5008, Cur Loss: 0.17743109, Cur Avg Loss: 0.27830489, Log Avg loss: 0.28191004, Global Avg Loss: 1.51073201, Time: 0.0210 Steps: 27990, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000323, Sample Num: 5168, Cur Loss: 0.49701366, Cur Avg Loss: 0.28048267, Log Avg loss: 0.34864716, Global Avg Loss: 1.51031698, Time: 0.0210 Steps: 28000, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000333, Sample Num: 5328, Cur Loss: 0.13565922, Cur Avg Loss: 0.27920717, Log Avg loss: 0.23800857, Global Avg Loss: 1.50986275, Time: 0.0210 Steps: 28010, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000343, Sample Num: 5488, Cur Loss: 0.88142920, Cur Avg Loss: 0.28359309, Log Avg loss: 0.42964438, Global Avg Loss: 1.50947723, Time: 0.0210 Steps: 28020, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000353, Sample Num: 5648, Cur Loss: 0.31495309, Cur Avg Loss: 0.28237510, Log Avg loss: 0.24059797, Global Avg Loss: 1.50902454, Time: 0.0210 Steps: 28030, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000363, Sample Num: 5808, Cur Loss: 0.49660537, Cur Avg Loss: 0.28328998, Log Avg loss: 0.31558529, Global Avg Loss: 1.50859892, Time: 0.0210 Steps: 28040, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000373, Sample Num: 5968, Cur Loss: 0.21503884, Cur Avg Loss: 0.28186414, Log Avg loss: 0.23010620, Global Avg Loss: 1.50814313, Time: 0.0210 Steps: 28050, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000383, Sample Num: 6128, Cur Loss: 0.27736723, Cur Avg Loss: 0.28258182, Log Avg loss: 0.30935118, Global Avg Loss: 1.50771591, Time: 0.0210 Steps: 28060, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000393, Sample Num: 6288, Cur Loss: 0.22339270, Cur Avg Loss: 0.28247753, Log Avg loss: 0.27848336, Global Avg Loss: 1.50727799, Time: 0.0210 Steps: 28070, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000403, Sample Num: 6448, Cur Loss: 0.24541560, Cur Avg Loss: 0.28392878, Log Avg loss: 0.34096277, Global Avg Loss: 1.50686264, Time: 0.0210 Steps: 28080, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000413, Sample Num: 6608, Cur Loss: 0.31955951, Cur Avg Loss: 0.28263601, Log Avg loss: 0.23053761, Global Avg Loss: 1.50640827, Time: 0.0210 Steps: 28090, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000423, Sample Num: 6768, Cur Loss: 0.20196071, Cur Avg Loss: 0.28199824, Log Avg loss: 0.25565803, Global Avg Loss: 1.50596316, Time: 0.0210 Steps: 28100, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000433, Sample Num: 6928, Cur Loss: 0.46940386, Cur Avg Loss: 0.28332829, Log Avg loss: 0.33958932, Global Avg Loss: 1.50554823, Time: 0.0210 Steps: 28110, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000443, Sample Num: 7088, Cur Loss: 0.43609846, Cur Avg Loss: 0.28261750, Log Avg loss: 0.25184052, Global Avg Loss: 1.50510239, Time: 0.0210 Steps: 28120, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000453, Sample Num: 7248, Cur Loss: 0.44433850, Cur Avg Loss: 0.28407028, Log Avg loss: 0.34842829, Global Avg Loss: 1.50469120, Time: 0.0211 Steps: 28130, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000463, Sample Num: 7408, Cur Loss: 0.30849582, Cur Avg Loss: 0.28513067, Log Avg loss: 0.33316653, Global Avg Loss: 1.50427488, Time: 0.0210 Steps: 28140, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000473, Sample Num: 7568, Cur Loss: 0.20478389, Cur Avg Loss: 0.28451632, Log Avg loss: 0.25607171, Global Avg Loss: 1.50383147, Time: 0.0210 Steps: 28150, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000483, Sample Num: 7728, Cur Loss: 0.35513085, Cur Avg Loss: 0.28389476, Log Avg loss: 0.25449516, Global Avg Loss: 1.50338781, Time: 0.0210 Steps: 28160, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000493, Sample Num: 7888, Cur Loss: 0.22843435, Cur Avg Loss: 0.28436622, Log Avg loss: 0.30713781, Global Avg Loss: 1.50296315, Time: 0.0210 Steps: 28170, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000503, Sample Num: 8048, Cur Loss: 0.11482019, Cur Avg Loss: 0.28621364, Log Avg loss: 0.37729137, Global Avg Loss: 1.50256370, Time: 0.0210 Steps: 28180, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000513, Sample Num: 8208, Cur Loss: 0.19564524, Cur Avg Loss: 0.28548591, Log Avg loss: 0.24888116, Global Avg Loss: 1.50211897, Time: 0.0248 Steps: 28190, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000523, Sample Num: 8368, Cur Loss: 0.04779470, Cur Avg Loss: 0.28555013, Log Avg loss: 0.28884430, Global Avg Loss: 1.50168873, Time: 0.0210 Steps: 28200, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000533, Sample Num: 8528, Cur Loss: 0.11038516, Cur Avg Loss: 0.28598724, Log Avg loss: 0.30884831, Global Avg Loss: 1.50126589, Time: 0.0210 Steps: 28210, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000543, Sample Num: 8688, Cur Loss: 0.54474175, Cur Avg Loss: 0.28512983, Log Avg loss: 0.23942971, Global Avg Loss: 1.50081875, Time: 0.0210 Steps: 28220, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000553, Sample Num: 8848, Cur Loss: 0.22824769, Cur Avg Loss: 0.28351153, Log Avg loss: 0.19563790, Global Avg Loss: 1.50035641, Time: 0.0211 Steps: 28230, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000563, Sample Num: 9008, Cur Loss: 0.30120713, Cur Avg Loss: 0.28358402, Log Avg loss: 0.28759251, Global Avg Loss: 1.49992696, Time: 0.0210 Steps: 28240, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000573, Sample Num: 9168, Cur Loss: 0.13218087, Cur Avg Loss: 0.28306421, Log Avg loss: 0.25379934, Global Avg Loss: 1.49948585, Time: 0.0210 Steps: 28250, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000583, Sample Num: 9328, Cur Loss: 0.23425931, Cur Avg Loss: 0.28256580, Log Avg loss: 0.25400670, Global Avg Loss: 1.49904513, Time: 0.0210 Steps: 28260, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000593, Sample Num: 9488, Cur Loss: 0.03759825, Cur Avg Loss: 0.28230260, Log Avg loss: 0.26695817, Global Avg Loss: 1.49860930, Time: 0.0211 Steps: 28270, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000603, Sample Num: 9648, Cur Loss: 0.13382147, Cur Avg Loss: 0.28070291, Log Avg loss: 0.18584106, Global Avg Loss: 1.49814510, Time: 0.0211 Steps: 28280, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000613, Sample Num: 9808, Cur Loss: 0.51829737, Cur Avg Loss: 0.28203559, Log Avg loss: 0.36239658, Global Avg Loss: 1.49774363, Time: 0.0211 Steps: 28290, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000623, Sample Num: 9968, Cur Loss: 0.06790362, Cur Avg Loss: 0.28117755, Log Avg loss: 0.22857923, Global Avg Loss: 1.49729516, Time: 0.0211 Steps: 28300, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000633, Sample Num: 10128, Cur Loss: 0.60511827, Cur Avg Loss: 0.28146613, Log Avg loss: 0.29944478, Global Avg Loss: 1.49687204, Time: 0.0208 Steps: 28310, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000643, Sample Num: 10288, Cur Loss: 0.59483498, Cur Avg Loss: 0.28111254, Log Avg loss: 0.25873033, Global Avg Loss: 1.49643485, Time: 0.0209 Steps: 28320, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000653, Sample Num: 10448, Cur Loss: 0.33931032, Cur Avg Loss: 0.28345981, Log Avg loss: 0.43438945, Global Avg Loss: 1.49605996, Time: 0.0208 Steps: 28330, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000663, Sample Num: 10608, Cur Loss: 0.07726000, Cur Avg Loss: 0.28279452, Log Avg loss: 0.23935119, Global Avg Loss: 1.49561652, Time: 0.0209 Steps: 28340, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000673, Sample Num: 10768, Cur Loss: 0.05049062, Cur Avg Loss: 0.28193155, Log Avg loss: 0.22471668, Global Avg Loss: 1.49516823, Time: 0.0209 Steps: 28350, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000683, Sample Num: 10928, Cur Loss: 0.30437279, Cur Avg Loss: 0.28122640, Log Avg loss: 0.23376969, Global Avg Loss: 1.49472345, Time: 0.0208 Steps: 28360, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000693, Sample Num: 11088, Cur Loss: 0.22538742, Cur Avg Loss: 0.28501842, Log Avg loss: 0.54401321, Global Avg Loss: 1.49438834, Time: 0.0209 Steps: 28370, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000703, Sample Num: 11248, Cur Loss: 0.22402060, Cur Avg Loss: 0.28428587, Log Avg loss: 0.23352048, Global Avg Loss: 1.49394406, Time: 0.0208 Steps: 28380, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000713, Sample Num: 11408, Cur Loss: 0.47348055, Cur Avg Loss: 0.28341651, Log Avg loss: 0.22230019, Global Avg Loss: 1.49349614, Time: 0.0209 Steps: 28390, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000723, Sample Num: 11568, Cur Loss: 0.31097347, Cur Avg Loss: 0.28225612, Log Avg loss: 0.19952067, Global Avg Loss: 1.49304052, Time: 0.0208 Steps: 28400, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000733, Sample Num: 11728, Cur Loss: 0.30143809, Cur Avg Loss: 0.28234496, Log Avg loss: 0.28876813, Global Avg Loss: 1.49261663, Time: 0.0208 Steps: 28410, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000743, Sample Num: 11888, Cur Loss: 0.20069462, Cur Avg Loss: 0.28271790, Log Avg loss: 0.31005417, Global Avg Loss: 1.49220053, Time: 0.0209 Steps: 28420, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000753, Sample Num: 12048, Cur Loss: 0.11154092, Cur Avg Loss: 0.28180695, Log Avg loss: 0.21412354, Global Avg Loss: 1.49175097, Time: 0.0209 Steps: 28430, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000763, Sample Num: 12208, Cur Loss: 0.52477252, Cur Avg Loss: 0.28155024, Log Avg loss: 0.26221961, Global Avg Loss: 1.49131865, Time: 0.0208 Steps: 28440, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000773, Sample Num: 12368, Cur Loss: 0.30121621, Cur Avg Loss: 0.28265157, Log Avg loss: 0.36668309, Global Avg Loss: 1.49092335, Time: 0.0246 Steps: 28450, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000783, Sample Num: 12528, Cur Loss: 1.61975789, Cur Avg Loss: 0.28667848, Log Avg loss: 0.59795884, Global Avg Loss: 1.49060958, Time: 0.0209 Steps: 28460, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000793, Sample Num: 12688, Cur Loss: 0.15138765, Cur Avg Loss: 0.28818539, Log Avg loss: 0.40617642, Global Avg Loss: 1.49022868, Time: 0.0209 Steps: 28470, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000803, Sample Num: 12848, Cur Loss: 0.58296931, Cur Avg Loss: 0.28909564, Log Avg loss: 0.36127825, Global Avg Loss: 1.48983228, Time: 0.0208 Steps: 28480, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000813, Sample Num: 13008, Cur Loss: 0.10610768, Cur Avg Loss: 0.29065999, Log Avg loss: 0.41627706, Global Avg Loss: 1.48945546, Time: 0.0208 Steps: 28490, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000823, Sample Num: 13168, Cur Loss: 0.39754963, Cur Avg Loss: 0.29009128, Log Avg loss: 0.24385546, Global Avg Loss: 1.48901841, Time: 0.0208 Steps: 28500, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000833, Sample Num: 13328, Cur Loss: 0.20885292, Cur Avg Loss: 0.28957774, Log Avg loss: 0.24731308, Global Avg Loss: 1.48858288, Time: 0.0209 Steps: 28510, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000843, Sample Num: 13488, Cur Loss: 0.24162544, Cur Avg Loss: 0.28920700, Log Avg loss: 0.25832452, Global Avg Loss: 1.48815151, Time: 0.0209 Steps: 28520, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000853, Sample Num: 13648, Cur Loss: 0.72573847, Cur Avg Loss: 0.29093174, Log Avg loss: 0.43632722, Global Avg Loss: 1.48778284, Time: 0.0209 Steps: 28530, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000863, Sample Num: 13808, Cur Loss: 0.21503749, Cur Avg Loss: 0.29189080, Log Avg loss: 0.37369907, Global Avg Loss: 1.48739248, Time: 0.0209 Steps: 28540, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000873, Sample Num: 13968, Cur Loss: 0.24222994, Cur Avg Loss: 0.29290314, Log Avg loss: 0.38026757, Global Avg Loss: 1.48700469, Time: 0.0208 Steps: 28550, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000883, Sample Num: 14128, Cur Loss: 0.51848602, Cur Avg Loss: 0.29341606, Log Avg loss: 0.33819395, Global Avg Loss: 1.48660245, Time: 0.0209 Steps: 28560, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000893, Sample Num: 14288, Cur Loss: 0.18655914, Cur Avg Loss: 0.29278305, Log Avg loss: 0.23688855, Global Avg Loss: 1.48616503, Time: 0.0208 Steps: 28570, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000903, Sample Num: 14448, Cur Loss: 0.39797783, Cur Avg Loss: 0.29345969, Log Avg loss: 0.35388379, Global Avg Loss: 1.48576885, Time: 0.0208 Steps: 28580, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000913, Sample Num: 14608, Cur Loss: 0.46257636, Cur Avg Loss: 0.29539386, Log Avg loss: 0.47004973, Global Avg Loss: 1.48541358, Time: 0.0208 Steps: 28590, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000923, Sample Num: 14768, Cur Loss: 0.32421976, Cur Avg Loss: 0.29545983, Log Avg loss: 0.30148222, Global Avg Loss: 1.48499961, Time: 0.0208 Steps: 28600, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000933, Sample Num: 14928, Cur Loss: 0.16522212, Cur Avg Loss: 0.29456565, Log Avg loss: 0.21203275, Global Avg Loss: 1.48455468, Time: 0.0208 Steps: 28610, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000943, Sample Num: 15088, Cur Loss: 0.07145856, Cur Avg Loss: 0.29379218, Log Avg loss: 0.22162749, Global Avg Loss: 1.48411340, Time: 0.0208 Steps: 28620, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000953, Sample Num: 15248, Cur Loss: 0.14696835, Cur Avg Loss: 0.29364660, Log Avg loss: 0.27991885, Global Avg Loss: 1.48369280, Time: 0.0208 Steps: 28630, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000963, Sample Num: 15408, Cur Loss: 0.74223083, Cur Avg Loss: 0.29369797, Log Avg loss: 0.29859368, Global Avg Loss: 1.48327900, Time: 0.0208 Steps: 28640, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000973, Sample Num: 15568, Cur Loss: 0.27938512, Cur Avg Loss: 0.29673965, Log Avg loss: 0.58965276, Global Avg Loss: 1.48296709, Time: 0.0208 Steps: 28650, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000983, Sample Num: 15728, Cur Loss: 0.68004131, Cur Avg Loss: 0.29658023, Log Avg loss: 0.28106932, Global Avg Loss: 1.48254773, Time: 0.0209 Steps: 28660, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000993, Sample Num: 15888, Cur Loss: 0.27971533, Cur Avg Loss: 0.29725332, Log Avg loss: 0.36341746, Global Avg Loss: 1.48215738, Time: 0.0208 Steps: 28670, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001003, Sample Num: 16048, Cur Loss: 0.18941945, Cur Avg Loss: 0.29705882, Log Avg loss: 0.27774537, Global Avg Loss: 1.48173743, Time: 0.0208 Steps: 28680, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001013, Sample Num: 16208, Cur Loss: 0.37689781, Cur Avg Loss: 0.29658920, Log Avg loss: 0.24948602, Global Avg Loss: 1.48130793, Time: 0.0208 Steps: 28690, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001023, Sample Num: 16368, Cur Loss: 0.38903290, Cur Avg Loss: 0.29811916, Log Avg loss: 0.45310428, Global Avg Loss: 1.48094967, Time: 0.0208 Steps: 28700, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001033, Sample Num: 16528, Cur Loss: 0.59747308, Cur Avg Loss: 0.29831533, Log Avg loss: 0.31838392, Global Avg Loss: 1.48054473, Time: 0.0209 Steps: 28710, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001043, Sample Num: 16688, Cur Loss: 0.14948839, Cur Avg Loss: 0.29781920, Log Avg loss: 0.24656810, Global Avg Loss: 1.48011507, Time: 0.0209 Steps: 28720, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001053, Sample Num: 16848, Cur Loss: 0.08527754, Cur Avg Loss: 0.29774840, Log Avg loss: 0.29036396, Global Avg Loss: 1.47970096, Time: 0.0209 Steps: 28730, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001063, Sample Num: 17008, Cur Loss: 0.42697883, Cur Avg Loss: 0.29732894, Log Avg loss: 0.25316054, Global Avg Loss: 1.47927419, Time: 0.0209 Steps: 28740, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001073, Sample Num: 17168, Cur Loss: 0.29716554, Cur Avg Loss: 0.29695754, Log Avg loss: 0.25747786, Global Avg Loss: 1.47884922, Time: 0.0209 Steps: 28750, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001083, Sample Num: 17328, Cur Loss: 0.07377551, Cur Avg Loss: 0.29679978, Log Avg loss: 0.27987216, Global Avg Loss: 1.47843233, Time: 0.0209 Steps: 28760, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001093, Sample Num: 17488, Cur Loss: 0.23088810, Cur Avg Loss: 0.29660785, Log Avg loss: 0.27582187, Global Avg Loss: 1.47801432, Time: 0.0208 Steps: 28770, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001103, Sample Num: 17648, Cur Loss: 0.16940126, Cur Avg Loss: 0.29635715, Log Avg loss: 0.26895507, Global Avg Loss: 1.47759421, Time: 0.0209 Steps: 28780, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001113, Sample Num: 17808, Cur Loss: 0.17256908, Cur Avg Loss: 0.29640194, Log Avg loss: 0.30134255, Global Avg Loss: 1.47718565, Time: 0.0208 Steps: 28790, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001123, Sample Num: 17968, Cur Loss: 0.17950498, Cur Avg Loss: 0.29632471, Log Avg loss: 0.28772921, Global Avg Loss: 1.47677265, Time: 0.0208 Steps: 28800, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001133, Sample Num: 18128, Cur Loss: 0.38191652, Cur Avg Loss: 0.29651186, Log Avg loss: 0.31752861, Global Avg Loss: 1.47637027, Time: 0.0209 Steps: 28810, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001143, Sample Num: 18288, Cur Loss: 0.31607345, Cur Avg Loss: 0.29664088, Log Avg loss: 0.31125876, Global Avg Loss: 1.47596600, Time: 0.0209 Steps: 28820, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001153, Sample Num: 18448, Cur Loss: 0.36888880, Cur Avg Loss: 0.29637839, Log Avg loss: 0.26637603, Global Avg Loss: 1.47554644, Time: 0.0209 Steps: 28830, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001163, Sample Num: 18608, Cur Loss: 0.28412443, Cur Avg Loss: 0.29617273, Log Avg loss: 0.27245935, Global Avg Loss: 1.47512928, Time: 0.0209 Steps: 28840, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001173, Sample Num: 18768, Cur Loss: 0.06097249, Cur Avg Loss: 0.29548493, Log Avg loss: 0.21549402, Global Avg Loss: 1.47469266, Time: 0.0209 Steps: 28850, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001183, Sample Num: 18928, Cur Loss: 0.31131095, Cur Avg Loss: 0.29569765, Log Avg loss: 0.32064949, Global Avg Loss: 1.47429279, Time: 0.0209 Steps: 28860, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001193, Sample Num: 19088, Cur Loss: 0.29812181, Cur Avg Loss: 0.29624546, Log Avg loss: 0.36105126, Global Avg Loss: 1.47390718, Time: 0.0209 Steps: 28870, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001203, Sample Num: 19248, Cur Loss: 0.43881524, Cur Avg Loss: 0.29655636, Log Avg loss: 0.33364698, Global Avg Loss: 1.47351236, Time: 0.0209 Steps: 28880, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001213, Sample Num: 19408, Cur Loss: 0.93376207, Cur Avg Loss: 0.29699046, Log Avg loss: 0.34921250, Global Avg Loss: 1.47312319, Time: 0.0209 Steps: 28890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001223, Sample Num: 19568, Cur Loss: 0.25513786, Cur Avg Loss: 0.29703376, Log Avg loss: 0.30228695, Global Avg Loss: 1.47271806, Time: 0.0209 Steps: 28900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001233, Sample Num: 19728, Cur Loss: 0.15201718, Cur Avg Loss: 0.29832076, Log Avg loss: 0.45572100, Global Avg Loss: 1.47236628, Time: 0.0209 Steps: 28910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001243, Sample Num: 19888, Cur Loss: 0.16691664, Cur Avg Loss: 0.29952210, Log Avg loss: 0.44764697, Global Avg Loss: 1.47201195, Time: 0.0209 Steps: 28920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001253, Sample Num: 20048, Cur Loss: 0.10910103, Cur Avg Loss: 0.29944445, Log Avg loss: 0.28979264, Global Avg Loss: 1.47160330, Time: 0.0209 Steps: 28930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001263, Sample Num: 20208, Cur Loss: 0.28484932, Cur Avg Loss: 0.29909283, Log Avg loss: 0.25503420, Global Avg Loss: 1.47118292, Time: 0.0209 Steps: 28940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001273, Sample Num: 20368, Cur Loss: 0.85476136, Cur Avg Loss: 0.29859173, Log Avg loss: 0.23530353, Global Avg Loss: 1.47075602, Time: 0.0209 Steps: 28950, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001283, Sample Num: 20528, Cur Loss: 0.40152693, Cur Avg Loss: 0.29902260, Log Avg loss: 0.35387257, Global Avg Loss: 1.47037036, Time: 0.0246 Steps: 28960, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001293, Sample Num: 20688, Cur Loss: 0.25386256, Cur Avg Loss: 0.29846357, Log Avg loss: 0.22673972, Global Avg Loss: 1.46994107, Time: 0.0209 Steps: 28970, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001303, Sample Num: 20848, Cur Loss: 0.15353319, Cur Avg Loss: 0.29798630, Log Avg loss: 0.23627524, Global Avg Loss: 1.46951538, Time: 0.0209 Steps: 28980, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001313, Sample Num: 21008, Cur Loss: 0.59591943, Cur Avg Loss: 0.30007068, Log Avg loss: 0.57166540, Global Avg Loss: 1.46920567, Time: 0.0209 Steps: 28990, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001323, Sample Num: 21168, Cur Loss: 0.14543039, Cur Avg Loss: 0.30072807, Log Avg loss: 0.38704360, Global Avg Loss: 1.46883251, Time: 0.0209 Steps: 29000, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001333, Sample Num: 21328, Cur Loss: 0.39260581, Cur Avg Loss: 0.30075425, Log Avg loss: 0.30421710, Global Avg Loss: 1.46843106, Time: 0.0209 Steps: 29010, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001343, Sample Num: 21488, Cur Loss: 0.24010383, Cur Avg Loss: 0.30028884, Log Avg loss: 0.23824982, Global Avg Loss: 1.46800715, Time: 0.0208 Steps: 29020, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001353, Sample Num: 21648, Cur Loss: 0.27778918, Cur Avg Loss: 0.30104091, Log Avg loss: 0.40204414, Global Avg Loss: 1.46763995, Time: 0.0208 Steps: 29030, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001363, Sample Num: 21808, Cur Loss: 0.09399386, Cur Avg Loss: 0.30116117, Log Avg loss: 0.31743279, Global Avg Loss: 1.46724388, Time: 0.0209 Steps: 29040, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001373, Sample Num: 21968, Cur Loss: 0.93756604, Cur Avg Loss: 0.30116500, Log Avg loss: 0.30168660, Global Avg Loss: 1.46684265, Time: 0.0209 Steps: 29050, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001383, Sample Num: 22128, Cur Loss: 0.24530232, Cur Avg Loss: 0.30178902, Log Avg loss: 0.38746658, Global Avg Loss: 1.46647122, Time: 0.0209 Steps: 29060, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001393, Sample Num: 22288, Cur Loss: 0.59005916, Cur Avg Loss: 0.30238213, Log Avg loss: 0.38441018, Global Avg Loss: 1.46609900, Time: 0.0209 Steps: 29070, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001403, Sample Num: 22448, Cur Loss: 0.18890415, Cur Avg Loss: 0.30225929, Log Avg loss: 0.28514666, Global Avg Loss: 1.46569289, Time: 0.0208 Steps: 29080, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001413, Sample Num: 22608, Cur Loss: 0.12855220, Cur Avg Loss: 0.30156804, Log Avg loss: 0.20458593, Global Avg Loss: 1.46525937, Time: 0.0209 Steps: 29090, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001423, Sample Num: 22768, Cur Loss: 0.08803609, Cur Avg Loss: 0.30170751, Log Avg loss: 0.32141430, Global Avg Loss: 1.46486630, Time: 0.0209 Steps: 29100, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001433, Sample Num: 22928, Cur Loss: 0.05113596, Cur Avg Loss: 0.30133446, Log Avg loss: 0.24824957, Global Avg Loss: 1.46444836, Time: 0.0209 Steps: 29110, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001443, Sample Num: 23088, Cur Loss: 0.36251259, Cur Avg Loss: 0.30120544, Log Avg loss: 0.28271695, Global Avg Loss: 1.46404255, Time: 0.0210 Steps: 29120, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001453, Sample Num: 23248, Cur Loss: 0.07057826, Cur Avg Loss: 0.30133074, Log Avg loss: 0.31941197, Global Avg Loss: 1.46364961, Time: 0.0209 Steps: 29130, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001463, Sample Num: 23408, Cur Loss: 0.07144567, Cur Avg Loss: 0.30096908, Log Avg loss: 0.24842001, Global Avg Loss: 1.46323258, Time: 0.0209 Steps: 29140, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001473, Sample Num: 23568, Cur Loss: 0.33792052, Cur Avg Loss: 0.30058959, Log Avg loss: 0.24506984, Global Avg Loss: 1.46281468, Time: 0.0209 Steps: 29150, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001483, Sample Num: 23728, Cur Loss: 0.15071200, Cur Avg Loss: 0.30005559, Log Avg loss: 0.22139807, Global Avg Loss: 1.46238896, Time: 0.0209 Steps: 29160, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001493, Sample Num: 23888, Cur Loss: 0.51696581, Cur Avg Loss: 0.30025082, Log Avg loss: 0.32920261, Global Avg Loss: 1.46200048, Time: 0.0209 Steps: 29170, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001503, Sample Num: 24048, Cur Loss: 0.29695821, Cur Avg Loss: 0.29972486, Log Avg loss: 0.22119877, Global Avg Loss: 1.46157526, Time: 0.0210 Steps: 29180, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001513, Sample Num: 24208, Cur Loss: 0.26416540, Cur Avg Loss: 0.29956123, Log Avg loss: 0.27496852, Global Avg Loss: 1.46116875, Time: 0.0209 Steps: 29190, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001523, Sample Num: 24368, Cur Loss: 0.23430720, Cur Avg Loss: 0.30019132, Log Avg loss: 0.39552426, Global Avg Loss: 1.46080380, Time: 0.0209 Steps: 29200, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001533, Sample Num: 24528, Cur Loss: 0.14317176, Cur Avg Loss: 0.30254195, Log Avg loss: 0.66054274, Global Avg Loss: 1.46052983, Time: 0.0209 Steps: 29210, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001543, Sample Num: 24688, Cur Loss: 0.06621110, Cur Avg Loss: 0.30207669, Log Avg loss: 0.23075182, Global Avg Loss: 1.46010896, Time: 0.0213 Steps: 29220, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001553, Sample Num: 24848, Cur Loss: 0.21417224, Cur Avg Loss: 0.30185457, Log Avg loss: 0.26758156, Global Avg Loss: 1.45970098, Time: 0.0209 Steps: 29230, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001563, Sample Num: 25008, Cur Loss: 0.62221658, Cur Avg Loss: 0.30183324, Log Avg loss: 0.29852019, Global Avg Loss: 1.45930386, Time: 0.0209 Steps: 29240, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001573, Sample Num: 25168, Cur Loss: 0.29396129, Cur Avg Loss: 0.30251551, Log Avg loss: 0.40915388, Global Avg Loss: 1.45894483, Time: 0.0209 Steps: 29250, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001583, Sample Num: 25328, Cur Loss: 0.44365034, Cur Avg Loss: 0.30296781, Log Avg loss: 0.37411556, Global Avg Loss: 1.45857408, Time: 0.0209 Steps: 29260, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001593, Sample Num: 25488, Cur Loss: 0.10998674, Cur Avg Loss: 0.30268791, Log Avg loss: 0.25838012, Global Avg Loss: 1.45816404, Time: 0.0208 Steps: 29270, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001603, Sample Num: 25648, Cur Loss: 0.10009877, Cur Avg Loss: 0.30259800, Log Avg loss: 0.28827489, Global Avg Loss: 1.45776448, Time: 0.0208 Steps: 29280, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001613, Sample Num: 25808, Cur Loss: 0.34360230, Cur Avg Loss: 0.30304966, Log Avg loss: 0.37545136, Global Avg Loss: 1.45739497, Time: 0.0209 Steps: 29290, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001623, Sample Num: 25968, Cur Loss: 0.10797814, Cur Avg Loss: 0.30265288, Log Avg loss: 0.23865189, Global Avg Loss: 1.45697902, Time: 0.0209 Steps: 29300, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001633, Sample Num: 26128, Cur Loss: 0.24464245, Cur Avg Loss: 0.30242207, Log Avg loss: 0.26496044, Global Avg Loss: 1.45657232, Time: 0.0209 Steps: 29310, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001643, Sample Num: 26288, Cur Loss: 0.15277122, Cur Avg Loss: 0.30244268, Log Avg loss: 0.30580885, Global Avg Loss: 1.45617984, Time: 0.0209 Steps: 29320, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001653, Sample Num: 26448, Cur Loss: 0.49962109, Cur Avg Loss: 0.30230551, Log Avg loss: 0.27976886, Global Avg Loss: 1.45577874, Time: 0.0209 Steps: 29330, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001663, Sample Num: 26608, Cur Loss: 0.17584249, Cur Avg Loss: 0.30185475, Log Avg loss: 0.22734434, Global Avg Loss: 1.45536005, Time: 0.0209 Steps: 29340, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001673, Sample Num: 26768, Cur Loss: 0.27649978, Cur Avg Loss: 0.30202202, Log Avg loss: 0.32983927, Global Avg Loss: 1.45497657, Time: 0.0209 Steps: 29350, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001683, Sample Num: 26928, Cur Loss: 0.13764438, Cur Avg Loss: 0.30133579, Log Avg loss: 0.18652948, Global Avg Loss: 1.45454454, Time: 0.0209 Steps: 29360, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001693, Sample Num: 27088, Cur Loss: 0.12813985, Cur Avg Loss: 0.30131594, Log Avg loss: 0.29797433, Global Avg Loss: 1.45415075, Time: 0.0209 Steps: 29370, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001703, Sample Num: 27248, Cur Loss: 1.03414798, Cur Avg Loss: 0.30194754, Log Avg loss: 0.40887738, Global Avg Loss: 1.45379497, Time: 0.0209 Steps: 29380, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001713, Sample Num: 27408, Cur Loss: 0.62962705, Cur Avg Loss: 0.30289810, Log Avg loss: 0.46477952, Global Avg Loss: 1.45345845, Time: 0.0209 Steps: 29390, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001723, Sample Num: 27568, Cur Loss: 1.05803490, Cur Avg Loss: 0.30548860, Log Avg loss: 0.74923991, Global Avg Loss: 1.45321892, Time: 0.0211 Steps: 29400, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001733, Sample Num: 27728, Cur Loss: 0.38164699, Cur Avg Loss: 0.30677456, Log Avg loss: 0.52834602, Global Avg Loss: 1.45290445, Time: 0.0209 Steps: 29410, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001743, Sample Num: 27888, Cur Loss: 0.18864772, Cur Avg Loss: 0.30666569, Log Avg loss: 0.28779797, Global Avg Loss: 1.45250842, Time: 0.0209 Steps: 29420, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001753, Sample Num: 28048, Cur Loss: 0.26166713, Cur Avg Loss: 0.30632013, Log Avg loss: 0.24609013, Global Avg Loss: 1.45209850, Time: 0.0209 Steps: 29430, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001763, Sample Num: 28208, Cur Loss: 0.28903800, Cur Avg Loss: 0.30624647, Log Avg loss: 0.29333366, Global Avg Loss: 1.45170489, Time: 0.0209 Steps: 29440, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001773, Sample Num: 28368, Cur Loss: 0.17549437, Cur Avg Loss: 0.30594831, Log Avg loss: 0.25338328, Global Avg Loss: 1.45129799, Time: 0.0209 Steps: 29450, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001783, Sample Num: 28528, Cur Loss: 0.35591200, Cur Avg Loss: 0.30576856, Log Avg loss: 0.27389830, Global Avg Loss: 1.45089833, Time: 0.0209 Steps: 29460, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001793, Sample Num: 28688, Cur Loss: 0.26817894, Cur Avg Loss: 0.30575623, Log Avg loss: 0.30355753, Global Avg Loss: 1.45050901, Time: 0.0245 Steps: 29470, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001803, Sample Num: 28848, Cur Loss: 0.60900044, Cur Avg Loss: 0.30581274, Log Avg loss: 0.31594501, Global Avg Loss: 1.45012415, Time: 0.0209 Steps: 29480, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001813, Sample Num: 29008, Cur Loss: 0.24051304, Cur Avg Loss: 0.30525754, Log Avg loss: 0.20515406, Global Avg Loss: 1.44970198, Time: 0.0209 Steps: 29490, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001823, Sample Num: 29168, Cur Loss: 0.19466227, Cur Avg Loss: 0.30535517, Log Avg loss: 0.32305599, Global Avg Loss: 1.44932007, Time: 0.0208 Steps: 29500, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001833, Sample Num: 29328, Cur Loss: 0.04570024, Cur Avg Loss: 0.30492047, Log Avg loss: 0.22567578, Global Avg Loss: 1.44890541, Time: 0.0209 Steps: 29510, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001843, Sample Num: 29488, Cur Loss: 0.26102903, Cur Avg Loss: 0.30502077, Log Avg loss: 0.32340484, Global Avg Loss: 1.44852415, Time: 0.0209 Steps: 29520, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001853, Sample Num: 29648, Cur Loss: 0.17877842, Cur Avg Loss: 0.30482420, Log Avg loss: 0.26859743, Global Avg Loss: 1.44812458, Time: 0.0209 Steps: 29530, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001863, Sample Num: 29808, Cur Loss: 1.03402209, Cur Avg Loss: 0.30527688, Log Avg loss: 0.38915704, Global Avg Loss: 1.44776609, Time: 0.0208 Steps: 29540, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001873, Sample Num: 29968, Cur Loss: 0.32317388, Cur Avg Loss: 0.30515291, Log Avg loss: 0.28205748, Global Avg Loss: 1.44737160, Time: 0.0209 Steps: 29550, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001883, Sample Num: 30128, Cur Loss: 0.20956773, Cur Avg Loss: 0.30492257, Log Avg loss: 0.26178045, Global Avg Loss: 1.44697053, Time: 0.0209 Steps: 29560, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001893, Sample Num: 30288, Cur Loss: 0.42291608, Cur Avg Loss: 0.30484254, Log Avg loss: 0.28977294, Global Avg Loss: 1.44657918, Time: 0.0208 Steps: 29570, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001903, Sample Num: 30448, Cur Loss: 0.52477080, Cur Avg Loss: 0.30474011, Log Avg loss: 0.28535056, Global Avg Loss: 1.44618661, Time: 0.0209 Steps: 29580, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001913, Sample Num: 30608, Cur Loss: 0.13309215, Cur Avg Loss: 0.30439339, Log Avg loss: 0.23841260, Global Avg Loss: 1.44577844, Time: 0.0208 Steps: 29590, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001923, Sample Num: 30768, Cur Loss: 0.34814692, Cur Avg Loss: 0.30432055, Log Avg loss: 0.29038551, Global Avg Loss: 1.44538811, Time: 0.0209 Steps: 29600, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001933, Sample Num: 30928, Cur Loss: 0.23561761, Cur Avg Loss: 0.30423868, Log Avg loss: 0.28849419, Global Avg Loss: 1.44499740, Time: 0.0209 Steps: 29610, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001943, Sample Num: 31088, Cur Loss: 0.22092186, Cur Avg Loss: 0.30416981, Log Avg loss: 0.29085893, Global Avg Loss: 1.44460775, Time: 0.0209 Steps: 29620, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001953, Sample Num: 31248, Cur Loss: 0.11051318, Cur Avg Loss: 0.30395846, Log Avg loss: 0.26289180, Global Avg Loss: 1.44420892, Time: 0.0208 Steps: 29630, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001963, Sample Num: 31408, Cur Loss: 0.36622632, Cur Avg Loss: 0.30382957, Log Avg loss: 0.27865724, Global Avg Loss: 1.44381569, Time: 0.0209 Steps: 29640, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001973, Sample Num: 31568, Cur Loss: 0.31668767, Cur Avg Loss: 0.30408909, Log Avg loss: 0.35503305, Global Avg Loss: 1.44344848, Time: 0.0209 Steps: 29650, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001983, Sample Num: 31728, Cur Loss: 0.10857505, Cur Avg Loss: 0.30425871, Log Avg loss: 0.33772470, Global Avg Loss: 1.44307568, Time: 0.0209 Steps: 29660, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001993, Sample Num: 31888, Cur Loss: 0.86748183, Cur Avg Loss: 0.30434449, Log Avg loss: 0.32135458, Global Avg Loss: 1.44269761, Time: 0.0209 Steps: 29670, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002003, Sample Num: 32048, Cur Loss: 0.18399490, Cur Avg Loss: 0.30445058, Log Avg loss: 0.32559561, Global Avg Loss: 1.44232123, Time: 0.0209 Steps: 29680, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002013, Sample Num: 32208, Cur Loss: 0.77066147, Cur Avg Loss: 0.30495942, Log Avg loss: 0.40687925, Global Avg Loss: 1.44197248, Time: 0.0208 Steps: 29690, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002023, Sample Num: 32368, Cur Loss: 0.38285223, Cur Avg Loss: 0.30518260, Log Avg loss: 0.35010871, Global Avg Loss: 1.44160485, Time: 0.0209 Steps: 29700, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002033, Sample Num: 32528, Cur Loss: 0.14127138, Cur Avg Loss: 0.30506149, Log Avg loss: 0.28056072, Global Avg Loss: 1.44121405, Time: 0.0208 Steps: 29710, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002043, Sample Num: 32688, Cur Loss: 0.45841828, Cur Avg Loss: 0.30537927, Log Avg loss: 0.36998401, Global Avg Loss: 1.44085361, Time: 0.0208 Steps: 29720, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002053, Sample Num: 32848, Cur Loss: 0.31599739, Cur Avg Loss: 0.30494902, Log Avg loss: 0.21704912, Global Avg Loss: 1.44044197, Time: 0.0246 Steps: 29730, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002063, Sample Num: 33008, Cur Loss: 0.24267782, Cur Avg Loss: 0.30460200, Log Avg loss: 0.23335976, Global Avg Loss: 1.44003609, Time: 0.0208 Steps: 29740, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002073, Sample Num: 33168, Cur Loss: 0.25006509, Cur Avg Loss: 0.30443160, Log Avg loss: 0.26927766, Global Avg Loss: 1.43964256, Time: 0.0209 Steps: 29750, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002083, Sample Num: 33328, Cur Loss: 0.26701477, Cur Avg Loss: 0.30396437, Log Avg loss: 0.20710701, Global Avg Loss: 1.43922840, Time: 0.0209 Steps: 29760, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002093, Sample Num: 33488, Cur Loss: 0.08856203, Cur Avg Loss: 0.30343598, Log Avg loss: 0.19337165, Global Avg Loss: 1.43880991, Time: 0.0209 Steps: 29770, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002103, Sample Num: 33648, Cur Loss: 0.33736825, Cur Avg Loss: 0.30350520, Log Avg loss: 0.31799407, Global Avg Loss: 1.43843354, Time: 0.0209 Steps: 29780, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002113, Sample Num: 33808, Cur Loss: 0.55425000, Cur Avg Loss: 0.30336395, Log Avg loss: 0.27365841, Global Avg Loss: 1.43804255, Time: 0.0209 Steps: 29790, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002123, Sample Num: 33968, Cur Loss: 0.17959160, Cur Avg Loss: 0.30298867, Log Avg loss: 0.22369315, Global Avg Loss: 1.43763505, Time: 0.0209 Steps: 29800, Updated lr: 0.000073 ***** Running evaluation checkpoint-29806 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-29806 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.747763, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.295098, "eval_total_loss": 207.453828, "eval_mae": 0.418736, "eval_mse": 0.295142, "eval_r2": 0.812388, "eval_sp_statistic": 0.89057, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.914552, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.222796, "test_total_loss": 111.843471, "test_mae": 0.319638, "test_mse": 0.222876, "test_r2": 0.856154, "test_sp_statistic": 0.885275, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.934802, "test_ps_pvalue": 0.0, "lr": 7.26827880512091e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4373969828768953, "train_cur_epoch_loss": 644.7749468032271, "train_cur_epoch_avg_loss": 0.3028534273382936, "train_cur_epoch_time": 44.74776268005371, "train_cur_epoch_avg_time": 0.021018206989222035, "epoch": 14, "step": 29806} ################################################## Training, Epoch: 0015, Batch: 000004, Sample Num: 64, Cur Loss: 0.07813133, Cur Avg Loss: 0.20342701, Log Avg loss: 0.23437009, Global Avg Loss: 1.43723140, Time: 0.0247 Steps: 29810, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000014, Sample Num: 224, Cur Loss: 0.26026148, Cur Avg Loss: 0.19767810, Log Avg loss: 0.19537854, Global Avg Loss: 1.43681496, Time: 0.0210 Steps: 29820, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000024, Sample Num: 384, Cur Loss: 0.10445238, Cur Avg Loss: 0.22080083, Log Avg loss: 0.25317265, Global Avg Loss: 1.43641816, Time: 0.0210 Steps: 29830, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000034, Sample Num: 544, Cur Loss: 0.63884181, Cur Avg Loss: 0.24839964, Log Avg loss: 0.31463678, Global Avg Loss: 1.43604223, Time: 0.0210 Steps: 29840, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000044, Sample Num: 704, Cur Loss: 0.52321088, Cur Avg Loss: 0.26765562, Log Avg loss: 0.33312594, Global Avg Loss: 1.43567274, Time: 0.0209 Steps: 29850, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000054, Sample Num: 864, Cur Loss: 0.38617310, Cur Avg Loss: 0.28360746, Log Avg loss: 0.35379559, Global Avg Loss: 1.43531042, Time: 0.0210 Steps: 29860, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000064, Sample Num: 1024, Cur Loss: 0.51059371, Cur Avg Loss: 0.29724146, Log Avg loss: 0.37086502, Global Avg Loss: 1.43495407, Time: 0.0210 Steps: 29870, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000074, Sample Num: 1184, Cur Loss: 0.12882842, Cur Avg Loss: 0.29005689, Log Avg loss: 0.24407570, Global Avg Loss: 1.43455551, Time: 0.0210 Steps: 29880, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000084, Sample Num: 1344, Cur Loss: 0.12865113, Cur Avg Loss: 0.28301863, Log Avg loss: 0.23093548, Global Avg Loss: 1.43415283, Time: 0.0210 Steps: 29890, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000094, Sample Num: 1504, Cur Loss: 0.33779031, Cur Avg Loss: 0.28542458, Log Avg loss: 0.30563455, Global Avg Loss: 1.43377540, Time: 0.0210 Steps: 29900, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000104, Sample Num: 1664, Cur Loss: 0.16701633, Cur Avg Loss: 0.27870611, Log Avg loss: 0.21555249, Global Avg Loss: 1.43336810, Time: 0.0210 Steps: 29910, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000114, Sample Num: 1824, Cur Loss: 0.74495560, Cur Avg Loss: 0.28296692, Log Avg loss: 0.32727938, Global Avg Loss: 1.43299842, Time: 0.0210 Steps: 29920, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000124, Sample Num: 1984, Cur Loss: 0.15924230, Cur Avg Loss: 0.29216030, Log Avg loss: 0.39696481, Global Avg Loss: 1.43265227, Time: 0.0210 Steps: 29930, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000134, Sample Num: 2144, Cur Loss: 0.15822601, Cur Avg Loss: 0.29594311, Log Avg loss: 0.34284998, Global Avg Loss: 1.43228827, Time: 0.0209 Steps: 29940, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000144, Sample Num: 2304, Cur Loss: 0.10777457, Cur Avg Loss: 0.30257602, Log Avg loss: 0.39145695, Global Avg Loss: 1.43194075, Time: 0.0210 Steps: 29950, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000154, Sample Num: 2464, Cur Loss: 0.14410019, Cur Avg Loss: 0.29597024, Log Avg loss: 0.20084705, Global Avg Loss: 1.43152984, Time: 0.0209 Steps: 29960, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000164, Sample Num: 2624, Cur Loss: 0.60726249, Cur Avg Loss: 0.29605301, Log Avg loss: 0.29732766, Global Avg Loss: 1.43115139, Time: 0.0210 Steps: 29970, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000174, Sample Num: 2784, Cur Loss: 0.10124547, Cur Avg Loss: 0.29046966, Log Avg loss: 0.19890274, Global Avg Loss: 1.43074037, Time: 0.0210 Steps: 29980, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000184, Sample Num: 2944, Cur Loss: 0.23756924, Cur Avg Loss: 0.29014454, Log Avg loss: 0.28448751, Global Avg Loss: 1.43035815, Time: 0.0210 Steps: 29990, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000194, Sample Num: 3104, Cur Loss: 0.19034600, Cur Avg Loss: 0.29139870, Log Avg loss: 0.31447516, Global Avg Loss: 1.42998619, Time: 0.0210 Steps: 30000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000204, Sample Num: 3264, Cur Loss: 0.15778375, Cur Avg Loss: 0.28569117, Log Avg loss: 0.17496515, Global Avg Loss: 1.42956799, Time: 0.0210 Steps: 30010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000214, Sample Num: 3424, Cur Loss: 0.07816148, Cur Avg Loss: 0.28436861, Log Avg loss: 0.25738826, Global Avg Loss: 1.42917753, Time: 0.0210 Steps: 30020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000224, Sample Num: 3584, Cur Loss: 0.20211628, Cur Avg Loss: 0.27931982, Log Avg loss: 0.17127577, Global Avg Loss: 1.42875865, Time: 0.0210 Steps: 30030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000234, Sample Num: 3744, Cur Loss: 0.30599275, Cur Avg Loss: 0.27905084, Log Avg loss: 0.27302573, Global Avg Loss: 1.42837391, Time: 0.0210 Steps: 30040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000244, Sample Num: 3904, Cur Loss: 0.08450921, Cur Avg Loss: 0.27474684, Log Avg loss: 0.17403313, Global Avg Loss: 1.42795650, Time: 0.0210 Steps: 30050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000254, Sample Num: 4064, Cur Loss: 0.43951952, Cur Avg Loss: 0.27498090, Log Avg loss: 0.28069213, Global Avg Loss: 1.42757484, Time: 0.0210 Steps: 30060, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000264, Sample Num: 4224, Cur Loss: 0.19534123, Cur Avg Loss: 0.27476144, Log Avg loss: 0.26918719, Global Avg Loss: 1.42718961, Time: 0.0212 Steps: 30070, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000274, Sample Num: 4384, Cur Loss: 0.36427349, Cur Avg Loss: 0.27242882, Log Avg loss: 0.21084764, Global Avg Loss: 1.42678524, Time: 0.0208 Steps: 30080, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000284, Sample Num: 4544, Cur Loss: 0.28876978, Cur Avg Loss: 0.27058789, Log Avg loss: 0.22014634, Global Avg Loss: 1.42638423, Time: 0.0207 Steps: 30090, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000294, Sample Num: 4704, Cur Loss: 0.44288510, Cur Avg Loss: 0.27261921, Log Avg loss: 0.33030878, Global Avg Loss: 1.42602008, Time: 0.0209 Steps: 30100, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000304, Sample Num: 4864, Cur Loss: 0.42407009, Cur Avg Loss: 0.27285038, Log Avg loss: 0.27964682, Global Avg Loss: 1.42563936, Time: 0.0209 Steps: 30110, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000314, Sample Num: 5024, Cur Loss: 0.17412901, Cur Avg Loss: 0.27170188, Log Avg loss: 0.23678741, Global Avg Loss: 1.42524465, Time: 0.0210 Steps: 30120, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000324, Sample Num: 5184, Cur Loss: 0.32054406, Cur Avg Loss: 0.27330543, Log Avg loss: 0.32365686, Global Avg Loss: 1.42487904, Time: 0.0209 Steps: 30130, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000334, Sample Num: 5344, Cur Loss: 0.10293566, Cur Avg Loss: 0.27121185, Log Avg loss: 0.20337991, Global Avg Loss: 1.42447376, Time: 0.0209 Steps: 30140, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000344, Sample Num: 5504, Cur Loss: 0.09887528, Cur Avg Loss: 0.26921045, Log Avg loss: 0.20236359, Global Avg Loss: 1.42406842, Time: 0.0208 Steps: 30150, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000354, Sample Num: 5664, Cur Loss: 0.19636419, Cur Avg Loss: 0.26906426, Log Avg loss: 0.26403532, Global Avg Loss: 1.42368379, Time: 0.0209 Steps: 30160, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000364, Sample Num: 5824, Cur Loss: 0.79784715, Cur Avg Loss: 0.27042046, Log Avg loss: 0.31842994, Global Avg Loss: 1.42331745, Time: 0.0209 Steps: 30170, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000374, Sample Num: 5984, Cur Loss: 0.08863883, Cur Avg Loss: 0.26955326, Log Avg loss: 0.23798708, Global Avg Loss: 1.42292470, Time: 0.0209 Steps: 30180, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000384, Sample Num: 6144, Cur Loss: 0.34094599, Cur Avg Loss: 0.26998770, Log Avg loss: 0.28623593, Global Avg Loss: 1.42254819, Time: 0.0209 Steps: 30190, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000394, Sample Num: 6304, Cur Loss: 0.54808927, Cur Avg Loss: 0.27062321, Log Avg loss: 0.29502668, Global Avg Loss: 1.42217483, Time: 0.0210 Steps: 30200, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000404, Sample Num: 6464, Cur Loss: 0.15639223, Cur Avg Loss: 0.27220880, Log Avg loss: 0.33468129, Global Avg Loss: 1.42181486, Time: 0.0209 Steps: 30210, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000414, Sample Num: 6624, Cur Loss: 0.13121676, Cur Avg Loss: 0.27139431, Log Avg loss: 0.23848864, Global Avg Loss: 1.42142329, Time: 0.0209 Steps: 30220, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000424, Sample Num: 6784, Cur Loss: 0.12167139, Cur Avg Loss: 0.26888832, Log Avg loss: 0.16514060, Global Avg Loss: 1.42100771, Time: 0.0209 Steps: 30230, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000434, Sample Num: 6944, Cur Loss: 0.07086344, Cur Avg Loss: 0.26787455, Log Avg loss: 0.22489049, Global Avg Loss: 1.42061217, Time: 0.0209 Steps: 30240, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000444, Sample Num: 7104, Cur Loss: 0.40586194, Cur Avg Loss: 0.26781147, Log Avg loss: 0.26507395, Global Avg Loss: 1.42023017, Time: 0.0209 Steps: 30250, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000454, Sample Num: 7264, Cur Loss: 0.39889792, Cur Avg Loss: 0.26724256, Log Avg loss: 0.24198305, Global Avg Loss: 1.41984080, Time: 0.0210 Steps: 30260, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000464, Sample Num: 7424, Cur Loss: 0.21020277, Cur Avg Loss: 0.26717291, Log Avg loss: 0.26401076, Global Avg Loss: 1.41945896, Time: 0.0209 Steps: 30270, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000474, Sample Num: 7584, Cur Loss: 0.18372054, Cur Avg Loss: 0.26698827, Log Avg loss: 0.25842103, Global Avg Loss: 1.41907553, Time: 0.0209 Steps: 30280, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000484, Sample Num: 7744, Cur Loss: 0.82254887, Cur Avg Loss: 0.26977815, Log Avg loss: 0.40201833, Global Avg Loss: 1.41873975, Time: 0.0209 Steps: 30290, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000494, Sample Num: 7904, Cur Loss: 0.41467822, Cur Avg Loss: 0.27077534, Log Avg loss: 0.31903906, Global Avg Loss: 1.41837681, Time: 0.0209 Steps: 30300, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000504, Sample Num: 8064, Cur Loss: 0.21586013, Cur Avg Loss: 0.26962958, Log Avg loss: 0.21302920, Global Avg Loss: 1.41797914, Time: 0.0209 Steps: 30310, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000514, Sample Num: 8224, Cur Loss: 0.22362830, Cur Avg Loss: 0.26833442, Log Avg loss: 0.20305824, Global Avg Loss: 1.41757844, Time: 0.0246 Steps: 30320, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000524, Sample Num: 8384, Cur Loss: 0.23617613, Cur Avg Loss: 0.26866981, Log Avg loss: 0.28590902, Global Avg Loss: 1.41720532, Time: 0.0210 Steps: 30330, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000534, Sample Num: 8544, Cur Loss: 0.12313230, Cur Avg Loss: 0.26776730, Log Avg loss: 0.22047567, Global Avg Loss: 1.41681088, Time: 0.0210 Steps: 30340, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000544, Sample Num: 8704, Cur Loss: 0.30515885, Cur Avg Loss: 0.26680302, Log Avg loss: 0.21531031, Global Avg Loss: 1.41641500, Time: 0.0209 Steps: 30350, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000554, Sample Num: 8864, Cur Loss: 0.28085548, Cur Avg Loss: 0.26679197, Log Avg loss: 0.26619096, Global Avg Loss: 1.41603614, Time: 0.0209 Steps: 30360, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000564, Sample Num: 9024, Cur Loss: 0.14117691, Cur Avg Loss: 0.26498388, Log Avg loss: 0.16481560, Global Avg Loss: 1.41562415, Time: 0.0209 Steps: 30370, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000574, Sample Num: 9184, Cur Loss: 0.17396531, Cur Avg Loss: 0.26497173, Log Avg loss: 0.26428645, Global Avg Loss: 1.41524517, Time: 0.0209 Steps: 30380, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000584, Sample Num: 9344, Cur Loss: 0.24518594, Cur Avg Loss: 0.26726911, Log Avg loss: 0.39913878, Global Avg Loss: 1.41491081, Time: 0.0208 Steps: 30390, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000594, Sample Num: 9504, Cur Loss: 0.08322138, Cur Avg Loss: 0.26585987, Log Avg loss: 0.18356069, Global Avg Loss: 1.41450576, Time: 0.0209 Steps: 30400, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000604, Sample Num: 9664, Cur Loss: 0.10503015, Cur Avg Loss: 0.26614920, Log Avg loss: 0.28333545, Global Avg Loss: 1.41413379, Time: 0.0209 Steps: 30410, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000614, Sample Num: 9824, Cur Loss: 0.12202944, Cur Avg Loss: 0.26549541, Log Avg loss: 0.22600599, Global Avg Loss: 1.41374322, Time: 0.0211 Steps: 30420, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000624, Sample Num: 9984, Cur Loss: 0.26784575, Cur Avg Loss: 0.26629696, Log Avg loss: 0.31551264, Global Avg Loss: 1.41338231, Time: 0.0210 Steps: 30430, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000634, Sample Num: 10144, Cur Loss: 0.14454845, Cur Avg Loss: 0.26583486, Log Avg loss: 0.23699976, Global Avg Loss: 1.41299585, Time: 0.0210 Steps: 30440, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000644, Sample Num: 10304, Cur Loss: 0.20742561, Cur Avg Loss: 0.26455164, Log Avg loss: 0.18319554, Global Avg Loss: 1.41259198, Time: 0.0211 Steps: 30450, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000654, Sample Num: 10464, Cur Loss: 0.24080795, Cur Avg Loss: 0.26416573, Log Avg loss: 0.23931262, Global Avg Loss: 1.41220679, Time: 0.0209 Steps: 30460, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000664, Sample Num: 10624, Cur Loss: 0.08007527, Cur Avg Loss: 0.26622932, Log Avg loss: 0.40118813, Global Avg Loss: 1.41187498, Time: 0.0210 Steps: 30470, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000674, Sample Num: 10784, Cur Loss: 0.12342836, Cur Avg Loss: 0.26592101, Log Avg loss: 0.24544961, Global Avg Loss: 1.41149230, Time: 0.0209 Steps: 30480, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000684, Sample Num: 10944, Cur Loss: 0.10057403, Cur Avg Loss: 0.26501546, Log Avg loss: 0.20398143, Global Avg Loss: 1.41109626, Time: 0.0210 Steps: 30490, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000694, Sample Num: 11104, Cur Loss: 0.17248601, Cur Avg Loss: 0.26459772, Log Avg loss: 0.23602416, Global Avg Loss: 1.41071099, Time: 0.0210 Steps: 30500, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000704, Sample Num: 11264, Cur Loss: 0.57099617, Cur Avg Loss: 0.26566323, Log Avg loss: 0.33960969, Global Avg Loss: 1.41035993, Time: 0.0209 Steps: 30510, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000714, Sample Num: 11424, Cur Loss: 0.46619850, Cur Avg Loss: 0.26715805, Log Avg loss: 0.37239359, Global Avg Loss: 1.41001983, Time: 0.0210 Steps: 30520, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000724, Sample Num: 11584, Cur Loss: 0.13234898, Cur Avg Loss: 0.26638211, Log Avg loss: 0.21097971, Global Avg Loss: 1.40962709, Time: 0.0210 Steps: 30530, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000734, Sample Num: 11744, Cur Loss: 0.21233888, Cur Avg Loss: 0.26745214, Log Avg loss: 0.34492232, Global Avg Loss: 1.40927847, Time: 0.0209 Steps: 30540, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000744, Sample Num: 11904, Cur Loss: 0.65039480, Cur Avg Loss: 0.26749661, Log Avg loss: 0.27076037, Global Avg Loss: 1.40890579, Time: 0.0209 Steps: 30550, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000754, Sample Num: 12064, Cur Loss: 0.08835530, Cur Avg Loss: 0.26714094, Log Avg loss: 0.24067952, Global Avg Loss: 1.40852352, Time: 0.0211 Steps: 30560, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000764, Sample Num: 12224, Cur Loss: 0.23747435, Cur Avg Loss: 0.26659162, Log Avg loss: 0.22517253, Global Avg Loss: 1.40813642, Time: 0.0210 Steps: 30570, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000774, Sample Num: 12384, Cur Loss: 0.07113676, Cur Avg Loss: 0.26727209, Log Avg loss: 0.31925998, Global Avg Loss: 1.40778035, Time: 0.0210 Steps: 30580, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000784, Sample Num: 12544, Cur Loss: 0.20750022, Cur Avg Loss: 0.26747670, Log Avg loss: 0.28331337, Global Avg Loss: 1.40741276, Time: 0.0209 Steps: 30590, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000794, Sample Num: 12704, Cur Loss: 0.36297259, Cur Avg Loss: 0.26803858, Log Avg loss: 0.31209044, Global Avg Loss: 1.40705481, Time: 0.0209 Steps: 30600, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000804, Sample Num: 12864, Cur Loss: 0.09633709, Cur Avg Loss: 0.26712772, Log Avg loss: 0.19480494, Global Avg Loss: 1.40665878, Time: 0.0209 Steps: 30610, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000814, Sample Num: 13024, Cur Loss: 0.38739610, Cur Avg Loss: 0.26826424, Log Avg loss: 0.35964087, Global Avg Loss: 1.40631684, Time: 0.0209 Steps: 30620, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000824, Sample Num: 13184, Cur Loss: 0.06011573, Cur Avg Loss: 0.26813085, Log Avg loss: 0.25727267, Global Avg Loss: 1.40594170, Time: 0.0209 Steps: 30630, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000834, Sample Num: 13344, Cur Loss: 0.08609330, Cur Avg Loss: 0.26703212, Log Avg loss: 0.17649677, Global Avg Loss: 1.40554045, Time: 0.0208 Steps: 30640, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000844, Sample Num: 13504, Cur Loss: 0.39228022, Cur Avg Loss: 0.26835035, Log Avg loss: 0.37829107, Global Avg Loss: 1.40520529, Time: 0.0209 Steps: 30650, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000854, Sample Num: 13664, Cur Loss: 0.26609316, Cur Avg Loss: 0.26972264, Log Avg loss: 0.38554378, Global Avg Loss: 1.40487272, Time: 0.0208 Steps: 30660, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000864, Sample Num: 13824, Cur Loss: 0.30321908, Cur Avg Loss: 0.26943136, Log Avg loss: 0.24455584, Global Avg Loss: 1.40449440, Time: 0.0209 Steps: 30670, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000874, Sample Num: 13984, Cur Loss: 0.27738795, Cur Avg Loss: 0.26903817, Log Avg loss: 0.23506662, Global Avg Loss: 1.40411323, Time: 0.0209 Steps: 30680, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000884, Sample Num: 14144, Cur Loss: 0.24133450, Cur Avg Loss: 0.26908787, Log Avg loss: 0.27343197, Global Avg Loss: 1.40374481, Time: 0.0209 Steps: 30690, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000894, Sample Num: 14304, Cur Loss: 0.13196261, Cur Avg Loss: 0.26809329, Log Avg loss: 0.18017207, Global Avg Loss: 1.40334625, Time: 0.0209 Steps: 30700, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000904, Sample Num: 14464, Cur Loss: 0.34152788, Cur Avg Loss: 0.26784868, Log Avg loss: 0.24598042, Global Avg Loss: 1.40296938, Time: 0.0209 Steps: 30710, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000914, Sample Num: 14624, Cur Loss: 0.68358290, Cur Avg Loss: 0.26893932, Log Avg loss: 0.36753369, Global Avg Loss: 1.40263232, Time: 0.0209 Steps: 30720, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000924, Sample Num: 14784, Cur Loss: 0.10319175, Cur Avg Loss: 0.26789696, Log Avg loss: 0.17262511, Global Avg Loss: 1.40223206, Time: 0.0209 Steps: 30730, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000934, Sample Num: 14944, Cur Loss: 0.17017707, Cur Avg Loss: 0.26805583, Log Avg loss: 0.28273536, Global Avg Loss: 1.40186788, Time: 0.0209 Steps: 30740, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000944, Sample Num: 15104, Cur Loss: 0.14109077, Cur Avg Loss: 0.26699590, Log Avg loss: 0.16799839, Global Avg Loss: 1.40146662, Time: 0.0210 Steps: 30750, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000954, Sample Num: 15264, Cur Loss: 0.14533456, Cur Avg Loss: 0.26608171, Log Avg loss: 0.17978206, Global Avg Loss: 1.40106945, Time: 0.0208 Steps: 30760, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000964, Sample Num: 15424, Cur Loss: 0.36317682, Cur Avg Loss: 0.26597273, Log Avg loss: 0.25557595, Global Avg Loss: 1.40069718, Time: 0.0209 Steps: 30770, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000974, Sample Num: 15584, Cur Loss: 0.23568901, Cur Avg Loss: 0.26642790, Log Avg loss: 0.31030688, Global Avg Loss: 1.40034293, Time: 0.0209 Steps: 30780, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000984, Sample Num: 15744, Cur Loss: 0.15056914, Cur Avg Loss: 0.26604346, Log Avg loss: 0.22859818, Global Avg Loss: 1.39996237, Time: 0.0210 Steps: 30790, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000994, Sample Num: 15904, Cur Loss: 0.48553503, Cur Avg Loss: 0.26666934, Log Avg loss: 0.32825614, Global Avg Loss: 1.39961441, Time: 0.0209 Steps: 30800, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001004, Sample Num: 16064, Cur Loss: 0.37939206, Cur Avg Loss: 0.26622847, Log Avg loss: 0.22240576, Global Avg Loss: 1.39923232, Time: 0.0209 Steps: 30810, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001014, Sample Num: 16224, Cur Loss: 0.50463724, Cur Avg Loss: 0.26703333, Log Avg loss: 0.34784153, Global Avg Loss: 1.39889118, Time: 0.0209 Steps: 30820, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001024, Sample Num: 16384, Cur Loss: 0.08916812, Cur Avg Loss: 0.26890131, Log Avg loss: 0.45831462, Global Avg Loss: 1.39858610, Time: 0.0254 Steps: 30830, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001034, Sample Num: 16544, Cur Loss: 0.25574598, Cur Avg Loss: 0.26892477, Log Avg loss: 0.27132713, Global Avg Loss: 1.39822058, Time: 0.0209 Steps: 30840, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001044, Sample Num: 16704, Cur Loss: 0.43758541, Cur Avg Loss: 0.26899130, Log Avg loss: 0.27587067, Global Avg Loss: 1.39785677, Time: 0.0210 Steps: 30850, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001054, Sample Num: 16864, Cur Loss: 0.38021949, Cur Avg Loss: 0.26864948, Log Avg loss: 0.23296289, Global Avg Loss: 1.39747929, Time: 0.0209 Steps: 30860, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001064, Sample Num: 17024, Cur Loss: 0.16026331, Cur Avg Loss: 0.26847549, Log Avg loss: 0.25013699, Global Avg Loss: 1.39710763, Time: 0.0210 Steps: 30870, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001074, Sample Num: 17184, Cur Loss: 0.63168502, Cur Avg Loss: 0.26915523, Log Avg loss: 0.34147995, Global Avg Loss: 1.39676578, Time: 0.0209 Steps: 30880, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001084, Sample Num: 17344, Cur Loss: 0.59003240, Cur Avg Loss: 0.27069516, Log Avg loss: 0.43608331, Global Avg Loss: 1.39645478, Time: 0.0209 Steps: 30890, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001094, Sample Num: 17504, Cur Loss: 0.25683868, Cur Avg Loss: 0.27262747, Log Avg loss: 0.48208986, Global Avg Loss: 1.39615886, Time: 0.0209 Steps: 30900, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001104, Sample Num: 17664, Cur Loss: 0.39842629, Cur Avg Loss: 0.27254038, Log Avg loss: 0.26301355, Global Avg Loss: 1.39579227, Time: 0.0209 Steps: 30910, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001114, Sample Num: 17824, Cur Loss: 0.12705450, Cur Avg Loss: 0.27264324, Log Avg loss: 0.28399883, Global Avg Loss: 1.39543270, Time: 0.0209 Steps: 30920, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001124, Sample Num: 17984, Cur Loss: 0.15778846, Cur Avg Loss: 0.27244140, Log Avg loss: 0.24995643, Global Avg Loss: 1.39506235, Time: 0.0209 Steps: 30930, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001134, Sample Num: 18144, Cur Loss: 0.26345199, Cur Avg Loss: 0.27154081, Log Avg loss: 0.17031446, Global Avg Loss: 1.39466651, Time: 0.0209 Steps: 30940, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001144, Sample Num: 18304, Cur Loss: 0.25005925, Cur Avg Loss: 0.27170903, Log Avg loss: 0.29078479, Global Avg Loss: 1.39430984, Time: 0.0210 Steps: 30950, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001154, Sample Num: 18464, Cur Loss: 0.34068266, Cur Avg Loss: 0.27161677, Log Avg loss: 0.26106221, Global Avg Loss: 1.39394381, Time: 0.0210 Steps: 30960, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001164, Sample Num: 18624, Cur Loss: 0.31270000, Cur Avg Loss: 0.27139359, Log Avg loss: 0.24563902, Global Avg Loss: 1.39357303, Time: 0.0209 Steps: 30970, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001174, Sample Num: 18784, Cur Loss: 0.40678281, Cur Avg Loss: 0.27137226, Log Avg loss: 0.26888869, Global Avg Loss: 1.39320999, Time: 0.0209 Steps: 30980, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001184, Sample Num: 18944, Cur Loss: 0.23757783, Cur Avg Loss: 0.27076197, Log Avg loss: 0.19911425, Global Avg Loss: 1.39282467, Time: 0.0208 Steps: 30990, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001194, Sample Num: 19104, Cur Loss: 0.13375087, Cur Avg Loss: 0.27066215, Log Avg loss: 0.25884309, Global Avg Loss: 1.39245887, Time: 0.0209 Steps: 31000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001204, Sample Num: 19264, Cur Loss: 0.28687626, Cur Avg Loss: 0.27061619, Log Avg loss: 0.26512943, Global Avg Loss: 1.39209534, Time: 0.0209 Steps: 31010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001214, Sample Num: 19424, Cur Loss: 0.19193429, Cur Avg Loss: 0.27174156, Log Avg loss: 0.40723523, Global Avg Loss: 1.39177784, Time: 0.0209 Steps: 31020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001224, Sample Num: 19584, Cur Loss: 0.37177101, Cur Avg Loss: 0.27262143, Log Avg loss: 0.37943783, Global Avg Loss: 1.39145160, Time: 0.0210 Steps: 31030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001234, Sample Num: 19744, Cur Loss: 0.14587839, Cur Avg Loss: 0.27248676, Log Avg loss: 0.25600296, Global Avg Loss: 1.39108580, Time: 0.0209 Steps: 31040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001244, Sample Num: 19904, Cur Loss: 0.25258771, Cur Avg Loss: 0.27211760, Log Avg loss: 0.22656390, Global Avg Loss: 1.39071075, Time: 0.0210 Steps: 31050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001254, Sample Num: 20064, Cur Loss: 0.06404619, Cur Avg Loss: 0.27154416, Log Avg loss: 0.20020866, Global Avg Loss: 1.39032746, Time: 0.0210 Steps: 31060, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001264, Sample Num: 20224, Cur Loss: 0.21229856, Cur Avg Loss: 0.27125176, Log Avg loss: 0.23458485, Global Avg Loss: 1.38995548, Time: 0.0210 Steps: 31070, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001274, Sample Num: 20384, Cur Loss: 0.22619361, Cur Avg Loss: 0.27138651, Log Avg loss: 0.28841843, Global Avg Loss: 1.38960106, Time: 0.0210 Steps: 31080, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001284, Sample Num: 20544, Cur Loss: 0.66594231, Cur Avg Loss: 0.27154681, Log Avg loss: 0.29196842, Global Avg Loss: 1.38924801, Time: 0.0246 Steps: 31090, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001294, Sample Num: 20704, Cur Loss: 0.43949276, Cur Avg Loss: 0.27207310, Log Avg loss: 0.33964927, Global Avg Loss: 1.38891052, Time: 0.0208 Steps: 31100, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001304, Sample Num: 20864, Cur Loss: 0.04169682, Cur Avg Loss: 0.27179037, Log Avg loss: 0.23520449, Global Avg Loss: 1.38853967, Time: 0.0208 Steps: 31110, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001314, Sample Num: 21024, Cur Loss: 0.42110223, Cur Avg Loss: 0.27151057, Log Avg loss: 0.23502526, Global Avg Loss: 1.38816900, Time: 0.0209 Steps: 31120, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001324, Sample Num: 21184, Cur Loss: 0.32844627, Cur Avg Loss: 0.27119281, Log Avg loss: 0.22943849, Global Avg Loss: 1.38779678, Time: 0.0209 Steps: 31130, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001334, Sample Num: 21344, Cur Loss: 0.40099111, Cur Avg Loss: 0.27249157, Log Avg loss: 0.44444837, Global Avg Loss: 1.38749384, Time: 0.0209 Steps: 31140, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001344, Sample Num: 21504, Cur Loss: 0.17203419, Cur Avg Loss: 0.27433733, Log Avg loss: 0.52056157, Global Avg Loss: 1.38721553, Time: 0.0209 Steps: 31150, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001354, Sample Num: 21664, Cur Loss: 0.14312768, Cur Avg Loss: 0.27345279, Log Avg loss: 0.15457015, Global Avg Loss: 1.38681995, Time: 0.0208 Steps: 31160, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001364, Sample Num: 21824, Cur Loss: 0.56679875, Cur Avg Loss: 0.27401128, Log Avg loss: 0.34963069, Global Avg Loss: 1.38648719, Time: 0.0209 Steps: 31170, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001374, Sample Num: 21984, Cur Loss: 0.31027573, Cur Avg Loss: 0.27391279, Log Avg loss: 0.26047883, Global Avg Loss: 1.38612606, Time: 0.0210 Steps: 31180, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001384, Sample Num: 22144, Cur Loss: 0.15916967, Cur Avg Loss: 0.27405282, Log Avg loss: 0.29329300, Global Avg Loss: 1.38577568, Time: 0.0210 Steps: 31190, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001394, Sample Num: 22304, Cur Loss: 0.34210652, Cur Avg Loss: 0.27370889, Log Avg loss: 0.22610914, Global Avg Loss: 1.38540400, Time: 0.0210 Steps: 31200, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001404, Sample Num: 22464, Cur Loss: 0.06238718, Cur Avg Loss: 0.27369473, Log Avg loss: 0.27172108, Global Avg Loss: 1.38504716, Time: 0.0209 Steps: 31210, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001414, Sample Num: 22624, Cur Loss: 0.16613787, Cur Avg Loss: 0.27358597, Log Avg loss: 0.25831584, Global Avg Loss: 1.38468626, Time: 0.0210 Steps: 31220, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001424, Sample Num: 22784, Cur Loss: 0.12651266, Cur Avg Loss: 0.27370119, Log Avg loss: 0.28999298, Global Avg Loss: 1.38433573, Time: 0.0209 Steps: 31230, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001434, Sample Num: 22944, Cur Loss: 0.26613152, Cur Avg Loss: 0.27359886, Log Avg loss: 0.25902689, Global Avg Loss: 1.38397552, Time: 0.0209 Steps: 31240, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001444, Sample Num: 23104, Cur Loss: 0.65026367, Cur Avg Loss: 0.27397779, Log Avg loss: 0.32831741, Global Avg Loss: 1.38363771, Time: 0.0209 Steps: 31250, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001454, Sample Num: 23264, Cur Loss: 0.78862041, Cur Avg Loss: 0.27395346, Log Avg loss: 0.27043937, Global Avg Loss: 1.38328160, Time: 0.0210 Steps: 31260, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001464, Sample Num: 23424, Cur Loss: 0.42661867, Cur Avg Loss: 0.27396693, Log Avg loss: 0.27592642, Global Avg Loss: 1.38292747, Time: 0.0209 Steps: 31270, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001474, Sample Num: 23584, Cur Loss: 0.27026677, Cur Avg Loss: 0.27324240, Log Avg loss: 0.16717042, Global Avg Loss: 1.38253880, Time: 0.0209 Steps: 31280, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001484, Sample Num: 23744, Cur Loss: 0.09210236, Cur Avg Loss: 0.27274257, Log Avg loss: 0.19906798, Global Avg Loss: 1.38216058, Time: 0.0209 Steps: 31290, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001494, Sample Num: 23904, Cur Loss: 0.17868137, Cur Avg Loss: 0.27386967, Log Avg loss: 0.44113114, Global Avg Loss: 1.38185993, Time: 0.0210 Steps: 31300, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001504, Sample Num: 24064, Cur Loss: 0.11210020, Cur Avg Loss: 0.27379082, Log Avg loss: 0.26201100, Global Avg Loss: 1.38150226, Time: 0.0209 Steps: 31310, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001514, Sample Num: 24224, Cur Loss: 0.87751615, Cur Avg Loss: 0.27409050, Log Avg loss: 0.31916137, Global Avg Loss: 1.38116307, Time: 0.0210 Steps: 31320, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001524, Sample Num: 24384, Cur Loss: 0.15607207, Cur Avg Loss: 0.27319489, Log Avg loss: 0.13759955, Global Avg Loss: 1.38076615, Time: 0.0209 Steps: 31330, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001534, Sample Num: 24544, Cur Loss: 0.14573941, Cur Avg Loss: 0.27280104, Log Avg loss: 0.21277817, Global Avg Loss: 1.38039347, Time: 0.0210 Steps: 31340, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001544, Sample Num: 24704, Cur Loss: 0.28167939, Cur Avg Loss: 0.27264258, Log Avg loss: 0.24833544, Global Avg Loss: 1.38003236, Time: 0.0211 Steps: 31350, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001554, Sample Num: 24864, Cur Loss: 0.05878127, Cur Avg Loss: 0.27240194, Log Avg loss: 0.23524652, Global Avg Loss: 1.37966732, Time: 0.0209 Steps: 31360, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001564, Sample Num: 25024, Cur Loss: 0.26700822, Cur Avg Loss: 0.27227815, Log Avg loss: 0.25304118, Global Avg Loss: 1.37930818, Time: 0.0209 Steps: 31370, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001574, Sample Num: 25184, Cur Loss: 0.21795392, Cur Avg Loss: 0.27276716, Log Avg loss: 0.34924980, Global Avg Loss: 1.37897992, Time: 0.0210 Steps: 31380, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001584, Sample Num: 25344, Cur Loss: 0.20727789, Cur Avg Loss: 0.27231813, Log Avg loss: 0.20163938, Global Avg Loss: 1.37860485, Time: 0.0209 Steps: 31390, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001594, Sample Num: 25504, Cur Loss: 0.16681015, Cur Avg Loss: 0.27226980, Log Avg loss: 0.26461435, Global Avg Loss: 1.37825008, Time: 0.0210 Steps: 31400, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001604, Sample Num: 25664, Cur Loss: 0.09664164, Cur Avg Loss: 0.27231508, Log Avg loss: 0.27953322, Global Avg Loss: 1.37790028, Time: 0.0209 Steps: 31410, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001614, Sample Num: 25824, Cur Loss: 0.33615774, Cur Avg Loss: 0.27218762, Log Avg loss: 0.25174301, Global Avg Loss: 1.37754186, Time: 0.0209 Steps: 31420, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001624, Sample Num: 25984, Cur Loss: 0.30227947, Cur Avg Loss: 0.27227232, Log Avg loss: 0.28594249, Global Avg Loss: 1.37719455, Time: 0.0209 Steps: 31430, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001634, Sample Num: 26144, Cur Loss: 0.18053651, Cur Avg Loss: 0.27248199, Log Avg loss: 0.30653279, Global Avg Loss: 1.37685401, Time: 0.0209 Steps: 31440, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001644, Sample Num: 26304, Cur Loss: 0.26449385, Cur Avg Loss: 0.27282312, Log Avg loss: 0.32856342, Global Avg Loss: 1.37652069, Time: 0.0209 Steps: 31450, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001654, Sample Num: 26464, Cur Loss: 0.25623852, Cur Avg Loss: 0.27343472, Log Avg loss: 0.37398221, Global Avg Loss: 1.37620202, Time: 0.0209 Steps: 31460, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001664, Sample Num: 26624, Cur Loss: 0.11001184, Cur Avg Loss: 0.27346803, Log Avg loss: 0.27897689, Global Avg Loss: 1.37585336, Time: 0.0209 Steps: 31470, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001674, Sample Num: 26784, Cur Loss: 0.29967755, Cur Avg Loss: 0.27331491, Log Avg loss: 0.24783710, Global Avg Loss: 1.37549503, Time: 0.0209 Steps: 31480, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001684, Sample Num: 26944, Cur Loss: 0.11931128, Cur Avg Loss: 0.27319840, Log Avg loss: 0.25369392, Global Avg Loss: 1.37513879, Time: 0.0209 Steps: 31490, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001694, Sample Num: 27104, Cur Loss: 0.14102662, Cur Avg Loss: 0.27314757, Log Avg loss: 0.26458713, Global Avg Loss: 1.37478624, Time: 0.0209 Steps: 31500, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001704, Sample Num: 27264, Cur Loss: 0.31086281, Cur Avg Loss: 0.27289172, Log Avg loss: 0.22955212, Global Avg Loss: 1.37442279, Time: 0.0209 Steps: 31510, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001714, Sample Num: 27424, Cur Loss: 0.82274961, Cur Avg Loss: 0.27419821, Log Avg loss: 0.49682302, Global Avg Loss: 1.37414436, Time: 0.0209 Steps: 31520, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001724, Sample Num: 27584, Cur Loss: 0.22363481, Cur Avg Loss: 0.27522514, Log Avg loss: 0.45124207, Global Avg Loss: 1.37385165, Time: 0.0209 Steps: 31530, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001734, Sample Num: 27744, Cur Loss: 0.22389472, Cur Avg Loss: 0.27486829, Log Avg loss: 0.21334688, Global Avg Loss: 1.37348371, Time: 0.0209 Steps: 31540, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001744, Sample Num: 27904, Cur Loss: 0.43682498, Cur Avg Loss: 0.27510203, Log Avg loss: 0.31563209, Global Avg Loss: 1.37314841, Time: 0.0209 Steps: 31550, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001754, Sample Num: 28064, Cur Loss: 0.18256949, Cur Avg Loss: 0.27483627, Log Avg loss: 0.22848830, Global Avg Loss: 1.37278572, Time: 0.0209 Steps: 31560, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001764, Sample Num: 28224, Cur Loss: 0.03751558, Cur Avg Loss: 0.27457207, Log Avg loss: 0.22823067, Global Avg Loss: 1.37242317, Time: 0.0209 Steps: 31570, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001774, Sample Num: 28384, Cur Loss: 0.27667305, Cur Avg Loss: 0.27403544, Log Avg loss: 0.17937498, Global Avg Loss: 1.37204539, Time: 0.0210 Steps: 31580, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001784, Sample Num: 28544, Cur Loss: 0.39277211, Cur Avg Loss: 0.27371110, Log Avg loss: 0.21617301, Global Avg Loss: 1.37167949, Time: 0.0209 Steps: 31590, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001794, Sample Num: 28704, Cur Loss: 0.38028127, Cur Avg Loss: 0.27365032, Log Avg loss: 0.26280644, Global Avg Loss: 1.37132858, Time: 0.0247 Steps: 31600, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001804, Sample Num: 28864, Cur Loss: 0.12704280, Cur Avg Loss: 0.27385858, Log Avg loss: 0.31122053, Global Avg Loss: 1.37099321, Time: 0.0209 Steps: 31610, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001814, Sample Num: 29024, Cur Loss: 0.44981122, Cur Avg Loss: 0.27359978, Log Avg loss: 0.22691236, Global Avg Loss: 1.37063139, Time: 0.0209 Steps: 31620, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001824, Sample Num: 29184, Cur Loss: 0.41343796, Cur Avg Loss: 0.27375350, Log Avg loss: 0.30163881, Global Avg Loss: 1.37029342, Time: 0.0209 Steps: 31630, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001834, Sample Num: 29344, Cur Loss: 0.31663549, Cur Avg Loss: 0.27337334, Log Avg loss: 0.20403247, Global Avg Loss: 1.36992482, Time: 0.0209 Steps: 31640, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001844, Sample Num: 29504, Cur Loss: 0.14788765, Cur Avg Loss: 0.27301182, Log Avg loss: 0.20670910, Global Avg Loss: 1.36955729, Time: 0.0209 Steps: 31650, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001854, Sample Num: 29664, Cur Loss: 0.44083399, Cur Avg Loss: 0.27353891, Log Avg loss: 0.37073400, Global Avg Loss: 1.36924181, Time: 0.0209 Steps: 31660, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001864, Sample Num: 29824, Cur Loss: 0.06241030, Cur Avg Loss: 0.27328795, Log Avg loss: 0.22676029, Global Avg Loss: 1.36888106, Time: 0.0209 Steps: 31670, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001874, Sample Num: 29984, Cur Loss: 0.31739607, Cur Avg Loss: 0.27341974, Log Avg loss: 0.29798406, Global Avg Loss: 1.36854303, Time: 0.0209 Steps: 31680, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001884, Sample Num: 30144, Cur Loss: 0.21989810, Cur Avg Loss: 0.27300738, Log Avg loss: 0.19573238, Global Avg Loss: 1.36817294, Time: 0.0209 Steps: 31690, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001894, Sample Num: 30304, Cur Loss: 0.42697749, Cur Avg Loss: 0.27298579, Log Avg loss: 0.26891821, Global Avg Loss: 1.36782617, Time: 0.0209 Steps: 31700, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001904, Sample Num: 30464, Cur Loss: 0.60130572, Cur Avg Loss: 0.27278067, Log Avg loss: 0.23393054, Global Avg Loss: 1.36746859, Time: 0.0209 Steps: 31710, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001914, Sample Num: 30624, Cur Loss: 0.09847283, Cur Avg Loss: 0.27242967, Log Avg loss: 0.20559917, Global Avg Loss: 1.36710230, Time: 0.0209 Steps: 31720, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001924, Sample Num: 30784, Cur Loss: 0.15315221, Cur Avg Loss: 0.27233458, Log Avg loss: 0.25413462, Global Avg Loss: 1.36675154, Time: 0.0209 Steps: 31730, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001934, Sample Num: 30944, Cur Loss: 0.50465101, Cur Avg Loss: 0.27248259, Log Avg loss: 0.30095967, Global Avg Loss: 1.36641575, Time: 0.0209 Steps: 31740, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001944, Sample Num: 31104, Cur Loss: 0.45221308, Cur Avg Loss: 0.27316889, Log Avg loss: 0.40589873, Global Avg Loss: 1.36611322, Time: 0.0209 Steps: 31750, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001954, Sample Num: 31264, Cur Loss: 0.40751362, Cur Avg Loss: 0.27329738, Log Avg loss: 0.29827686, Global Avg Loss: 1.36577700, Time: 0.0209 Steps: 31760, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001964, Sample Num: 31424, Cur Loss: 0.14031266, Cur Avg Loss: 0.27357685, Log Avg loss: 0.32818451, Global Avg Loss: 1.36545041, Time: 0.0210 Steps: 31770, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001974, Sample Num: 31584, Cur Loss: 0.10697412, Cur Avg Loss: 0.27381460, Log Avg loss: 0.32050919, Global Avg Loss: 1.36512160, Time: 0.0210 Steps: 31780, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001984, Sample Num: 31744, Cur Loss: 0.30035025, Cur Avg Loss: 0.27426881, Log Avg loss: 0.36392977, Global Avg Loss: 1.36480666, Time: 0.0209 Steps: 31790, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001994, Sample Num: 31904, Cur Loss: 0.12188260, Cur Avg Loss: 0.27452751, Log Avg loss: 0.32585239, Global Avg Loss: 1.36447995, Time: 0.0209 Steps: 31800, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002004, Sample Num: 32064, Cur Loss: 0.06160811, Cur Avg Loss: 0.27458379, Log Avg loss: 0.28580654, Global Avg Loss: 1.36414085, Time: 0.0209 Steps: 31810, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002014, Sample Num: 32224, Cur Loss: 0.10040158, Cur Avg Loss: 0.27440093, Log Avg loss: 0.23775673, Global Avg Loss: 1.36378686, Time: 0.0209 Steps: 31820, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002024, Sample Num: 32384, Cur Loss: 0.25153574, Cur Avg Loss: 0.27419702, Log Avg loss: 0.23312973, Global Avg Loss: 1.36343164, Time: 0.0209 Steps: 31830, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002034, Sample Num: 32544, Cur Loss: 0.12819606, Cur Avg Loss: 0.27444037, Log Avg loss: 0.32369404, Global Avg Loss: 1.36310509, Time: 0.0209 Steps: 31840, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002044, Sample Num: 32704, Cur Loss: 0.20318764, Cur Avg Loss: 0.27416965, Log Avg loss: 0.21910411, Global Avg Loss: 1.36274591, Time: 0.0209 Steps: 31850, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002054, Sample Num: 32864, Cur Loss: 0.25730953, Cur Avg Loss: 0.27422582, Log Avg loss: 0.28570687, Global Avg Loss: 1.36240786, Time: 0.0210 Steps: 31860, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002064, Sample Num: 33024, Cur Loss: 0.29706076, Cur Avg Loss: 0.27388119, Log Avg loss: 0.20309549, Global Avg Loss: 1.36204409, Time: 0.0210 Steps: 31870, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002074, Sample Num: 33184, Cur Loss: 0.16088161, Cur Avg Loss: 0.27341237, Log Avg loss: 0.17664640, Global Avg Loss: 1.36167226, Time: 0.0210 Steps: 31880, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002084, Sample Num: 33344, Cur Loss: 0.22715753, Cur Avg Loss: 0.27297954, Log Avg loss: 0.18321176, Global Avg Loss: 1.36130272, Time: 0.0210 Steps: 31890, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002094, Sample Num: 33504, Cur Loss: 0.17444673, Cur Avg Loss: 0.27272615, Log Avg loss: 0.21991915, Global Avg Loss: 1.36094492, Time: 0.0210 Steps: 31900, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002104, Sample Num: 33664, Cur Loss: 0.09568240, Cur Avg Loss: 0.27256343, Log Avg loss: 0.23848924, Global Avg Loss: 1.36059317, Time: 0.0210 Steps: 31910, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002114, Sample Num: 33824, Cur Loss: 0.18747886, Cur Avg Loss: 0.27265311, Log Avg loss: 0.29152288, Global Avg Loss: 1.36025824, Time: 0.0209 Steps: 31920, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002124, Sample Num: 33984, Cur Loss: 0.07192214, Cur Avg Loss: 0.27273783, Log Avg loss: 0.29064684, Global Avg Loss: 1.35992326, Time: 0.0210 Steps: 31930, Updated lr: 0.000071 ***** Running evaluation checkpoint-31935 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-31935 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.726331, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.292239, "eval_total_loss": 205.443853, "eval_mae": 0.359258, "eval_mse": 0.29234, "eval_r2": 0.81417, "eval_sp_statistic": 0.898656, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.915769, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.262516, "test_total_loss": 131.782912, "test_mae": 0.314695, "test_mse": 0.262609, "test_r2": 0.83051, "test_sp_statistic": 0.894845, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.940315, "test_ps_pvalue": 0.0, "lr": 7.066382171645329e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.35975002295807, "train_cur_epoch_loss": 580.562511537224, "train_cur_epoch_avg_loss": 0.27269258409451574, "train_cur_epoch_time": 44.72633147239685, "train_cur_epoch_avg_time": 0.021008140663408573, "epoch": 15, "step": 31935} ################################################## Training, Epoch: 0016, Batch: 000005, Sample Num: 80, Cur Loss: 0.49529994, Cur Avg Loss: 0.32524729, Log Avg loss: 0.28936020, Global Avg Loss: 1.35958808, Time: 0.0247 Steps: 31940, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000015, Sample Num: 240, Cur Loss: 0.66558814, Cur Avg Loss: 0.35851047, Log Avg loss: 0.37514206, Global Avg Loss: 1.35927996, Time: 0.0209 Steps: 31950, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000025, Sample Num: 400, Cur Loss: 0.58313942, Cur Avg Loss: 0.30703486, Log Avg loss: 0.22982145, Global Avg Loss: 1.35892656, Time: 0.0210 Steps: 31960, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000035, Sample Num: 560, Cur Loss: 0.12089301, Cur Avg Loss: 0.28462845, Log Avg loss: 0.22861242, Global Avg Loss: 1.35857301, Time: 0.0210 Steps: 31970, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000045, Sample Num: 720, Cur Loss: 0.11192150, Cur Avg Loss: 0.27709798, Log Avg loss: 0.25074134, Global Avg Loss: 1.35822659, Time: 0.0210 Steps: 31980, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000055, Sample Num: 880, Cur Loss: 0.32857302, Cur Avg Loss: 0.26589503, Log Avg loss: 0.21548175, Global Avg Loss: 1.35786937, Time: 0.0210 Steps: 31990, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000065, Sample Num: 1040, Cur Loss: 0.21510565, Cur Avg Loss: 0.27520809, Log Avg loss: 0.32642994, Global Avg Loss: 1.35754705, Time: 0.0210 Steps: 32000, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000075, Sample Num: 1200, Cur Loss: 0.27754074, Cur Avg Loss: 0.27142065, Log Avg loss: 0.24680223, Global Avg Loss: 1.35720005, Time: 0.0210 Steps: 32010, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000085, Sample Num: 1360, Cur Loss: 0.05090010, Cur Avg Loss: 0.25958411, Log Avg loss: 0.17081012, Global Avg Loss: 1.35682953, Time: 0.0210 Steps: 32020, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000095, Sample Num: 1520, Cur Loss: 0.62267721, Cur Avg Loss: 0.26131523, Log Avg loss: 0.27602971, Global Avg Loss: 1.35649210, Time: 0.0211 Steps: 32030, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000105, Sample Num: 1680, Cur Loss: 0.45529181, Cur Avg Loss: 0.25530221, Log Avg loss: 0.19817858, Global Avg Loss: 1.35613058, Time: 0.0210 Steps: 32040, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000115, Sample Num: 1840, Cur Loss: 0.14879936, Cur Avg Loss: 0.24940178, Log Avg loss: 0.18744724, Global Avg Loss: 1.35576593, Time: 0.0210 Steps: 32050, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000125, Sample Num: 2000, Cur Loss: 0.36170554, Cur Avg Loss: 0.25142658, Log Avg loss: 0.27471178, Global Avg Loss: 1.35542874, Time: 0.0211 Steps: 32060, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000135, Sample Num: 2160, Cur Loss: 0.10053018, Cur Avg Loss: 0.25956709, Log Avg loss: 0.36132347, Global Avg Loss: 1.35511876, Time: 0.0210 Steps: 32070, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000145, Sample Num: 2320, Cur Loss: 0.41471103, Cur Avg Loss: 0.25621495, Log Avg loss: 0.21096101, Global Avg Loss: 1.35476210, Time: 0.0210 Steps: 32080, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000155, Sample Num: 2480, Cur Loss: 0.15152863, Cur Avg Loss: 0.25340245, Log Avg loss: 0.21262123, Global Avg Loss: 1.35440618, Time: 0.0211 Steps: 32090, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000165, Sample Num: 2640, Cur Loss: 0.45601645, Cur Avg Loss: 0.25219651, Log Avg loss: 0.23350442, Global Avg Loss: 1.35405699, Time: 0.0211 Steps: 32100, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000175, Sample Num: 2800, Cur Loss: 0.09030325, Cur Avg Loss: 0.25101431, Log Avg loss: 0.23150797, Global Avg Loss: 1.35370740, Time: 0.0210 Steps: 32110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000185, Sample Num: 2960, Cur Loss: 0.35091287, Cur Avg Loss: 0.26054743, Log Avg loss: 0.42737699, Global Avg Loss: 1.35341900, Time: 0.0210 Steps: 32120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000195, Sample Num: 3120, Cur Loss: 0.09729417, Cur Avg Loss: 0.25599198, Log Avg loss: 0.17171623, Global Avg Loss: 1.35305121, Time: 0.0210 Steps: 32130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000205, Sample Num: 3280, Cur Loss: 0.12703049, Cur Avg Loss: 0.25716757, Log Avg loss: 0.28009156, Global Avg Loss: 1.35271737, Time: 0.0210 Steps: 32140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000215, Sample Num: 3440, Cur Loss: 0.10310832, Cur Avg Loss: 0.26278308, Log Avg loss: 0.37790110, Global Avg Loss: 1.35241416, Time: 0.0211 Steps: 32150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000225, Sample Num: 3600, Cur Loss: 0.60176444, Cur Avg Loss: 0.26628858, Log Avg loss: 0.34165671, Global Avg Loss: 1.35209987, Time: 0.0210 Steps: 32160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000235, Sample Num: 3760, Cur Loss: 0.47974670, Cur Avg Loss: 0.26921054, Log Avg loss: 0.33495475, Global Avg Loss: 1.35178369, Time: 0.0211 Steps: 32170, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000245, Sample Num: 3920, Cur Loss: 0.12580171, Cur Avg Loss: 0.26527865, Log Avg loss: 0.17287928, Global Avg Loss: 1.35141735, Time: 0.0210 Steps: 32180, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000255, Sample Num: 4080, Cur Loss: 0.14824641, Cur Avg Loss: 0.26384171, Log Avg loss: 0.22863648, Global Avg Loss: 1.35106855, Time: 0.0210 Steps: 32190, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000265, Sample Num: 4240, Cur Loss: 0.22165786, Cur Avg Loss: 0.26407174, Log Avg loss: 0.26993774, Global Avg Loss: 1.35073279, Time: 0.0210 Steps: 32200, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000275, Sample Num: 4400, Cur Loss: 0.22666480, Cur Avg Loss: 0.26364721, Log Avg loss: 0.25239702, Global Avg Loss: 1.35039180, Time: 0.0210 Steps: 32210, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000285, Sample Num: 4560, Cur Loss: 0.41663936, Cur Avg Loss: 0.26325015, Log Avg loss: 0.25233112, Global Avg Loss: 1.35005100, Time: 0.0211 Steps: 32220, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000295, Sample Num: 4720, Cur Loss: 0.21716821, Cur Avg Loss: 0.26196360, Log Avg loss: 0.22529673, Global Avg Loss: 1.34970202, Time: 0.0209 Steps: 32230, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000305, Sample Num: 4880, Cur Loss: 0.34738919, Cur Avg Loss: 0.26007231, Log Avg loss: 0.20427940, Global Avg Loss: 1.34934674, Time: 0.0209 Steps: 32240, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000315, Sample Num: 5040, Cur Loss: 0.51402259, Cur Avg Loss: 0.26069055, Log Avg loss: 0.27954682, Global Avg Loss: 1.34901502, Time: 0.0210 Steps: 32250, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000325, Sample Num: 5200, Cur Loss: 0.14314258, Cur Avg Loss: 0.26160557, Log Avg loss: 0.29042873, Global Avg Loss: 1.34868688, Time: 0.0209 Steps: 32260, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000335, Sample Num: 5360, Cur Loss: 0.10160471, Cur Avg Loss: 0.26122511, Log Avg loss: 0.24886013, Global Avg Loss: 1.34834606, Time: 0.0210 Steps: 32270, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000345, Sample Num: 5520, Cur Loss: 0.49883178, Cur Avg Loss: 0.26306785, Log Avg loss: 0.32479973, Global Avg Loss: 1.34802898, Time: 0.0210 Steps: 32280, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000355, Sample Num: 5680, Cur Loss: 0.21173710, Cur Avg Loss: 0.26467040, Log Avg loss: 0.31995823, Global Avg Loss: 1.34771059, Time: 0.0211 Steps: 32290, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000365, Sample Num: 5840, Cur Loss: 0.16822532, Cur Avg Loss: 0.26384928, Log Avg loss: 0.23469964, Global Avg Loss: 1.34736601, Time: 0.0210 Steps: 32300, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000375, Sample Num: 6000, Cur Loss: 0.22860040, Cur Avg Loss: 0.26090318, Log Avg loss: 0.15337058, Global Avg Loss: 1.34699646, Time: 0.0209 Steps: 32310, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000385, Sample Num: 6160, Cur Loss: 0.15799659, Cur Avg Loss: 0.26212987, Log Avg loss: 0.30813061, Global Avg Loss: 1.34667503, Time: 0.0208 Steps: 32320, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000395, Sample Num: 6320, Cur Loss: 0.49417603, Cur Avg Loss: 0.26379377, Log Avg loss: 0.32785409, Global Avg Loss: 1.34635990, Time: 0.0209 Steps: 32330, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000405, Sample Num: 6480, Cur Loss: 0.18341303, Cur Avg Loss: 0.26285621, Log Avg loss: 0.22582237, Global Avg Loss: 1.34601341, Time: 0.0211 Steps: 32340, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000415, Sample Num: 6640, Cur Loss: 0.16112982, Cur Avg Loss: 0.26593751, Log Avg loss: 0.39073023, Global Avg Loss: 1.34571812, Time: 0.0208 Steps: 32350, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000425, Sample Num: 6800, Cur Loss: 0.60362720, Cur Avg Loss: 0.26554944, Log Avg loss: 0.24944467, Global Avg Loss: 1.34537934, Time: 0.0208 Steps: 32360, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000435, Sample Num: 6960, Cur Loss: 0.13517520, Cur Avg Loss: 0.26589714, Log Avg loss: 0.28067408, Global Avg Loss: 1.34505042, Time: 0.0208 Steps: 32370, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000445, Sample Num: 7120, Cur Loss: 0.17702515, Cur Avg Loss: 0.26523227, Log Avg loss: 0.23631043, Global Avg Loss: 1.34470801, Time: 0.0208 Steps: 32380, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000455, Sample Num: 7280, Cur Loss: 0.20589232, Cur Avg Loss: 0.26376135, Log Avg loss: 0.19830550, Global Avg Loss: 1.34435407, Time: 0.0208 Steps: 32390, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000465, Sample Num: 7440, Cur Loss: 0.18988574, Cur Avg Loss: 0.26311643, Log Avg loss: 0.23377253, Global Avg Loss: 1.34401130, Time: 0.0210 Steps: 32400, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000475, Sample Num: 7600, Cur Loss: 0.50926828, Cur Avg Loss: 0.26349704, Log Avg loss: 0.28119547, Global Avg Loss: 1.34368337, Time: 0.0208 Steps: 32410, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000485, Sample Num: 7760, Cur Loss: 0.72150171, Cur Avg Loss: 0.26678783, Log Avg loss: 0.42310064, Global Avg Loss: 1.34339942, Time: 0.0208 Steps: 32420, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000495, Sample Num: 7920, Cur Loss: 0.49418083, Cur Avg Loss: 0.26855419, Log Avg loss: 0.35422250, Global Avg Loss: 1.34309440, Time: 0.0210 Steps: 32430, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000505, Sample Num: 8080, Cur Loss: 0.31905502, Cur Avg Loss: 0.26758630, Log Avg loss: 0.21967588, Global Avg Loss: 1.34274809, Time: 0.0210 Steps: 32440, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000515, Sample Num: 8240, Cur Loss: 0.45282048, Cur Avg Loss: 0.26732015, Log Avg loss: 0.25387921, Global Avg Loss: 1.34241254, Time: 0.0246 Steps: 32450, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000525, Sample Num: 8400, Cur Loss: 0.09374326, Cur Avg Loss: 0.26684678, Log Avg loss: 0.24246865, Global Avg Loss: 1.34207368, Time: 0.0208 Steps: 32460, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000535, Sample Num: 8560, Cur Loss: 0.07728452, Cur Avg Loss: 0.26639808, Log Avg loss: 0.24284089, Global Avg Loss: 1.34173514, Time: 0.0208 Steps: 32470, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000545, Sample Num: 8720, Cur Loss: 0.13100752, Cur Avg Loss: 0.26567177, Log Avg loss: 0.22681417, Global Avg Loss: 1.34139187, Time: 0.0208 Steps: 32480, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000555, Sample Num: 8880, Cur Loss: 0.12905709, Cur Avg Loss: 0.26592968, Log Avg loss: 0.27998623, Global Avg Loss: 1.34106519, Time: 0.0208 Steps: 32490, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000565, Sample Num: 9040, Cur Loss: 0.06260869, Cur Avg Loss: 0.26459166, Log Avg loss: 0.19033116, Global Avg Loss: 1.34071112, Time: 0.0208 Steps: 32500, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000575, Sample Num: 9200, Cur Loss: 0.14176200, Cur Avg Loss: 0.26453323, Log Avg loss: 0.26123199, Global Avg Loss: 1.34037907, Time: 0.0208 Steps: 32510, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000585, Sample Num: 9360, Cur Loss: 0.17390208, Cur Avg Loss: 0.26516456, Log Avg loss: 0.30146584, Global Avg Loss: 1.34005960, Time: 0.0208 Steps: 32520, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000595, Sample Num: 9520, Cur Loss: 0.34618026, Cur Avg Loss: 0.26650313, Log Avg loss: 0.34480953, Global Avg Loss: 1.33975365, Time: 0.0208 Steps: 32530, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000605, Sample Num: 9680, Cur Loss: 0.04989062, Cur Avg Loss: 0.26639624, Log Avg loss: 0.26003636, Global Avg Loss: 1.33942184, Time: 0.0208 Steps: 32540, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000615, Sample Num: 9840, Cur Loss: 0.12533005, Cur Avg Loss: 0.26472163, Log Avg loss: 0.16340802, Global Avg Loss: 1.33906055, Time: 0.0208 Steps: 32550, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000625, Sample Num: 10000, Cur Loss: 0.18703915, Cur Avg Loss: 0.26382140, Log Avg loss: 0.20845683, Global Avg Loss: 1.33871331, Time: 0.0213 Steps: 32560, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000635, Sample Num: 10160, Cur Loss: 0.35842872, Cur Avg Loss: 0.26358037, Log Avg loss: 0.24851616, Global Avg Loss: 1.33837859, Time: 0.0210 Steps: 32570, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000645, Sample Num: 10320, Cur Loss: 0.18222916, Cur Avg Loss: 0.26370259, Log Avg loss: 0.27146372, Global Avg Loss: 1.33805111, Time: 0.0208 Steps: 32580, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000655, Sample Num: 10480, Cur Loss: 0.05761559, Cur Avg Loss: 0.26220665, Log Avg loss: 0.16571855, Global Avg Loss: 1.33769139, Time: 0.0209 Steps: 32590, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000665, Sample Num: 10640, Cur Loss: 0.23340675, Cur Avg Loss: 0.26264584, Log Avg loss: 0.29141250, Global Avg Loss: 1.33737044, Time: 0.0211 Steps: 32600, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000675, Sample Num: 10800, Cur Loss: 0.20626481, Cur Avg Loss: 0.26388392, Log Avg loss: 0.34621641, Global Avg Loss: 1.33706650, Time: 0.0208 Steps: 32610, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000685, Sample Num: 10960, Cur Loss: 0.36521712, Cur Avg Loss: 0.26517551, Log Avg loss: 0.35235753, Global Avg Loss: 1.33676463, Time: 0.0211 Steps: 32620, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000695, Sample Num: 11120, Cur Loss: 0.25368959, Cur Avg Loss: 0.26419776, Log Avg loss: 0.19722200, Global Avg Loss: 1.33641540, Time: 0.0209 Steps: 32630, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000705, Sample Num: 11280, Cur Loss: 0.21458176, Cur Avg Loss: 0.26383338, Log Avg loss: 0.23850906, Global Avg Loss: 1.33607903, Time: 0.0209 Steps: 32640, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000715, Sample Num: 11440, Cur Loss: 0.30299288, Cur Avg Loss: 0.26364117, Log Avg loss: 0.25009021, Global Avg Loss: 1.33574641, Time: 0.0208 Steps: 32650, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000725, Sample Num: 11600, Cur Loss: 0.14269443, Cur Avg Loss: 0.26397329, Log Avg loss: 0.28772050, Global Avg Loss: 1.33542552, Time: 0.0208 Steps: 32660, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000735, Sample Num: 11760, Cur Loss: 0.14266238, Cur Avg Loss: 0.26428527, Log Avg loss: 0.28690351, Global Avg Loss: 1.33510458, Time: 0.0209 Steps: 32670, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000745, Sample Num: 11920, Cur Loss: 0.22299403, Cur Avg Loss: 0.26537099, Log Avg loss: 0.34517108, Global Avg Loss: 1.33480166, Time: 0.0209 Steps: 32680, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000755, Sample Num: 12080, Cur Loss: 0.43388554, Cur Avg Loss: 0.26712674, Log Avg loss: 0.39793033, Global Avg Loss: 1.33451507, Time: 0.0209 Steps: 32690, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000765, Sample Num: 12240, Cur Loss: 0.15389083, Cur Avg Loss: 0.26977250, Log Avg loss: 0.46952760, Global Avg Loss: 1.33425055, Time: 0.0208 Steps: 32700, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000775, Sample Num: 12400, Cur Loss: 0.10454869, Cur Avg Loss: 0.26857228, Log Avg loss: 0.17675543, Global Avg Loss: 1.33389668, Time: 0.0209 Steps: 32710, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000785, Sample Num: 12560, Cur Loss: 0.26635447, Cur Avg Loss: 0.27054824, Log Avg loss: 0.42368482, Global Avg Loss: 1.33361850, Time: 0.0209 Steps: 32720, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000795, Sample Num: 12720, Cur Loss: 0.17182684, Cur Avg Loss: 0.27292634, Log Avg loss: 0.45960773, Global Avg Loss: 1.33335146, Time: 0.0209 Steps: 32730, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000805, Sample Num: 12880, Cur Loss: 0.19391817, Cur Avg Loss: 0.27404154, Log Avg loss: 0.36269933, Global Avg Loss: 1.33305499, Time: 0.0208 Steps: 32740, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000815, Sample Num: 13040, Cur Loss: 0.45146713, Cur Avg Loss: 0.27469966, Log Avg loss: 0.32767822, Global Avg Loss: 1.33274801, Time: 0.0208 Steps: 32750, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000825, Sample Num: 13200, Cur Loss: 0.67169893, Cur Avg Loss: 0.27805215, Log Avg loss: 0.55128026, Global Avg Loss: 1.33250946, Time: 0.0208 Steps: 32760, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000835, Sample Num: 13360, Cur Loss: 0.20718370, Cur Avg Loss: 0.28076394, Log Avg loss: 0.50448704, Global Avg Loss: 1.33225679, Time: 0.0209 Steps: 32770, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000845, Sample Num: 13520, Cur Loss: 0.10533421, Cur Avg Loss: 0.27925195, Log Avg loss: 0.15300063, Global Avg Loss: 1.33189704, Time: 0.0209 Steps: 32780, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000855, Sample Num: 13680, Cur Loss: 0.14267042, Cur Avg Loss: 0.28095401, Log Avg loss: 0.42477828, Global Avg Loss: 1.33162039, Time: 0.0208 Steps: 32790, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000865, Sample Num: 13840, Cur Loss: 0.25877500, Cur Avg Loss: 0.28043113, Log Avg loss: 0.23572497, Global Avg Loss: 1.33128628, Time: 0.0208 Steps: 32800, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000875, Sample Num: 14000, Cur Loss: 0.52412128, Cur Avg Loss: 0.28162181, Log Avg loss: 0.38461519, Global Avg Loss: 1.33099775, Time: 0.0209 Steps: 32810, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000885, Sample Num: 14160, Cur Loss: 0.08340316, Cur Avg Loss: 0.28086879, Log Avg loss: 0.21497936, Global Avg Loss: 1.33065770, Time: 0.0208 Steps: 32820, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000895, Sample Num: 14320, Cur Loss: 0.25921273, Cur Avg Loss: 0.28142637, Log Avg loss: 0.33077231, Global Avg Loss: 1.33035314, Time: 0.0209 Steps: 32830, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000905, Sample Num: 14480, Cur Loss: 0.46763486, Cur Avg Loss: 0.28180811, Log Avg loss: 0.31597359, Global Avg Loss: 1.33004425, Time: 0.0208 Steps: 32840, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000915, Sample Num: 14640, Cur Loss: 0.21243754, Cur Avg Loss: 0.28188126, Log Avg loss: 0.28850212, Global Avg Loss: 1.32972719, Time: 0.0209 Steps: 32850, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000925, Sample Num: 14800, Cur Loss: 0.62354946, Cur Avg Loss: 0.28176674, Log Avg loss: 0.27128734, Global Avg Loss: 1.32940509, Time: 0.0208 Steps: 32860, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000935, Sample Num: 14960, Cur Loss: 0.43933865, Cur Avg Loss: 0.28178778, Log Avg loss: 0.28373451, Global Avg Loss: 1.32908697, Time: 0.0209 Steps: 32870, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000945, Sample Num: 15120, Cur Loss: 0.11288844, Cur Avg Loss: 0.28133422, Log Avg loss: 0.23892636, Global Avg Loss: 1.32875541, Time: 0.0208 Steps: 32880, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000955, Sample Num: 15280, Cur Loss: 0.06893898, Cur Avg Loss: 0.28121027, Log Avg loss: 0.26949642, Global Avg Loss: 1.32843335, Time: 0.0208 Steps: 32890, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000965, Sample Num: 15440, Cur Loss: 0.40148860, Cur Avg Loss: 0.28024101, Log Avg loss: 0.18767752, Global Avg Loss: 1.32808661, Time: 0.0208 Steps: 32900, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000975, Sample Num: 15600, Cur Loss: 0.48862010, Cur Avg Loss: 0.28162858, Log Avg loss: 0.41552916, Global Avg Loss: 1.32780932, Time: 0.0208 Steps: 32910, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000985, Sample Num: 15760, Cur Loss: 0.43442446, Cur Avg Loss: 0.28341265, Log Avg loss: 0.45735879, Global Avg Loss: 1.32754491, Time: 0.0208 Steps: 32920, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000995, Sample Num: 15920, Cur Loss: 0.06578787, Cur Avg Loss: 0.28440729, Log Avg loss: 0.38237981, Global Avg Loss: 1.32725789, Time: 0.0208 Steps: 32930, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001005, Sample Num: 16080, Cur Loss: 0.11351727, Cur Avg Loss: 0.28379833, Log Avg loss: 0.22320624, Global Avg Loss: 1.32692272, Time: 0.0208 Steps: 32940, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001015, Sample Num: 16240, Cur Loss: 0.12404894, Cur Avg Loss: 0.28269139, Log Avg loss: 0.17144454, Global Avg Loss: 1.32657204, Time: 0.0208 Steps: 32950, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001025, Sample Num: 16400, Cur Loss: 0.07728574, Cur Avg Loss: 0.28192292, Log Avg loss: 0.20392260, Global Avg Loss: 1.32623143, Time: 0.0246 Steps: 32960, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001035, Sample Num: 16560, Cur Loss: 0.04984433, Cur Avg Loss: 0.28067339, Log Avg loss: 0.15259729, Global Avg Loss: 1.32587546, Time: 0.0209 Steps: 32970, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001045, Sample Num: 16720, Cur Loss: 0.52418947, Cur Avg Loss: 0.28043329, Log Avg loss: 0.25558305, Global Avg Loss: 1.32555093, Time: 0.0208 Steps: 32980, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001055, Sample Num: 16880, Cur Loss: 0.56583059, Cur Avg Loss: 0.28092296, Log Avg loss: 0.33209253, Global Avg Loss: 1.32524979, Time: 0.0209 Steps: 32990, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001065, Sample Num: 17040, Cur Loss: 0.15982646, Cur Avg Loss: 0.28032448, Log Avg loss: 0.21718556, Global Avg Loss: 1.32491402, Time: 0.0208 Steps: 33000, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001075, Sample Num: 17200, Cur Loss: 0.13631015, Cur Avg Loss: 0.28006538, Log Avg loss: 0.25247089, Global Avg Loss: 1.32458913, Time: 0.0208 Steps: 33010, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001085, Sample Num: 17360, Cur Loss: 0.67897940, Cur Avg Loss: 0.28026626, Log Avg loss: 0.30186050, Global Avg Loss: 1.32427940, Time: 0.0208 Steps: 33020, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001095, Sample Num: 17520, Cur Loss: 0.31435513, Cur Avg Loss: 0.27953891, Log Avg loss: 0.20062222, Global Avg Loss: 1.32393921, Time: 0.0208 Steps: 33030, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001105, Sample Num: 17680, Cur Loss: 0.10223597, Cur Avg Loss: 0.27916267, Log Avg loss: 0.23796432, Global Avg Loss: 1.32361052, Time: 0.0209 Steps: 33040, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001115, Sample Num: 17840, Cur Loss: 0.30551875, Cur Avg Loss: 0.27842776, Log Avg loss: 0.19722031, Global Avg Loss: 1.32326971, Time: 0.0208 Steps: 33050, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001125, Sample Num: 18000, Cur Loss: 0.18653460, Cur Avg Loss: 0.27820639, Log Avg loss: 0.25352277, Global Avg Loss: 1.32294613, Time: 0.0208 Steps: 33060, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001135, Sample Num: 18160, Cur Loss: 0.39248025, Cur Avg Loss: 0.27849295, Log Avg loss: 0.31073170, Global Avg Loss: 1.32264005, Time: 0.0208 Steps: 33070, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001145, Sample Num: 18320, Cur Loss: 0.36867636, Cur Avg Loss: 0.27802040, Log Avg loss: 0.22438530, Global Avg Loss: 1.32230805, Time: 0.0209 Steps: 33080, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001155, Sample Num: 18480, Cur Loss: 0.18969676, Cur Avg Loss: 0.27744503, Log Avg loss: 0.21156508, Global Avg Loss: 1.32197238, Time: 0.0210 Steps: 33090, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001165, Sample Num: 18640, Cur Loss: 0.52332741, Cur Avg Loss: 0.27707589, Log Avg loss: 0.23444072, Global Avg Loss: 1.32164382, Time: 0.0208 Steps: 33100, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001175, Sample Num: 18800, Cur Loss: 0.07117166, Cur Avg Loss: 0.27618091, Log Avg loss: 0.17191556, Global Avg Loss: 1.32129657, Time: 0.0208 Steps: 33110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001185, Sample Num: 18960, Cur Loss: 0.13001269, Cur Avg Loss: 0.27668289, Log Avg loss: 0.33566523, Global Avg Loss: 1.32099898, Time: 0.0208 Steps: 33120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001195, Sample Num: 19120, Cur Loss: 0.19800276, Cur Avg Loss: 0.27733590, Log Avg loss: 0.35471801, Global Avg Loss: 1.32070732, Time: 0.0209 Steps: 33130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001205, Sample Num: 19280, Cur Loss: 0.16784710, Cur Avg Loss: 0.27686160, Log Avg loss: 0.22018248, Global Avg Loss: 1.32037523, Time: 0.0209 Steps: 33140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001215, Sample Num: 19440, Cur Loss: 0.12568521, Cur Avg Loss: 0.27593565, Log Avg loss: 0.16435875, Global Avg Loss: 1.32002651, Time: 0.0209 Steps: 33150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001225, Sample Num: 19600, Cur Loss: 0.19843557, Cur Avg Loss: 0.27579496, Log Avg loss: 0.25870099, Global Avg Loss: 1.31970645, Time: 0.0209 Steps: 33160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001235, Sample Num: 19760, Cur Loss: 0.20360816, Cur Avg Loss: 0.27472226, Log Avg loss: 0.14331629, Global Avg Loss: 1.31935179, Time: 0.0209 Steps: 33170, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001245, Sample Num: 19920, Cur Loss: 0.28537133, Cur Avg Loss: 0.27388732, Log Avg loss: 0.17077278, Global Avg Loss: 1.31900563, Time: 0.0209 Steps: 33180, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001255, Sample Num: 20080, Cur Loss: 0.16216686, Cur Avg Loss: 0.27401853, Log Avg loss: 0.29035446, Global Avg Loss: 1.31869570, Time: 0.0209 Steps: 33190, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001265, Sample Num: 20240, Cur Loss: 0.37032098, Cur Avg Loss: 0.27497333, Log Avg loss: 0.39480028, Global Avg Loss: 1.31841742, Time: 0.0209 Steps: 33200, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001275, Sample Num: 20400, Cur Loss: 0.63866246, Cur Avg Loss: 0.27544680, Log Avg loss: 0.33534156, Global Avg Loss: 1.31812140, Time: 0.0209 Steps: 33210, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001285, Sample Num: 20560, Cur Loss: 0.07055534, Cur Avg Loss: 0.27517356, Log Avg loss: 0.24033467, Global Avg Loss: 1.31779696, Time: 0.0211 Steps: 33220, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001295, Sample Num: 20720, Cur Loss: 0.26028961, Cur Avg Loss: 0.27554843, Log Avg loss: 0.32371979, Global Avg Loss: 1.31749781, Time: 0.0209 Steps: 33230, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001305, Sample Num: 20880, Cur Loss: 0.72812319, Cur Avg Loss: 0.27604444, Log Avg loss: 0.34027792, Global Avg Loss: 1.31720382, Time: 0.0210 Steps: 33240, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001315, Sample Num: 21040, Cur Loss: 0.42713639, Cur Avg Loss: 0.27668677, Log Avg loss: 0.36050963, Global Avg Loss: 1.31691609, Time: 0.0209 Steps: 33250, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001325, Sample Num: 21200, Cur Loss: 0.06831270, Cur Avg Loss: 0.27649525, Log Avg loss: 0.25131062, Global Avg Loss: 1.31659571, Time: 0.0209 Steps: 33260, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001335, Sample Num: 21360, Cur Loss: 0.57556802, Cur Avg Loss: 0.27664249, Log Avg loss: 0.29615261, Global Avg Loss: 1.31628899, Time: 0.0209 Steps: 33270, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001345, Sample Num: 21520, Cur Loss: 0.10050286, Cur Avg Loss: 0.27689429, Log Avg loss: 0.31050922, Global Avg Loss: 1.31598677, Time: 0.0210 Steps: 33280, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001355, Sample Num: 21680, Cur Loss: 0.74850315, Cur Avg Loss: 0.27724855, Log Avg loss: 0.32489666, Global Avg Loss: 1.31568906, Time: 0.0210 Steps: 33290, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001365, Sample Num: 21840, Cur Loss: 0.23677202, Cur Avg Loss: 0.27770720, Log Avg loss: 0.33985343, Global Avg Loss: 1.31539602, Time: 0.0211 Steps: 33300, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001375, Sample Num: 22000, Cur Loss: 0.53585446, Cur Avg Loss: 0.27946584, Log Avg loss: 0.51952045, Global Avg Loss: 1.31515709, Time: 0.0210 Steps: 33310, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001385, Sample Num: 22160, Cur Loss: 0.31624997, Cur Avg Loss: 0.28007656, Log Avg loss: 0.36405034, Global Avg Loss: 1.31487164, Time: 0.0210 Steps: 33320, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001395, Sample Num: 22320, Cur Loss: 0.19936547, Cur Avg Loss: 0.27966492, Log Avg loss: 0.22265321, Global Avg Loss: 1.31454394, Time: 0.0210 Steps: 33330, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001405, Sample Num: 22480, Cur Loss: 0.26871341, Cur Avg Loss: 0.27935381, Log Avg loss: 0.23595473, Global Avg Loss: 1.31422043, Time: 0.0210 Steps: 33340, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001415, Sample Num: 22640, Cur Loss: 0.83573455, Cur Avg Loss: 0.27910754, Log Avg loss: 0.24450669, Global Avg Loss: 1.31389967, Time: 0.0210 Steps: 33350, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001425, Sample Num: 22800, Cur Loss: 0.75490487, Cur Avg Loss: 0.28056479, Log Avg loss: 0.48676500, Global Avg Loss: 1.31365173, Time: 0.0210 Steps: 33360, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001435, Sample Num: 22960, Cur Loss: 0.07657950, Cur Avg Loss: 0.28064737, Log Avg loss: 0.29241477, Global Avg Loss: 1.31334570, Time: 0.0210 Steps: 33370, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001445, Sample Num: 23120, Cur Loss: 0.19343916, Cur Avg Loss: 0.28076243, Log Avg loss: 0.29727407, Global Avg Loss: 1.31304130, Time: 0.0210 Steps: 33380, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001455, Sample Num: 23280, Cur Loss: 0.17128883, Cur Avg Loss: 0.28041803, Log Avg loss: 0.23065135, Global Avg Loss: 1.31271714, Time: 0.0210 Steps: 33390, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001465, Sample Num: 23440, Cur Loss: 0.15086734, Cur Avg Loss: 0.28020804, Log Avg loss: 0.24965517, Global Avg Loss: 1.31239886, Time: 0.0209 Steps: 33400, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001475, Sample Num: 23600, Cur Loss: 0.60230803, Cur Avg Loss: 0.28072939, Log Avg loss: 0.35710672, Global Avg Loss: 1.31211293, Time: 0.0210 Steps: 33410, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001485, Sample Num: 23760, Cur Loss: 0.72484106, Cur Avg Loss: 0.28089640, Log Avg loss: 0.30553023, Global Avg Loss: 1.31181173, Time: 0.0210 Steps: 33420, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001495, Sample Num: 23920, Cur Loss: 0.16397420, Cur Avg Loss: 0.28106999, Log Avg loss: 0.30684913, Global Avg Loss: 1.31151112, Time: 0.0210 Steps: 33430, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001505, Sample Num: 24080, Cur Loss: 0.18232369, Cur Avg Loss: 0.28035771, Log Avg loss: 0.17387109, Global Avg Loss: 1.31117091, Time: 0.0210 Steps: 33440, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001515, Sample Num: 24240, Cur Loss: 0.28451407, Cur Avg Loss: 0.28063396, Log Avg loss: 0.32220931, Global Avg Loss: 1.31087526, Time: 0.0209 Steps: 33450, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001525, Sample Num: 24400, Cur Loss: 0.20342912, Cur Avg Loss: 0.28025919, Log Avg loss: 0.22348197, Global Avg Loss: 1.31055028, Time: 0.0210 Steps: 33460, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001535, Sample Num: 24560, Cur Loss: 0.23643743, Cur Avg Loss: 0.28010686, Log Avg loss: 0.25687684, Global Avg Loss: 1.31023547, Time: 0.0210 Steps: 33470, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001545, Sample Num: 24720, Cur Loss: 0.12055829, Cur Avg Loss: 0.28006126, Log Avg loss: 0.27306122, Global Avg Loss: 1.30992568, Time: 0.0209 Steps: 33480, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001555, Sample Num: 24880, Cur Loss: 0.19052687, Cur Avg Loss: 0.28024058, Log Avg loss: 0.30794517, Global Avg Loss: 1.30962649, Time: 0.0209 Steps: 33490, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001565, Sample Num: 25040, Cur Loss: 0.14395845, Cur Avg Loss: 0.28066886, Log Avg loss: 0.34726776, Global Avg Loss: 1.30933922, Time: 0.0209 Steps: 33500, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001575, Sample Num: 25200, Cur Loss: 0.21020329, Cur Avg Loss: 0.28034121, Log Avg loss: 0.22906281, Global Avg Loss: 1.30901684, Time: 0.0209 Steps: 33510, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001585, Sample Num: 25360, Cur Loss: 0.13704485, Cur Avg Loss: 0.28013233, Log Avg loss: 0.24723374, Global Avg Loss: 1.30870008, Time: 0.0209 Steps: 33520, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001595, Sample Num: 25520, Cur Loss: 0.17112464, Cur Avg Loss: 0.27965064, Log Avg loss: 0.20330387, Global Avg Loss: 1.30837041, Time: 0.0209 Steps: 33530, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001605, Sample Num: 25680, Cur Loss: 0.32332939, Cur Avg Loss: 0.28054538, Log Avg loss: 0.42325504, Global Avg Loss: 1.30810651, Time: 0.0209 Steps: 33540, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001615, Sample Num: 25840, Cur Loss: 0.14355445, Cur Avg Loss: 0.28013347, Log Avg loss: 0.21402329, Global Avg Loss: 1.30778040, Time: 0.0209 Steps: 33550, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001625, Sample Num: 26000, Cur Loss: 0.29851997, Cur Avg Loss: 0.27965981, Log Avg loss: 0.20316379, Global Avg Loss: 1.30745126, Time: 0.0208 Steps: 33560, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001635, Sample Num: 26160, Cur Loss: 0.23672245, Cur Avg Loss: 0.27941766, Log Avg loss: 0.24006818, Global Avg Loss: 1.30713330, Time: 0.0209 Steps: 33570, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001645, Sample Num: 26320, Cur Loss: 0.93060291, Cur Avg Loss: 0.27953536, Log Avg loss: 0.29877943, Global Avg Loss: 1.30683302, Time: 0.0209 Steps: 33580, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001655, Sample Num: 26480, Cur Loss: 0.24715471, Cur Avg Loss: 0.27921426, Log Avg loss: 0.22639197, Global Avg Loss: 1.30651136, Time: 0.0208 Steps: 33590, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001665, Sample Num: 26640, Cur Loss: 0.57014358, Cur Avg Loss: 0.27908543, Log Avg loss: 0.25776465, Global Avg Loss: 1.30619923, Time: 0.0209 Steps: 33600, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001675, Sample Num: 26800, Cur Loss: 0.17927554, Cur Avg Loss: 0.27900166, Log Avg loss: 0.26505460, Global Avg Loss: 1.30588946, Time: 0.0209 Steps: 33610, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001685, Sample Num: 26960, Cur Loss: 0.08465487, Cur Avg Loss: 0.27845821, Log Avg loss: 0.18742903, Global Avg Loss: 1.30555678, Time: 0.0208 Steps: 33620, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001695, Sample Num: 27120, Cur Loss: 0.23908985, Cur Avg Loss: 0.27792491, Log Avg loss: 0.18806483, Global Avg Loss: 1.30522449, Time: 0.0208 Steps: 33630, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001705, Sample Num: 27280, Cur Loss: 0.23488009, Cur Avg Loss: 0.27771536, Log Avg loss: 0.24219624, Global Avg Loss: 1.30490849, Time: 0.0210 Steps: 33640, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001715, Sample Num: 27440, Cur Loss: 0.38735870, Cur Avg Loss: 0.27778034, Log Avg loss: 0.28886019, Global Avg Loss: 1.30460655, Time: 0.0209 Steps: 33650, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001725, Sample Num: 27600, Cur Loss: 0.14673564, Cur Avg Loss: 0.27767691, Log Avg loss: 0.25993814, Global Avg Loss: 1.30429619, Time: 0.0209 Steps: 33660, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001735, Sample Num: 27760, Cur Loss: 0.14652309, Cur Avg Loss: 0.27706654, Log Avg loss: 0.17177812, Global Avg Loss: 1.30395983, Time: 0.0209 Steps: 33670, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001745, Sample Num: 27920, Cur Loss: 0.26153764, Cur Avg Loss: 0.27681442, Log Avg loss: 0.23307038, Global Avg Loss: 1.30364187, Time: 0.0209 Steps: 33680, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001755, Sample Num: 28080, Cur Loss: 0.24917802, Cur Avg Loss: 0.27641756, Log Avg loss: 0.20716646, Global Avg Loss: 1.30331641, Time: 0.0209 Steps: 33690, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001765, Sample Num: 28240, Cur Loss: 0.25617737, Cur Avg Loss: 0.27602856, Log Avg loss: 0.20775937, Global Avg Loss: 1.30299132, Time: 0.0209 Steps: 33700, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001775, Sample Num: 28400, Cur Loss: 0.06784141, Cur Avg Loss: 0.27593317, Log Avg loss: 0.25909587, Global Avg Loss: 1.30268165, Time: 0.0208 Steps: 33710, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001785, Sample Num: 28560, Cur Loss: 0.46781880, Cur Avg Loss: 0.27634206, Log Avg loss: 0.34892087, Global Avg Loss: 1.30239880, Time: 0.0208 Steps: 33720, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001795, Sample Num: 28720, Cur Loss: 0.24435493, Cur Avg Loss: 0.27644322, Log Avg loss: 0.29449972, Global Avg Loss: 1.30209999, Time: 0.0245 Steps: 33730, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001805, Sample Num: 28880, Cur Loss: 0.36773640, Cur Avg Loss: 0.27602617, Log Avg loss: 0.20116601, Global Avg Loss: 1.30177369, Time: 0.0209 Steps: 33740, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001815, Sample Num: 29040, Cur Loss: 0.18543963, Cur Avg Loss: 0.27603871, Log Avg loss: 0.27830170, Global Avg Loss: 1.30147044, Time: 0.0208 Steps: 33750, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001825, Sample Num: 29200, Cur Loss: 0.36088714, Cur Avg Loss: 0.27600104, Log Avg loss: 0.26916349, Global Avg Loss: 1.30116466, Time: 0.0208 Steps: 33760, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001835, Sample Num: 29360, Cur Loss: 0.36412206, Cur Avg Loss: 0.27636537, Log Avg loss: 0.34285619, Global Avg Loss: 1.30088088, Time: 0.0208 Steps: 33770, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001845, Sample Num: 29520, Cur Loss: 0.13602827, Cur Avg Loss: 0.27618692, Log Avg loss: 0.24344139, Global Avg Loss: 1.30056785, Time: 0.0208 Steps: 33780, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001855, Sample Num: 29680, Cur Loss: 0.54686958, Cur Avg Loss: 0.27634326, Log Avg loss: 0.30518766, Global Avg Loss: 1.30027327, Time: 0.0208 Steps: 33790, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001865, Sample Num: 29840, Cur Loss: 0.20260707, Cur Avg Loss: 0.27662355, Log Avg loss: 0.32861701, Global Avg Loss: 1.29998580, Time: 0.0208 Steps: 33800, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001875, Sample Num: 30000, Cur Loss: 0.24613813, Cur Avg Loss: 0.27645120, Log Avg loss: 0.24430813, Global Avg Loss: 1.29967356, Time: 0.0208 Steps: 33810, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001885, Sample Num: 30160, Cur Loss: 0.06121476, Cur Avg Loss: 0.27678548, Log Avg loss: 0.33946299, Global Avg Loss: 1.29938964, Time: 0.0208 Steps: 33820, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001895, Sample Num: 30320, Cur Loss: 0.09108637, Cur Avg Loss: 0.27675650, Log Avg loss: 0.27129422, Global Avg Loss: 1.29908574, Time: 0.0209 Steps: 33830, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001905, Sample Num: 30480, Cur Loss: 0.29911971, Cur Avg Loss: 0.27676065, Log Avg loss: 0.27754756, Global Avg Loss: 1.29878387, Time: 0.0208 Steps: 33840, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001915, Sample Num: 30640, Cur Loss: 0.42531866, Cur Avg Loss: 0.27664095, Log Avg loss: 0.25383802, Global Avg Loss: 1.29847517, Time: 0.0208 Steps: 33850, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001925, Sample Num: 30800, Cur Loss: 0.48169714, Cur Avg Loss: 0.27652974, Log Avg loss: 0.25523351, Global Avg Loss: 1.29816706, Time: 0.0208 Steps: 33860, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001935, Sample Num: 30960, Cur Loss: 0.35087705, Cur Avg Loss: 0.27752841, Log Avg loss: 0.46977197, Global Avg Loss: 1.29792248, Time: 0.0208 Steps: 33870, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001945, Sample Num: 31120, Cur Loss: 0.21178612, Cur Avg Loss: 0.27769666, Log Avg loss: 0.31025355, Global Avg Loss: 1.29763096, Time: 0.0208 Steps: 33880, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001955, Sample Num: 31280, Cur Loss: 0.54615510, Cur Avg Loss: 0.27744524, Log Avg loss: 0.22854222, Global Avg Loss: 1.29731550, Time: 0.0208 Steps: 33890, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001965, Sample Num: 31440, Cur Loss: 0.36236590, Cur Avg Loss: 0.27747655, Log Avg loss: 0.28359877, Global Avg Loss: 1.29701647, Time: 0.0208 Steps: 33900, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001975, Sample Num: 31600, Cur Loss: 0.23369473, Cur Avg Loss: 0.27711961, Log Avg loss: 0.20698048, Global Avg Loss: 1.29669502, Time: 0.0208 Steps: 33910, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001985, Sample Num: 31760, Cur Loss: 0.24191986, Cur Avg Loss: 0.27674591, Log Avg loss: 0.20294062, Global Avg Loss: 1.29637257, Time: 0.0208 Steps: 33920, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001995, Sample Num: 31920, Cur Loss: 0.24522470, Cur Avg Loss: 0.27634101, Log Avg loss: 0.19596778, Global Avg Loss: 1.29604826, Time: 0.0208 Steps: 33930, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002005, Sample Num: 32080, Cur Loss: 0.21818903, Cur Avg Loss: 0.27643038, Log Avg loss: 0.29425975, Global Avg Loss: 1.29575309, Time: 0.0208 Steps: 33940, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002015, Sample Num: 32240, Cur Loss: 0.20605372, Cur Avg Loss: 0.27622196, Log Avg loss: 0.23443439, Global Avg Loss: 1.29544048, Time: 0.0208 Steps: 33950, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002025, Sample Num: 32400, Cur Loss: 0.36261255, Cur Avg Loss: 0.27627652, Log Avg loss: 0.28727079, Global Avg Loss: 1.29514361, Time: 0.0208 Steps: 33960, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002035, Sample Num: 32560, Cur Loss: 0.22205806, Cur Avg Loss: 0.27634509, Log Avg loss: 0.29022893, Global Avg Loss: 1.29484778, Time: 0.0208 Steps: 33970, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002045, Sample Num: 32720, Cur Loss: 0.23642115, Cur Avg Loss: 0.27611520, Log Avg loss: 0.22933371, Global Avg Loss: 1.29453421, Time: 0.0208 Steps: 33980, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002055, Sample Num: 32880, Cur Loss: 0.22132012, Cur Avg Loss: 0.27604013, Log Avg loss: 0.26068855, Global Avg Loss: 1.29423005, Time: 0.0212 Steps: 33990, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002065, Sample Num: 33040, Cur Loss: 0.22529885, Cur Avg Loss: 0.27554563, Log Avg loss: 0.17392562, Global Avg Loss: 1.29390055, Time: 0.0209 Steps: 34000, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002075, Sample Num: 33200, Cur Loss: 0.12175854, Cur Avg Loss: 0.27568586, Log Avg loss: 0.30464355, Global Avg Loss: 1.29360968, Time: 0.0209 Steps: 34010, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002085, Sample Num: 33360, Cur Loss: 0.22654635, Cur Avg Loss: 0.27522751, Log Avg loss: 0.18011938, Global Avg Loss: 1.29328237, Time: 0.0210 Steps: 34020, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002095, Sample Num: 33520, Cur Loss: 0.21579251, Cur Avg Loss: 0.27477300, Log Avg loss: 0.18000743, Global Avg Loss: 1.29295523, Time: 0.0209 Steps: 34030, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002105, Sample Num: 33680, Cur Loss: 0.22670722, Cur Avg Loss: 0.27472294, Log Avg loss: 0.26423520, Global Avg Loss: 1.29265302, Time: 0.0209 Steps: 34040, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002115, Sample Num: 33840, Cur Loss: 0.25299907, Cur Avg Loss: 0.27437379, Log Avg loss: 0.20087871, Global Avg Loss: 1.29233238, Time: 0.0209 Steps: 34050, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002125, Sample Num: 34000, Cur Loss: 0.36827898, Cur Avg Loss: 0.27389110, Log Avg loss: 0.17180087, Global Avg Loss: 1.29200339, Time: 0.0209 Steps: 34060, Updated lr: 0.000069 ***** Running evaluation checkpoint-34064 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-34064 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.656107, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.286695, "eval_total_loss": 201.546691, "eval_mae": 0.411311, "eval_mse": 0.286744, "eval_r2": 0.817727, "eval_sp_statistic": 0.900558, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.917679, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.254564, "test_total_loss": 127.791223, "test_mae": 0.359563, "test_mse": 0.254654, "test_r2": 0.835644, "test_sp_statistic": 0.894429, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.936152, "test_ps_pvalue": 0.0, "lr": 6.86448553816975e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.291894354285817, "train_cur_epoch_loss": 583.4723012261093, "train_cur_epoch_avg_loss": 0.27405932420202406, "train_cur_epoch_time": 44.65610694885254, "train_cur_epoch_avg_time": 0.02097515591773252, "epoch": 16, "step": 34064} ################################################## Training, Epoch: 0017, Batch: 000006, Sample Num: 96, Cur Loss: 0.23110035, Cur Avg Loss: 0.24319019, Log Avg loss: 0.29128616, Global Avg Loss: 1.29170967, Time: 0.0247 Steps: 34070, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000016, Sample Num: 256, Cur Loss: 0.14704990, Cur Avg Loss: 0.18899493, Log Avg loss: 0.15647778, Global Avg Loss: 1.29137656, Time: 0.0209 Steps: 34080, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000026, Sample Num: 416, Cur Loss: 0.14004003, Cur Avg Loss: 0.20833225, Log Avg loss: 0.23927197, Global Avg Loss: 1.29106794, Time: 0.0210 Steps: 34090, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000036, Sample Num: 576, Cur Loss: 0.07952629, Cur Avg Loss: 0.20043299, Log Avg loss: 0.17989491, Global Avg Loss: 1.29074208, Time: 0.0208 Steps: 34100, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000046, Sample Num: 736, Cur Loss: 0.17031175, Cur Avg Loss: 0.19136637, Log Avg loss: 0.15872652, Global Avg Loss: 1.29041021, Time: 0.0209 Steps: 34110, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000056, Sample Num: 896, Cur Loss: 0.28329518, Cur Avg Loss: 0.21001352, Log Avg loss: 0.29579044, Global Avg Loss: 1.29011870, Time: 0.0210 Steps: 34120, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000066, Sample Num: 1056, Cur Loss: 0.36245185, Cur Avg Loss: 0.21493936, Log Avg loss: 0.24252403, Global Avg Loss: 1.28981176, Time: 0.0209 Steps: 34130, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000076, Sample Num: 1216, Cur Loss: 0.29015592, Cur Avg Loss: 0.21888456, Log Avg loss: 0.24492288, Global Avg Loss: 1.28950570, Time: 0.0209 Steps: 34140, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000086, Sample Num: 1376, Cur Loss: 0.14068210, Cur Avg Loss: 0.22559384, Log Avg loss: 0.27658437, Global Avg Loss: 1.28920909, Time: 0.0210 Steps: 34150, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000096, Sample Num: 1536, Cur Loss: 0.42392248, Cur Avg Loss: 0.22626792, Log Avg loss: 0.23206506, Global Avg Loss: 1.28889962, Time: 0.0209 Steps: 34160, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000106, Sample Num: 1696, Cur Loss: 0.06595203, Cur Avg Loss: 0.22584247, Log Avg loss: 0.22175817, Global Avg Loss: 1.28858732, Time: 0.0210 Steps: 34170, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000116, Sample Num: 1856, Cur Loss: 0.14542669, Cur Avg Loss: 0.23067709, Log Avg loss: 0.28192397, Global Avg Loss: 1.28829280, Time: 0.0209 Steps: 34180, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000126, Sample Num: 2016, Cur Loss: 0.04771139, Cur Avg Loss: 0.25392370, Log Avg loss: 0.52358438, Global Avg Loss: 1.28806913, Time: 0.0209 Steps: 34190, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000136, Sample Num: 2176, Cur Loss: 0.49859369, Cur Avg Loss: 0.25061538, Log Avg loss: 0.20893056, Global Avg Loss: 1.28775360, Time: 0.0209 Steps: 34200, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000146, Sample Num: 2336, Cur Loss: 0.07568024, Cur Avg Loss: 0.24410792, Log Avg loss: 0.15560652, Global Avg Loss: 1.28742266, Time: 0.0210 Steps: 34210, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000156, Sample Num: 2496, Cur Loss: 0.14165941, Cur Avg Loss: 0.23844435, Log Avg loss: 0.15575622, Global Avg Loss: 1.28709195, Time: 0.0209 Steps: 34220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000166, Sample Num: 2656, Cur Loss: 0.30610210, Cur Avg Loss: 0.23754497, Log Avg loss: 0.22351467, Global Avg Loss: 1.28678124, Time: 0.0210 Steps: 34230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000176, Sample Num: 2816, Cur Loss: 0.22610623, Cur Avg Loss: 0.23731352, Log Avg loss: 0.23347147, Global Avg Loss: 1.28647361, Time: 0.0208 Steps: 34240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000186, Sample Num: 2976, Cur Loss: 0.59806377, Cur Avg Loss: 0.23837396, Log Avg loss: 0.25703771, Global Avg Loss: 1.28617305, Time: 0.0210 Steps: 34250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000196, Sample Num: 3136, Cur Loss: 0.51133597, Cur Avg Loss: 0.24007041, Log Avg loss: 0.27162440, Global Avg Loss: 1.28587691, Time: 0.0208 Steps: 34260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000206, Sample Num: 3296, Cur Loss: 0.38992181, Cur Avg Loss: 0.24539434, Log Avg loss: 0.34974337, Global Avg Loss: 1.28560375, Time: 0.0209 Steps: 34270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000216, Sample Num: 3456, Cur Loss: 0.35087618, Cur Avg Loss: 0.24804822, Log Avg loss: 0.30271797, Global Avg Loss: 1.28531703, Time: 0.0209 Steps: 34280, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000226, Sample Num: 3616, Cur Loss: 0.25529605, Cur Avg Loss: 0.24813334, Log Avg loss: 0.24997204, Global Avg Loss: 1.28501509, Time: 0.0209 Steps: 34290, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000236, Sample Num: 3776, Cur Loss: 0.35548261, Cur Avg Loss: 0.24753788, Log Avg loss: 0.23408046, Global Avg Loss: 1.28470869, Time: 0.0209 Steps: 34300, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000246, Sample Num: 3936, Cur Loss: 0.16863972, Cur Avg Loss: 0.25205099, Log Avg loss: 0.35856049, Global Avg Loss: 1.28443876, Time: 0.0209 Steps: 34310, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000256, Sample Num: 4096, Cur Loss: 0.41835374, Cur Avg Loss: 0.25239869, Log Avg loss: 0.26095195, Global Avg Loss: 1.28414054, Time: 0.0254 Steps: 34320, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000266, Sample Num: 4256, Cur Loss: 0.16105598, Cur Avg Loss: 0.25184408, Log Avg loss: 0.23764622, Global Avg Loss: 1.28383571, Time: 0.0209 Steps: 34330, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000276, Sample Num: 4416, Cur Loss: 0.14588638, Cur Avg Loss: 0.25293320, Log Avg loss: 0.28190381, Global Avg Loss: 1.28354394, Time: 0.0208 Steps: 34340, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000286, Sample Num: 4576, Cur Loss: 0.17587236, Cur Avg Loss: 0.25444069, Log Avg loss: 0.29604719, Global Avg Loss: 1.28325646, Time: 0.0208 Steps: 34350, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000296, Sample Num: 4736, Cur Loss: 0.18891710, Cur Avg Loss: 0.25266692, Log Avg loss: 0.20193723, Global Avg Loss: 1.28294175, Time: 0.0208 Steps: 34360, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000306, Sample Num: 4896, Cur Loss: 0.33489752, Cur Avg Loss: 0.24972043, Log Avg loss: 0.16250433, Global Avg Loss: 1.28261576, Time: 0.0208 Steps: 34370, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000316, Sample Num: 5056, Cur Loss: 0.06871571, Cur Avg Loss: 0.24866490, Log Avg loss: 0.21636577, Global Avg Loss: 1.28230563, Time: 0.0209 Steps: 34380, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000326, Sample Num: 5216, Cur Loss: 0.12276482, Cur Avg Loss: 0.24672964, Log Avg loss: 0.18557524, Global Avg Loss: 1.28198672, Time: 0.0210 Steps: 34390, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000336, Sample Num: 5376, Cur Loss: 0.17045316, Cur Avg Loss: 0.24646334, Log Avg loss: 0.23778208, Global Avg Loss: 1.28168317, Time: 0.0210 Steps: 34400, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000346, Sample Num: 5536, Cur Loss: 0.13166758, Cur Avg Loss: 0.24425004, Log Avg loss: 0.16988312, Global Avg Loss: 1.28136006, Time: 0.0207 Steps: 34410, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000356, Sample Num: 5696, Cur Loss: 0.14532998, Cur Avg Loss: 0.24570509, Log Avg loss: 0.29604979, Global Avg Loss: 1.28107380, Time: 0.0209 Steps: 34420, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000366, Sample Num: 5856, Cur Loss: 0.30843365, Cur Avg Loss: 0.24585959, Log Avg loss: 0.25135997, Global Avg Loss: 1.28077473, Time: 0.0209 Steps: 34430, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000376, Sample Num: 6016, Cur Loss: 0.25566235, Cur Avg Loss: 0.24785762, Log Avg loss: 0.32098547, Global Avg Loss: 1.28049604, Time: 0.0210 Steps: 34440, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000386, Sample Num: 6176, Cur Loss: 0.42114189, Cur Avg Loss: 0.25054284, Log Avg loss: 0.35150688, Global Avg Loss: 1.28022638, Time: 0.0209 Steps: 34450, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000396, Sample Num: 6336, Cur Loss: 0.10604934, Cur Avg Loss: 0.25081615, Log Avg loss: 0.26136600, Global Avg Loss: 1.27993072, Time: 0.0210 Steps: 34460, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000406, Sample Num: 6496, Cur Loss: 0.14956567, Cur Avg Loss: 0.25115209, Log Avg loss: 0.26445535, Global Avg Loss: 1.27963612, Time: 0.0209 Steps: 34470, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000416, Sample Num: 6656, Cur Loss: 0.29341066, Cur Avg Loss: 0.25098263, Log Avg loss: 0.24410258, Global Avg Loss: 1.27933579, Time: 0.0210 Steps: 34480, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000426, Sample Num: 6816, Cur Loss: 0.12339005, Cur Avg Loss: 0.25182803, Log Avg loss: 0.28699671, Global Avg Loss: 1.27904807, Time: 0.0210 Steps: 34490, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000436, Sample Num: 6976, Cur Loss: 0.66169363, Cur Avg Loss: 0.25352306, Log Avg loss: 0.32573141, Global Avg Loss: 1.27877175, Time: 0.0208 Steps: 34500, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000446, Sample Num: 7136, Cur Loss: 0.24926287, Cur Avg Loss: 0.25768429, Log Avg loss: 0.43911361, Global Avg Loss: 1.27852844, Time: 0.0208 Steps: 34510, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000456, Sample Num: 7296, Cur Loss: 0.48718530, Cur Avg Loss: 0.25868469, Log Avg loss: 0.30330280, Global Avg Loss: 1.27824593, Time: 0.0209 Steps: 34520, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000466, Sample Num: 7456, Cur Loss: 0.08580584, Cur Avg Loss: 0.25887519, Log Avg loss: 0.26756208, Global Avg Loss: 1.27795323, Time: 0.0210 Steps: 34530, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000476, Sample Num: 7616, Cur Loss: 0.47892615, Cur Avg Loss: 0.25845948, Log Avg loss: 0.23908706, Global Avg Loss: 1.27765246, Time: 0.0209 Steps: 34540, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000486, Sample Num: 7776, Cur Loss: 0.29443416, Cur Avg Loss: 0.25867013, Log Avg loss: 0.26869709, Global Avg Loss: 1.27736043, Time: 0.0209 Steps: 34550, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000496, Sample Num: 7936, Cur Loss: 0.10097501, Cur Avg Loss: 0.25681435, Log Avg loss: 0.16662371, Global Avg Loss: 1.27703904, Time: 0.0210 Steps: 34560, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000506, Sample Num: 8096, Cur Loss: 0.29648501, Cur Avg Loss: 0.25652614, Log Avg loss: 0.24223060, Global Avg Loss: 1.27673970, Time: 0.0209 Steps: 34570, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000516, Sample Num: 8256, Cur Loss: 0.19281501, Cur Avg Loss: 0.25713687, Log Avg loss: 0.28803984, Global Avg Loss: 1.27645379, Time: 0.0248 Steps: 34580, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000526, Sample Num: 8416, Cur Loss: 0.23615870, Cur Avg Loss: 0.25748906, Log Avg loss: 0.27566195, Global Avg Loss: 1.27616446, Time: 0.0211 Steps: 34590, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000536, Sample Num: 8576, Cur Loss: 0.08774701, Cur Avg Loss: 0.26063906, Log Avg loss: 0.42632953, Global Avg Loss: 1.27591884, Time: 0.0211 Steps: 34600, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000546, Sample Num: 8736, Cur Loss: 0.57286930, Cur Avg Loss: 0.26353061, Log Avg loss: 0.41851764, Global Avg Loss: 1.27567111, Time: 0.0210 Steps: 34610, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000556, Sample Num: 8896, Cur Loss: 0.40633079, Cur Avg Loss: 0.26612981, Log Avg loss: 0.40804619, Global Avg Loss: 1.27542049, Time: 0.0211 Steps: 34620, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000566, Sample Num: 9056, Cur Loss: 0.31065550, Cur Avg Loss: 0.26764467, Log Avg loss: 0.35187039, Global Avg Loss: 1.27515380, Time: 0.0211 Steps: 34630, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000576, Sample Num: 9216, Cur Loss: 0.22365408, Cur Avg Loss: 0.26781316, Log Avg loss: 0.27734972, Global Avg Loss: 1.27486575, Time: 0.0211 Steps: 34640, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000586, Sample Num: 9376, Cur Loss: 0.11564155, Cur Avg Loss: 0.26581972, Log Avg loss: 0.15099784, Global Avg Loss: 1.27454140, Time: 0.0211 Steps: 34650, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000596, Sample Num: 9536, Cur Loss: 0.15087235, Cur Avg Loss: 0.26626847, Log Avg loss: 0.29256547, Global Avg Loss: 1.27425809, Time: 0.0211 Steps: 34660, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000606, Sample Num: 9696, Cur Loss: 0.12149650, Cur Avg Loss: 0.26697468, Log Avg loss: 0.30906434, Global Avg Loss: 1.27397969, Time: 0.0210 Steps: 34670, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000616, Sample Num: 9856, Cur Loss: 0.11553557, Cur Avg Loss: 0.26486979, Log Avg loss: 0.13731346, Global Avg Loss: 1.27365193, Time: 0.0208 Steps: 34680, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000626, Sample Num: 10016, Cur Loss: 0.10381506, Cur Avg Loss: 0.26385458, Log Avg loss: 0.20131760, Global Avg Loss: 1.27334281, Time: 0.0208 Steps: 34690, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000636, Sample Num: 10176, Cur Loss: 0.30723101, Cur Avg Loss: 0.26290307, Log Avg loss: 0.20333880, Global Avg Loss: 1.27303446, Time: 0.0208 Steps: 34700, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000646, Sample Num: 10336, Cur Loss: 0.22922310, Cur Avg Loss: 0.26376673, Log Avg loss: 0.31869559, Global Avg Loss: 1.27275951, Time: 0.0208 Steps: 34710, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000656, Sample Num: 10496, Cur Loss: 0.23787546, Cur Avg Loss: 0.26400081, Log Avg loss: 0.27912225, Global Avg Loss: 1.27247332, Time: 0.0208 Steps: 34720, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000666, Sample Num: 10656, Cur Loss: 0.20102532, Cur Avg Loss: 0.26232667, Log Avg loss: 0.15250313, Global Avg Loss: 1.27215085, Time: 0.0208 Steps: 34730, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000676, Sample Num: 10816, Cur Loss: 0.29088372, Cur Avg Loss: 0.26062703, Log Avg loss: 0.14743078, Global Avg Loss: 1.27182709, Time: 0.0208 Steps: 34740, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000686, Sample Num: 10976, Cur Loss: 0.38102669, Cur Avg Loss: 0.25978238, Log Avg loss: 0.20268447, Global Avg Loss: 1.27151942, Time: 0.0207 Steps: 34750, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000696, Sample Num: 11136, Cur Loss: 0.25294453, Cur Avg Loss: 0.25941115, Log Avg loss: 0.23394426, Global Avg Loss: 1.27122093, Time: 0.0209 Steps: 34760, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000706, Sample Num: 11296, Cur Loss: 0.14762054, Cur Avg Loss: 0.25891119, Log Avg loss: 0.22411413, Global Avg Loss: 1.27091978, Time: 0.0208 Steps: 34770, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000716, Sample Num: 11456, Cur Loss: 0.16670743, Cur Avg Loss: 0.25748894, Log Avg loss: 0.15707792, Global Avg Loss: 1.27059952, Time: 0.0208 Steps: 34780, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000726, Sample Num: 11616, Cur Loss: 0.59333897, Cur Avg Loss: 0.25823443, Log Avg loss: 0.31161216, Global Avg Loss: 1.27032387, Time: 0.0208 Steps: 34790, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000736, Sample Num: 11776, Cur Loss: 0.25472343, Cur Avg Loss: 0.26001734, Log Avg loss: 0.38945655, Global Avg Loss: 1.27007075, Time: 0.0208 Steps: 34800, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000746, Sample Num: 11936, Cur Loss: 0.19989502, Cur Avg Loss: 0.26066754, Log Avg loss: 0.30852169, Global Avg Loss: 1.26979452, Time: 0.0208 Steps: 34810, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000756, Sample Num: 12096, Cur Loss: 0.44033167, Cur Avg Loss: 0.26163070, Log Avg loss: 0.33348279, Global Avg Loss: 1.26952562, Time: 0.0208 Steps: 34820, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000766, Sample Num: 12256, Cur Loss: 0.55374265, Cur Avg Loss: 0.26476765, Log Avg loss: 0.50192105, Global Avg Loss: 1.26930523, Time: 0.0208 Steps: 34830, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000776, Sample Num: 12416, Cur Loss: 0.47968853, Cur Avg Loss: 0.26532770, Log Avg loss: 0.30822765, Global Avg Loss: 1.26902938, Time: 0.0209 Steps: 34840, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000786, Sample Num: 12576, Cur Loss: 0.23765461, Cur Avg Loss: 0.26483386, Log Avg loss: 0.22651200, Global Avg Loss: 1.26873024, Time: 0.0210 Steps: 34850, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000796, Sample Num: 12736, Cur Loss: 0.14234875, Cur Avg Loss: 0.26383089, Log Avg loss: 0.18499706, Global Avg Loss: 1.26841935, Time: 0.0209 Steps: 34860, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000806, Sample Num: 12896, Cur Loss: 0.42191684, Cur Avg Loss: 0.26331743, Log Avg loss: 0.22244593, Global Avg Loss: 1.26811939, Time: 0.0209 Steps: 34870, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000816, Sample Num: 13056, Cur Loss: 0.12543991, Cur Avg Loss: 0.26199499, Log Avg loss: 0.15540679, Global Avg Loss: 1.26780038, Time: 0.0209 Steps: 34880, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000826, Sample Num: 13216, Cur Loss: 0.19771251, Cur Avg Loss: 0.26345240, Log Avg loss: 0.38237714, Global Avg Loss: 1.26754660, Time: 0.0209 Steps: 34890, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000836, Sample Num: 13376, Cur Loss: 0.27804136, Cur Avg Loss: 0.26341099, Log Avg loss: 0.25998976, Global Avg Loss: 1.26725790, Time: 0.0209 Steps: 34900, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000846, Sample Num: 13536, Cur Loss: 0.25249094, Cur Avg Loss: 0.26320251, Log Avg loss: 0.24577394, Global Avg Loss: 1.26696530, Time: 0.0209 Steps: 34910, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000856, Sample Num: 13696, Cur Loss: 0.35577565, Cur Avg Loss: 0.26301147, Log Avg loss: 0.24684983, Global Avg Loss: 1.26667317, Time: 0.0209 Steps: 34920, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000866, Sample Num: 13856, Cur Loss: 0.46591336, Cur Avg Loss: 0.26371685, Log Avg loss: 0.32409704, Global Avg Loss: 1.26640332, Time: 0.0209 Steps: 34930, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000876, Sample Num: 14016, Cur Loss: 0.55004084, Cur Avg Loss: 0.26330674, Log Avg loss: 0.22779168, Global Avg Loss: 1.26610607, Time: 0.0209 Steps: 34940, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000886, Sample Num: 14176, Cur Loss: 0.16508070, Cur Avg Loss: 0.26209967, Log Avg loss: 0.15635988, Global Avg Loss: 1.26578854, Time: 0.0209 Steps: 34950, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000896, Sample Num: 14336, Cur Loss: 0.30546910, Cur Avg Loss: 0.26242956, Log Avg loss: 0.29165805, Global Avg Loss: 1.26550990, Time: 0.0209 Steps: 34960, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000906, Sample Num: 14496, Cur Loss: 0.08087492, Cur Avg Loss: 0.26193099, Log Avg loss: 0.21725933, Global Avg Loss: 1.26521014, Time: 0.0209 Steps: 34970, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000916, Sample Num: 14656, Cur Loss: 0.31076288, Cur Avg Loss: 0.26213592, Log Avg loss: 0.28070216, Global Avg Loss: 1.26492870, Time: 0.0209 Steps: 34980, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000926, Sample Num: 14816, Cur Loss: 0.30049232, Cur Avg Loss: 0.26238966, Log Avg loss: 0.28563252, Global Avg Loss: 1.26464882, Time: 0.0209 Steps: 34990, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000936, Sample Num: 14976, Cur Loss: 0.24657281, Cur Avg Loss: 0.26228007, Log Avg loss: 0.25213176, Global Avg Loss: 1.26435953, Time: 0.0209 Steps: 35000, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000946, Sample Num: 15136, Cur Loss: 0.21120328, Cur Avg Loss: 0.26244708, Log Avg loss: 0.27807955, Global Avg Loss: 1.26407781, Time: 0.0209 Steps: 35010, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000956, Sample Num: 15296, Cur Loss: 0.18077089, Cur Avg Loss: 0.26210609, Log Avg loss: 0.22984776, Global Avg Loss: 1.26378249, Time: 0.0209 Steps: 35020, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000966, Sample Num: 15456, Cur Loss: 0.06200349, Cur Avg Loss: 0.26151166, Log Avg loss: 0.20468445, Global Avg Loss: 1.26348015, Time: 0.0209 Steps: 35030, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000976, Sample Num: 15616, Cur Loss: 0.21135625, Cur Avg Loss: 0.26069283, Log Avg loss: 0.18159361, Global Avg Loss: 1.26317139, Time: 0.0209 Steps: 35040, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000986, Sample Num: 15776, Cur Loss: 0.14830258, Cur Avg Loss: 0.26013636, Log Avg loss: 0.20582510, Global Avg Loss: 1.26286972, Time: 0.0210 Steps: 35050, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000996, Sample Num: 15936, Cur Loss: 0.34011698, Cur Avg Loss: 0.26000273, Log Avg loss: 0.24682697, Global Avg Loss: 1.26257992, Time: 0.0209 Steps: 35060, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001006, Sample Num: 16096, Cur Loss: 0.22266424, Cur Avg Loss: 0.25988957, Log Avg loss: 0.24861908, Global Avg Loss: 1.26229080, Time: 0.0209 Steps: 35070, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001016, Sample Num: 16256, Cur Loss: 0.39421731, Cur Avg Loss: 0.26013093, Log Avg loss: 0.28441106, Global Avg Loss: 1.26201204, Time: 0.0209 Steps: 35080, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001026, Sample Num: 16416, Cur Loss: 0.12981948, Cur Avg Loss: 0.26016404, Log Avg loss: 0.26352793, Global Avg Loss: 1.26172749, Time: 0.0247 Steps: 35090, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001036, Sample Num: 16576, Cur Loss: 0.17167921, Cur Avg Loss: 0.25956867, Log Avg loss: 0.19848461, Global Avg Loss: 1.26142457, Time: 0.0209 Steps: 35100, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001046, Sample Num: 16736, Cur Loss: 0.31253746, Cur Avg Loss: 0.25930526, Log Avg loss: 0.23201542, Global Avg Loss: 1.26113138, Time: 0.0209 Steps: 35110, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001056, Sample Num: 16896, Cur Loss: 0.11953027, Cur Avg Loss: 0.25895286, Log Avg loss: 0.22209155, Global Avg Loss: 1.26083552, Time: 0.0209 Steps: 35120, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001066, Sample Num: 17056, Cur Loss: 0.26765242, Cur Avg Loss: 0.25907029, Log Avg loss: 0.27147130, Global Avg Loss: 1.26055389, Time: 0.0208 Steps: 35130, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001076, Sample Num: 17216, Cur Loss: 0.19654578, Cur Avg Loss: 0.25821436, Log Avg loss: 0.16697250, Global Avg Loss: 1.26024268, Time: 0.0209 Steps: 35140, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001086, Sample Num: 17376, Cur Loss: 0.12947908, Cur Avg Loss: 0.25760930, Log Avg loss: 0.19250505, Global Avg Loss: 1.25993892, Time: 0.0209 Steps: 35150, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001096, Sample Num: 17536, Cur Loss: 0.14980930, Cur Avg Loss: 0.25719417, Log Avg loss: 0.21211008, Global Avg Loss: 1.25964090, Time: 0.0208 Steps: 35160, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001106, Sample Num: 17696, Cur Loss: 0.13136573, Cur Avg Loss: 0.25668892, Log Avg loss: 0.20131377, Global Avg Loss: 1.25933998, Time: 0.0209 Steps: 35170, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001116, Sample Num: 17856, Cur Loss: 0.38131827, Cur Avg Loss: 0.25731495, Log Avg loss: 0.32655417, Global Avg Loss: 1.25907484, Time: 0.0209 Steps: 35180, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001126, Sample Num: 18016, Cur Loss: 0.05306643, Cur Avg Loss: 0.25650564, Log Avg loss: 0.16618686, Global Avg Loss: 1.25876427, Time: 0.0208 Steps: 35190, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001136, Sample Num: 18176, Cur Loss: 0.17903101, Cur Avg Loss: 0.25728021, Log Avg loss: 0.34449631, Global Avg Loss: 1.25850453, Time: 0.0209 Steps: 35200, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001146, Sample Num: 18336, Cur Loss: 0.29059932, Cur Avg Loss: 0.25675246, Log Avg loss: 0.19679967, Global Avg Loss: 1.25820300, Time: 0.0209 Steps: 35210, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001156, Sample Num: 18496, Cur Loss: 0.21307282, Cur Avg Loss: 0.25696810, Log Avg loss: 0.28168067, Global Avg Loss: 1.25792574, Time: 0.0209 Steps: 35220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001166, Sample Num: 18656, Cur Loss: 0.37634248, Cur Avg Loss: 0.25667847, Log Avg loss: 0.22319699, Global Avg Loss: 1.25763203, Time: 0.0208 Steps: 35230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001176, Sample Num: 18816, Cur Loss: 0.13062219, Cur Avg Loss: 0.25658035, Log Avg loss: 0.24514060, Global Avg Loss: 1.25734472, Time: 0.0209 Steps: 35240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001186, Sample Num: 18976, Cur Loss: 0.39470083, Cur Avg Loss: 0.25623255, Log Avg loss: 0.21533026, Global Avg Loss: 1.25704911, Time: 0.0209 Steps: 35250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001196, Sample Num: 19136, Cur Loss: 0.13429084, Cur Avg Loss: 0.25554185, Log Avg loss: 0.17362564, Global Avg Loss: 1.25674184, Time: 0.0208 Steps: 35260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001206, Sample Num: 19296, Cur Loss: 0.43793556, Cur Avg Loss: 0.25650344, Log Avg loss: 0.37150904, Global Avg Loss: 1.25649085, Time: 0.0209 Steps: 35270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001216, Sample Num: 19456, Cur Loss: 0.22279197, Cur Avg Loss: 0.25693194, Log Avg loss: 0.30860895, Global Avg Loss: 1.25622218, Time: 0.0209 Steps: 35280, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001226, Sample Num: 19616, Cur Loss: 0.14776810, Cur Avg Loss: 0.25678579, Log Avg loss: 0.23901388, Global Avg Loss: 1.25593394, Time: 0.0209 Steps: 35290, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001236, Sample Num: 19776, Cur Loss: 0.18720874, Cur Avg Loss: 0.25698348, Log Avg loss: 0.28122052, Global Avg Loss: 1.25565781, Time: 0.0209 Steps: 35300, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001246, Sample Num: 19936, Cur Loss: 0.76107144, Cur Avg Loss: 0.25865924, Log Avg loss: 0.46578277, Global Avg Loss: 1.25543412, Time: 0.0208 Steps: 35310, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001256, Sample Num: 20096, Cur Loss: 0.07663888, Cur Avg Loss: 0.25858809, Log Avg loss: 0.24972305, Global Avg Loss: 1.25514937, Time: 0.0208 Steps: 35320, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001266, Sample Num: 20256, Cur Loss: 0.25488496, Cur Avg Loss: 0.25849663, Log Avg loss: 0.24700926, Global Avg Loss: 1.25486403, Time: 0.0208 Steps: 35330, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001276, Sample Num: 20416, Cur Loss: 0.41416469, Cur Avg Loss: 0.25927597, Log Avg loss: 0.35794125, Global Avg Loss: 1.25461023, Time: 0.0208 Steps: 35340, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001286, Sample Num: 20576, Cur Loss: 0.09656927, Cur Avg Loss: 0.25840490, Log Avg loss: 0.14725642, Global Avg Loss: 1.25429697, Time: 0.0227 Steps: 35350, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001296, Sample Num: 20736, Cur Loss: 0.25960806, Cur Avg Loss: 0.25824036, Log Avg loss: 0.23707973, Global Avg Loss: 1.25400930, Time: 0.0209 Steps: 35360, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001306, Sample Num: 20896, Cur Loss: 0.17269613, Cur Avg Loss: 0.25804270, Log Avg loss: 0.23242562, Global Avg Loss: 1.25372047, Time: 0.0209 Steps: 35370, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001316, Sample Num: 21056, Cur Loss: 0.41389680, Cur Avg Loss: 0.25807992, Log Avg loss: 0.26294159, Global Avg Loss: 1.25344043, Time: 0.0208 Steps: 35380, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001326, Sample Num: 21216, Cur Loss: 0.29838443, Cur Avg Loss: 0.25794920, Log Avg loss: 0.24074685, Global Avg Loss: 1.25315428, Time: 0.0208 Steps: 35390, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001336, Sample Num: 21376, Cur Loss: 0.08323412, Cur Avg Loss: 0.25834002, Log Avg loss: 0.31016282, Global Avg Loss: 1.25288790, Time: 0.0209 Steps: 35400, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001346, Sample Num: 21536, Cur Loss: 0.35071099, Cur Avg Loss: 0.25828564, Log Avg loss: 0.25102019, Global Avg Loss: 1.25260496, Time: 0.0209 Steps: 35410, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001356, Sample Num: 21696, Cur Loss: 0.45590687, Cur Avg Loss: 0.25949892, Log Avg loss: 0.42280598, Global Avg Loss: 1.25237069, Time: 0.0208 Steps: 35420, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001366, Sample Num: 21856, Cur Loss: 0.25799879, Cur Avg Loss: 0.25950466, Log Avg loss: 0.26028285, Global Avg Loss: 1.25209068, Time: 0.0209 Steps: 35430, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001376, Sample Num: 22016, Cur Loss: 0.05282710, Cur Avg Loss: 0.25949653, Log Avg loss: 0.25838614, Global Avg Loss: 1.25181029, Time: 0.0209 Steps: 35440, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001386, Sample Num: 22176, Cur Loss: 0.08727438, Cur Avg Loss: 0.25908565, Log Avg loss: 0.20254825, Global Avg Loss: 1.25151430, Time: 0.0208 Steps: 35450, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001396, Sample Num: 22336, Cur Loss: 0.18855712, Cur Avg Loss: 0.25929615, Log Avg loss: 0.28847186, Global Avg Loss: 1.25124272, Time: 0.0209 Steps: 35460, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001406, Sample Num: 22496, Cur Loss: 0.44073534, Cur Avg Loss: 0.25925726, Log Avg loss: 0.25382864, Global Avg Loss: 1.25096152, Time: 0.0209 Steps: 35470, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001416, Sample Num: 22656, Cur Loss: 0.44972658, Cur Avg Loss: 0.25982856, Log Avg loss: 0.34015270, Global Avg Loss: 1.25070481, Time: 0.0209 Steps: 35480, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001426, Sample Num: 22816, Cur Loss: 0.27305937, Cur Avg Loss: 0.26002998, Log Avg loss: 0.28855142, Global Avg Loss: 1.25043370, Time: 0.0209 Steps: 35490, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001436, Sample Num: 22976, Cur Loss: 0.15259154, Cur Avg Loss: 0.26050384, Log Avg loss: 0.32807677, Global Avg Loss: 1.25017388, Time: 0.0209 Steps: 35500, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001446, Sample Num: 23136, Cur Loss: 0.43991002, Cur Avg Loss: 0.26032000, Log Avg loss: 0.23391950, Global Avg Loss: 1.24988769, Time: 0.0208 Steps: 35510, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001456, Sample Num: 23296, Cur Loss: 0.04898955, Cur Avg Loss: 0.26025225, Log Avg loss: 0.25045597, Global Avg Loss: 1.24960632, Time: 0.0208 Steps: 35520, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001466, Sample Num: 23456, Cur Loss: 0.18515965, Cur Avg Loss: 0.26014464, Log Avg loss: 0.24447649, Global Avg Loss: 1.24932343, Time: 0.0209 Steps: 35530, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001476, Sample Num: 23616, Cur Loss: 0.15336031, Cur Avg Loss: 0.26032957, Log Avg loss: 0.28744034, Global Avg Loss: 1.24905278, Time: 0.0209 Steps: 35540, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001486, Sample Num: 23776, Cur Loss: 0.12583406, Cur Avg Loss: 0.26026861, Log Avg loss: 0.25127062, Global Avg Loss: 1.24877211, Time: 0.0209 Steps: 35550, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001496, Sample Num: 23936, Cur Loss: 0.29608327, Cur Avg Loss: 0.26031638, Log Avg loss: 0.26741495, Global Avg Loss: 1.24849614, Time: 0.0209 Steps: 35560, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001506, Sample Num: 24096, Cur Loss: 0.31091541, Cur Avg Loss: 0.26043193, Log Avg loss: 0.27771855, Global Avg Loss: 1.24822322, Time: 0.0209 Steps: 35570, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001516, Sample Num: 24256, Cur Loss: 0.19905061, Cur Avg Loss: 0.26006901, Log Avg loss: 0.20541386, Global Avg Loss: 1.24793013, Time: 0.0209 Steps: 35580, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001526, Sample Num: 24416, Cur Loss: 0.20000805, Cur Avg Loss: 0.26017837, Log Avg loss: 0.27675667, Global Avg Loss: 1.24765725, Time: 0.0209 Steps: 35590, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001536, Sample Num: 24576, Cur Loss: 0.07732476, Cur Avg Loss: 0.25988355, Log Avg loss: 0.21489479, Global Avg Loss: 1.24736715, Time: 0.0254 Steps: 35600, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001546, Sample Num: 24736, Cur Loss: 0.55925471, Cur Avg Loss: 0.26024328, Log Avg loss: 0.31549724, Global Avg Loss: 1.24710546, Time: 0.0209 Steps: 35610, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001556, Sample Num: 24896, Cur Loss: 0.28527626, Cur Avg Loss: 0.26114260, Log Avg loss: 0.40017752, Global Avg Loss: 1.24686769, Time: 0.0209 Steps: 35620, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001566, Sample Num: 25056, Cur Loss: 0.55118531, Cur Avg Loss: 0.26220173, Log Avg loss: 0.42700173, Global Avg Loss: 1.24663759, Time: 0.0209 Steps: 35630, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001576, Sample Num: 25216, Cur Loss: 0.18294245, Cur Avg Loss: 0.26256358, Log Avg loss: 0.31922908, Global Avg Loss: 1.24637737, Time: 0.0209 Steps: 35640, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001586, Sample Num: 25376, Cur Loss: 0.12594062, Cur Avg Loss: 0.26393914, Log Avg loss: 0.48072805, Global Avg Loss: 1.24616260, Time: 0.0209 Steps: 35650, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001596, Sample Num: 25536, Cur Loss: 0.09496018, Cur Avg Loss: 0.26333318, Log Avg loss: 0.16722838, Global Avg Loss: 1.24586004, Time: 0.0208 Steps: 35660, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001606, Sample Num: 25696, Cur Loss: 0.29464784, Cur Avg Loss: 0.26325599, Log Avg loss: 0.25093593, Global Avg Loss: 1.24558112, Time: 0.0209 Steps: 35670, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001616, Sample Num: 25856, Cur Loss: 0.22192186, Cur Avg Loss: 0.26314352, Log Avg loss: 0.24508096, Global Avg Loss: 1.24530071, Time: 0.0209 Steps: 35680, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001626, Sample Num: 26016, Cur Loss: 0.19788052, Cur Avg Loss: 0.26284601, Log Avg loss: 0.21476831, Global Avg Loss: 1.24501196, Time: 0.0209 Steps: 35690, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001636, Sample Num: 26176, Cur Loss: 0.34463334, Cur Avg Loss: 0.26250858, Log Avg loss: 0.20764187, Global Avg Loss: 1.24472138, Time: 0.0209 Steps: 35700, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001646, Sample Num: 26336, Cur Loss: 0.23889190, Cur Avg Loss: 0.26218649, Log Avg loss: 0.20949303, Global Avg Loss: 1.24443148, Time: 0.0209 Steps: 35710, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001656, Sample Num: 26496, Cur Loss: 0.24473558, Cur Avg Loss: 0.26191259, Log Avg loss: 0.21682889, Global Avg Loss: 1.24414380, Time: 0.0209 Steps: 35720, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001666, Sample Num: 26656, Cur Loss: 0.27983731, Cur Avg Loss: 0.26201734, Log Avg loss: 0.27936391, Global Avg Loss: 1.24387378, Time: 0.0209 Steps: 35730, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001676, Sample Num: 26816, Cur Loss: 0.09557663, Cur Avg Loss: 0.26176716, Log Avg loss: 0.22008675, Global Avg Loss: 1.24358733, Time: 0.0209 Steps: 35740, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001686, Sample Num: 26976, Cur Loss: 0.11711542, Cur Avg Loss: 0.26087065, Log Avg loss: 0.11061573, Global Avg Loss: 1.24327041, Time: 0.0209 Steps: 35750, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001696, Sample Num: 27136, Cur Loss: 0.14189982, Cur Avg Loss: 0.26087955, Log Avg loss: 0.26237983, Global Avg Loss: 1.24299611, Time: 0.0209 Steps: 35760, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001706, Sample Num: 27296, Cur Loss: 0.16477254, Cur Avg Loss: 0.26137986, Log Avg loss: 0.34623230, Global Avg Loss: 1.24274541, Time: 0.0209 Steps: 35770, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001716, Sample Num: 27456, Cur Loss: 0.49971142, Cur Avg Loss: 0.26148318, Log Avg loss: 0.27910980, Global Avg Loss: 1.24247609, Time: 0.0209 Steps: 35780, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001726, Sample Num: 27616, Cur Loss: 0.25127101, Cur Avg Loss: 0.26110874, Log Avg loss: 0.19685459, Global Avg Loss: 1.24218393, Time: 0.0209 Steps: 35790, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001736, Sample Num: 27776, Cur Loss: 0.09848146, Cur Avg Loss: 0.26114441, Log Avg loss: 0.26730236, Global Avg Loss: 1.24191162, Time: 0.0209 Steps: 35800, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001746, Sample Num: 27936, Cur Loss: 0.23699898, Cur Avg Loss: 0.26119417, Log Avg loss: 0.26983192, Global Avg Loss: 1.24164016, Time: 0.0209 Steps: 35810, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001756, Sample Num: 28096, Cur Loss: 0.37350547, Cur Avg Loss: 0.26148194, Log Avg loss: 0.31172733, Global Avg Loss: 1.24138056, Time: 0.0209 Steps: 35820, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001766, Sample Num: 28256, Cur Loss: 0.12210606, Cur Avg Loss: 0.26200936, Log Avg loss: 0.35462394, Global Avg Loss: 1.24113307, Time: 0.0209 Steps: 35830, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001776, Sample Num: 28416, Cur Loss: 0.14747261, Cur Avg Loss: 0.26191771, Log Avg loss: 0.24573136, Global Avg Loss: 1.24085533, Time: 0.0209 Steps: 35840, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001786, Sample Num: 28576, Cur Loss: 0.17922094, Cur Avg Loss: 0.26190396, Log Avg loss: 0.25946325, Global Avg Loss: 1.24058158, Time: 0.0209 Steps: 35850, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001796, Sample Num: 28736, Cur Loss: 0.06398679, Cur Avg Loss: 0.26148139, Log Avg loss: 0.18600899, Global Avg Loss: 1.24028750, Time: 0.0248 Steps: 35860, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001806, Sample Num: 28896, Cur Loss: 0.03397556, Cur Avg Loss: 0.26091392, Log Avg loss: 0.15899756, Global Avg Loss: 1.23998606, Time: 0.0210 Steps: 35870, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001816, Sample Num: 29056, Cur Loss: 0.17929128, Cur Avg Loss: 0.26060838, Log Avg loss: 0.20542781, Global Avg Loss: 1.23969772, Time: 0.0210 Steps: 35880, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001826, Sample Num: 29216, Cur Loss: 0.33051944, Cur Avg Loss: 0.26070561, Log Avg loss: 0.27836146, Global Avg Loss: 1.23942986, Time: 0.0210 Steps: 35890, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001836, Sample Num: 29376, Cur Loss: 0.31747872, Cur Avg Loss: 0.26116865, Log Avg loss: 0.34571998, Global Avg Loss: 1.23918092, Time: 0.0210 Steps: 35900, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001846, Sample Num: 29536, Cur Loss: 0.60905969, Cur Avg Loss: 0.26107128, Log Avg loss: 0.24319437, Global Avg Loss: 1.23890356, Time: 0.0210 Steps: 35910, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001856, Sample Num: 29696, Cur Loss: 1.10949254, Cur Avg Loss: 0.26214497, Log Avg loss: 0.46034914, Global Avg Loss: 1.23868681, Time: 0.0210 Steps: 35920, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001866, Sample Num: 29856, Cur Loss: 0.26428670, Cur Avg Loss: 0.26190093, Log Avg loss: 0.21660657, Global Avg Loss: 1.23840235, Time: 0.0210 Steps: 35930, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001876, Sample Num: 30016, Cur Loss: 0.62474304, Cur Avg Loss: 0.26223982, Log Avg loss: 0.32547614, Global Avg Loss: 1.23814834, Time: 0.0211 Steps: 35940, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001886, Sample Num: 30176, Cur Loss: 0.26682660, Cur Avg Loss: 0.26294585, Log Avg loss: 0.39539702, Global Avg Loss: 1.23791391, Time: 0.0211 Steps: 35950, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001896, Sample Num: 30336, Cur Loss: 0.02355631, Cur Avg Loss: 0.26269565, Log Avg loss: 0.21550791, Global Avg Loss: 1.23762959, Time: 0.0210 Steps: 35960, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001906, Sample Num: 30496, Cur Loss: 0.26315227, Cur Avg Loss: 0.26266509, Log Avg loss: 0.25687106, Global Avg Loss: 1.23735693, Time: 0.0210 Steps: 35970, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001916, Sample Num: 30656, Cur Loss: 0.05650618, Cur Avg Loss: 0.26233123, Log Avg loss: 0.19869697, Global Avg Loss: 1.23706826, Time: 0.0211 Steps: 35980, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001926, Sample Num: 30816, Cur Loss: 0.26781720, Cur Avg Loss: 0.26239115, Log Avg loss: 0.27387180, Global Avg Loss: 1.23680063, Time: 0.0210 Steps: 35990, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001936, Sample Num: 30976, Cur Loss: 0.15620248, Cur Avg Loss: 0.26258924, Log Avg loss: 0.30074229, Global Avg Loss: 1.23654061, Time: 0.0210 Steps: 36000, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001946, Sample Num: 31136, Cur Loss: 0.26338956, Cur Avg Loss: 0.26301553, Log Avg loss: 0.34554496, Global Avg Loss: 1.23629318, Time: 0.0211 Steps: 36010, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001956, Sample Num: 31296, Cur Loss: 0.25014171, Cur Avg Loss: 0.26271245, Log Avg loss: 0.20373253, Global Avg Loss: 1.23600652, Time: 0.0210 Steps: 36020, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001966, Sample Num: 31456, Cur Loss: 0.28523898, Cur Avg Loss: 0.26271908, Log Avg loss: 0.26401590, Global Avg Loss: 1.23573675, Time: 0.0210 Steps: 36030, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001976, Sample Num: 31616, Cur Loss: 0.17625879, Cur Avg Loss: 0.26256363, Log Avg loss: 0.23200354, Global Avg Loss: 1.23545824, Time: 0.0210 Steps: 36040, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001986, Sample Num: 31776, Cur Loss: 0.31838584, Cur Avg Loss: 0.26276807, Log Avg loss: 0.30316493, Global Avg Loss: 1.23519963, Time: 0.0210 Steps: 36050, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001996, Sample Num: 31936, Cur Loss: 0.32847458, Cur Avg Loss: 0.26273474, Log Avg loss: 0.25611478, Global Avg Loss: 1.23492811, Time: 0.0211 Steps: 36060, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002006, Sample Num: 32096, Cur Loss: 0.29474512, Cur Avg Loss: 0.26292751, Log Avg loss: 0.30140522, Global Avg Loss: 1.23466931, Time: 0.0210 Steps: 36070, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002016, Sample Num: 32256, Cur Loss: 0.17523167, Cur Avg Loss: 0.26281493, Log Avg loss: 0.24023116, Global Avg Loss: 1.23439369, Time: 0.0211 Steps: 36080, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002026, Sample Num: 32416, Cur Loss: 0.09208041, Cur Avg Loss: 0.26247587, Log Avg loss: 0.19412054, Global Avg Loss: 1.23410544, Time: 0.0210 Steps: 36090, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002036, Sample Num: 32576, Cur Loss: 0.05878711, Cur Avg Loss: 0.26224119, Log Avg loss: 0.21469653, Global Avg Loss: 1.23382306, Time: 0.0210 Steps: 36100, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002046, Sample Num: 32736, Cur Loss: 0.31891525, Cur Avg Loss: 0.26278953, Log Avg loss: 0.37443029, Global Avg Loss: 1.23358506, Time: 0.0210 Steps: 36110, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002056, Sample Num: 32896, Cur Loss: 0.15169778, Cur Avg Loss: 0.26267738, Log Avg loss: 0.23973286, Global Avg Loss: 1.23330991, Time: 0.0210 Steps: 36120, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002066, Sample Num: 33056, Cur Loss: 0.58688009, Cur Avg Loss: 0.26299946, Log Avg loss: 0.32921843, Global Avg Loss: 1.23305968, Time: 0.0210 Steps: 36130, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002076, Sample Num: 33216, Cur Loss: 0.26465338, Cur Avg Loss: 0.26321414, Log Avg loss: 0.30756668, Global Avg Loss: 1.23280359, Time: 0.0209 Steps: 36140, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002086, Sample Num: 33376, Cur Loss: 0.09091225, Cur Avg Loss: 0.26317045, Log Avg loss: 0.25410089, Global Avg Loss: 1.23253286, Time: 0.0210 Steps: 36150, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002096, Sample Num: 33536, Cur Loss: 0.35126802, Cur Avg Loss: 0.26294869, Log Avg loss: 0.21668825, Global Avg Loss: 1.23225193, Time: 0.0209 Steps: 36160, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002106, Sample Num: 33696, Cur Loss: 0.16030695, Cur Avg Loss: 0.26267057, Log Avg loss: 0.20437771, Global Avg Loss: 1.23196775, Time: 0.0210 Steps: 36170, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002116, Sample Num: 33856, Cur Loss: 0.23696968, Cur Avg Loss: 0.26268802, Log Avg loss: 0.26636237, Global Avg Loss: 1.23170086, Time: 0.0209 Steps: 36180, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002126, Sample Num: 34016, Cur Loss: 0.71756923, Cur Avg Loss: 0.26288569, Log Avg loss: 0.30471344, Global Avg Loss: 1.23144472, Time: 0.0209 Steps: 36190, Updated lr: 0.000067 ***** Running evaluation checkpoint-36193 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-36193 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.755650, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.312498, "eval_total_loss": 219.686268, "eval_mae": 0.435523, "eval_mse": 0.312541, "eval_r2": 0.801329, "eval_sp_statistic": 0.903387, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.916341, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.276515, "test_total_loss": 138.810427, "test_mae": 0.339511, "test_mse": 0.276615, "test_r2": 0.82147, "test_sp_statistic": 0.901606, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.943203, "test_ps_pvalue": 0.0, "lr": 6.662588904694168e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2313778286810493, "train_cur_epoch_loss": 560.1684690611437, "train_cur_epoch_avg_loss": 0.26311341900476454, "train_cur_epoch_time": 44.75564980506897, "train_cur_epoch_avg_time": 0.02102191160407185, "epoch": 17, "step": 36193} ################################################## Training, Epoch: 0018, Batch: 000007, Sample Num: 112, Cur Loss: 0.37730923, Cur Avg Loss: 0.26810216, Log Avg loss: 0.31502042, Global Avg Loss: 1.23119156, Time: 0.0246 Steps: 36200, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000017, Sample Num: 272, Cur Loss: 0.37761682, Cur Avg Loss: 0.22894596, Log Avg loss: 0.20153662, Global Avg Loss: 1.23090720, Time: 0.0209 Steps: 36210, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000027, Sample Num: 432, Cur Loss: 0.64931655, Cur Avg Loss: 0.24062866, Log Avg loss: 0.26048924, Global Avg Loss: 1.23063928, Time: 0.0209 Steps: 36220, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000037, Sample Num: 592, Cur Loss: 0.52668488, Cur Avg Loss: 0.26990819, Log Avg loss: 0.34896291, Global Avg Loss: 1.23039592, Time: 0.0209 Steps: 36230, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000047, Sample Num: 752, Cur Loss: 2.19783211, Cur Avg Loss: 0.34036321, Log Avg loss: 0.60104682, Global Avg Loss: 1.23022226, Time: 0.0209 Steps: 36240, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000057, Sample Num: 912, Cur Loss: 0.51074100, Cur Avg Loss: 0.35003315, Log Avg loss: 0.39548182, Global Avg Loss: 1.22999199, Time: 0.0209 Steps: 36250, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000067, Sample Num: 1072, Cur Loss: 0.11377885, Cur Avg Loss: 0.36610968, Log Avg loss: 0.45774594, Global Avg Loss: 1.22977902, Time: 0.0209 Steps: 36260, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000077, Sample Num: 1232, Cur Loss: 0.04541960, Cur Avg Loss: 0.34222247, Log Avg loss: 0.18217812, Global Avg Loss: 1.22949018, Time: 0.0209 Steps: 36270, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000087, Sample Num: 1392, Cur Loss: 0.40273315, Cur Avg Loss: 0.32598783, Log Avg loss: 0.20098117, Global Avg Loss: 1.22920669, Time: 0.0209 Steps: 36280, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000097, Sample Num: 1552, Cur Loss: 0.21097100, Cur Avg Loss: 0.31567537, Log Avg loss: 0.22595693, Global Avg Loss: 1.22893024, Time: 0.0209 Steps: 36290, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000107, Sample Num: 1712, Cur Loss: 0.07995171, Cur Avg Loss: 0.30589994, Log Avg loss: 0.21107823, Global Avg Loss: 1.22864984, Time: 0.0209 Steps: 36300, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000117, Sample Num: 1872, Cur Loss: 0.15460670, Cur Avg Loss: 0.29617789, Log Avg loss: 0.19215205, Global Avg Loss: 1.22836438, Time: 0.0209 Steps: 36310, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000127, Sample Num: 2032, Cur Loss: 0.13208519, Cur Avg Loss: 0.28506454, Log Avg loss: 0.15503833, Global Avg Loss: 1.22806886, Time: 0.0210 Steps: 36320, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000137, Sample Num: 2192, Cur Loss: 0.27859458, Cur Avg Loss: 0.28201191, Log Avg loss: 0.24324348, Global Avg Loss: 1.22779778, Time: 0.0209 Steps: 36330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000147, Sample Num: 2352, Cur Loss: 0.30580229, Cur Avg Loss: 0.28254854, Log Avg loss: 0.28990037, Global Avg Loss: 1.22753969, Time: 0.0209 Steps: 36340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000157, Sample Num: 2512, Cur Loss: 0.29329294, Cur Avg Loss: 0.28270306, Log Avg loss: 0.28497448, Global Avg Loss: 1.22728039, Time: 0.0209 Steps: 36350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000167, Sample Num: 2672, Cur Loss: 0.20604773, Cur Avg Loss: 0.27581825, Log Avg loss: 0.16772681, Global Avg Loss: 1.22698898, Time: 0.0209 Steps: 36360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000177, Sample Num: 2832, Cur Loss: 0.31452581, Cur Avg Loss: 0.27596513, Log Avg loss: 0.27841796, Global Avg Loss: 1.22672817, Time: 0.0209 Steps: 36370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000187, Sample Num: 2992, Cur Loss: 0.75998449, Cur Avg Loss: 0.27487473, Log Avg loss: 0.25557462, Global Avg Loss: 1.22646122, Time: 0.0208 Steps: 36380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000197, Sample Num: 3152, Cur Loss: 0.14201969, Cur Avg Loss: 0.27147832, Log Avg loss: 0.20796543, Global Avg Loss: 1.22618134, Time: 0.0209 Steps: 36390, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000207, Sample Num: 3312, Cur Loss: 0.06994805, Cur Avg Loss: 0.26712451, Log Avg loss: 0.18135442, Global Avg Loss: 1.22589430, Time: 0.0208 Steps: 36400, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000217, Sample Num: 3472, Cur Loss: 0.14157793, Cur Avg Loss: 0.26750713, Log Avg loss: 0.27542743, Global Avg Loss: 1.22563325, Time: 0.0208 Steps: 36410, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000227, Sample Num: 3632, Cur Loss: 0.23601334, Cur Avg Loss: 0.26396423, Log Avg loss: 0.18708340, Global Avg Loss: 1.22534810, Time: 0.0208 Steps: 36420, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000237, Sample Num: 3792, Cur Loss: 0.38343024, Cur Avg Loss: 0.26436540, Log Avg loss: 0.27347185, Global Avg Loss: 1.22508681, Time: 0.0208 Steps: 36430, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000247, Sample Num: 3952, Cur Loss: 0.29624158, Cur Avg Loss: 0.26369182, Log Avg loss: 0.24772801, Global Avg Loss: 1.22481860, Time: 0.0209 Steps: 36440, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000257, Sample Num: 4112, Cur Loss: 0.47647274, Cur Avg Loss: 0.26271560, Log Avg loss: 0.23860296, Global Avg Loss: 1.22454803, Time: 0.0246 Steps: 36450, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000267, Sample Num: 4272, Cur Loss: 0.12426224, Cur Avg Loss: 0.26351905, Log Avg loss: 0.28416770, Global Avg Loss: 1.22429011, Time: 0.0209 Steps: 36460, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000277, Sample Num: 4432, Cur Loss: 0.06811469, Cur Avg Loss: 0.26407991, Log Avg loss: 0.27905487, Global Avg Loss: 1.22403093, Time: 0.0209 Steps: 36470, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000287, Sample Num: 4592, Cur Loss: 0.56272840, Cur Avg Loss: 0.26368850, Log Avg loss: 0.25284637, Global Avg Loss: 1.22376470, Time: 0.0209 Steps: 36480, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000297, Sample Num: 4752, Cur Loss: 0.06827457, Cur Avg Loss: 0.25900961, Log Avg loss: 0.12472568, Global Avg Loss: 1.22346351, Time: 0.0209 Steps: 36490, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000307, Sample Num: 4912, Cur Loss: 0.27414995, Cur Avg Loss: 0.25709390, Log Avg loss: 0.20019716, Global Avg Loss: 1.22318317, Time: 0.0209 Steps: 36500, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000317, Sample Num: 5072, Cur Loss: 0.14080474, Cur Avg Loss: 0.25428418, Log Avg loss: 0.16802583, Global Avg Loss: 1.22289416, Time: 0.0209 Steps: 36510, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000327, Sample Num: 5232, Cur Loss: 0.15760291, Cur Avg Loss: 0.25962165, Log Avg loss: 0.42881944, Global Avg Loss: 1.22267673, Time: 0.0209 Steps: 36520, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000337, Sample Num: 5392, Cur Loss: 0.18156369, Cur Avg Loss: 0.25762603, Log Avg loss: 0.19236920, Global Avg Loss: 1.22239468, Time: 0.0209 Steps: 36530, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000347, Sample Num: 5552, Cur Loss: 0.16932389, Cur Avg Loss: 0.25554161, Log Avg loss: 0.18529675, Global Avg Loss: 1.22211086, Time: 0.0210 Steps: 36540, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000357, Sample Num: 5712, Cur Loss: 0.21460265, Cur Avg Loss: 0.25398737, Log Avg loss: 0.20005530, Global Avg Loss: 1.22183122, Time: 0.0209 Steps: 36550, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000367, Sample Num: 5872, Cur Loss: 0.14036097, Cur Avg Loss: 0.25364474, Log Avg loss: 0.24141281, Global Avg Loss: 1.22156306, Time: 0.0209 Steps: 36560, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000377, Sample Num: 6032, Cur Loss: 0.21963453, Cur Avg Loss: 0.25516599, Log Avg loss: 0.31099580, Global Avg Loss: 1.22131406, Time: 0.0209 Steps: 36570, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000387, Sample Num: 6192, Cur Loss: 0.13969778, Cur Avg Loss: 0.25433634, Log Avg loss: 0.22305864, Global Avg Loss: 1.22104117, Time: 0.0209 Steps: 36580, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000397, Sample Num: 6352, Cur Loss: 0.19257304, Cur Avg Loss: 0.25486585, Log Avg loss: 0.27535774, Global Avg Loss: 1.22078271, Time: 0.0209 Steps: 36590, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000407, Sample Num: 6512, Cur Loss: 0.26713222, Cur Avg Loss: 0.25428134, Log Avg loss: 0.23107623, Global Avg Loss: 1.22051230, Time: 0.0210 Steps: 36600, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000417, Sample Num: 6672, Cur Loss: 0.18045457, Cur Avg Loss: 0.25238255, Log Avg loss: 0.17510189, Global Avg Loss: 1.22022675, Time: 0.0209 Steps: 36610, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000427, Sample Num: 6832, Cur Loss: 0.22382239, Cur Avg Loss: 0.25124494, Log Avg loss: 0.20380674, Global Avg Loss: 1.21994919, Time: 0.0210 Steps: 36620, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000437, Sample Num: 6992, Cur Loss: 0.33811921, Cur Avg Loss: 0.25081281, Log Avg loss: 0.23236075, Global Avg Loss: 1.21967958, Time: 0.0209 Steps: 36630, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000447, Sample Num: 7152, Cur Loss: 0.07928132, Cur Avg Loss: 0.24857315, Log Avg loss: 0.15070007, Global Avg Loss: 1.21938783, Time: 0.0209 Steps: 36640, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000457, Sample Num: 7312, Cur Loss: 0.23555136, Cur Avg Loss: 0.25046130, Log Avg loss: 0.33486131, Global Avg Loss: 1.21914648, Time: 0.0209 Steps: 36650, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000467, Sample Num: 7472, Cur Loss: 0.39219904, Cur Avg Loss: 0.25116167, Log Avg loss: 0.28316903, Global Avg Loss: 1.21889117, Time: 0.0210 Steps: 36660, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000477, Sample Num: 7632, Cur Loss: 0.13265520, Cur Avg Loss: 0.25199701, Log Avg loss: 0.29100718, Global Avg Loss: 1.21863813, Time: 0.0209 Steps: 36670, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000487, Sample Num: 7792, Cur Loss: 0.22025354, Cur Avg Loss: 0.24974279, Log Avg loss: 0.14221647, Global Avg Loss: 1.21834467, Time: 0.0209 Steps: 36680, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000497, Sample Num: 7952, Cur Loss: 0.23516576, Cur Avg Loss: 0.25041712, Log Avg loss: 0.28325695, Global Avg Loss: 1.21808981, Time: 0.0209 Steps: 36690, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000507, Sample Num: 8112, Cur Loss: 0.15922081, Cur Avg Loss: 0.24905679, Log Avg loss: 0.18144856, Global Avg Loss: 1.21780734, Time: 0.0209 Steps: 36700, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000517, Sample Num: 8272, Cur Loss: 0.14660868, Cur Avg Loss: 0.24820853, Log Avg loss: 0.20520166, Global Avg Loss: 1.21753151, Time: 0.0247 Steps: 36710, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000527, Sample Num: 8432, Cur Loss: 0.08584016, Cur Avg Loss: 0.24605882, Log Avg loss: 0.13491898, Global Avg Loss: 1.21723668, Time: 0.0209 Steps: 36720, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000537, Sample Num: 8592, Cur Loss: 0.26178622, Cur Avg Loss: 0.24634939, Log Avg loss: 0.26166225, Global Avg Loss: 1.21697651, Time: 0.0210 Steps: 36730, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000547, Sample Num: 8752, Cur Loss: 0.35993814, Cur Avg Loss: 0.24512281, Log Avg loss: 0.17925544, Global Avg Loss: 1.21669406, Time: 0.0210 Steps: 36740, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000557, Sample Num: 8912, Cur Loss: 0.55451530, Cur Avg Loss: 0.24521823, Log Avg loss: 0.25043788, Global Avg Loss: 1.21643114, Time: 0.0209 Steps: 36750, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000567, Sample Num: 9072, Cur Loss: 0.37342080, Cur Avg Loss: 0.24687055, Log Avg loss: 0.33890439, Global Avg Loss: 1.21619242, Time: 0.0210 Steps: 36760, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000577, Sample Num: 9232, Cur Loss: 0.13713647, Cur Avg Loss: 0.24815032, Log Avg loss: 0.32071368, Global Avg Loss: 1.21594888, Time: 0.0210 Steps: 36770, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000587, Sample Num: 9392, Cur Loss: 0.38400835, Cur Avg Loss: 0.24904869, Log Avg loss: 0.30088473, Global Avg Loss: 1.21570009, Time: 0.0210 Steps: 36780, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000597, Sample Num: 9552, Cur Loss: 0.56701875, Cur Avg Loss: 0.24837363, Log Avg loss: 0.20874722, Global Avg Loss: 1.21542639, Time: 0.0209 Steps: 36790, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000607, Sample Num: 9712, Cur Loss: 0.12252283, Cur Avg Loss: 0.24793703, Log Avg loss: 0.22187195, Global Avg Loss: 1.21515640, Time: 0.0209 Steps: 36800, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000617, Sample Num: 9872, Cur Loss: 0.19239971, Cur Avg Loss: 0.24721411, Log Avg loss: 0.20333319, Global Avg Loss: 1.21488152, Time: 0.0210 Steps: 36810, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000627, Sample Num: 10032, Cur Loss: 0.33641827, Cur Avg Loss: 0.24750907, Log Avg loss: 0.26570820, Global Avg Loss: 1.21462374, Time: 0.0209 Steps: 36820, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000637, Sample Num: 10192, Cur Loss: 0.16174686, Cur Avg Loss: 0.25208588, Log Avg loss: 0.53905187, Global Avg Loss: 1.21444031, Time: 0.0211 Steps: 36830, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000647, Sample Num: 10352, Cur Loss: 0.17707467, Cur Avg Loss: 0.25441112, Log Avg loss: 0.40252864, Global Avg Loss: 1.21421992, Time: 0.0209 Steps: 36840, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000657, Sample Num: 10512, Cur Loss: 0.25053522, Cur Avg Loss: 0.25528831, Log Avg loss: 0.31204260, Global Avg Loss: 1.21397509, Time: 0.0211 Steps: 36850, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000667, Sample Num: 10672, Cur Loss: 0.51694101, Cur Avg Loss: 0.25492602, Log Avg loss: 0.23112370, Global Avg Loss: 1.21370845, Time: 0.0209 Steps: 36860, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000677, Sample Num: 10832, Cur Loss: 0.07387093, Cur Avg Loss: 0.25575560, Log Avg loss: 0.31108809, Global Avg Loss: 1.21346364, Time: 0.0209 Steps: 36870, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000687, Sample Num: 10992, Cur Loss: 0.25633705, Cur Avg Loss: 0.25558045, Log Avg loss: 0.24372296, Global Avg Loss: 1.21320069, Time: 0.0210 Steps: 36880, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000697, Sample Num: 11152, Cur Loss: 0.20081593, Cur Avg Loss: 0.25509557, Log Avg loss: 0.22178472, Global Avg Loss: 1.21293194, Time: 0.0210 Steps: 36890, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000707, Sample Num: 11312, Cur Loss: 0.46898925, Cur Avg Loss: 0.25453967, Log Avg loss: 0.21579303, Global Avg Loss: 1.21266172, Time: 0.0209 Steps: 36900, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000717, Sample Num: 11472, Cur Loss: 0.13804358, Cur Avg Loss: 0.25383946, Log Avg loss: 0.20433491, Global Avg Loss: 1.21238853, Time: 0.0209 Steps: 36910, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000727, Sample Num: 11632, Cur Loss: 0.28259856, Cur Avg Loss: 0.25327627, Log Avg loss: 0.21289529, Global Avg Loss: 1.21211781, Time: 0.0210 Steps: 36920, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000737, Sample Num: 11792, Cur Loss: 0.12399518, Cur Avg Loss: 0.25324877, Log Avg loss: 0.25124969, Global Avg Loss: 1.21185763, Time: 0.0210 Steps: 36930, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000747, Sample Num: 11952, Cur Loss: 0.09939785, Cur Avg Loss: 0.25205691, Log Avg loss: 0.16421645, Global Avg Loss: 1.21157402, Time: 0.0210 Steps: 36940, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000757, Sample Num: 12112, Cur Loss: 0.22436906, Cur Avg Loss: 0.25145084, Log Avg loss: 0.20617772, Global Avg Loss: 1.21130192, Time: 0.0210 Steps: 36950, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000767, Sample Num: 12272, Cur Loss: 0.18040361, Cur Avg Loss: 0.25207929, Log Avg loss: 0.29965283, Global Avg Loss: 1.21105526, Time: 0.0209 Steps: 36960, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000777, Sample Num: 12432, Cur Loss: 0.12897189, Cur Avg Loss: 0.25157522, Log Avg loss: 0.21291320, Global Avg Loss: 1.21078528, Time: 0.0209 Steps: 36970, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000787, Sample Num: 12592, Cur Loss: 0.34317118, Cur Avg Loss: 0.25048686, Log Avg loss: 0.16592129, Global Avg Loss: 1.21050273, Time: 0.0209 Steps: 36980, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000797, Sample Num: 12752, Cur Loss: 0.16344537, Cur Avg Loss: 0.24928439, Log Avg loss: 0.15464993, Global Avg Loss: 1.21021729, Time: 0.0209 Steps: 36990, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000807, Sample Num: 12912, Cur Loss: 0.22411667, Cur Avg Loss: 0.24948742, Log Avg loss: 0.26566873, Global Avg Loss: 1.20996200, Time: 0.0209 Steps: 37000, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000817, Sample Num: 13072, Cur Loss: 0.22197108, Cur Avg Loss: 0.24854829, Log Avg loss: 0.17276052, Global Avg Loss: 1.20968175, Time: 0.0209 Steps: 37010, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000827, Sample Num: 13232, Cur Loss: 0.33955401, Cur Avg Loss: 0.24779868, Log Avg loss: 0.18655573, Global Avg Loss: 1.20940538, Time: 0.0209 Steps: 37020, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000837, Sample Num: 13392, Cur Loss: 0.53623378, Cur Avg Loss: 0.24797201, Log Avg loss: 0.26230675, Global Avg Loss: 1.20914962, Time: 0.0209 Steps: 37030, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000847, Sample Num: 13552, Cur Loss: 0.47692204, Cur Avg Loss: 0.24803641, Log Avg loss: 0.25342673, Global Avg Loss: 1.20889159, Time: 0.0209 Steps: 37040, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000857, Sample Num: 13712, Cur Loss: 0.38368610, Cur Avg Loss: 0.24847065, Log Avg loss: 0.28525033, Global Avg Loss: 1.20864230, Time: 0.0209 Steps: 37050, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000867, Sample Num: 13872, Cur Loss: 0.20811152, Cur Avg Loss: 0.24934818, Log Avg loss: 0.32455285, Global Avg Loss: 1.20840374, Time: 0.0208 Steps: 37060, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000877, Sample Num: 14032, Cur Loss: 0.25231260, Cur Avg Loss: 0.24879949, Log Avg loss: 0.20122740, Global Avg Loss: 1.20813204, Time: 0.0209 Steps: 37070, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000887, Sample Num: 14192, Cur Loss: 0.41141969, Cur Avg Loss: 0.24978903, Log Avg loss: 0.33657211, Global Avg Loss: 1.20789700, Time: 0.0208 Steps: 37080, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000897, Sample Num: 14352, Cur Loss: 0.46542317, Cur Avg Loss: 0.24975143, Log Avg loss: 0.24641594, Global Avg Loss: 1.20763777, Time: 0.0209 Steps: 37090, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000907, Sample Num: 14512, Cur Loss: 0.13469809, Cur Avg Loss: 0.24934825, Log Avg loss: 0.21318358, Global Avg Loss: 1.20736972, Time: 0.0210 Steps: 37100, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000917, Sample Num: 14672, Cur Loss: 0.21204911, Cur Avg Loss: 0.24878274, Log Avg loss: 0.19749102, Global Avg Loss: 1.20709759, Time: 0.0209 Steps: 37110, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000927, Sample Num: 14832, Cur Loss: 0.34880280, Cur Avg Loss: 0.24904263, Log Avg loss: 0.27287449, Global Avg Loss: 1.20684591, Time: 0.0209 Steps: 37120, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000937, Sample Num: 14992, Cur Loss: 0.34190536, Cur Avg Loss: 0.24910891, Log Avg loss: 0.25525292, Global Avg Loss: 1.20658963, Time: 0.0209 Steps: 37130, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000947, Sample Num: 15152, Cur Loss: 0.19280587, Cur Avg Loss: 0.24947156, Log Avg loss: 0.28345159, Global Avg Loss: 1.20634107, Time: 0.0208 Steps: 37140, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000957, Sample Num: 15312, Cur Loss: 0.13101012, Cur Avg Loss: 0.24999931, Log Avg loss: 0.29997713, Global Avg Loss: 1.20609710, Time: 0.0210 Steps: 37150, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000967, Sample Num: 15472, Cur Loss: 0.17833766, Cur Avg Loss: 0.24908499, Log Avg loss: 0.16158462, Global Avg Loss: 1.20581601, Time: 0.0209 Steps: 37160, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000977, Sample Num: 15632, Cur Loss: 0.04971369, Cur Avg Loss: 0.24951350, Log Avg loss: 0.29095082, Global Avg Loss: 1.20556988, Time: 0.0209 Steps: 37170, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000987, Sample Num: 15792, Cur Loss: 0.43432367, Cur Avg Loss: 0.24963412, Log Avg loss: 0.26141852, Global Avg Loss: 1.20531594, Time: 0.0209 Steps: 37180, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000997, Sample Num: 15952, Cur Loss: 0.13593325, Cur Avg Loss: 0.24915694, Log Avg loss: 0.20205906, Global Avg Loss: 1.20504617, Time: 0.0209 Steps: 37190, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001007, Sample Num: 16112, Cur Loss: 0.22475219, Cur Avg Loss: 0.24929943, Log Avg loss: 0.26350591, Global Avg Loss: 1.20479307, Time: 0.0210 Steps: 37200, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001017, Sample Num: 16272, Cur Loss: 0.19579366, Cur Avg Loss: 0.24894403, Log Avg loss: 0.21315547, Global Avg Loss: 1.20452657, Time: 0.0210 Steps: 37210, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001027, Sample Num: 16432, Cur Loss: 0.25382242, Cur Avg Loss: 0.25025499, Log Avg loss: 0.38357933, Global Avg Loss: 1.20430601, Time: 0.0242 Steps: 37220, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001037, Sample Num: 16592, Cur Loss: 0.45385695, Cur Avg Loss: 0.25038755, Log Avg loss: 0.26400134, Global Avg Loss: 1.20405344, Time: 0.0208 Steps: 37230, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001047, Sample Num: 16752, Cur Loss: 0.08762959, Cur Avg Loss: 0.25148031, Log Avg loss: 0.36479947, Global Avg Loss: 1.20382808, Time: 0.0208 Steps: 37240, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001057, Sample Num: 16912, Cur Loss: 0.29606897, Cur Avg Loss: 0.25205941, Log Avg loss: 0.31269132, Global Avg Loss: 1.20358885, Time: 0.0208 Steps: 37250, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001067, Sample Num: 17072, Cur Loss: 0.15156698, Cur Avg Loss: 0.25130920, Log Avg loss: 0.17201236, Global Avg Loss: 1.20331199, Time: 0.0208 Steps: 37260, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001077, Sample Num: 17232, Cur Loss: 0.08908346, Cur Avg Loss: 0.25101095, Log Avg loss: 0.21918767, Global Avg Loss: 1.20304794, Time: 0.0208 Steps: 37270, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001087, Sample Num: 17392, Cur Loss: 0.13172686, Cur Avg Loss: 0.25076692, Log Avg loss: 0.22448445, Global Avg Loss: 1.20278545, Time: 0.0208 Steps: 37280, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001097, Sample Num: 17552, Cur Loss: 0.15543285, Cur Avg Loss: 0.25012724, Log Avg loss: 0.18059443, Global Avg Loss: 1.20251133, Time: 0.0208 Steps: 37290, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001107, Sample Num: 17712, Cur Loss: 0.36623248, Cur Avg Loss: 0.25006937, Log Avg loss: 0.24372074, Global Avg Loss: 1.20225428, Time: 0.0208 Steps: 37300, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001117, Sample Num: 17872, Cur Loss: 0.29864576, Cur Avg Loss: 0.25113212, Log Avg loss: 0.36877902, Global Avg Loss: 1.20203089, Time: 0.0208 Steps: 37310, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001127, Sample Num: 18032, Cur Loss: 0.62961984, Cur Avg Loss: 0.25306091, Log Avg loss: 0.46850599, Global Avg Loss: 1.20183434, Time: 0.0207 Steps: 37320, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001137, Sample Num: 18192, Cur Loss: 0.13161205, Cur Avg Loss: 0.25414154, Log Avg loss: 0.37592921, Global Avg Loss: 1.20161309, Time: 0.0207 Steps: 37330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001147, Sample Num: 18352, Cur Loss: 0.09259962, Cur Avg Loss: 0.25345960, Log Avg loss: 0.17592219, Global Avg Loss: 1.20133840, Time: 0.0208 Steps: 37340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001157, Sample Num: 18512, Cur Loss: 0.36368161, Cur Avg Loss: 0.25338106, Log Avg loss: 0.24437311, Global Avg Loss: 1.20108219, Time: 0.0207 Steps: 37350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001167, Sample Num: 18672, Cur Loss: 0.68325323, Cur Avg Loss: 0.25339856, Log Avg loss: 0.25542307, Global Avg Loss: 1.20082907, Time: 0.0207 Steps: 37360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001177, Sample Num: 18832, Cur Loss: 0.35704440, Cur Avg Loss: 0.25347010, Log Avg loss: 0.26181854, Global Avg Loss: 1.20057779, Time: 0.0208 Steps: 37370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001187, Sample Num: 18992, Cur Loss: 0.22183436, Cur Avg Loss: 0.25356226, Log Avg loss: 0.26440952, Global Avg Loss: 1.20032734, Time: 0.0208 Steps: 37380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001197, Sample Num: 19152, Cur Loss: 0.19868791, Cur Avg Loss: 0.25382433, Log Avg loss: 0.28493224, Global Avg Loss: 1.20008252, Time: 0.0208 Steps: 37390, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001207, Sample Num: 19312, Cur Loss: 0.10923137, Cur Avg Loss: 0.25379487, Log Avg loss: 0.25026900, Global Avg Loss: 1.19982856, Time: 0.0208 Steps: 37400, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001217, Sample Num: 19472, Cur Loss: 0.08504007, Cur Avg Loss: 0.25328871, Log Avg loss: 0.19219449, Global Avg Loss: 1.19955921, Time: 0.0208 Steps: 37410, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001227, Sample Num: 19632, Cur Loss: 0.14865145, Cur Avg Loss: 0.25325943, Log Avg loss: 0.24969667, Global Avg Loss: 1.19930537, Time: 0.0208 Steps: 37420, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001237, Sample Num: 19792, Cur Loss: 0.05757470, Cur Avg Loss: 0.25360366, Log Avg loss: 0.29584106, Global Avg Loss: 1.19906400, Time: 0.0209 Steps: 37430, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001247, Sample Num: 19952, Cur Loss: 0.08015610, Cur Avg Loss: 0.25384444, Log Avg loss: 0.28362829, Global Avg Loss: 1.19881949, Time: 0.0208 Steps: 37440, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001257, Sample Num: 20112, Cur Loss: 0.20548941, Cur Avg Loss: 0.25336732, Log Avg loss: 0.19387080, Global Avg Loss: 1.19855115, Time: 0.0207 Steps: 37450, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001267, Sample Num: 20272, Cur Loss: 0.20390475, Cur Avg Loss: 0.25257466, Log Avg loss: 0.15293727, Global Avg Loss: 1.19827202, Time: 0.0208 Steps: 37460, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001277, Sample Num: 20432, Cur Loss: 0.16844694, Cur Avg Loss: 0.25239234, Log Avg loss: 0.22929268, Global Avg Loss: 1.19801342, Time: 0.0207 Steps: 37470, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001287, Sample Num: 20592, Cur Loss: 0.35076165, Cur Avg Loss: 0.25215378, Log Avg loss: 0.22168920, Global Avg Loss: 1.19775293, Time: 0.0209 Steps: 37480, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001297, Sample Num: 20752, Cur Loss: 0.33166403, Cur Avg Loss: 0.25215508, Log Avg loss: 0.25232184, Global Avg Loss: 1.19750074, Time: 0.0208 Steps: 37490, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001307, Sample Num: 20912, Cur Loss: 0.04972689, Cur Avg Loss: 0.25195156, Log Avg loss: 0.22555551, Global Avg Loss: 1.19724156, Time: 0.0208 Steps: 37500, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001317, Sample Num: 21072, Cur Loss: 0.18762372, Cur Avg Loss: 0.25227908, Log Avg loss: 0.29508642, Global Avg Loss: 1.19700105, Time: 0.0208 Steps: 37510, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001327, Sample Num: 21232, Cur Loss: 0.38048917, Cur Avg Loss: 0.25240967, Log Avg loss: 0.26960819, Global Avg Loss: 1.19675387, Time: 0.0208 Steps: 37520, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001337, Sample Num: 21392, Cur Loss: 0.33117288, Cur Avg Loss: 0.25201584, Log Avg loss: 0.19975438, Global Avg Loss: 1.19648822, Time: 0.0208 Steps: 37530, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001347, Sample Num: 21552, Cur Loss: 0.16328810, Cur Avg Loss: 0.25243120, Log Avg loss: 0.30796431, Global Avg Loss: 1.19625153, Time: 0.0209 Steps: 37540, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001357, Sample Num: 21712, Cur Loss: 0.20594552, Cur Avg Loss: 0.25292243, Log Avg loss: 0.31909093, Global Avg Loss: 1.19601794, Time: 0.0208 Steps: 37550, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001367, Sample Num: 21872, Cur Loss: 0.11321475, Cur Avg Loss: 0.25248516, Log Avg loss: 0.19314884, Global Avg Loss: 1.19575093, Time: 0.0208 Steps: 37560, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001377, Sample Num: 22032, Cur Loss: 0.13481545, Cur Avg Loss: 0.25211102, Log Avg loss: 0.20096554, Global Avg Loss: 1.19548615, Time: 0.0208 Steps: 37570, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001387, Sample Num: 22192, Cur Loss: 0.20448843, Cur Avg Loss: 0.25195916, Log Avg loss: 0.23104752, Global Avg Loss: 1.19522951, Time: 0.0209 Steps: 37580, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001397, Sample Num: 22352, Cur Loss: 0.05997803, Cur Avg Loss: 0.25163185, Log Avg loss: 0.20623421, Global Avg Loss: 1.19496641, Time: 0.0208 Steps: 37590, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001407, Sample Num: 22512, Cur Loss: 0.35566688, Cur Avg Loss: 0.25221204, Log Avg loss: 0.33326529, Global Avg Loss: 1.19473724, Time: 0.0208 Steps: 37600, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001417, Sample Num: 22672, Cur Loss: 0.62047625, Cur Avg Loss: 0.25410917, Log Avg loss: 0.52103446, Global Avg Loss: 1.19455811, Time: 0.0208 Steps: 37610, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001427, Sample Num: 22832, Cur Loss: 0.39808744, Cur Avg Loss: 0.25479622, Log Avg loss: 0.35215197, Global Avg Loss: 1.19433418, Time: 0.0208 Steps: 37620, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001437, Sample Num: 22992, Cur Loss: 0.20131186, Cur Avg Loss: 0.25474995, Log Avg loss: 0.24814704, Global Avg Loss: 1.19408274, Time: 0.0208 Steps: 37630, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001447, Sample Num: 23152, Cur Loss: 0.85953689, Cur Avg Loss: 0.25518529, Log Avg loss: 0.31774326, Global Avg Loss: 1.19384992, Time: 0.0208 Steps: 37640, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001457, Sample Num: 23312, Cur Loss: 0.27696723, Cur Avg Loss: 0.25521143, Log Avg loss: 0.25899341, Global Avg Loss: 1.19360161, Time: 0.0208 Steps: 37650, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001467, Sample Num: 23472, Cur Loss: 0.28551900, Cur Avg Loss: 0.25547399, Log Avg loss: 0.29372937, Global Avg Loss: 1.19336267, Time: 0.0209 Steps: 37660, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001477, Sample Num: 23632, Cur Loss: 0.37620211, Cur Avg Loss: 0.25646761, Log Avg loss: 0.40223179, Global Avg Loss: 1.19315265, Time: 0.0208 Steps: 37670, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001487, Sample Num: 23792, Cur Loss: 0.40436441, Cur Avg Loss: 0.25665013, Log Avg loss: 0.28360853, Global Avg Loss: 1.19291127, Time: 0.0208 Steps: 37680, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001497, Sample Num: 23952, Cur Loss: 0.35406578, Cur Avg Loss: 0.25724833, Log Avg loss: 0.34620022, Global Avg Loss: 1.19268661, Time: 0.0208 Steps: 37690, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001507, Sample Num: 24112, Cur Loss: 0.11646240, Cur Avg Loss: 0.25728149, Log Avg loss: 0.26224557, Global Avg Loss: 1.19243981, Time: 0.0208 Steps: 37700, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001517, Sample Num: 24272, Cur Loss: 0.12430081, Cur Avg Loss: 0.25710758, Log Avg loss: 0.23090002, Global Avg Loss: 1.19218483, Time: 0.0208 Steps: 37710, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001527, Sample Num: 24432, Cur Loss: 0.24766763, Cur Avg Loss: 0.25719997, Log Avg loss: 0.27121501, Global Avg Loss: 1.19194067, Time: 0.0208 Steps: 37720, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001537, Sample Num: 24592, Cur Loss: 0.15854198, Cur Avg Loss: 0.25708286, Log Avg loss: 0.23920017, Global Avg Loss: 1.19168816, Time: 0.0245 Steps: 37730, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001547, Sample Num: 24752, Cur Loss: 0.16056621, Cur Avg Loss: 0.25700299, Log Avg loss: 0.24472726, Global Avg Loss: 1.19143724, Time: 0.0208 Steps: 37740, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001557, Sample Num: 24912, Cur Loss: 0.15580553, Cur Avg Loss: 0.25781138, Log Avg loss: 0.38286847, Global Avg Loss: 1.19122305, Time: 0.0208 Steps: 37750, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001567, Sample Num: 25072, Cur Loss: 0.13533327, Cur Avg Loss: 0.25773402, Log Avg loss: 0.24568930, Global Avg Loss: 1.19097264, Time: 0.0209 Steps: 37760, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001577, Sample Num: 25232, Cur Loss: 0.28213602, Cur Avg Loss: 0.25849228, Log Avg loss: 0.37731150, Global Avg Loss: 1.19075722, Time: 0.0208 Steps: 37770, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001587, Sample Num: 25392, Cur Loss: 0.13366438, Cur Avg Loss: 0.25870011, Log Avg loss: 0.29147616, Global Avg Loss: 1.19051919, Time: 0.0208 Steps: 37780, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001597, Sample Num: 25552, Cur Loss: 0.31743529, Cur Avg Loss: 0.25866662, Log Avg loss: 0.25335118, Global Avg Loss: 1.19027119, Time: 0.0208 Steps: 37790, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001607, Sample Num: 25712, Cur Loss: 0.21322249, Cur Avg Loss: 0.25842772, Log Avg loss: 0.22027583, Global Avg Loss: 1.19001458, Time: 0.0208 Steps: 37800, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001617, Sample Num: 25872, Cur Loss: 0.05640540, Cur Avg Loss: 0.25799397, Log Avg loss: 0.18828980, Global Avg Loss: 1.18974964, Time: 0.0208 Steps: 37810, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001627, Sample Num: 26032, Cur Loss: 0.09383158, Cur Avg Loss: 0.25749095, Log Avg loss: 0.17615244, Global Avg Loss: 1.18948164, Time: 0.0208 Steps: 37820, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001637, Sample Num: 26192, Cur Loss: 0.42114282, Cur Avg Loss: 0.25739275, Log Avg loss: 0.24141628, Global Avg Loss: 1.18923103, Time: 0.0208 Steps: 37830, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001647, Sample Num: 26352, Cur Loss: 0.26794422, Cur Avg Loss: 0.25680394, Log Avg loss: 0.16041510, Global Avg Loss: 1.18895914, Time: 0.0208 Steps: 37840, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001657, Sample Num: 26512, Cur Loss: 0.13356024, Cur Avg Loss: 0.25671191, Log Avg loss: 0.24155470, Global Avg Loss: 1.18870883, Time: 0.0209 Steps: 37850, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001667, Sample Num: 26672, Cur Loss: 0.12350018, Cur Avg Loss: 0.25644276, Log Avg loss: 0.21184429, Global Avg Loss: 1.18845081, Time: 0.0208 Steps: 37860, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001677, Sample Num: 26832, Cur Loss: 0.12946695, Cur Avg Loss: 0.25595825, Log Avg loss: 0.17519110, Global Avg Loss: 1.18818325, Time: 0.0208 Steps: 37870, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001687, Sample Num: 26992, Cur Loss: 0.24694316, Cur Avg Loss: 0.25574900, Log Avg loss: 0.22065686, Global Avg Loss: 1.18792783, Time: 0.0208 Steps: 37880, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001697, Sample Num: 27152, Cur Loss: 0.29717350, Cur Avg Loss: 0.25564893, Log Avg loss: 0.23876863, Global Avg Loss: 1.18767733, Time: 0.0208 Steps: 37890, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001707, Sample Num: 27312, Cur Loss: 0.14825004, Cur Avg Loss: 0.25560419, Log Avg loss: 0.24801056, Global Avg Loss: 1.18742940, Time: 0.0208 Steps: 37900, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001717, Sample Num: 27472, Cur Loss: 0.14468992, Cur Avg Loss: 0.25537589, Log Avg loss: 0.21640546, Global Avg Loss: 1.18717326, Time: 0.0208 Steps: 37910, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001727, Sample Num: 27632, Cur Loss: 0.28404382, Cur Avg Loss: 0.25553717, Log Avg loss: 0.28322943, Global Avg Loss: 1.18693487, Time: 0.0208 Steps: 37920, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001737, Sample Num: 27792, Cur Loss: 0.16740024, Cur Avg Loss: 0.25547722, Log Avg loss: 0.24512382, Global Avg Loss: 1.18668657, Time: 0.0208 Steps: 37930, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001747, Sample Num: 27952, Cur Loss: 0.24378678, Cur Avg Loss: 0.25554921, Log Avg loss: 0.26805400, Global Avg Loss: 1.18644444, Time: 0.0208 Steps: 37940, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001757, Sample Num: 28112, Cur Loss: 0.15609582, Cur Avg Loss: 0.25512212, Log Avg loss: 0.18050848, Global Avg Loss: 1.18617938, Time: 0.0208 Steps: 37950, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001767, Sample Num: 28272, Cur Loss: 0.25771663, Cur Avg Loss: 0.25480907, Log Avg loss: 0.19980642, Global Avg Loss: 1.18591953, Time: 0.0209 Steps: 37960, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001777, Sample Num: 28432, Cur Loss: 0.10940279, Cur Avg Loss: 0.25510845, Log Avg loss: 0.30800889, Global Avg Loss: 1.18568832, Time: 0.0208 Steps: 37970, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001787, Sample Num: 28592, Cur Loss: 0.57359946, Cur Avg Loss: 0.25692505, Log Avg loss: 0.57973478, Global Avg Loss: 1.18552877, Time: 0.0208 Steps: 37980, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001797, Sample Num: 28752, Cur Loss: 0.15319780, Cur Avg Loss: 0.25784233, Log Avg loss: 0.42175988, Global Avg Loss: 1.18532773, Time: 0.0245 Steps: 37990, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001807, Sample Num: 28912, Cur Loss: 0.25423071, Cur Avg Loss: 0.25787509, Log Avg loss: 0.26376261, Global Avg Loss: 1.18508521, Time: 0.0209 Steps: 38000, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001817, Sample Num: 29072, Cur Loss: 0.10029881, Cur Avg Loss: 0.25718827, Log Avg loss: 0.13308038, Global Avg Loss: 1.18480844, Time: 0.0208 Steps: 38010, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001827, Sample Num: 29232, Cur Loss: 0.11684197, Cur Avg Loss: 0.25686173, Log Avg loss: 0.19752921, Global Avg Loss: 1.18454877, Time: 0.0208 Steps: 38020, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001837, Sample Num: 29392, Cur Loss: 0.31974691, Cur Avg Loss: 0.25696237, Log Avg loss: 0.27534893, Global Avg Loss: 1.18430969, Time: 0.0208 Steps: 38030, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001847, Sample Num: 29552, Cur Loss: 0.19424590, Cur Avg Loss: 0.25712481, Log Avg loss: 0.28696444, Global Avg Loss: 1.18407380, Time: 0.0208 Steps: 38040, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001857, Sample Num: 29712, Cur Loss: 0.35511714, Cur Avg Loss: 0.25695683, Log Avg loss: 0.22593240, Global Avg Loss: 1.18382199, Time: 0.0209 Steps: 38050, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001867, Sample Num: 29872, Cur Loss: 0.16699506, Cur Avg Loss: 0.25713982, Log Avg loss: 0.29112109, Global Avg Loss: 1.18358744, Time: 0.0208 Steps: 38060, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001877, Sample Num: 30032, Cur Loss: 0.09964450, Cur Avg Loss: 0.25701932, Log Avg loss: 0.23452191, Global Avg Loss: 1.18333814, Time: 0.0208 Steps: 38070, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001887, Sample Num: 30192, Cur Loss: 0.18580872, Cur Avg Loss: 0.25707710, Log Avg loss: 0.26792208, Global Avg Loss: 1.18309775, Time: 0.0208 Steps: 38080, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001897, Sample Num: 30352, Cur Loss: 0.25255612, Cur Avg Loss: 0.25748551, Log Avg loss: 0.33455205, Global Avg Loss: 1.18287497, Time: 0.0208 Steps: 38090, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001907, Sample Num: 30512, Cur Loss: 0.22565398, Cur Avg Loss: 0.25736154, Log Avg loss: 0.23384447, Global Avg Loss: 1.18262588, Time: 0.0208 Steps: 38100, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001917, Sample Num: 30672, Cur Loss: 0.32991147, Cur Avg Loss: 0.25692986, Log Avg loss: 0.17460934, Global Avg Loss: 1.18236138, Time: 0.0208 Steps: 38110, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001927, Sample Num: 30832, Cur Loss: 0.82569420, Cur Avg Loss: 0.25720901, Log Avg loss: 0.31072080, Global Avg Loss: 1.18213273, Time: 0.0208 Steps: 38120, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001937, Sample Num: 30992, Cur Loss: 0.22387078, Cur Avg Loss: 0.25783252, Log Avg loss: 0.37798293, Global Avg Loss: 1.18192183, Time: 0.0208 Steps: 38130, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001947, Sample Num: 31152, Cur Loss: 0.13878702, Cur Avg Loss: 0.25802003, Log Avg loss: 0.29434108, Global Avg Loss: 1.18168911, Time: 0.0208 Steps: 38140, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001957, Sample Num: 31312, Cur Loss: 0.20274360, Cur Avg Loss: 0.25859363, Log Avg loss: 0.37027273, Global Avg Loss: 1.18147642, Time: 0.0208 Steps: 38150, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001967, Sample Num: 31472, Cur Loss: 0.16080168, Cur Avg Loss: 0.25905483, Log Avg loss: 0.34931257, Global Avg Loss: 1.18125835, Time: 0.0208 Steps: 38160, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001977, Sample Num: 31632, Cur Loss: 0.34882677, Cur Avg Loss: 0.25897363, Log Avg loss: 0.24300249, Global Avg Loss: 1.18101254, Time: 0.0208 Steps: 38170, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001987, Sample Num: 31792, Cur Loss: 0.06272990, Cur Avg Loss: 0.25897053, Log Avg loss: 0.25835749, Global Avg Loss: 1.18077088, Time: 0.0208 Steps: 38180, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001997, Sample Num: 31952, Cur Loss: 0.20033924, Cur Avg Loss: 0.25910990, Log Avg loss: 0.28680241, Global Avg Loss: 1.18053680, Time: 0.0208 Steps: 38190, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002007, Sample Num: 32112, Cur Loss: 0.17210940, Cur Avg Loss: 0.25870759, Log Avg loss: 0.17836553, Global Avg Loss: 1.18027445, Time: 0.0208 Steps: 38200, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002017, Sample Num: 32272, Cur Loss: 0.22191688, Cur Avg Loss: 0.25856357, Log Avg loss: 0.22965838, Global Avg Loss: 1.18002566, Time: 0.0208 Steps: 38210, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002027, Sample Num: 32432, Cur Loss: 0.42309958, Cur Avg Loss: 0.25861339, Log Avg loss: 0.26866279, Global Avg Loss: 1.17978721, Time: 0.0208 Steps: 38220, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002037, Sample Num: 32592, Cur Loss: 0.44422507, Cur Avg Loss: 0.25848321, Log Avg loss: 0.23209489, Global Avg Loss: 1.17953932, Time: 0.0208 Steps: 38230, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002047, Sample Num: 32752, Cur Loss: 0.25678965, Cur Avg Loss: 0.25843884, Log Avg loss: 0.24940089, Global Avg Loss: 1.17929608, Time: 0.0208 Steps: 38240, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002057, Sample Num: 32912, Cur Loss: 0.06828845, Cur Avg Loss: 0.25817590, Log Avg loss: 0.20435281, Global Avg Loss: 1.17904119, Time: 0.0211 Steps: 38250, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002067, Sample Num: 33072, Cur Loss: 0.21160001, Cur Avg Loss: 0.25824469, Log Avg loss: 0.27239528, Global Avg Loss: 1.17880422, Time: 0.0210 Steps: 38260, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002077, Sample Num: 33232, Cur Loss: 0.16013975, Cur Avg Loss: 0.25798840, Log Avg loss: 0.20501284, Global Avg Loss: 1.17854977, Time: 0.0210 Steps: 38270, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002087, Sample Num: 33392, Cur Loss: 0.77137601, Cur Avg Loss: 0.25813923, Log Avg loss: 0.28946740, Global Avg Loss: 1.17831751, Time: 0.0210 Steps: 38280, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002097, Sample Num: 33552, Cur Loss: 0.08279189, Cur Avg Loss: 0.25799857, Log Avg loss: 0.22864185, Global Avg Loss: 1.17806949, Time: 0.0211 Steps: 38290, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002107, Sample Num: 33712, Cur Loss: 0.76322973, Cur Avg Loss: 0.25774389, Log Avg loss: 0.20433828, Global Avg Loss: 1.17781525, Time: 0.0210 Steps: 38300, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002117, Sample Num: 33872, Cur Loss: 0.16549546, Cur Avg Loss: 0.25785261, Log Avg loss: 0.28075821, Global Avg Loss: 1.17758109, Time: 0.0210 Steps: 38310, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002127, Sample Num: 34032, Cur Loss: 0.26940370, Cur Avg Loss: 0.25755198, Log Avg loss: 0.19390871, Global Avg Loss: 1.17732439, Time: 0.0210 Steps: 38320, Updated lr: 0.000065 ***** Running evaluation checkpoint-38322 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-38322 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.698847, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.390025, "eval_total_loss": 274.187714, "eval_mae": 0.523311, "eval_mse": 0.390059, "eval_r2": 0.752053, "eval_sp_statistic": 0.903774, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.923103, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.252137, "test_total_loss": 126.572757, "test_mae": 0.402729, "test_mse": 0.252215, "test_r2": 0.837219, "test_sp_statistic": 0.898416, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.941692, "test_ps_pvalue": 0.0, "lr": 6.460692271218587e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.177279595523017, "train_cur_epoch_loss": 548.4509061798453, "train_cur_epoch_avg_loss": 0.25760963183647034, "train_cur_epoch_time": 44.6988468170166, "train_cur_epoch_avg_time": 0.02099523100846247, "epoch": 18, "step": 38322} ################################################## Training, Epoch: 0019, Batch: 000008, Sample Num: 128, Cur Loss: 0.18724848, Cur Avg Loss: 0.25846581, Log Avg loss: 0.27055800, Global Avg Loss: 1.17708783, Time: 0.0211 Steps: 38330, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000018, Sample Num: 288, Cur Loss: 0.11854817, Cur Avg Loss: 0.25764814, Log Avg loss: 0.25699401, Global Avg Loss: 1.17684784, Time: 0.0209 Steps: 38340, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000028, Sample Num: 448, Cur Loss: 0.16494469, Cur Avg Loss: 0.26800041, Log Avg loss: 0.28663449, Global Avg Loss: 1.17661572, Time: 0.0208 Steps: 38350, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000038, Sample Num: 608, Cur Loss: 0.18540661, Cur Avg Loss: 0.26854268, Log Avg loss: 0.27006105, Global Avg Loss: 1.17637939, Time: 0.0209 Steps: 38360, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000048, Sample Num: 768, Cur Loss: 0.10621770, Cur Avg Loss: 0.26454492, Log Avg loss: 0.24935342, Global Avg Loss: 1.17613779, Time: 0.0208 Steps: 38370, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000058, Sample Num: 928, Cur Loss: 0.19979967, Cur Avg Loss: 0.25267103, Log Avg loss: 0.19567637, Global Avg Loss: 1.17588232, Time: 0.0209 Steps: 38380, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000068, Sample Num: 1088, Cur Loss: 0.52840263, Cur Avg Loss: 0.24489163, Log Avg loss: 0.19977108, Global Avg Loss: 1.17562806, Time: 0.0209 Steps: 38390, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000078, Sample Num: 1248, Cur Loss: 0.14999393, Cur Avg Loss: 0.23721799, Log Avg loss: 0.18503728, Global Avg Loss: 1.17537010, Time: 0.0209 Steps: 38400, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000088, Sample Num: 1408, Cur Loss: 0.24742442, Cur Avg Loss: 0.24771139, Log Avg loss: 0.32955987, Global Avg Loss: 1.17514989, Time: 0.0208 Steps: 38410, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000098, Sample Num: 1568, Cur Loss: 0.19999328, Cur Avg Loss: 0.24655669, Log Avg loss: 0.23639533, Global Avg Loss: 1.17490555, Time: 0.0209 Steps: 38420, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000108, Sample Num: 1728, Cur Loss: 0.24554020, Cur Avg Loss: 0.25848408, Log Avg loss: 0.37537255, Global Avg Loss: 1.17469750, Time: 0.0208 Steps: 38430, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000118, Sample Num: 1888, Cur Loss: 0.30955762, Cur Avg Loss: 0.26756913, Log Avg loss: 0.36568764, Global Avg Loss: 1.17448704, Time: 0.0209 Steps: 38440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000128, Sample Num: 2048, Cur Loss: 0.19761658, Cur Avg Loss: 0.26447545, Log Avg loss: 0.22796999, Global Avg Loss: 1.17424087, Time: 0.0210 Steps: 38450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000138, Sample Num: 2208, Cur Loss: 0.10037403, Cur Avg Loss: 0.25800205, Log Avg loss: 0.17514251, Global Avg Loss: 1.17398110, Time: 0.0208 Steps: 38460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000148, Sample Num: 2368, Cur Loss: 0.14365909, Cur Avg Loss: 0.25454347, Log Avg loss: 0.20681508, Global Avg Loss: 1.17372969, Time: 0.0209 Steps: 38470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000158, Sample Num: 2528, Cur Loss: 0.35837033, Cur Avg Loss: 0.25151048, Log Avg loss: 0.20662231, Global Avg Loss: 1.17347836, Time: 0.0209 Steps: 38480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000168, Sample Num: 2688, Cur Loss: 0.10645016, Cur Avg Loss: 0.24913593, Log Avg loss: 0.21161796, Global Avg Loss: 1.17322846, Time: 0.0208 Steps: 38490, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000178, Sample Num: 2848, Cur Loss: 0.06991878, Cur Avg Loss: 0.24662316, Log Avg loss: 0.20440862, Global Avg Loss: 1.17297682, Time: 0.0209 Steps: 38500, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000188, Sample Num: 3008, Cur Loss: 0.05964214, Cur Avg Loss: 0.24287869, Log Avg loss: 0.17622720, Global Avg Loss: 1.17271799, Time: 0.0210 Steps: 38510, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000198, Sample Num: 3168, Cur Loss: 0.06803124, Cur Avg Loss: 0.24060058, Log Avg loss: 0.19777201, Global Avg Loss: 1.17246489, Time: 0.0209 Steps: 38520, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000208, Sample Num: 3328, Cur Loss: 0.08036168, Cur Avg Loss: 0.23617665, Log Avg loss: 0.14858282, Global Avg Loss: 1.17219915, Time: 0.0210 Steps: 38530, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000218, Sample Num: 3488, Cur Loss: 0.38235241, Cur Avg Loss: 0.23840319, Log Avg loss: 0.28471525, Global Avg Loss: 1.17196888, Time: 0.0210 Steps: 38540, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000228, Sample Num: 3648, Cur Loss: 0.37167162, Cur Avg Loss: 0.23958517, Log Avg loss: 0.26535233, Global Avg Loss: 1.17173370, Time: 0.0209 Steps: 38550, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000238, Sample Num: 3808, Cur Loss: 0.28041321, Cur Avg Loss: 0.23734824, Log Avg loss: 0.18634628, Global Avg Loss: 1.17147815, Time: 0.0209 Steps: 38560, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000248, Sample Num: 3968, Cur Loss: 0.22397472, Cur Avg Loss: 0.23553750, Log Avg loss: 0.19244198, Global Avg Loss: 1.17122432, Time: 0.0208 Steps: 38570, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000258, Sample Num: 4128, Cur Loss: 0.10178852, Cur Avg Loss: 0.23385775, Log Avg loss: 0.19219987, Global Avg Loss: 1.17097055, Time: 0.0246 Steps: 38580, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000268, Sample Num: 4288, Cur Loss: 0.17489484, Cur Avg Loss: 0.23165270, Log Avg loss: 0.17476234, Global Avg Loss: 1.17071240, Time: 0.0209 Steps: 38590, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000278, Sample Num: 4448, Cur Loss: 0.27154726, Cur Avg Loss: 0.23152631, Log Avg loss: 0.22813917, Global Avg Loss: 1.17046821, Time: 0.0209 Steps: 38600, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000288, Sample Num: 4608, Cur Loss: 0.07118306, Cur Avg Loss: 0.23019721, Log Avg loss: 0.19324820, Global Avg Loss: 1.17021511, Time: 0.0209 Steps: 38610, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000298, Sample Num: 4768, Cur Loss: 0.34852085, Cur Avg Loss: 0.23103044, Log Avg loss: 0.25502740, Global Avg Loss: 1.16997814, Time: 0.0209 Steps: 38620, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000308, Sample Num: 4928, Cur Loss: 0.32429469, Cur Avg Loss: 0.22950596, Log Avg loss: 0.18407638, Global Avg Loss: 1.16972292, Time: 0.0210 Steps: 38630, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000318, Sample Num: 5088, Cur Loss: 0.64865160, Cur Avg Loss: 0.23883557, Log Avg loss: 0.52618775, Global Avg Loss: 1.16955638, Time: 0.0209 Steps: 38640, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000328, Sample Num: 5248, Cur Loss: 0.13211688, Cur Avg Loss: 0.24350019, Log Avg loss: 0.39183495, Global Avg Loss: 1.16935515, Time: 0.0210 Steps: 38650, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000338, Sample Num: 5408, Cur Loss: 0.21915215, Cur Avg Loss: 0.24196728, Log Avg loss: 0.19168803, Global Avg Loss: 1.16910227, Time: 0.0209 Steps: 38660, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000348, Sample Num: 5568, Cur Loss: 0.80024409, Cur Avg Loss: 0.24304213, Log Avg loss: 0.27937189, Global Avg Loss: 1.16887218, Time: 0.0210 Steps: 38670, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000358, Sample Num: 5728, Cur Loss: 0.17390871, Cur Avg Loss: 0.24771931, Log Avg loss: 0.41048529, Global Avg Loss: 1.16867612, Time: 0.0210 Steps: 38680, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000368, Sample Num: 5888, Cur Loss: 0.04896979, Cur Avg Loss: 0.24902885, Log Avg loss: 0.29591040, Global Avg Loss: 1.16845054, Time: 0.0209 Steps: 38690, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000378, Sample Num: 6048, Cur Loss: 0.33760720, Cur Avg Loss: 0.24896610, Log Avg loss: 0.24665666, Global Avg Loss: 1.16821235, Time: 0.0209 Steps: 38700, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000388, Sample Num: 6208, Cur Loss: 0.53861982, Cur Avg Loss: 0.25022971, Log Avg loss: 0.29799444, Global Avg Loss: 1.16798754, Time: 0.0210 Steps: 38710, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000398, Sample Num: 6368, Cur Loss: 0.26598895, Cur Avg Loss: 0.24991648, Log Avg loss: 0.23776299, Global Avg Loss: 1.16774730, Time: 0.0210 Steps: 38720, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000408, Sample Num: 6528, Cur Loss: 0.36187953, Cur Avg Loss: 0.24953713, Log Avg loss: 0.23443912, Global Avg Loss: 1.16750632, Time: 0.0209 Steps: 38730, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000418, Sample Num: 6688, Cur Loss: 0.18443033, Cur Avg Loss: 0.24850640, Log Avg loss: 0.20645245, Global Avg Loss: 1.16725824, Time: 0.0209 Steps: 38740, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000428, Sample Num: 6848, Cur Loss: 0.25643370, Cur Avg Loss: 0.24903021, Log Avg loss: 0.27092576, Global Avg Loss: 1.16702693, Time: 0.0209 Steps: 38750, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000438, Sample Num: 7008, Cur Loss: 0.15213194, Cur Avg Loss: 0.24958994, Log Avg loss: 0.27354627, Global Avg Loss: 1.16679642, Time: 0.0209 Steps: 38760, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000448, Sample Num: 7168, Cur Loss: 0.20886120, Cur Avg Loss: 0.24894079, Log Avg loss: 0.22050805, Global Avg Loss: 1.16655234, Time: 0.0209 Steps: 38770, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000458, Sample Num: 7328, Cur Loss: 0.29252294, Cur Avg Loss: 0.24957004, Log Avg loss: 0.27776048, Global Avg Loss: 1.16632315, Time: 0.0209 Steps: 38780, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000468, Sample Num: 7488, Cur Loss: 0.08194800, Cur Avg Loss: 0.24974869, Log Avg loss: 0.25793086, Global Avg Loss: 1.16608897, Time: 0.0209 Steps: 38790, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000478, Sample Num: 7648, Cur Loss: 0.10250171, Cur Avg Loss: 0.24859824, Log Avg loss: 0.19475689, Global Avg Loss: 1.16583862, Time: 0.0209 Steps: 38800, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000488, Sample Num: 7808, Cur Loss: 0.08662681, Cur Avg Loss: 0.24736052, Log Avg loss: 0.18819742, Global Avg Loss: 1.16558672, Time: 0.0209 Steps: 38810, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000498, Sample Num: 7968, Cur Loss: 0.08522037, Cur Avg Loss: 0.24749277, Log Avg loss: 0.25394673, Global Avg Loss: 1.16535188, Time: 0.0209 Steps: 38820, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000508, Sample Num: 8128, Cur Loss: 0.10164656, Cur Avg Loss: 0.24614267, Log Avg loss: 0.17890759, Global Avg Loss: 1.16509784, Time: 0.0209 Steps: 38830, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000518, Sample Num: 8288, Cur Loss: 0.15521605, Cur Avg Loss: 0.24512539, Log Avg loss: 0.19344761, Global Avg Loss: 1.16484767, Time: 0.0245 Steps: 38840, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000528, Sample Num: 8448, Cur Loss: 0.26126081, Cur Avg Loss: 0.24421941, Log Avg loss: 0.19728972, Global Avg Loss: 1.16459862, Time: 0.0209 Steps: 38850, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000538, Sample Num: 8608, Cur Loss: 0.16920193, Cur Avg Loss: 0.24351570, Log Avg loss: 0.20636004, Global Avg Loss: 1.16435204, Time: 0.0208 Steps: 38860, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000548, Sample Num: 8768, Cur Loss: 0.14934425, Cur Avg Loss: 0.24358002, Log Avg loss: 0.24704035, Global Avg Loss: 1.16411604, Time: 0.0208 Steps: 38870, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000558, Sample Num: 8928, Cur Loss: 0.13646638, Cur Avg Loss: 0.24382370, Log Avg loss: 0.25717733, Global Avg Loss: 1.16388277, Time: 0.0209 Steps: 38880, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000568, Sample Num: 9088, Cur Loss: 0.22152528, Cur Avg Loss: 0.24303082, Log Avg loss: 0.19878795, Global Avg Loss: 1.16363461, Time: 0.0208 Steps: 38890, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000578, Sample Num: 9248, Cur Loss: 0.07344224, Cur Avg Loss: 0.24218291, Log Avg loss: 0.19402148, Global Avg Loss: 1.16338536, Time: 0.0208 Steps: 38900, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000588, Sample Num: 9408, Cur Loss: 0.29242241, Cur Avg Loss: 0.24184147, Log Avg loss: 0.22210676, Global Avg Loss: 1.16314345, Time: 0.0208 Steps: 38910, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000598, Sample Num: 9568, Cur Loss: 0.18101403, Cur Avg Loss: 0.24122657, Log Avg loss: 0.20506997, Global Avg Loss: 1.16289728, Time: 0.0208 Steps: 38920, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000608, Sample Num: 9728, Cur Loss: 0.09744306, Cur Avg Loss: 0.24245407, Log Avg loss: 0.31585860, Global Avg Loss: 1.16267970, Time: 0.0208 Steps: 38930, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000618, Sample Num: 9888, Cur Loss: 0.09512228, Cur Avg Loss: 0.24264425, Log Avg loss: 0.25420735, Global Avg Loss: 1.16244640, Time: 0.0209 Steps: 38940, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000628, Sample Num: 10048, Cur Loss: 0.28800428, Cur Avg Loss: 0.24259180, Log Avg loss: 0.23935040, Global Avg Loss: 1.16220940, Time: 0.0209 Steps: 38950, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000638, Sample Num: 10208, Cur Loss: 0.09864280, Cur Avg Loss: 0.24241735, Log Avg loss: 0.23146161, Global Avg Loss: 1.16197051, Time: 0.0209 Steps: 38960, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000648, Sample Num: 10368, Cur Loss: 0.24659911, Cur Avg Loss: 0.24193989, Log Avg loss: 0.21147799, Global Avg Loss: 1.16172660, Time: 0.0209 Steps: 38970, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000658, Sample Num: 10528, Cur Loss: 0.12992953, Cur Avg Loss: 0.24082499, Log Avg loss: 0.16857999, Global Avg Loss: 1.16147182, Time: 0.0209 Steps: 38980, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000668, Sample Num: 10688, Cur Loss: 0.24606326, Cur Avg Loss: 0.24105370, Log Avg loss: 0.25610241, Global Avg Loss: 1.16123961, Time: 0.0209 Steps: 38990, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000678, Sample Num: 10848, Cur Loss: 0.17845294, Cur Avg Loss: 0.24039943, Log Avg loss: 0.19669424, Global Avg Loss: 1.16099229, Time: 0.0208 Steps: 39000, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000688, Sample Num: 11008, Cur Loss: 0.55547059, Cur Avg Loss: 0.24115029, Log Avg loss: 0.29205886, Global Avg Loss: 1.16076955, Time: 0.0209 Steps: 39010, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000698, Sample Num: 11168, Cur Loss: 0.02881691, Cur Avg Loss: 0.24254612, Log Avg loss: 0.33857875, Global Avg Loss: 1.16055884, Time: 0.0209 Steps: 39020, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000708, Sample Num: 11328, Cur Loss: 0.22529343, Cur Avg Loss: 0.24755616, Log Avg loss: 0.59725731, Global Avg Loss: 1.16041451, Time: 0.0209 Steps: 39030, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000718, Sample Num: 11488, Cur Loss: 0.08884721, Cur Avg Loss: 0.24910507, Log Avg loss: 0.35876804, Global Avg Loss: 1.16020917, Time: 0.0209 Steps: 39040, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000728, Sample Num: 11648, Cur Loss: 0.15903530, Cur Avg Loss: 0.24882207, Log Avg loss: 0.22850232, Global Avg Loss: 1.15997058, Time: 0.0209 Steps: 39050, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000738, Sample Num: 11808, Cur Loss: 0.24589089, Cur Avg Loss: 0.24798815, Log Avg loss: 0.18727870, Global Avg Loss: 1.15972155, Time: 0.0209 Steps: 39060, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000748, Sample Num: 11968, Cur Loss: 0.05663495, Cur Avg Loss: 0.24675663, Log Avg loss: 0.15587079, Global Avg Loss: 1.15946462, Time: 0.0209 Steps: 39070, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000758, Sample Num: 12128, Cur Loss: 0.37552822, Cur Avg Loss: 0.24902239, Log Avg loss: 0.41850090, Global Avg Loss: 1.15927502, Time: 0.0208 Steps: 39080, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000768, Sample Num: 12288, Cur Loss: 0.25117749, Cur Avg Loss: 0.24843541, Log Avg loss: 0.20394228, Global Avg Loss: 1.15903062, Time: 0.0254 Steps: 39090, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000778, Sample Num: 12448, Cur Loss: 0.13272741, Cur Avg Loss: 0.24738699, Log Avg loss: 0.16686870, Global Avg Loss: 1.15877687, Time: 0.0209 Steps: 39100, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000788, Sample Num: 12608, Cur Loss: 0.26337364, Cur Avg Loss: 0.24929885, Log Avg loss: 0.39804177, Global Avg Loss: 1.15858236, Time: 0.0209 Steps: 39110, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000798, Sample Num: 12768, Cur Loss: 0.30961314, Cur Avg Loss: 0.25133529, Log Avg loss: 0.41180641, Global Avg Loss: 1.15839147, Time: 0.0209 Steps: 39120, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000808, Sample Num: 12928, Cur Loss: 0.62725055, Cur Avg Loss: 0.25126495, Log Avg loss: 0.24565211, Global Avg Loss: 1.15815821, Time: 0.0208 Steps: 39130, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000818, Sample Num: 13088, Cur Loss: 0.15979907, Cur Avg Loss: 0.25332284, Log Avg loss: 0.41960005, Global Avg Loss: 1.15796951, Time: 0.0209 Steps: 39140, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000828, Sample Num: 13248, Cur Loss: 0.20127785, Cur Avg Loss: 0.25425849, Log Avg loss: 0.33079438, Global Avg Loss: 1.15775823, Time: 0.0209 Steps: 39150, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000838, Sample Num: 13408, Cur Loss: 0.17187561, Cur Avg Loss: 0.25557464, Log Avg loss: 0.36455242, Global Avg Loss: 1.15755567, Time: 0.0209 Steps: 39160, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000848, Sample Num: 13568, Cur Loss: 0.10572853, Cur Avg Loss: 0.25492299, Log Avg loss: 0.20031478, Global Avg Loss: 1.15731129, Time: 0.0208 Steps: 39170, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000858, Sample Num: 13728, Cur Loss: 0.25076801, Cur Avg Loss: 0.25416663, Log Avg loss: 0.19002728, Global Avg Loss: 1.15706441, Time: 0.0208 Steps: 39180, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000868, Sample Num: 13888, Cur Loss: 0.26285017, Cur Avg Loss: 0.25433589, Log Avg loss: 0.26885819, Global Avg Loss: 1.15683777, Time: 0.0208 Steps: 39190, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000878, Sample Num: 14048, Cur Loss: 0.31839114, Cur Avg Loss: 0.25406710, Log Avg loss: 0.23073648, Global Avg Loss: 1.15660152, Time: 0.0210 Steps: 39200, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000888, Sample Num: 14208, Cur Loss: 0.27901217, Cur Avg Loss: 0.25342752, Log Avg loss: 0.19727236, Global Avg Loss: 1.15635686, Time: 0.0210 Steps: 39210, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000898, Sample Num: 14368, Cur Loss: 0.29708916, Cur Avg Loss: 0.25361852, Log Avg loss: 0.27057894, Global Avg Loss: 1.15613101, Time: 0.0209 Steps: 39220, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000908, Sample Num: 14528, Cur Loss: 0.42498475, Cur Avg Loss: 0.25429430, Log Avg loss: 0.31497898, Global Avg Loss: 1.15591659, Time: 0.0210 Steps: 39230, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000918, Sample Num: 14688, Cur Loss: 0.15224314, Cur Avg Loss: 0.25440249, Log Avg loss: 0.26422688, Global Avg Loss: 1.15568935, Time: 0.0209 Steps: 39240, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000928, Sample Num: 14848, Cur Loss: 0.27572760, Cur Avg Loss: 0.25389472, Log Avg loss: 0.20728102, Global Avg Loss: 1.15544772, Time: 0.0210 Steps: 39250, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000938, Sample Num: 15008, Cur Loss: 0.15130113, Cur Avg Loss: 0.25408454, Log Avg loss: 0.27169947, Global Avg Loss: 1.15522262, Time: 0.0209 Steps: 39260, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000948, Sample Num: 15168, Cur Loss: 0.30760151, Cur Avg Loss: 0.25332086, Log Avg loss: 0.18168828, Global Avg Loss: 1.15497471, Time: 0.0208 Steps: 39270, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000958, Sample Num: 15328, Cur Loss: 0.18844590, Cur Avg Loss: 0.25310816, Log Avg loss: 0.23294392, Global Avg Loss: 1.15473998, Time: 0.0208 Steps: 39280, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000968, Sample Num: 15488, Cur Loss: 0.41955107, Cur Avg Loss: 0.25307192, Log Avg loss: 0.24960033, Global Avg Loss: 1.15450960, Time: 0.0208 Steps: 39290, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000978, Sample Num: 15648, Cur Loss: 0.22158036, Cur Avg Loss: 0.25387946, Log Avg loss: 0.33204962, Global Avg Loss: 1.15430033, Time: 0.0208 Steps: 39300, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000988, Sample Num: 15808, Cur Loss: 0.18916149, Cur Avg Loss: 0.25385752, Log Avg loss: 0.25171119, Global Avg Loss: 1.15407072, Time: 0.0208 Steps: 39310, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000998, Sample Num: 15968, Cur Loss: 0.10053948, Cur Avg Loss: 0.25318963, Log Avg loss: 0.18720229, Global Avg Loss: 1.15382482, Time: 0.0208 Steps: 39320, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001008, Sample Num: 16128, Cur Loss: 0.30333406, Cur Avg Loss: 0.25262924, Log Avg loss: 0.19670229, Global Avg Loss: 1.15358146, Time: 0.0208 Steps: 39330, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001018, Sample Num: 16288, Cur Loss: 0.16219433, Cur Avg Loss: 0.25212822, Log Avg loss: 0.20162503, Global Avg Loss: 1.15333948, Time: 0.0208 Steps: 39340, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001028, Sample Num: 16448, Cur Loss: 0.20651460, Cur Avg Loss: 0.25201613, Log Avg loss: 0.24060552, Global Avg Loss: 1.15310753, Time: 0.0227 Steps: 39350, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001038, Sample Num: 16608, Cur Loss: 0.30861533, Cur Avg Loss: 0.25245367, Log Avg loss: 0.29743344, Global Avg Loss: 1.15289013, Time: 0.0208 Steps: 39360, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001048, Sample Num: 16768, Cur Loss: 0.17745978, Cur Avg Loss: 0.25321314, Log Avg loss: 0.33204535, Global Avg Loss: 1.15268164, Time: 0.0208 Steps: 39370, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001058, Sample Num: 16928, Cur Loss: 0.65879482, Cur Avg Loss: 0.25368306, Log Avg loss: 0.30293073, Global Avg Loss: 1.15246585, Time: 0.0208 Steps: 39380, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001068, Sample Num: 17088, Cur Loss: 0.33117354, Cur Avg Loss: 0.25402788, Log Avg loss: 0.29051001, Global Avg Loss: 1.15224703, Time: 0.0208 Steps: 39390, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001078, Sample Num: 17248, Cur Loss: 0.39088541, Cur Avg Loss: 0.25462045, Log Avg loss: 0.31790688, Global Avg Loss: 1.15203527, Time: 0.0208 Steps: 39400, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001088, Sample Num: 17408, Cur Loss: 0.13851607, Cur Avg Loss: 0.25418368, Log Avg loss: 0.20710010, Global Avg Loss: 1.15179550, Time: 0.0208 Steps: 39410, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001098, Sample Num: 17568, Cur Loss: 0.12905730, Cur Avg Loss: 0.25411607, Log Avg loss: 0.24675956, Global Avg Loss: 1.15156591, Time: 0.0209 Steps: 39420, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001108, Sample Num: 17728, Cur Loss: 0.10180767, Cur Avg Loss: 0.25478534, Log Avg loss: 0.32827169, Global Avg Loss: 1.15135711, Time: 0.0208 Steps: 39430, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001118, Sample Num: 17888, Cur Loss: 0.30650929, Cur Avg Loss: 0.25453567, Log Avg loss: 0.22687215, Global Avg Loss: 1.15112271, Time: 0.0208 Steps: 39440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001128, Sample Num: 18048, Cur Loss: 0.19925985, Cur Avg Loss: 0.25432759, Log Avg loss: 0.23106413, Global Avg Loss: 1.15088948, Time: 0.0208 Steps: 39450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001138, Sample Num: 18208, Cur Loss: 0.78104460, Cur Avg Loss: 0.25492605, Log Avg loss: 0.32243254, Global Avg Loss: 1.15067954, Time: 0.0208 Steps: 39460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001148, Sample Num: 18368, Cur Loss: 0.41020548, Cur Avg Loss: 0.25567944, Log Avg loss: 0.34141493, Global Avg Loss: 1.15047450, Time: 0.0208 Steps: 39470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001158, Sample Num: 18528, Cur Loss: 0.15468754, Cur Avg Loss: 0.25616565, Log Avg loss: 0.31198257, Global Avg Loss: 1.15026212, Time: 0.0208 Steps: 39480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001168, Sample Num: 18688, Cur Loss: 0.10864379, Cur Avg Loss: 0.25591724, Log Avg loss: 0.22715193, Global Avg Loss: 1.15002836, Time: 0.0208 Steps: 39490, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001178, Sample Num: 18848, Cur Loss: 0.42531028, Cur Avg Loss: 0.25563028, Log Avg loss: 0.22211355, Global Avg Loss: 1.14979345, Time: 0.0208 Steps: 39500, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001188, Sample Num: 19008, Cur Loss: 0.19158708, Cur Avg Loss: 0.25608422, Log Avg loss: 0.30955746, Global Avg Loss: 1.14958078, Time: 0.0208 Steps: 39510, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001198, Sample Num: 19168, Cur Loss: 0.19202638, Cur Avg Loss: 0.25617693, Log Avg loss: 0.26719092, Global Avg Loss: 1.14935751, Time: 0.0208 Steps: 39520, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001208, Sample Num: 19328, Cur Loss: 0.24342483, Cur Avg Loss: 0.25625728, Log Avg loss: 0.26588331, Global Avg Loss: 1.14913401, Time: 0.0208 Steps: 39530, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001218, Sample Num: 19488, Cur Loss: 0.26862541, Cur Avg Loss: 0.25632873, Log Avg loss: 0.26496028, Global Avg Loss: 1.14891040, Time: 0.0209 Steps: 39540, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001228, Sample Num: 19648, Cur Loss: 0.37206444, Cur Avg Loss: 0.25594605, Log Avg loss: 0.20933560, Global Avg Loss: 1.14867283, Time: 0.0208 Steps: 39550, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001238, Sample Num: 19808, Cur Loss: 0.13053626, Cur Avg Loss: 0.25487787, Log Avg loss: 0.12370557, Global Avg Loss: 1.14841374, Time: 0.0208 Steps: 39560, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001248, Sample Num: 19968, Cur Loss: 0.09180339, Cur Avg Loss: 0.25463797, Log Avg loss: 0.22493756, Global Avg Loss: 1.14818036, Time: 0.0209 Steps: 39570, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001258, Sample Num: 20128, Cur Loss: 0.31818545, Cur Avg Loss: 0.25487490, Log Avg loss: 0.28444489, Global Avg Loss: 1.14796213, Time: 0.0208 Steps: 39580, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001268, Sample Num: 20288, Cur Loss: 0.23601815, Cur Avg Loss: 0.25505237, Log Avg loss: 0.27737707, Global Avg Loss: 1.14774223, Time: 0.0208 Steps: 39590, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001278, Sample Num: 20448, Cur Loss: 0.36338976, Cur Avg Loss: 0.25580060, Log Avg loss: 0.35067659, Global Avg Loss: 1.14754096, Time: 0.0208 Steps: 39600, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001288, Sample Num: 20608, Cur Loss: 0.18590505, Cur Avg Loss: 0.25609446, Log Avg loss: 0.29364957, Global Avg Loss: 1.14732538, Time: 0.0209 Steps: 39610, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001298, Sample Num: 20768, Cur Loss: 0.16165718, Cur Avg Loss: 0.25619476, Log Avg loss: 0.26911395, Global Avg Loss: 1.14710372, Time: 0.0209 Steps: 39620, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001308, Sample Num: 20928, Cur Loss: 0.52107072, Cur Avg Loss: 0.25625386, Log Avg loss: 0.26392417, Global Avg Loss: 1.14688087, Time: 0.0209 Steps: 39630, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001318, Sample Num: 21088, Cur Loss: 0.26542109, Cur Avg Loss: 0.25627406, Log Avg loss: 0.25891609, Global Avg Loss: 1.14665686, Time: 0.0208 Steps: 39640, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001328, Sample Num: 21248, Cur Loss: 0.21713340, Cur Avg Loss: 0.25658237, Log Avg loss: 0.29721822, Global Avg Loss: 1.14644262, Time: 0.0208 Steps: 39650, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001338, Sample Num: 21408, Cur Loss: 0.17170981, Cur Avg Loss: 0.25647975, Log Avg loss: 0.24285218, Global Avg Loss: 1.14621479, Time: 0.0208 Steps: 39660, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001348, Sample Num: 21568, Cur Loss: 0.31755757, Cur Avg Loss: 0.25614678, Log Avg loss: 0.21159446, Global Avg Loss: 1.14597919, Time: 0.0208 Steps: 39670, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001358, Sample Num: 21728, Cur Loss: 0.04785057, Cur Avg Loss: 0.25664500, Log Avg loss: 0.32380597, Global Avg Loss: 1.14577199, Time: 0.0208 Steps: 39680, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001368, Sample Num: 21888, Cur Loss: 0.39799708, Cur Avg Loss: 0.25674001, Log Avg loss: 0.26964150, Global Avg Loss: 1.14555125, Time: 0.0208 Steps: 39690, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001378, Sample Num: 22048, Cur Loss: 0.14052823, Cur Avg Loss: 0.25627939, Log Avg loss: 0.19326653, Global Avg Loss: 1.14531138, Time: 0.0208 Steps: 39700, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001388, Sample Num: 22208, Cur Loss: 0.37093461, Cur Avg Loss: 0.25605852, Log Avg loss: 0.22562341, Global Avg Loss: 1.14507978, Time: 0.0209 Steps: 39710, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001398, Sample Num: 22368, Cur Loss: 0.33648157, Cur Avg Loss: 0.25531593, Log Avg loss: 0.15224472, Global Avg Loss: 1.14482982, Time: 0.0209 Steps: 39720, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001408, Sample Num: 22528, Cur Loss: 0.13154483, Cur Avg Loss: 0.25551660, Log Avg loss: 0.28357030, Global Avg Loss: 1.14461304, Time: 0.0208 Steps: 39730, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001418, Sample Num: 22688, Cur Loss: 0.31524324, Cur Avg Loss: 0.25552663, Log Avg loss: 0.25693785, Global Avg Loss: 1.14438967, Time: 0.0209 Steps: 39740, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001428, Sample Num: 22848, Cur Loss: 0.19823259, Cur Avg Loss: 0.25521304, Log Avg loss: 0.21074629, Global Avg Loss: 1.14415479, Time: 0.0209 Steps: 39750, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001438, Sample Num: 23008, Cur Loss: 0.22711033, Cur Avg Loss: 0.25539870, Log Avg loss: 0.28191045, Global Avg Loss: 1.14393793, Time: 0.0209 Steps: 39760, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001448, Sample Num: 23168, Cur Loss: 0.45824429, Cur Avg Loss: 0.25540623, Log Avg loss: 0.25648969, Global Avg Loss: 1.14371478, Time: 0.0208 Steps: 39770, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001458, Sample Num: 23328, Cur Loss: 0.17522281, Cur Avg Loss: 0.25528821, Log Avg loss: 0.23819902, Global Avg Loss: 1.14348715, Time: 0.0208 Steps: 39780, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001468, Sample Num: 23488, Cur Loss: 0.89739537, Cur Avg Loss: 0.25603084, Log Avg loss: 0.36430573, Global Avg Loss: 1.14329133, Time: 0.0208 Steps: 39790, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001478, Sample Num: 23648, Cur Loss: 0.13233714, Cur Avg Loss: 0.25631006, Log Avg loss: 0.29730000, Global Avg Loss: 1.14307877, Time: 0.0208 Steps: 39800, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001488, Sample Num: 23808, Cur Loss: 0.04809389, Cur Avg Loss: 0.25627239, Log Avg loss: 0.25070431, Global Avg Loss: 1.14285461, Time: 0.0208 Steps: 39810, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001498, Sample Num: 23968, Cur Loss: 0.15363818, Cur Avg Loss: 0.25651489, Log Avg loss: 0.29259973, Global Avg Loss: 1.14264108, Time: 0.0208 Steps: 39820, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001508, Sample Num: 24128, Cur Loss: 0.16389352, Cur Avg Loss: 0.25671968, Log Avg loss: 0.28739752, Global Avg Loss: 1.14242636, Time: 0.0209 Steps: 39830, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001518, Sample Num: 24288, Cur Loss: 0.17672950, Cur Avg Loss: 0.25678358, Log Avg loss: 0.26641914, Global Avg Loss: 1.14220648, Time: 0.0208 Steps: 39840, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001528, Sample Num: 24448, Cur Loss: 0.87367082, Cur Avg Loss: 0.25785980, Log Avg loss: 0.42123016, Global Avg Loss: 1.14202556, Time: 0.0208 Steps: 39850, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001538, Sample Num: 24608, Cur Loss: 0.09389894, Cur Avg Loss: 0.25759442, Log Avg loss: 0.21704390, Global Avg Loss: 1.14179350, Time: 0.0247 Steps: 39860, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001548, Sample Num: 24768, Cur Loss: 0.22362953, Cur Avg Loss: 0.25738483, Log Avg loss: 0.22514949, Global Avg Loss: 1.14156359, Time: 0.0209 Steps: 39870, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001558, Sample Num: 24928, Cur Loss: 0.13474201, Cur Avg Loss: 0.25671674, Log Avg loss: 0.15329771, Global Avg Loss: 1.14131578, Time: 0.0208 Steps: 39880, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001568, Sample Num: 25088, Cur Loss: 0.12574090, Cur Avg Loss: 0.25636572, Log Avg loss: 0.20167638, Global Avg Loss: 1.14108022, Time: 0.0208 Steps: 39890, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001578, Sample Num: 25248, Cur Loss: 0.72069985, Cur Avg Loss: 0.25736581, Log Avg loss: 0.41417949, Global Avg Loss: 1.14089804, Time: 0.0208 Steps: 39900, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001588, Sample Num: 25408, Cur Loss: 0.20114538, Cur Avg Loss: 0.25830641, Log Avg loss: 0.40673319, Global Avg Loss: 1.14071409, Time: 0.0209 Steps: 39910, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001598, Sample Num: 25568, Cur Loss: 0.06195996, Cur Avg Loss: 0.25911695, Log Avg loss: 0.38783120, Global Avg Loss: 1.14052549, Time: 0.0208 Steps: 39920, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001608, Sample Num: 25728, Cur Loss: 0.05970596, Cur Avg Loss: 0.25868276, Log Avg loss: 0.18929941, Global Avg Loss: 1.14028727, Time: 0.0209 Steps: 39930, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001618, Sample Num: 25888, Cur Loss: 0.10820029, Cur Avg Loss: 0.25844312, Log Avg loss: 0.21990907, Global Avg Loss: 1.14005683, Time: 0.0208 Steps: 39940, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001628, Sample Num: 26048, Cur Loss: 0.52181029, Cur Avg Loss: 0.25873621, Log Avg loss: 0.30615739, Global Avg Loss: 1.13984809, Time: 0.0208 Steps: 39950, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001638, Sample Num: 26208, Cur Loss: 0.30908960, Cur Avg Loss: 0.25980562, Log Avg loss: 0.43390495, Global Avg Loss: 1.13967143, Time: 0.0209 Steps: 39960, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001648, Sample Num: 26368, Cur Loss: 0.17406577, Cur Avg Loss: 0.25959655, Log Avg loss: 0.22535081, Global Avg Loss: 1.13944268, Time: 0.0209 Steps: 39970, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001658, Sample Num: 26528, Cur Loss: 0.28234392, Cur Avg Loss: 0.25947083, Log Avg loss: 0.23875250, Global Avg Loss: 1.13921739, Time: 0.0209 Steps: 39980, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001668, Sample Num: 26688, Cur Loss: 0.06313214, Cur Avg Loss: 0.25891680, Log Avg loss: 0.16705884, Global Avg Loss: 1.13897429, Time: 0.0209 Steps: 39990, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001678, Sample Num: 26848, Cur Loss: 0.08739877, Cur Avg Loss: 0.25863508, Log Avg loss: 0.21164492, Global Avg Loss: 1.13874246, Time: 0.0209 Steps: 40000, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001688, Sample Num: 27008, Cur Loss: 0.27975512, Cur Avg Loss: 0.25833963, Log Avg loss: 0.20876296, Global Avg Loss: 1.13851002, Time: 0.0209 Steps: 40010, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001698, Sample Num: 27168, Cur Loss: 0.40615568, Cur Avg Loss: 0.25794854, Log Avg loss: 0.19193297, Global Avg Loss: 1.13827350, Time: 0.0209 Steps: 40020, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001708, Sample Num: 27328, Cur Loss: 0.36853009, Cur Avg Loss: 0.25778281, Log Avg loss: 0.22964065, Global Avg Loss: 1.13804651, Time: 0.0208 Steps: 40030, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001718, Sample Num: 27488, Cur Loss: 0.52040946, Cur Avg Loss: 0.25751027, Log Avg loss: 0.21096021, Global Avg Loss: 1.13781497, Time: 0.0208 Steps: 40040, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001728, Sample Num: 27648, Cur Loss: 0.42760557, Cur Avg Loss: 0.25699403, Log Avg loss: 0.16830423, Global Avg Loss: 1.13757289, Time: 0.0208 Steps: 40050, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001738, Sample Num: 27808, Cur Loss: 0.10430962, Cur Avg Loss: 0.25727610, Log Avg loss: 0.30601769, Global Avg Loss: 1.13736531, Time: 0.0208 Steps: 40060, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001748, Sample Num: 27968, Cur Loss: 0.26005286, Cur Avg Loss: 0.25731052, Log Avg loss: 0.26329352, Global Avg Loss: 1.13714718, Time: 0.0208 Steps: 40070, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001758, Sample Num: 28128, Cur Loss: 0.11915363, Cur Avg Loss: 0.25710673, Log Avg loss: 0.22148335, Global Avg Loss: 1.13691872, Time: 0.0208 Steps: 40080, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001768, Sample Num: 28288, Cur Loss: 0.41403592, Cur Avg Loss: 0.25711160, Log Avg loss: 0.25796822, Global Avg Loss: 1.13669948, Time: 0.0208 Steps: 40090, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001778, Sample Num: 28448, Cur Loss: 0.33443576, Cur Avg Loss: 0.25678136, Log Avg loss: 0.19839576, Global Avg Loss: 1.13646548, Time: 0.0208 Steps: 40100, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001788, Sample Num: 28608, Cur Loss: 0.40568125, Cur Avg Loss: 0.25718574, Log Avg loss: 0.32908301, Global Avg Loss: 1.13626419, Time: 0.0208 Steps: 40110, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001798, Sample Num: 28768, Cur Loss: 0.49732196, Cur Avg Loss: 0.25854946, Log Avg loss: 0.50238351, Global Avg Loss: 1.13610620, Time: 0.0209 Steps: 40120, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001808, Sample Num: 28928, Cur Loss: 0.25561967, Cur Avg Loss: 0.25952168, Log Avg loss: 0.43432748, Global Avg Loss: 1.13593132, Time: 0.0209 Steps: 40130, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001818, Sample Num: 29088, Cur Loss: 0.15052140, Cur Avg Loss: 0.25995743, Log Avg loss: 0.33874117, Global Avg Loss: 1.13573272, Time: 0.0211 Steps: 40140, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001828, Sample Num: 29248, Cur Loss: 0.16603389, Cur Avg Loss: 0.25965893, Log Avg loss: 0.20539160, Global Avg Loss: 1.13550100, Time: 0.0210 Steps: 40150, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001838, Sample Num: 29408, Cur Loss: 0.38336933, Cur Avg Loss: 0.25948489, Log Avg loss: 0.22766927, Global Avg Loss: 1.13527495, Time: 0.0209 Steps: 40160, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001848, Sample Num: 29568, Cur Loss: 0.21851717, Cur Avg Loss: 0.25934003, Log Avg loss: 0.23271466, Global Avg Loss: 1.13505026, Time: 0.0209 Steps: 40170, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001858, Sample Num: 29728, Cur Loss: 0.27745017, Cur Avg Loss: 0.25950159, Log Avg loss: 0.28935810, Global Avg Loss: 1.13483979, Time: 0.0209 Steps: 40180, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001868, Sample Num: 29888, Cur Loss: 0.33334470, Cur Avg Loss: 0.25922199, Log Avg loss: 0.20727322, Global Avg Loss: 1.13460899, Time: 0.0209 Steps: 40190, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001878, Sample Num: 30048, Cur Loss: 0.37020794, Cur Avg Loss: 0.25912238, Log Avg loss: 0.24051370, Global Avg Loss: 1.13438658, Time: 0.0209 Steps: 40200, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001888, Sample Num: 30208, Cur Loss: 0.17928165, Cur Avg Loss: 0.25964496, Log Avg loss: 0.35778599, Global Avg Loss: 1.13419344, Time: 0.0209 Steps: 40210, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001898, Sample Num: 30368, Cur Loss: 0.29192704, Cur Avg Loss: 0.26012821, Log Avg loss: 0.35136518, Global Avg Loss: 1.13399881, Time: 0.0210 Steps: 40220, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001908, Sample Num: 30528, Cur Loss: 0.14057066, Cur Avg Loss: 0.25987328, Log Avg loss: 0.21148818, Global Avg Loss: 1.13376950, Time: 0.0210 Steps: 40230, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001918, Sample Num: 30688, Cur Loss: 0.19522503, Cur Avg Loss: 0.25951639, Log Avg loss: 0.19142139, Global Avg Loss: 1.13353532, Time: 0.0210 Steps: 40240, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001928, Sample Num: 30848, Cur Loss: 0.31965101, Cur Avg Loss: 0.25957897, Log Avg loss: 0.27158173, Global Avg Loss: 1.13332117, Time: 0.0210 Steps: 40250, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001938, Sample Num: 31008, Cur Loss: 0.26981872, Cur Avg Loss: 0.25953046, Log Avg loss: 0.25017793, Global Avg Loss: 1.13310181, Time: 0.0209 Steps: 40260, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001948, Sample Num: 31168, Cur Loss: 0.09178518, Cur Avg Loss: 0.25933862, Log Avg loss: 0.22216079, Global Avg Loss: 1.13287560, Time: 0.0211 Steps: 40270, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001958, Sample Num: 31328, Cur Loss: 0.25052902, Cur Avg Loss: 0.25895843, Log Avg loss: 0.18489808, Global Avg Loss: 1.13264025, Time: 0.0210 Steps: 40280, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001968, Sample Num: 31488, Cur Loss: 0.22563419, Cur Avg Loss: 0.25888010, Log Avg loss: 0.24354206, Global Avg Loss: 1.13241958, Time: 0.0210 Steps: 40290, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001978, Sample Num: 31648, Cur Loss: 0.14227809, Cur Avg Loss: 0.25893008, Log Avg loss: 0.26876611, Global Avg Loss: 1.13220527, Time: 0.0210 Steps: 40300, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001988, Sample Num: 31808, Cur Loss: 0.07492464, Cur Avg Loss: 0.25860507, Log Avg loss: 0.19431809, Global Avg Loss: 1.13197260, Time: 0.0210 Steps: 40310, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001998, Sample Num: 31968, Cur Loss: 0.35847598, Cur Avg Loss: 0.25834237, Log Avg loss: 0.20611835, Global Avg Loss: 1.13174297, Time: 0.0210 Steps: 40320, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002008, Sample Num: 32128, Cur Loss: 0.06214936, Cur Avg Loss: 0.25790431, Log Avg loss: 0.17037876, Global Avg Loss: 1.13150460, Time: 0.0210 Steps: 40330, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002018, Sample Num: 32288, Cur Loss: 0.34471142, Cur Avg Loss: 0.25819127, Log Avg loss: 0.31581405, Global Avg Loss: 1.13130240, Time: 0.0209 Steps: 40340, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002028, Sample Num: 32448, Cur Loss: 0.14128095, Cur Avg Loss: 0.25823462, Log Avg loss: 0.26698154, Global Avg Loss: 1.13108819, Time: 0.0210 Steps: 40350, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002038, Sample Num: 32608, Cur Loss: 0.11192269, Cur Avg Loss: 0.25756034, Log Avg loss: 0.12081687, Global Avg Loss: 1.13083787, Time: 0.0210 Steps: 40360, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002048, Sample Num: 32768, Cur Loss: 0.40136829, Cur Avg Loss: 0.25741211, Log Avg loss: 0.22720345, Global Avg Loss: 1.13061404, Time: 0.0255 Steps: 40370, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002058, Sample Num: 32928, Cur Loss: 0.34930044, Cur Avg Loss: 0.25694898, Log Avg loss: 0.16209981, Global Avg Loss: 1.13037419, Time: 0.0209 Steps: 40380, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002068, Sample Num: 33088, Cur Loss: 0.45299992, Cur Avg Loss: 0.25694695, Log Avg loss: 0.25652917, Global Avg Loss: 1.13015784, Time: 0.0208 Steps: 40390, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002078, Sample Num: 33248, Cur Loss: 0.10952808, Cur Avg Loss: 0.25679397, Log Avg loss: 0.22515733, Global Avg Loss: 1.12993383, Time: 0.0209 Steps: 40400, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002088, Sample Num: 33408, Cur Loss: 0.06880813, Cur Avg Loss: 0.25655055, Log Avg loss: 0.20596747, Global Avg Loss: 1.12970518, Time: 0.0209 Steps: 40410, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002098, Sample Num: 33568, Cur Loss: 0.10710041, Cur Avg Loss: 0.25656177, Log Avg loss: 0.25890565, Global Avg Loss: 1.12948974, Time: 0.0209 Steps: 40420, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002108, Sample Num: 33728, Cur Loss: 0.15640563, Cur Avg Loss: 0.25620283, Log Avg loss: 0.18089623, Global Avg Loss: 1.12925511, Time: 0.0209 Steps: 40430, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002118, Sample Num: 33888, Cur Loss: 0.18164687, Cur Avg Loss: 0.25615400, Log Avg loss: 0.24586021, Global Avg Loss: 1.12903667, Time: 0.0209 Steps: 40440, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002128, Sample Num: 34048, Cur Loss: 0.29575026, Cur Avg Loss: 0.25630743, Log Avg loss: 0.28880359, Global Avg Loss: 1.12882895, Time: 0.0209 Steps: 40450, Updated lr: 0.000063 ***** Running evaluation checkpoint-40451 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-40451 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.693171, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.23802, "eval_total_loss": 167.328162, "eval_mae": 0.339341, "eval_mse": 0.238088, "eval_r2": 0.848656, "eval_sp_statistic": 0.906247, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.923548, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.205604, "test_total_loss": 103.212959, "test_mae": 0.298987, "test_mse": 0.205673, "test_r2": 0.867257, "test_sp_statistic": 0.897926, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.940994, "test_ps_pvalue": 0.0, "lr": 6.258795637743007e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1288037040114742, "train_cur_epoch_loss": 545.5299713350832, "train_cur_epoch_avg_loss": 0.2562376568037028, "train_cur_epoch_time": 44.69317054748535, "train_cur_epoch_avg_time": 0.02099256484146799, "epoch": 19, "step": 40451} ################################################## Training, Epoch: 0020, Batch: 000009, Sample Num: 144, Cur Loss: 0.38356823, Cur Avg Loss: 0.32281220, Log Avg loss: 0.30130793, Global Avg Loss: 1.12862442, Time: 0.0209 Steps: 40460, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000019, Sample Num: 304, Cur Loss: 0.20582917, Cur Avg Loss: 0.30309791, Log Avg loss: 0.28535505, Global Avg Loss: 1.12841605, Time: 0.0208 Steps: 40470, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000029, Sample Num: 464, Cur Loss: 0.11251001, Cur Avg Loss: 0.28260642, Log Avg loss: 0.24367258, Global Avg Loss: 1.12819749, Time: 0.0208 Steps: 40480, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000039, Sample Num: 624, Cur Loss: 0.13954610, Cur Avg Loss: 0.27830224, Log Avg loss: 0.26582014, Global Avg Loss: 1.12798450, Time: 0.0208 Steps: 40490, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000049, Sample Num: 784, Cur Loss: 0.46523979, Cur Avg Loss: 0.28886166, Log Avg loss: 0.33004341, Global Avg Loss: 1.12778748, Time: 0.0208 Steps: 40500, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000059, Sample Num: 944, Cur Loss: 0.04479329, Cur Avg Loss: 0.30037128, Log Avg loss: 0.35676838, Global Avg Loss: 1.12759715, Time: 0.0208 Steps: 40510, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000069, Sample Num: 1104, Cur Loss: 0.14141135, Cur Avg Loss: 0.27891221, Log Avg loss: 0.15230368, Global Avg Loss: 1.12735646, Time: 0.0209 Steps: 40520, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000079, Sample Num: 1264, Cur Loss: 0.30114233, Cur Avg Loss: 0.26239287, Log Avg loss: 0.14840944, Global Avg Loss: 1.12711492, Time: 0.0209 Steps: 40530, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000089, Sample Num: 1424, Cur Loss: 0.17848575, Cur Avg Loss: 0.25508587, Log Avg loss: 0.19736060, Global Avg Loss: 1.12688558, Time: 0.0208 Steps: 40540, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000099, Sample Num: 1584, Cur Loss: 0.33071393, Cur Avg Loss: 0.26733954, Log Avg loss: 0.37639722, Global Avg Loss: 1.12670050, Time: 0.0209 Steps: 40550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000109, Sample Num: 1744, Cur Loss: 0.26860774, Cur Avg Loss: 0.27612110, Log Avg loss: 0.36305849, Global Avg Loss: 1.12651222, Time: 0.0209 Steps: 40560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000119, Sample Num: 1904, Cur Loss: 0.07049061, Cur Avg Loss: 0.26423074, Log Avg loss: 0.13462581, Global Avg Loss: 1.12626774, Time: 0.0209 Steps: 40570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000129, Sample Num: 2064, Cur Loss: 0.25749725, Cur Avg Loss: 0.26667045, Log Avg loss: 0.29570305, Global Avg Loss: 1.12606306, Time: 0.0208 Steps: 40580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000139, Sample Num: 2224, Cur Loss: 0.20171931, Cur Avg Loss: 0.26744198, Log Avg loss: 0.27739469, Global Avg Loss: 1.12585398, Time: 0.0208 Steps: 40590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000149, Sample Num: 2384, Cur Loss: 0.33957431, Cur Avg Loss: 0.26682232, Log Avg loss: 0.25820904, Global Avg Loss: 1.12564027, Time: 0.0210 Steps: 40600, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000159, Sample Num: 2544, Cur Loss: 0.25143129, Cur Avg Loss: 0.26092861, Log Avg loss: 0.17311231, Global Avg Loss: 1.12540572, Time: 0.0209 Steps: 40610, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000169, Sample Num: 2704, Cur Loss: 0.10930100, Cur Avg Loss: 0.25698853, Log Avg loss: 0.19434124, Global Avg Loss: 1.12517651, Time: 0.0209 Steps: 40620, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000179, Sample Num: 2864, Cur Loss: 0.09000959, Cur Avg Loss: 0.25298558, Log Avg loss: 0.18533575, Global Avg Loss: 1.12494519, Time: 0.0209 Steps: 40630, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000189, Sample Num: 3024, Cur Loss: 0.10288191, Cur Avg Loss: 0.25019725, Log Avg loss: 0.20028608, Global Avg Loss: 1.12471767, Time: 0.0208 Steps: 40640, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000199, Sample Num: 3184, Cur Loss: 0.15043595, Cur Avg Loss: 0.25179888, Log Avg loss: 0.28206977, Global Avg Loss: 1.12451037, Time: 0.0209 Steps: 40650, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000209, Sample Num: 3344, Cur Loss: 0.27964032, Cur Avg Loss: 0.24949818, Log Avg loss: 0.20371421, Global Avg Loss: 1.12428391, Time: 0.0209 Steps: 40660, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000219, Sample Num: 3504, Cur Loss: 0.23712507, Cur Avg Loss: 0.24978630, Log Avg loss: 0.25580813, Global Avg Loss: 1.12407037, Time: 0.0209 Steps: 40670, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000229, Sample Num: 3664, Cur Loss: 0.30748641, Cur Avg Loss: 0.24636156, Log Avg loss: 0.17135958, Global Avg Loss: 1.12383617, Time: 0.0208 Steps: 40680, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000239, Sample Num: 3824, Cur Loss: 0.19219002, Cur Avg Loss: 0.24290993, Log Avg loss: 0.16386763, Global Avg Loss: 1.12360025, Time: 0.0209 Steps: 40690, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000249, Sample Num: 3984, Cur Loss: 0.08594489, Cur Avg Loss: 0.24111883, Log Avg loss: 0.19831168, Global Avg Loss: 1.12337290, Time: 0.0210 Steps: 40700, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000259, Sample Num: 4144, Cur Loss: 0.10683472, Cur Avg Loss: 0.23979080, Log Avg loss: 0.20672281, Global Avg Loss: 1.12314774, Time: 0.0245 Steps: 40710, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000269, Sample Num: 4304, Cur Loss: 0.18280755, Cur Avg Loss: 0.23712647, Log Avg loss: 0.16812034, Global Avg Loss: 1.12291320, Time: 0.0209 Steps: 40720, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000279, Sample Num: 4464, Cur Loss: 0.25321811, Cur Avg Loss: 0.23471890, Log Avg loss: 0.16995530, Global Avg Loss: 1.12267923, Time: 0.0210 Steps: 40730, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000289, Sample Num: 4624, Cur Loss: 0.34903157, Cur Avg Loss: 0.23305756, Log Avg loss: 0.18670612, Global Avg Loss: 1.12244949, Time: 0.0210 Steps: 40740, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000299, Sample Num: 4784, Cur Loss: 0.18280473, Cur Avg Loss: 0.23219627, Log Avg loss: 0.20730497, Global Avg Loss: 1.12222492, Time: 0.0209 Steps: 40750, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000309, Sample Num: 4944, Cur Loss: 0.05191590, Cur Avg Loss: 0.22992349, Log Avg loss: 0.16196741, Global Avg Loss: 1.12198933, Time: 0.0210 Steps: 40760, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000319, Sample Num: 5104, Cur Loss: 0.23365080, Cur Avg Loss: 0.23096347, Log Avg loss: 0.26309893, Global Avg Loss: 1.12177866, Time: 0.0209 Steps: 40770, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000329, Sample Num: 5264, Cur Loss: 0.07825676, Cur Avg Loss: 0.23140427, Log Avg loss: 0.24546578, Global Avg Loss: 1.12156377, Time: 0.0209 Steps: 40780, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000339, Sample Num: 5424, Cur Loss: 0.15389259, Cur Avg Loss: 0.23241140, Log Avg loss: 0.26554592, Global Avg Loss: 1.12135391, Time: 0.0209 Steps: 40790, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000349, Sample Num: 5584, Cur Loss: 0.66532862, Cur Avg Loss: 0.23281577, Log Avg loss: 0.24652394, Global Avg Loss: 1.12113949, Time: 0.0209 Steps: 40800, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000359, Sample Num: 5744, Cur Loss: 0.12928522, Cur Avg Loss: 0.23114343, Log Avg loss: 0.17277860, Global Avg Loss: 1.12090711, Time: 0.0210 Steps: 40810, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000369, Sample Num: 5904, Cur Loss: 0.15729457, Cur Avg Loss: 0.23220451, Log Avg loss: 0.27029726, Global Avg Loss: 1.12069873, Time: 0.0209 Steps: 40820, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000379, Sample Num: 6064, Cur Loss: 0.39394444, Cur Avg Loss: 0.23507081, Log Avg loss: 0.34083752, Global Avg Loss: 1.12050773, Time: 0.0209 Steps: 40830, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000389, Sample Num: 6224, Cur Loss: 0.80949628, Cur Avg Loss: 0.23671451, Log Avg loss: 0.29901061, Global Avg Loss: 1.12030658, Time: 0.0209 Steps: 40840, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000399, Sample Num: 6384, Cur Loss: 0.10140686, Cur Avg Loss: 0.23520014, Log Avg loss: 0.17629123, Global Avg Loss: 1.12007548, Time: 0.0210 Steps: 40850, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000409, Sample Num: 6544, Cur Loss: 0.05332331, Cur Avg Loss: 0.23592219, Log Avg loss: 0.26473175, Global Avg Loss: 1.11986615, Time: 0.0209 Steps: 40860, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000419, Sample Num: 6704, Cur Loss: 0.14019716, Cur Avg Loss: 0.23417751, Log Avg loss: 0.16282025, Global Avg Loss: 1.11963198, Time: 0.0209 Steps: 40870, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000429, Sample Num: 6864, Cur Loss: 0.45573971, Cur Avg Loss: 0.23384614, Log Avg loss: 0.21996172, Global Avg Loss: 1.11941190, Time: 0.0210 Steps: 40880, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000439, Sample Num: 7024, Cur Loss: 0.09086113, Cur Avg Loss: 0.23225544, Log Avg loss: 0.16401431, Global Avg Loss: 1.11917825, Time: 0.0209 Steps: 40890, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000449, Sample Num: 7184, Cur Loss: 0.09238322, Cur Avg Loss: 0.23011825, Log Avg loss: 0.13629573, Global Avg Loss: 1.11893794, Time: 0.0209 Steps: 40900, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000459, Sample Num: 7344, Cur Loss: 0.15756828, Cur Avg Loss: 0.22833388, Log Avg loss: 0.14821587, Global Avg Loss: 1.11870066, Time: 0.0209 Steps: 40910, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000469, Sample Num: 7504, Cur Loss: 0.11494432, Cur Avg Loss: 0.22800091, Log Avg loss: 0.21271728, Global Avg Loss: 1.11847925, Time: 0.0210 Steps: 40920, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000479, Sample Num: 7664, Cur Loss: 0.62461919, Cur Avg Loss: 0.22797529, Log Avg loss: 0.22677367, Global Avg Loss: 1.11826139, Time: 0.0209 Steps: 40930, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000489, Sample Num: 7824, Cur Loss: 0.20220971, Cur Avg Loss: 0.22643248, Log Avg loss: 0.15253210, Global Avg Loss: 1.11802550, Time: 0.0209 Steps: 40940, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000499, Sample Num: 7984, Cur Loss: 0.19921674, Cur Avg Loss: 0.22732650, Log Avg loss: 0.27104387, Global Avg Loss: 1.11781867, Time: 0.0209 Steps: 40950, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000509, Sample Num: 8144, Cur Loss: 0.13516760, Cur Avg Loss: 0.22794508, Log Avg loss: 0.25881251, Global Avg Loss: 1.11760895, Time: 0.0209 Steps: 40960, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000519, Sample Num: 8304, Cur Loss: 0.33963025, Cur Avg Loss: 0.23083433, Log Avg loss: 0.37789686, Global Avg Loss: 1.11742840, Time: 0.0213 Steps: 40970, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000529, Sample Num: 8464, Cur Loss: 0.21534976, Cur Avg Loss: 0.23041505, Log Avg loss: 0.20865453, Global Avg Loss: 1.11720664, Time: 0.0209 Steps: 40980, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000539, Sample Num: 8624, Cur Loss: 0.31192076, Cur Avg Loss: 0.22981335, Log Avg loss: 0.19798360, Global Avg Loss: 1.11698239, Time: 0.0210 Steps: 40990, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000549, Sample Num: 8784, Cur Loss: 0.15818673, Cur Avg Loss: 0.22987180, Log Avg loss: 0.23302200, Global Avg Loss: 1.11676679, Time: 0.0210 Steps: 41000, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000559, Sample Num: 8944, Cur Loss: 0.12550952, Cur Avg Loss: 0.23059022, Log Avg loss: 0.27003174, Global Avg Loss: 1.11656032, Time: 0.0210 Steps: 41010, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000569, Sample Num: 9104, Cur Loss: 0.62468749, Cur Avg Loss: 0.23354810, Log Avg loss: 0.39889346, Global Avg Loss: 1.11638536, Time: 0.0209 Steps: 41020, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000579, Sample Num: 9264, Cur Loss: 0.37429973, Cur Avg Loss: 0.23364597, Log Avg loss: 0.23921459, Global Avg Loss: 1.11617157, Time: 0.0209 Steps: 41030, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000589, Sample Num: 9424, Cur Loss: 0.39665079, Cur Avg Loss: 0.23450613, Log Avg loss: 0.28430934, Global Avg Loss: 1.11596888, Time: 0.0209 Steps: 41040, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000599, Sample Num: 9584, Cur Loss: 0.24199119, Cur Avg Loss: 0.23429251, Log Avg loss: 0.22171048, Global Avg Loss: 1.11575103, Time: 0.0209 Steps: 41050, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000609, Sample Num: 9744, Cur Loss: 0.10125195, Cur Avg Loss: 0.23305511, Log Avg loss: 0.15893466, Global Avg Loss: 1.11551800, Time: 0.0210 Steps: 41060, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000619, Sample Num: 9904, Cur Loss: 0.27308878, Cur Avg Loss: 0.23313161, Log Avg loss: 0.23779070, Global Avg Loss: 1.11530429, Time: 0.0209 Steps: 41070, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000629, Sample Num: 10064, Cur Loss: 0.18082638, Cur Avg Loss: 0.23257072, Log Avg loss: 0.19785177, Global Avg Loss: 1.11508095, Time: 0.0209 Steps: 41080, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000639, Sample Num: 10224, Cur Loss: 0.26637775, Cur Avg Loss: 0.23240736, Log Avg loss: 0.22213163, Global Avg Loss: 1.11486364, Time: 0.0209 Steps: 41090, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000649, Sample Num: 10384, Cur Loss: 0.12133783, Cur Avg Loss: 0.23164830, Log Avg loss: 0.18314476, Global Avg Loss: 1.11463694, Time: 0.0210 Steps: 41100, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000659, Sample Num: 10544, Cur Loss: 0.13881266, Cur Avg Loss: 0.23179433, Log Avg loss: 0.24127157, Global Avg Loss: 1.11442450, Time: 0.0209 Steps: 41110, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000669, Sample Num: 10704, Cur Loss: 0.14676060, Cur Avg Loss: 0.23092137, Log Avg loss: 0.17339344, Global Avg Loss: 1.11419565, Time: 0.0210 Steps: 41120, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000679, Sample Num: 10864, Cur Loss: 0.18494640, Cur Avg Loss: 0.23042026, Log Avg loss: 0.19689597, Global Avg Loss: 1.11397262, Time: 0.0209 Steps: 41130, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000689, Sample Num: 11024, Cur Loss: 0.54813671, Cur Avg Loss: 0.23000258, Log Avg loss: 0.20164225, Global Avg Loss: 1.11375086, Time: 0.0210 Steps: 41140, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000699, Sample Num: 11184, Cur Loss: 0.30216902, Cur Avg Loss: 0.22996537, Log Avg loss: 0.22740115, Global Avg Loss: 1.11353547, Time: 0.0209 Steps: 41150, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000709, Sample Num: 11344, Cur Loss: 0.16968989, Cur Avg Loss: 0.23004479, Log Avg loss: 0.23559624, Global Avg Loss: 1.11332217, Time: 0.0210 Steps: 41160, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000719, Sample Num: 11504, Cur Loss: 0.05210003, Cur Avg Loss: 0.23033776, Log Avg loss: 0.25110961, Global Avg Loss: 1.11311274, Time: 0.0209 Steps: 41170, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000729, Sample Num: 11664, Cur Loss: 0.14931829, Cur Avg Loss: 0.23001626, Log Avg loss: 0.20689989, Global Avg Loss: 1.11289268, Time: 0.0209 Steps: 41180, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000739, Sample Num: 11824, Cur Loss: 0.18335904, Cur Avg Loss: 0.22922494, Log Avg loss: 0.17153775, Global Avg Loss: 1.11266414, Time: 0.0209 Steps: 41190, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000749, Sample Num: 11984, Cur Loss: 0.10366490, Cur Avg Loss: 0.22930752, Log Avg loss: 0.23541039, Global Avg Loss: 1.11245121, Time: 0.0209 Steps: 41200, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000759, Sample Num: 12144, Cur Loss: 0.22079131, Cur Avg Loss: 0.22806999, Log Avg loss: 0.13537890, Global Avg Loss: 1.11221412, Time: 0.0209 Steps: 41210, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000769, Sample Num: 12304, Cur Loss: 0.22339152, Cur Avg Loss: 0.22722367, Log Avg loss: 0.16298843, Global Avg Loss: 1.11198383, Time: 0.0246 Steps: 41220, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000779, Sample Num: 12464, Cur Loss: 0.15691939, Cur Avg Loss: 0.22653681, Log Avg loss: 0.17371669, Global Avg Loss: 1.11175626, Time: 0.0209 Steps: 41230, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000789, Sample Num: 12624, Cur Loss: 0.27426648, Cur Avg Loss: 0.22648010, Log Avg loss: 0.22206238, Global Avg Loss: 1.11154053, Time: 0.0209 Steps: 41240, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000799, Sample Num: 12784, Cur Loss: 0.23217183, Cur Avg Loss: 0.22618282, Log Avg loss: 0.20272808, Global Avg Loss: 1.11132021, Time: 0.0209 Steps: 41250, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000809, Sample Num: 12944, Cur Loss: 0.41232425, Cur Avg Loss: 0.22635088, Log Avg loss: 0.23977851, Global Avg Loss: 1.11110898, Time: 0.0209 Steps: 41260, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000819, Sample Num: 13104, Cur Loss: 0.14782669, Cur Avg Loss: 0.22586514, Log Avg loss: 0.18656890, Global Avg Loss: 1.11088496, Time: 0.0210 Steps: 41270, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000829, Sample Num: 13264, Cur Loss: 0.21506524, Cur Avg Loss: 0.22601480, Log Avg loss: 0.23827220, Global Avg Loss: 1.11067357, Time: 0.0209 Steps: 41280, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000839, Sample Num: 13424, Cur Loss: 0.35944349, Cur Avg Loss: 0.22652756, Log Avg loss: 0.26903506, Global Avg Loss: 1.11046973, Time: 0.0209 Steps: 41290, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000849, Sample Num: 13584, Cur Loss: 0.13527012, Cur Avg Loss: 0.22692516, Log Avg loss: 0.26028369, Global Avg Loss: 1.11026388, Time: 0.0210 Steps: 41300, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000859, Sample Num: 13744, Cur Loss: 0.19866183, Cur Avg Loss: 0.22724379, Log Avg loss: 0.25429543, Global Avg Loss: 1.11005667, Time: 0.0209 Steps: 41310, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000869, Sample Num: 13904, Cur Loss: 0.14372291, Cur Avg Loss: 0.22712057, Log Avg loss: 0.21653637, Global Avg Loss: 1.10984043, Time: 0.0209 Steps: 41320, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000879, Sample Num: 14064, Cur Loss: 0.21777010, Cur Avg Loss: 0.22748782, Log Avg loss: 0.25940184, Global Avg Loss: 1.10963466, Time: 0.0209 Steps: 41330, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000889, Sample Num: 14224, Cur Loss: 0.04859988, Cur Avg Loss: 0.22726908, Log Avg loss: 0.20804157, Global Avg Loss: 1.10941657, Time: 0.0209 Steps: 41340, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000899, Sample Num: 14384, Cur Loss: 0.15574244, Cur Avg Loss: 0.22768257, Log Avg loss: 0.26444204, Global Avg Loss: 1.10921222, Time: 0.0210 Steps: 41350, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000909, Sample Num: 14544, Cur Loss: 0.29595634, Cur Avg Loss: 0.22846769, Log Avg loss: 0.29904980, Global Avg Loss: 1.10901634, Time: 0.0210 Steps: 41360, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000919, Sample Num: 14704, Cur Loss: 0.14423372, Cur Avg Loss: 0.22828302, Log Avg loss: 0.21149640, Global Avg Loss: 1.10879939, Time: 0.0209 Steps: 41370, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000929, Sample Num: 14864, Cur Loss: 0.12246630, Cur Avg Loss: 0.22798415, Log Avg loss: 0.20051789, Global Avg Loss: 1.10857989, Time: 0.0209 Steps: 41380, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000939, Sample Num: 15024, Cur Loss: 0.11036921, Cur Avg Loss: 0.22711969, Log Avg loss: 0.14681119, Global Avg Loss: 1.10834752, Time: 0.0209 Steps: 41390, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000949, Sample Num: 15184, Cur Loss: 0.12865825, Cur Avg Loss: 0.22687255, Log Avg loss: 0.20366620, Global Avg Loss: 1.10812900, Time: 0.0209 Steps: 41400, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000959, Sample Num: 15344, Cur Loss: 0.22832988, Cur Avg Loss: 0.22722169, Log Avg loss: 0.26035500, Global Avg Loss: 1.10792428, Time: 0.0210 Steps: 41410, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000969, Sample Num: 15504, Cur Loss: 0.13126987, Cur Avg Loss: 0.22670972, Log Avg loss: 0.17761177, Global Avg Loss: 1.10769967, Time: 0.0210 Steps: 41420, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000979, Sample Num: 15664, Cur Loss: 0.51381797, Cur Avg Loss: 0.22692186, Log Avg loss: 0.24747812, Global Avg Loss: 1.10749204, Time: 0.0210 Steps: 41430, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000989, Sample Num: 15824, Cur Loss: 0.33697572, Cur Avg Loss: 0.22676077, Log Avg loss: 0.21099049, Global Avg Loss: 1.10727570, Time: 0.0210 Steps: 41440, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000999, Sample Num: 15984, Cur Loss: 0.25331318, Cur Avg Loss: 0.22628181, Log Avg loss: 0.17891298, Global Avg Loss: 1.10705173, Time: 0.0210 Steps: 41450, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001009, Sample Num: 16144, Cur Loss: 0.23513116, Cur Avg Loss: 0.22646708, Log Avg loss: 0.24497525, Global Avg Loss: 1.10684380, Time: 0.0209 Steps: 41460, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001019, Sample Num: 16304, Cur Loss: 0.44955111, Cur Avg Loss: 0.22642815, Log Avg loss: 0.22250037, Global Avg Loss: 1.10663055, Time: 0.0209 Steps: 41470, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001029, Sample Num: 16464, Cur Loss: 0.88398957, Cur Avg Loss: 0.22813378, Log Avg loss: 0.40193684, Global Avg Loss: 1.10646066, Time: 0.0246 Steps: 41480, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001039, Sample Num: 16624, Cur Loss: 0.35202140, Cur Avg Loss: 0.22787416, Log Avg loss: 0.20115945, Global Avg Loss: 1.10624247, Time: 0.0209 Steps: 41490, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001049, Sample Num: 16784, Cur Loss: 0.14944525, Cur Avg Loss: 0.22753847, Log Avg loss: 0.19266035, Global Avg Loss: 1.10602232, Time: 0.0209 Steps: 41500, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001059, Sample Num: 16944, Cur Loss: 0.22135483, Cur Avg Loss: 0.22717694, Log Avg loss: 0.18925290, Global Avg Loss: 1.10580147, Time: 0.0209 Steps: 41510, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001069, Sample Num: 17104, Cur Loss: 0.20737939, Cur Avg Loss: 0.22752023, Log Avg loss: 0.26387454, Global Avg Loss: 1.10559869, Time: 0.0209 Steps: 41520, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001079, Sample Num: 17264, Cur Loss: 0.08763243, Cur Avg Loss: 0.22790323, Log Avg loss: 0.26884625, Global Avg Loss: 1.10539721, Time: 0.0209 Steps: 41530, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001089, Sample Num: 17424, Cur Loss: 0.18526615, Cur Avg Loss: 0.22834970, Log Avg loss: 0.27652320, Global Avg Loss: 1.10519768, Time: 0.0210 Steps: 41540, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001099, Sample Num: 17584, Cur Loss: 0.15337189, Cur Avg Loss: 0.22801098, Log Avg loss: 0.19112448, Global Avg Loss: 1.10497768, Time: 0.0209 Steps: 41550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001109, Sample Num: 17744, Cur Loss: 0.41973367, Cur Avg Loss: 0.22766443, Log Avg loss: 0.18957856, Global Avg Loss: 1.10475742, Time: 0.0209 Steps: 41560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001119, Sample Num: 17904, Cur Loss: 0.09897837, Cur Avg Loss: 0.22696587, Log Avg loss: 0.14949578, Global Avg Loss: 1.10452763, Time: 0.0210 Steps: 41570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001129, Sample Num: 18064, Cur Loss: 0.38648552, Cur Avg Loss: 0.22833268, Log Avg loss: 0.38127872, Global Avg Loss: 1.10435369, Time: 0.0208 Steps: 41580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001139, Sample Num: 18224, Cur Loss: 0.04767085, Cur Avg Loss: 0.22779818, Log Avg loss: 0.16745304, Global Avg Loss: 1.10412841, Time: 0.0208 Steps: 41590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001149, Sample Num: 18384, Cur Loss: 0.17125843, Cur Avg Loss: 0.22771151, Log Avg loss: 0.21784020, Global Avg Loss: 1.10391536, Time: 0.0208 Steps: 41600, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001159, Sample Num: 18544, Cur Loss: 0.15501153, Cur Avg Loss: 0.22749892, Log Avg loss: 0.20307152, Global Avg Loss: 1.10369887, Time: 0.0208 Steps: 41610, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001169, Sample Num: 18704, Cur Loss: 0.53272063, Cur Avg Loss: 0.22773489, Log Avg loss: 0.25508416, Global Avg Loss: 1.10349497, Time: 0.0208 Steps: 41620, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001179, Sample Num: 18864, Cur Loss: 0.40112394, Cur Avg Loss: 0.22810391, Log Avg loss: 0.27124283, Global Avg Loss: 1.10329506, Time: 0.0208 Steps: 41630, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001189, Sample Num: 19024, Cur Loss: 0.29497010, Cur Avg Loss: 0.22843273, Log Avg loss: 0.26719950, Global Avg Loss: 1.10309426, Time: 0.0208 Steps: 41640, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001199, Sample Num: 19184, Cur Loss: 0.15269311, Cur Avg Loss: 0.22780986, Log Avg loss: 0.15375154, Global Avg Loss: 1.10286633, Time: 0.0209 Steps: 41650, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001209, Sample Num: 19344, Cur Loss: 0.13050519, Cur Avg Loss: 0.22799433, Log Avg loss: 0.25011203, Global Avg Loss: 1.10266164, Time: 0.0208 Steps: 41660, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001219, Sample Num: 19504, Cur Loss: 0.34158540, Cur Avg Loss: 0.22830337, Log Avg loss: 0.26566571, Global Avg Loss: 1.10246077, Time: 0.0208 Steps: 41670, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001229, Sample Num: 19664, Cur Loss: 0.09987767, Cur Avg Loss: 0.22834052, Log Avg loss: 0.23286941, Global Avg Loss: 1.10225214, Time: 0.0208 Steps: 41680, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001239, Sample Num: 19824, Cur Loss: 0.20546250, Cur Avg Loss: 0.22845771, Log Avg loss: 0.24286057, Global Avg Loss: 1.10204600, Time: 0.0208 Steps: 41690, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001249, Sample Num: 19984, Cur Loss: 0.31567305, Cur Avg Loss: 0.22842294, Log Avg loss: 0.22411431, Global Avg Loss: 1.10183546, Time: 0.0209 Steps: 41700, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001259, Sample Num: 20144, Cur Loss: 0.24944234, Cur Avg Loss: 0.22813434, Log Avg loss: 0.19208857, Global Avg Loss: 1.10161735, Time: 0.0208 Steps: 41710, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001269, Sample Num: 20304, Cur Loss: 0.56300849, Cur Avg Loss: 0.22827019, Log Avg loss: 0.24537431, Global Avg Loss: 1.10141212, Time: 0.0208 Steps: 41720, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001279, Sample Num: 20464, Cur Loss: 0.20602605, Cur Avg Loss: 0.22780093, Log Avg loss: 0.16825183, Global Avg Loss: 1.10118850, Time: 0.0208 Steps: 41730, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001289, Sample Num: 20624, Cur Loss: 0.13073643, Cur Avg Loss: 0.22733877, Log Avg loss: 0.16822861, Global Avg Loss: 1.10096498, Time: 0.0208 Steps: 41740, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001299, Sample Num: 20784, Cur Loss: 0.11336678, Cur Avg Loss: 0.22726510, Log Avg loss: 0.21776819, Global Avg Loss: 1.10075344, Time: 0.0208 Steps: 41750, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001309, Sample Num: 20944, Cur Loss: 0.36978844, Cur Avg Loss: 0.22718889, Log Avg loss: 0.21728982, Global Avg Loss: 1.10054188, Time: 0.0209 Steps: 41760, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001319, Sample Num: 21104, Cur Loss: 0.33603939, Cur Avg Loss: 0.22736848, Log Avg loss: 0.25087642, Global Avg Loss: 1.10033846, Time: 0.0208 Steps: 41770, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001329, Sample Num: 21264, Cur Loss: 0.19172110, Cur Avg Loss: 0.22695989, Log Avg loss: 0.17306757, Global Avg Loss: 1.10011652, Time: 0.0208 Steps: 41780, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001339, Sample Num: 21424, Cur Loss: 0.03156972, Cur Avg Loss: 0.22697862, Log Avg loss: 0.22946745, Global Avg Loss: 1.09990818, Time: 0.0209 Steps: 41790, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001349, Sample Num: 21584, Cur Loss: 0.09889373, Cur Avg Loss: 0.22739726, Log Avg loss: 0.28345356, Global Avg Loss: 1.09971286, Time: 0.0209 Steps: 41800, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001359, Sample Num: 21744, Cur Loss: 0.22887582, Cur Avg Loss: 0.22788314, Log Avg loss: 0.29342812, Global Avg Loss: 1.09952001, Time: 0.0209 Steps: 41810, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001369, Sample Num: 21904, Cur Loss: 0.33059564, Cur Avg Loss: 0.22791371, Log Avg loss: 0.23206727, Global Avg Loss: 1.09931259, Time: 0.0208 Steps: 41820, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001379, Sample Num: 22064, Cur Loss: 0.16258301, Cur Avg Loss: 0.22755262, Log Avg loss: 0.17812049, Global Avg Loss: 1.09909237, Time: 0.0209 Steps: 41830, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001389, Sample Num: 22224, Cur Loss: 0.14494143, Cur Avg Loss: 0.22787008, Log Avg loss: 0.27164766, Global Avg Loss: 1.09889460, Time: 0.0209 Steps: 41840, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001399, Sample Num: 22384, Cur Loss: 0.05822786, Cur Avg Loss: 0.22816843, Log Avg loss: 0.26960855, Global Avg Loss: 1.09869645, Time: 0.0209 Steps: 41850, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001409, Sample Num: 22544, Cur Loss: 0.12005335, Cur Avg Loss: 0.22778085, Log Avg loss: 0.17355942, Global Avg Loss: 1.09847544, Time: 0.0209 Steps: 41860, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001419, Sample Num: 22704, Cur Loss: 0.49284399, Cur Avg Loss: 0.22911315, Log Avg loss: 0.41683355, Global Avg Loss: 1.09831264, Time: 0.0208 Steps: 41870, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001429, Sample Num: 22864, Cur Loss: 0.11452621, Cur Avg Loss: 0.22901704, Log Avg loss: 0.21537857, Global Avg Loss: 1.09810181, Time: 0.0209 Steps: 41880, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001439, Sample Num: 23024, Cur Loss: 0.18438101, Cur Avg Loss: 0.22936102, Log Avg loss: 0.27851625, Global Avg Loss: 1.09790616, Time: 0.0209 Steps: 41890, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001449, Sample Num: 23184, Cur Loss: 0.04931054, Cur Avg Loss: 0.22901538, Log Avg loss: 0.17927760, Global Avg Loss: 1.09768692, Time: 0.0209 Steps: 41900, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001459, Sample Num: 23344, Cur Loss: 0.33105582, Cur Avg Loss: 0.22883871, Log Avg loss: 0.20323882, Global Avg Loss: 1.09747350, Time: 0.0208 Steps: 41910, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001469, Sample Num: 23504, Cur Loss: 0.09243683, Cur Avg Loss: 0.22844634, Log Avg loss: 0.17120049, Global Avg Loss: 1.09725254, Time: 0.0208 Steps: 41920, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001479, Sample Num: 23664, Cur Loss: 0.30439141, Cur Avg Loss: 0.22818898, Log Avg loss: 0.19038213, Global Avg Loss: 1.09703625, Time: 0.0209 Steps: 41930, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001489, Sample Num: 23824, Cur Loss: 0.30780724, Cur Avg Loss: 0.22854421, Log Avg loss: 0.28108292, Global Avg Loss: 1.09684170, Time: 0.0208 Steps: 41940, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001499, Sample Num: 23984, Cur Loss: 0.17300513, Cur Avg Loss: 0.22835619, Log Avg loss: 0.20035983, Global Avg Loss: 1.09662800, Time: 0.0209 Steps: 41950, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001509, Sample Num: 24144, Cur Loss: 0.20691425, Cur Avg Loss: 0.22882998, Log Avg loss: 0.29985100, Global Avg Loss: 1.09643811, Time: 0.0209 Steps: 41960, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001519, Sample Num: 24304, Cur Loss: 0.15163533, Cur Avg Loss: 0.22841160, Log Avg loss: 0.16527771, Global Avg Loss: 1.09621625, Time: 0.0209 Steps: 41970, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001529, Sample Num: 24464, Cur Loss: 0.09833343, Cur Avg Loss: 0.22825488, Log Avg loss: 0.20444996, Global Avg Loss: 1.09600382, Time: 0.0209 Steps: 41980, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001539, Sample Num: 24624, Cur Loss: 0.19107723, Cur Avg Loss: 0.22810442, Log Avg loss: 0.20509859, Global Avg Loss: 1.09579165, Time: 0.0245 Steps: 41990, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001549, Sample Num: 24784, Cur Loss: 0.23004553, Cur Avg Loss: 0.22763873, Log Avg loss: 0.15596891, Global Avg Loss: 1.09556788, Time: 0.0208 Steps: 42000, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001559, Sample Num: 24944, Cur Loss: 0.21249354, Cur Avg Loss: 0.22743573, Log Avg loss: 0.19599211, Global Avg Loss: 1.09535375, Time: 0.0208 Steps: 42010, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001569, Sample Num: 25104, Cur Loss: 0.12644723, Cur Avg Loss: 0.22693161, Log Avg loss: 0.14833827, Global Avg Loss: 1.09512838, Time: 0.0209 Steps: 42020, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001579, Sample Num: 25264, Cur Loss: 0.17607988, Cur Avg Loss: 0.22719410, Log Avg loss: 0.26837855, Global Avg Loss: 1.09493167, Time: 0.0208 Steps: 42030, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001589, Sample Num: 25424, Cur Loss: 0.22783904, Cur Avg Loss: 0.22748759, Log Avg loss: 0.27383024, Global Avg Loss: 1.09473636, Time: 0.0208 Steps: 42040, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001599, Sample Num: 25584, Cur Loss: 0.08003379, Cur Avg Loss: 0.22680329, Log Avg loss: 0.11806740, Global Avg Loss: 1.09450409, Time: 0.0208 Steps: 42050, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001609, Sample Num: 25744, Cur Loss: 0.08703860, Cur Avg Loss: 0.22638200, Log Avg loss: 0.15901900, Global Avg Loss: 1.09428168, Time: 0.0208 Steps: 42060, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001619, Sample Num: 25904, Cur Loss: 0.35707796, Cur Avg Loss: 0.22664425, Log Avg loss: 0.26884042, Global Avg Loss: 1.09408547, Time: 0.0208 Steps: 42070, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001629, Sample Num: 26064, Cur Loss: 0.25808766, Cur Avg Loss: 0.22670739, Log Avg loss: 0.23692890, Global Avg Loss: 1.09388177, Time: 0.0208 Steps: 42080, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001639, Sample Num: 26224, Cur Loss: 0.10257319, Cur Avg Loss: 0.22632924, Log Avg loss: 0.16472857, Global Avg Loss: 1.09366102, Time: 0.0208 Steps: 42090, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001649, Sample Num: 26384, Cur Loss: 0.38174704, Cur Avg Loss: 0.22676150, Log Avg loss: 0.29760924, Global Avg Loss: 1.09347193, Time: 0.0209 Steps: 42100, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001659, Sample Num: 26544, Cur Loss: 0.33097157, Cur Avg Loss: 0.22670463, Log Avg loss: 0.21732741, Global Avg Loss: 1.09326387, Time: 0.0209 Steps: 42110, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001669, Sample Num: 26704, Cur Loss: 0.30608383, Cur Avg Loss: 0.22696782, Log Avg loss: 0.27063051, Global Avg Loss: 1.09306856, Time: 0.0208 Steps: 42120, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001679, Sample Num: 26864, Cur Loss: 0.25843501, Cur Avg Loss: 0.22704012, Log Avg loss: 0.23910750, Global Avg Loss: 1.09286587, Time: 0.0208 Steps: 42130, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001689, Sample Num: 27024, Cur Loss: 0.21708274, Cur Avg Loss: 0.22729264, Log Avg loss: 0.26969016, Global Avg Loss: 1.09267052, Time: 0.0208 Steps: 42140, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001699, Sample Num: 27184, Cur Loss: 0.39127111, Cur Avg Loss: 0.22768554, Log Avg loss: 0.29404624, Global Avg Loss: 1.09248105, Time: 0.0208 Steps: 42150, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001709, Sample Num: 27344, Cur Loss: 0.24771829, Cur Avg Loss: 0.22783523, Log Avg loss: 0.25326804, Global Avg Loss: 1.09228200, Time: 0.0208 Steps: 42160, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001719, Sample Num: 27504, Cur Loss: 0.21618627, Cur Avg Loss: 0.22799064, Log Avg loss: 0.25454914, Global Avg Loss: 1.09208334, Time: 0.0208 Steps: 42170, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001729, Sample Num: 27664, Cur Loss: 0.18391587, Cur Avg Loss: 0.22818075, Log Avg loss: 0.26086105, Global Avg Loss: 1.09188628, Time: 0.0208 Steps: 42180, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001739, Sample Num: 27824, Cur Loss: 0.14414966, Cur Avg Loss: 0.22852110, Log Avg loss: 0.28736698, Global Avg Loss: 1.09169559, Time: 0.0208 Steps: 42190, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001749, Sample Num: 27984, Cur Loss: 0.10042926, Cur Avg Loss: 0.22848063, Log Avg loss: 0.22144424, Global Avg Loss: 1.09148937, Time: 0.0208 Steps: 42200, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001759, Sample Num: 28144, Cur Loss: 0.13125803, Cur Avg Loss: 0.22827136, Log Avg loss: 0.19166887, Global Avg Loss: 1.09127619, Time: 0.0208 Steps: 42210, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001769, Sample Num: 28304, Cur Loss: 0.19153209, Cur Avg Loss: 0.22794996, Log Avg loss: 0.17141555, Global Avg Loss: 1.09105832, Time: 0.0208 Steps: 42220, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001779, Sample Num: 28464, Cur Loss: 0.11223592, Cur Avg Loss: 0.22835512, Log Avg loss: 0.30002849, Global Avg Loss: 1.09087100, Time: 0.0208 Steps: 42230, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001789, Sample Num: 28624, Cur Loss: 0.21513686, Cur Avg Loss: 0.22811545, Log Avg loss: 0.18547841, Global Avg Loss: 1.09065666, Time: 0.0209 Steps: 42240, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001799, Sample Num: 28784, Cur Loss: 0.43530688, Cur Avg Loss: 0.22837331, Log Avg loss: 0.27450480, Global Avg Loss: 1.09046348, Time: 0.0211 Steps: 42250, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001809, Sample Num: 28944, Cur Loss: 0.17745766, Cur Avg Loss: 0.22884845, Log Avg loss: 0.31432501, Global Avg Loss: 1.09027983, Time: 0.0209 Steps: 42260, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001819, Sample Num: 29104, Cur Loss: 0.05783219, Cur Avg Loss: 0.22837891, Log Avg loss: 0.14343978, Global Avg Loss: 1.09005583, Time: 0.0210 Steps: 42270, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001829, Sample Num: 29264, Cur Loss: 0.15142550, Cur Avg Loss: 0.22828355, Log Avg loss: 0.21093827, Global Avg Loss: 1.08984790, Time: 0.0210 Steps: 42280, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001839, Sample Num: 29424, Cur Loss: 0.53248537, Cur Avg Loss: 0.22852409, Log Avg loss: 0.27251814, Global Avg Loss: 1.08965463, Time: 0.0209 Steps: 42290, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001849, Sample Num: 29584, Cur Loss: 0.25136757, Cur Avg Loss: 0.22855526, Log Avg loss: 0.23428723, Global Avg Loss: 1.08945242, Time: 0.0210 Steps: 42300, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001859, Sample Num: 29744, Cur Loss: 0.97705984, Cur Avg Loss: 0.22922274, Log Avg loss: 0.35264036, Global Avg Loss: 1.08927827, Time: 0.0210 Steps: 42310, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001869, Sample Num: 29904, Cur Loss: 0.13261008, Cur Avg Loss: 0.22934310, Log Avg loss: 0.25171822, Global Avg Loss: 1.08908036, Time: 0.0209 Steps: 42320, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001879, Sample Num: 30064, Cur Loss: 0.16420950, Cur Avg Loss: 0.22927714, Log Avg loss: 0.21694778, Global Avg Loss: 1.08887433, Time: 0.0209 Steps: 42330, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001889, Sample Num: 30224, Cur Loss: 0.15500014, Cur Avg Loss: 0.22932404, Log Avg loss: 0.23813685, Global Avg Loss: 1.08867340, Time: 0.0210 Steps: 42340, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001899, Sample Num: 30384, Cur Loss: 0.26954773, Cur Avg Loss: 0.22944159, Log Avg loss: 0.25164712, Global Avg Loss: 1.08847575, Time: 0.0210 Steps: 42350, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001909, Sample Num: 30544, Cur Loss: 0.27848744, Cur Avg Loss: 0.22976462, Log Avg loss: 0.29110883, Global Avg Loss: 1.08828752, Time: 0.0209 Steps: 42360, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001919, Sample Num: 30704, Cur Loss: 0.42799672, Cur Avg Loss: 0.22979623, Log Avg loss: 0.23583074, Global Avg Loss: 1.08808633, Time: 0.0210 Steps: 42370, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001929, Sample Num: 30864, Cur Loss: 0.28288591, Cur Avg Loss: 0.23002355, Log Avg loss: 0.27364637, Global Avg Loss: 1.08789415, Time: 0.0209 Steps: 42380, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001939, Sample Num: 31024, Cur Loss: 0.24659446, Cur Avg Loss: 0.22969916, Log Avg loss: 0.16712397, Global Avg Loss: 1.08767694, Time: 0.0210 Steps: 42390, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001949, Sample Num: 31184, Cur Loss: 0.33072582, Cur Avg Loss: 0.22965540, Log Avg loss: 0.22117019, Global Avg Loss: 1.08747257, Time: 0.0210 Steps: 42400, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001959, Sample Num: 31344, Cur Loss: 0.26744851, Cur Avg Loss: 0.22964678, Log Avg loss: 0.22796567, Global Avg Loss: 1.08726990, Time: 0.0209 Steps: 42410, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001969, Sample Num: 31504, Cur Loss: 0.21418668, Cur Avg Loss: 0.22942265, Log Avg loss: 0.18551634, Global Avg Loss: 1.08705733, Time: 0.0210 Steps: 42420, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001979, Sample Num: 31664, Cur Loss: 0.06938218, Cur Avg Loss: 0.22939943, Log Avg loss: 0.22482799, Global Avg Loss: 1.08685412, Time: 0.0210 Steps: 42430, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001989, Sample Num: 31824, Cur Loss: 0.27275068, Cur Avg Loss: 0.22933245, Log Avg loss: 0.21607702, Global Avg Loss: 1.08664894, Time: 0.0210 Steps: 42440, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001999, Sample Num: 31984, Cur Loss: 0.28617722, Cur Avg Loss: 0.22931313, Log Avg loss: 0.22546965, Global Avg Loss: 1.08644607, Time: 0.0209 Steps: 42450, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002009, Sample Num: 32144, Cur Loss: 0.46209922, Cur Avg Loss: 0.22951770, Log Avg loss: 0.27041231, Global Avg Loss: 1.08625388, Time: 0.0209 Steps: 42460, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002019, Sample Num: 32304, Cur Loss: 0.11417265, Cur Avg Loss: 0.22960334, Log Avg loss: 0.24680664, Global Avg Loss: 1.08605622, Time: 0.0210 Steps: 42470, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002029, Sample Num: 32464, Cur Loss: 0.06602367, Cur Avg Loss: 0.22967458, Log Avg loss: 0.24405937, Global Avg Loss: 1.08585801, Time: 0.0210 Steps: 42480, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002039, Sample Num: 32624, Cur Loss: 0.29931045, Cur Avg Loss: 0.23039513, Log Avg loss: 0.37659344, Global Avg Loss: 1.08569109, Time: 0.0209 Steps: 42490, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002049, Sample Num: 32784, Cur Loss: 0.05034740, Cur Avg Loss: 0.23009225, Log Avg loss: 0.16833531, Global Avg Loss: 1.08547524, Time: 0.0247 Steps: 42500, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002059, Sample Num: 32944, Cur Loss: 0.62251866, Cur Avg Loss: 0.23012343, Log Avg loss: 0.23651356, Global Avg Loss: 1.08527553, Time: 0.0210 Steps: 42510, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002069, Sample Num: 33104, Cur Loss: 0.80326200, Cur Avg Loss: 0.23038045, Log Avg loss: 0.28330094, Global Avg Loss: 1.08508692, Time: 0.0210 Steps: 42520, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002079, Sample Num: 33264, Cur Loss: 0.22501969, Cur Avg Loss: 0.23053654, Log Avg loss: 0.26283094, Global Avg Loss: 1.08489358, Time: 0.0210 Steps: 42530, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002089, Sample Num: 33424, Cur Loss: 0.19264925, Cur Avg Loss: 0.23027559, Log Avg loss: 0.17602370, Global Avg Loss: 1.08467993, Time: 0.0210 Steps: 42540, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002099, Sample Num: 33584, Cur Loss: 0.16208361, Cur Avg Loss: 0.22989124, Log Avg loss: 0.14960056, Global Avg Loss: 1.08446017, Time: 0.0210 Steps: 42550, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002109, Sample Num: 33744, Cur Loss: 0.07184882, Cur Avg Loss: 0.22996814, Log Avg loss: 0.24610924, Global Avg Loss: 1.08426319, Time: 0.0210 Steps: 42560, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002119, Sample Num: 33904, Cur Loss: 0.28702444, Cur Avg Loss: 0.22982290, Log Avg loss: 0.19919189, Global Avg Loss: 1.08405528, Time: 0.0210 Steps: 42570, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002129, Sample Num: 34055, Cur Loss: 0.03386023, Cur Avg Loss: 0.23002942, Log Avg loss: 0.27379057, Global Avg Loss: 1.08386499, Time: 0.0102 Steps: 42580, Updated lr: 0.000061 ***** Running evaluation checkpoint-42580 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-42580 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.676041, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.243975, "eval_total_loss": 171.514539, "eval_mae": 0.338658, "eval_mse": 0.244052, "eval_r2": 0.844864, "eval_sp_statistic": 0.90774, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.922622, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.207825, "test_total_loss": 104.328132, "test_mae": 0.298537, "test_mse": 0.207897, "test_r2": 0.865821, "test_sp_statistic": 0.904349, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.938615, "test_ps_pvalue": 0.0, "lr": 6.056899004267425e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.083864989657521, "train_cur_epoch_loss": 489.73262864910066, "train_cur_epoch_avg_loss": 0.23002941693240989, "train_cur_epoch_time": 44.67604112625122, "train_cur_epoch_avg_time": 0.020984519082316213, "epoch": 20, "step": 42580} ################################################## Training, Epoch: 0021, Batch: 000010, Sample Num: 160, Cur Loss: 0.13487530, Cur Avg Loss: 0.22772885, Log Avg loss: 0.22772885, Global Avg Loss: 1.08366397, Time: 0.0209 Steps: 42590, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000020, Sample Num: 320, Cur Loss: 0.17263635, Cur Avg Loss: 0.25105476, Log Avg loss: 0.27438066, Global Avg Loss: 1.08347400, Time: 0.0210 Steps: 42600, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000030, Sample Num: 480, Cur Loss: 0.14416723, Cur Avg Loss: 0.24389974, Log Avg loss: 0.22958969, Global Avg Loss: 1.08327360, Time: 0.0209 Steps: 42610, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000040, Sample Num: 640, Cur Loss: 0.05922744, Cur Avg Loss: 0.25964028, Log Avg loss: 0.30686192, Global Avg Loss: 1.08309143, Time: 0.0210 Steps: 42620, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000050, Sample Num: 800, Cur Loss: 0.26578104, Cur Avg Loss: 0.25367294, Log Avg loss: 0.22980356, Global Avg Loss: 1.08289127, Time: 0.0209 Steps: 42630, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000060, Sample Num: 960, Cur Loss: 0.12976989, Cur Avg Loss: 0.23736727, Log Avg loss: 0.15583893, Global Avg Loss: 1.08267386, Time: 0.0210 Steps: 42640, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000070, Sample Num: 1120, Cur Loss: 0.34612092, Cur Avg Loss: 0.23242994, Log Avg loss: 0.20280594, Global Avg Loss: 1.08246756, Time: 0.0210 Steps: 42650, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000080, Sample Num: 1280, Cur Loss: 0.15928042, Cur Avg Loss: 0.23204979, Log Avg loss: 0.22938874, Global Avg Loss: 1.08226759, Time: 0.0210 Steps: 42660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000090, Sample Num: 1440, Cur Loss: 0.11855166, Cur Avg Loss: 0.23782529, Log Avg loss: 0.28402934, Global Avg Loss: 1.08208051, Time: 0.0210 Steps: 42670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000100, Sample Num: 1600, Cur Loss: 0.11624306, Cur Avg Loss: 0.23524559, Log Avg loss: 0.21202823, Global Avg Loss: 1.08187666, Time: 0.0210 Steps: 42680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000110, Sample Num: 1760, Cur Loss: 0.16960388, Cur Avg Loss: 0.22911237, Log Avg loss: 0.16778022, Global Avg Loss: 1.08166254, Time: 0.0209 Steps: 42690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000120, Sample Num: 1920, Cur Loss: 0.31423494, Cur Avg Loss: 0.22613619, Log Avg loss: 0.19339826, Global Avg Loss: 1.08145451, Time: 0.0209 Steps: 42700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000130, Sample Num: 2080, Cur Loss: 0.27911031, Cur Avg Loss: 0.22507133, Log Avg loss: 0.21229302, Global Avg Loss: 1.08125101, Time: 0.0210 Steps: 42710, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000140, Sample Num: 2240, Cur Loss: 0.48441190, Cur Avg Loss: 0.23395754, Log Avg loss: 0.34947827, Global Avg Loss: 1.08107971, Time: 0.0209 Steps: 42720, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000150, Sample Num: 2400, Cur Loss: 0.07312802, Cur Avg Loss: 0.23600591, Log Avg loss: 0.26468308, Global Avg Loss: 1.08088865, Time: 0.0210 Steps: 42730, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000160, Sample Num: 2560, Cur Loss: 0.31601986, Cur Avg Loss: 0.23573062, Log Avg loss: 0.23160124, Global Avg Loss: 1.08068994, Time: 0.0209 Steps: 42740, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000170, Sample Num: 2720, Cur Loss: 0.14518008, Cur Avg Loss: 0.24047503, Log Avg loss: 0.31638565, Global Avg Loss: 1.08051116, Time: 0.0209 Steps: 42750, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000180, Sample Num: 2880, Cur Loss: 0.21602674, Cur Avg Loss: 0.23712456, Log Avg loss: 0.18016654, Global Avg Loss: 1.08030060, Time: 0.0209 Steps: 42760, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000190, Sample Num: 3040, Cur Loss: 0.62389338, Cur Avg Loss: 0.23812610, Log Avg loss: 0.25615385, Global Avg Loss: 1.08010791, Time: 0.0209 Steps: 42770, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000200, Sample Num: 3200, Cur Loss: 0.17901978, Cur Avg Loss: 0.24341866, Log Avg loss: 0.34397719, Global Avg Loss: 1.07993583, Time: 0.0210 Steps: 42780, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000210, Sample Num: 3360, Cur Loss: 0.77354956, Cur Avg Loss: 0.24250725, Log Avg loss: 0.22427917, Global Avg Loss: 1.07973587, Time: 0.0210 Steps: 42790, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000220, Sample Num: 3520, Cur Loss: 0.19773906, Cur Avg Loss: 0.24056888, Log Avg loss: 0.19986305, Global Avg Loss: 1.07953029, Time: 0.0209 Steps: 42800, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000230, Sample Num: 3680, Cur Loss: 0.12626015, Cur Avg Loss: 0.23899453, Log Avg loss: 0.20435888, Global Avg Loss: 1.07932586, Time: 0.0209 Steps: 42810, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000240, Sample Num: 3840, Cur Loss: 0.16037577, Cur Avg Loss: 0.23844122, Log Avg loss: 0.22571509, Global Avg Loss: 1.07912651, Time: 0.0209 Steps: 42820, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000250, Sample Num: 4000, Cur Loss: 0.19513580, Cur Avg Loss: 0.24157838, Log Avg loss: 0.31687014, Global Avg Loss: 1.07894854, Time: 0.0210 Steps: 42830, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000260, Sample Num: 4160, Cur Loss: 0.27155912, Cur Avg Loss: 0.23953746, Log Avg loss: 0.18851434, Global Avg Loss: 1.07874069, Time: 0.0248 Steps: 42840, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000270, Sample Num: 4320, Cur Loss: 0.21675125, Cur Avg Loss: 0.23780316, Log Avg loss: 0.19271149, Global Avg Loss: 1.07853391, Time: 0.0212 Steps: 42850, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000280, Sample Num: 4480, Cur Loss: 0.16846988, Cur Avg Loss: 0.23699098, Log Avg loss: 0.21506209, Global Avg Loss: 1.07833245, Time: 0.0211 Steps: 42860, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000290, Sample Num: 4640, Cur Loss: 0.09001244, Cur Avg Loss: 0.23573472, Log Avg loss: 0.20055946, Global Avg Loss: 1.07812770, Time: 0.0211 Steps: 42870, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000300, Sample Num: 4800, Cur Loss: 0.06682496, Cur Avg Loss: 0.23322828, Log Avg loss: 0.16054154, Global Avg Loss: 1.07791371, Time: 0.0211 Steps: 42880, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000310, Sample Num: 4960, Cur Loss: 0.38657874, Cur Avg Loss: 0.23163673, Log Avg loss: 0.18389011, Global Avg Loss: 1.07770526, Time: 0.0210 Steps: 42890, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000320, Sample Num: 5120, Cur Loss: 0.36868092, Cur Avg Loss: 0.23220529, Log Avg loss: 0.24983065, Global Avg Loss: 1.07751228, Time: 0.0211 Steps: 42900, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000330, Sample Num: 5280, Cur Loss: 0.62618840, Cur Avg Loss: 0.23298870, Log Avg loss: 0.25805796, Global Avg Loss: 1.07732131, Time: 0.0211 Steps: 42910, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000340, Sample Num: 5440, Cur Loss: 0.16796130, Cur Avg Loss: 0.23120313, Log Avg loss: 0.17227924, Global Avg Loss: 1.07711045, Time: 0.0211 Steps: 42920, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000350, Sample Num: 5600, Cur Loss: 0.26158971, Cur Avg Loss: 0.23089519, Log Avg loss: 0.22042537, Global Avg Loss: 1.07691089, Time: 0.0210 Steps: 42930, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000360, Sample Num: 5760, Cur Loss: 0.24259540, Cur Avg Loss: 0.23182047, Log Avg loss: 0.26420508, Global Avg Loss: 1.07672163, Time: 0.0211 Steps: 42940, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000370, Sample Num: 5920, Cur Loss: 0.11130473, Cur Avg Loss: 0.22908167, Log Avg loss: 0.13048508, Global Avg Loss: 1.07650131, Time: 0.0211 Steps: 42950, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000380, Sample Num: 6080, Cur Loss: 0.11433595, Cur Avg Loss: 0.23068381, Log Avg loss: 0.28996292, Global Avg Loss: 1.07631823, Time: 0.0211 Steps: 42960, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000390, Sample Num: 6240, Cur Loss: 0.29256079, Cur Avg Loss: 0.23022426, Log Avg loss: 0.21276134, Global Avg Loss: 1.07611726, Time: 0.0211 Steps: 42970, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000400, Sample Num: 6400, Cur Loss: 0.16424644, Cur Avg Loss: 0.23070363, Log Avg loss: 0.24939884, Global Avg Loss: 1.07592491, Time: 0.0210 Steps: 42980, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000410, Sample Num: 6560, Cur Loss: 0.46052584, Cur Avg Loss: 0.23142496, Log Avg loss: 0.26027816, Global Avg Loss: 1.07573518, Time: 0.0211 Steps: 42990, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000420, Sample Num: 6720, Cur Loss: 0.10276672, Cur Avg Loss: 0.23021614, Log Avg loss: 0.18065455, Global Avg Loss: 1.07552702, Time: 0.0211 Steps: 43000, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000430, Sample Num: 6880, Cur Loss: 0.19009987, Cur Avg Loss: 0.22890255, Log Avg loss: 0.17373200, Global Avg Loss: 1.07531735, Time: 0.0210 Steps: 43010, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000440, Sample Num: 7040, Cur Loss: 0.23978795, Cur Avg Loss: 0.22784886, Log Avg loss: 0.18254027, Global Avg Loss: 1.07510983, Time: 0.0210 Steps: 43020, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000450, Sample Num: 7200, Cur Loss: 0.03077605, Cur Avg Loss: 0.22659623, Log Avg loss: 0.17148034, Global Avg Loss: 1.07489983, Time: 0.0211 Steps: 43030, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000460, Sample Num: 7360, Cur Loss: 0.31686041, Cur Avg Loss: 0.22563013, Log Avg loss: 0.18215589, Global Avg Loss: 1.07469241, Time: 0.0210 Steps: 43040, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000470, Sample Num: 7520, Cur Loss: 0.29203939, Cur Avg Loss: 0.22435163, Log Avg loss: 0.16554028, Global Avg Loss: 1.07448122, Time: 0.0210 Steps: 43050, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000480, Sample Num: 7680, Cur Loss: 0.19872695, Cur Avg Loss: 0.22381101, Log Avg loss: 0.19840213, Global Avg Loss: 1.07427776, Time: 0.0210 Steps: 43060, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000490, Sample Num: 7840, Cur Loss: 0.22065300, Cur Avg Loss: 0.22308723, Log Avg loss: 0.18834556, Global Avg Loss: 1.07407207, Time: 0.0210 Steps: 43070, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000500, Sample Num: 8000, Cur Loss: 0.10487738, Cur Avg Loss: 0.22196429, Log Avg loss: 0.16694039, Global Avg Loss: 1.07386150, Time: 0.0210 Steps: 43080, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000510, Sample Num: 8160, Cur Loss: 0.17018646, Cur Avg Loss: 0.22274813, Log Avg loss: 0.26194030, Global Avg Loss: 1.07367308, Time: 0.0210 Steps: 43090, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000520, Sample Num: 8320, Cur Loss: 0.06453662, Cur Avg Loss: 0.22189144, Log Avg loss: 0.17820013, Global Avg Loss: 1.07346531, Time: 0.0209 Steps: 43100, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000530, Sample Num: 8480, Cur Loss: 0.19369984, Cur Avg Loss: 0.22163997, Log Avg loss: 0.20856340, Global Avg Loss: 1.07326468, Time: 0.0209 Steps: 43110, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000540, Sample Num: 8640, Cur Loss: 0.11343619, Cur Avg Loss: 0.22141453, Log Avg loss: 0.20946630, Global Avg Loss: 1.07306436, Time: 0.0209 Steps: 43120, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000550, Sample Num: 8800, Cur Loss: 0.29767844, Cur Avg Loss: 0.22138852, Log Avg loss: 0.21998373, Global Avg Loss: 1.07286656, Time: 0.0209 Steps: 43130, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000560, Sample Num: 8960, Cur Loss: 0.20570558, Cur Avg Loss: 0.22302985, Log Avg loss: 0.31330315, Global Avg Loss: 1.07269050, Time: 0.0209 Steps: 43140, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000570, Sample Num: 9120, Cur Loss: 0.10334749, Cur Avg Loss: 0.22570492, Log Avg loss: 0.37550893, Global Avg Loss: 1.07252892, Time: 0.0209 Steps: 43150, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000580, Sample Num: 9280, Cur Loss: 0.07081073, Cur Avg Loss: 0.22525983, Log Avg loss: 0.19988988, Global Avg Loss: 1.07232674, Time: 0.0209 Steps: 43160, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000590, Sample Num: 9440, Cur Loss: 0.24738550, Cur Avg Loss: 0.22512075, Log Avg loss: 0.21705404, Global Avg Loss: 1.07212862, Time: 0.0209 Steps: 43170, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000600, Sample Num: 9600, Cur Loss: 0.22534984, Cur Avg Loss: 0.22479124, Log Avg loss: 0.20535017, Global Avg Loss: 1.07192788, Time: 0.0209 Steps: 43180, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000610, Sample Num: 9760, Cur Loss: 0.13303410, Cur Avg Loss: 0.22514549, Log Avg loss: 0.24640006, Global Avg Loss: 1.07173674, Time: 0.0209 Steps: 43190, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000620, Sample Num: 9920, Cur Loss: 0.19148311, Cur Avg Loss: 0.22526466, Log Avg loss: 0.23253410, Global Avg Loss: 1.07154248, Time: 0.0209 Steps: 43200, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000630, Sample Num: 10080, Cur Loss: 0.19037384, Cur Avg Loss: 0.22539378, Log Avg loss: 0.23339909, Global Avg Loss: 1.07134852, Time: 0.0209 Steps: 43210, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000640, Sample Num: 10240, Cur Loss: 0.18265513, Cur Avg Loss: 0.22487567, Log Avg loss: 0.19223501, Global Avg Loss: 1.07114511, Time: 0.0209 Steps: 43220, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000650, Sample Num: 10400, Cur Loss: 0.30361840, Cur Avg Loss: 0.22498348, Log Avg loss: 0.23188348, Global Avg Loss: 1.07095097, Time: 0.0209 Steps: 43230, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000660, Sample Num: 10560, Cur Loss: 0.03937803, Cur Avg Loss: 0.22394207, Log Avg loss: 0.15625031, Global Avg Loss: 1.07073943, Time: 0.0209 Steps: 43240, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000670, Sample Num: 10720, Cur Loss: 0.23399200, Cur Avg Loss: 0.22349280, Log Avg loss: 0.19384115, Global Avg Loss: 1.07053668, Time: 0.0208 Steps: 43250, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000680, Sample Num: 10880, Cur Loss: 0.08290109, Cur Avg Loss: 0.22316977, Log Avg loss: 0.20152628, Global Avg Loss: 1.07033580, Time: 0.0209 Steps: 43260, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000690, Sample Num: 11040, Cur Loss: 0.51047337, Cur Avg Loss: 0.22711784, Log Avg loss: 0.49558675, Global Avg Loss: 1.07020297, Time: 0.0209 Steps: 43270, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000700, Sample Num: 11200, Cur Loss: 0.40971574, Cur Avg Loss: 0.22692518, Log Avg loss: 0.21363163, Global Avg Loss: 1.07000506, Time: 0.0209 Steps: 43280, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000710, Sample Num: 11360, Cur Loss: 0.08205061, Cur Avg Loss: 0.22621955, Log Avg loss: 0.17682577, Global Avg Loss: 1.06979873, Time: 0.0209 Steps: 43290, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000720, Sample Num: 11520, Cur Loss: 0.41300714, Cur Avg Loss: 0.22642002, Log Avg loss: 0.24065325, Global Avg Loss: 1.06960724, Time: 0.0209 Steps: 43300, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000730, Sample Num: 11680, Cur Loss: 0.26969469, Cur Avg Loss: 0.22655891, Log Avg loss: 0.23655928, Global Avg Loss: 1.06941490, Time: 0.0208 Steps: 43310, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000740, Sample Num: 11840, Cur Loss: 0.14876924, Cur Avg Loss: 0.22697330, Log Avg loss: 0.25722325, Global Avg Loss: 1.06922741, Time: 0.0209 Steps: 43320, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000750, Sample Num: 12000, Cur Loss: 0.35850370, Cur Avg Loss: 0.22729937, Log Avg loss: 0.25142845, Global Avg Loss: 1.06903867, Time: 0.0208 Steps: 43330, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000760, Sample Num: 12160, Cur Loss: 0.34364226, Cur Avg Loss: 0.22766867, Log Avg loss: 0.25536655, Global Avg Loss: 1.06885093, Time: 0.0209 Steps: 43340, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000770, Sample Num: 12320, Cur Loss: 0.17742240, Cur Avg Loss: 0.22822164, Log Avg loss: 0.27024696, Global Avg Loss: 1.06866671, Time: 0.0246 Steps: 43350, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000780, Sample Num: 12480, Cur Loss: 0.60368502, Cur Avg Loss: 0.23026453, Log Avg loss: 0.38756729, Global Avg Loss: 1.06850963, Time: 0.0207 Steps: 43360, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000790, Sample Num: 12640, Cur Loss: 0.29773569, Cur Avg Loss: 0.23167861, Log Avg loss: 0.34197657, Global Avg Loss: 1.06834211, Time: 0.0208 Steps: 43370, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000800, Sample Num: 12800, Cur Loss: 0.22513688, Cur Avg Loss: 0.23209848, Log Avg loss: 0.26526836, Global Avg Loss: 1.06815699, Time: 0.0208 Steps: 43380, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000810, Sample Num: 12960, Cur Loss: 0.39591476, Cur Avg Loss: 0.23193230, Log Avg loss: 0.21863805, Global Avg Loss: 1.06796120, Time: 0.0208 Steps: 43390, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000820, Sample Num: 13120, Cur Loss: 0.08979014, Cur Avg Loss: 0.23203619, Log Avg loss: 0.24045133, Global Avg Loss: 1.06777053, Time: 0.0207 Steps: 43400, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000830, Sample Num: 13280, Cur Loss: 0.23054776, Cur Avg Loss: 0.23101969, Log Avg loss: 0.14766634, Global Avg Loss: 1.06755857, Time: 0.0207 Steps: 43410, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000840, Sample Num: 13440, Cur Loss: 0.13641645, Cur Avg Loss: 0.23030580, Log Avg loss: 0.17105312, Global Avg Loss: 1.06735210, Time: 0.0208 Steps: 43420, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000850, Sample Num: 13600, Cur Loss: 0.26403779, Cur Avg Loss: 0.23087966, Log Avg loss: 0.27908398, Global Avg Loss: 1.06717060, Time: 0.0208 Steps: 43430, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000860, Sample Num: 13760, Cur Loss: 0.21108423, Cur Avg Loss: 0.23037329, Log Avg loss: 0.18733171, Global Avg Loss: 1.06696805, Time: 0.0208 Steps: 43440, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000870, Sample Num: 13920, Cur Loss: 0.06670654, Cur Avg Loss: 0.23006896, Log Avg loss: 0.20389651, Global Avg Loss: 1.06676942, Time: 0.0207 Steps: 43450, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000880, Sample Num: 14080, Cur Loss: 0.16487801, Cur Avg Loss: 0.22948247, Log Avg loss: 0.17845770, Global Avg Loss: 1.06656502, Time: 0.0207 Steps: 43460, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000890, Sample Num: 14240, Cur Loss: 0.08537891, Cur Avg Loss: 0.22882966, Log Avg loss: 0.17138278, Global Avg Loss: 1.06635909, Time: 0.0208 Steps: 43470, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000900, Sample Num: 14400, Cur Loss: 0.06587876, Cur Avg Loss: 0.22811259, Log Avg loss: 0.16429310, Global Avg Loss: 1.06615162, Time: 0.0208 Steps: 43480, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000910, Sample Num: 14560, Cur Loss: 0.13283010, Cur Avg Loss: 0.22773069, Log Avg loss: 0.19335959, Global Avg Loss: 1.06595094, Time: 0.0207 Steps: 43490, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000920, Sample Num: 14720, Cur Loss: 0.19878843, Cur Avg Loss: 0.22789296, Log Avg loss: 0.24265965, Global Avg Loss: 1.06576167, Time: 0.0207 Steps: 43500, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000930, Sample Num: 14880, Cur Loss: 0.10000256, Cur Avg Loss: 0.22742824, Log Avg loss: 0.18467457, Global Avg Loss: 1.06555917, Time: 0.0208 Steps: 43510, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000940, Sample Num: 15040, Cur Loss: 0.12398057, Cur Avg Loss: 0.22766528, Log Avg loss: 0.24970990, Global Avg Loss: 1.06537171, Time: 0.0207 Steps: 43520, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000950, Sample Num: 15200, Cur Loss: 0.20659161, Cur Avg Loss: 0.22766525, Log Avg loss: 0.22766244, Global Avg Loss: 1.06517926, Time: 0.0207 Steps: 43530, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000960, Sample Num: 15360, Cur Loss: 0.12942651, Cur Avg Loss: 0.22658070, Log Avg loss: 0.12354846, Global Avg Loss: 1.06496299, Time: 0.0208 Steps: 43540, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000970, Sample Num: 15520, Cur Loss: 0.73355532, Cur Avg Loss: 0.22713801, Log Avg loss: 0.28063964, Global Avg Loss: 1.06478290, Time: 0.0208 Steps: 43550, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000980, Sample Num: 15680, Cur Loss: 0.18050084, Cur Avg Loss: 0.22762803, Log Avg loss: 0.27516030, Global Avg Loss: 1.06460162, Time: 0.0207 Steps: 43560, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000990, Sample Num: 15840, Cur Loss: 0.16754809, Cur Avg Loss: 0.22699039, Log Avg loss: 0.16450089, Global Avg Loss: 1.06439504, Time: 0.0207 Steps: 43570, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001000, Sample Num: 16000, Cur Loss: 0.09571069, Cur Avg Loss: 0.22599199, Log Avg loss: 0.12715070, Global Avg Loss: 1.06417997, Time: 0.0208 Steps: 43580, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001010, Sample Num: 16160, Cur Loss: 0.47245944, Cur Avg Loss: 0.22735038, Log Avg loss: 0.36318903, Global Avg Loss: 1.06401916, Time: 0.0208 Steps: 43590, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001020, Sample Num: 16320, Cur Loss: 0.23276839, Cur Avg Loss: 0.22765087, Log Avg loss: 0.25800075, Global Avg Loss: 1.06383429, Time: 0.0208 Steps: 43600, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001030, Sample Num: 16480, Cur Loss: 0.14873116, Cur Avg Loss: 0.22746005, Log Avg loss: 0.20799665, Global Avg Loss: 1.06363804, Time: 0.0245 Steps: 43610, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001040, Sample Num: 16640, Cur Loss: 0.11842484, Cur Avg Loss: 0.22807939, Log Avg loss: 0.29187100, Global Avg Loss: 1.06346111, Time: 0.0208 Steps: 43620, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001050, Sample Num: 16800, Cur Loss: 0.20876099, Cur Avg Loss: 0.22800949, Log Avg loss: 0.22073974, Global Avg Loss: 1.06326796, Time: 0.0208 Steps: 43630, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001060, Sample Num: 16960, Cur Loss: 0.20199277, Cur Avg Loss: 0.22732436, Log Avg loss: 0.15538612, Global Avg Loss: 1.06305992, Time: 0.0208 Steps: 43640, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001070, Sample Num: 17120, Cur Loss: 0.18649490, Cur Avg Loss: 0.22707600, Log Avg loss: 0.20074981, Global Avg Loss: 1.06286237, Time: 0.0208 Steps: 43650, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001080, Sample Num: 17280, Cur Loss: 0.12641788, Cur Avg Loss: 0.22685375, Log Avg loss: 0.20307340, Global Avg Loss: 1.06266544, Time: 0.0208 Steps: 43660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001090, Sample Num: 17440, Cur Loss: 0.10039465, Cur Avg Loss: 0.22706438, Log Avg loss: 0.24981170, Global Avg Loss: 1.06247931, Time: 0.0208 Steps: 43670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001100, Sample Num: 17600, Cur Loss: 0.07182182, Cur Avg Loss: 0.22784350, Log Avg loss: 0.31276830, Global Avg Loss: 1.06230767, Time: 0.0208 Steps: 43680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001110, Sample Num: 17760, Cur Loss: 0.12308446, Cur Avg Loss: 0.22742999, Log Avg loss: 0.18194391, Global Avg Loss: 1.06210617, Time: 0.0208 Steps: 43690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001120, Sample Num: 17920, Cur Loss: 0.04507827, Cur Avg Loss: 0.22746579, Log Avg loss: 0.23143914, Global Avg Loss: 1.06191609, Time: 0.0208 Steps: 43700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001130, Sample Num: 18080, Cur Loss: 0.06367580, Cur Avg Loss: 0.22657899, Log Avg loss: 0.12725747, Global Avg Loss: 1.06170225, Time: 0.0208 Steps: 43710, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001140, Sample Num: 18240, Cur Loss: 0.37101880, Cur Avg Loss: 0.22683954, Log Avg loss: 0.25628107, Global Avg Loss: 1.06151803, Time: 0.0207 Steps: 43720, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001150, Sample Num: 18400, Cur Loss: 0.18334571, Cur Avg Loss: 0.22641248, Log Avg loss: 0.17772842, Global Avg Loss: 1.06131593, Time: 0.0208 Steps: 43730, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001160, Sample Num: 18560, Cur Loss: 0.35697314, Cur Avg Loss: 0.22654206, Log Avg loss: 0.24144295, Global Avg Loss: 1.06112849, Time: 0.0208 Steps: 43740, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001170, Sample Num: 18720, Cur Loss: 0.78220892, Cur Avg Loss: 0.22715244, Log Avg loss: 0.29795726, Global Avg Loss: 1.06095405, Time: 0.0208 Steps: 43750, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001180, Sample Num: 18880, Cur Loss: 0.17519373, Cur Avg Loss: 0.22788918, Log Avg loss: 0.31408706, Global Avg Loss: 1.06078337, Time: 0.0208 Steps: 43760, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001190, Sample Num: 19040, Cur Loss: 0.26460195, Cur Avg Loss: 0.22813265, Log Avg loss: 0.25686219, Global Avg Loss: 1.06059971, Time: 0.0208 Steps: 43770, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001200, Sample Num: 19200, Cur Loss: 0.27174446, Cur Avg Loss: 0.22908767, Log Avg loss: 0.34273507, Global Avg Loss: 1.06043573, Time: 0.0208 Steps: 43780, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001210, Sample Num: 19360, Cur Loss: 0.53778732, Cur Avg Loss: 0.22995277, Log Avg loss: 0.33376534, Global Avg Loss: 1.06026979, Time: 0.0208 Steps: 43790, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001220, Sample Num: 19520, Cur Loss: 0.47285265, Cur Avg Loss: 0.23056630, Log Avg loss: 0.30480356, Global Avg Loss: 1.06009731, Time: 0.0207 Steps: 43800, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001230, Sample Num: 19680, Cur Loss: 0.12706950, Cur Avg Loss: 0.23061377, Log Avg loss: 0.23640430, Global Avg Loss: 1.05990929, Time: 0.0208 Steps: 43810, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001240, Sample Num: 19840, Cur Loss: 0.23267251, Cur Avg Loss: 0.23082030, Log Avg loss: 0.25622397, Global Avg Loss: 1.05972589, Time: 0.0208 Steps: 43820, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001250, Sample Num: 20000, Cur Loss: 0.07023408, Cur Avg Loss: 0.23058782, Log Avg loss: 0.20176024, Global Avg Loss: 1.05953014, Time: 0.0207 Steps: 43830, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001260, Sample Num: 20160, Cur Loss: 0.31212008, Cur Avg Loss: 0.23043571, Log Avg loss: 0.21142157, Global Avg Loss: 1.05933668, Time: 0.0208 Steps: 43840, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001270, Sample Num: 20320, Cur Loss: 0.15498227, Cur Avg Loss: 0.23097449, Log Avg loss: 0.29886132, Global Avg Loss: 1.05916326, Time: 0.0208 Steps: 43850, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001280, Sample Num: 20480, Cur Loss: 0.29629624, Cur Avg Loss: 0.23114229, Log Avg loss: 0.25245335, Global Avg Loss: 1.05897933, Time: 0.0253 Steps: 43860, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001290, Sample Num: 20640, Cur Loss: 0.39333621, Cur Avg Loss: 0.23128725, Log Avg loss: 0.24984183, Global Avg Loss: 1.05879489, Time: 0.0208 Steps: 43870, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001300, Sample Num: 20800, Cur Loss: 0.10348666, Cur Avg Loss: 0.23210797, Log Avg loss: 0.33798002, Global Avg Loss: 1.05863062, Time: 0.0207 Steps: 43880, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001310, Sample Num: 20960, Cur Loss: 0.24693799, Cur Avg Loss: 0.23195227, Log Avg loss: 0.21171240, Global Avg Loss: 1.05843766, Time: 0.0208 Steps: 43890, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001320, Sample Num: 21120, Cur Loss: 0.22315960, Cur Avg Loss: 0.23164616, Log Avg loss: 0.19154459, Global Avg Loss: 1.05824019, Time: 0.0208 Steps: 43900, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001330, Sample Num: 21280, Cur Loss: 0.20655844, Cur Avg Loss: 0.23161989, Log Avg loss: 0.22815288, Global Avg Loss: 1.05805114, Time: 0.0208 Steps: 43910, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001340, Sample Num: 21440, Cur Loss: 0.05747887, Cur Avg Loss: 0.23133422, Log Avg loss: 0.19334009, Global Avg Loss: 1.05785426, Time: 0.0207 Steps: 43920, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001350, Sample Num: 21600, Cur Loss: 0.11560287, Cur Avg Loss: 0.23093711, Log Avg loss: 0.17772421, Global Avg Loss: 1.05765391, Time: 0.0208 Steps: 43930, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001360, Sample Num: 21760, Cur Loss: 0.63434190, Cur Avg Loss: 0.23108997, Log Avg loss: 0.25172626, Global Avg Loss: 1.05747050, Time: 0.0207 Steps: 43940, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001370, Sample Num: 21920, Cur Loss: 0.13851343, Cur Avg Loss: 0.23170579, Log Avg loss: 0.31545698, Global Avg Loss: 1.05730167, Time: 0.0208 Steps: 43950, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001380, Sample Num: 22080, Cur Loss: 0.23722158, Cur Avg Loss: 0.23155964, Log Avg loss: 0.21153688, Global Avg Loss: 1.05710927, Time: 0.0207 Steps: 43960, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001390, Sample Num: 22240, Cur Loss: 0.17229062, Cur Avg Loss: 0.23111508, Log Avg loss: 0.16976564, Global Avg Loss: 1.05690746, Time: 0.0207 Steps: 43970, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001400, Sample Num: 22400, Cur Loss: 0.14706455, Cur Avg Loss: 0.23117122, Log Avg loss: 0.23897558, Global Avg Loss: 1.05672149, Time: 0.0208 Steps: 43980, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001410, Sample Num: 22560, Cur Loss: 0.25176364, Cur Avg Loss: 0.23113225, Log Avg loss: 0.22567682, Global Avg Loss: 1.05653257, Time: 0.0207 Steps: 43990, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001420, Sample Num: 22720, Cur Loss: 0.44763887, Cur Avg Loss: 0.23174951, Log Avg loss: 0.31878270, Global Avg Loss: 1.05636490, Time: 0.0208 Steps: 44000, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001430, Sample Num: 22880, Cur Loss: 0.03392822, Cur Avg Loss: 0.23155658, Log Avg loss: 0.20416043, Global Avg Loss: 1.05617126, Time: 0.0208 Steps: 44010, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001440, Sample Num: 23040, Cur Loss: 0.28524104, Cur Avg Loss: 0.23126079, Log Avg loss: 0.18896299, Global Avg Loss: 1.05597426, Time: 0.0208 Steps: 44020, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001450, Sample Num: 23200, Cur Loss: 0.12017021, Cur Avg Loss: 0.23128298, Log Avg loss: 0.23447808, Global Avg Loss: 1.05578768, Time: 0.0208 Steps: 44030, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001460, Sample Num: 23360, Cur Loss: 0.13825864, Cur Avg Loss: 0.23143449, Log Avg loss: 0.25340417, Global Avg Loss: 1.05560549, Time: 0.0208 Steps: 44040, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001470, Sample Num: 23520, Cur Loss: 0.29682419, Cur Avg Loss: 0.23169222, Log Avg loss: 0.26932072, Global Avg Loss: 1.05542699, Time: 0.0208 Steps: 44050, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001480, Sample Num: 23680, Cur Loss: 0.13990679, Cur Avg Loss: 0.23184396, Log Avg loss: 0.25414905, Global Avg Loss: 1.05524513, Time: 0.0207 Steps: 44060, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001490, Sample Num: 23840, Cur Loss: 0.40495431, Cur Avg Loss: 0.23167049, Log Avg loss: 0.20599684, Global Avg Loss: 1.05505242, Time: 0.0207 Steps: 44070, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001500, Sample Num: 24000, Cur Loss: 0.06428379, Cur Avg Loss: 0.23101964, Log Avg loss: 0.13404358, Global Avg Loss: 1.05484348, Time: 0.0208 Steps: 44080, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001510, Sample Num: 24160, Cur Loss: 0.16347167, Cur Avg Loss: 0.23086675, Log Avg loss: 0.20793217, Global Avg Loss: 1.05465140, Time: 0.0207 Steps: 44090, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001520, Sample Num: 24320, Cur Loss: 0.34992313, Cur Avg Loss: 0.23099306, Log Avg loss: 0.25006635, Global Avg Loss: 1.05446895, Time: 0.0207 Steps: 44100, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001530, Sample Num: 24480, Cur Loss: 0.08311792, Cur Avg Loss: 0.23099467, Log Avg loss: 0.23123927, Global Avg Loss: 1.05428232, Time: 0.0207 Steps: 44110, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001540, Sample Num: 24640, Cur Loss: 0.11625186, Cur Avg Loss: 0.23098327, Log Avg loss: 0.22923966, Global Avg Loss: 1.05409532, Time: 0.0232 Steps: 44120, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001550, Sample Num: 24800, Cur Loss: 0.15435944, Cur Avg Loss: 0.23043096, Log Avg loss: 0.14537523, Global Avg Loss: 1.05388940, Time: 0.0208 Steps: 44130, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001560, Sample Num: 24960, Cur Loss: 0.26486221, Cur Avg Loss: 0.23060671, Log Avg loss: 0.25784692, Global Avg Loss: 1.05370906, Time: 0.0207 Steps: 44140, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001570, Sample Num: 25120, Cur Loss: 0.24304128, Cur Avg Loss: 0.23059246, Log Avg loss: 0.22837011, Global Avg Loss: 1.05352212, Time: 0.0208 Steps: 44150, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001580, Sample Num: 25280, Cur Loss: 0.30094716, Cur Avg Loss: 0.23061225, Log Avg loss: 0.23371924, Global Avg Loss: 1.05333647, Time: 0.0207 Steps: 44160, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001590, Sample Num: 25440, Cur Loss: 0.46745789, Cur Avg Loss: 0.23066840, Log Avg loss: 0.23954077, Global Avg Loss: 1.05315223, Time: 0.0208 Steps: 44170, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001600, Sample Num: 25600, Cur Loss: 0.27048650, Cur Avg Loss: 0.23118414, Log Avg loss: 0.31318667, Global Avg Loss: 1.05298474, Time: 0.0208 Steps: 44180, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001610, Sample Num: 25760, Cur Loss: 0.37945378, Cur Avg Loss: 0.23161804, Log Avg loss: 0.30104157, Global Avg Loss: 1.05281458, Time: 0.0208 Steps: 44190, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001620, Sample Num: 25920, Cur Loss: 0.31200060, Cur Avg Loss: 0.23138977, Log Avg loss: 0.19463792, Global Avg Loss: 1.05262042, Time: 0.0208 Steps: 44200, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001630, Sample Num: 26080, Cur Loss: 0.11722469, Cur Avg Loss: 0.23120611, Log Avg loss: 0.20145415, Global Avg Loss: 1.05242789, Time: 0.0208 Steps: 44210, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001640, Sample Num: 26240, Cur Loss: 0.43915522, Cur Avg Loss: 0.23111776, Log Avg loss: 0.21671552, Global Avg Loss: 1.05223891, Time: 0.0207 Steps: 44220, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001650, Sample Num: 26400, Cur Loss: 0.65652382, Cur Avg Loss: 0.23200176, Log Avg loss: 0.37697921, Global Avg Loss: 1.05208623, Time: 0.0208 Steps: 44230, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001660, Sample Num: 26560, Cur Loss: 0.20566648, Cur Avg Loss: 0.23177609, Log Avg loss: 0.19453984, Global Avg Loss: 1.05189240, Time: 0.0207 Steps: 44240, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001670, Sample Num: 26720, Cur Loss: 0.64492708, Cur Avg Loss: 0.23150600, Log Avg loss: 0.18667043, Global Avg Loss: 1.05169686, Time: 0.0207 Steps: 44250, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001680, Sample Num: 26880, Cur Loss: 0.21879439, Cur Avg Loss: 0.23141069, Log Avg loss: 0.21549384, Global Avg Loss: 1.05150794, Time: 0.0207 Steps: 44260, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001690, Sample Num: 27040, Cur Loss: 0.08556939, Cur Avg Loss: 0.23107796, Log Avg loss: 0.17517918, Global Avg Loss: 1.05130998, Time: 0.0207 Steps: 44270, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001700, Sample Num: 27200, Cur Loss: 0.34127492, Cur Avg Loss: 0.23119333, Log Avg loss: 0.25069121, Global Avg Loss: 1.05112918, Time: 0.0207 Steps: 44280, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001710, Sample Num: 27360, Cur Loss: 0.14983012, Cur Avg Loss: 0.23080752, Log Avg loss: 0.16522063, Global Avg Loss: 1.05092915, Time: 0.0207 Steps: 44290, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001720, Sample Num: 27520, Cur Loss: 0.12964074, Cur Avg Loss: 0.23055422, Log Avg loss: 0.18723919, Global Avg Loss: 1.05073419, Time: 0.0208 Steps: 44300, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001730, Sample Num: 27680, Cur Loss: 0.12634364, Cur Avg Loss: 0.23059699, Log Avg loss: 0.23795419, Global Avg Loss: 1.05055076, Time: 0.0207 Steps: 44310, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001740, Sample Num: 27840, Cur Loss: 0.46588221, Cur Avg Loss: 0.23061630, Log Avg loss: 0.23395589, Global Avg Loss: 1.05036651, Time: 0.0207 Steps: 44320, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001750, Sample Num: 28000, Cur Loss: 0.31028312, Cur Avg Loss: 0.23066894, Log Avg loss: 0.23982815, Global Avg Loss: 1.05018367, Time: 0.0208 Steps: 44330, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001760, Sample Num: 28160, Cur Loss: 0.26890758, Cur Avg Loss: 0.23025179, Log Avg loss: 0.15725096, Global Avg Loss: 1.04998228, Time: 0.0208 Steps: 44340, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001770, Sample Num: 28320, Cur Loss: 0.45971864, Cur Avg Loss: 0.23035826, Log Avg loss: 0.24909761, Global Avg Loss: 1.04980170, Time: 0.0208 Steps: 44350, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001780, Sample Num: 28480, Cur Loss: 0.23533311, Cur Avg Loss: 0.23008072, Log Avg loss: 0.18095503, Global Avg Loss: 1.04960584, Time: 0.0207 Steps: 44360, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001790, Sample Num: 28640, Cur Loss: 0.24687879, Cur Avg Loss: 0.23020105, Log Avg loss: 0.25162089, Global Avg Loss: 1.04942599, Time: 0.0208 Steps: 44370, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001800, Sample Num: 28800, Cur Loss: 0.52717799, Cur Avg Loss: 0.23127455, Log Avg loss: 0.42343012, Global Avg Loss: 1.04928494, Time: 0.0208 Steps: 44380, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001810, Sample Num: 28960, Cur Loss: 0.22105445, Cur Avg Loss: 0.23154145, Log Avg loss: 0.27958470, Global Avg Loss: 1.04911154, Time: 0.0207 Steps: 44390, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001820, Sample Num: 29120, Cur Loss: 0.16899619, Cur Avg Loss: 0.23153280, Log Avg loss: 0.22996664, Global Avg Loss: 1.04892705, Time: 0.0207 Steps: 44400, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001830, Sample Num: 29280, Cur Loss: 0.13266721, Cur Avg Loss: 0.23145154, Log Avg loss: 0.21666190, Global Avg Loss: 1.04873964, Time: 0.0207 Steps: 44410, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001840, Sample Num: 29440, Cur Loss: 0.14529610, Cur Avg Loss: 0.23155925, Log Avg loss: 0.25127074, Global Avg Loss: 1.04856011, Time: 0.0208 Steps: 44420, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001850, Sample Num: 29600, Cur Loss: 0.03292135, Cur Avg Loss: 0.23135010, Log Avg loss: 0.19286699, Global Avg Loss: 1.04836752, Time: 0.0207 Steps: 44430, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001860, Sample Num: 29760, Cur Loss: 0.18510477, Cur Avg Loss: 0.23131180, Log Avg loss: 0.22422467, Global Avg Loss: 1.04818207, Time: 0.0207 Steps: 44440, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001870, Sample Num: 29920, Cur Loss: 0.20148477, Cur Avg Loss: 0.23134335, Log Avg loss: 0.23721275, Global Avg Loss: 1.04799962, Time: 0.0207 Steps: 44450, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001880, Sample Num: 30080, Cur Loss: 0.59336495, Cur Avg Loss: 0.23131389, Log Avg loss: 0.22580533, Global Avg Loss: 1.04781470, Time: 0.0207 Steps: 44460, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001890, Sample Num: 30240, Cur Loss: 0.03924554, Cur Avg Loss: 0.23115574, Log Avg loss: 0.20142372, Global Avg Loss: 1.04762437, Time: 0.0207 Steps: 44470, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001900, Sample Num: 30400, Cur Loss: 0.17564529, Cur Avg Loss: 0.23141454, Log Avg loss: 0.28032677, Global Avg Loss: 1.04745186, Time: 0.0207 Steps: 44480, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001910, Sample Num: 30560, Cur Loss: 0.73957968, Cur Avg Loss: 0.23171821, Log Avg loss: 0.28941490, Global Avg Loss: 1.04728148, Time: 0.0207 Steps: 44490, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001920, Sample Num: 30720, Cur Loss: 0.20860779, Cur Avg Loss: 0.23165625, Log Avg loss: 0.21982287, Global Avg Loss: 1.04709553, Time: 0.0207 Steps: 44500, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001930, Sample Num: 30880, Cur Loss: 0.38345593, Cur Avg Loss: 0.23145625, Log Avg loss: 0.19305525, Global Avg Loss: 1.04690366, Time: 0.0207 Steps: 44510, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001940, Sample Num: 31040, Cur Loss: 0.06213835, Cur Avg Loss: 0.23142743, Log Avg loss: 0.22586559, Global Avg Loss: 1.04671924, Time: 0.0207 Steps: 44520, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001950, Sample Num: 31200, Cur Loss: 0.12041262, Cur Avg Loss: 0.23091391, Log Avg loss: 0.13129073, Global Avg Loss: 1.04651366, Time: 0.0207 Steps: 44530, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001960, Sample Num: 31360, Cur Loss: 0.23380820, Cur Avg Loss: 0.23101975, Log Avg loss: 0.25165817, Global Avg Loss: 1.04633520, Time: 0.0207 Steps: 44540, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001970, Sample Num: 31520, Cur Loss: 0.29333547, Cur Avg Loss: 0.23079428, Log Avg loss: 0.18660340, Global Avg Loss: 1.04614222, Time: 0.0207 Steps: 44550, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001980, Sample Num: 31680, Cur Loss: 0.24299765, Cur Avg Loss: 0.23059194, Log Avg loss: 0.19073003, Global Avg Loss: 1.04595025, Time: 0.0207 Steps: 44560, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001990, Sample Num: 31840, Cur Loss: 0.25321224, Cur Avg Loss: 0.23049107, Log Avg loss: 0.21051987, Global Avg Loss: 1.04576281, Time: 0.0208 Steps: 44570, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002000, Sample Num: 32000, Cur Loss: 0.45899862, Cur Avg Loss: 0.23058224, Log Avg loss: 0.24872517, Global Avg Loss: 1.04558402, Time: 0.0207 Steps: 44580, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002010, Sample Num: 32160, Cur Loss: 0.20179783, Cur Avg Loss: 0.23064422, Log Avg loss: 0.24303922, Global Avg Loss: 1.04540404, Time: 0.0207 Steps: 44590, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002020, Sample Num: 32320, Cur Loss: 0.11089054, Cur Avg Loss: 0.23072785, Log Avg loss: 0.24753743, Global Avg Loss: 1.04522515, Time: 0.0207 Steps: 44600, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002030, Sample Num: 32480, Cur Loss: 0.32233310, Cur Avg Loss: 0.23058377, Log Avg loss: 0.20148066, Global Avg Loss: 1.04503601, Time: 0.0207 Steps: 44610, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002040, Sample Num: 32640, Cur Loss: 0.22689001, Cur Avg Loss: 0.23052084, Log Avg loss: 0.21774490, Global Avg Loss: 1.04485060, Time: 0.0207 Steps: 44620, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002050, Sample Num: 32800, Cur Loss: 0.11574985, Cur Avg Loss: 0.23017841, Log Avg loss: 0.16032419, Global Avg Loss: 1.04465241, Time: 0.0247 Steps: 44630, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002060, Sample Num: 32960, Cur Loss: 0.32765442, Cur Avg Loss: 0.23006183, Log Avg loss: 0.20616150, Global Avg Loss: 1.04446457, Time: 0.0209 Steps: 44640, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002070, Sample Num: 33120, Cur Loss: 0.31922463, Cur Avg Loss: 0.23019328, Log Avg loss: 0.25727197, Global Avg Loss: 1.04428827, Time: 0.0209 Steps: 44650, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002080, Sample Num: 33280, Cur Loss: 0.22241268, Cur Avg Loss: 0.22995177, Log Avg loss: 0.17995980, Global Avg Loss: 1.04409474, Time: 0.0209 Steps: 44660, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002090, Sample Num: 33440, Cur Loss: 0.17642877, Cur Avg Loss: 0.22956660, Log Avg loss: 0.14945152, Global Avg Loss: 1.04389446, Time: 0.0209 Steps: 44670, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002100, Sample Num: 33600, Cur Loss: 0.40375501, Cur Avg Loss: 0.22973171, Log Avg loss: 0.26423927, Global Avg Loss: 1.04371996, Time: 0.0209 Steps: 44680, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002110, Sample Num: 33760, Cur Loss: 0.81651473, Cur Avg Loss: 0.23043679, Log Avg loss: 0.37850289, Global Avg Loss: 1.04357111, Time: 0.0208 Steps: 44690, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002120, Sample Num: 33920, Cur Loss: 0.14079240, Cur Avg Loss: 0.23036491, Log Avg loss: 0.21519930, Global Avg Loss: 1.04338579, Time: 0.0209 Steps: 44700, Updated lr: 0.000059 ***** Running evaluation checkpoint-44709 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-44709 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.549698, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.372856, "eval_total_loss": 262.117634, "eval_mae": 0.399533, "eval_mse": 0.373007, "eval_r2": 0.762892, "eval_sp_statistic": 0.913778, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.921046, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.276362, "test_total_loss": 138.733764, "test_mae": 0.314003, "test_mse": 0.276463, "test_r2": 0.821569, "test_sp_statistic": 0.906809, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946312, "test_ps_pvalue": 0.0, "lr": 5.855002370791844e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.0432171628911973, "train_cur_epoch_loss": 490.2248760852963, "train_cur_epoch_avg_loss": 0.23026062756472349, "train_cur_epoch_time": 44.54969811439514, "train_cur_epoch_avg_time": 0.0209251752533561, "epoch": 21, "step": 44709} ################################################## Training, Epoch: 0022, Batch: 000001, Sample Num: 16, Cur Loss: 0.30693424, Cur Avg Loss: 0.30693424, Log Avg loss: 0.21581985, Global Avg Loss: 1.04320069, Time: 0.0248 Steps: 44710, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000011, Sample Num: 176, Cur Loss: 0.14508763, Cur Avg Loss: 0.19691071, Log Avg loss: 0.18590836, Global Avg Loss: 1.04300899, Time: 0.0208 Steps: 44720, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000021, Sample Num: 336, Cur Loss: 0.16971371, Cur Avg Loss: 0.19068576, Log Avg loss: 0.18383830, Global Avg Loss: 1.04281691, Time: 0.0209 Steps: 44730, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000031, Sample Num: 496, Cur Loss: 0.14163832, Cur Avg Loss: 0.19234609, Log Avg loss: 0.19583280, Global Avg Loss: 1.04262760, Time: 0.0209 Steps: 44740, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000041, Sample Num: 656, Cur Loss: 0.19659421, Cur Avg Loss: 0.18987248, Log Avg loss: 0.18220427, Global Avg Loss: 1.04243533, Time: 0.0210 Steps: 44750, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000051, Sample Num: 816, Cur Loss: 0.28669330, Cur Avg Loss: 0.19654944, Log Avg loss: 0.22392498, Global Avg Loss: 1.04225246, Time: 0.0210 Steps: 44760, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000061, Sample Num: 976, Cur Loss: 0.10460356, Cur Avg Loss: 0.19577209, Log Avg loss: 0.19180757, Global Avg Loss: 1.04206250, Time: 0.0209 Steps: 44770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000071, Sample Num: 1136, Cur Loss: 0.18824768, Cur Avg Loss: 0.20639766, Log Avg loss: 0.27121365, Global Avg Loss: 1.04189036, Time: 0.0210 Steps: 44780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000081, Sample Num: 1296, Cur Loss: 0.22414553, Cur Avg Loss: 0.20269364, Log Avg loss: 0.17639510, Global Avg Loss: 1.04169713, Time: 0.0209 Steps: 44790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000091, Sample Num: 1456, Cur Loss: 0.18078327, Cur Avg Loss: 0.19895873, Log Avg loss: 0.16870594, Global Avg Loss: 1.04150226, Time: 0.0209 Steps: 44800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000101, Sample Num: 1616, Cur Loss: 0.37796694, Cur Avg Loss: 0.19935178, Log Avg loss: 0.20292855, Global Avg Loss: 1.04131512, Time: 0.0210 Steps: 44810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000111, Sample Num: 1776, Cur Loss: 0.31896186, Cur Avg Loss: 0.20452433, Log Avg loss: 0.25676711, Global Avg Loss: 1.04114008, Time: 0.0209 Steps: 44820, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000121, Sample Num: 1936, Cur Loss: 0.34837198, Cur Avg Loss: 0.21674075, Log Avg loss: 0.35234296, Global Avg Loss: 1.04098643, Time: 0.0210 Steps: 44830, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000131, Sample Num: 2096, Cur Loss: 0.04396950, Cur Avg Loss: 0.21646767, Log Avg loss: 0.21316338, Global Avg Loss: 1.04080182, Time: 0.0209 Steps: 44840, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000141, Sample Num: 2256, Cur Loss: 0.09942234, Cur Avg Loss: 0.21495050, Log Avg loss: 0.19507557, Global Avg Loss: 1.04061325, Time: 0.0209 Steps: 44850, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000151, Sample Num: 2416, Cur Loss: 0.27267492, Cur Avg Loss: 0.21797486, Log Avg loss: 0.26061837, Global Avg Loss: 1.04043937, Time: 0.0209 Steps: 44860, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000161, Sample Num: 2576, Cur Loss: 0.15122133, Cur Avg Loss: 0.22212044, Log Avg loss: 0.28471874, Global Avg Loss: 1.04027095, Time: 0.0210 Steps: 44870, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000171, Sample Num: 2736, Cur Loss: 0.21240407, Cur Avg Loss: 0.22911294, Log Avg loss: 0.34169217, Global Avg Loss: 1.04011530, Time: 0.0210 Steps: 44880, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000181, Sample Num: 2896, Cur Loss: 0.33082277, Cur Avg Loss: 0.23174360, Log Avg loss: 0.27672798, Global Avg Loss: 1.03994524, Time: 0.0208 Steps: 44890, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000191, Sample Num: 3056, Cur Loss: 0.16472143, Cur Avg Loss: 0.23034461, Log Avg loss: 0.20502289, Global Avg Loss: 1.03975929, Time: 0.0210 Steps: 44900, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000201, Sample Num: 3216, Cur Loss: 0.20025006, Cur Avg Loss: 0.22616439, Log Avg loss: 0.14632210, Global Avg Loss: 1.03956035, Time: 0.0209 Steps: 44910, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000211, Sample Num: 3376, Cur Loss: 0.12808293, Cur Avg Loss: 0.22541806, Log Avg loss: 0.21041672, Global Avg Loss: 1.03937576, Time: 0.0209 Steps: 44920, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000221, Sample Num: 3536, Cur Loss: 0.12140220, Cur Avg Loss: 0.22238324, Log Avg loss: 0.15834873, Global Avg Loss: 1.03917968, Time: 0.0210 Steps: 44930, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000231, Sample Num: 3696, Cur Loss: 0.30624852, Cur Avg Loss: 0.22035199, Log Avg loss: 0.17546116, Global Avg Loss: 1.03898748, Time: 0.0209 Steps: 44940, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000241, Sample Num: 3856, Cur Loss: 0.84311777, Cur Avg Loss: 0.21996563, Log Avg loss: 0.21104082, Global Avg Loss: 1.03880329, Time: 0.0209 Steps: 44950, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000251, Sample Num: 4016, Cur Loss: 0.23408091, Cur Avg Loss: 0.21848476, Log Avg loss: 0.18279584, Global Avg Loss: 1.03861290, Time: 0.0210 Steps: 44960, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000261, Sample Num: 4176, Cur Loss: 0.05625200, Cur Avg Loss: 0.21690321, Log Avg loss: 0.17720633, Global Avg Loss: 1.03842134, Time: 0.0208 Steps: 44970, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000271, Sample Num: 4336, Cur Loss: 0.12159280, Cur Avg Loss: 0.21565104, Log Avg loss: 0.18296922, Global Avg Loss: 1.03823116, Time: 0.0207 Steps: 44980, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000281, Sample Num: 4496, Cur Loss: 0.27897522, Cur Avg Loss: 0.21535978, Log Avg loss: 0.20746679, Global Avg Loss: 1.03804650, Time: 0.0208 Steps: 44990, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000291, Sample Num: 4656, Cur Loss: 0.04562996, Cur Avg Loss: 0.21324868, Log Avg loss: 0.15392683, Global Avg Loss: 1.03785003, Time: 0.0208 Steps: 45000, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000301, Sample Num: 4816, Cur Loss: 0.54876906, Cur Avg Loss: 0.21704377, Log Avg loss: 0.32748073, Global Avg Loss: 1.03769221, Time: 0.0208 Steps: 45010, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000311, Sample Num: 4976, Cur Loss: 0.11079323, Cur Avg Loss: 0.21502023, Log Avg loss: 0.15411184, Global Avg Loss: 1.03749594, Time: 0.0208 Steps: 45020, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000321, Sample Num: 5136, Cur Loss: 0.30153647, Cur Avg Loss: 0.21466461, Log Avg loss: 0.20360466, Global Avg Loss: 1.03731076, Time: 0.0208 Steps: 45030, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000331, Sample Num: 5296, Cur Loss: 0.20513499, Cur Avg Loss: 0.21220556, Log Avg loss: 0.13327010, Global Avg Loss: 1.03711004, Time: 0.0208 Steps: 45040, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000341, Sample Num: 5456, Cur Loss: 0.12649253, Cur Avg Loss: 0.20990012, Log Avg loss: 0.13358995, Global Avg Loss: 1.03690948, Time: 0.0208 Steps: 45050, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000351, Sample Num: 5616, Cur Loss: 0.23079419, Cur Avg Loss: 0.20806978, Log Avg loss: 0.14565532, Global Avg Loss: 1.03671169, Time: 0.0208 Steps: 45060, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000361, Sample Num: 5776, Cur Loss: 0.22301883, Cur Avg Loss: 0.20774540, Log Avg loss: 0.19635976, Global Avg Loss: 1.03652523, Time: 0.0208 Steps: 45070, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000371, Sample Num: 5936, Cur Loss: 0.12646842, Cur Avg Loss: 0.20761674, Log Avg loss: 0.20297208, Global Avg Loss: 1.03634033, Time: 0.0208 Steps: 45080, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000381, Sample Num: 6096, Cur Loss: 0.14862262, Cur Avg Loss: 0.20944588, Log Avg loss: 0.27730684, Global Avg Loss: 1.03617199, Time: 0.0208 Steps: 45090, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000391, Sample Num: 6256, Cur Loss: 0.13354778, Cur Avg Loss: 0.20949356, Log Avg loss: 0.21131036, Global Avg Loss: 1.03598909, Time: 0.0207 Steps: 45100, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000401, Sample Num: 6416, Cur Loss: 0.23458247, Cur Avg Loss: 0.21091117, Log Avg loss: 0.26633969, Global Avg Loss: 1.03581848, Time: 0.0208 Steps: 45110, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000411, Sample Num: 6576, Cur Loss: 0.28342181, Cur Avg Loss: 0.21142010, Log Avg loss: 0.23182793, Global Avg Loss: 1.03564029, Time: 0.0208 Steps: 45120, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000421, Sample Num: 6736, Cur Loss: 0.23500144, Cur Avg Loss: 0.21048177, Log Avg loss: 0.17191658, Global Avg Loss: 1.03544890, Time: 0.0208 Steps: 45130, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000431, Sample Num: 6896, Cur Loss: 0.28955978, Cur Avg Loss: 0.21146073, Log Avg loss: 0.25267471, Global Avg Loss: 1.03527549, Time: 0.0208 Steps: 45140, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000441, Sample Num: 7056, Cur Loss: 0.16405225, Cur Avg Loss: 0.21004665, Log Avg loss: 0.14910002, Global Avg Loss: 1.03507922, Time: 0.0208 Steps: 45150, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000451, Sample Num: 7216, Cur Loss: 0.02635711, Cur Avg Loss: 0.20952397, Log Avg loss: 0.18647392, Global Avg Loss: 1.03489131, Time: 0.0208 Steps: 45160, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000461, Sample Num: 7376, Cur Loss: 0.27187100, Cur Avg Loss: 0.21067919, Log Avg loss: 0.26277958, Global Avg Loss: 1.03472037, Time: 0.0208 Steps: 45170, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000471, Sample Num: 7536, Cur Loss: 0.12553632, Cur Avg Loss: 0.21129701, Log Avg loss: 0.23977843, Global Avg Loss: 1.03454442, Time: 0.0208 Steps: 45180, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000481, Sample Num: 7696, Cur Loss: 0.11410130, Cur Avg Loss: 0.21048380, Log Avg loss: 0.17218180, Global Avg Loss: 1.03435359, Time: 0.0208 Steps: 45190, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000491, Sample Num: 7856, Cur Loss: 0.03759888, Cur Avg Loss: 0.21085508, Log Avg loss: 0.22871363, Global Avg Loss: 1.03417535, Time: 0.0208 Steps: 45200, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000501, Sample Num: 8016, Cur Loss: 0.25322416, Cur Avg Loss: 0.21276047, Log Avg loss: 0.30631472, Global Avg Loss: 1.03401436, Time: 0.0207 Steps: 45210, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000511, Sample Num: 8176, Cur Loss: 0.15177694, Cur Avg Loss: 0.21185942, Log Avg loss: 0.16671725, Global Avg Loss: 1.03382256, Time: 0.0207 Steps: 45220, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000521, Sample Num: 8336, Cur Loss: 0.33040127, Cur Avg Loss: 0.21332362, Log Avg loss: 0.28814412, Global Avg Loss: 1.03365770, Time: 0.0208 Steps: 45230, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000531, Sample Num: 8496, Cur Loss: 0.31949201, Cur Avg Loss: 0.21281586, Log Avg loss: 0.18636122, Global Avg Loss: 1.03347041, Time: 0.0207 Steps: 45240, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000541, Sample Num: 8656, Cur Loss: 0.05737347, Cur Avg Loss: 0.21259151, Log Avg loss: 0.20067850, Global Avg Loss: 1.03328637, Time: 0.0208 Steps: 45250, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000551, Sample Num: 8816, Cur Loss: 0.08639742, Cur Avg Loss: 0.21222929, Log Avg loss: 0.19263371, Global Avg Loss: 1.03310063, Time: 0.0207 Steps: 45260, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000561, Sample Num: 8976, Cur Loss: 0.26338452, Cur Avg Loss: 0.21241720, Log Avg loss: 0.22277080, Global Avg Loss: 1.03292163, Time: 0.0207 Steps: 45270, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000571, Sample Num: 9136, Cur Loss: 0.51475418, Cur Avg Loss: 0.21177366, Log Avg loss: 0.17567127, Global Avg Loss: 1.03273231, Time: 0.0207 Steps: 45280, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000581, Sample Num: 9296, Cur Loss: 0.23514087, Cur Avg Loss: 0.21312055, Log Avg loss: 0.29002776, Global Avg Loss: 1.03256832, Time: 0.0208 Steps: 45290, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000591, Sample Num: 9456, Cur Loss: 0.05384657, Cur Avg Loss: 0.21170052, Log Avg loss: 0.12919697, Global Avg Loss: 1.03236890, Time: 0.0208 Steps: 45300, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000601, Sample Num: 9616, Cur Loss: 0.07215494, Cur Avg Loss: 0.21127396, Log Avg loss: 0.18606427, Global Avg Loss: 1.03218212, Time: 0.0207 Steps: 45310, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000611, Sample Num: 9776, Cur Loss: 0.45184526, Cur Avg Loss: 0.21083163, Log Avg loss: 0.18424738, Global Avg Loss: 1.03199502, Time: 0.0207 Steps: 45320, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000621, Sample Num: 9936, Cur Loss: 0.12690321, Cur Avg Loss: 0.20929048, Log Avg loss: 0.11512617, Global Avg Loss: 1.03179275, Time: 0.0208 Steps: 45330, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000631, Sample Num: 10096, Cur Loss: 0.64285213, Cur Avg Loss: 0.20978236, Log Avg loss: 0.24032802, Global Avg Loss: 1.03161819, Time: 0.0207 Steps: 45340, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000641, Sample Num: 10256, Cur Loss: 0.58724999, Cur Avg Loss: 0.21022963, Log Avg loss: 0.23845235, Global Avg Loss: 1.03144329, Time: 0.0207 Steps: 45350, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000651, Sample Num: 10416, Cur Loss: 0.07992655, Cur Avg Loss: 0.20964005, Log Avg loss: 0.17184827, Global Avg Loss: 1.03125379, Time: 0.0208 Steps: 45360, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000661, Sample Num: 10576, Cur Loss: 0.21051790, Cur Avg Loss: 0.21057096, Log Avg loss: 0.27117289, Global Avg Loss: 1.03108626, Time: 0.0208 Steps: 45370, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000671, Sample Num: 10736, Cur Loss: 0.13937350, Cur Avg Loss: 0.21117804, Log Avg loss: 0.25130597, Global Avg Loss: 1.03091442, Time: 0.0208 Steps: 45380, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000681, Sample Num: 10896, Cur Loss: 0.20491134, Cur Avg Loss: 0.21131318, Log Avg loss: 0.22038147, Global Avg Loss: 1.03073585, Time: 0.0207 Steps: 45390, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000691, Sample Num: 11056, Cur Loss: 0.28414640, Cur Avg Loss: 0.21027940, Log Avg loss: 0.13987880, Global Avg Loss: 1.03053963, Time: 0.0208 Steps: 45400, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000701, Sample Num: 11216, Cur Loss: 0.32335013, Cur Avg Loss: 0.21130564, Log Avg loss: 0.28221857, Global Avg Loss: 1.03037484, Time: 0.0208 Steps: 45410, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000711, Sample Num: 11376, Cur Loss: 0.12676209, Cur Avg Loss: 0.21156941, Log Avg loss: 0.23005982, Global Avg Loss: 1.03019863, Time: 0.0208 Steps: 45420, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000721, Sample Num: 11536, Cur Loss: 0.05518623, Cur Avg Loss: 0.21104241, Log Avg loss: 0.17357317, Global Avg Loss: 1.03001008, Time: 0.0207 Steps: 45430, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000731, Sample Num: 11696, Cur Loss: 0.09247337, Cur Avg Loss: 0.21080124, Log Avg loss: 0.19341237, Global Avg Loss: 1.02982596, Time: 0.0207 Steps: 45440, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000741, Sample Num: 11856, Cur Loss: 0.23074798, Cur Avg Loss: 0.21004864, Log Avg loss: 0.15503413, Global Avg Loss: 1.02963349, Time: 0.0207 Steps: 45450, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000751, Sample Num: 12016, Cur Loss: 0.19398054, Cur Avg Loss: 0.21081313, Log Avg loss: 0.26746115, Global Avg Loss: 1.02946583, Time: 0.0207 Steps: 45460, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000761, Sample Num: 12176, Cur Loss: 0.17119052, Cur Avg Loss: 0.21070808, Log Avg loss: 0.20281952, Global Avg Loss: 1.02928403, Time: 0.0207 Steps: 45470, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000771, Sample Num: 12336, Cur Loss: 0.24789748, Cur Avg Loss: 0.21167769, Log Avg loss: 0.28546462, Global Avg Loss: 1.02912048, Time: 0.0244 Steps: 45480, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000781, Sample Num: 12496, Cur Loss: 0.07100650, Cur Avg Loss: 0.21109054, Log Avg loss: 0.16582133, Global Avg Loss: 1.02893071, Time: 0.0208 Steps: 45490, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000791, Sample Num: 12656, Cur Loss: 0.08151890, Cur Avg Loss: 0.21067110, Log Avg loss: 0.17791290, Global Avg Loss: 1.02874367, Time: 0.0207 Steps: 45500, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000801, Sample Num: 12816, Cur Loss: 0.28796577, Cur Avg Loss: 0.21039003, Log Avg loss: 0.18815752, Global Avg Loss: 1.02855897, Time: 0.0207 Steps: 45510, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000811, Sample Num: 12976, Cur Loss: 0.39596128, Cur Avg Loss: 0.21080677, Log Avg loss: 0.24418707, Global Avg Loss: 1.02838665, Time: 0.0207 Steps: 45520, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000821, Sample Num: 13136, Cur Loss: 0.10180530, Cur Avg Loss: 0.21102055, Log Avg loss: 0.22835852, Global Avg Loss: 1.02821094, Time: 0.0207 Steps: 45530, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000831, Sample Num: 13296, Cur Loss: 0.22942337, Cur Avg Loss: 0.21140874, Log Avg loss: 0.24327941, Global Avg Loss: 1.02803858, Time: 0.0207 Steps: 45540, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000841, Sample Num: 13456, Cur Loss: 0.08929895, Cur Avg Loss: 0.21068727, Log Avg loss: 0.15073302, Global Avg Loss: 1.02784597, Time: 0.0207 Steps: 45550, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000851, Sample Num: 13616, Cur Loss: 0.05007771, Cur Avg Loss: 0.20982972, Log Avg loss: 0.13770911, Global Avg Loss: 1.02765060, Time: 0.0209 Steps: 45560, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000861, Sample Num: 13776, Cur Loss: 0.06672338, Cur Avg Loss: 0.21004952, Log Avg loss: 0.22875463, Global Avg Loss: 1.02747529, Time: 0.0209 Steps: 45570, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000871, Sample Num: 13936, Cur Loss: 0.19561338, Cur Avg Loss: 0.20963484, Log Avg loss: 0.17393109, Global Avg Loss: 1.02728802, Time: 0.0207 Steps: 45580, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000881, Sample Num: 14096, Cur Loss: 0.18612489, Cur Avg Loss: 0.21038466, Log Avg loss: 0.27569432, Global Avg Loss: 1.02712316, Time: 0.0208 Steps: 45590, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000891, Sample Num: 14256, Cur Loss: 0.12555085, Cur Avg Loss: 0.20953685, Log Avg loss: 0.13484441, Global Avg Loss: 1.02692749, Time: 0.0208 Steps: 45600, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000901, Sample Num: 14416, Cur Loss: 0.21320295, Cur Avg Loss: 0.20905160, Log Avg loss: 0.16581614, Global Avg Loss: 1.02673869, Time: 0.0207 Steps: 45610, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000911, Sample Num: 14576, Cur Loss: 0.17896143, Cur Avg Loss: 0.20833013, Log Avg loss: 0.14332525, Global Avg Loss: 1.02654504, Time: 0.0207 Steps: 45620, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000921, Sample Num: 14736, Cur Loss: 0.04966261, Cur Avg Loss: 0.20803750, Log Avg loss: 0.18137914, Global Avg Loss: 1.02635982, Time: 0.0207 Steps: 45630, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000931, Sample Num: 14896, Cur Loss: 0.17550936, Cur Avg Loss: 0.20802897, Log Avg loss: 0.20724337, Global Avg Loss: 1.02618035, Time: 0.0207 Steps: 45640, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000941, Sample Num: 15056, Cur Loss: 0.07034902, Cur Avg Loss: 0.20808076, Log Avg loss: 0.21290237, Global Avg Loss: 1.02600219, Time: 0.0208 Steps: 45650, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000951, Sample Num: 15216, Cur Loss: 0.27506691, Cur Avg Loss: 0.20884547, Log Avg loss: 0.28080438, Global Avg Loss: 1.02583899, Time: 0.0208 Steps: 45660, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000961, Sample Num: 15376, Cur Loss: 0.22097605, Cur Avg Loss: 0.20941941, Log Avg loss: 0.26400119, Global Avg Loss: 1.02567217, Time: 0.0208 Steps: 45670, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000971, Sample Num: 15536, Cur Loss: 0.14953753, Cur Avg Loss: 0.20895596, Log Avg loss: 0.16441906, Global Avg Loss: 1.02548363, Time: 0.0208 Steps: 45680, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000981, Sample Num: 15696, Cur Loss: 0.36781523, Cur Avg Loss: 0.20937977, Log Avg loss: 0.25053173, Global Avg Loss: 1.02531402, Time: 0.0208 Steps: 45690, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000991, Sample Num: 15856, Cur Loss: 0.15794677, Cur Avg Loss: 0.20905287, Log Avg loss: 0.17698330, Global Avg Loss: 1.02512839, Time: 0.0208 Steps: 45700, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001001, Sample Num: 16016, Cur Loss: 0.40182239, Cur Avg Loss: 0.20987530, Log Avg loss: 0.29137837, Global Avg Loss: 1.02496787, Time: 0.0209 Steps: 45710, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001011, Sample Num: 16176, Cur Loss: 0.29225466, Cur Avg Loss: 0.21084280, Log Avg loss: 0.30768959, Global Avg Loss: 1.02481098, Time: 0.0208 Steps: 45720, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001021, Sample Num: 16336, Cur Loss: 0.15531097, Cur Avg Loss: 0.21061516, Log Avg loss: 0.18760094, Global Avg Loss: 1.02462791, Time: 0.0208 Steps: 45730, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001031, Sample Num: 16496, Cur Loss: 0.11475052, Cur Avg Loss: 0.21044063, Log Avg loss: 0.19262080, Global Avg Loss: 1.02444601, Time: 0.0211 Steps: 45740, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001041, Sample Num: 16656, Cur Loss: 0.13395210, Cur Avg Loss: 0.20978189, Log Avg loss: 0.14186633, Global Avg Loss: 1.02425309, Time: 0.0208 Steps: 45750, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001051, Sample Num: 16816, Cur Loss: 0.32619113, Cur Avg Loss: 0.20939388, Log Avg loss: 0.16900144, Global Avg Loss: 1.02406620, Time: 0.0208 Steps: 45760, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001061, Sample Num: 16976, Cur Loss: 0.19719429, Cur Avg Loss: 0.21006613, Log Avg loss: 0.28071972, Global Avg Loss: 1.02390379, Time: 0.0208 Steps: 45770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001071, Sample Num: 17136, Cur Loss: 0.17195788, Cur Avg Loss: 0.20960737, Log Avg loss: 0.16093279, Global Avg Loss: 1.02371528, Time: 0.0208 Steps: 45780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001081, Sample Num: 17296, Cur Loss: 0.17442426, Cur Avg Loss: 0.21020444, Log Avg loss: 0.27415092, Global Avg Loss: 1.02355159, Time: 0.0208 Steps: 45790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001091, Sample Num: 17456, Cur Loss: 0.14266591, Cur Avg Loss: 0.21055412, Log Avg loss: 0.24835454, Global Avg Loss: 1.02338233, Time: 0.0208 Steps: 45800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001101, Sample Num: 17616, Cur Loss: 0.15718783, Cur Avg Loss: 0.21064613, Log Avg loss: 0.22068467, Global Avg Loss: 1.02320711, Time: 0.0208 Steps: 45810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001111, Sample Num: 17776, Cur Loss: 0.26636958, Cur Avg Loss: 0.21074512, Log Avg loss: 0.22164384, Global Avg Loss: 1.02303217, Time: 0.0208 Steps: 45820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001121, Sample Num: 17936, Cur Loss: 0.22228497, Cur Avg Loss: 0.21036533, Log Avg loss: 0.16817046, Global Avg Loss: 1.02284564, Time: 0.0208 Steps: 45830, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001131, Sample Num: 18096, Cur Loss: 0.10661210, Cur Avg Loss: 0.21000244, Log Avg loss: 0.16932227, Global Avg Loss: 1.02265944, Time: 0.0208 Steps: 45840, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001141, Sample Num: 18256, Cur Loss: 0.17731838, Cur Avg Loss: 0.20994536, Log Avg loss: 0.20349023, Global Avg Loss: 1.02248078, Time: 0.0208 Steps: 45850, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001151, Sample Num: 18416, Cur Loss: 0.09266549, Cur Avg Loss: 0.21004820, Log Avg loss: 0.22178150, Global Avg Loss: 1.02230618, Time: 0.0208 Steps: 45860, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001161, Sample Num: 18576, Cur Loss: 0.26507479, Cur Avg Loss: 0.20938965, Log Avg loss: 0.13359137, Global Avg Loss: 1.02211244, Time: 0.0208 Steps: 45870, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001171, Sample Num: 18736, Cur Loss: 0.41914576, Cur Avg Loss: 0.21004916, Log Avg loss: 0.28661781, Global Avg Loss: 1.02195213, Time: 0.0208 Steps: 45880, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001181, Sample Num: 18896, Cur Loss: 0.16990867, Cur Avg Loss: 0.21033029, Log Avg loss: 0.24325033, Global Avg Loss: 1.02178244, Time: 0.0208 Steps: 45890, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001191, Sample Num: 19056, Cur Loss: 0.24792720, Cur Avg Loss: 0.20994169, Log Avg loss: 0.16404880, Global Avg Loss: 1.02159557, Time: 0.0208 Steps: 45900, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001201, Sample Num: 19216, Cur Loss: 0.09290823, Cur Avg Loss: 0.20970235, Log Avg loss: 0.18119591, Global Avg Loss: 1.02141252, Time: 0.0208 Steps: 45910, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001211, Sample Num: 19376, Cur Loss: 0.22792722, Cur Avg Loss: 0.20963940, Log Avg loss: 0.20207939, Global Avg Loss: 1.02123409, Time: 0.0208 Steps: 45920, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001221, Sample Num: 19536, Cur Loss: 0.16671595, Cur Avg Loss: 0.20975158, Log Avg loss: 0.22333687, Global Avg Loss: 1.02106037, Time: 0.0208 Steps: 45930, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001231, Sample Num: 19696, Cur Loss: 0.27795225, Cur Avg Loss: 0.21060280, Log Avg loss: 0.31453714, Global Avg Loss: 1.02090658, Time: 0.0208 Steps: 45940, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001241, Sample Num: 19856, Cur Loss: 0.13711642, Cur Avg Loss: 0.21082657, Log Avg loss: 0.23837212, Global Avg Loss: 1.02073628, Time: 0.0208 Steps: 45950, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001251, Sample Num: 20016, Cur Loss: 0.55874205, Cur Avg Loss: 0.21139448, Log Avg loss: 0.28187187, Global Avg Loss: 1.02057551, Time: 0.0209 Steps: 45960, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001261, Sample Num: 20176, Cur Loss: 0.32453096, Cur Avg Loss: 0.21222825, Log Avg loss: 0.31653326, Global Avg Loss: 1.02042236, Time: 0.0208 Steps: 45970, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001271, Sample Num: 20336, Cur Loss: 0.59458941, Cur Avg Loss: 0.21298394, Log Avg loss: 0.30827615, Global Avg Loss: 1.02026748, Time: 0.0208 Steps: 45980, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001281, Sample Num: 20496, Cur Loss: 0.09476888, Cur Avg Loss: 0.21326983, Log Avg loss: 0.24960662, Global Avg Loss: 1.02009991, Time: 0.0245 Steps: 45990, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001291, Sample Num: 20656, Cur Loss: 0.22143054, Cur Avg Loss: 0.21369660, Log Avg loss: 0.26836554, Global Avg Loss: 1.01993649, Time: 0.0208 Steps: 46000, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001301, Sample Num: 20816, Cur Loss: 0.39798394, Cur Avg Loss: 0.21320926, Log Avg loss: 0.15029390, Global Avg Loss: 1.01974748, Time: 0.0208 Steps: 46010, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001311, Sample Num: 20976, Cur Loss: 0.14518458, Cur Avg Loss: 0.21342798, Log Avg loss: 0.24188397, Global Avg Loss: 1.01957845, Time: 0.0208 Steps: 46020, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001321, Sample Num: 21136, Cur Loss: 0.27965322, Cur Avg Loss: 0.21337020, Log Avg loss: 0.20579480, Global Avg Loss: 1.01940165, Time: 0.0208 Steps: 46030, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001331, Sample Num: 21296, Cur Loss: 0.21589342, Cur Avg Loss: 0.21351852, Log Avg loss: 0.23311214, Global Avg Loss: 1.01923087, Time: 0.0207 Steps: 46040, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001341, Sample Num: 21456, Cur Loss: 0.07472541, Cur Avg Loss: 0.21336868, Log Avg loss: 0.19342465, Global Avg Loss: 1.01905154, Time: 0.0208 Steps: 46050, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001351, Sample Num: 21616, Cur Loss: 0.12454002, Cur Avg Loss: 0.21279179, Log Avg loss: 0.13543007, Global Avg Loss: 1.01885970, Time: 0.0207 Steps: 46060, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001361, Sample Num: 21776, Cur Loss: 0.26599756, Cur Avg Loss: 0.21278315, Log Avg loss: 0.21161578, Global Avg Loss: 1.01868448, Time: 0.0208 Steps: 46070, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001371, Sample Num: 21936, Cur Loss: 0.37568468, Cur Avg Loss: 0.21310333, Log Avg loss: 0.25668019, Global Avg Loss: 1.01851911, Time: 0.0208 Steps: 46080, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001381, Sample Num: 22096, Cur Loss: 0.44922918, Cur Avg Loss: 0.21401295, Log Avg loss: 0.33872212, Global Avg Loss: 1.01837162, Time: 0.0208 Steps: 46090, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001391, Sample Num: 22256, Cur Loss: 0.24494562, Cur Avg Loss: 0.21435209, Log Avg loss: 0.26118759, Global Avg Loss: 1.01820737, Time: 0.0208 Steps: 46100, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001401, Sample Num: 22416, Cur Loss: 0.43309760, Cur Avg Loss: 0.21415338, Log Avg loss: 0.18651211, Global Avg Loss: 1.01802700, Time: 0.0208 Steps: 46110, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001411, Sample Num: 22576, Cur Loss: 0.27795103, Cur Avg Loss: 0.21397223, Log Avg loss: 0.18859395, Global Avg Loss: 1.01784716, Time: 0.0208 Steps: 46120, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001421, Sample Num: 22736, Cur Loss: 0.08383086, Cur Avg Loss: 0.21354653, Log Avg loss: 0.15348057, Global Avg Loss: 1.01765978, Time: 0.0208 Steps: 46130, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001431, Sample Num: 22896, Cur Loss: 0.23900270, Cur Avg Loss: 0.21347216, Log Avg loss: 0.20290385, Global Avg Loss: 1.01748320, Time: 0.0208 Steps: 46140, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001441, Sample Num: 23056, Cur Loss: 0.16047066, Cur Avg Loss: 0.21357295, Log Avg loss: 0.22799622, Global Avg Loss: 1.01731213, Time: 0.0207 Steps: 46150, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001451, Sample Num: 23216, Cur Loss: 0.08623544, Cur Avg Loss: 0.21329285, Log Avg loss: 0.17292966, Global Avg Loss: 1.01712920, Time: 0.0208 Steps: 46160, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001461, Sample Num: 23376, Cur Loss: 0.11527012, Cur Avg Loss: 0.21313165, Log Avg loss: 0.18974130, Global Avg Loss: 1.01695000, Time: 0.0208 Steps: 46170, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001471, Sample Num: 23536, Cur Loss: 0.28266457, Cur Avg Loss: 0.21271518, Log Avg loss: 0.15186981, Global Avg Loss: 1.01676267, Time: 0.0208 Steps: 46180, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001481, Sample Num: 23696, Cur Loss: 0.27808341, Cur Avg Loss: 0.21279331, Log Avg loss: 0.22428602, Global Avg Loss: 1.01659110, Time: 0.0208 Steps: 46190, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001491, Sample Num: 23856, Cur Loss: 0.39467812, Cur Avg Loss: 0.21267704, Log Avg loss: 0.19545687, Global Avg Loss: 1.01641337, Time: 0.0208 Steps: 46200, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001501, Sample Num: 24016, Cur Loss: 0.14225200, Cur Avg Loss: 0.21345701, Log Avg loss: 0.32975100, Global Avg Loss: 1.01626477, Time: 0.0208 Steps: 46210, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001511, Sample Num: 24176, Cur Loss: 0.11861783, Cur Avg Loss: 0.21341062, Log Avg loss: 0.20644740, Global Avg Loss: 1.01608956, Time: 0.0208 Steps: 46220, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001521, Sample Num: 24336, Cur Loss: 0.17827219, Cur Avg Loss: 0.21343783, Log Avg loss: 0.21754914, Global Avg Loss: 1.01591683, Time: 0.0208 Steps: 46230, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001531, Sample Num: 24496, Cur Loss: 0.29479524, Cur Avg Loss: 0.21375293, Log Avg loss: 0.26168032, Global Avg Loss: 1.01575372, Time: 0.0208 Steps: 46240, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001541, Sample Num: 24656, Cur Loss: 0.13030298, Cur Avg Loss: 0.21511707, Log Avg loss: 0.42396562, Global Avg Loss: 1.01562576, Time: 0.0245 Steps: 46250, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001551, Sample Num: 24816, Cur Loss: 0.49406606, Cur Avg Loss: 0.21585424, Log Avg loss: 0.32945353, Global Avg Loss: 1.01547743, Time: 0.0208 Steps: 46260, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001561, Sample Num: 24976, Cur Loss: 0.33449367, Cur Avg Loss: 0.21608119, Log Avg loss: 0.25128112, Global Avg Loss: 1.01531227, Time: 0.0208 Steps: 46270, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001571, Sample Num: 25136, Cur Loss: 0.26942462, Cur Avg Loss: 0.21643565, Log Avg loss: 0.27176589, Global Avg Loss: 1.01515161, Time: 0.0208 Steps: 46280, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001581, Sample Num: 25296, Cur Loss: 0.51046920, Cur Avg Loss: 0.21673375, Log Avg loss: 0.26356627, Global Avg Loss: 1.01498925, Time: 0.0208 Steps: 46290, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001591, Sample Num: 25456, Cur Loss: 0.09863493, Cur Avg Loss: 0.21634604, Log Avg loss: 0.15504862, Global Avg Loss: 1.01480351, Time: 0.0208 Steps: 46300, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001601, Sample Num: 25616, Cur Loss: 0.17183374, Cur Avg Loss: 0.21648918, Log Avg loss: 0.23926287, Global Avg Loss: 1.01463605, Time: 0.0208 Steps: 46310, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001611, Sample Num: 25776, Cur Loss: 0.09847329, Cur Avg Loss: 0.21638552, Log Avg loss: 0.19978910, Global Avg Loss: 1.01446013, Time: 0.0208 Steps: 46320, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001621, Sample Num: 25936, Cur Loss: 0.19265151, Cur Avg Loss: 0.21618274, Log Avg loss: 0.18351420, Global Avg Loss: 1.01428078, Time: 0.0208 Steps: 46330, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001631, Sample Num: 26096, Cur Loss: 0.37976390, Cur Avg Loss: 0.21653961, Log Avg loss: 0.27438833, Global Avg Loss: 1.01412111, Time: 0.0208 Steps: 46340, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001641, Sample Num: 26256, Cur Loss: 0.08856803, Cur Avg Loss: 0.21735471, Log Avg loss: 0.35029752, Global Avg Loss: 1.01397789, Time: 0.0208 Steps: 46350, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001651, Sample Num: 26416, Cur Loss: 0.28732157, Cur Avg Loss: 0.21705923, Log Avg loss: 0.16857100, Global Avg Loss: 1.01379553, Time: 0.0208 Steps: 46360, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001661, Sample Num: 26576, Cur Loss: 0.18612716, Cur Avg Loss: 0.21659979, Log Avg loss: 0.14074672, Global Avg Loss: 1.01360725, Time: 0.0207 Steps: 46370, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001671, Sample Num: 26736, Cur Loss: 0.18695739, Cur Avg Loss: 0.21660779, Log Avg loss: 0.21793748, Global Avg Loss: 1.01343570, Time: 0.0208 Steps: 46380, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001681, Sample Num: 26896, Cur Loss: 0.22808832, Cur Avg Loss: 0.21677666, Log Avg loss: 0.24499463, Global Avg Loss: 1.01327005, Time: 0.0208 Steps: 46390, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001691, Sample Num: 27056, Cur Loss: 0.22973050, Cur Avg Loss: 0.21716184, Log Avg loss: 0.28191047, Global Avg Loss: 1.01311243, Time: 0.0208 Steps: 46400, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001701, Sample Num: 27216, Cur Loss: 0.06353509, Cur Avg Loss: 0.21774425, Log Avg loss: 0.31622943, Global Avg Loss: 1.01296227, Time: 0.0208 Steps: 46410, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001711, Sample Num: 27376, Cur Loss: 0.09975123, Cur Avg Loss: 0.21877874, Log Avg loss: 0.39474611, Global Avg Loss: 1.01282909, Time: 0.0208 Steps: 46420, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001721, Sample Num: 27536, Cur Loss: 0.33884862, Cur Avg Loss: 0.22006694, Log Avg loss: 0.44047664, Global Avg Loss: 1.01270582, Time: 0.0208 Steps: 46430, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001731, Sample Num: 27696, Cur Loss: 0.10955115, Cur Avg Loss: 0.21973935, Log Avg loss: 0.16336115, Global Avg Loss: 1.01252293, Time: 0.0208 Steps: 46440, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001741, Sample Num: 27856, Cur Loss: 0.08436817, Cur Avg Loss: 0.21958509, Log Avg loss: 0.19288321, Global Avg Loss: 1.01234648, Time: 0.0208 Steps: 46450, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001751, Sample Num: 28016, Cur Loss: 0.07460141, Cur Avg Loss: 0.21937617, Log Avg loss: 0.18300345, Global Avg Loss: 1.01216797, Time: 0.0207 Steps: 46460, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001761, Sample Num: 28176, Cur Loss: 0.09585510, Cur Avg Loss: 0.21918482, Log Avg loss: 0.18567895, Global Avg Loss: 1.01199011, Time: 0.0207 Steps: 46470, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001771, Sample Num: 28336, Cur Loss: 0.20281664, Cur Avg Loss: 0.21900976, Log Avg loss: 0.18818153, Global Avg Loss: 1.01181287, Time: 0.0207 Steps: 46480, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001781, Sample Num: 28496, Cur Loss: 0.12432982, Cur Avg Loss: 0.21966378, Log Avg loss: 0.33549168, Global Avg Loss: 1.01166740, Time: 0.0208 Steps: 46490, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001791, Sample Num: 28656, Cur Loss: 0.14944606, Cur Avg Loss: 0.21922090, Log Avg loss: 0.14034271, Global Avg Loss: 1.01148002, Time: 0.0208 Steps: 46500, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001801, Sample Num: 28816, Cur Loss: 0.27529132, Cur Avg Loss: 0.21956976, Log Avg loss: 0.28205124, Global Avg Loss: 1.01132318, Time: 0.0208 Steps: 46510, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001811, Sample Num: 28976, Cur Loss: 0.04105980, Cur Avg Loss: 0.21926029, Log Avg loss: 0.16352558, Global Avg Loss: 1.01114094, Time: 0.0208 Steps: 46520, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001821, Sample Num: 29136, Cur Loss: 0.18351763, Cur Avg Loss: 0.21933137, Log Avg loss: 0.23220263, Global Avg Loss: 1.01097353, Time: 0.0208 Steps: 46530, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001831, Sample Num: 29296, Cur Loss: 0.21833839, Cur Avg Loss: 0.21917755, Log Avg loss: 0.19116670, Global Avg Loss: 1.01079738, Time: 0.0208 Steps: 46540, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001841, Sample Num: 29456, Cur Loss: 0.34009463, Cur Avg Loss: 0.21924111, Log Avg loss: 0.23088055, Global Avg Loss: 1.01062984, Time: 0.0207 Steps: 46550, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001851, Sample Num: 29616, Cur Loss: 0.42787308, Cur Avg Loss: 0.21961138, Log Avg loss: 0.28777692, Global Avg Loss: 1.01047459, Time: 0.0208 Steps: 46560, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001861, Sample Num: 29776, Cur Loss: 0.39479023, Cur Avg Loss: 0.21970628, Log Avg loss: 0.23727197, Global Avg Loss: 1.01030856, Time: 0.0208 Steps: 46570, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001871, Sample Num: 29936, Cur Loss: 0.10566311, Cur Avg Loss: 0.21952097, Log Avg loss: 0.18503620, Global Avg Loss: 1.01013138, Time: 0.0208 Steps: 46580, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001881, Sample Num: 30096, Cur Loss: 0.17439029, Cur Avg Loss: 0.21938195, Log Avg loss: 0.19337109, Global Avg Loss: 1.00995608, Time: 0.0207 Steps: 46590, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001891, Sample Num: 30256, Cur Loss: 0.13265665, Cur Avg Loss: 0.21949752, Log Avg loss: 0.24123628, Global Avg Loss: 1.00979111, Time: 0.0208 Steps: 46600, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001901, Sample Num: 30416, Cur Loss: 0.14664626, Cur Avg Loss: 0.21933091, Log Avg loss: 0.18782532, Global Avg Loss: 1.00961477, Time: 0.0208 Steps: 46610, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001911, Sample Num: 30576, Cur Loss: 0.47919482, Cur Avg Loss: 0.21963694, Log Avg loss: 0.27781281, Global Avg Loss: 1.00945779, Time: 0.0208 Steps: 46620, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001921, Sample Num: 30736, Cur Loss: 0.25174606, Cur Avg Loss: 0.21979787, Log Avg loss: 0.25055167, Global Avg Loss: 1.00929504, Time: 0.0208 Steps: 46630, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001931, Sample Num: 30896, Cur Loss: 0.28439987, Cur Avg Loss: 0.21965693, Log Avg loss: 0.19258154, Global Avg Loss: 1.00911993, Time: 0.0207 Steps: 46640, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001941, Sample Num: 31056, Cur Loss: 0.48304254, Cur Avg Loss: 0.21984064, Log Avg loss: 0.25531448, Global Avg Loss: 1.00895835, Time: 0.0207 Steps: 46650, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001951, Sample Num: 31216, Cur Loss: 0.49718437, Cur Avg Loss: 0.22163860, Log Avg loss: 0.57062399, Global Avg Loss: 1.00886440, Time: 0.0208 Steps: 46660, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001961, Sample Num: 31376, Cur Loss: 0.18843791, Cur Avg Loss: 0.22179105, Log Avg loss: 0.25153375, Global Avg Loss: 1.00870213, Time: 0.0208 Steps: 46670, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001971, Sample Num: 31536, Cur Loss: 0.40098664, Cur Avg Loss: 0.22158777, Log Avg loss: 0.18172459, Global Avg Loss: 1.00852497, Time: 0.0208 Steps: 46680, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001981, Sample Num: 31696, Cur Loss: 0.13162746, Cur Avg Loss: 0.22151671, Log Avg loss: 0.20751077, Global Avg Loss: 1.00835341, Time: 0.0207 Steps: 46690, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001991, Sample Num: 31856, Cur Loss: 0.23483713, Cur Avg Loss: 0.22131508, Log Avg loss: 0.18137111, Global Avg Loss: 1.00817633, Time: 0.0208 Steps: 46700, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002001, Sample Num: 32016, Cur Loss: 0.08147855, Cur Avg Loss: 0.22117455, Log Avg loss: 0.19319650, Global Avg Loss: 1.00800185, Time: 0.0208 Steps: 46710, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002011, Sample Num: 32176, Cur Loss: 0.26818043, Cur Avg Loss: 0.22119527, Log Avg loss: 0.22534119, Global Avg Loss: 1.00783433, Time: 0.0208 Steps: 46720, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002021, Sample Num: 32336, Cur Loss: 0.08240435, Cur Avg Loss: 0.22086329, Log Avg loss: 0.15410182, Global Avg Loss: 1.00765163, Time: 0.0208 Steps: 46730, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002031, Sample Num: 32496, Cur Loss: 0.19877204, Cur Avg Loss: 0.22070655, Log Avg loss: 0.18902842, Global Avg Loss: 1.00747649, Time: 0.0208 Steps: 46740, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002041, Sample Num: 32656, Cur Loss: 0.46600038, Cur Avg Loss: 0.22039206, Log Avg loss: 0.15652035, Global Avg Loss: 1.00729447, Time: 0.0208 Steps: 46750, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002051, Sample Num: 32816, Cur Loss: 0.37485915, Cur Avg Loss: 0.22026367, Log Avg loss: 0.19405869, Global Avg Loss: 1.00712055, Time: 0.0247 Steps: 46760, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002061, Sample Num: 32976, Cur Loss: 0.04269297, Cur Avg Loss: 0.21979663, Log Avg loss: 0.12400732, Global Avg Loss: 1.00693173, Time: 0.0210 Steps: 46770, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002071, Sample Num: 33136, Cur Loss: 0.30921042, Cur Avg Loss: 0.21956137, Log Avg loss: 0.17107371, Global Avg Loss: 1.00675305, Time: 0.0210 Steps: 46780, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002081, Sample Num: 33296, Cur Loss: 0.07521015, Cur Avg Loss: 0.21919967, Log Avg loss: 0.14429198, Global Avg Loss: 1.00656873, Time: 0.0210 Steps: 46790, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002091, Sample Num: 33456, Cur Loss: 0.05761750, Cur Avg Loss: 0.21897753, Log Avg loss: 0.17275055, Global Avg Loss: 1.00639056, Time: 0.0209 Steps: 46800, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002101, Sample Num: 33616, Cur Loss: 0.08577168, Cur Avg Loss: 0.21922515, Log Avg loss: 0.27100215, Global Avg Loss: 1.00623346, Time: 0.0209 Steps: 46810, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002111, Sample Num: 33776, Cur Loss: 0.17451507, Cur Avg Loss: 0.21917847, Log Avg loss: 0.20937082, Global Avg Loss: 1.00606326, Time: 0.0209 Steps: 46820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002121, Sample Num: 33936, Cur Loss: 0.22794437, Cur Avg Loss: 0.21937097, Log Avg loss: 0.26000842, Global Avg Loss: 1.00590395, Time: 0.0210 Steps: 46830, Updated lr: 0.000057 ***** Running evaluation checkpoint-46838 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-46838 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.606809, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.257572, "eval_total_loss": 181.072876, "eval_mae": 0.334684, "eval_mse": 0.257663, "eval_r2": 0.836213, "eval_sp_statistic": 0.915904, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.925194, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.197752, "test_total_loss": 99.271573, "test_mae": 0.333918, "test_mse": 0.19782, "test_r2": 0.872325, "test_sp_statistic": 0.907003, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947986, "test_ps_pvalue": 0.0, "lr": 5.653105737316264e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.005759850611877, "train_cur_epoch_loss": 466.5837472565472, "train_cur_epoch_avg_loss": 0.21915629274614712, "train_cur_epoch_time": 44.60680866241455, "train_cur_epoch_avg_time": 0.02095200031113882, "epoch": 22, "step": 46838} ################################################## Training, Epoch: 0023, Batch: 000002, Sample Num: 32, Cur Loss: 0.33696252, Cur Avg Loss: 0.19317372, Log Avg loss: 0.16842590, Global Avg Loss: 1.00572515, Time: 0.0247 Steps: 46840, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000012, Sample Num: 192, Cur Loss: 0.04568134, Cur Avg Loss: 0.19397352, Log Avg loss: 0.19413348, Global Avg Loss: 1.00555192, Time: 0.0209 Steps: 46850, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000022, Sample Num: 352, Cur Loss: 0.11809091, Cur Avg Loss: 0.20678804, Log Avg loss: 0.22216547, Global Avg Loss: 1.00538475, Time: 0.0209 Steps: 46860, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000032, Sample Num: 512, Cur Loss: 0.21465473, Cur Avg Loss: 0.19511351, Log Avg loss: 0.16942953, Global Avg Loss: 1.00520639, Time: 0.0209 Steps: 46870, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000042, Sample Num: 672, Cur Loss: 0.04665982, Cur Avg Loss: 0.20043815, Log Avg loss: 0.21747699, Global Avg Loss: 1.00503836, Time: 0.0209 Steps: 46880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000052, Sample Num: 832, Cur Loss: 0.08965719, Cur Avg Loss: 0.20415173, Log Avg loss: 0.21974879, Global Avg Loss: 1.00487088, Time: 0.0209 Steps: 46890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000062, Sample Num: 992, Cur Loss: 0.34841785, Cur Avg Loss: 0.21165837, Log Avg loss: 0.25069289, Global Avg Loss: 1.00471008, Time: 0.0209 Steps: 46900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000072, Sample Num: 1152, Cur Loss: 0.04655503, Cur Avg Loss: 0.21230086, Log Avg loss: 0.21628431, Global Avg Loss: 1.00454201, Time: 0.0209 Steps: 46910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000082, Sample Num: 1312, Cur Loss: 0.10271621, Cur Avg Loss: 0.20384350, Log Avg loss: 0.14295052, Global Avg Loss: 1.00435838, Time: 0.0209 Steps: 46920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000092, Sample Num: 1472, Cur Loss: 0.26802045, Cur Avg Loss: 0.20612368, Log Avg loss: 0.22482109, Global Avg Loss: 1.00419227, Time: 0.0209 Steps: 46930, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000102, Sample Num: 1632, Cur Loss: 0.47864234, Cur Avg Loss: 0.20558707, Log Avg loss: 0.20065032, Global Avg Loss: 1.00402109, Time: 0.0209 Steps: 46940, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000112, Sample Num: 1792, Cur Loss: 0.13008170, Cur Avg Loss: 0.20524340, Log Avg loss: 0.20173799, Global Avg Loss: 1.00385021, Time: 0.0209 Steps: 46950, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000122, Sample Num: 1952, Cur Loss: 0.07638771, Cur Avg Loss: 0.20374830, Log Avg loss: 0.18700317, Global Avg Loss: 1.00367626, Time: 0.0209 Steps: 46960, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000132, Sample Num: 2112, Cur Loss: 0.19975147, Cur Avg Loss: 0.19954287, Log Avg loss: 0.14823661, Global Avg Loss: 1.00349414, Time: 0.0209 Steps: 46970, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000142, Sample Num: 2272, Cur Loss: 0.16816750, Cur Avg Loss: 0.19779186, Log Avg loss: 0.17467845, Global Avg Loss: 1.00331772, Time: 0.0209 Steps: 46980, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000152, Sample Num: 2432, Cur Loss: 0.12087669, Cur Avg Loss: 0.20032892, Log Avg loss: 0.23635516, Global Avg Loss: 1.00315450, Time: 0.0209 Steps: 46990, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000162, Sample Num: 2592, Cur Loss: 0.10787638, Cur Avg Loss: 0.19722105, Log Avg loss: 0.14998148, Global Avg Loss: 1.00297297, Time: 0.0208 Steps: 47000, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000172, Sample Num: 2752, Cur Loss: 0.26740941, Cur Avg Loss: 0.19434819, Log Avg loss: 0.14780795, Global Avg Loss: 1.00279106, Time: 0.0208 Steps: 47010, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000182, Sample Num: 2912, Cur Loss: 0.08685161, Cur Avg Loss: 0.19412524, Log Avg loss: 0.19029043, Global Avg Loss: 1.00261826, Time: 0.0209 Steps: 47020, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000192, Sample Num: 3072, Cur Loss: 0.52581239, Cur Avg Loss: 0.19515254, Log Avg loss: 0.21384944, Global Avg Loss: 1.00245055, Time: 0.0209 Steps: 47030, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000202, Sample Num: 3232, Cur Loss: 0.08017442, Cur Avg Loss: 0.19536396, Log Avg loss: 0.19942310, Global Avg Loss: 1.00227983, Time: 0.0209 Steps: 47040, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000212, Sample Num: 3392, Cur Loss: 0.29940146, Cur Avg Loss: 0.19571358, Log Avg loss: 0.20277592, Global Avg Loss: 1.00210991, Time: 0.0209 Steps: 47050, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000222, Sample Num: 3552, Cur Loss: 0.19747581, Cur Avg Loss: 0.19613077, Log Avg loss: 0.20497532, Global Avg Loss: 1.00194052, Time: 0.0209 Steps: 47060, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000232, Sample Num: 3712, Cur Loss: 0.24868515, Cur Avg Loss: 0.19499977, Log Avg loss: 0.16989144, Global Avg Loss: 1.00176375, Time: 0.0209 Steps: 47070, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000242, Sample Num: 3872, Cur Loss: 0.19269279, Cur Avg Loss: 0.19478453, Log Avg loss: 0.18979094, Global Avg Loss: 1.00159129, Time: 0.0209 Steps: 47080, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000252, Sample Num: 4032, Cur Loss: 0.33347082, Cur Avg Loss: 0.19627596, Log Avg loss: 0.23236862, Global Avg Loss: 1.00142793, Time: 0.0209 Steps: 47090, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000262, Sample Num: 4192, Cur Loss: 0.03659990, Cur Avg Loss: 0.19547354, Log Avg loss: 0.17525270, Global Avg Loss: 1.00125253, Time: 0.0209 Steps: 47100, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000272, Sample Num: 4352, Cur Loss: 0.17482266, Cur Avg Loss: 0.19575323, Log Avg loss: 0.20308097, Global Avg Loss: 1.00108310, Time: 0.0208 Steps: 47110, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000282, Sample Num: 4512, Cur Loss: 0.11210527, Cur Avg Loss: 0.19711350, Log Avg loss: 0.23411285, Global Avg Loss: 1.00092033, Time: 0.0208 Steps: 47120, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000292, Sample Num: 4672, Cur Loss: 0.14886788, Cur Avg Loss: 0.19907258, Log Avg loss: 0.25431872, Global Avg Loss: 1.00076192, Time: 0.0209 Steps: 47130, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000302, Sample Num: 4832, Cur Loss: 0.17606305, Cur Avg Loss: 0.19893532, Log Avg loss: 0.19492735, Global Avg Loss: 1.00059097, Time: 0.0208 Steps: 47140, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000312, Sample Num: 4992, Cur Loss: 0.16352698, Cur Avg Loss: 0.20068540, Log Avg loss: 0.25353784, Global Avg Loss: 1.00043253, Time: 0.0208 Steps: 47150, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000322, Sample Num: 5152, Cur Loss: 0.28203136, Cur Avg Loss: 0.20270550, Log Avg loss: 0.26573247, Global Avg Loss: 1.00027674, Time: 0.0207 Steps: 47160, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000332, Sample Num: 5312, Cur Loss: 0.29530174, Cur Avg Loss: 0.20314570, Log Avg loss: 0.21732025, Global Avg Loss: 1.00011075, Time: 0.0209 Steps: 47170, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000342, Sample Num: 5472, Cur Loss: 0.07292348, Cur Avg Loss: 0.20280836, Log Avg loss: 0.19160870, Global Avg Loss: 0.99993939, Time: 0.0207 Steps: 47180, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000352, Sample Num: 5632, Cur Loss: 0.24216042, Cur Avg Loss: 0.20189587, Log Avg loss: 0.17068874, Global Avg Loss: 0.99976366, Time: 0.0208 Steps: 47190, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000362, Sample Num: 5792, Cur Loss: 0.36078829, Cur Avg Loss: 0.20262378, Log Avg loss: 0.22824623, Global Avg Loss: 0.99960021, Time: 0.0209 Steps: 47200, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000372, Sample Num: 5952, Cur Loss: 0.19417030, Cur Avg Loss: 0.20263720, Log Avg loss: 0.20312290, Global Avg Loss: 0.99943150, Time: 0.0208 Steps: 47210, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000382, Sample Num: 6112, Cur Loss: 0.10900219, Cur Avg Loss: 0.20216196, Log Avg loss: 0.18448294, Global Avg Loss: 0.99925891, Time: 0.0207 Steps: 47220, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000392, Sample Num: 6272, Cur Loss: 0.22180921, Cur Avg Loss: 0.20360331, Log Avg loss: 0.25866279, Global Avg Loss: 0.99910210, Time: 0.0208 Steps: 47230, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000402, Sample Num: 6432, Cur Loss: 0.39666894, Cur Avg Loss: 0.20767618, Log Avg loss: 0.36733285, Global Avg Loss: 0.99896837, Time: 0.0209 Steps: 47240, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000412, Sample Num: 6592, Cur Loss: 0.20508155, Cur Avg Loss: 0.20879605, Log Avg loss: 0.25381474, Global Avg Loss: 0.99881066, Time: 0.0209 Steps: 47250, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000422, Sample Num: 6752, Cur Loss: 0.13267963, Cur Avg Loss: 0.20871657, Log Avg loss: 0.20544226, Global Avg Loss: 0.99864279, Time: 0.0208 Steps: 47260, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000432, Sample Num: 6912, Cur Loss: 0.31822616, Cur Avg Loss: 0.20817803, Log Avg loss: 0.18545164, Global Avg Loss: 0.99847076, Time: 0.0208 Steps: 47270, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000442, Sample Num: 7072, Cur Loss: 0.19543916, Cur Avg Loss: 0.20911024, Log Avg loss: 0.24938172, Global Avg Loss: 0.99831232, Time: 0.0208 Steps: 47280, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000452, Sample Num: 7232, Cur Loss: 0.10442168, Cur Avg Loss: 0.21027413, Log Avg loss: 0.26171763, Global Avg Loss: 0.99815656, Time: 0.0209 Steps: 47290, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000462, Sample Num: 7392, Cur Loss: 0.21196565, Cur Avg Loss: 0.21118010, Log Avg loss: 0.25213017, Global Avg Loss: 0.99799884, Time: 0.0209 Steps: 47300, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000472, Sample Num: 7552, Cur Loss: 0.06754713, Cur Avg Loss: 0.21057583, Log Avg loss: 0.18265851, Global Avg Loss: 0.99782650, Time: 0.0209 Steps: 47310, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000482, Sample Num: 7712, Cur Loss: 0.08129746, Cur Avg Loss: 0.21026989, Log Avg loss: 0.19582953, Global Avg Loss: 0.99765702, Time: 0.0209 Steps: 47320, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000492, Sample Num: 7872, Cur Loss: 0.24903253, Cur Avg Loss: 0.21002839, Log Avg loss: 0.19838833, Global Avg Loss: 0.99748814, Time: 0.0209 Steps: 47330, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000502, Sample Num: 8032, Cur Loss: 0.08473718, Cur Avg Loss: 0.20927386, Log Avg loss: 0.17215074, Global Avg Loss: 0.99731380, Time: 0.0208 Steps: 47340, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000512, Sample Num: 8192, Cur Loss: 0.07028775, Cur Avg Loss: 0.20826240, Log Avg loss: 0.15748693, Global Avg Loss: 0.99713644, Time: 0.0254 Steps: 47350, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000522, Sample Num: 8352, Cur Loss: 0.22597945, Cur Avg Loss: 0.20739665, Log Avg loss: 0.16307057, Global Avg Loss: 0.99696032, Time: 0.0207 Steps: 47360, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000532, Sample Num: 8512, Cur Loss: 0.15814430, Cur Avg Loss: 0.20695023, Log Avg loss: 0.18364673, Global Avg Loss: 0.99678863, Time: 0.0208 Steps: 47370, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000542, Sample Num: 8672, Cur Loss: 0.12342270, Cur Avg Loss: 0.20703184, Log Avg loss: 0.21137400, Global Avg Loss: 0.99662286, Time: 0.0208 Steps: 47380, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000552, Sample Num: 8832, Cur Loss: 0.26486129, Cur Avg Loss: 0.20687749, Log Avg loss: 0.19851153, Global Avg Loss: 0.99645445, Time: 0.0208 Steps: 47390, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000562, Sample Num: 8992, Cur Loss: 0.10335694, Cur Avg Loss: 0.21241696, Log Avg loss: 0.51819565, Global Avg Loss: 0.99635355, Time: 0.0208 Steps: 47400, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000572, Sample Num: 9152, Cur Loss: 0.20052746, Cur Avg Loss: 0.21322612, Log Avg loss: 0.25870094, Global Avg Loss: 0.99619796, Time: 0.0208 Steps: 47410, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000582, Sample Num: 9312, Cur Loss: 0.06392610, Cur Avg Loss: 0.21396570, Log Avg loss: 0.25626966, Global Avg Loss: 0.99604192, Time: 0.0208 Steps: 47420, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000592, Sample Num: 9472, Cur Loss: 0.17127474, Cur Avg Loss: 0.21559617, Log Avg loss: 0.31048968, Global Avg Loss: 0.99589738, Time: 0.0208 Steps: 47430, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000602, Sample Num: 9632, Cur Loss: 0.10399152, Cur Avg Loss: 0.21552737, Log Avg loss: 0.21145443, Global Avg Loss: 0.99573203, Time: 0.0208 Steps: 47440, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000612, Sample Num: 9792, Cur Loss: 0.17381190, Cur Avg Loss: 0.21485441, Log Avg loss: 0.17434216, Global Avg Loss: 0.99555892, Time: 0.0208 Steps: 47450, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000622, Sample Num: 9952, Cur Loss: 0.27394587, Cur Avg Loss: 0.21473962, Log Avg loss: 0.20771408, Global Avg Loss: 0.99539292, Time: 0.0208 Steps: 47460, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000632, Sample Num: 10112, Cur Loss: 0.12623236, Cur Avg Loss: 0.21400175, Log Avg loss: 0.16810655, Global Avg Loss: 0.99521864, Time: 0.0208 Steps: 47470, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000642, Sample Num: 10272, Cur Loss: 0.16713825, Cur Avg Loss: 0.21365977, Log Avg loss: 0.19204678, Global Avg Loss: 0.99504948, Time: 0.0208 Steps: 47480, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000652, Sample Num: 10432, Cur Loss: 0.26518810, Cur Avg Loss: 0.21319414, Log Avg loss: 0.18330020, Global Avg Loss: 0.99487855, Time: 0.0208 Steps: 47490, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000662, Sample Num: 10592, Cur Loss: 0.44514984, Cur Avg Loss: 0.21297166, Log Avg loss: 0.19846653, Global Avg Loss: 0.99471089, Time: 0.0208 Steps: 47500, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000672, Sample Num: 10752, Cur Loss: 0.11645252, Cur Avg Loss: 0.21302569, Log Avg loss: 0.21660190, Global Avg Loss: 0.99454711, Time: 0.0208 Steps: 47510, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000682, Sample Num: 10912, Cur Loss: 0.13974355, Cur Avg Loss: 0.21277851, Log Avg loss: 0.19616824, Global Avg Loss: 0.99437910, Time: 0.0208 Steps: 47520, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000692, Sample Num: 11072, Cur Loss: 0.19775419, Cur Avg Loss: 0.21308057, Log Avg loss: 0.23368124, Global Avg Loss: 0.99421905, Time: 0.0208 Steps: 47530, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000702, Sample Num: 11232, Cur Loss: 0.07282893, Cur Avg Loss: 0.21206783, Log Avg loss: 0.14198638, Global Avg Loss: 0.99403979, Time: 0.0207 Steps: 47540, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000712, Sample Num: 11392, Cur Loss: 0.15351567, Cur Avg Loss: 0.21219183, Log Avg loss: 0.22089627, Global Avg Loss: 0.99387719, Time: 0.0207 Steps: 47550, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000722, Sample Num: 11552, Cur Loss: 0.27090210, Cur Avg Loss: 0.21188734, Log Avg loss: 0.19020800, Global Avg Loss: 0.99370821, Time: 0.0208 Steps: 47560, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000732, Sample Num: 11712, Cur Loss: 0.56226182, Cur Avg Loss: 0.21399670, Log Avg loss: 0.36629219, Global Avg Loss: 0.99357632, Time: 0.0208 Steps: 47570, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000742, Sample Num: 11872, Cur Loss: 0.30999029, Cur Avg Loss: 0.21509561, Log Avg loss: 0.29553622, Global Avg Loss: 0.99342961, Time: 0.0208 Steps: 47580, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000752, Sample Num: 12032, Cur Loss: 0.05531491, Cur Avg Loss: 0.21454364, Log Avg loss: 0.17358688, Global Avg Loss: 0.99325734, Time: 0.0208 Steps: 47590, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000762, Sample Num: 12192, Cur Loss: 0.20949051, Cur Avg Loss: 0.21379882, Log Avg loss: 0.15778877, Global Avg Loss: 0.99308182, Time: 0.0208 Steps: 47600, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000772, Sample Num: 12352, Cur Loss: 0.20084487, Cur Avg Loss: 0.21319994, Log Avg loss: 0.16756544, Global Avg Loss: 0.99290843, Time: 0.0245 Steps: 47610, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000782, Sample Num: 12512, Cur Loss: 0.30563533, Cur Avg Loss: 0.21461484, Log Avg loss: 0.32384469, Global Avg Loss: 0.99276793, Time: 0.0207 Steps: 47620, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000792, Sample Num: 12672, Cur Loss: 0.24605629, Cur Avg Loss: 0.21531121, Log Avg loss: 0.26976730, Global Avg Loss: 0.99261613, Time: 0.0208 Steps: 47630, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000802, Sample Num: 12832, Cur Loss: 0.05134159, Cur Avg Loss: 0.21597251, Log Avg loss: 0.26834773, Global Avg Loss: 0.99246410, Time: 0.0208 Steps: 47640, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000812, Sample Num: 12992, Cur Loss: 0.26130193, Cur Avg Loss: 0.21614986, Log Avg loss: 0.23037304, Global Avg Loss: 0.99230417, Time: 0.0208 Steps: 47650, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000822, Sample Num: 13152, Cur Loss: 0.23408338, Cur Avg Loss: 0.21573476, Log Avg loss: 0.18202895, Global Avg Loss: 0.99213416, Time: 0.0207 Steps: 47660, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000832, Sample Num: 13312, Cur Loss: 0.24453381, Cur Avg Loss: 0.21554875, Log Avg loss: 0.20025846, Global Avg Loss: 0.99196804, Time: 0.0207 Steps: 47670, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000842, Sample Num: 13472, Cur Loss: 0.19083869, Cur Avg Loss: 0.21545462, Log Avg loss: 0.20762328, Global Avg Loss: 0.99180354, Time: 0.0208 Steps: 47680, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000852, Sample Num: 13632, Cur Loss: 0.07668267, Cur Avg Loss: 0.21501726, Log Avg loss: 0.17819149, Global Avg Loss: 0.99163293, Time: 0.0207 Steps: 47690, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000862, Sample Num: 13792, Cur Loss: 0.59022135, Cur Avg Loss: 0.21521997, Log Avg loss: 0.23249046, Global Avg Loss: 0.99147378, Time: 0.0208 Steps: 47700, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000872, Sample Num: 13952, Cur Loss: 0.26513779, Cur Avg Loss: 0.21458074, Log Avg loss: 0.15947969, Global Avg Loss: 0.99129940, Time: 0.0207 Steps: 47710, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000882, Sample Num: 14112, Cur Loss: 0.22355993, Cur Avg Loss: 0.21453921, Log Avg loss: 0.21091713, Global Avg Loss: 0.99113586, Time: 0.0207 Steps: 47720, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000892, Sample Num: 14272, Cur Loss: 0.32289553, Cur Avg Loss: 0.21439370, Log Avg loss: 0.20156000, Global Avg Loss: 0.99097044, Time: 0.0209 Steps: 47730, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000902, Sample Num: 14432, Cur Loss: 0.12952653, Cur Avg Loss: 0.21387418, Log Avg loss: 0.16753341, Global Avg Loss: 0.99079796, Time: 0.0207 Steps: 47740, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000912, Sample Num: 14592, Cur Loss: 0.12892810, Cur Avg Loss: 0.21314140, Log Avg loss: 0.14704433, Global Avg Loss: 0.99062125, Time: 0.0207 Steps: 47750, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000922, Sample Num: 14752, Cur Loss: 0.09953423, Cur Avg Loss: 0.21239646, Log Avg loss: 0.14445798, Global Avg Loss: 0.99044408, Time: 0.0208 Steps: 47760, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000932, Sample Num: 14912, Cur Loss: 0.61143416, Cur Avg Loss: 0.21315528, Log Avg loss: 0.28311857, Global Avg Loss: 0.99029601, Time: 0.0207 Steps: 47770, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000942, Sample Num: 15072, Cur Loss: 0.22693476, Cur Avg Loss: 0.21290231, Log Avg loss: 0.18932525, Global Avg Loss: 0.99012838, Time: 0.0207 Steps: 47780, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000952, Sample Num: 15232, Cur Loss: 0.19861555, Cur Avg Loss: 0.21258758, Log Avg loss: 0.18293975, Global Avg Loss: 0.98995947, Time: 0.0207 Steps: 47790, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000962, Sample Num: 15392, Cur Loss: 0.31931818, Cur Avg Loss: 0.21234100, Log Avg loss: 0.18886730, Global Avg Loss: 0.98979188, Time: 0.0207 Steps: 47800, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000972, Sample Num: 15552, Cur Loss: 0.12632078, Cur Avg Loss: 0.21181782, Log Avg loss: 0.16148725, Global Avg Loss: 0.98961863, Time: 0.0207 Steps: 47810, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000982, Sample Num: 15712, Cur Loss: 0.09708026, Cur Avg Loss: 0.21131409, Log Avg loss: 0.16235136, Global Avg Loss: 0.98944564, Time: 0.0207 Steps: 47820, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000992, Sample Num: 15872, Cur Loss: 0.44823617, Cur Avg Loss: 0.21178777, Log Avg loss: 0.25830397, Global Avg Loss: 0.98929277, Time: 0.0207 Steps: 47830, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001002, Sample Num: 16032, Cur Loss: 0.04037421, Cur Avg Loss: 0.21235718, Log Avg loss: 0.26884256, Global Avg Loss: 0.98914218, Time: 0.0207 Steps: 47840, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001012, Sample Num: 16192, Cur Loss: 0.32411194, Cur Avg Loss: 0.21229513, Log Avg loss: 0.20607717, Global Avg Loss: 0.98897853, Time: 0.0207 Steps: 47850, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001022, Sample Num: 16352, Cur Loss: 0.12092899, Cur Avg Loss: 0.21318564, Log Avg loss: 0.30330509, Global Avg Loss: 0.98883526, Time: 0.0207 Steps: 47860, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001032, Sample Num: 16512, Cur Loss: 0.32413337, Cur Avg Loss: 0.21341242, Log Avg loss: 0.23658954, Global Avg Loss: 0.98867812, Time: 0.0208 Steps: 47870, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001042, Sample Num: 16672, Cur Loss: 0.15099230, Cur Avg Loss: 0.21302809, Log Avg loss: 0.17336509, Global Avg Loss: 0.98850784, Time: 0.0207 Steps: 47880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001052, Sample Num: 16832, Cur Loss: 0.05141073, Cur Avg Loss: 0.21233772, Log Avg loss: 0.14040192, Global Avg Loss: 0.98833074, Time: 0.0207 Steps: 47890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001062, Sample Num: 16992, Cur Loss: 0.12584463, Cur Avg Loss: 0.21208994, Log Avg loss: 0.18602341, Global Avg Loss: 0.98816324, Time: 0.0207 Steps: 47900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001072, Sample Num: 17152, Cur Loss: 0.31040958, Cur Avg Loss: 0.21243913, Log Avg loss: 0.24952241, Global Avg Loss: 0.98800907, Time: 0.0207 Steps: 47910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001082, Sample Num: 17312, Cur Loss: 0.26704997, Cur Avg Loss: 0.21172205, Log Avg loss: 0.13485176, Global Avg Loss: 0.98783103, Time: 0.0207 Steps: 47920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001092, Sample Num: 17472, Cur Loss: 0.12371734, Cur Avg Loss: 0.21173221, Log Avg loss: 0.21283161, Global Avg Loss: 0.98766934, Time: 0.0208 Steps: 47930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001102, Sample Num: 17632, Cur Loss: 0.09721365, Cur Avg Loss: 0.21229895, Log Avg loss: 0.27418703, Global Avg Loss: 0.98752051, Time: 0.0207 Steps: 47940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001112, Sample Num: 17792, Cur Loss: 0.67961931, Cur Avg Loss: 0.21279293, Log Avg loss: 0.26722896, Global Avg Loss: 0.98737029, Time: 0.0208 Steps: 47950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001122, Sample Num: 17952, Cur Loss: 0.12817310, Cur Avg Loss: 0.21409428, Log Avg loss: 0.35880446, Global Avg Loss: 0.98723923, Time: 0.0207 Steps: 47960, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001132, Sample Num: 18112, Cur Loss: 0.09185304, Cur Avg Loss: 0.21364608, Log Avg loss: 0.16335855, Global Avg Loss: 0.98706748, Time: 0.0207 Steps: 47970, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001142, Sample Num: 18272, Cur Loss: 0.23161049, Cur Avg Loss: 0.21342574, Log Avg loss: 0.18848275, Global Avg Loss: 0.98690104, Time: 0.0208 Steps: 47980, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001152, Sample Num: 18432, Cur Loss: 0.24027808, Cur Avg Loss: 0.21378677, Log Avg loss: 0.25501607, Global Avg Loss: 0.98674854, Time: 0.0207 Steps: 47990, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001162, Sample Num: 18592, Cur Loss: 0.27043703, Cur Avg Loss: 0.21373360, Log Avg loss: 0.20760878, Global Avg Loss: 0.98658622, Time: 0.0207 Steps: 48000, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001172, Sample Num: 18752, Cur Loss: 0.18864141, Cur Avg Loss: 0.21331791, Log Avg loss: 0.16501506, Global Avg Loss: 0.98641509, Time: 0.0207 Steps: 48010, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001182, Sample Num: 18912, Cur Loss: 0.08943160, Cur Avg Loss: 0.21343860, Log Avg loss: 0.22758366, Global Avg Loss: 0.98625707, Time: 0.0208 Steps: 48020, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001192, Sample Num: 19072, Cur Loss: 0.28988761, Cur Avg Loss: 0.21328099, Log Avg loss: 0.19465144, Global Avg Loss: 0.98609225, Time: 0.0207 Steps: 48030, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001202, Sample Num: 19232, Cur Loss: 0.16529830, Cur Avg Loss: 0.21366937, Log Avg loss: 0.25996377, Global Avg Loss: 0.98594110, Time: 0.0208 Steps: 48040, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001212, Sample Num: 19392, Cur Loss: 0.76033330, Cur Avg Loss: 0.21480868, Log Avg loss: 0.35175396, Global Avg Loss: 0.98580912, Time: 0.0207 Steps: 48050, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001222, Sample Num: 19552, Cur Loss: 0.42940837, Cur Avg Loss: 0.21537872, Log Avg loss: 0.28446693, Global Avg Loss: 0.98566319, Time: 0.0208 Steps: 48060, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001232, Sample Num: 19712, Cur Loss: 0.44601774, Cur Avg Loss: 0.21686753, Log Avg loss: 0.39880056, Global Avg Loss: 0.98554110, Time: 0.0207 Steps: 48070, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001242, Sample Num: 19872, Cur Loss: 0.20656413, Cur Avg Loss: 0.21735442, Log Avg loss: 0.27733976, Global Avg Loss: 0.98539380, Time: 0.0207 Steps: 48080, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001252, Sample Num: 20032, Cur Loss: 0.46814311, Cur Avg Loss: 0.21820126, Log Avg loss: 0.32337804, Global Avg Loss: 0.98525614, Time: 0.0208 Steps: 48090, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001262, Sample Num: 20192, Cur Loss: 0.16727312, Cur Avg Loss: 0.21899356, Log Avg loss: 0.31818966, Global Avg Loss: 0.98511746, Time: 0.0208 Steps: 48100, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001272, Sample Num: 20352, Cur Loss: 0.08109215, Cur Avg Loss: 0.21828103, Log Avg loss: 0.12836013, Global Avg Loss: 0.98493938, Time: 0.0207 Steps: 48110, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001282, Sample Num: 20512, Cur Loss: 0.11957692, Cur Avg Loss: 0.21800311, Log Avg loss: 0.18265208, Global Avg Loss: 0.98477265, Time: 0.0246 Steps: 48120, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001292, Sample Num: 20672, Cur Loss: 0.14410155, Cur Avg Loss: 0.21802611, Log Avg loss: 0.22097411, Global Avg Loss: 0.98461395, Time: 0.0208 Steps: 48130, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001302, Sample Num: 20832, Cur Loss: 0.21501437, Cur Avg Loss: 0.21769551, Log Avg loss: 0.17498191, Global Avg Loss: 0.98444577, Time: 0.0208 Steps: 48140, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001312, Sample Num: 20992, Cur Loss: 0.34121042, Cur Avg Loss: 0.21801065, Log Avg loss: 0.25904227, Global Avg Loss: 0.98429512, Time: 0.0208 Steps: 48150, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001322, Sample Num: 21152, Cur Loss: 0.11489335, Cur Avg Loss: 0.21815280, Log Avg loss: 0.23680304, Global Avg Loss: 0.98413991, Time: 0.0208 Steps: 48160, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001332, Sample Num: 21312, Cur Loss: 0.24208203, Cur Avg Loss: 0.21816423, Log Avg loss: 0.21967421, Global Avg Loss: 0.98398120, Time: 0.0208 Steps: 48170, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001342, Sample Num: 21472, Cur Loss: 0.39812127, Cur Avg Loss: 0.21814090, Log Avg loss: 0.21503341, Global Avg Loss: 0.98382161, Time: 0.0208 Steps: 48180, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001352, Sample Num: 21632, Cur Loss: 0.26610735, Cur Avg Loss: 0.21784551, Log Avg loss: 0.17820401, Global Avg Loss: 0.98365443, Time: 0.0208 Steps: 48190, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001362, Sample Num: 21792, Cur Loss: 0.12880571, Cur Avg Loss: 0.21750876, Log Avg loss: 0.17198066, Global Avg Loss: 0.98348603, Time: 0.0208 Steps: 48200, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001372, Sample Num: 21952, Cur Loss: 0.15156263, Cur Avg Loss: 0.21671510, Log Avg loss: 0.10861922, Global Avg Loss: 0.98330456, Time: 0.0208 Steps: 48210, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001382, Sample Num: 22112, Cur Loss: 0.11053833, Cur Avg Loss: 0.21667336, Log Avg loss: 0.21094622, Global Avg Loss: 0.98314439, Time: 0.0208 Steps: 48220, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001392, Sample Num: 22272, Cur Loss: 0.16438293, Cur Avg Loss: 0.21633195, Log Avg loss: 0.16914886, Global Avg Loss: 0.98297562, Time: 0.0208 Steps: 48230, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001402, Sample Num: 22432, Cur Loss: 0.27381596, Cur Avg Loss: 0.21604465, Log Avg loss: 0.17605253, Global Avg Loss: 0.98280834, Time: 0.0208 Steps: 48240, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001412, Sample Num: 22592, Cur Loss: 0.22953403, Cur Avg Loss: 0.21597755, Log Avg loss: 0.20657064, Global Avg Loss: 0.98264747, Time: 0.0208 Steps: 48250, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001422, Sample Num: 22752, Cur Loss: 0.22099060, Cur Avg Loss: 0.21603721, Log Avg loss: 0.22446071, Global Avg Loss: 0.98249036, Time: 0.0208 Steps: 48260, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001432, Sample Num: 22912, Cur Loss: 0.13440369, Cur Avg Loss: 0.21595451, Log Avg loss: 0.20419466, Global Avg Loss: 0.98232912, Time: 0.0208 Steps: 48270, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001442, Sample Num: 23072, Cur Loss: 0.31597272, Cur Avg Loss: 0.21651607, Log Avg loss: 0.29693191, Global Avg Loss: 0.98218716, Time: 0.0208 Steps: 48280, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001452, Sample Num: 23232, Cur Loss: 0.16737670, Cur Avg Loss: 0.21703952, Log Avg loss: 0.29252056, Global Avg Loss: 0.98204434, Time: 0.0208 Steps: 48290, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001462, Sample Num: 23392, Cur Loss: 0.24484201, Cur Avg Loss: 0.21763287, Log Avg loss: 0.30378675, Global Avg Loss: 0.98190392, Time: 0.0208 Steps: 48300, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001472, Sample Num: 23552, Cur Loss: 0.07270193, Cur Avg Loss: 0.21802346, Log Avg loss: 0.27512826, Global Avg Loss: 0.98175762, Time: 0.0208 Steps: 48310, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001482, Sample Num: 23712, Cur Loss: 0.75803059, Cur Avg Loss: 0.21881723, Log Avg loss: 0.33565964, Global Avg Loss: 0.98162390, Time: 0.0208 Steps: 48320, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001492, Sample Num: 23872, Cur Loss: 0.12962426, Cur Avg Loss: 0.21844772, Log Avg loss: 0.16368753, Global Avg Loss: 0.98145466, Time: 0.0208 Steps: 48330, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001502, Sample Num: 24032, Cur Loss: 0.08216074, Cur Avg Loss: 0.21882661, Log Avg loss: 0.27535642, Global Avg Loss: 0.98130859, Time: 0.0208 Steps: 48340, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001512, Sample Num: 24192, Cur Loss: 0.18661886, Cur Avg Loss: 0.21938886, Log Avg loss: 0.30383802, Global Avg Loss: 0.98116848, Time: 0.0208 Steps: 48350, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001522, Sample Num: 24352, Cur Loss: 0.03497197, Cur Avg Loss: 0.21916000, Log Avg loss: 0.18455725, Global Avg Loss: 0.98100375, Time: 0.0207 Steps: 48360, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001532, Sample Num: 24512, Cur Loss: 0.29723680, Cur Avg Loss: 0.21873889, Log Avg loss: 0.15464575, Global Avg Loss: 0.98083291, Time: 0.0208 Steps: 48370, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001542, Sample Num: 24672, Cur Loss: 0.14782292, Cur Avg Loss: 0.21886939, Log Avg loss: 0.23886273, Global Avg Loss: 0.98067955, Time: 0.0209 Steps: 48380, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001552, Sample Num: 24832, Cur Loss: 0.20657797, Cur Avg Loss: 0.21874016, Log Avg loss: 0.19881201, Global Avg Loss: 0.98051797, Time: 0.0208 Steps: 48390, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001562, Sample Num: 24992, Cur Loss: 0.39633194, Cur Avg Loss: 0.21875940, Log Avg loss: 0.22174564, Global Avg Loss: 0.98036120, Time: 0.0207 Steps: 48400, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001572, Sample Num: 25152, Cur Loss: 0.05965316, Cur Avg Loss: 0.21852367, Log Avg loss: 0.18170237, Global Avg Loss: 0.98019622, Time: 0.0208 Steps: 48410, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001582, Sample Num: 25312, Cur Loss: 0.15981922, Cur Avg Loss: 0.21796153, Log Avg loss: 0.12959316, Global Avg Loss: 0.98002055, Time: 0.0208 Steps: 48420, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001592, Sample Num: 25472, Cur Loss: 0.38546699, Cur Avg Loss: 0.21827575, Log Avg loss: 0.26798480, Global Avg Loss: 0.97987353, Time: 0.0208 Steps: 48430, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001602, Sample Num: 25632, Cur Loss: 0.57080030, Cur Avg Loss: 0.21882356, Log Avg loss: 0.30603513, Global Avg Loss: 0.97973442, Time: 0.0208 Steps: 48440, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001612, Sample Num: 25792, Cur Loss: 0.15430161, Cur Avg Loss: 0.21889832, Log Avg loss: 0.23087499, Global Avg Loss: 0.97957985, Time: 0.0208 Steps: 48450, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001622, Sample Num: 25952, Cur Loss: 0.06500536, Cur Avg Loss: 0.21842483, Log Avg loss: 0.14209808, Global Avg Loss: 0.97940704, Time: 0.0208 Steps: 48460, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001632, Sample Num: 26112, Cur Loss: 0.47405159, Cur Avg Loss: 0.21909303, Log Avg loss: 0.32747629, Global Avg Loss: 0.97927253, Time: 0.0208 Steps: 48470, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001642, Sample Num: 26272, Cur Loss: 0.10739108, Cur Avg Loss: 0.21999610, Log Avg loss: 0.36737590, Global Avg Loss: 0.97914632, Time: 0.0208 Steps: 48480, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001652, Sample Num: 26432, Cur Loss: 0.48711777, Cur Avg Loss: 0.22044997, Log Avg loss: 0.29497648, Global Avg Loss: 0.97900522, Time: 0.0208 Steps: 48490, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001662, Sample Num: 26592, Cur Loss: 0.08140779, Cur Avg Loss: 0.22111222, Log Avg loss: 0.33051491, Global Avg Loss: 0.97887151, Time: 0.0208 Steps: 48500, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001672, Sample Num: 26752, Cur Loss: 0.26202428, Cur Avg Loss: 0.22061509, Log Avg loss: 0.13799267, Global Avg Loss: 0.97869817, Time: 0.0208 Steps: 48510, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001682, Sample Num: 26912, Cur Loss: 0.07819803, Cur Avg Loss: 0.22016163, Log Avg loss: 0.14434333, Global Avg Loss: 0.97852621, Time: 0.0208 Steps: 48520, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001692, Sample Num: 27072, Cur Loss: 0.13092542, Cur Avg Loss: 0.21996286, Log Avg loss: 0.18652945, Global Avg Loss: 0.97836301, Time: 0.0208 Steps: 48530, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001702, Sample Num: 27232, Cur Loss: 0.08731596, Cur Avg Loss: 0.21944245, Log Avg loss: 0.13138866, Global Avg Loss: 0.97818852, Time: 0.0208 Steps: 48540, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001712, Sample Num: 27392, Cur Loss: 0.14717507, Cur Avg Loss: 0.21935985, Log Avg loss: 0.20530135, Global Avg Loss: 0.97802933, Time: 0.0207 Steps: 48550, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001722, Sample Num: 27552, Cur Loss: 0.15972076, Cur Avg Loss: 0.21947714, Log Avg loss: 0.23955695, Global Avg Loss: 0.97787726, Time: 0.0208 Steps: 48560, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001732, Sample Num: 27712, Cur Loss: 0.28835970, Cur Avg Loss: 0.21941419, Log Avg loss: 0.20857490, Global Avg Loss: 0.97771886, Time: 0.0208 Steps: 48570, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001742, Sample Num: 27872, Cur Loss: 0.16145238, Cur Avg Loss: 0.21951692, Log Avg loss: 0.23731038, Global Avg Loss: 0.97756645, Time: 0.0208 Steps: 48580, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001752, Sample Num: 28032, Cur Loss: 0.07457489, Cur Avg Loss: 0.21910704, Log Avg loss: 0.14770498, Global Avg Loss: 0.97739567, Time: 0.0208 Steps: 48590, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001762, Sample Num: 28192, Cur Loss: 0.11037980, Cur Avg Loss: 0.21902240, Log Avg loss: 0.20419430, Global Avg Loss: 0.97723657, Time: 0.0208 Steps: 48600, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001772, Sample Num: 28352, Cur Loss: 0.19556603, Cur Avg Loss: 0.21889586, Log Avg loss: 0.19659915, Global Avg Loss: 0.97707598, Time: 0.0207 Steps: 48610, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001782, Sample Num: 28512, Cur Loss: 0.12786946, Cur Avg Loss: 0.21899989, Log Avg loss: 0.23743349, Global Avg Loss: 0.97692385, Time: 0.0208 Steps: 48620, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001792, Sample Num: 28672, Cur Loss: 0.10776777, Cur Avg Loss: 0.21883709, Log Avg loss: 0.18982638, Global Avg Loss: 0.97676200, Time: 0.0253 Steps: 48630, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001802, Sample Num: 28832, Cur Loss: 0.19414435, Cur Avg Loss: 0.21883748, Log Avg loss: 0.21890816, Global Avg Loss: 0.97660619, Time: 0.0220 Steps: 48640, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001812, Sample Num: 28992, Cur Loss: 0.15947312, Cur Avg Loss: 0.21935534, Log Avg loss: 0.31267340, Global Avg Loss: 0.97646972, Time: 0.0219 Steps: 48650, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001822, Sample Num: 29152, Cur Loss: 0.24268970, Cur Avg Loss: 0.21928679, Log Avg loss: 0.20686488, Global Avg Loss: 0.97631156, Time: 0.0219 Steps: 48660, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001832, Sample Num: 29312, Cur Loss: 0.10491064, Cur Avg Loss: 0.21919705, Log Avg loss: 0.20284741, Global Avg Loss: 0.97615264, Time: 0.0218 Steps: 48670, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001842, Sample Num: 29472, Cur Loss: 0.12616448, Cur Avg Loss: 0.21912954, Log Avg loss: 0.20676026, Global Avg Loss: 0.97599459, Time: 0.0219 Steps: 48680, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001852, Sample Num: 29632, Cur Loss: 0.15080628, Cur Avg Loss: 0.21905436, Log Avg loss: 0.20520666, Global Avg Loss: 0.97583628, Time: 0.0219 Steps: 48690, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001862, Sample Num: 29792, Cur Loss: 0.70946181, Cur Avg Loss: 0.21915684, Log Avg loss: 0.23813656, Global Avg Loss: 0.97568480, Time: 0.0219 Steps: 48700, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001872, Sample Num: 29952, Cur Loss: 0.09442360, Cur Avg Loss: 0.21930042, Log Avg loss: 0.24603490, Global Avg Loss: 0.97553501, Time: 0.0219 Steps: 48710, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001882, Sample Num: 30112, Cur Loss: 0.18028733, Cur Avg Loss: 0.21894612, Log Avg loss: 0.15262124, Global Avg Loss: 0.97536610, Time: 0.0219 Steps: 48720, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001892, Sample Num: 30272, Cur Loss: 0.18337157, Cur Avg Loss: 0.21894692, Log Avg loss: 0.21909803, Global Avg Loss: 0.97521091, Time: 0.0219 Steps: 48730, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001902, Sample Num: 30432, Cur Loss: 0.32099706, Cur Avg Loss: 0.21906828, Log Avg loss: 0.24202985, Global Avg Loss: 0.97506048, Time: 0.0219 Steps: 48740, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001912, Sample Num: 30592, Cur Loss: 0.44403431, Cur Avg Loss: 0.21897826, Log Avg loss: 0.20185615, Global Avg Loss: 0.97490187, Time: 0.0219 Steps: 48750, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001922, Sample Num: 30752, Cur Loss: 0.02285657, Cur Avg Loss: 0.21876064, Log Avg loss: 0.17715193, Global Avg Loss: 0.97473827, Time: 0.0219 Steps: 48760, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001932, Sample Num: 30912, Cur Loss: 0.12139932, Cur Avg Loss: 0.21856720, Log Avg loss: 0.18138709, Global Avg Loss: 0.97457559, Time: 0.0219 Steps: 48770, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001942, Sample Num: 31072, Cur Loss: 0.39829004, Cur Avg Loss: 0.21922174, Log Avg loss: 0.34567942, Global Avg Loss: 0.97444667, Time: 0.0219 Steps: 48780, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001952, Sample Num: 31232, Cur Loss: 0.13010487, Cur Avg Loss: 0.21902051, Log Avg loss: 0.17994134, Global Avg Loss: 0.97428383, Time: 0.0219 Steps: 48790, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001962, Sample Num: 31392, Cur Loss: 0.10224144, Cur Avg Loss: 0.21892997, Log Avg loss: 0.20125567, Global Avg Loss: 0.97412542, Time: 0.0219 Steps: 48800, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001972, Sample Num: 31552, Cur Loss: 0.05377159, Cur Avg Loss: 0.21958471, Log Avg loss: 0.34804465, Global Avg Loss: 0.97399715, Time: 0.0219 Steps: 48810, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001982, Sample Num: 31712, Cur Loss: 0.30501264, Cur Avg Loss: 0.21996942, Log Avg loss: 0.29583498, Global Avg Loss: 0.97385824, Time: 0.0219 Steps: 48820, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001992, Sample Num: 31872, Cur Loss: 0.49235153, Cur Avg Loss: 0.22010858, Log Avg loss: 0.24769072, Global Avg Loss: 0.97370953, Time: 0.0219 Steps: 48830, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002002, Sample Num: 32032, Cur Loss: 0.10176518, Cur Avg Loss: 0.22025185, Log Avg loss: 0.24878997, Global Avg Loss: 0.97356110, Time: 0.0219 Steps: 48840, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002012, Sample Num: 32192, Cur Loss: 0.23120144, Cur Avg Loss: 0.21992011, Log Avg loss: 0.15350608, Global Avg Loss: 0.97339323, Time: 0.0219 Steps: 48850, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002022, Sample Num: 32352, Cur Loss: 0.35107347, Cur Avg Loss: 0.22029156, Log Avg loss: 0.29502837, Global Avg Loss: 0.97325439, Time: 0.0219 Steps: 48860, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002032, Sample Num: 32512, Cur Loss: 0.07049151, Cur Avg Loss: 0.22024800, Log Avg loss: 0.21143877, Global Avg Loss: 0.97309850, Time: 0.0219 Steps: 48870, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002042, Sample Num: 32672, Cur Loss: 0.37774557, Cur Avg Loss: 0.22015948, Log Avg loss: 0.20217388, Global Avg Loss: 0.97294078, Time: 0.0219 Steps: 48880, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002052, Sample Num: 32832, Cur Loss: 0.29148585, Cur Avg Loss: 0.21996033, Log Avg loss: 0.17929256, Global Avg Loss: 0.97277845, Time: 0.0210 Steps: 48890, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002062, Sample Num: 32992, Cur Loss: 0.17403522, Cur Avg Loss: 0.22033299, Log Avg loss: 0.29680408, Global Avg Loss: 0.97264022, Time: 0.0208 Steps: 48900, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002072, Sample Num: 33152, Cur Loss: 0.17789148, Cur Avg Loss: 0.22018258, Log Avg loss: 0.18916804, Global Avg Loss: 0.97248003, Time: 0.0208 Steps: 48910, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002082, Sample Num: 33312, Cur Loss: 0.16113009, Cur Avg Loss: 0.21971323, Log Avg loss: 0.12246258, Global Avg Loss: 0.97230627, Time: 0.0208 Steps: 48920, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002092, Sample Num: 33472, Cur Loss: 0.44257486, Cur Avg Loss: 0.21958619, Log Avg loss: 0.19313759, Global Avg Loss: 0.97214703, Time: 0.0209 Steps: 48930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002102, Sample Num: 33632, Cur Loss: 0.07415114, Cur Avg Loss: 0.21942592, Log Avg loss: 0.18589645, Global Avg Loss: 0.97198637, Time: 0.0208 Steps: 48940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002112, Sample Num: 33792, Cur Loss: 0.59843439, Cur Avg Loss: 0.21949692, Log Avg loss: 0.23442130, Global Avg Loss: 0.97183570, Time: 0.0209 Steps: 48950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002122, Sample Num: 33952, Cur Loss: 0.07844189, Cur Avg Loss: 0.21918134, Log Avg loss: 0.15253116, Global Avg Loss: 0.97166836, Time: 0.0209 Steps: 48960, Updated lr: 0.000055 ***** Running evaluation checkpoint-48967 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-48967 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.582889, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.240694, "eval_total_loss": 169.207862, "eval_mae": 0.361685, "eval_mse": 0.240753, "eval_r2": 0.846961, "eval_sp_statistic": 0.911433, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.926049, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.201736, "test_total_loss": 101.271473, "test_mae": 0.326465, "test_mse": 0.201808, "test_r2": 0.869751, "test_sp_statistic": 0.906615, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.94118, "test_ps_pvalue": 0.0, "lr": 5.451209103840683e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9715512555455957, "train_cur_epoch_loss": 466.17044734209776, "train_cur_epoch_avg_loss": 0.21896216408741087, "train_cur_epoch_time": 44.58288931846619, "train_cur_epoch_avg_time": 0.020940765297541658, "epoch": 23, "step": 48967} ################################################## Training, Epoch: 0024, Batch: 000003, Sample Num: 48, Cur Loss: 0.24122909, Cur Avg Loss: 0.21528502, Log Avg loss: 0.17134971, Global Avg Loss: 0.97150493, Time: 0.0247 Steps: 48970, Updated lr: 0.000055 Training, Epoch: 0024, Batch: 000013, Sample Num: 208, Cur Loss: 0.11444682, Cur Avg Loss: 0.24078440, Log Avg loss: 0.24843421, Global Avg Loss: 0.97135730, Time: 0.0210 Steps: 48980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000023, Sample Num: 368, Cur Loss: 0.08958980, Cur Avg Loss: 0.20809863, Log Avg loss: 0.16560713, Global Avg Loss: 0.97119283, Time: 0.0209 Steps: 48990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000033, Sample Num: 528, Cur Loss: 0.11779199, Cur Avg Loss: 0.21500926, Log Avg loss: 0.23090372, Global Avg Loss: 0.97104175, Time: 0.0209 Steps: 49000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000043, Sample Num: 688, Cur Loss: 0.25109315, Cur Avg Loss: 0.21573432, Log Avg loss: 0.21812701, Global Avg Loss: 0.97088812, Time: 0.0209 Steps: 49010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000053, Sample Num: 848, Cur Loss: 0.19485943, Cur Avg Loss: 0.20724029, Log Avg loss: 0.17071597, Global Avg Loss: 0.97072489, Time: 0.0210 Steps: 49020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000063, Sample Num: 1008, Cur Loss: 0.12445312, Cur Avg Loss: 0.19897366, Log Avg loss: 0.15516048, Global Avg Loss: 0.97055855, Time: 0.0210 Steps: 49030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000073, Sample Num: 1168, Cur Loss: 0.09959115, Cur Avg Loss: 0.19538303, Log Avg loss: 0.17276207, Global Avg Loss: 0.97039587, Time: 0.0209 Steps: 49040, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000083, Sample Num: 1328, Cur Loss: 0.19585919, Cur Avg Loss: 0.19475487, Log Avg loss: 0.19016930, Global Avg Loss: 0.97023680, Time: 0.0210 Steps: 49050, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000093, Sample Num: 1488, Cur Loss: 0.10242927, Cur Avg Loss: 0.19512433, Log Avg loss: 0.19819085, Global Avg Loss: 0.97007943, Time: 0.0209 Steps: 49060, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000103, Sample Num: 1648, Cur Loss: 0.60745996, Cur Avg Loss: 0.19967497, Log Avg loss: 0.24199594, Global Avg Loss: 0.96993105, Time: 0.0210 Steps: 49070, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000113, Sample Num: 1808, Cur Loss: 0.09151528, Cur Avg Loss: 0.19837833, Log Avg loss: 0.18502293, Global Avg Loss: 0.96977113, Time: 0.0209 Steps: 49080, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000123, Sample Num: 1968, Cur Loss: 0.09947298, Cur Avg Loss: 0.19340973, Log Avg loss: 0.13726451, Global Avg Loss: 0.96960154, Time: 0.0210 Steps: 49090, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000133, Sample Num: 2128, Cur Loss: 0.24664448, Cur Avg Loss: 0.19139984, Log Avg loss: 0.16667820, Global Avg Loss: 0.96943801, Time: 0.0210 Steps: 49100, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000143, Sample Num: 2288, Cur Loss: 0.11421205, Cur Avg Loss: 0.19150504, Log Avg loss: 0.19290427, Global Avg Loss: 0.96927989, Time: 0.0210 Steps: 49110, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000153, Sample Num: 2448, Cur Loss: 0.25019652, Cur Avg Loss: 0.19628548, Log Avg loss: 0.26464570, Global Avg Loss: 0.96913644, Time: 0.0210 Steps: 49120, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000163, Sample Num: 2608, Cur Loss: 0.06519310, Cur Avg Loss: 0.19674104, Log Avg loss: 0.20371118, Global Avg Loss: 0.96898065, Time: 0.0209 Steps: 49130, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000173, Sample Num: 2768, Cur Loss: 0.24729496, Cur Avg Loss: 0.19595760, Log Avg loss: 0.18318742, Global Avg Loss: 0.96882074, Time: 0.0210 Steps: 49140, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000183, Sample Num: 2928, Cur Loss: 0.55429900, Cur Avg Loss: 0.19998097, Log Avg loss: 0.26958535, Global Avg Loss: 0.96867847, Time: 0.0209 Steps: 49150, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000193, Sample Num: 3088, Cur Loss: 0.08122806, Cur Avg Loss: 0.19942611, Log Avg loss: 0.18927209, Global Avg Loss: 0.96851993, Time: 0.0210 Steps: 49160, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000203, Sample Num: 3248, Cur Loss: 0.15333234, Cur Avg Loss: 0.19533413, Log Avg loss: 0.11635899, Global Avg Loss: 0.96834662, Time: 0.0210 Steps: 49170, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000213, Sample Num: 3408, Cur Loss: 0.40239212, Cur Avg Loss: 0.19459142, Log Avg loss: 0.17951435, Global Avg Loss: 0.96818622, Time: 0.0210 Steps: 49180, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000223, Sample Num: 3568, Cur Loss: 0.14415176, Cur Avg Loss: 0.19357156, Log Avg loss: 0.17184851, Global Avg Loss: 0.96802433, Time: 0.0209 Steps: 49190, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000233, Sample Num: 3728, Cur Loss: 0.17272964, Cur Avg Loss: 0.19341462, Log Avg loss: 0.18991495, Global Avg Loss: 0.96786618, Time: 0.0209 Steps: 49200, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000243, Sample Num: 3888, Cur Loss: 0.13407049, Cur Avg Loss: 0.19689659, Log Avg loss: 0.27802658, Global Avg Loss: 0.96772599, Time: 0.0209 Steps: 49210, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000253, Sample Num: 4048, Cur Loss: 0.22335745, Cur Avg Loss: 0.20277493, Log Avg loss: 0.34561862, Global Avg Loss: 0.96759960, Time: 0.0209 Steps: 49220, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000263, Sample Num: 4208, Cur Loss: 0.16746968, Cur Avg Loss: 0.20242555, Log Avg loss: 0.19358624, Global Avg Loss: 0.96744238, Time: 0.0211 Steps: 49230, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000273, Sample Num: 4368, Cur Loss: 0.20282830, Cur Avg Loss: 0.20290619, Log Avg loss: 0.21554688, Global Avg Loss: 0.96728968, Time: 0.0208 Steps: 49240, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000283, Sample Num: 4528, Cur Loss: 0.09727557, Cur Avg Loss: 0.20140842, Log Avg loss: 0.16051932, Global Avg Loss: 0.96712587, Time: 0.0208 Steps: 49250, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000293, Sample Num: 4688, Cur Loss: 0.13262907, Cur Avg Loss: 0.19929785, Log Avg loss: 0.13956884, Global Avg Loss: 0.96695787, Time: 0.0209 Steps: 49260, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000303, Sample Num: 4848, Cur Loss: 0.17225011, Cur Avg Loss: 0.19819606, Log Avg loss: 0.16591353, Global Avg Loss: 0.96679529, Time: 0.0209 Steps: 49270, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000313, Sample Num: 5008, Cur Loss: 0.23358096, Cur Avg Loss: 0.19953688, Log Avg loss: 0.24016356, Global Avg Loss: 0.96664784, Time: 0.0208 Steps: 49280, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000323, Sample Num: 5168, Cur Loss: 0.09422223, Cur Avg Loss: 0.20002142, Log Avg loss: 0.21518752, Global Avg Loss: 0.96649538, Time: 0.0209 Steps: 49290, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000333, Sample Num: 5328, Cur Loss: 0.11691269, Cur Avg Loss: 0.20031881, Log Avg loss: 0.20992470, Global Avg Loss: 0.96634192, Time: 0.0209 Steps: 49300, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000343, Sample Num: 5488, Cur Loss: 0.11450836, Cur Avg Loss: 0.19938970, Log Avg loss: 0.16845019, Global Avg Loss: 0.96618011, Time: 0.0208 Steps: 49310, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000353, Sample Num: 5648, Cur Loss: 0.15541579, Cur Avg Loss: 0.20160916, Log Avg loss: 0.27773687, Global Avg Loss: 0.96604052, Time: 0.0208 Steps: 49320, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000363, Sample Num: 5808, Cur Loss: 0.32560506, Cur Avg Loss: 0.20264138, Log Avg loss: 0.23907875, Global Avg Loss: 0.96589315, Time: 0.0208 Steps: 49330, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000373, Sample Num: 5968, Cur Loss: 0.29045585, Cur Avg Loss: 0.20261746, Log Avg loss: 0.20174891, Global Avg Loss: 0.96573828, Time: 0.0208 Steps: 49340, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000383, Sample Num: 6128, Cur Loss: 0.28952143, Cur Avg Loss: 0.20283506, Log Avg loss: 0.21095161, Global Avg Loss: 0.96558533, Time: 0.0209 Steps: 49350, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000393, Sample Num: 6288, Cur Loss: 0.11777239, Cur Avg Loss: 0.20389851, Log Avg loss: 0.24462850, Global Avg Loss: 0.96543927, Time: 0.0208 Steps: 49360, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000403, Sample Num: 6448, Cur Loss: 0.06006641, Cur Avg Loss: 0.20379792, Log Avg loss: 0.19984473, Global Avg Loss: 0.96528420, Time: 0.0209 Steps: 49370, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000413, Sample Num: 6608, Cur Loss: 0.16016208, Cur Avg Loss: 0.20317524, Log Avg loss: 0.17808139, Global Avg Loss: 0.96512478, Time: 0.0210 Steps: 49380, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000423, Sample Num: 6768, Cur Loss: 0.20551191, Cur Avg Loss: 0.20285794, Log Avg loss: 0.18975360, Global Avg Loss: 0.96496779, Time: 0.0209 Steps: 49390, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000433, Sample Num: 6928, Cur Loss: 0.04507124, Cur Avg Loss: 0.20187258, Log Avg loss: 0.16019183, Global Avg Loss: 0.96480488, Time: 0.0210 Steps: 49400, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000443, Sample Num: 7088, Cur Loss: 0.21457553, Cur Avg Loss: 0.20046712, Log Avg loss: 0.13961078, Global Avg Loss: 0.96463787, Time: 0.0208 Steps: 49410, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000453, Sample Num: 7248, Cur Loss: 0.11821697, Cur Avg Loss: 0.20005205, Log Avg loss: 0.18166447, Global Avg Loss: 0.96447944, Time: 0.0208 Steps: 49420, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000463, Sample Num: 7408, Cur Loss: 0.21248569, Cur Avg Loss: 0.20078388, Log Avg loss: 0.23393573, Global Avg Loss: 0.96433165, Time: 0.0208 Steps: 49430, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000473, Sample Num: 7568, Cur Loss: 0.06903256, Cur Avg Loss: 0.20133632, Log Avg loss: 0.22691397, Global Avg Loss: 0.96418249, Time: 0.0208 Steps: 49440, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000483, Sample Num: 7728, Cur Loss: 0.13767710, Cur Avg Loss: 0.20210081, Log Avg loss: 0.23826117, Global Avg Loss: 0.96403569, Time: 0.0208 Steps: 49450, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000493, Sample Num: 7888, Cur Loss: 0.09266640, Cur Avg Loss: 0.20185602, Log Avg loss: 0.19003296, Global Avg Loss: 0.96387920, Time: 0.0208 Steps: 49460, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000503, Sample Num: 8048, Cur Loss: 0.07490392, Cur Avg Loss: 0.20143131, Log Avg loss: 0.18049311, Global Avg Loss: 0.96372085, Time: 0.0209 Steps: 49470, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000513, Sample Num: 8208, Cur Loss: 0.27014911, Cur Avg Loss: 0.20035201, Log Avg loss: 0.14606324, Global Avg Loss: 0.96355560, Time: 0.0246 Steps: 49480, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000523, Sample Num: 8368, Cur Loss: 0.15939537, Cur Avg Loss: 0.19903470, Log Avg loss: 0.13145662, Global Avg Loss: 0.96338746, Time: 0.0209 Steps: 49490, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000533, Sample Num: 8528, Cur Loss: 0.48666316, Cur Avg Loss: 0.19912601, Log Avg loss: 0.20390170, Global Avg Loss: 0.96323403, Time: 0.0209 Steps: 49500, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000543, Sample Num: 8688, Cur Loss: 0.27061880, Cur Avg Loss: 0.19851966, Log Avg loss: 0.16620115, Global Avg Loss: 0.96307305, Time: 0.0209 Steps: 49510, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000553, Sample Num: 8848, Cur Loss: 0.30492178, Cur Avg Loss: 0.19939181, Log Avg loss: 0.24674958, Global Avg Loss: 0.96292839, Time: 0.0209 Steps: 49520, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000563, Sample Num: 9008, Cur Loss: 0.21088746, Cur Avg Loss: 0.20077973, Log Avg loss: 0.27753143, Global Avg Loss: 0.96279001, Time: 0.0209 Steps: 49530, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000573, Sample Num: 9168, Cur Loss: 0.17099945, Cur Avg Loss: 0.20104221, Log Avg loss: 0.21581998, Global Avg Loss: 0.96263923, Time: 0.0209 Steps: 49540, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000583, Sample Num: 9328, Cur Loss: 0.71067542, Cur Avg Loss: 0.20235383, Log Avg loss: 0.27750949, Global Avg Loss: 0.96250096, Time: 0.0209 Steps: 49550, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000593, Sample Num: 9488, Cur Loss: 0.17329702, Cur Avg Loss: 0.20364151, Log Avg loss: 0.27871351, Global Avg Loss: 0.96236299, Time: 0.0209 Steps: 49560, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000603, Sample Num: 9648, Cur Loss: 0.04803581, Cur Avg Loss: 0.20222177, Log Avg loss: 0.11803108, Global Avg Loss: 0.96219266, Time: 0.0209 Steps: 49570, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000613, Sample Num: 9808, Cur Loss: 0.36281091, Cur Avg Loss: 0.20224329, Log Avg loss: 0.20354082, Global Avg Loss: 0.96203964, Time: 0.0209 Steps: 49580, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000623, Sample Num: 9968, Cur Loss: 0.07424206, Cur Avg Loss: 0.20266700, Log Avg loss: 0.22864033, Global Avg Loss: 0.96189175, Time: 0.0209 Steps: 49590, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000633, Sample Num: 10128, Cur Loss: 0.11748771, Cur Avg Loss: 0.20336006, Log Avg loss: 0.24653774, Global Avg Loss: 0.96174753, Time: 0.0209 Steps: 49600, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000643, Sample Num: 10288, Cur Loss: 0.27522415, Cur Avg Loss: 0.20279891, Log Avg loss: 0.16727849, Global Avg Loss: 0.96158738, Time: 0.0210 Steps: 49610, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000653, Sample Num: 10448, Cur Loss: 0.24247505, Cur Avg Loss: 0.20176135, Log Avg loss: 0.13504616, Global Avg Loss: 0.96142081, Time: 0.0209 Steps: 49620, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000663, Sample Num: 10608, Cur Loss: 0.23823395, Cur Avg Loss: 0.20104180, Log Avg loss: 0.15405529, Global Avg Loss: 0.96125813, Time: 0.0209 Steps: 49630, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000673, Sample Num: 10768, Cur Loss: 0.18438017, Cur Avg Loss: 0.20107791, Log Avg loss: 0.20347148, Global Avg Loss: 0.96110547, Time: 0.0209 Steps: 49640, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000683, Sample Num: 10928, Cur Loss: 0.09850921, Cur Avg Loss: 0.20079450, Log Avg loss: 0.18172100, Global Avg Loss: 0.96094850, Time: 0.0209 Steps: 49650, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000693, Sample Num: 11088, Cur Loss: 0.22415921, Cur Avg Loss: 0.20119535, Log Avg loss: 0.22857387, Global Avg Loss: 0.96080102, Time: 0.0209 Steps: 49660, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000703, Sample Num: 11248, Cur Loss: 0.11542913, Cur Avg Loss: 0.20065230, Log Avg loss: 0.16301907, Global Avg Loss: 0.96064040, Time: 0.0209 Steps: 49670, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000713, Sample Num: 11408, Cur Loss: 0.05108845, Cur Avg Loss: 0.19941555, Log Avg loss: 0.11247145, Global Avg Loss: 0.96046968, Time: 0.0209 Steps: 49680, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000723, Sample Num: 11568, Cur Loss: 0.38701236, Cur Avg Loss: 0.19951103, Log Avg loss: 0.20631912, Global Avg Loss: 0.96031791, Time: 0.0209 Steps: 49690, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000733, Sample Num: 11728, Cur Loss: 0.09668820, Cur Avg Loss: 0.19899432, Log Avg loss: 0.16163597, Global Avg Loss: 0.96015721, Time: 0.0209 Steps: 49700, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000743, Sample Num: 11888, Cur Loss: 0.17433003, Cur Avg Loss: 0.20127085, Log Avg loss: 0.36814054, Global Avg Loss: 0.96003811, Time: 0.0209 Steps: 49710, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000753, Sample Num: 12048, Cur Loss: 0.69434345, Cur Avg Loss: 0.20249447, Log Avg loss: 0.29340950, Global Avg Loss: 0.95990404, Time: 0.0210 Steps: 49720, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000763, Sample Num: 12208, Cur Loss: 0.27165130, Cur Avg Loss: 0.20274249, Log Avg loss: 0.22141838, Global Avg Loss: 0.95975554, Time: 0.0209 Steps: 49730, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000773, Sample Num: 12368, Cur Loss: 0.21694981, Cur Avg Loss: 0.20388955, Log Avg loss: 0.29140990, Global Avg Loss: 0.95962117, Time: 0.0210 Steps: 49740, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000783, Sample Num: 12528, Cur Loss: 0.31167862, Cur Avg Loss: 0.20345497, Log Avg loss: 0.16986235, Global Avg Loss: 0.95946242, Time: 0.0209 Steps: 49750, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000793, Sample Num: 12688, Cur Loss: 0.16316974, Cur Avg Loss: 0.20267789, Log Avg loss: 0.14183251, Global Avg Loss: 0.95929811, Time: 0.0208 Steps: 49760, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000803, Sample Num: 12848, Cur Loss: 0.13590400, Cur Avg Loss: 0.20336806, Log Avg loss: 0.25809845, Global Avg Loss: 0.95915722, Time: 0.0209 Steps: 49770, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000813, Sample Num: 13008, Cur Loss: 0.15027110, Cur Avg Loss: 0.20272795, Log Avg loss: 0.15132734, Global Avg Loss: 0.95899494, Time: 0.0210 Steps: 49780, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000823, Sample Num: 13168, Cur Loss: 0.43415868, Cur Avg Loss: 0.20360579, Log Avg loss: 0.27497430, Global Avg Loss: 0.95885756, Time: 0.0208 Steps: 49790, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000833, Sample Num: 13328, Cur Loss: 0.15913461, Cur Avg Loss: 0.20321356, Log Avg loss: 0.17093295, Global Avg Loss: 0.95869934, Time: 0.0208 Steps: 49800, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000843, Sample Num: 13488, Cur Loss: 0.39649740, Cur Avg Loss: 0.20331146, Log Avg loss: 0.21146612, Global Avg Loss: 0.95854933, Time: 0.0209 Steps: 49810, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000853, Sample Num: 13648, Cur Loss: 0.27769667, Cur Avg Loss: 0.20510692, Log Avg loss: 0.35646474, Global Avg Loss: 0.95842847, Time: 0.0209 Steps: 49820, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000863, Sample Num: 13808, Cur Loss: 0.07679806, Cur Avg Loss: 0.20468100, Log Avg loss: 0.16834963, Global Avg Loss: 0.95826992, Time: 0.0209 Steps: 49830, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000873, Sample Num: 13968, Cur Loss: 0.38768926, Cur Avg Loss: 0.20481800, Log Avg loss: 0.21664125, Global Avg Loss: 0.95812112, Time: 0.0209 Steps: 49840, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000883, Sample Num: 14128, Cur Loss: 0.09198699, Cur Avg Loss: 0.20495253, Log Avg loss: 0.21669691, Global Avg Loss: 0.95797239, Time: 0.0209 Steps: 49850, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000893, Sample Num: 14288, Cur Loss: 0.15753488, Cur Avg Loss: 0.20546617, Log Avg loss: 0.25082069, Global Avg Loss: 0.95783056, Time: 0.0208 Steps: 49860, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000903, Sample Num: 14448, Cur Loss: 0.16818759, Cur Avg Loss: 0.20499645, Log Avg loss: 0.16305050, Global Avg Loss: 0.95767119, Time: 0.0208 Steps: 49870, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000913, Sample Num: 14608, Cur Loss: 0.24352127, Cur Avg Loss: 0.20470319, Log Avg loss: 0.17822202, Global Avg Loss: 0.95751492, Time: 0.0208 Steps: 49880, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000923, Sample Num: 14768, Cur Loss: 0.13916272, Cur Avg Loss: 0.20559260, Log Avg loss: 0.28679493, Global Avg Loss: 0.95738048, Time: 0.0208 Steps: 49890, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000933, Sample Num: 14928, Cur Loss: 0.07619278, Cur Avg Loss: 0.20492832, Log Avg loss: 0.14361571, Global Avg Loss: 0.95721740, Time: 0.0209 Steps: 49900, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000943, Sample Num: 15088, Cur Loss: 0.13282625, Cur Avg Loss: 0.20437614, Log Avg loss: 0.15285788, Global Avg Loss: 0.95705624, Time: 0.0208 Steps: 49910, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000953, Sample Num: 15248, Cur Loss: 0.12930684, Cur Avg Loss: 0.20409001, Log Avg loss: 0.17710806, Global Avg Loss: 0.95690000, Time: 0.0209 Steps: 49920, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000963, Sample Num: 15408, Cur Loss: 0.33581159, Cur Avg Loss: 0.20446784, Log Avg loss: 0.24047512, Global Avg Loss: 0.95675652, Time: 0.0209 Steps: 49930, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000973, Sample Num: 15568, Cur Loss: 0.21018693, Cur Avg Loss: 0.20533013, Log Avg loss: 0.28836793, Global Avg Loss: 0.95662268, Time: 0.0208 Steps: 49940, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000983, Sample Num: 15728, Cur Loss: 0.07147820, Cur Avg Loss: 0.20497795, Log Avg loss: 0.17071116, Global Avg Loss: 0.95646534, Time: 0.0209 Steps: 49950, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000993, Sample Num: 15888, Cur Loss: 0.13434891, Cur Avg Loss: 0.20577130, Log Avg loss: 0.28375736, Global Avg Loss: 0.95633069, Time: 0.0209 Steps: 49960, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001003, Sample Num: 16048, Cur Loss: 0.06381916, Cur Avg Loss: 0.20497620, Log Avg loss: 0.12602307, Global Avg Loss: 0.95616453, Time: 0.0209 Steps: 49970, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001013, Sample Num: 16208, Cur Loss: 0.15808362, Cur Avg Loss: 0.20460567, Log Avg loss: 0.16744171, Global Avg Loss: 0.95600672, Time: 0.0209 Steps: 49980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001023, Sample Num: 16368, Cur Loss: 0.09049769, Cur Avg Loss: 0.20393090, Log Avg loss: 0.13557670, Global Avg Loss: 0.95584260, Time: 0.0208 Steps: 49990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001033, Sample Num: 16528, Cur Loss: 0.32564485, Cur Avg Loss: 0.20395971, Log Avg loss: 0.20690676, Global Avg Loss: 0.95569281, Time: 0.0210 Steps: 50000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001043, Sample Num: 16688, Cur Loss: 0.74620020, Cur Avg Loss: 0.20441895, Log Avg loss: 0.25185864, Global Avg Loss: 0.95555208, Time: 0.0209 Steps: 50010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001053, Sample Num: 16848, Cur Loss: 0.33354002, Cur Avg Loss: 0.20399364, Log Avg loss: 0.15963325, Global Avg Loss: 0.95539296, Time: 0.0209 Steps: 50020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001063, Sample Num: 17008, Cur Loss: 0.17104706, Cur Avg Loss: 0.20406649, Log Avg loss: 0.21173773, Global Avg Loss: 0.95524431, Time: 0.0209 Steps: 50030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001073, Sample Num: 17168, Cur Loss: 0.26260394, Cur Avg Loss: 0.20418359, Log Avg loss: 0.21663179, Global Avg Loss: 0.95509671, Time: 0.0208 Steps: 50040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001083, Sample Num: 17328, Cur Loss: 0.34899640, Cur Avg Loss: 0.20366835, Log Avg loss: 0.14838316, Global Avg Loss: 0.95493553, Time: 0.0209 Steps: 50050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001093, Sample Num: 17488, Cur Loss: 0.22619864, Cur Avg Loss: 0.20326410, Log Avg loss: 0.15948404, Global Avg Loss: 0.95477663, Time: 0.0208 Steps: 50060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001103, Sample Num: 17648, Cur Loss: 0.33560324, Cur Avg Loss: 0.20298822, Log Avg loss: 0.17283399, Global Avg Loss: 0.95462046, Time: 0.0209 Steps: 50070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001113, Sample Num: 17808, Cur Loss: 0.22775397, Cur Avg Loss: 0.20292484, Log Avg loss: 0.19593417, Global Avg Loss: 0.95446896, Time: 0.0209 Steps: 50080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001123, Sample Num: 17968, Cur Loss: 0.31464779, Cur Avg Loss: 0.20345043, Log Avg loss: 0.26194864, Global Avg Loss: 0.95433071, Time: 0.0209 Steps: 50090, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001133, Sample Num: 18128, Cur Loss: 0.22814550, Cur Avg Loss: 0.20350458, Log Avg loss: 0.20958511, Global Avg Loss: 0.95418206, Time: 0.0209 Steps: 50100, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001143, Sample Num: 18288, Cur Loss: 0.19548324, Cur Avg Loss: 0.20506037, Log Avg loss: 0.38133134, Global Avg Loss: 0.95406774, Time: 0.0210 Steps: 50110, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001153, Sample Num: 18448, Cur Loss: 0.28121683, Cur Avg Loss: 0.20632215, Log Avg loss: 0.35054426, Global Avg Loss: 0.95394732, Time: 0.0209 Steps: 50120, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001163, Sample Num: 18608, Cur Loss: 0.17306888, Cur Avg Loss: 0.20742058, Log Avg loss: 0.33406892, Global Avg Loss: 0.95382367, Time: 0.0209 Steps: 50130, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001173, Sample Num: 18768, Cur Loss: 0.40810359, Cur Avg Loss: 0.20738394, Log Avg loss: 0.20312332, Global Avg Loss: 0.95367395, Time: 0.0209 Steps: 50140, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001183, Sample Num: 18928, Cur Loss: 0.12181407, Cur Avg Loss: 0.20763571, Log Avg loss: 0.23716803, Global Avg Loss: 0.95353107, Time: 0.0209 Steps: 50150, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001193, Sample Num: 19088, Cur Loss: 0.09276255, Cur Avg Loss: 0.20705743, Log Avg loss: 0.13864647, Global Avg Loss: 0.95336862, Time: 0.0208 Steps: 50160, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001203, Sample Num: 19248, Cur Loss: 0.08709093, Cur Avg Loss: 0.20676808, Log Avg loss: 0.17224875, Global Avg Loss: 0.95321292, Time: 0.0209 Steps: 50170, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001213, Sample Num: 19408, Cur Loss: 0.21899779, Cur Avg Loss: 0.20672066, Log Avg loss: 0.20101673, Global Avg Loss: 0.95306302, Time: 0.0209 Steps: 50180, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001223, Sample Num: 19568, Cur Loss: 0.06037217, Cur Avg Loss: 0.20693187, Log Avg loss: 0.23255105, Global Avg Loss: 0.95291947, Time: 0.0208 Steps: 50190, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001233, Sample Num: 19728, Cur Loss: 0.18311886, Cur Avg Loss: 0.20657826, Log Avg loss: 0.16333174, Global Avg Loss: 0.95276218, Time: 0.0209 Steps: 50200, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001243, Sample Num: 19888, Cur Loss: 0.09915097, Cur Avg Loss: 0.20570891, Log Avg loss: 0.09851823, Global Avg Loss: 0.95259204, Time: 0.0209 Steps: 50210, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001253, Sample Num: 20048, Cur Loss: 0.09063886, Cur Avg Loss: 0.20596913, Log Avg loss: 0.23831444, Global Avg Loss: 0.95244981, Time: 0.0209 Steps: 50220, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001263, Sample Num: 20208, Cur Loss: 0.26383975, Cur Avg Loss: 0.20592508, Log Avg loss: 0.20040628, Global Avg Loss: 0.95230009, Time: 0.0209 Steps: 50230, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001273, Sample Num: 20368, Cur Loss: 0.26644540, Cur Avg Loss: 0.20679009, Log Avg loss: 0.31604062, Global Avg Loss: 0.95217345, Time: 0.0209 Steps: 50240, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001283, Sample Num: 20528, Cur Loss: 0.12132568, Cur Avg Loss: 0.20676943, Log Avg loss: 0.20413884, Global Avg Loss: 0.95202459, Time: 0.0246 Steps: 50250, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001293, Sample Num: 20688, Cur Loss: 0.34001672, Cur Avg Loss: 0.20695728, Log Avg loss: 0.23105868, Global Avg Loss: 0.95188114, Time: 0.0209 Steps: 50260, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001303, Sample Num: 20848, Cur Loss: 0.25445724, Cur Avg Loss: 0.20691797, Log Avg loss: 0.20183583, Global Avg Loss: 0.95173194, Time: 0.0209 Steps: 50270, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001313, Sample Num: 21008, Cur Loss: 0.65052873, Cur Avg Loss: 0.20850108, Log Avg loss: 0.41477967, Global Avg Loss: 0.95162514, Time: 0.0209 Steps: 50280, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001323, Sample Num: 21168, Cur Loss: 0.10183519, Cur Avg Loss: 0.20855226, Log Avg loss: 0.21527170, Global Avg Loss: 0.95147872, Time: 0.0209 Steps: 50290, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001333, Sample Num: 21328, Cur Loss: 0.05561260, Cur Avg Loss: 0.20828652, Log Avg loss: 0.17313013, Global Avg Loss: 0.95132398, Time: 0.0208 Steps: 50300, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001343, Sample Num: 21488, Cur Loss: 0.41364831, Cur Avg Loss: 0.20832324, Log Avg loss: 0.21321744, Global Avg Loss: 0.95117727, Time: 0.0209 Steps: 50310, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001353, Sample Num: 21648, Cur Loss: 0.06191041, Cur Avg Loss: 0.20793930, Log Avg loss: 0.15637560, Global Avg Loss: 0.95101932, Time: 0.0209 Steps: 50320, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001363, Sample Num: 21808, Cur Loss: 0.07462963, Cur Avg Loss: 0.20759722, Log Avg loss: 0.16131462, Global Avg Loss: 0.95086241, Time: 0.0209 Steps: 50330, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001373, Sample Num: 21968, Cur Loss: 0.11596655, Cur Avg Loss: 0.20751910, Log Avg loss: 0.19687096, Global Avg Loss: 0.95071264, Time: 0.0209 Steps: 50340, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001383, Sample Num: 22128, Cur Loss: 0.04106956, Cur Avg Loss: 0.20771683, Log Avg loss: 0.23486498, Global Avg Loss: 0.95057046, Time: 0.0208 Steps: 50350, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001393, Sample Num: 22288, Cur Loss: 0.25433680, Cur Avg Loss: 0.20743752, Log Avg loss: 0.16880913, Global Avg Loss: 0.95041523, Time: 0.0208 Steps: 50360, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001403, Sample Num: 22448, Cur Loss: 0.29822123, Cur Avg Loss: 0.20699459, Log Avg loss: 0.14529417, Global Avg Loss: 0.95025538, Time: 0.0209 Steps: 50370, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001413, Sample Num: 22608, Cur Loss: 0.38599515, Cur Avg Loss: 0.20703363, Log Avg loss: 0.21251085, Global Avg Loss: 0.95010895, Time: 0.0208 Steps: 50380, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001423, Sample Num: 22768, Cur Loss: 0.09229381, Cur Avg Loss: 0.20771131, Log Avg loss: 0.30346774, Global Avg Loss: 0.94998062, Time: 0.0208 Steps: 50390, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001433, Sample Num: 22928, Cur Loss: 0.35834917, Cur Avg Loss: 0.20797090, Log Avg loss: 0.24491087, Global Avg Loss: 0.94984073, Time: 0.0209 Steps: 50400, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001443, Sample Num: 23088, Cur Loss: 0.51875019, Cur Avg Loss: 0.20867602, Log Avg loss: 0.30971953, Global Avg Loss: 0.94971374, Time: 0.0209 Steps: 50410, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001453, Sample Num: 23248, Cur Loss: 0.21770403, Cur Avg Loss: 0.20874816, Log Avg loss: 0.21915853, Global Avg Loss: 0.94956885, Time: 0.0209 Steps: 50420, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001463, Sample Num: 23408, Cur Loss: 0.20404874, Cur Avg Loss: 0.20897520, Log Avg loss: 0.24196330, Global Avg Loss: 0.94942854, Time: 0.0208 Steps: 50430, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001473, Sample Num: 23568, Cur Loss: 0.27849913, Cur Avg Loss: 0.20997252, Log Avg loss: 0.35588141, Global Avg Loss: 0.94931086, Time: 0.0209 Steps: 50440, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001483, Sample Num: 23728, Cur Loss: 0.22397637, Cur Avg Loss: 0.20998829, Log Avg loss: 0.21231091, Global Avg Loss: 0.94916478, Time: 0.0209 Steps: 50450, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001493, Sample Num: 23888, Cur Loss: 0.38474423, Cur Avg Loss: 0.20962138, Log Avg loss: 0.15520799, Global Avg Loss: 0.94900743, Time: 0.0209 Steps: 50460, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001503, Sample Num: 24048, Cur Loss: 0.30377507, Cur Avg Loss: 0.21012598, Log Avg loss: 0.28546315, Global Avg Loss: 0.94887596, Time: 0.0209 Steps: 50470, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001513, Sample Num: 24208, Cur Loss: 0.54952621, Cur Avg Loss: 0.21008514, Log Avg loss: 0.20394698, Global Avg Loss: 0.94872839, Time: 0.0209 Steps: 50480, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001523, Sample Num: 24368, Cur Loss: 0.18212086, Cur Avg Loss: 0.20971194, Log Avg loss: 0.15324724, Global Avg Loss: 0.94857084, Time: 0.0209 Steps: 50490, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001533, Sample Num: 24528, Cur Loss: 0.20376819, Cur Avg Loss: 0.20955631, Log Avg loss: 0.18585380, Global Avg Loss: 0.94841981, Time: 0.0208 Steps: 50500, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001543, Sample Num: 24688, Cur Loss: 0.15557140, Cur Avg Loss: 0.20888405, Log Avg loss: 0.10582595, Global Avg Loss: 0.94825299, Time: 0.0209 Steps: 50510, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001553, Sample Num: 24848, Cur Loss: 0.27663913, Cur Avg Loss: 0.20956166, Log Avg loss: 0.31411690, Global Avg Loss: 0.94812747, Time: 0.0209 Steps: 50520, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001563, Sample Num: 25008, Cur Loss: 0.30367497, Cur Avg Loss: 0.21020069, Log Avg loss: 0.30944258, Global Avg Loss: 0.94800107, Time: 0.0209 Steps: 50530, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001573, Sample Num: 25168, Cur Loss: 0.15883413, Cur Avg Loss: 0.21009917, Log Avg loss: 0.19423162, Global Avg Loss: 0.94785193, Time: 0.0209 Steps: 50540, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001583, Sample Num: 25328, Cur Loss: 0.24308747, Cur Avg Loss: 0.21028330, Log Avg loss: 0.23924721, Global Avg Loss: 0.94771175, Time: 0.0209 Steps: 50550, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001593, Sample Num: 25488, Cur Loss: 0.24430515, Cur Avg Loss: 0.21004682, Log Avg loss: 0.17261173, Global Avg Loss: 0.94755844, Time: 0.0209 Steps: 50560, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001603, Sample Num: 25648, Cur Loss: 0.22852902, Cur Avg Loss: 0.21055705, Log Avg loss: 0.29183665, Global Avg Loss: 0.94742878, Time: 0.0209 Steps: 50570, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001613, Sample Num: 25808, Cur Loss: 0.37711063, Cur Avg Loss: 0.21065009, Log Avg loss: 0.22556439, Global Avg Loss: 0.94728606, Time: 0.0209 Steps: 50580, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001623, Sample Num: 25968, Cur Loss: 0.21514806, Cur Avg Loss: 0.21068057, Log Avg loss: 0.21559683, Global Avg Loss: 0.94714143, Time: 0.0208 Steps: 50590, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001633, Sample Num: 26128, Cur Loss: 0.14371116, Cur Avg Loss: 0.21038130, Log Avg loss: 0.16181037, Global Avg Loss: 0.94698623, Time: 0.0209 Steps: 50600, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001643, Sample Num: 26288, Cur Loss: 0.21891317, Cur Avg Loss: 0.21015797, Log Avg loss: 0.17368822, Global Avg Loss: 0.94683343, Time: 0.0209 Steps: 50610, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001653, Sample Num: 26448, Cur Loss: 0.26306099, Cur Avg Loss: 0.20973576, Log Avg loss: 0.14036573, Global Avg Loss: 0.94667411, Time: 0.0210 Steps: 50620, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001663, Sample Num: 26608, Cur Loss: 0.05750592, Cur Avg Loss: 0.20962293, Log Avg loss: 0.19097245, Global Avg Loss: 0.94652485, Time: 0.0209 Steps: 50630, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001673, Sample Num: 26768, Cur Loss: 0.02518506, Cur Avg Loss: 0.20933654, Log Avg loss: 0.16171025, Global Avg Loss: 0.94636987, Time: 0.0209 Steps: 50640, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001683, Sample Num: 26928, Cur Loss: 0.10929798, Cur Avg Loss: 0.20931618, Log Avg loss: 0.20590939, Global Avg Loss: 0.94622368, Time: 0.0209 Steps: 50650, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001693, Sample Num: 27088, Cur Loss: 0.19458506, Cur Avg Loss: 0.20911101, Log Avg loss: 0.17458148, Global Avg Loss: 0.94607136, Time: 0.0209 Steps: 50660, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001703, Sample Num: 27248, Cur Loss: 0.14661650, Cur Avg Loss: 0.20918687, Log Avg loss: 0.22202927, Global Avg Loss: 0.94592847, Time: 0.0209 Steps: 50670, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001713, Sample Num: 27408, Cur Loss: 0.09220079, Cur Avg Loss: 0.20914278, Log Avg loss: 0.20163355, Global Avg Loss: 0.94578161, Time: 0.0209 Steps: 50680, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001723, Sample Num: 27568, Cur Loss: 0.08986787, Cur Avg Loss: 0.20945454, Log Avg loss: 0.26285965, Global Avg Loss: 0.94564688, Time: 0.0209 Steps: 50690, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001733, Sample Num: 27728, Cur Loss: 0.35875720, Cur Avg Loss: 0.20917914, Log Avg loss: 0.16172834, Global Avg Loss: 0.94549226, Time: 0.0209 Steps: 50700, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001743, Sample Num: 27888, Cur Loss: 0.25141788, Cur Avg Loss: 0.20953680, Log Avg loss: 0.27151937, Global Avg Loss: 0.94535936, Time: 0.0209 Steps: 50710, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001753, Sample Num: 28048, Cur Loss: 0.41450122, Cur Avg Loss: 0.20948757, Log Avg loss: 0.20090700, Global Avg Loss: 0.94521258, Time: 0.0209 Steps: 50720, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001763, Sample Num: 28208, Cur Loss: 0.30406612, Cur Avg Loss: 0.20972943, Log Avg loss: 0.25212738, Global Avg Loss: 0.94507596, Time: 0.0210 Steps: 50730, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001773, Sample Num: 28368, Cur Loss: 0.06278942, Cur Avg Loss: 0.21002932, Log Avg loss: 0.26289899, Global Avg Loss: 0.94494151, Time: 0.0209 Steps: 50740, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001783, Sample Num: 28528, Cur Loss: 0.29828781, Cur Avg Loss: 0.20978049, Log Avg loss: 0.16566300, Global Avg Loss: 0.94478796, Time: 0.0209 Steps: 50750, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001793, Sample Num: 28688, Cur Loss: 0.11877685, Cur Avg Loss: 0.20962829, Log Avg loss: 0.18249215, Global Avg Loss: 0.94463778, Time: 0.0245 Steps: 50760, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001803, Sample Num: 28848, Cur Loss: 0.20055796, Cur Avg Loss: 0.20988291, Log Avg loss: 0.25553571, Global Avg Loss: 0.94450205, Time: 0.0209 Steps: 50770, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001813, Sample Num: 29008, Cur Loss: 0.43679991, Cur Avg Loss: 0.21049530, Log Avg loss: 0.32090910, Global Avg Loss: 0.94437925, Time: 0.0209 Steps: 50780, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001823, Sample Num: 29168, Cur Loss: 0.34611109, Cur Avg Loss: 0.21075971, Log Avg loss: 0.25869746, Global Avg Loss: 0.94424425, Time: 0.0210 Steps: 50790, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001833, Sample Num: 29328, Cur Loss: 0.26014248, Cur Avg Loss: 0.21105116, Log Avg loss: 0.26418302, Global Avg Loss: 0.94411038, Time: 0.0209 Steps: 50800, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001843, Sample Num: 29488, Cur Loss: 0.34302574, Cur Avg Loss: 0.21115850, Log Avg loss: 0.23083356, Global Avg Loss: 0.94397000, Time: 0.0209 Steps: 50810, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001853, Sample Num: 29648, Cur Loss: 0.20886120, Cur Avg Loss: 0.21129262, Log Avg loss: 0.23601047, Global Avg Loss: 0.94383069, Time: 0.0209 Steps: 50820, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001863, Sample Num: 29808, Cur Loss: 0.12139939, Cur Avg Loss: 0.21109004, Log Avg loss: 0.17355282, Global Avg Loss: 0.94367915, Time: 0.0209 Steps: 50830, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001873, Sample Num: 29968, Cur Loss: 0.08316221, Cur Avg Loss: 0.21136186, Log Avg loss: 0.26200013, Global Avg Loss: 0.94354506, Time: 0.0209 Steps: 50840, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001883, Sample Num: 30128, Cur Loss: 0.12579623, Cur Avg Loss: 0.21114762, Log Avg loss: 0.17102098, Global Avg Loss: 0.94339314, Time: 0.0209 Steps: 50850, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001893, Sample Num: 30288, Cur Loss: 0.17555131, Cur Avg Loss: 0.21097674, Log Avg loss: 0.17880005, Global Avg Loss: 0.94324281, Time: 0.0209 Steps: 50860, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001903, Sample Num: 30448, Cur Loss: 0.18019009, Cur Avg Loss: 0.21114869, Log Avg loss: 0.24369830, Global Avg Loss: 0.94310529, Time: 0.0209 Steps: 50870, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001913, Sample Num: 30608, Cur Loss: 0.15300341, Cur Avg Loss: 0.21122826, Log Avg loss: 0.22637205, Global Avg Loss: 0.94296443, Time: 0.0208 Steps: 50880, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001923, Sample Num: 30768, Cur Loss: 0.14534795, Cur Avg Loss: 0.21116750, Log Avg loss: 0.19954270, Global Avg Loss: 0.94281834, Time: 0.0209 Steps: 50890, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001933, Sample Num: 30928, Cur Loss: 0.09352177, Cur Avg Loss: 0.21145142, Log Avg loss: 0.26605083, Global Avg Loss: 0.94268538, Time: 0.0209 Steps: 50900, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001943, Sample Num: 31088, Cur Loss: 0.11087369, Cur Avg Loss: 0.21122239, Log Avg loss: 0.16694920, Global Avg Loss: 0.94253301, Time: 0.0209 Steps: 50910, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001953, Sample Num: 31248, Cur Loss: 0.15907119, Cur Avg Loss: 0.21108063, Log Avg loss: 0.18353816, Global Avg Loss: 0.94238395, Time: 0.0208 Steps: 50920, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001963, Sample Num: 31408, Cur Loss: 0.60665953, Cur Avg Loss: 0.21130985, Log Avg loss: 0.25607610, Global Avg Loss: 0.94224920, Time: 0.0209 Steps: 50930, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001973, Sample Num: 31568, Cur Loss: 0.02072749, Cur Avg Loss: 0.21125316, Log Avg loss: 0.20012558, Global Avg Loss: 0.94210351, Time: 0.0209 Steps: 50940, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001983, Sample Num: 31728, Cur Loss: 0.28630304, Cur Avg Loss: 0.21091156, Log Avg loss: 0.14351268, Global Avg Loss: 0.94194677, Time: 0.0209 Steps: 50950, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001993, Sample Num: 31888, Cur Loss: 0.26195776, Cur Avg Loss: 0.21090487, Log Avg loss: 0.20957864, Global Avg Loss: 0.94180306, Time: 0.0209 Steps: 50960, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002003, Sample Num: 32048, Cur Loss: 0.22062902, Cur Avg Loss: 0.21072464, Log Avg loss: 0.17480429, Global Avg Loss: 0.94165258, Time: 0.0209 Steps: 50970, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002013, Sample Num: 32208, Cur Loss: 0.28290683, Cur Avg Loss: 0.21090728, Log Avg loss: 0.24749031, Global Avg Loss: 0.94151641, Time: 0.0209 Steps: 50980, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002023, Sample Num: 32368, Cur Loss: 0.07269064, Cur Avg Loss: 0.21060347, Log Avg loss: 0.14944638, Global Avg Loss: 0.94136107, Time: 0.0209 Steps: 50990, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002033, Sample Num: 32528, Cur Loss: 0.17426294, Cur Avg Loss: 0.21101785, Log Avg loss: 0.29484814, Global Avg Loss: 0.94123431, Time: 0.0209 Steps: 51000, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002043, Sample Num: 32688, Cur Loss: 0.03276137, Cur Avg Loss: 0.21085637, Log Avg loss: 0.17802706, Global Avg Loss: 0.94108469, Time: 0.0209 Steps: 51010, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002053, Sample Num: 32848, Cur Loss: 0.10260677, Cur Avg Loss: 0.21089748, Log Avg loss: 0.21929473, Global Avg Loss: 0.94094322, Time: 0.0248 Steps: 51020, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002063, Sample Num: 33008, Cur Loss: 0.13971247, Cur Avg Loss: 0.21096623, Log Avg loss: 0.22508257, Global Avg Loss: 0.94080293, Time: 0.0210 Steps: 51030, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002073, Sample Num: 33168, Cur Loss: 0.20517343, Cur Avg Loss: 0.21092321, Log Avg loss: 0.20204740, Global Avg Loss: 0.94065819, Time: 0.0209 Steps: 51040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002083, Sample Num: 33328, Cur Loss: 0.12282158, Cur Avg Loss: 0.21065577, Log Avg loss: 0.15521490, Global Avg Loss: 0.94050433, Time: 0.0209 Steps: 51050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002093, Sample Num: 33488, Cur Loss: 0.06599764, Cur Avg Loss: 0.21056360, Log Avg loss: 0.19136462, Global Avg Loss: 0.94035762, Time: 0.0208 Steps: 51060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002103, Sample Num: 33648, Cur Loss: 0.13326097, Cur Avg Loss: 0.21042627, Log Avg loss: 0.18168408, Global Avg Loss: 0.94020906, Time: 0.0209 Steps: 51070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002113, Sample Num: 33808, Cur Loss: 0.14532213, Cur Avg Loss: 0.21028802, Log Avg loss: 0.18121342, Global Avg Loss: 0.94006047, Time: 0.0211 Steps: 51080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002123, Sample Num: 33968, Cur Loss: 0.17275664, Cur Avg Loss: 0.21029203, Log Avg loss: 0.21113862, Global Avg Loss: 0.93991780, Time: 0.0210 Steps: 51090, Updated lr: 0.000052 ***** Running evaluation checkpoint-51096 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-51096 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.700372, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.268482, "eval_total_loss": 188.742939, "eval_mae": 0.33862, "eval_mse": 0.268585, "eval_r2": 0.829269, "eval_sp_statistic": 0.915867, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.927672, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.189109, "test_total_loss": 94.932543, "test_mae": 0.284982, "test_mse": 0.189175, "test_r2": 0.877905, "test_sp_statistic": 0.908361, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.944919, "test_ps_pvalue": 0.0, "lr": 5.249312470365102e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9398284226172801, "train_cur_epoch_loss": 447.5227517513558, "train_cur_epoch_avg_loss": 0.21020326526601965, "train_cur_epoch_time": 44.70037245750427, "train_cur_epoch_avg_time": 0.020995947608033947, "epoch": 24, "step": 51096} ################################################## Training, Epoch: 0025, Batch: 000004, Sample Num: 64, Cur Loss: 0.12687016, Cur Avg Loss: 0.16677698, Log Avg loss: 0.17398867, Global Avg Loss: 0.93976791, Time: 0.0230 Steps: 51100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000014, Sample Num: 224, Cur Loss: 0.15545255, Cur Avg Loss: 0.16642601, Log Avg loss: 0.16628562, Global Avg Loss: 0.93961657, Time: 0.0211 Steps: 51110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000024, Sample Num: 384, Cur Loss: 0.09834753, Cur Avg Loss: 0.19015503, Log Avg loss: 0.22337566, Global Avg Loss: 0.93947646, Time: 0.0211 Steps: 51120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000034, Sample Num: 544, Cur Loss: 0.14055178, Cur Avg Loss: 0.18453637, Log Avg loss: 0.17105159, Global Avg Loss: 0.93932617, Time: 0.0211 Steps: 51130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000044, Sample Num: 704, Cur Loss: 0.54090244, Cur Avg Loss: 0.19778848, Log Avg loss: 0.24284564, Global Avg Loss: 0.93918998, Time: 0.0211 Steps: 51140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000054, Sample Num: 864, Cur Loss: 0.30903870, Cur Avg Loss: 0.20674266, Log Avg loss: 0.24614108, Global Avg Loss: 0.93905449, Time: 0.0211 Steps: 51150, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000064, Sample Num: 1024, Cur Loss: 0.04812658, Cur Avg Loss: 0.19899170, Log Avg loss: 0.15713649, Global Avg Loss: 0.93890165, Time: 0.0212 Steps: 51160, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000074, Sample Num: 1184, Cur Loss: 0.27492514, Cur Avg Loss: 0.20269362, Log Avg loss: 0.22638593, Global Avg Loss: 0.93876241, Time: 0.0210 Steps: 51170, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000084, Sample Num: 1344, Cur Loss: 0.10648238, Cur Avg Loss: 0.19348972, Log Avg loss: 0.12538088, Global Avg Loss: 0.93860348, Time: 0.0212 Steps: 51180, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000094, Sample Num: 1504, Cur Loss: 0.26329315, Cur Avg Loss: 0.19252351, Log Avg loss: 0.18440726, Global Avg Loss: 0.93845615, Time: 0.0210 Steps: 51190, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000104, Sample Num: 1664, Cur Loss: 0.10562792, Cur Avg Loss: 0.19003041, Log Avg loss: 0.16659528, Global Avg Loss: 0.93830540, Time: 0.0210 Steps: 51200, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000114, Sample Num: 1824, Cur Loss: 0.71981800, Cur Avg Loss: 0.19883347, Log Avg loss: 0.29038535, Global Avg Loss: 0.93817887, Time: 0.0212 Steps: 51210, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000124, Sample Num: 1984, Cur Loss: 0.08152627, Cur Avg Loss: 0.19721813, Log Avg loss: 0.17880327, Global Avg Loss: 0.93803062, Time: 0.0211 Steps: 51220, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000134, Sample Num: 2144, Cur Loss: 0.17381226, Cur Avg Loss: 0.19234534, Log Avg loss: 0.13192273, Global Avg Loss: 0.93787326, Time: 0.0211 Steps: 51230, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000144, Sample Num: 2304, Cur Loss: 0.21556276, Cur Avg Loss: 0.19395455, Log Avg loss: 0.21551795, Global Avg Loss: 0.93773229, Time: 0.0212 Steps: 51240, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000154, Sample Num: 2464, Cur Loss: 0.32638815, Cur Avg Loss: 0.19240954, Log Avg loss: 0.17016141, Global Avg Loss: 0.93758252, Time: 0.0210 Steps: 51250, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000164, Sample Num: 2624, Cur Loss: 0.06399367, Cur Avg Loss: 0.19462891, Log Avg loss: 0.22880725, Global Avg Loss: 0.93744425, Time: 0.0212 Steps: 51260, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000174, Sample Num: 2784, Cur Loss: 0.14922068, Cur Avg Loss: 0.19526288, Log Avg loss: 0.20565994, Global Avg Loss: 0.93730152, Time: 0.0210 Steps: 51270, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000184, Sample Num: 2944, Cur Loss: 0.26811320, Cur Avg Loss: 0.19771431, Log Avg loss: 0.24036912, Global Avg Loss: 0.93716561, Time: 0.0211 Steps: 51280, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000194, Sample Num: 3104, Cur Loss: 0.50801229, Cur Avg Loss: 0.19932593, Log Avg loss: 0.22897981, Global Avg Loss: 0.93702754, Time: 0.0210 Steps: 51290, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000204, Sample Num: 3264, Cur Loss: 0.36117756, Cur Avg Loss: 0.20268727, Log Avg loss: 0.26789719, Global Avg Loss: 0.93689710, Time: 0.0211 Steps: 51300, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000214, Sample Num: 3424, Cur Loss: 0.32515684, Cur Avg Loss: 0.20311407, Log Avg loss: 0.21182079, Global Avg Loss: 0.93675579, Time: 0.0211 Steps: 51310, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000224, Sample Num: 3584, Cur Loss: 0.03964845, Cur Avg Loss: 0.20215154, Log Avg loss: 0.18155346, Global Avg Loss: 0.93660863, Time: 0.0211 Steps: 51320, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000234, Sample Num: 3744, Cur Loss: 0.21919796, Cur Avg Loss: 0.20119283, Log Avg loss: 0.17971780, Global Avg Loss: 0.93646118, Time: 0.0212 Steps: 51330, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000244, Sample Num: 3904, Cur Loss: 0.11958542, Cur Avg Loss: 0.19958841, Log Avg loss: 0.16204487, Global Avg Loss: 0.93631034, Time: 0.0210 Steps: 51340, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000254, Sample Num: 4064, Cur Loss: 0.18068349, Cur Avg Loss: 0.20232813, Log Avg loss: 0.26917734, Global Avg Loss: 0.93618042, Time: 0.0212 Steps: 51350, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000264, Sample Num: 4224, Cur Loss: 0.28338468, Cur Avg Loss: 0.20320646, Log Avg loss: 0.22551599, Global Avg Loss: 0.93604205, Time: 0.0210 Steps: 51360, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000274, Sample Num: 4384, Cur Loss: 0.08357465, Cur Avg Loss: 0.20354877, Log Avg loss: 0.21258569, Global Avg Loss: 0.93590122, Time: 0.0210 Steps: 51370, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000284, Sample Num: 4544, Cur Loss: 0.13663460, Cur Avg Loss: 0.20272058, Log Avg loss: 0.18002827, Global Avg Loss: 0.93575410, Time: 0.0210 Steps: 51380, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000294, Sample Num: 4704, Cur Loss: 0.05782047, Cur Avg Loss: 0.20183584, Log Avg loss: 0.17670935, Global Avg Loss: 0.93560640, Time: 0.0210 Steps: 51390, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000304, Sample Num: 4864, Cur Loss: 0.08201148, Cur Avg Loss: 0.20040891, Log Avg loss: 0.15845693, Global Avg Loss: 0.93545520, Time: 0.0211 Steps: 51400, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000314, Sample Num: 5024, Cur Loss: 0.30710006, Cur Avg Loss: 0.20289051, Log Avg loss: 0.27833129, Global Avg Loss: 0.93532738, Time: 0.0210 Steps: 51410, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000324, Sample Num: 5184, Cur Loss: 0.07430769, Cur Avg Loss: 0.20545928, Log Avg loss: 0.28611874, Global Avg Loss: 0.93520113, Time: 0.0210 Steps: 51420, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000334, Sample Num: 5344, Cur Loss: 0.15873122, Cur Avg Loss: 0.20465758, Log Avg loss: 0.17868254, Global Avg Loss: 0.93505403, Time: 0.0210 Steps: 51430, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000344, Sample Num: 5504, Cur Loss: 0.36163896, Cur Avg Loss: 0.20519040, Log Avg loss: 0.22298654, Global Avg Loss: 0.93491560, Time: 0.0210 Steps: 51440, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000354, Sample Num: 5664, Cur Loss: 0.20073676, Cur Avg Loss: 0.20652166, Log Avg loss: 0.25231694, Global Avg Loss: 0.93478293, Time: 0.0210 Steps: 51450, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000364, Sample Num: 5824, Cur Loss: 0.18332475, Cur Avg Loss: 0.20652600, Log Avg loss: 0.20667965, Global Avg Loss: 0.93464144, Time: 0.0210 Steps: 51460, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000374, Sample Num: 5984, Cur Loss: 0.67363346, Cur Avg Loss: 0.21069372, Log Avg loss: 0.36239872, Global Avg Loss: 0.93453026, Time: 0.0210 Steps: 51470, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000384, Sample Num: 6144, Cur Loss: 0.18485117, Cur Avg Loss: 0.21135252, Log Avg loss: 0.23599155, Global Avg Loss: 0.93439457, Time: 0.0210 Steps: 51480, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000394, Sample Num: 6304, Cur Loss: 0.17418715, Cur Avg Loss: 0.21042722, Log Avg loss: 0.17489583, Global Avg Loss: 0.93424707, Time: 0.0210 Steps: 51490, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000404, Sample Num: 6464, Cur Loss: 0.34723675, Cur Avg Loss: 0.21136146, Log Avg loss: 0.24817037, Global Avg Loss: 0.93411385, Time: 0.0210 Steps: 51500, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000414, Sample Num: 6624, Cur Loss: 0.30746692, Cur Avg Loss: 0.21053005, Log Avg loss: 0.17694100, Global Avg Loss: 0.93396685, Time: 0.0210 Steps: 51510, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000424, Sample Num: 6784, Cur Loss: 0.32739574, Cur Avg Loss: 0.20999546, Log Avg loss: 0.18786360, Global Avg Loss: 0.93382203, Time: 0.0210 Steps: 51520, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000434, Sample Num: 6944, Cur Loss: 0.15506116, Cur Avg Loss: 0.21001714, Log Avg loss: 0.21093634, Global Avg Loss: 0.93368175, Time: 0.0210 Steps: 51530, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000444, Sample Num: 7104, Cur Loss: 0.31711212, Cur Avg Loss: 0.20905438, Log Avg loss: 0.16727039, Global Avg Loss: 0.93353305, Time: 0.0210 Steps: 51540, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000454, Sample Num: 7264, Cur Loss: 0.16928527, Cur Avg Loss: 0.20894924, Log Avg loss: 0.20428108, Global Avg Loss: 0.93339158, Time: 0.0210 Steps: 51550, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000464, Sample Num: 7424, Cur Loss: 0.18158212, Cur Avg Loss: 0.20943197, Log Avg loss: 0.23134806, Global Avg Loss: 0.93325542, Time: 0.0210 Steps: 51560, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000474, Sample Num: 7584, Cur Loss: 0.06540501, Cur Avg Loss: 0.20834018, Log Avg loss: 0.15768097, Global Avg Loss: 0.93310503, Time: 0.0210 Steps: 51570, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000484, Sample Num: 7744, Cur Loss: 0.26858842, Cur Avg Loss: 0.20833897, Log Avg loss: 0.20828181, Global Avg Loss: 0.93296450, Time: 0.0210 Steps: 51580, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000494, Sample Num: 7904, Cur Loss: 0.18616900, Cur Avg Loss: 0.20727156, Log Avg loss: 0.15560868, Global Avg Loss: 0.93281382, Time: 0.0210 Steps: 51590, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000504, Sample Num: 8064, Cur Loss: 0.09908271, Cur Avg Loss: 0.20668717, Log Avg loss: 0.17781876, Global Avg Loss: 0.93266751, Time: 0.0210 Steps: 51600, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000514, Sample Num: 8224, Cur Loss: 0.09176275, Cur Avg Loss: 0.20590033, Log Avg loss: 0.16624339, Global Avg Loss: 0.93251901, Time: 0.0249 Steps: 51610, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000524, Sample Num: 8384, Cur Loss: 0.16712226, Cur Avg Loss: 0.20480600, Log Avg loss: 0.14855722, Global Avg Loss: 0.93236713, Time: 0.0210 Steps: 51620, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000534, Sample Num: 8544, Cur Loss: 0.67355317, Cur Avg Loss: 0.20581956, Log Avg loss: 0.25893042, Global Avg Loss: 0.93223670, Time: 0.0209 Steps: 51630, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000544, Sample Num: 8704, Cur Loss: 0.39741966, Cur Avg Loss: 0.20594329, Log Avg loss: 0.21255022, Global Avg Loss: 0.93209733, Time: 0.0209 Steps: 51640, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000554, Sample Num: 8864, Cur Loss: 0.21273470, Cur Avg Loss: 0.20582684, Log Avg loss: 0.19949222, Global Avg Loss: 0.93195549, Time: 0.0209 Steps: 51650, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000564, Sample Num: 9024, Cur Loss: 0.09145954, Cur Avg Loss: 0.20584231, Log Avg loss: 0.20669926, Global Avg Loss: 0.93181510, Time: 0.0209 Steps: 51660, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000574, Sample Num: 9184, Cur Loss: 0.19283235, Cur Avg Loss: 0.20608156, Log Avg loss: 0.21957543, Global Avg Loss: 0.93167726, Time: 0.0209 Steps: 51670, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000584, Sample Num: 9344, Cur Loss: 0.17114423, Cur Avg Loss: 0.20631969, Log Avg loss: 0.21998795, Global Avg Loss: 0.93153955, Time: 0.0209 Steps: 51680, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000594, Sample Num: 9504, Cur Loss: 0.43169248, Cur Avg Loss: 0.20655148, Log Avg loss: 0.22008826, Global Avg Loss: 0.93140191, Time: 0.0209 Steps: 51690, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000604, Sample Num: 9664, Cur Loss: 0.06131709, Cur Avg Loss: 0.20546781, Log Avg loss: 0.14109794, Global Avg Loss: 0.93124905, Time: 0.0209 Steps: 51700, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000614, Sample Num: 9824, Cur Loss: 0.23646666, Cur Avg Loss: 0.20560447, Log Avg loss: 0.21385836, Global Avg Loss: 0.93111031, Time: 0.0210 Steps: 51710, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000624, Sample Num: 9984, Cur Loss: 0.09061648, Cur Avg Loss: 0.20541665, Log Avg loss: 0.19388456, Global Avg Loss: 0.93096777, Time: 0.0209 Steps: 51720, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000634, Sample Num: 10144, Cur Loss: 0.11653678, Cur Avg Loss: 0.20459970, Log Avg loss: 0.15362199, Global Avg Loss: 0.93081750, Time: 0.0209 Steps: 51730, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000644, Sample Num: 10304, Cur Loss: 0.14435348, Cur Avg Loss: 0.20436392, Log Avg loss: 0.18941538, Global Avg Loss: 0.93067421, Time: 0.0209 Steps: 51740, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000654, Sample Num: 10464, Cur Loss: 0.16101107, Cur Avg Loss: 0.20395558, Log Avg loss: 0.17765874, Global Avg Loss: 0.93052870, Time: 0.0209 Steps: 51750, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000664, Sample Num: 10624, Cur Loss: 0.44477338, Cur Avg Loss: 0.20372369, Log Avg loss: 0.18855775, Global Avg Loss: 0.93038535, Time: 0.0209 Steps: 51760, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000674, Sample Num: 10784, Cur Loss: 0.18236648, Cur Avg Loss: 0.20329001, Log Avg loss: 0.17449425, Global Avg Loss: 0.93023934, Time: 0.0209 Steps: 51770, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000684, Sample Num: 10944, Cur Loss: 0.36953682, Cur Avg Loss: 0.20342131, Log Avg loss: 0.21227065, Global Avg Loss: 0.93010068, Time: 0.0209 Steps: 51780, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000694, Sample Num: 11104, Cur Loss: 0.16526276, Cur Avg Loss: 0.20405166, Log Avg loss: 0.24716777, Global Avg Loss: 0.92996882, Time: 0.0209 Steps: 51790, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000704, Sample Num: 11264, Cur Loss: 0.21142869, Cur Avg Loss: 0.20382635, Log Avg loss: 0.18818971, Global Avg Loss: 0.92982561, Time: 0.0209 Steps: 51800, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000714, Sample Num: 11424, Cur Loss: 0.03362991, Cur Avg Loss: 0.20312671, Log Avg loss: 0.15387184, Global Avg Loss: 0.92967585, Time: 0.0209 Steps: 51810, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000724, Sample Num: 11584, Cur Loss: 0.08700493, Cur Avg Loss: 0.20199531, Log Avg loss: 0.12121353, Global Avg Loss: 0.92951983, Time: 0.0210 Steps: 51820, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000734, Sample Num: 11744, Cur Loss: 0.35317498, Cur Avg Loss: 0.20222429, Log Avg loss: 0.21880248, Global Avg Loss: 0.92938271, Time: 0.0209 Steps: 51830, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000744, Sample Num: 11904, Cur Loss: 0.19622128, Cur Avg Loss: 0.20197478, Log Avg loss: 0.18366058, Global Avg Loss: 0.92923886, Time: 0.0209 Steps: 51840, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000754, Sample Num: 12064, Cur Loss: 0.20189282, Cur Avg Loss: 0.20156851, Log Avg loss: 0.17134185, Global Avg Loss: 0.92909269, Time: 0.0209 Steps: 51850, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000764, Sample Num: 12224, Cur Loss: 0.10802922, Cur Avg Loss: 0.20209394, Log Avg loss: 0.24171185, Global Avg Loss: 0.92896014, Time: 0.0209 Steps: 51860, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000774, Sample Num: 12384, Cur Loss: 0.16567120, Cur Avg Loss: 0.20172553, Log Avg loss: 0.17357878, Global Avg Loss: 0.92881451, Time: 0.0210 Steps: 51870, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000784, Sample Num: 12544, Cur Loss: 0.63153577, Cur Avg Loss: 0.20317004, Log Avg loss: 0.31497489, Global Avg Loss: 0.92869619, Time: 0.0210 Steps: 51880, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000794, Sample Num: 12704, Cur Loss: 0.41296655, Cur Avg Loss: 0.20320969, Log Avg loss: 0.20631889, Global Avg Loss: 0.92855698, Time: 0.0209 Steps: 51890, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000804, Sample Num: 12864, Cur Loss: 0.10903767, Cur Avg Loss: 0.20294905, Log Avg loss: 0.18225382, Global Avg Loss: 0.92841318, Time: 0.0210 Steps: 51900, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000814, Sample Num: 13024, Cur Loss: 0.49070340, Cur Avg Loss: 0.20279126, Log Avg loss: 0.19010482, Global Avg Loss: 0.92827095, Time: 0.0210 Steps: 51910, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000824, Sample Num: 13184, Cur Loss: 0.10513104, Cur Avg Loss: 0.20272188, Log Avg loss: 0.19707414, Global Avg Loss: 0.92813012, Time: 0.0210 Steps: 51920, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000834, Sample Num: 13344, Cur Loss: 0.08901622, Cur Avg Loss: 0.20232172, Log Avg loss: 0.16934873, Global Avg Loss: 0.92798401, Time: 0.0210 Steps: 51930, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000844, Sample Num: 13504, Cur Loss: 0.15004219, Cur Avg Loss: 0.20141071, Log Avg loss: 0.12543253, Global Avg Loss: 0.92782949, Time: 0.0210 Steps: 51940, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000854, Sample Num: 13664, Cur Loss: 0.21723559, Cur Avg Loss: 0.20086199, Log Avg loss: 0.15455006, Global Avg Loss: 0.92768064, Time: 0.0210 Steps: 51950, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000864, Sample Num: 13824, Cur Loss: 0.43412718, Cur Avg Loss: 0.20233098, Log Avg loss: 0.32778312, Global Avg Loss: 0.92756519, Time: 0.0210 Steps: 51960, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000874, Sample Num: 13984, Cur Loss: 0.31426686, Cur Avg Loss: 0.20525844, Log Avg loss: 0.45819079, Global Avg Loss: 0.92747487, Time: 0.0209 Steps: 51970, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000884, Sample Num: 14144, Cur Loss: 0.17967969, Cur Avg Loss: 0.20475815, Log Avg loss: 0.16103236, Global Avg Loss: 0.92732742, Time: 0.0209 Steps: 51980, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000894, Sample Num: 14304, Cur Loss: 0.32414341, Cur Avg Loss: 0.20490178, Log Avg loss: 0.21759929, Global Avg Loss: 0.92719091, Time: 0.0210 Steps: 51990, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000904, Sample Num: 14464, Cur Loss: 0.16726382, Cur Avg Loss: 0.20403978, Log Avg loss: 0.12697687, Global Avg Loss: 0.92703702, Time: 0.0210 Steps: 52000, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000914, Sample Num: 14624, Cur Loss: 0.33506632, Cur Avg Loss: 0.20421747, Log Avg loss: 0.22028052, Global Avg Loss: 0.92690113, Time: 0.0210 Steps: 52010, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000924, Sample Num: 14784, Cur Loss: 0.51853776, Cur Avg Loss: 0.20459596, Log Avg loss: 0.23919003, Global Avg Loss: 0.92676893, Time: 0.0210 Steps: 52020, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000934, Sample Num: 14944, Cur Loss: 0.14931427, Cur Avg Loss: 0.20427671, Log Avg loss: 0.17477759, Global Avg Loss: 0.92662440, Time: 0.0210 Steps: 52030, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000944, Sample Num: 15104, Cur Loss: 0.29263762, Cur Avg Loss: 0.20514910, Log Avg loss: 0.28663030, Global Avg Loss: 0.92650142, Time: 0.0210 Steps: 52040, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000954, Sample Num: 15264, Cur Loss: 0.04232628, Cur Avg Loss: 0.20696799, Log Avg loss: 0.37867121, Global Avg Loss: 0.92639617, Time: 0.0210 Steps: 52050, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000964, Sample Num: 15424, Cur Loss: 0.21201922, Cur Avg Loss: 0.20740151, Log Avg loss: 0.24875990, Global Avg Loss: 0.92626600, Time: 0.0210 Steps: 52060, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000974, Sample Num: 15584, Cur Loss: 0.30777931, Cur Avg Loss: 0.20697771, Log Avg loss: 0.16612338, Global Avg Loss: 0.92612002, Time: 0.0209 Steps: 52070, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000984, Sample Num: 15744, Cur Loss: 0.25208184, Cur Avg Loss: 0.20729932, Log Avg loss: 0.23862391, Global Avg Loss: 0.92598801, Time: 0.0210 Steps: 52080, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000994, Sample Num: 15904, Cur Loss: 0.07275249, Cur Avg Loss: 0.20697853, Log Avg loss: 0.17541241, Global Avg Loss: 0.92584392, Time: 0.0210 Steps: 52090, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001004, Sample Num: 16064, Cur Loss: 0.18338501, Cur Avg Loss: 0.20630887, Log Avg loss: 0.13974523, Global Avg Loss: 0.92569304, Time: 0.0210 Steps: 52100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001014, Sample Num: 16224, Cur Loss: 0.39065233, Cur Avg Loss: 0.20662707, Log Avg loss: 0.23857382, Global Avg Loss: 0.92556118, Time: 0.0210 Steps: 52110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001024, Sample Num: 16384, Cur Loss: 0.23448993, Cur Avg Loss: 0.20624503, Log Avg loss: 0.16750637, Global Avg Loss: 0.92541573, Time: 0.0255 Steps: 52120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001034, Sample Num: 16544, Cur Loss: 0.42933920, Cur Avg Loss: 0.20607150, Log Avg loss: 0.18830186, Global Avg Loss: 0.92527433, Time: 0.0210 Steps: 52130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001044, Sample Num: 16704, Cur Loss: 0.09917937, Cur Avg Loss: 0.20639939, Log Avg loss: 0.24030317, Global Avg Loss: 0.92514296, Time: 0.0210 Steps: 52140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001054, Sample Num: 16864, Cur Loss: 0.16228795, Cur Avg Loss: 0.20586804, Log Avg loss: 0.15039593, Global Avg Loss: 0.92499440, Time: 0.0210 Steps: 52150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001064, Sample Num: 17024, Cur Loss: 0.18985756, Cur Avg Loss: 0.20562101, Log Avg loss: 0.17958346, Global Avg Loss: 0.92485149, Time: 0.0210 Steps: 52160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001074, Sample Num: 17184, Cur Loss: 0.05159166, Cur Avg Loss: 0.20557782, Log Avg loss: 0.20098256, Global Avg Loss: 0.92471274, Time: 0.0210 Steps: 52170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001084, Sample Num: 17344, Cur Loss: 0.22377369, Cur Avg Loss: 0.20549735, Log Avg loss: 0.19685517, Global Avg Loss: 0.92457325, Time: 0.0209 Steps: 52180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001094, Sample Num: 17504, Cur Loss: 0.26436883, Cur Avg Loss: 0.20626609, Log Avg loss: 0.28959735, Global Avg Loss: 0.92445158, Time: 0.0210 Steps: 52190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001104, Sample Num: 17664, Cur Loss: 0.03387077, Cur Avg Loss: 0.20614443, Log Avg loss: 0.19283414, Global Avg Loss: 0.92431143, Time: 0.0210 Steps: 52200, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001114, Sample Num: 17824, Cur Loss: 0.13529843, Cur Avg Loss: 0.20586556, Log Avg loss: 0.17507857, Global Avg Loss: 0.92416792, Time: 0.0210 Steps: 52210, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001124, Sample Num: 17984, Cur Loss: 0.15991104, Cur Avg Loss: 0.20557383, Log Avg loss: 0.17307542, Global Avg Loss: 0.92402409, Time: 0.0210 Steps: 52220, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001134, Sample Num: 18144, Cur Loss: 0.30671000, Cur Avg Loss: 0.20532311, Log Avg loss: 0.17714220, Global Avg Loss: 0.92388109, Time: 0.0210 Steps: 52230, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001144, Sample Num: 18304, Cur Loss: 0.11617516, Cur Avg Loss: 0.20536343, Log Avg loss: 0.20993544, Global Avg Loss: 0.92374443, Time: 0.0210 Steps: 52240, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001154, Sample Num: 18464, Cur Loss: 0.60519159, Cur Avg Loss: 0.20540841, Log Avg loss: 0.21055420, Global Avg Loss: 0.92360793, Time: 0.0210 Steps: 52250, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001164, Sample Num: 18624, Cur Loss: 0.07654256, Cur Avg Loss: 0.20519254, Log Avg loss: 0.18028120, Global Avg Loss: 0.92346569, Time: 0.0210 Steps: 52260, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001174, Sample Num: 18784, Cur Loss: 0.15902719, Cur Avg Loss: 0.20493461, Log Avg loss: 0.17491176, Global Avg Loss: 0.92332249, Time: 0.0212 Steps: 52270, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001184, Sample Num: 18944, Cur Loss: 0.15911704, Cur Avg Loss: 0.20610283, Log Avg loss: 0.34325138, Global Avg Loss: 0.92321153, Time: 0.0210 Steps: 52280, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001194, Sample Num: 19104, Cur Loss: 0.13053848, Cur Avg Loss: 0.20635812, Log Avg loss: 0.23658477, Global Avg Loss: 0.92308022, Time: 0.0210 Steps: 52290, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001204, Sample Num: 19264, Cur Loss: 0.41755170, Cur Avg Loss: 0.20614460, Log Avg loss: 0.18065061, Global Avg Loss: 0.92293826, Time: 0.0210 Steps: 52300, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001214, Sample Num: 19424, Cur Loss: 0.16346765, Cur Avg Loss: 0.20559576, Log Avg loss: 0.13951481, Global Avg Loss: 0.92278850, Time: 0.0210 Steps: 52310, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001224, Sample Num: 19584, Cur Loss: 0.15887941, Cur Avg Loss: 0.20550479, Log Avg loss: 0.19446098, Global Avg Loss: 0.92264929, Time: 0.0209 Steps: 52320, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001234, Sample Num: 19744, Cur Loss: 0.04053136, Cur Avg Loss: 0.20514176, Log Avg loss: 0.16070674, Global Avg Loss: 0.92250369, Time: 0.0210 Steps: 52330, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001244, Sample Num: 19904, Cur Loss: 0.47752336, Cur Avg Loss: 0.20578868, Log Avg loss: 0.28561875, Global Avg Loss: 0.92238201, Time: 0.0210 Steps: 52340, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001254, Sample Num: 20064, Cur Loss: 0.24463688, Cur Avg Loss: 0.20553171, Log Avg loss: 0.17356492, Global Avg Loss: 0.92223897, Time: 0.0210 Steps: 52350, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001264, Sample Num: 20224, Cur Loss: 0.29818469, Cur Avg Loss: 0.20638738, Log Avg loss: 0.31368874, Global Avg Loss: 0.92212274, Time: 0.0209 Steps: 52360, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001274, Sample Num: 20384, Cur Loss: 0.74661744, Cur Avg Loss: 0.20848943, Log Avg loss: 0.47418799, Global Avg Loss: 0.92203721, Time: 0.0209 Steps: 52370, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001284, Sample Num: 20544, Cur Loss: 0.29825521, Cur Avg Loss: 0.20879537, Log Avg loss: 0.24777218, Global Avg Loss: 0.92190848, Time: 0.0246 Steps: 52380, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001294, Sample Num: 20704, Cur Loss: 0.55068904, Cur Avg Loss: 0.20982671, Log Avg loss: 0.34225038, Global Avg Loss: 0.92179784, Time: 0.0209 Steps: 52390, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001304, Sample Num: 20864, Cur Loss: 0.05257889, Cur Avg Loss: 0.21111104, Log Avg loss: 0.37730350, Global Avg Loss: 0.92169393, Time: 0.0209 Steps: 52400, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001314, Sample Num: 21024, Cur Loss: 0.32114577, Cur Avg Loss: 0.21205528, Log Avg loss: 0.33518443, Global Avg Loss: 0.92158202, Time: 0.0209 Steps: 52410, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001324, Sample Num: 21184, Cur Loss: 0.23217602, Cur Avg Loss: 0.21238554, Log Avg loss: 0.25578254, Global Avg Loss: 0.92145501, Time: 0.0209 Steps: 52420, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001334, Sample Num: 21344, Cur Loss: 0.06308993, Cur Avg Loss: 0.21251787, Log Avg loss: 0.23003715, Global Avg Loss: 0.92132313, Time: 0.0209 Steps: 52430, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001344, Sample Num: 21504, Cur Loss: 0.47965130, Cur Avg Loss: 0.21262477, Log Avg loss: 0.22688577, Global Avg Loss: 0.92119071, Time: 0.0209 Steps: 52440, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001354, Sample Num: 21664, Cur Loss: 0.12183937, Cur Avg Loss: 0.21258826, Log Avg loss: 0.20768088, Global Avg Loss: 0.92105467, Time: 0.0209 Steps: 52450, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001364, Sample Num: 21824, Cur Loss: 0.09690354, Cur Avg Loss: 0.21243998, Log Avg loss: 0.19236355, Global Avg Loss: 0.92091577, Time: 0.0209 Steps: 52460, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001374, Sample Num: 21984, Cur Loss: 0.09469937, Cur Avg Loss: 0.21243950, Log Avg loss: 0.21237339, Global Avg Loss: 0.92078073, Time: 0.0209 Steps: 52470, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001384, Sample Num: 22144, Cur Loss: 0.56624889, Cur Avg Loss: 0.21307832, Log Avg loss: 0.30085282, Global Avg Loss: 0.92066260, Time: 0.0209 Steps: 52480, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001394, Sample Num: 22304, Cur Loss: 0.33526772, Cur Avg Loss: 0.21286749, Log Avg loss: 0.18368845, Global Avg Loss: 0.92052220, Time: 0.0209 Steps: 52490, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001404, Sample Num: 22464, Cur Loss: 0.17757139, Cur Avg Loss: 0.21256968, Log Avg loss: 0.17105509, Global Avg Loss: 0.92037945, Time: 0.0209 Steps: 52500, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001414, Sample Num: 22624, Cur Loss: 0.40931928, Cur Avg Loss: 0.21284061, Log Avg loss: 0.25087901, Global Avg Loss: 0.92025195, Time: 0.0209 Steps: 52510, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001424, Sample Num: 22784, Cur Loss: 0.20768231, Cur Avg Loss: 0.21270545, Log Avg loss: 0.19359337, Global Avg Loss: 0.92011359, Time: 0.0209 Steps: 52520, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001434, Sample Num: 22944, Cur Loss: 0.08788499, Cur Avg Loss: 0.21267166, Log Avg loss: 0.20786033, Global Avg Loss: 0.91997800, Time: 0.0209 Steps: 52530, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001444, Sample Num: 23104, Cur Loss: 0.15935904, Cur Avg Loss: 0.21210624, Log Avg loss: 0.13102481, Global Avg Loss: 0.91982784, Time: 0.0209 Steps: 52540, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001454, Sample Num: 23264, Cur Loss: 0.30160558, Cur Avg Loss: 0.21221540, Log Avg loss: 0.22797891, Global Avg Loss: 0.91969618, Time: 0.0209 Steps: 52550, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001464, Sample Num: 23424, Cur Loss: 0.14788428, Cur Avg Loss: 0.21156836, Log Avg loss: 0.11748850, Global Avg Loss: 0.91954355, Time: 0.0209 Steps: 52560, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001474, Sample Num: 23584, Cur Loss: 0.23752943, Cur Avg Loss: 0.21139417, Log Avg loss: 0.18589291, Global Avg Loss: 0.91940400, Time: 0.0209 Steps: 52570, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001484, Sample Num: 23744, Cur Loss: 0.20994651, Cur Avg Loss: 0.21128983, Log Avg loss: 0.19590931, Global Avg Loss: 0.91926640, Time: 0.0209 Steps: 52580, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001494, Sample Num: 23904, Cur Loss: 0.11373606, Cur Avg Loss: 0.21123824, Log Avg loss: 0.20358317, Global Avg Loss: 0.91913031, Time: 0.0209 Steps: 52590, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001504, Sample Num: 24064, Cur Loss: 0.18771756, Cur Avg Loss: 0.21101184, Log Avg loss: 0.17718754, Global Avg Loss: 0.91898926, Time: 0.0210 Steps: 52600, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001514, Sample Num: 24224, Cur Loss: 0.03084037, Cur Avg Loss: 0.21064256, Log Avg loss: 0.15510321, Global Avg Loss: 0.91884406, Time: 0.0209 Steps: 52610, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001524, Sample Num: 24384, Cur Loss: 0.28555420, Cur Avg Loss: 0.21043458, Log Avg loss: 0.17894555, Global Avg Loss: 0.91870345, Time: 0.0209 Steps: 52620, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001534, Sample Num: 24544, Cur Loss: 0.19290651, Cur Avg Loss: 0.21047395, Log Avg loss: 0.21647373, Global Avg Loss: 0.91857002, Time: 0.0209 Steps: 52630, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001544, Sample Num: 24704, Cur Loss: 0.22864136, Cur Avg Loss: 0.21045549, Log Avg loss: 0.20762413, Global Avg Loss: 0.91843496, Time: 0.0209 Steps: 52640, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001554, Sample Num: 24864, Cur Loss: 0.07383991, Cur Avg Loss: 0.21046301, Log Avg loss: 0.21162405, Global Avg Loss: 0.91830071, Time: 0.0209 Steps: 52650, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001564, Sample Num: 25024, Cur Loss: 0.06676826, Cur Avg Loss: 0.21011339, Log Avg loss: 0.15578236, Global Avg Loss: 0.91815591, Time: 0.0209 Steps: 52660, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001574, Sample Num: 25184, Cur Loss: 0.11261186, Cur Avg Loss: 0.20978143, Log Avg loss: 0.15786369, Global Avg Loss: 0.91801156, Time: 0.0209 Steps: 52670, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001584, Sample Num: 25344, Cur Loss: 0.22166669, Cur Avg Loss: 0.20980709, Log Avg loss: 0.21384554, Global Avg Loss: 0.91787790, Time: 0.0209 Steps: 52680, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001594, Sample Num: 25504, Cur Loss: 0.05378643, Cur Avg Loss: 0.20946054, Log Avg loss: 0.15456695, Global Avg Loss: 0.91773303, Time: 0.0209 Steps: 52690, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001604, Sample Num: 25664, Cur Loss: 0.12580049, Cur Avg Loss: 0.20922567, Log Avg loss: 0.17178640, Global Avg Loss: 0.91759148, Time: 0.0209 Steps: 52700, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001614, Sample Num: 25824, Cur Loss: 0.28685704, Cur Avg Loss: 0.20921379, Log Avg loss: 0.20730957, Global Avg Loss: 0.91745673, Time: 0.0209 Steps: 52710, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001624, Sample Num: 25984, Cur Loss: 0.34313858, Cur Avg Loss: 0.20916704, Log Avg loss: 0.20162078, Global Avg Loss: 0.91732095, Time: 0.0209 Steps: 52720, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001634, Sample Num: 26144, Cur Loss: 0.14066169, Cur Avg Loss: 0.20911075, Log Avg loss: 0.19996917, Global Avg Loss: 0.91718491, Time: 0.0209 Steps: 52730, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001644, Sample Num: 26304, Cur Loss: 0.17543472, Cur Avg Loss: 0.20891076, Log Avg loss: 0.17623345, Global Avg Loss: 0.91704441, Time: 0.0209 Steps: 52740, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001654, Sample Num: 26464, Cur Loss: 0.10125054, Cur Avg Loss: 0.20846922, Log Avg loss: 0.13587965, Global Avg Loss: 0.91689633, Time: 0.0209 Steps: 52750, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001664, Sample Num: 26624, Cur Loss: 0.13749059, Cur Avg Loss: 0.20866763, Log Avg loss: 0.24148356, Global Avg Loss: 0.91676831, Time: 0.0209 Steps: 52760, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001674, Sample Num: 26784, Cur Loss: 0.43855414, Cur Avg Loss: 0.20898821, Log Avg loss: 0.26233399, Global Avg Loss: 0.91664429, Time: 0.0209 Steps: 52770, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001684, Sample Num: 26944, Cur Loss: 0.32330957, Cur Avg Loss: 0.20895668, Log Avg loss: 0.20367833, Global Avg Loss: 0.91650921, Time: 0.0209 Steps: 52780, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001694, Sample Num: 27104, Cur Loss: 0.14443743, Cur Avg Loss: 0.20848314, Log Avg loss: 0.12873883, Global Avg Loss: 0.91635998, Time: 0.0209 Steps: 52790, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001704, Sample Num: 27264, Cur Loss: 0.23164144, Cur Avg Loss: 0.20851938, Log Avg loss: 0.21465866, Global Avg Loss: 0.91622709, Time: 0.0210 Steps: 52800, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001714, Sample Num: 27424, Cur Loss: 0.48939517, Cur Avg Loss: 0.20835458, Log Avg loss: 0.18027174, Global Avg Loss: 0.91608773, Time: 0.0209 Steps: 52810, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001724, Sample Num: 27584, Cur Loss: 0.05194739, Cur Avg Loss: 0.20827818, Log Avg loss: 0.19518424, Global Avg Loss: 0.91595124, Time: 0.0210 Steps: 52820, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001734, Sample Num: 27744, Cur Loss: 0.08171827, Cur Avg Loss: 0.20831613, Log Avg loss: 0.21485879, Global Avg Loss: 0.91581854, Time: 0.0209 Steps: 52830, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001744, Sample Num: 27904, Cur Loss: 0.05805626, Cur Avg Loss: 0.20854823, Log Avg loss: 0.24879316, Global Avg Loss: 0.91569230, Time: 0.0210 Steps: 52840, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001754, Sample Num: 28064, Cur Loss: 0.05687620, Cur Avg Loss: 0.20844948, Log Avg loss: 0.19122867, Global Avg Loss: 0.91555522, Time: 0.0210 Steps: 52850, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001764, Sample Num: 28224, Cur Loss: 0.13083372, Cur Avg Loss: 0.20860650, Log Avg loss: 0.23614779, Global Avg Loss: 0.91542669, Time: 0.0209 Steps: 52860, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001774, Sample Num: 28384, Cur Loss: 0.43877578, Cur Avg Loss: 0.20882102, Log Avg loss: 0.24666185, Global Avg Loss: 0.91530020, Time: 0.0209 Steps: 52870, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001784, Sample Num: 28544, Cur Loss: 0.38362825, Cur Avg Loss: 0.20915817, Log Avg loss: 0.26896783, Global Avg Loss: 0.91517797, Time: 0.0209 Steps: 52880, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001794, Sample Num: 28704, Cur Loss: 0.45886242, Cur Avg Loss: 0.20970832, Log Avg loss: 0.30785565, Global Avg Loss: 0.91506315, Time: 0.0247 Steps: 52890, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001804, Sample Num: 28864, Cur Loss: 0.31634563, Cur Avg Loss: 0.20952959, Log Avg loss: 0.17746487, Global Avg Loss: 0.91492371, Time: 0.0209 Steps: 52900, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001814, Sample Num: 29024, Cur Loss: 0.23143417, Cur Avg Loss: 0.20948585, Log Avg loss: 0.20159552, Global Avg Loss: 0.91478889, Time: 0.0209 Steps: 52910, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001824, Sample Num: 29184, Cur Loss: 0.08133545, Cur Avg Loss: 0.20950293, Log Avg loss: 0.21260126, Global Avg Loss: 0.91465621, Time: 0.0209 Steps: 52920, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001834, Sample Num: 29344, Cur Loss: 0.38412154, Cur Avg Loss: 0.20961890, Log Avg loss: 0.23077299, Global Avg Loss: 0.91452700, Time: 0.0209 Steps: 52930, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001844, Sample Num: 29504, Cur Loss: 0.31670389, Cur Avg Loss: 0.20956446, Log Avg loss: 0.19957902, Global Avg Loss: 0.91439195, Time: 0.0209 Steps: 52940, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001854, Sample Num: 29664, Cur Loss: 0.27446425, Cur Avg Loss: 0.20923243, Log Avg loss: 0.14800581, Global Avg Loss: 0.91424721, Time: 0.0209 Steps: 52950, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001864, Sample Num: 29824, Cur Loss: 0.22052267, Cur Avg Loss: 0.20929911, Log Avg loss: 0.22166211, Global Avg Loss: 0.91411644, Time: 0.0209 Steps: 52960, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001874, Sample Num: 29984, Cur Loss: 0.07713121, Cur Avg Loss: 0.20941769, Log Avg loss: 0.23152097, Global Avg Loss: 0.91398757, Time: 0.0209 Steps: 52970, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001884, Sample Num: 30144, Cur Loss: 0.36704907, Cur Avg Loss: 0.21005008, Log Avg loss: 0.32856100, Global Avg Loss: 0.91387708, Time: 0.0209 Steps: 52980, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001894, Sample Num: 30304, Cur Loss: 0.26732242, Cur Avg Loss: 0.21006208, Log Avg loss: 0.21232252, Global Avg Loss: 0.91374468, Time: 0.0210 Steps: 52990, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001904, Sample Num: 30464, Cur Loss: 0.11396627, Cur Avg Loss: 0.20979333, Log Avg loss: 0.15889104, Global Avg Loss: 0.91360226, Time: 0.0209 Steps: 53000, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001914, Sample Num: 30624, Cur Loss: 0.03737275, Cur Avg Loss: 0.20950959, Log Avg loss: 0.15548513, Global Avg Loss: 0.91345924, Time: 0.0210 Steps: 53010, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001924, Sample Num: 30784, Cur Loss: 0.23819205, Cur Avg Loss: 0.20952044, Log Avg loss: 0.21159756, Global Avg Loss: 0.91332687, Time: 0.0209 Steps: 53020, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001934, Sample Num: 30944, Cur Loss: 0.19401008, Cur Avg Loss: 0.20980488, Log Avg loss: 0.26453081, Global Avg Loss: 0.91320452, Time: 0.0209 Steps: 53030, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001944, Sample Num: 31104, Cur Loss: 0.33452022, Cur Avg Loss: 0.20992142, Log Avg loss: 0.23246057, Global Avg Loss: 0.91307617, Time: 0.0211 Steps: 53040, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001954, Sample Num: 31264, Cur Loss: 0.45010850, Cur Avg Loss: 0.20987758, Log Avg loss: 0.20135624, Global Avg Loss: 0.91294201, Time: 0.0209 Steps: 53050, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001964, Sample Num: 31424, Cur Loss: 0.12600154, Cur Avg Loss: 0.20980899, Log Avg loss: 0.19640653, Global Avg Loss: 0.91280697, Time: 0.0210 Steps: 53060, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001974, Sample Num: 31584, Cur Loss: 0.09816553, Cur Avg Loss: 0.20972812, Log Avg loss: 0.19384535, Global Avg Loss: 0.91267150, Time: 0.0210 Steps: 53070, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001984, Sample Num: 31744, Cur Loss: 0.12720703, Cur Avg Loss: 0.20951456, Log Avg loss: 0.16735696, Global Avg Loss: 0.91253108, Time: 0.0209 Steps: 53080, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001994, Sample Num: 31904, Cur Loss: 0.12120205, Cur Avg Loss: 0.20911645, Log Avg loss: 0.13013172, Global Avg Loss: 0.91238371, Time: 0.0210 Steps: 53090, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002004, Sample Num: 32064, Cur Loss: 0.23953612, Cur Avg Loss: 0.20892004, Log Avg loss: 0.16975470, Global Avg Loss: 0.91224386, Time: 0.0210 Steps: 53100, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002014, Sample Num: 32224, Cur Loss: 0.32821894, Cur Avg Loss: 0.20885088, Log Avg loss: 0.19499126, Global Avg Loss: 0.91210881, Time: 0.0209 Steps: 53110, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002024, Sample Num: 32384, Cur Loss: 0.08077554, Cur Avg Loss: 0.20869162, Log Avg loss: 0.17661700, Global Avg Loss: 0.91197035, Time: 0.0209 Steps: 53120, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002034, Sample Num: 32544, Cur Loss: 0.65112543, Cur Avg Loss: 0.20918818, Log Avg loss: 0.30969304, Global Avg Loss: 0.91185699, Time: 0.0209 Steps: 53130, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002044, Sample Num: 32704, Cur Loss: 0.13371952, Cur Avg Loss: 0.20916112, Log Avg loss: 0.20365606, Global Avg Loss: 0.91172372, Time: 0.0209 Steps: 53140, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002054, Sample Num: 32864, Cur Loss: 0.04957981, Cur Avg Loss: 0.20901686, Log Avg loss: 0.17953133, Global Avg Loss: 0.91158596, Time: 0.0213 Steps: 53150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002064, Sample Num: 33024, Cur Loss: 0.12895440, Cur Avg Loss: 0.20877854, Log Avg loss: 0.15982764, Global Avg Loss: 0.91144454, Time: 0.0210 Steps: 53160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002074, Sample Num: 33184, Cur Loss: 0.16414982, Cur Avg Loss: 0.20870128, Log Avg loss: 0.19275507, Global Avg Loss: 0.91130938, Time: 0.0209 Steps: 53170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002084, Sample Num: 33344, Cur Loss: 0.09465042, Cur Avg Loss: 0.20869502, Log Avg loss: 0.20739550, Global Avg Loss: 0.91117701, Time: 0.0211 Steps: 53180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002094, Sample Num: 33504, Cur Loss: 0.12152143, Cur Avg Loss: 0.20941404, Log Avg loss: 0.35925817, Global Avg Loss: 0.91107325, Time: 0.0209 Steps: 53190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002104, Sample Num: 33664, Cur Loss: 0.05163526, Cur Avg Loss: 0.20948138, Log Avg loss: 0.22358279, Global Avg Loss: 0.91094402, Time: 0.0210 Steps: 53200, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002114, Sample Num: 33824, Cur Loss: 0.17489789, Cur Avg Loss: 0.20929862, Log Avg loss: 0.17084510, Global Avg Loss: 0.91080493, Time: 0.0209 Steps: 53210, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002124, Sample Num: 33984, Cur Loss: 0.05692894, Cur Avg Loss: 0.20916510, Log Avg loss: 0.18093828, Global Avg Loss: 0.91066779, Time: 0.0210 Steps: 53220, Updated lr: 0.000050 ***** Running evaluation checkpoint-53225 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-53225 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.785652, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.229196, "eval_total_loss": 161.125034, "eval_mae": 0.334104, "eval_mse": 0.229264, "eval_r2": 0.854264, "eval_sp_statistic": 0.912926, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.926638, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.182454, "test_total_loss": 91.591929, "test_mae": 0.29812, "test_mse": 0.18252, "test_r2": 0.8822, "test_sp_statistic": 0.908796, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.944314, "test_ps_pvalue": 0.0, "lr": 5.0474158368895205e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.9105922554736758, "train_cur_epoch_loss": 444.7997155338526, "train_cur_epoch_avg_loss": 0.2089242440271736, "train_cur_epoch_time": 44.78565216064453, "train_cur_epoch_avg_time": 0.021036003833088086, "epoch": 25, "step": 53225} ################################################## Training, Epoch: 0026, Batch: 000005, Sample Num: 80, Cur Loss: 0.26084083, Cur Avg Loss: 0.21875919, Log Avg loss: 0.16268482, Global Avg Loss: 0.91052727, Time: 0.0248 Steps: 53230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000015, Sample Num: 240, Cur Loss: 0.04835930, Cur Avg Loss: 0.17625841, Log Avg loss: 0.15500802, Global Avg Loss: 0.91038536, Time: 0.0210 Steps: 53240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000025, Sample Num: 400, Cur Loss: 0.07845268, Cur Avg Loss: 0.20090550, Log Avg loss: 0.23787614, Global Avg Loss: 0.91025907, Time: 0.0210 Steps: 53250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000035, Sample Num: 560, Cur Loss: 0.12715435, Cur Avg Loss: 0.18749876, Log Avg loss: 0.15398191, Global Avg Loss: 0.91011707, Time: 0.0211 Steps: 53260, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000045, Sample Num: 720, Cur Loss: 0.18328133, Cur Avg Loss: 0.18224672, Log Avg loss: 0.16386457, Global Avg Loss: 0.90997698, Time: 0.0210 Steps: 53270, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000055, Sample Num: 880, Cur Loss: 0.36470708, Cur Avg Loss: 0.20083384, Log Avg loss: 0.28447587, Global Avg Loss: 0.90985958, Time: 0.0211 Steps: 53280, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000065, Sample Num: 1040, Cur Loss: 0.15142198, Cur Avg Loss: 0.20438951, Log Avg loss: 0.22394568, Global Avg Loss: 0.90973087, Time: 0.0211 Steps: 53290, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000075, Sample Num: 1200, Cur Loss: 0.13225684, Cur Avg Loss: 0.19894317, Log Avg loss: 0.16354200, Global Avg Loss: 0.90959087, Time: 0.0210 Steps: 53300, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000085, Sample Num: 1360, Cur Loss: 0.36500469, Cur Avg Loss: 0.20250645, Log Avg loss: 0.22923106, Global Avg Loss: 0.90946325, Time: 0.0209 Steps: 53310, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000095, Sample Num: 1520, Cur Loss: 0.09893924, Cur Avg Loss: 0.19957575, Log Avg loss: 0.17466479, Global Avg Loss: 0.90932544, Time: 0.0210 Steps: 53320, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000105, Sample Num: 1680, Cur Loss: 0.10105537, Cur Avg Loss: 0.19325911, Log Avg loss: 0.13325103, Global Avg Loss: 0.90917992, Time: 0.0210 Steps: 53330, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000115, Sample Num: 1840, Cur Loss: 0.15109390, Cur Avg Loss: 0.19052347, Log Avg loss: 0.16179925, Global Avg Loss: 0.90903980, Time: 0.0210 Steps: 53340, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000125, Sample Num: 2000, Cur Loss: 0.12826280, Cur Avg Loss: 0.18470246, Log Avg loss: 0.11776079, Global Avg Loss: 0.90889148, Time: 0.0209 Steps: 53350, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000135, Sample Num: 2160, Cur Loss: 0.08271597, Cur Avg Loss: 0.18401700, Log Avg loss: 0.17544879, Global Avg Loss: 0.90875403, Time: 0.0210 Steps: 53360, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000145, Sample Num: 2320, Cur Loss: 0.29259780, Cur Avg Loss: 0.18185202, Log Avg loss: 0.15262476, Global Avg Loss: 0.90861235, Time: 0.0210 Steps: 53370, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000155, Sample Num: 2480, Cur Loss: 0.11577445, Cur Avg Loss: 0.18343016, Log Avg loss: 0.20631327, Global Avg Loss: 0.90848079, Time: 0.0210 Steps: 53380, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000165, Sample Num: 2640, Cur Loss: 0.30962935, Cur Avg Loss: 0.19191785, Log Avg loss: 0.32347693, Global Avg Loss: 0.90837122, Time: 0.0210 Steps: 53390, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000175, Sample Num: 2800, Cur Loss: 0.04864944, Cur Avg Loss: 0.18652418, Log Avg loss: 0.09752865, Global Avg Loss: 0.90821937, Time: 0.0209 Steps: 53400, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000185, Sample Num: 2960, Cur Loss: 0.30084300, Cur Avg Loss: 0.18813398, Log Avg loss: 0.21630547, Global Avg Loss: 0.90808983, Time: 0.0209 Steps: 53410, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000195, Sample Num: 3120, Cur Loss: 0.18130547, Cur Avg Loss: 0.18891337, Log Avg loss: 0.20333208, Global Avg Loss: 0.90795790, Time: 0.0210 Steps: 53420, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000205, Sample Num: 3280, Cur Loss: 0.04781031, Cur Avg Loss: 0.18589971, Log Avg loss: 0.12713344, Global Avg Loss: 0.90781176, Time: 0.0209 Steps: 53430, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000215, Sample Num: 3440, Cur Loss: 0.11145760, Cur Avg Loss: 0.18507934, Log Avg loss: 0.16826170, Global Avg Loss: 0.90767337, Time: 0.0209 Steps: 53440, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000225, Sample Num: 3600, Cur Loss: 0.45120305, Cur Avg Loss: 0.18532005, Log Avg loss: 0.19049527, Global Avg Loss: 0.90753919, Time: 0.0209 Steps: 53450, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000235, Sample Num: 3760, Cur Loss: 0.16713683, Cur Avg Loss: 0.18516292, Log Avg loss: 0.18162764, Global Avg Loss: 0.90740341, Time: 0.0210 Steps: 53460, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000245, Sample Num: 3920, Cur Loss: 0.19277951, Cur Avg Loss: 0.18552420, Log Avg loss: 0.19401421, Global Avg Loss: 0.90726999, Time: 0.0210 Steps: 53470, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000255, Sample Num: 4080, Cur Loss: 0.27959496, Cur Avg Loss: 0.18484537, Log Avg loss: 0.16821409, Global Avg Loss: 0.90713179, Time: 0.0209 Steps: 53480, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000265, Sample Num: 4240, Cur Loss: 0.21923587, Cur Avg Loss: 0.18359273, Log Avg loss: 0.15165033, Global Avg Loss: 0.90699056, Time: 0.0210 Steps: 53490, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000275, Sample Num: 4400, Cur Loss: 0.08970156, Cur Avg Loss: 0.18164354, Log Avg loss: 0.12998998, Global Avg Loss: 0.90684532, Time: 0.0210 Steps: 53500, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000285, Sample Num: 4560, Cur Loss: 0.46977621, Cur Avg Loss: 0.18257265, Log Avg loss: 0.20812308, Global Avg Loss: 0.90671474, Time: 0.0210 Steps: 53510, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000295, Sample Num: 4720, Cur Loss: 0.13315307, Cur Avg Loss: 0.18473476, Log Avg loss: 0.24635503, Global Avg Loss: 0.90659136, Time: 0.0209 Steps: 53520, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000305, Sample Num: 4880, Cur Loss: 0.09543063, Cur Avg Loss: 0.18472300, Log Avg loss: 0.18437620, Global Avg Loss: 0.90645644, Time: 0.0210 Steps: 53530, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000315, Sample Num: 5040, Cur Loss: 0.37700632, Cur Avg Loss: 0.18725780, Log Avg loss: 0.26456920, Global Avg Loss: 0.90633655, Time: 0.0210 Steps: 53540, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000325, Sample Num: 5200, Cur Loss: 0.17028368, Cur Avg Loss: 0.18607465, Log Avg loss: 0.14880536, Global Avg Loss: 0.90619509, Time: 0.0210 Steps: 53550, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000335, Sample Num: 5360, Cur Loss: 0.11941547, Cur Avg Loss: 0.18643122, Log Avg loss: 0.19801964, Global Avg Loss: 0.90606287, Time: 0.0211 Steps: 53560, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000345, Sample Num: 5520, Cur Loss: 0.27301189, Cur Avg Loss: 0.18568697, Log Avg loss: 0.16075456, Global Avg Loss: 0.90592374, Time: 0.0210 Steps: 53570, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000355, Sample Num: 5680, Cur Loss: 0.15881385, Cur Avg Loss: 0.18419431, Log Avg loss: 0.13269775, Global Avg Loss: 0.90577943, Time: 0.0210 Steps: 53580, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000365, Sample Num: 5840, Cur Loss: 0.08974978, Cur Avg Loss: 0.18476957, Log Avg loss: 0.20519119, Global Avg Loss: 0.90564870, Time: 0.0210 Steps: 53590, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000375, Sample Num: 6000, Cur Loss: 0.32422137, Cur Avg Loss: 0.18483645, Log Avg loss: 0.18727762, Global Avg Loss: 0.90551467, Time: 0.0210 Steps: 53600, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000385, Sample Num: 6160, Cur Loss: 0.11432591, Cur Avg Loss: 0.18431640, Log Avg loss: 0.16481452, Global Avg Loss: 0.90537651, Time: 0.0209 Steps: 53610, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000395, Sample Num: 6320, Cur Loss: 0.13598724, Cur Avg Loss: 0.18380253, Log Avg loss: 0.16401843, Global Avg Loss: 0.90523825, Time: 0.0211 Steps: 53620, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000405, Sample Num: 6480, Cur Loss: 0.30797639, Cur Avg Loss: 0.18303242, Log Avg loss: 0.15261300, Global Avg Loss: 0.90509791, Time: 0.0210 Steps: 53630, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000415, Sample Num: 6640, Cur Loss: 0.45891559, Cur Avg Loss: 0.18412596, Log Avg loss: 0.22841440, Global Avg Loss: 0.90497176, Time: 0.0210 Steps: 53640, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000425, Sample Num: 6800, Cur Loss: 0.23164821, Cur Avg Loss: 0.18608253, Log Avg loss: 0.26728035, Global Avg Loss: 0.90485290, Time: 0.0210 Steps: 53650, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000435, Sample Num: 6960, Cur Loss: 0.37902859, Cur Avg Loss: 0.18744369, Log Avg loss: 0.24529269, Global Avg Loss: 0.90472998, Time: 0.0210 Steps: 53660, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000445, Sample Num: 7120, Cur Loss: 0.06561579, Cur Avg Loss: 0.18740828, Log Avg loss: 0.18586810, Global Avg Loss: 0.90459604, Time: 0.0210 Steps: 53670, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000455, Sample Num: 7280, Cur Loss: 0.16323270, Cur Avg Loss: 0.18694906, Log Avg loss: 0.16651392, Global Avg Loss: 0.90445854, Time: 0.0210 Steps: 53680, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000465, Sample Num: 7440, Cur Loss: 0.10206905, Cur Avg Loss: 0.18737389, Log Avg loss: 0.20670355, Global Avg Loss: 0.90432858, Time: 0.0210 Steps: 53690, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000475, Sample Num: 7600, Cur Loss: 0.31296945, Cur Avg Loss: 0.18757271, Log Avg loss: 0.19681785, Global Avg Loss: 0.90419683, Time: 0.0210 Steps: 53700, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000485, Sample Num: 7760, Cur Loss: 0.18724167, Cur Avg Loss: 0.19111818, Log Avg loss: 0.35952792, Global Avg Loss: 0.90409542, Time: 0.0210 Steps: 53710, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000495, Sample Num: 7920, Cur Loss: 0.12504464, Cur Avg Loss: 0.19141752, Log Avg loss: 0.20593533, Global Avg Loss: 0.90396546, Time: 0.0210 Steps: 53720, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000505, Sample Num: 8080, Cur Loss: 0.21918875, Cur Avg Loss: 0.19136236, Log Avg loss: 0.18863219, Global Avg Loss: 0.90383232, Time: 0.0210 Steps: 53730, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000515, Sample Num: 8240, Cur Loss: 0.23327631, Cur Avg Loss: 0.19140369, Log Avg loss: 0.19349097, Global Avg Loss: 0.90370014, Time: 0.0247 Steps: 53740, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000525, Sample Num: 8400, Cur Loss: 0.37738076, Cur Avg Loss: 0.19092294, Log Avg loss: 0.16616437, Global Avg Loss: 0.90356293, Time: 0.0210 Steps: 53750, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000535, Sample Num: 8560, Cur Loss: 0.29539114, Cur Avg Loss: 0.19096394, Log Avg loss: 0.19311639, Global Avg Loss: 0.90343078, Time: 0.0210 Steps: 53760, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000545, Sample Num: 8720, Cur Loss: 0.06902187, Cur Avg Loss: 0.19011146, Log Avg loss: 0.14450364, Global Avg Loss: 0.90328963, Time: 0.0210 Steps: 53770, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000555, Sample Num: 8880, Cur Loss: 0.36949396, Cur Avg Loss: 0.19014146, Log Avg loss: 0.19177628, Global Avg Loss: 0.90315733, Time: 0.0210 Steps: 53780, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000565, Sample Num: 9040, Cur Loss: 0.12538463, Cur Avg Loss: 0.18930245, Log Avg loss: 0.14273734, Global Avg Loss: 0.90301596, Time: 0.0210 Steps: 53790, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000575, Sample Num: 9200, Cur Loss: 0.49702245, Cur Avg Loss: 0.19009836, Log Avg loss: 0.23506746, Global Avg Loss: 0.90289181, Time: 0.0210 Steps: 53800, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000585, Sample Num: 9360, Cur Loss: 0.13391593, Cur Avg Loss: 0.19010100, Log Avg loss: 0.19025311, Global Avg Loss: 0.90275937, Time: 0.0210 Steps: 53810, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000595, Sample Num: 9520, Cur Loss: 0.24807644, Cur Avg Loss: 0.19073253, Log Avg loss: 0.22767667, Global Avg Loss: 0.90263394, Time: 0.0209 Steps: 53820, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000605, Sample Num: 9680, Cur Loss: 0.11801523, Cur Avg Loss: 0.19045591, Log Avg loss: 0.17399738, Global Avg Loss: 0.90249858, Time: 0.0209 Steps: 53830, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000615, Sample Num: 9840, Cur Loss: 0.04301849, Cur Avg Loss: 0.19076380, Log Avg loss: 0.20939098, Global Avg Loss: 0.90236985, Time: 0.0210 Steps: 53840, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000625, Sample Num: 10000, Cur Loss: 0.35485199, Cur Avg Loss: 0.19115213, Log Avg loss: 0.21503449, Global Avg Loss: 0.90224221, Time: 0.0211 Steps: 53850, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000635, Sample Num: 10160, Cur Loss: 0.30636501, Cur Avg Loss: 0.19101446, Log Avg loss: 0.18240985, Global Avg Loss: 0.90210856, Time: 0.0210 Steps: 53860, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000645, Sample Num: 10320, Cur Loss: 0.09158869, Cur Avg Loss: 0.19055972, Log Avg loss: 0.16168361, Global Avg Loss: 0.90197111, Time: 0.0209 Steps: 53870, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000655, Sample Num: 10480, Cur Loss: 0.05437023, Cur Avg Loss: 0.19074136, Log Avg loss: 0.20245753, Global Avg Loss: 0.90184128, Time: 0.0210 Steps: 53880, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000665, Sample Num: 10640, Cur Loss: 0.25024056, Cur Avg Loss: 0.19088408, Log Avg loss: 0.20023188, Global Avg Loss: 0.90171109, Time: 0.0210 Steps: 53890, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000675, Sample Num: 10800, Cur Loss: 0.35395569, Cur Avg Loss: 0.19036865, Log Avg loss: 0.15609255, Global Avg Loss: 0.90157276, Time: 0.0210 Steps: 53900, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000685, Sample Num: 10960, Cur Loss: 0.07796171, Cur Avg Loss: 0.18940479, Log Avg loss: 0.12434424, Global Avg Loss: 0.90142859, Time: 0.0210 Steps: 53910, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000695, Sample Num: 11120, Cur Loss: 0.24064286, Cur Avg Loss: 0.19032994, Log Avg loss: 0.25370308, Global Avg Loss: 0.90130846, Time: 0.0210 Steps: 53920, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000705, Sample Num: 11280, Cur Loss: 0.11268567, Cur Avg Loss: 0.19032134, Log Avg loss: 0.18972346, Global Avg Loss: 0.90117651, Time: 0.0209 Steps: 53930, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000715, Sample Num: 11440, Cur Loss: 0.09635635, Cur Avg Loss: 0.18978808, Log Avg loss: 0.15219301, Global Avg Loss: 0.90103766, Time: 0.0210 Steps: 53940, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000725, Sample Num: 11600, Cur Loss: 0.08002681, Cur Avg Loss: 0.18910967, Log Avg loss: 0.14060362, Global Avg Loss: 0.90089671, Time: 0.0210 Steps: 53950, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000735, Sample Num: 11760, Cur Loss: 0.13594952, Cur Avg Loss: 0.18975022, Log Avg loss: 0.23618968, Global Avg Loss: 0.90077352, Time: 0.0209 Steps: 53960, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000745, Sample Num: 11920, Cur Loss: 0.36381507, Cur Avg Loss: 0.18893428, Log Avg loss: 0.12896342, Global Avg Loss: 0.90063051, Time: 0.0210 Steps: 53970, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000755, Sample Num: 12080, Cur Loss: 0.27450058, Cur Avg Loss: 0.18921497, Log Avg loss: 0.21012637, Global Avg Loss: 0.90050260, Time: 0.0209 Steps: 53980, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000765, Sample Num: 12240, Cur Loss: 0.11587356, Cur Avg Loss: 0.18841089, Log Avg loss: 0.12770228, Global Avg Loss: 0.90035946, Time: 0.0210 Steps: 53990, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000775, Sample Num: 12400, Cur Loss: 0.03556336, Cur Avg Loss: 0.18777024, Log Avg loss: 0.13876047, Global Avg Loss: 0.90021842, Time: 0.0210 Steps: 54000, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000785, Sample Num: 12560, Cur Loss: 0.11304181, Cur Avg Loss: 0.18745520, Log Avg loss: 0.16303958, Global Avg Loss: 0.90008193, Time: 0.0209 Steps: 54010, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000795, Sample Num: 12720, Cur Loss: 0.13235614, Cur Avg Loss: 0.18772181, Log Avg loss: 0.20865116, Global Avg Loss: 0.89995394, Time: 0.0209 Steps: 54020, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000805, Sample Num: 12880, Cur Loss: 0.07536016, Cur Avg Loss: 0.18733277, Log Avg loss: 0.15640361, Global Avg Loss: 0.89981632, Time: 0.0209 Steps: 54030, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000815, Sample Num: 13040, Cur Loss: 0.14775267, Cur Avg Loss: 0.18852325, Log Avg loss: 0.28435687, Global Avg Loss: 0.89970243, Time: 0.0209 Steps: 54040, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000825, Sample Num: 13200, Cur Loss: 0.03586665, Cur Avg Loss: 0.18842436, Log Avg loss: 0.18036489, Global Avg Loss: 0.89956934, Time: 0.0209 Steps: 54050, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000835, Sample Num: 13360, Cur Loss: 0.24529481, Cur Avg Loss: 0.18860482, Log Avg loss: 0.20349331, Global Avg Loss: 0.89944058, Time: 0.0209 Steps: 54060, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000845, Sample Num: 13520, Cur Loss: 0.17373294, Cur Avg Loss: 0.18822274, Log Avg loss: 0.15631894, Global Avg Loss: 0.89930314, Time: 0.0209 Steps: 54070, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000855, Sample Num: 13680, Cur Loss: 0.03080080, Cur Avg Loss: 0.18814223, Log Avg loss: 0.18133927, Global Avg Loss: 0.89917038, Time: 0.0209 Steps: 54080, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000865, Sample Num: 13840, Cur Loss: 0.21727289, Cur Avg Loss: 0.18798316, Log Avg loss: 0.17438203, Global Avg Loss: 0.89903639, Time: 0.0209 Steps: 54090, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000875, Sample Num: 14000, Cur Loss: 0.14014809, Cur Avg Loss: 0.18798890, Log Avg loss: 0.18848570, Global Avg Loss: 0.89890505, Time: 0.0209 Steps: 54100, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000885, Sample Num: 14160, Cur Loss: 0.13021713, Cur Avg Loss: 0.18736185, Log Avg loss: 0.13249520, Global Avg Loss: 0.89876341, Time: 0.0209 Steps: 54110, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000895, Sample Num: 14320, Cur Loss: 0.27513409, Cur Avg Loss: 0.18682158, Log Avg loss: 0.13900738, Global Avg Loss: 0.89862302, Time: 0.0209 Steps: 54120, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000905, Sample Num: 14480, Cur Loss: 0.26155087, Cur Avg Loss: 0.18695372, Log Avg loss: 0.19878024, Global Avg Loss: 0.89849374, Time: 0.0209 Steps: 54130, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000915, Sample Num: 14640, Cur Loss: 0.10733357, Cur Avg Loss: 0.18691823, Log Avg loss: 0.18370630, Global Avg Loss: 0.89836171, Time: 0.0209 Steps: 54140, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000925, Sample Num: 14800, Cur Loss: 0.13298388, Cur Avg Loss: 0.18696854, Log Avg loss: 0.19157227, Global Avg Loss: 0.89823119, Time: 0.0208 Steps: 54150, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000935, Sample Num: 14960, Cur Loss: 0.09171458, Cur Avg Loss: 0.18746659, Log Avg loss: 0.23353648, Global Avg Loss: 0.89810846, Time: 0.0209 Steps: 54160, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000945, Sample Num: 15120, Cur Loss: 0.09177762, Cur Avg Loss: 0.18783564, Log Avg loss: 0.22234154, Global Avg Loss: 0.89798371, Time: 0.0209 Steps: 54170, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000955, Sample Num: 15280, Cur Loss: 0.21868162, Cur Avg Loss: 0.18732034, Log Avg loss: 0.13862473, Global Avg Loss: 0.89784355, Time: 0.0209 Steps: 54180, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000965, Sample Num: 15440, Cur Loss: 0.24652293, Cur Avg Loss: 0.18681054, Log Avg loss: 0.13812447, Global Avg Loss: 0.89770336, Time: 0.0208 Steps: 54190, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000975, Sample Num: 15600, Cur Loss: 0.08215863, Cur Avg Loss: 0.18636471, Log Avg loss: 0.14334213, Global Avg Loss: 0.89756418, Time: 0.0209 Steps: 54200, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000985, Sample Num: 15760, Cur Loss: 0.26777548, Cur Avg Loss: 0.18653658, Log Avg loss: 0.20329335, Global Avg Loss: 0.89743611, Time: 0.0209 Steps: 54210, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000995, Sample Num: 15920, Cur Loss: 0.25233752, Cur Avg Loss: 0.18644531, Log Avg loss: 0.17745519, Global Avg Loss: 0.89730332, Time: 0.0209 Steps: 54220, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001005, Sample Num: 16080, Cur Loss: 0.46060988, Cur Avg Loss: 0.18723714, Log Avg loss: 0.26602478, Global Avg Loss: 0.89718691, Time: 0.0209 Steps: 54230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001015, Sample Num: 16240, Cur Loss: 0.21099173, Cur Avg Loss: 0.18796206, Log Avg loss: 0.26081598, Global Avg Loss: 0.89706958, Time: 0.0209 Steps: 54240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001025, Sample Num: 16400, Cur Loss: 0.28857470, Cur Avg Loss: 0.18822082, Log Avg loss: 0.21448574, Global Avg Loss: 0.89694376, Time: 0.0247 Steps: 54250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001035, Sample Num: 16560, Cur Loss: 0.40629193, Cur Avg Loss: 0.18841677, Log Avg loss: 0.20850165, Global Avg Loss: 0.89681688, Time: 0.0209 Steps: 54260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001045, Sample Num: 16720, Cur Loss: 0.55356228, Cur Avg Loss: 0.18894152, Log Avg loss: 0.24325303, Global Avg Loss: 0.89669646, Time: 0.0210 Steps: 54270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001055, Sample Num: 16880, Cur Loss: 0.16057271, Cur Avg Loss: 0.18948329, Log Avg loss: 0.24609837, Global Avg Loss: 0.89657660, Time: 0.0210 Steps: 54280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001065, Sample Num: 17040, Cur Loss: 0.22622472, Cur Avg Loss: 0.18953984, Log Avg loss: 0.19550588, Global Avg Loss: 0.89644746, Time: 0.0210 Steps: 54290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001075, Sample Num: 17200, Cur Loss: 0.06571258, Cur Avg Loss: 0.18920324, Log Avg loss: 0.15335527, Global Avg Loss: 0.89631061, Time: 0.0210 Steps: 54300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001085, Sample Num: 17360, Cur Loss: 0.19853032, Cur Avg Loss: 0.18965714, Log Avg loss: 0.23845123, Global Avg Loss: 0.89618948, Time: 0.0210 Steps: 54310, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001095, Sample Num: 17520, Cur Loss: 0.10285498, Cur Avg Loss: 0.18920693, Log Avg loss: 0.14035880, Global Avg Loss: 0.89605034, Time: 0.0210 Steps: 54320, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001105, Sample Num: 17680, Cur Loss: 0.44837761, Cur Avg Loss: 0.18972294, Log Avg loss: 0.24622634, Global Avg Loss: 0.89593073, Time: 0.0210 Steps: 54330, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001115, Sample Num: 17840, Cur Loss: 0.41383010, Cur Avg Loss: 0.18960398, Log Avg loss: 0.17645901, Global Avg Loss: 0.89579833, Time: 0.0210 Steps: 54340, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001125, Sample Num: 18000, Cur Loss: 0.18775809, Cur Avg Loss: 0.18915465, Log Avg loss: 0.13905396, Global Avg Loss: 0.89565909, Time: 0.0210 Steps: 54350, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001135, Sample Num: 18160, Cur Loss: 0.09593906, Cur Avg Loss: 0.18885982, Log Avg loss: 0.15569129, Global Avg Loss: 0.89552297, Time: 0.0210 Steps: 54360, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001145, Sample Num: 18320, Cur Loss: 0.04651801, Cur Avg Loss: 0.18868336, Log Avg loss: 0.16865563, Global Avg Loss: 0.89538928, Time: 0.0211 Steps: 54370, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001155, Sample Num: 18480, Cur Loss: 0.16159977, Cur Avg Loss: 0.18874308, Log Avg loss: 0.19558065, Global Avg Loss: 0.89526059, Time: 0.0210 Steps: 54380, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001165, Sample Num: 18640, Cur Loss: 0.17091329, Cur Avg Loss: 0.18931266, Log Avg loss: 0.25509928, Global Avg Loss: 0.89514289, Time: 0.0210 Steps: 54390, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001175, Sample Num: 18800, Cur Loss: 0.27013671, Cur Avg Loss: 0.18997641, Log Avg loss: 0.26730272, Global Avg Loss: 0.89502748, Time: 0.0210 Steps: 54400, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001185, Sample Num: 18960, Cur Loss: 0.07051300, Cur Avg Loss: 0.18978145, Log Avg loss: 0.16687389, Global Avg Loss: 0.89489366, Time: 0.0211 Steps: 54410, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001195, Sample Num: 19120, Cur Loss: 0.11472966, Cur Avg Loss: 0.18962707, Log Avg loss: 0.17133300, Global Avg Loss: 0.89476070, Time: 0.0210 Steps: 54420, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001205, Sample Num: 19280, Cur Loss: 0.10059386, Cur Avg Loss: 0.18924372, Log Avg loss: 0.14343332, Global Avg Loss: 0.89462266, Time: 0.0210 Steps: 54430, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001215, Sample Num: 19440, Cur Loss: 0.05298507, Cur Avg Loss: 0.18909410, Log Avg loss: 0.17106549, Global Avg Loss: 0.89448975, Time: 0.0210 Steps: 54440, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001225, Sample Num: 19600, Cur Loss: 0.14104207, Cur Avg Loss: 0.18892056, Log Avg loss: 0.16783575, Global Avg Loss: 0.89435630, Time: 0.0210 Steps: 54450, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001235, Sample Num: 19760, Cur Loss: 0.13658808, Cur Avg Loss: 0.18860174, Log Avg loss: 0.14954617, Global Avg Loss: 0.89421954, Time: 0.0210 Steps: 54460, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001245, Sample Num: 19920, Cur Loss: 0.65522760, Cur Avg Loss: 0.18935096, Log Avg loss: 0.28187939, Global Avg Loss: 0.89410712, Time: 0.0210 Steps: 54470, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001255, Sample Num: 20080, Cur Loss: 0.17988312, Cur Avg Loss: 0.18939399, Log Avg loss: 0.19475157, Global Avg Loss: 0.89397875, Time: 0.0211 Steps: 54480, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001265, Sample Num: 20240, Cur Loss: 0.17991698, Cur Avg Loss: 0.18950492, Log Avg loss: 0.20342649, Global Avg Loss: 0.89385202, Time: 0.0210 Steps: 54490, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001275, Sample Num: 20400, Cur Loss: 0.06604239, Cur Avg Loss: 0.18938504, Log Avg loss: 0.17421938, Global Avg Loss: 0.89371998, Time: 0.0210 Steps: 54500, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001285, Sample Num: 20560, Cur Loss: 0.10763925, Cur Avg Loss: 0.18965211, Log Avg loss: 0.22370380, Global Avg Loss: 0.89359706, Time: 0.0210 Steps: 54510, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001295, Sample Num: 20720, Cur Loss: 0.11504904, Cur Avg Loss: 0.18989696, Log Avg loss: 0.22136109, Global Avg Loss: 0.89347376, Time: 0.0209 Steps: 54520, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001305, Sample Num: 20880, Cur Loss: 0.14815433, Cur Avg Loss: 0.18996306, Log Avg loss: 0.19852261, Global Avg Loss: 0.89334632, Time: 0.0209 Steps: 54530, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001315, Sample Num: 21040, Cur Loss: 0.03623566, Cur Avg Loss: 0.19017337, Log Avg loss: 0.21761909, Global Avg Loss: 0.89322242, Time: 0.0209 Steps: 54540, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001325, Sample Num: 21200, Cur Loss: 0.13794625, Cur Avg Loss: 0.19012049, Log Avg loss: 0.18316666, Global Avg Loss: 0.89309225, Time: 0.0209 Steps: 54550, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001335, Sample Num: 21360, Cur Loss: 0.32121676, Cur Avg Loss: 0.19026530, Log Avg loss: 0.20945167, Global Avg Loss: 0.89296695, Time: 0.0209 Steps: 54560, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001345, Sample Num: 21520, Cur Loss: 0.14709342, Cur Avg Loss: 0.19047561, Log Avg loss: 0.21855239, Global Avg Loss: 0.89284337, Time: 0.0209 Steps: 54570, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001355, Sample Num: 21680, Cur Loss: 0.20891488, Cur Avg Loss: 0.19013383, Log Avg loss: 0.14416442, Global Avg Loss: 0.89270620, Time: 0.0209 Steps: 54580, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001365, Sample Num: 21840, Cur Loss: 0.34547320, Cur Avg Loss: 0.19000828, Log Avg loss: 0.17299652, Global Avg Loss: 0.89257436, Time: 0.0209 Steps: 54590, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001375, Sample Num: 22000, Cur Loss: 0.04244086, Cur Avg Loss: 0.19013091, Log Avg loss: 0.20686986, Global Avg Loss: 0.89244877, Time: 0.0209 Steps: 54600, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001385, Sample Num: 22160, Cur Loss: 0.48995477, Cur Avg Loss: 0.19022230, Log Avg loss: 0.20278857, Global Avg Loss: 0.89232248, Time: 0.0209 Steps: 54610, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001395, Sample Num: 22320, Cur Loss: 0.12079102, Cur Avg Loss: 0.19027011, Log Avg loss: 0.19689207, Global Avg Loss: 0.89219516, Time: 0.0209 Steps: 54620, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001405, Sample Num: 22480, Cur Loss: 0.14747538, Cur Avg Loss: 0.18985952, Log Avg loss: 0.13258108, Global Avg Loss: 0.89205611, Time: 0.0209 Steps: 54630, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001415, Sample Num: 22640, Cur Loss: 0.18908961, Cur Avg Loss: 0.18972681, Log Avg loss: 0.17108156, Global Avg Loss: 0.89192416, Time: 0.0209 Steps: 54640, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001425, Sample Num: 22800, Cur Loss: 0.09432155, Cur Avg Loss: 0.18974594, Log Avg loss: 0.19245297, Global Avg Loss: 0.89179617, Time: 0.0209 Steps: 54650, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001435, Sample Num: 22960, Cur Loss: 0.08391373, Cur Avg Loss: 0.18998698, Log Avg loss: 0.22433472, Global Avg Loss: 0.89167406, Time: 0.0209 Steps: 54660, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001445, Sample Num: 23120, Cur Loss: 0.05322166, Cur Avg Loss: 0.18962542, Log Avg loss: 0.13774186, Global Avg Loss: 0.89153615, Time: 0.0209 Steps: 54670, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001455, Sample Num: 23280, Cur Loss: 0.05203348, Cur Avg Loss: 0.18951787, Log Avg loss: 0.17397640, Global Avg Loss: 0.89140492, Time: 0.0209 Steps: 54680, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001465, Sample Num: 23440, Cur Loss: 0.22682330, Cur Avg Loss: 0.18942774, Log Avg loss: 0.17631514, Global Avg Loss: 0.89127417, Time: 0.0209 Steps: 54690, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001475, Sample Num: 23600, Cur Loss: 0.09124729, Cur Avg Loss: 0.18928935, Log Avg loss: 0.16901532, Global Avg Loss: 0.89114213, Time: 0.0209 Steps: 54700, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001485, Sample Num: 23760, Cur Loss: 0.12567319, Cur Avg Loss: 0.18919070, Log Avg loss: 0.17463908, Global Avg Loss: 0.89101117, Time: 0.0209 Steps: 54710, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001495, Sample Num: 23920, Cur Loss: 0.07670194, Cur Avg Loss: 0.18950264, Log Avg loss: 0.23582614, Global Avg Loss: 0.89089143, Time: 0.0209 Steps: 54720, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001505, Sample Num: 24080, Cur Loss: 0.32757699, Cur Avg Loss: 0.18980746, Log Avg loss: 0.23537707, Global Avg Loss: 0.89077166, Time: 0.0209 Steps: 54730, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001515, Sample Num: 24240, Cur Loss: 0.79430461, Cur Avg Loss: 0.19092146, Log Avg loss: 0.35857863, Global Avg Loss: 0.89067444, Time: 0.0209 Steps: 54740, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001525, Sample Num: 24400, Cur Loss: 0.31315595, Cur Avg Loss: 0.19138882, Log Avg loss: 0.26219464, Global Avg Loss: 0.89055965, Time: 0.0209 Steps: 54750, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001535, Sample Num: 24560, Cur Loss: 0.25588584, Cur Avg Loss: 0.19163848, Log Avg loss: 0.22971054, Global Avg Loss: 0.89043897, Time: 0.0209 Steps: 54760, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001545, Sample Num: 24720, Cur Loss: 0.37412757, Cur Avg Loss: 0.19204390, Log Avg loss: 0.25427596, Global Avg Loss: 0.89032282, Time: 0.0210 Steps: 54770, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001555, Sample Num: 24880, Cur Loss: 0.10461679, Cur Avg Loss: 0.19201440, Log Avg loss: 0.18745669, Global Avg Loss: 0.89019451, Time: 0.0210 Steps: 54780, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001565, Sample Num: 25040, Cur Loss: 0.16751239, Cur Avg Loss: 0.19173881, Log Avg loss: 0.14888576, Global Avg Loss: 0.89005921, Time: 0.0210 Steps: 54790, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001575, Sample Num: 25200, Cur Loss: 0.16362953, Cur Avg Loss: 0.19157354, Log Avg loss: 0.16570840, Global Avg Loss: 0.88992703, Time: 0.0210 Steps: 54800, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001585, Sample Num: 25360, Cur Loss: 0.19714889, Cur Avg Loss: 0.19139279, Log Avg loss: 0.16292417, Global Avg Loss: 0.88979439, Time: 0.0210 Steps: 54810, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001595, Sample Num: 25520, Cur Loss: 0.24304897, Cur Avg Loss: 0.19135470, Log Avg loss: 0.18531833, Global Avg Loss: 0.88966588, Time: 0.0210 Steps: 54820, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001605, Sample Num: 25680, Cur Loss: 0.12182410, Cur Avg Loss: 0.19149325, Log Avg loss: 0.21359116, Global Avg Loss: 0.88954258, Time: 0.0210 Steps: 54830, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001615, Sample Num: 25840, Cur Loss: 0.10743400, Cur Avg Loss: 0.19169467, Log Avg loss: 0.22402243, Global Avg Loss: 0.88942122, Time: 0.0210 Steps: 54840, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001625, Sample Num: 26000, Cur Loss: 0.27021053, Cur Avg Loss: 0.19156169, Log Avg loss: 0.17008596, Global Avg Loss: 0.88929007, Time: 0.0209 Steps: 54850, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001635, Sample Num: 26160, Cur Loss: 0.43718907, Cur Avg Loss: 0.19139130, Log Avg loss: 0.16370285, Global Avg Loss: 0.88915781, Time: 0.0210 Steps: 54860, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001645, Sample Num: 26320, Cur Loss: 0.02822771, Cur Avg Loss: 0.19108155, Log Avg loss: 0.14043673, Global Avg Loss: 0.88902136, Time: 0.0209 Steps: 54870, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001655, Sample Num: 26480, Cur Loss: 0.30087355, Cur Avg Loss: 0.19135577, Log Avg loss: 0.23646582, Global Avg Loss: 0.88890245, Time: 0.0210 Steps: 54880, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001665, Sample Num: 26640, Cur Loss: 0.31371230, Cur Avg Loss: 0.19157145, Log Avg loss: 0.22726585, Global Avg Loss: 0.88878191, Time: 0.0210 Steps: 54890, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001675, Sample Num: 26800, Cur Loss: 0.15073808, Cur Avg Loss: 0.19169083, Log Avg loss: 0.21156858, Global Avg Loss: 0.88865856, Time: 0.0210 Steps: 54900, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001685, Sample Num: 26960, Cur Loss: 0.20213979, Cur Avg Loss: 0.19170245, Log Avg loss: 0.19364807, Global Avg Loss: 0.88853199, Time: 0.0210 Steps: 54910, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001695, Sample Num: 27120, Cur Loss: 0.52317470, Cur Avg Loss: 0.19183764, Log Avg loss: 0.21461801, Global Avg Loss: 0.88840928, Time: 0.0210 Steps: 54920, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001705, Sample Num: 27280, Cur Loss: 0.24060249, Cur Avg Loss: 0.19153131, Log Avg loss: 0.13960748, Global Avg Loss: 0.88827296, Time: 0.0210 Steps: 54930, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001715, Sample Num: 27440, Cur Loss: 0.39550120, Cur Avg Loss: 0.19151659, Log Avg loss: 0.18900730, Global Avg Loss: 0.88814568, Time: 0.0209 Steps: 54940, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001725, Sample Num: 27600, Cur Loss: 0.13952282, Cur Avg Loss: 0.19135165, Log Avg loss: 0.16306345, Global Avg Loss: 0.88801373, Time: 0.0210 Steps: 54950, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001735, Sample Num: 27760, Cur Loss: 0.31556448, Cur Avg Loss: 0.19143621, Log Avg loss: 0.20602419, Global Avg Loss: 0.88788964, Time: 0.0210 Steps: 54960, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001745, Sample Num: 27920, Cur Loss: 0.11573704, Cur Avg Loss: 0.19168683, Log Avg loss: 0.23516933, Global Avg Loss: 0.88777090, Time: 0.0210 Steps: 54970, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001755, Sample Num: 28080, Cur Loss: 0.30000600, Cur Avg Loss: 0.19159648, Log Avg loss: 0.17583022, Global Avg Loss: 0.88764141, Time: 0.0210 Steps: 54980, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001765, Sample Num: 28240, Cur Loss: 0.34882066, Cur Avg Loss: 0.19167617, Log Avg loss: 0.20566130, Global Avg Loss: 0.88751739, Time: 0.0210 Steps: 54990, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001775, Sample Num: 28400, Cur Loss: 0.23678158, Cur Avg Loss: 0.19146968, Log Avg loss: 0.15502354, Global Avg Loss: 0.88738421, Time: 0.0210 Steps: 55000, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001785, Sample Num: 28560, Cur Loss: 0.15292573, Cur Avg Loss: 0.19136029, Log Avg loss: 0.17194452, Global Avg Loss: 0.88725415, Time: 0.0210 Steps: 55010, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001795, Sample Num: 28720, Cur Loss: 0.08738565, Cur Avg Loss: 0.19105820, Log Avg loss: 0.13713431, Global Avg Loss: 0.88711782, Time: 0.0247 Steps: 55020, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001805, Sample Num: 28880, Cur Loss: 0.07274947, Cur Avg Loss: 0.19078442, Log Avg loss: 0.14164158, Global Avg Loss: 0.88698235, Time: 0.0209 Steps: 55030, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001815, Sample Num: 29040, Cur Loss: 0.18906616, Cur Avg Loss: 0.19051251, Log Avg loss: 0.14143253, Global Avg Loss: 0.88684689, Time: 0.0211 Steps: 55040, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001825, Sample Num: 29200, Cur Loss: 0.22086725, Cur Avg Loss: 0.19048835, Log Avg loss: 0.18610385, Global Avg Loss: 0.88671960, Time: 0.0211 Steps: 55050, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001835, Sample Num: 29360, Cur Loss: 0.18390313, Cur Avg Loss: 0.19031631, Log Avg loss: 0.15891811, Global Avg Loss: 0.88658742, Time: 0.0210 Steps: 55060, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001845, Sample Num: 29520, Cur Loss: 0.21013388, Cur Avg Loss: 0.19013899, Log Avg loss: 0.15760195, Global Avg Loss: 0.88645504, Time: 0.0209 Steps: 55070, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001855, Sample Num: 29680, Cur Loss: 0.19856590, Cur Avg Loss: 0.19002278, Log Avg loss: 0.16858039, Global Avg Loss: 0.88632471, Time: 0.0209 Steps: 55080, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001865, Sample Num: 29840, Cur Loss: 0.12581660, Cur Avg Loss: 0.18999772, Log Avg loss: 0.18534944, Global Avg Loss: 0.88619747, Time: 0.0209 Steps: 55090, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001875, Sample Num: 30000, Cur Loss: 0.18783224, Cur Avg Loss: 0.19001792, Log Avg loss: 0.19378577, Global Avg Loss: 0.88607180, Time: 0.0209 Steps: 55100, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001885, Sample Num: 30160, Cur Loss: 0.44059885, Cur Avg Loss: 0.19053719, Log Avg loss: 0.28789977, Global Avg Loss: 0.88596326, Time: 0.0209 Steps: 55110, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001895, Sample Num: 30320, Cur Loss: 0.21710838, Cur Avg Loss: 0.19059562, Log Avg loss: 0.20160939, Global Avg Loss: 0.88583911, Time: 0.0210 Steps: 55120, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001905, Sample Num: 30480, Cur Loss: 0.17835930, Cur Avg Loss: 0.19111384, Log Avg loss: 0.28931748, Global Avg Loss: 0.88573090, Time: 0.0209 Steps: 55130, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001915, Sample Num: 30640, Cur Loss: 0.09023409, Cur Avg Loss: 0.19124557, Log Avg loss: 0.21634093, Global Avg Loss: 0.88560950, Time: 0.0210 Steps: 55140, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001925, Sample Num: 30800, Cur Loss: 0.18848194, Cur Avg Loss: 0.19114529, Log Avg loss: 0.17194002, Global Avg Loss: 0.88548010, Time: 0.0209 Steps: 55150, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001935, Sample Num: 30960, Cur Loss: 0.19739646, Cur Avg Loss: 0.19111547, Log Avg loss: 0.18537612, Global Avg Loss: 0.88535318, Time: 0.0209 Steps: 55160, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001945, Sample Num: 31120, Cur Loss: 0.02649694, Cur Avg Loss: 0.19130819, Log Avg loss: 0.22859855, Global Avg Loss: 0.88523413, Time: 0.0209 Steps: 55170, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001955, Sample Num: 31280, Cur Loss: 0.11470525, Cur Avg Loss: 0.19169210, Log Avg loss: 0.26636340, Global Avg Loss: 0.88512198, Time: 0.0209 Steps: 55180, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001965, Sample Num: 31440, Cur Loss: 0.05427983, Cur Avg Loss: 0.19174880, Log Avg loss: 0.20283368, Global Avg Loss: 0.88499835, Time: 0.0211 Steps: 55190, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001975, Sample Num: 31600, Cur Loss: 0.35653630, Cur Avg Loss: 0.19159420, Log Avg loss: 0.16121426, Global Avg Loss: 0.88486723, Time: 0.0211 Steps: 55200, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001985, Sample Num: 31760, Cur Loss: 0.04984359, Cur Avg Loss: 0.19132196, Log Avg loss: 0.13755461, Global Avg Loss: 0.88473188, Time: 0.0210 Steps: 55210, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001995, Sample Num: 31920, Cur Loss: 0.10030372, Cur Avg Loss: 0.19116589, Log Avg loss: 0.16018749, Global Avg Loss: 0.88460067, Time: 0.0210 Steps: 55220, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002005, Sample Num: 32080, Cur Loss: 0.18718630, Cur Avg Loss: 0.19140973, Log Avg loss: 0.24005480, Global Avg Loss: 0.88448396, Time: 0.0211 Steps: 55230, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002015, Sample Num: 32240, Cur Loss: 0.16636376, Cur Avg Loss: 0.19157151, Log Avg loss: 0.22400867, Global Avg Loss: 0.88436440, Time: 0.0210 Steps: 55240, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002025, Sample Num: 32400, Cur Loss: 0.15952738, Cur Avg Loss: 0.19166255, Log Avg loss: 0.21000685, Global Avg Loss: 0.88424234, Time: 0.0210 Steps: 55250, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002035, Sample Num: 32560, Cur Loss: 0.62276798, Cur Avg Loss: 0.19175854, Log Avg loss: 0.21119703, Global Avg Loss: 0.88412055, Time: 0.0210 Steps: 55260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002045, Sample Num: 32720, Cur Loss: 0.18438572, Cur Avg Loss: 0.19232074, Log Avg loss: 0.30672852, Global Avg Loss: 0.88401608, Time: 0.0210 Steps: 55270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002055, Sample Num: 32880, Cur Loss: 0.10762143, Cur Avg Loss: 0.19263691, Log Avg loss: 0.25729367, Global Avg Loss: 0.88390271, Time: 0.0238 Steps: 55280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002065, Sample Num: 33040, Cur Loss: 0.19093716, Cur Avg Loss: 0.19233827, Log Avg loss: 0.13096789, Global Avg Loss: 0.88376653, Time: 0.0209 Steps: 55290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002075, Sample Num: 33200, Cur Loss: 0.08061276, Cur Avg Loss: 0.19211585, Log Avg loss: 0.14618612, Global Avg Loss: 0.88363315, Time: 0.0209 Steps: 55300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002085, Sample Num: 33360, Cur Loss: 0.11282858, Cur Avg Loss: 0.19214247, Log Avg loss: 0.19766578, Global Avg Loss: 0.88350913, Time: 0.0209 Steps: 55310, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002095, Sample Num: 33520, Cur Loss: 0.08113372, Cur Avg Loss: 0.19194463, Log Avg loss: 0.15069571, Global Avg Loss: 0.88337666, Time: 0.0209 Steps: 55320, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002105, Sample Num: 33680, Cur Loss: 0.04922085, Cur Avg Loss: 0.19231448, Log Avg loss: 0.26979800, Global Avg Loss: 0.88326577, Time: 0.0210 Steps: 55330, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002115, Sample Num: 33840, Cur Loss: 0.14661908, Cur Avg Loss: 0.19227738, Log Avg loss: 0.18446697, Global Avg Loss: 0.88313949, Time: 0.0209 Steps: 55340, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002125, Sample Num: 34000, Cur Loss: 0.31089112, Cur Avg Loss: 0.19210494, Log Avg loss: 0.15563389, Global Avg Loss: 0.88300805, Time: 0.0209 Steps: 55350, Updated lr: 0.000048 ***** Running evaluation checkpoint-55354 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-55354 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.981267, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.223397, "eval_total_loss": 157.047904, "eval_mae": 0.314518, "eval_mse": 0.223475, "eval_r2": 0.857944, "eval_sp_statistic": 0.918736, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.928649, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.192038, "test_total_loss": 96.403264, "test_mae": 0.312489, "test_mse": 0.192108, "test_r2": 0.876012, "test_sp_statistic": 0.911131, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.942668, "test_ps_pvalue": 0.0, "lr": 4.845519203413941e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8829539264620436, "train_cur_epoch_loss": 408.7588477935642, "train_cur_epoch_avg_loss": 0.1919957011712373, "train_cur_epoch_time": 44.98126673698425, "train_cur_epoch_avg_time": 0.021127884798959256, "epoch": 26, "step": 55354} ################################################## Training, Epoch: 0027, Batch: 000006, Sample Num: 96, Cur Loss: 0.31047493, Cur Avg Loss: 0.13458172, Log Avg loss: 0.13433424, Global Avg Loss: 0.88287282, Time: 0.0211 Steps: 55360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000016, Sample Num: 256, Cur Loss: 0.20884758, Cur Avg Loss: 0.19263453, Log Avg loss: 0.22746621, Global Avg Loss: 0.88275445, Time: 0.0209 Steps: 55370, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000026, Sample Num: 416, Cur Loss: 0.20838037, Cur Avg Loss: 0.19738602, Log Avg loss: 0.20498841, Global Avg Loss: 0.88263206, Time: 0.0211 Steps: 55380, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000036, Sample Num: 576, Cur Loss: 0.06994495, Cur Avg Loss: 0.17109747, Log Avg loss: 0.10274725, Global Avg Loss: 0.88249126, Time: 0.0210 Steps: 55390, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000046, Sample Num: 736, Cur Loss: 0.18930803, Cur Avg Loss: 0.18399059, Log Avg loss: 0.23040581, Global Avg Loss: 0.88237356, Time: 0.0211 Steps: 55400, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000056, Sample Num: 896, Cur Loss: 0.16691469, Cur Avg Loss: 0.18791614, Log Avg loss: 0.20597369, Global Avg Loss: 0.88225149, Time: 0.0210 Steps: 55410, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000066, Sample Num: 1056, Cur Loss: 0.58678496, Cur Avg Loss: 0.19761463, Log Avg loss: 0.25192618, Global Avg Loss: 0.88213775, Time: 0.0211 Steps: 55420, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000076, Sample Num: 1216, Cur Loss: 0.11620240, Cur Avg Loss: 0.18768131, Log Avg loss: 0.12212140, Global Avg Loss: 0.88200064, Time: 0.0210 Steps: 55430, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000086, Sample Num: 1376, Cur Loss: 0.19383605, Cur Avg Loss: 0.19043725, Log Avg loss: 0.21138236, Global Avg Loss: 0.88187968, Time: 0.0210 Steps: 55440, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000096, Sample Num: 1536, Cur Loss: 0.18705186, Cur Avg Loss: 0.19375371, Log Avg loss: 0.22227527, Global Avg Loss: 0.88176072, Time: 0.0209 Steps: 55450, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000106, Sample Num: 1696, Cur Loss: 0.12371382, Cur Avg Loss: 0.19216008, Log Avg loss: 0.17686120, Global Avg Loss: 0.88163362, Time: 0.0210 Steps: 55460, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000116, Sample Num: 1856, Cur Loss: 0.04581500, Cur Avg Loss: 0.19353136, Log Avg loss: 0.20806693, Global Avg Loss: 0.88151219, Time: 0.0210 Steps: 55470, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000126, Sample Num: 2016, Cur Loss: 0.26649457, Cur Avg Loss: 0.18918069, Log Avg loss: 0.13871295, Global Avg Loss: 0.88137831, Time: 0.0210 Steps: 55480, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000136, Sample Num: 2176, Cur Loss: 0.10709846, Cur Avg Loss: 0.18657367, Log Avg loss: 0.15372516, Global Avg Loss: 0.88124717, Time: 0.0210 Steps: 55490, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000146, Sample Num: 2336, Cur Loss: 0.44307330, Cur Avg Loss: 0.18645638, Log Avg loss: 0.18486129, Global Avg Loss: 0.88112170, Time: 0.0210 Steps: 55500, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000156, Sample Num: 2496, Cur Loss: 0.14685616, Cur Avg Loss: 0.19130449, Log Avg loss: 0.26208690, Global Avg Loss: 0.88101018, Time: 0.0211 Steps: 55510, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000166, Sample Num: 2656, Cur Loss: 0.17009005, Cur Avg Loss: 0.19057099, Log Avg loss: 0.17912847, Global Avg Loss: 0.88088376, Time: 0.0210 Steps: 55520, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000176, Sample Num: 2816, Cur Loss: 0.18221790, Cur Avg Loss: 0.18994748, Log Avg loss: 0.17959712, Global Avg Loss: 0.88075747, Time: 0.0213 Steps: 55530, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000186, Sample Num: 2976, Cur Loss: 0.25905550, Cur Avg Loss: 0.19103740, Log Avg loss: 0.21022004, Global Avg Loss: 0.88063674, Time: 0.0210 Steps: 55540, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000196, Sample Num: 3136, Cur Loss: 0.26404533, Cur Avg Loss: 0.19314478, Log Avg loss: 0.23234196, Global Avg Loss: 0.88052004, Time: 0.0210 Steps: 55550, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000206, Sample Num: 3296, Cur Loss: 0.34410924, Cur Avg Loss: 0.19277367, Log Avg loss: 0.18549990, Global Avg Loss: 0.88039494, Time: 0.0211 Steps: 55560, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000216, Sample Num: 3456, Cur Loss: 1.00997531, Cur Avg Loss: 0.19642606, Log Avg loss: 0.27166532, Global Avg Loss: 0.88028540, Time: 0.0210 Steps: 55570, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000226, Sample Num: 3616, Cur Loss: 0.27778444, Cur Avg Loss: 0.19704974, Log Avg loss: 0.21052117, Global Avg Loss: 0.88016490, Time: 0.0211 Steps: 55580, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000236, Sample Num: 3776, Cur Loss: 0.10017063, Cur Avg Loss: 0.19612876, Log Avg loss: 0.17531471, Global Avg Loss: 0.88003810, Time: 0.0209 Steps: 55590, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000246, Sample Num: 3936, Cur Loss: 0.10943554, Cur Avg Loss: 0.19563209, Log Avg loss: 0.18391057, Global Avg Loss: 0.87991290, Time: 0.0210 Steps: 55600, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000256, Sample Num: 4096, Cur Loss: 0.14212964, Cur Avg Loss: 0.19553064, Log Avg loss: 0.19303514, Global Avg Loss: 0.87978938, Time: 0.0255 Steps: 55610, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000266, Sample Num: 4256, Cur Loss: 0.10095142, Cur Avg Loss: 0.19517298, Log Avg loss: 0.18601676, Global Avg Loss: 0.87966465, Time: 0.0210 Steps: 55620, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000276, Sample Num: 4416, Cur Loss: 0.13311443, Cur Avg Loss: 0.19327671, Log Avg loss: 0.14283607, Global Avg Loss: 0.87953220, Time: 0.0210 Steps: 55630, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000286, Sample Num: 4576, Cur Loss: 0.14548279, Cur Avg Loss: 0.19178867, Log Avg loss: 0.15071858, Global Avg Loss: 0.87940121, Time: 0.0210 Steps: 55640, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000296, Sample Num: 4736, Cur Loss: 0.08765060, Cur Avg Loss: 0.19117559, Log Avg loss: 0.17364155, Global Avg Loss: 0.87927439, Time: 0.0209 Steps: 55650, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000306, Sample Num: 4896, Cur Loss: 0.16778514, Cur Avg Loss: 0.19351941, Log Avg loss: 0.26289653, Global Avg Loss: 0.87916365, Time: 0.0209 Steps: 55660, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000316, Sample Num: 5056, Cur Loss: 0.17637324, Cur Avg Loss: 0.19180387, Log Avg loss: 0.13930821, Global Avg Loss: 0.87903075, Time: 0.0209 Steps: 55670, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000326, Sample Num: 5216, Cur Loss: 0.26761675, Cur Avg Loss: 0.19121165, Log Avg loss: 0.17249758, Global Avg Loss: 0.87890385, Time: 0.0209 Steps: 55680, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000336, Sample Num: 5376, Cur Loss: 0.28446406, Cur Avg Loss: 0.19238334, Log Avg loss: 0.23058044, Global Avg Loss: 0.87878744, Time: 0.0209 Steps: 55690, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000346, Sample Num: 5536, Cur Loss: 0.43178323, Cur Avg Loss: 0.19378114, Log Avg loss: 0.24074744, Global Avg Loss: 0.87867289, Time: 0.0209 Steps: 55700, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000356, Sample Num: 5696, Cur Loss: 0.09296955, Cur Avg Loss: 0.19251024, Log Avg loss: 0.14853696, Global Avg Loss: 0.87854183, Time: 0.0209 Steps: 55710, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000366, Sample Num: 5856, Cur Loss: 0.07111743, Cur Avg Loss: 0.19301903, Log Avg loss: 0.21113184, Global Avg Loss: 0.87842205, Time: 0.0209 Steps: 55720, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000376, Sample Num: 6016, Cur Loss: 0.25186062, Cur Avg Loss: 0.19392459, Log Avg loss: 0.22706825, Global Avg Loss: 0.87830517, Time: 0.0209 Steps: 55730, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000386, Sample Num: 6176, Cur Loss: 0.33648399, Cur Avg Loss: 0.19264917, Log Avg loss: 0.14469313, Global Avg Loss: 0.87817356, Time: 0.0210 Steps: 55740, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000396, Sample Num: 6336, Cur Loss: 0.16075769, Cur Avg Loss: 0.19353847, Log Avg loss: 0.22786561, Global Avg Loss: 0.87805691, Time: 0.0209 Steps: 55750, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000406, Sample Num: 6496, Cur Loss: 0.27207166, Cur Avg Loss: 0.19373260, Log Avg loss: 0.20142006, Global Avg Loss: 0.87793556, Time: 0.0210 Steps: 55760, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000416, Sample Num: 6656, Cur Loss: 0.20039573, Cur Avg Loss: 0.19265982, Log Avg loss: 0.14910486, Global Avg Loss: 0.87780488, Time: 0.0209 Steps: 55770, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000426, Sample Num: 6816, Cur Loss: 0.15059742, Cur Avg Loss: 0.19192795, Log Avg loss: 0.16148234, Global Avg Loss: 0.87767646, Time: 0.0209 Steps: 55780, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000436, Sample Num: 6976, Cur Loss: 0.07280244, Cur Avg Loss: 0.19354834, Log Avg loss: 0.26257711, Global Avg Loss: 0.87756621, Time: 0.0209 Steps: 55790, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000446, Sample Num: 7136, Cur Loss: 0.18669838, Cur Avg Loss: 0.19432880, Log Avg loss: 0.22835652, Global Avg Loss: 0.87744986, Time: 0.0209 Steps: 55800, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000456, Sample Num: 7296, Cur Loss: 0.13779548, Cur Avg Loss: 0.19371497, Log Avg loss: 0.16633832, Global Avg Loss: 0.87732245, Time: 0.0209 Steps: 55810, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000466, Sample Num: 7456, Cur Loss: 0.06163020, Cur Avg Loss: 0.19232977, Log Avg loss: 0.12916489, Global Avg Loss: 0.87718841, Time: 0.0209 Steps: 55820, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000476, Sample Num: 7616, Cur Loss: 0.06894501, Cur Avg Loss: 0.19136076, Log Avg loss: 0.14620483, Global Avg Loss: 0.87705748, Time: 0.0209 Steps: 55830, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000486, Sample Num: 7776, Cur Loss: 0.32646775, Cur Avg Loss: 0.19156605, Log Avg loss: 0.20133774, Global Avg Loss: 0.87693647, Time: 0.0209 Steps: 55840, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000496, Sample Num: 7936, Cur Loss: 0.09677535, Cur Avg Loss: 0.19402737, Log Avg loss: 0.31364769, Global Avg Loss: 0.87683562, Time: 0.0209 Steps: 55850, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000506, Sample Num: 8096, Cur Loss: 0.16018426, Cur Avg Loss: 0.19461084, Log Avg loss: 0.22355073, Global Avg Loss: 0.87671867, Time: 0.0209 Steps: 55860, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000516, Sample Num: 8256, Cur Loss: 0.05951428, Cur Avg Loss: 0.19458573, Log Avg loss: 0.19331505, Global Avg Loss: 0.87659635, Time: 0.0247 Steps: 55870, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000526, Sample Num: 8416, Cur Loss: 0.27775079, Cur Avg Loss: 0.19537392, Log Avg loss: 0.23604441, Global Avg Loss: 0.87648172, Time: 0.0210 Steps: 55880, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000536, Sample Num: 8576, Cur Loss: 0.28294328, Cur Avg Loss: 0.19597100, Log Avg loss: 0.22737784, Global Avg Loss: 0.87636558, Time: 0.0210 Steps: 55890, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000546, Sample Num: 8736, Cur Loss: 0.09509054, Cur Avg Loss: 0.19480874, Log Avg loss: 0.13251143, Global Avg Loss: 0.87623251, Time: 0.0210 Steps: 55900, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000556, Sample Num: 8896, Cur Loss: 0.07207192, Cur Avg Loss: 0.19350764, Log Avg loss: 0.12246743, Global Avg Loss: 0.87609769, Time: 0.0210 Steps: 55910, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000566, Sample Num: 9056, Cur Loss: 0.08747813, Cur Avg Loss: 0.19322344, Log Avg loss: 0.17742218, Global Avg Loss: 0.87597275, Time: 0.0210 Steps: 55920, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000576, Sample Num: 9216, Cur Loss: 0.10033408, Cur Avg Loss: 0.19327794, Log Avg loss: 0.19636242, Global Avg Loss: 0.87585124, Time: 0.0210 Steps: 55930, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000586, Sample Num: 9376, Cur Loss: 0.11104554, Cur Avg Loss: 0.19208684, Log Avg loss: 0.12347937, Global Avg Loss: 0.87571674, Time: 0.0209 Steps: 55940, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000596, Sample Num: 9536, Cur Loss: 0.34721476, Cur Avg Loss: 0.19240120, Log Avg loss: 0.21082313, Global Avg Loss: 0.87559790, Time: 0.0209 Steps: 55950, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000606, Sample Num: 9696, Cur Loss: 0.16361091, Cur Avg Loss: 0.19266519, Log Avg loss: 0.20839856, Global Avg Loss: 0.87547868, Time: 0.0209 Steps: 55960, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000616, Sample Num: 9856, Cur Loss: 0.18403292, Cur Avg Loss: 0.19260293, Log Avg loss: 0.18883040, Global Avg Loss: 0.87535600, Time: 0.0209 Steps: 55970, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000626, Sample Num: 10016, Cur Loss: 0.24945322, Cur Avg Loss: 0.19295135, Log Avg loss: 0.21441356, Global Avg Loss: 0.87523793, Time: 0.0210 Steps: 55980, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000636, Sample Num: 10176, Cur Loss: 0.17339113, Cur Avg Loss: 0.19314708, Log Avg loss: 0.20540029, Global Avg Loss: 0.87511829, Time: 0.0211 Steps: 55990, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000646, Sample Num: 10336, Cur Loss: 0.29756281, Cur Avg Loss: 0.19378107, Log Avg loss: 0.23410244, Global Avg Loss: 0.87500383, Time: 0.0210 Steps: 56000, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000656, Sample Num: 10496, Cur Loss: 0.15688975, Cur Avg Loss: 0.19334356, Log Avg loss: 0.16508074, Global Avg Loss: 0.87487708, Time: 0.0210 Steps: 56010, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000666, Sample Num: 10656, Cur Loss: 0.15165699, Cur Avg Loss: 0.19227159, Log Avg loss: 0.12195039, Global Avg Loss: 0.87474267, Time: 0.0210 Steps: 56020, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000676, Sample Num: 10816, Cur Loss: 0.30979946, Cur Avg Loss: 0.19195759, Log Avg loss: 0.17104528, Global Avg Loss: 0.87461708, Time: 0.0210 Steps: 56030, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000686, Sample Num: 10976, Cur Loss: 0.06220989, Cur Avg Loss: 0.19155112, Log Avg loss: 0.16407324, Global Avg Loss: 0.87449029, Time: 0.0211 Steps: 56040, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000696, Sample Num: 11136, Cur Loss: 0.50716597, Cur Avg Loss: 0.19184268, Log Avg loss: 0.21184371, Global Avg Loss: 0.87437206, Time: 0.0210 Steps: 56050, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000706, Sample Num: 11296, Cur Loss: 0.47446874, Cur Avg Loss: 0.19207586, Log Avg loss: 0.20830559, Global Avg Loss: 0.87425325, Time: 0.0211 Steps: 56060, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000716, Sample Num: 11456, Cur Loss: 0.13632588, Cur Avg Loss: 0.19206957, Log Avg loss: 0.19162559, Global Avg Loss: 0.87413150, Time: 0.0210 Steps: 56070, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000726, Sample Num: 11616, Cur Loss: 0.17685370, Cur Avg Loss: 0.19258821, Log Avg loss: 0.22972259, Global Avg Loss: 0.87401660, Time: 0.0210 Steps: 56080, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000736, Sample Num: 11776, Cur Loss: 0.07560033, Cur Avg Loss: 0.19217058, Log Avg loss: 0.16185054, Global Avg Loss: 0.87388963, Time: 0.0210 Steps: 56090, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000746, Sample Num: 11936, Cur Loss: 0.14652304, Cur Avg Loss: 0.19153487, Log Avg loss: 0.14474680, Global Avg Loss: 0.87375966, Time: 0.0210 Steps: 56100, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000756, Sample Num: 12096, Cur Loss: 0.21341534, Cur Avg Loss: 0.19081227, Log Avg loss: 0.13690597, Global Avg Loss: 0.87362833, Time: 0.0210 Steps: 56110, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000766, Sample Num: 12256, Cur Loss: 0.10604958, Cur Avg Loss: 0.19115741, Log Avg loss: 0.21724994, Global Avg Loss: 0.87351137, Time: 0.0210 Steps: 56120, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000776, Sample Num: 12416, Cur Loss: 0.13611081, Cur Avg Loss: 0.19147566, Log Avg loss: 0.21585354, Global Avg Loss: 0.87339421, Time: 0.0210 Steps: 56130, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000786, Sample Num: 12576, Cur Loss: 0.16454858, Cur Avg Loss: 0.19130787, Log Avg loss: 0.17828737, Global Avg Loss: 0.87327039, Time: 0.0209 Steps: 56140, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000796, Sample Num: 12736, Cur Loss: 0.10445511, Cur Avg Loss: 0.19302070, Log Avg loss: 0.32764907, Global Avg Loss: 0.87317322, Time: 0.0209 Steps: 56150, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000806, Sample Num: 12896, Cur Loss: 0.11439251, Cur Avg Loss: 0.19238262, Log Avg loss: 0.14159212, Global Avg Loss: 0.87304295, Time: 0.0209 Steps: 56160, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000816, Sample Num: 13056, Cur Loss: 0.23937106, Cur Avg Loss: 0.19224013, Log Avg loss: 0.18075476, Global Avg Loss: 0.87291970, Time: 0.0208 Steps: 56170, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000826, Sample Num: 13216, Cur Loss: 0.15861945, Cur Avg Loss: 0.19297670, Log Avg loss: 0.25308111, Global Avg Loss: 0.87280937, Time: 0.0209 Steps: 56180, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000836, Sample Num: 13376, Cur Loss: 0.24205565, Cur Avg Loss: 0.19355207, Log Avg loss: 0.24107805, Global Avg Loss: 0.87269694, Time: 0.0208 Steps: 56190, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000846, Sample Num: 13536, Cur Loss: 0.19065051, Cur Avg Loss: 0.19438907, Log Avg loss: 0.26436151, Global Avg Loss: 0.87258870, Time: 0.0209 Steps: 56200, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000856, Sample Num: 13696, Cur Loss: 0.13776088, Cur Avg Loss: 0.19455082, Log Avg loss: 0.20823539, Global Avg Loss: 0.87247051, Time: 0.0209 Steps: 56210, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000866, Sample Num: 13856, Cur Loss: 0.27219868, Cur Avg Loss: 0.19483976, Log Avg loss: 0.21957331, Global Avg Loss: 0.87235437, Time: 0.0209 Steps: 56220, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000876, Sample Num: 14016, Cur Loss: 0.30270085, Cur Avg Loss: 0.19498979, Log Avg loss: 0.20798163, Global Avg Loss: 0.87223622, Time: 0.0209 Steps: 56230, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000886, Sample Num: 14176, Cur Loss: 0.16471437, Cur Avg Loss: 0.19627252, Log Avg loss: 0.30863987, Global Avg Loss: 0.87213601, Time: 0.0209 Steps: 56240, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000896, Sample Num: 14336, Cur Loss: 0.20398927, Cur Avg Loss: 0.19847375, Log Avg loss: 0.39350252, Global Avg Loss: 0.87205092, Time: 0.0208 Steps: 56250, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000906, Sample Num: 14496, Cur Loss: 0.20619640, Cur Avg Loss: 0.19938935, Log Avg loss: 0.28142726, Global Avg Loss: 0.87194594, Time: 0.0209 Steps: 56260, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000916, Sample Num: 14656, Cur Loss: 0.32875705, Cur Avg Loss: 0.20059340, Log Avg loss: 0.30968055, Global Avg Loss: 0.87184601, Time: 0.0209 Steps: 56270, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000926, Sample Num: 14816, Cur Loss: 0.29739791, Cur Avg Loss: 0.20041409, Log Avg loss: 0.18398931, Global Avg Loss: 0.87172379, Time: 0.0208 Steps: 56280, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000936, Sample Num: 14976, Cur Loss: 0.13644412, Cur Avg Loss: 0.19977070, Log Avg loss: 0.14019321, Global Avg Loss: 0.87159384, Time: 0.0209 Steps: 56290, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000946, Sample Num: 15136, Cur Loss: 0.09241475, Cur Avg Loss: 0.19949659, Log Avg loss: 0.17383927, Global Avg Loss: 0.87146990, Time: 0.0208 Steps: 56300, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000956, Sample Num: 15296, Cur Loss: 0.30883765, Cur Avg Loss: 0.19943038, Log Avg loss: 0.19316678, Global Avg Loss: 0.87134944, Time: 0.0209 Steps: 56310, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000966, Sample Num: 15456, Cur Loss: 0.48522583, Cur Avg Loss: 0.19962913, Log Avg loss: 0.21863027, Global Avg Loss: 0.87123355, Time: 0.0209 Steps: 56320, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000976, Sample Num: 15616, Cur Loss: 0.23780648, Cur Avg Loss: 0.19982577, Log Avg loss: 0.21882049, Global Avg Loss: 0.87111773, Time: 0.0209 Steps: 56330, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000986, Sample Num: 15776, Cur Loss: 0.22726585, Cur Avg Loss: 0.20037124, Log Avg loss: 0.25360914, Global Avg Loss: 0.87100812, Time: 0.0209 Steps: 56340, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000996, Sample Num: 15936, Cur Loss: 0.09641619, Cur Avg Loss: 0.19973072, Log Avg loss: 0.13657569, Global Avg Loss: 0.87087779, Time: 0.0209 Steps: 56350, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001006, Sample Num: 16096, Cur Loss: 0.19942005, Cur Avg Loss: 0.19918708, Log Avg loss: 0.14504054, Global Avg Loss: 0.87074900, Time: 0.0209 Steps: 56360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001016, Sample Num: 16256, Cur Loss: 0.26309401, Cur Avg Loss: 0.19878134, Log Avg loss: 0.15796354, Global Avg Loss: 0.87062256, Time: 0.0208 Steps: 56370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001026, Sample Num: 16416, Cur Loss: 0.10622896, Cur Avg Loss: 0.19859338, Log Avg loss: 0.17949711, Global Avg Loss: 0.87049997, Time: 0.0248 Steps: 56380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001036, Sample Num: 16576, Cur Loss: 0.54882699, Cur Avg Loss: 0.19891879, Log Avg loss: 0.23230549, Global Avg Loss: 0.87038680, Time: 0.0209 Steps: 56390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001046, Sample Num: 16736, Cur Loss: 0.06141470, Cur Avg Loss: 0.19884609, Log Avg loss: 0.19131447, Global Avg Loss: 0.87026639, Time: 0.0210 Steps: 56400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001056, Sample Num: 16896, Cur Loss: 0.06421155, Cur Avg Loss: 0.19892273, Log Avg loss: 0.20693930, Global Avg Loss: 0.87014880, Time: 0.0209 Steps: 56410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001066, Sample Num: 17056, Cur Loss: 0.08078696, Cur Avg Loss: 0.19836509, Log Avg loss: 0.13947879, Global Avg Loss: 0.87001930, Time: 0.0210 Steps: 56420, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001076, Sample Num: 17216, Cur Loss: 0.08703842, Cur Avg Loss: 0.19808029, Log Avg loss: 0.16772048, Global Avg Loss: 0.86989484, Time: 0.0210 Steps: 56430, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001086, Sample Num: 17376, Cur Loss: 0.39872530, Cur Avg Loss: 0.19794797, Log Avg loss: 0.18370983, Global Avg Loss: 0.86977327, Time: 0.0210 Steps: 56440, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001096, Sample Num: 17536, Cur Loss: 0.19863404, Cur Avg Loss: 0.19789299, Log Avg loss: 0.19192283, Global Avg Loss: 0.86965319, Time: 0.0209 Steps: 56450, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001106, Sample Num: 17696, Cur Loss: 0.11756017, Cur Avg Loss: 0.19818830, Log Avg loss: 0.23055423, Global Avg Loss: 0.86953999, Time: 0.0210 Steps: 56460, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001116, Sample Num: 17856, Cur Loss: 0.14630799, Cur Avg Loss: 0.19819037, Log Avg loss: 0.19841858, Global Avg Loss: 0.86942115, Time: 0.0210 Steps: 56470, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001126, Sample Num: 18016, Cur Loss: 0.10985426, Cur Avg Loss: 0.19799117, Log Avg loss: 0.17576067, Global Avg Loss: 0.86929833, Time: 0.0210 Steps: 56480, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001136, Sample Num: 18176, Cur Loss: 0.12172291, Cur Avg Loss: 0.19739773, Log Avg loss: 0.13057681, Global Avg Loss: 0.86916756, Time: 0.0209 Steps: 56490, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001146, Sample Num: 18336, Cur Loss: 0.10305681, Cur Avg Loss: 0.19745243, Log Avg loss: 0.20366644, Global Avg Loss: 0.86904977, Time: 0.0209 Steps: 56500, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001156, Sample Num: 18496, Cur Loss: 0.15501180, Cur Avg Loss: 0.19699692, Log Avg loss: 0.14479533, Global Avg Loss: 0.86892161, Time: 0.0210 Steps: 56510, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001166, Sample Num: 18656, Cur Loss: 0.23549721, Cur Avg Loss: 0.19697441, Log Avg loss: 0.19437211, Global Avg Loss: 0.86880226, Time: 0.0210 Steps: 56520, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001176, Sample Num: 18816, Cur Loss: 0.10087965, Cur Avg Loss: 0.19711803, Log Avg loss: 0.21386453, Global Avg Loss: 0.86868640, Time: 0.0210 Steps: 56530, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001186, Sample Num: 18976, Cur Loss: 0.09161857, Cur Avg Loss: 0.19710963, Log Avg loss: 0.19612137, Global Avg Loss: 0.86856745, Time: 0.0210 Steps: 56540, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001196, Sample Num: 19136, Cur Loss: 0.05131103, Cur Avg Loss: 0.19739160, Log Avg loss: 0.23083277, Global Avg Loss: 0.86845468, Time: 0.0210 Steps: 56550, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001206, Sample Num: 19296, Cur Loss: 0.19885674, Cur Avg Loss: 0.19714716, Log Avg loss: 0.16791239, Global Avg Loss: 0.86833082, Time: 0.0210 Steps: 56560, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001216, Sample Num: 19456, Cur Loss: 0.09638697, Cur Avg Loss: 0.19729821, Log Avg loss: 0.21551465, Global Avg Loss: 0.86821542, Time: 0.0210 Steps: 56570, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001226, Sample Num: 19616, Cur Loss: 0.32512882, Cur Avg Loss: 0.19759801, Log Avg loss: 0.23405459, Global Avg Loss: 0.86810334, Time: 0.0209 Steps: 56580, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001236, Sample Num: 19776, Cur Loss: 0.06814902, Cur Avg Loss: 0.19759646, Log Avg loss: 0.19740659, Global Avg Loss: 0.86798482, Time: 0.0213 Steps: 56590, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001246, Sample Num: 19936, Cur Loss: 0.07934642, Cur Avg Loss: 0.19756241, Log Avg loss: 0.19335325, Global Avg Loss: 0.86786563, Time: 0.0210 Steps: 56600, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001256, Sample Num: 20096, Cur Loss: 0.11712070, Cur Avg Loss: 0.19742024, Log Avg loss: 0.17970580, Global Avg Loss: 0.86774406, Time: 0.0210 Steps: 56610, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001266, Sample Num: 20256, Cur Loss: 0.24662414, Cur Avg Loss: 0.19780984, Log Avg loss: 0.24674389, Global Avg Loss: 0.86763439, Time: 0.0209 Steps: 56620, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001276, Sample Num: 20416, Cur Loss: 0.25238705, Cur Avg Loss: 0.19760861, Log Avg loss: 0.17213228, Global Avg Loss: 0.86751157, Time: 0.0209 Steps: 56630, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001286, Sample Num: 20576, Cur Loss: 0.13100654, Cur Avg Loss: 0.19729090, Log Avg loss: 0.15675157, Global Avg Loss: 0.86738608, Time: 0.0210 Steps: 56640, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001296, Sample Num: 20736, Cur Loss: 0.08214944, Cur Avg Loss: 0.19722548, Log Avg loss: 0.18881244, Global Avg Loss: 0.86726630, Time: 0.0211 Steps: 56650, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001306, Sample Num: 20896, Cur Loss: 0.11767823, Cur Avg Loss: 0.19674809, Log Avg loss: 0.13487801, Global Avg Loss: 0.86713704, Time: 0.0208 Steps: 56660, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001316, Sample Num: 21056, Cur Loss: 0.20099092, Cur Avg Loss: 0.19669758, Log Avg loss: 0.19010107, Global Avg Loss: 0.86701757, Time: 0.0209 Steps: 56670, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001326, Sample Num: 21216, Cur Loss: 0.24358603, Cur Avg Loss: 0.19653132, Log Avg loss: 0.17465105, Global Avg Loss: 0.86689542, Time: 0.0208 Steps: 56680, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001336, Sample Num: 21376, Cur Loss: 0.50563788, Cur Avg Loss: 0.19659083, Log Avg loss: 0.20448306, Global Avg Loss: 0.86677857, Time: 0.0209 Steps: 56690, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001346, Sample Num: 21536, Cur Loss: 0.21174473, Cur Avg Loss: 0.19603159, Log Avg loss: 0.12131667, Global Avg Loss: 0.86664709, Time: 0.0208 Steps: 56700, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001356, Sample Num: 21696, Cur Loss: 0.10327355, Cur Avg Loss: 0.19587729, Log Avg loss: 0.17510804, Global Avg Loss: 0.86652515, Time: 0.0209 Steps: 56710, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001366, Sample Num: 21856, Cur Loss: 0.09215863, Cur Avg Loss: 0.19622140, Log Avg loss: 0.24288283, Global Avg Loss: 0.86641520, Time: 0.0208 Steps: 56720, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001376, Sample Num: 22016, Cur Loss: 0.13569818, Cur Avg Loss: 0.19605007, Log Avg loss: 0.17264695, Global Avg Loss: 0.86629291, Time: 0.0209 Steps: 56730, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001386, Sample Num: 22176, Cur Loss: 0.14332403, Cur Avg Loss: 0.19565139, Log Avg loss: 0.14079207, Global Avg Loss: 0.86616504, Time: 0.0208 Steps: 56740, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001396, Sample Num: 22336, Cur Loss: 0.04790967, Cur Avg Loss: 0.19567671, Log Avg loss: 0.19918639, Global Avg Loss: 0.86604751, Time: 0.0209 Steps: 56750, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001406, Sample Num: 22496, Cur Loss: 0.22029191, Cur Avg Loss: 0.19536827, Log Avg loss: 0.15231046, Global Avg Loss: 0.86592177, Time: 0.0208 Steps: 56760, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001416, Sample Num: 22656, Cur Loss: 0.17217028, Cur Avg Loss: 0.19563217, Log Avg loss: 0.23273578, Global Avg Loss: 0.86581023, Time: 0.0209 Steps: 56770, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001426, Sample Num: 22816, Cur Loss: 0.13598825, Cur Avg Loss: 0.19620097, Log Avg loss: 0.27674356, Global Avg Loss: 0.86570649, Time: 0.0208 Steps: 56780, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001436, Sample Num: 22976, Cur Loss: 0.03639751, Cur Avg Loss: 0.19589182, Log Avg loss: 0.15180784, Global Avg Loss: 0.86558078, Time: 0.0209 Steps: 56790, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001446, Sample Num: 23136, Cur Loss: 0.37324521, Cur Avg Loss: 0.19635845, Log Avg loss: 0.26336584, Global Avg Loss: 0.86547475, Time: 0.0208 Steps: 56800, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001456, Sample Num: 23296, Cur Loss: 0.10305516, Cur Avg Loss: 0.19667806, Log Avg loss: 0.24289342, Global Avg Loss: 0.86536516, Time: 0.0209 Steps: 56810, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001466, Sample Num: 23456, Cur Loss: 0.13296738, Cur Avg Loss: 0.19657587, Log Avg loss: 0.18169723, Global Avg Loss: 0.86524484, Time: 0.0208 Steps: 56820, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001476, Sample Num: 23616, Cur Loss: 0.09095265, Cur Avg Loss: 0.19662779, Log Avg loss: 0.20423964, Global Avg Loss: 0.86512853, Time: 0.0209 Steps: 56830, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001486, Sample Num: 23776, Cur Loss: 0.09445405, Cur Avg Loss: 0.19634941, Log Avg loss: 0.15526009, Global Avg Loss: 0.86500364, Time: 0.0209 Steps: 56840, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001496, Sample Num: 23936, Cur Loss: 0.21156508, Cur Avg Loss: 0.19627780, Log Avg loss: 0.18563704, Global Avg Loss: 0.86488414, Time: 0.0208 Steps: 56850, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001506, Sample Num: 24096, Cur Loss: 0.21080130, Cur Avg Loss: 0.19647750, Log Avg loss: 0.22635259, Global Avg Loss: 0.86477184, Time: 0.0208 Steps: 56860, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001516, Sample Num: 24256, Cur Loss: 0.18921827, Cur Avg Loss: 0.19639796, Log Avg loss: 0.18441821, Global Avg Loss: 0.86465221, Time: 0.0209 Steps: 56870, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001526, Sample Num: 24416, Cur Loss: 0.18412906, Cur Avg Loss: 0.19619190, Log Avg loss: 0.16495459, Global Avg Loss: 0.86452919, Time: 0.0209 Steps: 56880, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001536, Sample Num: 24576, Cur Loss: 0.11888336, Cur Avg Loss: 0.19651495, Log Avg loss: 0.24581149, Global Avg Loss: 0.86442044, Time: 0.0254 Steps: 56890, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001546, Sample Num: 24736, Cur Loss: 0.23423852, Cur Avg Loss: 0.19634167, Log Avg loss: 0.16972594, Global Avg Loss: 0.86429835, Time: 0.0208 Steps: 56900, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001556, Sample Num: 24896, Cur Loss: 0.10092725, Cur Avg Loss: 0.19652999, Log Avg loss: 0.22564405, Global Avg Loss: 0.86418612, Time: 0.0208 Steps: 56910, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001566, Sample Num: 25056, Cur Loss: 0.19912562, Cur Avg Loss: 0.19598482, Log Avg loss: 0.11115647, Global Avg Loss: 0.86405383, Time: 0.0208 Steps: 56920, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001576, Sample Num: 25216, Cur Loss: 0.04058958, Cur Avg Loss: 0.19559330, Log Avg loss: 0.13428212, Global Avg Loss: 0.86392564, Time: 0.0211 Steps: 56930, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001586, Sample Num: 25376, Cur Loss: 0.24587874, Cur Avg Loss: 0.19533007, Log Avg loss: 0.15384411, Global Avg Loss: 0.86380093, Time: 0.0210 Steps: 56940, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001596, Sample Num: 25536, Cur Loss: 0.14690426, Cur Avg Loss: 0.19530577, Log Avg loss: 0.19145190, Global Avg Loss: 0.86368287, Time: 0.0211 Steps: 56950, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001606, Sample Num: 25696, Cur Loss: 0.56605065, Cur Avg Loss: 0.19568339, Log Avg loss: 0.25595136, Global Avg Loss: 0.86357618, Time: 0.0211 Steps: 56960, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001616, Sample Num: 25856, Cur Loss: 0.21415398, Cur Avg Loss: 0.19561938, Log Avg loss: 0.18533894, Global Avg Loss: 0.86345713, Time: 0.0211 Steps: 56970, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001626, Sample Num: 26016, Cur Loss: 0.08991930, Cur Avg Loss: 0.19531219, Log Avg loss: 0.14567045, Global Avg Loss: 0.86333116, Time: 0.0211 Steps: 56980, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001636, Sample Num: 26176, Cur Loss: 0.28004640, Cur Avg Loss: 0.19542748, Log Avg loss: 0.21417415, Global Avg Loss: 0.86321725, Time: 0.0211 Steps: 56990, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001646, Sample Num: 26336, Cur Loss: 0.03936984, Cur Avg Loss: 0.19491912, Log Avg loss: 0.11175142, Global Avg Loss: 0.86308541, Time: 0.0210 Steps: 57000, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001656, Sample Num: 26496, Cur Loss: 0.14257845, Cur Avg Loss: 0.19505008, Log Avg loss: 0.21660650, Global Avg Loss: 0.86297202, Time: 0.0210 Steps: 57010, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001666, Sample Num: 26656, Cur Loss: 0.09293970, Cur Avg Loss: 0.19517129, Log Avg loss: 0.21524292, Global Avg Loss: 0.86285842, Time: 0.0211 Steps: 57020, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001676, Sample Num: 26816, Cur Loss: 0.23250312, Cur Avg Loss: 0.19498614, Log Avg loss: 0.16414036, Global Avg Loss: 0.86273590, Time: 0.0210 Steps: 57030, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001686, Sample Num: 26976, Cur Loss: 0.07628397, Cur Avg Loss: 0.19490529, Log Avg loss: 0.18135450, Global Avg Loss: 0.86261644, Time: 0.0211 Steps: 57040, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001696, Sample Num: 27136, Cur Loss: 0.12001078, Cur Avg Loss: 0.19485567, Log Avg loss: 0.18649095, Global Avg Loss: 0.86249793, Time: 0.0210 Steps: 57050, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001706, Sample Num: 27296, Cur Loss: 0.15665847, Cur Avg Loss: 0.19515208, Log Avg loss: 0.24542246, Global Avg Loss: 0.86238978, Time: 0.0211 Steps: 57060, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001716, Sample Num: 27456, Cur Loss: 0.26542324, Cur Avg Loss: 0.19529611, Log Avg loss: 0.21986704, Global Avg Loss: 0.86227720, Time: 0.0211 Steps: 57070, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001726, Sample Num: 27616, Cur Loss: 0.25129351, Cur Avg Loss: 0.19560996, Log Avg loss: 0.24946687, Global Avg Loss: 0.86216984, Time: 0.0210 Steps: 57080, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001736, Sample Num: 27776, Cur Loss: 0.39566392, Cur Avg Loss: 0.19592315, Log Avg loss: 0.24998016, Global Avg Loss: 0.86206261, Time: 0.0210 Steps: 57090, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001746, Sample Num: 27936, Cur Loss: 0.10322798, Cur Avg Loss: 0.19640537, Log Avg loss: 0.28011854, Global Avg Loss: 0.86196069, Time: 0.0210 Steps: 57100, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001756, Sample Num: 28096, Cur Loss: 0.06904498, Cur Avg Loss: 0.19613249, Log Avg loss: 0.14848759, Global Avg Loss: 0.86183576, Time: 0.0210 Steps: 57110, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001766, Sample Num: 28256, Cur Loss: 0.27035874, Cur Avg Loss: 0.19627748, Log Avg loss: 0.22173739, Global Avg Loss: 0.86172370, Time: 0.0210 Steps: 57120, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001776, Sample Num: 28416, Cur Loss: 0.11974931, Cur Avg Loss: 0.19610898, Log Avg loss: 0.16635321, Global Avg Loss: 0.86160198, Time: 0.0210 Steps: 57130, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001786, Sample Num: 28576, Cur Loss: 0.09086165, Cur Avg Loss: 0.19582933, Log Avg loss: 0.14616239, Global Avg Loss: 0.86147677, Time: 0.0211 Steps: 57140, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001796, Sample Num: 28736, Cur Loss: 0.11138017, Cur Avg Loss: 0.19587586, Log Avg loss: 0.20418694, Global Avg Loss: 0.86136176, Time: 0.0248 Steps: 57150, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001806, Sample Num: 28896, Cur Loss: 0.40443987, Cur Avg Loss: 0.19609094, Log Avg loss: 0.23471813, Global Avg Loss: 0.86125213, Time: 0.0211 Steps: 57160, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001816, Sample Num: 29056, Cur Loss: 0.15917826, Cur Avg Loss: 0.19621019, Log Avg loss: 0.21774672, Global Avg Loss: 0.86113957, Time: 0.0211 Steps: 57170, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001826, Sample Num: 29216, Cur Loss: 0.25342798, Cur Avg Loss: 0.19662959, Log Avg loss: 0.27279273, Global Avg Loss: 0.86103668, Time: 0.0210 Steps: 57180, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001836, Sample Num: 29376, Cur Loss: 0.15405466, Cur Avg Loss: 0.19694340, Log Avg loss: 0.25424535, Global Avg Loss: 0.86093058, Time: 0.0210 Steps: 57190, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001846, Sample Num: 29536, Cur Loss: 0.09746684, Cur Avg Loss: 0.19667051, Log Avg loss: 0.14656779, Global Avg Loss: 0.86080569, Time: 0.0210 Steps: 57200, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001856, Sample Num: 29696, Cur Loss: 0.21911596, Cur Avg Loss: 0.19664198, Log Avg loss: 0.19137518, Global Avg Loss: 0.86068868, Time: 0.0211 Steps: 57210, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001866, Sample Num: 29856, Cur Loss: 0.16956212, Cur Avg Loss: 0.19656636, Log Avg loss: 0.18253202, Global Avg Loss: 0.86057016, Time: 0.0210 Steps: 57220, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001876, Sample Num: 30016, Cur Loss: 0.23442692, Cur Avg Loss: 0.19651053, Log Avg loss: 0.18609293, Global Avg Loss: 0.86045230, Time: 0.0211 Steps: 57230, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001886, Sample Num: 30176, Cur Loss: 0.38542590, Cur Avg Loss: 0.19639640, Log Avg loss: 0.17498525, Global Avg Loss: 0.86033255, Time: 0.0211 Steps: 57240, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001896, Sample Num: 30336, Cur Loss: 0.46596915, Cur Avg Loss: 0.19695506, Log Avg loss: 0.30231868, Global Avg Loss: 0.86023508, Time: 0.0210 Steps: 57250, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001906, Sample Num: 30496, Cur Loss: 0.90371227, Cur Avg Loss: 0.19757392, Log Avg loss: 0.31490947, Global Avg Loss: 0.86013985, Time: 0.0210 Steps: 57260, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001916, Sample Num: 30656, Cur Loss: 0.34733465, Cur Avg Loss: 0.19804458, Log Avg loss: 0.28775259, Global Avg Loss: 0.86003990, Time: 0.0210 Steps: 57270, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001926, Sample Num: 30816, Cur Loss: 0.26173788, Cur Avg Loss: 0.19834698, Log Avg loss: 0.25628752, Global Avg Loss: 0.85993450, Time: 0.0211 Steps: 57280, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001936, Sample Num: 30976, Cur Loss: 0.12201826, Cur Avg Loss: 0.19800996, Log Avg loss: 0.13309863, Global Avg Loss: 0.85980763, Time: 0.0210 Steps: 57290, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001946, Sample Num: 31136, Cur Loss: 0.43768662, Cur Avg Loss: 0.19847755, Log Avg loss: 0.28900358, Global Avg Loss: 0.85970801, Time: 0.0210 Steps: 57300, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001956, Sample Num: 31296, Cur Loss: 0.20644112, Cur Avg Loss: 0.19866238, Log Avg loss: 0.23463037, Global Avg Loss: 0.85959894, Time: 0.0210 Steps: 57310, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001966, Sample Num: 31456, Cur Loss: 0.06632950, Cur Avg Loss: 0.19839825, Log Avg loss: 0.14673395, Global Avg Loss: 0.85947457, Time: 0.0211 Steps: 57320, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001976, Sample Num: 31616, Cur Loss: 0.23537952, Cur Avg Loss: 0.19835743, Log Avg loss: 0.19033328, Global Avg Loss: 0.85935786, Time: 0.0210 Steps: 57330, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001986, Sample Num: 31776, Cur Loss: 0.10289527, Cur Avg Loss: 0.19846082, Log Avg loss: 0.21888981, Global Avg Loss: 0.85924616, Time: 0.0210 Steps: 57340, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001996, Sample Num: 31936, Cur Loss: 0.28287789, Cur Avg Loss: 0.19823477, Log Avg loss: 0.15334048, Global Avg Loss: 0.85912307, Time: 0.0211 Steps: 57350, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002006, Sample Num: 32096, Cur Loss: 0.11427698, Cur Avg Loss: 0.19840948, Log Avg loss: 0.23328235, Global Avg Loss: 0.85901397, Time: 0.0210 Steps: 57360, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002016, Sample Num: 32256, Cur Loss: 0.02792567, Cur Avg Loss: 0.19819322, Log Avg loss: 0.15481141, Global Avg Loss: 0.85889122, Time: 0.0210 Steps: 57370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002026, Sample Num: 32416, Cur Loss: 0.05678177, Cur Avg Loss: 0.19816223, Log Avg loss: 0.19191539, Global Avg Loss: 0.85877498, Time: 0.0210 Steps: 57380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002036, Sample Num: 32576, Cur Loss: 0.20383653, Cur Avg Loss: 0.19839158, Log Avg loss: 0.24485675, Global Avg Loss: 0.85866801, Time: 0.0210 Steps: 57390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002046, Sample Num: 32736, Cur Loss: 0.12924770, Cur Avg Loss: 0.19868261, Log Avg loss: 0.25793607, Global Avg Loss: 0.85856335, Time: 0.0210 Steps: 57400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002056, Sample Num: 32896, Cur Loss: 0.14583288, Cur Avg Loss: 0.19852727, Log Avg loss: 0.16674467, Global Avg Loss: 0.85844284, Time: 0.0211 Steps: 57410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002066, Sample Num: 33056, Cur Loss: 0.06936850, Cur Avg Loss: 0.19836787, Log Avg loss: 0.16559692, Global Avg Loss: 0.85832218, Time: 0.0211 Steps: 57420, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002076, Sample Num: 33216, Cur Loss: 0.24374688, Cur Avg Loss: 0.19854662, Log Avg loss: 0.23547568, Global Avg Loss: 0.85821373, Time: 0.0211 Steps: 57430, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002086, Sample Num: 33376, Cur Loss: 0.18469384, Cur Avg Loss: 0.19835663, Log Avg loss: 0.15891411, Global Avg Loss: 0.85809198, Time: 0.0212 Steps: 57440, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002096, Sample Num: 33536, Cur Loss: 0.05270467, Cur Avg Loss: 0.19841722, Log Avg loss: 0.21105753, Global Avg Loss: 0.85797936, Time: 0.0211 Steps: 57450, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002106, Sample Num: 33696, Cur Loss: 0.52482527, Cur Avg Loss: 0.19853455, Log Avg loss: 0.22312548, Global Avg Loss: 0.85786887, Time: 0.0211 Steps: 57460, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002116, Sample Num: 33856, Cur Loss: 0.11854194, Cur Avg Loss: 0.19858537, Log Avg loss: 0.20928766, Global Avg Loss: 0.85775602, Time: 0.0212 Steps: 57470, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002126, Sample Num: 34016, Cur Loss: 0.12959912, Cur Avg Loss: 0.19858470, Log Avg loss: 0.19844308, Global Avg Loss: 0.85764131, Time: 0.0210 Steps: 57480, Updated lr: 0.000046 ***** Running evaluation checkpoint-57483 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-57483 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.856388, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.347641, "eval_total_loss": 244.391881, "eval_mae": 0.405136, "eval_mse": 0.347776, "eval_r2": 0.77893, "eval_sp_statistic": 0.921046, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.929752, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.226179, "test_total_loss": 113.541645, "test_mae": 0.313211, "test_mse": 0.22625, "test_r2": 0.853976, "test_sp_statistic": 0.913625, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.949825, "test_ps_pvalue": 0.0, "lr": 4.6436225699383594e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8576009158575043, "train_cur_epoch_loss": 422.4418008569628, "train_cur_epoch_avg_loss": 0.19842264013948463, "train_cur_epoch_time": 44.85638785362244, "train_cur_epoch_avg_time": 0.02106922867713595, "epoch": 27, "step": 57483} ################################################## Training, Epoch: 0028, Batch: 000007, Sample Num: 112, Cur Loss: 0.10697262, Cur Avg Loss: 0.16295397, Log Avg loss: 0.13914139, Global Avg Loss: 0.85751634, Time: 0.0211 Steps: 57490, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000017, Sample Num: 272, Cur Loss: 0.04741198, Cur Avg Loss: 0.15955930, Log Avg loss: 0.15718303, Global Avg Loss: 0.85739454, Time: 0.0212 Steps: 57500, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000027, Sample Num: 432, Cur Loss: 0.09194310, Cur Avg Loss: 0.15423343, Log Avg loss: 0.14517944, Global Avg Loss: 0.85727070, Time: 0.0212 Steps: 57510, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000037, Sample Num: 592, Cur Loss: 0.14772350, Cur Avg Loss: 0.15095760, Log Avg loss: 0.14211289, Global Avg Loss: 0.85714636, Time: 0.0211 Steps: 57520, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000047, Sample Num: 752, Cur Loss: 0.17999169, Cur Avg Loss: 0.15193945, Log Avg loss: 0.15557229, Global Avg Loss: 0.85702442, Time: 0.0212 Steps: 57530, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000057, Sample Num: 912, Cur Loss: 0.11223640, Cur Avg Loss: 0.14909303, Log Avg loss: 0.13571484, Global Avg Loss: 0.85689906, Time: 0.0212 Steps: 57540, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000067, Sample Num: 1072, Cur Loss: 0.34311780, Cur Avg Loss: 0.18077174, Log Avg loss: 0.36134039, Global Avg Loss: 0.85681295, Time: 0.0211 Steps: 57550, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000077, Sample Num: 1232, Cur Loss: 0.20074385, Cur Avg Loss: 0.17891011, Log Avg loss: 0.16643718, Global Avg Loss: 0.85669301, Time: 0.0212 Steps: 57560, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000087, Sample Num: 1392, Cur Loss: 0.24756849, Cur Avg Loss: 0.18245323, Log Avg loss: 0.20973526, Global Avg Loss: 0.85658063, Time: 0.0210 Steps: 57570, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000097, Sample Num: 1552, Cur Loss: 0.19857492, Cur Avg Loss: 0.17860553, Log Avg loss: 0.14513057, Global Avg Loss: 0.85645707, Time: 0.0209 Steps: 57580, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000107, Sample Num: 1712, Cur Loss: 0.21147069, Cur Avg Loss: 0.17836071, Log Avg loss: 0.17598595, Global Avg Loss: 0.85633891, Time: 0.0217 Steps: 57590, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000117, Sample Num: 1872, Cur Loss: 0.05240346, Cur Avg Loss: 0.17412442, Log Avg loss: 0.12879605, Global Avg Loss: 0.85621260, Time: 0.0211 Steps: 57600, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000127, Sample Num: 2032, Cur Loss: 0.23339212, Cur Avg Loss: 0.18539460, Log Avg loss: 0.31725579, Global Avg Loss: 0.85611905, Time: 0.0210 Steps: 57610, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000137, Sample Num: 2192, Cur Loss: 0.12314713, Cur Avg Loss: 0.18520575, Log Avg loss: 0.18280738, Global Avg Loss: 0.85600220, Time: 0.0211 Steps: 57620, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000147, Sample Num: 2352, Cur Loss: 0.46552238, Cur Avg Loss: 0.18833839, Log Avg loss: 0.23125554, Global Avg Loss: 0.85589379, Time: 0.0210 Steps: 57630, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000157, Sample Num: 2512, Cur Loss: 0.16446897, Cur Avg Loss: 0.18496498, Log Avg loss: 0.13537587, Global Avg Loss: 0.85576879, Time: 0.0210 Steps: 57640, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000167, Sample Num: 2672, Cur Loss: 0.17499843, Cur Avg Loss: 0.18156759, Log Avg loss: 0.12822848, Global Avg Loss: 0.85564259, Time: 0.0210 Steps: 57650, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000177, Sample Num: 2832, Cur Loss: 0.06874906, Cur Avg Loss: 0.18219301, Log Avg loss: 0.19263758, Global Avg Loss: 0.85552760, Time: 0.0209 Steps: 57660, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000187, Sample Num: 2992, Cur Loss: 0.19271578, Cur Avg Loss: 0.18480336, Log Avg loss: 0.23100655, Global Avg Loss: 0.85541931, Time: 0.0210 Steps: 57670, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000197, Sample Num: 3152, Cur Loss: 0.13645102, Cur Avg Loss: 0.18525429, Log Avg loss: 0.19368676, Global Avg Loss: 0.85530459, Time: 0.0210 Steps: 57680, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000207, Sample Num: 3312, Cur Loss: 0.44136456, Cur Avg Loss: 0.18714389, Log Avg loss: 0.22436885, Global Avg Loss: 0.85519522, Time: 0.0209 Steps: 57690, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000217, Sample Num: 3472, Cur Loss: 0.09043381, Cur Avg Loss: 0.18528848, Log Avg loss: 0.14688158, Global Avg Loss: 0.85507246, Time: 0.0208 Steps: 57700, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000227, Sample Num: 3632, Cur Loss: 0.14456664, Cur Avg Loss: 0.18561809, Log Avg loss: 0.19277056, Global Avg Loss: 0.85495770, Time: 0.0210 Steps: 57710, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000237, Sample Num: 3792, Cur Loss: 0.08840159, Cur Avg Loss: 0.18462424, Log Avg loss: 0.16206380, Global Avg Loss: 0.85483765, Time: 0.0211 Steps: 57720, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000247, Sample Num: 3952, Cur Loss: 0.15533969, Cur Avg Loss: 0.18258221, Log Avg loss: 0.13418615, Global Avg Loss: 0.85471282, Time: 0.0209 Steps: 57730, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000257, Sample Num: 4112, Cur Loss: 0.24692747, Cur Avg Loss: 0.18213456, Log Avg loss: 0.17107756, Global Avg Loss: 0.85459442, Time: 0.0247 Steps: 57740, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000267, Sample Num: 4272, Cur Loss: 0.04676622, Cur Avg Loss: 0.18174178, Log Avg loss: 0.17164744, Global Avg Loss: 0.85447616, Time: 0.0209 Steps: 57750, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000277, Sample Num: 4432, Cur Loss: 0.15131432, Cur Avg Loss: 0.18156757, Log Avg loss: 0.17691625, Global Avg Loss: 0.85435886, Time: 0.0210 Steps: 57760, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000287, Sample Num: 4592, Cur Loss: 0.13580497, Cur Avg Loss: 0.18007197, Log Avg loss: 0.13864384, Global Avg Loss: 0.85423497, Time: 0.0209 Steps: 57770, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000297, Sample Num: 4752, Cur Loss: 0.15473759, Cur Avg Loss: 0.17954462, Log Avg loss: 0.16440970, Global Avg Loss: 0.85411558, Time: 0.0209 Steps: 57780, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000307, Sample Num: 4912, Cur Loss: 0.31256145, Cur Avg Loss: 0.17975953, Log Avg loss: 0.18614238, Global Avg Loss: 0.85399999, Time: 0.0210 Steps: 57790, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000317, Sample Num: 5072, Cur Loss: 0.03248989, Cur Avg Loss: 0.18003333, Log Avg loss: 0.18843883, Global Avg Loss: 0.85388484, Time: 0.0209 Steps: 57800, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000327, Sample Num: 5232, Cur Loss: 0.19338663, Cur Avg Loss: 0.18044713, Log Avg loss: 0.19356465, Global Avg Loss: 0.85377062, Time: 0.0210 Steps: 57810, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000337, Sample Num: 5392, Cur Loss: 0.19628388, Cur Avg Loss: 0.18020058, Log Avg loss: 0.17213850, Global Avg Loss: 0.85365273, Time: 0.0210 Steps: 57820, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000347, Sample Num: 5552, Cur Loss: 0.16512527, Cur Avg Loss: 0.18244370, Log Avg loss: 0.25803673, Global Avg Loss: 0.85354974, Time: 0.0209 Steps: 57830, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000357, Sample Num: 5712, Cur Loss: 0.51192480, Cur Avg Loss: 0.18462450, Log Avg loss: 0.26029820, Global Avg Loss: 0.85344717, Time: 0.0209 Steps: 57840, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000367, Sample Num: 5872, Cur Loss: 0.27765283, Cur Avg Loss: 0.18465254, Log Avg loss: 0.18565359, Global Avg Loss: 0.85333174, Time: 0.0210 Steps: 57850, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000377, Sample Num: 6032, Cur Loss: 0.24785531, Cur Avg Loss: 0.18400287, Log Avg loss: 0.16016002, Global Avg Loss: 0.85321193, Time: 0.0209 Steps: 57860, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000387, Sample Num: 6192, Cur Loss: 0.15474178, Cur Avg Loss: 0.18419243, Log Avg loss: 0.19133873, Global Avg Loss: 0.85309756, Time: 0.0210 Steps: 57870, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000397, Sample Num: 6352, Cur Loss: 0.10223280, Cur Avg Loss: 0.18421361, Log Avg loss: 0.18503326, Global Avg Loss: 0.85298214, Time: 0.0211 Steps: 57880, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000407, Sample Num: 6512, Cur Loss: 0.08442254, Cur Avg Loss: 0.18411670, Log Avg loss: 0.18026934, Global Avg Loss: 0.85286593, Time: 0.0210 Steps: 57890, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000417, Sample Num: 6672, Cur Loss: 0.36627203, Cur Avg Loss: 0.18367818, Log Avg loss: 0.16583034, Global Avg Loss: 0.85274728, Time: 0.0209 Steps: 57900, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000427, Sample Num: 6832, Cur Loss: 0.24496481, Cur Avg Loss: 0.18272509, Log Avg loss: 0.14298136, Global Avg Loss: 0.85262471, Time: 0.0210 Steps: 57910, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000437, Sample Num: 6992, Cur Loss: 0.13458373, Cur Avg Loss: 0.18385019, Log Avg loss: 0.23189221, Global Avg Loss: 0.85251754, Time: 0.0210 Steps: 57920, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000447, Sample Num: 7152, Cur Loss: 0.10890844, Cur Avg Loss: 0.18344789, Log Avg loss: 0.16586713, Global Avg Loss: 0.85239901, Time: 0.0211 Steps: 57930, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000457, Sample Num: 7312, Cur Loss: 0.19016212, Cur Avg Loss: 0.18493356, Log Avg loss: 0.25134303, Global Avg Loss: 0.85229527, Time: 0.0210 Steps: 57940, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000467, Sample Num: 7472, Cur Loss: 0.17011115, Cur Avg Loss: 0.18470549, Log Avg loss: 0.17428254, Global Avg Loss: 0.85217827, Time: 0.0209 Steps: 57950, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000477, Sample Num: 7632, Cur Loss: 0.15429652, Cur Avg Loss: 0.18540549, Log Avg loss: 0.21809565, Global Avg Loss: 0.85206887, Time: 0.0210 Steps: 57960, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000487, Sample Num: 7792, Cur Loss: 0.11904848, Cur Avg Loss: 0.18466484, Log Avg loss: 0.14933573, Global Avg Loss: 0.85194765, Time: 0.0210 Steps: 57970, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000497, Sample Num: 7952, Cur Loss: 0.20142269, Cur Avg Loss: 0.18347556, Log Avg loss: 0.12555781, Global Avg Loss: 0.85182237, Time: 0.0209 Steps: 57980, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000507, Sample Num: 8112, Cur Loss: 0.19020343, Cur Avg Loss: 0.18335210, Log Avg loss: 0.17721602, Global Avg Loss: 0.85170603, Time: 0.0210 Steps: 57990, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000517, Sample Num: 8272, Cur Loss: 0.04521628, Cur Avg Loss: 0.18259066, Log Avg loss: 0.14398554, Global Avg Loss: 0.85158401, Time: 0.0225 Steps: 58000, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000527, Sample Num: 8432, Cur Loss: 0.17235424, Cur Avg Loss: 0.18224838, Log Avg loss: 0.16455293, Global Avg Loss: 0.85146558, Time: 0.0209 Steps: 58010, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000537, Sample Num: 8592, Cur Loss: 0.28835064, Cur Avg Loss: 0.18286526, Log Avg loss: 0.21537472, Global Avg Loss: 0.85135595, Time: 0.0209 Steps: 58020, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000547, Sample Num: 8752, Cur Loss: 0.24075463, Cur Avg Loss: 0.18236022, Log Avg loss: 0.15523932, Global Avg Loss: 0.85123599, Time: 0.0209 Steps: 58030, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000557, Sample Num: 8912, Cur Loss: 0.13293329, Cur Avg Loss: 0.18194423, Log Avg loss: 0.15918973, Global Avg Loss: 0.85111675, Time: 0.0208 Steps: 58040, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000567, Sample Num: 9072, Cur Loss: 0.05127706, Cur Avg Loss: 0.18273061, Log Avg loss: 0.22653203, Global Avg Loss: 0.85100916, Time: 0.0208 Steps: 58050, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000577, Sample Num: 9232, Cur Loss: 0.09807305, Cur Avg Loss: 0.18322049, Log Avg loss: 0.21099653, Global Avg Loss: 0.85089893, Time: 0.0208 Steps: 58060, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000587, Sample Num: 9392, Cur Loss: 0.13373950, Cur Avg Loss: 0.18434457, Log Avg loss: 0.24920418, Global Avg Loss: 0.85079531, Time: 0.0208 Steps: 58070, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000597, Sample Num: 9552, Cur Loss: 0.08668705, Cur Avg Loss: 0.18324798, Log Avg loss: 0.11887832, Global Avg Loss: 0.85066929, Time: 0.0208 Steps: 58080, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000607, Sample Num: 9712, Cur Loss: 0.10311361, Cur Avg Loss: 0.18274990, Log Avg loss: 0.15301443, Global Avg Loss: 0.85054919, Time: 0.0208 Steps: 58090, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000617, Sample Num: 9872, Cur Loss: 0.10331361, Cur Avg Loss: 0.18266095, Log Avg loss: 0.17726135, Global Avg Loss: 0.85043331, Time: 0.0208 Steps: 58100, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000627, Sample Num: 10032, Cur Loss: 0.10542638, Cur Avg Loss: 0.18266476, Log Avg loss: 0.18290026, Global Avg Loss: 0.85031843, Time: 0.0208 Steps: 58110, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000637, Sample Num: 10192, Cur Loss: 0.11441207, Cur Avg Loss: 0.18151878, Log Avg loss: 0.10966548, Global Avg Loss: 0.85019100, Time: 0.0207 Steps: 58120, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000647, Sample Num: 10352, Cur Loss: 0.33426553, Cur Avg Loss: 0.18293172, Log Avg loss: 0.27293588, Global Avg Loss: 0.85009170, Time: 0.0208 Steps: 58130, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000657, Sample Num: 10512, Cur Loss: 0.08275669, Cur Avg Loss: 0.18374615, Log Avg loss: 0.23643995, Global Avg Loss: 0.84998615, Time: 0.0209 Steps: 58140, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000667, Sample Num: 10672, Cur Loss: 0.09073142, Cur Avg Loss: 0.18380603, Log Avg loss: 0.18774005, Global Avg Loss: 0.84987226, Time: 0.0208 Steps: 58150, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000677, Sample Num: 10832, Cur Loss: 0.55684775, Cur Avg Loss: 0.18387324, Log Avg loss: 0.18835607, Global Avg Loss: 0.84975852, Time: 0.0209 Steps: 58160, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000687, Sample Num: 10992, Cur Loss: 0.19631684, Cur Avg Loss: 0.18357702, Log Avg loss: 0.16352284, Global Avg Loss: 0.84964055, Time: 0.0208 Steps: 58170, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000697, Sample Num: 11152, Cur Loss: 0.16347656, Cur Avg Loss: 0.18278675, Log Avg loss: 0.12849564, Global Avg Loss: 0.84951660, Time: 0.0208 Steps: 58180, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000707, Sample Num: 11312, Cur Loss: 0.12038628, Cur Avg Loss: 0.18262357, Log Avg loss: 0.17124994, Global Avg Loss: 0.84940004, Time: 0.0208 Steps: 58190, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000717, Sample Num: 11472, Cur Loss: 0.08047324, Cur Avg Loss: 0.18192313, Log Avg loss: 0.13240156, Global Avg Loss: 0.84927684, Time: 0.0208 Steps: 58200, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000727, Sample Num: 11632, Cur Loss: 0.10482117, Cur Avg Loss: 0.18214772, Log Avg loss: 0.19825116, Global Avg Loss: 0.84916500, Time: 0.0209 Steps: 58210, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000737, Sample Num: 11792, Cur Loss: 0.15862924, Cur Avg Loss: 0.18199607, Log Avg loss: 0.17097136, Global Avg Loss: 0.84904852, Time: 0.0208 Steps: 58220, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000747, Sample Num: 11952, Cur Loss: 0.11057049, Cur Avg Loss: 0.18202298, Log Avg loss: 0.18400630, Global Avg Loss: 0.84893431, Time: 0.0208 Steps: 58230, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000757, Sample Num: 12112, Cur Loss: 0.13553149, Cur Avg Loss: 0.18154908, Log Avg loss: 0.14614808, Global Avg Loss: 0.84881363, Time: 0.0208 Steps: 58240, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000767, Sample Num: 12272, Cur Loss: 0.37011683, Cur Avg Loss: 0.18154114, Log Avg loss: 0.18094020, Global Avg Loss: 0.84869898, Time: 0.0208 Steps: 58250, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000777, Sample Num: 12432, Cur Loss: 0.29358172, Cur Avg Loss: 0.18244389, Log Avg loss: 0.25168497, Global Avg Loss: 0.84859650, Time: 0.0208 Steps: 58260, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000787, Sample Num: 12592, Cur Loss: 0.09838854, Cur Avg Loss: 0.18319929, Log Avg loss: 0.24189387, Global Avg Loss: 0.84849239, Time: 0.0208 Steps: 58270, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000797, Sample Num: 12752, Cur Loss: 0.26615605, Cur Avg Loss: 0.18314857, Log Avg loss: 0.17915681, Global Avg Loss: 0.84837754, Time: 0.0208 Steps: 58280, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000807, Sample Num: 12912, Cur Loss: 0.11549883, Cur Avg Loss: 0.18237124, Log Avg loss: 0.12041853, Global Avg Loss: 0.84825265, Time: 0.0208 Steps: 58290, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000817, Sample Num: 13072, Cur Loss: 0.23117180, Cur Avg Loss: 0.18227059, Log Avg loss: 0.17414768, Global Avg Loss: 0.84813702, Time: 0.0208 Steps: 58300, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000827, Sample Num: 13232, Cur Loss: 0.15619731, Cur Avg Loss: 0.18189088, Log Avg loss: 0.15086853, Global Avg Loss: 0.84801744, Time: 0.0208 Steps: 58310, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000837, Sample Num: 13392, Cur Loss: 0.10823280, Cur Avg Loss: 0.18171581, Log Avg loss: 0.16723728, Global Avg Loss: 0.84790071, Time: 0.0208 Steps: 58320, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000847, Sample Num: 13552, Cur Loss: 0.12006220, Cur Avg Loss: 0.18133465, Log Avg loss: 0.14943162, Global Avg Loss: 0.84778097, Time: 0.0208 Steps: 58330, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000857, Sample Num: 13712, Cur Loss: 0.11536953, Cur Avg Loss: 0.18203137, Log Avg loss: 0.24104349, Global Avg Loss: 0.84767697, Time: 0.0208 Steps: 58340, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000867, Sample Num: 13872, Cur Loss: 0.28514662, Cur Avg Loss: 0.18204630, Log Avg loss: 0.18332620, Global Avg Loss: 0.84756311, Time: 0.0208 Steps: 58350, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000877, Sample Num: 14032, Cur Loss: 0.29595426, Cur Avg Loss: 0.18178750, Log Avg loss: 0.15934993, Global Avg Loss: 0.84744519, Time: 0.0208 Steps: 58360, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000887, Sample Num: 14192, Cur Loss: 0.17859459, Cur Avg Loss: 0.18147938, Log Avg loss: 0.15445650, Global Avg Loss: 0.84732646, Time: 0.0208 Steps: 58370, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000897, Sample Num: 14352, Cur Loss: 0.15857647, Cur Avg Loss: 0.18137135, Log Avg loss: 0.17178919, Global Avg Loss: 0.84721075, Time: 0.0209 Steps: 58380, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000907, Sample Num: 14512, Cur Loss: 0.08821031, Cur Avg Loss: 0.18050605, Log Avg loss: 0.10288879, Global Avg Loss: 0.84708328, Time: 0.0208 Steps: 58390, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000917, Sample Num: 14672, Cur Loss: 0.05442354, Cur Avg Loss: 0.18108127, Log Avg loss: 0.23325369, Global Avg Loss: 0.84697817, Time: 0.0208 Steps: 58400, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000927, Sample Num: 14832, Cur Loss: 0.13716084, Cur Avg Loss: 0.18125874, Log Avg loss: 0.19753250, Global Avg Loss: 0.84686698, Time: 0.0208 Steps: 58410, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000937, Sample Num: 14992, Cur Loss: 0.06003432, Cur Avg Loss: 0.18096230, Log Avg loss: 0.15348294, Global Avg Loss: 0.84674829, Time: 0.0208 Steps: 58420, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000947, Sample Num: 15152, Cur Loss: 0.17738612, Cur Avg Loss: 0.18119619, Log Avg loss: 0.20311170, Global Avg Loss: 0.84663814, Time: 0.0208 Steps: 58430, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000957, Sample Num: 15312, Cur Loss: 0.29392624, Cur Avg Loss: 0.18099382, Log Avg loss: 0.16182879, Global Avg Loss: 0.84652095, Time: 0.0208 Steps: 58440, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000967, Sample Num: 15472, Cur Loss: 0.35319000, Cur Avg Loss: 0.18150896, Log Avg loss: 0.23080830, Global Avg Loss: 0.84641561, Time: 0.0208 Steps: 58450, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000977, Sample Num: 15632, Cur Loss: 0.08977176, Cur Avg Loss: 0.18198593, Log Avg loss: 0.22810888, Global Avg Loss: 0.84630985, Time: 0.0208 Steps: 58460, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000987, Sample Num: 15792, Cur Loss: 0.18847898, Cur Avg Loss: 0.18214852, Log Avg loss: 0.19803317, Global Avg Loss: 0.84619897, Time: 0.0208 Steps: 58470, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000997, Sample Num: 15952, Cur Loss: 0.08458307, Cur Avg Loss: 0.18148100, Log Avg loss: 0.11559704, Global Avg Loss: 0.84607404, Time: 0.0208 Steps: 58480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001007, Sample Num: 16112, Cur Loss: 0.09330295, Cur Avg Loss: 0.18207727, Log Avg loss: 0.24152585, Global Avg Loss: 0.84597068, Time: 0.0208 Steps: 58490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001017, Sample Num: 16272, Cur Loss: 0.18858480, Cur Avg Loss: 0.18234465, Log Avg loss: 0.20926926, Global Avg Loss: 0.84586185, Time: 0.0208 Steps: 58500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001027, Sample Num: 16432, Cur Loss: 0.07353856, Cur Avg Loss: 0.18202207, Log Avg loss: 0.14921596, Global Avg Loss: 0.84574278, Time: 0.0245 Steps: 58510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001037, Sample Num: 16592, Cur Loss: 0.11886029, Cur Avg Loss: 0.18149546, Log Avg loss: 0.12741201, Global Avg Loss: 0.84562003, Time: 0.0208 Steps: 58520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001047, Sample Num: 16752, Cur Loss: 0.32002655, Cur Avg Loss: 0.18463492, Log Avg loss: 0.51019731, Global Avg Loss: 0.84556272, Time: 0.0208 Steps: 58530, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001057, Sample Num: 16912, Cur Loss: 0.22874261, Cur Avg Loss: 0.18526455, Log Avg loss: 0.25118697, Global Avg Loss: 0.84546119, Time: 0.0208 Steps: 58540, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001067, Sample Num: 17072, Cur Loss: 0.09621009, Cur Avg Loss: 0.18490282, Log Avg loss: 0.14666845, Global Avg Loss: 0.84534184, Time: 0.0208 Steps: 58550, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001077, Sample Num: 17232, Cur Loss: 0.31408986, Cur Avg Loss: 0.18464745, Log Avg loss: 0.15739908, Global Avg Loss: 0.84522436, Time: 0.0208 Steps: 58560, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001087, Sample Num: 17392, Cur Loss: 0.09665207, Cur Avg Loss: 0.18420076, Log Avg loss: 0.13609201, Global Avg Loss: 0.84510329, Time: 0.0208 Steps: 58570, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001097, Sample Num: 17552, Cur Loss: 0.20306939, Cur Avg Loss: 0.18482527, Log Avg loss: 0.25270948, Global Avg Loss: 0.84500216, Time: 0.0208 Steps: 58580, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001107, Sample Num: 17712, Cur Loss: 0.24972191, Cur Avg Loss: 0.18546267, Log Avg loss: 0.25538547, Global Avg Loss: 0.84490153, Time: 0.0208 Steps: 58590, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001117, Sample Num: 17872, Cur Loss: 0.12063989, Cur Avg Loss: 0.18564873, Log Avg loss: 0.20624576, Global Avg Loss: 0.84479254, Time: 0.0208 Steps: 58600, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001127, Sample Num: 18032, Cur Loss: 0.03703208, Cur Avg Loss: 0.18542101, Log Avg loss: 0.15998443, Global Avg Loss: 0.84467570, Time: 0.0208 Steps: 58610, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001137, Sample Num: 18192, Cur Loss: 0.17825380, Cur Avg Loss: 0.18500919, Log Avg loss: 0.13859723, Global Avg Loss: 0.84455525, Time: 0.0208 Steps: 58620, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001147, Sample Num: 18352, Cur Loss: 0.31154382, Cur Avg Loss: 0.18560099, Log Avg loss: 0.25288913, Global Avg Loss: 0.84445434, Time: 0.0208 Steps: 58630, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001157, Sample Num: 18512, Cur Loss: 0.42316768, Cur Avg Loss: 0.18555021, Log Avg loss: 0.17972521, Global Avg Loss: 0.84434098, Time: 0.0208 Steps: 58640, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001167, Sample Num: 18672, Cur Loss: 0.31190169, Cur Avg Loss: 0.18589208, Log Avg loss: 0.22544653, Global Avg Loss: 0.84423546, Time: 0.0208 Steps: 58650, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001177, Sample Num: 18832, Cur Loss: 0.09417839, Cur Avg Loss: 0.18533458, Log Avg loss: 0.12027444, Global Avg Loss: 0.84411204, Time: 0.0208 Steps: 58660, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001187, Sample Num: 18992, Cur Loss: 0.07037185, Cur Avg Loss: 0.18534814, Log Avg loss: 0.18694385, Global Avg Loss: 0.84400003, Time: 0.0208 Steps: 58670, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001197, Sample Num: 19152, Cur Loss: 0.70648330, Cur Avg Loss: 0.18558005, Log Avg loss: 0.21310843, Global Avg Loss: 0.84389251, Time: 0.0208 Steps: 58680, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001207, Sample Num: 19312, Cur Loss: 0.26379147, Cur Avg Loss: 0.18560781, Log Avg loss: 0.18892980, Global Avg Loss: 0.84378092, Time: 0.0208 Steps: 58690, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001217, Sample Num: 19472, Cur Loss: 0.17020041, Cur Avg Loss: 0.18555432, Log Avg loss: 0.17909846, Global Avg Loss: 0.84366768, Time: 0.0208 Steps: 58700, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001227, Sample Num: 19632, Cur Loss: 0.07379976, Cur Avg Loss: 0.18559864, Log Avg loss: 0.19099237, Global Avg Loss: 0.84355651, Time: 0.0208 Steps: 58710, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001237, Sample Num: 19792, Cur Loss: 0.13544425, Cur Avg Loss: 0.18556237, Log Avg loss: 0.18111183, Global Avg Loss: 0.84344370, Time: 0.0208 Steps: 58720, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001247, Sample Num: 19952, Cur Loss: 0.03271240, Cur Avg Loss: 0.18573608, Log Avg loss: 0.20722450, Global Avg Loss: 0.84333537, Time: 0.0208 Steps: 58730, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001257, Sample Num: 20112, Cur Loss: 0.16780323, Cur Avg Loss: 0.18592818, Log Avg loss: 0.20988226, Global Avg Loss: 0.84322753, Time: 0.0208 Steps: 58740, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001267, Sample Num: 20272, Cur Loss: 0.21024531, Cur Avg Loss: 0.18575655, Log Avg loss: 0.16418289, Global Avg Loss: 0.84311195, Time: 0.0208 Steps: 58750, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001277, Sample Num: 20432, Cur Loss: 0.08322380, Cur Avg Loss: 0.18585432, Log Avg loss: 0.19824260, Global Avg Loss: 0.84300220, Time: 0.0208 Steps: 58760, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001287, Sample Num: 20592, Cur Loss: 0.18098855, Cur Avg Loss: 0.18597638, Log Avg loss: 0.20156264, Global Avg Loss: 0.84289306, Time: 0.0209 Steps: 58770, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001297, Sample Num: 20752, Cur Loss: 0.12555705, Cur Avg Loss: 0.18589439, Log Avg loss: 0.17534238, Global Avg Loss: 0.84277949, Time: 0.0209 Steps: 58780, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001307, Sample Num: 20912, Cur Loss: 0.09813941, Cur Avg Loss: 0.18610446, Log Avg loss: 0.21335095, Global Avg Loss: 0.84267243, Time: 0.0208 Steps: 58790, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001317, Sample Num: 21072, Cur Loss: 0.03069107, Cur Avg Loss: 0.18602845, Log Avg loss: 0.17609386, Global Avg Loss: 0.84255906, Time: 0.0209 Steps: 58800, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001327, Sample Num: 21232, Cur Loss: 0.27535534, Cur Avg Loss: 0.18595095, Log Avg loss: 0.17574341, Global Avg Loss: 0.84244568, Time: 0.0210 Steps: 58810, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001337, Sample Num: 21392, Cur Loss: 0.10738514, Cur Avg Loss: 0.18538790, Log Avg loss: 0.11067175, Global Avg Loss: 0.84232127, Time: 0.0209 Steps: 58820, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001347, Sample Num: 21552, Cur Loss: 0.33806652, Cur Avg Loss: 0.18519692, Log Avg loss: 0.15966283, Global Avg Loss: 0.84220523, Time: 0.0209 Steps: 58830, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001357, Sample Num: 21712, Cur Loss: 0.07739484, Cur Avg Loss: 0.18502880, Log Avg loss: 0.16238363, Global Avg Loss: 0.84208969, Time: 0.0209 Steps: 58840, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001367, Sample Num: 21872, Cur Loss: 0.19892794, Cur Avg Loss: 0.18498586, Log Avg loss: 0.17915847, Global Avg Loss: 0.84197705, Time: 0.0209 Steps: 58850, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001377, Sample Num: 22032, Cur Loss: 0.24347728, Cur Avg Loss: 0.18500663, Log Avg loss: 0.18784610, Global Avg Loss: 0.84186591, Time: 0.0210 Steps: 58860, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001387, Sample Num: 22192, Cur Loss: 0.42519775, Cur Avg Loss: 0.18495120, Log Avg loss: 0.17731775, Global Avg Loss: 0.84175303, Time: 0.0209 Steps: 58870, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001397, Sample Num: 22352, Cur Loss: 0.25154483, Cur Avg Loss: 0.18513560, Log Avg loss: 0.21071172, Global Avg Loss: 0.84164585, Time: 0.0209 Steps: 58880, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001407, Sample Num: 22512, Cur Loss: 0.31571725, Cur Avg Loss: 0.18527567, Log Avg loss: 0.20484433, Global Avg Loss: 0.84153772, Time: 0.0209 Steps: 58890, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001417, Sample Num: 22672, Cur Loss: 0.10546517, Cur Avg Loss: 0.18550499, Log Avg loss: 0.21776958, Global Avg Loss: 0.84143182, Time: 0.0209 Steps: 58900, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001427, Sample Num: 22832, Cur Loss: 0.57060039, Cur Avg Loss: 0.18610539, Log Avg loss: 0.27118200, Global Avg Loss: 0.84133502, Time: 0.0209 Steps: 58910, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001437, Sample Num: 22992, Cur Loss: 0.23088576, Cur Avg Loss: 0.18572581, Log Avg loss: 0.13156035, Global Avg Loss: 0.84121455, Time: 0.0209 Steps: 58920, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001447, Sample Num: 23152, Cur Loss: 0.10122787, Cur Avg Loss: 0.18560635, Log Avg loss: 0.16844019, Global Avg Loss: 0.84110039, Time: 0.0209 Steps: 58930, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001457, Sample Num: 23312, Cur Loss: 0.18708354, Cur Avg Loss: 0.18592434, Log Avg loss: 0.23193668, Global Avg Loss: 0.84099703, Time: 0.0209 Steps: 58940, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001467, Sample Num: 23472, Cur Loss: 0.22790782, Cur Avg Loss: 0.18627460, Log Avg loss: 0.23730781, Global Avg Loss: 0.84089463, Time: 0.0209 Steps: 58950, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001477, Sample Num: 23632, Cur Loss: 0.45975596, Cur Avg Loss: 0.18655896, Log Avg loss: 0.22827546, Global Avg Loss: 0.84079072, Time: 0.0209 Steps: 58960, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001487, Sample Num: 23792, Cur Loss: 0.07699254, Cur Avg Loss: 0.18612948, Log Avg loss: 0.12269440, Global Avg Loss: 0.84066895, Time: 0.0208 Steps: 58970, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001497, Sample Num: 23952, Cur Loss: 0.16114074, Cur Avg Loss: 0.18577380, Log Avg loss: 0.13288384, Global Avg Loss: 0.84054895, Time: 0.0209 Steps: 58980, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001507, Sample Num: 24112, Cur Loss: 0.32991070, Cur Avg Loss: 0.18612410, Log Avg loss: 0.23856452, Global Avg Loss: 0.84044690, Time: 0.0209 Steps: 58990, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001517, Sample Num: 24272, Cur Loss: 0.18153414, Cur Avg Loss: 0.18719994, Log Avg loss: 0.34932844, Global Avg Loss: 0.84036366, Time: 0.0208 Steps: 59000, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001527, Sample Num: 24432, Cur Loss: 0.15064166, Cur Avg Loss: 0.18696889, Log Avg loss: 0.15191902, Global Avg Loss: 0.84024699, Time: 0.0209 Steps: 59010, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001537, Sample Num: 24592, Cur Loss: 0.24523559, Cur Avg Loss: 0.18689730, Log Avg loss: 0.17596575, Global Avg Loss: 0.84013444, Time: 0.0245 Steps: 59020, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001547, Sample Num: 24752, Cur Loss: 0.26585808, Cur Avg Loss: 0.18692660, Log Avg loss: 0.19142926, Global Avg Loss: 0.84002454, Time: 0.0208 Steps: 59030, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001557, Sample Num: 24912, Cur Loss: 0.23589522, Cur Avg Loss: 0.18744784, Log Avg loss: 0.26808389, Global Avg Loss: 0.83992767, Time: 0.0208 Steps: 59040, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001567, Sample Num: 25072, Cur Loss: 0.10405415, Cur Avg Loss: 0.18767638, Log Avg loss: 0.22326014, Global Avg Loss: 0.83982324, Time: 0.0208 Steps: 59050, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001577, Sample Num: 25232, Cur Loss: 0.14271456, Cur Avg Loss: 0.18769052, Log Avg loss: 0.18990744, Global Avg Loss: 0.83971320, Time: 0.0208 Steps: 59060, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001587, Sample Num: 25392, Cur Loss: 0.07706586, Cur Avg Loss: 0.18726761, Log Avg loss: 0.12057456, Global Avg Loss: 0.83959145, Time: 0.0208 Steps: 59070, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001597, Sample Num: 25552, Cur Loss: 0.22753084, Cur Avg Loss: 0.18709719, Log Avg loss: 0.16005124, Global Avg Loss: 0.83947643, Time: 0.0208 Steps: 59080, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001607, Sample Num: 25712, Cur Loss: 0.16176960, Cur Avg Loss: 0.18707715, Log Avg loss: 0.18387694, Global Avg Loss: 0.83936548, Time: 0.0209 Steps: 59090, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001617, Sample Num: 25872, Cur Loss: 0.54691374, Cur Avg Loss: 0.18700460, Log Avg loss: 0.17534526, Global Avg Loss: 0.83925313, Time: 0.0208 Steps: 59100, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001627, Sample Num: 26032, Cur Loss: 0.27327621, Cur Avg Loss: 0.18706655, Log Avg loss: 0.19708334, Global Avg Loss: 0.83914449, Time: 0.0209 Steps: 59110, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001637, Sample Num: 26192, Cur Loss: 0.16521633, Cur Avg Loss: 0.18678046, Log Avg loss: 0.14023360, Global Avg Loss: 0.83902627, Time: 0.0209 Steps: 59120, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001647, Sample Num: 26352, Cur Loss: 0.14097954, Cur Avg Loss: 0.18711592, Log Avg loss: 0.24203179, Global Avg Loss: 0.83892531, Time: 0.0208 Steps: 59130, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001657, Sample Num: 26512, Cur Loss: 0.27571121, Cur Avg Loss: 0.18776842, Log Avg loss: 0.29523535, Global Avg Loss: 0.83883337, Time: 0.0209 Steps: 59140, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001667, Sample Num: 26672, Cur Loss: 0.43352753, Cur Avg Loss: 0.18795953, Log Avg loss: 0.21962561, Global Avg Loss: 0.83872869, Time: 0.0208 Steps: 59150, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001677, Sample Num: 26832, Cur Loss: 0.38309932, Cur Avg Loss: 0.18818012, Log Avg loss: 0.22495336, Global Avg Loss: 0.83862494, Time: 0.0208 Steps: 59160, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001687, Sample Num: 26992, Cur Loss: 0.45364720, Cur Avg Loss: 0.18915458, Log Avg loss: 0.35257019, Global Avg Loss: 0.83854280, Time: 0.0208 Steps: 59170, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001697, Sample Num: 27152, Cur Loss: 0.16107923, Cur Avg Loss: 0.18956652, Log Avg loss: 0.25906072, Global Avg Loss: 0.83844488, Time: 0.0208 Steps: 59180, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001707, Sample Num: 27312, Cur Loss: 0.10742889, Cur Avg Loss: 0.18953411, Log Avg loss: 0.18403451, Global Avg Loss: 0.83833432, Time: 0.0208 Steps: 59190, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001717, Sample Num: 27472, Cur Loss: 0.12609075, Cur Avg Loss: 0.18938317, Log Avg loss: 0.16361885, Global Avg Loss: 0.83822034, Time: 0.0209 Steps: 59200, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001727, Sample Num: 27632, Cur Loss: 0.06260891, Cur Avg Loss: 0.18930802, Log Avg loss: 0.17640371, Global Avg Loss: 0.83810857, Time: 0.0208 Steps: 59210, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001737, Sample Num: 27792, Cur Loss: 0.19227481, Cur Avg Loss: 0.18934578, Log Avg loss: 0.19586689, Global Avg Loss: 0.83800012, Time: 0.0209 Steps: 59220, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001747, Sample Num: 27952, Cur Loss: 0.29514357, Cur Avg Loss: 0.18917551, Log Avg loss: 0.15959942, Global Avg Loss: 0.83788558, Time: 0.0208 Steps: 59230, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001757, Sample Num: 28112, Cur Loss: 0.23397152, Cur Avg Loss: 0.18928111, Log Avg loss: 0.20772944, Global Avg Loss: 0.83777921, Time: 0.0208 Steps: 59240, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001767, Sample Num: 28272, Cur Loss: 0.23469035, Cur Avg Loss: 0.18945819, Log Avg loss: 0.22057217, Global Avg Loss: 0.83767504, Time: 0.0210 Steps: 59250, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001777, Sample Num: 28432, Cur Loss: 0.19061020, Cur Avg Loss: 0.18912425, Log Avg loss: 0.13011598, Global Avg Loss: 0.83755564, Time: 0.0208 Steps: 59260, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001787, Sample Num: 28592, Cur Loss: 0.13759692, Cur Avg Loss: 0.18903257, Log Avg loss: 0.17274207, Global Avg Loss: 0.83744347, Time: 0.0208 Steps: 59270, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001797, Sample Num: 28752, Cur Loss: 0.21406952, Cur Avg Loss: 0.18962100, Log Avg loss: 0.29477251, Global Avg Loss: 0.83735193, Time: 0.0210 Steps: 59280, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001807, Sample Num: 28912, Cur Loss: 0.14421439, Cur Avg Loss: 0.18966865, Log Avg loss: 0.19823215, Global Avg Loss: 0.83724413, Time: 0.0210 Steps: 59290, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001817, Sample Num: 29072, Cur Loss: 0.18939513, Cur Avg Loss: 0.18962894, Log Avg loss: 0.18245275, Global Avg Loss: 0.83713371, Time: 0.0210 Steps: 59300, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001827, Sample Num: 29232, Cur Loss: 0.27838451, Cur Avg Loss: 0.18963822, Log Avg loss: 0.19132464, Global Avg Loss: 0.83702483, Time: 0.0210 Steps: 59310, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001837, Sample Num: 29392, Cur Loss: 0.09218564, Cur Avg Loss: 0.18958569, Log Avg loss: 0.17998875, Global Avg Loss: 0.83691407, Time: 0.0210 Steps: 59320, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001847, Sample Num: 29552, Cur Loss: 0.17825031, Cur Avg Loss: 0.18959991, Log Avg loss: 0.19221202, Global Avg Loss: 0.83680540, Time: 0.0208 Steps: 59330, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001857, Sample Num: 29712, Cur Loss: 0.10329814, Cur Avg Loss: 0.18970264, Log Avg loss: 0.20867705, Global Avg Loss: 0.83669955, Time: 0.0209 Steps: 59340, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001867, Sample Num: 29872, Cur Loss: 0.04790005, Cur Avg Loss: 0.18965369, Log Avg loss: 0.18056286, Global Avg Loss: 0.83658900, Time: 0.0209 Steps: 59350, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001877, Sample Num: 30032, Cur Loss: 0.44310394, Cur Avg Loss: 0.18966932, Log Avg loss: 0.19258817, Global Avg Loss: 0.83648050, Time: 0.0209 Steps: 59360, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001887, Sample Num: 30192, Cur Loss: 0.11178157, Cur Avg Loss: 0.18938583, Log Avg loss: 0.13617476, Global Avg Loss: 0.83636255, Time: 0.0208 Steps: 59370, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001897, Sample Num: 30352, Cur Loss: 0.27939266, Cur Avg Loss: 0.18919367, Log Avg loss: 0.15293231, Global Avg Loss: 0.83624745, Time: 0.0208 Steps: 59380, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001907, Sample Num: 30512, Cur Loss: 0.08464841, Cur Avg Loss: 0.18903683, Log Avg loss: 0.15928457, Global Avg Loss: 0.83613347, Time: 0.0209 Steps: 59390, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001917, Sample Num: 30672, Cur Loss: 0.20843922, Cur Avg Loss: 0.18931799, Log Avg loss: 0.24293532, Global Avg Loss: 0.83603360, Time: 0.0209 Steps: 59400, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001927, Sample Num: 30832, Cur Loss: 0.18877997, Cur Avg Loss: 0.18931915, Log Avg loss: 0.18954227, Global Avg Loss: 0.83592478, Time: 0.0209 Steps: 59410, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001937, Sample Num: 30992, Cur Loss: 0.17628086, Cur Avg Loss: 0.18955503, Log Avg loss: 0.23500771, Global Avg Loss: 0.83582365, Time: 0.0209 Steps: 59420, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001947, Sample Num: 31152, Cur Loss: 0.18296632, Cur Avg Loss: 0.18980279, Log Avg loss: 0.23779545, Global Avg Loss: 0.83572303, Time: 0.0209 Steps: 59430, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001957, Sample Num: 31312, Cur Loss: 0.22972217, Cur Avg Loss: 0.18974158, Log Avg loss: 0.17782306, Global Avg Loss: 0.83561234, Time: 0.0208 Steps: 59440, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001967, Sample Num: 31472, Cur Loss: 0.07608795, Cur Avg Loss: 0.18946050, Log Avg loss: 0.13445355, Global Avg Loss: 0.83549440, Time: 0.0209 Steps: 59450, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001977, Sample Num: 31632, Cur Loss: 0.16314295, Cur Avg Loss: 0.18938522, Log Avg loss: 0.17457694, Global Avg Loss: 0.83538325, Time: 0.0209 Steps: 59460, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001987, Sample Num: 31792, Cur Loss: 0.32677928, Cur Avg Loss: 0.18962588, Log Avg loss: 0.23720413, Global Avg Loss: 0.83528266, Time: 0.0209 Steps: 59470, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001997, Sample Num: 31952, Cur Loss: 0.20401800, Cur Avg Loss: 0.18989691, Log Avg loss: 0.24375143, Global Avg Loss: 0.83518321, Time: 0.0208 Steps: 59480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002007, Sample Num: 32112, Cur Loss: 0.39287287, Cur Avg Loss: 0.19015146, Log Avg loss: 0.24098398, Global Avg Loss: 0.83508333, Time: 0.0209 Steps: 59490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002017, Sample Num: 32272, Cur Loss: 0.25992078, Cur Avg Loss: 0.19022514, Log Avg loss: 0.20501444, Global Avg Loss: 0.83497744, Time: 0.0209 Steps: 59500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002027, Sample Num: 32432, Cur Loss: 0.24659234, Cur Avg Loss: 0.19042364, Log Avg loss: 0.23045933, Global Avg Loss: 0.83487586, Time: 0.0209 Steps: 59510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002037, Sample Num: 32592, Cur Loss: 0.14784384, Cur Avg Loss: 0.19053850, Log Avg loss: 0.21382240, Global Avg Loss: 0.83477151, Time: 0.0209 Steps: 59520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002047, Sample Num: 32752, Cur Loss: 0.11268693, Cur Avg Loss: 0.19067181, Log Avg loss: 0.21782536, Global Avg Loss: 0.83466788, Time: 0.0208 Steps: 59530, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002057, Sample Num: 32912, Cur Loss: 0.06301425, Cur Avg Loss: 0.19040712, Log Avg loss: 0.13622672, Global Avg Loss: 0.83455057, Time: 0.0210 Steps: 59540, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002067, Sample Num: 33072, Cur Loss: 0.31204173, Cur Avg Loss: 0.19047916, Log Avg loss: 0.20529606, Global Avg Loss: 0.83444490, Time: 0.0209 Steps: 59550, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002077, Sample Num: 33232, Cur Loss: 0.17687985, Cur Avg Loss: 0.19056819, Log Avg loss: 0.20897186, Global Avg Loss: 0.83433989, Time: 0.0209 Steps: 59560, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002087, Sample Num: 33392, Cur Loss: 0.68869042, Cur Avg Loss: 0.19078802, Log Avg loss: 0.23644656, Global Avg Loss: 0.83423952, Time: 0.0209 Steps: 59570, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002097, Sample Num: 33552, Cur Loss: 0.16833815, Cur Avg Loss: 0.19050293, Log Avg loss: 0.13100554, Global Avg Loss: 0.83412149, Time: 0.0209 Steps: 59580, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002107, Sample Num: 33712, Cur Loss: 0.28992206, Cur Avg Loss: 0.19042708, Log Avg loss: 0.17452008, Global Avg Loss: 0.83401080, Time: 0.0209 Steps: 59590, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002117, Sample Num: 33872, Cur Loss: 0.28146136, Cur Avg Loss: 0.19051945, Log Avg loss: 0.20998209, Global Avg Loss: 0.83390609, Time: 0.0209 Steps: 59600, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002127, Sample Num: 34032, Cur Loss: 0.18006055, Cur Avg Loss: 0.19030030, Log Avg loss: 0.14390708, Global Avg Loss: 0.83379034, Time: 0.0210 Steps: 59610, Updated lr: 0.000044 ***** Running evaluation checkpoint-59612 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-59612 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.689882, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.317509, "eval_total_loss": 223.208866, "eval_mae": 0.37038, "eval_mse": 0.317635, "eval_r2": 0.79809, "eval_sp_statistic": 0.921057, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.927908, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.207728, "test_total_loss": 104.279628, "test_mae": 0.291401, "test_mse": 0.207801, "test_r2": 0.865883, "test_sp_statistic": 0.91284, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.944697, "test_ps_pvalue": 0.0, "lr": 4.441725936462779e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8337716293614464, "train_cur_epoch_loss": 405.32092325761914, "train_cur_epoch_avg_loss": 0.19038089396788124, "train_cur_epoch_time": 44.689881801605225, "train_cur_epoch_avg_time": 0.02099102010408888, "epoch": 28, "step": 59612} ################################################## Training, Epoch: 0029, Batch: 000008, Sample Num: 128, Cur Loss: 0.10113421, Cur Avg Loss: 0.11340638, Log Avg loss: 0.14594279, Global Avg Loss: 0.83367497, Time: 0.0210 Steps: 59620, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000018, Sample Num: 288, Cur Loss: 0.03994346, Cur Avg Loss: 0.21550280, Log Avg loss: 0.29717994, Global Avg Loss: 0.83358500, Time: 0.0210 Steps: 59630, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000028, Sample Num: 448, Cur Loss: 0.08485791, Cur Avg Loss: 0.18425939, Log Avg loss: 0.12802124, Global Avg Loss: 0.83346669, Time: 0.0210 Steps: 59640, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000038, Sample Num: 608, Cur Loss: 0.11706463, Cur Avg Loss: 0.18304394, Log Avg loss: 0.17964069, Global Avg Loss: 0.83335708, Time: 0.0210 Steps: 59650, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000048, Sample Num: 768, Cur Loss: 0.13232516, Cur Avg Loss: 0.18169152, Log Avg loss: 0.17655234, Global Avg Loss: 0.83324699, Time: 0.0210 Steps: 59660, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000058, Sample Num: 928, Cur Loss: 0.18672666, Cur Avg Loss: 0.17755162, Log Avg loss: 0.15768010, Global Avg Loss: 0.83313378, Time: 0.0210 Steps: 59670, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000068, Sample Num: 1088, Cur Loss: 0.09884954, Cur Avg Loss: 0.17243469, Log Avg loss: 0.14275649, Global Avg Loss: 0.83301810, Time: 0.0210 Steps: 59680, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000078, Sample Num: 1248, Cur Loss: 0.24897110, Cur Avg Loss: 0.16818480, Log Avg loss: 0.13928553, Global Avg Loss: 0.83290187, Time: 0.0210 Steps: 59690, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000088, Sample Num: 1408, Cur Loss: 0.09282604, Cur Avg Loss: 0.16853672, Log Avg loss: 0.17128173, Global Avg Loss: 0.83279105, Time: 0.0210 Steps: 59700, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000098, Sample Num: 1568, Cur Loss: 0.41927385, Cur Avg Loss: 0.17620144, Log Avg loss: 0.24365094, Global Avg Loss: 0.83269238, Time: 0.0210 Steps: 59710, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000108, Sample Num: 1728, Cur Loss: 0.09660198, Cur Avg Loss: 0.17900922, Log Avg loss: 0.20652546, Global Avg Loss: 0.83258753, Time: 0.0210 Steps: 59720, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000118, Sample Num: 1888, Cur Loss: 0.13082656, Cur Avg Loss: 0.17659158, Log Avg loss: 0.15048107, Global Avg Loss: 0.83247333, Time: 0.0210 Steps: 59730, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000128, Sample Num: 2048, Cur Loss: 0.20556241, Cur Avg Loss: 0.17905077, Log Avg loss: 0.20806928, Global Avg Loss: 0.83236881, Time: 0.0210 Steps: 59740, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000138, Sample Num: 2208, Cur Loss: 0.10937318, Cur Avg Loss: 0.17680073, Log Avg loss: 0.14800018, Global Avg Loss: 0.83225427, Time: 0.0210 Steps: 59750, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000148, Sample Num: 2368, Cur Loss: 0.49159300, Cur Avg Loss: 0.18459764, Log Avg loss: 0.29219498, Global Avg Loss: 0.83216390, Time: 0.0210 Steps: 59760, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000158, Sample Num: 2528, Cur Loss: 0.12185486, Cur Avg Loss: 0.18286333, Log Avg loss: 0.15719556, Global Avg Loss: 0.83205098, Time: 0.0210 Steps: 59770, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000168, Sample Num: 2688, Cur Loss: 0.13347098, Cur Avg Loss: 0.18312422, Log Avg loss: 0.18724622, Global Avg Loss: 0.83194311, Time: 0.0209 Steps: 59780, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000178, Sample Num: 2848, Cur Loss: 0.47138134, Cur Avg Loss: 0.18454285, Log Avg loss: 0.20837583, Global Avg Loss: 0.83183882, Time: 0.0209 Steps: 59790, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000188, Sample Num: 3008, Cur Loss: 0.17336106, Cur Avg Loss: 0.18321188, Log Avg loss: 0.15952059, Global Avg Loss: 0.83172639, Time: 0.0210 Steps: 59800, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000198, Sample Num: 3168, Cur Loss: 0.46781680, Cur Avg Loss: 0.18506892, Log Avg loss: 0.21998145, Global Avg Loss: 0.83162411, Time: 0.0210 Steps: 59810, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000208, Sample Num: 3328, Cur Loss: 0.26453349, Cur Avg Loss: 0.18504545, Log Avg loss: 0.18458069, Global Avg Loss: 0.83151594, Time: 0.0210 Steps: 59820, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000218, Sample Num: 3488, Cur Loss: 0.17072663, Cur Avg Loss: 0.18497228, Log Avg loss: 0.18345028, Global Avg Loss: 0.83140763, Time: 0.0210 Steps: 59830, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000228, Sample Num: 3648, Cur Loss: 0.05051470, Cur Avg Loss: 0.18242263, Log Avg loss: 0.12684022, Global Avg Loss: 0.83128989, Time: 0.0210 Steps: 59840, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000238, Sample Num: 3808, Cur Loss: 0.05185500, Cur Avg Loss: 0.18077041, Log Avg loss: 0.14309994, Global Avg Loss: 0.83117490, Time: 0.0210 Steps: 59850, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000248, Sample Num: 3968, Cur Loss: 0.08447589, Cur Avg Loss: 0.17950810, Log Avg loss: 0.14946509, Global Avg Loss: 0.83106102, Time: 0.0211 Steps: 59860, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000258, Sample Num: 4128, Cur Loss: 0.08835790, Cur Avg Loss: 0.17793311, Log Avg loss: 0.13887323, Global Avg Loss: 0.83094540, Time: 0.0247 Steps: 59870, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000268, Sample Num: 4288, Cur Loss: 0.15665202, Cur Avg Loss: 0.17921034, Log Avg loss: 0.21216307, Global Avg Loss: 0.83084206, Time: 0.0209 Steps: 59880, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000278, Sample Num: 4448, Cur Loss: 0.27978355, Cur Avg Loss: 0.17848765, Log Avg loss: 0.15911930, Global Avg Loss: 0.83072990, Time: 0.0209 Steps: 59890, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000288, Sample Num: 4608, Cur Loss: 0.18158764, Cur Avg Loss: 0.18040976, Log Avg loss: 0.23384462, Global Avg Loss: 0.83063026, Time: 0.0209 Steps: 59900, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000298, Sample Num: 4768, Cur Loss: 0.03744707, Cur Avg Loss: 0.17976733, Log Avg loss: 0.16126542, Global Avg Loss: 0.83051853, Time: 0.0209 Steps: 59910, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000308, Sample Num: 4928, Cur Loss: 0.05650375, Cur Avg Loss: 0.17790967, Log Avg loss: 0.12255127, Global Avg Loss: 0.83040038, Time: 0.0209 Steps: 59920, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000318, Sample Num: 5088, Cur Loss: 0.17364204, Cur Avg Loss: 0.17701548, Log Avg loss: 0.14947444, Global Avg Loss: 0.83028676, Time: 0.0209 Steps: 59930, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000328, Sample Num: 5248, Cur Loss: 0.07322633, Cur Avg Loss: 0.17744657, Log Avg loss: 0.19115518, Global Avg Loss: 0.83018013, Time: 0.0209 Steps: 59940, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000338, Sample Num: 5408, Cur Loss: 0.19104062, Cur Avg Loss: 0.17812222, Log Avg loss: 0.20028359, Global Avg Loss: 0.83007506, Time: 0.0209 Steps: 59950, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000348, Sample Num: 5568, Cur Loss: 0.21990240, Cur Avg Loss: 0.17977862, Log Avg loss: 0.23576489, Global Avg Loss: 0.82997594, Time: 0.0209 Steps: 59960, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000358, Sample Num: 5728, Cur Loss: 0.60178995, Cur Avg Loss: 0.18160494, Log Avg loss: 0.24516075, Global Avg Loss: 0.82987842, Time: 0.0209 Steps: 59970, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000368, Sample Num: 5888, Cur Loss: 0.18482083, Cur Avg Loss: 0.18126256, Log Avg loss: 0.16900567, Global Avg Loss: 0.82976824, Time: 0.0209 Steps: 59980, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000378, Sample Num: 6048, Cur Loss: 0.11172670, Cur Avg Loss: 0.18031256, Log Avg loss: 0.14535243, Global Avg Loss: 0.82965415, Time: 0.0209 Steps: 59990, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000388, Sample Num: 6208, Cur Loss: 0.30540502, Cur Avg Loss: 0.18092162, Log Avg loss: 0.20394417, Global Avg Loss: 0.82954987, Time: 0.0209 Steps: 60000, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000398, Sample Num: 6368, Cur Loss: 0.20527431, Cur Avg Loss: 0.18082981, Log Avg loss: 0.17726753, Global Avg Loss: 0.82944117, Time: 0.0209 Steps: 60010, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000408, Sample Num: 6528, Cur Loss: 0.33016291, Cur Avg Loss: 0.18102955, Log Avg loss: 0.18897929, Global Avg Loss: 0.82933446, Time: 0.0209 Steps: 60020, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000418, Sample Num: 6688, Cur Loss: 0.18070701, Cur Avg Loss: 0.18169043, Log Avg loss: 0.20865411, Global Avg Loss: 0.82923107, Time: 0.0209 Steps: 60030, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000428, Sample Num: 6848, Cur Loss: 0.32992566, Cur Avg Loss: 0.18191054, Log Avg loss: 0.19111106, Global Avg Loss: 0.82912478, Time: 0.0209 Steps: 60040, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000438, Sample Num: 7008, Cur Loss: 0.15212226, Cur Avg Loss: 0.18229137, Log Avg loss: 0.19859107, Global Avg Loss: 0.82901978, Time: 0.0209 Steps: 60050, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000448, Sample Num: 7168, Cur Loss: 0.07804453, Cur Avg Loss: 0.18186669, Log Avg loss: 0.16326578, Global Avg Loss: 0.82890894, Time: 0.0209 Steps: 60060, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000458, Sample Num: 7328, Cur Loss: 0.16891740, Cur Avg Loss: 0.18208785, Log Avg loss: 0.19199560, Global Avg Loss: 0.82880291, Time: 0.0209 Steps: 60070, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000468, Sample Num: 7488, Cur Loss: 0.30553380, Cur Avg Loss: 0.18456139, Log Avg loss: 0.29784970, Global Avg Loss: 0.82871453, Time: 0.0209 Steps: 60080, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000478, Sample Num: 7648, Cur Loss: 0.18002343, Cur Avg Loss: 0.18357840, Log Avg loss: 0.13757456, Global Avg Loss: 0.82859951, Time: 0.0209 Steps: 60090, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000488, Sample Num: 7808, Cur Loss: 0.29378676, Cur Avg Loss: 0.18332480, Log Avg loss: 0.17120246, Global Avg Loss: 0.82849013, Time: 0.0209 Steps: 60100, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000498, Sample Num: 7968, Cur Loss: 0.09928794, Cur Avg Loss: 0.18227421, Log Avg loss: 0.13100562, Global Avg Loss: 0.82837410, Time: 0.0210 Steps: 60110, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000508, Sample Num: 8128, Cur Loss: 0.24305074, Cur Avg Loss: 0.18321869, Log Avg loss: 0.23025361, Global Avg Loss: 0.82827461, Time: 0.0209 Steps: 60120, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000518, Sample Num: 8288, Cur Loss: 0.42608497, Cur Avg Loss: 0.18295558, Log Avg loss: 0.16958990, Global Avg Loss: 0.82816507, Time: 0.0246 Steps: 60130, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000528, Sample Num: 8448, Cur Loss: 0.11838681, Cur Avg Loss: 0.18240550, Log Avg loss: 0.15391107, Global Avg Loss: 0.82805295, Time: 0.0210 Steps: 60140, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000538, Sample Num: 8608, Cur Loss: 0.29654622, Cur Avg Loss: 0.18327002, Log Avg loss: 0.22891691, Global Avg Loss: 0.82795334, Time: 0.0209 Steps: 60150, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000548, Sample Num: 8768, Cur Loss: 0.27519900, Cur Avg Loss: 0.18391417, Log Avg loss: 0.21856922, Global Avg Loss: 0.82785205, Time: 0.0209 Steps: 60160, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000558, Sample Num: 8928, Cur Loss: 0.14963672, Cur Avg Loss: 0.18342765, Log Avg loss: 0.15676642, Global Avg Loss: 0.82774052, Time: 0.0209 Steps: 60170, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000568, Sample Num: 9088, Cur Loss: 0.21134862, Cur Avg Loss: 0.18327398, Log Avg loss: 0.17469934, Global Avg Loss: 0.82763200, Time: 0.0208 Steps: 60180, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000578, Sample Num: 9248, Cur Loss: 0.09911277, Cur Avg Loss: 0.18392640, Log Avg loss: 0.22098384, Global Avg Loss: 0.82753121, Time: 0.0209 Steps: 60190, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000588, Sample Num: 9408, Cur Loss: 0.15141600, Cur Avg Loss: 0.18328421, Log Avg loss: 0.14616571, Global Avg Loss: 0.82741803, Time: 0.0209 Steps: 60200, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000598, Sample Num: 9568, Cur Loss: 0.13667278, Cur Avg Loss: 0.18285828, Log Avg loss: 0.15781358, Global Avg Loss: 0.82730682, Time: 0.0209 Steps: 60210, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000608, Sample Num: 9728, Cur Loss: 0.08682041, Cur Avg Loss: 0.18232699, Log Avg loss: 0.15055571, Global Avg Loss: 0.82719444, Time: 0.0209 Steps: 60220, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000618, Sample Num: 9888, Cur Loss: 0.13495849, Cur Avg Loss: 0.18336011, Log Avg loss: 0.24617402, Global Avg Loss: 0.82709797, Time: 0.0209 Steps: 60230, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000628, Sample Num: 10048, Cur Loss: 0.05985225, Cur Avg Loss: 0.18230625, Log Avg loss: 0.11717727, Global Avg Loss: 0.82698012, Time: 0.0209 Steps: 60240, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000638, Sample Num: 10208, Cur Loss: 0.16426866, Cur Avg Loss: 0.18179426, Log Avg loss: 0.14964140, Global Avg Loss: 0.82686770, Time: 0.0209 Steps: 60250, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000648, Sample Num: 10368, Cur Loss: 0.07108594, Cur Avg Loss: 0.18153553, Log Avg loss: 0.16502866, Global Avg Loss: 0.82675787, Time: 0.0209 Steps: 60260, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000658, Sample Num: 10528, Cur Loss: 0.14622207, Cur Avg Loss: 0.18136805, Log Avg loss: 0.17051551, Global Avg Loss: 0.82664899, Time: 0.0209 Steps: 60270, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000668, Sample Num: 10688, Cur Loss: 0.17358874, Cur Avg Loss: 0.18080897, Log Avg loss: 0.14402120, Global Avg Loss: 0.82653575, Time: 0.0209 Steps: 60280, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000678, Sample Num: 10848, Cur Loss: 0.26483184, Cur Avg Loss: 0.18085150, Log Avg loss: 0.18369228, Global Avg Loss: 0.82642912, Time: 0.0210 Steps: 60290, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000688, Sample Num: 11008, Cur Loss: 0.22672424, Cur Avg Loss: 0.18179735, Log Avg loss: 0.24592640, Global Avg Loss: 0.82633285, Time: 0.0209 Steps: 60300, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000698, Sample Num: 11168, Cur Loss: 0.06354678, Cur Avg Loss: 0.18075443, Log Avg loss: 0.10900120, Global Avg Loss: 0.82621391, Time: 0.0209 Steps: 60310, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000708, Sample Num: 11328, Cur Loss: 0.06456370, Cur Avg Loss: 0.18030362, Log Avg loss: 0.14883757, Global Avg Loss: 0.82610161, Time: 0.0209 Steps: 60320, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000718, Sample Num: 11488, Cur Loss: 0.05103451, Cur Avg Loss: 0.18043077, Log Avg loss: 0.18943251, Global Avg Loss: 0.82599608, Time: 0.0209 Steps: 60330, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000728, Sample Num: 11648, Cur Loss: 0.22261900, Cur Avg Loss: 0.18040218, Log Avg loss: 0.17834976, Global Avg Loss: 0.82588875, Time: 0.0209 Steps: 60340, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000738, Sample Num: 11808, Cur Loss: 0.16358933, Cur Avg Loss: 0.18046934, Log Avg loss: 0.18535837, Global Avg Loss: 0.82578261, Time: 0.0209 Steps: 60350, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000748, Sample Num: 11968, Cur Loss: 0.07364181, Cur Avg Loss: 0.18057401, Log Avg loss: 0.18829856, Global Avg Loss: 0.82567700, Time: 0.0209 Steps: 60360, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000758, Sample Num: 12128, Cur Loss: 0.02813498, Cur Avg Loss: 0.18058379, Log Avg loss: 0.18131535, Global Avg Loss: 0.82557026, Time: 0.0209 Steps: 60370, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000768, Sample Num: 12288, Cur Loss: 0.27857035, Cur Avg Loss: 0.18070186, Log Avg loss: 0.18965140, Global Avg Loss: 0.82546495, Time: 0.0254 Steps: 60380, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000778, Sample Num: 12448, Cur Loss: 0.07326584, Cur Avg Loss: 0.18063184, Log Avg loss: 0.17525422, Global Avg Loss: 0.82535728, Time: 0.0209 Steps: 60390, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000788, Sample Num: 12608, Cur Loss: 0.22568494, Cur Avg Loss: 0.18150262, Log Avg loss: 0.24924963, Global Avg Loss: 0.82526189, Time: 0.0209 Steps: 60400, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000798, Sample Num: 12768, Cur Loss: 0.32747209, Cur Avg Loss: 0.18209043, Log Avg loss: 0.22840970, Global Avg Loss: 0.82516309, Time: 0.0209 Steps: 60410, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000808, Sample Num: 12928, Cur Loss: 0.76520628, Cur Avg Loss: 0.18261048, Log Avg loss: 0.22411087, Global Avg Loss: 0.82506362, Time: 0.0210 Steps: 60420, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000818, Sample Num: 13088, Cur Loss: 0.07120323, Cur Avg Loss: 0.18267777, Log Avg loss: 0.18811457, Global Avg Loss: 0.82495821, Time: 0.0209 Steps: 60430, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000828, Sample Num: 13248, Cur Loss: 0.40621606, Cur Avg Loss: 0.18282601, Log Avg loss: 0.19495169, Global Avg Loss: 0.82485398, Time: 0.0209 Steps: 60440, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000838, Sample Num: 13408, Cur Loss: 0.17759441, Cur Avg Loss: 0.18277888, Log Avg loss: 0.17887700, Global Avg Loss: 0.82474711, Time: 0.0209 Steps: 60450, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000848, Sample Num: 13568, Cur Loss: 0.32225880, Cur Avg Loss: 0.18251742, Log Avg loss: 0.16060676, Global Avg Loss: 0.82463727, Time: 0.0209 Steps: 60460, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000858, Sample Num: 13728, Cur Loss: 0.14075577, Cur Avg Loss: 0.18230749, Log Avg loss: 0.16450586, Global Avg Loss: 0.82452810, Time: 0.0209 Steps: 60470, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000868, Sample Num: 13888, Cur Loss: 0.05537075, Cur Avg Loss: 0.18156522, Log Avg loss: 0.11787831, Global Avg Loss: 0.82441126, Time: 0.0210 Steps: 60480, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000878, Sample Num: 14048, Cur Loss: 0.25504801, Cur Avg Loss: 0.18189685, Log Avg loss: 0.21068211, Global Avg Loss: 0.82430980, Time: 0.0209 Steps: 60490, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000888, Sample Num: 14208, Cur Loss: 0.19029517, Cur Avg Loss: 0.18228357, Log Avg loss: 0.21623767, Global Avg Loss: 0.82420929, Time: 0.0209 Steps: 60500, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000898, Sample Num: 14368, Cur Loss: 0.17946762, Cur Avg Loss: 0.18222575, Log Avg loss: 0.17709143, Global Avg Loss: 0.82410235, Time: 0.0209 Steps: 60510, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000908, Sample Num: 14528, Cur Loss: 0.04008469, Cur Avg Loss: 0.18217899, Log Avg loss: 0.17797958, Global Avg Loss: 0.82399559, Time: 0.0210 Steps: 60520, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000918, Sample Num: 14688, Cur Loss: 0.18432289, Cur Avg Loss: 0.18221640, Log Avg loss: 0.18561391, Global Avg Loss: 0.82389012, Time: 0.0210 Steps: 60530, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000928, Sample Num: 14848, Cur Loss: 0.08115826, Cur Avg Loss: 0.18279248, Log Avg loss: 0.23567649, Global Avg Loss: 0.82379296, Time: 0.0210 Steps: 60540, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000938, Sample Num: 15008, Cur Loss: 0.30977702, Cur Avg Loss: 0.18286498, Log Avg loss: 0.18959317, Global Avg Loss: 0.82368822, Time: 0.0210 Steps: 60550, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000948, Sample Num: 15168, Cur Loss: 0.03498601, Cur Avg Loss: 0.18271611, Log Avg loss: 0.16875146, Global Avg Loss: 0.82358007, Time: 0.0210 Steps: 60560, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000958, Sample Num: 15328, Cur Loss: 0.17243040, Cur Avg Loss: 0.18291715, Log Avg loss: 0.20197584, Global Avg Loss: 0.82347745, Time: 0.0210 Steps: 60570, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000968, Sample Num: 15488, Cur Loss: 0.37170985, Cur Avg Loss: 0.18322903, Log Avg loss: 0.21310704, Global Avg Loss: 0.82337669, Time: 0.0209 Steps: 60580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000978, Sample Num: 15648, Cur Loss: 0.15437017, Cur Avg Loss: 0.18331961, Log Avg loss: 0.19208775, Global Avg Loss: 0.82327250, Time: 0.0210 Steps: 60590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000988, Sample Num: 15808, Cur Loss: 0.12365133, Cur Avg Loss: 0.18343246, Log Avg loss: 0.19446986, Global Avg Loss: 0.82316874, Time: 0.0210 Steps: 60600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000998, Sample Num: 15968, Cur Loss: 0.12131190, Cur Avg Loss: 0.18334165, Log Avg loss: 0.17436905, Global Avg Loss: 0.82306170, Time: 0.0209 Steps: 60610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001008, Sample Num: 16128, Cur Loss: 0.08469412, Cur Avg Loss: 0.18270525, Log Avg loss: 0.11919227, Global Avg Loss: 0.82294558, Time: 0.0210 Steps: 60620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001018, Sample Num: 16288, Cur Loss: 0.31232062, Cur Avg Loss: 0.18277023, Log Avg loss: 0.18932017, Global Avg Loss: 0.82284108, Time: 0.0210 Steps: 60630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001028, Sample Num: 16448, Cur Loss: 0.38374260, Cur Avg Loss: 0.18277549, Log Avg loss: 0.18331142, Global Avg Loss: 0.82273561, Time: 0.0247 Steps: 60640, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001038, Sample Num: 16608, Cur Loss: 0.13779469, Cur Avg Loss: 0.18237046, Log Avg loss: 0.14073287, Global Avg Loss: 0.82262316, Time: 0.0210 Steps: 60650, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001048, Sample Num: 16768, Cur Loss: 0.02212649, Cur Avg Loss: 0.18334854, Log Avg loss: 0.28487333, Global Avg Loss: 0.82253451, Time: 0.0209 Steps: 60660, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001058, Sample Num: 16928, Cur Loss: 0.20751105, Cur Avg Loss: 0.18352425, Log Avg loss: 0.20193935, Global Avg Loss: 0.82243222, Time: 0.0210 Steps: 60670, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001068, Sample Num: 17088, Cur Loss: 0.31933853, Cur Avg Loss: 0.18336394, Log Avg loss: 0.16640320, Global Avg Loss: 0.82232411, Time: 0.0210 Steps: 60680, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001078, Sample Num: 17248, Cur Loss: 0.07041992, Cur Avg Loss: 0.18351464, Log Avg loss: 0.19960957, Global Avg Loss: 0.82222151, Time: 0.0209 Steps: 60690, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001088, Sample Num: 17408, Cur Loss: 0.13704965, Cur Avg Loss: 0.18395184, Log Avg loss: 0.23108184, Global Avg Loss: 0.82212412, Time: 0.0210 Steps: 60700, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001098, Sample Num: 17568, Cur Loss: 0.19546580, Cur Avg Loss: 0.18369931, Log Avg loss: 0.15622339, Global Avg Loss: 0.82201443, Time: 0.0209 Steps: 60710, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001108, Sample Num: 17728, Cur Loss: 0.03306221, Cur Avg Loss: 0.18311284, Log Avg loss: 0.11871901, Global Avg Loss: 0.82189861, Time: 0.0209 Steps: 60720, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001118, Sample Num: 17888, Cur Loss: 0.04194305, Cur Avg Loss: 0.18219494, Log Avg loss: 0.08049087, Global Avg Loss: 0.82177652, Time: 0.0209 Steps: 60730, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001128, Sample Num: 18048, Cur Loss: 0.13326125, Cur Avg Loss: 0.18143889, Log Avg loss: 0.09691331, Global Avg Loss: 0.82165719, Time: 0.0209 Steps: 60740, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001138, Sample Num: 18208, Cur Loss: 0.11369642, Cur Avg Loss: 0.18121842, Log Avg loss: 0.15634867, Global Avg Loss: 0.82154767, Time: 0.0210 Steps: 60750, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001148, Sample Num: 18368, Cur Loss: 0.35460076, Cur Avg Loss: 0.18140466, Log Avg loss: 0.20259961, Global Avg Loss: 0.82144580, Time: 0.0210 Steps: 60760, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001158, Sample Num: 18528, Cur Loss: 0.12572210, Cur Avg Loss: 0.18132213, Log Avg loss: 0.17184747, Global Avg Loss: 0.82133891, Time: 0.0210 Steps: 60770, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001168, Sample Num: 18688, Cur Loss: 0.09479620, Cur Avg Loss: 0.18181470, Log Avg loss: 0.23885362, Global Avg Loss: 0.82124307, Time: 0.0209 Steps: 60780, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001178, Sample Num: 18848, Cur Loss: 0.27381849, Cur Avg Loss: 0.18189269, Log Avg loss: 0.19100208, Global Avg Loss: 0.82113940, Time: 0.0209 Steps: 60790, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001188, Sample Num: 19008, Cur Loss: 0.43872839, Cur Avg Loss: 0.18292955, Log Avg loss: 0.30507177, Global Avg Loss: 0.82105452, Time: 0.0211 Steps: 60800, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001198, Sample Num: 19168, Cur Loss: 0.10073124, Cur Avg Loss: 0.18276798, Log Avg loss: 0.16357419, Global Avg Loss: 0.82094640, Time: 0.0210 Steps: 60810, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001208, Sample Num: 19328, Cur Loss: 0.19094482, Cur Avg Loss: 0.18271939, Log Avg loss: 0.17689815, Global Avg Loss: 0.82084050, Time: 0.0210 Steps: 60820, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001218, Sample Num: 19488, Cur Loss: 0.14648983, Cur Avg Loss: 0.18322431, Log Avg loss: 0.24421785, Global Avg Loss: 0.82074571, Time: 0.0209 Steps: 60830, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001228, Sample Num: 19648, Cur Loss: 0.41084397, Cur Avg Loss: 0.18293395, Log Avg loss: 0.14756861, Global Avg Loss: 0.82063506, Time: 0.0210 Steps: 60840, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001238, Sample Num: 19808, Cur Loss: 0.17395440, Cur Avg Loss: 0.18366154, Log Avg loss: 0.27300893, Global Avg Loss: 0.82054507, Time: 0.0209 Steps: 60850, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001248, Sample Num: 19968, Cur Loss: 0.18011361, Cur Avg Loss: 0.18380533, Log Avg loss: 0.20160746, Global Avg Loss: 0.82044337, Time: 0.0210 Steps: 60860, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001258, Sample Num: 20128, Cur Loss: 0.28505951, Cur Avg Loss: 0.18410544, Log Avg loss: 0.22155851, Global Avg Loss: 0.82034498, Time: 0.0209 Steps: 60870, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001268, Sample Num: 20288, Cur Loss: 0.17574953, Cur Avg Loss: 0.18407727, Log Avg loss: 0.18053392, Global Avg Loss: 0.82023989, Time: 0.0210 Steps: 60880, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001278, Sample Num: 20448, Cur Loss: 0.20566945, Cur Avg Loss: 0.18397753, Log Avg loss: 0.17133080, Global Avg Loss: 0.82013332, Time: 0.0210 Steps: 60890, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001288, Sample Num: 20608, Cur Loss: 0.28723183, Cur Avg Loss: 0.18432175, Log Avg loss: 0.22831298, Global Avg Loss: 0.82003614, Time: 0.0210 Steps: 60900, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001298, Sample Num: 20768, Cur Loss: 0.09501021, Cur Avg Loss: 0.18508845, Log Avg loss: 0.28383893, Global Avg Loss: 0.81994811, Time: 0.0209 Steps: 60910, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001308, Sample Num: 20928, Cur Loss: 0.12344299, Cur Avg Loss: 0.18599248, Log Avg loss: 0.30333590, Global Avg Loss: 0.81986330, Time: 0.0209 Steps: 60920, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001318, Sample Num: 21088, Cur Loss: 0.11641355, Cur Avg Loss: 0.18588736, Log Avg loss: 0.17213750, Global Avg Loss: 0.81975700, Time: 0.0208 Steps: 60930, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001328, Sample Num: 21248, Cur Loss: 0.12311727, Cur Avg Loss: 0.18551126, Log Avg loss: 0.13594110, Global Avg Loss: 0.81964479, Time: 0.0209 Steps: 60940, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001338, Sample Num: 21408, Cur Loss: 0.12686272, Cur Avg Loss: 0.18539041, Log Avg loss: 0.16934127, Global Avg Loss: 0.81953809, Time: 0.0209 Steps: 60950, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001348, Sample Num: 21568, Cur Loss: 0.21032096, Cur Avg Loss: 0.18575874, Log Avg loss: 0.23504203, Global Avg Loss: 0.81944221, Time: 0.0212 Steps: 60960, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001358, Sample Num: 21728, Cur Loss: 0.16644679, Cur Avg Loss: 0.18547149, Log Avg loss: 0.14674986, Global Avg Loss: 0.81933188, Time: 0.0211 Steps: 60970, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001368, Sample Num: 21888, Cur Loss: 0.17443103, Cur Avg Loss: 0.18511771, Log Avg loss: 0.13707456, Global Avg Loss: 0.81922000, Time: 0.0209 Steps: 60980, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001378, Sample Num: 22048, Cur Loss: 0.27319622, Cur Avg Loss: 0.18537341, Log Avg loss: 0.22035312, Global Avg Loss: 0.81912181, Time: 0.0209 Steps: 60990, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001388, Sample Num: 22208, Cur Loss: 0.19067593, Cur Avg Loss: 0.18576567, Log Avg loss: 0.23981881, Global Avg Loss: 0.81902684, Time: 0.0209 Steps: 61000, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001398, Sample Num: 22368, Cur Loss: 0.17618060, Cur Avg Loss: 0.18550966, Log Avg loss: 0.14997574, Global Avg Loss: 0.81891718, Time: 0.0209 Steps: 61010, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001408, Sample Num: 22528, Cur Loss: 0.07946568, Cur Avg Loss: 0.18562850, Log Avg loss: 0.20224178, Global Avg Loss: 0.81881611, Time: 0.0208 Steps: 61020, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001418, Sample Num: 22688, Cur Loss: 0.16512500, Cur Avg Loss: 0.18568876, Log Avg loss: 0.19417398, Global Avg Loss: 0.81871376, Time: 0.0208 Steps: 61030, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001428, Sample Num: 22848, Cur Loss: 0.09617083, Cur Avg Loss: 0.18640142, Log Avg loss: 0.28745622, Global Avg Loss: 0.81862673, Time: 0.0208 Steps: 61040, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001438, Sample Num: 23008, Cur Loss: 0.16146277, Cur Avg Loss: 0.18672545, Log Avg loss: 0.23299667, Global Avg Loss: 0.81853080, Time: 0.0210 Steps: 61050, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001448, Sample Num: 23168, Cur Loss: 0.09539366, Cur Avg Loss: 0.18684109, Log Avg loss: 0.20347029, Global Avg Loss: 0.81843007, Time: 0.0209 Steps: 61060, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001458, Sample Num: 23328, Cur Loss: 0.18426397, Cur Avg Loss: 0.18680452, Log Avg loss: 0.18151021, Global Avg Loss: 0.81832578, Time: 0.0209 Steps: 61070, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001468, Sample Num: 23488, Cur Loss: 0.39811638, Cur Avg Loss: 0.18696981, Log Avg loss: 0.21106879, Global Avg Loss: 0.81822636, Time: 0.0208 Steps: 61080, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001478, Sample Num: 23648, Cur Loss: 0.11407551, Cur Avg Loss: 0.18636464, Log Avg loss: 0.09752539, Global Avg Loss: 0.81810839, Time: 0.0209 Steps: 61090, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001488, Sample Num: 23808, Cur Loss: 0.11822141, Cur Avg Loss: 0.18643015, Log Avg loss: 0.19611269, Global Avg Loss: 0.81800659, Time: 0.0209 Steps: 61100, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001498, Sample Num: 23968, Cur Loss: 0.11862471, Cur Avg Loss: 0.18623040, Log Avg loss: 0.15650713, Global Avg Loss: 0.81789834, Time: 0.0211 Steps: 61110, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001508, Sample Num: 24128, Cur Loss: 0.24402842, Cur Avg Loss: 0.18636329, Log Avg loss: 0.20627039, Global Avg Loss: 0.81779827, Time: 0.0209 Steps: 61120, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001518, Sample Num: 24288, Cur Loss: 0.37375540, Cur Avg Loss: 0.18655361, Log Avg loss: 0.21525357, Global Avg Loss: 0.81769970, Time: 0.0209 Steps: 61130, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001528, Sample Num: 24448, Cur Loss: 0.20837170, Cur Avg Loss: 0.18664515, Log Avg loss: 0.20054116, Global Avg Loss: 0.81759876, Time: 0.0210 Steps: 61140, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001538, Sample Num: 24608, Cur Loss: 0.25155786, Cur Avg Loss: 0.18642342, Log Avg loss: 0.15254252, Global Avg Loss: 0.81749000, Time: 0.0247 Steps: 61150, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001548, Sample Num: 24768, Cur Loss: 0.13244477, Cur Avg Loss: 0.18606890, Log Avg loss: 0.13154357, Global Avg Loss: 0.81737785, Time: 0.0209 Steps: 61160, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001558, Sample Num: 24928, Cur Loss: 0.41913426, Cur Avg Loss: 0.18619686, Log Avg loss: 0.20600535, Global Avg Loss: 0.81727790, Time: 0.0209 Steps: 61170, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001568, Sample Num: 25088, Cur Loss: 0.06850123, Cur Avg Loss: 0.18648261, Log Avg loss: 0.23100298, Global Avg Loss: 0.81718207, Time: 0.0210 Steps: 61180, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001578, Sample Num: 25248, Cur Loss: 0.31186938, Cur Avg Loss: 0.18735285, Log Avg loss: 0.32380591, Global Avg Loss: 0.81710144, Time: 0.0209 Steps: 61190, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001588, Sample Num: 25408, Cur Loss: 0.63303792, Cur Avg Loss: 0.18788780, Log Avg loss: 0.27230332, Global Avg Loss: 0.81701242, Time: 0.0210 Steps: 61200, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001598, Sample Num: 25568, Cur Loss: 0.08686505, Cur Avg Loss: 0.18792437, Log Avg loss: 0.19373186, Global Avg Loss: 0.81691059, Time: 0.0209 Steps: 61210, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001608, Sample Num: 25728, Cur Loss: 0.22554420, Cur Avg Loss: 0.18780364, Log Avg loss: 0.16851061, Global Avg Loss: 0.81680468, Time: 0.0209 Steps: 61220, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001618, Sample Num: 25888, Cur Loss: 0.04669001, Cur Avg Loss: 0.18789246, Log Avg loss: 0.20217502, Global Avg Loss: 0.81670430, Time: 0.0209 Steps: 61230, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001628, Sample Num: 26048, Cur Loss: 0.09814538, Cur Avg Loss: 0.18753061, Log Avg loss: 0.12898298, Global Avg Loss: 0.81659200, Time: 0.0209 Steps: 61240, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001638, Sample Num: 26208, Cur Loss: 0.18145302, Cur Avg Loss: 0.18741084, Log Avg loss: 0.16791183, Global Avg Loss: 0.81648609, Time: 0.0209 Steps: 61250, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001648, Sample Num: 26368, Cur Loss: 0.16960657, Cur Avg Loss: 0.18724814, Log Avg loss: 0.16059906, Global Avg Loss: 0.81637903, Time: 0.0209 Steps: 61260, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001658, Sample Num: 26528, Cur Loss: 0.05315638, Cur Avg Loss: 0.18707836, Log Avg loss: 0.15909791, Global Avg Loss: 0.81627175, Time: 0.0209 Steps: 61270, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001668, Sample Num: 26688, Cur Loss: 0.20815474, Cur Avg Loss: 0.18697694, Log Avg loss: 0.17016120, Global Avg Loss: 0.81616632, Time: 0.0209 Steps: 61280, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001678, Sample Num: 26848, Cur Loss: 0.11158437, Cur Avg Loss: 0.18698076, Log Avg loss: 0.18761909, Global Avg Loss: 0.81606376, Time: 0.0209 Steps: 61290, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001688, Sample Num: 27008, Cur Loss: 0.04775941, Cur Avg Loss: 0.18696816, Log Avg loss: 0.18485272, Global Avg Loss: 0.81596079, Time: 0.0209 Steps: 61300, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001698, Sample Num: 27168, Cur Loss: 0.41724634, Cur Avg Loss: 0.18746471, Log Avg loss: 0.27128220, Global Avg Loss: 0.81587195, Time: 0.0209 Steps: 61310, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001708, Sample Num: 27328, Cur Loss: 0.12328938, Cur Avg Loss: 0.18762710, Log Avg loss: 0.21520230, Global Avg Loss: 0.81577400, Time: 0.0209 Steps: 61320, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001718, Sample Num: 27488, Cur Loss: 0.31846413, Cur Avg Loss: 0.18761378, Log Avg loss: 0.18533832, Global Avg Loss: 0.81567120, Time: 0.0209 Steps: 61330, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001728, Sample Num: 27648, Cur Loss: 0.27961010, Cur Avg Loss: 0.18744004, Log Avg loss: 0.15759213, Global Avg Loss: 0.81556392, Time: 0.0209 Steps: 61340, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001738, Sample Num: 27808, Cur Loss: 0.05709933, Cur Avg Loss: 0.18703034, Log Avg loss: 0.11623262, Global Avg Loss: 0.81544993, Time: 0.0209 Steps: 61350, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001748, Sample Num: 27968, Cur Loss: 0.22440812, Cur Avg Loss: 0.18685940, Log Avg loss: 0.15715070, Global Avg Loss: 0.81534264, Time: 0.0209 Steps: 61360, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001758, Sample Num: 28128, Cur Loss: 0.14315529, Cur Avg Loss: 0.18709510, Log Avg loss: 0.22829638, Global Avg Loss: 0.81524699, Time: 0.0209 Steps: 61370, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001768, Sample Num: 28288, Cur Loss: 0.21080793, Cur Avg Loss: 0.18831402, Log Avg loss: 0.40260003, Global Avg Loss: 0.81517976, Time: 0.0209 Steps: 61380, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001778, Sample Num: 28448, Cur Loss: 0.19004996, Cur Avg Loss: 0.18849694, Log Avg loss: 0.22083679, Global Avg Loss: 0.81508294, Time: 0.0210 Steps: 61390, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001788, Sample Num: 28608, Cur Loss: 0.10082042, Cur Avg Loss: 0.18850260, Log Avg loss: 0.18950939, Global Avg Loss: 0.81498106, Time: 0.0209 Steps: 61400, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001798, Sample Num: 28768, Cur Loss: 0.20321831, Cur Avg Loss: 0.18847980, Log Avg loss: 0.18440210, Global Avg Loss: 0.81487838, Time: 0.0210 Steps: 61410, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001808, Sample Num: 28928, Cur Loss: 0.31397948, Cur Avg Loss: 0.18883138, Log Avg loss: 0.25204511, Global Avg Loss: 0.81478674, Time: 0.0209 Steps: 61420, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001818, Sample Num: 29088, Cur Loss: 0.26888505, Cur Avg Loss: 0.18874074, Log Avg loss: 0.17235295, Global Avg Loss: 0.81468216, Time: 0.0209 Steps: 61430, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001828, Sample Num: 29248, Cur Loss: 0.06997427, Cur Avg Loss: 0.18869267, Log Avg loss: 0.17995479, Global Avg Loss: 0.81457885, Time: 0.0208 Steps: 61440, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001838, Sample Num: 29408, Cur Loss: 0.15363704, Cur Avg Loss: 0.18899340, Log Avg loss: 0.24396618, Global Avg Loss: 0.81448599, Time: 0.0208 Steps: 61450, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001848, Sample Num: 29568, Cur Loss: 0.06104401, Cur Avg Loss: 0.18873791, Log Avg loss: 0.14177863, Global Avg Loss: 0.81437654, Time: 0.0208 Steps: 61460, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001858, Sample Num: 29728, Cur Loss: 0.03900864, Cur Avg Loss: 0.18830270, Log Avg loss: 0.10787611, Global Avg Loss: 0.81426160, Time: 0.0209 Steps: 61470, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001868, Sample Num: 29888, Cur Loss: 0.32211879, Cur Avg Loss: 0.18853311, Log Avg loss: 0.23134288, Global Avg Loss: 0.81416679, Time: 0.0208 Steps: 61480, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001878, Sample Num: 30048, Cur Loss: 0.17746054, Cur Avg Loss: 0.18859142, Log Avg loss: 0.19948383, Global Avg Loss: 0.81406682, Time: 0.0209 Steps: 61490, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001888, Sample Num: 30208, Cur Loss: 0.19810265, Cur Avg Loss: 0.18895971, Log Avg loss: 0.25812417, Global Avg Loss: 0.81397643, Time: 0.0208 Steps: 61500, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001898, Sample Num: 30368, Cur Loss: 0.21837261, Cur Avg Loss: 0.18883960, Log Avg loss: 0.16616425, Global Avg Loss: 0.81387111, Time: 0.0208 Steps: 61510, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001908, Sample Num: 30528, Cur Loss: 0.20077051, Cur Avg Loss: 0.18857431, Log Avg loss: 0.13822253, Global Avg Loss: 0.81376128, Time: 0.0208 Steps: 61520, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001918, Sample Num: 30688, Cur Loss: 0.13641265, Cur Avg Loss: 0.18831659, Log Avg loss: 0.13914210, Global Avg Loss: 0.81365164, Time: 0.0209 Steps: 61530, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001928, Sample Num: 30848, Cur Loss: 0.34925210, Cur Avg Loss: 0.18860447, Log Avg loss: 0.24382116, Global Avg Loss: 0.81355905, Time: 0.0209 Steps: 61540, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001938, Sample Num: 31008, Cur Loss: 0.05848305, Cur Avg Loss: 0.18850842, Log Avg loss: 0.16999024, Global Avg Loss: 0.81345449, Time: 0.0209 Steps: 61550, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001948, Sample Num: 31168, Cur Loss: 0.41809264, Cur Avg Loss: 0.18865928, Log Avg loss: 0.21789480, Global Avg Loss: 0.81335774, Time: 0.0209 Steps: 61560, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001958, Sample Num: 31328, Cur Loss: 0.07178667, Cur Avg Loss: 0.18861811, Log Avg loss: 0.18059924, Global Avg Loss: 0.81325497, Time: 0.0209 Steps: 61570, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001968, Sample Num: 31488, Cur Loss: 0.05642123, Cur Avg Loss: 0.18840092, Log Avg loss: 0.14587522, Global Avg Loss: 0.81314660, Time: 0.0208 Steps: 61580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001978, Sample Num: 31648, Cur Loss: 0.19281697, Cur Avg Loss: 0.18861946, Log Avg loss: 0.23162670, Global Avg Loss: 0.81305218, Time: 0.0209 Steps: 61590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001988, Sample Num: 31808, Cur Loss: 0.12297207, Cur Avg Loss: 0.18876486, Log Avg loss: 0.21752486, Global Avg Loss: 0.81295550, Time: 0.0208 Steps: 61600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001998, Sample Num: 31968, Cur Loss: 0.12473918, Cur Avg Loss: 0.18859747, Log Avg loss: 0.15532078, Global Avg Loss: 0.81284876, Time: 0.0209 Steps: 61610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002008, Sample Num: 32128, Cur Loss: 0.30081397, Cur Avg Loss: 0.18880146, Log Avg loss: 0.22955923, Global Avg Loss: 0.81275410, Time: 0.0208 Steps: 61620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002018, Sample Num: 32288, Cur Loss: 0.36855114, Cur Avg Loss: 0.18882284, Log Avg loss: 0.19311660, Global Avg Loss: 0.81265356, Time: 0.0209 Steps: 61630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002028, Sample Num: 32448, Cur Loss: 0.51482606, Cur Avg Loss: 0.18914735, Log Avg loss: 0.25463329, Global Avg Loss: 0.81256303, Time: 0.0209 Steps: 61640, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002038, Sample Num: 32608, Cur Loss: 0.14729095, Cur Avg Loss: 0.18959063, Log Avg loss: 0.27948751, Global Avg Loss: 0.81247656, Time: 0.0209 Steps: 61650, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002048, Sample Num: 32768, Cur Loss: 0.13204198, Cur Avg Loss: 0.18971527, Log Avg loss: 0.21511680, Global Avg Loss: 0.81237968, Time: 0.0254 Steps: 61660, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002058, Sample Num: 32928, Cur Loss: 0.07530916, Cur Avg Loss: 0.18958600, Log Avg loss: 0.16311212, Global Avg Loss: 0.81227440, Time: 0.0210 Steps: 61670, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002068, Sample Num: 33088, Cur Loss: 0.64936471, Cur Avg Loss: 0.18962976, Log Avg loss: 0.19863369, Global Avg Loss: 0.81217491, Time: 0.0209 Steps: 61680, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002078, Sample Num: 33248, Cur Loss: 0.24319132, Cur Avg Loss: 0.18959742, Log Avg loss: 0.18290946, Global Avg Loss: 0.81207291, Time: 0.0209 Steps: 61690, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002088, Sample Num: 33408, Cur Loss: 0.02102974, Cur Avg Loss: 0.18938001, Log Avg loss: 0.14420244, Global Avg Loss: 0.81196466, Time: 0.0209 Steps: 61700, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002098, Sample Num: 33568, Cur Loss: 0.06017814, Cur Avg Loss: 0.18922221, Log Avg loss: 0.15627517, Global Avg Loss: 0.81185841, Time: 0.0209 Steps: 61710, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002108, Sample Num: 33728, Cur Loss: 0.02396935, Cur Avg Loss: 0.18923316, Log Avg loss: 0.19152881, Global Avg Loss: 0.81175790, Time: 0.0209 Steps: 61720, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002118, Sample Num: 33888, Cur Loss: 0.14149931, Cur Avg Loss: 0.18886779, Log Avg loss: 0.11184943, Global Avg Loss: 0.81164452, Time: 0.0209 Steps: 61730, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002128, Sample Num: 34048, Cur Loss: 0.16233501, Cur Avg Loss: 0.18903366, Log Avg loss: 0.22416395, Global Avg Loss: 0.81154937, Time: 0.0209 Steps: 61740, Updated lr: 0.000042 ***** Running evaluation checkpoint-61741 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-61741 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.738932, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.238708, "eval_total_loss": 167.811467, "eval_mae": 0.320697, "eval_mse": 0.238793, "eval_r2": 0.848207, "eval_sp_statistic": 0.919546, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.93052, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.18344, "test_total_loss": 92.087095, "test_mae": 0.293655, "test_mse": 0.183507, "test_r2": 0.881563, "test_sp_statistic": 0.914251, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946554, "test_ps_pvalue": 0.0, "lr": 4.239829302987198e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.8115364596417765, "train_cur_epoch_loss": 402.27818524837494, "train_cur_epoch_avg_loss": 0.1889517074910169, "train_cur_epoch_time": 44.73893213272095, "train_cur_epoch_avg_time": 0.02101405924505446, "epoch": 29, "step": 61741} ################################################## Training, Epoch: 0030, Batch: 000009, Sample Num: 144, Cur Loss: 0.17896381, Cur Avg Loss: 0.19235227, Log Avg loss: 0.17457281, Global Avg Loss: 0.81144621, Time: 0.0210 Steps: 61750, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000019, Sample Num: 304, Cur Loss: 0.07363413, Cur Avg Loss: 0.15811805, Log Avg loss: 0.12730725, Global Avg Loss: 0.81133544, Time: 0.0210 Steps: 61760, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000029, Sample Num: 464, Cur Loss: 0.56966937, Cur Avg Loss: 0.16138119, Log Avg loss: 0.16758116, Global Avg Loss: 0.81123122, Time: 0.0209 Steps: 61770, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000039, Sample Num: 624, Cur Loss: 0.39082503, Cur Avg Loss: 0.17388624, Log Avg loss: 0.21015088, Global Avg Loss: 0.81113393, Time: 0.0210 Steps: 61780, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000049, Sample Num: 784, Cur Loss: 0.19096106, Cur Avg Loss: 0.18108955, Log Avg loss: 0.20918246, Global Avg Loss: 0.81103651, Time: 0.0209 Steps: 61790, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000059, Sample Num: 944, Cur Loss: 0.46094137, Cur Avg Loss: 0.18360935, Log Avg loss: 0.19595639, Global Avg Loss: 0.81093698, Time: 0.0209 Steps: 61800, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000069, Sample Num: 1104, Cur Loss: 0.08605294, Cur Avg Loss: 0.17858508, Log Avg loss: 0.14894183, Global Avg Loss: 0.81082988, Time: 0.0209 Steps: 61810, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000079, Sample Num: 1264, Cur Loss: 0.05607274, Cur Avg Loss: 0.18082840, Log Avg loss: 0.19630731, Global Avg Loss: 0.81073048, Time: 0.0210 Steps: 61820, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000089, Sample Num: 1424, Cur Loss: 0.10371672, Cur Avg Loss: 0.17676505, Log Avg loss: 0.14466465, Global Avg Loss: 0.81062275, Time: 0.0210 Steps: 61830, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000099, Sample Num: 1584, Cur Loss: 0.11946766, Cur Avg Loss: 0.17519486, Log Avg loss: 0.16122018, Global Avg Loss: 0.81051774, Time: 0.0210 Steps: 61840, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000109, Sample Num: 1744, Cur Loss: 0.20561282, Cur Avg Loss: 0.17615258, Log Avg loss: 0.18563393, Global Avg Loss: 0.81041670, Time: 0.0209 Steps: 61850, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000119, Sample Num: 1904, Cur Loss: 0.23228475, Cur Avg Loss: 0.17674207, Log Avg loss: 0.18316757, Global Avg Loss: 0.81031531, Time: 0.0209 Steps: 61860, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000129, Sample Num: 2064, Cur Loss: 0.20477313, Cur Avg Loss: 0.17977854, Log Avg loss: 0.21591248, Global Avg Loss: 0.81021923, Time: 0.0209 Steps: 61870, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000139, Sample Num: 2224, Cur Loss: 0.14077736, Cur Avg Loss: 0.18133387, Log Avg loss: 0.20139765, Global Avg Loss: 0.81012085, Time: 0.0210 Steps: 61880, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000149, Sample Num: 2384, Cur Loss: 0.06250943, Cur Avg Loss: 0.17825227, Log Avg loss: 0.13541807, Global Avg Loss: 0.81001183, Time: 0.0209 Steps: 61890, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000159, Sample Num: 2544, Cur Loss: 0.05353427, Cur Avg Loss: 0.17414863, Log Avg loss: 0.11300443, Global Avg Loss: 0.80989923, Time: 0.0209 Steps: 61900, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000169, Sample Num: 2704, Cur Loss: 0.21613921, Cur Avg Loss: 0.17742132, Log Avg loss: 0.22945702, Global Avg Loss: 0.80980547, Time: 0.0210 Steps: 61910, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000179, Sample Num: 2864, Cur Loss: 0.23104200, Cur Avg Loss: 0.17801314, Log Avg loss: 0.18801491, Global Avg Loss: 0.80970505, Time: 0.0209 Steps: 61920, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000189, Sample Num: 3024, Cur Loss: 0.26874310, Cur Avg Loss: 0.17643300, Log Avg loss: 0.14814849, Global Avg Loss: 0.80959823, Time: 0.0210 Steps: 61930, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000199, Sample Num: 3184, Cur Loss: 0.16711444, Cur Avg Loss: 0.17478141, Log Avg loss: 0.14356646, Global Avg Loss: 0.80949070, Time: 0.0209 Steps: 61940, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000209, Sample Num: 3344, Cur Loss: 0.13571081, Cur Avg Loss: 0.17142013, Log Avg loss: 0.10453054, Global Avg Loss: 0.80937691, Time: 0.0210 Steps: 61950, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000219, Sample Num: 3504, Cur Loss: 0.07143359, Cur Avg Loss: 0.16999273, Log Avg loss: 0.14016017, Global Avg Loss: 0.80926890, Time: 0.0209 Steps: 61960, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000229, Sample Num: 3664, Cur Loss: 0.03500870, Cur Avg Loss: 0.17048006, Log Avg loss: 0.18115241, Global Avg Loss: 0.80916754, Time: 0.0210 Steps: 61970, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000239, Sample Num: 3824, Cur Loss: 0.29824406, Cur Avg Loss: 0.17247455, Log Avg loss: 0.21814844, Global Avg Loss: 0.80907218, Time: 0.0209 Steps: 61980, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000249, Sample Num: 3984, Cur Loss: 0.07712179, Cur Avg Loss: 0.17029405, Log Avg loss: 0.11818014, Global Avg Loss: 0.80896073, Time: 0.0209 Steps: 61990, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000259, Sample Num: 4144, Cur Loss: 0.10146987, Cur Avg Loss: 0.16841446, Log Avg loss: 0.12161257, Global Avg Loss: 0.80884987, Time: 0.0248 Steps: 62000, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000269, Sample Num: 4304, Cur Loss: 0.09126728, Cur Avg Loss: 0.16787352, Log Avg loss: 0.15386339, Global Avg Loss: 0.80874424, Time: 0.0210 Steps: 62010, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000279, Sample Num: 4464, Cur Loss: 0.10337344, Cur Avg Loss: 0.16738053, Log Avg loss: 0.15411895, Global Avg Loss: 0.80863869, Time: 0.0210 Steps: 62020, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000289, Sample Num: 4624, Cur Loss: 0.17478265, Cur Avg Loss: 0.16701936, Log Avg loss: 0.15694272, Global Avg Loss: 0.80853363, Time: 0.0210 Steps: 62030, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000299, Sample Num: 4784, Cur Loss: 0.14014168, Cur Avg Loss: 0.16535180, Log Avg loss: 0.11715948, Global Avg Loss: 0.80842219, Time: 0.0210 Steps: 62040, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000309, Sample Num: 4944, Cur Loss: 0.17570506, Cur Avg Loss: 0.16651061, Log Avg loss: 0.20115879, Global Avg Loss: 0.80832432, Time: 0.0210 Steps: 62050, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000319, Sample Num: 5104, Cur Loss: 0.19621849, Cur Avg Loss: 0.16696001, Log Avg loss: 0.18084662, Global Avg Loss: 0.80822322, Time: 0.0210 Steps: 62060, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000329, Sample Num: 5264, Cur Loss: 0.12608019, Cur Avg Loss: 0.16899014, Log Avg loss: 0.23375129, Global Avg Loss: 0.80813066, Time: 0.0210 Steps: 62070, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000339, Sample Num: 5424, Cur Loss: 0.10874505, Cur Avg Loss: 0.16987939, Log Avg loss: 0.19913558, Global Avg Loss: 0.80803257, Time: 0.0209 Steps: 62080, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000349, Sample Num: 5584, Cur Loss: 0.14596190, Cur Avg Loss: 0.17069691, Log Avg loss: 0.19841105, Global Avg Loss: 0.80793438, Time: 0.0210 Steps: 62090, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000359, Sample Num: 5744, Cur Loss: 0.05276884, Cur Avg Loss: 0.17122756, Log Avg loss: 0.18974706, Global Avg Loss: 0.80783483, Time: 0.0210 Steps: 62100, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000369, Sample Num: 5904, Cur Loss: 0.17787050, Cur Avg Loss: 0.17048743, Log Avg loss: 0.14391673, Global Avg Loss: 0.80772794, Time: 0.0210 Steps: 62110, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000379, Sample Num: 6064, Cur Loss: 0.18915381, Cur Avg Loss: 0.17030747, Log Avg loss: 0.16366706, Global Avg Loss: 0.80762426, Time: 0.0210 Steps: 62120, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000389, Sample Num: 6224, Cur Loss: 0.13184994, Cur Avg Loss: 0.16970171, Log Avg loss: 0.14674351, Global Avg Loss: 0.80751789, Time: 0.0210 Steps: 62130, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000399, Sample Num: 6384, Cur Loss: 0.29918033, Cur Avg Loss: 0.16958558, Log Avg loss: 0.16506817, Global Avg Loss: 0.80741450, Time: 0.0210 Steps: 62140, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000409, Sample Num: 6544, Cur Loss: 0.11201006, Cur Avg Loss: 0.16877958, Log Avg loss: 0.13662005, Global Avg Loss: 0.80730657, Time: 0.0212 Steps: 62150, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000419, Sample Num: 6704, Cur Loss: 0.08186843, Cur Avg Loss: 0.16984849, Log Avg loss: 0.21356664, Global Avg Loss: 0.80721105, Time: 0.0210 Steps: 62160, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000429, Sample Num: 6864, Cur Loss: 0.11516624, Cur Avg Loss: 0.17092538, Log Avg loss: 0.21604723, Global Avg Loss: 0.80711596, Time: 0.0210 Steps: 62170, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000439, Sample Num: 7024, Cur Loss: 0.12477116, Cur Avg Loss: 0.17127551, Log Avg loss: 0.18629601, Global Avg Loss: 0.80701612, Time: 0.0210 Steps: 62180, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000449, Sample Num: 7184, Cur Loss: 0.11335867, Cur Avg Loss: 0.17035430, Log Avg loss: 0.12991339, Global Avg Loss: 0.80690725, Time: 0.0210 Steps: 62190, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000459, Sample Num: 7344, Cur Loss: 0.05546077, Cur Avg Loss: 0.16868883, Log Avg loss: 0.09390935, Global Avg Loss: 0.80679262, Time: 0.0210 Steps: 62200, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000469, Sample Num: 7504, Cur Loss: 0.42000079, Cur Avg Loss: 0.16961792, Log Avg loss: 0.21226302, Global Avg Loss: 0.80669705, Time: 0.0210 Steps: 62210, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000479, Sample Num: 7664, Cur Loss: 0.20199692, Cur Avg Loss: 0.16939097, Log Avg loss: 0.15874674, Global Avg Loss: 0.80659291, Time: 0.0210 Steps: 62220, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000489, Sample Num: 7824, Cur Loss: 0.15923545, Cur Avg Loss: 0.16965734, Log Avg loss: 0.18241652, Global Avg Loss: 0.80649261, Time: 0.0212 Steps: 62230, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000499, Sample Num: 7984, Cur Loss: 0.33606267, Cur Avg Loss: 0.16950307, Log Avg loss: 0.16195940, Global Avg Loss: 0.80638905, Time: 0.0210 Steps: 62240, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000509, Sample Num: 8144, Cur Loss: 0.34087110, Cur Avg Loss: 0.17006415, Log Avg loss: 0.19806220, Global Avg Loss: 0.80629133, Time: 0.0210 Steps: 62250, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000519, Sample Num: 8304, Cur Loss: 0.08882532, Cur Avg Loss: 0.16947511, Log Avg loss: 0.13949260, Global Avg Loss: 0.80618423, Time: 0.0208 Steps: 62260, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000529, Sample Num: 8464, Cur Loss: 0.19553250, Cur Avg Loss: 0.16877841, Log Avg loss: 0.13261979, Global Avg Loss: 0.80607606, Time: 0.0209 Steps: 62270, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000539, Sample Num: 8624, Cur Loss: 0.19936559, Cur Avg Loss: 0.16897158, Log Avg loss: 0.17919042, Global Avg Loss: 0.80597541, Time: 0.0209 Steps: 62280, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000549, Sample Num: 8784, Cur Loss: 0.27155480, Cur Avg Loss: 0.16855385, Log Avg loss: 0.14603838, Global Avg Loss: 0.80586946, Time: 0.0208 Steps: 62290, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000559, Sample Num: 8944, Cur Loss: 0.09366217, Cur Avg Loss: 0.16787669, Log Avg loss: 0.13070045, Global Avg Loss: 0.80576109, Time: 0.0209 Steps: 62300, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000569, Sample Num: 9104, Cur Loss: 0.37613770, Cur Avg Loss: 0.16978922, Log Avg loss: 0.27669939, Global Avg Loss: 0.80567618, Time: 0.0209 Steps: 62310, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000579, Sample Num: 9264, Cur Loss: 0.04976230, Cur Avg Loss: 0.16903085, Log Avg loss: 0.12587981, Global Avg Loss: 0.80556710, Time: 0.0209 Steps: 62320, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000589, Sample Num: 9424, Cur Loss: 0.09776439, Cur Avg Loss: 0.16885368, Log Avg loss: 0.15859534, Global Avg Loss: 0.80546330, Time: 0.0208 Steps: 62330, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000599, Sample Num: 9584, Cur Loss: 0.17294824, Cur Avg Loss: 0.16995254, Log Avg loss: 0.23467578, Global Avg Loss: 0.80537174, Time: 0.0208 Steps: 62340, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000609, Sample Num: 9744, Cur Loss: 0.13212559, Cur Avg Loss: 0.16988731, Log Avg loss: 0.16597976, Global Avg Loss: 0.80526919, Time: 0.0208 Steps: 62350, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000619, Sample Num: 9904, Cur Loss: 0.16617227, Cur Avg Loss: 0.16926828, Log Avg loss: 0.13156938, Global Avg Loss: 0.80516115, Time: 0.0208 Steps: 62360, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000629, Sample Num: 10064, Cur Loss: 0.29244030, Cur Avg Loss: 0.16890525, Log Avg loss: 0.14643344, Global Avg Loss: 0.80505554, Time: 0.0208 Steps: 62370, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000639, Sample Num: 10224, Cur Loss: 0.10598299, Cur Avg Loss: 0.16840842, Log Avg loss: 0.13715812, Global Avg Loss: 0.80494847, Time: 0.0209 Steps: 62380, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000649, Sample Num: 10384, Cur Loss: 0.29602385, Cur Avg Loss: 0.16959644, Log Avg loss: 0.24551110, Global Avg Loss: 0.80485880, Time: 0.0208 Steps: 62390, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000659, Sample Num: 10544, Cur Loss: 0.33894479, Cur Avg Loss: 0.16958696, Log Avg loss: 0.16897140, Global Avg Loss: 0.80475690, Time: 0.0209 Steps: 62400, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000669, Sample Num: 10704, Cur Loss: 0.14900737, Cur Avg Loss: 0.16981764, Log Avg loss: 0.18501975, Global Avg Loss: 0.80465760, Time: 0.0209 Steps: 62410, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000679, Sample Num: 10864, Cur Loss: 0.04788124, Cur Avg Loss: 0.16973918, Log Avg loss: 0.16449019, Global Avg Loss: 0.80455504, Time: 0.0209 Steps: 62420, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000689, Sample Num: 11024, Cur Loss: 0.07834744, Cur Avg Loss: 0.17009108, Log Avg loss: 0.19398459, Global Avg Loss: 0.80445724, Time: 0.0209 Steps: 62430, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000699, Sample Num: 11184, Cur Loss: 0.10569885, Cur Avg Loss: 0.16998001, Log Avg loss: 0.16232795, Global Avg Loss: 0.80435440, Time: 0.0209 Steps: 62440, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000709, Sample Num: 11344, Cur Loss: 0.37774709, Cur Avg Loss: 0.17003446, Log Avg loss: 0.17384047, Global Avg Loss: 0.80425343, Time: 0.0208 Steps: 62450, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000719, Sample Num: 11504, Cur Loss: 0.11017293, Cur Avg Loss: 0.17049518, Log Avg loss: 0.20315979, Global Avg Loss: 0.80415720, Time: 0.0208 Steps: 62460, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000729, Sample Num: 11664, Cur Loss: 0.08310190, Cur Avg Loss: 0.17012712, Log Avg loss: 0.14366345, Global Avg Loss: 0.80405147, Time: 0.0208 Steps: 62470, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000739, Sample Num: 11824, Cur Loss: 0.28999391, Cur Avg Loss: 0.17039615, Log Avg loss: 0.19000856, Global Avg Loss: 0.80395319, Time: 0.0209 Steps: 62480, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000749, Sample Num: 11984, Cur Loss: 0.08712412, Cur Avg Loss: 0.17084275, Log Avg loss: 0.20384673, Global Avg Loss: 0.80385716, Time: 0.0208 Steps: 62490, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000759, Sample Num: 12144, Cur Loss: 0.12160249, Cur Avg Loss: 0.17055699, Log Avg loss: 0.14915338, Global Avg Loss: 0.80375240, Time: 0.0209 Steps: 62500, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000769, Sample Num: 12304, Cur Loss: 0.09134646, Cur Avg Loss: 0.17146209, Log Avg loss: 0.24015901, Global Avg Loss: 0.80366224, Time: 0.0245 Steps: 62510, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000779, Sample Num: 12464, Cur Loss: 0.11055569, Cur Avg Loss: 0.17242662, Log Avg loss: 0.24659938, Global Avg Loss: 0.80357314, Time: 0.0209 Steps: 62520, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000789, Sample Num: 12624, Cur Loss: 0.12061951, Cur Avg Loss: 0.17241621, Log Avg loss: 0.17160515, Global Avg Loss: 0.80347208, Time: 0.0209 Steps: 62530, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000799, Sample Num: 12784, Cur Loss: 0.03971276, Cur Avg Loss: 0.17155460, Log Avg loss: 0.10357348, Global Avg Loss: 0.80336016, Time: 0.0209 Steps: 62540, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000809, Sample Num: 12944, Cur Loss: 0.07464965, Cur Avg Loss: 0.17134647, Log Avg loss: 0.15471686, Global Avg Loss: 0.80325646, Time: 0.0209 Steps: 62550, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000819, Sample Num: 13104, Cur Loss: 0.14821869, Cur Avg Loss: 0.17107945, Log Avg loss: 0.14947748, Global Avg Loss: 0.80315196, Time: 0.0208 Steps: 62560, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000829, Sample Num: 13264, Cur Loss: 0.35109657, Cur Avg Loss: 0.17162399, Log Avg loss: 0.21622186, Global Avg Loss: 0.80305816, Time: 0.0208 Steps: 62570, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000839, Sample Num: 13424, Cur Loss: 0.21766776, Cur Avg Loss: 0.17171539, Log Avg loss: 0.17929279, Global Avg Loss: 0.80295848, Time: 0.0208 Steps: 62580, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000849, Sample Num: 13584, Cur Loss: 0.14072511, Cur Avg Loss: 0.17176337, Log Avg loss: 0.17578835, Global Avg Loss: 0.80285828, Time: 0.0209 Steps: 62590, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000859, Sample Num: 13744, Cur Loss: 0.87609446, Cur Avg Loss: 0.17247353, Log Avg loss: 0.23276613, Global Avg Loss: 0.80276721, Time: 0.0208 Steps: 62600, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000869, Sample Num: 13904, Cur Loss: 0.14567903, Cur Avg Loss: 0.17256863, Log Avg loss: 0.18073799, Global Avg Loss: 0.80266786, Time: 0.0209 Steps: 62610, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000879, Sample Num: 14064, Cur Loss: 0.16452000, Cur Avg Loss: 0.17312547, Log Avg loss: 0.22151455, Global Avg Loss: 0.80257505, Time: 0.0209 Steps: 62620, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000889, Sample Num: 14224, Cur Loss: 0.32651755, Cur Avg Loss: 0.17398005, Log Avg loss: 0.24909828, Global Avg Loss: 0.80248668, Time: 0.0209 Steps: 62630, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000899, Sample Num: 14384, Cur Loss: 0.04411942, Cur Avg Loss: 0.17388147, Log Avg loss: 0.16511786, Global Avg Loss: 0.80238493, Time: 0.0208 Steps: 62640, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000909, Sample Num: 14544, Cur Loss: 0.04442441, Cur Avg Loss: 0.17393062, Log Avg loss: 0.17834915, Global Avg Loss: 0.80228532, Time: 0.0209 Steps: 62650, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000919, Sample Num: 14704, Cur Loss: 0.14927760, Cur Avg Loss: 0.17433505, Log Avg loss: 0.21109703, Global Avg Loss: 0.80219097, Time: 0.0208 Steps: 62660, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000929, Sample Num: 14864, Cur Loss: 0.36334282, Cur Avg Loss: 0.17412828, Log Avg loss: 0.15512647, Global Avg Loss: 0.80208773, Time: 0.0208 Steps: 62670, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000939, Sample Num: 15024, Cur Loss: 0.06331341, Cur Avg Loss: 0.17348340, Log Avg loss: 0.11357428, Global Avg Loss: 0.80197788, Time: 0.0209 Steps: 62680, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000949, Sample Num: 15184, Cur Loss: 0.02830885, Cur Avg Loss: 0.17308199, Log Avg loss: 0.13538978, Global Avg Loss: 0.80187155, Time: 0.0208 Steps: 62690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000959, Sample Num: 15344, Cur Loss: 0.08358724, Cur Avg Loss: 0.17295829, Log Avg loss: 0.16121851, Global Avg Loss: 0.80176937, Time: 0.0208 Steps: 62700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000969, Sample Num: 15504, Cur Loss: 0.03524523, Cur Avg Loss: 0.17279542, Log Avg loss: 0.15717670, Global Avg Loss: 0.80166658, Time: 0.0208 Steps: 62710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000979, Sample Num: 15664, Cur Loss: 0.06979541, Cur Avg Loss: 0.17269176, Log Avg loss: 0.16264656, Global Avg Loss: 0.80156470, Time: 0.0208 Steps: 62720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000989, Sample Num: 15824, Cur Loss: 0.12582117, Cur Avg Loss: 0.17355870, Log Avg loss: 0.25843270, Global Avg Loss: 0.80147811, Time: 0.0208 Steps: 62730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000999, Sample Num: 15984, Cur Loss: 0.12996921, Cur Avg Loss: 0.17325247, Log Avg loss: 0.14296608, Global Avg Loss: 0.80137316, Time: 0.0209 Steps: 62740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001009, Sample Num: 16144, Cur Loss: 0.18484034, Cur Avg Loss: 0.17307411, Log Avg loss: 0.15525558, Global Avg Loss: 0.80127019, Time: 0.0208 Steps: 62750, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001019, Sample Num: 16304, Cur Loss: 0.36934292, Cur Avg Loss: 0.17333962, Log Avg loss: 0.20013000, Global Avg Loss: 0.80117440, Time: 0.0209 Steps: 62760, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001029, Sample Num: 16464, Cur Loss: 0.30569643, Cur Avg Loss: 0.17325535, Log Avg loss: 0.16466790, Global Avg Loss: 0.80107300, Time: 0.0245 Steps: 62770, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001039, Sample Num: 16624, Cur Loss: 0.03234192, Cur Avg Loss: 0.17315283, Log Avg loss: 0.16260357, Global Avg Loss: 0.80097130, Time: 0.0209 Steps: 62780, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001049, Sample Num: 16784, Cur Loss: 0.17049575, Cur Avg Loss: 0.17317562, Log Avg loss: 0.17554403, Global Avg Loss: 0.80087170, Time: 0.0208 Steps: 62790, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001059, Sample Num: 16944, Cur Loss: 0.12758966, Cur Avg Loss: 0.17447787, Log Avg loss: 0.31108365, Global Avg Loss: 0.80079370, Time: 0.0208 Steps: 62800, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001069, Sample Num: 17104, Cur Loss: 0.17631777, Cur Avg Loss: 0.17399594, Log Avg loss: 0.12295986, Global Avg Loss: 0.80068579, Time: 0.0209 Steps: 62810, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001079, Sample Num: 17264, Cur Loss: 0.18276802, Cur Avg Loss: 0.17378433, Log Avg loss: 0.15116269, Global Avg Loss: 0.80058239, Time: 0.0208 Steps: 62820, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001089, Sample Num: 17424, Cur Loss: 0.35377282, Cur Avg Loss: 0.17405270, Log Avg loss: 0.20300990, Global Avg Loss: 0.80048728, Time: 0.0209 Steps: 62830, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001099, Sample Num: 17584, Cur Loss: 0.18038893, Cur Avg Loss: 0.17415794, Log Avg loss: 0.18561843, Global Avg Loss: 0.80038944, Time: 0.0209 Steps: 62840, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001109, Sample Num: 17744, Cur Loss: 0.24411269, Cur Avg Loss: 0.17399919, Log Avg loss: 0.15655326, Global Avg Loss: 0.80028700, Time: 0.0208 Steps: 62850, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001119, Sample Num: 17904, Cur Loss: 0.10222326, Cur Avg Loss: 0.17407727, Log Avg loss: 0.18273604, Global Avg Loss: 0.80018875, Time: 0.0208 Steps: 62860, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001129, Sample Num: 18064, Cur Loss: 0.28716692, Cur Avg Loss: 0.17412271, Log Avg loss: 0.17920683, Global Avg Loss: 0.80008998, Time: 0.0208 Steps: 62870, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001139, Sample Num: 18224, Cur Loss: 0.17091040, Cur Avg Loss: 0.17374011, Log Avg loss: 0.13054531, Global Avg Loss: 0.79998350, Time: 0.0209 Steps: 62880, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001149, Sample Num: 18384, Cur Loss: 0.27812615, Cur Avg Loss: 0.17347998, Log Avg loss: 0.14385110, Global Avg Loss: 0.79987917, Time: 0.0208 Steps: 62890, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001159, Sample Num: 18544, Cur Loss: 0.18684892, Cur Avg Loss: 0.17360078, Log Avg loss: 0.18748029, Global Avg Loss: 0.79978181, Time: 0.0208 Steps: 62900, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001169, Sample Num: 18704, Cur Loss: 0.03115555, Cur Avg Loss: 0.17376713, Log Avg loss: 0.19304737, Global Avg Loss: 0.79968537, Time: 0.0208 Steps: 62910, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001179, Sample Num: 18864, Cur Loss: 0.63814056, Cur Avg Loss: 0.17405083, Log Avg loss: 0.20721484, Global Avg Loss: 0.79959120, Time: 0.0208 Steps: 62920, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001189, Sample Num: 19024, Cur Loss: 0.16014424, Cur Avg Loss: 0.17467042, Log Avg loss: 0.24772013, Global Avg Loss: 0.79950351, Time: 0.0208 Steps: 62930, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001199, Sample Num: 19184, Cur Loss: 0.16631788, Cur Avg Loss: 0.17512118, Log Avg loss: 0.22871749, Global Avg Loss: 0.79941282, Time: 0.0210 Steps: 62940, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001209, Sample Num: 19344, Cur Loss: 0.27696455, Cur Avg Loss: 0.17534176, Log Avg loss: 0.20178857, Global Avg Loss: 0.79931788, Time: 0.0209 Steps: 62950, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001219, Sample Num: 19504, Cur Loss: 0.11947411, Cur Avg Loss: 0.17508318, Log Avg loss: 0.14382100, Global Avg Loss: 0.79921377, Time: 0.0208 Steps: 62960, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001229, Sample Num: 19664, Cur Loss: 0.05023319, Cur Avg Loss: 0.17493304, Log Avg loss: 0.15663060, Global Avg Loss: 0.79911172, Time: 0.0208 Steps: 62970, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001239, Sample Num: 19824, Cur Loss: 0.14450394, Cur Avg Loss: 0.17506111, Log Avg loss: 0.19080092, Global Avg Loss: 0.79901514, Time: 0.0208 Steps: 62980, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001249, Sample Num: 19984, Cur Loss: 0.22254714, Cur Avg Loss: 0.17467117, Log Avg loss: 0.12635810, Global Avg Loss: 0.79890835, Time: 0.0208 Steps: 62990, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001259, Sample Num: 20144, Cur Loss: 0.15049969, Cur Avg Loss: 0.17482376, Log Avg loss: 0.19388224, Global Avg Loss: 0.79881231, Time: 0.0208 Steps: 63000, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001269, Sample Num: 20304, Cur Loss: 0.10444526, Cur Avg Loss: 0.17480146, Log Avg loss: 0.17199420, Global Avg Loss: 0.79871283, Time: 0.0208 Steps: 63010, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001279, Sample Num: 20464, Cur Loss: 0.03844709, Cur Avg Loss: 0.17490317, Log Avg loss: 0.18780913, Global Avg Loss: 0.79861590, Time: 0.0208 Steps: 63020, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001289, Sample Num: 20624, Cur Loss: 0.16548620, Cur Avg Loss: 0.17450528, Log Avg loss: 0.12361593, Global Avg Loss: 0.79850880, Time: 0.0209 Steps: 63030, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001299, Sample Num: 20784, Cur Loss: 0.08757323, Cur Avg Loss: 0.17432612, Log Avg loss: 0.15123229, Global Avg Loss: 0.79840613, Time: 0.0208 Steps: 63040, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001309, Sample Num: 20944, Cur Loss: 0.56716943, Cur Avg Loss: 0.17469536, Log Avg loss: 0.22265938, Global Avg Loss: 0.79831481, Time: 0.0208 Steps: 63050, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001319, Sample Num: 21104, Cur Loss: 0.07313401, Cur Avg Loss: 0.17446966, Log Avg loss: 0.14492599, Global Avg Loss: 0.79821120, Time: 0.0208 Steps: 63060, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001329, Sample Num: 21264, Cur Loss: 0.07129721, Cur Avg Loss: 0.17409483, Log Avg loss: 0.12465452, Global Avg Loss: 0.79810440, Time: 0.0209 Steps: 63070, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001339, Sample Num: 21424, Cur Loss: 0.11577594, Cur Avg Loss: 0.17420525, Log Avg loss: 0.18887978, Global Avg Loss: 0.79800782, Time: 0.0209 Steps: 63080, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001349, Sample Num: 21584, Cur Loss: 0.08251794, Cur Avg Loss: 0.17432081, Log Avg loss: 0.18979446, Global Avg Loss: 0.79791142, Time: 0.0209 Steps: 63090, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001359, Sample Num: 21744, Cur Loss: 0.07360981, Cur Avg Loss: 0.17418990, Log Avg loss: 0.15653085, Global Avg Loss: 0.79780977, Time: 0.0208 Steps: 63100, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001369, Sample Num: 21904, Cur Loss: 0.12767026, Cur Avg Loss: 0.17428889, Log Avg loss: 0.18774079, Global Avg Loss: 0.79771310, Time: 0.0208 Steps: 63110, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001379, Sample Num: 22064, Cur Loss: 0.08525987, Cur Avg Loss: 0.17412589, Log Avg loss: 0.15181120, Global Avg Loss: 0.79761078, Time: 0.0208 Steps: 63120, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001389, Sample Num: 22224, Cur Loss: 0.21829605, Cur Avg Loss: 0.17420175, Log Avg loss: 0.18466249, Global Avg Loss: 0.79751368, Time: 0.0209 Steps: 63130, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001399, Sample Num: 22384, Cur Loss: 0.28445187, Cur Avg Loss: 0.17430757, Log Avg loss: 0.18900683, Global Avg Loss: 0.79741731, Time: 0.0209 Steps: 63140, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001409, Sample Num: 22544, Cur Loss: 0.13004659, Cur Avg Loss: 0.17416066, Log Avg loss: 0.15360743, Global Avg Loss: 0.79731536, Time: 0.0208 Steps: 63150, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001419, Sample Num: 22704, Cur Loss: 0.39030886, Cur Avg Loss: 0.17421954, Log Avg loss: 0.18251622, Global Avg Loss: 0.79721802, Time: 0.0208 Steps: 63160, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001429, Sample Num: 22864, Cur Loss: 0.07004584, Cur Avg Loss: 0.17465140, Log Avg loss: 0.23593179, Global Avg Loss: 0.79712917, Time: 0.0209 Steps: 63170, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001439, Sample Num: 23024, Cur Loss: 0.04880916, Cur Avg Loss: 0.17465623, Log Avg loss: 0.17534637, Global Avg Loss: 0.79703075, Time: 0.0208 Steps: 63180, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001449, Sample Num: 23184, Cur Loss: 0.18863055, Cur Avg Loss: 0.17506244, Log Avg loss: 0.23351680, Global Avg Loss: 0.79694157, Time: 0.0209 Steps: 63190, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001459, Sample Num: 23344, Cur Loss: 0.10285729, Cur Avg Loss: 0.17494664, Log Avg loss: 0.15816708, Global Avg Loss: 0.79684050, Time: 0.0208 Steps: 63200, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001469, Sample Num: 23504, Cur Loss: 0.17907739, Cur Avg Loss: 0.17466363, Log Avg loss: 0.13337225, Global Avg Loss: 0.79673554, Time: 0.0208 Steps: 63210, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001479, Sample Num: 23664, Cur Loss: 0.17553985, Cur Avg Loss: 0.17501972, Log Avg loss: 0.22732978, Global Avg Loss: 0.79664547, Time: 0.0208 Steps: 63220, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001489, Sample Num: 23824, Cur Loss: 0.06534707, Cur Avg Loss: 0.17477441, Log Avg loss: 0.13849233, Global Avg Loss: 0.79654138, Time: 0.0209 Steps: 63230, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001499, Sample Num: 23984, Cur Loss: 0.14008060, Cur Avg Loss: 0.17461980, Log Avg loss: 0.15159892, Global Avg Loss: 0.79643940, Time: 0.0208 Steps: 63240, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001509, Sample Num: 24144, Cur Loss: 0.16027482, Cur Avg Loss: 0.17430733, Log Avg loss: 0.12746783, Global Avg Loss: 0.79633363, Time: 0.0209 Steps: 63250, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001519, Sample Num: 24304, Cur Loss: 0.53280389, Cur Avg Loss: 0.17475684, Log Avg loss: 0.24258859, Global Avg Loss: 0.79624610, Time: 0.0208 Steps: 63260, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001529, Sample Num: 24464, Cur Loss: 0.04832795, Cur Avg Loss: 0.17441917, Log Avg loss: 0.12312561, Global Avg Loss: 0.79613971, Time: 0.0208 Steps: 63270, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001539, Sample Num: 24624, Cur Loss: 0.26475209, Cur Avg Loss: 0.17491826, Log Avg loss: 0.25123000, Global Avg Loss: 0.79605360, Time: 0.0245 Steps: 63280, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001549, Sample Num: 24784, Cur Loss: 0.48746604, Cur Avg Loss: 0.17540375, Log Avg loss: 0.25011998, Global Avg Loss: 0.79596734, Time: 0.0209 Steps: 63290, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001559, Sample Num: 24944, Cur Loss: 0.08384013, Cur Avg Loss: 0.17634479, Log Avg loss: 0.32211247, Global Avg Loss: 0.79589248, Time: 0.0209 Steps: 63300, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001569, Sample Num: 25104, Cur Loss: 0.05395809, Cur Avg Loss: 0.17610410, Log Avg loss: 0.13858066, Global Avg Loss: 0.79578866, Time: 0.0209 Steps: 63310, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001579, Sample Num: 25264, Cur Loss: 0.10136206, Cur Avg Loss: 0.17606106, Log Avg loss: 0.16930825, Global Avg Loss: 0.79568972, Time: 0.0209 Steps: 63320, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001589, Sample Num: 25424, Cur Loss: 0.22136262, Cur Avg Loss: 0.17615395, Log Avg loss: 0.19082025, Global Avg Loss: 0.79559421, Time: 0.0209 Steps: 63330, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001599, Sample Num: 25584, Cur Loss: 0.19207068, Cur Avg Loss: 0.17636209, Log Avg loss: 0.20943700, Global Avg Loss: 0.79550167, Time: 0.0209 Steps: 63340, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001609, Sample Num: 25744, Cur Loss: 0.20356782, Cur Avg Loss: 0.17625847, Log Avg loss: 0.15968902, Global Avg Loss: 0.79540130, Time: 0.0208 Steps: 63350, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001619, Sample Num: 25904, Cur Loss: 0.16408211, Cur Avg Loss: 0.17610375, Log Avg loss: 0.15120969, Global Avg Loss: 0.79529963, Time: 0.0210 Steps: 63360, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001629, Sample Num: 26064, Cur Loss: 0.08554666, Cur Avg Loss: 0.17617575, Log Avg loss: 0.18783165, Global Avg Loss: 0.79520377, Time: 0.0209 Steps: 63370, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001639, Sample Num: 26224, Cur Loss: 0.32241842, Cur Avg Loss: 0.17611319, Log Avg loss: 0.16592260, Global Avg Loss: 0.79510448, Time: 0.0209 Steps: 63380, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001649, Sample Num: 26384, Cur Loss: 0.22246093, Cur Avg Loss: 0.17603832, Log Avg loss: 0.16376632, Global Avg Loss: 0.79500489, Time: 0.0208 Steps: 63390, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001659, Sample Num: 26544, Cur Loss: 0.04790108, Cur Avg Loss: 0.17608826, Log Avg loss: 0.18432360, Global Avg Loss: 0.79490856, Time: 0.0208 Steps: 63400, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001669, Sample Num: 26704, Cur Loss: 0.23734368, Cur Avg Loss: 0.17639062, Log Avg loss: 0.22655189, Global Avg Loss: 0.79481893, Time: 0.0208 Steps: 63410, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001679, Sample Num: 26864, Cur Loss: 0.28763244, Cur Avg Loss: 0.17652371, Log Avg loss: 0.19873664, Global Avg Loss: 0.79472494, Time: 0.0208 Steps: 63420, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001689, Sample Num: 27024, Cur Loss: 0.28024310, Cur Avg Loss: 0.17641219, Log Avg loss: 0.15768909, Global Avg Loss: 0.79462451, Time: 0.0209 Steps: 63430, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001699, Sample Num: 27184, Cur Loss: 0.22764535, Cur Avg Loss: 0.17647819, Log Avg loss: 0.18762422, Global Avg Loss: 0.79452883, Time: 0.0209 Steps: 63440, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001709, Sample Num: 27344, Cur Loss: 0.14704964, Cur Avg Loss: 0.17650507, Log Avg loss: 0.18107253, Global Avg Loss: 0.79443215, Time: 0.0209 Steps: 63450, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001719, Sample Num: 27504, Cur Loss: 0.05615842, Cur Avg Loss: 0.17627207, Log Avg loss: 0.13645192, Global Avg Loss: 0.79432846, Time: 0.0208 Steps: 63460, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001729, Sample Num: 27664, Cur Loss: 0.13884869, Cur Avg Loss: 0.17600934, Log Avg loss: 0.13084675, Global Avg Loss: 0.79422393, Time: 0.0209 Steps: 63470, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001739, Sample Num: 27824, Cur Loss: 0.14134222, Cur Avg Loss: 0.17615572, Log Avg loss: 0.20146547, Global Avg Loss: 0.79413055, Time: 0.0208 Steps: 63480, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001749, Sample Num: 27984, Cur Loss: 0.23792890, Cur Avg Loss: 0.17640300, Log Avg loss: 0.21940504, Global Avg Loss: 0.79404003, Time: 0.0209 Steps: 63490, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001759, Sample Num: 28144, Cur Loss: 0.16389903, Cur Avg Loss: 0.17607312, Log Avg loss: 0.11837727, Global Avg Loss: 0.79393362, Time: 0.0208 Steps: 63500, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001769, Sample Num: 28304, Cur Loss: 0.14001226, Cur Avg Loss: 0.17635835, Log Avg loss: 0.22652966, Global Avg Loss: 0.79384428, Time: 0.0208 Steps: 63510, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001779, Sample Num: 28464, Cur Loss: 0.46939948, Cur Avg Loss: 0.17686971, Log Avg loss: 0.26732905, Global Avg Loss: 0.79376139, Time: 0.0208 Steps: 63520, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001789, Sample Num: 28624, Cur Loss: 0.19496310, Cur Avg Loss: 0.17694762, Log Avg loss: 0.19080703, Global Avg Loss: 0.79366649, Time: 0.0208 Steps: 63530, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001799, Sample Num: 28784, Cur Loss: 0.37841237, Cur Avg Loss: 0.17672488, Log Avg loss: 0.13687741, Global Avg Loss: 0.79356312, Time: 0.0220 Steps: 63540, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001809, Sample Num: 28944, Cur Loss: 0.17027494, Cur Avg Loss: 0.17646519, Log Avg loss: 0.12974786, Global Avg Loss: 0.79345866, Time: 0.0219 Steps: 63550, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001819, Sample Num: 29104, Cur Loss: 0.07038322, Cur Avg Loss: 0.17645563, Log Avg loss: 0.17472514, Global Avg Loss: 0.79336132, Time: 0.0219 Steps: 63560, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001829, Sample Num: 29264, Cur Loss: 0.14143679, Cur Avg Loss: 0.17634682, Log Avg loss: 0.15655421, Global Avg Loss: 0.79326114, Time: 0.0219 Steps: 63570, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001839, Sample Num: 29424, Cur Loss: 0.09353070, Cur Avg Loss: 0.17661124, Log Avg loss: 0.22497436, Global Avg Loss: 0.79317176, Time: 0.0219 Steps: 63580, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001849, Sample Num: 29584, Cur Loss: 0.24221817, Cur Avg Loss: 0.17720141, Log Avg loss: 0.28573393, Global Avg Loss: 0.79309196, Time: 0.0219 Steps: 63590, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001859, Sample Num: 29744, Cur Loss: 0.14808932, Cur Avg Loss: 0.17697057, Log Avg loss: 0.13428734, Global Avg Loss: 0.79298838, Time: 0.0220 Steps: 63600, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001869, Sample Num: 29904, Cur Loss: 0.36333925, Cur Avg Loss: 0.17668932, Log Avg loss: 0.12440507, Global Avg Loss: 0.79288327, Time: 0.0219 Steps: 63610, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001879, Sample Num: 30064, Cur Loss: 0.12587799, Cur Avg Loss: 0.17672807, Log Avg loss: 0.18397005, Global Avg Loss: 0.79278756, Time: 0.0219 Steps: 63620, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001889, Sample Num: 30224, Cur Loss: 0.14937493, Cur Avg Loss: 0.17704126, Log Avg loss: 0.23589099, Global Avg Loss: 0.79270004, Time: 0.0219 Steps: 63630, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001899, Sample Num: 30384, Cur Loss: 0.08241346, Cur Avg Loss: 0.17691088, Log Avg loss: 0.15228125, Global Avg Loss: 0.79259941, Time: 0.0219 Steps: 63640, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001909, Sample Num: 30544, Cur Loss: 0.14242066, Cur Avg Loss: 0.17690671, Log Avg loss: 0.17611522, Global Avg Loss: 0.79250255, Time: 0.0218 Steps: 63650, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001919, Sample Num: 30704, Cur Loss: 0.09834066, Cur Avg Loss: 0.17684243, Log Avg loss: 0.16457063, Global Avg Loss: 0.79240391, Time: 0.0219 Steps: 63660, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001929, Sample Num: 30864, Cur Loss: 0.31920525, Cur Avg Loss: 0.17745483, Log Avg loss: 0.29497421, Global Avg Loss: 0.79232579, Time: 0.0219 Steps: 63670, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001939, Sample Num: 31024, Cur Loss: 0.08164561, Cur Avg Loss: 0.17734101, Log Avg loss: 0.15538608, Global Avg Loss: 0.79222577, Time: 0.0219 Steps: 63680, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001949, Sample Num: 31184, Cur Loss: 0.22871540, Cur Avg Loss: 0.17723199, Log Avg loss: 0.15609300, Global Avg Loss: 0.79212589, Time: 0.0219 Steps: 63690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001959, Sample Num: 31344, Cur Loss: 0.37730128, Cur Avg Loss: 0.17764439, Log Avg loss: 0.25802069, Global Avg Loss: 0.79204204, Time: 0.0219 Steps: 63700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001969, Sample Num: 31504, Cur Loss: 0.30292916, Cur Avg Loss: 0.17857281, Log Avg loss: 0.36045123, Global Avg Loss: 0.79197430, Time: 0.0219 Steps: 63710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001979, Sample Num: 31664, Cur Loss: 0.28023857, Cur Avg Loss: 0.17884054, Log Avg loss: 0.23155507, Global Avg Loss: 0.79188635, Time: 0.0219 Steps: 63720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001989, Sample Num: 31824, Cur Loss: 0.10736504, Cur Avg Loss: 0.17872391, Log Avg loss: 0.15564438, Global Avg Loss: 0.79178651, Time: 0.0219 Steps: 63730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001999, Sample Num: 31984, Cur Loss: 0.07120338, Cur Avg Loss: 0.17883347, Log Avg loss: 0.20062471, Global Avg Loss: 0.79169377, Time: 0.0219 Steps: 63740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 002009, Sample Num: 32144, Cur Loss: 0.10277718, Cur Avg Loss: 0.17913860, Log Avg loss: 0.24013264, Global Avg Loss: 0.79160725, Time: 0.0219 Steps: 63750, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002019, Sample Num: 32304, Cur Loss: 0.07173693, Cur Avg Loss: 0.17927684, Log Avg loss: 0.20705034, Global Avg Loss: 0.79151557, Time: 0.0219 Steps: 63760, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002029, Sample Num: 32464, Cur Loss: 0.01914001, Cur Avg Loss: 0.17956750, Log Avg loss: 0.23825140, Global Avg Loss: 0.79142881, Time: 0.0219 Steps: 63770, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002039, Sample Num: 32624, Cur Loss: 0.34769899, Cur Avg Loss: 0.17961179, Log Avg loss: 0.18859846, Global Avg Loss: 0.79133429, Time: 0.0219 Steps: 63780, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002049, Sample Num: 32784, Cur Loss: 0.15865600, Cur Avg Loss: 0.17955876, Log Avg loss: 0.16874618, Global Avg Loss: 0.79123669, Time: 0.0245 Steps: 63790, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002059, Sample Num: 32944, Cur Loss: 0.24154264, Cur Avg Loss: 0.17942823, Log Avg loss: 0.15268173, Global Avg Loss: 0.79113660, Time: 0.0208 Steps: 63800, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002069, Sample Num: 33104, Cur Loss: 0.21024284, Cur Avg Loss: 0.17912595, Log Avg loss: 0.11688780, Global Avg Loss: 0.79103094, Time: 0.0208 Steps: 63810, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002079, Sample Num: 33264, Cur Loss: 0.13284004, Cur Avg Loss: 0.17908293, Log Avg loss: 0.17018059, Global Avg Loss: 0.79093366, Time: 0.0210 Steps: 63820, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002089, Sample Num: 33424, Cur Loss: 0.14493231, Cur Avg Loss: 0.17901208, Log Avg loss: 0.16428311, Global Avg Loss: 0.79083548, Time: 0.0208 Steps: 63830, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002099, Sample Num: 33584, Cur Loss: 0.05841571, Cur Avg Loss: 0.17909725, Log Avg loss: 0.19688932, Global Avg Loss: 0.79074244, Time: 0.0208 Steps: 63840, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002109, Sample Num: 33744, Cur Loss: 0.13841674, Cur Avg Loss: 0.17890079, Log Avg loss: 0.13766302, Global Avg Loss: 0.79064016, Time: 0.0208 Steps: 63850, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002119, Sample Num: 33904, Cur Loss: 0.22990873, Cur Avg Loss: 0.17906798, Log Avg loss: 0.21432839, Global Avg Loss: 0.79054992, Time: 0.0208 Steps: 63860, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002129, Sample Num: 34055, Cur Loss: 0.11940827, Cur Avg Loss: 0.17963264, Log Avg loss: 0.29928464, Global Avg Loss: 0.79047300, Time: 0.0101 Steps: 63870, Updated lr: 0.000040 ***** Running evaluation checkpoint-63870 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-63870 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.800528, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.239382, "eval_total_loss": 168.28539, "eval_mae": 0.373733, "eval_mse": 0.239429, "eval_r2": 0.847803, "eval_sp_statistic": 0.9243, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.931033, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.286912, "test_total_loss": 144.029598, "test_mae": 0.448024, "test_mse": 0.286992, "test_r2": 0.814773, "test_sp_statistic": 0.91149, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.941254, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7904729989458489, "train_cur_epoch_loss": 382.4378879284486, "train_cur_epoch_avg_loss": 0.17963263876394955, "train_cur_epoch_time": 44.800528049468994, "train_cur_epoch_avg_time": 0.021042991098858148, "epoch": 30, "step": 63870} ################################################## Training, Epoch: 0031, Batch: 000010, Sample Num: 160, Cur Loss: 0.07599482, Cur Avg Loss: 0.30291052, Log Avg loss: 0.30291052, Global Avg Loss: 0.79039667, Time: 0.0209 Steps: 63880, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000020, Sample Num: 320, Cur Loss: 0.10605906, Cur Avg Loss: 0.23738053, Log Avg loss: 0.17185054, Global Avg Loss: 0.79029986, Time: 0.0209 Steps: 63890, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000030, Sample Num: 480, Cur Loss: 0.20612241, Cur Avg Loss: 0.22604473, Log Avg loss: 0.20337314, Global Avg Loss: 0.79020801, Time: 0.0209 Steps: 63900, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000040, Sample Num: 640, Cur Loss: 0.26893044, Cur Avg Loss: 0.20608455, Log Avg loss: 0.14620402, Global Avg Loss: 0.79010724, Time: 0.0209 Steps: 63910, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000050, Sample Num: 800, Cur Loss: 0.12433706, Cur Avg Loss: 0.19003856, Log Avg loss: 0.12585460, Global Avg Loss: 0.79000332, Time: 0.0209 Steps: 63920, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000060, Sample Num: 960, Cur Loss: 0.04674845, Cur Avg Loss: 0.18093098, Log Avg loss: 0.13539310, Global Avg Loss: 0.78990093, Time: 0.0210 Steps: 63930, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000070, Sample Num: 1120, Cur Loss: 0.23784201, Cur Avg Loss: 0.18318856, Log Avg loss: 0.19673398, Global Avg Loss: 0.78980816, Time: 0.0210 Steps: 63940, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000080, Sample Num: 1280, Cur Loss: 0.08393225, Cur Avg Loss: 0.18405400, Log Avg loss: 0.19011212, Global Avg Loss: 0.78971438, Time: 0.0209 Steps: 63950, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000090, Sample Num: 1440, Cur Loss: 0.11957679, Cur Avg Loss: 0.17706734, Log Avg loss: 0.12117407, Global Avg Loss: 0.78960986, Time: 0.0209 Steps: 63960, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000100, Sample Num: 1600, Cur Loss: 0.15872994, Cur Avg Loss: 0.17702121, Log Avg loss: 0.17660598, Global Avg Loss: 0.78951403, Time: 0.0209 Steps: 63970, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000110, Sample Num: 1760, Cur Loss: 0.04013746, Cur Avg Loss: 0.17197086, Log Avg loss: 0.12146736, Global Avg Loss: 0.78940962, Time: 0.0209 Steps: 63980, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000120, Sample Num: 1920, Cur Loss: 0.35836148, Cur Avg Loss: 0.17304359, Log Avg loss: 0.18484370, Global Avg Loss: 0.78931514, Time: 0.0210 Steps: 63990, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000130, Sample Num: 2080, Cur Loss: 0.32858783, Cur Avg Loss: 0.17359461, Log Avg loss: 0.18020677, Global Avg Loss: 0.78921996, Time: 0.0209 Steps: 64000, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000140, Sample Num: 2240, Cur Loss: 0.04231760, Cur Avg Loss: 0.17090214, Log Avg loss: 0.13590005, Global Avg Loss: 0.78911790, Time: 0.0209 Steps: 64010, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000150, Sample Num: 2400, Cur Loss: 0.51677370, Cur Avg Loss: 0.17394110, Log Avg loss: 0.21648660, Global Avg Loss: 0.78902845, Time: 0.0209 Steps: 64020, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000160, Sample Num: 2560, Cur Loss: 0.12828538, Cur Avg Loss: 0.17178068, Log Avg loss: 0.13937431, Global Avg Loss: 0.78892699, Time: 0.0209 Steps: 64030, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000170, Sample Num: 2720, Cur Loss: 0.10233748, Cur Avg Loss: 0.17086724, Log Avg loss: 0.15625228, Global Avg Loss: 0.78882820, Time: 0.0208 Steps: 64040, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000180, Sample Num: 2880, Cur Loss: 0.15114211, Cur Avg Loss: 0.17152085, Log Avg loss: 0.18263225, Global Avg Loss: 0.78873355, Time: 0.0209 Steps: 64050, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000190, Sample Num: 3040, Cur Loss: 0.37629968, Cur Avg Loss: 0.17222381, Log Avg loss: 0.18487703, Global Avg Loss: 0.78863929, Time: 0.0210 Steps: 64060, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000200, Sample Num: 3200, Cur Loss: 0.11243865, Cur Avg Loss: 0.17895658, Log Avg loss: 0.30687925, Global Avg Loss: 0.78856410, Time: 0.0209 Steps: 64070, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000210, Sample Num: 3360, Cur Loss: 0.11608104, Cur Avg Loss: 0.17944370, Log Avg loss: 0.18918613, Global Avg Loss: 0.78847056, Time: 0.0210 Steps: 64080, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000220, Sample Num: 3520, Cur Loss: 0.17377767, Cur Avg Loss: 0.17795878, Log Avg loss: 0.14677527, Global Avg Loss: 0.78837044, Time: 0.0209 Steps: 64090, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000230, Sample Num: 3680, Cur Loss: 0.10348509, Cur Avg Loss: 0.17545980, Log Avg loss: 0.12048228, Global Avg Loss: 0.78826624, Time: 0.0209 Steps: 64100, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000240, Sample Num: 3840, Cur Loss: 0.30196512, Cur Avg Loss: 0.17528088, Log Avg loss: 0.17116576, Global Avg Loss: 0.78816999, Time: 0.0209 Steps: 64110, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000250, Sample Num: 4000, Cur Loss: 0.05640965, Cur Avg Loss: 0.17505879, Log Avg loss: 0.16972854, Global Avg Loss: 0.78807354, Time: 0.0208 Steps: 64120, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000260, Sample Num: 4160, Cur Loss: 0.02635507, Cur Avg Loss: 0.17308370, Log Avg loss: 0.12370644, Global Avg Loss: 0.78796994, Time: 0.0246 Steps: 64130, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000270, Sample Num: 4320, Cur Loss: 0.25520608, Cur Avg Loss: 0.17351047, Log Avg loss: 0.18460664, Global Avg Loss: 0.78787587, Time: 0.0209 Steps: 64140, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000280, Sample Num: 4480, Cur Loss: 0.19491768, Cur Avg Loss: 0.17200353, Log Avg loss: 0.13131622, Global Avg Loss: 0.78777352, Time: 0.0211 Steps: 64150, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000290, Sample Num: 4640, Cur Loss: 0.09743783, Cur Avg Loss: 0.17104413, Log Avg loss: 0.14418083, Global Avg Loss: 0.78767321, Time: 0.0209 Steps: 64160, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000300, Sample Num: 4800, Cur Loss: 0.12656322, Cur Avg Loss: 0.16976914, Log Avg loss: 0.13279434, Global Avg Loss: 0.78757116, Time: 0.0210 Steps: 64170, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000310, Sample Num: 4960, Cur Loss: 0.18198888, Cur Avg Loss: 0.16746059, Log Avg loss: 0.09820423, Global Avg Loss: 0.78746375, Time: 0.0209 Steps: 64180, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000320, Sample Num: 5120, Cur Loss: 0.21161973, Cur Avg Loss: 0.16690624, Log Avg loss: 0.14972136, Global Avg Loss: 0.78736439, Time: 0.0210 Steps: 64190, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000330, Sample Num: 5280, Cur Loss: 0.10769795, Cur Avg Loss: 0.16640314, Log Avg loss: 0.15030406, Global Avg Loss: 0.78726516, Time: 0.0209 Steps: 64200, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000340, Sample Num: 5440, Cur Loss: 0.04205035, Cur Avg Loss: 0.16917832, Log Avg loss: 0.26075896, Global Avg Loss: 0.78718317, Time: 0.0209 Steps: 64210, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000350, Sample Num: 5600, Cur Loss: 0.18065490, Cur Avg Loss: 0.16854572, Log Avg loss: 0.14703760, Global Avg Loss: 0.78708349, Time: 0.0209 Steps: 64220, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000360, Sample Num: 5760, Cur Loss: 0.07053689, Cur Avg Loss: 0.16797385, Log Avg loss: 0.14795844, Global Avg Loss: 0.78698398, Time: 0.0210 Steps: 64230, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000370, Sample Num: 5920, Cur Loss: 0.04002429, Cur Avg Loss: 0.16839150, Log Avg loss: 0.18342676, Global Avg Loss: 0.78689003, Time: 0.0209 Steps: 64240, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000380, Sample Num: 6080, Cur Loss: 0.21983542, Cur Avg Loss: 0.16830581, Log Avg loss: 0.16513516, Global Avg Loss: 0.78679326, Time: 0.0209 Steps: 64250, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000390, Sample Num: 6240, Cur Loss: 0.09405307, Cur Avg Loss: 0.16812858, Log Avg loss: 0.16139406, Global Avg Loss: 0.78669593, Time: 0.0210 Steps: 64260, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000400, Sample Num: 6400, Cur Loss: 0.50624859, Cur Avg Loss: 0.17036029, Log Avg loss: 0.25739688, Global Avg Loss: 0.78661358, Time: 0.0209 Steps: 64270, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000410, Sample Num: 6560, Cur Loss: 0.36851996, Cur Avg Loss: 0.17206966, Log Avg loss: 0.24044451, Global Avg Loss: 0.78652861, Time: 0.0209 Steps: 64280, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000420, Sample Num: 6720, Cur Loss: 0.06284343, Cur Avg Loss: 0.17210632, Log Avg loss: 0.17360938, Global Avg Loss: 0.78643327, Time: 0.0210 Steps: 64290, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000430, Sample Num: 6880, Cur Loss: 0.15976879, Cur Avg Loss: 0.17370306, Log Avg loss: 0.24076595, Global Avg Loss: 0.78634841, Time: 0.0209 Steps: 64300, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000440, Sample Num: 7040, Cur Loss: 0.25594795, Cur Avg Loss: 0.17364764, Log Avg loss: 0.17126464, Global Avg Loss: 0.78625277, Time: 0.0209 Steps: 64310, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000450, Sample Num: 7200, Cur Loss: 0.13284831, Cur Avg Loss: 0.17331988, Log Avg loss: 0.15889863, Global Avg Loss: 0.78615523, Time: 0.0210 Steps: 64320, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000460, Sample Num: 7360, Cur Loss: 0.18811375, Cur Avg Loss: 0.17336528, Log Avg loss: 0.17540819, Global Avg Loss: 0.78606029, Time: 0.0209 Steps: 64330, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000470, Sample Num: 7520, Cur Loss: 0.23770234, Cur Avg Loss: 0.17280055, Log Avg loss: 0.14682277, Global Avg Loss: 0.78596094, Time: 0.0209 Steps: 64340, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000480, Sample Num: 7680, Cur Loss: 0.03728857, Cur Avg Loss: 0.17171603, Log Avg loss: 0.12074359, Global Avg Loss: 0.78585756, Time: 0.0209 Steps: 64350, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000490, Sample Num: 7840, Cur Loss: 0.17308608, Cur Avg Loss: 0.17290532, Log Avg loss: 0.22999144, Global Avg Loss: 0.78577119, Time: 0.0210 Steps: 64360, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000500, Sample Num: 8000, Cur Loss: 0.26179579, Cur Avg Loss: 0.17419959, Log Avg loss: 0.23761867, Global Avg Loss: 0.78568604, Time: 0.0208 Steps: 64370, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000510, Sample Num: 8160, Cur Loss: 0.16551417, Cur Avg Loss: 0.17337076, Log Avg loss: 0.13192956, Global Avg Loss: 0.78558449, Time: 0.0210 Steps: 64380, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000520, Sample Num: 8320, Cur Loss: 0.43175131, Cur Avg Loss: 0.17462655, Log Avg loss: 0.23867152, Global Avg Loss: 0.78549955, Time: 0.0210 Steps: 64390, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000530, Sample Num: 8480, Cur Loss: 0.12914911, Cur Avg Loss: 0.17407363, Log Avg loss: 0.14532198, Global Avg Loss: 0.78540015, Time: 0.0210 Steps: 64400, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000540, Sample Num: 8640, Cur Loss: 0.15880485, Cur Avg Loss: 0.17311018, Log Avg loss: 0.12204717, Global Avg Loss: 0.78529716, Time: 0.0210 Steps: 64410, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000550, Sample Num: 8800, Cur Loss: 0.19861758, Cur Avg Loss: 0.17239863, Log Avg loss: 0.13397483, Global Avg Loss: 0.78519605, Time: 0.0209 Steps: 64420, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000560, Sample Num: 8960, Cur Loss: 0.24776395, Cur Avg Loss: 0.17301277, Log Avg loss: 0.20679051, Global Avg Loss: 0.78510628, Time: 0.0210 Steps: 64430, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000570, Sample Num: 9120, Cur Loss: 0.04611069, Cur Avg Loss: 0.17303200, Log Avg loss: 0.17410929, Global Avg Loss: 0.78501146, Time: 0.0210 Steps: 64440, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000580, Sample Num: 9280, Cur Loss: 0.16798706, Cur Avg Loss: 0.17243094, Log Avg loss: 0.13817023, Global Avg Loss: 0.78491110, Time: 0.0211 Steps: 64450, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000590, Sample Num: 9440, Cur Loss: 0.31171435, Cur Avg Loss: 0.17218872, Log Avg loss: 0.15813994, Global Avg Loss: 0.78481387, Time: 0.0210 Steps: 64460, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000600, Sample Num: 9600, Cur Loss: 0.23592283, Cur Avg Loss: 0.17228312, Log Avg loss: 0.17785295, Global Avg Loss: 0.78471972, Time: 0.0210 Steps: 64470, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000610, Sample Num: 9760, Cur Loss: 0.15818080, Cur Avg Loss: 0.17128148, Log Avg loss: 0.11118312, Global Avg Loss: 0.78461526, Time: 0.0210 Steps: 64480, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000620, Sample Num: 9920, Cur Loss: 0.07297941, Cur Avg Loss: 0.17057236, Log Avg loss: 0.12731610, Global Avg Loss: 0.78451334, Time: 0.0210 Steps: 64490, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000630, Sample Num: 10080, Cur Loss: 0.07068985, Cur Avg Loss: 0.17007972, Log Avg loss: 0.13953570, Global Avg Loss: 0.78441334, Time: 0.0210 Steps: 64500, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000640, Sample Num: 10240, Cur Loss: 0.06105185, Cur Avg Loss: 0.17035654, Log Avg loss: 0.18779603, Global Avg Loss: 0.78432086, Time: 0.0210 Steps: 64510, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000650, Sample Num: 10400, Cur Loss: 0.07780752, Cur Avg Loss: 0.17018175, Log Avg loss: 0.15899558, Global Avg Loss: 0.78422394, Time: 0.0210 Steps: 64520, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000660, Sample Num: 10560, Cur Loss: 0.02191271, Cur Avg Loss: 0.16988091, Log Avg loss: 0.15032633, Global Avg Loss: 0.78412571, Time: 0.0210 Steps: 64530, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000670, Sample Num: 10720, Cur Loss: 0.07386118, Cur Avg Loss: 0.16981028, Log Avg loss: 0.16514876, Global Avg Loss: 0.78402980, Time: 0.0209 Steps: 64540, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000680, Sample Num: 10880, Cur Loss: 0.17355452, Cur Avg Loss: 0.17062676, Log Avg loss: 0.22533070, Global Avg Loss: 0.78394325, Time: 0.0210 Steps: 64550, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000690, Sample Num: 11040, Cur Loss: 0.17922974, Cur Avg Loss: 0.17045937, Log Avg loss: 0.15907694, Global Avg Loss: 0.78384646, Time: 0.0209 Steps: 64560, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000700, Sample Num: 11200, Cur Loss: 0.23886275, Cur Avg Loss: 0.17085090, Log Avg loss: 0.19786663, Global Avg Loss: 0.78375571, Time: 0.0210 Steps: 64570, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000710, Sample Num: 11360, Cur Loss: 0.21765085, Cur Avg Loss: 0.17087888, Log Avg loss: 0.17283733, Global Avg Loss: 0.78366111, Time: 0.0209 Steps: 64580, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000720, Sample Num: 11520, Cur Loss: 0.07816230, Cur Avg Loss: 0.17113193, Log Avg loss: 0.18909808, Global Avg Loss: 0.78356906, Time: 0.0210 Steps: 64590, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000730, Sample Num: 11680, Cur Loss: 0.08064897, Cur Avg Loss: 0.17118007, Log Avg loss: 0.17464666, Global Avg Loss: 0.78347480, Time: 0.0209 Steps: 64600, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000740, Sample Num: 11840, Cur Loss: 0.47942191, Cur Avg Loss: 0.17147806, Log Avg loss: 0.19323132, Global Avg Loss: 0.78338344, Time: 0.0211 Steps: 64610, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000750, Sample Num: 12000, Cur Loss: 0.19242808, Cur Avg Loss: 0.17200156, Log Avg loss: 0.21074072, Global Avg Loss: 0.78329483, Time: 0.0210 Steps: 64620, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000760, Sample Num: 12160, Cur Loss: 0.07759742, Cur Avg Loss: 0.17272554, Log Avg loss: 0.22702370, Global Avg Loss: 0.78320876, Time: 0.0210 Steps: 64630, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000770, Sample Num: 12320, Cur Loss: 0.15912542, Cur Avg Loss: 0.17255965, Log Avg loss: 0.15995234, Global Avg Loss: 0.78311234, Time: 0.0247 Steps: 64640, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000780, Sample Num: 12480, Cur Loss: 0.13896126, Cur Avg Loss: 0.17219997, Log Avg loss: 0.14450406, Global Avg Loss: 0.78301356, Time: 0.0209 Steps: 64650, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000790, Sample Num: 12640, Cur Loss: 0.04969528, Cur Avg Loss: 0.17155389, Log Avg loss: 0.12115967, Global Avg Loss: 0.78291120, Time: 0.0209 Steps: 64660, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000800, Sample Num: 12800, Cur Loss: 0.16013288, Cur Avg Loss: 0.17151809, Log Avg loss: 0.16869039, Global Avg Loss: 0.78281622, Time: 0.0209 Steps: 64670, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000810, Sample Num: 12960, Cur Loss: 0.24268109, Cur Avg Loss: 0.17127183, Log Avg loss: 0.15157055, Global Avg Loss: 0.78271862, Time: 0.0209 Steps: 64680, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000820, Sample Num: 13120, Cur Loss: 0.10989363, Cur Avg Loss: 0.17077262, Log Avg loss: 0.13033642, Global Avg Loss: 0.78261778, Time: 0.0209 Steps: 64690, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000830, Sample Num: 13280, Cur Loss: 0.34156922, Cur Avg Loss: 0.17099728, Log Avg loss: 0.18941960, Global Avg Loss: 0.78252609, Time: 0.0209 Steps: 64700, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000840, Sample Num: 13440, Cur Loss: 0.13586046, Cur Avg Loss: 0.17114882, Log Avg loss: 0.18372672, Global Avg Loss: 0.78243356, Time: 0.0209 Steps: 64710, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000850, Sample Num: 13600, Cur Loss: 0.10695416, Cur Avg Loss: 0.17131785, Log Avg loss: 0.18551669, Global Avg Loss: 0.78234133, Time: 0.0209 Steps: 64720, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000860, Sample Num: 13760, Cur Loss: 0.32748625, Cur Avg Loss: 0.17161104, Log Avg loss: 0.19653179, Global Avg Loss: 0.78225083, Time: 0.0209 Steps: 64730, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000870, Sample Num: 13920, Cur Loss: 0.26972297, Cur Avg Loss: 0.17305678, Log Avg loss: 0.29739067, Global Avg Loss: 0.78217593, Time: 0.0210 Steps: 64740, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000880, Sample Num: 14080, Cur Loss: 0.11080213, Cur Avg Loss: 0.17306992, Log Avg loss: 0.17421294, Global Avg Loss: 0.78208204, Time: 0.0209 Steps: 64750, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000890, Sample Num: 14240, Cur Loss: 0.06448621, Cur Avg Loss: 0.17350700, Log Avg loss: 0.21196986, Global Avg Loss: 0.78199400, Time: 0.0210 Steps: 64760, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000900, Sample Num: 14400, Cur Loss: 0.09000757, Cur Avg Loss: 0.17333230, Log Avg loss: 0.15778410, Global Avg Loss: 0.78189763, Time: 0.0209 Steps: 64770, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000910, Sample Num: 14560, Cur Loss: 0.15356588, Cur Avg Loss: 0.17334557, Log Avg loss: 0.17453963, Global Avg Loss: 0.78180387, Time: 0.0209 Steps: 64780, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000920, Sample Num: 14720, Cur Loss: 0.22611061, Cur Avg Loss: 0.17284263, Log Avg loss: 0.12707510, Global Avg Loss: 0.78170282, Time: 0.0209 Steps: 64790, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000930, Sample Num: 14880, Cur Loss: 0.06406568, Cur Avg Loss: 0.17219310, Log Avg loss: 0.11243630, Global Avg Loss: 0.78159954, Time: 0.0209 Steps: 64800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000940, Sample Num: 15040, Cur Loss: 0.23461431, Cur Avg Loss: 0.17201903, Log Avg loss: 0.15583120, Global Avg Loss: 0.78150298, Time: 0.0210 Steps: 64810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000950, Sample Num: 15200, Cur Loss: 0.56585038, Cur Avg Loss: 0.17244671, Log Avg loss: 0.21264816, Global Avg Loss: 0.78141522, Time: 0.0209 Steps: 64820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000960, Sample Num: 15360, Cur Loss: 0.06017559, Cur Avg Loss: 0.17217552, Log Avg loss: 0.14641223, Global Avg Loss: 0.78131727, Time: 0.0209 Steps: 64830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000970, Sample Num: 15520, Cur Loss: 0.09158892, Cur Avg Loss: 0.17253414, Log Avg loss: 0.20696201, Global Avg Loss: 0.78122869, Time: 0.0209 Steps: 64840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000980, Sample Num: 15680, Cur Loss: 0.08631764, Cur Avg Loss: 0.17285390, Log Avg loss: 0.20387049, Global Avg Loss: 0.78113966, Time: 0.0209 Steps: 64850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000990, Sample Num: 15840, Cur Loss: 0.24575087, Cur Avg Loss: 0.17336346, Log Avg loss: 0.22330083, Global Avg Loss: 0.78105366, Time: 0.0209 Steps: 64860, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001000, Sample Num: 16000, Cur Loss: 0.40203798, Cur Avg Loss: 0.17380596, Log Avg loss: 0.21761367, Global Avg Loss: 0.78096680, Time: 0.0209 Steps: 64870, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001010, Sample Num: 16160, Cur Loss: 0.47887972, Cur Avg Loss: 0.17374291, Log Avg loss: 0.16743729, Global Avg Loss: 0.78087224, Time: 0.0209 Steps: 64880, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001020, Sample Num: 16320, Cur Loss: 0.12556329, Cur Avg Loss: 0.17402382, Log Avg loss: 0.20239579, Global Avg Loss: 0.78078309, Time: 0.0209 Steps: 64890, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001030, Sample Num: 16480, Cur Loss: 0.54789674, Cur Avg Loss: 0.17415257, Log Avg loss: 0.18728510, Global Avg Loss: 0.78069164, Time: 0.0209 Steps: 64900, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001040, Sample Num: 16640, Cur Loss: 0.41632318, Cur Avg Loss: 0.17464983, Log Avg loss: 0.22586762, Global Avg Loss: 0.78060617, Time: 0.0208 Steps: 64910, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001050, Sample Num: 16800, Cur Loss: 0.63964629, Cur Avg Loss: 0.17489125, Log Avg loss: 0.19999944, Global Avg Loss: 0.78051673, Time: 0.0209 Steps: 64920, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001060, Sample Num: 16960, Cur Loss: 0.06427122, Cur Avg Loss: 0.17497983, Log Avg loss: 0.18428054, Global Avg Loss: 0.78042490, Time: 0.0209 Steps: 64930, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001070, Sample Num: 17120, Cur Loss: 0.23242766, Cur Avg Loss: 0.17507088, Log Avg loss: 0.18472146, Global Avg Loss: 0.78033317, Time: 0.0209 Steps: 64940, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001080, Sample Num: 17280, Cur Loss: 0.07649531, Cur Avg Loss: 0.17499601, Log Avg loss: 0.16698586, Global Avg Loss: 0.78023874, Time: 0.0209 Steps: 64950, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001090, Sample Num: 17440, Cur Loss: 0.20394491, Cur Avg Loss: 0.17443378, Log Avg loss: 0.11371254, Global Avg Loss: 0.78013613, Time: 0.0209 Steps: 64960, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001100, Sample Num: 17600, Cur Loss: 0.54050142, Cur Avg Loss: 0.17481792, Log Avg loss: 0.21668911, Global Avg Loss: 0.78004941, Time: 0.0208 Steps: 64970, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001110, Sample Num: 17760, Cur Loss: 0.17909372, Cur Avg Loss: 0.17522932, Log Avg loss: 0.22048363, Global Avg Loss: 0.77996330, Time: 0.0209 Steps: 64980, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001120, Sample Num: 17920, Cur Loss: 0.10085154, Cur Avg Loss: 0.17532909, Log Avg loss: 0.18640289, Global Avg Loss: 0.77987197, Time: 0.0209 Steps: 64990, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001130, Sample Num: 18080, Cur Loss: 0.11850125, Cur Avg Loss: 0.17498881, Log Avg loss: 0.13687734, Global Avg Loss: 0.77977304, Time: 0.0209 Steps: 65000, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001140, Sample Num: 18240, Cur Loss: 0.14349526, Cur Avg Loss: 0.17527575, Log Avg loss: 0.20770014, Global Avg Loss: 0.77968505, Time: 0.0209 Steps: 65010, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001150, Sample Num: 18400, Cur Loss: 0.06342050, Cur Avg Loss: 0.17505068, Log Avg loss: 0.14939278, Global Avg Loss: 0.77958811, Time: 0.0209 Steps: 65020, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001160, Sample Num: 18560, Cur Loss: 0.18929949, Cur Avg Loss: 0.17531671, Log Avg loss: 0.20591011, Global Avg Loss: 0.77949989, Time: 0.0209 Steps: 65030, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001170, Sample Num: 18720, Cur Loss: 0.17955193, Cur Avg Loss: 0.17500180, Log Avg loss: 0.13847198, Global Avg Loss: 0.77940133, Time: 0.0209 Steps: 65040, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001180, Sample Num: 18880, Cur Loss: 0.33118570, Cur Avg Loss: 0.17513818, Log Avg loss: 0.19109556, Global Avg Loss: 0.77931089, Time: 0.0209 Steps: 65050, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001190, Sample Num: 19040, Cur Loss: 0.04368455, Cur Avg Loss: 0.17496643, Log Avg loss: 0.15469927, Global Avg Loss: 0.77921489, Time: 0.0209 Steps: 65060, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001200, Sample Num: 19200, Cur Loss: 0.33549774, Cur Avg Loss: 0.17520954, Log Avg loss: 0.20414034, Global Avg Loss: 0.77912651, Time: 0.0209 Steps: 65070, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001210, Sample Num: 19360, Cur Loss: 0.19828323, Cur Avg Loss: 0.17563285, Log Avg loss: 0.22642916, Global Avg Loss: 0.77904158, Time: 0.0209 Steps: 65080, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001220, Sample Num: 19520, Cur Loss: 0.05471794, Cur Avg Loss: 0.17657661, Log Avg loss: 0.29077231, Global Avg Loss: 0.77896657, Time: 0.0208 Steps: 65090, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001230, Sample Num: 19680, Cur Loss: 0.08145439, Cur Avg Loss: 0.17609228, Log Avg loss: 0.11700318, Global Avg Loss: 0.77886488, Time: 0.0209 Steps: 65100, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001240, Sample Num: 19840, Cur Loss: 0.27754682, Cur Avg Loss: 0.17607968, Log Avg loss: 0.17453083, Global Avg Loss: 0.77877207, Time: 0.0209 Steps: 65110, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001250, Sample Num: 20000, Cur Loss: 0.23128365, Cur Avg Loss: 0.17573954, Log Avg loss: 0.13356120, Global Avg Loss: 0.77867299, Time: 0.0208 Steps: 65120, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001260, Sample Num: 20160, Cur Loss: 0.12461329, Cur Avg Loss: 0.17613578, Log Avg loss: 0.22566650, Global Avg Loss: 0.77858808, Time: 0.0209 Steps: 65130, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001270, Sample Num: 20320, Cur Loss: 0.25890297, Cur Avg Loss: 0.17629206, Log Avg loss: 0.19598289, Global Avg Loss: 0.77849864, Time: 0.0209 Steps: 65140, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001280, Sample Num: 20480, Cur Loss: 0.53438145, Cur Avg Loss: 0.17664829, Log Avg loss: 0.22189001, Global Avg Loss: 0.77841320, Time: 0.0254 Steps: 65150, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001290, Sample Num: 20640, Cur Loss: 0.21006519, Cur Avg Loss: 0.17698164, Log Avg loss: 0.21965014, Global Avg Loss: 0.77832745, Time: 0.0209 Steps: 65160, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001300, Sample Num: 20800, Cur Loss: 0.37492758, Cur Avg Loss: 0.17736356, Log Avg loss: 0.22663133, Global Avg Loss: 0.77824280, Time: 0.0209 Steps: 65170, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001310, Sample Num: 20960, Cur Loss: 0.18696544, Cur Avg Loss: 0.17729709, Log Avg loss: 0.16865612, Global Avg Loss: 0.77814927, Time: 0.0209 Steps: 65180, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001320, Sample Num: 21120, Cur Loss: 0.62260360, Cur Avg Loss: 0.17777036, Log Avg loss: 0.23976915, Global Avg Loss: 0.77806669, Time: 0.0209 Steps: 65190, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001330, Sample Num: 21280, Cur Loss: 0.29208189, Cur Avg Loss: 0.17748199, Log Avg loss: 0.13941659, Global Avg Loss: 0.77796873, Time: 0.0209 Steps: 65200, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001340, Sample Num: 21440, Cur Loss: 0.47270855, Cur Avg Loss: 0.17753814, Log Avg loss: 0.18500575, Global Avg Loss: 0.77787780, Time: 0.0209 Steps: 65210, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001350, Sample Num: 21600, Cur Loss: 0.10650267, Cur Avg Loss: 0.17727529, Log Avg loss: 0.14205342, Global Avg Loss: 0.77778031, Time: 0.0209 Steps: 65220, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001360, Sample Num: 21760, Cur Loss: 0.27091825, Cur Avg Loss: 0.17694649, Log Avg loss: 0.13255871, Global Avg Loss: 0.77768140, Time: 0.0209 Steps: 65230, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001370, Sample Num: 21920, Cur Loss: 0.22460808, Cur Avg Loss: 0.17667818, Log Avg loss: 0.14018778, Global Avg Loss: 0.77758368, Time: 0.0209 Steps: 65240, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001380, Sample Num: 22080, Cur Loss: 0.50343496, Cur Avg Loss: 0.17714932, Log Avg loss: 0.24169580, Global Avg Loss: 0.77750156, Time: 0.0209 Steps: 65250, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001390, Sample Num: 22240, Cur Loss: 0.14405566, Cur Avg Loss: 0.17718872, Log Avg loss: 0.18262619, Global Avg Loss: 0.77741040, Time: 0.0209 Steps: 65260, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001400, Sample Num: 22400, Cur Loss: 0.30832368, Cur Avg Loss: 0.17699439, Log Avg loss: 0.14998297, Global Avg Loss: 0.77731427, Time: 0.0209 Steps: 65270, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001410, Sample Num: 22560, Cur Loss: 0.15699266, Cur Avg Loss: 0.17703613, Log Avg loss: 0.18287944, Global Avg Loss: 0.77722321, Time: 0.0209 Steps: 65280, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001420, Sample Num: 22720, Cur Loss: 0.14555715, Cur Avg Loss: 0.17693639, Log Avg loss: 0.16287312, Global Avg Loss: 0.77712912, Time: 0.0210 Steps: 65290, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001430, Sample Num: 22880, Cur Loss: 0.07972819, Cur Avg Loss: 0.17689444, Log Avg loss: 0.17093774, Global Avg Loss: 0.77703629, Time: 0.0209 Steps: 65300, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001440, Sample Num: 23040, Cur Loss: 0.21651959, Cur Avg Loss: 0.17686169, Log Avg loss: 0.17217828, Global Avg Loss: 0.77694367, Time: 0.0209 Steps: 65310, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001450, Sample Num: 23200, Cur Loss: 0.02789994, Cur Avg Loss: 0.17717428, Log Avg loss: 0.22218632, Global Avg Loss: 0.77685874, Time: 0.0209 Steps: 65320, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001460, Sample Num: 23360, Cur Loss: 0.27839726, Cur Avg Loss: 0.17730677, Log Avg loss: 0.19651810, Global Avg Loss: 0.77676991, Time: 0.0209 Steps: 65330, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001470, Sample Num: 23520, Cur Loss: 0.17674522, Cur Avg Loss: 0.17712371, Log Avg loss: 0.15039711, Global Avg Loss: 0.77667405, Time: 0.0209 Steps: 65340, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001480, Sample Num: 23680, Cur Loss: 0.09708235, Cur Avg Loss: 0.17713324, Log Avg loss: 0.17853393, Global Avg Loss: 0.77658252, Time: 0.0209 Steps: 65350, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001490, Sample Num: 23840, Cur Loss: 0.10484059, Cur Avg Loss: 0.17710769, Log Avg loss: 0.17332688, Global Avg Loss: 0.77649022, Time: 0.0209 Steps: 65360, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001500, Sample Num: 24000, Cur Loss: 0.06366356, Cur Avg Loss: 0.17686868, Log Avg loss: 0.14125591, Global Avg Loss: 0.77639305, Time: 0.0209 Steps: 65370, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001510, Sample Num: 24160, Cur Loss: 0.15998147, Cur Avg Loss: 0.17685595, Log Avg loss: 0.17494610, Global Avg Loss: 0.77630105, Time: 0.0209 Steps: 65380, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001520, Sample Num: 24320, Cur Loss: 0.31388861, Cur Avg Loss: 0.17689021, Log Avg loss: 0.18206341, Global Avg Loss: 0.77621018, Time: 0.0209 Steps: 65390, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001530, Sample Num: 24480, Cur Loss: 0.08296952, Cur Avg Loss: 0.17650301, Log Avg loss: 0.11764894, Global Avg Loss: 0.77610948, Time: 0.0209 Steps: 65400, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001540, Sample Num: 24640, Cur Loss: 0.23470907, Cur Avg Loss: 0.17660944, Log Avg loss: 0.19289351, Global Avg Loss: 0.77602032, Time: 0.0247 Steps: 65410, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001550, Sample Num: 24800, Cur Loss: 0.10971126, Cur Avg Loss: 0.17636030, Log Avg loss: 0.13799230, Global Avg Loss: 0.77592279, Time: 0.0209 Steps: 65420, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001560, Sample Num: 24960, Cur Loss: 0.16199584, Cur Avg Loss: 0.17660115, Log Avg loss: 0.21393350, Global Avg Loss: 0.77583690, Time: 0.0209 Steps: 65430, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001570, Sample Num: 25120, Cur Loss: 0.07581805, Cur Avg Loss: 0.17630128, Log Avg loss: 0.12952129, Global Avg Loss: 0.77573813, Time: 0.0209 Steps: 65440, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001580, Sample Num: 25280, Cur Loss: 0.10120970, Cur Avg Loss: 0.17614622, Log Avg loss: 0.15180191, Global Avg Loss: 0.77564280, Time: 0.0209 Steps: 65450, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001590, Sample Num: 25440, Cur Loss: 0.24973190, Cur Avg Loss: 0.17576690, Log Avg loss: 0.11583482, Global Avg Loss: 0.77554201, Time: 0.0209 Steps: 65460, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001600, Sample Num: 25600, Cur Loss: 0.46686271, Cur Avg Loss: 0.17557407, Log Avg loss: 0.14491313, Global Avg Loss: 0.77544568, Time: 0.0209 Steps: 65470, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001610, Sample Num: 25760, Cur Loss: 0.15834096, Cur Avg Loss: 0.17575971, Log Avg loss: 0.20546186, Global Avg Loss: 0.77535864, Time: 0.0209 Steps: 65480, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001620, Sample Num: 25920, Cur Loss: 0.20822525, Cur Avg Loss: 0.17607311, Log Avg loss: 0.22653069, Global Avg Loss: 0.77527483, Time: 0.0209 Steps: 65490, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001630, Sample Num: 26080, Cur Loss: 0.05226744, Cur Avg Loss: 0.17592927, Log Avg loss: 0.15262721, Global Avg Loss: 0.77517977, Time: 0.0209 Steps: 65500, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001640, Sample Num: 26240, Cur Loss: 0.17353736, Cur Avg Loss: 0.17592138, Log Avg loss: 0.17463632, Global Avg Loss: 0.77508810, Time: 0.0208 Steps: 65510, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001650, Sample Num: 26400, Cur Loss: 0.21700706, Cur Avg Loss: 0.17593307, Log Avg loss: 0.17784998, Global Avg Loss: 0.77499695, Time: 0.0209 Steps: 65520, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001660, Sample Num: 26560, Cur Loss: 0.43676206, Cur Avg Loss: 0.17624604, Log Avg loss: 0.22788491, Global Avg Loss: 0.77491346, Time: 0.0209 Steps: 65530, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001670, Sample Num: 26720, Cur Loss: 0.41495264, Cur Avg Loss: 0.17636228, Log Avg loss: 0.19565947, Global Avg Loss: 0.77482508, Time: 0.0210 Steps: 65540, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001680, Sample Num: 26880, Cur Loss: 0.13240506, Cur Avg Loss: 0.17619266, Log Avg loss: 0.14786585, Global Avg Loss: 0.77472943, Time: 0.0209 Steps: 65550, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001690, Sample Num: 27040, Cur Loss: 0.05939115, Cur Avg Loss: 0.17621270, Log Avg loss: 0.17957873, Global Avg Loss: 0.77463865, Time: 0.0208 Steps: 65560, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001700, Sample Num: 27200, Cur Loss: 0.16086788, Cur Avg Loss: 0.17650966, Log Avg loss: 0.22669648, Global Avg Loss: 0.77455508, Time: 0.0208 Steps: 65570, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001710, Sample Num: 27360, Cur Loss: 0.21727341, Cur Avg Loss: 0.17646037, Log Avg loss: 0.16808131, Global Avg Loss: 0.77446261, Time: 0.0210 Steps: 65580, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001720, Sample Num: 27520, Cur Loss: 0.06142575, Cur Avg Loss: 0.17613085, Log Avg loss: 0.11978302, Global Avg Loss: 0.77436279, Time: 0.0209 Steps: 65590, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001730, Sample Num: 27680, Cur Loss: 0.12706915, Cur Avg Loss: 0.17628767, Log Avg loss: 0.20325921, Global Avg Loss: 0.77427573, Time: 0.0208 Steps: 65600, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001740, Sample Num: 27840, Cur Loss: 0.15132254, Cur Avg Loss: 0.17634318, Log Avg loss: 0.18594709, Global Avg Loss: 0.77418606, Time: 0.0209 Steps: 65610, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001750, Sample Num: 28000, Cur Loss: 0.05311878, Cur Avg Loss: 0.17620373, Log Avg loss: 0.15193919, Global Avg Loss: 0.77409124, Time: 0.0209 Steps: 65620, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001760, Sample Num: 28160, Cur Loss: 0.17472924, Cur Avg Loss: 0.17677899, Log Avg loss: 0.27745003, Global Avg Loss: 0.77401556, Time: 0.0209 Steps: 65630, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001770, Sample Num: 28320, Cur Loss: 0.10225989, Cur Avg Loss: 0.17699895, Log Avg loss: 0.21571095, Global Avg Loss: 0.77393051, Time: 0.0208 Steps: 65640, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001780, Sample Num: 28480, Cur Loss: 0.25071716, Cur Avg Loss: 0.17681718, Log Avg loss: 0.14464465, Global Avg Loss: 0.77383465, Time: 0.0209 Steps: 65650, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001790, Sample Num: 28640, Cur Loss: 0.11620793, Cur Avg Loss: 0.17662586, Log Avg loss: 0.14257028, Global Avg Loss: 0.77373851, Time: 0.0209 Steps: 65660, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001800, Sample Num: 28800, Cur Loss: 0.19404571, Cur Avg Loss: 0.17685288, Log Avg loss: 0.21748951, Global Avg Loss: 0.77365381, Time: 0.0209 Steps: 65670, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001810, Sample Num: 28960, Cur Loss: 0.18109253, Cur Avg Loss: 0.17684908, Log Avg loss: 0.17616600, Global Avg Loss: 0.77356284, Time: 0.0208 Steps: 65680, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001820, Sample Num: 29120, Cur Loss: 0.22632472, Cur Avg Loss: 0.17691750, Log Avg loss: 0.18930067, Global Avg Loss: 0.77347390, Time: 0.0208 Steps: 65690, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001830, Sample Num: 29280, Cur Loss: 0.26787460, Cur Avg Loss: 0.17704564, Log Avg loss: 0.20036704, Global Avg Loss: 0.77338667, Time: 0.0208 Steps: 65700, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001840, Sample Num: 29440, Cur Loss: 0.25098175, Cur Avg Loss: 0.17709955, Log Avg loss: 0.18696558, Global Avg Loss: 0.77329742, Time: 0.0209 Steps: 65710, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001850, Sample Num: 29600, Cur Loss: 0.22559696, Cur Avg Loss: 0.17716628, Log Avg loss: 0.18944532, Global Avg Loss: 0.77320858, Time: 0.0209 Steps: 65720, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001860, Sample Num: 29760, Cur Loss: 0.02941321, Cur Avg Loss: 0.17744984, Log Avg loss: 0.22990871, Global Avg Loss: 0.77312593, Time: 0.0208 Steps: 65730, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001870, Sample Num: 29920, Cur Loss: 0.06887411, Cur Avg Loss: 0.17765244, Log Avg loss: 0.21533437, Global Avg Loss: 0.77304108, Time: 0.0209 Steps: 65740, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001880, Sample Num: 30080, Cur Loss: 0.16417377, Cur Avg Loss: 0.17738849, Log Avg loss: 0.12803147, Global Avg Loss: 0.77294298, Time: 0.0208 Steps: 65750, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001890, Sample Num: 30240, Cur Loss: 0.05508600, Cur Avg Loss: 0.17746860, Log Avg loss: 0.19252787, Global Avg Loss: 0.77285472, Time: 0.0208 Steps: 65760, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001900, Sample Num: 30400, Cur Loss: 0.20462494, Cur Avg Loss: 0.17750010, Log Avg loss: 0.18345487, Global Avg Loss: 0.77276510, Time: 0.0208 Steps: 65770, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001910, Sample Num: 30560, Cur Loss: 0.09476844, Cur Avg Loss: 0.17777029, Log Avg loss: 0.22910575, Global Avg Loss: 0.77268245, Time: 0.0208 Steps: 65780, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001920, Sample Num: 30720, Cur Loss: 0.21503299, Cur Avg Loss: 0.17780921, Log Avg loss: 0.18524290, Global Avg Loss: 0.77259316, Time: 0.0208 Steps: 65790, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001930, Sample Num: 30880, Cur Loss: 0.35616055, Cur Avg Loss: 0.17777179, Log Avg loss: 0.17058686, Global Avg Loss: 0.77250167, Time: 0.0209 Steps: 65800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001940, Sample Num: 31040, Cur Loss: 0.08893467, Cur Avg Loss: 0.17791039, Log Avg loss: 0.20465960, Global Avg Loss: 0.77241539, Time: 0.0208 Steps: 65810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001950, Sample Num: 31200, Cur Loss: 0.23532687, Cur Avg Loss: 0.17800971, Log Avg loss: 0.19727921, Global Avg Loss: 0.77232801, Time: 0.0209 Steps: 65820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001960, Sample Num: 31360, Cur Loss: 0.43575948, Cur Avg Loss: 0.17842390, Log Avg loss: 0.25919117, Global Avg Loss: 0.77225006, Time: 0.0208 Steps: 65830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001970, Sample Num: 31520, Cur Loss: 0.16266313, Cur Avg Loss: 0.17833536, Log Avg loss: 0.16098100, Global Avg Loss: 0.77215722, Time: 0.0209 Steps: 65840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001980, Sample Num: 31680, Cur Loss: 0.08946791, Cur Avg Loss: 0.17835900, Log Avg loss: 0.18301543, Global Avg Loss: 0.77206775, Time: 0.0209 Steps: 65850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001990, Sample Num: 31840, Cur Loss: 0.21188331, Cur Avg Loss: 0.17881563, Log Avg loss: 0.26922920, Global Avg Loss: 0.77199140, Time: 0.0208 Steps: 65860, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002000, Sample Num: 32000, Cur Loss: 0.37887013, Cur Avg Loss: 0.17880827, Log Avg loss: 0.17734367, Global Avg Loss: 0.77190112, Time: 0.0208 Steps: 65870, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002010, Sample Num: 32160, Cur Loss: 0.24564317, Cur Avg Loss: 0.17885930, Log Avg loss: 0.18906567, Global Avg Loss: 0.77181265, Time: 0.0209 Steps: 65880, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002020, Sample Num: 32320, Cur Loss: 0.24301492, Cur Avg Loss: 0.17883419, Log Avg loss: 0.17378705, Global Avg Loss: 0.77172189, Time: 0.0208 Steps: 65890, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002030, Sample Num: 32480, Cur Loss: 0.48685768, Cur Avg Loss: 0.17908529, Log Avg loss: 0.22980686, Global Avg Loss: 0.77163966, Time: 0.0208 Steps: 65900, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002040, Sample Num: 32640, Cur Loss: 0.03339306, Cur Avg Loss: 0.17908037, Log Avg loss: 0.17808103, Global Avg Loss: 0.77154960, Time: 0.0208 Steps: 65910, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002050, Sample Num: 32800, Cur Loss: 0.05226674, Cur Avg Loss: 0.17873357, Log Avg loss: 0.10798721, Global Avg Loss: 0.77144894, Time: 0.0248 Steps: 65920, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002060, Sample Num: 32960, Cur Loss: 0.11950232, Cur Avg Loss: 0.17872957, Log Avg loss: 0.17790815, Global Avg Loss: 0.77135892, Time: 0.0208 Steps: 65930, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002070, Sample Num: 33120, Cur Loss: 0.27332163, Cur Avg Loss: 0.17887261, Log Avg loss: 0.20834064, Global Avg Loss: 0.77127353, Time: 0.0210 Steps: 65940, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002080, Sample Num: 33280, Cur Loss: 0.03863474, Cur Avg Loss: 0.17844765, Log Avg loss: 0.09047937, Global Avg Loss: 0.77117030, Time: 0.0209 Steps: 65950, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002090, Sample Num: 33440, Cur Loss: 0.27123624, Cur Avg Loss: 0.17885975, Log Avg loss: 0.26457644, Global Avg Loss: 0.77109350, Time: 0.0210 Steps: 65960, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002100, Sample Num: 33600, Cur Loss: 0.10370547, Cur Avg Loss: 0.17919276, Log Avg loss: 0.24879323, Global Avg Loss: 0.77101433, Time: 0.0210 Steps: 65970, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002110, Sample Num: 33760, Cur Loss: 0.15759151, Cur Avg Loss: 0.17909181, Log Avg loss: 0.15789106, Global Avg Loss: 0.77092140, Time: 0.0211 Steps: 65980, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002120, Sample Num: 33920, Cur Loss: 0.08409055, Cur Avg Loss: 0.17896298, Log Avg loss: 0.15178002, Global Avg Loss: 0.77082758, Time: 0.0209 Steps: 65990, Updated lr: 0.000038 ***** Running evaluation checkpoint-65999 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-65999 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.661025, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.238277, "eval_total_loss": 167.508474, "eval_mae": 0.321472, "eval_mse": 0.23836, "eval_r2": 0.848482, "eval_sp_statistic": 0.919481, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.929908, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.204421, "test_total_loss": 102.619178, "test_mae": 0.353325, "test_mse": 0.204489, "test_r2": 0.868021, "test_sp_statistic": 0.914596, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947768, "test_ps_pvalue": 0.0, "lr": 3.8360360360360366e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7707401507388926, "train_cur_epoch_loss": 380.56876594480127, "train_cur_epoch_avg_loss": 0.17875470453020256, "train_cur_epoch_time": 44.661025047302246, "train_cur_epoch_avg_time": 0.020977465968671793, "epoch": 31, "step": 65999} ################################################## Training, Epoch: 0032, Batch: 000001, Sample Num: 16, Cur Loss: 0.55265355, Cur Avg Loss: 0.55265355, Log Avg loss: 0.17199076, Global Avg Loss: 0.77073685, Time: 0.0247 Steps: 66000, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000011, Sample Num: 176, Cur Loss: 0.21559089, Cur Avg Loss: 0.16725748, Log Avg loss: 0.12871787, Global Avg Loss: 0.77063959, Time: 0.0210 Steps: 66010, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000021, Sample Num: 336, Cur Loss: 0.16016388, Cur Avg Loss: 0.15427160, Log Avg loss: 0.13998714, Global Avg Loss: 0.77054406, Time: 0.0210 Steps: 66020, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000031, Sample Num: 496, Cur Loss: 0.05033614, Cur Avg Loss: 0.14946439, Log Avg loss: 0.13936924, Global Avg Loss: 0.77044847, Time: 0.0211 Steps: 66030, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000041, Sample Num: 656, Cur Loss: 0.03298377, Cur Avg Loss: 0.14307234, Log Avg loss: 0.12325699, Global Avg Loss: 0.77035047, Time: 0.0210 Steps: 66040, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000051, Sample Num: 816, Cur Loss: 0.05233847, Cur Avg Loss: 0.13115942, Log Avg loss: 0.08231647, Global Avg Loss: 0.77024630, Time: 0.0211 Steps: 66050, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000061, Sample Num: 976, Cur Loss: 0.10300335, Cur Avg Loss: 0.13570460, Log Avg loss: 0.15888499, Global Avg Loss: 0.77015376, Time: 0.0211 Steps: 66060, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000071, Sample Num: 1136, Cur Loss: 0.21811211, Cur Avg Loss: 0.14553428, Log Avg loss: 0.20549533, Global Avg Loss: 0.77006829, Time: 0.0211 Steps: 66070, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000081, Sample Num: 1296, Cur Loss: 0.10552065, Cur Avg Loss: 0.15067121, Log Avg loss: 0.18714343, Global Avg Loss: 0.76998008, Time: 0.0211 Steps: 66080, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000091, Sample Num: 1456, Cur Loss: 0.11937303, Cur Avg Loss: 0.15526321, Log Avg loss: 0.19245838, Global Avg Loss: 0.76989269, Time: 0.0210 Steps: 66090, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000101, Sample Num: 1616, Cur Loss: 0.15795314, Cur Avg Loss: 0.15606700, Log Avg loss: 0.16338147, Global Avg Loss: 0.76980094, Time: 0.0211 Steps: 66100, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000111, Sample Num: 1776, Cur Loss: 0.21413074, Cur Avg Loss: 0.16414003, Log Avg loss: 0.24567771, Global Avg Loss: 0.76972166, Time: 0.0210 Steps: 66110, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000121, Sample Num: 1936, Cur Loss: 0.06976653, Cur Avg Loss: 0.16488724, Log Avg loss: 0.17318127, Global Avg Loss: 0.76963144, Time: 0.0211 Steps: 66120, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000131, Sample Num: 2096, Cur Loss: 0.06930521, Cur Avg Loss: 0.16374533, Log Avg loss: 0.14992818, Global Avg Loss: 0.76953773, Time: 0.0210 Steps: 66130, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000141, Sample Num: 2256, Cur Loss: 0.24148402, Cur Avg Loss: 0.16059144, Log Avg loss: 0.11927550, Global Avg Loss: 0.76943941, Time: 0.0211 Steps: 66140, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000151, Sample Num: 2416, Cur Loss: 0.11959226, Cur Avg Loss: 0.15924707, Log Avg loss: 0.14029138, Global Avg Loss: 0.76934430, Time: 0.0210 Steps: 66150, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000161, Sample Num: 2576, Cur Loss: 0.10668956, Cur Avg Loss: 0.15627054, Log Avg loss: 0.11132498, Global Avg Loss: 0.76924484, Time: 0.0210 Steps: 66160, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000171, Sample Num: 2736, Cur Loss: 0.08655442, Cur Avg Loss: 0.15784109, Log Avg loss: 0.18312700, Global Avg Loss: 0.76915626, Time: 0.0210 Steps: 66170, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000181, Sample Num: 2896, Cur Loss: 0.38103265, Cur Avg Loss: 0.15848884, Log Avg loss: 0.16956533, Global Avg Loss: 0.76906566, Time: 0.0210 Steps: 66180, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000191, Sample Num: 3056, Cur Loss: 0.16668318, Cur Avg Loss: 0.16748769, Log Avg loss: 0.33036693, Global Avg Loss: 0.76899939, Time: 0.0210 Steps: 66190, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000201, Sample Num: 3216, Cur Loss: 0.17248662, Cur Avg Loss: 0.16635137, Log Avg loss: 0.14464760, Global Avg Loss: 0.76890507, Time: 0.0210 Steps: 66200, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000211, Sample Num: 3376, Cur Loss: 0.13613880, Cur Avg Loss: 0.16646370, Log Avg loss: 0.16872162, Global Avg Loss: 0.76881442, Time: 0.0210 Steps: 66210, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000221, Sample Num: 3536, Cur Loss: 0.20045111, Cur Avg Loss: 0.16647930, Log Avg loss: 0.16680845, Global Avg Loss: 0.76872351, Time: 0.0210 Steps: 66220, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000231, Sample Num: 3696, Cur Loss: 0.55447018, Cur Avg Loss: 0.16837641, Log Avg loss: 0.21030238, Global Avg Loss: 0.76863920, Time: 0.0210 Steps: 66230, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000241, Sample Num: 3856, Cur Loss: 0.07133113, Cur Avg Loss: 0.16575463, Log Avg loss: 0.10519161, Global Avg Loss: 0.76853904, Time: 0.0211 Steps: 66240, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000251, Sample Num: 4016, Cur Loss: 0.19143757, Cur Avg Loss: 0.16589623, Log Avg loss: 0.16930889, Global Avg Loss: 0.76844859, Time: 0.0210 Steps: 66250, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000261, Sample Num: 4176, Cur Loss: 0.19675347, Cur Avg Loss: 0.16560311, Log Avg loss: 0.15824567, Global Avg Loss: 0.76835650, Time: 0.0247 Steps: 66260, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000271, Sample Num: 4336, Cur Loss: 0.09426210, Cur Avg Loss: 0.16651178, Log Avg loss: 0.19022799, Global Avg Loss: 0.76826926, Time: 0.0210 Steps: 66270, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000281, Sample Num: 4496, Cur Loss: 0.32189634, Cur Avg Loss: 0.16694199, Log Avg loss: 0.17860075, Global Avg Loss: 0.76818029, Time: 0.0210 Steps: 66280, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000291, Sample Num: 4656, Cur Loss: 0.09725708, Cur Avg Loss: 0.16419346, Log Avg loss: 0.08695968, Global Avg Loss: 0.76807753, Time: 0.0210 Steps: 66290, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000301, Sample Num: 4816, Cur Loss: 0.09062959, Cur Avg Loss: 0.16502830, Log Avg loss: 0.18932237, Global Avg Loss: 0.76799024, Time: 0.0210 Steps: 66300, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000311, Sample Num: 4976, Cur Loss: 0.20856774, Cur Avg Loss: 0.16478157, Log Avg loss: 0.15735493, Global Avg Loss: 0.76789815, Time: 0.0210 Steps: 66310, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000321, Sample Num: 5136, Cur Loss: 0.15611650, Cur Avg Loss: 0.16528995, Log Avg loss: 0.18110057, Global Avg Loss: 0.76780967, Time: 0.0210 Steps: 66320, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000331, Sample Num: 5296, Cur Loss: 0.10781202, Cur Avg Loss: 0.16469938, Log Avg loss: 0.14574204, Global Avg Loss: 0.76771589, Time: 0.0210 Steps: 66330, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000341, Sample Num: 5456, Cur Loss: 0.07440872, Cur Avg Loss: 0.16743879, Log Avg loss: 0.25811325, Global Avg Loss: 0.76763907, Time: 0.0210 Steps: 66340, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000351, Sample Num: 5616, Cur Loss: 0.15289533, Cur Avg Loss: 0.16877072, Log Avg loss: 0.21418966, Global Avg Loss: 0.76755566, Time: 0.0210 Steps: 66350, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000361, Sample Num: 5776, Cur Loss: 0.14789683, Cur Avg Loss: 0.16822264, Log Avg loss: 0.14898502, Global Avg Loss: 0.76746244, Time: 0.0210 Steps: 66360, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000371, Sample Num: 5936, Cur Loss: 0.10539593, Cur Avg Loss: 0.16746251, Log Avg loss: 0.14002154, Global Avg Loss: 0.76736790, Time: 0.0210 Steps: 66370, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000381, Sample Num: 6096, Cur Loss: 0.20183754, Cur Avg Loss: 0.16634511, Log Avg loss: 0.12488965, Global Avg Loss: 0.76727112, Time: 0.0210 Steps: 66380, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000391, Sample Num: 6256, Cur Loss: 0.33264774, Cur Avg Loss: 0.16764631, Log Avg loss: 0.21722214, Global Avg Loss: 0.76718827, Time: 0.0210 Steps: 66390, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000401, Sample Num: 6416, Cur Loss: 0.19885811, Cur Avg Loss: 0.16814756, Log Avg loss: 0.18774658, Global Avg Loss: 0.76710100, Time: 0.0210 Steps: 66400, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000411, Sample Num: 6576, Cur Loss: 0.08742604, Cur Avg Loss: 0.16710874, Log Avg loss: 0.12545202, Global Avg Loss: 0.76700438, Time: 0.0210 Steps: 66410, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000421, Sample Num: 6736, Cur Loss: 0.07653525, Cur Avg Loss: 0.16755605, Log Avg loss: 0.18594014, Global Avg Loss: 0.76691690, Time: 0.0210 Steps: 66420, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000431, Sample Num: 6896, Cur Loss: 0.04243610, Cur Avg Loss: 0.16880924, Log Avg loss: 0.22156895, Global Avg Loss: 0.76683480, Time: 0.0210 Steps: 66430, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000441, Sample Num: 7056, Cur Loss: 0.07593115, Cur Avg Loss: 0.16897758, Log Avg loss: 0.17623266, Global Avg Loss: 0.76674591, Time: 0.0209 Steps: 66440, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000451, Sample Num: 7216, Cur Loss: 0.20779699, Cur Avg Loss: 0.16830160, Log Avg loss: 0.13849123, Global Avg Loss: 0.76665137, Time: 0.0210 Steps: 66450, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000461, Sample Num: 7376, Cur Loss: 0.32314938, Cur Avg Loss: 0.16887386, Log Avg loss: 0.19468278, Global Avg Loss: 0.76656530, Time: 0.0210 Steps: 66460, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000471, Sample Num: 7536, Cur Loss: 0.15262938, Cur Avg Loss: 0.16973751, Log Avg loss: 0.20955167, Global Avg Loss: 0.76648150, Time: 0.0210 Steps: 66470, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000481, Sample Num: 7696, Cur Loss: 0.19272238, Cur Avg Loss: 0.16844914, Log Avg loss: 0.10776668, Global Avg Loss: 0.76638242, Time: 0.0210 Steps: 66480, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000491, Sample Num: 7856, Cur Loss: 0.10695867, Cur Avg Loss: 0.16829581, Log Avg loss: 0.16092056, Global Avg Loss: 0.76629136, Time: 0.0210 Steps: 66490, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000501, Sample Num: 8016, Cur Loss: 0.08239028, Cur Avg Loss: 0.16759772, Log Avg loss: 0.13332160, Global Avg Loss: 0.76619618, Time: 0.0210 Steps: 66500, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000511, Sample Num: 8176, Cur Loss: 0.20871359, Cur Avg Loss: 0.16746481, Log Avg loss: 0.16080622, Global Avg Loss: 0.76610515, Time: 0.0210 Steps: 66510, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000521, Sample Num: 8336, Cur Loss: 0.11834746, Cur Avg Loss: 0.16733717, Log Avg loss: 0.16081473, Global Avg Loss: 0.76601416, Time: 0.0211 Steps: 66520, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000531, Sample Num: 8496, Cur Loss: 0.10231200, Cur Avg Loss: 0.16736852, Log Avg loss: 0.16900171, Global Avg Loss: 0.76592442, Time: 0.0210 Steps: 66530, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000541, Sample Num: 8656, Cur Loss: 0.12510005, Cur Avg Loss: 0.16744038, Log Avg loss: 0.17125628, Global Avg Loss: 0.76583505, Time: 0.0210 Steps: 66540, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000551, Sample Num: 8816, Cur Loss: 0.22758846, Cur Avg Loss: 0.16950276, Log Avg loss: 0.28107745, Global Avg Loss: 0.76576221, Time: 0.0210 Steps: 66550, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000561, Sample Num: 8976, Cur Loss: 0.26411512, Cur Avg Loss: 0.16959848, Log Avg loss: 0.17487291, Global Avg Loss: 0.76567344, Time: 0.0210 Steps: 66560, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000571, Sample Num: 9136, Cur Loss: 0.04815753, Cur Avg Loss: 0.16897186, Log Avg loss: 0.13381834, Global Avg Loss: 0.76557852, Time: 0.0210 Steps: 66570, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000581, Sample Num: 9296, Cur Loss: 0.05369551, Cur Avg Loss: 0.16828204, Log Avg loss: 0.12889298, Global Avg Loss: 0.76548289, Time: 0.0210 Steps: 66580, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000591, Sample Num: 9456, Cur Loss: 0.04553615, Cur Avg Loss: 0.16767517, Log Avg loss: 0.13241626, Global Avg Loss: 0.76538782, Time: 0.0210 Steps: 66590, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000601, Sample Num: 9616, Cur Loss: 0.04246671, Cur Avg Loss: 0.16772624, Log Avg loss: 0.17074438, Global Avg Loss: 0.76529854, Time: 0.0210 Steps: 66600, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000611, Sample Num: 9776, Cur Loss: 0.24751090, Cur Avg Loss: 0.16772706, Log Avg loss: 0.16777675, Global Avg Loss: 0.76520883, Time: 0.0210 Steps: 66610, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000621, Sample Num: 9936, Cur Loss: 0.02432969, Cur Avg Loss: 0.16760608, Log Avg loss: 0.16021416, Global Avg Loss: 0.76511802, Time: 0.0210 Steps: 66620, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000631, Sample Num: 10096, Cur Loss: 0.03883388, Cur Avg Loss: 0.16685995, Log Avg loss: 0.12052521, Global Avg Loss: 0.76502128, Time: 0.0210 Steps: 66630, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000641, Sample Num: 10256, Cur Loss: 0.03605420, Cur Avg Loss: 0.16711001, Log Avg loss: 0.18288879, Global Avg Loss: 0.76493392, Time: 0.0210 Steps: 66640, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000651, Sample Num: 10416, Cur Loss: 0.09342784, Cur Avg Loss: 0.16681586, Log Avg loss: 0.14796080, Global Avg Loss: 0.76484136, Time: 0.0210 Steps: 66650, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000661, Sample Num: 10576, Cur Loss: 0.11961311, Cur Avg Loss: 0.16624149, Log Avg loss: 0.12884977, Global Avg Loss: 0.76474595, Time: 0.0210 Steps: 66660, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000671, Sample Num: 10736, Cur Loss: 0.07471988, Cur Avg Loss: 0.16653962, Log Avg loss: 0.18624626, Global Avg Loss: 0.76465918, Time: 0.0210 Steps: 66670, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000681, Sample Num: 10896, Cur Loss: 0.18488476, Cur Avg Loss: 0.16606681, Log Avg loss: 0.13434146, Global Avg Loss: 0.76456465, Time: 0.0210 Steps: 66680, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000691, Sample Num: 11056, Cur Loss: 0.12260833, Cur Avg Loss: 0.16676331, Log Avg loss: 0.21419465, Global Avg Loss: 0.76448212, Time: 0.0210 Steps: 66690, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000701, Sample Num: 11216, Cur Loss: 0.60038441, Cur Avg Loss: 0.16809269, Log Avg loss: 0.25995263, Global Avg Loss: 0.76440648, Time: 0.0210 Steps: 66700, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000711, Sample Num: 11376, Cur Loss: 0.07110910, Cur Avg Loss: 0.16754115, Log Avg loss: 0.12887810, Global Avg Loss: 0.76431121, Time: 0.0210 Steps: 66710, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000721, Sample Num: 11536, Cur Loss: 0.07112700, Cur Avg Loss: 0.16702123, Log Avg loss: 0.13005523, Global Avg Loss: 0.76421615, Time: 0.0210 Steps: 66720, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000731, Sample Num: 11696, Cur Loss: 0.05887940, Cur Avg Loss: 0.16641182, Log Avg loss: 0.12247328, Global Avg Loss: 0.76411998, Time: 0.0210 Steps: 66730, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000741, Sample Num: 11856, Cur Loss: 0.16561837, Cur Avg Loss: 0.16627101, Log Avg loss: 0.15597780, Global Avg Loss: 0.76402886, Time: 0.0210 Steps: 66740, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000751, Sample Num: 12016, Cur Loss: 0.51784950, Cur Avg Loss: 0.16657378, Log Avg loss: 0.18900881, Global Avg Loss: 0.76394271, Time: 0.0211 Steps: 66750, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000761, Sample Num: 12176, Cur Loss: 0.14591826, Cur Avg Loss: 0.16637427, Log Avg loss: 0.15139115, Global Avg Loss: 0.76385096, Time: 0.0211 Steps: 66760, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000771, Sample Num: 12336, Cur Loss: 0.14451551, Cur Avg Loss: 0.16667477, Log Avg loss: 0.18954279, Global Avg Loss: 0.76376495, Time: 0.0247 Steps: 66770, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000781, Sample Num: 12496, Cur Loss: 0.04801548, Cur Avg Loss: 0.16674001, Log Avg loss: 0.17177000, Global Avg Loss: 0.76367630, Time: 0.0210 Steps: 66780, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000791, Sample Num: 12656, Cur Loss: 0.28340855, Cur Avg Loss: 0.16650138, Log Avg loss: 0.14786496, Global Avg Loss: 0.76358410, Time: 0.0210 Steps: 66790, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000801, Sample Num: 12816, Cur Loss: 0.13458988, Cur Avg Loss: 0.16655575, Log Avg loss: 0.17085637, Global Avg Loss: 0.76349536, Time: 0.0210 Steps: 66800, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000811, Sample Num: 12976, Cur Loss: 0.04630736, Cur Avg Loss: 0.16653097, Log Avg loss: 0.16454574, Global Avg Loss: 0.76340572, Time: 0.0210 Steps: 66810, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000821, Sample Num: 13136, Cur Loss: 0.08588594, Cur Avg Loss: 0.16640127, Log Avg loss: 0.15588308, Global Avg Loss: 0.76331480, Time: 0.0211 Steps: 66820, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000831, Sample Num: 13296, Cur Loss: 0.32563233, Cur Avg Loss: 0.16651485, Log Avg loss: 0.17583968, Global Avg Loss: 0.76322689, Time: 0.0210 Steps: 66830, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000841, Sample Num: 13456, Cur Loss: 0.18695700, Cur Avg Loss: 0.16697372, Log Avg loss: 0.20510523, Global Avg Loss: 0.76314339, Time: 0.0210 Steps: 66840, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000851, Sample Num: 13616, Cur Loss: 0.14070535, Cur Avg Loss: 0.16695344, Log Avg loss: 0.16524794, Global Avg Loss: 0.76305395, Time: 0.0210 Steps: 66850, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000861, Sample Num: 13776, Cur Loss: 0.08185204, Cur Avg Loss: 0.16685514, Log Avg loss: 0.15848966, Global Avg Loss: 0.76296353, Time: 0.0210 Steps: 66860, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000871, Sample Num: 13936, Cur Loss: 0.08036399, Cur Avg Loss: 0.16638369, Log Avg loss: 0.12579210, Global Avg Loss: 0.76286824, Time: 0.0210 Steps: 66870, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000881, Sample Num: 14096, Cur Loss: 0.09707236, Cur Avg Loss: 0.16603252, Log Avg loss: 0.13544603, Global Avg Loss: 0.76277443, Time: 0.0210 Steps: 66880, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000891, Sample Num: 14256, Cur Loss: 0.27003068, Cur Avg Loss: 0.16624786, Log Avg loss: 0.18521861, Global Avg Loss: 0.76268809, Time: 0.0210 Steps: 66890, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000901, Sample Num: 14416, Cur Loss: 0.24667141, Cur Avg Loss: 0.16652069, Log Avg loss: 0.19082998, Global Avg Loss: 0.76260261, Time: 0.0210 Steps: 66900, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000911, Sample Num: 14576, Cur Loss: 0.14574686, Cur Avg Loss: 0.16674420, Log Avg loss: 0.18688311, Global Avg Loss: 0.76251656, Time: 0.0210 Steps: 66910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000921, Sample Num: 14736, Cur Loss: 0.12461945, Cur Avg Loss: 0.16662403, Log Avg loss: 0.15567595, Global Avg Loss: 0.76242588, Time: 0.0210 Steps: 66920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000931, Sample Num: 14896, Cur Loss: 0.14334574, Cur Avg Loss: 0.16632191, Log Avg loss: 0.13849666, Global Avg Loss: 0.76233266, Time: 0.0210 Steps: 66930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000941, Sample Num: 15056, Cur Loss: 0.24262023, Cur Avg Loss: 0.16640810, Log Avg loss: 0.17443256, Global Avg Loss: 0.76224483, Time: 0.0210 Steps: 66940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000951, Sample Num: 15216, Cur Loss: 0.05720842, Cur Avg Loss: 0.16605066, Log Avg loss: 0.13241567, Global Avg Loss: 0.76215076, Time: 0.0210 Steps: 66950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000961, Sample Num: 15376, Cur Loss: 0.07145934, Cur Avg Loss: 0.16638369, Log Avg loss: 0.19805516, Global Avg Loss: 0.76206652, Time: 0.0210 Steps: 66960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000971, Sample Num: 15536, Cur Loss: 0.42570934, Cur Avg Loss: 0.16707388, Log Avg loss: 0.23340036, Global Avg Loss: 0.76198758, Time: 0.0210 Steps: 66970, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000981, Sample Num: 15696, Cur Loss: 0.38953823, Cur Avg Loss: 0.16703245, Log Avg loss: 0.16300964, Global Avg Loss: 0.76189815, Time: 0.0210 Steps: 66980, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000991, Sample Num: 15856, Cur Loss: 0.13881919, Cur Avg Loss: 0.16649175, Log Avg loss: 0.11344925, Global Avg Loss: 0.76180135, Time: 0.0210 Steps: 66990, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001001, Sample Num: 16016, Cur Loss: 0.08670218, Cur Avg Loss: 0.16799418, Log Avg loss: 0.31688469, Global Avg Loss: 0.76173495, Time: 0.0210 Steps: 67000, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001011, Sample Num: 16176, Cur Loss: 0.05168279, Cur Avg Loss: 0.16806059, Log Avg loss: 0.17470863, Global Avg Loss: 0.76164734, Time: 0.0210 Steps: 67010, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001021, Sample Num: 16336, Cur Loss: 0.34643349, Cur Avg Loss: 0.16858707, Log Avg loss: 0.22181477, Global Avg Loss: 0.76156680, Time: 0.0210 Steps: 67020, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001031, Sample Num: 16496, Cur Loss: 0.19725275, Cur Avg Loss: 0.16836517, Log Avg loss: 0.14570871, Global Avg Loss: 0.76147492, Time: 0.0210 Steps: 67030, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001041, Sample Num: 16656, Cur Loss: 0.15873127, Cur Avg Loss: 0.16861378, Log Avg loss: 0.19424521, Global Avg Loss: 0.76139031, Time: 0.0210 Steps: 67040, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001051, Sample Num: 16816, Cur Loss: 0.05706012, Cur Avg Loss: 0.16828341, Log Avg loss: 0.13389186, Global Avg Loss: 0.76129672, Time: 0.0210 Steps: 67050, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001061, Sample Num: 16976, Cur Loss: 0.31942499, Cur Avg Loss: 0.16863274, Log Avg loss: 0.20534749, Global Avg Loss: 0.76121382, Time: 0.0210 Steps: 67060, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001071, Sample Num: 17136, Cur Loss: 0.13940021, Cur Avg Loss: 0.16903671, Log Avg loss: 0.21189789, Global Avg Loss: 0.76113191, Time: 0.0210 Steps: 67070, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001081, Sample Num: 17296, Cur Loss: 0.07776564, Cur Avg Loss: 0.16902439, Log Avg loss: 0.16770557, Global Avg Loss: 0.76104345, Time: 0.0210 Steps: 67080, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001091, Sample Num: 17456, Cur Loss: 0.06436154, Cur Avg Loss: 0.16851791, Log Avg loss: 0.11376685, Global Avg Loss: 0.76094697, Time: 0.0210 Steps: 67090, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001101, Sample Num: 17616, Cur Loss: 0.23521718, Cur Avg Loss: 0.16862222, Log Avg loss: 0.18000281, Global Avg Loss: 0.76086039, Time: 0.0210 Steps: 67100, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001111, Sample Num: 17776, Cur Loss: 0.11922543, Cur Avg Loss: 0.16915567, Log Avg loss: 0.22788868, Global Avg Loss: 0.76078097, Time: 0.0210 Steps: 67110, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001121, Sample Num: 17936, Cur Loss: 0.14298098, Cur Avg Loss: 0.16918982, Log Avg loss: 0.17298358, Global Avg Loss: 0.76069340, Time: 0.0210 Steps: 67120, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001131, Sample Num: 18096, Cur Loss: 0.10637224, Cur Avg Loss: 0.16908673, Log Avg loss: 0.15753005, Global Avg Loss: 0.76060355, Time: 0.0210 Steps: 67130, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001141, Sample Num: 18256, Cur Loss: 0.38328481, Cur Avg Loss: 0.16968093, Log Avg loss: 0.23688517, Global Avg Loss: 0.76052555, Time: 0.0210 Steps: 67140, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001151, Sample Num: 18416, Cur Loss: 0.25324750, Cur Avg Loss: 0.16910058, Log Avg loss: 0.10288284, Global Avg Loss: 0.76042761, Time: 0.0210 Steps: 67150, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001161, Sample Num: 18576, Cur Loss: 0.42422333, Cur Avg Loss: 0.16975446, Log Avg loss: 0.24501541, Global Avg Loss: 0.76035087, Time: 0.0210 Steps: 67160, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001171, Sample Num: 18736, Cur Loss: 0.14720240, Cur Avg Loss: 0.17038341, Log Avg loss: 0.24340512, Global Avg Loss: 0.76027390, Time: 0.0210 Steps: 67170, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001181, Sample Num: 18896, Cur Loss: 0.09518564, Cur Avg Loss: 0.17008306, Log Avg loss: 0.13491235, Global Avg Loss: 0.76018082, Time: 0.0210 Steps: 67180, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001191, Sample Num: 19056, Cur Loss: 0.24556462, Cur Avg Loss: 0.16990001, Log Avg loss: 0.14828113, Global Avg Loss: 0.76008975, Time: 0.0210 Steps: 67190, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001201, Sample Num: 19216, Cur Loss: 0.02413342, Cur Avg Loss: 0.17003772, Log Avg loss: 0.18643891, Global Avg Loss: 0.76000438, Time: 0.0210 Steps: 67200, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001211, Sample Num: 19376, Cur Loss: 0.03143472, Cur Avg Loss: 0.16960665, Log Avg loss: 0.11783503, Global Avg Loss: 0.75990884, Time: 0.0209 Steps: 67210, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001221, Sample Num: 19536, Cur Loss: 0.24987413, Cur Avg Loss: 0.16931494, Log Avg loss: 0.13398967, Global Avg Loss: 0.75981572, Time: 0.0210 Steps: 67220, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001231, Sample Num: 19696, Cur Loss: 0.10856999, Cur Avg Loss: 0.16917929, Log Avg loss: 0.15261550, Global Avg Loss: 0.75972540, Time: 0.0210 Steps: 67230, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001241, Sample Num: 19856, Cur Loss: 0.19331002, Cur Avg Loss: 0.16928074, Log Avg loss: 0.18176944, Global Avg Loss: 0.75963945, Time: 0.0210 Steps: 67240, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001251, Sample Num: 20016, Cur Loss: 0.12093066, Cur Avg Loss: 0.16957493, Log Avg loss: 0.20608386, Global Avg Loss: 0.75955714, Time: 0.0210 Steps: 67250, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001261, Sample Num: 20176, Cur Loss: 0.24017024, Cur Avg Loss: 0.16998376, Log Avg loss: 0.22112923, Global Avg Loss: 0.75947708, Time: 0.0210 Steps: 67260, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001271, Sample Num: 20336, Cur Loss: 0.03330578, Cur Avg Loss: 0.16954173, Log Avg loss: 0.11380065, Global Avg Loss: 0.75938110, Time: 0.0210 Steps: 67270, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001281, Sample Num: 20496, Cur Loss: 0.06386831, Cur Avg Loss: 0.16916895, Log Avg loss: 0.12178909, Global Avg Loss: 0.75928634, Time: 0.0245 Steps: 67280, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001291, Sample Num: 20656, Cur Loss: 0.16139767, Cur Avg Loss: 0.16900676, Log Avg loss: 0.14823058, Global Avg Loss: 0.75919553, Time: 0.0209 Steps: 67290, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001301, Sample Num: 20816, Cur Loss: 0.12932494, Cur Avg Loss: 0.16886561, Log Avg loss: 0.15064310, Global Avg Loss: 0.75910510, Time: 0.0209 Steps: 67300, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001311, Sample Num: 20976, Cur Loss: 0.21295890, Cur Avg Loss: 0.16856589, Log Avg loss: 0.12957196, Global Avg Loss: 0.75901157, Time: 0.0209 Steps: 67310, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001321, Sample Num: 21136, Cur Loss: 0.09418081, Cur Avg Loss: 0.16887282, Log Avg loss: 0.20911171, Global Avg Loss: 0.75892989, Time: 0.0209 Steps: 67320, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001331, Sample Num: 21296, Cur Loss: 0.06726488, Cur Avg Loss: 0.16905947, Log Avg loss: 0.19371603, Global Avg Loss: 0.75884594, Time: 0.0209 Steps: 67330, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001341, Sample Num: 21456, Cur Loss: 0.22267859, Cur Avg Loss: 0.16891751, Log Avg loss: 0.15002277, Global Avg Loss: 0.75875553, Time: 0.0209 Steps: 67340, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001351, Sample Num: 21616, Cur Loss: 0.22520189, Cur Avg Loss: 0.16903917, Log Avg loss: 0.18535288, Global Avg Loss: 0.75867040, Time: 0.0209 Steps: 67350, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001361, Sample Num: 21776, Cur Loss: 0.39388400, Cur Avg Loss: 0.16912051, Log Avg loss: 0.18010977, Global Avg Loss: 0.75858450, Time: 0.0209 Steps: 67360, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001371, Sample Num: 21936, Cur Loss: 0.07740727, Cur Avg Loss: 0.16896816, Log Avg loss: 0.14823303, Global Avg Loss: 0.75849391, Time: 0.0208 Steps: 67370, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001381, Sample Num: 22096, Cur Loss: 0.17354895, Cur Avg Loss: 0.16897889, Log Avg loss: 0.17045033, Global Avg Loss: 0.75840663, Time: 0.0209 Steps: 67380, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001391, Sample Num: 22256, Cur Loss: 0.24611709, Cur Avg Loss: 0.16933186, Log Avg loss: 0.21807779, Global Avg Loss: 0.75832646, Time: 0.0208 Steps: 67390, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001401, Sample Num: 22416, Cur Loss: 0.24268526, Cur Avg Loss: 0.16965205, Log Avg loss: 0.21418922, Global Avg Loss: 0.75824572, Time: 0.0209 Steps: 67400, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001411, Sample Num: 22576, Cur Loss: 0.59408391, Cur Avg Loss: 0.17011600, Log Avg loss: 0.23511630, Global Avg Loss: 0.75816812, Time: 0.0209 Steps: 67410, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001421, Sample Num: 22736, Cur Loss: 0.16921720, Cur Avg Loss: 0.17029537, Log Avg loss: 0.19560389, Global Avg Loss: 0.75808468, Time: 0.0209 Steps: 67420, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001431, Sample Num: 22896, Cur Loss: 0.41050279, Cur Avg Loss: 0.17053782, Log Avg loss: 0.20499037, Global Avg Loss: 0.75800265, Time: 0.0209 Steps: 67430, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001441, Sample Num: 23056, Cur Loss: 0.06650759, Cur Avg Loss: 0.17037678, Log Avg loss: 0.14733240, Global Avg Loss: 0.75791210, Time: 0.0208 Steps: 67440, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001451, Sample Num: 23216, Cur Loss: 0.04791221, Cur Avg Loss: 0.17010517, Log Avg loss: 0.13096509, Global Avg Loss: 0.75781915, Time: 0.0209 Steps: 67450, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001461, Sample Num: 23376, Cur Loss: 0.03711373, Cur Avg Loss: 0.16982132, Log Avg loss: 0.12863565, Global Avg Loss: 0.75772588, Time: 0.0209 Steps: 67460, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001471, Sample Num: 23536, Cur Loss: 0.15952426, Cur Avg Loss: 0.16986980, Log Avg loss: 0.17695249, Global Avg Loss: 0.75763981, Time: 0.0208 Steps: 67470, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001481, Sample Num: 23696, Cur Loss: 0.23931396, Cur Avg Loss: 0.16976803, Log Avg loss: 0.15479775, Global Avg Loss: 0.75755047, Time: 0.0209 Steps: 67480, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001491, Sample Num: 23856, Cur Loss: 0.19652161, Cur Avg Loss: 0.16969963, Log Avg loss: 0.15956928, Global Avg Loss: 0.75746187, Time: 0.0208 Steps: 67490, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001501, Sample Num: 24016, Cur Loss: 0.29451019, Cur Avg Loss: 0.16934990, Log Avg loss: 0.11720471, Global Avg Loss: 0.75736701, Time: 0.0209 Steps: 67500, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001511, Sample Num: 24176, Cur Loss: 0.02463539, Cur Avg Loss: 0.16947334, Log Avg loss: 0.18800275, Global Avg Loss: 0.75728268, Time: 0.0209 Steps: 67510, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001521, Sample Num: 24336, Cur Loss: 0.23090190, Cur Avg Loss: 0.16962498, Log Avg loss: 0.19253773, Global Avg Loss: 0.75719903, Time: 0.0208 Steps: 67520, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001531, Sample Num: 24496, Cur Loss: 0.24636690, Cur Avg Loss: 0.16986738, Log Avg loss: 0.20673565, Global Avg Loss: 0.75711752, Time: 0.0209 Steps: 67530, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001541, Sample Num: 24656, Cur Loss: 0.04707258, Cur Avg Loss: 0.16998929, Log Avg loss: 0.18865443, Global Avg Loss: 0.75703335, Time: 0.0210 Steps: 67540, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001551, Sample Num: 24816, Cur Loss: 0.19424057, Cur Avg Loss: 0.16967777, Log Avg loss: 0.12167177, Global Avg Loss: 0.75693930, Time: 0.0209 Steps: 67550, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001561, Sample Num: 24976, Cur Loss: 0.18405963, Cur Avg Loss: 0.16989810, Log Avg loss: 0.20407107, Global Avg Loss: 0.75685746, Time: 0.0209 Steps: 67560, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001571, Sample Num: 25136, Cur Loss: 0.06376033, Cur Avg Loss: 0.16943541, Log Avg loss: 0.09720936, Global Avg Loss: 0.75675984, Time: 0.0210 Steps: 67570, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001581, Sample Num: 25296, Cur Loss: 0.17029975, Cur Avg Loss: 0.16928719, Log Avg loss: 0.14600249, Global Avg Loss: 0.75666946, Time: 0.0209 Steps: 67580, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001591, Sample Num: 25456, Cur Loss: 0.11146098, Cur Avg Loss: 0.16907710, Log Avg loss: 0.13586104, Global Avg Loss: 0.75657761, Time: 0.0209 Steps: 67590, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001601, Sample Num: 25616, Cur Loss: 0.06814194, Cur Avg Loss: 0.16930935, Log Avg loss: 0.20626089, Global Avg Loss: 0.75649621, Time: 0.0209 Steps: 67600, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001611, Sample Num: 25776, Cur Loss: 0.03340145, Cur Avg Loss: 0.16882620, Log Avg loss: 0.09147436, Global Avg Loss: 0.75639784, Time: 0.0209 Steps: 67610, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001621, Sample Num: 25936, Cur Loss: 0.58339256, Cur Avg Loss: 0.16966923, Log Avg loss: 0.30548057, Global Avg Loss: 0.75633116, Time: 0.0209 Steps: 67620, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001631, Sample Num: 26096, Cur Loss: 0.21781573, Cur Avg Loss: 0.17064677, Log Avg loss: 0.32910631, Global Avg Loss: 0.75626799, Time: 0.0210 Steps: 67630, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001641, Sample Num: 26256, Cur Loss: 0.21012016, Cur Avg Loss: 0.17105549, Log Avg loss: 0.23771803, Global Avg Loss: 0.75619133, Time: 0.0209 Steps: 67640, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001651, Sample Num: 26416, Cur Loss: 0.18930808, Cur Avg Loss: 0.17148909, Log Avg loss: 0.24264358, Global Avg Loss: 0.75611541, Time: 0.0209 Steps: 67650, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001661, Sample Num: 26576, Cur Loss: 0.04696344, Cur Avg Loss: 0.17152248, Log Avg loss: 0.17703400, Global Avg Loss: 0.75602983, Time: 0.0209 Steps: 67660, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001671, Sample Num: 26736, Cur Loss: 0.21438429, Cur Avg Loss: 0.17212720, Log Avg loss: 0.27257202, Global Avg Loss: 0.75595838, Time: 0.0209 Steps: 67670, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001681, Sample Num: 26896, Cur Loss: 0.29021919, Cur Avg Loss: 0.17265255, Log Avg loss: 0.26043853, Global Avg Loss: 0.75588517, Time: 0.0210 Steps: 67680, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001691, Sample Num: 27056, Cur Loss: 0.21802992, Cur Avg Loss: 0.17251264, Log Avg loss: 0.14899289, Global Avg Loss: 0.75579551, Time: 0.0209 Steps: 67690, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001701, Sample Num: 27216, Cur Loss: 0.10712782, Cur Avg Loss: 0.17292398, Log Avg loss: 0.24248123, Global Avg Loss: 0.75571969, Time: 0.0209 Steps: 67700, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001711, Sample Num: 27376, Cur Loss: 0.36302641, Cur Avg Loss: 0.17290521, Log Avg loss: 0.16971316, Global Avg Loss: 0.75563314, Time: 0.0210 Steps: 67710, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001721, Sample Num: 27536, Cur Loss: 0.09803830, Cur Avg Loss: 0.17285109, Log Avg loss: 0.16359174, Global Avg Loss: 0.75554572, Time: 0.0209 Steps: 67720, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001731, Sample Num: 27696, Cur Loss: 0.08193675, Cur Avg Loss: 0.17260418, Log Avg loss: 0.13011031, Global Avg Loss: 0.75545337, Time: 0.0210 Steps: 67730, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001741, Sample Num: 27856, Cur Loss: 0.30818072, Cur Avg Loss: 0.17249910, Log Avg loss: 0.15430900, Global Avg Loss: 0.75536463, Time: 0.0209 Steps: 67740, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001751, Sample Num: 28016, Cur Loss: 0.14788565, Cur Avg Loss: 0.17269504, Log Avg loss: 0.20680975, Global Avg Loss: 0.75528366, Time: 0.0209 Steps: 67750, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001761, Sample Num: 28176, Cur Loss: 0.06232096, Cur Avg Loss: 0.17241442, Log Avg loss: 0.12327725, Global Avg Loss: 0.75519039, Time: 0.0209 Steps: 67760, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001771, Sample Num: 28336, Cur Loss: 0.19787544, Cur Avg Loss: 0.17243490, Log Avg loss: 0.17604203, Global Avg Loss: 0.75510493, Time: 0.0209 Steps: 67770, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001781, Sample Num: 28496, Cur Loss: 0.10797608, Cur Avg Loss: 0.17205081, Log Avg loss: 0.10402839, Global Avg Loss: 0.75500888, Time: 0.0209 Steps: 67780, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001791, Sample Num: 28656, Cur Loss: 0.16607988, Cur Avg Loss: 0.17163934, Log Avg loss: 0.09835579, Global Avg Loss: 0.75491201, Time: 0.0209 Steps: 67790, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001801, Sample Num: 28816, Cur Loss: 0.07915539, Cur Avg Loss: 0.17149711, Log Avg loss: 0.14602353, Global Avg Loss: 0.75482221, Time: 0.0209 Steps: 67800, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001811, Sample Num: 28976, Cur Loss: 0.14392167, Cur Avg Loss: 0.17156294, Log Avg loss: 0.18341992, Global Avg Loss: 0.75473794, Time: 0.0209 Steps: 67810, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001821, Sample Num: 29136, Cur Loss: 0.45323491, Cur Avg Loss: 0.17199117, Log Avg loss: 0.24954377, Global Avg Loss: 0.75466345, Time: 0.0209 Steps: 67820, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001831, Sample Num: 29296, Cur Loss: 0.24426349, Cur Avg Loss: 0.17237539, Log Avg loss: 0.24234030, Global Avg Loss: 0.75458792, Time: 0.0209 Steps: 67830, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001841, Sample Num: 29456, Cur Loss: 0.09008013, Cur Avg Loss: 0.17243553, Log Avg loss: 0.18344775, Global Avg Loss: 0.75450373, Time: 0.0209 Steps: 67840, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001851, Sample Num: 29616, Cur Loss: 0.20880526, Cur Avg Loss: 0.17232272, Log Avg loss: 0.15155411, Global Avg Loss: 0.75441486, Time: 0.0209 Steps: 67850, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001861, Sample Num: 29776, Cur Loss: 0.07510076, Cur Avg Loss: 0.17265956, Log Avg loss: 0.23500817, Global Avg Loss: 0.75433832, Time: 0.0209 Steps: 67860, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001871, Sample Num: 29936, Cur Loss: 0.58277178, Cur Avg Loss: 0.17282352, Log Avg loss: 0.20333707, Global Avg Loss: 0.75425714, Time: 0.0209 Steps: 67870, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001881, Sample Num: 30096, Cur Loss: 0.09172323, Cur Avg Loss: 0.17290990, Log Avg loss: 0.18907179, Global Avg Loss: 0.75417388, Time: 0.0208 Steps: 67880, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001891, Sample Num: 30256, Cur Loss: 0.21184894, Cur Avg Loss: 0.17303134, Log Avg loss: 0.19587373, Global Avg Loss: 0.75409164, Time: 0.0208 Steps: 67890, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001901, Sample Num: 30416, Cur Loss: 0.15159658, Cur Avg Loss: 0.17282028, Log Avg loss: 0.13290913, Global Avg Loss: 0.75400016, Time: 0.0208 Steps: 67900, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001911, Sample Num: 30576, Cur Loss: 0.04252451, Cur Avg Loss: 0.17263290, Log Avg loss: 0.13701242, Global Avg Loss: 0.75390930, Time: 0.0208 Steps: 67910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001921, Sample Num: 30736, Cur Loss: 0.09074490, Cur Avg Loss: 0.17250009, Log Avg loss: 0.14711940, Global Avg Loss: 0.75381996, Time: 0.0209 Steps: 67920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001931, Sample Num: 30896, Cur Loss: 0.09756412, Cur Avg Loss: 0.17272238, Log Avg loss: 0.21542416, Global Avg Loss: 0.75374071, Time: 0.0208 Steps: 67930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001941, Sample Num: 31056, Cur Loss: 0.24731115, Cur Avg Loss: 0.17302288, Log Avg loss: 0.23104937, Global Avg Loss: 0.75366377, Time: 0.0210 Steps: 67940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001951, Sample Num: 31216, Cur Loss: 0.12534150, Cur Avg Loss: 0.17324355, Log Avg loss: 0.21607610, Global Avg Loss: 0.75358466, Time: 0.0209 Steps: 67950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001961, Sample Num: 31376, Cur Loss: 0.07491533, Cur Avg Loss: 0.17339411, Log Avg loss: 0.20276793, Global Avg Loss: 0.75350361, Time: 0.0209 Steps: 67960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001971, Sample Num: 31536, Cur Loss: 0.17950307, Cur Avg Loss: 0.17329242, Log Avg loss: 0.15335117, Global Avg Loss: 0.75341531, Time: 0.0208 Steps: 67970, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001981, Sample Num: 31696, Cur Loss: 0.22282818, Cur Avg Loss: 0.17312972, Log Avg loss: 0.14106199, Global Avg Loss: 0.75332523, Time: 0.0208 Steps: 67980, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001991, Sample Num: 31856, Cur Loss: 0.13164325, Cur Avg Loss: 0.17316977, Log Avg loss: 0.18110351, Global Avg Loss: 0.75324107, Time: 0.0209 Steps: 67990, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002001, Sample Num: 32016, Cur Loss: 0.08314073, Cur Avg Loss: 0.17341934, Log Avg loss: 0.22310819, Global Avg Loss: 0.75316311, Time: 0.0208 Steps: 68000, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002011, Sample Num: 32176, Cur Loss: 0.05452966, Cur Avg Loss: 0.17320270, Log Avg loss: 0.12985451, Global Avg Loss: 0.75307146, Time: 0.0209 Steps: 68010, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002021, Sample Num: 32336, Cur Loss: 0.17112413, Cur Avg Loss: 0.17305282, Log Avg loss: 0.14291180, Global Avg Loss: 0.75298175, Time: 0.0208 Steps: 68020, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002031, Sample Num: 32496, Cur Loss: 0.18992171, Cur Avg Loss: 0.17318405, Log Avg loss: 0.19970547, Global Avg Loss: 0.75290043, Time: 0.0209 Steps: 68030, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002041, Sample Num: 32656, Cur Loss: 0.17152286, Cur Avg Loss: 0.17332304, Log Avg loss: 0.20155126, Global Avg Loss: 0.75281939, Time: 0.0209 Steps: 68040, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002051, Sample Num: 32816, Cur Loss: 0.23712407, Cur Avg Loss: 0.17351054, Log Avg loss: 0.21177940, Global Avg Loss: 0.75273989, Time: 0.0248 Steps: 68050, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002061, Sample Num: 32976, Cur Loss: 0.14033791, Cur Avg Loss: 0.17345088, Log Avg loss: 0.16121549, Global Avg Loss: 0.75265298, Time: 0.0211 Steps: 68060, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002071, Sample Num: 33136, Cur Loss: 0.24154142, Cur Avg Loss: 0.17373575, Log Avg loss: 0.23244645, Global Avg Loss: 0.75257655, Time: 0.0211 Steps: 68070, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002081, Sample Num: 33296, Cur Loss: 0.33980122, Cur Avg Loss: 0.17403719, Log Avg loss: 0.23646674, Global Avg Loss: 0.75250074, Time: 0.0211 Steps: 68080, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002091, Sample Num: 33456, Cur Loss: 0.14030646, Cur Avg Loss: 0.17400106, Log Avg loss: 0.16648209, Global Avg Loss: 0.75241468, Time: 0.0211 Steps: 68090, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002101, Sample Num: 33616, Cur Loss: 0.19285023, Cur Avg Loss: 0.17365347, Log Avg loss: 0.10097237, Global Avg Loss: 0.75231902, Time: 0.0211 Steps: 68100, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002111, Sample Num: 33776, Cur Loss: 0.14441003, Cur Avg Loss: 0.17354601, Log Avg loss: 0.15096711, Global Avg Loss: 0.75223073, Time: 0.0211 Steps: 68110, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002121, Sample Num: 33936, Cur Loss: 0.21537039, Cur Avg Loss: 0.17333187, Log Avg loss: 0.12812812, Global Avg Loss: 0.75213911, Time: 0.0211 Steps: 68120, Updated lr: 0.000036 ***** Running evaluation checkpoint-68128 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-68128 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.769255, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.333518, "eval_total_loss": 234.462967, "eval_mae": 0.378089, "eval_mse": 0.333652, "eval_r2": 0.787909, "eval_sp_statistic": 0.923242, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.928843, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.242593, "test_total_loss": 121.781535, "test_mae": 0.314523, "test_mse": 0.242674, "test_r2": 0.843376, "test_sp_statistic": 0.914282, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.948156, "test_ps_pvalue": 0.0, "lr": 3.634139402560455e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7520659468566245, "train_cur_epoch_loss": 368.66961883194745, "train_cur_epoch_avg_loss": 0.1731656265063163, "train_cur_epoch_time": 44.7692551612854, "train_cur_epoch_avg_time": 0.021028302095483984, "epoch": 32, "step": 68128} ################################################## Training, Epoch: 0033, Batch: 000002, Sample Num: 32, Cur Loss: 0.11490881, Cur Avg Loss: 0.10704840, Log Avg loss: 0.12468161, Global Avg Loss: 0.75204701, Time: 0.0248 Steps: 68130, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000012, Sample Num: 192, Cur Loss: 0.40649420, Cur Avg Loss: 0.19018624, Log Avg loss: 0.20681381, Global Avg Loss: 0.75196700, Time: 0.0210 Steps: 68140, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000022, Sample Num: 352, Cur Loss: 0.10263400, Cur Avg Loss: 0.20321268, Log Avg loss: 0.21884442, Global Avg Loss: 0.75188877, Time: 0.0210 Steps: 68150, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000032, Sample Num: 512, Cur Loss: 0.10839279, Cur Avg Loss: 0.19298395, Log Avg loss: 0.17048074, Global Avg Loss: 0.75180347, Time: 0.0209 Steps: 68160, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000042, Sample Num: 672, Cur Loss: 0.19242626, Cur Avg Loss: 0.18899219, Log Avg loss: 0.17621855, Global Avg Loss: 0.75171903, Time: 0.0209 Steps: 68170, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000052, Sample Num: 832, Cur Loss: 0.15078762, Cur Avg Loss: 0.18205142, Log Avg loss: 0.15290019, Global Avg Loss: 0.75163120, Time: 0.0209 Steps: 68180, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000062, Sample Num: 992, Cur Loss: 0.17939150, Cur Avg Loss: 0.18545821, Log Avg loss: 0.20317355, Global Avg Loss: 0.75155077, Time: 0.0209 Steps: 68190, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000072, Sample Num: 1152, Cur Loss: 0.15764219, Cur Avg Loss: 0.18554911, Log Avg loss: 0.18611266, Global Avg Loss: 0.75146786, Time: 0.0209 Steps: 68200, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000082, Sample Num: 1312, Cur Loss: 0.09511954, Cur Avg Loss: 0.18297533, Log Avg loss: 0.16444410, Global Avg Loss: 0.75138180, Time: 0.0210 Steps: 68210, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000092, Sample Num: 1472, Cur Loss: 0.18103974, Cur Avg Loss: 0.18036834, Log Avg loss: 0.15899105, Global Avg Loss: 0.75129497, Time: 0.0209 Steps: 68220, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000102, Sample Num: 1632, Cur Loss: 0.18892357, Cur Avg Loss: 0.18127954, Log Avg loss: 0.18966259, Global Avg Loss: 0.75121265, Time: 0.0209 Steps: 68230, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000112, Sample Num: 1792, Cur Loss: 0.17106675, Cur Avg Loss: 0.18004682, Log Avg loss: 0.16747299, Global Avg Loss: 0.75112711, Time: 0.0209 Steps: 68240, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000122, Sample Num: 1952, Cur Loss: 0.12372554, Cur Avg Loss: 0.17987516, Log Avg loss: 0.17795264, Global Avg Loss: 0.75104313, Time: 0.0209 Steps: 68250, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000132, Sample Num: 2112, Cur Loss: 0.23581286, Cur Avg Loss: 0.17663759, Log Avg loss: 0.13713926, Global Avg Loss: 0.75095319, Time: 0.0209 Steps: 68260, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000142, Sample Num: 2272, Cur Loss: 0.15865034, Cur Avg Loss: 0.17174032, Log Avg loss: 0.10709627, Global Avg Loss: 0.75085888, Time: 0.0210 Steps: 68270, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000152, Sample Num: 2432, Cur Loss: 0.15970564, Cur Avg Loss: 0.17267752, Log Avg loss: 0.18598586, Global Avg Loss: 0.75077615, Time: 0.0209 Steps: 68280, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000162, Sample Num: 2592, Cur Loss: 0.12951177, Cur Avg Loss: 0.17333044, Log Avg loss: 0.18325476, Global Avg Loss: 0.75069305, Time: 0.0209 Steps: 68290, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000172, Sample Num: 2752, Cur Loss: 0.20195283, Cur Avg Loss: 0.17619549, Log Avg loss: 0.22260926, Global Avg Loss: 0.75061573, Time: 0.0209 Steps: 68300, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000182, Sample Num: 2912, Cur Loss: 0.16256134, Cur Avg Loss: 0.17422923, Log Avg loss: 0.14040967, Global Avg Loss: 0.75052640, Time: 0.0210 Steps: 68310, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000192, Sample Num: 3072, Cur Loss: 0.70028126, Cur Avg Loss: 0.17428954, Log Avg loss: 0.17538721, Global Avg Loss: 0.75044222, Time: 0.0209 Steps: 68320, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000202, Sample Num: 3232, Cur Loss: 0.10882310, Cur Avg Loss: 0.17048002, Log Avg loss: 0.09733717, Global Avg Loss: 0.75034664, Time: 0.0210 Steps: 68330, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000212, Sample Num: 3392, Cur Loss: 0.08413355, Cur Avg Loss: 0.16735736, Log Avg loss: 0.10427953, Global Avg Loss: 0.75025210, Time: 0.0209 Steps: 68340, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000222, Sample Num: 3552, Cur Loss: 0.08526827, Cur Avg Loss: 0.16492830, Log Avg loss: 0.11343228, Global Avg Loss: 0.75015893, Time: 0.0209 Steps: 68350, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000232, Sample Num: 3712, Cur Loss: 0.08230480, Cur Avg Loss: 0.16461995, Log Avg loss: 0.15777460, Global Avg Loss: 0.75007227, Time: 0.0209 Steps: 68360, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000242, Sample Num: 3872, Cur Loss: 0.04072649, Cur Avg Loss: 0.16420452, Log Avg loss: 0.15456652, Global Avg Loss: 0.74998517, Time: 0.0208 Steps: 68370, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000252, Sample Num: 4032, Cur Loss: 0.05855327, Cur Avg Loss: 0.16687663, Log Avg loss: 0.23154180, Global Avg Loss: 0.74990936, Time: 0.0209 Steps: 68380, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000262, Sample Num: 4192, Cur Loss: 0.18409403, Cur Avg Loss: 0.17086714, Log Avg loss: 0.27142791, Global Avg Loss: 0.74983939, Time: 0.0231 Steps: 68390, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000272, Sample Num: 4352, Cur Loss: 0.10523762, Cur Avg Loss: 0.17139415, Log Avg loss: 0.18520185, Global Avg Loss: 0.74975684, Time: 0.0208 Steps: 68400, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000282, Sample Num: 4512, Cur Loss: 0.25583309, Cur Avg Loss: 0.17266635, Log Avg loss: 0.20727012, Global Avg Loss: 0.74967754, Time: 0.0208 Steps: 68410, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000292, Sample Num: 4672, Cur Loss: 0.17578460, Cur Avg Loss: 0.17335040, Log Avg loss: 0.19264063, Global Avg Loss: 0.74959613, Time: 0.0208 Steps: 68420, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000302, Sample Num: 4832, Cur Loss: 0.20212333, Cur Avg Loss: 0.17460145, Log Avg loss: 0.21113212, Global Avg Loss: 0.74951744, Time: 0.0208 Steps: 68430, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000312, Sample Num: 4992, Cur Loss: 0.07937087, Cur Avg Loss: 0.17301763, Log Avg loss: 0.12518616, Global Avg Loss: 0.74942622, Time: 0.0208 Steps: 68440, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000322, Sample Num: 5152, Cur Loss: 0.15510145, Cur Avg Loss: 0.17282754, Log Avg loss: 0.16689681, Global Avg Loss: 0.74934111, Time: 0.0208 Steps: 68450, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000332, Sample Num: 5312, Cur Loss: 0.42248505, Cur Avg Loss: 0.17381140, Log Avg loss: 0.20549175, Global Avg Loss: 0.74926167, Time: 0.0208 Steps: 68460, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000342, Sample Num: 5472, Cur Loss: 0.08975517, Cur Avg Loss: 0.17444328, Log Avg loss: 0.19542177, Global Avg Loss: 0.74918079, Time: 0.0208 Steps: 68470, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000352, Sample Num: 5632, Cur Loss: 0.07329177, Cur Avg Loss: 0.17645433, Log Avg loss: 0.24523217, Global Avg Loss: 0.74910720, Time: 0.0208 Steps: 68480, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000362, Sample Num: 5792, Cur Loss: 0.07443471, Cur Avg Loss: 0.17590252, Log Avg loss: 0.15647873, Global Avg Loss: 0.74902067, Time: 0.0208 Steps: 68490, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000372, Sample Num: 5952, Cur Loss: 0.12471707, Cur Avg Loss: 0.17570549, Log Avg loss: 0.16857294, Global Avg Loss: 0.74893593, Time: 0.0208 Steps: 68500, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000382, Sample Num: 6112, Cur Loss: 0.15373276, Cur Avg Loss: 0.17645129, Log Avg loss: 0.20419528, Global Avg Loss: 0.74885642, Time: 0.0208 Steps: 68510, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000392, Sample Num: 6272, Cur Loss: 0.16964695, Cur Avg Loss: 0.17624536, Log Avg loss: 0.16837862, Global Avg Loss: 0.74877170, Time: 0.0208 Steps: 68520, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000402, Sample Num: 6432, Cur Loss: 0.13128363, Cur Avg Loss: 0.17532149, Log Avg loss: 0.13910599, Global Avg Loss: 0.74868274, Time: 0.0209 Steps: 68530, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000412, Sample Num: 6592, Cur Loss: 0.03115850, Cur Avg Loss: 0.17431701, Log Avg loss: 0.13393683, Global Avg Loss: 0.74859305, Time: 0.0208 Steps: 68540, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000422, Sample Num: 6752, Cur Loss: 0.30229762, Cur Avg Loss: 0.17378064, Log Avg loss: 0.15168222, Global Avg Loss: 0.74850597, Time: 0.0208 Steps: 68550, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000432, Sample Num: 6912, Cur Loss: 0.52089894, Cur Avg Loss: 0.17401675, Log Avg loss: 0.18398047, Global Avg Loss: 0.74842363, Time: 0.0208 Steps: 68560, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000442, Sample Num: 7072, Cur Loss: 0.02458307, Cur Avg Loss: 0.17341367, Log Avg loss: 0.14736065, Global Avg Loss: 0.74833597, Time: 0.0208 Steps: 68570, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000452, Sample Num: 7232, Cur Loss: 0.17140216, Cur Avg Loss: 0.17253014, Log Avg loss: 0.13347807, Global Avg Loss: 0.74824632, Time: 0.0208 Steps: 68580, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000462, Sample Num: 7392, Cur Loss: 0.22284386, Cur Avg Loss: 0.17104709, Log Avg loss: 0.10401342, Global Avg Loss: 0.74815239, Time: 0.0208 Steps: 68590, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000472, Sample Num: 7552, Cur Loss: 0.08307542, Cur Avg Loss: 0.17006109, Log Avg loss: 0.12450793, Global Avg Loss: 0.74806148, Time: 0.0208 Steps: 68600, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000482, Sample Num: 7712, Cur Loss: 0.42648241, Cur Avg Loss: 0.17060092, Log Avg loss: 0.19608087, Global Avg Loss: 0.74798103, Time: 0.0208 Steps: 68610, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000492, Sample Num: 7872, Cur Loss: 0.02552924, Cur Avg Loss: 0.16883820, Log Avg loss: 0.08387514, Global Avg Loss: 0.74788425, Time: 0.0208 Steps: 68620, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000502, Sample Num: 8032, Cur Loss: 0.38649932, Cur Avg Loss: 0.16835799, Log Avg loss: 0.14473145, Global Avg Loss: 0.74779637, Time: 0.0208 Steps: 68630, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000512, Sample Num: 8192, Cur Loss: 0.05705273, Cur Avg Loss: 0.16712970, Log Avg loss: 0.10546954, Global Avg Loss: 0.74770279, Time: 0.0254 Steps: 68640, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000522, Sample Num: 8352, Cur Loss: 0.10398085, Cur Avg Loss: 0.16777984, Log Avg loss: 0.20106701, Global Avg Loss: 0.74762316, Time: 0.0208 Steps: 68650, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000532, Sample Num: 8512, Cur Loss: 0.15101540, Cur Avg Loss: 0.16917464, Log Avg loss: 0.24198321, Global Avg Loss: 0.74754952, Time: 0.0208 Steps: 68660, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000542, Sample Num: 8672, Cur Loss: 0.14173093, Cur Avg Loss: 0.16827687, Log Avg loss: 0.12051575, Global Avg Loss: 0.74745820, Time: 0.0208 Steps: 68670, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000552, Sample Num: 8832, Cur Loss: 0.08300798, Cur Avg Loss: 0.16945260, Log Avg loss: 0.23317717, Global Avg Loss: 0.74738332, Time: 0.0208 Steps: 68680, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000562, Sample Num: 8992, Cur Loss: 0.28044468, Cur Avg Loss: 0.16970465, Log Avg loss: 0.18361733, Global Avg Loss: 0.74730125, Time: 0.0208 Steps: 68690, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000572, Sample Num: 9152, Cur Loss: 0.03739563, Cur Avg Loss: 0.16914982, Log Avg loss: 0.13796865, Global Avg Loss: 0.74721255, Time: 0.0208 Steps: 68700, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000582, Sample Num: 9312, Cur Loss: 0.20316362, Cur Avg Loss: 0.16986755, Log Avg loss: 0.21092169, Global Avg Loss: 0.74713450, Time: 0.0208 Steps: 68710, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000592, Sample Num: 9472, Cur Loss: 0.18283519, Cur Avg Loss: 0.16980082, Log Avg loss: 0.16591700, Global Avg Loss: 0.74704993, Time: 0.0208 Steps: 68720, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000602, Sample Num: 9632, Cur Loss: 0.07864411, Cur Avg Loss: 0.16950386, Log Avg loss: 0.15192414, Global Avg Loss: 0.74696334, Time: 0.0208 Steps: 68730, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000612, Sample Num: 9792, Cur Loss: 0.24127075, Cur Avg Loss: 0.16937748, Log Avg loss: 0.16176933, Global Avg Loss: 0.74687821, Time: 0.0208 Steps: 68740, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000622, Sample Num: 9952, Cur Loss: 0.16867593, Cur Avg Loss: 0.17027354, Log Avg loss: 0.22511199, Global Avg Loss: 0.74680231, Time: 0.0208 Steps: 68750, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000632, Sample Num: 10112, Cur Loss: 0.07426637, Cur Avg Loss: 0.17035900, Log Avg loss: 0.17567482, Global Avg Loss: 0.74671925, Time: 0.0208 Steps: 68760, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000642, Sample Num: 10272, Cur Loss: 0.09879404, Cur Avg Loss: 0.16946794, Log Avg loss: 0.11315286, Global Avg Loss: 0.74662712, Time: 0.0209 Steps: 68770, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000652, Sample Num: 10432, Cur Loss: 0.16658923, Cur Avg Loss: 0.16996946, Log Avg loss: 0.20216749, Global Avg Loss: 0.74654796, Time: 0.0208 Steps: 68780, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000662, Sample Num: 10592, Cur Loss: 0.11716761, Cur Avg Loss: 0.16928698, Log Avg loss: 0.12478933, Global Avg Loss: 0.74645758, Time: 0.0208 Steps: 68790, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000672, Sample Num: 10752, Cur Loss: 0.32128865, Cur Avg Loss: 0.16963709, Log Avg loss: 0.19281381, Global Avg Loss: 0.74637711, Time: 0.0208 Steps: 68800, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000682, Sample Num: 10912, Cur Loss: 0.05239275, Cur Avg Loss: 0.16963952, Log Avg loss: 0.16980287, Global Avg Loss: 0.74629331, Time: 0.0208 Steps: 68810, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000692, Sample Num: 11072, Cur Loss: 0.26381999, Cur Avg Loss: 0.16959563, Log Avg loss: 0.16660227, Global Avg Loss: 0.74620908, Time: 0.0208 Steps: 68820, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000702, Sample Num: 11232, Cur Loss: 0.04553260, Cur Avg Loss: 0.16918670, Log Avg loss: 0.14088868, Global Avg Loss: 0.74612114, Time: 0.0208 Steps: 68830, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000712, Sample Num: 11392, Cur Loss: 0.06832910, Cur Avg Loss: 0.16894110, Log Avg loss: 0.15170012, Global Avg Loss: 0.74603479, Time: 0.0208 Steps: 68840, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000722, Sample Num: 11552, Cur Loss: 0.17128401, Cur Avg Loss: 0.16885562, Log Avg loss: 0.16276936, Global Avg Loss: 0.74595007, Time: 0.0208 Steps: 68850, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000732, Sample Num: 11712, Cur Loss: 0.12300692, Cur Avg Loss: 0.16893363, Log Avg loss: 0.17456634, Global Avg Loss: 0.74586710, Time: 0.0208 Steps: 68860, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000742, Sample Num: 11872, Cur Loss: 0.37642375, Cur Avg Loss: 0.16834970, Log Avg loss: 0.12560570, Global Avg Loss: 0.74577703, Time: 0.0208 Steps: 68870, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000752, Sample Num: 12032, Cur Loss: 0.07774327, Cur Avg Loss: 0.16750128, Log Avg loss: 0.10454867, Global Avg Loss: 0.74568394, Time: 0.0208 Steps: 68880, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000762, Sample Num: 12192, Cur Loss: 0.20394978, Cur Avg Loss: 0.16748335, Log Avg loss: 0.16613484, Global Avg Loss: 0.74559981, Time: 0.0208 Steps: 68890, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000772, Sample Num: 12352, Cur Loss: 0.23332272, Cur Avg Loss: 0.16772686, Log Avg loss: 0.18628226, Global Avg Loss: 0.74551864, Time: 0.0227 Steps: 68900, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000782, Sample Num: 12512, Cur Loss: 0.03508738, Cur Avg Loss: 0.16713281, Log Avg loss: 0.12127281, Global Avg Loss: 0.74542805, Time: 0.0208 Steps: 68910, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000792, Sample Num: 12672, Cur Loss: 0.58396828, Cur Avg Loss: 0.16765743, Log Avg loss: 0.20868271, Global Avg Loss: 0.74535017, Time: 0.0208 Steps: 68920, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000802, Sample Num: 12832, Cur Loss: 0.30123731, Cur Avg Loss: 0.16802207, Log Avg loss: 0.19690087, Global Avg Loss: 0.74527060, Time: 0.0209 Steps: 68930, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000812, Sample Num: 12992, Cur Loss: 0.10055643, Cur Avg Loss: 0.16821679, Log Avg loss: 0.18383402, Global Avg Loss: 0.74518916, Time: 0.0208 Steps: 68940, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000822, Sample Num: 13152, Cur Loss: 0.53889757, Cur Avg Loss: 0.16919595, Log Avg loss: 0.24870307, Global Avg Loss: 0.74511716, Time: 0.0208 Steps: 68950, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000832, Sample Num: 13312, Cur Loss: 0.23226905, Cur Avg Loss: 0.17024332, Log Avg loss: 0.25633714, Global Avg Loss: 0.74504628, Time: 0.0208 Steps: 68960, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000842, Sample Num: 13472, Cur Loss: 0.12848930, Cur Avg Loss: 0.17020420, Log Avg loss: 0.16695004, Global Avg Loss: 0.74496246, Time: 0.0209 Steps: 68970, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000852, Sample Num: 13632, Cur Loss: 0.01389799, Cur Avg Loss: 0.16947560, Log Avg loss: 0.10812750, Global Avg Loss: 0.74487014, Time: 0.0209 Steps: 68980, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000862, Sample Num: 13792, Cur Loss: 0.04912544, Cur Avg Loss: 0.16996272, Log Avg loss: 0.21146520, Global Avg Loss: 0.74479282, Time: 0.0208 Steps: 68990, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000872, Sample Num: 13952, Cur Loss: 0.09393886, Cur Avg Loss: 0.16998917, Log Avg loss: 0.17226859, Global Avg Loss: 0.74470985, Time: 0.0208 Steps: 69000, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000882, Sample Num: 14112, Cur Loss: 0.26009542, Cur Avg Loss: 0.17073559, Log Avg loss: 0.23582404, Global Avg Loss: 0.74463611, Time: 0.0209 Steps: 69010, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000892, Sample Num: 14272, Cur Loss: 0.34889948, Cur Avg Loss: 0.17153404, Log Avg loss: 0.24195686, Global Avg Loss: 0.74456327, Time: 0.0208 Steps: 69020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000902, Sample Num: 14432, Cur Loss: 0.33634919, Cur Avg Loss: 0.17149598, Log Avg loss: 0.16810099, Global Avg Loss: 0.74447977, Time: 0.0208 Steps: 69030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000912, Sample Num: 14592, Cur Loss: 0.08497451, Cur Avg Loss: 0.17246952, Log Avg loss: 0.26028281, Global Avg Loss: 0.74440963, Time: 0.0208 Steps: 69040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000922, Sample Num: 14752, Cur Loss: 0.12891267, Cur Avg Loss: 0.17355536, Log Avg loss: 0.27258407, Global Avg Loss: 0.74434130, Time: 0.0208 Steps: 69050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000932, Sample Num: 14912, Cur Loss: 0.10327217, Cur Avg Loss: 0.17299647, Log Avg loss: 0.12146698, Global Avg Loss: 0.74425111, Time: 0.0208 Steps: 69060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000942, Sample Num: 15072, Cur Loss: 0.19084845, Cur Avg Loss: 0.17268207, Log Avg loss: 0.14338015, Global Avg Loss: 0.74416411, Time: 0.0209 Steps: 69070, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000952, Sample Num: 15232, Cur Loss: 0.20954335, Cur Avg Loss: 0.17234045, Log Avg loss: 0.14015934, Global Avg Loss: 0.74407668, Time: 0.0208 Steps: 69080, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000962, Sample Num: 15392, Cur Loss: 0.22454050, Cur Avg Loss: 0.17218756, Log Avg loss: 0.15763283, Global Avg Loss: 0.74399180, Time: 0.0208 Steps: 69090, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000972, Sample Num: 15552, Cur Loss: 0.20477161, Cur Avg Loss: 0.17202337, Log Avg loss: 0.15622786, Global Avg Loss: 0.74390674, Time: 0.0208 Steps: 69100, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000982, Sample Num: 15712, Cur Loss: 0.29961053, Cur Avg Loss: 0.17150944, Log Avg loss: 0.12155627, Global Avg Loss: 0.74381669, Time: 0.0208 Steps: 69110, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000992, Sample Num: 15872, Cur Loss: 0.18986973, Cur Avg Loss: 0.17204125, Log Avg loss: 0.22426448, Global Avg Loss: 0.74374152, Time: 0.0208 Steps: 69120, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001002, Sample Num: 16032, Cur Loss: 0.10971490, Cur Avg Loss: 0.17174003, Log Avg loss: 0.14185942, Global Avg Loss: 0.74365445, Time: 0.0208 Steps: 69130, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001012, Sample Num: 16192, Cur Loss: 0.06462809, Cur Avg Loss: 0.17129354, Log Avg loss: 0.12655531, Global Avg Loss: 0.74356520, Time: 0.0208 Steps: 69140, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001022, Sample Num: 16352, Cur Loss: 0.03763844, Cur Avg Loss: 0.17047239, Log Avg loss: 0.08737178, Global Avg Loss: 0.74347031, Time: 0.0208 Steps: 69150, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001032, Sample Num: 16512, Cur Loss: 0.04460521, Cur Avg Loss: 0.16988368, Log Avg loss: 0.10971688, Global Avg Loss: 0.74337867, Time: 0.0210 Steps: 69160, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001042, Sample Num: 16672, Cur Loss: 0.26666492, Cur Avg Loss: 0.17025109, Log Avg loss: 0.20816838, Global Avg Loss: 0.74330129, Time: 0.0208 Steps: 69170, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001052, Sample Num: 16832, Cur Loss: 0.17049921, Cur Avg Loss: 0.17042064, Log Avg loss: 0.18808773, Global Avg Loss: 0.74322104, Time: 0.0208 Steps: 69180, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001062, Sample Num: 16992, Cur Loss: 0.05192360, Cur Avg Loss: 0.17040497, Log Avg loss: 0.16875591, Global Avg Loss: 0.74313801, Time: 0.0209 Steps: 69190, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001072, Sample Num: 17152, Cur Loss: 0.12871467, Cur Avg Loss: 0.17014457, Log Avg loss: 0.14249080, Global Avg Loss: 0.74305121, Time: 0.0208 Steps: 69200, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001082, Sample Num: 17312, Cur Loss: 0.14594276, Cur Avg Loss: 0.16987101, Log Avg loss: 0.14054554, Global Avg Loss: 0.74296416, Time: 0.0209 Steps: 69210, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001092, Sample Num: 17472, Cur Loss: 0.19740337, Cur Avg Loss: 0.16957810, Log Avg loss: 0.13788434, Global Avg Loss: 0.74287674, Time: 0.0208 Steps: 69220, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001102, Sample Num: 17632, Cur Loss: 0.12947845, Cur Avg Loss: 0.16927812, Log Avg loss: 0.13652073, Global Avg Loss: 0.74278916, Time: 0.0208 Steps: 69230, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001112, Sample Num: 17792, Cur Loss: 0.12970801, Cur Avg Loss: 0.16923813, Log Avg loss: 0.16483122, Global Avg Loss: 0.74270568, Time: 0.0208 Steps: 69240, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001122, Sample Num: 17952, Cur Loss: 0.09875397, Cur Avg Loss: 0.16911281, Log Avg loss: 0.15517760, Global Avg Loss: 0.74262084, Time: 0.0208 Steps: 69250, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001132, Sample Num: 18112, Cur Loss: 0.28986475, Cur Avg Loss: 0.16933078, Log Avg loss: 0.19378664, Global Avg Loss: 0.74254160, Time: 0.0208 Steps: 69260, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001142, Sample Num: 18272, Cur Loss: 0.07247569, Cur Avg Loss: 0.16890844, Log Avg loss: 0.12109933, Global Avg Loss: 0.74245189, Time: 0.0208 Steps: 69270, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001152, Sample Num: 18432, Cur Loss: 0.67699242, Cur Avg Loss: 0.16954992, Log Avg loss: 0.24280709, Global Avg Loss: 0.74237977, Time: 0.0208 Steps: 69280, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001162, Sample Num: 18592, Cur Loss: 0.14838116, Cur Avg Loss: 0.17116098, Log Avg loss: 0.35675518, Global Avg Loss: 0.74232411, Time: 0.0208 Steps: 69290, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001172, Sample Num: 18752, Cur Loss: 0.08764537, Cur Avg Loss: 0.17131881, Log Avg loss: 0.18965923, Global Avg Loss: 0.74224436, Time: 0.0208 Steps: 69300, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001182, Sample Num: 18912, Cur Loss: 0.02323889, Cur Avg Loss: 0.17121459, Log Avg loss: 0.15900007, Global Avg Loss: 0.74216021, Time: 0.0208 Steps: 69310, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001192, Sample Num: 19072, Cur Loss: 0.25219217, Cur Avg Loss: 0.17126769, Log Avg loss: 0.17754401, Global Avg Loss: 0.74207876, Time: 0.0209 Steps: 69320, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001202, Sample Num: 19232, Cur Loss: 0.03925653, Cur Avg Loss: 0.17170686, Log Avg loss: 0.22405587, Global Avg Loss: 0.74200405, Time: 0.0208 Steps: 69330, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001212, Sample Num: 19392, Cur Loss: 0.42728978, Cur Avg Loss: 0.17183162, Log Avg loss: 0.18682764, Global Avg Loss: 0.74192398, Time: 0.0208 Steps: 69340, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001222, Sample Num: 19552, Cur Loss: 0.39134789, Cur Avg Loss: 0.17211936, Log Avg loss: 0.20699260, Global Avg Loss: 0.74184684, Time: 0.0208 Steps: 69350, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001232, Sample Num: 19712, Cur Loss: 0.10494897, Cur Avg Loss: 0.17237111, Log Avg loss: 0.20313576, Global Avg Loss: 0.74176918, Time: 0.0208 Steps: 69360, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001242, Sample Num: 19872, Cur Loss: 0.04128212, Cur Avg Loss: 0.17244836, Log Avg loss: 0.18196491, Global Avg Loss: 0.74168848, Time: 0.0208 Steps: 69370, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001252, Sample Num: 20032, Cur Loss: 0.11594343, Cur Avg Loss: 0.17262810, Log Avg loss: 0.19495165, Global Avg Loss: 0.74160967, Time: 0.0208 Steps: 69380, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001262, Sample Num: 20192, Cur Loss: 0.21188712, Cur Avg Loss: 0.17274328, Log Avg loss: 0.18716462, Global Avg Loss: 0.74152977, Time: 0.0208 Steps: 69390, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001272, Sample Num: 20352, Cur Loss: 0.03961781, Cur Avg Loss: 0.17280685, Log Avg loss: 0.18082903, Global Avg Loss: 0.74144898, Time: 0.0208 Steps: 69400, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001282, Sample Num: 20512, Cur Loss: 0.08460193, Cur Avg Loss: 0.17268930, Log Avg loss: 0.15773713, Global Avg Loss: 0.74136488, Time: 0.0247 Steps: 69410, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001292, Sample Num: 20672, Cur Loss: 0.10981716, Cur Avg Loss: 0.17256596, Log Avg loss: 0.15675388, Global Avg Loss: 0.74128067, Time: 0.0209 Steps: 69420, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001302, Sample Num: 20832, Cur Loss: 0.29258895, Cur Avg Loss: 0.17239048, Log Avg loss: 0.14971822, Global Avg Loss: 0.74119547, Time: 0.0208 Steps: 69430, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001312, Sample Num: 20992, Cur Loss: 0.08059829, Cur Avg Loss: 0.17233679, Log Avg loss: 0.16534597, Global Avg Loss: 0.74111254, Time: 0.0208 Steps: 69440, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001322, Sample Num: 21152, Cur Loss: 0.14477277, Cur Avg Loss: 0.17241041, Log Avg loss: 0.18207006, Global Avg Loss: 0.74103204, Time: 0.0208 Steps: 69450, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001332, Sample Num: 21312, Cur Loss: 0.45518416, Cur Avg Loss: 0.17312815, Log Avg loss: 0.26801280, Global Avg Loss: 0.74096394, Time: 0.0210 Steps: 69460, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001342, Sample Num: 21472, Cur Loss: 0.12198730, Cur Avg Loss: 0.17291690, Log Avg loss: 0.14477904, Global Avg Loss: 0.74087812, Time: 0.0209 Steps: 69470, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001352, Sample Num: 21632, Cur Loss: 0.05952504, Cur Avg Loss: 0.17255432, Log Avg loss: 0.12389545, Global Avg Loss: 0.74078932, Time: 0.0208 Steps: 69480, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001362, Sample Num: 21792, Cur Loss: 0.08968203, Cur Avg Loss: 0.17250155, Log Avg loss: 0.16536761, Global Avg Loss: 0.74070652, Time: 0.0208 Steps: 69490, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001372, Sample Num: 21952, Cur Loss: 0.07423005, Cur Avg Loss: 0.17266616, Log Avg loss: 0.19508513, Global Avg Loss: 0.74062801, Time: 0.0209 Steps: 69500, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001382, Sample Num: 22112, Cur Loss: 0.05054872, Cur Avg Loss: 0.17201799, Log Avg loss: 0.08309025, Global Avg Loss: 0.74053342, Time: 0.0209 Steps: 69510, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001392, Sample Num: 22272, Cur Loss: 0.10864432, Cur Avg Loss: 0.17198384, Log Avg loss: 0.16726315, Global Avg Loss: 0.74045095, Time: 0.0209 Steps: 69520, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001402, Sample Num: 22432, Cur Loss: 0.25389260, Cur Avg Loss: 0.17217990, Log Avg loss: 0.19947243, Global Avg Loss: 0.74037315, Time: 0.0208 Steps: 69530, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001412, Sample Num: 22592, Cur Loss: 0.14421979, Cur Avg Loss: 0.17238609, Log Avg loss: 0.20129349, Global Avg Loss: 0.74029563, Time: 0.0209 Steps: 69540, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001422, Sample Num: 22752, Cur Loss: 0.09328973, Cur Avg Loss: 0.17270042, Log Avg loss: 0.21708329, Global Avg Loss: 0.74022040, Time: 0.0209 Steps: 69550, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001432, Sample Num: 22912, Cur Loss: 0.06539394, Cur Avg Loss: 0.17256418, Log Avg loss: 0.15319154, Global Avg Loss: 0.74013601, Time: 0.0208 Steps: 69560, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001442, Sample Num: 23072, Cur Loss: 0.23778099, Cur Avg Loss: 0.17295759, Log Avg loss: 0.22929381, Global Avg Loss: 0.74006258, Time: 0.0209 Steps: 69570, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001452, Sample Num: 23232, Cur Loss: 0.08855529, Cur Avg Loss: 0.17290850, Log Avg loss: 0.16583020, Global Avg Loss: 0.73998005, Time: 0.0208 Steps: 69580, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001462, Sample Num: 23392, Cur Loss: 0.31150097, Cur Avg Loss: 0.17288621, Log Avg loss: 0.16964955, Global Avg Loss: 0.73989810, Time: 0.0209 Steps: 69590, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001472, Sample Num: 23552, Cur Loss: 0.22966632, Cur Avg Loss: 0.17345200, Log Avg loss: 0.25617039, Global Avg Loss: 0.73982859, Time: 0.0209 Steps: 69600, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001482, Sample Num: 23712, Cur Loss: 0.06900255, Cur Avg Loss: 0.17345707, Log Avg loss: 0.17420315, Global Avg Loss: 0.73974734, Time: 0.0209 Steps: 69610, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001492, Sample Num: 23872, Cur Loss: 0.11313210, Cur Avg Loss: 0.17356368, Log Avg loss: 0.18936267, Global Avg Loss: 0.73966828, Time: 0.0208 Steps: 69620, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001502, Sample Num: 24032, Cur Loss: 0.04155193, Cur Avg Loss: 0.17355412, Log Avg loss: 0.17212778, Global Avg Loss: 0.73958677, Time: 0.0209 Steps: 69630, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001512, Sample Num: 24192, Cur Loss: 0.17956898, Cur Avg Loss: 0.17346031, Log Avg loss: 0.15937024, Global Avg Loss: 0.73950346, Time: 0.0208 Steps: 69640, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001522, Sample Num: 24352, Cur Loss: 0.02759811, Cur Avg Loss: 0.17305489, Log Avg loss: 0.11175515, Global Avg Loss: 0.73941333, Time: 0.0209 Steps: 69650, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001532, Sample Num: 24512, Cur Loss: 0.24986732, Cur Avg Loss: 0.17319626, Log Avg loss: 0.19471381, Global Avg Loss: 0.73933513, Time: 0.0209 Steps: 69660, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001542, Sample Num: 24672, Cur Loss: 0.23239586, Cur Avg Loss: 0.17320624, Log Avg loss: 0.17473449, Global Avg Loss: 0.73925410, Time: 0.0210 Steps: 69670, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001552, Sample Num: 24832, Cur Loss: 0.34774190, Cur Avg Loss: 0.17344610, Log Avg loss: 0.21043297, Global Avg Loss: 0.73917820, Time: 0.0209 Steps: 69680, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001562, Sample Num: 24992, Cur Loss: 0.31229696, Cur Avg Loss: 0.17345392, Log Avg loss: 0.17466766, Global Avg Loss: 0.73909720, Time: 0.0209 Steps: 69690, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001572, Sample Num: 25152, Cur Loss: 0.10322590, Cur Avg Loss: 0.17308670, Log Avg loss: 0.11572662, Global Avg Loss: 0.73900776, Time: 0.0209 Steps: 69700, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001582, Sample Num: 25312, Cur Loss: 0.21861634, Cur Avg Loss: 0.17333465, Log Avg loss: 0.21231247, Global Avg Loss: 0.73893221, Time: 0.0209 Steps: 69710, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001592, Sample Num: 25472, Cur Loss: 0.20350985, Cur Avg Loss: 0.17338646, Log Avg loss: 0.18158225, Global Avg Loss: 0.73885227, Time: 0.0209 Steps: 69720, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001602, Sample Num: 25632, Cur Loss: 0.04171145, Cur Avg Loss: 0.17349873, Log Avg loss: 0.19137245, Global Avg Loss: 0.73877375, Time: 0.0209 Steps: 69730, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001612, Sample Num: 25792, Cur Loss: 0.16294676, Cur Avg Loss: 0.17318513, Log Avg loss: 0.12294695, Global Avg Loss: 0.73868545, Time: 0.0210 Steps: 69740, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001622, Sample Num: 25952, Cur Loss: 0.31778383, Cur Avg Loss: 0.17351124, Log Avg loss: 0.22607945, Global Avg Loss: 0.73861196, Time: 0.0209 Steps: 69750, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001632, Sample Num: 26112, Cur Loss: 0.11463265, Cur Avg Loss: 0.17318895, Log Avg loss: 0.12091343, Global Avg Loss: 0.73852341, Time: 0.0209 Steps: 69760, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001642, Sample Num: 26272, Cur Loss: 0.10527232, Cur Avg Loss: 0.17295606, Log Avg loss: 0.13494817, Global Avg Loss: 0.73843690, Time: 0.0209 Steps: 69770, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001652, Sample Num: 26432, Cur Loss: 0.18732612, Cur Avg Loss: 0.17277902, Log Avg loss: 0.14371062, Global Avg Loss: 0.73835167, Time: 0.0209 Steps: 69780, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001662, Sample Num: 26592, Cur Loss: 0.20085546, Cur Avg Loss: 0.17270657, Log Avg loss: 0.16073744, Global Avg Loss: 0.73826891, Time: 0.0209 Steps: 69790, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001672, Sample Num: 26752, Cur Loss: 0.15651053, Cur Avg Loss: 0.17252361, Log Avg loss: 0.14211507, Global Avg Loss: 0.73818350, Time: 0.0209 Steps: 69800, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001682, Sample Num: 26912, Cur Loss: 0.34829560, Cur Avg Loss: 0.17256380, Log Avg loss: 0.17928400, Global Avg Loss: 0.73810344, Time: 0.0209 Steps: 69810, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001692, Sample Num: 27072, Cur Loss: 0.45367542, Cur Avg Loss: 0.17314692, Log Avg loss: 0.27122810, Global Avg Loss: 0.73803657, Time: 0.0209 Steps: 69820, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001702, Sample Num: 27232, Cur Loss: 0.05343451, Cur Avg Loss: 0.17364340, Log Avg loss: 0.25764676, Global Avg Loss: 0.73796778, Time: 0.0209 Steps: 69830, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001712, Sample Num: 27392, Cur Loss: 0.15892194, Cur Avg Loss: 0.17364258, Log Avg loss: 0.17350370, Global Avg Loss: 0.73788695, Time: 0.0209 Steps: 69840, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001722, Sample Num: 27552, Cur Loss: 0.20597802, Cur Avg Loss: 0.17365979, Log Avg loss: 0.17660671, Global Avg Loss: 0.73780660, Time: 0.0209 Steps: 69850, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001732, Sample Num: 27712, Cur Loss: 0.10520347, Cur Avg Loss: 0.17348890, Log Avg loss: 0.14406072, Global Avg Loss: 0.73772161, Time: 0.0209 Steps: 69860, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001742, Sample Num: 27872, Cur Loss: 0.08073081, Cur Avg Loss: 0.17334566, Log Avg loss: 0.14853660, Global Avg Loss: 0.73763728, Time: 0.0209 Steps: 69870, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001752, Sample Num: 28032, Cur Loss: 0.31088930, Cur Avg Loss: 0.17381252, Log Avg loss: 0.25513946, Global Avg Loss: 0.73756824, Time: 0.0209 Steps: 69880, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001762, Sample Num: 28192, Cur Loss: 0.09688491, Cur Avg Loss: 0.17366061, Log Avg loss: 0.14704602, Global Avg Loss: 0.73748374, Time: 0.0209 Steps: 69890, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001772, Sample Num: 28352, Cur Loss: 0.08784753, Cur Avg Loss: 0.17357876, Log Avg loss: 0.15915655, Global Avg Loss: 0.73740101, Time: 0.0209 Steps: 69900, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001782, Sample Num: 28512, Cur Loss: 0.12472641, Cur Avg Loss: 0.17368139, Log Avg loss: 0.19186773, Global Avg Loss: 0.73732297, Time: 0.0209 Steps: 69910, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001792, Sample Num: 28672, Cur Loss: 0.09174581, Cur Avg Loss: 0.17362119, Log Avg loss: 0.16289396, Global Avg Loss: 0.73724082, Time: 0.0254 Steps: 69920, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001802, Sample Num: 28832, Cur Loss: 0.19714536, Cur Avg Loss: 0.17335624, Log Avg loss: 0.12587708, Global Avg Loss: 0.73715339, Time: 0.0209 Steps: 69930, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001812, Sample Num: 28992, Cur Loss: 0.08216008, Cur Avg Loss: 0.17351434, Log Avg loss: 0.20200440, Global Avg Loss: 0.73707688, Time: 0.0209 Steps: 69940, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001822, Sample Num: 29152, Cur Loss: 0.07246566, Cur Avg Loss: 0.17371828, Log Avg loss: 0.21067094, Global Avg Loss: 0.73700162, Time: 0.0209 Steps: 69950, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001832, Sample Num: 29312, Cur Loss: 0.18806614, Cur Avg Loss: 0.17365703, Log Avg loss: 0.16249839, Global Avg Loss: 0.73691950, Time: 0.0209 Steps: 69960, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001842, Sample Num: 29472, Cur Loss: 0.16135532, Cur Avg Loss: 0.17388323, Log Avg loss: 0.21532176, Global Avg Loss: 0.73684496, Time: 0.0209 Steps: 69970, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001852, Sample Num: 29632, Cur Loss: 0.04881983, Cur Avg Loss: 0.17387512, Log Avg loss: 0.17238175, Global Avg Loss: 0.73676430, Time: 0.0209 Steps: 69980, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001862, Sample Num: 29792, Cur Loss: 0.11066673, Cur Avg Loss: 0.17371990, Log Avg loss: 0.14497400, Global Avg Loss: 0.73667974, Time: 0.0209 Steps: 69990, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001872, Sample Num: 29952, Cur Loss: 0.03942667, Cur Avg Loss: 0.17329524, Log Avg loss: 0.09422263, Global Avg Loss: 0.73658796, Time: 0.0209 Steps: 70000, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001882, Sample Num: 30112, Cur Loss: 0.13380376, Cur Avg Loss: 0.17317952, Log Avg loss: 0.15151767, Global Avg Loss: 0.73650439, Time: 0.0209 Steps: 70010, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001892, Sample Num: 30272, Cur Loss: 0.27166218, Cur Avg Loss: 0.17338719, Log Avg loss: 0.21247039, Global Avg Loss: 0.73642955, Time: 0.0209 Steps: 70020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001902, Sample Num: 30432, Cur Loss: 0.02529218, Cur Avg Loss: 0.17309481, Log Avg loss: 0.11777687, Global Avg Loss: 0.73634121, Time: 0.0209 Steps: 70030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001912, Sample Num: 30592, Cur Loss: 0.08756855, Cur Avg Loss: 0.17342332, Log Avg loss: 0.23590585, Global Avg Loss: 0.73626976, Time: 0.0209 Steps: 70040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001922, Sample Num: 30752, Cur Loss: 0.25830621, Cur Avg Loss: 0.17362284, Log Avg loss: 0.21177065, Global Avg Loss: 0.73619489, Time: 0.0209 Steps: 70050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001932, Sample Num: 30912, Cur Loss: 0.14446889, Cur Avg Loss: 0.17341679, Log Avg loss: 0.13381359, Global Avg Loss: 0.73610891, Time: 0.0209 Steps: 70060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001942, Sample Num: 31072, Cur Loss: 0.11895448, Cur Avg Loss: 0.17340939, Log Avg loss: 0.17197905, Global Avg Loss: 0.73602840, Time: 0.0209 Steps: 70070, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001952, Sample Num: 31232, Cur Loss: 0.14856774, Cur Avg Loss: 0.17328600, Log Avg loss: 0.14932457, Global Avg Loss: 0.73594468, Time: 0.0209 Steps: 70080, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001962, Sample Num: 31392, Cur Loss: 0.10122461, Cur Avg Loss: 0.17308966, Log Avg loss: 0.13476420, Global Avg Loss: 0.73585891, Time: 0.0209 Steps: 70090, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001972, Sample Num: 31552, Cur Loss: 0.06779858, Cur Avg Loss: 0.17332453, Log Avg loss: 0.21940549, Global Avg Loss: 0.73578523, Time: 0.0209 Steps: 70100, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001982, Sample Num: 31712, Cur Loss: 0.07372802, Cur Avg Loss: 0.17389392, Log Avg loss: 0.28617828, Global Avg Loss: 0.73572110, Time: 0.0209 Steps: 70110, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001992, Sample Num: 31872, Cur Loss: 0.11219250, Cur Avg Loss: 0.17419109, Log Avg loss: 0.23309028, Global Avg Loss: 0.73564942, Time: 0.0209 Steps: 70120, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002002, Sample Num: 32032, Cur Loss: 0.10287160, Cur Avg Loss: 0.17427046, Log Avg loss: 0.19008066, Global Avg Loss: 0.73557163, Time: 0.0209 Steps: 70130, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002012, Sample Num: 32192, Cur Loss: 0.27027607, Cur Avg Loss: 0.17426670, Log Avg loss: 0.17351448, Global Avg Loss: 0.73549149, Time: 0.0209 Steps: 70140, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002022, Sample Num: 32352, Cur Loss: 0.12562908, Cur Avg Loss: 0.17395224, Log Avg loss: 0.11068129, Global Avg Loss: 0.73540243, Time: 0.0209 Steps: 70150, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002032, Sample Num: 32512, Cur Loss: 0.09596477, Cur Avg Loss: 0.17389194, Log Avg loss: 0.16170071, Global Avg Loss: 0.73532066, Time: 0.0209 Steps: 70160, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002042, Sample Num: 32672, Cur Loss: 0.32207823, Cur Avg Loss: 0.17406877, Log Avg loss: 0.21000057, Global Avg Loss: 0.73524579, Time: 0.0209 Steps: 70170, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002052, Sample Num: 32832, Cur Loss: 0.03942936, Cur Avg Loss: 0.17419667, Log Avg loss: 0.20031433, Global Avg Loss: 0.73516957, Time: 0.0247 Steps: 70180, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002062, Sample Num: 32992, Cur Loss: 0.31917477, Cur Avg Loss: 0.17413330, Log Avg loss: 0.16112926, Global Avg Loss: 0.73508779, Time: 0.0209 Steps: 70190, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002072, Sample Num: 33152, Cur Loss: 0.16983160, Cur Avg Loss: 0.17423548, Log Avg loss: 0.19530472, Global Avg Loss: 0.73501089, Time: 0.0209 Steps: 70200, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002082, Sample Num: 33312, Cur Loss: 0.08165845, Cur Avg Loss: 0.17421860, Log Avg loss: 0.17072082, Global Avg Loss: 0.73493052, Time: 0.0209 Steps: 70210, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002092, Sample Num: 33472, Cur Loss: 0.10495288, Cur Avg Loss: 0.17422531, Log Avg loss: 0.17562289, Global Avg Loss: 0.73485087, Time: 0.0210 Steps: 70220, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002102, Sample Num: 33632, Cur Loss: 0.18445833, Cur Avg Loss: 0.17395322, Log Avg loss: 0.11703200, Global Avg Loss: 0.73476290, Time: 0.0209 Steps: 70230, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002112, Sample Num: 33792, Cur Loss: 0.44411981, Cur Avg Loss: 0.17387686, Log Avg loss: 0.15782645, Global Avg Loss: 0.73468076, Time: 0.0210 Steps: 70240, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002122, Sample Num: 33952, Cur Loss: 0.05478169, Cur Avg Loss: 0.17399843, Log Avg loss: 0.19967306, Global Avg Loss: 0.73460460, Time: 0.0209 Steps: 70250, Updated lr: 0.000034 ***** Running evaluation checkpoint-70257 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-70257 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.618561, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.257756, "eval_total_loss": 181.202587, "eval_mae": 0.328322, "eval_mse": 0.257856, "eval_r2": 0.83609, "eval_sp_statistic": 0.925691, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.930896, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.177325, "test_total_loss": 89.017132, "test_mae": 0.295489, "test_mse": 0.177389, "test_r2": 0.885511, "test_sp_statistic": 0.918739, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.949079, "test_ps_pvalue": 0.0, "lr": 3.432242769084875e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7345465688282089, "train_cur_epoch_loss": 370.2894587153569, "train_cur_epoch_avg_loss": 0.17392647191890884, "train_cur_epoch_time": 44.618560791015625, "train_cur_epoch_avg_time": 0.020957520333966945, "epoch": 33, "step": 70257} ################################################## Training, Epoch: 0034, Batch: 000003, Sample Num: 48, Cur Loss: 0.02765389, Cur Avg Loss: 0.06075777, Log Avg loss: 0.12470639, Global Avg Loss: 0.73451780, Time: 0.0248 Steps: 70260, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000013, Sample Num: 208, Cur Loss: 0.09650163, Cur Avg Loss: 0.10025412, Log Avg loss: 0.11210302, Global Avg Loss: 0.73442922, Time: 0.0209 Steps: 70270, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000023, Sample Num: 368, Cur Loss: 0.13695838, Cur Avg Loss: 0.12676788, Log Avg loss: 0.16123577, Global Avg Loss: 0.73434767, Time: 0.0209 Steps: 70280, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000033, Sample Num: 528, Cur Loss: 0.09744190, Cur Avg Loss: 0.13392278, Log Avg loss: 0.15037907, Global Avg Loss: 0.73426459, Time: 0.0209 Steps: 70290, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000043, Sample Num: 688, Cur Loss: 0.12473988, Cur Avg Loss: 0.13152521, Log Avg loss: 0.12361321, Global Avg Loss: 0.73417772, Time: 0.0209 Steps: 70300, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000053, Sample Num: 848, Cur Loss: 0.08751304, Cur Avg Loss: 0.13421014, Log Avg loss: 0.14575535, Global Avg Loss: 0.73409403, Time: 0.0209 Steps: 70310, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000063, Sample Num: 1008, Cur Loss: 0.19422561, Cur Avg Loss: 0.13550248, Log Avg loss: 0.14235186, Global Avg Loss: 0.73400988, Time: 0.0210 Steps: 70320, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000073, Sample Num: 1168, Cur Loss: 0.15859725, Cur Avg Loss: 0.13515476, Log Avg loss: 0.13296412, Global Avg Loss: 0.73392442, Time: 0.0209 Steps: 70330, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000083, Sample Num: 1328, Cur Loss: 0.10188055, Cur Avg Loss: 0.13145852, Log Avg loss: 0.10447598, Global Avg Loss: 0.73383494, Time: 0.0210 Steps: 70340, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000093, Sample Num: 1488, Cur Loss: 0.32231334, Cur Avg Loss: 0.13144771, Log Avg loss: 0.13135801, Global Avg Loss: 0.73374930, Time: 0.0209 Steps: 70350, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000103, Sample Num: 1648, Cur Loss: 0.21234703, Cur Avg Loss: 0.13104176, Log Avg loss: 0.12726644, Global Avg Loss: 0.73366310, Time: 0.0210 Steps: 70360, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000113, Sample Num: 1808, Cur Loss: 0.15842012, Cur Avg Loss: 0.13965782, Log Avg loss: 0.22840319, Global Avg Loss: 0.73359130, Time: 0.0210 Steps: 70370, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000123, Sample Num: 1968, Cur Loss: 0.33295387, Cur Avg Loss: 0.14541150, Log Avg loss: 0.21042808, Global Avg Loss: 0.73351696, Time: 0.0210 Steps: 70380, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000133, Sample Num: 2128, Cur Loss: 0.21565990, Cur Avg Loss: 0.14363690, Log Avg loss: 0.12180940, Global Avg Loss: 0.73343006, Time: 0.0209 Steps: 70390, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000143, Sample Num: 2288, Cur Loss: 0.35127017, Cur Avg Loss: 0.14684053, Log Avg loss: 0.18944870, Global Avg Loss: 0.73335279, Time: 0.0209 Steps: 70400, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000153, Sample Num: 2448, Cur Loss: 0.09713234, Cur Avg Loss: 0.14916812, Log Avg loss: 0.18245280, Global Avg Loss: 0.73327455, Time: 0.0208 Steps: 70410, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000163, Sample Num: 2608, Cur Loss: 0.32063913, Cur Avg Loss: 0.15206888, Log Avg loss: 0.19645039, Global Avg Loss: 0.73319832, Time: 0.0208 Steps: 70420, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000173, Sample Num: 2768, Cur Loss: 0.06634857, Cur Avg Loss: 0.14806282, Log Avg loss: 0.08276413, Global Avg Loss: 0.73310597, Time: 0.0209 Steps: 70430, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000183, Sample Num: 2928, Cur Loss: 0.10728654, Cur Avg Loss: 0.14965222, Log Avg loss: 0.17714871, Global Avg Loss: 0.73302704, Time: 0.0210 Steps: 70440, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000193, Sample Num: 3088, Cur Loss: 0.27074009, Cur Avg Loss: 0.15244688, Log Avg loss: 0.20358917, Global Avg Loss: 0.73295189, Time: 0.0210 Steps: 70450, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000203, Sample Num: 3248, Cur Loss: 0.08472882, Cur Avg Loss: 0.15417705, Log Avg loss: 0.18756942, Global Avg Loss: 0.73287449, Time: 0.0210 Steps: 70460, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000213, Sample Num: 3408, Cur Loss: 0.16861266, Cur Avg Loss: 0.15726993, Log Avg loss: 0.22005543, Global Avg Loss: 0.73280171, Time: 0.0208 Steps: 70470, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000223, Sample Num: 3568, Cur Loss: 0.17476329, Cur Avg Loss: 0.15711280, Log Avg loss: 0.15376583, Global Avg Loss: 0.73271956, Time: 0.0208 Steps: 70480, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000233, Sample Num: 3728, Cur Loss: 0.21587008, Cur Avg Loss: 0.15773951, Log Avg loss: 0.17171523, Global Avg Loss: 0.73263997, Time: 0.0208 Steps: 70490, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000243, Sample Num: 3888, Cur Loss: 0.03175017, Cur Avg Loss: 0.15651498, Log Avg loss: 0.12798340, Global Avg Loss: 0.73255420, Time: 0.0209 Steps: 70500, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000253, Sample Num: 4048, Cur Loss: 0.07627411, Cur Avg Loss: 0.15584237, Log Avg loss: 0.13949786, Global Avg Loss: 0.73247010, Time: 0.0209 Steps: 70510, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000263, Sample Num: 4208, Cur Loss: 0.19869718, Cur Avg Loss: 0.15674328, Log Avg loss: 0.17953632, Global Avg Loss: 0.73239169, Time: 0.0210 Steps: 70520, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000273, Sample Num: 4368, Cur Loss: 0.14816916, Cur Avg Loss: 0.16134597, Log Avg loss: 0.28239667, Global Avg Loss: 0.73232789, Time: 0.0210 Steps: 70530, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000283, Sample Num: 4528, Cur Loss: 0.09323521, Cur Avg Loss: 0.16168186, Log Avg loss: 0.17085176, Global Avg Loss: 0.73224829, Time: 0.0210 Steps: 70540, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000293, Sample Num: 4688, Cur Loss: 0.09141642, Cur Avg Loss: 0.16021137, Log Avg loss: 0.11859641, Global Avg Loss: 0.73216131, Time: 0.0210 Steps: 70550, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000303, Sample Num: 4848, Cur Loss: 0.34464601, Cur Avg Loss: 0.16124024, Log Avg loss: 0.19138616, Global Avg Loss: 0.73208467, Time: 0.0210 Steps: 70560, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000313, Sample Num: 5008, Cur Loss: 0.71563935, Cur Avg Loss: 0.16255526, Log Avg loss: 0.20240055, Global Avg Loss: 0.73200961, Time: 0.0211 Steps: 70570, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000323, Sample Num: 5168, Cur Loss: 0.17396811, Cur Avg Loss: 0.16234282, Log Avg loss: 0.15569345, Global Avg Loss: 0.73192795, Time: 0.0211 Steps: 70580, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000333, Sample Num: 5328, Cur Loss: 0.13661171, Cur Avg Loss: 0.16060649, Log Avg loss: 0.10452285, Global Avg Loss: 0.73183907, Time: 0.0209 Steps: 70590, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000343, Sample Num: 5488, Cur Loss: 0.31651554, Cur Avg Loss: 0.16019375, Log Avg loss: 0.14644949, Global Avg Loss: 0.73175616, Time: 0.0209 Steps: 70600, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000353, Sample Num: 5648, Cur Loss: 0.10188636, Cur Avg Loss: 0.15919013, Log Avg loss: 0.12476593, Global Avg Loss: 0.73167019, Time: 0.0208 Steps: 70610, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000363, Sample Num: 5808, Cur Loss: 0.12799296, Cur Avg Loss: 0.15843047, Log Avg loss: 0.13161445, Global Avg Loss: 0.73158522, Time: 0.0209 Steps: 70620, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000373, Sample Num: 5968, Cur Loss: 0.21067199, Cur Avg Loss: 0.15669353, Log Avg loss: 0.09364270, Global Avg Loss: 0.73149490, Time: 0.0209 Steps: 70630, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000383, Sample Num: 6128, Cur Loss: 0.11356603, Cur Avg Loss: 0.15612333, Log Avg loss: 0.13485485, Global Avg Loss: 0.73141044, Time: 0.0209 Steps: 70640, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000393, Sample Num: 6288, Cur Loss: 0.19786125, Cur Avg Loss: 0.15575259, Log Avg loss: 0.14155321, Global Avg Loss: 0.73132695, Time: 0.0208 Steps: 70650, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000403, Sample Num: 6448, Cur Loss: 0.18770343, Cur Avg Loss: 0.15495495, Log Avg loss: 0.12360772, Global Avg Loss: 0.73124094, Time: 0.0208 Steps: 70660, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000413, Sample Num: 6608, Cur Loss: 0.20690629, Cur Avg Loss: 0.15473748, Log Avg loss: 0.14597339, Global Avg Loss: 0.73115813, Time: 0.0208 Steps: 70670, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000423, Sample Num: 6768, Cur Loss: 0.12180836, Cur Avg Loss: 0.15453831, Log Avg loss: 0.14631277, Global Avg Loss: 0.73107538, Time: 0.0209 Steps: 70680, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000433, Sample Num: 6928, Cur Loss: 0.17067125, Cur Avg Loss: 0.15505662, Log Avg loss: 0.17698124, Global Avg Loss: 0.73099700, Time: 0.0208 Steps: 70690, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000443, Sample Num: 7088, Cur Loss: 0.05664876, Cur Avg Loss: 0.15601388, Log Avg loss: 0.19746322, Global Avg Loss: 0.73092153, Time: 0.0209 Steps: 70700, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000453, Sample Num: 7248, Cur Loss: 0.25085473, Cur Avg Loss: 0.15601541, Log Avg loss: 0.15608284, Global Avg Loss: 0.73084024, Time: 0.0209 Steps: 70710, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000463, Sample Num: 7408, Cur Loss: 0.04535298, Cur Avg Loss: 0.15658002, Log Avg loss: 0.18215700, Global Avg Loss: 0.73076265, Time: 0.0208 Steps: 70720, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000473, Sample Num: 7568, Cur Loss: 0.08745773, Cur Avg Loss: 0.15798014, Log Avg loss: 0.22280591, Global Avg Loss: 0.73069084, Time: 0.0208 Steps: 70730, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000483, Sample Num: 7728, Cur Loss: 0.05867955, Cur Avg Loss: 0.15694123, Log Avg loss: 0.10780045, Global Avg Loss: 0.73060278, Time: 0.0208 Steps: 70740, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000493, Sample Num: 7888, Cur Loss: 0.15868479, Cur Avg Loss: 0.15733453, Log Avg loss: 0.17633135, Global Avg Loss: 0.73052444, Time: 0.0209 Steps: 70750, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000503, Sample Num: 8048, Cur Loss: 0.27654961, Cur Avg Loss: 0.15812434, Log Avg loss: 0.19706160, Global Avg Loss: 0.73044905, Time: 0.0208 Steps: 70760, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000513, Sample Num: 8208, Cur Loss: 0.24243358, Cur Avg Loss: 0.15813317, Log Avg loss: 0.15857741, Global Avg Loss: 0.73036824, Time: 0.0245 Steps: 70770, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000523, Sample Num: 8368, Cur Loss: 0.35219404, Cur Avg Loss: 0.15806694, Log Avg loss: 0.15466949, Global Avg Loss: 0.73028691, Time: 0.0208 Steps: 70780, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000533, Sample Num: 8528, Cur Loss: 0.18205905, Cur Avg Loss: 0.15770928, Log Avg loss: 0.13900347, Global Avg Loss: 0.73020338, Time: 0.0208 Steps: 70790, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000543, Sample Num: 8688, Cur Loss: 0.09813219, Cur Avg Loss: 0.15728820, Log Avg loss: 0.13484469, Global Avg Loss: 0.73011929, Time: 0.0208 Steps: 70800, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000553, Sample Num: 8848, Cur Loss: 0.31898707, Cur Avg Loss: 0.15823929, Log Avg loss: 0.20988373, Global Avg Loss: 0.73004582, Time: 0.0209 Steps: 70810, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000563, Sample Num: 9008, Cur Loss: 0.60670853, Cur Avg Loss: 0.15832021, Log Avg loss: 0.16279483, Global Avg Loss: 0.72996572, Time: 0.0208 Steps: 70820, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000573, Sample Num: 9168, Cur Loss: 0.05006261, Cur Avg Loss: 0.15851110, Log Avg loss: 0.16925801, Global Avg Loss: 0.72988656, Time: 0.0208 Steps: 70830, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000583, Sample Num: 9328, Cur Loss: 0.16186103, Cur Avg Loss: 0.15926695, Log Avg loss: 0.20257735, Global Avg Loss: 0.72981212, Time: 0.0209 Steps: 70840, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000593, Sample Num: 9488, Cur Loss: 0.33342916, Cur Avg Loss: 0.15972677, Log Avg loss: 0.18653423, Global Avg Loss: 0.72973544, Time: 0.0208 Steps: 70850, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000603, Sample Num: 9648, Cur Loss: 0.22337998, Cur Avg Loss: 0.15892720, Log Avg loss: 0.11151300, Global Avg Loss: 0.72964820, Time: 0.0209 Steps: 70860, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000613, Sample Num: 9808, Cur Loss: 0.12229027, Cur Avg Loss: 0.15843191, Log Avg loss: 0.12856559, Global Avg Loss: 0.72956338, Time: 0.0209 Steps: 70870, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000623, Sample Num: 9968, Cur Loss: 0.16781083, Cur Avg Loss: 0.15885507, Log Avg loss: 0.18479505, Global Avg Loss: 0.72948653, Time: 0.0209 Steps: 70880, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000633, Sample Num: 10128, Cur Loss: 0.09076873, Cur Avg Loss: 0.16004546, Log Avg loss: 0.23420675, Global Avg Loss: 0.72941666, Time: 0.0208 Steps: 70890, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000643, Sample Num: 10288, Cur Loss: 0.10983101, Cur Avg Loss: 0.15987114, Log Avg loss: 0.14883662, Global Avg Loss: 0.72933477, Time: 0.0209 Steps: 70900, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000653, Sample Num: 10448, Cur Loss: 0.10009205, Cur Avg Loss: 0.15925711, Log Avg loss: 0.11977483, Global Avg Loss: 0.72924881, Time: 0.0208 Steps: 70910, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000663, Sample Num: 10608, Cur Loss: 0.22423059, Cur Avg Loss: 0.15952549, Log Avg loss: 0.17705081, Global Avg Loss: 0.72917095, Time: 0.0208 Steps: 70920, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000673, Sample Num: 10768, Cur Loss: 0.12031890, Cur Avg Loss: 0.15935745, Log Avg loss: 0.14821606, Global Avg Loss: 0.72908904, Time: 0.0209 Steps: 70930, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000683, Sample Num: 10928, Cur Loss: 0.05190772, Cur Avg Loss: 0.15871237, Log Avg loss: 0.11529888, Global Avg Loss: 0.72900252, Time: 0.0209 Steps: 70940, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000693, Sample Num: 11088, Cur Loss: 0.12557724, Cur Avg Loss: 0.15850083, Log Avg loss: 0.14405258, Global Avg Loss: 0.72892008, Time: 0.0209 Steps: 70950, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000703, Sample Num: 11248, Cur Loss: 0.17174785, Cur Avg Loss: 0.15873864, Log Avg loss: 0.17521917, Global Avg Loss: 0.72884205, Time: 0.0209 Steps: 70960, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000713, Sample Num: 11408, Cur Loss: 0.36189419, Cur Avg Loss: 0.15966895, Log Avg loss: 0.22506914, Global Avg Loss: 0.72877106, Time: 0.0209 Steps: 70970, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000723, Sample Num: 11568, Cur Loss: 0.14833263, Cur Avg Loss: 0.16007003, Log Avg loss: 0.18866726, Global Avg Loss: 0.72869497, Time: 0.0209 Steps: 70980, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000733, Sample Num: 11728, Cur Loss: 0.09851375, Cur Avg Loss: 0.15962569, Log Avg loss: 0.12749994, Global Avg Loss: 0.72861028, Time: 0.0208 Steps: 70990, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000743, Sample Num: 11888, Cur Loss: 0.10884278, Cur Avg Loss: 0.16044008, Log Avg loss: 0.22013462, Global Avg Loss: 0.72853867, Time: 0.0209 Steps: 71000, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000753, Sample Num: 12048, Cur Loss: 0.04444015, Cur Avg Loss: 0.16041074, Log Avg loss: 0.15823141, Global Avg Loss: 0.72845835, Time: 0.0208 Steps: 71010, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000763, Sample Num: 12208, Cur Loss: 0.28791690, Cur Avg Loss: 0.16032091, Log Avg loss: 0.15355613, Global Avg Loss: 0.72837740, Time: 0.0209 Steps: 71020, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000773, Sample Num: 12368, Cur Loss: 0.06162813, Cur Avg Loss: 0.15967479, Log Avg loss: 0.11037602, Global Avg Loss: 0.72829040, Time: 0.0246 Steps: 71030, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000783, Sample Num: 12528, Cur Loss: 0.26414704, Cur Avg Loss: 0.15946484, Log Avg loss: 0.14323586, Global Avg Loss: 0.72820804, Time: 0.0209 Steps: 71040, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000793, Sample Num: 12688, Cur Loss: 0.19592744, Cur Avg Loss: 0.15924297, Log Avg loss: 0.14187003, Global Avg Loss: 0.72812552, Time: 0.0209 Steps: 71050, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000803, Sample Num: 12848, Cur Loss: 0.14882679, Cur Avg Loss: 0.15930173, Log Avg loss: 0.16396140, Global Avg Loss: 0.72804612, Time: 0.0209 Steps: 71060, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000813, Sample Num: 13008, Cur Loss: 0.10132615, Cur Avg Loss: 0.15973086, Log Avg loss: 0.19419025, Global Avg Loss: 0.72797101, Time: 0.0209 Steps: 71070, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000823, Sample Num: 13168, Cur Loss: 0.16997789, Cur Avg Loss: 0.16075873, Log Avg loss: 0.24432434, Global Avg Loss: 0.72790296, Time: 0.0209 Steps: 71080, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000833, Sample Num: 13328, Cur Loss: 0.04090241, Cur Avg Loss: 0.16030835, Log Avg loss: 0.12324208, Global Avg Loss: 0.72781791, Time: 0.0208 Steps: 71090, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000843, Sample Num: 13488, Cur Loss: 0.04350512, Cur Avg Loss: 0.15956411, Log Avg loss: 0.09756893, Global Avg Loss: 0.72772927, Time: 0.0210 Steps: 71100, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000853, Sample Num: 13648, Cur Loss: 0.17159276, Cur Avg Loss: 0.15951701, Log Avg loss: 0.15554706, Global Avg Loss: 0.72764880, Time: 0.0209 Steps: 71110, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000863, Sample Num: 13808, Cur Loss: 0.05777566, Cur Avg Loss: 0.15913249, Log Avg loss: 0.12633275, Global Avg Loss: 0.72756425, Time: 0.0208 Steps: 71120, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000873, Sample Num: 13968, Cur Loss: 0.10355698, Cur Avg Loss: 0.15883022, Log Avg loss: 0.13274444, Global Avg Loss: 0.72748063, Time: 0.0209 Steps: 71130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000883, Sample Num: 14128, Cur Loss: 0.25529116, Cur Avg Loss: 0.15880403, Log Avg loss: 0.15651751, Global Avg Loss: 0.72740037, Time: 0.0208 Steps: 71140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000893, Sample Num: 14288, Cur Loss: 0.18878993, Cur Avg Loss: 0.15882730, Log Avg loss: 0.16088164, Global Avg Loss: 0.72732075, Time: 0.0208 Steps: 71150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000903, Sample Num: 14448, Cur Loss: 0.26184586, Cur Avg Loss: 0.15886907, Log Avg loss: 0.16259958, Global Avg Loss: 0.72724139, Time: 0.0209 Steps: 71160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000913, Sample Num: 14608, Cur Loss: 0.29685640, Cur Avg Loss: 0.15926253, Log Avg loss: 0.19479173, Global Avg Loss: 0.72716657, Time: 0.0208 Steps: 71170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000923, Sample Num: 14768, Cur Loss: 0.13019632, Cur Avg Loss: 0.15877097, Log Avg loss: 0.11389150, Global Avg Loss: 0.72708041, Time: 0.0208 Steps: 71180, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000933, Sample Num: 14928, Cur Loss: 0.28598440, Cur Avg Loss: 0.16005370, Log Avg loss: 0.27844940, Global Avg Loss: 0.72701740, Time: 0.0208 Steps: 71190, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000943, Sample Num: 15088, Cur Loss: 0.09748396, Cur Avg Loss: 0.16056649, Log Avg loss: 0.20841015, Global Avg Loss: 0.72694456, Time: 0.0208 Steps: 71200, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000953, Sample Num: 15248, Cur Loss: 0.05023153, Cur Avg Loss: 0.16058476, Log Avg loss: 0.16230797, Global Avg Loss: 0.72686527, Time: 0.0208 Steps: 71210, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000963, Sample Num: 15408, Cur Loss: 0.08498041, Cur Avg Loss: 0.16018701, Log Avg loss: 0.12228125, Global Avg Loss: 0.72678038, Time: 0.0208 Steps: 71220, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000973, Sample Num: 15568, Cur Loss: 0.17789993, Cur Avg Loss: 0.16034711, Log Avg loss: 0.17576476, Global Avg Loss: 0.72670302, Time: 0.0208 Steps: 71230, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000983, Sample Num: 15728, Cur Loss: 0.09112190, Cur Avg Loss: 0.16065379, Log Avg loss: 0.19049362, Global Avg Loss: 0.72662775, Time: 0.0209 Steps: 71240, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000993, Sample Num: 15888, Cur Loss: 0.34281889, Cur Avg Loss: 0.16082332, Log Avg loss: 0.17748864, Global Avg Loss: 0.72655068, Time: 0.0208 Steps: 71250, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001003, Sample Num: 16048, Cur Loss: 0.15081179, Cur Avg Loss: 0.16130418, Log Avg loss: 0.20905308, Global Avg Loss: 0.72647806, Time: 0.0208 Steps: 71260, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001013, Sample Num: 16208, Cur Loss: 0.20372382, Cur Avg Loss: 0.16114195, Log Avg loss: 0.14487074, Global Avg Loss: 0.72639645, Time: 0.0208 Steps: 71270, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001023, Sample Num: 16368, Cur Loss: 0.26412228, Cur Avg Loss: 0.16103720, Log Avg loss: 0.15042604, Global Avg Loss: 0.72631565, Time: 0.0208 Steps: 71280, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001033, Sample Num: 16528, Cur Loss: 0.08734530, Cur Avg Loss: 0.16100759, Log Avg loss: 0.15797841, Global Avg Loss: 0.72623593, Time: 0.0208 Steps: 71290, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001043, Sample Num: 16688, Cur Loss: 0.33770210, Cur Avg Loss: 0.16099380, Log Avg loss: 0.15956919, Global Avg Loss: 0.72615645, Time: 0.0209 Steps: 71300, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001053, Sample Num: 16848, Cur Loss: 0.11492085, Cur Avg Loss: 0.16127550, Log Avg loss: 0.19065701, Global Avg Loss: 0.72608135, Time: 0.0208 Steps: 71310, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001063, Sample Num: 17008, Cur Loss: 0.17074262, Cur Avg Loss: 0.16139951, Log Avg loss: 0.17445699, Global Avg Loss: 0.72600401, Time: 0.0208 Steps: 71320, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001073, Sample Num: 17168, Cur Loss: 0.14101537, Cur Avg Loss: 0.16131588, Log Avg loss: 0.15242621, Global Avg Loss: 0.72592360, Time: 0.0208 Steps: 71330, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001083, Sample Num: 17328, Cur Loss: 0.12747793, Cur Avg Loss: 0.16194836, Log Avg loss: 0.22981412, Global Avg Loss: 0.72585406, Time: 0.0209 Steps: 71340, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001093, Sample Num: 17488, Cur Loss: 0.16109914, Cur Avg Loss: 0.16214144, Log Avg loss: 0.18305123, Global Avg Loss: 0.72577798, Time: 0.0208 Steps: 71350, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001103, Sample Num: 17648, Cur Loss: 0.15385598, Cur Avg Loss: 0.16183185, Log Avg loss: 0.12799356, Global Avg Loss: 0.72569421, Time: 0.0208 Steps: 71360, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001113, Sample Num: 17808, Cur Loss: 0.38262984, Cur Avg Loss: 0.16155163, Log Avg loss: 0.13064422, Global Avg Loss: 0.72561083, Time: 0.0208 Steps: 71370, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001123, Sample Num: 17968, Cur Loss: 0.17432816, Cur Avg Loss: 0.16137636, Log Avg loss: 0.14186877, Global Avg Loss: 0.72552905, Time: 0.0208 Steps: 71380, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001133, Sample Num: 18128, Cur Loss: 0.08263816, Cur Avg Loss: 0.16115325, Log Avg loss: 0.13609799, Global Avg Loss: 0.72544649, Time: 0.0209 Steps: 71390, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001143, Sample Num: 18288, Cur Loss: 0.33215714, Cur Avg Loss: 0.16199385, Log Avg loss: 0.25723293, Global Avg Loss: 0.72538091, Time: 0.0209 Steps: 71400, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001153, Sample Num: 18448, Cur Loss: 0.13118815, Cur Avg Loss: 0.16230897, Log Avg loss: 0.19832740, Global Avg Loss: 0.72530711, Time: 0.0209 Steps: 71410, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001163, Sample Num: 18608, Cur Loss: 0.02964137, Cur Avg Loss: 0.16216932, Log Avg loss: 0.14606761, Global Avg Loss: 0.72522600, Time: 0.0209 Steps: 71420, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001173, Sample Num: 18768, Cur Loss: 0.21242537, Cur Avg Loss: 0.16211528, Log Avg loss: 0.15583080, Global Avg Loss: 0.72514629, Time: 0.0209 Steps: 71430, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001183, Sample Num: 18928, Cur Loss: 0.12424649, Cur Avg Loss: 0.16175425, Log Avg loss: 0.11940518, Global Avg Loss: 0.72506150, Time: 0.0208 Steps: 71440, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001193, Sample Num: 19088, Cur Loss: 0.22077945, Cur Avg Loss: 0.16176145, Log Avg loss: 0.16261343, Global Avg Loss: 0.72498278, Time: 0.0208 Steps: 71450, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001203, Sample Num: 19248, Cur Loss: 0.10931678, Cur Avg Loss: 0.16184272, Log Avg loss: 0.17153794, Global Avg Loss: 0.72490533, Time: 0.0209 Steps: 71460, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001213, Sample Num: 19408, Cur Loss: 0.08378977, Cur Avg Loss: 0.16155443, Log Avg loss: 0.12687394, Global Avg Loss: 0.72482166, Time: 0.0209 Steps: 71470, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001223, Sample Num: 19568, Cur Loss: 0.21818206, Cur Avg Loss: 0.16147112, Log Avg loss: 0.15136469, Global Avg Loss: 0.72474143, Time: 0.0208 Steps: 71480, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001233, Sample Num: 19728, Cur Loss: 0.19323628, Cur Avg Loss: 0.16199837, Log Avg loss: 0.22648101, Global Avg Loss: 0.72467173, Time: 0.0210 Steps: 71490, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001243, Sample Num: 19888, Cur Loss: 0.26233137, Cur Avg Loss: 0.16251861, Log Avg loss: 0.22666475, Global Avg Loss: 0.72460208, Time: 0.0208 Steps: 71500, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001253, Sample Num: 20048, Cur Loss: 0.10384057, Cur Avg Loss: 0.16220997, Log Avg loss: 0.12384647, Global Avg Loss: 0.72451807, Time: 0.0208 Steps: 71510, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001263, Sample Num: 20208, Cur Loss: 0.16302602, Cur Avg Loss: 0.16209175, Log Avg loss: 0.14727761, Global Avg Loss: 0.72443736, Time: 0.0208 Steps: 71520, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001273, Sample Num: 20368, Cur Loss: 0.13624182, Cur Avg Loss: 0.16194359, Log Avg loss: 0.14323131, Global Avg Loss: 0.72435611, Time: 0.0208 Steps: 71530, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001283, Sample Num: 20528, Cur Loss: 0.05373595, Cur Avg Loss: 0.16215712, Log Avg loss: 0.18934030, Global Avg Loss: 0.72428132, Time: 0.0246 Steps: 71540, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001293, Sample Num: 20688, Cur Loss: 0.05493443, Cur Avg Loss: 0.16233248, Log Avg loss: 0.18483073, Global Avg Loss: 0.72420593, Time: 0.0208 Steps: 71550, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001303, Sample Num: 20848, Cur Loss: 0.03073012, Cur Avg Loss: 0.16245133, Log Avg loss: 0.17781903, Global Avg Loss: 0.72412957, Time: 0.0208 Steps: 71560, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001313, Sample Num: 21008, Cur Loss: 0.08568586, Cur Avg Loss: 0.16291082, Log Avg loss: 0.22278177, Global Avg Loss: 0.72405952, Time: 0.0208 Steps: 71570, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001323, Sample Num: 21168, Cur Loss: 0.45056403, Cur Avg Loss: 0.16310059, Log Avg loss: 0.18801774, Global Avg Loss: 0.72398464, Time: 0.0208 Steps: 71580, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001333, Sample Num: 21328, Cur Loss: 0.14466512, Cur Avg Loss: 0.16330713, Log Avg loss: 0.19063154, Global Avg Loss: 0.72391014, Time: 0.0208 Steps: 71590, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001343, Sample Num: 21488, Cur Loss: 0.07201634, Cur Avg Loss: 0.16344217, Log Avg loss: 0.18144330, Global Avg Loss: 0.72383437, Time: 0.0208 Steps: 71600, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001353, Sample Num: 21648, Cur Loss: 0.09466236, Cur Avg Loss: 0.16336901, Log Avg loss: 0.15354357, Global Avg Loss: 0.72375473, Time: 0.0208 Steps: 71610, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001363, Sample Num: 21808, Cur Loss: 0.07267778, Cur Avg Loss: 0.16386257, Log Avg loss: 0.23064170, Global Avg Loss: 0.72368588, Time: 0.0208 Steps: 71620, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001373, Sample Num: 21968, Cur Loss: 0.22843352, Cur Avg Loss: 0.16411787, Log Avg loss: 0.19891562, Global Avg Loss: 0.72361262, Time: 0.0208 Steps: 71630, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001383, Sample Num: 22128, Cur Loss: 0.09493959, Cur Avg Loss: 0.16419193, Log Avg loss: 0.17435922, Global Avg Loss: 0.72353595, Time: 0.0208 Steps: 71640, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001393, Sample Num: 22288, Cur Loss: 0.45457956, Cur Avg Loss: 0.16455467, Log Avg loss: 0.21472252, Global Avg Loss: 0.72346494, Time: 0.0208 Steps: 71650, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001403, Sample Num: 22448, Cur Loss: 0.26826018, Cur Avg Loss: 0.16471097, Log Avg loss: 0.18648360, Global Avg Loss: 0.72339001, Time: 0.0208 Steps: 71660, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001413, Sample Num: 22608, Cur Loss: 0.09028787, Cur Avg Loss: 0.16492327, Log Avg loss: 0.19470840, Global Avg Loss: 0.72331624, Time: 0.0208 Steps: 71670, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001423, Sample Num: 22768, Cur Loss: 0.21613201, Cur Avg Loss: 0.16476517, Log Avg loss: 0.14242636, Global Avg Loss: 0.72323520, Time: 0.0208 Steps: 71680, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001433, Sample Num: 22928, Cur Loss: 0.20971894, Cur Avg Loss: 0.16460283, Log Avg loss: 0.14150100, Global Avg Loss: 0.72315405, Time: 0.0208 Steps: 71690, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001443, Sample Num: 23088, Cur Loss: 0.02918117, Cur Avg Loss: 0.16471901, Log Avg loss: 0.18136761, Global Avg Loss: 0.72307849, Time: 0.0208 Steps: 71700, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001453, Sample Num: 23248, Cur Loss: 0.20575581, Cur Avg Loss: 0.16531382, Log Avg loss: 0.25114519, Global Avg Loss: 0.72301268, Time: 0.0208 Steps: 71710, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001463, Sample Num: 23408, Cur Loss: 0.16234592, Cur Avg Loss: 0.16523968, Log Avg loss: 0.15446708, Global Avg Loss: 0.72293341, Time: 0.0208 Steps: 71720, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001473, Sample Num: 23568, Cur Loss: 0.09203079, Cur Avg Loss: 0.16492234, Log Avg loss: 0.11849564, Global Avg Loss: 0.72284914, Time: 0.0208 Steps: 71730, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001483, Sample Num: 23728, Cur Loss: 0.09033172, Cur Avg Loss: 0.16442080, Log Avg loss: 0.09054449, Global Avg Loss: 0.72276100, Time: 0.0209 Steps: 71740, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001493, Sample Num: 23888, Cur Loss: 0.16718039, Cur Avg Loss: 0.16460339, Log Avg loss: 0.19168059, Global Avg Loss: 0.72268698, Time: 0.0208 Steps: 71750, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001503, Sample Num: 24048, Cur Loss: 0.21434951, Cur Avg Loss: 0.16480466, Log Avg loss: 0.19485447, Global Avg Loss: 0.72261343, Time: 0.0208 Steps: 71760, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001513, Sample Num: 24208, Cur Loss: 0.25041032, Cur Avg Loss: 0.16481202, Log Avg loss: 0.16591844, Global Avg Loss: 0.72253586, Time: 0.0208 Steps: 71770, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001523, Sample Num: 24368, Cur Loss: 0.07794327, Cur Avg Loss: 0.16470956, Log Avg loss: 0.14920757, Global Avg Loss: 0.72245599, Time: 0.0208 Steps: 71780, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001533, Sample Num: 24528, Cur Loss: 0.37869716, Cur Avg Loss: 0.16483260, Log Avg loss: 0.18357069, Global Avg Loss: 0.72238093, Time: 0.0208 Steps: 71790, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001543, Sample Num: 24688, Cur Loss: 0.11144085, Cur Avg Loss: 0.16458150, Log Avg loss: 0.12608812, Global Avg Loss: 0.72229788, Time: 0.0210 Steps: 71800, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001553, Sample Num: 24848, Cur Loss: 0.20275095, Cur Avg Loss: 0.16446331, Log Avg loss: 0.14622646, Global Avg Loss: 0.72221765, Time: 0.0210 Steps: 71810, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001563, Sample Num: 25008, Cur Loss: 0.21673916, Cur Avg Loss: 0.16430517, Log Avg loss: 0.13974583, Global Avg Loss: 0.72213655, Time: 0.0211 Steps: 71820, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001573, Sample Num: 25168, Cur Loss: 0.14460006, Cur Avg Loss: 0.16467828, Log Avg loss: 0.22299547, Global Avg Loss: 0.72206706, Time: 0.0210 Steps: 71830, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001583, Sample Num: 25328, Cur Loss: 0.23330560, Cur Avg Loss: 0.16481731, Log Avg loss: 0.18668728, Global Avg Loss: 0.72199254, Time: 0.0210 Steps: 71840, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001593, Sample Num: 25488, Cur Loss: 0.15010406, Cur Avg Loss: 0.16466139, Log Avg loss: 0.13997891, Global Avg Loss: 0.72191154, Time: 0.0210 Steps: 71850, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001603, Sample Num: 25648, Cur Loss: 0.09122852, Cur Avg Loss: 0.16453421, Log Avg loss: 0.14427453, Global Avg Loss: 0.72183115, Time: 0.0210 Steps: 71860, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001613, Sample Num: 25808, Cur Loss: 0.03118362, Cur Avg Loss: 0.16419543, Log Avg loss: 0.10988861, Global Avg Loss: 0.72174601, Time: 0.0210 Steps: 71870, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001623, Sample Num: 25968, Cur Loss: 0.38963577, Cur Avg Loss: 0.16431638, Log Avg loss: 0.18382635, Global Avg Loss: 0.72167117, Time: 0.0210 Steps: 71880, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001633, Sample Num: 26128, Cur Loss: 0.29348296, Cur Avg Loss: 0.16467367, Log Avg loss: 0.22266243, Global Avg Loss: 0.72160176, Time: 0.0210 Steps: 71890, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001643, Sample Num: 26288, Cur Loss: 0.11036578, Cur Avg Loss: 0.16470221, Log Avg loss: 0.16936150, Global Avg Loss: 0.72152495, Time: 0.0210 Steps: 71900, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001653, Sample Num: 26448, Cur Loss: 0.11052995, Cur Avg Loss: 0.16460867, Log Avg loss: 0.14924064, Global Avg Loss: 0.72144537, Time: 0.0210 Steps: 71910, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001663, Sample Num: 26608, Cur Loss: 0.22699407, Cur Avg Loss: 0.16431701, Log Avg loss: 0.11610503, Global Avg Loss: 0.72136120, Time: 0.0210 Steps: 71920, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001673, Sample Num: 26768, Cur Loss: 0.04266800, Cur Avg Loss: 0.16424465, Log Avg loss: 0.15221197, Global Avg Loss: 0.72128207, Time: 0.0210 Steps: 71930, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001683, Sample Num: 26928, Cur Loss: 0.09347583, Cur Avg Loss: 0.16406891, Log Avg loss: 0.13466780, Global Avg Loss: 0.72120053, Time: 0.0210 Steps: 71940, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001693, Sample Num: 27088, Cur Loss: 0.14914230, Cur Avg Loss: 0.16404584, Log Avg loss: 0.16016266, Global Avg Loss: 0.72112256, Time: 0.0210 Steps: 71950, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001703, Sample Num: 27248, Cur Loss: 0.29711396, Cur Avg Loss: 0.16406974, Log Avg loss: 0.16811609, Global Avg Loss: 0.72104571, Time: 0.0210 Steps: 71960, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001713, Sample Num: 27408, Cur Loss: 0.03349148, Cur Avg Loss: 0.16414219, Log Avg loss: 0.17648080, Global Avg Loss: 0.72097004, Time: 0.0210 Steps: 71970, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001723, Sample Num: 27568, Cur Loss: 0.08232279, Cur Avg Loss: 0.16409962, Log Avg loss: 0.15680708, Global Avg Loss: 0.72089166, Time: 0.0210 Steps: 71980, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001733, Sample Num: 27728, Cur Loss: 0.68257236, Cur Avg Loss: 0.16438557, Log Avg loss: 0.21365522, Global Avg Loss: 0.72082120, Time: 0.0210 Steps: 71990, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001743, Sample Num: 27888, Cur Loss: 0.10610536, Cur Avg Loss: 0.16409520, Log Avg loss: 0.11377384, Global Avg Loss: 0.72073689, Time: 0.0211 Steps: 72000, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001753, Sample Num: 28048, Cur Loss: 0.28373235, Cur Avg Loss: 0.16405901, Log Avg loss: 0.15775105, Global Avg Loss: 0.72065871, Time: 0.0210 Steps: 72010, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001763, Sample Num: 28208, Cur Loss: 0.09506024, Cur Avg Loss: 0.16400772, Log Avg loss: 0.15501632, Global Avg Loss: 0.72058017, Time: 0.0210 Steps: 72020, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001773, Sample Num: 28368, Cur Loss: 0.06784415, Cur Avg Loss: 0.16395801, Log Avg loss: 0.15519356, Global Avg Loss: 0.72050168, Time: 0.0210 Steps: 72030, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001783, Sample Num: 28528, Cur Loss: 0.21597587, Cur Avg Loss: 0.16436785, Log Avg loss: 0.23703261, Global Avg Loss: 0.72043457, Time: 0.0210 Steps: 72040, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001793, Sample Num: 28688, Cur Loss: 0.09986177, Cur Avg Loss: 0.16450105, Log Avg loss: 0.18825099, Global Avg Loss: 0.72036070, Time: 0.0248 Steps: 72050, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001803, Sample Num: 28848, Cur Loss: 0.29701328, Cur Avg Loss: 0.16470902, Log Avg loss: 0.20199743, Global Avg Loss: 0.72028877, Time: 0.0210 Steps: 72060, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001813, Sample Num: 29008, Cur Loss: 0.28570384, Cur Avg Loss: 0.16527119, Log Avg loss: 0.26663047, Global Avg Loss: 0.72022582, Time: 0.0210 Steps: 72070, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001823, Sample Num: 29168, Cur Loss: 0.13832918, Cur Avg Loss: 0.16548659, Log Avg loss: 0.20453922, Global Avg Loss: 0.72015428, Time: 0.0210 Steps: 72080, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001833, Sample Num: 29328, Cur Loss: 0.27558914, Cur Avg Loss: 0.16559705, Log Avg loss: 0.18573467, Global Avg Loss: 0.72008015, Time: 0.0210 Steps: 72090, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001843, Sample Num: 29488, Cur Loss: 0.24339920, Cur Avg Loss: 0.16536033, Log Avg loss: 0.12196883, Global Avg Loss: 0.71999719, Time: 0.0210 Steps: 72100, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001853, Sample Num: 29648, Cur Loss: 0.15352054, Cur Avg Loss: 0.16542637, Log Avg loss: 0.17759709, Global Avg Loss: 0.71992197, Time: 0.0210 Steps: 72110, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001863, Sample Num: 29808, Cur Loss: 0.10714649, Cur Avg Loss: 0.16544190, Log Avg loss: 0.16832047, Global Avg Loss: 0.71984549, Time: 0.0210 Steps: 72120, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001873, Sample Num: 29968, Cur Loss: 0.06556787, Cur Avg Loss: 0.16533605, Log Avg loss: 0.14561599, Global Avg Loss: 0.71976588, Time: 0.0210 Steps: 72130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001883, Sample Num: 30128, Cur Loss: 0.07407505, Cur Avg Loss: 0.16551515, Log Avg loss: 0.19906118, Global Avg Loss: 0.71969370, Time: 0.0211 Steps: 72140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001893, Sample Num: 30288, Cur Loss: 0.19349632, Cur Avg Loss: 0.16532914, Log Avg loss: 0.13030187, Global Avg Loss: 0.71961201, Time: 0.0210 Steps: 72150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001903, Sample Num: 30448, Cur Loss: 0.15750907, Cur Avg Loss: 0.16500547, Log Avg loss: 0.10373477, Global Avg Loss: 0.71952666, Time: 0.0210 Steps: 72160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001913, Sample Num: 30608, Cur Loss: 0.02098109, Cur Avg Loss: 0.16518888, Log Avg loss: 0.20009358, Global Avg Loss: 0.71945469, Time: 0.0210 Steps: 72170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001923, Sample Num: 30768, Cur Loss: 0.18712042, Cur Avg Loss: 0.16536828, Log Avg loss: 0.19968715, Global Avg Loss: 0.71938268, Time: 0.0210 Steps: 72180, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001933, Sample Num: 30928, Cur Loss: 0.05328247, Cur Avg Loss: 0.16556528, Log Avg loss: 0.20344714, Global Avg Loss: 0.71931121, Time: 0.0210 Steps: 72190, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001943, Sample Num: 31088, Cur Loss: 0.02366932, Cur Avg Loss: 0.16577703, Log Avg loss: 0.20670809, Global Avg Loss: 0.71924021, Time: 0.0210 Steps: 72200, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001953, Sample Num: 31248, Cur Loss: 0.30424213, Cur Avg Loss: 0.16552902, Log Avg loss: 0.11734095, Global Avg Loss: 0.71915685, Time: 0.0210 Steps: 72210, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001963, Sample Num: 31408, Cur Loss: 0.09953003, Cur Avg Loss: 0.16523593, Log Avg loss: 0.10799593, Global Avg Loss: 0.71907223, Time: 0.0211 Steps: 72220, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001973, Sample Num: 31568, Cur Loss: 0.25130534, Cur Avg Loss: 0.16530771, Log Avg loss: 0.17939892, Global Avg Loss: 0.71899751, Time: 0.0210 Steps: 72230, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001983, Sample Num: 31728, Cur Loss: 0.06212762, Cur Avg Loss: 0.16548973, Log Avg loss: 0.20140207, Global Avg Loss: 0.71892586, Time: 0.0210 Steps: 72240, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001993, Sample Num: 31888, Cur Loss: 0.19490203, Cur Avg Loss: 0.16574848, Log Avg loss: 0.21705825, Global Avg Loss: 0.71885640, Time: 0.0210 Steps: 72250, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002003, Sample Num: 32048, Cur Loss: 0.38083306, Cur Avg Loss: 0.16575458, Log Avg loss: 0.16696937, Global Avg Loss: 0.71878003, Time: 0.0210 Steps: 72260, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002013, Sample Num: 32208, Cur Loss: 0.03838116, Cur Avg Loss: 0.16572520, Log Avg loss: 0.15984124, Global Avg Loss: 0.71870269, Time: 0.0210 Steps: 72270, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002023, Sample Num: 32368, Cur Loss: 0.14318323, Cur Avg Loss: 0.16558500, Log Avg loss: 0.13736256, Global Avg Loss: 0.71862226, Time: 0.0210 Steps: 72280, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002033, Sample Num: 32528, Cur Loss: 0.26600391, Cur Avg Loss: 0.16574935, Log Avg loss: 0.19899812, Global Avg Loss: 0.71855038, Time: 0.0210 Steps: 72290, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002043, Sample Num: 32688, Cur Loss: 0.34153599, Cur Avg Loss: 0.16601149, Log Avg loss: 0.21930364, Global Avg Loss: 0.71848132, Time: 0.0210 Steps: 72300, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002053, Sample Num: 32848, Cur Loss: 0.07049927, Cur Avg Loss: 0.16596424, Log Avg loss: 0.15631126, Global Avg Loss: 0.71840358, Time: 0.0233 Steps: 72310, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002063, Sample Num: 33008, Cur Loss: 0.28784063, Cur Avg Loss: 0.16584734, Log Avg loss: 0.14184800, Global Avg Loss: 0.71832386, Time: 0.0210 Steps: 72320, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002073, Sample Num: 33168, Cur Loss: 0.09687003, Cur Avg Loss: 0.16565748, Log Avg loss: 0.12648858, Global Avg Loss: 0.71824203, Time: 0.0210 Steps: 72330, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002083, Sample Num: 33328, Cur Loss: 0.10897103, Cur Avg Loss: 0.16565807, Log Avg loss: 0.16578057, Global Avg Loss: 0.71816566, Time: 0.0211 Steps: 72340, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002093, Sample Num: 33488, Cur Loss: 0.29676524, Cur Avg Loss: 0.16586107, Log Avg loss: 0.20814556, Global Avg Loss: 0.71809517, Time: 0.0211 Steps: 72350, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002103, Sample Num: 33648, Cur Loss: 0.15130000, Cur Avg Loss: 0.16568163, Log Avg loss: 0.12812649, Global Avg Loss: 0.71801364, Time: 0.0210 Steps: 72360, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002113, Sample Num: 33808, Cur Loss: 0.21418688, Cur Avg Loss: 0.16548201, Log Avg loss: 0.12350111, Global Avg Loss: 0.71793149, Time: 0.0210 Steps: 72370, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002123, Sample Num: 33968, Cur Loss: 0.11478771, Cur Avg Loss: 0.16532669, Log Avg loss: 0.13250864, Global Avg Loss: 0.71785061, Time: 0.0211 Steps: 72380, Updated lr: 0.000032 ***** Running evaluation checkpoint-72386 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-72386 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.768298, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.263292, "eval_total_loss": 185.09423, "eval_mae": 0.340068, "eval_mse": 0.263395, "eval_r2": 0.832569, "eval_sp_statistic": 0.926863, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.932728, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.175137, "test_total_loss": 87.918881, "test_mae": 0.279172, "test_mse": 0.175201, "test_r2": 0.886924, "test_sp_statistic": 0.918131, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946311, "test_ps_pvalue": 0.0, "lr": 3.2303461356092936e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7178039850160428, "train_cur_epoch_loss": 351.9209732078016, "train_cur_epoch_avg_loss": 0.16529871921456157, "train_cur_epoch_time": 44.76829767227173, "train_cur_epoch_avg_time": 0.021027852358981552, "epoch": 34, "step": 72386} ################################################## Training, Epoch: 0035, Batch: 000004, Sample Num: 64, Cur Loss: 0.14232388, Cur Avg Loss: 0.25624467, Log Avg loss: 0.19573787, Global Avg Loss: 0.71777848, Time: 0.0249 Steps: 72390, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000014, Sample Num: 224, Cur Loss: 0.07078445, Cur Avg Loss: 0.14435654, Log Avg loss: 0.09960128, Global Avg Loss: 0.71769310, Time: 0.0212 Steps: 72400, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000024, Sample Num: 384, Cur Loss: 0.04170702, Cur Avg Loss: 0.12813582, Log Avg loss: 0.10542680, Global Avg Loss: 0.71760854, Time: 0.0211 Steps: 72410, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000034, Sample Num: 544, Cur Loss: 0.07370691, Cur Avg Loss: 0.14271080, Log Avg loss: 0.17769075, Global Avg Loss: 0.71753399, Time: 0.0211 Steps: 72420, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000044, Sample Num: 704, Cur Loss: 0.14531867, Cur Avg Loss: 0.14330376, Log Avg loss: 0.14531981, Global Avg Loss: 0.71745499, Time: 0.0211 Steps: 72430, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000054, Sample Num: 864, Cur Loss: 0.07909560, Cur Avg Loss: 0.15986767, Log Avg loss: 0.23274889, Global Avg Loss: 0.71738807, Time: 0.0211 Steps: 72440, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000064, Sample Num: 1024, Cur Loss: 0.13916364, Cur Avg Loss: 0.16787928, Log Avg loss: 0.21114200, Global Avg Loss: 0.71731820, Time: 0.0211 Steps: 72450, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000074, Sample Num: 1184, Cur Loss: 0.34103116, Cur Avg Loss: 0.17001051, Log Avg loss: 0.18365035, Global Avg Loss: 0.71724455, Time: 0.0212 Steps: 72460, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000084, Sample Num: 1344, Cur Loss: 0.40500277, Cur Avg Loss: 0.17444182, Log Avg loss: 0.20723356, Global Avg Loss: 0.71717417, Time: 0.0211 Steps: 72470, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000094, Sample Num: 1504, Cur Loss: 0.28374776, Cur Avg Loss: 0.16789340, Log Avg loss: 0.11288663, Global Avg Loss: 0.71709080, Time: 0.0211 Steps: 72480, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000104, Sample Num: 1664, Cur Loss: 0.08435371, Cur Avg Loss: 0.16761056, Log Avg loss: 0.16495188, Global Avg Loss: 0.71701463, Time: 0.0211 Steps: 72490, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000114, Sample Num: 1824, Cur Loss: 0.41271535, Cur Avg Loss: 0.16445499, Log Avg loss: 0.13163711, Global Avg Loss: 0.71693389, Time: 0.0212 Steps: 72500, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000124, Sample Num: 1984, Cur Loss: 0.12350065, Cur Avg Loss: 0.16124884, Log Avg loss: 0.12469871, Global Avg Loss: 0.71685222, Time: 0.0211 Steps: 72510, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000134, Sample Num: 2144, Cur Loss: 0.36368942, Cur Avg Loss: 0.15827917, Log Avg loss: 0.12145517, Global Avg Loss: 0.71677011, Time: 0.0211 Steps: 72520, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000144, Sample Num: 2304, Cur Loss: 0.28176811, Cur Avg Loss: 0.16117978, Log Avg loss: 0.20004794, Global Avg Loss: 0.71669887, Time: 0.0211 Steps: 72530, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000154, Sample Num: 2464, Cur Loss: 0.02824692, Cur Avg Loss: 0.16133495, Log Avg loss: 0.16356945, Global Avg Loss: 0.71662262, Time: 0.0211 Steps: 72540, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000164, Sample Num: 2624, Cur Loss: 0.07893921, Cur Avg Loss: 0.15816340, Log Avg loss: 0.10932158, Global Avg Loss: 0.71653891, Time: 0.0211 Steps: 72550, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000174, Sample Num: 2784, Cur Loss: 0.16143821, Cur Avg Loss: 0.15881100, Log Avg loss: 0.16943156, Global Avg Loss: 0.71646351, Time: 0.0211 Steps: 72560, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000184, Sample Num: 2944, Cur Loss: 0.75052005, Cur Avg Loss: 0.15955485, Log Avg loss: 0.17249790, Global Avg Loss: 0.71638855, Time: 0.0211 Steps: 72570, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000194, Sample Num: 3104, Cur Loss: 0.05338078, Cur Avg Loss: 0.16044926, Log Avg loss: 0.17690639, Global Avg Loss: 0.71631422, Time: 0.0211 Steps: 72580, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000204, Sample Num: 3264, Cur Loss: 0.03317767, Cur Avg Loss: 0.15983222, Log Avg loss: 0.14786156, Global Avg Loss: 0.71623591, Time: 0.0213 Steps: 72590, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000214, Sample Num: 3424, Cur Loss: 0.01933786, Cur Avg Loss: 0.15988907, Log Avg loss: 0.16104893, Global Avg Loss: 0.71615944, Time: 0.0211 Steps: 72600, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000224, Sample Num: 3584, Cur Loss: 0.32985610, Cur Avg Loss: 0.16187572, Log Avg loss: 0.20439008, Global Avg Loss: 0.71608896, Time: 0.0211 Steps: 72610, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000234, Sample Num: 3744, Cur Loss: 0.25537670, Cur Avg Loss: 0.16011557, Log Avg loss: 0.12068805, Global Avg Loss: 0.71600697, Time: 0.0211 Steps: 72620, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000244, Sample Num: 3904, Cur Loss: 0.13056196, Cur Avg Loss: 0.16039711, Log Avg loss: 0.16698532, Global Avg Loss: 0.71593138, Time: 0.0211 Steps: 72630, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000254, Sample Num: 4064, Cur Loss: 0.16701180, Cur Avg Loss: 0.15896156, Log Avg loss: 0.12393401, Global Avg Loss: 0.71584988, Time: 0.0219 Steps: 72640, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000264, Sample Num: 4224, Cur Loss: 0.08242480, Cur Avg Loss: 0.16059824, Log Avg loss: 0.20217003, Global Avg Loss: 0.71577918, Time: 0.0212 Steps: 72650, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000274, Sample Num: 4384, Cur Loss: 0.10369022, Cur Avg Loss: 0.15845509, Log Avg loss: 0.10187582, Global Avg Loss: 0.71569469, Time: 0.0210 Steps: 72660, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000284, Sample Num: 4544, Cur Loss: 0.10667044, Cur Avg Loss: 0.15976322, Log Avg loss: 0.19560604, Global Avg Loss: 0.71562312, Time: 0.0210 Steps: 72670, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000294, Sample Num: 4704, Cur Loss: 0.13263932, Cur Avg Loss: 0.15788736, Log Avg loss: 0.10461302, Global Avg Loss: 0.71553905, Time: 0.0211 Steps: 72680, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000304, Sample Num: 4864, Cur Loss: 0.13447446, Cur Avg Loss: 0.15823442, Log Avg loss: 0.16843781, Global Avg Loss: 0.71546378, Time: 0.0210 Steps: 72690, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000314, Sample Num: 5024, Cur Loss: 0.23308197, Cur Avg Loss: 0.15950280, Log Avg loss: 0.19806147, Global Avg Loss: 0.71539262, Time: 0.0211 Steps: 72700, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000324, Sample Num: 5184, Cur Loss: 0.08337426, Cur Avg Loss: 0.16012954, Log Avg loss: 0.17980921, Global Avg Loss: 0.71531896, Time: 0.0210 Steps: 72710, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000334, Sample Num: 5344, Cur Loss: 0.22391126, Cur Avg Loss: 0.16204401, Log Avg loss: 0.22407308, Global Avg Loss: 0.71525140, Time: 0.0210 Steps: 72720, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000344, Sample Num: 5504, Cur Loss: 0.11164943, Cur Avg Loss: 0.16111494, Log Avg loss: 0.13008392, Global Avg Loss: 0.71517094, Time: 0.0210 Steps: 72730, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000354, Sample Num: 5664, Cur Loss: 0.16224526, Cur Avg Loss: 0.16014017, Log Avg loss: 0.12660810, Global Avg Loss: 0.71509003, Time: 0.0211 Steps: 72740, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000364, Sample Num: 5824, Cur Loss: 0.05177311, Cur Avg Loss: 0.15999114, Log Avg loss: 0.15471541, Global Avg Loss: 0.71501300, Time: 0.0210 Steps: 72750, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000374, Sample Num: 5984, Cur Loss: 0.01804325, Cur Avg Loss: 0.16039492, Log Avg loss: 0.17509235, Global Avg Loss: 0.71493880, Time: 0.0210 Steps: 72760, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000384, Sample Num: 6144, Cur Loss: 0.07694011, Cur Avg Loss: 0.15878019, Log Avg loss: 0.09838930, Global Avg Loss: 0.71485407, Time: 0.0211 Steps: 72770, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000394, Sample Num: 6304, Cur Loss: 0.05980294, Cur Avg Loss: 0.15824912, Log Avg loss: 0.13785601, Global Avg Loss: 0.71477479, Time: 0.0211 Steps: 72780, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000404, Sample Num: 6464, Cur Loss: 0.24992995, Cur Avg Loss: 0.15795031, Log Avg loss: 0.14617745, Global Avg Loss: 0.71469668, Time: 0.0211 Steps: 72790, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000414, Sample Num: 6624, Cur Loss: 0.10804843, Cur Avg Loss: 0.15769774, Log Avg loss: 0.14749373, Global Avg Loss: 0.71461877, Time: 0.0211 Steps: 72800, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000424, Sample Num: 6784, Cur Loss: 0.24446584, Cur Avg Loss: 0.16019456, Log Avg loss: 0.26356303, Global Avg Loss: 0.71455682, Time: 0.0211 Steps: 72810, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000434, Sample Num: 6944, Cur Loss: 0.08936357, Cur Avg Loss: 0.16126156, Log Avg loss: 0.20650226, Global Avg Loss: 0.71448705, Time: 0.0210 Steps: 72820, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000444, Sample Num: 7104, Cur Loss: 0.17738348, Cur Avg Loss: 0.16102307, Log Avg loss: 0.15067248, Global Avg Loss: 0.71440963, Time: 0.0210 Steps: 72830, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000454, Sample Num: 7264, Cur Loss: 0.18860486, Cur Avg Loss: 0.16073784, Log Avg loss: 0.14807358, Global Avg Loss: 0.71433188, Time: 0.0211 Steps: 72840, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000464, Sample Num: 7424, Cur Loss: 0.06576855, Cur Avg Loss: 0.16094112, Log Avg loss: 0.17017008, Global Avg Loss: 0.71425719, Time: 0.0210 Steps: 72850, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000474, Sample Num: 7584, Cur Loss: 0.06900484, Cur Avg Loss: 0.15987651, Log Avg loss: 0.11047874, Global Avg Loss: 0.71417432, Time: 0.0211 Steps: 72860, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000484, Sample Num: 7744, Cur Loss: 0.05420920, Cur Avg Loss: 0.16045999, Log Avg loss: 0.18811720, Global Avg Loss: 0.71410213, Time: 0.0211 Steps: 72870, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000494, Sample Num: 7904, Cur Loss: 0.23512395, Cur Avg Loss: 0.16037623, Log Avg loss: 0.15632223, Global Avg Loss: 0.71402559, Time: 0.0211 Steps: 72880, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000504, Sample Num: 8064, Cur Loss: 0.06044715, Cur Avg Loss: 0.15958530, Log Avg loss: 0.12051313, Global Avg Loss: 0.71394417, Time: 0.0211 Steps: 72890, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000514, Sample Num: 8224, Cur Loss: 0.04150841, Cur Avg Loss: 0.15953069, Log Avg loss: 0.15677857, Global Avg Loss: 0.71386774, Time: 0.0248 Steps: 72900, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000524, Sample Num: 8384, Cur Loss: 0.12892179, Cur Avg Loss: 0.15915947, Log Avg loss: 0.14007847, Global Avg Loss: 0.71378904, Time: 0.0211 Steps: 72910, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000534, Sample Num: 8544, Cur Loss: 0.18596424, Cur Avg Loss: 0.15967927, Log Avg loss: 0.18691686, Global Avg Loss: 0.71371679, Time: 0.0211 Steps: 72920, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000544, Sample Num: 8704, Cur Loss: 0.16099715, Cur Avg Loss: 0.15992460, Log Avg loss: 0.17302541, Global Avg Loss: 0.71364265, Time: 0.0211 Steps: 72930, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000554, Sample Num: 8864, Cur Loss: 0.17387363, Cur Avg Loss: 0.15976287, Log Avg loss: 0.15096467, Global Avg Loss: 0.71356550, Time: 0.0211 Steps: 72940, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000564, Sample Num: 9024, Cur Loss: 0.13544467, Cur Avg Loss: 0.16027064, Log Avg loss: 0.18840092, Global Avg Loss: 0.71349351, Time: 0.0211 Steps: 72950, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000574, Sample Num: 9184, Cur Loss: 0.25951529, Cur Avg Loss: 0.16002131, Log Avg loss: 0.14595946, Global Avg Loss: 0.71341573, Time: 0.0211 Steps: 72960, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000584, Sample Num: 9344, Cur Loss: 0.05644171, Cur Avg Loss: 0.16037626, Log Avg loss: 0.18075014, Global Avg Loss: 0.71334273, Time: 0.0211 Steps: 72970, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000594, Sample Num: 9504, Cur Loss: 0.04269571, Cur Avg Loss: 0.16016344, Log Avg loss: 0.14773456, Global Avg Loss: 0.71326523, Time: 0.0211 Steps: 72980, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000604, Sample Num: 9664, Cur Loss: 0.09188531, Cur Avg Loss: 0.16050213, Log Avg loss: 0.18062076, Global Avg Loss: 0.71319225, Time: 0.0211 Steps: 72990, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000614, Sample Num: 9824, Cur Loss: 0.07972012, Cur Avg Loss: 0.16051689, Log Avg loss: 0.16140842, Global Avg Loss: 0.71311667, Time: 0.0211 Steps: 73000, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000624, Sample Num: 9984, Cur Loss: 0.13615201, Cur Avg Loss: 0.15993149, Log Avg loss: 0.12398754, Global Avg Loss: 0.71303597, Time: 0.0211 Steps: 73010, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000634, Sample Num: 10144, Cur Loss: 0.55897796, Cur Avg Loss: 0.16077171, Log Avg loss: 0.21320144, Global Avg Loss: 0.71296752, Time: 0.0211 Steps: 73020, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000644, Sample Num: 10304, Cur Loss: 0.44046354, Cur Avg Loss: 0.16041284, Log Avg loss: 0.13766038, Global Avg Loss: 0.71288875, Time: 0.0211 Steps: 73030, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000654, Sample Num: 10464, Cur Loss: 0.22415057, Cur Avg Loss: 0.16051922, Log Avg loss: 0.16737058, Global Avg Loss: 0.71281406, Time: 0.0211 Steps: 73040, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000664, Sample Num: 10624, Cur Loss: 0.05481891, Cur Avg Loss: 0.16080094, Log Avg loss: 0.17922490, Global Avg Loss: 0.71274101, Time: 0.0211 Steps: 73050, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000674, Sample Num: 10784, Cur Loss: 0.07548182, Cur Avg Loss: 0.16016354, Log Avg loss: 0.11784049, Global Avg Loss: 0.71265959, Time: 0.0211 Steps: 73060, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000684, Sample Num: 10944, Cur Loss: 0.09421550, Cur Avg Loss: 0.15967403, Log Avg loss: 0.12668120, Global Avg Loss: 0.71257939, Time: 0.0211 Steps: 73070, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000694, Sample Num: 11104, Cur Loss: 0.26148036, Cur Avg Loss: 0.15953597, Log Avg loss: 0.15009251, Global Avg Loss: 0.71250243, Time: 0.0211 Steps: 73080, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000704, Sample Num: 11264, Cur Loss: 0.16146272, Cur Avg Loss: 0.15910996, Log Avg loss: 0.12954451, Global Avg Loss: 0.71242267, Time: 0.0211 Steps: 73090, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000714, Sample Num: 11424, Cur Loss: 0.47626644, Cur Avg Loss: 0.15927341, Log Avg loss: 0.17078074, Global Avg Loss: 0.71234857, Time: 0.0211 Steps: 73100, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000724, Sample Num: 11584, Cur Loss: 0.25164896, Cur Avg Loss: 0.15963819, Log Avg loss: 0.18568304, Global Avg Loss: 0.71227653, Time: 0.0211 Steps: 73110, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000734, Sample Num: 11744, Cur Loss: 0.03155306, Cur Avg Loss: 0.15913880, Log Avg loss: 0.12298352, Global Avg Loss: 0.71219594, Time: 0.0211 Steps: 73120, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000744, Sample Num: 11904, Cur Loss: 0.14018600, Cur Avg Loss: 0.15896833, Log Avg loss: 0.14645577, Global Avg Loss: 0.71211858, Time: 0.0211 Steps: 73130, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000754, Sample Num: 12064, Cur Loss: 0.11104922, Cur Avg Loss: 0.15877434, Log Avg loss: 0.14434108, Global Avg Loss: 0.71204095, Time: 0.0211 Steps: 73140, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000764, Sample Num: 12224, Cur Loss: 0.08337883, Cur Avg Loss: 0.15972851, Log Avg loss: 0.23167313, Global Avg Loss: 0.71197528, Time: 0.0211 Steps: 73150, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000774, Sample Num: 12384, Cur Loss: 0.24956366, Cur Avg Loss: 0.15962985, Log Avg loss: 0.15209183, Global Avg Loss: 0.71189875, Time: 0.0212 Steps: 73160, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000784, Sample Num: 12544, Cur Loss: 0.22953776, Cur Avg Loss: 0.16043909, Log Avg loss: 0.22307483, Global Avg Loss: 0.71183195, Time: 0.0211 Steps: 73170, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000794, Sample Num: 12704, Cur Loss: 0.15292889, Cur Avg Loss: 0.16095602, Log Avg loss: 0.20148278, Global Avg Loss: 0.71176221, Time: 0.0210 Steps: 73180, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000804, Sample Num: 12864, Cur Loss: 0.09405207, Cur Avg Loss: 0.16093921, Log Avg loss: 0.15960464, Global Avg Loss: 0.71168677, Time: 0.0210 Steps: 73190, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000814, Sample Num: 13024, Cur Loss: 0.03030786, Cur Avg Loss: 0.16041615, Log Avg loss: 0.11836216, Global Avg Loss: 0.71160571, Time: 0.0211 Steps: 73200, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000824, Sample Num: 13184, Cur Loss: 0.10320698, Cur Avg Loss: 0.15976782, Log Avg loss: 0.10699386, Global Avg Loss: 0.71152312, Time: 0.0210 Steps: 73210, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000834, Sample Num: 13344, Cur Loss: 0.14770120, Cur Avg Loss: 0.15985258, Log Avg loss: 0.16683661, Global Avg Loss: 0.71144873, Time: 0.0210 Steps: 73220, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000844, Sample Num: 13504, Cur Loss: 0.16779159, Cur Avg Loss: 0.15994368, Log Avg loss: 0.16754183, Global Avg Loss: 0.71137446, Time: 0.0210 Steps: 73230, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000854, Sample Num: 13664, Cur Loss: 0.28460562, Cur Avg Loss: 0.16015659, Log Avg loss: 0.17812606, Global Avg Loss: 0.71130165, Time: 0.0210 Steps: 73240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000864, Sample Num: 13824, Cur Loss: 0.08134541, Cur Avg Loss: 0.15973712, Log Avg loss: 0.12391422, Global Avg Loss: 0.71122146, Time: 0.0210 Steps: 73250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000874, Sample Num: 13984, Cur Loss: 0.20325345, Cur Avg Loss: 0.15967313, Log Avg loss: 0.15414489, Global Avg Loss: 0.71114542, Time: 0.0210 Steps: 73260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000884, Sample Num: 14144, Cur Loss: 0.06382366, Cur Avg Loss: 0.15931914, Log Avg loss: 0.12838038, Global Avg Loss: 0.71106588, Time: 0.0210 Steps: 73270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000894, Sample Num: 14304, Cur Loss: 0.14156583, Cur Avg Loss: 0.15888177, Log Avg loss: 0.12021793, Global Avg Loss: 0.71098526, Time: 0.0210 Steps: 73280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000904, Sample Num: 14464, Cur Loss: 0.26593965, Cur Avg Loss: 0.15898360, Log Avg loss: 0.16808723, Global Avg Loss: 0.71091118, Time: 0.0210 Steps: 73290, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000914, Sample Num: 14624, Cur Loss: 0.07289986, Cur Avg Loss: 0.15897593, Log Avg loss: 0.15828261, Global Avg Loss: 0.71083579, Time: 0.0210 Steps: 73300, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000924, Sample Num: 14784, Cur Loss: 0.16886352, Cur Avg Loss: 0.15837960, Log Avg loss: 0.10387466, Global Avg Loss: 0.71075299, Time: 0.0211 Steps: 73310, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000934, Sample Num: 14944, Cur Loss: 0.36304143, Cur Avg Loss: 0.15856571, Log Avg loss: 0.17576245, Global Avg Loss: 0.71068003, Time: 0.0210 Steps: 73320, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000944, Sample Num: 15104, Cur Loss: 0.05801505, Cur Avg Loss: 0.15831592, Log Avg loss: 0.13498524, Global Avg Loss: 0.71060152, Time: 0.0211 Steps: 73330, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000954, Sample Num: 15264, Cur Loss: 0.05217771, Cur Avg Loss: 0.15798568, Log Avg loss: 0.12681133, Global Avg Loss: 0.71052192, Time: 0.0210 Steps: 73340, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000964, Sample Num: 15424, Cur Loss: 0.19918661, Cur Avg Loss: 0.15772825, Log Avg loss: 0.13316925, Global Avg Loss: 0.71044321, Time: 0.0210 Steps: 73350, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000974, Sample Num: 15584, Cur Loss: 0.08079091, Cur Avg Loss: 0.15747785, Log Avg loss: 0.13333983, Global Avg Loss: 0.71036454, Time: 0.0210 Steps: 73360, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000984, Sample Num: 15744, Cur Loss: 0.43215179, Cur Avg Loss: 0.15825731, Log Avg loss: 0.23417647, Global Avg Loss: 0.71029964, Time: 0.0210 Steps: 73370, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000994, Sample Num: 15904, Cur Loss: 0.23395683, Cur Avg Loss: 0.15874236, Log Avg loss: 0.20647093, Global Avg Loss: 0.71023098, Time: 0.0210 Steps: 73380, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001004, Sample Num: 16064, Cur Loss: 0.20162283, Cur Avg Loss: 0.15957132, Log Avg loss: 0.24197016, Global Avg Loss: 0.71016717, Time: 0.0211 Steps: 73390, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001014, Sample Num: 16224, Cur Loss: 0.35295889, Cur Avg Loss: 0.16002642, Log Avg loss: 0.20571888, Global Avg Loss: 0.71009845, Time: 0.0210 Steps: 73400, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001024, Sample Num: 16384, Cur Loss: 0.18066075, Cur Avg Loss: 0.16024704, Log Avg loss: 0.18261753, Global Avg Loss: 0.71002659, Time: 0.0255 Steps: 73410, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001034, Sample Num: 16544, Cur Loss: 0.06492499, Cur Avg Loss: 0.16020238, Log Avg loss: 0.15562929, Global Avg Loss: 0.70995108, Time: 0.0212 Steps: 73420, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001044, Sample Num: 16704, Cur Loss: 0.16878423, Cur Avg Loss: 0.16052625, Log Avg loss: 0.19401420, Global Avg Loss: 0.70988082, Time: 0.0211 Steps: 73430, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001054, Sample Num: 16864, Cur Loss: 0.04704337, Cur Avg Loss: 0.16042480, Log Avg loss: 0.14983330, Global Avg Loss: 0.70980456, Time: 0.0211 Steps: 73440, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001064, Sample Num: 17024, Cur Loss: 0.05197573, Cur Avg Loss: 0.16062975, Log Avg loss: 0.18223197, Global Avg Loss: 0.70973273, Time: 0.0211 Steps: 73450, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001074, Sample Num: 17184, Cur Loss: 0.06039271, Cur Avg Loss: 0.16041999, Log Avg loss: 0.13810141, Global Avg Loss: 0.70965492, Time: 0.0212 Steps: 73460, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001084, Sample Num: 17344, Cur Loss: 0.18104836, Cur Avg Loss: 0.16022582, Log Avg loss: 0.13937198, Global Avg Loss: 0.70957730, Time: 0.0211 Steps: 73470, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001094, Sample Num: 17504, Cur Loss: 0.26972193, Cur Avg Loss: 0.15999739, Log Avg loss: 0.13523559, Global Avg Loss: 0.70949913, Time: 0.0211 Steps: 73480, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001104, Sample Num: 17664, Cur Loss: 0.04157187, Cur Avg Loss: 0.16086780, Log Avg loss: 0.25609063, Global Avg Loss: 0.70943744, Time: 0.0211 Steps: 73490, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001114, Sample Num: 17824, Cur Loss: 0.11751120, Cur Avg Loss: 0.16054689, Log Avg loss: 0.12511885, Global Avg Loss: 0.70935794, Time: 0.0211 Steps: 73500, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001124, Sample Num: 17984, Cur Loss: 0.32935432, Cur Avg Loss: 0.16053561, Log Avg loss: 0.15927858, Global Avg Loss: 0.70928311, Time: 0.0212 Steps: 73510, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001134, Sample Num: 18144, Cur Loss: 0.21498746, Cur Avg Loss: 0.16107769, Log Avg loss: 0.22200745, Global Avg Loss: 0.70921683, Time: 0.0212 Steps: 73520, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001144, Sample Num: 18304, Cur Loss: 0.19594194, Cur Avg Loss: 0.16134959, Log Avg loss: 0.19218279, Global Avg Loss: 0.70914651, Time: 0.0211 Steps: 73530, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001154, Sample Num: 18464, Cur Loss: 0.06121139, Cur Avg Loss: 0.16081108, Log Avg loss: 0.09920530, Global Avg Loss: 0.70906357, Time: 0.0211 Steps: 73540, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001164, Sample Num: 18624, Cur Loss: 0.12492343, Cur Avg Loss: 0.16082507, Log Avg loss: 0.16244016, Global Avg Loss: 0.70898925, Time: 0.0211 Steps: 73550, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001174, Sample Num: 18784, Cur Loss: 0.18009001, Cur Avg Loss: 0.16108875, Log Avg loss: 0.19178084, Global Avg Loss: 0.70891894, Time: 0.0211 Steps: 73560, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001184, Sample Num: 18944, Cur Loss: 0.11097512, Cur Avg Loss: 0.16098624, Log Avg loss: 0.14895184, Global Avg Loss: 0.70884283, Time: 0.0211 Steps: 73570, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001194, Sample Num: 19104, Cur Loss: 0.11275300, Cur Avg Loss: 0.16061728, Log Avg loss: 0.11693270, Global Avg Loss: 0.70876239, Time: 0.0212 Steps: 73580, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001204, Sample Num: 19264, Cur Loss: 0.20721902, Cur Avg Loss: 0.16072703, Log Avg loss: 0.17383018, Global Avg Loss: 0.70868969, Time: 0.0211 Steps: 73590, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001214, Sample Num: 19424, Cur Loss: 0.14835101, Cur Avg Loss: 0.16076393, Log Avg loss: 0.16520746, Global Avg Loss: 0.70861585, Time: 0.0212 Steps: 73600, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001224, Sample Num: 19584, Cur Loss: 0.23680806, Cur Avg Loss: 0.16089436, Log Avg loss: 0.17672786, Global Avg Loss: 0.70854359, Time: 0.0211 Steps: 73610, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001234, Sample Num: 19744, Cur Loss: 0.05258256, Cur Avg Loss: 0.16066242, Log Avg loss: 0.13227398, Global Avg Loss: 0.70846532, Time: 0.0212 Steps: 73620, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001244, Sample Num: 19904, Cur Loss: 0.12771258, Cur Avg Loss: 0.16067232, Log Avg loss: 0.16189314, Global Avg Loss: 0.70839109, Time: 0.0211 Steps: 73630, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001254, Sample Num: 20064, Cur Loss: 0.25695309, Cur Avg Loss: 0.16049055, Log Avg loss: 0.13787879, Global Avg Loss: 0.70831361, Time: 0.0211 Steps: 73640, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001264, Sample Num: 20224, Cur Loss: 0.07797480, Cur Avg Loss: 0.16064081, Log Avg loss: 0.17948311, Global Avg Loss: 0.70824181, Time: 0.0212 Steps: 73650, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001274, Sample Num: 20384, Cur Loss: 0.06026120, Cur Avg Loss: 0.16115412, Log Avg loss: 0.22603668, Global Avg Loss: 0.70817635, Time: 0.0212 Steps: 73660, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001284, Sample Num: 20544, Cur Loss: 0.16207889, Cur Avg Loss: 0.16118623, Log Avg loss: 0.16527760, Global Avg Loss: 0.70810265, Time: 0.0248 Steps: 73670, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001294, Sample Num: 20704, Cur Loss: 0.24076046, Cur Avg Loss: 0.16126907, Log Avg loss: 0.17190535, Global Avg Loss: 0.70802988, Time: 0.0212 Steps: 73680, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001304, Sample Num: 20864, Cur Loss: 0.34008807, Cur Avg Loss: 0.16143841, Log Avg loss: 0.18335021, Global Avg Loss: 0.70795868, Time: 0.0212 Steps: 73690, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001314, Sample Num: 21024, Cur Loss: 0.22357738, Cur Avg Loss: 0.16115762, Log Avg loss: 0.12454338, Global Avg Loss: 0.70787952, Time: 0.0212 Steps: 73700, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001324, Sample Num: 21184, Cur Loss: 0.13874426, Cur Avg Loss: 0.16163837, Log Avg loss: 0.22480894, Global Avg Loss: 0.70781398, Time: 0.0211 Steps: 73710, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001334, Sample Num: 21344, Cur Loss: 0.12202638, Cur Avg Loss: 0.16145923, Log Avg loss: 0.13774122, Global Avg Loss: 0.70773665, Time: 0.0212 Steps: 73720, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001344, Sample Num: 21504, Cur Loss: 0.13414401, Cur Avg Loss: 0.16171138, Log Avg loss: 0.19534851, Global Avg Loss: 0.70766716, Time: 0.0212 Steps: 73730, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001354, Sample Num: 21664, Cur Loss: 0.17728800, Cur Avg Loss: 0.16236792, Log Avg loss: 0.25060654, Global Avg Loss: 0.70760517, Time: 0.0213 Steps: 73740, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001364, Sample Num: 21824, Cur Loss: 0.28666037, Cur Avg Loss: 0.16303546, Log Avg loss: 0.25342021, Global Avg Loss: 0.70754359, Time: 0.0212 Steps: 73750, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001374, Sample Num: 21984, Cur Loss: 0.10643119, Cur Avg Loss: 0.16284550, Log Avg loss: 0.13693489, Global Avg Loss: 0.70746623, Time: 0.0212 Steps: 73760, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001384, Sample Num: 22144, Cur Loss: 0.13727280, Cur Avg Loss: 0.16327126, Log Avg loss: 0.22177057, Global Avg Loss: 0.70740039, Time: 0.0212 Steps: 73770, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001394, Sample Num: 22304, Cur Loss: 0.06960464, Cur Avg Loss: 0.16316286, Log Avg loss: 0.14815970, Global Avg Loss: 0.70732459, Time: 0.0212 Steps: 73780, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001404, Sample Num: 22464, Cur Loss: 0.35075003, Cur Avg Loss: 0.16365036, Log Avg loss: 0.23160880, Global Avg Loss: 0.70726012, Time: 0.0212 Steps: 73790, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001414, Sample Num: 22624, Cur Loss: 0.33568817, Cur Avg Loss: 0.16403981, Log Avg loss: 0.21871893, Global Avg Loss: 0.70719392, Time: 0.0212 Steps: 73800, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001424, Sample Num: 22784, Cur Loss: 0.05470605, Cur Avg Loss: 0.16387647, Log Avg loss: 0.14077979, Global Avg Loss: 0.70711718, Time: 0.0212 Steps: 73810, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001434, Sample Num: 22944, Cur Loss: 0.17047691, Cur Avg Loss: 0.16389218, Log Avg loss: 0.16612957, Global Avg Loss: 0.70704390, Time: 0.0212 Steps: 73820, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001444, Sample Num: 23104, Cur Loss: 0.50429606, Cur Avg Loss: 0.16386408, Log Avg loss: 0.15983345, Global Avg Loss: 0.70696978, Time: 0.0212 Steps: 73830, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001454, Sample Num: 23264, Cur Loss: 0.06760044, Cur Avg Loss: 0.16342137, Log Avg loss: 0.09949437, Global Avg Loss: 0.70688751, Time: 0.0212 Steps: 73840, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001464, Sample Num: 23424, Cur Loss: 0.07440989, Cur Avg Loss: 0.16333591, Log Avg loss: 0.15091006, Global Avg Loss: 0.70681223, Time: 0.0212 Steps: 73850, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001474, Sample Num: 23584, Cur Loss: 0.10770857, Cur Avg Loss: 0.16292454, Log Avg loss: 0.10270019, Global Avg Loss: 0.70673044, Time: 0.0212 Steps: 73860, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001484, Sample Num: 23744, Cur Loss: 0.08058621, Cur Avg Loss: 0.16256096, Log Avg loss: 0.10896941, Global Avg Loss: 0.70664952, Time: 0.0212 Steps: 73870, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001494, Sample Num: 23904, Cur Loss: 0.11965229, Cur Avg Loss: 0.16285010, Log Avg loss: 0.20575836, Global Avg Loss: 0.70658172, Time: 0.0212 Steps: 73880, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001504, Sample Num: 24064, Cur Loss: 0.14284423, Cur Avg Loss: 0.16262918, Log Avg loss: 0.12962448, Global Avg Loss: 0.70650363, Time: 0.0212 Steps: 73890, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001514, Sample Num: 24224, Cur Loss: 0.13026845, Cur Avg Loss: 0.16236243, Log Avg loss: 0.12224177, Global Avg Loss: 0.70642457, Time: 0.0212 Steps: 73900, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001524, Sample Num: 24384, Cur Loss: 0.22038789, Cur Avg Loss: 0.16285224, Log Avg loss: 0.23701066, Global Avg Loss: 0.70636106, Time: 0.0212 Steps: 73910, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001534, Sample Num: 24544, Cur Loss: 0.21015114, Cur Avg Loss: 0.16315486, Log Avg loss: 0.20927350, Global Avg Loss: 0.70629382, Time: 0.0212 Steps: 73920, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001544, Sample Num: 24704, Cur Loss: 0.12041093, Cur Avg Loss: 0.16365077, Log Avg loss: 0.23972354, Global Avg Loss: 0.70623071, Time: 0.0212 Steps: 73930, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001554, Sample Num: 24864, Cur Loss: 0.09114716, Cur Avg Loss: 0.16359163, Log Avg loss: 0.15446071, Global Avg Loss: 0.70615608, Time: 0.0211 Steps: 73940, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001564, Sample Num: 25024, Cur Loss: 0.22811425, Cur Avg Loss: 0.16319035, Log Avg loss: 0.10083155, Global Avg Loss: 0.70607423, Time: 0.0211 Steps: 73950, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001574, Sample Num: 25184, Cur Loss: 0.06313438, Cur Avg Loss: 0.16330299, Log Avg loss: 0.18091874, Global Avg Loss: 0.70600322, Time: 0.0212 Steps: 73960, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001584, Sample Num: 25344, Cur Loss: 0.20165715, Cur Avg Loss: 0.16319968, Log Avg loss: 0.14693900, Global Avg Loss: 0.70592764, Time: 0.0211 Steps: 73970, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001594, Sample Num: 25504, Cur Loss: 0.12000529, Cur Avg Loss: 0.16327784, Log Avg loss: 0.17565879, Global Avg Loss: 0.70585596, Time: 0.0211 Steps: 73980, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001604, Sample Num: 25664, Cur Loss: 0.15834279, Cur Avg Loss: 0.16324257, Log Avg loss: 0.15762088, Global Avg Loss: 0.70578187, Time: 0.0211 Steps: 73990, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001614, Sample Num: 25824, Cur Loss: 0.27031511, Cur Avg Loss: 0.16299114, Log Avg loss: 0.12266080, Global Avg Loss: 0.70570307, Time: 0.0211 Steps: 74000, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001624, Sample Num: 25984, Cur Loss: 0.31067449, Cur Avg Loss: 0.16333349, Log Avg loss: 0.21858906, Global Avg Loss: 0.70563725, Time: 0.0211 Steps: 74010, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001634, Sample Num: 26144, Cur Loss: 0.22291742, Cur Avg Loss: 0.16307237, Log Avg loss: 0.12066693, Global Avg Loss: 0.70555822, Time: 0.0211 Steps: 74020, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001644, Sample Num: 26304, Cur Loss: 0.03732746, Cur Avg Loss: 0.16267744, Log Avg loss: 0.09814499, Global Avg Loss: 0.70547617, Time: 0.0210 Steps: 74030, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001654, Sample Num: 26464, Cur Loss: 0.22484645, Cur Avg Loss: 0.16289398, Log Avg loss: 0.19849341, Global Avg Loss: 0.70540770, Time: 0.0211 Steps: 74040, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001664, Sample Num: 26624, Cur Loss: 0.26130819, Cur Avg Loss: 0.16318669, Log Avg loss: 0.21160106, Global Avg Loss: 0.70534101, Time: 0.0212 Steps: 74050, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001674, Sample Num: 26784, Cur Loss: 0.16430850, Cur Avg Loss: 0.16347168, Log Avg loss: 0.21089496, Global Avg Loss: 0.70527425, Time: 0.0213 Steps: 74060, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001684, Sample Num: 26944, Cur Loss: 0.06085299, Cur Avg Loss: 0.16308223, Log Avg loss: 0.09788820, Global Avg Loss: 0.70519225, Time: 0.0212 Steps: 74070, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001694, Sample Num: 27104, Cur Loss: 0.08369541, Cur Avg Loss: 0.16291516, Log Avg loss: 0.13478093, Global Avg Loss: 0.70511525, Time: 0.0212 Steps: 74080, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001704, Sample Num: 27264, Cur Loss: 0.10870299, Cur Avg Loss: 0.16282564, Log Avg loss: 0.14765943, Global Avg Loss: 0.70504001, Time: 0.0211 Steps: 74090, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001714, Sample Num: 27424, Cur Loss: 0.37863648, Cur Avg Loss: 0.16307769, Log Avg loss: 0.20602712, Global Avg Loss: 0.70497266, Time: 0.0211 Steps: 74100, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001724, Sample Num: 27584, Cur Loss: 0.11738563, Cur Avg Loss: 0.16304883, Log Avg loss: 0.15810208, Global Avg Loss: 0.70489887, Time: 0.0211 Steps: 74110, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001734, Sample Num: 27744, Cur Loss: 0.22989073, Cur Avg Loss: 0.16301681, Log Avg loss: 0.15749728, Global Avg Loss: 0.70482502, Time: 0.0212 Steps: 74120, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001744, Sample Num: 27904, Cur Loss: 0.15455596, Cur Avg Loss: 0.16322389, Log Avg loss: 0.19913212, Global Avg Loss: 0.70475680, Time: 0.0211 Steps: 74130, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001754, Sample Num: 28064, Cur Loss: 0.10101468, Cur Avg Loss: 0.16315181, Log Avg loss: 0.15058108, Global Avg Loss: 0.70468205, Time: 0.0211 Steps: 74140, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001764, Sample Num: 28224, Cur Loss: 0.05807334, Cur Avg Loss: 0.16298575, Log Avg loss: 0.13385892, Global Avg Loss: 0.70460507, Time: 0.0211 Steps: 74150, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001774, Sample Num: 28384, Cur Loss: 0.16039360, Cur Avg Loss: 0.16292338, Log Avg loss: 0.15192029, Global Avg Loss: 0.70453055, Time: 0.0211 Steps: 74160, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001784, Sample Num: 28544, Cur Loss: 0.19091466, Cur Avg Loss: 0.16277490, Log Avg loss: 0.13643423, Global Avg Loss: 0.70445395, Time: 0.0210 Steps: 74170, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001794, Sample Num: 28704, Cur Loss: 0.44782019, Cur Avg Loss: 0.16317120, Log Avg loss: 0.23387187, Global Avg Loss: 0.70439051, Time: 0.0249 Steps: 74180, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001804, Sample Num: 28864, Cur Loss: 0.30535132, Cur Avg Loss: 0.16310919, Log Avg loss: 0.15198399, Global Avg Loss: 0.70431606, Time: 0.0211 Steps: 74190, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001814, Sample Num: 29024, Cur Loss: 0.28154036, Cur Avg Loss: 0.16323546, Log Avg loss: 0.18601426, Global Avg Loss: 0.70424620, Time: 0.0211 Steps: 74200, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001824, Sample Num: 29184, Cur Loss: 0.17453271, Cur Avg Loss: 0.16320588, Log Avg loss: 0.15784173, Global Avg Loss: 0.70417258, Time: 0.0211 Steps: 74210, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001834, Sample Num: 29344, Cur Loss: 0.28048369, Cur Avg Loss: 0.16319024, Log Avg loss: 0.16033738, Global Avg Loss: 0.70409930, Time: 0.0211 Steps: 74220, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001844, Sample Num: 29504, Cur Loss: 0.15013196, Cur Avg Loss: 0.16310695, Log Avg loss: 0.14783156, Global Avg Loss: 0.70402436, Time: 0.0211 Steps: 74230, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001854, Sample Num: 29664, Cur Loss: 0.29146650, Cur Avg Loss: 0.16315024, Log Avg loss: 0.17113179, Global Avg Loss: 0.70395258, Time: 0.0210 Steps: 74240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001864, Sample Num: 29824, Cur Loss: 0.22249065, Cur Avg Loss: 0.16335344, Log Avg loss: 0.20102708, Global Avg Loss: 0.70388485, Time: 0.0211 Steps: 74250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001874, Sample Num: 29984, Cur Loss: 0.17339879, Cur Avg Loss: 0.16368906, Log Avg loss: 0.22624948, Global Avg Loss: 0.70382053, Time: 0.0210 Steps: 74260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001884, Sample Num: 30144, Cur Loss: 0.04517737, Cur Avg Loss: 0.16417303, Log Avg loss: 0.25486792, Global Avg Loss: 0.70376008, Time: 0.0211 Steps: 74270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001894, Sample Num: 30304, Cur Loss: 0.02391467, Cur Avg Loss: 0.16412288, Log Avg loss: 0.15467449, Global Avg Loss: 0.70368616, Time: 0.0211 Steps: 74280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001904, Sample Num: 30464, Cur Loss: 0.15015036, Cur Avg Loss: 0.16439883, Log Avg loss: 0.21666357, Global Avg Loss: 0.70362060, Time: 0.0211 Steps: 74290, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001914, Sample Num: 30624, Cur Loss: 0.06151265, Cur Avg Loss: 0.16435786, Log Avg loss: 0.15655754, Global Avg Loss: 0.70354697, Time: 0.0211 Steps: 74300, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001924, Sample Num: 30784, Cur Loss: 0.10100178, Cur Avg Loss: 0.16406179, Log Avg loss: 0.10739454, Global Avg Loss: 0.70346675, Time: 0.0211 Steps: 74310, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001934, Sample Num: 30944, Cur Loss: 0.10298976, Cur Avg Loss: 0.16404696, Log Avg loss: 0.16119382, Global Avg Loss: 0.70339378, Time: 0.0211 Steps: 74320, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001944, Sample Num: 31104, Cur Loss: 0.16166502, Cur Avg Loss: 0.16413988, Log Avg loss: 0.18211072, Global Avg Loss: 0.70332365, Time: 0.0211 Steps: 74330, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001954, Sample Num: 31264, Cur Loss: 0.21519889, Cur Avg Loss: 0.16434214, Log Avg loss: 0.20366096, Global Avg Loss: 0.70325644, Time: 0.0211 Steps: 74340, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001964, Sample Num: 31424, Cur Loss: 0.09351198, Cur Avg Loss: 0.16454888, Log Avg loss: 0.20494529, Global Avg Loss: 0.70318942, Time: 0.0211 Steps: 74350, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001974, Sample Num: 31584, Cur Loss: 0.08847785, Cur Avg Loss: 0.16458500, Log Avg loss: 0.17167994, Global Avg Loss: 0.70311794, Time: 0.0211 Steps: 74360, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001984, Sample Num: 31744, Cur Loss: 0.13054042, Cur Avg Loss: 0.16509710, Log Avg loss: 0.26618602, Global Avg Loss: 0.70305919, Time: 0.0211 Steps: 74370, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001994, Sample Num: 31904, Cur Loss: 0.21544501, Cur Avg Loss: 0.16516308, Log Avg loss: 0.17825284, Global Avg Loss: 0.70298863, Time: 0.0210 Steps: 74380, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002004, Sample Num: 32064, Cur Loss: 0.09089930, Cur Avg Loss: 0.16490458, Log Avg loss: 0.11335944, Global Avg Loss: 0.70290937, Time: 0.0211 Steps: 74390, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002014, Sample Num: 32224, Cur Loss: 0.11791806, Cur Avg Loss: 0.16482758, Log Avg loss: 0.14939728, Global Avg Loss: 0.70283497, Time: 0.0211 Steps: 74400, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002024, Sample Num: 32384, Cur Loss: 0.19797963, Cur Avg Loss: 0.16469319, Log Avg loss: 0.13762704, Global Avg Loss: 0.70275901, Time: 0.0211 Steps: 74410, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002034, Sample Num: 32544, Cur Loss: 0.02095028, Cur Avg Loss: 0.16440080, Log Avg loss: 0.10522036, Global Avg Loss: 0.70267872, Time: 0.0211 Steps: 74420, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002044, Sample Num: 32704, Cur Loss: 0.23547611, Cur Avg Loss: 0.16444991, Log Avg loss: 0.17443898, Global Avg Loss: 0.70260775, Time: 0.0212 Steps: 74430, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002054, Sample Num: 32864, Cur Loss: 0.09665045, Cur Avg Loss: 0.16429721, Log Avg loss: 0.13308495, Global Avg Loss: 0.70253124, Time: 0.0249 Steps: 74440, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002064, Sample Num: 33024, Cur Loss: 0.12567607, Cur Avg Loss: 0.16418091, Log Avg loss: 0.14029351, Global Avg Loss: 0.70245572, Time: 0.0211 Steps: 74450, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002074, Sample Num: 33184, Cur Loss: 0.02864396, Cur Avg Loss: 0.16398955, Log Avg loss: 0.12449249, Global Avg Loss: 0.70237810, Time: 0.0210 Steps: 74460, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002084, Sample Num: 33344, Cur Loss: 0.24991107, Cur Avg Loss: 0.16394711, Log Avg loss: 0.15514453, Global Avg Loss: 0.70230462, Time: 0.0211 Steps: 74470, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002094, Sample Num: 33504, Cur Loss: 0.41368797, Cur Avg Loss: 0.16403951, Log Avg loss: 0.18329599, Global Avg Loss: 0.70223494, Time: 0.0211 Steps: 74480, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002104, Sample Num: 33664, Cur Loss: 0.41263473, Cur Avg Loss: 0.16408795, Log Avg loss: 0.17423208, Global Avg Loss: 0.70216405, Time: 0.0211 Steps: 74490, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002114, Sample Num: 33824, Cur Loss: 0.06227775, Cur Avg Loss: 0.16389622, Log Avg loss: 0.12355585, Global Avg Loss: 0.70208639, Time: 0.0211 Steps: 74500, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002124, Sample Num: 33984, Cur Loss: 0.07951397, Cur Avg Loss: 0.16383411, Log Avg loss: 0.15070461, Global Avg Loss: 0.70201239, Time: 0.0211 Steps: 74510, Updated lr: 0.000030 ***** Running evaluation checkpoint-74515 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-74515 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 45.080488, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.318094, "eval_total_loss": 223.619893, "eval_mae": 0.36985, "eval_mse": 0.318222, "eval_r2": 0.797717, "eval_sp_statistic": 0.92605, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.930936, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.198352, "test_total_loss": 99.572654, "test_mae": 0.324623, "test_mse": 0.198422, "test_r2": 0.871937, "test_sp_statistic": 0.91758, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947006, "test_ps_pvalue": 0.0, "lr": 3.0284495021337127e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.7019795318128317, "train_cur_epoch_loss": 349.0455536618829, "train_cur_epoch_avg_loss": 0.16394812290365565, "train_cur_epoch_time": 45.080487966537476, "train_cur_epoch_avg_time": 0.02117448941594057, "epoch": 35, "step": 74515} ################################################## Training, Epoch: 0036, Batch: 000005, Sample Num: 80, Cur Loss: 0.03873838, Cur Avg Loss: 0.16416144, Log Avg loss: 0.18827077, Global Avg Loss: 0.70194345, Time: 0.0232 Steps: 74520, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000015, Sample Num: 240, Cur Loss: 0.06911376, Cur Avg Loss: 0.13964544, Log Avg loss: 0.12738744, Global Avg Loss: 0.70186636, Time: 0.0211 Steps: 74530, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000025, Sample Num: 400, Cur Loss: 0.10382821, Cur Avg Loss: 0.15100618, Log Avg loss: 0.16804730, Global Avg Loss: 0.70179474, Time: 0.0212 Steps: 74540, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000035, Sample Num: 560, Cur Loss: 0.20220464, Cur Avg Loss: 0.14770349, Log Avg loss: 0.13944674, Global Avg Loss: 0.70171931, Time: 0.0211 Steps: 74550, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000045, Sample Num: 720, Cur Loss: 0.19752485, Cur Avg Loss: 0.15168018, Log Avg loss: 0.16559860, Global Avg Loss: 0.70164740, Time: 0.0212 Steps: 74560, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000055, Sample Num: 880, Cur Loss: 0.04957137, Cur Avg Loss: 0.17050878, Log Avg loss: 0.25523746, Global Avg Loss: 0.70158754, Time: 0.0212 Steps: 74570, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000065, Sample Num: 1040, Cur Loss: 0.08771889, Cur Avg Loss: 0.16152951, Log Avg loss: 0.11214357, Global Avg Loss: 0.70150850, Time: 0.0211 Steps: 74580, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000075, Sample Num: 1200, Cur Loss: 0.18082023, Cur Avg Loss: 0.17034674, Log Avg loss: 0.22765872, Global Avg Loss: 0.70144498, Time: 0.0213 Steps: 74590, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000085, Sample Num: 1360, Cur Loss: 0.09468452, Cur Avg Loss: 0.17791313, Log Avg loss: 0.23466103, Global Avg Loss: 0.70138241, Time: 0.0212 Steps: 74600, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000095, Sample Num: 1520, Cur Loss: 0.26773375, Cur Avg Loss: 0.17584274, Log Avg loss: 0.15824440, Global Avg Loss: 0.70130961, Time: 0.0211 Steps: 74610, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000105, Sample Num: 1680, Cur Loss: 0.31137323, Cur Avg Loss: 0.17592988, Log Avg loss: 0.17675779, Global Avg Loss: 0.70123931, Time: 0.0212 Steps: 74620, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000115, Sample Num: 1840, Cur Loss: 0.17163186, Cur Avg Loss: 0.17386160, Log Avg loss: 0.15214466, Global Avg Loss: 0.70116574, Time: 0.0212 Steps: 74630, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000125, Sample Num: 2000, Cur Loss: 0.14583899, Cur Avg Loss: 0.17792476, Log Avg loss: 0.22465101, Global Avg Loss: 0.70110189, Time: 0.0212 Steps: 74640, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000135, Sample Num: 2160, Cur Loss: 0.08654539, Cur Avg Loss: 0.17774914, Log Avg loss: 0.17555398, Global Avg Loss: 0.70103149, Time: 0.0211 Steps: 74650, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000145, Sample Num: 2320, Cur Loss: 0.13993537, Cur Avg Loss: 0.17750043, Log Avg loss: 0.17414287, Global Avg Loss: 0.70096092, Time: 0.0212 Steps: 74660, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000155, Sample Num: 2480, Cur Loss: 0.11346927, Cur Avg Loss: 0.17337149, Log Avg loss: 0.11350172, Global Avg Loss: 0.70088225, Time: 0.0212 Steps: 74670, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000165, Sample Num: 2640, Cur Loss: 0.07045370, Cur Avg Loss: 0.16972325, Log Avg loss: 0.11317567, Global Avg Loss: 0.70080355, Time: 0.0212 Steps: 74680, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000175, Sample Num: 2800, Cur Loss: 0.19114934, Cur Avg Loss: 0.16667790, Log Avg loss: 0.11642960, Global Avg Loss: 0.70072531, Time: 0.0212 Steps: 74690, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000185, Sample Num: 2960, Cur Loss: 0.05911164, Cur Avg Loss: 0.16504125, Log Avg loss: 0.13639989, Global Avg Loss: 0.70064976, Time: 0.0212 Steps: 74700, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000195, Sample Num: 3120, Cur Loss: 0.11698297, Cur Avg Loss: 0.16230213, Log Avg loss: 0.11162839, Global Avg Loss: 0.70057092, Time: 0.0212 Steps: 74710, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000205, Sample Num: 3280, Cur Loss: 0.06613927, Cur Avg Loss: 0.16005650, Log Avg loss: 0.11626672, Global Avg Loss: 0.70049272, Time: 0.0212 Steps: 74720, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000215, Sample Num: 3440, Cur Loss: 0.05470499, Cur Avg Loss: 0.15910014, Log Avg loss: 0.13949477, Global Avg Loss: 0.70041765, Time: 0.0212 Steps: 74730, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000225, Sample Num: 3600, Cur Loss: 0.05527581, Cur Avg Loss: 0.15909636, Log Avg loss: 0.15901512, Global Avg Loss: 0.70034522, Time: 0.0211 Steps: 74740, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000235, Sample Num: 3760, Cur Loss: 0.18616623, Cur Avg Loss: 0.16123126, Log Avg loss: 0.20926634, Global Avg Loss: 0.70027952, Time: 0.0211 Steps: 74750, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000245, Sample Num: 3920, Cur Loss: 0.23017776, Cur Avg Loss: 0.16008247, Log Avg loss: 0.13308589, Global Avg Loss: 0.70020365, Time: 0.0212 Steps: 74760, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000255, Sample Num: 4080, Cur Loss: 0.25843543, Cur Avg Loss: 0.16026109, Log Avg loss: 0.16463734, Global Avg Loss: 0.70013202, Time: 0.0212 Steps: 74770, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000265, Sample Num: 4240, Cur Loss: 0.17459163, Cur Avg Loss: 0.16069897, Log Avg loss: 0.17186488, Global Avg Loss: 0.70006138, Time: 0.0211 Steps: 74780, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000275, Sample Num: 4400, Cur Loss: 0.08221406, Cur Avg Loss: 0.16009275, Log Avg loss: 0.14402796, Global Avg Loss: 0.69998703, Time: 0.0211 Steps: 74790, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000285, Sample Num: 4560, Cur Loss: 0.04894812, Cur Avg Loss: 0.15854014, Log Avg loss: 0.11584346, Global Avg Loss: 0.69990894, Time: 0.0211 Steps: 74800, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000295, Sample Num: 4720, Cur Loss: 0.17614666, Cur Avg Loss: 0.15756044, Log Avg loss: 0.12963886, Global Avg Loss: 0.69983271, Time: 0.0211 Steps: 74810, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000305, Sample Num: 4880, Cur Loss: 0.06334298, Cur Avg Loss: 0.15614395, Log Avg loss: 0.11435751, Global Avg Loss: 0.69975446, Time: 0.0211 Steps: 74820, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000315, Sample Num: 5040, Cur Loss: 0.24455434, Cur Avg Loss: 0.15485615, Log Avg loss: 0.11557829, Global Avg Loss: 0.69967639, Time: 0.0209 Steps: 74830, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000325, Sample Num: 5200, Cur Loss: 0.06621360, Cur Avg Loss: 0.15392316, Log Avg loss: 0.12453400, Global Avg Loss: 0.69959954, Time: 0.0209 Steps: 74840, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000335, Sample Num: 5360, Cur Loss: 0.03029274, Cur Avg Loss: 0.15346233, Log Avg loss: 0.13848532, Global Avg Loss: 0.69952458, Time: 0.0210 Steps: 74850, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000345, Sample Num: 5520, Cur Loss: 0.08255060, Cur Avg Loss: 0.15300672, Log Avg loss: 0.13774385, Global Avg Loss: 0.69944953, Time: 0.0208 Steps: 74860, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000355, Sample Num: 5680, Cur Loss: 0.08563238, Cur Avg Loss: 0.15272724, Log Avg loss: 0.14308517, Global Avg Loss: 0.69937522, Time: 0.0208 Steps: 74870, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000365, Sample Num: 5840, Cur Loss: 0.09066267, Cur Avg Loss: 0.15160342, Log Avg loss: 0.11170780, Global Avg Loss: 0.69929674, Time: 0.0208 Steps: 74880, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000375, Sample Num: 6000, Cur Loss: 0.15534541, Cur Avg Loss: 0.15203121, Log Avg loss: 0.16764571, Global Avg Loss: 0.69922575, Time: 0.0208 Steps: 74890, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000385, Sample Num: 6160, Cur Loss: 0.17908281, Cur Avg Loss: 0.15293480, Log Avg loss: 0.18681921, Global Avg Loss: 0.69915734, Time: 0.0208 Steps: 74900, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000395, Sample Num: 6320, Cur Loss: 0.14738293, Cur Avg Loss: 0.15538917, Log Avg loss: 0.24988252, Global Avg Loss: 0.69909736, Time: 0.0208 Steps: 74910, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000405, Sample Num: 6480, Cur Loss: 0.11007976, Cur Avg Loss: 0.15474015, Log Avg loss: 0.12910363, Global Avg Loss: 0.69902128, Time: 0.0208 Steps: 74920, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000415, Sample Num: 6640, Cur Loss: 0.14554279, Cur Avg Loss: 0.15381840, Log Avg loss: 0.11648761, Global Avg Loss: 0.69894354, Time: 0.0208 Steps: 74930, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000425, Sample Num: 6800, Cur Loss: 0.24443586, Cur Avg Loss: 0.15336701, Log Avg loss: 0.13463463, Global Avg Loss: 0.69886824, Time: 0.0208 Steps: 74940, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000435, Sample Num: 6960, Cur Loss: 0.13162315, Cur Avg Loss: 0.15343487, Log Avg loss: 0.15631891, Global Avg Loss: 0.69879585, Time: 0.0209 Steps: 74950, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000445, Sample Num: 7120, Cur Loss: 0.16704018, Cur Avg Loss: 0.15274722, Log Avg loss: 0.12283435, Global Avg Loss: 0.69871901, Time: 0.0208 Steps: 74960, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000455, Sample Num: 7280, Cur Loss: 0.31655210, Cur Avg Loss: 0.15342487, Log Avg loss: 0.18357998, Global Avg Loss: 0.69865030, Time: 0.0208 Steps: 74970, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000465, Sample Num: 7440, Cur Loss: 0.05330199, Cur Avg Loss: 0.15326322, Log Avg loss: 0.14590828, Global Avg Loss: 0.69857658, Time: 0.0208 Steps: 74980, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000475, Sample Num: 7600, Cur Loss: 0.09386327, Cur Avg Loss: 0.15506994, Log Avg loss: 0.23908227, Global Avg Loss: 0.69851531, Time: 0.0208 Steps: 74990, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000485, Sample Num: 7760, Cur Loss: 0.12882838, Cur Avg Loss: 0.15512036, Log Avg loss: 0.15751531, Global Avg Loss: 0.69844318, Time: 0.0208 Steps: 75000, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000495, Sample Num: 7920, Cur Loss: 0.12014069, Cur Avg Loss: 0.15617530, Log Avg loss: 0.20734009, Global Avg Loss: 0.69837770, Time: 0.0208 Steps: 75010, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000505, Sample Num: 8080, Cur Loss: 0.22510767, Cur Avg Loss: 0.15652560, Log Avg loss: 0.17386562, Global Avg Loss: 0.69830779, Time: 0.0208 Steps: 75020, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000515, Sample Num: 8240, Cur Loss: 0.13407077, Cur Avg Loss: 0.15687532, Log Avg loss: 0.17453585, Global Avg Loss: 0.69823798, Time: 0.0245 Steps: 75030, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000525, Sample Num: 8400, Cur Loss: 0.16870295, Cur Avg Loss: 0.15619324, Log Avg loss: 0.12106644, Global Avg Loss: 0.69816106, Time: 0.0209 Steps: 75040, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000535, Sample Num: 8560, Cur Loss: 0.19988187, Cur Avg Loss: 0.15571247, Log Avg loss: 0.13047189, Global Avg Loss: 0.69808542, Time: 0.0208 Steps: 75050, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000545, Sample Num: 8720, Cur Loss: 0.20209715, Cur Avg Loss: 0.15624334, Log Avg loss: 0.18464484, Global Avg Loss: 0.69801702, Time: 0.0208 Steps: 75060, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000555, Sample Num: 8880, Cur Loss: 0.13346380, Cur Avg Loss: 0.15598685, Log Avg loss: 0.14200796, Global Avg Loss: 0.69794295, Time: 0.0209 Steps: 75070, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000565, Sample Num: 9040, Cur Loss: 0.09872046, Cur Avg Loss: 0.15572905, Log Avg loss: 0.14142149, Global Avg Loss: 0.69786883, Time: 0.0208 Steps: 75080, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000575, Sample Num: 9200, Cur Loss: 0.06286467, Cur Avg Loss: 0.15518613, Log Avg loss: 0.12451125, Global Avg Loss: 0.69779247, Time: 0.0209 Steps: 75090, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000585, Sample Num: 9360, Cur Loss: 0.06821109, Cur Avg Loss: 0.15615355, Log Avg loss: 0.21178004, Global Avg Loss: 0.69772776, Time: 0.0209 Steps: 75100, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000595, Sample Num: 9520, Cur Loss: 0.12502077, Cur Avg Loss: 0.15664575, Log Avg loss: 0.18543915, Global Avg Loss: 0.69765955, Time: 0.0208 Steps: 75110, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000605, Sample Num: 9680, Cur Loss: 0.10315218, Cur Avg Loss: 0.15636797, Log Avg loss: 0.13984011, Global Avg Loss: 0.69758530, Time: 0.0209 Steps: 75120, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000615, Sample Num: 9840, Cur Loss: 0.07484901, Cur Avg Loss: 0.15562318, Log Avg loss: 0.11056355, Global Avg Loss: 0.69750716, Time: 0.0209 Steps: 75130, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000625, Sample Num: 10000, Cur Loss: 0.19950488, Cur Avg Loss: 0.15585455, Log Avg loss: 0.17008393, Global Avg Loss: 0.69743697, Time: 0.0208 Steps: 75140, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000635, Sample Num: 10160, Cur Loss: 0.09625152, Cur Avg Loss: 0.15541988, Log Avg loss: 0.12825274, Global Avg Loss: 0.69736123, Time: 0.0209 Steps: 75150, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000645, Sample Num: 10320, Cur Loss: 0.22212671, Cur Avg Loss: 0.15661298, Log Avg loss: 0.23237481, Global Avg Loss: 0.69729936, Time: 0.0208 Steps: 75160, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000655, Sample Num: 10480, Cur Loss: 0.19393766, Cur Avg Loss: 0.15676285, Log Avg loss: 0.16642947, Global Avg Loss: 0.69722874, Time: 0.0208 Steps: 75170, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000665, Sample Num: 10640, Cur Loss: 0.15116039, Cur Avg Loss: 0.15662837, Log Avg loss: 0.14782023, Global Avg Loss: 0.69715566, Time: 0.0209 Steps: 75180, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000675, Sample Num: 10800, Cur Loss: 0.20608670, Cur Avg Loss: 0.15667102, Log Avg loss: 0.15950701, Global Avg Loss: 0.69708416, Time: 0.0208 Steps: 75190, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000685, Sample Num: 10960, Cur Loss: 0.24042606, Cur Avg Loss: 0.15717901, Log Avg loss: 0.19146871, Global Avg Loss: 0.69701692, Time: 0.0209 Steps: 75200, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000695, Sample Num: 11120, Cur Loss: 0.11427743, Cur Avg Loss: 0.15658855, Log Avg loss: 0.11614162, Global Avg Loss: 0.69693969, Time: 0.0208 Steps: 75210, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000705, Sample Num: 11280, Cur Loss: 0.05417643, Cur Avg Loss: 0.15650392, Log Avg loss: 0.15062246, Global Avg Loss: 0.69686706, Time: 0.0208 Steps: 75220, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000715, Sample Num: 11440, Cur Loss: 0.27466977, Cur Avg Loss: 0.15723121, Log Avg loss: 0.20850501, Global Avg Loss: 0.69680214, Time: 0.0208 Steps: 75230, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000725, Sample Num: 11600, Cur Loss: 0.62751675, Cur Avg Loss: 0.15795642, Log Avg loss: 0.20980923, Global Avg Loss: 0.69673742, Time: 0.0209 Steps: 75240, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000735, Sample Num: 11760, Cur Loss: 0.06667177, Cur Avg Loss: 0.15734202, Log Avg loss: 0.11279800, Global Avg Loss: 0.69665982, Time: 0.0209 Steps: 75250, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000745, Sample Num: 11920, Cur Loss: 0.14831328, Cur Avg Loss: 0.15815986, Log Avg loss: 0.21827097, Global Avg Loss: 0.69659625, Time: 0.0209 Steps: 75260, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000755, Sample Num: 12080, Cur Loss: 0.03855995, Cur Avg Loss: 0.15826362, Log Avg loss: 0.16599330, Global Avg Loss: 0.69652576, Time: 0.0209 Steps: 75270, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000765, Sample Num: 12240, Cur Loss: 0.25790387, Cur Avg Loss: 0.15934096, Log Avg loss: 0.24068078, Global Avg Loss: 0.69646521, Time: 0.0209 Steps: 75280, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000775, Sample Num: 12400, Cur Loss: 0.28647465, Cur Avg Loss: 0.16015005, Log Avg loss: 0.22204531, Global Avg Loss: 0.69640219, Time: 0.0209 Steps: 75290, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000785, Sample Num: 12560, Cur Loss: 0.22719866, Cur Avg Loss: 0.16044207, Log Avg loss: 0.18307318, Global Avg Loss: 0.69633402, Time: 0.0208 Steps: 75300, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000795, Sample Num: 12720, Cur Loss: 0.13563004, Cur Avg Loss: 0.16132001, Log Avg loss: 0.23023819, Global Avg Loss: 0.69627213, Time: 0.0208 Steps: 75310, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000805, Sample Num: 12880, Cur Loss: 0.04764967, Cur Avg Loss: 0.16119327, Log Avg loss: 0.15111758, Global Avg Loss: 0.69619975, Time: 0.0209 Steps: 75320, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000815, Sample Num: 13040, Cur Loss: 0.22604910, Cur Avg Loss: 0.16096186, Log Avg loss: 0.14233378, Global Avg Loss: 0.69612623, Time: 0.0209 Steps: 75330, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000825, Sample Num: 13200, Cur Loss: 0.17046957, Cur Avg Loss: 0.16114002, Log Avg loss: 0.17565977, Global Avg Loss: 0.69605715, Time: 0.0209 Steps: 75340, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000835, Sample Num: 13360, Cur Loss: 0.09502624, Cur Avg Loss: 0.16194803, Log Avg loss: 0.22860876, Global Avg Loss: 0.69599511, Time: 0.0208 Steps: 75350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000845, Sample Num: 13520, Cur Loss: 0.28114426, Cur Avg Loss: 0.16300753, Log Avg loss: 0.25147580, Global Avg Loss: 0.69593612, Time: 0.0210 Steps: 75360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000855, Sample Num: 13680, Cur Loss: 0.14999470, Cur Avg Loss: 0.16292580, Log Avg loss: 0.15602011, Global Avg Loss: 0.69586449, Time: 0.0208 Steps: 75370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000865, Sample Num: 13840, Cur Loss: 0.19326864, Cur Avg Loss: 0.16386614, Log Avg loss: 0.24426497, Global Avg Loss: 0.69580458, Time: 0.0209 Steps: 75380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000875, Sample Num: 14000, Cur Loss: 0.23696141, Cur Avg Loss: 0.16374505, Log Avg loss: 0.15327031, Global Avg Loss: 0.69573261, Time: 0.0208 Steps: 75390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000885, Sample Num: 14160, Cur Loss: 0.34657732, Cur Avg Loss: 0.16419944, Log Avg loss: 0.20395884, Global Avg Loss: 0.69566739, Time: 0.0208 Steps: 75400, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000895, Sample Num: 14320, Cur Loss: 0.09764929, Cur Avg Loss: 0.16492244, Log Avg loss: 0.22890813, Global Avg Loss: 0.69560550, Time: 0.0208 Steps: 75410, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000905, Sample Num: 14480, Cur Loss: 0.13731955, Cur Avg Loss: 0.16452729, Log Avg loss: 0.12916149, Global Avg Loss: 0.69553039, Time: 0.0209 Steps: 75420, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000915, Sample Num: 14640, Cur Loss: 0.25109124, Cur Avg Loss: 0.16446088, Log Avg loss: 0.15845033, Global Avg Loss: 0.69545919, Time: 0.0208 Steps: 75430, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000925, Sample Num: 14800, Cur Loss: 0.10764241, Cur Avg Loss: 0.16412169, Log Avg loss: 0.13308621, Global Avg Loss: 0.69538464, Time: 0.0209 Steps: 75440, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000935, Sample Num: 14960, Cur Loss: 0.31322727, Cur Avg Loss: 0.16428361, Log Avg loss: 0.17926139, Global Avg Loss: 0.69531624, Time: 0.0208 Steps: 75450, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000945, Sample Num: 15120, Cur Loss: 0.08316150, Cur Avg Loss: 0.16397345, Log Avg loss: 0.13497287, Global Avg Loss: 0.69524198, Time: 0.0209 Steps: 75460, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000955, Sample Num: 15280, Cur Loss: 0.13087277, Cur Avg Loss: 0.16391807, Log Avg loss: 0.15868471, Global Avg Loss: 0.69517088, Time: 0.0208 Steps: 75470, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000965, Sample Num: 15440, Cur Loss: 0.10848982, Cur Avg Loss: 0.16386619, Log Avg loss: 0.15891171, Global Avg Loss: 0.69509984, Time: 0.0208 Steps: 75480, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000975, Sample Num: 15600, Cur Loss: 0.38969508, Cur Avg Loss: 0.16430549, Log Avg loss: 0.20669828, Global Avg Loss: 0.69503514, Time: 0.0208 Steps: 75490, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000985, Sample Num: 15760, Cur Loss: 0.09333920, Cur Avg Loss: 0.16413392, Log Avg loss: 0.14740599, Global Avg Loss: 0.69496261, Time: 0.0208 Steps: 75500, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000995, Sample Num: 15920, Cur Loss: 0.15292922, Cur Avg Loss: 0.16391243, Log Avg loss: 0.14209563, Global Avg Loss: 0.69488939, Time: 0.0209 Steps: 75510, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001005, Sample Num: 16080, Cur Loss: 0.63608408, Cur Avg Loss: 0.16438205, Log Avg loss: 0.21110865, Global Avg Loss: 0.69482533, Time: 0.0208 Steps: 75520, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001015, Sample Num: 16240, Cur Loss: 0.12821047, Cur Avg Loss: 0.16405414, Log Avg loss: 0.13109982, Global Avg Loss: 0.69475069, Time: 0.0208 Steps: 75530, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001025, Sample Num: 16400, Cur Loss: 0.32934687, Cur Avg Loss: 0.16466296, Log Avg loss: 0.22645805, Global Avg Loss: 0.69468870, Time: 0.0245 Steps: 75540, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001035, Sample Num: 16560, Cur Loss: 0.10386553, Cur Avg Loss: 0.16454995, Log Avg loss: 0.15296601, Global Avg Loss: 0.69461700, Time: 0.0209 Steps: 75550, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001045, Sample Num: 16720, Cur Loss: 0.12539342, Cur Avg Loss: 0.16463682, Log Avg loss: 0.17362855, Global Avg Loss: 0.69454805, Time: 0.0209 Steps: 75560, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001055, Sample Num: 16880, Cur Loss: 0.05145181, Cur Avg Loss: 0.16403048, Log Avg loss: 0.10066790, Global Avg Loss: 0.69446946, Time: 0.0208 Steps: 75570, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001065, Sample Num: 17040, Cur Loss: 0.25505567, Cur Avg Loss: 0.16336180, Log Avg loss: 0.09281525, Global Avg Loss: 0.69438985, Time: 0.0208 Steps: 75580, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001075, Sample Num: 17200, Cur Loss: 0.09346294, Cur Avg Loss: 0.16321547, Log Avg loss: 0.14763140, Global Avg Loss: 0.69431752, Time: 0.0208 Steps: 75590, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001085, Sample Num: 17360, Cur Loss: 0.12686814, Cur Avg Loss: 0.16350290, Log Avg loss: 0.19440171, Global Avg Loss: 0.69425139, Time: 0.0208 Steps: 75600, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001095, Sample Num: 17520, Cur Loss: 0.10413449, Cur Avg Loss: 0.16366664, Log Avg loss: 0.18143310, Global Avg Loss: 0.69418357, Time: 0.0208 Steps: 75610, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001105, Sample Num: 17680, Cur Loss: 0.15131801, Cur Avg Loss: 0.16408301, Log Avg loss: 0.20967484, Global Avg Loss: 0.69411950, Time: 0.0208 Steps: 75620, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001115, Sample Num: 17840, Cur Loss: 0.26286072, Cur Avg Loss: 0.16390696, Log Avg loss: 0.14445321, Global Avg Loss: 0.69404682, Time: 0.0208 Steps: 75630, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001125, Sample Num: 18000, Cur Loss: 0.22819938, Cur Avg Loss: 0.16416288, Log Avg loss: 0.19269823, Global Avg Loss: 0.69398054, Time: 0.0210 Steps: 75640, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001135, Sample Num: 18160, Cur Loss: 0.03657363, Cur Avg Loss: 0.16387252, Log Avg loss: 0.13120769, Global Avg Loss: 0.69390615, Time: 0.0208 Steps: 75650, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001145, Sample Num: 18320, Cur Loss: 0.22903542, Cur Avg Loss: 0.16429041, Log Avg loss: 0.21172082, Global Avg Loss: 0.69384242, Time: 0.0208 Steps: 75660, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001155, Sample Num: 18480, Cur Loss: 0.08404137, Cur Avg Loss: 0.16380352, Log Avg loss: 0.10805470, Global Avg Loss: 0.69376500, Time: 0.0209 Steps: 75670, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001165, Sample Num: 18640, Cur Loss: 0.07811181, Cur Avg Loss: 0.16361048, Log Avg loss: 0.14131391, Global Avg Loss: 0.69369201, Time: 0.0208 Steps: 75680, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001175, Sample Num: 18800, Cur Loss: 0.16692078, Cur Avg Loss: 0.16329533, Log Avg loss: 0.12657993, Global Avg Loss: 0.69361708, Time: 0.0208 Steps: 75690, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001185, Sample Num: 18960, Cur Loss: 0.32731763, Cur Avg Loss: 0.16316501, Log Avg loss: 0.14785299, Global Avg Loss: 0.69354498, Time: 0.0208 Steps: 75700, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001195, Sample Num: 19120, Cur Loss: 0.13313226, Cur Avg Loss: 0.16263858, Log Avg loss: 0.10025656, Global Avg Loss: 0.69346662, Time: 0.0208 Steps: 75710, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001205, Sample Num: 19280, Cur Loss: 0.18121073, Cur Avg Loss: 0.16274722, Log Avg loss: 0.17572898, Global Avg Loss: 0.69339825, Time: 0.0209 Steps: 75720, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001215, Sample Num: 19440, Cur Loss: 0.08726221, Cur Avg Loss: 0.16240517, Log Avg loss: 0.12118816, Global Avg Loss: 0.69332269, Time: 0.0208 Steps: 75730, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001225, Sample Num: 19600, Cur Loss: 0.11616301, Cur Avg Loss: 0.16297564, Log Avg loss: 0.23228771, Global Avg Loss: 0.69326182, Time: 0.0208 Steps: 75740, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001235, Sample Num: 19760, Cur Loss: 0.27082109, Cur Avg Loss: 0.16334888, Log Avg loss: 0.20907193, Global Avg Loss: 0.69319790, Time: 0.0208 Steps: 75750, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001245, Sample Num: 19920, Cur Loss: 0.32606065, Cur Avg Loss: 0.16327455, Log Avg loss: 0.15409398, Global Avg Loss: 0.69312674, Time: 0.0209 Steps: 75760, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001255, Sample Num: 20080, Cur Loss: 0.04784904, Cur Avg Loss: 0.16366831, Log Avg loss: 0.21269232, Global Avg Loss: 0.69306333, Time: 0.0208 Steps: 75770, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001265, Sample Num: 20240, Cur Loss: 0.17465249, Cur Avg Loss: 0.16354820, Log Avg loss: 0.14847404, Global Avg Loss: 0.69299147, Time: 0.0208 Steps: 75780, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001275, Sample Num: 20400, Cur Loss: 0.55286688, Cur Avg Loss: 0.16415821, Log Avg loss: 0.24132372, Global Avg Loss: 0.69293187, Time: 0.0208 Steps: 75790, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001285, Sample Num: 20560, Cur Loss: 0.25572416, Cur Avg Loss: 0.16412597, Log Avg loss: 0.16001614, Global Avg Loss: 0.69286157, Time: 0.0246 Steps: 75800, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001295, Sample Num: 20720, Cur Loss: 0.16952115, Cur Avg Loss: 0.16408307, Log Avg loss: 0.15856976, Global Avg Loss: 0.69279109, Time: 0.0208 Steps: 75810, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001305, Sample Num: 20880, Cur Loss: 0.34737688, Cur Avg Loss: 0.16435953, Log Avg loss: 0.20016104, Global Avg Loss: 0.69272611, Time: 0.0209 Steps: 75820, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001315, Sample Num: 21040, Cur Loss: 0.13026746, Cur Avg Loss: 0.16477668, Log Avg loss: 0.21921534, Global Avg Loss: 0.69266367, Time: 0.0209 Steps: 75830, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001325, Sample Num: 21200, Cur Loss: 0.14603907, Cur Avg Loss: 0.16456795, Log Avg loss: 0.13712032, Global Avg Loss: 0.69259042, Time: 0.0212 Steps: 75840, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001335, Sample Num: 21360, Cur Loss: 0.42247051, Cur Avg Loss: 0.16441995, Log Avg loss: 0.14480879, Global Avg Loss: 0.69251820, Time: 0.0211 Steps: 75850, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001345, Sample Num: 21520, Cur Loss: 0.04628598, Cur Avg Loss: 0.16431916, Log Avg loss: 0.15086415, Global Avg Loss: 0.69244680, Time: 0.0209 Steps: 75860, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001355, Sample Num: 21680, Cur Loss: 0.06375042, Cur Avg Loss: 0.16434569, Log Avg loss: 0.16791440, Global Avg Loss: 0.69237766, Time: 0.0209 Steps: 75870, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001365, Sample Num: 21840, Cur Loss: 0.12506388, Cur Avg Loss: 0.16406539, Log Avg loss: 0.12608419, Global Avg Loss: 0.69230303, Time: 0.0209 Steps: 75880, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001375, Sample Num: 22000, Cur Loss: 0.07002760, Cur Avg Loss: 0.16407808, Log Avg loss: 0.16581097, Global Avg Loss: 0.69223366, Time: 0.0209 Steps: 75890, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001385, Sample Num: 22160, Cur Loss: 0.05277855, Cur Avg Loss: 0.16397491, Log Avg loss: 0.14978839, Global Avg Loss: 0.69216219, Time: 0.0209 Steps: 75900, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001395, Sample Num: 22320, Cur Loss: 0.19938543, Cur Avg Loss: 0.16386052, Log Avg loss: 0.14801788, Global Avg Loss: 0.69209051, Time: 0.0208 Steps: 75910, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001405, Sample Num: 22480, Cur Loss: 0.12918352, Cur Avg Loss: 0.16418841, Log Avg loss: 0.20992834, Global Avg Loss: 0.69202700, Time: 0.0209 Steps: 75920, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001415, Sample Num: 22640, Cur Loss: 0.17915541, Cur Avg Loss: 0.16413454, Log Avg loss: 0.15656666, Global Avg Loss: 0.69195648, Time: 0.0209 Steps: 75930, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001425, Sample Num: 22800, Cur Loss: 0.06642140, Cur Avg Loss: 0.16403443, Log Avg loss: 0.14986863, Global Avg Loss: 0.69188509, Time: 0.0209 Steps: 75940, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001435, Sample Num: 22960, Cur Loss: 0.19351742, Cur Avg Loss: 0.16422393, Log Avg loss: 0.19122831, Global Avg Loss: 0.69181917, Time: 0.0209 Steps: 75950, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001445, Sample Num: 23120, Cur Loss: 0.04359936, Cur Avg Loss: 0.16451086, Log Avg loss: 0.20568487, Global Avg Loss: 0.69175517, Time: 0.0209 Steps: 75960, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001455, Sample Num: 23280, Cur Loss: 0.10708550, Cur Avg Loss: 0.16464638, Log Avg loss: 0.18422901, Global Avg Loss: 0.69168837, Time: 0.0209 Steps: 75970, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001465, Sample Num: 23440, Cur Loss: 0.30321583, Cur Avg Loss: 0.16457819, Log Avg loss: 0.15465642, Global Avg Loss: 0.69161769, Time: 0.0209 Steps: 75980, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001475, Sample Num: 23600, Cur Loss: 0.06662820, Cur Avg Loss: 0.16447787, Log Avg loss: 0.14978125, Global Avg Loss: 0.69154638, Time: 0.0209 Steps: 75990, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001485, Sample Num: 23760, Cur Loss: 0.17897841, Cur Avg Loss: 0.16452868, Log Avg loss: 0.17202223, Global Avg Loss: 0.69147802, Time: 0.0209 Steps: 76000, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001495, Sample Num: 23920, Cur Loss: 0.18632144, Cur Avg Loss: 0.16431122, Log Avg loss: 0.13201921, Global Avg Loss: 0.69140442, Time: 0.0209 Steps: 76010, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001505, Sample Num: 24080, Cur Loss: 0.09011439, Cur Avg Loss: 0.16410783, Log Avg loss: 0.13370126, Global Avg Loss: 0.69133106, Time: 0.0209 Steps: 76020, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001515, Sample Num: 24240, Cur Loss: 0.10185511, Cur Avg Loss: 0.16415437, Log Avg loss: 0.17115766, Global Avg Loss: 0.69126264, Time: 0.0209 Steps: 76030, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001525, Sample Num: 24400, Cur Loss: 0.06147721, Cur Avg Loss: 0.16443850, Log Avg loss: 0.20748521, Global Avg Loss: 0.69119902, Time: 0.0209 Steps: 76040, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001535, Sample Num: 24560, Cur Loss: 0.08927984, Cur Avg Loss: 0.16395006, Log Avg loss: 0.08946210, Global Avg Loss: 0.69111990, Time: 0.0209 Steps: 76050, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001545, Sample Num: 24720, Cur Loss: 0.15235090, Cur Avg Loss: 0.16378172, Log Avg loss: 0.13794223, Global Avg Loss: 0.69104717, Time: 0.0209 Steps: 76060, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001555, Sample Num: 24880, Cur Loss: 0.15628934, Cur Avg Loss: 0.16358149, Log Avg loss: 0.13264541, Global Avg Loss: 0.69097376, Time: 0.0209 Steps: 76070, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001565, Sample Num: 25040, Cur Loss: 0.06293950, Cur Avg Loss: 0.16340157, Log Avg loss: 0.13542402, Global Avg Loss: 0.69090074, Time: 0.0209 Steps: 76080, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001575, Sample Num: 25200, Cur Loss: 0.07943390, Cur Avg Loss: 0.16308880, Log Avg loss: 0.11414040, Global Avg Loss: 0.69082494, Time: 0.0208 Steps: 76090, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001585, Sample Num: 25360, Cur Loss: 0.14648072, Cur Avg Loss: 0.16278736, Log Avg loss: 0.11531034, Global Avg Loss: 0.69074931, Time: 0.0209 Steps: 76100, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001595, Sample Num: 25520, Cur Loss: 0.17996642, Cur Avg Loss: 0.16283099, Log Avg loss: 0.16974599, Global Avg Loss: 0.69068086, Time: 0.0209 Steps: 76110, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001605, Sample Num: 25680, Cur Loss: 0.15225729, Cur Avg Loss: 0.16311121, Log Avg loss: 0.20780686, Global Avg Loss: 0.69061742, Time: 0.0209 Steps: 76120, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001615, Sample Num: 25840, Cur Loss: 0.09786762, Cur Avg Loss: 0.16282182, Log Avg loss: 0.11637497, Global Avg Loss: 0.69054199, Time: 0.0209 Steps: 76130, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001625, Sample Num: 26000, Cur Loss: 0.25256634, Cur Avg Loss: 0.16295992, Log Avg loss: 0.18526215, Global Avg Loss: 0.69047563, Time: 0.0209 Steps: 76140, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001635, Sample Num: 26160, Cur Loss: 0.09673107, Cur Avg Loss: 0.16287783, Log Avg loss: 0.14953882, Global Avg Loss: 0.69040460, Time: 0.0209 Steps: 76150, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001645, Sample Num: 26320, Cur Loss: 0.08178787, Cur Avg Loss: 0.16286420, Log Avg loss: 0.16063558, Global Avg Loss: 0.69033504, Time: 0.0209 Steps: 76160, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001655, Sample Num: 26480, Cur Loss: 0.26001477, Cur Avg Loss: 0.16284679, Log Avg loss: 0.15998262, Global Avg Loss: 0.69026541, Time: 0.0209 Steps: 76170, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001665, Sample Num: 26640, Cur Loss: 0.13431583, Cur Avg Loss: 0.16294597, Log Avg loss: 0.17936090, Global Avg Loss: 0.69019834, Time: 0.0208 Steps: 76180, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001675, Sample Num: 26800, Cur Loss: 0.18990061, Cur Avg Loss: 0.16313646, Log Avg loss: 0.19485277, Global Avg Loss: 0.69013333, Time: 0.0209 Steps: 76190, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001685, Sample Num: 26960, Cur Loss: 0.35589522, Cur Avg Loss: 0.16328429, Log Avg loss: 0.18804675, Global Avg Loss: 0.69006744, Time: 0.0209 Steps: 76200, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001695, Sample Num: 27120, Cur Loss: 0.28009242, Cur Avg Loss: 0.16397607, Log Avg loss: 0.28053982, Global Avg Loss: 0.69001370, Time: 0.0209 Steps: 76210, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001705, Sample Num: 27280, Cur Loss: 0.26238471, Cur Avg Loss: 0.16409546, Log Avg loss: 0.18433181, Global Avg Loss: 0.68994736, Time: 0.0209 Steps: 76220, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001715, Sample Num: 27440, Cur Loss: 0.18034211, Cur Avg Loss: 0.16444792, Log Avg loss: 0.22454247, Global Avg Loss: 0.68988630, Time: 0.0209 Steps: 76230, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001725, Sample Num: 27600, Cur Loss: 0.06139599, Cur Avg Loss: 0.16450131, Log Avg loss: 0.17365761, Global Avg Loss: 0.68981859, Time: 0.0209 Steps: 76240, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001735, Sample Num: 27760, Cur Loss: 0.08476612, Cur Avg Loss: 0.16432025, Log Avg loss: 0.13308793, Global Avg Loss: 0.68974558, Time: 0.0209 Steps: 76250, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001745, Sample Num: 27920, Cur Loss: 0.06336144, Cur Avg Loss: 0.16411765, Log Avg loss: 0.12896611, Global Avg Loss: 0.68967204, Time: 0.0209 Steps: 76260, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001755, Sample Num: 28080, Cur Loss: 0.18133831, Cur Avg Loss: 0.16422802, Log Avg loss: 0.18348743, Global Avg Loss: 0.68960568, Time: 0.0209 Steps: 76270, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001765, Sample Num: 28240, Cur Loss: 0.08056676, Cur Avg Loss: 0.16390249, Log Avg loss: 0.10677216, Global Avg Loss: 0.68952927, Time: 0.0209 Steps: 76280, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001775, Sample Num: 28400, Cur Loss: 0.04372907, Cur Avg Loss: 0.16410983, Log Avg loss: 0.20070517, Global Avg Loss: 0.68946520, Time: 0.0209 Steps: 76290, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001785, Sample Num: 28560, Cur Loss: 0.05146684, Cur Avg Loss: 0.16395485, Log Avg loss: 0.13644675, Global Avg Loss: 0.68939272, Time: 0.0209 Steps: 76300, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001795, Sample Num: 28720, Cur Loss: 0.08801377, Cur Avg Loss: 0.16368675, Log Avg loss: 0.11583112, Global Avg Loss: 0.68931755, Time: 0.0245 Steps: 76310, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001805, Sample Num: 28880, Cur Loss: 0.25564611, Cur Avg Loss: 0.16362730, Log Avg loss: 0.15295514, Global Avg Loss: 0.68924728, Time: 0.0209 Steps: 76320, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001815, Sample Num: 29040, Cur Loss: 0.14591196, Cur Avg Loss: 0.16352078, Log Avg loss: 0.14429516, Global Avg Loss: 0.68917588, Time: 0.0208 Steps: 76330, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001825, Sample Num: 29200, Cur Loss: 0.06268758, Cur Avg Loss: 0.16359366, Log Avg loss: 0.17682124, Global Avg Loss: 0.68910877, Time: 0.0209 Steps: 76340, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001835, Sample Num: 29360, Cur Loss: 0.16865134, Cur Avg Loss: 0.16364332, Log Avg loss: 0.17270631, Global Avg Loss: 0.68904113, Time: 0.0208 Steps: 76350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001845, Sample Num: 29520, Cur Loss: 0.31047821, Cur Avg Loss: 0.16418038, Log Avg loss: 0.26272950, Global Avg Loss: 0.68898530, Time: 0.0209 Steps: 76360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001855, Sample Num: 29680, Cur Loss: 0.15037230, Cur Avg Loss: 0.16467414, Log Avg loss: 0.25577323, Global Avg Loss: 0.68892858, Time: 0.0208 Steps: 76370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001865, Sample Num: 29840, Cur Loss: 0.13120802, Cur Avg Loss: 0.16482903, Log Avg loss: 0.19356238, Global Avg Loss: 0.68886372, Time: 0.0208 Steps: 76380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001875, Sample Num: 30000, Cur Loss: 0.09050836, Cur Avg Loss: 0.16496533, Log Avg loss: 0.19038398, Global Avg Loss: 0.68879847, Time: 0.0209 Steps: 76390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001885, Sample Num: 30160, Cur Loss: 0.15293019, Cur Avg Loss: 0.16513871, Log Avg loss: 0.19764763, Global Avg Loss: 0.68873418, Time: 0.0209 Steps: 76400, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001895, Sample Num: 30320, Cur Loss: 0.30548233, Cur Avg Loss: 0.16520237, Log Avg loss: 0.17720197, Global Avg Loss: 0.68866723, Time: 0.0208 Steps: 76410, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001905, Sample Num: 30480, Cur Loss: 0.08996974, Cur Avg Loss: 0.16521881, Log Avg loss: 0.16833422, Global Avg Loss: 0.68859914, Time: 0.0209 Steps: 76420, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001915, Sample Num: 30640, Cur Loss: 0.04640435, Cur Avg Loss: 0.16497685, Log Avg loss: 0.11888460, Global Avg Loss: 0.68852460, Time: 0.0209 Steps: 76430, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001925, Sample Num: 30800, Cur Loss: 0.18620914, Cur Avg Loss: 0.16499496, Log Avg loss: 0.16846188, Global Avg Loss: 0.68845657, Time: 0.0209 Steps: 76440, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001935, Sample Num: 30960, Cur Loss: 0.19762233, Cur Avg Loss: 0.16494268, Log Avg loss: 0.15487956, Global Avg Loss: 0.68838677, Time: 0.0209 Steps: 76450, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001945, Sample Num: 31120, Cur Loss: 0.06005451, Cur Avg Loss: 0.16471316, Log Avg loss: 0.12030032, Global Avg Loss: 0.68831248, Time: 0.0208 Steps: 76460, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001955, Sample Num: 31280, Cur Loss: 0.06853503, Cur Avg Loss: 0.16483564, Log Avg loss: 0.18865878, Global Avg Loss: 0.68824714, Time: 0.0209 Steps: 76470, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001965, Sample Num: 31440, Cur Loss: 0.33751005, Cur Avg Loss: 0.16505174, Log Avg loss: 0.20729859, Global Avg Loss: 0.68818425, Time: 0.0209 Steps: 76480, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001975, Sample Num: 31600, Cur Loss: 0.15461093, Cur Avg Loss: 0.16486034, Log Avg loss: 0.12725023, Global Avg Loss: 0.68811092, Time: 0.0209 Steps: 76490, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001985, Sample Num: 31760, Cur Loss: 0.10019868, Cur Avg Loss: 0.16484768, Log Avg loss: 0.16234750, Global Avg Loss: 0.68804219, Time: 0.0208 Steps: 76500, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001995, Sample Num: 31920, Cur Loss: 0.07791518, Cur Avg Loss: 0.16456923, Log Avg loss: 0.10929641, Global Avg Loss: 0.68796655, Time: 0.0209 Steps: 76510, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002005, Sample Num: 32080, Cur Loss: 0.18051462, Cur Avg Loss: 0.16463742, Log Avg loss: 0.17824121, Global Avg Loss: 0.68789993, Time: 0.0209 Steps: 76520, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002015, Sample Num: 32240, Cur Loss: 0.07415309, Cur Avg Loss: 0.16439559, Log Avg loss: 0.11590957, Global Avg Loss: 0.68782519, Time: 0.0208 Steps: 76530, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002025, Sample Num: 32400, Cur Loss: 0.23350555, Cur Avg Loss: 0.16450536, Log Avg loss: 0.18662465, Global Avg Loss: 0.68775971, Time: 0.0209 Steps: 76540, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002035, Sample Num: 32560, Cur Loss: 0.15772462, Cur Avg Loss: 0.16451949, Log Avg loss: 0.16737997, Global Avg Loss: 0.68769173, Time: 0.0208 Steps: 76550, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002045, Sample Num: 32720, Cur Loss: 0.13213806, Cur Avg Loss: 0.16429020, Log Avg loss: 0.11763004, Global Avg Loss: 0.68761727, Time: 0.0208 Steps: 76560, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002055, Sample Num: 32880, Cur Loss: 0.26142630, Cur Avg Loss: 0.16398720, Log Avg loss: 0.10202418, Global Avg Loss: 0.68754079, Time: 0.0209 Steps: 76570, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002065, Sample Num: 33040, Cur Loss: 0.14219528, Cur Avg Loss: 0.16382102, Log Avg loss: 0.12966942, Global Avg Loss: 0.68746794, Time: 0.0209 Steps: 76580, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002075, Sample Num: 33200, Cur Loss: 0.21911958, Cur Avg Loss: 0.16417411, Log Avg loss: 0.23708740, Global Avg Loss: 0.68740914, Time: 0.0209 Steps: 76590, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002085, Sample Num: 33360, Cur Loss: 0.05323101, Cur Avg Loss: 0.16425675, Log Avg loss: 0.18140600, Global Avg Loss: 0.68734308, Time: 0.0209 Steps: 76600, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002095, Sample Num: 33520, Cur Loss: 0.12782595, Cur Avg Loss: 0.16415902, Log Avg loss: 0.14378114, Global Avg Loss: 0.68727213, Time: 0.0209 Steps: 76610, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002105, Sample Num: 33680, Cur Loss: 0.28353444, Cur Avg Loss: 0.16435031, Log Avg loss: 0.20442620, Global Avg Loss: 0.68720911, Time: 0.0209 Steps: 76620, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002115, Sample Num: 33840, Cur Loss: 0.05424073, Cur Avg Loss: 0.16409678, Log Avg loss: 0.11072842, Global Avg Loss: 0.68713388, Time: 0.0209 Steps: 76630, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002125, Sample Num: 34000, Cur Loss: 0.04558067, Cur Avg Loss: 0.16385562, Log Avg loss: 0.11285123, Global Avg Loss: 0.68705895, Time: 0.0209 Steps: 76640, Updated lr: 0.000028 ***** Running evaluation checkpoint-76644 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-76644 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.700641, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.249631, "eval_total_loss": 175.490243, "eval_mae": 0.332797, "eval_mse": 0.249726, "eval_r2": 0.841258, "eval_sp_statistic": 0.924626, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.93377, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.169369, "test_total_loss": 85.023021, "test_mae": 0.283632, "test_mse": 0.169431, "test_r2": 0.890648, "test_sp_statistic": 0.917498, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947712, "test_ps_pvalue": 0.0, "lr": 2.826552868658132e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6870310853880923, "train_cur_epoch_loss": 348.80569545179605, "train_cur_epoch_avg_loss": 0.16383546052221515, "train_cur_epoch_time": 44.70064091682434, "train_cur_epoch_avg_time": 0.020996073704473622, "epoch": 36, "step": 76644} ################################################## Training, Epoch: 0037, Batch: 000006, Sample Num: 96, Cur Loss: 0.11415906, Cur Avg Loss: 0.18427880, Log Avg loss: 0.17181678, Global Avg Loss: 0.68699173, Time: 0.0208 Steps: 76650, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000016, Sample Num: 256, Cur Loss: 0.11542119, Cur Avg Loss: 0.16888250, Log Avg loss: 0.15964472, Global Avg Loss: 0.68692294, Time: 0.0208 Steps: 76660, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000026, Sample Num: 416, Cur Loss: 0.19562073, Cur Avg Loss: 0.16074888, Log Avg loss: 0.14773509, Global Avg Loss: 0.68685261, Time: 0.0207 Steps: 76670, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000036, Sample Num: 576, Cur Loss: 0.16552402, Cur Avg Loss: 0.17765292, Log Avg loss: 0.22160341, Global Avg Loss: 0.68679194, Time: 0.0207 Steps: 76680, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000046, Sample Num: 736, Cur Loss: 0.14244708, Cur Avg Loss: 0.17533788, Log Avg loss: 0.16700374, Global Avg Loss: 0.68672416, Time: 0.0207 Steps: 76690, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000056, Sample Num: 896, Cur Loss: 0.05226123, Cur Avg Loss: 0.17426317, Log Avg loss: 0.16931949, Global Avg Loss: 0.68665670, Time: 0.0208 Steps: 76700, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000066, Sample Num: 1056, Cur Loss: 0.09179581, Cur Avg Loss: 0.16942289, Log Avg loss: 0.14231736, Global Avg Loss: 0.68658574, Time: 0.0208 Steps: 76710, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000076, Sample Num: 1216, Cur Loss: 0.14903873, Cur Avg Loss: 0.15989327, Log Avg loss: 0.09699778, Global Avg Loss: 0.68650889, Time: 0.0208 Steps: 76720, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000086, Sample Num: 1376, Cur Loss: 0.05652802, Cur Avg Loss: 0.15631881, Log Avg loss: 0.12915292, Global Avg Loss: 0.68643626, Time: 0.0208 Steps: 76730, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000096, Sample Num: 1536, Cur Loss: 0.08097385, Cur Avg Loss: 0.15530841, Log Avg loss: 0.14661897, Global Avg Loss: 0.68636591, Time: 0.0207 Steps: 76740, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000106, Sample Num: 1696, Cur Loss: 0.07687251, Cur Avg Loss: 0.15588046, Log Avg loss: 0.16137213, Global Avg Loss: 0.68629751, Time: 0.0207 Steps: 76750, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000116, Sample Num: 1856, Cur Loss: 0.09418669, Cur Avg Loss: 0.15768097, Log Avg loss: 0.17676632, Global Avg Loss: 0.68623113, Time: 0.0207 Steps: 76760, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000126, Sample Num: 2016, Cur Loss: 0.22643104, Cur Avg Loss: 0.16159459, Log Avg loss: 0.20699267, Global Avg Loss: 0.68616870, Time: 0.0208 Steps: 76770, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000136, Sample Num: 2176, Cur Loss: 0.20928191, Cur Avg Loss: 0.15893113, Log Avg loss: 0.12537151, Global Avg Loss: 0.68609566, Time: 0.0207 Steps: 76780, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000146, Sample Num: 2336, Cur Loss: 0.46023434, Cur Avg Loss: 0.16240665, Log Avg loss: 0.20967363, Global Avg Loss: 0.68603362, Time: 0.0207 Steps: 76790, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000156, Sample Num: 2496, Cur Loss: 0.09097502, Cur Avg Loss: 0.15918616, Log Avg loss: 0.11216706, Global Avg Loss: 0.68595890, Time: 0.0208 Steps: 76800, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000166, Sample Num: 2656, Cur Loss: 0.15601030, Cur Avg Loss: 0.15866578, Log Avg loss: 0.15054779, Global Avg Loss: 0.68588919, Time: 0.0208 Steps: 76810, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000176, Sample Num: 2816, Cur Loss: 0.11631534, Cur Avg Loss: 0.15668294, Log Avg loss: 0.12376789, Global Avg Loss: 0.68581602, Time: 0.0208 Steps: 76820, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000186, Sample Num: 2976, Cur Loss: 0.13403104, Cur Avg Loss: 0.15869684, Log Avg loss: 0.19414147, Global Avg Loss: 0.68575203, Time: 0.0208 Steps: 76830, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000196, Sample Num: 3136, Cur Loss: 0.13151842, Cur Avg Loss: 0.15884190, Log Avg loss: 0.16154006, Global Avg Loss: 0.68568380, Time: 0.0207 Steps: 76840, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000206, Sample Num: 3296, Cur Loss: 0.16869298, Cur Avg Loss: 0.15809413, Log Avg loss: 0.14343782, Global Avg Loss: 0.68561325, Time: 0.0207 Steps: 76850, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000216, Sample Num: 3456, Cur Loss: 0.11578753, Cur Avg Loss: 0.15945530, Log Avg loss: 0.18749545, Global Avg Loss: 0.68554844, Time: 0.0208 Steps: 76860, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000226, Sample Num: 3616, Cur Loss: 0.28593263, Cur Avg Loss: 0.16001771, Log Avg loss: 0.17216568, Global Avg Loss: 0.68548165, Time: 0.0207 Steps: 76870, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000236, Sample Num: 3776, Cur Loss: 0.15269914, Cur Avg Loss: 0.15906233, Log Avg loss: 0.13747084, Global Avg Loss: 0.68541037, Time: 0.0208 Steps: 76880, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000246, Sample Num: 3936, Cur Loss: 0.09274039, Cur Avg Loss: 0.15697842, Log Avg loss: 0.10779796, Global Avg Loss: 0.68533525, Time: 0.0208 Steps: 76890, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000256, Sample Num: 4096, Cur Loss: 0.04186586, Cur Avg Loss: 0.15598537, Log Avg loss: 0.13155650, Global Avg Loss: 0.68526323, Time: 0.0254 Steps: 76900, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000266, Sample Num: 4256, Cur Loss: 0.03407548, Cur Avg Loss: 0.15521975, Log Avg loss: 0.13561969, Global Avg Loss: 0.68519177, Time: 0.0208 Steps: 76910, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000276, Sample Num: 4416, Cur Loss: 0.41540217, Cur Avg Loss: 0.15643121, Log Avg loss: 0.18865609, Global Avg Loss: 0.68512722, Time: 0.0208 Steps: 76920, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000286, Sample Num: 4576, Cur Loss: 0.33965492, Cur Avg Loss: 0.15699010, Log Avg loss: 0.17241546, Global Avg Loss: 0.68506057, Time: 0.0208 Steps: 76930, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000296, Sample Num: 4736, Cur Loss: 0.08858199, Cur Avg Loss: 0.15555480, Log Avg loss: 0.11450544, Global Avg Loss: 0.68498641, Time: 0.0207 Steps: 76940, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000306, Sample Num: 4896, Cur Loss: 0.14497545, Cur Avg Loss: 0.15714548, Log Avg loss: 0.20422951, Global Avg Loss: 0.68492394, Time: 0.0208 Steps: 76950, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000316, Sample Num: 5056, Cur Loss: 0.06226630, Cur Avg Loss: 0.15603419, Log Avg loss: 0.12202856, Global Avg Loss: 0.68485080, Time: 0.0208 Steps: 76960, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000326, Sample Num: 5216, Cur Loss: 0.07669494, Cur Avg Loss: 0.15604329, Log Avg loss: 0.15633092, Global Avg Loss: 0.68478213, Time: 0.0208 Steps: 76970, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000336, Sample Num: 5376, Cur Loss: 0.14166938, Cur Avg Loss: 0.15543712, Log Avg loss: 0.13567595, Global Avg Loss: 0.68471080, Time: 0.0207 Steps: 76980, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000346, Sample Num: 5536, Cur Loss: 0.28985125, Cur Avg Loss: 0.15659495, Log Avg loss: 0.19549809, Global Avg Loss: 0.68464726, Time: 0.0208 Steps: 76990, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000356, Sample Num: 5696, Cur Loss: 0.05132276, Cur Avg Loss: 0.15545222, Log Avg loss: 0.11591366, Global Avg Loss: 0.68457340, Time: 0.0208 Steps: 77000, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000366, Sample Num: 5856, Cur Loss: 0.11081878, Cur Avg Loss: 0.15411562, Log Avg loss: 0.10653276, Global Avg Loss: 0.68449834, Time: 0.0208 Steps: 77010, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000376, Sample Num: 6016, Cur Loss: 0.11423788, Cur Avg Loss: 0.15365414, Log Avg loss: 0.13676406, Global Avg Loss: 0.68442722, Time: 0.0208 Steps: 77020, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000386, Sample Num: 6176, Cur Loss: 0.20116645, Cur Avg Loss: 0.15432005, Log Avg loss: 0.17935813, Global Avg Loss: 0.68436165, Time: 0.0208 Steps: 77030, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000396, Sample Num: 6336, Cur Loss: 0.14100729, Cur Avg Loss: 0.15327142, Log Avg loss: 0.11279448, Global Avg Loss: 0.68428746, Time: 0.0208 Steps: 77040, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000406, Sample Num: 6496, Cur Loss: 0.09291872, Cur Avg Loss: 0.15418622, Log Avg loss: 0.19041212, Global Avg Loss: 0.68422336, Time: 0.0208 Steps: 77050, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000416, Sample Num: 6656, Cur Loss: 0.19942445, Cur Avg Loss: 0.15453287, Log Avg loss: 0.16860702, Global Avg Loss: 0.68415645, Time: 0.0208 Steps: 77060, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000426, Sample Num: 6816, Cur Loss: 0.19588861, Cur Avg Loss: 0.15627423, Log Avg loss: 0.22871491, Global Avg Loss: 0.68409736, Time: 0.0208 Steps: 77070, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000436, Sample Num: 6976, Cur Loss: 0.12964511, Cur Avg Loss: 0.15588723, Log Avg loss: 0.13940091, Global Avg Loss: 0.68402669, Time: 0.0208 Steps: 77080, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000446, Sample Num: 7136, Cur Loss: 0.57486808, Cur Avg Loss: 0.15669412, Log Avg loss: 0.19187443, Global Avg Loss: 0.68396285, Time: 0.0208 Steps: 77090, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000456, Sample Num: 7296, Cur Loss: 0.16507560, Cur Avg Loss: 0.15553266, Log Avg loss: 0.10373172, Global Avg Loss: 0.68388759, Time: 0.0208 Steps: 77100, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000466, Sample Num: 7456, Cur Loss: 0.11923847, Cur Avg Loss: 0.15421678, Log Avg loss: 0.09421272, Global Avg Loss: 0.68381112, Time: 0.0207 Steps: 77110, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000476, Sample Num: 7616, Cur Loss: 0.11450400, Cur Avg Loss: 0.15338859, Log Avg loss: 0.11479463, Global Avg Loss: 0.68373734, Time: 0.0208 Steps: 77120, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000486, Sample Num: 7776, Cur Loss: 0.07098290, Cur Avg Loss: 0.15288992, Log Avg loss: 0.12915348, Global Avg Loss: 0.68366544, Time: 0.0208 Steps: 77130, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000496, Sample Num: 7936, Cur Loss: 0.05693954, Cur Avg Loss: 0.15275328, Log Avg loss: 0.14611245, Global Avg Loss: 0.68359575, Time: 0.0208 Steps: 77140, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000506, Sample Num: 8096, Cur Loss: 0.17660052, Cur Avg Loss: 0.15321194, Log Avg loss: 0.17596160, Global Avg Loss: 0.68352995, Time: 0.0208 Steps: 77150, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000516, Sample Num: 8256, Cur Loss: 0.08501908, Cur Avg Loss: 0.15284819, Log Avg loss: 0.13444206, Global Avg Loss: 0.68345879, Time: 0.0209 Steps: 77160, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000526, Sample Num: 8416, Cur Loss: 0.12639600, Cur Avg Loss: 0.15330221, Log Avg loss: 0.17672987, Global Avg Loss: 0.68339313, Time: 0.0208 Steps: 77170, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000536, Sample Num: 8576, Cur Loss: 0.09887376, Cur Avg Loss: 0.15284577, Log Avg loss: 0.12883701, Global Avg Loss: 0.68332127, Time: 0.0208 Steps: 77180, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000546, Sample Num: 8736, Cur Loss: 0.07508801, Cur Avg Loss: 0.15225119, Log Avg loss: 0.12038198, Global Avg Loss: 0.68324834, Time: 0.0208 Steps: 77190, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000556, Sample Num: 8896, Cur Loss: 0.07246791, Cur Avg Loss: 0.15127413, Log Avg loss: 0.09792614, Global Avg Loss: 0.68317252, Time: 0.0208 Steps: 77200, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000566, Sample Num: 9056, Cur Loss: 0.29183817, Cur Avg Loss: 0.15127821, Log Avg loss: 0.15150549, Global Avg Loss: 0.68310367, Time: 0.0208 Steps: 77210, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000576, Sample Num: 9216, Cur Loss: 0.04859097, Cur Avg Loss: 0.15147046, Log Avg loss: 0.16235150, Global Avg Loss: 0.68303623, Time: 0.0208 Steps: 77220, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000586, Sample Num: 9376, Cur Loss: 0.20930888, Cur Avg Loss: 0.15144545, Log Avg loss: 0.15000519, Global Avg Loss: 0.68296721, Time: 0.0207 Steps: 77230, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000596, Sample Num: 9536, Cur Loss: 0.23536134, Cur Avg Loss: 0.15128122, Log Avg loss: 0.14165700, Global Avg Loss: 0.68289713, Time: 0.0208 Steps: 77240, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000606, Sample Num: 9696, Cur Loss: 0.07299718, Cur Avg Loss: 0.15099742, Log Avg loss: 0.13408292, Global Avg Loss: 0.68282608, Time: 0.0208 Steps: 77250, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000616, Sample Num: 9856, Cur Loss: 0.10697849, Cur Avg Loss: 0.15039214, Log Avg loss: 0.11371234, Global Avg Loss: 0.68275242, Time: 0.0208 Steps: 77260, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000626, Sample Num: 10016, Cur Loss: 0.24625963, Cur Avg Loss: 0.15023256, Log Avg loss: 0.14040260, Global Avg Loss: 0.68268223, Time: 0.0208 Steps: 77270, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000636, Sample Num: 10176, Cur Loss: 0.36258399, Cur Avg Loss: 0.15171456, Log Avg loss: 0.24448741, Global Avg Loss: 0.68262553, Time: 0.0208 Steps: 77280, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000646, Sample Num: 10336, Cur Loss: 0.14236122, Cur Avg Loss: 0.15142369, Log Avg loss: 0.13292462, Global Avg Loss: 0.68255441, Time: 0.0208 Steps: 77290, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000656, Sample Num: 10496, Cur Loss: 0.25938103, Cur Avg Loss: 0.15167043, Log Avg loss: 0.16760992, Global Avg Loss: 0.68248779, Time: 0.0207 Steps: 77300, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000666, Sample Num: 10656, Cur Loss: 0.07300366, Cur Avg Loss: 0.15190778, Log Avg loss: 0.16747749, Global Avg Loss: 0.68242118, Time: 0.0208 Steps: 77310, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000676, Sample Num: 10816, Cur Loss: 0.08800165, Cur Avg Loss: 0.15161770, Log Avg loss: 0.13229892, Global Avg Loss: 0.68235003, Time: 0.0208 Steps: 77320, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000686, Sample Num: 10976, Cur Loss: 0.04673040, Cur Avg Loss: 0.15152168, Log Avg loss: 0.14503030, Global Avg Loss: 0.68228054, Time: 0.0208 Steps: 77330, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000696, Sample Num: 11136, Cur Loss: 0.14789692, Cur Avg Loss: 0.15089882, Log Avg loss: 0.10817112, Global Avg Loss: 0.68220631, Time: 0.0208 Steps: 77340, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000706, Sample Num: 11296, Cur Loss: 0.13154826, Cur Avg Loss: 0.15082565, Log Avg loss: 0.14573236, Global Avg Loss: 0.68213695, Time: 0.0208 Steps: 77350, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000716, Sample Num: 11456, Cur Loss: 0.12232509, Cur Avg Loss: 0.15041180, Log Avg loss: 0.12119405, Global Avg Loss: 0.68206444, Time: 0.0208 Steps: 77360, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000726, Sample Num: 11616, Cur Loss: 0.18167879, Cur Avg Loss: 0.15034304, Log Avg loss: 0.14542033, Global Avg Loss: 0.68199508, Time: 0.0208 Steps: 77370, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000736, Sample Num: 11776, Cur Loss: 0.47415656, Cur Avg Loss: 0.15124167, Log Avg loss: 0.21648184, Global Avg Loss: 0.68193492, Time: 0.0207 Steps: 77380, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000746, Sample Num: 11936, Cur Loss: 0.04813565, Cur Avg Loss: 0.15160917, Log Avg loss: 0.17865766, Global Avg Loss: 0.68186989, Time: 0.0208 Steps: 77390, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000756, Sample Num: 12096, Cur Loss: 0.10181037, Cur Avg Loss: 0.15152094, Log Avg loss: 0.14493852, Global Avg Loss: 0.68180052, Time: 0.0208 Steps: 77400, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000766, Sample Num: 12256, Cur Loss: 0.24489820, Cur Avg Loss: 0.15178151, Log Avg loss: 0.17148040, Global Avg Loss: 0.68173460, Time: 0.0208 Steps: 77410, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000776, Sample Num: 12416, Cur Loss: 0.20517515, Cur Avg Loss: 0.15234638, Log Avg loss: 0.19561559, Global Avg Loss: 0.68167181, Time: 0.0208 Steps: 77420, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000786, Sample Num: 12576, Cur Loss: 0.24161462, Cur Avg Loss: 0.15247643, Log Avg loss: 0.16256854, Global Avg Loss: 0.68160477, Time: 0.0211 Steps: 77430, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000796, Sample Num: 12736, Cur Loss: 0.10549213, Cur Avg Loss: 0.15274557, Log Avg loss: 0.17389981, Global Avg Loss: 0.68153920, Time: 0.0208 Steps: 77440, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000806, Sample Num: 12896, Cur Loss: 0.09399701, Cur Avg Loss: 0.15309303, Log Avg loss: 0.18075092, Global Avg Loss: 0.68147454, Time: 0.0209 Steps: 77450, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000816, Sample Num: 13056, Cur Loss: 0.08695944, Cur Avg Loss: 0.15266786, Log Avg loss: 0.11839882, Global Avg Loss: 0.68140185, Time: 0.0208 Steps: 77460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000826, Sample Num: 13216, Cur Loss: 0.11940986, Cur Avg Loss: 0.15247975, Log Avg loss: 0.13713039, Global Avg Loss: 0.68133160, Time: 0.0207 Steps: 77470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000836, Sample Num: 13376, Cur Loss: 0.09516969, Cur Avg Loss: 0.15258581, Log Avg loss: 0.16134641, Global Avg Loss: 0.68126448, Time: 0.0208 Steps: 77480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000846, Sample Num: 13536, Cur Loss: 0.29034418, Cur Avg Loss: 0.15278609, Log Avg loss: 0.16952934, Global Avg Loss: 0.68119845, Time: 0.0209 Steps: 77490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000856, Sample Num: 13696, Cur Loss: 0.09956752, Cur Avg Loss: 0.15284428, Log Avg loss: 0.15776683, Global Avg Loss: 0.68113091, Time: 0.0208 Steps: 77500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000866, Sample Num: 13856, Cur Loss: 0.24290913, Cur Avg Loss: 0.15262223, Log Avg loss: 0.13361550, Global Avg Loss: 0.68106027, Time: 0.0209 Steps: 77510, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000876, Sample Num: 14016, Cur Loss: 0.05403211, Cur Avg Loss: 0.15235801, Log Avg loss: 0.12947600, Global Avg Loss: 0.68098911, Time: 0.0208 Steps: 77520, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000886, Sample Num: 14176, Cur Loss: 0.23183724, Cur Avg Loss: 0.15251407, Log Avg loss: 0.16618475, Global Avg Loss: 0.68092271, Time: 0.0208 Steps: 77530, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000896, Sample Num: 14336, Cur Loss: 0.30425477, Cur Avg Loss: 0.15244952, Log Avg loss: 0.14673094, Global Avg Loss: 0.68085382, Time: 0.0209 Steps: 77540, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000906, Sample Num: 14496, Cur Loss: 0.39829594, Cur Avg Loss: 0.15307883, Log Avg loss: 0.20946517, Global Avg Loss: 0.68079304, Time: 0.0208 Steps: 77550, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000916, Sample Num: 14656, Cur Loss: 0.02802387, Cur Avg Loss: 0.15298422, Log Avg loss: 0.14441186, Global Avg Loss: 0.68072388, Time: 0.0209 Steps: 77560, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000926, Sample Num: 14816, Cur Loss: 0.16162644, Cur Avg Loss: 0.15303855, Log Avg loss: 0.15801541, Global Avg Loss: 0.68065649, Time: 0.0207 Steps: 77570, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000936, Sample Num: 14976, Cur Loss: 0.22241214, Cur Avg Loss: 0.15290878, Log Avg loss: 0.14089202, Global Avg Loss: 0.68058692, Time: 0.0208 Steps: 77580, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000946, Sample Num: 15136, Cur Loss: 0.28700662, Cur Avg Loss: 0.15294230, Log Avg loss: 0.15607953, Global Avg Loss: 0.68051932, Time: 0.0212 Steps: 77590, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000956, Sample Num: 15296, Cur Loss: 0.19574171, Cur Avg Loss: 0.15280757, Log Avg loss: 0.14006289, Global Avg Loss: 0.68044967, Time: 0.0207 Steps: 77600, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000966, Sample Num: 15456, Cur Loss: 0.07700056, Cur Avg Loss: 0.15300534, Log Avg loss: 0.17191133, Global Avg Loss: 0.68038415, Time: 0.0207 Steps: 77610, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000976, Sample Num: 15616, Cur Loss: 0.22148748, Cur Avg Loss: 0.15298373, Log Avg loss: 0.15089688, Global Avg Loss: 0.68031593, Time: 0.0208 Steps: 77620, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000986, Sample Num: 15776, Cur Loss: 0.38424703, Cur Avg Loss: 0.15351508, Log Avg loss: 0.20537463, Global Avg Loss: 0.68025475, Time: 0.0208 Steps: 77630, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000996, Sample Num: 15936, Cur Loss: 0.13786444, Cur Avg Loss: 0.15331475, Log Avg loss: 0.13356187, Global Avg Loss: 0.68018434, Time: 0.0208 Steps: 77640, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001006, Sample Num: 16096, Cur Loss: 0.11628383, Cur Avg Loss: 0.15345367, Log Avg loss: 0.16729027, Global Avg Loss: 0.68011829, Time: 0.0208 Steps: 77650, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001016, Sample Num: 16256, Cur Loss: 0.25870919, Cur Avg Loss: 0.15319857, Log Avg loss: 0.12753596, Global Avg Loss: 0.68004713, Time: 0.0208 Steps: 77660, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001026, Sample Num: 16416, Cur Loss: 0.08219508, Cur Avg Loss: 0.15258412, Log Avg loss: 0.09015567, Global Avg Loss: 0.67997118, Time: 0.0246 Steps: 77670, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001036, Sample Num: 16576, Cur Loss: 0.04619345, Cur Avg Loss: 0.15251879, Log Avg loss: 0.14581624, Global Avg Loss: 0.67990242, Time: 0.0209 Steps: 77680, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001046, Sample Num: 16736, Cur Loss: 0.42608401, Cur Avg Loss: 0.15304041, Log Avg loss: 0.20708045, Global Avg Loss: 0.67984156, Time: 0.0207 Steps: 77690, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001056, Sample Num: 16896, Cur Loss: 0.16036490, Cur Avg Loss: 0.15313819, Log Avg loss: 0.16336498, Global Avg Loss: 0.67977509, Time: 0.0208 Steps: 77700, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001066, Sample Num: 17056, Cur Loss: 0.05509592, Cur Avg Loss: 0.15325764, Log Avg loss: 0.16587225, Global Avg Loss: 0.67970896, Time: 0.0208 Steps: 77710, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001076, Sample Num: 17216, Cur Loss: 0.11117131, Cur Avg Loss: 0.15313031, Log Avg loss: 0.13955679, Global Avg Loss: 0.67963946, Time: 0.0208 Steps: 77720, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001086, Sample Num: 17376, Cur Loss: 0.02356699, Cur Avg Loss: 0.15296945, Log Avg loss: 0.13566072, Global Avg Loss: 0.67956948, Time: 0.0208 Steps: 77730, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001096, Sample Num: 17536, Cur Loss: 0.14801231, Cur Avg Loss: 0.15335318, Log Avg loss: 0.19502642, Global Avg Loss: 0.67950715, Time: 0.0208 Steps: 77740, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001106, Sample Num: 17696, Cur Loss: 0.03740928, Cur Avg Loss: 0.15324826, Log Avg loss: 0.14174932, Global Avg Loss: 0.67943798, Time: 0.0208 Steps: 77750, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001116, Sample Num: 17856, Cur Loss: 0.06040201, Cur Avg Loss: 0.15315633, Log Avg loss: 0.14298802, Global Avg Loss: 0.67936899, Time: 0.0209 Steps: 77760, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001126, Sample Num: 18016, Cur Loss: 0.04726060, Cur Avg Loss: 0.15329612, Log Avg loss: 0.16889735, Global Avg Loss: 0.67930336, Time: 0.0209 Steps: 77770, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001136, Sample Num: 18176, Cur Loss: 0.06724117, Cur Avg Loss: 0.15312681, Log Avg loss: 0.13406184, Global Avg Loss: 0.67923325, Time: 0.0209 Steps: 77780, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001146, Sample Num: 18336, Cur Loss: 0.15465027, Cur Avg Loss: 0.15299511, Log Avg loss: 0.13803491, Global Avg Loss: 0.67916368, Time: 0.0209 Steps: 77790, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001156, Sample Num: 18496, Cur Loss: 0.68903565, Cur Avg Loss: 0.15381116, Log Avg loss: 0.24733026, Global Avg Loss: 0.67910818, Time: 0.0209 Steps: 77800, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001166, Sample Num: 18656, Cur Loss: 0.11239044, Cur Avg Loss: 0.15436601, Log Avg loss: 0.21850681, Global Avg Loss: 0.67904898, Time: 0.0209 Steps: 77810, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001176, Sample Num: 18816, Cur Loss: 0.11587033, Cur Avg Loss: 0.15431126, Log Avg loss: 0.14792703, Global Avg Loss: 0.67898073, Time: 0.0208 Steps: 77820, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001186, Sample Num: 18976, Cur Loss: 0.23445952, Cur Avg Loss: 0.15418477, Log Avg loss: 0.13930977, Global Avg Loss: 0.67891139, Time: 0.0209 Steps: 77830, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001196, Sample Num: 19136, Cur Loss: 0.27661926, Cur Avg Loss: 0.15388807, Log Avg loss: 0.11869942, Global Avg Loss: 0.67883942, Time: 0.0209 Steps: 77840, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001206, Sample Num: 19296, Cur Loss: 0.08738966, Cur Avg Loss: 0.15386667, Log Avg loss: 0.15130655, Global Avg Loss: 0.67877166, Time: 0.0210 Steps: 77850, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001216, Sample Num: 19456, Cur Loss: 0.33388090, Cur Avg Loss: 0.15409561, Log Avg loss: 0.18170660, Global Avg Loss: 0.67870782, Time: 0.0209 Steps: 77860, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001226, Sample Num: 19616, Cur Loss: 0.02948545, Cur Avg Loss: 0.15427697, Log Avg loss: 0.17633014, Global Avg Loss: 0.67864330, Time: 0.0209 Steps: 77870, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001236, Sample Num: 19776, Cur Loss: 0.21538304, Cur Avg Loss: 0.15394054, Log Avg loss: 0.11269457, Global Avg Loss: 0.67857063, Time: 0.0208 Steps: 77880, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001246, Sample Num: 19936, Cur Loss: 0.14712200, Cur Avg Loss: 0.15396362, Log Avg loss: 0.15681581, Global Avg Loss: 0.67850365, Time: 0.0209 Steps: 77890, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001256, Sample Num: 20096, Cur Loss: 0.10621235, Cur Avg Loss: 0.15371107, Log Avg loss: 0.12224404, Global Avg Loss: 0.67843224, Time: 0.0209 Steps: 77900, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001266, Sample Num: 20256, Cur Loss: 0.08469737, Cur Avg Loss: 0.15374099, Log Avg loss: 0.15749782, Global Avg Loss: 0.67836538, Time: 0.0209 Steps: 77910, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001276, Sample Num: 20416, Cur Loss: 0.15531327, Cur Avg Loss: 0.15356238, Log Avg loss: 0.13095109, Global Avg Loss: 0.67829512, Time: 0.0209 Steps: 77920, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001286, Sample Num: 20576, Cur Loss: 0.05975534, Cur Avg Loss: 0.15324655, Log Avg loss: 0.11294684, Global Avg Loss: 0.67822258, Time: 0.0247 Steps: 77930, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001296, Sample Num: 20736, Cur Loss: 0.24767411, Cur Avg Loss: 0.15344910, Log Avg loss: 0.17949709, Global Avg Loss: 0.67815859, Time: 0.0209 Steps: 77940, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001306, Sample Num: 20896, Cur Loss: 0.03929359, Cur Avg Loss: 0.15346353, Log Avg loss: 0.15533267, Global Avg Loss: 0.67809152, Time: 0.0209 Steps: 77950, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001316, Sample Num: 21056, Cur Loss: 0.05475826, Cur Avg Loss: 0.15355397, Log Avg loss: 0.16536610, Global Avg Loss: 0.67802575, Time: 0.0209 Steps: 77960, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001326, Sample Num: 21216, Cur Loss: 0.06056000, Cur Avg Loss: 0.15317890, Log Avg loss: 0.10381991, Global Avg Loss: 0.67795211, Time: 0.0209 Steps: 77970, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001336, Sample Num: 21376, Cur Loss: 0.14417922, Cur Avg Loss: 0.15289238, Log Avg loss: 0.11489900, Global Avg Loss: 0.67787990, Time: 0.0209 Steps: 77980, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001346, Sample Num: 21536, Cur Loss: 0.22132164, Cur Avg Loss: 0.15278264, Log Avg loss: 0.13812119, Global Avg Loss: 0.67781069, Time: 0.0209 Steps: 77990, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001356, Sample Num: 21696, Cur Loss: 0.12786506, Cur Avg Loss: 0.15283011, Log Avg loss: 0.15921964, Global Avg Loss: 0.67774421, Time: 0.0209 Steps: 78000, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001366, Sample Num: 21856, Cur Loss: 0.15037912, Cur Avg Loss: 0.15286994, Log Avg loss: 0.15827155, Global Avg Loss: 0.67767762, Time: 0.0209 Steps: 78010, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001376, Sample Num: 22016, Cur Loss: 0.13350588, Cur Avg Loss: 0.15252856, Log Avg loss: 0.10589629, Global Avg Loss: 0.67760433, Time: 0.0209 Steps: 78020, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001386, Sample Num: 22176, Cur Loss: 0.15136118, Cur Avg Loss: 0.15262161, Log Avg loss: 0.16542540, Global Avg Loss: 0.67753869, Time: 0.0209 Steps: 78030, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001396, Sample Num: 22336, Cur Loss: 0.08600691, Cur Avg Loss: 0.15277979, Log Avg loss: 0.17470269, Global Avg Loss: 0.67747426, Time: 0.0209 Steps: 78040, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001406, Sample Num: 22496, Cur Loss: 0.05683101, Cur Avg Loss: 0.15296457, Log Avg loss: 0.17875999, Global Avg Loss: 0.67741036, Time: 0.0209 Steps: 78050, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001416, Sample Num: 22656, Cur Loss: 0.08678320, Cur Avg Loss: 0.15308536, Log Avg loss: 0.17006818, Global Avg Loss: 0.67734537, Time: 0.0209 Steps: 78060, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001426, Sample Num: 22816, Cur Loss: 0.05651313, Cur Avg Loss: 0.15292683, Log Avg loss: 0.13047959, Global Avg Loss: 0.67727532, Time: 0.0209 Steps: 78070, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001436, Sample Num: 22976, Cur Loss: 0.19177340, Cur Avg Loss: 0.15289983, Log Avg loss: 0.14904974, Global Avg Loss: 0.67720767, Time: 0.0209 Steps: 78080, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001446, Sample Num: 23136, Cur Loss: 0.31078625, Cur Avg Loss: 0.15340204, Log Avg loss: 0.22551904, Global Avg Loss: 0.67714983, Time: 0.0209 Steps: 78090, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001456, Sample Num: 23296, Cur Loss: 0.03589922, Cur Avg Loss: 0.15347447, Log Avg loss: 0.16394810, Global Avg Loss: 0.67708411, Time: 0.0209 Steps: 78100, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001466, Sample Num: 23456, Cur Loss: 0.36125633, Cur Avg Loss: 0.15365040, Log Avg loss: 0.17926509, Global Avg Loss: 0.67702038, Time: 0.0209 Steps: 78110, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001476, Sample Num: 23616, Cur Loss: 0.17903960, Cur Avg Loss: 0.15367317, Log Avg loss: 0.15701162, Global Avg Loss: 0.67695382, Time: 0.0209 Steps: 78120, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001486, Sample Num: 23776, Cur Loss: 0.19163574, Cur Avg Loss: 0.15384511, Log Avg loss: 0.17922384, Global Avg Loss: 0.67689011, Time: 0.0209 Steps: 78130, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001496, Sample Num: 23936, Cur Loss: 0.11834906, Cur Avg Loss: 0.15414167, Log Avg loss: 0.19821062, Global Avg Loss: 0.67682885, Time: 0.0209 Steps: 78140, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001506, Sample Num: 24096, Cur Loss: 0.05505386, Cur Avg Loss: 0.15428524, Log Avg loss: 0.17576349, Global Avg Loss: 0.67676474, Time: 0.0209 Steps: 78150, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001516, Sample Num: 24256, Cur Loss: 0.03993467, Cur Avg Loss: 0.15427094, Log Avg loss: 0.15211702, Global Avg Loss: 0.67669761, Time: 0.0209 Steps: 78160, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001526, Sample Num: 24416, Cur Loss: 0.25252047, Cur Avg Loss: 0.15450055, Log Avg loss: 0.18930867, Global Avg Loss: 0.67663526, Time: 0.0209 Steps: 78170, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001536, Sample Num: 24576, Cur Loss: 0.05897429, Cur Avg Loss: 0.15434913, Log Avg loss: 0.13124307, Global Avg Loss: 0.67656550, Time: 0.0254 Steps: 78180, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001546, Sample Num: 24736, Cur Loss: 0.18243717, Cur Avg Loss: 0.15454248, Log Avg loss: 0.18424034, Global Avg Loss: 0.67650253, Time: 0.0209 Steps: 78190, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001556, Sample Num: 24896, Cur Loss: 0.16556507, Cur Avg Loss: 0.15447243, Log Avg loss: 0.14364298, Global Avg Loss: 0.67643439, Time: 0.0209 Steps: 78200, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001566, Sample Num: 25056, Cur Loss: 0.20877218, Cur Avg Loss: 0.15418588, Log Avg loss: 0.10959839, Global Avg Loss: 0.67636192, Time: 0.0209 Steps: 78210, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001576, Sample Num: 25216, Cur Loss: 0.06937689, Cur Avg Loss: 0.15388053, Log Avg loss: 0.10606409, Global Avg Loss: 0.67628901, Time: 0.0209 Steps: 78220, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001586, Sample Num: 25376, Cur Loss: 0.06788742, Cur Avg Loss: 0.15396081, Log Avg loss: 0.16661235, Global Avg Loss: 0.67622386, Time: 0.0209 Steps: 78230, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001596, Sample Num: 25536, Cur Loss: 0.09892327, Cur Avg Loss: 0.15396090, Log Avg loss: 0.15397521, Global Avg Loss: 0.67615711, Time: 0.0209 Steps: 78240, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001606, Sample Num: 25696, Cur Loss: 0.11694943, Cur Avg Loss: 0.15387102, Log Avg loss: 0.13952677, Global Avg Loss: 0.67608853, Time: 0.0209 Steps: 78250, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001616, Sample Num: 25856, Cur Loss: 0.23692134, Cur Avg Loss: 0.15395901, Log Avg loss: 0.16808978, Global Avg Loss: 0.67602362, Time: 0.0209 Steps: 78260, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001626, Sample Num: 26016, Cur Loss: 0.14714518, Cur Avg Loss: 0.15417384, Log Avg loss: 0.18889065, Global Avg Loss: 0.67596138, Time: 0.0209 Steps: 78270, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001636, Sample Num: 26176, Cur Loss: 0.17642818, Cur Avg Loss: 0.15438544, Log Avg loss: 0.18879140, Global Avg Loss: 0.67589915, Time: 0.0209 Steps: 78280, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001646, Sample Num: 26336, Cur Loss: 0.06532273, Cur Avg Loss: 0.15404164, Log Avg loss: 0.09779630, Global Avg Loss: 0.67582530, Time: 0.0209 Steps: 78290, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001656, Sample Num: 26496, Cur Loss: 0.19204713, Cur Avg Loss: 0.15377228, Log Avg loss: 0.10943475, Global Avg Loss: 0.67575297, Time: 0.0209 Steps: 78300, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001666, Sample Num: 26656, Cur Loss: 0.08955906, Cur Avg Loss: 0.15370010, Log Avg loss: 0.14174728, Global Avg Loss: 0.67568478, Time: 0.0209 Steps: 78310, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001676, Sample Num: 26816, Cur Loss: 0.30274433, Cur Avg Loss: 0.15375570, Log Avg loss: 0.16301873, Global Avg Loss: 0.67561932, Time: 0.0209 Steps: 78320, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001686, Sample Num: 26976, Cur Loss: 0.18230259, Cur Avg Loss: 0.15352692, Log Avg loss: 0.11518348, Global Avg Loss: 0.67554777, Time: 0.0209 Steps: 78330, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001696, Sample Num: 27136, Cur Loss: 0.08406424, Cur Avg Loss: 0.15346358, Log Avg loss: 0.14278438, Global Avg Loss: 0.67547976, Time: 0.0209 Steps: 78340, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001706, Sample Num: 27296, Cur Loss: 0.20745400, Cur Avg Loss: 0.15376554, Log Avg loss: 0.20497733, Global Avg Loss: 0.67541971, Time: 0.0209 Steps: 78350, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001716, Sample Num: 27456, Cur Loss: 0.08822228, Cur Avg Loss: 0.15409350, Log Avg loss: 0.21004335, Global Avg Loss: 0.67536032, Time: 0.0208 Steps: 78360, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001726, Sample Num: 27616, Cur Loss: 0.09453504, Cur Avg Loss: 0.15429068, Log Avg loss: 0.18812741, Global Avg Loss: 0.67529815, Time: 0.0208 Steps: 78370, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001736, Sample Num: 27776, Cur Loss: 0.25629699, Cur Avg Loss: 0.15425330, Log Avg loss: 0.14780159, Global Avg Loss: 0.67523085, Time: 0.0208 Steps: 78380, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001746, Sample Num: 27936, Cur Loss: 0.19264480, Cur Avg Loss: 0.15408369, Log Avg loss: 0.12464016, Global Avg Loss: 0.67516062, Time: 0.0208 Steps: 78390, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001756, Sample Num: 28096, Cur Loss: 0.18727645, Cur Avg Loss: 0.15431544, Log Avg loss: 0.19477754, Global Avg Loss: 0.67509934, Time: 0.0209 Steps: 78400, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001766, Sample Num: 28256, Cur Loss: 0.04908447, Cur Avg Loss: 0.15450319, Log Avg loss: 0.18747209, Global Avg Loss: 0.67503715, Time: 0.0208 Steps: 78410, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001776, Sample Num: 28416, Cur Loss: 0.12253845, Cur Avg Loss: 0.15486417, Log Avg loss: 0.21861461, Global Avg Loss: 0.67497895, Time: 0.0208 Steps: 78420, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001786, Sample Num: 28576, Cur Loss: 0.18891387, Cur Avg Loss: 0.15484965, Log Avg loss: 0.15226931, Global Avg Loss: 0.67491230, Time: 0.0208 Steps: 78430, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001796, Sample Num: 28736, Cur Loss: 0.10729634, Cur Avg Loss: 0.15488627, Log Avg loss: 0.16142676, Global Avg Loss: 0.67484684, Time: 0.0209 Steps: 78440, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001806, Sample Num: 28896, Cur Loss: 0.34879923, Cur Avg Loss: 0.15506014, Log Avg loss: 0.18628770, Global Avg Loss: 0.67478456, Time: 0.0208 Steps: 78450, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001816, Sample Num: 29056, Cur Loss: 0.19820395, Cur Avg Loss: 0.15548998, Log Avg loss: 0.23311886, Global Avg Loss: 0.67472827, Time: 0.0208 Steps: 78460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001826, Sample Num: 29216, Cur Loss: 0.37681532, Cur Avg Loss: 0.15548071, Log Avg loss: 0.15379776, Global Avg Loss: 0.67466189, Time: 0.0209 Steps: 78470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001836, Sample Num: 29376, Cur Loss: 0.08355910, Cur Avg Loss: 0.15540255, Log Avg loss: 0.14112996, Global Avg Loss: 0.67459390, Time: 0.0209 Steps: 78480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001846, Sample Num: 29536, Cur Loss: 0.03754519, Cur Avg Loss: 0.15520016, Log Avg loss: 0.11804226, Global Avg Loss: 0.67452300, Time: 0.0208 Steps: 78490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001856, Sample Num: 29696, Cur Loss: 0.18244493, Cur Avg Loss: 0.15539281, Log Avg loss: 0.19095598, Global Avg Loss: 0.67446140, Time: 0.0207 Steps: 78500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001866, Sample Num: 29856, Cur Loss: 0.05743879, Cur Avg Loss: 0.15551480, Log Avg loss: 0.17815569, Global Avg Loss: 0.67439818, Time: 0.0207 Steps: 78510, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001876, Sample Num: 30016, Cur Loss: 0.19020978, Cur Avg Loss: 0.15569582, Log Avg loss: 0.18947495, Global Avg Loss: 0.67433642, Time: 0.0207 Steps: 78520, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001886, Sample Num: 30176, Cur Loss: 0.04696327, Cur Avg Loss: 0.15594138, Log Avg loss: 0.20200778, Global Avg Loss: 0.67427628, Time: 0.0208 Steps: 78530, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001896, Sample Num: 30336, Cur Loss: 0.27103373, Cur Avg Loss: 0.15621401, Log Avg loss: 0.20763271, Global Avg Loss: 0.67421686, Time: 0.0207 Steps: 78540, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001906, Sample Num: 30496, Cur Loss: 0.09952266, Cur Avg Loss: 0.15609250, Log Avg loss: 0.13305296, Global Avg Loss: 0.67414797, Time: 0.0207 Steps: 78550, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001916, Sample Num: 30656, Cur Loss: 0.28288111, Cur Avg Loss: 0.15607778, Log Avg loss: 0.15327280, Global Avg Loss: 0.67408166, Time: 0.0207 Steps: 78560, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001926, Sample Num: 30816, Cur Loss: 0.13542694, Cur Avg Loss: 0.15597554, Log Avg loss: 0.13638537, Global Avg Loss: 0.67401323, Time: 0.0208 Steps: 78570, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001936, Sample Num: 30976, Cur Loss: 0.37577462, Cur Avg Loss: 0.15616888, Log Avg loss: 0.19340634, Global Avg Loss: 0.67395207, Time: 0.0209 Steps: 78580, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001946, Sample Num: 31136, Cur Loss: 0.11343199, Cur Avg Loss: 0.15621598, Log Avg loss: 0.16533572, Global Avg Loss: 0.67388735, Time: 0.0208 Steps: 78590, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001956, Sample Num: 31296, Cur Loss: 0.15354168, Cur Avg Loss: 0.15625465, Log Avg loss: 0.16377968, Global Avg Loss: 0.67382245, Time: 0.0208 Steps: 78600, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001966, Sample Num: 31456, Cur Loss: 0.33380154, Cur Avg Loss: 0.15634532, Log Avg loss: 0.17407989, Global Avg Loss: 0.67375888, Time: 0.0207 Steps: 78610, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001976, Sample Num: 31616, Cur Loss: 0.17873693, Cur Avg Loss: 0.15633535, Log Avg loss: 0.15437460, Global Avg Loss: 0.67369282, Time: 0.0208 Steps: 78620, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001986, Sample Num: 31776, Cur Loss: 0.27777871, Cur Avg Loss: 0.15633627, Log Avg loss: 0.15651803, Global Avg Loss: 0.67362704, Time: 0.0208 Steps: 78630, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001996, Sample Num: 31936, Cur Loss: 0.04810355, Cur Avg Loss: 0.15612700, Log Avg loss: 0.11456734, Global Avg Loss: 0.67355595, Time: 0.0209 Steps: 78640, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002006, Sample Num: 32096, Cur Loss: 0.32477438, Cur Avg Loss: 0.15639870, Log Avg loss: 0.21062881, Global Avg Loss: 0.67349709, Time: 0.0209 Steps: 78650, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002016, Sample Num: 32256, Cur Loss: 0.05550556, Cur Avg Loss: 0.15646557, Log Avg loss: 0.16988058, Global Avg Loss: 0.67343307, Time: 0.0209 Steps: 78660, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002026, Sample Num: 32416, Cur Loss: 0.36921287, Cur Avg Loss: 0.15679301, Log Avg loss: 0.22280507, Global Avg Loss: 0.67337579, Time: 0.0209 Steps: 78670, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002036, Sample Num: 32576, Cur Loss: 0.15151420, Cur Avg Loss: 0.15668055, Log Avg loss: 0.13389638, Global Avg Loss: 0.67330722, Time: 0.0208 Steps: 78680, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002046, Sample Num: 32736, Cur Loss: 0.38354787, Cur Avg Loss: 0.15669854, Log Avg loss: 0.16035996, Global Avg Loss: 0.67324203, Time: 0.0209 Steps: 78690, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002056, Sample Num: 32896, Cur Loss: 0.31360155, Cur Avg Loss: 0.15687085, Log Avg loss: 0.19212566, Global Avg Loss: 0.67318090, Time: 0.0208 Steps: 78700, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002066, Sample Num: 33056, Cur Loss: 0.18096133, Cur Avg Loss: 0.15677657, Log Avg loss: 0.13739338, Global Avg Loss: 0.67311283, Time: 0.0208 Steps: 78710, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002076, Sample Num: 33216, Cur Loss: 0.15498987, Cur Avg Loss: 0.15664697, Log Avg loss: 0.12987176, Global Avg Loss: 0.67304382, Time: 0.0208 Steps: 78720, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002086, Sample Num: 33376, Cur Loss: 0.17201965, Cur Avg Loss: 0.15657093, Log Avg loss: 0.14078537, Global Avg Loss: 0.67297622, Time: 0.0209 Steps: 78730, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002096, Sample Num: 33536, Cur Loss: 0.21469671, Cur Avg Loss: 0.15640466, Log Avg loss: 0.12172086, Global Avg Loss: 0.67290621, Time: 0.0208 Steps: 78740, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002106, Sample Num: 33696, Cur Loss: 0.10491127, Cur Avg Loss: 0.15617271, Log Avg loss: 0.10755604, Global Avg Loss: 0.67283442, Time: 0.0208 Steps: 78750, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002116, Sample Num: 33856, Cur Loss: 0.14315943, Cur Avg Loss: 0.15639031, Log Avg loss: 0.20221626, Global Avg Loss: 0.67277466, Time: 0.0209 Steps: 78760, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002126, Sample Num: 34016, Cur Loss: 0.04708955, Cur Avg Loss: 0.15636133, Log Avg loss: 0.15022824, Global Avg Loss: 0.67270832, Time: 0.0208 Steps: 78770, Updated lr: 0.000026 ***** Running evaluation checkpoint-78773 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-78773 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.539107, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.307008, "eval_total_loss": 215.826752, "eval_mae": 0.387753, "eval_mse": 0.307124, "eval_r2": 0.804772, "eval_sp_statistic": 0.926605, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.933218, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.172376, "test_total_loss": 86.532577, "test_mae": 0.28325, "test_mse": 0.172439, "test_r2": 0.888706, "test_sp_statistic": 0.918491, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946134, "test_ps_pvalue": 0.0, "lr": 2.624656235182551e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6726871581346907, "train_cur_epoch_loss": 332.77499925903976, "train_cur_epoch_avg_loss": 0.15630577701223097, "train_cur_epoch_time": 44.53910684585571, "train_cur_epoch_avg_time": 0.020920200491242703, "epoch": 37, "step": 78773} ################################################## Training, Epoch: 0038, Batch: 000007, Sample Num: 112, Cur Loss: 0.16125062, Cur Avg Loss: 0.15754288, Log Avg loss: 0.14536174, Global Avg Loss: 0.67264138, Time: 0.0226 Steps: 78780, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000017, Sample Num: 272, Cur Loss: 0.42256817, Cur Avg Loss: 0.18543646, Log Avg loss: 0.20496196, Global Avg Loss: 0.67258203, Time: 0.0209 Steps: 78790, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000027, Sample Num: 432, Cur Loss: 0.14565414, Cur Avg Loss: 0.17212603, Log Avg loss: 0.14949831, Global Avg Loss: 0.67251565, Time: 0.0210 Steps: 78800, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000037, Sample Num: 592, Cur Loss: 0.11902289, Cur Avg Loss: 0.15969994, Log Avg loss: 0.12614951, Global Avg Loss: 0.67244632, Time: 0.0209 Steps: 78810, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000047, Sample Num: 752, Cur Loss: 0.22464325, Cur Avg Loss: 0.15077390, Log Avg loss: 0.11774754, Global Avg Loss: 0.67237594, Time: 0.0210 Steps: 78820, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000057, Sample Num: 912, Cur Loss: 0.20796946, Cur Avg Loss: 0.15096973, Log Avg loss: 0.15189011, Global Avg Loss: 0.67230992, Time: 0.0209 Steps: 78830, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000067, Sample Num: 1072, Cur Loss: 0.09358718, Cur Avg Loss: 0.14962522, Log Avg loss: 0.14196151, Global Avg Loss: 0.67224265, Time: 0.0210 Steps: 78840, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000077, Sample Num: 1232, Cur Loss: 0.07303250, Cur Avg Loss: 0.15262225, Log Avg loss: 0.17270237, Global Avg Loss: 0.67217930, Time: 0.0209 Steps: 78850, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000087, Sample Num: 1392, Cur Loss: 0.03851994, Cur Avg Loss: 0.14591130, Log Avg loss: 0.09423697, Global Avg Loss: 0.67210601, Time: 0.0211 Steps: 78860, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000097, Sample Num: 1552, Cur Loss: 0.33825037, Cur Avg Loss: 0.14656403, Log Avg loss: 0.15224285, Global Avg Loss: 0.67204009, Time: 0.0209 Steps: 78870, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000107, Sample Num: 1712, Cur Loss: 0.14798462, Cur Avg Loss: 0.14441602, Log Avg loss: 0.12358022, Global Avg Loss: 0.67197056, Time: 0.0210 Steps: 78880, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000117, Sample Num: 1872, Cur Loss: 0.29767600, Cur Avg Loss: 0.14849880, Log Avg loss: 0.19218460, Global Avg Loss: 0.67190975, Time: 0.0209 Steps: 78890, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000127, Sample Num: 2032, Cur Loss: 0.08703218, Cur Avg Loss: 0.14908307, Log Avg loss: 0.15591905, Global Avg Loss: 0.67184435, Time: 0.0211 Steps: 78900, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000137, Sample Num: 2192, Cur Loss: 0.23855746, Cur Avg Loss: 0.14792685, Log Avg loss: 0.13324283, Global Avg Loss: 0.67177609, Time: 0.0211 Steps: 78910, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000147, Sample Num: 2352, Cur Loss: 0.05366855, Cur Avg Loss: 0.14962224, Log Avg loss: 0.17284912, Global Avg Loss: 0.67171287, Time: 0.0209 Steps: 78920, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000157, Sample Num: 2512, Cur Loss: 0.11199261, Cur Avg Loss: 0.15061072, Log Avg loss: 0.16514128, Global Avg Loss: 0.67164869, Time: 0.0211 Steps: 78930, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000167, Sample Num: 2672, Cur Loss: 0.10513923, Cur Avg Loss: 0.15105255, Log Avg loss: 0.15798927, Global Avg Loss: 0.67158362, Time: 0.0210 Steps: 78940, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000177, Sample Num: 2832, Cur Loss: 0.04017221, Cur Avg Loss: 0.14944460, Log Avg loss: 0.12259196, Global Avg Loss: 0.67151409, Time: 0.0209 Steps: 78950, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000187, Sample Num: 2992, Cur Loss: 0.20761776, Cur Avg Loss: 0.14882771, Log Avg loss: 0.13790876, Global Avg Loss: 0.67144651, Time: 0.0210 Steps: 78960, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000197, Sample Num: 3152, Cur Loss: 0.20631459, Cur Avg Loss: 0.14759347, Log Avg loss: 0.12451320, Global Avg Loss: 0.67137725, Time: 0.0211 Steps: 78970, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000207, Sample Num: 3312, Cur Loss: 0.12061450, Cur Avg Loss: 0.14935132, Log Avg loss: 0.18398098, Global Avg Loss: 0.67131554, Time: 0.0209 Steps: 78980, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000217, Sample Num: 3472, Cur Loss: 0.11429891, Cur Avg Loss: 0.14842794, Log Avg loss: 0.12931390, Global Avg Loss: 0.67124692, Time: 0.0211 Steps: 78990, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000227, Sample Num: 3632, Cur Loss: 0.14149220, Cur Avg Loss: 0.14709517, Log Avg loss: 0.11817395, Global Avg Loss: 0.67117691, Time: 0.0209 Steps: 79000, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000237, Sample Num: 3792, Cur Loss: 0.12254061, Cur Avg Loss: 0.14800174, Log Avg loss: 0.16858093, Global Avg Loss: 0.67111330, Time: 0.0210 Steps: 79010, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000247, Sample Num: 3952, Cur Loss: 0.03528085, Cur Avg Loss: 0.14805678, Log Avg loss: 0.14936118, Global Avg Loss: 0.67104727, Time: 0.0210 Steps: 79020, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000257, Sample Num: 4112, Cur Loss: 0.12671196, Cur Avg Loss: 0.15066022, Log Avg loss: 0.21496526, Global Avg Loss: 0.67098956, Time: 0.0245 Steps: 79030, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000267, Sample Num: 4272, Cur Loss: 0.21809742, Cur Avg Loss: 0.15040761, Log Avg loss: 0.14391563, Global Avg Loss: 0.67092288, Time: 0.0209 Steps: 79040, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000277, Sample Num: 4432, Cur Loss: 0.03266112, Cur Avg Loss: 0.14993750, Log Avg loss: 0.13738539, Global Avg Loss: 0.67085539, Time: 0.0208 Steps: 79050, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000287, Sample Num: 4592, Cur Loss: 0.07597511, Cur Avg Loss: 0.14822644, Log Avg loss: 0.10083025, Global Avg Loss: 0.67078328, Time: 0.0208 Steps: 79060, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000297, Sample Num: 4752, Cur Loss: 0.16238546, Cur Avg Loss: 0.14853459, Log Avg loss: 0.15737832, Global Avg Loss: 0.67071835, Time: 0.0208 Steps: 79070, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000307, Sample Num: 4912, Cur Loss: 0.07977252, Cur Avg Loss: 0.14969130, Log Avg loss: 0.18404574, Global Avg Loss: 0.67065681, Time: 0.0209 Steps: 79080, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000317, Sample Num: 5072, Cur Loss: 0.20363192, Cur Avg Loss: 0.15057481, Log Avg loss: 0.17769868, Global Avg Loss: 0.67059448, Time: 0.0208 Steps: 79090, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000327, Sample Num: 5232, Cur Loss: 0.07503958, Cur Avg Loss: 0.14988874, Log Avg loss: 0.12814014, Global Avg Loss: 0.67052591, Time: 0.0208 Steps: 79100, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000337, Sample Num: 5392, Cur Loss: 0.15636781, Cur Avg Loss: 0.15045512, Log Avg loss: 0.16897580, Global Avg Loss: 0.67046251, Time: 0.0208 Steps: 79110, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000347, Sample Num: 5552, Cur Loss: 0.16104178, Cur Avg Loss: 0.15003629, Log Avg loss: 0.13592183, Global Avg Loss: 0.67039495, Time: 0.0208 Steps: 79120, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000357, Sample Num: 5712, Cur Loss: 0.13653548, Cur Avg Loss: 0.15019918, Log Avg loss: 0.15585123, Global Avg Loss: 0.67032992, Time: 0.0209 Steps: 79130, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000367, Sample Num: 5872, Cur Loss: 0.09071846, Cur Avg Loss: 0.14975002, Log Avg loss: 0.13371498, Global Avg Loss: 0.67026211, Time: 0.0209 Steps: 79140, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000377, Sample Num: 6032, Cur Loss: 0.09925893, Cur Avg Loss: 0.14997678, Log Avg loss: 0.15829899, Global Avg Loss: 0.67019743, Time: 0.0208 Steps: 79150, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000387, Sample Num: 6192, Cur Loss: 0.18767966, Cur Avg Loss: 0.14924715, Log Avg loss: 0.12174027, Global Avg Loss: 0.67012815, Time: 0.0209 Steps: 79160, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000397, Sample Num: 6352, Cur Loss: 0.13131472, Cur Avg Loss: 0.14916741, Log Avg loss: 0.14608141, Global Avg Loss: 0.67006195, Time: 0.0209 Steps: 79170, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000407, Sample Num: 6512, Cur Loss: 0.11001558, Cur Avg Loss: 0.14836247, Log Avg loss: 0.11640641, Global Avg Loss: 0.66999203, Time: 0.0211 Steps: 79180, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000417, Sample Num: 6672, Cur Loss: 0.08140466, Cur Avg Loss: 0.14822184, Log Avg loss: 0.14249809, Global Avg Loss: 0.66992542, Time: 0.0209 Steps: 79190, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000427, Sample Num: 6832, Cur Loss: 0.33525434, Cur Avg Loss: 0.15059729, Log Avg loss: 0.24965341, Global Avg Loss: 0.66987236, Time: 0.0210 Steps: 79200, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000437, Sample Num: 6992, Cur Loss: 0.22742413, Cur Avg Loss: 0.15101532, Log Avg loss: 0.16886544, Global Avg Loss: 0.66980910, Time: 0.0211 Steps: 79210, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000447, Sample Num: 7152, Cur Loss: 0.07690071, Cur Avg Loss: 0.15256516, Log Avg loss: 0.22029304, Global Avg Loss: 0.66975236, Time: 0.0211 Steps: 79220, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000457, Sample Num: 7312, Cur Loss: 0.15045811, Cur Avg Loss: 0.15289960, Log Avg loss: 0.16784926, Global Avg Loss: 0.66968901, Time: 0.0211 Steps: 79230, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000467, Sample Num: 7472, Cur Loss: 0.13007455, Cur Avg Loss: 0.15275169, Log Avg loss: 0.14599222, Global Avg Loss: 0.66962292, Time: 0.0211 Steps: 79240, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000477, Sample Num: 7632, Cur Loss: 0.16604038, Cur Avg Loss: 0.15308755, Log Avg loss: 0.16877200, Global Avg Loss: 0.66955973, Time: 0.0211 Steps: 79250, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000487, Sample Num: 7792, Cur Loss: 0.07869437, Cur Avg Loss: 0.15236667, Log Avg loss: 0.11798052, Global Avg Loss: 0.66949013, Time: 0.0211 Steps: 79260, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000497, Sample Num: 7952, Cur Loss: 0.11387314, Cur Avg Loss: 0.15168967, Log Avg loss: 0.11871998, Global Avg Loss: 0.66942065, Time: 0.0211 Steps: 79270, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000507, Sample Num: 8112, Cur Loss: 0.27373296, Cur Avg Loss: 0.15182104, Log Avg loss: 0.15835013, Global Avg Loss: 0.66935619, Time: 0.0210 Steps: 79280, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000517, Sample Num: 8272, Cur Loss: 0.15107363, Cur Avg Loss: 0.15224446, Log Avg loss: 0.17371185, Global Avg Loss: 0.66929368, Time: 0.0211 Steps: 79290, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000527, Sample Num: 8432, Cur Loss: 0.27063185, Cur Avg Loss: 0.15340356, Log Avg loss: 0.21332891, Global Avg Loss: 0.66923618, Time: 0.0211 Steps: 79300, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000537, Sample Num: 8592, Cur Loss: 0.16969705, Cur Avg Loss: 0.15373639, Log Avg loss: 0.17127683, Global Avg Loss: 0.66917339, Time: 0.0211 Steps: 79310, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000547, Sample Num: 8752, Cur Loss: 0.48488766, Cur Avg Loss: 0.15400515, Log Avg loss: 0.16843759, Global Avg Loss: 0.66911027, Time: 0.0211 Steps: 79320, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000557, Sample Num: 8912, Cur Loss: 0.14086102, Cur Avg Loss: 0.15302386, Log Avg loss: 0.09934709, Global Avg Loss: 0.66903844, Time: 0.0211 Steps: 79330, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000567, Sample Num: 9072, Cur Loss: 0.06414824, Cur Avg Loss: 0.15279270, Log Avg loss: 0.13991695, Global Avg Loss: 0.66897175, Time: 0.0211 Steps: 79340, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000577, Sample Num: 9232, Cur Loss: 0.02715472, Cur Avg Loss: 0.15221688, Log Avg loss: 0.11956792, Global Avg Loss: 0.66890252, Time: 0.0211 Steps: 79350, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000587, Sample Num: 9392, Cur Loss: 0.21002121, Cur Avg Loss: 0.15167836, Log Avg loss: 0.12060602, Global Avg Loss: 0.66883343, Time: 0.0211 Steps: 79360, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000597, Sample Num: 9552, Cur Loss: 0.15232542, Cur Avg Loss: 0.15082372, Log Avg loss: 0.10065635, Global Avg Loss: 0.66876184, Time: 0.0211 Steps: 79370, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000607, Sample Num: 9712, Cur Loss: 0.80118752, Cur Avg Loss: 0.15235439, Log Avg loss: 0.24373518, Global Avg Loss: 0.66870830, Time: 0.0211 Steps: 79380, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000617, Sample Num: 9872, Cur Loss: 0.23906793, Cur Avg Loss: 0.15152294, Log Avg loss: 0.10105426, Global Avg Loss: 0.66863680, Time: 0.0211 Steps: 79390, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000627, Sample Num: 10032, Cur Loss: 0.11842597, Cur Avg Loss: 0.15143336, Log Avg loss: 0.14590569, Global Avg Loss: 0.66857096, Time: 0.0211 Steps: 79400, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000637, Sample Num: 10192, Cur Loss: 0.06062640, Cur Avg Loss: 0.15109056, Log Avg loss: 0.12959759, Global Avg Loss: 0.66850309, Time: 0.0211 Steps: 79410, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000647, Sample Num: 10352, Cur Loss: 0.08778324, Cur Avg Loss: 0.15110679, Log Avg loss: 0.15214004, Global Avg Loss: 0.66843807, Time: 0.0211 Steps: 79420, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000657, Sample Num: 10512, Cur Loss: 0.35424981, Cur Avg Loss: 0.15093607, Log Avg loss: 0.13989068, Global Avg Loss: 0.66837153, Time: 0.0211 Steps: 79430, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000667, Sample Num: 10672, Cur Loss: 0.03546137, Cur Avg Loss: 0.15137490, Log Avg loss: 0.18020630, Global Avg Loss: 0.66831008, Time: 0.0211 Steps: 79440, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000677, Sample Num: 10832, Cur Loss: 0.06709786, Cur Avg Loss: 0.15179532, Log Avg loss: 0.17983701, Global Avg Loss: 0.66824860, Time: 0.0211 Steps: 79450, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000687, Sample Num: 10992, Cur Loss: 0.08176069, Cur Avg Loss: 0.15116271, Log Avg loss: 0.10833495, Global Avg Loss: 0.66817813, Time: 0.0211 Steps: 79460, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000697, Sample Num: 11152, Cur Loss: 0.23210061, Cur Avg Loss: 0.15156452, Log Avg loss: 0.17916896, Global Avg Loss: 0.66811660, Time: 0.0211 Steps: 79470, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000707, Sample Num: 11312, Cur Loss: 0.17264228, Cur Avg Loss: 0.15073159, Log Avg loss: 0.09267643, Global Avg Loss: 0.66804420, Time: 0.0211 Steps: 79480, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000717, Sample Num: 11472, Cur Loss: 0.06979914, Cur Avg Loss: 0.15060360, Log Avg loss: 0.14155505, Global Avg Loss: 0.66797796, Time: 0.0211 Steps: 79490, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000727, Sample Num: 11632, Cur Loss: 0.29617751, Cur Avg Loss: 0.15121748, Log Avg loss: 0.19523230, Global Avg Loss: 0.66791850, Time: 0.0211 Steps: 79500, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000737, Sample Num: 11792, Cur Loss: 0.14852320, Cur Avg Loss: 0.15089986, Log Avg loss: 0.12780896, Global Avg Loss: 0.66785057, Time: 0.0211 Steps: 79510, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000747, Sample Num: 11952, Cur Loss: 0.19823653, Cur Avg Loss: 0.15105368, Log Avg loss: 0.16239056, Global Avg Loss: 0.66778700, Time: 0.0211 Steps: 79520, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000757, Sample Num: 12112, Cur Loss: 0.10195212, Cur Avg Loss: 0.15041335, Log Avg loss: 0.10258021, Global Avg Loss: 0.66771594, Time: 0.0211 Steps: 79530, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000767, Sample Num: 12272, Cur Loss: 0.14543578, Cur Avg Loss: 0.15098414, Log Avg loss: 0.19419285, Global Avg Loss: 0.66765640, Time: 0.0210 Steps: 79540, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000777, Sample Num: 12432, Cur Loss: 0.19150345, Cur Avg Loss: 0.15113976, Log Avg loss: 0.16307615, Global Avg Loss: 0.66759297, Time: 0.0211 Steps: 79550, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000787, Sample Num: 12592, Cur Loss: 0.03142024, Cur Avg Loss: 0.15129763, Log Avg loss: 0.16356370, Global Avg Loss: 0.66752962, Time: 0.0211 Steps: 79560, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000797, Sample Num: 12752, Cur Loss: 0.22833410, Cur Avg Loss: 0.15164963, Log Avg loss: 0.17935221, Global Avg Loss: 0.66746827, Time: 0.0211 Steps: 79570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000807, Sample Num: 12912, Cur Loss: 0.27302992, Cur Avg Loss: 0.15200898, Log Avg loss: 0.18064920, Global Avg Loss: 0.66740710, Time: 0.0211 Steps: 79580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000817, Sample Num: 13072, Cur Loss: 0.10675523, Cur Avg Loss: 0.15158093, Log Avg loss: 0.11703708, Global Avg Loss: 0.66733795, Time: 0.0211 Steps: 79590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000827, Sample Num: 13232, Cur Loss: 0.08335921, Cur Avg Loss: 0.15219624, Log Avg loss: 0.20246758, Global Avg Loss: 0.66727955, Time: 0.0211 Steps: 79600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000837, Sample Num: 13392, Cur Loss: 0.15214279, Cur Avg Loss: 0.15202211, Log Avg loss: 0.13762120, Global Avg Loss: 0.66721301, Time: 0.0211 Steps: 79610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000847, Sample Num: 13552, Cur Loss: 0.04731171, Cur Avg Loss: 0.15198626, Log Avg loss: 0.14898558, Global Avg Loss: 0.66714793, Time: 0.0211 Steps: 79620, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000857, Sample Num: 13712, Cur Loss: 0.13619703, Cur Avg Loss: 0.15221246, Log Avg loss: 0.17137218, Global Avg Loss: 0.66708567, Time: 0.0211 Steps: 79630, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000867, Sample Num: 13872, Cur Loss: 0.15831220, Cur Avg Loss: 0.15237875, Log Avg loss: 0.16662920, Global Avg Loss: 0.66702283, Time: 0.0211 Steps: 79640, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000877, Sample Num: 14032, Cur Loss: 0.18912384, Cur Avg Loss: 0.15245831, Log Avg loss: 0.15935613, Global Avg Loss: 0.66695909, Time: 0.0211 Steps: 79650, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000887, Sample Num: 14192, Cur Loss: 0.23582628, Cur Avg Loss: 0.15250379, Log Avg loss: 0.15649233, Global Avg Loss: 0.66689501, Time: 0.0212 Steps: 79660, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000897, Sample Num: 14352, Cur Loss: 0.07752858, Cur Avg Loss: 0.15180837, Log Avg loss: 0.09012540, Global Avg Loss: 0.66682261, Time: 0.0211 Steps: 79670, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000907, Sample Num: 14512, Cur Loss: 0.19019485, Cur Avg Loss: 0.15198546, Log Avg loss: 0.16787028, Global Avg Loss: 0.66675999, Time: 0.0211 Steps: 79680, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000917, Sample Num: 14672, Cur Loss: 0.13578624, Cur Avg Loss: 0.15149048, Log Avg loss: 0.10659532, Global Avg Loss: 0.66668970, Time: 0.0211 Steps: 79690, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000927, Sample Num: 14832, Cur Loss: 0.22732119, Cur Avg Loss: 0.15160182, Log Avg loss: 0.16181172, Global Avg Loss: 0.66662635, Time: 0.0211 Steps: 79700, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000937, Sample Num: 14992, Cur Loss: 0.12447818, Cur Avg Loss: 0.15189180, Log Avg loss: 0.17877271, Global Avg Loss: 0.66656515, Time: 0.0211 Steps: 79710, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000947, Sample Num: 15152, Cur Loss: 0.09036141, Cur Avg Loss: 0.15172648, Log Avg loss: 0.13623620, Global Avg Loss: 0.66649863, Time: 0.0210 Steps: 79720, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000957, Sample Num: 15312, Cur Loss: 0.11156493, Cur Avg Loss: 0.15165768, Log Avg loss: 0.14514229, Global Avg Loss: 0.66643324, Time: 0.0211 Steps: 79730, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000967, Sample Num: 15472, Cur Loss: 0.15357450, Cur Avg Loss: 0.15194878, Log Avg loss: 0.17980733, Global Avg Loss: 0.66637221, Time: 0.0211 Steps: 79740, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000977, Sample Num: 15632, Cur Loss: 0.02154782, Cur Avg Loss: 0.15171107, Log Avg loss: 0.12872434, Global Avg Loss: 0.66630479, Time: 0.0211 Steps: 79750, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000987, Sample Num: 15792, Cur Loss: 0.11478809, Cur Avg Loss: 0.15139737, Log Avg loss: 0.12074904, Global Avg Loss: 0.66623639, Time: 0.0211 Steps: 79760, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000997, Sample Num: 15952, Cur Loss: 0.32745847, Cur Avg Loss: 0.15171791, Log Avg loss: 0.18335510, Global Avg Loss: 0.66617586, Time: 0.0211 Steps: 79770, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001007, Sample Num: 16112, Cur Loss: 0.62944126, Cur Avg Loss: 0.15187340, Log Avg loss: 0.16737556, Global Avg Loss: 0.66611334, Time: 0.0210 Steps: 79780, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001017, Sample Num: 16272, Cur Loss: 0.23522072, Cur Avg Loss: 0.15235098, Log Avg loss: 0.20044313, Global Avg Loss: 0.66605497, Time: 0.0210 Steps: 79790, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001027, Sample Num: 16432, Cur Loss: 0.08024619, Cur Avg Loss: 0.15215149, Log Avg loss: 0.13186429, Global Avg Loss: 0.66598803, Time: 0.0249 Steps: 79800, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001037, Sample Num: 16592, Cur Loss: 0.20658799, Cur Avg Loss: 0.15193440, Log Avg loss: 0.12963840, Global Avg Loss: 0.66592083, Time: 0.0211 Steps: 79810, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001047, Sample Num: 16752, Cur Loss: 0.22342420, Cur Avg Loss: 0.15207949, Log Avg loss: 0.16712577, Global Avg Loss: 0.66585834, Time: 0.0211 Steps: 79820, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001057, Sample Num: 16912, Cur Loss: 0.18783841, Cur Avg Loss: 0.15186161, Log Avg loss: 0.12904989, Global Avg Loss: 0.66579110, Time: 0.0210 Steps: 79830, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001067, Sample Num: 17072, Cur Loss: 0.15527934, Cur Avg Loss: 0.15228568, Log Avg loss: 0.19710980, Global Avg Loss: 0.66573239, Time: 0.0211 Steps: 79840, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001077, Sample Num: 17232, Cur Loss: 0.03902019, Cur Avg Loss: 0.15270531, Log Avg loss: 0.19747937, Global Avg Loss: 0.66567375, Time: 0.0211 Steps: 79850, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001087, Sample Num: 17392, Cur Loss: 0.12598480, Cur Avg Loss: 0.15280259, Log Avg loss: 0.16327982, Global Avg Loss: 0.66561084, Time: 0.0211 Steps: 79860, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001097, Sample Num: 17552, Cur Loss: 0.12623860, Cur Avg Loss: 0.15256012, Log Avg loss: 0.12620340, Global Avg Loss: 0.66554331, Time: 0.0211 Steps: 79870, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001107, Sample Num: 17712, Cur Loss: 0.05019906, Cur Avg Loss: 0.15222841, Log Avg loss: 0.11583991, Global Avg Loss: 0.66547449, Time: 0.0211 Steps: 79880, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001117, Sample Num: 17872, Cur Loss: 0.06060243, Cur Avg Loss: 0.15243781, Log Avg loss: 0.17561867, Global Avg Loss: 0.66541317, Time: 0.0211 Steps: 79890, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001127, Sample Num: 18032, Cur Loss: 0.07340734, Cur Avg Loss: 0.15235567, Log Avg loss: 0.14318020, Global Avg Loss: 0.66534781, Time: 0.0210 Steps: 79900, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001137, Sample Num: 18192, Cur Loss: 0.11787777, Cur Avg Loss: 0.15207495, Log Avg loss: 0.12043796, Global Avg Loss: 0.66527962, Time: 0.0210 Steps: 79910, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001147, Sample Num: 18352, Cur Loss: 0.03576518, Cur Avg Loss: 0.15216408, Log Avg loss: 0.16229777, Global Avg Loss: 0.66521669, Time: 0.0210 Steps: 79920, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001157, Sample Num: 18512, Cur Loss: 0.05342060, Cur Avg Loss: 0.15199596, Log Avg loss: 0.13271322, Global Avg Loss: 0.66515007, Time: 0.0210 Steps: 79930, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001167, Sample Num: 18672, Cur Loss: 0.20494248, Cur Avg Loss: 0.15190406, Log Avg loss: 0.14127057, Global Avg Loss: 0.66508453, Time: 0.0213 Steps: 79940, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001177, Sample Num: 18832, Cur Loss: 0.17616075, Cur Avg Loss: 0.15204318, Log Avg loss: 0.16827858, Global Avg Loss: 0.66502239, Time: 0.0210 Steps: 79950, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001187, Sample Num: 18992, Cur Loss: 0.06699784, Cur Avg Loss: 0.15227334, Log Avg loss: 0.17936369, Global Avg Loss: 0.66496166, Time: 0.0210 Steps: 79960, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001197, Sample Num: 19152, Cur Loss: 0.09271592, Cur Avg Loss: 0.15201575, Log Avg loss: 0.12143937, Global Avg Loss: 0.66489369, Time: 0.0210 Steps: 79970, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001207, Sample Num: 19312, Cur Loss: 0.19857240, Cur Avg Loss: 0.15206561, Log Avg loss: 0.15803431, Global Avg Loss: 0.66483032, Time: 0.0210 Steps: 79980, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001217, Sample Num: 19472, Cur Loss: 0.08376223, Cur Avg Loss: 0.15252633, Log Avg loss: 0.20813533, Global Avg Loss: 0.66477322, Time: 0.0210 Steps: 79990, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001227, Sample Num: 19632, Cur Loss: 0.16469114, Cur Avg Loss: 0.15249935, Log Avg loss: 0.14921537, Global Avg Loss: 0.66470878, Time: 0.0210 Steps: 80000, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001237, Sample Num: 19792, Cur Loss: 0.28177544, Cur Avg Loss: 0.15249166, Log Avg loss: 0.15154792, Global Avg Loss: 0.66464464, Time: 0.0210 Steps: 80010, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001247, Sample Num: 19952, Cur Loss: 0.16943952, Cur Avg Loss: 0.15250290, Log Avg loss: 0.15389350, Global Avg Loss: 0.66458081, Time: 0.0210 Steps: 80020, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001257, Sample Num: 20112, Cur Loss: 0.53123987, Cur Avg Loss: 0.15295623, Log Avg loss: 0.20948684, Global Avg Loss: 0.66452395, Time: 0.0210 Steps: 80030, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001267, Sample Num: 20272, Cur Loss: 0.27819699, Cur Avg Loss: 0.15299177, Log Avg loss: 0.15745920, Global Avg Loss: 0.66446060, Time: 0.0210 Steps: 80040, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001277, Sample Num: 20432, Cur Loss: 0.07608251, Cur Avg Loss: 0.15284056, Log Avg loss: 0.13368190, Global Avg Loss: 0.66439429, Time: 0.0209 Steps: 80050, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001287, Sample Num: 20592, Cur Loss: 0.09249003, Cur Avg Loss: 0.15284958, Log Avg loss: 0.15400225, Global Avg Loss: 0.66433054, Time: 0.0209 Steps: 80060, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001297, Sample Num: 20752, Cur Loss: 0.14788270, Cur Avg Loss: 0.15280427, Log Avg loss: 0.14697274, Global Avg Loss: 0.66426593, Time: 0.0209 Steps: 80070, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001307, Sample Num: 20912, Cur Loss: 0.18652371, Cur Avg Loss: 0.15304190, Log Avg loss: 0.18386251, Global Avg Loss: 0.66420594, Time: 0.0209 Steps: 80080, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001317, Sample Num: 21072, Cur Loss: 0.13093710, Cur Avg Loss: 0.15291391, Log Avg loss: 0.13618482, Global Avg Loss: 0.66414001, Time: 0.0209 Steps: 80090, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001327, Sample Num: 21232, Cur Loss: 0.14971159, Cur Avg Loss: 0.15304292, Log Avg loss: 0.17003435, Global Avg Loss: 0.66407832, Time: 0.0209 Steps: 80100, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001337, Sample Num: 21392, Cur Loss: 0.31325197, Cur Avg Loss: 0.15330692, Log Avg loss: 0.18833939, Global Avg Loss: 0.66401893, Time: 0.0209 Steps: 80110, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001347, Sample Num: 21552, Cur Loss: 0.06788322, Cur Avg Loss: 0.15388344, Log Avg loss: 0.23096350, Global Avg Loss: 0.66396488, Time: 0.0209 Steps: 80120, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001357, Sample Num: 21712, Cur Loss: 0.19428283, Cur Avg Loss: 0.15363187, Log Avg loss: 0.11974594, Global Avg Loss: 0.66389697, Time: 0.0209 Steps: 80130, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001367, Sample Num: 21872, Cur Loss: 0.05371177, Cur Avg Loss: 0.15353275, Log Avg loss: 0.14008265, Global Avg Loss: 0.66383160, Time: 0.0209 Steps: 80140, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001377, Sample Num: 22032, Cur Loss: 0.07150590, Cur Avg Loss: 0.15365787, Log Avg loss: 0.17076084, Global Avg Loss: 0.66377009, Time: 0.0209 Steps: 80150, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001387, Sample Num: 22192, Cur Loss: 0.18226451, Cur Avg Loss: 0.15323636, Log Avg loss: 0.09519528, Global Avg Loss: 0.66369916, Time: 0.0209 Steps: 80160, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001397, Sample Num: 22352, Cur Loss: 0.07446084, Cur Avg Loss: 0.15298507, Log Avg loss: 0.11813075, Global Avg Loss: 0.66363110, Time: 0.0209 Steps: 80170, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001407, Sample Num: 22512, Cur Loss: 0.45985976, Cur Avg Loss: 0.15285121, Log Avg loss: 0.13415033, Global Avg Loss: 0.66356507, Time: 0.0210 Steps: 80180, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001417, Sample Num: 22672, Cur Loss: 0.23107472, Cur Avg Loss: 0.15277866, Log Avg loss: 0.14257119, Global Avg Loss: 0.66350010, Time: 0.0210 Steps: 80190, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001427, Sample Num: 22832, Cur Loss: 0.19787879, Cur Avg Loss: 0.15277505, Log Avg loss: 0.15226315, Global Avg Loss: 0.66343635, Time: 0.0209 Steps: 80200, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001437, Sample Num: 22992, Cur Loss: 0.41051683, Cur Avg Loss: 0.15309523, Log Avg loss: 0.19878530, Global Avg Loss: 0.66337842, Time: 0.0209 Steps: 80210, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001447, Sample Num: 23152, Cur Loss: 0.21461661, Cur Avg Loss: 0.15333991, Log Avg loss: 0.18850109, Global Avg Loss: 0.66331923, Time: 0.0209 Steps: 80220, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001457, Sample Num: 23312, Cur Loss: 0.10550611, Cur Avg Loss: 0.15347158, Log Avg loss: 0.17252404, Global Avg Loss: 0.66325805, Time: 0.0209 Steps: 80230, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001467, Sample Num: 23472, Cur Loss: 0.43436086, Cur Avg Loss: 0.15330218, Log Avg loss: 0.12861972, Global Avg Loss: 0.66319142, Time: 0.0210 Steps: 80240, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001477, Sample Num: 23632, Cur Loss: 0.07556840, Cur Avg Loss: 0.15321551, Log Avg loss: 0.14050241, Global Avg Loss: 0.66312629, Time: 0.0209 Steps: 80250, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001487, Sample Num: 23792, Cur Loss: 0.10711957, Cur Avg Loss: 0.15379874, Log Avg loss: 0.23994108, Global Avg Loss: 0.66307356, Time: 0.0209 Steps: 80260, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001497, Sample Num: 23952, Cur Loss: 0.04166814, Cur Avg Loss: 0.15374558, Log Avg loss: 0.14584124, Global Avg Loss: 0.66300913, Time: 0.0210 Steps: 80270, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001507, Sample Num: 24112, Cur Loss: 0.35527611, Cur Avg Loss: 0.15415292, Log Avg loss: 0.21513173, Global Avg Loss: 0.66295334, Time: 0.0210 Steps: 80280, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001517, Sample Num: 24272, Cur Loss: 0.14812118, Cur Avg Loss: 0.15410429, Log Avg loss: 0.14677529, Global Avg Loss: 0.66288905, Time: 0.0210 Steps: 80290, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001527, Sample Num: 24432, Cur Loss: 0.22601467, Cur Avg Loss: 0.15393699, Log Avg loss: 0.12855757, Global Avg Loss: 0.66282251, Time: 0.0209 Steps: 80300, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001537, Sample Num: 24592, Cur Loss: 0.15868567, Cur Avg Loss: 0.15358071, Log Avg loss: 0.09917661, Global Avg Loss: 0.66275232, Time: 0.0246 Steps: 80310, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001547, Sample Num: 24752, Cur Loss: 0.06369484, Cur Avg Loss: 0.15392310, Log Avg loss: 0.20654812, Global Avg Loss: 0.66269552, Time: 0.0210 Steps: 80320, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001557, Sample Num: 24912, Cur Loss: 0.15289064, Cur Avg Loss: 0.15418239, Log Avg loss: 0.19429570, Global Avg Loss: 0.66263722, Time: 0.0209 Steps: 80330, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001567, Sample Num: 25072, Cur Loss: 0.20491570, Cur Avg Loss: 0.15440373, Log Avg loss: 0.18886603, Global Avg Loss: 0.66257824, Time: 0.0209 Steps: 80340, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001577, Sample Num: 25232, Cur Loss: 0.07971073, Cur Avg Loss: 0.15450575, Log Avg loss: 0.17049162, Global Avg Loss: 0.66251700, Time: 0.0209 Steps: 80350, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001587, Sample Num: 25392, Cur Loss: 0.07025381, Cur Avg Loss: 0.15438081, Log Avg loss: 0.13467759, Global Avg Loss: 0.66245132, Time: 0.0209 Steps: 80360, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001597, Sample Num: 25552, Cur Loss: 0.20809965, Cur Avg Loss: 0.15434729, Log Avg loss: 0.14902754, Global Avg Loss: 0.66238743, Time: 0.0209 Steps: 80370, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001607, Sample Num: 25712, Cur Loss: 0.08654906, Cur Avg Loss: 0.15437629, Log Avg loss: 0.15900783, Global Avg Loss: 0.66232481, Time: 0.0209 Steps: 80380, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001617, Sample Num: 25872, Cur Loss: 0.22822008, Cur Avg Loss: 0.15426457, Log Avg loss: 0.13631072, Global Avg Loss: 0.66225938, Time: 0.0209 Steps: 80390, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001627, Sample Num: 26032, Cur Loss: 0.27707347, Cur Avg Loss: 0.15436876, Log Avg loss: 0.17121778, Global Avg Loss: 0.66219830, Time: 0.0209 Steps: 80400, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001637, Sample Num: 26192, Cur Loss: 0.20269710, Cur Avg Loss: 0.15466663, Log Avg loss: 0.20312976, Global Avg Loss: 0.66214121, Time: 0.0209 Steps: 80410, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001647, Sample Num: 26352, Cur Loss: 0.28808016, Cur Avg Loss: 0.15447419, Log Avg loss: 0.12297192, Global Avg Loss: 0.66207417, Time: 0.0209 Steps: 80420, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001657, Sample Num: 26512, Cur Loss: 0.32931840, Cur Avg Loss: 0.15457601, Log Avg loss: 0.17134575, Global Avg Loss: 0.66201315, Time: 0.0209 Steps: 80430, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001667, Sample Num: 26672, Cur Loss: 0.15866184, Cur Avg Loss: 0.15464815, Log Avg loss: 0.16660180, Global Avg Loss: 0.66195157, Time: 0.0209 Steps: 80440, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001677, Sample Num: 26832, Cur Loss: 0.43188751, Cur Avg Loss: 0.15482917, Log Avg loss: 0.18500510, Global Avg Loss: 0.66189228, Time: 0.0209 Steps: 80450, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001687, Sample Num: 26992, Cur Loss: 0.21020238, Cur Avg Loss: 0.15497195, Log Avg loss: 0.17891574, Global Avg Loss: 0.66183225, Time: 0.0210 Steps: 80460, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001697, Sample Num: 27152, Cur Loss: 0.10236613, Cur Avg Loss: 0.15474461, Log Avg loss: 0.11639291, Global Avg Loss: 0.66176447, Time: 0.0208 Steps: 80470, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001707, Sample Num: 27312, Cur Loss: 0.08446288, Cur Avg Loss: 0.15461192, Log Avg loss: 0.13209414, Global Avg Loss: 0.66169866, Time: 0.0209 Steps: 80480, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001717, Sample Num: 27472, Cur Loss: 0.11275352, Cur Avg Loss: 0.15445035, Log Avg loss: 0.12687017, Global Avg Loss: 0.66163221, Time: 0.0209 Steps: 80490, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001727, Sample Num: 27632, Cur Loss: 0.11178884, Cur Avg Loss: 0.15445627, Log Avg loss: 0.15547314, Global Avg Loss: 0.66156934, Time: 0.0209 Steps: 80500, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001737, Sample Num: 27792, Cur Loss: 0.24530050, Cur Avg Loss: 0.15452502, Log Avg loss: 0.16639677, Global Avg Loss: 0.66150783, Time: 0.0209 Steps: 80510, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001747, Sample Num: 27952, Cur Loss: 0.14454147, Cur Avg Loss: 0.15457303, Log Avg loss: 0.16291279, Global Avg Loss: 0.66144591, Time: 0.0210 Steps: 80520, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001757, Sample Num: 28112, Cur Loss: 0.03349928, Cur Avg Loss: 0.15431218, Log Avg loss: 0.10874156, Global Avg Loss: 0.66137728, Time: 0.0209 Steps: 80530, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001767, Sample Num: 28272, Cur Loss: 0.24696413, Cur Avg Loss: 0.15448287, Log Avg loss: 0.18447349, Global Avg Loss: 0.66131806, Time: 0.0209 Steps: 80540, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001777, Sample Num: 28432, Cur Loss: 0.09355759, Cur Avg Loss: 0.15452254, Log Avg loss: 0.16153282, Global Avg Loss: 0.66125602, Time: 0.0209 Steps: 80550, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001787, Sample Num: 28592, Cur Loss: 0.15213922, Cur Avg Loss: 0.15451821, Log Avg loss: 0.15374865, Global Avg Loss: 0.66119302, Time: 0.0209 Steps: 80560, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001797, Sample Num: 28752, Cur Loss: 0.36595249, Cur Avg Loss: 0.15474713, Log Avg loss: 0.19565391, Global Avg Loss: 0.66113524, Time: 0.0246 Steps: 80570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001807, Sample Num: 28912, Cur Loss: 0.07117605, Cur Avg Loss: 0.15465459, Log Avg loss: 0.13802658, Global Avg Loss: 0.66107032, Time: 0.0210 Steps: 80580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001817, Sample Num: 29072, Cur Loss: 0.12377731, Cur Avg Loss: 0.15499343, Log Avg loss: 0.21622072, Global Avg Loss: 0.66101512, Time: 0.0210 Steps: 80590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001827, Sample Num: 29232, Cur Loss: 0.07195457, Cur Avg Loss: 0.15485664, Log Avg loss: 0.13000258, Global Avg Loss: 0.66094924, Time: 0.0210 Steps: 80600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001837, Sample Num: 29392, Cur Loss: 0.07618555, Cur Avg Loss: 0.15457227, Log Avg loss: 0.10261772, Global Avg Loss: 0.66087997, Time: 0.0210 Steps: 80610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001847, Sample Num: 29552, Cur Loss: 0.16028851, Cur Avg Loss: 0.15479494, Log Avg loss: 0.19569865, Global Avg Loss: 0.66082227, Time: 0.0210 Steps: 80620, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001857, Sample Num: 29712, Cur Loss: 0.07100671, Cur Avg Loss: 0.15459162, Log Avg loss: 0.11703862, Global Avg Loss: 0.66075483, Time: 0.0210 Steps: 80630, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001867, Sample Num: 29872, Cur Loss: 0.08842322, Cur Avg Loss: 0.15462896, Log Avg loss: 0.16156318, Global Avg Loss: 0.66069293, Time: 0.0210 Steps: 80640, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001877, Sample Num: 30032, Cur Loss: 0.05251558, Cur Avg Loss: 0.15456765, Log Avg loss: 0.14312154, Global Avg Loss: 0.66062875, Time: 0.0209 Steps: 80650, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001887, Sample Num: 30192, Cur Loss: 0.25147772, Cur Avg Loss: 0.15475336, Log Avg loss: 0.18961111, Global Avg Loss: 0.66057036, Time: 0.0209 Steps: 80660, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001897, Sample Num: 30352, Cur Loss: 0.25321159, Cur Avg Loss: 0.15496492, Log Avg loss: 0.19488671, Global Avg Loss: 0.66051263, Time: 0.0209 Steps: 80670, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001907, Sample Num: 30512, Cur Loss: 0.10355776, Cur Avg Loss: 0.15505425, Log Avg loss: 0.17199965, Global Avg Loss: 0.66045208, Time: 0.0209 Steps: 80680, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001917, Sample Num: 30672, Cur Loss: 0.09169450, Cur Avg Loss: 0.15540570, Log Avg loss: 0.22242705, Global Avg Loss: 0.66039780, Time: 0.0210 Steps: 80690, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001927, Sample Num: 30832, Cur Loss: 0.08386531, Cur Avg Loss: 0.15521254, Log Avg loss: 0.11818490, Global Avg Loss: 0.66033061, Time: 0.0210 Steps: 80700, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001937, Sample Num: 30992, Cur Loss: 0.26221788, Cur Avg Loss: 0.15497364, Log Avg loss: 0.10893681, Global Avg Loss: 0.66026229, Time: 0.0209 Steps: 80710, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001947, Sample Num: 31152, Cur Loss: 0.19720259, Cur Avg Loss: 0.15511739, Log Avg loss: 0.18296228, Global Avg Loss: 0.66020316, Time: 0.0209 Steps: 80720, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001957, Sample Num: 31312, Cur Loss: 0.06338365, Cur Avg Loss: 0.15495959, Log Avg loss: 0.12423522, Global Avg Loss: 0.66013677, Time: 0.0210 Steps: 80730, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001967, Sample Num: 31472, Cur Loss: 0.19430393, Cur Avg Loss: 0.15500616, Log Avg loss: 0.16411906, Global Avg Loss: 0.66007534, Time: 0.0209 Steps: 80740, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001977, Sample Num: 31632, Cur Loss: 0.19535868, Cur Avg Loss: 0.15490221, Log Avg loss: 0.13445543, Global Avg Loss: 0.66001024, Time: 0.0210 Steps: 80750, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001987, Sample Num: 31792, Cur Loss: 0.15073948, Cur Avg Loss: 0.15510626, Log Avg loss: 0.19544732, Global Avg Loss: 0.65995272, Time: 0.0209 Steps: 80760, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001997, Sample Num: 31952, Cur Loss: 0.03597582, Cur Avg Loss: 0.15538433, Log Avg loss: 0.21063655, Global Avg Loss: 0.65989709, Time: 0.0210 Steps: 80770, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002007, Sample Num: 32112, Cur Loss: 0.21437337, Cur Avg Loss: 0.15537593, Log Avg loss: 0.15369881, Global Avg Loss: 0.65983443, Time: 0.0209 Steps: 80780, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002017, Sample Num: 32272, Cur Loss: 0.16271514, Cur Avg Loss: 0.15510094, Log Avg loss: 0.09991036, Global Avg Loss: 0.65976512, Time: 0.0209 Steps: 80790, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002027, Sample Num: 32432, Cur Loss: 0.46835423, Cur Avg Loss: 0.15521398, Log Avg loss: 0.17801367, Global Avg Loss: 0.65970550, Time: 0.0210 Steps: 80800, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002037, Sample Num: 32592, Cur Loss: 0.10625098, Cur Avg Loss: 0.15527021, Log Avg loss: 0.16666804, Global Avg Loss: 0.65964449, Time: 0.0209 Steps: 80810, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002047, Sample Num: 32752, Cur Loss: 0.07889234, Cur Avg Loss: 0.15540646, Log Avg loss: 0.18316177, Global Avg Loss: 0.65958553, Time: 0.0209 Steps: 80820, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002057, Sample Num: 32912, Cur Loss: 0.15325487, Cur Avg Loss: 0.15556114, Log Avg loss: 0.18722398, Global Avg Loss: 0.65952709, Time: 0.0209 Steps: 80830, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002067, Sample Num: 33072, Cur Loss: 0.51405036, Cur Avg Loss: 0.15576501, Log Avg loss: 0.19770175, Global Avg Loss: 0.65946996, Time: 0.0209 Steps: 80840, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002077, Sample Num: 33232, Cur Loss: 0.07357329, Cur Avg Loss: 0.15555811, Log Avg loss: 0.11279188, Global Avg Loss: 0.65940235, Time: 0.0209 Steps: 80850, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002087, Sample Num: 33392, Cur Loss: 0.17705476, Cur Avg Loss: 0.15539369, Log Avg loss: 0.12124371, Global Avg Loss: 0.65933579, Time: 0.0209 Steps: 80860, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002097, Sample Num: 33552, Cur Loss: 0.27751586, Cur Avg Loss: 0.15540233, Log Avg loss: 0.15720511, Global Avg Loss: 0.65927370, Time: 0.0209 Steps: 80870, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002107, Sample Num: 33712, Cur Loss: 0.08008287, Cur Avg Loss: 0.15531870, Log Avg loss: 0.13778126, Global Avg Loss: 0.65920922, Time: 0.0210 Steps: 80880, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002117, Sample Num: 33872, Cur Loss: 0.14441068, Cur Avg Loss: 0.15519899, Log Avg loss: 0.12997521, Global Avg Loss: 0.65914380, Time: 0.0210 Steps: 80890, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002127, Sample Num: 34032, Cur Loss: 0.10135336, Cur Avg Loss: 0.15530730, Log Avg loss: 0.17823751, Global Avg Loss: 0.65908435, Time: 0.0209 Steps: 80900, Updated lr: 0.000024 ***** Running evaluation checkpoint-80902 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-80902 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.878249, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.246254, "eval_total_loss": 173.116495, "eval_mae": 0.322368, "eval_mse": 0.246349, "eval_r2": 0.843404, "eval_sp_statistic": 0.927966, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.933017, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.177546, "test_total_loss": 89.127861, "test_mae": 0.301672, "test_mse": 0.177609, "test_r2": 0.885369, "test_sp_statistic": 0.9198, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.948269, "test_ps_pvalue": 0.0, "lr": 2.4227596017069703e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6590740039974994, "train_cur_epoch_loss": 330.819563661702, "train_cur_epoch_avg_loss": 0.15538730092141945, "train_cur_epoch_time": 44.878249168395996, "train_cur_epoch_avg_time": 0.021079497026019727, "epoch": 38, "step": 80902} ################################################## Training, Epoch: 0039, Batch: 000008, Sample Num: 128, Cur Loss: 0.15379539, Cur Avg Loss: 0.16493852, Log Avg loss: 0.18004405, Global Avg Loss: 0.65902515, Time: 0.0211 Steps: 80910, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000018, Sample Num: 288, Cur Loss: 0.07407039, Cur Avg Loss: 0.13817022, Log Avg loss: 0.11675559, Global Avg Loss: 0.65895813, Time: 0.0210 Steps: 80920, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000028, Sample Num: 448, Cur Loss: 0.17493710, Cur Avg Loss: 0.12788906, Log Avg loss: 0.10938296, Global Avg Loss: 0.65889023, Time: 0.0210 Steps: 80930, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000038, Sample Num: 608, Cur Loss: 0.12689547, Cur Avg Loss: 0.12539678, Log Avg loss: 0.11841841, Global Avg Loss: 0.65882345, Time: 0.0210 Steps: 80940, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000048, Sample Num: 768, Cur Loss: 0.05010149, Cur Avg Loss: 0.12770774, Log Avg loss: 0.13648939, Global Avg Loss: 0.65875893, Time: 0.0210 Steps: 80950, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000058, Sample Num: 928, Cur Loss: 0.19043219, Cur Avg Loss: 0.13387638, Log Avg loss: 0.16348584, Global Avg Loss: 0.65869775, Time: 0.0210 Steps: 80960, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000068, Sample Num: 1088, Cur Loss: 0.06596956, Cur Avg Loss: 0.13172041, Log Avg loss: 0.11921577, Global Avg Loss: 0.65863112, Time: 0.0210 Steps: 80970, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000078, Sample Num: 1248, Cur Loss: 0.16966408, Cur Avg Loss: 0.13188426, Log Avg loss: 0.13299843, Global Avg Loss: 0.65856621, Time: 0.0210 Steps: 80980, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000088, Sample Num: 1408, Cur Loss: 0.43640542, Cur Avg Loss: 0.13663081, Log Avg loss: 0.17365389, Global Avg Loss: 0.65850634, Time: 0.0210 Steps: 80990, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000098, Sample Num: 1568, Cur Loss: 0.10972448, Cur Avg Loss: 0.14049832, Log Avg loss: 0.17453242, Global Avg Loss: 0.65844659, Time: 0.0210 Steps: 81000, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000108, Sample Num: 1728, Cur Loss: 0.16554430, Cur Avg Loss: 0.13929136, Log Avg loss: 0.12746312, Global Avg Loss: 0.65838105, Time: 0.0210 Steps: 81010, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000118, Sample Num: 1888, Cur Loss: 0.04095289, Cur Avg Loss: 0.13852073, Log Avg loss: 0.13019798, Global Avg Loss: 0.65831585, Time: 0.0210 Steps: 81020, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000128, Sample Num: 2048, Cur Loss: 0.44138494, Cur Avg Loss: 0.14015883, Log Avg loss: 0.15948843, Global Avg Loss: 0.65825429, Time: 0.0210 Steps: 81030, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000138, Sample Num: 2208, Cur Loss: 0.03073639, Cur Avg Loss: 0.14050716, Log Avg loss: 0.14496583, Global Avg Loss: 0.65819096, Time: 0.0210 Steps: 81040, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000148, Sample Num: 2368, Cur Loss: 0.07413995, Cur Avg Loss: 0.14319416, Log Avg loss: 0.18027477, Global Avg Loss: 0.65813199, Time: 0.0210 Steps: 81050, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000158, Sample Num: 2528, Cur Loss: 0.19188268, Cur Avg Loss: 0.14416093, Log Avg loss: 0.15846901, Global Avg Loss: 0.65807035, Time: 0.0210 Steps: 81060, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000168, Sample Num: 2688, Cur Loss: 0.11534706, Cur Avg Loss: 0.14272549, Log Avg loss: 0.12004555, Global Avg Loss: 0.65800398, Time: 0.0210 Steps: 81070, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000178, Sample Num: 2848, Cur Loss: 0.37834287, Cur Avg Loss: 0.14871946, Log Avg loss: 0.24941811, Global Avg Loss: 0.65795359, Time: 0.0210 Steps: 81080, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000188, Sample Num: 3008, Cur Loss: 0.10275275, Cur Avg Loss: 0.14760886, Log Avg loss: 0.12784034, Global Avg Loss: 0.65788822, Time: 0.0209 Steps: 81090, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000198, Sample Num: 3168, Cur Loss: 0.10269160, Cur Avg Loss: 0.14551215, Log Avg loss: 0.10609401, Global Avg Loss: 0.65782018, Time: 0.0210 Steps: 81100, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000208, Sample Num: 3328, Cur Loss: 0.12897559, Cur Avg Loss: 0.14767222, Log Avg loss: 0.19044146, Global Avg Loss: 0.65776256, Time: 0.0210 Steps: 81110, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000218, Sample Num: 3488, Cur Loss: 0.11574239, Cur Avg Loss: 0.14812606, Log Avg loss: 0.15756589, Global Avg Loss: 0.65770089, Time: 0.0210 Steps: 81120, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000228, Sample Num: 3648, Cur Loss: 0.70538533, Cur Avg Loss: 0.15205324, Log Avg loss: 0.23766585, Global Avg Loss: 0.65764912, Time: 0.0210 Steps: 81130, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000238, Sample Num: 3808, Cur Loss: 0.11680090, Cur Avg Loss: 0.15299909, Log Avg loss: 0.17456446, Global Avg Loss: 0.65758958, Time: 0.0210 Steps: 81140, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000248, Sample Num: 3968, Cur Loss: 0.04302449, Cur Avg Loss: 0.15116057, Log Avg loss: 0.10740377, Global Avg Loss: 0.65752179, Time: 0.0210 Steps: 81150, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000258, Sample Num: 4128, Cur Loss: 0.54518104, Cur Avg Loss: 0.14987904, Log Avg loss: 0.11809720, Global Avg Loss: 0.65745532, Time: 0.0246 Steps: 81160, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000268, Sample Num: 4288, Cur Loss: 0.08746557, Cur Avg Loss: 0.14950350, Log Avg loss: 0.13981458, Global Avg Loss: 0.65739155, Time: 0.0209 Steps: 81170, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000278, Sample Num: 4448, Cur Loss: 0.10945356, Cur Avg Loss: 0.14856161, Log Avg loss: 0.12331878, Global Avg Loss: 0.65732576, Time: 0.0209 Steps: 81180, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000288, Sample Num: 4608, Cur Loss: 0.22840238, Cur Avg Loss: 0.14933801, Log Avg loss: 0.17092216, Global Avg Loss: 0.65726585, Time: 0.0209 Steps: 81190, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000298, Sample Num: 4768, Cur Loss: 0.05021263, Cur Avg Loss: 0.14843463, Log Avg loss: 0.12241711, Global Avg Loss: 0.65719998, Time: 0.0209 Steps: 81200, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000308, Sample Num: 4928, Cur Loss: 0.08591966, Cur Avg Loss: 0.14891488, Log Avg loss: 0.16322645, Global Avg Loss: 0.65713916, Time: 0.0209 Steps: 81210, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000318, Sample Num: 5088, Cur Loss: 0.11066893, Cur Avg Loss: 0.14909642, Log Avg loss: 0.15468765, Global Avg Loss: 0.65707729, Time: 0.0209 Steps: 81220, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000328, Sample Num: 5248, Cur Loss: 0.08768249, Cur Avg Loss: 0.14918675, Log Avg loss: 0.15205955, Global Avg Loss: 0.65701512, Time: 0.0209 Steps: 81230, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000338, Sample Num: 5408, Cur Loss: 0.07946071, Cur Avg Loss: 0.14840614, Log Avg loss: 0.12280201, Global Avg Loss: 0.65694936, Time: 0.0208 Steps: 81240, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000348, Sample Num: 5568, Cur Loss: 0.10085362, Cur Avg Loss: 0.14744047, Log Avg loss: 0.11480074, Global Avg Loss: 0.65688264, Time: 0.0209 Steps: 81250, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000358, Sample Num: 5728, Cur Loss: 0.22459766, Cur Avg Loss: 0.14868410, Log Avg loss: 0.19196240, Global Avg Loss: 0.65682542, Time: 0.0208 Steps: 81260, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000368, Sample Num: 5888, Cur Loss: 0.07564747, Cur Avg Loss: 0.14941327, Log Avg loss: 0.17551746, Global Avg Loss: 0.65676620, Time: 0.0209 Steps: 81270, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000378, Sample Num: 6048, Cur Loss: 0.16178605, Cur Avg Loss: 0.14911580, Log Avg loss: 0.13816913, Global Avg Loss: 0.65670240, Time: 0.0209 Steps: 81280, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000388, Sample Num: 6208, Cur Loss: 0.08401199, Cur Avg Loss: 0.14887593, Log Avg loss: 0.13980891, Global Avg Loss: 0.65663881, Time: 0.0209 Steps: 81290, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000398, Sample Num: 6368, Cur Loss: 0.12296313, Cur Avg Loss: 0.14780511, Log Avg loss: 0.10625726, Global Avg Loss: 0.65657111, Time: 0.0209 Steps: 81300, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000408, Sample Num: 6528, Cur Loss: 0.08725820, Cur Avg Loss: 0.14811364, Log Avg loss: 0.16039310, Global Avg Loss: 0.65651009, Time: 0.0209 Steps: 81310, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000418, Sample Num: 6688, Cur Loss: 0.17749834, Cur Avg Loss: 0.14824575, Log Avg loss: 0.15363566, Global Avg Loss: 0.65644825, Time: 0.0208 Steps: 81320, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000428, Sample Num: 6848, Cur Loss: 0.16764157, Cur Avg Loss: 0.14844474, Log Avg loss: 0.15676268, Global Avg Loss: 0.65638681, Time: 0.0209 Steps: 81330, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000438, Sample Num: 7008, Cur Loss: 0.03896974, Cur Avg Loss: 0.14852306, Log Avg loss: 0.15187518, Global Avg Loss: 0.65632479, Time: 0.0209 Steps: 81340, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000448, Sample Num: 7168, Cur Loss: 0.20190929, Cur Avg Loss: 0.14750520, Log Avg loss: 0.10292268, Global Avg Loss: 0.65625676, Time: 0.0209 Steps: 81350, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000458, Sample Num: 7328, Cur Loss: 0.07732090, Cur Avg Loss: 0.14650411, Log Avg loss: 0.10165534, Global Avg Loss: 0.65618859, Time: 0.0209 Steps: 81360, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000468, Sample Num: 7488, Cur Loss: 0.24510285, Cur Avg Loss: 0.14622002, Log Avg loss: 0.13320901, Global Avg Loss: 0.65612432, Time: 0.0209 Steps: 81370, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000478, Sample Num: 7648, Cur Loss: 0.20533095, Cur Avg Loss: 0.14576608, Log Avg loss: 0.12452157, Global Avg Loss: 0.65605900, Time: 0.0209 Steps: 81380, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000488, Sample Num: 7808, Cur Loss: 0.25557524, Cur Avg Loss: 0.14512955, Log Avg loss: 0.11470322, Global Avg Loss: 0.65599248, Time: 0.0209 Steps: 81390, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000498, Sample Num: 7968, Cur Loss: 0.06527841, Cur Avg Loss: 0.14496187, Log Avg loss: 0.13677919, Global Avg Loss: 0.65592870, Time: 0.0209 Steps: 81400, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000508, Sample Num: 8128, Cur Loss: 0.48114386, Cur Avg Loss: 0.14596072, Log Avg loss: 0.19570330, Global Avg Loss: 0.65587217, Time: 0.0209 Steps: 81410, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000518, Sample Num: 8288, Cur Loss: 0.10910772, Cur Avg Loss: 0.14643638, Log Avg loss: 0.17060033, Global Avg Loss: 0.65581257, Time: 0.0210 Steps: 81420, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000528, Sample Num: 8448, Cur Loss: 0.11269040, Cur Avg Loss: 0.14685932, Log Avg loss: 0.16876760, Global Avg Loss: 0.65575275, Time: 0.0209 Steps: 81430, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000538, Sample Num: 8608, Cur Loss: 0.08919396, Cur Avg Loss: 0.14624641, Log Avg loss: 0.11388448, Global Avg Loss: 0.65568622, Time: 0.0209 Steps: 81440, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000548, Sample Num: 8768, Cur Loss: 0.06116688, Cur Avg Loss: 0.14669366, Log Avg loss: 0.17075575, Global Avg Loss: 0.65562668, Time: 0.0209 Steps: 81450, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000558, Sample Num: 8928, Cur Loss: 0.10032488, Cur Avg Loss: 0.14659725, Log Avg loss: 0.14131382, Global Avg Loss: 0.65556354, Time: 0.0209 Steps: 81460, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000568, Sample Num: 9088, Cur Loss: 0.12893194, Cur Avg Loss: 0.14591785, Log Avg loss: 0.10800776, Global Avg Loss: 0.65549634, Time: 0.0209 Steps: 81470, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000578, Sample Num: 9248, Cur Loss: 0.10745145, Cur Avg Loss: 0.14578587, Log Avg loss: 0.13828918, Global Avg Loss: 0.65543286, Time: 0.0209 Steps: 81480, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000588, Sample Num: 9408, Cur Loss: 0.43229955, Cur Avg Loss: 0.14585254, Log Avg loss: 0.14970580, Global Avg Loss: 0.65537080, Time: 0.0209 Steps: 81490, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000598, Sample Num: 9568, Cur Loss: 0.12744604, Cur Avg Loss: 0.14548987, Log Avg loss: 0.12416481, Global Avg Loss: 0.65530562, Time: 0.0209 Steps: 81500, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000608, Sample Num: 9728, Cur Loss: 0.28980672, Cur Avg Loss: 0.14566113, Log Avg loss: 0.15590261, Global Avg Loss: 0.65524435, Time: 0.0209 Steps: 81510, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000618, Sample Num: 9888, Cur Loss: 0.29663187, Cur Avg Loss: 0.14541463, Log Avg loss: 0.13042780, Global Avg Loss: 0.65517997, Time: 0.0209 Steps: 81520, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000628, Sample Num: 10048, Cur Loss: 0.23162505, Cur Avg Loss: 0.14558383, Log Avg loss: 0.15604012, Global Avg Loss: 0.65511875, Time: 0.0210 Steps: 81530, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000638, Sample Num: 10208, Cur Loss: 0.05206218, Cur Avg Loss: 0.14661460, Log Avg loss: 0.21134675, Global Avg Loss: 0.65506433, Time: 0.0209 Steps: 81540, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000648, Sample Num: 10368, Cur Loss: 0.05330594, Cur Avg Loss: 0.14593843, Log Avg loss: 0.10279890, Global Avg Loss: 0.65499661, Time: 0.0209 Steps: 81550, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000658, Sample Num: 10528, Cur Loss: 0.19505905, Cur Avg Loss: 0.14557027, Log Avg loss: 0.12171382, Global Avg Loss: 0.65493122, Time: 0.0209 Steps: 81560, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000668, Sample Num: 10688, Cur Loss: 0.09408246, Cur Avg Loss: 0.14503912, Log Avg loss: 0.11008957, Global Avg Loss: 0.65486443, Time: 0.0209 Steps: 81570, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000678, Sample Num: 10848, Cur Loss: 0.17966843, Cur Avg Loss: 0.14475631, Log Avg loss: 0.12586461, Global Avg Loss: 0.65479958, Time: 0.0209 Steps: 81580, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000688, Sample Num: 11008, Cur Loss: 0.04839552, Cur Avg Loss: 0.14506269, Log Avg loss: 0.16583497, Global Avg Loss: 0.65473965, Time: 0.0209 Steps: 81590, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000698, Sample Num: 11168, Cur Loss: 0.32675165, Cur Avg Loss: 0.14573421, Log Avg loss: 0.19193450, Global Avg Loss: 0.65468294, Time: 0.0209 Steps: 81600, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000708, Sample Num: 11328, Cur Loss: 0.28330326, Cur Avg Loss: 0.14743906, Log Avg loss: 0.26643803, Global Avg Loss: 0.65463536, Time: 0.0209 Steps: 81610, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000718, Sample Num: 11488, Cur Loss: 0.05264190, Cur Avg Loss: 0.14667409, Log Avg loss: 0.09251384, Global Avg Loss: 0.65456649, Time: 0.0209 Steps: 81620, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000728, Sample Num: 11648, Cur Loss: 0.15313594, Cur Avg Loss: 0.14735446, Log Avg loss: 0.19620528, Global Avg Loss: 0.65451034, Time: 0.0209 Steps: 81630, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000738, Sample Num: 11808, Cur Loss: 0.27527720, Cur Avg Loss: 0.14749332, Log Avg loss: 0.15760225, Global Avg Loss: 0.65444948, Time: 0.0209 Steps: 81640, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000748, Sample Num: 11968, Cur Loss: 0.17633657, Cur Avg Loss: 0.14715108, Log Avg loss: 0.12189361, Global Avg Loss: 0.65438425, Time: 0.0209 Steps: 81650, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000758, Sample Num: 12128, Cur Loss: 0.11983953, Cur Avg Loss: 0.14781141, Log Avg loss: 0.19720460, Global Avg Loss: 0.65432827, Time: 0.0209 Steps: 81660, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000768, Sample Num: 12288, Cur Loss: 0.05438981, Cur Avg Loss: 0.14747551, Log Avg loss: 0.12201369, Global Avg Loss: 0.65426309, Time: 0.0255 Steps: 81670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000778, Sample Num: 12448, Cur Loss: 0.13049856, Cur Avg Loss: 0.14811834, Log Avg loss: 0.19748816, Global Avg Loss: 0.65420716, Time: 0.0208 Steps: 81680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000788, Sample Num: 12608, Cur Loss: 0.06158414, Cur Avg Loss: 0.14769761, Log Avg loss: 0.11496466, Global Avg Loss: 0.65414115, Time: 0.0209 Steps: 81690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000798, Sample Num: 12768, Cur Loss: 0.03303470, Cur Avg Loss: 0.14705823, Log Avg loss: 0.09667471, Global Avg Loss: 0.65407292, Time: 0.0208 Steps: 81700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000808, Sample Num: 12928, Cur Loss: 0.04612289, Cur Avg Loss: 0.14776721, Log Avg loss: 0.20434410, Global Avg Loss: 0.65401788, Time: 0.0208 Steps: 81710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000818, Sample Num: 13088, Cur Loss: 0.06080066, Cur Avg Loss: 0.14815659, Log Avg loss: 0.17961872, Global Avg Loss: 0.65395983, Time: 0.0208 Steps: 81720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000828, Sample Num: 13248, Cur Loss: 0.08998674, Cur Avg Loss: 0.14820973, Log Avg loss: 0.15255667, Global Avg Loss: 0.65389848, Time: 0.0208 Steps: 81730, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000838, Sample Num: 13408, Cur Loss: 0.10812223, Cur Avg Loss: 0.14801474, Log Avg loss: 0.13186894, Global Avg Loss: 0.65383461, Time: 0.0208 Steps: 81740, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000848, Sample Num: 13568, Cur Loss: 0.03247739, Cur Avg Loss: 0.14756434, Log Avg loss: 0.10982143, Global Avg Loss: 0.65376807, Time: 0.0208 Steps: 81750, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000858, Sample Num: 13728, Cur Loss: 0.07363315, Cur Avg Loss: 0.14745652, Log Avg loss: 0.13831286, Global Avg Loss: 0.65370502, Time: 0.0209 Steps: 81760, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000868, Sample Num: 13888, Cur Loss: 0.15283227, Cur Avg Loss: 0.14747956, Log Avg loss: 0.14945624, Global Avg Loss: 0.65364336, Time: 0.0208 Steps: 81770, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000878, Sample Num: 14048, Cur Loss: 0.29641774, Cur Avg Loss: 0.14776609, Log Avg loss: 0.17263691, Global Avg Loss: 0.65358454, Time: 0.0208 Steps: 81780, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000888, Sample Num: 14208, Cur Loss: 0.13115662, Cur Avg Loss: 0.14740756, Log Avg loss: 0.11592914, Global Avg Loss: 0.65351880, Time: 0.0208 Steps: 81790, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000898, Sample Num: 14368, Cur Loss: 0.19548127, Cur Avg Loss: 0.14750654, Log Avg loss: 0.15629596, Global Avg Loss: 0.65345802, Time: 0.0209 Steps: 81800, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000908, Sample Num: 14528, Cur Loss: 0.09710635, Cur Avg Loss: 0.14724543, Log Avg loss: 0.12379763, Global Avg Loss: 0.65339328, Time: 0.0208 Steps: 81810, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000918, Sample Num: 14688, Cur Loss: 0.15153533, Cur Avg Loss: 0.14707544, Log Avg loss: 0.13164002, Global Avg Loss: 0.65332951, Time: 0.0208 Steps: 81820, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000928, Sample Num: 14848, Cur Loss: 0.10939777, Cur Avg Loss: 0.14728564, Log Avg loss: 0.16658229, Global Avg Loss: 0.65327002, Time: 0.0208 Steps: 81830, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000938, Sample Num: 15008, Cur Loss: 0.19391437, Cur Avg Loss: 0.14742020, Log Avg loss: 0.15990755, Global Avg Loss: 0.65320974, Time: 0.0208 Steps: 81840, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000948, Sample Num: 15168, Cur Loss: 0.07887040, Cur Avg Loss: 0.14774869, Log Avg loss: 0.17856062, Global Avg Loss: 0.65315175, Time: 0.0208 Steps: 81850, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000958, Sample Num: 15328, Cur Loss: 0.34405062, Cur Avg Loss: 0.14828163, Log Avg loss: 0.19880470, Global Avg Loss: 0.65309625, Time: 0.0208 Steps: 81860, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000968, Sample Num: 15488, Cur Loss: 0.13013501, Cur Avg Loss: 0.14797135, Log Avg loss: 0.11824674, Global Avg Loss: 0.65303092, Time: 0.0208 Steps: 81870, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000978, Sample Num: 15648, Cur Loss: 0.06889402, Cur Avg Loss: 0.14752509, Log Avg loss: 0.10432716, Global Avg Loss: 0.65296391, Time: 0.0208 Steps: 81880, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000988, Sample Num: 15808, Cur Loss: 0.25393152, Cur Avg Loss: 0.14761462, Log Avg loss: 0.15637028, Global Avg Loss: 0.65290326, Time: 0.0208 Steps: 81890, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000998, Sample Num: 15968, Cur Loss: 0.05119967, Cur Avg Loss: 0.14730316, Log Avg loss: 0.11653129, Global Avg Loss: 0.65283777, Time: 0.0208 Steps: 81900, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001008, Sample Num: 16128, Cur Loss: 0.47017264, Cur Avg Loss: 0.14767047, Log Avg loss: 0.18432780, Global Avg Loss: 0.65278058, Time: 0.0208 Steps: 81910, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001018, Sample Num: 16288, Cur Loss: 0.50325662, Cur Avg Loss: 0.14783835, Log Avg loss: 0.16476085, Global Avg Loss: 0.65272100, Time: 0.0208 Steps: 81920, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001028, Sample Num: 16448, Cur Loss: 0.14752686, Cur Avg Loss: 0.14830572, Log Avg loss: 0.19588395, Global Avg Loss: 0.65266524, Time: 0.0247 Steps: 81930, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001038, Sample Num: 16608, Cur Loss: 0.24301451, Cur Avg Loss: 0.14801060, Log Avg loss: 0.11767215, Global Avg Loss: 0.65259995, Time: 0.0209 Steps: 81940, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001048, Sample Num: 16768, Cur Loss: 0.09528758, Cur Avg Loss: 0.14777313, Log Avg loss: 0.12312315, Global Avg Loss: 0.65253534, Time: 0.0209 Steps: 81950, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001058, Sample Num: 16928, Cur Loss: 0.37102786, Cur Avg Loss: 0.14771619, Log Avg loss: 0.14174890, Global Avg Loss: 0.65247302, Time: 0.0209 Steps: 81960, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001068, Sample Num: 17088, Cur Loss: 0.13164051, Cur Avg Loss: 0.14717590, Log Avg loss: 0.09001355, Global Avg Loss: 0.65240440, Time: 0.0209 Steps: 81970, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001078, Sample Num: 17248, Cur Loss: 0.24111703, Cur Avg Loss: 0.14718160, Log Avg loss: 0.14779048, Global Avg Loss: 0.65234285, Time: 0.0209 Steps: 81980, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001088, Sample Num: 17408, Cur Loss: 0.03786776, Cur Avg Loss: 0.14702757, Log Avg loss: 0.13042348, Global Avg Loss: 0.65227919, Time: 0.0209 Steps: 81990, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001098, Sample Num: 17568, Cur Loss: 0.17007893, Cur Avg Loss: 0.14724695, Log Avg loss: 0.17111460, Global Avg Loss: 0.65222051, Time: 0.0210 Steps: 82000, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001108, Sample Num: 17728, Cur Loss: 0.09815834, Cur Avg Loss: 0.14726030, Log Avg loss: 0.14872672, Global Avg Loss: 0.65215912, Time: 0.0209 Steps: 82010, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001118, Sample Num: 17888, Cur Loss: 0.02651352, Cur Avg Loss: 0.14713032, Log Avg loss: 0.13272836, Global Avg Loss: 0.65209579, Time: 0.0209 Steps: 82020, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001128, Sample Num: 18048, Cur Loss: 0.14119415, Cur Avg Loss: 0.14708349, Log Avg loss: 0.14184840, Global Avg Loss: 0.65203359, Time: 0.0209 Steps: 82030, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001138, Sample Num: 18208, Cur Loss: 0.14803296, Cur Avg Loss: 0.14690829, Log Avg loss: 0.12714587, Global Avg Loss: 0.65196961, Time: 0.0210 Steps: 82040, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001148, Sample Num: 18368, Cur Loss: 0.06200785, Cur Avg Loss: 0.14734030, Log Avg loss: 0.19650261, Global Avg Loss: 0.65191410, Time: 0.0209 Steps: 82050, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001158, Sample Num: 18528, Cur Loss: 0.12230164, Cur Avg Loss: 0.14767866, Log Avg loss: 0.18652205, Global Avg Loss: 0.65185738, Time: 0.0209 Steps: 82060, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001168, Sample Num: 18688, Cur Loss: 0.18220904, Cur Avg Loss: 0.14792484, Log Avg loss: 0.17643231, Global Avg Loss: 0.65179946, Time: 0.0209 Steps: 82070, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001178, Sample Num: 18848, Cur Loss: 0.03876188, Cur Avg Loss: 0.14823885, Log Avg loss: 0.18491616, Global Avg Loss: 0.65174257, Time: 0.0209 Steps: 82080, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001188, Sample Num: 19008, Cur Loss: 0.19124852, Cur Avg Loss: 0.14875867, Log Avg loss: 0.20999278, Global Avg Loss: 0.65168876, Time: 0.0210 Steps: 82090, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001198, Sample Num: 19168, Cur Loss: 0.16778010, Cur Avg Loss: 0.14885252, Log Avg loss: 0.16000184, Global Avg Loss: 0.65162887, Time: 0.0209 Steps: 82100, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001208, Sample Num: 19328, Cur Loss: 0.11818682, Cur Avg Loss: 0.14867005, Log Avg loss: 0.12681044, Global Avg Loss: 0.65156496, Time: 0.0209 Steps: 82110, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001218, Sample Num: 19488, Cur Loss: 0.18549713, Cur Avg Loss: 0.14904804, Log Avg loss: 0.19470924, Global Avg Loss: 0.65150932, Time: 0.0213 Steps: 82120, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001228, Sample Num: 19648, Cur Loss: 0.25694504, Cur Avg Loss: 0.14961797, Log Avg loss: 0.21903583, Global Avg Loss: 0.65145667, Time: 0.0209 Steps: 82130, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001238, Sample Num: 19808, Cur Loss: 0.11684300, Cur Avg Loss: 0.14944419, Log Avg loss: 0.12810394, Global Avg Loss: 0.65139295, Time: 0.0209 Steps: 82140, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001248, Sample Num: 19968, Cur Loss: 0.05710396, Cur Avg Loss: 0.14956608, Log Avg loss: 0.16465584, Global Avg Loss: 0.65133370, Time: 0.0209 Steps: 82150, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001258, Sample Num: 20128, Cur Loss: 0.08182941, Cur Avg Loss: 0.14984236, Log Avg loss: 0.18432132, Global Avg Loss: 0.65127686, Time: 0.0209 Steps: 82160, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001268, Sample Num: 20288, Cur Loss: 0.04082943, Cur Avg Loss: 0.14968350, Log Avg loss: 0.12969937, Global Avg Loss: 0.65121338, Time: 0.0209 Steps: 82170, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001278, Sample Num: 20448, Cur Loss: 0.09688563, Cur Avg Loss: 0.14972295, Log Avg loss: 0.15472535, Global Avg Loss: 0.65115297, Time: 0.0208 Steps: 82180, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001288, Sample Num: 20608, Cur Loss: 0.18429536, Cur Avg Loss: 0.14991470, Log Avg loss: 0.17442015, Global Avg Loss: 0.65109497, Time: 0.0209 Steps: 82190, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001298, Sample Num: 20768, Cur Loss: 0.12054683, Cur Avg Loss: 0.15032701, Log Avg loss: 0.20343282, Global Avg Loss: 0.65104051, Time: 0.0208 Steps: 82200, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001308, Sample Num: 20928, Cur Loss: 0.07653689, Cur Avg Loss: 0.14978766, Log Avg loss: 0.07977942, Global Avg Loss: 0.65097102, Time: 0.0209 Steps: 82210, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001318, Sample Num: 21088, Cur Loss: 0.08126763, Cur Avg Loss: 0.14950264, Log Avg loss: 0.11222203, Global Avg Loss: 0.65090549, Time: 0.0209 Steps: 82220, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001328, Sample Num: 21248, Cur Loss: 0.05703150, Cur Avg Loss: 0.14934600, Log Avg loss: 0.12870091, Global Avg Loss: 0.65084199, Time: 0.0208 Steps: 82230, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001338, Sample Num: 21408, Cur Loss: 0.09268828, Cur Avg Loss: 0.14949844, Log Avg loss: 0.16974288, Global Avg Loss: 0.65078349, Time: 0.0208 Steps: 82240, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001348, Sample Num: 21568, Cur Loss: 0.43332574, Cur Avg Loss: 0.14980969, Log Avg loss: 0.19145456, Global Avg Loss: 0.65072764, Time: 0.0208 Steps: 82250, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001358, Sample Num: 21728, Cur Loss: 0.15888579, Cur Avg Loss: 0.14966618, Log Avg loss: 0.13032208, Global Avg Loss: 0.65066438, Time: 0.0209 Steps: 82260, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001368, Sample Num: 21888, Cur Loss: 0.18261386, Cur Avg Loss: 0.14989112, Log Avg loss: 0.18043713, Global Avg Loss: 0.65060722, Time: 0.0208 Steps: 82270, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001378, Sample Num: 22048, Cur Loss: 0.19675346, Cur Avg Loss: 0.15013366, Log Avg loss: 0.18331376, Global Avg Loss: 0.65055043, Time: 0.0209 Steps: 82280, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001388, Sample Num: 22208, Cur Loss: 0.06333985, Cur Avg Loss: 0.14998259, Log Avg loss: 0.12916474, Global Avg Loss: 0.65048707, Time: 0.0208 Steps: 82290, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001398, Sample Num: 22368, Cur Loss: 0.16346151, Cur Avg Loss: 0.15000636, Log Avg loss: 0.15330577, Global Avg Loss: 0.65042666, Time: 0.0208 Steps: 82300, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001408, Sample Num: 22528, Cur Loss: 0.13029078, Cur Avg Loss: 0.14990140, Log Avg loss: 0.13522818, Global Avg Loss: 0.65036407, Time: 0.0208 Steps: 82310, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001418, Sample Num: 22688, Cur Loss: 0.19755772, Cur Avg Loss: 0.15012597, Log Avg loss: 0.18174571, Global Avg Loss: 0.65030714, Time: 0.0208 Steps: 82320, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001428, Sample Num: 22848, Cur Loss: 0.37514353, Cur Avg Loss: 0.15038806, Log Avg loss: 0.18755242, Global Avg Loss: 0.65025093, Time: 0.0208 Steps: 82330, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001438, Sample Num: 23008, Cur Loss: 0.08156763, Cur Avg Loss: 0.15039425, Log Avg loss: 0.15127806, Global Avg Loss: 0.65019033, Time: 0.0208 Steps: 82340, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001448, Sample Num: 23168, Cur Loss: 0.19643348, Cur Avg Loss: 0.15048849, Log Avg loss: 0.16403920, Global Avg Loss: 0.65013130, Time: 0.0209 Steps: 82350, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001458, Sample Num: 23328, Cur Loss: 0.03002287, Cur Avg Loss: 0.15012625, Log Avg loss: 0.09767430, Global Avg Loss: 0.65006422, Time: 0.0208 Steps: 82360, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001468, Sample Num: 23488, Cur Loss: 0.31619513, Cur Avg Loss: 0.15090340, Log Avg loss: 0.26421168, Global Avg Loss: 0.65001738, Time: 0.0208 Steps: 82370, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001478, Sample Num: 23648, Cur Loss: 0.27791488, Cur Avg Loss: 0.15102079, Log Avg loss: 0.16825424, Global Avg Loss: 0.64995890, Time: 0.0209 Steps: 82380, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001488, Sample Num: 23808, Cur Loss: 0.12629628, Cur Avg Loss: 0.15098820, Log Avg loss: 0.14617131, Global Avg Loss: 0.64989775, Time: 0.0209 Steps: 82390, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001498, Sample Num: 23968, Cur Loss: 0.14432967, Cur Avg Loss: 0.15084464, Log Avg loss: 0.12948203, Global Avg Loss: 0.64983459, Time: 0.0208 Steps: 82400, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001508, Sample Num: 24128, Cur Loss: 0.13596752, Cur Avg Loss: 0.15057973, Log Avg loss: 0.11089670, Global Avg Loss: 0.64976919, Time: 0.0208 Steps: 82410, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001518, Sample Num: 24288, Cur Loss: 0.06430510, Cur Avg Loss: 0.15048921, Log Avg loss: 0.13683868, Global Avg Loss: 0.64970696, Time: 0.0208 Steps: 82420, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001528, Sample Num: 24448, Cur Loss: 0.38657737, Cur Avg Loss: 0.15047195, Log Avg loss: 0.14785285, Global Avg Loss: 0.64964608, Time: 0.0208 Steps: 82430, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001538, Sample Num: 24608, Cur Loss: 0.12432554, Cur Avg Loss: 0.15055092, Log Avg loss: 0.16261642, Global Avg Loss: 0.64958700, Time: 0.0247 Steps: 82440, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001548, Sample Num: 24768, Cur Loss: 0.21976018, Cur Avg Loss: 0.15064961, Log Avg loss: 0.16582926, Global Avg Loss: 0.64952833, Time: 0.0209 Steps: 82450, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001558, Sample Num: 24928, Cur Loss: 0.08891498, Cur Avg Loss: 0.15071853, Log Avg loss: 0.16138700, Global Avg Loss: 0.64946913, Time: 0.0208 Steps: 82460, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001568, Sample Num: 25088, Cur Loss: 0.13316697, Cur Avg Loss: 0.15081451, Log Avg loss: 0.16576834, Global Avg Loss: 0.64941048, Time: 0.0209 Steps: 82470, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001578, Sample Num: 25248, Cur Loss: 0.05096796, Cur Avg Loss: 0.15076792, Log Avg loss: 0.14346196, Global Avg Loss: 0.64934914, Time: 0.0209 Steps: 82480, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001588, Sample Num: 25408, Cur Loss: 0.09581594, Cur Avg Loss: 0.15068368, Log Avg loss: 0.13739153, Global Avg Loss: 0.64928707, Time: 0.0209 Steps: 82490, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001598, Sample Num: 25568, Cur Loss: 0.46525985, Cur Avg Loss: 0.15118170, Log Avg loss: 0.23026592, Global Avg Loss: 0.64923628, Time: 0.0209 Steps: 82500, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001608, Sample Num: 25728, Cur Loss: 0.08021864, Cur Avg Loss: 0.15096845, Log Avg loss: 0.11689208, Global Avg Loss: 0.64917177, Time: 0.0209 Steps: 82510, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001618, Sample Num: 25888, Cur Loss: 0.09491305, Cur Avg Loss: 0.15111609, Log Avg loss: 0.17485606, Global Avg Loss: 0.64911429, Time: 0.0209 Steps: 82520, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001628, Sample Num: 26048, Cur Loss: 0.08893217, Cur Avg Loss: 0.15085576, Log Avg loss: 0.10873463, Global Avg Loss: 0.64904881, Time: 0.0209 Steps: 82530, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001638, Sample Num: 26208, Cur Loss: 0.17454050, Cur Avg Loss: 0.15075631, Log Avg loss: 0.13456641, Global Avg Loss: 0.64898648, Time: 0.0209 Steps: 82540, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001648, Sample Num: 26368, Cur Loss: 0.04371292, Cur Avg Loss: 0.15053569, Log Avg loss: 0.11439824, Global Avg Loss: 0.64892172, Time: 0.0208 Steps: 82550, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001658, Sample Num: 26528, Cur Loss: 0.35720068, Cur Avg Loss: 0.15050344, Log Avg loss: 0.14518742, Global Avg Loss: 0.64886070, Time: 0.0209 Steps: 82560, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001668, Sample Num: 26688, Cur Loss: 0.33465046, Cur Avg Loss: 0.15079796, Log Avg loss: 0.19963029, Global Avg Loss: 0.64880630, Time: 0.0209 Steps: 82570, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001678, Sample Num: 26848, Cur Loss: 0.03798014, Cur Avg Loss: 0.15051467, Log Avg loss: 0.10326209, Global Avg Loss: 0.64874024, Time: 0.0209 Steps: 82580, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001688, Sample Num: 27008, Cur Loss: 0.08337779, Cur Avg Loss: 0.15057325, Log Avg loss: 0.16040178, Global Avg Loss: 0.64868111, Time: 0.0209 Steps: 82590, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001698, Sample Num: 27168, Cur Loss: 0.15026203, Cur Avg Loss: 0.15052755, Log Avg loss: 0.14281390, Global Avg Loss: 0.64861987, Time: 0.0209 Steps: 82600, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001708, Sample Num: 27328, Cur Loss: 0.09162819, Cur Avg Loss: 0.15027024, Log Avg loss: 0.10657906, Global Avg Loss: 0.64855425, Time: 0.0209 Steps: 82610, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001718, Sample Num: 27488, Cur Loss: 0.28093436, Cur Avg Loss: 0.15028339, Log Avg loss: 0.15252998, Global Avg Loss: 0.64849421, Time: 0.0209 Steps: 82620, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001728, Sample Num: 27648, Cur Loss: 0.10588125, Cur Avg Loss: 0.15017698, Log Avg loss: 0.13189489, Global Avg Loss: 0.64843169, Time: 0.0209 Steps: 82630, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001738, Sample Num: 27808, Cur Loss: 0.23119390, Cur Avg Loss: 0.15033360, Log Avg loss: 0.17739718, Global Avg Loss: 0.64837470, Time: 0.0209 Steps: 82640, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001748, Sample Num: 27968, Cur Loss: 0.06149813, Cur Avg Loss: 0.15037845, Log Avg loss: 0.15817327, Global Avg Loss: 0.64831539, Time: 0.0209 Steps: 82650, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001758, Sample Num: 28128, Cur Loss: 0.13928467, Cur Avg Loss: 0.15053735, Log Avg loss: 0.17831404, Global Avg Loss: 0.64825853, Time: 0.0209 Steps: 82660, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001768, Sample Num: 28288, Cur Loss: 0.18719745, Cur Avg Loss: 0.15082245, Log Avg loss: 0.20094328, Global Avg Loss: 0.64820442, Time: 0.0209 Steps: 82670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001778, Sample Num: 28448, Cur Loss: 0.09776143, Cur Avg Loss: 0.15124664, Log Avg loss: 0.22624283, Global Avg Loss: 0.64815338, Time: 0.0209 Steps: 82680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001788, Sample Num: 28608, Cur Loss: 0.08873285, Cur Avg Loss: 0.15109376, Log Avg loss: 0.12391212, Global Avg Loss: 0.64808998, Time: 0.0209 Steps: 82690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001798, Sample Num: 28768, Cur Loss: 0.05563104, Cur Avg Loss: 0.15084959, Log Avg loss: 0.10719118, Global Avg Loss: 0.64802458, Time: 0.0210 Steps: 82700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001808, Sample Num: 28928, Cur Loss: 0.10236497, Cur Avg Loss: 0.15066437, Log Avg loss: 0.11736180, Global Avg Loss: 0.64796042, Time: 0.0210 Steps: 82710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001818, Sample Num: 29088, Cur Loss: 0.07477719, Cur Avg Loss: 0.15079827, Log Avg loss: 0.17500853, Global Avg Loss: 0.64790324, Time: 0.0209 Steps: 82720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001828, Sample Num: 29248, Cur Loss: 0.52529681, Cur Avg Loss: 0.15121895, Log Avg loss: 0.22769819, Global Avg Loss: 0.64785245, Time: 0.0209 Steps: 82730, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001838, Sample Num: 29408, Cur Loss: 0.10298862, Cur Avg Loss: 0.15113196, Log Avg loss: 0.13522942, Global Avg Loss: 0.64779050, Time: 0.0209 Steps: 82740, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001848, Sample Num: 29568, Cur Loss: 0.04283110, Cur Avg Loss: 0.15118683, Log Avg loss: 0.16127304, Global Avg Loss: 0.64773170, Time: 0.0210 Steps: 82750, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001858, Sample Num: 29728, Cur Loss: 0.02755983, Cur Avg Loss: 0.15113758, Log Avg loss: 0.14203504, Global Avg Loss: 0.64767060, Time: 0.0209 Steps: 82760, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001868, Sample Num: 29888, Cur Loss: 0.24303487, Cur Avg Loss: 0.15120839, Log Avg loss: 0.16436457, Global Avg Loss: 0.64761221, Time: 0.0209 Steps: 82770, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001878, Sample Num: 30048, Cur Loss: 0.04704816, Cur Avg Loss: 0.15105185, Log Avg loss: 0.12181115, Global Avg Loss: 0.64754869, Time: 0.0209 Steps: 82780, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001888, Sample Num: 30208, Cur Loss: 0.07560991, Cur Avg Loss: 0.15097292, Log Avg loss: 0.13615044, Global Avg Loss: 0.64748692, Time: 0.0208 Steps: 82790, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001898, Sample Num: 30368, Cur Loss: 0.25644678, Cur Avg Loss: 0.15120712, Log Avg loss: 0.19542301, Global Avg Loss: 0.64743232, Time: 0.0208 Steps: 82800, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001908, Sample Num: 30528, Cur Loss: 0.25734288, Cur Avg Loss: 0.15146437, Log Avg loss: 0.20029069, Global Avg Loss: 0.64737832, Time: 0.0209 Steps: 82810, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001918, Sample Num: 30688, Cur Loss: 0.22254315, Cur Avg Loss: 0.15169571, Log Avg loss: 0.19583513, Global Avg Loss: 0.64732380, Time: 0.0209 Steps: 82820, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001928, Sample Num: 30848, Cur Loss: 0.13744928, Cur Avg Loss: 0.15169878, Log Avg loss: 0.15228711, Global Avg Loss: 0.64726404, Time: 0.0209 Steps: 82830, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001938, Sample Num: 31008, Cur Loss: 0.16618398, Cur Avg Loss: 0.15190691, Log Avg loss: 0.19203473, Global Avg Loss: 0.64720909, Time: 0.0209 Steps: 82840, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001948, Sample Num: 31168, Cur Loss: 0.13360456, Cur Avg Loss: 0.15216031, Log Avg loss: 0.20127024, Global Avg Loss: 0.64715526, Time: 0.0208 Steps: 82850, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001958, Sample Num: 31328, Cur Loss: 0.07382478, Cur Avg Loss: 0.15254069, Log Avg loss: 0.22663913, Global Avg Loss: 0.64710451, Time: 0.0209 Steps: 82860, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001968, Sample Num: 31488, Cur Loss: 0.10480323, Cur Avg Loss: 0.15236177, Log Avg loss: 0.11732761, Global Avg Loss: 0.64704058, Time: 0.0211 Steps: 82870, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001978, Sample Num: 31648, Cur Loss: 0.20785254, Cur Avg Loss: 0.15260162, Log Avg loss: 0.19980519, Global Avg Loss: 0.64698662, Time: 0.0209 Steps: 82880, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001988, Sample Num: 31808, Cur Loss: 0.21240315, Cur Avg Loss: 0.15228562, Log Avg loss: 0.08978075, Global Avg Loss: 0.64691940, Time: 0.0208 Steps: 82890, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001998, Sample Num: 31968, Cur Loss: 0.10350595, Cur Avg Loss: 0.15244667, Log Avg loss: 0.18446261, Global Avg Loss: 0.64686361, Time: 0.0209 Steps: 82900, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002008, Sample Num: 32128, Cur Loss: 0.04524805, Cur Avg Loss: 0.15242274, Log Avg loss: 0.14764121, Global Avg Loss: 0.64680340, Time: 0.0209 Steps: 82910, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002018, Sample Num: 32288, Cur Loss: 0.17165518, Cur Avg Loss: 0.15223890, Log Avg loss: 0.11532487, Global Avg Loss: 0.64673931, Time: 0.0208 Steps: 82920, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002028, Sample Num: 32448, Cur Loss: 0.14690077, Cur Avg Loss: 0.15240933, Log Avg loss: 0.18680172, Global Avg Loss: 0.64668384, Time: 0.0208 Steps: 82930, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002038, Sample Num: 32608, Cur Loss: 0.13351822, Cur Avg Loss: 0.15248754, Log Avg loss: 0.16834867, Global Avg Loss: 0.64662617, Time: 0.0209 Steps: 82940, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002048, Sample Num: 32768, Cur Loss: 0.19726633, Cur Avg Loss: 0.15268905, Log Avg loss: 0.19375774, Global Avg Loss: 0.64657158, Time: 0.0254 Steps: 82950, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002058, Sample Num: 32928, Cur Loss: 0.12785444, Cur Avg Loss: 0.15260311, Log Avg loss: 0.13500134, Global Avg Loss: 0.64650991, Time: 0.0208 Steps: 82960, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002068, Sample Num: 33088, Cur Loss: 0.33638400, Cur Avg Loss: 0.15259079, Log Avg loss: 0.15005537, Global Avg Loss: 0.64645008, Time: 0.0208 Steps: 82970, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002078, Sample Num: 33248, Cur Loss: 0.07077758, Cur Avg Loss: 0.15255587, Log Avg loss: 0.14533488, Global Avg Loss: 0.64638969, Time: 0.0208 Steps: 82980, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002088, Sample Num: 33408, Cur Loss: 0.15714735, Cur Avg Loss: 0.15257730, Log Avg loss: 0.15703091, Global Avg Loss: 0.64633072, Time: 0.0208 Steps: 82990, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002098, Sample Num: 33568, Cur Loss: 0.36176121, Cur Avg Loss: 0.15279004, Log Avg loss: 0.19720885, Global Avg Loss: 0.64627661, Time: 0.0209 Steps: 83000, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002108, Sample Num: 33728, Cur Loss: 0.09495270, Cur Avg Loss: 0.15268697, Log Avg loss: 0.13106399, Global Avg Loss: 0.64621454, Time: 0.0209 Steps: 83010, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002118, Sample Num: 33888, Cur Loss: 0.11655318, Cur Avg Loss: 0.15295891, Log Avg loss: 0.21028289, Global Avg Loss: 0.64616203, Time: 0.0209 Steps: 83020, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002128, Sample Num: 34048, Cur Loss: 0.09212695, Cur Avg Loss: 0.15253938, Log Avg loss: 0.06368288, Global Avg Loss: 0.64609188, Time: 0.0209 Steps: 83030, Updated lr: 0.000022 ***** Running evaluation checkpoint-83031 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-83031 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.694566, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.208639, "eval_total_loss": 146.673276, "eval_mae": 0.299284, "eval_mse": 0.208713, "eval_r2": 0.867328, "eval_sp_statistic": 0.928614, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.935336, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.256507, "test_total_loss": 128.766453, "test_mae": 0.41418, "test_mse": 0.25658, "test_r2": 0.834401, "test_sp_statistic": 0.920014, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946481, "test_ps_pvalue": 0.0, "lr": 2.2208629682313894e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6460849251014746, "train_cur_epoch_loss": 324.6723446948454, "train_cur_epoch_avg_loss": 0.15249992705253423, "train_cur_epoch_time": 44.694565534591675, "train_cur_epoch_avg_time": 0.02099322007261234, "epoch": 39, "step": 83031} ################################################## Training, Epoch: 0040, Batch: 000009, Sample Num: 144, Cur Loss: 0.17753553, Cur Avg Loss: 0.18818277, Log Avg loss: 0.17621946, Global Avg Loss: 0.64603530, Time: 0.0210 Steps: 83040, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000019, Sample Num: 304, Cur Loss: 0.22175834, Cur Avg Loss: 0.18502985, Log Avg loss: 0.18219222, Global Avg Loss: 0.64597945, Time: 0.0210 Steps: 83050, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000029, Sample Num: 464, Cur Loss: 0.06238884, Cur Avg Loss: 0.16499521, Log Avg loss: 0.12692939, Global Avg Loss: 0.64591695, Time: 0.0210 Steps: 83060, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000039, Sample Num: 624, Cur Loss: 0.10108206, Cur Avg Loss: 0.14513452, Log Avg loss: 0.08753852, Global Avg Loss: 0.64584974, Time: 0.0210 Steps: 83070, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000049, Sample Num: 784, Cur Loss: 0.06850395, Cur Avg Loss: 0.13118436, Log Avg loss: 0.07677875, Global Avg Loss: 0.64578124, Time: 0.0210 Steps: 83080, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000059, Sample Num: 944, Cur Loss: 0.14114206, Cur Avg Loss: 0.14309319, Log Avg loss: 0.20144644, Global Avg Loss: 0.64572776, Time: 0.0210 Steps: 83090, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000069, Sample Num: 1104, Cur Loss: 0.24168062, Cur Avg Loss: 0.14581993, Log Avg loss: 0.16190771, Global Avg Loss: 0.64566954, Time: 0.0210 Steps: 83100, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000079, Sample Num: 1264, Cur Loss: 0.06296119, Cur Avg Loss: 0.14548132, Log Avg loss: 0.14314489, Global Avg Loss: 0.64560908, Time: 0.0210 Steps: 83110, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000089, Sample Num: 1424, Cur Loss: 0.19035396, Cur Avg Loss: 0.14598003, Log Avg loss: 0.14991981, Global Avg Loss: 0.64554944, Time: 0.0210 Steps: 83120, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000099, Sample Num: 1584, Cur Loss: 0.05433400, Cur Avg Loss: 0.14524758, Log Avg loss: 0.13872882, Global Avg Loss: 0.64548847, Time: 0.0210 Steps: 83130, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000109, Sample Num: 1744, Cur Loss: 0.12519327, Cur Avg Loss: 0.14464001, Log Avg loss: 0.13862508, Global Avg Loss: 0.64542751, Time: 0.0209 Steps: 83140, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000119, Sample Num: 1904, Cur Loss: 0.17302501, Cur Avg Loss: 0.15424715, Log Avg loss: 0.25896495, Global Avg Loss: 0.64538103, Time: 0.0210 Steps: 83150, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000129, Sample Num: 2064, Cur Loss: 0.11423022, Cur Avg Loss: 0.15979266, Log Avg loss: 0.22578419, Global Avg Loss: 0.64533058, Time: 0.0210 Steps: 83160, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000139, Sample Num: 2224, Cur Loss: 0.11016840, Cur Avg Loss: 0.15806246, Log Avg loss: 0.13574291, Global Avg Loss: 0.64526931, Time: 0.0210 Steps: 83170, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000149, Sample Num: 2384, Cur Loss: 0.16701992, Cur Avg Loss: 0.15616178, Log Avg loss: 0.12974239, Global Avg Loss: 0.64520733, Time: 0.0210 Steps: 83180, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000159, Sample Num: 2544, Cur Loss: 0.07677204, Cur Avg Loss: 0.15487822, Log Avg loss: 0.13575313, Global Avg Loss: 0.64514609, Time: 0.0210 Steps: 83190, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000169, Sample Num: 2704, Cur Loss: 0.11246057, Cur Avg Loss: 0.15135870, Log Avg loss: 0.09539838, Global Avg Loss: 0.64508001, Time: 0.0209 Steps: 83200, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000179, Sample Num: 2864, Cur Loss: 0.09639838, Cur Avg Loss: 0.15200592, Log Avg loss: 0.16294388, Global Avg Loss: 0.64502207, Time: 0.0210 Steps: 83210, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000189, Sample Num: 3024, Cur Loss: 0.19668822, Cur Avg Loss: 0.14990968, Log Avg loss: 0.11238708, Global Avg Loss: 0.64495807, Time: 0.0210 Steps: 83220, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000199, Sample Num: 3184, Cur Loss: 0.16718224, Cur Avg Loss: 0.15161109, Log Avg loss: 0.18376761, Global Avg Loss: 0.64490266, Time: 0.0210 Steps: 83230, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000209, Sample Num: 3344, Cur Loss: 0.09044208, Cur Avg Loss: 0.15123863, Log Avg loss: 0.14382672, Global Avg Loss: 0.64484246, Time: 0.0210 Steps: 83240, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000219, Sample Num: 3504, Cur Loss: 0.22475109, Cur Avg Loss: 0.15244137, Log Avg loss: 0.17757861, Global Avg Loss: 0.64478633, Time: 0.0210 Steps: 83250, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000229, Sample Num: 3664, Cur Loss: 0.10651164, Cur Avg Loss: 0.15212428, Log Avg loss: 0.14518003, Global Avg Loss: 0.64472633, Time: 0.0210 Steps: 83260, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000239, Sample Num: 3824, Cur Loss: 0.07213183, Cur Avg Loss: 0.14992517, Log Avg loss: 0.09956552, Global Avg Loss: 0.64466086, Time: 0.0210 Steps: 83270, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000249, Sample Num: 3984, Cur Loss: 0.03944220, Cur Avg Loss: 0.14916180, Log Avg loss: 0.13091720, Global Avg Loss: 0.64459917, Time: 0.0209 Steps: 83280, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000259, Sample Num: 4144, Cur Loss: 0.02407342, Cur Avg Loss: 0.14780475, Log Avg loss: 0.11401417, Global Avg Loss: 0.64453546, Time: 0.0247 Steps: 83290, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000269, Sample Num: 4304, Cur Loss: 0.08304418, Cur Avg Loss: 0.14771757, Log Avg loss: 0.14545972, Global Avg Loss: 0.64447555, Time: 0.0210 Steps: 83300, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000279, Sample Num: 4464, Cur Loss: 0.07785219, Cur Avg Loss: 0.14701308, Log Avg loss: 0.12806239, Global Avg Loss: 0.64441356, Time: 0.0210 Steps: 83310, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000289, Sample Num: 4624, Cur Loss: 0.03841690, Cur Avg Loss: 0.14643923, Log Avg loss: 0.13042865, Global Avg Loss: 0.64435188, Time: 0.0210 Steps: 83320, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000299, Sample Num: 4784, Cur Loss: 0.18236408, Cur Avg Loss: 0.14687486, Log Avg loss: 0.15946451, Global Avg Loss: 0.64429369, Time: 0.0210 Steps: 83330, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000309, Sample Num: 4944, Cur Loss: 0.24127579, Cur Avg Loss: 0.14702343, Log Avg loss: 0.15146591, Global Avg Loss: 0.64423455, Time: 0.0210 Steps: 83340, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000319, Sample Num: 5104, Cur Loss: 0.02746924, Cur Avg Loss: 0.14813699, Log Avg loss: 0.18254589, Global Avg Loss: 0.64417916, Time: 0.0210 Steps: 83350, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000329, Sample Num: 5264, Cur Loss: 0.10618010, Cur Avg Loss: 0.14767026, Log Avg loss: 0.13278151, Global Avg Loss: 0.64411781, Time: 0.0210 Steps: 83360, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000339, Sample Num: 5424, Cur Loss: 0.19401221, Cur Avg Loss: 0.14895293, Log Avg loss: 0.19115286, Global Avg Loss: 0.64406348, Time: 0.0210 Steps: 83370, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000349, Sample Num: 5584, Cur Loss: 0.17492414, Cur Avg Loss: 0.14878921, Log Avg loss: 0.14323906, Global Avg Loss: 0.64400342, Time: 0.0210 Steps: 83380, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000359, Sample Num: 5744, Cur Loss: 0.08717253, Cur Avg Loss: 0.14785366, Log Avg loss: 0.11520296, Global Avg Loss: 0.64394000, Time: 0.0210 Steps: 83390, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000369, Sample Num: 5904, Cur Loss: 0.20040695, Cur Avg Loss: 0.14705596, Log Avg loss: 0.11841838, Global Avg Loss: 0.64387699, Time: 0.0210 Steps: 83400, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000379, Sample Num: 6064, Cur Loss: 0.15774480, Cur Avg Loss: 0.14650113, Log Avg loss: 0.12602817, Global Avg Loss: 0.64381491, Time: 0.0210 Steps: 83410, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000389, Sample Num: 6224, Cur Loss: 0.02978626, Cur Avg Loss: 0.14633780, Log Avg loss: 0.14014757, Global Avg Loss: 0.64375453, Time: 0.0210 Steps: 83420, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000399, Sample Num: 6384, Cur Loss: 0.16927642, Cur Avg Loss: 0.14623263, Log Avg loss: 0.14214163, Global Avg Loss: 0.64369441, Time: 0.0210 Steps: 83430, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000409, Sample Num: 6544, Cur Loss: 0.07869524, Cur Avg Loss: 0.14512586, Log Avg loss: 0.10096569, Global Avg Loss: 0.64362936, Time: 0.0210 Steps: 83440, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000419, Sample Num: 6704, Cur Loss: 0.08158291, Cur Avg Loss: 0.14424276, Log Avg loss: 0.10812376, Global Avg Loss: 0.64356519, Time: 0.0210 Steps: 83450, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000429, Sample Num: 6864, Cur Loss: 0.13028623, Cur Avg Loss: 0.14370450, Log Avg loss: 0.12115151, Global Avg Loss: 0.64350260, Time: 0.0210 Steps: 83460, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000439, Sample Num: 7024, Cur Loss: 0.30394155, Cur Avg Loss: 0.14365839, Log Avg loss: 0.14168006, Global Avg Loss: 0.64344248, Time: 0.0210 Steps: 83470, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000449, Sample Num: 7184, Cur Loss: 0.20432618, Cur Avg Loss: 0.14327513, Log Avg loss: 0.12645032, Global Avg Loss: 0.64338055, Time: 0.0210 Steps: 83480, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000459, Sample Num: 7344, Cur Loss: 0.57999337, Cur Avg Loss: 0.14421444, Log Avg loss: 0.18638920, Global Avg Loss: 0.64332581, Time: 0.0210 Steps: 83490, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000469, Sample Num: 7504, Cur Loss: 0.09031174, Cur Avg Loss: 0.14476523, Log Avg loss: 0.17004664, Global Avg Loss: 0.64326913, Time: 0.0210 Steps: 83500, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000479, Sample Num: 7664, Cur Loss: 0.33486259, Cur Avg Loss: 0.14664100, Log Avg loss: 0.23461475, Global Avg Loss: 0.64322019, Time: 0.0210 Steps: 83510, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000489, Sample Num: 7824, Cur Loss: 0.17613401, Cur Avg Loss: 0.14708817, Log Avg loss: 0.16850767, Global Avg Loss: 0.64316336, Time: 0.0210 Steps: 83520, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000499, Sample Num: 7984, Cur Loss: 0.39623946, Cur Avg Loss: 0.14769211, Log Avg loss: 0.17722433, Global Avg Loss: 0.64310758, Time: 0.0210 Steps: 83530, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000509, Sample Num: 8144, Cur Loss: 0.07221660, Cur Avg Loss: 0.14761303, Log Avg loss: 0.14366721, Global Avg Loss: 0.64304779, Time: 0.0210 Steps: 83540, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000519, Sample Num: 8304, Cur Loss: 0.10941183, Cur Avg Loss: 0.14764219, Log Avg loss: 0.14912636, Global Avg Loss: 0.64298867, Time: 0.0211 Steps: 83550, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000529, Sample Num: 8464, Cur Loss: 0.12767181, Cur Avg Loss: 0.14890545, Log Avg loss: 0.21446860, Global Avg Loss: 0.64293739, Time: 0.0209 Steps: 83560, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000539, Sample Num: 8624, Cur Loss: 0.04019422, Cur Avg Loss: 0.14883194, Log Avg loss: 0.14494320, Global Avg Loss: 0.64287780, Time: 0.0210 Steps: 83570, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000549, Sample Num: 8784, Cur Loss: 0.04330745, Cur Avg Loss: 0.14955657, Log Avg loss: 0.18861415, Global Avg Loss: 0.64282345, Time: 0.0210 Steps: 83580, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000559, Sample Num: 8944, Cur Loss: 0.04120977, Cur Avg Loss: 0.14946658, Log Avg loss: 0.14452637, Global Avg Loss: 0.64276384, Time: 0.0210 Steps: 83590, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000569, Sample Num: 9104, Cur Loss: 0.14346212, Cur Avg Loss: 0.14952573, Log Avg loss: 0.15283187, Global Avg Loss: 0.64270523, Time: 0.0210 Steps: 83600, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000579, Sample Num: 9264, Cur Loss: 0.21822079, Cur Avg Loss: 0.14944322, Log Avg loss: 0.14474871, Global Avg Loss: 0.64264568, Time: 0.0210 Steps: 83610, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000589, Sample Num: 9424, Cur Loss: 0.08656365, Cur Avg Loss: 0.14912759, Log Avg loss: 0.13085234, Global Avg Loss: 0.64258447, Time: 0.0209 Steps: 83620, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000599, Sample Num: 9584, Cur Loss: 0.37505513, Cur Avg Loss: 0.14962282, Log Avg loss: 0.17879191, Global Avg Loss: 0.64252901, Time: 0.0210 Steps: 83630, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000609, Sample Num: 9744, Cur Loss: 0.11645182, Cur Avg Loss: 0.14899274, Log Avg loss: 0.11125108, Global Avg Loss: 0.64246549, Time: 0.0210 Steps: 83640, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000619, Sample Num: 9904, Cur Loss: 0.17453034, Cur Avg Loss: 0.14859425, Log Avg loss: 0.12432644, Global Avg Loss: 0.64240355, Time: 0.0209 Steps: 83650, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000629, Sample Num: 10064, Cur Loss: 0.06726100, Cur Avg Loss: 0.14853334, Log Avg loss: 0.14476313, Global Avg Loss: 0.64234407, Time: 0.0210 Steps: 83660, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000639, Sample Num: 10224, Cur Loss: 0.02618347, Cur Avg Loss: 0.14818855, Log Avg loss: 0.12650070, Global Avg Loss: 0.64228242, Time: 0.0210 Steps: 83670, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000649, Sample Num: 10384, Cur Loss: 0.04982067, Cur Avg Loss: 0.14807902, Log Avg loss: 0.14108061, Global Avg Loss: 0.64222252, Time: 0.0210 Steps: 83680, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000659, Sample Num: 10544, Cur Loss: 0.09834822, Cur Avg Loss: 0.14870832, Log Avg loss: 0.18954943, Global Avg Loss: 0.64216843, Time: 0.0210 Steps: 83690, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000669, Sample Num: 10704, Cur Loss: 0.11418885, Cur Avg Loss: 0.14907068, Log Avg loss: 0.17295046, Global Avg Loss: 0.64211237, Time: 0.0210 Steps: 83700, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000679, Sample Num: 10864, Cur Loss: 0.10347391, Cur Avg Loss: 0.14818384, Log Avg loss: 0.08885401, Global Avg Loss: 0.64204628, Time: 0.0210 Steps: 83710, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000689, Sample Num: 11024, Cur Loss: 0.06703441, Cur Avg Loss: 0.14789566, Log Avg loss: 0.12832822, Global Avg Loss: 0.64198492, Time: 0.0210 Steps: 83720, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000699, Sample Num: 11184, Cur Loss: 0.08383532, Cur Avg Loss: 0.14716797, Log Avg loss: 0.09703061, Global Avg Loss: 0.64191984, Time: 0.0210 Steps: 83730, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000709, Sample Num: 11344, Cur Loss: 0.05500592, Cur Avg Loss: 0.14668933, Log Avg loss: 0.11323195, Global Avg Loss: 0.64185670, Time: 0.0210 Steps: 83740, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000719, Sample Num: 11504, Cur Loss: 0.21846218, Cur Avg Loss: 0.14714125, Log Avg loss: 0.17918230, Global Avg Loss: 0.64180146, Time: 0.0210 Steps: 83750, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000729, Sample Num: 11664, Cur Loss: 0.08269757, Cur Avg Loss: 0.14670410, Log Avg loss: 0.11527338, Global Avg Loss: 0.64173859, Time: 0.0210 Steps: 83760, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000739, Sample Num: 11824, Cur Loss: 0.17501685, Cur Avg Loss: 0.14711692, Log Avg loss: 0.17721124, Global Avg Loss: 0.64168314, Time: 0.0211 Steps: 83770, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000749, Sample Num: 11984, Cur Loss: 0.09814588, Cur Avg Loss: 0.14733702, Log Avg loss: 0.16360242, Global Avg Loss: 0.64162608, Time: 0.0209 Steps: 83780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000759, Sample Num: 12144, Cur Loss: 0.11716059, Cur Avg Loss: 0.14734717, Log Avg loss: 0.14810782, Global Avg Loss: 0.64156718, Time: 0.0210 Steps: 83790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000769, Sample Num: 12304, Cur Loss: 0.16485475, Cur Avg Loss: 0.14727345, Log Avg loss: 0.14167805, Global Avg Loss: 0.64150753, Time: 0.0247 Steps: 83800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000779, Sample Num: 12464, Cur Loss: 0.31504205, Cur Avg Loss: 0.14721970, Log Avg loss: 0.14308618, Global Avg Loss: 0.64144806, Time: 0.0210 Steps: 83810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000789, Sample Num: 12624, Cur Loss: 0.07481409, Cur Avg Loss: 0.14771506, Log Avg loss: 0.18630377, Global Avg Loss: 0.64139376, Time: 0.0210 Steps: 83820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000799, Sample Num: 12784, Cur Loss: 0.24162166, Cur Avg Loss: 0.14843003, Log Avg loss: 0.20484081, Global Avg Loss: 0.64134168, Time: 0.0210 Steps: 83830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000809, Sample Num: 12944, Cur Loss: 0.07247004, Cur Avg Loss: 0.14838155, Log Avg loss: 0.14450779, Global Avg Loss: 0.64128242, Time: 0.0210 Steps: 83840, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000819, Sample Num: 13104, Cur Loss: 0.32937485, Cur Avg Loss: 0.14912974, Log Avg loss: 0.20965839, Global Avg Loss: 0.64123094, Time: 0.0210 Steps: 83850, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000829, Sample Num: 13264, Cur Loss: 0.07050076, Cur Avg Loss: 0.14901621, Log Avg loss: 0.13971829, Global Avg Loss: 0.64117114, Time: 0.0210 Steps: 83860, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000839, Sample Num: 13424, Cur Loss: 0.21142326, Cur Avg Loss: 0.14955956, Log Avg loss: 0.19460333, Global Avg Loss: 0.64111790, Time: 0.0210 Steps: 83870, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000849, Sample Num: 13584, Cur Loss: 0.04839896, Cur Avg Loss: 0.14957493, Log Avg loss: 0.15086430, Global Avg Loss: 0.64105945, Time: 0.0210 Steps: 83880, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000859, Sample Num: 13744, Cur Loss: 0.08612577, Cur Avg Loss: 0.14936803, Log Avg loss: 0.13180198, Global Avg Loss: 0.64099874, Time: 0.0209 Steps: 83890, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000869, Sample Num: 13904, Cur Loss: 0.16839904, Cur Avg Loss: 0.14918046, Log Avg loss: 0.13306852, Global Avg Loss: 0.64093820, Time: 0.0210 Steps: 83900, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000879, Sample Num: 14064, Cur Loss: 0.15194976, Cur Avg Loss: 0.14954617, Log Avg loss: 0.18132637, Global Avg Loss: 0.64088343, Time: 0.0211 Steps: 83910, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000889, Sample Num: 14224, Cur Loss: 0.15072447, Cur Avg Loss: 0.14949503, Log Avg loss: 0.14499975, Global Avg Loss: 0.64082434, Time: 0.0209 Steps: 83920, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000899, Sample Num: 14384, Cur Loss: 0.34468412, Cur Avg Loss: 0.14935790, Log Avg loss: 0.13716728, Global Avg Loss: 0.64076433, Time: 0.0210 Steps: 83930, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000909, Sample Num: 14544, Cur Loss: 0.23212804, Cur Avg Loss: 0.14910807, Log Avg loss: 0.12664810, Global Avg Loss: 0.64070308, Time: 0.0209 Steps: 83940, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000919, Sample Num: 14704, Cur Loss: 0.06858495, Cur Avg Loss: 0.14895936, Log Avg loss: 0.13544147, Global Avg Loss: 0.64064290, Time: 0.0209 Steps: 83950, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000929, Sample Num: 14864, Cur Loss: 0.37696955, Cur Avg Loss: 0.14956487, Log Avg loss: 0.20521134, Global Avg Loss: 0.64059103, Time: 0.0209 Steps: 83960, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000939, Sample Num: 15024, Cur Loss: 0.10903159, Cur Avg Loss: 0.14916224, Log Avg loss: 0.11175848, Global Avg Loss: 0.64052805, Time: 0.0210 Steps: 83970, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000949, Sample Num: 15184, Cur Loss: 0.22514036, Cur Avg Loss: 0.14914469, Log Avg loss: 0.14749630, Global Avg Loss: 0.64046935, Time: 0.0210 Steps: 83980, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000959, Sample Num: 15344, Cur Loss: 0.01605045, Cur Avg Loss: 0.14903869, Log Avg loss: 0.13897954, Global Avg Loss: 0.64040964, Time: 0.0210 Steps: 83990, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000969, Sample Num: 15504, Cur Loss: 0.08012135, Cur Avg Loss: 0.14864485, Log Avg loss: 0.11087574, Global Avg Loss: 0.64034660, Time: 0.0210 Steps: 84000, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000979, Sample Num: 15664, Cur Loss: 0.04158913, Cur Avg Loss: 0.14834159, Log Avg loss: 0.11895517, Global Avg Loss: 0.64028454, Time: 0.0209 Steps: 84010, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000989, Sample Num: 15824, Cur Loss: 0.36295512, Cur Avg Loss: 0.14844614, Log Avg loss: 0.15868155, Global Avg Loss: 0.64022722, Time: 0.0210 Steps: 84020, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000999, Sample Num: 15984, Cur Loss: 0.11808440, Cur Avg Loss: 0.14800638, Log Avg loss: 0.10451474, Global Avg Loss: 0.64016346, Time: 0.0209 Steps: 84030, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001009, Sample Num: 16144, Cur Loss: 0.05432818, Cur Avg Loss: 0.14788348, Log Avg loss: 0.13560490, Global Avg Loss: 0.64010343, Time: 0.0210 Steps: 84040, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001019, Sample Num: 16304, Cur Loss: 0.15809344, Cur Avg Loss: 0.14792934, Log Avg loss: 0.15255662, Global Avg Loss: 0.64004542, Time: 0.0209 Steps: 84050, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001029, Sample Num: 16464, Cur Loss: 0.13800824, Cur Avg Loss: 0.14818179, Log Avg loss: 0.17390697, Global Avg Loss: 0.63998997, Time: 0.0246 Steps: 84060, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001039, Sample Num: 16624, Cur Loss: 0.21241277, Cur Avg Loss: 0.14813302, Log Avg loss: 0.14311460, Global Avg Loss: 0.63993086, Time: 0.0209 Steps: 84070, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001049, Sample Num: 16784, Cur Loss: 0.05580805, Cur Avg Loss: 0.14770394, Log Avg loss: 0.10312279, Global Avg Loss: 0.63986702, Time: 0.0209 Steps: 84080, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001059, Sample Num: 16944, Cur Loss: 0.08507653, Cur Avg Loss: 0.14767520, Log Avg loss: 0.14466015, Global Avg Loss: 0.63980813, Time: 0.0209 Steps: 84090, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001069, Sample Num: 17104, Cur Loss: 0.25657639, Cur Avg Loss: 0.14732659, Log Avg loss: 0.11040888, Global Avg Loss: 0.63974518, Time: 0.0209 Steps: 84100, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001079, Sample Num: 17264, Cur Loss: 0.02552011, Cur Avg Loss: 0.14734766, Log Avg loss: 0.14960007, Global Avg Loss: 0.63968690, Time: 0.0209 Steps: 84110, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001089, Sample Num: 17424, Cur Loss: 0.07431952, Cur Avg Loss: 0.14744807, Log Avg loss: 0.15828191, Global Avg Loss: 0.63962968, Time: 0.0209 Steps: 84120, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001099, Sample Num: 17584, Cur Loss: 0.10997113, Cur Avg Loss: 0.14748454, Log Avg loss: 0.15145615, Global Avg Loss: 0.63957165, Time: 0.0209 Steps: 84130, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001109, Sample Num: 17744, Cur Loss: 0.18094559, Cur Avg Loss: 0.14741381, Log Avg loss: 0.13964064, Global Avg Loss: 0.63951223, Time: 0.0209 Steps: 84140, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001119, Sample Num: 17904, Cur Loss: 0.20900506, Cur Avg Loss: 0.14767507, Log Avg loss: 0.17664876, Global Avg Loss: 0.63945723, Time: 0.0208 Steps: 84150, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001129, Sample Num: 18064, Cur Loss: 0.16465424, Cur Avg Loss: 0.14783408, Log Avg loss: 0.16562699, Global Avg Loss: 0.63940093, Time: 0.0209 Steps: 84160, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001139, Sample Num: 18224, Cur Loss: 0.17558898, Cur Avg Loss: 0.14773044, Log Avg loss: 0.13603004, Global Avg Loss: 0.63934112, Time: 0.0209 Steps: 84170, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001149, Sample Num: 18384, Cur Loss: 0.02900465, Cur Avg Loss: 0.14756885, Log Avg loss: 0.12916399, Global Avg Loss: 0.63928052, Time: 0.0208 Steps: 84180, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001159, Sample Num: 18544, Cur Loss: 0.12390477, Cur Avg Loss: 0.14775112, Log Avg loss: 0.16869304, Global Avg Loss: 0.63922462, Time: 0.0209 Steps: 84190, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001169, Sample Num: 18704, Cur Loss: 0.17452380, Cur Avg Loss: 0.14797764, Log Avg loss: 0.17423137, Global Avg Loss: 0.63916940, Time: 0.0208 Steps: 84200, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001179, Sample Num: 18864, Cur Loss: 0.25845751, Cur Avg Loss: 0.14780933, Log Avg loss: 0.12813385, Global Avg Loss: 0.63910871, Time: 0.0209 Steps: 84210, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001189, Sample Num: 19024, Cur Loss: 0.18355788, Cur Avg Loss: 0.14847781, Log Avg loss: 0.22729184, Global Avg Loss: 0.63905981, Time: 0.0209 Steps: 84220, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001199, Sample Num: 19184, Cur Loss: 0.09364371, Cur Avg Loss: 0.14812762, Log Avg loss: 0.10649001, Global Avg Loss: 0.63899659, Time: 0.0209 Steps: 84230, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001209, Sample Num: 19344, Cur Loss: 0.02584403, Cur Avg Loss: 0.14776698, Log Avg loss: 0.10452604, Global Avg Loss: 0.63893314, Time: 0.0208 Steps: 84240, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001219, Sample Num: 19504, Cur Loss: 0.28241229, Cur Avg Loss: 0.14780776, Log Avg loss: 0.15273897, Global Avg Loss: 0.63887543, Time: 0.0209 Steps: 84250, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001229, Sample Num: 19664, Cur Loss: 0.14089665, Cur Avg Loss: 0.14813186, Log Avg loss: 0.18763940, Global Avg Loss: 0.63882188, Time: 0.0209 Steps: 84260, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001239, Sample Num: 19824, Cur Loss: 0.11826050, Cur Avg Loss: 0.14798563, Log Avg loss: 0.13001321, Global Avg Loss: 0.63876150, Time: 0.0208 Steps: 84270, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001249, Sample Num: 19984, Cur Loss: 0.02761327, Cur Avg Loss: 0.14799586, Log Avg loss: 0.14926386, Global Avg Loss: 0.63870342, Time: 0.0209 Steps: 84280, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001259, Sample Num: 20144, Cur Loss: 0.12459388, Cur Avg Loss: 0.14803856, Log Avg loss: 0.15337223, Global Avg Loss: 0.63864584, Time: 0.0208 Steps: 84290, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001269, Sample Num: 20304, Cur Loss: 0.10178618, Cur Avg Loss: 0.14813325, Log Avg loss: 0.16005396, Global Avg Loss: 0.63858907, Time: 0.0209 Steps: 84300, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001279, Sample Num: 20464, Cur Loss: 0.03878888, Cur Avg Loss: 0.14836200, Log Avg loss: 0.17739092, Global Avg Loss: 0.63853437, Time: 0.0209 Steps: 84310, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001289, Sample Num: 20624, Cur Loss: 0.07162409, Cur Avg Loss: 0.14833485, Log Avg loss: 0.14486191, Global Avg Loss: 0.63847582, Time: 0.0209 Steps: 84320, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001299, Sample Num: 20784, Cur Loss: 0.09966023, Cur Avg Loss: 0.14829002, Log Avg loss: 0.14251224, Global Avg Loss: 0.63841701, Time: 0.0209 Steps: 84330, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001309, Sample Num: 20944, Cur Loss: 0.22747764, Cur Avg Loss: 0.14842065, Log Avg loss: 0.16538866, Global Avg Loss: 0.63836092, Time: 0.0209 Steps: 84340, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001319, Sample Num: 21104, Cur Loss: 0.12958416, Cur Avg Loss: 0.14832812, Log Avg loss: 0.13621622, Global Avg Loss: 0.63830139, Time: 0.0209 Steps: 84350, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001329, Sample Num: 21264, Cur Loss: 0.31029347, Cur Avg Loss: 0.14829324, Log Avg loss: 0.14369282, Global Avg Loss: 0.63824276, Time: 0.0209 Steps: 84360, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001339, Sample Num: 21424, Cur Loss: 0.13255289, Cur Avg Loss: 0.14800034, Log Avg loss: 0.10907323, Global Avg Loss: 0.63818004, Time: 0.0209 Steps: 84370, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001349, Sample Num: 21584, Cur Loss: 0.05446244, Cur Avg Loss: 0.14767106, Log Avg loss: 0.10358115, Global Avg Loss: 0.63811668, Time: 0.0209 Steps: 84380, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001359, Sample Num: 21744, Cur Loss: 0.12603477, Cur Avg Loss: 0.14762800, Log Avg loss: 0.14181905, Global Avg Loss: 0.63805787, Time: 0.0209 Steps: 84390, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001369, Sample Num: 21904, Cur Loss: 0.13839391, Cur Avg Loss: 0.14755406, Log Avg loss: 0.13750590, Global Avg Loss: 0.63799857, Time: 0.0209 Steps: 84400, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001379, Sample Num: 22064, Cur Loss: 0.08711462, Cur Avg Loss: 0.14763016, Log Avg loss: 0.15804796, Global Avg Loss: 0.63794171, Time: 0.0209 Steps: 84410, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001389, Sample Num: 22224, Cur Loss: 0.40431279, Cur Avg Loss: 0.14806541, Log Avg loss: 0.20808635, Global Avg Loss: 0.63789079, Time: 0.0209 Steps: 84420, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001399, Sample Num: 22384, Cur Loss: 0.13413933, Cur Avg Loss: 0.14844910, Log Avg loss: 0.20174353, Global Avg Loss: 0.63783913, Time: 0.0209 Steps: 84430, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001409, Sample Num: 22544, Cur Loss: 0.05072068, Cur Avg Loss: 0.14821479, Log Avg loss: 0.11543523, Global Avg Loss: 0.63777726, Time: 0.0209 Steps: 84440, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001419, Sample Num: 22704, Cur Loss: 0.12635469, Cur Avg Loss: 0.14838379, Log Avg loss: 0.17219547, Global Avg Loss: 0.63772213, Time: 0.0209 Steps: 84450, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001429, Sample Num: 22864, Cur Loss: 0.04784363, Cur Avg Loss: 0.14840964, Log Avg loss: 0.15207846, Global Avg Loss: 0.63766463, Time: 0.0209 Steps: 84460, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001439, Sample Num: 23024, Cur Loss: 0.11089126, Cur Avg Loss: 0.14858190, Log Avg loss: 0.17319671, Global Avg Loss: 0.63760965, Time: 0.0209 Steps: 84470, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001449, Sample Num: 23184, Cur Loss: 0.06064283, Cur Avg Loss: 0.14820649, Log Avg loss: 0.09418609, Global Avg Loss: 0.63754532, Time: 0.0209 Steps: 84480, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001459, Sample Num: 23344, Cur Loss: 0.09569719, Cur Avg Loss: 0.14819489, Log Avg loss: 0.14651350, Global Avg Loss: 0.63748720, Time: 0.0209 Steps: 84490, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001469, Sample Num: 23504, Cur Loss: 0.40003553, Cur Avg Loss: 0.14849493, Log Avg loss: 0.19227024, Global Avg Loss: 0.63743451, Time: 0.0208 Steps: 84500, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001479, Sample Num: 23664, Cur Loss: 0.08692079, Cur Avg Loss: 0.14836744, Log Avg loss: 0.12963979, Global Avg Loss: 0.63737443, Time: 0.0209 Steps: 84510, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001489, Sample Num: 23824, Cur Loss: 0.33283928, Cur Avg Loss: 0.14844548, Log Avg loss: 0.15998709, Global Avg Loss: 0.63731795, Time: 0.0209 Steps: 84520, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001499, Sample Num: 23984, Cur Loss: 0.29283550, Cur Avg Loss: 0.14875191, Log Avg loss: 0.19437916, Global Avg Loss: 0.63726555, Time: 0.0209 Steps: 84530, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001509, Sample Num: 24144, Cur Loss: 0.04592042, Cur Avg Loss: 0.14876499, Log Avg loss: 0.15072581, Global Avg Loss: 0.63720799, Time: 0.0209 Steps: 84540, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001519, Sample Num: 24304, Cur Loss: 0.40675646, Cur Avg Loss: 0.14870356, Log Avg loss: 0.13943430, Global Avg Loss: 0.63714912, Time: 0.0210 Steps: 84550, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001529, Sample Num: 24464, Cur Loss: 0.08849286, Cur Avg Loss: 0.14878367, Log Avg loss: 0.16095274, Global Avg Loss: 0.63709281, Time: 0.0209 Steps: 84560, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001539, Sample Num: 24624, Cur Loss: 0.08687099, Cur Avg Loss: 0.14847250, Log Avg loss: 0.10089390, Global Avg Loss: 0.63702940, Time: 0.0247 Steps: 84570, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001549, Sample Num: 24784, Cur Loss: 0.14094663, Cur Avg Loss: 0.14844523, Log Avg loss: 0.14424914, Global Avg Loss: 0.63697114, Time: 0.0209 Steps: 84580, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001559, Sample Num: 24944, Cur Loss: 0.02887131, Cur Avg Loss: 0.14824973, Log Avg loss: 0.11796661, Global Avg Loss: 0.63690979, Time: 0.0209 Steps: 84590, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001569, Sample Num: 25104, Cur Loss: 0.16266657, Cur Avg Loss: 0.14806249, Log Avg loss: 0.11887208, Global Avg Loss: 0.63684855, Time: 0.0209 Steps: 84600, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001579, Sample Num: 25264, Cur Loss: 0.22364512, Cur Avg Loss: 0.14847243, Log Avg loss: 0.21279167, Global Avg Loss: 0.63679843, Time: 0.0209 Steps: 84610, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001589, Sample Num: 25424, Cur Loss: 0.03265422, Cur Avg Loss: 0.14824063, Log Avg loss: 0.11163915, Global Avg Loss: 0.63673637, Time: 0.0209 Steps: 84620, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001599, Sample Num: 25584, Cur Loss: 0.08478194, Cur Avg Loss: 0.14798937, Log Avg loss: 0.10806342, Global Avg Loss: 0.63667390, Time: 0.0209 Steps: 84630, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001609, Sample Num: 25744, Cur Loss: 0.37563419, Cur Avg Loss: 0.14805890, Log Avg loss: 0.15917772, Global Avg Loss: 0.63661749, Time: 0.0209 Steps: 84640, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001619, Sample Num: 25904, Cur Loss: 0.16748473, Cur Avg Loss: 0.14841824, Log Avg loss: 0.20623629, Global Avg Loss: 0.63656665, Time: 0.0209 Steps: 84650, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001629, Sample Num: 26064, Cur Loss: 0.28188637, Cur Avg Loss: 0.14844655, Log Avg loss: 0.15302976, Global Avg Loss: 0.63650953, Time: 0.0209 Steps: 84660, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001639, Sample Num: 26224, Cur Loss: 0.24427068, Cur Avg Loss: 0.14844942, Log Avg loss: 0.14891661, Global Avg Loss: 0.63645194, Time: 0.0208 Steps: 84670, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001649, Sample Num: 26384, Cur Loss: 0.07872187, Cur Avg Loss: 0.14825727, Log Avg loss: 0.11676414, Global Avg Loss: 0.63639057, Time: 0.0209 Steps: 84680, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001659, Sample Num: 26544, Cur Loss: 0.10402253, Cur Avg Loss: 0.14818919, Log Avg loss: 0.13696244, Global Avg Loss: 0.63633160, Time: 0.0209 Steps: 84690, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001669, Sample Num: 26704, Cur Loss: 0.42464876, Cur Avg Loss: 0.14797841, Log Avg loss: 0.11301027, Global Avg Loss: 0.63626982, Time: 0.0208 Steps: 84700, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001679, Sample Num: 26864, Cur Loss: 0.21654341, Cur Avg Loss: 0.14821762, Log Avg loss: 0.18814145, Global Avg Loss: 0.63621691, Time: 0.0209 Steps: 84710, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001689, Sample Num: 27024, Cur Loss: 0.20890671, Cur Avg Loss: 0.14835171, Log Avg loss: 0.17086554, Global Avg Loss: 0.63616199, Time: 0.0209 Steps: 84720, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001699, Sample Num: 27184, Cur Loss: 0.14745390, Cur Avg Loss: 0.14864713, Log Avg loss: 0.19854449, Global Avg Loss: 0.63611034, Time: 0.0209 Steps: 84730, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001709, Sample Num: 27344, Cur Loss: 0.40991551, Cur Avg Loss: 0.14848970, Log Avg loss: 0.12174234, Global Avg Loss: 0.63604964, Time: 0.0209 Steps: 84740, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001719, Sample Num: 27504, Cur Loss: 0.06511388, Cur Avg Loss: 0.14843921, Log Avg loss: 0.13981021, Global Avg Loss: 0.63599108, Time: 0.0209 Steps: 84750, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001729, Sample Num: 27664, Cur Loss: 0.07630260, Cur Avg Loss: 0.14854155, Log Avg loss: 0.16613396, Global Avg Loss: 0.63593565, Time: 0.0209 Steps: 84760, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001739, Sample Num: 27824, Cur Loss: 0.11713043, Cur Avg Loss: 0.14858683, Log Avg loss: 0.15641517, Global Avg Loss: 0.63587908, Time: 0.0209 Steps: 84770, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001749, Sample Num: 27984, Cur Loss: 0.08087880, Cur Avg Loss: 0.14870161, Log Avg loss: 0.16866173, Global Avg Loss: 0.63582397, Time: 0.0209 Steps: 84780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001759, Sample Num: 28144, Cur Loss: 0.11878275, Cur Avg Loss: 0.14852349, Log Avg loss: 0.11737026, Global Avg Loss: 0.63576283, Time: 0.0209 Steps: 84790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001769, Sample Num: 28304, Cur Loss: 0.28759775, Cur Avg Loss: 0.14882291, Log Avg loss: 0.20149119, Global Avg Loss: 0.63571162, Time: 0.0208 Steps: 84800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001779, Sample Num: 28464, Cur Loss: 0.07597852, Cur Avg Loss: 0.14874755, Log Avg loss: 0.13541598, Global Avg Loss: 0.63565263, Time: 0.0209 Steps: 84810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001789, Sample Num: 28624, Cur Loss: 0.29231781, Cur Avg Loss: 0.14850905, Log Avg loss: 0.10607941, Global Avg Loss: 0.63559019, Time: 0.0209 Steps: 84820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001799, Sample Num: 28784, Cur Loss: 0.17276169, Cur Avg Loss: 0.14844342, Log Avg loss: 0.13670369, Global Avg Loss: 0.63553138, Time: 0.0209 Steps: 84830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001809, Sample Num: 28944, Cur Loss: 0.36116114, Cur Avg Loss: 0.14880185, Log Avg loss: 0.21328232, Global Avg Loss: 0.63548161, Time: 0.0209 Steps: 84840, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001819, Sample Num: 29104, Cur Loss: 0.11830804, Cur Avg Loss: 0.14886006, Log Avg loss: 0.15939086, Global Avg Loss: 0.63542550, Time: 0.0208 Steps: 84850, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001829, Sample Num: 29264, Cur Loss: 0.19760697, Cur Avg Loss: 0.14907351, Log Avg loss: 0.18790061, Global Avg Loss: 0.63537277, Time: 0.0208 Steps: 84860, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001839, Sample Num: 29424, Cur Loss: 0.07463356, Cur Avg Loss: 0.14884057, Log Avg loss: 0.10623415, Global Avg Loss: 0.63531042, Time: 0.0208 Steps: 84870, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001849, Sample Num: 29584, Cur Loss: 0.14564611, Cur Avg Loss: 0.14879005, Log Avg loss: 0.13950007, Global Avg Loss: 0.63525201, Time: 0.0209 Steps: 84880, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001859, Sample Num: 29744, Cur Loss: 0.09931540, Cur Avg Loss: 0.14889920, Log Avg loss: 0.16908140, Global Avg Loss: 0.63519709, Time: 0.0209 Steps: 84890, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001869, Sample Num: 29904, Cur Loss: 0.09149340, Cur Avg Loss: 0.14877266, Log Avg loss: 0.12524843, Global Avg Loss: 0.63513703, Time: 0.0209 Steps: 84900, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001879, Sample Num: 30064, Cur Loss: 0.20628941, Cur Avg Loss: 0.14899533, Log Avg loss: 0.19061272, Global Avg Loss: 0.63508467, Time: 0.0209 Steps: 84910, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001889, Sample Num: 30224, Cur Loss: 0.13629600, Cur Avg Loss: 0.14903125, Log Avg loss: 0.15578004, Global Avg Loss: 0.63502823, Time: 0.0209 Steps: 84920, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001899, Sample Num: 30384, Cur Loss: 0.04169168, Cur Avg Loss: 0.14876062, Log Avg loss: 0.09763827, Global Avg Loss: 0.63496496, Time: 0.0208 Steps: 84930, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001909, Sample Num: 30544, Cur Loss: 0.36492467, Cur Avg Loss: 0.14897022, Log Avg loss: 0.18877483, Global Avg Loss: 0.63491243, Time: 0.0208 Steps: 84940, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001919, Sample Num: 30704, Cur Loss: 0.03992501, Cur Avg Loss: 0.14903122, Log Avg loss: 0.16067557, Global Avg Loss: 0.63485660, Time: 0.0208 Steps: 84950, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001929, Sample Num: 30864, Cur Loss: 0.05987551, Cur Avg Loss: 0.14887345, Log Avg loss: 0.11859765, Global Avg Loss: 0.63479584, Time: 0.0208 Steps: 84960, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001939, Sample Num: 31024, Cur Loss: 0.20025650, Cur Avg Loss: 0.14894693, Log Avg loss: 0.16312011, Global Avg Loss: 0.63474033, Time: 0.0208 Steps: 84970, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001949, Sample Num: 31184, Cur Loss: 0.27815765, Cur Avg Loss: 0.14885011, Log Avg loss: 0.13007734, Global Avg Loss: 0.63468094, Time: 0.0209 Steps: 84980, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001959, Sample Num: 31344, Cur Loss: 0.12123308, Cur Avg Loss: 0.14900559, Log Avg loss: 0.17930881, Global Avg Loss: 0.63462736, Time: 0.0209 Steps: 84990, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001969, Sample Num: 31504, Cur Loss: 0.08007150, Cur Avg Loss: 0.14926894, Log Avg loss: 0.20085876, Global Avg Loss: 0.63457633, Time: 0.0208 Steps: 85000, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001979, Sample Num: 31664, Cur Loss: 0.03464179, Cur Avg Loss: 0.14899461, Log Avg loss: 0.09497913, Global Avg Loss: 0.63451285, Time: 0.0209 Steps: 85010, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001989, Sample Num: 31824, Cur Loss: 0.05422942, Cur Avg Loss: 0.14902490, Log Avg loss: 0.15501934, Global Avg Loss: 0.63445646, Time: 0.0209 Steps: 85020, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001999, Sample Num: 31984, Cur Loss: 0.39168179, Cur Avg Loss: 0.14898609, Log Avg loss: 0.14126781, Global Avg Loss: 0.63439845, Time: 0.0208 Steps: 85030, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002009, Sample Num: 32144, Cur Loss: 0.09438303, Cur Avg Loss: 0.14915203, Log Avg loss: 0.18232348, Global Avg Loss: 0.63434529, Time: 0.0209 Steps: 85040, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002019, Sample Num: 32304, Cur Loss: 0.15779281, Cur Avg Loss: 0.14904985, Log Avg loss: 0.12852020, Global Avg Loss: 0.63428582, Time: 0.0209 Steps: 85050, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002029, Sample Num: 32464, Cur Loss: 0.11540630, Cur Avg Loss: 0.14920149, Log Avg loss: 0.17981903, Global Avg Loss: 0.63423239, Time: 0.0208 Steps: 85060, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002039, Sample Num: 32624, Cur Loss: 0.06413183, Cur Avg Loss: 0.14932302, Log Avg loss: 0.17398072, Global Avg Loss: 0.63417829, Time: 0.0209 Steps: 85070, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002049, Sample Num: 32784, Cur Loss: 0.04163414, Cur Avg Loss: 0.14978754, Log Avg loss: 0.24450339, Global Avg Loss: 0.63413249, Time: 0.0245 Steps: 85080, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002059, Sample Num: 32944, Cur Loss: 0.09946388, Cur Avg Loss: 0.14964415, Log Avg loss: 0.12026412, Global Avg Loss: 0.63407210, Time: 0.0209 Steps: 85090, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002069, Sample Num: 33104, Cur Loss: 0.07391715, Cur Avg Loss: 0.14940981, Log Avg loss: 0.10115946, Global Avg Loss: 0.63400948, Time: 0.0209 Steps: 85100, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002079, Sample Num: 33264, Cur Loss: 0.08629704, Cur Avg Loss: 0.14946368, Log Avg loss: 0.16060768, Global Avg Loss: 0.63395385, Time: 0.0208 Steps: 85110, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002089, Sample Num: 33424, Cur Loss: 0.13126604, Cur Avg Loss: 0.14910360, Log Avg loss: 0.07424432, Global Avg Loss: 0.63388810, Time: 0.0209 Steps: 85120, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002099, Sample Num: 33584, Cur Loss: 0.05962432, Cur Avg Loss: 0.14906140, Log Avg loss: 0.14024582, Global Avg Loss: 0.63383011, Time: 0.0209 Steps: 85130, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002109, Sample Num: 33744, Cur Loss: 0.32348150, Cur Avg Loss: 0.14919834, Log Avg loss: 0.17794114, Global Avg Loss: 0.63377656, Time: 0.0209 Steps: 85140, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002119, Sample Num: 33904, Cur Loss: 0.16822699, Cur Avg Loss: 0.14917553, Log Avg loss: 0.14436432, Global Avg Loss: 0.63371909, Time: 0.0209 Steps: 85150, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002129, Sample Num: 34055, Cur Loss: 0.05673045, Cur Avg Loss: 0.14952132, Log Avg loss: 0.22279455, Global Avg Loss: 0.63367083, Time: 0.0101 Steps: 85160, Updated lr: 0.000020 ***** Running evaluation checkpoint-85160 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-85160 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.758763, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.261016, "eval_total_loss": 183.493914, "eval_mae": 0.335994, "eval_mse": 0.261117, "eval_r2": 0.834017, "eval_sp_statistic": 0.928281, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.934045, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.185601, "test_total_loss": 93.171683, "test_mae": 0.316633, "test_mse": 0.185667, "test_r2": 0.880169, "test_sp_statistic": 0.919629, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.94744, "test_ps_pvalue": 0.0, "lr": 2.0189663347558086e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6336708349170502, "train_cur_epoch_loss": 318.33088543545455, "train_cur_epoch_avg_loss": 0.1495213177244972, "train_cur_epoch_time": 44.75876331329346, "train_cur_epoch_avg_time": 0.021023374031608012, "epoch": 40, "step": 85160} ################################################## Training, Epoch: 0041, Batch: 000010, Sample Num: 160, Cur Loss: 0.10442813, Cur Avg Loss: 0.20842395, Log Avg loss: 0.20842395, Global Avg Loss: 0.63362091, Time: 0.0211 Steps: 85170, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000020, Sample Num: 320, Cur Loss: 0.19585699, Cur Avg Loss: 0.15367172, Log Avg loss: 0.09891949, Global Avg Loss: 0.63355813, Time: 0.0209 Steps: 85180, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000030, Sample Num: 480, Cur Loss: 0.09673762, Cur Avg Loss: 0.15136521, Log Avg loss: 0.14675219, Global Avg Loss: 0.63350099, Time: 0.0210 Steps: 85190, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000040, Sample Num: 640, Cur Loss: 0.05266429, Cur Avg Loss: 0.15487583, Log Avg loss: 0.16540769, Global Avg Loss: 0.63344605, Time: 0.0209 Steps: 85200, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000050, Sample Num: 800, Cur Loss: 0.04142579, Cur Avg Loss: 0.15789947, Log Avg loss: 0.16999402, Global Avg Loss: 0.63339166, Time: 0.0210 Steps: 85210, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000060, Sample Num: 960, Cur Loss: 0.31834167, Cur Avg Loss: 0.16205028, Log Avg loss: 0.18280436, Global Avg Loss: 0.63333879, Time: 0.0209 Steps: 85220, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000070, Sample Num: 1120, Cur Loss: 0.07317754, Cur Avg Loss: 0.15441118, Log Avg loss: 0.10857657, Global Avg Loss: 0.63327722, Time: 0.0210 Steps: 85230, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000080, Sample Num: 1280, Cur Loss: 0.06407014, Cur Avg Loss: 0.14994110, Log Avg loss: 0.11865053, Global Avg Loss: 0.63321684, Time: 0.0209 Steps: 85240, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000090, Sample Num: 1440, Cur Loss: 0.15019959, Cur Avg Loss: 0.15197815, Log Avg loss: 0.16827455, Global Avg Loss: 0.63316230, Time: 0.0209 Steps: 85250, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000100, Sample Num: 1600, Cur Loss: 0.03923720, Cur Avg Loss: 0.15098643, Log Avg loss: 0.14206099, Global Avg Loss: 0.63310470, Time: 0.0210 Steps: 85260, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000110, Sample Num: 1760, Cur Loss: 0.03574305, Cur Avg Loss: 0.14592943, Log Avg loss: 0.09535937, Global Avg Loss: 0.63304164, Time: 0.0209 Steps: 85270, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000120, Sample Num: 1920, Cur Loss: 0.16842669, Cur Avg Loss: 0.14578400, Log Avg loss: 0.14418432, Global Avg Loss: 0.63298431, Time: 0.0209 Steps: 85280, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000130, Sample Num: 2080, Cur Loss: 0.05380036, Cur Avg Loss: 0.14368476, Log Avg loss: 0.11849381, Global Avg Loss: 0.63292399, Time: 0.0210 Steps: 85290, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000140, Sample Num: 2240, Cur Loss: 0.33183816, Cur Avg Loss: 0.14642380, Log Avg loss: 0.18203134, Global Avg Loss: 0.63287113, Time: 0.0210 Steps: 85300, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000150, Sample Num: 2400, Cur Loss: 0.12650545, Cur Avg Loss: 0.14354620, Log Avg loss: 0.10325978, Global Avg Loss: 0.63280905, Time: 0.0210 Steps: 85310, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000160, Sample Num: 2560, Cur Loss: 0.10283593, Cur Avg Loss: 0.14479576, Log Avg loss: 0.16353923, Global Avg Loss: 0.63275405, Time: 0.0210 Steps: 85320, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000170, Sample Num: 2720, Cur Loss: 0.35677227, Cur Avg Loss: 0.14414808, Log Avg loss: 0.13378522, Global Avg Loss: 0.63269558, Time: 0.0210 Steps: 85330, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000180, Sample Num: 2880, Cur Loss: 0.26975191, Cur Avg Loss: 0.14282018, Log Avg loss: 0.12024582, Global Avg Loss: 0.63263553, Time: 0.0210 Steps: 85340, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000190, Sample Num: 3040, Cur Loss: 0.07781658, Cur Avg Loss: 0.14053102, Log Avg loss: 0.09932612, Global Avg Loss: 0.63257304, Time: 0.0210 Steps: 85350, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000200, Sample Num: 3200, Cur Loss: 0.06520109, Cur Avg Loss: 0.13912161, Log Avg loss: 0.11234279, Global Avg Loss: 0.63251210, Time: 0.0210 Steps: 85360, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000210, Sample Num: 3360, Cur Loss: 0.02287286, Cur Avg Loss: 0.14219865, Log Avg loss: 0.20373948, Global Avg Loss: 0.63246187, Time: 0.0210 Steps: 85370, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000220, Sample Num: 3520, Cur Loss: 0.22377312, Cur Avg Loss: 0.14129643, Log Avg loss: 0.12234987, Global Avg Loss: 0.63240213, Time: 0.0210 Steps: 85380, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000230, Sample Num: 3680, Cur Loss: 0.16774502, Cur Avg Loss: 0.14176133, Log Avg loss: 0.15198903, Global Avg Loss: 0.63234586, Time: 0.0210 Steps: 85390, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000240, Sample Num: 3840, Cur Loss: 0.18878755, Cur Avg Loss: 0.14346439, Log Avg loss: 0.18263491, Global Avg Loss: 0.63229321, Time: 0.0210 Steps: 85400, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000250, Sample Num: 4000, Cur Loss: 0.04772716, Cur Avg Loss: 0.14326616, Log Avg loss: 0.13850871, Global Avg Loss: 0.63223539, Time: 0.0210 Steps: 85410, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000260, Sample Num: 4160, Cur Loss: 0.19885275, Cur Avg Loss: 0.14284674, Log Avg loss: 0.13236107, Global Avg Loss: 0.63217687, Time: 0.0249 Steps: 85420, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000270, Sample Num: 4320, Cur Loss: 0.15585843, Cur Avg Loss: 0.14410390, Log Avg loss: 0.17679021, Global Avg Loss: 0.63212357, Time: 0.0211 Steps: 85430, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000280, Sample Num: 4480, Cur Loss: 0.04054384, Cur Avg Loss: 0.14289896, Log Avg loss: 0.11036552, Global Avg Loss: 0.63206250, Time: 0.0211 Steps: 85440, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000290, Sample Num: 4640, Cur Loss: 0.32545012, Cur Avg Loss: 0.14460798, Log Avg loss: 0.19246046, Global Avg Loss: 0.63201105, Time: 0.0211 Steps: 85450, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000300, Sample Num: 4800, Cur Loss: 0.27177921, Cur Avg Loss: 0.14482543, Log Avg loss: 0.15113140, Global Avg Loss: 0.63195479, Time: 0.0211 Steps: 85460, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000310, Sample Num: 4960, Cur Loss: 0.03975291, Cur Avg Loss: 0.14459854, Log Avg loss: 0.13779207, Global Avg Loss: 0.63189697, Time: 0.0211 Steps: 85470, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000320, Sample Num: 5120, Cur Loss: 0.12943378, Cur Avg Loss: 0.14438429, Log Avg loss: 0.13774260, Global Avg Loss: 0.63183916, Time: 0.0211 Steps: 85480, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000330, Sample Num: 5280, Cur Loss: 0.12114910, Cur Avg Loss: 0.14282357, Log Avg loss: 0.09288026, Global Avg Loss: 0.63177612, Time: 0.0211 Steps: 85490, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000340, Sample Num: 5440, Cur Loss: 0.06592519, Cur Avg Loss: 0.14269276, Log Avg loss: 0.13837618, Global Avg Loss: 0.63171841, Time: 0.0211 Steps: 85500, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000350, Sample Num: 5600, Cur Loss: 0.10273421, Cur Avg Loss: 0.14245914, Log Avg loss: 0.13451586, Global Avg Loss: 0.63166026, Time: 0.0211 Steps: 85510, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000360, Sample Num: 5760, Cur Loss: 0.10612758, Cur Avg Loss: 0.14266096, Log Avg loss: 0.14972480, Global Avg Loss: 0.63160391, Time: 0.0211 Steps: 85520, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000370, Sample Num: 5920, Cur Loss: 0.03548833, Cur Avg Loss: 0.14252171, Log Avg loss: 0.13750875, Global Avg Loss: 0.63154614, Time: 0.0210 Steps: 85530, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000380, Sample Num: 6080, Cur Loss: 0.16011512, Cur Avg Loss: 0.14130956, Log Avg loss: 0.09646017, Global Avg Loss: 0.63148359, Time: 0.0211 Steps: 85540, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000390, Sample Num: 6240, Cur Loss: 0.09776910, Cur Avg Loss: 0.14161967, Log Avg loss: 0.15340382, Global Avg Loss: 0.63142770, Time: 0.0211 Steps: 85550, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000400, Sample Num: 6400, Cur Loss: 0.04535687, Cur Avg Loss: 0.14165965, Log Avg loss: 0.14321868, Global Avg Loss: 0.63137064, Time: 0.0211 Steps: 85560, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000410, Sample Num: 6560, Cur Loss: 0.30758667, Cur Avg Loss: 0.14133457, Log Avg loss: 0.12833158, Global Avg Loss: 0.63131186, Time: 0.0211 Steps: 85570, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000420, Sample Num: 6720, Cur Loss: 0.24790916, Cur Avg Loss: 0.14069559, Log Avg loss: 0.11449714, Global Avg Loss: 0.63125147, Time: 0.0211 Steps: 85580, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000430, Sample Num: 6880, Cur Loss: 0.19340071, Cur Avg Loss: 0.14121811, Log Avg loss: 0.16316412, Global Avg Loss: 0.63119678, Time: 0.0211 Steps: 85590, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000440, Sample Num: 7040, Cur Loss: 0.06180581, Cur Avg Loss: 0.14145691, Log Avg loss: 0.15172515, Global Avg Loss: 0.63114076, Time: 0.0211 Steps: 85600, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000450, Sample Num: 7200, Cur Loss: 0.22857112, Cur Avg Loss: 0.14222278, Log Avg loss: 0.17592111, Global Avg Loss: 0.63108759, Time: 0.0211 Steps: 85610, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000460, Sample Num: 7360, Cur Loss: 0.18424812, Cur Avg Loss: 0.14227670, Log Avg loss: 0.14470312, Global Avg Loss: 0.63103078, Time: 0.0211 Steps: 85620, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000470, Sample Num: 7520, Cur Loss: 0.50719905, Cur Avg Loss: 0.14296503, Log Avg loss: 0.17462822, Global Avg Loss: 0.63097748, Time: 0.0211 Steps: 85630, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000480, Sample Num: 7680, Cur Loss: 0.08766317, Cur Avg Loss: 0.14401208, Log Avg loss: 0.19322359, Global Avg Loss: 0.63092637, Time: 0.0210 Steps: 85640, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000490, Sample Num: 7840, Cur Loss: 0.17400654, Cur Avg Loss: 0.14405757, Log Avg loss: 0.14624099, Global Avg Loss: 0.63086978, Time: 0.0210 Steps: 85650, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000500, Sample Num: 8000, Cur Loss: 0.26165992, Cur Avg Loss: 0.14387211, Log Avg loss: 0.13478458, Global Avg Loss: 0.63081187, Time: 0.0211 Steps: 85660, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000510, Sample Num: 8160, Cur Loss: 0.07870726, Cur Avg Loss: 0.14335738, Log Avg loss: 0.11762103, Global Avg Loss: 0.63075196, Time: 0.0211 Steps: 85670, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000520, Sample Num: 8320, Cur Loss: 0.05721359, Cur Avg Loss: 0.14320565, Log Avg loss: 0.13546695, Global Avg Loss: 0.63069416, Time: 0.0210 Steps: 85680, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000530, Sample Num: 8480, Cur Loss: 0.29045632, Cur Avg Loss: 0.14296235, Log Avg loss: 0.13031107, Global Avg Loss: 0.63063576, Time: 0.0209 Steps: 85690, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000540, Sample Num: 8640, Cur Loss: 0.05262892, Cur Avg Loss: 0.14299041, Log Avg loss: 0.14447751, Global Avg Loss: 0.63057903, Time: 0.0209 Steps: 85700, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000550, Sample Num: 8800, Cur Loss: 0.40772599, Cur Avg Loss: 0.14304970, Log Avg loss: 0.14625132, Global Avg Loss: 0.63052253, Time: 0.0209 Steps: 85710, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000560, Sample Num: 8960, Cur Loss: 0.05263901, Cur Avg Loss: 0.14311723, Log Avg loss: 0.14683124, Global Avg Loss: 0.63046610, Time: 0.0209 Steps: 85720, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000570, Sample Num: 9120, Cur Loss: 0.04622281, Cur Avg Loss: 0.14304216, Log Avg loss: 0.13883864, Global Avg Loss: 0.63040875, Time: 0.0209 Steps: 85730, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000580, Sample Num: 9280, Cur Loss: 0.19303760, Cur Avg Loss: 0.14289510, Log Avg loss: 0.13451230, Global Avg Loss: 0.63035092, Time: 0.0209 Steps: 85740, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000590, Sample Num: 9440, Cur Loss: 0.03852215, Cur Avg Loss: 0.14304225, Log Avg loss: 0.15157686, Global Avg Loss: 0.63029508, Time: 0.0209 Steps: 85750, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000600, Sample Num: 9600, Cur Loss: 0.09584890, Cur Avg Loss: 0.14252916, Log Avg loss: 0.11225696, Global Avg Loss: 0.63023468, Time: 0.0209 Steps: 85760, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000610, Sample Num: 9760, Cur Loss: 0.24708450, Cur Avg Loss: 0.14189115, Log Avg loss: 0.10361053, Global Avg Loss: 0.63017328, Time: 0.0210 Steps: 85770, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000620, Sample Num: 9920, Cur Loss: 0.07302520, Cur Avg Loss: 0.14249455, Log Avg loss: 0.17930184, Global Avg Loss: 0.63012071, Time: 0.0210 Steps: 85780, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000630, Sample Num: 10080, Cur Loss: 0.07928172, Cur Avg Loss: 0.14388226, Log Avg loss: 0.22992087, Global Avg Loss: 0.63007407, Time: 0.0209 Steps: 85790, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000640, Sample Num: 10240, Cur Loss: 0.07316538, Cur Avg Loss: 0.14469304, Log Avg loss: 0.19577164, Global Avg Loss: 0.63002345, Time: 0.0208 Steps: 85800, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000650, Sample Num: 10400, Cur Loss: 0.03595225, Cur Avg Loss: 0.14415071, Log Avg loss: 0.10944150, Global Avg Loss: 0.62996278, Time: 0.0209 Steps: 85810, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000660, Sample Num: 10560, Cur Loss: 0.18301730, Cur Avg Loss: 0.14365459, Log Avg loss: 0.11140740, Global Avg Loss: 0.62990236, Time: 0.0209 Steps: 85820, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000670, Sample Num: 10720, Cur Loss: 0.48010743, Cur Avg Loss: 0.14369586, Log Avg loss: 0.14641916, Global Avg Loss: 0.62984603, Time: 0.0209 Steps: 85830, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000680, Sample Num: 10880, Cur Loss: 0.12423896, Cur Avg Loss: 0.14317955, Log Avg loss: 0.10858688, Global Avg Loss: 0.62978530, Time: 0.0210 Steps: 85840, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000690, Sample Num: 11040, Cur Loss: 0.18785742, Cur Avg Loss: 0.14348962, Log Avg loss: 0.16457436, Global Avg Loss: 0.62973111, Time: 0.0210 Steps: 85850, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000700, Sample Num: 11200, Cur Loss: 0.18620837, Cur Avg Loss: 0.14333804, Log Avg loss: 0.13287946, Global Avg Loss: 0.62967325, Time: 0.0210 Steps: 85860, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000710, Sample Num: 11360, Cur Loss: 0.14099069, Cur Avg Loss: 0.14296401, Log Avg loss: 0.11678163, Global Avg Loss: 0.62961352, Time: 0.0209 Steps: 85870, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000720, Sample Num: 11520, Cur Loss: 0.18761599, Cur Avg Loss: 0.14329774, Log Avg loss: 0.16699287, Global Avg Loss: 0.62955965, Time: 0.0209 Steps: 85880, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000730, Sample Num: 11680, Cur Loss: 0.14162834, Cur Avg Loss: 0.14320047, Log Avg loss: 0.13619696, Global Avg Loss: 0.62950221, Time: 0.0209 Steps: 85890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000740, Sample Num: 11840, Cur Loss: 0.03512849, Cur Avg Loss: 0.14284075, Log Avg loss: 0.11658098, Global Avg Loss: 0.62944250, Time: 0.0208 Steps: 85900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000750, Sample Num: 12000, Cur Loss: 0.02780909, Cur Avg Loss: 0.14280210, Log Avg loss: 0.13994162, Global Avg Loss: 0.62938552, Time: 0.0209 Steps: 85910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000760, Sample Num: 12160, Cur Loss: 0.37040678, Cur Avg Loss: 0.14307482, Log Avg loss: 0.16352924, Global Avg Loss: 0.62933130, Time: 0.0208 Steps: 85920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000770, Sample Num: 12320, Cur Loss: 0.08440731, Cur Avg Loss: 0.14371558, Log Avg loss: 0.19241292, Global Avg Loss: 0.62928045, Time: 0.0247 Steps: 85930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000780, Sample Num: 12480, Cur Loss: 0.13178937, Cur Avg Loss: 0.14383371, Log Avg loss: 0.15292983, Global Avg Loss: 0.62922502, Time: 0.0209 Steps: 85940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000790, Sample Num: 12640, Cur Loss: 0.18104947, Cur Avg Loss: 0.14365528, Log Avg loss: 0.12973815, Global Avg Loss: 0.62916691, Time: 0.0211 Steps: 85950, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000800, Sample Num: 12800, Cur Loss: 0.02687873, Cur Avg Loss: 0.14351327, Log Avg loss: 0.13229469, Global Avg Loss: 0.62910911, Time: 0.0209 Steps: 85960, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000810, Sample Num: 12960, Cur Loss: 0.13915546, Cur Avg Loss: 0.14316466, Log Avg loss: 0.11527521, Global Avg Loss: 0.62904934, Time: 0.0209 Steps: 85970, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000820, Sample Num: 13120, Cur Loss: 0.21191427, Cur Avg Loss: 0.14296635, Log Avg loss: 0.12690372, Global Avg Loss: 0.62899094, Time: 0.0209 Steps: 85980, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000830, Sample Num: 13280, Cur Loss: 0.05822618, Cur Avg Loss: 0.14251046, Log Avg loss: 0.10512731, Global Avg Loss: 0.62893001, Time: 0.0209 Steps: 85990, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000840, Sample Num: 13440, Cur Loss: 0.13315047, Cur Avg Loss: 0.14204513, Log Avg loss: 0.10342269, Global Avg Loss: 0.62886891, Time: 0.0209 Steps: 86000, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000850, Sample Num: 13600, Cur Loss: 0.08357367, Cur Avg Loss: 0.14259546, Log Avg loss: 0.18882317, Global Avg Loss: 0.62881775, Time: 0.0209 Steps: 86010, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000860, Sample Num: 13760, Cur Loss: 0.07427064, Cur Avg Loss: 0.14216146, Log Avg loss: 0.10527118, Global Avg Loss: 0.62875688, Time: 0.0209 Steps: 86020, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000870, Sample Num: 13920, Cur Loss: 0.03939662, Cur Avg Loss: 0.14168006, Log Avg loss: 0.10028033, Global Avg Loss: 0.62869545, Time: 0.0209 Steps: 86030, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000880, Sample Num: 14080, Cur Loss: 0.10395341, Cur Avg Loss: 0.14172326, Log Avg loss: 0.14548145, Global Avg Loss: 0.62863929, Time: 0.0209 Steps: 86040, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000890, Sample Num: 14240, Cur Loss: 0.02040264, Cur Avg Loss: 0.14190905, Log Avg loss: 0.15825841, Global Avg Loss: 0.62858463, Time: 0.0210 Steps: 86050, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000900, Sample Num: 14400, Cur Loss: 0.08872581, Cur Avg Loss: 0.14166102, Log Avg loss: 0.11958620, Global Avg Loss: 0.62852548, Time: 0.0209 Steps: 86060, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000910, Sample Num: 14560, Cur Loss: 0.06516518, Cur Avg Loss: 0.14170423, Log Avg loss: 0.14559318, Global Avg Loss: 0.62846938, Time: 0.0209 Steps: 86070, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000920, Sample Num: 14720, Cur Loss: 0.07633074, Cur Avg Loss: 0.14211370, Log Avg loss: 0.17937549, Global Avg Loss: 0.62841720, Time: 0.0210 Steps: 86080, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000930, Sample Num: 14880, Cur Loss: 0.11181276, Cur Avg Loss: 0.14229431, Log Avg loss: 0.15891013, Global Avg Loss: 0.62836267, Time: 0.0209 Steps: 86090, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000940, Sample Num: 15040, Cur Loss: 0.07845124, Cur Avg Loss: 0.14279535, Log Avg loss: 0.18939207, Global Avg Loss: 0.62831168, Time: 0.0209 Steps: 86100, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000950, Sample Num: 15200, Cur Loss: 0.23664039, Cur Avg Loss: 0.14277736, Log Avg loss: 0.14108715, Global Avg Loss: 0.62825510, Time: 0.0209 Steps: 86110, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000960, Sample Num: 15360, Cur Loss: 0.10596828, Cur Avg Loss: 0.14264944, Log Avg loss: 0.13049689, Global Avg Loss: 0.62819730, Time: 0.0209 Steps: 86120, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000970, Sample Num: 15520, Cur Loss: 0.08089401, Cur Avg Loss: 0.14286728, Log Avg loss: 0.16377947, Global Avg Loss: 0.62814338, Time: 0.0209 Steps: 86130, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000980, Sample Num: 15680, Cur Loss: 0.36409527, Cur Avg Loss: 0.14329162, Log Avg loss: 0.18445233, Global Avg Loss: 0.62809187, Time: 0.0209 Steps: 86140, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000990, Sample Num: 15840, Cur Loss: 0.15143228, Cur Avg Loss: 0.14377701, Log Avg loss: 0.19134569, Global Avg Loss: 0.62804118, Time: 0.0209 Steps: 86150, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001000, Sample Num: 16000, Cur Loss: 0.08083990, Cur Avg Loss: 0.14387461, Log Avg loss: 0.15353713, Global Avg Loss: 0.62798611, Time: 0.0209 Steps: 86160, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001010, Sample Num: 16160, Cur Loss: 0.10285643, Cur Avg Loss: 0.14371140, Log Avg loss: 0.12738992, Global Avg Loss: 0.62792801, Time: 0.0209 Steps: 86170, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001020, Sample Num: 16320, Cur Loss: 0.24341476, Cur Avg Loss: 0.14400591, Log Avg loss: 0.17375191, Global Avg Loss: 0.62787531, Time: 0.0209 Steps: 86180, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001030, Sample Num: 16480, Cur Loss: 0.14823556, Cur Avg Loss: 0.14474109, Log Avg loss: 0.21972950, Global Avg Loss: 0.62782796, Time: 0.0244 Steps: 86190, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001040, Sample Num: 16640, Cur Loss: 0.06582753, Cur Avg Loss: 0.14468203, Log Avg loss: 0.13859853, Global Avg Loss: 0.62777120, Time: 0.0209 Steps: 86200, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001050, Sample Num: 16800, Cur Loss: 0.07950933, Cur Avg Loss: 0.14463944, Log Avg loss: 0.14021063, Global Avg Loss: 0.62771465, Time: 0.0209 Steps: 86210, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001060, Sample Num: 16960, Cur Loss: 0.14588058, Cur Avg Loss: 0.14416644, Log Avg loss: 0.09450115, Global Avg Loss: 0.62765280, Time: 0.0210 Steps: 86220, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001070, Sample Num: 17120, Cur Loss: 0.20731999, Cur Avg Loss: 0.14464010, Log Avg loss: 0.19484777, Global Avg Loss: 0.62760261, Time: 0.0209 Steps: 86230, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001080, Sample Num: 17280, Cur Loss: 0.04732016, Cur Avg Loss: 0.14509118, Log Avg loss: 0.19335709, Global Avg Loss: 0.62755226, Time: 0.0210 Steps: 86240, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001090, Sample Num: 17440, Cur Loss: 0.21988872, Cur Avg Loss: 0.14566098, Log Avg loss: 0.20719917, Global Avg Loss: 0.62750352, Time: 0.0210 Steps: 86250, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001100, Sample Num: 17600, Cur Loss: 0.07990219, Cur Avg Loss: 0.14585627, Log Avg loss: 0.16714230, Global Avg Loss: 0.62745015, Time: 0.0209 Steps: 86260, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001110, Sample Num: 17760, Cur Loss: 0.18272088, Cur Avg Loss: 0.14545146, Log Avg loss: 0.10092304, Global Avg Loss: 0.62738912, Time: 0.0209 Steps: 86270, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001120, Sample Num: 17920, Cur Loss: 0.05069428, Cur Avg Loss: 0.14504614, Log Avg loss: 0.10005564, Global Avg Loss: 0.62732800, Time: 0.0210 Steps: 86280, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001130, Sample Num: 18080, Cur Loss: 0.14277843, Cur Avg Loss: 0.14514384, Log Avg loss: 0.15608655, Global Avg Loss: 0.62727339, Time: 0.0209 Steps: 86290, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001140, Sample Num: 18240, Cur Loss: 0.04582054, Cur Avg Loss: 0.14613395, Log Avg loss: 0.25801619, Global Avg Loss: 0.62723060, Time: 0.0210 Steps: 86300, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001150, Sample Num: 18400, Cur Loss: 0.08532815, Cur Avg Loss: 0.14649991, Log Avg loss: 0.18821889, Global Avg Loss: 0.62717974, Time: 0.0209 Steps: 86310, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001160, Sample Num: 18560, Cur Loss: 0.24135114, Cur Avg Loss: 0.14695002, Log Avg loss: 0.19871268, Global Avg Loss: 0.62713010, Time: 0.0209 Steps: 86320, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001170, Sample Num: 18720, Cur Loss: 0.15764260, Cur Avg Loss: 0.14694282, Log Avg loss: 0.14610737, Global Avg Loss: 0.62707438, Time: 0.0209 Steps: 86330, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001180, Sample Num: 18880, Cur Loss: 0.20396182, Cur Avg Loss: 0.14684243, Log Avg loss: 0.13509737, Global Avg Loss: 0.62701740, Time: 0.0209 Steps: 86340, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001190, Sample Num: 19040, Cur Loss: 0.31290716, Cur Avg Loss: 0.14683942, Log Avg loss: 0.14648425, Global Avg Loss: 0.62696175, Time: 0.0209 Steps: 86350, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001200, Sample Num: 19200, Cur Loss: 0.09365690, Cur Avg Loss: 0.14660878, Log Avg loss: 0.11916239, Global Avg Loss: 0.62690295, Time: 0.0209 Steps: 86360, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001210, Sample Num: 19360, Cur Loss: 0.37336254, Cur Avg Loss: 0.14676010, Log Avg loss: 0.16491847, Global Avg Loss: 0.62684946, Time: 0.0209 Steps: 86370, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001220, Sample Num: 19520, Cur Loss: 0.04276234, Cur Avg Loss: 0.14705619, Log Avg loss: 0.18288332, Global Avg Loss: 0.62679807, Time: 0.0209 Steps: 86380, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001230, Sample Num: 19680, Cur Loss: 0.05974305, Cur Avg Loss: 0.14676071, Log Avg loss: 0.11071156, Global Avg Loss: 0.62673833, Time: 0.0210 Steps: 86390, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001240, Sample Num: 19840, Cur Loss: 0.05619716, Cur Avg Loss: 0.14684572, Log Avg loss: 0.15730188, Global Avg Loss: 0.62668399, Time: 0.0209 Steps: 86400, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001250, Sample Num: 20000, Cur Loss: 0.30721378, Cur Avg Loss: 0.14690195, Log Avg loss: 0.15387456, Global Avg Loss: 0.62662928, Time: 0.0210 Steps: 86410, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001260, Sample Num: 20160, Cur Loss: 0.09423077, Cur Avg Loss: 0.14654341, Log Avg loss: 0.10172622, Global Avg Loss: 0.62656854, Time: 0.0209 Steps: 86420, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001270, Sample Num: 20320, Cur Loss: 0.20382498, Cur Avg Loss: 0.14665524, Log Avg loss: 0.16074655, Global Avg Loss: 0.62651464, Time: 0.0209 Steps: 86430, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001280, Sample Num: 20480, Cur Loss: 0.24688773, Cur Avg Loss: 0.14662549, Log Avg loss: 0.14284625, Global Avg Loss: 0.62645869, Time: 0.0255 Steps: 86440, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001290, Sample Num: 20640, Cur Loss: 0.06295533, Cur Avg Loss: 0.14650776, Log Avg loss: 0.13143825, Global Avg Loss: 0.62640143, Time: 0.0210 Steps: 86450, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001300, Sample Num: 20800, Cur Loss: 0.34854680, Cur Avg Loss: 0.14671941, Log Avg loss: 0.17402333, Global Avg Loss: 0.62634910, Time: 0.0210 Steps: 86460, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001310, Sample Num: 20960, Cur Loss: 0.15777293, Cur Avg Loss: 0.14700740, Log Avg loss: 0.18444596, Global Avg Loss: 0.62629800, Time: 0.0210 Steps: 86470, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001320, Sample Num: 21120, Cur Loss: 0.13453116, Cur Avg Loss: 0.14711898, Log Avg loss: 0.16173483, Global Avg Loss: 0.62624428, Time: 0.0209 Steps: 86480, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001330, Sample Num: 21280, Cur Loss: 0.18355992, Cur Avg Loss: 0.14720711, Log Avg loss: 0.15884076, Global Avg Loss: 0.62619024, Time: 0.0210 Steps: 86490, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001340, Sample Num: 21440, Cur Loss: 0.06415924, Cur Avg Loss: 0.14703896, Log Avg loss: 0.12467473, Global Avg Loss: 0.62613226, Time: 0.0209 Steps: 86500, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001350, Sample Num: 21600, Cur Loss: 0.26256385, Cur Avg Loss: 0.14709473, Log Avg loss: 0.15456773, Global Avg Loss: 0.62607775, Time: 0.0210 Steps: 86510, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001360, Sample Num: 21760, Cur Loss: 0.12667127, Cur Avg Loss: 0.14713972, Log Avg loss: 0.15321406, Global Avg Loss: 0.62602310, Time: 0.0209 Steps: 86520, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001370, Sample Num: 21920, Cur Loss: 0.01997884, Cur Avg Loss: 0.14685355, Log Avg loss: 0.10793476, Global Avg Loss: 0.62596322, Time: 0.0210 Steps: 86530, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001380, Sample Num: 22080, Cur Loss: 0.09889694, Cur Avg Loss: 0.14690910, Log Avg loss: 0.15451836, Global Avg Loss: 0.62590875, Time: 0.0210 Steps: 86540, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001390, Sample Num: 22240, Cur Loss: 0.21723606, Cur Avg Loss: 0.14682430, Log Avg loss: 0.13512238, Global Avg Loss: 0.62585204, Time: 0.0210 Steps: 86550, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001400, Sample Num: 22400, Cur Loss: 0.17024703, Cur Avg Loss: 0.14699011, Log Avg loss: 0.17003784, Global Avg Loss: 0.62579938, Time: 0.0210 Steps: 86560, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001410, Sample Num: 22560, Cur Loss: 0.04941906, Cur Avg Loss: 0.14704101, Log Avg loss: 0.15416665, Global Avg Loss: 0.62574490, Time: 0.0210 Steps: 86570, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001420, Sample Num: 22720, Cur Loss: 0.06844626, Cur Avg Loss: 0.14688837, Log Avg loss: 0.12536695, Global Avg Loss: 0.62568711, Time: 0.0210 Steps: 86580, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001430, Sample Num: 22880, Cur Loss: 0.11243328, Cur Avg Loss: 0.14696715, Log Avg loss: 0.15815361, Global Avg Loss: 0.62563311, Time: 0.0210 Steps: 86590, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001440, Sample Num: 23040, Cur Loss: 0.22295478, Cur Avg Loss: 0.14671664, Log Avg loss: 0.11089428, Global Avg Loss: 0.62557368, Time: 0.0209 Steps: 86600, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001450, Sample Num: 23200, Cur Loss: 0.16643630, Cur Avg Loss: 0.14646964, Log Avg loss: 0.11090056, Global Avg Loss: 0.62551425, Time: 0.0210 Steps: 86610, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001460, Sample Num: 23360, Cur Loss: 0.09704158, Cur Avg Loss: 0.14642062, Log Avg loss: 0.13931296, Global Avg Loss: 0.62545812, Time: 0.0209 Steps: 86620, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001470, Sample Num: 23520, Cur Loss: 0.25553566, Cur Avg Loss: 0.14622922, Log Avg loss: 0.11828569, Global Avg Loss: 0.62539958, Time: 0.0210 Steps: 86630, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001480, Sample Num: 23680, Cur Loss: 0.16524577, Cur Avg Loss: 0.14623981, Log Avg loss: 0.14779580, Global Avg Loss: 0.62534445, Time: 0.0210 Steps: 86640, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001490, Sample Num: 23840, Cur Loss: 0.16985148, Cur Avg Loss: 0.14641295, Log Avg loss: 0.17203838, Global Avg Loss: 0.62529214, Time: 0.0210 Steps: 86650, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001500, Sample Num: 24000, Cur Loss: 0.43581173, Cur Avg Loss: 0.14671309, Log Avg loss: 0.19143374, Global Avg Loss: 0.62524207, Time: 0.0210 Steps: 86660, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001510, Sample Num: 24160, Cur Loss: 0.07204273, Cur Avg Loss: 0.14670996, Log Avg loss: 0.14623930, Global Avg Loss: 0.62518680, Time: 0.0209 Steps: 86670, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001520, Sample Num: 24320, Cur Loss: 0.03811888, Cur Avg Loss: 0.14663750, Log Avg loss: 0.13569657, Global Avg Loss: 0.62513033, Time: 0.0210 Steps: 86680, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001530, Sample Num: 24480, Cur Loss: 0.09292820, Cur Avg Loss: 0.14657844, Log Avg loss: 0.13760115, Global Avg Loss: 0.62507410, Time: 0.0209 Steps: 86690, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001540, Sample Num: 24640, Cur Loss: 0.06896807, Cur Avg Loss: 0.14660641, Log Avg loss: 0.15088675, Global Avg Loss: 0.62501940, Time: 0.0214 Steps: 86700, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001550, Sample Num: 24800, Cur Loss: 0.09827602, Cur Avg Loss: 0.14657207, Log Avg loss: 0.14128305, Global Avg Loss: 0.62496361, Time: 0.0212 Steps: 86710, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001560, Sample Num: 24960, Cur Loss: 0.30118251, Cur Avg Loss: 0.14677507, Log Avg loss: 0.17823977, Global Avg Loss: 0.62491210, Time: 0.0212 Steps: 86720, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001570, Sample Num: 25120, Cur Loss: 0.12030205, Cur Avg Loss: 0.14650057, Log Avg loss: 0.10367972, Global Avg Loss: 0.62485200, Time: 0.0212 Steps: 86730, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001580, Sample Num: 25280, Cur Loss: 0.09773566, Cur Avg Loss: 0.14641953, Log Avg loss: 0.13369623, Global Avg Loss: 0.62479538, Time: 0.0212 Steps: 86740, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001590, Sample Num: 25440, Cur Loss: 0.04234098, Cur Avg Loss: 0.14654334, Log Avg loss: 0.16610504, Global Avg Loss: 0.62474250, Time: 0.0212 Steps: 86750, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001600, Sample Num: 25600, Cur Loss: 0.12365772, Cur Avg Loss: 0.14629490, Log Avg loss: 0.10679232, Global Avg Loss: 0.62468280, Time: 0.0211 Steps: 86760, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001610, Sample Num: 25760, Cur Loss: 0.12272419, Cur Avg Loss: 0.14642873, Log Avg loss: 0.16784159, Global Avg Loss: 0.62463016, Time: 0.0212 Steps: 86770, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001620, Sample Num: 25920, Cur Loss: 0.07592090, Cur Avg Loss: 0.14644078, Log Avg loss: 0.14838160, Global Avg Loss: 0.62457528, Time: 0.0211 Steps: 86780, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001630, Sample Num: 26080, Cur Loss: 0.15218291, Cur Avg Loss: 0.14639657, Log Avg loss: 0.13923409, Global Avg Loss: 0.62451935, Time: 0.0212 Steps: 86790, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001640, Sample Num: 26240, Cur Loss: 0.18243614, Cur Avg Loss: 0.14670596, Log Avg loss: 0.19713620, Global Avg Loss: 0.62447012, Time: 0.0211 Steps: 86800, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001650, Sample Num: 26400, Cur Loss: 0.53117996, Cur Avg Loss: 0.14690204, Log Avg loss: 0.17905989, Global Avg Loss: 0.62441881, Time: 0.0211 Steps: 86810, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001660, Sample Num: 26560, Cur Loss: 0.23460779, Cur Avg Loss: 0.14685883, Log Avg loss: 0.13972863, Global Avg Loss: 0.62436298, Time: 0.0211 Steps: 86820, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001670, Sample Num: 26720, Cur Loss: 0.08757769, Cur Avg Loss: 0.14700457, Log Avg loss: 0.17119755, Global Avg Loss: 0.62431079, Time: 0.0211 Steps: 86830, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001680, Sample Num: 26880, Cur Loss: 0.23697209, Cur Avg Loss: 0.14692535, Log Avg loss: 0.13369597, Global Avg Loss: 0.62425429, Time: 0.0211 Steps: 86840, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001690, Sample Num: 27040, Cur Loss: 0.09872349, Cur Avg Loss: 0.14705065, Log Avg loss: 0.16810043, Global Avg Loss: 0.62420177, Time: 0.0211 Steps: 86850, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001700, Sample Num: 27200, Cur Loss: 0.17008524, Cur Avg Loss: 0.14745888, Log Avg loss: 0.21644986, Global Avg Loss: 0.62415483, Time: 0.0211 Steps: 86860, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001710, Sample Num: 27360, Cur Loss: 0.25396788, Cur Avg Loss: 0.14734095, Log Avg loss: 0.12729272, Global Avg Loss: 0.62409763, Time: 0.0211 Steps: 86870, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001720, Sample Num: 27520, Cur Loss: 0.08428739, Cur Avg Loss: 0.14734811, Log Avg loss: 0.14857249, Global Avg Loss: 0.62404290, Time: 0.0211 Steps: 86880, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001730, Sample Num: 27680, Cur Loss: 0.11898704, Cur Avg Loss: 0.14724459, Log Avg loss: 0.12943989, Global Avg Loss: 0.62398598, Time: 0.0212 Steps: 86890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001740, Sample Num: 27840, Cur Loss: 0.10432141, Cur Avg Loss: 0.14714467, Log Avg loss: 0.12985750, Global Avg Loss: 0.62392911, Time: 0.0212 Steps: 86900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001750, Sample Num: 28000, Cur Loss: 0.07362461, Cur Avg Loss: 0.14700190, Log Avg loss: 0.12216043, Global Avg Loss: 0.62387138, Time: 0.0212 Steps: 86910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001760, Sample Num: 28160, Cur Loss: 0.20027077, Cur Avg Loss: 0.14694109, Log Avg loss: 0.13629985, Global Avg Loss: 0.62381529, Time: 0.0212 Steps: 86920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001770, Sample Num: 28320, Cur Loss: 0.34779191, Cur Avg Loss: 0.14694332, Log Avg loss: 0.14733518, Global Avg Loss: 0.62376047, Time: 0.0211 Steps: 86930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001780, Sample Num: 28480, Cur Loss: 0.19802886, Cur Avg Loss: 0.14693811, Log Avg loss: 0.14601572, Global Avg Loss: 0.62370552, Time: 0.0213 Steps: 86940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001790, Sample Num: 28640, Cur Loss: 0.15186158, Cur Avg Loss: 0.14690095, Log Avg loss: 0.14028692, Global Avg Loss: 0.62364993, Time: 0.0211 Steps: 86950, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001800, Sample Num: 28800, Cur Loss: 0.25276446, Cur Avg Loss: 0.14707796, Log Avg loss: 0.17876176, Global Avg Loss: 0.62359877, Time: 0.0210 Steps: 86960, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001810, Sample Num: 28960, Cur Loss: 0.07485618, Cur Avg Loss: 0.14686986, Log Avg loss: 0.10941262, Global Avg Loss: 0.62353964, Time: 0.0209 Steps: 86970, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001820, Sample Num: 29120, Cur Loss: 0.14779866, Cur Avg Loss: 0.14667122, Log Avg loss: 0.11071702, Global Avg Loss: 0.62348068, Time: 0.0209 Steps: 86980, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001830, Sample Num: 29280, Cur Loss: 0.06450623, Cur Avg Loss: 0.14679680, Log Avg loss: 0.16965260, Global Avg Loss: 0.62342851, Time: 0.0210 Steps: 86990, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001840, Sample Num: 29440, Cur Loss: 0.13879094, Cur Avg Loss: 0.14654748, Log Avg loss: 0.10092152, Global Avg Loss: 0.62336846, Time: 0.0210 Steps: 87000, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001850, Sample Num: 29600, Cur Loss: 0.04564861, Cur Avg Loss: 0.14629462, Log Avg loss: 0.09976854, Global Avg Loss: 0.62330828, Time: 0.0210 Steps: 87010, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001860, Sample Num: 29760, Cur Loss: 0.10373898, Cur Avg Loss: 0.14622971, Log Avg loss: 0.13422229, Global Avg Loss: 0.62325208, Time: 0.0210 Steps: 87020, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001870, Sample Num: 29920, Cur Loss: 0.30915266, Cur Avg Loss: 0.14640299, Log Avg loss: 0.17863333, Global Avg Loss: 0.62320099, Time: 0.0210 Steps: 87030, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001880, Sample Num: 30080, Cur Loss: 0.04068355, Cur Avg Loss: 0.14632170, Log Avg loss: 0.13111980, Global Avg Loss: 0.62314445, Time: 0.0210 Steps: 87040, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001890, Sample Num: 30240, Cur Loss: 0.16291755, Cur Avg Loss: 0.14619848, Log Avg loss: 0.12303334, Global Avg Loss: 0.62308700, Time: 0.0210 Steps: 87050, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001900, Sample Num: 30400, Cur Loss: 0.72267842, Cur Avg Loss: 0.14695134, Log Avg loss: 0.28924103, Global Avg Loss: 0.62304865, Time: 0.0210 Steps: 87060, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001910, Sample Num: 30560, Cur Loss: 0.35267809, Cur Avg Loss: 0.14705447, Log Avg loss: 0.16665006, Global Avg Loss: 0.62299624, Time: 0.0209 Steps: 87070, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001920, Sample Num: 30720, Cur Loss: 0.14719507, Cur Avg Loss: 0.14689127, Log Avg loss: 0.11571974, Global Avg Loss: 0.62293798, Time: 0.0209 Steps: 87080, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001930, Sample Num: 30880, Cur Loss: 0.43127185, Cur Avg Loss: 0.14742569, Log Avg loss: 0.25003462, Global Avg Loss: 0.62289516, Time: 0.0210 Steps: 87090, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001940, Sample Num: 31040, Cur Loss: 0.07850688, Cur Avg Loss: 0.14745343, Log Avg loss: 0.15280704, Global Avg Loss: 0.62284119, Time: 0.0210 Steps: 87100, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001950, Sample Num: 31200, Cur Loss: 0.09788503, Cur Avg Loss: 0.14766197, Log Avg loss: 0.18811820, Global Avg Loss: 0.62279129, Time: 0.0209 Steps: 87110, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001960, Sample Num: 31360, Cur Loss: 0.47212386, Cur Avg Loss: 0.14787707, Log Avg loss: 0.18982151, Global Avg Loss: 0.62274159, Time: 0.0210 Steps: 87120, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001970, Sample Num: 31520, Cur Loss: 0.42603317, Cur Avg Loss: 0.14787276, Log Avg loss: 0.14702868, Global Avg Loss: 0.62268699, Time: 0.0209 Steps: 87130, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001980, Sample Num: 31680, Cur Loss: 0.29348719, Cur Avg Loss: 0.14800625, Log Avg loss: 0.17430295, Global Avg Loss: 0.62263554, Time: 0.0210 Steps: 87140, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001990, Sample Num: 31840, Cur Loss: 0.06549259, Cur Avg Loss: 0.14804190, Log Avg loss: 0.15510106, Global Avg Loss: 0.62258189, Time: 0.0209 Steps: 87150, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002000, Sample Num: 32000, Cur Loss: 0.51085770, Cur Avg Loss: 0.14812993, Log Avg loss: 0.16564733, Global Avg Loss: 0.62252946, Time: 0.0209 Steps: 87160, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002010, Sample Num: 32160, Cur Loss: 0.08781194, Cur Avg Loss: 0.14797140, Log Avg loss: 0.11626617, Global Avg Loss: 0.62247139, Time: 0.0209 Steps: 87170, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002020, Sample Num: 32320, Cur Loss: 0.14987753, Cur Avg Loss: 0.14818083, Log Avg loss: 0.19027636, Global Avg Loss: 0.62242181, Time: 0.0210 Steps: 87180, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002030, Sample Num: 32480, Cur Loss: 0.09516972, Cur Avg Loss: 0.14830881, Log Avg loss: 0.17416051, Global Avg Loss: 0.62237040, Time: 0.0210 Steps: 87190, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002040, Sample Num: 32640, Cur Loss: 0.13035387, Cur Avg Loss: 0.14830295, Log Avg loss: 0.14711318, Global Avg Loss: 0.62231590, Time: 0.0209 Steps: 87200, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002050, Sample Num: 32800, Cur Loss: 0.02151345, Cur Avg Loss: 0.14838215, Log Avg loss: 0.16454002, Global Avg Loss: 0.62226341, Time: 0.0248 Steps: 87210, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002060, Sample Num: 32960, Cur Loss: 0.04664308, Cur Avg Loss: 0.14838599, Log Avg loss: 0.14917233, Global Avg Loss: 0.62220917, Time: 0.0210 Steps: 87220, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002070, Sample Num: 33120, Cur Loss: 0.11214515, Cur Avg Loss: 0.14832420, Log Avg loss: 0.13559643, Global Avg Loss: 0.62215338, Time: 0.0211 Steps: 87230, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002080, Sample Num: 33280, Cur Loss: 0.20805818, Cur Avg Loss: 0.14817025, Log Avg loss: 0.11630148, Global Avg Loss: 0.62209540, Time: 0.0209 Steps: 87240, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002090, Sample Num: 33440, Cur Loss: 0.14714600, Cur Avg Loss: 0.14796504, Log Avg loss: 0.10528098, Global Avg Loss: 0.62203616, Time: 0.0209 Steps: 87250, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002100, Sample Num: 33600, Cur Loss: 0.06748770, Cur Avg Loss: 0.14770193, Log Avg loss: 0.09271200, Global Avg Loss: 0.62197550, Time: 0.0210 Steps: 87260, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002110, Sample Num: 33760, Cur Loss: 0.39410114, Cur Avg Loss: 0.14786711, Log Avg loss: 0.18255443, Global Avg Loss: 0.62192515, Time: 0.0210 Steps: 87270, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002120, Sample Num: 33920, Cur Loss: 0.12869442, Cur Avg Loss: 0.14803128, Log Avg loss: 0.18267278, Global Avg Loss: 0.62187482, Time: 0.0210 Steps: 87280, Updated lr: 0.000018 ***** Running evaluation checkpoint-87289 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-87289 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.817245, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.346838, "eval_total_loss": 243.82714, "eval_mae": 0.408159, "eval_mse": 0.346974, "eval_r2": 0.77944, "eval_sp_statistic": 0.92804, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.933325, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.188801, "test_total_loss": 94.778005, "test_mae": 0.281098, "test_mse": 0.18887, "test_r2": 0.878102, "test_sp_statistic": 0.918694, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947108, "test_ps_pvalue": 0.0, "lr": 1.8170697012802277e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6218200328894529, "train_cur_epoch_loss": 314.64054935146123, "train_cur_epoch_avg_loss": 0.14778795178556187, "train_cur_epoch_time": 44.8172447681427, "train_cur_epoch_avg_time": 0.021050843009930813, "epoch": 41, "step": 87289} ################################################## Training, Epoch: 0042, Batch: 000001, Sample Num: 16, Cur Loss: 0.07296065, Cur Avg Loss: 0.07296065, Log Avg loss: 0.08871892, Global Avg Loss: 0.62181375, Time: 0.0249 Steps: 87290, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000011, Sample Num: 176, Cur Loss: 0.06681786, Cur Avg Loss: 0.10823762, Log Avg loss: 0.11176531, Global Avg Loss: 0.62175532, Time: 0.0211 Steps: 87300, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000021, Sample Num: 336, Cur Loss: 0.15342660, Cur Avg Loss: 0.10978998, Log Avg loss: 0.11149757, Global Avg Loss: 0.62169688, Time: 0.0209 Steps: 87310, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000031, Sample Num: 496, Cur Loss: 0.20964153, Cur Avg Loss: 0.12201076, Log Avg loss: 0.14767441, Global Avg Loss: 0.62164259, Time: 0.0210 Steps: 87320, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000041, Sample Num: 656, Cur Loss: 0.15520331, Cur Avg Loss: 0.12399814, Log Avg loss: 0.13015901, Global Avg Loss: 0.62158631, Time: 0.0210 Steps: 87330, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000051, Sample Num: 816, Cur Loss: 0.05527736, Cur Avg Loss: 0.12178856, Log Avg loss: 0.11272928, Global Avg Loss: 0.62152805, Time: 0.0209 Steps: 87340, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000061, Sample Num: 976, Cur Loss: 0.11607705, Cur Avg Loss: 0.12020710, Log Avg loss: 0.11214163, Global Avg Loss: 0.62146974, Time: 0.0210 Steps: 87350, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000071, Sample Num: 1136, Cur Loss: 0.19668892, Cur Avg Loss: 0.12267543, Log Avg loss: 0.13773228, Global Avg Loss: 0.62141436, Time: 0.0209 Steps: 87360, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000081, Sample Num: 1296, Cur Loss: 0.05285138, Cur Avg Loss: 0.12451112, Log Avg loss: 0.13754449, Global Avg Loss: 0.62135898, Time: 0.0209 Steps: 87370, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000091, Sample Num: 1456, Cur Loss: 0.15744875, Cur Avg Loss: 0.12769797, Log Avg loss: 0.15351146, Global Avg Loss: 0.62130544, Time: 0.0210 Steps: 87380, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000101, Sample Num: 1616, Cur Loss: 0.12170291, Cur Avg Loss: 0.12732757, Log Avg loss: 0.12395690, Global Avg Loss: 0.62124853, Time: 0.0209 Steps: 87390, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000111, Sample Num: 1776, Cur Loss: 0.45690984, Cur Avg Loss: 0.12892387, Log Avg loss: 0.14504652, Global Avg Loss: 0.62119404, Time: 0.0209 Steps: 87400, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000121, Sample Num: 1936, Cur Loss: 0.08657214, Cur Avg Loss: 0.12698690, Log Avg loss: 0.10548652, Global Avg Loss: 0.62113504, Time: 0.0210 Steps: 87410, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000131, Sample Num: 2096, Cur Loss: 0.12924375, Cur Avg Loss: 0.13362868, Log Avg loss: 0.21399432, Global Avg Loss: 0.62108847, Time: 0.0210 Steps: 87420, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000141, Sample Num: 2256, Cur Loss: 0.07497960, Cur Avg Loss: 0.13143018, Log Avg loss: 0.10262981, Global Avg Loss: 0.62102917, Time: 0.0210 Steps: 87430, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000151, Sample Num: 2416, Cur Loss: 0.14184222, Cur Avg Loss: 0.13065180, Log Avg loss: 0.11967661, Global Avg Loss: 0.62097184, Time: 0.0209 Steps: 87440, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000161, Sample Num: 2576, Cur Loss: 0.06303461, Cur Avg Loss: 0.13087675, Log Avg loss: 0.13427354, Global Avg Loss: 0.62091618, Time: 0.0209 Steps: 87450, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000171, Sample Num: 2736, Cur Loss: 0.13586064, Cur Avg Loss: 0.12986412, Log Avg loss: 0.11356066, Global Avg Loss: 0.62085817, Time: 0.0209 Steps: 87460, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000181, Sample Num: 2896, Cur Loss: 0.12373546, Cur Avg Loss: 0.12807836, Log Avg loss: 0.09754183, Global Avg Loss: 0.62079834, Time: 0.0210 Steps: 87470, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000191, Sample Num: 3056, Cur Loss: 0.31800920, Cur Avg Loss: 0.12994281, Log Avg loss: 0.16368942, Global Avg Loss: 0.62074609, Time: 0.0209 Steps: 87480, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000201, Sample Num: 3216, Cur Loss: 0.27011919, Cur Avg Loss: 0.13369465, Log Avg loss: 0.20535482, Global Avg Loss: 0.62069861, Time: 0.0210 Steps: 87490, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000211, Sample Num: 3376, Cur Loss: 0.11343981, Cur Avg Loss: 0.13909799, Log Avg loss: 0.24770514, Global Avg Loss: 0.62065598, Time: 0.0209 Steps: 87500, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000221, Sample Num: 3536, Cur Loss: 0.09470527, Cur Avg Loss: 0.13771848, Log Avg loss: 0.10861075, Global Avg Loss: 0.62059747, Time: 0.0210 Steps: 87510, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000231, Sample Num: 3696, Cur Loss: 0.06618637, Cur Avg Loss: 0.13950710, Log Avg loss: 0.17903555, Global Avg Loss: 0.62054702, Time: 0.0210 Steps: 87520, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000241, Sample Num: 3856, Cur Loss: 0.35455781, Cur Avg Loss: 0.14054808, Log Avg loss: 0.16459489, Global Avg Loss: 0.62049493, Time: 0.0209 Steps: 87530, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000251, Sample Num: 4016, Cur Loss: 0.13261005, Cur Avg Loss: 0.13933508, Log Avg loss: 0.11010170, Global Avg Loss: 0.62043662, Time: 0.0209 Steps: 87540, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000261, Sample Num: 4176, Cur Loss: 0.05095387, Cur Avg Loss: 0.13913202, Log Avg loss: 0.13403525, Global Avg Loss: 0.62038107, Time: 0.0210 Steps: 87550, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000271, Sample Num: 4336, Cur Loss: 0.39467490, Cur Avg Loss: 0.13991059, Log Avg loss: 0.16023112, Global Avg Loss: 0.62032851, Time: 0.0209 Steps: 87560, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000281, Sample Num: 4496, Cur Loss: 0.05458177, Cur Avg Loss: 0.13845287, Log Avg loss: 0.09894868, Global Avg Loss: 0.62026897, Time: 0.0209 Steps: 87570, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000291, Sample Num: 4656, Cur Loss: 0.23449844, Cur Avg Loss: 0.13866039, Log Avg loss: 0.14449179, Global Avg Loss: 0.62021465, Time: 0.0208 Steps: 87580, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000301, Sample Num: 4816, Cur Loss: 0.04033580, Cur Avg Loss: 0.13917878, Log Avg loss: 0.15426383, Global Avg Loss: 0.62016145, Time: 0.0208 Steps: 87590, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000311, Sample Num: 4976, Cur Loss: 0.05138656, Cur Avg Loss: 0.13852549, Log Avg loss: 0.11886157, Global Avg Loss: 0.62010423, Time: 0.0208 Steps: 87600, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000321, Sample Num: 5136, Cur Loss: 0.21392362, Cur Avg Loss: 0.13885831, Log Avg loss: 0.14920912, Global Avg Loss: 0.62005048, Time: 0.0209 Steps: 87610, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000331, Sample Num: 5296, Cur Loss: 0.10473566, Cur Avg Loss: 0.13855607, Log Avg loss: 0.12885398, Global Avg Loss: 0.61999442, Time: 0.0209 Steps: 87620, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000341, Sample Num: 5456, Cur Loss: 0.19314435, Cur Avg Loss: 0.13827694, Log Avg loss: 0.12903767, Global Avg Loss: 0.61993839, Time: 0.0208 Steps: 87630, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000351, Sample Num: 5616, Cur Loss: 0.05230837, Cur Avg Loss: 0.13812898, Log Avg loss: 0.13308380, Global Avg Loss: 0.61988284, Time: 0.0209 Steps: 87640, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000361, Sample Num: 5776, Cur Loss: 0.07751930, Cur Avg Loss: 0.13762565, Log Avg loss: 0.11995873, Global Avg Loss: 0.61982580, Time: 0.0209 Steps: 87650, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000371, Sample Num: 5936, Cur Loss: 0.16404748, Cur Avg Loss: 0.13815978, Log Avg loss: 0.15744192, Global Avg Loss: 0.61977306, Time: 0.0210 Steps: 87660, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000381, Sample Num: 6096, Cur Loss: 0.18931317, Cur Avg Loss: 0.13842976, Log Avg loss: 0.14844600, Global Avg Loss: 0.61971929, Time: 0.0209 Steps: 87670, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000391, Sample Num: 6256, Cur Loss: 0.37866974, Cur Avg Loss: 0.13880065, Log Avg loss: 0.15293141, Global Avg Loss: 0.61966606, Time: 0.0209 Steps: 87680, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000401, Sample Num: 6416, Cur Loss: 0.33363527, Cur Avg Loss: 0.13917835, Log Avg loss: 0.15394660, Global Avg Loss: 0.61961295, Time: 0.0208 Steps: 87690, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000411, Sample Num: 6576, Cur Loss: 0.03392172, Cur Avg Loss: 0.13968630, Log Avg loss: 0.16005493, Global Avg Loss: 0.61956055, Time: 0.0209 Steps: 87700, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000421, Sample Num: 6736, Cur Loss: 0.02311044, Cur Avg Loss: 0.13836289, Log Avg loss: 0.08397059, Global Avg Loss: 0.61949948, Time: 0.0208 Steps: 87710, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000431, Sample Num: 6896, Cur Loss: 0.13153349, Cur Avg Loss: 0.13769894, Log Avg loss: 0.10974690, Global Avg Loss: 0.61944137, Time: 0.0209 Steps: 87720, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000441, Sample Num: 7056, Cur Loss: 0.02297002, Cur Avg Loss: 0.13713399, Log Avg loss: 0.11278443, Global Avg Loss: 0.61938362, Time: 0.0208 Steps: 87730, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000451, Sample Num: 7216, Cur Loss: 0.01672851, Cur Avg Loss: 0.13763050, Log Avg loss: 0.15952672, Global Avg Loss: 0.61933121, Time: 0.0208 Steps: 87740, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000461, Sample Num: 7376, Cur Loss: 0.08604241, Cur Avg Loss: 0.13693774, Log Avg loss: 0.10569413, Global Avg Loss: 0.61927267, Time: 0.0208 Steps: 87750, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000471, Sample Num: 7536, Cur Loss: 0.04268110, Cur Avg Loss: 0.13709798, Log Avg loss: 0.14448496, Global Avg Loss: 0.61921857, Time: 0.0208 Steps: 87760, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000481, Sample Num: 7696, Cur Loss: 0.13608482, Cur Avg Loss: 0.13880755, Log Avg loss: 0.21932866, Global Avg Loss: 0.61917301, Time: 0.0208 Steps: 87770, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000491, Sample Num: 7856, Cur Loss: 0.06097192, Cur Avg Loss: 0.13746130, Log Avg loss: 0.07270633, Global Avg Loss: 0.61911076, Time: 0.0208 Steps: 87780, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000501, Sample Num: 8016, Cur Loss: 0.18879162, Cur Avg Loss: 0.13758152, Log Avg loss: 0.14348446, Global Avg Loss: 0.61905658, Time: 0.0209 Steps: 87790, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000511, Sample Num: 8176, Cur Loss: 0.22138649, Cur Avg Loss: 0.13797890, Log Avg loss: 0.15788750, Global Avg Loss: 0.61900406, Time: 0.0208 Steps: 87800, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000521, Sample Num: 8336, Cur Loss: 0.04573144, Cur Avg Loss: 0.13833212, Log Avg loss: 0.15638174, Global Avg Loss: 0.61895137, Time: 0.0211 Steps: 87810, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000531, Sample Num: 8496, Cur Loss: 0.13413848, Cur Avg Loss: 0.13794067, Log Avg loss: 0.11754649, Global Avg Loss: 0.61889428, Time: 0.0210 Steps: 87820, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000541, Sample Num: 8656, Cur Loss: 0.28393522, Cur Avg Loss: 0.13794431, Log Avg loss: 0.13813711, Global Avg Loss: 0.61883954, Time: 0.0210 Steps: 87830, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000551, Sample Num: 8816, Cur Loss: 0.21032155, Cur Avg Loss: 0.13714945, Log Avg loss: 0.09414772, Global Avg Loss: 0.61877981, Time: 0.0210 Steps: 87840, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000561, Sample Num: 8976, Cur Loss: 0.05158393, Cur Avg Loss: 0.13751363, Log Avg loss: 0.15758002, Global Avg Loss: 0.61872731, Time: 0.0210 Steps: 87850, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000571, Sample Num: 9136, Cur Loss: 0.16819394, Cur Avg Loss: 0.13788829, Log Avg loss: 0.15890673, Global Avg Loss: 0.61867497, Time: 0.0210 Steps: 87860, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000581, Sample Num: 9296, Cur Loss: 0.25111264, Cur Avg Loss: 0.13804147, Log Avg loss: 0.14678813, Global Avg Loss: 0.61862127, Time: 0.0210 Steps: 87870, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000591, Sample Num: 9456, Cur Loss: 0.03732473, Cur Avg Loss: 0.13804979, Log Avg loss: 0.13853309, Global Avg Loss: 0.61856664, Time: 0.0210 Steps: 87880, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000601, Sample Num: 9616, Cur Loss: 0.20143473, Cur Avg Loss: 0.13889582, Log Avg loss: 0.18889635, Global Avg Loss: 0.61851775, Time: 0.0211 Steps: 87890, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000611, Sample Num: 9776, Cur Loss: 0.03519918, Cur Avg Loss: 0.13883460, Log Avg loss: 0.13515494, Global Avg Loss: 0.61846276, Time: 0.0210 Steps: 87900, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000621, Sample Num: 9936, Cur Loss: 0.05885571, Cur Avg Loss: 0.13896316, Log Avg loss: 0.14681828, Global Avg Loss: 0.61840911, Time: 0.0210 Steps: 87910, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000631, Sample Num: 10096, Cur Loss: 0.21157184, Cur Avg Loss: 0.13924130, Log Avg loss: 0.15651371, Global Avg Loss: 0.61835658, Time: 0.0210 Steps: 87920, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000641, Sample Num: 10256, Cur Loss: 0.15932886, Cur Avg Loss: 0.13895230, Log Avg loss: 0.12071674, Global Avg Loss: 0.61829998, Time: 0.0210 Steps: 87930, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000651, Sample Num: 10416, Cur Loss: 0.10833665, Cur Avg Loss: 0.13907023, Log Avg loss: 0.14662932, Global Avg Loss: 0.61824634, Time: 0.0210 Steps: 87940, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000661, Sample Num: 10576, Cur Loss: 0.12226316, Cur Avg Loss: 0.13882155, Log Avg loss: 0.12263252, Global Avg Loss: 0.61818999, Time: 0.0210 Steps: 87950, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000671, Sample Num: 10736, Cur Loss: 0.08179936, Cur Avg Loss: 0.13872114, Log Avg loss: 0.13208405, Global Avg Loss: 0.61813473, Time: 0.0210 Steps: 87960, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000681, Sample Num: 10896, Cur Loss: 0.07496098, Cur Avg Loss: 0.13963936, Log Avg loss: 0.20125189, Global Avg Loss: 0.61808734, Time: 0.0210 Steps: 87970, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000691, Sample Num: 11056, Cur Loss: 0.07508208, Cur Avg Loss: 0.13935188, Log Avg loss: 0.11977426, Global Avg Loss: 0.61803070, Time: 0.0210 Steps: 87980, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000701, Sample Num: 11216, Cur Loss: 0.07975315, Cur Avg Loss: 0.13916878, Log Avg loss: 0.12651680, Global Avg Loss: 0.61797484, Time: 0.0210 Steps: 87990, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000711, Sample Num: 11376, Cur Loss: 0.13823661, Cur Avg Loss: 0.13883085, Log Avg loss: 0.11514174, Global Avg Loss: 0.61791770, Time: 0.0211 Steps: 88000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000721, Sample Num: 11536, Cur Loss: 0.30090711, Cur Avg Loss: 0.13902365, Log Avg loss: 0.15273204, Global Avg Loss: 0.61786484, Time: 0.0210 Steps: 88010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000731, Sample Num: 11696, Cur Loss: 0.28084326, Cur Avg Loss: 0.13913864, Log Avg loss: 0.14742925, Global Avg Loss: 0.61781140, Time: 0.0210 Steps: 88020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000741, Sample Num: 11856, Cur Loss: 0.27690825, Cur Avg Loss: 0.13915883, Log Avg loss: 0.14063464, Global Avg Loss: 0.61775719, Time: 0.0210 Steps: 88030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000751, Sample Num: 12016, Cur Loss: 0.13748363, Cur Avg Loss: 0.13939824, Log Avg loss: 0.15713874, Global Avg Loss: 0.61770487, Time: 0.0210 Steps: 88040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000761, Sample Num: 12176, Cur Loss: 0.13572468, Cur Avg Loss: 0.13929246, Log Avg loss: 0.13134807, Global Avg Loss: 0.61764964, Time: 0.0210 Steps: 88050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000771, Sample Num: 12336, Cur Loss: 0.07329442, Cur Avg Loss: 0.13945049, Log Avg loss: 0.15147679, Global Avg Loss: 0.61759670, Time: 0.0245 Steps: 88060, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000781, Sample Num: 12496, Cur Loss: 0.08798921, Cur Avg Loss: 0.13934448, Log Avg loss: 0.13117117, Global Avg Loss: 0.61754147, Time: 0.0209 Steps: 88070, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000791, Sample Num: 12656, Cur Loss: 0.14142987, Cur Avg Loss: 0.13928532, Log Avg loss: 0.13466452, Global Avg Loss: 0.61748664, Time: 0.0208 Steps: 88080, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000801, Sample Num: 12816, Cur Loss: 0.14296740, Cur Avg Loss: 0.13926398, Log Avg loss: 0.13757619, Global Avg Loss: 0.61743216, Time: 0.0208 Steps: 88090, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000811, Sample Num: 12976, Cur Loss: 0.05702934, Cur Avg Loss: 0.13901789, Log Avg loss: 0.11930611, Global Avg Loss: 0.61737562, Time: 0.0208 Steps: 88100, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000821, Sample Num: 13136, Cur Loss: 0.06853899, Cur Avg Loss: 0.13854681, Log Avg loss: 0.10034268, Global Avg Loss: 0.61731694, Time: 0.0208 Steps: 88110, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000831, Sample Num: 13296, Cur Loss: 0.09626681, Cur Avg Loss: 0.13843752, Log Avg loss: 0.12946422, Global Avg Loss: 0.61726158, Time: 0.0208 Steps: 88120, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000841, Sample Num: 13456, Cur Loss: 0.10086325, Cur Avg Loss: 0.13889513, Log Avg loss: 0.17692295, Global Avg Loss: 0.61721162, Time: 0.0208 Steps: 88130, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000851, Sample Num: 13616, Cur Loss: 0.05976782, Cur Avg Loss: 0.13908812, Log Avg loss: 0.15531857, Global Avg Loss: 0.61715921, Time: 0.0208 Steps: 88140, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000861, Sample Num: 13776, Cur Loss: 0.05955807, Cur Avg Loss: 0.13898229, Log Avg loss: 0.12997607, Global Avg Loss: 0.61710394, Time: 0.0208 Steps: 88150, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000871, Sample Num: 13936, Cur Loss: 0.04284209, Cur Avg Loss: 0.13862327, Log Avg loss: 0.10771154, Global Avg Loss: 0.61704616, Time: 0.0208 Steps: 88160, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000881, Sample Num: 14096, Cur Loss: 0.09238990, Cur Avg Loss: 0.13812936, Log Avg loss: 0.09510971, Global Avg Loss: 0.61698697, Time: 0.0208 Steps: 88170, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000891, Sample Num: 14256, Cur Loss: 0.38127819, Cur Avg Loss: 0.13841321, Log Avg loss: 0.16342064, Global Avg Loss: 0.61693553, Time: 0.0208 Steps: 88180, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000901, Sample Num: 14416, Cur Loss: 0.10355011, Cur Avg Loss: 0.13817773, Log Avg loss: 0.11719674, Global Avg Loss: 0.61687886, Time: 0.0208 Steps: 88190, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000911, Sample Num: 14576, Cur Loss: 0.23946047, Cur Avg Loss: 0.13894379, Log Avg loss: 0.20796521, Global Avg Loss: 0.61683250, Time: 0.0208 Steps: 88200, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000921, Sample Num: 14736, Cur Loss: 0.09126633, Cur Avg Loss: 0.13866956, Log Avg loss: 0.11368709, Global Avg Loss: 0.61677546, Time: 0.0208 Steps: 88210, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000931, Sample Num: 14896, Cur Loss: 0.18831736, Cur Avg Loss: 0.13900803, Log Avg loss: 0.17018142, Global Avg Loss: 0.61672484, Time: 0.0208 Steps: 88220, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000941, Sample Num: 15056, Cur Loss: 0.12319895, Cur Avg Loss: 0.13891470, Log Avg loss: 0.13022585, Global Avg Loss: 0.61666970, Time: 0.0208 Steps: 88230, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000951, Sample Num: 15216, Cur Loss: 0.15350862, Cur Avg Loss: 0.13937702, Log Avg loss: 0.18288110, Global Avg Loss: 0.61662054, Time: 0.0208 Steps: 88240, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000961, Sample Num: 15376, Cur Loss: 0.21051393, Cur Avg Loss: 0.13974331, Log Avg loss: 0.17457747, Global Avg Loss: 0.61657045, Time: 0.0208 Steps: 88250, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000971, Sample Num: 15536, Cur Loss: 0.27583683, Cur Avg Loss: 0.13980705, Log Avg loss: 0.14593280, Global Avg Loss: 0.61651713, Time: 0.0208 Steps: 88260, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000981, Sample Num: 15696, Cur Loss: 0.18151897, Cur Avg Loss: 0.14020029, Log Avg loss: 0.17838392, Global Avg Loss: 0.61646749, Time: 0.0208 Steps: 88270, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000991, Sample Num: 15856, Cur Loss: 0.07980935, Cur Avg Loss: 0.14044605, Log Avg loss: 0.16455516, Global Avg Loss: 0.61641630, Time: 0.0208 Steps: 88280, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001001, Sample Num: 16016, Cur Loss: 0.08038679, Cur Avg Loss: 0.14015399, Log Avg loss: 0.11121069, Global Avg Loss: 0.61635908, Time: 0.0208 Steps: 88290, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001011, Sample Num: 16176, Cur Loss: 0.12516479, Cur Avg Loss: 0.14049698, Log Avg loss: 0.17483013, Global Avg Loss: 0.61630907, Time: 0.0208 Steps: 88300, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001021, Sample Num: 16336, Cur Loss: 0.11986530, Cur Avg Loss: 0.14027526, Log Avg loss: 0.11785988, Global Avg Loss: 0.61625263, Time: 0.0208 Steps: 88310, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001031, Sample Num: 16496, Cur Loss: 0.12905739, Cur Avg Loss: 0.14028232, Log Avg loss: 0.14100311, Global Avg Loss: 0.61619882, Time: 0.0210 Steps: 88320, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001041, Sample Num: 16656, Cur Loss: 0.22899446, Cur Avg Loss: 0.14072763, Log Avg loss: 0.18663820, Global Avg Loss: 0.61615019, Time: 0.0208 Steps: 88330, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001051, Sample Num: 16816, Cur Loss: 0.27857420, Cur Avg Loss: 0.14085413, Log Avg loss: 0.15402356, Global Avg Loss: 0.61609788, Time: 0.0208 Steps: 88340, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001061, Sample Num: 16976, Cur Loss: 0.32663584, Cur Avg Loss: 0.14074988, Log Avg loss: 0.12979284, Global Avg Loss: 0.61604284, Time: 0.0209 Steps: 88350, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001071, Sample Num: 17136, Cur Loss: 0.16792834, Cur Avg Loss: 0.14051644, Log Avg loss: 0.11574867, Global Avg Loss: 0.61598622, Time: 0.0208 Steps: 88360, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001081, Sample Num: 17296, Cur Loss: 0.04998785, Cur Avg Loss: 0.14010959, Log Avg loss: 0.09653536, Global Avg Loss: 0.61592743, Time: 0.0208 Steps: 88370, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001091, Sample Num: 17456, Cur Loss: 0.10351469, Cur Avg Loss: 0.13998942, Log Avg loss: 0.12699949, Global Avg Loss: 0.61587211, Time: 0.0208 Steps: 88380, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001101, Sample Num: 17616, Cur Loss: 0.11720795, Cur Avg Loss: 0.13975378, Log Avg loss: 0.11404523, Global Avg Loss: 0.61581534, Time: 0.0208 Steps: 88390, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001111, Sample Num: 17776, Cur Loss: 0.15891869, Cur Avg Loss: 0.13966497, Log Avg loss: 0.12988774, Global Avg Loss: 0.61576037, Time: 0.0208 Steps: 88400, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001121, Sample Num: 17936, Cur Loss: 0.04214077, Cur Avg Loss: 0.13988178, Log Avg loss: 0.16396916, Global Avg Loss: 0.61570927, Time: 0.0208 Steps: 88410, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001131, Sample Num: 18096, Cur Loss: 0.14677575, Cur Avg Loss: 0.13996926, Log Avg loss: 0.14977531, Global Avg Loss: 0.61565657, Time: 0.0208 Steps: 88420, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001141, Sample Num: 18256, Cur Loss: 0.11300948, Cur Avg Loss: 0.13999057, Log Avg loss: 0.14240071, Global Avg Loss: 0.61560305, Time: 0.0208 Steps: 88430, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001151, Sample Num: 18416, Cur Loss: 0.03477994, Cur Avg Loss: 0.14002506, Log Avg loss: 0.14396077, Global Avg Loss: 0.61554973, Time: 0.0208 Steps: 88440, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001161, Sample Num: 18576, Cur Loss: 0.04579553, Cur Avg Loss: 0.13983544, Log Avg loss: 0.11800957, Global Avg Loss: 0.61549347, Time: 0.0208 Steps: 88450, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001171, Sample Num: 18736, Cur Loss: 0.15767792, Cur Avg Loss: 0.13964157, Log Avg loss: 0.11713351, Global Avg Loss: 0.61543714, Time: 0.0208 Steps: 88460, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001181, Sample Num: 18896, Cur Loss: 0.06927276, Cur Avg Loss: 0.13969427, Log Avg loss: 0.14586601, Global Avg Loss: 0.61538406, Time: 0.0208 Steps: 88470, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001191, Sample Num: 19056, Cur Loss: 0.15925616, Cur Avg Loss: 0.13942204, Log Avg loss: 0.10727117, Global Avg Loss: 0.61532663, Time: 0.0208 Steps: 88480, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001201, Sample Num: 19216, Cur Loss: 0.18670186, Cur Avg Loss: 0.14041092, Log Avg loss: 0.25818648, Global Avg Loss: 0.61528627, Time: 0.0208 Steps: 88490, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001211, Sample Num: 19376, Cur Loss: 0.09605685, Cur Avg Loss: 0.14122110, Log Avg loss: 0.23852358, Global Avg Loss: 0.61524370, Time: 0.0208 Steps: 88500, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001221, Sample Num: 19536, Cur Loss: 0.11580785, Cur Avg Loss: 0.14107449, Log Avg loss: 0.12332056, Global Avg Loss: 0.61518812, Time: 0.0208 Steps: 88510, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001231, Sample Num: 19696, Cur Loss: 0.12037112, Cur Avg Loss: 0.14155238, Log Avg loss: 0.19990305, Global Avg Loss: 0.61514121, Time: 0.0208 Steps: 88520, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001241, Sample Num: 19856, Cur Loss: 0.10230385, Cur Avg Loss: 0.14151644, Log Avg loss: 0.13709168, Global Avg Loss: 0.61508721, Time: 0.0208 Steps: 88530, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001251, Sample Num: 20016, Cur Loss: 0.15656750, Cur Avg Loss: 0.14161285, Log Avg loss: 0.15357764, Global Avg Loss: 0.61503509, Time: 0.0208 Steps: 88540, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001261, Sample Num: 20176, Cur Loss: 0.08156707, Cur Avg Loss: 0.14134063, Log Avg loss: 0.10728506, Global Avg Loss: 0.61497775, Time: 0.0208 Steps: 88550, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001271, Sample Num: 20336, Cur Loss: 0.13456260, Cur Avg Loss: 0.14162742, Log Avg loss: 0.17779185, Global Avg Loss: 0.61492838, Time: 0.0208 Steps: 88560, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001281, Sample Num: 20496, Cur Loss: 0.05090801, Cur Avg Loss: 0.14160808, Log Avg loss: 0.13915019, Global Avg Loss: 0.61487466, Time: 0.0245 Steps: 88570, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001291, Sample Num: 20656, Cur Loss: 0.03045424, Cur Avg Loss: 0.14149377, Log Avg loss: 0.12685046, Global Avg Loss: 0.61481957, Time: 0.0208 Steps: 88580, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001301, Sample Num: 20816, Cur Loss: 0.17344204, Cur Avg Loss: 0.14158601, Log Avg loss: 0.15349399, Global Avg Loss: 0.61476749, Time: 0.0208 Steps: 88590, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001311, Sample Num: 20976, Cur Loss: 0.23762906, Cur Avg Loss: 0.14182031, Log Avg loss: 0.17230341, Global Avg Loss: 0.61471755, Time: 0.0208 Steps: 88600, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001321, Sample Num: 21136, Cur Loss: 0.06147649, Cur Avg Loss: 0.14188587, Log Avg loss: 0.15048023, Global Avg Loss: 0.61466516, Time: 0.0208 Steps: 88610, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001331, Sample Num: 21296, Cur Loss: 0.12542105, Cur Avg Loss: 0.14192349, Log Avg loss: 0.14689388, Global Avg Loss: 0.61461238, Time: 0.0208 Steps: 88620, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001341, Sample Num: 21456, Cur Loss: 0.33738321, Cur Avg Loss: 0.14189560, Log Avg loss: 0.13818272, Global Avg Loss: 0.61455862, Time: 0.0208 Steps: 88630, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001351, Sample Num: 21616, Cur Loss: 0.13388607, Cur Avg Loss: 0.14158233, Log Avg loss: 0.09957373, Global Avg Loss: 0.61450053, Time: 0.0208 Steps: 88640, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001361, Sample Num: 21776, Cur Loss: 0.06833903, Cur Avg Loss: 0.14170517, Log Avg loss: 0.15829955, Global Avg Loss: 0.61444906, Time: 0.0208 Steps: 88650, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001371, Sample Num: 21936, Cur Loss: 0.07534122, Cur Avg Loss: 0.14174469, Log Avg loss: 0.14712373, Global Avg Loss: 0.61439635, Time: 0.0208 Steps: 88660, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001381, Sample Num: 22096, Cur Loss: 0.34015971, Cur Avg Loss: 0.14188950, Log Avg loss: 0.16174342, Global Avg Loss: 0.61434531, Time: 0.0208 Steps: 88670, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001391, Sample Num: 22256, Cur Loss: 0.03391886, Cur Avg Loss: 0.14174767, Log Avg loss: 0.12216023, Global Avg Loss: 0.61428980, Time: 0.0208 Steps: 88680, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001401, Sample Num: 22416, Cur Loss: 0.10828145, Cur Avg Loss: 0.14140601, Log Avg loss: 0.09388182, Global Avg Loss: 0.61423113, Time: 0.0208 Steps: 88690, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001411, Sample Num: 22576, Cur Loss: 0.02961549, Cur Avg Loss: 0.14121518, Log Avg loss: 0.11447992, Global Avg Loss: 0.61417479, Time: 0.0208 Steps: 88700, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001421, Sample Num: 22736, Cur Loss: 0.06969634, Cur Avg Loss: 0.14128138, Log Avg loss: 0.15062244, Global Avg Loss: 0.61412253, Time: 0.0208 Steps: 88710, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001431, Sample Num: 22896, Cur Loss: 0.07787939, Cur Avg Loss: 0.14091271, Log Avg loss: 0.08852401, Global Avg Loss: 0.61406329, Time: 0.0208 Steps: 88720, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001441, Sample Num: 23056, Cur Loss: 0.08815627, Cur Avg Loss: 0.14086194, Log Avg loss: 0.13359681, Global Avg Loss: 0.61400914, Time: 0.0208 Steps: 88730, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001451, Sample Num: 23216, Cur Loss: 0.18027055, Cur Avg Loss: 0.14109337, Log Avg loss: 0.17444256, Global Avg Loss: 0.61395960, Time: 0.0208 Steps: 88740, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001461, Sample Num: 23376, Cur Loss: 0.07365253, Cur Avg Loss: 0.14103582, Log Avg loss: 0.13268487, Global Avg Loss: 0.61390538, Time: 0.0208 Steps: 88750, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001471, Sample Num: 23536, Cur Loss: 0.10466683, Cur Avg Loss: 0.14109577, Log Avg loss: 0.14985549, Global Avg Loss: 0.61385310, Time: 0.0208 Steps: 88760, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001481, Sample Num: 23696, Cur Loss: 0.12513025, Cur Avg Loss: 0.14108822, Log Avg loss: 0.13997710, Global Avg Loss: 0.61379971, Time: 0.0208 Steps: 88770, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001491, Sample Num: 23856, Cur Loss: 0.14512505, Cur Avg Loss: 0.14115706, Log Avg loss: 0.15135205, Global Avg Loss: 0.61374762, Time: 0.0208 Steps: 88780, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001501, Sample Num: 24016, Cur Loss: 0.18581757, Cur Avg Loss: 0.14127969, Log Avg loss: 0.15956445, Global Avg Loss: 0.61369647, Time: 0.0208 Steps: 88790, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001511, Sample Num: 24176, Cur Loss: 0.21789250, Cur Avg Loss: 0.14166797, Log Avg loss: 0.19994869, Global Avg Loss: 0.61364988, Time: 0.0208 Steps: 88800, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001521, Sample Num: 24336, Cur Loss: 0.14533833, Cur Avg Loss: 0.14177752, Log Avg loss: 0.15833074, Global Avg Loss: 0.61359861, Time: 0.0208 Steps: 88810, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001531, Sample Num: 24496, Cur Loss: 0.13899127, Cur Avg Loss: 0.14206997, Log Avg loss: 0.18655078, Global Avg Loss: 0.61355053, Time: 0.0208 Steps: 88820, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001541, Sample Num: 24656, Cur Loss: 0.30899271, Cur Avg Loss: 0.14225211, Log Avg loss: 0.17013848, Global Avg Loss: 0.61350061, Time: 0.0246 Steps: 88830, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001551, Sample Num: 24816, Cur Loss: 0.06978391, Cur Avg Loss: 0.14253838, Log Avg loss: 0.18665174, Global Avg Loss: 0.61345256, Time: 0.0208 Steps: 88840, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001561, Sample Num: 24976, Cur Loss: 0.14689612, Cur Avg Loss: 0.14282431, Log Avg loss: 0.18717291, Global Avg Loss: 0.61340459, Time: 0.0208 Steps: 88850, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001571, Sample Num: 25136, Cur Loss: 0.14101741, Cur Avg Loss: 0.14294495, Log Avg loss: 0.16177664, Global Avg Loss: 0.61335376, Time: 0.0208 Steps: 88860, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001581, Sample Num: 25296, Cur Loss: 0.07939363, Cur Avg Loss: 0.14279739, Log Avg loss: 0.11961529, Global Avg Loss: 0.61329821, Time: 0.0208 Steps: 88870, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001591, Sample Num: 25456, Cur Loss: 0.08543497, Cur Avg Loss: 0.14252590, Log Avg loss: 0.09960408, Global Avg Loss: 0.61324041, Time: 0.0208 Steps: 88880, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001601, Sample Num: 25616, Cur Loss: 0.22950467, Cur Avg Loss: 0.14240619, Log Avg loss: 0.12335968, Global Avg Loss: 0.61318530, Time: 0.0208 Steps: 88890, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001611, Sample Num: 25776, Cur Loss: 0.03301670, Cur Avg Loss: 0.14230961, Log Avg loss: 0.12684740, Global Avg Loss: 0.61313059, Time: 0.0208 Steps: 88900, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001621, Sample Num: 25936, Cur Loss: 0.08249360, Cur Avg Loss: 0.14233211, Log Avg loss: 0.14595654, Global Avg Loss: 0.61307805, Time: 0.0208 Steps: 88910, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001631, Sample Num: 26096, Cur Loss: 0.04961420, Cur Avg Loss: 0.14251286, Log Avg loss: 0.17181320, Global Avg Loss: 0.61302842, Time: 0.0208 Steps: 88920, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001641, Sample Num: 26256, Cur Loss: 0.22388175, Cur Avg Loss: 0.14237040, Log Avg loss: 0.11913515, Global Avg Loss: 0.61297289, Time: 0.0208 Steps: 88930, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001651, Sample Num: 26416, Cur Loss: 0.16730398, Cur Avg Loss: 0.14254760, Log Avg loss: 0.17162536, Global Avg Loss: 0.61292326, Time: 0.0208 Steps: 88940, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001661, Sample Num: 26576, Cur Loss: 0.10212589, Cur Avg Loss: 0.14271317, Log Avg loss: 0.17004852, Global Avg Loss: 0.61287347, Time: 0.0208 Steps: 88950, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001671, Sample Num: 26736, Cur Loss: 0.13197508, Cur Avg Loss: 0.14284656, Log Avg loss: 0.16500308, Global Avg Loss: 0.61282313, Time: 0.0208 Steps: 88960, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001681, Sample Num: 26896, Cur Loss: 0.15366599, Cur Avg Loss: 0.14283732, Log Avg loss: 0.14129305, Global Avg Loss: 0.61277013, Time: 0.0208 Steps: 88970, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001691, Sample Num: 27056, Cur Loss: 0.16926531, Cur Avg Loss: 0.14286392, Log Avg loss: 0.14733637, Global Avg Loss: 0.61271782, Time: 0.0208 Steps: 88980, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001701, Sample Num: 27216, Cur Loss: 0.13895503, Cur Avg Loss: 0.14279555, Log Avg loss: 0.13123420, Global Avg Loss: 0.61266372, Time: 0.0208 Steps: 88990, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001711, Sample Num: 27376, Cur Loss: 0.08873241, Cur Avg Loss: 0.14258264, Log Avg loss: 0.10636556, Global Avg Loss: 0.61260683, Time: 0.0208 Steps: 89000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001721, Sample Num: 27536, Cur Loss: 0.24040565, Cur Avg Loss: 0.14260323, Log Avg loss: 0.14612647, Global Avg Loss: 0.61255442, Time: 0.0208 Steps: 89010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001731, Sample Num: 27696, Cur Loss: 0.17214963, Cur Avg Loss: 0.14279375, Log Avg loss: 0.17558262, Global Avg Loss: 0.61250533, Time: 0.0208 Steps: 89020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001741, Sample Num: 27856, Cur Loss: 0.23360223, Cur Avg Loss: 0.14276753, Log Avg loss: 0.13822797, Global Avg Loss: 0.61245206, Time: 0.0208 Steps: 89030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001751, Sample Num: 28016, Cur Loss: 0.04373582, Cur Avg Loss: 0.14249965, Log Avg loss: 0.09586298, Global Avg Loss: 0.61239404, Time: 0.0208 Steps: 89040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001761, Sample Num: 28176, Cur Loss: 0.08312750, Cur Avg Loss: 0.14258388, Log Avg loss: 0.15733115, Global Avg Loss: 0.61234294, Time: 0.0208 Steps: 89050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001771, Sample Num: 28336, Cur Loss: 0.30739853, Cur Avg Loss: 0.14262128, Log Avg loss: 0.14920760, Global Avg Loss: 0.61229094, Time: 0.0209 Steps: 89060, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001781, Sample Num: 28496, Cur Loss: 0.25607330, Cur Avg Loss: 0.14274685, Log Avg loss: 0.16498660, Global Avg Loss: 0.61224072, Time: 0.0209 Steps: 89070, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001791, Sample Num: 28656, Cur Loss: 0.11263438, Cur Avg Loss: 0.14280959, Log Avg loss: 0.15398199, Global Avg Loss: 0.61218928, Time: 0.0209 Steps: 89080, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001801, Sample Num: 28816, Cur Loss: 0.09341969, Cur Avg Loss: 0.14284596, Log Avg loss: 0.14936093, Global Avg Loss: 0.61213733, Time: 0.0210 Steps: 89090, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001811, Sample Num: 28976, Cur Loss: 0.09489121, Cur Avg Loss: 0.14271506, Log Avg loss: 0.11913977, Global Avg Loss: 0.61208200, Time: 0.0209 Steps: 89100, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001821, Sample Num: 29136, Cur Loss: 0.32961953, Cur Avg Loss: 0.14298907, Log Avg loss: 0.19261226, Global Avg Loss: 0.61203492, Time: 0.0209 Steps: 89110, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001831, Sample Num: 29296, Cur Loss: 0.02586396, Cur Avg Loss: 0.14313405, Log Avg loss: 0.16953556, Global Avg Loss: 0.61198527, Time: 0.0209 Steps: 89120, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001841, Sample Num: 29456, Cur Loss: 0.28582093, Cur Avg Loss: 0.14298445, Log Avg loss: 0.11559182, Global Avg Loss: 0.61192958, Time: 0.0209 Steps: 89130, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001851, Sample Num: 29616, Cur Loss: 0.16658345, Cur Avg Loss: 0.14313004, Log Avg loss: 0.16993395, Global Avg Loss: 0.61187999, Time: 0.0209 Steps: 89140, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001861, Sample Num: 29776, Cur Loss: 0.16219546, Cur Avg Loss: 0.14294852, Log Avg loss: 0.10934879, Global Avg Loss: 0.61182362, Time: 0.0209 Steps: 89150, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001871, Sample Num: 29936, Cur Loss: 0.08871979, Cur Avg Loss: 0.14269782, Log Avg loss: 0.09604164, Global Avg Loss: 0.61176577, Time: 0.0209 Steps: 89160, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001881, Sample Num: 30096, Cur Loss: 0.10593884, Cur Avg Loss: 0.14283232, Log Avg loss: 0.16799704, Global Avg Loss: 0.61171601, Time: 0.0209 Steps: 89170, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001891, Sample Num: 30256, Cur Loss: 0.13944018, Cur Avg Loss: 0.14291199, Log Avg loss: 0.15789959, Global Avg Loss: 0.61166512, Time: 0.0209 Steps: 89180, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001901, Sample Num: 30416, Cur Loss: 0.11784523, Cur Avg Loss: 0.14307388, Log Avg loss: 0.17368679, Global Avg Loss: 0.61161601, Time: 0.0209 Steps: 89190, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001911, Sample Num: 30576, Cur Loss: 0.25070146, Cur Avg Loss: 0.14302420, Log Avg loss: 0.13358012, Global Avg Loss: 0.61156242, Time: 0.0209 Steps: 89200, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001921, Sample Num: 30736, Cur Loss: 0.29977474, Cur Avg Loss: 0.14310975, Log Avg loss: 0.15945844, Global Avg Loss: 0.61151174, Time: 0.0209 Steps: 89210, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001931, Sample Num: 30896, Cur Loss: 0.18948418, Cur Avg Loss: 0.14359892, Log Avg loss: 0.23756819, Global Avg Loss: 0.61146983, Time: 0.0214 Steps: 89220, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001941, Sample Num: 31056, Cur Loss: 0.10683236, Cur Avg Loss: 0.14395953, Log Avg loss: 0.21359229, Global Avg Loss: 0.61142524, Time: 0.0209 Steps: 89230, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001951, Sample Num: 31216, Cur Loss: 0.05663147, Cur Avg Loss: 0.14406266, Log Avg loss: 0.16408119, Global Avg Loss: 0.61137511, Time: 0.0209 Steps: 89240, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001961, Sample Num: 31376, Cur Loss: 0.13255897, Cur Avg Loss: 0.14399149, Log Avg loss: 0.13010705, Global Avg Loss: 0.61132119, Time: 0.0208 Steps: 89250, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001971, Sample Num: 31536, Cur Loss: 0.06420850, Cur Avg Loss: 0.14401834, Log Avg loss: 0.14928194, Global Avg Loss: 0.61126943, Time: 0.0209 Steps: 89260, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001981, Sample Num: 31696, Cur Loss: 0.15381983, Cur Avg Loss: 0.14414563, Log Avg loss: 0.16923457, Global Avg Loss: 0.61121991, Time: 0.0209 Steps: 89270, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001991, Sample Num: 31856, Cur Loss: 0.46219429, Cur Avg Loss: 0.14425155, Log Avg loss: 0.16523419, Global Avg Loss: 0.61116996, Time: 0.0209 Steps: 89280, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002001, Sample Num: 32016, Cur Loss: 0.13810185, Cur Avg Loss: 0.14431541, Log Avg loss: 0.15702983, Global Avg Loss: 0.61111909, Time: 0.0209 Steps: 89290, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002011, Sample Num: 32176, Cur Loss: 0.09972663, Cur Avg Loss: 0.14436702, Log Avg loss: 0.15469562, Global Avg Loss: 0.61106798, Time: 0.0209 Steps: 89300, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002021, Sample Num: 32336, Cur Loss: 0.04680905, Cur Avg Loss: 0.14453242, Log Avg loss: 0.17779388, Global Avg Loss: 0.61101947, Time: 0.0209 Steps: 89310, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002031, Sample Num: 32496, Cur Loss: 0.03345763, Cur Avg Loss: 0.14442014, Log Avg loss: 0.12172752, Global Avg Loss: 0.61096469, Time: 0.0209 Steps: 89320, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002041, Sample Num: 32656, Cur Loss: 0.10360652, Cur Avg Loss: 0.14473926, Log Avg loss: 0.20955358, Global Avg Loss: 0.61091975, Time: 0.0209 Steps: 89330, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002051, Sample Num: 32816, Cur Loss: 0.17874172, Cur Avg Loss: 0.14468324, Log Avg loss: 0.13324915, Global Avg Loss: 0.61086629, Time: 0.0246 Steps: 89340, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002061, Sample Num: 32976, Cur Loss: 0.06181385, Cur Avg Loss: 0.14479799, Log Avg loss: 0.16833378, Global Avg Loss: 0.61081676, Time: 0.0209 Steps: 89350, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002071, Sample Num: 33136, Cur Loss: 0.09866025, Cur Avg Loss: 0.14465148, Log Avg loss: 0.11445597, Global Avg Loss: 0.61076121, Time: 0.0209 Steps: 89360, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002081, Sample Num: 33296, Cur Loss: 0.12659825, Cur Avg Loss: 0.14450930, Log Avg loss: 0.11506339, Global Avg Loss: 0.61070575, Time: 0.0208 Steps: 89370, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002091, Sample Num: 33456, Cur Loss: 0.19222105, Cur Avg Loss: 0.14439272, Log Avg loss: 0.12013196, Global Avg Loss: 0.61065086, Time: 0.0209 Steps: 89380, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002101, Sample Num: 33616, Cur Loss: 0.17568675, Cur Avg Loss: 0.14456056, Log Avg loss: 0.17965704, Global Avg Loss: 0.61060265, Time: 0.0209 Steps: 89390, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002111, Sample Num: 33776, Cur Loss: 0.13652791, Cur Avg Loss: 0.14440918, Log Avg loss: 0.11260424, Global Avg Loss: 0.61054694, Time: 0.0209 Steps: 89400, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002121, Sample Num: 33936, Cur Loss: 0.08352163, Cur Avg Loss: 0.14426110, Log Avg loss: 0.11300012, Global Avg Loss: 0.61049129, Time: 0.0208 Steps: 89410, Updated lr: 0.000016 ***** Running evaluation checkpoint-89418 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-89418 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.653174, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.233461, "eval_total_loss": 164.123357, "eval_mae": 0.312959, "eval_mse": 0.233552, "eval_r2": 0.851539, "eval_sp_statistic": 0.930633, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.935225, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.223558, "test_total_loss": 112.226174, "test_mae": 0.37233, "test_mse": 0.223628, "test_r2": 0.855668, "test_sp_statistic": 0.918056, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.945825, "test_ps_pvalue": 0.0, "lr": 1.6151730678046468e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.6104481506082485, "train_cur_epoch_loss": 307.0038802009076, "train_cur_epoch_avg_loss": 0.1442009770788669, "train_cur_epoch_time": 44.65317368507385, "train_cur_epoch_avg_time": 0.02097377815174911, "epoch": 42, "step": 89418} ################################################## Training, Epoch: 0043, Batch: 000002, Sample Num: 32, Cur Loss: 0.05401443, Cur Avg Loss: 0.03943976, Log Avg loss: 0.11049705, Global Avg Loss: 0.61043538, Time: 0.0247 Steps: 89420, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000012, Sample Num: 192, Cur Loss: 0.34366041, Cur Avg Loss: 0.14045368, Log Avg loss: 0.16065647, Global Avg Loss: 0.61038509, Time: 0.0209 Steps: 89430, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000022, Sample Num: 352, Cur Loss: 0.14468081, Cur Avg Loss: 0.14237293, Log Avg loss: 0.14467602, Global Avg Loss: 0.61033302, Time: 0.0210 Steps: 89440, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000032, Sample Num: 512, Cur Loss: 0.29570103, Cur Avg Loss: 0.13780858, Log Avg loss: 0.12776700, Global Avg Loss: 0.61027907, Time: 0.0209 Steps: 89450, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000042, Sample Num: 672, Cur Loss: 0.05912065, Cur Avg Loss: 0.12930474, Log Avg loss: 0.10209247, Global Avg Loss: 0.61022226, Time: 0.0209 Steps: 89460, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000052, Sample Num: 832, Cur Loss: 0.11461687, Cur Avg Loss: 0.12204645, Log Avg loss: 0.09156160, Global Avg Loss: 0.61016429, Time: 0.0211 Steps: 89470, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000062, Sample Num: 992, Cur Loss: 0.11907073, Cur Avg Loss: 0.12646029, Log Avg loss: 0.14941229, Global Avg Loss: 0.61011280, Time: 0.0209 Steps: 89480, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000072, Sample Num: 1152, Cur Loss: 0.06922327, Cur Avg Loss: 0.13148795, Log Avg loss: 0.16265946, Global Avg Loss: 0.61006280, Time: 0.0209 Steps: 89490, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000082, Sample Num: 1312, Cur Loss: 0.24887788, Cur Avg Loss: 0.13576378, Log Avg loss: 0.16654976, Global Avg Loss: 0.61001324, Time: 0.0209 Steps: 89500, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000092, Sample Num: 1472, Cur Loss: 0.40066147, Cur Avg Loss: 0.13472279, Log Avg loss: 0.12618662, Global Avg Loss: 0.60995919, Time: 0.0209 Steps: 89510, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000102, Sample Num: 1632, Cur Loss: 0.11222342, Cur Avg Loss: 0.13661539, Log Avg loss: 0.15402736, Global Avg Loss: 0.60990826, Time: 0.0209 Steps: 89520, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000112, Sample Num: 1792, Cur Loss: 0.17444594, Cur Avg Loss: 0.13667703, Log Avg loss: 0.13730577, Global Avg Loss: 0.60985547, Time: 0.0209 Steps: 89530, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000122, Sample Num: 1952, Cur Loss: 0.04816580, Cur Avg Loss: 0.13707239, Log Avg loss: 0.14150031, Global Avg Loss: 0.60980317, Time: 0.0209 Steps: 89540, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000132, Sample Num: 2112, Cur Loss: 0.05573700, Cur Avg Loss: 0.13568072, Log Avg loss: 0.11870242, Global Avg Loss: 0.60974833, Time: 0.0209 Steps: 89550, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000142, Sample Num: 2272, Cur Loss: 0.03898990, Cur Avg Loss: 0.14097350, Log Avg loss: 0.21083820, Global Avg Loss: 0.60970378, Time: 0.0210 Steps: 89560, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000152, Sample Num: 2432, Cur Loss: 0.11957218, Cur Avg Loss: 0.14026485, Log Avg loss: 0.13020196, Global Avg Loss: 0.60965025, Time: 0.0210 Steps: 89570, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000162, Sample Num: 2592, Cur Loss: 0.10374182, Cur Avg Loss: 0.14050943, Log Avg loss: 0.14422709, Global Avg Loss: 0.60959829, Time: 0.0209 Steps: 89580, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000172, Sample Num: 2752, Cur Loss: 0.31060755, Cur Avg Loss: 0.14033583, Log Avg loss: 0.13752355, Global Avg Loss: 0.60954560, Time: 0.0210 Steps: 89590, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000182, Sample Num: 2912, Cur Loss: 0.11202525, Cur Avg Loss: 0.14018955, Log Avg loss: 0.13767352, Global Avg Loss: 0.60949294, Time: 0.0211 Steps: 89600, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000192, Sample Num: 3072, Cur Loss: 0.06999188, Cur Avg Loss: 0.14279926, Log Avg loss: 0.19029590, Global Avg Loss: 0.60944616, Time: 0.0210 Steps: 89610, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000202, Sample Num: 3232, Cur Loss: 0.10608427, Cur Avg Loss: 0.14039810, Log Avg loss: 0.09429579, Global Avg Loss: 0.60938868, Time: 0.0210 Steps: 89620, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000212, Sample Num: 3392, Cur Loss: 0.07505819, Cur Avg Loss: 0.14123080, Log Avg loss: 0.15805133, Global Avg Loss: 0.60933832, Time: 0.0210 Steps: 89630, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000222, Sample Num: 3552, Cur Loss: 0.04501245, Cur Avg Loss: 0.13962598, Log Avg loss: 0.10560388, Global Avg Loss: 0.60928213, Time: 0.0209 Steps: 89640, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000232, Sample Num: 3712, Cur Loss: 0.08906760, Cur Avg Loss: 0.13909700, Log Avg loss: 0.12735361, Global Avg Loss: 0.60922837, Time: 0.0209 Steps: 89650, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000242, Sample Num: 3872, Cur Loss: 0.11014940, Cur Avg Loss: 0.13800183, Log Avg loss: 0.11259388, Global Avg Loss: 0.60917298, Time: 0.0209 Steps: 89660, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000252, Sample Num: 4032, Cur Loss: 0.16378483, Cur Avg Loss: 0.13852501, Log Avg loss: 0.15118605, Global Avg Loss: 0.60912190, Time: 0.0210 Steps: 89670, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000262, Sample Num: 4192, Cur Loss: 0.15168488, Cur Avg Loss: 0.14005759, Log Avg loss: 0.17867851, Global Avg Loss: 0.60907391, Time: 0.0245 Steps: 89680, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000272, Sample Num: 4352, Cur Loss: 0.24078566, Cur Avg Loss: 0.14076354, Log Avg loss: 0.15925948, Global Avg Loss: 0.60902375, Time: 0.0209 Steps: 89690, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000282, Sample Num: 4512, Cur Loss: 0.08818822, Cur Avg Loss: 0.13926056, Log Avg loss: 0.09837954, Global Avg Loss: 0.60896683, Time: 0.0209 Steps: 89700, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000292, Sample Num: 4672, Cur Loss: 0.07088945, Cur Avg Loss: 0.13819029, Log Avg loss: 0.10800866, Global Avg Loss: 0.60891098, Time: 0.0209 Steps: 89710, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000302, Sample Num: 4832, Cur Loss: 0.19546531, Cur Avg Loss: 0.13957352, Log Avg loss: 0.17996382, Global Avg Loss: 0.60886317, Time: 0.0210 Steps: 89720, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000312, Sample Num: 4992, Cur Loss: 0.05062307, Cur Avg Loss: 0.14095331, Log Avg loss: 0.18262293, Global Avg Loss: 0.60881567, Time: 0.0209 Steps: 89730, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000322, Sample Num: 5152, Cur Loss: 0.23997363, Cur Avg Loss: 0.14199775, Log Avg loss: 0.17458433, Global Avg Loss: 0.60876728, Time: 0.0209 Steps: 89740, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000332, Sample Num: 5312, Cur Loss: 0.10470018, Cur Avg Loss: 0.14157259, Log Avg loss: 0.12788250, Global Avg Loss: 0.60871370, Time: 0.0208 Steps: 89750, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000342, Sample Num: 5472, Cur Loss: 0.18480992, Cur Avg Loss: 0.14125891, Log Avg loss: 0.13084461, Global Avg Loss: 0.60866046, Time: 0.0209 Steps: 89760, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000352, Sample Num: 5632, Cur Loss: 0.07634075, Cur Avg Loss: 0.14066648, Log Avg loss: 0.12040535, Global Avg Loss: 0.60860607, Time: 0.0209 Steps: 89770, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000362, Sample Num: 5792, Cur Loss: 0.04423258, Cur Avg Loss: 0.14001601, Log Avg loss: 0.11711960, Global Avg Loss: 0.60855133, Time: 0.0209 Steps: 89780, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000372, Sample Num: 5952, Cur Loss: 0.16609353, Cur Avg Loss: 0.13997699, Log Avg loss: 0.13856454, Global Avg Loss: 0.60849899, Time: 0.0209 Steps: 89790, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000382, Sample Num: 6112, Cur Loss: 0.23145284, Cur Avg Loss: 0.13963203, Log Avg loss: 0.12679927, Global Avg Loss: 0.60844535, Time: 0.0209 Steps: 89800, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000392, Sample Num: 6272, Cur Loss: 0.05584971, Cur Avg Loss: 0.13889064, Log Avg loss: 0.11056954, Global Avg Loss: 0.60838991, Time: 0.0209 Steps: 89810, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000402, Sample Num: 6432, Cur Loss: 0.09956232, Cur Avg Loss: 0.13790617, Log Avg loss: 0.09931502, Global Avg Loss: 0.60833323, Time: 0.0209 Steps: 89820, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000412, Sample Num: 6592, Cur Loss: 0.10309680, Cur Avg Loss: 0.13726684, Log Avg loss: 0.11156586, Global Avg Loss: 0.60827793, Time: 0.0209 Steps: 89830, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000422, Sample Num: 6752, Cur Loss: 0.03144622, Cur Avg Loss: 0.13628779, Log Avg loss: 0.09595083, Global Avg Loss: 0.60822091, Time: 0.0209 Steps: 89840, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000432, Sample Num: 6912, Cur Loss: 0.18402764, Cur Avg Loss: 0.13622407, Log Avg loss: 0.13353522, Global Avg Loss: 0.60816807, Time: 0.0209 Steps: 89850, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000442, Sample Num: 7072, Cur Loss: 0.27708378, Cur Avg Loss: 0.13750755, Log Avg loss: 0.19295397, Global Avg Loss: 0.60812187, Time: 0.0209 Steps: 89860, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000452, Sample Num: 7232, Cur Loss: 0.15039049, Cur Avg Loss: 0.13683812, Log Avg loss: 0.10724935, Global Avg Loss: 0.60806614, Time: 0.0209 Steps: 89870, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000462, Sample Num: 7392, Cur Loss: 0.18396461, Cur Avg Loss: 0.14066373, Log Avg loss: 0.31358125, Global Avg Loss: 0.60803337, Time: 0.0209 Steps: 89880, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000472, Sample Num: 7552, Cur Loss: 0.09361242, Cur Avg Loss: 0.14157050, Log Avg loss: 0.18346315, Global Avg Loss: 0.60798614, Time: 0.0209 Steps: 89890, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000482, Sample Num: 7712, Cur Loss: 0.15304908, Cur Avg Loss: 0.14140931, Log Avg loss: 0.13380100, Global Avg Loss: 0.60793339, Time: 0.0209 Steps: 89900, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000492, Sample Num: 7872, Cur Loss: 0.15845841, Cur Avg Loss: 0.14154212, Log Avg loss: 0.14794353, Global Avg Loss: 0.60788223, Time: 0.0209 Steps: 89910, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000502, Sample Num: 8032, Cur Loss: 0.39899987, Cur Avg Loss: 0.14155882, Log Avg loss: 0.14238075, Global Avg Loss: 0.60783046, Time: 0.0209 Steps: 89920, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000512, Sample Num: 8192, Cur Loss: 0.02970472, Cur Avg Loss: 0.14169540, Log Avg loss: 0.14855179, Global Avg Loss: 0.60777939, Time: 0.0253 Steps: 89930, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000522, Sample Num: 8352, Cur Loss: 0.15506595, Cur Avg Loss: 0.14128740, Log Avg loss: 0.12039755, Global Avg Loss: 0.60772520, Time: 0.0208 Steps: 89940, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000532, Sample Num: 8512, Cur Loss: 0.19611348, Cur Avg Loss: 0.14115211, Log Avg loss: 0.13408994, Global Avg Loss: 0.60767255, Time: 0.0208 Steps: 89950, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000542, Sample Num: 8672, Cur Loss: 0.10921191, Cur Avg Loss: 0.14077647, Log Avg loss: 0.12079226, Global Avg Loss: 0.60761843, Time: 0.0208 Steps: 89960, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000552, Sample Num: 8832, Cur Loss: 0.34745973, Cur Avg Loss: 0.14204811, Log Avg loss: 0.21097097, Global Avg Loss: 0.60757434, Time: 0.0208 Steps: 89970, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000562, Sample Num: 8992, Cur Loss: 0.42920923, Cur Avg Loss: 0.14306062, Log Avg loss: 0.19895143, Global Avg Loss: 0.60752893, Time: 0.0208 Steps: 89980, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000572, Sample Num: 9152, Cur Loss: 0.02470026, Cur Avg Loss: 0.14277036, Log Avg loss: 0.12645756, Global Avg Loss: 0.60747547, Time: 0.0208 Steps: 89990, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000582, Sample Num: 9312, Cur Loss: 0.17982444, Cur Avg Loss: 0.14315893, Log Avg loss: 0.16538514, Global Avg Loss: 0.60742635, Time: 0.0208 Steps: 90000, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000592, Sample Num: 9472, Cur Loss: 0.22596046, Cur Avg Loss: 0.14324195, Log Avg loss: 0.14807363, Global Avg Loss: 0.60737531, Time: 0.0208 Steps: 90010, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000602, Sample Num: 9632, Cur Loss: 0.15022722, Cur Avg Loss: 0.14382463, Log Avg loss: 0.17831961, Global Avg Loss: 0.60732765, Time: 0.0208 Steps: 90020, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000612, Sample Num: 9792, Cur Loss: 0.02998631, Cur Avg Loss: 0.14419032, Log Avg loss: 0.16620502, Global Avg Loss: 0.60727865, Time: 0.0208 Steps: 90030, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000622, Sample Num: 9952, Cur Loss: 0.07011205, Cur Avg Loss: 0.14281228, Log Avg loss: 0.05847595, Global Avg Loss: 0.60721770, Time: 0.0208 Steps: 90040, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000632, Sample Num: 10112, Cur Loss: 0.03233714, Cur Avg Loss: 0.14198535, Log Avg loss: 0.09055058, Global Avg Loss: 0.60716033, Time: 0.0208 Steps: 90050, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000642, Sample Num: 10272, Cur Loss: 0.26880381, Cur Avg Loss: 0.14134950, Log Avg loss: 0.10116346, Global Avg Loss: 0.60710414, Time: 0.0208 Steps: 90060, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000652, Sample Num: 10432, Cur Loss: 0.07666111, Cur Avg Loss: 0.14127212, Log Avg loss: 0.13630451, Global Avg Loss: 0.60705187, Time: 0.0208 Steps: 90070, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000662, Sample Num: 10592, Cur Loss: 0.06591627, Cur Avg Loss: 0.14116889, Log Avg loss: 0.13443799, Global Avg Loss: 0.60699941, Time: 0.0208 Steps: 90080, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000672, Sample Num: 10752, Cur Loss: 0.19796462, Cur Avg Loss: 0.14099624, Log Avg loss: 0.12956691, Global Avg Loss: 0.60694641, Time: 0.0208 Steps: 90090, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000682, Sample Num: 10912, Cur Loss: 0.02039200, Cur Avg Loss: 0.14104651, Log Avg loss: 0.14442478, Global Avg Loss: 0.60689508, Time: 0.0208 Steps: 90100, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000692, Sample Num: 11072, Cur Loss: 0.18677859, Cur Avg Loss: 0.14087721, Log Avg loss: 0.12933091, Global Avg Loss: 0.60684208, Time: 0.0208 Steps: 90110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000702, Sample Num: 11232, Cur Loss: 0.05612099, Cur Avg Loss: 0.14118243, Log Avg loss: 0.16230345, Global Avg Loss: 0.60679275, Time: 0.0208 Steps: 90120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000712, Sample Num: 11392, Cur Loss: 0.07215921, Cur Avg Loss: 0.14100145, Log Avg loss: 0.12829683, Global Avg Loss: 0.60673966, Time: 0.0208 Steps: 90130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000722, Sample Num: 11552, Cur Loss: 0.04414867, Cur Avg Loss: 0.14072820, Log Avg loss: 0.12127271, Global Avg Loss: 0.60668581, Time: 0.0208 Steps: 90140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000732, Sample Num: 11712, Cur Loss: 0.02626470, Cur Avg Loss: 0.14105934, Log Avg loss: 0.16496762, Global Avg Loss: 0.60663681, Time: 0.0208 Steps: 90150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000742, Sample Num: 11872, Cur Loss: 0.04205469, Cur Avg Loss: 0.14078235, Log Avg loss: 0.12050658, Global Avg Loss: 0.60658289, Time: 0.0208 Steps: 90160, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000752, Sample Num: 12032, Cur Loss: 0.41623834, Cur Avg Loss: 0.14093101, Log Avg loss: 0.15196199, Global Avg Loss: 0.60653247, Time: 0.0208 Steps: 90170, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000762, Sample Num: 12192, Cur Loss: 0.05731762, Cur Avg Loss: 0.14083125, Log Avg loss: 0.13332883, Global Avg Loss: 0.60648000, Time: 0.0208 Steps: 90180, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000772, Sample Num: 12352, Cur Loss: 0.10753840, Cur Avg Loss: 0.14043525, Log Avg loss: 0.11026036, Global Avg Loss: 0.60642498, Time: 0.0246 Steps: 90190, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000782, Sample Num: 12512, Cur Loss: 0.18715201, Cur Avg Loss: 0.14024716, Log Avg loss: 0.12572683, Global Avg Loss: 0.60637169, Time: 0.0207 Steps: 90200, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000792, Sample Num: 12672, Cur Loss: 0.07097137, Cur Avg Loss: 0.14022786, Log Avg loss: 0.13871882, Global Avg Loss: 0.60631984, Time: 0.0208 Steps: 90210, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000802, Sample Num: 12832, Cur Loss: 0.25358975, Cur Avg Loss: 0.14055370, Log Avg loss: 0.16635948, Global Avg Loss: 0.60627108, Time: 0.0208 Steps: 90220, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000812, Sample Num: 12992, Cur Loss: 0.03510632, Cur Avg Loss: 0.14038106, Log Avg loss: 0.12653610, Global Avg Loss: 0.60621791, Time: 0.0208 Steps: 90230, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000822, Sample Num: 13152, Cur Loss: 0.26701728, Cur Avg Loss: 0.14019428, Log Avg loss: 0.12502765, Global Avg Loss: 0.60616459, Time: 0.0208 Steps: 90240, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000832, Sample Num: 13312, Cur Loss: 0.15251209, Cur Avg Loss: 0.14034430, Log Avg loss: 0.15267604, Global Avg Loss: 0.60611434, Time: 0.0207 Steps: 90250, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000842, Sample Num: 13472, Cur Loss: 0.11697935, Cur Avg Loss: 0.14018073, Log Avg loss: 0.12657115, Global Avg Loss: 0.60606121, Time: 0.0208 Steps: 90260, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000852, Sample Num: 13632, Cur Loss: 0.04990070, Cur Avg Loss: 0.14025029, Log Avg loss: 0.14610752, Global Avg Loss: 0.60601026, Time: 0.0208 Steps: 90270, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000862, Sample Num: 13792, Cur Loss: 0.07363535, Cur Avg Loss: 0.14005791, Log Avg loss: 0.12366683, Global Avg Loss: 0.60595683, Time: 0.0207 Steps: 90280, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000872, Sample Num: 13952, Cur Loss: 0.28838629, Cur Avg Loss: 0.13998232, Log Avg loss: 0.13346684, Global Avg Loss: 0.60590450, Time: 0.0208 Steps: 90290, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000882, Sample Num: 14112, Cur Loss: 0.17822105, Cur Avg Loss: 0.14014574, Log Avg loss: 0.15439569, Global Avg Loss: 0.60585450, Time: 0.0208 Steps: 90300, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000892, Sample Num: 14272, Cur Loss: 0.12667945, Cur Avg Loss: 0.13964539, Log Avg loss: 0.09551469, Global Avg Loss: 0.60579799, Time: 0.0208 Steps: 90310, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000902, Sample Num: 14432, Cur Loss: 0.09151533, Cur Avg Loss: 0.13970426, Log Avg loss: 0.14495566, Global Avg Loss: 0.60574697, Time: 0.0208 Steps: 90320, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000912, Sample Num: 14592, Cur Loss: 0.05337770, Cur Avg Loss: 0.13938997, Log Avg loss: 0.11104054, Global Avg Loss: 0.60569220, Time: 0.0208 Steps: 90330, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000922, Sample Num: 14752, Cur Loss: 0.21742333, Cur Avg Loss: 0.14048562, Log Avg loss: 0.24040926, Global Avg Loss: 0.60565177, Time: 0.0208 Steps: 90340, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000932, Sample Num: 14912, Cur Loss: 0.17330649, Cur Avg Loss: 0.14046371, Log Avg loss: 0.13844376, Global Avg Loss: 0.60560005, Time: 0.0208 Steps: 90350, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000942, Sample Num: 15072, Cur Loss: 0.08360319, Cur Avg Loss: 0.14015976, Log Avg loss: 0.11183141, Global Avg Loss: 0.60554541, Time: 0.0208 Steps: 90360, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000952, Sample Num: 15232, Cur Loss: 0.15403196, Cur Avg Loss: 0.14052380, Log Avg loss: 0.17481645, Global Avg Loss: 0.60549775, Time: 0.0208 Steps: 90370, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000962, Sample Num: 15392, Cur Loss: 0.09305743, Cur Avg Loss: 0.14041183, Log Avg loss: 0.12975204, Global Avg Loss: 0.60544511, Time: 0.0208 Steps: 90380, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000972, Sample Num: 15552, Cur Loss: 0.11116073, Cur Avg Loss: 0.14007785, Log Avg loss: 0.10794928, Global Avg Loss: 0.60539007, Time: 0.0208 Steps: 90390, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000982, Sample Num: 15712, Cur Loss: 0.20843026, Cur Avg Loss: 0.14022218, Log Avg loss: 0.15425098, Global Avg Loss: 0.60534016, Time: 0.0208 Steps: 90400, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000992, Sample Num: 15872, Cur Loss: 0.15061764, Cur Avg Loss: 0.14004965, Log Avg loss: 0.12310663, Global Avg Loss: 0.60528683, Time: 0.0208 Steps: 90410, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001002, Sample Num: 16032, Cur Loss: 0.13291948, Cur Avg Loss: 0.14017123, Log Avg loss: 0.15223202, Global Avg Loss: 0.60523672, Time: 0.0208 Steps: 90420, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001012, Sample Num: 16192, Cur Loss: 0.08385636, Cur Avg Loss: 0.14002280, Log Avg loss: 0.12515023, Global Avg Loss: 0.60518363, Time: 0.0208 Steps: 90430, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001022, Sample Num: 16352, Cur Loss: 0.11978994, Cur Avg Loss: 0.14024796, Log Avg loss: 0.16303480, Global Avg Loss: 0.60513474, Time: 0.0208 Steps: 90440, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001032, Sample Num: 16512, Cur Loss: 0.13464651, Cur Avg Loss: 0.14001943, Log Avg loss: 0.11666298, Global Avg Loss: 0.60508074, Time: 0.0209 Steps: 90450, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001042, Sample Num: 16672, Cur Loss: 0.15994492, Cur Avg Loss: 0.13990221, Log Avg loss: 0.12780483, Global Avg Loss: 0.60502798, Time: 0.0208 Steps: 90460, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001052, Sample Num: 16832, Cur Loss: 0.09532156, Cur Avg Loss: 0.13975093, Log Avg loss: 0.12398833, Global Avg Loss: 0.60497481, Time: 0.0208 Steps: 90470, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001062, Sample Num: 16992, Cur Loss: 0.16288851, Cur Avg Loss: 0.14027346, Log Avg loss: 0.19524296, Global Avg Loss: 0.60492952, Time: 0.0209 Steps: 90480, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001072, Sample Num: 17152, Cur Loss: 0.10993402, Cur Avg Loss: 0.14026923, Log Avg loss: 0.13982069, Global Avg Loss: 0.60487812, Time: 0.0208 Steps: 90490, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001082, Sample Num: 17312, Cur Loss: 0.30556440, Cur Avg Loss: 0.14032698, Log Avg loss: 0.14651769, Global Avg Loss: 0.60482748, Time: 0.0208 Steps: 90500, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001092, Sample Num: 17472, Cur Loss: 0.10972898, Cur Avg Loss: 0.13995006, Log Avg loss: 0.09916712, Global Avg Loss: 0.60477161, Time: 0.0208 Steps: 90510, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001102, Sample Num: 17632, Cur Loss: 0.04826232, Cur Avg Loss: 0.13972625, Log Avg loss: 0.11528662, Global Avg Loss: 0.60471753, Time: 0.0208 Steps: 90520, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001112, Sample Num: 17792, Cur Loss: 0.02901565, Cur Avg Loss: 0.13955885, Log Avg loss: 0.12111053, Global Avg Loss: 0.60466411, Time: 0.0208 Steps: 90530, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001122, Sample Num: 17952, Cur Loss: 0.13678710, Cur Avg Loss: 0.13937024, Log Avg loss: 0.11839745, Global Avg Loss: 0.60461041, Time: 0.0208 Steps: 90540, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001132, Sample Num: 18112, Cur Loss: 0.10671132, Cur Avg Loss: 0.13980718, Log Avg loss: 0.18883109, Global Avg Loss: 0.60456449, Time: 0.0208 Steps: 90550, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001142, Sample Num: 18272, Cur Loss: 0.09817053, Cur Avg Loss: 0.13979614, Log Avg loss: 0.13854645, Global Avg Loss: 0.60451303, Time: 0.0208 Steps: 90560, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001152, Sample Num: 18432, Cur Loss: 0.09650420, Cur Avg Loss: 0.14026459, Log Avg loss: 0.19376182, Global Avg Loss: 0.60446768, Time: 0.0208 Steps: 90570, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001162, Sample Num: 18592, Cur Loss: 0.15344632, Cur Avg Loss: 0.14052759, Log Avg loss: 0.17082497, Global Avg Loss: 0.60441980, Time: 0.0208 Steps: 90580, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001172, Sample Num: 18752, Cur Loss: 0.08105537, Cur Avg Loss: 0.14033539, Log Avg loss: 0.11800178, Global Avg Loss: 0.60436611, Time: 0.0208 Steps: 90590, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001182, Sample Num: 18912, Cur Loss: 0.06288279, Cur Avg Loss: 0.13999019, Log Avg loss: 0.09953263, Global Avg Loss: 0.60431039, Time: 0.0208 Steps: 90600, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001192, Sample Num: 19072, Cur Loss: 0.02606361, Cur Avg Loss: 0.13992100, Log Avg loss: 0.13174307, Global Avg Loss: 0.60425823, Time: 0.0208 Steps: 90610, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001202, Sample Num: 19232, Cur Loss: 0.38947496, Cur Avg Loss: 0.14025090, Log Avg loss: 0.17957512, Global Avg Loss: 0.60421137, Time: 0.0208 Steps: 90620, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001212, Sample Num: 19392, Cur Loss: 0.03526202, Cur Avg Loss: 0.14025777, Log Avg loss: 0.14108405, Global Avg Loss: 0.60416027, Time: 0.0208 Steps: 90630, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001222, Sample Num: 19552, Cur Loss: 0.14339864, Cur Avg Loss: 0.14004518, Log Avg loss: 0.11427896, Global Avg Loss: 0.60410622, Time: 0.0210 Steps: 90640, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001232, Sample Num: 19712, Cur Loss: 0.31449330, Cur Avg Loss: 0.14092599, Log Avg loss: 0.24856068, Global Avg Loss: 0.60406700, Time: 0.0208 Steps: 90650, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001242, Sample Num: 19872, Cur Loss: 0.05103043, Cur Avg Loss: 0.14081004, Log Avg loss: 0.12652503, Global Avg Loss: 0.60401433, Time: 0.0208 Steps: 90660, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001252, Sample Num: 20032, Cur Loss: 0.24818611, Cur Avg Loss: 0.14110021, Log Avg loss: 0.17713993, Global Avg Loss: 0.60396725, Time: 0.0208 Steps: 90670, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001262, Sample Num: 20192, Cur Loss: 0.23572524, Cur Avg Loss: 0.14120420, Log Avg loss: 0.15422328, Global Avg Loss: 0.60391765, Time: 0.0208 Steps: 90680, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001272, Sample Num: 20352, Cur Loss: 0.06156574, Cur Avg Loss: 0.14130482, Log Avg loss: 0.15400305, Global Avg Loss: 0.60386804, Time: 0.0208 Steps: 90690, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001282, Sample Num: 20512, Cur Loss: 0.07791737, Cur Avg Loss: 0.14104695, Log Avg loss: 0.10824613, Global Avg Loss: 0.60381339, Time: 0.0246 Steps: 90700, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001292, Sample Num: 20672, Cur Loss: 0.17931601, Cur Avg Loss: 0.14108954, Log Avg loss: 0.14654991, Global Avg Loss: 0.60376299, Time: 0.0208 Steps: 90710, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001302, Sample Num: 20832, Cur Loss: 0.02307917, Cur Avg Loss: 0.14105356, Log Avg loss: 0.13640503, Global Avg Loss: 0.60371147, Time: 0.0208 Steps: 90720, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001312, Sample Num: 20992, Cur Loss: 0.36767721, Cur Avg Loss: 0.14155971, Log Avg loss: 0.20745997, Global Avg Loss: 0.60366780, Time: 0.0207 Steps: 90730, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001322, Sample Num: 21152, Cur Loss: 0.16753191, Cur Avg Loss: 0.14151346, Log Avg loss: 0.13544521, Global Avg Loss: 0.60361619, Time: 0.0208 Steps: 90740, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001332, Sample Num: 21312, Cur Loss: 0.20790961, Cur Avg Loss: 0.14119948, Log Avg loss: 0.09969205, Global Avg Loss: 0.60356067, Time: 0.0208 Steps: 90750, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001342, Sample Num: 21472, Cur Loss: 0.15928890, Cur Avg Loss: 0.14101516, Log Avg loss: 0.11646278, Global Avg Loss: 0.60350700, Time: 0.0208 Steps: 90760, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001352, Sample Num: 21632, Cur Loss: 0.10220932, Cur Avg Loss: 0.14103709, Log Avg loss: 0.14398103, Global Avg Loss: 0.60345637, Time: 0.0208 Steps: 90770, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001362, Sample Num: 21792, Cur Loss: 0.15069677, Cur Avg Loss: 0.14150776, Log Avg loss: 0.20514196, Global Avg Loss: 0.60341250, Time: 0.0208 Steps: 90780, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001372, Sample Num: 21952, Cur Loss: 0.20146239, Cur Avg Loss: 0.14174325, Log Avg loss: 0.17381730, Global Avg Loss: 0.60336518, Time: 0.0208 Steps: 90790, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001382, Sample Num: 22112, Cur Loss: 0.17727503, Cur Avg Loss: 0.14179881, Log Avg loss: 0.14942084, Global Avg Loss: 0.60331518, Time: 0.0208 Steps: 90800, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001392, Sample Num: 22272, Cur Loss: 0.21557149, Cur Avg Loss: 0.14189822, Log Avg loss: 0.15563719, Global Avg Loss: 0.60326589, Time: 0.0208 Steps: 90810, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001402, Sample Num: 22432, Cur Loss: 0.16765895, Cur Avg Loss: 0.14209134, Log Avg loss: 0.16897357, Global Avg Loss: 0.60321807, Time: 0.0208 Steps: 90820, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001412, Sample Num: 22592, Cur Loss: 0.04347942, Cur Avg Loss: 0.14180613, Log Avg loss: 0.10181974, Global Avg Loss: 0.60316286, Time: 0.0208 Steps: 90830, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001422, Sample Num: 22752, Cur Loss: 0.06714860, Cur Avg Loss: 0.14167449, Log Avg loss: 0.12308658, Global Avg Loss: 0.60311002, Time: 0.0208 Steps: 90840, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001432, Sample Num: 22912, Cur Loss: 0.04728384, Cur Avg Loss: 0.14132154, Log Avg loss: 0.09113233, Global Avg Loss: 0.60305366, Time: 0.0208 Steps: 90850, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001442, Sample Num: 23072, Cur Loss: 0.09115600, Cur Avg Loss: 0.14114379, Log Avg loss: 0.11568982, Global Avg Loss: 0.60300002, Time: 0.0208 Steps: 90860, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001452, Sample Num: 23232, Cur Loss: 0.37925223, Cur Avg Loss: 0.14173460, Log Avg loss: 0.22692907, Global Avg Loss: 0.60295864, Time: 0.0208 Steps: 90870, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001462, Sample Num: 23392, Cur Loss: 0.12457884, Cur Avg Loss: 0.14193102, Log Avg loss: 0.17045170, Global Avg Loss: 0.60291105, Time: 0.0207 Steps: 90880, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001472, Sample Num: 23552, Cur Loss: 0.15487903, Cur Avg Loss: 0.14156397, Log Avg loss: 0.08790180, Global Avg Loss: 0.60285438, Time: 0.0210 Steps: 90890, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001482, Sample Num: 23712, Cur Loss: 0.15411106, Cur Avg Loss: 0.14176967, Log Avg loss: 0.17204851, Global Avg Loss: 0.60280699, Time: 0.0208 Steps: 90900, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001492, Sample Num: 23872, Cur Loss: 0.10538206, Cur Avg Loss: 0.14207174, Log Avg loss: 0.18683771, Global Avg Loss: 0.60276123, Time: 0.0208 Steps: 90910, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001502, Sample Num: 24032, Cur Loss: 0.10323277, Cur Avg Loss: 0.14183667, Log Avg loss: 0.10676446, Global Avg Loss: 0.60270668, Time: 0.0207 Steps: 90920, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001512, Sample Num: 24192, Cur Loss: 0.06903704, Cur Avg Loss: 0.14197475, Log Avg loss: 0.16271422, Global Avg Loss: 0.60265829, Time: 0.0208 Steps: 90930, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001522, Sample Num: 24352, Cur Loss: 0.10071222, Cur Avg Loss: 0.14185617, Log Avg loss: 0.12392646, Global Avg Loss: 0.60260565, Time: 0.0208 Steps: 90940, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001532, Sample Num: 24512, Cur Loss: 0.09904400, Cur Avg Loss: 0.14192925, Log Avg loss: 0.15305210, Global Avg Loss: 0.60255622, Time: 0.0208 Steps: 90950, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001542, Sample Num: 24672, Cur Loss: 0.25887260, Cur Avg Loss: 0.14198208, Log Avg loss: 0.15007675, Global Avg Loss: 0.60250648, Time: 0.0246 Steps: 90960, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001552, Sample Num: 24832, Cur Loss: 0.02543327, Cur Avg Loss: 0.14210952, Log Avg loss: 0.16176017, Global Avg Loss: 0.60245803, Time: 0.0208 Steps: 90970, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001562, Sample Num: 24992, Cur Loss: 0.02000663, Cur Avg Loss: 0.14180938, Log Avg loss: 0.09522795, Global Avg Loss: 0.60240228, Time: 0.0208 Steps: 90980, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001572, Sample Num: 25152, Cur Loss: 0.66780186, Cur Avg Loss: 0.14196058, Log Avg loss: 0.16557785, Global Avg Loss: 0.60235427, Time: 0.0208 Steps: 90990, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001582, Sample Num: 25312, Cur Loss: 0.11881610, Cur Avg Loss: 0.14166145, Log Avg loss: 0.09463867, Global Avg Loss: 0.60229847, Time: 0.0208 Steps: 91000, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001592, Sample Num: 25472, Cur Loss: 0.26140761, Cur Avg Loss: 0.14168365, Log Avg loss: 0.14519499, Global Avg Loss: 0.60224825, Time: 0.0208 Steps: 91010, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001602, Sample Num: 25632, Cur Loss: 0.11823305, Cur Avg Loss: 0.14199484, Log Avg loss: 0.19153643, Global Avg Loss: 0.60220313, Time: 0.0208 Steps: 91020, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001612, Sample Num: 25792, Cur Loss: 0.01932018, Cur Avg Loss: 0.14165904, Log Avg loss: 0.08786447, Global Avg Loss: 0.60214662, Time: 0.0208 Steps: 91030, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001622, Sample Num: 25952, Cur Loss: 0.17323950, Cur Avg Loss: 0.14142759, Log Avg loss: 0.10411717, Global Avg Loss: 0.60209192, Time: 0.0208 Steps: 91040, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001632, Sample Num: 26112, Cur Loss: 0.05542927, Cur Avg Loss: 0.14161108, Log Avg loss: 0.17137309, Global Avg Loss: 0.60204461, Time: 0.0208 Steps: 91050, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001642, Sample Num: 26272, Cur Loss: 0.21700045, Cur Avg Loss: 0.14144245, Log Avg loss: 0.11392285, Global Avg Loss: 0.60199101, Time: 0.0208 Steps: 91060, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001652, Sample Num: 26432, Cur Loss: 0.12943915, Cur Avg Loss: 0.14139115, Log Avg loss: 0.13296724, Global Avg Loss: 0.60193951, Time: 0.0208 Steps: 91070, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001662, Sample Num: 26592, Cur Loss: 0.19810100, Cur Avg Loss: 0.14140379, Log Avg loss: 0.14349147, Global Avg Loss: 0.60188917, Time: 0.0208 Steps: 91080, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001672, Sample Num: 26752, Cur Loss: 0.22422613, Cur Avg Loss: 0.14169176, Log Avg loss: 0.18955213, Global Avg Loss: 0.60184391, Time: 0.0208 Steps: 91090, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001682, Sample Num: 26912, Cur Loss: 0.23787232, Cur Avg Loss: 0.14160101, Log Avg loss: 0.12642865, Global Avg Loss: 0.60179172, Time: 0.0208 Steps: 91100, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001692, Sample Num: 27072, Cur Loss: 0.14336255, Cur Avg Loss: 0.14137274, Log Avg loss: 0.10297667, Global Avg Loss: 0.60173697, Time: 0.0208 Steps: 91110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001702, Sample Num: 27232, Cur Loss: 0.17484474, Cur Avg Loss: 0.14134155, Log Avg loss: 0.13606414, Global Avg Loss: 0.60168587, Time: 0.0208 Steps: 91120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001712, Sample Num: 27392, Cur Loss: 0.42784840, Cur Avg Loss: 0.14193106, Log Avg loss: 0.24226658, Global Avg Loss: 0.60164642, Time: 0.0208 Steps: 91130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001722, Sample Num: 27552, Cur Loss: 0.05579277, Cur Avg Loss: 0.14214372, Log Avg loss: 0.17855091, Global Avg Loss: 0.60160000, Time: 0.0208 Steps: 91140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001732, Sample Num: 27712, Cur Loss: 0.11644569, Cur Avg Loss: 0.14210626, Log Avg loss: 0.13565627, Global Avg Loss: 0.60154888, Time: 0.0208 Steps: 91150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001742, Sample Num: 27872, Cur Loss: 0.09550510, Cur Avg Loss: 0.14204650, Log Avg loss: 0.13169600, Global Avg Loss: 0.60149734, Time: 0.0208 Steps: 91160, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001752, Sample Num: 28032, Cur Loss: 0.10743406, Cur Avg Loss: 0.14199361, Log Avg loss: 0.13277926, Global Avg Loss: 0.60144593, Time: 0.0208 Steps: 91170, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001762, Sample Num: 28192, Cur Loss: 0.25368953, Cur Avg Loss: 0.14204449, Log Avg loss: 0.15095845, Global Avg Loss: 0.60139652, Time: 0.0208 Steps: 91180, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001772, Sample Num: 28352, Cur Loss: 0.03772689, Cur Avg Loss: 0.14196239, Log Avg loss: 0.12749627, Global Avg Loss: 0.60134456, Time: 0.0208 Steps: 91190, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001782, Sample Num: 28512, Cur Loss: 0.09845524, Cur Avg Loss: 0.14186539, Log Avg loss: 0.12467676, Global Avg Loss: 0.60129229, Time: 0.0208 Steps: 91200, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001792, Sample Num: 28672, Cur Loss: 0.11245230, Cur Avg Loss: 0.14214379, Log Avg loss: 0.19175474, Global Avg Loss: 0.60124739, Time: 0.0253 Steps: 91210, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001802, Sample Num: 28832, Cur Loss: 0.12794191, Cur Avg Loss: 0.14229915, Log Avg loss: 0.17014033, Global Avg Loss: 0.60120013, Time: 0.0210 Steps: 91220, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001812, Sample Num: 28992, Cur Loss: 0.10698811, Cur Avg Loss: 0.14218244, Log Avg loss: 0.12115161, Global Avg Loss: 0.60114751, Time: 0.0208 Steps: 91230, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001822, Sample Num: 29152, Cur Loss: 0.03572598, Cur Avg Loss: 0.14200128, Log Avg loss: 0.10917562, Global Avg Loss: 0.60109359, Time: 0.0207 Steps: 91240, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001832, Sample Num: 29312, Cur Loss: 0.08010378, Cur Avg Loss: 0.14216491, Log Avg loss: 0.17197798, Global Avg Loss: 0.60104656, Time: 0.0207 Steps: 91250, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001842, Sample Num: 29472, Cur Loss: 0.44964856, Cur Avg Loss: 0.14275523, Log Avg loss: 0.25090120, Global Avg Loss: 0.60100819, Time: 0.0208 Steps: 91260, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001852, Sample Num: 29632, Cur Loss: 0.28153881, Cur Avg Loss: 0.14252582, Log Avg loss: 0.10026833, Global Avg Loss: 0.60095333, Time: 0.0207 Steps: 91270, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001862, Sample Num: 29792, Cur Loss: 0.08791151, Cur Avg Loss: 0.14276961, Log Avg loss: 0.18791994, Global Avg Loss: 0.60090808, Time: 0.0207 Steps: 91280, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001872, Sample Num: 29952, Cur Loss: 0.22367889, Cur Avg Loss: 0.14264330, Log Avg loss: 0.11912410, Global Avg Loss: 0.60085531, Time: 0.0207 Steps: 91290, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001882, Sample Num: 30112, Cur Loss: 0.02534569, Cur Avg Loss: 0.14270385, Log Avg loss: 0.15403852, Global Avg Loss: 0.60080637, Time: 0.0208 Steps: 91300, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001892, Sample Num: 30272, Cur Loss: 0.14658026, Cur Avg Loss: 0.14295335, Log Avg loss: 0.18991044, Global Avg Loss: 0.60076137, Time: 0.0208 Steps: 91310, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001902, Sample Num: 30432, Cur Loss: 0.31649679, Cur Avg Loss: 0.14334255, Log Avg loss: 0.21697934, Global Avg Loss: 0.60071934, Time: 0.0208 Steps: 91320, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001912, Sample Num: 30592, Cur Loss: 0.12301067, Cur Avg Loss: 0.14328779, Log Avg loss: 0.13287170, Global Avg Loss: 0.60066812, Time: 0.0208 Steps: 91330, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001922, Sample Num: 30752, Cur Loss: 0.41699177, Cur Avg Loss: 0.14371904, Log Avg loss: 0.22617463, Global Avg Loss: 0.60062712, Time: 0.0208 Steps: 91340, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001932, Sample Num: 30912, Cur Loss: 0.28380564, Cur Avg Loss: 0.14371939, Log Avg loss: 0.14378614, Global Avg Loss: 0.60057711, Time: 0.0208 Steps: 91350, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001942, Sample Num: 31072, Cur Loss: 0.06736134, Cur Avg Loss: 0.14348823, Log Avg loss: 0.09882763, Global Avg Loss: 0.60052219, Time: 0.0208 Steps: 91360, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001952, Sample Num: 31232, Cur Loss: 0.31681266, Cur Avg Loss: 0.14351541, Log Avg loss: 0.14879308, Global Avg Loss: 0.60047275, Time: 0.0208 Steps: 91370, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001962, Sample Num: 31392, Cur Loss: 0.06163345, Cur Avg Loss: 0.14339396, Log Avg loss: 0.11968877, Global Avg Loss: 0.60042013, Time: 0.0207 Steps: 91380, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001972, Sample Num: 31552, Cur Loss: 0.11673936, Cur Avg Loss: 0.14320037, Log Avg loss: 0.10521717, Global Avg Loss: 0.60036595, Time: 0.0207 Steps: 91390, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001982, Sample Num: 31712, Cur Loss: 0.21567664, Cur Avg Loss: 0.14307450, Log Avg loss: 0.11825209, Global Avg Loss: 0.60031320, Time: 0.0207 Steps: 91400, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001992, Sample Num: 31872, Cur Loss: 0.03488450, Cur Avg Loss: 0.14305855, Log Avg loss: 0.13989757, Global Avg Loss: 0.60026283, Time: 0.0207 Steps: 91410, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002002, Sample Num: 32032, Cur Loss: 0.11049846, Cur Avg Loss: 0.14317743, Log Avg loss: 0.16685955, Global Avg Loss: 0.60021542, Time: 0.0207 Steps: 91420, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002012, Sample Num: 32192, Cur Loss: 0.20211773, Cur Avg Loss: 0.14357974, Log Avg loss: 0.22412056, Global Avg Loss: 0.60017429, Time: 0.0208 Steps: 91430, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002022, Sample Num: 32352, Cur Loss: 0.38633218, Cur Avg Loss: 0.14375774, Log Avg loss: 0.17957303, Global Avg Loss: 0.60012829, Time: 0.0207 Steps: 91440, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002032, Sample Num: 32512, Cur Loss: 0.60222697, Cur Avg Loss: 0.14403941, Log Avg loss: 0.20099312, Global Avg Loss: 0.60008465, Time: 0.0208 Steps: 91450, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002042, Sample Num: 32672, Cur Loss: 0.09266856, Cur Avg Loss: 0.14401688, Log Avg loss: 0.13943883, Global Avg Loss: 0.60003428, Time: 0.0207 Steps: 91460, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002052, Sample Num: 32832, Cur Loss: 0.09386241, Cur Avg Loss: 0.14390020, Log Avg loss: 0.12007375, Global Avg Loss: 0.59998181, Time: 0.0246 Steps: 91470, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002062, Sample Num: 32992, Cur Loss: 0.14040159, Cur Avg Loss: 0.14405326, Log Avg loss: 0.17546142, Global Avg Loss: 0.59993540, Time: 0.0208 Steps: 91480, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002072, Sample Num: 33152, Cur Loss: 0.11650038, Cur Avg Loss: 0.14405552, Log Avg loss: 0.14452098, Global Avg Loss: 0.59988562, Time: 0.0208 Steps: 91490, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002082, Sample Num: 33312, Cur Loss: 0.15482980, Cur Avg Loss: 0.14411087, Log Avg loss: 0.15557919, Global Avg Loss: 0.59983707, Time: 0.0209 Steps: 91500, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002092, Sample Num: 33472, Cur Loss: 0.10867267, Cur Avg Loss: 0.14395070, Log Avg loss: 0.11060258, Global Avg Loss: 0.59978360, Time: 0.0208 Steps: 91510, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002102, Sample Num: 33632, Cur Loss: 0.04683225, Cur Avg Loss: 0.14371662, Log Avg loss: 0.09474811, Global Avg Loss: 0.59972842, Time: 0.0208 Steps: 91520, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002112, Sample Num: 33792, Cur Loss: 0.03543477, Cur Avg Loss: 0.14357092, Log Avg loss: 0.11294501, Global Avg Loss: 0.59967524, Time: 0.0208 Steps: 91530, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002122, Sample Num: 33952, Cur Loss: 0.09448995, Cur Avg Loss: 0.14356831, Log Avg loss: 0.14301744, Global Avg Loss: 0.59962535, Time: 0.0209 Steps: 91540, Updated lr: 0.000014 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.574091, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.274912, "eval_total_loss": 193.262794, "eval_mae": 0.350853, "eval_mse": 0.275018, "eval_r2": 0.82518, "eval_sp_statistic": 0.931543, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.935381, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.170281, "test_total_loss": 85.481211, "test_mae": 0.291757, "test_mse": 0.170343, "test_r2": 0.890059, "test_sp_statistic": 0.920282, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.949269, "test_ps_pvalue": 0.0, "lr": 1.413276434329066e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5995875184016084, "train_cur_epoch_loss": 305.38581602368504, "train_cur_epoch_avg_loss": 0.14344096572272666, "train_cur_epoch_time": 44.57409143447876, "train_cur_epoch_avg_time": 0.020936632895480863, "epoch": 43, "step": 91547} ################################################## Training, Epoch: 0044, Batch: 000003, Sample Num: 48, Cur Loss: 0.06750768, Cur Avg Loss: 0.13813604, Log Avg loss: 0.11482618, Global Avg Loss: 0.59957240, Time: 0.0246 Steps: 91550, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000013, Sample Num: 208, Cur Loss: 0.10085527, Cur Avg Loss: 0.10785465, Log Avg loss: 0.09877024, Global Avg Loss: 0.59951770, Time: 0.0209 Steps: 91560, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000023, Sample Num: 368, Cur Loss: 0.08963938, Cur Avg Loss: 0.10643456, Log Avg loss: 0.10458845, Global Avg Loss: 0.59946365, Time: 0.0209 Steps: 91570, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000033, Sample Num: 528, Cur Loss: 0.08135776, Cur Avg Loss: 0.11984275, Log Avg loss: 0.15068158, Global Avg Loss: 0.59941465, Time: 0.0209 Steps: 91580, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000043, Sample Num: 688, Cur Loss: 0.12574270, Cur Avg Loss: 0.12554989, Log Avg loss: 0.14438346, Global Avg Loss: 0.59936497, Time: 0.0208 Steps: 91590, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000053, Sample Num: 848, Cur Loss: 0.05083540, Cur Avg Loss: 0.12373340, Log Avg loss: 0.11592249, Global Avg Loss: 0.59931219, Time: 0.0209 Steps: 91600, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000063, Sample Num: 1008, Cur Loss: 0.02129408, Cur Avg Loss: 0.12328506, Log Avg loss: 0.12090883, Global Avg Loss: 0.59925997, Time: 0.0209 Steps: 91610, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000073, Sample Num: 1168, Cur Loss: 0.15061021, Cur Avg Loss: 0.12379829, Log Avg loss: 0.12703165, Global Avg Loss: 0.59920842, Time: 0.0209 Steps: 91620, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000083, Sample Num: 1328, Cur Loss: 0.07134507, Cur Avg Loss: 0.13000031, Log Avg loss: 0.17527506, Global Avg Loss: 0.59916216, Time: 0.0208 Steps: 91630, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000093, Sample Num: 1488, Cur Loss: 0.07639115, Cur Avg Loss: 0.13187666, Log Avg loss: 0.14745040, Global Avg Loss: 0.59911287, Time: 0.0209 Steps: 91640, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000103, Sample Num: 1648, Cur Loss: 0.02538717, Cur Avg Loss: 0.13051562, Log Avg loss: 0.11785788, Global Avg Loss: 0.59906036, Time: 0.0209 Steps: 91650, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000113, Sample Num: 1808, Cur Loss: 0.06041070, Cur Avg Loss: 0.13373071, Log Avg loss: 0.16684613, Global Avg Loss: 0.59901320, Time: 0.0208 Steps: 91660, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000123, Sample Num: 1968, Cur Loss: 0.18341450, Cur Avg Loss: 0.13226150, Log Avg loss: 0.11565942, Global Avg Loss: 0.59896047, Time: 0.0209 Steps: 91670, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000133, Sample Num: 2128, Cur Loss: 0.21130617, Cur Avg Loss: 0.13083050, Log Avg loss: 0.11322926, Global Avg Loss: 0.59890749, Time: 0.0209 Steps: 91680, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000143, Sample Num: 2288, Cur Loss: 0.05674731, Cur Avg Loss: 0.12880157, Log Avg loss: 0.10181672, Global Avg Loss: 0.59885328, Time: 0.0209 Steps: 91690, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000153, Sample Num: 2448, Cur Loss: 0.10329732, Cur Avg Loss: 0.12954553, Log Avg loss: 0.14018422, Global Avg Loss: 0.59880326, Time: 0.0209 Steps: 91700, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000163, Sample Num: 2608, Cur Loss: 0.08936127, Cur Avg Loss: 0.12861351, Log Avg loss: 0.11435367, Global Avg Loss: 0.59875044, Time: 0.0209 Steps: 91710, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000173, Sample Num: 2768, Cur Loss: 0.04124252, Cur Avg Loss: 0.13053505, Log Avg loss: 0.16185602, Global Avg Loss: 0.59870280, Time: 0.0209 Steps: 91720, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000183, Sample Num: 2928, Cur Loss: 0.06005300, Cur Avg Loss: 0.13261728, Log Avg loss: 0.16863986, Global Avg Loss: 0.59865592, Time: 0.0209 Steps: 91730, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000193, Sample Num: 3088, Cur Loss: 0.11921421, Cur Avg Loss: 0.13258460, Log Avg loss: 0.13198656, Global Avg Loss: 0.59860505, Time: 0.0209 Steps: 91740, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000203, Sample Num: 3248, Cur Loss: 0.07388970, Cur Avg Loss: 0.13151799, Log Avg loss: 0.11093250, Global Avg Loss: 0.59855190, Time: 0.0209 Steps: 91750, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000213, Sample Num: 3408, Cur Loss: 0.06404965, Cur Avg Loss: 0.13355902, Log Avg loss: 0.17499189, Global Avg Loss: 0.59850574, Time: 0.0209 Steps: 91760, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000223, Sample Num: 3568, Cur Loss: 0.04283824, Cur Avg Loss: 0.13931445, Log Avg loss: 0.26190505, Global Avg Loss: 0.59846906, Time: 0.0209 Steps: 91770, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000233, Sample Num: 3728, Cur Loss: 0.06442343, Cur Avg Loss: 0.13971375, Log Avg loss: 0.14861812, Global Avg Loss: 0.59842005, Time: 0.0209 Steps: 91780, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000243, Sample Num: 3888, Cur Loss: 0.18099836, Cur Avg Loss: 0.13833068, Log Avg loss: 0.10610534, Global Avg Loss: 0.59836641, Time: 0.0209 Steps: 91790, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000253, Sample Num: 4048, Cur Loss: 0.28607365, Cur Avg Loss: 0.14073270, Log Avg loss: 0.19910156, Global Avg Loss: 0.59832292, Time: 0.0209 Steps: 91800, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000263, Sample Num: 4208, Cur Loss: 0.16954565, Cur Avg Loss: 0.13921588, Log Avg loss: 0.10084044, Global Avg Loss: 0.59826873, Time: 0.0209 Steps: 91810, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000273, Sample Num: 4368, Cur Loss: 0.05014115, Cur Avg Loss: 0.13754860, Log Avg loss: 0.09369926, Global Avg Loss: 0.59821378, Time: 0.0208 Steps: 91820, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000283, Sample Num: 4528, Cur Loss: 0.04602190, Cur Avg Loss: 0.13630494, Log Avg loss: 0.10235298, Global Avg Loss: 0.59815978, Time: 0.0208 Steps: 91830, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000293, Sample Num: 4688, Cur Loss: 0.13416155, Cur Avg Loss: 0.13677445, Log Avg loss: 0.15006162, Global Avg Loss: 0.59811099, Time: 0.0208 Steps: 91840, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000303, Sample Num: 4848, Cur Loss: 0.09583227, Cur Avg Loss: 0.13644972, Log Avg loss: 0.12693494, Global Avg Loss: 0.59805969, Time: 0.0209 Steps: 91850, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000313, Sample Num: 5008, Cur Loss: 0.06638604, Cur Avg Loss: 0.13723446, Log Avg loss: 0.16101228, Global Avg Loss: 0.59801212, Time: 0.0208 Steps: 91860, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000323, Sample Num: 5168, Cur Loss: 0.24001467, Cur Avg Loss: 0.13638778, Log Avg loss: 0.10988652, Global Avg Loss: 0.59795898, Time: 0.0208 Steps: 91870, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000333, Sample Num: 5328, Cur Loss: 0.19485037, Cur Avg Loss: 0.13621990, Log Avg loss: 0.13079755, Global Avg Loss: 0.59790814, Time: 0.0208 Steps: 91880, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000343, Sample Num: 5488, Cur Loss: 0.22928333, Cur Avg Loss: 0.13591969, Log Avg loss: 0.12592261, Global Avg Loss: 0.59785677, Time: 0.0209 Steps: 91890, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000353, Sample Num: 5648, Cur Loss: 0.28932831, Cur Avg Loss: 0.13596869, Log Avg loss: 0.13764938, Global Avg Loss: 0.59780670, Time: 0.0208 Steps: 91900, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000363, Sample Num: 5808, Cur Loss: 0.09571686, Cur Avg Loss: 0.13633935, Log Avg loss: 0.14942357, Global Avg Loss: 0.59775791, Time: 0.0209 Steps: 91910, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000373, Sample Num: 5968, Cur Loss: 0.20901084, Cur Avg Loss: 0.13725823, Log Avg loss: 0.17061344, Global Avg Loss: 0.59771144, Time: 0.0208 Steps: 91920, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000383, Sample Num: 6128, Cur Loss: 0.17057741, Cur Avg Loss: 0.13738085, Log Avg loss: 0.14195477, Global Avg Loss: 0.59766187, Time: 0.0208 Steps: 91930, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000393, Sample Num: 6288, Cur Loss: 0.18321255, Cur Avg Loss: 0.13707489, Log Avg loss: 0.12535670, Global Avg Loss: 0.59761050, Time: 0.0208 Steps: 91940, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000403, Sample Num: 6448, Cur Loss: 0.17015576, Cur Avg Loss: 0.13776227, Log Avg loss: 0.16477612, Global Avg Loss: 0.59756342, Time: 0.0208 Steps: 91950, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000413, Sample Num: 6608, Cur Loss: 0.21483929, Cur Avg Loss: 0.13804226, Log Avg loss: 0.14932590, Global Avg Loss: 0.59751468, Time: 0.0208 Steps: 91960, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000423, Sample Num: 6768, Cur Loss: 0.04331036, Cur Avg Loss: 0.13773865, Log Avg loss: 0.12519954, Global Avg Loss: 0.59746332, Time: 0.0208 Steps: 91970, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000433, Sample Num: 6928, Cur Loss: 0.12743330, Cur Avg Loss: 0.13752963, Log Avg loss: 0.12868828, Global Avg Loss: 0.59741236, Time: 0.0208 Steps: 91980, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000443, Sample Num: 7088, Cur Loss: 0.08926094, Cur Avg Loss: 0.13645588, Log Avg loss: 0.08996229, Global Avg Loss: 0.59735720, Time: 0.0208 Steps: 91990, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000453, Sample Num: 7248, Cur Loss: 0.09090026, Cur Avg Loss: 0.13607311, Log Avg loss: 0.11911634, Global Avg Loss: 0.59730521, Time: 0.0209 Steps: 92000, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000463, Sample Num: 7408, Cur Loss: 0.06085548, Cur Avg Loss: 0.13595691, Log Avg loss: 0.13069302, Global Avg Loss: 0.59725450, Time: 0.0208 Steps: 92010, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000473, Sample Num: 7568, Cur Loss: 0.07244443, Cur Avg Loss: 0.13671919, Log Avg loss: 0.17201293, Global Avg Loss: 0.59720829, Time: 0.0208 Steps: 92020, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000483, Sample Num: 7728, Cur Loss: 0.05571716, Cur Avg Loss: 0.13623448, Log Avg loss: 0.11330759, Global Avg Loss: 0.59715571, Time: 0.0208 Steps: 92030, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000493, Sample Num: 7888, Cur Loss: 0.29970142, Cur Avg Loss: 0.13856921, Log Avg loss: 0.25133671, Global Avg Loss: 0.59711814, Time: 0.0209 Steps: 92040, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000503, Sample Num: 8048, Cur Loss: 0.16903511, Cur Avg Loss: 0.13849651, Log Avg loss: 0.13491228, Global Avg Loss: 0.59706792, Time: 0.0209 Steps: 92050, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000513, Sample Num: 8208, Cur Loss: 0.25890729, Cur Avg Loss: 0.13870051, Log Avg loss: 0.14896187, Global Avg Loss: 0.59701925, Time: 0.0246 Steps: 92060, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000523, Sample Num: 8368, Cur Loss: 0.04130455, Cur Avg Loss: 0.13828585, Log Avg loss: 0.11701368, Global Avg Loss: 0.59696711, Time: 0.0209 Steps: 92070, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000533, Sample Num: 8528, Cur Loss: 0.13947012, Cur Avg Loss: 0.13844171, Log Avg loss: 0.14659355, Global Avg Loss: 0.59691820, Time: 0.0208 Steps: 92080, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000543, Sample Num: 8688, Cur Loss: 0.25064680, Cur Avg Loss: 0.13852682, Log Avg loss: 0.14306270, Global Avg Loss: 0.59686892, Time: 0.0208 Steps: 92090, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000553, Sample Num: 8848, Cur Loss: 0.02437551, Cur Avg Loss: 0.13781709, Log Avg loss: 0.09927903, Global Avg Loss: 0.59681489, Time: 0.0209 Steps: 92100, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000563, Sample Num: 9008, Cur Loss: 0.07924633, Cur Avg Loss: 0.13723306, Log Avg loss: 0.10493600, Global Avg Loss: 0.59676149, Time: 0.0208 Steps: 92110, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000573, Sample Num: 9168, Cur Loss: 0.10060930, Cur Avg Loss: 0.13677048, Log Avg loss: 0.11072717, Global Avg Loss: 0.59670873, Time: 0.0209 Steps: 92120, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000583, Sample Num: 9328, Cur Loss: 0.12974353, Cur Avg Loss: 0.13613133, Log Avg loss: 0.09950852, Global Avg Loss: 0.59665476, Time: 0.0208 Steps: 92130, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000593, Sample Num: 9488, Cur Loss: 0.19274612, Cur Avg Loss: 0.13611167, Log Avg loss: 0.13496525, Global Avg Loss: 0.59660465, Time: 0.0208 Steps: 92140, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000603, Sample Num: 9648, Cur Loss: 0.15534604, Cur Avg Loss: 0.13594810, Log Avg loss: 0.12624831, Global Avg Loss: 0.59655361, Time: 0.0208 Steps: 92150, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000613, Sample Num: 9808, Cur Loss: 0.07860035, Cur Avg Loss: 0.13615393, Log Avg loss: 0.14856559, Global Avg Loss: 0.59650500, Time: 0.0208 Steps: 92160, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000623, Sample Num: 9968, Cur Loss: 0.14843914, Cur Avg Loss: 0.13633913, Log Avg loss: 0.14769180, Global Avg Loss: 0.59645631, Time: 0.0208 Steps: 92170, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000633, Sample Num: 10128, Cur Loss: 0.16585687, Cur Avg Loss: 0.13579141, Log Avg loss: 0.10166852, Global Avg Loss: 0.59640263, Time: 0.0209 Steps: 92180, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000643, Sample Num: 10288, Cur Loss: 0.04557550, Cur Avg Loss: 0.13693820, Log Avg loss: 0.20952997, Global Avg Loss: 0.59636067, Time: 0.0208 Steps: 92190, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000653, Sample Num: 10448, Cur Loss: 0.12083788, Cur Avg Loss: 0.13748716, Log Avg loss: 0.17278533, Global Avg Loss: 0.59631473, Time: 0.0209 Steps: 92200, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000663, Sample Num: 10608, Cur Loss: 0.33085081, Cur Avg Loss: 0.13731980, Log Avg loss: 0.12639106, Global Avg Loss: 0.59626376, Time: 0.0209 Steps: 92210, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000673, Sample Num: 10768, Cur Loss: 0.08258726, Cur Avg Loss: 0.13741020, Log Avg loss: 0.14340389, Global Avg Loss: 0.59621466, Time: 0.0208 Steps: 92220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000683, Sample Num: 10928, Cur Loss: 0.04372042, Cur Avg Loss: 0.13753061, Log Avg loss: 0.14563437, Global Avg Loss: 0.59616580, Time: 0.0208 Steps: 92230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000693, Sample Num: 11088, Cur Loss: 0.08681946, Cur Avg Loss: 0.13760172, Log Avg loss: 0.14245857, Global Avg Loss: 0.59611661, Time: 0.0208 Steps: 92240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000703, Sample Num: 11248, Cur Loss: 0.03622518, Cur Avg Loss: 0.13750649, Log Avg loss: 0.13090660, Global Avg Loss: 0.59606619, Time: 0.0208 Steps: 92250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000713, Sample Num: 11408, Cur Loss: 0.22055718, Cur Avg Loss: 0.13770901, Log Avg loss: 0.15194641, Global Avg Loss: 0.59601805, Time: 0.0209 Steps: 92260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000723, Sample Num: 11568, Cur Loss: 0.14567405, Cur Avg Loss: 0.13747536, Log Avg loss: 0.12081623, Global Avg Loss: 0.59596655, Time: 0.0209 Steps: 92270, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000733, Sample Num: 11728, Cur Loss: 0.08135147, Cur Avg Loss: 0.13681655, Log Avg loss: 0.08918455, Global Avg Loss: 0.59591163, Time: 0.0208 Steps: 92280, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000743, Sample Num: 11888, Cur Loss: 0.13186982, Cur Avg Loss: 0.13711396, Log Avg loss: 0.15891382, Global Avg Loss: 0.59586428, Time: 0.0208 Steps: 92290, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000753, Sample Num: 12048, Cur Loss: 0.21227714, Cur Avg Loss: 0.13743332, Log Avg loss: 0.16116187, Global Avg Loss: 0.59581718, Time: 0.0208 Steps: 92300, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000763, Sample Num: 12208, Cur Loss: 0.07292154, Cur Avg Loss: 0.13712645, Log Avg loss: 0.11401940, Global Avg Loss: 0.59576499, Time: 0.0208 Steps: 92310, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000773, Sample Num: 12368, Cur Loss: 0.05863403, Cur Avg Loss: 0.13733592, Log Avg loss: 0.15331794, Global Avg Loss: 0.59571706, Time: 0.0247 Steps: 92320, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000783, Sample Num: 12528, Cur Loss: 0.12524444, Cur Avg Loss: 0.13675530, Log Avg loss: 0.09187371, Global Avg Loss: 0.59566249, Time: 0.0209 Steps: 92330, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000793, Sample Num: 12688, Cur Loss: 0.13714731, Cur Avg Loss: 0.13672445, Log Avg loss: 0.13430905, Global Avg Loss: 0.59561253, Time: 0.0209 Steps: 92340, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000803, Sample Num: 12848, Cur Loss: 0.08552922, Cur Avg Loss: 0.13701299, Log Avg loss: 0.15989427, Global Avg Loss: 0.59556535, Time: 0.0209 Steps: 92350, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000813, Sample Num: 13008, Cur Loss: 0.43329686, Cur Avg Loss: 0.13737528, Log Avg loss: 0.16646660, Global Avg Loss: 0.59551889, Time: 0.0209 Steps: 92360, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000823, Sample Num: 13168, Cur Loss: 0.25257981, Cur Avg Loss: 0.13792299, Log Avg loss: 0.18245224, Global Avg Loss: 0.59547417, Time: 0.0209 Steps: 92370, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000833, Sample Num: 13328, Cur Loss: 0.03251621, Cur Avg Loss: 0.13740444, Log Avg loss: 0.09472809, Global Avg Loss: 0.59541997, Time: 0.0209 Steps: 92380, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000843, Sample Num: 13488, Cur Loss: 0.17905998, Cur Avg Loss: 0.13770179, Log Avg loss: 0.16247108, Global Avg Loss: 0.59537310, Time: 0.0209 Steps: 92390, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000853, Sample Num: 13648, Cur Loss: 0.40261650, Cur Avg Loss: 0.13819948, Log Avg loss: 0.18015399, Global Avg Loss: 0.59532817, Time: 0.0209 Steps: 92400, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000863, Sample Num: 13808, Cur Loss: 0.08734733, Cur Avg Loss: 0.13789032, Log Avg loss: 0.11151895, Global Avg Loss: 0.59527581, Time: 0.0209 Steps: 92410, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000873, Sample Num: 13968, Cur Loss: 0.12410325, Cur Avg Loss: 0.13796028, Log Avg loss: 0.14399829, Global Avg Loss: 0.59522698, Time: 0.0209 Steps: 92420, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000883, Sample Num: 14128, Cur Loss: 0.07704423, Cur Avg Loss: 0.13772325, Log Avg loss: 0.11703005, Global Avg Loss: 0.59517525, Time: 0.0209 Steps: 92430, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000893, Sample Num: 14288, Cur Loss: 0.09405769, Cur Avg Loss: 0.13799011, Log Avg loss: 0.16155456, Global Avg Loss: 0.59512834, Time: 0.0209 Steps: 92440, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000903, Sample Num: 14448, Cur Loss: 0.10582432, Cur Avg Loss: 0.13791192, Log Avg loss: 0.13092968, Global Avg Loss: 0.59507813, Time: 0.0209 Steps: 92450, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000913, Sample Num: 14608, Cur Loss: 0.22374651, Cur Avg Loss: 0.13857316, Log Avg loss: 0.19828282, Global Avg Loss: 0.59503521, Time: 0.0209 Steps: 92460, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000923, Sample Num: 14768, Cur Loss: 0.19147554, Cur Avg Loss: 0.13870460, Log Avg loss: 0.15070461, Global Avg Loss: 0.59498716, Time: 0.0209 Steps: 92470, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000933, Sample Num: 14928, Cur Loss: 0.03506754, Cur Avg Loss: 0.13824959, Log Avg loss: 0.09625290, Global Avg Loss: 0.59493323, Time: 0.0210 Steps: 92480, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000943, Sample Num: 15088, Cur Loss: 0.10077866, Cur Avg Loss: 0.13850260, Log Avg loss: 0.16210813, Global Avg Loss: 0.59488644, Time: 0.0209 Steps: 92490, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000953, Sample Num: 15248, Cur Loss: 0.07994019, Cur Avg Loss: 0.13827470, Log Avg loss: 0.11678345, Global Avg Loss: 0.59483475, Time: 0.0209 Steps: 92500, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000963, Sample Num: 15408, Cur Loss: 0.06117446, Cur Avg Loss: 0.13797278, Log Avg loss: 0.10920038, Global Avg Loss: 0.59478225, Time: 0.0209 Steps: 92510, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000973, Sample Num: 15568, Cur Loss: 0.27793998, Cur Avg Loss: 0.13811270, Log Avg loss: 0.15158647, Global Avg Loss: 0.59473435, Time: 0.0209 Steps: 92520, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000983, Sample Num: 15728, Cur Loss: 0.20665768, Cur Avg Loss: 0.13773674, Log Avg loss: 0.10115578, Global Avg Loss: 0.59468101, Time: 0.0209 Steps: 92530, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000993, Sample Num: 15888, Cur Loss: 0.18333474, Cur Avg Loss: 0.13774389, Log Avg loss: 0.13844648, Global Avg Loss: 0.59463171, Time: 0.0210 Steps: 92540, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001003, Sample Num: 16048, Cur Loss: 0.22979294, Cur Avg Loss: 0.13770024, Log Avg loss: 0.13336654, Global Avg Loss: 0.59458187, Time: 0.0210 Steps: 92550, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001013, Sample Num: 16208, Cur Loss: 0.07413039, Cur Avg Loss: 0.13784036, Log Avg loss: 0.15189410, Global Avg Loss: 0.59453404, Time: 0.0210 Steps: 92560, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001023, Sample Num: 16368, Cur Loss: 0.06059377, Cur Avg Loss: 0.13818818, Log Avg loss: 0.17342199, Global Avg Loss: 0.59448855, Time: 0.0210 Steps: 92570, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001033, Sample Num: 16528, Cur Loss: 0.18158434, Cur Avg Loss: 0.13845014, Log Avg loss: 0.16524865, Global Avg Loss: 0.59444219, Time: 0.0210 Steps: 92580, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001043, Sample Num: 16688, Cur Loss: 0.03991914, Cur Avg Loss: 0.13799558, Log Avg loss: 0.09103962, Global Avg Loss: 0.59438782, Time: 0.0210 Steps: 92590, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001053, Sample Num: 16848, Cur Loss: 0.12599598, Cur Avg Loss: 0.13797859, Log Avg loss: 0.13620664, Global Avg Loss: 0.59433834, Time: 0.0209 Steps: 92600, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001063, Sample Num: 17008, Cur Loss: 0.14464384, Cur Avg Loss: 0.13808156, Log Avg loss: 0.14892413, Global Avg Loss: 0.59429024, Time: 0.0210 Steps: 92610, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001073, Sample Num: 17168, Cur Loss: 0.24455899, Cur Avg Loss: 0.13810546, Log Avg loss: 0.14064661, Global Avg Loss: 0.59424126, Time: 0.0209 Steps: 92620, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001083, Sample Num: 17328, Cur Loss: 0.06883413, Cur Avg Loss: 0.13829820, Log Avg loss: 0.15897866, Global Avg Loss: 0.59419427, Time: 0.0209 Steps: 92630, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001093, Sample Num: 17488, Cur Loss: 0.18853945, Cur Avg Loss: 0.13808363, Log Avg loss: 0.11484597, Global Avg Loss: 0.59414253, Time: 0.0209 Steps: 92640, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001103, Sample Num: 17648, Cur Loss: 0.07983039, Cur Avg Loss: 0.13780652, Log Avg loss: 0.10751797, Global Avg Loss: 0.59409001, Time: 0.0210 Steps: 92650, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001113, Sample Num: 17808, Cur Loss: 0.22122440, Cur Avg Loss: 0.13795245, Log Avg loss: 0.15404919, Global Avg Loss: 0.59404252, Time: 0.0209 Steps: 92660, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001123, Sample Num: 17968, Cur Loss: 0.19116476, Cur Avg Loss: 0.13833519, Log Avg loss: 0.18093395, Global Avg Loss: 0.59399794, Time: 0.0209 Steps: 92670, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001133, Sample Num: 18128, Cur Loss: 0.29629567, Cur Avg Loss: 0.13829441, Log Avg loss: 0.13371472, Global Avg Loss: 0.59394827, Time: 0.0210 Steps: 92680, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001143, Sample Num: 18288, Cur Loss: 0.18294923, Cur Avg Loss: 0.13844751, Log Avg loss: 0.15579418, Global Avg Loss: 0.59390100, Time: 0.0209 Steps: 92690, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001153, Sample Num: 18448, Cur Loss: 0.46763656, Cur Avg Loss: 0.13869279, Log Avg loss: 0.16672803, Global Avg Loss: 0.59385492, Time: 0.0209 Steps: 92700, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001163, Sample Num: 18608, Cur Loss: 0.04703394, Cur Avg Loss: 0.13865078, Log Avg loss: 0.13380678, Global Avg Loss: 0.59380530, Time: 0.0210 Steps: 92710, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001173, Sample Num: 18768, Cur Loss: 0.12961477, Cur Avg Loss: 0.13900210, Log Avg loss: 0.17986089, Global Avg Loss: 0.59376066, Time: 0.0209 Steps: 92720, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001183, Sample Num: 18928, Cur Loss: 0.05811981, Cur Avg Loss: 0.13928650, Log Avg loss: 0.17264663, Global Avg Loss: 0.59371524, Time: 0.0209 Steps: 92730, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001193, Sample Num: 19088, Cur Loss: 0.17754082, Cur Avg Loss: 0.13933966, Log Avg loss: 0.14562859, Global Avg Loss: 0.59366693, Time: 0.0210 Steps: 92740, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001203, Sample Num: 19248, Cur Loss: 0.14641654, Cur Avg Loss: 0.13949512, Log Avg loss: 0.15804105, Global Avg Loss: 0.59361996, Time: 0.0209 Steps: 92750, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001213, Sample Num: 19408, Cur Loss: 0.40783378, Cur Avg Loss: 0.13971199, Log Avg loss: 0.16580156, Global Avg Loss: 0.59357384, Time: 0.0209 Steps: 92760, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001223, Sample Num: 19568, Cur Loss: 0.09702762, Cur Avg Loss: 0.13964993, Log Avg loss: 0.13212225, Global Avg Loss: 0.59352410, Time: 0.0210 Steps: 92770, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001233, Sample Num: 19728, Cur Loss: 0.04267213, Cur Avg Loss: 0.13965420, Log Avg loss: 0.14017606, Global Avg Loss: 0.59347523, Time: 0.0209 Steps: 92780, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001243, Sample Num: 19888, Cur Loss: 0.23518981, Cur Avg Loss: 0.13948845, Log Avg loss: 0.11905135, Global Avg Loss: 0.59342410, Time: 0.0209 Steps: 92790, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001253, Sample Num: 20048, Cur Loss: 0.12052163, Cur Avg Loss: 0.13999995, Log Avg loss: 0.20357963, Global Avg Loss: 0.59338210, Time: 0.0210 Steps: 92800, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001263, Sample Num: 20208, Cur Loss: 0.03613281, Cur Avg Loss: 0.13948461, Log Avg loss: 0.07491278, Global Avg Loss: 0.59332623, Time: 0.0210 Steps: 92810, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001273, Sample Num: 20368, Cur Loss: 0.15490943, Cur Avg Loss: 0.13985465, Log Avg loss: 0.18659056, Global Avg Loss: 0.59328241, Time: 0.0209 Steps: 92820, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001283, Sample Num: 20528, Cur Loss: 0.07712452, Cur Avg Loss: 0.13956416, Log Avg loss: 0.10258506, Global Avg Loss: 0.59322955, Time: 0.0246 Steps: 92830, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001293, Sample Num: 20688, Cur Loss: 0.08409709, Cur Avg Loss: 0.13920335, Log Avg loss: 0.09291125, Global Avg Loss: 0.59317566, Time: 0.0209 Steps: 92840, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001303, Sample Num: 20848, Cur Loss: 0.37023464, Cur Avg Loss: 0.13907746, Log Avg loss: 0.12279960, Global Avg Loss: 0.59312500, Time: 0.0210 Steps: 92850, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001313, Sample Num: 21008, Cur Loss: 0.16587386, Cur Avg Loss: 0.13903684, Log Avg loss: 0.13374387, Global Avg Loss: 0.59307553, Time: 0.0210 Steps: 92860, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001323, Sample Num: 21168, Cur Loss: 0.29196674, Cur Avg Loss: 0.13930201, Log Avg loss: 0.17411980, Global Avg Loss: 0.59303042, Time: 0.0210 Steps: 92870, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001333, Sample Num: 21328, Cur Loss: 0.16865188, Cur Avg Loss: 0.13924870, Log Avg loss: 0.13219570, Global Avg Loss: 0.59298080, Time: 0.0209 Steps: 92880, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001343, Sample Num: 21488, Cur Loss: 0.08798317, Cur Avg Loss: 0.13941427, Log Avg loss: 0.16148428, Global Avg Loss: 0.59293435, Time: 0.0209 Steps: 92890, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001353, Sample Num: 21648, Cur Loss: 0.20817573, Cur Avg Loss: 0.13929035, Log Avg loss: 0.12264784, Global Avg Loss: 0.59288373, Time: 0.0209 Steps: 92900, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001363, Sample Num: 21808, Cur Loss: 0.07430704, Cur Avg Loss: 0.13915370, Log Avg loss: 0.12066502, Global Avg Loss: 0.59283290, Time: 0.0209 Steps: 92910, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001373, Sample Num: 21968, Cur Loss: 0.18706343, Cur Avg Loss: 0.13905387, Log Avg loss: 0.12544751, Global Avg Loss: 0.59278260, Time: 0.0209 Steps: 92920, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001383, Sample Num: 22128, Cur Loss: 0.14876592, Cur Avg Loss: 0.13927599, Log Avg loss: 0.16977283, Global Avg Loss: 0.59273708, Time: 0.0209 Steps: 92930, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001393, Sample Num: 22288, Cur Loss: 0.03697162, Cur Avg Loss: 0.13919165, Log Avg loss: 0.12752752, Global Avg Loss: 0.59268703, Time: 0.0209 Steps: 92940, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001403, Sample Num: 22448, Cur Loss: 0.15187731, Cur Avg Loss: 0.13891907, Log Avg loss: 0.10094771, Global Avg Loss: 0.59263413, Time: 0.0209 Steps: 92950, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001413, Sample Num: 22608, Cur Loss: 0.11079163, Cur Avg Loss: 0.13887722, Log Avg loss: 0.13300599, Global Avg Loss: 0.59258468, Time: 0.0209 Steps: 92960, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001423, Sample Num: 22768, Cur Loss: 0.08534259, Cur Avg Loss: 0.13879903, Log Avg loss: 0.12775073, Global Avg Loss: 0.59253468, Time: 0.0210 Steps: 92970, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001433, Sample Num: 22928, Cur Loss: 0.09967045, Cur Avg Loss: 0.13889693, Log Avg loss: 0.15282789, Global Avg Loss: 0.59248739, Time: 0.0210 Steps: 92980, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001443, Sample Num: 23088, Cur Loss: 0.13275531, Cur Avg Loss: 0.13881566, Log Avg loss: 0.12717035, Global Avg Loss: 0.59243735, Time: 0.0209 Steps: 92990, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001453, Sample Num: 23248, Cur Loss: 0.64218807, Cur Avg Loss: 0.13920862, Log Avg loss: 0.19591262, Global Avg Loss: 0.59239472, Time: 0.0209 Steps: 93000, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001463, Sample Num: 23408, Cur Loss: 0.18041050, Cur Avg Loss: 0.13968137, Log Avg loss: 0.20837179, Global Avg Loss: 0.59235343, Time: 0.0209 Steps: 93010, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001473, Sample Num: 23568, Cur Loss: 0.17899674, Cur Avg Loss: 0.13985772, Log Avg loss: 0.16565759, Global Avg Loss: 0.59230756, Time: 0.0209 Steps: 93020, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001483, Sample Num: 23728, Cur Loss: 0.06794888, Cur Avg Loss: 0.13970274, Log Avg loss: 0.11687426, Global Avg Loss: 0.59225645, Time: 0.0209 Steps: 93030, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001493, Sample Num: 23888, Cur Loss: 0.02858867, Cur Avg Loss: 0.14002761, Log Avg loss: 0.18820637, Global Avg Loss: 0.59221302, Time: 0.0210 Steps: 93040, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001503, Sample Num: 24048, Cur Loss: 0.18465057, Cur Avg Loss: 0.13994233, Log Avg loss: 0.12720997, Global Avg Loss: 0.59216305, Time: 0.0209 Steps: 93050, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001513, Sample Num: 24208, Cur Loss: 0.10316955, Cur Avg Loss: 0.14024312, Log Avg loss: 0.18545130, Global Avg Loss: 0.59211935, Time: 0.0209 Steps: 93060, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001523, Sample Num: 24368, Cur Loss: 0.09361551, Cur Avg Loss: 0.14023562, Log Avg loss: 0.13910103, Global Avg Loss: 0.59207067, Time: 0.0209 Steps: 93070, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001533, Sample Num: 24528, Cur Loss: 0.24089983, Cur Avg Loss: 0.14068590, Log Avg loss: 0.20926391, Global Avg Loss: 0.59202954, Time: 0.0209 Steps: 93080, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001543, Sample Num: 24688, Cur Loss: 0.06496824, Cur Avg Loss: 0.14061536, Log Avg loss: 0.12980158, Global Avg Loss: 0.59197989, Time: 0.0213 Steps: 93090, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001553, Sample Num: 24848, Cur Loss: 0.13759692, Cur Avg Loss: 0.14103254, Log Avg loss: 0.20540280, Global Avg Loss: 0.59193837, Time: 0.0209 Steps: 93100, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001563, Sample Num: 25008, Cur Loss: 0.23058173, Cur Avg Loss: 0.14173839, Log Avg loss: 0.25135798, Global Avg Loss: 0.59190179, Time: 0.0210 Steps: 93110, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001573, Sample Num: 25168, Cur Loss: 0.09116784, Cur Avg Loss: 0.14186507, Log Avg loss: 0.16166478, Global Avg Loss: 0.59185559, Time: 0.0209 Steps: 93120, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001583, Sample Num: 25328, Cur Loss: 0.16909209, Cur Avg Loss: 0.14172334, Log Avg loss: 0.11942817, Global Avg Loss: 0.59180486, Time: 0.0209 Steps: 93130, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001593, Sample Num: 25488, Cur Loss: 0.03452040, Cur Avg Loss: 0.14152660, Log Avg loss: 0.11038397, Global Avg Loss: 0.59175317, Time: 0.0209 Steps: 93140, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001603, Sample Num: 25648, Cur Loss: 0.23370022, Cur Avg Loss: 0.14172132, Log Avg loss: 0.17273941, Global Avg Loss: 0.59170819, Time: 0.0209 Steps: 93150, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001613, Sample Num: 25808, Cur Loss: 0.14293066, Cur Avg Loss: 0.14158688, Log Avg loss: 0.12003663, Global Avg Loss: 0.59165756, Time: 0.0209 Steps: 93160, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001623, Sample Num: 25968, Cur Loss: 0.10155042, Cur Avg Loss: 0.14157242, Log Avg loss: 0.13924035, Global Avg Loss: 0.59160900, Time: 0.0209 Steps: 93170, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001633, Sample Num: 26128, Cur Loss: 0.04743839, Cur Avg Loss: 0.14167469, Log Avg loss: 0.15827186, Global Avg Loss: 0.59156250, Time: 0.0210 Steps: 93180, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001643, Sample Num: 26288, Cur Loss: 0.25348383, Cur Avg Loss: 0.14174650, Log Avg loss: 0.15347406, Global Avg Loss: 0.59151549, Time: 0.0210 Steps: 93190, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001653, Sample Num: 26448, Cur Loss: 0.07360540, Cur Avg Loss: 0.14182748, Log Avg loss: 0.15513269, Global Avg Loss: 0.59146866, Time: 0.0209 Steps: 93200, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001663, Sample Num: 26608, Cur Loss: 0.13777235, Cur Avg Loss: 0.14159010, Log Avg loss: 0.10235009, Global Avg Loss: 0.59141619, Time: 0.0209 Steps: 93210, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001673, Sample Num: 26768, Cur Loss: 0.16597798, Cur Avg Loss: 0.14133045, Log Avg loss: 0.09815166, Global Avg Loss: 0.59136327, Time: 0.0209 Steps: 93220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001683, Sample Num: 26928, Cur Loss: 0.07784817, Cur Avg Loss: 0.14137133, Log Avg loss: 0.14821050, Global Avg Loss: 0.59131574, Time: 0.0209 Steps: 93230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001693, Sample Num: 27088, Cur Loss: 0.10660680, Cur Avg Loss: 0.14124426, Log Avg loss: 0.11985797, Global Avg Loss: 0.59126518, Time: 0.0210 Steps: 93240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001703, Sample Num: 27248, Cur Loss: 0.06238338, Cur Avg Loss: 0.14116823, Log Avg loss: 0.12829691, Global Avg Loss: 0.59121553, Time: 0.0210 Steps: 93250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001713, Sample Num: 27408, Cur Loss: 0.06921540, Cur Avg Loss: 0.14114158, Log Avg loss: 0.13660310, Global Avg Loss: 0.59116678, Time: 0.0209 Steps: 93260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001723, Sample Num: 27568, Cur Loss: 0.05043603, Cur Avg Loss: 0.14111921, Log Avg loss: 0.13728582, Global Avg Loss: 0.59111812, Time: 0.0209 Steps: 93270, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001733, Sample Num: 27728, Cur Loss: 0.15795141, Cur Avg Loss: 0.14115288, Log Avg loss: 0.14695449, Global Avg Loss: 0.59107050, Time: 0.0209 Steps: 93280, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001743, Sample Num: 27888, Cur Loss: 0.09561902, Cur Avg Loss: 0.14099749, Log Avg loss: 0.11406869, Global Avg Loss: 0.59101937, Time: 0.0209 Steps: 93290, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001753, Sample Num: 28048, Cur Loss: 0.16424163, Cur Avg Loss: 0.14084252, Log Avg loss: 0.11383163, Global Avg Loss: 0.59096823, Time: 0.0209 Steps: 93300, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001763, Sample Num: 28208, Cur Loss: 0.03288002, Cur Avg Loss: 0.14073274, Log Avg loss: 0.12148732, Global Avg Loss: 0.59091791, Time: 0.0209 Steps: 93310, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001773, Sample Num: 28368, Cur Loss: 0.25922880, Cur Avg Loss: 0.14079314, Log Avg loss: 0.15144308, Global Avg Loss: 0.59087082, Time: 0.0209 Steps: 93320, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001783, Sample Num: 28528, Cur Loss: 0.32873252, Cur Avg Loss: 0.14070819, Log Avg loss: 0.12564572, Global Avg Loss: 0.59082097, Time: 0.0209 Steps: 93330, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001793, Sample Num: 28688, Cur Loss: 0.21640307, Cur Avg Loss: 0.14097701, Log Avg loss: 0.18890734, Global Avg Loss: 0.59077791, Time: 0.0245 Steps: 93340, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001803, Sample Num: 28848, Cur Loss: 0.01779665, Cur Avg Loss: 0.14096095, Log Avg loss: 0.13808220, Global Avg Loss: 0.59072942, Time: 0.0208 Steps: 93350, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001813, Sample Num: 29008, Cur Loss: 0.14657664, Cur Avg Loss: 0.14107447, Log Avg loss: 0.16154276, Global Avg Loss: 0.59068345, Time: 0.0210 Steps: 93360, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001823, Sample Num: 29168, Cur Loss: 0.35435623, Cur Avg Loss: 0.14133033, Log Avg loss: 0.18771635, Global Avg Loss: 0.59064029, Time: 0.0209 Steps: 93370, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001833, Sample Num: 29328, Cur Loss: 0.07519417, Cur Avg Loss: 0.14149841, Log Avg loss: 0.17214073, Global Avg Loss: 0.59059547, Time: 0.0209 Steps: 93380, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001843, Sample Num: 29488, Cur Loss: 0.14818768, Cur Avg Loss: 0.14132036, Log Avg loss: 0.10868240, Global Avg Loss: 0.59054387, Time: 0.0209 Steps: 93390, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001853, Sample Num: 29648, Cur Loss: 0.06581325, Cur Avg Loss: 0.14149077, Log Avg loss: 0.17289769, Global Avg Loss: 0.59049915, Time: 0.0209 Steps: 93400, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001863, Sample Num: 29808, Cur Loss: 0.11735968, Cur Avg Loss: 0.14176822, Log Avg loss: 0.19317903, Global Avg Loss: 0.59045662, Time: 0.0210 Steps: 93410, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001873, Sample Num: 29968, Cur Loss: 0.42546624, Cur Avg Loss: 0.14199074, Log Avg loss: 0.18344722, Global Avg Loss: 0.59041305, Time: 0.0209 Steps: 93420, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001883, Sample Num: 30128, Cur Loss: 0.10364050, Cur Avg Loss: 0.14187516, Log Avg loss: 0.12022726, Global Avg Loss: 0.59036273, Time: 0.0209 Steps: 93430, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001893, Sample Num: 30288, Cur Loss: 0.06090031, Cur Avg Loss: 0.14178368, Log Avg loss: 0.12455761, Global Avg Loss: 0.59031288, Time: 0.0209 Steps: 93440, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001903, Sample Num: 30448, Cur Loss: 0.15792117, Cur Avg Loss: 0.14195194, Log Avg loss: 0.17380344, Global Avg Loss: 0.59026830, Time: 0.0209 Steps: 93450, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001913, Sample Num: 30608, Cur Loss: 0.25661334, Cur Avg Loss: 0.14199734, Log Avg loss: 0.15063736, Global Avg Loss: 0.59022127, Time: 0.0210 Steps: 93460, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001923, Sample Num: 30768, Cur Loss: 0.01650188, Cur Avg Loss: 0.14174914, Log Avg loss: 0.09426828, Global Avg Loss: 0.59016821, Time: 0.0210 Steps: 93470, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001933, Sample Num: 30928, Cur Loss: 0.41602641, Cur Avg Loss: 0.14160689, Log Avg loss: 0.11425233, Global Avg Loss: 0.59011729, Time: 0.0209 Steps: 93480, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001943, Sample Num: 31088, Cur Loss: 0.02261702, Cur Avg Loss: 0.14176703, Log Avg loss: 0.17272107, Global Avg Loss: 0.59007265, Time: 0.0209 Steps: 93490, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001953, Sample Num: 31248, Cur Loss: 0.27796641, Cur Avg Loss: 0.14169421, Log Avg loss: 0.12754585, Global Avg Loss: 0.59002318, Time: 0.0210 Steps: 93500, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001963, Sample Num: 31408, Cur Loss: 0.15876089, Cur Avg Loss: 0.14174358, Log Avg loss: 0.15138534, Global Avg Loss: 0.58997627, Time: 0.0210 Steps: 93510, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001973, Sample Num: 31568, Cur Loss: 0.15606043, Cur Avg Loss: 0.14202626, Log Avg loss: 0.19751710, Global Avg Loss: 0.58993431, Time: 0.0210 Steps: 93520, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001983, Sample Num: 31728, Cur Loss: 0.06975517, Cur Avg Loss: 0.14202089, Log Avg loss: 0.14096121, Global Avg Loss: 0.58988630, Time: 0.0209 Steps: 93530, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001993, Sample Num: 31888, Cur Loss: 0.07591095, Cur Avg Loss: 0.14179319, Log Avg loss: 0.09663938, Global Avg Loss: 0.58983357, Time: 0.0210 Steps: 93540, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002003, Sample Num: 32048, Cur Loss: 0.06334744, Cur Avg Loss: 0.14183178, Log Avg loss: 0.14952358, Global Avg Loss: 0.58978651, Time: 0.0209 Steps: 93550, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002013, Sample Num: 32208, Cur Loss: 0.29101056, Cur Avg Loss: 0.14202391, Log Avg loss: 0.18050795, Global Avg Loss: 0.58974276, Time: 0.0210 Steps: 93560, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002023, Sample Num: 32368, Cur Loss: 0.10062505, Cur Avg Loss: 0.14214574, Log Avg loss: 0.16666966, Global Avg Loss: 0.58969755, Time: 0.0209 Steps: 93570, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002033, Sample Num: 32528, Cur Loss: 0.08947854, Cur Avg Loss: 0.14215850, Log Avg loss: 0.14474084, Global Avg Loss: 0.58965000, Time: 0.0209 Steps: 93580, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002043, Sample Num: 32688, Cur Loss: 0.14743283, Cur Avg Loss: 0.14208420, Log Avg loss: 0.12697723, Global Avg Loss: 0.58960056, Time: 0.0209 Steps: 93590, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002053, Sample Num: 32848, Cur Loss: 0.08073831, Cur Avg Loss: 0.14190891, Log Avg loss: 0.10609887, Global Avg Loss: 0.58954891, Time: 0.0247 Steps: 93600, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002063, Sample Num: 33008, Cur Loss: 0.32018870, Cur Avg Loss: 0.14182774, Log Avg loss: 0.12516234, Global Avg Loss: 0.58949930, Time: 0.0209 Steps: 93610, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002073, Sample Num: 33168, Cur Loss: 0.25007591, Cur Avg Loss: 0.14184655, Log Avg loss: 0.14572802, Global Avg Loss: 0.58945190, Time: 0.0209 Steps: 93620, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002083, Sample Num: 33328, Cur Loss: 0.37073836, Cur Avg Loss: 0.14191303, Log Avg loss: 0.15569275, Global Avg Loss: 0.58940557, Time: 0.0209 Steps: 93630, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002093, Sample Num: 33488, Cur Loss: 0.15766031, Cur Avg Loss: 0.14176217, Log Avg loss: 0.11033909, Global Avg Loss: 0.58935441, Time: 0.0209 Steps: 93640, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002103, Sample Num: 33648, Cur Loss: 0.23585534, Cur Avg Loss: 0.14156227, Log Avg loss: 0.09972204, Global Avg Loss: 0.58930212, Time: 0.0211 Steps: 93650, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002113, Sample Num: 33808, Cur Loss: 0.02649392, Cur Avg Loss: 0.14156916, Log Avg loss: 0.14301912, Global Avg Loss: 0.58925448, Time: 0.0209 Steps: 93660, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002123, Sample Num: 33968, Cur Loss: 0.08489397, Cur Avg Loss: 0.14146295, Log Avg loss: 0.11902094, Global Avg Loss: 0.58920427, Time: 0.0209 Steps: 93670, Updated lr: 0.000012 ***** Running evaluation checkpoint-93676 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-93676 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.702730, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.269961, "eval_total_loss": 189.782478, "eval_mae": 0.351752, "eval_mse": 0.270065, "eval_r2": 0.828329, "eval_sp_statistic": 0.931747, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.935686, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.18598, "test_total_loss": 93.362151, "test_mae": 0.312792, "test_mse": 0.186046, "test_r2": 0.879924, "test_sp_statistic": 0.919385, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946452, "test_ps_pvalue": 0.0, "lr": 1.2113798008534852e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5891776830211283, "train_cur_epoch_loss": 301.3700875751674, "train_cur_epoch_avg_loss": 0.1415547616604826, "train_cur_epoch_time": 44.70273017883301, "train_cur_epoch_avg_time": 0.020997055039376707, "epoch": 44, "step": 93676} ################################################## Training, Epoch: 0045, Batch: 000004, Sample Num: 64, Cur Loss: 0.24903062, Cur Avg Loss: 0.17721296, Log Avg loss: 0.17530953, Global Avg Loss: 0.58916009, Time: 0.0248 Steps: 93680, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000014, Sample Num: 224, Cur Loss: 0.18244538, Cur Avg Loss: 0.17815521, Log Avg loss: 0.17853211, Global Avg Loss: 0.58911626, Time: 0.0210 Steps: 93690, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000024, Sample Num: 384, Cur Loss: 0.12826052, Cur Avg Loss: 0.15968519, Log Avg loss: 0.13382717, Global Avg Loss: 0.58906767, Time: 0.0210 Steps: 93700, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000034, Sample Num: 544, Cur Loss: 0.03431676, Cur Avg Loss: 0.15279026, Log Avg loss: 0.13624241, Global Avg Loss: 0.58901935, Time: 0.0210 Steps: 93710, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000044, Sample Num: 704, Cur Loss: 0.14528368, Cur Avg Loss: 0.16649500, Log Avg loss: 0.21309113, Global Avg Loss: 0.58897924, Time: 0.0210 Steps: 93720, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000054, Sample Num: 864, Cur Loss: 0.24012290, Cur Avg Loss: 0.15735586, Log Avg loss: 0.11714363, Global Avg Loss: 0.58892890, Time: 0.0210 Steps: 93730, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000064, Sample Num: 1024, Cur Loss: 0.09925645, Cur Avg Loss: 0.15325701, Log Avg loss: 0.13112322, Global Avg Loss: 0.58888006, Time: 0.0210 Steps: 93740, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000074, Sample Num: 1184, Cur Loss: 0.04622606, Cur Avg Loss: 0.15271269, Log Avg loss: 0.14922904, Global Avg Loss: 0.58883317, Time: 0.0210 Steps: 93750, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000084, Sample Num: 1344, Cur Loss: 0.04518157, Cur Avg Loss: 0.15566821, Log Avg loss: 0.17753903, Global Avg Loss: 0.58878930, Time: 0.0211 Steps: 93760, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000094, Sample Num: 1504, Cur Loss: 0.20808661, Cur Avg Loss: 0.15212396, Log Avg loss: 0.12235232, Global Avg Loss: 0.58873956, Time: 0.0210 Steps: 93770, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000104, Sample Num: 1664, Cur Loss: 0.13004728, Cur Avg Loss: 0.15116698, Log Avg loss: 0.14217136, Global Avg Loss: 0.58869194, Time: 0.0211 Steps: 93780, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000114, Sample Num: 1824, Cur Loss: 0.09906048, Cur Avg Loss: 0.15111879, Log Avg loss: 0.15061756, Global Avg Loss: 0.58864523, Time: 0.0210 Steps: 93790, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000124, Sample Num: 1984, Cur Loss: 0.10717856, Cur Avg Loss: 0.14756786, Log Avg loss: 0.10708729, Global Avg Loss: 0.58859389, Time: 0.0211 Steps: 93800, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000134, Sample Num: 2144, Cur Loss: 0.17328092, Cur Avg Loss: 0.14975743, Log Avg loss: 0.17690812, Global Avg Loss: 0.58855001, Time: 0.0210 Steps: 93810, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000144, Sample Num: 2304, Cur Loss: 0.13421658, Cur Avg Loss: 0.14738428, Log Avg loss: 0.11558401, Global Avg Loss: 0.58849959, Time: 0.0211 Steps: 93820, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000154, Sample Num: 2464, Cur Loss: 0.10326807, Cur Avg Loss: 0.14717256, Log Avg loss: 0.14412378, Global Avg Loss: 0.58845223, Time: 0.0211 Steps: 93830, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000164, Sample Num: 2624, Cur Loss: 0.07910492, Cur Avg Loss: 0.14740194, Log Avg loss: 0.15093449, Global Avg Loss: 0.58840561, Time: 0.0210 Steps: 93840, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000174, Sample Num: 2784, Cur Loss: 0.12786791, Cur Avg Loss: 0.14493305, Log Avg loss: 0.10444318, Global Avg Loss: 0.58835404, Time: 0.0210 Steps: 93850, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000184, Sample Num: 2944, Cur Loss: 0.03399187, Cur Avg Loss: 0.14057896, Log Avg loss: 0.06481780, Global Avg Loss: 0.58829827, Time: 0.0210 Steps: 93860, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000194, Sample Num: 3104, Cur Loss: 0.15980345, Cur Avg Loss: 0.13996782, Log Avg loss: 0.12872277, Global Avg Loss: 0.58824931, Time: 0.0210 Steps: 93870, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000204, Sample Num: 3264, Cur Loss: 0.35014802, Cur Avg Loss: 0.13977856, Log Avg loss: 0.13610711, Global Avg Loss: 0.58820114, Time: 0.0210 Steps: 93880, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000214, Sample Num: 3424, Cur Loss: 0.09578442, Cur Avg Loss: 0.13909798, Log Avg loss: 0.12521403, Global Avg Loss: 0.58815183, Time: 0.0210 Steps: 93890, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000224, Sample Num: 3584, Cur Loss: 0.05051467, Cur Avg Loss: 0.13709747, Log Avg loss: 0.09428666, Global Avg Loss: 0.58809924, Time: 0.0211 Steps: 93900, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000234, Sample Num: 3744, Cur Loss: 0.02753372, Cur Avg Loss: 0.13421073, Log Avg loss: 0.06954756, Global Avg Loss: 0.58804402, Time: 0.0210 Steps: 93910, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000244, Sample Num: 3904, Cur Loss: 0.04522458, Cur Avg Loss: 0.13453815, Log Avg loss: 0.14219998, Global Avg Loss: 0.58799655, Time: 0.0211 Steps: 93920, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000254, Sample Num: 4064, Cur Loss: 0.07426029, Cur Avg Loss: 0.13507884, Log Avg loss: 0.14827162, Global Avg Loss: 0.58794974, Time: 0.0210 Steps: 93930, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000264, Sample Num: 4224, Cur Loss: 0.10981039, Cur Avg Loss: 0.13469706, Log Avg loss: 0.12499979, Global Avg Loss: 0.58790045, Time: 0.0209 Steps: 93940, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000274, Sample Num: 4384, Cur Loss: 0.12530124, Cur Avg Loss: 0.13363148, Log Avg loss: 0.10550006, Global Avg Loss: 0.58784911, Time: 0.0209 Steps: 93950, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000284, Sample Num: 4544, Cur Loss: 0.20515652, Cur Avg Loss: 0.13425494, Log Avg loss: 0.15133789, Global Avg Loss: 0.58780265, Time: 0.0209 Steps: 93960, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000294, Sample Num: 4704, Cur Loss: 0.09040810, Cur Avg Loss: 0.13356586, Log Avg loss: 0.11399584, Global Avg Loss: 0.58775223, Time: 0.0210 Steps: 93970, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000304, Sample Num: 4864, Cur Loss: 0.13535529, Cur Avg Loss: 0.13225529, Log Avg loss: 0.09372472, Global Avg Loss: 0.58769966, Time: 0.0209 Steps: 93980, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000314, Sample Num: 5024, Cur Loss: 0.21903129, Cur Avg Loss: 0.13414276, Log Avg loss: 0.19152167, Global Avg Loss: 0.58765751, Time: 0.0209 Steps: 93990, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000324, Sample Num: 5184, Cur Loss: 0.07959394, Cur Avg Loss: 0.13568110, Log Avg loss: 0.18398499, Global Avg Loss: 0.58761457, Time: 0.0209 Steps: 94000, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000334, Sample Num: 5344, Cur Loss: 0.06558477, Cur Avg Loss: 0.13722417, Log Avg loss: 0.18721990, Global Avg Loss: 0.58757198, Time: 0.0210 Steps: 94010, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000344, Sample Num: 5504, Cur Loss: 0.07783462, Cur Avg Loss: 0.13693286, Log Avg loss: 0.12720309, Global Avg Loss: 0.58752301, Time: 0.0210 Steps: 94020, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000354, Sample Num: 5664, Cur Loss: 0.06479216, Cur Avg Loss: 0.13591116, Log Avg loss: 0.10076443, Global Avg Loss: 0.58747125, Time: 0.0209 Steps: 94030, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000364, Sample Num: 5824, Cur Loss: 0.07832778, Cur Avg Loss: 0.13444544, Log Avg loss: 0.08255916, Global Avg Loss: 0.58741755, Time: 0.0209 Steps: 94040, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000374, Sample Num: 5984, Cur Loss: 0.13980220, Cur Avg Loss: 0.13415721, Log Avg loss: 0.12366543, Global Avg Loss: 0.58736824, Time: 0.0209 Steps: 94050, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000384, Sample Num: 6144, Cur Loss: 0.20373988, Cur Avg Loss: 0.13444334, Log Avg loss: 0.14514461, Global Avg Loss: 0.58732123, Time: 0.0209 Steps: 94060, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000394, Sample Num: 6304, Cur Loss: 0.26197758, Cur Avg Loss: 0.13426432, Log Avg loss: 0.12739014, Global Avg Loss: 0.58727234, Time: 0.0209 Steps: 94070, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000404, Sample Num: 6464, Cur Loss: 0.04525670, Cur Avg Loss: 0.13416212, Log Avg loss: 0.13013535, Global Avg Loss: 0.58722375, Time: 0.0210 Steps: 94080, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000414, Sample Num: 6624, Cur Loss: 0.23321445, Cur Avg Loss: 0.13614646, Log Avg loss: 0.21631389, Global Avg Loss: 0.58718433, Time: 0.0209 Steps: 94090, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000424, Sample Num: 6784, Cur Loss: 0.05053276, Cur Avg Loss: 0.13512563, Log Avg loss: 0.09286326, Global Avg Loss: 0.58713179, Time: 0.0209 Steps: 94100, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000434, Sample Num: 6944, Cur Loss: 0.08816274, Cur Avg Loss: 0.13499886, Log Avg loss: 0.12962384, Global Avg Loss: 0.58708318, Time: 0.0208 Steps: 94110, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000444, Sample Num: 7104, Cur Loss: 0.17593051, Cur Avg Loss: 0.13504823, Log Avg loss: 0.13719098, Global Avg Loss: 0.58703538, Time: 0.0210 Steps: 94120, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000454, Sample Num: 7264, Cur Loss: 0.04556653, Cur Avg Loss: 0.13483699, Log Avg loss: 0.12545750, Global Avg Loss: 0.58698634, Time: 0.0209 Steps: 94130, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000464, Sample Num: 7424, Cur Loss: 0.09755372, Cur Avg Loss: 0.13488963, Log Avg loss: 0.13727967, Global Avg Loss: 0.58693857, Time: 0.0209 Steps: 94140, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000474, Sample Num: 7584, Cur Loss: 0.03139722, Cur Avg Loss: 0.13481261, Log Avg loss: 0.13123913, Global Avg Loss: 0.58689017, Time: 0.0209 Steps: 94150, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000484, Sample Num: 7744, Cur Loss: 0.07786620, Cur Avg Loss: 0.13470374, Log Avg loss: 0.12954296, Global Avg Loss: 0.58684160, Time: 0.0210 Steps: 94160, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000494, Sample Num: 7904, Cur Loss: 0.07150196, Cur Avg Loss: 0.13455392, Log Avg loss: 0.12730297, Global Avg Loss: 0.58679280, Time: 0.0210 Steps: 94170, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000504, Sample Num: 8064, Cur Loss: 0.17014179, Cur Avg Loss: 0.13451461, Log Avg loss: 0.13257239, Global Avg Loss: 0.58674457, Time: 0.0209 Steps: 94180, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000514, Sample Num: 8224, Cur Loss: 0.08989939, Cur Avg Loss: 0.13372739, Log Avg loss: 0.09405162, Global Avg Loss: 0.58669227, Time: 0.0247 Steps: 94190, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000524, Sample Num: 8384, Cur Loss: 0.08684313, Cur Avg Loss: 0.13374969, Log Avg loss: 0.13489570, Global Avg Loss: 0.58664430, Time: 0.0210 Steps: 94200, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000534, Sample Num: 8544, Cur Loss: 0.04469951, Cur Avg Loss: 0.13253675, Log Avg loss: 0.06897870, Global Avg Loss: 0.58658936, Time: 0.0210 Steps: 94210, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000544, Sample Num: 8704, Cur Loss: 0.25637493, Cur Avg Loss: 0.13247863, Log Avg loss: 0.12937533, Global Avg Loss: 0.58654083, Time: 0.0210 Steps: 94220, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000554, Sample Num: 8864, Cur Loss: 0.23261863, Cur Avg Loss: 0.13240233, Log Avg loss: 0.12825136, Global Avg Loss: 0.58649219, Time: 0.0210 Steps: 94230, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000564, Sample Num: 9024, Cur Loss: 0.14958066, Cur Avg Loss: 0.13228946, Log Avg loss: 0.12603643, Global Avg Loss: 0.58644333, Time: 0.0209 Steps: 94240, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000574, Sample Num: 9184, Cur Loss: 0.19680640, Cur Avg Loss: 0.13192298, Log Avg loss: 0.11125385, Global Avg Loss: 0.58639292, Time: 0.0210 Steps: 94250, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000584, Sample Num: 9344, Cur Loss: 0.07493635, Cur Avg Loss: 0.13224861, Log Avg loss: 0.15093957, Global Avg Loss: 0.58634672, Time: 0.0210 Steps: 94260, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000594, Sample Num: 9504, Cur Loss: 0.04124555, Cur Avg Loss: 0.13228515, Log Avg loss: 0.13441899, Global Avg Loss: 0.58629878, Time: 0.0209 Steps: 94270, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000604, Sample Num: 9664, Cur Loss: 0.18329215, Cur Avg Loss: 0.13265456, Log Avg loss: 0.15459735, Global Avg Loss: 0.58625299, Time: 0.0209 Steps: 94280, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000614, Sample Num: 9824, Cur Loss: 0.03599334, Cur Avg Loss: 0.13248523, Log Avg loss: 0.12225818, Global Avg Loss: 0.58620378, Time: 0.0210 Steps: 94290, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000624, Sample Num: 9984, Cur Loss: 0.18418683, Cur Avg Loss: 0.13222061, Log Avg loss: 0.11597287, Global Avg Loss: 0.58615392, Time: 0.0210 Steps: 94300, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000634, Sample Num: 10144, Cur Loss: 0.13502575, Cur Avg Loss: 0.13143927, Log Avg loss: 0.08268358, Global Avg Loss: 0.58610053, Time: 0.0209 Steps: 94310, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000644, Sample Num: 10304, Cur Loss: 0.04582842, Cur Avg Loss: 0.13139586, Log Avg loss: 0.12864361, Global Avg Loss: 0.58605203, Time: 0.0210 Steps: 94320, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000654, Sample Num: 10464, Cur Loss: 0.14263164, Cur Avg Loss: 0.13137744, Log Avg loss: 0.13019123, Global Avg Loss: 0.58600370, Time: 0.0210 Steps: 94330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000664, Sample Num: 10624, Cur Loss: 0.07262415, Cur Avg Loss: 0.13176195, Log Avg loss: 0.15690881, Global Avg Loss: 0.58595822, Time: 0.0209 Steps: 94340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000674, Sample Num: 10784, Cur Loss: 0.05394546, Cur Avg Loss: 0.13195963, Log Avg loss: 0.14508598, Global Avg Loss: 0.58591149, Time: 0.0210 Steps: 94350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000684, Sample Num: 10944, Cur Loss: 0.19805495, Cur Avg Loss: 0.13300428, Log Avg loss: 0.20341353, Global Avg Loss: 0.58587096, Time: 0.0210 Steps: 94360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000694, Sample Num: 11104, Cur Loss: 0.04802748, Cur Avg Loss: 0.13307092, Log Avg loss: 0.13762919, Global Avg Loss: 0.58582346, Time: 0.0210 Steps: 94370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000704, Sample Num: 11264, Cur Loss: 0.06584813, Cur Avg Loss: 0.13285776, Log Avg loss: 0.11806426, Global Avg Loss: 0.58577390, Time: 0.0210 Steps: 94380, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000714, Sample Num: 11424, Cur Loss: 0.10892418, Cur Avg Loss: 0.13325352, Log Avg loss: 0.16111528, Global Avg Loss: 0.58572891, Time: 0.0209 Steps: 94390, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000724, Sample Num: 11584, Cur Loss: 0.05958172, Cur Avg Loss: 0.13300196, Log Avg loss: 0.11504018, Global Avg Loss: 0.58567905, Time: 0.0210 Steps: 94400, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000734, Sample Num: 11744, Cur Loss: 0.32043138, Cur Avg Loss: 0.13375239, Log Avg loss: 0.18808358, Global Avg Loss: 0.58563693, Time: 0.0210 Steps: 94410, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000744, Sample Num: 11904, Cur Loss: 0.59884727, Cur Avg Loss: 0.13449330, Log Avg loss: 0.18887604, Global Avg Loss: 0.58559491, Time: 0.0209 Steps: 94420, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000754, Sample Num: 12064, Cur Loss: 0.09174617, Cur Avg Loss: 0.13451057, Log Avg loss: 0.13579545, Global Avg Loss: 0.58554728, Time: 0.0210 Steps: 94430, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000764, Sample Num: 12224, Cur Loss: 0.07792536, Cur Avg Loss: 0.13470103, Log Avg loss: 0.14906190, Global Avg Loss: 0.58550106, Time: 0.0210 Steps: 94440, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000774, Sample Num: 12384, Cur Loss: 0.07517944, Cur Avg Loss: 0.13440206, Log Avg loss: 0.11156094, Global Avg Loss: 0.58545088, Time: 0.0209 Steps: 94450, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000784, Sample Num: 12544, Cur Loss: 0.42295900, Cur Avg Loss: 0.13450070, Log Avg loss: 0.14213547, Global Avg Loss: 0.58540395, Time: 0.0209 Steps: 94460, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000794, Sample Num: 12704, Cur Loss: 0.21345216, Cur Avg Loss: 0.13470355, Log Avg loss: 0.15060660, Global Avg Loss: 0.58535793, Time: 0.0209 Steps: 94470, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000804, Sample Num: 12864, Cur Loss: 0.08078662, Cur Avg Loss: 0.13498533, Log Avg loss: 0.15735886, Global Avg Loss: 0.58531263, Time: 0.0209 Steps: 94480, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000814, Sample Num: 13024, Cur Loss: 0.12202200, Cur Avg Loss: 0.13497532, Log Avg loss: 0.13416997, Global Avg Loss: 0.58526488, Time: 0.0210 Steps: 94490, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000824, Sample Num: 13184, Cur Loss: 0.23042750, Cur Avg Loss: 0.13484113, Log Avg loss: 0.12391860, Global Avg Loss: 0.58521606, Time: 0.0209 Steps: 94500, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000834, Sample Num: 13344, Cur Loss: 0.01994111, Cur Avg Loss: 0.13509833, Log Avg loss: 0.15629111, Global Avg Loss: 0.58517068, Time: 0.0209 Steps: 94510, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000844, Sample Num: 13504, Cur Loss: 0.02930747, Cur Avg Loss: 0.13485133, Log Avg loss: 0.11425189, Global Avg Loss: 0.58512085, Time: 0.0209 Steps: 94520, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000854, Sample Num: 13664, Cur Loss: 0.05868069, Cur Avg Loss: 0.13520324, Log Avg loss: 0.16490461, Global Avg Loss: 0.58507640, Time: 0.0209 Steps: 94530, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000864, Sample Num: 13824, Cur Loss: 0.03258344, Cur Avg Loss: 0.13535811, Log Avg loss: 0.14858369, Global Avg Loss: 0.58503023, Time: 0.0209 Steps: 94540, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000874, Sample Num: 13984, Cur Loss: 0.10817249, Cur Avg Loss: 0.13519852, Log Avg loss: 0.12141024, Global Avg Loss: 0.58498120, Time: 0.0209 Steps: 94550, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000884, Sample Num: 14144, Cur Loss: 0.08883161, Cur Avg Loss: 0.13517862, Log Avg loss: 0.13343910, Global Avg Loss: 0.58493344, Time: 0.0209 Steps: 94560, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000894, Sample Num: 14304, Cur Loss: 0.64682424, Cur Avg Loss: 0.13600420, Log Avg loss: 0.20898554, Global Avg Loss: 0.58489369, Time: 0.0209 Steps: 94570, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000904, Sample Num: 14464, Cur Loss: 0.03763935, Cur Avg Loss: 0.13557093, Log Avg loss: 0.09683632, Global Avg Loss: 0.58484209, Time: 0.0211 Steps: 94580, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000914, Sample Num: 14624, Cur Loss: 0.13998590, Cur Avg Loss: 0.13538273, Log Avg loss: 0.11836955, Global Avg Loss: 0.58479277, Time: 0.0209 Steps: 94590, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000924, Sample Num: 14784, Cur Loss: 0.03592962, Cur Avg Loss: 0.13480930, Log Avg loss: 0.08239781, Global Avg Loss: 0.58473967, Time: 0.0209 Steps: 94600, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000934, Sample Num: 14944, Cur Loss: 0.08118120, Cur Avg Loss: 0.13471672, Log Avg loss: 0.12616219, Global Avg Loss: 0.58469120, Time: 0.0209 Steps: 94610, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000944, Sample Num: 15104, Cur Loss: 0.02301442, Cur Avg Loss: 0.13494610, Log Avg loss: 0.15637078, Global Avg Loss: 0.58464593, Time: 0.0209 Steps: 94620, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000954, Sample Num: 15264, Cur Loss: 0.09364529, Cur Avg Loss: 0.13494078, Log Avg loss: 0.13443798, Global Avg Loss: 0.58459835, Time: 0.0209 Steps: 94630, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000964, Sample Num: 15424, Cur Loss: 0.18535110, Cur Avg Loss: 0.13470411, Log Avg loss: 0.11212633, Global Avg Loss: 0.58454843, Time: 0.0210 Steps: 94640, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000974, Sample Num: 15584, Cur Loss: 0.22439991, Cur Avg Loss: 0.13515116, Log Avg loss: 0.17824661, Global Avg Loss: 0.58450550, Time: 0.0209 Steps: 94650, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000984, Sample Num: 15744, Cur Loss: 0.05660088, Cur Avg Loss: 0.13490616, Log Avg loss: 0.11104327, Global Avg Loss: 0.58445549, Time: 0.0209 Steps: 94660, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000994, Sample Num: 15904, Cur Loss: 0.02892751, Cur Avg Loss: 0.13501988, Log Avg loss: 0.14621013, Global Avg Loss: 0.58440919, Time: 0.0209 Steps: 94670, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001004, Sample Num: 16064, Cur Loss: 0.06457460, Cur Avg Loss: 0.13518205, Log Avg loss: 0.15130108, Global Avg Loss: 0.58436345, Time: 0.0209 Steps: 94680, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001014, Sample Num: 16224, Cur Loss: 0.08973298, Cur Avg Loss: 0.13539514, Log Avg loss: 0.15678974, Global Avg Loss: 0.58431829, Time: 0.0209 Steps: 94690, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001024, Sample Num: 16384, Cur Loss: 0.07287816, Cur Avg Loss: 0.13514467, Log Avg loss: 0.10974747, Global Avg Loss: 0.58426818, Time: 0.0254 Steps: 94700, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001034, Sample Num: 16544, Cur Loss: 0.29943109, Cur Avg Loss: 0.13558366, Log Avg loss: 0.18053558, Global Avg Loss: 0.58422555, Time: 0.0210 Steps: 94710, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001044, Sample Num: 16704, Cur Loss: 0.31379461, Cur Avg Loss: 0.13648175, Log Avg loss: 0.22934432, Global Avg Loss: 0.58418809, Time: 0.0209 Steps: 94720, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001054, Sample Num: 16864, Cur Loss: 0.27057284, Cur Avg Loss: 0.13631865, Log Avg loss: 0.11929111, Global Avg Loss: 0.58413901, Time: 0.0209 Steps: 94730, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001064, Sample Num: 17024, Cur Loss: 0.09721850, Cur Avg Loss: 0.13633779, Log Avg loss: 0.13835483, Global Avg Loss: 0.58409196, Time: 0.0209 Steps: 94740, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001074, Sample Num: 17184, Cur Loss: 0.05562841, Cur Avg Loss: 0.13644667, Log Avg loss: 0.14803163, Global Avg Loss: 0.58404594, Time: 0.0209 Steps: 94750, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001084, Sample Num: 17344, Cur Loss: 0.26845768, Cur Avg Loss: 0.13653744, Log Avg loss: 0.14628667, Global Avg Loss: 0.58399974, Time: 0.0209 Steps: 94760, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001094, Sample Num: 17504, Cur Loss: 0.20720002, Cur Avg Loss: 0.13642179, Log Avg loss: 0.12388487, Global Avg Loss: 0.58395119, Time: 0.0209 Steps: 94770, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001104, Sample Num: 17664, Cur Loss: 0.08458510, Cur Avg Loss: 0.13644962, Log Avg loss: 0.13949400, Global Avg Loss: 0.58390429, Time: 0.0209 Steps: 94780, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001114, Sample Num: 17824, Cur Loss: 0.12834910, Cur Avg Loss: 0.13659228, Log Avg loss: 0.15234259, Global Avg Loss: 0.58385877, Time: 0.0209 Steps: 94790, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001124, Sample Num: 17984, Cur Loss: 0.13440999, Cur Avg Loss: 0.13734979, Log Avg loss: 0.22173632, Global Avg Loss: 0.58382057, Time: 0.0210 Steps: 94800, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001134, Sample Num: 18144, Cur Loss: 0.08604162, Cur Avg Loss: 0.13713810, Log Avg loss: 0.11334382, Global Avg Loss: 0.58377094, Time: 0.0209 Steps: 94810, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001144, Sample Num: 18304, Cur Loss: 0.06518878, Cur Avg Loss: 0.13717366, Log Avg loss: 0.14120682, Global Avg Loss: 0.58372427, Time: 0.0209 Steps: 94820, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001154, Sample Num: 18464, Cur Loss: 0.23167059, Cur Avg Loss: 0.13730028, Log Avg loss: 0.15178516, Global Avg Loss: 0.58367872, Time: 0.0209 Steps: 94830, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001164, Sample Num: 18624, Cur Loss: 0.15709640, Cur Avg Loss: 0.13751914, Log Avg loss: 0.16277565, Global Avg Loss: 0.58363434, Time: 0.0210 Steps: 94840, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001174, Sample Num: 18784, Cur Loss: 0.07566724, Cur Avg Loss: 0.13722541, Log Avg loss: 0.10303507, Global Avg Loss: 0.58358367, Time: 0.0209 Steps: 94850, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001184, Sample Num: 18944, Cur Loss: 0.27874601, Cur Avg Loss: 0.13729675, Log Avg loss: 0.14567252, Global Avg Loss: 0.58353751, Time: 0.0209 Steps: 94860, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001194, Sample Num: 19104, Cur Loss: 0.06913953, Cur Avg Loss: 0.13704099, Log Avg loss: 0.10675896, Global Avg Loss: 0.58348725, Time: 0.0209 Steps: 94870, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001204, Sample Num: 19264, Cur Loss: 0.07458558, Cur Avg Loss: 0.13671543, Log Avg loss: 0.09784285, Global Avg Loss: 0.58343607, Time: 0.0209 Steps: 94880, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001214, Sample Num: 19424, Cur Loss: 0.11549373, Cur Avg Loss: 0.13650029, Log Avg loss: 0.11059731, Global Avg Loss: 0.58338624, Time: 0.0209 Steps: 94890, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001224, Sample Num: 19584, Cur Loss: 0.04309030, Cur Avg Loss: 0.13680570, Log Avg loss: 0.17388293, Global Avg Loss: 0.58334309, Time: 0.0209 Steps: 94900, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001234, Sample Num: 19744, Cur Loss: 0.11322694, Cur Avg Loss: 0.13669946, Log Avg loss: 0.12369538, Global Avg Loss: 0.58329466, Time: 0.0209 Steps: 94910, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001244, Sample Num: 19904, Cur Loss: 0.21991627, Cur Avg Loss: 0.13653882, Log Avg loss: 0.11671608, Global Avg Loss: 0.58324550, Time: 0.0209 Steps: 94920, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001254, Sample Num: 20064, Cur Loss: 0.46594861, Cur Avg Loss: 0.13681768, Log Avg loss: 0.17150808, Global Avg Loss: 0.58320213, Time: 0.0209 Steps: 94930, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001264, Sample Num: 20224, Cur Loss: 0.16491693, Cur Avg Loss: 0.13673696, Log Avg loss: 0.12661506, Global Avg Loss: 0.58315404, Time: 0.0209 Steps: 94940, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001274, Sample Num: 20384, Cur Loss: 0.17790537, Cur Avg Loss: 0.13659061, Log Avg loss: 0.11809149, Global Avg Loss: 0.58310506, Time: 0.0210 Steps: 94950, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001284, Sample Num: 20544, Cur Loss: 0.10628984, Cur Avg Loss: 0.13667690, Log Avg loss: 0.14767012, Global Avg Loss: 0.58305920, Time: 0.0247 Steps: 94960, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001294, Sample Num: 20704, Cur Loss: 0.07740219, Cur Avg Loss: 0.13687078, Log Avg loss: 0.16176471, Global Avg Loss: 0.58301484, Time: 0.0209 Steps: 94970, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001304, Sample Num: 20864, Cur Loss: 0.03701035, Cur Avg Loss: 0.13670698, Log Avg loss: 0.11551113, Global Avg Loss: 0.58296562, Time: 0.0209 Steps: 94980, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001314, Sample Num: 21024, Cur Loss: 0.08749063, Cur Avg Loss: 0.13639201, Log Avg loss: 0.09532089, Global Avg Loss: 0.58291428, Time: 0.0210 Steps: 94990, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001324, Sample Num: 21184, Cur Loss: 0.18711343, Cur Avg Loss: 0.13638684, Log Avg loss: 0.13570724, Global Avg Loss: 0.58286721, Time: 0.0209 Steps: 95000, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001334, Sample Num: 21344, Cur Loss: 0.02180261, Cur Avg Loss: 0.13689638, Log Avg loss: 0.20435900, Global Avg Loss: 0.58282737, Time: 0.0209 Steps: 95010, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001344, Sample Num: 21504, Cur Loss: 0.20001495, Cur Avg Loss: 0.13680278, Log Avg loss: 0.12431656, Global Avg Loss: 0.58277912, Time: 0.0209 Steps: 95020, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001354, Sample Num: 21664, Cur Loss: 0.07583205, Cur Avg Loss: 0.13680564, Log Avg loss: 0.13719081, Global Avg Loss: 0.58273223, Time: 0.0209 Steps: 95030, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001364, Sample Num: 21824, Cur Loss: 0.10653809, Cur Avg Loss: 0.13691707, Log Avg loss: 0.15200393, Global Avg Loss: 0.58268691, Time: 0.0209 Steps: 95040, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001374, Sample Num: 21984, Cur Loss: 0.06808931, Cur Avg Loss: 0.13707130, Log Avg loss: 0.15810886, Global Avg Loss: 0.58264224, Time: 0.0209 Steps: 95050, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001384, Sample Num: 22144, Cur Loss: 0.07495145, Cur Avg Loss: 0.13720583, Log Avg loss: 0.15568992, Global Avg Loss: 0.58259732, Time: 0.0209 Steps: 95060, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001394, Sample Num: 22304, Cur Loss: 0.18845740, Cur Avg Loss: 0.13724091, Log Avg loss: 0.14209623, Global Avg Loss: 0.58255099, Time: 0.0209 Steps: 95070, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001404, Sample Num: 22464, Cur Loss: 0.11531886, Cur Avg Loss: 0.13709761, Log Avg loss: 0.11712099, Global Avg Loss: 0.58250204, Time: 0.0209 Steps: 95080, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001414, Sample Num: 22624, Cur Loss: 0.11704206, Cur Avg Loss: 0.13721567, Log Avg loss: 0.15379223, Global Avg Loss: 0.58245695, Time: 0.0209 Steps: 95090, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001424, Sample Num: 22784, Cur Loss: 0.14384159, Cur Avg Loss: 0.13753318, Log Avg loss: 0.18242862, Global Avg Loss: 0.58241489, Time: 0.0209 Steps: 95100, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001434, Sample Num: 22944, Cur Loss: 0.03146693, Cur Avg Loss: 0.13720482, Log Avg loss: 0.09044599, Global Avg Loss: 0.58236316, Time: 0.0209 Steps: 95110, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001444, Sample Num: 23104, Cur Loss: 0.11051466, Cur Avg Loss: 0.13721774, Log Avg loss: 0.13907006, Global Avg Loss: 0.58231656, Time: 0.0210 Steps: 95120, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001454, Sample Num: 23264, Cur Loss: 0.27368817, Cur Avg Loss: 0.13765645, Log Avg loss: 0.20100730, Global Avg Loss: 0.58227648, Time: 0.0209 Steps: 95130, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001464, Sample Num: 23424, Cur Loss: 0.03853982, Cur Avg Loss: 0.13760747, Log Avg loss: 0.13048487, Global Avg Loss: 0.58222899, Time: 0.0209 Steps: 95140, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001474, Sample Num: 23584, Cur Loss: 0.11560620, Cur Avg Loss: 0.13773771, Log Avg loss: 0.15680477, Global Avg Loss: 0.58218428, Time: 0.0209 Steps: 95150, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001484, Sample Num: 23744, Cur Loss: 0.26557854, Cur Avg Loss: 0.13759230, Log Avg loss: 0.11615960, Global Avg Loss: 0.58213530, Time: 0.0209 Steps: 95160, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001494, Sample Num: 23904, Cur Loss: 0.15364298, Cur Avg Loss: 0.13784797, Log Avg loss: 0.17578909, Global Avg Loss: 0.58209261, Time: 0.0209 Steps: 95170, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001504, Sample Num: 24064, Cur Loss: 0.36768514, Cur Avg Loss: 0.13772465, Log Avg loss: 0.11930055, Global Avg Loss: 0.58204399, Time: 0.0209 Steps: 95180, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001514, Sample Num: 24224, Cur Loss: 0.14591075, Cur Avg Loss: 0.13794591, Log Avg loss: 0.17122383, Global Avg Loss: 0.58200083, Time: 0.0209 Steps: 95190, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001524, Sample Num: 24384, Cur Loss: 0.06303432, Cur Avg Loss: 0.13773757, Log Avg loss: 0.10619528, Global Avg Loss: 0.58195085, Time: 0.0209 Steps: 95200, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001534, Sample Num: 24544, Cur Loss: 0.19965726, Cur Avg Loss: 0.13781984, Log Avg loss: 0.15035723, Global Avg Loss: 0.58190552, Time: 0.0209 Steps: 95210, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001544, Sample Num: 24704, Cur Loss: 0.12058914, Cur Avg Loss: 0.13762704, Log Avg loss: 0.10805170, Global Avg Loss: 0.58185575, Time: 0.0210 Steps: 95220, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001554, Sample Num: 24864, Cur Loss: 0.13908394, Cur Avg Loss: 0.13758929, Log Avg loss: 0.13176116, Global Avg Loss: 0.58180849, Time: 0.0209 Steps: 95230, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001564, Sample Num: 25024, Cur Loss: 0.17826128, Cur Avg Loss: 0.13754847, Log Avg loss: 0.13120376, Global Avg Loss: 0.58176118, Time: 0.0209 Steps: 95240, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001574, Sample Num: 25184, Cur Loss: 0.24885264, Cur Avg Loss: 0.13746552, Log Avg loss: 0.12449205, Global Avg Loss: 0.58171317, Time: 0.0209 Steps: 95250, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001584, Sample Num: 25344, Cur Loss: 0.14468071, Cur Avg Loss: 0.13747613, Log Avg loss: 0.13914702, Global Avg Loss: 0.58166671, Time: 0.0209 Steps: 95260, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001594, Sample Num: 25504, Cur Loss: 0.13803783, Cur Avg Loss: 0.13765525, Log Avg loss: 0.16602708, Global Avg Loss: 0.58162308, Time: 0.0208 Steps: 95270, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001604, Sample Num: 25664, Cur Loss: 0.18463005, Cur Avg Loss: 0.13800891, Log Avg loss: 0.19438253, Global Avg Loss: 0.58158244, Time: 0.0209 Steps: 95280, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001614, Sample Num: 25824, Cur Loss: 0.26850140, Cur Avg Loss: 0.13807372, Log Avg loss: 0.14846923, Global Avg Loss: 0.58153699, Time: 0.0209 Steps: 95290, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001624, Sample Num: 25984, Cur Loss: 0.15306886, Cur Avg Loss: 0.13813156, Log Avg loss: 0.14746774, Global Avg Loss: 0.58149144, Time: 0.0209 Steps: 95300, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001634, Sample Num: 26144, Cur Loss: 0.05524618, Cur Avg Loss: 0.13812015, Log Avg loss: 0.13626620, Global Avg Loss: 0.58144473, Time: 0.0209 Steps: 95310, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001644, Sample Num: 26304, Cur Loss: 0.28267509, Cur Avg Loss: 0.13826159, Log Avg loss: 0.16137362, Global Avg Loss: 0.58140066, Time: 0.0209 Steps: 95320, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001654, Sample Num: 26464, Cur Loss: 0.03219772, Cur Avg Loss: 0.13821724, Log Avg loss: 0.13092576, Global Avg Loss: 0.58135340, Time: 0.0209 Steps: 95330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001664, Sample Num: 26624, Cur Loss: 0.10042638, Cur Avg Loss: 0.13810211, Log Avg loss: 0.11906042, Global Avg Loss: 0.58130491, Time: 0.0209 Steps: 95340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001674, Sample Num: 26784, Cur Loss: 0.07388864, Cur Avg Loss: 0.13822687, Log Avg loss: 0.15898616, Global Avg Loss: 0.58126062, Time: 0.0209 Steps: 95350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001684, Sample Num: 26944, Cur Loss: 0.07689073, Cur Avg Loss: 0.13802694, Log Avg loss: 0.10455914, Global Avg Loss: 0.58121063, Time: 0.0208 Steps: 95360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001694, Sample Num: 27104, Cur Loss: 0.13062140, Cur Avg Loss: 0.13813896, Log Avg loss: 0.15700280, Global Avg Loss: 0.58116615, Time: 0.0209 Steps: 95370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001704, Sample Num: 27264, Cur Loss: 0.06071952, Cur Avg Loss: 0.13795958, Log Avg loss: 0.10757229, Global Avg Loss: 0.58111650, Time: 0.0209 Steps: 95380, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001714, Sample Num: 27424, Cur Loss: 0.10671332, Cur Avg Loss: 0.13799527, Log Avg loss: 0.14407625, Global Avg Loss: 0.58107068, Time: 0.0209 Steps: 95390, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001724, Sample Num: 27584, Cur Loss: 0.14103386, Cur Avg Loss: 0.13783910, Log Avg loss: 0.11107249, Global Avg Loss: 0.58102142, Time: 0.0209 Steps: 95400, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001734, Sample Num: 27744, Cur Loss: 0.20772129, Cur Avg Loss: 0.13766676, Log Avg loss: 0.10795511, Global Avg Loss: 0.58097184, Time: 0.0209 Steps: 95410, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001744, Sample Num: 27904, Cur Loss: 0.20763883, Cur Avg Loss: 0.13749088, Log Avg loss: 0.10699265, Global Avg Loss: 0.58092216, Time: 0.0208 Steps: 95420, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001754, Sample Num: 28064, Cur Loss: 0.14154719, Cur Avg Loss: 0.13750110, Log Avg loss: 0.13928492, Global Avg Loss: 0.58087588, Time: 0.0209 Steps: 95430, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001764, Sample Num: 28224, Cur Loss: 0.27610648, Cur Avg Loss: 0.13779481, Log Avg loss: 0.18931078, Global Avg Loss: 0.58083486, Time: 0.0209 Steps: 95440, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001774, Sample Num: 28384, Cur Loss: 0.14615659, Cur Avg Loss: 0.13793521, Log Avg loss: 0.16270131, Global Avg Loss: 0.58079105, Time: 0.0209 Steps: 95450, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001784, Sample Num: 28544, Cur Loss: 0.10981669, Cur Avg Loss: 0.13798299, Log Avg loss: 0.14645926, Global Avg Loss: 0.58074555, Time: 0.0209 Steps: 95460, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001794, Sample Num: 28704, Cur Loss: 0.14965087, Cur Avg Loss: 0.13825377, Log Avg loss: 0.18656131, Global Avg Loss: 0.58070426, Time: 0.0247 Steps: 95470, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001804, Sample Num: 28864, Cur Loss: 0.03158342, Cur Avg Loss: 0.13798044, Log Avg loss: 0.08894472, Global Avg Loss: 0.58065276, Time: 0.0209 Steps: 95480, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001814, Sample Num: 29024, Cur Loss: 0.27766246, Cur Avg Loss: 0.13801914, Log Avg loss: 0.14500016, Global Avg Loss: 0.58060714, Time: 0.0210 Steps: 95490, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001824, Sample Num: 29184, Cur Loss: 0.06098550, Cur Avg Loss: 0.13788300, Log Avg loss: 0.11318762, Global Avg Loss: 0.58055819, Time: 0.0210 Steps: 95500, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001834, Sample Num: 29344, Cur Loss: 0.12609835, Cur Avg Loss: 0.13783034, Log Avg loss: 0.12822613, Global Avg Loss: 0.58051083, Time: 0.0210 Steps: 95510, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001844, Sample Num: 29504, Cur Loss: 0.07628623, Cur Avg Loss: 0.13820552, Log Avg loss: 0.20701291, Global Avg Loss: 0.58047173, Time: 0.0210 Steps: 95520, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001854, Sample Num: 29664, Cur Loss: 0.07104626, Cur Avg Loss: 0.13823652, Log Avg loss: 0.14395373, Global Avg Loss: 0.58042604, Time: 0.0209 Steps: 95530, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001864, Sample Num: 29824, Cur Loss: 0.05019401, Cur Avg Loss: 0.13831784, Log Avg loss: 0.15339349, Global Avg Loss: 0.58038134, Time: 0.0210 Steps: 95540, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001874, Sample Num: 29984, Cur Loss: 0.39213139, Cur Avg Loss: 0.13836868, Log Avg loss: 0.14784587, Global Avg Loss: 0.58033607, Time: 0.0211 Steps: 95550, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001884, Sample Num: 30144, Cur Loss: 0.15038043, Cur Avg Loss: 0.13812193, Log Avg loss: 0.09188061, Global Avg Loss: 0.58028496, Time: 0.0209 Steps: 95560, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001894, Sample Num: 30304, Cur Loss: 0.27113447, Cur Avg Loss: 0.13826966, Log Avg loss: 0.16610161, Global Avg Loss: 0.58024162, Time: 0.0210 Steps: 95570, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001904, Sample Num: 30464, Cur Loss: 0.32016253, Cur Avg Loss: 0.13828387, Log Avg loss: 0.14097658, Global Avg Loss: 0.58019566, Time: 0.0209 Steps: 95580, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001914, Sample Num: 30624, Cur Loss: 0.09387209, Cur Avg Loss: 0.13833838, Log Avg loss: 0.14871700, Global Avg Loss: 0.58015052, Time: 0.0210 Steps: 95590, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001924, Sample Num: 30784, Cur Loss: 0.43226588, Cur Avg Loss: 0.13843911, Log Avg loss: 0.15771874, Global Avg Loss: 0.58010633, Time: 0.0209 Steps: 95600, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001934, Sample Num: 30944, Cur Loss: 0.26605794, Cur Avg Loss: 0.13836400, Log Avg loss: 0.12391304, Global Avg Loss: 0.58005862, Time: 0.0211 Steps: 95610, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001944, Sample Num: 31104, Cur Loss: 0.30267215, Cur Avg Loss: 0.13848385, Log Avg loss: 0.16166180, Global Avg Loss: 0.58001486, Time: 0.0209 Steps: 95620, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001954, Sample Num: 31264, Cur Loss: 0.23903839, Cur Avg Loss: 0.13834383, Log Avg loss: 0.11112503, Global Avg Loss: 0.57996583, Time: 0.0210 Steps: 95630, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001964, Sample Num: 31424, Cur Loss: 0.09819111, Cur Avg Loss: 0.13831359, Log Avg loss: 0.13240300, Global Avg Loss: 0.57991904, Time: 0.0209 Steps: 95640, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001974, Sample Num: 31584, Cur Loss: 0.04561888, Cur Avg Loss: 0.13844100, Log Avg loss: 0.16346593, Global Avg Loss: 0.57987550, Time: 0.0209 Steps: 95650, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001984, Sample Num: 31744, Cur Loss: 0.07480409, Cur Avg Loss: 0.13829826, Log Avg loss: 0.11012133, Global Avg Loss: 0.57982639, Time: 0.0210 Steps: 95660, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001994, Sample Num: 31904, Cur Loss: 0.12350111, Cur Avg Loss: 0.13849112, Log Avg loss: 0.17675326, Global Avg Loss: 0.57978426, Time: 0.0209 Steps: 95670, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002004, Sample Num: 32064, Cur Loss: 0.04522054, Cur Avg Loss: 0.13856988, Log Avg loss: 0.15427546, Global Avg Loss: 0.57973979, Time: 0.0209 Steps: 95680, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002014, Sample Num: 32224, Cur Loss: 0.33130124, Cur Avg Loss: 0.13884323, Log Avg loss: 0.19362170, Global Avg Loss: 0.57969943, Time: 0.0209 Steps: 95690, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002024, Sample Num: 32384, Cur Loss: 0.13288590, Cur Avg Loss: 0.13856015, Log Avg loss: 0.08154905, Global Avg Loss: 0.57964738, Time: 0.0209 Steps: 95700, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002034, Sample Num: 32544, Cur Loss: 0.18424636, Cur Avg Loss: 0.13858967, Log Avg loss: 0.14456465, Global Avg Loss: 0.57960192, Time: 0.0210 Steps: 95710, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002044, Sample Num: 32704, Cur Loss: 0.07892790, Cur Avg Loss: 0.13917492, Log Avg loss: 0.25821491, Global Avg Loss: 0.57956835, Time: 0.0210 Steps: 95720, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002054, Sample Num: 32864, Cur Loss: 0.04056402, Cur Avg Loss: 0.13906577, Log Avg loss: 0.11675396, Global Avg Loss: 0.57952000, Time: 0.0211 Steps: 95730, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002064, Sample Num: 33024, Cur Loss: 0.24276325, Cur Avg Loss: 0.13918558, Log Avg loss: 0.16379483, Global Avg Loss: 0.57947658, Time: 0.0211 Steps: 95740, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002074, Sample Num: 33184, Cur Loss: 0.06435746, Cur Avg Loss: 0.13912385, Log Avg loss: 0.12638250, Global Avg Loss: 0.57942926, Time: 0.0210 Steps: 95750, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002084, Sample Num: 33344, Cur Loss: 0.15948628, Cur Avg Loss: 0.13899952, Log Avg loss: 0.11321498, Global Avg Loss: 0.57938057, Time: 0.0210 Steps: 95760, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002094, Sample Num: 33504, Cur Loss: 0.11155054, Cur Avg Loss: 0.13890254, Log Avg loss: 0.11869170, Global Avg Loss: 0.57933247, Time: 0.0210 Steps: 95770, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002104, Sample Num: 33664, Cur Loss: 0.15530959, Cur Avg Loss: 0.13905040, Log Avg loss: 0.17001153, Global Avg Loss: 0.57928973, Time: 0.0210 Steps: 95780, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002114, Sample Num: 33824, Cur Loss: 0.16006666, Cur Avg Loss: 0.13891734, Log Avg loss: 0.11092221, Global Avg Loss: 0.57924084, Time: 0.0209 Steps: 95790, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002124, Sample Num: 33984, Cur Loss: 0.18555900, Cur Avg Loss: 0.13881394, Log Avg loss: 0.11695472, Global Avg Loss: 0.57919258, Time: 0.0210 Steps: 95800, Updated lr: 0.000010 ***** Running evaluation checkpoint-95805 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-95805 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.793445, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.221406, "eval_total_loss": 155.648371, "eval_mae": 0.306715, "eval_mse": 0.221491, "eval_r2": 0.859206, "eval_sp_statistic": 0.930653, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.93607, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.194163, "test_total_loss": 97.469636, "test_mae": 0.327945, "test_mse": 0.19423, "test_r2": 0.874643, "test_sp_statistic": 0.920903, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946972, "test_ps_pvalue": 0.0, "lr": 1.0094831673779043e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5791657769279245, "train_cur_epoch_loss": 295.1686238925904, "train_cur_epoch_avg_loss": 0.1386419088269565, "train_cur_epoch_time": 44.793444871902466, "train_cur_epoch_avg_time": 0.021039664101410272, "epoch": 45, "step": 95805} ################################################## Training, Epoch: 0046, Batch: 000005, Sample Num: 80, Cur Loss: 0.50384802, Cur Avg Loss: 0.28723836, Log Avg loss: 0.17640051, Global Avg Loss: 0.57915054, Time: 0.0218 Steps: 95810, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000015, Sample Num: 240, Cur Loss: 0.19728810, Cur Avg Loss: 0.19210412, Log Avg loss: 0.14453700, Global Avg Loss: 0.57910518, Time: 0.0209 Steps: 95820, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000025, Sample Num: 400, Cur Loss: 0.35056007, Cur Avg Loss: 0.16187548, Log Avg loss: 0.11653253, Global Avg Loss: 0.57905691, Time: 0.0208 Steps: 95830, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000035, Sample Num: 560, Cur Loss: 0.12261184, Cur Avg Loss: 0.15387444, Log Avg loss: 0.13387184, Global Avg Loss: 0.57901046, Time: 0.0209 Steps: 95840, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000045, Sample Num: 720, Cur Loss: 0.10694343, Cur Avg Loss: 0.15310895, Log Avg loss: 0.15042971, Global Avg Loss: 0.57896575, Time: 0.0209 Steps: 95850, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000055, Sample Num: 880, Cur Loss: 0.36056179, Cur Avg Loss: 0.14696714, Log Avg loss: 0.11932900, Global Avg Loss: 0.57891780, Time: 0.0209 Steps: 95860, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000065, Sample Num: 1040, Cur Loss: 0.02882386, Cur Avg Loss: 0.14263072, Log Avg loss: 0.11878039, Global Avg Loss: 0.57886981, Time: 0.0209 Steps: 95870, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000075, Sample Num: 1200, Cur Loss: 0.13923426, Cur Avg Loss: 0.13691392, Log Avg loss: 0.09975473, Global Avg Loss: 0.57881984, Time: 0.0209 Steps: 95880, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000085, Sample Num: 1360, Cur Loss: 0.22227731, Cur Avg Loss: 0.14381388, Log Avg loss: 0.19556363, Global Avg Loss: 0.57877987, Time: 0.0209 Steps: 95890, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000095, Sample Num: 1520, Cur Loss: 0.07652560, Cur Avg Loss: 0.14216676, Log Avg loss: 0.12816617, Global Avg Loss: 0.57873288, Time: 0.0210 Steps: 95900, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000105, Sample Num: 1680, Cur Loss: 0.14346179, Cur Avg Loss: 0.13948811, Log Avg loss: 0.11404093, Global Avg Loss: 0.57868443, Time: 0.0210 Steps: 95910, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000115, Sample Num: 1840, Cur Loss: 0.05325932, Cur Avg Loss: 0.13396910, Log Avg loss: 0.07601951, Global Avg Loss: 0.57863202, Time: 0.0219 Steps: 95920, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000125, Sample Num: 2000, Cur Loss: 0.10176437, Cur Avg Loss: 0.13225946, Log Avg loss: 0.11259866, Global Avg Loss: 0.57858344, Time: 0.0219 Steps: 95930, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000135, Sample Num: 2160, Cur Loss: 0.13779993, Cur Avg Loss: 0.13174666, Log Avg loss: 0.12533667, Global Avg Loss: 0.57853620, Time: 0.0209 Steps: 95940, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000145, Sample Num: 2320, Cur Loss: 0.24494702, Cur Avg Loss: 0.13233726, Log Avg loss: 0.14031026, Global Avg Loss: 0.57849053, Time: 0.0219 Steps: 95950, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000155, Sample Num: 2480, Cur Loss: 0.03447831, Cur Avg Loss: 0.13302185, Log Avg loss: 0.14294851, Global Avg Loss: 0.57844514, Time: 0.0209 Steps: 95960, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000165, Sample Num: 2640, Cur Loss: 0.06290367, Cur Avg Loss: 0.13055195, Log Avg loss: 0.09226839, Global Avg Loss: 0.57839448, Time: 0.0219 Steps: 95970, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000175, Sample Num: 2800, Cur Loss: 0.07543183, Cur Avg Loss: 0.12827576, Log Avg loss: 0.09071861, Global Avg Loss: 0.57834367, Time: 0.0209 Steps: 95980, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000185, Sample Num: 2960, Cur Loss: 0.10136815, Cur Avg Loss: 0.12915354, Log Avg loss: 0.14451481, Global Avg Loss: 0.57829848, Time: 0.0209 Steps: 95990, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000195, Sample Num: 3120, Cur Loss: 0.06090069, Cur Avg Loss: 0.12740985, Log Avg loss: 0.09515156, Global Avg Loss: 0.57824815, Time: 0.0219 Steps: 96000, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000205, Sample Num: 3280, Cur Loss: 0.13510013, Cur Avg Loss: 0.12644648, Log Avg loss: 0.10766081, Global Avg Loss: 0.57819913, Time: 0.0219 Steps: 96010, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000215, Sample Num: 3440, Cur Loss: 0.32503429, Cur Avg Loss: 0.12637711, Log Avg loss: 0.12495501, Global Avg Loss: 0.57815193, Time: 0.0219 Steps: 96020, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000225, Sample Num: 3600, Cur Loss: 0.11563424, Cur Avg Loss: 0.13059124, Log Avg loss: 0.22119505, Global Avg Loss: 0.57811476, Time: 0.0209 Steps: 96030, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000235, Sample Num: 3760, Cur Loss: 0.03110647, Cur Avg Loss: 0.13174525, Log Avg loss: 0.15771044, Global Avg Loss: 0.57807098, Time: 0.0210 Steps: 96040, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000245, Sample Num: 3920, Cur Loss: 0.21307087, Cur Avg Loss: 0.13065543, Log Avg loss: 0.10504456, Global Avg Loss: 0.57802174, Time: 0.0219 Steps: 96050, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000255, Sample Num: 4080, Cur Loss: 0.15382546, Cur Avg Loss: 0.12990576, Log Avg loss: 0.11153880, Global Avg Loss: 0.57797318, Time: 0.0209 Steps: 96060, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000265, Sample Num: 4240, Cur Loss: 0.05907335, Cur Avg Loss: 0.12798416, Log Avg loss: 0.07898354, Global Avg Loss: 0.57792124, Time: 0.0208 Steps: 96070, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000275, Sample Num: 4400, Cur Loss: 0.03599659, Cur Avg Loss: 0.12777885, Log Avg loss: 0.12233801, Global Avg Loss: 0.57787382, Time: 0.0208 Steps: 96080, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000285, Sample Num: 4560, Cur Loss: 0.09958541, Cur Avg Loss: 0.12822513, Log Avg loss: 0.14049780, Global Avg Loss: 0.57782830, Time: 0.0208 Steps: 96090, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000295, Sample Num: 4720, Cur Loss: 0.09576941, Cur Avg Loss: 0.12926385, Log Avg loss: 0.15886754, Global Avg Loss: 0.57778470, Time: 0.0209 Steps: 96100, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000305, Sample Num: 4880, Cur Loss: 0.19583943, Cur Avg Loss: 0.12958822, Log Avg loss: 0.13915719, Global Avg Loss: 0.57773907, Time: 0.0208 Steps: 96110, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000315, Sample Num: 5040, Cur Loss: 0.09508637, Cur Avg Loss: 0.12950577, Log Avg loss: 0.12699085, Global Avg Loss: 0.57769217, Time: 0.0208 Steps: 96120, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000325, Sample Num: 5200, Cur Loss: 0.18229510, Cur Avg Loss: 0.12896660, Log Avg loss: 0.11198278, Global Avg Loss: 0.57764373, Time: 0.0208 Steps: 96130, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000335, Sample Num: 5360, Cur Loss: 0.12205024, Cur Avg Loss: 0.12914449, Log Avg loss: 0.13492584, Global Avg Loss: 0.57759768, Time: 0.0208 Steps: 96140, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000345, Sample Num: 5520, Cur Loss: 0.18330383, Cur Avg Loss: 0.12872215, Log Avg loss: 0.11457369, Global Avg Loss: 0.57754952, Time: 0.0208 Steps: 96150, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000355, Sample Num: 5680, Cur Loss: 0.54487699, Cur Avg Loss: 0.12864087, Log Avg loss: 0.12583685, Global Avg Loss: 0.57750255, Time: 0.0208 Steps: 96160, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000365, Sample Num: 5840, Cur Loss: 0.02628548, Cur Avg Loss: 0.13057751, Log Avg loss: 0.19932815, Global Avg Loss: 0.57746322, Time: 0.0208 Steps: 96170, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000375, Sample Num: 6000, Cur Loss: 0.24625415, Cur Avg Loss: 0.12984334, Log Avg loss: 0.10304631, Global Avg Loss: 0.57741390, Time: 0.0212 Steps: 96180, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000385, Sample Num: 6160, Cur Loss: 0.06686470, Cur Avg Loss: 0.12984913, Log Avg loss: 0.13006622, Global Avg Loss: 0.57736739, Time: 0.0208 Steps: 96190, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000395, Sample Num: 6320, Cur Loss: 0.03954540, Cur Avg Loss: 0.13083765, Log Avg loss: 0.16889565, Global Avg Loss: 0.57732493, Time: 0.0208 Steps: 96200, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000405, Sample Num: 6480, Cur Loss: 0.11708744, Cur Avg Loss: 0.13118519, Log Avg loss: 0.14491314, Global Avg Loss: 0.57727998, Time: 0.0208 Steps: 96210, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000415, Sample Num: 6640, Cur Loss: 0.07487334, Cur Avg Loss: 0.13065529, Log Avg loss: 0.10919413, Global Avg Loss: 0.57723134, Time: 0.0208 Steps: 96220, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000425, Sample Num: 6800, Cur Loss: 0.11418796, Cur Avg Loss: 0.12971179, Log Avg loss: 0.09055671, Global Avg Loss: 0.57718076, Time: 0.0208 Steps: 96230, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000435, Sample Num: 6960, Cur Loss: 0.06134801, Cur Avg Loss: 0.12944085, Log Avg loss: 0.11792572, Global Avg Loss: 0.57713304, Time: 0.0210 Steps: 96240, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000445, Sample Num: 7120, Cur Loss: 0.23597372, Cur Avg Loss: 0.12976066, Log Avg loss: 0.14367260, Global Avg Loss: 0.57708801, Time: 0.0208 Steps: 96250, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000455, Sample Num: 7280, Cur Loss: 0.10369778, Cur Avg Loss: 0.12976260, Log Avg loss: 0.12984894, Global Avg Loss: 0.57704155, Time: 0.0208 Steps: 96260, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000465, Sample Num: 7440, Cur Loss: 0.05547797, Cur Avg Loss: 0.12938944, Log Avg loss: 0.11241051, Global Avg Loss: 0.57699328, Time: 0.0209 Steps: 96270, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000475, Sample Num: 7600, Cur Loss: 0.16044161, Cur Avg Loss: 0.12885362, Log Avg loss: 0.10393819, Global Avg Loss: 0.57694415, Time: 0.0208 Steps: 96280, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000485, Sample Num: 7760, Cur Loss: 0.25626540, Cur Avg Loss: 0.12929818, Log Avg loss: 0.15041468, Global Avg Loss: 0.57689985, Time: 0.0208 Steps: 96290, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000495, Sample Num: 7920, Cur Loss: 0.17626137, Cur Avg Loss: 0.12957626, Log Avg loss: 0.14306311, Global Avg Loss: 0.57685480, Time: 0.0208 Steps: 96300, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000505, Sample Num: 8080, Cur Loss: 0.15243028, Cur Avg Loss: 0.13032537, Log Avg loss: 0.16740627, Global Avg Loss: 0.57681229, Time: 0.0208 Steps: 96310, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000515, Sample Num: 8240, Cur Loss: 0.10865956, Cur Avg Loss: 0.13067506, Log Avg loss: 0.14833449, Global Avg Loss: 0.57676780, Time: 0.0245 Steps: 96320, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000525, Sample Num: 8400, Cur Loss: 0.20697038, Cur Avg Loss: 0.13058198, Log Avg loss: 0.12578801, Global Avg Loss: 0.57672099, Time: 0.0208 Steps: 96330, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000535, Sample Num: 8560, Cur Loss: 0.31756091, Cur Avg Loss: 0.13054884, Log Avg loss: 0.12880943, Global Avg Loss: 0.57667450, Time: 0.0208 Steps: 96340, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000545, Sample Num: 8720, Cur Loss: 0.13111678, Cur Avg Loss: 0.13085798, Log Avg loss: 0.14739666, Global Avg Loss: 0.57662994, Time: 0.0208 Steps: 96350, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000555, Sample Num: 8880, Cur Loss: 0.18577060, Cur Avg Loss: 0.13070111, Log Avg loss: 0.12215157, Global Avg Loss: 0.57658278, Time: 0.0208 Steps: 96360, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000565, Sample Num: 9040, Cur Loss: 0.06692674, Cur Avg Loss: 0.12949561, Log Avg loss: 0.06259076, Global Avg Loss: 0.57652944, Time: 0.0209 Steps: 96370, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000575, Sample Num: 9200, Cur Loss: 0.06375653, Cur Avg Loss: 0.13006666, Log Avg loss: 0.16233060, Global Avg Loss: 0.57648647, Time: 0.0208 Steps: 96380, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000585, Sample Num: 9360, Cur Loss: 0.31100196, Cur Avg Loss: 0.13045060, Log Avg loss: 0.15252740, Global Avg Loss: 0.57644248, Time: 0.0208 Steps: 96390, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000595, Sample Num: 9520, Cur Loss: 0.09570724, Cur Avg Loss: 0.12993027, Log Avg loss: 0.09949108, Global Avg Loss: 0.57639301, Time: 0.0208 Steps: 96400, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000605, Sample Num: 9680, Cur Loss: 0.08677033, Cur Avg Loss: 0.12929856, Log Avg loss: 0.09171175, Global Avg Loss: 0.57634273, Time: 0.0208 Steps: 96410, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000615, Sample Num: 9840, Cur Loss: 0.03941225, Cur Avg Loss: 0.12826843, Log Avg loss: 0.06594541, Global Avg Loss: 0.57628980, Time: 0.0208 Steps: 96420, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000625, Sample Num: 10000, Cur Loss: 0.26638353, Cur Avg Loss: 0.12898188, Log Avg loss: 0.17285901, Global Avg Loss: 0.57624796, Time: 0.0208 Steps: 96430, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000635, Sample Num: 10160, Cur Loss: 0.34204531, Cur Avg Loss: 0.12880285, Log Avg loss: 0.11761331, Global Avg Loss: 0.57620041, Time: 0.0208 Steps: 96440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000645, Sample Num: 10320, Cur Loss: 0.14936814, Cur Avg Loss: 0.12892167, Log Avg loss: 0.13646701, Global Avg Loss: 0.57615481, Time: 0.0208 Steps: 96450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000655, Sample Num: 10480, Cur Loss: 0.05258102, Cur Avg Loss: 0.12864671, Log Avg loss: 0.11091184, Global Avg Loss: 0.57610658, Time: 0.0208 Steps: 96460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000665, Sample Num: 10640, Cur Loss: 0.28097647, Cur Avg Loss: 0.12910192, Log Avg loss: 0.15891833, Global Avg Loss: 0.57606334, Time: 0.0208 Steps: 96470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000675, Sample Num: 10800, Cur Loss: 0.57553142, Cur Avg Loss: 0.12980114, Log Avg loss: 0.17629892, Global Avg Loss: 0.57602190, Time: 0.0209 Steps: 96480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000685, Sample Num: 10960, Cur Loss: 0.23934305, Cur Avg Loss: 0.13008105, Log Avg loss: 0.14897526, Global Avg Loss: 0.57597764, Time: 0.0208 Steps: 96490, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000695, Sample Num: 11120, Cur Loss: 0.03732121, Cur Avg Loss: 0.12960682, Log Avg loss: 0.09712222, Global Avg Loss: 0.57592802, Time: 0.0208 Steps: 96500, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000705, Sample Num: 11280, Cur Loss: 0.17089254, Cur Avg Loss: 0.12960832, Log Avg loss: 0.12971229, Global Avg Loss: 0.57588179, Time: 0.0208 Steps: 96510, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000715, Sample Num: 11440, Cur Loss: 0.10255109, Cur Avg Loss: 0.12972487, Log Avg loss: 0.13794197, Global Avg Loss: 0.57583641, Time: 0.0208 Steps: 96520, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000725, Sample Num: 11600, Cur Loss: 0.07753417, Cur Avg Loss: 0.12893502, Log Avg loss: 0.07246071, Global Avg Loss: 0.57578427, Time: 0.0208 Steps: 96530, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000735, Sample Num: 11760, Cur Loss: 0.01954448, Cur Avg Loss: 0.12898992, Log Avg loss: 0.13297007, Global Avg Loss: 0.57573840, Time: 0.0209 Steps: 96540, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000745, Sample Num: 11920, Cur Loss: 0.05387419, Cur Avg Loss: 0.12998183, Log Avg loss: 0.20288727, Global Avg Loss: 0.57569978, Time: 0.0209 Steps: 96550, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000755, Sample Num: 12080, Cur Loss: 0.12870693, Cur Avg Loss: 0.13076364, Log Avg loss: 0.18900791, Global Avg Loss: 0.57565973, Time: 0.0208 Steps: 96560, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000765, Sample Num: 12240, Cur Loss: 0.14232501, Cur Avg Loss: 0.13078112, Log Avg loss: 0.13210084, Global Avg Loss: 0.57561380, Time: 0.0209 Steps: 96570, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000775, Sample Num: 12400, Cur Loss: 0.08072285, Cur Avg Loss: 0.13122159, Log Avg loss: 0.16491791, Global Avg Loss: 0.57557128, Time: 0.0209 Steps: 96580, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000785, Sample Num: 12560, Cur Loss: 0.21385758, Cur Avg Loss: 0.13157991, Log Avg loss: 0.15934936, Global Avg Loss: 0.57552819, Time: 0.0209 Steps: 96590, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000795, Sample Num: 12720, Cur Loss: 0.08395845, Cur Avg Loss: 0.13097259, Log Avg loss: 0.08329792, Global Avg Loss: 0.57547723, Time: 0.0208 Steps: 96600, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000805, Sample Num: 12880, Cur Loss: 0.15132785, Cur Avg Loss: 0.13117360, Log Avg loss: 0.14715419, Global Avg Loss: 0.57543290, Time: 0.0209 Steps: 96610, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000815, Sample Num: 13040, Cur Loss: 0.07341595, Cur Avg Loss: 0.13136528, Log Avg loss: 0.14679521, Global Avg Loss: 0.57538853, Time: 0.0209 Steps: 96620, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000825, Sample Num: 13200, Cur Loss: 0.07571809, Cur Avg Loss: 0.13140216, Log Avg loss: 0.13440853, Global Avg Loss: 0.57534290, Time: 0.0209 Steps: 96630, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000835, Sample Num: 13360, Cur Loss: 0.03144196, Cur Avg Loss: 0.13148518, Log Avg loss: 0.13833414, Global Avg Loss: 0.57529768, Time: 0.0208 Steps: 96640, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000845, Sample Num: 13520, Cur Loss: 0.08354456, Cur Avg Loss: 0.13156828, Log Avg loss: 0.13850712, Global Avg Loss: 0.57525248, Time: 0.0208 Steps: 96650, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000855, Sample Num: 13680, Cur Loss: 0.02067772, Cur Avg Loss: 0.13116412, Log Avg loss: 0.09701270, Global Avg Loss: 0.57520301, Time: 0.0209 Steps: 96660, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000865, Sample Num: 13840, Cur Loss: 0.04982637, Cur Avg Loss: 0.13138818, Log Avg loss: 0.15054508, Global Avg Loss: 0.57515908, Time: 0.0208 Steps: 96670, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000875, Sample Num: 14000, Cur Loss: 0.08093328, Cur Avg Loss: 0.13166085, Log Avg loss: 0.15524703, Global Avg Loss: 0.57511564, Time: 0.0208 Steps: 96680, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000885, Sample Num: 14160, Cur Loss: 0.36853272, Cur Avg Loss: 0.13246277, Log Avg loss: 0.20263100, Global Avg Loss: 0.57507712, Time: 0.0208 Steps: 96690, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000895, Sample Num: 14320, Cur Loss: 0.05953649, Cur Avg Loss: 0.13226368, Log Avg loss: 0.11464379, Global Avg Loss: 0.57502951, Time: 0.0208 Steps: 96700, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000905, Sample Num: 14480, Cur Loss: 0.31535515, Cur Avg Loss: 0.13259655, Log Avg loss: 0.16238820, Global Avg Loss: 0.57498684, Time: 0.0208 Steps: 96710, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000915, Sample Num: 14640, Cur Loss: 0.12742205, Cur Avg Loss: 0.13246290, Log Avg loss: 0.12036762, Global Avg Loss: 0.57493983, Time: 0.0208 Steps: 96720, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000925, Sample Num: 14800, Cur Loss: 0.21853274, Cur Avg Loss: 0.13277812, Log Avg loss: 0.16162091, Global Avg Loss: 0.57489711, Time: 0.0208 Steps: 96730, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000935, Sample Num: 14960, Cur Loss: 0.15018176, Cur Avg Loss: 0.13293094, Log Avg loss: 0.14706675, Global Avg Loss: 0.57485288, Time: 0.0208 Steps: 96740, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000945, Sample Num: 15120, Cur Loss: 0.09618421, Cur Avg Loss: 0.13263734, Log Avg loss: 0.10518595, Global Avg Loss: 0.57480434, Time: 0.0208 Steps: 96750, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000955, Sample Num: 15280, Cur Loss: 0.18956491, Cur Avg Loss: 0.13264640, Log Avg loss: 0.13350212, Global Avg Loss: 0.57475873, Time: 0.0208 Steps: 96760, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000965, Sample Num: 15440, Cur Loss: 0.07277746, Cur Avg Loss: 0.13280999, Log Avg loss: 0.14843363, Global Avg Loss: 0.57471467, Time: 0.0208 Steps: 96770, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000975, Sample Num: 15600, Cur Loss: 0.03123777, Cur Avg Loss: 0.13238343, Log Avg loss: 0.09121994, Global Avg Loss: 0.57466471, Time: 0.0208 Steps: 96780, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000985, Sample Num: 15760, Cur Loss: 0.15108645, Cur Avg Loss: 0.13213419, Log Avg loss: 0.10783356, Global Avg Loss: 0.57461648, Time: 0.0208 Steps: 96790, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000995, Sample Num: 15920, Cur Loss: 0.16774362, Cur Avg Loss: 0.13218559, Log Avg loss: 0.13724808, Global Avg Loss: 0.57457130, Time: 0.0208 Steps: 96800, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001005, Sample Num: 16080, Cur Loss: 0.06621650, Cur Avg Loss: 0.13225244, Log Avg loss: 0.13890371, Global Avg Loss: 0.57452630, Time: 0.0208 Steps: 96810, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001015, Sample Num: 16240, Cur Loss: 0.19380587, Cur Avg Loss: 0.13262368, Log Avg loss: 0.16993397, Global Avg Loss: 0.57448451, Time: 0.0208 Steps: 96820, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001025, Sample Num: 16400, Cur Loss: 0.36383995, Cur Avg Loss: 0.13262673, Log Avg loss: 0.13293564, Global Avg Loss: 0.57443891, Time: 0.0245 Steps: 96830, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001035, Sample Num: 16560, Cur Loss: 0.13105005, Cur Avg Loss: 0.13244504, Log Avg loss: 0.11382207, Global Avg Loss: 0.57439135, Time: 0.0209 Steps: 96840, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001045, Sample Num: 16720, Cur Loss: 0.16127342, Cur Avg Loss: 0.13274002, Log Avg loss: 0.16327034, Global Avg Loss: 0.57434890, Time: 0.0208 Steps: 96850, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001055, Sample Num: 16880, Cur Loss: 0.39740664, Cur Avg Loss: 0.13322482, Log Avg loss: 0.18388641, Global Avg Loss: 0.57430858, Time: 0.0209 Steps: 96860, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001065, Sample Num: 17040, Cur Loss: 0.05282286, Cur Avg Loss: 0.13317334, Log Avg loss: 0.12774297, Global Avg Loss: 0.57426248, Time: 0.0209 Steps: 96870, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001075, Sample Num: 17200, Cur Loss: 0.09014864, Cur Avg Loss: 0.13290394, Log Avg loss: 0.10421259, Global Avg Loss: 0.57421397, Time: 0.0208 Steps: 96880, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001085, Sample Num: 17360, Cur Loss: 0.04464393, Cur Avg Loss: 0.13265515, Log Avg loss: 0.10591018, Global Avg Loss: 0.57416563, Time: 0.0207 Steps: 96890, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001095, Sample Num: 17520, Cur Loss: 0.02355888, Cur Avg Loss: 0.13290693, Log Avg loss: 0.16022436, Global Avg Loss: 0.57412291, Time: 0.0208 Steps: 96900, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001105, Sample Num: 17680, Cur Loss: 0.08597980, Cur Avg Loss: 0.13260172, Log Avg loss: 0.09918201, Global Avg Loss: 0.57407391, Time: 0.0209 Steps: 96910, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001115, Sample Num: 17840, Cur Loss: 0.04480495, Cur Avg Loss: 0.13216869, Log Avg loss: 0.08431862, Global Avg Loss: 0.57402337, Time: 0.0209 Steps: 96920, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001125, Sample Num: 18000, Cur Loss: 0.08125994, Cur Avg Loss: 0.13215375, Log Avg loss: 0.13048825, Global Avg Loss: 0.57397762, Time: 0.0209 Steps: 96930, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001135, Sample Num: 18160, Cur Loss: 0.06161147, Cur Avg Loss: 0.13239169, Log Avg loss: 0.15915907, Global Avg Loss: 0.57393482, Time: 0.0208 Steps: 96940, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001145, Sample Num: 18320, Cur Loss: 0.32274163, Cur Avg Loss: 0.13258560, Log Avg loss: 0.15459463, Global Avg Loss: 0.57389157, Time: 0.0208 Steps: 96950, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001155, Sample Num: 18480, Cur Loss: 0.09330171, Cur Avg Loss: 0.13237015, Log Avg loss: 0.10770138, Global Avg Loss: 0.57384349, Time: 0.0208 Steps: 96960, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001165, Sample Num: 18640, Cur Loss: 0.12673980, Cur Avg Loss: 0.13265940, Log Avg loss: 0.16606765, Global Avg Loss: 0.57380144, Time: 0.0208 Steps: 96970, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001175, Sample Num: 18800, Cur Loss: 0.09993720, Cur Avg Loss: 0.13239449, Log Avg loss: 0.10153238, Global Avg Loss: 0.57375274, Time: 0.0208 Steps: 96980, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001185, Sample Num: 18960, Cur Loss: 0.07888216, Cur Avg Loss: 0.13214203, Log Avg loss: 0.10247848, Global Avg Loss: 0.57370415, Time: 0.0208 Steps: 96990, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001195, Sample Num: 19120, Cur Loss: 0.05706657, Cur Avg Loss: 0.13220426, Log Avg loss: 0.13957812, Global Avg Loss: 0.57365940, Time: 0.0208 Steps: 97000, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001205, Sample Num: 19280, Cur Loss: 0.07659861, Cur Avg Loss: 0.13193795, Log Avg loss: 0.10011370, Global Avg Loss: 0.57361058, Time: 0.0207 Steps: 97010, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001215, Sample Num: 19440, Cur Loss: 0.07005879, Cur Avg Loss: 0.13194777, Log Avg loss: 0.13313119, Global Avg Loss: 0.57356518, Time: 0.0208 Steps: 97020, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001225, Sample Num: 19600, Cur Loss: 0.19522339, Cur Avg Loss: 0.13252816, Log Avg loss: 0.20304621, Global Avg Loss: 0.57352699, Time: 0.0208 Steps: 97030, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001235, Sample Num: 19760, Cur Loss: 0.11855017, Cur Avg Loss: 0.13242607, Log Avg loss: 0.11991998, Global Avg Loss: 0.57348025, Time: 0.0208 Steps: 97040, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001245, Sample Num: 19920, Cur Loss: 0.18504366, Cur Avg Loss: 0.13233073, Log Avg loss: 0.12055551, Global Avg Loss: 0.57343358, Time: 0.0209 Steps: 97050, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001255, Sample Num: 20080, Cur Loss: 0.07378801, Cur Avg Loss: 0.13250463, Log Avg loss: 0.15415599, Global Avg Loss: 0.57339038, Time: 0.0208 Steps: 97060, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001265, Sample Num: 20240, Cur Loss: 0.08062303, Cur Avg Loss: 0.13286539, Log Avg loss: 0.17814086, Global Avg Loss: 0.57334967, Time: 0.0208 Steps: 97070, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001275, Sample Num: 20400, Cur Loss: 0.03491881, Cur Avg Loss: 0.13302261, Log Avg loss: 0.15291087, Global Avg Loss: 0.57330636, Time: 0.0208 Steps: 97080, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001285, Sample Num: 20560, Cur Loss: 0.07045233, Cur Avg Loss: 0.13347352, Log Avg loss: 0.19096398, Global Avg Loss: 0.57326698, Time: 0.0209 Steps: 97090, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001295, Sample Num: 20720, Cur Loss: 0.12017868, Cur Avg Loss: 0.13338913, Log Avg loss: 0.12254542, Global Avg Loss: 0.57322056, Time: 0.0208 Steps: 97100, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001305, Sample Num: 20880, Cur Loss: 0.02231203, Cur Avg Loss: 0.13360148, Log Avg loss: 0.16110087, Global Avg Loss: 0.57317812, Time: 0.0208 Steps: 97110, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001315, Sample Num: 21040, Cur Loss: 0.03508566, Cur Avg Loss: 0.13326791, Log Avg loss: 0.08973705, Global Avg Loss: 0.57312834, Time: 0.0208 Steps: 97120, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001325, Sample Num: 21200, Cur Loss: 0.23830356, Cur Avg Loss: 0.13306928, Log Avg loss: 0.10694902, Global Avg Loss: 0.57308035, Time: 0.0207 Steps: 97130, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001335, Sample Num: 21360, Cur Loss: 0.28339413, Cur Avg Loss: 0.13290405, Log Avg loss: 0.11101137, Global Avg Loss: 0.57303278, Time: 0.0207 Steps: 97140, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001345, Sample Num: 21520, Cur Loss: 0.11368989, Cur Avg Loss: 0.13309596, Log Avg loss: 0.15871555, Global Avg Loss: 0.57299013, Time: 0.0207 Steps: 97150, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001355, Sample Num: 21680, Cur Loss: 0.07441172, Cur Avg Loss: 0.13295991, Log Avg loss: 0.11466184, Global Avg Loss: 0.57294296, Time: 0.0208 Steps: 97160, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001365, Sample Num: 21840, Cur Loss: 0.05586183, Cur Avg Loss: 0.13314557, Log Avg loss: 0.15830228, Global Avg Loss: 0.57290029, Time: 0.0207 Steps: 97170, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001375, Sample Num: 22000, Cur Loss: 0.10128096, Cur Avg Loss: 0.13279702, Log Avg loss: 0.08521998, Global Avg Loss: 0.57285010, Time: 0.0208 Steps: 97180, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001385, Sample Num: 22160, Cur Loss: 0.08144883, Cur Avg Loss: 0.13290278, Log Avg loss: 0.14744443, Global Avg Loss: 0.57280633, Time: 0.0208 Steps: 97190, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001395, Sample Num: 22320, Cur Loss: 0.09784287, Cur Avg Loss: 0.13309596, Log Avg loss: 0.15985105, Global Avg Loss: 0.57276385, Time: 0.0207 Steps: 97200, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001405, Sample Num: 22480, Cur Loss: 0.21770376, Cur Avg Loss: 0.13324080, Log Avg loss: 0.15344605, Global Avg Loss: 0.57272071, Time: 0.0209 Steps: 97210, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001415, Sample Num: 22640, Cur Loss: 0.36432105, Cur Avg Loss: 0.13340136, Log Avg loss: 0.15595995, Global Avg Loss: 0.57267785, Time: 0.0209 Steps: 97220, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001425, Sample Num: 22800, Cur Loss: 0.09117097, Cur Avg Loss: 0.13311799, Log Avg loss: 0.09302170, Global Avg Loss: 0.57262851, Time: 0.0208 Steps: 97230, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001435, Sample Num: 22960, Cur Loss: 0.31176156, Cur Avg Loss: 0.13335711, Log Avg loss: 0.16743141, Global Avg Loss: 0.57258684, Time: 0.0209 Steps: 97240, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001445, Sample Num: 23120, Cur Loss: 0.06207877, Cur Avg Loss: 0.13323723, Log Avg loss: 0.11603415, Global Avg Loss: 0.57253990, Time: 0.0209 Steps: 97250, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001455, Sample Num: 23280, Cur Loss: 0.52040929, Cur Avg Loss: 0.13343968, Log Avg loss: 0.16269361, Global Avg Loss: 0.57249776, Time: 0.0207 Steps: 97260, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001465, Sample Num: 23440, Cur Loss: 0.13236600, Cur Avg Loss: 0.13365608, Log Avg loss: 0.16514325, Global Avg Loss: 0.57245588, Time: 0.0207 Steps: 97270, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001475, Sample Num: 23600, Cur Loss: 0.04653680, Cur Avg Loss: 0.13357424, Log Avg loss: 0.12158485, Global Avg Loss: 0.57240953, Time: 0.0209 Steps: 97280, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001485, Sample Num: 23760, Cur Loss: 0.19099131, Cur Avg Loss: 0.13334659, Log Avg loss: 0.09976833, Global Avg Loss: 0.57236095, Time: 0.0208 Steps: 97290, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001495, Sample Num: 23920, Cur Loss: 0.13261056, Cur Avg Loss: 0.13324443, Log Avg loss: 0.11807364, Global Avg Loss: 0.57231426, Time: 0.0209 Steps: 97300, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001505, Sample Num: 24080, Cur Loss: 0.13407825, Cur Avg Loss: 0.13331105, Log Avg loss: 0.14326949, Global Avg Loss: 0.57227017, Time: 0.0208 Steps: 97310, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001515, Sample Num: 24240, Cur Loss: 0.11291621, Cur Avg Loss: 0.13337630, Log Avg loss: 0.14319748, Global Avg Loss: 0.57222608, Time: 0.0207 Steps: 97320, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001525, Sample Num: 24400, Cur Loss: 0.08969294, Cur Avg Loss: 0.13360884, Log Avg loss: 0.16883738, Global Avg Loss: 0.57218464, Time: 0.0207 Steps: 97330, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001535, Sample Num: 24560, Cur Loss: 0.06664852, Cur Avg Loss: 0.13389359, Log Avg loss: 0.17731936, Global Avg Loss: 0.57214407, Time: 0.0207 Steps: 97340, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001545, Sample Num: 24720, Cur Loss: 0.15185465, Cur Avg Loss: 0.13396849, Log Avg loss: 0.14546538, Global Avg Loss: 0.57210024, Time: 0.0207 Steps: 97350, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001555, Sample Num: 24880, Cur Loss: 0.09813960, Cur Avg Loss: 0.13364569, Log Avg loss: 0.08377235, Global Avg Loss: 0.57205009, Time: 0.0207 Steps: 97360, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001565, Sample Num: 25040, Cur Loss: 0.10304181, Cur Avg Loss: 0.13381744, Log Avg loss: 0.16052503, Global Avg Loss: 0.57200782, Time: 0.0207 Steps: 97370, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001575, Sample Num: 25200, Cur Loss: 0.26772624, Cur Avg Loss: 0.13382786, Log Avg loss: 0.13545906, Global Avg Loss: 0.57196299, Time: 0.0208 Steps: 97380, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001585, Sample Num: 25360, Cur Loss: 0.13085034, Cur Avg Loss: 0.13389989, Log Avg loss: 0.14524339, Global Avg Loss: 0.57191918, Time: 0.0209 Steps: 97390, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001595, Sample Num: 25520, Cur Loss: 0.11506005, Cur Avg Loss: 0.13374806, Log Avg loss: 0.10968383, Global Avg Loss: 0.57187172, Time: 0.0208 Steps: 97400, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001605, Sample Num: 25680, Cur Loss: 0.22651607, Cur Avg Loss: 0.13397134, Log Avg loss: 0.16958381, Global Avg Loss: 0.57183042, Time: 0.0207 Steps: 97410, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001615, Sample Num: 25840, Cur Loss: 0.15801293, Cur Avg Loss: 0.13400436, Log Avg loss: 0.13930420, Global Avg Loss: 0.57178602, Time: 0.0208 Steps: 97420, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001625, Sample Num: 26000, Cur Loss: 0.18108009, Cur Avg Loss: 0.13411928, Log Avg loss: 0.15267965, Global Avg Loss: 0.57174301, Time: 0.0208 Steps: 97430, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001635, Sample Num: 26160, Cur Loss: 0.23582885, Cur Avg Loss: 0.13417066, Log Avg loss: 0.14251899, Global Avg Loss: 0.57169896, Time: 0.0208 Steps: 97440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001645, Sample Num: 26320, Cur Loss: 0.16586146, Cur Avg Loss: 0.13404426, Log Avg loss: 0.11337918, Global Avg Loss: 0.57165192, Time: 0.0208 Steps: 97450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001655, Sample Num: 26480, Cur Loss: 0.10592024, Cur Avg Loss: 0.13393670, Log Avg loss: 0.11624291, Global Avg Loss: 0.57160520, Time: 0.0208 Steps: 97460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001665, Sample Num: 26640, Cur Loss: 0.03362667, Cur Avg Loss: 0.13399892, Log Avg loss: 0.14429525, Global Avg Loss: 0.57156136, Time: 0.0208 Steps: 97470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001675, Sample Num: 26800, Cur Loss: 0.37375334, Cur Avg Loss: 0.13436727, Log Avg loss: 0.19569812, Global Avg Loss: 0.57152280, Time: 0.0207 Steps: 97480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001685, Sample Num: 26960, Cur Loss: 0.04429856, Cur Avg Loss: 0.13450352, Log Avg loss: 0.15732514, Global Avg Loss: 0.57148031, Time: 0.0207 Steps: 97490, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001695, Sample Num: 27120, Cur Loss: 0.19588664, Cur Avg Loss: 0.13435454, Log Avg loss: 0.10925075, Global Avg Loss: 0.57143290, Time: 0.0207 Steps: 97500, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001705, Sample Num: 27280, Cur Loss: 0.09476538, Cur Avg Loss: 0.13432491, Log Avg loss: 0.12930301, Global Avg Loss: 0.57138756, Time: 0.0208 Steps: 97510, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001715, Sample Num: 27440, Cur Loss: 0.03138406, Cur Avg Loss: 0.13426994, Log Avg loss: 0.12489717, Global Avg Loss: 0.57134178, Time: 0.0207 Steps: 97520, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001725, Sample Num: 27600, Cur Loss: 0.28255522, Cur Avg Loss: 0.13438516, Log Avg loss: 0.15414549, Global Avg Loss: 0.57129900, Time: 0.0208 Steps: 97530, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001735, Sample Num: 27760, Cur Loss: 0.20424938, Cur Avg Loss: 0.13464221, Log Avg loss: 0.17898416, Global Avg Loss: 0.57125878, Time: 0.0207 Steps: 97540, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001745, Sample Num: 27920, Cur Loss: 0.03438312, Cur Avg Loss: 0.13457899, Log Avg loss: 0.12361000, Global Avg Loss: 0.57121289, Time: 0.0208 Steps: 97550, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001755, Sample Num: 28080, Cur Loss: 0.22306696, Cur Avg Loss: 0.13457402, Log Avg loss: 0.13370620, Global Avg Loss: 0.57116805, Time: 0.0208 Steps: 97560, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001765, Sample Num: 28240, Cur Loss: 0.12246874, Cur Avg Loss: 0.13442094, Log Avg loss: 0.10755688, Global Avg Loss: 0.57112053, Time: 0.0207 Steps: 97570, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001775, Sample Num: 28400, Cur Loss: 0.14048204, Cur Avg Loss: 0.13450555, Log Avg loss: 0.14943785, Global Avg Loss: 0.57107732, Time: 0.0208 Steps: 97580, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001785, Sample Num: 28560, Cur Loss: 0.11594079, Cur Avg Loss: 0.13444344, Log Avg loss: 0.12341870, Global Avg Loss: 0.57103145, Time: 0.0208 Steps: 97590, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001795, Sample Num: 28720, Cur Loss: 0.09445551, Cur Avg Loss: 0.13429133, Log Avg loss: 0.10714053, Global Avg Loss: 0.57098392, Time: 0.0244 Steps: 97600, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001805, Sample Num: 28880, Cur Loss: 0.05296102, Cur Avg Loss: 0.13449524, Log Avg loss: 0.17109728, Global Avg Loss: 0.57094295, Time: 0.0207 Steps: 97610, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001815, Sample Num: 29040, Cur Loss: 0.09313866, Cur Avg Loss: 0.13449301, Log Avg loss: 0.13409021, Global Avg Loss: 0.57089820, Time: 0.0208 Steps: 97620, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001825, Sample Num: 29200, Cur Loss: 0.10135352, Cur Avg Loss: 0.13440778, Log Avg loss: 0.11893921, Global Avg Loss: 0.57085190, Time: 0.0207 Steps: 97630, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001835, Sample Num: 29360, Cur Loss: 0.06542394, Cur Avg Loss: 0.13444634, Log Avg loss: 0.14148258, Global Avg Loss: 0.57080793, Time: 0.0207 Steps: 97640, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001845, Sample Num: 29520, Cur Loss: 0.10763100, Cur Avg Loss: 0.13447828, Log Avg loss: 0.14033886, Global Avg Loss: 0.57076385, Time: 0.0209 Steps: 97650, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001855, Sample Num: 29680, Cur Loss: 0.08216684, Cur Avg Loss: 0.13457076, Log Avg loss: 0.15163430, Global Avg Loss: 0.57072093, Time: 0.0207 Steps: 97660, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001865, Sample Num: 29840, Cur Loss: 0.24905600, Cur Avg Loss: 0.13479334, Log Avg loss: 0.17608208, Global Avg Loss: 0.57068052, Time: 0.0207 Steps: 97670, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001875, Sample Num: 30000, Cur Loss: 0.16127133, Cur Avg Loss: 0.13489995, Log Avg loss: 0.15478169, Global Avg Loss: 0.57063795, Time: 0.0207 Steps: 97680, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001885, Sample Num: 30160, Cur Loss: 0.11876447, Cur Avg Loss: 0.13479768, Log Avg loss: 0.11562361, Global Avg Loss: 0.57059137, Time: 0.0208 Steps: 97690, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001895, Sample Num: 30320, Cur Loss: 0.30831814, Cur Avg Loss: 0.13493281, Log Avg loss: 0.16040307, Global Avg Loss: 0.57054939, Time: 0.0207 Steps: 97700, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001905, Sample Num: 30480, Cur Loss: 0.17068717, Cur Avg Loss: 0.13494191, Log Avg loss: 0.13666729, Global Avg Loss: 0.57050498, Time: 0.0207 Steps: 97710, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001915, Sample Num: 30640, Cur Loss: 0.30449778, Cur Avg Loss: 0.13503508, Log Avg loss: 0.15278346, Global Avg Loss: 0.57046223, Time: 0.0207 Steps: 97720, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001925, Sample Num: 30800, Cur Loss: 0.10362259, Cur Avg Loss: 0.13515589, Log Avg loss: 0.15829161, Global Avg Loss: 0.57042006, Time: 0.0207 Steps: 97730, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001935, Sample Num: 30960, Cur Loss: 0.12672463, Cur Avg Loss: 0.13536212, Log Avg loss: 0.17506193, Global Avg Loss: 0.57037961, Time: 0.0207 Steps: 97740, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001945, Sample Num: 31120, Cur Loss: 0.10463193, Cur Avg Loss: 0.13535240, Log Avg loss: 0.13347004, Global Avg Loss: 0.57033491, Time: 0.0207 Steps: 97750, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001955, Sample Num: 31280, Cur Loss: 0.07243542, Cur Avg Loss: 0.13564553, Log Avg loss: 0.19266105, Global Avg Loss: 0.57029628, Time: 0.0207 Steps: 97760, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001965, Sample Num: 31440, Cur Loss: 0.08223673, Cur Avg Loss: 0.13534993, Log Avg loss: 0.07755857, Global Avg Loss: 0.57024588, Time: 0.0207 Steps: 97770, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001975, Sample Num: 31600, Cur Loss: 0.14328074, Cur Avg Loss: 0.13520799, Log Avg loss: 0.10731754, Global Avg Loss: 0.57019854, Time: 0.0207 Steps: 97780, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001985, Sample Num: 31760, Cur Loss: 0.06407765, Cur Avg Loss: 0.13512071, Log Avg loss: 0.11788190, Global Avg Loss: 0.57015228, Time: 0.0207 Steps: 97790, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001995, Sample Num: 31920, Cur Loss: 0.25469232, Cur Avg Loss: 0.13524836, Log Avg loss: 0.16058843, Global Avg Loss: 0.57011041, Time: 0.0207 Steps: 97800, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002005, Sample Num: 32080, Cur Loss: 0.05705892, Cur Avg Loss: 0.13557693, Log Avg loss: 0.20112681, Global Avg Loss: 0.57007268, Time: 0.0207 Steps: 97810, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002015, Sample Num: 32240, Cur Loss: 0.29027805, Cur Avg Loss: 0.13594052, Log Avg loss: 0.20883937, Global Avg Loss: 0.57003575, Time: 0.0208 Steps: 97820, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002025, Sample Num: 32400, Cur Loss: 0.09922548, Cur Avg Loss: 0.13610998, Log Avg loss: 0.17025712, Global Avg Loss: 0.56999489, Time: 0.0208 Steps: 97830, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002035, Sample Num: 32560, Cur Loss: 0.07441696, Cur Avg Loss: 0.13612770, Log Avg loss: 0.13971508, Global Avg Loss: 0.56995091, Time: 0.0208 Steps: 97840, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002045, Sample Num: 32720, Cur Loss: 0.10784832, Cur Avg Loss: 0.13613298, Log Avg loss: 0.13720805, Global Avg Loss: 0.56990669, Time: 0.0208 Steps: 97850, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002055, Sample Num: 32880, Cur Loss: 0.06637031, Cur Avg Loss: 0.13596799, Log Avg loss: 0.10222710, Global Avg Loss: 0.56985890, Time: 0.0210 Steps: 97860, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002065, Sample Num: 33040, Cur Loss: 0.03555095, Cur Avg Loss: 0.13570935, Log Avg loss: 0.08255896, Global Avg Loss: 0.56980910, Time: 0.0208 Steps: 97870, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002075, Sample Num: 33200, Cur Loss: 0.05979740, Cur Avg Loss: 0.13565850, Log Avg loss: 0.12515773, Global Avg Loss: 0.56976368, Time: 0.0208 Steps: 97880, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002085, Sample Num: 33360, Cur Loss: 0.08533916, Cur Avg Loss: 0.13557594, Log Avg loss: 0.11844424, Global Avg Loss: 0.56971757, Time: 0.0209 Steps: 97890, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002095, Sample Num: 33520, Cur Loss: 0.56402528, Cur Avg Loss: 0.13577136, Log Avg loss: 0.17651750, Global Avg Loss: 0.56967741, Time: 0.0208 Steps: 97900, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002105, Sample Num: 33680, Cur Loss: 0.05934229, Cur Avg Loss: 0.13579900, Log Avg loss: 0.14159000, Global Avg Loss: 0.56963369, Time: 0.0208 Steps: 97910, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002115, Sample Num: 33840, Cur Loss: 0.30091631, Cur Avg Loss: 0.13587295, Log Avg loss: 0.15143782, Global Avg Loss: 0.56959098, Time: 0.0209 Steps: 97920, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002125, Sample Num: 34000, Cur Loss: 0.12742136, Cur Avg Loss: 0.13574746, Log Avg loss: 0.10920645, Global Avg Loss: 0.56954397, Time: 0.0208 Steps: 97930, Updated lr: 0.000008 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.564349, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.230114, "eval_total_loss": 161.770306, "eval_mae": 0.31452, "eval_mse": 0.230203, "eval_r2": 0.853668, "eval_sp_statistic": 0.931355, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.936445, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.191294, "test_total_loss": 96.029549, "test_mae": 0.323322, "test_mse": 0.19136, "test_r2": 0.876494, "test_sp_statistic": 0.920988, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.946973, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5695254734005623, "train_cur_epoch_loss": 288.93045343086123, "train_cur_epoch_avg_loss": 0.13571181466926313, "train_cur_epoch_time": 44.56434917449951, "train_cur_epoch_avg_time": 0.02093205691615759, "epoch": 46, "step": 97934} ################################################## Training, Epoch: 0047, Batch: 000006, Sample Num: 96, Cur Loss: 0.32066357, Cur Avg Loss: 0.14997700, Log Avg loss: 0.13669688, Global Avg Loss: 0.56949977, Time: 0.0211 Steps: 97940, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000016, Sample Num: 256, Cur Loss: 0.09641442, Cur Avg Loss: 0.12525869, Log Avg loss: 0.11042771, Global Avg Loss: 0.56945290, Time: 0.0219 Steps: 97950, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000026, Sample Num: 416, Cur Loss: 0.10773408, Cur Avg Loss: 0.11558182, Log Avg loss: 0.10009883, Global Avg Loss: 0.56940499, Time: 0.0218 Steps: 97960, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000036, Sample Num: 576, Cur Loss: 0.05792027, Cur Avg Loss: 0.11447587, Log Avg loss: 0.11160038, Global Avg Loss: 0.56935826, Time: 0.0210 Steps: 97970, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000046, Sample Num: 736, Cur Loss: 0.09656119, Cur Avg Loss: 0.12555101, Log Avg loss: 0.16542151, Global Avg Loss: 0.56931703, Time: 0.0210 Steps: 97980, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000056, Sample Num: 896, Cur Loss: 0.10108246, Cur Avg Loss: 0.12543247, Log Avg loss: 0.12488720, Global Avg Loss: 0.56927168, Time: 0.0211 Steps: 97990, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000066, Sample Num: 1056, Cur Loss: 0.16047487, Cur Avg Loss: 0.12527078, Log Avg loss: 0.12436531, Global Avg Loss: 0.56922628, Time: 0.0210 Steps: 98000, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000076, Sample Num: 1216, Cur Loss: 0.10632338, Cur Avg Loss: 0.12689261, Log Avg loss: 0.13759668, Global Avg Loss: 0.56918224, Time: 0.0219 Steps: 98010, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000086, Sample Num: 1376, Cur Loss: 0.22421281, Cur Avg Loss: 0.12460428, Log Avg loss: 0.10721300, Global Avg Loss: 0.56913511, Time: 0.0219 Steps: 98020, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000096, Sample Num: 1536, Cur Loss: 0.10279316, Cur Avg Loss: 0.12267015, Log Avg loss: 0.10603662, Global Avg Loss: 0.56908787, Time: 0.0219 Steps: 98030, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000106, Sample Num: 1696, Cur Loss: 0.17413247, Cur Avg Loss: 0.12296700, Log Avg loss: 0.12581677, Global Avg Loss: 0.56904266, Time: 0.0210 Steps: 98040, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000116, Sample Num: 1856, Cur Loss: 0.29545203, Cur Avg Loss: 0.12491564, Log Avg loss: 0.14557122, Global Avg Loss: 0.56899947, Time: 0.0210 Steps: 98050, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000126, Sample Num: 2016, Cur Loss: 0.05317803, Cur Avg Loss: 0.12262947, Log Avg loss: 0.09610986, Global Avg Loss: 0.56895124, Time: 0.0219 Steps: 98060, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000136, Sample Num: 2176, Cur Loss: 0.10750341, Cur Avg Loss: 0.12445000, Log Avg loss: 0.14738873, Global Avg Loss: 0.56890826, Time: 0.0210 Steps: 98070, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000146, Sample Num: 2336, Cur Loss: 0.08275047, Cur Avg Loss: 0.12358722, Log Avg loss: 0.11185340, Global Avg Loss: 0.56886166, Time: 0.0219 Steps: 98080, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000156, Sample Num: 2496, Cur Loss: 0.22221516, Cur Avg Loss: 0.12454254, Log Avg loss: 0.13849021, Global Avg Loss: 0.56881778, Time: 0.0219 Steps: 98090, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000166, Sample Num: 2656, Cur Loss: 0.09317633, Cur Avg Loss: 0.12483086, Log Avg loss: 0.12932858, Global Avg Loss: 0.56877298, Time: 0.0211 Steps: 98100, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000176, Sample Num: 2816, Cur Loss: 0.18615466, Cur Avg Loss: 0.12858432, Log Avg loss: 0.19089190, Global Avg Loss: 0.56873447, Time: 0.0210 Steps: 98110, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000186, Sample Num: 2976, Cur Loss: 0.02635469, Cur Avg Loss: 0.12635667, Log Avg loss: 0.08714992, Global Avg Loss: 0.56868539, Time: 0.0210 Steps: 98120, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000196, Sample Num: 3136, Cur Loss: 0.09735818, Cur Avg Loss: 0.12531228, Log Avg loss: 0.10588663, Global Avg Loss: 0.56863822, Time: 0.0211 Steps: 98130, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000206, Sample Num: 3296, Cur Loss: 0.16844961, Cur Avg Loss: 0.12668940, Log Avg loss: 0.15368100, Global Avg Loss: 0.56859594, Time: 0.0210 Steps: 98140, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000216, Sample Num: 3456, Cur Loss: 0.01819297, Cur Avg Loss: 0.12539356, Log Avg loss: 0.09869922, Global Avg Loss: 0.56854807, Time: 0.0209 Steps: 98150, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000226, Sample Num: 3616, Cur Loss: 0.07696455, Cur Avg Loss: 0.12524172, Log Avg loss: 0.12196201, Global Avg Loss: 0.56850257, Time: 0.0219 Steps: 98160, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000236, Sample Num: 3776, Cur Loss: 0.10631654, Cur Avg Loss: 0.12537230, Log Avg loss: 0.12832332, Global Avg Loss: 0.56845773, Time: 0.0209 Steps: 98170, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000246, Sample Num: 3936, Cur Loss: 0.09773988, Cur Avg Loss: 0.12822268, Log Avg loss: 0.19549181, Global Avg Loss: 0.56841974, Time: 0.0219 Steps: 98180, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000256, Sample Num: 4096, Cur Loss: 0.07911257, Cur Avg Loss: 0.12899744, Log Avg loss: 0.14805631, Global Avg Loss: 0.56837693, Time: 0.0254 Steps: 98190, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000266, Sample Num: 4256, Cur Loss: 0.08790623, Cur Avg Loss: 0.12659280, Log Avg loss: 0.06503411, Global Avg Loss: 0.56832568, Time: 0.0207 Steps: 98200, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000276, Sample Num: 4416, Cur Loss: 0.24946955, Cur Avg Loss: 0.12662584, Log Avg loss: 0.12750468, Global Avg Loss: 0.56828079, Time: 0.0207 Steps: 98210, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000286, Sample Num: 4576, Cur Loss: 0.08013391, Cur Avg Loss: 0.12783491, Log Avg loss: 0.16120533, Global Avg Loss: 0.56823935, Time: 0.0208 Steps: 98220, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000296, Sample Num: 4736, Cur Loss: 0.05342864, Cur Avg Loss: 0.12739451, Log Avg loss: 0.11479891, Global Avg Loss: 0.56819318, Time: 0.0208 Steps: 98230, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000306, Sample Num: 4896, Cur Loss: 0.08378670, Cur Avg Loss: 0.12773370, Log Avg loss: 0.13777373, Global Avg Loss: 0.56814937, Time: 0.0210 Steps: 98240, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000316, Sample Num: 5056, Cur Loss: 0.15950863, Cur Avg Loss: 0.12751117, Log Avg loss: 0.12070175, Global Avg Loss: 0.56810383, Time: 0.0210 Steps: 98250, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000326, Sample Num: 5216, Cur Loss: 0.04613205, Cur Avg Loss: 0.12824821, Log Avg loss: 0.15153871, Global Avg Loss: 0.56806144, Time: 0.0208 Steps: 98260, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000336, Sample Num: 5376, Cur Loss: 0.16060054, Cur Avg Loss: 0.12948804, Log Avg loss: 0.16990662, Global Avg Loss: 0.56802092, Time: 0.0208 Steps: 98270, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000346, Sample Num: 5536, Cur Loss: 0.17170921, Cur Avg Loss: 0.13062423, Log Avg loss: 0.16880031, Global Avg Loss: 0.56798030, Time: 0.0208 Steps: 98280, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000356, Sample Num: 5696, Cur Loss: 0.04723896, Cur Avg Loss: 0.12999040, Log Avg loss: 0.10805958, Global Avg Loss: 0.56793351, Time: 0.0208 Steps: 98290, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000366, Sample Num: 5856, Cur Loss: 0.14732139, Cur Avg Loss: 0.13213210, Log Avg loss: 0.20837680, Global Avg Loss: 0.56789693, Time: 0.0209 Steps: 98300, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000376, Sample Num: 6016, Cur Loss: 0.02711881, Cur Avg Loss: 0.13222748, Log Avg loss: 0.13571838, Global Avg Loss: 0.56785297, Time: 0.0208 Steps: 98310, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000386, Sample Num: 6176, Cur Loss: 0.08017129, Cur Avg Loss: 0.13160433, Log Avg loss: 0.10817400, Global Avg Loss: 0.56780621, Time: 0.0210 Steps: 98320, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000396, Sample Num: 6336, Cur Loss: 0.06018316, Cur Avg Loss: 0.13089411, Log Avg loss: 0.10347955, Global Avg Loss: 0.56775899, Time: 0.0211 Steps: 98330, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000406, Sample Num: 6496, Cur Loss: 0.01814698, Cur Avg Loss: 0.13091509, Log Avg loss: 0.13174573, Global Avg Loss: 0.56771466, Time: 0.0209 Steps: 98340, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000416, Sample Num: 6656, Cur Loss: 0.10489498, Cur Avg Loss: 0.13084966, Log Avg loss: 0.12819332, Global Avg Loss: 0.56766997, Time: 0.0209 Steps: 98350, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000426, Sample Num: 6816, Cur Loss: 0.12127230, Cur Avg Loss: 0.13051884, Log Avg loss: 0.11675685, Global Avg Loss: 0.56762412, Time: 0.0208 Steps: 98360, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000436, Sample Num: 6976, Cur Loss: 0.26830524, Cur Avg Loss: 0.13102175, Log Avg loss: 0.15244567, Global Avg Loss: 0.56758192, Time: 0.0208 Steps: 98370, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000446, Sample Num: 7136, Cur Loss: 0.06843096, Cur Avg Loss: 0.13049605, Log Avg loss: 0.10757538, Global Avg Loss: 0.56753516, Time: 0.0209 Steps: 98380, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000456, Sample Num: 7296, Cur Loss: 0.01821732, Cur Avg Loss: 0.12989226, Log Avg loss: 0.10296329, Global Avg Loss: 0.56748794, Time: 0.0210 Steps: 98390, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000466, Sample Num: 7456, Cur Loss: 0.10974848, Cur Avg Loss: 0.13030577, Log Avg loss: 0.14916188, Global Avg Loss: 0.56744543, Time: 0.0210 Steps: 98400, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000476, Sample Num: 7616, Cur Loss: 0.14829777, Cur Avg Loss: 0.12973973, Log Avg loss: 0.10336206, Global Avg Loss: 0.56739827, Time: 0.0208 Steps: 98410, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000486, Sample Num: 7776, Cur Loss: 0.08846930, Cur Avg Loss: 0.12905623, Log Avg loss: 0.09652183, Global Avg Loss: 0.56735043, Time: 0.0208 Steps: 98420, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000496, Sample Num: 7936, Cur Loss: 0.13471767, Cur Avg Loss: 0.12875082, Log Avg loss: 0.11390797, Global Avg Loss: 0.56730436, Time: 0.0207 Steps: 98430, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000506, Sample Num: 8096, Cur Loss: 0.05919564, Cur Avg Loss: 0.12868035, Log Avg loss: 0.12518499, Global Avg Loss: 0.56725945, Time: 0.0208 Steps: 98440, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000516, Sample Num: 8256, Cur Loss: 0.11288536, Cur Avg Loss: 0.12908264, Log Avg loss: 0.14943823, Global Avg Loss: 0.56721701, Time: 0.0214 Steps: 98450, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000526, Sample Num: 8416, Cur Loss: 0.04206248, Cur Avg Loss: 0.12915721, Log Avg loss: 0.13300510, Global Avg Loss: 0.56717291, Time: 0.0208 Steps: 98460, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000536, Sample Num: 8576, Cur Loss: 0.14015172, Cur Avg Loss: 0.12872875, Log Avg loss: 0.10619204, Global Avg Loss: 0.56712609, Time: 0.0208 Steps: 98470, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000546, Sample Num: 8736, Cur Loss: 0.06582657, Cur Avg Loss: 0.12788783, Log Avg loss: 0.08281437, Global Avg Loss: 0.56707691, Time: 0.0208 Steps: 98480, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000556, Sample Num: 8896, Cur Loss: 0.03881110, Cur Avg Loss: 0.12767963, Log Avg loss: 0.11631181, Global Avg Loss: 0.56703115, Time: 0.0208 Steps: 98490, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000566, Sample Num: 9056, Cur Loss: 0.11235134, Cur Avg Loss: 0.12905411, Log Avg loss: 0.20547555, Global Avg Loss: 0.56699444, Time: 0.0207 Steps: 98500, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000576, Sample Num: 9216, Cur Loss: 0.13951991, Cur Avg Loss: 0.12886527, Log Avg loss: 0.11817693, Global Avg Loss: 0.56694888, Time: 0.0208 Steps: 98510, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000586, Sample Num: 9376, Cur Loss: 0.12027981, Cur Avg Loss: 0.12887470, Log Avg loss: 0.12941741, Global Avg Loss: 0.56690447, Time: 0.0208 Steps: 98520, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000596, Sample Num: 9536, Cur Loss: 0.07598411, Cur Avg Loss: 0.12856230, Log Avg loss: 0.11025568, Global Avg Loss: 0.56685812, Time: 0.0208 Steps: 98530, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000606, Sample Num: 9696, Cur Loss: 0.08128205, Cur Avg Loss: 0.12792550, Log Avg loss: 0.08997238, Global Avg Loss: 0.56680973, Time: 0.0208 Steps: 98540, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000616, Sample Num: 9856, Cur Loss: 0.11572909, Cur Avg Loss: 0.12771849, Log Avg loss: 0.11517390, Global Avg Loss: 0.56676390, Time: 0.0208 Steps: 98550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000626, Sample Num: 10016, Cur Loss: 0.45332670, Cur Avg Loss: 0.12951170, Log Avg loss: 0.23997341, Global Avg Loss: 0.56673074, Time: 0.0207 Steps: 98560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000636, Sample Num: 10176, Cur Loss: 0.11536069, Cur Avg Loss: 0.12955199, Log Avg loss: 0.13207425, Global Avg Loss: 0.56668665, Time: 0.0208 Steps: 98570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000646, Sample Num: 10336, Cur Loss: 0.03702283, Cur Avg Loss: 0.12872751, Log Avg loss: 0.07629027, Global Avg Loss: 0.56663690, Time: 0.0208 Steps: 98580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000656, Sample Num: 10496, Cur Loss: 0.09001531, Cur Avg Loss: 0.12885174, Log Avg loss: 0.13687715, Global Avg Loss: 0.56659331, Time: 0.0207 Steps: 98590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000666, Sample Num: 10656, Cur Loss: 0.06648564, Cur Avg Loss: 0.12933929, Log Avg loss: 0.16132257, Global Avg Loss: 0.56655221, Time: 0.0208 Steps: 98600, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000676, Sample Num: 10816, Cur Loss: 0.30795452, Cur Avg Loss: 0.12972670, Log Avg loss: 0.15552812, Global Avg Loss: 0.56651053, Time: 0.0208 Steps: 98610, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000686, Sample Num: 10976, Cur Loss: 0.02673983, Cur Avg Loss: 0.12979927, Log Avg loss: 0.13470514, Global Avg Loss: 0.56646674, Time: 0.0208 Steps: 98620, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000696, Sample Num: 11136, Cur Loss: 0.06032337, Cur Avg Loss: 0.12993935, Log Avg loss: 0.13954902, Global Avg Loss: 0.56642346, Time: 0.0208 Steps: 98630, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000706, Sample Num: 11296, Cur Loss: 0.14874208, Cur Avg Loss: 0.12978532, Log Avg loss: 0.11906450, Global Avg Loss: 0.56637810, Time: 0.0208 Steps: 98640, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000716, Sample Num: 11456, Cur Loss: 0.03293991, Cur Avg Loss: 0.12953833, Log Avg loss: 0.11210109, Global Avg Loss: 0.56633205, Time: 0.0208 Steps: 98650, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000726, Sample Num: 11616, Cur Loss: 0.02075964, Cur Avg Loss: 0.12894354, Log Avg loss: 0.08635625, Global Avg Loss: 0.56628340, Time: 0.0208 Steps: 98660, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000736, Sample Num: 11776, Cur Loss: 0.10132928, Cur Avg Loss: 0.12878520, Log Avg loss: 0.11728971, Global Avg Loss: 0.56623790, Time: 0.0208 Steps: 98670, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000746, Sample Num: 11936, Cur Loss: 0.08386707, Cur Avg Loss: 0.12855817, Log Avg loss: 0.11184872, Global Avg Loss: 0.56619185, Time: 0.0208 Steps: 98680, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000756, Sample Num: 12096, Cur Loss: 0.03312564, Cur Avg Loss: 0.12883150, Log Avg loss: 0.14922170, Global Avg Loss: 0.56614960, Time: 0.0208 Steps: 98690, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000766, Sample Num: 12256, Cur Loss: 0.11746627, Cur Avg Loss: 0.12876250, Log Avg loss: 0.12354639, Global Avg Loss: 0.56610476, Time: 0.0208 Steps: 98700, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000776, Sample Num: 12416, Cur Loss: 0.19873041, Cur Avg Loss: 0.12928559, Log Avg loss: 0.16935468, Global Avg Loss: 0.56606457, Time: 0.0209 Steps: 98710, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000786, Sample Num: 12576, Cur Loss: 0.15156969, Cur Avg Loss: 0.12954830, Log Avg loss: 0.14993425, Global Avg Loss: 0.56602241, Time: 0.0207 Steps: 98720, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000796, Sample Num: 12736, Cur Loss: 0.06639789, Cur Avg Loss: 0.12961326, Log Avg loss: 0.13471894, Global Avg Loss: 0.56597873, Time: 0.0208 Steps: 98730, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000806, Sample Num: 12896, Cur Loss: 0.14793655, Cur Avg Loss: 0.12973288, Log Avg loss: 0.13925440, Global Avg Loss: 0.56593551, Time: 0.0207 Steps: 98740, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000816, Sample Num: 13056, Cur Loss: 0.11061019, Cur Avg Loss: 0.12938090, Log Avg loss: 0.10101157, Global Avg Loss: 0.56588843, Time: 0.0207 Steps: 98750, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000826, Sample Num: 13216, Cur Loss: 0.27757409, Cur Avg Loss: 0.12991590, Log Avg loss: 0.17357173, Global Avg Loss: 0.56584871, Time: 0.0207 Steps: 98760, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000836, Sample Num: 13376, Cur Loss: 0.26878232, Cur Avg Loss: 0.13002088, Log Avg loss: 0.13869232, Global Avg Loss: 0.56580546, Time: 0.0207 Steps: 98770, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000846, Sample Num: 13536, Cur Loss: 0.08824833, Cur Avg Loss: 0.13043847, Log Avg loss: 0.16534901, Global Avg Loss: 0.56576492, Time: 0.0207 Steps: 98780, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000856, Sample Num: 13696, Cur Loss: 0.22648127, Cur Avg Loss: 0.13076610, Log Avg loss: 0.15848406, Global Avg Loss: 0.56572369, Time: 0.0208 Steps: 98790, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000866, Sample Num: 13856, Cur Loss: 0.04060901, Cur Avg Loss: 0.13104105, Log Avg loss: 0.15457677, Global Avg Loss: 0.56568208, Time: 0.0207 Steps: 98800, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000876, Sample Num: 14016, Cur Loss: 0.13677433, Cur Avg Loss: 0.13143038, Log Avg loss: 0.16514647, Global Avg Loss: 0.56564154, Time: 0.0208 Steps: 98810, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000886, Sample Num: 14176, Cur Loss: 0.01669312, Cur Avg Loss: 0.13165026, Log Avg loss: 0.15091123, Global Avg Loss: 0.56559957, Time: 0.0207 Steps: 98820, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000896, Sample Num: 14336, Cur Loss: 0.12498281, Cur Avg Loss: 0.13237075, Log Avg loss: 0.19620628, Global Avg Loss: 0.56556220, Time: 0.0208 Steps: 98830, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000906, Sample Num: 14496, Cur Loss: 0.08251335, Cur Avg Loss: 0.13217781, Log Avg loss: 0.11488997, Global Avg Loss: 0.56551660, Time: 0.0207 Steps: 98840, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000916, Sample Num: 14656, Cur Loss: 0.04026602, Cur Avg Loss: 0.13181873, Log Avg loss: 0.09928626, Global Avg Loss: 0.56546944, Time: 0.0207 Steps: 98850, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000926, Sample Num: 14816, Cur Loss: 0.03450848, Cur Avg Loss: 0.13183461, Log Avg loss: 0.13328971, Global Avg Loss: 0.56542572, Time: 0.0207 Steps: 98860, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000936, Sample Num: 14976, Cur Loss: 0.10343187, Cur Avg Loss: 0.13214947, Log Avg loss: 0.16130566, Global Avg Loss: 0.56538484, Time: 0.0208 Steps: 98870, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000946, Sample Num: 15136, Cur Loss: 0.13721445, Cur Avg Loss: 0.13160445, Log Avg loss: 0.08058989, Global Avg Loss: 0.56533582, Time: 0.0207 Steps: 98880, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000956, Sample Num: 15296, Cur Loss: 0.03378851, Cur Avg Loss: 0.13240496, Log Avg loss: 0.20813335, Global Avg Loss: 0.56529970, Time: 0.0208 Steps: 98890, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000966, Sample Num: 15456, Cur Loss: 0.11435015, Cur Avg Loss: 0.13213708, Log Avg loss: 0.10652820, Global Avg Loss: 0.56525331, Time: 0.0207 Steps: 98900, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000976, Sample Num: 15616, Cur Loss: 0.10563429, Cur Avg Loss: 0.13175935, Log Avg loss: 0.09527051, Global Avg Loss: 0.56520579, Time: 0.0208 Steps: 98910, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000986, Sample Num: 15776, Cur Loss: 0.17713653, Cur Avg Loss: 0.13144769, Log Avg loss: 0.10102950, Global Avg Loss: 0.56515887, Time: 0.0208 Steps: 98920, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000996, Sample Num: 15936, Cur Loss: 0.06050928, Cur Avg Loss: 0.13117621, Log Avg loss: 0.10440851, Global Avg Loss: 0.56511229, Time: 0.0208 Steps: 98930, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001006, Sample Num: 16096, Cur Loss: 0.25048372, Cur Avg Loss: 0.13093885, Log Avg loss: 0.10729768, Global Avg Loss: 0.56506602, Time: 0.0208 Steps: 98940, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001016, Sample Num: 16256, Cur Loss: 0.06795672, Cur Avg Loss: 0.13078390, Log Avg loss: 0.11519572, Global Avg Loss: 0.56502056, Time: 0.0208 Steps: 98950, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001026, Sample Num: 16416, Cur Loss: 0.17094505, Cur Avg Loss: 0.13067675, Log Avg loss: 0.11979015, Global Avg Loss: 0.56497557, Time: 0.0246 Steps: 98960, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001036, Sample Num: 16576, Cur Loss: 0.08333038, Cur Avg Loss: 0.13085388, Log Avg loss: 0.14902810, Global Avg Loss: 0.56493354, Time: 0.0209 Steps: 98970, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001046, Sample Num: 16736, Cur Loss: 0.10377055, Cur Avg Loss: 0.13060438, Log Avg loss: 0.10475573, Global Avg Loss: 0.56488705, Time: 0.0209 Steps: 98980, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001056, Sample Num: 16896, Cur Loss: 0.05791580, Cur Avg Loss: 0.13106065, Log Avg loss: 0.17878646, Global Avg Loss: 0.56484804, Time: 0.0209 Steps: 98990, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001066, Sample Num: 17056, Cur Loss: 0.07843876, Cur Avg Loss: 0.13152436, Log Avg loss: 0.18049229, Global Avg Loss: 0.56480922, Time: 0.0209 Steps: 99000, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001076, Sample Num: 17216, Cur Loss: 0.09566706, Cur Avg Loss: 0.13149538, Log Avg loss: 0.12840617, Global Avg Loss: 0.56476514, Time: 0.0209 Steps: 99010, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001086, Sample Num: 17376, Cur Loss: 0.07075293, Cur Avg Loss: 0.13126592, Log Avg loss: 0.10657606, Global Avg Loss: 0.56471887, Time: 0.0208 Steps: 99020, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001096, Sample Num: 17536, Cur Loss: 0.16951083, Cur Avg Loss: 0.13143701, Log Avg loss: 0.15001676, Global Avg Loss: 0.56467699, Time: 0.0209 Steps: 99030, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001106, Sample Num: 17696, Cur Loss: 0.03270961, Cur Avg Loss: 0.13146632, Log Avg loss: 0.13467905, Global Avg Loss: 0.56463358, Time: 0.0209 Steps: 99040, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001116, Sample Num: 17856, Cur Loss: 0.07640176, Cur Avg Loss: 0.13152381, Log Avg loss: 0.13788243, Global Avg Loss: 0.56459049, Time: 0.0209 Steps: 99050, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001126, Sample Num: 18016, Cur Loss: 0.10018412, Cur Avg Loss: 0.13151173, Log Avg loss: 0.13016336, Global Avg Loss: 0.56454664, Time: 0.0208 Steps: 99060, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001136, Sample Num: 18176, Cur Loss: 0.05973359, Cur Avg Loss: 0.13144731, Log Avg loss: 0.12419325, Global Avg Loss: 0.56450219, Time: 0.0209 Steps: 99070, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001146, Sample Num: 18336, Cur Loss: 0.14052604, Cur Avg Loss: 0.13129222, Log Avg loss: 0.11367419, Global Avg Loss: 0.56445669, Time: 0.0209 Steps: 99080, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001156, Sample Num: 18496, Cur Loss: 0.02735986, Cur Avg Loss: 0.13137472, Log Avg loss: 0.14082905, Global Avg Loss: 0.56441394, Time: 0.0209 Steps: 99090, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001166, Sample Num: 18656, Cur Loss: 0.19931901, Cur Avg Loss: 0.13174723, Log Avg loss: 0.17481033, Global Avg Loss: 0.56437462, Time: 0.0209 Steps: 99100, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001176, Sample Num: 18816, Cur Loss: 0.35817081, Cur Avg Loss: 0.13206113, Log Avg loss: 0.16866091, Global Avg Loss: 0.56433469, Time: 0.0209 Steps: 99110, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001186, Sample Num: 18976, Cur Loss: 0.25066984, Cur Avg Loss: 0.13215968, Log Avg loss: 0.14374992, Global Avg Loss: 0.56429226, Time: 0.0210 Steps: 99120, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001196, Sample Num: 19136, Cur Loss: 0.09872406, Cur Avg Loss: 0.13210766, Log Avg loss: 0.12593831, Global Avg Loss: 0.56424804, Time: 0.0208 Steps: 99130, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001206, Sample Num: 19296, Cur Loss: 0.07725736, Cur Avg Loss: 0.13209465, Log Avg loss: 0.13053845, Global Avg Loss: 0.56420430, Time: 0.0209 Steps: 99140, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001216, Sample Num: 19456, Cur Loss: 0.17592429, Cur Avg Loss: 0.13208691, Log Avg loss: 0.13115285, Global Avg Loss: 0.56416062, Time: 0.0209 Steps: 99150, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001226, Sample Num: 19616, Cur Loss: 0.03141486, Cur Avg Loss: 0.13259868, Log Avg loss: 0.19482972, Global Avg Loss: 0.56412337, Time: 0.0209 Steps: 99160, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001236, Sample Num: 19776, Cur Loss: 0.04029827, Cur Avg Loss: 0.13273131, Log Avg loss: 0.14899251, Global Avg Loss: 0.56408151, Time: 0.0208 Steps: 99170, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001246, Sample Num: 19936, Cur Loss: 0.07343457, Cur Avg Loss: 0.13256982, Log Avg loss: 0.11260941, Global Avg Loss: 0.56403599, Time: 0.0208 Steps: 99180, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001256, Sample Num: 20096, Cur Loss: 0.04984796, Cur Avg Loss: 0.13236879, Log Avg loss: 0.10732098, Global Avg Loss: 0.56398995, Time: 0.0209 Steps: 99190, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001266, Sample Num: 20256, Cur Loss: 0.07247986, Cur Avg Loss: 0.13235854, Log Avg loss: 0.13107043, Global Avg Loss: 0.56394631, Time: 0.0209 Steps: 99200, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001276, Sample Num: 20416, Cur Loss: 0.08518549, Cur Avg Loss: 0.13237409, Log Avg loss: 0.13434257, Global Avg Loss: 0.56390300, Time: 0.0210 Steps: 99210, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001286, Sample Num: 20576, Cur Loss: 0.16181165, Cur Avg Loss: 0.13237425, Log Avg loss: 0.13239504, Global Avg Loss: 0.56385951, Time: 0.0210 Steps: 99220, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001296, Sample Num: 20736, Cur Loss: 0.13739023, Cur Avg Loss: 0.13239304, Log Avg loss: 0.13480905, Global Avg Loss: 0.56381628, Time: 0.0208 Steps: 99230, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001306, Sample Num: 20896, Cur Loss: 0.16863228, Cur Avg Loss: 0.13225172, Log Avg loss: 0.11393753, Global Avg Loss: 0.56377094, Time: 0.0208 Steps: 99240, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001316, Sample Num: 21056, Cur Loss: 0.05551776, Cur Avg Loss: 0.13233974, Log Avg loss: 0.14383519, Global Avg Loss: 0.56372863, Time: 0.0209 Steps: 99250, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001326, Sample Num: 21216, Cur Loss: 0.14114811, Cur Avg Loss: 0.13283602, Log Avg loss: 0.19814600, Global Avg Loss: 0.56369180, Time: 0.0209 Steps: 99260, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001336, Sample Num: 21376, Cur Loss: 0.28323075, Cur Avg Loss: 0.13324121, Log Avg loss: 0.18696981, Global Avg Loss: 0.56365385, Time: 0.0208 Steps: 99270, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001346, Sample Num: 21536, Cur Loss: 0.15691242, Cur Avg Loss: 0.13302537, Log Avg loss: 0.10418918, Global Avg Loss: 0.56360757, Time: 0.0209 Steps: 99280, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001356, Sample Num: 21696, Cur Loss: 0.07679439, Cur Avg Loss: 0.13306033, Log Avg loss: 0.13776593, Global Avg Loss: 0.56356468, Time: 0.0209 Steps: 99290, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001366, Sample Num: 21856, Cur Loss: 0.25407177, Cur Avg Loss: 0.13315050, Log Avg loss: 0.14537663, Global Avg Loss: 0.56352257, Time: 0.0209 Steps: 99300, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001376, Sample Num: 22016, Cur Loss: 0.08211419, Cur Avg Loss: 0.13333939, Log Avg loss: 0.15914250, Global Avg Loss: 0.56348185, Time: 0.0208 Steps: 99310, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001386, Sample Num: 22176, Cur Loss: 0.05945406, Cur Avg Loss: 0.13323083, Log Avg loss: 0.11829326, Global Avg Loss: 0.56343703, Time: 0.0209 Steps: 99320, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001396, Sample Num: 22336, Cur Loss: 0.03528767, Cur Avg Loss: 0.13307950, Log Avg loss: 0.11210479, Global Avg Loss: 0.56339159, Time: 0.0209 Steps: 99330, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001406, Sample Num: 22496, Cur Loss: 0.04813311, Cur Avg Loss: 0.13305308, Log Avg loss: 0.12936413, Global Avg Loss: 0.56334790, Time: 0.0209 Steps: 99340, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001416, Sample Num: 22656, Cur Loss: 0.04404698, Cur Avg Loss: 0.13286498, Log Avg loss: 0.10641793, Global Avg Loss: 0.56330191, Time: 0.0209 Steps: 99350, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001426, Sample Num: 22816, Cur Loss: 0.01750311, Cur Avg Loss: 0.13252777, Log Avg loss: 0.08477926, Global Avg Loss: 0.56325375, Time: 0.0209 Steps: 99360, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001436, Sample Num: 22976, Cur Loss: 0.08940231, Cur Avg Loss: 0.13248551, Log Avg loss: 0.12645975, Global Avg Loss: 0.56320979, Time: 0.0208 Steps: 99370, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001446, Sample Num: 23136, Cur Loss: 0.34323555, Cur Avg Loss: 0.13251946, Log Avg loss: 0.13739408, Global Avg Loss: 0.56316694, Time: 0.0208 Steps: 99380, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001456, Sample Num: 23296, Cur Loss: 0.11083797, Cur Avg Loss: 0.13244156, Log Avg loss: 0.12117714, Global Avg Loss: 0.56312247, Time: 0.0208 Steps: 99390, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001466, Sample Num: 23456, Cur Loss: 0.31033289, Cur Avg Loss: 0.13255081, Log Avg loss: 0.14845728, Global Avg Loss: 0.56308076, Time: 0.0209 Steps: 99400, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001476, Sample Num: 23616, Cur Loss: 0.10837515, Cur Avg Loss: 0.13249473, Log Avg loss: 0.12427444, Global Avg Loss: 0.56303662, Time: 0.0208 Steps: 99410, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001486, Sample Num: 23776, Cur Loss: 0.06089246, Cur Avg Loss: 0.13223251, Log Avg loss: 0.09352818, Global Avg Loss: 0.56298939, Time: 0.0210 Steps: 99420, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001496, Sample Num: 23936, Cur Loss: 0.07112078, Cur Avg Loss: 0.13243633, Log Avg loss: 0.16272399, Global Avg Loss: 0.56294913, Time: 0.0208 Steps: 99430, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001506, Sample Num: 24096, Cur Loss: 0.25526470, Cur Avg Loss: 0.13261052, Log Avg loss: 0.15866907, Global Avg Loss: 0.56290848, Time: 0.0208 Steps: 99440, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001516, Sample Num: 24256, Cur Loss: 0.05218484, Cur Avg Loss: 0.13264813, Log Avg loss: 0.13831272, Global Avg Loss: 0.56286578, Time: 0.0209 Steps: 99450, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001526, Sample Num: 24416, Cur Loss: 0.20518786, Cur Avg Loss: 0.13287282, Log Avg loss: 0.16693579, Global Avg Loss: 0.56282598, Time: 0.0209 Steps: 99460, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001536, Sample Num: 24576, Cur Loss: 0.06385711, Cur Avg Loss: 0.13270616, Log Avg loss: 0.10727358, Global Avg Loss: 0.56278018, Time: 0.0254 Steps: 99470, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001546, Sample Num: 24736, Cur Loss: 0.08392590, Cur Avg Loss: 0.13261584, Log Avg loss: 0.11874261, Global Avg Loss: 0.56273554, Time: 0.0209 Steps: 99480, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001556, Sample Num: 24896, Cur Loss: 0.08631046, Cur Avg Loss: 0.13267013, Log Avg loss: 0.14106408, Global Avg Loss: 0.56269316, Time: 0.0209 Steps: 99490, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001566, Sample Num: 25056, Cur Loss: 0.14669043, Cur Avg Loss: 0.13250700, Log Avg loss: 0.10712354, Global Avg Loss: 0.56264737, Time: 0.0209 Steps: 99500, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001576, Sample Num: 25216, Cur Loss: 0.08786486, Cur Avg Loss: 0.13244615, Log Avg loss: 0.12291712, Global Avg Loss: 0.56260318, Time: 0.0209 Steps: 99510, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001586, Sample Num: 25376, Cur Loss: 0.15480921, Cur Avg Loss: 0.13253622, Log Avg loss: 0.14673077, Global Avg Loss: 0.56256140, Time: 0.0209 Steps: 99520, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001596, Sample Num: 25536, Cur Loss: 0.05195130, Cur Avg Loss: 0.13266833, Log Avg loss: 0.15362197, Global Avg Loss: 0.56252031, Time: 0.0209 Steps: 99530, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001606, Sample Num: 25696, Cur Loss: 0.04469377, Cur Avg Loss: 0.13277493, Log Avg loss: 0.14978718, Global Avg Loss: 0.56247885, Time: 0.0208 Steps: 99540, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001616, Sample Num: 25856, Cur Loss: 0.07961785, Cur Avg Loss: 0.13287427, Log Avg loss: 0.14882947, Global Avg Loss: 0.56243729, Time: 0.0209 Steps: 99550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001626, Sample Num: 26016, Cur Loss: 0.01903326, Cur Avg Loss: 0.13284472, Log Avg loss: 0.12806820, Global Avg Loss: 0.56239366, Time: 0.0208 Steps: 99560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001636, Sample Num: 26176, Cur Loss: 0.12156919, Cur Avg Loss: 0.13268691, Log Avg loss: 0.10702839, Global Avg Loss: 0.56234793, Time: 0.0209 Steps: 99570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001646, Sample Num: 26336, Cur Loss: 0.27010989, Cur Avg Loss: 0.13292049, Log Avg loss: 0.17113397, Global Avg Loss: 0.56230864, Time: 0.0208 Steps: 99580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001656, Sample Num: 26496, Cur Loss: 0.06847495, Cur Avg Loss: 0.13309852, Log Avg loss: 0.16240218, Global Avg Loss: 0.56226849, Time: 0.0209 Steps: 99590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001666, Sample Num: 26656, Cur Loss: 0.04146852, Cur Avg Loss: 0.13277898, Log Avg loss: 0.07986232, Global Avg Loss: 0.56222006, Time: 0.0209 Steps: 99600, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001676, Sample Num: 26816, Cur Loss: 0.17422020, Cur Avg Loss: 0.13253696, Log Avg loss: 0.09221608, Global Avg Loss: 0.56217287, Time: 0.0209 Steps: 99610, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001686, Sample Num: 26976, Cur Loss: 0.16514380, Cur Avg Loss: 0.13293167, Log Avg loss: 0.19908573, Global Avg Loss: 0.56213642, Time: 0.0209 Steps: 99620, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001696, Sample Num: 27136, Cur Loss: 0.13753685, Cur Avg Loss: 0.13286857, Log Avg loss: 0.12223036, Global Avg Loss: 0.56209227, Time: 0.0209 Steps: 99630, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001706, Sample Num: 27296, Cur Loss: 0.04177665, Cur Avg Loss: 0.13282016, Log Avg loss: 0.12461000, Global Avg Loss: 0.56204836, Time: 0.0209 Steps: 99640, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001716, Sample Num: 27456, Cur Loss: 0.03327976, Cur Avg Loss: 0.13288759, Log Avg loss: 0.14439091, Global Avg Loss: 0.56200645, Time: 0.0209 Steps: 99650, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001726, Sample Num: 27616, Cur Loss: 0.17259645, Cur Avg Loss: 0.13288594, Log Avg loss: 0.13260254, Global Avg Loss: 0.56196336, Time: 0.0209 Steps: 99660, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001736, Sample Num: 27776, Cur Loss: 0.22075628, Cur Avg Loss: 0.13272925, Log Avg loss: 0.10568411, Global Avg Loss: 0.56191758, Time: 0.0210 Steps: 99670, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001746, Sample Num: 27936, Cur Loss: 0.22943461, Cur Avg Loss: 0.13276170, Log Avg loss: 0.13839579, Global Avg Loss: 0.56187510, Time: 0.0209 Steps: 99680, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001756, Sample Num: 28096, Cur Loss: 0.15202320, Cur Avg Loss: 0.13315644, Log Avg loss: 0.20207692, Global Avg Loss: 0.56183901, Time: 0.0208 Steps: 99690, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001766, Sample Num: 28256, Cur Loss: 0.24206509, Cur Avg Loss: 0.13318622, Log Avg loss: 0.13841709, Global Avg Loss: 0.56179654, Time: 0.0209 Steps: 99700, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001776, Sample Num: 28416, Cur Loss: 0.05640151, Cur Avg Loss: 0.13322333, Log Avg loss: 0.13977593, Global Avg Loss: 0.56175421, Time: 0.0209 Steps: 99710, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001786, Sample Num: 28576, Cur Loss: 0.21852422, Cur Avg Loss: 0.13342374, Log Avg loss: 0.16901609, Global Avg Loss: 0.56171483, Time: 0.0209 Steps: 99720, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001796, Sample Num: 28736, Cur Loss: 0.39385027, Cur Avg Loss: 0.13336057, Log Avg loss: 0.12207838, Global Avg Loss: 0.56167074, Time: 0.0244 Steps: 99730, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001806, Sample Num: 28896, Cur Loss: 0.06892927, Cur Avg Loss: 0.13325943, Log Avg loss: 0.11509538, Global Avg Loss: 0.56162597, Time: 0.0210 Steps: 99740, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001816, Sample Num: 29056, Cur Loss: 0.06181547, Cur Avg Loss: 0.13312961, Log Avg loss: 0.10968401, Global Avg Loss: 0.56158066, Time: 0.0210 Steps: 99750, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001826, Sample Num: 29216, Cur Loss: 0.05756449, Cur Avg Loss: 0.13299959, Log Avg loss: 0.10938847, Global Avg Loss: 0.56153533, Time: 0.0210 Steps: 99760, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001836, Sample Num: 29376, Cur Loss: 0.26925236, Cur Avg Loss: 0.13340376, Log Avg loss: 0.20720505, Global Avg Loss: 0.56149982, Time: 0.0210 Steps: 99770, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001846, Sample Num: 29536, Cur Loss: 0.51036501, Cur Avg Loss: 0.13373992, Log Avg loss: 0.19545883, Global Avg Loss: 0.56146313, Time: 0.0210 Steps: 99780, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001856, Sample Num: 29696, Cur Loss: 0.21959797, Cur Avg Loss: 0.13377200, Log Avg loss: 0.13969446, Global Avg Loss: 0.56142087, Time: 0.0210 Steps: 99790, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001866, Sample Num: 29856, Cur Loss: 0.13067602, Cur Avg Loss: 0.13410999, Log Avg loss: 0.19684104, Global Avg Loss: 0.56138434, Time: 0.0210 Steps: 99800, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001876, Sample Num: 30016, Cur Loss: 0.18481758, Cur Avg Loss: 0.13409514, Log Avg loss: 0.13132281, Global Avg Loss: 0.56134125, Time: 0.0210 Steps: 99810, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001886, Sample Num: 30176, Cur Loss: 0.04258817, Cur Avg Loss: 0.13393680, Log Avg loss: 0.10423342, Global Avg Loss: 0.56129546, Time: 0.0210 Steps: 99820, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001896, Sample Num: 30336, Cur Loss: 0.30312401, Cur Avg Loss: 0.13406096, Log Avg loss: 0.15747609, Global Avg Loss: 0.56125501, Time: 0.0210 Steps: 99830, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001906, Sample Num: 30496, Cur Loss: 0.06104539, Cur Avg Loss: 0.13428695, Log Avg loss: 0.17713520, Global Avg Loss: 0.56121653, Time: 0.0210 Steps: 99840, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001916, Sample Num: 30656, Cur Loss: 0.10766984, Cur Avg Loss: 0.13414840, Log Avg loss: 0.10774139, Global Avg Loss: 0.56117112, Time: 0.0210 Steps: 99850, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001926, Sample Num: 30816, Cur Loss: 0.26739195, Cur Avg Loss: 0.13445368, Log Avg loss: 0.19294590, Global Avg Loss: 0.56113424, Time: 0.0210 Steps: 99860, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001936, Sample Num: 30976, Cur Loss: 0.09356596, Cur Avg Loss: 0.13429437, Log Avg loss: 0.10361125, Global Avg Loss: 0.56108843, Time: 0.0210 Steps: 99870, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001946, Sample Num: 31136, Cur Loss: 0.06800030, Cur Avg Loss: 0.13428699, Log Avg loss: 0.13285827, Global Avg Loss: 0.56104556, Time: 0.0211 Steps: 99880, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001956, Sample Num: 31296, Cur Loss: 0.05874804, Cur Avg Loss: 0.13434420, Log Avg loss: 0.14547649, Global Avg Loss: 0.56100395, Time: 0.0210 Steps: 99890, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001966, Sample Num: 31456, Cur Loss: 0.06954814, Cur Avg Loss: 0.13442021, Log Avg loss: 0.14928763, Global Avg Loss: 0.56096274, Time: 0.0210 Steps: 99900, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001976, Sample Num: 31616, Cur Loss: 0.23050278, Cur Avg Loss: 0.13459858, Log Avg loss: 0.16966590, Global Avg Loss: 0.56092358, Time: 0.0209 Steps: 99910, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001986, Sample Num: 31776, Cur Loss: 0.42477927, Cur Avg Loss: 0.13467209, Log Avg loss: 0.14919881, Global Avg Loss: 0.56088237, Time: 0.0210 Steps: 99920, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001996, Sample Num: 31936, Cur Loss: 0.24332593, Cur Avg Loss: 0.13476275, Log Avg loss: 0.15276691, Global Avg Loss: 0.56084153, Time: 0.0210 Steps: 99930, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002006, Sample Num: 32096, Cur Loss: 0.10190656, Cur Avg Loss: 0.13473399, Log Avg loss: 0.12899318, Global Avg Loss: 0.56079832, Time: 0.0210 Steps: 99940, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002016, Sample Num: 32256, Cur Loss: 0.09159979, Cur Avg Loss: 0.13483154, Log Avg loss: 0.15440083, Global Avg Loss: 0.56075766, Time: 0.0210 Steps: 99950, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002026, Sample Num: 32416, Cur Loss: 0.19269827, Cur Avg Loss: 0.13482440, Log Avg loss: 0.13338539, Global Avg Loss: 0.56071491, Time: 0.0210 Steps: 99960, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002036, Sample Num: 32576, Cur Loss: 0.14485434, Cur Avg Loss: 0.13486705, Log Avg loss: 0.14350775, Global Avg Loss: 0.56067317, Time: 0.0210 Steps: 99970, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002046, Sample Num: 32736, Cur Loss: 0.22003691, Cur Avg Loss: 0.13484474, Log Avg loss: 0.13030161, Global Avg Loss: 0.56063013, Time: 0.0210 Steps: 99980, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002056, Sample Num: 32896, Cur Loss: 0.03581448, Cur Avg Loss: 0.13458727, Log Avg loss: 0.08190818, Global Avg Loss: 0.56058225, Time: 0.0210 Steps: 99990, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002066, Sample Num: 33056, Cur Loss: 0.27724424, Cur Avg Loss: 0.13455385, Log Avg loss: 0.12768336, Global Avg Loss: 0.56053896, Time: 0.0209 Steps: 100000, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002076, Sample Num: 33216, Cur Loss: 0.11585838, Cur Avg Loss: 0.13470204, Log Avg loss: 0.16531749, Global Avg Loss: 0.56049944, Time: 0.0209 Steps: 100010, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002086, Sample Num: 33376, Cur Loss: 0.16184506, Cur Avg Loss: 0.13456857, Log Avg loss: 0.10686177, Global Avg Loss: 0.56045409, Time: 0.0209 Steps: 100020, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002096, Sample Num: 33536, Cur Loss: 0.02564554, Cur Avg Loss: 0.13427943, Log Avg loss: 0.07396304, Global Avg Loss: 0.56040545, Time: 0.0208 Steps: 100030, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002106, Sample Num: 33696, Cur Loss: 0.09217389, Cur Avg Loss: 0.13414599, Log Avg loss: 0.10617754, Global Avg Loss: 0.56036005, Time: 0.0208 Steps: 100040, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002116, Sample Num: 33856, Cur Loss: 0.24505663, Cur Avg Loss: 0.13442485, Log Avg loss: 0.19315291, Global Avg Loss: 0.56032335, Time: 0.0209 Steps: 100050, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002126, Sample Num: 34016, Cur Loss: 0.15897025, Cur Avg Loss: 0.13437255, Log Avg loss: 0.12330648, Global Avg Loss: 0.56027967, Time: 0.0209 Steps: 100060, Updated lr: 0.000006 ***** Running evaluation checkpoint-100063 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-100063 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.678521, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.302198, "eval_total_loss": 212.4453, "eval_mae": 0.38547, "eval_mse": 0.302314, "eval_r2": 0.807829, "eval_sp_statistic": 0.932493, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.936353, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.166254, "test_total_loss": 83.459465, "test_mae": 0.275707, "test_mse": 0.166315, "test_r2": 0.892659, "test_sp_statistic": 0.921234, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.948558, "test_ps_pvalue": 0.0, "lr": 6.056899004267426e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5602659174426501, "train_cur_epoch_loss": 285.98078505322337, "train_cur_epoch_avg_loss": 0.1343263433786864, "train_cur_epoch_time": 44.678520917892456, "train_cur_epoch_avg_time": 0.020985683850583588, "epoch": 47, "step": 100063} ################################################## Training, Epoch: 0048, Batch: 000007, Sample Num: 112, Cur Loss: 0.07539117, Cur Avg Loss: 0.14583385, Log Avg loss: 0.13255778, Global Avg Loss: 0.56023693, Time: 0.0233 Steps: 100070, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000017, Sample Num: 272, Cur Loss: 0.03816530, Cur Avg Loss: 0.14534165, Log Avg loss: 0.14499712, Global Avg Loss: 0.56019544, Time: 0.0211 Steps: 100080, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000027, Sample Num: 432, Cur Loss: 0.10164575, Cur Avg Loss: 0.13053722, Log Avg loss: 0.10536969, Global Avg Loss: 0.56015000, Time: 0.0210 Steps: 100090, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000037, Sample Num: 592, Cur Loss: 0.14080325, Cur Avg Loss: 0.12246852, Log Avg loss: 0.10068303, Global Avg Loss: 0.56010409, Time: 0.0210 Steps: 100100, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000047, Sample Num: 752, Cur Loss: 0.09479036, Cur Avg Loss: 0.11643652, Log Avg loss: 0.09411814, Global Avg Loss: 0.56005755, Time: 0.0209 Steps: 100110, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000057, Sample Num: 912, Cur Loss: 0.05218140, Cur Avg Loss: 0.11486767, Log Avg loss: 0.10749404, Global Avg Loss: 0.56001234, Time: 0.0210 Steps: 100120, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000067, Sample Num: 1072, Cur Loss: 0.07297975, Cur Avg Loss: 0.11395086, Log Avg loss: 0.10872508, Global Avg Loss: 0.55996727, Time: 0.0210 Steps: 100130, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000077, Sample Num: 1232, Cur Loss: 0.12395255, Cur Avg Loss: 0.11747115, Log Avg loss: 0.14105706, Global Avg Loss: 0.55992544, Time: 0.0209 Steps: 100140, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000087, Sample Num: 1392, Cur Loss: 0.03637505, Cur Avg Loss: 0.11553167, Log Avg loss: 0.10059771, Global Avg Loss: 0.55987958, Time: 0.0210 Steps: 100150, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000097, Sample Num: 1552, Cur Loss: 0.67947143, Cur Avg Loss: 0.12569553, Log Avg loss: 0.21412107, Global Avg Loss: 0.55984506, Time: 0.0209 Steps: 100160, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000107, Sample Num: 1712, Cur Loss: 0.06529618, Cur Avg Loss: 0.12668562, Log Avg loss: 0.13628955, Global Avg Loss: 0.55980277, Time: 0.0211 Steps: 100170, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000117, Sample Num: 1872, Cur Loss: 0.20570891, Cur Avg Loss: 0.12778505, Log Avg loss: 0.13954893, Global Avg Loss: 0.55976082, Time: 0.0211 Steps: 100180, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000127, Sample Num: 2032, Cur Loss: 0.07448884, Cur Avg Loss: 0.12592859, Log Avg loss: 0.10420802, Global Avg Loss: 0.55971536, Time: 0.0210 Steps: 100190, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000137, Sample Num: 2192, Cur Loss: 0.25576562, Cur Avg Loss: 0.12615742, Log Avg loss: 0.12906350, Global Avg Loss: 0.55967238, Time: 0.0209 Steps: 100200, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000147, Sample Num: 2352, Cur Loss: 0.14003924, Cur Avg Loss: 0.12817701, Log Avg loss: 0.15584548, Global Avg Loss: 0.55963208, Time: 0.0210 Steps: 100210, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000157, Sample Num: 2512, Cur Loss: 0.17149839, Cur Avg Loss: 0.13161870, Log Avg loss: 0.18221152, Global Avg Loss: 0.55959442, Time: 0.0211 Steps: 100220, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000167, Sample Num: 2672, Cur Loss: 0.23333113, Cur Avg Loss: 0.13078007, Log Avg loss: 0.11761355, Global Avg Loss: 0.55955032, Time: 0.0209 Steps: 100230, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000177, Sample Num: 2832, Cur Loss: 0.13165319, Cur Avg Loss: 0.13017618, Log Avg loss: 0.12009119, Global Avg Loss: 0.55950648, Time: 0.0211 Steps: 100240, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000187, Sample Num: 2992, Cur Loss: 0.05209664, Cur Avg Loss: 0.12681534, Log Avg loss: 0.06732842, Global Avg Loss: 0.55945739, Time: 0.0210 Steps: 100250, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000197, Sample Num: 3152, Cur Loss: 0.29152071, Cur Avg Loss: 0.12770149, Log Avg loss: 0.14427256, Global Avg Loss: 0.55941598, Time: 0.0209 Steps: 100260, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000207, Sample Num: 3312, Cur Loss: 0.06734718, Cur Avg Loss: 0.12553640, Log Avg loss: 0.08288419, Global Avg Loss: 0.55936845, Time: 0.0210 Steps: 100270, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000217, Sample Num: 3472, Cur Loss: 0.12896483, Cur Avg Loss: 0.12819848, Log Avg loss: 0.18330340, Global Avg Loss: 0.55933095, Time: 0.0209 Steps: 100280, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000227, Sample Num: 3632, Cur Loss: 0.10936649, Cur Avg Loss: 0.12647289, Log Avg loss: 0.08902764, Global Avg Loss: 0.55928405, Time: 0.0210 Steps: 100290, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000237, Sample Num: 3792, Cur Loss: 0.19121549, Cur Avg Loss: 0.12592048, Log Avg loss: 0.11338089, Global Avg Loss: 0.55923960, Time: 0.0210 Steps: 100300, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000247, Sample Num: 3952, Cur Loss: 0.16607125, Cur Avg Loss: 0.12530071, Log Avg loss: 0.11061218, Global Avg Loss: 0.55919487, Time: 0.0211 Steps: 100310, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000257, Sample Num: 4112, Cur Loss: 0.09183823, Cur Avg Loss: 0.12699245, Log Avg loss: 0.16877822, Global Avg Loss: 0.55915596, Time: 0.0246 Steps: 100320, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000267, Sample Num: 4272, Cur Loss: 0.04777228, Cur Avg Loss: 0.12655502, Log Avg loss: 0.11531311, Global Avg Loss: 0.55911172, Time: 0.0209 Steps: 100330, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000277, Sample Num: 4432, Cur Loss: 0.13118705, Cur Avg Loss: 0.12599069, Log Avg loss: 0.11092328, Global Avg Loss: 0.55906705, Time: 0.0209 Steps: 100340, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000287, Sample Num: 4592, Cur Loss: 0.08826232, Cur Avg Loss: 0.12664407, Log Avg loss: 0.14474267, Global Avg Loss: 0.55902576, Time: 0.0209 Steps: 100350, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000297, Sample Num: 4752, Cur Loss: 0.57008368, Cur Avg Loss: 0.12758413, Log Avg loss: 0.15456361, Global Avg Loss: 0.55898546, Time: 0.0209 Steps: 100360, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000307, Sample Num: 4912, Cur Loss: 0.06089318, Cur Avg Loss: 0.12823373, Log Avg loss: 0.14752704, Global Avg Loss: 0.55894447, Time: 0.0210 Steps: 100370, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000317, Sample Num: 5072, Cur Loss: 0.07491378, Cur Avg Loss: 0.12983955, Log Avg loss: 0.17913809, Global Avg Loss: 0.55890663, Time: 0.0209 Steps: 100380, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000327, Sample Num: 5232, Cur Loss: 0.05147469, Cur Avg Loss: 0.12974530, Log Avg loss: 0.12675768, Global Avg Loss: 0.55886358, Time: 0.0209 Steps: 100390, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000337, Sample Num: 5392, Cur Loss: 0.18422066, Cur Avg Loss: 0.12949936, Log Avg loss: 0.12145708, Global Avg Loss: 0.55882002, Time: 0.0209 Steps: 100400, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000347, Sample Num: 5552, Cur Loss: 0.05306705, Cur Avg Loss: 0.12883571, Log Avg loss: 0.10647086, Global Avg Loss: 0.55877497, Time: 0.0209 Steps: 100410, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000357, Sample Num: 5712, Cur Loss: 0.19581026, Cur Avg Loss: 0.12883932, Log Avg loss: 0.12896461, Global Avg Loss: 0.55873217, Time: 0.0209 Steps: 100420, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000367, Sample Num: 5872, Cur Loss: 0.10343207, Cur Avg Loss: 0.12854525, Log Avg loss: 0.11804669, Global Avg Loss: 0.55868829, Time: 0.0209 Steps: 100430, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000377, Sample Num: 6032, Cur Loss: 0.05678437, Cur Avg Loss: 0.12820806, Log Avg loss: 0.11583324, Global Avg Loss: 0.55864419, Time: 0.0209 Steps: 100440, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000387, Sample Num: 6192, Cur Loss: 0.14957242, Cur Avg Loss: 0.12750217, Log Avg loss: 0.10089019, Global Avg Loss: 0.55859862, Time: 0.0209 Steps: 100450, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000397, Sample Num: 6352, Cur Loss: 0.29449028, Cur Avg Loss: 0.12753297, Log Avg loss: 0.12872475, Global Avg Loss: 0.55855583, Time: 0.0209 Steps: 100460, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000407, Sample Num: 6512, Cur Loss: 0.03466008, Cur Avg Loss: 0.12714930, Log Avg loss: 0.11191794, Global Avg Loss: 0.55851138, Time: 0.0209 Steps: 100470, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000417, Sample Num: 6672, Cur Loss: 0.16889131, Cur Avg Loss: 0.12670668, Log Avg loss: 0.10869193, Global Avg Loss: 0.55846661, Time: 0.0209 Steps: 100480, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000427, Sample Num: 6832, Cur Loss: 0.08940330, Cur Avg Loss: 0.12807451, Log Avg loss: 0.18511305, Global Avg Loss: 0.55842946, Time: 0.0209 Steps: 100490, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000437, Sample Num: 6992, Cur Loss: 0.08701965, Cur Avg Loss: 0.12807021, Log Avg loss: 0.12788642, Global Avg Loss: 0.55838662, Time: 0.0209 Steps: 100500, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000447, Sample Num: 7152, Cur Loss: 0.08053553, Cur Avg Loss: 0.12784975, Log Avg loss: 0.11821553, Global Avg Loss: 0.55834282, Time: 0.0209 Steps: 100510, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000457, Sample Num: 7312, Cur Loss: 0.08972238, Cur Avg Loss: 0.12765072, Log Avg loss: 0.11875415, Global Avg Loss: 0.55829909, Time: 0.0209 Steps: 100520, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000467, Sample Num: 7472, Cur Loss: 0.08274279, Cur Avg Loss: 0.12787301, Log Avg loss: 0.13803170, Global Avg Loss: 0.55825729, Time: 0.0209 Steps: 100530, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000477, Sample Num: 7632, Cur Loss: 0.09284255, Cur Avg Loss: 0.12766842, Log Avg loss: 0.11811437, Global Avg Loss: 0.55821351, Time: 0.0209 Steps: 100540, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000487, Sample Num: 7792, Cur Loss: 0.04492998, Cur Avg Loss: 0.12700713, Log Avg loss: 0.09546328, Global Avg Loss: 0.55816749, Time: 0.0209 Steps: 100550, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000497, Sample Num: 7952, Cur Loss: 0.22081956, Cur Avg Loss: 0.12718699, Log Avg loss: 0.13594614, Global Avg Loss: 0.55812550, Time: 0.0209 Steps: 100560, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000507, Sample Num: 8112, Cur Loss: 0.03946268, Cur Avg Loss: 0.12710775, Log Avg loss: 0.12316950, Global Avg Loss: 0.55808225, Time: 0.0209 Steps: 100570, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000517, Sample Num: 8272, Cur Loss: 0.02421965, Cur Avg Loss: 0.12678145, Log Avg loss: 0.11023808, Global Avg Loss: 0.55803773, Time: 0.0247 Steps: 100580, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000527, Sample Num: 8432, Cur Loss: 0.09044015, Cur Avg Loss: 0.12739049, Log Avg loss: 0.15887803, Global Avg Loss: 0.55799804, Time: 0.0209 Steps: 100590, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000537, Sample Num: 8592, Cur Loss: 0.12505293, Cur Avg Loss: 0.12792557, Log Avg loss: 0.15612431, Global Avg Loss: 0.55795810, Time: 0.0209 Steps: 100600, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000547, Sample Num: 8752, Cur Loss: 0.27678472, Cur Avg Loss: 0.12800767, Log Avg loss: 0.13241639, Global Avg Loss: 0.55791580, Time: 0.0209 Steps: 100610, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000557, Sample Num: 8912, Cur Loss: 0.12250300, Cur Avg Loss: 0.12703795, Log Avg loss: 0.07399444, Global Avg Loss: 0.55786771, Time: 0.0209 Steps: 100620, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000567, Sample Num: 9072, Cur Loss: 0.21495146, Cur Avg Loss: 0.12735083, Log Avg loss: 0.14477829, Global Avg Loss: 0.55782666, Time: 0.0209 Steps: 100630, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000577, Sample Num: 9232, Cur Loss: 0.07915226, Cur Avg Loss: 0.12659484, Log Avg loss: 0.08373000, Global Avg Loss: 0.55777955, Time: 0.0209 Steps: 100640, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000587, Sample Num: 9392, Cur Loss: 0.17726174, Cur Avg Loss: 0.12601955, Log Avg loss: 0.09282555, Global Avg Loss: 0.55773335, Time: 0.0209 Steps: 100650, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000597, Sample Num: 9552, Cur Loss: 0.34253359, Cur Avg Loss: 0.12599467, Log Avg loss: 0.12453417, Global Avg Loss: 0.55769032, Time: 0.0209 Steps: 100660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000607, Sample Num: 9712, Cur Loss: 0.21833088, Cur Avg Loss: 0.12541904, Log Avg loss: 0.09105341, Global Avg Loss: 0.55764396, Time: 0.0209 Steps: 100670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000617, Sample Num: 9872, Cur Loss: 0.13763121, Cur Avg Loss: 0.12524812, Log Avg loss: 0.11487371, Global Avg Loss: 0.55759999, Time: 0.0210 Steps: 100680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000627, Sample Num: 10032, Cur Loss: 0.17029901, Cur Avg Loss: 0.12572943, Log Avg loss: 0.15542583, Global Avg Loss: 0.55756004, Time: 0.0209 Steps: 100690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000637, Sample Num: 10192, Cur Loss: 0.31665510, Cur Avg Loss: 0.12638246, Log Avg loss: 0.16732751, Global Avg Loss: 0.55752129, Time: 0.0209 Steps: 100700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000647, Sample Num: 10352, Cur Loss: 0.10554874, Cur Avg Loss: 0.12637507, Log Avg loss: 0.12590475, Global Avg Loss: 0.55747843, Time: 0.0209 Steps: 100710, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000657, Sample Num: 10512, Cur Loss: 0.24336955, Cur Avg Loss: 0.12623221, Log Avg loss: 0.11698896, Global Avg Loss: 0.55743470, Time: 0.0209 Steps: 100720, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000667, Sample Num: 10672, Cur Loss: 0.04817093, Cur Avg Loss: 0.12606929, Log Avg loss: 0.11536547, Global Avg Loss: 0.55739081, Time: 0.0209 Steps: 100730, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000677, Sample Num: 10832, Cur Loss: 0.26217619, Cur Avg Loss: 0.12596590, Log Avg loss: 0.11906985, Global Avg Loss: 0.55734730, Time: 0.0209 Steps: 100740, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000687, Sample Num: 10992, Cur Loss: 0.03827693, Cur Avg Loss: 0.12543784, Log Avg loss: 0.08968805, Global Avg Loss: 0.55730089, Time: 0.0210 Steps: 100750, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000697, Sample Num: 11152, Cur Loss: 0.07008035, Cur Avg Loss: 0.12592095, Log Avg loss: 0.15911053, Global Avg Loss: 0.55726137, Time: 0.0209 Steps: 100760, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000707, Sample Num: 11312, Cur Loss: 0.17959212, Cur Avg Loss: 0.12643474, Log Avg loss: 0.16224606, Global Avg Loss: 0.55722217, Time: 0.0209 Steps: 100770, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000717, Sample Num: 11472, Cur Loss: 0.24386790, Cur Avg Loss: 0.12627540, Log Avg loss: 0.11501027, Global Avg Loss: 0.55717829, Time: 0.0209 Steps: 100780, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000727, Sample Num: 11632, Cur Loss: 0.05358344, Cur Avg Loss: 0.12629225, Log Avg loss: 0.12749988, Global Avg Loss: 0.55713566, Time: 0.0209 Steps: 100790, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000737, Sample Num: 11792, Cur Loss: 0.09432461, Cur Avg Loss: 0.12708510, Log Avg loss: 0.18472523, Global Avg Loss: 0.55709871, Time: 0.0209 Steps: 100800, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000747, Sample Num: 11952, Cur Loss: 0.16567208, Cur Avg Loss: 0.12741531, Log Avg loss: 0.15175208, Global Avg Loss: 0.55705850, Time: 0.0209 Steps: 100810, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000757, Sample Num: 12112, Cur Loss: 0.14073862, Cur Avg Loss: 0.12759883, Log Avg loss: 0.14130768, Global Avg Loss: 0.55701727, Time: 0.0209 Steps: 100820, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000767, Sample Num: 12272, Cur Loss: 0.09424016, Cur Avg Loss: 0.12745684, Log Avg loss: 0.11670837, Global Avg Loss: 0.55697360, Time: 0.0209 Steps: 100830, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000777, Sample Num: 12432, Cur Loss: 0.07642254, Cur Avg Loss: 0.12864433, Log Avg loss: 0.21972456, Global Avg Loss: 0.55694015, Time: 0.0210 Steps: 100840, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000787, Sample Num: 12592, Cur Loss: 0.13041329, Cur Avg Loss: 0.12887027, Log Avg loss: 0.14642599, Global Avg Loss: 0.55689945, Time: 0.0208 Steps: 100850, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000797, Sample Num: 12752, Cur Loss: 0.21311101, Cur Avg Loss: 0.12907301, Log Avg loss: 0.14502888, Global Avg Loss: 0.55685861, Time: 0.0208 Steps: 100860, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000807, Sample Num: 12912, Cur Loss: 0.06194530, Cur Avg Loss: 0.12873032, Log Avg loss: 0.10141798, Global Avg Loss: 0.55681346, Time: 0.0208 Steps: 100870, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000817, Sample Num: 13072, Cur Loss: 0.17684172, Cur Avg Loss: 0.12876659, Log Avg loss: 0.13169357, Global Avg Loss: 0.55677132, Time: 0.0208 Steps: 100880, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000827, Sample Num: 13232, Cur Loss: 0.11457808, Cur Avg Loss: 0.12858853, Log Avg loss: 0.11404111, Global Avg Loss: 0.55672744, Time: 0.0208 Steps: 100890, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000837, Sample Num: 13392, Cur Loss: 0.08493314, Cur Avg Loss: 0.12864219, Log Avg loss: 0.13307971, Global Avg Loss: 0.55668545, Time: 0.0209 Steps: 100900, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000847, Sample Num: 13552, Cur Loss: 0.20540567, Cur Avg Loss: 0.12858714, Log Avg loss: 0.12397929, Global Avg Loss: 0.55664257, Time: 0.0210 Steps: 100910, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000857, Sample Num: 13712, Cur Loss: 0.07669257, Cur Avg Loss: 0.12837976, Log Avg loss: 0.11081442, Global Avg Loss: 0.55659839, Time: 0.0209 Steps: 100920, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000867, Sample Num: 13872, Cur Loss: 0.30048597, Cur Avg Loss: 0.12860045, Log Avg loss: 0.14751370, Global Avg Loss: 0.55655786, Time: 0.0209 Steps: 100930, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000877, Sample Num: 14032, Cur Loss: 0.13263957, Cur Avg Loss: 0.12840197, Log Avg loss: 0.11119386, Global Avg Loss: 0.55651374, Time: 0.0209 Steps: 100940, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000887, Sample Num: 14192, Cur Loss: 0.29408517, Cur Avg Loss: 0.12806678, Log Avg loss: 0.09867040, Global Avg Loss: 0.55646839, Time: 0.0209 Steps: 100950, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000897, Sample Num: 14352, Cur Loss: 0.20531911, Cur Avg Loss: 0.12880143, Log Avg loss: 0.19396499, Global Avg Loss: 0.55643248, Time: 0.0210 Steps: 100960, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000907, Sample Num: 14512, Cur Loss: 0.22201914, Cur Avg Loss: 0.12845215, Log Avg loss: 0.09712158, Global Avg Loss: 0.55638699, Time: 0.0209 Steps: 100970, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000917, Sample Num: 14672, Cur Loss: 0.05826137, Cur Avg Loss: 0.12851851, Log Avg loss: 0.13453784, Global Avg Loss: 0.55634522, Time: 0.0208 Steps: 100980, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000927, Sample Num: 14832, Cur Loss: 0.11747067, Cur Avg Loss: 0.12848144, Log Avg loss: 0.12508158, Global Avg Loss: 0.55630251, Time: 0.0209 Steps: 100990, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000937, Sample Num: 14992, Cur Loss: 0.11838551, Cur Avg Loss: 0.12851524, Log Avg loss: 0.13164899, Global Avg Loss: 0.55626047, Time: 0.0208 Steps: 101000, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000947, Sample Num: 15152, Cur Loss: 0.02702101, Cur Avg Loss: 0.12827729, Log Avg loss: 0.10598096, Global Avg Loss: 0.55621589, Time: 0.0208 Steps: 101010, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000957, Sample Num: 15312, Cur Loss: 0.10696436, Cur Avg Loss: 0.12862505, Log Avg loss: 0.16155841, Global Avg Loss: 0.55617682, Time: 0.0208 Steps: 101020, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000967, Sample Num: 15472, Cur Loss: 0.03529365, Cur Avg Loss: 0.12868733, Log Avg loss: 0.13464691, Global Avg Loss: 0.55613510, Time: 0.0208 Steps: 101030, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000977, Sample Num: 15632, Cur Loss: 0.10537838, Cur Avg Loss: 0.12903576, Log Avg loss: 0.16272951, Global Avg Loss: 0.55609616, Time: 0.0208 Steps: 101040, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000987, Sample Num: 15792, Cur Loss: 0.09893243, Cur Avg Loss: 0.12924601, Log Avg loss: 0.14978726, Global Avg Loss: 0.55605596, Time: 0.0209 Steps: 101050, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000997, Sample Num: 15952, Cur Loss: 0.02623262, Cur Avg Loss: 0.12946911, Log Avg loss: 0.15148877, Global Avg Loss: 0.55601592, Time: 0.0209 Steps: 101060, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001007, Sample Num: 16112, Cur Loss: 0.04663548, Cur Avg Loss: 0.12949077, Log Avg loss: 0.13165095, Global Avg Loss: 0.55597394, Time: 0.0208 Steps: 101070, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001017, Sample Num: 16272, Cur Loss: 0.18043926, Cur Avg Loss: 0.12945683, Log Avg loss: 0.12603895, Global Avg Loss: 0.55593140, Time: 0.0210 Steps: 101080, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001027, Sample Num: 16432, Cur Loss: 0.18685165, Cur Avg Loss: 0.12923559, Log Avg loss: 0.10673530, Global Avg Loss: 0.55588697, Time: 0.0247 Steps: 101090, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001037, Sample Num: 16592, Cur Loss: 0.10576719, Cur Avg Loss: 0.12946657, Log Avg loss: 0.15318805, Global Avg Loss: 0.55584713, Time: 0.0211 Steps: 101100, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001047, Sample Num: 16752, Cur Loss: 0.05182155, Cur Avg Loss: 0.12945144, Log Avg loss: 0.12788281, Global Avg Loss: 0.55580481, Time: 0.0211 Steps: 101110, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001057, Sample Num: 16912, Cur Loss: 0.12810677, Cur Avg Loss: 0.13026765, Log Avg loss: 0.21572503, Global Avg Loss: 0.55577118, Time: 0.0210 Steps: 101120, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001067, Sample Num: 17072, Cur Loss: 0.07667065, Cur Avg Loss: 0.13088611, Log Avg loss: 0.19625675, Global Avg Loss: 0.55573563, Time: 0.0210 Steps: 101130, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001077, Sample Num: 17232, Cur Loss: 0.15110591, Cur Avg Loss: 0.13073864, Log Avg loss: 0.11500426, Global Avg Loss: 0.55569205, Time: 0.0211 Steps: 101140, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001087, Sample Num: 17392, Cur Loss: 0.03598304, Cur Avg Loss: 0.13090509, Log Avg loss: 0.14883146, Global Avg Loss: 0.55565183, Time: 0.0211 Steps: 101150, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001097, Sample Num: 17552, Cur Loss: 0.15395141, Cur Avg Loss: 0.13094532, Log Avg loss: 0.13531781, Global Avg Loss: 0.55561028, Time: 0.0211 Steps: 101160, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001107, Sample Num: 17712, Cur Loss: 0.09528911, Cur Avg Loss: 0.13099388, Log Avg loss: 0.13632106, Global Avg Loss: 0.55556883, Time: 0.0210 Steps: 101170, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001117, Sample Num: 17872, Cur Loss: 0.07352570, Cur Avg Loss: 0.13143987, Log Avg loss: 0.18081120, Global Avg Loss: 0.55553179, Time: 0.0211 Steps: 101180, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001127, Sample Num: 18032, Cur Loss: 0.06006470, Cur Avg Loss: 0.13124045, Log Avg loss: 0.10896471, Global Avg Loss: 0.55548766, Time: 0.0211 Steps: 101190, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001137, Sample Num: 18192, Cur Loss: 0.22743280, Cur Avg Loss: 0.13155244, Log Avg loss: 0.16671423, Global Avg Loss: 0.55544925, Time: 0.0211 Steps: 101200, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001147, Sample Num: 18352, Cur Loss: 0.14044553, Cur Avg Loss: 0.13157114, Log Avg loss: 0.13369726, Global Avg Loss: 0.55540757, Time: 0.0211 Steps: 101210, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001157, Sample Num: 18512, Cur Loss: 0.21451928, Cur Avg Loss: 0.13114660, Log Avg loss: 0.08245205, Global Avg Loss: 0.55536085, Time: 0.0211 Steps: 101220, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001167, Sample Num: 18672, Cur Loss: 0.12144136, Cur Avg Loss: 0.13092540, Log Avg loss: 0.10533241, Global Avg Loss: 0.55531639, Time: 0.0210 Steps: 101230, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001177, Sample Num: 18832, Cur Loss: 0.04976761, Cur Avg Loss: 0.13118567, Log Avg loss: 0.16155929, Global Avg Loss: 0.55527750, Time: 0.0211 Steps: 101240, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001187, Sample Num: 18992, Cur Loss: 0.02528447, Cur Avg Loss: 0.13086961, Log Avg loss: 0.09366962, Global Avg Loss: 0.55523191, Time: 0.0211 Steps: 101250, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001197, Sample Num: 19152, Cur Loss: 0.05193241, Cur Avg Loss: 0.13076359, Log Avg loss: 0.11817803, Global Avg Loss: 0.55518875, Time: 0.0210 Steps: 101260, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001207, Sample Num: 19312, Cur Loss: 0.18100914, Cur Avg Loss: 0.13053879, Log Avg loss: 0.10363138, Global Avg Loss: 0.55514416, Time: 0.0211 Steps: 101270, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001217, Sample Num: 19472, Cur Loss: 0.12602453, Cur Avg Loss: 0.13047652, Log Avg loss: 0.12296050, Global Avg Loss: 0.55510149, Time: 0.0211 Steps: 101280, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001227, Sample Num: 19632, Cur Loss: 0.19414875, Cur Avg Loss: 0.13049184, Log Avg loss: 0.13235571, Global Avg Loss: 0.55505975, Time: 0.0211 Steps: 101290, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001237, Sample Num: 19792, Cur Loss: 0.10387557, Cur Avg Loss: 0.13048946, Log Avg loss: 0.13019711, Global Avg Loss: 0.55501781, Time: 0.0211 Steps: 101300, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001247, Sample Num: 19952, Cur Loss: 0.06923754, Cur Avg Loss: 0.13041290, Log Avg loss: 0.12094332, Global Avg Loss: 0.55497496, Time: 0.0211 Steps: 101310, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001257, Sample Num: 20112, Cur Loss: 0.03760127, Cur Avg Loss: 0.13056715, Log Avg loss: 0.14980199, Global Avg Loss: 0.55493497, Time: 0.0210 Steps: 101320, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001267, Sample Num: 20272, Cur Loss: 0.06632815, Cur Avg Loss: 0.13062259, Log Avg loss: 0.13759095, Global Avg Loss: 0.55489379, Time: 0.0208 Steps: 101330, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001277, Sample Num: 20432, Cur Loss: 0.18319416, Cur Avg Loss: 0.13030659, Log Avg loss: 0.09026889, Global Avg Loss: 0.55484794, Time: 0.0209 Steps: 101340, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001287, Sample Num: 20592, Cur Loss: 0.07283650, Cur Avg Loss: 0.13015263, Log Avg loss: 0.11049236, Global Avg Loss: 0.55480409, Time: 0.0210 Steps: 101350, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001297, Sample Num: 20752, Cur Loss: 0.11292005, Cur Avg Loss: 0.13009402, Log Avg loss: 0.12255142, Global Avg Loss: 0.55476145, Time: 0.0208 Steps: 101360, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001307, Sample Num: 20912, Cur Loss: 0.04869993, Cur Avg Loss: 0.13035770, Log Avg loss: 0.16455648, Global Avg Loss: 0.55472296, Time: 0.0208 Steps: 101370, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001317, Sample Num: 21072, Cur Loss: 0.18722206, Cur Avg Loss: 0.13054260, Log Avg loss: 0.15470880, Global Avg Loss: 0.55468350, Time: 0.0208 Steps: 101380, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001327, Sample Num: 21232, Cur Loss: 0.10579699, Cur Avg Loss: 0.13048374, Log Avg loss: 0.12273225, Global Avg Loss: 0.55464090, Time: 0.0209 Steps: 101390, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001337, Sample Num: 21392, Cur Loss: 0.05070898, Cur Avg Loss: 0.13020648, Log Avg loss: 0.09341396, Global Avg Loss: 0.55459541, Time: 0.0209 Steps: 101400, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001347, Sample Num: 21552, Cur Loss: 0.15871820, Cur Avg Loss: 0.13016180, Log Avg loss: 0.12418845, Global Avg Loss: 0.55455297, Time: 0.0209 Steps: 101410, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001357, Sample Num: 21712, Cur Loss: 0.05720506, Cur Avg Loss: 0.12981023, Log Avg loss: 0.08245298, Global Avg Loss: 0.55450642, Time: 0.0210 Steps: 101420, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001367, Sample Num: 21872, Cur Loss: 0.13399938, Cur Avg Loss: 0.12997262, Log Avg loss: 0.15201005, Global Avg Loss: 0.55446674, Time: 0.0209 Steps: 101430, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001377, Sample Num: 22032, Cur Loss: 0.22462186, Cur Avg Loss: 0.13019131, Log Avg loss: 0.16008584, Global Avg Loss: 0.55442786, Time: 0.0209 Steps: 101440, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001387, Sample Num: 22192, Cur Loss: 0.05184076, Cur Avg Loss: 0.13045847, Log Avg loss: 0.16724595, Global Avg Loss: 0.55438969, Time: 0.0209 Steps: 101450, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001397, Sample Num: 22352, Cur Loss: 0.07905973, Cur Avg Loss: 0.13057065, Log Avg loss: 0.14613015, Global Avg Loss: 0.55434945, Time: 0.0209 Steps: 101460, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001407, Sample Num: 22512, Cur Loss: 0.09959690, Cur Avg Loss: 0.13054491, Log Avg loss: 0.12694862, Global Avg Loss: 0.55430733, Time: 0.0210 Steps: 101470, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001417, Sample Num: 22672, Cur Loss: 0.31327409, Cur Avg Loss: 0.13059933, Log Avg loss: 0.13825693, Global Avg Loss: 0.55426634, Time: 0.0209 Steps: 101480, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001427, Sample Num: 22832, Cur Loss: 0.06860699, Cur Avg Loss: 0.13060766, Log Avg loss: 0.13178804, Global Avg Loss: 0.55422471, Time: 0.0209 Steps: 101490, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001437, Sample Num: 22992, Cur Loss: 0.33901089, Cur Avg Loss: 0.13063509, Log Avg loss: 0.13454886, Global Avg Loss: 0.55418336, Time: 0.0209 Steps: 101500, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001447, Sample Num: 23152, Cur Loss: 0.19412549, Cur Avg Loss: 0.13075852, Log Avg loss: 0.14849500, Global Avg Loss: 0.55414340, Time: 0.0209 Steps: 101510, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001457, Sample Num: 23312, Cur Loss: 0.15500669, Cur Avg Loss: 0.13109334, Log Avg loss: 0.17954176, Global Avg Loss: 0.55410650, Time: 0.0209 Steps: 101520, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001467, Sample Num: 23472, Cur Loss: 0.39084136, Cur Avg Loss: 0.13122479, Log Avg loss: 0.15037786, Global Avg Loss: 0.55406673, Time: 0.0210 Steps: 101530, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001477, Sample Num: 23632, Cur Loss: 0.09302215, Cur Avg Loss: 0.13107759, Log Avg loss: 0.10948396, Global Avg Loss: 0.55402295, Time: 0.0209 Steps: 101540, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001487, Sample Num: 23792, Cur Loss: 0.08120362, Cur Avg Loss: 0.13111777, Log Avg loss: 0.13705102, Global Avg Loss: 0.55398189, Time: 0.0209 Steps: 101550, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001497, Sample Num: 23952, Cur Loss: 0.06164826, Cur Avg Loss: 0.13113352, Log Avg loss: 0.13347604, Global Avg Loss: 0.55394048, Time: 0.0209 Steps: 101560, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001507, Sample Num: 24112, Cur Loss: 0.10073067, Cur Avg Loss: 0.13111860, Log Avg loss: 0.12888463, Global Avg Loss: 0.55389863, Time: 0.0209 Steps: 101570, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001517, Sample Num: 24272, Cur Loss: 0.06706744, Cur Avg Loss: 0.13099479, Log Avg loss: 0.11233750, Global Avg Loss: 0.55385516, Time: 0.0209 Steps: 101580, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001527, Sample Num: 24432, Cur Loss: 0.06301248, Cur Avg Loss: 0.13092171, Log Avg loss: 0.11983535, Global Avg Loss: 0.55381244, Time: 0.0209 Steps: 101590, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001537, Sample Num: 24592, Cur Loss: 0.10148453, Cur Avg Loss: 0.13125194, Log Avg loss: 0.18167788, Global Avg Loss: 0.55377581, Time: 0.0246 Steps: 101600, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001547, Sample Num: 24752, Cur Loss: 0.12073483, Cur Avg Loss: 0.13116860, Log Avg loss: 0.11835882, Global Avg Loss: 0.55373296, Time: 0.0210 Steps: 101610, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001557, Sample Num: 24912, Cur Loss: 0.17473096, Cur Avg Loss: 0.13119000, Log Avg loss: 0.13450082, Global Avg Loss: 0.55369171, Time: 0.0209 Steps: 101620, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001567, Sample Num: 25072, Cur Loss: 0.10662995, Cur Avg Loss: 0.13082850, Log Avg loss: 0.07454355, Global Avg Loss: 0.55364456, Time: 0.0208 Steps: 101630, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001577, Sample Num: 25232, Cur Loss: 0.07039414, Cur Avg Loss: 0.13107520, Log Avg loss: 0.16973331, Global Avg Loss: 0.55360679, Time: 0.0209 Steps: 101640, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001587, Sample Num: 25392, Cur Loss: 0.20277780, Cur Avg Loss: 0.13135614, Log Avg loss: 0.17565990, Global Avg Loss: 0.55356961, Time: 0.0210 Steps: 101650, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001597, Sample Num: 25552, Cur Loss: 0.18277282, Cur Avg Loss: 0.13148985, Log Avg loss: 0.15270915, Global Avg Loss: 0.55353018, Time: 0.0209 Steps: 101660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001607, Sample Num: 25712, Cur Loss: 0.20844203, Cur Avg Loss: 0.13173433, Log Avg loss: 0.17077843, Global Avg Loss: 0.55349253, Time: 0.0209 Steps: 101670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001617, Sample Num: 25872, Cur Loss: 0.15614241, Cur Avg Loss: 0.13163595, Log Avg loss: 0.11582640, Global Avg Loss: 0.55344949, Time: 0.0209 Steps: 101680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001627, Sample Num: 26032, Cur Loss: 0.07664349, Cur Avg Loss: 0.13179851, Log Avg loss: 0.15808365, Global Avg Loss: 0.55341061, Time: 0.0209 Steps: 101690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001637, Sample Num: 26192, Cur Loss: 0.18311216, Cur Avg Loss: 0.13174644, Log Avg loss: 0.12327468, Global Avg Loss: 0.55336831, Time: 0.0209 Steps: 101700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001647, Sample Num: 26352, Cur Loss: 0.05203469, Cur Avg Loss: 0.13164004, Log Avg loss: 0.11422241, Global Avg Loss: 0.55332514, Time: 0.0209 Steps: 101710, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001657, Sample Num: 26512, Cur Loss: 0.08793954, Cur Avg Loss: 0.13162841, Log Avg loss: 0.12971357, Global Avg Loss: 0.55328349, Time: 0.0209 Steps: 101720, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001667, Sample Num: 26672, Cur Loss: 0.20971897, Cur Avg Loss: 0.13177938, Log Avg loss: 0.15679517, Global Avg Loss: 0.55324452, Time: 0.0208 Steps: 101730, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001677, Sample Num: 26832, Cur Loss: 0.21395366, Cur Avg Loss: 0.13177233, Log Avg loss: 0.13059623, Global Avg Loss: 0.55320298, Time: 0.0209 Steps: 101740, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001687, Sample Num: 26992, Cur Loss: 0.08977982, Cur Avg Loss: 0.13149251, Log Avg loss: 0.08456646, Global Avg Loss: 0.55315692, Time: 0.0209 Steps: 101750, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001697, Sample Num: 27152, Cur Loss: 0.07291690, Cur Avg Loss: 0.13136375, Log Avg loss: 0.10964349, Global Avg Loss: 0.55311333, Time: 0.0207 Steps: 101760, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001707, Sample Num: 27312, Cur Loss: 0.19147253, Cur Avg Loss: 0.13121693, Log Avg loss: 0.10630029, Global Avg Loss: 0.55306943, Time: 0.0209 Steps: 101770, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001717, Sample Num: 27472, Cur Loss: 0.06158116, Cur Avg Loss: 0.13139683, Log Avg loss: 0.16210556, Global Avg Loss: 0.55303102, Time: 0.0209 Steps: 101780, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001727, Sample Num: 27632, Cur Loss: 0.38769448, Cur Avg Loss: 0.13143127, Log Avg loss: 0.13734505, Global Avg Loss: 0.55299018, Time: 0.0209 Steps: 101790, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001737, Sample Num: 27792, Cur Loss: 0.14729431, Cur Avg Loss: 0.13124549, Log Avg loss: 0.09916131, Global Avg Loss: 0.55294560, Time: 0.0210 Steps: 101800, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001747, Sample Num: 27952, Cur Loss: 0.06130550, Cur Avg Loss: 0.13104589, Log Avg loss: 0.09637555, Global Avg Loss: 0.55290075, Time: 0.0209 Steps: 101810, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001757, Sample Num: 28112, Cur Loss: 0.23901299, Cur Avg Loss: 0.13107337, Log Avg loss: 0.13587463, Global Avg Loss: 0.55285980, Time: 0.0208 Steps: 101820, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001767, Sample Num: 28272, Cur Loss: 0.05273986, Cur Avg Loss: 0.13091628, Log Avg loss: 0.10331482, Global Avg Loss: 0.55281565, Time: 0.0208 Steps: 101830, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001777, Sample Num: 28432, Cur Loss: 0.30561313, Cur Avg Loss: 0.13117738, Log Avg loss: 0.17731475, Global Avg Loss: 0.55277878, Time: 0.0209 Steps: 101840, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001787, Sample Num: 28592, Cur Loss: 0.29195464, Cur Avg Loss: 0.13123418, Log Avg loss: 0.14132718, Global Avg Loss: 0.55273838, Time: 0.0208 Steps: 101850, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001797, Sample Num: 28752, Cur Loss: 0.03362940, Cur Avg Loss: 0.13134954, Log Avg loss: 0.15196421, Global Avg Loss: 0.55269903, Time: 0.0209 Steps: 101860, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001807, Sample Num: 28912, Cur Loss: 0.27845722, Cur Avg Loss: 0.13136492, Log Avg loss: 0.13412798, Global Avg Loss: 0.55265795, Time: 0.0209 Steps: 101870, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001817, Sample Num: 29072, Cur Loss: 0.06460777, Cur Avg Loss: 0.13171910, Log Avg loss: 0.19572066, Global Avg Loss: 0.55262291, Time: 0.0209 Steps: 101880, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001827, Sample Num: 29232, Cur Loss: 0.23647447, Cur Avg Loss: 0.13191000, Log Avg loss: 0.16659556, Global Avg Loss: 0.55258502, Time: 0.0209 Steps: 101890, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001837, Sample Num: 29392, Cur Loss: 0.12417236, Cur Avg Loss: 0.13186417, Log Avg loss: 0.12349191, Global Avg Loss: 0.55254291, Time: 0.0209 Steps: 101900, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001847, Sample Num: 29552, Cur Loss: 0.39399189, Cur Avg Loss: 0.13190982, Log Avg loss: 0.14029538, Global Avg Loss: 0.55250246, Time: 0.0209 Steps: 101910, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001857, Sample Num: 29712, Cur Loss: 0.33552724, Cur Avg Loss: 0.13207253, Log Avg loss: 0.16212487, Global Avg Loss: 0.55246416, Time: 0.0209 Steps: 101920, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001867, Sample Num: 29872, Cur Loss: 0.20746391, Cur Avg Loss: 0.13208549, Log Avg loss: 0.13449288, Global Avg Loss: 0.55242315, Time: 0.0210 Steps: 101930, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001877, Sample Num: 30032, Cur Loss: 0.10547052, Cur Avg Loss: 0.13214678, Log Avg loss: 0.14358851, Global Avg Loss: 0.55238305, Time: 0.0209 Steps: 101940, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001887, Sample Num: 30192, Cur Loss: 0.09608969, Cur Avg Loss: 0.13230650, Log Avg loss: 0.16228540, Global Avg Loss: 0.55234479, Time: 0.0209 Steps: 101950, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001897, Sample Num: 30352, Cur Loss: 0.07498276, Cur Avg Loss: 0.13239365, Log Avg loss: 0.14883994, Global Avg Loss: 0.55230521, Time: 0.0209 Steps: 101960, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001907, Sample Num: 30512, Cur Loss: 0.06945336, Cur Avg Loss: 0.13240079, Log Avg loss: 0.13375446, Global Avg Loss: 0.55226416, Time: 0.0209 Steps: 101970, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001917, Sample Num: 30672, Cur Loss: 0.11359112, Cur Avg Loss: 0.13228773, Log Avg loss: 0.11072726, Global Avg Loss: 0.55222087, Time: 0.0209 Steps: 101980, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001927, Sample Num: 30832, Cur Loss: 0.08160242, Cur Avg Loss: 0.13215091, Log Avg loss: 0.10592344, Global Avg Loss: 0.55217711, Time: 0.0209 Steps: 101990, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001937, Sample Num: 30992, Cur Loss: 0.12942344, Cur Avg Loss: 0.13205705, Log Avg loss: 0.11396918, Global Avg Loss: 0.55213415, Time: 0.0209 Steps: 102000, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001947, Sample Num: 31152, Cur Loss: 0.46191764, Cur Avg Loss: 0.13252708, Log Avg loss: 0.22357336, Global Avg Loss: 0.55210194, Time: 0.0209 Steps: 102010, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001957, Sample Num: 31312, Cur Loss: 0.07959767, Cur Avg Loss: 0.13258016, Log Avg loss: 0.14291342, Global Avg Loss: 0.55206183, Time: 0.0209 Steps: 102020, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001967, Sample Num: 31472, Cur Loss: 0.03024200, Cur Avg Loss: 0.13241869, Log Avg loss: 0.10081889, Global Avg Loss: 0.55201760, Time: 0.0210 Steps: 102030, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001977, Sample Num: 31632, Cur Loss: 0.04041212, Cur Avg Loss: 0.13222144, Log Avg loss: 0.09342231, Global Avg Loss: 0.55197266, Time: 0.0209 Steps: 102040, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001987, Sample Num: 31792, Cur Loss: 0.19610175, Cur Avg Loss: 0.13247500, Log Avg loss: 0.18260526, Global Avg Loss: 0.55193647, Time: 0.0209 Steps: 102050, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001997, Sample Num: 31952, Cur Loss: 0.07900411, Cur Avg Loss: 0.13243006, Log Avg loss: 0.12350055, Global Avg Loss: 0.55189449, Time: 0.0209 Steps: 102060, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002007, Sample Num: 32112, Cur Loss: 0.27252841, Cur Avg Loss: 0.13263667, Log Avg loss: 0.17389607, Global Avg Loss: 0.55185745, Time: 0.0209 Steps: 102070, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002017, Sample Num: 32272, Cur Loss: 0.04309125, Cur Avg Loss: 0.13271626, Log Avg loss: 0.14868941, Global Avg Loss: 0.55181796, Time: 0.0209 Steps: 102080, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002027, Sample Num: 32432, Cur Loss: 0.14127137, Cur Avg Loss: 0.13259810, Log Avg loss: 0.10876567, Global Avg Loss: 0.55177456, Time: 0.0209 Steps: 102090, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002037, Sample Num: 32592, Cur Loss: 0.11119612, Cur Avg Loss: 0.13287904, Log Avg loss: 0.18982581, Global Avg Loss: 0.55173911, Time: 0.0209 Steps: 102100, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002047, Sample Num: 32752, Cur Loss: 0.22007635, Cur Avg Loss: 0.13272990, Log Avg loss: 0.10235059, Global Avg Loss: 0.55169510, Time: 0.0209 Steps: 102110, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002057, Sample Num: 32912, Cur Loss: 0.17733072, Cur Avg Loss: 0.13273746, Log Avg loss: 0.13428513, Global Avg Loss: 0.55165423, Time: 0.0209 Steps: 102120, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002067, Sample Num: 33072, Cur Loss: 0.29613712, Cur Avg Loss: 0.13273203, Log Avg loss: 0.13161467, Global Avg Loss: 0.55161310, Time: 0.0209 Steps: 102130, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002077, Sample Num: 33232, Cur Loss: 0.31740853, Cur Avg Loss: 0.13273160, Log Avg loss: 0.13264212, Global Avg Loss: 0.55157208, Time: 0.0209 Steps: 102140, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002087, Sample Num: 33392, Cur Loss: 0.08891194, Cur Avg Loss: 0.13257458, Log Avg loss: 0.09996242, Global Avg Loss: 0.55152787, Time: 0.0209 Steps: 102150, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002097, Sample Num: 33552, Cur Loss: 0.47195134, Cur Avg Loss: 0.13297249, Log Avg loss: 0.21601456, Global Avg Loss: 0.55149503, Time: 0.0210 Steps: 102160, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002107, Sample Num: 33712, Cur Loss: 0.14823599, Cur Avg Loss: 0.13275066, Log Avg loss: 0.08623416, Global Avg Loss: 0.55144949, Time: 0.0209 Steps: 102170, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002117, Sample Num: 33872, Cur Loss: 0.08032657, Cur Avg Loss: 0.13267151, Log Avg loss: 0.11599400, Global Avg Loss: 0.55140687, Time: 0.0209 Steps: 102180, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002127, Sample Num: 34032, Cur Loss: 0.04755911, Cur Avg Loss: 0.13247360, Log Avg loss: 0.09057587, Global Avg Loss: 0.55136178, Time: 0.0209 Steps: 102190, Updated lr: 0.000004 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.667694, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.251471, "eval_total_loss": 176.783813, "eval_mae": 0.333753, "eval_mse": 0.251568, "eval_r2": 0.840087, "eval_sp_statistic": 0.932646, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.93678, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.17646, "test_total_loss": 88.583148, "test_mae": 0.300161, "test_mse": 0.176524, "test_r2": 0.88607, "test_sp_statistic": 0.921825, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947996, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5513524230792626, "train_cur_epoch_loss": 281.91832225210965, "train_cur_epoch_avg_loss": 0.1324181880000515, "train_cur_epoch_time": 44.66769361495972, "train_cur_epoch_avg_time": 0.020980598222151113, "epoch": 48, "step": 102192} ################################################## Training, Epoch: 0049, Batch: 000008, Sample Num: 128, Cur Loss: 0.18443367, Cur Avg Loss: 0.12550654, Log Avg loss: 0.11510320, Global Avg Loss: 0.55131909, Time: 0.0212 Steps: 102200, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000018, Sample Num: 288, Cur Loss: 0.08243909, Cur Avg Loss: 0.11395313, Log Avg loss: 0.10471040, Global Avg Loss: 0.55127539, Time: 0.0209 Steps: 102210, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000028, Sample Num: 448, Cur Loss: 0.08443008, Cur Avg Loss: 0.10919408, Log Avg loss: 0.10062780, Global Avg Loss: 0.55123131, Time: 0.0210 Steps: 102220, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000038, Sample Num: 608, Cur Loss: 0.01212564, Cur Avg Loss: 0.10529285, Log Avg loss: 0.09436941, Global Avg Loss: 0.55118662, Time: 0.0210 Steps: 102230, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000048, Sample Num: 768, Cur Loss: 0.04110705, Cur Avg Loss: 0.10347768, Log Avg loss: 0.09658004, Global Avg Loss: 0.55114215, Time: 0.0210 Steps: 102240, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000058, Sample Num: 928, Cur Loss: 0.03630399, Cur Avg Loss: 0.11128495, Log Avg loss: 0.14875981, Global Avg Loss: 0.55110280, Time: 0.0210 Steps: 102250, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000068, Sample Num: 1088, Cur Loss: 0.17675689, Cur Avg Loss: 0.11342733, Log Avg loss: 0.12585318, Global Avg Loss: 0.55106122, Time: 0.0210 Steps: 102260, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000078, Sample Num: 1248, Cur Loss: 0.29959616, Cur Avg Loss: 0.11263433, Log Avg loss: 0.10724192, Global Avg Loss: 0.55101782, Time: 0.0209 Steps: 102270, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000088, Sample Num: 1408, Cur Loss: 0.10493183, Cur Avg Loss: 0.11171020, Log Avg loss: 0.10450199, Global Avg Loss: 0.55097416, Time: 0.0210 Steps: 102280, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000098, Sample Num: 1568, Cur Loss: 0.06351510, Cur Avg Loss: 0.11335350, Log Avg loss: 0.12781449, Global Avg Loss: 0.55093279, Time: 0.0210 Steps: 102290, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000108, Sample Num: 1728, Cur Loss: 0.12104675, Cur Avg Loss: 0.11532289, Log Avg loss: 0.13462298, Global Avg Loss: 0.55089210, Time: 0.0210 Steps: 102300, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000118, Sample Num: 1888, Cur Loss: 0.09342469, Cur Avg Loss: 0.11692314, Log Avg loss: 0.13420584, Global Avg Loss: 0.55085137, Time: 0.0209 Steps: 102310, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000128, Sample Num: 2048, Cur Loss: 0.03806785, Cur Avg Loss: 0.11530187, Log Avg loss: 0.09617083, Global Avg Loss: 0.55080693, Time: 0.0209 Steps: 102320, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000138, Sample Num: 2208, Cur Loss: 0.05914572, Cur Avg Loss: 0.11561416, Log Avg loss: 0.11961156, Global Avg Loss: 0.55076480, Time: 0.0209 Steps: 102330, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000148, Sample Num: 2368, Cur Loss: 0.19953364, Cur Avg Loss: 0.11653603, Log Avg loss: 0.12925775, Global Avg Loss: 0.55072361, Time: 0.0209 Steps: 102340, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000158, Sample Num: 2528, Cur Loss: 0.12217595, Cur Avg Loss: 0.11861641, Log Avg loss: 0.14940604, Global Avg Loss: 0.55068440, Time: 0.0210 Steps: 102350, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000168, Sample Num: 2688, Cur Loss: 0.03725950, Cur Avg Loss: 0.12251565, Log Avg loss: 0.18412368, Global Avg Loss: 0.55064859, Time: 0.0210 Steps: 102360, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000178, Sample Num: 2848, Cur Loss: 0.19208287, Cur Avg Loss: 0.12209061, Log Avg loss: 0.11495000, Global Avg Loss: 0.55060603, Time: 0.0210 Steps: 102370, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000188, Sample Num: 3008, Cur Loss: 0.33755457, Cur Avg Loss: 0.12655241, Log Avg loss: 0.20597236, Global Avg Loss: 0.55057236, Time: 0.0209 Steps: 102380, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000198, Sample Num: 3168, Cur Loss: 0.09572195, Cur Avg Loss: 0.12625741, Log Avg loss: 0.12071140, Global Avg Loss: 0.55053038, Time: 0.0210 Steps: 102390, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000208, Sample Num: 3328, Cur Loss: 0.19474182, Cur Avg Loss: 0.12495590, Log Avg loss: 0.09918598, Global Avg Loss: 0.55048631, Time: 0.0209 Steps: 102400, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000218, Sample Num: 3488, Cur Loss: 0.30393672, Cur Avg Loss: 0.12510520, Log Avg loss: 0.12821074, Global Avg Loss: 0.55044507, Time: 0.0210 Steps: 102410, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000228, Sample Num: 3648, Cur Loss: 0.21681914, Cur Avg Loss: 0.12422577, Log Avg loss: 0.10505416, Global Avg Loss: 0.55040158, Time: 0.0210 Steps: 102420, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000238, Sample Num: 3808, Cur Loss: 0.08726121, Cur Avg Loss: 0.12397773, Log Avg loss: 0.11832248, Global Avg Loss: 0.55035940, Time: 0.0209 Steps: 102430, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000248, Sample Num: 3968, Cur Loss: 0.12331432, Cur Avg Loss: 0.12402521, Log Avg loss: 0.12515511, Global Avg Loss: 0.55031789, Time: 0.0209 Steps: 102440, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000258, Sample Num: 4128, Cur Loss: 0.14443491, Cur Avg Loss: 0.12335314, Log Avg loss: 0.10668589, Global Avg Loss: 0.55027459, Time: 0.0247 Steps: 102450, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000268, Sample Num: 4288, Cur Loss: 0.33120674, Cur Avg Loss: 0.12385041, Log Avg loss: 0.13667982, Global Avg Loss: 0.55023423, Time: 0.0209 Steps: 102460, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000278, Sample Num: 4448, Cur Loss: 0.14349586, Cur Avg Loss: 0.12414235, Log Avg loss: 0.13196644, Global Avg Loss: 0.55019341, Time: 0.0208 Steps: 102470, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000288, Sample Num: 4608, Cur Loss: 0.13699740, Cur Avg Loss: 0.12328454, Log Avg loss: 0.09943729, Global Avg Loss: 0.55014942, Time: 0.0208 Steps: 102480, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000298, Sample Num: 4768, Cur Loss: 0.19068795, Cur Avg Loss: 0.12368671, Log Avg loss: 0.13526934, Global Avg Loss: 0.55010894, Time: 0.0209 Steps: 102490, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000308, Sample Num: 4928, Cur Loss: 0.22532697, Cur Avg Loss: 0.12429672, Log Avg loss: 0.14247496, Global Avg Loss: 0.55006917, Time: 0.0209 Steps: 102500, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000318, Sample Num: 5088, Cur Loss: 0.13078104, Cur Avg Loss: 0.12374454, Log Avg loss: 0.10673736, Global Avg Loss: 0.55002593, Time: 0.0208 Steps: 102510, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000328, Sample Num: 5248, Cur Loss: 0.22120094, Cur Avg Loss: 0.12509850, Log Avg loss: 0.16815441, Global Avg Loss: 0.54998868, Time: 0.0209 Steps: 102520, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000338, Sample Num: 5408, Cur Loss: 0.13883755, Cur Avg Loss: 0.12524750, Log Avg loss: 0.13013479, Global Avg Loss: 0.54994773, Time: 0.0208 Steps: 102530, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000348, Sample Num: 5568, Cur Loss: 0.08715021, Cur Avg Loss: 0.12545442, Log Avg loss: 0.13244837, Global Avg Loss: 0.54990701, Time: 0.0208 Steps: 102540, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000358, Sample Num: 5728, Cur Loss: 0.25637835, Cur Avg Loss: 0.12620223, Log Avg loss: 0.15222582, Global Avg Loss: 0.54986823, Time: 0.0211 Steps: 102550, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000368, Sample Num: 5888, Cur Loss: 0.02908771, Cur Avg Loss: 0.12584903, Log Avg loss: 0.11320470, Global Avg Loss: 0.54982566, Time: 0.0208 Steps: 102560, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000378, Sample Num: 6048, Cur Loss: 0.04061674, Cur Avg Loss: 0.12573612, Log Avg loss: 0.12158094, Global Avg Loss: 0.54978390, Time: 0.0208 Steps: 102570, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000388, Sample Num: 6208, Cur Loss: 0.13283053, Cur Avg Loss: 0.12570823, Log Avg loss: 0.12465396, Global Avg Loss: 0.54974246, Time: 0.0209 Steps: 102580, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000398, Sample Num: 6368, Cur Loss: 0.14865574, Cur Avg Loss: 0.12522381, Log Avg loss: 0.10642834, Global Avg Loss: 0.54969925, Time: 0.0209 Steps: 102590, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000408, Sample Num: 6528, Cur Loss: 0.29085675, Cur Avg Loss: 0.12573583, Log Avg loss: 0.14611419, Global Avg Loss: 0.54965991, Time: 0.0208 Steps: 102600, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000418, Sample Num: 6688, Cur Loss: 0.06943974, Cur Avg Loss: 0.12562631, Log Avg loss: 0.12115782, Global Avg Loss: 0.54961815, Time: 0.0209 Steps: 102610, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000428, Sample Num: 6848, Cur Loss: 0.21785069, Cur Avg Loss: 0.12597804, Log Avg loss: 0.14068057, Global Avg Loss: 0.54957830, Time: 0.0209 Steps: 102620, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000438, Sample Num: 7008, Cur Loss: 0.05027862, Cur Avg Loss: 0.12780140, Log Avg loss: 0.20584120, Global Avg Loss: 0.54954481, Time: 0.0209 Steps: 102630, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000448, Sample Num: 7168, Cur Loss: 0.11464809, Cur Avg Loss: 0.12726609, Log Avg loss: 0.10381962, Global Avg Loss: 0.54950138, Time: 0.0208 Steps: 102640, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000458, Sample Num: 7328, Cur Loss: 0.05221514, Cur Avg Loss: 0.12588330, Log Avg loss: 0.06393416, Global Avg Loss: 0.54945408, Time: 0.0209 Steps: 102650, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000468, Sample Num: 7488, Cur Loss: 0.13324767, Cur Avg Loss: 0.12610145, Log Avg loss: 0.13609274, Global Avg Loss: 0.54941382, Time: 0.0209 Steps: 102660, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000478, Sample Num: 7648, Cur Loss: 0.10568023, Cur Avg Loss: 0.12624436, Log Avg loss: 0.13293249, Global Avg Loss: 0.54937325, Time: 0.0208 Steps: 102670, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000488, Sample Num: 7808, Cur Loss: 0.11148108, Cur Avg Loss: 0.12600467, Log Avg loss: 0.11454748, Global Avg Loss: 0.54933090, Time: 0.0208 Steps: 102680, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000498, Sample Num: 7968, Cur Loss: 0.12918936, Cur Avg Loss: 0.12577144, Log Avg loss: 0.11438974, Global Avg Loss: 0.54928855, Time: 0.0209 Steps: 102690, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000508, Sample Num: 8128, Cur Loss: 0.31756565, Cur Avg Loss: 0.12532386, Log Avg loss: 0.10303454, Global Avg Loss: 0.54924510, Time: 0.0208 Steps: 102700, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000518, Sample Num: 8288, Cur Loss: 0.04700561, Cur Avg Loss: 0.12598246, Log Avg loss: 0.15943911, Global Avg Loss: 0.54920714, Time: 0.0210 Steps: 102710, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000528, Sample Num: 8448, Cur Loss: 0.02196933, Cur Avg Loss: 0.12651023, Log Avg loss: 0.15384884, Global Avg Loss: 0.54916865, Time: 0.0209 Steps: 102720, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000538, Sample Num: 8608, Cur Loss: 0.09533744, Cur Avg Loss: 0.12679845, Log Avg loss: 0.14201632, Global Avg Loss: 0.54912902, Time: 0.0209 Steps: 102730, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000548, Sample Num: 8768, Cur Loss: 0.09091982, Cur Avg Loss: 0.12619314, Log Avg loss: 0.09362746, Global Avg Loss: 0.54908469, Time: 0.0208 Steps: 102740, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000558, Sample Num: 8928, Cur Loss: 0.01419135, Cur Avg Loss: 0.12563623, Log Avg loss: 0.09511800, Global Avg Loss: 0.54904050, Time: 0.0209 Steps: 102750, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000568, Sample Num: 9088, Cur Loss: 0.05250761, Cur Avg Loss: 0.12523260, Log Avg loss: 0.10270961, Global Avg Loss: 0.54899707, Time: 0.0209 Steps: 102760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000578, Sample Num: 9248, Cur Loss: 0.21258865, Cur Avg Loss: 0.12444374, Log Avg loss: 0.07963672, Global Avg Loss: 0.54895140, Time: 0.0210 Steps: 102770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000588, Sample Num: 9408, Cur Loss: 0.34616652, Cur Avg Loss: 0.12533837, Log Avg loss: 0.17704782, Global Avg Loss: 0.54891521, Time: 0.0209 Steps: 102780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000598, Sample Num: 9568, Cur Loss: 0.04244469, Cur Avg Loss: 0.12513182, Log Avg loss: 0.11298707, Global Avg Loss: 0.54887281, Time: 0.0209 Steps: 102790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000608, Sample Num: 9728, Cur Loss: 0.14657353, Cur Avg Loss: 0.12502815, Log Avg loss: 0.11882860, Global Avg Loss: 0.54883097, Time: 0.0209 Steps: 102800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000618, Sample Num: 9888, Cur Loss: 0.10190596, Cur Avg Loss: 0.12431334, Log Avg loss: 0.08085272, Global Avg Loss: 0.54878545, Time: 0.0210 Steps: 102810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000628, Sample Num: 10048, Cur Loss: 0.06392652, Cur Avg Loss: 0.12415679, Log Avg loss: 0.11448171, Global Avg Loss: 0.54874321, Time: 0.0209 Steps: 102820, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000638, Sample Num: 10208, Cur Loss: 0.44429323, Cur Avg Loss: 0.12488857, Log Avg loss: 0.17084476, Global Avg Loss: 0.54870646, Time: 0.0210 Steps: 102830, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000648, Sample Num: 10368, Cur Loss: 0.08600594, Cur Avg Loss: 0.12571066, Log Avg loss: 0.17815978, Global Avg Loss: 0.54867043, Time: 0.0209 Steps: 102840, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000658, Sample Num: 10528, Cur Loss: 0.10769258, Cur Avg Loss: 0.12504033, Log Avg loss: 0.08160291, Global Avg Loss: 0.54862502, Time: 0.0208 Steps: 102850, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000668, Sample Num: 10688, Cur Loss: 0.06138454, Cur Avg Loss: 0.12444134, Log Avg loss: 0.08502808, Global Avg Loss: 0.54857995, Time: 0.0210 Steps: 102860, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000678, Sample Num: 10848, Cur Loss: 0.04983312, Cur Avg Loss: 0.12431119, Log Avg loss: 0.11561680, Global Avg Loss: 0.54853786, Time: 0.0209 Steps: 102870, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000688, Sample Num: 11008, Cur Loss: 0.24645798, Cur Avg Loss: 0.12466599, Log Avg loss: 0.14872181, Global Avg Loss: 0.54849900, Time: 0.0210 Steps: 102880, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000698, Sample Num: 11168, Cur Loss: 0.47796190, Cur Avg Loss: 0.12499820, Log Avg loss: 0.14785373, Global Avg Loss: 0.54846006, Time: 0.0209 Steps: 102890, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000708, Sample Num: 11328, Cur Loss: 0.08463603, Cur Avg Loss: 0.12575958, Log Avg loss: 0.17890394, Global Avg Loss: 0.54842415, Time: 0.0209 Steps: 102900, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000718, Sample Num: 11488, Cur Loss: 0.59169286, Cur Avg Loss: 0.12648974, Log Avg loss: 0.17818525, Global Avg Loss: 0.54838817, Time: 0.0209 Steps: 102910, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000728, Sample Num: 11648, Cur Loss: 0.05981990, Cur Avg Loss: 0.12597856, Log Avg loss: 0.08927629, Global Avg Loss: 0.54834356, Time: 0.0209 Steps: 102920, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000738, Sample Num: 11808, Cur Loss: 0.03728907, Cur Avg Loss: 0.12549574, Log Avg loss: 0.09034594, Global Avg Loss: 0.54829906, Time: 0.0209 Steps: 102930, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000748, Sample Num: 11968, Cur Loss: 0.05936331, Cur Avg Loss: 0.12547616, Log Avg loss: 0.12403157, Global Avg Loss: 0.54825785, Time: 0.0209 Steps: 102940, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000758, Sample Num: 12128, Cur Loss: 0.05571210, Cur Avg Loss: 0.12535387, Log Avg loss: 0.11620612, Global Avg Loss: 0.54821588, Time: 0.0210 Steps: 102950, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000768, Sample Num: 12288, Cur Loss: 0.14529030, Cur Avg Loss: 0.12513898, Log Avg loss: 0.10885034, Global Avg Loss: 0.54817321, Time: 0.0254 Steps: 102960, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000778, Sample Num: 12448, Cur Loss: 0.19188112, Cur Avg Loss: 0.12567915, Log Avg loss: 0.16716436, Global Avg Loss: 0.54813621, Time: 0.0209 Steps: 102970, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000788, Sample Num: 12608, Cur Loss: 0.19971828, Cur Avg Loss: 0.12577284, Log Avg loss: 0.13306195, Global Avg Loss: 0.54809590, Time: 0.0209 Steps: 102980, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000798, Sample Num: 12768, Cur Loss: 0.06585695, Cur Avg Loss: 0.12571034, Log Avg loss: 0.12078547, Global Avg Loss: 0.54805441, Time: 0.0209 Steps: 102990, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000808, Sample Num: 12928, Cur Loss: 0.07942408, Cur Avg Loss: 0.12568424, Log Avg loss: 0.12360155, Global Avg Loss: 0.54801320, Time: 0.0209 Steps: 103000, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000818, Sample Num: 13088, Cur Loss: 0.10625979, Cur Avg Loss: 0.12566918, Log Avg loss: 0.12445199, Global Avg Loss: 0.54797208, Time: 0.0209 Steps: 103010, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000828, Sample Num: 13248, Cur Loss: 0.20601101, Cur Avg Loss: 0.12649377, Log Avg loss: 0.19394556, Global Avg Loss: 0.54793772, Time: 0.0209 Steps: 103020, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000838, Sample Num: 13408, Cur Loss: 0.14303935, Cur Avg Loss: 0.12666263, Log Avg loss: 0.14064431, Global Avg Loss: 0.54789819, Time: 0.0208 Steps: 103030, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000848, Sample Num: 13568, Cur Loss: 0.17101565, Cur Avg Loss: 0.12691165, Log Avg loss: 0.14777885, Global Avg Loss: 0.54785935, Time: 0.0209 Steps: 103040, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000858, Sample Num: 13728, Cur Loss: 0.24053621, Cur Avg Loss: 0.12704484, Log Avg loss: 0.13833968, Global Avg Loss: 0.54781961, Time: 0.0209 Steps: 103050, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000868, Sample Num: 13888, Cur Loss: 0.06924838, Cur Avg Loss: 0.12732531, Log Avg loss: 0.15138930, Global Avg Loss: 0.54778115, Time: 0.0210 Steps: 103060, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000878, Sample Num: 14048, Cur Loss: 0.08364908, Cur Avg Loss: 0.12717151, Log Avg loss: 0.11382167, Global Avg Loss: 0.54773905, Time: 0.0209 Steps: 103070, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000888, Sample Num: 14208, Cur Loss: 0.14715883, Cur Avg Loss: 0.12777859, Log Avg loss: 0.18108085, Global Avg Loss: 0.54770348, Time: 0.0210 Steps: 103080, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000898, Sample Num: 14368, Cur Loss: 0.29078984, Cur Avg Loss: 0.12779219, Log Avg loss: 0.12899959, Global Avg Loss: 0.54766286, Time: 0.0210 Steps: 103090, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000908, Sample Num: 14528, Cur Loss: 0.16897224, Cur Avg Loss: 0.12785482, Log Avg loss: 0.13347934, Global Avg Loss: 0.54762269, Time: 0.0210 Steps: 103100, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000918, Sample Num: 14688, Cur Loss: 0.11194696, Cur Avg Loss: 0.12776436, Log Avg loss: 0.11955010, Global Avg Loss: 0.54758117, Time: 0.0210 Steps: 103110, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000928, Sample Num: 14848, Cur Loss: 0.07178569, Cur Avg Loss: 0.12745177, Log Avg loss: 0.09875602, Global Avg Loss: 0.54753765, Time: 0.0210 Steps: 103120, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000938, Sample Num: 15008, Cur Loss: 0.04995707, Cur Avg Loss: 0.12750112, Log Avg loss: 0.13208099, Global Avg Loss: 0.54749736, Time: 0.0210 Steps: 103130, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000948, Sample Num: 15168, Cur Loss: 0.33601567, Cur Avg Loss: 0.12781854, Log Avg loss: 0.15759233, Global Avg Loss: 0.54745956, Time: 0.0210 Steps: 103140, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000958, Sample Num: 15328, Cur Loss: 0.11387648, Cur Avg Loss: 0.12761071, Log Avg loss: 0.10790822, Global Avg Loss: 0.54741694, Time: 0.0210 Steps: 103150, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000968, Sample Num: 15488, Cur Loss: 0.03792986, Cur Avg Loss: 0.12733583, Log Avg loss: 0.10100258, Global Avg Loss: 0.54737367, Time: 0.0213 Steps: 103160, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000978, Sample Num: 15648, Cur Loss: 0.21500663, Cur Avg Loss: 0.12738610, Log Avg loss: 0.13225225, Global Avg Loss: 0.54733343, Time: 0.0213 Steps: 103170, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000988, Sample Num: 15808, Cur Loss: 0.04406036, Cur Avg Loss: 0.12762449, Log Avg loss: 0.15093907, Global Avg Loss: 0.54729502, Time: 0.0212 Steps: 103180, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000998, Sample Num: 15968, Cur Loss: 0.10552710, Cur Avg Loss: 0.12751137, Log Avg loss: 0.11633551, Global Avg Loss: 0.54725325, Time: 0.0210 Steps: 103190, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001008, Sample Num: 16128, Cur Loss: 0.05754745, Cur Avg Loss: 0.12684245, Log Avg loss: 0.06008378, Global Avg Loss: 0.54720605, Time: 0.0210 Steps: 103200, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001018, Sample Num: 16288, Cur Loss: 0.04750488, Cur Avg Loss: 0.12685216, Log Avg loss: 0.12783097, Global Avg Loss: 0.54716541, Time: 0.0210 Steps: 103210, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001028, Sample Num: 16448, Cur Loss: 0.33280534, Cur Avg Loss: 0.12722948, Log Avg loss: 0.16564071, Global Avg Loss: 0.54712845, Time: 0.0247 Steps: 103220, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001038, Sample Num: 16608, Cur Loss: 0.17109936, Cur Avg Loss: 0.12705277, Log Avg loss: 0.10888663, Global Avg Loss: 0.54708600, Time: 0.0210 Steps: 103230, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001048, Sample Num: 16768, Cur Loss: 0.05880930, Cur Avg Loss: 0.12706599, Log Avg loss: 0.12843811, Global Avg Loss: 0.54704545, Time: 0.0209 Steps: 103240, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001058, Sample Num: 16928, Cur Loss: 0.19743466, Cur Avg Loss: 0.12708609, Log Avg loss: 0.12919312, Global Avg Loss: 0.54700498, Time: 0.0210 Steps: 103250, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001068, Sample Num: 17088, Cur Loss: 0.04785689, Cur Avg Loss: 0.12713883, Log Avg loss: 0.13271877, Global Avg Loss: 0.54696486, Time: 0.0209 Steps: 103260, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001078, Sample Num: 17248, Cur Loss: 0.18884587, Cur Avg Loss: 0.12690756, Log Avg loss: 0.10220734, Global Avg Loss: 0.54692179, Time: 0.0210 Steps: 103270, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001088, Sample Num: 17408, Cur Loss: 0.11540359, Cur Avg Loss: 0.12663639, Log Avg loss: 0.09740456, Global Avg Loss: 0.54687827, Time: 0.0210 Steps: 103280, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001098, Sample Num: 17568, Cur Loss: 0.07044575, Cur Avg Loss: 0.12779561, Log Avg loss: 0.25391879, Global Avg Loss: 0.54684990, Time: 0.0210 Steps: 103290, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001108, Sample Num: 17728, Cur Loss: 0.10303658, Cur Avg Loss: 0.12771317, Log Avg loss: 0.11866164, Global Avg Loss: 0.54680845, Time: 0.0210 Steps: 103300, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001118, Sample Num: 17888, Cur Loss: 0.06675769, Cur Avg Loss: 0.12747951, Log Avg loss: 0.10158969, Global Avg Loss: 0.54676536, Time: 0.0209 Steps: 103310, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001128, Sample Num: 18048, Cur Loss: 0.15203987, Cur Avg Loss: 0.12739142, Log Avg loss: 0.11754355, Global Avg Loss: 0.54672381, Time: 0.0210 Steps: 103320, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001138, Sample Num: 18208, Cur Loss: 0.08131164, Cur Avg Loss: 0.12731694, Log Avg loss: 0.11891449, Global Avg Loss: 0.54668241, Time: 0.0210 Steps: 103330, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001148, Sample Num: 18368, Cur Loss: 0.06922887, Cur Avg Loss: 0.12737876, Log Avg loss: 0.13441494, Global Avg Loss: 0.54664252, Time: 0.0210 Steps: 103340, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001158, Sample Num: 18528, Cur Loss: 0.13064182, Cur Avg Loss: 0.12727400, Log Avg loss: 0.11524712, Global Avg Loss: 0.54660078, Time: 0.0210 Steps: 103350, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001168, Sample Num: 18688, Cur Loss: 0.13309731, Cur Avg Loss: 0.12707594, Log Avg loss: 0.10414094, Global Avg Loss: 0.54655797, Time: 0.0209 Steps: 103360, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001178, Sample Num: 18848, Cur Loss: 0.10446614, Cur Avg Loss: 0.12681624, Log Avg loss: 0.09648310, Global Avg Loss: 0.54651443, Time: 0.0210 Steps: 103370, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001188, Sample Num: 19008, Cur Loss: 0.04797300, Cur Avg Loss: 0.12658471, Log Avg loss: 0.09930966, Global Avg Loss: 0.54647117, Time: 0.0209 Steps: 103380, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001198, Sample Num: 19168, Cur Loss: 0.13397226, Cur Avg Loss: 0.12646684, Log Avg loss: 0.11246397, Global Avg Loss: 0.54642919, Time: 0.0209 Steps: 103390, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001208, Sample Num: 19328, Cur Loss: 0.21920288, Cur Avg Loss: 0.12687909, Log Avg loss: 0.17626709, Global Avg Loss: 0.54639339, Time: 0.0209 Steps: 103400, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001218, Sample Num: 19488, Cur Loss: 0.08148644, Cur Avg Loss: 0.12693772, Log Avg loss: 0.13402028, Global Avg Loss: 0.54635351, Time: 0.0209 Steps: 103410, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001228, Sample Num: 19648, Cur Loss: 0.15324308, Cur Avg Loss: 0.12710683, Log Avg loss: 0.14770452, Global Avg Loss: 0.54631497, Time: 0.0210 Steps: 103420, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001238, Sample Num: 19808, Cur Loss: 0.11929527, Cur Avg Loss: 0.12739956, Log Avg loss: 0.16334696, Global Avg Loss: 0.54627794, Time: 0.0209 Steps: 103430, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001248, Sample Num: 19968, Cur Loss: 0.28587991, Cur Avg Loss: 0.12749394, Log Avg loss: 0.13917753, Global Avg Loss: 0.54623859, Time: 0.0209 Steps: 103440, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001258, Sample Num: 20128, Cur Loss: 0.24269283, Cur Avg Loss: 0.12799064, Log Avg loss: 0.18997973, Global Avg Loss: 0.54620415, Time: 0.0210 Steps: 103450, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001268, Sample Num: 20288, Cur Loss: 0.06194362, Cur Avg Loss: 0.12806798, Log Avg loss: 0.13779692, Global Avg Loss: 0.54616467, Time: 0.0210 Steps: 103460, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001278, Sample Num: 20448, Cur Loss: 0.48548296, Cur Avg Loss: 0.12808908, Log Avg loss: 0.13076462, Global Avg Loss: 0.54612453, Time: 0.0209 Steps: 103470, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001288, Sample Num: 20608, Cur Loss: 0.12296104, Cur Avg Loss: 0.12814180, Log Avg loss: 0.13487885, Global Avg Loss: 0.54608478, Time: 0.0208 Steps: 103480, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001298, Sample Num: 20768, Cur Loss: 0.04722725, Cur Avg Loss: 0.12802328, Log Avg loss: 0.11275857, Global Avg Loss: 0.54604291, Time: 0.0210 Steps: 103490, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001308, Sample Num: 20928, Cur Loss: 0.09886611, Cur Avg Loss: 0.12783411, Log Avg loss: 0.10327970, Global Avg Loss: 0.54600013, Time: 0.0208 Steps: 103500, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001318, Sample Num: 21088, Cur Loss: 0.09368987, Cur Avg Loss: 0.12741255, Log Avg loss: 0.07227182, Global Avg Loss: 0.54595437, Time: 0.0208 Steps: 103510, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001328, Sample Num: 21248, Cur Loss: 0.09484532, Cur Avg Loss: 0.12730429, Log Avg loss: 0.11303562, Global Avg Loss: 0.54591255, Time: 0.0209 Steps: 103520, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001338, Sample Num: 21408, Cur Loss: 0.03873802, Cur Avg Loss: 0.12756657, Log Avg loss: 0.16239745, Global Avg Loss: 0.54587550, Time: 0.0208 Steps: 103530, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001348, Sample Num: 21568, Cur Loss: 0.03451496, Cur Avg Loss: 0.12762754, Log Avg loss: 0.13578520, Global Avg Loss: 0.54583590, Time: 0.0208 Steps: 103540, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001358, Sample Num: 21728, Cur Loss: 0.08980912, Cur Avg Loss: 0.12763130, Log Avg loss: 0.12813834, Global Avg Loss: 0.54579556, Time: 0.0210 Steps: 103550, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001368, Sample Num: 21888, Cur Loss: 0.11955943, Cur Avg Loss: 0.12738657, Log Avg loss: 0.09415309, Global Avg Loss: 0.54575195, Time: 0.0210 Steps: 103560, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001378, Sample Num: 22048, Cur Loss: 0.08701551, Cur Avg Loss: 0.12764192, Log Avg loss: 0.16257333, Global Avg Loss: 0.54571495, Time: 0.0210 Steps: 103570, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001388, Sample Num: 22208, Cur Loss: 0.09001830, Cur Avg Loss: 0.12732731, Log Avg loss: 0.08397420, Global Avg Loss: 0.54567037, Time: 0.0208 Steps: 103580, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001398, Sample Num: 22368, Cur Loss: 0.25919151, Cur Avg Loss: 0.12742407, Log Avg loss: 0.14085471, Global Avg Loss: 0.54563129, Time: 0.0208 Steps: 103590, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001408, Sample Num: 22528, Cur Loss: 0.01424226, Cur Avg Loss: 0.12762919, Log Avg loss: 0.15630410, Global Avg Loss: 0.54559371, Time: 0.0208 Steps: 103600, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001418, Sample Num: 22688, Cur Loss: 0.13080625, Cur Avg Loss: 0.12784910, Log Avg loss: 0.15881317, Global Avg Loss: 0.54555638, Time: 0.0207 Steps: 103610, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001428, Sample Num: 22848, Cur Loss: 0.42860413, Cur Avg Loss: 0.12812356, Log Avg loss: 0.16704105, Global Avg Loss: 0.54551985, Time: 0.0208 Steps: 103620, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001438, Sample Num: 23008, Cur Loss: 0.06808434, Cur Avg Loss: 0.12840664, Log Avg loss: 0.16883109, Global Avg Loss: 0.54548350, Time: 0.0208 Steps: 103630, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001448, Sample Num: 23168, Cur Loss: 0.09817746, Cur Avg Loss: 0.12842322, Log Avg loss: 0.13080805, Global Avg Loss: 0.54544349, Time: 0.0209 Steps: 103640, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001458, Sample Num: 23328, Cur Loss: 0.02923092, Cur Avg Loss: 0.12817456, Log Avg loss: 0.09216780, Global Avg Loss: 0.54539976, Time: 0.0208 Steps: 103650, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001468, Sample Num: 23488, Cur Loss: 0.09335302, Cur Avg Loss: 0.12816247, Log Avg loss: 0.12639994, Global Avg Loss: 0.54535934, Time: 0.0207 Steps: 103660, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001478, Sample Num: 23648, Cur Loss: 0.06242010, Cur Avg Loss: 0.12854707, Log Avg loss: 0.18500555, Global Avg Loss: 0.54532458, Time: 0.0208 Steps: 103670, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001488, Sample Num: 23808, Cur Loss: 0.07669770, Cur Avg Loss: 0.12836014, Log Avg loss: 0.10073297, Global Avg Loss: 0.54528170, Time: 0.0208 Steps: 103680, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001498, Sample Num: 23968, Cur Loss: 0.40235654, Cur Avg Loss: 0.12850374, Log Avg loss: 0.14987171, Global Avg Loss: 0.54524357, Time: 0.0208 Steps: 103690, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001508, Sample Num: 24128, Cur Loss: 0.03385101, Cur Avg Loss: 0.12868797, Log Avg loss: 0.15628546, Global Avg Loss: 0.54520606, Time: 0.0210 Steps: 103700, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001518, Sample Num: 24288, Cur Loss: 0.18085870, Cur Avg Loss: 0.12882149, Log Avg loss: 0.14895640, Global Avg Loss: 0.54516785, Time: 0.0208 Steps: 103710, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001528, Sample Num: 24448, Cur Loss: 0.03695951, Cur Avg Loss: 0.12857982, Log Avg loss: 0.09189432, Global Avg Loss: 0.54512415, Time: 0.0208 Steps: 103720, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001538, Sample Num: 24608, Cur Loss: 0.09709993, Cur Avg Loss: 0.12857256, Log Avg loss: 0.12746218, Global Avg Loss: 0.54508389, Time: 0.0248 Steps: 103730, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001548, Sample Num: 24768, Cur Loss: 0.20667957, Cur Avg Loss: 0.12861535, Log Avg loss: 0.13519688, Global Avg Loss: 0.54504437, Time: 0.0210 Steps: 103740, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001558, Sample Num: 24928, Cur Loss: 0.07858210, Cur Avg Loss: 0.12840799, Log Avg loss: 0.09630913, Global Avg Loss: 0.54500112, Time: 0.0210 Steps: 103750, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001568, Sample Num: 25088, Cur Loss: 0.17409244, Cur Avg Loss: 0.12834457, Log Avg loss: 0.11846378, Global Avg Loss: 0.54496001, Time: 0.0209 Steps: 103760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001578, Sample Num: 25248, Cur Loss: 0.11990994, Cur Avg Loss: 0.12809559, Log Avg loss: 0.08905473, Global Avg Loss: 0.54491608, Time: 0.0209 Steps: 103770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001588, Sample Num: 25408, Cur Loss: 0.08802234, Cur Avg Loss: 0.12834597, Log Avg loss: 0.16785662, Global Avg Loss: 0.54487975, Time: 0.0209 Steps: 103780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001598, Sample Num: 25568, Cur Loss: 0.13927542, Cur Avg Loss: 0.12806486, Log Avg loss: 0.08342383, Global Avg Loss: 0.54483529, Time: 0.0209 Steps: 103790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001608, Sample Num: 25728, Cur Loss: 0.09266923, Cur Avg Loss: 0.12826134, Log Avg loss: 0.15965904, Global Avg Loss: 0.54479818, Time: 0.0209 Steps: 103800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001618, Sample Num: 25888, Cur Loss: 0.08645896, Cur Avg Loss: 0.12812356, Log Avg loss: 0.10596948, Global Avg Loss: 0.54475591, Time: 0.0209 Steps: 103810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001628, Sample Num: 26048, Cur Loss: 0.03798467, Cur Avg Loss: 0.12849620, Log Avg loss: 0.18878920, Global Avg Loss: 0.54472162, Time: 0.0209 Steps: 103820, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001638, Sample Num: 26208, Cur Loss: 0.28814486, Cur Avg Loss: 0.12852036, Log Avg loss: 0.13245364, Global Avg Loss: 0.54468191, Time: 0.0209 Steps: 103830, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001648, Sample Num: 26368, Cur Loss: 0.08292936, Cur Avg Loss: 0.12836537, Log Avg loss: 0.10297766, Global Avg Loss: 0.54463938, Time: 0.0209 Steps: 103840, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001658, Sample Num: 26528, Cur Loss: 0.14648633, Cur Avg Loss: 0.12865869, Log Avg loss: 0.17699750, Global Avg Loss: 0.54460398, Time: 0.0209 Steps: 103850, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001668, Sample Num: 26688, Cur Loss: 0.05136975, Cur Avg Loss: 0.12852971, Log Avg loss: 0.10714448, Global Avg Loss: 0.54456186, Time: 0.0209 Steps: 103860, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001678, Sample Num: 26848, Cur Loss: 0.14754851, Cur Avg Loss: 0.12861839, Log Avg loss: 0.14341045, Global Avg Loss: 0.54452324, Time: 0.0209 Steps: 103870, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001688, Sample Num: 27008, Cur Loss: 0.14110850, Cur Avg Loss: 0.12847478, Log Avg loss: 0.10437711, Global Avg Loss: 0.54448086, Time: 0.0209 Steps: 103880, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001698, Sample Num: 27168, Cur Loss: 0.04472899, Cur Avg Loss: 0.12843599, Log Avg loss: 0.12188793, Global Avg Loss: 0.54444019, Time: 0.0209 Steps: 103890, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001708, Sample Num: 27328, Cur Loss: 0.20294583, Cur Avg Loss: 0.12854277, Log Avg loss: 0.14667465, Global Avg Loss: 0.54440190, Time: 0.0209 Steps: 103900, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001718, Sample Num: 27488, Cur Loss: 0.16949698, Cur Avg Loss: 0.12856396, Log Avg loss: 0.13218224, Global Avg Loss: 0.54436223, Time: 0.0209 Steps: 103910, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001728, Sample Num: 27648, Cur Loss: 0.33419174, Cur Avg Loss: 0.12882974, Log Avg loss: 0.17449105, Global Avg Loss: 0.54432664, Time: 0.0209 Steps: 103920, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001738, Sample Num: 27808, Cur Loss: 0.21685520, Cur Avg Loss: 0.12910852, Log Avg loss: 0.17728145, Global Avg Loss: 0.54429133, Time: 0.0209 Steps: 103930, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001748, Sample Num: 27968, Cur Loss: 0.21735662, Cur Avg Loss: 0.12918575, Log Avg loss: 0.14260986, Global Avg Loss: 0.54425268, Time: 0.0209 Steps: 103940, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001758, Sample Num: 28128, Cur Loss: 0.13349777, Cur Avg Loss: 0.12919036, Log Avg loss: 0.12999561, Global Avg Loss: 0.54421283, Time: 0.0209 Steps: 103950, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001768, Sample Num: 28288, Cur Loss: 0.10892353, Cur Avg Loss: 0.12944399, Log Avg loss: 0.17403143, Global Avg Loss: 0.54417722, Time: 0.0209 Steps: 103960, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001778, Sample Num: 28448, Cur Loss: 0.03879536, Cur Avg Loss: 0.12940869, Log Avg loss: 0.12316795, Global Avg Loss: 0.54413673, Time: 0.0209 Steps: 103970, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001788, Sample Num: 28608, Cur Loss: 0.16438186, Cur Avg Loss: 0.12915436, Log Avg loss: 0.08393382, Global Avg Loss: 0.54409247, Time: 0.0209 Steps: 103980, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001798, Sample Num: 28768, Cur Loss: 0.11577123, Cur Avg Loss: 0.12935212, Log Avg loss: 0.16471254, Global Avg Loss: 0.54405599, Time: 0.0241 Steps: 103990, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001808, Sample Num: 28928, Cur Loss: 0.10468330, Cur Avg Loss: 0.12948985, Log Avg loss: 0.15425375, Global Avg Loss: 0.54401850, Time: 0.0209 Steps: 104000, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001818, Sample Num: 29088, Cur Loss: 0.04316563, Cur Avg Loss: 0.12945016, Log Avg loss: 0.12227432, Global Avg Loss: 0.54397796, Time: 0.0209 Steps: 104010, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001828, Sample Num: 29248, Cur Loss: 0.17373726, Cur Avg Loss: 0.12959612, Log Avg loss: 0.15613191, Global Avg Loss: 0.54394067, Time: 0.0209 Steps: 104020, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001838, Sample Num: 29408, Cur Loss: 0.20455423, Cur Avg Loss: 0.12996478, Log Avg loss: 0.19735537, Global Avg Loss: 0.54390735, Time: 0.0209 Steps: 104030, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001848, Sample Num: 29568, Cur Loss: 0.08590753, Cur Avg Loss: 0.12987629, Log Avg loss: 0.11361249, Global Avg Loss: 0.54386600, Time: 0.0210 Steps: 104040, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001858, Sample Num: 29728, Cur Loss: 0.09292141, Cur Avg Loss: 0.12996912, Log Avg loss: 0.14712299, Global Avg Loss: 0.54382787, Time: 0.0209 Steps: 104050, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001868, Sample Num: 29888, Cur Loss: 0.13384867, Cur Avg Loss: 0.13011872, Log Avg loss: 0.15791543, Global Avg Loss: 0.54379078, Time: 0.0209 Steps: 104060, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001878, Sample Num: 30048, Cur Loss: 0.28120601, Cur Avg Loss: 0.13018366, Log Avg loss: 0.14231430, Global Avg Loss: 0.54375220, Time: 0.0210 Steps: 104070, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001888, Sample Num: 30208, Cur Loss: 0.43343490, Cur Avg Loss: 0.13008422, Log Avg loss: 0.11140810, Global Avg Loss: 0.54371066, Time: 0.0209 Steps: 104080, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001898, Sample Num: 30368, Cur Loss: 0.09241274, Cur Avg Loss: 0.13005054, Log Avg loss: 0.12369230, Global Avg Loss: 0.54367031, Time: 0.0209 Steps: 104090, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001908, Sample Num: 30528, Cur Loss: 0.10822231, Cur Avg Loss: 0.12991484, Log Avg loss: 0.10415825, Global Avg Loss: 0.54362809, Time: 0.0212 Steps: 104100, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001918, Sample Num: 30688, Cur Loss: 0.16551691, Cur Avg Loss: 0.12985457, Log Avg loss: 0.11835660, Global Avg Loss: 0.54358724, Time: 0.0209 Steps: 104110, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001928, Sample Num: 30848, Cur Loss: 0.02907939, Cur Avg Loss: 0.12976169, Log Avg loss: 0.11194575, Global Avg Loss: 0.54354579, Time: 0.0209 Steps: 104120, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001938, Sample Num: 31008, Cur Loss: 0.06527118, Cur Avg Loss: 0.12987852, Log Avg loss: 0.15240465, Global Avg Loss: 0.54350822, Time: 0.0209 Steps: 104130, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001948, Sample Num: 31168, Cur Loss: 0.14802420, Cur Avg Loss: 0.12966125, Log Avg loss: 0.08755429, Global Avg Loss: 0.54346444, Time: 0.0208 Steps: 104140, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001958, Sample Num: 31328, Cur Loss: 0.24338946, Cur Avg Loss: 0.12983458, Log Avg loss: 0.16359918, Global Avg Loss: 0.54342797, Time: 0.0209 Steps: 104150, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001968, Sample Num: 31488, Cur Loss: 0.32426438, Cur Avg Loss: 0.13000269, Log Avg loss: 0.16291917, Global Avg Loss: 0.54339144, Time: 0.0208 Steps: 104160, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001978, Sample Num: 31648, Cur Loss: 0.16133675, Cur Avg Loss: 0.12996589, Log Avg loss: 0.12272248, Global Avg Loss: 0.54335105, Time: 0.0208 Steps: 104170, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001988, Sample Num: 31808, Cur Loss: 0.19784380, Cur Avg Loss: 0.12994787, Log Avg loss: 0.12638316, Global Avg Loss: 0.54331103, Time: 0.0208 Steps: 104180, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001998, Sample Num: 31968, Cur Loss: 0.25819764, Cur Avg Loss: 0.12994488, Log Avg loss: 0.12935192, Global Avg Loss: 0.54327130, Time: 0.0208 Steps: 104190, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002008, Sample Num: 32128, Cur Loss: 0.20090526, Cur Avg Loss: 0.13029026, Log Avg loss: 0.19929700, Global Avg Loss: 0.54323829, Time: 0.0208 Steps: 104200, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002018, Sample Num: 32288, Cur Loss: 0.16116527, Cur Avg Loss: 0.13055609, Log Avg loss: 0.18393354, Global Avg Loss: 0.54320381, Time: 0.0208 Steps: 104210, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002028, Sample Num: 32448, Cur Loss: 0.32426330, Cur Avg Loss: 0.13062099, Log Avg loss: 0.14371897, Global Avg Loss: 0.54316548, Time: 0.0208 Steps: 104220, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002038, Sample Num: 32608, Cur Loss: 0.09438185, Cur Avg Loss: 0.13067630, Log Avg loss: 0.14189224, Global Avg Loss: 0.54312698, Time: 0.0208 Steps: 104230, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002048, Sample Num: 32768, Cur Loss: 0.35333392, Cur Avg Loss: 0.13088356, Log Avg loss: 0.17312436, Global Avg Loss: 0.54309148, Time: 0.0253 Steps: 104240, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002058, Sample Num: 32928, Cur Loss: 0.18257450, Cur Avg Loss: 0.13094091, Log Avg loss: 0.14268469, Global Avg Loss: 0.54305308, Time: 0.0210 Steps: 104250, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002068, Sample Num: 33088, Cur Loss: 0.08434249, Cur Avg Loss: 0.13096305, Log Avg loss: 0.13552053, Global Avg Loss: 0.54301399, Time: 0.0209 Steps: 104260, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002078, Sample Num: 33248, Cur Loss: 0.02801859, Cur Avg Loss: 0.13081874, Log Avg loss: 0.10097550, Global Avg Loss: 0.54297159, Time: 0.0209 Steps: 104270, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002088, Sample Num: 33408, Cur Loss: 0.25796273, Cur Avg Loss: 0.13071093, Log Avg loss: 0.10830718, Global Avg Loss: 0.54292991, Time: 0.0209 Steps: 104280, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002098, Sample Num: 33568, Cur Loss: 0.04723825, Cur Avg Loss: 0.13053582, Log Avg loss: 0.09397253, Global Avg Loss: 0.54288686, Time: 0.0209 Steps: 104290, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002108, Sample Num: 33728, Cur Loss: 0.21918714, Cur Avg Loss: 0.13049160, Log Avg loss: 0.12121504, Global Avg Loss: 0.54284643, Time: 0.0209 Steps: 104300, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002118, Sample Num: 33888, Cur Loss: 0.17969488, Cur Avg Loss: 0.13044476, Log Avg loss: 0.12057035, Global Avg Loss: 0.54280595, Time: 0.0209 Steps: 104310, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002128, Sample Num: 34048, Cur Loss: 0.16673639, Cur Avg Loss: 0.13051630, Log Avg loss: 0.14566844, Global Avg Loss: 0.54276788, Time: 0.0211 Steps: 104320, Updated lr: 0.000002 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.828596, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.249419, "eval_total_loss": 175.341598, "eval_mae": 0.330862, "eval_mse": 0.249516, "eval_r2": 0.841391, "eval_sp_statistic": 0.93284, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.937126, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.174699, "test_total_loss": 87.698664, "test_mae": 0.294155, "test_mse": 0.174762, "test_r2": 0.887207, "test_sp_statistic": 0.921977, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.94748, "test_ps_pvalue": 0.0, "lr": 2.0189663347558085e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5427634668661746, "train_cur_epoch_loss": 277.8208076301962, "train_cur_epoch_avg_loss": 0.1304935686379503, "train_cur_epoch_time": 44.828595876693726, "train_cur_epoch_avg_time": 0.02105617467200269, "epoch": 49, "step": 104321} ################################################## Training, Epoch: 0050, Batch: 000009, Sample Num: 144, Cur Loss: 0.05744734, Cur Avg Loss: 0.07756729, Log Avg loss: 0.07802325, Global Avg Loss: 0.54272334, Time: 0.0210 Steps: 104330, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000019, Sample Num: 304, Cur Loss: 0.12005423, Cur Avg Loss: 0.09355448, Log Avg loss: 0.10794294, Global Avg Loss: 0.54268167, Time: 0.0209 Steps: 104340, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000029, Sample Num: 464, Cur Loss: 0.09531712, Cur Avg Loss: 0.09073270, Log Avg loss: 0.08537133, Global Avg Loss: 0.54263784, Time: 0.0209 Steps: 104350, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000039, Sample Num: 624, Cur Loss: 0.04917262, Cur Avg Loss: 0.08844797, Log Avg loss: 0.08182226, Global Avg Loss: 0.54259369, Time: 0.0209 Steps: 104360, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000049, Sample Num: 784, Cur Loss: 0.43916073, Cur Avg Loss: 0.10046492, Log Avg loss: 0.14733099, Global Avg Loss: 0.54255581, Time: 0.0209 Steps: 104370, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000059, Sample Num: 944, Cur Loss: 0.46695411, Cur Avg Loss: 0.11182066, Log Avg loss: 0.16746381, Global Avg Loss: 0.54251988, Time: 0.0209 Steps: 104380, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000069, Sample Num: 1104, Cur Loss: 0.05715051, Cur Avg Loss: 0.11000445, Log Avg loss: 0.09928879, Global Avg Loss: 0.54247742, Time: 0.0209 Steps: 104390, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000079, Sample Num: 1264, Cur Loss: 0.07452676, Cur Avg Loss: 0.11065649, Log Avg loss: 0.11515560, Global Avg Loss: 0.54243649, Time: 0.0210 Steps: 104400, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000089, Sample Num: 1424, Cur Loss: 0.10810970, Cur Avg Loss: 0.11171425, Log Avg loss: 0.12007056, Global Avg Loss: 0.54239604, Time: 0.0210 Steps: 104410, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000099, Sample Num: 1584, Cur Loss: 0.08848447, Cur Avg Loss: 0.10968175, Log Avg loss: 0.09159246, Global Avg Loss: 0.54235286, Time: 0.0209 Steps: 104420, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000109, Sample Num: 1744, Cur Loss: 0.02529038, Cur Avg Loss: 0.11320336, Log Avg loss: 0.14806726, Global Avg Loss: 0.54231511, Time: 0.0209 Steps: 104430, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000119, Sample Num: 1904, Cur Loss: 0.08302332, Cur Avg Loss: 0.11567832, Log Avg loss: 0.14265548, Global Avg Loss: 0.54227684, Time: 0.0210 Steps: 104440, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000129, Sample Num: 2064, Cur Loss: 0.04631409, Cur Avg Loss: 0.11721462, Log Avg loss: 0.13549660, Global Avg Loss: 0.54223790, Time: 0.0209 Steps: 104450, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000139, Sample Num: 2224, Cur Loss: 0.27786061, Cur Avg Loss: 0.11984049, Log Avg loss: 0.15371415, Global Avg Loss: 0.54220070, Time: 0.0209 Steps: 104460, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000149, Sample Num: 2384, Cur Loss: 0.17364171, Cur Avg Loss: 0.12160220, Log Avg loss: 0.14608994, Global Avg Loss: 0.54216279, Time: 0.0209 Steps: 104470, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000159, Sample Num: 2544, Cur Loss: 0.02616716, Cur Avg Loss: 0.11807193, Log Avg loss: 0.06547095, Global Avg Loss: 0.54211716, Time: 0.0209 Steps: 104480, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000169, Sample Num: 2704, Cur Loss: 0.25016761, Cur Avg Loss: 0.11985069, Log Avg loss: 0.14813295, Global Avg Loss: 0.54207946, Time: 0.0209 Steps: 104490, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000179, Sample Num: 2864, Cur Loss: 0.05986875, Cur Avg Loss: 0.11773573, Log Avg loss: 0.08199295, Global Avg Loss: 0.54203543, Time: 0.0211 Steps: 104500, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000189, Sample Num: 3024, Cur Loss: 0.12404844, Cur Avg Loss: 0.11818674, Log Avg loss: 0.12625983, Global Avg Loss: 0.54199565, Time: 0.0211 Steps: 104510, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000199, Sample Num: 3184, Cur Loss: 0.05194284, Cur Avg Loss: 0.11815284, Log Avg loss: 0.11751217, Global Avg Loss: 0.54195503, Time: 0.0210 Steps: 104520, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000209, Sample Num: 3344, Cur Loss: 0.10324705, Cur Avg Loss: 0.11771655, Log Avg loss: 0.10903440, Global Avg Loss: 0.54191362, Time: 0.0211 Steps: 104530, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000219, Sample Num: 3504, Cur Loss: 0.04369171, Cur Avg Loss: 0.11988889, Log Avg loss: 0.16529076, Global Avg Loss: 0.54187759, Time: 0.0211 Steps: 104540, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000229, Sample Num: 3664, Cur Loss: 0.05685779, Cur Avg Loss: 0.12047935, Log Avg loss: 0.13341031, Global Avg Loss: 0.54183852, Time: 0.0211 Steps: 104550, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000239, Sample Num: 3824, Cur Loss: 0.03310244, Cur Avg Loss: 0.12090125, Log Avg loss: 0.13056279, Global Avg Loss: 0.54179919, Time: 0.0211 Steps: 104560, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000249, Sample Num: 3984, Cur Loss: 0.03270022, Cur Avg Loss: 0.12364164, Log Avg loss: 0.18913696, Global Avg Loss: 0.54176546, Time: 0.0211 Steps: 104570, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000259, Sample Num: 4144, Cur Loss: 0.05181361, Cur Avg Loss: 0.12294841, Log Avg loss: 0.10568707, Global Avg Loss: 0.54172376, Time: 0.0247 Steps: 104580, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000269, Sample Num: 4304, Cur Loss: 0.06355283, Cur Avg Loss: 0.12341182, Log Avg loss: 0.13541398, Global Avg Loss: 0.54168492, Time: 0.0210 Steps: 104590, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000279, Sample Num: 4464, Cur Loss: 0.08389843, Cur Avg Loss: 0.12382548, Log Avg loss: 0.13495313, Global Avg Loss: 0.54164603, Time: 0.0210 Steps: 104600, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000289, Sample Num: 4624, Cur Loss: 0.12671903, Cur Avg Loss: 0.12288717, Log Avg loss: 0.09670814, Global Avg Loss: 0.54160350, Time: 0.0210 Steps: 104610, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000299, Sample Num: 4784, Cur Loss: 0.21189037, Cur Avg Loss: 0.12357653, Log Avg loss: 0.14349908, Global Avg Loss: 0.54156545, Time: 0.0210 Steps: 104620, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000309, Sample Num: 4944, Cur Loss: 0.07791446, Cur Avg Loss: 0.12327478, Log Avg loss: 0.11425248, Global Avg Loss: 0.54152461, Time: 0.0211 Steps: 104630, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000319, Sample Num: 5104, Cur Loss: 0.06629134, Cur Avg Loss: 0.12281492, Log Avg loss: 0.10860537, Global Avg Loss: 0.54148323, Time: 0.0211 Steps: 104640, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000329, Sample Num: 5264, Cur Loss: 0.08738835, Cur Avg Loss: 0.12191768, Log Avg loss: 0.09329553, Global Avg Loss: 0.54144041, Time: 0.0212 Steps: 104650, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000339, Sample Num: 5424, Cur Loss: 0.09884560, Cur Avg Loss: 0.12318253, Log Avg loss: 0.16479626, Global Avg Loss: 0.54140442, Time: 0.0211 Steps: 104660, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000349, Sample Num: 5584, Cur Loss: 0.27811247, Cur Avg Loss: 0.12646735, Log Avg loss: 0.23782261, Global Avg Loss: 0.54137542, Time: 0.0210 Steps: 104670, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000359, Sample Num: 5744, Cur Loss: 0.13273774, Cur Avg Loss: 0.12670612, Log Avg loss: 0.13503912, Global Avg Loss: 0.54133660, Time: 0.0211 Steps: 104680, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000369, Sample Num: 5904, Cur Loss: 0.03664129, Cur Avg Loss: 0.12618630, Log Avg loss: 0.10752496, Global Avg Loss: 0.54129516, Time: 0.0211 Steps: 104690, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000379, Sample Num: 6064, Cur Loss: 0.20927641, Cur Avg Loss: 0.12607959, Log Avg loss: 0.12214176, Global Avg Loss: 0.54125513, Time: 0.0211 Steps: 104700, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000389, Sample Num: 6224, Cur Loss: 0.14114250, Cur Avg Loss: 0.12575008, Log Avg loss: 0.11326195, Global Avg Loss: 0.54121425, Time: 0.0212 Steps: 104710, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000399, Sample Num: 6384, Cur Loss: 0.12359366, Cur Avg Loss: 0.12502924, Log Avg loss: 0.09698838, Global Avg Loss: 0.54117183, Time: 0.0210 Steps: 104720, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000409, Sample Num: 6544, Cur Loss: 0.16019425, Cur Avg Loss: 0.12587610, Log Avg loss: 0.15966571, Global Avg Loss: 0.54113540, Time: 0.0210 Steps: 104730, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000419, Sample Num: 6704, Cur Loss: 0.12224588, Cur Avg Loss: 0.12563189, Log Avg loss: 0.11564381, Global Avg Loss: 0.54109478, Time: 0.0211 Steps: 104740, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000429, Sample Num: 6864, Cur Loss: 0.21650480, Cur Avg Loss: 0.12549935, Log Avg loss: 0.11994607, Global Avg Loss: 0.54105458, Time: 0.0211 Steps: 104750, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000439, Sample Num: 7024, Cur Loss: 0.14975128, Cur Avg Loss: 0.12554898, Log Avg loss: 0.12767809, Global Avg Loss: 0.54101512, Time: 0.0212 Steps: 104760, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000449, Sample Num: 7184, Cur Loss: 0.10202295, Cur Avg Loss: 0.12615700, Log Avg loss: 0.15284918, Global Avg Loss: 0.54097807, Time: 0.0211 Steps: 104770, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000459, Sample Num: 7344, Cur Loss: 0.03408808, Cur Avg Loss: 0.12690676, Log Avg loss: 0.16057096, Global Avg Loss: 0.54094176, Time: 0.0211 Steps: 104780, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000469, Sample Num: 7504, Cur Loss: 0.08621076, Cur Avg Loss: 0.12593414, Log Avg loss: 0.08129086, Global Avg Loss: 0.54089790, Time: 0.0211 Steps: 104790, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000479, Sample Num: 7664, Cur Loss: 0.05172816, Cur Avg Loss: 0.12551541, Log Avg loss: 0.10587680, Global Avg Loss: 0.54085639, Time: 0.0210 Steps: 104800, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000489, Sample Num: 7824, Cur Loss: 0.11375080, Cur Avg Loss: 0.12463421, Log Avg loss: 0.08242487, Global Avg Loss: 0.54081265, Time: 0.0210 Steps: 104810, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000499, Sample Num: 7984, Cur Loss: 0.08575492, Cur Avg Loss: 0.12482682, Log Avg loss: 0.13424541, Global Avg Loss: 0.54077386, Time: 0.0212 Steps: 104820, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000509, Sample Num: 8144, Cur Loss: 0.17509907, Cur Avg Loss: 0.12416080, Log Avg loss: 0.09092647, Global Avg Loss: 0.54073095, Time: 0.0209 Steps: 104830, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000519, Sample Num: 8304, Cur Loss: 0.17491955, Cur Avg Loss: 0.12354106, Log Avg loss: 0.09199632, Global Avg Loss: 0.54068815, Time: 0.0211 Steps: 104840, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000529, Sample Num: 8464, Cur Loss: 0.03248581, Cur Avg Loss: 0.12338237, Log Avg loss: 0.11514648, Global Avg Loss: 0.54064756, Time: 0.0210 Steps: 104850, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000539, Sample Num: 8624, Cur Loss: 0.04211513, Cur Avg Loss: 0.12372372, Log Avg loss: 0.14178069, Global Avg Loss: 0.54060952, Time: 0.0211 Steps: 104860, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000549, Sample Num: 8784, Cur Loss: 0.06126028, Cur Avg Loss: 0.12349881, Log Avg loss: 0.11137640, Global Avg Loss: 0.54056859, Time: 0.0211 Steps: 104870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000559, Sample Num: 8944, Cur Loss: 0.44004530, Cur Avg Loss: 0.12354575, Log Avg loss: 0.12612292, Global Avg Loss: 0.54052908, Time: 0.0211 Steps: 104880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000569, Sample Num: 9104, Cur Loss: 0.18812883, Cur Avg Loss: 0.12301757, Log Avg loss: 0.09349223, Global Avg Loss: 0.54048646, Time: 0.0211 Steps: 104890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000579, Sample Num: 9264, Cur Loss: 0.19889297, Cur Avg Loss: 0.12336912, Log Avg loss: 0.14337197, Global Avg Loss: 0.54044860, Time: 0.0211 Steps: 104900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000589, Sample Num: 9424, Cur Loss: 0.09717166, Cur Avg Loss: 0.12339205, Log Avg loss: 0.12471999, Global Avg Loss: 0.54040897, Time: 0.0211 Steps: 104910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000599, Sample Num: 9584, Cur Loss: 0.44007984, Cur Avg Loss: 0.12458802, Log Avg loss: 0.19503050, Global Avg Loss: 0.54037606, Time: 0.0211 Steps: 104920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000609, Sample Num: 9744, Cur Loss: 0.02319489, Cur Avg Loss: 0.12520782, Log Avg loss: 0.16233376, Global Avg Loss: 0.54034003, Time: 0.0212 Steps: 104930, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000619, Sample Num: 9904, Cur Loss: 0.11767366, Cur Avg Loss: 0.12544304, Log Avg loss: 0.13976783, Global Avg Loss: 0.54030186, Time: 0.0211 Steps: 104940, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000629, Sample Num: 10064, Cur Loss: 0.09085661, Cur Avg Loss: 0.12542829, Log Avg loss: 0.12451581, Global Avg Loss: 0.54026224, Time: 0.0211 Steps: 104950, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000639, Sample Num: 10224, Cur Loss: 0.28809020, Cur Avg Loss: 0.12578141, Log Avg loss: 0.14799243, Global Avg Loss: 0.54022487, Time: 0.0212 Steps: 104960, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000649, Sample Num: 10384, Cur Loss: 0.03246707, Cur Avg Loss: 0.12614485, Log Avg loss: 0.14936848, Global Avg Loss: 0.54018763, Time: 0.0211 Steps: 104970, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000659, Sample Num: 10544, Cur Loss: 0.06689335, Cur Avg Loss: 0.12607843, Log Avg loss: 0.12176794, Global Avg Loss: 0.54014777, Time: 0.0211 Steps: 104980, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000669, Sample Num: 10704, Cur Loss: 0.07253776, Cur Avg Loss: 0.12616341, Log Avg loss: 0.13176381, Global Avg Loss: 0.54010888, Time: 0.0212 Steps: 104990, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000679, Sample Num: 10864, Cur Loss: 0.10822360, Cur Avg Loss: 0.12621865, Log Avg loss: 0.12991418, Global Avg Loss: 0.54006981, Time: 0.0211 Steps: 105000, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000689, Sample Num: 11024, Cur Loss: 0.17262274, Cur Avg Loss: 0.12666536, Log Avg loss: 0.15699683, Global Avg Loss: 0.54003333, Time: 0.0211 Steps: 105010, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000699, Sample Num: 11184, Cur Loss: 0.00775260, Cur Avg Loss: 0.12582480, Log Avg loss: 0.06790990, Global Avg Loss: 0.53998838, Time: 0.0210 Steps: 105020, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000709, Sample Num: 11344, Cur Loss: 0.04841816, Cur Avg Loss: 0.12564233, Log Avg loss: 0.11288815, Global Avg Loss: 0.53994771, Time: 0.0211 Steps: 105030, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000719, Sample Num: 11504, Cur Loss: 0.08427167, Cur Avg Loss: 0.12556592, Log Avg loss: 0.12014837, Global Avg Loss: 0.53990774, Time: 0.0212 Steps: 105040, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000729, Sample Num: 11664, Cur Loss: 0.11627705, Cur Avg Loss: 0.12557337, Log Avg loss: 0.12610860, Global Avg Loss: 0.53986835, Time: 0.0211 Steps: 105050, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000739, Sample Num: 11824, Cur Loss: 0.13350083, Cur Avg Loss: 0.12490495, Log Avg loss: 0.07617757, Global Avg Loss: 0.53982422, Time: 0.0211 Steps: 105060, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000749, Sample Num: 11984, Cur Loss: 0.26634246, Cur Avg Loss: 0.12471509, Log Avg loss: 0.11068388, Global Avg Loss: 0.53978338, Time: 0.0212 Steps: 105070, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000759, Sample Num: 12144, Cur Loss: 0.06178207, Cur Avg Loss: 0.12456073, Log Avg loss: 0.11299940, Global Avg Loss: 0.53974276, Time: 0.0211 Steps: 105080, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000769, Sample Num: 12304, Cur Loss: 0.03878459, Cur Avg Loss: 0.12396772, Log Avg loss: 0.07895861, Global Avg Loss: 0.53969891, Time: 0.0247 Steps: 105090, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000779, Sample Num: 12464, Cur Loss: 0.26647604, Cur Avg Loss: 0.12519611, Log Avg loss: 0.21965927, Global Avg Loss: 0.53966846, Time: 0.0211 Steps: 105100, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000789, Sample Num: 12624, Cur Loss: 0.16716199, Cur Avg Loss: 0.12530466, Log Avg loss: 0.13376057, Global Avg Loss: 0.53962984, Time: 0.0211 Steps: 105110, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000799, Sample Num: 12784, Cur Loss: 0.09840825, Cur Avg Loss: 0.12495560, Log Avg loss: 0.09741465, Global Avg Loss: 0.53958778, Time: 0.0211 Steps: 105120, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000809, Sample Num: 12944, Cur Loss: 0.09285609, Cur Avg Loss: 0.12487581, Log Avg loss: 0.11850040, Global Avg Loss: 0.53954772, Time: 0.0210 Steps: 105130, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000819, Sample Num: 13104, Cur Loss: 0.14254552, Cur Avg Loss: 0.12529707, Log Avg loss: 0.15937696, Global Avg Loss: 0.53951156, Time: 0.0210 Steps: 105140, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000829, Sample Num: 13264, Cur Loss: 0.07041489, Cur Avg Loss: 0.12544405, Log Avg loss: 0.13748197, Global Avg Loss: 0.53947333, Time: 0.0210 Steps: 105150, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000839, Sample Num: 13424, Cur Loss: 0.03017552, Cur Avg Loss: 0.12564344, Log Avg loss: 0.14217255, Global Avg Loss: 0.53943555, Time: 0.0210 Steps: 105160, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000849, Sample Num: 13584, Cur Loss: 0.07369724, Cur Avg Loss: 0.12544367, Log Avg loss: 0.10868315, Global Avg Loss: 0.53939459, Time: 0.0210 Steps: 105170, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000859, Sample Num: 13744, Cur Loss: 0.07051313, Cur Avg Loss: 0.12520843, Log Avg loss: 0.10523704, Global Avg Loss: 0.53935331, Time: 0.0210 Steps: 105180, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000869, Sample Num: 13904, Cur Loss: 0.16447288, Cur Avg Loss: 0.12506957, Log Avg loss: 0.11314144, Global Avg Loss: 0.53931280, Time: 0.0211 Steps: 105190, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000879, Sample Num: 14064, Cur Loss: 0.04722591, Cur Avg Loss: 0.12460218, Log Avg loss: 0.08398603, Global Avg Loss: 0.53926951, Time: 0.0210 Steps: 105200, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000889, Sample Num: 14224, Cur Loss: 0.05627171, Cur Avg Loss: 0.12524344, Log Avg loss: 0.18160974, Global Avg Loss: 0.53923552, Time: 0.0210 Steps: 105210, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000899, Sample Num: 14384, Cur Loss: 0.08510898, Cur Avg Loss: 0.12536869, Log Avg loss: 0.13650352, Global Avg Loss: 0.53919724, Time: 0.0210 Steps: 105220, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000909, Sample Num: 14544, Cur Loss: 0.38553980, Cur Avg Loss: 0.12578273, Log Avg loss: 0.16300513, Global Avg Loss: 0.53916150, Time: 0.0210 Steps: 105230, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000919, Sample Num: 14704, Cur Loss: 0.13838890, Cur Avg Loss: 0.12564873, Log Avg loss: 0.11346794, Global Avg Loss: 0.53912105, Time: 0.0210 Steps: 105240, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000929, Sample Num: 14864, Cur Loss: 0.13634638, Cur Avg Loss: 0.12572646, Log Avg loss: 0.13286984, Global Avg Loss: 0.53908245, Time: 0.0210 Steps: 105250, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000939, Sample Num: 15024, Cur Loss: 0.07604302, Cur Avg Loss: 0.12522618, Log Avg loss: 0.07875013, Global Avg Loss: 0.53903871, Time: 0.0211 Steps: 105260, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000949, Sample Num: 15184, Cur Loss: 0.08200225, Cur Avg Loss: 0.12513464, Log Avg loss: 0.11653948, Global Avg Loss: 0.53899858, Time: 0.0210 Steps: 105270, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000959, Sample Num: 15344, Cur Loss: 0.32034719, Cur Avg Loss: 0.12523307, Log Avg loss: 0.13457397, Global Avg Loss: 0.53896016, Time: 0.0210 Steps: 105280, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000969, Sample Num: 15504, Cur Loss: 0.11235789, Cur Avg Loss: 0.12521261, Log Avg loss: 0.12325040, Global Avg Loss: 0.53892068, Time: 0.0210 Steps: 105290, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000979, Sample Num: 15664, Cur Loss: 0.13548881, Cur Avg Loss: 0.12555885, Log Avg loss: 0.15910910, Global Avg Loss: 0.53888461, Time: 0.0210 Steps: 105300, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000989, Sample Num: 15824, Cur Loss: 0.18058726, Cur Avg Loss: 0.12571315, Log Avg loss: 0.14081900, Global Avg Loss: 0.53884681, Time: 0.0210 Steps: 105310, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000999, Sample Num: 15984, Cur Loss: 0.02848544, Cur Avg Loss: 0.12528023, Log Avg loss: 0.08246530, Global Avg Loss: 0.53880348, Time: 0.0211 Steps: 105320, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001009, Sample Num: 16144, Cur Loss: 0.11942533, Cur Avg Loss: 0.12507514, Log Avg loss: 0.10458588, Global Avg Loss: 0.53876226, Time: 0.0210 Steps: 105330, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001019, Sample Num: 16304, Cur Loss: 0.10485728, Cur Avg Loss: 0.12536278, Log Avg loss: 0.15438634, Global Avg Loss: 0.53872577, Time: 0.0210 Steps: 105340, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001029, Sample Num: 16464, Cur Loss: 0.03828265, Cur Avg Loss: 0.12522198, Log Avg loss: 0.11087423, Global Avg Loss: 0.53868515, Time: 0.0212 Steps: 105350, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001039, Sample Num: 16624, Cur Loss: 0.16712528, Cur Avg Loss: 0.12560757, Log Avg loss: 0.16528431, Global Avg Loss: 0.53864971, Time: 0.0210 Steps: 105360, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001049, Sample Num: 16784, Cur Loss: 0.04469575, Cur Avg Loss: 0.12528073, Log Avg loss: 0.09132246, Global Avg Loss: 0.53860726, Time: 0.0210 Steps: 105370, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001059, Sample Num: 16944, Cur Loss: 0.01800902, Cur Avg Loss: 0.12546487, Log Avg loss: 0.14478144, Global Avg Loss: 0.53856989, Time: 0.0211 Steps: 105380, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001069, Sample Num: 17104, Cur Loss: 0.24633323, Cur Avg Loss: 0.12584625, Log Avg loss: 0.16623413, Global Avg Loss: 0.53853456, Time: 0.0210 Steps: 105390, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001079, Sample Num: 17264, Cur Loss: 0.05256998, Cur Avg Loss: 0.12611720, Log Avg loss: 0.15508117, Global Avg Loss: 0.53849818, Time: 0.0210 Steps: 105400, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001089, Sample Num: 17424, Cur Loss: 0.02337148, Cur Avg Loss: 0.12565353, Log Avg loss: 0.07562363, Global Avg Loss: 0.53845427, Time: 0.0210 Steps: 105410, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001099, Sample Num: 17584, Cur Loss: 0.18503042, Cur Avg Loss: 0.12557507, Log Avg loss: 0.11703080, Global Avg Loss: 0.53841429, Time: 0.0210 Steps: 105420, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001109, Sample Num: 17744, Cur Loss: 0.11167235, Cur Avg Loss: 0.12547939, Log Avg loss: 0.11496471, Global Avg Loss: 0.53837413, Time: 0.0210 Steps: 105430, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001119, Sample Num: 17904, Cur Loss: 0.12376755, Cur Avg Loss: 0.12523201, Log Avg loss: 0.09779680, Global Avg Loss: 0.53833234, Time: 0.0210 Steps: 105440, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001129, Sample Num: 18064, Cur Loss: 0.11576563, Cur Avg Loss: 0.12558634, Log Avg loss: 0.16523608, Global Avg Loss: 0.53829696, Time: 0.0211 Steps: 105450, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001139, Sample Num: 18224, Cur Loss: 0.16221763, Cur Avg Loss: 0.12605556, Log Avg loss: 0.17903080, Global Avg Loss: 0.53826290, Time: 0.0210 Steps: 105460, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001149, Sample Num: 18384, Cur Loss: 0.03908768, Cur Avg Loss: 0.12586202, Log Avg loss: 0.10381808, Global Avg Loss: 0.53822170, Time: 0.0211 Steps: 105470, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001159, Sample Num: 18544, Cur Loss: 0.03665701, Cur Avg Loss: 0.12577451, Log Avg loss: 0.11571888, Global Avg Loss: 0.53818165, Time: 0.0211 Steps: 105480, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001169, Sample Num: 18704, Cur Loss: 0.17383842, Cur Avg Loss: 0.12534098, Log Avg loss: 0.07509501, Global Avg Loss: 0.53813775, Time: 0.0210 Steps: 105490, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001179, Sample Num: 18864, Cur Loss: 0.08555073, Cur Avg Loss: 0.12527289, Log Avg loss: 0.11731319, Global Avg Loss: 0.53809786, Time: 0.0211 Steps: 105500, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001189, Sample Num: 19024, Cur Loss: 0.21481267, Cur Avg Loss: 0.12518669, Log Avg loss: 0.11502389, Global Avg Loss: 0.53805776, Time: 0.0211 Steps: 105510, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001199, Sample Num: 19184, Cur Loss: 0.17831200, Cur Avg Loss: 0.12529166, Log Avg loss: 0.13777223, Global Avg Loss: 0.53801983, Time: 0.0210 Steps: 105520, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001209, Sample Num: 19344, Cur Loss: 0.14511764, Cur Avg Loss: 0.12571354, Log Avg loss: 0.17629747, Global Avg Loss: 0.53798555, Time: 0.0210 Steps: 105530, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001219, Sample Num: 19504, Cur Loss: 0.18240564, Cur Avg Loss: 0.12574469, Log Avg loss: 0.12951090, Global Avg Loss: 0.53794685, Time: 0.0210 Steps: 105540, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001229, Sample Num: 19664, Cur Loss: 0.12024786, Cur Avg Loss: 0.12598920, Log Avg loss: 0.15579418, Global Avg Loss: 0.53791064, Time: 0.0210 Steps: 105550, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001239, Sample Num: 19824, Cur Loss: 0.15083817, Cur Avg Loss: 0.12610898, Log Avg loss: 0.14083111, Global Avg Loss: 0.53787303, Time: 0.0211 Steps: 105560, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001249, Sample Num: 19984, Cur Loss: 0.17136666, Cur Avg Loss: 0.12602844, Log Avg loss: 0.11604892, Global Avg Loss: 0.53783307, Time: 0.0210 Steps: 105570, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001259, Sample Num: 20144, Cur Loss: 0.29554325, Cur Avg Loss: 0.12605573, Log Avg loss: 0.12946403, Global Avg Loss: 0.53779439, Time: 0.0210 Steps: 105580, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001269, Sample Num: 20304, Cur Loss: 0.13879447, Cur Avg Loss: 0.12657757, Log Avg loss: 0.19227752, Global Avg Loss: 0.53776167, Time: 0.0210 Steps: 105590, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001279, Sample Num: 20464, Cur Loss: 0.02644261, Cur Avg Loss: 0.12644333, Log Avg loss: 0.10940832, Global Avg Loss: 0.53772110, Time: 0.0210 Steps: 105600, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001289, Sample Num: 20624, Cur Loss: 0.15569663, Cur Avg Loss: 0.12619964, Log Avg loss: 0.09503101, Global Avg Loss: 0.53767919, Time: 0.0211 Steps: 105610, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001299, Sample Num: 20784, Cur Loss: 0.05000746, Cur Avg Loss: 0.12636756, Log Avg loss: 0.14801328, Global Avg Loss: 0.53764229, Time: 0.0210 Steps: 105620, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001309, Sample Num: 20944, Cur Loss: 0.09297011, Cur Avg Loss: 0.12605425, Log Avg loss: 0.08535547, Global Avg Loss: 0.53759948, Time: 0.0210 Steps: 105630, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001319, Sample Num: 21104, Cur Loss: 0.06522775, Cur Avg Loss: 0.12581447, Log Avg loss: 0.09442640, Global Avg Loss: 0.53755752, Time: 0.0210 Steps: 105640, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001329, Sample Num: 21264, Cur Loss: 0.09139918, Cur Avg Loss: 0.12574416, Log Avg loss: 0.11647121, Global Avg Loss: 0.53751767, Time: 0.0210 Steps: 105650, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001339, Sample Num: 21424, Cur Loss: 0.24977279, Cur Avg Loss: 0.12609993, Log Avg loss: 0.17338070, Global Avg Loss: 0.53748320, Time: 0.0210 Steps: 105660, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001349, Sample Num: 21584, Cur Loss: 0.02609103, Cur Avg Loss: 0.12599630, Log Avg loss: 0.11212143, Global Avg Loss: 0.53744295, Time: 0.0210 Steps: 105670, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001359, Sample Num: 21744, Cur Loss: 0.35546848, Cur Avg Loss: 0.12614168, Log Avg loss: 0.14575331, Global Avg Loss: 0.53740589, Time: 0.0210 Steps: 105680, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001369, Sample Num: 21904, Cur Loss: 0.20764074, Cur Avg Loss: 0.12585117, Log Avg loss: 0.08637064, Global Avg Loss: 0.53736321, Time: 0.0210 Steps: 105690, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001379, Sample Num: 22064, Cur Loss: 0.03536414, Cur Avg Loss: 0.12573556, Log Avg loss: 0.10990856, Global Avg Loss: 0.53732277, Time: 0.0210 Steps: 105700, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001389, Sample Num: 22224, Cur Loss: 0.15942208, Cur Avg Loss: 0.12583571, Log Avg loss: 0.13964547, Global Avg Loss: 0.53728515, Time: 0.0210 Steps: 105710, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001399, Sample Num: 22384, Cur Loss: 0.02638880, Cur Avg Loss: 0.12575888, Log Avg loss: 0.11508778, Global Avg Loss: 0.53724522, Time: 0.0210 Steps: 105720, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001409, Sample Num: 22544, Cur Loss: 0.30911881, Cur Avg Loss: 0.12578692, Log Avg loss: 0.12971047, Global Avg Loss: 0.53720667, Time: 0.0210 Steps: 105730, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001419, Sample Num: 22704, Cur Loss: 0.35672680, Cur Avg Loss: 0.12580335, Log Avg loss: 0.12811729, Global Avg Loss: 0.53716798, Time: 0.0210 Steps: 105740, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001429, Sample Num: 22864, Cur Loss: 0.13446930, Cur Avg Loss: 0.12571250, Log Avg loss: 0.11282153, Global Avg Loss: 0.53712786, Time: 0.0210 Steps: 105750, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001439, Sample Num: 23024, Cur Loss: 0.10092264, Cur Avg Loss: 0.12549321, Log Avg loss: 0.09415570, Global Avg Loss: 0.53708597, Time: 0.0210 Steps: 105760, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001449, Sample Num: 23184, Cur Loss: 0.15682513, Cur Avg Loss: 0.12557086, Log Avg loss: 0.13674521, Global Avg Loss: 0.53704812, Time: 0.0210 Steps: 105770, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001459, Sample Num: 23344, Cur Loss: 0.17292885, Cur Avg Loss: 0.12609115, Log Avg loss: 0.20148109, Global Avg Loss: 0.53701640, Time: 0.0210 Steps: 105780, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001469, Sample Num: 23504, Cur Loss: 0.02285773, Cur Avg Loss: 0.12612291, Log Avg loss: 0.13075743, Global Avg Loss: 0.53697800, Time: 0.0210 Steps: 105790, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001479, Sample Num: 23664, Cur Loss: 0.17929889, Cur Avg Loss: 0.12651203, Log Avg loss: 0.18367306, Global Avg Loss: 0.53694460, Time: 0.0210 Steps: 105800, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001489, Sample Num: 23824, Cur Loss: 0.13696805, Cur Avg Loss: 0.12652527, Log Avg loss: 0.12848429, Global Avg Loss: 0.53690600, Time: 0.0210 Steps: 105810, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001499, Sample Num: 23984, Cur Loss: 0.10556015, Cur Avg Loss: 0.12651610, Log Avg loss: 0.12515070, Global Avg Loss: 0.53686709, Time: 0.0210 Steps: 105820, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001509, Sample Num: 24144, Cur Loss: 0.04383689, Cur Avg Loss: 0.12656237, Log Avg loss: 0.13349786, Global Avg Loss: 0.53682897, Time: 0.0210 Steps: 105830, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001519, Sample Num: 24304, Cur Loss: 0.17361306, Cur Avg Loss: 0.12669027, Log Avg loss: 0.14598951, Global Avg Loss: 0.53679205, Time: 0.0210 Steps: 105840, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001529, Sample Num: 24464, Cur Loss: 0.02447177, Cur Avg Loss: 0.12687686, Log Avg loss: 0.15522077, Global Avg Loss: 0.53675600, Time: 0.0210 Steps: 105850, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001539, Sample Num: 24624, Cur Loss: 0.06964280, Cur Avg Loss: 0.12687307, Log Avg loss: 0.12629268, Global Avg Loss: 0.53671722, Time: 0.0247 Steps: 105860, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001549, Sample Num: 24784, Cur Loss: 0.12104057, Cur Avg Loss: 0.12697010, Log Avg loss: 0.14190312, Global Avg Loss: 0.53667993, Time: 0.0211 Steps: 105870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001559, Sample Num: 24944, Cur Loss: 0.21032250, Cur Avg Loss: 0.12702245, Log Avg loss: 0.13513275, Global Avg Loss: 0.53664201, Time: 0.0210 Steps: 105880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001569, Sample Num: 25104, Cur Loss: 0.15196139, Cur Avg Loss: 0.12695320, Log Avg loss: 0.11615633, Global Avg Loss: 0.53660230, Time: 0.0210 Steps: 105890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001579, Sample Num: 25264, Cur Loss: 0.12737551, Cur Avg Loss: 0.12696779, Log Avg loss: 0.12925749, Global Avg Loss: 0.53656383, Time: 0.0210 Steps: 105900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001589, Sample Num: 25424, Cur Loss: 0.03419119, Cur Avg Loss: 0.12682754, Log Avg loss: 0.10468114, Global Avg Loss: 0.53652305, Time: 0.0210 Steps: 105910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001599, Sample Num: 25584, Cur Loss: 0.06483284, Cur Avg Loss: 0.12706301, Log Avg loss: 0.16447899, Global Avg Loss: 0.53648793, Time: 0.0210 Steps: 105920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001609, Sample Num: 25744, Cur Loss: 0.05812568, Cur Avg Loss: 0.12684806, Log Avg loss: 0.09247864, Global Avg Loss: 0.53644601, Time: 0.0210 Steps: 105930, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001619, Sample Num: 25904, Cur Loss: 0.02923017, Cur Avg Loss: 0.12678745, Log Avg loss: 0.11703529, Global Avg Loss: 0.53640642, Time: 0.0210 Steps: 105940, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001629, Sample Num: 26064, Cur Loss: 0.09445328, Cur Avg Loss: 0.12662887, Log Avg loss: 0.10095469, Global Avg Loss: 0.53636532, Time: 0.0210 Steps: 105950, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001639, Sample Num: 26224, Cur Loss: 0.06108825, Cur Avg Loss: 0.12702345, Log Avg loss: 0.19129965, Global Avg Loss: 0.53633276, Time: 0.0211 Steps: 105960, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001649, Sample Num: 26384, Cur Loss: 0.09065954, Cur Avg Loss: 0.12689448, Log Avg loss: 0.10575738, Global Avg Loss: 0.53629213, Time: 0.0212 Steps: 105970, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001659, Sample Num: 26544, Cur Loss: 0.08938778, Cur Avg Loss: 0.12670150, Log Avg loss: 0.09487798, Global Avg Loss: 0.53625048, Time: 0.0210 Steps: 105980, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001669, Sample Num: 26704, Cur Loss: 0.14381717, Cur Avg Loss: 0.12661591, Log Avg loss: 0.11241673, Global Avg Loss: 0.53621049, Time: 0.0210 Steps: 105990, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001679, Sample Num: 26864, Cur Loss: 0.07826014, Cur Avg Loss: 0.12673605, Log Avg loss: 0.14678709, Global Avg Loss: 0.53617375, Time: 0.0210 Steps: 106000, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001689, Sample Num: 27024, Cur Loss: 0.15484211, Cur Avg Loss: 0.12651471, Log Avg loss: 0.08935254, Global Avg Loss: 0.53613160, Time: 0.0210 Steps: 106010, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001699, Sample Num: 27184, Cur Loss: 0.09401041, Cur Avg Loss: 0.12669231, Log Avg loss: 0.15668897, Global Avg Loss: 0.53609581, Time: 0.0210 Steps: 106020, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001709, Sample Num: 27344, Cur Loss: 0.03942212, Cur Avg Loss: 0.12649422, Log Avg loss: 0.09283783, Global Avg Loss: 0.53605401, Time: 0.0211 Steps: 106030, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001719, Sample Num: 27504, Cur Loss: 0.06521401, Cur Avg Loss: 0.12650974, Log Avg loss: 0.12916211, Global Avg Loss: 0.53601563, Time: 0.0210 Steps: 106040, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001729, Sample Num: 27664, Cur Loss: 0.03963622, Cur Avg Loss: 0.12657211, Log Avg loss: 0.13729504, Global Avg Loss: 0.53597804, Time: 0.0210 Steps: 106050, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001739, Sample Num: 27824, Cur Loss: 0.04639576, Cur Avg Loss: 0.12683510, Log Avg loss: 0.17230519, Global Avg Loss: 0.53594375, Time: 0.0210 Steps: 106060, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001749, Sample Num: 27984, Cur Loss: 0.09317572, Cur Avg Loss: 0.12660237, Log Avg loss: 0.08613070, Global Avg Loss: 0.53590134, Time: 0.0210 Steps: 106070, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001759, Sample Num: 28144, Cur Loss: 0.02897998, Cur Avg Loss: 0.12655450, Log Avg loss: 0.11818286, Global Avg Loss: 0.53586196, Time: 0.0210 Steps: 106080, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001769, Sample Num: 28304, Cur Loss: 0.26610881, Cur Avg Loss: 0.12690949, Log Avg loss: 0.18935110, Global Avg Loss: 0.53582930, Time: 0.0210 Steps: 106090, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001779, Sample Num: 28464, Cur Loss: 0.31101435, Cur Avg Loss: 0.12707266, Log Avg loss: 0.15593696, Global Avg Loss: 0.53579350, Time: 0.0210 Steps: 106100, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001789, Sample Num: 28624, Cur Loss: 0.14969823, Cur Avg Loss: 0.12715694, Log Avg loss: 0.14215055, Global Avg Loss: 0.53575640, Time: 0.0210 Steps: 106110, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001799, Sample Num: 28784, Cur Loss: 0.17015862, Cur Avg Loss: 0.12717637, Log Avg loss: 0.13065262, Global Avg Loss: 0.53571822, Time: 0.0210 Steps: 106120, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001809, Sample Num: 28944, Cur Loss: 0.05226730, Cur Avg Loss: 0.12703816, Log Avg loss: 0.10217517, Global Avg Loss: 0.53567737, Time: 0.0212 Steps: 106130, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001819, Sample Num: 29104, Cur Loss: 0.10205866, Cur Avg Loss: 0.12699834, Log Avg loss: 0.11979500, Global Avg Loss: 0.53563819, Time: 0.0210 Steps: 106140, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001829, Sample Num: 29264, Cur Loss: 0.02710492, Cur Avg Loss: 0.12670082, Log Avg loss: 0.07258047, Global Avg Loss: 0.53559457, Time: 0.0210 Steps: 106150, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001839, Sample Num: 29424, Cur Loss: 0.13286929, Cur Avg Loss: 0.12667752, Log Avg loss: 0.12241704, Global Avg Loss: 0.53555565, Time: 0.0210 Steps: 106160, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001849, Sample Num: 29584, Cur Loss: 0.10883924, Cur Avg Loss: 0.12668961, Log Avg loss: 0.12891209, Global Avg Loss: 0.53551735, Time: 0.0211 Steps: 106170, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001859, Sample Num: 29744, Cur Loss: 0.11195384, Cur Avg Loss: 0.12686824, Log Avg loss: 0.15989777, Global Avg Loss: 0.53548197, Time: 0.0211 Steps: 106180, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001869, Sample Num: 29904, Cur Loss: 0.02811490, Cur Avg Loss: 0.12682253, Log Avg loss: 0.11832506, Global Avg Loss: 0.53544269, Time: 0.0211 Steps: 106190, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001879, Sample Num: 30064, Cur Loss: 0.16992447, Cur Avg Loss: 0.12688354, Log Avg loss: 0.13828545, Global Avg Loss: 0.53540529, Time: 0.0211 Steps: 106200, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001889, Sample Num: 30224, Cur Loss: 0.04333822, Cur Avg Loss: 0.12740668, Log Avg loss: 0.22570521, Global Avg Loss: 0.53537613, Time: 0.0211 Steps: 106210, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001899, Sample Num: 30384, Cur Loss: 0.17952852, Cur Avg Loss: 0.12740907, Log Avg loss: 0.12786051, Global Avg Loss: 0.53533777, Time: 0.0210 Steps: 106220, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001909, Sample Num: 30544, Cur Loss: 0.04920274, Cur Avg Loss: 0.12739306, Log Avg loss: 0.12435284, Global Avg Loss: 0.53529908, Time: 0.0210 Steps: 106230, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001919, Sample Num: 30704, Cur Loss: 0.16425727, Cur Avg Loss: 0.12752157, Log Avg loss: 0.15205468, Global Avg Loss: 0.53526300, Time: 0.0210 Steps: 106240, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001929, Sample Num: 30864, Cur Loss: 0.17169352, Cur Avg Loss: 0.12754725, Log Avg loss: 0.13247539, Global Avg Loss: 0.53522509, Time: 0.0210 Steps: 106250, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001939, Sample Num: 31024, Cur Loss: 0.07411332, Cur Avg Loss: 0.12757941, Log Avg loss: 0.13378144, Global Avg Loss: 0.53518732, Time: 0.0210 Steps: 106260, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001949, Sample Num: 31184, Cur Loss: 0.22022027, Cur Avg Loss: 0.12758153, Log Avg loss: 0.12799335, Global Avg Loss: 0.53514900, Time: 0.0211 Steps: 106270, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001959, Sample Num: 31344, Cur Loss: 0.05935992, Cur Avg Loss: 0.12769354, Log Avg loss: 0.14952356, Global Avg Loss: 0.53511271, Time: 0.0211 Steps: 106280, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001969, Sample Num: 31504, Cur Loss: 0.16505775, Cur Avg Loss: 0.12766161, Log Avg loss: 0.12140737, Global Avg Loss: 0.53507379, Time: 0.0210 Steps: 106290, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001979, Sample Num: 31664, Cur Loss: 0.30864203, Cur Avg Loss: 0.12763535, Log Avg loss: 0.12246553, Global Avg Loss: 0.53503498, Time: 0.0211 Steps: 106300, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001989, Sample Num: 31824, Cur Loss: 0.08266293, Cur Avg Loss: 0.12755288, Log Avg loss: 0.11123142, Global Avg Loss: 0.53499511, Time: 0.0210 Steps: 106310, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001999, Sample Num: 31984, Cur Loss: 0.18793446, Cur Avg Loss: 0.12757864, Log Avg loss: 0.13270188, Global Avg Loss: 0.53495727, Time: 0.0211 Steps: 106320, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002009, Sample Num: 32144, Cur Loss: 0.06172808, Cur Avg Loss: 0.12779867, Log Avg loss: 0.17178211, Global Avg Loss: 0.53492312, Time: 0.0211 Steps: 106330, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002019, Sample Num: 32304, Cur Loss: 0.32110870, Cur Avg Loss: 0.12773565, Log Avg loss: 0.11507489, Global Avg Loss: 0.53488364, Time: 0.0210 Steps: 106340, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002029, Sample Num: 32464, Cur Loss: 0.21839045, Cur Avg Loss: 0.12818186, Log Avg loss: 0.21827281, Global Avg Loss: 0.53485387, Time: 0.0211 Steps: 106350, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002039, Sample Num: 32624, Cur Loss: 0.13425040, Cur Avg Loss: 0.12845254, Log Avg loss: 0.18337304, Global Avg Loss: 0.53482082, Time: 0.0210 Steps: 106360, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002049, Sample Num: 32784, Cur Loss: 0.08817829, Cur Avg Loss: 0.12876620, Log Avg loss: 0.19272167, Global Avg Loss: 0.53478866, Time: 0.0248 Steps: 106370, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002059, Sample Num: 32944, Cur Loss: 0.05018264, Cur Avg Loss: 0.12871144, Log Avg loss: 0.11749213, Global Avg Loss: 0.53474943, Time: 0.0211 Steps: 106380, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002069, Sample Num: 33104, Cur Loss: 0.02894182, Cur Avg Loss: 0.12852090, Log Avg loss: 0.08928698, Global Avg Loss: 0.53470756, Time: 0.0211 Steps: 106390, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002079, Sample Num: 33264, Cur Loss: 0.07308393, Cur Avg Loss: 0.12862008, Log Avg loss: 0.14914111, Global Avg Loss: 0.53467132, Time: 0.0211 Steps: 106400, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002089, Sample Num: 33424, Cur Loss: 0.03946493, Cur Avg Loss: 0.12846120, Log Avg loss: 0.09543022, Global Avg Loss: 0.53463004, Time: 0.0211 Steps: 106410, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002099, Sample Num: 33584, Cur Loss: 0.10076293, Cur Avg Loss: 0.12848591, Log Avg loss: 0.13364865, Global Avg Loss: 0.53459237, Time: 0.0211 Steps: 106420, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002109, Sample Num: 33744, Cur Loss: 0.07787394, Cur Avg Loss: 0.12827319, Log Avg loss: 0.08362248, Global Avg Loss: 0.53454999, Time: 0.0211 Steps: 106430, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002119, Sample Num: 33904, Cur Loss: 0.09158507, Cur Avg Loss: 0.12831777, Log Avg loss: 0.13771855, Global Avg Loss: 0.53451271, Time: 0.0211 Steps: 106440, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002129, Sample Num: 34055, Cur Loss: 0.22113577, Cur Avg Loss: 0.12846203, Log Avg loss: 0.15903072, Global Avg Loss: 0.53447744, Time: 0.0102 Steps: 106450, Updated lr: 0.000000 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 44.998014, Avg time per batch (s): 0.020000 {"eval_avg_loss": 0.249143, "eval_total_loss": 175.14732, "eval_mae": 0.32941, "eval_mse": 0.24924, "eval_r2": 0.841567, "eval_sp_statistic": 0.932775, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.937114, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.174149, "test_total_loss": 87.422953, "test_mae": 0.293061, "test_mse": 0.174213, "test_r2": 0.887562, "test_sp_statistic": 0.922217, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.947612, "test_ps_pvalue": 0.0, "lr": 0.0, "cur_epoch_step": 2129, "train_global_avg_loss": 0.5344774380367959, "train_cur_epoch_loss": 273.49565207073465, "train_cur_epoch_avg_loss": 0.12846202539724502, "train_cur_epoch_time": 44.998013973236084, "train_cur_epoch_avg_time": 0.021135751044263074, "epoch": 50, "step": 106450} ################################################## #########################Best Metric######################### {"epoch": 49, "global_step": 104321, "eval_avg_loss": 0.249419, "eval_total_loss": 175.341598, "eval_mae": 0.330862, "eval_mse": 0.249516, "eval_r2": 0.841391, "eval_sp_statistic": 0.93284, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.937126, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.174699, "test_total_loss": 87.698664, "test_mae": 0.294155, "test_mse": 0.174762, "test_r2": 0.887207, "test_sp_statistic": 0.921977, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.94748, "test_ps_pvalue": 0.0} ################################################## Total Time: 38849.600471, Avg time per epoch(50 epochs): 776.990000 ++++++++++++Validation+++++++++++++ best sp_statistic global step: 104321 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250225170110/checkpoint-104321 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## {"evaluation_avg_loss_104321": 0.249419, "evaluation_total_loss_104321": 175.341598, "evaluation_mae_104321": 0.330862, "evaluation_mse_104321": 0.249516, "evaluation_r2_104321": 0.841391, "evaluation_sp_statistic_104321": 0.93284, "evaluation_sp_pvalue_104321": 0.0, "evaluation_ps_statistic_104321": 0.937126, "evaluation_ps_pvalue_104321": 0.0} ++++++++++++Testing+++++++++++++ best sp_statistic global step: 104321 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250225170110/checkpoint-104321 ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## {"evaluation_avg_loss_104321": 0.174699, "evaluation_total_loss_104321": 87.698664, "evaluation_mae_104321": 0.294155, "evaluation_mse_104321": 0.174762, "evaluation_r2_104321": 0.887207, "evaluation_sp_statistic_104321": 0.921977, "evaluation_sp_pvalue_104321": 0.0, "evaluation_ps_statistic_104321": 0.94748, "evaluation_ps_pvalue_104321": 0.0}